/home/fresvfqn/waterdamagerestorationandrepairsmithtown.com/Compressed/python3.9.tar
site-packages/_distutils_hack/__pycache__/override.cpython-39.pyc000064400000000434150564745010021125 0ustar00a

(�Re,�@sed���dS)Z_distutils_hackN)�
__import__Zdo_override�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/_distutils_hack/override.py�<module>�site-packages/_distutils_hack/__pycache__/__init__.cpython-39.pyc000064400000012041150564745010021042 0ustar00a

(�Reh�@s�ddlZddlZddlZddlZddlZdejvZe�dde�dd�Z	dd�Z
d	d
�Zdd�Zd
d�Z
Gdd�d�Ze�Zdd�Zdd�ZdS)�NZ__pypy__�ignorez.+ distutils\b.+ deprecatedcCs.dtjvrdStr tjdkr dSt�d�dS)N�	distutils)��a�Distutils was imported before Setuptools, but importing Setuptools also replaces the `distutils` module in `sys.modules`. This may lead to undesirable behaviors or errors. To avoid these issues, avoid using distutils directly, ensure that setuptools is installed in the traditional way (e.g. not an editable install), and/or make sure that setuptools is always imported before distutils.)�sys�modules�is_pypy�version_info�warnings�warn�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/_distutils_hack/__init__.py�warn_distutils_presents
�rcCs>dtjvrdSt�d�dd�tjD�}|D]}tj|=q,dS)Nrz"Setuptools is replacing distutils.cSsg|]}t�d|�r|�qS)zdistutils\b)�re�match)�.0�namerrr
�
<listcomp>$�z#clear_distutils.<locals>.<listcomp>)rrr
r)Zmodsrrrr
�clear_distutils s

rcCstj�dd�}|dkS)z?
    Allow selection of distutils by environment variable.
    ZSETUPTOOLS_USE_DISTUTILS�stdlib�local)�os�environ�get)�whichrrr
�enabled)srcCsBt�t�d�}d|_|tjd<t�d�}d|jvs>J|j��dS)N�setuptools._distutilsrzdistutils.core�
_distutils)r�	importlib�
import_module�__name__rr�__file__)r�corerrr
�ensure_local_distutils1s


r$cCst�rt�t�dS)z�
    Ensure that the local copy of distutils is preferred over stdlib.

    See https://github.com/pypa/setuptools/issues/417#issuecomment-392298401
    for more motivation.
    N)rrr$rrrr
�do_override<sr%c@s2eZdZd
dd�Zdd�Zdd�Zedd	��ZdS)�DistutilsMetaFinderNcCs4|durdSdjfit���}t||dd��}|�S)Nzspec_for_{fullname}cSsdS�Nrrrrr
�<lambda>Nrz/DistutilsMetaFinder.find_spec.<locals>.<lambda>)�format�locals�getattr)�self�fullname�path�target�method_name�methodrrr
�	find_specIs
zDistutilsMetaFinder.find_speccs8ddl�ddl�G�fdd�d�jj�}�j�d|��S)Nrcs eZdZ�fdd�Zdd�ZdS)z?DistutilsMetaFinder.spec_for_distutils.<locals>.DistutilsLoadercs
��d�S)Nr)r )r,�spec�rrr
�
create_moduleWszMDistutilsMetaFinder.spec_for_distutils.<locals>.DistutilsLoader.create_modulecSsdSr'r)r,�modulerrr
�exec_moduleZszKDistutilsMetaFinder.spec_for_distutils.<locals>.DistutilsLoader.exec_moduleN)r!�
__module__�__qualname__r5r7rr4rr
�DistutilsLoaderUsr:r)�
importlib.abc�importlib.util�abc�Loader�util�spec_from_loader)r,r:rr4r
�spec_for_distutilsQsz&DistutilsMetaFinder.spec_for_distutilscCs |��rdSt�dd�|_dS)zj
        Ensure stdlib distutils when running under pip.
        See pypa/pip#8761 for rationale.
        NcSsdSr'rrrrr
r(grz2DistutilsMetaFinder.spec_for_pip.<locals>.<lambda>)�pip_imported_during_buildrrA)r,rrr
�spec_for_pip_sz DistutilsMetaFinder.spec_for_pipcCs ddl}tdd�|�d�D��S)zO
        Detect if pip is being imported in a build script. Ref #2355.
        rNcss"|]\}}|jd�d�VqdS)r"zsetup.pyN)�	f_globals�endswith)r�frame�linerrr
�	<genexpr>os�z@DistutilsMetaFinder.pip_imported_during_build.<locals>.<genexpr>)�	traceback�any�
walk_stack)rIrrr
rBis�z-DistutilsMetaFinder.pip_imported_during_build)N)r!r8r9r2rArC�staticmethodrBrrrr
r&Hs


r&cCstj�dt�dS)Nr)r�	meta_path�insert�DISTUTILS_FINDERrrrr
�add_shimxsrPcCs(ztj�t�Wnty"Yn0dSr')rrM�removerO�
ValueErrorrrrr
�remove_shim|srS)rrrrr
�builtin_module_namesr�filterwarnings�DeprecationWarningrrrr$r%r&rOrPrSrrrr
�<module>s$
�	-site-packages/_distutils_hack/override.py000064400000000054150564745010014634 0ustar00__import__('_distutils_hack').do_override()
site-packages/_distutils_hack/__init__.py000064400000007150150564745010014560 0ustar00import sys
import os
import re
import importlib
import warnings


is_pypy = '__pypy__' in sys.builtin_module_names


warnings.filterwarnings('ignore',
                        r'.+ distutils\b.+ deprecated',
                        DeprecationWarning)


def warn_distutils_present():
    if 'distutils' not in sys.modules:
        return
    if is_pypy and sys.version_info < (3, 7):
        # PyPy for 3.6 unconditionally imports distutils, so bypass the warning
        # https://foss.heptapod.net/pypy/pypy/-/blob/be829135bc0d758997b3566062999ee8b23872b4/lib-python/3/site.py#L250
        return
    warnings.warn(
        "Distutils was imported before Setuptools, but importing Setuptools "
        "also replaces the `distutils` module in `sys.modules`. This may lead "
        "to undesirable behaviors or errors. To avoid these issues, avoid "
        "using distutils directly, ensure that setuptools is installed in the "
        "traditional way (e.g. not an editable install), and/or make sure "
        "that setuptools is always imported before distutils.")


def clear_distutils():
    if 'distutils' not in sys.modules:
        return
    warnings.warn("Setuptools is replacing distutils.")
    mods = [name for name in sys.modules if re.match(r'distutils\b', name)]
    for name in mods:
        del sys.modules[name]


def enabled():
    """
    Allow selection of distutils by environment variable.
    """
    which = os.environ.get('SETUPTOOLS_USE_DISTUTILS', 'stdlib')
    return which == 'local'


def ensure_local_distutils():
    clear_distutils()
    distutils = importlib.import_module('setuptools._distutils')
    distutils.__name__ = 'distutils'
    sys.modules['distutils'] = distutils

    # sanity check that submodules load as expected
    core = importlib.import_module('distutils.core')
    assert '_distutils' in core.__file__, core.__file__


def do_override():
    """
    Ensure that the local copy of distutils is preferred over stdlib.

    See https://github.com/pypa/setuptools/issues/417#issuecomment-392298401
    for more motivation.
    """
    if enabled():
        warn_distutils_present()
        ensure_local_distutils()


class DistutilsMetaFinder:
    def find_spec(self, fullname, path, target=None):
        if path is not None:
            return

        method_name = 'spec_for_{fullname}'.format(**locals())
        method = getattr(self, method_name, lambda: None)
        return method()

    def spec_for_distutils(self):
        import importlib.abc
        import importlib.util

        class DistutilsLoader(importlib.abc.Loader):

            def create_module(self, spec):
                return importlib.import_module('setuptools._distutils')

            def exec_module(self, module):
                pass

        return importlib.util.spec_from_loader('distutils', DistutilsLoader())

    def spec_for_pip(self):
        """
        Ensure stdlib distutils when running under pip.
        See pypa/pip#8761 for rationale.
        """
        if self.pip_imported_during_build():
            return
        clear_distutils()
        self.spec_for_distutils = lambda: None

    @staticmethod
    def pip_imported_during_build():
        """
        Detect if pip is being imported in a build script. Ref #2355.
        """
        import traceback
        return any(
            frame.f_globals['__file__'].endswith('setup.py')
            for frame, line in traceback.walk_stack(None)
        )


DISTUTILS_FINDER = DistutilsMetaFinder()


def add_shim():
    sys.meta_path.insert(0, DISTUTILS_FINDER)


def remove_shim():
    try:
        sys.meta_path.remove(DISTUTILS_FINDER)
    except ValueError:
        pass
site-packages/pip-21.3.1.dist-info/WHEEL000064400000000134150564745010013263 0ustar00Wheel-Version: 1.0
Generator: bdist_wheel (0.37.0)
Root-Is-Purelib: true
Tag: py3-none-any

site-packages/pip-21.3.1.dist-info/direct_url.json000064400000000147150564745010015527 0ustar00{"archive_info": {}, "url": "file:///builddir/build/BUILD/pip-21.3.1/dist/pip-21.3.1-py3-none-any.whl"}site-packages/pip-21.3.1.dist-info/RECORD000064400000173112150564745010013404 0ustar00../../../bin/pip,sha256=KlJDQt1hwlg3F2zaPoYxPfFpRr9irBB68fDPoKVECkY,234
../../../bin/pip3,sha256=KlJDQt1hwlg3F2zaPoYxPfFpRr9irBB68fDPoKVECkY,234
../../../bin/pip3.9,sha256=KlJDQt1hwlg3F2zaPoYxPfFpRr9irBB68fDPoKVECkY,234
pip-21.3.1.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
pip-21.3.1.dist-info/LICENSE.txt,sha256=I6c2HCsVgQKLxiO52ivSSZeryqR4Gs5q1ESjeUT42uE,1090
pip-21.3.1.dist-info/METADATA,sha256=PjWcvFEqJd4gOfiQam8il34_wPNKxf8ubyYI2wYm7tc,4216
pip-21.3.1.dist-info/RECORD,,
pip-21.3.1.dist-info/REQUESTED,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip-21.3.1.dist-info/WHEEL,sha256=ewwEueio1C2XeHTvT17n8dZUJgOvyCWCt0WVNLClP9o,92
pip-21.3.1.dist-info/direct_url.json,sha256=Vl58sluk6OBfGWPA_3cYakPqWvzPs77FuaR2htQJdG0,103
pip-21.3.1.dist-info/entry_points.txt,sha256=5ExSa1s54zSPNA_1epJn5SX06786S8k5YHwskMvVYzw,125
pip-21.3.1.dist-info/top_level.txt,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
pip/__init__.py,sha256=798yhPIf6eMHi7R5Ogb3BJ5ALJ0Id8IwEuOSU2DFlp0,357
pip/__main__.py,sha256=mXwWDftNLMKfwVqKFWGE_uuBZvGSIiUELhLkeysIuZc,1198
pip/__pycache__/__init__.cpython-39.pyc,,
pip/__pycache__/__main__.cpython-39.pyc,,
pip/_internal/__init__.py,sha256=FU88-ODK1jGlVF2Qxp1gJqTwg0wujj12LayXQgFhlh8,587
pip/_internal/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/__pycache__/build_env.cpython-39.pyc,,
pip/_internal/__pycache__/cache.cpython-39.pyc,,
pip/_internal/__pycache__/configuration.cpython-39.pyc,,
pip/_internal/__pycache__/exceptions.cpython-39.pyc,,
pip/_internal/__pycache__/main.cpython-39.pyc,,
pip/_internal/__pycache__/pyproject.cpython-39.pyc,,
pip/_internal/__pycache__/self_outdated_check.cpython-39.pyc,,
pip/_internal/__pycache__/wheel_builder.cpython-39.pyc,,
pip/_internal/build_env.py,sha256=uIg4HJDgZK542FXVTl3jkPDNbklNgb8Rj6DeZef_oS8,9950
pip/_internal/cache.py,sha256=71eaYwrls34HJ6gzbmmYiotiKhPNFTM_tqYJXD5nf3s,9441
pip/_internal/cli/__init__.py,sha256=FkHBgpxxb-_gd6r1FjnNhfMOzAUYyXoXKJ6abijfcFU,132
pip/_internal/cli/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/cli/__pycache__/autocompletion.cpython-39.pyc,,
pip/_internal/cli/__pycache__/base_command.cpython-39.pyc,,
pip/_internal/cli/__pycache__/cmdoptions.cpython-39.pyc,,
pip/_internal/cli/__pycache__/command_context.cpython-39.pyc,,
pip/_internal/cli/__pycache__/main.cpython-39.pyc,,
pip/_internal/cli/__pycache__/main_parser.cpython-39.pyc,,
pip/_internal/cli/__pycache__/parser.cpython-39.pyc,,
pip/_internal/cli/__pycache__/progress_bars.cpython-39.pyc,,
pip/_internal/cli/__pycache__/req_command.cpython-39.pyc,,
pip/_internal/cli/__pycache__/spinners.cpython-39.pyc,,
pip/_internal/cli/__pycache__/status_codes.cpython-39.pyc,,
pip/_internal/cli/autocompletion.py,sha256=NK5yqe49SgExZOCFVEUT5Bf0QV2CuITGK27WSo2MWg8,6399
pip/_internal/cli/base_command.py,sha256=oFuvjLsYE17V67L1dHeTo-YePZN97RKpOuGEXwCKwLc,7790
pip/_internal/cli/cmdoptions.py,sha256=o6hueHSc3VWZ-_do9eeoZKEaxqh18zlXKAzVZ00Kg-o,28391
pip/_internal/cli/command_context.py,sha256=a1pBBvvGLDiZ1Kw64_4tT6HmRTwYDoYy8JFgG5Czn7s,760
pip/_internal/cli/main.py,sha256=ioJ8IVlb2K1qLOxR-tXkee9lURhYV89CDM71MKag7YY,2472
pip/_internal/cli/main_parser.py,sha256=Q9TnytfuC5Z2JSjBFWVGtEdYLFy7rukNIb04movHdAo,2614
pip/_internal/cli/parser.py,sha256=CDXTuFr2UD8ozOlZYf1KDziQdo9-X_IaYOiUcyJQwrA,10788
pip/_internal/cli/progress_bars.py,sha256=ha8wowclY8_PaoM0cz4G6qK37zjnzuxQ-ydOtzx4EMI,8300
pip/_internal/cli/req_command.py,sha256=La6J8YonTxoPtJ8HMPN4RTKyzg0VS_R4vxfVf_HmFZw,17097
pip/_internal/cli/spinners.py,sha256=TFhjxtOnLeNJ5YmRvQm4eKPgPbJNkZiqO8jOXuxRaYU,5076
pip/_internal/cli/status_codes.py,sha256=sEFHUaUJbqv8iArL3HAtcztWZmGOFX01hTesSytDEh0,116
pip/_internal/commands/__init__.py,sha256=Vc1HjsLEtyCh7506OozPHPKXe2Hk-z9cFkFF3BMj1lM,3736
pip/_internal/commands/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/commands/__pycache__/cache.cpython-39.pyc,,
pip/_internal/commands/__pycache__/check.cpython-39.pyc,,
pip/_internal/commands/__pycache__/completion.cpython-39.pyc,,
pip/_internal/commands/__pycache__/configuration.cpython-39.pyc,,
pip/_internal/commands/__pycache__/debug.cpython-39.pyc,,
pip/_internal/commands/__pycache__/download.cpython-39.pyc,,
pip/_internal/commands/__pycache__/freeze.cpython-39.pyc,,
pip/_internal/commands/__pycache__/hash.cpython-39.pyc,,
pip/_internal/commands/__pycache__/help.cpython-39.pyc,,
pip/_internal/commands/__pycache__/index.cpython-39.pyc,,
pip/_internal/commands/__pycache__/install.cpython-39.pyc,,
pip/_internal/commands/__pycache__/list.cpython-39.pyc,,
pip/_internal/commands/__pycache__/search.cpython-39.pyc,,
pip/_internal/commands/__pycache__/show.cpython-39.pyc,,
pip/_internal/commands/__pycache__/uninstall.cpython-39.pyc,,
pip/_internal/commands/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/commands/cache.py,sha256=p9gvc6W_xgxE2zO0o8NXqO1gGJEinEK42qEC-a7Cnuk,7524
pip/_internal/commands/check.py,sha256=0gjXR7j36xJT5cs2heYU_dfOfpnFfzX8OoPNNoKhqdM,1685
pip/_internal/commands/completion.py,sha256=kTG_I1VR3N5kGC4Ma9pQTSoY9Q1URCrNyseHSQ-rCL4,2958
pip/_internal/commands/configuration.py,sha256=arE8vLstjBg-Ar1krXF-bBmT1qBtnL7Fpk-NVh38a0U,8944
pip/_internal/commands/debug.py,sha256=krET-y45CnQzXwKR1qA3M_tJE4LE2vnQtm3yfGyDSnE,6629
pip/_internal/commands/download.py,sha256=p4lmYDgawRrwDFUpde_-1Gld45FnsMNHUFtOWFUCcSE,4904
pip/_internal/commands/freeze.py,sha256=gCjoD6foBZPBAAYx5t8zZLkJhsF_ZRtnb3dPuD7beO8,2951
pip/_internal/commands/hash.py,sha256=EVVOuvGtoPEdFi8SNnmdqlCQrhCxV-kJsdwtdcCnXGQ,1703
pip/_internal/commands/help.py,sha256=gcc6QDkcgHMOuAn5UxaZwAStsRBrnGSn_yxjS57JIoM,1132
pip/_internal/commands/index.py,sha256=1VVXXj5MsI2qH-N7uniQQyVkg-KCn_RdjiyiUmkUS5U,4762
pip/_internal/commands/install.py,sha256=HTWdTb72Bcrm2tA_d55_hX6yQbchnr_XRdA2Xs8uApU,27851
pip/_internal/commands/list.py,sha256=SnCh19e5zQKonNP7j25c_xru0Wm7wWWF8j49f-Dy9Bw,12203
pip/_internal/commands/search.py,sha256=sbBZiARRc050QquOKcCvOr2K3XLsoYebLKZGRi__iUI,5697
pip/_internal/commands/show.py,sha256=OREbPHF6UzvQiGLC1UIjG52Kc_jYDgcXZMYzgKXMbBI,8064
pip/_internal/commands/uninstall.py,sha256=DNTYAGJNljMO_YYBxrpcwj0FEl7lo_P55_98O6g2TNk,3526
pip/_internal/commands/wheel.py,sha256=xGSwLPYUM7jP_McD-wnM4D3zsP0n-NSkHFp4d0mAWIg,6168
pip/_internal/configuration.py,sha256=dKHBEl8aXnqVuRB0NW7Nz7lyYMwr7XCfkMZvUORaSRo,13153
pip/_internal/distributions/__init__.py,sha256=Hq6kt6gXBgjNit5hTTWLAzeCNOKoB-N0pGYSqehrli8,858
pip/_internal/distributions/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/distributions/__pycache__/base.cpython-39.pyc,,
pip/_internal/distributions/__pycache__/installed.cpython-39.pyc,,
pip/_internal/distributions/__pycache__/sdist.cpython-39.pyc,,
pip/_internal/distributions/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/distributions/base.py,sha256=3FUYD8Gb4YuSu3pggC_FRctZBDbpm5ZK89tPksIUjoE,1172
pip/_internal/distributions/installed.py,sha256=QObf6KALGtwGx-Ap3Ua5FfcfaRMXWOk_wcrm7n5gYII,767
pip/_internal/distributions/sdist.py,sha256=3fsErGhAWdGzuO7Wea0F_8b9fKyUL1PoYet273OoAoM,5598
pip/_internal/distributions/wheel.py,sha256=-NgzdIs-w_hcer_U81yzgpVTljJRg5m79xufqvbjv0s,1115
pip/_internal/exceptions.py,sha256=XyfiRZn2X8WR61X-JF50BU72TdmVkneWPy9cnuKv2Rg,12762
pip/_internal/index/__init__.py,sha256=vpt-JeTZefh8a-FC22ZeBSXFVbuBcXSGiILhQZJaNpQ,30
pip/_internal/index/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/index/__pycache__/collector.cpython-39.pyc,,
pip/_internal/index/__pycache__/package_finder.cpython-39.pyc,,
pip/_internal/index/__pycache__/sources.cpython-39.pyc,,
pip/_internal/index/collector.py,sha256=7rhUeH0IU_dUMk13-lBAN9czRuJ6dbG76Un7xuQ36Ck,17534
pip/_internal/index/package_finder.py,sha256=_N9LIcwAXbGDN3BUDlikSB93WI9PHv3MvkJ4YapfrPY,36344
pip/_internal/index/sources.py,sha256=SVyPitv08-Qalh2_Bk5diAJ9GAA_d-a93koouQodAG0,6557
pip/_internal/locations/__init__.py,sha256=CpH6Cz9HSZ0csN_KPtOcvS9TGYLb7ZNGtCAAmVtjXW0,14444
pip/_internal/locations/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/locations/__pycache__/_distutils.cpython-39.pyc,,
pip/_internal/locations/__pycache__/_sysconfig.cpython-39.pyc,,
pip/_internal/locations/__pycache__/base.cpython-39.pyc,,
pip/_internal/locations/_distutils.py,sha256=Sk7tw8ZP1DWMYJ8MibABsa8IME2Ejv1PKeGlYQCBTZc,5871
pip/_internal/locations/_sysconfig.py,sha256=LQNKTJKyjVqxXaPntlBwdUqTG1xwYf6GVCKMbyRJx5M,7918
pip/_internal/locations/base.py,sha256=x5D1ONktmPJd8nnUTh-ELsAJ7fiXA-k-0a_vhfi2_Us,1579
pip/_internal/main.py,sha256=r-UnUe8HLo5XFJz8inTcOOTiu_sxNhgHb6VwlGUllOI,340
pip/_internal/metadata/__init__.py,sha256=HzTS3lRukzn-MJaEZkUQhAFe6ulxvNe7nNoBvUzy-DU,1660
pip/_internal/metadata/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/metadata/__pycache__/base.cpython-39.pyc,,
pip/_internal/metadata/__pycache__/pkg_resources.cpython-39.pyc,,
pip/_internal/metadata/base.py,sha256=gbNbb9blWO5hejmror-2n4_wLuYVrTyqwUluY9OmnMg,11103
pip/_internal/metadata/pkg_resources.py,sha256=-LiuojtAfl3yhNx8rnUKYN3ECBVCVcDWszCupithXAw,5089
pip/_internal/models/__init__.py,sha256=3DHUd_qxpPozfzouoqa9g9ts1Czr5qaHfFxbnxriepM,63
pip/_internal/models/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/models/__pycache__/candidate.cpython-39.pyc,,
pip/_internal/models/__pycache__/direct_url.cpython-39.pyc,,
pip/_internal/models/__pycache__/format_control.cpython-39.pyc,,
pip/_internal/models/__pycache__/index.cpython-39.pyc,,
pip/_internal/models/__pycache__/link.cpython-39.pyc,,
pip/_internal/models/__pycache__/scheme.cpython-39.pyc,,
pip/_internal/models/__pycache__/search_scope.cpython-39.pyc,,
pip/_internal/models/__pycache__/selection_prefs.cpython-39.pyc,,
pip/_internal/models/__pycache__/target_python.cpython-39.pyc,,
pip/_internal/models/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/models/candidate.py,sha256=6pcABsaR7CfIHlbJbr2_kMkVJFL_yrYjTx6SVWUnCPQ,990
pip/_internal/models/direct_url.py,sha256=7XtGQSLLDQb5ZywI2EMnnLcddtf5CJLx44lMtTHPxFw,6350
pip/_internal/models/format_control.py,sha256=DJpMYjxeYKKQdwNcML2_F0vtAh-qnKTYe-CpTxQe-4g,2520
pip/_internal/models/index.py,sha256=tYnL8oxGi4aSNWur0mG8DAP7rC6yuha_MwJO8xw0crI,1030
pip/_internal/models/link.py,sha256=hoT_qsOBAgLBm9GKqpBrNF_mrEXeGXQE-aH_RX2cGgg,9817
pip/_internal/models/scheme.py,sha256=3EFQp_ICu_shH1-TBqhl0QAusKCPDFOlgHFeN4XowWs,738
pip/_internal/models/search_scope.py,sha256=LwloG0PJAmtI1hFXIypsD95kWE9xfR5hf_a2v1Vw7sk,4520
pip/_internal/models/selection_prefs.py,sha256=KZdi66gsR-_RUXUr9uejssk3rmTHrQVJWeNA2sV-VSY,1907
pip/_internal/models/target_python.py,sha256=qKpZox7J8NAaPmDs5C_aniwfPDxzvpkrCKqfwndG87k,3858
pip/_internal/models/wheel.py,sha256=hN9Ub-m-cAJCajCcQHyQNsqpcDCbPPDlEzBDwaBMc14,3500
pip/_internal/network/__init__.py,sha256=jf6Tt5nV_7zkARBrKojIXItgejvoegVJVKUbhAa5Ioc,50
pip/_internal/network/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/network/__pycache__/auth.cpython-39.pyc,,
pip/_internal/network/__pycache__/cache.cpython-39.pyc,,
pip/_internal/network/__pycache__/download.cpython-39.pyc,,
pip/_internal/network/__pycache__/lazy_wheel.cpython-39.pyc,,
pip/_internal/network/__pycache__/session.cpython-39.pyc,,
pip/_internal/network/__pycache__/utils.cpython-39.pyc,,
pip/_internal/network/__pycache__/xmlrpc.cpython-39.pyc,,
pip/_internal/network/auth.py,sha256=a3C7Xaa8kTJjXkdi_wrUjqaySc8Z9Yz7U6QIbXfzMyc,12190
pip/_internal/network/cache.py,sha256=HoprMCecwd4IS2wDZowc9B_OpaBlFjJYJl4xOxvtuwU,2100
pip/_internal/network/download.py,sha256=VmiR-KKIBugShZS4JlD7N8mq3hErx-0fK-D8aTYU3Og,6016
pip/_internal/network/lazy_wheel.py,sha256=1b8ZJ1w4bSBzpGzGwJR_CL2yQ6AFIwWQkS1vbPPw2XU,7627
pip/_internal/network/session.py,sha256=38IKGKC64MTVUIH5XOR1hr2pOCzp39RccykdmGAvqRU,16729
pip/_internal/network/utils.py,sha256=igLlTu_-q0LmL8FdJKq-Uj7AT_owrQ-T9FfyarkhK5U,4059
pip/_internal/network/xmlrpc.py,sha256=AzQgG4GgS152_cqmGr_Oz2MIXsCal-xfsis7fA7nmU0,1791
pip/_internal/operations/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/operations/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/operations/__pycache__/check.cpython-39.pyc,,
pip/_internal/operations/__pycache__/freeze.cpython-39.pyc,,
pip/_internal/operations/__pycache__/prepare.cpython-39.pyc,,
pip/_internal/operations/build/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/operations/build/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/metadata.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/metadata_editable.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/metadata_legacy.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/wheel_editable.cpython-39.pyc,,
pip/_internal/operations/build/__pycache__/wheel_legacy.cpython-39.pyc,,
pip/_internal/operations/build/metadata.py,sha256=KEsyrRFOBs2jhR-AcjyJyeV5GlsK1ubQqAB1j-b0Zu4,1119
pip/_internal/operations/build/metadata_editable.py,sha256=RnA8UgQqZwtBjBdqi1DW1gI3xaZ7qhKp1Xd-0YTktSk,1177
pip/_internal/operations/build/metadata_legacy.py,sha256=hjAJ75iKuJfKQYALZD0U6wJ7ElJ_BAEvjDxF8b9_l5k,1945
pip/_internal/operations/build/wheel.py,sha256=AO9XnTGhTgHtZmU8Dkbfo1OGr41rBuSDjIgAa4zUKgE,1063
pip/_internal/operations/build/wheel_editable.py,sha256=TVETY-L_M_dSEKBhTIcQOP75zKVXw8tuq1U354Mm30A,1405
pip/_internal/operations/build/wheel_legacy.py,sha256=aFMVOvyG-_CAIuXEVxuPJkz5UfCppSeu9FBPzn2tWvI,3047
pip/_internal/operations/check.py,sha256=ca4O9CkPt9Em9sLCf3H0iVt1GIcW7M8C0U5XooaBuT4,5109
pip/_internal/operations/freeze.py,sha256=ZiYw5GlUpLVx4VJHz4S1AP2JFNyvH0iq5kpcYj2ovyw,9770
pip/_internal/operations/install/__init__.py,sha256=mX7hyD2GNBO2mFGokDQ30r_GXv7Y_PLdtxcUv144e-s,51
pip/_internal/operations/install/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/operations/install/__pycache__/editable_legacy.cpython-39.pyc,,
pip/_internal/operations/install/__pycache__/legacy.cpython-39.pyc,,
pip/_internal/operations/install/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/operations/install/editable_legacy.py,sha256=J4VCOHvk_BgA_wG02WmlDtSWLwZJ5S_g9SXBkjYojaw,1298
pip/_internal/operations/install/legacy.py,sha256=YKrZvH894Iqf2oEkYqF9O7CK1DjTgfZCP3R9Azpjeqo,4158
pip/_internal/operations/install/wheel.py,sha256=QuQyCZE-XjuJjDYRixo40oUt2ucFhNmSrCbcXY7A9aE,27412
pip/_internal/operations/prepare.py,sha256=Dg-lFYsFhYeib8NuQvGOxd0wxcmTqXfe_c5zYb3ep64,23838
pip/_internal/pyproject.py,sha256=YgcyleTgyuh7NwGH9j8_21htqnF_VxgKiPc4ecLBWKk,7215
pip/_internal/req/__init__.py,sha256=A7mUvT1KAcCYP3H7gUOTx2GRMlgoDur3H68Q0OJqM5A,2793
pip/_internal/req/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/req/__pycache__/constructors.cpython-39.pyc,,
pip/_internal/req/__pycache__/req_file.cpython-39.pyc,,
pip/_internal/req/__pycache__/req_install.cpython-39.pyc,,
pip/_internal/req/__pycache__/req_set.cpython-39.pyc,,
pip/_internal/req/__pycache__/req_tracker.cpython-39.pyc,,
pip/_internal/req/__pycache__/req_uninstall.cpython-39.pyc,,
pip/_internal/req/constructors.py,sha256=FVWkWeGt3fK0DTC3Gurd2jglp_Z10CK-abd6yM3HD-A,15285
pip/_internal/req/req_file.py,sha256=5N8OTouPCof-305StC2YK9HBxQMw-xO46skRoBPbkZo,17421
pip/_internal/req/req_install.py,sha256=BLPPMm9obLThHs1Fp8mvaO02jHRrEZ6De4oAU_QDkx4,33866
pip/_internal/req/req_set.py,sha256=kHYiLvkKRx21WaLTwOI-54Ng0SSzZZ9SE7FD0PsfvYA,7584
pip/_internal/req/req_tracker.py,sha256=jK7JDu-Wt73X-gqozrFtgJVlUlnQo0P4IQ4x4_gPlfM,4117
pip/_internal/req/req_uninstall.py,sha256=Uf8Kx-PgoQIudFq9Y7sFP-uz_I6x1gEfPpJJxujOf14,23748
pip/_internal/resolution/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/resolution/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/resolution/__pycache__/base.cpython-39.pyc,,
pip/_internal/resolution/base.py,sha256=qlmh325SBVfvG6Me9gc5Nsh5sdwHBwzHBq6aEXtKsLA,583
pip/_internal/resolution/legacy/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/resolution/legacy/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/resolution/legacy/__pycache__/resolver.cpython-39.pyc,,
pip/_internal/resolution/legacy/resolver.py,sha256=IcQ57uIyki-hCBm4e5H8zZt840VzfxdWQFZNT9oYBH8,18449
pip/_internal/resolution/resolvelib/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/resolution/resolvelib/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/base.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/candidates.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/factory.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/found_candidates.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/provider.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/reporter.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/requirements.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/__pycache__/resolver.cpython-39.pyc,,
pip/_internal/resolution/resolvelib/base.py,sha256=u1O4fkvCO4mhmu5i32xrDv9AX5NgUci_eYVyBDQhTIM,5220
pip/_internal/resolution/resolvelib/candidates.py,sha256=5q66J90AoMKKwy1HsdXvEeleOJG8QkAbo8OidFekee0,18210
pip/_internal/resolution/resolvelib/factory.py,sha256=fa97Ja0EbOYkyz5haXTUPbnR3H5MOLvouY6n8OjupSs,27391
pip/_internal/resolution/resolvelib/found_candidates.py,sha256=hvL3Hoa9VaYo-qEOZkBi2Iqw251UDxPz-uMHVaWmLpE,5705
pip/_internal/resolution/resolvelib/provider.py,sha256=HUMHvkU001rtlqvs11NPmMtlyMMLlVQfAl6qXdsLxZQ,9205
pip/_internal/resolution/resolvelib/reporter.py,sha256=3ZVVYrs5PqvLFJkGLcuXoMK5mTInFzl31xjUpDBpZZk,2526
pip/_internal/resolution/resolvelib/requirements.py,sha256=pcsnwz7txyDNZUEOWJOZEfivy3COWHPf_DIU7fwZ-Kk,5455
pip/_internal/resolution/resolvelib/resolver.py,sha256=bkrMZs_jJHP_KFAbg36-lcN4Ums7ESgllup8piHXOz0,9580
pip/_internal/self_outdated_check.py,sha256=nVLSc0nl4JZ9VI7GsZvblE-zzT-T5ofmMgplned8s_s,6393
pip/_internal/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_internal/utils/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/utils/__pycache__/_log.cpython-39.pyc,,
pip/_internal/utils/__pycache__/appdirs.cpython-39.pyc,,
pip/_internal/utils/__pycache__/compat.cpython-39.pyc,,
pip/_internal/utils/__pycache__/compatibility_tags.cpython-39.pyc,,
pip/_internal/utils/__pycache__/datetime.cpython-39.pyc,,
pip/_internal/utils/__pycache__/deprecation.cpython-39.pyc,,
pip/_internal/utils/__pycache__/direct_url_helpers.cpython-39.pyc,,
pip/_internal/utils/__pycache__/distutils_args.cpython-39.pyc,,
pip/_internal/utils/__pycache__/egg_link.cpython-39.pyc,,
pip/_internal/utils/__pycache__/encoding.cpython-39.pyc,,
pip/_internal/utils/__pycache__/entrypoints.cpython-39.pyc,,
pip/_internal/utils/__pycache__/filesystem.cpython-39.pyc,,
pip/_internal/utils/__pycache__/filetypes.cpython-39.pyc,,
pip/_internal/utils/__pycache__/glibc.cpython-39.pyc,,
pip/_internal/utils/__pycache__/hashes.cpython-39.pyc,,
pip/_internal/utils/__pycache__/inject_securetransport.cpython-39.pyc,,
pip/_internal/utils/__pycache__/logging.cpython-39.pyc,,
pip/_internal/utils/__pycache__/misc.cpython-39.pyc,,
pip/_internal/utils/__pycache__/models.cpython-39.pyc,,
pip/_internal/utils/__pycache__/packaging.cpython-39.pyc,,
pip/_internal/utils/__pycache__/parallel.cpython-39.pyc,,
pip/_internal/utils/__pycache__/pkg_resources.cpython-39.pyc,,
pip/_internal/utils/__pycache__/setuptools_build.cpython-39.pyc,,
pip/_internal/utils/__pycache__/subprocess.cpython-39.pyc,,
pip/_internal/utils/__pycache__/temp_dir.cpython-39.pyc,,
pip/_internal/utils/__pycache__/unpacking.cpython-39.pyc,,
pip/_internal/utils/__pycache__/urls.cpython-39.pyc,,
pip/_internal/utils/__pycache__/virtualenv.cpython-39.pyc,,
pip/_internal/utils/__pycache__/wheel.cpython-39.pyc,,
pip/_internal/utils/_log.py,sha256=-jHLOE_THaZz5BFcCnoSL9EYAtJ0nXem49s9of4jvKw,1015
pip/_internal/utils/appdirs.py,sha256=swgcTKOm3daLeXTW6v5BUS2Ti2RvEnGRQYH_yDXklAo,1665
pip/_internal/utils/compat.py,sha256=ACyBfLgj3_XG-iA5omEDrXqDM0cQKzi8h8HRBInzG6Q,1884
pip/_internal/utils/compatibility_tags.py,sha256=ydin8QG8BHqYRsPY4OL6cmb44CbqXl1T0xxS97VhHkk,5377
pip/_internal/utils/datetime.py,sha256=m21Y3wAtQc-ji6Veb6k_M5g6A0ZyFI4egchTdnwh-pQ,242
pip/_internal/utils/deprecation.py,sha256=NKo8VqLioJ4nnXXGmW4KdasxF90EFHkZaHeX1fT08C8,3627
pip/_internal/utils/direct_url_helpers.py,sha256=6F1tc2rcKaCZmgfVwsE6ObIe_Pux23mUVYA-2D9wCFc,3206
pip/_internal/utils/distutils_args.py,sha256=mcAscyp80vTt3xAGTipnpgc83V-_wCvydNELVXLq7JI,1249
pip/_internal/utils/egg_link.py,sha256=5MVlpz5LirT4iLQq86OYzjXaYF0D4Qk1dprEI7ThST4,2203
pip/_internal/utils/encoding.py,sha256=bdZ3YgUpaOEBI5MP4-DEXiQarCW3V0rxw1kRz-TaU1Q,1169
pip/_internal/utils/entrypoints.py,sha256=g4QKi37JkZ2r6A9WbYLMwVzOiE3dn_lsqVhaFy0Erq0,1130
pip/_internal/utils/filesystem.py,sha256=rrl-rY1w8TYyKYndUyZlE9ffkQyA4-jI9x_59zXkn5s,5893
pip/_internal/utils/filetypes.py,sha256=i8XAQ0eFCog26Fw9yV0Yb1ygAqKYB1w9Cz9n0fj8gZU,716
pip/_internal/utils/glibc.py,sha256=tDfwVYnJCOC0BNVpItpy8CGLP9BjkxFHdl0mTS0J7fc,3110
pip/_internal/utils/hashes.py,sha256=anpZfFGIT6HcIj2td9NHtE8AWg6GeAIhwpP8GPvZE0E,4811
pip/_internal/utils/inject_securetransport.py,sha256=o-QRVMGiENrTJxw3fAhA7uxpdEdw6M41TjHYtSVRrcg,795
pip/_internal/utils/logging.py,sha256=oEkBvjj2A6NtVo75_Q-sL7qqH0bMFuY0pK4d8t40SKg,11532
pip/_internal/utils/misc.py,sha256=BmZJygU2e8NSUD-Iz4A1zzIL52GDhikAwCeZik7K7V4,20778
pip/_internal/utils/models.py,sha256=5GoYU586SrxURMvDn_jBMJInitviJg4O5-iOU-6I0WY,1193
pip/_internal/utils/packaging.py,sha256=wA29RPW_KkorI2PIfkm9cWCytpcVbk-wubwUE8YTmbQ,2952
pip/_internal/utils/parallel.py,sha256=Z-vNgYsyiAx8JfZYbD6ZSzkkPfpk0ANQI_YpCBE0Pxo,3196
pip/_internal/utils/pkg_resources.py,sha256=A7HUm5lSk7n1_7qypyI4QkXErXgb5iXDlKPXo8r_1Hk,987
pip/_internal/utils/setuptools_build.py,sha256=yDrfmxUgd0A9SDKV-7UuSTA3YLmVav5J86G9Fym-2FE,4697
pip/_internal/utils/subprocess.py,sha256=cy2c6XRuYkX3XJF_lIjY5nQL2XygBHLJr6WXwTsjfnc,10058
pip/_internal/utils/temp_dir.py,sha256=zob3PYMVevONkheOMUp_4jDofrEY3HIu5DHK78cSspI,7662
pip/_internal/utils/unpacking.py,sha256=HUFlMEyCa9dPwdLh6sWeh95DeKytV8rsOyKShEw9y6g,8906
pip/_internal/utils/urls.py,sha256=AhaesUGl-9it6uvG6fsFPOr9ynFpGaTMk4t5XTX7Z_Q,1759
pip/_internal/utils/virtualenv.py,sha256=4_48qMzCwB_F5jIK5BC_ua7uiAMVifmQWU9NdaGUoVA,3459
pip/_internal/utils/wheel.py,sha256=YwsLfuDzPJhFLuGotZ69i0bxJVGSweGuIHG2SxZvZtM,6163
pip/_internal/vcs/__init__.py,sha256=UAqvzpbi0VbZo3Ub6skEeZAw-ooIZR-zX_WpCbxyCoU,596
pip/_internal/vcs/__pycache__/__init__.cpython-39.pyc,,
pip/_internal/vcs/__pycache__/bazaar.cpython-39.pyc,,
pip/_internal/vcs/__pycache__/git.cpython-39.pyc,,
pip/_internal/vcs/__pycache__/mercurial.cpython-39.pyc,,
pip/_internal/vcs/__pycache__/subversion.cpython-39.pyc,,
pip/_internal/vcs/__pycache__/versioncontrol.cpython-39.pyc,,
pip/_internal/vcs/bazaar.py,sha256=pNMHrCLx1jSJzu1t1ycDVwhXQ23XI4Q483cvewaTUDs,2857
pip/_internal/vcs/git.py,sha256=Ph_hThbfTG040GpJRz1z0ByiNkj5eHgF_shCCbNnCw0,17804
pip/_internal/vcs/mercurial.py,sha256=Mtk-Bqjnp3wlaOdHfNSxq86vgCwNc3-df6UqgIXvMjE,4945
pip/_internal/vcs/subversion.py,sha256=h4_nYmYN9kcfeTPp9wjkHhIeTpFZwoCp1UVm4hbBq90,11596
pip/_internal/vcs/versioncontrol.py,sha256=W1zLW32PeuYiCV1I_dhqlk_n74B_GFTjNC5xdxs-1Ek,22414
pip/_internal/wheel_builder.py,sha256=ZakEA7CEJyp70yHoX0QLE8TAwM7vxF9PYPtjBxT3F1I,12247
pip/_vendor/__init__.py,sha256=xjcBX0EP50pkaMdCssrsBXoZgo2hTtYxlcH1CIyA3T4,4708
pip/_vendor/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/__pycache__/distro.cpython-39.pyc,,
pip/_vendor/__pycache__/pyparsing.cpython-39.pyc,,
pip/_vendor/__pycache__/six.cpython-39.pyc,,
pip/_vendor/cachecontrol/__init__.py,sha256=pJtAaUxOsMPnytI1A3juAJkXYDr8krdSnsg4Yg3OBEg,302
pip/_vendor/cachecontrol/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/_cmd.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/adapter.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/cache.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/controller.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/filewrapper.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/heuristics.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/serialize.cpython-39.pyc,,
pip/_vendor/cachecontrol/__pycache__/wrapper.cpython-39.pyc,,
pip/_vendor/cachecontrol/_cmd.py,sha256=URGE0KrA87QekCG3SGPatlSPT571dZTDjNa-ZXX3pDc,1295
pip/_vendor/cachecontrol/adapter.py,sha256=sSwaSYd93IIfCFU4tOMgSo6b2LCt_gBSaQUj8ktJFOA,4882
pip/_vendor/cachecontrol/cache.py,sha256=1fc4wJP8HYt1ycnJXeEw5pCpeBL2Cqxx6g9Fb0AYDWQ,805
pip/_vendor/cachecontrol/caches/__init__.py,sha256=-gHNKYvaeD0kOk5M74eOrsSgIKUtC6i6GfbmugGweEo,86
pip/_vendor/cachecontrol/caches/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/cachecontrol/caches/__pycache__/file_cache.cpython-39.pyc,,
pip/_vendor/cachecontrol/caches/__pycache__/redis_cache.cpython-39.pyc,,
pip/_vendor/cachecontrol/caches/file_cache.py,sha256=nYVKsJtXh6gJXvdn1iWyrhxvkwpQrK-eKoMRzuiwkKk,4153
pip/_vendor/cachecontrol/caches/redis_cache.py,sha256=HxelMpNCo-dYr2fiJDwM3hhhRmxUYtB5tXm1GpAAT4Y,856
pip/_vendor/cachecontrol/compat.py,sha256=kHNvMRdt6s_Xwqq_9qJmr9ou3wYMOMUMxPPcwNxT8Mc,695
pip/_vendor/cachecontrol/controller.py,sha256=CWEX3pedIM9s60suf4zZPtm_JvVgnvogMGK_OiBG5F8,14149
pip/_vendor/cachecontrol/filewrapper.py,sha256=vACKO8Llzu_ZWyjV1Fxn1MA4TGU60N5N3GSrAFdAY2Q,2533
pip/_vendor/cachecontrol/heuristics.py,sha256=BFGHJ3yQcxvZizfo90LLZ04T_Z5XSCXvFotrp7Us0sc,4070
pip/_vendor/cachecontrol/serialize.py,sha256=vIa4jvq4x_KSOLdEIedoknX2aXYHQujLDFV4-F21Dno,7091
pip/_vendor/cachecontrol/wrapper.py,sha256=5LX0uJwkNQUtYSEw3aGmGu9WY8wGipd81mJ8lG0d0M4,690
pip/_vendor/certifi/__init__.py,sha256=-b78tXibbl0qtgCzv9tc9v6ozwcNX915lT9Tf4a9lds,62
pip/_vendor/certifi/__main__.py,sha256=1k3Cr95vCxxGRGDljrW3wMdpZdL3Nhf0u1n-k2qdsCY,255
pip/_vendor/certifi/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/certifi/__pycache__/__main__.cpython-39.pyc,,
pip/_vendor/certifi/__pycache__/core.cpython-39.pyc,,
pip/_vendor/certifi/core.py,sha256=xcXTRTzkcnYHuGFEjzUFUvg9MOnYkD1gjqhsWOUiB0Y,2846
pip/_vendor/chardet/__init__.py,sha256=mWZaWmvZkhwfBEAT9O1Y6nRTfKzhT7FHhQTTAujbqUA,3271
pip/_vendor/chardet/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/big5freq.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/big5prober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/chardistribution.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/charsetgroupprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/charsetprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/codingstatemachine.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/cp949prober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/enums.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/escprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/escsm.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/eucjpprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/euckrfreq.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/euckrprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/euctwfreq.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/euctwprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/gb2312freq.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/gb2312prober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/hebrewprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/jisfreq.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/jpcntx.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langbulgarianmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langgreekmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langhebrewmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langhungarianmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langrussianmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langthaimodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/langturkishmodel.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/latin1prober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/mbcharsetprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/mbcsgroupprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/mbcssm.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/sbcharsetprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/sbcsgroupprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/sjisprober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/universaldetector.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/utf8prober.cpython-39.pyc,,
pip/_vendor/chardet/__pycache__/version.cpython-39.pyc,,
pip/_vendor/chardet/big5freq.py,sha256=D_zK5GyzoVsRes0HkLJziltFQX0bKCLOrFe9_xDvO_8,31254
pip/_vendor/chardet/big5prober.py,sha256=kBxHbdetBpPe7xrlb-e990iot64g_eGSLd32lB7_h3M,1757
pip/_vendor/chardet/chardistribution.py,sha256=3woWS62KrGooKyqz4zQSnjFbJpa6V7g02daAibTwcl8,9411
pip/_vendor/chardet/charsetgroupprober.py,sha256=GZLReHP6FRRn43hvSOoGCxYamErKzyp6RgOQxVeC3kg,3839
pip/_vendor/chardet/charsetprober.py,sha256=KSmwJErjypyj0bRZmC5F5eM7c8YQgLYIjZXintZNstg,5110
pip/_vendor/chardet/cli/__init__.py,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
pip/_vendor/chardet/cli/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/chardet/cli/__pycache__/chardetect.cpython-39.pyc,,
pip/_vendor/chardet/cli/chardetect.py,sha256=XK5zqjUG2a4-y6eLHZ8ThYcp6WWUrdlmELxNypcc2SE,2747
pip/_vendor/chardet/codingstatemachine.py,sha256=VYp_6cyyki5sHgXDSZnXW4q1oelHc3cu9AyQTX7uug8,3590
pip/_vendor/chardet/compat.py,sha256=40zr6wICZwknxyuLGGcIOPyve8DTebBCbbvttvnmp5Q,1200
pip/_vendor/chardet/cp949prober.py,sha256=TZ434QX8zzBsnUvL_8wm4AQVTZ2ZkqEEQL_lNw9f9ow,1855
pip/_vendor/chardet/enums.py,sha256=Aimwdb9as1dJKZaFNUH2OhWIVBVd6ZkJJ_WK5sNY8cU,1661
pip/_vendor/chardet/escprober.py,sha256=kkyqVg1Yw3DIOAMJ2bdlyQgUFQhuHAW8dUGskToNWSc,3950
pip/_vendor/chardet/escsm.py,sha256=RuXlgNvTIDarndvllNCk5WZBIpdCxQ0kcd9EAuxUh84,10510
pip/_vendor/chardet/eucjpprober.py,sha256=iD8Jdp0ISRjgjiVN7f0e8xGeQJ5GM2oeZ1dA8nbSeUw,3749
pip/_vendor/chardet/euckrfreq.py,sha256=-7GdmvgWez4-eO4SuXpa7tBiDi5vRXQ8WvdFAzVaSfo,13546
pip/_vendor/chardet/euckrprober.py,sha256=MqFMTQXxW4HbzIpZ9lKDHB3GN8SP4yiHenTmf8g_PxY,1748
pip/_vendor/chardet/euctwfreq.py,sha256=No1WyduFOgB5VITUA7PLyC5oJRNzRyMbBxaKI1l16MA,31621
pip/_vendor/chardet/euctwprober.py,sha256=13p6EP4yRaxqnP4iHtxHOJ6R2zxHq1_m8hTRjzVZ95c,1747
pip/_vendor/chardet/gb2312freq.py,sha256=JX8lsweKLmnCwmk8UHEQsLgkr_rP_kEbvivC4qPOrlc,20715
pip/_vendor/chardet/gb2312prober.py,sha256=gGvIWi9WhDjE-xQXHvNIyrnLvEbMAYgyUSZ65HUfylw,1754
pip/_vendor/chardet/hebrewprober.py,sha256=c3SZ-K7hvyzGY6JRAZxJgwJ_sUS9k0WYkvMY00YBYFo,13838
pip/_vendor/chardet/jisfreq.py,sha256=vpmJv2Bu0J8gnMVRPHMFefTRvo_ha1mryLig8CBwgOg,25777
pip/_vendor/chardet/jpcntx.py,sha256=PYlNqRUQT8LM3cT5FmHGP0iiscFlTWED92MALvBungo,19643
pip/_vendor/chardet/langbulgarianmodel.py,sha256=Ys1irPhpi61ywJpv6a-6JQOLwn7Ed7YqTBpGZ7adhOk,105675
pip/_vendor/chardet/langgreekmodel.py,sha256=en5Np_mAnOAR1Akoh0562eSe7phGWuq7H2_KoEv4eQU,99549
pip/_vendor/chardet/langhebrewmodel.py,sha256=NHkt2JO49NbzkgPzgYgrGIKvpO040WOjUMEPRzkZVDo,98754
pip/_vendor/chardet/langhungarianmodel.py,sha256=IA_31i27hFJwCRhHQ_0CdUIeBr_qndYmj2SJqDdw1n4,102476
pip/_vendor/chardet/langrussianmodel.py,sha256=QzrKRvQInhNxnMkYTugNqEfvVFZl_WI1bj9rFLdjYgQ,131158
pip/_vendor/chardet/langthaimodel.py,sha256=Ju2fDTy26CdiZYkioD8I5ZzKEGe8Et-49zon1_OOZI4,103290
pip/_vendor/chardet/langturkishmodel.py,sha256=ycEZsKGFiLl-FgoZLiUmWeS8F7F2ts3RrNdgRzArDlU,95924
pip/_vendor/chardet/latin1prober.py,sha256=S2IoORhFk39FEFOlSFWtgVybRiP6h7BlLldHVclNkU8,5370
pip/_vendor/chardet/mbcharsetprober.py,sha256=AR95eFH9vuqSfvLQZN-L5ijea25NOBCoXqw8s5O9xLQ,3413
pip/_vendor/chardet/mbcsgroupprober.py,sha256=h6TRnnYq2OxG1WdD5JOyxcdVpn7dG0q-vB8nWr5mbh4,2012
pip/_vendor/chardet/mbcssm.py,sha256=SY32wVIF3HzcjY3BaEspy9metbNSKxIIB0RKPn7tjpI,25481
pip/_vendor/chardet/metadata/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/chardet/metadata/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/chardet/metadata/__pycache__/languages.cpython-39.pyc,,
pip/_vendor/chardet/metadata/languages.py,sha256=Bn9KbnR_sctMbPDMw9IwxOyYeOGuE4tcVNjvE2FOEmA,19452
pip/_vendor/chardet/sbcharsetprober.py,sha256=nmyMyuxzG87DN6K3Rk2MUzJLMLR69MrWpdnHzOwVUwQ,6136
pip/_vendor/chardet/sbcsgroupprober.py,sha256=hqefQuXmiFyDBArOjujH6hd6WFXlOD1kWCsxDhjx5Vc,4309
pip/_vendor/chardet/sjisprober.py,sha256=IIt-lZj0WJqK4rmUZzKZP4GJlE8KUEtFYVuY96ek5MQ,3774
pip/_vendor/chardet/universaldetector.py,sha256=DpZTXCX0nUHXxkQ9sr4GZxGB_hveZ6hWt3uM94cgWKs,12503
pip/_vendor/chardet/utf8prober.py,sha256=IdD8v3zWOsB8OLiyPi-y_fqwipRFxV9Nc1eKBLSuIEw,2766
pip/_vendor/chardet/version.py,sha256=A4CILFAd8MRVG1HoXPp45iK9RLlWyV73a1EtwE8Tvn8,242
pip/_vendor/colorama/__init__.py,sha256=pCdErryzLSzDW5P-rRPBlPLqbBtIRNJB6cMgoeJns5k,239
pip/_vendor/colorama/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/colorama/__pycache__/ansi.cpython-39.pyc,,
pip/_vendor/colorama/__pycache__/ansitowin32.cpython-39.pyc,,
pip/_vendor/colorama/__pycache__/initialise.cpython-39.pyc,,
pip/_vendor/colorama/__pycache__/win32.cpython-39.pyc,,
pip/_vendor/colorama/__pycache__/winterm.cpython-39.pyc,,
pip/_vendor/colorama/ansi.py,sha256=Top4EeEuaQdBWdteKMEcGOTeKeF19Q-Wo_6_Cj5kOzQ,2522
pip/_vendor/colorama/ansitowin32.py,sha256=yV7CEmCb19MjnJKODZEEvMH_fnbJhwnpzo4sxZuGXmA,10517
pip/_vendor/colorama/initialise.py,sha256=PprovDNxMTrvoNHFcL2NZjpH2XzDc8BLxLxiErfUl4k,1915
pip/_vendor/colorama/win32.py,sha256=bJ8Il9jwaBN5BJ8bmN6FoYZ1QYuMKv2j8fGrXh7TJjw,5404
pip/_vendor/colorama/winterm.py,sha256=2y_2b7Zsv34feAsP67mLOVc-Bgq51mdYGo571VprlrM,6438
pip/_vendor/distlib/__init__.py,sha256=HTGLP7dnTRTQCbEZNGUxBq-0sobr0KQUMn3yd6uEObA,581
pip/_vendor/distlib/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/database.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/index.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/locators.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/manifest.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/markers.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/metadata.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/resources.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/scripts.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/util.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/version.cpython-39.pyc,,
pip/_vendor/distlib/__pycache__/wheel.cpython-39.pyc,,
pip/_vendor/distlib/_backport/__init__.py,sha256=bqS_dTOH6uW9iGgd0uzfpPjo6vZ4xpPZ7kyfZJ2vNaw,274
pip/_vendor/distlib/_backport/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/distlib/_backport/__pycache__/misc.cpython-39.pyc,,
pip/_vendor/distlib/_backport/__pycache__/shutil.cpython-39.pyc,,
pip/_vendor/distlib/_backport/__pycache__/sysconfig.cpython-39.pyc,,
pip/_vendor/distlib/_backport/__pycache__/tarfile.cpython-39.pyc,,
pip/_vendor/distlib/_backport/misc.py,sha256=KWecINdbFNOxSOP1fGF680CJnaC6S4fBRgEtaYTw0ig,971
pip/_vendor/distlib/_backport/shutil.py,sha256=IX_G2NPqwecJibkIDje04bqu0xpHkfSQ2GaGdEVqM5Y,25707
pip/_vendor/distlib/_backport/sysconfig.cfg,sha256=swZKxq9RY5e9r3PXCrlvQPMsvOdiWZBTHLEbqS8LJLU,2617
pip/_vendor/distlib/_backport/sysconfig.py,sha256=BQHFlb6pubCl_dvT1NjtzIthylofjKisox239stDg0U,26854
pip/_vendor/distlib/_backport/tarfile.py,sha256=Ihp7rXRcjbIKw8COm9wSePV9ARGXbSF9gGXAMn2Q-KU,92628
pip/_vendor/distlib/compat.py,sha256=fbsxc5PfJ2wBx1K4k6mQ2goAYs-GZW0tcOPIlE_vf0I,41495
pip/_vendor/distlib/database.py,sha256=Kl0YvPQKc4OcpVi7k5cFziydM1xOK8iqdxLGXgbZHV4,51059
pip/_vendor/distlib/index.py,sha256=UfcimNW19AB7IKWam4VaJbXuCBvArKfSxhV16EwavzE,20739
pip/_vendor/distlib/locators.py,sha256=AKlB3oZvfOTg4E0CtfwOzujFL19X5V4XUA4eHdKOu44,51965
pip/_vendor/distlib/manifest.py,sha256=nQEhYmgoreaBZzyFzwYsXxJARu3fo4EkunU163U16iE,14811
pip/_vendor/distlib/markers.py,sha256=9c70ISEKwBjmUOHuIdOygVnRVESOKdNYp9a2TVn4qrI,4989
pip/_vendor/distlib/metadata.py,sha256=vatoxFdmBr6ie-sTVXVNPOPG3uwMDWJTnEECnm7xDCw,39109
pip/_vendor/distlib/resources.py,sha256=LwbPksc0A1JMbi6XnuPdMBUn83X7BPuFNWqPGEKI698,10820
pip/_vendor/distlib/scripts.py,sha256=tjSwENINeV91ROZxec5zTSMRg2jEeKc4enyCHDzNvEE,17720
pip/_vendor/distlib/util.py,sha256=0Uq_qa63FCLtdyNdWvMnmPbiSvVa-ykHM2E8HT7LSIU,67766
pip/_vendor/distlib/version.py,sha256=WG__LyAa2GwmA6qSoEJtvJE8REA1LZpbSizy8WvhJLk,23513
pip/_vendor/distlib/wheel.py,sha256=pj5VVCjqZMcHvgizORWwAFPS7hOk61CZ59dxP8laQ4E,42943
pip/_vendor/distro.py,sha256=O1EeHMq1-xAO373JI2_6pYEtd09yEkxtmrYkdY-9S-w,48414
pip/_vendor/html5lib/__init__.py,sha256=BYzcKCqeEii52xDrqBFruhnmtmkiuHXFyFh-cglQ8mk,1160
pip/_vendor/html5lib/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/_ihatexml.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/_inputstream.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/_tokenizer.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/_utils.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/constants.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/html5parser.cpython-39.pyc,,
pip/_vendor/html5lib/__pycache__/serializer.cpython-39.pyc,,
pip/_vendor/html5lib/_ihatexml.py,sha256=ifOwF7pXqmyThIXc3boWc96s4MDezqRrRVp7FwDYUFs,16728
pip/_vendor/html5lib/_inputstream.py,sha256=jErNASMlkgs7MpOM9Ve_VdLDJyFFweAjLuhVutZz33U,32353
pip/_vendor/html5lib/_tokenizer.py,sha256=04mgA2sNTniutl2fxFv-ei5bns4iRaPxVXXHh_HrV_4,77040
pip/_vendor/html5lib/_trie/__init__.py,sha256=nqfgO910329BEVJ5T4psVwQtjd2iJyEXQ2-X8c1YxwU,109
pip/_vendor/html5lib/_trie/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/_trie/__pycache__/_base.cpython-39.pyc,,
pip/_vendor/html5lib/_trie/__pycache__/py.cpython-39.pyc,,
pip/_vendor/html5lib/_trie/_base.py,sha256=CaybYyMro8uERQYjby2tTeSUatnWDfWroUN9N7ety5w,1013
pip/_vendor/html5lib/_trie/py.py,sha256=wXmQLrZRf4MyWNyg0m3h81m9InhLR7GJ002mIIZh-8o,1775
pip/_vendor/html5lib/_utils.py,sha256=Dx9AKntksRjFT1veBj7I362pf5OgIaT0zglwq43RnfU,4931
pip/_vendor/html5lib/constants.py,sha256=Ll-yzLU_jcjyAI_h57zkqZ7aQWE5t5xA4y_jQgoUUhw,83464
pip/_vendor/html5lib/filters/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/html5lib/filters/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/alphabeticalattributes.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/base.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/inject_meta_charset.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/lint.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/optionaltags.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/sanitizer.cpython-39.pyc,,
pip/_vendor/html5lib/filters/__pycache__/whitespace.cpython-39.pyc,,
pip/_vendor/html5lib/filters/alphabeticalattributes.py,sha256=lViZc2JMCclXi_5gduvmdzrRxtO5Xo9ONnbHBVCsykU,919
pip/_vendor/html5lib/filters/base.py,sha256=z-IU9ZAYjpsVsqmVt7kuWC63jR11hDMr6CVrvuao8W0,286
pip/_vendor/html5lib/filters/inject_meta_charset.py,sha256=egDXUEHXmAG9504xz0K6ALDgYkvUrC2q15YUVeNlVQg,2945
pip/_vendor/html5lib/filters/lint.py,sha256=jk6q56xY0ojiYfvpdP-OZSm9eTqcAdRqhCoPItemPYA,3643
pip/_vendor/html5lib/filters/optionaltags.py,sha256=8lWT75J0aBOHmPgfmqTHSfPpPMp01T84NKu0CRedxcE,10588
pip/_vendor/html5lib/filters/sanitizer.py,sha256=m6oGmkBhkGAnn2nV6D4hE78SCZ6WEnK9rKdZB3uXBIc,26897
pip/_vendor/html5lib/filters/whitespace.py,sha256=8eWqZxd4UC4zlFGW6iyY6f-2uuT8pOCSALc3IZt7_t4,1214
pip/_vendor/html5lib/html5parser.py,sha256=anr-aXre_ImfrkQ35c_rftKXxC80vJCREKe06Tq15HA,117186
pip/_vendor/html5lib/serializer.py,sha256=_PpvcZF07cwE7xr9uKkZqh5f4UEaI8ltCU2xPJzaTpk,15759
pip/_vendor/html5lib/treeadapters/__init__.py,sha256=A0rY5gXIe4bJOiSGRO_j_tFhngRBO8QZPzPtPw5dFzo,679
pip/_vendor/html5lib/treeadapters/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/treeadapters/__pycache__/genshi.cpython-39.pyc,,
pip/_vendor/html5lib/treeadapters/__pycache__/sax.cpython-39.pyc,,
pip/_vendor/html5lib/treeadapters/genshi.py,sha256=CH27pAsDKmu4ZGkAUrwty7u0KauGLCZRLPMzaO3M5vo,1715
pip/_vendor/html5lib/treeadapters/sax.py,sha256=BKS8woQTnKiqeffHsxChUqL4q2ZR_wb5fc9MJ3zQC8s,1776
pip/_vendor/html5lib/treebuilders/__init__.py,sha256=AysSJyvPfikCMMsTVvaxwkgDieELD5dfR8FJIAuq7hY,3592
pip/_vendor/html5lib/treebuilders/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/treebuilders/__pycache__/base.cpython-39.pyc,,
pip/_vendor/html5lib/treebuilders/__pycache__/dom.cpython-39.pyc,,
pip/_vendor/html5lib/treebuilders/__pycache__/etree.cpython-39.pyc,,
pip/_vendor/html5lib/treebuilders/__pycache__/etree_lxml.cpython-39.pyc,,
pip/_vendor/html5lib/treebuilders/base.py,sha256=z-o51vt9r_l2IDG5IioTOKGzZne4Fy3_Fc-7ztrOh4I,14565
pip/_vendor/html5lib/treebuilders/dom.py,sha256=22whb0C71zXIsai5mamg6qzBEiigcBIvaDy4Asw3at0,8925
pip/_vendor/html5lib/treebuilders/etree.py,sha256=w5ZFpKk6bAxnrwD2_BrF5EVC7vzz0L3LMi9Sxrbc_8w,12836
pip/_vendor/html5lib/treebuilders/etree_lxml.py,sha256=9gqDjs-IxsPhBYa5cpvv2FZ1KZlG83Giusy2lFmvIkE,14766
pip/_vendor/html5lib/treewalkers/__init__.py,sha256=OBPtc1TU5mGyy18QDMxKEyYEz0wxFUUNj5v0-XgmYhY,5719
pip/_vendor/html5lib/treewalkers/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/__pycache__/base.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/__pycache__/dom.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/__pycache__/etree.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/__pycache__/etree_lxml.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/__pycache__/genshi.cpython-39.pyc,,
pip/_vendor/html5lib/treewalkers/base.py,sha256=ouiOsuSzvI0KgzdWP8PlxIaSNs9falhbiinAEc_UIJY,7476
pip/_vendor/html5lib/treewalkers/dom.py,sha256=EHyFR8D8lYNnyDU9lx_IKigVJRyecUGua0mOi7HBukc,1413
pip/_vendor/html5lib/treewalkers/etree.py,sha256=xo1L5m9VtkfpFJK0pFmkLVajhqYYVisVZn3k9kYpPkI,4551
pip/_vendor/html5lib/treewalkers/etree_lxml.py,sha256=_b0LAVWLcVu9WaU_-w3D8f0IRSpCbjf667V-3NRdhTw,6357
pip/_vendor/html5lib/treewalkers/genshi.py,sha256=4D2PECZ5n3ZN3qu3jMl9yY7B81jnQApBQSVlfaIuYbA,2309
pip/_vendor/idna/__init__.py,sha256=KJQN1eQBr8iIK5SKrJ47lXvxG0BJ7Lm38W4zT0v_8lk,849
pip/_vendor/idna/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/codec.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/core.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/idnadata.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/intranges.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/package_data.cpython-39.pyc,,
pip/_vendor/idna/__pycache__/uts46data.cpython-39.pyc,,
pip/_vendor/idna/codec.py,sha256=QsPFD3Je8gN17rfs14e7zTGRWlnL7bNf2ZqcHTRVYHs,3453
pip/_vendor/idna/compat.py,sha256=5A9xR04puRHCsyjBNewZlVSiarth7K1bZqyEOeob1fA,360
pip/_vendor/idna/core.py,sha256=icq2P13S6JMjoXgKhhd6ihhby7QsnZlNfniH6fLyf6U,12826
pip/_vendor/idna/idnadata.py,sha256=cl4x9RLdw1ZMtEEbvKwAsX-Id3AdIjO5U3HaoKM6VGs,42350
pip/_vendor/idna/intranges.py,sha256=EqgXwyATAn-CTACInqH9tYsYAitGB2VcQ50RZt_Cpjs,1933
pip/_vendor/idna/package_data.py,sha256=_028B4fvadRIaXMwMYjhuQPP3AxTIt1IRE7X6RDR4Mk,21
pip/_vendor/idna/uts46data.py,sha256=DGzwDQv8JijY17I_7ondo3stjFjNnjvVAbA-z0k1XOE,201849
pip/_vendor/msgpack/__init__.py,sha256=2gJwcsTIaAtCM0GMi2rU-_Y6kILeeQuqRkrQ22jSANc,1118
pip/_vendor/msgpack/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/msgpack/__pycache__/_version.cpython-39.pyc,,
pip/_vendor/msgpack/__pycache__/exceptions.cpython-39.pyc,,
pip/_vendor/msgpack/__pycache__/ext.cpython-39.pyc,,
pip/_vendor/msgpack/__pycache__/fallback.cpython-39.pyc,,
pip/_vendor/msgpack/_version.py,sha256=dFR03oACnj4lsKd1RnwD7BPMiVI_FMygdOL1TOBEw_U,20
pip/_vendor/msgpack/exceptions.py,sha256=dCTWei8dpkrMsQDcjQk74ATl9HsIBH0ybt8zOPNqMYc,1081
pip/_vendor/msgpack/ext.py,sha256=4l356Y4sVEcvCla2dh_cL57vh4GMhZfa3kuWHFHYz6A,6088
pip/_vendor/msgpack/fallback.py,sha256=Rpv1Ldey8f8ueRnQznD4ARKBn9dxM2PywVNkXI8IEeE,38026
pip/_vendor/packaging/__about__.py,sha256=p_OQloqH2saadcbUQmWEsWK857dI6_ff5E3aSiCqGFA,661
pip/_vendor/packaging/__init__.py,sha256=b9Kk5MF7KxhhLgcDmiUWukN-LatWFxPdNug0joPhHSk,497
pip/_vendor/packaging/__pycache__/__about__.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/_manylinux.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/_musllinux.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/_structures.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/markers.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/requirements.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/tags.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/utils.cpython-39.pyc,,
pip/_vendor/packaging/__pycache__/version.cpython-39.pyc,,
pip/_vendor/packaging/_manylinux.py,sha256=XcbiXB-qcjv3bcohp6N98TMpOP4_j3m-iOA8ptK2GWY,11488
pip/_vendor/packaging/_musllinux.py,sha256=z5yeG1ygOPx4uUyLdqj-p8Dk5UBb5H_b0NIjW9yo8oA,4378
pip/_vendor/packaging/_structures.py,sha256=TMiAgFbdUOPmIfDIfiHc3KFhSJ8kMjof2QS5I-2NyQ8,1629
pip/_vendor/packaging/markers.py,sha256=AJBOcY8Oq0kYc570KuuPTkvuqjAlhufaE2c9sCUbm64,8487
pip/_vendor/packaging/requirements.py,sha256=NtDlPBtojpn1IUC85iMjPNsUmufjpSlwnNA-Xb4m5NA,4676
pip/_vendor/packaging/specifiers.py,sha256=MZ-fYcNL3u7pNrt-6g2EQO7AbRXkjc-SPEYwXMQbLmc,30964
pip/_vendor/packaging/tags.py,sha256=akIerYw8W0sz4OW9HHozgawWnbt2GGOPm3sviW0jowY,15714
pip/_vendor/packaging/utils.py,sha256=dJjeat3BS-TYn1RrUFVwufUMasbtzLfYRoy_HXENeFQ,4200
pip/_vendor/packaging/version.py,sha256=AOLsewGcHCskMRdz2ykO3Gv7U7YZwgzE5Pj9ifjd6S8,14479
pip/_vendor/pep517/__init__.py,sha256=Y1bATL2qbFNN6M_DQa4yyrwqjpIiL-j9T6kBmR0DS14,130
pip/_vendor/pep517/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/build.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/check.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/colorlog.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/dirtools.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/envbuild.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/meta.cpython-39.pyc,,
pip/_vendor/pep517/__pycache__/wrappers.cpython-39.pyc,,
pip/_vendor/pep517/build.py,sha256=2bar6EdjwIz2Dlfy94qdxn3oA9mVnnny40mfoT5f-qI,3457
pip/_vendor/pep517/check.py,sha256=bCORq1WrHjhpTONa-zpAqG0EB9rHNuhO1ORu6DsDuL8,6084
pip/_vendor/pep517/colorlog.py,sha256=Tk9AuYm_cLF3BKTBoSTJt9bRryn0aFojIQOwbfVUTxQ,4098
pip/_vendor/pep517/compat.py,sha256=NmLImE5oiDT3gbEhJ4w7xeoMFcpAPrGu_NltBytSJUY,1253
pip/_vendor/pep517/dirtools.py,sha256=2mkAkAL0mRz_elYFjRKuekTJVipH1zTn4tbf1EDev84,1129
pip/_vendor/pep517/envbuild.py,sha256=zFde--rmzjXMLXcm7SA_3hDtgk5VCTA8hjpk88RbF6E,6100
pip/_vendor/pep517/in_process/__init__.py,sha256=MyWoAi8JHdcBv7yXuWpUSVADbx6LSB9rZh7kTIgdA8Y,563
pip/_vendor/pep517/in_process/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/pep517/in_process/__pycache__/_in_process.cpython-39.pyc,,
pip/_vendor/pep517/in_process/_in_process.py,sha256=D3waguyNSGcwosociD5USfcycYr2RCzCjYtxX5UHQmQ,11201
pip/_vendor/pep517/meta.py,sha256=8mnM5lDnT4zXQpBTliJbRGfesH7iioHwozbDxALPS9Y,2463
pip/_vendor/pep517/wrappers.py,sha256=impq7Cz_LL1iDF1iiOzYWB4MaEu6O6Gps7TJ5qsJz1Q,13429
pip/_vendor/pkg_resources/__init__.py,sha256=NnpQ3g6BCHzpMgOR_OLBmYtniY4oOzdKpwqghfq_6ug,108287
pip/_vendor/pkg_resources/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/pkg_resources/__pycache__/py31compat.cpython-39.pyc,,
pip/_vendor/pkg_resources/py31compat.py,sha256=CRk8fkiPRDLsbi5pZcKsHI__Pbmh_94L8mr9Qy9Ab2U,562
pip/_vendor/platformdirs/__init__.py,sha256=3iz938Grn-6IRg8gSuMxJtgiBfH0xqRqAlMBo-vPGUw,12859
pip/_vendor/platformdirs/__main__.py,sha256=SzGvNkYWuosrWXs2yL2VqcXEh-kivWq3-53-BpTco0o,1140
pip/_vendor/platformdirs/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/__main__.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/android.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/api.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/macos.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/unix.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/version.cpython-39.pyc,,
pip/_vendor/platformdirs/__pycache__/windows.cpython-39.pyc,,
pip/_vendor/platformdirs/android.py,sha256=dadYfG2oc900YVi5AONQWw2WEvk-kmgkZs5iiNSiWiE,3994
pip/_vendor/platformdirs/api.py,sha256=yhRR6RkcZzPBfJD4Sn90vCHZbRMQ9nwtnRaa93X1wR8,4922
pip/_vendor/platformdirs/macos.py,sha256=vIowPYKkHksJcWVjqHQoa-oI1i2D0S7gsSdyFzZDJEA,2619
pip/_vendor/platformdirs/unix.py,sha256=7JdDnsyTFn2IHC8IFdiNYH7_R8VS-rPx8ivh4_dT1DU,6905
pip/_vendor/platformdirs/version.py,sha256=uUssQTtUqVP-PxbOSNBzNGRW27X5u1GvOllg--kzyuw,80
pip/_vendor/platformdirs/windows.py,sha256=91nNccR0CSxX_myMppSvUT1qtQao6kaO96e6ior8-Xw,6416
pip/_vendor/progress/__init__.py,sha256=1HejNZtv2ouUNQeStUDAtZrtwkz_3FmYKQ476hJ7zOs,5294
pip/_vendor/progress/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/progress/__pycache__/bar.cpython-39.pyc,,
pip/_vendor/progress/__pycache__/colors.cpython-39.pyc,,
pip/_vendor/progress/__pycache__/counter.cpython-39.pyc,,
pip/_vendor/progress/__pycache__/spinner.cpython-39.pyc,,
pip/_vendor/progress/bar.py,sha256=GbedY0oZ-Q1duXjmvVLO0tSf-uTSH7hJ3zzyI91Esws,2942
pip/_vendor/progress/colors.py,sha256=cCYXQnYFYVmQKKmYEbQ_lj6SPSFzdw4FN98F2x2kR-U,2655
pip/_vendor/progress/counter.py,sha256=zYt9DWH0_05s8Q9TrJwHVud-WwsyyaR3PwYtk5hxwwQ,1613
pip/_vendor/progress/spinner.py,sha256=u5ElzW94XEiLGH-aAlr54VJtKfeK745xr6UfGvvflzU,1461
pip/_vendor/pyparsing.py,sha256=J1b4z3S_KwyJW7hKGnoN-hXW9pgMIzIP6QThyY5yJq4,273394
pip/_vendor/requests/__init__.py,sha256=g4Bh1QYh6JKjMS4YLobx0uOLq-41sINaXjvbhX2VI8g,5113
pip/_vendor/requests/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/__version__.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/_internal_utils.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/adapters.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/api.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/auth.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/certs.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/compat.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/cookies.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/exceptions.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/help.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/hooks.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/models.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/packages.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/sessions.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/status_codes.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/structures.cpython-39.pyc,,
pip/_vendor/requests/__pycache__/utils.cpython-39.pyc,,
pip/_vendor/requests/__version__.py,sha256=PZEyPTSIN_jRIAIB51wV7pw81m3qAw0InSR7OrKZUnE,441
pip/_vendor/requests/_internal_utils.py,sha256=Zx3PnEUccyfsB-ie11nZVAW8qClJy0gx1qNME7rgT18,1096
pip/_vendor/requests/adapters.py,sha256=e-bmKEApNVqFdylxuMJJfiaHdlmS_zhWhIMEzlHvGuc,21548
pip/_vendor/requests/api.py,sha256=hjuoP79IAEmX6Dysrw8t032cLfwLHxbI_wM4gC5G9t0,6402
pip/_vendor/requests/auth.py,sha256=OMoJIVKyRLy9THr91y8rxysZuclwPB-K1Xg1zBomUhQ,10207
pip/_vendor/requests/certs.py,sha256=nXRVq9DtGmv_1AYbwjTu9UrgAcdJv05ZvkNeaoLOZxY,465
pip/_vendor/requests/compat.py,sha256=LQWuCR4qXk6w7-qQopXyz0WNHUdAD40k0mKnaAEf1-g,2045
pip/_vendor/requests/cookies.py,sha256=Y-bKX6TvW3FnYlE6Au0SXtVVWcaNdFvuAwQxw-G0iTI,18430
pip/_vendor/requests/exceptions.py,sha256=dwIi512RCDqXJ2T81nLC88mqPNhUFnOI_CgKKDXhTO8,3250
pip/_vendor/requests/help.py,sha256=dyhe3lcmHXnFCzDiZVjcGmVvvO_jtsfAm-AC542ndw8,3972
pip/_vendor/requests/hooks.py,sha256=QReGyy0bRcr5rkwCuObNakbYsc7EkiKeBwG4qHekr2Q,757
pip/_vendor/requests/models.py,sha256=9_LS_t1t6HbbaWFE3ZkxGmmHN2V8BgxziiOU84rrQ50,34924
pip/_vendor/requests/packages.py,sha256=njJmVifY4aSctuW3PP5EFRCxjEwMRDO6J_feG2dKWsI,695
pip/_vendor/requests/sessions.py,sha256=57O4ud9yRL6eLYh-dtFbqC1kO4d_EwZcCgYXEkujlfs,30168
pip/_vendor/requests/status_codes.py,sha256=gT79Pbs_cQjBgp-fvrUgg1dn2DQO32bDj4TInjnMPSc,4188
pip/_vendor/requests/structures.py,sha256=msAtr9mq1JxHd-JRyiILfdFlpbJwvvFuP3rfUQT_QxE,3005
pip/_vendor/requests/utils.py,sha256=U_-i6WxLw-67KEij43xHbcvL0DdeQ5Jbd4hfifWJzQY,31394
pip/_vendor/resolvelib/__init__.py,sha256=fzWkeoLV8ol6l2fvBVRZZLylOePc9w9tKRvUb8RJsCY,537
pip/_vendor/resolvelib/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/resolvelib/__pycache__/providers.cpython-39.pyc,,
pip/_vendor/resolvelib/__pycache__/reporters.cpython-39.pyc,,
pip/_vendor/resolvelib/__pycache__/resolvers.cpython-39.pyc,,
pip/_vendor/resolvelib/__pycache__/structs.cpython-39.pyc,,
pip/_vendor/resolvelib/compat/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/resolvelib/compat/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/resolvelib/compat/__pycache__/collections_abc.cpython-39.pyc,,
pip/_vendor/resolvelib/compat/collections_abc.py,sha256=uy8xUZ-NDEw916tugUXm8HgwCGiMO0f-RcdnpkfXfOs,156
pip/_vendor/resolvelib/providers.py,sha256=roVmFBItQJ0TkhNua65h8LdNny7rmeqVEXZu90QiP4o,5872
pip/_vendor/resolvelib/reporters.py,sha256=hQvvXuuEBOyEWO8KDfLsWKVjX55UFMAUwO0YZMNpzAw,1364
pip/_vendor/resolvelib/resolvers.py,sha256=UjFUEVrUa1hCzfEEakmjHEjYAL9J5ACJmwZyHFdmzvE,17540
pip/_vendor/resolvelib/structs.py,sha256=IVIYof6sA_N4ZEiE1C1UhzTX495brCNnyCdgq6CYq28,4794
pip/_vendor/six.py,sha256=TOOfQi7nFGfMrIvtdr6wX4wyHH8M7aknmuLfo2cBBrM,34549
pip/_vendor/tenacity/__init__.py,sha256=GLLsTFD4Bd5VDgTR6mU_FxyOsrxc48qONorVaRebeD4,18257
pip/_vendor/tenacity/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/_asyncio.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/_utils.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/after.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/before.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/before_sleep.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/nap.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/retry.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/stop.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/tornadoweb.cpython-39.pyc,,
pip/_vendor/tenacity/__pycache__/wait.cpython-39.pyc,,
pip/_vendor/tenacity/_asyncio.py,sha256=HEb0BVJEeBJE9P-m9XBxh1KcaF96BwoeqkJCL5sbVcQ,3314
pip/_vendor/tenacity/_utils.py,sha256=-y68scDcyoqvTJuJJ0GTfjdSCljEYlbCYvgk7nM4NdM,1944
pip/_vendor/tenacity/after.py,sha256=dlmyxxFy2uqpLXDr838DiEd7jgv2AGthsWHGYcGYsaI,1496
pip/_vendor/tenacity/before.py,sha256=7XtvRmO0dRWUp8SVn24OvIiGFj8-4OP5muQRUiWgLh0,1376
pip/_vendor/tenacity/before_sleep.py,sha256=ThyDvqKU5yle_IvYQz_b6Tp6UjUS0PhVp6zgqYl9U6Y,1908
pip/_vendor/tenacity/nap.py,sha256=fRWvnz1aIzbIq9Ap3gAkAZgDH6oo5zxMrU6ZOVByq0I,1383
pip/_vendor/tenacity/retry.py,sha256=62R71W59bQjuNyFKsDM7hE2aEkEPtwNBRA0tnsEvgSk,6645
pip/_vendor/tenacity/stop.py,sha256=sKHmHaoSaW6sKu3dTxUVKr1-stVkY7lw4Y9yjZU30zQ,2790
pip/_vendor/tenacity/tornadoweb.py,sha256=E8lWO2nwe6dJgoB-N2HhQprYLDLB_UdSgFnv-EN6wKE,2145
pip/_vendor/tenacity/wait.py,sha256=e_Saa6I2tsNLpCL1t9897wN2fGb0XQMQlE4bU2t9V2w,6691
pip/_vendor/tomli/__init__.py,sha256=z1Elt0nLAqU5Y0DOn9p__8QnLWavlEOpRyQikdYgKro,230
pip/_vendor/tomli/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/tomli/__pycache__/_parser.cpython-39.pyc,,
pip/_vendor/tomli/__pycache__/_re.cpython-39.pyc,,
pip/_vendor/tomli/_parser.py,sha256=50BD4o9YbzFAGAYyZLqZC8F81DQ7iWWyJnrHNwBKa6A,22415
pip/_vendor/tomli/_re.py,sha256=5GPfgXKteg7wRFCF-DzlkAPI2ilHbkMK2-JC49F-AJQ,2681
pip/_vendor/urllib3/__init__.py,sha256=j3yzHIbmW7CS-IKQJ9-PPQf_YKO8EOAey_rMW0UR7us,2763
pip/_vendor/urllib3/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/_collections.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/_version.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/connection.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/connectionpool.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/exceptions.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/fields.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/filepost.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/poolmanager.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/request.cpython-39.pyc,,
pip/_vendor/urllib3/__pycache__/response.cpython-39.pyc,,
pip/_vendor/urllib3/_collections.py,sha256=Rp1mVyBgc_UlAcp6M3at1skJBXR5J43NawRTvW2g_XY,10811
pip/_vendor/urllib3/_version.py,sha256=CA4bKbKLwUBfKitbVR-44Whe53HWyInIVElDQQniAJU,63
pip/_vendor/urllib3/connection.py,sha256=8TiEbQrJMgySqOllKNeX5tMv8nluKRjNj5j9hyzS6x0,20080
pip/_vendor/urllib3/connectionpool.py,sha256=FQoodlNAP1KeUi4htGdl5TJEvKL5LWisCbmFNewxRpg,37587
pip/_vendor/urllib3/contrib/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/urllib3/contrib/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/_appengine_environ.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/appengine.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/ntlmpool.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/pyopenssl.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/securetransport.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/__pycache__/socks.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/_appengine_environ.py,sha256=bDbyOEhW2CKLJcQqAKAyrEHN-aklsyHFKq6vF8ZFsmk,957
pip/_vendor/urllib3/contrib/_securetransport/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/urllib3/contrib/_securetransport/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/_securetransport/__pycache__/bindings.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/_securetransport/__pycache__/low_level.cpython-39.pyc,,
pip/_vendor/urllib3/contrib/_securetransport/bindings.py,sha256=eRy1Mj-wpg7sR6-OSvnSV4jUbjMT464dLN_CWxbIRVw,17649
pip/_vendor/urllib3/contrib/_securetransport/low_level.py,sha256=B2JBB2_NRP02xK6DCa1Pa9IuxrPwxzDzZbixQkb7U9M,13922
pip/_vendor/urllib3/contrib/appengine.py,sha256=lfzpHFmJiO82shClLEm3QB62SYgHWnjpZOH_2JhU5Tc,11034
pip/_vendor/urllib3/contrib/ntlmpool.py,sha256=ej9gGvfAb2Gt00lafFp45SIoRz-QwrQ4WChm6gQmAlM,4538
pip/_vendor/urllib3/contrib/pyopenssl.py,sha256=DD4pInv_3OEEGffEFynBoirc8ldR789sLmGSKukzA0E,16900
pip/_vendor/urllib3/contrib/securetransport.py,sha256=4qUKo7PUV-vVIqXmr2BD-sH7qplB918jiD5eNsRI9vU,34449
pip/_vendor/urllib3/contrib/socks.py,sha256=aRi9eWXo9ZEb95XUxef4Z21CFlnnjbEiAo9HOseoMt4,7097
pip/_vendor/urllib3/exceptions.py,sha256=0Mnno3KHTNfXRfY7638NufOPkUb6mXOm-Lqj-4x2w8A,8217
pip/_vendor/urllib3/fields.py,sha256=kvLDCg_JmH1lLjUUEY_FLS8UhY7hBvDPuVETbY8mdrM,8579
pip/_vendor/urllib3/filepost.py,sha256=5b_qqgRHVlL7uLtdAYBzBh-GHmU5AfJVt_2N0XS3PeY,2440
pip/_vendor/urllib3/packages/__init__.py,sha256=h4BLhD4tLaBx1adaDtKXfupsgqY0wWLXb_f1_yVlV6A,108
pip/_vendor/urllib3/packages/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/packages/__pycache__/six.cpython-39.pyc,,
pip/_vendor/urllib3/packages/backports/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pip/_vendor/urllib3/packages/backports/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/packages/backports/__pycache__/makefile.cpython-39.pyc,,
pip/_vendor/urllib3/packages/backports/makefile.py,sha256=nbzt3i0agPVP07jqqgjhaYjMmuAi_W5E0EywZivVO8E,1417
pip/_vendor/urllib3/packages/six.py,sha256=1LVW7ljqRirFlfExjwl-v1B7vSAUNTmzGMs-qays2zg,34666
pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py,sha256=ZVMwCkHx-py8ERsxxM3Il-MiREZktV-8iLBmCfRRHI4,927
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/_implementation.cpython-39.pyc,,
pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.py,sha256=6dZ-q074g7XhsJ27MFCgkct8iVNZB3sMZvKhf-KUVy0,5679
pip/_vendor/urllib3/poolmanager.py,sha256=whzlX6UTEgODMOCy0ZDMUONRBCz5wyIM8Z9opXAY-Lk,19763
pip/_vendor/urllib3/request.py,sha256=ZFSIqX0C6WizixecChZ3_okyu7BEv0lZu1VT0s6h4SM,5985
pip/_vendor/urllib3/response.py,sha256=hGhGBh7TkEkh_IQg5C1W_xuPNrgIKv5BUXPyE-q0LuE,28203
pip/_vendor/urllib3/util/__init__.py,sha256=JEmSmmqqLyaw8P51gUImZh8Gwg9i1zSe-DoqAitn2nc,1155
pip/_vendor/urllib3/util/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/connection.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/proxy.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/queue.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/request.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/response.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/retry.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/ssl_.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/ssltransport.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/timeout.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/url.cpython-39.pyc,,
pip/_vendor/urllib3/util/__pycache__/wait.cpython-39.pyc,,
pip/_vendor/urllib3/util/connection.py,sha256=KykjNIXzUZEzeKEOpl5xvKs6IsESXP9o9eTrjE0W_Ys,4920
pip/_vendor/urllib3/util/proxy.py,sha256=zUvPPCJrp6dOF0N4GAVbOcl6o-4uXKSrGiTkkr5vUS4,1605
pip/_vendor/urllib3/util/queue.py,sha256=nRgX8_eX-_VkvxoX096QWoz8Ps0QHUAExILCY_7PncM,498
pip/_vendor/urllib3/util/request.py,sha256=NnzaEKQ1Pauw5MFMV6HmgEMHITf0Aua9fQuzi2uZzGc,4123
pip/_vendor/urllib3/util/response.py,sha256=GJpg3Egi9qaJXRwBh5wv-MNuRWan5BIu40oReoxWP28,3510
pip/_vendor/urllib3/util/retry.py,sha256=tOWfZpLsuc7Vbk5nWpMwkHdMoXCp90IAvH4xtjSDRqQ,21391
pip/_vendor/urllib3/util/ssl_.py,sha256=X4-AqW91aYPhPx6-xbf66yHFQKbqqfC_5Zt4WkLX1Hc,17177
pip/_vendor/urllib3/util/ssltransport.py,sha256=F_UncOXGcc-MgeWFTA1H4QCt_RRNQXRbF6onje3SyHY,6931
pip/_vendor/urllib3/util/timeout.py,sha256=QSbBUNOB9yh6AnDn61SrLQ0hg5oz0I9-uXEG91AJuIg,10003
pip/_vendor/urllib3/util/url.py,sha256=QVEzcbHipbXyCWwH6R4K4TR-N8T4LM55WEMwNUTBmLE,14047
pip/_vendor/urllib3/util/wait.py,sha256=3MUKRSAUJDB2tgco7qRUskW0zXGAWYvRRE4Q1_6xlLs,5404
pip/_vendor/vendor.txt,sha256=vux9Tgc3pSRZZnXz9TNDdn514NdkDdnb-QPC0LCHkK4,432
pip/_vendor/webencodings/__init__.py,sha256=qOBJIuPy_4ByYH6W_bNgJF-qYQ2DoU-dKsDu5yRWCXg,10579
pip/_vendor/webencodings/__pycache__/__init__.cpython-39.pyc,,
pip/_vendor/webencodings/__pycache__/labels.cpython-39.pyc,,
pip/_vendor/webencodings/__pycache__/mklabels.cpython-39.pyc,,
pip/_vendor/webencodings/__pycache__/tests.cpython-39.pyc,,
pip/_vendor/webencodings/__pycache__/x_user_defined.cpython-39.pyc,,
pip/_vendor/webencodings/labels.py,sha256=4AO_KxTddqGtrL9ns7kAPjb0CcN6xsCIxbK37HY9r3E,8979
pip/_vendor/webencodings/mklabels.py,sha256=GYIeywnpaLnP0GSic8LFWgd0UVvO_l1Nc6YoF-87R_4,1305
pip/_vendor/webencodings/tests.py,sha256=OtGLyjhNY1fvkW1GvLJ_FV9ZoqC9Anyjr7q3kxTbzNs,6563
pip/_vendor/webencodings/x_user_defined.py,sha256=yOqWSdmpytGfUgh_Z6JYgDNhoc-BAHyyeeT15Fr42tM,4307
pip/py.typed,sha256=EBVvvPRTn_eIpz5e5QztSCdrMX7Qwd7VP93RSoIlZ2I,286
site-packages/pip-21.3.1.dist-info/entry_points.txt000064400000000175150564745010015777 0ustar00[console_scripts]
pip = pip._internal.cli.main:main
pip3 = pip._internal.cli.main:main
pip3.9 = pip._internal.cli.main:main

site-packages/pip-21.3.1.dist-info/INSTALLER000064400000000004150564745010013750 0ustar00rpm
site-packages/pip-21.3.1.dist-info/REQUESTED000064400000000000150564745010013750 0ustar00site-packages/pip-21.3.1.dist-info/METADATA000064400000010170150564745010013600 0ustar00Metadata-Version: 2.1
Name: pip
Version: 21.3.1
Summary: The PyPA recommended tool for installing Python packages.
Home-page: https://pip.pypa.io/
Author: The pip developers
Author-email: distutils-sig@python.org
License: MIT
Project-URL: Documentation, https://pip.pypa.io
Project-URL: Source, https://github.com/pypa/pip
Project-URL: Changelog, https://pip.pypa.io/en/stable/news/
Platform: UNKNOWN
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: MIT License
Classifier: Topic :: Software Development :: Build Tools
Classifier: Programming Language :: Python
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3 :: Only
Classifier: Programming Language :: Python :: 3.6
Classifier: Programming Language :: Python :: 3.7
Classifier: Programming Language :: Python :: 3.8
Classifier: Programming Language :: Python :: 3.9
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: Implementation :: CPython
Classifier: Programming Language :: Python :: Implementation :: PyPy
Requires-Python: >=3.6
License-File: LICENSE.txt

pip - The Python Package Installer
==================================

.. image:: https://img.shields.io/pypi/v/pip.svg
   :target: https://pypi.org/project/pip/

.. image:: https://readthedocs.org/projects/pip/badge/?version=latest
   :target: https://pip.pypa.io/en/latest

pip is the `package installer`_ for Python. You can use pip to install packages from the `Python Package Index`_ and other indexes.

Please take a look at our documentation for how to install and use pip:

* `Installation`_
* `Usage`_

We release updates regularly, with a new version every 3 months. Find more details in our documentation:

* `Release notes`_
* `Release process`_

In pip 20.3, we've `made a big improvement to the heart of pip`_; `learn more`_. We want your input, so `sign up for our user experience research studies`_ to help us do it right.

**Note**: pip 21.0, in January 2021, removed Python 2 support, per pip's `Python 2 support policy`_. Please migrate to Python 3.

If you find bugs, need help, or want to talk to the developers, please use our mailing lists or chat rooms:

* `Issue tracking`_
* `Discourse channel`_
* `User IRC`_

If you want to get involved head over to GitHub to get the source code, look at our development documentation and feel free to jump on the developer mailing lists and chat rooms:

* `GitHub page`_
* `Development documentation`_
* `Development mailing list`_
* `Development IRC`_

Code of Conduct
---------------

Everyone interacting in the pip project's codebases, issue trackers, chat
rooms, and mailing lists is expected to follow the `PSF Code of Conduct`_.

.. _package installer: https://packaging.python.org/guides/tool-recommendations/
.. _Python Package Index: https://pypi.org
.. _Installation: https://pip.pypa.io/en/stable/installation/
.. _Usage: https://pip.pypa.io/en/stable/
.. _Release notes: https://pip.pypa.io/en/stable/news.html
.. _Release process: https://pip.pypa.io/en/latest/development/release-process/
.. _GitHub page: https://github.com/pypa/pip
.. _Development documentation: https://pip.pypa.io/en/latest/development
.. _made a big improvement to the heart of pip: https://pyfound.blogspot.com/2020/11/pip-20-3-new-resolver.html
.. _learn more: https://pip.pypa.io/en/latest/user_guide/#changes-to-the-pip-dependency-resolver-in-20-3-2020
.. _sign up for our user experience research studies: https://pyfound.blogspot.com/2020/03/new-pip-resolver-to-roll-out-this-year.html
.. _Python 2 support policy: https://pip.pypa.io/en/latest/development/release-process/#python-2-support
.. _Issue tracking: https://github.com/pypa/pip/issues
.. _Discourse channel: https://discuss.python.org/c/packaging
.. _Development mailing list: https://mail.python.org/mailman3/lists/distutils-sig.python.org/
.. _User IRC: https://kiwiirc.com/nextclient/#ircs://irc.libera.chat:+6697/pypa
.. _Development IRC: https://kiwiirc.com/nextclient/#ircs://irc.libera.chat:+6697/pypa-dev
.. _PSF Code of Conduct: https://github.com/pypa/.github/blob/main/CODE_OF_CONDUCT.md


site-packages/pip-21.3.1.dist-info/top_level.txt000064400000000004150564745010015222 0ustar00pip
site-packages/pip-21.3.1.dist-info/LICENSE.txt000064400000002102150564745010014314 0ustar00Copyright (c) 2008-2021 The pip developers (see AUTHORS.txt file)

Permission is hereby granted, free of charge, to any person obtaining
a copy of this software and associated documentation files (the
"Software"), to deal in the Software without restriction, including
without limitation the rights to use, copy, modify, merge, publish,
distribute, sublicense, and/or sell copies of the Software, and to
permit persons to whom the Software is furnished to do so, subject to
the following conditions:

The above copyright notice and this permission notice shall be
included in all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
site-packages/distutils-precedence.pth000064400000000230150564745010014122 0ustar00import os; var = 'SETUPTOOLS_USE_DISTUTILS'; enabled = os.environ.get(var, 'stdlib') == 'local'; enabled and __import__('_distutils_hack').add_shim(); 
site-packages/setuptools/glob.py000064400000011411150564745010013007 0ustar00"""
Filename globbing utility. Mostly a copy of `glob` from Python 3.5.

Changes include:
 * `yield from` and PEP3102 `*` removed.
 * Hidden files are not ignored.
"""

import os
import re
import fnmatch

__all__ = ["glob", "iglob", "escape"]


def glob(pathname, recursive=False):
    """Return a list of paths matching a pathname pattern.

    The pattern may contain simple shell-style wildcards a la
    fnmatch. However, unlike fnmatch, filenames starting with a
    dot are special cases that are not matched by '*' and '?'
    patterns.

    If recursive is true, the pattern '**' will match any files and
    zero or more directories and subdirectories.
    """
    return list(iglob(pathname, recursive=recursive))


def iglob(pathname, recursive=False):
    """Return an iterator which yields the paths matching a pathname pattern.

    The pattern may contain simple shell-style wildcards a la
    fnmatch. However, unlike fnmatch, filenames starting with a
    dot are special cases that are not matched by '*' and '?'
    patterns.

    If recursive is true, the pattern '**' will match any files and
    zero or more directories and subdirectories.
    """
    it = _iglob(pathname, recursive)
    if recursive and _isrecursive(pathname):
        s = next(it)  # skip empty string
        assert not s
    return it


def _iglob(pathname, recursive):
    dirname, basename = os.path.split(pathname)
    glob_in_dir = glob2 if recursive and _isrecursive(basename) else glob1

    if not has_magic(pathname):
        if basename:
            if os.path.lexists(pathname):
                yield pathname
        else:
            # Patterns ending with a slash should match only directories
            if os.path.isdir(dirname):
                yield pathname
        return

    if not dirname:
        yield from glob_in_dir(dirname, basename)
        return
    # `os.path.split()` returns the argument itself as a dirname if it is a
    # drive or UNC path.  Prevent an infinite recursion if a drive or UNC path
    # contains magic characters (i.e. r'\\?\C:').
    if dirname != pathname and has_magic(dirname):
        dirs = _iglob(dirname, recursive)
    else:
        dirs = [dirname]
    if not has_magic(basename):
        glob_in_dir = glob0
    for dirname in dirs:
        for name in glob_in_dir(dirname, basename):
            yield os.path.join(dirname, name)


# These 2 helper functions non-recursively glob inside a literal directory.
# They return a list of basenames. `glob1` accepts a pattern while `glob0`
# takes a literal basename (so it only has to check for its existence).


def glob1(dirname, pattern):
    if not dirname:
        if isinstance(pattern, bytes):
            dirname = os.curdir.encode('ASCII')
        else:
            dirname = os.curdir
    try:
        names = os.listdir(dirname)
    except OSError:
        return []
    return fnmatch.filter(names, pattern)


def glob0(dirname, basename):
    if not basename:
        # `os.path.split()` returns an empty basename for paths ending with a
        # directory separator.  'q*x/' should match only directories.
        if os.path.isdir(dirname):
            return [basename]
    else:
        if os.path.lexists(os.path.join(dirname, basename)):
            return [basename]
    return []


# This helper function recursively yields relative pathnames inside a literal
# directory.


def glob2(dirname, pattern):
    assert _isrecursive(pattern)
    yield pattern[:0]
    for x in _rlistdir(dirname):
        yield x


# Recursively yields relative pathnames inside a literal directory.
def _rlistdir(dirname):
    if not dirname:
        if isinstance(dirname, bytes):
            dirname = os.curdir.encode('ASCII')
        else:
            dirname = os.curdir
    try:
        names = os.listdir(dirname)
    except os.error:
        return
    for x in names:
        yield x
        path = os.path.join(dirname, x) if dirname else x
        for y in _rlistdir(path):
            yield os.path.join(x, y)


magic_check = re.compile('([*?[])')
magic_check_bytes = re.compile(b'([*?[])')


def has_magic(s):
    if isinstance(s, bytes):
        match = magic_check_bytes.search(s)
    else:
        match = magic_check.search(s)
    return match is not None


def _isrecursive(pattern):
    if isinstance(pattern, bytes):
        return pattern == b'**'
    else:
        return pattern == '**'


def escape(pathname):
    """Escape all special characters.
    """
    # Escaping is done by wrapping any of "*?[" between square brackets.
    # Metacharacters do not work in the drive part and shouldn't be escaped.
    drive, pathname = os.path.splitdrive(pathname)
    if isinstance(pathname, bytes):
        pathname = magic_check_bytes.sub(br'[\1]', pathname)
    else:
        pathname = magic_check.sub(r'[\1]', pathname)
    return drive + pathname
site-packages/setuptools/archive_util.py000064400000015645150564745010014557 0ustar00"""Utilities for extracting common archive formats"""

import zipfile
import tarfile
import os
import shutil
import posixpath
import contextlib
from distutils.errors import DistutilsError

from pkg_resources import ensure_directory

__all__ = [
    "unpack_archive", "unpack_zipfile", "unpack_tarfile", "default_filter",
    "UnrecognizedFormat", "extraction_drivers", "unpack_directory",
]


class UnrecognizedFormat(DistutilsError):
    """Couldn't recognize the archive type"""


def default_filter(src, dst):
    """The default progress/filter callback; returns True for all files"""
    return dst


def unpack_archive(
        filename, extract_dir, progress_filter=default_filter,
        drivers=None):
    """Unpack `filename` to `extract_dir`, or raise ``UnrecognizedFormat``

    `progress_filter` is a function taking two arguments: a source path
    internal to the archive ('/'-separated), and a filesystem path where it
    will be extracted.  The callback must return the desired extract path
    (which may be the same as the one passed in), or else ``None`` to skip
    that file or directory.  The callback can thus be used to report on the
    progress of the extraction, as well as to filter the items extracted or
    alter their extraction paths.

    `drivers`, if supplied, must be a non-empty sequence of functions with the
    same signature as this function (minus the `drivers` argument), that raise
    ``UnrecognizedFormat`` if they do not support extracting the designated
    archive type.  The `drivers` are tried in sequence until one is found that
    does not raise an error, or until all are exhausted (in which case
    ``UnrecognizedFormat`` is raised).  If you do not supply a sequence of
    drivers, the module's ``extraction_drivers`` constant will be used, which
    means that ``unpack_zipfile`` and ``unpack_tarfile`` will be tried, in that
    order.
    """
    for driver in drivers or extraction_drivers:
        try:
            driver(filename, extract_dir, progress_filter)
        except UnrecognizedFormat:
            continue
        else:
            return
    else:
        raise UnrecognizedFormat(
            "Not a recognized archive type: %s" % filename
        )


def unpack_directory(filename, extract_dir, progress_filter=default_filter):
    """"Unpack" a directory, using the same interface as for archives

    Raises ``UnrecognizedFormat`` if `filename` is not a directory
    """
    if not os.path.isdir(filename):
        raise UnrecognizedFormat("%s is not a directory" % filename)

    paths = {
        filename: ('', extract_dir),
    }
    for base, dirs, files in os.walk(filename):
        src, dst = paths[base]
        for d in dirs:
            paths[os.path.join(base, d)] = src + d + '/', os.path.join(dst, d)
        for f in files:
            target = os.path.join(dst, f)
            target = progress_filter(src + f, target)
            if not target:
                # skip non-files
                continue
            ensure_directory(target)
            f = os.path.join(base, f)
            shutil.copyfile(f, target)
            shutil.copystat(f, target)


def unpack_zipfile(filename, extract_dir, progress_filter=default_filter):
    """Unpack zip `filename` to `extract_dir`

    Raises ``UnrecognizedFormat`` if `filename` is not a zipfile (as determined
    by ``zipfile.is_zipfile()``).  See ``unpack_archive()`` for an explanation
    of the `progress_filter` argument.
    """

    if not zipfile.is_zipfile(filename):
        raise UnrecognizedFormat("%s is not a zip file" % (filename,))

    with zipfile.ZipFile(filename) as z:
        for info in z.infolist():
            name = info.filename

            # don't extract absolute paths or ones with .. in them
            if name.startswith('/') or '..' in name.split('/'):
                continue

            target = os.path.join(extract_dir, *name.split('/'))
            target = progress_filter(name, target)
            if not target:
                continue
            if name.endswith('/'):
                # directory
                ensure_directory(target)
            else:
                # file
                ensure_directory(target)
                data = z.read(info.filename)
                with open(target, 'wb') as f:
                    f.write(data)
            unix_attributes = info.external_attr >> 16
            if unix_attributes:
                os.chmod(target, unix_attributes)


def _resolve_tar_file_or_dir(tar_obj, tar_member_obj):
    """Resolve any links and extract link targets as normal files."""
    while tar_member_obj is not None and (
            tar_member_obj.islnk() or tar_member_obj.issym()):
        linkpath = tar_member_obj.linkname
        if tar_member_obj.issym():
            base = posixpath.dirname(tar_member_obj.name)
            linkpath = posixpath.join(base, linkpath)
            linkpath = posixpath.normpath(linkpath)
        tar_member_obj = tar_obj._getmember(linkpath)

    is_file_or_dir = (
        tar_member_obj is not None and
        (tar_member_obj.isfile() or tar_member_obj.isdir())
    )
    if is_file_or_dir:
        return tar_member_obj

    raise LookupError('Got unknown file type')


def _iter_open_tar(tar_obj, extract_dir, progress_filter):
    """Emit member-destination pairs from a tar archive."""
    # don't do any chowning!
    tar_obj.chown = lambda *args: None

    with contextlib.closing(tar_obj):
        for member in tar_obj:
            name = member.name
            # don't extract absolute paths or ones with .. in them
            if name.startswith('/') or '..' in name.split('/'):
                continue

            prelim_dst = os.path.join(extract_dir, *name.split('/'))

            try:
                member = _resolve_tar_file_or_dir(tar_obj, member)
            except LookupError:
                continue

            final_dst = progress_filter(name, prelim_dst)
            if not final_dst:
                continue

            if final_dst.endswith(os.sep):
                final_dst = final_dst[:-1]

            yield member, final_dst


def unpack_tarfile(filename, extract_dir, progress_filter=default_filter):
    """Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir`

    Raises ``UnrecognizedFormat`` if `filename` is not a tarfile (as determined
    by ``tarfile.open()``).  See ``unpack_archive()`` for an explanation
    of the `progress_filter` argument.
    """
    try:
        tarobj = tarfile.open(filename)
    except tarfile.TarError as e:
        raise UnrecognizedFormat(
            "%s is not a compressed or uncompressed tar file" % (filename,)
        ) from e

    for member, final_dst in _iter_open_tar(
            tarobj, extract_dir, progress_filter,
    ):
        try:
            # XXX Ugh
            tarobj._extract_member(member, final_dst)
        except tarfile.ExtractError:
            # chown/chmod/mkfifo/mknode/makedev failed
            pass

    return True


extraction_drivers = unpack_directory, unpack_zipfile, unpack_tarfile
site-packages/setuptools/_deprecation_warning.py000064400000000332150564745010016245 0ustar00class SetuptoolsDeprecationWarning(Warning):
    """
    Base class for warning deprecations in ``setuptools``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    """
site-packages/setuptools/__pycache__/unicode_utils.cpython-39.pyc000064400000002212150564745010021220 0ustar00a

(�Re��@s,ddlZddlZdd�Zdd�Zdd�ZdS)�NcCsRt|t�rt�d|�Sz$|�d�}t�d|�}|�d�}WntyLYn0|S)NZNFD�utf-8)�
isinstance�str�unicodedata�	normalize�decode�encode�UnicodeError)�path�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/unicode_utils.py�	decomposes

r
c	CsXt|t�r|St��pd}|df}|D],}z|�|�WStyPYq&Yq&0q&dS)zY
    Ensure that the given path is decoded,
    NONE when no expected encoding works
    rN)rr�sys�getfilesystemencodingr�UnicodeDecodeError)r
Zfs_enc�
candidates�encrrr�filesys_decodes
rcCs&z|�|�WSty YdS0dS)z/turn unicode encoding into a functional routineN)r�UnicodeEncodeError)�stringrrrr�
try_encode%sr)rrr
rrrrrr�<module>ssite-packages/setuptools/__pycache__/py34compat.cpython-39.pyc000064400000001020150564745010020351 0ustar00a

(�Re��@sTddlZzddlZWney&Yn0zejjZWneyNdd�ZYn0dS)�NcCs|j�|j�S)N)�loader�load_module�name)�spec�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/py34compat.py�module_from_specsr)�	importlib�importlib.util�ImportError�utilr�AttributeErrorrrrr�<module>ssite-packages/setuptools/__pycache__/depends.cpython-39.pyc000064400000012261150564745010020001 0ustar00a

(�Reb�@s�ddlZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
ddlmZgd�ZGdd�d�Z
d	d
�Zddd
�Zddd�Zdd�Ze�dS)�N)�
StrictVersion�)�find_module�PY_COMPILED�	PY_FROZEN�	PY_SOURCE)�_imp)�Requirer�get_module_constant�extract_constantc@sHeZdZdZddd�Zdd�Zdd	�Zddd�Zdd
d�Zddd�Z	dS)r	z7A prerequisite to building or installing a distribution�NcCsF|dur|durt}|dur0||�}|dur0d}|j�t��|`dS)N�__version__)r�__dict__�update�locals�self)r�name�requested_version�moduleZhomepage�	attribute�format�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/depends.py�__init__szRequire.__init__cCs |jdurd|j|jfS|jS)z0Return full package/distribution name, w/versionNz%s-%s)rr)rrrr�	full_name"s
zRequire.full_namecCs*|jdup(|jdup(t|�dko(||jkS)z%Is 'version' sufficiently up-to-date?N�unknown)rr�strr)r�versionrrr�
version_ok(s�zRequire.version_okrcCs~|jdurDz$t|j|�\}}}|r*|��|WStyBYdS0t|j|j||�}|durz||urz|jdurz|�|�S|S)a�Get version number of installed module, 'None', or 'default'

        Search 'paths' for module.  If not found, return 'None'.  If found,
        return the extracted version attribute, or 'default' if no version
        attribute was specified, or the value cannot be determined without
        importing the module.  The version is formatted according to the
        requirement's version format (if any), unless it is 'None' or the
        supplied 'default'.
        N)rrr�close�ImportErrorr
r)r�paths�default�f�p�i�vrrr�get_version-s

zRequire.get_versioncCs|�|�duS)z/Return true if dependency is present on 'paths'N)r')rr!rrr�
is_presentHszRequire.is_presentcCs |�|�}|durdS|�|�S)z>Return true if dependency is present and up-to-date on 'paths'NF)r'r)rr!rrrr�
is_currentLs
zRequire.is_current)rNN)Nr)N)N)
�__name__�
__module__�__qualname__�__doc__rrrr'r(r)rrrrr	s�


r	cCs"tjdd��}|s|�St�|�S)Ncss
dVdS)Nrrrrr�emptyUszmaybe_close.<locals>.empty)�
contextlib�contextmanager�closing)r#r.rrr�maybe_closeTs

r2���cCs�zt||�\}}\}}}}	Wnty2YdS0t|���|tkr\|�d�t�|�}
nX|tkrrt�	||�}
nB|t
kr�t|��|d�}
n(t�|||	�}t
||d�Wd�SWd�n1s�0Yt|
||�S)z�Find 'module' by searching 'paths', and extract 'symbol'

    Return 'None' if 'module' does not exist on 'paths', or it does not define
    'symbol'.  If the module defines 'symbol' as a constant, return the
    constant.  Otherwise, return 'default'.N��exec)rr r2r�read�marshal�loadrr�get_frozen_objectr�compileZ
get_module�getattrr)r�symbolr"r!r#�path�suffix�mode�kind�info�codeZimportedrrrr
_s

8r
cCs�||jvrdSt|j��|�}d}d}d}|}t�|�D]H}|j}	|j}
|	|kr\|j|
}q8|
|kr||	|kst|	|kr||S|}q8dS)aExtract the constant value of 'symbol' from 'code'

    If the name 'symbol' is bound to a constant value by the Python code
    object 'code', return that value.  If 'symbol' is bound to an expression,
    return 'default'.  Otherwise, return 'None'.

    Return value is based on the first assignment to 'symbol'.  'symbol' must
    be a global, or at least a non-"fast" local in the code block.  That is,
    only 'STORE_NAME' and 'STORE_GLOBAL' opcodes are checked, and 'symbol'
    must be present in 'code.co_names'.
    N�Z�a�d)�co_names�list�index�dis�Bytecode�opcode�arg�	co_consts)rBr<r"Zname_idx�
STORE_NAME�STORE_GLOBAL�
LOAD_CONST�constZ	byte_code�oprLrrrr|s
rcCs>tj�d�stjdkrdSd}|D]}t�|=t�|�q"dS)z�
    Patch the globals to remove the objects not available on some platforms.

    XXX it'd be better to test assertions about bytecode instead.
    �java�cliN)rr
)�sys�platform�
startswith�globals�__all__�remove)Zincompatiblerrrr�_update_globals�sr[)r3N)r3)rUr7r/rIZdistutils.versionrrrrrrrrYr	r2r
rr[rrrr�<module>sD

$site-packages/setuptools/__pycache__/_deprecation_warning.cpython-39.pyc000064400000001126150564745010022536 0ustar00a

(�Re��@sGdd�de�ZdS)c@seZdZdZdS)�SetuptoolsDeprecationWarningz�
    Base class for warning deprecations in ``setuptools``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_deprecation_warning.pyrsrN)�Warningrrrrr�<module>�site-packages/setuptools/__pycache__/dep_util.cpython-39.pyc000064400000001611150564745010020161 0ustar00a

(�Re��@sddlmZdd�ZdS)�)�newer_groupcCsht|�t|�krtd��g}g}tt|��D]2}t||||�r,|�||�|�||�q,||fS)z�Walk both arguments in parallel, testing if each source group is newer
    than its corresponding target. Returns a pair of lists (sources_groups,
    targets) where sources is newer than target, according to the semantics
    of 'newer_group()'.
    z5'sources_group' and 'targets' must be the same length)�len�
ValueError�ranger�append)Zsources_groups�targets�	n_sources�	n_targets�i�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/dep_util.py�newer_pairwise_groups�r
N)�distutils.dep_utilrr
rrrr�<module>ssite-packages/setuptools/__pycache__/windows_support.cpython-39.pyc000064400000002057150564745010021647 0ustar00a

(�Re��@s(ddlZddlZdd�Zedd��ZdS)�NcCst��dkrdd�S|S)N�Windowsc_sdS)N�)�args�kwargsrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/windows_support.py�<lambda>�zwindows_only.<locals>.<lambda>)�platform�system)�funcrrr�windows_onlysrcCsLtd�tjjj}tjjtjjf|_tjj	|_
d}|||�}|sHt���dS)z�
    Set the hidden attribute on a file or directory.

    From http://stackoverflow.com/questions/19622133/

    `path` must be text.
    zctypes.wintypes�N)�
__import__�ctypes�windllZkernel32ZSetFileAttributesWZwintypesZLPWSTRZDWORD�argtypesZBOOL�restypeZWinError)�pathZSetFileAttributes�FILE_ATTRIBUTE_HIDDEN�retrrr�	hide_files	


r)r	rrrrrrr�<module>ssite-packages/setuptools/__pycache__/version.cpython-39.pyc000064400000000564150564745010020047 0ustar00a

(�Re��@s4ddlZze�d�jZWney.dZYn0dS)�N�
setuptools�unknown)�
pkg_resources�get_distribution�version�__version__�	Exception�r	r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/version.py�<module>ssite-packages/setuptools/__pycache__/monkey.cpython-39.pyc000064400000011065150564745010017662 0ustar00a

(�Rea�@s�dZddlZddlZddlZddlZddlZddlmZddl	Z	ddl
Z
gZdd�Zdd�Z
dd	�Zd
d�Zdd
�Zdd�Zdd�Zdd�ZdS)z
Monkey patching of distutils.
�N)�
import_modulecCs"t��dkr|f|jSt�|�S)am
    Returns the bases classes for cls sorted by the MRO.

    Works around an issue on Jython where inspect.getmro will not return all
    base classes if multiple classes share the same name. Instead, this
    function will return a tuple containing the class itself, and the contents
    of cls.__bases__. See https://github.com/pypa/setuptools/issues/1024.
    �Jython)�platform�python_implementation�	__bases__�inspect�getmro)�cls�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/monkey.py�_get_mros	rcCs.t|t�rtnt|tj�rtndd�}||�S)NcSsdS)Nr
)�itemr
r
r�<lambda>(�zget_unpatched.<locals>.<lambda>)�
isinstance�type�get_unpatched_class�types�FunctionType�get_unpatched_function)r
�lookupr
r
r�
get_unpatched$s
�rcCs:dd�t|�D�}t|�}|j�d�s6d|}t|��|S)z�Protect against re-patching the distutils if reloaded

    Also ensures that no other distutils extension monkeypatched the distutils
    first.
    css|]}|j�d�s|VqdS)�
setuptoolsN)�
__module__�
startswith)�.0r	r
r
r�	<genexpr>3s�z&get_unpatched_class.<locals>.<genexpr>�	distutilsz(distutils has already been patched by %r)r�nextrr�AssertionError)r	Zexternal_bases�base�msgr
r
rr-s�rcCs�tjtj_tjdk}|r"tjtj_tjdkp^dtjko@dknp^dtjkoZdkn}|rrd}|tjj	_
t�tjtjtj
fD]}tjj|_q�tjjtj_tjjtj_dtjvr�tjjtjd_t�dS)N)��r")���
)r"�)r"r'�)r"r#zhttps://upload.pypi.org/legacy/zdistutils.command.build_ext)r�Commandr�core�sys�version_info�findall�filelist�config�
PyPIRCCommand�DEFAULT_REPOSITORY�_patch_distribution_metadata�dist�cmd�Distribution�	extension�	Extension�modules�#patch_for_msvc_specialized_compiler)Zhas_issue_12885Zneeds_warehouseZ	warehouse�moduler
r
r�	patch_all?s*



��

�
r;cCs*dD] }ttj|�}ttjj||�qdS)zDPatch write_pkg_file and read_pkg_file for higher metadata standards)�write_pkg_file�
read_pkg_fileZget_metadata_versionN)�getattrrr3�setattrr�DistributionMetadata)�attr�new_valr
r
rr2fsr2cCs*t||�}t|��d|�t|||�dS)z�
    Patch func_name in target_mod with replacement

    Important - original must be resolved by name to avoid
    patching an already patched function.
    �	unpatchedN)r>�vars�
setdefaultr?)�replacementZ
target_mod�	func_name�originalr
r
r�
patch_funcms
rIcCs
t|d�S)NrC)r>)�	candidater
r
rr~srcs�td��t��dkrdS�fdd�}t�|d�}t�|d�}zt|d��t|d	��WntyjYn0zt|d
��Wnty�Yn0zt|d��Wnty�Yn0dS)z\
    Patch functions in distutils to use standalone Microsoft Visual C++
    compilers.
    zsetuptools.msvc�WindowsNcsLd|vrdnd}||�d�}t�|�}t|�}t||�sBt|��|||fS)zT
        Prepare the parameters for patch_func to patch indicated function.
        �msvc9Zmsvc9_Zmsvc14_�_)�lstripr>r�hasattr�ImportError)�mod_namerGZrepl_prefixZ	repl_name�repl�mod�Zmsvcr
r�patch_params�s

z9patch_for_msvc_specialized_compiler.<locals>.patch_paramszdistutils.msvc9compilerzdistutils._msvccompilerZfind_vcvarsallZquery_vcvarsallZ_get_vc_envZgen_lib_options)rr�system�	functools�partialrIrP)rUrLZmsvc14r
rTrr9�s&
r9)�__doc__r+�distutils.filelistrrrrW�	importlibrrr�__all__rrrr;r2rIrr9r
r
r
r�<module>s"	'site-packages/setuptools/__pycache__/installer.cpython-39.pyc000064400000005403150564745010020354 0ustar00a

(�Re�
�@spddlZddlZddlZddlZddlZddlmZddlmZddl	Z	ddl
mZdd�Zdd�Z
d	d
�ZdS)�N)�log)�DistutilsError)�WheelcCs(t|t�r|��St|ttf�s$J�|S)z8Ensure find-links option end-up being a list of strings.)�
isinstance�str�split�tuple�list)�
find_links�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/installer.py�_fixup_find_links
s
r
cCsVzt�d�Wn"tjy0|�dtj�Yn0t|�}|�d�}d|vrTtd��dt	j
vofdt	j
v}dt	j
vrxd	}nd
|vr�|d
d}nd	}d|vr�t|dd�d	d	�ng}|jr�|�
|j�t	j�|���}t��}t�|�D]}||vr�|�|�r�|Sq�t����,}	tjd
ddddd|	g}
|�r<|
�d�|d	u�rT|
�
d|f�|�p\gD]}|
�
d|f��q^|
�|j�p�t|��zt�|
�Wn6tj�y�}ztt|��|�WYd	}~n
d	}~00tt�t	j� |	d��d�}
t	j� ||
�!��}|
�"|�t�#|t	j� |d��}tj$j%||d�}|Wd	�S1�sH0Yd	S)zLFetch an egg needed for building.

    Use pip/wheel to fetch/build a wheel.�wheelz,WARNING: The wheel package is not available.�easy_installZallow_hostszQthe `allow-hosts` option is not supported when using pip to install requirements.Z	PIP_QUIETZPIP_VERBOSEZ
PIP_INDEX_URLN�	index_url�r
z-m�pipz--disable-pip-version-checkz	--no-depsz-wz--quietz--index-urlz--find-linksz*.whlrzEGG-INFO)�metadata)&�
pkg_resources�get_distribution�DistributionNotFound�announcer�WARN�strip_marker�get_option_dictr�os�environr
Zdependency_links�extend�path�realpathZget_egg_cache_dir�Environment�find_distributions�can_add�tempfile�TemporaryDirectory�sys�
executable�append�urlr�
subprocess�
check_call�CalledProcessErrorr�glob�join�egg_nameZinstall_as_egg�PathMetadata�Distribution�
from_filename)�dist�req�opts�quietrr
Zeggs_dir�environmentZegg_distZtmpdir�cmd�link�er�
dist_locationZ
dist_metadatarrr�fetch_build_eggsf

 �
�

$
��r;cCstj�t|��}d|_|S)z�
    Return a new requirement without the environment marker to avoid
    calling pip with something like `babel; extra == "i18n"`, which
    would always be ignored.
    N)r�Requirement�parser�marker)r3rrrrXsr)r,rr)r%r#�	distutilsr�distutils.errorsrrZsetuptools.wheelrr
r;rrrrr�<module>sCsite-packages/setuptools/__pycache__/archive_util.cpython-39.pyc000064400000013346150564745010021042 0ustar00a

(�Re��@s�dZddlZddlZddlZddlZddlZddlZddlmZddl	m
Z
gd�ZGdd�de�Zdd	�Z
e
dfd
d�Ze
fdd
�Ze
fdd�Zdd�Zdd�Ze
fdd�ZeeefZdS)z/Utilities for extracting common archive formats�N)�DistutilsError)�ensure_directory)�unpack_archive�unpack_zipfile�unpack_tarfile�default_filter�UnrecognizedFormat�extraction_drivers�unpack_directoryc@seZdZdZdS)rz#Couldn't recognize the archive typeN)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/archive_util.pyrsrcCs|S)z@The default progress/filter callback; returns True for all filesr)�src�dstrrrrsrc	CsL|ptD]2}z||||�Wnty2YqYq0dSqtd|��dS)a�Unpack `filename` to `extract_dir`, or raise ``UnrecognizedFormat``

    `progress_filter` is a function taking two arguments: a source path
    internal to the archive ('/'-separated), and a filesystem path where it
    will be extracted.  The callback must return the desired extract path
    (which may be the same as the one passed in), or else ``None`` to skip
    that file or directory.  The callback can thus be used to report on the
    progress of the extraction, as well as to filter the items extracted or
    alter their extraction paths.

    `drivers`, if supplied, must be a non-empty sequence of functions with the
    same signature as this function (minus the `drivers` argument), that raise
    ``UnrecognizedFormat`` if they do not support extracting the designated
    archive type.  The `drivers` are tried in sequence until one is found that
    does not raise an error, or until all are exhausted (in which case
    ``UnrecognizedFormat`` is raised).  If you do not supply a sequence of
    drivers, the module's ``extraction_drivers`` constant will be used, which
    means that ``unpack_zipfile`` and ``unpack_tarfile`` will be tried, in that
    order.
    Nz!Not a recognized archive type: %s)r	r)�filename�extract_dir�progress_filterZdriversZdriverrrrrs
�rcCs�tj�|�std|��|d|fi}t�|�D]�\}}}||\}}|D],}	||	dtj�||	�f|tj�||	�<qH|D]T}
tj�||
�}|||
|�}|s�qzt|�tj�||
�}
t�|
|�t�	|
|�qzq.dS)z�"Unpack" a directory, using the same interface as for archives

    Raises ``UnrecognizedFormat`` if `filename` is not a directory
    z%s is not a directory��/N)
�os�path�isdirr�walk�joinr�shutil�copyfile�copystat)rrr�paths�base�dirs�filesrr�d�f�targetrrrr
@s"�*r
c

Cst�|�std|f��t�|���}|��D]�}|j}|�d�s,d|�d�vrPq,tj	j
|g|�d��R�}|||�}|szq,|�d�r�t|�nHt|�|�
|j�}t|d��}|�|�Wd�n1s�0Y|jd?}	|	r,t�||	�q,Wd�n1�s0YdS)z�Unpack zip `filename` to `extract_dir`

    Raises ``UnrecognizedFormat`` if `filename` is not a zipfile (as determined
    by ``zipfile.is_zipfile()``).  See ``unpack_archive()`` for an explanation
    of the `progress_filter` argument.
    z%s is not a zip filer�..�wbN�)�zipfile�
is_zipfiler�ZipFile�infolistr�
startswith�splitrrr�endswithr�read�open�write�
external_attr�chmod)
rrr�z�info�namer&�datar%Zunix_attributesrrrr[s(



(
rcCs�|durT|��s|��rT|j}|��rHt�|j�}t�||�}t�|�}|�|�}q|duoj|�	�pj|�
�}|rt|Std��dS)z;Resolve any links and extract link targets as normal files.NzGot unknown file type)�islnk�issym�linkname�	posixpath�dirnamer8r�normpath�
_getmember�isfiler�LookupError)�tar_objZtar_member_obj�linkpathr!Zis_file_or_dirrrr�_resolve_tar_file_or_dir�s"��
�rEc
cs�dd�|_t�|���|D]�}|j}|�d�sd|�d�vr>qtjj|g|�d��R�}zt	||�}Wnt
y|YqYn0|||�}|s�q|�tj�r�|dd�}||fVqWd�n1s�0YdS)z1Emit member-destination pairs from a tar archive.cWsdS)Nr)�argsrrr�<lambda>��z _iter_open_tar.<locals>.<lambda>rr'N���)
�chown�
contextlib�closingr8r.r/rrrrErBr0�sep)rCrr�memberr8Z
prelim_dst�	final_dstrrr�_iter_open_tar�s"


rPc
Cs�zt�|�}Wn6tjyD}ztd|f�|�WYd}~n
d}~00t|||�D].\}}z|�||�WqRtjy~YqR0qRdS)z�Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir`

    Raises ``UnrecognizedFormat`` if `filename` is not a tarfile (as determined
    by ``tarfile.open()``).  See ``unpack_archive()`` for an explanation
    of the `progress_filter` argument.
    z/%s is not a compressed or uncompressed tar fileNT)�tarfiler2�TarErrorrrP�_extract_member�ExtractError)rrr�tarobj�erNrOrrrr�s ���r)rr*rQrrr=rK�distutils.errorsr�
pkg_resourcesr�__all__rrrr
rrErPrr	rrrr�<module>s(�
$%site-packages/setuptools/__pycache__/msvc.cpython-39.pyc000064400000123607150564745010017336 0ustar00a

(�Re���@s�dZddlZddlmZddlmZmZddlmZm	Z	m
Z
mZddlZddl
Z
ddlZddlZddlZddlZddlmZddlmZdd	lmZe��d
kr�ddlZddlmZnGdd
�d
�Ze�ZeejjfZ zddl!m"Z"Wne y�Yn0dd�Z#d/dd�Z$dd�Z%dd�Z&ddddd�Z'dd�Z(dd�Z)d d!�Z*d"d#�Z+d0d%d&�Z,Gd'd(�d(�Z-Gd)d*�d*�Z.Gd+d,�d,�Z/Gd-d.�d.�Z0dS)1a�
Improved support for Microsoft Visual C++ compilers.

Known supported compilers:
--------------------------
Microsoft Visual C++ 9.0:
    Microsoft Visual C++ Compiler for Python 2.7 (x86, amd64)
    Microsoft Windows SDK 6.1 (x86, x64, ia64)
    Microsoft Windows SDK 7.0 (x86, x64, ia64)

Microsoft Visual C++ 10.0:
    Microsoft Windows SDK 7.1 (x86, x64, ia64)

Microsoft Visual C++ 14.X:
    Microsoft Visual C++ Build Tools 2015 (x86, x64, arm)
    Microsoft Visual Studio Build Tools 2017 (x86, x64, arm, arm64)
    Microsoft Visual Studio Build Tools 2019 (x86, x64, arm, arm64)

This may also support compilers shipped with compatible Visual Studio versions.
�N)�open)�listdir�pathsep)�join�isfile�isdir�dirname)�
LegacyVersion)�unique_everseen�)�
get_unpatched�Windows)�environc@seZdZdZdZdZdZdS)�winregN)�__name__�
__module__�__qualname__�
HKEY_USERS�HKEY_CURRENT_USER�HKEY_LOCAL_MACHINE�HKEY_CLASSES_ROOT�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/msvc.pyr+sr)�RegcCs�d}|d|f}zt�|d�}WnFtyfz|d|f}t�|d�}Wnty`d}Yn0Yn0|r�t|d�}t|�r�|Stt�|�S)a�
    Patched "distutils.msvc9compiler.find_vcvarsall" to use the standalone
    compiler build for Python
    (VCForPython / Microsoft Visual C++ Compiler for Python 2.7).

    Fall back to original behavior when the standalone compiler is not
    available.

    Redirect the path of "vcvarsall.bat".

    Parameters
    ----------
    version: float
        Required Microsoft Visual C++ version.

    Return
    ------
    str
        vcvarsall.bat path
    z-Software\%sMicrosoft\DevDiv\VCForPython\%0.1f��
installdirzWow6432Node\N�
vcvarsall.bat)r�	get_value�KeyErrorrrr�msvc9_find_vcvarsall)�versionZvc_base�key�
productdir�	vcvarsallrrrrBs
r�x86c
Os�z"tt�}|||g|�Ri|��WStjjy8YntyHYn0zt||���WStjjy�}zt|||��WYd}~n
d}~00dS)ao
    Patched "distutils.msvc9compiler.query_vcvarsall" for support extra
    Microsoft Visual C++ 9.0 and 10.0 compilers.

    Set environment without use of "vcvarsall.bat".

    Parameters
    ----------
    ver: float
        Required Microsoft Visual C++ version.
    arch: str
        Target architecture.

    Return
    ------
    dict
        environment
    N)	r�msvc9_query_vcvarsall�	distutils�errors�DistutilsPlatformError�
ValueError�EnvironmentInfo�
return_env�_augment_exception)�ver�arch�args�kwargs�orig�excrrrr%lsr%cCszt�tjddtjtjB�}Wnty2YdS0d}d}|��t��D]�}zt�||�\}}}Wnty|Yq�Yn0|rJ|tj	krJt
|�rJztt|��}Wnt
tfy�YqJYn0|dkrJ||krJ||}}qJWd�n1s�0Y||fS)�0Python 3.8 "distutils/_msvccompiler.py" backportz'Software\Microsoft\VisualStudio\SxS\VC7r�NNN�)r�OpenKeyr�KEY_READZKEY_WOW64_32KEY�OSError�	itertools�count�	EnumValue�REG_SZr�int�floatr)�	TypeError)r!�best_version�best_dir�i�vZvc_dirZvtr rrr�_msvc14_find_vc2015�s2
�
*rDcCs�t�d�pt�d�}|sdSz>t�t|ddd�ddd	d
dd
dd
dddg�jddd���}Wntjtt	fyvYdS0t|ddd�}t
|�r�d|fSdS)a�Python 3.8 "distutils/_msvccompiler.py" backport

    Returns "15, path" based on the result of invoking vswhere.exe
    If no install is found, returns "None, None"

    The version is returned to avoid unnecessarily changing the function
    result. It may be ignored when the path is not None.

    If vswhere.exe is not available, by definition, VS 2017 is not
    installed.
    �ProgramFiles(x86)�ProgramFilesr4zMicrosoft Visual StudioZ	Installerzvswhere.exez-latestz-prereleasez-requiresAnyz	-requiresz1Microsoft.VisualStudio.Component.VC.Tools.x86.x64z)Microsoft.VisualStudio.Workload.WDExpressz	-property�installationPathz	-products�*�mbcs�strict)�encodingr'�VCZ	AuxiliaryZBuild�)r�get�
subprocess�check_outputr�decode�strip�CalledProcessErrorr8�UnicodeDecodeErrorr)�root�pathrrr�_msvc14_find_vc2017�s,�	�
rW�x64�arm�arm64)r$Z	x86_amd64Zx86_armZ	x86_arm64c	
Cs�t�\}}d}|tvr t|}nd|vr,dnd}|r�t|ddddd|d	d
�	}zddl}|j|dd
�d}Wntttfy�d}Yn0|s�t�\}}|r�t|d|dd
�}|s�dSt|d�}t|�s�dS|r�t|�s�d}||fS)r3N�amd64rXr$z..�redistZMSVCz**zMicrosoft.VC14*.CRTzvcruntime140.dllrT)�	recursive���zMicrosoft.VC140.CRTr4r)	rW�PLAT_SPEC_TO_RUNTIMEr�glob�ImportErrorr8�LookupErrorrDr)	�	plat_spec�_rA�	vcruntimeZvcruntime_platZvcredistr`r@r#rrr�_msvc14_find_vcvarsall�s:

�

�
rfc
Cs�dtvrdd�t��D�St|�\}}|s6tj�d��z&tjd�||�tj	d�j
ddd	�}Wn<tjy�}z"tj�d
�|j��|�WYd}~n
d}~00dd�d
d�|�
�D�D�}|r�||d<|S)r3ZDISTUTILS_USE_SDKcSsi|]\}}|��|�qSr��lower)�.0r!�valuerrr�
<dictcomp>s�z&_msvc14_get_vc_env.<locals>.<dictcomp>zUnable to find vcvarsall.batzcmd /u /c "{}" {} && set)�stderrzutf-16le�replace)r'zError executing {}NcSs$i|]\}}}|r|r|��|�qSrrg)rir!rdrjrrrrks�css|]}|�d�VqdS)�=N)�	partition)ri�linerrr�	<genexpr>�z%_msvc14_get_vc_env.<locals>.<genexpr>�py_vcruntime_redist)r�itemsrfr&r'r(rOrP�format�STDOUTrQrS�cmd�
splitlines)rcr#re�outr2�envrrr�_msvc14_get_vc_envs8��
��

���r{c
CsDz
t|�WStjjy>}zt|d��WYd}~n
d}~00dS)a*
    Patched "distutils._msvccompiler._get_vc_env" for support extra
    Microsoft Visual C++ 14.X compilers.

    Set environment without use of "vcvarsall.bat".

    Parameters
    ----------
    plat_spec: str
        Target architecture.

    Return
    ------
    dict
        environment
    �,@N)r{r&r'r(r,)rcr2rrr�msvc14_get_vc_env(s


r}cOsJdtjvr8ddl}t|j�td�kr8|jjj|i|��Stt	�|i|��S)z�
    Patched "distutils._msvccompiler.gen_lib_options" for fix
    compatibility between "numpy.distutils" and "distutils._msvccompiler"
    (for Numpy < 1.11.2)
    znumpy.distutilsrNz1.11.2)
�sys�modulesZnumpyr	�__version__r&Z	ccompilerZgen_lib_optionsr�msvc14_gen_lib_options)r/r0�nprrrr�Bs

r�rcCs�|jd}d|��vs"d|��vr�d}|jfit���}d}|dkrj|���d�dkr`|d	7}q�|d
7}n.|dkr�|d7}||d
7}n|dkr�|d7}|f|_dS)zl
    Add details to the exception message to help guide the user
    as to what action will resolve it.
    rr#zvisual cz;Microsoft Visual C++ {version:0.1f} or greater is required.z-www.microsoft.com/download/details.aspx?id=%d�"@Zia64r^z( Get it with "Microsoft Windows SDK 7.0"z% Get it from http://aka.ms/vcpython27�$@z* Get it with "Microsoft Windows SDK 7.1": iW r|zd Get it with "Microsoft C++ Build Tools": https://visualstudio.microsoft.com/visual-cpp-build-tools/N)r/rhru�locals�find)r2r r.�message�tmplZ
msdownloadrrrr,Os


r,c@sbeZdZdZe�dd���Zdd�Ze	dd��Z
dd	�Zd
d�Zdd
d�Z
ddd�Zddd�ZdS)�PlatformInfoz�
    Current and Target Architectures information.

    Parameters
    ----------
    arch: str
        Target architecture.
    Zprocessor_architecturercCs|���dd�|_dS)NrXr[)rhrmr.)�selfr.rrr�__init__szPlatformInfo.__init__cCs|j|j�d�dd�S)zs
        Return Target CPU architecture.

        Return
        ------
        str
            Target CPU
        rdrN)r.r��r�rrr�
target_cpu�s
zPlatformInfo.target_cpucCs
|jdkS)z�
        Return True if target CPU is x86 32 bits..

        Return
        ------
        bool
            CPU is x86 32 bits
        r$�r�r�rrr�
target_is_x86�s	zPlatformInfo.target_is_x86cCs
|jdkS)z�
        Return True if current CPU is x86 32 bits..

        Return
        ------
        bool
            CPU is x86 32 bits
        r$��current_cpur�rrr�current_is_x86�s	zPlatformInfo.current_is_x86FcCs.|jdkr|rdS|jdkr$|r$dSd|jS)uk
        Current platform specific subfolder.

        Parameters
        ----------
        hidex86: bool
            return '' and not '†' if architecture is x86.
        x64: bool
            return 'd' and not 'md64' if architecture is amd64.

        Return
        ------
        str
            subfolder: '	arget', or '' (see hidex86 parameter)
        r$rr[�\x64�\%sr��r��hidex86rXrrr�current_dir�s�zPlatformInfo.current_dircCs.|jdkr|rdS|jdkr$|r$dSd|jS)ar
        Target platform specific subfolder.

        Parameters
        ----------
        hidex86: bool
            return '' and not '\x86' if architecture is x86.
        x64: bool
            return '\x64' and not '\amd64' if architecture is amd64.

        Return
        ------
        str
            subfolder: '\current', or '' (see hidex86 parameter)
        r$rr[r�r�r�r�rrr�
target_dir�s�zPlatformInfo.target_dircCs0|rdn|j}|j|krdS|���dd|�S)ap
        Cross platform specific subfolder.

        Parameters
        ----------
        forcex86: bool
            Use 'x86' as current architecture even if current architecture is
            not x86.

        Return
        ------
        str
            subfolder: '' if target architecture is current architecture,
            '\current_target' if not.
        r$r�\z\%s_)r�r�r�rm)r��forcex86�currentrrr�	cross_dir�s�zPlatformInfo.cross_dirN)FF)FF)F)rrr�__doc__rrNrhr�r��propertyr�r�r�r�r�r�rrrrr�ts


r�c@s�eZdZdZejejejejfZ	dd�Z
edd��Zedd��Z
edd	��Zed
d��Zedd
��Zedd��Zedd��Zedd��Zedd��Zddd�Zdd�ZdS)�RegistryInfoz�
    Microsoft Visual Studio related registry information.

    Parameters
    ----------
    platform_info: PlatformInfo
        "PlatformInfo" instance.
    cCs
||_dS�N)�pi)r�Z
platform_inforrrr��szRegistryInfo.__init__cCsdS)z�
        Microsoft Visual Studio root registry key.

        Return
        ------
        str
            Registry key
        ZVisualStudiorr�rrr�visualstudio�s
zRegistryInfo.visualstudiocCst|jd�S)z
        Microsoft Visual Studio SxS registry key.

        Return
        ------
        str
            Registry key
        ZSxS)rr�r�rrr�sxss
zRegistryInfo.sxscCst|jd�S)z|
        Microsoft Visual C++ VC7 registry key.

        Return
        ------
        str
            Registry key
        ZVC7�rr�r�rrr�vcs
zRegistryInfo.vccCst|jd�S)z
        Microsoft Visual Studio VS7 registry key.

        Return
        ------
        str
            Registry key
        ZVS7r�r�rrr�vss
zRegistryInfo.vscCsdS)z�
        Microsoft Visual C++ for Python registry key.

        Return
        ------
        str
            Registry key
        zDevDiv\VCForPythonrr�rrr�
vc_for_python(s
zRegistryInfo.vc_for_pythoncCsdS)zq
        Microsoft SDK registry key.

        Return
        ------
        str
            Registry key
        zMicrosoft SDKsrr�rrr�
microsoft_sdk4s
zRegistryInfo.microsoft_sdkcCst|jd�S)z�
        Microsoft Windows/Platform SDK registry key.

        Return
        ------
        str
            Registry key
        r
�rr�r�rrr�windows_sdk@s
zRegistryInfo.windows_sdkcCst|jd�S)z�
        Microsoft .NET Framework SDK registry key.

        Return
        ------
        str
            Registry key
        ZNETFXSDKr�r�rrr�	netfx_sdkLs
zRegistryInfo.netfx_sdkcCsdS)z�
        Microsoft Windows Kits Roots registry key.

        Return
        ------
        str
            Registry key
        zWindows Kits\Installed Rootsrr�rrr�windows_kits_rootsXs
zRegistryInfo.windows_kits_rootsFcCs$|j��s|rdnd}td|d|�S)a
        Return key in Microsoft software registry.

        Parameters
        ----------
        key: str
            Registry key path where look.
        x86: str
            Force x86 software registry.

        Return
        ------
        str
            Registry key
        rZWow6432NodeZSoftware�	Microsoft)r�r�r)r�r!r$Znode64rrr�	microsoftdszRegistryInfo.microsoftc	
Cs�tj}tj}tj}|j}|jD]�}d}z||||�d|�}Wn\ttfy�|j�	�s�z||||d�d|�}Wq�ttfy�YYqYq�0nYqYn0zLz$t�
||�dWW|r�||�Sttfy�Yn0W|r�||�q|r�||�0qdS)a
        Look for values in registry in Microsoft software registry.

        Parameters
        ----------
        key: str
            Registry key path where look.
        name: str
            Value name to find.

        Return
        ------
        str
            value
        NrT)rr7r6ZCloseKeyr��HKEYSr8�IOErrorr�r��QueryValueEx)	r�r!�nameZkey_readZopenkeyZclosekey�ms�hkey�bkeyrrr�lookupws4


�
�zRegistryInfo.lookupN)F)rrrr�rrrrrr�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr��s6�









r�c@s<eZdZdZe�dd�Ze�dd�Ze�de�Zd7dd�Z	d	d
�Z
dd�Zd
d�Ze
dd��Zedd��Zedd��Zdd�Zdd�Zedd��Zedd��Zedd��Zedd ��Zed!d"��Zed#d$��Zed%d&��Zed'd(��Zed)d*��Zed+d,��Zed-d.��Zed/d0��Zed1d2��Z d3d4�Z!e
d8d5d6��Z"dS)9�
SystemInfoz�
    Microsoft Windows and Visual Studio related system information.

    Parameters
    ----------
    registry_info: RegistryInfo
        "RegistryInfo" instance.
    vc_ver: float
        Required Microsoft Visual C++ version.
    �WinDirrrFrENcCs2||_|jj|_|��|_|p$|��|_|_dSr�)�rir��find_programdata_vs_vers�known_vs_paths�_find_latest_available_vs_ver�vs_ver�vc_ver)r�Z
registry_infor�rrrr��s



�zSystemInfo.__init__cCs>|��}|s|jstj�d��t|�}|�|j�t|�dS)zm
        Find the latest VC version

        Return
        ------
        float
            version
        z%No Microsoft Visual C++ version foundr^)�find_reg_vs_versr�r&r'r(�set�update�sorted)r�Zreg_vc_versZvc_versrrrr��s	
�z(SystemInfo._find_latest_available_vs_vercCsn|jj}|jj|jj|jjf}g}t�|jj|�D�]0\}}zt�	|||�dtj
�}WnttfypYq2Yn0|��t�
|�\}}}	t|�D]T}
t�t��6tt�||
�d�}||vr�|�|�Wd�q�1s�0Yq�t|�D]T}
t�t��4tt�||
��}||v�r"|�|�Wd�q�1�s80Yq�Wd�q21�sZ0Yq2t|�S)z�
        Find Microsoft Visual Studio versions available in registry.

        Return
        ------
        list of float
            Versions
        rN)r�r�r�r�r�r9�productr�rr6r7r8r�ZQueryInfoKey�range�
contextlib�suppressr)r>r;�append�EnumKeyr�)r�r�ZvckeysZvs_versr�r!r�Zsubkeys�valuesrdrBr-rrrr��s*	
*
NzSystemInfo.find_reg_vs_versc	Cs�i}d}zt|�}Wnttfy.|YS0|D]�}zpt||d�}t|ddd��}t�|�}Wd�n1st0Y|d}tt|d��|||�|d	�<Wq4tttfy�Yq4Yq40q4|S)
z�
        Find Visual studio 2017+ versions from information in
        "C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances".

        Return
        ------
        dict
            float version as key, path as value.
        z9C:\ProgramData\Microsoft\VisualStudio\Packages\_Instancesz
state.json�rtzutf-8)rKNrG�
VC\Tools\MSVCZinstallationVersion)	rr8r�rr�json�load�_as_float_versionr)	r�Zvs_versionsZ
instances_dirZhashed_namesr�Z
state_pathZ
state_file�stateZvs_pathrrrr��s*
�
(��z#SystemInfo.find_programdata_vs_verscCstd�|�d�dd���S)z�
        Return a string version as a simplified float version (major.minor)

        Parameters
        ----------
        version: str
            Version.

        Return
        ------
        float
            version
        �.N�)r>r�split)r rrrr�szSystemInfo._as_float_versioncCs.t|jd|j�}|j�|jjd|j�p,|S)zp
        Microsoft Visual Studio directory.

        Return
        ------
        str
            path
        zMicrosoft Visual Studio %0.1f�%0.1f)r�ProgramFilesx86r�r�r�r�)r��defaultrrr�VSInstallDir)s�zSystemInfo.VSInstallDircCs,|��p|��}t|�s(d}tj�|��|S)zm
        Microsoft Visual C++ directory.

        Return
        ------
        str
            path
        z(Microsoft Visual C++ directory not found)�	_guess_vc�_guess_vc_legacyrr&r'r()r�rV�msgrrr�VCInstallDir:s

zSystemInfo.VCInstallDirc
Cs�|jdkrdSz|j|j}Wnty6|j}Yn0t|d�}z$t|�d}|�|�|_t||�WStt	t
fy�YdS0dS)zl
        Locate Visual C++ for VS2017+.

        Return
        ------
        str
            path
        r|rr�r^N)r�r�rr�rrr�r�r8r��
IndexError)r�Zvs_dirZguess_vcr�rrrr�Ls	

zSystemInfo._guess_vccCsbt|jd|j�}t|jjd|j�}|j�|d�}|rBt|d�n|}|j�|jjd|j�p`|S)z{
        Locate Visual C++ for versions prior to 2017.

        Return
        ------
        str
            path
        z Microsoft Visual Studio %0.1f\VCr�rrL)rr�r�r�r�r�r�)r�r�Zreg_pathZ	python_vcZ
default_vcrrrr�js	�zSystemInfo._guess_vc_legacycCsJ|jdkrdS|jdkrdS|jdkr*dS|jdkr8dS|jd	krFd
SdS)z�
        Microsoft Windows SDK versions for specified MSVC++ version.

        Return
        ------
        tuple of str
            versions
        r�)z7.0z6.1z6.0ar�)z7.1z7.0a�&@)z8.0z8.0a�(@)�8.1z8.1ar|)z10.0r�N�r�r�rrr�WindowsSdkVersion~s





zSystemInfo.WindowsSdkVersioncCs|�t|jd��S)zt
        Microsoft Windows SDK last version.

        Return
        ------
        str
            version
        �lib)�_use_last_dir_namer�
WindowsSdkDirr�rrr�WindowsSdkLastVersion�s
z SystemInfo.WindowsSdkLastVersioncCsd}|jD],}t|jjd|�}|j�|d�}|r
q8q
|rDt|�stt|jjd|j�}|j�|d�}|rtt|d�}|r�t|�s�|jD]6}|d|�d��}d	|}t|j	|�}t|�r�|}q�|r�t|�s�|jD]$}d
|}t|j	|�}t|�r�|}q�|�st|j
d�}|S)zn
        Microsoft Windows SDK directory.

        Return
        ------
        str
            path
        rzv%s�installationfolderr�rZWinSDKNr�zMicrosoft SDKs\Windows Kits\%szMicrosoft SDKs\Windows\v%sZPlatformSDK)r�rr�r�r�rr�r��rfindrFr�)r��sdkdirr-�locrV�install_baseZintver�drrrr��s6




zSystemInfo.WindowsSdkDirc	Cs�|jdkrd}d}n&d}|jdkr&dnd}|jjd|d�}d	||�d
d�f}g}|jdkr~|jD]}|t|jj||�g7}qb|jD]}|t|jj	d
||�g7}q�|D]}|j�
|d�}|r�|Sq�dS)zy
        Microsoft Windows SDK executable directory.

        Return
        ------
        str
            path
        r��#r�(r�TF)rXr�zWinSDK-NetFx%dTools%sr��-r|zv%sAr�N)r�r�r�rm�NetFxSdkVersionrr�r�r�r�r�)	r�Znetfxverr.r�ZfxZregpathsr-rVZexecpathrrr�WindowsSDKExecutablePath�s"



z#SystemInfo.WindowsSDKExecutablePathcCs&t|jjd|j�}|j�|d�p$dS)zl
        Microsoft Visual F# directory.

        Return
        ------
        str
            path
        z%0.1f\Setup\F#r"r)rr�r�r�r�)r�rVrrr�FSharpInstallDir�s
zSystemInfo.FSharpInstallDircCsF|jdkrdnd}|D]*}|j�|jjd|�}|r|p:dSqdS)zt
        Microsoft Universal CRT SDK directory.

        Return
        ------
        str
            path
        r|)�10Z81rz
kitsroot%srN)r�r�r�r�)r��versr-r�rrr�UniversalCRTSdkDir�s�zSystemInfo.UniversalCRTSdkDircCs|�t|jd��S)z�
        Microsoft Universal C Runtime SDK last version.

        Return
        ------
        str
            version
        r�)r�rr�r�rrr�UniversalCRTSdkLastVersions
z%SystemInfo.UniversalCRTSdkLastVersioncCs|jdkrdSdS)z�
        Microsoft .NET Framework SDK versions.

        Return
        ------
        tuple of str
            versions
        r|)	z4.7.2z4.7.1z4.7z4.6.2z4.6.1z4.6z4.5.2z4.5.1z4.5rr�r�rrrr�s��zSystemInfo.NetFxSdkVersioncCs8d}|jD](}t|jj|�}|j�|d�}|r
q4q
|S)zu
        Microsoft .NET Framework SDK directory.

        Return
        ------
        str
            path
        rZkitsinstallationfolder)r�rr�r�r�)r�r�r-r�rrr�NetFxSdkDir*s

zSystemInfo.NetFxSdkDircCs"t|jd�}|j�|jjd�p |S)zw
        Microsoft .NET Framework 32bit directory.

        Return
        ------
        str
            path
        zMicrosoft.NET\FrameworkZframeworkdir32�rr�r�r�r��r�Zguess_fwrrr�FrameworkDir32<szSystemInfo.FrameworkDir32cCs"t|jd�}|j�|jjd�p |S)zw
        Microsoft .NET Framework 64bit directory.

        Return
        ------
        str
            path
        zMicrosoft.NET\Framework64Zframeworkdir64r�r�rrr�FrameworkDir64LszSystemInfo.FrameworkDir64cCs
|�d�S)z�
        Microsoft .NET Framework 32bit versions.

        Return
        ------
        tuple of str
            versions
        � ��_find_dot_net_versionsr�rrr�FrameworkVersion32\s
zSystemInfo.FrameworkVersion32cCs
|�d�S)z�
        Microsoft .NET Framework 64bit versions.

        Return
        ------
        tuple of str
            versions
        �@r�r�rrr�FrameworkVersion64hs
zSystemInfo.FrameworkVersion64cCs�|j�|jjd|�}t|d|�}|p6|�|d�p6d}|jdkrJ|dfS|jdkrt|��dd	�d
krldn|dfS|jd
kr�dS|jdkr�dSdS)z�
        Find Microsoft .NET Framework versions.

        Parameters
        ----------
        bits: int
            Platform number of bits: 32 or 64.

        Return
        ------
        tuple of str
            versions
        zframeworkver%dzFrameworkDir%drCrr�zv4.0r�Nr�Zv4z
v4.0.30319�v3.5r�)r�
v2.0.50727g @)zv3.0r)r�r�r��getattrr�r�rh)r��bitsZreg_verZdot_net_dirr-rrrr�ts

 

z!SystemInfo._find_dot_net_versionscs*��fdd�tt���D�}t|d�p(dS)a)
        Return name of the last dir in path or '' if no dir found.

        Parameters
        ----------
        path: str
            Use dirs in this path
        prefix: str
            Use only dirs starting by this prefix

        Return
        ------
        str
            name
        c3s*|]"}tt�|��r|���r|VqdSr�)rr�
startswith)ri�dir_name�rV�prefixrrrq�s
�z0SystemInfo._use_last_dir_name.<locals>.<genexpr>Nr)�reversedr�next)rVrZ
matching_dirsrrrr��s
�zSystemInfo._use_last_dir_name)N)r)#rrrr�rrNr�rFr�r�r�r�r��staticmethodr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr��sZ


*





*
"








r�c@sTeZdZdZd=dd�Zedd��Zedd	��Zed
d��Zedd
��Z	edd��Z
edd��Zedd��Zedd��Z
edd��Zedd��Zedd��Zdd�Zedd��Zed d!��Zed"d#��Zed$d%��Zed&d'��Zed(d)��Zed*d+��Zed,d-��Zed.d/��Zed0d1��Zed2d3��Zed4d5��Zed6d7��Zd>d9d:�Zd;d<�Z dS)?r*aY
    Return environment variables for specified Microsoft Visual C++ version
    and platform : Lib, Include, Path and libpath.

    This function is compatible with Microsoft Visual C++ 9.0 to 14.X.

    Script created by analysing Microsoft environment configuration files like
    "vcvars[...].bat", "SetEnv.Cmd", "vcbuildtools.bat", ...

    Parameters
    ----------
    arch: str
        Target architecture.
    vc_ver: float
        Required Microsoft Visual C++ version. If not set, autodetect the last
        version.
    vc_min_ver: float
        Minimum Microsoft Visual C++ version.
    NrcCsBt|�|_t|j�|_t|j|�|_|j|kr>d}tj�	|��dS)Nz.No suitable Microsoft Visual C++ version found)
r�r�r�r�r��sir�r&r'r()r�r.r�Z
vc_min_ver�errrrrr��s

zEnvironmentInfo.__init__cCs|jjS)zk
        Microsoft Visual Studio.

        Return
        ------
        float
            version
        )rr�r�rrrr��s
zEnvironmentInfo.vs_vercCs|jjS)zp
        Microsoft Visual C++ version.

        Return
        ------
        float
            version
        )rr�r�rrrr��s
zEnvironmentInfo.vc_vercsVddg}�jdkrD�jjddd�}|dg7}|dg7}|d|g7}�fd	d
�|D�S)zu
        Microsoft Visual Studio Tools.

        Return
        ------
        list of str
            paths
        zCommon7\IDEz
Common7\Toolsr|T�r�rXz1Common7\IDE\CommonExtensions\Microsoft\TestWindowzTeam Tools\Performance ToolszTeam Tools\Performance Tools%scsg|]}t�jj|��qSr�rrr��rirVr�rr�
<listcomp>�rrz+EnvironmentInfo.VSTools.<locals>.<listcomp>)r�r�r�)r��paths�arch_subdirrr�r�VSTools�s



zEnvironmentInfo.VSToolscCst|jjd�t|jjd�gS)z�
        Microsoft Visual C++ & Microsoft Foundation Class Includes.

        Return
        ------
        list of str
            paths
        �IncludezATLMFC\Include�rrr�r�rrr�
VCIncludes�s
�zEnvironmentInfo.VCIncludescsb�jdkr�jjdd�}n�jjdd�}d|d|g}�jdkrP|d|g7}�fd	d
�|D�S)z�
        Microsoft Visual C++ & Microsoft Foundation Class Libraries.

        Return
        ------
        list of str
            paths
        �.@T�rX�r��Lib%szATLMFC\Lib%sr|zLib\store%scsg|]}t�jj|��qSrrrr�rrrrrz/EnvironmentInfo.VCLibraries.<locals>.<listcomp>)r�r�r�)r�rrrr�r�VCLibrariess


zEnvironmentInfo.VCLibrariescCs|jdkrgSt|jjd�gS)z�
        Microsoft Visual C++ store references Libraries.

        Return
        ------
        list of str
            paths
        r|zLib\store\references)r�rrr�r�rrr�VCStoreRefss

zEnvironmentInfo.VCStoreRefscCs|j}t|jd�g}|jdkr"dnd}|j�|�}|rL|t|jd|�g7}|jdkr|d|jjdd�}|t|j|�g7}n�|jdkr�|j��r�d	nd
}|t|j||jjdd��g7}|jj	|jj
kr�|t|j||jjdd��g7}n|t|jd�g7}|S)
zr
        Microsoft Visual C++ Tools.

        Return
        ------
        list of str
            paths
        Z
VCPackagesr�TF�Bin%sr|rrz
bin\HostX86%sz
bin\HostX64%sr�Bin)rrr�r�r�r�r�r�r�r�r�)r�r�toolsr�rrVZhost_dirrrr�VCTools(s,


���
zEnvironmentInfo.VCToolscCsh|jdkr.|jjddd�}t|jjd|�gS|jjdd�}t|jjd�}|j}t|d||f�gSdS)	zw
        Microsoft Windows SDK Libraries.

        Return
        ------
        list of str
            paths
        r�Tr
rrr�z%sum%sN)r�r�r�rrr��_sdk_subdir)r�rr�Zlibverrrr�OSLibrariesMs

zEnvironmentInfo.OSLibrariescCsht|jjd�}|jdkr&|t|d�gS|jdkr8|j}nd}t|d|�t|d|�t|d|�gSd	S)
zu
        Microsoft Windows SDK Include.

        Return
        ------
        list of str
            paths
        �includer��glr|rz%ssharedz%sumz%swinrtN)rrr�r�r!)r�r#�sdkverrrr�
OSIncludesas


�zEnvironmentInfo.OSIncludescCs�t|jjd�}g}|jdkr&||j7}|jdkr@|t|d�g7}|jdkr�||t|jjd�t|dd�t|d	d�t|d
d�t|jjddd
|jddd�g7}|S)z}
        Microsoft Windows SDK Libraries Paths.

        Return
        ------
        list of str
            paths
        Z
Referencesr�r�zCommonConfiguration\Neutralr|Z
UnionMetadataz'Windows.Foundation.UniversalApiContractz1.0.0.0z%Windows.Foundation.FoundationContractz,Windows.Networking.Connectivity.WwanContractZ
ExtensionSDKszMicrosoft.VCLibsr�ZCommonConfigurationZneutral)rrr�r�r")r��ref�libpathrrr�	OSLibpathys2




�
�
��zEnvironmentInfo.OSLibpathcCst|���S)zs
        Microsoft Windows SDK Tools.

        Return
        ------
        list of str
            paths
        )�list�
_sdk_toolsr�rrr�SdkTools�s
zEnvironmentInfo.SdkToolsccs�|jdkr,|jdkrdnd}t|jj|�V|j��s\|jjdd�}d|}t|jj|�V|jdvr�|j��rvd	}n|jjddd
�}d|}t|jj|�VnB|jdkr�t|jjd�}|jjdd�}|jj}t|d||f�V|jj	r�|jj	Vd
S)z�
        Microsoft Windows SDK Tools paths generator.

        Return
        ------
        generator of str
            paths
        rr�rzBin\x86Trr)r�r�rr
zBin\NETFX 4.0 Tools%sz%s%sN)
r�rrr�r�r�r�r�r�r�)r��bin_dirrrVr%rrrr+�s(	




zEnvironmentInfo._sdk_toolscCs|jj}|rd|SdS)zu
        Microsoft Windows SDK version subdir.

        Return
        ------
        str
            subdir
        �%s\r)rr��r��ucrtverrrrr!�s
zEnvironmentInfo._sdk_subdircCs|jdkrgSt|jjd�gS)zs
        Microsoft Windows SDK Setup.

        Return
        ------
        list of str
            paths
        r��Setup)r�rrr�r�rrr�SdkSetup�s

zEnvironmentInfo.SdkSetupcs�|j}|j�|jdkr0d}|��o,|��}n$|��p>|��}|jdkpR|jdk}g}|rt|�fdd��jD�7}|r�|�fdd��jD�7}|S)zv
        Microsoft .NET Framework Tools.

        Return
        ------
        list of str
            paths
        r�Tr[csg|]}t�j|��qSr)rr��rir-�rrrr�s�z+EnvironmentInfo.FxTools.<locals>.<listcomp>csg|]}t�j|��qSr)rr�r3r4rrrs�)	r�rr�r�r�r�r�r�r�)r�r�Z	include32Z	include64rrr4r�FxTools�s"

��zEnvironmentInfo.FxToolscCs8|jdks|jjsgS|jjdd�}t|jjd|�gS)z~
        Microsoft .Net Framework SDK Libraries.

        Return
        ------
        list of str
            paths
        r|Trzlib\um%s)r�rr�r�r�r)r�rrrr�NetFxSDKLibrariess
z!EnvironmentInfo.NetFxSDKLibrariescCs&|jdks|jjsgSt|jjd�gS)z}
        Microsoft .Net Framework SDK Includes.

        Return
        ------
        list of str
            paths
        r|z
include\um)r�rr�rr�rrr�NetFxSDKIncludess
z EnvironmentInfo.NetFxSDKIncludescCst|jjd�gS)z�
        Microsoft Visual Studio Team System Database.

        Return
        ------
        list of str
            paths
        z
VSTSDB\Deployrr�rrr�VsTDb$s
zEnvironmentInfo.VsTDbcCsv|jdkrgS|jdkr0|jj}|jjdd�}n|jj}d}d|j|f}t||�g}|jdkrr|t||d�g7}|S)zn
        Microsoft Build Engine.

        Return
        ------
        list of str
            paths
        r�rTrrzMSBuild\%0.1f\bin%sZRoslyn)r�rr�r�r�r�r)r��	base_pathrrV�buildrrr�MSBuild0s



zEnvironmentInfo.MSBuildcCs|jdkrgSt|jjd�gS)zt
        Microsoft HTML Help Workshop.

        Return
        ------
        list of str
            paths
        r�zHTML Help Workshop)r�rrr�r�rrr�HTMLHelpWorkshopLs

z EnvironmentInfo.HTMLHelpWorkshopcCsD|jdkrgS|jjdd�}t|jjd�}|j}t|d||f�gS)z�
        Microsoft Universal C Runtime SDK Libraries.

        Return
        ------
        list of str
            paths
        r|Trr�z%sucrt%s)r�r�r�rrr��_ucrt_subdir)r�rr�r0rrr�
UCRTLibraries[s

zEnvironmentInfo.UCRTLibrariescCs.|jdkrgSt|jjd�}t|d|j�gS)z�
        Microsoft Universal C Runtime SDK Include.

        Return
        ------
        list of str
            paths
        r|r#z%sucrt)r�rrr�r=)r�r#rrr�UCRTIncludesms

zEnvironmentInfo.UCRTIncludescCs|jj}|rd|SdS)z�
        Microsoft Universal C Runtime SDK version subdir.

        Return
        ------
        str
            subdir
        r.r)rr�r/rrrr=}s
zEnvironmentInfo._ucrt_subdircCs(d|jkrdkrnngS|jjgS)zk
        Microsoft Visual F#.

        Return
        ------
        list of str
            paths
        r�r�)r�rr�r�rrr�FSharp�s
zEnvironmentInfo.FSharpc
Cs�d|j}|jjdd��d�}g}|jj}t|�dd��}t|�rft	|t
|�d�}||t	|d�g7}|t	|d	�g7}d
|jdd
t|j�df}t
�||�D]&\}}t	||||�}	t|	�r�|	Sq�dS)
z
        Microsoft Visual C++ runtime redistributable dll.

        Return
        ------
        str
            path
        zvcruntime%d0.dllTrr�z\Toolsz\Redistr^Zonecorer\zMicrosoft.VC%d.CRT�
N)r�r�r�rRrr�rrmrrrr=r�r9r�r)
r�rer�prefixesZ
tools_pathZredist_pathZcrt_dirsrZcrt_dirrVrrr�VCRuntimeRedist�s 

�zEnvironmentInfo.VCRuntimeRedistTcCs�t|�d|j|j|j|jg|�|�d|j|j|j|j	|j
g|�|�d|j|j|j|jg|�|�d|j
|j|j|j|j|j|j|j|jg	|�d�}|jdkr�t|j�r�|j|d<|S)z�
        Return environment dict.

        Parameters
        ----------
        exists: bool
            It True, only return existing paths.

        Return
        ------
        dict
            environment
        r#r�r(rV)r#r�r(rVr5rs)�dict�_build_pathsrr&r?r7rr"r5r>r6rr)r rr8r,r2r;r<r@r�rrC)r��existsrzrrrr+�sV�������	�� 
zEnvironmentInfo.return_envc
Csntj�|�}t�|d��t�}t�||�}|r<ttt	|��n|}|s\d|�
�}tj�
|��t|�}	t�|	�S)aC
        Given an environment variable name and specified paths,
        return a pathsep-separated string of paths containing
        unique, extant, directories from those paths and from
        the environment variable. Raise an error if no paths
        are resolved.

        Parameters
        ----------
        name: str
            Environment variable name
        spec_path_lists: list of str
            Paths
        exists: bool
            It True, only return existing paths.

        Return
        ------
        str
            Pathsep-separated paths
        rz %s environment variable is empty)r9�chain�
from_iterablerrNr�rr*�filterr�upperr&r'r(r
r)
r�r�Zspec_path_listsrFZ
spec_pathsZ	env_pathsrZextant_pathsr�Zunique_pathsrrrrE�szEnvironmentInfo._build_paths)Nr)T)!rrrr�r�r�r�r�rrrrr r"r&r)r,r+r!r2r5r6r7r8r;r<r>r?r=r@rCr+rErrrrr*�sj
	







$


#
#












"
2r*)r$)r)1r�r��ior�osrr�os.pathrrrrr~r��platformr9rO�distutils.errorsr&Z#setuptools.extern.packaging.versionr	Z setuptools.extern.more_itertoolsr
Zmonkeyr�systemrrrDrar'r(Z_msvc9_suppress_errorsZdistutils.msvc9compilerrrr%rDrWr_rfr{r}r�r,r�r�r�r*rrrr�<module>s`�	*
&&�'$
%s:site-packages/setuptools/__pycache__/wheel.cpython-39.pyc000064400000016240150564745010017464 0ustar00a

(�Re` �@s�dZddlmZddlmZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddlm
Z
ddlmZddlmZddlmZe	�d	e	j�jZd
Zdd�ZGd
d�d�ZdS)zWheels support.�)�get_platform)�logN)�
parse_version)�sys_tags)�canonicalize_name)�write_requirementsz�^(?P<project_name>.+?)-(?P<version>\d.*?)
    ((-(?P<build>\d.*?))?-(?P<py_version>.+?)-(?P<abi>.+?)-(?P<platform>.+?)
    )\.whl$z8__import__('pkg_resources').declare_namespace(__name__)
cCs�t�|�D]�\}}}tj�||�}|D].}tj�||�}tj�|||�}t�||�q&ttt|���D]D\}	}
tj�||
�}tj�|||
�}tj�	|�sft�||�||	=qfq
tj|dd�D]\}}}|r�J�t�
|�q�dS)zDMove everything under `src_dir` to `dst_dir`, and delete the former.T)�topdownN)�os�walk�path�relpath�join�renames�reversed�list�	enumerate�exists�rmdir)�src_dirZdst_dir�dirpath�dirnames�	filenames�subdir�f�src�dst�n�d�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/wheel.py�unpacks
r c@sheZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	e
dd��Ze
dd��Ze
dd��Z
dS)�WheelcCsPttj�|��}|dur$td|��||_|����D]\}}t|||�q6dS)Nzinvalid wheel name: %r)	�
WHEEL_NAMEr	r�basename�
ValueError�filename�	groupdict�items�setattr)�selfr%�match�k�vrrr�__init__6szWheel.__init__cCs&t�|j�d�|j�d�|j�d��S)z>List tags (py_version, abi, platform) supported by this wheel.�.)�	itertools�product�
py_version�split�abi�platform�r)rrr�tags>s



�z
Wheel.tagscs0tdd�t�D���t�fdd�|��D�d�S)z5Is the wheel is compatible with the current platform?css|]}|j|j|jfVqdS�N)�interpreterr3r4��.0�trrr�	<genexpr>Hsz&Wheel.is_compatible.<locals>.<genexpr>c3s|]}|�vrdVqdS)TNrr9��supported_tagsrrr<J�F)�setr�nextr6r5rr=r�
is_compatibleFs�zWheel.is_compatiblecCs,tj|j|j|jdkrdnt�d���dS)N�any)�project_name�versionr4z.egg)�
pkg_resources�DistributionrDrEr4r�egg_namer5rrrrHLs��zWheel.egg_namecCsJ|��D]4}t�|�}|�d�rt|��t|j��r|Sqtd��dS)Nz
.dist-infoz.unsupported wheel format. .dist-info not found)�namelist�	posixpath�dirname�endswithr�
startswithrDr$)r)�zf�memberrKrrr�
get_dist_infoRs

��
zWheel.get_dist_infocCs<t�|j��}|�||�Wd�n1s.0YdS)z"Install wheel as an egg directory.N)�zipfile�ZipFiler%�_install_as_egg)r)�destination_eggdirrNrrr�install_as_egg\szWheel.install_as_eggcCs\d|j|jf}|�|�}d|}tj�|d�}|�||||�|�||�|�||�dS)Nz%s-%sz%s.data�EGG-INFO)	rDrErPr	rr
�_convert_metadata�_move_data_entries�_fix_namespace_packages)r)rTrNZ
dist_basename�	dist_info�	dist_data�egg_inforrrrSas
zWheel._install_as_eggc	sV��fdd�}|d�}t|�d��}td�|ko>td�kn}|sTtd|��t�|���|�tj�|���tj	j
|�t�|��d��d	d
��tt
t�����������fdd��jD�}t��|�t�tj�|d
�tj�|d��tj	t�|d�d�}	tjj}
t�tj�z*t|	�d�dtj�|d��Wt�|
�nt�|
�0dS)NcsT��t��|���.}|���d�}tj���|�Wd�S1sF0YdS)Nzutf-8)	�openrJr
�read�decode�email�parser�Parser�parsestr)�name�fp�value)rZrNrr�get_metadatamsz-Wheel._convert_metadata.<locals>.get_metadataZWHEELz
Wheel-Versionz1.0z2.0dev0z$unsupported wheel format version: %s)�metadatacSsd|_t|�Sr7)�marker�str)�reqrrr�raw_req�sz(Wheel._convert_metadata.<locals>.raw_reqc
s2i|]*}|t�fdd�t���|f��D���qS)c3s|]}|�vr|VqdSr7r)r:rk)�install_requiresrrr<�s�z5Wheel._convert_metadata.<locals>.<dictcomp>.<genexpr>)�sorted�map�requires)r:�extra)�distrmrlrr�
<dictcomp>�s��z+Wheel._convert_metadata.<locals>.<dictcomp>�METADATAzPKG-INFO)rm�extras_require)�attrsr\zrequires.txt)r�getr$r	�mkdir�
extractallrr
rFrG�
from_location�PathMetadatarrnrorp�extras�rename�
setuptools�dictr�_global_log�	threshold�
set_threshold�WARNr�get_command_obj)rNrTrZr\rg�wheel_metadata�
wheel_versionZwheel_v1ruZ
setup_distZ
log_thresholdr)rrrZrmrlrNrrWksR��


������zWheel._convert_metadatacs�tj�|���tj��d�}tj�|�r�tj�|dd�}t�|�t�|�D]D}|�d�rpt�tj�||��qLt�tj�||�tj�||��qLt�	|�t
tjj�fdd�dD��D]}t||�q�tj���r�t�	��dS)z,Move data entries to their correct location.�scriptsrVz.pycc3s|]}tj��|�VqdSr7)r	rr
)r:r�r[rrr<�s�z+Wheel._move_data_entries.<locals>.<genexpr>)�data�headers�purelib�platlibN)r	rr
rrx�listdirrL�unlinkr}r�filterr )rTr[Zdist_data_scriptsZegg_info_scripts�entryrrr�rrX�s*�

�
�zWheel._move_data_entriesc	Cs�tj�|d�}tj�|�r�t|��}|����}Wd�n1sD0Y|D]�}tjj|g|�d��R�}tj�|d�}tj�|�s�t�|�tj�|�sRt|d��}|�t	�Wd�qR1s�0YqRdS)Nznamespace_packages.txtr.z__init__.py�w)
r	rr
rr]r^r2rx�write�NAMESPACE_PACKAGE_INIT)r\rTZnamespace_packagesre�modZmod_dirZmod_initrrrrY�s�
*
zWheel._fix_namespace_packagesN)�__name__�
__module__�__qualname__r-r6rBrHrPrUrS�staticmethodrWrXrYrrrrr!4s


?
r!)�__doc__�distutils.utilr�	distutilsrr`r/r	rJ�rerQrFr~rZ setuptools.extern.packaging.tagsrZ!setuptools.extern.packaging.utilsrZsetuptools.command.egg_infor�compile�VERBOSEr*r"r�r r!rrrr�<module>s,��site-packages/setuptools/__pycache__/extension.cpython-39.pyc000064400000003710150564745010020372 0ustar00a

(�Re��@spddlZddlZddlZddlZddlZddlmZdd�ZeZ	eej
j�ZGdd�de�ZGdd	�d	e�Z
dS)
�N�)�
get_unpatchedcCs2d}zt|dgd�jWdSty,Yn0dS)z0
    Return True if Cython can be imported.
    zCython.Distutils.build_ext�	build_ext)�fromlistTF)�
__import__r�	Exception)Zcython_impl�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/extension.py�_have_cython
sr
c@s eZdZdZdd�Zdd�ZdS)�	Extensionz7Extension that uses '.c' files in place of '.pyx' filescOs.|�dd�|_tj|||g|�Ri|��dS)N�py_limited_apiF)�popr�
_Extension�__init__)�self�name�sources�args�kwrrr	r!szExtension.__init__cCsNt�r
dS|jpd}|��dkr$dnd}t�tjd|�}tt||j	��|_	dS)z�
        Replace sources with .pyx extensions to sources with the target
        language extension. This mechanism allows language authors to supply
        pre-converted sources but to prefer the .pyx sources.
        N�zc++z.cppz.cz.pyx$)
r
�language�lower�	functools�partial�re�sub�list�mapr)r�langZ
target_extrrrr	�_convert_pyx_sources_to_lang's
z&Extension._convert_pyx_sources_to_langN)�__name__�
__module__�__qualname__�__doc__rrrrrr	rsrc@seZdZdZdS)�Libraryz=Just like a regular Extension, but built as a library insteadN)r r!r"r#rrrr	r$6sr$)rr�distutils.core�	distutils�distutils.errors�distutils.extensionZmonkeyrr
Z
have_pyrex�corerrr$rrrr	�<module>ssite-packages/setuptools/__pycache__/build_meta.cpython-39.pyc000064400000021641150564745010020466 0ustar00a

(�Re((�@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
mZgd�ZGdd�de
�ZGdd�dejj�Zejd	d
��Zdd�Zd
d�Zdd�ZGdd�de�ZGdd�de�Ze�ZejZejZejZejZejZe�ZdS)a-A PEP 517 interface to setuptools

Previously, when a user or a command line tool (let's call it a "frontend")
needed to make a request of setuptools to take a certain action, for
example, generating a list of installation requirements, the frontend would
would call "setup.py egg_info" or "setup.py bdist_wheel" on the command line.

PEP 517 defines a different method of interfacing with setuptools. Rather
than calling "setup.py" directly, the frontend should:

  1. Set the current directory to the directory with a setup.py file
  2. Import this module into a safe python interpreter (one in which
     setuptools can potentially set global variables or crash hard).
  3. Call one of the functions defined in PEP 517.

What each function does is defined in PEP 517. However, here is a "casual"
definition of the functions (this definition should not be relied on for
bug reports or API stability):

  - `build_wheel`: build a wheel in the folder and return the basename
  - `get_requires_for_build_wheel`: get the `setup_requires` to build
  - `prepare_metadata_for_build_wheel`: get the `install_requires`
  - `build_sdist`: build an sdist in the folder and return the basename
  - `get_requires_for_build_sdist`: get the `setup_requires` to build

Again, this is not a formal definition! Just a "taste" of the module.
�N)�parse_requirements)�get_requires_for_build_sdist�get_requires_for_build_wheel� prepare_metadata_for_build_wheel�build_wheel�build_sdist�
__legacy__�SetupRequirementsErrorc@seZdZdd�ZdS)r	cCs
||_dS�N)�
specifiers)�selfr�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/build_meta.py�__init__4szSetupRequirementsError.__init__N)�__name__�
__module__�__qualname__rr
r
r
rr	3sr	c@s&eZdZdd�Zeejdd���ZdS)�DistributioncCstttt|���}t|��dSr
)�list�map�strrr	)rrZspecifier_listr
r
r�fetch_build_eggs9szDistribution.fetch_build_eggsccs2tjj}|tj_zdVW|tj_n
|tj_0dS)zw
        Replace
        distutils.dist.Distribution with this class
        for the duration of this context.
        N)�	distutils�corer)�cls�origr
r
r�patch>s
zDistribution.patchN)rrrr�classmethod�
contextlib�contextmanagerrr
r
r
rr8srccs.tj}dd�t_zdVW|t_n|t_0dS)a
Temporarily disable installing setup_requires

    Under PEP 517, the backend reports build dependencies to the frontend,
    and the frontend is responsible for ensuring they're installed.
    So setuptools (acting as a backend) should not try to install them.
    cSsdSr
r
)�attrsr
r
r�<lambda>W�z+no_install_setup_requires.<locals>.<lambda>N)�
setuptoolsZ_install_setup_requires)rr
r
r�no_install_setup_requiresNs

r$cs�fdd�t���D�S)Ncs&g|]}tj�tj��|��r|�qSr
)�os�path�isdir�join)�.0�name�Za_dirr
r�
<listcomp>_s�z1_get_immediate_subdirectories.<locals>.<listcomp>)r%�listdirr+r
r+r�_get_immediate_subdirectories^sr.csB�fdd�t�|�D�}z
|\}Wnty<td��Yn0|S)Nc3s|]}|���r|VqdSr
��endswith�r)�f��	extensionr
r�	<genexpr>ds
�z'_file_with_extension.<locals>.<genexpr>z[No distribution was found. Ensure that `setup.py` is not empty and that it calls `setup()`.)r%r-�
ValueError)�	directoryr4Zmatching�filer
r3r�_file_with_extensioncs
�
�
r9cCs&tj�|�st�d�Sttdt�|�S)Nz%from setuptools import setup; setup()�open)r%r&�exists�io�StringIO�getattr�tokenizer:��setup_scriptr
r
r�_open_setup_scriptqs
rBc@s`eZdZdd�Zdd�Zddd�Zdd	d
�Zddd�Zdd
d�Zdd�Z	ddd�Z
ddd�ZdS)�_BuildMetaBackendcCs|pi}|�dg�|S)N�--global-option)�
setdefault�r�config_settingsr
r
r�_fix_config{sz_BuildMetaBackend._fix_configc
Cs�|�|�}tjdd�dg|dt_z4t���|��Wd�n1sP0YWn.ty�}z||j7}WYd}~n
d}~00|S)N��egg_inforD)rH�sys�argvrr�	run_setupr	r)rrG�requirements�er
r
r�_get_build_requires�s
�
* z%_BuildMetaBackend._get_build_requires�setup.pycCsX|}d}t|�� }|���dd�}Wd�n1s60Ytt||d�t��dS)N�__main__z\r\nz\n�exec)rB�read�replacerS�compile�locals)rrA�__file__rr2�coder
r
rrM�s

.z_BuildMetaBackend.run_setupNcCs|�|�}|j|dgd�S)N�wheel�rN�rHrPrFr
r
rr�s
�z._BuildMetaBackend.get_requires_for_build_wheelcCs|�|�}|j|gd�S)Nr[r\rFr
r
rr�s
z._BuildMetaBackend.get_requires_for_build_sdistcCs�tjdd�dd|gt_t��|��Wd�n1s>0Y|}dd�t�|�D�}t|�dkr�tt|��dkr�tj�	|t�|�d�}qLt|�dks�J�q�qL||kr�t
�tj�	||d�|�t
j|dd�|dS)	NrIZ	dist_infoz
--egg-basecSsg|]}|�d�r|�qS)z
.dist-infor/r1r
r
rr,�s
�zF_BuildMetaBackend.prepare_metadata_for_build_wheel.<locals>.<listcomp>rT)�
ignore_errors)
rKrLr$rMr%r-�lenr.r&r(�shutil�move�rmtree)r�metadata_directoryrGZdist_info_directoryZ
dist_infosr
r
rr�s0�&
����z2_BuildMetaBackend.prepare_metadata_for_build_wheelc	Cs�|�|�}tj�|�}tj|dd�tj|d���}tjdd�|d|g|dt_t	��|�
�Wd�n1sz0Yt||�}tj�||�}tj�
|�r�t�|�t�tj�||�|�Wd�n1s�0Y|S)NT)�exist_ok)�dirrIz
--dist-dirrD)rHr%r&�abspath�makedirs�tempfile�TemporaryDirectoryrKrLr$rMr9r(r;�remove�rename)rZ
setup_commandZresult_extensionZresult_directoryrGZtmp_dist_dirZresult_basename�result_pathr
r
r�_build_with_temp_dir�s&
��&�
4z&_BuildMetaBackend._build_with_temp_dircCs|�dgd||�S)N�bdist_wheelz.whl�rl)r�wheel_directoryrGrbr
r
rr�s
�z_BuildMetaBackend.build_wheelcCs|�gd�d||�S)N)�sdistz	--formats�gztarz.tar.gzrn)r�sdist_directoryrGr
r
rr�s
�z_BuildMetaBackend.build_sdist)rQ)N)N)N)NN)N)rrrrHrPrMrrrrlrrr
r
r
rrCys


�
"�
rCcs"eZdZdZd�fdd�	Z�ZS)�_BuildMetaLegacyBackendaOCompatibility backend for setuptools

    This is a version of setuptools.build_meta that endeavors
    to maintain backwards
    compatibility with pre-PEP 517 modes of invocation. It
    exists as a temporary
    bridge between the old packaging mechanism and the new
    packaging mechanism,
    and will eventually be removed.
    rQc
s�ttj�}tj�tj�|��}|tjvr6tj�d|�tjd}|tjd<z.tt	|�j
|d�W|tjdd�<|tjd<n|tjdd�<|tjd<0dS)Nrr@)rrKr&r%�dirnamere�insertrL�superrsrM)rrA�sys_pathZ
script_dirZ
sys_argv_0��	__class__r
rrM�s 



���z!_BuildMetaLegacyBackend.run_setup)rQ)rrr�__doc__rM�
__classcell__r
r
rxrrs�s
rs) rzr<r%rKr?r_rrgr#r�
pkg_resourcesr�__all__�
BaseExceptionr	�distrrr$r.r9rB�objectrCrsZ_BACKENDrrrrrrr
r
r
r�<module>s6	
m)site-packages/setuptools/__pycache__/dist.cpython-39.pyc000064400000107217150564745010017330 0ustar00a

(�ReO��@s�dgZddlZddlZddlZddlZddlZddlZddlZddl	Zddl
ZddlZddlZddl
mZddlmZddlmZddlmZddlZddlZddlmZmZmZddlmZdd	lmZdd
lm Z m!Z!ddl
m"Z"ddl#m$Z$dd
l%m&Z&ddl%m'Z'ddl(m)Z)ddl*m+Z+ddl,Z,ddl-Z,ddl,m.Z.ddl/m0Z0ddl1m2Z2ddl3Z3e�rpddl4m5Z5e6d�e6d�dd�Z7dd�Z8e9e9d�dd�Z:de9ee9d �d!d"�Z;de9ee9d �d#d$�Z<de9eee9d �d%d&�Z=dee9d'�d(d)�Z>d*d+�Z?d,d-�Z@d.d/�ZAeBeCfZDd0d1�ZEd2d3�ZFd4d5�ZGd6d7�ZHd8d9�ZId:d;�ZJd<d=�ZKd>d?�ZLd@dA�ZMdBdC�ZNdDdE�ZOdFdG�ZPdHdI�ZQe0ejRjS�ZTGdJd�deT�ZSGdKdL�dLe+�ZUdS)M�Distribution�N)�	strtobool)�DEBUG��translate_longopt)�iglob)�List�Optional�
TYPE_CHECKING)�defaultdict)�message_from_file)�DistutilsOptionError�DistutilsSetupError)�
rfc822_escape)�
StrictVersion)�	packaging)�ordered_set)�unique_everseen�)�SetuptoolsDeprecationWarning)�windows_support)�
get_unpatched)�parse_configuration)�Messagez&setuptools.extern.packaging.specifiersz#setuptools.extern.packaging.versioncCst�dt�t|�S)NzDo not call this function)�warnings�warn�DistDeprecationWarningr)�cls�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/dist.py�_get_unpatched2sr cCs&t|dd�}|dur"td�}||_|S)N�metadata_version�2.1)�getattrrr!)�self�mvrrr�get_metadata_version7s
r&)�content�returnc
CsJ|��}t|�dkr |d��Sd�|d��t�d�|dd���f�S)zFReverse RFC-822 escaping by removing leading whitespaces from content.rr�
N)�
splitlines�len�lstrip�join�textwrap�dedent)r'�linesrrr�rfc822_unescape?sr1r)�msg�fieldr(cCs||}|dkrdS|S)zRead Message header field.�UNKNOWNNr�r2r3�valuerrr�_read_field_from_msgGsr7cCst||�}|dur|St|�S)z4Read Message header field and apply rfc822_unescape.N)r7r1r5rrr�_read_field_unescaped_from_msgOs
r8cCs|�|d�}|gkrdS|S)z9Read Message header field and return all results as list.N)�get_all)r2r3�valuesrrr�_read_list_from_msgWsr;)r2r(cCs|����}|dkrdS|S)Nr4)�get_payload�strip)r2r6rrr�_read_payload_from_msg_sr>cCsVt|�}t|d�|_t|d�|_t|d�|_t|d�|_t|d�|_d|_t|d�|_	d|_
t|d�|_t|d	�|_
d
|vr�t|d
�|_nd|_t|d�|_|jdur�|jtd�kr�t|�|_t|d�|_d
|vr�t|d
��d�|_t|d�|_t|d�|_|jtd�k�r4t|d�|_t|d�|_t|d�|_nd|_d|_d|_t|d�|_dS)z-Reads the metadata values from a file object.zmetadata-version�name�version�summary�authorNzauthor-emailz	home-page�licensezdownload-url�descriptionr"�keywords�,�platform�
classifierz1.1�requires�provides�	obsoleteszlicense-file)rrr!r7r?r@rDrB�
maintainer�author_email�maintainer_email�urlr8rC�download_url�long_descriptionr>�splitrEr;�	platforms�classifiersrIrJrK�
license_files)r$�filer2rrr�
read_pkg_filefs<
rWcCs"d|vrt�d�|�dd�}|S)Nr)z1newlines not allowed and will break in the future� )rr�replace)�valrrr�single_line�s
r[c
s�|��}�fdd�}|dt|��|d|���|d|���|dt|����|d|���d}|D]&\}}t||d	�}|d	urf|||�qft|�	��}|d
|�|j
r�|d|j
�|j��D]}	|dd
|	�q�d�
|���}
|
r�|d|
�|��D]}|d|�q�|��d|���|��d|���|��d|���|��d|���t|d��rh|d|j�|j�r||d|j�|j�r�|jD]}|d|��q�|��d|j�p�g���d|���d	S)z0Write the PKG-INFO format data to a file object.cs��d||f�dS)Nz%s: %s
)�write)�keyr6�rVrr�write_field�sz#write_pkg_file.<locals>.write_fieldzMetadata-Version�Name�VersionZSummaryz	Home-page))ZAuthorrB)zAuthor-emailrM)Z
MaintainerrL)zMaintainer-emailrNNZLicensezDownload-URLzProject-URLz%s, %srFZKeywords�Platform�
Classifier�Requires�Provides�	Obsoletes�python_requireszRequires-PythonzDescription-Content-TypezProvides-ExtrazLicense-Filez
%s

)r&�str�get_name�get_versionr[�get_description�get_urlr#r�get_licenserP�project_urls�itemsr-�get_keywords�
get_platforms�_write_list�get_classifiers�get_requires�get_provides�
get_obsoletes�hasattrrg�long_description_content_type�provides_extrasrUr\�get_long_description)
r$rVr@r_Zoptional_fieldsr3�attrZattr_valrC�project_urlrErG�extrarr^r�write_pkg_file�sH


r~cCsbztj�d|�}|jrJ�Wn>ttttfy\}ztd||f�|�WYd}~n
d}~00dS)Nzx=z4%r must be importable 'module:attrs' string (got %r))	�
pkg_resources�
EntryPoint�parse�extras�	TypeError�
ValueError�AttributeError�AssertionErrorr)�distr{r6�ep�errr�check_importable�s
��r�cCslz(t|ttf�sJ�d�|�|ks&J�Wn>ttttfyf}ztd||f�|�WYd}~n
d}~00dS)z"Verify that value is a string list�z%%r must be a list of strings (got %r)N)	�
isinstance�list�tupler-r�r�r�r�r�r�r{r6r�rrr�assert_string_list�s
��r�cCsd|}t|||�|D]J}|�|�s2tdd|��|�d�\}}}|r||vrtj�d||�qdS)z(Verify that namespace packages are validz1Distribution contains no modules or packages for znamespace package %r�.z^WARNING: %r is declared as a package namespace, but %r is not: please correct this in setup.pyN)r��has_contents_forr�
rpartition�	distutils�logr)r�r{r6Zns_packages�nsp�parent�sep�childrrr�	check_nsp�s 
���r�c
CsRztt�t|����Wn4tttfyL}ztd�|�WYd}~n
d}~00dS)z+Verify that extras_require mapping is validz�'extras_require' must be a dictionary whose values are strings or lists of strings containing valid project/version requirement specifiers.N)	r��	itertools�starmap�_check_extraror�r�r�rr�rrr�check_extras
s��r�cCs<|�d�\}}}|r*t�|�r*td|��tt�|��dS)N�:zInvalid environment marker: )�	partitionr�invalid_markerrr��parse_requirements)r}�reqsr?r��markerrrrr�sr�cCs&t|�|kr"d}t|j||d���dS)z)Verify that value is True, False, 0, or 1z0{attr!r} must be a boolean value (got {value!r}))r{r6N)�boolr�format)r�r{r6�tmplrrr�assert_boolsr�cCs,|st�|�d�t�dSt|�d���dS)Nz is ignored.z is invalid.)rrrr�r�r{r6rrr�invalid_unless_false$sr�c
Csnz(tt�|��t|ttf�r&td��Wn@ttfyh}z$d}t|j	||d��|�WYd}~n
d}~00dS)z9Verify that install_requires is a valid requirements listzUnordered types are not allowedzm{attr!r} must be a string or list of strings containing valid project/version requirement specifiers; {error}�r{�errorN)
r�rr�r��dict�setr�r�rr��r�r{r6r�r�rrr�check_requirements+s�r�c
CsZztj�|�WnDtjjtfyT}z$d}t|j||d��|�WYd}~n
d}~00dS)z.Verify that value is a valid version specifierzF{attr!r} must be a string containing valid version specifiers; {error}r�N)r�
specifiers�SpecifierSet�InvalidSpecifierr�rr�r�rrr�check_specifier9s�r�c
CsDztj�|�Wn.ty>}zt|�|�WYd}~n
d}~00dS)z)Verify that entry_points map is parseableN)rr��	parse_mapr�rr�rrr�check_entry_pointsDsr�cCst|t�std��dS)Nztest_suite must be a string)r�rhrr�rrr�check_test_suiteLs
r�cCsZt|t�std�|���|��D]4\}}t|t�sBtd�||���t|d�|�|�q dS)z@Verify that value is a dictionary of package names to glob listszT{!r} must be a dictionary mapping package names to lists of string wildcard patternsz,keys of {!r} dict must be strings (got {!r})zvalues of {!r} dictN)r�r�rr�rorhr�)r�r{r6�k�vrrr�check_package_dataQs
��

�r�cCs(|D]}t�d|�stj�d|�qdS)Nz\w+(\.\w+)*z[WARNING: %r not a valid package name; please use only .-separated package names in setup.py)�re�matchr�r�r)r�r{r6�pkgnamerrr�check_packages`s�r�c@s~eZdZdZdd�eejdd�dd�d�ZdZdd	�Z	dUd
d�Z
dd
�Zedd��Z
edd��Zdd�Zdd�Zedd��Zdd�Zdd�Zdd�Zedd��ZdVd d!�Zd"d#�Zd$d%�Zd&d'�ZdWd(d)�ZdXd+d,�Zd-d.�Zd/d0�Zed1d2��Zd3d4�Zd5d6�Z d7d8�Z!d9d:�Z"d;d<�Z#d=d>�Z$d?d@�Z%dAdB�Z&dCdD�Z'dEdF�Z(dGdH�Z)dIdJ�Z*dKdL�Z+dMdN�Z,dOdP�Z-dQdR�Z.dSdT�Z/dS)YraGDistribution with support for tests and package data

    This is an enhanced version of 'distutils.dist.Distribution' that
    effectively adds the following new optional keyword arguments to 'setup()':

     'install_requires' -- a string or sequence of strings specifying project
        versions that the distribution requires when installed, in the format
        used by 'pkg_resources.require()'.  They will be installed
        automatically when the package is installed.  If you wish to use
        packages that are not available in PyPI, or want to give your users an
        alternate download location, you can add a 'find_links' option to the
        '[easy_install]' section of your project's 'setup.cfg' file, and then
        setuptools will scan the listed web pages for links that satisfy the
        requirements.

     'extras_require' -- a dictionary mapping names of optional "extras" to the
        additional requirement(s) that using those extras incurs. For example,
        this::

            extras_require = dict(reST = ["docutils>=0.3", "reSTedit"])

        indicates that the distribution can optionally provide an extra
        capability called "reST", but it can only be used if docutils and
        reSTedit are installed.  If the user installs your package using
        EasyInstall and requests one of your extras, the corresponding
        additional requirements will be installed if needed.

     'test_suite' -- the name of a test suite to run for the 'test' command.
        If the user runs 'python setup.py test', the package will be installed,
        and the named test suite will be run.  The format is the same as
        would be used on a 'unittest.py' command line.  That is, it is the
        dotted name of an object to import and call to generate a test suite.

     'package_data' -- a dictionary mapping package names to lists of filenames
        or globs to use to find data files contained in the named packages.
        If the dictionary has filenames or globs listed under '""' (the empty
        string), those names will be searched for in every package, in addition
        to any names for the specific package.  Data files found using these
        names/globs will be installed along with the package, in the same
        location as the package.  Note that globs are allowed to reference
        the contents of non-package subdirectories, as long as you use '/' as
        a path separator.  (Globs are automatically converted to
        platform-specific paths at runtime.)

    In addition to these new keywords, this class also has several new methods
    for manipulating the distribution's contents.  For example, the 'include()'
    and 'exclude()' methods can be thought of as in-place add and subtract
    commands that add or remove packages, modules, extensions, and so on from
    the distribution.
    cCsdS�Nrrrrr�<lambda>��zDistribution.<lambda>cCsdSr�rrrrrr��r�cCsdSr�rrrrrr��r�)rxrnry�license_filerUNcCsl|rd|vsd|vrdSt�t|d����}tjj�|�}|durh|�d�sht�t|d��|_	||_
dS)Nr?r@zPKG-INFO)r�	safe_namerh�lower�working_set�by_key�get�has_metadata�safe_version�_version�
_patched_dist)r$�attrsr]r�rrr�patch_missing_pkg_info�sz#Distribution.patch_missing_pkg_infocs�t�d�}|si�_|pi}g�_|�dd��_��|�|�dg��_|�dg��_t�	d�D]}t
���|jd�q`t
���fdd�|��D����|������jj���j_���dS)N�package_data�src_root�dependency_links�setup_requires�distutils.setup_keywordscs i|]\}}|�jvr||�qSr)�_DISTUTILS_UNSUPPORTED_METADATA��.0r�r��r$rr�
<dictcomp>�s
�z)Distribution.__init__.<locals>.<dictcomp>)rwr��
dist_files�popr�r�r�r�r�iter_entry_points�vars�
setdefaultr?�
_Distribution�__init__ro�_set_metadata_defaults�_normalize_version�_validate_version�metadatar@�_finalize_requires)r$r�Zhave_package_datar�rr�rr��s,


��	
�zDistribution.__init__cCs4|j��D]$\}}t|j��||�||���q
dS)z�
        Fill-in missing metadata fields not supported by distutils.
        Some fields may have been set by other tools (e.g. pbr).
        Those fields (vars(self.metadata)) take precedence to
        supplied attrs.
        N)r�ror�r�r�r�)r$r��option�defaultrrrr��sz#Distribution._set_metadata_defaultscCsTt|tj�s|dur|Sttj�|��}||krPd}t�|j	fit
����|S|S)Nz)Normalizing '{version}' to '{normalized}')r��
setuptools�sicrhrr@rarrr��locals)r@�
normalizedr�rrrr��szDistribution._normalize_versionc	Csdt|tj�rt|�}|dur`ztj�|�Wn2tjjtfy^t	�
d|�t�|�YS0|S)Nz�The version specified (%r) is an invalid version, this may not work as expected with newer versions of setuptools, pip, and PyPI. Please see PEP 440 for more details.)
r��numbers�Numberrhrr@ra�InvalidVersionr�rrr�r�)r@rrrr��s��zDistribution._validate_versioncCsft|dd�r|j|j_t|dd�rR|j��D]$}|�d�d}|r,|jj�|�q,|��|�	�dS)z�
        Set `metadata.python_requires` and fix environment markers
        in `install_requires` and `extras_require`.
        rgN�extras_requirer�r)
r#rgr�r��keysrRry�add�_convert_extras_requirements�"_move_install_requirements_markers)r$r}rrrr��s
zDistribution._finalize_requirescCsht|dd�pi}tt�|_|��D]@\}}|j|t�|�D]"}|�|�}|j||�|�q>q"dS)z�
        Convert requirements in `extras_require` of the form
        `"extra": ["barbazquux; {marker}"]` to
        `"extra:{marker}": ["barbazquux"]`.
        r�N)	r#rr��_tmp_extras_requirerorr��_suffix_for�append)r$Z
spec_ext_reqs�sectionr��r�suffixrrrr�s


z)Distribution._convert_extras_requirementscCs|jrdt|j�SdS)ze
        For a requirement, return the 'extras_require' suffix for
        that requirement.
        r�r�)r�rh��reqrrrr�!szDistribution._suffix_forcs�dd�}t�dd�pd}tt�|��}t||�}t�||�}ttt|���_	|D]}�j
dt|j��|�qPt
�fdd��j
��D���_dS)	zv
        Move requirements in `install_requires` that are using environment
        markers `extras_require`.
        cSs|jSr��r�r�rrr�
is_simple_req3szFDistribution._move_install_requirements_markers.<locals>.is_simple_req�install_requiresNrr�c3s,|]$\}}|dd�t�j|�D�fVqdS)cSsg|]}t|��qSr)rh)r�r�rrr�
<listcomp>?r�zMDistribution._move_install_requirements_markers.<locals>.<genexpr>.<listcomp>N)�map�
_clean_reqr�r�rr�	<genexpr>>s�zBDistribution._move_install_requirements_markers.<locals>.<genexpr>)r#r�rr��filterr��filterfalser�rhr�r�r�r�r�ror�)r$r�Zspec_inst_reqsZ	inst_reqsZsimple_reqsZcomplex_reqsr�rr�rr�)s

�z/Distribution._move_install_requirements_markerscCs
d|_|S)zP
        Given a Requirement, remove environment markers and return it.
        Nr�)r$r�rrrr�CszDistribution._clean_reqcCs`|jj}|r|ng}|jj}|r2||vr2|�|�|durF|durFd}tt|�|���|j_dS)z<Compute names of all license files which should be included.N)zLICEN[CS]E*zCOPYING*zNOTICE*zAUTHORS*)r�rUr�r�r�r�_expand_patterns)r$rU�patternsr�rrr�_finalize_license_filesJs
�z$Distribution._finalize_license_filescCsdd�|D�S)z�
        >>> list(Distribution._expand_patterns(['LICENSE']))
        ['LICENSE']
        >>> list(Distribution._expand_patterns(['setup.cfg', 'LIC*']))
        ['setup.cfg', 'LICENSE']
        css:|]2}tt|��D] }|�d�stj�|�r|VqqdS)�~N)�sortedr�endswith�os�path�isfile)r��patternr
rrrres�z0Distribution._expand_patterns.<locals>.<genexpr>r)rrrrr]s�zDistribution._expand_patternscCs�ddlm}tjtjkrgngd�}t|�}|dur<|��}trJ|�d�|�}t	|_
|D]�}tj|dd��6}tr�|�dj
fit����|�|�Wd�n1s�0Y|��D]d}|�|�}|�|�}	|D]F}
|
d	ks�|
|vr�q�|�||
�}|�|
|�}
|�|
|�}
||f|	|
<q�q�|��qZd
|jv�r:dS|jd
��D]�\}
\}}|j�|
�}
|
�rrt|�}n|
dv�r�t|�}zt||
�p�|
|�Wn0t�y�}zt|�|�WYd}~n
d}~00�qHdS)z�
        Adapted from distutils.dist.Distribution.parse_config_files,
        this method provides the same functionality in subtly-improved
        ways.
        r)�ConfigParser)
zinstall-basezinstall-platbasezinstall-libzinstall-platlibzinstall-purelibzinstall-headerszinstall-scriptszinstall-data�prefixzexec-prefix�home�user�rootNz"Distribution.parse_config_files():�utf-8)�encodingz  reading {filename}�__name__�global)�verbose�dry_run)�configparserr
�sysr�base_prefix�	frozenset�find_config_filesr�announcerh�optionxform�io�openr�r��	read_file�sections�options�get_option_dictr��warn_dash_deprecation�make_option_lowercaser��command_optionsro�negative_optr�setattrr�r
)r$�	filenamesr
�ignore_options�parser�filename�readerr�r#�opt_dict�optrZ�src�aliasr�rrr�_parse_config_filesmsP
��
(



z Distribution._parse_config_filescCsd|dvr|S|�dd�}tjj|��}|�d�sF|dkrF||vrF|Sd|vr`t�d||f�|S)N)zoptions.extras_requirezoptions.data_files�-�_r#r�zrUsage of dash-separated '%s' will not be supported in future versions. Please use the underscore name '%s' instead)rYr��command�__all__�_setuptools_commands�
startswithrr)r$r0r�Zunderscore_opt�commandsrrrr%�s$�����z"Distribution.warn_dash_deprecationcCs8zt�d�}t|�d��WStjy2gYS0dS)Nr��distutils.commands)r�get_distributionr��
get_entry_map�DistributionNotFound)r$r�rrrr8�s

z!Distribution._setuptools_commandscCs4|dks|��r|S|��}t�d|||f�|S)Nr�zlUsage of uppercase key '%s' in '%s' will be deprecated in future versions. Please use lowercase '%s' instead)�islowerr�rr)r$r0r�Z
lowercase_optrrrr&�s��z"Distribution.make_option_lowercasecCs\|��}|dur|�|�}tr,|�d|�|��D�] \}\}}trZ|�d|||f�zdd�|jD�}Wnty�g}Yn0z
|j}Wnty�i}Yn0z|t|t	�}	||vr�|	r�t
|||t|��nJ||vr�|	r�t
||t|��n,t||��rt
|||�nt
d|||f��Wq4t�yT}
zt
|
�|
�WYd}
~
q4d}
~
00q4dS)a�
        Set the options for 'command_obj' from 'option_dict'.  Basically
        this means copying elements of a dictionary ('option_dict') to
        attributes of an instance ('command').

        'command_obj' must be a Command instance.  If 'option_dict' is not
        supplied, uses the standard option dictionary for this command
        (from 'self.command_options').

        (Adopted from distutils.dist.Distribution._set_command_options)
        Nz#  setting options for '%s' command:z    %s = %s (from %s)cSsg|]}t|��qSrr)r��orrrr��r�z5Distribution._set_command_options.<locals>.<listcomp>z1error in %s: command '%s' has no such option '%s')�get_command_namer$rrro�boolean_optionsr�r(r�rhr)rrwr
r�)r$�command_obj�option_dict�command_namer��sourcer6�	bool_opts�neg_opt�	is_stringr�rrr�_set_command_options�s>




��z!Distribution._set_command_optionsFcCs0|j|d�t||j|d�|��|��dS)zYParses configuration files from various levels
        and loads configuration.

        )r*)�ignore_option_errorsN)r3rr'r�r)r$r*rKrrr�parse_config_filess�zDistribution.parse_config_filescCs8tjjt�|�|jdd�}|D]}tjj|dd�q|S)zResolve pre-setup requirementsT)�	installer�replace_conflicting)rY)rr��resolver��fetch_build_eggr�)r$rIZresolved_distsr�rrr�fetch_build_eggs$s�zDistribution.fetch_build_eggscCsPd}dd�}t�|�}t�|j|�}tdd�|�}t||d�D]}||�q>dS)z�
        Allow plugins to apply arbitrary operations to the
        distribution. Each hook may optionally define a 'order'
        to influence the order of execution. Smaller numbers
        go first and the default is 0.
        z(setuptools.finalize_distribution_optionscSst|dd�S)N�orderr)r#)�hookrrr�by_order8sz/Distribution.finalize_options.<locals>.by_ordercSs|��Sr�)�load)r�rrrr�=r�z/Distribution.finalize_options.<locals>.<lambda>)r]N)rr�r�r�_removedr�r)r$�grouprTZdefined�filteredZloadedr�rrr�finalize_options/s
zDistribution.finalize_optionscCsdh}|j|vS)z�
        When removing an entry point, if metadata is loaded
        from an older version of Setuptools, that removed
        entry point will attempt to be loaded and will fail.
        See #2765 for more details.
        Z
2to3_doctests)r?)r��removedrrrrVAs
�zDistribution._removedcCsJt�d�D]:}t||jd�}|dur
|j|jd�|��||j|�q
dS)Nr��rM)rr�r#r?�requirerPrU)r$r�r6rrr�_finalize_setup_keywordsOs
z%Distribution._finalize_setup_keywordscCs�tj�tjd�}tj�|�s�t�|�t�|�tj�|d�}t|d��.}|�	d�|�	d�|�	d�Wd�n1s|0Y|S)Nz.eggsz
README.txt�wzcThis directory contains eggs that were downloaded by setuptools to build, test, and run plug-ins.

zAThis directory caches those eggs to prevent repeated downloads.

z/However, it is safe to delete this directory.

)
r	r
r-�curdir�exists�mkdirrZ	hide_filer r\)r$Z
egg_cache_dirZreadme_txt_filename�frrr�get_egg_cache_dirVs

��(zDistribution.get_egg_cache_dircCsddlm}|||�S)z Fetch an egg needed for buildingr)rP)Zsetuptools.installerrP)r$r�rPrrrrPiszDistribution.fetch_build_eggcCs\||jvr|j|St�d|�}|D]*}|j|jd�|��|j|<}|St�||�S)z(Pluggable version of get_command_class()r;r[N)�cmdclassrr�r\rPrUr��get_command_class)r$r6Zepsr�rdrrrreos

zDistribution.get_command_classcCs:t�d�D]$}|j|jvr
|��}||j|j<q
t�|�S�Nr;)rr�r?rdrOr��print_commands�r$r�rdrrrrg|s
zDistribution.print_commandscCs:t�d�D]$}|j|jvr
|��}||j|j<q
t�|�Srf)rr�r?rdrOr��get_command_listrhrrrri�s
zDistribution.get_command_listcKs@|��D]2\}}t|d|d�}|r.||�q|�||�qdS)a�Add items to distribution that are named in keyword arguments

        For example, 'dist.include(py_modules=["x"])' would add 'x' to
        the distribution's 'py_modules' attribute, if it was not already
        there.

        Currently, this method only supports inclusion for attributes that are
        lists or tuples.  If you need to add support for adding to other
        attributes in this or a subclass, you can add an '_include_X' method,
        where 'X' is the name of the attribute.  The method will be called with
        the value passed to 'include()'.  So, 'dist.include(foo={"bar":"baz"})'
        will try to call 'dist._include_foo({"bar":"baz"})', which can then
        handle whatever special inclusion logic is needed.
        Z	_include_N)ror#�
_include_misc)r$r�r�r��includerrrrk�s

zDistribution.includecsf�d�|jr&��fdd�|jD�|_|jrD��fdd�|jD�|_|jrb��fdd�|jD�|_dS)z9Remove packages, modules, and extensions in named packager�cs"g|]}|�kr|���s|�qSr�r9�r��p��package�pfxrrr��sz0Distribution.exclude_package.<locals>.<listcomp>cs"g|]}|�kr|���s|�qSrrlrmrorrr��scs&g|]}|j�kr|j���s|�qSr)r?r9rmrorrr��s�N)�packages�
py_modules�ext_modules)r$rprror�exclude_package�s���zDistribution.exclude_packagecCs2|d}|��D]}||ks&|�|�rdSqdS)z<Return true if 'exclude_package(package)' would do somethingr�TN)�iter_distribution_namesr9)r$rprqrnrrrr��szDistribution.has_contents_forc
s�t�t�std|�f��zt||�}Wn2tyZ}ztd|�|�WYd}~n
d}~00|dur|t|t�s|t|d��n|r�t||�fdd�|D��dS)zAHandle 'exclude()' for list/tuple attrs without a special handlerz(%s: setting must be a list or tuple (%r)� %s: No such distribution settingN�4: this setting cannot be changed via include/excludecsg|]}|�vr|�qSrr�r��item�r6rrr��r�z.Distribution._exclude_misc.<locals>.<listcomp>�r��sequencerr#r�r))r$r?r6�oldr�rr{r�
_exclude_misc�s

�$�zDistribution._exclude_miscc
s�t|t�std||f��zt||��Wn2tyZ}ztd|�|�WYd}~n
d}~00�durrt|||�n:t�t�s�t|d��n"�fdd�|D�}t||�|�dS)zAHandle 'include()' for list/tuple attrs without a special handlerz%s: setting must be a list (%r)rwNrxcsg|]}|�vr|�qSrrry�r~rrr��r�z.Distribution._include_misc.<locals>.<listcomp>r|)r$r?r6r��newrr�rrj�s
$
�zDistribution._include_misccKs@|��D]2\}}t|d|d�}|r.||�q|�||�qdS)aRemove items from distribution that are named in keyword arguments

        For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from
        the distribution's 'py_modules' attribute.  Excluding packages uses
        the 'exclude_package()' method, so all of the package's contained
        packages, modules, and extensions are also excluded.

        Currently, this method only supports exclusion from attributes that are
        lists or tuples.  If you need to add support for excluding from other
        attributes in this or a subclass, you can add an '_exclude_X' method,
        where 'X' is the name of the attribute.  The method will be called with
        the value passed to 'exclude()'.  So, 'dist.exclude(foo={"bar":"baz"})'
        will try to call 'dist._exclude_foo({"bar":"baz"})', which can then
        handle whatever special exclusion logic is needed.
        Z	_exclude_N)ror#r)r$r�r�r��excluderrrr��s

zDistribution.excludecCs,t|t�std|f��tt|j|��dS)Nz.packages: setting must be a list or tuple (%r))r�r}rr�r�ru)r$rrrrr�_exclude_packages�s

�zDistribution._exclude_packagesc
Cs�|jj|_|jj|_|d}|�d�}||vrf||\}}||=ddl}|�|d�|dd�<|d}q&t�|||�}|�|�}	t	|	dd�r�d|f|�|�d<|dur�gS|S)Nr�aliasesTrZcommand_consumes_arguments�command line�args)
�	__class__�global_optionsr(r$�shlexrRr��_parse_command_optsrer#)
r$r,r�r6r�r1r2r��nargs�	cmd_classrrrr�s"




z Distribution._parse_command_optscCs�i}|j��D]�\}}|��D]�\}\}}|dkr4q|�dd�}|dkr�|�|�}|j��}|�t|di��|��D]\}	}
|
|krv|	}d}q�qvtd��n|dkr�d}||�	|i�|<qq|S)	ahReturn a '{cmd: {opt:val}}' map of all command-line options

        Option names are all long, but do not include the leading '--', and
        contain dashes rather than underscores.  If the option doesn't take
        an argument (e.g. '--quiet'), the 'val' is 'None'.

        Note that options provided by config files are intentionally excluded.
        r�r5r4rr(NzShouldn't be able to get herer)
r'rorY�get_command_objr(�copy�updater#r�r�)r$�d�cmd�optsr0r1rZZcmdobjrH�neg�posrrr�get_cmdline_optionss(



z Distribution.get_cmdline_optionsccsv|jpdD]
}|Vq
|jpdD]
}|Vq |jp4dD]:}t|t�rN|\}}n|j}|�d�rj|dd�}|Vq6dS)z@Yield all packages, modules, and extension names in distributionr�moduleNi����)rrrsrtr�r�r?r)r$�pkgr��extr?Z	buildinforrrrvEs


z$Distribution.iter_distribution_namesc
Cs�ddl}|jrt�||�St|jtj�s4t�||�S|jj�	�dvrPt�||�S|jj}|jj
}|jdkrndppd}|jj}t�|j�
�d|||�|_z(t�||�Wt�|j�
�||||�|_St�|j�
�||||�|_0dS)z�If there were any non-global "display-only" options
        (--help-commands or the metadata display options) on the command
        line, display the requested info and return true; else return
        false.
        rN)r�utf8�win32r)r)r�
help_commandsr��handle_display_optionsr��stdoutr�
TextIOWrapperrr��errorsrG�line_buffering�detach)r$�option_orderrrr��newliner�rrrr�Ws2�����z#Distribution.handle_display_options)N)N)N)NF)0r�
__module__�__qualname__�__doc__r�rZ
OrderedSetr�r�r�r�r��staticmethodr�r�r�r�r�r�r�rrr3r%r8r&rJrLrQrYrVr]rcrPrergrirkrur�rrjr�r�r�r�rvr�rrrrrmsf4�






O
.




	(c@seZdZdZdS)rzrClass for warning about deprecations in dist in
    setuptools. Not ignored by default, unlike DeprecationWarning.N)rr�r�r�rrrrr|sr)Vr7rrr�r	rr�Z
distutils.logr��distutils.core�
distutils.cmd�distutils.dist�distutils.command�distutils.utilr�distutils.debugr�distutils.fancy_getoptr�globrr�r.�typingrr	r
�collectionsr�emailr�distutils.errorsr
rrZdistutils.versionrZsetuptools.externrrZ setuptools.extern.more_itertoolsrr�rr�Zsetuptools.commandrZsetuptools.monkeyrZsetuptools.configrr�
email.messager�
__import__r r&rhr1r7r8r;r>rWr[r~r�r�r}r�r�r�r�r�r�r�r�r�r�r�r�r��corerr�rrrrr�<module>s�-
>

site-packages/setuptools/__pycache__/sandbox.cpython-39.pyc000064400000036674150564745010020033 0ustar00a

(�Re8�@s�ddlZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
ddlmZej�d�r�ddlmmmmZnejejZzeZWney�dZYn0eZgd�Zd-dd�Zejd.dd	��Z ejd
d��Z!ejdd
��Z"ejdd��Z#Gdd�de$�Z%Gdd�d�Z&ejdd��Z'dd�Z(ejdd��Z)ejdd��Z*hd�Z+dd�Z,dd �Z-d!d"�Z.Gd#d$�d$�Z/e0ed%��r�ej1gZ2ngZ2Gd&d'�d'e/�Z3e�4ej5d(d)�d*�6�D��Z7Gd+d,�d,e
�Z8dS)/�N)�DistutilsError)�working_set�java)�AbstractSandbox�DirectorySandbox�SandboxViolation�	run_setupcCs^d}t||��}|��}Wd�n1s,0Y|durB|}t||d�}t|||�dS)z.
    Python 3 implementation of execfile.
    �rbN�exec)�open�read�compiler
)�filename�globals�locals�mode�stream�script�code�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/sandbox.py�	_execfile$s&rc
csRtjdd�}|dur$|tjdd�<z|VW|tjdd�<n|tjdd�<0dS�N)�sys�argv)�repl�savedrrr�	save_argv1src
cs<tjdd�}z|VW|tjdd�<n|tjdd�<0dSr)r�path�rrrr�	save_path<sr ccs8tj|dd�tj}|t_zdVW|t_n|t_0dS)zL
    Monkey-patch tempfile.tempdir with replacement, ensuring it exists
    T)�exist_okN)�os�makedirs�tempfile�tempdir)�replacementrrrr�
override_tempEsr'c	cs8t��}t�|�z|VWt�|�nt�|�0dSr)r"�getcwd�chdir)�targetrrrr�pushdVs

r+c@seZdZdZedd��ZdS)�UnpickleableExceptionzP
    An exception representing another Exception that could not be pickled.
    c	CsNzt�|�t�|�fWStyHddlm}|�||t|���YS0dS)z�
        Always return a dumped (pickled) type and exc. If exc can't be pickled,
        wrap it in UnpickleableException first.
        r)r,N)�pickle�dumps�	Exception�setuptools.sandboxr,�dump�repr)�type�exc�clsrrrr1es
zUnpickleableException.dumpN)�__name__�
__module__�__qualname__�__doc__�staticmethodr1rrrrr,`sr,c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�ExceptionSaverz^
    A Context Manager that will save an exception, serialized, and restore it
    later.
    cCs|Srr��selfrrr�	__enter__zszExceptionSaver.__enter__cCs |sdSt�||�|_||_dS�NT)r,r1�_saved�_tb)r=r3r4�tbrrr�__exit__}s
zExceptionSaver.__exit__cCs2dt|�vrdSttj|j�\}}|�|j��dS)z"restore and re-raise any exceptionr@N)�vars�mapr-�loadsr@�with_tracebackrA)r=r3r4rrr�resume�szExceptionSaver.resumeN)r6r7r8r9r>rCrHrrrrr;tsr;c#sjtj���t��}�VWd�n1s,0Ytj����fdd�tjD�}t|�|��dS)z�
    Context in which imported modules are saved.

    Translates exceptions internal to the context into the equivalent exception
    outside the context.
    Nc3s$|]}|�vr|�d�s|VqdS)z
encodings.N��
startswith)�.0�mod_namerrr�	<genexpr>�s
�zsave_modules.<locals>.<genexpr>)r�modules�copyr;�update�_clear_modulesrH)�	saved_excZdel_modulesrrr�save_modules�s
$
�rScCst|�D]}tj|=qdSr)�listrrN)Zmodule_namesrLrrrrQ�srQc	cs.t��}z|VWt�|�nt�|�0dSr)�
pkg_resources�__getstate__�__setstate__rrrr�save_pkg_resources_state�srXc
cstj�|d�}t���t���t���t�t��nt|��Ft	|��t
d�dVWd�n1sj0YWd�n1s�0YWd�n1s�0YWd�n1s�0YWd�n1s�0YWd�n1�s0YdS)N�temp�
setuptools)r"r�joinrXrSr �hide_setuptoolsrr'r+�
__import__)�	setup_dir�temp_dirrrr�
setup_context�s

r`>�	distutilsrUZCython�_distutils_hackrZcCs|�dd�d}|tvS)aH
    >>> _needs_hiding('setuptools')
    True
    >>> _needs_hiding('pkg_resources')
    True
    >>> _needs_hiding('setuptools_plugin')
    False
    >>> _needs_hiding('setuptools.__init__')
    True
    >>> _needs_hiding('distutils')
    True
    >>> _needs_hiding('os')
    False
    >>> _needs_hiding('Cython')
    True
    �.�r)�split�_MODULES_TO_HIDE)rL�base_modulerrr�
_needs_hiding�srhcCs6tj�dd�}|dur|��tttj�}t|�dS)a%
    Remove references to setuptools' modules from sys.modules to allow the
    invocation to import the most appropriate setuptools. This technique is
    necessary to avoid issues such as #315 where setuptools upgrading itself
    would fail to find a function declared in the metadata.
    rbN)rrN�getZremove_shim�filterrhrQ)rbrNrrrr\�s
r\cCs�tj�tj�|��}t|���z�|gt|�tjdd�<tj�d|�t	�
�t	j�dd��t
|��&t|dd�}t||�Wd�n1s�0YWn6ty�}z|jr�|jdr��WYd}~n
d}~00Wd�n1s�0YdS)z8Run a distutils setup script, sandboxed in its directoryNrcSs|��Sr)�activate)�distrrr�<lambda>��zrun_setup.<locals>.<lambda>�__main__)�__file__r6)r"r�abspath�dirnamer`rTrr�insertr�__init__�	callbacks�appendr�dictr�
SystemExit�args)Zsetup_scriptryr^�ns�vrrrr�s

,rc@seZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dD]Zee
e�rDe
e�e�e<qDd$dd�Zerzede�Zede�ZdD]Zee
e�r�ee�e�e<q�dd�ZdD]Zee
e�r�ee�e�e<q�dd�ZdD]Zee
e�r�ee�e�e<q�dd�Zdd�Zd d!�Zd"d#�ZdS)%rzDWrap 'os' module and 'open()' builtin for virtualizing setup scriptsFcs�fdd�tt�D��_dS)Ncs$g|]}|�d�st�|�r|�qS)�_)rJ�hasattr)rK�namer<rr�
<listcomp>s�z,AbstractSandbox.__init__.<locals>.<listcomp>)�dir�_os�_attrsr<rr<rrts
�zAbstractSandbox.__init__cCs"|jD]}tt|t||��qdSr)r��setattrr"�getattr)r=�sourcer~rrr�_copys
zAbstractSandbox._copycCs(|�|�tr|jt_|jt_d|_dSr?)r��_file�builtins�file�_openr�_activer<rrrr>s

zAbstractSandbox.__enter__cCs$d|_trtt_tt_|�t�dS�NF)r�r�r�r�r�rr�r�)r=�exc_type�	exc_value�	tracebackrrrrC!s
zAbstractSandbox.__exit__cCs.|�|�Wd�S1s 0YdS)zRun 'func' under os sandboxingNr)r=�funcrrr�run(szAbstractSandbox.runcstt�����fdd�}|S)Ncs>|jr&|j�||g|�Ri|��\}}�||g|�Ri|��Sr)r��_remap_pair)r=�src�dstry�kw�r~�originalrr�wrap0s z3AbstractSandbox._mk_dual_path_wrapper.<locals>.wrap�r�r��r~r�rr�r�_mk_dual_path_wrapper-s
z%AbstractSandbox._mk_dual_path_wrapper)�rename�link�symlinkNcs �ptt�����fdd�}|S)Ncs6|jr |j�|g|�Ri|��}�|g|�Ri|��Sr)r��_remap_input�r=rryr�r�rrr�>sz5AbstractSandbox._mk_single_path_wrapper.<locals>.wrapr�)r~r�r�rr�r�_mk_single_path_wrapper;sz'AbstractSandbox._mk_single_path_wrapperr�r)�stat�listdirr)r�chmod�chown�mkdir�remove�unlink�rmdir�utime�lchown�chroot�lstatZ	startfile�mkfifo�mknod�pathconf�accesscstt�����fdd�}|S)NcsT|jr>|j�|g|�Ri|��}|���|g|�Ri|���S�|g|�Ri|��Sr)r�r��
_remap_outputr�r�rrr�csz4AbstractSandbox._mk_single_with_return.<locals>.wrapr�r�rr�r�_mk_single_with_return`s
z&AbstractSandbox._mk_single_with_return)�readlink�tempnamcstt�����fdd�}|S)Ncs$�|i|��}|jr |��|�S|Sr)r�r�)r=ryr��retvalr�rrr�rsz'AbstractSandbox._mk_query.<locals>.wrapr�r�rr�r�	_mk_queryos
zAbstractSandbox._mk_query)r(�tmpnamcCs|S)z=Called to remap or validate any path, whether input or outputr)r=rrrr�_validate_path~szAbstractSandbox._validate_pathcOs
|�|�S�zCalled for path inputs�r��r=�	operationrryr�rrrr��szAbstractSandbox._remap_inputcCs
|�|�S)zCalled for path outputsr�)r=r�rrrrr��szAbstractSandbox._remap_outputcOs<|j|d|g|�Ri|��|j|d|g|�Ri|��fS)�?Called for path pairs like rename, link, and symlink operationsz-fromz-to)r��r=r�r�r�ryr�rrrr��s�zAbstractSandbox._remap_pair)N)r6r7r8r9r�rtr�r>rCr�r�r~r}r�rr�r�r�r�r�r�r�r�r�rrrrr
s<








r�devnullc@s�eZdZdZe�gd��ZgZefdd�Z	dd�Z
er@ddd	�Zdd
d�Zdd
�Z
dd�Zdd�Zdd�Zdd�Zddd�ZdS)rz<Restrict operations to a single subdirectory - pseudo-chroot)
rr�r�r�r�r�r�r�r�r�r�r�r�cCsFtj�tj�|��|_tj�|jd�|_dd�|D�|_t�	|�dS)N�cSs g|]}tj�tj�|���qSr)r"r�normcase�realpath)rKrrrrr�sz-DirectorySandbox.__init__.<locals>.<listcomp>)
r"rr�r��_sandboxr[�_prefix�_exceptionsrrt)r=Zsandbox�
exceptionsrrrrt�s�zDirectorySandbox.__init__cOsddlm}||||��dS)Nr)r)r0r)r=r�ryr�rrrr�
_violation�szDirectorySandbox._violation�rcOsF|dvr.|�|�s.|jd||g|�Ri|��t||g|�Ri|��S)N�r��rtr	ZrU�Ur�)�_okr�r��r=rrryr�rrrr��szDirectorySandbox._filecOsF|dvr.|�|�s.|jd||g|�Ri|��t||g|�Ri|��S)Nr�r)r�r�r�r�rrrr��szDirectorySandbox._opencCs|�d�dS)Nr�)r�r<rrrr��szDirectorySandbox.tmpnamcCsV|j}zBd|_tj�tj�|��}|�|�p@||jkp@|�|j�W||_S||_0dSr�)	r�r"rr�r��	_exemptedr�rJr�)r=r�activer�rrrr��s
�
��zDirectorySandbox._okcs<�fdd�|jD�}�fdd�|jD�}t�||�}t|�S)Nc3s|]}��|�VqdSrrI)rK�	exception��filepathrrrM�sz-DirectorySandbox._exempted.<locals>.<genexpr>c3s|]}t�|��VqdSr)�re�match)rK�patternr�rrrM�s)r��_exception_patterns�	itertools�chain�any)r=r�Z
start_matchesZpattern_matches�
candidatesrr�rr��s
�
�zDirectorySandbox._exemptedcOs:||jvr6|�|�s6|j|tj�|�g|�Ri|��|Sr�)�	write_opsr�r�r"rr�r�rrrr��s"zDirectorySandbox._remap_inputcOs8|�|�r|�|�s0|j|||g|�Ri|��||fS)r�)r�r�r�rrrr��szDirectorySandbox._remap_pair�cOsL|t@r0|�|�s0|jd|||g|�Ri|��tj|||g|�Ri|��S)zCalled for low-level os.open()zos.open)�WRITE_FLAGSr�r�r�r)r=r��flagsrryr�rrrr�szDirectorySandbox.openN)r�)r�)r�)r6r7r8r9rw�fromkeysr�r��_EXCEPTIONSrtr�r�r�r�r�r�r�r�rrrrrr�s �



rcCsg|]}tt|d��qS)rr�)rK�arrrr�s�rz4O_WRONLY O_RDWR O_APPEND O_CREAT O_TRUNC O_TEMPORARYc@s&eZdZdZe�d���Zdd�ZdS)rzEA setup script attempted to modify the filesystem outside the sandboxa
        SandboxViolation: {cmd}{args!r} {kwargs}

        The package setup script has attempted to modify files on your system
        that are not within the EasyInstall build area, and has been aborted.

        This package cannot be safely installed by EasyInstall, and may not
        support alternate installation locations even if you run its setup
        script by hand.  Please inform the package's author and the EasyInstall
        maintainers to find out if a fix or workaround is available.
        cCs |j\}}}|jjfit���Sr)ry�tmpl�formatr)r=�cmdry�kwargsrrr�__str__szSandboxViolation.__str__N)	r6r7r8r9�textwrap�dedent�lstripr�r�rrrrr�s
�r)N)N)9r"rr$�operator�	functoolsr�r��
contextlibr-r�r�rU�distutils.errorsrr�platformrJZ$org.python.modules.posix.PosixModule�pythonrN�posixZPosixModuler�r~r�r��	NameErrorrr��__all__r�contextmanagerrr r'r+r/r,r;rSrQrXr`rfrhr\rrr}r�r�r�reduce�or_rer�rrrrr�<module>sr 






	


		
^��	site-packages/setuptools/__pycache__/namespaces.cpython-39.pyc000064400000007100150564745010020472 0ustar00a

(�Re�@sFddlZddlmZddlZejjZGdd�d�ZGdd�de�ZdS)�N)�logc@sTeZdZdZdd�Zdd�Zdd�ZdZd	Zd
d�Z	dd
�Z
dd�Zedd��Z
dS)�	Installerz
-nspkg.pthcCs�|��}|sdStj�|���\}}||j7}|j�|�t�	d|�t
|j|�}|jrdt
|�dSt|d��}|�|�Wd�n1s�0YdS)Nz
Installing %s�wt)�_get_all_ns_packages�os�path�splitext�_get_target�	nspkg_ext�outputs�appendr�info�map�_gen_nspkg_line�dry_run�list�open�
writelines)�self�nsp�filename�ext�lines�f�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/namespaces.py�install_namespaces
s
zInstaller.install_namespacescCsHtj�|���\}}||j7}tj�|�s.dSt�d|�t�|�dS)NzRemoving %s)	rrrr	r
�existsrr
�remove)rrrrrr�uninstall_namespacess
zInstaller.uninstall_namespacescCs|jS�N)�target�rrrrr	'szInstaller._get_target)	zimport sys, types, osz#has_mfs = sys.version_info > (3, 5)z$p = os.path.join(%(root)s, *%(pth)r)z4importlib = has_mfs and __import__('importlib.util')z-has_mfs and __import__('importlib.machinery')z�m = has_mfs and sys.modules.setdefault(%(pkg)r, importlib.util.module_from_spec(importlib.machinery.PathFinder.find_spec(%(pkg)r, [os.path.dirname(p)])))zCm = m or sys.modules.setdefault(%(pkg)r, types.ModuleType(%(pkg)r))z7mp = (m or []) and m.__dict__.setdefault('__path__',[])z(p not in mp) and mp.append(p))z4m and setattr(sys.modules[%(parent)r], %(child)r, m)cCsdS)Nz$sys._getframe(1).f_locals['sitedir']rr"rrr�	_get_rootEszInstaller._get_rootcCsNt|�d��}|��}|j}|�d�\}}}|r:||j7}d�|�t�dS)N�.�;�
)�tuple�splitr#�_nspkg_tmpl�
rpartition�_nspkg_tmpl_multi�join�locals)r�pkg�pth�rootZ
tmpl_lines�parent�sep�childrrrrHs
zInstaller._gen_nspkg_linecCs |jjp
g}ttt|j|���S)z,Return sorted list of all package namespaces)�distributionZnamespace_packages�sorted�flattenr�
_pkg_names)r�pkgsrrrrQszInstaller._get_all_ns_packagesccs(|�d�}|r$d�|�V|��q
dS)z�
        Given a namespace package, yield the components of that
        package.

        >>> names = Installer._pkg_names('a.b.c')
        >>> set(names) == set(['a', 'a.b', 'a.b.c'])
        True
        r$N)r(r,�pop)r.�partsrrrr7Vs

zInstaller._pkg_namesN)�__name__�
__module__�__qualname__r
rrr	r)r+r#rr�staticmethodr7rrrrr	s	rc@seZdZdd�Zdd�ZdS)�DevelopInstallercCstt|j��Sr )�repr�strZegg_pathr"rrrr#gszDevelopInstaller._get_rootcCs|jSr )�egg_linkr"rrrr	jszDevelopInstaller._get_targetN)r;r<r=r#r	rrrrr?fsr?)	r�	distutilsr�	itertools�chain�
from_iterabler6rr?rrrr�<module>s
]site-packages/setuptools/__pycache__/package_index.cpython-39.pyc000064400000077654150564745010021162 0ustar00a

(�Re�@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
ZddlZddlZddlZddlmZddlZddlmZmZmZmZmZmZmZmZmZm Z m!Z!m"Z"m#Z#ddl$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+dd	l,m-Z-e�.d
�Z/e�.dej0�Z1e�.d�Z2e�.d
ej0�j3Z4d�5�Z6gd�Z7dZ8dZ9e9j:dj:ej;�ed�Z<dd�Z=dd�Z>dd�Z?dBdd�Z@dCdd�ZAdDdd�ZBdedfd d!�ZCd"d#�ZDe�.d$ej0�ZEeDd%d&��ZFGd'd(�d(�ZGGd)d*�d*eG�ZHGd+d,�d,e�ZIe�.d-�jJZKd.d/�ZLd0d1�ZMdEd2d3�ZNd4d5�ZOGd6d7�d7�ZPGd8d9�d9ejQ�ZRejSjTfd:d;�ZUd<d=�ZVeNe8�eU�ZUd>d?�ZWd@dA�ZXdS)Fz#PyPI and direct package downloading�N��wraps)
�
CHECKOUT_DIST�Distribution�BINARY_DIST�normalize_path�SOURCE_DIST�Environment�find_distributions�	safe_name�safe_version�to_filename�Requirement�DEVELOP_DIST�EGG_DIST)�log)�DistutilsError)�	translate)�Wheel��unique_everseenz^egg=([-A-Za-z0-9_.+!]+)$zhref\s*=\s*['"]?([^'"> ]+)z�<a href="([^"#]+)">([^<]+)</a>\n\s+\(<a (?:title="MD5 hash"\n\s+)href="[^?]+\?:action=show_md5&amp;digest=([0-9a-f]{32})">md5</a>\)z([-+.a-z0-9]{2,}):z.tar.gz .tar.bz2 .tar .zip .tgz)�PackageIndex�distros_for_url�parse_bdist_wininst�interpret_distro_name�z<setuptools/{setuptools.__version__} Python-urllib/{py_major}z{}.{})Zpy_major�
setuptoolsc
CsFzt�|�WSty@}ztd|f�|�WYd}~n
d}~00dS)Nz1Not a URL, existing file, or requirement spec: %r)r�parse�
ValueErrorr)�spec�e�r!��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/package_index.py�parse_requirement_arg6s��r#cCs�|��}d\}}}|�d�r�|�d�r8|dd�}d}nn|�dd�rb|d	d
�}|dd�}d}nD|�d�r~|dd�}d
}n(|�dd�r�|d	d
�}|dd�}d
}|||fS)z=Return (base,pyversion) or (None,None) for possible .exe name)NNN�.exez
.win32.exeNi�����win32z	.win32-pyi�i�������z.win-amd64.exei�z	win-amd64z
.win-amd64-pyi��)�lower�endswith�
startswith)�namer'�base�py_ver�platr!r!r"r?s$



rc	Csxtj�|�}|\}}}}}}tj�|�d�d�}|dkrX|dkrXtj�|�d�d�}d|vrp|�dd�\}}||fS)N�/���zsourceforge.net�download����#�)�urllibr�urlparse�unquote�split)	�url�parts�scheme�server�path�
parameters�query�fragmentr+r!r!r"�egg_info_for_urlWsr@ccs\t|�\}}t|||�D]
}|Vq|rXt�|�}|rXt||�d�|td�D]
}|VqLdS)zEYield egg or source distribution objects that might be found at a URLr3)�
precedenceN)r@�distros_for_location�EGG_FRAGMENT�matchr�groupr)r8�metadatar+r?�distrDr!r!r"rbs
�
rcCs�|�d�r|dd�}|�d�r8d|vr8t�|||�gS|�d�rxd|vrxt|�}|��s^gSt||j|jtdd�gS|�d	�r�t|�\}}}|dur�t	||||t
|�StD]0}|�|�r�|dt|��}t	|||�Sq�gS)
z:Yield egg or source distribution objects based on basename�.egg.zipNr&z.egg�-z.whlr3)�location�project_name�versionrAr$)
r(r�
from_locationr�
is_compatiblerKrLrrrr�
EXTENSIONS�len)rJ�basenamerF�wheelZwin_baser,�platform�extr!r!r"rBps2
�
�
rBcCstt|�tj�|�|�S)zEYield possible egg or source distribution objects based on a filename)rBr�osr<rQ)�filenamerFr!r!r"�distros_for_filename�s�rWc
cs||�d�}|s,tdd�|dd�D��r,dStdt|�d�D]8}t||d�|d|��d�||d��|||d�Vq>dS)z�Generate alternative interpretations of a source distro name

    Note: if `location` is a filesystem filename, you should call
    ``pkg_resources.normalize_path()`` on it before passing it to this
    routine!
    rIcss|]}t�d|�VqdS)z	py\d\.\d$N)�rerD)�.0�pr!r!r"�	<genexpr>��z(interpret_distro_name.<locals>.<genexpr>�Nr3)�
py_versionrArS)r7�any�rangerPr�join)rJrQrFr^rArSr9rZr!r!r"r�s
$�rcst���fdd��}|S)zs
    Wrap a function returning an iterable such that the resulting iterable
    only ever yields unique items.
    cst�|i|���S�Nr)�args�kwargs��funcr!r"�wrapper�szunique_values.<locals>.wrapperr)rfrgr!rer"�
unique_values�srhz(<([^>]*\srel\s*=\s*['"]?([^'">]+)[^>]*)>c	cs�t�|�D]d}|��\}}tttj|���d���}d|vsDd|vr
t	�|�D]}t
j�|t
|�d���VqNq
dD]@}|�|�}|dkrtt	�||�}|rtt
j�|t
|�d���VqtdS)zEFind rel="homepage" and rel="download" links in `page`, yielding URLs�,Zhomepager0r3)z
<th>Home Pagez<th>Download URLr/N)�REL�finditer�groups�set�map�str�stripr'r7�HREFr4r�urljoin�
htmldecoderE�find�search)r8�pagerD�tag�relZrels�posr!r!r"�find_external_links�s
rzc@s(eZdZdZdd�Zdd�Zdd�ZdS)	�ContentCheckerzP
    A null content checker that defines the interface for checking content
    cCsdS)z3
        Feed a block of data to the hash.
        Nr!��self�blockr!r!r"�feed�szContentChecker.feedcCsdS)zC
        Check the hash. Return False if validation fails.
        Tr!�r}r!r!r"�is_valid�szContentChecker.is_validcCsdS)zu
        Call reporter with information about the checker (hash name)
        substituted into the template.
        Nr!)r}�reporter�templater!r!r"�report�szContentChecker.reportN)�__name__�
__module__�__qualname__�__doc__rr�r�r!r!r!r"r{�sr{c@sBeZdZe�d�Zdd�Zedd��Zdd�Z	dd	�Z
d
d�ZdS)
�HashCheckerzK(?P<hash_name>sha1|sha224|sha384|sha256|sha512|md5)=(?P<expected>[a-f0-9]+)cCs||_t�|�|_||_dSrb)�	hash_name�hashlib�new�hash�expected)r}r�r�r!r!r"�__init__�szHashChecker.__init__cCsBtj�|�d}|st�S|j�|�}|s0t�S|fi|����S)z5Construct a (possibly null) ContentChecker from a URLr/)r4rr5r{�patternru�	groupdict)�clsr8r?rDr!r!r"�from_urlszHashChecker.from_urlcCs|j�|�dSrb)r��updater|r!r!r"rszHashChecker.feedcCs|j��|jkSrb)r��	hexdigestr�r�r!r!r"r�szHashChecker.is_validcCs||j}||�Srb)r�)r}r�r��msgr!r!r"r�s
zHashChecker.reportN)r�r�r�rX�compiler�r��classmethodr�rr�r�r!r!r!r"r��s�

r�csDeZdZdZdLdd�ZdMd	d
�ZdNdd�ZdOd
d�Zdd�Zdd�Z	dd�Z
dd�Zdd�ZdPdd�Z
dd�ZdQ�fdd�	Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�ZdRd)d*�ZdSd+d,�Zd-d.�Zd/Zd0d1�Zd2d3�ZdTd4d5�Zd6d7�Zd8d9�Zd:d;�Zd<d=�Zd>d?�Z e!dUd@dA��Z"dBdC�Z#dDdE�Z$dFdG�Z%dHdI�Z&dJdK�Z'�Z(S)Vrz;A distribution index that scans web pages for download URLs�https://pypi.org/simple/��*NTcOsrtj|g|�Ri|��|dd|�d��|_i|_i|_i|_t�d�	t
t|���j|_
g|_tjj|_dS)Nr.�|)r	r�r(�	index_url�scanned_urls�fetched_urls�
package_pagesrXr�rarnrrD�allows�to_scanr4�request�urlopen�opener)r}r��hostsZ	ca_bundleZ
verify_sslrc�kwr!r!r"r�szPackageIndex.__init__FcCs�||jvr|sdSd|j|<t|�s2|�|�dStt|��}|r\|�|�sPdS|�d|�|sn|rn||jvr�tt|j	|��dS|�|�s�d|j|<dS|�
d|�d|j|<d}|�|||�}|dur�dSt|t
jj�r�|jdkr�|�
d|j�d|j|j<d|j�d	d
���v�r(|��dS|j}|��}t|t��svt|t
jj��rXd}n|j�d��phd}|�|d
�}|��t�|�D](}	t
j�|t|	� d���}
|�!|
��q�|�"|j#��r�t$|dd�dk�r�|�%||�}dS)z<Evaluate a URL as a possible download, and maybe retrieve itNTzFound link: %sz
Reading %sz<Download error on %s: %%s -- Some packages may not be found!i�zAuthentication error: %s�html�content-type�zlatin-1�charset�ignorer3�code�)&r��
URL_SCHEME�process_filename�listr�url_ok�debugr�rn�add�info�open_url�
isinstancer4�error�	HTTPErrorr�r�r8�headers�getr'�close�readro�	get_param�decoderqrkrrrrsrE�process_urlr)r��getattr�
process_index)r}r8�retrieve�dists�tmpl�fr+rvr�rD�linkr!r!r"r�*sT





 zPackageIndex.process_urlcCs�tj�|�s|�d|�dStj�|�r^|s^tj�|�}t�|�D]}|�tj�||�d�qBt	|�}|r�|�
d|�tt|j
|��dS)Nz
Not found: %sTz	Found: %s)rUr<�exists�warn�isdir�realpath�listdirr�rarWr�r�rnr�)r}�fn�nestedr<�itemr�r!r!r"r�_szPackageIndex.process_filenamecCsbt|�}|o|�d���dk}|s8|�tj�|�d�r<dSd}|rRt||��n|�||�dS)Nr3�fileTzN
Note: Bypassing %s (disallowed host; see http://bit.ly/2hrImnY for details).
)	r�rEr'r�r4rr5rr�)r}r8�fatal�s�is_filer�r!r!r"r�os�zPackageIndex.url_okcCs2ttjj|�}dd�|D�}tt�|j|��dS)Ncss0|](}t�|�D]}|�d�r||fVqqdS)z	.egg-linkN)rUr�r()rYr<�entryr!r!r"r[~s
�z.PackageIndex.scan_egg_links.<locals>.<genexpr>)�filterrUr<r�r��	itertools�starmap�
scan_egg_link)r}�search_path�dirsZ	egg_linksr!r!r"�scan_egg_links|s
�zPackageIndex.scan_egg_linkscCs�ttj�||���&}ttdttj|���}Wd�n1s>0Yt	|�dkrXdS|\}}t
tj�||��D]*}tjj|g|�R�|_t|_
|�|�qrdS)Nr])�openrUr<rar�r�rnrorprPr
rJrrAr�)r}r<r�Z	raw_lines�linesZegg_pathZ
setup_pathrGr!r!r"r��s4zPackageIndex.scan_egg_linkcCs�d}|�|j�s|Stttjj|t|j�d��d���}t|�dksRd|dvrV|St	|d�}t
|d�}d|j�|�
�i�|<t|�t|�fS)N)NNr.r]r2r3rT)r)r�r�rnr4rr6rPr7rrr��
setdefaultr'r
)r}r�ZNO_MATCH_SENTINELr9�pkg�verr!r!r"�_scan�s�zPackageIndex._scanc	
Cs�t�|�D]:}z"|�tj�|t|�d����Wq
tyBYq
0q
|�|�\}}|s\dSt	||�D]H}t
|�\}}|�d�r�|s�|r�|d||f7}n
|�|�|�
|�qft�dd�|�S)z#Process the contents of a PyPI pager3r��.pyz
#egg=%s-%scSsd|�ddd�S)Nz<a href="%s#md5=%s">%s</a>r3�r])rE)�mr!r!r"�<lambda>�r\z,PackageIndex.process_index.<locals>.<lambda>)rqrkr�r4rrrrsrErrzr@r(�need_version_info�scan_url�PYPI_MD5�sub)	r}r8rvrDr�r��new_urlr+�fragr!r!r"r��s$"
�zPackageIndex.process_indexcCs|�d|�dS)NzPPage at %s links to .py file(s) without version info; an index scan is required.)�scan_all�r}r8r!r!r"r��s�zPackageIndex.need_version_infocGs<|j|jvr,|r"|j|g|�R�|�d�|�|j�dS)Nz6Scanning index of all packages (this may take a while))r�r�r�r�r��r}r�rcr!r!r"r��s�zPackageIndex.scan_allcCsz|�|j|jd�|j�|j�s:|�|j|jd�|j�|j�sR|�|�t|j�|jd��D]}|�|�qfdS)Nr.r!)	r�r��unsafe_namer�r��keyrK�not_found_in_indexr�)r}�requirementr8r!r!r"�
find_packages�s
zPackageIndex.find_packagescsR|��|�|�||jD]"}||vr0|S|�d||�qtt|��||�S)Nz%s does not match %s)�prescanr�r�r��superr�obtain)r}r��	installerrG��	__class__r!r"r��s
zPackageIndex.obtaincCsL|�|jd|�|��sH|��t�|�td|jjtj	�
|�f��dS)z-
        checker is a ContentChecker
        zValidating %%s checksum for %sz7%s validation failed for %s; possible download problem?N)r�r�r�r�rU�unlinkrr�r*r<rQ)r}�checkerrV�tfpr!r!r"�
check_hash�s�
��zPackageIndex.check_hashcCsN|D]D}|jdus0t|�r0|�d�s0tt|��r<|�|�q|j�|�qdS)z;Add `urls` to the list that will be prescanned for searchesN�file:)r�r�r)r�rr��append)r}�urlsr8r!r!r"�add_find_links�s���
�zPackageIndex.add_find_linkscCs"|jrtt|j|j��d|_dS)z7Scan urls scheduled for prescanning (e.g. --find-links)N)r�r�rnr�r�r!r!r"r�szPackageIndex.prescancCs<||jr|jd}}n|jd}}|||j�|��dS)Nz#Couldn't retrieve index page for %rz3Couldn't find index page for %r (maybe misspelled?))r�r�r�r�r�)r}r��methr�r!r!r"r�s
�zPackageIndex.not_found_in_indexcCs~t|t�sjt|�}|rR|�|�d�||�}t|�\}}|�d�rN|�|||�}|Stj	�
|�rb|St|�}t|�
||�dd�S)aLocate and/or download `spec` to `tmpdir`, returning a local path

        `spec` may be a ``Requirement`` object, or a string containing a URL,
        an existing local filename, or a project/version requirement spec
        (i.e. the string form of a ``Requirement`` object).  If it is the URL
        of a .py file with an unambiguous ``#egg=name-version`` tag (i.e., one
        that escapes ``-`` as ``_`` throughout), a trivial ``setup.py`` is
        automatically created alongside the downloaded file.

        If `spec` is a ``Requirement`` object or a string containing a
        project/version requirement spec, this method returns the location of
        a matching distribution (possibly after downloading it to `tmpdir`).
        If `spec` is a locally existing file or directory name, it is simply
        returned unchanged.  If `spec` is a URL, it is downloaded to a subpath
        of `tmpdir`, and the local filename is returned.  Various errors may be
        raised if a problem occurs during downloading.
        r3r�rJN)r�rr��
_download_urlrEr@r(�	gen_setuprUr<r�r#r��fetch_distribution)r}r�tmpdirr:�foundr+r?r!r!r"r0s

zPackageIndex.downloadc	s���d|�i�d}d
�����fdd�	}|rH�����|�||�}|s^|dur^|||�}|dur��jdurx���||�}|dur�|s���|�||�}|dur���d�r�dp�d|�n��d|�|j|jd	�SdS)a|Obtain a distribution suitable for fulfilling `requirement`

        `requirement` must be a ``pkg_resources.Requirement`` instance.
        If necessary, or if the `force_scan` flag is set, the requirement is
        searched for in the (online) package index as well as the locally
        installed packages.  If a distribution matching `requirement` is found,
        the returned distribution's ``location`` is the value you would have
        gotten from calling the ``download()`` method with the matching
        distribution's URL or filename.  If no matching distribution is found,
        ``None`` is returned.

        If the `source` flag is set, only source distributions and source
        checkout links will be considered.  Unless the `develop_ok` flag is
        set, development and system eggs (i.e., those using the ``.egg-info``
        format) will be ignored.
        zSearching for %sNcs�|dur�}||jD]v}|jtkrF�sF|�vr��d|�d�|<q||vo\|jtkp\�}|r��|j��}||_tj	�
|j�r|SqdS)Nz&Skipping development or system egg: %sr3)r�rArr�rr0rJ�download_locationrUr<r�)�req�envrG�test�loc��
develop_okr}�skipped�sourcerr!r"rtUs$��z-PackageIndex.fetch_distribution.<locals>.findz:No local packages or working download links found for %s%sza source distribution of r�zBest match: %s)rJ)N)r�r�r�r�r��cloner)	r}r�r�
force_scanrrZlocal_indexrGrtr!r
r"r=s2




�zPackageIndex.fetch_distributioncCs"|�||||�}|dur|jSdS)a3Obtain a file suitable for fulfilling `requirement`

        DEPRECATED; use the ``fetch_distribution()`` method now instead.  For
        backward compatibility, this routine is identical but returns the
        ``location`` of the downloaded distribution instead of a distribution
        object.
        N)rrJ)r}r�rrrrGr!r!r"�fetch�szPackageIndex.fetchc
	Cst�|�}|r*dd�t||�d�d�D�p,g}t|�dkr�tj�|�}tj�|�|kr�tj�	||�}ddl
m}|||�s�t�
||�|}ttj�	|d�d��<}	|	�d|dj|djtj�|�df�Wd�n1s�0Y|S|�rtd	||f��ntd
��dS)NcSsg|]}|jr|�qSr!)rL)rY�dr!r!r"�
<listcomp>�s�z*PackageIndex.gen_setup.<locals>.<listcomp>r3r)�samefilezsetup.py�wzIfrom setuptools import setup
setup(name=%r, version=%r, py_modules=[%r])
z�Can't unambiguously interpret project/version identifier %r; any dashes in the name or version should be escaped using underscores. %rzpCan't process plain .py files without an '#egg=name-version' suffix to enable automatic setup script generation.)rCrDrrErPrUr<rQ�dirnameraZsetuptools.command.easy_installr�shutil�copy2r��writerKrL�splitextr)
r}rVr?rrDr�rQ�dstrr�r!r!r"r�s@

��
���"���zPackageIndex.gen_setupi cCs:|�d|�d}�zt�|�}|�|�}t|tjj�rLtd||j	|j
f��|��}d}|j}d}d|vr�|�d�}	t
tt|	��}|�|||||�t|d��`}
|�|�}|r�|�|�|
�|�|d7}|�|||||�q�q�q�|�|||
�Wd�n1�s0Y|W|�r"|��Sn|�r4|��0dS)	NzDownloading %szCan't download %s: %s %srr/zcontent-lengthzContent-Length�wbr3)r�r�r�r�r�r4r�r�rr�r��dl_blocksize�get_all�maxrn�int�
reporthookr�r�rrr�r�)r}r8rV�fpr�r��blocknum�bs�size�sizesr�r~r!r!r"�_download_to�sD

�



.��zPackageIndex._download_tocCsdSrbr!)r}r8rVr%Zblksizer'r!r!r"r#�szPackageIndex.reporthookc
Cs�|�d�rt|�Szt||j�WSttjjfy�}zHd�dd�|j	D��}|r`|�
||�ntd||f�|�WYd}~�n*d}~0tj
jy�}z|WYd}~Sd}~0tj
j�y}z4|r�|�
||j�ntd||jf�|�WYd}~n�d}~0tjj�yT}z6|�r,|�
||j�ntd||jf�|�WYd}~n^d}~0tjjtj
f�y�}z2|�r�|�
||�ntd||f�|�WYd}~n
d}~00dS)Nr�� cSsg|]}t|��qSr!)ro)rY�argr!r!r"r�r\z)PackageIndex.open_url.<locals>.<listcomp>z%s %szDownload error for %s: %sz;%s returned a bad status line. The server might be down, %s)r)�
local_open�open_with_authr�r�http�client�
InvalidURLrarcr�rr4r�r��URLError�reason�
BadStatusLine�line�
HTTPException�socket)r}r8�warning�vr�r!r!r"r��sJ
(�������zPackageIndex.open_urlcCs�t|�\}}|r0d|vr4|�dd��dd�}qnd}|�d�rJ|dd�}tj�||�}|dksj|�d	�rv|�||�S|d
ks�|�d�r�|�||�S|�d�r�|�	||�S|d
kr�t
j�t
j
�|�d�S|�|d�|�||�SdS)Nz..�.�\�_Z__downloaded__rHr&�svnzsvn+�gitzgit+zhg+r�r]T)r@�replacer(rUr<rar)�
_download_svn�
_download_git�_download_hgr4r��url2pathnamerr5r��_attempt_download)r}r:r8rr*r?rVr!r!r"rs$

zPackageIndex._download_urlcCs|�|d�dS)NT)r�r�r!r!r"r�)szPackageIndex.scan_urlcCs6|�||�}d|�dd���vr.|�|||�S|SdS)Nr�r�r�)r)r�r'�_download_html)r}r8rVr�r!r!r"rC,szPackageIndex._attempt_downloadcCsnt|�}|D]>}|��rt�d|�rF|��t�|�|�||�SqLq|��t�|�td|��dS)Nz <title>([^- ]+ - )?Revision \d+:zUnexpected HTML page found at )	r�rprXrur�rUr�r?r)r}r8r�rVr�r4r!r!r"rD3s

zPackageIndex._download_htmlcCst�dt�|�dd�d}d}|���d�r�d|vr�tj�|�\}}}}}}	|s�|�d�r�d	|d
d�vr�|d
d��d	d�\}}t	|�\}
}|
r�d|
vr�|
�dd�\}}
d||
f}nd
|
}|}||||||	f}tj�
|�}|�d||�t�
d|||f�|S)Nz"SVN download support is deprecatedr2r3rr�zsvn:�@z//r.r]�:z --username=%s --password=%sz --username=z'Doing subversion checkout from %s to %szsvn checkout%s -q %s %s)�warningsr��UserWarningr7r'r)r4rr5�
_splituser�
urlunparser�rU�system)r}r8rV�credsr:�netlocr<rZ�qr��auth�host�user�pwr9r!r!r"r?Bs&zPackageIndex._download_svncCsptj�|�\}}}}}|�dd�d}|�dd�d}d}d|vrR|�dd�\}}tj�||||df�}||fS)N�+r3r/r2rrEr�)r4r�urlsplitr7�rsplit�
urlunsplit)r8�
pop_prefixr:rMr<r>r��revr!r!r"�_vcs_split_rev_from_urlXsz$PackageIndex._vcs_split_rev_from_urlcCsl|�dd�d}|j|dd�\}}|�d||�t�d||f�|durh|�d|�t�d	||f�|S)
Nr2r3rT�rWzDoing git clone from %s to %szgit clone --quiet %s %szChecking out %szgit -C %s checkout --quiet %s�r7rYr�rUrK�r}r8rVrXr!r!r"r@js�zPackageIndex._download_gitcCsl|�dd�d}|j|dd�\}}|�d||�t�d||f�|durh|�d|�t�d	||f�|S)
Nr2r3rTrZzDoing hg clone from %s to %szhg clone --quiet %s %szUpdating to %szhg --cwd %s up -C -r %s -qr[r\r!r!r"rAzs�zPackageIndex._download_hgcGstj|g|�R�dSrb)rr�r�r!r!r"r��szPackageIndex.debugcGstj|g|�R�dSrb)rr�r�r!r!r"r��szPackageIndex.infocGstj|g|�R�dSrb)rr�r�r!r!r"r��szPackageIndex.warn)r�r�NT)F)F)F)N)N)FFFN)FF)N)F))r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rr�r�r0rrrrr)r#r�rr�rCrDr?�staticmethodrYr@rAr�r�r��
__classcell__r!r!r�r"rsR�

5




		
#�
L

)$
#rz!&(#(\d+|x[\da-fA-F]+)|[\w.:-]+);?cCs|�d�}t�|�S)Nr)rEr��unescape)rD�whatr!r!r"�
decode_entity�s
racCs
tt|�S)a
    Decode HTML entities in the given text.

    >>> htmldecode(
    ...     'https://../package_name-0.1.2.tar.gz'
    ...     '?tokena=A&amp;tokenb=B">package_name-0.1.2.tar.gz')
    'https://../package_name-0.1.2.tar.gz?tokena=A&tokenb=B">package_name-0.1.2.tar.gz'
    )�
entity_subra)�textr!r!r"rs�s	rscs�fdd�}|S)Ncs��fdd�}|S)Nc	s>t��}t���z�|i|��Wt�|�St�|�0dSrb)r6�getdefaulttimeout�setdefaulttimeout)rcrdZold_timeout)rf�timeoutr!r"�_socket_timeout�s

�z@socket_timeout.<locals>._socket_timeout.<locals>._socket_timeoutr!)rfrg�rfrer"rg�sz'socket_timeout.<locals>._socket_timeoutr!)rfrgr!rhr"�socket_timeout�sricCs2tj�|�}|��}t�|�}|��}|�dd�S)a9
    Encode auth from a URL suitable for an HTTP header.
    >>> str(_encode_auth('username%3Apassword'))
    'dXNlcm5hbWU6cGFzc3dvcmQ='

    Long auth strings should not cause a newline to be inserted.
    >>> long_auth = 'username:' + 'password'*10
    >>> chr(10) in str(_encode_auth(long_auth))
    False
    �
r�)r4rr6�encode�base64�	b64encoder�r>)rOZauth_sZ
auth_bytesZ
encoded_bytes�encodedr!r!r"�_encode_auth�s

roc@s(eZdZdZdd�Zdd�Zdd�ZdS)	�
Credentialz:
    A username/password pair. Use like a namedtuple.
    cCs||_||_dSrb��username�password)r}rrrsr!r!r"r��szCredential.__init__ccs|jV|jVdSrbrqr�r!r!r"�__iter__�szCredential.__iter__cCsdt|�S)Nz%(username)s:%(password)s)�varsr�r!r!r"�__str__�szCredential.__str__N)r�r�r�r�r�rtrvr!r!r!r"rp�srpc@s0eZdZdd�Zedd��Zdd�Zdd�Zd	S)
�
PyPIConfigcCsNt�gd�d�}tj�||�tj�tj�d�d�}tj�	|�rJ|�
|�dS)z%
        Load from ~/.pypirc
        )rrrs�
repositoryr��~z.pypircN)�dict�fromkeys�configparser�RawConfigParserr�rUr<ra�
expanduserr�r�)r}�defaults�rcr!r!r"r��s
zPyPIConfig.__init__cs&�fdd����D�}tt�j|��S)Ncs g|]}��|d���r|�qS)rx)r�rp)rY�sectionr�r!r"r�s�z2PyPIConfig.creds_by_repository.<locals>.<listcomp>)�sectionsrzrn�_get_repo_cred)r}Zsections_with_repositoriesr!r�r"�creds_by_repository�s
�zPyPIConfig.creds_by_repositorycCs6|�|d���}|t|�|d���|�|d����fS)Nrxrrrs)r�rprp)r}r��repor!r!r"r��s
�zPyPIConfig._get_repo_credcCs*|j��D]\}}|�|�r
|Sq
dS)z�
        If the URL indicated appears to be a repository defined in this
        config, return the credential for that repository.
        N)r��itemsr))r}r8rx�credr!r!r"�find_credential�s
zPyPIConfig.find_credentialN)r�r�r�r��propertyr�r�r�r!r!r!r"rw�s

rwcCs@tj�|�}|\}}}}}}|�d�r2tj�d��|dvrHt|�\}	}
nd}	|	s�t��	|�}|r�t
|�}	|j|f}tj
dg|�R�|	r�dt|	�}	||
||||f}
tj�|
�}tj�|�}|�d|	�ntj�|�}|�dt�||�}|	�r<tj�|j�\}}}}}}||k�r<||
k�r<||||||f}
tj�|
�|_|S)	z4Open a urllib2 request, handling HTTP authenticationrFznonnumeric port: '')r.�httpsNz*Authenticating as %s for %s (from .pypirc)zBasic �
Authorizationz
User-Agent)r4rr5r(r.r/r0rIrwr�rorrrr�rorJr��Request�
add_header�
user_agentr8)r8r��parsedr:rMr<�paramsr>r�rO�addressr�r�r9r�r�r$�s2�h2Zpath2Zparam2Zquery2Zfrag2r!r!r"r-s8

r-cCs |�d�\}}}|r|nd|fS)zNsplituser('user[:passwd]@host[:port]')
    --> 'user[:passwd]', 'host[:port]'.rEN)�
rpartition)rPrQ�delimr!r!r"rI4srIcCs|Srbr!)r8r!r!r"�
fix_sf_url?sr�c	Cs,tj�|�\}}}}}}tj�|�}tj�|�r<tj�|�S|�	d�r�tj�
|�r�g}t�|�D]x}	tj�||	�}
|	dkr�t
|
d��}|��}Wd�n1s�0Yq�ntj�
|
�r�|	d7}	|�dj|	d��q`d}
|
j|d�|�d	�}d
\}}n
d\}}}dd
i}t�|�}tj�|||||�S)z7Read a local path, with special support for directoriesr.z
index.html�rNz<a href="{name}">{name}</a>)r*zB<html><head><title>{url}</title></head><body>{files}</body></html>rj)r8�files)���OK)r�zPath not foundz	Not foundr�z	text/html)r4rr5r�rBrUr<�isfiler�r(r�r�rar�r�r��format�io�StringIOr�r�)r8r:r;r<�paramr>r�rVr�r��filepathr$�bodyr��status�messager�Zbody_streamr!r!r"r,Cs.&�


r,)N)N)N)r)Yr��sysrUrXr�rr6rlr�r�rGr|r��http.clientr.�urllib.parser4�urllib.request�urllib.error�	functoolsrr�
pkg_resourcesrrrrrr	r
rrr
rrr�	distutilsr�distutils.errorsr�fnmatchrZsetuptools.wheelrZ setuptools.extern.more_itertoolsrr�rC�Irqr�rDr�r7rO�__all__Z_SOCKET_TIMEOUTZ_tmplr��version_infor�r#rr@rrBrWrrhrjrzr{r�rr�rbrarsrirorpr}rwr�r�r-rIr�r,r!r!r!r"�<module>s�<
��	

!
�
#

!
&/site-packages/setuptools/__pycache__/glob.cpython-39.pyc000064400000007236150564745010017310 0ustar00a

(�Re	�@s�dZddlZddlZddlZgd�Zddd�Zddd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Ze�d�Z
e�d�Zdd�Zdd�Zdd�ZdS)z�
Filename globbing utility. Mostly a copy of `glob` from Python 3.5.

Changes include:
 * `yield from` and PEP3102 `*` removed.
 * Hidden files are not ignored.
�N)�glob�iglob�escapeFcCstt||d��S)ayReturn a list of paths matching a pathname pattern.

    The pattern may contain simple shell-style wildcards a la
    fnmatch. However, unlike fnmatch, filenames starting with a
    dot are special cases that are not matched by '*' and '?'
    patterns.

    If recursive is true, the pattern '**' will match any files and
    zero or more directories and subdirectories.
    )�	recursive)�listr)�pathnamer�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/glob.pyrsrcCs*t||�}|r&t|�r&t|�}|r&J�|S)a�Return an iterator which yields the paths matching a pathname pattern.

    The pattern may contain simple shell-style wildcards a la
    fnmatch. However, unlike fnmatch, filenames starting with a
    dot are special cases that are not matched by '*' and '?'
    patterns.

    If recursive is true, the pattern '**' will match any files and
    zero or more directories and subdirectories.
    )�_iglob�_isrecursive�next)rr�it�srrr	rs

rccs�tj�|�\}}|r t|�r tnt}t|�sZ|rDtj�|�rV|Vntj�|�rV|VdS|sr|||�EdHdS||kr�t|�r�t	||�}n|g}t|�s�t
}|D]$}|||�D]}tj�||�Vq�q�dS�N)�os�path�splitr�glob2�glob1�	has_magic�lexists�isdirr
�glob0�join)rr�dirname�basename�glob_in_dir�dirs�namerrr	r
0s(r
cCsT|s"t|t�rtj�d�}ntj}zt�|�}WntyFgYS0t�||�S�N�ASCII)	�
isinstance�bytesr�curdir�encode�listdir�OSError�fnmatch�filter)r�pattern�namesrrr	rTs

rcCs8|stj�|�r4|gSntj�tj�||��r4|gSgSr)rrrrr)rrrrr	rasrccs2t|�sJ�|dd�Vt|�D]
}|Vq"dS)Nr)r�	_rlistdir)rr)�xrrr	rqsrccs�|s"t|t�rtj�d�}ntj}zt�|�}WntjyFYdS0|D]>}|V|rhtj�||�n|}t	|�D]}tj�||�VqtqLdSr)
r!r"rr#r$r%�errorrrr+)rr*r,r�yrrr	r+ys
r+z([*?[])s([*?[])cCs(t|t�rt�|�}n
t�|�}|duSr)r!r"�magic_check_bytes�search�magic_check)r�matchrrr	r�s

rcCst|t�r|dkS|dkSdS)Ns**z**)r!r")r)rrr	r�s
rcCs<tj�|�\}}t|t�r(t�d|�}nt�d|�}||S)z#Escape all special characters.
    s[\1]z[\1])rr�
splitdriver!r"r/�subr1)r�driverrr	r�s

r)F)F)�__doc__r�rer'�__all__rrr
rrrr+�compiler1r/rrrrrrr	�<module>s 

$


site-packages/setuptools/__pycache__/config.cpython-39.pyc000064400000050612150564745010017626 0ustar00a

(�ReSZ�@sddlZddlZddlZddlZddlZddlZddlZddlmZddlm	Z	ddlm
Z
ddlmZddl
Z
ddlmZmZddlmZmZddlmZGd	d
�d
�Ze
jdd��Zddd�Zdd�Zdd�Zddd�ZGdd�d�ZGdd�de�ZGdd�de�ZdS)�N)�defaultdict)�partial��wraps)�iglob)�DistutilsOptionError�DistutilsFileError)�
LegacyVersion�parse)�SpecifierSetc@s eZdZdZdd�Zdd�ZdS)�StaticModulez0
    Attempt to load the module by the name
    cCs`tj�|�}t|j��}|��}Wd�n1s40Yt�|�}t|��	t
��|`dS�N)�	importlib�util�	find_spec�open�origin�read�astr
�vars�update�locals�self)r�name�spec�strm�src�module�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/config.py�__init__s&
zStaticModule.__init__c
s^zt�fdd�|jjD��WStyX}z$tdjfit����|�WYd}~n
d}~00dS)Nc3sH|]@}t|tj�r|jD](}t|tj�r|j�krt�|j�VqqdSr
)�
isinstancer�Assign�targets�Name�id�literal_eval�value)�.0Z	statement�target��attrrr�	<genexpr>#s

�z+StaticModule.__getattr__.<locals>.<genexpr>z#{self.name} has no attribute {attr})�nextr�body�	Exception�AttributeError�formatr)rr+�err*r�__getattr__!s�
��zStaticModule.__getattr__N)�__name__�
__module__�__qualname__�__doc__r r3rrrrrsrc	cs8z$tj�d|�dVWtj�|�ntj�|�0dS)zH
    Add path to front of sys.path for the duration of the context.
    rN)�sys�path�insert�remove)r9rrr�
patch_path0sr<Fc		Cs�ddlm}m}tj�|�}tj�|�s4td|��t��}t�	tj�
|��zT|�}|rb|��ng}||vrx|�|�|j
||d�t||j|d�}Wt�	|�nt�	|�0t|�S)a,Read given configuration file and returns options from it as a dict.

    :param str|unicode filepath: Path to configuration file
        to get options from.

    :param bool find_others: Whether to search for other configuration files
        which could be on in various places.

    :param bool ignore_option_errors: Whether to silently ignore
        options, values of which could not be resolved (e.g. due to exceptions
        in directives such as file:, attr:, etc.).
        If False exceptions are propagated as expected.

    :rtype: dict
    r)�Distribution�
_Distributionz%Configuration file %s does not exist.)�	filenames)�ignore_option_errors)Zsetuptools.distr=r>�osr9�abspath�isfiler�getcwd�chdir�dirname�find_config_files�append�parse_config_files�parse_configuration�command_options�configuration_to_dict)	�filepathZfind_othersr@r=r>Zcurrent_directory�distr?�handlersrrr�read_configuration<s"
�rPcCs2djfit���}t�t||�}t|||�}|�S)z�
    Given a target object and option key, get that option from
    the target object, either through a get_{key} method or
    from an attribute directly.
    z	get_{key})r1r�	functoolsr�getattr)�
target_obj�keyZgetter_nameZby_attribute�getterrrr�_get_optionisrVcCs<tt�}|D]*}|jD]}t|j|�}|||j|<qq|S)z�Returns configuration data gathered by given handlers as a dict.

    :param list[ConfigHandler] handlers: Handlers list,
        usually from parse_configuration()

    :rtype: dict
    )r�dict�set_optionsrVrS�section_prefix)rOZconfig_dict�handler�optionr'rrrrLus
rLcCs6t|||�}|��t|j|||j�}|��||fS)a�Performs additional parsing of configuration options
    for a distribution.

    Returns a list of used option handlers.

    :param Distribution distribution:
    :param dict command_options:
    :param bool ignore_option_errors: Whether to silently ignore
        options, values of which could not be resolved (e.g. due to exceptions
        in directives such as file:, attr:, etc.).
        If False exceptions are propagated as expected.
    :rtype: list
    )�ConfigOptionsHandlerr
�ConfigMetadataHandler�metadata�package_dir)�distributionrKr@�options�metarrrrJ�s�rJc@s�eZdZdZdZiZd'dd�Zedd��Zdd	�Z	e
d(dd��Ze
d)d
d��Ze
dd��Z
e
dd��Ze
dd��Ze
dd��Zedd��Zedd��Ze
d*dd��Ze
dd��Ze
d+dd ��Zd!d"�Zd#d$�Zd%d&�ZdS),�
ConfigHandlerz1Handles metadata supplied in configuration files.NFcCs^i}|j}|��D].\}}|�|�s&q|�|d��d�}|||<q||_||_||_g|_dS)N��.)	rY�items�
startswith�replace�stripr@rS�sectionsrX)rrSrar@rjrY�section_name�section_optionsrrrr �s

zConfigHandler.__init__cCstd|jj��dS)�.Metadata item name to parser function mapping.z!%s must provide .parsers propertyN)�NotImplementedError�	__class__r4)rrrr�parsers�s
�zConfigHandler.parsersc	Cs�t�}|j}|j�||�}t|||�}||ur6t|��|r>dSd}|j�|�}|r~z||�}Wnty|d}|jsx�Yn0|r�dSt|d|d�}|dur�t	|||�n||�|j
�|�dS)NFTzset_%s)�tuplerS�aliases�getrR�KeyErrorrpr/r@�setattrrXrH)	rZoption_namer'�unknownrSZ
current_valueZskip_option�parser�setterrrr�__setitem__�s0zConfigHandler.__setitem__�,cCs8t|t�r|Sd|vr |��}n
|�|�}dd�|D�S)z�Represents value as a list.

        Value is split either by separator (defaults to comma) or by lines.

        :param value:
        :param separator: List items separator character.
        :rtype: list
        �
cSsg|]}|��r|���qSr)ri)r(�chunkrrr�
<listcomp>�z-ConfigHandler._parse_list.<locals>.<listcomp>)r!�list�
splitlines�split)�clsr'�	separatorrrr�_parse_list�s



zConfigHandler._parse_listc	sjd}|j�|d�}g}|D]J�t�fdd�|D��rZ|�tdd�ttj����D���q|���q|S)aEquivalent to _parse_list() but expands any glob patterns using glob().

        However, unlike with glob() calls, the results remain relative paths.

        :param value:
        :param separator: List items separator character.
        :rtype: list
        )�*�?�[�]�{�}�r�c3s|]}|�vVqdSr
r)r(�char�r'rrr,r~z1ConfigHandler._parse_list_glob.<locals>.<genexpr>css |]}tj�|t���VqdSr
)rAr9�relpathrD�r(r9rrrr,s�)	r��any�extend�sortedrrAr9rBrH)r�r'r�Zglob_characters�valuesZexpanded_valuesrr�r�_parse_list_globs
�zConfigHandler._parse_list_globcCsPd}i}|�|�D]8}|�|�\}}}||kr:td|��|��||��<q|S)zPRepresents value as a dict.

        :param value:
        :rtype: dict
        �=z(Unable to parse option value to dict: %s)r��	partitionrri)r�r'r��result�linerT�sep�valrrr�_parse_dict s�zConfigHandler._parse_dictcCs|��}|dvS)zQRepresents value as boolean.

        :param value:
        :rtype: bool
        )�1�true�yes)�lower)r�r'rrr�_parse_bool3szConfigHandler._parse_boolcs�fdd�}|S)z�Returns a parser function to make sure field inputs
        are not files.

        Parses a value after getting the key so error messages are
        more informative.

        :param key:
        :rtype: callable
        cs d}|�|�rtd�����|S)N�file:zCOnly strings are accepted for the {0} field, files are not accepted)rg�
ValueErrorr1)r'Zexclude_directive�rTrrrwIs
��z3ConfigHandler._exclude_files_parser.<locals>.parserr)r�rTrwrr�r�_exclude_files_parser=s	z#ConfigHandler._exclude_files_parsercs\d}t|t�s|S|�|�s |S|t|�d�}dd�|�d�D�}d��fdd�|D��S)aORepresents value as a string, allowing including text
        from nearest files using `file:` directive.

        Directive is sandboxed and won't reach anything outside
        directory with setup.py.

        Examples:
            file: README.rst, CHANGELOG.md, src/file.txt

        :param str value:
        :rtype: str
        r�Ncss|]}tj�|���VqdSr
)rAr9rBrir�rrrr,kr~z,ConfigHandler._parse_file.<locals>.<genexpr>rzr{c3s.|]&}��|�stj�|�r��|�VqdS)TN)�
_assert_localrAr9rC�
_read_filer��r�rrr,ls�)r!�strrg�lenr��join)r�r'Zinclude_directiverZ	filepathsrr�r�_parse_fileTs

�zConfigHandler._parse_filecCs|�t���std|��dS)Nz#`file:` directive can not access %s)rgrArDr)rMrrrr�rszConfigHandler._assert_localcCs:tj|dd��}|��Wd�S1s,0YdS)Nzutf-8)�encoding)�iorr)rM�frrrr�wszConfigHandler._read_filec	Cs4d}|�|�s|S|�|d����d�}|��}d�|�}|p@d}t��}|r�|d|vr�||d}|�dd�}	t	|	�dkr�tj
�t��|	d�}|	d}q�|}nd|vr�tj
�t��|d�}t|��Nztt
|�|�WWd�St�yt�|�}
Yn0Wd�n1�s 0Yt|
|�S)	z�Represents value as a module attribute.

        Examples:
            attr: package.attr
            attr: package.module.attr

        :param str value:
        :rtype: str
        zattr:rdrer r�/�N)rgrhrir��popr�rArD�rsplitr�r9r<rRrr/r�
import_module)r�r'r_Zattr_directiveZ
attrs_path�	attr_name�module_name�parent_pathZcustom_path�partsrrrr�_parse_attr|s0



0zConfigHandler._parse_attrcs�fdd�}|S)z�Returns parser function to represents value as a list.

        Parses a value applying given methods one after another.

        :param parse_methods:
        :rtype: callable
        cs|}�D]}||�}q|Sr
r)r'�parsed�method��
parse_methodsrrr
�s
z1ConfigHandler._get_parser_compound.<locals>.parser)r�r�r
rr�r�_get_parser_compound�s
z"ConfigHandler._get_parser_compoundcCs6i}|pdd�}|��D]\}\}}||�||<q|S)z�Parses section options into a dictionary.

        Optionally applies a given parser to values.

        :param dict section_options:
        :param callable values_parser:
        :rtype: dict
        cSs|Sr
r)r�rrr�<lambda>�r~z6ConfigHandler._parse_section_to_dict.<locals>.<lambda>)rf)r�rlZ
values_parserr'rT�_r�rrr�_parse_section_to_dict�s

z$ConfigHandler._parse_section_to_dictc	Cs:|��D],\}\}}z|||<Wqty2Yq0qdS)zQParses configuration file section.

        :param dict section_options:
        N)rfrt)rrlrr�r'rrr�
parse_section�s
zConfigHandler.parse_sectioncCsb|j��D]R\}}d}|r"d|}t|d|�dd�d�}|durTtd|j|f��||�q
dS)zTParses configuration file items from one
        or more related sections.

        rdz_%szparse_section%sre�__Nz0Unsupported distribution option section: [%s.%s])rjrfrRrhrrY)rrkrlZmethod_postfixZsection_parser_methodrrrr
�s ���zConfigHandler.parsecst�����fdd��}|S)z�this function will wrap around parameters that are deprecated

        :param msg: deprecation message
        :param warning_class: class of warning exception to be raised
        :param func: function to be wrapped around
        cst�����|i|��Sr
)�warnings�warn)�args�kwargs��func�msg�
warning_classrr�config_handler�sz@ConfigHandler._deprecated_config_handler.<locals>.config_handlerr)rr�r�r�r�rr�r�_deprecated_config_handler�sz(ConfigHandler._deprecated_config_handler)F)rz)rz)N)N)r4r5r6r7rYrrr �propertyrpry�classmethodr�r�r�r�r�r��staticmethodr�r�r�r�r�r�r
r�rrrrrc�s@

&

	



-
rccsHeZdZdZddddd�ZdZd�fd	d
�	Zedd��Zd
d�Z	�Z
S)r]r^�url�description�classifiers�	platforms)Z	home_page�summary�
classifier�platformFNcstt|��|||�||_dSr
)�superr]r r_)rrSrar@r_�rorrr s
�zConfigMetadataHandler.__init__cCs^|j}|j}|j}|j}||||�|dt�||�||�|d�|�|d�dt�||||j|d�
S)rmz[The requires parameter is deprecated, please use install_requires for runtime dependencies.�license�license_filezDThe license_file parameter is deprecated, use license_files instead.)
r��keywords�provides�requires�	obsoletesr�r�r�Z
license_filesr��long_description�versionZproject_urls)r�r�r�r�r��DeprecationWarningr��_parse_version)r�
parse_listZ
parse_file�
parse_dictZexclude_files_parserrrrrps4�
��zConfigMetadataHandler.parserscCs�|�|�}||krF|��}tt|�t�rBd}t|jfit�����|S|�||j	�}t
|�rb|�}t|t�s�t|d�r�d�
tt|��}nd|}|S)zSParses `version` option value.

        :param value:
        :rtype: str

        zCVersion loaded from {value} does not comply with PEP 440: {version}�__iter__rez%s)r�rir!r
r	rr1rr�r_�callabler��hasattrr��map)rr'r��tmplrrrr�?s 
�

z$ConfigMetadataHandler._parse_version)FN)r4r5r6rYrrZstrict_moder r�rpr��
__classcell__rrr�rr]s��
!r]c@sdeZdZdZedd��Zdd�Zdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�ZdS)r\racCsN|j}t|jdd�}|j}|j}|j}|||||||||||j|j|t|d�S)rm�;r�)Zzip_safeZinclude_package_datar_�scriptsZeager_resourcesZdependency_linksZnamespace_packagesZinstall_requiresZsetup_requiresZ
tests_require�packages�entry_points�
py_modulesZpython_requires�cmdclass)r�rr�r��_parse_cmdclass�_parse_packagesr�r)rr�Zparse_list_semicolonZ
parse_boolr�Zparse_cmdclassrrrrpgs*�zConfigOptionsHandler.parserscs$dd���fdd�|�|���D�S)NcSs8|�d�}||dd�}|d|�}t|�}t||�S)Nrer�)�rfind�
__import__rR)Zqualified_class_name�idx�
class_name�pkg_namerrrr�
resolve_class�s

z;ConfigOptionsHandler._parse_cmdclass.<locals>.resolve_classcsi|]\}}|�|��qSrr�r(�k�v�r�rr�
<dictcomp>�r~z8ConfigOptionsHandler._parse_cmdclass.<locals>.<dictcomp>)r�rf)rr'rr�rr��s	z$ConfigOptionsHandler._parse_cmdclasscCsnddg}|��}||vr"|�|�S||dk}|�|j�di��}|rTddlm}nddlm}|fi|��S)zTParses `packages` option value.

        :param value:
        :rtype: list
        zfind:zfind_namespace:r�z
packages.findr)�find_namespace_packages)�
find_packages)rir��parse_section_packages__findrjrs�
setuptoolsr�r�)rr'Zfind_directivesZ
trimmed_valueZfindns�find_kwargsr�rrrr��s
�z$ConfigOptionsHandler._parse_packagescsR|�||j�}gd��t�fdd�|��D��}|�d�}|durN|d|d<|S)z�Parses `packages.find` configuration file section.

        To be used in conjunction with _parse_packages().

        :param dict section_options:
        )�where�include�excludecs$g|]\}}|�vr|r||f�qSrrr��Z
valid_keysrrr}�r~zEConfigOptionsHandler.parse_section_packages__find.<locals>.<listcomp>rNr)r�r�rWrfrs)rrlZsection_datarrrrrr�s�
z1ConfigOptionsHandler.parse_section_packages__findcCs|�||j�}||d<dS)z`Parses `entry_points` configuration file section.

        :param dict section_options:
        r�N)r�r��rrlr�rrr�parse_section_entry_points�sz/ConfigOptionsHandler.parse_section_entry_pointscCs.|�||j�}|�d�}|r*||d<|d=|S)Nr�rd)r�r�rs)rrlr��rootrrr�_parse_package_data�s
z(ConfigOptionsHandler._parse_package_datacCs|�|�|d<dS)z`Parses `package_data` configuration file section.

        :param dict section_options:
        �package_dataN�r
�rrlrrr�parse_section_package_data�sz/ConfigOptionsHandler.parse_section_package_datacCs|�|�|d<dS)zhParses `exclude_package_data` configuration file section.

        :param dict section_options:
        Zexclude_package_dataNrr
rrr�"parse_section_exclude_package_data�sz7ConfigOptionsHandler.parse_section_exclude_package_datacCs"t|jdd�}|�||�|d<dS)zbParses `extras_require` configuration file section.

        :param dict section_options:
        r�r�Zextras_requireN)rr�r�)rrlr�rrr�parse_section_extras_require�s�z1ConfigOptionsHandler.parse_section_extras_requirecCs(|�||j�}dd�|��D�|d<dS)z^Parses `data_files` configuration file section.

        :param dict section_options:
        cSsg|]\}}||f�qSrrr�rrrr}�r~zAConfigOptionsHandler.parse_section_data_files.<locals>.<listcomp>�
data_filesN)r�r�rfrrrr�parse_section_data_files�sz-ConfigOptionsHandler.parse_section_data_filesN)r4r5r6rYr�rpr�r�rrr
rrrrrrrrr\cs


r\)FF)F) rr�rAr8r�rQr�collectionsrrr�globr�
contextlib�distutils.errorsrrZ#setuptools.extern.packaging.versionr	r
Z&setuptools.extern.packaging.specifiersrr�contextmanagerr<rPrVrLrJrcr]r\rrrr�<module>s4

-
c_site-packages/setuptools/__pycache__/__init__.cpython-39.pyc000064400000021052150564745010020114 0ustar00a

(�Re��@sFdZddlmZddlZddlZddlZddlZddlZddl	Z
ddlmZddl
mZddlmZddlZddlmZdd	lmZdd
lmZddlmZgd�ZejjZdZGd
d�d�ZGdd�de�Z ej!Z"e j!Z#dd�Z$dd�Z%e
j&j%je%_e�'e
j&j(�Z)Gdd�de)�Z(dd�Z*ej+fdd�Z,Gdd�de-�Z.e�/�dS)z@Extensions to the 'distutils' for large or complex distributions���fnmatchcaseN)�DistutilsOptionError)�convert_path�)�SetuptoolsDeprecationWarning)�	Extension)�Distribution)�Require)�monkey)�setupr	�Commandrr
r�
find_packages�find_namespace_packagesc@sBeZdZdZeddd��Zedd��Zed	d
��Zedd��Z	d
S)�
PackageFinderzI
    Generate a list of all Python packages found within a directory
    �.���*cCs,t|�t|�|jddg|�R�|j|���S)a	Return a list all Python packages found within directory 'where'

        'where' is the root directory which will be searched for packages.  It
        should be supplied as a "cross-platform" (i.e. URL-style) path; it will
        be converted to the appropriate local path syntax.

        'exclude' is a sequence of package names to exclude; '*' can be used
        as a wildcard in the names, such that 'foo.*' will exclude all
        subpackages of 'foo' (but not 'foo' itself).

        'include' is a sequence of package names to include.  If it's
        specified, only the named packages will be included.  If it's not
        specified, all found packages will be included.  'include' can contain
        shell style wildcard patterns just like 'exclude'.
        Zez_setupz*__pycache__)�list�_find_packages_iterr�
_build_filter)�cls�where�exclude�includerr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/__init__.py�find-s��zPackageFinder.findccs�tj|dd�D]�\}}}|dd�}g|dd�<|D]d}tj�||�}	tj�|	|�}
|
�tjjd�}d|vs4|�|	�sxq4||�r�||�s�|V|�|�q4qdS)zy
        All the packages found in 'where' that pass the 'include' filter, but
        not the 'exclude' filter.
        T��followlinksNr)	�os�walk�path�join�relpath�replace�sep�_looks_like_package�append)rrrr�root�dirs�filesZall_dirs�dir�	full_path�rel_path�packagerrrrGsz!PackageFinder._find_packages_itercCstj�tj�|d��S)z%Does a directory look like a package?z__init__.py)r r"�isfiler#�r"rrrr'csz!PackageFinder._looks_like_packagecs�fdd�S)z�
        Given a list of patterns, return a callable that will be true only if
        the input matches at least one of the patterns.
        cst�fdd��D��S)Nc3s|]}t�|d�VqdS))�patNr)�.0r2��namerr�	<genexpr>n�z@PackageFinder._build_filter.<locals>.<lambda>.<locals>.<genexpr>)�anyr4��patternsr4r�<lambda>nr7z-PackageFinder._build_filter.<locals>.<lambda>rr9rr9rrhszPackageFinder._build_filterN)rrr)
�__name__�
__module__�__qualname__�__doc__�classmethodrr�staticmethodr'rrrrrr(s

rc@seZdZedd��ZdS)�PEP420PackageFindercCsdS)NTrr1rrrr'rsz'PEP420PackageFinder._looks_like_packageN)r<r=r>rAr'rrrrrBqsrBcCsJGdd�dtjj�}||�}|jdd�|jrFt�dt�|�|j�dS)Nc@s eZdZdZdd�Zdd�ZdS)z4_install_setup_requires.<locals>.MinimalDistributionzl
        A minimal version of a distribution for supporting the
        fetch_build_eggs interface.
        cs6d}�fdd�t|�t��@D�}tjj�||�dS)N)Zdependency_links�setup_requirescsi|]}|�|�qSrr)r3�k��attrsrr�
<dictcomp>�r7zQ_install_setup_requires.<locals>.MinimalDistribution.__init__.<locals>.<dictcomp>)�set�	distutils�corer	�__init__)�selfrFZ_incl�filteredrrErrK�sz=_install_setup_requires.<locals>.MinimalDistribution.__init__cSsdS)zl
            Disable finalize_options to avoid building the working set.
            Ref #2158.
            Nr)rLrrr�finalize_options�szE_install_setup_requires.<locals>.MinimalDistribution.finalize_optionsN)r<r=r>r?rKrNrrrr�MinimalDistribution~srOT)Zignore_option_errorszdsetup_requires is deprecated. Supply build dependencies using PEP 517 pyproject.toml build-requires.)	rIrJr	�parse_config_filesrC�warnings�warnrZfetch_build_eggs)rFrO�distrrr�_install_setup_requires{s�rTcKst|�tjjfi|��S�N)rTrIrJrrErrrr�src@s:eZdZejZdZdd�Zddd�Zdd�Zd
d
d�Z	dS)r
FcKst�||�t|��|�dS)zj
        Construct the command for dist, updating
        vars(self) with any keyword parameters.
        N)�_CommandrK�vars�update)rLrS�kwrrrrK�szCommand.__init__NcCsBt||�}|dur"t|||�|St|t�s>td|||f��|S)Nz'%s' must be a %s (got `%s`))�getattr�setattr�
isinstance�strr)rL�option�what�default�valrrr�_ensure_stringlike�s

�zCommand._ensure_stringlikecCspt||�}|durdSt|t�r6t||t�d|��n6t|t�rTtdd�|D��}nd}|sltd||f��dS)z�Ensure that 'option' is a list of strings.  If 'option' is
        currently a string, we split it either on /,\s*/ or /\s+/, so
        "foo bar baz", "foo,bar,baz", and "foo,   bar baz" all become
        ["foo", "bar", "baz"].
        Nz,\s*|\s+css|]}t|t�VqdSrU)r\r])r3�vrrrr6�r7z-Command.ensure_string_list.<locals>.<genexpr>Fz''%s' must be a list of strings (got %r))	rZr\r]r[�re�splitr�allr)rLr^ra�okrrr�ensure_string_list�s



�zCommand.ensure_string_listrcKs t�|||�}t|��|�|SrU)rV�reinitialize_commandrWrX)rL�command�reinit_subcommandsrY�cmdrrrri�szCommand.reinitialize_command)N)r)
r<r=r>rVr?Zcommand_consumes_argumentsrKrbrhrirrrrr
�s
r
cCs&dd�tj|dd�D�}ttjj|�S)z%
    Find all files under 'path'
    css,|]$\}}}|D]}tj�||�VqqdSrU)r r"r#)r3�baser*r+�filerrrr6�s�z#_find_all_simple.<locals>.<genexpr>Tr)r r!�filterr"r0)r"�resultsrrr�_find_all_simple�s�rqcCs6t|�}|tjkr.tjtjj|d�}t||�}t|�S)z�
    Find all files under 'dir' and return the list of full filenames.
    Unless dir is '.', return full filenames with dir prepended.
    )�start)	rqr �curdir�	functools�partialr"r$�mapr)r,r+Zmake_relrrr�findall�s


rwc@seZdZdZdS)�sicz;Treat this string as-is (https://en.wikipedia.org/wiki/Sic)N)r<r=r>r?rrrrrx�srx)0r?�fnmatchrrtr rdrQZ_distutils_hack.overrideZ_distutils_hack�distutils.corerI�distutils.errorsr�distutils.utilrZ_deprecation_warningrZsetuptools.version�
setuptoolsZsetuptools.extensionrZsetuptools.distr	Zsetuptools.dependsr
�r�__all__�version�__version__Zbootstrap_install_fromrrBrrrrTrrJZ
get_unpatchedr
rVrqrsrwr]rxZ	patch_allrrrr�<module>s>I!3site-packages/setuptools/__pycache__/launch.cpython-39.pyc000064400000001665150564745010017637 0ustar00a

(�Re,�@s.dZddlZddlZdd�Zedkr*e�dS)z[
Launch the Python script on the command line after
setuptools is bootstrapped via import.
�NcCs�ttjd}t|ddd�}tjdd�tjdd�<ttdt�}||��}|��}Wd�n1sf0Y|�dd�}t	||d�}t
||�dS)	zP
    Run the script in sys.argv[1] as if it had
    been invoked naturally.
    ��__main__N)�__file__�__name__�__doc__�openz\r\nz\n�exec)�__builtins__�sys�argv�dict�getattr�tokenizer�read�replace�compiler)�script_name�	namespace�open_Zfid�scriptZnorm_script�code�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/launch.py�run
s
�
&rr)rrr
rrrrrr�<module>s
site-packages/setuptools/__pycache__/_imp.cpython-39.pyc000064400000004125150564745010017303 0ustar00a

(�ReX	�@sddZddlZddlZddlZddlmZdZdZdZ	dZ
dZd	d
�Zddd�Z
dd
d�Zdd�ZdS)zX
Re-implementation of find_module and get_frozen_object
from the deprecated imp module.
�N�)�module_from_spec����cCs(t|t�rtj��jntjj}|||�S)N)�
isinstance�list�	importlib�	machinery�
PathFinder�	find_spec�util)�module�paths�finder�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_imp.pyr
s
��r
c	CsRt||�}|durtd|��|js>t|d�r>tj�d|j�}d}d}t|jt	�}|j
dksp|r�t|jtjj
�r�t}d}d}}n�|j
dks�|r�t|jtjj�r�t}d}d}}n�|j�r6|j
}tj�|�d	}|tjjvr�d
nd}|tjjvr�t}n&|tjjv�r
t}n|tjjv�rt}|tthv�rBt||�}nd}d}}|||||ffS)z7Just like 'imp.find_module()', but with package supportN�
Can't find %s�submodule_search_locationsz__init__.py����frozen�zbuilt-inr�r�rb)r
�ImportError�has_location�hasattrr
r�spec_from_loader�loaderr�type�origin�
issubclassr�FrozenImporter�	PY_FROZEN�BuiltinImporter�	C_BUILTIN�os�path�splitext�SOURCE_SUFFIXES�	PY_SOURCE�BYTECODE_SUFFIXES�PY_COMPILED�EXTENSION_SUFFIXES�C_EXTENSION�open)	rr�spec�kind�fileZstaticr(�suffix�moderrr�find_modulesF

�

�
r6cCs&t||�}|std|��|j�|�S�Nr)r
rr�get_code)rrr1rrr�get_frozen_objectGs
r9cCs"t||�}|std|��t|�Sr7)r
rr)rr�infor1rrr�
get_moduleNs
r;)N)N)�__doc__r'�importlib.utilr
�importlib.machineryZ
py34compatrr+r-r/r&r$r
r6r9r;rrrr�<module>s	
*
site-packages/setuptools/__pycache__/errors.cpython-39.pyc000064400000001602150564745010017670 0ustar00a

(�Re�@s&dZddlmZGdd�dee�ZdS)zCsetuptools.errors

Provides exceptions used by setuptools modules.
�)�DistutilsErrorc@seZdZdZdS)�RemovedCommandErroraOError used for commands that have been removed in setuptools.

    Since ``setuptools`` is built on ``distutils``, simply removing a command
    from ``setuptools`` will make the behavior fall back to ``distutils``; this
    error is raised if a command exists in ``distutils`` but has been actively
    removed in ``setuptools``.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/errors.pyr	srN)r�distutils.errorsr�RuntimeErrorrrrrr	�<module>ssite-packages/setuptools/build_meta.py000064400000024050150564745010014174 0ustar00"""A PEP 517 interface to setuptools

Previously, when a user or a command line tool (let's call it a "frontend")
needed to make a request of setuptools to take a certain action, for
example, generating a list of installation requirements, the frontend would
would call "setup.py egg_info" or "setup.py bdist_wheel" on the command line.

PEP 517 defines a different method of interfacing with setuptools. Rather
than calling "setup.py" directly, the frontend should:

  1. Set the current directory to the directory with a setup.py file
  2. Import this module into a safe python interpreter (one in which
     setuptools can potentially set global variables or crash hard).
  3. Call one of the functions defined in PEP 517.

What each function does is defined in PEP 517. However, here is a "casual"
definition of the functions (this definition should not be relied on for
bug reports or API stability):

  - `build_wheel`: build a wheel in the folder and return the basename
  - `get_requires_for_build_wheel`: get the `setup_requires` to build
  - `prepare_metadata_for_build_wheel`: get the `install_requires`
  - `build_sdist`: build an sdist in the folder and return the basename
  - `get_requires_for_build_sdist`: get the `setup_requires` to build

Again, this is not a formal definition! Just a "taste" of the module.
"""

import io
import os
import sys
import tokenize
import shutil
import contextlib
import tempfile

import setuptools
import distutils

from pkg_resources import parse_requirements

__all__ = ['get_requires_for_build_sdist',
           'get_requires_for_build_wheel',
           'prepare_metadata_for_build_wheel',
           'build_wheel',
           'build_sdist',
           '__legacy__',
           'SetupRequirementsError']


class SetupRequirementsError(BaseException):
    def __init__(self, specifiers):
        self.specifiers = specifiers


class Distribution(setuptools.dist.Distribution):
    def fetch_build_eggs(self, specifiers):
        specifier_list = list(map(str, parse_requirements(specifiers)))

        raise SetupRequirementsError(specifier_list)

    @classmethod
    @contextlib.contextmanager
    def patch(cls):
        """
        Replace
        distutils.dist.Distribution with this class
        for the duration of this context.
        """
        orig = distutils.core.Distribution
        distutils.core.Distribution = cls
        try:
            yield
        finally:
            distutils.core.Distribution = orig


@contextlib.contextmanager
def no_install_setup_requires():
    """Temporarily disable installing setup_requires

    Under PEP 517, the backend reports build dependencies to the frontend,
    and the frontend is responsible for ensuring they're installed.
    So setuptools (acting as a backend) should not try to install them.
    """
    orig = setuptools._install_setup_requires
    setuptools._install_setup_requires = lambda attrs: None
    try:
        yield
    finally:
        setuptools._install_setup_requires = orig


def _get_immediate_subdirectories(a_dir):
    return [name for name in os.listdir(a_dir)
            if os.path.isdir(os.path.join(a_dir, name))]


def _file_with_extension(directory, extension):
    matching = (
        f for f in os.listdir(directory)
        if f.endswith(extension)
    )
    try:
        file, = matching
    except ValueError:
        raise ValueError(
            'No distribution was found. Ensure that `setup.py` '
            'is not empty and that it calls `setup()`.')
    return file


def _open_setup_script(setup_script):
    if not os.path.exists(setup_script):
        # Supply a default setup.py
        return io.StringIO(u"from setuptools import setup; setup()")

    return getattr(tokenize, 'open', open)(setup_script)


class _BuildMetaBackend(object):

    def _fix_config(self, config_settings):
        config_settings = config_settings or {}
        config_settings.setdefault('--global-option', [])
        return config_settings

    def _get_build_requires(self, config_settings, requirements):
        config_settings = self._fix_config(config_settings)

        sys.argv = sys.argv[:1] + ['egg_info'] + \
            config_settings["--global-option"]
        try:
            with Distribution.patch():
                self.run_setup()
        except SetupRequirementsError as e:
            requirements += e.specifiers

        return requirements

    def run_setup(self, setup_script='setup.py'):
        # Note that we can reuse our build directory between calls
        # Correctness comes first, then optimization later
        __file__ = setup_script
        __name__ = '__main__'

        with _open_setup_script(__file__) as f:
            code = f.read().replace(r'\r\n', r'\n')

        exec(compile(code, __file__, 'exec'), locals())

    def get_requires_for_build_wheel(self, config_settings=None):
        config_settings = self._fix_config(config_settings)
        return self._get_build_requires(
            config_settings, requirements=['wheel'])

    def get_requires_for_build_sdist(self, config_settings=None):
        config_settings = self._fix_config(config_settings)
        return self._get_build_requires(config_settings, requirements=[])

    def prepare_metadata_for_build_wheel(self, metadata_directory,
                                         config_settings=None):
        sys.argv = sys.argv[:1] + [
            'dist_info', '--egg-base', metadata_directory]
        with no_install_setup_requires():
            self.run_setup()

        dist_info_directory = metadata_directory
        while True:
            dist_infos = [f for f in os.listdir(dist_info_directory)
                          if f.endswith('.dist-info')]

            if (
                len(dist_infos) == 0 and
                len(_get_immediate_subdirectories(dist_info_directory)) == 1
            ):

                dist_info_directory = os.path.join(
                    dist_info_directory, os.listdir(dist_info_directory)[0])
                continue

            assert len(dist_infos) == 1
            break

        # PEP 517 requires that the .dist-info directory be placed in the
        # metadata_directory. To comply, we MUST copy the directory to the root
        if dist_info_directory != metadata_directory:
            shutil.move(
                os.path.join(dist_info_directory, dist_infos[0]),
                metadata_directory)
            shutil.rmtree(dist_info_directory, ignore_errors=True)

        return dist_infos[0]

    def _build_with_temp_dir(self, setup_command, result_extension,
                             result_directory, config_settings):
        config_settings = self._fix_config(config_settings)
        result_directory = os.path.abspath(result_directory)

        # Build in a temporary directory, then copy to the target.
        os.makedirs(result_directory, exist_ok=True)
        with tempfile.TemporaryDirectory(dir=result_directory) as tmp_dist_dir:
            sys.argv = (sys.argv[:1] + setup_command +
                        ['--dist-dir', tmp_dist_dir] +
                        config_settings["--global-option"])
            with no_install_setup_requires():
                self.run_setup()

            result_basename = _file_with_extension(
                tmp_dist_dir, result_extension)
            result_path = os.path.join(result_directory, result_basename)
            if os.path.exists(result_path):
                # os.rename will fail overwriting on non-Unix.
                os.remove(result_path)
            os.rename(os.path.join(tmp_dist_dir, result_basename), result_path)

        return result_basename

    def build_wheel(self, wheel_directory, config_settings=None,
                    metadata_directory=None):
        return self._build_with_temp_dir(['bdist_wheel'], '.whl',
                                         wheel_directory, config_settings)

    def build_sdist(self, sdist_directory, config_settings=None):
        return self._build_with_temp_dir(['sdist', '--formats', 'gztar'],
                                         '.tar.gz', sdist_directory,
                                         config_settings)


class _BuildMetaLegacyBackend(_BuildMetaBackend):
    """Compatibility backend for setuptools

    This is a version of setuptools.build_meta that endeavors
    to maintain backwards
    compatibility with pre-PEP 517 modes of invocation. It
    exists as a temporary
    bridge between the old packaging mechanism and the new
    packaging mechanism,
    and will eventually be removed.
    """
    def run_setup(self, setup_script='setup.py'):
        # In order to maintain compatibility with scripts assuming that
        # the setup.py script is in a directory on the PYTHONPATH, inject
        # '' into sys.path. (pypa/setuptools#1642)
        sys_path = list(sys.path)           # Save the original path

        script_dir = os.path.dirname(os.path.abspath(setup_script))
        if script_dir not in sys.path:
            sys.path.insert(0, script_dir)

        # Some setup.py scripts (e.g. in pygame and numpy) use sys.argv[0] to
        # get the directory of the source code. They expect it to refer to the
        # setup.py script.
        sys_argv_0 = sys.argv[0]
        sys.argv[0] = setup_script

        try:
            super(_BuildMetaLegacyBackend,
                  self).run_setup(setup_script=setup_script)
        finally:
            # While PEP 517 frontends should be calling each hook in a fresh
            # subprocess according to the standard (and thus it should not be
            # strictly necessary to restore the old sys.path), we'll restore
            # the original path so that the path manipulation does not persist
            # within the hook after run_setup is called.
            sys.path[:] = sys_path
            sys.argv[0] = sys_argv_0


# The primary backend
_BACKEND = _BuildMetaBackend()

get_requires_for_build_wheel = _BACKEND.get_requires_for_build_wheel
get_requires_for_build_sdist = _BACKEND.get_requires_for_build_sdist
prepare_metadata_for_build_wheel = _BACKEND.prepare_metadata_for_build_wheel
build_wheel = _BACKEND.build_wheel
build_sdist = _BACKEND.build_sdist


# The legacy backend
__legacy__ = _BuildMetaLegacyBackend()
site-packages/setuptools/_distutils/text_file.py000064400000030303150564745010016233 0ustar00"""text_file

provides the TextFile class, which gives an interface to text files
that (optionally) takes care of stripping comments, ignoring blank
lines, and joining lines with backslashes."""

import sys, io


class TextFile:
    """Provides a file-like object that takes care of all the things you
       commonly want to do when processing a text file that has some
       line-by-line syntax: strip comments (as long as "#" is your
       comment character), skip blank lines, join adjacent lines by
       escaping the newline (ie. backslash at end of line), strip
       leading and/or trailing whitespace.  All of these are optional
       and independently controllable.

       Provides a 'warn()' method so you can generate warning messages that
       report physical line number, even if the logical line in question
       spans multiple physical lines.  Also provides 'unreadline()' for
       implementing line-at-a-time lookahead.

       Constructor is called as:

           TextFile (filename=None, file=None, **options)

       It bombs (RuntimeError) if both 'filename' and 'file' are None;
       'filename' should be a string, and 'file' a file object (or
       something that provides 'readline()' and 'close()' methods).  It is
       recommended that you supply at least 'filename', so that TextFile
       can include it in warning messages.  If 'file' is not supplied,
       TextFile creates its own using 'io.open()'.

       The options are all boolean, and affect the value returned by
       'readline()':
         strip_comments [default: true]
           strip from "#" to end-of-line, as well as any whitespace
           leading up to the "#" -- unless it is escaped by a backslash
         lstrip_ws [default: false]
           strip leading whitespace from each line before returning it
         rstrip_ws [default: true]
           strip trailing whitespace (including line terminator!) from
           each line before returning it
         skip_blanks [default: true}
           skip lines that are empty *after* stripping comments and
           whitespace.  (If both lstrip_ws and rstrip_ws are false,
           then some lines may consist of solely whitespace: these will
           *not* be skipped, even if 'skip_blanks' is true.)
         join_lines [default: false]
           if a backslash is the last non-newline character on a line
           after stripping comments and whitespace, join the following line
           to it to form one "logical line"; if N consecutive lines end
           with a backslash, then N+1 physical lines will be joined to
           form one logical line.
         collapse_join [default: false]
           strip leading whitespace from lines that are joined to their
           predecessor; only matters if (join_lines and not lstrip_ws)
         errors [default: 'strict']
           error handler used to decode the file content

       Note that since 'rstrip_ws' can strip the trailing newline, the
       semantics of 'readline()' must differ from those of the builtin file
       object's 'readline()' method!  In particular, 'readline()' returns
       None for end-of-file: an empty string might just be a blank line (or
       an all-whitespace line), if 'rstrip_ws' is true but 'skip_blanks' is
       not."""

    default_options = { 'strip_comments': 1,
                        'skip_blanks':    1,
                        'lstrip_ws':      0,
                        'rstrip_ws':      1,
                        'join_lines':     0,
                        'collapse_join':  0,
                        'errors':         'strict',
                      }

    def __init__(self, filename=None, file=None, **options):
        """Construct a new TextFile object.  At least one of 'filename'
           (a string) and 'file' (a file-like object) must be supplied.
           They keyword argument options are described above and affect
           the values returned by 'readline()'."""
        if filename is None and file is None:
            raise RuntimeError("you must supply either or both of 'filename' and 'file'")

        # set values for all options -- either from client option hash
        # or fallback to default_options
        for opt in self.default_options.keys():
            if opt in options:
                setattr(self, opt, options[opt])
            else:
                setattr(self, opt, self.default_options[opt])

        # sanity check client option hash
        for opt in options.keys():
            if opt not in self.default_options:
                raise KeyError("invalid TextFile option '%s'" % opt)

        if file is None:
            self.open(filename)
        else:
            self.filename = filename
            self.file = file
            self.current_line = 0       # assuming that file is at BOF!

        # 'linebuf' is a stack of lines that will be emptied before we
        # actually read from the file; it's only populated by an
        # 'unreadline()' operation
        self.linebuf = []

    def open(self, filename):
        """Open a new file named 'filename'.  This overrides both the
           'filename' and 'file' arguments to the constructor."""
        self.filename = filename
        self.file = io.open(self.filename, 'r', errors=self.errors)
        self.current_line = 0

    def close(self):
        """Close the current file and forget everything we know about it
           (filename, current line number)."""
        file = self.file
        self.file = None
        self.filename = None
        self.current_line = None
        file.close()

    def gen_error(self, msg, line=None):
        outmsg = []
        if line is None:
            line = self.current_line
        outmsg.append(self.filename + ", ")
        if isinstance(line, (list, tuple)):
            outmsg.append("lines %d-%d: " % tuple(line))
        else:
            outmsg.append("line %d: " % line)
        outmsg.append(str(msg))
        return "".join(outmsg)

    def error(self, msg, line=None):
        raise ValueError("error: " + self.gen_error(msg, line))

    def warn(self, msg, line=None):
        """Print (to stderr) a warning message tied to the current logical
           line in the current file.  If the current logical line in the
           file spans multiple physical lines, the warning refers to the
           whole range, eg. "lines 3-5".  If 'line' supplied, it overrides
           the current line number; it may be a list or tuple to indicate a
           range of physical lines, or an integer for a single physical
           line."""
        sys.stderr.write("warning: " + self.gen_error(msg, line) + "\n")

    def readline(self):
        """Read and return a single logical line from the current file (or
           from an internal buffer if lines have previously been "unread"
           with 'unreadline()').  If the 'join_lines' option is true, this
           may involve reading multiple physical lines concatenated into a
           single string.  Updates the current line number, so calling
           'warn()' after 'readline()' emits a warning about the physical
           line(s) just read.  Returns None on end-of-file, since the empty
           string can occur if 'rstrip_ws' is true but 'strip_blanks' is
           not."""
        # If any "unread" lines waiting in 'linebuf', return the top
        # one.  (We don't actually buffer read-ahead data -- lines only
        # get put in 'linebuf' if the client explicitly does an
        # 'unreadline()'.
        if self.linebuf:
            line = self.linebuf[-1]
            del self.linebuf[-1]
            return line

        buildup_line = ''

        while True:
            # read the line, make it None if EOF
            line = self.file.readline()
            if line == '':
                line = None

            if self.strip_comments and line:

                # Look for the first "#" in the line.  If none, never
                # mind.  If we find one and it's the first character, or
                # is not preceded by "\", then it starts a comment --
                # strip the comment, strip whitespace before it, and
                # carry on.  Otherwise, it's just an escaped "#", so
                # unescape it (and any other escaped "#"'s that might be
                # lurking in there) and otherwise leave the line alone.

                pos = line.find("#")
                if pos == -1: # no "#" -- no comments
                    pass

                # It's definitely a comment -- either "#" is the first
                # character, or it's elsewhere and unescaped.
                elif pos == 0 or line[pos-1] != "\\":
                    # Have to preserve the trailing newline, because it's
                    # the job of a later step (rstrip_ws) to remove it --
                    # and if rstrip_ws is false, we'd better preserve it!
                    # (NB. this means that if the final line is all comment
                    # and has no trailing newline, we will think that it's
                    # EOF; I think that's OK.)
                    eol = (line[-1] == '\n') and '\n' or ''
                    line = line[0:pos] + eol

                    # If all that's left is whitespace, then skip line
                    # *now*, before we try to join it to 'buildup_line' --
                    # that way constructs like
                    #   hello \\
                    #   # comment that should be ignored
                    #   there
                    # result in "hello there".
                    if line.strip() == "":
                        continue
                else: # it's an escaped "#"
                    line = line.replace("\\#", "#")

            # did previous line end with a backslash? then accumulate
            if self.join_lines and buildup_line:
                # oops: end of file
                if line is None:
                    self.warn("continuation line immediately precedes "
                              "end-of-file")
                    return buildup_line

                if self.collapse_join:
                    line = line.lstrip()
                line = buildup_line + line

                # careful: pay attention to line number when incrementing it
                if isinstance(self.current_line, list):
                    self.current_line[1] = self.current_line[1] + 1
                else:
                    self.current_line = [self.current_line,
                                         self.current_line + 1]
            # just an ordinary line, read it as usual
            else:
                if line is None: # eof
                    return None

                # still have to be careful about incrementing the line number!
                if isinstance(self.current_line, list):
                    self.current_line = self.current_line[1] + 1
                else:
                    self.current_line = self.current_line + 1

            # strip whitespace however the client wants (leading and
            # trailing, or one or the other, or neither)
            if self.lstrip_ws and self.rstrip_ws:
                line = line.strip()
            elif self.lstrip_ws:
                line = line.lstrip()
            elif self.rstrip_ws:
                line = line.rstrip()

            # blank line (whether we rstrip'ed or not)? skip to next line
            # if appropriate
            if (line == '' or line == '\n') and self.skip_blanks:
                continue

            if self.join_lines:
                if line[-1] == '\\':
                    buildup_line = line[:-1]
                    continue

                if line[-2:] == '\\\n':
                    buildup_line = line[0:-2] + '\n'
                    continue

            # well, I guess there's some actual content there: return it
            return line

    def readlines(self):
        """Read and return the list of all logical lines remaining in the
           current file."""
        lines = []
        while True:
            line = self.readline()
            if line is None:
                return lines
            lines.append(line)

    def unreadline(self, line):
        """Push 'line' (a string) onto an internal buffer that will be
           checked by future 'readline()' calls.  Handy for implementing
           a parser with line-at-a-time lookahead."""
        self.linebuf.append(line)
site-packages/setuptools/_distutils/archive_util.py000064400000020574150564745010016737 0ustar00"""distutils.archive_util

Utility functions for creating archive files (tarballs, zip files,
that sort of thing)."""

import os
from warnings import warn
import sys

try:
    import zipfile
except ImportError:
    zipfile = None


from distutils.errors import DistutilsExecError
from distutils.spawn import spawn
from distutils.dir_util import mkpath
from distutils import log

try:
    from pwd import getpwnam
except ImportError:
    getpwnam = None

try:
    from grp import getgrnam
except ImportError:
    getgrnam = None

def _get_gid(name):
    """Returns a gid, given a group name."""
    if getgrnam is None or name is None:
        return None
    try:
        result = getgrnam(name)
    except KeyError:
        result = None
    if result is not None:
        return result[2]
    return None

def _get_uid(name):
    """Returns an uid, given a user name."""
    if getpwnam is None or name is None:
        return None
    try:
        result = getpwnam(name)
    except KeyError:
        result = None
    if result is not None:
        return result[2]
    return None

def make_tarball(base_name, base_dir, compress="gzip", verbose=0, dry_run=0,
                 owner=None, group=None):
    """Create a (possibly compressed) tar file from all the files under
    'base_dir'.

    'compress' must be "gzip" (the default), "bzip2", "xz", "compress", or
    None.  ("compress" will be deprecated in Python 3.2)

    'owner' and 'group' can be used to define an owner and a group for the
    archive that is being built. If not provided, the current owner and group
    will be used.

    The output tar file will be named 'base_dir' +  ".tar", possibly plus
    the appropriate compression extension (".gz", ".bz2", ".xz" or ".Z").

    Returns the output filename.
    """
    tar_compression = {'gzip': 'gz', 'bzip2': 'bz2', 'xz': 'xz', None: '',
                       'compress': ''}
    compress_ext = {'gzip': '.gz', 'bzip2': '.bz2', 'xz': '.xz',
                    'compress': '.Z'}

    # flags for compression program, each element of list will be an argument
    if compress is not None and compress not in compress_ext.keys():
        raise ValueError(
              "bad value for 'compress': must be None, 'gzip', 'bzip2', "
              "'xz' or 'compress'")

    archive_name = base_name + '.tar'
    if compress != 'compress':
        archive_name += compress_ext.get(compress, '')

    mkpath(os.path.dirname(archive_name), dry_run=dry_run)

    # creating the tarball
    import tarfile  # late import so Python build itself doesn't break

    log.info('Creating tar archive')

    uid = _get_uid(owner)
    gid = _get_gid(group)

    def _set_uid_gid(tarinfo):
        if gid is not None:
            tarinfo.gid = gid
            tarinfo.gname = group
        if uid is not None:
            tarinfo.uid = uid
            tarinfo.uname = owner
        return tarinfo

    if not dry_run:
        tar = tarfile.open(archive_name, 'w|%s' % tar_compression[compress])
        try:
            tar.add(base_dir, filter=_set_uid_gid)
        finally:
            tar.close()

    # compression using `compress`
    if compress == 'compress':
        warn("'compress' will be deprecated.", PendingDeprecationWarning)
        # the option varies depending on the platform
        compressed_name = archive_name + compress_ext[compress]
        if sys.platform == 'win32':
            cmd = [compress, archive_name, compressed_name]
        else:
            cmd = [compress, '-f', archive_name]
        spawn(cmd, dry_run=dry_run)
        return compressed_name

    return archive_name

def make_zipfile(base_name, base_dir, verbose=0, dry_run=0):
    """Create a zip file from all the files under 'base_dir'.

    The output zip file will be named 'base_name' + ".zip".  Uses either the
    "zipfile" Python module (if available) or the InfoZIP "zip" utility
    (if installed and found on the default search path).  If neither tool is
    available, raises DistutilsExecError.  Returns the name of the output zip
    file.
    """
    zip_filename = base_name + ".zip"
    mkpath(os.path.dirname(zip_filename), dry_run=dry_run)

    # If zipfile module is not available, try spawning an external
    # 'zip' command.
    if zipfile is None:
        if verbose:
            zipoptions = "-r"
        else:
            zipoptions = "-rq"

        try:
            spawn(["zip", zipoptions, zip_filename, base_dir],
                  dry_run=dry_run)
        except DistutilsExecError:
            # XXX really should distinguish between "couldn't find
            # external 'zip' command" and "zip failed".
            raise DistutilsExecError(("unable to create zip file '%s': "
                   "could neither import the 'zipfile' module nor "
                   "find a standalone zip utility") % zip_filename)

    else:
        log.info("creating '%s' and adding '%s' to it",
                 zip_filename, base_dir)

        if not dry_run:
            try:
                zip = zipfile.ZipFile(zip_filename, "w",
                                      compression=zipfile.ZIP_DEFLATED)
            except RuntimeError:
                zip = zipfile.ZipFile(zip_filename, "w",
                                      compression=zipfile.ZIP_STORED)

            with zip:
                if base_dir != os.curdir:
                    path = os.path.normpath(os.path.join(base_dir, ''))
                    zip.write(path, path)
                    log.info("adding '%s'", path)
                for dirpath, dirnames, filenames in os.walk(base_dir):
                    for name in dirnames:
                        path = os.path.normpath(os.path.join(dirpath, name, ''))
                        zip.write(path, path)
                        log.info("adding '%s'", path)
                    for name in filenames:
                        path = os.path.normpath(os.path.join(dirpath, name))
                        if os.path.isfile(path):
                            zip.write(path, path)
                            log.info("adding '%s'", path)

    return zip_filename

ARCHIVE_FORMATS = {
    'gztar': (make_tarball, [('compress', 'gzip')], "gzip'ed tar-file"),
    'bztar': (make_tarball, [('compress', 'bzip2')], "bzip2'ed tar-file"),
    'xztar': (make_tarball, [('compress', 'xz')], "xz'ed tar-file"),
    'ztar':  (make_tarball, [('compress', 'compress')], "compressed tar file"),
    'tar':   (make_tarball, [('compress', None)], "uncompressed tar file"),
    'zip':   (make_zipfile, [],"ZIP file")
    }

def check_archive_formats(formats):
    """Returns the first format from the 'format' list that is unknown.

    If all formats are known, returns None
    """
    for format in formats:
        if format not in ARCHIVE_FORMATS:
            return format
    return None

def make_archive(base_name, format, root_dir=None, base_dir=None, verbose=0,
                 dry_run=0, owner=None, group=None):
    """Create an archive file (eg. zip or tar).

    'base_name' is the name of the file to create, minus any format-specific
    extension; 'format' is the archive format: one of "zip", "tar", "gztar",
    "bztar", "xztar", or "ztar".

    'root_dir' is a directory that will be the root directory of the
    archive; ie. we typically chdir into 'root_dir' before creating the
    archive.  'base_dir' is the directory where we start archiving from;
    ie. 'base_dir' will be the common prefix of all files and
    directories in the archive.  'root_dir' and 'base_dir' both default
    to the current directory.  Returns the name of the archive file.

    'owner' and 'group' are used when creating a tar archive. By default,
    uses the current owner and group.
    """
    save_cwd = os.getcwd()
    if root_dir is not None:
        log.debug("changing into '%s'", root_dir)
        base_name = os.path.abspath(base_name)
        if not dry_run:
            os.chdir(root_dir)

    if base_dir is None:
        base_dir = os.curdir

    kwargs = {'dry_run': dry_run}

    try:
        format_info = ARCHIVE_FORMATS[format]
    except KeyError:
        raise ValueError("unknown archive format '%s'" % format)

    func = format_info[0]
    for arg, val in format_info[1]:
        kwargs[arg] = val

    if format != 'zip':
        kwargs['owner'] = owner
        kwargs['group'] = group

    try:
        filename = func(base_name, base_dir, **kwargs)
    finally:
        if root_dir is not None:
            log.debug("changing back to '%s'", save_cwd)
            os.chdir(save_cwd)

    return filename
site-packages/setuptools/_distutils/msvccompiler.py000064400000055764150564745010016775 0ustar00"""distutils.msvccompiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for the Microsoft Visual Studio.
"""

# Written by Perry Stoll
# hacked by Robin Becker and Thomas Heller to do a better job of
#   finding DevStudio (through the registry)

import sys, os
from distutils.errors import \
     DistutilsExecError, DistutilsPlatformError, \
     CompileError, LibError, LinkError
from distutils.ccompiler import \
     CCompiler, gen_lib_options
from distutils import log

_can_read_reg = False
try:
    import winreg

    _can_read_reg = True
    hkey_mod = winreg

    RegOpenKeyEx = winreg.OpenKeyEx
    RegEnumKey = winreg.EnumKey
    RegEnumValue = winreg.EnumValue
    RegError = winreg.error

except ImportError:
    try:
        import win32api
        import win32con
        _can_read_reg = True
        hkey_mod = win32con

        RegOpenKeyEx = win32api.RegOpenKeyEx
        RegEnumKey = win32api.RegEnumKey
        RegEnumValue = win32api.RegEnumValue
        RegError = win32api.error
    except ImportError:
        log.info("Warning: Can't read registry to find the "
                 "necessary compiler setting\n"
                 "Make sure that Python modules winreg, "
                 "win32api or win32con are installed.")
        pass

if _can_read_reg:
    HKEYS = (hkey_mod.HKEY_USERS,
             hkey_mod.HKEY_CURRENT_USER,
             hkey_mod.HKEY_LOCAL_MACHINE,
             hkey_mod.HKEY_CLASSES_ROOT)

def read_keys(base, key):
    """Return list of registry keys."""
    try:
        handle = RegOpenKeyEx(base, key)
    except RegError:
        return None
    L = []
    i = 0
    while True:
        try:
            k = RegEnumKey(handle, i)
        except RegError:
            break
        L.append(k)
        i += 1
    return L

def read_values(base, key):
    """Return dict of registry keys and values.

    All names are converted to lowercase.
    """
    try:
        handle = RegOpenKeyEx(base, key)
    except RegError:
        return None
    d = {}
    i = 0
    while True:
        try:
            name, value, type = RegEnumValue(handle, i)
        except RegError:
            break
        name = name.lower()
        d[convert_mbcs(name)] = convert_mbcs(value)
        i += 1
    return d

def convert_mbcs(s):
    dec = getattr(s, "decode", None)
    if dec is not None:
        try:
            s = dec("mbcs")
        except UnicodeError:
            pass
    return s

class MacroExpander:
    def __init__(self, version):
        self.macros = {}
        self.load_macros(version)

    def set_macro(self, macro, path, key):
        for base in HKEYS:
            d = read_values(base, path)
            if d:
                self.macros["$(%s)" % macro] = d[key]
                break

    def load_macros(self, version):
        vsbase = r"Software\Microsoft\VisualStudio\%0.1f" % version
        self.set_macro("VCInstallDir", vsbase + r"\Setup\VC", "productdir")
        self.set_macro("VSInstallDir", vsbase + r"\Setup\VS", "productdir")
        net = r"Software\Microsoft\.NETFramework"
        self.set_macro("FrameworkDir", net, "installroot")
        try:
            if version > 7.0:
                self.set_macro("FrameworkSDKDir", net, "sdkinstallrootv1.1")
            else:
                self.set_macro("FrameworkSDKDir", net, "sdkinstallroot")
        except KeyError as exc: #
            raise DistutilsPlatformError(
            """Python was built with Visual Studio 2003;
extensions must be built with a compiler than can generate compatible binaries.
Visual Studio 2003 was not found on this system. If you have Cygwin installed,
you can try compiling with MingW32, by passing "-c mingw32" to setup.py.""")

        p = r"Software\Microsoft\NET Framework Setup\Product"
        for base in HKEYS:
            try:
                h = RegOpenKeyEx(base, p)
            except RegError:
                continue
            key = RegEnumKey(h, 0)
            d = read_values(base, r"%s\%s" % (p, key))
            self.macros["$(FrameworkVersion)"] = d["version"]

    def sub(self, s):
        for k, v in self.macros.items():
            s = s.replace(k, v)
        return s

def get_build_version():
    """Return the version of MSVC that was used to build Python.

    For Python 2.3 and up, the version number is included in
    sys.version.  For earlier versions, assume the compiler is MSVC 6.
    """
    prefix = "MSC v."
    i = sys.version.find(prefix)
    if i == -1:
        return 6
    i = i + len(prefix)
    s, rest = sys.version[i:].split(" ", 1)
    majorVersion = int(s[:-2]) - 6
    if majorVersion >= 13:
        # v13 was skipped and should be v14
        majorVersion += 1
    minorVersion = int(s[2:3]) / 10.0
    # I don't think paths are affected by minor version in version 6
    if majorVersion == 6:
        minorVersion = 0
    if majorVersion >= 6:
        return majorVersion + minorVersion
    # else we don't know what version of the compiler this is
    return None

def get_build_architecture():
    """Return the processor architecture.

    Possible results are "Intel" or "AMD64".
    """

    prefix = " bit ("
    i = sys.version.find(prefix)
    if i == -1:
        return "Intel"
    j = sys.version.find(")", i)
    return sys.version[i+len(prefix):j]

def normalize_and_reduce_paths(paths):
    """Return a list of normalized paths with duplicates removed.

    The current order of paths is maintained.
    """
    # Paths are normalized so things like:  /a and /a/ aren't both preserved.
    reduced_paths = []
    for p in paths:
        np = os.path.normpath(p)
        # XXX(nnorwitz): O(n**2), if reduced_paths gets long perhaps use a set.
        if np not in reduced_paths:
            reduced_paths.append(np)
    return reduced_paths


class MSVCCompiler(CCompiler) :
    """Concrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class."""

    compiler_type = 'msvc'

    # Just set this so CCompiler's constructor doesn't barf.  We currently
    # don't use the 'set_executables()' bureaucracy provided by CCompiler,
    # as it really isn't necessary for this sort of single-compiler class.
    # Would be nice to have a consistent interface with UnixCCompiler,
    # though, so it's worth thinking about.
    executables = {}

    # Private class data (need to distinguish C from C++ source for compiler)
    _c_extensions = ['.c']
    _cpp_extensions = ['.cc', '.cpp', '.cxx']
    _rc_extensions = ['.rc']
    _mc_extensions = ['.mc']

    # Needed for the filename generation methods provided by the
    # base class, CCompiler.
    src_extensions = (_c_extensions + _cpp_extensions +
                      _rc_extensions + _mc_extensions)
    res_extension = '.res'
    obj_extension = '.obj'
    static_lib_extension = '.lib'
    shared_lib_extension = '.dll'
    static_lib_format = shared_lib_format = '%s%s'
    exe_extension = '.exe'

    def __init__(self, verbose=0, dry_run=0, force=0):
        CCompiler.__init__ (self, verbose, dry_run, force)
        self.__version = get_build_version()
        self.__arch = get_build_architecture()
        if self.__arch == "Intel":
            # x86
            if self.__version >= 7:
                self.__root = r"Software\Microsoft\VisualStudio"
                self.__macros = MacroExpander(self.__version)
            else:
                self.__root = r"Software\Microsoft\Devstudio"
            self.__product = "Visual Studio version %s" % self.__version
        else:
            # Win64. Assume this was built with the platform SDK
            self.__product = "Microsoft SDK compiler %s" % (self.__version + 6)

        self.initialized = False

    def initialize(self):
        self.__paths = []
        if "DISTUTILS_USE_SDK" in os.environ and "MSSdk" in os.environ and self.find_exe("cl.exe"):
            # Assume that the SDK set up everything alright; don't try to be
            # smarter
            self.cc = "cl.exe"
            self.linker = "link.exe"
            self.lib = "lib.exe"
            self.rc = "rc.exe"
            self.mc = "mc.exe"
        else:
            self.__paths = self.get_msvc_paths("path")

            if len(self.__paths) == 0:
                raise DistutilsPlatformError("Python was built with %s, "
                       "and extensions need to be built with the same "
                       "version of the compiler, but it isn't installed."
                       % self.__product)

            self.cc = self.find_exe("cl.exe")
            self.linker = self.find_exe("link.exe")
            self.lib = self.find_exe("lib.exe")
            self.rc = self.find_exe("rc.exe")   # resource compiler
            self.mc = self.find_exe("mc.exe")   # message compiler
            self.set_path_env_var('lib')
            self.set_path_env_var('include')

        # extend the MSVC path with the current path
        try:
            for p in os.environ['path'].split(';'):
                self.__paths.append(p)
        except KeyError:
            pass
        self.__paths = normalize_and_reduce_paths(self.__paths)
        os.environ['path'] = ";".join(self.__paths)

        self.preprocess_options = None
        if self.__arch == "Intel":
            self.compile_options = [ '/nologo', '/O2', '/MD', '/W3', '/GX' ,
                                     '/DNDEBUG']
            self.compile_options_debug = ['/nologo', '/Od', '/MDd', '/W3', '/GX',
                                          '/Z7', '/D_DEBUG']
        else:
            # Win64
            self.compile_options = [ '/nologo', '/O2', '/MD', '/W3', '/GS-' ,
                                     '/DNDEBUG']
            self.compile_options_debug = ['/nologo', '/Od', '/MDd', '/W3', '/GS-',
                                          '/Z7', '/D_DEBUG']

        self.ldflags_shared = ['/DLL', '/nologo', '/INCREMENTAL:NO']
        if self.__version >= 7:
            self.ldflags_shared_debug = [
                '/DLL', '/nologo', '/INCREMENTAL:no', '/DEBUG'
                ]
        else:
            self.ldflags_shared_debug = [
                '/DLL', '/nologo', '/INCREMENTAL:no', '/pdb:None', '/DEBUG'
                ]
        self.ldflags_static = [ '/nologo']

        self.initialized = True

    # -- Worker methods ------------------------------------------------

    def object_filenames(self,
                         source_filenames,
                         strip_dir=0,
                         output_dir=''):
        # Copied from ccompiler.py, extended to return .res as 'object'-file
        # for .rc input file
        if output_dir is None: output_dir = ''
        obj_names = []
        for src_name in source_filenames:
            (base, ext) = os.path.splitext (src_name)
            base = os.path.splitdrive(base)[1] # Chop off the drive
            base = base[os.path.isabs(base):]  # If abs, chop off leading /
            if ext not in self.src_extensions:
                # Better to raise an exception instead of silently continuing
                # and later complain about sources and targets having
                # different lengths
                raise CompileError ("Don't know how to compile %s" % src_name)
            if strip_dir:
                base = os.path.basename (base)
            if ext in self._rc_extensions:
                obj_names.append (os.path.join (output_dir,
                                                base + self.res_extension))
            elif ext in self._mc_extensions:
                obj_names.append (os.path.join (output_dir,
                                                base + self.res_extension))
            else:
                obj_names.append (os.path.join (output_dir,
                                                base + self.obj_extension))
        return obj_names


    def compile(self, sources,
                output_dir=None, macros=None, include_dirs=None, debug=0,
                extra_preargs=None, extra_postargs=None, depends=None):

        if not self.initialized:
            self.initialize()
        compile_info = self._setup_compile(output_dir, macros, include_dirs,
                                           sources, depends, extra_postargs)
        macros, objects, extra_postargs, pp_opts, build = compile_info

        compile_opts = extra_preargs or []
        compile_opts.append ('/c')
        if debug:
            compile_opts.extend(self.compile_options_debug)
        else:
            compile_opts.extend(self.compile_options)

        for obj in objects:
            try:
                src, ext = build[obj]
            except KeyError:
                continue
            if debug:
                # pass the full pathname to MSVC in debug mode,
                # this allows the debugger to find the source file
                # without asking the user to browse for it
                src = os.path.abspath(src)

            if ext in self._c_extensions:
                input_opt = "/Tc" + src
            elif ext in self._cpp_extensions:
                input_opt = "/Tp" + src
            elif ext in self._rc_extensions:
                # compile .RC to .RES file
                input_opt = src
                output_opt = "/fo" + obj
                try:
                    self.spawn([self.rc] + pp_opts +
                               [output_opt] + [input_opt])
                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            elif ext in self._mc_extensions:
                # Compile .MC to .RC file to .RES file.
                #   * '-h dir' specifies the directory for the
                #     generated include file
                #   * '-r dir' specifies the target directory of the
                #     generated RC file and the binary message resource
                #     it includes
                #
                # For now (since there are no options to change this),
                # we use the source-directory for the include file and
                # the build directory for the RC file and message
                # resources. This works at least for win32all.
                h_dir = os.path.dirname(src)
                rc_dir = os.path.dirname(obj)
                try:
                    # first compile .MC to .RC and .H file
                    self.spawn([self.mc] +
                               ['-h', h_dir, '-r', rc_dir] + [src])
                    base, _ = os.path.splitext (os.path.basename (src))
                    rc_file = os.path.join (rc_dir, base + '.rc')
                    # then compile .RC to .RES file
                    self.spawn([self.rc] +
                               ["/fo" + obj] + [rc_file])

                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            else:
                # how to handle this file?
                raise CompileError("Don't know how to compile %s to %s"
                                   % (src, obj))

            output_opt = "/Fo" + obj
            try:
                self.spawn([self.cc] + compile_opts + pp_opts +
                           [input_opt, output_opt] +
                           extra_postargs)
            except DistutilsExecError as msg:
                raise CompileError(msg)

        return objects


    def create_static_lib(self,
                          objects,
                          output_libname,
                          output_dir=None,
                          debug=0,
                          target_lang=None):

        if not self.initialized:
            self.initialize()
        (objects, output_dir) = self._fix_object_args(objects, output_dir)
        output_filename = self.library_filename(output_libname,
                                                output_dir=output_dir)

        if self._need_link(objects, output_filename):
            lib_args = objects + ['/OUT:' + output_filename]
            if debug:
                pass # XXX what goes here?
            try:
                self.spawn([self.lib] + lib_args)
            except DistutilsExecError as msg:
                raise LibError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)


    def link(self,
             target_desc,
             objects,
             output_filename,
             output_dir=None,
             libraries=None,
             library_dirs=None,
             runtime_library_dirs=None,
             export_symbols=None,
             debug=0,
             extra_preargs=None,
             extra_postargs=None,
             build_temp=None,
             target_lang=None):

        if not self.initialized:
            self.initialize()
        (objects, output_dir) = self._fix_object_args(objects, output_dir)
        fixed_args = self._fix_lib_args(libraries, library_dirs,
                                        runtime_library_dirs)
        (libraries, library_dirs, runtime_library_dirs) = fixed_args

        if runtime_library_dirs:
            self.warn ("I don't know what to do with 'runtime_library_dirs': "
                       + str (runtime_library_dirs))

        lib_opts = gen_lib_options(self,
                                   library_dirs, runtime_library_dirs,
                                   libraries)
        if output_dir is not None:
            output_filename = os.path.join(output_dir, output_filename)

        if self._need_link(objects, output_filename):
            if target_desc == CCompiler.EXECUTABLE:
                if debug:
                    ldflags = self.ldflags_shared_debug[1:]
                else:
                    ldflags = self.ldflags_shared[1:]
            else:
                if debug:
                    ldflags = self.ldflags_shared_debug
                else:
                    ldflags = self.ldflags_shared

            export_opts = []
            for sym in (export_symbols or []):
                export_opts.append("/EXPORT:" + sym)

            ld_args = (ldflags + lib_opts + export_opts +
                       objects + ['/OUT:' + output_filename])

            # The MSVC linker generates .lib and .exp files, which cannot be
            # suppressed by any linker switches. The .lib files may even be
            # needed! Make sure they are generated in the temporary build
            # directory. Since they have different names for debug and release
            # builds, they can go into the same directory.
            if export_symbols is not None:
                (dll_name, dll_ext) = os.path.splitext(
                    os.path.basename(output_filename))
                implib_file = os.path.join(
                    os.path.dirname(objects[0]),
                    self.library_filename(dll_name))
                ld_args.append ('/IMPLIB:' + implib_file)

            if extra_preargs:
                ld_args[:0] = extra_preargs
            if extra_postargs:
                ld_args.extend(extra_postargs)

            self.mkpath(os.path.dirname(output_filename))
            try:
                self.spawn([self.linker] + ld_args)
            except DistutilsExecError as msg:
                raise LinkError(msg)

        else:
            log.debug("skipping %s (up-to-date)", output_filename)


    # -- Miscellaneous methods -----------------------------------------
    # These are all used by the 'gen_lib_options() function, in
    # ccompiler.py.

    def library_dir_option(self, dir):
        return "/LIBPATH:" + dir

    def runtime_library_dir_option(self, dir):
        raise DistutilsPlatformError(
              "don't know how to set runtime library search path for MSVC++")

    def library_option(self, lib):
        return self.library_filename(lib)


    def find_library_file(self, dirs, lib, debug=0):
        # Prefer a debugging library if found (and requested), but deal
        # with it if we don't have one.
        if debug:
            try_names = [lib + "_d", lib]
        else:
            try_names = [lib]
        for dir in dirs:
            for name in try_names:
                libfile = os.path.join(dir, self.library_filename (name))
                if os.path.exists(libfile):
                    return libfile
        else:
            # Oops, didn't find it in *any* of 'dirs'
            return None

    # Helper methods for using the MSVC registry settings

    def find_exe(self, exe):
        """Return path to an MSVC executable program.

        Tries to find the program in several places: first, one of the
        MSVC program search paths from the registry; next, the directories
        in the PATH environment variable.  If any of those work, return an
        absolute path that is known to exist.  If none of them work, just
        return the original program name, 'exe'.
        """
        for p in self.__paths:
            fn = os.path.join(os.path.abspath(p), exe)
            if os.path.isfile(fn):
                return fn

        # didn't find it; try existing path
        for p in os.environ['Path'].split(';'):
            fn = os.path.join(os.path.abspath(p),exe)
            if os.path.isfile(fn):
                return fn

        return exe

    def get_msvc_paths(self, path, platform='x86'):
        """Get a list of devstudio directories (include, lib or path).

        Return a list of strings.  The list will be empty if unable to
        access the registry or appropriate registry keys not found.
        """
        if not _can_read_reg:
            return []

        path = path + " dirs"
        if self.__version >= 7:
            key = (r"%s\%0.1f\VC\VC_OBJECTS_PLATFORM_INFO\Win32\Directories"
                   % (self.__root, self.__version))
        else:
            key = (r"%s\6.0\Build System\Components\Platforms"
                   r"\Win32 (%s)\Directories" % (self.__root, platform))

        for base in HKEYS:
            d = read_values(base, key)
            if d:
                if self.__version >= 7:
                    return self.__macros.sub(d[path]).split(";")
                else:
                    return d[path].split(";")
        # MSVC 6 seems to create the registry entries we need only when
        # the GUI is run.
        if self.__version == 6:
            for base in HKEYS:
                if read_values(base, r"%s\6.0" % self.__root) is not None:
                    self.warn("It seems you have Visual Studio 6 installed, "
                        "but the expected registry settings are not present.\n"
                        "You must at least run the Visual Studio GUI once "
                        "so that these entries are created.")
                    break
        return []

    def set_path_env_var(self, name):
        """Set environment variable 'name' to an MSVC path type value.

        This is equivalent to a SET command prior to execution of spawned
        commands.
        """

        if name == "lib":
            p = self.get_msvc_paths("library")
        else:
            p = self.get_msvc_paths(name)
        if p:
            os.environ[name] = ';'.join(p)


if get_build_version() >= 8.0:
    log.debug("Importing new compiler from distutils.msvc9compiler")
    OldMSVCCompiler = MSVCCompiler
    from distutils.msvc9compiler import MSVCCompiler
    # get_build_architecture not really relevant now we support cross-compile
    from distutils.msvc9compiler import MacroExpander
site-packages/setuptools/_distutils/__pycache__/versionpredicate.cpython-39.pyc000064400000012157150564745010024114 0ustar00a

(�Re
�@s�dZddlZddlZddlZe�dej�Ze�d�Ze�d�Z	dd�Z
ejejej
ejejejd�ZGd	d
�d
�Zdadd�ZdS)
zBModule for parsing and testing package version predicate strings.
�Nz'(?i)^\s*([a-z_]\w*(?:\.[a-z_]\w*)*)(.*)z^\s*\((.*)\)\s*$z%^\s*(<=|>=|<|>|!=|==)\s*([^\s,]+)\s*$cCs6t�|�}|std|��|��\}}|tj�|�fS)zVParse a single version comparison.

    Return (comparison string, StrictVersion)
    z"bad package restriction syntax: %r)�re_splitComparison�match�
ValueError�groups�	distutils�version�
StrictVersion)�pred�res�compZverStr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/versionpredicate.py�splitUps

r)�<z<=z==�>z>=z!=c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�VersionPredicatea�Parse and test package version predicates.

    >>> v = VersionPredicate('pyepat.abc (>1.0, <3333.3a1, !=1555.1b3)')

    The `name` attribute provides the full dotted name that is given::

    >>> v.name
    'pyepat.abc'

    The str() of a `VersionPredicate` provides a normalized
    human-readable version of the expression::

    >>> print(v)
    pyepat.abc (> 1.0, < 3333.3a1, != 1555.1b3)

    The `satisfied_by()` method can be used to determine with a given
    version number is included in the set described by the version
    restrictions::

    >>> v.satisfied_by('1.1')
    True
    >>> v.satisfied_by('1.4')
    True
    >>> v.satisfied_by('1.0')
    False
    >>> v.satisfied_by('4444.4')
    False
    >>> v.satisfied_by('1555.1b3')
    False

    `VersionPredicate` is flexible in accepting extra whitespace::

    >>> v = VersionPredicate(' pat( ==  0.1  )  ')
    >>> v.name
    'pat'
    >>> v.satisfied_by('0.1')
    True
    >>> v.satisfied_by('0.2')
    False

    If any version numbers passed in do not conform to the
    restrictions of `StrictVersion`, a `ValueError` is raised::

    >>> v = VersionPredicate('p1.p2.p3.p4(>=1.0, <=1.3a1, !=1.2zb3)')
    Traceback (most recent call last):
      ...
    ValueError: invalid version number '1.2zb3'

    It the module or package name given does not conform to what's
    allowed as a legal module or package name, `ValueError` is
    raised::

    >>> v = VersionPredicate('foo-bar')
    Traceback (most recent call last):
      ...
    ValueError: expected parenthesized list: '-bar'

    >>> v = VersionPredicate('foo bar (12.21)')
    Traceback (most recent call last):
      ...
    ValueError: expected parenthesized list: 'bar (12.21)'

    cCs�|��}|std��t�|�}|s.td|��|��\|_}|��}|r�t�|�}|sbtd|��|��d}dd�|�d�D�|_|js�td|��ng|_d	S)
z*Parse a version predicate string.
        zempty package restrictionzbad package name in %rzexpected parenthesized list: %rrcSsg|]}t|��qSr)r)�.0ZaPredrrr
�
<listcomp>t�z-VersionPredicate.__init__.<locals>.<listcomp>�,zempty parenthesized list in %rN)	�stripr�re_validPackagerr�name�re_paren�splitr	)�selfZversionPredicateStrrZparen�strrrr
�__init__`s&

�zVersionPredicate.__init__cCs8|jr.dd�|jD�}|jdd�|�dS|jSdS)NcSs g|]\}}|dt|��qS)� )r)r�cond�verrrr
r}rz,VersionPredicate.__str__.<locals>.<listcomp>z (z, �))r	r�join)r�seqrrr
�__str__{szVersionPredicate.__str__cCs(|jD]\}}t|||�sdSqdS)z�True if version is compatible with all the predicates in self.
        The parameter version must be acceptable to the StrictVersion
        constructor.  It may be either a string or StrictVersion.
        FT)r	�compmap)rrrr rrr
�satisfied_by�szVersionPredicate.satisfied_byN)�__name__�
__module__�__qualname__�__doc__rr$r&rrrr
rs@rcCsdtdurt�dtj�a|��}t�|�}|s8td|��|�d�pDd}|rVtj	�
|�}|�d�|fS)a9Return the name and optional version number of a provision.

    The version number, if given, will be returned as a `StrictVersion`
    instance, otherwise it will be `None`.

    >>> split_provision('mypkg')
    ('mypkg', None)
    >>> split_provision(' mypkg( 1.2 ) ')
    ('mypkg', StrictVersion ('1.2'))
    Nz=([a-zA-Z_]\w*(?:\.[a-zA-Z_]\w*)*)(?:\s*\(\s*([^)\s]+)\s*\))?$z"illegal provides specification: %r��)�
_provision_rx�re�compile�ASCIIrrr�grouprrr)�value�mr rrr
�split_provision�s�
r4)r*r.Zdistutils.versionr�operatorr/r0rrrr�lt�le�eq�gt�ge�ner%rr-r4rrrr
�<module>s�

�nsite-packages/setuptools/_distutils/__pycache__/py38compat.cpython-39.pyc000064400000000731150564745010022550 0ustar00a

(�Re��@sdd�ZdS)cCs4zddl}|��WSty$Yn0d|||fS)N�z%s-%s.%s)�_aix_support�aix_platform�ImportError)�osname�version�releaser�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/py38compat.pyrs
rN)rrrrr	�<module>�site-packages/setuptools/_distutils/__pycache__/py35compat.cpython-39.pyc000064400000001246150564745010022547 0ustar00a

(�Re��@s(ddlZddlZdd�Zeede�ZdS)�NcCs*g}tjj}|dkr&|�dd|�|S)zgReturn a list of command-line arguments reproducing the current
    optimization settings in sys.flags.r�-�O)�sys�flags�optimize�append)�args�value�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/py35compat.py�#__optim_args_from_interpreter_flagss
r�"_optim_args_from_interpreter_flags)r�
subprocessr�getattrr
r
r
r
r�<module>s
�site-packages/setuptools/_distutils/__pycache__/unixccompiler.cpython-39.pyc000064400000015357150564745010023434 0ustar00a

(�Re�8�@s�dZddlZddlZddlZddlZddlmZddlmZddl	m
Z
mZmZddl
mZmZmZmZddlmZejdkr�ddlZGd	d
�d
e
�ZdS)a9distutils.unixccompiler

Contains the UnixCCompiler class, a subclass of CCompiler that handles
the "typical" Unix-style command-line C compiler:
  * macros defined with -Dname[=value]
  * macros undefined with -Uname
  * include search directories specified with -Idir
  * libraries specified with -lllib
  * library search directories specified with -Ldir
  * compile handled by 'cc' (or similar) executable with -c option:
    compiles .c to .o
  * link static library handled by 'ar' command (possibly with 'ranlib')
  * link shared library handled by 'cc -shared'
�N)�	sysconfig)�newer)�	CCompiler�gen_preprocess_options�gen_lib_options)�DistutilsExecError�CompileError�LibError�	LinkError)�log�darwinc
@s�eZdZdZddgdgdgddgdgddgdd�Zejdd�d	krNd
ged
<gd�ZdZd
Z	dZ
dZdZdZ
ZZeZejdkr�dZd'dd�Zdd�Zd(dd�Zd)dd�Zdd�Zdd �Zd!d"�Zd#d$�Zd*d%d&�ZdS)+�
UnixCCompiler�unixN�ccz-shared�arz-cr)�preprocessor�compiler�compiler_so�compiler_cxx�	linker_so�
linker_exe�archiver�ranlib�rr)z.cz.Cz.ccz.cxxz.cppz.mz.oz.az.soz.dylibz.tbdzlib%s%s�cygwinz.exec
Cs�|�d||�}|\}}}t||�}	|j|	}
|r>|
�d|g�|rN||
dd�<|r\|
�|�|
�|�|js~|dus~t||�r�|r�|�tj	�
|��z|�|
�Wn,ty�}zt
|��WYd}~n
d}~00dS)N�-or)Z_fix_compile_argsrr�extend�append�forcer�mkpath�os�path�dirname�spawnrr)�self�sourceZoutput_fileZmacros�include_dirs�
extra_preargs�extra_postargs�
fixed_args�ignore�pp_optsZpp_args�msg�r-��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/unixccompiler.py�
preprocessUs$




zUnixCCompiler.preprocessc	
Csr|j}tjdkr t�|||�}z |�|||d|g|�Wn,tyl}zt|��WYd}~n
d}~00dS)Nrr)r�sys�platform�_osx_support�compiler_fixupr#rr)	r$�obj�src�extZcc_argsr(r+rr,r-r-r.�_compileos
��
zUnixCCompiler._compilerc
Cs�|�||�\}}|j||d�}|�||�r�|�tj�|��|�|j|g||j	�|j
r�z|�|j
|g�Wq�ty�}zt|��WYd}~q�d}~00nt
�d|�dS)N)�
output_dir�skipping %s (up-to-date))�_fix_object_args�library_filename�
_need_linkrr r!r"r#r�objectsrrr	r�debug)r$r=Zoutput_libnamer8r>�target_lang�output_filenamer,r-r-r.�create_static_libzs$����	 zUnixCCompiler.create_static_libc
Cs�|�||�\}}|�|||�}|\}}}t||||�}t|ttd�f�sPtd��|durftj�	||�}|�
||��r�||j|d|g}|	r�dg|dd�<|
r�|
|dd�<|r�|�|�|�
tj�|��z�|tjkr�|jdd�}n|jdd�}|
dk�rr|j�rrd}tj�|d�dk�r@d}d||v�r@|d7}�q&tj�||�d	k�r\d}nd}|j||||<tjd
k�r�t�||�}|�||�Wn.t�y�}zt|��WYd}~n
d}~00nt�d|�dS)Nz%'output_dir' must be a string or Nonerz-grzc++�env��=Z	ld_so_aixrr9)r:Z
_fix_lib_argsr�
isinstance�str�type�	TypeErrorr r!�joinr<r=rrr"rZ
EXECUTABLErrr�basenamer0r1r2r3r#rr
rr>)r$Ztarget_descr=r@r8�	libraries�library_dirs�runtime_library_dirs�export_symbolsr>r'r(�
build_tempr?r)Zlib_optsZld_argsZlinker�i�offsetr,r-r-r.�link�sZ�
���

 zUnixCCompiler.linkcCsd|S)N�-Lr-)r$�dirr-r-r.�library_dir_option�sz UnixCCompiler.library_dir_optioncCsd|vpd|vS)N�gcczg++r-)r$Z
compiler_namer-r-r.�_is_gcc�szUnixCCompiler._is_gcccCs�tj�t�t�d��d�}tjdd�dkrjddl	m
}m}|�}|r`||�ddgkr`d|Sd	|SnNtjdd
�dkr�d|Stjdd�d
kr�|�|�r�dd	|gSdd	|gSt�d�dkr�d|Sd|SdS)N�CCrrr)�get_macosx_target_ver�
split_version�
�z-Wl,-rpath,rS�Zfreebsdz-Wl,-rpath=zhp-uxz-Wl,+sz+s�GNULD�yesz-Wl,--enable-new-dtags,-Rz-Wl,-R)
r r!rJ�shlex�splitr�get_config_varr0r1�distutils.utilrYrZrW)r$rTrrYrZZmacosx_target_verr-r-r.�runtime_library_dir_option�s 

z(UnixCCompiler.runtime_library_dir_optioncCsd|S)Nz-lr-)r$�libr-r-r.�library_optionszUnixCCompiler.library_optioncCs�|j|dd�}|j|dd�}|j|dd�}|j|dd�}tjdkrpt�d�}t�d|�}	|	durfd	}
n
|	�d
�}
|D�] }tj	�
||�}tj	�
||�}
tj	�
||�}tj	�
||�}tjdk�r@|�d�s�|�d��r@|�d
��s@tj	�
|
|d
d�|�}tj	�
|
|d
d�|�}
tj	�
|
|d
d�|�}tj	�
|
|d
d�|�}tj	�|
��rV|
Stj	�|��rl|Stj	�|��r�|Stj	�|�rt|SqtdS)N�shared)Zlib_type�dylib�
xcode_stub�staticr�CFLAGSz-isysroot\s*(\S+)�/rCz/System/z/usr/z/usr/local/)
r;r0r1rrb�re�search�groupr r!rI�
startswith�exists)r$�dirsrer>Zshared_fZdylib_fZxcode_stub_fZstatic_f�cflags�mZsysrootrTrgrhrjrir-r-r.�find_library_filesF



���
zUnixCCompiler.find_library_file)NNNNN)NrN)
NNNNNrNNNN)r)�__name__�
__module__�__qualname__�
compiler_typeZexecutablesr0r1Zsrc_extensionsZ
obj_extensionZstatic_lib_extension�shared_lib_extensionZdylib_lib_extensionZxcode_stub_lib_extensionZstatic_lib_formatZshared_lib_formatZdylib_lib_formatZxcode_stub_lib_formatZ
exe_extensionr/r7rArRrUrWrdrfrur-r-r-r.r
-sL�


�
�
�
B'r
)�__doc__r r0rmr`�	distutilsr�distutils.dep_utilr�distutils.ccompilerrrr�distutils.errorsrrr	r
rr1r2r
r-r-r-r.�<module>s 
site-packages/setuptools/_distutils/__pycache__/msvc9compiler.cpython-39.pyc000064400000042322150564745010023337 0ustar00a

(�Re�v�@sNdZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
ddlmZm
Z
ddlmZddlmZddlZejZejZejZejZejejejejfZej dko�ej!dkZ"e"r�d	Z#d
Z$dZ%ndZ#d
Z$dZ%ddd�Z&Gdd�d�Z'Gdd�d�Z(dd�Z)dd�Z*dd�Z+dd�Z,d$dd�Z-e)�Z.e.d k�r:ed!e.��Gd"d#�d#e�Z/dS)%adistutils.msvc9compiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for the Microsoft Visual Studio 2008.

The module is compatible with VS 2005 and VS 2008. You can find legacy support
for older versions of VS in distutils.msvccompiler.
�N)�DistutilsExecError�DistutilsPlatformError�CompileError�LibError�	LinkError)�	CCompiler�gen_lib_options)�log)�get_platform�win32lz1Software\Wow6432Node\Microsoft\VisualStudio\%0.1fz5Software\Wow6432Node\Microsoft\Microsoft SDKs\Windowsz,Software\Wow6432Node\Microsoft\.NETFrameworkz%Software\Microsoft\VisualStudio\%0.1fz)Software\Microsoft\Microsoft SDKs\Windowsz Software\Microsoft\.NETFramework�x86�amd64�rz	win-amd64c@sPeZdZdZdd�Zee�Zdd�Zee�Zdd�Zee�Zdd	�Ze	e�Zd
S)�Regz2Helper class to read values from the registry
    cCs:tD](}|�||�}|r||vr||Sqt|��dS�N)�HKEYS�read_values�KeyError)�cls�path�key�base�d�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/msvc9compiler.py�	get_value?s
z
Reg.get_valuecCsjzt||�}Wnty"YdS0g}d}zt||�}WntyPYqfYn0|�|�|d7}q,|S)zReturn list of registry keys.Nr�)�RegOpenKeyEx�RegError�
RegEnumKey�append)rrr�handle�L�i�krrr�	read_keysGs


z
Reg.read_keysc	Cs�zt||�}Wnty"YdS0i}d}zt||�\}}}WntyVYq~Yn0|��}|�|�||�|�<|d7}q,|S)z`Return dict of registry keys and values.

        All names are converted to lowercase.
        Nrr)rr�RegEnumValue�lower�convert_mbcs)	rrrr!rr#�name�value�typerrrrYs

zReg.read_valuescCs8t|dd�}|dur4z|d�}Wnty2Yn0|S)N�decode�mbcs)�getattr�UnicodeError)�s�decrrrr(oszReg.convert_mbcsN)
�__name__�
__module__�__qualname__�__doc__r�classmethodr%rr(�staticmethodrrrrr;src@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�
MacroExpandercCsi|_t||_|�|�dSr)�macros�VS_BASE�vsbase�load_macros)�self�versionrrr�__init__{s
zMacroExpander.__init__cCst�||�|jd|<dS)Nz$(%s))rrr9)r=Zmacrorrrrr�	set_macro�szMacroExpander.set_macroc	Cs|�d|jdd�|�d|jdd�|�dtd�z$|dkrP|�d	td
�ntd
��Wntyttd��Yn0|dkr�|�d
|jd�|�dtd�n`d}tD]V}zt||�}Wnty�Yq�Yn0t	|d�}t
�|d||f�}|d|jd<q�dS)NZVCInstallDirz	\Setup\VC�
productdirZVSInstallDirz	\Setup\VSZFrameworkDirZinstallroot� @ZFrameworkSDKDirzsdkinstallrootv2.0aPython was built with Visual Studio 2008;
extensions must be built with a compiler than can generate compatible binaries.
Visual Studio 2008 was not found on this system. If you have Cygwin installed,
you can try compiling with MingW32, by passing "-c mingw32" to setup.py.g"@ZFrameworkVersionzclr versionZ
WindowsSdkDirZcurrentinstallfolderz.Software\Microsoft\NET Framework Setup\Productrz%s\%sr>z$(FrameworkVersion))
r@r;�NET_BASErr�WINSDK_BASErrrrrrr9)r=r>�pr�hrrrrrr<�s2��


zMacroExpander.load_macroscCs$|j��D]\}}|�||�}q
|Sr)r9�items�replace)r=r0r$�vrrr�sub�szMacroExpander.subN)r2r3r4r?r@r<rJrrrrr8ysr8cCs�d}tj�|�}|dkrdS|t|�}tj|d��dd�\}}t|dd��d}|dkrf|d7}t|d	d
��d}|dkr�d}|dkr�||SdS)
z�Return the version of MSVC that was used to build Python.

    For Python 2.3 and up, the version number is included in
    sys.version.  For earlier versions, assume the compiler is MSVC 6.
    zMSC v.����N� r����
��g$@r)�sysr>�find�len�split�int)�prefixr#r0�restZmajorVersionZminorVersionrrr�get_build_version�srYcCs0g}|D]"}tj�|�}||vr|�|�q|S)znReturn a list of normalized paths with duplicates removed.

    The current order of paths is maintained.
    )�osr�normpathr )�pathsZ
reduced_pathsrE�nprrr�normalize_and_reduce_paths�sr^cCs<|�tj�}g}|D]}||vr|�|�qtj�|�}|S)z8Remove duplicate values of an environment variable.
    )rUrZ�pathsepr �join)�variableZoldListZnewListr#ZnewVariablerrr�removeDuplicates�srbcCst|}zt�d|d�}Wn ty<t�d�d}Yn0|rNtj�|�s�d|}tj	�
|d�}|r�tj�|�r�tj�|tjtjd�}tj�
|�}tj�|�s�t�d|�dSnt�d|�|s�t�d	�dStj�|d
�}tj�|�r�|St�d�dS)z�Find the vcvarsall.bat file

    At first it tries to find the productdir of VS 2008 in the registry. If
    that fails it falls back to the VS90COMNTOOLS env var.
    z%s\Setup\VCrAz%Unable to find productdir in registryNzVS%0.f0COMNTOOLSZVCz%s is not a valid directoryz Env var %s is not set or invalidzNo productdir foundz
vcvarsall.bat�Unable to find vcvarsall.bat)r:rrrr	�debugrZr�isdir�environ�getr`�pardir�abspath�isfile)r>r;rAZtoolskeyZtoolsdir�	vcvarsallrrr�find_vcvarsall�s4
�



rlcCsHt|�}hd�}i}|dur$td��t�d||�tjd||ftjtjd�}z�|��\}}|��dkrvt|�	d���|�	d�}|�
d	�D]d}t�|�}d
|vr�q�|�
�}|�
d
d�\}	}
|	��}	|	|vr�|
�tj�r�|
dd�}
t|
�||	<q�W|j��|j��n|j��|j��0t|�t|�k�rDttt|������|S)
zDLaunch vcvarsall.bat and read the settings from its environment
    >Zlibpath�lib�includerNrcz'Calling 'vcvarsall.bat %s' (version=%s)z
"%s" %s & set)�stdout�stderrrr-�
�=rrK)rlrr	rd�
subprocess�Popen�PIPE�communicate�waitr,rUrr(�stripr'�endswithrZr_rbro�closerprT�
ValueError�str�list�keys)r>�archrk�interesting�result�popenrorp�linerr*rrr�query_vcvarsall�sB�


�
r�rBz(VC %0.1f is not supported by this modulec
@s�eZdZdZdZiZdgZgd�ZdgZdgZ	eeee	Z
dZdZd	Z
d
ZdZZdZd,dd�Zd-dd�Zd.dd�Zd/dd�Zd0dd�Zd1dd�Zdd�Zdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd2d(d)�Zd*d+�ZdS)3�MSVCCompilerzwConcrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class.Zmsvcz.c)z.ccz.cppz.cxx�.rcz.mcz.resz.objz.libz.dllz%s%sz.exercCs8t�||||�t|_d|_g|_d|_d|_d|_dS)NzSoftware\Microsoft\VisualStudioF)	rr?�VERSION�_MSVCCompiler__versionZ_MSVCCompiler__root�_MSVCCompiler__paths�	plat_name�_MSVCCompiler__arch�initialized)r=�verbose�dry_run�forcerrrr?HszMSVCCompiler.__init__NcCs|jrJd��|durt�}d}||vr6td|f��dtjvrtdtjvrt|�d�rtd|_d|_d|_d	|_	d
|_
n�|t�ks�|dkr�t|}ntt�dt|}tt
|�}|d
�tj�|_|dtjd<|dtjd<t|j�dkr�td|j��|�d�|_|�d�|_|�d�|_|�d	�|_	|�d
�|_
z(tjd
�d�D]}|j�|��qHWnt�yrYn0t|j�|_d�|j�tjd
<d|_|jdk�r�gd�|_gd�|_ngd�|_gd�|_gd�|_|jdk�r�gd�|_dg|_d|_dS)Nzdon't init multiple timesrz--plat-name must be one of %sZDISTUTILS_USE_SDKZMSSdkzcl.exezlink.exezlib.exezrc.exezmc.exer�_rrmrnrzxPython was built with %s, and extensions need to be built with the same version of the compiler, but it isn't installed.�;r)�/nologo�/O2�/MD�/W3�/DNDEBUG)r��/Od�/MDdr��/Z7�/D_DEBUG)r�r�r�r��/GS-r�)r�r�r�r�r�r�r�)�/DLLr�z/INCREMENTAL:NO�)r�r�z/INCREMENTAL:noz/DEBUGr�T)r�r
rrZrf�find_exe�cc�linkerrm�rc�mc�PLAT_TO_VCVARSr�r�rUr_r�rTZ_MSVCCompiler__productr rr^r`Zpreprocess_optionsr��compile_options�compile_options_debug�ldflags_sharedr��ldflags_shared_debugZldflags_static)r=r�Zok_platsZ	plat_specZvc_envrErrr�
initializeSsd�
�
�




zMSVCCompiler.initialize�cCs�|durd}g}|D]�}tj�|�\}}tj�|�d}|tj�|�d�}||jvrbtd|��|rrtj�|�}||jvr�|�	tj�
|||j��q||jvr�|�	tj�
|||j��q|�	tj�
|||j
��q|S)Nr�rzDon't know how to compile %s)rZr�splitext�
splitdrive�isabs�src_extensionsr�basename�_rc_extensionsr r`�
res_extension�_mc_extensions�
obj_extension)r=Zsource_filenamesZ	strip_dir�
output_dirZ	obj_names�src_namer�extrrr�object_filenames�s,

�
��zMSVCCompiler.object_filenamesc	Cst|js|��|�||||||�}	|	\}}
}}}|p6g}
|
�d�|rT|
�|j�n|
�|j�|
D�]}z||\}}Wnty�YqdYn0|r�tj	�
|�}||jvr�d|}�nX||jvr�d|}�nB||j
v�r<|}d|}z"|�|jg||g|g�Wqdt�y6}zt|��WYd}~qdd}~00qdn�||jv�rtj	�|�}tj	�|�}zl|�|jgd|d|g|g�tj	�tj	�|��\}}tj	�||d�}|�|jgd|g|g�Wqdt�y�}zt|��WYd}~qdd}~00qdntd||f��d	|}z&|�|jg|
|||g|�Wqdt�yl}zt|��WYd}~qdd}~00qd|
S)
Nz/cz/Tcz/Tpz/foz-hz-rr�z"Don't know how to compile %s to %sz/Fo)r�r�Z_setup_compiler �extendr�r�rrZrri�
_c_extensions�_cpp_extensionsr��spawnr�rrr��dirnamer�r�r�r`r�)r=�sourcesr�r9�include_dirsrd�
extra_preargs�extra_postargs�dependsZcompile_info�objectsZpp_opts�buildZcompile_opts�obj�srcr�Z	input_optZ
output_opt�msgZh_dirZrc_dirrr�Zrc_filerrr�compile�s�
�




��


��
��
���
 zMSVCCompiler.compilec	
Cs�|js|��|�||�\}}|j||d�}|�||�r�|d|g}|rJz|�|jg|�Wq�ty�}zt|��WYd}~q�d}~00nt	�
d|�dS)N)r��/OUT:�skipping %s (up-to-date))r�r��_fix_object_args�library_filename�
_need_linkr�rmrrr	rd)	r=r�Zoutput_libnamer�rd�target_lang�output_filenameZlib_argsr�rrr�create_static_libs� zMSVCCompiler.create_static_libc
CsX|js|��|�||�\}}|�|||�}|\}}}|rL|�dt|��t||||�}|durptj�	||�}|�
||��rH|tjkr�|	r�|j
dd�}q�|jdd�}n|	r�|j
}n|j}g}|p�gD]}|�d|�q�||||d|g}tj�|d�}|du�rLtj�tj�|��\}}tj�	||�|��}|�d|�|�|||�|
�rl|
|dd�<|�r||�|�|�tj�|��z|�|jg|�Wn.t�y�}zt|��WYd}~n
d}~00|�||�}|du�rT|\}}d||f}z|�dd	d
||g�Wn.t�yD}zt|��WYd}~n
d}~00nt�d|�dS)Nz5I don't know what to do with 'runtime_library_dirs': rz/EXPORT:r�rz/IMPLIB:z-outputresource:%s;%szmt.exez-nologoz	-manifestr�)r�r�r�Z
_fix_lib_args�warnr|rrZrr`r�r�
EXECUTABLEr�r�r r�r�r�r��manifest_setup_ldargsr��mkpathr�r�rr�manifest_get_embed_infor	rd)r=�target_descr�r�r��	libraries�library_dirs�runtime_library_dirs�export_symbolsrdr�r��
build_tempr�Z
fixed_argsZlib_optsZldflagsZexport_opts�sym�ld_argsZdll_nameZdll_extZimplib_filer�ZmfinfoZ
mffilename�mfidZout_argrrr�link5s��
��

��

��


�
 zMSVCCompiler.linkcCs,tj�|tj�|�d�}|�d|�dS)Nz	.manifest�/MANIFESTFILE:)rZrr`r�r )r=r�r�r��
temp_manifestrrrr��s
�z"MSVCCompiler.manifest_setup_ldargscCs^|D]"}|�d�r|�dd�d}q,qdS|tjkr<d}nd}|�|�}|durVdS||fS)Nr��:rrP)�
startswithrUrr��_remove_visual_c_ref)r=r�r��argr�r�rrrr��s


z$MSVCCompiler.manifest_get_embed_infocCs�z�t|�}z|��}W|��n
|��0t�dtj�}t�|d|�}d}t�|d|�}t�dtj�}t�||�dur|WdSt|d�}z|�|�|W|��WS|��0Wnt	y�Yn0dS)NzU<assemblyIdentity.*?name=("|')Microsoft\.VC\d{2}\.CRT("|').*?(/>|</assemblyIdentity>)r�z*<dependentAssembly>\s*</dependentAssembly>zI<assemblyIdentity.*?name=(?:"|')(.+?)(?:"|').*?(?:/>|</assemblyIdentity>)�w)
�open�readrz�rer��DOTALLrJ�search�write�OSError)r=Z
manifest_fileZ
manifest_fZmanifest_buf�patternrrrr��s6	
��

�z!MSVCCompiler._remove_visual_c_refcCsd|S)Nz	/LIBPATH:r�r=�dirrrr�library_dir_option�szMSVCCompiler.library_dir_optioncCstd��dS)Nz<don't know how to set runtime library search path for MSVC++)rr�rrr�runtime_library_dir_option�s�z'MSVCCompiler.runtime_library_dir_optioncCs
|�|�Sr)r�)r=rmrrr�library_option�szMSVCCompiler.library_optioncCs\|r|d|g}n|g}|D]:}|D]0}tj�||�|��}tj�|�r$|Sq$qdS)NZ_d)rZrr`r��exists)r=�dirsrmrdZ	try_namesr�r)Zlibfilerrr�find_library_file�szMSVCCompiler.find_library_filecCsz|jD].}tj�tj�|�|�}tj�|�r|Sqtjd�d�D].}tj�tj�|�|�}tj�|�rF|SqF|S)a�Return path to an MSVC executable program.

        Tries to find the program in several places: first, one of the
        MSVC program search paths from the registry; next, the directories
        in the PATH environment variable.  If any of those work, return an
        absolute path that is known to exist.  If none of them work, just
        return the original program name, 'exe'.
        �Pathr�)r�rZrr`rirjrfrU)r=ZexerE�fnrrrr�s	


zMSVCCompiler.find_exe)rrr)N)rr�)NNNrNNN)NrN)
NNNNNrNNNN)r) r2r3r4r5�
compiler_typeZexecutablesr�r�r�r�r�r�r�Zstatic_lib_extension�shared_lib_extensionZstatic_lib_formatZshared_lib_formatZ
exe_extensionr?r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr�*sb��

W�
 �
X�
�
_+
r�)r)0r5rZrsrRr��distutils.errorsrrrrr�distutils.ccompilerrr�	distutilsr	�distutils.utilr
�winreg�	OpenKeyExr�EnumKeyrZ	EnumValuer&�errorrZ
HKEY_USERS�HKEY_CURRENT_USER�HKEY_LOCAL_MACHINE�HKEY_CLASSES_ROOTr�platform�maxsizeZNATIVE_WIN64r:rDrCr�rr8rYr^rbrlr�r�r�rrrr�<module>sP��>.#
)
site-packages/setuptools/_distutils/__pycache__/dir_util.cpython-39.pyc000064400000013443150564745010022360 0ustar00a

(�Reb�@spdZddlZddlZddlmZmZddlmZiaddd�Z	dd	d
�Z
ddd�Zd
d�Zddd�Z
dd�ZdS)zWdistutils.dir_util

Utility functions for manipulating directories and directory trees.�N)�DistutilsFileError�DistutilsInternalError)�log��cCsht|t�std|f��tj�|�}g}tj�|�s<|dkr@|St�tj�	|��rV|Stj�
|�\}}|g}|r�|r�tj�|�s�tj�
|�\}}|�d|�ql|D]�}tj�||�}tj�	|�}	t�|	�r�q�|dkr�t
�d|�|�sZzt�||�WnXt�yN}
z>|
jtjk�r$tj�|��s:td||
jdf��WYd}
~
n
d}
~
00|�|�dt|	<q�|S)	a�Create a directory and any missing ancestor directories.

    If the directory already exists (or if 'name' is the empty string, which
    means the current directory, which of course exists), then do nothing.
    Raise DistutilsFileError if unable to create some directory along the way
    (eg. some sub-path exists, but is a file rather than a directory).
    If 'verbose' is true, print a one-line summary of each mkdir to stdout.
    Return the list of directories actually created.
    z(mkpath: 'name' must be a string (got %r)�rrzcreating %szcould not create '%s': %s���N)�
isinstance�strr�os�path�normpath�isdir�
_path_created�get�abspath�split�insert�joinr�info�mkdir�OSError�errno�EEXISTr�args�append)�name�mode�verbose�dry_run�created_dirs�head�tail�tails�d�abs_head�exc�r'��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/dir_util.py�mkpathsB
�
�

r)c	CsNt�}|D] }|�tj�|tj�|���q
t|�D]}t||||d�q4dS)a�Create all the empty directories under 'base_dir' needed to put 'files'
    there.

    'base_dir' is just the name of a directory which doesn't necessarily
    exist yet; 'files' is a list of filenames to be interpreted relative to
    'base_dir'.  'base_dir' + the directory portion of every file in 'files'
    will be created if it doesn't already exist.  'mode', 'verbose' and
    'dry_run' flags are as for 'mkpath()'.
    �rrN)�set�addrrr�dirname�sortedr))�base_dir�filesrrr�need_dir�file�dirr'r'r(�create_treePs
r4c
Csbddlm}|s(tj�|�s(td|��zt�|�}	Wn@tyv}
z(|rPg}	ntd||
jf��WYd}
~
n
d}
~
00|s�t	||d�g}|	D]�}tj�
||�}
tj�
||�}|�d�r�q�|�rtj�|
��rt�
|
�}|dkr�t�d	||�|�st�||�|�|�q�tj�|
��r<|�t|
|||||||d
��q�||
||||||d
�|�|�q�|S)aCopy an entire directory tree 'src' to a new location 'dst'.

    Both 'src' and 'dst' must be directory names.  If 'src' is not a
    directory, raise DistutilsFileError.  If 'dst' does not exist, it is
    created with 'mkpath()'.  The end result of the copy is that every
    file in 'src' is copied to 'dst', and directories under 'src' are
    recursively copied to 'dst'.  Return the list of files that were
    copied or might have been copied, using their output name.  The
    return value is unaffected by 'update' or 'dry_run': it is simply
    the list of all files under 'src', with the names changed to be
    under 'dst'.

    'preserve_mode' and 'preserve_times' are the same as for
    'copy_file'; note that they only apply to regular files, not to
    directories.  If 'preserve_symlinks' is true, symlinks will be
    copied as symlinks (on platforms that support them!); otherwise
    (the default), the destination of the symlink will be copied.
    'update' and 'verbose' are the same as for 'copy_file'.
    r)�	copy_filez&cannot copy tree '%s': not a directoryzerror listing files in '%s': %sN)rz.nfsrzlinking %s -> %sr*)�distutils.file_utilr5rrrr�listdirr�strerrorr)r�
startswith�islink�readlinkrr�symlinkr�extend�	copy_tree)�src�dst�
preserve_mode�preserve_times�preserve_symlinks�updaterrr5�names�e�outputs�n�src_name�dst_name�	link_destr'r'r(r>csR��

���r>cCsft�|�D]F}tj�||�}tj�|�r@tj�|�s@t||�q
|�tj|f�q
|�tj	|f�dS)zHelper for remove_tree().N)
rr7rrrr:�_build_cmdtupler�remove�rmdir)r�	cmdtuples�f�real_fr'r'r(rL�srLcCs�|dkrt�d|�|rdSg}t||�|D]j}z2|d|d�tj�|d�}|tvrbt|=Wq.ty�}zt�d||�WYd}~q.d}~00q.dS)z�Recursively remove an entire directory tree.

    Any errors are ignored (apart from being reported to stdout if 'verbose'
    is true).
    rz'removing '%s' (and everything under it)Nrzerror removing %s: %s)	rrrLrrrrr�warn)�	directoryrrrO�cmdrr&r'r'r(�remove_tree�s

rUcCs6tj�|�\}}|dd�tjkr2||dd�}|S)z�Take the full path 'path', and make it a relative path.

    This is useful to make 'path' the second argument to os.path.join().
    rrN)rr�
splitdrive�sep)r�driver'r'r(�ensure_relative�srY)rrr)rrr)rrrrrr)rr)�__doc__rr�distutils.errorsrr�	distutilsrrr)r4r>rLrUrYr'r'r'r(�<module>s
?
�
E

site-packages/setuptools/_distutils/__pycache__/util.cpython-39.pyc000064400000033717150564745010021530 0ustar00a

(�Re�O�@s(dZddlZddlZddlZddlZddlZddlmZddl	m
Z
ddlmZddl
mZddlmZdd	lmZd
d�Zdd
�Zejdkr�dadZdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdadd�Zdd�Zd/d!d"�Z da!a"a#d#d$�Z$d%d&�Z%d0d'd(�Z&d)d*�Z'd1d+d,�Z(d-d.�Z)dS)2zudistutils.util

Miscellaneous utility functions -- anything that doesn't fit into
one of the other *util.py modules.
�N)�DistutilsPlatformError)�newer)�spawn)�log)�DistutilsByteCompileError�)�"_optim_args_from_interpreter_flagscCs�tjdkrFdtj��vrdSdtj��vr.dSdtj��vr@dStjSdtjvrZtjdStjd	ksnttd
�sttjSt��\}}}}}|���	dd�}|�	d
d�}|�	dd�}|dd�dkr�d||fS|dd�dk�r,|ddk�r�d}dt
|d�d|dd�f}ddd�}|d|tj7}n�|dd�dk�rVd d!lm
}||||�S|dd"�d#k�r�d#}t�d$tj�}|�|�}|�r�|��}n>|dd"�d%k�r�ddl}	ddl}
|	�|
j��|||�\}}}d&|||fS)'a�Return a string that identifies the current platform.  This is used mainly to
    distinguish platform-specific build directories and platform-specific built
    distributions.  Typically includes the OS name and version and the
    architecture (as supplied by 'os.uname()'), although the exact information
    included depends on the OS; eg. on Linux, the kernel version isn't
    particularly important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.

    �nt�amd64�	win-amd64z(arm)�	win-arm32z(arm64)�	win-arm64�_PYTHON_HOST_PLATFORM�posix�uname�/�� �_�-N��linuxz%s-%s�sunosr�5�solarisz%d.%s���32bit�64bit)i���l����z.%s�aixr)�aix_platform��cygwinz[\d.]+�darwinz%s-%s-%s)�os�name�sys�version�lower�platform�environ�hasattrr�replace�int�maxsizeZ
py38compatr �re�compile�ASCII�match�group�_osx_support�distutils.sysconfig�get_platform_osx�	sysconfig�get_config_vars)�osname�host�releaser'�machine�bitnessr �rel_re�mr4�	distutils�rA��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/util.py�get_host_platformsP


 


�
rCcCs:tjdkr0ddddd�}|�tj�d��p.t�St�SdS)Nr	�win32rrr
)�x86�x64�arm�arm64�VSCMD_ARG_TGT_ARCH)r$r%�getr*rC)�TARGET_TO_PLATrArArB�get_platformds
�rLr#�MACOSX_DEPLOYMENT_TARGETcCsdadS)zFor testing only. Do not call.N)�_syscfg_macosx_verrArArArB�_clear_cached_macosx_verusrOcCs.tdur*ddlm}|�t�p d}|r*|atS)z�Get the version of macOS latched in the Python interpreter configuration.
    Returns the version as a string or None if can't obtain one. Cached.Nr)r7r)rNr@r7�get_config_var�MACOSX_VERSION_VAR)r7�verrArArB�!get_macosx_target_ver_from_syscfgzsrScCs^t�}tj�t�}|rZ|rVt|�ddgkrVt|�ddgkrVdtd||f}t|��|S|S)aReturn the version of macOS for which we are building.

    The target version defaults to the version in sysconfig latched at time
    the Python interpreter was built, unless overridden by an environment
    variable. If neither source has a value, then None is returned�
r�$zE mismatch: now "%s" but "%s" during configure; must use 10.3 or later)rSr$r*rJrQ�
split_versionr)Z
syscfg_verZenv_ver�my_msgrArArB�get_macosx_target_ver�s��rXcCsdd�|�d�D�S)zEConvert a dot-separated string into a list of numbers for comparisonscSsg|]}t|��qSrA)r-)�.0�nrArArB�
<listcomp>��z!split_version.<locals>.<listcomp>�.)�split)�srArArBrV�srVcCsztjdkr|S|s|S|ddkr.td|��|ddkrFtd|��|�d�}d|vrd|�d�qP|sntjStjj|�S)a�Return 'pathname' as a name that will work on the native filesystem,
    i.e. split it on '/' and put it back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    rrzpath '%s' cannot be absolute���zpath '%s' cannot end with '/'r])r$�sep�
ValueErrorr^�remove�curdir�path�join)�pathname�pathsrArArB�convert_path�s	

ricCs�tjdkr<tj�|�s$tj�||�Stj�||dd��SnNtjdkr|tj�|�\}}|ddkrn|dd�}tj�||�Stdtj��dS)a	Return 'pathname' with 'new_root' prepended.  If 'pathname' is
    relative, this is equivalent to "os.path.join(new_root,pathname)".
    Otherwise, it requires making 'pathname' relative and then joining the
    two, which is tricky on DOS/Windows and Mac OS.
    rrNr	r�\z!nothing known about platform '%s')r$r%re�isabsrf�
splitdriver)�new_rootrg�drivererArArB�change_root�s

roc	CsvtrdStjdkrXdtjvrXz$ddl}|�t���dtjd<WnttfyVYn0dtjvrnt	�tjd<dadS)aLEnsure that 'os.environ' has all the environment variables we
    guarantee that users can use in config files, command-line options,
    etc.  Currently this includes:
      HOME - user's home directory (Unix only)
      PLAT - description of the current platform, including hardware
             and OS (see 'get_platform()')
    Nr�HOMErr�PLATr)
�_environ_checkedr$r%r*�pwd�getpwuid�getuid�ImportError�KeyErrorrL)rsrArArB�
check_environ�s	
rxc
CsXt�|fdd�}zt�d||�WStyR}ztd|��WYd}~n
d}~00dS)a�Perform shell/Perl-style variable substitution on 'string'.  Every
    occurrence of '$' followed by a name is considered a variable, and
    variable is substituted by the value found in the 'local_vars'
    dictionary, or in 'os.environ' if it's not in 'local_vars'.
    'os.environ' is first checked/augmented to guarantee that it contains
    certain values: see 'check_environ()'.  Raise ValueError for any
    variables not found in either 'local_vars' or 'os.environ'.
    cSs,|�d�}||vrt||�Stj|SdS)Nr)r3�strr$r*)r2�
local_vars�var_namerArArB�_subst�s
zsubst_vars.<locals>._substz\$([a-zA-Z_][a-zA-Z_0-9]*)zinvalid variable '$%s'N)rxr/�subrwrb)r_rzr|�varrArArB�
subst_vars�s	r�error: cCs|t|�S)N)ry)�exc�prefixrArArB�grok_environment_error
sr�cCs(t�dtj�at�d�at�d�adS)Nz
[^\\\'\"%s ]*z'(?:[^'\\]|\\.)*'z"(?:[^"\\]|\\.)*")r/r0�string�
whitespace�
_wordchars_re�
_squote_re�
_dquote_rerArArArB�_init_regexs
r�cCs�tdurt�|��}g}d}|�r�t�||�}|��}|t|�krZ|�|d|���q�||tjvr�|�|d|��||d��	�}d}n�||dkr�|d|�||dd�}|d}n�||dkr�t
�||�}n*||dkr�t�||�}ntd||��|du�r t
d||��|��\}}|d|�||d|d�||d�}|��d	}|t|�kr|�|��q�q|S)
aSplit a string up according to Unix shell-like rules for quotes and
    backslashes.  In short: words are delimited by spaces, as long as those
    spaces are not escaped by a backslash, or inside a quoted string.
    Single and double quotes are equivalent, and the quote characters can
    be backslash-escaped.  The backslash is stripped from any two-character
    escape sequence, leaving only the escaped character.  The quote
    characters are stripped from any quoted string.  Returns a list of
    words.
    Nrrjr�'�"z!this can't happen (bad char '%c')z"bad string (mismatched %s quotes?)r)r�r��stripr2�end�len�appendr�r��lstripr�r��RuntimeErrorrb�span)r_�words�posr?r��begrArArB�split_quoteds>

,
r�cCsP|dur6d|j|f}|dd�dkr6|dd�d}t�|�|sL||�dS)a�Perform some action that affects the outside world (eg.  by
    writing to the filesystem).  Such actions are special because they
    are disabled by the 'dry_run' flag.  This method takes care of all
    that bureaucracy for you; all you have to do is supply the
    function to call and an argument tuple for it (to embody the
    "external action" being performed), and an optional message to
    print.
    Nz%s%r���z,)r�))�__name__r�info)�func�args�msg�verbose�dry_runrArArB�executeYs	
r�cCs2|��}|dvrdS|dvr dStd|f��dS)z�Convert a string representation of truth to true (1) or false (0).

    True values are 'y', 'yes', 't', 'true', 'on', and '1'; false values
    are 'n', 'no', 'f', 'false', 'off', and '0'.  Raises ValueError if
    'val' is anything else.
    )�y�yes�t�true�on�1r)rZ�no�f�false�off�0rzinvalid truth value %rN)r(rb)�valrArArB�	strtoboollsr�c	Cshddl}tjrtd��|dur*do(|dk}|�sTzddlm}	|	d�\}
}Wn,tyxddlm}d|d�}
}Yn0t�	d|�|�s|
dur�t
�|
d	�}
n
t|d	�}
|
�L|
�
d
�|
�
d�tt|��d�|
�
d
|||||f�Wd�n1�s0Ytjg}|�t��|�|�t||d�tt
j|fd||d��nddlm}|D]�}|dd�dk�r~�qd|dk�r�|dk�r�dn|}tjj||d�}ntj�|�}|}|�r�|dt|��|k�r�td||f��|t|�d�}|�rt
j�||�}t
j� |�}|�rd|�s0t!||��rRt�	d||�|�s`||||�nt�"d||��qddS)a~Byte-compile a collection of Python source files to .pyc
    files in a __pycache__ subdirectory.  'py_files' is a list
    of files to compile; any files that don't end in ".py" are silently
    skipped.  'optimize' must be one of the following:
      0 - don't optimize
      1 - normal optimization (like "python -O")
      2 - extra optimization (like "python -OO")
    If 'force' is true, all files are recompiled regardless of
    timestamps.

    The source filename encoded in each bytecode file defaults to the
    filenames listed in 'py_files'; you can modify these with 'prefix' and
    'basedir'.  'prefix' is a string that will be stripped off of each
    source filename, and 'base_dir' is a directory name that will be
    prepended (after 'prefix' is stripped).  You can supply either or both
    (or neither) of 'prefix' and 'base_dir', as you wish.

    If 'dry_run' is true, doesn't actually do anything that would
    affect the filesystem.

    Byte-compilation is either done directly in this interpreter process
    with the standard py_compile module, or indirectly by writing a
    temporary script and executing it.  Normally, you should let
    'byte_compile()' figure out to use direct compilation or not (see
    the source for details).  The 'direct' flag is used by the script
    generated in indirect mode; unless you know what you're doing, leave
    it set to None.
    rNzbyte-compiling is disabled.T)�mkstempz.py)�mktempz$writing byte-compilation script '%s'�wz2from distutils.util import byte_compile
files = [
z,
z]
z�
byte_compile(files, optimize=%r, force=%r,
             prefix=%r, base_dir=%r,
             verbose=%r, dry_run=0,
             direct=1)
)r�zremoving %s)r0���r)�optimizationz1invalid prefix: filename %r doesn't start with %rzbyte-compiling %s to %sz%skipping byte-compilation of %s to %s)#�
subprocessr&�dont_write_bytecoder�tempfiler�rvr�rr�r$�fdopen�open�writerf�map�repr�
executable�extendrr�rr�rc�
py_compiler0�	importlib�util�cache_from_sourcer�rbre�basenamer�debug)�py_files�optimize�forcer��base_dirr�r��directr�r��	script_fd�script_namer��script�cmdr0�file�opt�cfile�dfile�
cfile_baserArArB�byte_compile|st$

�&
�

���r�cCs|�d�}d}|�|�S)z�Return a version of the string escaped for inclusion in an
    RFC-822 header, by ensuring there are 8 spaces space after each newline.
    �
z	
        )r^rf)�header�linesrarArArB�
rfc822_escapes
r�)r�)Nrr)rrNNrrN)*�__doc__r$r/�importlib.utilr�r�r&�distutils.errorsr�distutils.dep_utilr�distutils.spawnrr@rrZ
py35compatrrCrLr)rNrQrOrSrXrVrirorrrxrr�r�r�r�r�r�r�r�r�r�rArArArB�<module>sLP

=
�
site-packages/setuptools/_distutils/__pycache__/text_file.cpython-39.pyc000064400000020543150564745010022527 0ustar00a

(�Re�0�@s&dZddlZddlZGdd�d�ZdS)z�text_file

provides the TextFile class, which gives an interface to text files
that (optionally) takes care of stripping comments, ignoring blank
lines, and joining lines with backslashes.�Nc@steZdZdZdddddddd�Zddd�Zd	d
�Zdd�Zdd
d�Zddd�Z	ddd�Z
dd�Zdd�Zdd�Z
dS)�TextFilea�Provides a file-like object that takes care of all the things you
       commonly want to do when processing a text file that has some
       line-by-line syntax: strip comments (as long as "#" is your
       comment character), skip blank lines, join adjacent lines by
       escaping the newline (ie. backslash at end of line), strip
       leading and/or trailing whitespace.  All of these are optional
       and independently controllable.

       Provides a 'warn()' method so you can generate warning messages that
       report physical line number, even if the logical line in question
       spans multiple physical lines.  Also provides 'unreadline()' for
       implementing line-at-a-time lookahead.

       Constructor is called as:

           TextFile (filename=None, file=None, **options)

       It bombs (RuntimeError) if both 'filename' and 'file' are None;
       'filename' should be a string, and 'file' a file object (or
       something that provides 'readline()' and 'close()' methods).  It is
       recommended that you supply at least 'filename', so that TextFile
       can include it in warning messages.  If 'file' is not supplied,
       TextFile creates its own using 'io.open()'.

       The options are all boolean, and affect the value returned by
       'readline()':
         strip_comments [default: true]
           strip from "#" to end-of-line, as well as any whitespace
           leading up to the "#" -- unless it is escaped by a backslash
         lstrip_ws [default: false]
           strip leading whitespace from each line before returning it
         rstrip_ws [default: true]
           strip trailing whitespace (including line terminator!) from
           each line before returning it
         skip_blanks [default: true}
           skip lines that are empty *after* stripping comments and
           whitespace.  (If both lstrip_ws and rstrip_ws are false,
           then some lines may consist of solely whitespace: these will
           *not* be skipped, even if 'skip_blanks' is true.)
         join_lines [default: false]
           if a backslash is the last non-newline character on a line
           after stripping comments and whitespace, join the following line
           to it to form one "logical line"; if N consecutive lines end
           with a backslash, then N+1 physical lines will be joined to
           form one logical line.
         collapse_join [default: false]
           strip leading whitespace from lines that are joined to their
           predecessor; only matters if (join_lines and not lstrip_ws)
         errors [default: 'strict']
           error handler used to decode the file content

       Note that since 'rstrip_ws' can strip the trailing newline, the
       semantics of 'readline()' must differ from those of the builtin file
       object's 'readline()' method!  In particular, 'readline()' returns
       None for end-of-file: an empty string might just be a blank line (or
       an all-whitespace line), if 'rstrip_ws' is true but 'skip_blanks' is
       not.�r�strict)�strip_comments�skip_blanks�	lstrip_ws�	rstrip_ws�
join_lines�
collapse_join�errorsNcKs�|dur|durtd��|j��D]0}||vr@t||||�q"t|||j|�q"|��D]}||jvr\td|��q\|dur�|�|�n||_||_d|_g|_	dS)z�Construct a new TextFile object.  At least one of 'filename'
           (a string) and 'file' (a file-like object) must be supplied.
           They keyword argument options are described above and affect
           the values returned by 'readline()'.Nz7you must supply either or both of 'filename' and 'file'zinvalid TextFile option '%s'r)
�RuntimeError�default_options�keys�setattr�KeyError�open�filename�file�current_line�linebuf)�selfrr�options�opt�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/text_file.py�__init__Ns
zTextFile.__init__cCs&||_tj|jd|jd�|_d|_dS)zyOpen a new file named 'filename'.  This overrides both the
           'filename' and 'file' arguments to the constructor.�r)rrN)r�iorrrr)rrrrrrosz
TextFile.opencCs$|j}d|_d|_d|_|��dS)ziClose the current file and forget everything we know about it
           (filename, current line number).N)rrr�close)rrrrrrvs
zTextFile.closecCsjg}|dur|j}|�|jd�t|ttf�rD|�dt|��n|�d|�|�t|��d�|�S)Nz, z
lines %d-%d: z	line %d: �)r�appendr�
isinstance�list�tuple�str�join)r�msg�lineZoutmsgrrr�	gen_errorszTextFile.gen_errorcCstd|�||���dS)Nzerror: )�
ValueErrorr(�rr&r'rrr�error�szTextFile.errorcCs tj�d|�||�d�dS)a�Print (to stderr) a warning message tied to the current logical
           line in the current file.  If the current logical line in the
           file spans multiple physical lines, the warning refers to the
           whole range, eg. "lines 3-5".  If 'line' supplied, it overrides
           the current line number; it may be a list or tuple to indicate a
           range of physical lines, or an integer for a single physical
           line.z	warning: �
N)�sys�stderr�writer(r*rrr�warn�sz
TextFile.warncCs�|jr|jd}|jd=|Sd}|j��}|dkr6d}|jr�|r�|�d�}|dkrTnX|dksl||ddkr�|ddkr|dp~d}|d|�|}|��dkr�q n|�d	d�}|j�r|�r|dur�|�d
�|S|j	r�|�
�}||}t|jt
��r
|jdd|jd<n|j|jdg|_n:|du�r,dSt|jt
��rL|jdd|_n|jd|_|j�rr|j�rr|��}n"|j�r�|�
�}n|j�r�|��}|dk�s�|dk�r�|j�r�q |j�r�|ddk�r�|dd�}q |dd�dk�r�|dd�d}q |S)
aURead and return a single logical line from the current file (or
           from an internal buffer if lines have previously been "unread"
           with 'unreadline()').  If the 'join_lines' option is true, this
           may involve reading multiple physical lines concatenated into a
           single string.  Updates the current line number, so calling
           'warn()' after 'readline()' emits a warning about the physical
           line(s) just read.  Returns None on end-of-file, since the empty
           string can occur if 'rstrip_ws' is true but 'strip_blanks' is
           not.���rN�#rr�\r,z\#z2continuation line immediately precedes end-of-file���z\
)rr�readliner�find�strip�replacer	r0r
�lstripr!rr"rr�rstripr)rr'Zbuildup_line�pos�eolrrrr5�sf




	
�


zTextFile.readlinecCs(g}|��}|dur|S|�|�qdS)zWRead and return the list of all logical lines remaining in the
           current file.N)r5r )r�linesr'rrr�	readliness
zTextFile.readlinescCs|j�|�dS)z�Push 'line' (a string) onto an internal buffer that will be
           checked by future 'readline()' calls.  Handy for implementing
           a parser with line-at-a-time lookahead.N)rr )rr'rrr�
unreadlineszTextFile.unreadline)NN)N)N)N)�__name__�
__module__�__qualname__�__doc__r
rrrr(r+r0r5r>r?rrrrr
s$:�	
!	



x
r)rCr-rrrrrr�<module>ssite-packages/setuptools/_distutils/__pycache__/dep_util.cpython-39.pyc000064400000005406150564745010022352 0ustar00a

(�Re�
�@s6dZddlZddlmZdd�Zdd�Zdd	d
�ZdS)z�distutils.dep_util

Utility functions for simple, timestamp-based dependency of files
and groups of files; also, function based entirely on such
timestamp dependency analysis.�N)�DistutilsFileErrorcCs`tj�|�s tdtj�|���tj�|�s0dSddlm}t�|�|}t�|�|}||kS)aReturn true if 'source' exists and is more recently modified than
    'target', or if 'source' exists and 'target' doesn't.  Return false if
    both exist and 'target' is the same age or younger than 'source'.
    Raise DistutilsFileError if 'source' does not exist.
    zfile '%s' does not exist�r��ST_MTIME)�os�path�existsr�abspath�statr)�source�targetr�mtime1�mtime2�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/dep_util.py�newers
�rcCsht|�t|�krtd��g}g}tt|��D]2}t||||�r,|�||�|�||�q,||fS)z�Walk two filename lists in parallel, testing if each source is newer
    than its corresponding target.  Return a pair of lists (sources,
    targets) where source is newer than target, according to the semantics
    of 'newer()'.
    z+'sources' and 'targets' must be same length)�len�
ValueError�ranger�append)�sources�targets�	n_sources�	n_targets�irrr�newer_pairwise sr�errorcCs�tj�|�sdSddlm}t�|�|}|D]P}tj�|�sb|dkrHn|dkrTq.n|dkrbdSt�|�|}||kr.dSq.dS)a�Return true if 'target' is out-of-date with respect to any file
    listed in 'sources'.  In other words, if 'target' exists and is newer
    than every file in 'sources', return false; otherwise return true.
    'missing' controls what we do when a source file is missing; the
    default ("error") is to blow up with an OSError from inside 'stat()';
    if it is "ignore", we silently drop any missing source files; if it is
    "newer", any missing source files make us assume that 'target' is
    out-of-date (this is handy in "dry-run" mode: it'll make you pretend to
    carry out commands that wouldn't work because inputs are missing, but
    that doesn't matter because you're not actually going to run the
    commands).
    rrrr�ignorerN)rrrr
r)rr�missingr�target_mtimer�source_mtimerrr�newer_group6s r!)r)�__doc__r�distutils.errorsrrrr!rrrr�<module>s
site-packages/setuptools/_distutils/__pycache__/debug.cpython-39.pyc000064400000000456150564745010021633 0ustar00a

(�Re��@sddlZej�d�ZdS)�NZDISTUTILS_DEBUG)�os�environ�get�DEBUG�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/debug.py�<module>ssite-packages/setuptools/_distutils/__pycache__/cygwinccompiler.cpython-39.pyc000064400000021210150564745010023732 0ustar00a

(�Re*B�@s�dZddlZddlZddlZddlmZmZmZddlZddl	m
Z
ddlmZddl
mZmZmZmZddlmZddlmZd	d
�ZGdd�de
�ZGd
d�de�ZdZdZdZdd�Ze�d�Zdd�Zdd�Z dd�Z!dS)adistutils.cygwinccompiler

Provides the CygwinCCompiler class, a subclass of UnixCCompiler that
handles the Cygwin port of the GNU C compiler to Windows.  It also contains
the Mingw32CCompiler class which handles the mingw32 port of GCC (same as
cygwin in no-cygwin mode).
�N)�Popen�PIPE�check_output)�
UnixCCompiler)�
write_file)�DistutilsExecError�CCompilerError�CompileError�UnknownFileError)�LooseVersion)�find_executablecCs�tj�d�}|dkr|tj|d|d�}|dkr8dgS|dkrFdgS|d	krTd
gS|dkrbdgS|d
krpdgStd|��dS)zaInclude the appropriate MSVC runtime library if Python was built
    with MSVC 7.0 or later.
    zMSC v.�����
Z1300Zmsvcr70Z1310Zmsvcr71Z1400Zmsvcr80Z1500Zmsvcr90Z1600Zmsvcr100zUnknown MS Compiler version %s N)�sys�version�find�
ValueError)Zmsc_posZmsc_ver�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/cygwinccompiler.py�	get_msvcr?src
@sReZdZdZdZdZdZdZdZdZ	dZ
dd
d�Zdd
�Zddd�Z
ddd�ZdS)�CygwinCCompilerz? Handles the Cygwin port of the GNU C compiler to Windows.
    �cygwinz.o�.az.dllzlib%s%sz%s%sz.exercCsHt�||||�t�\}}|�d||f�|turB|�d|�tj�dd�|_	tj�dd�|_
d|j	vr�t�\|_|_
|_|�|jd|j|j
|jf�|j
dkr�|j	|_nd	|_|j
d
kr�d}q�d}n|j	|_d}|jd
|j	d|j	d
|j
d|j	d|j|fd�d|j	v�r<|jdk�r<dg|_|�d�nt�|_dS)Nz%Python's GCC status: %s (details: %s)z�Python's pyconfig.h doesn't seem to support your compiler. Reason: %s. Compiling may fail because of undefined preprocessor macros.�CC�gcc�CXXzg++z: gcc %s, ld %s, dllwrap %s
z2.10.90�dllwrap�2.13�-shared�
-mdll -staticz%s -mcygwin -O -Wallz%s -mcygwin -mdll -O -Wallz%s -mcygwinz%s -mcygwin %s��compiler�compiler_so�compiler_cxx�
linker_exe�	linker_so�2.91.57�msvcrtz,Consider upgrading to a newer version of gcc)r�__init__�check_config_h�debug_print�CONFIG_H_OK�warn�os�environ�get�cc�cxx�get_versions�gcc_version�
ld_versionZdllwrap_version�
compiler_type�
linker_dll�set_executables�
dll_librariesr)�self�verbose�dry_run�force�status�details�
shared_optionrrrr)dsX
���
�
��



���zCygwinCCompiler.__init__c
Cs�|dks|dkrXz|�dd|d|g�Wq�tyT}zt|��WYd}~q�d}~00nPz"|�|j||d|g|�Wn,ty�}zt|��WYd}~n
d}~00dS)z:Compiles the source by spawning GCC and windres if needed.�.rc�.resZwindresz-iz-oN)�spawnrr	r#)r:�obj�src�extZcc_args�extra_postargsZpp_opts�msgrrr�_compile�s �
zCygwinCCompiler._compileNcCsPt�|
p
g�}
t�|pg�}t�|p&g�}|�|j�|du�r||jksV|jdk�rtj�|d�}tj�tj�	|��\}}tj�
||d�}tj�
|d|d�}dtj�	|�dg}|D]}|�|�q�|�t
||fd	|�|jd
k�r|
�d|g�|
�d|g�n
|�|�|	�s(|
�d
�t�||||||||d|	|
|||
�dS)zLink the objects.Nrrz.def�librz
LIBRARY %sZEXPORTSz
writing %srz--output-libz--defz-s)�copy�extendr9Z
EXECUTABLEr7r.�path�dirname�splitext�basename�join�append�executerr�link)r:Ztarget_descZobjectsZoutput_filename�
output_dir�	libraries�library_dirs�runtime_library_dirs�export_symbols�debugZ
extra_preargsrG�
build_tempZtarget_lang�temp_dirZdll_nameZ
dll_extensionZdef_fileZlib_file�contents�symrrrrT�sH
��

���

�zCygwinCCompiler.link�cCs�|durd}g}|D]�}tj�tj�|��\}}||jddgvrRtd||f��|rbtj�|�}|dvr�|�tj�||||j	��q|�tj�|||j	��q|S)z#Adds supports for rc and res files.Nr_rArBz"unknown file type '%s' (from '%s'))rBrA)
r.rMrO�normcaseZsrc_extensionsr
rPrRrQ�
obj_extension)r:Zsource_filenamesZ	strip_dirrUZ	obj_names�src_name�baserFrrr�object_filenamess&���z CygwinCCompiler.object_filenames)rrr)
NNNNNrNNNN)rr_)�__name__�
__module__�__qualname__�__doc__r6raZstatic_lib_extension�shared_lib_extensionZstatic_lib_formatZshared_lib_formatZ
exe_extensionr)rIrTrdrrrrrYs 
@�
Nrc@seZdZdZdZddd�ZdS)�Mingw32CCompilerz@ Handles the Mingw32 port of the GNU C compiler to Windows.
    Zmingw32rc	Cs�t�||||�d|jvr*|jdkr*d}nd}d|jvrH|jdkrHd}nd}t|j�r^td��|jd	|jd
|jd	|jd|jd|j	||fd
�g|_
t�|_
dS)Nrrr rr'z--entry _DllMain@12r_z1Cygwin gcc cannot be used with --compiler=mingw32z%s -O -Wallz%s -mdll -O -Wallz%sz%s %s %sr!)rr)r1r5r4�is_cygwinccrr8r2r7r9r)r:r;r<r=r@�entry_pointrrrr)s.
����zMingw32CCompiler.__init__N)rrr)rerfrgrhr6r)rrrrrjsrj�okznot okZ	uncertainc
Cs�ddlm}dtjvrtdfSdtjvr0tdfS|��}z`t|�}zHd|��vrhtd|fW|��WSt	d	|fW|��WSW|��n
|��0Wn6t
y�}ztd
||jffWYd}~Sd}~00dS)awCheck if the current Python installation appears amenable to building
    extensions with GCC.

    Returns a tuple (status, details), where 'status' is one of the following
    constants:

    - CONFIG_H_OK: all is well, go ahead and compile
    - CONFIG_H_NOTOK: doesn't look good
    - CONFIG_H_UNCERTAIN: not sure -- unable to read pyconfig.h

    'details' is a human-readable string explaining the situation.

    Note there are two ways to conclude "OK": either 'sys.version' contains
    the string "GCC" (implying that this Python was built with GCC), or the
    installed "pyconfig.h" contains the string "__GNUC__".
    r)�	sysconfigZGCCzsys.version mentions 'GCC'ZClangzsys.version mentions 'Clang'Z__GNUC__z'%s' mentions '__GNUC__'z '%s' does not mention '__GNUC__'zcouldn't read '%s': %sN)
�	distutilsrnrrr,�get_config_h_filename�open�read�close�CONFIG_H_NOTOK�OSError�CONFIG_H_UNCERTAIN�strerror)rn�fn�config_h�excrrrr*Ms*

���r*s(\d+\.\d+(\.\d+)*)cCst|��d}t|�durdSt|dtd�j}z|��}W|��n
|��0t�|�}|durbdSt	|�
d����S)z�Find the version of an executable by running `cmd` in the shell.

    If the command is not found, or the output does not match
    `RE_VERSION`, returns None.
    rNT)�shell�stdout�)�splitrrrr|rrrs�
RE_VERSION�searchr�group�decode)�cmd�
executable�out�
out_string�resultrrr�_find_exe_version~s

r�cCsgd�}tdd�|D��S)zg Try to find out the versions of gcc, ld and dllwrap.

    If not possible it returns None for it.
    )zgcc -dumpversionzld -vzdllwrap --versioncSsg|]}t|��qSr)r�)�.0r�rrr�
<listcomp>��z get_versions.<locals>.<listcomp>)�tuple)�commandsrrrr3�sr3cCst|dg�}|���d�S)zCTry to determine if the compiler that would be used is from cygwin.z-dumpmachinescygwin)r�strip�endswith)r1r�rrrrk�srk)"rhr.rrK�
subprocessrrr�reZdistutils.unixccompilerr�distutils.file_utilr�distutils.errorsrrr	r
Zdistutils.versionr�distutils.spawnrrrrjr,rtrvr*�compilerr�r3rkrrrr�<module>s,1@1/
site-packages/setuptools/_distutils/__pycache__/spawn.cpython-39.pyc000064400000005605150564745010021676 0ustar00a

(�Re�
�@s\dZddlZddlZddlZddlmZmZddlmZddl	m
Z
ddd�Zdd	d
�ZdS)
z�distutils.spawn

Provides the 'spawn()' function, a front-end to various platform-
specific functions for launching another program in a sub-process.
Also provides the 'find_executable()' to search the path for a given
executable name.
�N)�DistutilsPlatformError�DistutilsExecError)�DEBUG)�log�c
Cst|�}t�t�|��|r dS|r@t|d�}|dur@||d<|durL|nttj�}t	j
dkr�ddlm}m
}|�}|r�|||<z tj||d�}	|	��|	j}
WnHty�}z0ts�|d}td||jdf�|�WYd}~n
d}~00|
�rts�|d}td||
f��dS)	a�Run another program, specified as a command list 'cmd', in a new process.

    'cmd' is just the argument list for the new process, ie.
    cmd[0] is the program to run and cmd[1:] are the rest of its arguments.
    There is no way to run a program with a name different from that of its
    executable.

    If 'search_path' is true (the default), the system's executable
    search path will be used to find the program; otherwise, cmd[0]
    must be the exact path to the executable.  If 'dry_run' is true,
    the command will not actually be run.

    Raise DistutilsExecError if running the program fails in any way; just
    return on success.
    Nr�darwin)�MACOSX_VERSION_VAR�get_macosx_target_ver)�envzcommand %r failed: %s���z#command %r failed with exit code %s)�listr�info�
subprocess�list2cmdline�find_executable�dict�os�environ�sys�platform�distutils.utilrr	�Popen�wait�
returncode�OSErrorrr�args)�cmd�search_path�verbose�dry_runr
�
executablerr	Zmacosx_target_ver�proc�exitcode�exc�r$��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/spawn.py�spawns@

��
�r&c	Cs�tj�|�\}}tjdkr*|dkr*|d}tj�|�r:|S|dur�tj�dd�}|dur�zt�d�}Wnt	t
fy�tj}Yn0|s�dS|�tj
�}|D]&}tj�||�}tj�|�r�|Sq�dS)z�Tries to find 'executable' in the directories listed in 'path'.

    A string listing directories separated by 'os.pathsep'; defaults to
    os.environ['PATH'].  Returns the complete filename or None if not found.
    �win32z.exeN�PATH�CS_PATH)r�path�splitextrr�isfiler�get�confstr�AttributeError�
ValueError�defpath�split�pathsep�join)r r*�_�ext�paths�p�fr$r$r%rHs(
r)rrrN)N)
�__doc__rrr�distutils.errorsrr�distutils.debugr�	distutilsrr&rr$r$r$r%�<module>s
6site-packages/setuptools/_distutils/__pycache__/version.cpython-39.pyc000064400000016427150564745010022237 0ustar00a

(�Re�0�@s>dZddlZGdd�d�ZGdd�de�ZGdd�de�ZdS)	a�Provides classes to represent module version numbers (one class for
each style of version numbering).  There are currently two such classes
implemented: StrictVersion and LooseVersion.

Every version number class implements the following interface:
  * the 'parse' method takes a string and parses it to some internal
    representation; if the string is an invalid version number,
    'parse' raises a ValueError exception
  * the class constructor takes an optional string argument which,
    if supplied, is passed to 'parse'
  * __str__ reconstructs the string that was passed to 'parse' (or
    an equivalent string -- ie. one that will generate an equivalent
    version number instance)
  * __repr__ generates Python code to recreate the version number instance
  * _cmp compares the current instance with either another instance
    of the same class or a string (which will be parsed to an instance
    of the same class, thus must follow the same rules)
�Nc@sJeZdZdZddd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dS)�Versionz�Abstract base class for version numbering classes.  Just provides
    constructor (__init__) and reproducer (__repr__), because those
    seem to be the same for all version numbering classes; and route
    rich comparisons to _cmp.
    NcCs|r|�|�dS�N��parse��self�vstring�r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/version.py�__init__&szVersion.__init__cCsd|jjt|�fS)Nz	%s ('%s'))�	__class__�__name__�str�rr	r	r
�__repr__*szVersion.__repr__cCs|�|�}|tur|S|dkS�Nr��_cmp�NotImplemented�r�other�cr	r	r
�__eq__-s
zVersion.__eq__cCs|�|�}|tur|S|dkSrrrr	r	r
�__lt__3s
zVersion.__lt__cCs|�|�}|tur|S|dkSrrrr	r	r
�__le__9s
zVersion.__le__cCs|�|�}|tur|S|dkSrrrr	r	r
�__gt__?s
zVersion.__gt__cCs|�|�}|tur|S|dkSrrrr	r	r
�__ge__Es
zVersion.__ge__)N)r
�
__module__�__qualname__�__doc__rrrrrrrr	r	r	r
rs
rc@s<eZdZdZe�dejejB�Zdd�Z	dd�Z
dd�Zd	S)
�
StrictVersiona?Version numbering for anal retentives and software idealists.
    Implements the standard interface for version number classes as
    described above.  A version number consists of two or three
    dot-separated numeric components, with an optional "pre-release" tag
    on the end.  The pre-release tag consists of the letter 'a' or 'b'
    followed by a number.  If the numeric components of two version
    numbers are equal, then one with a pre-release tag will always
    be deemed earlier (lesser) than one without.

    The following are valid version numbers (shown in the order that
    would be obtained by sorting according to the supplied cmp function):

        0.4       0.4.0  (these two are equivalent)
        0.4.1
        0.5a1
        0.5b3
        0.5
        0.9.6
        1.0
        1.0.4a3
        1.0.4b1
        1.0.4

    The following are examples of invalid version numbers:

        1
        2.7.2.2
        1.3.a4
        1.3pl1
        1.3c4

    The rationale for this version numbering system will be explained
    in the distutils documentation.
    z)^(\d+) \. (\d+) (\. (\d+))? ([ab](\d+))?$cCs�|j�|�}|std|��|�ddddd�\}}}}}|rTttt|||g��|_nttt||g��d|_|r�|dt|�f|_nd|_dS)	Nzinvalid version number '%s'�����)rr)	�
version_re�match�
ValueError�group�tuple�map�int�version�
prerelease)rrr'�major�minor�patchr.Zprerelease_numr	r	r
r�s�zStrictVersion.parsecCsb|jddkr*d�tt|jdd���}nd�tt|j��}|jr^||jdt|jd�}|S)Nr"r�.r!)r-�joinr+rr.rr	r	r
�__str__�szStrictVersion.__str__cCs�t|t�rt|�}nt|t�s"tS|j|jkrB|j|jkr>dSdS|jsR|jsRdS|jrb|jsbdS|jsr|jrrdS|jr�|jr�|j|jkr�dS|j|jkr�dSdSnds�Jd��dS)N���r!rFznever get here)�
isinstancerr rr-r.�rrr	r	r
r�s*


zStrictVersion._cmpN)r
rrr�re�compile�VERBOSE�ASCIIr&rr4rr	r	r	r
r ]s#
�
r c@sHeZdZdZe�dej�Zddd�Zdd�Z	dd	�Z
d
d�Zdd
�ZdS)�LooseVersiona�Version numbering for anarchists and software realists.
    Implements the standard interface for version number classes as
    described above.  A version number consists of a series of numbers,
    separated by either periods or strings of letters.  When comparing
    version numbers, the numeric components will be compared
    numerically, and the alphabetic components lexically.  The following
    are all valid version numbers, in no particular order:

        1.5.1
        1.5.2b2
        161
        3.10a
        8.02
        3.4j
        1996.07.12
        3.2.pl0
        3.1.1.6
        2g6
        11g
        0.960923
        2.2beta29
        1.13++
        5.5.kw
        2.0b1pl0

    In fact, there is no such thing as an invalid version number under
    this scheme; the rules for comparison are simple and predictable,
    but may not always give the results you want (for some definition
    of "want").
    z(\d+ | [a-z]+ | \.)NcCs|r|�|�dSrrrr	r	r
r0szLooseVersion.__init__c	Cs\||_dd�|j�|�D�}t|�D],\}}zt|�||<Wq$tyNYq$0q$||_dS)NcSsg|]}|r|dkr|�qS)r2r	)�.0�xr	r	r
�
<listcomp>:s�z&LooseVersion.parse.<locals>.<listcomp>)r�component_re�split�	enumerater,r(r-)rr�
components�i�objr	r	r
r5szLooseVersion.parsecCs|jSr)rrr	r	r
r4EszLooseVersion.__str__cCsdt|�S)NzLooseVersion ('%s'))rrr	r	r
rIszLooseVersion.__repr__cCsVt|t�rt|�}nt|t�s"tS|j|jkr2dS|j|jkrBdS|j|jkrRdSdS)Nrr5r!)r6rr<rr-r7r	r	r
rMs


zLooseVersion._cmp)N)
r
rrrr8r9r:r@rrr4rrr	r	r	r
r<
s
r<)rr8rr r<r	r	r	r
�<module>
s
>1site-packages/setuptools/_distutils/__pycache__/fancy_getopt.cpython-39.pyc000064400000024753150564745010023235 0ustar00a

(�RexE�@s�dZddlZddlZddlZddlZddlTdZe�de�Ze�deef�Z	e
�dd�ZGd	d
�d
�Z
dd�Zd
d�ejD�Zdd�Zdd�ZGdd�d�Zedkr�dZdD]*Zede�ed�eee���e�q�dS)a6distutils.fancy_getopt

Wrapper around the standard getopt module that provides the following
additional features:
  * short and long options are tied together
  * options have help strings, so fancy_getopt could potentially
    create a complete usage summary
  * options set attributes of a passed-in object
�N)�*z[a-zA-Z](?:[a-zA-Z0-9-]*)z^%s$z^(%s)=!(%s)$�-�_c@s�eZdZdZddd�Zdd�Zdd�Zd d	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
d!dd�Zdd�Zd"dd�Zd#dd�ZdS)$�FancyGetopta�Wrapper around the standard 'getopt()' module that provides some
    handy extra functionality:
      * short and long options are tied together
      * options have help strings, and help text can be assembled
        from them
      * options set attributes of a passed-in object
      * boolean options can have "negative aliases" -- eg. if
        --quiet is the "negative alias" of --verbose, then "--quiet"
        on the command line sets 'verbose' to false
    NcCsN||_i|_|jr|��i|_i|_g|_g|_i|_i|_i|_	g|_
dS�N)�option_table�option_index�_build_index�alias�negative_alias�
short_opts�	long_opts�
short2long�	attr_name�	takes_arg�option_order��selfr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/fancy_getopt.py�__init__)s	zFancyGetopt.__init__cCs(|j��|jD]}||j|d<qdS)Nr)r�clearr)r�optionrrrr	Qs

zFancyGetopt._build_indexcCs||_|��dSr)rr	rrrr�set_option_tableVszFancyGetopt.set_option_tablecCs<||jvrtd|��n |||f}|j�|�||j|<dS)Nz'option conflict: already an option '%s')r�DistutilsGetoptErrorr�append)r�long_option�short_option�help_stringrrrr�
add_optionZs
�
zFancyGetopt.add_optioncCs
||jvS)zcReturn true if the option table for this parser has an
        option with long name 'long_option'.)r�rrrrr�
has_optioncszFancyGetopt.has_optioncCs
|�t�S)z�Translate long option name 'long_option' to the form it
        has as an attribute of some object: ie., translate hyphens
        to underscores.��	translate�
longopt_xlater rrr�
get_attr_namehszFancyGetopt.get_attr_namecCs\t|t�sJ�|��D]@\}}||jvr:td|||f��||jvrtd|||f��qdS)Nz(invalid %s '%s': option '%s' not definedz0invalid %s '%s': aliased option '%s' not defined)�
isinstance�dict�itemsrr)r�aliases�whatr
�optrrr�_check_alias_dictns
�
�zFancyGetopt._check_alias_dictcCs|�|d�||_dS)z'Set the aliases for this option parser.r
N)r,r
)rr
rrr�set_aliasesxszFancyGetopt.set_aliasescCs|�|d�||_dS)z�Set the negative aliases for this option parser.
        'negative_alias' should be a dictionary mapping option names to
        option names, both the key and value must already be defined
        in the option table.znegative aliasN)r,r)rrrrr�set_negative_aliases}sz FancyGetopt.set_negative_aliasescCs�g|_g|_|j��i|_|jD�]�}t|�dkrD|\}}}d}n(t|�dkr^|\}}}}ntd|f��t|t	�r�t|�dkr�t
d|��|dus�t|t	�r�t|�dks�t
d	|��||j|<|j�|�|d
dkr�|r�|d}|dd
�}d|j|<nF|j
�|�}|du�r:|j|�r0t
d
||f��||jd
<d|j|<|j�|�}|du�r�|j||j|k�r�t
d||f��t�|��s�t
d|��|�|�|j|<|r"|j�|�||j|d<q"dS)z�Populate the various data structures that keep tabs on the
        option table.  Called by 'getopt()' before it can do anything
        worthwhile.
        �r�zinvalid option tuple: %r�z9invalid long option '%s': must be a string of length >= 2N�z:invalid short option '%s': must a single character or None����=�:z>invalid negative alias '%s': aliased option '%s' takes a valuezginvalid alias '%s': inconsistent with aliased option '%s' (one of them takes a value, the other doesn'tzEinvalid long option name '%s' (must be letters, numbers, hyphens only)r
rrr�repeatr�len�
ValueErrorr&�strrrrr�getr
�
longopt_re�matchr%r)rr�long�short�helpr6�alias_torrr�_grok_option_table�sr

��
��

��


��
��zFancyGetopt._grok_option_tablec
Cs�|durtjdd�}|dur*t�}d}nd}|��d�|j�}zt�|||j�\}}Wn.tjy�}zt	|��WYd}~n
d}~00|D]�\}}t
|�dkr�|ddkr�|j|d}n,t
|�dkr�|dd�d	ks�J�|dd�}|j�
|�}	|	r�|	}|j|�s<|d
k�sJd��|j�
|�}	|	�r8|	}d}nd}|j|}
|�rn|j�
|
�du�rnt||
d�d}t||
|�|j�||f�q�|�r�||fS|SdS)aParse command-line options in args. Store as attributes on object.

        If 'args' is None or not supplied, uses 'sys.argv[1:]'.  If
        'object' is None or not supplied, creates a new OptionDummy
        object, stores option values there, and returns a tuple (args,
        object).  If 'object' is supplied, it is modified in place and
        'getopt()' just returns 'args'; in both cases, the returned
        'args' is a modified copy of the passed-in 'args' list, which
        is left untouched.
        Nr2TF� r1rrz--�zboolean option can't have value)�sys�argv�OptionDummyrA�joinr�getoptr
�error�DistutilsArgErrorr7rr
r:rrrr6�getattr�setattrrr)r�args�object�created_objectr�opts�msgr+�valr
�attrrrrrH�sF 
zFancyGetopt.getoptcCs|jdurtd��n|jSdS)z�Returns the list of (option, value) tuples processed by the
        previous run of 'getopt()'.  Raises RuntimeError if
        'getopt()' hasn't been called yet.
        Nz!'getopt()' hasn't been called yet)r�RuntimeError)rrrr�get_option_orders

zFancyGetopt.get_option_ordercCsjd}|jD]L}|d}|d}t|�}|ddkr:|d}|durJ|d}||kr
|}q
|ddd}d}||}	d	|}
|r�|g}nd
g}|jD]�}|dd�\}}}t||	�}
|ddkr�|dd�}|du�r|
r�|�d|||
df�n|�d
||f�n:d||f}|
�r4|�d|||
df�n|�d|�|
dd�D]}|�|
|��qNq�|S)z�Generate help text (a list of strings, one per suggested line of
        output) from the option table for this FancyGetopt object.
        rr2r3r4N�r1�NrBzOption summary:r/z  --%-*s  %sz
  --%-*s  z%s (-%s)z  --%-*s)rr7�	wrap_textr)r�header�max_optrr=r>�l�	opt_width�
line_width�
text_width�
big_indent�linesr?�text�	opt_namesrrr�
generate_helpsH



�zFancyGetopt.generate_helpcCs0|durtj}|�|�D]}|�|d�qdS)N�
)rD�stdoutrc�write)rrY�file�linerrr�
print_helphszFancyGetopt.print_help)N)NN)NN)N)NN)�__name__�
__module__�__qualname__�__doc__rr	rrr!r%r,r-r.rArHrUrcrirrrrrs
(
	
M
=

OrcCst|�}|�|�|�||�Sr)rr.rH)�options�negative_optrNrM�parserrrr�fancy_getoptos
rqcCsi|]}t|�d�qS)rB)�ord)�.0Z_wscharrrr�
<dictcomp>u�rtcCs|durgSt|�|kr|gS|��}|�t�}t�d|�}dd�|D�}g}|�rg}d}|r�t|d�}|||kr�|�|d�|d=||}q\|r�|dddkr�|d=q�q\|�r|dkr�|�|dd|��|d|d�|d<|dddk�r|d=|�d�|��qN|S)	z�wrap_text(text : string, width : int) -> [string]

    Split 'text' into multiple lines of no more than 'width' characters
    each, and return the list of strings that results.
    Nz( +|-+)cSsg|]}|r|�qSrr)rs�chrrr�
<listcomp>�ruzwrap_text.<locals>.<listcomp>rr3rBrC)r7�
expandtabsr#�WS_TRANS�re�splitrrG)ra�width�chunksr`�cur_line�cur_lenr[rrrrXws:

rXcCs
|�t�S)zXConvert a long option name to a valid Python identifier by
    changing "-" to "_".
    r")r+rrr�translate_longopt�sr�c@seZdZdZgfdd�ZdS)rFz_Dummy class just used as a place to hold command-line option
    values as instance attributes.cCs|D]}t||d�qdS)zkCreate a new OptionDummy instance.  The attributes listed in
        'options' will be initialized to None.N)rL)rrnr+rrrr�szOptionDummy.__init__N)rjrkrlrmrrrrrrF�srF�__main__z�Tra-la-la, supercalifragilisticexpialidocious.
How *do* you spell that odd word, anyways?
(Someone ask Mary -- she'll know [or she'll
say, "How should I know?"].))�
���(z	width: %drd)rmrD�stringrzrH�distutils.errors�longopt_pat�compiler;�neg_alias_rer9�	maketransr$rrq�
whitespaceryrXr�rFrjra�w�printrGrrrr�<module>s*
T6site-packages/setuptools/_distutils/__pycache__/archive_util.cpython-39.pyc000064400000015045150564745010023223 0ustar00a

(�Re|!�@s>dZddlZddlmZddlZzddlZWneyBdZYn0ddlmZddl	m
Z
ddlmZddl
mZzddlmZWney�dZYn0zdd	lmZWney�dZYn0d
d�Zdd
�Zd#dd�Zd$dd�Zedgdfedgdfedgdfedgdfedgdfegdfd�Zdd �Zd%d!d"�ZdS)&zodistutils.archive_util

Utility functions for creating archive files (tarballs, zip files,
that sort of thing).�N)�warn)�DistutilsExecError)�spawn)�mkpath)�log)�getpwnam)�getgrnamcCsLtdus|durdSzt|�}Wnty6d}Yn0|durH|dSdS)z"Returns a gid, given a group name.N�)r�KeyError��name�result�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/archive_util.py�_get_gids
rcCsLtdus|durdSzt|�}Wnty6d}Yn0|durH|dSdS)z"Returns an uid, given a user name.Nr	)rr
rrrr�_get_uid+s
r�gzipcs6dddddd�}dddd	d
�}|dur:||��vr:td��|d
}	|dkrZ|	|�|d�7}	ttj�|	�|d�ddl}
t�	d�t
���t�������fdd�}|s�|
�|	d||�}z|j
||d�W|��n
|��0|dk�r2tdt�|	||}
tjdk�r||	|
g}n
|d|	g}t||d�|
S|	S)a=Create a (possibly compressed) tar file from all the files under
    'base_dir'.

    'compress' must be "gzip" (the default), "bzip2", "xz", "compress", or
    None.  ("compress" will be deprecated in Python 3.2)

    'owner' and 'group' can be used to define an owner and a group for the
    archive that is being built. If not provided, the current owner and group
    will be used.

    The output tar file will be named 'base_dir' +  ".tar", possibly plus
    the appropriate compression extension (".gz", ".bz2", ".xz" or ".Z").

    Returns the output filename.
    �gz�bz2�xz�)r�bzip2rN�compressz.gzz.bz2z.xzz.Z)rrrrNzKbad value for 'compress': must be None, 'gzip', 'bzip2', 'xz' or 'compress'z.tarr��dry_runrzCreating tar archivecs,�dur�|_�|_�dur(�|_�|_|S)N)�gid�gname�uid�uname)�tarinfo�r�group�ownerrrr�_set_uid_gidasz"make_tarball.<locals>._set_uid_gidzw|%s)�filterz'compress' will be deprecated.�win32z-f)�keys�
ValueError�getr�os�path�dirname�tarfiler�inforr�open�add�closer�PendingDeprecationWarning�sys�platformr)�	base_name�base_dirr�verboserr"r!�tar_compression�compress_ext�archive_namer,r#�tar�compressed_name�cmdrr r�make_tarball7sB���
	


r=c
Cs�|d}ttj�|�|d�tdurn|r.d}nd}ztd|||g|d�Wntyhtd|��Yn0�nJt�d||�|�s�ztj	|d	tj
d
�}Wn$ty�tj	|d	tjd
�}Yn0|��|tj
kr�tj�tj�|d��}|�||�t�d|�t�|�D]�\}}	}
|	D]6}tj�tj�||d��}|�||�t�d|��q|
D]B}tj�tj�||��}tj�|��rP|�||�t�d|��qP�qWd�n1�s�0Y|S)
avCreate a zip file from all the files under 'base_dir'.

    The output zip file will be named 'base_name' + ".zip".  Uses either the
    "zipfile" Python module (if available) or the InfoZIP "zip" utility
    (if installed and found on the default search path).  If neither tool is
    available, raises DistutilsExecError.  Returns the name of the output zip
    file.
    z.ziprNz-rz-rq�zipzkunable to create zip file '%s': could neither import the 'zipfile' module nor find a standalone zip utilityz#creating '%s' and adding '%s' to it�w)�compressionrzadding '%s')rr)r*r+�zipfilerrrr-�ZipFile�ZIP_DEFLATED�RuntimeError�
ZIP_STORED�curdir�normpath�join�write�walk�isfile)r4r5r6r�zip_filename�
zipoptionsr>r*�dirpath�dirnames�	filenamesrrrr�make_zipfilesT	�
���
�
4rQ)rrzgzip'ed tar-file)rrzbzip2'ed tar-file)rrzxz'ed tar-file)rrzcompressed tar file)rNzuncompressed tar filezZIP file)�gztar�bztar�xztar�ztarr:r>cCs|D]}|tvr|SqdS)zqReturns the first format from the 'format' list that is unknown.

    If all formats are known, returns None
    N)�ARCHIVE_FORMATS)�formats�formatrrr�check_archive_formats�s
rYc
Cst��}|dur6t�d|�tj�|�}|s6t�|�|durDtj}d|i}	zt|}
Wnt	yvt
d|��Yn0|
d}|
dD]\}}
|
|	|<q�|dkr�||	d<||	d	<z4|||fi|	��}W|dur�t�d
|�t�|�n"|du�rt�d
|�t�|�0|S)a�Create an archive file (eg. zip or tar).

    'base_name' is the name of the file to create, minus any format-specific
    extension; 'format' is the archive format: one of "zip", "tar", "gztar",
    "bztar", "xztar", or "ztar".

    'root_dir' is a directory that will be the root directory of the
    archive; ie. we typically chdir into 'root_dir' before creating the
    archive.  'base_dir' is the directory where we start archiving from;
    ie. 'base_dir' will be the common prefix of all files and
    directories in the archive.  'root_dir' and 'base_dir' both default
    to the current directory.  Returns the name of the archive file.

    'owner' and 'group' are used when creating a tar archive. By default,
    uses the current owner and group.
    Nzchanging into '%s'rzunknown archive format '%s'r�r>r"r!zchanging back to '%s')r)�getcwdr�debugr*�abspath�chdirrFrVr
r')r4rX�root_dirr5r6rr"r!�save_cwd�kwargs�format_info�func�arg�val�filenamerrr�make_archive�s8

�
rg)rrrNN)rr)NNrrNN)�__doc__r)�warningsrr2rA�ImportError�distutils.errorsr�distutils.spawnr�distutils.dir_utilr�	distutilsr�pwdr�grprrrr=rQrVrYrgrrrr�<module>sH


�
H
=




�	
�site-packages/setuptools/_distutils/__pycache__/filelist.cpython-39.pyc000064400000025201150564745010022353 0ustar00a

(�Re_4�@s�dZddlZddlZddlZddlZddlmZddlmZm	Z	ddl
mZGdd�d�Zdd	�Z
Gd
d�de�Zejfdd
�Zdd�Zddd�ZdS)zsdistutils.filelist

Provides the FileList class, used for poking about the filesystem
and building lists of files.
�N��convert_path)�DistutilsTemplateError�DistutilsInternalError)�logc@s|eZdZdZddd�Zdd�Zejfdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zddd�Zddd�ZdS) �FileLista�A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.

    Instance attributes:
      dir
        directory from which files will be taken -- only used if
        'allfiles' not supplied to constructor
      files
        list of filenames currently being built/filtered/manipulated
      allfiles
        complete list of files under consideration (ie. without any
        filtering applied)
    NcCsd|_g|_dS�N)�allfiles�files)�self�warn�debug_print�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/filelist.py�__init__ szFileList.__init__cCs
||_dSr)r	)rr	rrr�set_allfiles&szFileList.set_allfilescCst|�|_dSr)�findallr	)r�dirrrrr)szFileList.findallcCsddlm}|rt|�dS)z~Print 'msg' to stdout if the global DEBUG (taken from the
        DISTUTILS_DEBUG environment variable) flag is true.
        r)�DEBUGN)�distutils.debugr�print)r�msgrrrrr
,szFileList.debug_printcCs|j�|�dSr)r
�append)r�itemrrrr6szFileList.appendcCs|j�|�dSr)r
�extend)r�itemsrrrr9szFileList.extendcCs<tttjj|j��}g|_|D]}|j�tjj|��qdSr)�sorted�map�os�path�splitr
r�join)rZsortable_filesZ
sort_tuplerrr�sort<sz
FileList.sortcCs@tt|j�ddd�D]$}|j||j|dkr|j|=qdS)N�r���)�range�lenr
)r�irrr�remove_duplicatesEszFileList.remove_duplicatescCs�|��}|d}d}}}|dvrTt|�dkr<td|��dd�|dd�D�}n~|dvr�t|�d	krttd
|��t|d�}dd�|dd�D�}n:|dvr�t|�dkr�td
|��t|d�}ntd|��||||fS)Nr)�include�exclude�global-include�global-exclude�z&'%s' expects <pattern1> <pattern2> ...cSsg|]}t|��qSrr��.0�wrrr�
<listcomp>X�z1FileList._parse_template_line.<locals>.<listcomp>r#)�recursive-include�recursive-exclude�z,'%s' expects <dir> <pattern1> <pattern2> ...cSsg|]}t|��qSrrr.rrrr1^r2)�graft�prunez#'%s' expects a single <dir_pattern>zunknown action '%s')r r&rr)r�line�words�action�patternsr�dir_patternrrr�_parse_template_lineMs0���zFileList._parse_template_linecCsD|�|�\}}}}|dkrV|�dd�|��|D]}|j|dd�s2t�d|�q2�n�|dkr�|�dd�|��|D]}|j|dd�svt�d	|�qv�n�|d
kr�|�dd�|��|D]}|j|dd�s�t�d
|�q��nb|dk�r(|�dd�|��|D]"}|j|dd��st�d|��q�n|dk�rz|�d|d�|�f�|D](}|j||d��sNd}t�|||��qNn�|dk�r�|�d|d�|�f�|D]$}|j||d��s�t�d||��q�nx|dk�r�|�d|�|jd|d��s@t�d|�nB|dk�r4|�d|�|jd|d��s@t�d|�ntd|��dS)Nr)zinclude � r#)�anchorz%warning: no files found matching '%s'r*zexclude z9warning: no previously-included files found matching '%s'r+zglobal-include rz>warning: no files found matching '%s' anywhere in distributionr,zglobal-exclude zRwarning: no previously-included files matching '%s' found anywhere in distributionr3zrecursive-include %s %s)�prefixz:warning: no files found matching '%s' under directory '%s'r4zrecursive-exclude %s %szNwarning: no previously-included files matching '%s' found under directory '%s'r6zgraft z+warning: no directories found matching '%s'r7zprune z6no previously-included directories found matching '%s'z'this cannot happen: invalid action '%s')r=r
r!�include_patternrr�exclude_patternr)rr8r:r;rr<�patternrrrr�process_template_lineis|�
�
�

�
��
��

�
��zFileList.process_template_liner#rcCsld}t||||�}|�d|j�|jdur4|��|jD],}|�|�r:|�d|�|j�|�d}q:|S)a�Select strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.  Patterns
        are not quite the same as implemented by the 'fnmatch' module: '*'
        and '?'  match non-special characters, where "special" is platform-
        dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found, False otherwise.
        Fz%include_pattern: applying regex r'%s'Nz adding T)�translate_patternr
rCr	r�searchr
r)rrCr?r@�is_regex�files_found�
pattern_re�namerrrrA�s�


zFileList.include_patterncCsrd}t||||�}|�d|j�tt|j�ddd�D]4}|�|j|�r8|�d|j|�|j|=d}q8|S)aRemove strings (presumably filenames) from 'files' that match
        'pattern'.  Other parameters are the same as for
        'include_pattern()', above.
        The list 'self.files' is modified in place.
        Return True if files are found, False otherwise.
        Fz%exclude_pattern: applying regex r'%s'r#r$z
 removing T)rEr
rCr%r&r
rF)rrCr?r@rGrHrIr'rrrrB�s�zFileList.exclude_pattern)NN)r#Nr)r#Nr)�__name__�
__module__�__qualname__�__doc__rrr�curdirrr
rrr"r(r=rDrArBrrrrrs

	M
+�rcCs0t�tj|dd��}dd�|D�}ttjj|�S)z%
    Find all files under 'path'
    T)�followlinkscss,|]$\}}}|D]}tj�||�VqqdSr)rrr!)r/�base�dirsr
�filerrr�	<genexpr>�s�z#_find_all_simple.<locals>.<genexpr>)�_UniqueDirs�filterr�walkr�isfile)rZ
all_unique�resultsrrr�_find_all_simple�s
�rZc@s$eZdZdZdd�Zedd��ZdS)rUz�
    Exclude previously-seen dirs from walk results,
    avoiding infinite recursion.
    Ref https://bugs.python.org/issue44497.
    cCsF|\}}}t�|�}|j|jf}||v}|r6|dd�=|�|�|S)z�
        Given an item from an os.walk result, determine
        if the item represents a unique dir for this instance
        and if not, prevent further traversal.
        N)r�stat�st_dev�st_ino�add)rZ	walk_itemrQrRr
r[�	candidate�foundrrr�__call__	s



z_UniqueDirs.__call__cCst|�|�Sr)rV)�clsrrrrrVsz_UniqueDirs.filterN)rKrLrMrNra�classmethodrVrrrrrUsrUcCs6t|�}|tjkr.tjtjj|d�}t||�}t|�S)z�
    Find all files under 'dir' and return the list of full filenames.
    Unless dir is '.', return full filenames with dir prepended.
    )�start)	rZrrO�	functools�partialr�relpathr�list)rr
Zmake_relrrrrs


rcCs8t�|�}tj}tjdkrd}d|}t�d||�}|S)z�Translate a shell-like glob pattern to a regular expression; return
    a string containing the regex.  Differs from 'fnmatch.translate()' in
    that '*' does not match "special characters" (which are
    platform-specific).
    �\z\\\\z\1[^%s]z((?<!\\)(\\\\)*)\.)�fnmatch�	translater�sep�re�sub)rCrIrl�escapedrrr�
glob_to_re)s

rpr#c
Cs
|rt|t�rt�|�S|Std��d�\}}}|rVt|�}|�|�rP|�|�sZJ�nd}|dur�t|�}|�|�r~|�|�s�J�|t|�t|�t|��}t	j
}	t	j
dkr�d}	|t|�t|�t|��}d|||	||f}n|�rd||t|�d�f}t�|�S)aTranslate a shell-like wildcard pattern to a compiled regular
    expression.  Return the compiled regex.  If 'is_regex' true,
    then 'pattern' is directly compiled to a regex (if it's a string)
    or just returned as-is (assumes it's a regex object).
    �_�Nriz\\z%s\A%s%s.*%s%sz%s\A%s)�
isinstance�strrm�compilerp�	partition�
startswith�endswithr&rrl)
rCr?r@rGrdrq�endrIZ	prefix_rerlrrrrE@s.



�rE)r#Nr)rNrrmrjre�distutils.utilr�distutils.errorsrr�	distutilsrrrZ�setrUrOrrprErrrr�<module>sf
site-packages/setuptools/_distutils/__pycache__/extension.cpython-39.pyc000064400000015557150564745010022571 0ustar00a

(�Re)�@s.dZddlZddlZGdd�d�Zdd�ZdS)zmdistutils.extension

Provides the Extension class, used to describe C/C++ extension
modules in setup scripts.�Nc@s"eZdZdZddd�Zdd�ZdS)�	Extensiona�Just a collection of attributes that describes an extension
    module and everything needed to build it (hopefully in a portable
    way, but there are hooks that let you be as unportable as you need).

    Instance attributes:
      name : string
        the full name of the extension, including any packages -- ie.
        *not* a filename or pathname, but Python dotted name
      sources : [string]
        list of source filenames, relative to the distribution root
        (where the setup script lives), in Unix form (slash-separated)
        for portability.  Source files may be C, C++, SWIG (.i),
        platform-specific resource files, or whatever else is recognized
        by the "build_ext" command as source for a Python extension.
      include_dirs : [string]
        list of directories to search for C/C++ header files (in Unix
        form for portability)
      define_macros : [(name : string, value : string|None)]
        list of macros to define; each macro is defined using a 2-tuple,
        where 'value' is either the string to define it to or None to
        define it without a particular value (equivalent of "#define
        FOO" in source or -DFOO on Unix C compiler command line)
      undef_macros : [string]
        list of macros to undefine explicitly
      library_dirs : [string]
        list of directories to search for C/C++ libraries at link time
      libraries : [string]
        list of library names (not filenames or paths) to link against
      runtime_library_dirs : [string]
        list of directories to search for C/C++ libraries at run time
        (for shared extensions, this is when the extension is loaded)
      extra_objects : [string]
        list of extra files to link with (eg. object files not implied
        by 'sources', static library that must be explicitly specified,
        binary resource files, etc.)
      extra_compile_args : [string]
        any extra platform- and compiler-specific information to use
        when compiling the source files in 'sources'.  For platforms and
        compilers where "command line" makes sense, this is typically a
        list of command-line arguments, but for other platforms it could
        be anything.
      extra_link_args : [string]
        any extra platform- and compiler-specific information to use
        when linking object files together to create the extension (or
        to create a new static Python interpreter).  Similar
        interpretation as for 'extra_compile_args'.
      export_symbols : [string]
        list of symbols to be exported from a shared extension.  Not
        used on all platforms, and not generally necessary for Python
        extensions, which typically export exactly one symbol: "init" +
        extension_name.
      swig_opts : [string]
        any extra options to pass to SWIG if a source file has the .i
        extension.
      depends : [string]
        list of files that the extension depends on
      language : string
        extension language (i.e. "c", "c++", "objc"). Will be detected
        from the source extensions if not provided.
      optional : boolean
        specifies that a build failure in the extension should not abort the
        build process, but simply not install the failing extension.
    NcKst|t�std��t|t�r.tdd�|D��s6td��||_||_|pHg|_|pRg|_|p\g|_	|pfg|_
|ppg|_|pzg|_|	p�g|_
|
p�g|_|p�g|_|p�g|_|
p�g|_|p�g|_||_||_t|�dk�rdd�|D�}d�t|��}d	|}t�|�dS)
Nz'name' must be a stringcss|]}t|t�VqdS)N)�
isinstance�str)�.0�v�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/extension.py�	<genexpr>j�z%Extension.__init__.<locals>.<genexpr>z#'sources' must be a list of stringsrcSsg|]}t|��qSr)�repr)r�optionrrr�
<listcomp>�r
z&Extension.__init__.<locals>.<listcomp>z, zUnknown Extension options: %s)rr�AssertionError�list�all�name�sources�include_dirs�
define_macros�undef_macros�library_dirs�	libraries�runtime_library_dirs�
extra_objects�extra_compile_args�extra_link_args�export_symbols�	swig_opts�depends�language�optional�len�join�sorted�warnings�warn)�selfrrrrrrrrrrrrrrrr �kw�options�msgrrr�__init__Vs6

�











zExtension.__init__cCsd|jj|jj|jt|�fS)Nz<%s.%s(%r) at %#x>)�	__class__�
__module__�__qualname__r�id)r&rrr�__repr__�s�zExtension.__repr__)NNNNNNNNNNNNNN)�__name__r,r-�__doc__r*r/rrrrrs"C�
/rcCs�ddlm}m}m}ddlm}ddlm}||�}||dddddd�}�zfg}|��}	|	durd�q�|�	|	�rpqP|	d|	dkr�d	kr�nn|�
d
|	�qP||	|�}	||	�}
|
d}t|g�}d}
|
dd�D�]�}|
dur�|
�|�d}
q�t
j�|�d}|dd�}|dd�}|dv�r2|j�|�q�|d
k�rJ|j�|�q�|dk�r�|�d�}|dk�rz|j�|df�n$|j�|d|�||dd�f�q�|dk�r�|j�|�q�|dk�r�|j�|�q�|dk�r�|j�|�q�|dk�r|j�|�q�|dk�r|j�|�q�|dk�r*|j}
q�|dk�r<|j}
q�|dk�rN|j}
q�|dk�rr|j�|�|�s�|j}
q�|dv�r�|j�|�q�|�
d|�q�|�|�qPW|��n
|��0|S)z3Reads a Setup file and returns Extension instances.r)�parse_makefile�expand_makefile_vars�_variable_rx)�TextFile)�split_quoted�)�strip_comments�skip_blanks�
join_lines�	lstrip_ws�	rstrip_wsN����*z'%s' lines not handled yet�)z.cz.ccz.cppz.cxxz.c++z.mz.mmz-Iz-D�=z-Uz-Cz-lz-Lz-Rz-rpathz-Xlinkerz
-Xcompilerz-u)z.az.soz.slz.oz.dylibzunrecognized argument '%s')�distutils.sysconfigr2r3r4�distutils.text_filer5�distutils.utilr6�readline�matchr%r�append�os�path�splitextrr�findrrrrrrrr�close)�filenamer2r3r4r5r6�vars�file�
extensions�line�words�module�ext�append_next_word�word�suffix�switch�value�equalsrrr�read_setup_file�s��
 







�









rZ)r1rGr$rrZrrrr�<module>szsite-packages/setuptools/_distutils/__pycache__/file_util.cpython-39.pyc000064400000013711150564745010022517 0ustar00a

(�Re��@sZdZddlZddlmZddlmZdddd�Zdd
d�Zdd
d�Zddd�Z	dd�Z
dS)zFdistutils.file_util

Utility functions for operating on single files.
�N)�DistutilsFileError)�log�copyingzhard linkingzsymbolically linking)N�hard�sym�@c
Cs�d}d}�z�zt|d�}Wn6tyP}ztd||jf��WYd}~n
d}~00tj�|�r�zt�|�Wn6ty�}ztd||jf��WYd}~n
d}~00zt|d�}Wn6ty�}ztd||jf��WYd}~n
d}~00z|�|�}Wn8t�y0}ztd||jf��WYd}~n
d}~00|�s<�q�z|�	|�Wq�t�y�}ztd||jf��WYd}~q�d}~00q�W|�r�|�
�|�r�|�
�n|�r�|�
�|�r�|�
�0dS)	a5Copy the file 'src' to 'dst'; both must be filenames.  Any error
    opening either file, reading from 'src', or writing to 'dst', raises
    DistutilsFileError.  Data is read/written in chunks of 'buffer_size'
    bytes (default 16k).  No attempt is made to handle anything apart from
    regular files.
    N�rbzcould not open '%s': %szcould not delete '%s': %s�wbzcould not create '%s': %szcould not read from '%s': %szcould not write to '%s': %s)�open�OSErrorr�strerror�os�path�exists�unlink�read�write�close)�src�dst�buffer_size�fsrc�fdst�e�buf�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/file_util.py�_copy_file_contentssT	(����
�r�cCs
ddlm}ddlm}	m}
m}m}tj�	|�s<t
d|��tj�|�rd|}
tj�|tj�
|��}ntj�|�}
|r�|||�s�|dkr�t�d|�|dfSzt|}Wnty�td|��Yn0|dk�r
tj�
|�tj�
|�kr�t�d|||
�nt�d|||�|�r|dfS|d	k�rntj�|��r@tj�||��s�zt�||�|dfWSt�yjYn0n<|d
k�r�tj�|��r�tj�||��s�t�||�|dfSt||�|�s�|�rt�|�}|�r�t�|||	||
f�|�rt�||||��|dfS)aCopy a file 'src' to 'dst'.  If 'dst' is a directory, then 'src' is
    copied there with the same name; otherwise, it must be a filename.  (If
    the file exists, it will be ruthlessly clobbered.)  If 'preserve_mode'
    is true (the default), the file's mode (type and permission bits, or
    whatever is analogous on the current platform) is copied.  If
    'preserve_times' is true (the default), the last-modified and
    last-access times are copied as well.  If 'update' is true, 'src' will
    only be copied if 'dst' does not exist, or if 'dst' does exist but is
    older than 'src'.

    'link' allows you to make hard links (os.link) or symbolic links
    (os.symlink) instead of copying: set it to "hard" or "sym"; if it is
    None (the default), files are copied.  Don't set 'link' on systems that
    don't support it: 'copy_file()' doesn't check if hard or symbolic
    linking is available. If hardlink fails, falls back to
    _copy_file_contents().

    Under Mac OS, uses the native file copy function in macostools; on
    other systems, uses '_copy_file_contents()' to copy file contents.

    Return a tuple (dest_name, copied): 'dest_name' is the actual name of
    the output file, and 'copied' is true if the file was copied (or would
    have been copied, if 'dry_run' true).
    r)�newer)�ST_ATIME�ST_MTIME�ST_MODE�S_IMODEz4can't copy '%s': doesn't exist or not a regular filerz"not copying %s (output up-to-date)z&invalid value '%s' for 'link' argumentz%s %s -> %srr)�distutils.dep_utilr�statr r!r"r#r
r�isfiler�isdir�join�basename�dirnamer�debug�_copy_action�KeyError�
ValueError�infor�samefile�linkr�symlinkr�utime�chmod)rr�
preserve_mode�preserve_times�updater1�verbose�dry_runrr r!r"r#�dir�action�strrr�	copy_fileCsV!�





r=cCs�ddlm}m}m}m}m}ddl}	|dkr:t�d||�|rB|S||�sVt	d|��||�rrt
j�|||��}n||�r�t	d||f��|||��s�t	d||f��d	}
zt
�
||�WnRt�y}z8|j\}}
||	jkr�d
}
nt	d|||
f��WYd}~n
d}~00|
�r�t|||d�zt
�|�Wnht�y�}zN|j\}}
zt
�|�Wnt�ynYn0t	d
||||
f��WYd}~n
d}~00|S)a%Move a file 'src' to 'dst'.  If 'dst' is a directory, the file will
    be moved into it with the same name; otherwise, 'src' is just renamed
    to 'dst'.  Return the new full name of the file.

    Handles cross-device moves on Unix using 'copy_file()'.  What about
    other systems???
    r)rr&r'r)r*Nrzmoving %s -> %sz#can't move '%s': not a regular filez0can't move '%s': destination '%s' already existsz2can't move '%s': destination '%s' not a valid pathFTzcouldn't move '%s' to '%s': %s)r8zAcouldn't move '%s' to '%s' by copy/delete: delete '%s' failed: %s)�os.pathrr&r'r)r*�errnorr/rr
rr(�renamer�args�EXDEVr=r)rrr8r9rr&r'r)r*r?�copy_itr�num�msgrrr�	move_file�s`����

�

��rFcCs>t|d�}z$|D]}|�|d�qW|��n
|��0dS)z{Create a file with the specified name and write 'contents' (a
    sequence of strings without line terminators) to it.
    �w�
N)r
rr)�filename�contents�f�linerrr�
write_file�s

rM)r)rrrNrr)rr)�__doc__r
�distutils.errorsr�	distutilsrr,rr=rFrMrrrr�<module>s�
3�
d�
?site-packages/setuptools/_distutils/__pycache__/dist.cpython-39.pyc000064400000103275150564745010021513 0ustar00a

(�Re���@s�dZddlZddlZddlZddlmZzddlZWneyJdZYn0ddlTddl	m
Z
mZddlm
Z
mZmZddlmZddlmZe�d	�Zd
d�ZGdd
�d
�ZGdd�d�Zdd�ZdS)z}distutils.dist

Provides the Distribution class, which represents the module distribution
being built/installed/distributed.
�N)�message_from_file)�*)�FancyGetopt�translate_longopt)�
check_environ�	strtobool�
rfc822_escape��log)�DEBUGz^[a-zA-Z]([a-zA-Z0-9_]*)$cCsPt|t�rn@t|t�sLt|�j}d}|jfit���}t�tj|�t|�}|S)Nz>Warning: '{fieldname}' should be a list, got type '{typename}')	�
isinstance�str�list�type�__name__�format�localsr
�WARN)�value�	fieldname�typename�msg�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/dist.py�_ensure_lists


rc@sDeZdZdZgd�ZdZgd�Zdd�eD�ZddiZdId
d�Z	dd
�Z
dJdd�Zdd�ZdKdd�Z
dd�Zdd�Zdd�Zdd�Zddgfdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�ZdLd,d-�ZdMd.d/�ZdNd1d2�Zejfd3d4�Zd5d6�Zd7d8�Z d9d:�Z!d;d<�Z"d=d>�Z#d?d@�Z$dAdB�Z%dCdD�Z&dEdF�Z'dGdH�Z(d	S)O�Distributiona�The core of the Distutils.  Most of the work hiding behind 'setup'
    is really done within a Distribution instance, which farms the work out
    to the Distutils commands specified on the command line.

    Setup scripts will almost never instantiate Distribution directly,
    unless the 'setup()' function is totally inadequate to their needs.
    However, it is conceivable that a setup script might wish to subclass
    Distribution for some specialized purpose, and then pass the subclass
    to 'setup()' as the 'distclass' keyword argument.  If so, it is
    necessary to respect the expectations that 'setup' has of Distribution.
    See the code for 'setup()', in core.py, for details.
    ))�verbose�vzrun verbosely (default)�)�quiet�qz!run quietly (turns verbosity off))zdry-run�nzdon't actually do anything)�help�hzshow detailed help message)zno-user-cfgNz-ignore pydistutils.cfg in your home directoryz�Common commands: (see '--help-commands' for more)

  setup.py build      will build the package underneath 'build/'
  setup.py install    will install the package
))z
help-commandsNzlist all available commands)�nameNzprint package name)�version�Vzprint package version)�fullnameNzprint <package name>-<version>)�authorNzprint the author's name)�author-emailNz print the author's email address)�
maintainerNzprint the maintainer's name)zmaintainer-emailNz$print the maintainer's email address)�contactNz7print the maintainer's name if known, else the author's)z
contact-emailNz@print the maintainer's email address if known, else the author's)�urlNzprint the URL for this package)�licenseNz print the license of the package)�licenceNzalias for --license)�descriptionNzprint the package description)zlong-descriptionNz"print the long package description)�	platformsNzprint the list of platforms)�classifiersNzprint the list of classifiers)�keywordsNzprint the list of keywords)�providesNz+print the list of packages/modules provided)�requiresNz+print the list of packages/modules required)�	obsoletesNz0print the list of packages/modules made obsoletecCsg|]}t|d��qS)r�r)�.0�xrrr�
<listcomp>��zDistribution.<listcomp>rrNcCs\d|_d|_d|_|jD]}t||d�qt�|_|jjD] }d|}t||t|j|��q:i|_	d|_
d|_d|_i|_
g|_d|_i|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_i|_i|_|�r|�d�}|du�r8|d=|��D]4\}}|� |�}|��D]\}	}
d|
f||	<�q�qd|v�r~|d|d	<|d=d
}t!du�rnt!�"|�nt#j$�%|d�|��D]�\}}
t&|jd|��r�t|jd|�|
�nNt&|j|��r�t|j||
�n0t&||��r�t|||
�nd
t'|�}t!�"|��q�d|_(|jdu�rP|jD].}
|
�)d��s6�qP|
dk�r d|_(�qP�q |�*�dS)a0Construct a new Distribution instance: initialize all the
        attributes of a Distribution, and then use 'attrs' (a dictionary
        mapping attribute names to values) to assign some of those
        attributes their "real" values.  (Any attributes not mentioned in
        'attrs' will be assigned to some null value: 0, None, an empty list
        or dictionary, etc.)  Most importantly, initialize the
        'command_obj' attribute to the empty dictionary; this will be
        filled in with real command objects by 'parse_command_line()'.
        rr�get_N��optionszsetup scriptr.r-z:'licence' distribution option is deprecated; use 'license'�
�set_zUnknown distribution option: %sT�-z
--no-user-cfgF)+r�dry_runr"�display_option_names�setattr�DistributionMetadata�metadata�_METHOD_BASENAMES�getattr�cmdclass�command_packages�script_name�script_args�command_options�
dist_files�packages�package_data�package_dir�
py_modules�	libraries�headers�ext_modules�ext_package�include_dirs�
extra_path�scripts�
data_files�password�command_obj�have_run�get�items�get_option_dict�warnings�warn�sys�stderr�write�hasattr�repr�
want_user_cfg�
startswith�finalize_options)�self�attrs�attr�basename�method_namer=�command�cmd_options�opt_dict�opt�valr�key�argrrr�__init__�s~








zDistribution.__init__cCs&|j�|�}|dur"i}|j|<|S)z�Get the option dictionary for a given command.  If that
        command's option dictionary hasn't been created yet, then create it
        and return the new dictionary; otherwise, return the existing
        option dictionary.
        N)rLr])rjro�dictrrrr_'szDistribution.get_option_dictr<c	Cs�ddlm}|dur"t|j���}|dur@|�||�|d}|sV|�|d�dS|D]h}|j�|�}|dur�|�|d|�qZ|�|d|�||�}|�d�D]}|�|d|�q�qZdS)Nr)�pformatz  zno commands known yetzno option dict for '%s' commandzoption dict for '%s' command:r>)�pprintrx�sortedrL�keys�announcer]�split)	rj�header�commands�indentrx�cmd_namerq�out�linerrr�dump_option_dicts2s*��zDistribution.dump_option_dictscCs�g}t�tj�tjdj�}tj�|d�}tj�|�rB|�	|�tj
dkrRd}nd}|jr�tj�tj�d�|�}tj�|�r�|�	|�d}tj�|�r�|�	|�t
r�|�dd	�|��|S)
a�Find as many configuration files as should be processed for this
        platform, and return a list of filenames in the order in which they
        should be parsed.  The filenames returned are guaranteed to exist
        (modulo nasty race conditions).

        There are three possible config files: distutils.cfg in the
        Distutils installation directory (ie. where the top-level
        Distutils __inst__.py file lives), a file in the user's home
        directory named .pydistutils.cfg on Unix and pydistutils.cfg
        on Windows/Mac; and setup.cfg in the current directory.

        The file in the user's home directory can be disabled with the
        --no-user-cfg option.
        �	distutilsz
distutils.cfg�posixz.pydistutils.cfgzpydistutils.cfg�~z	setup.cfgzusing config files: %sz, )r�os�path�dirnamerb�modules�__file__�join�isfile�appendr$rg�
expanduserrr|)rj�files�sys_dir�sys_file�
user_filename�	user_file�
local_filerrr�find_config_filesNs&



zDistribution.find_config_filescCs�ddlm}tjtjkr"gd�}ng}t|�}|dur>|��}trL|�d�|�}|D]�}trl|�d|�|�	|�|�
�D]V}|�|�}|�|�}|D]8}	|	dkr�|	|vr�|�
||	�}
|	�dd�}	||
f||	<q�q~|��qVd	|jv�r�|jd	��D]�\}	\}}
|j�
|	�}zF|�r.t||t|
��n(|	d
v�rJt||	t|
��nt||	|
�Wq�t�y�}
zt|
��WYd}
~
q�d}
~
00q�dS)Nr)�ConfigParser)
zinstall-basezinstall-platbasezinstall-libzinstall-platlibzinstall-purelibzinstall-headerszinstall-scriptszinstall-data�prefixzexec-prefix�home�user�rootz"Distribution.parse_config_files():z  reading %srr@�_�global)rrA)�configparserr�rbr��base_prefix�	frozensetr�rr|�read�sectionsr=r_r]�replacervrLr^�negative_optrCr�
ValueError�DistutilsOptionError)rj�	filenamesr��ignore_options�parser�filename�sectionr=rqrrrs�src�aliasrrrr�parse_config_files~sD






zDistribution.parse_config_filescCs�|��}g|_t||j�}|�|j�|�ddi�|j|j|d�}|�	�}t
�|j�|�
|�rhdS|r�|�||�}|durhdSqh|jr�|j|t|j�dk|jd�dS|js�td��dS)	a�Parse the setup script's command line, taken from the
        'script_args' instance attribute (which defaults to 'sys.argv[1:]'
        -- see 'setup()' in core.py).  This list is first processed for
        "global options" -- options that set attributes of the Distribution
        instance.  Then, it is alternately scanned for Distutils commands
        and options for that command.  Each new command terminates the
        options for the previous command.  The allowed options for a
        command are determined by the 'user_options' attribute of the
        command class -- thus, we have to be able to load command classes
        in order to parse the command line.  Any error in that 'options'
        attribute raises DistutilsGetoptError; any error on the
        command-line raises DistutilsArgError.  If no Distutils commands
        were found on the command line, raises DistutilsArgError.  Return
        true if command-line was successfully parsed and we should carry
        on with executing commands; false if no errors but we shouldn't
        execute commands (currently, this only happens if user asks for
        help).
        r.r-)�args�objectNr��display_optionsrzno commands suppliedT)�_get_toplevel_optionsrrr��set_negative_aliasesr��set_aliases�getoptrK�get_option_orderr
�
set_verbosityr�handle_display_options�_parse_command_optsr"�
_show_help�len�DistutilsArgError)rj�toplevel_optionsr�r��option_orderrrr�parse_command_line�s.	
�zDistribution.parse_command_linecCs|jdgS)z�Return the non-display options recognized at the top level.

        This includes options that are recognized *only* at the top
        level as well as options recognized for commands.
        )zcommand-packages=Nz0list of packages that provide distutils commands)�global_options�rjrrrr��s�z"Distribution._get_toplevel_optionsc
Csddlm}|d}t�|�s*td|��|j�|�z|�|�}Wn,typ}zt	|��WYd}~n
d}~00t
||�s�td|��t|d�r�t
|jt�s�d}t||��|j}t|d�r�|��}|�|j�t|d	�r�t
|jt�r�t|j�}ng}|�|j|j|�|�|�|�|d
d��\}}	t|	d��rX|	j�rX|j|d|gd�dSt|d	��r�t
|jt��r�d}
|jD]F\}}}
}t|	|�|���r|d
}
t|��r�|�ntd
||f���q||
�r�dS|�|�}t|	���D]\}}d|f||<�q�|S)a�Parse the command-line options for a single command.
        'parser' must be a FancyGetopt instance; 'args' must be the list
        of arguments, starting with the current command (whose options
        we are about to parse).  Returns a new version of 'args' with
        the next command at the front of the list; will be the empty
        list if there are no more commands on the command line.  Returns
        None if the user asked for help on this command.
        r��Commandzinvalid command name '%s'Nz&command class %s must subclass Command�user_optionszIcommand class %s must provide 'user_options' attribute (a list of tuples)r��help_optionsrr"r�zYinvalid help function %r for help option '%s': must be a callable object (function, etc.)zcommand line) �
distutils.cmdr��
command_re�match�
SystemExitrr��get_command_class�DistutilsModuleErrorr��
issubclass�DistutilsClassErrorrerr�rr��copy�updater��fix_help_options�set_option_tabler�r�r�r"r��
get_attr_name�callabler_�varsr^)rjr�r�r�ro�	cmd_classrr�r��opts�help_option_found�help_option�short�desc�funcrqr$rrrrr�sr


�

�


���

�
��
z Distribution._parse_command_optscCsPdD]F}t|j|�}|durqt|t�rdd�|�d�D�}t|j||�qdS)z�Set final values for all the options on the Distribution
        instance, analogous to the .finalize_options() method of Command
        objects.
        �r2r0NcSsg|]}|���qSr��strip)r7�elmrrrr9kr:z1Distribution.finalize_options.<locals>.<listcomp>�,)rGrErr
r}rC)rjrlrrrrrias
zDistribution.finalize_optionsrc
Csddlm}ddlm}|rR|r*|��}n|j}|�|�|�|jd�t	d�|rt|�|j
�|�d�t	d�|jD]z}t|t
�r�t||�r�|}	n
|�|�}	t|	d�r�t|	jt�r�|�|	jt|	j��n|�|	j�|�d|	j�t	d�qzt	||j��d	S)
abShow help for the setup script command-line in the form of
        several lists of command-line options.  'parser' should be a
        FancyGetopt instance; do not expect it to be returned in the
        same state, as its option table will be reset to make it
        generate the correct help text.

        If 'global_options' is true, lists the global options:
        --verbose, --dry-run, etc.  If 'display_options' is true, lists
        the "display-only" options: --name, --version, etc.  Finally,
        lists per-command help for every command name or command class
        in 'commands'.
        r��	gen_usager�z
Global options:r<zKInformation display options (just display information, ignore any commands)r�zOptions for '%s' command:N)�distutils.corer�r�r�r�r�r��
print_help�common_usage�printr�rrrr�r�rer�rr�r�rrJ)
rjr�r�r�rr�r�r=ro�klassrrrr�ns:

�



��
zDistribution._show_helpc	Cs�ddlm}|jr4|��td�t||j��dSd}i}|jD]}d||d<qB|D]l\}}|rX|�|�rXt|�}t	|j
d|��}|dvr�td�|��n |dvr�td	�|��nt|�d}qX|S)
z�If there were any non-global "display-only" options
        (--help-commands or the metadata display options) on the command
        line, display the requested info and return true; else return
        false.
        rr�r<rr;r�r�)r1r3r4r5r>)r�r��
help_commands�print_commandsr�rJr�r]rrGrEr�)	rjr�r��any_display_options�is_display_option�optionrrrsrrrrr��s*
z#Distribution.handle_display_optionsc	Csht|d�|D]R}|j�|�}|s.|�|�}z
|j}WntyNd}Yn0td|||f�qdS)zZPrint a subset of the list of all commands -- used by
        'print_commands()'.
        �:�(no description available)z
  %-*s  %sN)r�rHr]r�r/�AttributeError)rjrr~�
max_length�cmdr�r/rrr�print_command_list�s


zDistribution.print_command_listcCs�ddl}|jj}i}|D]}d||<qg}|j��D]}|�|�s4|�|�q4d}||D]}t|�|krZt|�}qZ|�|d|�|r�t	�|�|d|�dS)anPrint out a help message listing all available commands with a
        description of each.  The list is divided into "standard commands"
        (listed in distutils.command.__all__) and "extra commands"
        (mentioned in self.cmdclass, but not a standard command).  The
        descriptions come from the command class attribute
        'description'.
        rNrzStandard commandszExtra commands)
�distutils.commandro�__all__rHr{r]r�r�r�r�)rjr��std_commands�is_stdr��extra_commandsr�rrrr��s.


��zDistribution.print_commandsc		Cs�ddl}|jj}i}|D]}d||<qg}|j��D]}|�|�s4|�|�q4g}||D]N}|j�|�}|sx|�|�}z
|j}Wnt	y�d}Yn0|�||f�qZ|S)a>Get a list of (command, description) tuples.
        The list is divided into "standard commands" (listed in
        distutils.command.__all__) and "extra commands" (mentioned in
        self.cmdclass, but not a standard command).  The descriptions come
        from the command class attribute 'description'.
        rNrr�)
r�ror�rHr{r]r�r�r/r�)	rjr�r�r�r�r��rvr�r/rrr�get_command_list�s(	




zDistribution.get_command_listcCsN|j}t|t�sJ|durd}dd�|�d�D�}d|vrD|�dd�||_|S)z9Return a list of packages from which commands are loaded.Nr<cSsg|]}|dkr|���qS)r<r�)r7�pkgrrrr9"r:z5Distribution.get_command_packages.<locals>.<listcomp>r�zdistutils.commandr)rIrrr}�insert)rj�pkgsrrr�get_command_packagess
z!Distribution.get_command_packagesc	Cs�|j�|�}|r|S|��D]�}d||f}|}zt|�tj|}Wnty\YqYn0zt||�}Wn$ty�t	d|||f��Yn0||j|<|St	d|��dS)aoReturn the class that implements the Distutils command named by
        'command'.  First we check the 'cmdclass' dictionary; if the
        command is mentioned there, we fetch the class object from the
        dictionary and return it.  Otherwise we load the command module
        ("distutils.command." + command) and fetch the command class from
        the module.  The loaded class is also stored in 'cmdclass'
        to speed future calls to 'get_command_class()'.

        Raises DistutilsModuleError if the expected module could not be
        found, or if that module does not define the expected class.
        z%s.%sz3invalid command '%s' (no class '%s' in module '%s')zinvalid command '%s'N)
rHr]r��
__import__rbr��ImportErrorrGr�r�)rjror��pkgname�module_name�
klass_name�modulerrrr�(s,
��

zDistribution.get_command_classcCsl|j�|�}|sh|rhtr&|�d|�|�|�}||�}|j|<d|j|<|j�|�}|rh|�||�|S)aReturn the command object for 'command'.  Normally this object
        is cached on a previous call to 'get_command_obj()'; if no command
        object for 'command' is in the cache, then we either create and
        return it (if 'create' is true) or return None.
        z<Distribution.get_command_obj(): creating '%s' command objectr)r[r]rr|r�r\rL�_set_command_options)rjro�create�cmd_objr�r=rrr�get_command_objNs�

zDistribution.get_command_objcCsZ|��}|dur|�|�}tr,|�d|�|��D�]\}\}}trZ|�d|||f�zdd�|jD�}Wnty�g}Yn0z
|j}Wnty�i}Yn0z|t|t	�}	||vr�|	r�t
|||t|��nJ||vr�|	r�t
||t|��n,t||��rt
|||�nt
d|||f��Wq4t�yR}
zt
|
��WYd}
~
q4d}
~
00q4dS)aySet the options for 'command_obj' from 'option_dict'.  Basically
        this means copying elements of a dictionary ('option_dict') to
        attributes of an instance ('command').

        'command_obj' must be a Command instance.  If 'option_dict' is not
        supplied, uses the standard option dictionary for this command
        (from 'self.command_options').
        Nz#  setting options for '%s' command:z    %s = %s (from %s)cSsg|]}t|��qSrr6)r7�orrrr9}s�z5Distribution._set_command_options.<locals>.<listcomp>z1error in %s: command '%s' has no such option '%s')�get_command_namer_rr|r^�boolean_optionsr�r�rr
rCrrer�r�)rjr[�option_dict�command_namer��sourcer�	bool_opts�neg_opt�	is_stringrrrrrisF	

��




��z!Distribution._set_command_optionsrcCs|ddlm}t||�s&|}|�|�}n|��}|js8|S|��d|_d|j|<|�|�|rx|�	�D]}|�
||�qf|S)a�Reinitializes a command to the state it was in when first
        returned by 'get_command_obj()': ie., initialized but not yet
        finalized.  This provides the opportunity to sneak option
        values in programmatically, overriding or supplementing
        user-supplied values from the config files and command line.
        You'll have to re-finalize the command object (by calling
        'finalize_options()' or 'ensure_finalized()') before using it for
        real.

        'command' should be a command name (string) or command object.  If
        'reinit_subcommands' is true, also reinitializes the command's
        sub-commands, as declared by the 'sub_commands' class attribute (if
        it has one).  See the "install" command for an example.  Only
        reinitializes the sub-commands that actually matter, ie. those
        whose test predicates return true.

        Returns the reinitialized command object.
        rr�)r�r�rr	r�	finalized�initialize_optionsr\r�get_sub_commands�reinitialize_command)rjro�reinit_subcommandsr�r�subrrrr�s


z!Distribution.reinitialize_commandcCst�||�dS�Nr	)rjr�levelrrrr|�szDistribution.announcecCs|jD]}|�|�qdS)z�Run each command that was seen on the setup script command line.
        Uses the list of commands found and cache of command objects
        created by 'get_command_obj()'.
        N)r�run_command)rjr�rrr�run_commands�s
zDistribution.run_commandscCsD|j�|�rdSt�d|�|�|�}|��|��d|j|<dS)a�Do whatever it takes to run a command (including nothing at all,
        if the command has already been run).  Specifically: if we have
        already created and run the command named by 'command', return
        silently without doing anything.  If the command named by 'command'
        doesn't even have a command object yet, create one.  Then invoke
        'run()' on that command object (or an existing one).
        Nz
running %sr)r\r]r
�infor	�ensure_finalized�run)rjrorrrrr�s	
zDistribution.run_commandcCst|jp|jpg�dkS�Nr)r�rNrQr�rrr�has_pure_modules�szDistribution.has_pure_modulescCs|jot|j�dkSr )rTr�r�rrr�has_ext_modules�szDistribution.has_ext_modulescCs|jot|j�dkSr )rRr�r�rrr�has_c_libraries�szDistribution.has_c_librariescCs|��p|��Sr)r!r"r�rrr�has_modules�szDistribution.has_modulescCs|jot|j�dkSr )rSr�r�rrr�has_headers�szDistribution.has_headerscCs|jot|j�dkSr )rXr�r�rrr�has_scripts�szDistribution.has_scriptscCs|jot|j�dkSr )rYr�r�rrr�has_data_files�szDistribution.has_data_filescCs|��o|��o|��Sr)r!r"r#r�rrr�is_pure�s
��zDistribution.is_pure)N)NNr<)N)r)N)r))r�
__module__�__qualname__�__doc__r�r�r�rBr�rvr_r�r�r�r�r�r�rir�r�r�r�r�r�r�r	rrr
�INFOr|rrr!r"r#r$r%r&r'r(rrrrr-sN,

0
:C[
�
2(!"&

,
)
rc@seZdZdZdZdBdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zd d!�Zd"d#�ZeZd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zd2d3�Zd4d5�Zd6d7�Zd8d9�Z d:d;�Z!d<d=�Z"d>d?�Z#d@dA�Z$dS)CrDz]Dummy class to hold the distribution meta-data: name, version,
    author, and so forth.
    )r$r%r(�author_emailr*�maintainer_emailr,r-r/�long_descriptionr2r0r'r+�
contact_emailr1�download_urlr3r4r5NcCs�|dur|�t|��nfd|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_d|_d|_
d|_d|_d|_d|_d|_dSr)�
read_pkg_file�openr$r%r(r-r*r.r,r-r/r/r2r0r1r1r3r4r5)rjr�rrrrvs&zDistributionMetadata.__init__cst|���fdd�}�fdd�}�d}|d�|_|d�|_|d�|_|d	�|_d
|_|d�|_d
|_|d�|_|d
�|_	d�vr�|d�|_
nd
|_
|d�|_|d�|_d�vr�|d��d�|_
|d�|_|d�|_|dkr�|d�|_|d�|_|d�|_nd
|_d
|_d
|_d
S)z-Reads the metadata values from a file object.cs�|}|dkrdS|S�N�UNKNOWNr)r$r�rrr�_read_field)sz7DistributionMetadata.read_pkg_file.<locals>._read_fieldcs��|d�}|gkrdS|Sr)�get_all)r$�valuesr6rr�
_read_list/sz6DistributionMetadata.read_pkg_file.<locals>._read_listzmetadata-versionr$r%�summaryr(Nr)z	home-pager-zdownload-urlr/r2r��platform�
classifier�1.1r4r3r5)rr$r%r/r(r*r-r.r,r-r1r/r}r2r0r1r4r3r5)rj�filer7r:�metadata_versionrr6rr2%s:












z"DistributionMetadata.read_pkg_filecCsFttj�|d�ddd��}|�|�Wd�n1s80YdS)z7Write the PKG-INFO file into the release tree.
        zPKG-INFO�wzUTF-8)�encodingN)r3r�r�r��write_pkg_file)rj�base_dir�pkg_inforrr�write_pkg_infoYs
�z#DistributionMetadata.write_pkg_infocCsbd}|js"|js"|js"|js"|jr&d}|�d|�|�d|���|�d|���|�d|���|�d|�	��|�d|�
��|�d	|���|�d
|���|jr�|�d|j�t
|���}|�d|�d
�|���}|�r|�d|�|�|d|���|�|d|���|�|d|���|�|d|���|�|d|���dS)z9Write the PKG-INFO format data to a file object.
        z1.0r>zMetadata-Version: %s
z	Name: %s
zVersion: %s
zSummary: %s
zHome-page: %s
zAuthor: %s
zAuthor-email: %s
zLicense: %s
zDownload-URL: %s
zDescription: %s
r�z
Keywords: %s
�Platform�
Classifier�Requires�Provides�	ObsoletesN)r3r4r5r1r1rd�get_name�get_version�get_description�get_url�get_contact�get_contact_email�get_licenser�get_long_descriptionr��get_keywords�_write_list�
get_platforms�get_classifiers�get_requires�get_provides�
get_obsoletes)rjr?r%�	long_descr2rrrrC`s6��z#DistributionMetadata.write_pkg_filecCs |D]}|�d||f�qdS)Nz%s: %s
)rd)rjr?r$r9rrrrrU�sz DistributionMetadata._write_listcCs
|jpdSr4)r$r�rrrrL�szDistributionMetadata.get_namecCs
|jpdS)Nz0.0.0)r%r�rrrrM�sz DistributionMetadata.get_versioncCsd|��|��fS)Nz%s-%s)rLrMr�rrr�get_fullname�sz!DistributionMetadata.get_fullnamecCs
|jpdSr4)r(r�rrr�
get_author�szDistributionMetadata.get_authorcCs
|jpdSr4)r-r�rrr�get_author_email�sz%DistributionMetadata.get_author_emailcCs
|jpdSr4)r*r�rrr�get_maintainer�sz#DistributionMetadata.get_maintainercCs
|jpdSr4)r.r�rrr�get_maintainer_email�sz)DistributionMetadata.get_maintainer_emailcCs|jp|jpdSr4)r*r(r�rrrrP�sz DistributionMetadata.get_contactcCs|jp|jpdSr4)r.r-r�rrrrQ�sz&DistributionMetadata.get_contact_emailcCs
|jpdSr4)r,r�rrrrO�szDistributionMetadata.get_urlcCs
|jpdSr4)r-r�rrrrR�sz DistributionMetadata.get_licensecCs
|jpdSr4)r/r�rrrrN�sz$DistributionMetadata.get_descriptioncCs
|jpdSr4)r/r�rrrrS�sz)DistributionMetadata.get_long_descriptioncCs
|jpgSr)r2r�rrrrT�sz!DistributionMetadata.get_keywordscCst|d�|_dS)Nr2)rr2�rjrrrr�set_keywords�sz!DistributionMetadata.set_keywordscCs|jp
dgSr4)r0r�rrrrV�sz"DistributionMetadata.get_platformscCst|d�|_dS)Nr0)rr0rarrr�
set_platforms�sz"DistributionMetadata.set_platformscCs
|jpgSr)r1r�rrrrW�sz$DistributionMetadata.get_classifierscCst|d�|_dS)Nr1)rr1rarrr�set_classifiers�sz$DistributionMetadata.set_classifierscCs
|jpdSr4)r1r�rrr�get_download_url�sz%DistributionMetadata.get_download_urlcCs
|jpgSr)r4r�rrrrX�sz!DistributionMetadata.get_requirescCs,ddl}|D]}|j�|�qt|�|_dSr )�distutils.versionpredicate�versionpredicate�VersionPredicaterr4�rjrr�rrrr�set_requires�sz!DistributionMetadata.set_requirescCs
|jpgSr)r3r�rrrrY�sz!DistributionMetadata.get_providescCs6dd�|D�}|D]}ddl}|j�|�q||_dS)NcSsg|]}|���qSrr�)r7rrrrr9�r:z5DistributionMetadata.set_provides.<locals>.<listcomp>r)rfrg�split_provisionr3)rjrrr�rrr�set_provides�s
z!DistributionMetadata.set_providescCs
|jpgSr)r5r�rrrrZ�sz"DistributionMetadata.get_obsoletescCs,ddl}|D]}|j�|�qt|�|_dSr )rfrgrhrr5rirrr�
set_obsoletes�sz"DistributionMetadata.set_obsoletes)N)%rr)r*r+rFrvr2rFrCrUrLrMr\r]r^r_r`rPrQrOrR�get_licencerNrSrTrbrVrcrWrdrerXrjrYrlrZrmrrrrrDsD	
4"rDcCs$g}|D]}|�|dd��q|S)zConvert a 4-tuple 'help_options' list as found in various command
    classes to the 3-tuple form required by FancyGetopt.
    r�)r�)r=�new_options�
help_tuplerrrr��sr�)r+rbr��re�emailrr`r�distutils.errors�distutils.fancy_getoptrr�distutils.utilrrrr�r
�distutils.debugr�compiler�rrrDr�rrrr�<module>s4

Zcsite-packages/setuptools/_distutils/__pycache__/_msvccompiler.cpython-39.pyc000064400000033045150564745010023407 0ustar00a

(�ReMQ�@sdZddlZddlZddlZddlZddlZe�e��ddl	Z	Wd�n1sT0Yddl
mZmZm
Z
mZmZddlmZmZddlmZddlmZddlmZdd	�Zd
d�Zdd
ddd�Zdd�Zdd�Zddd�Zddddd�ZGdd�de�Z dS)adistutils._msvccompiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for Microsoft Visual Studio 2015.

The module is compatible with VS 2015 and later. You can find legacy support
for older versions in distutils.msvc9compiler and distutils.msvccompiler.
�N)�DistutilsExecError�DistutilsPlatformError�CompileError�LibError�	LinkError)�	CCompiler�gen_lib_options)�log)�get_platform)�countcCsztjtjdtjtjBd�}Wnty<t�d�YdS0d}d}|��t�D]�}zt�	||�\}}}Wnty�Yq�Yn0|rR|tj
krRtj�
|�rRztt|��}Wnttfy�YqRYn0|dkrR||krR||}}qRWd�n1�s�0Y||fS)Nz'Software\Microsoft\VisualStudio\SxS\VC7)�accesszVisual C++ is not registered�NNr�)�winreg�	OpenKeyEx�HKEY_LOCAL_MACHINEZKEY_READZKEY_WOW64_32KEY�OSErrorr	�debugrZ	EnumValue�REG_SZ�os�path�isdir�int�float�
ValueError�	TypeError)�key�best_version�best_dir�i�vZvc_dirZvt�version�r"��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/_msvccompiler.py�_find_vc2015 s2
�



,r$c
Cs�tj�d�ptj�d�}|s dSz8tjtj�|ddd�ddd	d
ddd
dg	ddd���}Wntjt	t
fytYdS0tj�|ddd�}tj�|�r�d|fSdS)aJReturns "15, path" based on the result of invoking vswhere.exe
    If no install is found, returns "None, None"

    The version is returned to avoid unnecessarily changing the function
    result. It may be ignored when the path is not None.

    If vswhere.exe is not available, by definition, VS 2017 is not
    installed.
    zProgramFiles(x86)ZProgramFilesr
zMicrosoft Visual StudioZ	Installerzvswhere.exez-latestz-prereleasez	-requiresz1Microsoft.VisualStudio.Component.VC.Tools.x86.x64z	-propertyZinstallationPathz	-products�*�mbcs�strict)�encoding�errorsZVCZ	AuxiliaryZBuild�)r�environ�get�
subprocess�check_outputr�join�strip�CalledProcessErrorr�UnicodeDecodeErrorr)�rootrr"r"r#�_find_vc2017<s(
��r4�x86�x64�arm�arm64)r5�	x86_amd64�x86_arm�	x86_arm64cCs\t�\}}|st�\}}|s*t�d�dStj�|d�}tj�|�sTt�d|�dS|dfS)Nz$No suitable Visual C++ version foundr
z
vcvarsall.batz%s cannot be found)r4r$r	rrrr/�isfile)�	plat_spec�_rr�	vcvarsallr"r"r#�_find_vcvarsallcs


r@c
Cs�t�d�rdd�tj��D�St|�\}}|s6td��z&tjd�||�tj	d�j
ddd	�}WnBtjy�}z(t�
|j�td
�|j���WYd}~n
d}~00dd�dd
�|��D�D�}|S)NZDISTUTILS_USE_SDKcSsi|]\}}|��|�qSr"��lower)�.0r�valuer"r"r#�
<dictcomp>ws�z_get_vc_env.<locals>.<dictcomp>zUnable to find vcvarsall.batzcmd /u /c "{}" {} && set)�stderrzutf-16le�replace)r)zError executing {}cSs$i|]\}}}|r|r|��|�qSr"rA)rCrr>rDr"r"r#rE�s�css|]}|�d�VqdS)�=N)�	partition)rC�liner"r"r#�	<genexpr>��z_get_vc_env.<locals>.<genexpr>)r�getenvr+�itemsr@rr-r.�format�STDOUT�decoder1r	�error�output�cmd�
splitlines)r=r?r>�out�exc�envr"r"r#�_get_vc_envus.
�
��
��rYcCsN|st�d��tj�}|D].}tj�tj�|�|�}tj�|�r|Sq|S)atReturn path to an MSVC executable program.

    Tries to find the program in several places: first, one of the
    MSVC program search paths from the registry; next, the directories
    in the PATH environment variable.  If any of those work, return an
    absolute path that is known to exist.  If none of them work, just
    return the original program name, 'exe'.
    r)rrM�split�pathseprr/�abspathr<)Zexe�paths�p�fnr"r"r#�	_find_exe�s	
r`r9r:r;)�win32z	win-amd64z	win-arm32z	win-arm64c
s�eZdZdZdZiZdgZgd�ZdgZdgZ	eeee	Z
dZdZd	Z
d
ZdZZdZd(dd�Zd)dd�Zd*dd�Zd+dd�Zd,dd�Zd-dd�Z�fdd�Zej�fdd��Zd d!�Zd"d#�Zd$d%�Zd.d&d'�Z�Z S)/�MSVCCompilerzwConcrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class.Zmsvcz.c)z.ccz.cppz.cxx�.rcz.mcz.resz.objz.libz.dllz%s%sz.exercCs t�||||�d|_d|_dS)NF)r�__init__�	plat_name�initialized)�self�verbose�dry_run�forcer"r"r#rd�szMSVCCompiler.__init__NcCs|jrJd��|durt�}|tvr6td�tt����t|}t|�}|sRtd��|�dd�|_|j�	t
j�}td|�|_
td|�|_td|�|_td	|�|_td
|�|_td|�|_|�dd��	t
j�D]}|r�|�|�t
j��q�|�d
d��	t
j�D]}|r�|�|�t
j��q�d|_gd�|_gd�|_gd�}gd�}g|�d�|_g|�d�|_g|�d�d�d�|_g|�d�d�d�|_g|�|_g|�|_t j!df|jt j!df|jt j!df|jt j"df|jt j"df|jt j"df|jt j#df|jt j#df|jt j#df|ji	|_$d|_dS)Nzdon't init multiple timesz--plat-name must be one of {}z7Unable to find a compatible Visual Studio installation.r�zcl.exezlink.exezlib.exezrc.exezmc.exezmt.exe�include�lib)�/nologoz/O2�/W3z/GLz/DNDEBUGz/MD)rnz/Odz/MDdz/Ziroz/D_DEBUG)rn�/INCREMENTAL:NO�/LTCG)rnrprqz/DEBUG:FULLz/MANIFEST:EMBED,ID=1z/DLLz/MANIFEST:EMBED,ID=2z/MANIFESTUAC:NOFT)%rfr
�PLAT_TO_VCVARSrrO�tuplerYr,�_pathsrZrr[r`�cc�linkerrm�rc�mc�mtZadd_include_dir�rstrip�sepZadd_library_dirZpreprocess_options�compile_options�compile_options_debugZldflags_exeZldflags_exe_debugZldflags_sharedZldflags_shared_debugZldflags_staticZldflags_static_debugrZ
EXECUTABLEZ
SHARED_OBJECTZSHARED_LIBRARY�_ldflags)rgrer=Zvc_envr]�dir�ldflagsZ
ldflags_debugr"r"r#�
initialize�s^�



�zMSVCCompiler.initializerkcsXi�fdd��jD���fdd��j�jD����p8d����fdd�}tt||��S)Ncsi|]}|�j�qSr")�
obj_extension�rC�ext�rgr"r#rE&rLz1MSVCCompiler.object_filenames.<locals>.<dictcomp>csi|]}|�j�qSr")�
res_extensionr�r�r"r#rE'rLrkcs�tj�|�\}}�r"tj�|�}n2tj�|�\}}|�tjjtjjf�rT|dd�}ztj��|�|�WSt	y�t
d�|���Yn0dS)N�zDon't know how to compile {})rr�splitext�basename�
splitdrive�
startswithr{�altsepr/�LookupErrorrrO)r^�baser�r>)�ext_map�
output_dir�	strip_dirr"r#�
make_out_path,sz4MSVCCompiler.object_filenames.<locals>.make_out_path)�src_extensions�_rc_extensions�_mc_extensions�list�map)rgZsource_filenamesr�r�r�r")r�r�rgr�r#�object_filenames!s��zMSVCCompiler.object_filenamesc	Cs�|js|��|�||||||�}	|	\}}
}}}|p6g}
|
�d�|rT|
�|j�n|
�|j�d}|
D�]}z||\}}Wnty�YqhYn0|r�tj	�
|�}||jvr�d|}�nH||jvr�d|}d}�n.||j
v�r@|}d|}z|�|jg|||g�Wqht�y:}zt|��WYd}~qhd}~00qhn�||jv�r�tj	�|�}tj	�|�}z\|�|jd|d||g�tj	�tj	�|��\}}tj	�||d	�}|�|jd||g�Wqht�y�}zt|��WYd}~qhd}~00qhntd
�||���|jg|
|}|�r$|�d�|�|�|�d|�|�|�z|�|�Wqht�y�}zt|��WYd}~qhd}~00qh|
S)
Nz/cFz/Tcz/TpTz/foz-hz-rrcz"Don't know how to compile {} to {}z/EHscz/Fo)rfr�Z_setup_compile�append�extendr}r|�KeyErrorrrr\�
_c_extensions�_cpp_extensionsr��spawnrwrrr��dirnamerxr�r�r/rOru)rg�sourcesr�Zmacros�include_dirsr�
extra_preargs�extra_postargs�dependsZcompile_info�objectsZpp_opts�buildZcompile_optsZadd_cpp_opts�obj�srcr�Z	input_optZ
output_opt�msgZh_dirZrc_dirr�r>Zrc_file�argsr"r"r#�compileBsr
�




�


 zMSVCCompiler.compilec	
Cs�|js|��|�||�\}}|j||d�}|�||�r�|d|g}|rJz,t�d|jd�|��|�	|jg|�Wq�t
y�}zt|��WYd}~q�d}~00nt�d|�dS)N)r��/OUT:�Executing "%s" %s� �skipping %s (up-to-date))rfr��_fix_object_args�library_filename�
_need_linkr	rrmr/r�rr)	rgr�Zoutput_libnamer�r�target_lang�output_filenameZlib_argsr�r"r"r#�create_static_lib�s� zMSVCCompiler.create_static_libc
Cs�|js|��|�||�\}}|�|||�}|\}}}|rL|�dt|��t||||�}|durptj�	||�}|�
||��r�|j||	f}dd�|p�gD�}||||d|g}tj�|d�}|du�rtj�
tj�|��\}}tj�	||�|��}|�d|�|
�r|
|dd�<|�r.|�|�tj�tj�|��}|�|�z,t�d|jd�	|��|�|jg|�Wn.t�y�}zt|��WYd}~n
d}~00nt�d	|�dS)
Nz5I don't know what to do with 'runtime_library_dirs': cSsg|]}d|�qS)z/EXPORT:r")rC�symr"r"r#�
<listcomp>�rLz%MSVCCompiler.link.<locals>.<listcomp>r�rz/IMPLIB:r�r�r�)rfr�r�Z
_fix_lib_args�warn�strrrrr/r�r~r�r�r�r�r�r�r\�mkpathr	rrvr�rr)rgZtarget_descr�r�r��	libraries�library_dirs�runtime_library_dirs�export_symbolsrr�r��
build_tempr�Z
fixed_argsZlib_optsr�Zexport_optsZld_argsZdll_nameZdll_extZimplib_filer�r"r"r#�link�s^�
��
��

��

 zMSVCCompiler.linkcsRttj|jd�}|�||�� }t�j||d�Wd�S1sB0Y|jS)N)�PATH)rX)�dictrr+rt�_fallback_spawn�superr�rD)rgrTrX�fallback��	__class__r"r#r��s.zMSVCCompiler.spawnc
#s�tddi��}z
|VWn2tyJ}zdt|�vr6�WYd}~nd}~00dSt�d�tj�d|��t��	|�|_
Wd�n1s�0YdS)z�
        Discovered in pypa/distutils#15, some tools monkeypatch the compiler,
        so the 'env' kwarg causes a TypeError. Detect this condition and
        restore the legacy, unsafe behavior.
        ZBagr"z!unexpected keyword argument 'env'Nz>Fallback spawn triggered. Please update distutils monkeypatch.z
os.environ)�typerr��warningsr��unittestZmock�patchr�r�rD)rgrTrXZbagrWr�r"r#r�s
�zMSVCCompiler._fallback_spawncCsd|S)Nz	/LIBPATH:r"�rgrr"r"r#�library_dir_optionszMSVCCompiler.library_dir_optioncCstd��dS)Nz:don't know how to set runtime library search path for MSVC)rr�r"r"r#�runtime_library_dir_options�z'MSVCCompiler.runtime_library_dir_optioncCs
|�|�S)N)r�)rgrmr"r"r#�library_option szMSVCCompiler.library_optioncCs\|r|d|g}n|g}|D]:}|D]0}tj�||�|��}tj�|�r$|Sq$qdS)NZ_d)rrr/r�r<)rg�dirsrmrZ	try_namesr�nameZlibfiler"r"r#�find_library_file#szMSVCCompiler.find_library_file)rrr)N)rrk)NNNrNNN)NrN)
NNNNNrNNNN)r)!�__name__�
__module__�__qualname__�__doc__�
compiler_typeZexecutablesr�r�r�r�r�r�r�Zstatic_lib_extension�shared_lib_extensionZstatic_lib_formatZshared_lib_formatZ
exe_extensionrdr�r�r�r�r�r��
contextlib�contextmanagerr�r�r�r�r��
__classcell__r"r"r�r#rb�s`��

P�
"�
]�
�
Erb)N)!r�rr-r�r�Z
unittest.mockr��suppress�ImportErrorr�distutils.errorsrrrrr�distutils.ccompilerrr�	distutilsr	�distutils.utilr
�	itertoolsrr$r4ZPLAT_SPEC_TO_RUNTIMEr@rYr`rrrbr"r"r"r#�<module>s8&!�
�site-packages/setuptools/_distutils/__pycache__/bcppcompiler.cpython-39.pyc000064400000014714150564745010023226 0ustar00a

(�Re.:�@spdZddlZddlmZmZmZmZmZddlm	Z	m
Z
ddlmZddl
mZddlmZGdd	�d	e	�ZdS)
z�distutils.bcppcompiler

Contains BorlandCCompiler, an implementation of the abstract CCompiler class
for the Borland C++ compiler.
�N)�DistutilsExecError�CompileError�LibError�	LinkError�UnknownFileError)�	CCompiler�gen_preprocess_options)�
write_file)�newer)�logc
@s�eZdZdZdZiZdgZgd�ZeeZdZ	dZ
dZdZZ
d	Zddd�Zddd�Zddd�Zddd�Zddd�Zd dd�Zd!dd�Zd
S)"�BCPPCompilerzConcrete class that implements an interface to the Borland C/C++
    compiler, as defined by the CCompiler abstract class.
    Zbcppz.c)z.ccz.cppz.cxxz.objz.libz.dllz%s%sz.exercCsnt�||||�d|_d|_d|_d|_gd�|_gd�|_gd�|_gd�|_	g|_
gd�|_gd�|_dS)	Nz	bcc32.exezilink32.exeztlib.exe)�/tWMz/O2�/q�/g0)r
z/Odrr)z/Tpd�/Gnr�/x)rrr)rrrz/r)
r�__init__�cc�linker�libZpreprocess_options�compile_options�compile_options_debug�ldflags_shared�ldflags_shared_debugZldflags_static�ldflags_exe�ldflags_exe_debug)�self�verbose�dry_run�force�r ��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/bcppcompiler.pyr5s




zBCPPCompiler.__init__Nc	Cs�|�||||||�\}}	}}
}|p$g}|�d�|rB|�|j�n|�|j�|	D�]>}
z||
\}}Wnty~YqRYn0tj�|�}tj�|
�}
|�	tj�
|
��|dkr�qR|dk�rz|�dd|
|g�WqRt�y}zt
|��WYd}~qRd}~00qR||jv�rd}n||jv�r*d}nd}d|
}z,|�|jg||
||g||g�WqRt�y�}zt
|��WYd}~qRd}~00qR|	S)	Nz-c�.res�.rcZbrcc32z-fo�z-P�-o)Z_setup_compile�append�extendrr�KeyError�os�path�normpath�mkpath�dirname�spawnrr�
_c_extensions�_cpp_extensionsr)r�sources�
output_dir�macros�include_dirs�debug�
extra_preargs�extra_postargs�depends�objects�pp_opts�buildZcompile_opts�obj�src�ext�msgZ	input_optZ
output_optr r r!�compileQsT��



���
 zBCPPCompiler.compilec	
Cs�|�||�\}}|j||d�}|�||�r�|dg|}|r:z|�|jg|�Wq�ty|}zt|��WYd}~q�d}~00nt�d|�dS)N)r2z/u�skipping %s (up-to-date))	�_fix_object_args�library_filename�
_need_linkr.rrrrr5)	rr9Zoutput_libnamer2r5�target_lang�output_filenameZlib_argsr?r r r!�create_static_lib�s� zBCPPCompiler.create_static_libc 
Cs�|�||�\}}|�|||�\}}}|r8t�dt|��|durNtj�||�}|�||��r�|t	j
kr�d}|	r~|jdd�}q�|jdd�}n&d}|	r�|j
dd�}n|jdd�}|dur�d}n�tj�|�\}}tj�|�\}}tj�|d�}tj�|d|�}dg}|�pgD]}|�d||f��q|�t||fd	|�ttjj|�}|g}g}|D]>}tj�tj�|��\}}|d
k�r�|�|�n
|�|��q`|D]}|�dtj�|���q�|�d�|�|�|�d
|g�|�d�|D]4}|�|||	�}|du�r|�|�n
|�|��q�|�d�|�d�|�d
|g�|�d
�|�|�|
�rp|
|dd�<|�r�|�|�|�tj�|��z|�|jg|�Wn.t�y�}zt|��WYd}~n
d}~00nt�d|�dS)Nz7I don't know what to do with 'runtime_library_dirs': %sZc0w32Zc0d32r$rz%s.defZEXPORTSz  %s=_%sz
writing %sr"z/L%sz/L.�,z,,Zimport32Zcw32mtrA) rBZ
_fix_lib_argsr�warn�strr)r*�joinrDrZ
EXECUTABLErrrr�split�splitextr-r&�executer	�mapr+�normcaser'�find_library_filer,r.rrrr5) rZtarget_descr9rFr2�	libraries�library_dirs�runtime_library_dirs�export_symbolsr5r6r7�
build_temprEZstartup_objZld_argsZdef_file�head�tail�modnamer>�temp_dir�contents�symZobjects2�	resources�file�base�lr�libfiler?r r r!�link�s���
�










 zBCPPCompiler.linkc	Csr|r"|d}|d|d||f}n|d|f}|D]:}|D]0}tj�||�|��}tj�|�r:|Sq:q2dS)NZ_dZ_bcpp)r)r*rKrC�exists)	r�dirsrr5ZdlibZ	try_names�dir�namerar r r!rQ4s
zBCPPCompiler.find_library_filer$cCs�|durd}g}|D]�}tj�tj�|��\}}||jddgvrRtd||f��|rbtj�|�}|dkr�|�tj�|||��q|dkr�|�tj�||d��q|�tj�|||j	��q|S)Nr$r#r"z"unknown file type '%s' (from '%s'))
r)r*rMrP�src_extensionsr�basenamer&rK�
obj_extension)rZsource_filenamesZ	strip_dirr2Z	obj_names�src_namer_r>r r r!�object_filenamesNs$��zBCPPCompiler.object_filenamesc
Cs�|�d||�\}}}t||�}dg|}	|dur>|	�d|�|rN||	dd�<|r\|	�|�|	�|�|js~|dus~t||�r�|r�|�tj�	|��z|�
|	�Wn4ty�}
zt|
�t
|
��WYd}
~
n
d}
~
00dS)Nz	cpp32.exer%r)Z_fix_compile_argsrr&r'rr
r,r)r*r-r.r�printr)r�sourceZoutput_filer3r4r6r7�_r:Zpp_argsr?r r r!�
preprocessis&	�



zBCPPCompiler.preprocess)rrr)NNNrNNN)NrN)
NNNNNrNNNN)r)rr$)NNNNN)�__name__�
__module__�__qualname__�__doc__�
compiler_typeZexecutablesr/r0rgriZstatic_lib_extension�shared_lib_extensionZstatic_lib_formatZshared_lib_formatZ
exe_extensionrr@rGrbrQrkror r r r!rsX�
�
D�
�


�
�r)rsr)�distutils.errorsrrrrr�distutils.ccompilerrr�distutils.file_utilr	�distutils.dep_utilr
�	distutilsrrr r r r!�<module>ssite-packages/setuptools/_distutils/__pycache__/log.cpython-39.pyc000064400000004565150564745010021333 0ustar00a

(�Re��@sldZdZdZdZdZdZddlZGdd	�d	�Ze�Zej	Z	ej
Z
ejZejZej
Z
ejZd
d�Zdd
�ZdS)z,A simple log mechanism styled after PEP 282.������Nc@sPeZdZefdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�ZdS)�LogcCs
||_dS�N)�	threshold)�selfr	�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/log.py�__init__szLog.__init__cCs�|tttttfvr"tdt|���||jkr�|r8||}|tttfvrNtj	}ntj
}z|�d|�Wn8ty�|j
}|�|d��|�}|�d|�Yn0|��dS)Nz%s wrong log levelz%s
�backslashreplace)�DEBUG�INFO�WARN�ERROR�FATAL�
ValueError�strr	�sys�stderr�stdout�write�UnicodeEncodeError�encoding�encode�decode�flush)r
�level�msg�args�streamrrrr�_logs
zLog._logcGs|�|||�dSr)r#)r
rr r!rrr�log'szLog.logcGs|�t||�dSr)r#r�r
r r!rrr�debug*sz	Log.debugcGs|�t||�dSr)r#rr%rrr�info-szLog.infocGs|�t||�dSr)r#rr%rrr�warn0szLog.warncGs|�t||�dSr)r#rr%rrr�error3sz	Log.errorcGs|�t||�dSr)r#rr%rrr�fatal6sz	Log.fatalN)�__name__�
__module__�__qualname__rr
r#r$r&r'r(r)r*rrrrrsrcCstj}|t_|Sr)�_global_logr	)r�oldrrr�
set_thresholdAsr0cCs8|dkrtt�n"|dkr$tt�n|dkr4tt�dS)Nrrr)r0rrr)�vrrr�
set_verbosityGs

r2)�__doc__rrrrrrrr.r$r&r'r(r)r*r0r2rrrr�<module>s +site-packages/setuptools/_distutils/__pycache__/cmd.cpython-39.pyc000064400000033320150564745010021304 0ustar00a

(�Re�F�@sbdZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
mZddlmZGdd�d�Z
dS)ztdistutils.cmd

Provides the Command class, the base class for the command classes
in the distutils.command package.
�N)�DistutilsOptionError)�util�dir_util�	file_util�archive_util�dep_util��logc@s"eZdZdZgZdd�Zdd�Zdd�Zdd	�Zd
d�Z	dCdd�Z
dd�ZdDdd�Zdd�Z
dEdd�ZdFdd�Zdd�ZdGdd�Zdd �Zd!d"�Zd#d$�Zd%d&�ZdHd'd(�ZdId*d+�Zd,d-�Zd.d/�Zd0d1�ZdJd2d3�ZdKd5d6�ZdLd7d8�ZdMd9d:�ZdNd;d<�ZdOd=d>�Z dPd?d@�Z!dQdAdB�Z"dS)R�Commanda}Abstract base class for defining command classes, the "worker bees"
    of the Distutils.  A useful analogy for command classes is to think of
    them as subroutines with local variables called "options".  The options
    are "declared" in 'initialize_options()' and "defined" (given their
    final values, aka "finalized") in 'finalize_options()', both of which
    must be defined by every command class.  The distinction between the
    two is necessary because option values might come from the outside
    world (command line, config file, ...), and any options dependent on
    other options must be computed *after* these outside influences have
    been processed -- hence 'finalize_options()'.  The "body" of the
    subroutine, where it does all its work based on the values of its
    options, is the 'run()' method, which must also be implemented by every
    command class.
    cCsbddlm}t||�std��|jtur0td��||_|��d|_	|j
|_
d|_d|_d|_
dS)z�Create and initialize a new Command object.  Most importantly,
        invokes the 'initialize_options()' method, which is the real
        initializer and depends on the actual command being
        instantiated.
        r)�Distributionz$dist must be a Distribution instancezCommand is an abstract classN)�distutils.distr�
isinstance�	TypeError�	__class__r
�RuntimeError�distribution�initialize_options�_dry_run�verbose�force�help�	finalized)�self�distr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/cmd.py�__init__/s


zCommand.__init__cCs<|dkr0t|d|�}|dur*t|j|�S|Snt|��dS)N�dry_run�_)�getattrr�AttributeError)r�attr�myvalrrr�__getattr___szCommand.__getattr__cCs|js|��d|_dS)N�)r�finalize_options�rrrr�ensure_finalizediszCommand.ensure_finalizedcCstd|j��dS)a�Set default values for all the options that this command
        supports.  Note that these defaults may be overridden by other
        commands, by the setup script, by config files, or by the
        command-line.  Thus, this is not the place to code dependencies
        between options; generally, 'initialize_options()' implementations
        are just a bunch of "self.foo = None" assignments.

        This method must be implemented by all command classes.
        �,abstract method -- subclass %s must overrideN�rrr&rrrr{s
�zCommand.initialize_optionscCstd|j��dS)aSet final values for all the options that this command supports.
        This is always called as late as possible, ie.  after any option
        assignments from the command-line or from other commands have been
        done.  Thus, this is the place to code option dependencies: if
        'foo' depends on 'bar', then it is safe to set 'foo' from 'bar' as
        long as 'foo' still has the same value it was assigned in
        'initialize_options()'.

        This method must be implemented by all command classes.
        r(Nr)r&rrrr%�s�zCommand.finalize_optionsN�cCs�ddlm}|dur d|��}|j||tjd�|d}|jD]R\}}}|�|�}|ddkrn|dd�}t||�}|j|d||ftjd�qBdS)	Nr)�
longopt_xlatezcommand options for '%s':)�levelz  ����=z%s = %s)	�distutils.fancy_getoptr+�get_command_name�announcer	�INFO�user_options�	translater)r�header�indentr+�optionr�valuerrr�dump_options�s

�zCommand.dump_optionscCstd|j��dS)a�A command's raison d'etre: carry out the action it exists to
        perform, controlled by the options initialized in
        'initialize_options()', customized by other commands, the setup
        script, the command-line, and config files, and finalized in
        'finalize_options()'.  All terminal output and filesystem
        interaction should be done by 'run()'.

        This method must be implemented by all command classes.
        r(Nr)r&rrr�run�s
�zCommand.runr$cCst�||�dS)zmIf the current verbosity level is of greater than or equal to
        'level' print 'msg' to stdout.
        Nr)r�msgr,rrrr1�szCommand.announcecCs&ddlm}|r"t|�tj��dS)z~Print 'msg' to stdout if the global DEBUG (taken from the
        DISTUTILS_DEBUG environment variable) flag is true.
        r)�DEBUGN)�distutils.debugr<�print�sys�stdout�flush)rr;r<rrr�debug_print�szCommand.debug_printcCsBt||�}|dur"t|||�|St|t�s>td|||f��|S)Nz'%s' must be a %s (got `%s`))r�setattrr
�strr)rr7�what�default�valrrr�_ensure_stringlike�s

�zCommand._ensure_stringlikecCs|�|d|�dS)zWEnsure that 'option' is a string; if not defined, set it to
        'default'.
        �stringN)rH)rr7rFrrr�
ensure_string�szCommand.ensure_stringcCspt||�}|durdSt|t�r6t||t�d|��n6t|t�rTtdd�|D��}nd}|sltd||f��dS)z�Ensure that 'option' is a list of strings.  If 'option' is
        currently a string, we split it either on /,\s*/ or /\s+/, so
        "foo bar baz", "foo,bar,baz", and "foo,   bar baz" all become
        ["foo", "bar", "baz"].
        Nz,\s*|\s+css|]}t|t�VqdS�N)r
rD)�.0�vrrr�	<genexpr>��z-Command.ensure_string_list.<locals>.<genexpr>Fz''%s' must be a list of strings (got %r))	rr
rDrC�re�split�list�allr)rr7rG�okrrr�ensure_string_list�s


��zCommand.ensure_string_listcCs6|�|||�}|dur2||�s2td|||f��dS)Nzerror in '%s' option: )rHr)rr7�testerrE�	error_fmtrFrGrrr�_ensure_tested_string�s
�zCommand._ensure_tested_stringcCs|�|tjjdd�dS)z5Ensure that 'option' is the name of an existing file.�filenamez$'%s' does not exist or is not a fileN)rX�os�path�isfile�rr7rrr�ensure_filename�s�zCommand.ensure_filenamecCs|�|tjjdd�dS)Nzdirectory namez)'%s' does not exist or is not a directory)rXrZr[�isdirr]rrr�ensure_dirnames�zCommand.ensure_dirnamecCst|d�r|jS|jjSdS)N�command_name)�hasattrrar�__name__r&rrrr0	s
zCommand.get_command_namecGsF|j�|�}|��|D](\}}t||�durt||t||��qdS)a>Set the values of any "undefined" options from corresponding
        option values in some other command object.  "Undefined" here means
        "is None", which is the convention used to indicate that an option
        has not been changed between 'initialize_options()' and
        'finalize_options()'.  Usually called from 'finalize_options()' for
        options that depend on some other command rather than another
        option of the same command.  'src_cmd' is the other command from
        which option values will be taken (a command object will be created
        for it if necessary); the remaining arguments are
        '(src_option,dst_option)' tuples which mean "take the value of
        'src_option' in the 'src_cmd' command object, and copy it to
        'dst_option' in the current command object".
        N)r�get_command_objr'rrC)r�src_cmd�option_pairs�src_cmd_obj�
src_option�
dst_optionrrr�set_undefined_optionss
zCommand.set_undefined_optionscCs|j�||�}|��|S)z�Wrapper around Distribution's 'get_command_obj()' method: find
        (create if necessary and 'create' is true) the command object for
        'command', call its 'ensure_finalized()' method, and return the
        finalized command object.
        )rrdr')r�command�create�cmd_objrrr�get_finalized_command$szCommand.get_finalized_commandrcCs|j�||�SrK)r�reinitialize_command)rrk�reinit_subcommandsrrrro0s�zCommand.reinitialize_commandcCs|j�|�dS)z�Run some other command: uses the 'run_command()' method of
        Distribution, which creates and finalizes the command object if
        necessary and then invokes its 'run()' method.
        N)r�run_command)rrkrrrrq4szCommand.run_commandcCs2g}|jD]"\}}|dus"||�r
|�|�q
|S)akDetermine the sub-commands that are relevant in the current
        distribution (ie., that need to be run).  This is based on the
        'sub_commands' class attribute: each tuple in that list may include
        a method that we call to determine if the subcommand needs to be
        run for the current distribution.  Return a list of command names.
        N)�sub_commands�append)r�commands�cmd_name�methodrrr�get_sub_commands;s
zCommand.get_sub_commandscCst�d|��|�dS)Nzwarning: %s: %s
)r	�warnr0)rr;rrrrxKszCommand.warncCstj||||jd�dS�N�r)r�executer)r�func�argsr;r,rrrr{NszCommand.execute�cCstj|||jd�dSry)r�mkpathr)r�name�moderrrrQszCommand.mkpathc	Cstj|||||j||jd�S)z�Copy a file respecting verbose, dry-run and force flags.  (The
        former two default to whatever is in the Distribution object, and
        the latter defaults to false for commands that don't define it.)rz)r�	copy_filerr)r�infile�outfile�
preserve_mode�preserve_times�linkr,rrrr�Ts

�zCommand.copy_filec	Cstj||||||j|jd�S)z\Copy an entire directory tree respecting verbose, dry-run,
        and force flags.
        rz)r�	copy_treerr)rr�r�r�r��preserve_symlinksr,rrrr�]s

�zCommand.copy_treecCstj|||jd�S)z$Move a file respecting dry-run flag.rz)r�	move_filer)r�src�dstr,rrrr�fszCommand.move_filecCs ddlm}||||jd�dS)z2Spawn an external command respecting dry-run flag.r)�spawnrzN)�distutils.spawnr�r)r�cmd�search_pathr,r�rrrr�jsz
Command.spawnc	Cstj|||||j||d�S)N)r�owner�group)r�make_archiver)r�	base_name�format�root_dir�base_dirr�r�rrrr�os�zCommand.make_archivecCs�|durd|}t|t�r"|f}nt|ttf�s8td��|durRd|d�|�f}|jsdt�||�rv|�	||||�n
t
�|�dS)a�Special case of 'execute()' for operations that process one or
        more input files and generate one output file.  Works just like
        'execute()', except the operation is skipped and a different
        message printed if 'outfile' already exists and is newer than all
        files listed in 'infiles'.  If the command defined 'self.force',
        and it is true, then the command is unconditionally run -- does no
        timestamp checks.
        Nzskipping %s (inputs unchanged)z9'infiles' must be a string, or a list or tuple of stringszgenerating %s from %sz, )r
rDrR�tupler�joinrr�newer_groupr{r	�debug)r�infilesr�r|r}�exec_msg�skip_msgr,rrr�	make_fileus

�zCommand.make_file)Nr*)r$)N)N)N)r$)r)Nr$)r~)r$r$Nr$)r$r$rr$)r$)r$r$)NNNN)NNr$)#rc�
__module__�__qualname__�__doc__rrrr#r'rr%r9r:r1rBrHrJrUrXr^r`r0rjrnrorqrwrxr{rr�r�r�r�r�r�rrrrr
sP0






�




�
	�
	

�
�r
)r�r?rZrP�distutils.errorsr�	distutilsrrrrrr	r
rrrr�<module>s
site-packages/setuptools/_distutils/__pycache__/config.cpython-39.pyc000064400000007063150564745010022013 0ustar00a

(�Re��@s<dZddlZddlmZddlmZdZGdd�de�ZdS)z�distutils.pypirc

Provides the PyPIRCCommand class, the base class for the command classes
that uses .pypirc in the distutils.command package.
�N)�RawConfigParser)�CommandzE[distutils]
index-servers =
    pypi

[pypi]
username:%s
password:%s
c@sheZdZdZdZdZdZdZdddefdgZd	gZ	d
d�Z
dd
�Zdd�Zdd�Z
dd�Zdd�ZdS)�
PyPIRCCommandz;Base command that knows how to handle the .pypirc file
    zhttps://upload.pypi.org/legacy/�pypiNzrepository=�rzurl of repository [default: %s])�
show-responseNz&display full response text from serverrcCstj�tj�d�d�S)zReturns rc file path.�~z.pypirc)�os�path�join�
expanduser��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/config.py�_get_rc_file&szPyPIRCCommand._get_rc_filecCs\|��}t�t�|tjtjBd�d��"}|�t||f�Wd�n1sN0YdS)zCreates a default .pypirc file.i��wN)rr	�fdopen�open�O_CREAT�O_WRONLY�write�DEFAULT_PYPIRC)r�username�password�rc�frrr�
_store_pypirc*s zPyPIRCCommand._store_pypirccCs�|��}tj�|��r�|�d|�|jp.|j}t�}|�|�|�	�}d|v�rF|�
dd�}dd�|�d�D�}|gkr�d|vr�dg}niS|D]�}d|i}|�
|d	�|d	<d
|jfd|jfdfD].\}	}
|�
||	�r�|�
||	�||	<q�|
||	<q�|dk�r ||jdfv�r |j|d
<|S|d|k�s:|d
|kr�|Sq�nRd
|v�r�d
}|�
|d
��rp|�
|d
�}n|j}|�
|d	�|�
|d�|||jd�SiS)zReads the .pypirc file.zUsing PyPI login from %s�	distutilsz
index-serverscSs g|]}|��dkr|���qS)�)�strip)�.0�serverrrr�
<listcomp>=s�z.PyPIRCCommand._read_pypirc.<locals>.<listcomp>�
rr"r�
repository�realm)rNzserver-loginr)rrr%r"r&)rr	r
�exists�announcer%�DEFAULT_REPOSITORYr�read�sections�get�split�
DEFAULT_REALM�
has_option)rrr%�configr+�
index_servers�_serversr"�current�key�defaultrrr�_read_pypirc0sb

���

�

�


�zPyPIRCCommand._read_pypirccCs8ddl}|�dd�}|�|�d�dd�}|���|�S)z%Read and decode a PyPI HTTP response.rNzcontent-typez
text/plain��charset�ascii)�cgi�	getheader�parse_headerr,r*�decode)r�responser:�content_type�encodingrrr�_read_pypi_responsepsz!PyPIRCCommand._read_pypi_responsecCsd|_d|_d|_dS)zInitialize options.Nr)r%r&�
show_responser
rrr�initialize_optionswsz PyPIRCCommand.initialize_optionscCs(|jdur|j|_|jdur$|j|_dS)zFinalizes options.N)r%r)r&r.r
rrr�finalize_options}s

zPyPIRCCommand.finalize_options)�__name__�
__module__�__qualname__�__doc__r)r.r%r&�user_options�boolean_optionsrrr6rArCrDrrrrrs&���@r)rHr	�configparserr�
distutils.cmdrrrrrrr�<module>s

site-packages/setuptools/_distutils/__pycache__/__init__.cpython-39.pyc000064400000000772150564745010022305 0ustar00a

(�Re��@s*dZddlZejdej�d��ZdZdS)z�distutils

The main package for the Python Module Distribution Utilities.  Normally
used from a setup script as

   from distutils.core import setup

   setup (...)
�N� T)�__doc__�sys�version�index�__version__�local�r	r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/__init__.py�<module>s
site-packages/setuptools/_distutils/__pycache__/ccompiler.cpython-39.pyc000064400000101300150564745010022510 0ustar00a

(�Re���@s�dZddlZddlZddlZddlTddlmZddlmZddl	m
Z
ddlmZddl
mZmZdd	lmZGd
d�d�ZdZdd
d�Zdddddd�Zdd�Zddd�Zdd�Zdd�ZdS)z�distutils.ccompiler

Contains CCompiler, an abstract base class that defines the interface
for the Distutils compiler abstraction model.�N)�*)�spawn)�	move_file)�mkpath)�newer_group)�split_quoted�execute)�logc
@seZdZdZdZdZdZdZdZdZ	dZ
dZdddddd�Zgd�Z
drd	d
�Zdd�Zd
d�Zdd�Zdd�Zdsdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�Zd)d*�Zd+d,�Zd-d.�Z d/d0�Z!dtd1d2�Z"d3d4�Z#d5d6�Z$d7d8�Z%d9d:�Z&dud;d<�Z'dvd=d>�Z(d?d@�Z)dwdAdB�Z*dCZ+dDZ,dEZ-dxdFdG�Z.dydHdI�Z/dzdJdK�Z0d{dLdM�Z1dNdO�Z2dPdQ�Z3dRdS�Z4d|dTdU�Z5d}dVdW�Z6d~dYdZ�Z7dd[d\�Z8d�d]d^�Z9d�d`da�Z:d�dcdd�Z;dedf�Z<dgdh�Z=d�didj�Z>dkdl�Z?dmdn�Z@d�dpdq�ZAdS)��	CCompilera�Abstract base class to define the interface that must be implemented
    by real compiler classes.  Also has some utility methods used by
    several compiler classes.

    The basic idea behind a compiler abstraction class is that each
    instance can be used for all the compile/link steps in building a
    single project.  Thus, attributes common to all of those compile and
    link steps -- include directories, macros to define, libraries to link
    against, etc. -- are attributes of the compiler instance.  To allow for
    variability in how individual files are treated, most of those
    attributes may be varied on a per-compilation or per-link basis.
    N�c�c++�objc)�.cz.ccz.cppz.cxxz.m)rr
rrcCsb||_||_||_d|_g|_g|_g|_g|_g|_g|_	|j
��D]}|�||j
|�qFdS�N)
�dry_run�force�verbose�
output_dir�macros�include_dirs�	libraries�library_dirs�runtime_library_dirs�objects�executables�keys�set_executable)�selfrrr�key�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/ccompiler.py�__init__UszCCompiler.__init__cKs<|D]2}||jvr&td||jjf��|�|||�qdS)a�Define the executables (and options for them) that will be run
        to perform the various stages of compilation.  The exact set of
        executables that may be specified here depends on the compiler
        class (via the 'executables' class attribute), but most will have:
          compiler      the C/C++ compiler
          linker_so     linker used to create shared objects and libraries
          linker_exe    linker used to create binary executables
          archiver      static library creator

        On platforms with a command-line (Unix, DOS/Windows), each of these
        is a string that will be split into executable name and (optional)
        list of arguments.  (Splitting the string is done similarly to how
        Unix shells operate: words are delimited by spaces, but quotes and
        backslashes can override this.  See
        'distutils.util.split_quoted()'.)
        z$unknown executable '%s' for class %sN)r�
ValueError�	__class__�__name__r)r�kwargsrrrr �set_executablesys

�zCCompiler.set_executablescCs,t|t�rt||t|��nt|||�dSr)�
isinstance�str�setattrr)rr�valuerrr r�s
zCCompiler.set_executablecCs0d}|jD] }|d|kr"|S|d7}q
dS)Nr�)r)r�name�i�defnrrr �_find_macro�s

zCCompiler._find_macrocCs`|D]V}t|t�rFt|�dvrFt|dt�s8|ddurFt|dt�std|dd��qdS)z�Ensures that every element of 'definitions' is a valid macro
        definition, ie. either (name,value) 2-tuple or a (name,) tuple.  Do
        nothing if all definitions are OK, raise TypeError otherwise.
        )r+�r+Nrzinvalid macro definition '%s': z.must be tuple (string,), (string, string), or z(string, None))r'�tuple�lenr(�	TypeError)rZdefinitionsr.rrr �_check_macro_definitions�s

��
����z"CCompiler._check_macro_definitionscCs.|�|�}|dur|j|=|j�||f�dS)a_Define a preprocessor macro for all compilations driven by this
        compiler object.  The optional parameter 'value' should be a
        string; if it is not supplied, then the macro will be defined
        without an explicit value and the exact outcome depends on the
        compiler used (XXX true? does ANSI say anything about this?)
        N�r/r�append)rr,r*r-rrr �define_macro�s	
zCCompiler.define_macrocCs0|�|�}|dur|j|=|f}|j�|�dS)a�Undefine a preprocessor macro for all compilations driven by
        this compiler object.  If the same macro is defined by
        'define_macro()' and undefined by 'undefine_macro()' the last call
        takes precedence (including multiple redefinitions or
        undefinitions).  If the macro is redefined/undefined on a
        per-compilation basis (ie. in the call to 'compile()'), then that
        takes precedence.
        Nr5)rr,r-Zundefnrrr �undefine_macro�s

zCCompiler.undefine_macrocCs|j�|�dS)z�Add 'dir' to the list of directories that will be searched for
        header files.  The compiler is instructed to search directories in
        the order in which they are supplied by successive calls to
        'add_include_dir()'.
        N)rr6�r�dirrrr �add_include_dir�szCCompiler.add_include_dircCs|dd�|_dS)aySet the list of directories that will be searched to 'dirs' (a
        list of strings).  Overrides any preceding calls to
        'add_include_dir()'; subsequence calls to 'add_include_dir()' add
        to the list passed to 'set_include_dirs()'.  This does not affect
        any list of standard include directories that the compiler may
        search by default.
        N�r�r�dirsrrr �set_include_dirs�szCCompiler.set_include_dirscCs|j�|�dS)a�Add 'libname' to the list of libraries that will be included in
        all links driven by this compiler object.  Note that 'libname'
        should *not* be the name of a file containing a library, but the
        name of the library itself: the actual filename will be inferred by
        the linker, the compiler, or the compiler class (depending on the
        platform).

        The linker will be instructed to link against libraries in the
        order they were supplied to 'add_library()' and/or
        'set_libraries()'.  It is perfectly valid to duplicate library
        names; the linker will be instructed to link against libraries as
        many times as they are mentioned.
        N)rr6)r�libnamerrr �add_library�szCCompiler.add_librarycCs|dd�|_dS)z�Set the list of libraries to be included in all links driven by
        this compiler object to 'libnames' (a list of strings).  This does
        not affect any standard system libraries that the linker may
        include by default.
        N)r)rZlibnamesrrr �
set_libraries�szCCompiler.set_librariescCs|j�|�dS)a'Add 'dir' to the list of directories that will be searched for
        libraries specified to 'add_library()' and 'set_libraries()'.  The
        linker will be instructed to search for libraries in the order they
        are supplied to 'add_library_dir()' and/or 'set_library_dirs()'.
        N)rr6r9rrr �add_library_dirszCCompiler.add_library_dircCs|dd�|_dS)z�Set the list of library search directories to 'dirs' (a list of
        strings).  This does not affect any standard library search path
        that the linker may search by default.
        N)rr=rrr �set_library_dirsszCCompiler.set_library_dirscCs|j�|�dS)zlAdd 'dir' to the list of directories that will be searched for
        shared libraries at runtime.
        N)rr6r9rrr �add_runtime_library_dirsz!CCompiler.add_runtime_library_dircCs|dd�|_dS)z�Set the list of directories to search for shared libraries at
        runtime to 'dirs' (a list of strings).  This does not affect any
        standard search path that the runtime linker may search by
        default.
        N)rr=rrr �set_runtime_library_dirssz"CCompiler.set_runtime_library_dirscCs|j�|�dS)z�Add 'object' to the list of object files (or analogues, such as
        explicitly named library files or the output of "resource
        compilers") to be included in every link driven by this compiler
        object.
        N)rr6)r�objectrrr �add_link_object szCCompiler.add_link_objectcCs|dd�|_dS)z�Set the list of object files (or analogues) to be included in
        every link to 'objects'.  This does not affect any standard object
        files that the linker may include by default (such as system
        libraries).
        N)r)rrrrr �set_link_objects(szCCompiler.set_link_objectscCs*|dur|j}nt|t�s"td��|dur2|j}n"t|t�rL||jpFg}ntd��|durd|j}n*t|ttf�r�t|�|jp�g}ntd��|dur�g}|j|d|d�}t	|�t	|�ks�J�t
||�}i}	tt	|��D]B}
||
}||
}tj
�|�d}
|�tj
�|��||
f|	|<q�|||||	fS)z;Process arguments and decide which source files to compile.N�%'output_dir' must be a string or None�/'macros' (if supplied) must be a list of tuples�6'include_dirs' (if supplied) must be a list of stringsr)�	strip_dirrr+)rr'r(r3r�listrr1�object_filenamesr2�gen_preprocess_options�range�os�path�splitextr�dirname)rZoutdirrZincdirs�sources�depends�extrar�pp_opts�buildr-�src�obj�extrrr �_setup_compile6s>

��
zCCompiler._setup_compilecCs0|dg}|rdg|dd�<|r,||dd�<|S)Nz-cz-grr)rrY�debug�before�cc_argsrrr �_get_cc_argsas
zCCompiler._get_cc_argscCs�|dur|j}nt|t�s"td��|dur2|j}n"t|t�rL||jpFg}ntd��|durd|j}n*t|ttf�r�t|�|jp�g}ntd��|||fS)a'Typecheck and fix-up some of the arguments to the 'compile()'
        method, and return fixed-up values.  Specifically: if 'output_dir'
        is None, replaces it with 'self.output_dir'; ensures that 'macros'
        is a list, and augments it with 'self.macros'; ensures that
        'include_dirs' is a list, and augments it with 'self.include_dirs'.
        Guarantees that the returned values are of the correct type,
        i.e. for 'output_dir' either string or None, and for 'macros' and
        'include_dirs' either list or None.
        NrJrKrL)rr'r(r3rrNrr1)rrrrrrr �_fix_compile_argsjs"


�zCCompiler._fix_compile_argscCs*|j||d�}t|�t|�ks"J�|ifS)a,Decide which source files must be recompiled.

        Determine the list of object files corresponding to 'sources',
        and figure out which ones really need to be recompiled.
        Return a list of all object files and a dictionary telling
        which source files can be skipped.
        )r)rOr2)rrVrrWrrrr �
_prep_compile�s	zCCompiler._prep_compilecCsHt|ttf�std��t|�}|dur.|j}nt|t�s@td��||fS)z�Typecheck and fix up some arguments supplied to various methods.
        Specifically: ensure that 'objects' is a list; if output_dir is
        None, replace with self.output_dir.  Return fixed versions of
        'objects' and 'output_dir'.
        z,'objects' must be a list or tuple of stringsNrJ)r'rNr1r3rr()rrrrrr �_fix_object_args�s
zCCompiler._fix_object_argscCs�|dur|j}n*t|ttf�r2t|�|jp,g}ntd��|durJ|j}n*t|ttf�rlt|�|jpfg}ntd��|dur�|j}n*t|ttf�r�t|�|jp�g}ntd��|||fS)a;Typecheck and fix up some of the arguments supplied to the
        'link_*' methods.  Specifically: ensure that all arguments are
        lists, and augment them with their permanent versions
        (eg. 'self.libraries' augments 'libraries').  Return a tuple with
        fixed versions of all arguments.
        Nz3'libraries' (if supplied) must be a list of stringsz6'library_dirs' (if supplied) must be a list of stringsz>'runtime_library_dirs' (if supplied) must be a list of strings)rr'rNr1r3rr)rrrrrrr �
_fix_lib_args�s,���zCCompiler._fix_lib_argscCs2|jr
dS|jr t||dd�}n
t||�}|SdS)zjReturn true if we need to relink the files listed in 'objects'
        to recreate 'output_file'.
        T�newer)�missingN)rrr)rr�output_filergrrr �
_need_link�s
zCCompiler._need_linkc		Cs|t|t�s|g}d}t|j�}|D]T}tj�|�\}}|j�|�}z |j�	|�}||kr`|}|}Wq"t
ytYq"0q"|S)z|Detect the language of a given file, or list of files. Uses
        language_map, and language_order to do the job.
        N)r'rNr2�language_orderrRrSrT�language_map�get�indexr")	rrV�langrn�source�baser]ZextlangZextindexrrr �detect_language�s

zCCompiler.detect_languagecCsdS)a�Preprocess a single C/C++ source file, named in 'source'.
        Output will be written to file named 'output_file', or stdout if
        'output_file' not supplied.  'macros' is a list of macro
        definitions as for 'compile()', which will augment the macros set
        with 'define_macro()' and 'undefine_macro()'.  'include_dirs' is a
        list of directory names that will be added to the default list.

        Raises PreprocessError on failure.
        Nr)rrprirr�
extra_preargs�extra_postargsrrr �
preprocess�szCCompiler.preprocessc		Csv|�||||||�\}}	}}
}|�|
||�}|	D]@}
z||
\}}WntyZYq0Yn0|�|
|||||
�q0|	S)aK	Compile one or more source files.

        'sources' must be a list of filenames, most likely C/C++
        files, but in reality anything that can be handled by a
        particular compiler and compiler class (eg. MSVCCompiler can
        handle resource files in 'sources').  Return a list of object
        filenames, one per source filename in 'sources'.  Depending on
        the implementation, not all source files will necessarily be
        compiled, but all corresponding object filenames will be
        returned.

        If 'output_dir' is given, object files will be put under it, while
        retaining their original path component.  That is, "foo/bar.c"
        normally compiles to "foo/bar.o" (for a Unix implementation); if
        'output_dir' is "build", then it would compile to
        "build/foo/bar.o".

        'macros', if given, must be a list of macro definitions.  A macro
        definition is either a (name, value) 2-tuple or a (name,) 1-tuple.
        The former defines a macro; if the value is None, the macro is
        defined without an explicit value.  The 1-tuple case undefines a
        macro.  Later definitions/redefinitions/ undefinitions take
        precedence.

        'include_dirs', if given, must be a list of strings, the
        directories to add to the default include file search path for this
        compilation only.

        'debug' is a boolean; if true, the compiler will be instructed to
        output debug symbols in (or alongside) the object file(s).

        'extra_preargs' and 'extra_postargs' are implementation- dependent.
        On platforms that have the notion of a command-line (e.g. Unix,
        DOS/Windows), they are most likely lists of strings: extra
        command-line arguments to prepend/append to the compiler command
        line.  On other platforms, consult the implementation class
        documentation.  In any event, they are intended as an escape hatch
        for those occasions when the abstract compiler framework doesn't
        cut the mustard.

        'depends', if given, is a list of filenames that all targets
        depend on.  If a source file is older than any file in
        depends, then the source file will be recompiled.  This
        supports dependency tracking, but only at a coarse
        granularity.

        Raises CompileError on failure.
        )r^rb�KeyError�_compile)rrVrrrr_rsrtrWrrYrZrar\r[r]rrr �compile�s6��
zCCompiler.compilecCsdS)zCompile 'src' to product 'obj'.Nr)rr\r[r]rartrYrrr rwCszCCompiler._compilecCsdS)a&Link a bunch of stuff together to create a static library file.
        The "bunch of stuff" consists of the list of object files supplied
        as 'objects', the extra object files supplied to
        'add_link_object()' and/or 'set_link_objects()', the libraries
        supplied to 'add_library()' and/or 'set_libraries()', and the
        libraries supplied as 'libraries' (if any).

        'output_libname' should be a library name, not a filename; the
        filename will be inferred from the library name.  'output_dir' is
        the directory where the library file will be put.

        'debug' is a boolean; if true, debugging information will be
        included in the library (note that on most platforms, it is the
        compile step where this matters: the 'debug' flag is included here
        just for consistency).

        'target_lang' is the target language for which the given objects
        are being compiled. This allows specific linkage time treatment of
        certain languages.

        Raises LibError on failure.
        Nr)rr�output_libnamerr_�target_langrrr �create_static_libIszCCompiler.create_static_libZ
shared_objectZshared_library�
executablecCst�dS)auLink a bunch of stuff together to create an executable or
        shared library file.

        The "bunch of stuff" consists of the list of object files supplied
        as 'objects'.  'output_filename' should be a filename.  If
        'output_dir' is supplied, 'output_filename' is relative to it
        (i.e. 'output_filename' can provide directory components if
        needed).

        'libraries' is a list of libraries to link against.  These are
        library names, not filenames, since they're translated into
        filenames in a platform-specific way (eg. "foo" becomes "libfoo.a"
        on Unix and "foo.lib" on DOS/Windows).  However, they can include a
        directory component, which means the linker will look in that
        specific directory rather than searching all the normal locations.

        'library_dirs', if supplied, should be a list of directories to
        search for libraries that were specified as bare library names
        (ie. no directory component).  These are on top of the system
        default and those supplied to 'add_library_dir()' and/or
        'set_library_dirs()'.  'runtime_library_dirs' is a list of
        directories that will be embedded into the shared library and used
        to search for other shared libraries that *it* depends on at
        run-time.  (This may only be relevant on Unix.)

        'export_symbols' is a list of symbols that the shared library will
        export.  (This appears to be relevant only on Windows.)

        'debug' is as for 'compile()' and 'create_static_lib()', with the
        slight distinction that it actually matters on most platforms (as
        opposed to 'create_static_lib()', which includes a 'debug' flag
        mostly for form's sake).

        'extra_preargs' and 'extra_postargs' are as for 'compile()' (except
        of course that they supply command-line arguments for the
        particular linker being used).

        'target_lang' is the target language for which the given objects
        are being compiled. This allows specific linkage time treatment of
        certain languages.

        Raises LinkError on failure.
        N��NotImplementedError)rZtarget_descr�output_filenamerrrr�export_symbolsr_rsrt�
build_temprzrrr �linkis9zCCompiler.linkc

Cs2|�tj||j|dd�|||||||	|
||�
dS)N�shared)�lib_type)r�r
�SHARED_LIBRARY�library_filename)
rrryrrrrr�r_rsrtr�rzrrr �link_shared_lib�s
�zCCompiler.link_shared_libc

Cs(|�tj|||||||||	|
||�
dSr)r�r
�
SHARED_OBJECT)
rrrrrrrr�r_rsrtr�rzrrr �link_shared_object�s
�zCCompiler.link_shared_objectcCs.|�tj||�|�||||d|||	d|
�
dSr)r�r
�
EXECUTABLE�executable_filename)rrZoutput_prognamerrrrr_rsrtrzrrr �link_executable�s



�zCCompiler.link_executablecCst�dS)zkReturn the compiler option to add 'dir' to the list of
        directories searched for libraries.
        Nr}r9rrr �library_dir_option�szCCompiler.library_dir_optioncCst�dS)zsReturn the compiler option to add 'dir' to the list of
        directories searched for runtime libraries.
        Nr}r9rrr �runtime_library_dir_option�sz$CCompiler.runtime_library_dir_optioncCst�dS)zReturn the compiler option to add 'lib' to the list of libraries
        linked into the shared library or executable.
        Nr})r�librrr �library_option�szCCompiler.library_optionc

Cstddl}|durg}|dur g}|dur,g}|dur8g}|jd|dd�\}}t�|d�}	z2|D]}
|	�d|
�q^|	�d|�W|	��n
|	��0zDz|j|g|d	�}Wn ty�YWt�|�d
S0Wt�|�nt�|�0znz|j	|d||d�Wn2t
tf�y2YW|D]}t�|��qd
S0t�d�W|D]}t�|��qDn|D]}t�|��q\0dS)
z�Return a boolean indicating whether funcname is supported on
        the current platform.  The optional arguments can be used to
        augment the compilation environment.
        rNrT)�text�wz#include "%s"
z=int main (int argc, char **argv) {
    %s();
    return 0;
}
r<Fza.out)rr)�tempfile�mkstemprR�fdopen�write�closerx�CompileError�remover��	LinkErrorr3)
r�funcnameZincludesrrrr��fd�fname�fZinclr�fnrrr �has_function�sR	�
��
��zCCompiler.has_functioncCst�dS)aHSearch the specified list of directories for a static or shared
        library file 'lib' and return the full path to that file.  If
        'debug' true, look for a debugging version (if that makes sense on
        the current platform).  Return None if 'lib' wasn't found in any of
        the specified directories.
        Nr})rr>r�r_rrr �find_library_file+szCCompiler.find_library_file�cCs�|durd}g}|D]|}tj�|�\}}tj�|�d}|tj�|�d�}||jvrftd||f��|rvtj�|�}|�tj�	|||j
��q|S)Nr�r+z"unknown file type '%s' (from '%s'))rRrSrT�
splitdrive�isabs�src_extensions�UnknownFileError�basenamer6�join�
obj_extension)rZsource_filenamesrMrZ	obj_names�src_namerqr]rrr rOVs"

��zCCompiler.object_filenamescCs0|dusJ�|rtj�|�}tj�|||j�Sr)rRrSr�r��shared_lib_extension�rr�rMrrrr �shared_object_filenamegsz CCompiler.shared_object_filenamecCs4|dusJ�|rtj�|�}tj�|||jp.d�S)Nr�)rRrSr�r��
exe_extensionr�rrr r�mszCCompiler.executable_filename�staticc
Csl|dusJ�|dvrtd��t||d�}t||d�}tj�|�\}}|||f}	|r\d}tj�|||	�S)N)r�r�ZdylibZ
xcode_stubz?'lib_type' must be "static", "shared", "dylib", or "xcode_stub"Z_lib_formatZ_lib_extensionr�)r"�getattrrRrS�splitr�)
rr@r�rMr�fmtr]r:rq�filenamerrr r�ss�zCCompiler.library_filenamer+cCst�|�dSr)r	r_)r�msg�levelrrr �announce�szCCompiler.announcecCsddlm}|rt|�dS)Nr)�DEBUG)�distutils.debugr��print)rr�r�rrr �debug_print�szCCompiler.debug_printcCstj�d|�dS)Nzwarning: %s
)�sys�stderrr�)rr�rrr �warn�szCCompiler.warncCst||||j�dSr)rr)r�func�argsr�r�rrr r�szCCompiler.executecKst|fd|ji|��dS)Nr)rr)r�cmdr%rrr r�szCCompiler.spawncCst|||jd�S�N)r)rr)rr[�dstrrr r�szCCompiler.move_file�cCst|||jd�dSr�)rr)rr,�moderrr r�szCCompiler.mkpath)rrr)N)N)NNNNN)NNNrNNN)NrN)
NNNNNrNNNN)
NNNNNrNNNN)
NNNNNrNNNN)NNNNrNNN)NNNN)r)rr�)rr�)rr�)r�rr�)r+)Nr+)r�)Br$�
__module__�__qualname__�__doc__�
compiler_typer�r�Zstatic_lib_extensionr�Zstatic_lib_formatZshared_lib_formatr�rlrkr!r&rr/r4r7r8r;r?rArBrCrDrErFrHrIr^rbrcrdrerfrjrrrurxrwr{r�r�r�r�r�r�r�r�r�r�r�r�rOr�r�r�r�r�r�rrrrrrrr r
s��
$ 

+	 
"
�

�
D�
�
A�
�
�
�
3
+


�


r
))zcygwin.*�unix)�posixr�)�nt�msvccCsV|durtj}|durtj}tD]0\}}t�||�dusHt�||�dur |Sq dS)akDetermine the default compiler to use for the given platform.

       osname should be one of the standard Python OS names (i.e. the
       ones returned by os.name) and platform the common value
       returned by sys.platform for the platform in question.

       The default values are os.name and sys.platform in case the
       parameters are not given.
    Nr�)rRr,r��platform�_default_compilers�re�match)�osnamer��pattern�compilerrrr �get_default_compiler�s
�
r�)Z
unixccompilerZ
UnixCCompilerzstandard UNIX-style compiler)Z
_msvccompilerZMSVCCompilerzMicrosoft Visual C++)�cygwinccompilerZCygwinCCompilerz'Cygwin port of GNU C Compiler for Win32)r�ZMingw32CCompilerz(Mingw32 port of GNU C Compiler for Win32)ZbcppcompilerZBCPPCompilerzBorland C++ Compiler)r�r��cygwinZmingw32ZbcppcCsXddlm}g}t��D] }|�d|dt|df�q|��||�}|�d�dS)zyPrint list of available compilers (used by the "--help-compiler"
    options to "build", "build_ext", "build_clib").
    r)�FancyGetoptz	compiler=Nr0zList of available compilers:)�distutils.fancy_getoptr��compiler_classrr6�sort�
print_help)r�Z	compilersr�Zpretty_printerrrr �show_compilers�s
�r�cCs�|durtj}z"|dur t|�}t|\}}}Wn6tyfd|}|durZ|d|}t|��Yn0z*d|}t|�tj|}	t	|	�|}
Wn>t
y�td|��Yn"ty�td||f��Yn0|
d||�S)a[Generate an instance of some CCompiler subclass for the supplied
    platform/compiler combination.  'plat' defaults to 'os.name'
    (eg. 'posix', 'nt'), and 'compiler' defaults to the default compiler
    for that platform.  Currently only 'posix' and 'nt' are supported, and
    the default compilers are "traditional Unix interface" (UnixCCompiler
    class) and Visual C++ (MSVCCompiler class).  Note that it's perfectly
    possible to ask for a Unix compiler object under Windows, and a
    Microsoft compiler object under Unix -- if you supply a value for
    'compiler', 'plat' is ignored.
    Nz5don't know how to compile C/C++ code on platform '%s'z with '%s' compilerz
distutils.z4can't compile C/C++ code: unable to load module '%s'zBcan't compile C/C++ code: unable to find class '%s' in module '%s')rRr,r�r�rv�DistutilsPlatformError�
__import__r��modules�vars�ImportError�DistutilsModuleError)�platr�rrr�module_name�
class_name�long_descriptionr��module�klassrrr �new_compiler�s:
����
r�cCs�g}|D]�}t|t�r0dt|�kr.dks<ntd|��t|�dkr\|�d|d�qt|�dkr|ddur�|�d|d�q|�d|�q|D]}|�d	|�q�|S)
aGenerate C pre-processor options (-D, -U, -I) as used by at least
    two types of compilers: the typical Unix compiler and Visual C++.
    'macros' is the usual thing, a list of 1- or 2-tuples, where (name,)
    means undefine (-U) macro 'name', and (name,value) means define (-D)
    macro 'name' to 'value'.  'include_dirs' is just a list of directory
    names to be added to the header file search path (-I).  Returns a list
    of command-line options suitable for either Unix compilers or Visual
    C++.
    r+r0zPbad macro definition '%s': each element of 'macros' list must be a 1- or 2-tuplez-U%srNz-D%sz-D%s=%sz-I%s)r'r1r2r3r6)rrrYZmacror:rrr rPs"$��rPcCs�g}|D]}|�|�|��q|D],}|�|�}t|t�rD||}q"|�|�q"|D]V}tj�|�\}}	|r�|�|g|	�}
|
r�|�|
�q�|�	d|�qT|�|�
|��qT|S)acGenerate linker options for searching library directories and
    linking with specific libraries.  'libraries' and 'library_dirs' are,
    respectively, lists of library names (not filenames!) and search
    directories.  Returns a list of command-line options suitable for use
    with some compiler (depending on the two format strings passed in).
    z6no library file corresponding to '%s' found (skipping))r6r�r�r'rNrRrSr�r�r�r�)r�rrrZlib_optsr:�optr��lib_dirZlib_nameZlib_filerrr �gen_lib_options?s&


�r�)NN)NNrrr)r�r�rRr��distutils.errors�distutils.spawnr�distutils.file_utilr�distutils.dir_utilr�distutils.dep_utilr�distutils.utilrr�	distutilsr	r
r�r�r�r�r�rPr�rrrr �<module>s8
�
--site-packages/setuptools/_distutils/__pycache__/core.cpython-39.pyc000064400000015150150564745010021472 0ustar00a

(�Re�"�@s�dZddlZddlZddlmZddlTddlmZddlm	Z	ddl
mZddlm
Z
d	Zd
d�ZdadadZd
Zdd�Zddd�ZdS)a#distutils.core

The only module that needs to be imported to use the Distutils; provides
the 'setup' function (which is to be called from the setup script).  Also
indirectly provides the Distribution and Command classes, although they are
really defined in distutils.dist and distutils.cmd.
�N)�DEBUG)�*)�Distribution)�Command)�
PyPIRCCommand)�	Extensionz�usage: %(script)s [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...]
   or: %(script)s --help [cmd1 cmd2 ...]
   or: %(script)s --help-commands
   or: %(script)s cmd --help
cCstj�|�}tt�S)N)�os�path�basename�USAGE�vars)�script_name�script�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/core.py�	gen_usage sr)�	distclassr
�script_args�options�name�version�author�author_email�
maintainer�maintainer_email�url�license�description�long_description�keywords�	platforms�classifiers�download_url�requires�provides�	obsoletes)r�sources�include_dirs�
define_macros�undef_macros�library_dirs�	libraries�runtime_library_dirs�
extra_objects�extra_compile_args�extra_link_args�	swig_opts�export_symbols�depends�languagec
Ks|�d�}|r|d=nt}d|vr8tj�tjd�|d<d|vrRtjdd�|d<z||�a}WnNty�}z6d|vr�t	d|��nt	d	|d|f��WYd}~n
d}~00t
d
kr�|S|��tr�t
d�|��t
dkr�|Sz|��}Wn<t�y.}z"t	t|j�d
|��WYd}~n
d}~00t�rFt
d�|��t
dk�rT|S|�rz|��Wn�t�y�t	d��Yn�t�y�}z6t�r�tj�d|f��nt	d|f��WYd}~nLd}~0ttf�y}z&t�r�nt	dt|���WYd}~n
d}~00|S)a�The gateway to the Distutils: do everything your setup script needs
    to do, in a highly flexible and user-driven way.  Briefly: create a
    Distribution instance; find and parse config files; parse the command
    line; run each Distutils command found there, customized by the options
    supplied to 'setup()' (as keyword arguments), in config files, and on
    the command line.

    The Distribution instance might be an instance of a class supplied via
    the 'distclass' keyword argument to 'setup'; if no such class is
    supplied, then the Distribution class (in dist.py) is instantiated.
    All other arguments to 'setup' (except for 'cmdclass') are used to set
    attributes of the Distribution instance.

    The 'cmdclass' argument, if supplied, is a dictionary mapping command
    names to command classes.  Each command encountered on the command line
    will be turned into a command class, which is in turn instantiated; any
    class found in 'cmdclass' is used in place of the default, which is
    (for command 'foo_bar') class 'foo_bar' in module
    'distutils.command.foo_bar'.  The command class must provide a
    'user_options' attribute which is a list of option specifiers for
    'distutils.fancy_getopt'.  Any command-line options between the current
    and the next command are used to set attributes of the current command
    object.

    When the entire command-line has been successfully parsed, calls the
    'run()' method on each command object in turn.  This method will be
    driven entirely by the Distribution object (which each command object
    has a reference to, thanks to its constructor), and the
    command-specific options that became attributes of each command
    object.
    rr
rr�Nrzerror in setup command: %szerror in %s setup command: %s�initz%options (after parsing config files):�configz

error: %sz%options (after parsing command line):�commandline�interruptedz
error: %s
z	error: %szerror: )�getrrr	r
�sys�argv�_setup_distribution�DistutilsSetupError�
SystemExit�_setup_stop_after�parse_config_filesr�print�dump_option_dicts�parse_command_line�DistutilsArgErrorrr
�run_commands�KeyboardInterrupt�OSError�stderr�write�DistutilsError�CCompilerError�str)�attrs�klass�dist�msg�ok�excrrr�setup9sd%

�,
"�&rS�runcCs�|dvrtd|f��|atj��}d|i}zxzf|tjd<|durP|tjdd�<t|d��}t|��|�Wd�n1s~0YW|t_dan|t_da0Wnty�Yn0t	dur�t
d|��t	S)	a.Run a setup script in a somewhat controlled environment, and
    return the Distribution instance that drives things.  This is useful
    if you need to find out the distribution meta-data (passed as
    keyword args from 'script' to 'setup()', or the contents of the
    config files or command-line.

    'script_name' is a file that will be read and run with 'exec()';
    'sys.argv[0]' will be replaced with 'script' for the duration of the
    call.  'script_args' is a list of strings; if supplied,
    'sys.argv[1:]' will be replaced by 'script_args' for the duration of
    the call.

    'stop_after' tells 'setup()' when to stop processing; possible
    values:
      init
        stop after the Distribution instance has been created and
        populated with the keyword arguments to 'setup()'
      config
        stop after config files have been parsed (and their data
        stored in the Distribution instance)
      commandline
        stop after the command-line ('sys.argv[1:]' or 'script_args')
        have been parsed (and the data stored in the Distribution)
      run [default]
        stop after all commands have been run (the same as if 'setup()'
        had been called in the usual way

    Returns the Distribution instance, which provides all information
    used to drive the Distutils.
    )r5r6r7rTz"invalid value for 'stop_after': %r�__file__rNr4�rbzZ'distutils.core.setup()' was never called -- perhaps '%s' is not a Distutils setup script?)�
ValueErrorr?r:r;�copy�open�exec�readr>r<�RuntimeError)r
r�
stop_after�	save_argv�g�frrr�	run_setup�s.

.�
�ra)NrT)�__doc__rr:�distutils.debugr�distutils.errors�distutils.distr�
distutils.cmdr�distutils.configr�distutils.extensionrrrr?r<�setup_keywords�extension_keywordsrSrarrrr�<module>s 	qsite-packages/setuptools/_distutils/__pycache__/sysconfig.cpython-39.pyc000064400000030531150564745010022546 0ustar00a

(�Re~T�@s�dZddlZddlZddlZddlZddlmZdejvZej	�
ej�Zej	�
ej
�Zej	�
ej�Zej	�
ej�Zdejvr�ej	�ejd�Zn&ejr�ej	�ej	�ej��Zne��Zdd�Zeed	d�Zejd
kr�dd�Zee�Zee�Zd
d�Ze�ZdZ ze�sej!Z Wne"�y"Yn0dd�Z#d-dd�Z$d.dd�Z%dd�Z&dd�Z'dd�Z(d/dd�Z)e�*d�Z+e�*d�Z,e�*d �Z-d0d!d"�Z.d#d$�Z/da0d%d&�Z1d'd(�Z2d)d*�Z3d+d,�Z4dS)1a�Provide access to Python's configuration information.  The specific
configuration variables available depend heavily on the platform and
configuration.  The values may be retrieved using
get_config_var(name), and the list of variables is available via
get_config_vars().keys().  Additional convenience functions are also
available.

Written by:   Fred L. Drake, Jr.
Email:        <fdrake@acm.org>
�N�)�DistutilsPlatformErrorZ__pypy__Z_PYTHON_PROJECT_BASEcCs,dD]"}tj�tj�|d|��rdSqdS)N)�SetupzSetup.local�ModulesTF)�os�path�isfile�join)�d�fn�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/sysconfig.py�_is_python_source_dir,sr�_home�ntcCs0|r,tj�|��tj�tj�td���r,tS|S)NZPCbuild)rr�normcase�
startswithr	�PREFIX)r
rrr
�_fix_pcbuild5s
�rcCstrtt�Stt�S)N)�	_sys_homer�project_baserrrr
�
_python_build=sr�cCsdtjdd�S)z�Return a string containing the major and minor Python version,
    leaving off the patchlevel.  Sample return values could be '1.5'
    or '2.2'.
    z%d.%dN�)�sys�version_inforrrr
�get_python_versionQsrcCs�|dur|rtpt}tjdkr�tr:tjdkr:tj�|d�St	rh|rJt
pHtStj�td�d�}tj�
|�Strpdnd}|t�t}tj�|d|�Stjd	kr�t	r�tj�|d�tjjtj�|d
�Stj�|d�Stdtj��dS)a�Return the directory containing installed Python header files.

    If 'plat_specific' is false (the default), this is the path to the
    non-platform-specific header files, i.e. Python.h and so on;
    otherwise, this is the path to platform-specific header files
    (namely pyconfig.h).

    If 'prefix' is supplied, use it instead of sys.base_prefix or
    sys.base_exec_prefix -- i.e., ignore 'plat_specific'.
    N�posix����include�srcdir�Include�pypy�pythonr�PCzFI don't know where Python installs its C header files on platform '%s')�BASE_EXEC_PREFIX�BASE_PREFIXr�name�IS_PYPYrrrr	�python_buildrr�get_config_var�normpathr�build_flags�pathsepr)�
plat_specific�prefix�incdir�implementation�
python_dirrrr
�get_python_incYs0

���r5cCstrBtjdkrB|durt}|r4tj�|dtjd�Stj�|d�S|durh|r\|rVtpXt	}n|rdt
pft}tjdkr�|sz|r�ttdd�}nd}tr�d	nd
}tj�|||t
��}|r�|Stj�|d�Sn<tjdkr�|r�tj�|d�Stj�|dd�Sntd
tj��dS)aSReturn the directory containing the Python library (standard or
    site additions).

    If 'plat_specific' is true, return the directory containing
    platform-specific modules, i.e. any module from a non-pure-Python
    module distribution; otherwise, return the platform-shared library
    directory.  If 'standard_lib' is true, return the directory
    containing standard Python library modules; otherwise, return the
    directory for site-specific modules.

    If 'prefix' is supplied, use it instead of sys.base_prefix or
    sys.base_exec_prefix -- i.e., ignore 'plat_specific'.
    rNz
lib-pythonrz
site-packagesr�
platlibdir�libr$r%r�Libz?I don't know where Python installs its library on platform '%s')r*rrrrrr	�versionr'r(�EXEC_PREFIXr)�getattrrr)r0�standard_libr1�libdirr3�	libpythonrrr
�get_python_lib�s<

�
��r?c	Cs|jdk�rtjdkr8td�s8ddl}|�t�dtd<tddd	d
ddd
d�\}}}}}}}}	dtj	vr�tj	d}
dtj	vr�|�
|�r�|
|t|�d�}|
}dtj	vr�tj	d}dtj	vr�tj	d}dtj	vr�tj	d}n|d}dtj	v�r|dtj	d}d	tj	v�r2|dtj	d	}|dtj	d	}dtj	v�rt|dtj	d}|dtj	d}|dtj	d}d
tj	v�r�tj	d
}dtj	v�r�|dtj	d}n|d|	}|d|}
|j||
|
d|||||d�dtj	v�r|j
�dd��r|jtj	dd�||_dS)z�Do any platform-specific customization of a CCompiler instance.

    Mainly needed on Unix, so we can plug in the information that
    varies across Unices and is stored in Python's Makefile.
    �unix�darwin�CUSTOMIZED_OSX_COMPILERrN�True�CC�CXX�CFLAGS�CCSHARED�LDSHARED�SHLIB_SUFFIX�AR�ARFLAGS�CPPz -E�LDFLAGS� �CPPFLAGS)�preprocessor�compiler�compiler_so�compiler_cxx�	linker_so�
linker_exe�archiverZRANLIB�ranlib)rW)�
compiler_typer�platformr,�_osx_support�customize_compiler�_config_vars�get_config_varsr�environr�len�set_executablesZexecutables�get�shared_lib_extension)rQrZ�cc�cxx�cflags�ccshared�ldshared�shlib_suffix�ar�ar_flags�newcc�cpprV�cc_cmdrrr
r[�sf


��


�






�	r[cCsDtr,tjdkr"tj�tptd�}q6tp(t}n
tdd�}tj�|d�S)z2Return full pathname of installed pyconfig.h file.rr&r�r0z
pyconfig.h)r+rr)rr	rrr5)�inc_dirrrr
�get_config_h_filenames


rpcCs\trtj�tptd�Stddd�}d�t�t	�}t
tjd�rL|dtjj
7}tj�||d�S)zAReturn full pathname of installed Makefile from the Python build.�Makefilerr�r0r<zconfig-{}{}�
_multiarchz-%s)r+rrr	rrr?�formatrr.�hasattrrr3rs)�lib_dir�config_filerrr
�get_makefile_filenamesrxcCs�|duri}t�d�}t�d�}|��}|s.q�|�|�}|rv|�dd�\}}zt|�}WntyjYn0|||<q |�|�}|r d||�d�<q |S)z�Parse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    Nz"#define ([A-Z][A-Za-z0-9_]+) (.*)
z&/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/
rrr)�re�compile�readline�match�group�int�
ValueError)�fp�g�	define_rx�undef_rx�line�m�n�vrrr
�parse_config_hs"




r�z"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)z\$\(([A-Za-z][A-Za-z0-9_]*)\)z\${([A-Za-z][A-Za-z0-9_]*)}c	Cs�ddlm}||ddddd�}|dur*i}i}i}|��}|durDq�t�|�}|r2|�dd�\}}	|	��}	|	�dd	�}
d
|
vr�|	||<q2zt|	�}	Wn"t	y�|	�dd
�||<Yq20|	||<q2d}|�rpt
|�D�]�}||}
t�|
�p�t
�|
�}|�rf|�d�}d}||v�r"t||�}n�||v�r2d
}nx|tjv�rJtj|}n`||v�r�|�d��rx|dd�|v�rxd	}n$d||v�r�d
}nt|d|�}nd	||<}|�rl|
|��d�}|
d|���||}
d
|v�r�|
||<nxzt|
�}
Wn t	�y|
��||<Yn
0|
||<||=|�d��rl|dd�|v�rl|dd�}||v�rl|
||<q�||=q�q�|��|��D]"\}}	t|	t��r�|	��||<�q�|�|�|S)z�Parse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    r)�TextFiler�surrogateescape)�strip_comments�skip_blanks�
join_lines�errorsNrz$$r�$)rFrMrOTF�PY_r)�distutils.text_filer�r{�_variable_rxr|r}�strip�replacer~r�list�_findvar1_rx�search�_findvar2_rx�strrr^r�end�start�close�items�
isinstance�update)rr�r�r��done�notdoner�r�r�r��tmpv�renamed_variablesr)�value�found�item�after�krrr
�parse_makefileBs�








�



r�cCsVt�|�pt�|�}|rR|��\}}|d|�|�|�d��||d�}qqRq|S)a�Expand Makefile-style variables -- "${foo}" or "$(foo)" -- in
    'string' according to 'vars' (a dictionary mapping variable names to
    values).  Variables not present in 'vars' are silently expanded to the
    empty string.  The variable values in 'vars' should not contain further
    variable expansions; if 'vars' is the output of 'parse_makefile()',
    you're fine.  Returns a variable-expanded version of 's'.
    rrN)r�r�r��spanrar})�s�varsr��begr�rrr
�expand_makefile_vars�s*r�c
Cs�tj�ddjtjtjttjdd�d��}zt	|t
�t�dgd�}Wn(tyjt	dt
�t�dgd�}Yn0|j
}iat�|�d	S)
z7Initialize the module as appropriate for POSIX systems.�_PYTHON_SYSCONFIGDATA_NAMEz+_sysconfigdata_{abi}_{platform}_{multiarch}rsr)�abirY�	multiarch�build_time_varsrZ_sysconfigdataN)rr^rartr�abiflagsrYr;r3�
__import__�globals�locals�ImportErrorr�r\r�)r)�_tempr�rrr
�_init_posix�s ���
r�cCs~i}tddd�|d<tddd�|d<tdd�|d<t��d|d<d	|d
<t��dd�|d
<tj�tj�	t
j��|d<|adS)z+Initialize the module as appropriate for NTrrrr�LIBDEST�
BINLIBDESTrn�	INCLUDEPY�
EXT_SUFFIXz.exe�EXE�.r�VERSION�BINDIRN)
r?r5�_imp�extension_suffixesrr�rr�dirname�abspathr�
executabler\)r�rrr
�_init_nt�sr�cGsbtdu�r0t��dtj�}|r(|�niattd<ttd<t�st�d�}|dur\|td<t�dt�}tjdkr�t	r�tj
�t��}tj
�
||�}ntj
�t��}tj
�tj
�|��td<t	�rtjdk�rt}tj
�td��s|t��k�rtj
�
|td�}tj
�|�td<tjd	k�r0d
dl}|�t�|�rZg}|D]}|�t�|���q>|StSdS)a�With no arguments, return a dictionary of all configuration
    variables relevant for the current platform.  Generally this includes
    everything needed to build extensions and install both pure modules and
    extensions.  On Unix, this means every variable defined in Python's
    installed Makefile; on Windows it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    N�_init_r1�exec_prefixr��SOr"rrAr)r\r�rarr)rr:r*rr+rr�rxr	r�r-�isabs�getcwdrrYrZ�customize_config_vars�append)�args�funcr�r"�baserZ�valsr)rrr
r]�sD



�
r]cCs*|dkrddl}|�dtd�t��|�S)z�Return the value of a single variable using the dictionary
    returned by 'get_config_vars()'.  Equivalent to
    get_config_vars().get(name)
    r�rNz SO is deprecated, use EXT_SUFFIXr)�warnings�warn�DeprecationWarningr]ra)r)r�rrr
r,:sr,)rN)rrN)N)N)5�__doc__r�rryrr�r�builtin_module_namesr*rr-r1rr�r:�base_prefixr(�base_exec_prefixr'r^r�rr�r�r�rr;rr)rrr+r.r��AttributeErrorrr5r?r[rprxr�rzr�r�r�r�r�r\r�r�r]r,rrrr
�<module>s\




+
8K





jKsite-packages/setuptools/_distutils/__pycache__/msvccompiler.cpython-39.pyc000064400000034735150564745010023257 0ustar00a

(�Re�[�@s~dZddlZddlZddlmZmZmZmZmZddl	m
Z
mZddlm
Z
dZz,ddlZdZeZejZejZejZejZWndey�z4ddlZddlZdZeZejZejZejZejZWney�e
�d�Yn0Yn0er�ejejejej fZ!d	d
�Z"dd�Z#d
d�Z$Gdd�d�Z%dd�Z&dd�Z'dd�Z(Gdd�de
�Z)e&�dk�rze
�*d�e)Z+ddl,m)Z)ddl,m%Z%dS)z�distutils.msvccompiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for the Microsoft Visual Studio.
�N)�DistutilsExecError�DistutilsPlatformError�CompileError�LibError�	LinkError)�	CCompiler�gen_lib_options)�logFTz�Warning: Can't read registry to find the necessary compiler setting
Make sure that Python modules winreg, win32api or win32con are installed.cCsjzt||�}Wnty"YdS0g}d}zt||�}WntyPYqfYn0|�|�|d7}q,|S)zReturn list of registry keys.Nr�)�RegOpenKeyEx�RegError�
RegEnumKey�append)�base�key�handle�L�i�k�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/msvccompiler.py�	read_keys7s


rcCs~zt||�}Wnty"YdS0i}d}zt||�\}}}WntyVYqzYn0|��}t|�|t|�<|d7}q,|S)zXReturn dict of registry keys and values.

    All names are converted to lowercase.
    Nrr
)rr�RegEnumValue�lower�convert_mbcs)rrr�dr�name�value�typerrr�read_valuesHs

rcCs8t|dd�}|dur4z|d�}Wnty2Yn0|S)N�decode�mbcs)�getattr�UnicodeError)�s�decrrrr]src@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�
MacroExpandercCsi|_|�|�dS�N)�macros�load_macros)�self�versionrrr�__init__gszMacroExpander.__init__cCs2tD](}t||�}|r|||jd|<q.qdS)Nz$(%s))�HKEYSrr()r*Zmacro�pathrrrrrr�	set_macroks

zMacroExpander.set_macroc

Cs�d|}|�d|dd�|�d|dd�d}|�d|d	�z*|d
krX|�d|d�n|�d|d
�Wn,ty�}ztd��WYd}~n
d}~00d}tD]T}zt||�}Wnty�Yq�Yn0t|d�}t|d||f�}	|	d|jd<q�dS)Nz%Software\Microsoft\VisualStudio\%0.1fZVCInstallDirz	\Setup\VCZ
productdirZVSInstallDirz	\Setup\VSz Software\Microsoft\.NETFrameworkZFrameworkDirZinstallrootg@ZFrameworkSDKDirzsdkinstallrootv1.1ZsdkinstallrootaPython was built with Visual Studio 2003;
extensions must be built with a compiler than can generate compatible binaries.
Visual Studio 2003 was not found on this system. If you have Cygwin installed,
you can try compiling with MingW32, by passing "-c mingw32" to setup.py.z.Software\Microsoft\NET Framework Setup\Productrz%s\%sr+z$(FrameworkVersion))	r/�KeyErrorrr-rrr
rr()
r*r+Zvsbase�net�exc�pr�hrrrrrr)rs,�

zMacroExpander.load_macroscCs$|j��D]\}}|�||�}q
|Sr')r(�items�replace)r*r$r�vrrr�sub�szMacroExpander.subN)�__name__�
__module__�__qualname__r,r/r)r8rrrrr&fsr&cCs�d}tj�|�}|dkrdS|t|�}tj|d��dd�\}}t|dd��d}|dkrf|d7}t|d	d
��d}|dkr�d}|dkr�||SdS)
z�Return the version of MSVC that was used to build Python.

    For Python 2.3 and up, the version number is included in
    sys.version.  For earlier versions, assume the compiler is MSVC 6.
    zMSC v.����N� r
����
��g$@r)�sysr+�find�len�split�int)�prefixrr$�restZmajorVersionZminorVersionrrr�get_build_version�srJcCs@d}tj�|�}|dkrdStj�d|�}tj|t|�|�S)zUReturn the processor architecture.

    Possible results are "Intel" or "AMD64".
    z bit (r<�Intel�))rCr+rDrE)rHr�jrrr�get_build_architecture�srNcCs0g}|D]"}tj�|�}||vr|�|�q|S)znReturn a list of normalized paths with duplicates removed.

    The current order of paths is maintained.
    )�osr.�normpathr)�pathsZ
reduced_pathsr3�nprrr�normalize_and_reduce_paths�srSc
@s�eZdZdZdZiZdgZgd�ZdgZdgZ	eeee	Z
dZdZd	Z
d
ZdZZdZd+dd�Zdd�Zd,dd�Zd-dd�Zd.dd�Zd/dd�Zdd�Zdd�Zd d!�Zd0d"d#�Zd$d%�Zd1d'd(�Zd)d*�ZdS)2�MSVCCompilerzwConcrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class.Zmsvcz.c)z.ccz.cppz.cxx�.rcz.mcz.resz.objz.libz.dllz%s%sz.exercCsvt�||||�t�|_t�|_|jdkr\|jdkrHd|_t|j�|_nd|_d|j|_	nd|jd|_	d|_
dS)	NrK�zSoftware\Microsoft\VisualStudiozSoftware\Microsoft\DevstudiozVisual Studio version %szMicrosoft SDK compiler %sr=F)rr,rJ�_MSVCCompiler__versionrN�_MSVCCompiler__arch�_MSVCCompiler__rootr&�_MSVCCompiler__macros�_MSVCCompiler__product�initialized)r*�verbose�dry_run�forcerrrr,�s

zMSVCCompiler.__init__cCs�g|_dtjvrDdtjvrD|�d�rDd|_d|_d|_d|_d|_nx|�	d�|_t
|j�d	krltd
|j��|�d�|_|�d�|_|�d�|_|�d�|_|�d�|_|�
d�|�
d�z&tjd�d
�D]}|j�|�q�Wnty�Yn0t|j�|_d
�|j�tjd<d|_|jdk�r<gd�|_gd�|_ngd�|_gd�|_gd�|_|jdk�rrgd�|_n
gd�|_dg|_d|_dS)NZDISTUTILS_USE_SDKZMSSdkzcl.exezlink.exezlib.exezrc.exezmc.exer.rzxPython was built with %s, and extensions need to be built with the same version of the compiler, but it isn't installed.�lib�include�;rK)�/nologo�/O2�/MD�/W3�/GX�/DNDEBUG)rc�/Od�/MDdrfrg�/Z7�/D_DEBUG)rcrdrerf�/GS-rh)rcrirjrfrmrkrl)�/DLLrcz/INCREMENTAL:NOrV)rnrc�/INCREMENTAL:no�/DEBUG)rnrcroz	/pdb:NonerprcT)�_MSVCCompiler__pathsrO�environ�find_exe�cc�linkerr`�rc�mc�get_msvc_pathsrErr[�set_path_env_varrFrr0rS�joinZpreprocess_optionsrX�compile_options�compile_options_debug�ldflags_sharedrW�ldflags_shared_debugZldflags_staticr\)r*r3rrr�
initialize�sL�






zMSVCCompiler.initialize�cCs�|durd}g}|D]�}tj�|�\}}tj�|�d}|tj�|�d�}||jvrbtd|��|rrtj�|�}||jvr�|�	tj�
|||j��q||jvr�|�	tj�
|||j��q|�	tj�
|||j
��q|S)Nr�r
zDon't know how to compile %s)rOr.�splitext�
splitdrive�isabs�src_extensionsr�basename�_rc_extensionsrrz�
res_extension�_mc_extensions�
obj_extension)r*Zsource_filenamesZ	strip_dir�
output_dirZ	obj_names�src_namer�extrrr�object_filenames8s,

�
��zMSVCCompiler.object_filenamesNc	Cst|js|��|�||||||�}	|	\}}
}}}|p6g}
|
�d�|rT|
�|j�n|
�|j�|
D�]}z||\}}Wnty�YqdYn0|r�tj	�
|�}||jvr�d|}�nX||jvr�d|}�nB||j
v�r<|}d|}z"|�|jg||g|g�Wqdt�y6}zt|��WYd}~qdd}~00qdn�||jv�rtj	�|�}tj	�|�}zl|�|jgd|d|g|g�tj	�tj	�|��\}}tj	�||d�}|�|jgd|g|g�Wqdt�y�}zt|��WYd}~qdd}~00qdntd||f��d	|}z&|�|jg|
|||g|�Wqdt�yl}zt|��WYd}~qdd}~00qd|
S)
Nz/cz/Tcz/Tpz/foz-hz-rrUz"Don't know how to compile %s to %sz/Fo)r\rZ_setup_compiler�extendr|r{r0rOr.�abspath�
_c_extensions�_cpp_extensionsr��spawnrvrrr��dirnamerwr�r�rzrt)r*�sourcesr�r(�include_dirs�debug�
extra_preargs�extra_postargs�dependsZcompile_info�objectsZpp_opts�buildZcompile_opts�obj�srcr�Z	input_optZ
output_opt�msgZh_dirZrc_dirr�_Zrc_filerrr�compileWs�
�




��


��
��
���
 zMSVCCompiler.compilec	
Cs�|js|��|�||�\}}|j||d�}|�||�r�|d|g}|rJz|�|jg|�Wq�ty�}zt|��WYd}~q�d}~00nt	�
d|�dS)N)r��/OUT:�skipping %s (up-to-date))r\r�_fix_object_args�library_filename�
_need_linkr�r`rrr	r�)	r*r�Zoutput_libnamer�r��target_lang�output_filenameZlib_argsr�rrr�create_static_lib�s� zMSVCCompiler.create_static_libc
Cs�|js|��|�||�\}}|�|||�}|\}}}|rL|�dt|��t||||�}|durptj�	||�}|�
||��r�|tjkr�|	r�|j
dd�}q�|jdd�}n|	r�|j
}n|j}g}|p�gD]}|�d|�q�||||d|g}|du�rHtj�tj�|��\}}tj�	tj�|d�|�|��}|�d|�|
�rZ|
|dd�<|�rj|�|�|�tj�|��z|�|jg|�Wn.t�y�}zt|��WYd}~n
d}~00nt�d|�dS)Nz5I don't know what to do with 'runtime_library_dirs': r
z/EXPORT:r�rz/IMPLIB:r�)r\rr�Z
_fix_lib_args�warn�strrrOr.rzr�rZ
EXECUTABLEr~r}rr�r�r�r�r��mkpathr�rurrr	r�)r*Ztarget_descr�r�r��	libraries�library_dirs�runtime_library_dirs�export_symbolsr�r�r��
build_tempr�Z
fixed_argsZlib_optsZldflagsZexport_opts�symZld_argsZdll_nameZdll_extZimplib_filer�rrr�link�sh�
��

��

��
 zMSVCCompiler.linkcCsd|S)Nz	/LIBPATH:r�r*�dirrrr�library_dir_optionszMSVCCompiler.library_dir_optioncCstd��dS)Nz<don't know how to set runtime library search path for MSVC++)rr�rrr�runtime_library_dir_options�z'MSVCCompiler.runtime_library_dir_optioncCs
|�|�Sr')r�)r*r`rrr�library_optionszMSVCCompiler.library_optioncCs\|r|d|g}n|g}|D]:}|D]0}tj�||�|��}tj�|�r$|Sq$qdS)NZ_d)rOr.rzr��exists)r*�dirsr`r�Z	try_namesr�rZlibfilerrr�find_library_file#szMSVCCompiler.find_library_filecCsz|jD].}tj�tj�|�|�}tj�|�r|Sqtjd�d�D].}tj�tj�|�|�}tj�|�rF|SqF|S)a�Return path to an MSVC executable program.

        Tries to find the program in several places: first, one of the
        MSVC program search paths from the registry; next, the directories
        in the PATH environment variable.  If any of those work, return an
        absolute path that is known to exist.  If none of them work, just
        return the original program name, 'exe'.
        �Pathrb)rqrOr.rzr��isfilerrrF)r*Zexer3�fnrrrrs5s	


zMSVCCompiler.find_exe�x86cCs�tsgS|d}|jdkr,d|j|jf}nd|j|f}tD]H}t||�}|r>|jdkrt|j�||��d�S||�d�Sq>|jdkr�tD]&}t|d|j�dur�|�d	�q�q�gS)
z�Get a list of devstudio directories (include, lib or path).

        Return a list of strings.  The list will be empty if unable to
        access the registry or appropriate registry keys not found.
        z dirsrVz6%s\%0.1f\VC\VC_OBJECTS_PLATFORM_INFO\Win32\Directoriesz?%s\6.0\Build System\Components\Platforms\Win32 (%s)\Directoriesrbr=z%s\6.0Nz�It seems you have Visual Studio 6 installed, but the expected registry settings are not present.
You must at least run the Visual Studio GUI once so that these entries are created.)	�
_can_read_regrWrYr-rrZr8rFr�)r*r.�platformrrrrrrrxKs,

��



zMSVCCompiler.get_msvc_pathscCs6|dkr|�d�}n
|�|�}|r2d�|�tj|<dS)z�Set environment variable 'name' to an MSVC path type value.

        This is equivalent to a SET command prior to execution of spawned
        commands.
        r`�libraryrbN)rxrzrOrr)r*rr3rrrryos

zMSVCCompiler.set_path_env_var)rrr)rr�)NNNrNNN)NrN)
NNNNNrNNNN)r)r�)r9r:r;�__doc__�
compiler_typeZexecutablesr�r�r�r�r�r�r�Zstatic_lib_extension�shared_lib_extensionZstatic_lib_formatZshared_lib_formatZ
exe_extensionr,rr�r�r�r�r�r�r�r�rsrxryrrrrrT�s`��
B�
 �
X�
�
S

$rTg @z3Importing new compiler from distutils.msvc9compiler)rT)r&)-r�rCrO�distutils.errorsrrrrr�distutils.ccompilerrr�	distutilsr	r��winregZhkey_mod�	OpenKeyExr�EnumKeyr
Z	EnumValuer�errorr�ImportErrorZwin32apiZwin32con�infoZ
HKEY_USERS�HKEY_CURRENT_USER�HKEY_LOCAL_MACHINE�HKEY_CLASSES_ROOTr-rrrr&rJrNrSrTr�ZOldMSVCCompilerZdistutils.msvc9compilerrrrr�<module>s`



�	-
9
site-packages/setuptools/_distutils/__pycache__/errors.cpython-39.pyc000064400000012356150564745010022063 0ustar00a

(�Re�
�@s8dZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGd	d
�d
e�ZGdd�de�ZGd
d�de�ZGdd�de�Z	Gdd�de�Z
Gdd�de�ZGdd�de�ZGdd�de�Z
Gdd�de�ZGdd�de�ZGdd�de�ZGdd �d e�ZGd!d"�d"e�ZGd#d$�d$e�ZGd%d&�d&e�Zd'S)(a�distutils.errors

Provides exceptions used by the Distutils modules.  Note that Distutils
modules may raise standard exceptions; in particular, SystemExit is
usually raised for errors that are obviously the end-user's fault
(eg. bad command-line arguments).

This module is safe to use in "from ... import *" mode; it only exports
symbols whose names start with "Distutils" and end with "Error".c@seZdZdZdS)�DistutilsErrorzThe root of all Distutils evil.N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/errors.pyrsrc@seZdZdZdS)�DistutilsModuleErrorz�Unable to load an expected module, or to find an expected class
    within some module (in particular, command modules and classes).Nrrrrrr	sr	c@seZdZdZdS)�DistutilsClassErrorz�Some command class (or possibly distribution class, if anyone
    feels a need to subclass Distribution) is found not to be holding
    up its end of the bargain, ie. implementing some part of the
    "command "interface.Nrrrrrr
sr
c@seZdZdZdS)�DistutilsGetoptErrorz7The option table provided to 'fancy_getopt()' is bogus.Nrrrrrrsrc@seZdZdZdS)�DistutilsArgErrorzaRaised by fancy_getopt in response to getopt.error -- ie. an
    error in the command line usage.Nrrrrrrsrc@seZdZdZdS)�DistutilsFileErrorz�Any problems in the filesystem: expected file not found, etc.
    Typically this is for problems that we detect before OSError
    could be raised.Nrrrrrr
$sr
c@seZdZdZdS)�DistutilsOptionErrora�Syntactic/semantic errors in command options, such as use of
    mutually conflicting options, or inconsistent options,
    badly-spelled values, etc.  No distinction is made between option
    values originating in the setup script, the command line, config
    files, or what-have-you -- but if we *know* something originated in
    the setup script, we'll raise DistutilsSetupError instead.Nrrrrrr*src@seZdZdZdS)�DistutilsSetupErrorzqFor errors that can be definitely blamed on the setup script,
    such as invalid keyword arguments to 'setup()'.Nrrrrrr3src@seZdZdZdS)�DistutilsPlatformErrorz�We don't know how to do something on the current platform (but
    we do know how to do it on some platform) -- eg. trying to compile
    C files on a platform not supported by a CCompiler subclass.Nrrrrrr8src@seZdZdZdS)�DistutilsExecErrorz`Any problems executing an external program (such as the C
    compiler, when compiling C files).Nrrrrrr>src@seZdZdZdS)�DistutilsInternalErrorzoInternal inconsistencies or impossibilities (obviously, this
    should never be seen if the code is working!).NrrrrrrCsrc@seZdZdZdS)�DistutilsTemplateErrorz%Syntax error in a file list template.NrrrrrrHsrc@seZdZdZdS)�DistutilsByteCompileErrorzByte compile error.NrrrrrrKsrc@seZdZdZdS)�CCompilerErrorz#Some compile/link operation failed.NrrrrrrOsrc@seZdZdZdS)�PreprocessErrorz.Failure to preprocess one or more C/C++ files.NrrrrrrRsrc@seZdZdZdS)�CompileErrorz2Failure to compile one or more C/C++ source files.NrrrrrrUsrc@seZdZdZdS)�LibErrorzKFailure to create a static library from one or more C/C++ object
    files.NrrrrrrXsrc@seZdZdZdS)�	LinkErrorz]Failure to link one or more C/C++ object files into an executable
    or shared library file.Nrrrrrr\src@seZdZdZdS)�UnknownFileErrorz(Attempt to process an unknown file type.Nrrrrrr`srN)r�	Exceptionrr	r
rrr
rrrrrrrrrrrrrrrrr�<module>s&
	site-packages/setuptools/_distutils/bcppcompiler.py000064400000035056150564745010016741 0ustar00"""distutils.bcppcompiler

Contains BorlandCCompiler, an implementation of the abstract CCompiler class
for the Borland C++ compiler.
"""

# This implementation by Lyle Johnson, based on the original msvccompiler.py
# module and using the directions originally published by Gordon Williams.

# XXX looks like there's a LOT of overlap between these two classes:
# someone should sit down and factor out the common code as
# WindowsCCompiler!  --GPW


import os
from distutils.errors import \
     DistutilsExecError, \
     CompileError, LibError, LinkError, UnknownFileError
from distutils.ccompiler import \
     CCompiler, gen_preprocess_options
from distutils.file_util import write_file
from distutils.dep_util import newer
from distutils import log

class BCPPCompiler(CCompiler) :
    """Concrete class that implements an interface to the Borland C/C++
    compiler, as defined by the CCompiler abstract class.
    """

    compiler_type = 'bcpp'

    # Just set this so CCompiler's constructor doesn't barf.  We currently
    # don't use the 'set_executables()' bureaucracy provided by CCompiler,
    # as it really isn't necessary for this sort of single-compiler class.
    # Would be nice to have a consistent interface with UnixCCompiler,
    # though, so it's worth thinking about.
    executables = {}

    # Private class data (need to distinguish C from C++ source for compiler)
    _c_extensions = ['.c']
    _cpp_extensions = ['.cc', '.cpp', '.cxx']

    # Needed for the filename generation methods provided by the
    # base class, CCompiler.
    src_extensions = _c_extensions + _cpp_extensions
    obj_extension = '.obj'
    static_lib_extension = '.lib'
    shared_lib_extension = '.dll'
    static_lib_format = shared_lib_format = '%s%s'
    exe_extension = '.exe'


    def __init__ (self,
                  verbose=0,
                  dry_run=0,
                  force=0):

        CCompiler.__init__ (self, verbose, dry_run, force)

        # These executables are assumed to all be in the path.
        # Borland doesn't seem to use any special registry settings to
        # indicate their installation locations.

        self.cc = "bcc32.exe"
        self.linker = "ilink32.exe"
        self.lib = "tlib.exe"

        self.preprocess_options = None
        self.compile_options = ['/tWM', '/O2', '/q', '/g0']
        self.compile_options_debug = ['/tWM', '/Od', '/q', '/g0']

        self.ldflags_shared = ['/Tpd', '/Gn', '/q', '/x']
        self.ldflags_shared_debug = ['/Tpd', '/Gn', '/q', '/x']
        self.ldflags_static = []
        self.ldflags_exe = ['/Gn', '/q', '/x']
        self.ldflags_exe_debug = ['/Gn', '/q', '/x','/r']


    # -- Worker methods ------------------------------------------------

    def compile(self, sources,
                output_dir=None, macros=None, include_dirs=None, debug=0,
                extra_preargs=None, extra_postargs=None, depends=None):

        macros, objects, extra_postargs, pp_opts, build = \
                self._setup_compile(output_dir, macros, include_dirs, sources,
                                    depends, extra_postargs)
        compile_opts = extra_preargs or []
        compile_opts.append ('-c')
        if debug:
            compile_opts.extend (self.compile_options_debug)
        else:
            compile_opts.extend (self.compile_options)

        for obj in objects:
            try:
                src, ext = build[obj]
            except KeyError:
                continue
            # XXX why do the normpath here?
            src = os.path.normpath(src)
            obj = os.path.normpath(obj)
            # XXX _setup_compile() did a mkpath() too but before the normpath.
            # Is it possible to skip the normpath?
            self.mkpath(os.path.dirname(obj))

            if ext == '.res':
                # This is already a binary file -- skip it.
                continue # the 'for' loop
            if ext == '.rc':
                # This needs to be compiled to a .res file -- do it now.
                try:
                    self.spawn (["brcc32", "-fo", obj, src])
                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue # the 'for' loop

            # The next two are both for the real compiler.
            if ext in self._c_extensions:
                input_opt = ""
            elif ext in self._cpp_extensions:
                input_opt = "-P"
            else:
                # Unknown file type -- no extra options.  The compiler
                # will probably fail, but let it just in case this is a
                # file the compiler recognizes even if we don't.
                input_opt = ""

            output_opt = "-o" + obj

            # Compiler command line syntax is: "bcc32 [options] file(s)".
            # Note that the source file names must appear at the end of
            # the command line.
            try:
                self.spawn ([self.cc] + compile_opts + pp_opts +
                            [input_opt, output_opt] +
                            extra_postargs + [src])
            except DistutilsExecError as msg:
                raise CompileError(msg)

        return objects

    # compile ()


    def create_static_lib (self,
                           objects,
                           output_libname,
                           output_dir=None,
                           debug=0,
                           target_lang=None):

        (objects, output_dir) = self._fix_object_args (objects, output_dir)
        output_filename = \
            self.library_filename (output_libname, output_dir=output_dir)

        if self._need_link (objects, output_filename):
            lib_args = [output_filename, '/u'] + objects
            if debug:
                pass                    # XXX what goes here?
            try:
                self.spawn ([self.lib] + lib_args)
            except DistutilsExecError as msg:
                raise LibError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    # create_static_lib ()


    def link (self,
              target_desc,
              objects,
              output_filename,
              output_dir=None,
              libraries=None,
              library_dirs=None,
              runtime_library_dirs=None,
              export_symbols=None,
              debug=0,
              extra_preargs=None,
              extra_postargs=None,
              build_temp=None,
              target_lang=None):

        # XXX this ignores 'build_temp'!  should follow the lead of
        # msvccompiler.py

        (objects, output_dir) = self._fix_object_args (objects, output_dir)
        (libraries, library_dirs, runtime_library_dirs) = \
            self._fix_lib_args (libraries, library_dirs, runtime_library_dirs)

        if runtime_library_dirs:
            log.warn("I don't know what to do with 'runtime_library_dirs': %s",
                     str(runtime_library_dirs))

        if output_dir is not None:
            output_filename = os.path.join (output_dir, output_filename)

        if self._need_link (objects, output_filename):

            # Figure out linker args based on type of target.
            if target_desc == CCompiler.EXECUTABLE:
                startup_obj = 'c0w32'
                if debug:
                    ld_args = self.ldflags_exe_debug[:]
                else:
                    ld_args = self.ldflags_exe[:]
            else:
                startup_obj = 'c0d32'
                if debug:
                    ld_args = self.ldflags_shared_debug[:]
                else:
                    ld_args = self.ldflags_shared[:]


            # Create a temporary exports file for use by the linker
            if export_symbols is None:
                def_file = ''
            else:
                head, tail = os.path.split (output_filename)
                modname, ext = os.path.splitext (tail)
                temp_dir = os.path.dirname(objects[0]) # preserve tree structure
                def_file = os.path.join (temp_dir, '%s.def' % modname)
                contents = ['EXPORTS']
                for sym in (export_symbols or []):
                    contents.append('  %s=_%s' % (sym, sym))
                self.execute(write_file, (def_file, contents),
                             "writing %s" % def_file)

            # Borland C++ has problems with '/' in paths
            objects2 = map(os.path.normpath, objects)
            # split objects in .obj and .res files
            # Borland C++ needs them at different positions in the command line
            objects = [startup_obj]
            resources = []
            for file in objects2:
                (base, ext) = os.path.splitext(os.path.normcase(file))
                if ext == '.res':
                    resources.append(file)
                else:
                    objects.append(file)


            for l in library_dirs:
                ld_args.append("/L%s" % os.path.normpath(l))
            ld_args.append("/L.") # we sometimes use relative paths

            # list of object files
            ld_args.extend(objects)

            # XXX the command-line syntax for Borland C++ is a bit wonky;
            # certain filenames are jammed together in one big string, but
            # comma-delimited.  This doesn't mesh too well with the
            # Unix-centric attitude (with a DOS/Windows quoting hack) of
            # 'spawn()', so constructing the argument list is a bit
            # awkward.  Note that doing the obvious thing and jamming all
            # the filenames and commas into one argument would be wrong,
            # because 'spawn()' would quote any filenames with spaces in
            # them.  Arghghh!.  Apparently it works fine as coded...

            # name of dll/exe file
            ld_args.extend([',',output_filename])
            # no map file and start libraries
            ld_args.append(',,')

            for lib in libraries:
                # see if we find it and if there is a bcpp specific lib
                # (xxx_bcpp.lib)
                libfile = self.find_library_file(library_dirs, lib, debug)
                if libfile is None:
                    ld_args.append(lib)
                    # probably a BCPP internal library -- don't warn
                else:
                    # full name which prefers bcpp_xxx.lib over xxx.lib
                    ld_args.append(libfile)

            # some default libraries
            ld_args.append ('import32')
            ld_args.append ('cw32mt')

            # def file for export symbols
            ld_args.extend([',',def_file])
            # add resource files
            ld_args.append(',')
            ld_args.extend(resources)


            if extra_preargs:
                ld_args[:0] = extra_preargs
            if extra_postargs:
                ld_args.extend(extra_postargs)

            self.mkpath (os.path.dirname (output_filename))
            try:
                self.spawn ([self.linker] + ld_args)
            except DistutilsExecError as msg:
                raise LinkError(msg)

        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    # link ()

    # -- Miscellaneous methods -----------------------------------------


    def find_library_file (self, dirs, lib, debug=0):
        # List of effective library names to try, in order of preference:
        # xxx_bcpp.lib is better than xxx.lib
        # and xxx_d.lib is better than xxx.lib if debug is set
        #
        # The "_bcpp" suffix is to handle a Python installation for people
        # with multiple compilers (primarily Distutils hackers, I suspect
        # ;-).  The idea is they'd have one static library for each
        # compiler they care about, since (almost?) every Windows compiler
        # seems to have a different format for static libraries.
        if debug:
            dlib = (lib + "_d")
            try_names = (dlib + "_bcpp", lib + "_bcpp", dlib, lib)
        else:
            try_names = (lib + "_bcpp", lib)

        for dir in dirs:
            for name in try_names:
                libfile = os.path.join(dir, self.library_filename(name))
                if os.path.exists(libfile):
                    return libfile
        else:
            # Oops, didn't find it in *any* of 'dirs'
            return None

    # overwrite the one from CCompiler to support rc and res-files
    def object_filenames (self,
                          source_filenames,
                          strip_dir=0,
                          output_dir=''):
        if output_dir is None: output_dir = ''
        obj_names = []
        for src_name in source_filenames:
            # use normcase to make sure '.rc' is really '.rc' and not '.RC'
            (base, ext) = os.path.splitext (os.path.normcase(src_name))
            if ext not in (self.src_extensions + ['.rc','.res']):
                raise UnknownFileError("unknown file type '%s' (from '%s')" % \
                      (ext, src_name))
            if strip_dir:
                base = os.path.basename (base)
            if ext == '.res':
                # these can go unchanged
                obj_names.append (os.path.join (output_dir, base + ext))
            elif ext == '.rc':
                # these need to be compiled to .res-files
                obj_names.append (os.path.join (output_dir, base + '.res'))
            else:
                obj_names.append (os.path.join (output_dir,
                                            base + self.obj_extension))
        return obj_names

    # object_filenames ()

    def preprocess (self,
                    source,
                    output_file=None,
                    macros=None,
                    include_dirs=None,
                    extra_preargs=None,
                    extra_postargs=None):

        (_, macros, include_dirs) = \
            self._fix_compile_args(None, macros, include_dirs)
        pp_opts = gen_preprocess_options(macros, include_dirs)
        pp_args = ['cpp32.exe'] + pp_opts
        if output_file is not None:
            pp_args.append('-o' + output_file)
        if extra_preargs:
            pp_args[:0] = extra_preargs
        if extra_postargs:
            pp_args.extend(extra_postargs)
        pp_args.append(source)

        # We need to preprocess: either we're being forced to, or the
        # source file is newer than the target (or the target doesn't
        # exist).
        if self.force or output_file is None or newer(source, output_file):
            if output_file:
                self.mkpath(os.path.dirname(output_file))
            try:
                self.spawn(pp_args)
            except DistutilsExecError as msg:
                print(msg)
                raise CompileError(msg)

    # preprocess()
site-packages/setuptools/_distutils/cmd.py000064400000043237150564745010015025 0ustar00"""distutils.cmd

Provides the Command class, the base class for the command classes
in the distutils.command package.
"""

import sys, os, re
from distutils.errors import DistutilsOptionError
from distutils import util, dir_util, file_util, archive_util, dep_util
from distutils import log

class Command:
    """Abstract base class for defining command classes, the "worker bees"
    of the Distutils.  A useful analogy for command classes is to think of
    them as subroutines with local variables called "options".  The options
    are "declared" in 'initialize_options()' and "defined" (given their
    final values, aka "finalized") in 'finalize_options()', both of which
    must be defined by every command class.  The distinction between the
    two is necessary because option values might come from the outside
    world (command line, config file, ...), and any options dependent on
    other options must be computed *after* these outside influences have
    been processed -- hence 'finalize_options()'.  The "body" of the
    subroutine, where it does all its work based on the values of its
    options, is the 'run()' method, which must also be implemented by every
    command class.
    """

    # 'sub_commands' formalizes the notion of a "family" of commands,
    # eg. "install" as the parent with sub-commands "install_lib",
    # "install_headers", etc.  The parent of a family of commands
    # defines 'sub_commands' as a class attribute; it's a list of
    #    (command_name : string, predicate : unbound_method | string | None)
    # tuples, where 'predicate' is a method of the parent command that
    # determines whether the corresponding command is applicable in the
    # current situation.  (Eg. we "install_headers" is only applicable if
    # we have any C header files to install.)  If 'predicate' is None,
    # that command is always applicable.
    #
    # 'sub_commands' is usually defined at the *end* of a class, because
    # predicates can be unbound methods, so they must already have been
    # defined.  The canonical example is the "install" command.
    sub_commands = []


    # -- Creation/initialization methods -------------------------------

    def __init__(self, dist):
        """Create and initialize a new Command object.  Most importantly,
        invokes the 'initialize_options()' method, which is the real
        initializer and depends on the actual command being
        instantiated.
        """
        # late import because of mutual dependence between these classes
        from distutils.dist import Distribution

        if not isinstance(dist, Distribution):
            raise TypeError("dist must be a Distribution instance")
        if self.__class__ is Command:
            raise RuntimeError("Command is an abstract class")

        self.distribution = dist
        self.initialize_options()

        # Per-command versions of the global flags, so that the user can
        # customize Distutils' behaviour command-by-command and let some
        # commands fall back on the Distribution's behaviour.  None means
        # "not defined, check self.distribution's copy", while 0 or 1 mean
        # false and true (duh).  Note that this means figuring out the real
        # value of each flag is a touch complicated -- hence "self._dry_run"
        # will be handled by __getattr__, below.
        # XXX This needs to be fixed.
        self._dry_run = None

        # verbose is largely ignored, but needs to be set for
        # backwards compatibility (I think)?
        self.verbose = dist.verbose

        # Some commands define a 'self.force' option to ignore file
        # timestamps, but methods defined *here* assume that
        # 'self.force' exists for all commands.  So define it here
        # just to be safe.
        self.force = None

        # The 'help' flag is just used for command-line parsing, so
        # none of that complicated bureaucracy is needed.
        self.help = 0

        # 'finalized' records whether or not 'finalize_options()' has been
        # called.  'finalize_options()' itself should not pay attention to
        # this flag: it is the business of 'ensure_finalized()', which
        # always calls 'finalize_options()', to respect/update it.
        self.finalized = 0

    # XXX A more explicit way to customize dry_run would be better.
    def __getattr__(self, attr):
        if attr == 'dry_run':
            myval = getattr(self, "_" + attr)
            if myval is None:
                return getattr(self.distribution, attr)
            else:
                return myval
        else:
            raise AttributeError(attr)

    def ensure_finalized(self):
        if not self.finalized:
            self.finalize_options()
        self.finalized = 1

    # Subclasses must define:
    #   initialize_options()
    #     provide default values for all options; may be customized by
    #     setup script, by options from config file(s), or by command-line
    #     options
    #   finalize_options()
    #     decide on the final values for all options; this is called
    #     after all possible intervention from the outside world
    #     (command-line, option file, etc.) has been processed
    #   run()
    #     run the command: do whatever it is we're here to do,
    #     controlled by the command's various option values

    def initialize_options(self):
        """Set default values for all the options that this command
        supports.  Note that these defaults may be overridden by other
        commands, by the setup script, by config files, or by the
        command-line.  Thus, this is not the place to code dependencies
        between options; generally, 'initialize_options()' implementations
        are just a bunch of "self.foo = None" assignments.

        This method must be implemented by all command classes.
        """
        raise RuntimeError("abstract method -- subclass %s must override"
                           % self.__class__)

    def finalize_options(self):
        """Set final values for all the options that this command supports.
        This is always called as late as possible, ie.  after any option
        assignments from the command-line or from other commands have been
        done.  Thus, this is the place to code option dependencies: if
        'foo' depends on 'bar', then it is safe to set 'foo' from 'bar' as
        long as 'foo' still has the same value it was assigned in
        'initialize_options()'.

        This method must be implemented by all command classes.
        """
        raise RuntimeError("abstract method -- subclass %s must override"
                           % self.__class__)


    def dump_options(self, header=None, indent=""):
        from distutils.fancy_getopt import longopt_xlate
        if header is None:
            header = "command options for '%s':" % self.get_command_name()
        self.announce(indent + header, level=log.INFO)
        indent = indent + "  "
        for (option, _, _) in self.user_options:
            option = option.translate(longopt_xlate)
            if option[-1] == "=":
                option = option[:-1]
            value = getattr(self, option)
            self.announce(indent + "%s = %s" % (option, value),
                          level=log.INFO)

    def run(self):
        """A command's raison d'etre: carry out the action it exists to
        perform, controlled by the options initialized in
        'initialize_options()', customized by other commands, the setup
        script, the command-line, and config files, and finalized in
        'finalize_options()'.  All terminal output and filesystem
        interaction should be done by 'run()'.

        This method must be implemented by all command classes.
        """
        raise RuntimeError("abstract method -- subclass %s must override"
                           % self.__class__)

    def announce(self, msg, level=1):
        """If the current verbosity level is of greater than or equal to
        'level' print 'msg' to stdout.
        """
        log.log(level, msg)

    def debug_print(self, msg):
        """Print 'msg' to stdout if the global DEBUG (taken from the
        DISTUTILS_DEBUG environment variable) flag is true.
        """
        from distutils.debug import DEBUG
        if DEBUG:
            print(msg)
            sys.stdout.flush()


    # -- Option validation methods -------------------------------------
    # (these are very handy in writing the 'finalize_options()' method)
    #
    # NB. the general philosophy here is to ensure that a particular option
    # value meets certain type and value constraints.  If not, we try to
    # force it into conformance (eg. if we expect a list but have a string,
    # split the string on comma and/or whitespace).  If we can't force the
    # option into conformance, raise DistutilsOptionError.  Thus, command
    # classes need do nothing more than (eg.)
    #   self.ensure_string_list('foo')
    # and they can be guaranteed that thereafter, self.foo will be
    # a list of strings.

    def _ensure_stringlike(self, option, what, default=None):
        val = getattr(self, option)
        if val is None:
            setattr(self, option, default)
            return default
        elif not isinstance(val, str):
            raise DistutilsOptionError("'%s' must be a %s (got `%s`)"
                                       % (option, what, val))
        return val

    def ensure_string(self, option, default=None):
        """Ensure that 'option' is a string; if not defined, set it to
        'default'.
        """
        self._ensure_stringlike(option, "string", default)

    def ensure_string_list(self, option):
        r"""Ensure that 'option' is a list of strings.  If 'option' is
        currently a string, we split it either on /,\s*/ or /\s+/, so
        "foo bar baz", "foo,bar,baz", and "foo,   bar baz" all become
        ["foo", "bar", "baz"].
        """
        val = getattr(self, option)
        if val is None:
            return
        elif isinstance(val, str):
            setattr(self, option, re.split(r',\s*|\s+', val))
        else:
            if isinstance(val, list):
                ok = all(isinstance(v, str) for v in val)
            else:
                ok = False
            if not ok:
                raise DistutilsOptionError(
                      "'%s' must be a list of strings (got %r)"
                      % (option, val))

    def _ensure_tested_string(self, option, tester, what, error_fmt,
                              default=None):
        val = self._ensure_stringlike(option, what, default)
        if val is not None and not tester(val):
            raise DistutilsOptionError(("error in '%s' option: " + error_fmt)
                                       % (option, val))

    def ensure_filename(self, option):
        """Ensure that 'option' is the name of an existing file."""
        self._ensure_tested_string(option, os.path.isfile,
                                   "filename",
                                   "'%s' does not exist or is not a file")

    def ensure_dirname(self, option):
        self._ensure_tested_string(option, os.path.isdir,
                                   "directory name",
                                   "'%s' does not exist or is not a directory")


    # -- Convenience methods for commands ------------------------------

    def get_command_name(self):
        if hasattr(self, 'command_name'):
            return self.command_name
        else:
            return self.__class__.__name__

    def set_undefined_options(self, src_cmd, *option_pairs):
        """Set the values of any "undefined" options from corresponding
        option values in some other command object.  "Undefined" here means
        "is None", which is the convention used to indicate that an option
        has not been changed between 'initialize_options()' and
        'finalize_options()'.  Usually called from 'finalize_options()' for
        options that depend on some other command rather than another
        option of the same command.  'src_cmd' is the other command from
        which option values will be taken (a command object will be created
        for it if necessary); the remaining arguments are
        '(src_option,dst_option)' tuples which mean "take the value of
        'src_option' in the 'src_cmd' command object, and copy it to
        'dst_option' in the current command object".
        """
        # Option_pairs: list of (src_option, dst_option) tuples
        src_cmd_obj = self.distribution.get_command_obj(src_cmd)
        src_cmd_obj.ensure_finalized()
        for (src_option, dst_option) in option_pairs:
            if getattr(self, dst_option) is None:
                setattr(self, dst_option, getattr(src_cmd_obj, src_option))

    def get_finalized_command(self, command, create=1):
        """Wrapper around Distribution's 'get_command_obj()' method: find
        (create if necessary and 'create' is true) the command object for
        'command', call its 'ensure_finalized()' method, and return the
        finalized command object.
        """
        cmd_obj = self.distribution.get_command_obj(command, create)
        cmd_obj.ensure_finalized()
        return cmd_obj

    # XXX rename to 'get_reinitialized_command()'? (should do the
    # same in dist.py, if so)
    def reinitialize_command(self, command, reinit_subcommands=0):
        return self.distribution.reinitialize_command(command,
                                                      reinit_subcommands)

    def run_command(self, command):
        """Run some other command: uses the 'run_command()' method of
        Distribution, which creates and finalizes the command object if
        necessary and then invokes its 'run()' method.
        """
        self.distribution.run_command(command)

    def get_sub_commands(self):
        """Determine the sub-commands that are relevant in the current
        distribution (ie., that need to be run).  This is based on the
        'sub_commands' class attribute: each tuple in that list may include
        a method that we call to determine if the subcommand needs to be
        run for the current distribution.  Return a list of command names.
        """
        commands = []
        for (cmd_name, method) in self.sub_commands:
            if method is None or method(self):
                commands.append(cmd_name)
        return commands


    # -- External world manipulation -----------------------------------

    def warn(self, msg):
        log.warn("warning: %s: %s\n", self.get_command_name(), msg)

    def execute(self, func, args, msg=None, level=1):
        util.execute(func, args, msg, dry_run=self.dry_run)

    def mkpath(self, name, mode=0o777):
        dir_util.mkpath(name, mode, dry_run=self.dry_run)

    def copy_file(self, infile, outfile, preserve_mode=1, preserve_times=1,
                  link=None, level=1):
        """Copy a file respecting verbose, dry-run and force flags.  (The
        former two default to whatever is in the Distribution object, and
        the latter defaults to false for commands that don't define it.)"""
        return file_util.copy_file(infile, outfile, preserve_mode,
                                   preserve_times, not self.force, link,
                                   dry_run=self.dry_run)

    def copy_tree(self, infile, outfile, preserve_mode=1, preserve_times=1,
                   preserve_symlinks=0, level=1):
        """Copy an entire directory tree respecting verbose, dry-run,
        and force flags.
        """
        return dir_util.copy_tree(infile, outfile, preserve_mode,
                                  preserve_times, preserve_symlinks,
                                  not self.force, dry_run=self.dry_run)

    def move_file (self, src, dst, level=1):
        """Move a file respecting dry-run flag."""
        return file_util.move_file(src, dst, dry_run=self.dry_run)

    def spawn(self, cmd, search_path=1, level=1):
        """Spawn an external command respecting dry-run flag."""
        from distutils.spawn import spawn
        spawn(cmd, search_path, dry_run=self.dry_run)

    def make_archive(self, base_name, format, root_dir=None, base_dir=None,
                     owner=None, group=None):
        return archive_util.make_archive(base_name, format, root_dir, base_dir,
                                         dry_run=self.dry_run,
                                         owner=owner, group=group)

    def make_file(self, infiles, outfile, func, args,
                  exec_msg=None, skip_msg=None, level=1):
        """Special case of 'execute()' for operations that process one or
        more input files and generate one output file.  Works just like
        'execute()', except the operation is skipped and a different
        message printed if 'outfile' already exists and is newer than all
        files listed in 'infiles'.  If the command defined 'self.force',
        and it is true, then the command is unconditionally run -- does no
        timestamp checks.
        """
        if skip_msg is None:
            skip_msg = "skipping %s (inputs unchanged)" % outfile

        # Allow 'infiles' to be a single string
        if isinstance(infiles, str):
            infiles = (infiles,)
        elif not isinstance(infiles, (list, tuple)):
            raise TypeError(
                  "'infiles' must be a string, or a list or tuple of strings")

        if exec_msg is None:
            exec_msg = "generating %s from %s" % (outfile, ', '.join(infiles))

        # If 'outfile' must be regenerated (either because it doesn't
        # exist, is out-of-date, or the 'force' flag is true) then
        # perform the action that presumably regenerates it
        if self.force or dep_util.newer_group(infiles, outfile):
            self.execute(func, args, exec_msg, level)
        # Otherwise, print the "skip" message
        else:
            log.debug(skip_msg)
site-packages/setuptools/_distutils/py35compat.py000064400000000707150564745010016261 0ustar00import sys
import subprocess


def __optim_args_from_interpreter_flags():
    """Return a list of command-line arguments reproducing the current
    optimization settings in sys.flags."""
    args = []
    value = sys.flags.optimize
    if value > 0:
        args.append("-" + "O" * value)
    return args


_optim_args_from_interpreter_flags = getattr(
    subprocess,
    "_optim_args_from_interpreter_flags",
    __optim_args_from_interpreter_flags,
)
site-packages/setuptools/_distutils/command/upload.py000064400000016655150564745010017170 0ustar00"""
distutils.command.upload

Implements the Distutils 'upload' subcommand (upload package to a package
index).
"""

import os
import io
import hashlib
from base64 import standard_b64encode
from urllib.request import urlopen, Request, HTTPError
from urllib.parse import urlparse
from distutils.errors import DistutilsError, DistutilsOptionError
from distutils.core import PyPIRCCommand
from distutils.spawn import spawn
from distutils import log


# PyPI Warehouse supports MD5, SHA256, and Blake2 (blake2-256)
# https://bugs.python.org/issue40698
_FILE_CONTENT_DIGESTS = {
    "md5_digest": getattr(hashlib, "md5", None),
    "sha256_digest": getattr(hashlib, "sha256", None),
    "blake2_256_digest": getattr(hashlib, "blake2b", None),
}


class upload(PyPIRCCommand):

    description = "upload binary package to PyPI"

    user_options = PyPIRCCommand.user_options + [
        ('sign', 's',
         'sign files to upload using gpg'),
        ('identity=', 'i', 'GPG identity used to sign files'),
        ]

    boolean_options = PyPIRCCommand.boolean_options + ['sign']

    def initialize_options(self):
        PyPIRCCommand.initialize_options(self)
        self.username = ''
        self.password = ''
        self.show_response = 0
        self.sign = False
        self.identity = None

    def finalize_options(self):
        PyPIRCCommand.finalize_options(self)
        if self.identity and not self.sign:
            raise DistutilsOptionError(
                "Must use --sign for --identity to have meaning"
            )
        config = self._read_pypirc()
        if config != {}:
            self.username = config['username']
            self.password = config['password']
            self.repository = config['repository']
            self.realm = config['realm']

        # getting the password from the distribution
        # if previously set by the register command
        if not self.password and self.distribution.password:
            self.password = self.distribution.password

    def run(self):
        if not self.distribution.dist_files:
            msg = ("Must create and upload files in one command "
                   "(e.g. setup.py sdist upload)")
            raise DistutilsOptionError(msg)
        for command, pyversion, filename in self.distribution.dist_files:
            self.upload_file(command, pyversion, filename)

    def upload_file(self, command, pyversion, filename):
        # Makes sure the repository URL is compliant
        schema, netloc, url, params, query, fragments = \
            urlparse(self.repository)
        if params or query or fragments:
            raise AssertionError("Incompatible url %s" % self.repository)

        if schema not in ('http', 'https'):
            raise AssertionError("unsupported schema " + schema)

        # Sign if requested
        if self.sign:
            gpg_args = ["gpg", "--detach-sign", "-a", filename]
            if self.identity:
                gpg_args[2:2] = ["--local-user", self.identity]
            spawn(gpg_args,
                  dry_run=self.dry_run)

        # Fill in the data - send all the meta-data in case we need to
        # register a new release
        f = open(filename,'rb')
        try:
            content = f.read()
        finally:
            f.close()

        meta = self.distribution.metadata
        data = {
            # action
            ':action': 'file_upload',
            'protocol_version': '1',

            # identify release
            'name': meta.get_name(),
            'version': meta.get_version(),

            # file content
            'content': (os.path.basename(filename),content),
            'filetype': command,
            'pyversion': pyversion,

            # additional meta-data
            'metadata_version': '1.0',
            'summary': meta.get_description(),
            'home_page': meta.get_url(),
            'author': meta.get_contact(),
            'author_email': meta.get_contact_email(),
            'license': meta.get_licence(),
            'description': meta.get_long_description(),
            'keywords': meta.get_keywords(),
            'platform': meta.get_platforms(),
            'classifiers': meta.get_classifiers(),
            'download_url': meta.get_download_url(),
            # PEP 314
            'provides': meta.get_provides(),
            'requires': meta.get_requires(),
            'obsoletes': meta.get_obsoletes(),
            }

        data['comment'] = ''

        # file content digests
        for digest_name, digest_cons in _FILE_CONTENT_DIGESTS.items():
            if digest_cons is None:
                continue
            try:
                data[digest_name] = digest_cons(content).hexdigest()
            except ValueError:
                # hash digest not available or blocked by security policy
                pass

        if self.sign:
            with open(filename + ".asc", "rb") as f:
                data['gpg_signature'] = (os.path.basename(filename) + ".asc",
                                         f.read())

        # set up the authentication
        user_pass = (self.username + ":" + self.password).encode('ascii')
        # The exact encoding of the authentication string is debated.
        # Anyway PyPI only accepts ascii for both username or password.
        auth = "Basic " + standard_b64encode(user_pass).decode('ascii')

        # Build up the MIME payload for the POST data
        boundary = '--------------GHSKFJDLGDS7543FJKLFHRE75642756743254'
        sep_boundary = b'\r\n--' + boundary.encode('ascii')
        end_boundary = sep_boundary + b'--\r\n'
        body = io.BytesIO()
        for key, value in data.items():
            title = '\r\nContent-Disposition: form-data; name="%s"' % key
            # handle multiple entries for the same name
            if not isinstance(value, list):
                value = [value]
            for value in value:
                if type(value) is tuple:
                    title += '; filename="%s"' % value[0]
                    value = value[1]
                else:
                    value = str(value).encode('utf-8')
                body.write(sep_boundary)
                body.write(title.encode('utf-8'))
                body.write(b"\r\n\r\n")
                body.write(value)
        body.write(end_boundary)
        body = body.getvalue()

        msg = "Submitting %s to %s" % (filename, self.repository)
        self.announce(msg, log.INFO)

        # build the Request
        headers = {
            'Content-type': 'multipart/form-data; boundary=%s' % boundary,
            'Content-length': str(len(body)),
            'Authorization': auth,
        }

        request = Request(self.repository, data=body,
                          headers=headers)
        # send the data
        try:
            result = urlopen(request)
            status = result.getcode()
            reason = result.msg
        except HTTPError as e:
            status = e.code
            reason = e.msg
        except OSError as e:
            self.announce(str(e), log.ERROR)
            raise

        if status == 200:
            self.announce('Server response (%s): %s' % (status, reason),
                          log.INFO)
            if self.show_response:
                text = self._read_pypi_response(result)
                msg = '\n'.join(('-' * 75, text, '-' * 75))
                self.announce(msg, log.INFO)
        else:
            msg = 'Upload failed (%s): %s' % (status, reason)
            self.announce(msg, log.ERROR)
            raise DistutilsError(msg)
site-packages/setuptools/_distutils/command/build_ext.py000064400000075703150564745010017662 0ustar00"""distutils.command.build_ext

Implements the Distutils 'build_ext' command, for building extension
modules (currently limited to C extensions, should accommodate C++
extensions ASAP)."""

import contextlib
import os
import re
import sys
from distutils.core import Command
from distutils.errors import *
from distutils.sysconfig import customize_compiler, get_python_version
from distutils.sysconfig import get_config_h_filename
from distutils.dep_util import newer_group
from distutils.extension import Extension
from distutils.util import get_platform
from distutils import log
from . import py37compat

from site import USER_BASE

# An extension name is just a dot-separated list of Python NAMEs (ie.
# the same as a fully-qualified module name).
extension_name_re = re.compile \
    (r'^[a-zA-Z_][a-zA-Z_0-9]*(\.[a-zA-Z_][a-zA-Z_0-9]*)*$')


def show_compilers ():
    from distutils.ccompiler import show_compilers
    show_compilers()


class build_ext(Command):

    description = "build C/C++ extensions (compile/link to build directory)"

    # XXX thoughts on how to deal with complex command-line options like
    # these, i.e. how to make it so fancy_getopt can suck them off the
    # command line and make it look like setup.py defined the appropriate
    # lists of tuples of what-have-you.
    #   - each command needs a callback to process its command-line options
    #   - Command.__init__() needs access to its share of the whole
    #     command line (must ultimately come from
    #     Distribution.parse_command_line())
    #   - it then calls the current command class' option-parsing
    #     callback to deal with weird options like -D, which have to
    #     parse the option text and churn out some custom data
    #     structure
    #   - that data structure (in this case, a list of 2-tuples)
    #     will then be present in the command object by the time
    #     we get to finalize_options() (i.e. the constructor
    #     takes care of both command-line and client options
    #     in between initialize_options() and finalize_options())

    sep_by = " (separated by '%s')" % os.pathsep
    user_options = [
        ('build-lib=', 'b',
         "directory for compiled extension modules"),
        ('build-temp=', 't',
         "directory for temporary files (build by-products)"),
        ('plat-name=', 'p',
         "platform name to cross-compile for, if supported "
         "(default: %s)" % get_platform()),
        ('inplace', 'i',
         "ignore build-lib and put compiled extensions into the source " +
         "directory alongside your pure Python modules"),
        ('include-dirs=', 'I',
         "list of directories to search for header files" + sep_by),
        ('define=', 'D',
         "C preprocessor macros to define"),
        ('undef=', 'U',
         "C preprocessor macros to undefine"),
        ('libraries=', 'l',
         "external C libraries to link with"),
        ('library-dirs=', 'L',
         "directories to search for external C libraries" + sep_by),
        ('rpath=', 'R',
         "directories to search for shared C libraries at runtime"),
        ('link-objects=', 'O',
         "extra explicit link objects to include in the link"),
        ('debug', 'g',
         "compile/link with debugging information"),
        ('force', 'f',
         "forcibly build everything (ignore file timestamps)"),
        ('compiler=', 'c',
         "specify the compiler type"),
        ('parallel=', 'j',
         "number of parallel build jobs"),
        ('swig-cpp', None,
         "make SWIG create C++ files (default is C)"),
        ('swig-opts=', None,
         "list of SWIG command line options"),
        ('swig=', None,
         "path to the SWIG executable"),
        ('user', None,
         "add user include, library and rpath")
        ]

    boolean_options = ['inplace', 'debug', 'force', 'swig-cpp', 'user']

    help_options = [
        ('help-compiler', None,
         "list available compilers", show_compilers),
        ]

    def initialize_options(self):
        self.extensions = None
        self.build_lib = None
        self.plat_name = None
        self.build_temp = None
        self.inplace = 0
        self.package = None

        self.include_dirs = None
        self.define = None
        self.undef = None
        self.libraries = None
        self.library_dirs = None
        self.rpath = None
        self.link_objects = None
        self.debug = None
        self.force = None
        self.compiler = None
        self.swig = None
        self.swig_cpp = None
        self.swig_opts = None
        self.user = None
        self.parallel = None

    def finalize_options(self):
        from distutils import sysconfig

        self.set_undefined_options('build',
                                   ('build_lib', 'build_lib'),
                                   ('build_temp', 'build_temp'),
                                   ('compiler', 'compiler'),
                                   ('debug', 'debug'),
                                   ('force', 'force'),
                                   ('parallel', 'parallel'),
                                   ('plat_name', 'plat_name'),
                                   )

        if self.package is None:
            self.package = self.distribution.ext_package

        self.extensions = self.distribution.ext_modules

        # Make sure Python's include directories (for Python.h, pyconfig.h,
        # etc.) are in the include search path.
        py_include = sysconfig.get_python_inc()
        plat_py_include = sysconfig.get_python_inc(plat_specific=1)
        if self.include_dirs is None:
            self.include_dirs = self.distribution.include_dirs or []
        if isinstance(self.include_dirs, str):
            self.include_dirs = self.include_dirs.split(os.pathsep)

        # If in a virtualenv, add its include directory
        # Issue 16116
        if sys.exec_prefix != sys.base_exec_prefix:
            self.include_dirs.append(os.path.join(sys.exec_prefix, 'include'))

        # Put the Python "system" include dir at the end, so that
        # any local include dirs take precedence.
        self.include_dirs.extend(py_include.split(os.path.pathsep))
        if plat_py_include != py_include:
            self.include_dirs.extend(
                plat_py_include.split(os.path.pathsep))

        self.ensure_string_list('libraries')
        self.ensure_string_list('link_objects')

        # Life is easier if we're not forever checking for None, so
        # simplify these options to empty lists if unset
        if self.libraries is None:
            self.libraries = []
        if self.library_dirs is None:
            self.library_dirs = []
        elif isinstance(self.library_dirs, str):
            self.library_dirs = self.library_dirs.split(os.pathsep)

        if self.rpath is None:
            self.rpath = []
        elif isinstance(self.rpath, str):
            self.rpath = self.rpath.split(os.pathsep)

        # for extensions under windows use different directories
        # for Release and Debug builds.
        # also Python's library directory must be appended to library_dirs
        if os.name == 'nt':
            # the 'libs' directory is for binary installs - we assume that
            # must be the *native* platform.  But we don't really support
            # cross-compiling via a binary install anyway, so we let it go.
            self.library_dirs.append(os.path.join(sys.exec_prefix, 'libs'))
            if sys.base_exec_prefix != sys.prefix:  # Issue 16116
                self.library_dirs.append(os.path.join(sys.base_exec_prefix, 'libs'))
            if self.debug:
                self.build_temp = os.path.join(self.build_temp, "Debug")
            else:
                self.build_temp = os.path.join(self.build_temp, "Release")

            # Append the source distribution include and library directories,
            # this allows distutils on windows to work in the source tree
            self.include_dirs.append(os.path.dirname(get_config_h_filename()))
            _sys_home = getattr(sys, '_home', None)
            if _sys_home:
                self.library_dirs.append(_sys_home)

            # Use the .lib files for the correct architecture
            if self.plat_name == 'win32':
                suffix = 'win32'
            else:
                # win-amd64
                suffix = self.plat_name[4:]
            new_lib = os.path.join(sys.exec_prefix, 'PCbuild')
            if suffix:
                new_lib = os.path.join(new_lib, suffix)
            self.library_dirs.append(new_lib)

        # For extensions under Cygwin, Python's library directory must be
        # appended to library_dirs
        if sys.platform[:6] == 'cygwin':
            if not sysconfig.python_build:
                # building third party extensions
                self.library_dirs.append(os.path.join(sys.prefix, "lib",
                                                      "python" + get_python_version(),
                                                      "config"))
            else:
                # building python standard extensions
                self.library_dirs.append('.')

        # For building extensions with a shared Python library,
        # Python's library directory must be appended to library_dirs
        # See Issues: #1600860, #4366
        if (sysconfig.get_config_var('Py_ENABLE_SHARED')):
            if not sysconfig.python_build:
                # building third party extensions
                self.library_dirs.append(sysconfig.get_config_var('LIBDIR'))
            else:
                # building python standard extensions
                self.library_dirs.append('.')

        # The argument parsing will result in self.define being a string, but
        # it has to be a list of 2-tuples.  All the preprocessor symbols
        # specified by the 'define' option will be set to '1'.  Multiple
        # symbols can be separated with commas.

        if self.define:
            defines = self.define.split(',')
            self.define = [(symbol, '1') for symbol in defines]

        # The option for macros to undefine is also a string from the
        # option parsing, but has to be a list.  Multiple symbols can also
        # be separated with commas here.
        if self.undef:
            self.undef = self.undef.split(',')

        if self.swig_opts is None:
            self.swig_opts = []
        else:
            self.swig_opts = self.swig_opts.split(' ')

        # Finally add the user include and library directories if requested
        if self.user:
            user_include = os.path.join(USER_BASE, "include")
            user_lib = os.path.join(USER_BASE, "lib")
            if os.path.isdir(user_include):
                self.include_dirs.append(user_include)
            if os.path.isdir(user_lib):
                self.library_dirs.append(user_lib)
                self.rpath.append(user_lib)

        if isinstance(self.parallel, str):
            try:
                self.parallel = int(self.parallel)
            except ValueError:
                raise DistutilsOptionError("parallel should be an integer")

    def run(self):
        from distutils.ccompiler import new_compiler

        # 'self.extensions', as supplied by setup.py, is a list of
        # Extension instances.  See the documentation for Extension (in
        # distutils.extension) for details.
        #
        # For backwards compatibility with Distutils 0.8.2 and earlier, we
        # also allow the 'extensions' list to be a list of tuples:
        #    (ext_name, build_info)
        # where build_info is a dictionary containing everything that
        # Extension instances do except the name, with a few things being
        # differently named.  We convert these 2-tuples to Extension
        # instances as needed.

        if not self.extensions:
            return

        # If we were asked to build any C/C++ libraries, make sure that the
        # directory where we put them is in the library search path for
        # linking extensions.
        if self.distribution.has_c_libraries():
            build_clib = self.get_finalized_command('build_clib')
            self.libraries.extend(build_clib.get_library_names() or [])
            self.library_dirs.append(build_clib.build_clib)

        # Setup the CCompiler object that we'll use to do all the
        # compiling and linking
        self.compiler = new_compiler(compiler=self.compiler,
                                     verbose=self.verbose,
                                     dry_run=self.dry_run,
                                     force=self.force)
        customize_compiler(self.compiler)
        # If we are cross-compiling, init the compiler now (if we are not
        # cross-compiling, init would not hurt, but people may rely on
        # late initialization of compiler even if they shouldn't...)
        if os.name == 'nt' and self.plat_name != get_platform():
            self.compiler.initialize(self.plat_name)

        # And make sure that any compile/link-related options (which might
        # come from the command-line or from the setup script) are set in
        # that CCompiler object -- that way, they automatically apply to
        # all compiling and linking done here.
        if self.include_dirs is not None:
            self.compiler.set_include_dirs(self.include_dirs)
        if self.define is not None:
            # 'define' option is a list of (name,value) tuples
            for (name, value) in self.define:
                self.compiler.define_macro(name, value)
        if self.undef is not None:
            for macro in self.undef:
                self.compiler.undefine_macro(macro)
        if self.libraries is not None:
            self.compiler.set_libraries(self.libraries)
        if self.library_dirs is not None:
            self.compiler.set_library_dirs(self.library_dirs)
        if self.rpath is not None:
            self.compiler.set_runtime_library_dirs(self.rpath)
        if self.link_objects is not None:
            self.compiler.set_link_objects(self.link_objects)

        # Now actually compile and link everything.
        self.build_extensions()

    def check_extensions_list(self, extensions):
        """Ensure that the list of extensions (presumably provided as a
        command option 'extensions') is valid, i.e. it is a list of
        Extension objects.  We also support the old-style list of 2-tuples,
        where the tuples are (ext_name, build_info), which are converted to
        Extension instances here.

        Raise DistutilsSetupError if the structure is invalid anywhere;
        just returns otherwise.
        """
        if not isinstance(extensions, list):
            raise DistutilsSetupError(
                  "'ext_modules' option must be a list of Extension instances")

        for i, ext in enumerate(extensions):
            if isinstance(ext, Extension):
                continue                # OK! (assume type-checking done
                                        # by Extension constructor)

            if not isinstance(ext, tuple) or len(ext) != 2:
                raise DistutilsSetupError(
                       "each element of 'ext_modules' option must be an "
                       "Extension instance or 2-tuple")

            ext_name, build_info = ext

            log.warn("old-style (ext_name, build_info) tuple found in "
                     "ext_modules for extension '%s' "
                     "-- please convert to Extension instance", ext_name)

            if not (isinstance(ext_name, str) and
                    extension_name_re.match(ext_name)):
                raise DistutilsSetupError(
                       "first element of each tuple in 'ext_modules' "
                       "must be the extension name (a string)")

            if not isinstance(build_info, dict):
                raise DistutilsSetupError(
                       "second element of each tuple in 'ext_modules' "
                       "must be a dictionary (build info)")

            # OK, the (ext_name, build_info) dict is type-safe: convert it
            # to an Extension instance.
            ext = Extension(ext_name, build_info['sources'])

            # Easy stuff: one-to-one mapping from dict elements to
            # instance attributes.
            for key in ('include_dirs', 'library_dirs', 'libraries',
                        'extra_objects', 'extra_compile_args',
                        'extra_link_args'):
                val = build_info.get(key)
                if val is not None:
                    setattr(ext, key, val)

            # Medium-easy stuff: same syntax/semantics, different names.
            ext.runtime_library_dirs = build_info.get('rpath')
            if 'def_file' in build_info:
                log.warn("'def_file' element of build info dict "
                         "no longer supported")

            # Non-trivial stuff: 'macros' split into 'define_macros'
            # and 'undef_macros'.
            macros = build_info.get('macros')
            if macros:
                ext.define_macros = []
                ext.undef_macros = []
                for macro in macros:
                    if not (isinstance(macro, tuple) and len(macro) in (1, 2)):
                        raise DistutilsSetupError(
                              "'macros' element of build info dict "
                              "must be 1- or 2-tuple")
                    if len(macro) == 1:
                        ext.undef_macros.append(macro[0])
                    elif len(macro) == 2:
                        ext.define_macros.append(macro)

            extensions[i] = ext

    def get_source_files(self):
        self.check_extensions_list(self.extensions)
        filenames = []

        # Wouldn't it be neat if we knew the names of header files too...
        for ext in self.extensions:
            filenames.extend(ext.sources)
        return filenames

    def get_outputs(self):
        # Sanity check the 'extensions' list -- can't assume this is being
        # done in the same run as a 'build_extensions()' call (in fact, we
        # can probably assume that it *isn't*!).
        self.check_extensions_list(self.extensions)

        # And build the list of output (built) filenames.  Note that this
        # ignores the 'inplace' flag, and assumes everything goes in the
        # "build" tree.
        outputs = []
        for ext in self.extensions:
            outputs.append(self.get_ext_fullpath(ext.name))
        return outputs

    def build_extensions(self):
        # First, sanity-check the 'extensions' list
        self.check_extensions_list(self.extensions)
        if self.parallel:
            self._build_extensions_parallel()
        else:
            self._build_extensions_serial()

    def _build_extensions_parallel(self):
        workers = self.parallel
        if self.parallel is True:
            workers = os.cpu_count()  # may return None
        try:
            from concurrent.futures import ThreadPoolExecutor
        except ImportError:
            workers = None

        if workers is None:
            self._build_extensions_serial()
            return

        with ThreadPoolExecutor(max_workers=workers) as executor:
            futures = [executor.submit(self.build_extension, ext)
                       for ext in self.extensions]
            for ext, fut in zip(self.extensions, futures):
                with self._filter_build_errors(ext):
                    fut.result()

    def _build_extensions_serial(self):
        for ext in self.extensions:
            with self._filter_build_errors(ext):
                self.build_extension(ext)

    @contextlib.contextmanager
    def _filter_build_errors(self, ext):
        try:
            yield
        except (CCompilerError, DistutilsError, CompileError) as e:
            if not ext.optional:
                raise
            self.warn('building extension "%s" failed: %s' %
                      (ext.name, e))

    def build_extension(self, ext):
        sources = ext.sources
        if sources is None or not isinstance(sources, (list, tuple)):
            raise DistutilsSetupError(
                  "in 'ext_modules' option (extension '%s'), "
                  "'sources' must be present and must be "
                  "a list of source filenames" % ext.name)
        # sort to make the resulting .so file build reproducible
        sources = sorted(sources)

        ext_path = self.get_ext_fullpath(ext.name)
        depends = sources + ext.depends
        if not (self.force or newer_group(depends, ext_path, 'newer')):
            log.debug("skipping '%s' extension (up-to-date)", ext.name)
            return
        else:
            log.info("building '%s' extension", ext.name)

        # First, scan the sources for SWIG definition files (.i), run
        # SWIG on 'em to create .c files, and modify the sources list
        # accordingly.
        sources = self.swig_sources(sources, ext)

        # Next, compile the source code to object files.

        # XXX not honouring 'define_macros' or 'undef_macros' -- the
        # CCompiler API needs to change to accommodate this, and I
        # want to do one thing at a time!

        # Two possible sources for extra compiler arguments:
        #   - 'extra_compile_args' in Extension object
        #   - CFLAGS environment variable (not particularly
        #     elegant, but people seem to expect it and I
        #     guess it's useful)
        # The environment variable should take precedence, and
        # any sensible compiler will give precedence to later
        # command line args.  Hence we combine them in order:
        extra_args = ext.extra_compile_args or []

        macros = ext.define_macros[:]
        for undef in ext.undef_macros:
            macros.append((undef,))

        objects = self.compiler.compile(sources,
                                         output_dir=self.build_temp,
                                         macros=macros,
                                         include_dirs=ext.include_dirs,
                                         debug=self.debug,
                                         extra_postargs=extra_args,
                                         depends=ext.depends)

        # XXX outdated variable, kept here in case third-part code
        # needs it.
        self._built_objects = objects[:]

        # Now link the object files together into a "shared object" --
        # of course, first we have to figure out all the other things
        # that go into the mix.
        if ext.extra_objects:
            objects.extend(ext.extra_objects)
        extra_args = ext.extra_link_args or []

        # Detect target language, if not provided
        language = ext.language or self.compiler.detect_language(sources)

        self.compiler.link_shared_object(
            objects, ext_path,
            libraries=self.get_libraries(ext),
            library_dirs=ext.library_dirs,
            runtime_library_dirs=ext.runtime_library_dirs,
            extra_postargs=extra_args,
            export_symbols=self.get_export_symbols(ext),
            debug=self.debug,
            build_temp=self.build_temp,
            target_lang=language)

    def swig_sources(self, sources, extension):
        """Walk the list of source files in 'sources', looking for SWIG
        interface (.i) files.  Run SWIG on all that are found, and
        return a modified 'sources' list with SWIG source files replaced
        by the generated C (or C++) files.
        """
        new_sources = []
        swig_sources = []
        swig_targets = {}

        # XXX this drops generated C/C++ files into the source tree, which
        # is fine for developers who want to distribute the generated
        # source -- but there should be an option to put SWIG output in
        # the temp dir.

        if self.swig_cpp:
            log.warn("--swig-cpp is deprecated - use --swig-opts=-c++")

        if self.swig_cpp or ('-c++' in self.swig_opts) or \
           ('-c++' in extension.swig_opts):
            target_ext = '.cpp'
        else:
            target_ext = '.c'

        for source in sources:
            (base, ext) = os.path.splitext(source)
            if ext == ".i":             # SWIG interface file
                new_sources.append(base + '_wrap' + target_ext)
                swig_sources.append(source)
                swig_targets[source] = new_sources[-1]
            else:
                new_sources.append(source)

        if not swig_sources:
            return new_sources

        swig = self.swig or self.find_swig()
        swig_cmd = [swig, "-python"]
        swig_cmd.extend(self.swig_opts)
        if self.swig_cpp:
            swig_cmd.append("-c++")

        # Do not override commandline arguments
        if not self.swig_opts:
            for o in extension.swig_opts:
                swig_cmd.append(o)

        for source in swig_sources:
            target = swig_targets[source]
            log.info("swigging %s to %s", source, target)
            self.spawn(swig_cmd + ["-o", target, source])

        return new_sources

    def find_swig(self):
        """Return the name of the SWIG executable.  On Unix, this is
        just "swig" -- it should be in the PATH.  Tries a bit harder on
        Windows.
        """
        if os.name == "posix":
            return "swig"
        elif os.name == "nt":
            # Look for SWIG in its standard installation directory on
            # Windows (or so I presume!).  If we find it there, great;
            # if not, act like Unix and assume it's in the PATH.
            for vers in ("1.3", "1.2", "1.1"):
                fn = os.path.join("c:\\swig%s" % vers, "swig.exe")
                if os.path.isfile(fn):
                    return fn
            else:
                return "swig.exe"
        else:
            raise DistutilsPlatformError(
                  "I don't know how to find (much less run) SWIG "
                  "on platform '%s'" % os.name)

    # -- Name generators -----------------------------------------------
    # (extension names, filenames, whatever)
    def get_ext_fullpath(self, ext_name):
        """Returns the path of the filename for a given extension.

        The file is located in `build_lib` or directly in the package
        (inplace option).
        """
        fullname = self.get_ext_fullname(ext_name)
        modpath = fullname.split('.')
        filename = self.get_ext_filename(modpath[-1])

        if not self.inplace:
            # no further work needed
            # returning :
            #   build_dir/package/path/filename
            filename = os.path.join(*modpath[:-1]+[filename])
            return os.path.join(self.build_lib, filename)

        # the inplace option requires to find the package directory
        # using the build_py command for that
        package = '.'.join(modpath[0:-1])
        build_py = self.get_finalized_command('build_py')
        package_dir = os.path.abspath(build_py.get_package_dir(package))

        # returning
        #   package_dir/filename
        return os.path.join(package_dir, filename)

    def get_ext_fullname(self, ext_name):
        """Returns the fullname of a given extension name.

        Adds the `package.` prefix"""
        if self.package is None:
            return ext_name
        else:
            return self.package + '.' + ext_name

    def get_ext_filename(self, ext_name):
        r"""Convert the name of an extension (eg. "foo.bar") into the name
        of the file from which it will be loaded (eg. "foo/bar.so", or
        "foo\bar.pyd").
        """
        from distutils.sysconfig import get_config_var
        ext_path = ext_name.split('.')
        ext_suffix = get_config_var('EXT_SUFFIX')
        return os.path.join(*ext_path) + ext_suffix

    def get_export_symbols(self, ext):
        """Return the list of symbols that a shared extension has to
        export.  This either uses 'ext.export_symbols' or, if it's not
        provided, "PyInit_" + module_name.  Only relevant on Windows, where
        the .pyd file (DLL) must export the module "PyInit_" function.
        """
        name = ext.name.split('.')[-1]
        try:
            # Unicode module name support as defined in PEP-489
            # https://www.python.org/dev/peps/pep-0489/#export-hook-name
            name.encode('ascii')
        except UnicodeEncodeError:
            suffix = 'U_' + name.encode('punycode').replace(b'-', b'_').decode('ascii')
        else:
            suffix = "_" + name

        initfunc_name = "PyInit" + suffix
        if initfunc_name not in ext.export_symbols:
            ext.export_symbols.append(initfunc_name)
        return ext.export_symbols

    def get_libraries(self, ext):
        """Return the list of libraries to link against when building a
        shared extension.  On most platforms, this is just 'ext.libraries';
        on Windows, we add the Python library (eg. python20.dll).
        """
        # The python library is always needed on Windows.  For MSVC, this
        # is redundant, since the library is mentioned in a pragma in
        # pyconfig.h that MSVC groks.  The other Windows compilers all seem
        # to need it mentioned explicitly, though, so that's what we do.
        # Append '_d' to the python import library on debug builds.
        if sys.platform == "win32":
            from distutils._msvccompiler import MSVCCompiler
            if not isinstance(self.compiler, MSVCCompiler):
                template = "python%d%d"
                if self.debug:
                    template = template + '_d'
                pythonlib = (template %
                       (sys.hexversion >> 24, (sys.hexversion >> 16) & 0xff))
                # don't extend ext.libraries, it may be shared with other
                # extensions, it is a reference to the original list
                return ext.libraries + [pythonlib]
        else:
            # On Android only the main executable and LD_PRELOADs are considered
            # to be RTLD_GLOBAL, all the dependencies of the main executable
            # remain RTLD_LOCAL and so the shared libraries must be linked with
            # libpython when python is built with a shared python library (issue
            # bpo-21536).
            # On Cygwin (and if required, other POSIX-like platforms based on
            # Windows like MinGW) it is simply necessary that all symbols in
            # shared libraries are resolved at link time.
            from distutils.sysconfig import get_config_var
            link_libpython = False
            if get_config_var('Py_ENABLE_SHARED'):
                # A native build on an Android device or on Cygwin
                if hasattr(sys, 'getandroidapilevel'):
                    link_libpython = True
                elif sys.platform == 'cygwin':
                    link_libpython = True
                elif '_PYTHON_HOST_PLATFORM' in os.environ:
                    # We are cross-compiling for one of the relevant platforms
                    if get_config_var('ANDROID_API_LEVEL') != 0:
                        link_libpython = True
                    elif get_config_var('MACHDEP') == 'cygwin':
                        link_libpython = True

            if link_libpython:
                ldversion = get_config_var('LDVERSION')
                return ext.libraries + ['python' + ldversion]

        return ext.libraries + py37compat.pythonlib()
site-packages/setuptools/_distutils/command/__pycache__/bdist_wininst.cpython-39.pyc000064400000020713150564745010025041 0ustar00a

(�Re�>�@stdZddlZddlZddlZddlmZddlmZddlm	Z	ddl
TddlmZddl
mZGd	d
�d
e�ZdS)zzdistutils.command.bdist_wininst

Implements the Distutils 'bdist_wininst' command: create a windows installer
exe-program.�N)�Command)�get_platform)�remove_tree)�*)�get_python_version)�logc
s�eZdZdZdddde�fdddd	d
ddd
dddg
Zgd�ZejdkZ	�fdd�Z
dd�Zdd�Zdd�Z
dd�Zd$dd�Zd d!�Zd"d#�Z�ZS)%�
bdist_wininstz-create an executable installer for MS Windows)z
bdist-dir=Nz1temporary directory for creating the distributionz
plat-name=�pz;platform name to embed in generated filenames (default: %s))�	keep-temp�kzPkeep the pseudo-installation tree around after creating the distribution archive)ztarget-version=Nz6require a specific python version on the target system)�no-target-compile�cz/do not compile .py to .pyc on the target system)�no-target-optimize�oz;do not compile .py to .pyo (optimized) on the target system)z	dist-dir=�dz-directory to put final built distributions in)zbitmap=�bz>bitmap to use for the installer instead of python-powered logo)ztitle=�tz?title to display on the installer background instead of default)�
skip-buildNz2skip rebuilding everything (for testing/debugging))zinstall-script=NzUbasename of installation script to be run after installation or before deinstallation)zpre-install-script=Nz{Fully qualified filename of a script to be run before any files are installed.  This script need not be in the distribution)zuser-access-control=Nz�specify Vista's UAC handling - 'none'/default=no handling, 'auto'=use UAC if target Python installed for all users, 'force'=always use UAC)r
rrr�win32cs$t�j|i|��t�dtd�dS)Nz^bdist_wininst command is deprecated since Python 3.8, use bdist_wheel (wheel packages) instead�)�super�__init__�warnings�warn�DeprecationWarning)�self�args�kw��	__class__���/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/bdist_wininst.pyr?s�zbdist_wininst.__init__cCsRd|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_d|_dS)Nr)
�	bdist_dir�	plat_name�	keep_temp�no_target_compile�no_target_optimize�target_version�dist_dir�bitmap�title�
skip_build�install_script�pre_install_script�user_access_control)rr r r!�initialize_optionsEsz bdist_wininst.initialize_optionscCs�|�dd�|jdurR|jr6|jr6|j�d�}|j|_|�d�j}tj	�
|d�|_|js^d|_|js�|j��r�t
�}|jr�|j|kr�td|f��||_|�ddd�|jr�|jjD]}|jtj	�|�kr�q�q�td|j��dS)	N�bdist)r+r+Zwininst�zMtarget version can only be %s, or the '--skip-build' option must be specified)r(r()r#r#z(install_script '%s' not found in scripts)�set_undefined_optionsr"r+r#�distribution�get_command_obj�get_finalized_command�
bdist_base�os�path�joinr'�has_ext_modulesr�DistutilsOptionErrorr,�scripts�basename)rr0r6Z
short_version�scriptr r r!�finalize_optionsUs>
�����zbdist_wininst.finalize_optionsc
Cs�tjdkr&|j��s|j��r&td��|js6|�d�|jddd�}|j	|_
|j|_d|_|j|_|�d�}d|_
d|_|j��r�|j}|s�|js�Jd	��d
tjdd�}d|j|f}|�d�}tj�|jd
|�|_dD],}|��}|dkr�|d}t|d||�q�t�d|j	�|��tj�dtj�|j	d��|��tjd=ddlm}|�}	|j� �}
|j!|	d|j	d�}|�"||
|j#�|j���r�t$�}nd}|jj%�&d||�'|
�f�t�(d|�t�)|�|j*�s�t+|j	|j,d�dS)Nrz^distribution contains extensions and/or C libraries; must be compiled on a Windows 32 platform�build�install�)�reinit_subcommandsr�install_libz Should have already checked thisz%d.%drz.%s-%s�lib)�purelib�platlib�headersr<�datarHz/Include/$dist_name�install_zinstalling to %sZPURELIB)�mktemp�zip)�root_dir�anyrzremoving temporary file '%s')�dry_run)-�sys�platformr3r:�has_c_libraries�DistutilsPlatformErrorr+�run_command�reinitialize_commandr"�root�warn_dirr#�compile�optimizer'�version_infor5r7r8r9�
build_base�	build_lib�upper�setattrr�info�ensure_finalized�insert�run�tempfilerK�get_fullname�make_archive�
create_exer)r�
dist_files�append�get_installer_filename�debug�remover$rrO)
rrArDr'�plat_specifierr@�key�valuerKZarchive_basename�fullname�arcnameZ	pyversionr r r!rb{st
���




��
��
zbdist_wininst.runcCsZg}|jj}|�d�|jpdd}dd�}dD]B}t||d�}|r0|d|��||�f}|�d|||�f�q0|�d	�|jr�|�d
|j�|�d||��|�d|j�|�d
|j�|j	r�|�d|j	�|j
r�|�d|j
�|j�p|j��}|�d||��ddl
}ddl}	d|�|�
��|	jf}
|�d|
�d�|�S)Nz
[metadata]r1�
cSs|�dd�S)Nrqz\n)�replace)�sr r r!�escape�sz)bdist_wininst.get_inidata.<locals>.escape)�author�author_email�description�
maintainer�maintainer_email�name�url�versionz
    %s: %sz%s=%sz
[Setup]zinstall_script=%szinfo=%sztarget_compile=%dztarget_optimize=%dztarget_version=%szuser_access_control=%sztitle=%srzBuilt %s with distutils-%sz
build_info=%s)r3�metadatarh�long_description�getattr�
capitalizer,r%r&r'r.r*rd�time�	distutils�ctime�__version__r9)r�linesr}r_rtrzrIr*r�r�Z
build_infor r r!�get_inidata�s>
�
�zbdist_wininst.get_inidataNc	Cs�ddl}|�|j�|��}|�|�}|�d|�|rtt|d��}|��}Wd�n1s`0Yt|�}	nd}	t|d���}
|
�	|�
��|r�|
�	|�t|t�r�|�
d�}|d}|j�rt|jddd	��}|���
d�}Wd�n1s�0Y||d
}n|d}|
�	|�|�ddt|�|	�}
|
�	|
�t|d��}|
�	|���Wd�n1�sr0YWd�n1�s�0YdS)
Nrzcreating %s�rb�wb�mbcs��rzlatin-1)�encodings
z<iiii{V4)�struct�mkpathr(r�ri�announce�open�read�len�write�
get_exe_bytes�
isinstance�str�encoder-�pack)rrpror)r�Zcfgdata�installer_name�fZ
bitmapdataZ	bitmaplen�filer>Zscript_data�headerr r r!rf�sD
&



�,
�
zbdist_wininst.create_execCsD|jr&tj�|jd||j|jf�}ntj�|jd||jf�}|S)Nz%s.%s-py%s.exez	%s.%s.exe)r'r7r8r9r(r#)rror�r r r!ri1s
��
�z$bdist_wininst.get_installer_filenamec	Cs(t�}|jrl|j|krl|jdkr&d}q�|jdkr6d}q�|jdkrFd}q�|jdkrVd}q�|jdkrfd	}q�d
}n>zddlm}Wnty�d
}Yn0|�d
�d}|d}tj�t	�}|j
dkr�|j
dd�dkr�|j
dd�}nd}tj�|d||f�}t|d�}z|�
�W|��S|��0dS)Nz2.4z6.0z7.1z2.5z8.0z3.2z9.0z3.4z10.0z14.0r)�CRT_ASSEMBLY_VERSION�.z.0r��winr1zwininst-%s%s.exer�)rr'�msvcrtr��ImportError�	partitionr7r8�dirname�__file__r#r9r�r��close)	rZcur_versionZbvr��major�	directoryZsfix�filenamer�r r r!r�>s<	






�zbdist_wininst.get_exe_bytes)N)�__name__�
__module__�__qualname__rwr�user_options�boolean_optionsrPrQZ_unsupportedrr/r?rbr�rfrir��
__classcell__r r rr!rs:���%
&Q.
7
r)�__doc__r7rPr�distutils.corer�distutils.utilr�distutils.dir_utilr�distutils.errors�distutils.sysconfigrr�rrr r r r!�<module>ssite-packages/setuptools/_distutils/command/__pycache__/build_py.cpython-39.pyc000064400000023255150564745010023774 0ustar00a

(�Reo@�@sddZddlZddlZddlZddlZddlmZddlTddl	m
Z
ddlmZGdd�de�Z
dS)	zHdistutils.command.build_py

Implements the Distutils 'build_py' command.�N)�Command)�*)�convert_path)�logc@s�eZdZdZgd�ZddgZddiZdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd d!�Zd.d#d$�Zd%d&�Zd'd(�Zd)d*�Zd+d,�Zd-S)/�build_pyz5"build" pure Python modules (copy to build directory)))z
build-lib=�dzdirectory to "build" (copy) to)�compile�czcompile .py to .pyc)�
no-compileNz!don't compile .py files [default])z	optimize=�Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])�force�fz2forcibly build everything (ignore file timestamps)rrr
cCs4d|_d|_d|_d|_d|_d|_d|_d|_dS)Nr)�	build_lib�
py_modules�package�package_data�package_dirr�optimizer��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/build_py.py�initialize_options szbuild_py.initialize_optionsc	Cs�|�ddd�|jj|_|jj|_|jj|_i|_|jjr^|jj��D]\}}t|�|j|<qF|��|_	t
|jt�s�z,t|j�|_d|jkr�dks�nJ�Wnt
tfy�td��Yn0dS)N�build)rr)rrr�zoptimize must be 0, 1, or 2)�set_undefined_options�distribution�packagesrrr�itemsr�get_data_files�
data_files�
isinstancer�int�
ValueError�AssertionError�DistutilsOptionError)r�name�pathrrr�finalize_options*s$�



 zbuild_py.finalize_optionscCs:|jr|��|jr$|��|��|�|jdd��dS)Nr)�include_bytecode)r�
build_modulesr�build_packages�build_package_data�byte_compile�get_outputsrrrr�runCszbuild_py.runcs�g}|js|S|jD]h}|�|�}tjj|jg|�d��}d�|rPt|�d��fdd�|�||�D�}|�	||||f�q|S)z?Generate list of '(package,src_dir,build_dir,filenames)' tuples�.r�csg|]}|�d��qS�Nr)�.0�file��plenrr�
<listcomp>ssz+build_py.get_data_files.<locals>.<listcomp>)
r�get_package_dir�osr'�joinr�split�len�find_data_files�append)r�datar�src_dir�	build_dir�	filenamesrr5rras



�zbuild_py.get_data_filescsd|j�dg�|j�|g�}g�|D]:}t�tj�t�|�t|���}���fdd�|D��q$�S)z6Return filenames for package's data files in 'src_dir'�cs$g|]}|�vrtj�|�r|�qSr)r9r'�isfile)r3�fn��filesrrr7�s�z,build_py.find_data_files.<locals>.<listcomp>)	r�get�globr9r'r:�escaper�extend)rrr@Zglobs�pattern�filelistrrFrr=ys�zbuild_py.find_data_filescCs`d}|jD]P\}}}}|D]>}tj�||�}|�tj�|��|jtj�||�|dd�qq
dS)z$Copy data files into build directoryNF��
preserve_mode)r r9r'r:�mkpath�dirname�	copy_file)rZlastdirrr@rArB�filename�targetrrrr,�s�zbuild_py.build_package_datacCs�|�d�}|js&|r tjj|�SdSn�g}|r�z|jd�|�}Wn(tyj|�d|d�|d=Yq*0|�d|�tjj|�Sq*|j�d�}|dur�|�d|�|r�tjj|�SdSdS)z�Return the directory, relative to the top of the source
           distribution, where package 'package' should be found
           (at least according to the 'package_dir' option, if any).r0rCr���N)r;rr9r'r:�KeyError�insertrH)rrr'�tailZpdirrrrr8�s(
	zbuild_py.get_package_dircCsj|dkr8tj�|�s td|��tj�|�s8td|��|rftj�|d�}tj�|�rZ|St�d|�dS)NrCz%package directory '%s' does not existz>supposed package directory '%s' exists, but is not a directoryz__init__.pyz8package init file '%s' not found (or not a regular file))	r9r'�exists�DistutilsFileError�isdirr:rDr�warn)rrr�init_pyrrr�
check_package�s&����zbuild_py.check_packagecCs&tj�|�st�d||�dSdSdS)Nz!file %s (for module %s) not foundFT)r9r'rDrr\)r�module�module_filerrr�check_module�szbuild_py.check_modulec	Cs�|�||�t�tj�t�|�d��}g}tj�|jj�}|D]P}tj�|�}||kr�tj�	tj�
|��d}|�|||f�q>|�d|�q>|S)Nz*.pyrzexcluding %s)
r^rIr9r'r:rJ�abspathr�script_name�splitext�basenamer>�debug_print)	rrrZmodule_files�modulesZsetup_scriptr
Zabs_fr_rrr�find_package_modules�szbuild_py.find_package_modulesc	Cs�i}g}|jD]�}|�d�}d�|dd��}|d}z||\}}Wn tyf|�|�}d}Yn0|s�|�||�}	|df||<|	r�|�|d|	f�tj�||d�}
|�	||
�s�q|�|||
f�q|S)a�Finds individually-specified Python modules, ie. those listed by
        module name in 'self.py_modules'.  Returns a list of tuples (package,
        module_base, filename): 'package' is a tuple of the path through
        package-space to the module; 'module_base' is the bare (no
        packages, no dots) module name, and 'filename' is the path to the
        ".py" file (relative to the distribution root) that implements the
        module.
        r0rrUr1�__init__�.py)
rr;r:rVr8r^r>r9r'ra)rrrgr_r'rZmodule_baser�checkedr]r`rrr�find_modules�s*



zbuild_py.find_modulescCsNg}|jr|�|���|jrJ|jD]$}|�|�}|�||�}|�|�q$|S)a4Compute the list of all modules that will be built, whether
        they are specified one-module-at-a-time ('self.py_modules') or
        by whole packages ('self.packages').  Return a list of tuples
        (package, module, module_file), just like 'find_modules()' and
        'find_package_modules()' do.)rrKrlrr8rh)rrgrr�mrrr�find_all_moduless

zbuild_py.find_all_modulescCsdd�|��D�S)NcSsg|]}|d�qS)rUr)r3r_rrrr7-�z-build_py.get_source_files.<locals>.<listcomp>)rnrrrr�get_source_files,szbuild_py.get_source_filescCs$|gt|�|dg}tjj|�S)Nrj)�listr9r'r:)rrArr_Zoutfile_pathrrr�get_module_outfile/szbuild_py.get_module_outfiler1cCs�|��}g}|D]p\}}}|�d�}|�|j||�}|�|�|r|jr^|�tjj|dd��|j	dkr|�tjj||j	d��q|dd�|j
D�7}|S)Nr0rC)�optimizationrcSs,g|]$\}}}}|D]}tj�||��qqSr)r9r'r:)r3rr@rArBrSrrrr7Bs
�z(build_py.get_outputs.<locals>.<listcomp>)rnr;rrrr>r�	importlib�util�cache_from_sourcerr )rr)rg�outputsrr_r`rSrrrr.3s&


�

�
�zbuild_py.get_outputscCsbt|t�r|�d�}nt|ttf�s,td��|�|j||�}tj	�
|�}|�|�|j||dd�S)Nr0z:'package' must be a string (dot-separated), list, or tuplerrN)
r!�strr;rq�tuple�	TypeErrorrrrr9r'rQrPrR)rr_r`r�outfile�dirrrr�build_moduleJs
�
zbuild_py.build_modulecCs*|��}|D]\}}}|�|||�qdSr2)rlr})rrgrr_r`rrrr*Yszbuild_py.build_modulescCsP|jD]D}|�|�}|�||�}|D]$\}}}||ks:J�|�|||�q$qdSr2)rr8rhr})rrrrgZpackage_r_r`rrrr+bs


zbuild_py.build_packagescCs�tjr|�d�dSddlm}|j}|dtjkr>|tj}|jrZ||d|j	||j
d�|jdkr||||j|j	||j
d�dS)Nz%byte-compiling is disabled, skipping.r)r-rU)rr�prefix�dry_run)�sys�dont_write_bytecoder\�distutils.utilr-rr9�seprrrr)rrGr-r~rrrr-vs


�

�zbuild_py.byte_compileN)r1)�__name__�
__module__�__qualname__�description�user_options�boolean_options�negative_optrr(r/rr=r,r8r^rarhrlrnrprrr.r}r*r+r-rrrrrs.



'4
	r)�__doc__r9�importlib.utilrtr�rI�distutils.corer�distutils.errorsr�r�	distutilsrrrrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/install.cpython-39.pyc000064400000033152150564745010023630 0ustar00a

(�Re�k�
@s:dZddlZddlZddlmZddlmZddlmZddl	m
Z
ddlmZddl
mZdd	lmZmZmZdd
lmZddlmZddlmZdd
lmZdZdddddd�Zdddddd�dddddd�edddddd�dddddd�d�Ze�r"dddd d!d�ed"<ddd#d$d!d�ed%<dZGd&d'�d'e�ZdS)(zFdistutils.command.install

Implements the Distutils 'install' command.�N)�log)�Command)�DEBUG)�get_config_vars)�DistutilsPlatformError)�
write_file)�convert_path�
subst_vars�change_root)�get_platform)�DistutilsOptionError)�	USER_BASE)�	USER_SITETz$base/Lib/site-packagesz$base/Include/$dist_namez
$base/Scriptsz$base)�purelib�platlib�headers�scripts�dataz/$base/lib/python$py_version_short/site-packagesz;$platbase/$platlibdir/python$py_version_short/site-packagesz9$base/include/python$py_version_short$abiflags/$dist_namez	$base/binz$base/lib/pythonz$base/$platlibdir/pythonz$base/include/python/$dist_namez$base/site-packagesz$base/include/$dist_name)�unix_prefix�	unix_home�nt�pypy�pypy_ntz	$usersitez4$userbase/Python$py_version_nodot/Include/$dist_namez)$userbase/Python$py_version_nodot/Scriptsz	$userbase�nt_userz=$userbase/include/python$py_version_short$abiflags/$dist_namez
$userbase/bin�	unix_userc@seZdZdZgd�Zgd�Zer>e�dddef�e�d�ddiZ	d	d
�Z
dd�Zd
d�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�Zd)d*�Zd+d,�Zd-d.�Zd/d0�Zd1d2�Zd3efd4efd5efd6efd7d8d9�fgZdS):�installz'install everything from build directory))zprefix=Nzinstallation prefix)zexec-prefix=Nz.(Unix only) prefix for platform-specific files)zhome=Nz+(Unix only) home directory to install under)z
install-base=Nz;base installation directory (instead of --prefix or --home))zinstall-platbase=Nz\base installation directory for platform-specific files (instead of --exec-prefix or --home))zroot=Nz<install everything relative to this alternate root directory)zinstall-purelib=Nz;installation directory for pure Python module distributions)zinstall-platlib=Nz8installation directory for non-pure module distributions)zinstall-lib=Nzginstallation directory for all module distributions (overrides --install-purelib and --install-platlib))zinstall-headers=Nz(installation directory for C/C++ headers)zinstall-scripts=Nz)installation directory for Python scripts)z
install-data=Nz%installation directory for data files)�compile�czcompile .py to .pyc [default])�
no-compileNzdon't compile .py files)z	optimize=�Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])�force�fz1force installation (overwrite any existing files))�
skip-buildNz2skip rebuilding everything (for testing/debugging))zrecord=Nz3filename in which to record list of installed files)rr r"�userNz!install in user site-package '%s'rrcCs�d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_d|_t
|_t|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)zInitializes options.Nr�)�prefix�exec_prefix�homer#�install_base�install_platbase�root�install_purelib�install_platlib�install_headers�install_lib�install_scripts�install_datar
�install_userbaser�install_usersiter�optimize�
extra_path�install_path_filer �
skip_build�warn_dir�
build_base�	build_lib�record��self�r=��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install.py�initialize_options�s2zinstall.initialize_optionscCs�|js|js|jr&|js|jr&td��|jr@|js8|jr@td��|jrl|jsd|jsd|jsd|jsd|jrltd��tjdkr�|jr�|�	d�d|_|�
d�tjdkr�|��n|��|�
d�t
j��d	}td
d�\}}z
t
j}Wnty�d}Yn0|j��|j��|j��|d
t
jdd�dt
jdd�|||||tt
dd�d�|_t�rn|j|jd<|j|jd<|��|�
d�|j|jd<|j|jd<t�r�d	dlm}t d�||j�|�!�|�
d�|j�r�|�"�|j#du�r|j�$��r|j%|_#n|j&|_#|�'dddddddd�|�(�|j#|_)tj*�+|j#|j,�|_#|j-du�rf|�.d dddddd�|�
d!�|�/d"d#d$�dS)%zFinalizes options.zWmust supply either prefix/exec-prefix/home or install-base/install-platbase -- not bothz9must supply either home or prefix/exec-prefix -- not bothzGcan't combine user with prefix, exec_prefix/home, or install_(plat)base�posixz+exec-prefix option ignored on this platformNzpre-finalize_{unix,other}zpost-finalize_{unix,other}()rr%r&�z%d.%d�z%d%d�
platlibdir�lib)�	dist_name�dist_version�
dist_fullname�
py_version�py_version_short�py_version_nodot�
sys_prefixr%�sys_exec_prefixr&�abiflagsrC�userbase�usersitezpost-expand_basedirs()�base�platbase)�pprintzconfig vars:zpost-expand_dirs()rrrrr�libbasezafter prepending root�build)r8r8)r9r9)0r%r&r'r(r)rr#�os�name�warn�	dump_dirs�
finalize_unix�finalize_other�sys�version�splitrrM�AttributeError�distribution�get_name�get_version�get_fullname�version_info�getattr�config_vars�
HAS_USER_SITEr1r2�expand_basedirsrrR�print�expand_dirs�create_home_pathr.�has_ext_modulesr,r+�
convert_paths�handle_extra_path�install_libbase�path�join�
extra_dirsr*�change_roots�set_undefined_options)r<rHr%r&rMrRr=r=r>�finalize_options�s�������









�




�	�
�zinstall.finalize_optionscCs�tsdSddlm}t�|d�|jD]r}|d}|ddkrL|dd�}||jvrx|j|}|�|�}t||�}n|�|�}t||�}t�d||�q(dS)zDumps the list of user options.Nr)�
longopt_xlate�:����=z  %s: %s)	r�distutils.fancy_getoptrur�debug�user_options�negative_opt�	translaterd)r<�msgru�opt�opt_name�valr=r=r>rX�s





zinstall.dump_dirscCs"|jdus|jdur\|jdur2|jdur2|jdusP|jdusP|jdusP|jdurXtd��dS|j	r�|j
durttd��|j
|_|_|�d�n�|j
dur�|j
|_|_|�d�nl|jdur�|jdur�td��tj�tj�|_tj�tj�|_n|jdu�r|j|_|j|_|j|_|�d�dS)z&Finalizes options for posix platforms.NzPinstall-base or install-platbase supplied, but installation scheme is incomplete�$User base directory is not specifiedrrz*must not supply exec-prefix without prefixr)r(r)r.r+r,r-r/r0rr#r1r�
select_schemer'r%r&rUro�normpathr[r;r=r=r>rY�sL
������
�


�zinstall.finalize_unixcCs�|jr8|jdurtd��|j|_|_|�tjd�n~|jdur\|j|_|_|�d�nZ|j	durvtj
�tj	�|_	|j	|_|_z|�tj�Wn t
y�tdtj��Yn0dS)z)Finalizes options for non-posix platformsNr��_userrz)I don't know how to install stuff on '%s')r#r1rr(r)r�rUrVr'r%ror�r[�KeyErrorr;r=r=r>rZ�s&
�

�zinstall.finalize_othercCsnttd�r2tjdkr2|�d�s2tjdkr.d}nd}t|}tD]*}d|}t||�dur>t	||||�q>dS)	z=Sets the install directories by applying the install schemes.�pypy_version_info)��)r��_homerrr�install_N)
�hasattrr[rc�endswithrUrV�INSTALL_SCHEMES�SCHEME_KEYSrd�setattr)r<rV�scheme�key�attrnamer=r=r>r��s
��
zinstall.select_schemecCsX|D]N}t||�}|durtjdks.tjdkr:tj�|�}t||j�}t|||�qdS)Nr@r)rdrUrVro�
expanduserr	rer�)r<�attrs�attrr�r=r=r>�
_expand_attrs�s
zinstall._expand_attrscCs|�gd��dS)zNCalls `os.path.expanduser` on install_base, install_platbase and
        root.)r(r)r*N�r�r;r=r=r>rg�szinstall.expand_basedirscCs|�gd��dS)z+Calls `os.path.expanduser` on install dirs.)r+r,r.r-r/r0Nr�r;r=r=r>ri�szinstall.expand_dirscGs,|D]"}d|}t||tt||���qdS)z!Call `convert_path` over `names`.r�N)r�rrd�r<�namesrVr�r=r=r>rl�szinstall.convert_pathscCs�|jdur|jj|_|jdur�t�d�t|jt�rB|j�d�|_t|j�dkr`|jd}}n"t|j�dkrz|j\}}ntd��t	|�}nd}d}||_
||_dS)	z4Set `path_file` and `extra_dirs` using `extra_path`.NzIDistribution option extra_path is deprecated. See issue27919 for details.�,r$rrBzY'extra_path' option must be a list, tuple, or comma-separated string with 1 or 2 elementsrA)r4r_rrW�
isinstance�strr]�lenrr�	path_filerq)r<r�rqr=r=r>rm�s(


��
zinstall.handle_extra_pathc	Gs0|D]&}d|}t||t|jt||���qdS)z:Change the install directories pointed by name using root.r�N)r�r
r*rdr�r=r=r>rr!szinstall.change_rootscCsb|js
dSttj�d��}|j��D]8\}}|�|�r$tj�|�s$|�	d|�t�
|d�q$dS)zCreate directories under ~.N�~zos.makedirs('%s', 0o700)i�)r#rrUror�re�items�
startswith�isdir�debug_print�makedirs)r<r'rVror=r=r>rj'szinstall.create_home_pathcCs"|js6|�d�|j�d�j}|jr6|t�kr6td��|��D]}|�|�q>|j	r\|�
�|jr�|��}|j
r�t|j
�}tt|��D]}|||d�||<q�|�t|j|fd|j�ttjjtj�}ttjj|�}tj�tj�|j��}|j�r|j	�r|j�s||v�rt�d|j�dS)zRuns the command.rTz"Can't install when cross-compilingNz'writing list of installed files to '%s'z�modules installed to '%s', which is not in Python's module search path (sys.path) -- you'll have to change the search path yourself)r6�run_commandr_�get_command_obj�	plat_namer7rr�get_sub_commandsr��create_path_filer:�get_outputsr*r��range�executer�maprUror�r[�normcaser.r5rrz)r<�
build_plat�cmd_name�outputs�root_len�counter�sys_pathr.r=r=r>�run3sD

������zinstall.runcCsJtj�|j|jd�}|jr8|�t||jgfd|�n|�	d|�dS)zCreates the .pth file�.pthzcreating %szpath file '%s' not createdN)
rUrorprnr�r5r�rrqrW)r<�filenamer=r=r>r�_s
�
�zinstall.create_path_filecCshg}|��D].}|�|�}|��D]}||vr"|�|�q"q|jrd|jrd|�tj�|j	|jd��|S)z.Assembles the outputs of all the sub-commands.r�)
r��get_finalized_commandr��appendr�r5rUrorprn)r<r�r��cmdr�r=r=r>r�ms
�zinstall.get_outputscCs.g}|��D]}|�|�}|�|���q|S)z*Returns the inputs of all the sub-commands)r�r��extend�
get_inputs)r<�inputsr�r�r=r=r>r�~s

zinstall.get_inputscCs|j��p|j��S)zSReturns true if the current distribution has any Python
        modules to install.)r_�has_pure_modulesrkr;r=r=r>�has_lib�s
�zinstall.has_libcCs
|j��S)zLReturns true if the current distribution has any headers to
        install.)r_�has_headersr;r=r=r>r��szinstall.has_headerscCs
|j��S)zMReturns true if the current distribution has any scripts to.
        install.)r_�has_scriptsr;r=r=r>r��szinstall.has_scriptscCs
|j��S)zJReturns true if the current distribution has any data to.
        install.)r_�has_data_filesr;r=r=r>�has_data�szinstall.has_datar.r-r/r0�install_egg_infocCsdS)NTr=r;r=r=r>�<lambda>��zinstall.<lambda>) �__name__�
__module__�__qualname__�descriptionr{�boolean_optionsrfr�rr|r?rtrXrYrZr�r�rgrirlrmrrrjr�r�r�r�r�r�r�r��sub_commandsr=r=r=r>rWsJ;�
N(	",
�r)�__doc__r[rU�	distutilsr�distutils.corer�distutils.debugr�distutils.sysconfigr�distutils.errorsr�distutils.file_utilr�distutils.utilrr	r
rr�siter
rrf�WINDOWS_SCHEMEr�r�rr=r=r=r>�<module>sz�
�����!�
	�
site-packages/setuptools/_distutils/command/__pycache__/build_clib.cpython-39.pyc000064400000011452150564745010024251 0ustar00a

(�ReV�@sTdZddlZddlmZddlTddlmZddlmZdd�Z	Gd	d
�d
e�Z
dS)z�distutils.command.build_clib

Implements the Distutils 'build_clib' command, to build a C/C++ library
that is included in the module distribution and needed by an extension
module.�N)�Command)�*)�customize_compiler)�logcCsddlm}|�dS)Nr��show_compilers)�distutils.ccompilerrr�r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/build_clib.pyrsrc@sfeZdZdZgd�ZddgZdddefgZdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�ZdS)�
build_clibz/build C/C++ libraries used by Python extensions))zbuild-clib=�bz%directory to build C/C++ libraries to)zbuild-temp=�tz,directory to put temporary build by-products)�debug�gz"compile with debugging information)�force�fz2forcibly build everything (ignore file timestamps))z	compiler=�czspecify the compiler typerrz
help-compilerNzlist available compilerscCs:d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)Nr)	r�
build_temp�	libraries�include_dirs�define�undefrr�compiler��selfr	r	r
�initialize_options4szbuild_clib.initialize_optionscCsh|�dddddd�|jj|_|jr0|�|j�|jdurH|jjpDg|_t|jt�rd|j�tj	�|_dS)N�build)rr)rr)rr)rr)rr)
�set_undefined_options�distributionr�check_library_listr�
isinstance�str�split�os�pathseprr	r	r
�finalize_optionsDs�

zbuild_clib.finalize_optionscCs�|js
dSddlm}||j|j|jd�|_t|j�|jdurN|j�|j�|j	durv|j	D]\}}|j�
||�q^|jdur�|jD]}|j�|�q�|�
|j�dS)Nr)�new_compiler)r�dry_runr)rrr&rr'rrrZset_include_dirsrZdefine_macrorZundefine_macro�build_libraries)rr&�name�valueZmacror	r	r
�run^s"�




zbuild_clib.runcCs�t|t�std��|D]z}t|t�s8t|�dkr8td��|\}}t|t�sRtd��d|vsntjdkr~tj|vr~td|d��t|t�std��qd	S)
a`Ensure that the list of libraries is valid.

        `library` is presumably provided as a command option 'libraries'.
        This method checks that it is a list of 2-tuples, where the tuples
        are (library_name, build_info_dict).

        Raise DistutilsSetupError if the structure is invalid anywhere;
        just returns otherwise.
        z+'libraries' option must be a list of tuples�z*each element of 'libraries' must a 2-tuplezNfirst element of each tuple in 'libraries' must be a string (the library name)�/z;bad library name '%s': may not contain directory separatorsrzMsecond element of each tuple in 'libraries' must be a dictionary (build info)N)	r �list�DistutilsSetupError�tuple�lenr!r#�sep�dict)rr�libr)�
build_infor	r	r
rvs,

��
��
�zbuild_clib.check_library_listcCs,|js
dSg}|jD]\}}|�|�q|S)N)r�append)rZ	lib_names�lib_namer5r	r	r
�get_library_names�szbuild_clib.get_library_namescCsZ|�|j�g}|jD]>\}}|�d�}|dus>t|ttf�sJtd|��|�|�q|S)N�sources�fin 'libraries' option (library '%s'), 'sources' must be present and must be a list of source filenames)rr�getr r.r0r/�extend)r�	filenamesr7r5r9r	r	r
�get_source_files�s
��zbuild_clib.get_source_filescCs�|D]�\}}|�d�}|dus,t|ttf�s8td|��t|�}t�d|�|�d�}|�d�}|jj||j	|||j
d�}|jj|||j|j
d�qdS)Nr9r:zbuilding '%s' library�macrosr)�
output_dirr?rr)r@r)
r;r r.r0r/r�infor�compilerrZcreate_static_libr)rrr7r5r9r?rZobjectsr	r	r
r(�s,
��

�	
�zbuild_clib.build_libraries)�__name__�
__module__�__qualname__�description�user_options�boolean_optionsr�help_optionsrr%r+rr8r>r(r	r	r	r
rs
��$r)�__doc__r#�distutils.corer�distutils.errors�distutils.sysconfigr�	distutilsrrrr	r	r	r
�<module>ssite-packages/setuptools/_distutils/command/__pycache__/py37compat.cpython-39.pyc000064400000002072150564745010024165 0ustar00a

(�Re��@sPddlZdd�Zdd�ZejdkrHejdkrHejdd�d	krHeee�neZdS)
�NccsDddlm}|�d�sdSd�tjd?tjd?d@|�d	��VdS)
zj
    On Python 3.7 and earlier, distutils would include the Python
    library. See pypa/distutils#9.
    r��	sysconfigZPy_ENABLED_SHAREDNz
python{}.{}{}����ABIFLAGS)�	distutilsr�get_config_var�format�sys�
hexversionr�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/py37compat.py�_pythonlib_compats
�rcs��fdd�S)Ncs��|i|���S)Nr
)�args�kwargs��f1�f2r
r�<lambda>�zcompose.<locals>.<lambda>r
rr
rr�composesr)���darwinr�aix)rrr�version_info�platform�listZ	pythonlibr
r
r
r�<module>s����site-packages/setuptools/_distutils/command/__pycache__/install_scripts.cpython-39.pyc000064400000004277150564745010025405 0ustar00a

(�Re��@sDdZddlZddlmZddlmZddlmZGdd�de�ZdS)zudistutils.command.install_scripts

Implements the Distutils 'install_scripts' command, for installing
Python scripts.�N)�Command)�log)�ST_MODEc@sHeZdZdZgd�ZddgZdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dS)�install_scriptsz%install scripts (Python or otherwise)))zinstall-dir=�dzdirectory to install scripts to)z
build-dir=�bz'build directory (where to install from))�force�fz-force installation (overwrite existing files))�
skip-buildNzskip the build stepsrr
cCsd|_d|_d|_d|_dS)Nr)�install_dirr�	build_dir�
skip_build��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install_scripts.py�initialize_optionssz"install_scripts.initialize_optionscCs |�dd�|�dddd�dS)N�build)�
build_scriptsr�install)rr)rr)r
r
)�set_undefined_optionsrrrr�finalize_options!s�z install_scripts.finalize_optionscCs�|js|�d�|�|j|j�|_tjdkr~|��D]H}|j	rLt
�d|�q4t�|�t
dBd@}t
�d||�t�||�q4dS)Nr�posixzchanging mode of %simi�zchanging mode of %s to %o)r
�run_command�	copy_treerr�outfiles�os�name�get_outputs�dry_runr�info�statr�chmod)r�file�moderrr�run)s

zinstall_scripts.runcCs|jjp
gS�N)�distribution�scriptsrrrr�
get_inputs8szinstall_scripts.get_inputscCs
|jpgSr&)rrrrrr;szinstall_scripts.get_outputsN)�__name__�
__module__�__qualname__�description�user_options�boolean_optionsrrr%r)rrrrrrsr)	�__doc__r�distutils.corer�	distutilsrr!rrrrrr�<module>s
site-packages/setuptools/_distutils/command/__pycache__/bdist_msi.cpython-39.pyc000064400000046644150564745010024151 0ustar00a

(�Re���@s�dZddlZddlZddlZddlmZddlmZddlm	Z	ddl
mZddlm
Z
ddlmZdd	lmZddlZdd
lmZmZmZddlmZmZmZmZGdd
�d
e�ZGdd�de�ZdS)z#
Implements the bdist_msi command.
�N)�Command)�remove_tree)�get_python_version)�
StrictVersion)�DistutilsOptionError)�get_platform)�log)�schema�sequence�text)�	Directory�Feature�Dialog�add_datac@sFeZdZdZdd�Zdd�Zddd	�Zddd�Zddd�Zdd�Z	dS)�PyDialogz�Dialog class with a fixed layout: controls at the top, then a ruler,
    then a list of buttons: back, next, cancel. Optionally a bitmap at the
    left.cOs@tj|g|�R�|jd}d|d}|�dd||jd�dS)zbDialog(database, name, x, y, w, h, attributes, title, first,
        default, cancel, bitmap=true)�$�iHZ
BottomLinerN)r�__init__�h�line�w)�self�args�kwZrulerZbmwidth�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/bdist_msi.pyrs
zPyDialog.__init__c
Cs|�ddddddd|�dS)	z,Set the title text of the dialog at the top.�Title��
�@�<�z{\VerdanaBold10}%sN)r)r�titlerrrr"%s�zPyDialog.title�Back�c
Cs,|r
d}nd}|�|d|jddd|||�S)z�Add a back button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associated�r$���8���
pushbuttonr�rr"�next�name�active�flagsrrr�back,sz
PyDialog.back�Cancelc
Cs,|r
d}nd}|�|d|jddd|||�S)z�Add a cancel button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associatedr%r$i0r'r(r)r*r,rrr�cancel7szPyDialog.cancel�Nextc
Cs,|r
d}nd}|�|d|jddd|||�S)z�Add a Next button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associatedr%r$��r'r(r)r*r,rrrr-Bsz
PyDialog.nextc
Cs,|�|t|j|d�|jdddd||�S)z�Add a button with a given title, the tab-next button,
        its name in the Control table, giving its x position; the
        y-position is aligned with the other buttons.

        Return the button, so that events can be associated�r'r(r)r%)r+�intrr)rr.r"r-Zxposrrr�xbuttonMszPyDialog.xbuttonN)r#r$)r2r$)r4r$)
�__name__�
__module__�__qualname__�__doc__rr"r1r3r-r8rrrrrs



rc
s�eZdZdZdddde�fdddd	d
ddd
g
Zgd�Zgd�ZdZ�fdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd �Zd!d"�Z�ZS)#�	bdist_msiz7create a Microsoft Installer (.msi) binary distribution)z
bdist-dir=Nz1temporary directory for creating the distributionz
plat-name=�pz;platform name to embed in generated filenames (default: %s))�	keep-temp�kzPkeep the pseudo-installation tree around after creating the distribution archive)ztarget-version=Nz6require a specific python version on the target system)�no-target-compile�cz/do not compile .py to .pyc on the target system)�no-target-optimize�oz;do not compile .py to .pyo (optimized) on the target system)z	dist-dir=�dz-directory to put final built distributions in)�
skip-buildNz2skip rebuilding everything (for testing/debugging))zinstall-script=NzUbasename of installation script to be run after installation or before deinstallation)zpre-install-script=Nz{Fully qualified filename of a script to be run before any files are installed.  This script need not be in the distribution)r?rArCrF)z2.0z2.1z2.2z2.3z2.4z2.5z2.6z2.7z2.8z2.9z3.0z3.1z3.2z3.3z3.4z3.5z3.6z3.7z3.8z3.9�Xcs$t�j|i|��t�dtd�dS)NzZbdist_msi command is deprecated since Python 3.9, use bdist_wheel (wheel packages) instead�)�superr�warnings�warn�DeprecationWarning)rrr��	__class__rrrs�zbdist_msi.__init__cCsFd|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
dS)Nr)�	bdist_dir�	plat_name�	keep_tempZno_target_compileZno_target_optimize�target_version�dist_dir�
skip_build�install_script�pre_install_script�versions)rrrr�initialize_options�szbdist_msi.initialize_optionscCs�|�dd�|jdur2|�d�j}tj�|d�|_t�}|jsN|j	�
�rN||_|jr�|jg|_|js�|j	�
�r�|j|kr�t
d|f��nt|j�|_|�ddd�|jr�t
d��|jr�|j	jD]}|jtj�|�kr�q�q�t
d|j��d|_dS)	N�bdist)rTrTZmsizMtarget version can only be %s, or the '--skip-build' option must be specified)rSrS)rPrPz5the pre-install-script feature is not yet implementedz(install_script '%s' not found in scripts)�set_undefined_optionsrO�get_finalized_command�
bdist_base�os�path�joinrrR�distribution�has_ext_modulesrWrTr�list�all_versionsrVrU�scripts�basename�install_script_key)rr\Z
short_version�scriptrrr�finalize_options�sH

�������zbdist_msi.finalize_optionscCs�|js|�d�|jddd�}|j|_|j|_d|_|�d�}d|_d|_|j�	�r�|j
}|s~|jslJd��dtjdd	�}d
|j
|f}|�d�}tj�|jd|�|_t�d|j�|��tj�dtj�|jd
��|��tjd=|�|j�|j��}|�|�}tj�|�}tj�|��r0t�|�|jj }|j!}	|	�sJ|j"}	|	�sTd}	|�#�}
dt$|
�j%}|j��}|j
�r�d|j
|f}nd|}t&�'|t(|t&�)�||	�|_*t&�+|j*t,�d|
fg}
|j-�p�|j.}|�r�|
�/d|f�|j0�r|
�/d|j0f�|
�rt1|j*d|
�|�2�|�3�|�4�|�5�|j*�6�t7|jd��rld|j
�pXd|f}|jj8�/|�|j9�s�t:|j|j;d�dS)N�build�installr$)�reinit_subcommandsr�install_libz Should have already checked thisz%d.%drHz.%s-%s�libzinstalling to %sZPURELIB�UNKNOWNz%d.%d.%dzPython %s %sz	Python %sZDistVersionZ
ARPCONTACTZARPURLINFOABOUT�Property�
dist_filesr=�any)�dry_run)<rT�run_command�reinitialize_commandrO�prefix�warn_dir�compile�optimizer`rarR�sys�version_inforPr[r]r^r_�
build_base�	build_libr�info�ensure_finalized�insert�run�mkpathrS�get_fullname�get_installer_filename�abspath�exists�unlink�metadata�author�
maintainer�get_versionr�version�msilibZ
init_databaser	Zgen_uuid�dbZ
add_tablesr
�author_email�maintainer_email�append�urlr�add_find_python�	add_files�add_scripts�add_ui�Commit�hasattrrprQrrr)rrjrlrR�plat_specifierri�fullname�installer_namer�r�r�ZsversionZproduct_name�props�email�tuprrrr��s�




�


�

z
bdist_msi.runc
Cs|j}t�d�}tj�|j�}t||d|dd�}t|ddddddd�}||d	fg}|j	|j
gD]t}d|}d|}	}
d}||j
ur�d
}d}
nd|}d}
t||	||d|
|d�}t||||||
�}|�|||f�q`|��i}|D�]\}}}|g}|�r�|�
�}t�|j�D]�}tj�|j|�}tj�|��rld
|�|�|f}||}
t|||||
|�}|�|�n�|j�s�|�|j|d�||v�r�|�|�}||<||jk�r�|j�r�td|��d||_n*||}t|jd|||j|d|jfg��qq�|��q�|�|�dS)NZ	distfiles�	TARGETDIRZ	SourceDir�PythonZ
Everythingrr$)�	directory�zPython from another locationrHzPython %s from registryz%s|%szMultiple files with name %sz[#%s]Z
DuplicateFile)r�r�ZCABr]r^r�rOrr
rW�
other_versionr�r��pop�listdir�absoluter_�isdirZ
make_short�	componentZstart_component�logicalZadd_filerUrfrr�commit)rr�Zcab�rootdir�root�f�itemsr��targetr.�default�descr"�level�dir�seen�feature�todo�fileZafile�shortZnewdir�keyrrrr�sb

�

��

zbdist_msi.add_filescCs�d}|jD�]v}d|}d|}d|}d|}d|}d|}d|}	d	|}
d
|}d|}tjrld}
nd
}
t|jd|d
|d|
f|d|d|
fg�t|jd||f||fg�t|jd|d|d|df|	d|d|df|
d|d|dfg�t|jd|||f|	||df|
d|d
fg�t|jd|||f|	||df|
d|d
fg�t|jdd|dd|fg�|d7}|dks
J�q
dS)asAdds code to the installer to compute the location of Python.

        Properties PYTHON.MACHINE.X.Y and PYTHON.USER.X.Y will be set from the
        registry for each version of Python.

        Properties TARGETDIRX.Y will be set from PYTHON.USER.X.Y if defined,
        else from PYTHON.MACHINE.X.Y.

        Properties PYTHONX.Y will be set to TARGETDIRX.Y\python.exei�z)SOFTWARE\Python\PythonCore\%s\InstallPathzpython.machine.zpython.user.zPYTHON.MACHINE.zPYTHON.USER.ZPythonFromMachineZPythonFromUserZ	PythonExer��PYTHON�rHZ
RegLocatorNr$Z	AppSearch�CustomActioni3�[�]z]\python.exe�InstallExecuteSequence�InstallUISequence�	Conditionr�rz
NOT TARGETDIR�i�)rWr�ZWin64rr�)r�start�verZinstall_pathZmachine_regZuser_regZmachine_propZ	user_propZmachine_actionZuser_actionZ
exe_actionZtarget_dir_prop�exe_prop�Typerrrr�Ks`�����������zbdist_msi.add_find_pythonc	Cs,|jrjd}|j|jgD]P}d|}d|}t|jd|d||jfg�t|jd|d||fg�|d7}q|j�r(tj�	|j
d	�}t|d
��R}|�d�t|j��}|�|�
��Wd�n1s�0YWd�n1s�0Yt|jdd
t�|�fg�t|jddg�t|jddg�dS)Ni�zinstall_script.r�r��2r�z&Python%s=3r$zpreinstall.batrzrem ="""
%1 %0
exit
"""
�Binary�
PreInstall)r�rHr�N)r�z
NOT Installedi�)rUrWr�rr�rfrVr]r^r_rO�open�write�readr�r�)rr�r�Zinstall_actionr�Zscriptfnr�Zfinrrrr��s6��
	
J���zbdist_msi.add_scriptscCs�
|j}d}}d}d}d}d}d}d}	t|dgd	��t|d
gd��t|dgd
��t|dtj�t|dtj�t|d||||||ddd�}
|
�d�|
jdddd�|
jdddd�|
�ddddddd�|
�ddd dd!dd"�|
j	dddd#�}|�
d$d%�t|d&||||||ddd�}|�d'�|jdddd�|jdddd�|�ddddddd(�|�ddd dd!dd"�|j	dddd#�}|�
d$d%�t|d)||||||ddd�}
|
�d*�|
jdddd�|
jdddd�|
�d+dd,dd!dd"�|
j	dddd#�}|�
d$d-�t|d.||||d/|d0d0d0d1d2�}|�d3dd4d5ddd6�|�d+d!d7d8d!dd9�|�d:d!d;d<ddd=�|�d>d?d!d@d<dAdBdCddd�|jd%dDd%d#�}|�
d$d%�|j	dDd0dDd#�}|�
d$dD�|jd0d%d0d#�}|�
d$d0�t|dEddFd<dGdH|dIdd�}|�dIddJd8dKddL�|�
dMdNdOdPdQddRd��
d$dS�|�
dTdUdOdPdQddVd��
d$dW�|�
dXddOdPdQddYd��
d$dZ�|�
d[d\dOdPdQddd��
d$d]�|�
d^dPdOdPdQddDd��
d$d_�|�
d`dadOdPdQddbd��
d$dc�|�
dddedOdPdQdd0d��
d$df�t|dgddFdhdid|dRdRdR�}|�d:dKddjdkddl�|�
dVdOdmdndoddVdR�}|�
d$d%�|�
dRdpdmdndoddRdV�}|�
d$d-�t|dqddFdhdi||d-d-d-�}|�d:dKddjdkddr�|�
d-dsdmdndodd-d�}|�
d$d%�t|dt||||||ddd�}|�d+dddduddv�|�dw�|�dddxdd!ddy�}|�dd:�|�dzdd{ddkdd�}|�dzd:�|jdddd�|j	d|ddd�|�dd�}|�
d}dg�t|d~||||||d|d|d�}|�d�|�d�ddkdd!dd�|j���|jdddd�|�	d�d�}d}|j
d�d�|d��|j|jgD](}|d7}|j
d�d�|d�||d���q�|j
d�dq|dd��|j
d$d-|d�d��|�dd��}|�
d}dg�|�d�d�dd�ddNdd�dd�d�}|�
d�d��|j}d�|}d�|}|�d�dd5dddd��}|�d�|�|�d�|�|�d�|�|�d�|�|�d�d�dd�dd�dd�|dd|d�}|�d�|�|�d�|�|�d�|�|�d�|�t|d�||||||d�d�d�d1d2�}|�d3dd4d5ddd��|�d+d!d!d8d!dd��|�d:d!d�d<d�dd��|�d�d�d!d�d<d�d�dd�dd�|�d�dbdd���
d$d-�t|d�||||||d�d|d�}|�d��|�d�dd�dhddd�dLd|�	}|�d�dd�d�d!d��|�d�dd�d�d!d��|jdddd�|�	d�d�}|�
d�d�d�d�|j
d$d-d�d��|�dd��}|�
d}dg�t|d�||||||dddd1d2�}|�d3d!dd5ddd��|�d:d�d�ddkdd��|�d�d�d�d�d!dd��|�ddd�|dd!ddy�}|�dd:�|�d�d�d�dNddFd�dd�dd�}|�d�d��|jdd|d1d�|j	d�dd1d�|�dd��
d}dg�t|d�||||||d|d|d�}|�d��|�d�dd�d<d\ddġ|�d�dd�d<d�dd�dLd|�	}|�d�dd�d5dodʡ|�d�dd�d5dod͡|jddd1d�|�	dd�}|�
d�d�d�d��|�
d�d�d�d4�|�
d�d�d�dB�|�
d�d�d�dա|�
d�d�d�dء|�
d�d�d�dڡ|�
d�d�d�dܡ|�
d�d�d�dݡ|�
d$d-d�d!�|�ddš�
d}dg�dS)�Nr�iri,z[ProductName] Setupr%r$� ro))Z
DefaultUIFont�DlgFont8)ZErrorDialog�ErrorDlg)Z	Progress1ZInstall)Z	Progress2Zinstalls)�MaintenanceForm_Action�Repair)�
WhichUsers�ALLZ	TextStyle))r��Tahoma�	Nr)ZDlgFontBold8r��Nr$)Z
VerdanaBold10�VerdanarNr$)ZVerdanaRed9r�r��rr�))�
PrepareDlgz(Not Privileged or Windows9x or Installed�)�
WhichUsersDlgz.Privileged and not Windows9x and not Installed�)�SelectFeaturesDlgz
Not Installedi�)�MaintenanceTypeDlgz,Installed AND NOT RESUME AND NOT Preselectedi�)�ProgressDlgNi�
ActionText�UITextZ
FatalErrorZFinishz)[ProductName] Installer ended prematurelyz< Backr)r/r2r#ZDescription1r�Fr�Pr!z�[ProductName] setup ended prematurely because of an error.  Your system has not been modified.  To install this program at a later time, please run the installation again.ZDescription2��z.Click the Finish button to exit the Installer.)r.Z	EndDialogZExitZUserExitz'[ProductName] Installer was interruptedz�[ProductName] setup was interrupted.  Your system has not been modified.  To install this program at a later time, please run the installation again.Z
ExitDialogz&Completing the [ProductName] Installer�Description���ReturnZ
FilesInUse��RetryF)Zbitmapr���z{\DlgFontBold8}Files in Use�iz8Some files that need to be updated are currently in use.�Text�7iJz�The following applications are using files that need to be updated by this setup. Close these applications and then click Retry to continue the installation or Cancel to exit it.�ListZListBox�k��ZFileInUseProcess�Ignorer�r�eiZ	ErrorTextr��0r��N�x�H�Q�ZNoZErrorNo�Y��ZYesZErrorYes�AZAbortZ
ErrorAbort�C�*ZErrorCancel�IZErrorIgnore�O�ZOkZErrorOk�R��Z
ErrorRetryZ	CancelDlgi�U���z;Are you sure you want to cancel [ProductName] installation?�9r(r)�ZWaitForCostingDlgzRPlease wait while the installer finishes determining your disk space requirements.�fr��(zOPlease wait while the Installer prepares to guide you through the installation.z&Welcome to the [ProductName] Installer�nzPondering...Z
ActionData�r4ZSpawnDialogr�zSelect Python InstallationsZHintz9Select the Python locations where %s should be installed.zNext >z[TARGETDIR]z[SourceDir])Zorderingz
[TARGETDIR%s]z FEATURE_SELECTED AND &Python%s=3ZSpawnWaitDialogrHZFeaturesZ
SelectionTreer ZFEATUREZPathEditz[FEATURE_SELECTED]�1z!FEATURE_SELECTED AND &Python%s<>3ZOtherz$Provide an alternate Python locationZEnableZShowZDisableZHide���r�ZDiskCostDlg�OKz&{\DlgFontBold8}Disk Space RequirementszFThe disk space required for the installation of the selected features.�5aThe highlighted volumes (if any) do not have enough disk space available for the currently selected features.  You can either remove some files from the highlighted volumes, or choose to install less features onto local drive(s), or select different destination drive(s).Z
VolumeListZVolumeCostList�d�iz{120}{70}{70}{70}{70}g�?r�ZAdminInstallzGSelect whether to install [ProductName] for all users of this computer.r�r��zInstall for all usersZJUSTME�zInstall just for mez
[ALLUSERS]zWhichUsers="ALL"r�z({\DlgFontBold8}[Progress1] [ProductName]�#�AzYPlease wait while the Installer [Progress2] [ProductName]. This may take several minutes.ZStatusLabelzStatus:ZProgressBariz
Progress doneZSetProgress�Progressr�z)Welcome to the [ProductName] Setup WizardZBodyText�?z:Select whether you want to repair or remove [ProductName].ZRepairRadioGroup�lr�r�r�z&Repair [ProductName]ZRemoverzRe&move [ProductName]z[REINSTALL]zMaintenanceForm_Action="Repair"z[Progress1]Z	Repairingz[Progress2]ZrepairsZ	Reinstallr�z[REMOVE]zMaintenanceForm_Action="Remove"�ZRemoving�Zremoves�
�z MaintenanceForm_Action<>"Change")r�rrr�r�rr"r1r3r-�event�controlrr+�mappingr`r�rWr��	conditionr8Z
radiogroup�add)rr��x�yrrr"�modalZmodelessZtrack_disk_space�fatalrBZ	user_exitZexit_dialogZinuse�errorr3Zcosting�prepZseldlg�orderr�r�Zinstall_other_condZdont_install_other_condZcostZ
whichusers�g�progressZmaintrrrr��s��
�	�
�
���
���
�������       ������
�
���
��������
�
������
��zbdist_msi.add_uicCs<|jrd||j|jf}nd||jf}tj�|j|�}|S)Nz%s.%s-py%s.msiz	%s.%s.msi)rRrPr]r^r_rS)rr��	base_namer�rrrr��s�z bdist_msi.get_installer_filename)r9r:r;�descriptionr�user_options�boolean_optionsrcr�rrXrhr�r�r�r�r�r��
__classcell__rrrMrr=Us<���
([66&@r=)r<r]ryrJ�distutils.corer�distutils.dir_utilr�distutils.sysconfigrZdistutils.versionr�distutils.errorsr�distutils.utilr�	distutilsrr�r	r
rrr
rrrr=rrrr�<module>s>site-packages/setuptools/_distutils/command/__pycache__/build.cpython-39.pyc000064400000007617150564745010023270 0ustar00a

(�Re��@sTdZddlZddlZddlmZddlmZddlmZdd�Z	Gdd	�d	e�Z
dS)
zBdistutils.command.build

Implements the Distutils 'build' command.�N)�Command)�DistutilsOptionError)�get_platformcCsddlm}|�dS)Nr��show_compilers)�distutils.ccompilerrr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/build.pyrsrc@s�eZdZdZdddddddd	d
e�fddd
ddgZddgZdddefgZdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd �Zd!d"�Zd#efd$e
fd%efd&efgZdS)'�buildz"build everything needed to install)zbuild-base=�bz base directory for build library)zbuild-purelib=Nz2build directory for platform-neutral distributions)zbuild-platlib=Nz3build directory for platform-specific distributions)z
build-lib=NzWbuild directory for all distribution (defaults to either build-purelib or build-platlib)zbuild-scripts=Nzbuild directory for scripts)zbuild-temp=�tztemporary build directoryz
plat-name=�pz6platform name to build for, if supported (default: %s))z	compiler=�czspecify the compiler type)z	parallel=�jznumber of parallel build jobs)�debug�gz;compile extensions and libraries with debugging information)�force�fz2forcibly build everything (ignore file timestamps))zexecutable=�ez5specify final destination interpreter path (build.py)rrz
help-compilerNzlist available compilerscCsLd|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_dS)Nr
r)�
build_base�
build_purelib�
build_platlib�	build_lib�
build_temp�
build_scripts�compiler�	plat_namerr�
executable�parallel��selfrrr	�initialize_options8szbuild.initialize_optionscCsf|jdurt�|_ntjdkr&td��d|jgtjdd��R}ttd�rT|d7}|jdurptj	�
|jd�|_|jdur�tj	�
|jd|�|_|j
dur�|j��r�|j|_
n|j|_
|jdur�tj	�
|jd|�|_|jdu�rtj	�
|jd	tjdd��|_|jdu�r&tj�r&tj	�tj�|_t|jt��rbzt|j�|_Wnt�y`td
��Yn0dS)N�ntzW--plat-name only supported on Windows (try using './configure --help' on your platform)z	.%s-%d.%d��gettotalrefcountz-pydebug�lib�tempz
scripts-%d.%dzparallel should be an integer)rr�os�namer�sys�version_info�hasattrr�path�joinrrr�distribution�has_ext_modulesrrr�normpath�
isinstancer�str�int�
ValueError)r �plat_specifierrrr	�finalize_optionsHsD


�



�




�
�zbuild.finalize_optionscCs|��D]}|�|�qdS�N)�get_sub_commands�run_command)r �cmd_namerrr	�run�sz	build.runcCs
|j��Sr7)r.�has_pure_modulesrrrr	r<�szbuild.has_pure_modulescCs
|j��Sr7)r.�has_c_librariesrrrr	r=�szbuild.has_c_librariescCs
|j��Sr7)r.r/rrrr	r/�szbuild.has_ext_modulescCs
|j��Sr7)r.�has_scriptsrrrr	r>�szbuild.has_scripts�build_py�
build_clib�	build_extr)�__name__�
__module__�__qualname__�descriptionr�user_options�boolean_optionsr�help_optionsr!r6r;r<r=r/r>�sub_commandsrrrr	r
sF�����8�r
)�__doc__r)r'�distutils.corer�distutils.errorsr�distutils.utilrrr
rrrr	�<module>ssite-packages/setuptools/_distutils/command/__pycache__/install_headers.cpython-39.pyc000064400000003430150564745010025317 0ustar00a

(�Re�@s$dZddlmZGdd�de�ZdS)z�distutils.command.install_headers

Implements the Distutils 'install_headers' command, to install C/C++ header
files to the Python include directory.�)�Commandc@sFeZdZdZddgZdgZdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dS)�install_headerszinstall C/C++ header files)zinstall-dir=�dz$directory to install header files to)�force�fz-force installation (overwrite existing files)rcCsd|_d|_g|_dS)Nr)�install_dirr�outfiles��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install_headers.py�initialize_optionssz"install_headers.initialize_optionscCs|�ddd�dS)N�install)rr)rr)�set_undefined_optionsr	rrr�finalize_optionss�z install_headers.finalize_optionscCsH|jj}|sdS|�|j�|D]"}|�||j�\}}|j�|�q dS�N)�distribution�headers�mkpathr�	copy_filer�append)r
r�header�out�_rrr�run!szinstall_headers.runcCs|jjp
gSr)rrr	rrr�
get_inputs+szinstall_headers.get_inputscCs|jSr)rr	rrr�get_outputs.szinstall_headers.get_outputsN)�__name__�
__module__�__qualname__�description�user_options�boolean_optionsr
rrrrrrrrr
s�
rN)�__doc__�distutils.corerrrrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/check.cpython-39.pyc000064400000011627150564745010023242 0ustar00a

(�Re�@s�dZddlmZddlmZzHddlmZddlmZddl	m
Z
ddl	mZGdd	�d	e�Zd
Z
WneyzdZ
Yn0Gdd
�d
e�ZdS)zCdistutils.command.check

Implements the Distutils 'check' command.
�)�Command)�DistutilsSetupError)�Reporter)�Parser)�frontend)�nodesc@seZdZd	dd�Zdd�ZdS)
�SilentReporterNr�ascii�replacec
Cs"g|_t�||||||||�dS�N)�messagesr�__init__)�self�source�report_level�
halt_level�stream�debug�encoding�
error_handler�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/check.pyr
s�zSilentReporter.__init__cOs8|j�||||f�tj|g|�R||j|d�|��S)N)�level�type)r�appendr�system_messageZlevels)rr�message�children�kwargsrrrrs���zSilentReporter.system_message)Nrr	r
)�__name__�
__module__�__qualname__r
rrrrrrs�
rTFc@s\eZdZdZdZgd�Zgd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dS)�checkz6This command checks the meta-data of the package.
    z"perform some checks on the package))�metadata�mzVerify meta-data)�restructuredtext�rzEChecks if long string meta-data syntax are reStructuredText-compliant)�strict�sz(Will exit with an error if a check fails)r#r%r'cCsd|_d|_d|_d|_dS)z Sets default values for options.r�N)r%r#r'�	_warnings�rrrr�initialize_options0szcheck.initialize_optionscCsdSrrr+rrr�finalize_options7szcheck.finalize_optionscCs|jd7_t�||�S)z*Counts the number of warnings that occurs.r))r*r�warn)r�msgrrrr.:sz
check.warncCsL|jr|��|jr0tr"|��n|jr0td��|jrH|jdkrHtd��dS)zRuns the command.zThe docutils package is needed.rzPlease correct your package.N)r#�check_metadatar%�HAS_DOCUTILS�check_restructuredtextr'rr*r+rrr�run?s
z	check.runcCs�|jj}g}dD]"}t||�r(t||�s|�|�q|rL|�dd�|��|jrd|js�|�d�n"|j	r||j
s�|�d�n
|�d�dS)aEnsures that all required elements of meta-data are supplied.

        Required fields:
            name, version, URL

        Recommended fields:
            (author and author_email) or (maintainer and maintainer_email))

        Warns if any are missing.
        )�name�version�urlzmissing required meta-data: %sz, zNmissing meta-data: if 'author' supplied, 'author_email' should be supplied toozVmissing meta-data: if 'maintainer' supplied, 'maintainer_email' should be supplied toozkmissing meta-data: either (author and author_email) or (maintainer and maintainer_email) should be suppliedN)�distributionr#�hasattr�getattrrr.�join�author�author_email�
maintainer�maintainer_email)rr#�missing�attrrrrr0Oszcheck.check_metadatacCsX|j��}|�|�D]>}|d�d�}|dur8|d}nd|d|f}|�|�qdS)z4Checks if the long string fields are reST-compliant.����lineNr)z%s (line %s))r7�get_long_description�_check_rst_data�getr.)r�data�warningrBrrrr2ps

zcheck.check_restructuredtextc
Cs�|jjp
d}t�}tjtfd���}d|_d|_d|_t	||j
|j|j|j
|j|jd�}tj|||d�}|�|d�z|�||�Wn<ty�}z$|j�dd|d	if�WYd}~n
d}~00|jS)
z8Returns warnings when the provided data doesn't compile.zsetup.py)�
components�N)rrrr)rrAz!Could not finish the parsing: %s.�)r7�script_namerr�OptionParser�get_default_valuesZ	tab_widthZpep_referencesZrfc_referencesrrrZwarning_streamrZerror_encodingZerror_encoding_error_handlerr�documentZnote_source�parse�AttributeErrorrr)rrF�source_path�parser�settings�reporterrN�errrrD{s.��zcheck._check_rst_dataN)rr r!�__doc__�description�user_options�boolean_optionsr,r-r.r3r0r2rDrrrrr"#s!r"N)rV�distutils.corer�distutils.errorsrZdocutils.utilsrZdocutils.parsers.rstrZdocutilsrrrr1�	Exceptionr"rrrr�<module>s
site-packages/setuptools/_distutils/command/__pycache__/install_egg_info.cpython-39.pyc000064400000006066150564745010025471 0ustar00a

(�Re+
�@sddZddlmZddlmZmZddlZddlZddlZGdd�de�Z	dd�Z
d	d
�Zdd�ZdS)
z�distutils.command.install_egg_info

Implements the Distutils 'install_egg_info' command, for installing
a package's PKG-INFO metadata.�)�Command)�log�dir_utilNc@s:eZdZdZdZdgZdd�Zdd�Zdd	�Zd
d�Z	dS)
�install_egg_infoz)Install an .egg-info file for the packagez8Install package's PKG-INFO metadata as an .egg-info file)zinstall-dir=�dzdirectory to install tocCs
d|_dS�N)�install_dir��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install_egg_info.py�initialize_optionssz#install_egg_info.initialize_optionscCsd|�dd�dtt|j����tt|j����gtjdd��R}t	j
�|j|�|_
|j
g|_dS)N�install_lib)rrz%s-%s-py%d.%d.egg-info�)�set_undefined_options�to_filename�	safe_name�distribution�get_name�safe_version�get_version�sys�version_info�os�path�joinr�target�outputs)r
�basenamerrr�finalize_optionss��z!install_egg_info.finalize_optionscCs�|j}tj�|�r0tj�|�s0tj||jd�nNtj�|�rV|�	tj
|jfd|�n(tj�|j�s~|�	tj|jfd|j�t
�d|�|js�t|ddd��}|jj�|�Wd�n1s�0YdS)N)�dry_runz	Removing z	Creating z
Writing %s�wzUTF-8)�encoding)rrr�isdir�islinkr�remove_treer �exists�execute�unlinkr�makedirsr�info�openr�metadata�write_pkg_file)r
r�frrr�run s�zinstall_egg_info.runcCs|jSr)rr	rrr�get_outputs.szinstall_egg_info.get_outputsN)
�__name__�
__module__�__qualname__�__doc__�description�user_optionsr
rr/r0rrrrrs�
rcCst�dd|�S)z�Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    �[^A-Za-z0-9.]+�-)�re�sub��namerrrr6srcCs|�dd�}t�dd|�S)z�Convert an arbitrary string to a standard version string

    Spaces become dots, and all other non-alphanumeric characters become
    dashes, with runs of multiple dashes condensed to a single dash.
    � �.r7r8)�replacer9r:)�versionrrrr>srcCs|�dd�S)z|Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    r8�_)r?r;rrrrHsr)
r4�
distutils.cmdr�	distutilsrrrrr9rrrrrrrr�<module>s+
site-packages/setuptools/_distutils/command/__pycache__/clean.cpython-39.pyc000064400000004214150564745010023241 0ustar00a

(�Re�
�@sDdZddlZddlmZddlmZddlmZGdd�de�ZdS)zBdistutils.command.clean

Implements the Distutils 'clean' command.�N)�Command)�remove_tree)�logc@s6eZdZdZgd�ZdgZdd�Zdd�Zdd	�Zd
S)�cleanz-clean up temporary files from 'build' command))zbuild-base=�bz2base build directory (default: 'build.build-base'))z
build-lib=Nz<build directory for all modules (default: 'build.build-lib'))zbuild-temp=�tz7temporary build directory (default: 'build.build-temp'))zbuild-scripts=Nz<build directory for scripts (default: 'build.build-scripts'))zbdist-base=Nz+temporary directory for built distributions)�all�az7remove all build output, not just temporary by-productsrcCs(d|_d|_d|_d|_d|_d|_dS)N)�
build_base�	build_lib�
build_temp�
build_scripts�
bdist_baser��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/clean.py�initialize_options szclean.initialize_optionscCs"|�ddddd�|�dd�dS)N�build)r
r
)rr)r
r
)rr�bdist)rr)�set_undefined_optionsrrrr�finalize_options(s��zclean.finalize_optionscCs�tj�|j�r t|j|jd�nt�d|j�|jrr|j	|j
|jfD],}tj�|�rdt||jd�qDt�d|�qD|js�zt�
|j�t�d|j�Wnty�Yn0dS)N)�dry_runz%'%s' does not exist -- can't clean itz
removing '%s')�os�path�existsrrrr�debugrrrr
�warn�rmdirr
�info�OSError)r�	directoryrrr�run1s*���z	clean.runN)	�__name__�
__module__�__qualname__�description�user_options�boolean_optionsrrr"rrrrrs	r)	�__doc__r�distutils.corer�distutils.dir_utilr�	distutilsrrrrrr�<module>s
site-packages/setuptools/_distutils/command/__pycache__/upload.cpython-39.pyc000064400000012275150564745010023451 0ustar00a

(�Re��@s�dZddlZddlZddlZddlmZddlmZmZm	Z	ddl
mZddlm
Z
mZddlmZddlmZdd	lmZeed
d�eedd�eedd�d
�ZGdd�de�ZdS)zm
distutils.command.upload

Implements the Distutils 'upload' subcommand (upload package to a package
index).
�N)�standard_b64encode)�urlopen�Request�	HTTPError)�urlparse)�DistutilsError�DistutilsOptionError)�
PyPIRCCommand)�spawn)�log�md5�sha256�blake2b)Z
md5_digestZ
sha256_digestZblake2_256_digestc@sJeZdZdZejddgZejdgZdd�Zdd�Zd	d
�Z	dd�Z
d
S)�uploadzupload binary package to PyPI)�sign�szsign files to upload using gpg)z	identity=�izGPG identity used to sign filesrcCs,t�|�d|_d|_d|_d|_d|_dS)N�rF)r	�initialize_options�username�password�
show_responser�identity)�self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/upload.pyr)s
zupload.initialize_optionscCsrt�|�|jr|jstd��|��}|ikrV|d|_|d|_|d|_|d|_	|jsn|j
jrn|j
j|_dS)Nz.Must use --sign for --identity to have meaningrr�
repository�realm)r	�finalize_optionsrrr�_read_pypircrrrr�distribution)r�configrrrr1s
�



zupload.finalize_optionscCs:|jjsd}t|��|jjD]\}}}|�|||�qdS)NzHMust create and upload files in one command (e.g. setup.py sdist upload))r �
dist_filesr�upload_file)r�msg�command�	pyversion�filenamerrr�runCs
z
upload.runc"Cs
t|j�\}}}}}}	|s"|s"|	r0td|j��|dvrDtd|��|jr|ddd|g}
|jrnd|jg|
dd�<t|
|jd	�t|d
�}z|��}W|�	�n
|�	�0|j
j}
dd|
��|
�
�tj�|�|f||d
|
��|
��|
��|
��|
��|
��|
��|
��|
��|
��|
��|
��|
��d�}d|d<t��D]B\}}|du�rL�q6z||�� �||<Wnt!�ytYn0�q6|j�r�t|dd
��,}tj�|�d|��f|d<Wd�n1�s�0Y|j"d|j#�$d�}dt%|��&d�}d}d|�$d�}|d}t'�(�}|��D]�\}}d|}t)|t*��sB|g}|D]j}t+|�t,u�rr|d|d7}|d}nt-|��$d�}|�.|�|�.|�$d��|�.d�|�.|��qF�q |�.|�|�/�}d||jf}|�0|t1j2�d |t-t3|��|d!�}t4|j||d"�}zt5|�}|�6�}|j7}Wnjt8�yX} z| j9}| j7}WYd} ~ nBd} ~ 0t:�y�} z |�0t-| �t1j;��WYd} ~ n
d} ~ 00|d#k�r�|�0d$||ft1j2�|j<�r|�=|�}!d%�>d&|!d&f�}|�0|t1j2�n"d'||f}|�0|t1j;�t?|��dS)(NzIncompatible url %s)�http�httpszunsupported schema Zgpgz
--detach-signz-az--local-user�)�dry_run�rbZfile_upload�1z1.0)z:actionZprotocol_version�name�version�content�filetyper&�metadata_version�summaryZ	home_page�author�author_email�license�description�keywords�platform�classifiers�download_url�provides�requires�	obsoletesr�commentz.ascZ
gpg_signature�:�asciizBasic z3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s
--s--
z+
Content-Disposition: form-data; name="%s"z; filename="%s"r�zutf-8s

zSubmitting %s to %sz multipart/form-data; boundary=%s)zContent-typezContent-length�
Authorization)�data�headers��zServer response (%s): %s�
zK---------------------------------------------------------------------------zUpload failed (%s): %s)@rr�AssertionErrorrrr
r,�open�read�closer �metadata�get_name�get_version�os�path�basename�get_description�get_url�get_contact�get_contact_email�get_licence�get_long_description�get_keywords�
get_platforms�get_classifiers�get_download_url�get_provides�get_requires�
get_obsoletes�_FILE_CONTENT_DIGESTS�items�	hexdigest�
ValueErrorrr�encoder�decode�io�BytesIO�
isinstance�list�type�tuple�str�write�getvalue�announcer�INFO�lenrr�getcoder$r�code�OSError�ERRORr�_read_pypi_response�joinr)"rr%r&r'Zschema�netloc�url�params�query�	fragmentsZgpg_args�fr1�metarEZdigest_name�digest_cons�	user_pass�auth�boundaryZsep_boundaryZend_boundary�body�key�value�titler$rF�request�result�status�reason�e�textrrrr#Ks���

� 

�(




��

�
zupload.upload_fileN)�__name__�
__module__�__qualname__r8r	�user_options�boolean_optionsrrr(r#rrrrrs�r)�__doc__rPrf�hashlib�base64r�urllib.requestrrr�urllib.parser�distutils.errorsrr�distutils.corer	�distutils.spawnr
�	distutilsr�getattrr`rrrrr�<module>s


�site-packages/setuptools/_distutils/command/__pycache__/register.cpython-39.pyc000064400000020557150564745010024013 0ustar00a

(�Re�-�@sddZddlZddlZddlZddlZddlmZddlm	Z	ddl
TddlmZGdd�de	�Z
dS)	zhdistutils.command.register

Implements the Distutils 'register' command (register with the repository).
�N)�warn)�
PyPIRCCommand)�*)�logc@s�eZdZdZejddgZejgd�Zddd�fgZdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zddd�ZdS)�registerz7register the distribution with the Python package index)�list-classifiersNz list the valid Trove classifiers)�strictNzBWill stop the registering if the meta-data are not fully compliant)�verifyrr�checkcCsdS)NT���selfrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/register.py�<lambda>�zregister.<lambda>cCst�|�d|_d|_dS)Nr)r�initialize_options�list_classifiersrrrrrrs
zregister.initialize_optionscCs*t�|�d|jfdd�}||jjd<dS)Nr)r�)r�restructuredtextr
)r�finalize_optionsr�distribution�command_options)r
�
check_optionsrrrr$s

�zregister.finalize_optionscCsT|��|��|��D]}|�|�q|jr8|��n|jrH|��n|��dS)N)	r�_set_config�get_sub_commands�run_command�dry_run�verify_metadatar�classifiers�
send_metadata)r
�cmd_namerrr�run+s

zregister.runcCs8tdt�|j�d�}|��|j|_d|_|��dS)zDeprecated API.zddistutils.command.register.check_metadata is deprecated,               use the check command insteadr
rN)r�PendingDeprecationWarningr�get_command_obj�ensure_finalizedrrr!)r
r
rrr�check_metadata:s�zregister.check_metadatacCsz|��}|ikr@|d|_|d|_|d|_|d|_d|_n6|jd|jfvr^td|j��|jdkrp|j|_d|_d	S)
z: Reads the configuration file and set attributes.
        �username�password�
repository�realmT�pypiz%s not found in .pypircFN)�_read_pypircr&r'r(r)�
has_config�DEFAULT_REPOSITORY�
ValueError)r
�configrrrrDs




zregister._set_configcCs*|jd}tj�|�}t�|�|��dS)z8 Fetch the list of classifiers from the server.
        z?:action=list_classifiersN)r(�urllib�request�urlopenr�info�_read_pypi_response)r
�url�responserrrrUs
zregister.classifierscCs&|�|�d��\}}t�d||�dS)zF Send the metadata to the package index server to be checked.
        r	�Server response (%s): %sN)�post_to_server�build_post_datarr3)r
�code�resultrrrr\szregister.verify_metadatac
Cs�|jrd}|j}|j}nd}d}}d��}||vrd|�dtj�t�}|sRd}q,||vr,td�q,|dk�rl|s|td�}qn|s�t	�	d�}q|t
j��}t
j
�|j�d	}|�|j|||�|�|�d
�|�\}}|�d||ftj�|dk�r�|j�r||j_nf|�d
tj�|�d|��tj�d}|��dv�rNtd�}|�s*d}�q*|��dk�r�|�||��nl|dk�r�ddi}	d|	d<|	d<|	d<d|	d<|	d�s�td�|	d<�q�|	d|	dk�r0|	d�s�t	�	d�|	d<�q�|	d�st	�	d�|	d<�q�|	d|	dk�r�d|	d<d|	d<td��q�|	d�sJtd�|	d<�q0|�|	�\}}|dk�rrt�d||�nt�d�t�d �nP|d!k�r�dd"i}	d|	d<|	d�s�td#�|	d<�q�|�|	�\}}t�d||�dS)$a_ Send the metadata to the package index server.

            Well, do the following:
            1. figure who the user is, and then
            2. send the data as a Basic auth'ed POST.

            First we try to read the username/password from $HOME/.pypirc,
            which is a ConfigParser-formatted file with a section
            [distutils] containing username and password entries (both
            in clear text). Eg:

                [distutils]
                index-servers =
                    pypi

                [pypi]
                username: fred
                password: sekrit

            Otherwise, to figure who the user is, we offer the user three
            choices:

             1. use existing login,
             2. register as a new user, or
             3. set the password to a random string and email the user.

        �1�x�z1 2 3 4z�We need to know who you are, so please choose either:
 1. use your existing login,
 2. register as a new user,
 3. have the server generate a new password for you (and email it to you), or
 4. quit
Your selection [default 1]: z&Please choose one of the four options!z
Username: z
Password: r�submitr7��zAI can store your PyPI login so future submissions will be faster.z (the login will be stored in %s)�XZynzSave your login (y/N)?�n�y�2�:action�user�namer'�emailNZconfirmz
 Confirm: z!Password and confirm don't match!z
   EMail: z"You will receive an email shortly.z7Follow the instructions in it to complete registration.�3Zpassword_resetzYour email address: )r,r&r'�split�announcer�INFO�input�print�getpassr0r1�HTTPPasswordMgr�parse�urlparser(�add_passwordr)r8r9r�_get_rc_file�lower�
_store_pypircr3)
r
�choicer&r'�choices�auth�hostr:r;�datarrrrcs��



��

���








zregister.send_metadatacCs�|jj}|d|��|��|��|��|��|��|��|�	�|�
�|��|��|�
�|��|��|��d�}|ds�|ds�|dr�d|d<|S)Nz1.0)rE�metadata_versionrG�version�summaryZ	home_page�author�author_email�license�description�keywords�platformr�download_url�provides�requires�	obsoletesrfrgrhz1.1r\)r�metadata�get_name�get_version�get_description�get_url�get_contact�get_contact_email�get_licence�get_long_description�get_keywords�
get_platforms�get_classifiers�get_download_url�get_provides�get_requires�
get_obsoletes)r
�action�metar[rrrr9�s,�zregister.build_post_dataNc
Cs�d|vr$|�d|d|jftj�d}d|}|d}t��}|��D]~\}}t|�tg�td�fvrn|g}|D]R}t|�}|�	|�|�	d|�|�	d�|�	|�|rr|d	d
krr|�	d�qrqH|�	|�|�	d�|�
��d�}d
|tt|��d�}	t
j�|j||	�}
t
j�t
jj|d��}d}z|�|
�}Wn|t
jj�y�}
z*|j�rb|
j��}|
j|
jf}WYd}
~
nTd}
~
0t
jj�y�}
zdt|
�f}WYd}
~
n d}
~
00|j�r�|�|�}d}|j�r�d�d|df�}|�|tj�|S)zC Post a query to the server, and return a string response.
        rGzRegistering %s to %sz3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254z
--z--rz*
Content-Disposition: form-data; name="%s"z

����
�
zutf-8z/multipart/form-data; boundary=%s; charset=utf-8)zContent-typezContent-length)�password_mgrr>Ni�)r@�OKzK---------------------------------------------------------------------------)rKr(rrL�io�StringIO�items�type�str�write�getvalue�encode�lenr0r1�Request�build_opener�HTTPBasicAuthHandler�open�error�	HTTPError�
show_response�fp�readr:�msg�URLErrorr4�join)r
r[rY�boundaryZsep_boundaryZend_boundary�body�key�value�headers�req�openerr;�er�rrrr8�s^��





��
 "
zregister.post_to_server)N)�__name__�
__module__�__qualname__rbr�user_options�boolean_options�sub_commandsrrr!r%rrrrr9r8rrrrrs"�
zr)�__doc__rOr��urllib.parser0�urllib.request�warningsr�distutils.corer�distutils.errors�	distutilsrrrrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/bdist.cpython-39.pyc000064400000007177150564745010023277 0ustar00a

(�Re��@sHdZddlZddlmZddlTddlmZdd�ZGdd	�d	e�ZdS)
zidistutils.command.bdist

Implements the Distutils 'bdist' command (create a built [binary]
distribution).�N)�Command)�*)�get_platformcCsPddlm}g}tjD]"}|�d|dtj|df�q||�}|�d�dS)zFPrint list of available formats (arguments to "--format" option).
    r)�FancyGetopt�formats=N�z'List of available distribution formats:)�distutils.fancy_getoptr�bdist�format_commands�append�format_command�
print_help)r�formats�formatZpretty_printer�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/bdist.py�show_formatss
�rc
@s�eZdZdZdddde�fdddd	d
gZdgZdd
defgZdZ	ddd�Z
gd�Zdddddddddd�	Zdd�Z
dd �Zd!d"�Zd
S)#r	z$create a built (binary) distribution)zbdist-base=�bz4temporary directory for creating built distributionsz
plat-name=�pz;platform name to embed in generated filenames (default: %s))rNz/formats for distribution (comma-separated list))z	dist-dir=�dz=directory to put final built distributions in [default: dist])�
skip-buildNz2skip rebuilding everything (for testing/debugging))zowner=�uz@Owner name used when creating a tar file [default: current user])zgroup=�gzAGroup name used when creating a tar file [default: current group]rzhelp-formatsNz$lists available distribution formats)�	bdist_rpm�gztar�zip)�posix�nt)	Zrpmr�bztar�xztar�ztar�tarZwininstrZmsi)rzRPM distribution)�
bdist_dumbzgzip'ed tar file)r"zbzip2'ed tar file)r"zxz'ed tar file)r"zcompressed tar file)r"ztar file)�
bdist_wininstzWindows executable installer)r"zZIP file)Z	bdist_msizMicrosoft InstallercCs.d|_d|_d|_d|_d|_d|_d|_dS)Nr)�
bdist_base�	plat_namer�dist_dir�
skip_build�group�owner)�selfrrr�initialize_optionsQszbdist.initialize_optionscCs�|jdur(|jrt�|_n|�d�j|_|jdurT|�d�j}tj�|d|j�|_|�	d�|j
dur�z|jtjg|_
Wn t
y�tdtj��Yn0|jdur�d|_dS)N�buildzbdist.rz;don't know how to create built distributions on platform %s�dist)r%r'r�get_finalized_commandr$�
build_base�os�path�join�ensure_string_listr�default_format�name�KeyError�DistutilsPlatformErrorr&)r*r/rrr�finalize_optionsZs*


�

��

zbdist.finalize_optionsc	Cs�g}|jD]<}z|�|j|d�Wq
tyDtd|��Yq
0q
tt|j��D]h}||}|�|�}||jvr�|j||_	|dkr�|j
|_
|j|_|||dd�vr�d|_|�
|�qVdS)Nrzinvalid format '%s'r"r)rrrr6�DistutilsOptionError�range�len�reinitialize_command�no_format_optionrr)r(Z	keep_temp�run_command)r*�commandsr�i�cmd_nameZsub_cmdrrr�runvs"


z	bdist.run)�__name__�
__module__�__qualname__�descriptionr�user_options�boolean_optionsr�help_optionsr=r4r
rr+r8rBrrrrr	sH�������
	r	)	�__doc__r0�distutils.corer�distutils.errors�distutils.utilrrr	rrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/build_scripts.cpython-39.pyc000064400000007742150564745010025036 0ustar00a

(�ReK�@s�dZddlZddlZddlmZddlmZddlmZddl	m
Z
ddlmZddlm
Z
ddlZe�d	�ZGd
d�de�ZdS)zRdistutils.command.build_scripts

Implements the Distutils 'build_scripts' command.�N)�ST_MODE)�	sysconfig)�Command)�newer)�convert_path)�logs^#!.*python[0-9.]*([ 	].*)?$c@sFeZdZdZgd�ZdgZdd�Zdd�Zdd	�Zd
d�Z	dd
�Z
dS)�
build_scriptsz("build" scripts (copy and fixup #! line)))z
build-dir=�dzdirectory to "build" (copy) to)�force�fz1forcibly build everything (ignore file timestamps)zexecutable=�ez*specify final destination interpreter pathr
cCs"d|_d|_d|_d|_d|_dS�N)�	build_dir�scriptsr
�
executable�outfiles��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/build_scripts.py�initialize_optionss
z build_scripts.initialize_optionscCs|�dddd�|jj|_dS)N�build)rr)r
r
)rr)�set_undefined_options�distributionrrrrr�finalize_options%s�zbuild_scripts.finalize_optionscCs|jSr
)rrrrr�get_source_files,szbuild_scripts.get_source_filescCs|js
dS|��dSr
)r�copy_scriptsrrrr�run/szbuild_scripts.runc	Cs�|�|j�g}g}|jD�],}d}t|�}tj�|jtj�|��}|�|�|j	slt
||�slt�d|�qzt
|d�}Wnty�|js��d}YnX0t�|j�\}}|�d�|��}	|	s�|�d|�qt�|	�}
|
r�d}|
�d�p�d	}|�r$t�d
||j�|�|�|j�stj�s(|j}n(tj�t�d�dt�d
�t�d�f�}t�|�}d||d}
z|
�d�Wn"t�y�t d�!|
���Yn0z|
�|�Wn$t�y�t d�!|
|���Yn0t
|d��(}|�"|
�|�#|�$��Wd�n1�s
0Y|�rH|�%�q|�r2|�%�|�|�|�&||�qtj'dk�r�|D]`}|j�rtt�d|�nDt�(|�t)d@}|dBd@}||k�rZt�d|||�t�*||��qZ||fS)a"Copy each script listed in 'self.scripts'; if it's marked as a
        Python script in the Unix way (first line matches 'first_line_re',
        ie. starts with "\#!" and contains "python"), then adjust the first
        line to refer to the current Python interpreter as we copy.
        Fznot copying %s (up-to-date)�rbNrz%s is an empty file (skipping)T��zcopying and adjusting %s -> %s�BINDIRz
python%s%s�VERSION�EXEs#!�
zutf-8z.The shebang ({!r}) is not decodable from utf-8zAThe shebang ({!r}) is not decodable from the script encoding ({})�wb�posixzchanging mode of %si�imz!changing mode of %s from %o to %o)+�mkpathrrr�os�path�join�basename�appendr
rr�debug�open�OSError�dry_run�tokenize�detect_encoding�readline�seek�warn�
first_line_re�match�group�infor�python_buildr�get_config_var�fsencode�decode�UnicodeDecodeError�
ValueError�format�write�
writelines�	readlines�close�	copy_file�name�statr�chmod)rrZ
updated_files�script�adjust�outfiler�encoding�lines�
first_liner7�post_interpr�shebang�outf�fileZoldmodeZnewmoderrrr5s�



�

��
��
��

.


�zbuild_scripts.copy_scriptsN)�__name__�
__module__�__qualname__�description�user_options�boolean_optionsrrrrrrrrrrsr)�__doc__r(�rerGr�	distutilsr�distutils.corer�distutils.dep_utilr�distutils.utilrrr1�compiler6rrrrr�<module>s
site-packages/setuptools/_distutils/command/__pycache__/build_ext.cpython-39.pyc000064400000037676150564745010024160 0ustar00a

(�Re�{�@s�dZddlZddlZddlZddlZddlmZddlTddlm	Z	m
Z
ddlmZddlm
Z
ddlmZdd	lmZdd
lmZddlmZdd
lmZe�d�Zdd�ZGdd�de�ZdS)z�distutils.command.build_ext

Implements the Distutils 'build_ext' command, for building extension
modules (currently limited to C extensions, should accommodate C++
extensions ASAP).�N)�Command)�*)�customize_compiler�get_python_version)�get_config_h_filename)�newer_group)�	Extension)�get_platform)�log�)�
py37compat)�	USER_BASEz3^[a-zA-Z_][a-zA-Z_0-9]*(\.[a-zA-Z_][a-zA-Z_0-9]*)*$cCsddlm}|�dS)Nr��show_compilers)�distutils.ccompilerrr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/build_ext.pyrsrc@seZdZdZdejZddddde�fdd	d
defdd
ddddefddddddddddgZgd�Z	ddde
fgZd d!�Zd"d#�Z
d$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zejd2d3��Zd4d5�Zd6d7�Zd8d9�Zd:d;�Zd<d=�Zd>d?�Zd@dA�ZdBdC�ZdS)D�	build_extz8build C/C++ extensions (compile/link to build directory)z (separated by '%s'))z
build-lib=�bz(directory for compiled extension modules)zbuild-temp=�tz1directory for temporary files (build by-products)z
plat-name=�pz>platform name to cross-compile for, if supported (default: %s))�inplace�iziignore build-lib and put compiled extensions into the source directory alongside your pure Python modulesz
include-dirs=�Iz.list of directories to search for header files)zdefine=�DzC preprocessor macros to define)zundef=�Uz!C preprocessor macros to undefine)z
libraries=�lz!external C libraries to link withz
library-dirs=�Lz.directories to search for external C libraries)zrpath=�Rz7directories to search for shared C libraries at runtime)z
link-objects=�Oz2extra explicit link objects to include in the link)�debug�gz'compile/link with debugging information)�force�fz2forcibly build everything (ignore file timestamps))z	compiler=�czspecify the compiler type)z	parallel=�jznumber of parallel build jobs)�swig-cppNz)make SWIG create C++ files (default is C))z
swig-opts=Nz!list of SWIG command line options)zswig=Nzpath to the SWIG executable)�userNz#add user include, library and rpath)rr r"r&r'z
help-compilerNzlist available compilerscCs�d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_d|_d|_
d|_d|_d|_d|_d|_d|_d|_dS)Nr)�
extensions�	build_lib�	plat_name�
build_tempr�package�include_dirs�define�undef�	libraries�library_dirs�rpath�link_objectsr r"�compiler�swig�swig_cpp�	swig_optsr'�parallel��selfrrr�initialize_optionsks*zbuild_ext.initialize_optionsc

Cs�ddlm}|�ddddddd	d
�|jdur8|jj|_|jj|_|��}|jdd�}|j	durn|jj	pjg|_	t
|j	t�r�|j	�t
j�|_	tjtjkr�|j	�t
j�tjd
��|j	�|�t
jj��||kr�|j	�|�t
jj��|�d�|�d�|jdu�rg|_|jdu�rg|_nt
|jt��r:|j�t
j�|_|jdu�rNg|_nt
|jt��rl|j�t
j�|_t
jdk�rh|j�t
j�tjd��tjtjk�r�|j�t
j�tjd��|j�r�t
j�|jd�|_nt
j�|jd�|_|j	�t
j�t���t tdd�}|�r|j�|�|j!dk�r*d}n|j!dd�}t
j�tjd�}|�r\t
j�||�}|j�|�tj"dd�dk�r�|j#�s�|j�t
j�tjddt$�d��n|j�d�|�%d��r�|j#�s�|j�|�%d��n|j�d�|j&�r|j&�d �}d!d"�|D�|_&|j'�r"|j'�d �|_'|j(du�r6g|_(n|j(�d#�|_(|j)�r�t
j�t*d
�}t
j�t*d�}	t
j�+|��r�|j	�|�t
j�+|	��r�|j�|	�|j�|	�t
|j,t��r�zt-|j,�|_,Wnt.�y�t/d$��Yn0dS)%Nr)�	sysconfig�build)r)r))r+r+)r4r4)r r )r"r")r8r8)r*r*r)�
plat_specific�includer0r3�ntZlibsZDebugZRelease�_home�win32�ZPCbuild��cygwin�lib�python�config�.�Py_ENABLE_SHARED�LIBDIR�,cSsg|]}|df�qS)�1r)�.0�symbolrrr�
<listcomp>��z.build_ext.finalize_options.<locals>.<listcomp>� zparallel should be an integer)0�	distutilsr<�set_undefined_optionsr,�distribution�ext_package�ext_modulesr(�get_python_incr-�
isinstance�str�split�os�pathsep�sys�exec_prefix�base_exec_prefix�append�path�join�extend�ensure_string_listr0r1r2�name�prefixr r+�dirnamer�getattrr*�platform�python_buildr�get_config_varr.r/r7r'r
�isdirr8�int�
ValueError�DistutilsOptionError)
r:r<Z
py_includeZplat_py_include�	_sys_home�suffixZnew_libZdefinesZuser_includeZuser_librrr�finalize_options�s��




�

�zbuild_ext.finalize_optionscCsjddlm}|jsdS|j��rL|�d�}|j�|��p:g�|j	�
|j�||j|j
|j|jd�|_t|j�tjdkr�|jt�kr�|j�|j�|jdur�|j�|j�|jdur�|jD]\}}|j�||�q�|jdur�|jD]}|j�|�q�|jdu�r|j�|j�|j	du�r*|j�|j	�|jdu�rD|j�|j�|j du�r^|j�!|j �|�"�dS)Nr)�new_compiler�
build_clib)r4�verbose�dry_runr"r@)#rrtr(rU�has_c_libraries�get_finalized_commandr0rdZget_library_namesr1rarur4rvrwr"rr\rfr*r	Z
initializer-Zset_include_dirsr.Zdefine_macror/Zundefine_macroZ
set_librariesZset_library_dirsr2Zset_runtime_library_dirsr3Zset_link_objects�build_extensions)r:rtrurf�value�macrorrr�runs@

�




z
build_ext.runc
Csvt|t�std��t|�D�]T\}}t|t�r0qt|t�rFt|�dkrNtd��|\}}t�d|�t|t	�rvt
�|�s~td��t|t�s�td��t||d�}dD]"}|�
|�}|d	ur�t|||�q�|�
d
�|_d|vr�t�d�|�
d
�}|�rhg|_g|_|D]b}	t|	t��r"t|	�dv�s*td��t|	�dk�rJ|j�|	d�nt|	�dk�r|j�|	��q|||<qd	S)a�Ensure that the list of extensions (presumably provided as a
        command option 'extensions') is valid, i.e. it is a list of
        Extension objects.  We also support the old-style list of 2-tuples,
        where the tuples are (ext_name, build_info), which are converted to
        Extension instances here.

        Raise DistutilsSetupError if the structure is invalid anywhere;
        just returns otherwise.
        z:'ext_modules' option must be a list of Extension instances�zMeach element of 'ext_modules' option must be an Extension instance or 2-tuplezvold-style (ext_name, build_info) tuple found in ext_modules for extension '%s' -- please convert to Extension instancezRfirst element of each tuple in 'ext_modules' must be the extension name (a string)zOsecond element of each tuple in 'ext_modules' must be a dictionary (build info)�sources)r-r1r0�
extra_objects�extra_compile_args�extra_link_argsNr2Zdef_filez9'def_file' element of build info dict no longer supported�macros)rr~z9'macros' element of build info dict must be 1- or 2-tuplerr)rY�list�DistutilsSetupError�	enumerater�tuple�lenr
�warnrZ�extension_name_re�match�dict�get�setattr�runtime_library_dirs�
define_macros�undef_macrosra)
r:r(r�ext�ext_nameZ
build_info�key�valr�r|rrr�check_extensions_listWs^

�
��
��
�


�zbuild_ext.check_extensions_listcCs,|�|j�g}|jD]}|�|j�q|S�N)r�r(rdr)r:�	filenamesr�rrr�get_source_files�s

zbuild_ext.get_source_filescCs2|�|j�g}|jD]}|�|�|j��q|Sr�)r�r(ra�get_ext_fullpathrf)r:�outputsr�rrr�get_outputs�s

zbuild_ext.get_outputscCs(|�|j�|jr|��n|��dSr�)r�r(r8�_build_extensions_parallel�_build_extensions_serialr9rrrrz�s
zbuild_ext.build_extensionsc
s��j}�jdurt��}zddlm}Wnty>d}Yn0|durT���dS||d��n���fdd��jD�}t�j|�D]:\}}��	|��|�
�Wd�q�1s�0Yq�Wd�n1s�0YdS)NTr)�ThreadPoolExecutor)�max_workerscsg|]}���j|��qSr)�submit�build_extension)rNr���executorr:rrrP�s�z8build_ext._build_extensions_parallel.<locals>.<listcomp>)r8r\�	cpu_count�concurrent.futuresr��ImportErrorr�r(�zip�_filter_build_errors�result)r:�workersr��futuresr��futrr�rr��s"

�z$build_ext._build_extensions_parallelc	CsD|jD]8}|�|��|�|�Wd�q1s40YqdSr�)r(r�r�)r:r�rrrr��s
z"build_ext._build_extensions_serialc
csVz
dVWnFtttfyP}z(|js(�|�d|j|f�WYd}~n
d}~00dS)Nz"building extension "%s" failed: %s)�CCompilerError�DistutilsError�CompileError�optionalr�rf)r:r��errrr��s
�zbuild_ext._filter_build_errorsc
CsP|j}|dust|ttf�s*td|j��t|�}|�|j�}||j}|j	slt
||d�slt�d|j�dSt�
d|j�|�||�}|jp�g}|jdd�}|jD]}|�|f�q�|jj||j||j|j||jd�}|dd�|_|jr�|�|j�|j�pg}|j�p|j�|�}	|jj|||�|�|j|j ||�!|�|j|j|	d�
dS)Nzjin 'ext_modules' option (extension '%s'), 'sources' must be present and must be a list of source filenames�newerz$skipping '%s' extension (up-to-date)zbuilding '%s' extension)�
output_dirr�r-r �extra_postargs�depends)r0r1r�r��export_symbolsr r+Ztarget_lang)"rrYr�r�r�rf�sortedr�r�r"rr
r �info�swig_sourcesr�r�r�rar4�compiler+r-Z_built_objectsr�rdr��languageZdetect_languageZlink_shared_object�
get_librariesr1r��get_export_symbols)
r:r�r�ext_pathr��
extra_argsr�r/Zobjectsr�rrrr��sV��


�
�zbuild_ext.build_extensioncCs$g}g}i}|jrt�d�|js6d|jvs6d|jvr<d}nd}|D]P}tj�|�\}}	|	dkr�|�|d|�|�|�|d||<qD|�|�qD|s�|S|jp�|�	�}
|
dg}|�
|j�|jr�|�d�|js�|jD]}|�|�q�|D].}||}
t�d	||
�|�|d
|
|g�q�|S)z�Walk the list of source files in 'sources', looking for SWIG
        interface (.i) files.  Run SWIG on all that are found, and
        return a modified 'sources' list with SWIG source files replaced
        by the generated C (or C++) files.
        z/--swig-cpp is deprecated - use --swig-opts=-c++z-c++z.cppz.cz.i�_wrap���z-pythonzswigging %s to %sz-o)
r6r
r�r7r\rb�splitextrar5�	find_swigrdr��spawn)r:r�	extensionZnew_sourcesr�Zswig_targetsZ
target_ext�source�baser�r5Zswig_cmd�o�targetrrrr�3s@
�


zbuild_ext.swig_sourcescCs^tjdkrdStjdkrLdD]*}tj�d|d�}tj�|�r|SqdStdtj��dS)	z�Return the name of the SWIG executable.  On Unix, this is
        just "swig" -- it should be in the PATH.  Tries a bit harder on
        Windows.
        �posixr5r@)z1.3z1.2z1.1z	c:\swig%szswig.exez>I don't know how to find (much less run) SWIG on platform '%s'N)r\rfrbrc�isfile�DistutilsPlatformError)r:�vers�fnrrrr�is


��zbuild_ext.find_swigcCs�|�|�}|�d�}|�|d�}|jsRtjj|dd�|g�}tj�|j|�Sd�|dd��}|�d�}tj�	|�
|��}tj�||�S)z�Returns the path of the filename for a given extension.

        The file is located in `build_lib` or directly in the package
        (inplace option).
        rIr�Nr�build_py)�get_ext_fullnamer[�get_ext_filenamerr\rbrcr)ry�abspathZget_package_dir)r:r��fullname�modpath�filenamer,r��package_dirrrrr��s


zbuild_ext.get_ext_fullpathcCs |jdur|S|jd|SdS)zSReturns the fullname of a given extension name.

        Adds the `package.` prefixNrI)r,)r:r�rrrr��s
zbuild_ext.get_ext_fullnamecCs.ddlm}|�d�}|d�}tjj|�|S)z�Convert the name of an extension (eg. "foo.bar") into the name
        of the file from which it will be loaded (eg. "foo/bar.so", or
        "foo\bar.pyd").
        r�rlrI�
EXT_SUFFIX)�distutils.sysconfigrlr[r\rbrc)r:r�rlr�Z
ext_suffixrrrr��s
zbuild_ext.get_ext_filenamecCsz|j�d�d}z|�d�Wn.tyLd|�d��dd��d�}Yn
0d|}d	|}||jvrt|j�|�|jS)
aReturn the list of symbols that a shared extension has to
        export.  This either uses 'ext.export_symbols' or, if it's not
        provided, "PyInit_" + module_name.  Only relevant on Windows, where
        the .pyd file (DLL) must export the module "PyInit_" function.
        rIr��asciiZU_�punycode�-�_�_ZPyInit)rfr[�encode�UnicodeEncodeError�replace�decoder�ra)r:r�rfrrZ
initfunc_namerrrr��s"
zbuild_ext.get_export_symbolscCs�tjdkr^ddlm}t|j|�s�d}|jr4|d}|tjd?tjd?d@f}|j|gSn�dd	l	m
}d
}|d�r�ttd�r�d
}n<tjdkr�d
}n,dtj
vr�|d�dkr�d
}n|d�dkr�d
}|r�|d�}|jd|gS|jt��S)z�Return the list of libraries to link against when building a
        shared extension.  On most platforms, this is just 'ext.libraries';
        on Windows, we add the Python library (eg. python20.dll).
        rBr)�MSVCCompilerz
python%d%dZ_d���r�FrJ�getandroidapilevelTrE�_PYTHON_HOST_PLATFORM�ANDROID_API_LEVEL�MACHDEP�	LDVERSIONrG)r^rjZdistutils._msvccompilerr�rYr4r �
hexversionr0r�rl�hasattrr\�environr�	pythonlib)r:r�r��templater�rlZlink_libpython�	ldversionrrrr��s4

�



zbuild_ext.get_libraries) �__name__�
__module__�__qualname__�descriptionr\r]Zsep_byr	�user_options�boolean_optionsr�help_optionsr;rsr}r�r�r�rzr�r��
contextlib�contextmanagerr�r�r�r�r�r�r�r�r�rrrrr"sn
�����+��@N	
	L6	
r)�__doc__r�r\�rer^�distutils.corer�distutils.errorsr�rrr�distutils.dep_utilr�distutils.extensionr�distutils.utilr	rSr
�r�siter
r�r�rrrrrr�<module>s&�site-packages/setuptools/_distutils/command/__pycache__/bdist_rpm.cpython-39.pyc000064400000030053150564745010024142 0ustar00a

(�Re!T�@stdZddlZddlZddlZddlmZddlmZddlm	Z	ddl
TddlmZddl
mZGd	d
�d
e�ZdS)zwdistutils.command.bdist_rpm

Implements the Distutils 'bdist_rpm' command (create RPM source and binary
distributions).�N)�Command)�DEBUG)�
write_file)�*)�get_python_version)�logc@sdeZdZdZgd�Zgd�Zdddd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Zdd�Z
dS)�	bdist_rpmzcreate an RPM distribution)))zbdist-base=Nz/base directory for creating built distributions)z	rpm-base=Nzdbase directory for creating RPMs (defaults to "rpm" under --bdist-base; must be specified for RPM 2))z	dist-dir=�dzDdirectory to put final RPM files in (and .spec files if --spec-only))zpython=NzMpath to Python interpreter to hard-code in the .spec file (default: "python"))z
fix-pythonNzLhard-code the exact path to the current Python interpreter in the .spec file)z	spec-onlyNzonly regenerate spec file)zsource-onlyNzonly generate source RPM)zbinary-onlyNzonly generate binary RPM)z	use-bzip2Nz7use bzip2 instead of gzip to create source distribution)zdistribution-name=Nzgname of the (Linux) distribution to which this RPM applies (*not* the name of the module distribution!))zgroup=Nz9package classification [default: "Development/Libraries"])zrelease=NzRPM release number)zserial=NzRPM serial number)zvendor=NzaRPM "vendor" (eg. "Joe Blow <joe@example.com>") [default: maintainer or author from setup script])z	packager=NzBRPM packager (eg. "Jane Doe <jane@example.net>") [default: vendor])z
doc-files=Nz6list of documentation files (space or comma-separated))z
changelog=Nz
RPM changelog)zicon=Nzname of icon file)z	provides=Nz%capabilities provided by this package)z	requires=Nz%capabilities required by this package)z
conflicts=Nz-capabilities which conflict with this package)zbuild-requires=Nz+capabilities required to build this package)z
obsoletes=Nz*capabilities made obsolete by this package)�
no-autoreqNz+do not automatically calculate dependencies)�	keep-temp�kz"don't clean up RPM build directory)�no-keep-tempNz&clean up RPM build directory [default])�use-rpm-opt-flagsNz8compile with RPM_OPT_FLAGS when building from source RPM)�no-rpm-opt-flagsNz&do not pass any RPM CFLAGS to compiler)�	rpm3-modeNz"RPM 3 compatibility mode (default))�	rpm2-modeNzRPM 2 compatibility mode)zprep-script=Nz3Specify a script for the PREP phase of RPM building)z
build-script=Nz4Specify a script for the BUILD phase of RPM building)zpre-install=Nz:Specify a script for the pre-INSTALL phase of RPM building)zinstall-script=Nz6Specify a script for the INSTALL phase of RPM building)z
post-install=Nz;Specify a script for the post-INSTALL phase of RPM building)zpre-uninstall=Nz<Specify a script for the pre-UNINSTALL phase of RPM building)zpost-uninstall=Nz=Specify a script for the post-UNINSTALL phase of RPM building)z
clean-script=Nz4Specify a script for the CLEAN phase of RPM building)zverify-script=Nz6Specify a script for the VERIFY phase of the RPM build)zforce-arch=Nz0Force an architecture onto the RPM build process)�quiet�qz3Run the INSTALL phase of RPM building in quiet mode)rrrr
rrrr)r
rrcCs�d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_	d|_
d|_d|_d|_
d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_d|_ d|_!d|_"d|_#d|_$d|_%d|_&dS)Nr�)'�
bdist_base�rpm_base�dist_dir�python�
fix_python�	spec_only�binary_only�source_only�	use_bzip2�distribution_name�group�release�serial�vendor�packager�	doc_files�	changelog�icon�prep_script�build_script�install_script�clean_script�
verify_script�pre_install�post_install�
pre_uninstall�post_uninstall�prep�provides�requires�	conflicts�build_requires�	obsoletes�	keep_temp�use_rpm_opt_flags�	rpm3_mode�
no_autoreq�
force_archr��self�r=��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/bdist_rpm.py�initialize_options�sNzbdist_rpm.initialize_optionscCs�|�dd�|jdur6|js$td��tj�|jd�|_|jdurX|j	rPt
j|_qfd|_n|j	rftd��tjdkr~t
dtj��|jr�|jr�td	��|j��s�d
|_|�dd�|��dS)N�bdist)rrz)you must specify --rpm-base in RPM 2 mode�rpmZpython3z8--python and --fix-python are mutually exclusive options�posixz9don't know how to create RPM distributions on platform %sz6cannot supply both '--source-only' and '--binary-only'r)rr)�set_undefined_optionsrr8�DistutilsOptionError�os�path�joinrrr�sys�
executable�name�DistutilsPlatformErrorrr�distribution�has_ext_modulesr7�finalize_package_datar;r=r=r>�finalize_options�s6
�

�
��
zbdist_rpm.finalize_optionscCsT|�dd�|�dd|j��|j��f�|�d�|�d�t|jt�rxdD]&}tj	�
|�rP||jvrP|j�|�qP|�dd	�|�d
�|�d�|�d�|�|j
�|_
|�d
�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�|�d�dS)NrzDevelopment/Librariesr"z%s <%s>r#r$)ZREADMEz
README.txtr �1r!rr%r&r'r(r)r*r+r,r-r.r/r1r2r3r4r5r:)�
ensure_stringrL�get_contact�get_contact_email�ensure_string_list�
isinstancer$�listrErF�exists�append�_format_changelogr%�ensure_filename)r<Zreadmer=r=r>rN�sB
��



















zbdist_rpm.finalize_package_datacCs�tr<td�td|j�td|j�td|j�td|j�|jrT|j}|�|�n8i}dD]&}t	j
�|j|�||<|�||�q\|d}t	j
�|d|j
���}|�t||��fd	|�|jr�dS|j
jdd�}|�d
�}|jr�dg|_ndg|_|�d
�||j
_|��d
}|d}|�||�|j�rbt	j
�|j��rT|�|j|�ntd|j��t�d�dg}	|j�r�|	�d�n|j �r�|	�d�n
|	�d�|	�!dd|j"g�|j#�r�|	�!ddt	j
�$|j�g�|j%�s�|	�d�|j&�r|	�d�|	�|�d}
|
d}d|
d}d|||f}
t	�'|
�}z�g}d}|�(�}|�sV�q�|�)��*�}t+|�dk�stJ�|�|d �|du�rD|d
}�qD|�,�}|�r�t-d!t.|
���W|�,�n
|�,�0|�/|	�|j0�s�|j
�1��r�t2�}nd"}|j �sTt	j
�|d#|�}t	j
�|��s"J�|�3||j�t	j
�|j|�}|j
j�d$||f�|j�s�|D]`}t	j
�|d%|�}t	j
�|��r`|�3||j�t	j
�|jt	j
�4|��}|j
j�d$||f��q`dS)&Nzbefore _get_package_data():zvendor =z
packager =zdoc_files =zchangelog =)�SOURCES�SPECS�BUILD�RPMS�SRPMSr\z%s.speczwriting '%s'�sdist�bztar�gztarrr[zicon file '%s' does not existz
building RPMsZrpmbuildz-bsz-bbz-baz--definez__python %sz
_topdir %sz--cleanz--quietz%{name}-%{version}-%{release}z.src.rpmz%{arch}/z.%{arch}.rpmz%rpm -q --qf '%s %s\n' --specfile '%s'�rzFailed to execute: %s�anyr_rr^)5r�printr"r#r$r%rr�mkpathrErFrGrrL�get_name�executer�_make_spec_file�
dist_files�reinitialize_commandr�formats�run_commandZget_archive_files�	copy_filer&rW�DistutilsFileErrorr�inforrXr�extendrr8�abspathr6r�popen�readline�strip�split�len�close�DistutilsExecError�repr�spawn�dry_runrMr�	move_file�basename)r<Zspec_dirZrpm_dirr	Z	spec_pathZsaved_dist_filesr`�source�
source_dirZrpm_cmdZ
nvr_stringZsrc_rpmZnon_src_rpmZq_cmd�outZbinary_rpmsZ
source_rpm�line�l�statusZ	pyversionZsrpm�filenamerAr=r=r>�runs����


�

�


�


�

��z
bdist_rpm.runcCstj�|jtj�|��S)N)rErFrGrr~)r<rFr=r=r>�
_dist_path�szbdist_rpm._dist_pathc	Cs\d|j��d|j���dd�d|j��d|j�dd�dd|j��g}t�d	�}d
�dd�|�	�D��}d
}d}|�||�}||kr�|�
d�|�
d|d
�|�gd��|jr�|�
d�n
|�
d�|�d|j�
�d|jddg�|j�s|j���s$|�
d�n|�
d|j�dD]V}t||���}t|t��r`|�
d|d�|�f�n|du�r(|�
d||f��q(|j��dk�r�|�
d|j���|j�r�|�
d |j�|j�r�|�
d!d�|j��|j�r�|�
d"tj�|j��|j�r|�
d#�|�dd$|j��g�d%|jtj�tj d&�f}d'|}	|j!�rVd(|	}	d)|}
d*d+d,|	fd-d.|
fd/d0d1d2d3d4g	}|D]�\}}
}t||
�}|�s�|�r�|�dd5|g�|�r�t"|��$}|�|�#��$d
��Wd�n1�s�0Yn
|�
|��q�|�gd6��|j%�r6|�
d7d�|j%��|j&�rX|�dd8g�|�|j&�|S)9ziGenerate the text of an RPM spec file and return it as a
        list of strings (one per line).
        z
%define name z%define version �-�_z%define unmangled_version z%define release �z	Summary: zrpm --eval %{__os_install_post}�
cSsg|]}d|���qS)z  %s \)ru)�.0r�r=r=r>�
<listcomp>�s�z-bdist_rpm._make_spec_file.<locals>.<listcomp>zbrp-python-bytecompile \
z%brp-python-bytecompile %{__python} \
z2# Workaround for http://bugs.python.org/issue14443z%define __os_install_post )z
Name: %{name}zVersion: %{version}zRelease: %{release}z-Source0: %{name}-%{unmangled_version}.tar.bz2z,Source0: %{name}-%{unmangled_version}.tar.gzz	License: zGroup: z>BuildRoot: %{_tmppath}/%{name}-%{version}-%{release}-buildrootzPrefix: %{_prefix}zBuildArch: noarchz
BuildArch: %s)ZVendorZPackager�Provides�RequiresZ	Conflicts�	Obsoletesz%s: %s� N�UNKNOWNzUrl: zDistribution: zBuildRequires: zIcon: z
AutoReq: 0z%descriptionz%s %srz%s buildzenv CFLAGS="$RPM_OPT_FLAGS" z>%s install -O1 --root=$RPM_BUILD_ROOT --record=INSTALLED_FILES)r0r'z&%setup -n %{name}-%{unmangled_version}�buildr(�installr))�cleanr*zrm -rf $RPM_BUILD_ROOT)Zverifyscriptr+N)�prer,N)�postr-N)Zpreunr.N)Zpostunr/N�%)r�z%files -f INSTALLED_FILESz%defattr(-,root,root)z%doc z
%changelog)'rLrg�get_version�replacer �get_description�
subprocess�	getoutputrG�
splitlinesrXrqr�get_licenserr:rM�getattr�lowerrUrV�get_urlrr4r&rErFr~r9�get_long_descriptionrrH�argvr7�open�readrvr$r%)r<Z	spec_fileZvendor_hook�problem�fixedZ
fixed_hook�field�valZdef_setup_callZ	def_buildZinstall_cmdZscript_optionsZrpm_opt�attr�default�fr=r=r>ri�s��

�
	
�

�
���
�
6�zbdist_rpm._make_spec_filecCs||s|Sg}|���d�D]N}|��}|ddkrB|�d|g�q|ddkrZ|�|�q|�d|�q|dsx|d=|S)zKFormat the changelog correctly and convert it to a list of strings
        r�rrr�r�z  )rurvrqrX)r<r%Z
new_changelogr�r=r=r>rY0szbdist_rpm._format_changelogN)�__name__�
__module__�__qualname__�description�user_options�boolean_options�negative_optr?rOrNr�r�rirYr=r=r=r>rs m�--*r)�__doc__r�rHrE�distutils.corer�distutils.debugr�distutils.file_utilr�distutils.errors�distutils.sysconfigr�	distutilsrrr=r=r=r>�<module>ssite-packages/setuptools/_distutils/command/__pycache__/sdist.cpython-39.pyc000064400000034372150564745010023315 0ustar00a

(�Re=J�@s�dZddlZddlZddlmZddlmZddlmZddlm	Z	ddlm
Z
ddlmZdd	lm
Z
dd
lmZddlmZddlmZdd
lmZmZdd�ZGdd�de�ZdS)zadistutils.command.sdist

Implements the Distutils 'sdist' command (create a source distribution).�N)�glob)�warn)�Command)�dir_util)�	file_util)�archive_util)�TextFile)�FileList)�log)�convert_path)�DistutilsTemplateError�DistutilsOptionErrorcCs`ddlm}ddlm}g}|��D] }|�d|d||df�q$|��||��d�dS)zoPrint all possible values for the 'formats' option (used by
    the "--help-formats" command-line option).
    r)�FancyGetopt)�ARCHIVE_FORMATS�formats=N�z.List of available source distribution formats:)�distutils.fancy_getoptrZdistutils.archive_utilr�keys�append�sort�
print_help)rr�formats�format�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/sdist.py�show_formatss
��rc@seZdZdZdd�Zgd�Zgd�ZdddefgZd	d
d�Z	defgZ
d
Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zedd��Zdd�Zdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zd2d3�Zd4d5�Z d6d7�Z!d8d9�Z"dS):�sdistz6create a source distribution (tarball, zip file, etc.)cCs|jS)zYCallable used for the check sub-command.

        Placed here so user_options can view it)�metadata_check��selfrrr�checking_metadata(szsdist.checking_metadata))z	template=�tz5name of manifest template file [default: MANIFEST.in])z	manifest=�mz)name of manifest file [default: MANIFEST])�use-defaultsNzRinclude the default file set in the manifest [default; disable with --no-defaults])�no-defaultsNz"don't include the default file set)�pruneNz�specifically exclude files/directories that should not be distributed (build tree, RCS/CVS dirs, etc.) [default; disable with --no-prune])�no-pruneNz$don't automatically exclude anything)�
manifest-only�ozEjust regenerate the manifest and then stop (implies --force-manifest))�force-manifest�fzkforcibly regenerate the manifest and carry on as usual. Deprecated: now the manifest is always regenerated.)rNz6formats for source distribution (comma-separated list))�	keep-temp�kz@keep the distribution tree around after creating archive file(s))z	dist-dir=�dzFdirectory to put the source distribution archive(s) in [default: dist])�metadata-checkNz[Ensure that all required elements of meta-data are supplied. Warn if any missing. [default])zowner=�uz@Owner name used when creating a tar file [default: current user])zgroup=�gzAGroup name used when creating a tar file [default: current group])r#r%r'r)r+r.zhelp-formatsNz#list available distribution formatsr#r%)r$r&�check)ZREADMEz
README.txtz
README.rstcCsTd|_d|_d|_d|_d|_d|_dg|_d|_d|_d|_	d|_
d|_d|_dS)N�r�gztar)
�template�manifest�use_defaultsr%�
manifest_onlyZforce_manifestr�	keep_temp�dist_dir�
archive_filesr�owner�grouprrrr�initialize_optionseszsdist.initialize_optionscCsZ|jdurd|_|jdur d|_|�d�t�|j�}|rFtd|��|jdurVd|_dS)NZMANIFESTzMANIFEST.inrzunknown archive format '%s'�dist)r5r4�ensure_string_listr�check_archive_formatsrr
r9)rZ
bad_formatrrr�finalize_options|s


�
zsdist.finalize_optionscCs>t�|_|��D]}|�|�q|��|jr2dS|��dS�N)r	�filelist�get_sub_commands�run_command�
get_file_listr7�make_distribution)r�cmd_namerrr�run�sz	sdist.runcCs*tdt�|j�d�}|��|��dS)zDeprecated API.zadistutils.command.sdist.check_metadata is deprecated,               use the check command insteadr1N)r�PendingDeprecationWarning�distribution�get_command_obj�ensure_finalizedrI)rr1rrr�check_metadata�s�zsdist.check_metadatacCs�tj�|j�}|s:|��r:|��|j��|j��dS|sN|�	d|j�|j�
�|jrf|��|rr|�
�|jr�|��|j��|j��|��dS)aCFigure out the list of files to include in the source
        distribution, and put it in 'self.filelist'.  This might involve
        reading the manifest template (and writing the manifest), or just
        reading the manifest, or just using the default file set -- it all
        depends on the user's options.
        Nz?manifest template '%s' does not exist (using default file list))�os�path�isfiler4�_manifest_is_not_generated�
read_manifestrCrZremove_duplicatesr�findallr6�add_defaults�
read_templater%�prune_file_list�write_manifest)rZtemplate_existsrrrrF�s(

�


zsdist.get_file_listcCs<|��|��|��|��|��|��|��dS)a9Add all the default files to self.filelist:
          - README or README.txt
          - setup.py
          - test/test*.py
          - all pure Python modules mentioned in setup script
          - all files pointed by package_data (build_py)
          - all files defined in data_files.
          - all files defined as scripts.
          - all C sources listed as part of extensions or C libraries
            in the setup script (doesn't catch C headers!)
        Warns if (README or README.txt) or setup.py are missing; everything
        else is optional.
        N)�_add_defaults_standards�_add_defaults_optional�_add_defaults_python�_add_defaults_data_files�_add_defaults_ext�_add_defaults_c_libs�_add_defaults_scriptsrrrrrU�szsdist.add_defaultscCs:tj�|�sdStj�|�}tj�|�\}}|t�|�vS)z�
        Case-sensitive path existence check

        >>> sdist._cs_path_exists(__file__)
        True
        >>> sdist._cs_path_exists(__file__.upper())
        False
        F)rOrP�exists�abspath�split�listdir)�fspathra�	directory�filenamerrr�_cs_path_exists�s

zsdist._cs_path_existscCs�|j|jjg}|D]~}t|t�rj|}d}|D]"}|�|�r,d}|j�|�qPq,|s�|�dd�	|��q|�|�r�|j�|�q|�d|�qdS)NFTz,standard file not found: should have one of z, zstandard file '%s' not found)
�READMESrK�script_name�
isinstance�tuplergrCrr�join)rZ	standards�fnZaltsZgot_itrrrrY�s"

�
zsdist._add_defaults_standardscCs4ddg}|D]"}ttjjt|��}|j�|�qdS)Nz
test/test*.pyz	setup.cfg)�filterrOrPrQrrC�extend)r�optional�pattern�filesrrrrZszsdist._add_defaults_optionalcCs\|�d�}|j��r$|j�|���|jD],\}}}}|D]}|j�tj	�
||��q:q*dS)N�build_py)�get_finalized_commandrK�has_pure_modulesrCro�get_source_files�
data_filesrrOrPrl)rrs�pkg�src_dir�	build_dir�	filenamesrfrrrr[s

zsdist._add_defaults_pythoncCsz|j��rv|jjD]b}t|t�rBt|�}tj�|�rt|j	�
|�q|\}}|D]$}t|�}tj�|�rN|j	�
|�qNqdSrB)rK�has_data_filesrwrj�strrrOrPrQrCr)r�item�dirnamer{r*rrrr\$s

zsdist._add_defaults_data_filescCs(|j��r$|�d�}|j�|���dS)N�	build_ext)rK�has_ext_modulesrtrCrorv)rr�rrrr]5s

zsdist._add_defaults_extcCs(|j��r$|�d�}|j�|���dS)N�
build_clib)rK�has_c_librariesrtrCrorv)rr�rrrr^:s

zsdist._add_defaults_c_libscCs(|j��r$|�d�}|j�|���dS)N�
build_scripts)rK�has_scriptsrtrCrorv)rr�rrrr_?s

zsdist._add_defaults_scriptsc
Cs�t�d|j�t|jddddddd�}zr|��}|dur:q�z|j�|�Wq(ttfy�}z$|�	d|j
|j|f�WYd}~q(d}~00q(W|��n
|��0dS)z�Read and parse manifest template file named by self.template.

        (usually "MANIFEST.in") The parsing and processing is done by
        'self.filelist', which updates itself accordingly.
        zreading manifest template '%s'r2)�strip_comments�skip_blanks�
join_lines�	lstrip_ws�	rstrip_wsZ
collapse_joinNz%s, line %d: %s)
r
�infor4r�readlinerC�process_template_liner�
ValueErrorrrf�current_line�close)rr4�line�msgrrrrVDs"
�
�"zsdist.read_templatecCsz|�d�}|j��}|jjd|jd�|jjd|d�tjdkrFd}nd}gd�}d|d	�|�|f}|jj|d
d�dS)avPrune off branches that might slip into the file list as created
        by 'read_template()', but really don't belong there:
          * the build tree (typically "build")
          * the release tree itself (only an issue if we ran "sdist"
            previously with --keep-temp, or it aborted)
          * any RCS, CVS, .svn, .hg, .git, .bzr, _darcs directories
        �buildN)�prefix�win32z/|\\�/)�RCS�CVSz\.svnz\.hgz\.gitz\.bzr�_darcsz(^|%s)(%s)(%s).*�|r2)Zis_regex)	rtrK�get_fullnamerCZexclude_pattern�
build_base�sys�platformrl)rr��base_dir�sepsZvcs_dirsZvcs_ptrnrrrrWas


zsdist.prune_file_listcCsX|��rt�d|j�dS|jjdd�}|�dd�|�tj	|j|fd|j�dS)z�Write the file list in 'self.filelist' (presumably as filled in
        by 'add_defaults()' and 'read_template()') to the manifest file
        named by 'self.manifest'.
        z5not writing to manually maintained manifest file '%s'Nrz*# file GENERATED by distutils, do NOT editzwriting manifest file '%s')
rRr
r�r5rCrr�insert�executer�
write_file)r�contentrrrrXys��zsdist.write_manifestcCsDtj�|j�sdSt|j�}z|��}W|��n
|��0|dkS)NFz+# file GENERATED by distutils, do NOT edit
)rOrPrQr5�openr�r�)r�fp�
first_linerrrrR�s

z sdist._manifest_is_not_generatedcCsjt�d|j�t|j��>}|D](}|��}|�d�s|s:q|j�|�qWd�n1s\0YdS)z�Read the manifest file (named by 'self.manifest') and use it to
        fill in 'self.filelist', the list of files to include in the source
        distribution.
        zreading manifest file '%s'�#N)r
r�r5r��strip�
startswithrCr)rr5r�rrrrS�szsdist.read_manifestcCs�|�|�tj|||jd�ttd�r4d}d|}nd}d|}|sPt�d�n
t�|�|D]<}tj	�
|�s|t�d|�q^tj	�||�}|j|||d	�q^|j
j�|�dS)
a�Create the directory tree that will become the source
        distribution archive.  All directories implied by the filenames in
        'files' are created under 'base_dir', and then we hard link or copy
        (if hard linking is unavailable) those files into place.
        Essentially, this duplicates the developer's source tree, but in a
        directory named after the distribution, containing only the files
        to be distributed.
        ��dry_run�link�hardzmaking hard links in %s...Nzcopying files to %s...z)no files to distribute -- empty manifest?z#'%s' not a regular file -- skipping)r�)�mkpathr�create_treer��hasattrrOr
rr�rPrQrl�	copy_filerK�metadata�write_pkg_info)rr�rrr�r��file�destrrr�make_release_tree�s 
	


zsdist.make_release_treecCs�|j��}tj�|j|�}|�||jj�g}d|j	vrT|j	�
|j	�|j	�d���|j	D]:}|j
||||j|jd�}|�
|�|jj�
dd|f�qZ||_|js�tj||jd�dS)a�Create the source distribution(s).  First, we create the release
        tree with 'make_release_tree()'; then, we create all required
        archive files (according to 'self.formats') from the release tree.
        Finally, we clean up by blowing away the release tree (unless
        'self.keep_temp' is true).  The list of archive files created is
        stored so it can be retrieved later by 'get_archive_files()'.
        �tar)r�r;r<r�r�N)rKr�rOrPrlr9r�rCrrrr�pop�index�make_archiver;r<�
dist_filesr:r8r�remove_treer�)rr��	base_namer:�fmtr�rrrrG�s




�
zsdist.make_distributioncCs|jS)zzReturn the list of archive files created when the command
        was run, or None if the command hasn't run yet.
        )r:rrrr�get_archive_files�szsdist.get_archive_files)#�__name__�
__module__�__qualname__�descriptionr �user_options�boolean_optionsr�help_options�negative_opt�sub_commandsrhr=rArIrNrFrU�staticmethodrgrYrZr[r\r]r^r_rVrWrXrRrSr�rGr�rrrrr$sH'���
(
*r)�__doc__rOr�r�warningsr�distutils.corer�	distutilsrrr�distutils.text_filer�distutils.filelistr	r
�distutils.utilr�distutils.errorsrr
rrrrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/bdist_dumb.cpython-39.pyc000064400000007156150564745010024303 0ustar00a

(�Re1�@shdZddlZddlmZddlmZddlmZmZddl	Tddl
mZddlm
Z
Gd	d
�d
e�ZdS)z�distutils.command.bdist_dumb

Implements the Distutils 'bdist_dumb' command (create a "dumb" built
distribution -- i.e., just an archive to be unpacked under $prefix or
$exec_prefix).�N)�Command)�get_platform)�remove_tree�ensure_relative)�*)�get_python_version)�logc	@s\eZdZdZdddde�fdddd	d
ddg	Zgd
�Zddd�Zdd�Zdd�Z	dd�Z
dS)�
bdist_dumbz"create a "dumb" built distribution)z
bdist-dir=�dz1temporary directory for creating the distributionz
plat-name=�pz;platform name to embed in generated filenames (default: %s))zformat=�fz>archive format to create (tar, gztar, bztar, xztar, ztar, zip))�	keep-temp�kzPkeep the pseudo-installation tree around after creating the distribution archive)z	dist-dir=r
z-directory to put final built distributions in)�
skip-buildNz2skip rebuilding everything (for testing/debugging))�relativeNz7build the archive using relative paths (default: false))zowner=�uz@Owner name used when creating a tar file [default: current user])zgroup=�gzAGroup name used when creating a tar file [default: current group])r
rr�gztar�zip)�posix�ntcCs:d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)Nr)	�	bdist_dir�	plat_name�format�	keep_temp�dist_dir�
skip_buildr�owner�group)�self�r ��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/bdist_dumb.py�initialize_options2szbdist_dumb.initialize_optionscCsx|jdur&|�d�j}tj�|d�|_|jdurdz|jtj|_Wn t	ybt
dtj��Yn0|�dddd�dS)N�bdistZdumbz@don't know how to create dumb built distributions on platform %s)rr)rr)rr)r�get_finalized_command�
bdist_base�os�path�joinr�default_format�name�KeyError�DistutilsPlatformError�set_undefined_options)rr%r r r!�finalize_options=s"

��
�zbdist_dumb.finalize_optionscCs(|js|�d�|jddd�}|j|_|j|_d|_t�d|j�|�d�d|j�	�|j
f}tj�
|j|�}|js~|j}nJ|j��r�|j|jkr�tdt|j�t|j�f��ntj�
|jt|j��}|j||j||j|jd	�}|j��r�t�}nd
}|jj�d||f�|j�s$t|j|jd�dS)
N�build�install�)�reinit_subcommandsrzinstalling to %sz%s.%szScan't make a dumb built distribution where base and platbase are different (%s, %s))�root_dirrr�anyr	)�dry_run) r�run_command�reinitialize_commandr�root�warn_dirr�info�distribution�get_fullnamerr&r'r(rr�has_ext_modules�install_base�install_platbaser,�reprr�make_archiverrrr�
dist_files�appendrrr5)rr0Zarchive_basenameZpseudoinstall_rootZarchive_root�filenameZ	pyversionr r r!�runOsN


�

����
��
�zbdist_dumb.runN)�__name__�
__module__�__qualname__�descriptionr�user_options�boolean_optionsr)r"r.rEr r r r!r	s,����r	)�__doc__r&�distutils.corer�distutils.utilr�distutils.dir_utilrr�distutils.errors�distutils.sysconfigr�	distutilsrr	r r r r!�<module>ssite-packages/setuptools/_distutils/command/__pycache__/config.cpython-39.pyc000064400000024116150564745010023427 0ustar00a

(�Re=3�@sldZddlZddlZddlmZddlmZddlmZddl	m
Z
ddd	�ZGd
d�de�Zddd
�Z
dS)a�distutils.command.config

Implements the Distutils 'config' command, a (mostly) empty command class
that exists mainly to be sub-classed by specific module distributions and
applications.  The idea is that while every "config" command is different,
at least they're all named the same, and users always see "config" in the
list of standard commands.  Also, this is a good place to put common
configure-like tasks: "try to compile this C code", or "figure out where
this header file lives".
�N)�Command)�DistutilsExecError)�customize_compiler)�logz.cz.cxx)�czc++c@s�eZdZdZgd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
d(dd�Zd)dd�Zd*dd�Zd+dd�Zd,dd �Zd-d"d#�Zdddgfd$d%�Zd.d&d'�ZdS)/�configzprepare to build)	)z	compiler=Nzspecify the compiler type)zcc=Nzspecify the compiler executable)z
include-dirs=�Iz.list of directories to search for header files)zdefine=�DzC preprocessor macros to define)zundef=�Uz!C preprocessor macros to undefine)z
libraries=�lz!external C libraries to link with)z
library-dirs=�Lz.directories to search for external C libraries)�noisyNz1show every action (compile, link, run, ...) taken)zdump-sourceNz=dump generated source files before attempting to compile themcCs4d|_d|_d|_d|_d|_d|_d|_g|_dS)N�)�compiler�cc�include_dirs�	libraries�library_dirsr
�dump_source�
temp_files��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/config.py�initialize_options3szconfig.initialize_optionscCs�|jdur|jjpg|_nt|jt�r6|j�tj�|_|jdurHg|_nt|jt�r^|jg|_|jdurpg|_nt|jt�r�|j�tj�|_dS�N)	r�distribution�
isinstance�str�split�os�pathseprrrrrr�finalize_optionsBs



zconfig.finalize_optionscCsdSrrrrrr�runRsz
config.runcCszddlm}m}t|j|�sv||j|jdd�|_t|j�|jrN|j�|j�|j	rb|j�
|j	�|jrv|j�|j�dS)z^Check that 'self.compiler' really is a CCompiler object;
        if not, make it one.
        r)�	CCompiler�new_compilerr)r�dry_run�forceN)
�distutils.ccompilerr$r%rrr&rrZset_include_dirsrZ
set_librariesrZset_library_dirs)rr$r%rrr�_check_compilerYs�
zconfig._check_compilercCs�dt|}t|d��V}|r>|D]}|�d|�q |�d�|�|�|ddkr^|�d�Wd�n1sr0Y|S)NZ_configtest�wz#include <%s>
�
���)�LANG_EXT�open�write)r�body�headers�lang�filename�file�headerrrr�_gen_temp_sourcefileks

(zconfig._gen_temp_sourcefilecCs<|�|||�}d}|j�||g�|jj|||d�||fS)Nz
_configtest.i�r)r6r�extendr�
preprocess)rr0r1rr2�src�outrrr�_preprocessws
zconfig._preprocesscCs\|�|||�}|jr"t|d|�|j�|g�\}|j�||g�|jj|g|d�||fS)Nzcompiling '%s':r7)r6r�	dump_filerZobject_filenamesrr8�compile)rr0r1rr2r:�objrrr�_compile~szconfig._compilec
Csr|�||||�\}}tj�tj�|��d}	|jj|g|	|||d�|jjdur\|	|jj}	|j�	|	�|||	fS)Nr)rrZtarget_lang)
r@r �path�splitext�basenamerZlink_executableZ
exe_extensionr�append)
rr0r1rrrr2r:r?�progrrr�_link�s�zconfig._linkc	GsR|s|j}g|_t�dd�|��|D]&}zt�|�Wq&tyJYq&0q&dS)Nzremoving: %s� )rr�info�joinr �remove�OSError)r�	filenamesr3rrr�_clean�sz
config._cleanNrcCsPddlm}|��d}z|�||||�Wn|yBd}Yn0|��|S)aQConstruct a source file from 'body' (a string containing lines
        of C/C++ code) and 'headers' (a list of header files to include)
        and run it through the preprocessor.  Return true if the
        preprocessor succeeded, false if there were any errors.
        ('body' probably isn't of much use, but what the heck.)
        r��CompileErrorTF)r(rOr)r<rM�rr0r1rr2rO�okrrr�try_cpp�s
zconfig.try_cppcCs�|��|�||||�\}}t|t�r0t�|�}t|��8}d}	|��}
|
dkrPqb|�|
�r>d}	qbq>Wd�n1sv0Y|�	�|	S)a�Construct a source file (just like 'try_cpp()'), run it through
        the preprocessor, and return true if any line of the output matches
        'pattern'.  'pattern' should either be a compiled regex object or a
        string containing a regex.  If both 'body' and 'headers' are None,
        preprocesses an empty file -- which can be useful to determine the
        symbols the preprocessor and compiler set by default.
        F�TN)
r)r<rr�rer>r.�readline�searchrM)r�patternr0r1rr2r:r;r4�match�linerrr�
search_cpp�s	



"zconfig.search_cppcCsbddlm}|��z|�||||�d}Wn|yBd}Yn0t�|rPdpRd�|��|S)zwTry to compile a source file built from 'body' and 'headers'.
        Return true on success, false otherwise.
        rrNTF�success!�failure.)r(rOr)r@rrHrMrPrrr�try_compile�s
zconfig.try_compilec
	Csnddlm}m}|��z|�||||||�d}	Wn||fyNd}	Yn0t�|	r\dp^d�|��|	S)z�Try to compile and link a source file, built from 'body' and
        'headers', to executable form.  Return true on success, false
        otherwise.
        r�rO�	LinkErrorTFr[r\)r(rOr_r)rFrrHrM)
rr0r1rrrr2rOr_rQrrr�try_link�s
�
zconfig.try_linkc

Cs�ddlm}m}|��z.|�||||||�\}	}
}|�|g�d}Wn||tfybd}Yn0t�|rpdprd�|�	�|S)z�Try to compile, link to an executable, and run a program
        built from 'body' and 'headers'.  Return true on success, false
        otherwise.
        rr^TFr[r\)
r(rOr_r)rF�spawnrrrHrM)
rr0r1rrrr2rOr_r:r?ZexerQrrr�try_run�s
�

zconfig.try_runrc	Cst|��g}|r|�d|�|�d�|r<|�d|�n|�d|�|�d�d�|�d}|�|||||�S)a�Determine if function 'func' is available by constructing a
        source file that refers to 'func', and compiles and links it.
        If everything succeeds, returns true; otherwise returns false.

        The constructed source file starts out by including the header
        files listed in 'headers'.  If 'decl' is true, it then declares
        'func' (as "int func()"); you probably shouldn't supply 'headers'
        and set 'decl' true in the same call, or you might get errors about
        a conflicting declarations for 'func'.  Finally, the constructed
        'main()' function either references 'func' or (if 'call' is true)
        calls it.  'libraries' and 'library_dirs' are used when
        linking.
        z
int %s ();z
int main () {z  %s();z  %s;�}r+)r)rDrIr`)	r�funcr1rrr�decl�callr0rrr�
check_funcs


�zconfig.check_funccCs |��|�d|||g||�S)a�Determine if 'library' is available to be linked against,
        without actually checking that any particular symbols are provided
        by it.  'headers' will be used in constructing the source file to
        be compiled, but the only effect of this is to check if all the
        header files listed are available.  Any libraries listed in
        'other_libraries' will be included in the link, in case 'library'
        has symbols that depend on other libraries.
        zint main (void) { })r)r`)r�libraryrr1rZother_librariesrrr�	check_lib4s


�zconfig.check_libcCs|jd|g|d�S)z�Determine if the system header file named by 'header_file'
        exists and can be found by the preprocessor; return true if so,
        false otherwise.
        z
/* No body */)r0r1r)rR)rr5rrr2rrr�check_headerBs
�zconfig.check_header)NNNr)NNNr)NNr)NNNNr)NNNNr)NNNNrr)NNr)�__name__�
__module__�__qualname__�description�user_optionsrr"r#r)r6r<r@rFrMrRrZr]r`rbrgrirjrrrrrs>	
�

�
�
�
�
�rcCsR|durt�d|�n
t�|�t|�}zt�|���W|��n
|��0dS)zjDumps a file content into log.info.

    If head is not None, will be dumped before the file content.
    Nz%s)rrHr.�read�close)r3�headr4rrrr=Ks
r=)N)�__doc__r rT�distutils.corer�distutils.errorsr�distutils.sysconfigr�	distutilsrr-rr=rrrr�<module>s
8site-packages/setuptools/_distutils/command/__pycache__/__init__.cpython-39.pyc000064400000001105150564745010023712 0ustar00a

(�Re�@sdZgd�ZdS)z\distutils.command

Package containing implementation of all the standard Distutils
commands.)�build�build_py�	build_ext�
build_clib�
build_scripts�clean�install�install_lib�install_headers�install_scripts�install_data�sdist�register�bdist�
bdist_dumb�	bdist_rpm�
bdist_wininst�check�uploadN)�__doc__�__all__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/__init__.py�<module>ssite-packages/setuptools/_distutils/command/__pycache__/install_data.cpython-39.pyc000064400000004527150564745010024625 0ustar00a

(�Re�@s<dZddlZddlmZddlmZmZGdd�de�ZdS)z�distutils.command.install_data

Implements the Distutils 'install_data' command, for installing
platform-independent data files.�N)�Command)�change_root�convert_pathc@sFeZdZdZgd�ZdgZdd�Zdd�Zdd	�Zd
d�Z	dd
�Z
dS)�install_datazinstall data files))zinstall-dir=�dzIbase directory for installing data files (default: installation base dir))zroot=Nz<install everything relative to this alternate root directory)�force�fz-force installation (overwrite existing files)rcCs,d|_g|_d|_d|_|jj|_d|_dS)Nr�)�install_dir�outfiles�rootr�distribution�
data_files�warn_dir��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install_data.py�initialize_optionss
zinstall_data.initialize_optionscCs|�dddd�dS)N�install)rr
)rr)rr)�set_undefined_optionsrrrr�finalize_options#s
�zinstall_data.finalize_optionscCs�|�|j�|jD]�}t|t�rbt|�}|jrB|�d||jf�|�||j�\}}|j	�
|�qt|d�}tj�
|�s�tj�|j|�}n|jr�t|j|�}|�|�|dgkr�|j	�
|�q|dD](}t|�}|�||�\}}|j	�
|�q�qdS)NzMsetup script did not provide a directory for '%s' -- installing right in '%s'rr	)�mkpathr
r�
isinstance�strrr�warn�	copy_filer�append�os�path�isabs�joinrr)rr�out�_�dir�datarrr�run*s,

�
zinstall_data.runcCs
|jpgS�N)rrrrr�
get_inputsKszinstall_data.get_inputscCs|jSr')rrrrr�get_outputsNszinstall_data.get_outputsN)�__name__�
__module__�__qualname__�description�user_options�boolean_optionsrrr&r(r)rrrrrs	!r)�__doc__r�distutils.corer�distutils.utilrrrrrrr�<module>ssite-packages/setuptools/_distutils/command/__pycache__/install_lib.cpython-39.pyc000064400000012104150564745010024450 0ustar00a

(�Re� �@sLdZddlZddlZddlZddlmZddlmZdZ	Gdd�de�Z
dS)zkdistutils.command.install_lib

Implements the Distutils 'install_lib' command
(install all Python modules).�N)�Command)�DistutilsOptionErrorz.pyc@sxeZdZdZgd�Zgd�ZddiZdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�ZdS)�install_libz7install all Python modules (extensions and pure Python)))zinstall-dir=�dzdirectory to install to)z
build-dir=�bz'build directory (where to install from))�force�fz-force installation (overwrite existing files))�compile�czcompile .py to .pyc [default])�
no-compileNzdon't compile .py files)z	optimize=�Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])�
skip-buildNzskip the build steps)rr	r
rr	cCs(d|_d|_d|_d|_d|_d|_dS)Nr)�install_dir�	build_dirrr	�optimize�
skip_build��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_distutils/command/install_lib.py�initialize_options3szinstall_lib.initialize_optionsc	Cs�|�ddddddd�|jdur&d|_|jdur6d	|_t|jt�s�zt|j�|_|jd
vr^t�Wnttfy~td��Yn0dS)N�install)�	build_libr)rr)rr)r	r	)rr)rrTF)r��zoptimize must be 0, 1, or 2)�set_undefined_optionsr	r�
isinstance�int�AssertionError�
ValueErrorrrrrr�finalize_options<s&�	


zinstall_lib.finalize_optionscCs0|��|��}|dur,|j��r,|�|�dS�N)�buildr�distribution�has_pure_modules�byte_compile�rZoutfilesrrr�runVszinstall_lib.runcCs2|js.|j��r|�d�|j��r.|�d�dS)N�build_py�	build_ext)rr#r$�run_command�has_ext_modulesrrrrr"fs



zinstall_lib.buildcCs8tj�|j�r |�|j|j�}n|�d|j�dS|S)Nz3'%s' does not exist -- no Python modules to install)�os�path�isdirr�	copy_treer�warnr&rrrrms�zinstall_lib.installcCsrtjr|�d�dSddlm}|�d�j}|jrH||d|j||j	d�|j
dkrn|||j
|j||j|j	d�dS)Nz%byte-compiling is disabled, skipping.r)r%r)rr�prefix�dry_run)rrr1�verboser2)�sys�dont_write_bytecoder0�distutils.utilr%�get_finalized_command�rootr	rr2rr3)r�filesr%Zinstall_rootrrrr%vs
�
�zinstall_lib.byte_compilec
	Csd|sgS|�|�}|��}t||�}t|�ttj�}g}|D] }	|�tj�||	|d���q>|Sr!)	r7�get_outputs�getattr�lenr,�sep�appendr-�join)
rZhas_anyZ	build_cmdZ
cmd_option�
output_dirZbuild_filesr�
prefix_len�outputs�filerrr�_mutate_outputs�s

zinstall_lib._mutate_outputscCsrg}|D]d}tj�tj�|��d}|tkr.q|jrJ|�tjj	|dd��|j
dkr|�tjj	||j
d��q|S)Nr�)�optimizationr)r,r-�splitext�normcase�PYTHON_SOURCE_EXTENSIONr	r>�	importlib�util�cache_from_sourcer)rZpy_filenamesZbytecode_filesZpy_file�extrrr�_bytecode_filenames�s
�

�
zinstall_lib._bytecode_filenamescCsR|�|j��dd|j�}|jr*|�|�}ng}|�|j��dd|j�}|||S)z�Return the list of files that would be installed if this command
        were actually run.  Not affected by the "dry-run" flag or whether
        modules have actually been built yet.
        r(rr))rDr#r$rr	rNr+)rZpure_outputsZbytecode_outputsZext_outputsrrrr:�s����zinstall_lib.get_outputscCsLg}|j��r&|�d�}|�|���|j��rH|�d�}|�|���|S)z�Get the list of files that are input to this command, ie. the
        files that get installed as they are named in the build tree.
        The files in this list correspond one-to-one to the output
        filenames returned by 'get_outputs()'.
        r(r))r#r$r7�extendr:r+)r�inputsr(r)rrr�
get_inputs�s



zinstall_lib.get_inputsN)�__name__�
__module__�__qualname__�description�user_options�boolean_options�negative_optrr r'r"rr%rDrNr:rQrrrrrs		r)�__doc__r,�importlib.utilrJr4�distutils.corer�distutils.errorsrrIrrrrr�<module>ssite-packages/setuptools/_distutils/command/check.py000064400000013005150564745010016743 0ustar00"""distutils.command.check

Implements the Distutils 'check' command.
"""
from distutils.core import Command
from distutils.errors import DistutilsSetupError

try:
    # docutils is installed
    from docutils.utils import Reporter
    from docutils.parsers.rst import Parser
    from docutils import frontend
    from docutils import nodes

    class SilentReporter(Reporter):

        def __init__(self, source, report_level, halt_level, stream=None,
                     debug=0, encoding='ascii', error_handler='replace'):
            self.messages = []
            Reporter.__init__(self, source, report_level, halt_level, stream,
                              debug, encoding, error_handler)

        def system_message(self, level, message, *children, **kwargs):
            self.messages.append((level, message, children, kwargs))
            return nodes.system_message(message, level=level,
                                        type=self.levels[level],
                                        *children, **kwargs)

    HAS_DOCUTILS = True
except Exception:
    # Catch all exceptions because exceptions besides ImportError probably
    # indicate that docutils is not ported to Py3k.
    HAS_DOCUTILS = False

class check(Command):
    """This command checks the meta-data of the package.
    """
    description = ("perform some checks on the package")
    user_options = [('metadata', 'm', 'Verify meta-data'),
                    ('restructuredtext', 'r',
                     ('Checks if long string meta-data syntax '
                      'are reStructuredText-compliant')),
                    ('strict', 's',
                     'Will exit with an error if a check fails')]

    boolean_options = ['metadata', 'restructuredtext', 'strict']

    def initialize_options(self):
        """Sets default values for options."""
        self.restructuredtext = 0
        self.metadata = 1
        self.strict = 0
        self._warnings = 0

    def finalize_options(self):
        pass

    def warn(self, msg):
        """Counts the number of warnings that occurs."""
        self._warnings += 1
        return Command.warn(self, msg)

    def run(self):
        """Runs the command."""
        # perform the various tests
        if self.metadata:
            self.check_metadata()
        if self.restructuredtext:
            if HAS_DOCUTILS:
                self.check_restructuredtext()
            elif self.strict:
                raise DistutilsSetupError('The docutils package is needed.')

        # let's raise an error in strict mode, if we have at least
        # one warning
        if self.strict and self._warnings > 0:
            raise DistutilsSetupError('Please correct your package.')

    def check_metadata(self):
        """Ensures that all required elements of meta-data are supplied.

        Required fields:
            name, version, URL

        Recommended fields:
            (author and author_email) or (maintainer and maintainer_email))

        Warns if any are missing.
        """
        metadata = self.distribution.metadata

        missing = []
        for attr in ('name', 'version', 'url'):
            if not (hasattr(metadata, attr) and getattr(metadata, attr)):
                missing.append(attr)

        if missing:
            self.warn("missing required meta-data: %s"  % ', '.join(missing))
        if metadata.author:
            if not metadata.author_email:
                self.warn("missing meta-data: if 'author' supplied, " +
                          "'author_email' should be supplied too")
        elif metadata.maintainer:
            if not metadata.maintainer_email:
                self.warn("missing meta-data: if 'maintainer' supplied, " +
                          "'maintainer_email' should be supplied too")
        else:
            self.warn("missing meta-data: either (author and author_email) " +
                      "or (maintainer and maintainer_email) " +
                      "should be supplied")

    def check_restructuredtext(self):
        """Checks if the long string fields are reST-compliant."""
        data = self.distribution.get_long_description()
        for warning in self._check_rst_data(data):
            line = warning[-1].get('line')
            if line is None:
                warning = warning[1]
            else:
                warning = '%s (line %s)' % (warning[1], line)
            self.warn(warning)

    def _check_rst_data(self, data):
        """Returns warnings when the provided data doesn't compile."""
        # the include and csv_table directives need this to be a path
        source_path = self.distribution.script_name or 'setup.py'
        parser = Parser()
        settings = frontend.OptionParser(components=(Parser,)).get_default_values()
        settings.tab_width = 4
        settings.pep_references = None
        settings.rfc_references = None
        reporter = SilentReporter(source_path,
                          settings.report_level,
                          settings.halt_level,
                          stream=settings.warning_stream,
                          debug=settings.debug,
                          encoding=settings.error_encoding,
                          error_handler=settings.error_encoding_error_handler)

        document = nodes.document(settings, reporter, source=source_path)
        document.note_source(source_path, -1)
        try:
            parser.parse(data, document)
        except AttributeError as e:
            reporter.messages.append(
                (-1, 'Could not finish the parsing: %s.' % e, '', {}))

        return reporter.messages
site-packages/setuptools/_distutils/command/bdist_msi.py000064400000105373150564745010017655 0ustar00# Copyright (C) 2005, 2006 Martin von Löwis
# Licensed to PSF under a Contributor Agreement.
# The bdist_wininst command proper
# based on bdist_wininst
"""
Implements the bdist_msi command.
"""

import os
import sys
import warnings
from distutils.core import Command
from distutils.dir_util import remove_tree
from distutils.sysconfig import get_python_version
from distutils.version import StrictVersion
from distutils.errors import DistutilsOptionError
from distutils.util import get_platform
from distutils import log
import msilib
from msilib import schema, sequence, text
from msilib import Directory, Feature, Dialog, add_data

class PyDialog(Dialog):
    """Dialog class with a fixed layout: controls at the top, then a ruler,
    then a list of buttons: back, next, cancel. Optionally a bitmap at the
    left."""
    def __init__(self, *args, **kw):
        """Dialog(database, name, x, y, w, h, attributes, title, first,
        default, cancel, bitmap=true)"""
        Dialog.__init__(self, *args)
        ruler = self.h - 36
        bmwidth = 152*ruler/328
        #if kw.get("bitmap", True):
        #    self.bitmap("Bitmap", 0, 0, bmwidth, ruler, "PythonWin")
        self.line("BottomLine", 0, ruler, self.w, 0)

    def title(self, title):
        "Set the title text of the dialog at the top."
        # name, x, y, w, h, flags=Visible|Enabled|Transparent|NoPrefix,
        # text, in VerdanaBold10
        self.text("Title", 15, 10, 320, 60, 0x30003,
                  r"{\VerdanaBold10}%s" % title)

    def back(self, title, next, name = "Back", active = 1):
        """Add a back button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associated"""
        if active:
            flags = 3 # Visible|Enabled
        else:
            flags = 1 # Visible
        return self.pushbutton(name, 180, self.h-27 , 56, 17, flags, title, next)

    def cancel(self, title, next, name = "Cancel", active = 1):
        """Add a cancel button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associated"""
        if active:
            flags = 3 # Visible|Enabled
        else:
            flags = 1 # Visible
        return self.pushbutton(name, 304, self.h-27, 56, 17, flags, title, next)

    def next(self, title, next, name = "Next", active = 1):
        """Add a Next button with a given title, the tab-next button,
        its name in the Control table, possibly initially disabled.

        Return the button, so that events can be associated"""
        if active:
            flags = 3 # Visible|Enabled
        else:
            flags = 1 # Visible
        return self.pushbutton(name, 236, self.h-27, 56, 17, flags, title, next)

    def xbutton(self, name, title, next, xpos):
        """Add a button with a given title, the tab-next button,
        its name in the Control table, giving its x position; the
        y-position is aligned with the other buttons.

        Return the button, so that events can be associated"""
        return self.pushbutton(name, int(self.w*xpos - 28), self.h-27, 56, 17, 3, title, next)

class bdist_msi(Command):

    description = "create a Microsoft Installer (.msi) binary distribution"

    user_options = [('bdist-dir=', None,
                     "temporary directory for creating the distribution"),
                    ('plat-name=', 'p',
                     "platform name to embed in generated filenames "
                     "(default: %s)" % get_platform()),
                    ('keep-temp', 'k',
                     "keep the pseudo-installation tree around after " +
                     "creating the distribution archive"),
                    ('target-version=', None,
                     "require a specific python version" +
                     " on the target system"),
                    ('no-target-compile', 'c',
                     "do not compile .py to .pyc on the target system"),
                    ('no-target-optimize', 'o',
                     "do not compile .py to .pyo (optimized) "
                     "on the target system"),
                    ('dist-dir=', 'd',
                     "directory to put final built distributions in"),
                    ('skip-build', None,
                     "skip rebuilding everything (for testing/debugging)"),
                    ('install-script=', None,
                     "basename of installation script to be run after "
                     "installation or before deinstallation"),
                    ('pre-install-script=', None,
                     "Fully qualified filename of a script to be run before "
                     "any files are installed.  This script need not be in the "
                     "distribution"),
                   ]

    boolean_options = ['keep-temp', 'no-target-compile', 'no-target-optimize',
                       'skip-build']

    all_versions = ['2.0', '2.1', '2.2', '2.3', '2.4',
                    '2.5', '2.6', '2.7', '2.8', '2.9',
                    '3.0', '3.1', '3.2', '3.3', '3.4',
                    '3.5', '3.6', '3.7', '3.8', '3.9']
    other_version = 'X'

    def __init__(self, *args, **kw):
        super().__init__(*args, **kw)
        warnings.warn("bdist_msi command is deprecated since Python 3.9, "
                      "use bdist_wheel (wheel packages) instead",
                      DeprecationWarning, 2)

    def initialize_options(self):
        self.bdist_dir = None
        self.plat_name = None
        self.keep_temp = 0
        self.no_target_compile = 0
        self.no_target_optimize = 0
        self.target_version = None
        self.dist_dir = None
        self.skip_build = None
        self.install_script = None
        self.pre_install_script = None
        self.versions = None

    def finalize_options(self):
        self.set_undefined_options('bdist', ('skip_build', 'skip_build'))

        if self.bdist_dir is None:
            bdist_base = self.get_finalized_command('bdist').bdist_base
            self.bdist_dir = os.path.join(bdist_base, 'msi')

        short_version = get_python_version()
        if (not self.target_version) and self.distribution.has_ext_modules():
            self.target_version = short_version

        if self.target_version:
            self.versions = [self.target_version]
            if not self.skip_build and self.distribution.has_ext_modules()\
               and self.target_version != short_version:
                raise DistutilsOptionError(
                      "target version can only be %s, or the '--skip-build'"
                      " option must be specified" % (short_version,))
        else:
            self.versions = list(self.all_versions)

        self.set_undefined_options('bdist',
                                   ('dist_dir', 'dist_dir'),
                                   ('plat_name', 'plat_name'),
                                   )

        if self.pre_install_script:
            raise DistutilsOptionError(
                  "the pre-install-script feature is not yet implemented")

        if self.install_script:
            for script in self.distribution.scripts:
                if self.install_script == os.path.basename(script):
                    break
            else:
                raise DistutilsOptionError(
                      "install_script '%s' not found in scripts"
                      % self.install_script)
        self.install_script_key = None

    def run(self):
        if not self.skip_build:
            self.run_command('build')

        install = self.reinitialize_command('install', reinit_subcommands=1)
        install.prefix = self.bdist_dir
        install.skip_build = self.skip_build
        install.warn_dir = 0

        install_lib = self.reinitialize_command('install_lib')
        # we do not want to include pyc or pyo files
        install_lib.compile = 0
        install_lib.optimize = 0

        if self.distribution.has_ext_modules():
            # If we are building an installer for a Python version other
            # than the one we are currently running, then we need to ensure
            # our build_lib reflects the other Python version rather than ours.
            # Note that for target_version!=sys.version, we must have skipped the
            # build step, so there is no issue with enforcing the build of this
            # version.
            target_version = self.target_version
            if not target_version:
                assert self.skip_build, "Should have already checked this"
                target_version = '%d.%d' % sys.version_info[:2]
            plat_specifier = ".%s-%s" % (self.plat_name, target_version)
            build = self.get_finalized_command('build')
            build.build_lib = os.path.join(build.build_base,
                                           'lib' + plat_specifier)

        log.info("installing to %s", self.bdist_dir)
        install.ensure_finalized()

        # avoid warning of 'install_lib' about installing
        # into a directory not in sys.path
        sys.path.insert(0, os.path.join(self.bdist_dir, 'PURELIB'))

        install.run()

        del sys.path[0]

        self.mkpath(self.dist_dir)
        fullname = self.distribution.get_fullname()
        installer_name = self.get_installer_filename(fullname)
        installer_name = os.path.abspath(installer_name)
        if os.path.exists(installer_name): os.unlink(installer_name)

        metadata = self.distribution.metadata
        author = metadata.author
        if not author:
            author = metadata.maintainer
        if not author:
            author = "UNKNOWN"
        version = metadata.get_version()
        # ProductVersion must be strictly numeric
        # XXX need to deal with prerelease versions
        sversion = "%d.%d.%d" % StrictVersion(version).version
        # Prefix ProductName with Python x.y, so that
        # it sorts together with the other Python packages
        # in Add-Remove-Programs (APR)
        fullname = self.distribution.get_fullname()
        if self.target_version:
            product_name = "Python %s %s" % (self.target_version, fullname)
        else:
            product_name = "Python %s" % (fullname)
        self.db = msilib.init_database(installer_name, schema,
                product_name, msilib.gen_uuid(),
                sversion, author)
        msilib.add_tables(self.db, sequence)
        props = [('DistVersion', version)]
        email = metadata.author_email or metadata.maintainer_email
        if email:
            props.append(("ARPCONTACT", email))
        if metadata.url:
            props.append(("ARPURLINFOABOUT", metadata.url))
        if props:
            add_data(self.db, 'Property', props)

        self.add_find_python()
        self.add_files()
        self.add_scripts()
        self.add_ui()
        self.db.Commit()

        if hasattr(self.distribution, 'dist_files'):
            tup = 'bdist_msi', self.target_version or 'any', fullname
            self.distribution.dist_files.append(tup)

        if not self.keep_temp:
            remove_tree(self.bdist_dir, dry_run=self.dry_run)

    def add_files(self):
        db = self.db
        cab = msilib.CAB("distfiles")
        rootdir = os.path.abspath(self.bdist_dir)

        root = Directory(db, cab, None, rootdir, "TARGETDIR", "SourceDir")
        f = Feature(db, "Python", "Python", "Everything",
                    0, 1, directory="TARGETDIR")

        items = [(f, root, '')]
        for version in self.versions + [self.other_version]:
            target = "TARGETDIR" + version
            name = default = "Python" + version
            desc = "Everything"
            if version is self.other_version:
                title = "Python from another location"
                level = 2
            else:
                title = "Python %s from registry" % version
                level = 1
            f = Feature(db, name, title, desc, 1, level, directory=target)
            dir = Directory(db, cab, root, rootdir, target, default)
            items.append((f, dir, version))
        db.Commit()

        seen = {}
        for feature, dir, version in items:
            todo = [dir]
            while todo:
                dir = todo.pop()
                for file in os.listdir(dir.absolute):
                    afile = os.path.join(dir.absolute, file)
                    if os.path.isdir(afile):
                        short = "%s|%s" % (dir.make_short(file), file)
                        default = file + version
                        newdir = Directory(db, cab, dir, file, default, short)
                        todo.append(newdir)
                    else:
                        if not dir.component:
                            dir.start_component(dir.logical, feature, 0)
                        if afile not in seen:
                            key = seen[afile] = dir.add_file(file)
                            if file==self.install_script:
                                if self.install_script_key:
                                    raise DistutilsOptionError(
                                          "Multiple files with name %s" % file)
                                self.install_script_key = '[#%s]' % key
                        else:
                            key = seen[afile]
                            add_data(self.db, "DuplicateFile",
                                [(key + version, dir.component, key, None, dir.logical)])
            db.Commit()
        cab.commit(db)

    def add_find_python(self):
        """Adds code to the installer to compute the location of Python.

        Properties PYTHON.MACHINE.X.Y and PYTHON.USER.X.Y will be set from the
        registry for each version of Python.

        Properties TARGETDIRX.Y will be set from PYTHON.USER.X.Y if defined,
        else from PYTHON.MACHINE.X.Y.

        Properties PYTHONX.Y will be set to TARGETDIRX.Y\\python.exe"""

        start = 402
        for ver in self.versions:
            install_path = r"SOFTWARE\Python\PythonCore\%s\InstallPath" % ver
            machine_reg = "python.machine." + ver
            user_reg = "python.user." + ver
            machine_prop = "PYTHON.MACHINE." + ver
            user_prop = "PYTHON.USER." + ver
            machine_action = "PythonFromMachine" + ver
            user_action = "PythonFromUser" + ver
            exe_action = "PythonExe" + ver
            target_dir_prop = "TARGETDIR" + ver
            exe_prop = "PYTHON" + ver
            if msilib.Win64:
                # type: msidbLocatorTypeRawValue + msidbLocatorType64bit
                Type = 2+16
            else:
                Type = 2
            add_data(self.db, "RegLocator",
                    [(machine_reg, 2, install_path, None, Type),
                     (user_reg, 1, install_path, None, Type)])
            add_data(self.db, "AppSearch",
                    [(machine_prop, machine_reg),
                     (user_prop, user_reg)])
            add_data(self.db, "CustomAction",
                    [(machine_action, 51+256, target_dir_prop, "[" + machine_prop + "]"),
                     (user_action, 51+256, target_dir_prop, "[" + user_prop + "]"),
                     (exe_action, 51+256, exe_prop, "[" + target_dir_prop + "]\\python.exe"),
                    ])
            add_data(self.db, "InstallExecuteSequence",
                    [(machine_action, machine_prop, start),
                     (user_action, user_prop, start + 1),
                     (exe_action, None, start + 2),
                    ])
            add_data(self.db, "InstallUISequence",
                    [(machine_action, machine_prop, start),
                     (user_action, user_prop, start + 1),
                     (exe_action, None, start + 2),
                    ])
            add_data(self.db, "Condition",
                    [("Python" + ver, 0, "NOT TARGETDIR" + ver)])
            start += 4
            assert start < 500

    def add_scripts(self):
        if self.install_script:
            start = 6800
            for ver in self.versions + [self.other_version]:
                install_action = "install_script." + ver
                exe_prop = "PYTHON" + ver
                add_data(self.db, "CustomAction",
                        [(install_action, 50, exe_prop, self.install_script_key)])
                add_data(self.db, "InstallExecuteSequence",
                        [(install_action, "&Python%s=3" % ver, start)])
                start += 1
        # XXX pre-install scripts are currently refused in finalize_options()
        #     but if this feature is completed, it will also need to add
        #     entries for each version as the above code does
        if self.pre_install_script:
            scriptfn = os.path.join(self.bdist_dir, "preinstall.bat")
            with open(scriptfn, "w") as f:
                # The batch file will be executed with [PYTHON], so that %1
                # is the path to the Python interpreter; %0 will be the path
                # of the batch file.
                # rem ="""
                # %1 %0
                # exit
                # """
                # <actual script>
                f.write('rem ="""\n%1 %0\nexit\n"""\n')
                with open(self.pre_install_script) as fin:
                    f.write(fin.read())
            add_data(self.db, "Binary",
                [("PreInstall", msilib.Binary(scriptfn))
                ])
            add_data(self.db, "CustomAction",
                [("PreInstall", 2, "PreInstall", None)
                ])
            add_data(self.db, "InstallExecuteSequence",
                    [("PreInstall", "NOT Installed", 450)])


    def add_ui(self):
        db = self.db
        x = y = 50
        w = 370
        h = 300
        title = "[ProductName] Setup"

        # see "Dialog Style Bits"
        modal = 3      # visible | modal
        modeless = 1   # visible
        track_disk_space = 32

        # UI customization properties
        add_data(db, "Property",
                 # See "DefaultUIFont Property"
                 [("DefaultUIFont", "DlgFont8"),
                  # See "ErrorDialog Style Bit"
                  ("ErrorDialog", "ErrorDlg"),
                  ("Progress1", "Install"),   # modified in maintenance type dlg
                  ("Progress2", "installs"),
                  ("MaintenanceForm_Action", "Repair"),
                  # possible values: ALL, JUSTME
                  ("WhichUsers", "ALL")
                 ])

        # Fonts, see "TextStyle Table"
        add_data(db, "TextStyle",
                 [("DlgFont8", "Tahoma", 9, None, 0),
                  ("DlgFontBold8", "Tahoma", 8, None, 1), #bold
                  ("VerdanaBold10", "Verdana", 10, None, 1),
                  ("VerdanaRed9", "Verdana", 9, 255, 0),
                 ])

        # UI Sequences, see "InstallUISequence Table", "Using a Sequence Table"
        # Numbers indicate sequence; see sequence.py for how these action integrate
        add_data(db, "InstallUISequence",
                 [("PrepareDlg", "Not Privileged or Windows9x or Installed", 140),
                  ("WhichUsersDlg", "Privileged and not Windows9x and not Installed", 141),
                  # In the user interface, assume all-users installation if privileged.
                  ("SelectFeaturesDlg", "Not Installed", 1230),
                  # XXX no support for resume installations yet
                  #("ResumeDlg", "Installed AND (RESUME OR Preselected)", 1240),
                  ("MaintenanceTypeDlg", "Installed AND NOT RESUME AND NOT Preselected", 1250),
                  ("ProgressDlg", None, 1280)])

        add_data(db, 'ActionText', text.ActionText)
        add_data(db, 'UIText', text.UIText)
        #####################################################################
        # Standard dialogs: FatalError, UserExit, ExitDialog
        fatal=PyDialog(db, "FatalError", x, y, w, h, modal, title,
                     "Finish", "Finish", "Finish")
        fatal.title("[ProductName] Installer ended prematurely")
        fatal.back("< Back", "Finish", active = 0)
        fatal.cancel("Cancel", "Back", active = 0)
        fatal.text("Description1", 15, 70, 320, 80, 0x30003,
                   "[ProductName] setup ended prematurely because of an error.  Your system has not been modified.  To install this program at a later time, please run the installation again.")
        fatal.text("Description2", 15, 155, 320, 20, 0x30003,
                   "Click the Finish button to exit the Installer.")
        c=fatal.next("Finish", "Cancel", name="Finish")
        c.event("EndDialog", "Exit")

        user_exit=PyDialog(db, "UserExit", x, y, w, h, modal, title,
                     "Finish", "Finish", "Finish")
        user_exit.title("[ProductName] Installer was interrupted")
        user_exit.back("< Back", "Finish", active = 0)
        user_exit.cancel("Cancel", "Back", active = 0)
        user_exit.text("Description1", 15, 70, 320, 80, 0x30003,
                   "[ProductName] setup was interrupted.  Your system has not been modified.  "
                   "To install this program at a later time, please run the installation again.")
        user_exit.text("Description2", 15, 155, 320, 20, 0x30003,
                   "Click the Finish button to exit the Installer.")
        c = user_exit.next("Finish", "Cancel", name="Finish")
        c.event("EndDialog", "Exit")

        exit_dialog = PyDialog(db, "ExitDialog", x, y, w, h, modal, title,
                             "Finish", "Finish", "Finish")
        exit_dialog.title("Completing the [ProductName] Installer")
        exit_dialog.back("< Back", "Finish", active = 0)
        exit_dialog.cancel("Cancel", "Back", active = 0)
        exit_dialog.text("Description", 15, 235, 320, 20, 0x30003,
                   "Click the Finish button to exit the Installer.")
        c = exit_dialog.next("Finish", "Cancel", name="Finish")
        c.event("EndDialog", "Return")

        #####################################################################
        # Required dialog: FilesInUse, ErrorDlg
        inuse = PyDialog(db, "FilesInUse",
                         x, y, w, h,
                         19,                # KeepModeless|Modal|Visible
                         title,
                         "Retry", "Retry", "Retry", bitmap=False)
        inuse.text("Title", 15, 6, 200, 15, 0x30003,
                   r"{\DlgFontBold8}Files in Use")
        inuse.text("Description", 20, 23, 280, 20, 0x30003,
               "Some files that need to be updated are currently in use.")
        inuse.text("Text", 20, 55, 330, 50, 3,
                   "The following applications are using files that need to be updated by this setup. Close these applications and then click Retry to continue the installation or Cancel to exit it.")
        inuse.control("List", "ListBox", 20, 107, 330, 130, 7, "FileInUseProcess",
                      None, None, None)
        c=inuse.back("Exit", "Ignore", name="Exit")
        c.event("EndDialog", "Exit")
        c=inuse.next("Ignore", "Retry", name="Ignore")
        c.event("EndDialog", "Ignore")
        c=inuse.cancel("Retry", "Exit", name="Retry")
        c.event("EndDialog","Retry")

        # See "Error Dialog". See "ICE20" for the required names of the controls.
        error = Dialog(db, "ErrorDlg",
                       50, 10, 330, 101,
                       65543,       # Error|Minimize|Modal|Visible
                       title,
                       "ErrorText", None, None)
        error.text("ErrorText", 50,9,280,48,3, "")
        #error.control("ErrorIcon", "Icon", 15, 9, 24, 24, 5242881, None, "py.ico", None, None)
        error.pushbutton("N",120,72,81,21,3,"No",None).event("EndDialog","ErrorNo")
        error.pushbutton("Y",240,72,81,21,3,"Yes",None).event("EndDialog","ErrorYes")
        error.pushbutton("A",0,72,81,21,3,"Abort",None).event("EndDialog","ErrorAbort")
        error.pushbutton("C",42,72,81,21,3,"Cancel",None).event("EndDialog","ErrorCancel")
        error.pushbutton("I",81,72,81,21,3,"Ignore",None).event("EndDialog","ErrorIgnore")
        error.pushbutton("O",159,72,81,21,3,"Ok",None).event("EndDialog","ErrorOk")
        error.pushbutton("R",198,72,81,21,3,"Retry",None).event("EndDialog","ErrorRetry")

        #####################################################################
        # Global "Query Cancel" dialog
        cancel = Dialog(db, "CancelDlg", 50, 10, 260, 85, 3, title,
                        "No", "No", "No")
        cancel.text("Text", 48, 15, 194, 30, 3,
                    "Are you sure you want to cancel [ProductName] installation?")
        #cancel.control("Icon", "Icon", 15, 15, 24, 24, 5242881, None,
        #               "py.ico", None, None)
        c=cancel.pushbutton("Yes", 72, 57, 56, 17, 3, "Yes", "No")
        c.event("EndDialog", "Exit")

        c=cancel.pushbutton("No", 132, 57, 56, 17, 3, "No", "Yes")
        c.event("EndDialog", "Return")

        #####################################################################
        # Global "Wait for costing" dialog
        costing = Dialog(db, "WaitForCostingDlg", 50, 10, 260, 85, modal, title,
                         "Return", "Return", "Return")
        costing.text("Text", 48, 15, 194, 30, 3,
                     "Please wait while the installer finishes determining your disk space requirements.")
        c = costing.pushbutton("Return", 102, 57, 56, 17, 3, "Return", None)
        c.event("EndDialog", "Exit")

        #####################################################################
        # Preparation dialog: no user input except cancellation
        prep = PyDialog(db, "PrepareDlg", x, y, w, h, modeless, title,
                        "Cancel", "Cancel", "Cancel")
        prep.text("Description", 15, 70, 320, 40, 0x30003,
                  "Please wait while the Installer prepares to guide you through the installation.")
        prep.title("Welcome to the [ProductName] Installer")
        c=prep.text("ActionText", 15, 110, 320, 20, 0x30003, "Pondering...")
        c.mapping("ActionText", "Text")
        c=prep.text("ActionData", 15, 135, 320, 30, 0x30003, None)
        c.mapping("ActionData", "Text")
        prep.back("Back", None, active=0)
        prep.next("Next", None, active=0)
        c=prep.cancel("Cancel", None)
        c.event("SpawnDialog", "CancelDlg")

        #####################################################################
        # Feature (Python directory) selection
        seldlg = PyDialog(db, "SelectFeaturesDlg", x, y, w, h, modal, title,
                        "Next", "Next", "Cancel")
        seldlg.title("Select Python Installations")

        seldlg.text("Hint", 15, 30, 300, 20, 3,
                    "Select the Python locations where %s should be installed."
                    % self.distribution.get_fullname())

        seldlg.back("< Back", None, active=0)
        c = seldlg.next("Next >", "Cancel")
        order = 1
        c.event("[TARGETDIR]", "[SourceDir]", ordering=order)
        for version in self.versions + [self.other_version]:
            order += 1
            c.event("[TARGETDIR]", "[TARGETDIR%s]" % version,
                    "FEATURE_SELECTED AND &Python%s=3" % version,
                    ordering=order)
        c.event("SpawnWaitDialog", "WaitForCostingDlg", ordering=order + 1)
        c.event("EndDialog", "Return", ordering=order + 2)
        c = seldlg.cancel("Cancel", "Features")
        c.event("SpawnDialog", "CancelDlg")

        c = seldlg.control("Features", "SelectionTree", 15, 60, 300, 120, 3,
                           "FEATURE", None, "PathEdit", None)
        c.event("[FEATURE_SELECTED]", "1")
        ver = self.other_version
        install_other_cond = "FEATURE_SELECTED AND &Python%s=3" % ver
        dont_install_other_cond = "FEATURE_SELECTED AND &Python%s<>3" % ver

        c = seldlg.text("Other", 15, 200, 300, 15, 3,
                        "Provide an alternate Python location")
        c.condition("Enable", install_other_cond)
        c.condition("Show", install_other_cond)
        c.condition("Disable", dont_install_other_cond)
        c.condition("Hide", dont_install_other_cond)

        c = seldlg.control("PathEdit", "PathEdit", 15, 215, 300, 16, 1,
                           "TARGETDIR" + ver, None, "Next", None)
        c.condition("Enable", install_other_cond)
        c.condition("Show", install_other_cond)
        c.condition("Disable", dont_install_other_cond)
        c.condition("Hide", dont_install_other_cond)

        #####################################################################
        # Disk cost
        cost = PyDialog(db, "DiskCostDlg", x, y, w, h, modal, title,
                        "OK", "OK", "OK", bitmap=False)
        cost.text("Title", 15, 6, 200, 15, 0x30003,
                 r"{\DlgFontBold8}Disk Space Requirements")
        cost.text("Description", 20, 20, 280, 20, 0x30003,
                  "The disk space required for the installation of the selected features.")
        cost.text("Text", 20, 53, 330, 60, 3,
                  "The highlighted volumes (if any) do not have enough disk space "
              "available for the currently selected features.  You can either "
              "remove some files from the highlighted volumes, or choose to "
              "install less features onto local drive(s), or select different "
              "destination drive(s).")
        cost.control("VolumeList", "VolumeCostList", 20, 100, 330, 150, 393223,
                     None, "{120}{70}{70}{70}{70}", None, None)
        cost.xbutton("OK", "Ok", None, 0.5).event("EndDialog", "Return")

        #####################################################################
        # WhichUsers Dialog. Only available on NT, and for privileged users.
        # This must be run before FindRelatedProducts, because that will
        # take into account whether the previous installation was per-user
        # or per-machine. We currently don't support going back to this
        # dialog after "Next" was selected; to support this, we would need to
        # find how to reset the ALLUSERS property, and how to re-run
        # FindRelatedProducts.
        # On Windows9x, the ALLUSERS property is ignored on the command line
        # and in the Property table, but installer fails according to the documentation
        # if a dialog attempts to set ALLUSERS.
        whichusers = PyDialog(db, "WhichUsersDlg", x, y, w, h, modal, title,
                            "AdminInstall", "Next", "Cancel")
        whichusers.title("Select whether to install [ProductName] for all users of this computer.")
        # A radio group with two options: allusers, justme
        g = whichusers.radiogroup("AdminInstall", 15, 60, 260, 50, 3,
                                  "WhichUsers", "", "Next")
        g.add("ALL", 0, 5, 150, 20, "Install for all users")
        g.add("JUSTME", 0, 25, 150, 20, "Install just for me")

        whichusers.back("Back", None, active=0)

        c = whichusers.next("Next >", "Cancel")
        c.event("[ALLUSERS]", "1", 'WhichUsers="ALL"', 1)
        c.event("EndDialog", "Return", ordering = 2)

        c = whichusers.cancel("Cancel", "AdminInstall")
        c.event("SpawnDialog", "CancelDlg")

        #####################################################################
        # Installation Progress dialog (modeless)
        progress = PyDialog(db, "ProgressDlg", x, y, w, h, modeless, title,
                            "Cancel", "Cancel", "Cancel", bitmap=False)
        progress.text("Title", 20, 15, 200, 15, 0x30003,
                     r"{\DlgFontBold8}[Progress1] [ProductName]")
        progress.text("Text", 35, 65, 300, 30, 3,
                      "Please wait while the Installer [Progress2] [ProductName]. "
                      "This may take several minutes.")
        progress.text("StatusLabel", 35, 100, 35, 20, 3, "Status:")

        c=progress.text("ActionText", 70, 100, w-70, 20, 3, "Pondering...")
        c.mapping("ActionText", "Text")

        #c=progress.text("ActionData", 35, 140, 300, 20, 3, None)
        #c.mapping("ActionData", "Text")

        c=progress.control("ProgressBar", "ProgressBar", 35, 120, 300, 10, 65537,
                           None, "Progress done", None, None)
        c.mapping("SetProgress", "Progress")

        progress.back("< Back", "Next", active=False)
        progress.next("Next >", "Cancel", active=False)
        progress.cancel("Cancel", "Back").event("SpawnDialog", "CancelDlg")

        ###################################################################
        # Maintenance type: repair/uninstall
        maint = PyDialog(db, "MaintenanceTypeDlg", x, y, w, h, modal, title,
                         "Next", "Next", "Cancel")
        maint.title("Welcome to the [ProductName] Setup Wizard")
        maint.text("BodyText", 15, 63, 330, 42, 3,
                   "Select whether you want to repair or remove [ProductName].")
        g=maint.radiogroup("RepairRadioGroup", 15, 108, 330, 60, 3,
                            "MaintenanceForm_Action", "", "Next")
        #g.add("Change", 0, 0, 200, 17, "&Change [ProductName]")
        g.add("Repair", 0, 18, 200, 17, "&Repair [ProductName]")
        g.add("Remove", 0, 36, 200, 17, "Re&move [ProductName]")

        maint.back("< Back", None, active=False)
        c=maint.next("Finish", "Cancel")
        # Change installation: Change progress dialog to "Change", then ask
        # for feature selection
        #c.event("[Progress1]", "Change", 'MaintenanceForm_Action="Change"', 1)
        #c.event("[Progress2]", "changes", 'MaintenanceForm_Action="Change"', 2)

        # Reinstall: Change progress dialog to "Repair", then invoke reinstall
        # Also set list of reinstalled features to "ALL"
        c.event("[REINSTALL]", "ALL", 'MaintenanceForm_Action="Repair"', 5)
        c.event("[Progress1]", "Repairing", 'MaintenanceForm_Action="Repair"', 6)
        c.event("[Progress2]", "repairs", 'MaintenanceForm_Action="Repair"', 7)
        c.event("Reinstall", "ALL", 'MaintenanceForm_Action="Repair"', 8)

        # Uninstall: Change progress to "Remove", then invoke uninstall
        # Also set list of removed features to "ALL"
        c.event("[REMOVE]", "ALL", 'MaintenanceForm_Action="Remove"', 11)
        c.event("[Progress1]", "Removing", 'MaintenanceForm_Action="Remove"', 12)
        c.event("[Progress2]", "removes", 'MaintenanceForm_Action="Remove"', 13)
        c.event("Remove", "ALL", 'MaintenanceForm_Action="Remove"', 14)

        # Close dialog when maintenance action scheduled
        c.event("EndDialog", "Return", 'MaintenanceForm_Action<>"Change"', 20)
        #c.event("NewDialog", "SelectFeaturesDlg", 'MaintenanceForm_Action="Change"', 21)

        maint.cancel("Cancel", "RepairRadioGroup").event("SpawnDialog", "CancelDlg")

    def get_installer_filename(self, fullname):
        # Factored out to allow overriding in subclasses
        if self.target_version:
            base_name = "%s.%s-py%s.msi" % (fullname, self.plat_name,
                                            self.target_version)
        else:
            base_name = "%s.%s.msi" % (fullname, self.plat_name)
        installer_name = os.path.join(self.dist_dir, base_name)
        return installer_name
site-packages/setuptools/_distutils/command/install_lib.py000064400000020315150564745010020164 0ustar00"""distutils.command.install_lib

Implements the Distutils 'install_lib' command
(install all Python modules)."""

import os
import importlib.util
import sys

from distutils.core import Command
from distutils.errors import DistutilsOptionError


# Extension for Python source files.
PYTHON_SOURCE_EXTENSION = ".py"

class install_lib(Command):

    description = "install all Python modules (extensions and pure Python)"

    # The byte-compilation options are a tad confusing.  Here are the
    # possible scenarios:
    #   1) no compilation at all (--no-compile --no-optimize)
    #   2) compile .pyc only (--compile --no-optimize; default)
    #   3) compile .pyc and "opt-1" .pyc (--compile --optimize)
    #   4) compile "opt-1" .pyc only (--no-compile --optimize)
    #   5) compile .pyc and "opt-2" .pyc (--compile --optimize-more)
    #   6) compile "opt-2" .pyc only (--no-compile --optimize-more)
    #
    # The UI for this is two options, 'compile' and 'optimize'.
    # 'compile' is strictly boolean, and only decides whether to
    # generate .pyc files.  'optimize' is three-way (0, 1, or 2), and
    # decides both whether to generate .pyc files and what level of
    # optimization to use.

    user_options = [
        ('install-dir=', 'd', "directory to install to"),
        ('build-dir=','b', "build directory (where to install from)"),
        ('force', 'f', "force installation (overwrite existing files)"),
        ('compile', 'c', "compile .py to .pyc [default]"),
        ('no-compile', None, "don't compile .py files"),
        ('optimize=', 'O',
         "also compile with optimization: -O1 for \"python -O\", "
         "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"),
        ('skip-build', None, "skip the build steps"),
        ]

    boolean_options = ['force', 'compile', 'skip-build']
    negative_opt = {'no-compile' : 'compile'}

    def initialize_options(self):
        # let the 'install' command dictate our installation directory
        self.install_dir = None
        self.build_dir = None
        self.force = 0
        self.compile = None
        self.optimize = None
        self.skip_build = None

    def finalize_options(self):
        # Get all the information we need to install pure Python modules
        # from the umbrella 'install' command -- build (source) directory,
        # install (target) directory, and whether to compile .py files.
        self.set_undefined_options('install',
                                   ('build_lib', 'build_dir'),
                                   ('install_lib', 'install_dir'),
                                   ('force', 'force'),
                                   ('compile', 'compile'),
                                   ('optimize', 'optimize'),
                                   ('skip_build', 'skip_build'),
                                  )

        if self.compile is None:
            self.compile = True
        if self.optimize is None:
            self.optimize = False

        if not isinstance(self.optimize, int):
            try:
                self.optimize = int(self.optimize)
                if self.optimize not in (0, 1, 2):
                    raise AssertionError
            except (ValueError, AssertionError):
                raise DistutilsOptionError("optimize must be 0, 1, or 2")

    def run(self):
        # Make sure we have built everything we need first
        self.build()

        # Install everything: simply dump the entire contents of the build
        # directory to the installation directory (that's the beauty of
        # having a build directory!)
        outfiles = self.install()

        # (Optionally) compile .py to .pyc
        if outfiles is not None and self.distribution.has_pure_modules():
            self.byte_compile(outfiles)

    # -- Top-level worker functions ------------------------------------
    # (called from 'run()')

    def build(self):
        if not self.skip_build:
            if self.distribution.has_pure_modules():
                self.run_command('build_py')
            if self.distribution.has_ext_modules():
                self.run_command('build_ext')

    def install(self):
        if os.path.isdir(self.build_dir):
            outfiles = self.copy_tree(self.build_dir, self.install_dir)
        else:
            self.warn("'%s' does not exist -- no Python modules to install" %
                      self.build_dir)
            return
        return outfiles

    def byte_compile(self, files):
        if sys.dont_write_bytecode:
            self.warn('byte-compiling is disabled, skipping.')
            return

        from distutils.util import byte_compile

        # Get the "--root" directory supplied to the "install" command,
        # and use it as a prefix to strip off the purported filename
        # encoded in bytecode files.  This is far from complete, but it
        # should at least generate usable bytecode in RPM distributions.
        install_root = self.get_finalized_command('install').root

        if self.compile:
            byte_compile(files, optimize=0,
                         force=self.force, prefix=install_root,
                         dry_run=self.dry_run)
        if self.optimize > 0:
            byte_compile(files, optimize=self.optimize,
                         force=self.force, prefix=install_root,
                         verbose=self.verbose, dry_run=self.dry_run)


    # -- Utility methods -----------------------------------------------

    def _mutate_outputs(self, has_any, build_cmd, cmd_option, output_dir):
        if not has_any:
            return []

        build_cmd = self.get_finalized_command(build_cmd)
        build_files = build_cmd.get_outputs()
        build_dir = getattr(build_cmd, cmd_option)

        prefix_len = len(build_dir) + len(os.sep)
        outputs = []
        for file in build_files:
            outputs.append(os.path.join(output_dir, file[prefix_len:]))

        return outputs

    def _bytecode_filenames(self, py_filenames):
        bytecode_files = []
        for py_file in py_filenames:
            # Since build_py handles package data installation, the
            # list of outputs can contain more than just .py files.
            # Make sure we only report bytecode for the .py files.
            ext = os.path.splitext(os.path.normcase(py_file))[1]
            if ext != PYTHON_SOURCE_EXTENSION:
                continue
            if self.compile:
                bytecode_files.append(importlib.util.cache_from_source(
                    py_file, optimization=''))
            if self.optimize > 0:
                bytecode_files.append(importlib.util.cache_from_source(
                    py_file, optimization=self.optimize))

        return bytecode_files


    # -- External interface --------------------------------------------
    # (called by outsiders)

    def get_outputs(self):
        """Return the list of files that would be installed if this command
        were actually run.  Not affected by the "dry-run" flag or whether
        modules have actually been built yet.
        """
        pure_outputs = \
            self._mutate_outputs(self.distribution.has_pure_modules(),
                                 'build_py', 'build_lib',
                                 self.install_dir)
        if self.compile:
            bytecode_outputs = self._bytecode_filenames(pure_outputs)
        else:
            bytecode_outputs = []

        ext_outputs = \
            self._mutate_outputs(self.distribution.has_ext_modules(),
                                 'build_ext', 'build_lib',
                                 self.install_dir)

        return pure_outputs + bytecode_outputs + ext_outputs

    def get_inputs(self):
        """Get the list of files that are input to this command, ie. the
        files that get installed as they are named in the build tree.
        The files in this list correspond one-to-one to the output
        filenames returned by 'get_outputs()'.
        """
        inputs = []

        if self.distribution.has_pure_modules():
            build_py = self.get_finalized_command('build_py')
            inputs.extend(build_py.get_outputs())

        if self.distribution.has_ext_modules():
            build_ext = self.get_finalized_command('build_ext')
            inputs.extend(build_ext.get_outputs())

        return inputs
site-packages/setuptools/_distutils/command/install_data.py000064400000005406150564745010020333 0ustar00"""distutils.command.install_data

Implements the Distutils 'install_data' command, for installing
platform-independent data files."""

# contributed by Bastian Kleineidam

import os
from distutils.core import Command
from distutils.util import change_root, convert_path

class install_data(Command):

    description = "install data files"

    user_options = [
        ('install-dir=', 'd',
         "base directory for installing data files "
         "(default: installation base dir)"),
        ('root=', None,
         "install everything relative to this alternate root directory"),
        ('force', 'f', "force installation (overwrite existing files)"),
        ]

    boolean_options = ['force']

    def initialize_options(self):
        self.install_dir = None
        self.outfiles = []
        self.root = None
        self.force = 0
        self.data_files = self.distribution.data_files
        self.warn_dir = 1

    def finalize_options(self):
        self.set_undefined_options('install',
                                   ('install_data', 'install_dir'),
                                   ('root', 'root'),
                                   ('force', 'force'),
                                  )

    def run(self):
        self.mkpath(self.install_dir)
        for f in self.data_files:
            if isinstance(f, str):
                # it's a simple file, so copy it
                f = convert_path(f)
                if self.warn_dir:
                    self.warn("setup script did not provide a directory for "
                              "'%s' -- installing right in '%s'" %
                              (f, self.install_dir))
                (out, _) = self.copy_file(f, self.install_dir)
                self.outfiles.append(out)
            else:
                # it's a tuple with path to install to and a list of files
                dir = convert_path(f[0])
                if not os.path.isabs(dir):
                    dir = os.path.join(self.install_dir, dir)
                elif self.root:
                    dir = change_root(self.root, dir)
                self.mkpath(dir)

                if f[1] == []:
                    # If there are no files listed, the user must be
                    # trying to create an empty directory, so add the
                    # directory to the list of output files.
                    self.outfiles.append(dir)
                else:
                    # Copy files, adding them to the list of output files.
                    for data in f[1]:
                        data = convert_path(data)
                        (out, _) = self.copy_file(data, dir)
                        self.outfiles.append(out)

    def get_inputs(self):
        return self.data_files or []

    def get_outputs(self):
        return self.outfiles
site-packages/setuptools/_distutils/command/build_py.py000064400000040157150564745010017505 0ustar00"""distutils.command.build_py

Implements the Distutils 'build_py' command."""

import os
import importlib.util
import sys
import glob

from distutils.core import Command
from distutils.errors import *
from distutils.util import convert_path
from distutils import log

class build_py (Command):

    description = "\"build\" pure Python modules (copy to build directory)"

    user_options = [
        ('build-lib=', 'd', "directory to \"build\" (copy) to"),
        ('compile', 'c', "compile .py to .pyc"),
        ('no-compile', None, "don't compile .py files [default]"),
        ('optimize=', 'O',
         "also compile with optimization: -O1 for \"python -O\", "
         "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"),
        ('force', 'f', "forcibly build everything (ignore file timestamps)"),
        ]

    boolean_options = ['compile', 'force']
    negative_opt = {'no-compile' : 'compile'}

    def initialize_options(self):
        self.build_lib = None
        self.py_modules = None
        self.package = None
        self.package_data = None
        self.package_dir = None
        self.compile = 0
        self.optimize = 0
        self.force = None

    def finalize_options(self):
        self.set_undefined_options('build',
                                   ('build_lib', 'build_lib'),
                                   ('force', 'force'))

        # Get the distribution options that are aliases for build_py
        # options -- list of packages and list of modules.
        self.packages = self.distribution.packages
        self.py_modules = self.distribution.py_modules
        self.package_data = self.distribution.package_data
        self.package_dir = {}
        if self.distribution.package_dir:
            for name, path in self.distribution.package_dir.items():
                self.package_dir[name] = convert_path(path)
        self.data_files = self.get_data_files()

        # Ick, copied straight from install_lib.py (fancy_getopt needs a
        # type system!  Hell, *everything* needs a type system!!!)
        if not isinstance(self.optimize, int):
            try:
                self.optimize = int(self.optimize)
                assert 0 <= self.optimize <= 2
            except (ValueError, AssertionError):
                raise DistutilsOptionError("optimize must be 0, 1, or 2")

    def run(self):
        # XXX copy_file by default preserves atime and mtime.  IMHO this is
        # the right thing to do, but perhaps it should be an option -- in
        # particular, a site administrator might want installed files to
        # reflect the time of installation rather than the last
        # modification time before the installed release.

        # XXX copy_file by default preserves mode, which appears to be the
        # wrong thing to do: if a file is read-only in the working
        # directory, we want it to be installed read/write so that the next
        # installation of the same module distribution can overwrite it
        # without problems.  (This might be a Unix-specific issue.)  Thus
        # we turn off 'preserve_mode' when copying to the build directory,
        # since the build directory is supposed to be exactly what the
        # installation will look like (ie. we preserve mode when
        # installing).

        # Two options control which modules will be installed: 'packages'
        # and 'py_modules'.  The former lets us work with whole packages, not
        # specifying individual modules at all; the latter is for
        # specifying modules one-at-a-time.

        if self.py_modules:
            self.build_modules()
        if self.packages:
            self.build_packages()
            self.build_package_data()

        self.byte_compile(self.get_outputs(include_bytecode=0))

    def get_data_files(self):
        """Generate list of '(package,src_dir,build_dir,filenames)' tuples"""
        data = []
        if not self.packages:
            return data
        for package in self.packages:
            # Locate package source directory
            src_dir = self.get_package_dir(package)

            # Compute package build directory
            build_dir = os.path.join(*([self.build_lib] + package.split('.')))

            # Length of path to strip from found files
            plen = 0
            if src_dir:
                plen = len(src_dir)+1

            # Strip directory from globbed filenames
            filenames = [
                file[plen:] for file in self.find_data_files(package, src_dir)
                ]
            data.append((package, src_dir, build_dir, filenames))
        return data

    def find_data_files(self, package, src_dir):
        """Return filenames for package's data files in 'src_dir'"""
        globs = (self.package_data.get('', [])
                 + self.package_data.get(package, []))
        files = []
        for pattern in globs:
            # Each pattern has to be converted to a platform-specific path
            filelist = glob.glob(os.path.join(glob.escape(src_dir), convert_path(pattern)))
            # Files that match more than one pattern are only added once
            files.extend([fn for fn in filelist if fn not in files
                and os.path.isfile(fn)])
        return files

    def build_package_data(self):
        """Copy data files into build directory"""
        lastdir = None
        for package, src_dir, build_dir, filenames in self.data_files:
            for filename in filenames:
                target = os.path.join(build_dir, filename)
                self.mkpath(os.path.dirname(target))
                self.copy_file(os.path.join(src_dir, filename), target,
                               preserve_mode=False)

    def get_package_dir(self, package):
        """Return the directory, relative to the top of the source
           distribution, where package 'package' should be found
           (at least according to the 'package_dir' option, if any)."""
        path = package.split('.')

        if not self.package_dir:
            if path:
                return os.path.join(*path)
            else:
                return ''
        else:
            tail = []
            while path:
                try:
                    pdir = self.package_dir['.'.join(path)]
                except KeyError:
                    tail.insert(0, path[-1])
                    del path[-1]
                else:
                    tail.insert(0, pdir)
                    return os.path.join(*tail)
            else:
                # Oops, got all the way through 'path' without finding a
                # match in package_dir.  If package_dir defines a directory
                # for the root (nameless) package, then fallback on it;
                # otherwise, we might as well have not consulted
                # package_dir at all, as we just use the directory implied
                # by 'tail' (which should be the same as the original value
                # of 'path' at this point).
                pdir = self.package_dir.get('')
                if pdir is not None:
                    tail.insert(0, pdir)

                if tail:
                    return os.path.join(*tail)
                else:
                    return ''

    def check_package(self, package, package_dir):
        # Empty dir name means current directory, which we can probably
        # assume exists.  Also, os.path.exists and isdir don't know about
        # my "empty string means current dir" convention, so we have to
        # circumvent them.
        if package_dir != "":
            if not os.path.exists(package_dir):
                raise DistutilsFileError(
                      "package directory '%s' does not exist" % package_dir)
            if not os.path.isdir(package_dir):
                raise DistutilsFileError(
                       "supposed package directory '%s' exists, "
                       "but is not a directory" % package_dir)

        # Require __init__.py for all but the "root package"
        if package:
            init_py = os.path.join(package_dir, "__init__.py")
            if os.path.isfile(init_py):
                return init_py
            else:
                log.warn(("package init file '%s' not found " +
                          "(or not a regular file)"), init_py)

        # Either not in a package at all (__init__.py not expected), or
        # __init__.py doesn't exist -- so don't return the filename.
        return None

    def check_module(self, module, module_file):
        if not os.path.isfile(module_file):
            log.warn("file %s (for module %s) not found", module_file, module)
            return False
        else:
            return True

    def find_package_modules(self, package, package_dir):
        self.check_package(package, package_dir)
        module_files = glob.glob(os.path.join(glob.escape(package_dir), "*.py"))
        modules = []
        setup_script = os.path.abspath(self.distribution.script_name)

        for f in module_files:
            abs_f = os.path.abspath(f)
            if abs_f != setup_script:
                module = os.path.splitext(os.path.basename(f))[0]
                modules.append((package, module, f))
            else:
                self.debug_print("excluding %s" % setup_script)
        return modules

    def find_modules(self):
        """Finds individually-specified Python modules, ie. those listed by
        module name in 'self.py_modules'.  Returns a list of tuples (package,
        module_base, filename): 'package' is a tuple of the path through
        package-space to the module; 'module_base' is the bare (no
        packages, no dots) module name, and 'filename' is the path to the
        ".py" file (relative to the distribution root) that implements the
        module.
        """
        # Map package names to tuples of useful info about the package:
        #    (package_dir, checked)
        # package_dir - the directory where we'll find source files for
        #   this package
        # checked - true if we have checked that the package directory
        #   is valid (exists, contains __init__.py, ... ?)
        packages = {}

        # List of (package, module, filename) tuples to return
        modules = []

        # We treat modules-in-packages almost the same as toplevel modules,
        # just the "package" for a toplevel is empty (either an empty
        # string or empty list, depending on context).  Differences:
        #   - don't check for __init__.py in directory for empty package
        for module in self.py_modules:
            path = module.split('.')
            package = '.'.join(path[0:-1])
            module_base = path[-1]

            try:
                (package_dir, checked) = packages[package]
            except KeyError:
                package_dir = self.get_package_dir(package)
                checked = 0

            if not checked:
                init_py = self.check_package(package, package_dir)
                packages[package] = (package_dir, 1)
                if init_py:
                    modules.append((package, "__init__", init_py))

            # XXX perhaps we should also check for just .pyc files
            # (so greedy closed-source bastards can distribute Python
            # modules too)
            module_file = os.path.join(package_dir, module_base + ".py")
            if not self.check_module(module, module_file):
                continue

            modules.append((package, module_base, module_file))

        return modules

    def find_all_modules(self):
        """Compute the list of all modules that will be built, whether
        they are specified one-module-at-a-time ('self.py_modules') or
        by whole packages ('self.packages').  Return a list of tuples
        (package, module, module_file), just like 'find_modules()' and
        'find_package_modules()' do."""
        modules = []
        if self.py_modules:
            modules.extend(self.find_modules())
        if self.packages:
            for package in self.packages:
                package_dir = self.get_package_dir(package)
                m = self.find_package_modules(package, package_dir)
                modules.extend(m)
        return modules

    def get_source_files(self):
        return [module[-1] for module in self.find_all_modules()]

    def get_module_outfile(self, build_dir, package, module):
        outfile_path = [build_dir] + list(package) + [module + ".py"]
        return os.path.join(*outfile_path)

    def get_outputs(self, include_bytecode=1):
        modules = self.find_all_modules()
        outputs = []
        for (package, module, module_file) in modules:
            package = package.split('.')
            filename = self.get_module_outfile(self.build_lib, package, module)
            outputs.append(filename)
            if include_bytecode:
                if self.compile:
                    outputs.append(importlib.util.cache_from_source(
                        filename, optimization=''))
                if self.optimize > 0:
                    outputs.append(importlib.util.cache_from_source(
                        filename, optimization=self.optimize))

        outputs += [
            os.path.join(build_dir, filename)
            for package, src_dir, build_dir, filenames in self.data_files
            for filename in filenames
            ]

        return outputs

    def build_module(self, module, module_file, package):
        if isinstance(package, str):
            package = package.split('.')
        elif not isinstance(package, (list, tuple)):
            raise TypeError(
                  "'package' must be a string (dot-separated), list, or tuple")

        # Now put the module source file into the "build" area -- this is
        # easy, we just copy it somewhere under self.build_lib (the build
        # directory for Python source).
        outfile = self.get_module_outfile(self.build_lib, package, module)
        dir = os.path.dirname(outfile)
        self.mkpath(dir)
        return self.copy_file(module_file, outfile, preserve_mode=0)

    def build_modules(self):
        modules = self.find_modules()
        for (package, module, module_file) in modules:
            # Now "build" the module -- ie. copy the source file to
            # self.build_lib (the build directory for Python source).
            # (Actually, it gets copied to the directory for this package
            # under self.build_lib.)
            self.build_module(module, module_file, package)

    def build_packages(self):
        for package in self.packages:
            # Get list of (package, module, module_file) tuples based on
            # scanning the package directory.  'package' is only included
            # in the tuple so that 'find_modules()' and
            # 'find_package_tuples()' have a consistent interface; it's
            # ignored here (apart from a sanity check).  Also, 'module' is
            # the *unqualified* module name (ie. no dots, no package -- we
            # already know its package!), and 'module_file' is the path to
            # the .py file, relative to the current directory
            # (ie. including 'package_dir').
            package_dir = self.get_package_dir(package)
            modules = self.find_package_modules(package, package_dir)

            # Now loop over the modules we found, "building" each one (just
            # copy it to self.build_lib).
            for (package_, module, module_file) in modules:
                assert package == package_
                self.build_module(module, module_file, package)

    def byte_compile(self, files):
        if sys.dont_write_bytecode:
            self.warn('byte-compiling is disabled, skipping.')
            return

        from distutils.util import byte_compile
        prefix = self.build_lib
        if prefix[-1] != os.sep:
            prefix = prefix + os.sep

        # XXX this code is essentially the same as the 'byte_compile()
        # method of the "install_lib" command, except for the determination
        # of the 'prefix' string.  Hmmm.
        if self.compile:
            byte_compile(files, optimize=0,
                         force=self.force, prefix=prefix, dry_run=self.dry_run)
        if self.optimize > 0:
            byte_compile(files, optimize=self.optimize,
                         force=self.force, prefix=prefix, dry_run=self.dry_run)
site-packages/setuptools/_distutils/command/install_scripts.py000064400000003741150564745010021111 0ustar00"""distutils.command.install_scripts

Implements the Distutils 'install_scripts' command, for installing
Python scripts."""

# contributed by Bastian Kleineidam

import os
from distutils.core import Command
from distutils import log
from stat import ST_MODE


class install_scripts(Command):

    description = "install scripts (Python or otherwise)"

    user_options = [
        ('install-dir=', 'd', "directory to install scripts to"),
        ('build-dir=','b', "build directory (where to install from)"),
        ('force', 'f', "force installation (overwrite existing files)"),
        ('skip-build', None, "skip the build steps"),
    ]

    boolean_options = ['force', 'skip-build']

    def initialize_options(self):
        self.install_dir = None
        self.force = 0
        self.build_dir = None
        self.skip_build = None

    def finalize_options(self):
        self.set_undefined_options('build', ('build_scripts', 'build_dir'))
        self.set_undefined_options('install',
                                   ('install_scripts', 'install_dir'),
                                   ('force', 'force'),
                                   ('skip_build', 'skip_build'),
                                  )

    def run(self):
        if not self.skip_build:
            self.run_command('build_scripts')
        self.outfiles = self.copy_tree(self.build_dir, self.install_dir)
        if os.name == 'posix':
            # Set the executable bits (owner, group, and world) on
            # all the scripts we just installed.
            for file in self.get_outputs():
                if self.dry_run:
                    log.info("changing mode of %s", file)
                else:
                    mode = ((os.stat(file)[ST_MODE]) | 0o555) & 0o7777
                    log.info("changing mode of %s to %o", file, mode)
                    os.chmod(file, mode)

    def get_inputs(self):
        return self.distribution.scripts or []

    def get_outputs(self):
        return self.outfiles or []
site-packages/setuptools/_distutils/command/bdist.py000064400000012672150564745010017004 0ustar00"""distutils.command.bdist

Implements the Distutils 'bdist' command (create a built [binary]
distribution)."""

import os
from distutils.core import Command
from distutils.errors import *
from distutils.util import get_platform


def show_formats():
    """Print list of available formats (arguments to "--format" option).
    """
    from distutils.fancy_getopt import FancyGetopt
    formats = []
    for format in bdist.format_commands:
        formats.append(("formats=" + format, None,
                        bdist.format_command[format][1]))
    pretty_printer = FancyGetopt(formats)
    pretty_printer.print_help("List of available distribution formats:")


class bdist(Command):

    description = "create a built (binary) distribution"

    user_options = [('bdist-base=', 'b',
                     "temporary directory for creating built distributions"),
                    ('plat-name=', 'p',
                     "platform name to embed in generated filenames "
                     "(default: %s)" % get_platform()),
                    ('formats=', None,
                     "formats for distribution (comma-separated list)"),
                    ('dist-dir=', 'd',
                     "directory to put final built distributions in "
                     "[default: dist]"),
                    ('skip-build', None,
                     "skip rebuilding everything (for testing/debugging)"),
                    ('owner=', 'u',
                     "Owner name used when creating a tar file"
                     " [default: current user]"),
                    ('group=', 'g',
                     "Group name used when creating a tar file"
                     " [default: current group]"),
                   ]

    boolean_options = ['skip-build']

    help_options = [
        ('help-formats', None,
         "lists available distribution formats", show_formats),
        ]

    # The following commands do not take a format option from bdist
    no_format_option = ('bdist_rpm',)

    # This won't do in reality: will need to distinguish RPM-ish Linux,
    # Debian-ish Linux, Solaris, FreeBSD, ..., Windows, Mac OS.
    default_format = {'posix': 'gztar',
                      'nt': 'zip'}

    # Establish the preferred order (for the --help-formats option).
    format_commands = ['rpm', 'gztar', 'bztar', 'xztar', 'ztar', 'tar',
                       'wininst', 'zip', 'msi']

    # And the real information.
    format_command = {'rpm':   ('bdist_rpm',  "RPM distribution"),
                      'gztar': ('bdist_dumb', "gzip'ed tar file"),
                      'bztar': ('bdist_dumb', "bzip2'ed tar file"),
                      'xztar': ('bdist_dumb', "xz'ed tar file"),
                      'ztar':  ('bdist_dumb', "compressed tar file"),
                      'tar':   ('bdist_dumb', "tar file"),
                      'wininst': ('bdist_wininst',
                                  "Windows executable installer"),
                      'zip':   ('bdist_dumb', "ZIP file"),
                      'msi':   ('bdist_msi',  "Microsoft Installer")
                      }


    def initialize_options(self):
        self.bdist_base = None
        self.plat_name = None
        self.formats = None
        self.dist_dir = None
        self.skip_build = 0
        self.group = None
        self.owner = None

    def finalize_options(self):
        # have to finalize 'plat_name' before 'bdist_base'
        if self.plat_name is None:
            if self.skip_build:
                self.plat_name = get_platform()
            else:
                self.plat_name = self.get_finalized_command('build').plat_name

        # 'bdist_base' -- parent of per-built-distribution-format
        # temporary directories (eg. we'll probably have
        # "build/bdist.<plat>/dumb", "build/bdist.<plat>/rpm", etc.)
        if self.bdist_base is None:
            build_base = self.get_finalized_command('build').build_base
            self.bdist_base = os.path.join(build_base,
                                           'bdist.' + self.plat_name)

        self.ensure_string_list('formats')
        if self.formats is None:
            try:
                self.formats = [self.default_format[os.name]]
            except KeyError:
                raise DistutilsPlatformError(
                      "don't know how to create built distributions "
                      "on platform %s" % os.name)

        if self.dist_dir is None:
            self.dist_dir = "dist"

    def run(self):
        # Figure out which sub-commands we need to run.
        commands = []
        for format in self.formats:
            try:
                commands.append(self.format_command[format][0])
            except KeyError:
                raise DistutilsOptionError("invalid format '%s'" % format)

        # Reinitialize and run each command.
        for i in range(len(self.formats)):
            cmd_name = commands[i]
            sub_cmd = self.reinitialize_command(cmd_name)
            if cmd_name not in self.no_format_option:
                sub_cmd.format = self.formats[i]

            # passing the owner and group names for tar archiving
            if cmd_name == 'bdist_dumb':
                sub_cmd.owner = self.owner
                sub_cmd.group = self.group

            # If we're going to need to run this command again, tell it to
            # keep its temporary files around so subsequent runs go faster.
            if cmd_name in commands[i+1:]:
                sub_cmd.keep_temp = 1
            self.run_command(cmd_name)
site-packages/setuptools/_distutils/command/install.py000064400000065616150564745010017353 0ustar00"""distutils.command.install

Implements the Distutils 'install' command."""

import sys
import os

from distutils import log
from distutils.core import Command
from distutils.debug import DEBUG
from distutils.sysconfig import get_config_vars
from distutils.errors import DistutilsPlatformError
from distutils.file_util import write_file
from distutils.util import convert_path, subst_vars, change_root
from distutils.util import get_platform
from distutils.errors import DistutilsOptionError

from site import USER_BASE
from site import USER_SITE
HAS_USER_SITE = True

WINDOWS_SCHEME = {
    'purelib': '$base/Lib/site-packages',
    'platlib': '$base/Lib/site-packages',
    'headers': '$base/Include/$dist_name',
    'scripts': '$base/Scripts',
    'data'   : '$base',
}

INSTALL_SCHEMES = {
    'unix_prefix': {
        'purelib': '$base/lib/python$py_version_short/site-packages',
        'platlib': '$platbase/$platlibdir/python$py_version_short/site-packages',
        'headers': '$base/include/python$py_version_short$abiflags/$dist_name',
        'scripts': '$base/bin',
        'data'   : '$base',
        },
    'unix_home': {
        'purelib': '$base/lib/python',
        'platlib': '$base/$platlibdir/python',
        'headers': '$base/include/python/$dist_name',
        'scripts': '$base/bin',
        'data'   : '$base',
        },
    'nt': WINDOWS_SCHEME,
    'pypy': {
        'purelib': '$base/site-packages',
        'platlib': '$base/site-packages',
        'headers': '$base/include/$dist_name',
        'scripts': '$base/bin',
        'data'   : '$base',
        },
    'pypy_nt': {
        'purelib': '$base/site-packages',
        'platlib': '$base/site-packages',
        'headers': '$base/include/$dist_name',
        'scripts': '$base/Scripts',
        'data'   : '$base',
        },
    }

# user site schemes
if HAS_USER_SITE:
    INSTALL_SCHEMES['nt_user'] = {
        'purelib': '$usersite',
        'platlib': '$usersite',
        'headers': '$userbase/Python$py_version_nodot/Include/$dist_name',
        'scripts': '$userbase/Python$py_version_nodot/Scripts',
        'data'   : '$userbase',
        }

    INSTALL_SCHEMES['unix_user'] = {
        'purelib': '$usersite',
        'platlib': '$usersite',
        'headers':
            '$userbase/include/python$py_version_short$abiflags/$dist_name',
        'scripts': '$userbase/bin',
        'data'   : '$userbase',
        }

# The keys to an installation scheme; if any new types of files are to be
# installed, be sure to add an entry to every installation scheme above,
# and to SCHEME_KEYS here.
SCHEME_KEYS = ('purelib', 'platlib', 'headers', 'scripts', 'data')


class install(Command):

    description = "install everything from build directory"

    user_options = [
        # Select installation scheme and set base director(y|ies)
        ('prefix=', None,
         "installation prefix"),
        ('exec-prefix=', None,
         "(Unix only) prefix for platform-specific files"),
        ('home=', None,
         "(Unix only) home directory to install under"),

        # Or, just set the base director(y|ies)
        ('install-base=', None,
         "base installation directory (instead of --prefix or --home)"),
        ('install-platbase=', None,
         "base installation directory for platform-specific files " +
         "(instead of --exec-prefix or --home)"),
        ('root=', None,
         "install everything relative to this alternate root directory"),

        # Or, explicitly set the installation scheme
        ('install-purelib=', None,
         "installation directory for pure Python module distributions"),
        ('install-platlib=', None,
         "installation directory for non-pure module distributions"),
        ('install-lib=', None,
         "installation directory for all module distributions " +
         "(overrides --install-purelib and --install-platlib)"),

        ('install-headers=', None,
         "installation directory for C/C++ headers"),
        ('install-scripts=', None,
         "installation directory for Python scripts"),
        ('install-data=', None,
         "installation directory for data files"),

        # Byte-compilation options -- see install_lib.py for details, as
        # these are duplicated from there (but only install_lib does
        # anything with them).
        ('compile', 'c', "compile .py to .pyc [default]"),
        ('no-compile', None, "don't compile .py files"),
        ('optimize=', 'O',
         "also compile with optimization: -O1 for \"python -O\", "
         "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"),

        # Miscellaneous control options
        ('force', 'f',
         "force installation (overwrite any existing files)"),
        ('skip-build', None,
         "skip rebuilding everything (for testing/debugging)"),

        # Where to install documentation (eventually!)
        #('doc-format=', None, "format of documentation to generate"),
        #('install-man=', None, "directory for Unix man pages"),
        #('install-html=', None, "directory for HTML documentation"),
        #('install-info=', None, "directory for GNU info files"),

        ('record=', None,
         "filename in which to record list of installed files"),
        ]

    boolean_options = ['compile', 'force', 'skip-build']

    if HAS_USER_SITE:
        user_options.append(('user', None,
                             "install in user site-package '%s'" % USER_SITE))
        boolean_options.append('user')

    negative_opt = {'no-compile' : 'compile'}


    def initialize_options(self):
        """Initializes options."""
        # High-level options: these select both an installation base
        # and scheme.
        self.prefix = None
        self.exec_prefix = None
        self.home = None
        self.user = 0

        # These select only the installation base; it's up to the user to
        # specify the installation scheme (currently, that means supplying
        # the --install-{platlib,purelib,scripts,data} options).
        self.install_base = None
        self.install_platbase = None
        self.root = None

        # These options are the actual installation directories; if not
        # supplied by the user, they are filled in using the installation
        # scheme implied by prefix/exec-prefix/home and the contents of
        # that installation scheme.
        self.install_purelib = None     # for pure module distributions
        self.install_platlib = None     # non-pure (dists w/ extensions)
        self.install_headers = None     # for C/C++ headers
        self.install_lib = None         # set to either purelib or platlib
        self.install_scripts = None
        self.install_data = None
        self.install_userbase = USER_BASE
        self.install_usersite = USER_SITE

        self.compile = None
        self.optimize = None

        # Deprecated
        # These two are for putting non-packagized distributions into their
        # own directory and creating a .pth file if it makes sense.
        # 'extra_path' comes from the setup file; 'install_path_file' can
        # be turned off if it makes no sense to install a .pth file.  (But
        # better to install it uselessly than to guess wrong and not
        # install it when it's necessary and would be used!)  Currently,
        # 'install_path_file' is always true unless some outsider meddles
        # with it.
        self.extra_path = None
        self.install_path_file = 1

        # 'force' forces installation, even if target files are not
        # out-of-date.  'skip_build' skips running the "build" command,
        # handy if you know it's not necessary.  'warn_dir' (which is *not*
        # a user option, it's just there so the bdist_* commands can turn
        # it off) determines whether we warn about installing to a
        # directory not in sys.path.
        self.force = 0
        self.skip_build = 0
        self.warn_dir = 1

        # These are only here as a conduit from the 'build' command to the
        # 'install_*' commands that do the real work.  ('build_base' isn't
        # actually used anywhere, but it might be useful in future.)  They
        # are not user options, because if the user told the install
        # command where the build directory is, that wouldn't affect the
        # build command.
        self.build_base = None
        self.build_lib = None

        # Not defined yet because we don't know anything about
        # documentation yet.
        #self.install_man = None
        #self.install_html = None
        #self.install_info = None

        self.record = None


    # -- Option finalizing methods -------------------------------------
    # (This is rather more involved than for most commands,
    # because this is where the policy for installing third-
    # party Python modules on various platforms given a wide
    # array of user input is decided.  Yes, it's quite complex!)

    def finalize_options(self):
        """Finalizes options."""
        # This method (and its helpers, like 'finalize_unix()',
        # 'finalize_other()', and 'select_scheme()') is where the default
        # installation directories for modules, extension modules, and
        # anything else we care to install from a Python module
        # distribution.  Thus, this code makes a pretty important policy
        # statement about how third-party stuff is added to a Python
        # installation!  Note that the actual work of installation is done
        # by the relatively simple 'install_*' commands; they just take
        # their orders from the installation directory options determined
        # here.

        # Check for errors/inconsistencies in the options; first, stuff
        # that's wrong on any platform.

        if ((self.prefix or self.exec_prefix or self.home) and
            (self.install_base or self.install_platbase)):
            raise DistutilsOptionError(
                   "must supply either prefix/exec-prefix/home or " +
                   "install-base/install-platbase -- not both")

        if self.home and (self.prefix or self.exec_prefix):
            raise DistutilsOptionError(
                  "must supply either home or prefix/exec-prefix -- not both")

        if self.user and (self.prefix or self.exec_prefix or self.home or
                self.install_base or self.install_platbase):
            raise DistutilsOptionError("can't combine user with prefix, "
                                       "exec_prefix/home, or install_(plat)base")

        # Next, stuff that's wrong (or dubious) only on certain platforms.
        if os.name != "posix":
            if self.exec_prefix:
                self.warn("exec-prefix option ignored on this platform")
                self.exec_prefix = None

        # Now the interesting logic -- so interesting that we farm it out
        # to other methods.  The goal of these methods is to set the final
        # values for the install_{lib,scripts,data,...}  options, using as
        # input a heady brew of prefix, exec_prefix, home, install_base,
        # install_platbase, user-supplied versions of
        # install_{purelib,platlib,lib,scripts,data,...}, and the
        # INSTALL_SCHEME dictionary above.  Phew!

        self.dump_dirs("pre-finalize_{unix,other}")

        if os.name == 'posix':
            self.finalize_unix()
        else:
            self.finalize_other()

        self.dump_dirs("post-finalize_{unix,other}()")

        # Expand configuration variables, tilde, etc. in self.install_base
        # and self.install_platbase -- that way, we can use $base or
        # $platbase in the other installation directories and not worry
        # about needing recursive variable expansion (shudder).

        py_version = sys.version.split()[0]
        (prefix, exec_prefix) = get_config_vars('prefix', 'exec_prefix')
        try:
            abiflags = sys.abiflags
        except AttributeError:
            # sys.abiflags may not be defined on all platforms.
            abiflags = ''
        self.config_vars = {'dist_name': self.distribution.get_name(),
                            'dist_version': self.distribution.get_version(),
                            'dist_fullname': self.distribution.get_fullname(),
                            'py_version': py_version,
                            'py_version_short': '%d.%d' % sys.version_info[:2],
                            'py_version_nodot': '%d%d' % sys.version_info[:2],
                            'sys_prefix': prefix,
                            'prefix': prefix,
                            'sys_exec_prefix': exec_prefix,
                            'exec_prefix': exec_prefix,
                            'abiflags': abiflags,
                            'platlibdir': getattr(sys, 'platlibdir', 'lib'),
                           }

        if HAS_USER_SITE:
            self.config_vars['userbase'] = self.install_userbase
            self.config_vars['usersite'] = self.install_usersite

        self.expand_basedirs()

        self.dump_dirs("post-expand_basedirs()")

        # Now define config vars for the base directories so we can expand
        # everything else.
        self.config_vars['base'] = self.install_base
        self.config_vars['platbase'] = self.install_platbase

        if DEBUG:
            from pprint import pprint
            print("config vars:")
            pprint(self.config_vars)

        # Expand "~" and configuration variables in the installation
        # directories.
        self.expand_dirs()

        self.dump_dirs("post-expand_dirs()")

        # Create directories in the home dir:
        if self.user:
            self.create_home_path()

        # Pick the actual directory to install all modules to: either
        # install_purelib or install_platlib, depending on whether this
        # module distribution is pure or not.  Of course, if the user
        # already specified install_lib, use their selection.
        if self.install_lib is None:
            if self.distribution.has_ext_modules(): # has extensions: non-pure
                self.install_lib = self.install_platlib
            else:
                self.install_lib = self.install_purelib


        # Convert directories from Unix /-separated syntax to the local
        # convention.
        self.convert_paths('lib', 'purelib', 'platlib',
                           'scripts', 'data', 'headers',
                           'userbase', 'usersite')

        # Deprecated
        # Well, we're not actually fully completely finalized yet: we still
        # have to deal with 'extra_path', which is the hack for allowing
        # non-packagized module distributions (hello, Numerical Python!) to
        # get their own directories.
        self.handle_extra_path()
        self.install_libbase = self.install_lib # needed for .pth file
        self.install_lib = os.path.join(self.install_lib, self.extra_dirs)

        # If a new root directory was supplied, make all the installation
        # dirs relative to it.
        if self.root is not None:
            self.change_roots('libbase', 'lib', 'purelib', 'platlib',
                              'scripts', 'data', 'headers')

        self.dump_dirs("after prepending root")

        # Find out the build directories, ie. where to install from.
        self.set_undefined_options('build',
                                   ('build_base', 'build_base'),
                                   ('build_lib', 'build_lib'))

        # Punt on doc directories for now -- after all, we're punting on
        # documentation completely!

    def dump_dirs(self, msg):
        """Dumps the list of user options."""
        if not DEBUG:
            return
        from distutils.fancy_getopt import longopt_xlate
        log.debug(msg + ":")
        for opt in self.user_options:
            opt_name = opt[0]
            if opt_name[-1] == "=":
                opt_name = opt_name[0:-1]
            if opt_name in self.negative_opt:
                opt_name = self.negative_opt[opt_name]
                opt_name = opt_name.translate(longopt_xlate)
                val = not getattr(self, opt_name)
            else:
                opt_name = opt_name.translate(longopt_xlate)
                val = getattr(self, opt_name)
            log.debug("  %s: %s", opt_name, val)

    def finalize_unix(self):
        """Finalizes options for posix platforms."""
        if self.install_base is not None or self.install_platbase is not None:
            if ((self.install_lib is None and
                 self.install_purelib is None and
                 self.install_platlib is None) or
                self.install_headers is None or
                self.install_scripts is None or
                self.install_data is None):
                raise DistutilsOptionError(
                      "install-base or install-platbase supplied, but "
                      "installation scheme is incomplete")
            return

        if self.user:
            if self.install_userbase is None:
                raise DistutilsPlatformError(
                    "User base directory is not specified")
            self.install_base = self.install_platbase = self.install_userbase
            self.select_scheme("unix_user")
        elif self.home is not None:
            self.install_base = self.install_platbase = self.home
            self.select_scheme("unix_home")
        else:
            if self.prefix is None:
                if self.exec_prefix is not None:
                    raise DistutilsOptionError(
                          "must not supply exec-prefix without prefix")

                self.prefix = os.path.normpath(sys.prefix)
                self.exec_prefix = os.path.normpath(sys.exec_prefix)

            else:
                if self.exec_prefix is None:
                    self.exec_prefix = self.prefix

            self.install_base = self.prefix
            self.install_platbase = self.exec_prefix
            self.select_scheme("unix_prefix")

    def finalize_other(self):
        """Finalizes options for non-posix platforms"""
        if self.user:
            if self.install_userbase is None:
                raise DistutilsPlatformError(
                    "User base directory is not specified")
            self.install_base = self.install_platbase = self.install_userbase
            self.select_scheme(os.name + "_user")
        elif self.home is not None:
            self.install_base = self.install_platbase = self.home
            self.select_scheme("unix_home")
        else:
            if self.prefix is None:
                self.prefix = os.path.normpath(sys.prefix)

            self.install_base = self.install_platbase = self.prefix
            try:
                self.select_scheme(os.name)
            except KeyError:
                raise DistutilsPlatformError(
                      "I don't know how to install stuff on '%s'" % os.name)

    def select_scheme(self, name):
        """Sets the install directories by applying the install schemes."""
        # it's the caller's problem if they supply a bad name!
        if (hasattr(sys, 'pypy_version_info') and
                sys.version_info < (3, 8) and
                not name.endswith(('_user', '_home'))):
            if os.name == 'nt':
                name = 'pypy_nt'
            else:
                name = 'pypy'
        scheme = INSTALL_SCHEMES[name]
        for key in SCHEME_KEYS:
            attrname = 'install_' + key
            if getattr(self, attrname) is None:
                setattr(self, attrname, scheme[key])

    def _expand_attrs(self, attrs):
        for attr in attrs:
            val = getattr(self, attr)
            if val is not None:
                if os.name == 'posix' or os.name == 'nt':
                    val = os.path.expanduser(val)
                val = subst_vars(val, self.config_vars)
                setattr(self, attr, val)

    def expand_basedirs(self):
        """Calls `os.path.expanduser` on install_base, install_platbase and
        root."""
        self._expand_attrs(['install_base', 'install_platbase', 'root'])

    def expand_dirs(self):
        """Calls `os.path.expanduser` on install dirs."""
        self._expand_attrs(['install_purelib', 'install_platlib',
                            'install_lib', 'install_headers',
                            'install_scripts', 'install_data',])

    def convert_paths(self, *names):
        """Call `convert_path` over `names`."""
        for name in names:
            attr = "install_" + name
            setattr(self, attr, convert_path(getattr(self, attr)))

    def handle_extra_path(self):
        """Set `path_file` and `extra_dirs` using `extra_path`."""
        if self.extra_path is None:
            self.extra_path = self.distribution.extra_path

        if self.extra_path is not None:
            log.warn(
                "Distribution option extra_path is deprecated. "
                "See issue27919 for details."
            )
            if isinstance(self.extra_path, str):
                self.extra_path = self.extra_path.split(',')

            if len(self.extra_path) == 1:
                path_file = extra_dirs = self.extra_path[0]
            elif len(self.extra_path) == 2:
                path_file, extra_dirs = self.extra_path
            else:
                raise DistutilsOptionError(
                      "'extra_path' option must be a list, tuple, or "
                      "comma-separated string with 1 or 2 elements")

            # convert to local form in case Unix notation used (as it
            # should be in setup scripts)
            extra_dirs = convert_path(extra_dirs)
        else:
            path_file = None
            extra_dirs = ''

        # XXX should we warn if path_file and not extra_dirs? (in which
        # case the path file would be harmless but pointless)
        self.path_file = path_file
        self.extra_dirs = extra_dirs

    def change_roots(self, *names):
        """Change the install directories pointed by name using root."""
        for name in names:
            attr = "install_" + name
            setattr(self, attr, change_root(self.root, getattr(self, attr)))

    def create_home_path(self):
        """Create directories under ~."""
        if not self.user:
            return
        home = convert_path(os.path.expanduser("~"))
        for name, path in self.config_vars.items():
            if path.startswith(home) and not os.path.isdir(path):
                self.debug_print("os.makedirs('%s', 0o700)" % path)
                os.makedirs(path, 0o700)

    # -- Command execution methods -------------------------------------

    def run(self):
        """Runs the command."""
        # Obviously have to build before we can install
        if not self.skip_build:
            self.run_command('build')
            # If we built for any other platform, we can't install.
            build_plat = self.distribution.get_command_obj('build').plat_name
            # check warn_dir - it is a clue that the 'install' is happening
            # internally, and not to sys.path, so we don't check the platform
            # matches what we are running.
            if self.warn_dir and build_plat != get_platform():
                raise DistutilsPlatformError("Can't install when "
                                             "cross-compiling")

        # Run all sub-commands (at least those that need to be run)
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)

        if self.path_file:
            self.create_path_file()

        # write list of installed files, if requested.
        if self.record:
            outputs = self.get_outputs()
            if self.root:               # strip any package prefix
                root_len = len(self.root)
                for counter in range(len(outputs)):
                    outputs[counter] = outputs[counter][root_len:]
            self.execute(write_file,
                         (self.record, outputs),
                         "writing list of installed files to '%s'" %
                         self.record)

        sys_path = map(os.path.normpath, sys.path)
        sys_path = map(os.path.normcase, sys_path)
        install_lib = os.path.normcase(os.path.normpath(self.install_lib))
        if (self.warn_dir and
            not (self.path_file and self.install_path_file) and
            install_lib not in sys_path):
            log.debug(("modules installed to '%s', which is not in "
                       "Python's module search path (sys.path) -- "
                       "you'll have to change the search path yourself"),
                       self.install_lib)

    def create_path_file(self):
        """Creates the .pth file"""
        filename = os.path.join(self.install_libbase,
                                self.path_file + ".pth")
        if self.install_path_file:
            self.execute(write_file,
                         (filename, [self.extra_dirs]),
                         "creating %s" % filename)
        else:
            self.warn("path file '%s' not created" % filename)


    # -- Reporting methods ---------------------------------------------

    def get_outputs(self):
        """Assembles the outputs of all the sub-commands."""
        outputs = []
        for cmd_name in self.get_sub_commands():
            cmd = self.get_finalized_command(cmd_name)
            # Add the contents of cmd.get_outputs(), ensuring
            # that outputs doesn't contain duplicate entries
            for filename in cmd.get_outputs():
                if filename not in outputs:
                    outputs.append(filename)

        if self.path_file and self.install_path_file:
            outputs.append(os.path.join(self.install_libbase,
                                        self.path_file + ".pth"))

        return outputs

    def get_inputs(self):
        """Returns the inputs of all the sub-commands"""
        # XXX gee, this looks familiar ;-(
        inputs = []
        for cmd_name in self.get_sub_commands():
            cmd = self.get_finalized_command(cmd_name)
            inputs.extend(cmd.get_inputs())

        return inputs

    # -- Predicates for sub-command list -------------------------------

    def has_lib(self):
        """Returns true if the current distribution has any Python
        modules to install."""
        return (self.distribution.has_pure_modules() or
                self.distribution.has_ext_modules())

    def has_headers(self):
        """Returns true if the current distribution has any headers to
        install."""
        return self.distribution.has_headers()

    def has_scripts(self):
        """Returns true if the current distribution has any scripts to.
        install."""
        return self.distribution.has_scripts()

    def has_data(self):
        """Returns true if the current distribution has any data to.
        install."""
        return self.distribution.has_data_files()

    # 'sub_commands': a list of commands this command might have to run to
    # get its work done.  See cmd.py for more info.
    sub_commands = [('install_lib',     has_lib),
                    ('install_headers', has_headers),
                    ('install_scripts', has_scripts),
                    ('install_data',    has_data),
                    ('install_egg_info', lambda self:True),
                   ]
site-packages/setuptools/_distutils/command/build_clib.py000064400000017526150564745010017772 0ustar00"""distutils.command.build_clib

Implements the Distutils 'build_clib' command, to build a C/C++ library
that is included in the module distribution and needed by an extension
module."""


# XXX this module has *lots* of code ripped-off quite transparently from
# build_ext.py -- not surprisingly really, as the work required to build
# a static library from a collection of C source files is not really all
# that different from what's required to build a shared object file from
# a collection of C source files.  Nevertheless, I haven't done the
# necessary refactoring to account for the overlap in code between the
# two modules, mainly because a number of subtle details changed in the
# cut 'n paste.  Sigh.

import os
from distutils.core import Command
from distutils.errors import *
from distutils.sysconfig import customize_compiler
from distutils import log

def show_compilers():
    from distutils.ccompiler import show_compilers
    show_compilers()


class build_clib(Command):

    description = "build C/C++ libraries used by Python extensions"

    user_options = [
        ('build-clib=', 'b',
         "directory to build C/C++ libraries to"),
        ('build-temp=', 't',
         "directory to put temporary build by-products"),
        ('debug', 'g',
         "compile with debugging information"),
        ('force', 'f',
         "forcibly build everything (ignore file timestamps)"),
        ('compiler=', 'c',
         "specify the compiler type"),
        ]

    boolean_options = ['debug', 'force']

    help_options = [
        ('help-compiler', None,
         "list available compilers", show_compilers),
        ]

    def initialize_options(self):
        self.build_clib = None
        self.build_temp = None

        # List of libraries to build
        self.libraries = None

        # Compilation options for all libraries
        self.include_dirs = None
        self.define = None
        self.undef = None
        self.debug = None
        self.force = 0
        self.compiler = None


    def finalize_options(self):
        # This might be confusing: both build-clib and build-temp default
        # to build-temp as defined by the "build" command.  This is because
        # I think that C libraries are really just temporary build
        # by-products, at least from the point of view of building Python
        # extensions -- but I want to keep my options open.
        self.set_undefined_options('build',
                                   ('build_temp', 'build_clib'),
                                   ('build_temp', 'build_temp'),
                                   ('compiler', 'compiler'),
                                   ('debug', 'debug'),
                                   ('force', 'force'))

        self.libraries = self.distribution.libraries
        if self.libraries:
            self.check_library_list(self.libraries)

        if self.include_dirs is None:
            self.include_dirs = self.distribution.include_dirs or []
        if isinstance(self.include_dirs, str):
            self.include_dirs = self.include_dirs.split(os.pathsep)

        # XXX same as for build_ext -- what about 'self.define' and
        # 'self.undef' ?


    def run(self):
        if not self.libraries:
            return

        # Yech -- this is cut 'n pasted from build_ext.py!
        from distutils.ccompiler import new_compiler
        self.compiler = new_compiler(compiler=self.compiler,
                                     dry_run=self.dry_run,
                                     force=self.force)
        customize_compiler(self.compiler)

        if self.include_dirs is not None:
            self.compiler.set_include_dirs(self.include_dirs)
        if self.define is not None:
            # 'define' option is a list of (name,value) tuples
            for (name,value) in self.define:
                self.compiler.define_macro(name, value)
        if self.undef is not None:
            for macro in self.undef:
                self.compiler.undefine_macro(macro)

        self.build_libraries(self.libraries)


    def check_library_list(self, libraries):
        """Ensure that the list of libraries is valid.

        `library` is presumably provided as a command option 'libraries'.
        This method checks that it is a list of 2-tuples, where the tuples
        are (library_name, build_info_dict).

        Raise DistutilsSetupError if the structure is invalid anywhere;
        just returns otherwise.
        """
        if not isinstance(libraries, list):
            raise DistutilsSetupError(
                  "'libraries' option must be a list of tuples")

        for lib in libraries:
            if not isinstance(lib, tuple) and len(lib) != 2:
                raise DistutilsSetupError(
                      "each element of 'libraries' must a 2-tuple")

            name, build_info = lib

            if not isinstance(name, str):
                raise DistutilsSetupError(
                      "first element of each tuple in 'libraries' "
                      "must be a string (the library name)")

            if '/' in name or (os.sep != '/' and os.sep in name):
                raise DistutilsSetupError("bad library name '%s': "
                       "may not contain directory separators" % lib[0])

            if not isinstance(build_info, dict):
                raise DistutilsSetupError(
                      "second element of each tuple in 'libraries' "
                      "must be a dictionary (build info)")


    def get_library_names(self):
        # Assume the library list is valid -- 'check_library_list()' is
        # called from 'finalize_options()', so it should be!
        if not self.libraries:
            return None

        lib_names = []
        for (lib_name, build_info) in self.libraries:
            lib_names.append(lib_name)
        return lib_names


    def get_source_files(self):
        self.check_library_list(self.libraries)
        filenames = []
        for (lib_name, build_info) in self.libraries:
            sources = build_info.get('sources')
            if sources is None or not isinstance(sources, (list, tuple)):
                raise DistutilsSetupError(
                       "in 'libraries' option (library '%s'), "
                       "'sources' must be present and must be "
                       "a list of source filenames" % lib_name)

            filenames.extend(sources)
        return filenames


    def build_libraries(self, libraries):
        for (lib_name, build_info) in libraries:
            sources = build_info.get('sources')
            if sources is None or not isinstance(sources, (list, tuple)):
                raise DistutilsSetupError(
                       "in 'libraries' option (library '%s'), "
                       "'sources' must be present and must be "
                       "a list of source filenames" % lib_name)
            sources = list(sources)

            log.info("building '%s' library", lib_name)

            # First, compile the source code to object files in the library
            # directory.  (This should probably change to putting object
            # files in a temporary build directory.)
            macros = build_info.get('macros')
            include_dirs = build_info.get('include_dirs')
            objects = self.compiler.compile(sources,
                                            output_dir=self.build_temp,
                                            macros=macros,
                                            include_dirs=include_dirs,
                                            debug=self.debug)

            # Now "link" the object files together into a static library.
            # (On Unix at least, this isn't really linking -- it just
            # builds an archive.  Whatever.)
            self.compiler.create_static_lib(objects, lib_name,
                                            output_dir=self.build_clib,
                                            debug=self.debug)
site-packages/setuptools/_distutils/command/py37compat.py000064400000001237150564745010017700 0ustar00import sys


def _pythonlib_compat():
    """
    On Python 3.7 and earlier, distutils would include the Python
    library. See pypa/distutils#9.
    """
    from distutils import sysconfig
    if not sysconfig.get_config_var('Py_ENABLED_SHARED'):
        return

    yield 'python{}.{}{}'.format(
        sys.hexversion >> 24,
        (sys.hexversion >> 16) & 0xff,
        sysconfig.get_config_var('ABIFLAGS'),
    )


def compose(f1, f2):
    return lambda *args, **kwargs: f1(f2(*args, **kwargs))


pythonlib = (
    compose(list, _pythonlib_compat)
    if sys.version_info < (3, 8)
    and sys.platform != 'darwin'
    and sys.platform[:3] != 'aix'
    else list
)
site-packages/setuptools/_distutils/command/build.py000064400000013215150564745010016770 0ustar00"""distutils.command.build

Implements the Distutils 'build' command."""

import sys, os
from distutils.core import Command
from distutils.errors import DistutilsOptionError
from distutils.util import get_platform


def show_compilers():
    from distutils.ccompiler import show_compilers
    show_compilers()


class build(Command):

    description = "build everything needed to install"

    user_options = [
        ('build-base=', 'b',
         "base directory for build library"),
        ('build-purelib=', None,
         "build directory for platform-neutral distributions"),
        ('build-platlib=', None,
         "build directory for platform-specific distributions"),
        ('build-lib=', None,
         "build directory for all distribution (defaults to either " +
         "build-purelib or build-platlib"),
        ('build-scripts=', None,
         "build directory for scripts"),
        ('build-temp=', 't',
         "temporary build directory"),
        ('plat-name=', 'p',
         "platform name to build for, if supported "
         "(default: %s)" % get_platform()),
        ('compiler=', 'c',
         "specify the compiler type"),
        ('parallel=', 'j',
         "number of parallel build jobs"),
        ('debug', 'g',
         "compile extensions and libraries with debugging information"),
        ('force', 'f',
         "forcibly build everything (ignore file timestamps)"),
        ('executable=', 'e',
         "specify final destination interpreter path (build.py)"),
        ]

    boolean_options = ['debug', 'force']

    help_options = [
        ('help-compiler', None,
         "list available compilers", show_compilers),
        ]

    def initialize_options(self):
        self.build_base = 'build'
        # these are decided only after 'build_base' has its final value
        # (unless overridden by the user or client)
        self.build_purelib = None
        self.build_platlib = None
        self.build_lib = None
        self.build_temp = None
        self.build_scripts = None
        self.compiler = None
        self.plat_name = None
        self.debug = None
        self.force = 0
        self.executable = None
        self.parallel = None

    def finalize_options(self):
        if self.plat_name is None:
            self.plat_name = get_platform()
        else:
            # plat-name only supported for windows (other platforms are
            # supported via ./configure flags, if at all).  Avoid misleading
            # other platforms.
            if os.name != 'nt':
                raise DistutilsOptionError(
                            "--plat-name only supported on Windows (try "
                            "using './configure --help' on your platform)")

        plat_specifier = ".%s-%d.%d" % (self.plat_name, *sys.version_info[:2])

        # Make it so Python 2.x and Python 2.x with --with-pydebug don't
        # share the same build directories. Doing so confuses the build
        # process for C modules
        if hasattr(sys, 'gettotalrefcount'):
            plat_specifier += '-pydebug'

        # 'build_purelib' and 'build_platlib' just default to 'lib' and
        # 'lib.<plat>' under the base build directory.  We only use one of
        # them for a given distribution, though --
        if self.build_purelib is None:
            self.build_purelib = os.path.join(self.build_base, 'lib')
        if self.build_platlib is None:
            self.build_platlib = os.path.join(self.build_base,
                                              'lib' + plat_specifier)

        # 'build_lib' is the actual directory that we will use for this
        # particular module distribution -- if user didn't supply it, pick
        # one of 'build_purelib' or 'build_platlib'.
        if self.build_lib is None:
            if self.distribution.has_ext_modules():
                self.build_lib = self.build_platlib
            else:
                self.build_lib = self.build_purelib

        # 'build_temp' -- temporary directory for compiler turds,
        # "build/temp.<plat>"
        if self.build_temp is None:
            self.build_temp = os.path.join(self.build_base,
                                           'temp' + plat_specifier)
        if self.build_scripts is None:
            self.build_scripts = os.path.join(self.build_base,
                                              'scripts-%d.%d' % sys.version_info[:2])

        if self.executable is None and sys.executable:
            self.executable = os.path.normpath(sys.executable)

        if isinstance(self.parallel, str):
            try:
                self.parallel = int(self.parallel)
            except ValueError:
                raise DistutilsOptionError("parallel should be an integer")

    def run(self):
        # Run all relevant sub-commands.  This will be some subset of:
        #  - build_py      - pure Python modules
        #  - build_clib    - standalone C libraries
        #  - build_ext     - Python extensions
        #  - build_scripts - (Python) scripts
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)


    # -- Predicates for the sub-command list ---------------------------

    def has_pure_modules(self):
        return self.distribution.has_pure_modules()

    def has_c_libraries(self):
        return self.distribution.has_c_libraries()

    def has_ext_modules(self):
        return self.distribution.has_ext_modules()

    def has_scripts(self):
        return self.distribution.has_scripts()


    sub_commands = [('build_py',      has_pure_modules),
                    ('build_clib',    has_c_libraries),
                    ('build_ext',     has_ext_modules),
                    ('build_scripts', has_scripts),
                   ]
site-packages/setuptools/_distutils/command/__init__.py000064400000001437150564745010017433 0ustar00"""distutils.command

Package containing implementation of all the standard Distutils
commands."""

__all__ = ['build',
           'build_py',
           'build_ext',
           'build_clib',
           'build_scripts',
           'clean',
           'install',
           'install_lib',
           'install_headers',
           'install_scripts',
           'install_data',
           'sdist',
           'register',
           'bdist',
           'bdist_dumb',
           'bdist_rpm',
           'bdist_wininst',
           'check',
           'upload',
           # These two are reserved for future use:
           #'bdist_sdux',
           #'bdist_pkgtool',
           # Note:
           # bdist_packager is not included because it only provides
           # an abstract base class
          ]
site-packages/setuptools/_distutils/command/config.py000064400000031475150564745010017146 0ustar00"""distutils.command.config

Implements the Distutils 'config' command, a (mostly) empty command class
that exists mainly to be sub-classed by specific module distributions and
applications.  The idea is that while every "config" command is different,
at least they're all named the same, and users always see "config" in the
list of standard commands.  Also, this is a good place to put common
configure-like tasks: "try to compile this C code", or "figure out where
this header file lives".
"""

import os, re

from distutils.core import Command
from distutils.errors import DistutilsExecError
from distutils.sysconfig import customize_compiler
from distutils import log

LANG_EXT = {"c": ".c", "c++": ".cxx"}

class config(Command):

    description = "prepare to build"

    user_options = [
        ('compiler=', None,
         "specify the compiler type"),
        ('cc=', None,
         "specify the compiler executable"),
        ('include-dirs=', 'I',
         "list of directories to search for header files"),
        ('define=', 'D',
         "C preprocessor macros to define"),
        ('undef=', 'U',
         "C preprocessor macros to undefine"),
        ('libraries=', 'l',
         "external C libraries to link with"),
        ('library-dirs=', 'L',
         "directories to search for external C libraries"),

        ('noisy', None,
         "show every action (compile, link, run, ...) taken"),
        ('dump-source', None,
         "dump generated source files before attempting to compile them"),
        ]


    # The three standard command methods: since the "config" command
    # does nothing by default, these are empty.

    def initialize_options(self):
        self.compiler = None
        self.cc = None
        self.include_dirs = None
        self.libraries = None
        self.library_dirs = None

        # maximal output for now
        self.noisy = 1
        self.dump_source = 1

        # list of temporary files generated along-the-way that we have
        # to clean at some point
        self.temp_files = []

    def finalize_options(self):
        if self.include_dirs is None:
            self.include_dirs = self.distribution.include_dirs or []
        elif isinstance(self.include_dirs, str):
            self.include_dirs = self.include_dirs.split(os.pathsep)

        if self.libraries is None:
            self.libraries = []
        elif isinstance(self.libraries, str):
            self.libraries = [self.libraries]

        if self.library_dirs is None:
            self.library_dirs = []
        elif isinstance(self.library_dirs, str):
            self.library_dirs = self.library_dirs.split(os.pathsep)

    def run(self):
        pass

    # Utility methods for actual "config" commands.  The interfaces are
    # loosely based on Autoconf macros of similar names.  Sub-classes
    # may use these freely.

    def _check_compiler(self):
        """Check that 'self.compiler' really is a CCompiler object;
        if not, make it one.
        """
        # We do this late, and only on-demand, because this is an expensive
        # import.
        from distutils.ccompiler import CCompiler, new_compiler
        if not isinstance(self.compiler, CCompiler):
            self.compiler = new_compiler(compiler=self.compiler,
                                         dry_run=self.dry_run, force=1)
            customize_compiler(self.compiler)
            if self.include_dirs:
                self.compiler.set_include_dirs(self.include_dirs)
            if self.libraries:
                self.compiler.set_libraries(self.libraries)
            if self.library_dirs:
                self.compiler.set_library_dirs(self.library_dirs)

    def _gen_temp_sourcefile(self, body, headers, lang):
        filename = "_configtest" + LANG_EXT[lang]
        with open(filename, "w") as file:
            if headers:
                for header in headers:
                    file.write("#include <%s>\n" % header)
                file.write("\n")
            file.write(body)
            if body[-1] != "\n":
                file.write("\n")
        return filename

    def _preprocess(self, body, headers, include_dirs, lang):
        src = self._gen_temp_sourcefile(body, headers, lang)
        out = "_configtest.i"
        self.temp_files.extend([src, out])
        self.compiler.preprocess(src, out, include_dirs=include_dirs)
        return (src, out)

    def _compile(self, body, headers, include_dirs, lang):
        src = self._gen_temp_sourcefile(body, headers, lang)
        if self.dump_source:
            dump_file(src, "compiling '%s':" % src)
        (obj,) = self.compiler.object_filenames([src])
        self.temp_files.extend([src, obj])
        self.compiler.compile([src], include_dirs=include_dirs)
        return (src, obj)

    def _link(self, body, headers, include_dirs, libraries, library_dirs,
              lang):
        (src, obj) = self._compile(body, headers, include_dirs, lang)
        prog = os.path.splitext(os.path.basename(src))[0]
        self.compiler.link_executable([obj], prog,
                                      libraries=libraries,
                                      library_dirs=library_dirs,
                                      target_lang=lang)

        if self.compiler.exe_extension is not None:
            prog = prog + self.compiler.exe_extension
        self.temp_files.append(prog)

        return (src, obj, prog)

    def _clean(self, *filenames):
        if not filenames:
            filenames = self.temp_files
            self.temp_files = []
        log.info("removing: %s", ' '.join(filenames))
        for filename in filenames:
            try:
                os.remove(filename)
            except OSError:
                pass


    # XXX these ignore the dry-run flag: what to do, what to do? even if
    # you want a dry-run build, you still need some sort of configuration
    # info.  My inclination is to make it up to the real config command to
    # consult 'dry_run', and assume a default (minimal) configuration if
    # true.  The problem with trying to do it here is that you'd have to
    # return either true or false from all the 'try' methods, neither of
    # which is correct.

    # XXX need access to the header search path and maybe default macros.

    def try_cpp(self, body=None, headers=None, include_dirs=None, lang="c"):
        """Construct a source file from 'body' (a string containing lines
        of C/C++ code) and 'headers' (a list of header files to include)
        and run it through the preprocessor.  Return true if the
        preprocessor succeeded, false if there were any errors.
        ('body' probably isn't of much use, but what the heck.)
        """
        from distutils.ccompiler import CompileError
        self._check_compiler()
        ok = True
        try:
            self._preprocess(body, headers, include_dirs, lang)
        except CompileError:
            ok = False

        self._clean()
        return ok

    def search_cpp(self, pattern, body=None, headers=None, include_dirs=None,
                   lang="c"):
        """Construct a source file (just like 'try_cpp()'), run it through
        the preprocessor, and return true if any line of the output matches
        'pattern'.  'pattern' should either be a compiled regex object or a
        string containing a regex.  If both 'body' and 'headers' are None,
        preprocesses an empty file -- which can be useful to determine the
        symbols the preprocessor and compiler set by default.
        """
        self._check_compiler()
        src, out = self._preprocess(body, headers, include_dirs, lang)

        if isinstance(pattern, str):
            pattern = re.compile(pattern)

        with open(out) as file:
            match = False
            while True:
                line = file.readline()
                if line == '':
                    break
                if pattern.search(line):
                    match = True
                    break

        self._clean()
        return match

    def try_compile(self, body, headers=None, include_dirs=None, lang="c"):
        """Try to compile a source file built from 'body' and 'headers'.
        Return true on success, false otherwise.
        """
        from distutils.ccompiler import CompileError
        self._check_compiler()
        try:
            self._compile(body, headers, include_dirs, lang)
            ok = True
        except CompileError:
            ok = False

        log.info(ok and "success!" or "failure.")
        self._clean()
        return ok

    def try_link(self, body, headers=None, include_dirs=None, libraries=None,
                 library_dirs=None, lang="c"):
        """Try to compile and link a source file, built from 'body' and
        'headers', to executable form.  Return true on success, false
        otherwise.
        """
        from distutils.ccompiler import CompileError, LinkError
        self._check_compiler()
        try:
            self._link(body, headers, include_dirs,
                       libraries, library_dirs, lang)
            ok = True
        except (CompileError, LinkError):
            ok = False

        log.info(ok and "success!" or "failure.")
        self._clean()
        return ok

    def try_run(self, body, headers=None, include_dirs=None, libraries=None,
                library_dirs=None, lang="c"):
        """Try to compile, link to an executable, and run a program
        built from 'body' and 'headers'.  Return true on success, false
        otherwise.
        """
        from distutils.ccompiler import CompileError, LinkError
        self._check_compiler()
        try:
            src, obj, exe = self._link(body, headers, include_dirs,
                                       libraries, library_dirs, lang)
            self.spawn([exe])
            ok = True
        except (CompileError, LinkError, DistutilsExecError):
            ok = False

        log.info(ok and "success!" or "failure.")
        self._clean()
        return ok


    # -- High-level methods --------------------------------------------
    # (these are the ones that are actually likely to be useful
    # when implementing a real-world config command!)

    def check_func(self, func, headers=None, include_dirs=None,
                   libraries=None, library_dirs=None, decl=0, call=0):
        """Determine if function 'func' is available by constructing a
        source file that refers to 'func', and compiles and links it.
        If everything succeeds, returns true; otherwise returns false.

        The constructed source file starts out by including the header
        files listed in 'headers'.  If 'decl' is true, it then declares
        'func' (as "int func()"); you probably shouldn't supply 'headers'
        and set 'decl' true in the same call, or you might get errors about
        a conflicting declarations for 'func'.  Finally, the constructed
        'main()' function either references 'func' or (if 'call' is true)
        calls it.  'libraries' and 'library_dirs' are used when
        linking.
        """
        self._check_compiler()
        body = []
        if decl:
            body.append("int %s ();" % func)
        body.append("int main () {")
        if call:
            body.append("  %s();" % func)
        else:
            body.append("  %s;" % func)
        body.append("}")
        body = "\n".join(body) + "\n"

        return self.try_link(body, headers, include_dirs,
                             libraries, library_dirs)

    def check_lib(self, library, library_dirs=None, headers=None,
                  include_dirs=None, other_libraries=[]):
        """Determine if 'library' is available to be linked against,
        without actually checking that any particular symbols are provided
        by it.  'headers' will be used in constructing the source file to
        be compiled, but the only effect of this is to check if all the
        header files listed are available.  Any libraries listed in
        'other_libraries' will be included in the link, in case 'library'
        has symbols that depend on other libraries.
        """
        self._check_compiler()
        return self.try_link("int main (void) { }", headers, include_dirs,
                             [library] + other_libraries, library_dirs)

    def check_header(self, header, include_dirs=None, library_dirs=None,
                     lang="c"):
        """Determine if the system header file named by 'header_file'
        exists and can be found by the preprocessor; return true if so,
        false otherwise.
        """
        return self.try_cpp(body="/* No body */", headers=[header],
                            include_dirs=include_dirs)

def dump_file(filename, head=None):
    """Dumps a file content into log.info.

    If head is not None, will be dumped before the file content.
    """
    if head is None:
        log.info('%s', filename)
    else:
        log.info(head)
    file = open(filename)
    try:
        log.info(file.read())
    finally:
        file.close()
site-packages/setuptools/_distutils/command/build_scripts.py000064400000013513150564745010020540 0ustar00"""distutils.command.build_scripts

Implements the Distutils 'build_scripts' command."""

import os, re
from stat import ST_MODE
from distutils import sysconfig
from distutils.core import Command
from distutils.dep_util import newer
from distutils.util import convert_path
from distutils import log
import tokenize

# check if Python is called on the first line with this expression
first_line_re = re.compile(b'^#!.*python[0-9.]*([ \t].*)?$')

class build_scripts(Command):

    description = "\"build\" scripts (copy and fixup #! line)"

    user_options = [
        ('build-dir=', 'd', "directory to \"build\" (copy) to"),
        ('force', 'f', "forcibly build everything (ignore file timestamps"),
        ('executable=', 'e', "specify final destination interpreter path"),
        ]

    boolean_options = ['force']


    def initialize_options(self):
        self.build_dir = None
        self.scripts = None
        self.force = None
        self.executable = None
        self.outfiles = None

    def finalize_options(self):
        self.set_undefined_options('build',
                                   ('build_scripts', 'build_dir'),
                                   ('force', 'force'),
                                   ('executable', 'executable'))
        self.scripts = self.distribution.scripts

    def get_source_files(self):
        return self.scripts

    def run(self):
        if not self.scripts:
            return
        self.copy_scripts()


    def copy_scripts(self):
        r"""Copy each script listed in 'self.scripts'; if it's marked as a
        Python script in the Unix way (first line matches 'first_line_re',
        ie. starts with "\#!" and contains "python"), then adjust the first
        line to refer to the current Python interpreter as we copy.
        """
        self.mkpath(self.build_dir)
        outfiles = []
        updated_files = []
        for script in self.scripts:
            adjust = False
            script = convert_path(script)
            outfile = os.path.join(self.build_dir, os.path.basename(script))
            outfiles.append(outfile)

            if not self.force and not newer(script, outfile):
                log.debug("not copying %s (up-to-date)", script)
                continue

            # Always open the file, but ignore failures in dry-run mode --
            # that way, we'll get accurate feedback if we can read the
            # script.
            try:
                f = open(script, "rb")
            except OSError:
                if not self.dry_run:
                    raise
                f = None
            else:
                encoding, lines = tokenize.detect_encoding(f.readline)
                f.seek(0)
                first_line = f.readline()
                if not first_line:
                    self.warn("%s is an empty file (skipping)" % script)
                    continue

                match = first_line_re.match(first_line)
                if match:
                    adjust = True
                    post_interp = match.group(1) or b''

            if adjust:
                log.info("copying and adjusting %s -> %s", script,
                         self.build_dir)
                updated_files.append(outfile)
                if not self.dry_run:
                    if not sysconfig.python_build:
                        executable = self.executable
                    else:
                        executable = os.path.join(
                            sysconfig.get_config_var("BINDIR"),
                           "python%s%s" % (sysconfig.get_config_var("VERSION"),
                                           sysconfig.get_config_var("EXE")))
                    executable = os.fsencode(executable)
                    shebang = b"#!" + executable + post_interp + b"\n"
                    # Python parser starts to read a script using UTF-8 until
                    # it gets a #coding:xxx cookie. The shebang has to be the
                    # first line of a file, the #coding:xxx cookie cannot be
                    # written before. So the shebang has to be decodable from
                    # UTF-8.
                    try:
                        shebang.decode('utf-8')
                    except UnicodeDecodeError:
                        raise ValueError(
                            "The shebang ({!r}) is not decodable "
                            "from utf-8".format(shebang))
                    # If the script is encoded to a custom encoding (use a
                    # #coding:xxx cookie), the shebang has to be decodable from
                    # the script encoding too.
                    try:
                        shebang.decode(encoding)
                    except UnicodeDecodeError:
                        raise ValueError(
                            "The shebang ({!r}) is not decodable "
                            "from the script encoding ({})"
                            .format(shebang, encoding))
                    with open(outfile, "wb") as outf:
                        outf.write(shebang)
                        outf.writelines(f.readlines())
                if f:
                    f.close()
            else:
                if f:
                    f.close()
                updated_files.append(outfile)
                self.copy_file(script, outfile)

        if os.name == 'posix':
            for file in outfiles:
                if self.dry_run:
                    log.info("changing mode of %s", file)
                else:
                    oldmode = os.stat(file)[ST_MODE] & 0o7777
                    newmode = (oldmode | 0o555) & 0o7777
                    if newmode != oldmode:
                        log.info("changing mode of %s from %o to %o",
                                 file, oldmode, newmode)
                        os.chmod(file, newmode)
        # XXX should we modify self.outfiles?
        return outfiles, updated_files
site-packages/setuptools/_distutils/command/install_headers.py000064400000002422150564745010021030 0ustar00"""distutils.command.install_headers

Implements the Distutils 'install_headers' command, to install C/C++ header
files to the Python include directory."""

from distutils.core import Command


# XXX force is never used
class install_headers(Command):

    description = "install C/C++ header files"

    user_options = [('install-dir=', 'd',
                     "directory to install header files to"),
                    ('force', 'f',
                     "force installation (overwrite existing files)"),
                   ]

    boolean_options = ['force']

    def initialize_options(self):
        self.install_dir = None
        self.force = 0
        self.outfiles = []

    def finalize_options(self):
        self.set_undefined_options('install',
                                   ('install_headers', 'install_dir'),
                                   ('force', 'force'))


    def run(self):
        headers = self.distribution.headers
        if not headers:
            return

        self.mkpath(self.install_dir)
        for header in headers:
            (out, _) = self.copy_file(header, self.install_dir)
            self.outfiles.append(out)

    def get_inputs(self):
        return self.distribution.headers or []

    def get_outputs(self):
        return self.outfiles
site-packages/setuptools/_distutils/command/bdist_rpm.py000064400000052041150564745010017654 0ustar00"""distutils.command.bdist_rpm

Implements the Distutils 'bdist_rpm' command (create RPM source and binary
distributions)."""

import subprocess, sys, os
from distutils.core import Command
from distutils.debug import DEBUG
from distutils.file_util import write_file
from distutils.errors import *
from distutils.sysconfig import get_python_version
from distutils import log

class bdist_rpm(Command):

    description = "create an RPM distribution"

    user_options = [
        ('bdist-base=', None,
         "base directory for creating built distributions"),
        ('rpm-base=', None,
         "base directory for creating RPMs (defaults to \"rpm\" under "
         "--bdist-base; must be specified for RPM 2)"),
        ('dist-dir=', 'd',
         "directory to put final RPM files in "
         "(and .spec files if --spec-only)"),
        ('python=', None,
         "path to Python interpreter to hard-code in the .spec file "
         "(default: \"python\")"),
        ('fix-python', None,
         "hard-code the exact path to the current Python interpreter in "
         "the .spec file"),
        ('spec-only', None,
         "only regenerate spec file"),
        ('source-only', None,
         "only generate source RPM"),
        ('binary-only', None,
         "only generate binary RPM"),
        ('use-bzip2', None,
         "use bzip2 instead of gzip to create source distribution"),

        # More meta-data: too RPM-specific to put in the setup script,
        # but needs to go in the .spec file -- so we make these options
        # to "bdist_rpm".  The idea is that packagers would put this
        # info in setup.cfg, although they are of course free to
        # supply it on the command line.
        ('distribution-name=', None,
         "name of the (Linux) distribution to which this "
         "RPM applies (*not* the name of the module distribution!)"),
        ('group=', None,
         "package classification [default: \"Development/Libraries\"]"),
        ('release=', None,
         "RPM release number"),
        ('serial=', None,
         "RPM serial number"),
        ('vendor=', None,
         "RPM \"vendor\" (eg. \"Joe Blow <joe@example.com>\") "
         "[default: maintainer or author from setup script]"),
        ('packager=', None,
         "RPM packager (eg. \"Jane Doe <jane@example.net>\") "
         "[default: vendor]"),
        ('doc-files=', None,
         "list of documentation files (space or comma-separated)"),
        ('changelog=', None,
         "RPM changelog"),
        ('icon=', None,
         "name of icon file"),
        ('provides=', None,
         "capabilities provided by this package"),
        ('requires=', None,
         "capabilities required by this package"),
        ('conflicts=', None,
         "capabilities which conflict with this package"),
        ('build-requires=', None,
         "capabilities required to build this package"),
        ('obsoletes=', None,
         "capabilities made obsolete by this package"),
        ('no-autoreq', None,
         "do not automatically calculate dependencies"),

        # Actions to take when building RPM
        ('keep-temp', 'k',
         "don't clean up RPM build directory"),
        ('no-keep-temp', None,
         "clean up RPM build directory [default]"),
        ('use-rpm-opt-flags', None,
         "compile with RPM_OPT_FLAGS when building from source RPM"),
        ('no-rpm-opt-flags', None,
         "do not pass any RPM CFLAGS to compiler"),
        ('rpm3-mode', None,
         "RPM 3 compatibility mode (default)"),
        ('rpm2-mode', None,
         "RPM 2 compatibility mode"),

        # Add the hooks necessary for specifying custom scripts
        ('prep-script=', None,
         "Specify a script for the PREP phase of RPM building"),
        ('build-script=', None,
         "Specify a script for the BUILD phase of RPM building"),

        ('pre-install=', None,
         "Specify a script for the pre-INSTALL phase of RPM building"),
        ('install-script=', None,
         "Specify a script for the INSTALL phase of RPM building"),
        ('post-install=', None,
         "Specify a script for the post-INSTALL phase of RPM building"),

        ('pre-uninstall=', None,
         "Specify a script for the pre-UNINSTALL phase of RPM building"),
        ('post-uninstall=', None,
         "Specify a script for the post-UNINSTALL phase of RPM building"),

        ('clean-script=', None,
         "Specify a script for the CLEAN phase of RPM building"),

        ('verify-script=', None,
         "Specify a script for the VERIFY phase of the RPM build"),

        # Allow a packager to explicitly force an architecture
        ('force-arch=', None,
         "Force an architecture onto the RPM build process"),

        ('quiet', 'q',
         "Run the INSTALL phase of RPM building in quiet mode"),
        ]

    boolean_options = ['keep-temp', 'use-rpm-opt-flags', 'rpm3-mode',
                       'no-autoreq', 'quiet']

    negative_opt = {'no-keep-temp': 'keep-temp',
                    'no-rpm-opt-flags': 'use-rpm-opt-flags',
                    'rpm2-mode': 'rpm3-mode'}


    def initialize_options(self):
        self.bdist_base = None
        self.rpm_base = None
        self.dist_dir = None
        self.python = None
        self.fix_python = None
        self.spec_only = None
        self.binary_only = None
        self.source_only = None
        self.use_bzip2 = None

        self.distribution_name = None
        self.group = None
        self.release = None
        self.serial = None
        self.vendor = None
        self.packager = None
        self.doc_files = None
        self.changelog = None
        self.icon = None

        self.prep_script = None
        self.build_script = None
        self.install_script = None
        self.clean_script = None
        self.verify_script = None
        self.pre_install = None
        self.post_install = None
        self.pre_uninstall = None
        self.post_uninstall = None
        self.prep = None
        self.provides = None
        self.requires = None
        self.conflicts = None
        self.build_requires = None
        self.obsoletes = None

        self.keep_temp = 0
        self.use_rpm_opt_flags = 1
        self.rpm3_mode = 1
        self.no_autoreq = 0

        self.force_arch = None
        self.quiet = 0

    def finalize_options(self):
        self.set_undefined_options('bdist', ('bdist_base', 'bdist_base'))
        if self.rpm_base is None:
            if not self.rpm3_mode:
                raise DistutilsOptionError(
                      "you must specify --rpm-base in RPM 2 mode")
            self.rpm_base = os.path.join(self.bdist_base, "rpm")

        if self.python is None:
            if self.fix_python:
                self.python = sys.executable
            else:
                self.python = "python3"
        elif self.fix_python:
            raise DistutilsOptionError(
                  "--python and --fix-python are mutually exclusive options")

        if os.name != 'posix':
            raise DistutilsPlatformError("don't know how to create RPM "
                   "distributions on platform %s" % os.name)
        if self.binary_only and self.source_only:
            raise DistutilsOptionError(
                  "cannot supply both '--source-only' and '--binary-only'")

        # don't pass CFLAGS to pure python distributions
        if not self.distribution.has_ext_modules():
            self.use_rpm_opt_flags = 0

        self.set_undefined_options('bdist', ('dist_dir', 'dist_dir'))
        self.finalize_package_data()

    def finalize_package_data(self):
        self.ensure_string('group', "Development/Libraries")
        self.ensure_string('vendor',
                           "%s <%s>" % (self.distribution.get_contact(),
                                        self.distribution.get_contact_email()))
        self.ensure_string('packager')
        self.ensure_string_list('doc_files')
        if isinstance(self.doc_files, list):
            for readme in ('README', 'README.txt'):
                if os.path.exists(readme) and readme not in self.doc_files:
                    self.doc_files.append(readme)

        self.ensure_string('release', "1")
        self.ensure_string('serial')   # should it be an int?

        self.ensure_string('distribution_name')

        self.ensure_string('changelog')
          # Format changelog correctly
        self.changelog = self._format_changelog(self.changelog)

        self.ensure_filename('icon')

        self.ensure_filename('prep_script')
        self.ensure_filename('build_script')
        self.ensure_filename('install_script')
        self.ensure_filename('clean_script')
        self.ensure_filename('verify_script')
        self.ensure_filename('pre_install')
        self.ensure_filename('post_install')
        self.ensure_filename('pre_uninstall')
        self.ensure_filename('post_uninstall')

        # XXX don't forget we punted on summaries and descriptions -- they
        # should be handled here eventually!

        # Now *this* is some meta-data that belongs in the setup script...
        self.ensure_string_list('provides')
        self.ensure_string_list('requires')
        self.ensure_string_list('conflicts')
        self.ensure_string_list('build_requires')
        self.ensure_string_list('obsoletes')

        self.ensure_string('force_arch')

    def run(self):
        if DEBUG:
            print("before _get_package_data():")
            print("vendor =", self.vendor)
            print("packager =", self.packager)
            print("doc_files =", self.doc_files)
            print("changelog =", self.changelog)

        # make directories
        if self.spec_only:
            spec_dir = self.dist_dir
            self.mkpath(spec_dir)
        else:
            rpm_dir = {}
            for d in ('SOURCES', 'SPECS', 'BUILD', 'RPMS', 'SRPMS'):
                rpm_dir[d] = os.path.join(self.rpm_base, d)
                self.mkpath(rpm_dir[d])
            spec_dir = rpm_dir['SPECS']

        # Spec file goes into 'dist_dir' if '--spec-only specified',
        # build/rpm.<plat> otherwise.
        spec_path = os.path.join(spec_dir,
                                 "%s.spec" % self.distribution.get_name())
        self.execute(write_file,
                     (spec_path,
                      self._make_spec_file()),
                     "writing '%s'" % spec_path)

        if self.spec_only: # stop if requested
            return

        # Make a source distribution and copy to SOURCES directory with
        # optional icon.
        saved_dist_files = self.distribution.dist_files[:]
        sdist = self.reinitialize_command('sdist')
        if self.use_bzip2:
            sdist.formats = ['bztar']
        else:
            sdist.formats = ['gztar']
        self.run_command('sdist')
        self.distribution.dist_files = saved_dist_files

        source = sdist.get_archive_files()[0]
        source_dir = rpm_dir['SOURCES']
        self.copy_file(source, source_dir)

        if self.icon:
            if os.path.exists(self.icon):
                self.copy_file(self.icon, source_dir)
            else:
                raise DistutilsFileError(
                      "icon file '%s' does not exist" % self.icon)

        # build package
        log.info("building RPMs")
        rpm_cmd = ['rpmbuild']

        if self.source_only: # what kind of RPMs?
            rpm_cmd.append('-bs')
        elif self.binary_only:
            rpm_cmd.append('-bb')
        else:
            rpm_cmd.append('-ba')
        rpm_cmd.extend(['--define', '__python %s' % self.python])
        if self.rpm3_mode:
            rpm_cmd.extend(['--define',
                             '_topdir %s' % os.path.abspath(self.rpm_base)])
        if not self.keep_temp:
            rpm_cmd.append('--clean')

        if self.quiet:
            rpm_cmd.append('--quiet')

        rpm_cmd.append(spec_path)
        # Determine the binary rpm names that should be built out of this spec
        # file
        # Note that some of these may not be really built (if the file
        # list is empty)
        nvr_string = "%{name}-%{version}-%{release}"
        src_rpm = nvr_string + ".src.rpm"
        non_src_rpm = "%{arch}/" + nvr_string + ".%{arch}.rpm"
        q_cmd = r"rpm -q --qf '%s %s\n' --specfile '%s'" % (
            src_rpm, non_src_rpm, spec_path)

        out = os.popen(q_cmd)
        try:
            binary_rpms = []
            source_rpm = None
            while True:
                line = out.readline()
                if not line:
                    break
                l = line.strip().split()
                assert(len(l) == 2)
                binary_rpms.append(l[1])
                # The source rpm is named after the first entry in the spec file
                if source_rpm is None:
                    source_rpm = l[0]

            status = out.close()
            if status:
                raise DistutilsExecError("Failed to execute: %s" % repr(q_cmd))

        finally:
            out.close()

        self.spawn(rpm_cmd)

        if not self.dry_run:
            if self.distribution.has_ext_modules():
                pyversion = get_python_version()
            else:
                pyversion = 'any'

            if not self.binary_only:
                srpm = os.path.join(rpm_dir['SRPMS'], source_rpm)
                assert(os.path.exists(srpm))
                self.move_file(srpm, self.dist_dir)
                filename = os.path.join(self.dist_dir, source_rpm)
                self.distribution.dist_files.append(
                    ('bdist_rpm', pyversion, filename))

            if not self.source_only:
                for rpm in binary_rpms:
                    rpm = os.path.join(rpm_dir['RPMS'], rpm)
                    if os.path.exists(rpm):
                        self.move_file(rpm, self.dist_dir)
                        filename = os.path.join(self.dist_dir,
                                                os.path.basename(rpm))
                        self.distribution.dist_files.append(
                            ('bdist_rpm', pyversion, filename))

    def _dist_path(self, path):
        return os.path.join(self.dist_dir, os.path.basename(path))

    def _make_spec_file(self):
        """Generate the text of an RPM spec file and return it as a
        list of strings (one per line).
        """
        # definitions and headers
        spec_file = [
            '%define name ' + self.distribution.get_name(),
            '%define version ' + self.distribution.get_version().replace('-','_'),
            '%define unmangled_version ' + self.distribution.get_version(),
            '%define release ' + self.release.replace('-','_'),
            '',
            'Summary: ' + self.distribution.get_description(),
            ]

        # Workaround for #14443 which affects some RPM based systems such as
        # RHEL6 (and probably derivatives)
        vendor_hook = subprocess.getoutput('rpm --eval %{__os_install_post}')
        # Generate a potential replacement value for __os_install_post (whilst
        # normalizing the whitespace to simplify the test for whether the
        # invocation of brp-python-bytecompile passes in __python):
        vendor_hook = '\n'.join(['  %s \\' % line.strip()
                                 for line in vendor_hook.splitlines()])
        problem = "brp-python-bytecompile \\\n"
        fixed = "brp-python-bytecompile %{__python} \\\n"
        fixed_hook = vendor_hook.replace(problem, fixed)
        if fixed_hook != vendor_hook:
            spec_file.append('# Workaround for http://bugs.python.org/issue14443')
            spec_file.append('%define __os_install_post ' + fixed_hook + '\n')

        # put locale summaries into spec file
        # XXX not supported for now (hard to put a dictionary
        # in a config file -- arg!)
        #for locale in self.summaries.keys():
        #    spec_file.append('Summary(%s): %s' % (locale,
        #                                          self.summaries[locale]))

        spec_file.extend([
            'Name: %{name}',
            'Version: %{version}',
            'Release: %{release}',])

        # XXX yuck! this filename is available from the "sdist" command,
        # but only after it has run: and we create the spec file before
        # running "sdist", in case of --spec-only.
        if self.use_bzip2:
            spec_file.append('Source0: %{name}-%{unmangled_version}.tar.bz2')
        else:
            spec_file.append('Source0: %{name}-%{unmangled_version}.tar.gz')

        spec_file.extend([
            'License: ' + self.distribution.get_license(),
            'Group: ' + self.group,
            'BuildRoot: %{_tmppath}/%{name}-%{version}-%{release}-buildroot',
            'Prefix: %{_prefix}', ])

        if not self.force_arch:
            # noarch if no extension modules
            if not self.distribution.has_ext_modules():
                spec_file.append('BuildArch: noarch')
        else:
            spec_file.append( 'BuildArch: %s' % self.force_arch )

        for field in ('Vendor',
                      'Packager',
                      'Provides',
                      'Requires',
                      'Conflicts',
                      'Obsoletes',
                      ):
            val = getattr(self, field.lower())
            if isinstance(val, list):
                spec_file.append('%s: %s' % (field, ' '.join(val)))
            elif val is not None:
                spec_file.append('%s: %s' % (field, val))


        if self.distribution.get_url() != 'UNKNOWN':
            spec_file.append('Url: ' + self.distribution.get_url())

        if self.distribution_name:
            spec_file.append('Distribution: ' + self.distribution_name)

        if self.build_requires:
            spec_file.append('BuildRequires: ' +
                             ' '.join(self.build_requires))

        if self.icon:
            spec_file.append('Icon: ' + os.path.basename(self.icon))

        if self.no_autoreq:
            spec_file.append('AutoReq: 0')

        spec_file.extend([
            '',
            '%description',
            self.distribution.get_long_description()
            ])

        # put locale descriptions into spec file
        # XXX again, suppressed because config file syntax doesn't
        # easily support this ;-(
        #for locale in self.descriptions.keys():
        #    spec_file.extend([
        #        '',
        #        '%description -l ' + locale,
        #        self.descriptions[locale],
        #        ])

        # rpm scripts
        # figure out default build script
        def_setup_call = "%s %s" % (self.python,os.path.basename(sys.argv[0]))
        def_build = "%s build" % def_setup_call
        if self.use_rpm_opt_flags:
            def_build = 'env CFLAGS="$RPM_OPT_FLAGS" ' + def_build

        # insert contents of files

        # XXX this is kind of misleading: user-supplied options are files
        # that we open and interpolate into the spec file, but the defaults
        # are just text that we drop in as-is.  Hmmm.

        install_cmd = ('%s install -O1 --root=$RPM_BUILD_ROOT '
                       '--record=INSTALLED_FILES') % def_setup_call

        script_options = [
            ('prep', 'prep_script', "%setup -n %{name}-%{unmangled_version}"),
            ('build', 'build_script', def_build),
            ('install', 'install_script', install_cmd),
            ('clean', 'clean_script', "rm -rf $RPM_BUILD_ROOT"),
            ('verifyscript', 'verify_script', None),
            ('pre', 'pre_install', None),
            ('post', 'post_install', None),
            ('preun', 'pre_uninstall', None),
            ('postun', 'post_uninstall', None),
        ]

        for (rpm_opt, attr, default) in script_options:
            # Insert contents of file referred to, if no file is referred to
            # use 'default' as contents of script
            val = getattr(self, attr)
            if val or default:
                spec_file.extend([
                    '',
                    '%' + rpm_opt,])
                if val:
                    with open(val) as f:
                        spec_file.extend(f.read().split('\n'))
                else:
                    spec_file.append(default)


        # files section
        spec_file.extend([
            '',
            '%files -f INSTALLED_FILES',
            '%defattr(-,root,root)',
            ])

        if self.doc_files:
            spec_file.append('%doc ' + ' '.join(self.doc_files))

        if self.changelog:
            spec_file.extend([
                '',
                '%changelog',])
            spec_file.extend(self.changelog)

        return spec_file

    def _format_changelog(self, changelog):
        """Format the changelog correctly and convert it to a list of strings
        """
        if not changelog:
            return changelog
        new_changelog = []
        for line in changelog.strip().split('\n'):
            line = line.strip()
            if line[0] == '*':
                new_changelog.extend(['', line])
            elif line[0] == '-':
                new_changelog.append(line)
            else:
                new_changelog.append('  ' + line)

        # strip trailing newline inserted by first changelog entry
        if not new_changelog[0]:
            del new_changelog[0]

        return new_changelog
site-packages/setuptools/_distutils/command/install_egg_info.py000064400000005053150564745010021175 0ustar00"""distutils.command.install_egg_info

Implements the Distutils 'install_egg_info' command, for installing
a package's PKG-INFO metadata."""


from distutils.cmd import Command
from distutils import log, dir_util
import os, sys, re

class install_egg_info(Command):
    """Install an .egg-info file for the package"""

    description = "Install package's PKG-INFO metadata as an .egg-info file"
    user_options = [
        ('install-dir=', 'd', "directory to install to"),
    ]

    def initialize_options(self):
        self.install_dir = None

    def finalize_options(self):
        self.set_undefined_options('install_lib',('install_dir','install_dir'))
        basename = "%s-%s-py%d.%d.egg-info" % (
            to_filename(safe_name(self.distribution.get_name())),
            to_filename(safe_version(self.distribution.get_version())),
            *sys.version_info[:2]
        )
        self.target = os.path.join(self.install_dir, basename)
        self.outputs = [self.target]

    def run(self):
        target = self.target
        if os.path.isdir(target) and not os.path.islink(target):
            dir_util.remove_tree(target, dry_run=self.dry_run)
        elif os.path.exists(target):
            self.execute(os.unlink,(self.target,),"Removing "+target)
        elif not os.path.isdir(self.install_dir):
            self.execute(os.makedirs, (self.install_dir,),
                         "Creating "+self.install_dir)
        log.info("Writing %s", target)
        if not self.dry_run:
            with open(target, 'w', encoding='UTF-8') as f:
                self.distribution.metadata.write_pkg_file(f)

    def get_outputs(self):
        return self.outputs


# The following routines are taken from setuptools' pkg_resources module and
# can be replaced by importing them from pkg_resources once it is included
# in the stdlib.

def safe_name(name):
    """Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    """
    return re.sub('[^A-Za-z0-9.]+', '-', name)


def safe_version(version):
    """Convert an arbitrary string to a standard version string

    Spaces become dots, and all other non-alphanumeric characters become
    dashes, with runs of multiple dashes condensed to a single dash.
    """
    version = version.replace(' ','.')
    return re.sub('[^A-Za-z0-9.]+', '-', version)


def to_filename(name):
    """Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    """
    return name.replace('-','_')
site-packages/setuptools/_distutils/command/bdist_dumb.py000064400000011461150564745010020006 0ustar00"""distutils.command.bdist_dumb

Implements the Distutils 'bdist_dumb' command (create a "dumb" built
distribution -- i.e., just an archive to be unpacked under $prefix or
$exec_prefix)."""

import os
from distutils.core import Command
from distutils.util import get_platform
from distutils.dir_util import remove_tree, ensure_relative
from distutils.errors import *
from distutils.sysconfig import get_python_version
from distutils import log

class bdist_dumb(Command):

    description = "create a \"dumb\" built distribution"

    user_options = [('bdist-dir=', 'd',
                     "temporary directory for creating the distribution"),
                    ('plat-name=', 'p',
                     "platform name to embed in generated filenames "
                     "(default: %s)" % get_platform()),
                    ('format=', 'f',
                     "archive format to create (tar, gztar, bztar, xztar, "
                     "ztar, zip)"),
                    ('keep-temp', 'k',
                     "keep the pseudo-installation tree around after " +
                     "creating the distribution archive"),
                    ('dist-dir=', 'd',
                     "directory to put final built distributions in"),
                    ('skip-build', None,
                     "skip rebuilding everything (for testing/debugging)"),
                    ('relative', None,
                     "build the archive using relative paths "
                     "(default: false)"),
                    ('owner=', 'u',
                     "Owner name used when creating a tar file"
                     " [default: current user]"),
                    ('group=', 'g',
                     "Group name used when creating a tar file"
                     " [default: current group]"),
                   ]

    boolean_options = ['keep-temp', 'skip-build', 'relative']

    default_format = { 'posix': 'gztar',
                       'nt': 'zip' }

    def initialize_options(self):
        self.bdist_dir = None
        self.plat_name = None
        self.format = None
        self.keep_temp = 0
        self.dist_dir = None
        self.skip_build = None
        self.relative = 0
        self.owner = None
        self.group = None

    def finalize_options(self):
        if self.bdist_dir is None:
            bdist_base = self.get_finalized_command('bdist').bdist_base
            self.bdist_dir = os.path.join(bdist_base, 'dumb')

        if self.format is None:
            try:
                self.format = self.default_format[os.name]
            except KeyError:
                raise DistutilsPlatformError(
                       "don't know how to create dumb built distributions "
                       "on platform %s" % os.name)

        self.set_undefined_options('bdist',
                                   ('dist_dir', 'dist_dir'),
                                   ('plat_name', 'plat_name'),
                                   ('skip_build', 'skip_build'))

    def run(self):
        if not self.skip_build:
            self.run_command('build')

        install = self.reinitialize_command('install', reinit_subcommands=1)
        install.root = self.bdist_dir
        install.skip_build = self.skip_build
        install.warn_dir = 0

        log.info("installing to %s", self.bdist_dir)
        self.run_command('install')

        # And make an archive relative to the root of the
        # pseudo-installation tree.
        archive_basename = "%s.%s" % (self.distribution.get_fullname(),
                                      self.plat_name)

        pseudoinstall_root = os.path.join(self.dist_dir, archive_basename)
        if not self.relative:
            archive_root = self.bdist_dir
        else:
            if (self.distribution.has_ext_modules() and
                (install.install_base != install.install_platbase)):
                raise DistutilsPlatformError(
                       "can't make a dumb built distribution where "
                       "base and platbase are different (%s, %s)"
                       % (repr(install.install_base),
                          repr(install.install_platbase)))
            else:
                archive_root = os.path.join(self.bdist_dir,
                                   ensure_relative(install.install_base))

        # Make the archive
        filename = self.make_archive(pseudoinstall_root,
                                     self.format, root_dir=archive_root,
                                     owner=self.owner, group=self.group)
        if self.distribution.has_ext_modules():
            pyversion = get_python_version()
        else:
            pyversion = 'any'
        self.distribution.dist_files.append(('bdist_dumb', pyversion,
                                             filename))

        if not self.keep_temp:
            remove_tree(self.bdist_dir, dry_run=self.dry_run)
site-packages/setuptools/_distutils/command/bdist_wininst.py000064400000037236150564745010020562 0ustar00"""distutils.command.bdist_wininst

Implements the Distutils 'bdist_wininst' command: create a windows installer
exe-program."""

import os
import sys
import warnings
from distutils.core import Command
from distutils.util import get_platform
from distutils.dir_util import remove_tree
from distutils.errors import *
from distutils.sysconfig import get_python_version
from distutils import log

class bdist_wininst(Command):

    description = "create an executable installer for MS Windows"

    user_options = [('bdist-dir=', None,
                     "temporary directory for creating the distribution"),
                    ('plat-name=', 'p',
                     "platform name to embed in generated filenames "
                     "(default: %s)" % get_platform()),
                    ('keep-temp', 'k',
                     "keep the pseudo-installation tree around after " +
                     "creating the distribution archive"),
                    ('target-version=', None,
                     "require a specific python version" +
                     " on the target system"),
                    ('no-target-compile', 'c',
                     "do not compile .py to .pyc on the target system"),
                    ('no-target-optimize', 'o',
                     "do not compile .py to .pyo (optimized) "
                     "on the target system"),
                    ('dist-dir=', 'd',
                     "directory to put final built distributions in"),
                    ('bitmap=', 'b',
                     "bitmap to use for the installer instead of python-powered logo"),
                    ('title=', 't',
                     "title to display on the installer background instead of default"),
                    ('skip-build', None,
                     "skip rebuilding everything (for testing/debugging)"),
                    ('install-script=', None,
                     "basename of installation script to be run after "
                     "installation or before deinstallation"),
                    ('pre-install-script=', None,
                     "Fully qualified filename of a script to be run before "
                     "any files are installed.  This script need not be in the "
                     "distribution"),
                    ('user-access-control=', None,
                     "specify Vista's UAC handling - 'none'/default=no "
                     "handling, 'auto'=use UAC if target Python installed for "
                     "all users, 'force'=always use UAC"),
                   ]

    boolean_options = ['keep-temp', 'no-target-compile', 'no-target-optimize',
                       'skip-build']

    # bpo-10945: bdist_wininst requires mbcs encoding only available on Windows
    _unsupported = (sys.platform != "win32")

    def __init__(self, *args, **kw):
        super().__init__(*args, **kw)
        warnings.warn("bdist_wininst command is deprecated since Python 3.8, "
                      "use bdist_wheel (wheel packages) instead",
                      DeprecationWarning, 2)

    def initialize_options(self):
        self.bdist_dir = None
        self.plat_name = None
        self.keep_temp = 0
        self.no_target_compile = 0
        self.no_target_optimize = 0
        self.target_version = None
        self.dist_dir = None
        self.bitmap = None
        self.title = None
        self.skip_build = None
        self.install_script = None
        self.pre_install_script = None
        self.user_access_control = None


    def finalize_options(self):
        self.set_undefined_options('bdist', ('skip_build', 'skip_build'))

        if self.bdist_dir is None:
            if self.skip_build and self.plat_name:
                # If build is skipped and plat_name is overridden, bdist will
                # not see the correct 'plat_name' - so set that up manually.
                bdist = self.distribution.get_command_obj('bdist')
                bdist.plat_name = self.plat_name
                # next the command will be initialized using that name
            bdist_base = self.get_finalized_command('bdist').bdist_base
            self.bdist_dir = os.path.join(bdist_base, 'wininst')

        if not self.target_version:
            self.target_version = ""

        if not self.skip_build and self.distribution.has_ext_modules():
            short_version = get_python_version()
            if self.target_version and self.target_version != short_version:
                raise DistutilsOptionError(
                      "target version can only be %s, or the '--skip-build'" \
                      " option must be specified" % (short_version,))
            self.target_version = short_version

        self.set_undefined_options('bdist',
                                   ('dist_dir', 'dist_dir'),
                                   ('plat_name', 'plat_name'),
                                  )

        if self.install_script:
            for script in self.distribution.scripts:
                if self.install_script == os.path.basename(script):
                    break
            else:
                raise DistutilsOptionError(
                      "install_script '%s' not found in scripts"
                      % self.install_script)

    def run(self):
        if (sys.platform != "win32" and
            (self.distribution.has_ext_modules() or
             self.distribution.has_c_libraries())):
            raise DistutilsPlatformError \
                  ("distribution contains extensions and/or C libraries; "
                   "must be compiled on a Windows 32 platform")

        if not self.skip_build:
            self.run_command('build')

        install = self.reinitialize_command('install', reinit_subcommands=1)
        install.root = self.bdist_dir
        install.skip_build = self.skip_build
        install.warn_dir = 0
        install.plat_name = self.plat_name

        install_lib = self.reinitialize_command('install_lib')
        # we do not want to include pyc or pyo files
        install_lib.compile = 0
        install_lib.optimize = 0

        if self.distribution.has_ext_modules():
            # If we are building an installer for a Python version other
            # than the one we are currently running, then we need to ensure
            # our build_lib reflects the other Python version rather than ours.
            # Note that for target_version!=sys.version, we must have skipped the
            # build step, so there is no issue with enforcing the build of this
            # version.
            target_version = self.target_version
            if not target_version:
                assert self.skip_build, "Should have already checked this"
                target_version = '%d.%d' % sys.version_info[:2]
            plat_specifier = ".%s-%s" % (self.plat_name, target_version)
            build = self.get_finalized_command('build')
            build.build_lib = os.path.join(build.build_base,
                                           'lib' + plat_specifier)

        # Use a custom scheme for the zip-file, because we have to decide
        # at installation time which scheme to use.
        for key in ('purelib', 'platlib', 'headers', 'scripts', 'data'):
            value = key.upper()
            if key == 'headers':
                value = value + '/Include/$dist_name'
            setattr(install,
                    'install_' + key,
                    value)

        log.info("installing to %s", self.bdist_dir)
        install.ensure_finalized()

        # avoid warning of 'install_lib' about installing
        # into a directory not in sys.path
        sys.path.insert(0, os.path.join(self.bdist_dir, 'PURELIB'))

        install.run()

        del sys.path[0]

        # And make an archive relative to the root of the
        # pseudo-installation tree.
        from tempfile import mktemp
        archive_basename = mktemp()
        fullname = self.distribution.get_fullname()
        arcname = self.make_archive(archive_basename, "zip",
                                    root_dir=self.bdist_dir)
        # create an exe containing the zip-file
        self.create_exe(arcname, fullname, self.bitmap)
        if self.distribution.has_ext_modules():
            pyversion = get_python_version()
        else:
            pyversion = 'any'
        self.distribution.dist_files.append(('bdist_wininst', pyversion,
                                             self.get_installer_filename(fullname)))
        # remove the zip-file again
        log.debug("removing temporary file '%s'", arcname)
        os.remove(arcname)

        if not self.keep_temp:
            remove_tree(self.bdist_dir, dry_run=self.dry_run)

    def get_inidata(self):
        # Return data describing the installation.
        lines = []
        metadata = self.distribution.metadata

        # Write the [metadata] section.
        lines.append("[metadata]")

        # 'info' will be displayed in the installer's dialog box,
        # describing the items to be installed.
        info = (metadata.long_description or '') + '\n'

        # Escape newline characters
        def escape(s):
            return s.replace("\n", "\\n")

        for name in ["author", "author_email", "description", "maintainer",
                     "maintainer_email", "name", "url", "version"]:
            data = getattr(metadata, name, "")
            if data:
                info = info + ("\n    %s: %s" % \
                               (name.capitalize(), escape(data)))
                lines.append("%s=%s" % (name, escape(data)))

        # The [setup] section contains entries controlling
        # the installer runtime.
        lines.append("\n[Setup]")
        if self.install_script:
            lines.append("install_script=%s" % self.install_script)
        lines.append("info=%s" % escape(info))
        lines.append("target_compile=%d" % (not self.no_target_compile))
        lines.append("target_optimize=%d" % (not self.no_target_optimize))
        if self.target_version:
            lines.append("target_version=%s" % self.target_version)
        if self.user_access_control:
            lines.append("user_access_control=%s" % self.user_access_control)

        title = self.title or self.distribution.get_fullname()
        lines.append("title=%s" % escape(title))
        import time
        import distutils
        build_info = "Built %s with distutils-%s" % \
                     (time.ctime(time.time()), distutils.__version__)
        lines.append("build_info=%s" % build_info)
        return "\n".join(lines)

    def create_exe(self, arcname, fullname, bitmap=None):
        import struct

        self.mkpath(self.dist_dir)

        cfgdata = self.get_inidata()

        installer_name = self.get_installer_filename(fullname)
        self.announce("creating %s" % installer_name)

        if bitmap:
            with open(bitmap, "rb") as f:
                bitmapdata = f.read()
            bitmaplen = len(bitmapdata)
        else:
            bitmaplen = 0

        with open(installer_name, "wb") as file:
            file.write(self.get_exe_bytes())
            if bitmap:
                file.write(bitmapdata)

            # Convert cfgdata from unicode to ascii, mbcs encoded
            if isinstance(cfgdata, str):
                cfgdata = cfgdata.encode("mbcs")

            # Append the pre-install script
            cfgdata = cfgdata + b"\0"
            if self.pre_install_script:
                # We need to normalize newlines, so we open in text mode and
                # convert back to bytes. "latin-1" simply avoids any possible
                # failures.
                with open(self.pre_install_script, "r",
                          encoding="latin-1") as script:
                    script_data = script.read().encode("latin-1")
                cfgdata = cfgdata + script_data + b"\n\0"
            else:
                # empty pre-install script
                cfgdata = cfgdata + b"\0"
            file.write(cfgdata)

            # The 'magic number' 0x1234567B is used to make sure that the
            # binary layout of 'cfgdata' is what the wininst.exe binary
            # expects.  If the layout changes, increment that number, make
            # the corresponding changes to the wininst.exe sources, and
            # recompile them.
            header = struct.pack("<iii",
                                0x1234567B,       # tag
                                len(cfgdata),     # length
                                bitmaplen,        # number of bytes in bitmap
                                )
            file.write(header)
            with open(arcname, "rb") as f:
                file.write(f.read())

    def get_installer_filename(self, fullname):
        # Factored out to allow overriding in subclasses
        if self.target_version:
            # if we create an installer for a specific python version,
            # it's better to include this in the name
            installer_name = os.path.join(self.dist_dir,
                                          "%s.%s-py%s.exe" %
                                           (fullname, self.plat_name, self.target_version))
        else:
            installer_name = os.path.join(self.dist_dir,
                                          "%s.%s.exe" % (fullname, self.plat_name))
        return installer_name

    def get_exe_bytes(self):
        # If a target-version other than the current version has been
        # specified, then using the MSVC version from *this* build is no good.
        # Without actually finding and executing the target version and parsing
        # its sys.version, we just hard-code our knowledge of old versions.
        # NOTE: Possible alternative is to allow "--target-version" to
        # specify a Python executable rather than a simple version string.
        # We can then execute this program to obtain any info we need, such
        # as the real sys.version string for the build.
        cur_version = get_python_version()

        # If the target version is *later* than us, then we assume they
        # use what we use
        # string compares seem wrong, but are what sysconfig.py itself uses
        if self.target_version and self.target_version < cur_version:
            if self.target_version < "2.4":
                bv = '6.0'
            elif self.target_version == "2.4":
                bv = '7.1'
            elif self.target_version == "2.5":
                bv = '8.0'
            elif self.target_version <= "3.2":
                bv = '9.0'
            elif self.target_version <= "3.4":
                bv = '10.0'
            else:
                bv = '14.0'
        else:
            # for current version - use authoritative check.
            try:
                from msvcrt import CRT_ASSEMBLY_VERSION
            except ImportError:
                # cross-building, so assume the latest version
                bv = '14.0'
            else:
                # as far as we know, CRT is binary compatible based on
                # the first field, so assume 'x.0' until proven otherwise
                major = CRT_ASSEMBLY_VERSION.partition('.')[0]
                bv = major + '.0'


        # wininst-x.y.exe is in the same directory as this file
        directory = os.path.dirname(__file__)
        # we must use a wininst-x.y.exe built with the same C compiler
        # used for python.  XXX What about mingw, borland, and so on?

        # if plat_name starts with "win" but is not "win32"
        # we want to strip "win" and leave the rest (e.g. -amd64)
        # for all other cases, we don't want any suffix
        if self.plat_name != 'win32' and self.plat_name[:3] == 'win':
            sfix = self.plat_name[3:]
        else:
            sfix = ''

        filename = os.path.join(directory, "wininst-%s%s.exe" % (bv, sfix))
        f = open(filename, "rb")
        try:
            return f.read()
        finally:
            f.close()
site-packages/setuptools/_distutils/command/register.py000064400000026700150564745010017520 0ustar00"""distutils.command.register

Implements the Distutils 'register' command (register with the repository).
"""

# created 2002/10/21, Richard Jones

import getpass
import io
import urllib.parse, urllib.request
from warnings import warn

from distutils.core import PyPIRCCommand
from distutils.errors import *
from distutils import log

class register(PyPIRCCommand):

    description = ("register the distribution with the Python package index")
    user_options = PyPIRCCommand.user_options + [
        ('list-classifiers', None,
         'list the valid Trove classifiers'),
        ('strict', None ,
         'Will stop the registering if the meta-data are not fully compliant')
        ]
    boolean_options = PyPIRCCommand.boolean_options + [
        'verify', 'list-classifiers', 'strict']

    sub_commands = [('check', lambda self: True)]

    def initialize_options(self):
        PyPIRCCommand.initialize_options(self)
        self.list_classifiers = 0
        self.strict = 0

    def finalize_options(self):
        PyPIRCCommand.finalize_options(self)
        # setting options for the `check` subcommand
        check_options = {'strict': ('register', self.strict),
                         'restructuredtext': ('register', 1)}
        self.distribution.command_options['check'] = check_options

    def run(self):
        self.finalize_options()
        self._set_config()

        # Run sub commands
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)

        if self.dry_run:
            self.verify_metadata()
        elif self.list_classifiers:
            self.classifiers()
        else:
            self.send_metadata()

    def check_metadata(self):
        """Deprecated API."""
        warn("distutils.command.register.check_metadata is deprecated, \
              use the check command instead", PendingDeprecationWarning)
        check = self.distribution.get_command_obj('check')
        check.ensure_finalized()
        check.strict = self.strict
        check.restructuredtext = 1
        check.run()

    def _set_config(self):
        ''' Reads the configuration file and set attributes.
        '''
        config = self._read_pypirc()
        if config != {}:
            self.username = config['username']
            self.password = config['password']
            self.repository = config['repository']
            self.realm = config['realm']
            self.has_config = True
        else:
            if self.repository not in ('pypi', self.DEFAULT_REPOSITORY):
                raise ValueError('%s not found in .pypirc' % self.repository)
            if self.repository == 'pypi':
                self.repository = self.DEFAULT_REPOSITORY
            self.has_config = False

    def classifiers(self):
        ''' Fetch the list of classifiers from the server.
        '''
        url = self.repository+'?:action=list_classifiers'
        response = urllib.request.urlopen(url)
        log.info(self._read_pypi_response(response))

    def verify_metadata(self):
        ''' Send the metadata to the package index server to be checked.
        '''
        # send the info to the server and report the result
        (code, result) = self.post_to_server(self.build_post_data('verify'))
        log.info('Server response (%s): %s', code, result)

    def send_metadata(self):
        ''' Send the metadata to the package index server.

            Well, do the following:
            1. figure who the user is, and then
            2. send the data as a Basic auth'ed POST.

            First we try to read the username/password from $HOME/.pypirc,
            which is a ConfigParser-formatted file with a section
            [distutils] containing username and password entries (both
            in clear text). Eg:

                [distutils]
                index-servers =
                    pypi

                [pypi]
                username: fred
                password: sekrit

            Otherwise, to figure who the user is, we offer the user three
            choices:

             1. use existing login,
             2. register as a new user, or
             3. set the password to a random string and email the user.

        '''
        # see if we can short-cut and get the username/password from the
        # config
        if self.has_config:
            choice = '1'
            username = self.username
            password = self.password
        else:
            choice = 'x'
            username = password = ''

        # get the user's login info
        choices = '1 2 3 4'.split()
        while choice not in choices:
            self.announce('''\
We need to know who you are, so please choose either:
 1. use your existing login,
 2. register as a new user,
 3. have the server generate a new password for you (and email it to you), or
 4. quit
Your selection [default 1]: ''', log.INFO)
            choice = input()
            if not choice:
                choice = '1'
            elif choice not in choices:
                print('Please choose one of the four options!')

        if choice == '1':
            # get the username and password
            while not username:
                username = input('Username: ')
            while not password:
                password = getpass.getpass('Password: ')

            # set up the authentication
            auth = urllib.request.HTTPPasswordMgr()
            host = urllib.parse.urlparse(self.repository)[1]
            auth.add_password(self.realm, host, username, password)
            # send the info to the server and report the result
            code, result = self.post_to_server(self.build_post_data('submit'),
                auth)
            self.announce('Server response (%s): %s' % (code, result),
                          log.INFO)

            # possibly save the login
            if code == 200:
                if self.has_config:
                    # sharing the password in the distribution instance
                    # so the upload command can reuse it
                    self.distribution.password = password
                else:
                    self.announce(('I can store your PyPI login so future '
                                   'submissions will be faster.'), log.INFO)
                    self.announce('(the login will be stored in %s)' % \
                                  self._get_rc_file(), log.INFO)
                    choice = 'X'
                    while choice.lower() not in 'yn':
                        choice = input('Save your login (y/N)?')
                        if not choice:
                            choice = 'n'
                    if choice.lower() == 'y':
                        self._store_pypirc(username, password)

        elif choice == '2':
            data = {':action': 'user'}
            data['name'] = data['password'] = data['email'] = ''
            data['confirm'] = None
            while not data['name']:
                data['name'] = input('Username: ')
            while data['password'] != data['confirm']:
                while not data['password']:
                    data['password'] = getpass.getpass('Password: ')
                while not data['confirm']:
                    data['confirm'] = getpass.getpass(' Confirm: ')
                if data['password'] != data['confirm']:
                    data['password'] = ''
                    data['confirm'] = None
                    print("Password and confirm don't match!")
            while not data['email']:
                data['email'] = input('   EMail: ')
            code, result = self.post_to_server(data)
            if code != 200:
                log.info('Server response (%s): %s', code, result)
            else:
                log.info('You will receive an email shortly.')
                log.info(('Follow the instructions in it to '
                          'complete registration.'))
        elif choice == '3':
            data = {':action': 'password_reset'}
            data['email'] = ''
            while not data['email']:
                data['email'] = input('Your email address: ')
            code, result = self.post_to_server(data)
            log.info('Server response (%s): %s', code, result)

    def build_post_data(self, action):
        # figure the data to send - the metadata plus some additional
        # information used by the package server
        meta = self.distribution.metadata
        data = {
            ':action': action,
            'metadata_version' : '1.0',
            'name': meta.get_name(),
            'version': meta.get_version(),
            'summary': meta.get_description(),
            'home_page': meta.get_url(),
            'author': meta.get_contact(),
            'author_email': meta.get_contact_email(),
            'license': meta.get_licence(),
            'description': meta.get_long_description(),
            'keywords': meta.get_keywords(),
            'platform': meta.get_platforms(),
            'classifiers': meta.get_classifiers(),
            'download_url': meta.get_download_url(),
            # PEP 314
            'provides': meta.get_provides(),
            'requires': meta.get_requires(),
            'obsoletes': meta.get_obsoletes(),
        }
        if data['provides'] or data['requires'] or data['obsoletes']:
            data['metadata_version'] = '1.1'
        return data

    def post_to_server(self, data, auth=None):
        ''' Post a query to the server, and return a string response.
        '''
        if 'name' in data:
            self.announce('Registering %s to %s' % (data['name'],
                                                    self.repository),
                                                    log.INFO)
        # Build up the MIME payload for the urllib2 POST data
        boundary = '--------------GHSKFJDLGDS7543FJKLFHRE75642756743254'
        sep_boundary = '\n--' + boundary
        end_boundary = sep_boundary + '--'
        body = io.StringIO()
        for key, value in data.items():
            # handle multiple entries for the same name
            if type(value) not in (type([]), type( () )):
                value = [value]
            for value in value:
                value = str(value)
                body.write(sep_boundary)
                body.write('\nContent-Disposition: form-data; name="%s"'%key)
                body.write("\n\n")
                body.write(value)
                if value and value[-1] == '\r':
                    body.write('\n')  # write an extra newline (lurve Macs)
        body.write(end_boundary)
        body.write("\n")
        body = body.getvalue().encode("utf-8")

        # build the Request
        headers = {
            'Content-type': 'multipart/form-data; boundary=%s; charset=utf-8'%boundary,
            'Content-length': str(len(body))
        }
        req = urllib.request.Request(self.repository, body, headers)

        # handle HTTP and include the Basic Auth handler
        opener = urllib.request.build_opener(
            urllib.request.HTTPBasicAuthHandler(password_mgr=auth)
        )
        data = ''
        try:
            result = opener.open(req)
        except urllib.error.HTTPError as e:
            if self.show_response:
                data = e.fp.read()
            result = e.code, e.msg
        except urllib.error.URLError as e:
            result = 500, str(e)
        else:
            if self.show_response:
                data = self._read_pypi_response(result)
            result = 200, 'OK'
        if self.show_response:
            msg = '\n'.join(('-' * 75, data, '-' * 75))
            self.announce(msg, log.INFO)
        return result
site-packages/setuptools/_distutils/command/sdist.py000064400000045075150564745010017030 0ustar00"""distutils.command.sdist

Implements the Distutils 'sdist' command (create a source distribution)."""

import os
import sys
from glob import glob
from warnings import warn

from distutils.core import Command
from distutils import dir_util
from distutils import file_util
from distutils import archive_util
from distutils.text_file import TextFile
from distutils.filelist import FileList
from distutils import log
from distutils.util import convert_path
from distutils.errors import DistutilsTemplateError, DistutilsOptionError


def show_formats():
    """Print all possible values for the 'formats' option (used by
    the "--help-formats" command-line option).
    """
    from distutils.fancy_getopt import FancyGetopt
    from distutils.archive_util import ARCHIVE_FORMATS
    formats = []
    for format in ARCHIVE_FORMATS.keys():
        formats.append(("formats=" + format, None,
                        ARCHIVE_FORMATS[format][2]))
    formats.sort()
    FancyGetopt(formats).print_help(
        "List of available source distribution formats:")


class sdist(Command):

    description = "create a source distribution (tarball, zip file, etc.)"

    def checking_metadata(self):
        """Callable used for the check sub-command.

        Placed here so user_options can view it"""
        return self.metadata_check

    user_options = [
        ('template=', 't',
         "name of manifest template file [default: MANIFEST.in]"),
        ('manifest=', 'm',
         "name of manifest file [default: MANIFEST]"),
        ('use-defaults', None,
         "include the default file set in the manifest "
         "[default; disable with --no-defaults]"),
        ('no-defaults', None,
         "don't include the default file set"),
        ('prune', None,
         "specifically exclude files/directories that should not be "
         "distributed (build tree, RCS/CVS dirs, etc.) "
         "[default; disable with --no-prune]"),
        ('no-prune', None,
         "don't automatically exclude anything"),
        ('manifest-only', 'o',
         "just regenerate the manifest and then stop "
         "(implies --force-manifest)"),
        ('force-manifest', 'f',
         "forcibly regenerate the manifest and carry on as usual. "
         "Deprecated: now the manifest is always regenerated."),
        ('formats=', None,
         "formats for source distribution (comma-separated list)"),
        ('keep-temp', 'k',
         "keep the distribution tree around after creating " +
         "archive file(s)"),
        ('dist-dir=', 'd',
         "directory to put the source distribution archive(s) in "
         "[default: dist]"),
        ('metadata-check', None,
         "Ensure that all required elements of meta-data "
         "are supplied. Warn if any missing. [default]"),
        ('owner=', 'u',
         "Owner name used when creating a tar file [default: current user]"),
        ('group=', 'g',
         "Group name used when creating a tar file [default: current group]"),
        ]

    boolean_options = ['use-defaults', 'prune',
                       'manifest-only', 'force-manifest',
                       'keep-temp', 'metadata-check']

    help_options = [
        ('help-formats', None,
         "list available distribution formats", show_formats),
        ]

    negative_opt = {'no-defaults': 'use-defaults',
                    'no-prune': 'prune' }

    sub_commands = [('check', checking_metadata)]

    READMES = ('README', 'README.txt', 'README.rst')

    def initialize_options(self):
        # 'template' and 'manifest' are, respectively, the names of
        # the manifest template and manifest file.
        self.template = None
        self.manifest = None

        # 'use_defaults': if true, we will include the default file set
        # in the manifest
        self.use_defaults = 1
        self.prune = 1

        self.manifest_only = 0
        self.force_manifest = 0

        self.formats = ['gztar']
        self.keep_temp = 0
        self.dist_dir = None

        self.archive_files = None
        self.metadata_check = 1
        self.owner = None
        self.group = None

    def finalize_options(self):
        if self.manifest is None:
            self.manifest = "MANIFEST"
        if self.template is None:
            self.template = "MANIFEST.in"

        self.ensure_string_list('formats')

        bad_format = archive_util.check_archive_formats(self.formats)
        if bad_format:
            raise DistutilsOptionError(
                  "unknown archive format '%s'" % bad_format)

        if self.dist_dir is None:
            self.dist_dir = "dist"

    def run(self):
        # 'filelist' contains the list of files that will make up the
        # manifest
        self.filelist = FileList()

        # Run sub commands
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)

        # Do whatever it takes to get the list of files to process
        # (process the manifest template, read an existing manifest,
        # whatever).  File list is accumulated in 'self.filelist'.
        self.get_file_list()

        # If user just wanted us to regenerate the manifest, stop now.
        if self.manifest_only:
            return

        # Otherwise, go ahead and create the source distribution tarball,
        # or zipfile, or whatever.
        self.make_distribution()

    def check_metadata(self):
        """Deprecated API."""
        warn("distutils.command.sdist.check_metadata is deprecated, \
              use the check command instead", PendingDeprecationWarning)
        check = self.distribution.get_command_obj('check')
        check.ensure_finalized()
        check.run()

    def get_file_list(self):
        """Figure out the list of files to include in the source
        distribution, and put it in 'self.filelist'.  This might involve
        reading the manifest template (and writing the manifest), or just
        reading the manifest, or just using the default file set -- it all
        depends on the user's options.
        """
        # new behavior when using a template:
        # the file list is recalculated every time because
        # even if MANIFEST.in or setup.py are not changed
        # the user might have added some files in the tree that
        # need to be included.
        #
        #  This makes --force the default and only behavior with templates.
        template_exists = os.path.isfile(self.template)
        if not template_exists and self._manifest_is_not_generated():
            self.read_manifest()
            self.filelist.sort()
            self.filelist.remove_duplicates()
            return

        if not template_exists:
            self.warn(("manifest template '%s' does not exist " +
                        "(using default file list)") %
                        self.template)
        self.filelist.findall()

        if self.use_defaults:
            self.add_defaults()

        if template_exists:
            self.read_template()

        if self.prune:
            self.prune_file_list()

        self.filelist.sort()
        self.filelist.remove_duplicates()
        self.write_manifest()

    def add_defaults(self):
        """Add all the default files to self.filelist:
          - README or README.txt
          - setup.py
          - test/test*.py
          - all pure Python modules mentioned in setup script
          - all files pointed by package_data (build_py)
          - all files defined in data_files.
          - all files defined as scripts.
          - all C sources listed as part of extensions or C libraries
            in the setup script (doesn't catch C headers!)
        Warns if (README or README.txt) or setup.py are missing; everything
        else is optional.
        """
        self._add_defaults_standards()
        self._add_defaults_optional()
        self._add_defaults_python()
        self._add_defaults_data_files()
        self._add_defaults_ext()
        self._add_defaults_c_libs()
        self._add_defaults_scripts()

    @staticmethod
    def _cs_path_exists(fspath):
        """
        Case-sensitive path existence check

        >>> sdist._cs_path_exists(__file__)
        True
        >>> sdist._cs_path_exists(__file__.upper())
        False
        """
        if not os.path.exists(fspath):
            return False
        # make absolute so we always have a directory
        abspath = os.path.abspath(fspath)
        directory, filename = os.path.split(abspath)
        return filename in os.listdir(directory)

    def _add_defaults_standards(self):
        standards = [self.READMES, self.distribution.script_name]
        for fn in standards:
            if isinstance(fn, tuple):
                alts = fn
                got_it = False
                for fn in alts:
                    if self._cs_path_exists(fn):
                        got_it = True
                        self.filelist.append(fn)
                        break

                if not got_it:
                    self.warn("standard file not found: should have one of " +
                              ', '.join(alts))
            else:
                if self._cs_path_exists(fn):
                    self.filelist.append(fn)
                else:
                    self.warn("standard file '%s' not found" % fn)

    def _add_defaults_optional(self):
        optional = ['test/test*.py', 'setup.cfg']
        for pattern in optional:
            files = filter(os.path.isfile, glob(pattern))
            self.filelist.extend(files)

    def _add_defaults_python(self):
        # build_py is used to get:
        #  - python modules
        #  - files defined in package_data
        build_py = self.get_finalized_command('build_py')

        # getting python files
        if self.distribution.has_pure_modules():
            self.filelist.extend(build_py.get_source_files())

        # getting package_data files
        # (computed in build_py.data_files by build_py.finalize_options)
        for pkg, src_dir, build_dir, filenames in build_py.data_files:
            for filename in filenames:
                self.filelist.append(os.path.join(src_dir, filename))

    def _add_defaults_data_files(self):
        # getting distribution.data_files
        if self.distribution.has_data_files():
            for item in self.distribution.data_files:
                if isinstance(item, str):
                    # plain file
                    item = convert_path(item)
                    if os.path.isfile(item):
                        self.filelist.append(item)
                else:
                    # a (dirname, filenames) tuple
                    dirname, filenames = item
                    for f in filenames:
                        f = convert_path(f)
                        if os.path.isfile(f):
                            self.filelist.append(f)

    def _add_defaults_ext(self):
        if self.distribution.has_ext_modules():
            build_ext = self.get_finalized_command('build_ext')
            self.filelist.extend(build_ext.get_source_files())

    def _add_defaults_c_libs(self):
        if self.distribution.has_c_libraries():
            build_clib = self.get_finalized_command('build_clib')
            self.filelist.extend(build_clib.get_source_files())

    def _add_defaults_scripts(self):
        if self.distribution.has_scripts():
            build_scripts = self.get_finalized_command('build_scripts')
            self.filelist.extend(build_scripts.get_source_files())

    def read_template(self):
        """Read and parse manifest template file named by self.template.

        (usually "MANIFEST.in") The parsing and processing is done by
        'self.filelist', which updates itself accordingly.
        """
        log.info("reading manifest template '%s'", self.template)
        template = TextFile(self.template, strip_comments=1, skip_blanks=1,
                            join_lines=1, lstrip_ws=1, rstrip_ws=1,
                            collapse_join=1)

        try:
            while True:
                line = template.readline()
                if line is None:            # end of file
                    break

                try:
                    self.filelist.process_template_line(line)
                # the call above can raise a DistutilsTemplateError for
                # malformed lines, or a ValueError from the lower-level
                # convert_path function
                except (DistutilsTemplateError, ValueError) as msg:
                    self.warn("%s, line %d: %s" % (template.filename,
                                                   template.current_line,
                                                   msg))
        finally:
            template.close()

    def prune_file_list(self):
        """Prune off branches that might slip into the file list as created
        by 'read_template()', but really don't belong there:
          * the build tree (typically "build")
          * the release tree itself (only an issue if we ran "sdist"
            previously with --keep-temp, or it aborted)
          * any RCS, CVS, .svn, .hg, .git, .bzr, _darcs directories
        """
        build = self.get_finalized_command('build')
        base_dir = self.distribution.get_fullname()

        self.filelist.exclude_pattern(None, prefix=build.build_base)
        self.filelist.exclude_pattern(None, prefix=base_dir)

        if sys.platform == 'win32':
            seps = r'/|\\'
        else:
            seps = '/'

        vcs_dirs = ['RCS', 'CVS', r'\.svn', r'\.hg', r'\.git', r'\.bzr',
                    '_darcs']
        vcs_ptrn = r'(^|%s)(%s)(%s).*' % (seps, '|'.join(vcs_dirs), seps)
        self.filelist.exclude_pattern(vcs_ptrn, is_regex=1)

    def write_manifest(self):
        """Write the file list in 'self.filelist' (presumably as filled in
        by 'add_defaults()' and 'read_template()') to the manifest file
        named by 'self.manifest'.
        """
        if self._manifest_is_not_generated():
            log.info("not writing to manually maintained "
                     "manifest file '%s'" % self.manifest)
            return

        content = self.filelist.files[:]
        content.insert(0, '# file GENERATED by distutils, do NOT edit')
        self.execute(file_util.write_file, (self.manifest, content),
                     "writing manifest file '%s'" % self.manifest)

    def _manifest_is_not_generated(self):
        # check for special comment used in 3.1.3 and higher
        if not os.path.isfile(self.manifest):
            return False

        fp = open(self.manifest)
        try:
            first_line = fp.readline()
        finally:
            fp.close()
        return first_line != '# file GENERATED by distutils, do NOT edit\n'

    def read_manifest(self):
        """Read the manifest file (named by 'self.manifest') and use it to
        fill in 'self.filelist', the list of files to include in the source
        distribution.
        """
        log.info("reading manifest file '%s'", self.manifest)
        with open(self.manifest) as manifest:
            for line in manifest:
                # ignore comments and blank lines
                line = line.strip()
                if line.startswith('#') or not line:
                    continue
                self.filelist.append(line)

    def make_release_tree(self, base_dir, files):
        """Create the directory tree that will become the source
        distribution archive.  All directories implied by the filenames in
        'files' are created under 'base_dir', and then we hard link or copy
        (if hard linking is unavailable) those files into place.
        Essentially, this duplicates the developer's source tree, but in a
        directory named after the distribution, containing only the files
        to be distributed.
        """
        # Create all the directories under 'base_dir' necessary to
        # put 'files' there; the 'mkpath()' is just so we don't die
        # if the manifest happens to be empty.
        self.mkpath(base_dir)
        dir_util.create_tree(base_dir, files, dry_run=self.dry_run)

        # And walk over the list of files, either making a hard link (if
        # os.link exists) to each one that doesn't already exist in its
        # corresponding location under 'base_dir', or copying each file
        # that's out-of-date in 'base_dir'.  (Usually, all files will be
        # out-of-date, because by default we blow away 'base_dir' when
        # we're done making the distribution archives.)

        if hasattr(os, 'link'):        # can make hard links on this system
            link = 'hard'
            msg = "making hard links in %s..." % base_dir
        else:                           # nope, have to copy
            link = None
            msg = "copying files to %s..." % base_dir

        if not files:
            log.warn("no files to distribute -- empty manifest?")
        else:
            log.info(msg)
        for file in files:
            if not os.path.isfile(file):
                log.warn("'%s' not a regular file -- skipping", file)
            else:
                dest = os.path.join(base_dir, file)
                self.copy_file(file, dest, link=link)

        self.distribution.metadata.write_pkg_info(base_dir)

    def make_distribution(self):
        """Create the source distribution(s).  First, we create the release
        tree with 'make_release_tree()'; then, we create all required
        archive files (according to 'self.formats') from the release tree.
        Finally, we clean up by blowing away the release tree (unless
        'self.keep_temp' is true).  The list of archive files created is
        stored so it can be retrieved later by 'get_archive_files()'.
        """
        # Don't warn about missing meta-data here -- should be (and is!)
        # done elsewhere.
        base_dir = self.distribution.get_fullname()
        base_name = os.path.join(self.dist_dir, base_dir)

        self.make_release_tree(base_dir, self.filelist.files)
        archive_files = []              # remember names of files we create
        # tar archive must be created last to avoid overwrite and remove
        if 'tar' in self.formats:
            self.formats.append(self.formats.pop(self.formats.index('tar')))

        for fmt in self.formats:
            file = self.make_archive(base_name, fmt, base_dir=base_dir,
                                     owner=self.owner, group=self.group)
            archive_files.append(file)
            self.distribution.dist_files.append(('sdist', '', file))

        self.archive_files = archive_files

        if not self.keep_temp:
            dir_util.remove_tree(base_dir, dry_run=self.dry_run)

    def get_archive_files(self):
        """Return the list of archive files created when the command
        was run, or None if the command hasn't run yet.
        """
        return self.archive_files
site-packages/setuptools/_distutils/command/clean.py000064400000005330150564745010016752 0ustar00"""distutils.command.clean

Implements the Distutils 'clean' command."""

# contributed by Bastian Kleineidam <calvin@cs.uni-sb.de>, added 2000-03-18

import os
from distutils.core import Command
from distutils.dir_util import remove_tree
from distutils import log

class clean(Command):

    description = "clean up temporary files from 'build' command"
    user_options = [
        ('build-base=', 'b',
         "base build directory (default: 'build.build-base')"),
        ('build-lib=', None,
         "build directory for all modules (default: 'build.build-lib')"),
        ('build-temp=', 't',
         "temporary build directory (default: 'build.build-temp')"),
        ('build-scripts=', None,
         "build directory for scripts (default: 'build.build-scripts')"),
        ('bdist-base=', None,
         "temporary directory for built distributions"),
        ('all', 'a',
         "remove all build output, not just temporary by-products")
    ]

    boolean_options = ['all']

    def initialize_options(self):
        self.build_base = None
        self.build_lib = None
        self.build_temp = None
        self.build_scripts = None
        self.bdist_base = None
        self.all = None

    def finalize_options(self):
        self.set_undefined_options('build',
                                   ('build_base', 'build_base'),
                                   ('build_lib', 'build_lib'),
                                   ('build_scripts', 'build_scripts'),
                                   ('build_temp', 'build_temp'))
        self.set_undefined_options('bdist',
                                   ('bdist_base', 'bdist_base'))

    def run(self):
        # remove the build/temp.<plat> directory (unless it's already
        # gone)
        if os.path.exists(self.build_temp):
            remove_tree(self.build_temp, dry_run=self.dry_run)
        else:
            log.debug("'%s' does not exist -- can't clean it",
                      self.build_temp)

        if self.all:
            # remove build directories
            for directory in (self.build_lib,
                              self.bdist_base,
                              self.build_scripts):
                if os.path.exists(directory):
                    remove_tree(directory, dry_run=self.dry_run)
                else:
                    log.warn("'%s' does not exist -- can't clean it",
                             directory)

        # just for the heck of it, try to remove the base build directory:
        # we might have emptied it right now, but if not we don't care
        if not self.dry_run:
            try:
                os.rmdir(self.build_base)
                log.info("removing '%s'", self.build_base)
            except OSError:
                pass
site-packages/setuptools/_distutils/fancy_getopt.py000064400000042570150564745010016743 0ustar00"""distutils.fancy_getopt

Wrapper around the standard getopt module that provides the following
additional features:
  * short and long options are tied together
  * options have help strings, so fancy_getopt could potentially
    create a complete usage summary
  * options set attributes of a passed-in object
"""

import sys, string, re
import getopt
from distutils.errors import *

# Much like command_re in distutils.core, this is close to but not quite
# the same as a Python NAME -- except, in the spirit of most GNU
# utilities, we use '-' in place of '_'.  (The spirit of LISP lives on!)
# The similarities to NAME are again not a coincidence...
longopt_pat = r'[a-zA-Z](?:[a-zA-Z0-9-]*)'
longopt_re = re.compile(r'^%s$' % longopt_pat)

# For recognizing "negative alias" options, eg. "quiet=!verbose"
neg_alias_re = re.compile("^(%s)=!(%s)$" % (longopt_pat, longopt_pat))

# This is used to translate long options to legitimate Python identifiers
# (for use as attributes of some object).
longopt_xlate = str.maketrans('-', '_')

class FancyGetopt:
    """Wrapper around the standard 'getopt()' module that provides some
    handy extra functionality:
      * short and long options are tied together
      * options have help strings, and help text can be assembled
        from them
      * options set attributes of a passed-in object
      * boolean options can have "negative aliases" -- eg. if
        --quiet is the "negative alias" of --verbose, then "--quiet"
        on the command line sets 'verbose' to false
    """

    def __init__(self, option_table=None):
        # The option table is (currently) a list of tuples.  The
        # tuples may have 3 or four values:
        #   (long_option, short_option, help_string [, repeatable])
        # if an option takes an argument, its long_option should have '='
        # appended; short_option should just be a single character, no ':'
        # in any case.  If a long_option doesn't have a corresponding
        # short_option, short_option should be None.  All option tuples
        # must have long options.
        self.option_table = option_table

        # 'option_index' maps long option names to entries in the option
        # table (ie. those 3-tuples).
        self.option_index = {}
        if self.option_table:
            self._build_index()

        # 'alias' records (duh) alias options; {'foo': 'bar'} means
        # --foo is an alias for --bar
        self.alias = {}

        # 'negative_alias' keeps track of options that are the boolean
        # opposite of some other option
        self.negative_alias = {}

        # These keep track of the information in the option table.  We
        # don't actually populate these structures until we're ready to
        # parse the command-line, since the 'option_table' passed in here
        # isn't necessarily the final word.
        self.short_opts = []
        self.long_opts = []
        self.short2long = {}
        self.attr_name = {}
        self.takes_arg = {}

        # And 'option_order' is filled up in 'getopt()'; it records the
        # original order of options (and their values) on the command-line,
        # but expands short options, converts aliases, etc.
        self.option_order = []

    def _build_index(self):
        self.option_index.clear()
        for option in self.option_table:
            self.option_index[option[0]] = option

    def set_option_table(self, option_table):
        self.option_table = option_table
        self._build_index()

    def add_option(self, long_option, short_option=None, help_string=None):
        if long_option in self.option_index:
            raise DistutilsGetoptError(
                  "option conflict: already an option '%s'" % long_option)
        else:
            option = (long_option, short_option, help_string)
            self.option_table.append(option)
            self.option_index[long_option] = option

    def has_option(self, long_option):
        """Return true if the option table for this parser has an
        option with long name 'long_option'."""
        return long_option in self.option_index

    def get_attr_name(self, long_option):
        """Translate long option name 'long_option' to the form it
        has as an attribute of some object: ie., translate hyphens
        to underscores."""
        return long_option.translate(longopt_xlate)

    def _check_alias_dict(self, aliases, what):
        assert isinstance(aliases, dict)
        for (alias, opt) in aliases.items():
            if alias not in self.option_index:
                raise DistutilsGetoptError(("invalid %s '%s': "
                       "option '%s' not defined") % (what, alias, alias))
            if opt not in self.option_index:
                raise DistutilsGetoptError(("invalid %s '%s': "
                       "aliased option '%s' not defined") % (what, alias, opt))

    def set_aliases(self, alias):
        """Set the aliases for this option parser."""
        self._check_alias_dict(alias, "alias")
        self.alias = alias

    def set_negative_aliases(self, negative_alias):
        """Set the negative aliases for this option parser.
        'negative_alias' should be a dictionary mapping option names to
        option names, both the key and value must already be defined
        in the option table."""
        self._check_alias_dict(negative_alias, "negative alias")
        self.negative_alias = negative_alias

    def _grok_option_table(self):
        """Populate the various data structures that keep tabs on the
        option table.  Called by 'getopt()' before it can do anything
        worthwhile.
        """
        self.long_opts = []
        self.short_opts = []
        self.short2long.clear()
        self.repeat = {}

        for option in self.option_table:
            if len(option) == 3:
                long, short, help = option
                repeat = 0
            elif len(option) == 4:
                long, short, help, repeat = option
            else:
                # the option table is part of the code, so simply
                # assert that it is correct
                raise ValueError("invalid option tuple: %r" % (option,))

            # Type- and value-check the option names
            if not isinstance(long, str) or len(long) < 2:
                raise DistutilsGetoptError(("invalid long option '%s': "
                       "must be a string of length >= 2") % long)

            if (not ((short is None) or
                     (isinstance(short, str) and len(short) == 1))):
                raise DistutilsGetoptError("invalid short option '%s': "
                       "must a single character or None" % short)

            self.repeat[long] = repeat
            self.long_opts.append(long)

            if long[-1] == '=':             # option takes an argument?
                if short: short = short + ':'
                long = long[0:-1]
                self.takes_arg[long] = 1
            else:
                # Is option is a "negative alias" for some other option (eg.
                # "quiet" == "!verbose")?
                alias_to = self.negative_alias.get(long)
                if alias_to is not None:
                    if self.takes_arg[alias_to]:
                        raise DistutilsGetoptError(
                              "invalid negative alias '%s': "
                              "aliased option '%s' takes a value"
                              % (long, alias_to))

                    self.long_opts[-1] = long # XXX redundant?!
                self.takes_arg[long] = 0

            # If this is an alias option, make sure its "takes arg" flag is
            # the same as the option it's aliased to.
            alias_to = self.alias.get(long)
            if alias_to is not None:
                if self.takes_arg[long] != self.takes_arg[alias_to]:
                    raise DistutilsGetoptError(
                          "invalid alias '%s': inconsistent with "
                          "aliased option '%s' (one of them takes a value, "
                          "the other doesn't"
                          % (long, alias_to))

            # Now enforce some bondage on the long option name, so we can
            # later translate it to an attribute name on some object.  Have
            # to do this a bit late to make sure we've removed any trailing
            # '='.
            if not longopt_re.match(long):
                raise DistutilsGetoptError(
                       "invalid long option name '%s' "
                       "(must be letters, numbers, hyphens only" % long)

            self.attr_name[long] = self.get_attr_name(long)
            if short:
                self.short_opts.append(short)
                self.short2long[short[0]] = long

    def getopt(self, args=None, object=None):
        """Parse command-line options in args. Store as attributes on object.

        If 'args' is None or not supplied, uses 'sys.argv[1:]'.  If
        'object' is None or not supplied, creates a new OptionDummy
        object, stores option values there, and returns a tuple (args,
        object).  If 'object' is supplied, it is modified in place and
        'getopt()' just returns 'args'; in both cases, the returned
        'args' is a modified copy of the passed-in 'args' list, which
        is left untouched.
        """
        if args is None:
            args = sys.argv[1:]
        if object is None:
            object = OptionDummy()
            created_object = True
        else:
            created_object = False

        self._grok_option_table()

        short_opts = ' '.join(self.short_opts)
        try:
            opts, args = getopt.getopt(args, short_opts, self.long_opts)
        except getopt.error as msg:
            raise DistutilsArgError(msg)

        for opt, val in opts:
            if len(opt) == 2 and opt[0] == '-': # it's a short option
                opt = self.short2long[opt[1]]
            else:
                assert len(opt) > 2 and opt[:2] == '--'
                opt = opt[2:]

            alias = self.alias.get(opt)
            if alias:
                opt = alias

            if not self.takes_arg[opt]:     # boolean option?
                assert val == '', "boolean option can't have value"
                alias = self.negative_alias.get(opt)
                if alias:
                    opt = alias
                    val = 0
                else:
                    val = 1

            attr = self.attr_name[opt]
            # The only repeating option at the moment is 'verbose'.
            # It has a negative option -q quiet, which should set verbose = 0.
            if val and self.repeat.get(attr) is not None:
                val = getattr(object, attr, 0) + 1
            setattr(object, attr, val)
            self.option_order.append((opt, val))

        # for opts
        if created_object:
            return args, object
        else:
            return args

    def get_option_order(self):
        """Returns the list of (option, value) tuples processed by the
        previous run of 'getopt()'.  Raises RuntimeError if
        'getopt()' hasn't been called yet.
        """
        if self.option_order is None:
            raise RuntimeError("'getopt()' hasn't been called yet")
        else:
            return self.option_order

    def generate_help(self, header=None):
        """Generate help text (a list of strings, one per suggested line of
        output) from the option table for this FancyGetopt object.
        """
        # Blithely assume the option table is good: probably wouldn't call
        # 'generate_help()' unless you've already called 'getopt()'.

        # First pass: determine maximum length of long option names
        max_opt = 0
        for option in self.option_table:
            long = option[0]
            short = option[1]
            l = len(long)
            if long[-1] == '=':
                l = l - 1
            if short is not None:
                l = l + 5                   # " (-x)" where short == 'x'
            if l > max_opt:
                max_opt = l

        opt_width = max_opt + 2 + 2 + 2     # room for indent + dashes + gutter

        # Typical help block looks like this:
        #   --foo       controls foonabulation
        # Help block for longest option looks like this:
        #   --flimflam  set the flim-flam level
        # and with wrapped text:
        #   --flimflam  set the flim-flam level (must be between
        #               0 and 100, except on Tuesdays)
        # Options with short names will have the short name shown (but
        # it doesn't contribute to max_opt):
        #   --foo (-f)  controls foonabulation
        # If adding the short option would make the left column too wide,
        # we push the explanation off to the next line
        #   --flimflam (-l)
        #               set the flim-flam level
        # Important parameters:
        #   - 2 spaces before option block start lines
        #   - 2 dashes for each long option name
        #   - min. 2 spaces between option and explanation (gutter)
        #   - 5 characters (incl. space) for short option name

        # Now generate lines of help text.  (If 80 columns were good enough
        # for Jesus, then 78 columns are good enough for me!)
        line_width = 78
        text_width = line_width - opt_width
        big_indent = ' ' * opt_width
        if header:
            lines = [header]
        else:
            lines = ['Option summary:']

        for option in self.option_table:
            long, short, help = option[:3]
            text = wrap_text(help, text_width)
            if long[-1] == '=':
                long = long[0:-1]

            # Case 1: no short option at all (makes life easy)
            if short is None:
                if text:
                    lines.append("  --%-*s  %s" % (max_opt, long, text[0]))
                else:
                    lines.append("  --%-*s  " % (max_opt, long))

            # Case 2: we have a short option, so we have to include it
            # just after the long option
            else:
                opt_names = "%s (-%s)" % (long, short)
                if text:
                    lines.append("  --%-*s  %s" %
                                 (max_opt, opt_names, text[0]))
                else:
                    lines.append("  --%-*s" % opt_names)

            for l in text[1:]:
                lines.append(big_indent + l)
        return lines

    def print_help(self, header=None, file=None):
        if file is None:
            file = sys.stdout
        for line in self.generate_help(header):
            file.write(line + "\n")


def fancy_getopt(options, negative_opt, object, args):
    parser = FancyGetopt(options)
    parser.set_negative_aliases(negative_opt)
    return parser.getopt(args, object)


WS_TRANS = {ord(_wschar) : ' ' for _wschar in string.whitespace}

def wrap_text(text, width):
    """wrap_text(text : string, width : int) -> [string]

    Split 'text' into multiple lines of no more than 'width' characters
    each, and return the list of strings that results.
    """
    if text is None:
        return []
    if len(text) <= width:
        return [text]

    text = text.expandtabs()
    text = text.translate(WS_TRANS)
    chunks = re.split(r'( +|-+)', text)
    chunks = [ch for ch in chunks if ch] # ' - ' results in empty strings
    lines = []

    while chunks:
        cur_line = []                   # list of chunks (to-be-joined)
        cur_len = 0                     # length of current line

        while chunks:
            l = len(chunks[0])
            if cur_len + l <= width:    # can squeeze (at least) this chunk in
                cur_line.append(chunks[0])
                del chunks[0]
                cur_len = cur_len + l
            else:                       # this line is full
                # drop last chunk if all space
                if cur_line and cur_line[-1][0] == ' ':
                    del cur_line[-1]
                break

        if chunks:                      # any chunks left to process?
            # if the current line is still empty, then we had a single
            # chunk that's too big too fit on a line -- so we break
            # down and break it up at the line width
            if cur_len == 0:
                cur_line.append(chunks[0][0:width])
                chunks[0] = chunks[0][width:]

            # all-whitespace chunks at the end of a line can be discarded
            # (and we know from the re.split above that if a chunk has
            # *any* whitespace, it is *all* whitespace)
            if chunks[0][0] == ' ':
                del chunks[0]

        # and store this line in the list-of-all-lines -- as a single
        # string, of course!
        lines.append(''.join(cur_line))

    return lines


def translate_longopt(opt):
    """Convert a long option name to a valid Python identifier by
    changing "-" to "_".
    """
    return opt.translate(longopt_xlate)


class OptionDummy:
    """Dummy class just used as a place to hold command-line option
    values as instance attributes."""

    def __init__(self, options=[]):
        """Create a new OptionDummy instance.  The attributes listed in
        'options' will be initialized to None."""
        for opt in options:
            setattr(self, opt, None)


if __name__ == "__main__":
    text = """\
Tra-la-la, supercalifragilisticexpialidocious.
How *do* you spell that odd word, anyways?
(Someone ask Mary -- she'll know [or she'll
say, "How should I know?"].)"""

    for w in (10, 20, 30, 40):
        print("width: %d" % w)
        print("\n".join(wrap_text(text, w)))
        print()
site-packages/setuptools/_distutils/ccompiler.py000064400000134767150564745010016250 0ustar00"""distutils.ccompiler

Contains CCompiler, an abstract base class that defines the interface
for the Distutils compiler abstraction model."""

import sys, os, re
from distutils.errors import *
from distutils.spawn import spawn
from distutils.file_util import move_file
from distutils.dir_util import mkpath
from distutils.dep_util import newer_group
from distutils.util import split_quoted, execute
from distutils import log

class CCompiler:
    """Abstract base class to define the interface that must be implemented
    by real compiler classes.  Also has some utility methods used by
    several compiler classes.

    The basic idea behind a compiler abstraction class is that each
    instance can be used for all the compile/link steps in building a
    single project.  Thus, attributes common to all of those compile and
    link steps -- include directories, macros to define, libraries to link
    against, etc. -- are attributes of the compiler instance.  To allow for
    variability in how individual files are treated, most of those
    attributes may be varied on a per-compilation or per-link basis.
    """

    # 'compiler_type' is a class attribute that identifies this class.  It
    # keeps code that wants to know what kind of compiler it's dealing with
    # from having to import all possible compiler classes just to do an
    # 'isinstance'.  In concrete CCompiler subclasses, 'compiler_type'
    # should really, really be one of the keys of the 'compiler_class'
    # dictionary (see below -- used by the 'new_compiler()' factory
    # function) -- authors of new compiler interface classes are
    # responsible for updating 'compiler_class'!
    compiler_type = None

    # XXX things not handled by this compiler abstraction model:
    #   * client can't provide additional options for a compiler,
    #     e.g. warning, optimization, debugging flags.  Perhaps this
    #     should be the domain of concrete compiler abstraction classes
    #     (UnixCCompiler, MSVCCompiler, etc.) -- or perhaps the base
    #     class should have methods for the common ones.
    #   * can't completely override the include or library searchg
    #     path, ie. no "cc -I -Idir1 -Idir2" or "cc -L -Ldir1 -Ldir2".
    #     I'm not sure how widely supported this is even by Unix
    #     compilers, much less on other platforms.  And I'm even less
    #     sure how useful it is; maybe for cross-compiling, but
    #     support for that is a ways off.  (And anyways, cross
    #     compilers probably have a dedicated binary with the
    #     right paths compiled in.  I hope.)
    #   * can't do really freaky things with the library list/library
    #     dirs, e.g. "-Ldir1 -lfoo -Ldir2 -lfoo" to link against
    #     different versions of libfoo.a in different locations.  I
    #     think this is useless without the ability to null out the
    #     library search path anyways.


    # Subclasses that rely on the standard filename generation methods
    # implemented below should override these; see the comment near
    # those methods ('object_filenames()' et. al.) for details:
    src_extensions = None               # list of strings
    obj_extension = None                # string
    static_lib_extension = None
    shared_lib_extension = None         # string
    static_lib_format = None            # format string
    shared_lib_format = None            # prob. same as static_lib_format
    exe_extension = None                # string

    # Default language settings. language_map is used to detect a source
    # file or Extension target language, checking source filenames.
    # language_order is used to detect the language precedence, when deciding
    # what language to use when mixing source types. For example, if some
    # extension has two files with ".c" extension, and one with ".cpp", it
    # is still linked as c++.
    language_map = {".c"   : "c",
                    ".cc"  : "c++",
                    ".cpp" : "c++",
                    ".cxx" : "c++",
                    ".m"   : "objc",
                   }
    language_order = ["c++", "objc", "c"]

    def __init__(self, verbose=0, dry_run=0, force=0):
        self.dry_run = dry_run
        self.force = force
        self.verbose = verbose

        # 'output_dir': a common output directory for object, library,
        # shared object, and shared library files
        self.output_dir = None

        # 'macros': a list of macro definitions (or undefinitions).  A
        # macro definition is a 2-tuple (name, value), where the value is
        # either a string or None (no explicit value).  A macro
        # undefinition is a 1-tuple (name,).
        self.macros = []

        # 'include_dirs': a list of directories to search for include files
        self.include_dirs = []

        # 'libraries': a list of libraries to include in any link
        # (library names, not filenames: eg. "foo" not "libfoo.a")
        self.libraries = []

        # 'library_dirs': a list of directories to search for libraries
        self.library_dirs = []

        # 'runtime_library_dirs': a list of directories to search for
        # shared libraries/objects at runtime
        self.runtime_library_dirs = []

        # 'objects': a list of object files (or similar, such as explicitly
        # named library files) to include on any link
        self.objects = []

        for key in self.executables.keys():
            self.set_executable(key, self.executables[key])

    def set_executables(self, **kwargs):
        """Define the executables (and options for them) that will be run
        to perform the various stages of compilation.  The exact set of
        executables that may be specified here depends on the compiler
        class (via the 'executables' class attribute), but most will have:
          compiler      the C/C++ compiler
          linker_so     linker used to create shared objects and libraries
          linker_exe    linker used to create binary executables
          archiver      static library creator

        On platforms with a command-line (Unix, DOS/Windows), each of these
        is a string that will be split into executable name and (optional)
        list of arguments.  (Splitting the string is done similarly to how
        Unix shells operate: words are delimited by spaces, but quotes and
        backslashes can override this.  See
        'distutils.util.split_quoted()'.)
        """

        # Note that some CCompiler implementation classes will define class
        # attributes 'cpp', 'cc', etc. with hard-coded executable names;
        # this is appropriate when a compiler class is for exactly one
        # compiler/OS combination (eg. MSVCCompiler).  Other compiler
        # classes (UnixCCompiler, in particular) are driven by information
        # discovered at run-time, since there are many different ways to do
        # basically the same things with Unix C compilers.

        for key in kwargs:
            if key not in self.executables:
                raise ValueError("unknown executable '%s' for class %s" %
                      (key, self.__class__.__name__))
            self.set_executable(key, kwargs[key])

    def set_executable(self, key, value):
        if isinstance(value, str):
            setattr(self, key, split_quoted(value))
        else:
            setattr(self, key, value)

    def _find_macro(self, name):
        i = 0
        for defn in self.macros:
            if defn[0] == name:
                return i
            i += 1
        return None

    def _check_macro_definitions(self, definitions):
        """Ensures that every element of 'definitions' is a valid macro
        definition, ie. either (name,value) 2-tuple or a (name,) tuple.  Do
        nothing if all definitions are OK, raise TypeError otherwise.
        """
        for defn in definitions:
            if not (isinstance(defn, tuple) and
                    (len(defn) in (1, 2) and
                      (isinstance (defn[1], str) or defn[1] is None)) and
                    isinstance (defn[0], str)):
                raise TypeError(("invalid macro definition '%s': " % defn) + \
                      "must be tuple (string,), (string, string), or " + \
                      "(string, None)")


    # -- Bookkeeping methods -------------------------------------------

    def define_macro(self, name, value=None):
        """Define a preprocessor macro for all compilations driven by this
        compiler object.  The optional parameter 'value' should be a
        string; if it is not supplied, then the macro will be defined
        without an explicit value and the exact outcome depends on the
        compiler used (XXX true? does ANSI say anything about this?)
        """
        # Delete from the list of macro definitions/undefinitions if
        # already there (so that this one will take precedence).
        i = self._find_macro (name)
        if i is not None:
            del self.macros[i]

        self.macros.append((name, value))

    def undefine_macro(self, name):
        """Undefine a preprocessor macro for all compilations driven by
        this compiler object.  If the same macro is defined by
        'define_macro()' and undefined by 'undefine_macro()' the last call
        takes precedence (including multiple redefinitions or
        undefinitions).  If the macro is redefined/undefined on a
        per-compilation basis (ie. in the call to 'compile()'), then that
        takes precedence.
        """
        # Delete from the list of macro definitions/undefinitions if
        # already there (so that this one will take precedence).
        i = self._find_macro (name)
        if i is not None:
            del self.macros[i]

        undefn = (name,)
        self.macros.append(undefn)

    def add_include_dir(self, dir):
        """Add 'dir' to the list of directories that will be searched for
        header files.  The compiler is instructed to search directories in
        the order in which they are supplied by successive calls to
        'add_include_dir()'.
        """
        self.include_dirs.append(dir)

    def set_include_dirs(self, dirs):
        """Set the list of directories that will be searched to 'dirs' (a
        list of strings).  Overrides any preceding calls to
        'add_include_dir()'; subsequence calls to 'add_include_dir()' add
        to the list passed to 'set_include_dirs()'.  This does not affect
        any list of standard include directories that the compiler may
        search by default.
        """
        self.include_dirs = dirs[:]

    def add_library(self, libname):
        """Add 'libname' to the list of libraries that will be included in
        all links driven by this compiler object.  Note that 'libname'
        should *not* be the name of a file containing a library, but the
        name of the library itself: the actual filename will be inferred by
        the linker, the compiler, or the compiler class (depending on the
        platform).

        The linker will be instructed to link against libraries in the
        order they were supplied to 'add_library()' and/or
        'set_libraries()'.  It is perfectly valid to duplicate library
        names; the linker will be instructed to link against libraries as
        many times as they are mentioned.
        """
        self.libraries.append(libname)

    def set_libraries(self, libnames):
        """Set the list of libraries to be included in all links driven by
        this compiler object to 'libnames' (a list of strings).  This does
        not affect any standard system libraries that the linker may
        include by default.
        """
        self.libraries = libnames[:]

    def add_library_dir(self, dir):
        """Add 'dir' to the list of directories that will be searched for
        libraries specified to 'add_library()' and 'set_libraries()'.  The
        linker will be instructed to search for libraries in the order they
        are supplied to 'add_library_dir()' and/or 'set_library_dirs()'.
        """
        self.library_dirs.append(dir)

    def set_library_dirs(self, dirs):
        """Set the list of library search directories to 'dirs' (a list of
        strings).  This does not affect any standard library search path
        that the linker may search by default.
        """
        self.library_dirs = dirs[:]

    def add_runtime_library_dir(self, dir):
        """Add 'dir' to the list of directories that will be searched for
        shared libraries at runtime.
        """
        self.runtime_library_dirs.append(dir)

    def set_runtime_library_dirs(self, dirs):
        """Set the list of directories to search for shared libraries at
        runtime to 'dirs' (a list of strings).  This does not affect any
        standard search path that the runtime linker may search by
        default.
        """
        self.runtime_library_dirs = dirs[:]

    def add_link_object(self, object):
        """Add 'object' to the list of object files (or analogues, such as
        explicitly named library files or the output of "resource
        compilers") to be included in every link driven by this compiler
        object.
        """
        self.objects.append(object)

    def set_link_objects(self, objects):
        """Set the list of object files (or analogues) to be included in
        every link to 'objects'.  This does not affect any standard object
        files that the linker may include by default (such as system
        libraries).
        """
        self.objects = objects[:]


    # -- Private utility methods --------------------------------------
    # (here for the convenience of subclasses)

    # Helper method to prep compiler in subclass compile() methods

    def _setup_compile(self, outdir, macros, incdirs, sources, depends,
                       extra):
        """Process arguments and decide which source files to compile."""
        if outdir is None:
            outdir = self.output_dir
        elif not isinstance(outdir, str):
            raise TypeError("'output_dir' must be a string or None")

        if macros is None:
            macros = self.macros
        elif isinstance(macros, list):
            macros = macros + (self.macros or [])
        else:
            raise TypeError("'macros' (if supplied) must be a list of tuples")

        if incdirs is None:
            incdirs = self.include_dirs
        elif isinstance(incdirs, (list, tuple)):
            incdirs = list(incdirs) + (self.include_dirs or [])
        else:
            raise TypeError(
                  "'include_dirs' (if supplied) must be a list of strings")

        if extra is None:
            extra = []

        # Get the list of expected output (object) files
        objects = self.object_filenames(sources, strip_dir=0,
                                        output_dir=outdir)
        assert len(objects) == len(sources)

        pp_opts = gen_preprocess_options(macros, incdirs)

        build = {}
        for i in range(len(sources)):
            src = sources[i]
            obj = objects[i]
            ext = os.path.splitext(src)[1]
            self.mkpath(os.path.dirname(obj))
            build[obj] = (src, ext)

        return macros, objects, extra, pp_opts, build

    def _get_cc_args(self, pp_opts, debug, before):
        # works for unixccompiler, cygwinccompiler
        cc_args = pp_opts + ['-c']
        if debug:
            cc_args[:0] = ['-g']
        if before:
            cc_args[:0] = before
        return cc_args

    def _fix_compile_args(self, output_dir, macros, include_dirs):
        """Typecheck and fix-up some of the arguments to the 'compile()'
        method, and return fixed-up values.  Specifically: if 'output_dir'
        is None, replaces it with 'self.output_dir'; ensures that 'macros'
        is a list, and augments it with 'self.macros'; ensures that
        'include_dirs' is a list, and augments it with 'self.include_dirs'.
        Guarantees that the returned values are of the correct type,
        i.e. for 'output_dir' either string or None, and for 'macros' and
        'include_dirs' either list or None.
        """
        if output_dir is None:
            output_dir = self.output_dir
        elif not isinstance(output_dir, str):
            raise TypeError("'output_dir' must be a string or None")

        if macros is None:
            macros = self.macros
        elif isinstance(macros, list):
            macros = macros + (self.macros or [])
        else:
            raise TypeError("'macros' (if supplied) must be a list of tuples")

        if include_dirs is None:
            include_dirs = self.include_dirs
        elif isinstance(include_dirs, (list, tuple)):
            include_dirs = list(include_dirs) + (self.include_dirs or [])
        else:
            raise TypeError(
                  "'include_dirs' (if supplied) must be a list of strings")

        return output_dir, macros, include_dirs

    def _prep_compile(self, sources, output_dir, depends=None):
        """Decide which source files must be recompiled.

        Determine the list of object files corresponding to 'sources',
        and figure out which ones really need to be recompiled.
        Return a list of all object files and a dictionary telling
        which source files can be skipped.
        """
        # Get the list of expected output (object) files
        objects = self.object_filenames(sources, output_dir=output_dir)
        assert len(objects) == len(sources)

        # Return an empty dict for the "which source files can be skipped"
        # return value to preserve API compatibility.
        return objects, {}

    def _fix_object_args(self, objects, output_dir):
        """Typecheck and fix up some arguments supplied to various methods.
        Specifically: ensure that 'objects' is a list; if output_dir is
        None, replace with self.output_dir.  Return fixed versions of
        'objects' and 'output_dir'.
        """
        if not isinstance(objects, (list, tuple)):
            raise TypeError("'objects' must be a list or tuple of strings")
        objects = list(objects)

        if output_dir is None:
            output_dir = self.output_dir
        elif not isinstance(output_dir, str):
            raise TypeError("'output_dir' must be a string or None")

        return (objects, output_dir)

    def _fix_lib_args(self, libraries, library_dirs, runtime_library_dirs):
        """Typecheck and fix up some of the arguments supplied to the
        'link_*' methods.  Specifically: ensure that all arguments are
        lists, and augment them with their permanent versions
        (eg. 'self.libraries' augments 'libraries').  Return a tuple with
        fixed versions of all arguments.
        """
        if libraries is None:
            libraries = self.libraries
        elif isinstance(libraries, (list, tuple)):
            libraries = list (libraries) + (self.libraries or [])
        else:
            raise TypeError(
                  "'libraries' (if supplied) must be a list of strings")

        if library_dirs is None:
            library_dirs = self.library_dirs
        elif isinstance(library_dirs, (list, tuple)):
            library_dirs = list (library_dirs) + (self.library_dirs or [])
        else:
            raise TypeError(
                  "'library_dirs' (if supplied) must be a list of strings")

        if runtime_library_dirs is None:
            runtime_library_dirs = self.runtime_library_dirs
        elif isinstance(runtime_library_dirs, (list, tuple)):
            runtime_library_dirs = (list(runtime_library_dirs) +
                                    (self.runtime_library_dirs or []))
        else:
            raise TypeError("'runtime_library_dirs' (if supplied) "
                            "must be a list of strings")

        return (libraries, library_dirs, runtime_library_dirs)

    def _need_link(self, objects, output_file):
        """Return true if we need to relink the files listed in 'objects'
        to recreate 'output_file'.
        """
        if self.force:
            return True
        else:
            if self.dry_run:
                newer = newer_group (objects, output_file, missing='newer')
            else:
                newer = newer_group (objects, output_file)
            return newer

    def detect_language(self, sources):
        """Detect the language of a given file, or list of files. Uses
        language_map, and language_order to do the job.
        """
        if not isinstance(sources, list):
            sources = [sources]
        lang = None
        index = len(self.language_order)
        for source in sources:
            base, ext = os.path.splitext(source)
            extlang = self.language_map.get(ext)
            try:
                extindex = self.language_order.index(extlang)
                if extindex < index:
                    lang = extlang
                    index = extindex
            except ValueError:
                pass
        return lang


    # -- Worker methods ------------------------------------------------
    # (must be implemented by subclasses)

    def preprocess(self, source, output_file=None, macros=None,
                   include_dirs=None, extra_preargs=None, extra_postargs=None):
        """Preprocess a single C/C++ source file, named in 'source'.
        Output will be written to file named 'output_file', or stdout if
        'output_file' not supplied.  'macros' is a list of macro
        definitions as for 'compile()', which will augment the macros set
        with 'define_macro()' and 'undefine_macro()'.  'include_dirs' is a
        list of directory names that will be added to the default list.

        Raises PreprocessError on failure.
        """
        pass

    def compile(self, sources, output_dir=None, macros=None,
                include_dirs=None, debug=0, extra_preargs=None,
                extra_postargs=None, depends=None):
        """Compile one or more source files.

        'sources' must be a list of filenames, most likely C/C++
        files, but in reality anything that can be handled by a
        particular compiler and compiler class (eg. MSVCCompiler can
        handle resource files in 'sources').  Return a list of object
        filenames, one per source filename in 'sources'.  Depending on
        the implementation, not all source files will necessarily be
        compiled, but all corresponding object filenames will be
        returned.

        If 'output_dir' is given, object files will be put under it, while
        retaining their original path component.  That is, "foo/bar.c"
        normally compiles to "foo/bar.o" (for a Unix implementation); if
        'output_dir' is "build", then it would compile to
        "build/foo/bar.o".

        'macros', if given, must be a list of macro definitions.  A macro
        definition is either a (name, value) 2-tuple or a (name,) 1-tuple.
        The former defines a macro; if the value is None, the macro is
        defined without an explicit value.  The 1-tuple case undefines a
        macro.  Later definitions/redefinitions/ undefinitions take
        precedence.

        'include_dirs', if given, must be a list of strings, the
        directories to add to the default include file search path for this
        compilation only.

        'debug' is a boolean; if true, the compiler will be instructed to
        output debug symbols in (or alongside) the object file(s).

        'extra_preargs' and 'extra_postargs' are implementation- dependent.
        On platforms that have the notion of a command-line (e.g. Unix,
        DOS/Windows), they are most likely lists of strings: extra
        command-line arguments to prepend/append to the compiler command
        line.  On other platforms, consult the implementation class
        documentation.  In any event, they are intended as an escape hatch
        for those occasions when the abstract compiler framework doesn't
        cut the mustard.

        'depends', if given, is a list of filenames that all targets
        depend on.  If a source file is older than any file in
        depends, then the source file will be recompiled.  This
        supports dependency tracking, but only at a coarse
        granularity.

        Raises CompileError on failure.
        """
        # A concrete compiler class can either override this method
        # entirely or implement _compile().
        macros, objects, extra_postargs, pp_opts, build = \
                self._setup_compile(output_dir, macros, include_dirs, sources,
                                    depends, extra_postargs)
        cc_args = self._get_cc_args(pp_opts, debug, extra_preargs)

        for obj in objects:
            try:
                src, ext = build[obj]
            except KeyError:
                continue
            self._compile(obj, src, ext, cc_args, extra_postargs, pp_opts)

        # Return *all* object filenames, not just the ones we just built.
        return objects

    def _compile(self, obj, src, ext, cc_args, extra_postargs, pp_opts):
        """Compile 'src' to product 'obj'."""
        # A concrete compiler class that does not override compile()
        # should implement _compile().
        pass

    def create_static_lib(self, objects, output_libname, output_dir=None,
                          debug=0, target_lang=None):
        """Link a bunch of stuff together to create a static library file.
        The "bunch of stuff" consists of the list of object files supplied
        as 'objects', the extra object files supplied to
        'add_link_object()' and/or 'set_link_objects()', the libraries
        supplied to 'add_library()' and/or 'set_libraries()', and the
        libraries supplied as 'libraries' (if any).

        'output_libname' should be a library name, not a filename; the
        filename will be inferred from the library name.  'output_dir' is
        the directory where the library file will be put.

        'debug' is a boolean; if true, debugging information will be
        included in the library (note that on most platforms, it is the
        compile step where this matters: the 'debug' flag is included here
        just for consistency).

        'target_lang' is the target language for which the given objects
        are being compiled. This allows specific linkage time treatment of
        certain languages.

        Raises LibError on failure.
        """
        pass


    # values for target_desc parameter in link()
    SHARED_OBJECT = "shared_object"
    SHARED_LIBRARY = "shared_library"
    EXECUTABLE = "executable"

    def link(self,
             target_desc,
             objects,
             output_filename,
             output_dir=None,
             libraries=None,
             library_dirs=None,
             runtime_library_dirs=None,
             export_symbols=None,
             debug=0,
             extra_preargs=None,
             extra_postargs=None,
             build_temp=None,
             target_lang=None):
        """Link a bunch of stuff together to create an executable or
        shared library file.

        The "bunch of stuff" consists of the list of object files supplied
        as 'objects'.  'output_filename' should be a filename.  If
        'output_dir' is supplied, 'output_filename' is relative to it
        (i.e. 'output_filename' can provide directory components if
        needed).

        'libraries' is a list of libraries to link against.  These are
        library names, not filenames, since they're translated into
        filenames in a platform-specific way (eg. "foo" becomes "libfoo.a"
        on Unix and "foo.lib" on DOS/Windows).  However, they can include a
        directory component, which means the linker will look in that
        specific directory rather than searching all the normal locations.

        'library_dirs', if supplied, should be a list of directories to
        search for libraries that were specified as bare library names
        (ie. no directory component).  These are on top of the system
        default and those supplied to 'add_library_dir()' and/or
        'set_library_dirs()'.  'runtime_library_dirs' is a list of
        directories that will be embedded into the shared library and used
        to search for other shared libraries that *it* depends on at
        run-time.  (This may only be relevant on Unix.)

        'export_symbols' is a list of symbols that the shared library will
        export.  (This appears to be relevant only on Windows.)

        'debug' is as for 'compile()' and 'create_static_lib()', with the
        slight distinction that it actually matters on most platforms (as
        opposed to 'create_static_lib()', which includes a 'debug' flag
        mostly for form's sake).

        'extra_preargs' and 'extra_postargs' are as for 'compile()' (except
        of course that they supply command-line arguments for the
        particular linker being used).

        'target_lang' is the target language for which the given objects
        are being compiled. This allows specific linkage time treatment of
        certain languages.

        Raises LinkError on failure.
        """
        raise NotImplementedError


    # Old 'link_*()' methods, rewritten to use the new 'link()' method.

    def link_shared_lib(self,
                        objects,
                        output_libname,
                        output_dir=None,
                        libraries=None,
                        library_dirs=None,
                        runtime_library_dirs=None,
                        export_symbols=None,
                        debug=0,
                        extra_preargs=None,
                        extra_postargs=None,
                        build_temp=None,
                        target_lang=None):
        self.link(CCompiler.SHARED_LIBRARY, objects,
                  self.library_filename(output_libname, lib_type='shared'),
                  output_dir,
                  libraries, library_dirs, runtime_library_dirs,
                  export_symbols, debug,
                  extra_preargs, extra_postargs, build_temp, target_lang)


    def link_shared_object(self,
                           objects,
                           output_filename,
                           output_dir=None,
                           libraries=None,
                           library_dirs=None,
                           runtime_library_dirs=None,
                           export_symbols=None,
                           debug=0,
                           extra_preargs=None,
                           extra_postargs=None,
                           build_temp=None,
                           target_lang=None):
        self.link(CCompiler.SHARED_OBJECT, objects,
                  output_filename, output_dir,
                  libraries, library_dirs, runtime_library_dirs,
                  export_symbols, debug,
                  extra_preargs, extra_postargs, build_temp, target_lang)


    def link_executable(self,
                        objects,
                        output_progname,
                        output_dir=None,
                        libraries=None,
                        library_dirs=None,
                        runtime_library_dirs=None,
                        debug=0,
                        extra_preargs=None,
                        extra_postargs=None,
                        target_lang=None):
        self.link(CCompiler.EXECUTABLE, objects,
                  self.executable_filename(output_progname), output_dir,
                  libraries, library_dirs, runtime_library_dirs, None,
                  debug, extra_preargs, extra_postargs, None, target_lang)


    # -- Miscellaneous methods -----------------------------------------
    # These are all used by the 'gen_lib_options() function; there is
    # no appropriate default implementation so subclasses should
    # implement all of these.

    def library_dir_option(self, dir):
        """Return the compiler option to add 'dir' to the list of
        directories searched for libraries.
        """
        raise NotImplementedError

    def runtime_library_dir_option(self, dir):
        """Return the compiler option to add 'dir' to the list of
        directories searched for runtime libraries.
        """
        raise NotImplementedError

    def library_option(self, lib):
        """Return the compiler option to add 'lib' to the list of libraries
        linked into the shared library or executable.
        """
        raise NotImplementedError

    def has_function(self, funcname, includes=None, include_dirs=None,
                     libraries=None, library_dirs=None):
        """Return a boolean indicating whether funcname is supported on
        the current platform.  The optional arguments can be used to
        augment the compilation environment.
        """
        # this can't be included at module scope because it tries to
        # import math which might not be available at that point - maybe
        # the necessary logic should just be inlined?
        import tempfile
        if includes is None:
            includes = []
        if include_dirs is None:
            include_dirs = []
        if libraries is None:
            libraries = []
        if library_dirs is None:
            library_dirs = []
        fd, fname = tempfile.mkstemp(".c", funcname, text=True)
        f = os.fdopen(fd, "w")
        try:
            for incl in includes:
                f.write("""#include "%s"\n""" % incl)
            f.write("""\
int main (int argc, char **argv) {
    %s();
    return 0;
}
""" % funcname)
        finally:
            f.close()
        try:
            objects = self.compile([fname], include_dirs=include_dirs)
        except CompileError:
            return False
        finally:
            os.remove(fname)

        try:
            self.link_executable(objects, "a.out",
                                 libraries=libraries,
                                 library_dirs=library_dirs)
        except (LinkError, TypeError):
            return False
        else:
            os.remove("a.out")
        finally:
            for fn in objects:
                os.remove(fn)
        return True

    def find_library_file (self, dirs, lib, debug=0):
        """Search the specified list of directories for a static or shared
        library file 'lib' and return the full path to that file.  If
        'debug' true, look for a debugging version (if that makes sense on
        the current platform).  Return None if 'lib' wasn't found in any of
        the specified directories.
        """
        raise NotImplementedError

    # -- Filename generation methods -----------------------------------

    # The default implementation of the filename generating methods are
    # prejudiced towards the Unix/DOS/Windows view of the world:
    #   * object files are named by replacing the source file extension
    #     (eg. .c/.cpp -> .o/.obj)
    #   * library files (shared or static) are named by plugging the
    #     library name and extension into a format string, eg.
    #     "lib%s.%s" % (lib_name, ".a") for Unix static libraries
    #   * executables are named by appending an extension (possibly
    #     empty) to the program name: eg. progname + ".exe" for
    #     Windows
    #
    # To reduce redundant code, these methods expect to find
    # several attributes in the current object (presumably defined
    # as class attributes):
    #   * src_extensions -
    #     list of C/C++ source file extensions, eg. ['.c', '.cpp']
    #   * obj_extension -
    #     object file extension, eg. '.o' or '.obj'
    #   * static_lib_extension -
    #     extension for static library files, eg. '.a' or '.lib'
    #   * shared_lib_extension -
    #     extension for shared library/object files, eg. '.so', '.dll'
    #   * static_lib_format -
    #     format string for generating static library filenames,
    #     eg. 'lib%s.%s' or '%s.%s'
    #   * shared_lib_format
    #     format string for generating shared library filenames
    #     (probably same as static_lib_format, since the extension
    #     is one of the intended parameters to the format string)
    #   * exe_extension -
    #     extension for executable files, eg. '' or '.exe'

    def object_filenames(self, source_filenames, strip_dir=0, output_dir=''):
        if output_dir is None:
            output_dir = ''
        obj_names = []
        for src_name in source_filenames:
            base, ext = os.path.splitext(src_name)
            base = os.path.splitdrive(base)[1] # Chop off the drive
            base = base[os.path.isabs(base):]  # If abs, chop off leading /
            if ext not in self.src_extensions:
                raise UnknownFileError(
                      "unknown file type '%s' (from '%s')" % (ext, src_name))
            if strip_dir:
                base = os.path.basename(base)
            obj_names.append(os.path.join(output_dir,
                                          base + self.obj_extension))
        return obj_names

    def shared_object_filename(self, basename, strip_dir=0, output_dir=''):
        assert output_dir is not None
        if strip_dir:
            basename = os.path.basename(basename)
        return os.path.join(output_dir, basename + self.shared_lib_extension)

    def executable_filename(self, basename, strip_dir=0, output_dir=''):
        assert output_dir is not None
        if strip_dir:
            basename = os.path.basename(basename)
        return os.path.join(output_dir, basename + (self.exe_extension or ''))

    def library_filename(self, libname, lib_type='static',     # or 'shared'
                         strip_dir=0, output_dir=''):
        assert output_dir is not None
        if lib_type not in ("static", "shared", "dylib", "xcode_stub"):
            raise ValueError(
                  "'lib_type' must be \"static\", \"shared\", \"dylib\", or \"xcode_stub\"")
        fmt = getattr(self, lib_type + "_lib_format")
        ext = getattr(self, lib_type + "_lib_extension")

        dir, base = os.path.split(libname)
        filename = fmt % (base, ext)
        if strip_dir:
            dir = ''

        return os.path.join(output_dir, dir, filename)


    # -- Utility methods -----------------------------------------------

    def announce(self, msg, level=1):
        log.debug(msg)

    def debug_print(self, msg):
        from distutils.debug import DEBUG
        if DEBUG:
            print(msg)

    def warn(self, msg):
        sys.stderr.write("warning: %s\n" % msg)

    def execute(self, func, args, msg=None, level=1):
        execute(func, args, msg, self.dry_run)

    def spawn(self, cmd, **kwargs):
        spawn(cmd, dry_run=self.dry_run, **kwargs)

    def move_file(self, src, dst):
        return move_file(src, dst, dry_run=self.dry_run)

    def mkpath (self, name, mode=0o777):
        mkpath(name, mode, dry_run=self.dry_run)


# Map a sys.platform/os.name ('posix', 'nt') to the default compiler
# type for that platform. Keys are interpreted as re match
# patterns. Order is important; platform mappings are preferred over
# OS names.
_default_compilers = (

    # Platform string mappings

    # on a cygwin built python we can use gcc like an ordinary UNIXish
    # compiler
    ('cygwin.*', 'unix'),

    # OS name mappings
    ('posix', 'unix'),
    ('nt', 'msvc'),

    )

def get_default_compiler(osname=None, platform=None):
    """Determine the default compiler to use for the given platform.

       osname should be one of the standard Python OS names (i.e. the
       ones returned by os.name) and platform the common value
       returned by sys.platform for the platform in question.

       The default values are os.name and sys.platform in case the
       parameters are not given.
    """
    if osname is None:
        osname = os.name
    if platform is None:
        platform = sys.platform
    for pattern, compiler in _default_compilers:
        if re.match(pattern, platform) is not None or \
           re.match(pattern, osname) is not None:
            return compiler
    # Default to Unix compiler
    return 'unix'

# Map compiler types to (module_name, class_name) pairs -- ie. where to
# find the code that implements an interface to this compiler.  (The module
# is assumed to be in the 'distutils' package.)
compiler_class = { 'unix':    ('unixccompiler', 'UnixCCompiler',
                               "standard UNIX-style compiler"),
                   'msvc':    ('_msvccompiler', 'MSVCCompiler',
                               "Microsoft Visual C++"),
                   'cygwin':  ('cygwinccompiler', 'CygwinCCompiler',
                               "Cygwin port of GNU C Compiler for Win32"),
                   'mingw32': ('cygwinccompiler', 'Mingw32CCompiler',
                               "Mingw32 port of GNU C Compiler for Win32"),
                   'bcpp':    ('bcppcompiler', 'BCPPCompiler',
                               "Borland C++ Compiler"),
                 }

def show_compilers():
    """Print list of available compilers (used by the "--help-compiler"
    options to "build", "build_ext", "build_clib").
    """
    # XXX this "knows" that the compiler option it's describing is
    # "--compiler", which just happens to be the case for the three
    # commands that use it.
    from distutils.fancy_getopt import FancyGetopt
    compilers = []
    for compiler in compiler_class.keys():
        compilers.append(("compiler="+compiler, None,
                          compiler_class[compiler][2]))
    compilers.sort()
    pretty_printer = FancyGetopt(compilers)
    pretty_printer.print_help("List of available compilers:")


def new_compiler(plat=None, compiler=None, verbose=0, dry_run=0, force=0):
    """Generate an instance of some CCompiler subclass for the supplied
    platform/compiler combination.  'plat' defaults to 'os.name'
    (eg. 'posix', 'nt'), and 'compiler' defaults to the default compiler
    for that platform.  Currently only 'posix' and 'nt' are supported, and
    the default compilers are "traditional Unix interface" (UnixCCompiler
    class) and Visual C++ (MSVCCompiler class).  Note that it's perfectly
    possible to ask for a Unix compiler object under Windows, and a
    Microsoft compiler object under Unix -- if you supply a value for
    'compiler', 'plat' is ignored.
    """
    if plat is None:
        plat = os.name

    try:
        if compiler is None:
            compiler = get_default_compiler(plat)

        (module_name, class_name, long_description) = compiler_class[compiler]
    except KeyError:
        msg = "don't know how to compile C/C++ code on platform '%s'" % plat
        if compiler is not None:
            msg = msg + " with '%s' compiler" % compiler
        raise DistutilsPlatformError(msg)

    try:
        module_name = "distutils." + module_name
        __import__ (module_name)
        module = sys.modules[module_name]
        klass = vars(module)[class_name]
    except ImportError:
        raise DistutilsModuleError(
              "can't compile C/C++ code: unable to load module '%s'" % \
              module_name)
    except KeyError:
        raise DistutilsModuleError(
               "can't compile C/C++ code: unable to find class '%s' "
               "in module '%s'" % (class_name, module_name))

    # XXX The None is necessary to preserve backwards compatibility
    # with classes that expect verbose to be the first positional
    # argument.
    return klass(None, dry_run, force)


def gen_preprocess_options(macros, include_dirs):
    """Generate C pre-processor options (-D, -U, -I) as used by at least
    two types of compilers: the typical Unix compiler and Visual C++.
    'macros' is the usual thing, a list of 1- or 2-tuples, where (name,)
    means undefine (-U) macro 'name', and (name,value) means define (-D)
    macro 'name' to 'value'.  'include_dirs' is just a list of directory
    names to be added to the header file search path (-I).  Returns a list
    of command-line options suitable for either Unix compilers or Visual
    C++.
    """
    # XXX it would be nice (mainly aesthetic, and so we don't generate
    # stupid-looking command lines) to go over 'macros' and eliminate
    # redundant definitions/undefinitions (ie. ensure that only the
    # latest mention of a particular macro winds up on the command
    # line).  I don't think it's essential, though, since most (all?)
    # Unix C compilers only pay attention to the latest -D or -U
    # mention of a macro on their command line.  Similar situation for
    # 'include_dirs'.  I'm punting on both for now.  Anyways, weeding out
    # redundancies like this should probably be the province of
    # CCompiler, since the data structures used are inherited from it
    # and therefore common to all CCompiler classes.
    pp_opts = []
    for macro in macros:
        if not (isinstance(macro, tuple) and 1 <= len(macro) <= 2):
            raise TypeError(
                  "bad macro definition '%s': "
                  "each element of 'macros' list must be a 1- or 2-tuple"
                  % macro)

        if len(macro) == 1:        # undefine this macro
            pp_opts.append("-U%s" % macro[0])
        elif len(macro) == 2:
            if macro[1] is None:    # define with no explicit value
                pp_opts.append("-D%s" % macro[0])
            else:
                # XXX *don't* need to be clever about quoting the
                # macro value here, because we're going to avoid the
                # shell at all costs when we spawn the command!
                pp_opts.append("-D%s=%s" % macro)

    for dir in include_dirs:
        pp_opts.append("-I%s" % dir)
    return pp_opts


def gen_lib_options (compiler, library_dirs, runtime_library_dirs, libraries):
    """Generate linker options for searching library directories and
    linking with specific libraries.  'libraries' and 'library_dirs' are,
    respectively, lists of library names (not filenames!) and search
    directories.  Returns a list of command-line options suitable for use
    with some compiler (depending on the two format strings passed in).
    """
    lib_opts = []

    for dir in library_dirs:
        lib_opts.append(compiler.library_dir_option(dir))

    for dir in runtime_library_dirs:
        opt = compiler.runtime_library_dir_option(dir)
        if isinstance(opt, list):
            lib_opts = lib_opts + opt
        else:
            lib_opts.append(opt)

    # XXX it's important that we *not* remove redundant library mentions!
    # sometimes you really do have to say "-lfoo -lbar -lfoo" in order to
    # resolve all symbols.  I just hope we never have to say "-lfoo obj.o
    # -lbar" to get things to work -- that's certainly a possibility, but a
    # pretty nasty way to arrange your C code.

    for lib in libraries:
        (lib_dir, lib_name) = os.path.split(lib)
        if lib_dir:
            lib_file = compiler.find_library_file([lib_dir], lib_name)
            if lib_file:
                lib_opts.append(lib_file)
            else:
                compiler.warn("no library file corresponding to "
                              "'%s' found (skipping)" % lib)
        else:
            lib_opts.append(compiler.library_option (lib))
    return lib_opts
site-packages/setuptools/_distutils/py38compat.py000064400000000324150564745010016257 0ustar00def aix_platform(osname, version, release):
    try:
        import _aix_support
        return _aix_support.aix_platform()
    except ImportError:
        pass
    return "%s-%s.%s" % (osname, version, release)
site-packages/setuptools/_distutils/spawn.py000064400000006652150564745010015412 0ustar00"""distutils.spawn

Provides the 'spawn()' function, a front-end to various platform-
specific functions for launching another program in a sub-process.
Also provides the 'find_executable()' to search the path for a given
executable name.
"""

import sys
import os
import subprocess

from distutils.errors import DistutilsPlatformError, DistutilsExecError
from distutils.debug import DEBUG
from distutils import log


def spawn(cmd, search_path=1, verbose=0, dry_run=0, env=None):
    """Run another program, specified as a command list 'cmd', in a new process.

    'cmd' is just the argument list for the new process, ie.
    cmd[0] is the program to run and cmd[1:] are the rest of its arguments.
    There is no way to run a program with a name different from that of its
    executable.

    If 'search_path' is true (the default), the system's executable
    search path will be used to find the program; otherwise, cmd[0]
    must be the exact path to the executable.  If 'dry_run' is true,
    the command will not actually be run.

    Raise DistutilsExecError if running the program fails in any way; just
    return on success.
    """
    # cmd is documented as a list, but just in case some code passes a tuple
    # in, protect our %-formatting code against horrible death
    cmd = list(cmd)

    log.info(subprocess.list2cmdline(cmd))
    if dry_run:
        return

    if search_path:
        executable = find_executable(cmd[0])
        if executable is not None:
            cmd[0] = executable

    env = env if env is not None else dict(os.environ)

    if sys.platform == 'darwin':
        from distutils.util import MACOSX_VERSION_VAR, get_macosx_target_ver
        macosx_target_ver = get_macosx_target_ver()
        if macosx_target_ver:
            env[MACOSX_VERSION_VAR] = macosx_target_ver

    try:
        proc = subprocess.Popen(cmd, env=env)
        proc.wait()
        exitcode = proc.returncode
    except OSError as exc:
        if not DEBUG:
            cmd = cmd[0]
        raise DistutilsExecError(
            "command %r failed: %s" % (cmd, exc.args[-1])) from exc

    if exitcode:
        if not DEBUG:
            cmd = cmd[0]
        raise DistutilsExecError(
              "command %r failed with exit code %s" % (cmd, exitcode))


def find_executable(executable, path=None):
    """Tries to find 'executable' in the directories listed in 'path'.

    A string listing directories separated by 'os.pathsep'; defaults to
    os.environ['PATH'].  Returns the complete filename or None if not found.
    """
    _, ext = os.path.splitext(executable)
    if (sys.platform == 'win32') and (ext != '.exe'):
        executable = executable + '.exe'

    if os.path.isfile(executable):
        return executable

    if path is None:
        path = os.environ.get('PATH', None)
        if path is None:
            try:
                path = os.confstr("CS_PATH")
            except (AttributeError, ValueError):
                # os.confstr() or CS_PATH is not available
                path = os.defpath
        # bpo-35755: Don't use os.defpath if the PATH environment variable is
        # set to an empty string

    # PATH='' doesn't match, whereas PATH=':' looks in the current directory
    if not path:
        return None

    paths = path.split(os.pathsep)
    for p in paths:
        f = os.path.join(p, executable)
        if os.path.isfile(f):
            # the file exists, we have a shot at spawn working
            return f
    return None
site-packages/setuptools/_distutils/debug.py000064400000000213150564745010015333 0ustar00import os

# If DISTUTILS_DEBUG is anything other than the empty string, we run in
# debug mode.
DEBUG = os.environ.get('DISTUTILS_DEBUG')
site-packages/setuptools/_distutils/_msvccompiler.py000064400000050515150564745010017121 0ustar00"""distutils._msvccompiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for Microsoft Visual Studio 2015.

The module is compatible with VS 2015 and later. You can find legacy support
for older versions in distutils.msvc9compiler and distutils.msvccompiler.
"""

# Written by Perry Stoll
# hacked by Robin Becker and Thomas Heller to do a better job of
#   finding DevStudio (through the registry)
# ported to VS 2005 and VS 2008 by Christian Heimes
# ported to VS 2015 by Steve Dower

import os
import subprocess
import contextlib
import warnings
import unittest.mock
with contextlib.suppress(ImportError):
    import winreg

from distutils.errors import DistutilsExecError, DistutilsPlatformError, \
                             CompileError, LibError, LinkError
from distutils.ccompiler import CCompiler, gen_lib_options
from distutils import log
from distutils.util import get_platform

from itertools import count

def _find_vc2015():
    try:
        key = winreg.OpenKeyEx(
            winreg.HKEY_LOCAL_MACHINE,
            r"Software\Microsoft\VisualStudio\SxS\VC7",
            access=winreg.KEY_READ | winreg.KEY_WOW64_32KEY
        )
    except OSError:
        log.debug("Visual C++ is not registered")
        return None, None

    best_version = 0
    best_dir = None
    with key:
        for i in count():
            try:
                v, vc_dir, vt = winreg.EnumValue(key, i)
            except OSError:
                break
            if v and vt == winreg.REG_SZ and os.path.isdir(vc_dir):
                try:
                    version = int(float(v))
                except (ValueError, TypeError):
                    continue
                if version >= 14 and version > best_version:
                    best_version, best_dir = version, vc_dir
    return best_version, best_dir

def _find_vc2017():
    """Returns "15, path" based on the result of invoking vswhere.exe
    If no install is found, returns "None, None"

    The version is returned to avoid unnecessarily changing the function
    result. It may be ignored when the path is not None.

    If vswhere.exe is not available, by definition, VS 2017 is not
    installed.
    """
    root = os.environ.get("ProgramFiles(x86)") or os.environ.get("ProgramFiles")
    if not root:
        return None, None

    try:
        path = subprocess.check_output([
            os.path.join(root, "Microsoft Visual Studio", "Installer", "vswhere.exe"),
            "-latest",
            "-prerelease",
            "-requires", "Microsoft.VisualStudio.Component.VC.Tools.x86.x64",
            "-property", "installationPath",
            "-products", "*",
        ], encoding="mbcs", errors="strict").strip()
    except (subprocess.CalledProcessError, OSError, UnicodeDecodeError):
        return None, None

    path = os.path.join(path, "VC", "Auxiliary", "Build")
    if os.path.isdir(path):
        return 15, path

    return None, None

PLAT_SPEC_TO_RUNTIME = {
    'x86' : 'x86',
    'x86_amd64' : 'x64',
    'x86_arm' : 'arm',
    'x86_arm64' : 'arm64'
}

def _find_vcvarsall(plat_spec):
    # bpo-38597: Removed vcruntime return value
    _, best_dir = _find_vc2017()

    if not best_dir:
        best_version, best_dir = _find_vc2015()

    if not best_dir:
        log.debug("No suitable Visual C++ version found")
        return None, None

    vcvarsall = os.path.join(best_dir, "vcvarsall.bat")
    if not os.path.isfile(vcvarsall):
        log.debug("%s cannot be found", vcvarsall)
        return None, None

    return vcvarsall, None

def _get_vc_env(plat_spec):
    if os.getenv("DISTUTILS_USE_SDK"):
        return {
            key.lower(): value
            for key, value in os.environ.items()
        }

    vcvarsall, _ = _find_vcvarsall(plat_spec)
    if not vcvarsall:
        raise DistutilsPlatformError("Unable to find vcvarsall.bat")

    try:
        out = subprocess.check_output(
            'cmd /u /c "{}" {} && set'.format(vcvarsall, plat_spec),
            stderr=subprocess.STDOUT,
        ).decode('utf-16le', errors='replace')
    except subprocess.CalledProcessError as exc:
        log.error(exc.output)
        raise DistutilsPlatformError("Error executing {}"
                .format(exc.cmd))

    env = {
        key.lower(): value
        for key, _, value in
        (line.partition('=') for line in out.splitlines())
        if key and value
    }

    return env

def _find_exe(exe, paths=None):
    """Return path to an MSVC executable program.

    Tries to find the program in several places: first, one of the
    MSVC program search paths from the registry; next, the directories
    in the PATH environment variable.  If any of those work, return an
    absolute path that is known to exist.  If none of them work, just
    return the original program name, 'exe'.
    """
    if not paths:
        paths = os.getenv('path').split(os.pathsep)
    for p in paths:
        fn = os.path.join(os.path.abspath(p), exe)
        if os.path.isfile(fn):
            return fn
    return exe

# A map keyed by get_platform() return values to values accepted by
# 'vcvarsall.bat'. Always cross-compile from x86 to work with the
# lighter-weight MSVC installs that do not include native 64-bit tools.
PLAT_TO_VCVARS = {
    'win32' : 'x86',
    'win-amd64' : 'x86_amd64',
    'win-arm32' : 'x86_arm',
    'win-arm64' : 'x86_arm64'
}

class MSVCCompiler(CCompiler) :
    """Concrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class."""

    compiler_type = 'msvc'

    # Just set this so CCompiler's constructor doesn't barf.  We currently
    # don't use the 'set_executables()' bureaucracy provided by CCompiler,
    # as it really isn't necessary for this sort of single-compiler class.
    # Would be nice to have a consistent interface with UnixCCompiler,
    # though, so it's worth thinking about.
    executables = {}

    # Private class data (need to distinguish C from C++ source for compiler)
    _c_extensions = ['.c']
    _cpp_extensions = ['.cc', '.cpp', '.cxx']
    _rc_extensions = ['.rc']
    _mc_extensions = ['.mc']

    # Needed for the filename generation methods provided by the
    # base class, CCompiler.
    src_extensions = (_c_extensions + _cpp_extensions +
                      _rc_extensions + _mc_extensions)
    res_extension = '.res'
    obj_extension = '.obj'
    static_lib_extension = '.lib'
    shared_lib_extension = '.dll'
    static_lib_format = shared_lib_format = '%s%s'
    exe_extension = '.exe'


    def __init__(self, verbose=0, dry_run=0, force=0):
        CCompiler.__init__ (self, verbose, dry_run, force)
        # target platform (.plat_name is consistent with 'bdist')
        self.plat_name = None
        self.initialized = False

    def initialize(self, plat_name=None):
        # multi-init means we would need to check platform same each time...
        assert not self.initialized, "don't init multiple times"
        if plat_name is None:
            plat_name = get_platform()
        # sanity check for platforms to prevent obscure errors later.
        if plat_name not in PLAT_TO_VCVARS:
            raise DistutilsPlatformError("--plat-name must be one of {}"
                                         .format(tuple(PLAT_TO_VCVARS)))

        # Get the vcvarsall.bat spec for the requested platform.
        plat_spec = PLAT_TO_VCVARS[plat_name]

        vc_env = _get_vc_env(plat_spec)
        if not vc_env:
            raise DistutilsPlatformError("Unable to find a compatible "
                "Visual Studio installation.")

        self._paths = vc_env.get('path', '')
        paths = self._paths.split(os.pathsep)
        self.cc = _find_exe("cl.exe", paths)
        self.linker = _find_exe("link.exe", paths)
        self.lib = _find_exe("lib.exe", paths)
        self.rc = _find_exe("rc.exe", paths)   # resource compiler
        self.mc = _find_exe("mc.exe", paths)   # message compiler
        self.mt = _find_exe("mt.exe", paths)   # message compiler

        for dir in vc_env.get('include', '').split(os.pathsep):
            if dir:
                self.add_include_dir(dir.rstrip(os.sep))

        for dir in vc_env.get('lib', '').split(os.pathsep):
            if dir:
                self.add_library_dir(dir.rstrip(os.sep))

        self.preprocess_options = None
        # bpo-38597: Always compile with dynamic linking
        # Future releases of Python 3.x will include all past
        # versions of vcruntime*.dll for compatibility.
        self.compile_options = [
            '/nologo', '/O2', '/W3', '/GL', '/DNDEBUG', '/MD'
        ]

        self.compile_options_debug = [
            '/nologo', '/Od', '/MDd', '/Zi', '/W3', '/D_DEBUG'
        ]

        ldflags = [
            '/nologo', '/INCREMENTAL:NO', '/LTCG'
        ]

        ldflags_debug = [
            '/nologo', '/INCREMENTAL:NO', '/LTCG', '/DEBUG:FULL'
        ]

        self.ldflags_exe = [*ldflags, '/MANIFEST:EMBED,ID=1']
        self.ldflags_exe_debug = [*ldflags_debug, '/MANIFEST:EMBED,ID=1']
        self.ldflags_shared = [*ldflags, '/DLL', '/MANIFEST:EMBED,ID=2', '/MANIFESTUAC:NO']
        self.ldflags_shared_debug = [*ldflags_debug, '/DLL', '/MANIFEST:EMBED,ID=2', '/MANIFESTUAC:NO']
        self.ldflags_static = [*ldflags]
        self.ldflags_static_debug = [*ldflags_debug]

        self._ldflags = {
            (CCompiler.EXECUTABLE, None): self.ldflags_exe,
            (CCompiler.EXECUTABLE, False): self.ldflags_exe,
            (CCompiler.EXECUTABLE, True): self.ldflags_exe_debug,
            (CCompiler.SHARED_OBJECT, None): self.ldflags_shared,
            (CCompiler.SHARED_OBJECT, False): self.ldflags_shared,
            (CCompiler.SHARED_OBJECT, True): self.ldflags_shared_debug,
            (CCompiler.SHARED_LIBRARY, None): self.ldflags_static,
            (CCompiler.SHARED_LIBRARY, False): self.ldflags_static,
            (CCompiler.SHARED_LIBRARY, True): self.ldflags_static_debug,
        }

        self.initialized = True

    # -- Worker methods ------------------------------------------------

    def object_filenames(self,
                         source_filenames,
                         strip_dir=0,
                         output_dir=''):
        ext_map = {
            **{ext: self.obj_extension for ext in self.src_extensions},
            **{ext: self.res_extension for ext in self._rc_extensions + self._mc_extensions},
        }

        output_dir = output_dir or ''

        def make_out_path(p):
            base, ext = os.path.splitext(p)
            if strip_dir:
                base = os.path.basename(base)
            else:
                _, base = os.path.splitdrive(base)
                if base.startswith((os.path.sep, os.path.altsep)):
                    base = base[1:]
            try:
                # XXX: This may produce absurdly long paths. We should check
                # the length of the result and trim base until we fit within
                # 260 characters.
                return os.path.join(output_dir, base + ext_map[ext])
            except LookupError:
                # Better to raise an exception instead of silently continuing
                # and later complain about sources and targets having
                # different lengths
                raise CompileError("Don't know how to compile {}".format(p))

        return list(map(make_out_path, source_filenames))


    def compile(self, sources,
                output_dir=None, macros=None, include_dirs=None, debug=0,
                extra_preargs=None, extra_postargs=None, depends=None):

        if not self.initialized:
            self.initialize()
        compile_info = self._setup_compile(output_dir, macros, include_dirs,
                                           sources, depends, extra_postargs)
        macros, objects, extra_postargs, pp_opts, build = compile_info

        compile_opts = extra_preargs or []
        compile_opts.append('/c')
        if debug:
            compile_opts.extend(self.compile_options_debug)
        else:
            compile_opts.extend(self.compile_options)


        add_cpp_opts = False

        for obj in objects:
            try:
                src, ext = build[obj]
            except KeyError:
                continue
            if debug:
                # pass the full pathname to MSVC in debug mode,
                # this allows the debugger to find the source file
                # without asking the user to browse for it
                src = os.path.abspath(src)

            if ext in self._c_extensions:
                input_opt = "/Tc" + src
            elif ext in self._cpp_extensions:
                input_opt = "/Tp" + src
                add_cpp_opts = True
            elif ext in self._rc_extensions:
                # compile .RC to .RES file
                input_opt = src
                output_opt = "/fo" + obj
                try:
                    self.spawn([self.rc] + pp_opts + [output_opt, input_opt])
                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            elif ext in self._mc_extensions:
                # Compile .MC to .RC file to .RES file.
                #   * '-h dir' specifies the directory for the
                #     generated include file
                #   * '-r dir' specifies the target directory of the
                #     generated RC file and the binary message resource
                #     it includes
                #
                # For now (since there are no options to change this),
                # we use the source-directory for the include file and
                # the build directory for the RC file and message
                # resources. This works at least for win32all.
                h_dir = os.path.dirname(src)
                rc_dir = os.path.dirname(obj)
                try:
                    # first compile .MC to .RC and .H file
                    self.spawn([self.mc, '-h', h_dir, '-r', rc_dir, src])
                    base, _ = os.path.splitext(os.path.basename (src))
                    rc_file = os.path.join(rc_dir, base + '.rc')
                    # then compile .RC to .RES file
                    self.spawn([self.rc, "/fo" + obj, rc_file])

                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            else:
                # how to handle this file?
                raise CompileError("Don't know how to compile {} to {}"
                                   .format(src, obj))

            args = [self.cc] + compile_opts + pp_opts
            if add_cpp_opts:
                args.append('/EHsc')
            args.append(input_opt)
            args.append("/Fo" + obj)
            args.extend(extra_postargs)

            try:
                self.spawn(args)
            except DistutilsExecError as msg:
                raise CompileError(msg)

        return objects


    def create_static_lib(self,
                          objects,
                          output_libname,
                          output_dir=None,
                          debug=0,
                          target_lang=None):

        if not self.initialized:
            self.initialize()
        objects, output_dir = self._fix_object_args(objects, output_dir)
        output_filename = self.library_filename(output_libname,
                                                output_dir=output_dir)

        if self._need_link(objects, output_filename):
            lib_args = objects + ['/OUT:' + output_filename]
            if debug:
                pass # XXX what goes here?
            try:
                log.debug('Executing "%s" %s', self.lib, ' '.join(lib_args))
                self.spawn([self.lib] + lib_args)
            except DistutilsExecError as msg:
                raise LibError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)


    def link(self,
             target_desc,
             objects,
             output_filename,
             output_dir=None,
             libraries=None,
             library_dirs=None,
             runtime_library_dirs=None,
             export_symbols=None,
             debug=0,
             extra_preargs=None,
             extra_postargs=None,
             build_temp=None,
             target_lang=None):

        if not self.initialized:
            self.initialize()
        objects, output_dir = self._fix_object_args(objects, output_dir)
        fixed_args = self._fix_lib_args(libraries, library_dirs,
                                        runtime_library_dirs)
        libraries, library_dirs, runtime_library_dirs = fixed_args

        if runtime_library_dirs:
            self.warn("I don't know what to do with 'runtime_library_dirs': "
                       + str(runtime_library_dirs))

        lib_opts = gen_lib_options(self,
                                   library_dirs, runtime_library_dirs,
                                   libraries)
        if output_dir is not None:
            output_filename = os.path.join(output_dir, output_filename)

        if self._need_link(objects, output_filename):
            ldflags = self._ldflags[target_desc, debug]

            export_opts = ["/EXPORT:" + sym for sym in (export_symbols or [])]

            ld_args = (ldflags + lib_opts + export_opts +
                       objects + ['/OUT:' + output_filename])

            # The MSVC linker generates .lib and .exp files, which cannot be
            # suppressed by any linker switches. The .lib files may even be
            # needed! Make sure they are generated in the temporary build
            # directory. Since they have different names for debug and release
            # builds, they can go into the same directory.
            build_temp = os.path.dirname(objects[0])
            if export_symbols is not None:
                (dll_name, dll_ext) = os.path.splitext(
                    os.path.basename(output_filename))
                implib_file = os.path.join(
                    build_temp,
                    self.library_filename(dll_name))
                ld_args.append ('/IMPLIB:' + implib_file)

            if extra_preargs:
                ld_args[:0] = extra_preargs
            if extra_postargs:
                ld_args.extend(extra_postargs)

            output_dir = os.path.dirname(os.path.abspath(output_filename))
            self.mkpath(output_dir)
            try:
                log.debug('Executing "%s" %s', self.linker, ' '.join(ld_args))
                self.spawn([self.linker] + ld_args)
            except DistutilsExecError as msg:
                raise LinkError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    def spawn(self, cmd):
        env = dict(os.environ, PATH=self._paths)
        with self._fallback_spawn(cmd, env) as fallback:
            return super().spawn(cmd, env=env)
        return fallback.value

    @contextlib.contextmanager
    def _fallback_spawn(self, cmd, env):
        """
        Discovered in pypa/distutils#15, some tools monkeypatch the compiler,
        so the 'env' kwarg causes a TypeError. Detect this condition and
        restore the legacy, unsafe behavior.
        """
        bag = type('Bag', (), {})()
        try:
            yield bag
        except TypeError as exc:
            if "unexpected keyword argument 'env'" not in str(exc):
                raise
        else:
            return
        warnings.warn(
            "Fallback spawn triggered. Please update distutils monkeypatch.")
        with unittest.mock.patch('os.environ', env):
            bag.value = super().spawn(cmd)

    # -- Miscellaneous methods -----------------------------------------
    # These are all used by the 'gen_lib_options() function, in
    # ccompiler.py.

    def library_dir_option(self, dir):
        return "/LIBPATH:" + dir

    def runtime_library_dir_option(self, dir):
        raise DistutilsPlatformError(
              "don't know how to set runtime library search path for MSVC")

    def library_option(self, lib):
        return self.library_filename(lib)

    def find_library_file(self, dirs, lib, debug=0):
        # Prefer a debugging library if found (and requested), but deal
        # with it if we don't have one.
        if debug:
            try_names = [lib + "_d", lib]
        else:
            try_names = [lib]
        for dir in dirs:
            for name in try_names:
                libfile = os.path.join(dir, self.library_filename(name))
                if os.path.isfile(libfile):
                    return libfile
        else:
            # Oops, didn't find it in *any* of 'dirs'
            return None
site-packages/setuptools/_distutils/dep_util.py000064400000006643150564745010016067 0ustar00"""distutils.dep_util

Utility functions for simple, timestamp-based dependency of files
and groups of files; also, function based entirely on such
timestamp dependency analysis."""

import os
from distutils.errors import DistutilsFileError


def newer (source, target):
    """Return true if 'source' exists and is more recently modified than
    'target', or if 'source' exists and 'target' doesn't.  Return false if
    both exist and 'target' is the same age or younger than 'source'.
    Raise DistutilsFileError if 'source' does not exist.
    """
    if not os.path.exists(source):
        raise DistutilsFileError("file '%s' does not exist" %
                                 os.path.abspath(source))
    if not os.path.exists(target):
        return 1

    from stat import ST_MTIME
    mtime1 = os.stat(source)[ST_MTIME]
    mtime2 = os.stat(target)[ST_MTIME]

    return mtime1 > mtime2

# newer ()


def newer_pairwise (sources, targets):
    """Walk two filename lists in parallel, testing if each source is newer
    than its corresponding target.  Return a pair of lists (sources,
    targets) where source is newer than target, according to the semantics
    of 'newer()'.
    """
    if len(sources) != len(targets):
        raise ValueError("'sources' and 'targets' must be same length")

    # build a pair of lists (sources, targets) where  source is newer
    n_sources = []
    n_targets = []
    for i in range(len(sources)):
        if newer(sources[i], targets[i]):
            n_sources.append(sources[i])
            n_targets.append(targets[i])

    return (n_sources, n_targets)

# newer_pairwise ()


def newer_group (sources, target, missing='error'):
    """Return true if 'target' is out-of-date with respect to any file
    listed in 'sources'.  In other words, if 'target' exists and is newer
    than every file in 'sources', return false; otherwise return true.
    'missing' controls what we do when a source file is missing; the
    default ("error") is to blow up with an OSError from inside 'stat()';
    if it is "ignore", we silently drop any missing source files; if it is
    "newer", any missing source files make us assume that 'target' is
    out-of-date (this is handy in "dry-run" mode: it'll make you pretend to
    carry out commands that wouldn't work because inputs are missing, but
    that doesn't matter because you're not actually going to run the
    commands).
    """
    # If the target doesn't even exist, then it's definitely out-of-date.
    if not os.path.exists(target):
        return 1

    # Otherwise we have to find out the hard way: if *any* source file
    # is more recent than 'target', then 'target' is out-of-date and
    # we can immediately return true.  If we fall through to the end
    # of the loop, then 'target' is up-to-date and we return false.
    from stat import ST_MTIME
    target_mtime = os.stat(target)[ST_MTIME]
    for source in sources:
        if not os.path.exists(source):
            if missing == 'error':      # blow up when we stat() the file
                pass
            elif missing == 'ignore':   # missing source dropped from
                continue                #  target's dependency list
            elif missing == 'newer':    # missing source means target is
                return 1                #  out-of-date

        source_mtime = os.stat(source)[ST_MTIME]
        if source_mtime > target_mtime:
            return 1
    else:
        return 0

# newer_group ()
site-packages/setuptools/_distutils/filelist.py000064400000032137150564745010016072 0ustar00"""distutils.filelist

Provides the FileList class, used for poking about the filesystem
and building lists of files.
"""

import os
import re
import fnmatch
import functools

from distutils.util import convert_path
from distutils.errors import DistutilsTemplateError, DistutilsInternalError
from distutils import log


class FileList:
    """A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.

    Instance attributes:
      dir
        directory from which files will be taken -- only used if
        'allfiles' not supplied to constructor
      files
        list of filenames currently being built/filtered/manipulated
      allfiles
        complete list of files under consideration (ie. without any
        filtering applied)
    """

    def __init__(self, warn=None, debug_print=None):
        # ignore argument to FileList, but keep them for backwards
        # compatibility
        self.allfiles = None
        self.files = []

    def set_allfiles(self, allfiles):
        self.allfiles = allfiles

    def findall(self, dir=os.curdir):
        self.allfiles = findall(dir)

    def debug_print(self, msg):
        """Print 'msg' to stdout if the global DEBUG (taken from the
        DISTUTILS_DEBUG environment variable) flag is true.
        """
        from distutils.debug import DEBUG
        if DEBUG:
            print(msg)

    # Collection methods

    def append(self, item):
        self.files.append(item)

    def extend(self, items):
        self.files.extend(items)

    def sort(self):
        # Not a strict lexical sort!
        sortable_files = sorted(map(os.path.split, self.files))
        self.files = []
        for sort_tuple in sortable_files:
            self.files.append(os.path.join(*sort_tuple))

    # Other miscellaneous utility methods

    def remove_duplicates(self):
        # Assumes list has been sorted!
        for i in range(len(self.files) - 1, 0, -1):
            if self.files[i] == self.files[i - 1]:
                del self.files[i]

    # "File template" methods

    def _parse_template_line(self, line):
        words = line.split()
        action = words[0]

        patterns = dir = dir_pattern = None

        if action in ('include', 'exclude',
                      'global-include', 'global-exclude'):
            if len(words) < 2:
                raise DistutilsTemplateError(
                      "'%s' expects <pattern1> <pattern2> ..." % action)
            patterns = [convert_path(w) for w in words[1:]]
        elif action in ('recursive-include', 'recursive-exclude'):
            if len(words) < 3:
                raise DistutilsTemplateError(
                      "'%s' expects <dir> <pattern1> <pattern2> ..." % action)
            dir = convert_path(words[1])
            patterns = [convert_path(w) for w in words[2:]]
        elif action in ('graft', 'prune'):
            if len(words) != 2:
                raise DistutilsTemplateError(
                      "'%s' expects a single <dir_pattern>" % action)
            dir_pattern = convert_path(words[1])
        else:
            raise DistutilsTemplateError("unknown action '%s'" % action)

        return (action, patterns, dir, dir_pattern)

    def process_template_line(self, line):
        # Parse the line: split it up, make sure the right number of words
        # is there, and return the relevant words.  'action' is always
        # defined: it's the first word of the line.  Which of the other
        # three are defined depends on the action; it'll be either
        # patterns, (dir and patterns), or (dir_pattern).
        (action, patterns, dir, dir_pattern) = self._parse_template_line(line)

        # OK, now we know that the action is valid and we have the
        # right number of words on the line for that action -- so we
        # can proceed with minimal error-checking.
        if action == 'include':
            self.debug_print("include " + ' '.join(patterns))
            for pattern in patterns:
                if not self.include_pattern(pattern, anchor=1):
                    log.warn("warning: no files found matching '%s'",
                             pattern)

        elif action == 'exclude':
            self.debug_print("exclude " + ' '.join(patterns))
            for pattern in patterns:
                if not self.exclude_pattern(pattern, anchor=1):
                    log.warn(("warning: no previously-included files "
                              "found matching '%s'"), pattern)

        elif action == 'global-include':
            self.debug_print("global-include " + ' '.join(patterns))
            for pattern in patterns:
                if not self.include_pattern(pattern, anchor=0):
                    log.warn(("warning: no files found matching '%s' "
                              "anywhere in distribution"), pattern)

        elif action == 'global-exclude':
            self.debug_print("global-exclude " + ' '.join(patterns))
            for pattern in patterns:
                if not self.exclude_pattern(pattern, anchor=0):
                    log.warn(("warning: no previously-included files matching "
                              "'%s' found anywhere in distribution"),
                             pattern)

        elif action == 'recursive-include':
            self.debug_print("recursive-include %s %s" %
                             (dir, ' '.join(patterns)))
            for pattern in patterns:
                if not self.include_pattern(pattern, prefix=dir):
                    msg = (
                        "warning: no files found matching '%s' "
                        "under directory '%s'"
                    )
                    log.warn(msg, pattern, dir)

        elif action == 'recursive-exclude':
            self.debug_print("recursive-exclude %s %s" %
                             (dir, ' '.join(patterns)))
            for pattern in patterns:
                if not self.exclude_pattern(pattern, prefix=dir):
                    log.warn(("warning: no previously-included files matching "
                              "'%s' found under directory '%s'"),
                             pattern, dir)

        elif action == 'graft':
            self.debug_print("graft " + dir_pattern)
            if not self.include_pattern(None, prefix=dir_pattern):
                log.warn("warning: no directories found matching '%s'",
                         dir_pattern)

        elif action == 'prune':
            self.debug_print("prune " + dir_pattern)
            if not self.exclude_pattern(None, prefix=dir_pattern):
                log.warn(("no previously-included directories found "
                          "matching '%s'"), dir_pattern)
        else:
            raise DistutilsInternalError(
                  "this cannot happen: invalid action '%s'" % action)

    # Filtering/selection methods

    def include_pattern(self, pattern, anchor=1, prefix=None, is_regex=0):
        """Select strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.  Patterns
        are not quite the same as implemented by the 'fnmatch' module: '*'
        and '?'  match non-special characters, where "special" is platform-
        dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found, False otherwise.
        """
        # XXX docstring lying about what the special chars are?
        files_found = False
        pattern_re = translate_pattern(pattern, anchor, prefix, is_regex)
        self.debug_print("include_pattern: applying regex r'%s'" %
                         pattern_re.pattern)

        # delayed loading of allfiles list
        if self.allfiles is None:
            self.findall()

        for name in self.allfiles:
            if pattern_re.search(name):
                self.debug_print(" adding " + name)
                self.files.append(name)
                files_found = True
        return files_found

    def exclude_pattern(
            self, pattern, anchor=1, prefix=None, is_regex=0):
        """Remove strings (presumably filenames) from 'files' that match
        'pattern'.  Other parameters are the same as for
        'include_pattern()', above.
        The list 'self.files' is modified in place.
        Return True if files are found, False otherwise.
        """
        files_found = False
        pattern_re = translate_pattern(pattern, anchor, prefix, is_regex)
        self.debug_print("exclude_pattern: applying regex r'%s'" %
                         pattern_re.pattern)
        for i in range(len(self.files)-1, -1, -1):
            if pattern_re.search(self.files[i]):
                self.debug_print(" removing " + self.files[i])
                del self.files[i]
                files_found = True
        return files_found


# Utility functions

def _find_all_simple(path):
    """
    Find all files under 'path'
    """
    all_unique = _UniqueDirs.filter(os.walk(path, followlinks=True))
    results = (
        os.path.join(base, file)
        for base, dirs, files in all_unique
        for file in files
    )
    return filter(os.path.isfile, results)


class _UniqueDirs(set):
    """
    Exclude previously-seen dirs from walk results,
    avoiding infinite recursion.
    Ref https://bugs.python.org/issue44497.
    """
    def __call__(self, walk_item):
        """
        Given an item from an os.walk result, determine
        if the item represents a unique dir for this instance
        and if not, prevent further traversal.
        """
        base, dirs, files = walk_item
        stat = os.stat(base)
        candidate = stat.st_dev, stat.st_ino
        found = candidate in self
        if found:
            del dirs[:]
        self.add(candidate)
        return not found

    @classmethod
    def filter(cls, items):
        return filter(cls(), items)


def findall(dir=os.curdir):
    """
    Find all files under 'dir' and return the list of full filenames.
    Unless dir is '.', return full filenames with dir prepended.
    """
    files = _find_all_simple(dir)
    if dir == os.curdir:
        make_rel = functools.partial(os.path.relpath, start=dir)
        files = map(make_rel, files)
    return list(files)


def glob_to_re(pattern):
    """Translate a shell-like glob pattern to a regular expression; return
    a string containing the regex.  Differs from 'fnmatch.translate()' in
    that '*' does not match "special characters" (which are
    platform-specific).
    """
    pattern_re = fnmatch.translate(pattern)

    # '?' and '*' in the glob pattern become '.' and '.*' in the RE, which
    # IMHO is wrong -- '?' and '*' aren't supposed to match slash in Unix,
    # and by extension they shouldn't match such "special characters" under
    # any OS.  So change all non-escaped dots in the RE to match any
    # character except the special characters (currently: just os.sep).
    sep = os.sep
    if os.sep == '\\':
        # we're using a regex to manipulate a regex, so we need
        # to escape the backslash twice
        sep = r'\\\\'
    escaped = r'\1[^%s]' % sep
    pattern_re = re.sub(r'((?<!\\)(\\\\)*)\.', escaped, pattern_re)
    return pattern_re


def translate_pattern(pattern, anchor=1, prefix=None, is_regex=0):
    """Translate a shell-like wildcard pattern to a compiled regular
    expression.  Return the compiled regex.  If 'is_regex' true,
    then 'pattern' is directly compiled to a regex (if it's a string)
    or just returned as-is (assumes it's a regex object).
    """
    if is_regex:
        if isinstance(pattern, str):
            return re.compile(pattern)
        else:
            return pattern

    # ditch start and end characters
    start, _, end = glob_to_re('_').partition('_')

    if pattern:
        pattern_re = glob_to_re(pattern)
        assert pattern_re.startswith(start) and pattern_re.endswith(end)
    else:
        pattern_re = ''

    if prefix is not None:
        prefix_re = glob_to_re(prefix)
        assert prefix_re.startswith(start) and prefix_re.endswith(end)
        prefix_re = prefix_re[len(start): len(prefix_re) - len(end)]
        sep = os.sep
        if os.sep == '\\':
            sep = r'\\'
        pattern_re = pattern_re[len(start): len(pattern_re) - len(end)]
        pattern_re = r'%s\A%s%s.*%s%s' % (
            start, prefix_re, sep, pattern_re, end)
    else:                               # no prefix -- respect anchor flag
        if anchor:
            pattern_re = r'%s\A%s' % (start, pattern_re[len(start):])

    return re.compile(pattern_re)
site-packages/setuptools/_distutils/version.py000064400000030342150564745010015740 0ustar00#
# distutils/version.py
#
# Implements multiple version numbering conventions for the
# Python Module Distribution Utilities.
#
# $Id$
#

"""Provides classes to represent module version numbers (one class for
each style of version numbering).  There are currently two such classes
implemented: StrictVersion and LooseVersion.

Every version number class implements the following interface:
  * the 'parse' method takes a string and parses it to some internal
    representation; if the string is an invalid version number,
    'parse' raises a ValueError exception
  * the class constructor takes an optional string argument which,
    if supplied, is passed to 'parse'
  * __str__ reconstructs the string that was passed to 'parse' (or
    an equivalent string -- ie. one that will generate an equivalent
    version number instance)
  * __repr__ generates Python code to recreate the version number instance
  * _cmp compares the current instance with either another instance
    of the same class or a string (which will be parsed to an instance
    of the same class, thus must follow the same rules)
"""

import re

class Version:
    """Abstract base class for version numbering classes.  Just provides
    constructor (__init__) and reproducer (__repr__), because those
    seem to be the same for all version numbering classes; and route
    rich comparisons to _cmp.
    """

    def __init__ (self, vstring=None):
        if vstring:
            self.parse(vstring)

    def __repr__ (self):
        return "%s ('%s')" % (self.__class__.__name__, str(self))

    def __eq__(self, other):
        c = self._cmp(other)
        if c is NotImplemented:
            return c
        return c == 0

    def __lt__(self, other):
        c = self._cmp(other)
        if c is NotImplemented:
            return c
        return c < 0

    def __le__(self, other):
        c = self._cmp(other)
        if c is NotImplemented:
            return c
        return c <= 0

    def __gt__(self, other):
        c = self._cmp(other)
        if c is NotImplemented:
            return c
        return c > 0

    def __ge__(self, other):
        c = self._cmp(other)
        if c is NotImplemented:
            return c
        return c >= 0


# Interface for version-number classes -- must be implemented
# by the following classes (the concrete ones -- Version should
# be treated as an abstract class).
#    __init__ (string) - create and take same action as 'parse'
#                        (string parameter is optional)
#    parse (string)    - convert a string representation to whatever
#                        internal representation is appropriate for
#                        this style of version numbering
#    __str__ (self)    - convert back to a string; should be very similar
#                        (if not identical to) the string supplied to parse
#    __repr__ (self)   - generate Python code to recreate
#                        the instance
#    _cmp (self, other) - compare two version numbers ('other' may
#                        be an unparsed version string, or another
#                        instance of your version class)


class StrictVersion (Version):

    """Version numbering for anal retentives and software idealists.
    Implements the standard interface for version number classes as
    described above.  A version number consists of two or three
    dot-separated numeric components, with an optional "pre-release" tag
    on the end.  The pre-release tag consists of the letter 'a' or 'b'
    followed by a number.  If the numeric components of two version
    numbers are equal, then one with a pre-release tag will always
    be deemed earlier (lesser) than one without.

    The following are valid version numbers (shown in the order that
    would be obtained by sorting according to the supplied cmp function):

        0.4       0.4.0  (these two are equivalent)
        0.4.1
        0.5a1
        0.5b3
        0.5
        0.9.6
        1.0
        1.0.4a3
        1.0.4b1
        1.0.4

    The following are examples of invalid version numbers:

        1
        2.7.2.2
        1.3.a4
        1.3pl1
        1.3c4

    The rationale for this version numbering system will be explained
    in the distutils documentation.
    """

    version_re = re.compile(r'^(\d+) \. (\d+) (\. (\d+))? ([ab](\d+))?$',
                            re.VERBOSE | re.ASCII)


    def parse (self, vstring):
        match = self.version_re.match(vstring)
        if not match:
            raise ValueError("invalid version number '%s'" % vstring)

        (major, minor, patch, prerelease, prerelease_num) = \
            match.group(1, 2, 4, 5, 6)

        if patch:
            self.version = tuple(map(int, [major, minor, patch]))
        else:
            self.version = tuple(map(int, [major, minor])) + (0,)

        if prerelease:
            self.prerelease = (prerelease[0], int(prerelease_num))
        else:
            self.prerelease = None


    def __str__ (self):

        if self.version[2] == 0:
            vstring = '.'.join(map(str, self.version[0:2]))
        else:
            vstring = '.'.join(map(str, self.version))

        if self.prerelease:
            vstring = vstring + self.prerelease[0] + str(self.prerelease[1])

        return vstring


    def _cmp (self, other):
        if isinstance(other, str):
            other = StrictVersion(other)
        elif not isinstance(other, StrictVersion):
            return NotImplemented

        if self.version != other.version:
            # numeric versions don't match
            # prerelease stuff doesn't matter
            if self.version < other.version:
                return -1
            else:
                return 1

        # have to compare prerelease
        # case 1: neither has prerelease; they're equal
        # case 2: self has prerelease, other doesn't; other is greater
        # case 3: self doesn't have prerelease, other does: self is greater
        # case 4: both have prerelease: must compare them!

        if (not self.prerelease and not other.prerelease):
            return 0
        elif (self.prerelease and not other.prerelease):
            return -1
        elif (not self.prerelease and other.prerelease):
            return 1
        elif (self.prerelease and other.prerelease):
            if self.prerelease == other.prerelease:
                return 0
            elif self.prerelease < other.prerelease:
                return -1
            else:
                return 1
        else:
            assert False, "never get here"

# end class StrictVersion


# The rules according to Greg Stein:
# 1) a version number has 1 or more numbers separated by a period or by
#    sequences of letters. If only periods, then these are compared
#    left-to-right to determine an ordering.
# 2) sequences of letters are part of the tuple for comparison and are
#    compared lexicographically
# 3) recognize the numeric components may have leading zeroes
#
# The LooseVersion class below implements these rules: a version number
# string is split up into a tuple of integer and string components, and
# comparison is a simple tuple comparison.  This means that version
# numbers behave in a predictable and obvious way, but a way that might
# not necessarily be how people *want* version numbers to behave.  There
# wouldn't be a problem if people could stick to purely numeric version
# numbers: just split on period and compare the numbers as tuples.
# However, people insist on putting letters into their version numbers;
# the most common purpose seems to be:
#   - indicating a "pre-release" version
#     ('alpha', 'beta', 'a', 'b', 'pre', 'p')
#   - indicating a post-release patch ('p', 'pl', 'patch')
# but of course this can't cover all version number schemes, and there's
# no way to know what a programmer means without asking him.
#
# The problem is what to do with letters (and other non-numeric
# characters) in a version number.  The current implementation does the
# obvious and predictable thing: keep them as strings and compare
# lexically within a tuple comparison.  This has the desired effect if
# an appended letter sequence implies something "post-release":
# eg. "0.99" < "0.99pl14" < "1.0", and "5.001" < "5.001m" < "5.002".
#
# However, if letters in a version number imply a pre-release version,
# the "obvious" thing isn't correct.  Eg. you would expect that
# "1.5.1" < "1.5.2a2" < "1.5.2", but under the tuple/lexical comparison
# implemented here, this just isn't so.
#
# Two possible solutions come to mind.  The first is to tie the
# comparison algorithm to a particular set of semantic rules, as has
# been done in the StrictVersion class above.  This works great as long
# as everyone can go along with bondage and discipline.  Hopefully a
# (large) subset of Python module programmers will agree that the
# particular flavour of bondage and discipline provided by StrictVersion
# provides enough benefit to be worth using, and will submit their
# version numbering scheme to its domination.  The free-thinking
# anarchists in the lot will never give in, though, and something needs
# to be done to accommodate them.
#
# Perhaps a "moderately strict" version class could be implemented that
# lets almost anything slide (syntactically), and makes some heuristic
# assumptions about non-digits in version number strings.  This could
# sink into special-case-hell, though; if I was as talented and
# idiosyncratic as Larry Wall, I'd go ahead and implement a class that
# somehow knows that "1.2.1" < "1.2.2a2" < "1.2.2" < "1.2.2pl3", and is
# just as happy dealing with things like "2g6" and "1.13++".  I don't
# think I'm smart enough to do it right though.
#
# In any case, I've coded the test suite for this module (see
# ../test/test_version.py) specifically to fail on things like comparing
# "1.2a2" and "1.2".  That's not because the *code* is doing anything
# wrong, it's because the simple, obvious design doesn't match my
# complicated, hairy expectations for real-world version numbers.  It
# would be a snap to fix the test suite to say, "Yep, LooseVersion does
# the Right Thing" (ie. the code matches the conception).  But I'd rather
# have a conception that matches common notions about version numbers.

class LooseVersion (Version):

    """Version numbering for anarchists and software realists.
    Implements the standard interface for version number classes as
    described above.  A version number consists of a series of numbers,
    separated by either periods or strings of letters.  When comparing
    version numbers, the numeric components will be compared
    numerically, and the alphabetic components lexically.  The following
    are all valid version numbers, in no particular order:

        1.5.1
        1.5.2b2
        161
        3.10a
        8.02
        3.4j
        1996.07.12
        3.2.pl0
        3.1.1.6
        2g6
        11g
        0.960923
        2.2beta29
        1.13++
        5.5.kw
        2.0b1pl0

    In fact, there is no such thing as an invalid version number under
    this scheme; the rules for comparison are simple and predictable,
    but may not always give the results you want (for some definition
    of "want").
    """

    component_re = re.compile(r'(\d+ | [a-z]+ | \.)', re.VERBOSE)

    def __init__ (self, vstring=None):
        if vstring:
            self.parse(vstring)


    def parse (self, vstring):
        # I've given up on thinking I can reconstruct the version string
        # from the parsed tuple -- so I just store the string here for
        # use by __str__
        self.vstring = vstring
        components = [x for x in self.component_re.split(vstring)
                              if x and x != '.']
        for i, obj in enumerate(components):
            try:
                components[i] = int(obj)
            except ValueError:
                pass

        self.version = components


    def __str__ (self):
        return self.vstring


    def __repr__ (self):
        return "LooseVersion ('%s')" % str(self)


    def _cmp (self, other):
        if isinstance(other, str):
            other = LooseVersion(other)
        elif not isinstance(other, LooseVersion):
            return NotImplemented

        if self.version == other.version:
            return 0
        if self.version < other.version:
            return -1
        if self.version > other.version:
            return 1


# end class LooseVersion
site-packages/setuptools/_distutils/versionpredicate.py000064400000012015150564745010017616 0ustar00"""Module for parsing and testing package version predicate strings.
"""
import re
import distutils.version
import operator


re_validPackage = re.compile(r"(?i)^\s*([a-z_]\w*(?:\.[a-z_]\w*)*)(.*)",
    re.ASCII)
# (package) (rest)

re_paren = re.compile(r"^\s*\((.*)\)\s*$") # (list) inside of parentheses
re_splitComparison = re.compile(r"^\s*(<=|>=|<|>|!=|==)\s*([^\s,]+)\s*$")
# (comp) (version)


def splitUp(pred):
    """Parse a single version comparison.

    Return (comparison string, StrictVersion)
    """
    res = re_splitComparison.match(pred)
    if not res:
        raise ValueError("bad package restriction syntax: %r" % pred)
    comp, verStr = res.groups()
    return (comp, distutils.version.StrictVersion(verStr))

compmap = {"<": operator.lt, "<=": operator.le, "==": operator.eq,
           ">": operator.gt, ">=": operator.ge, "!=": operator.ne}

class VersionPredicate:
    """Parse and test package version predicates.

    >>> v = VersionPredicate('pyepat.abc (>1.0, <3333.3a1, !=1555.1b3)')

    The `name` attribute provides the full dotted name that is given::

    >>> v.name
    'pyepat.abc'

    The str() of a `VersionPredicate` provides a normalized
    human-readable version of the expression::

    >>> print(v)
    pyepat.abc (> 1.0, < 3333.3a1, != 1555.1b3)

    The `satisfied_by()` method can be used to determine with a given
    version number is included in the set described by the version
    restrictions::

    >>> v.satisfied_by('1.1')
    True
    >>> v.satisfied_by('1.4')
    True
    >>> v.satisfied_by('1.0')
    False
    >>> v.satisfied_by('4444.4')
    False
    >>> v.satisfied_by('1555.1b3')
    False

    `VersionPredicate` is flexible in accepting extra whitespace::

    >>> v = VersionPredicate(' pat( ==  0.1  )  ')
    >>> v.name
    'pat'
    >>> v.satisfied_by('0.1')
    True
    >>> v.satisfied_by('0.2')
    False

    If any version numbers passed in do not conform to the
    restrictions of `StrictVersion`, a `ValueError` is raised::

    >>> v = VersionPredicate('p1.p2.p3.p4(>=1.0, <=1.3a1, !=1.2zb3)')
    Traceback (most recent call last):
      ...
    ValueError: invalid version number '1.2zb3'

    It the module or package name given does not conform to what's
    allowed as a legal module or package name, `ValueError` is
    raised::

    >>> v = VersionPredicate('foo-bar')
    Traceback (most recent call last):
      ...
    ValueError: expected parenthesized list: '-bar'

    >>> v = VersionPredicate('foo bar (12.21)')
    Traceback (most recent call last):
      ...
    ValueError: expected parenthesized list: 'bar (12.21)'

    """

    def __init__(self, versionPredicateStr):
        """Parse a version predicate string.
        """
        # Fields:
        #    name:  package name
        #    pred:  list of (comparison string, StrictVersion)

        versionPredicateStr = versionPredicateStr.strip()
        if not versionPredicateStr:
            raise ValueError("empty package restriction")
        match = re_validPackage.match(versionPredicateStr)
        if not match:
            raise ValueError("bad package name in %r" % versionPredicateStr)
        self.name, paren = match.groups()
        paren = paren.strip()
        if paren:
            match = re_paren.match(paren)
            if not match:
                raise ValueError("expected parenthesized list: %r" % paren)
            str = match.groups()[0]
            self.pred = [splitUp(aPred) for aPred in str.split(",")]
            if not self.pred:
                raise ValueError("empty parenthesized list in %r"
                                 % versionPredicateStr)
        else:
            self.pred = []

    def __str__(self):
        if self.pred:
            seq = [cond + " " + str(ver) for cond, ver in self.pred]
            return self.name + " (" + ", ".join(seq) + ")"
        else:
            return self.name

    def satisfied_by(self, version):
        """True if version is compatible with all the predicates in self.
        The parameter version must be acceptable to the StrictVersion
        constructor.  It may be either a string or StrictVersion.
        """
        for cond, ver in self.pred:
            if not compmap[cond](version, ver):
                return False
        return True


_provision_rx = None

def split_provision(value):
    """Return the name and optional version number of a provision.

    The version number, if given, will be returned as a `StrictVersion`
    instance, otherwise it will be `None`.

    >>> split_provision('mypkg')
    ('mypkg', None)
    >>> split_provision(' mypkg( 1.2 ) ')
    ('mypkg', StrictVersion ('1.2'))
    """
    global _provision_rx
    if _provision_rx is None:
        _provision_rx = re.compile(
            r"([a-zA-Z_]\w*(?:\.[a-zA-Z_]\w*)*)(?:\s*\(\s*([^)\s]+)\s*\))?$",
            re.ASCII)
    value = value.strip()
    m = _provision_rx.match(value)
    if not m:
        raise ValueError("illegal provides specification: %r" % value)
    ver = m.group(2) or None
    if ver:
        ver = distutils.version.StrictVersion(ver)
    return m.group(1), ver
site-packages/setuptools/_distutils/__init__.py000064400000000372150564745010016012 0ustar00"""distutils

The main package for the Python Module Distribution Utilities.  Normally
used from a setup script as

   from distutils.core import setup

   setup (...)
"""

import sys

__version__ = sys.version[:sys.version.index(' ')]

local = True
site-packages/setuptools/_distutils/util.py000064400000047627150564745010015246 0ustar00"""distutils.util

Miscellaneous utility functions -- anything that doesn't fit into
one of the other *util.py modules.
"""

import os
import re
import importlib.util
import string
import sys
from distutils.errors import DistutilsPlatformError
from distutils.dep_util import newer
from distutils.spawn import spawn
from distutils import log
from distutils.errors import DistutilsByteCompileError
from .py35compat import _optim_args_from_interpreter_flags


def get_host_platform():
    """Return a string that identifies the current platform.  This is used mainly to
    distinguish platform-specific build directories and platform-specific built
    distributions.  Typically includes the OS name and version and the
    architecture (as supplied by 'os.uname()'), although the exact information
    included depends on the OS; eg. on Linux, the kernel version isn't
    particularly important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.

    """
    if os.name == 'nt':
        if 'amd64' in sys.version.lower():
            return 'win-amd64'
        if '(arm)' in sys.version.lower():
            return 'win-arm32'
        if '(arm64)' in sys.version.lower():
            return 'win-arm64'
        return sys.platform

    # Set for cross builds explicitly
    if "_PYTHON_HOST_PLATFORM" in os.environ:
        return os.environ["_PYTHON_HOST_PLATFORM"]

    if os.name != "posix" or not hasattr(os, 'uname'):
        # XXX what about the architecture? NT is Intel or Alpha,
        # Mac OS is M68k or PPC, etc.
        return sys.platform

    # Try to distinguish various flavours of Unix

    (osname, host, release, version, machine) = os.uname()

    # Convert the OS name to lowercase, remove '/' characters, and translate
    # spaces (for "Power Macintosh")
    osname = osname.lower().replace('/', '')
    machine = machine.replace(' ', '_')
    machine = machine.replace('/', '-')

    if osname[:5] == "linux":
        # At least on Linux/Intel, 'machine' is the processor --
        # i386, etc.
        # XXX what about Alpha, SPARC, etc?
        return  "%s-%s" % (osname, machine)
    elif osname[:5] == "sunos":
        if release[0] >= "5":           # SunOS 5 == Solaris 2
            osname = "solaris"
            release = "%d.%s" % (int(release[0]) - 3, release[2:])
            # We can't use "platform.architecture()[0]" because a
            # bootstrap problem. We use a dict to get an error
            # if some suspicious happens.
            bitness = {2147483647:"32bit", 9223372036854775807:"64bit"}
            machine += ".%s" % bitness[sys.maxsize]
        # fall through to standard osname-release-machine representation
    elif osname[:3] == "aix":
        from .py38compat import aix_platform
        return aix_platform(osname, version, release)
    elif osname[:6] == "cygwin":
        osname = "cygwin"
        rel_re = re.compile (r'[\d.]+', re.ASCII)
        m = rel_re.match(release)
        if m:
            release = m.group()
    elif osname[:6] == "darwin":
        import _osx_support, distutils.sysconfig
        osname, release, machine = _osx_support.get_platform_osx(
                                        distutils.sysconfig.get_config_vars(),
                                        osname, release, machine)

    return "%s-%s-%s" % (osname, release, machine)

def get_platform():
    if os.name == 'nt':
        TARGET_TO_PLAT = {
            'x86' : 'win32',
            'x64' : 'win-amd64',
            'arm' : 'win-arm32',
            'arm64': 'win-arm64',
        }
        return TARGET_TO_PLAT.get(os.environ.get('VSCMD_ARG_TGT_ARCH')) or get_host_platform()
    else:
        return get_host_platform()


if sys.platform == 'darwin':
    _syscfg_macosx_ver = None # cache the version pulled from sysconfig
MACOSX_VERSION_VAR = 'MACOSX_DEPLOYMENT_TARGET'

def _clear_cached_macosx_ver():
    """For testing only. Do not call."""
    global _syscfg_macosx_ver
    _syscfg_macosx_ver = None

def get_macosx_target_ver_from_syscfg():
    """Get the version of macOS latched in the Python interpreter configuration.
    Returns the version as a string or None if can't obtain one. Cached."""
    global _syscfg_macosx_ver
    if _syscfg_macosx_ver is None:
        from distutils import sysconfig
        ver = sysconfig.get_config_var(MACOSX_VERSION_VAR) or ''
        if ver:
            _syscfg_macosx_ver = ver
    return _syscfg_macosx_ver

def get_macosx_target_ver():
    """Return the version of macOS for which we are building.

    The target version defaults to the version in sysconfig latched at time
    the Python interpreter was built, unless overridden by an environment
    variable. If neither source has a value, then None is returned"""

    syscfg_ver = get_macosx_target_ver_from_syscfg()
    env_ver = os.environ.get(MACOSX_VERSION_VAR)

    if env_ver:
        # Validate overridden version against sysconfig version, if have both.
        # Ensure that the deployment target of the build process is not less
        # than 10.3 if the interpreter was built for 10.3 or later.  This
        # ensures extension modules are built with correct compatibility
        # values, specifically LDSHARED which can use
        # '-undefined dynamic_lookup' which only works on >= 10.3.
        if syscfg_ver and split_version(syscfg_ver) >= [10, 3] and \
            split_version(env_ver) < [10, 3]:
            my_msg = ('$' + MACOSX_VERSION_VAR + ' mismatch: '
                      'now "%s" but "%s" during configure; '
                      'must use 10.3 or later'
                      % (env_ver, syscfg_ver))
            raise DistutilsPlatformError(my_msg)
        return env_ver
    return syscfg_ver


def split_version(s):
    """Convert a dot-separated string into a list of numbers for comparisons"""
    return [int(n) for n in s.split('.')]


def convert_path (pathname):
    """Return 'pathname' as a name that will work on the native filesystem,
    i.e. split it on '/' and put it back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    """
    if os.sep == '/':
        return pathname
    if not pathname:
        return pathname
    if pathname[0] == '/':
        raise ValueError("path '%s' cannot be absolute" % pathname)
    if pathname[-1] == '/':
        raise ValueError("path '%s' cannot end with '/'" % pathname)

    paths = pathname.split('/')
    while '.' in paths:
        paths.remove('.')
    if not paths:
        return os.curdir
    return os.path.join(*paths)

# convert_path ()


def change_root (new_root, pathname):
    """Return 'pathname' with 'new_root' prepended.  If 'pathname' is
    relative, this is equivalent to "os.path.join(new_root,pathname)".
    Otherwise, it requires making 'pathname' relative and then joining the
    two, which is tricky on DOS/Windows and Mac OS.
    """
    if os.name == 'posix':
        if not os.path.isabs(pathname):
            return os.path.join(new_root, pathname)
        else:
            return os.path.join(new_root, pathname[1:])

    elif os.name == 'nt':
        (drive, path) = os.path.splitdrive(pathname)
        if path[0] == '\\':
            path = path[1:]
        return os.path.join(new_root, path)

    else:
        raise DistutilsPlatformError("nothing known about platform '%s'" % os.name)


_environ_checked = 0
def check_environ ():
    """Ensure that 'os.environ' has all the environment variables we
    guarantee that users can use in config files, command-line options,
    etc.  Currently this includes:
      HOME - user's home directory (Unix only)
      PLAT - description of the current platform, including hardware
             and OS (see 'get_platform()')
    """
    global _environ_checked
    if _environ_checked:
        return

    if os.name == 'posix' and 'HOME' not in os.environ:
        try:
            import pwd
            os.environ['HOME'] = pwd.getpwuid(os.getuid())[5]
        except (ImportError, KeyError):
            # bpo-10496: if the current user identifier doesn't exist in the
            # password database, do nothing
            pass

    if 'PLAT' not in os.environ:
        os.environ['PLAT'] = get_platform()

    _environ_checked = 1


def subst_vars (s, local_vars):
    """Perform shell/Perl-style variable substitution on 'string'.  Every
    occurrence of '$' followed by a name is considered a variable, and
    variable is substituted by the value found in the 'local_vars'
    dictionary, or in 'os.environ' if it's not in 'local_vars'.
    'os.environ' is first checked/augmented to guarantee that it contains
    certain values: see 'check_environ()'.  Raise ValueError for any
    variables not found in either 'local_vars' or 'os.environ'.
    """
    check_environ()
    def _subst (match, local_vars=local_vars):
        var_name = match.group(1)
        if var_name in local_vars:
            return str(local_vars[var_name])
        else:
            return os.environ[var_name]

    try:
        return re.sub(r'\$([a-zA-Z_][a-zA-Z_0-9]*)', _subst, s)
    except KeyError as var:
        raise ValueError("invalid variable '$%s'" % var)

# subst_vars ()


def grok_environment_error (exc, prefix="error: "):
    # Function kept for backward compatibility.
    # Used to try clever things with EnvironmentErrors,
    # but nowadays str(exception) produces good messages.
    return prefix + str(exc)


# Needed by 'split_quoted()'
_wordchars_re = _squote_re = _dquote_re = None
def _init_regex():
    global _wordchars_re, _squote_re, _dquote_re
    _wordchars_re = re.compile(r'[^\\\'\"%s ]*' % string.whitespace)
    _squote_re = re.compile(r"'(?:[^'\\]|\\.)*'")
    _dquote_re = re.compile(r'"(?:[^"\\]|\\.)*"')

def split_quoted (s):
    """Split a string up according to Unix shell-like rules for quotes and
    backslashes.  In short: words are delimited by spaces, as long as those
    spaces are not escaped by a backslash, or inside a quoted string.
    Single and double quotes are equivalent, and the quote characters can
    be backslash-escaped.  The backslash is stripped from any two-character
    escape sequence, leaving only the escaped character.  The quote
    characters are stripped from any quoted string.  Returns a list of
    words.
    """

    # This is a nice algorithm for splitting up a single string, since it
    # doesn't require character-by-character examination.  It was a little
    # bit of a brain-bender to get it working right, though...
    if _wordchars_re is None: _init_regex()

    s = s.strip()
    words = []
    pos = 0

    while s:
        m = _wordchars_re.match(s, pos)
        end = m.end()
        if end == len(s):
            words.append(s[:end])
            break

        if s[end] in string.whitespace: # unescaped, unquoted whitespace: now
            words.append(s[:end])       # we definitely have a word delimiter
            s = s[end:].lstrip()
            pos = 0

        elif s[end] == '\\':            # preserve whatever is being escaped;
                                        # will become part of the current word
            s = s[:end] + s[end+1:]
            pos = end+1

        else:
            if s[end] == "'":           # slurp singly-quoted string
                m = _squote_re.match(s, end)
            elif s[end] == '"':         # slurp doubly-quoted string
                m = _dquote_re.match(s, end)
            else:
                raise RuntimeError("this can't happen (bad char '%c')" % s[end])

            if m is None:
                raise ValueError("bad string (mismatched %s quotes?)" % s[end])

            (beg, end) = m.span()
            s = s[:beg] + s[beg+1:end-1] + s[end:]
            pos = m.end() - 2

        if pos >= len(s):
            words.append(s)
            break

    return words

# split_quoted ()


def execute (func, args, msg=None, verbose=0, dry_run=0):
    """Perform some action that affects the outside world (eg.  by
    writing to the filesystem).  Such actions are special because they
    are disabled by the 'dry_run' flag.  This method takes care of all
    that bureaucracy for you; all you have to do is supply the
    function to call and an argument tuple for it (to embody the
    "external action" being performed), and an optional message to
    print.
    """
    if msg is None:
        msg = "%s%r" % (func.__name__, args)
        if msg[-2:] == ',)':        # correct for singleton tuple
            msg = msg[0:-2] + ')'

    log.info(msg)
    if not dry_run:
        func(*args)


def strtobool (val):
    """Convert a string representation of truth to true (1) or false (0).

    True values are 'y', 'yes', 't', 'true', 'on', and '1'; false values
    are 'n', 'no', 'f', 'false', 'off', and '0'.  Raises ValueError if
    'val' is anything else.
    """
    val = val.lower()
    if val in ('y', 'yes', 't', 'true', 'on', '1'):
        return 1
    elif val in ('n', 'no', 'f', 'false', 'off', '0'):
        return 0
    else:
        raise ValueError("invalid truth value %r" % (val,))


def byte_compile (py_files,
                  optimize=0, force=0,
                  prefix=None, base_dir=None,
                  verbose=1, dry_run=0,
                  direct=None):
    """Byte-compile a collection of Python source files to .pyc
    files in a __pycache__ subdirectory.  'py_files' is a list
    of files to compile; any files that don't end in ".py" are silently
    skipped.  'optimize' must be one of the following:
      0 - don't optimize
      1 - normal optimization (like "python -O")
      2 - extra optimization (like "python -OO")
    If 'force' is true, all files are recompiled regardless of
    timestamps.

    The source filename encoded in each bytecode file defaults to the
    filenames listed in 'py_files'; you can modify these with 'prefix' and
    'basedir'.  'prefix' is a string that will be stripped off of each
    source filename, and 'base_dir' is a directory name that will be
    prepended (after 'prefix' is stripped).  You can supply either or both
    (or neither) of 'prefix' and 'base_dir', as you wish.

    If 'dry_run' is true, doesn't actually do anything that would
    affect the filesystem.

    Byte-compilation is either done directly in this interpreter process
    with the standard py_compile module, or indirectly by writing a
    temporary script and executing it.  Normally, you should let
    'byte_compile()' figure out to use direct compilation or not (see
    the source for details).  The 'direct' flag is used by the script
    generated in indirect mode; unless you know what you're doing, leave
    it set to None.
    """

    # Late import to fix a bootstrap issue: _posixsubprocess is built by
    # setup.py, but setup.py uses distutils.
    import subprocess

    # nothing is done if sys.dont_write_bytecode is True
    if sys.dont_write_bytecode:
        raise DistutilsByteCompileError('byte-compiling is disabled.')

    # First, if the caller didn't force us into direct or indirect mode,
    # figure out which mode we should be in.  We take a conservative
    # approach: choose direct mode *only* if the current interpreter is
    # in debug mode and optimize is 0.  If we're not in debug mode (-O
    # or -OO), we don't know which level of optimization this
    # interpreter is running with, so we can't do direct
    # byte-compilation and be certain that it's the right thing.  Thus,
    # always compile indirectly if the current interpreter is in either
    # optimize mode, or if either optimization level was requested by
    # the caller.
    if direct is None:
        direct = (__debug__ and optimize == 0)

    # "Indirect" byte-compilation: write a temporary script and then
    # run it with the appropriate flags.
    if not direct:
        try:
            from tempfile import mkstemp
            (script_fd, script_name) = mkstemp(".py")
        except ImportError:
            from tempfile import mktemp
            (script_fd, script_name) = None, mktemp(".py")
        log.info("writing byte-compilation script '%s'", script_name)
        if not dry_run:
            if script_fd is not None:
                script = os.fdopen(script_fd, "w")
            else:
                script = open(script_name, "w")

            with script:
                script.write("""\
from distutils.util import byte_compile
files = [
""")

                # XXX would be nice to write absolute filenames, just for
                # safety's sake (script should be more robust in the face of
                # chdir'ing before running it).  But this requires abspath'ing
                # 'prefix' as well, and that breaks the hack in build_lib's
                # 'byte_compile()' method that carefully tacks on a trailing
                # slash (os.sep really) to make sure the prefix here is "just
                # right".  This whole prefix business is rather delicate -- the
                # problem is that it's really a directory, but I'm treating it
                # as a dumb string, so trailing slashes and so forth matter.

                #py_files = map(os.path.abspath, py_files)
                #if prefix:
                #    prefix = os.path.abspath(prefix)

                script.write(",\n".join(map(repr, py_files)) + "]\n")
                script.write("""
byte_compile(files, optimize=%r, force=%r,
             prefix=%r, base_dir=%r,
             verbose=%r, dry_run=0,
             direct=1)
""" % (optimize, force, prefix, base_dir, verbose))

        cmd = [sys.executable]
        cmd.extend(_optim_args_from_interpreter_flags())
        cmd.append(script_name)
        spawn(cmd, dry_run=dry_run)
        execute(os.remove, (script_name,), "removing %s" % script_name,
                dry_run=dry_run)

    # "Direct" byte-compilation: use the py_compile module to compile
    # right here, right now.  Note that the script generated in indirect
    # mode simply calls 'byte_compile()' in direct mode, a weird sort of
    # cross-process recursion.  Hey, it works!
    else:
        from py_compile import compile

        for file in py_files:
            if file[-3:] != ".py":
                # This lets us be lazy and not filter filenames in
                # the "install_lib" command.
                continue

            # Terminology from the py_compile module:
            #   cfile - byte-compiled file
            #   dfile - purported source filename (same as 'file' by default)
            if optimize >= 0:
                opt = '' if optimize == 0 else optimize
                cfile = importlib.util.cache_from_source(
                    file, optimization=opt)
            else:
                cfile = importlib.util.cache_from_source(file)
            dfile = file
            if prefix:
                if file[:len(prefix)] != prefix:
                    raise ValueError("invalid prefix: filename %r doesn't start with %r"
                           % (file, prefix))
                dfile = dfile[len(prefix):]
            if base_dir:
                dfile = os.path.join(base_dir, dfile)

            cfile_base = os.path.basename(cfile)
            if direct:
                if force or newer(file, cfile):
                    log.info("byte-compiling %s to %s", file, cfile_base)
                    if not dry_run:
                        compile(file, cfile, dfile)
                else:
                    log.debug("skipping byte-compilation of %s to %s",
                              file, cfile_base)

# byte_compile ()

def rfc822_escape (header):
    """Return a version of the string escaped for inclusion in an
    RFC-822 header, by ensuring there are 8 spaces space after each newline.
    """
    lines = header.split('\n')
    sep = '\n' + 8 * ' '
    return sep.join(lines)
site-packages/setuptools/_distutils/log.py000064400000003661150564745010015040 0ustar00"""A simple log mechanism styled after PEP 282."""

# The class here is styled after PEP 282 so that it could later be
# replaced with a standard Python logging implementation.

DEBUG = 1
INFO = 2
WARN = 3
ERROR = 4
FATAL = 5

import sys

class Log:

    def __init__(self, threshold=WARN):
        self.threshold = threshold

    def _log(self, level, msg, args):
        if level not in (DEBUG, INFO, WARN, ERROR, FATAL):
            raise ValueError('%s wrong log level' % str(level))

        if level >= self.threshold:
            if args:
                msg = msg % args
            if level in (WARN, ERROR, FATAL):
                stream = sys.stderr
            else:
                stream = sys.stdout
            try:
                stream.write('%s\n' % msg)
            except UnicodeEncodeError:
                # emulate backslashreplace error handler
                encoding = stream.encoding
                msg = msg.encode(encoding, "backslashreplace").decode(encoding)
                stream.write('%s\n' % msg)
            stream.flush()

    def log(self, level, msg, *args):
        self._log(level, msg, args)

    def debug(self, msg, *args):
        self._log(DEBUG, msg, args)

    def info(self, msg, *args):
        self._log(INFO, msg, args)

    def warn(self, msg, *args):
        self._log(WARN, msg, args)

    def error(self, msg, *args):
        self._log(ERROR, msg, args)

    def fatal(self, msg, *args):
        self._log(FATAL, msg, args)

_global_log = Log()
log = _global_log.log
debug = _global_log.debug
info = _global_log.info
warn = _global_log.warn
error = _global_log.error
fatal = _global_log.fatal

def set_threshold(level):
    # return the old threshold for use from tests
    old = _global_log.threshold
    _global_log.threshold = level
    return old

def set_verbosity(v):
    if v <= 0:
        set_threshold(WARN)
    elif v == 1:
        set_threshold(INFO)
    elif v >= 2:
        set_threshold(DEBUG)
site-packages/setuptools/_distutils/dist.py000064400000142365150564745010015227 0ustar00"""distutils.dist

Provides the Distribution class, which represents the module distribution
being built/installed/distributed.
"""

import sys
import os
import re
from email import message_from_file

try:
    import warnings
except ImportError:
    warnings = None

from distutils.errors import *
from distutils.fancy_getopt import FancyGetopt, translate_longopt
from distutils.util import check_environ, strtobool, rfc822_escape
from distutils import log
from distutils.debug import DEBUG

# Regex to define acceptable Distutils command names.  This is not *quite*
# the same as a Python NAME -- I don't allow leading underscores.  The fact
# that they're very similar is no coincidence; the default naming scheme is
# to look for a Python module named after the command.
command_re = re.compile(r'^[a-zA-Z]([a-zA-Z0-9_]*)$')


def _ensure_list(value, fieldname):
    if isinstance(value, str):
        # a string containing comma separated values is okay.  It will
        # be converted to a list by Distribution.finalize_options().
        pass
    elif not isinstance(value, list):
        # passing a tuple or an iterator perhaps, warn and convert
        typename = type(value).__name__
        msg = "Warning: '{fieldname}' should be a list, got type '{typename}'"
        msg = msg.format(**locals())
        log.log(log.WARN, msg)
        value = list(value)
    return value


class Distribution:
    """The core of the Distutils.  Most of the work hiding behind 'setup'
    is really done within a Distribution instance, which farms the work out
    to the Distutils commands specified on the command line.

    Setup scripts will almost never instantiate Distribution directly,
    unless the 'setup()' function is totally inadequate to their needs.
    However, it is conceivable that a setup script might wish to subclass
    Distribution for some specialized purpose, and then pass the subclass
    to 'setup()' as the 'distclass' keyword argument.  If so, it is
    necessary to respect the expectations that 'setup' has of Distribution.
    See the code for 'setup()', in core.py, for details.
    """

    # 'global_options' describes the command-line options that may be
    # supplied to the setup script prior to any actual commands.
    # Eg. "./setup.py -n" or "./setup.py --quiet" both take advantage of
    # these global options.  This list should be kept to a bare minimum,
    # since every global option is also valid as a command option -- and we
    # don't want to pollute the commands with too many options that they
    # have minimal control over.
    # The fourth entry for verbose means that it can be repeated.
    global_options = [
        ('verbose', 'v', "run verbosely (default)", 1),
        ('quiet', 'q', "run quietly (turns verbosity off)"),
        ('dry-run', 'n', "don't actually do anything"),
        ('help', 'h', "show detailed help message"),
        ('no-user-cfg', None,
            'ignore pydistutils.cfg in your home directory'),
    ]

    # 'common_usage' is a short (2-3 line) string describing the common
    # usage of the setup script.
    common_usage = """\
Common commands: (see '--help-commands' for more)

  setup.py build      will build the package underneath 'build/'
  setup.py install    will install the package
"""

    # options that are not propagated to the commands
    display_options = [
        ('help-commands', None,
         "list all available commands"),
        ('name', None,
         "print package name"),
        ('version', 'V',
         "print package version"),
        ('fullname', None,
         "print <package name>-<version>"),
        ('author', None,
         "print the author's name"),
        ('author-email', None,
         "print the author's email address"),
        ('maintainer', None,
         "print the maintainer's name"),
        ('maintainer-email', None,
         "print the maintainer's email address"),
        ('contact', None,
         "print the maintainer's name if known, else the author's"),
        ('contact-email', None,
         "print the maintainer's email address if known, else the author's"),
        ('url', None,
         "print the URL for this package"),
        ('license', None,
         "print the license of the package"),
        ('licence', None,
         "alias for --license"),
        ('description', None,
         "print the package description"),
        ('long-description', None,
         "print the long package description"),
        ('platforms', None,
         "print the list of platforms"),
        ('classifiers', None,
         "print the list of classifiers"),
        ('keywords', None,
         "print the list of keywords"),
        ('provides', None,
         "print the list of packages/modules provided"),
        ('requires', None,
         "print the list of packages/modules required"),
        ('obsoletes', None,
         "print the list of packages/modules made obsolete")
        ]
    display_option_names = [translate_longopt(x[0]) for x in display_options]

    # negative options are options that exclude other options
    negative_opt = {'quiet': 'verbose'}

    # -- Creation/initialization methods -------------------------------

    def __init__(self, attrs=None):
        """Construct a new Distribution instance: initialize all the
        attributes of a Distribution, and then use 'attrs' (a dictionary
        mapping attribute names to values) to assign some of those
        attributes their "real" values.  (Any attributes not mentioned in
        'attrs' will be assigned to some null value: 0, None, an empty list
        or dictionary, etc.)  Most importantly, initialize the
        'command_obj' attribute to the empty dictionary; this will be
        filled in with real command objects by 'parse_command_line()'.
        """

        # Default values for our command-line options
        self.verbose = 1
        self.dry_run = 0
        self.help = 0
        for attr in self.display_option_names:
            setattr(self, attr, 0)

        # Store the distribution meta-data (name, version, author, and so
        # forth) in a separate object -- we're getting to have enough
        # information here (and enough command-line options) that it's
        # worth it.  Also delegate 'get_XXX()' methods to the 'metadata'
        # object in a sneaky and underhanded (but efficient!) way.
        self.metadata = DistributionMetadata()
        for basename in self.metadata._METHOD_BASENAMES:
            method_name = "get_" + basename
            setattr(self, method_name, getattr(self.metadata, method_name))

        # 'cmdclass' maps command names to class objects, so we
        # can 1) quickly figure out which class to instantiate when
        # we need to create a new command object, and 2) have a way
        # for the setup script to override command classes
        self.cmdclass = {}

        # 'command_packages' is a list of packages in which commands
        # are searched for.  The factory for command 'foo' is expected
        # to be named 'foo' in the module 'foo' in one of the packages
        # named here.  This list is searched from the left; an error
        # is raised if no named package provides the command being
        # searched for.  (Always access using get_command_packages().)
        self.command_packages = None

        # 'script_name' and 'script_args' are usually set to sys.argv[0]
        # and sys.argv[1:], but they can be overridden when the caller is
        # not necessarily a setup script run from the command-line.
        self.script_name = None
        self.script_args = None

        # 'command_options' is where we store command options between
        # parsing them (from config files, the command-line, etc.) and when
        # they are actually needed -- ie. when the command in question is
        # instantiated.  It is a dictionary of dictionaries of 2-tuples:
        #   command_options = { command_name : { option : (source, value) } }
        self.command_options = {}

        # 'dist_files' is the list of (command, pyversion, file) that
        # have been created by any dist commands run so far. This is
        # filled regardless of whether the run is dry or not. pyversion
        # gives sysconfig.get_python_version() if the dist file is
        # specific to a Python version, 'any' if it is good for all
        # Python versions on the target platform, and '' for a source
        # file. pyversion should not be used to specify minimum or
        # maximum required Python versions; use the metainfo for that
        # instead.
        self.dist_files = []

        # These options are really the business of various commands, rather
        # than of the Distribution itself.  We provide aliases for them in
        # Distribution as a convenience to the developer.
        self.packages = None
        self.package_data = {}
        self.package_dir = None
        self.py_modules = None
        self.libraries = None
        self.headers = None
        self.ext_modules = None
        self.ext_package = None
        self.include_dirs = None
        self.extra_path = None
        self.scripts = None
        self.data_files = None
        self.password = ''

        # And now initialize bookkeeping stuff that can't be supplied by
        # the caller at all.  'command_obj' maps command names to
        # Command instances -- that's how we enforce that every command
        # class is a singleton.
        self.command_obj = {}

        # 'have_run' maps command names to boolean values; it keeps track
        # of whether we have actually run a particular command, to make it
        # cheap to "run" a command whenever we think we might need to -- if
        # it's already been done, no need for expensive filesystem
        # operations, we just check the 'have_run' dictionary and carry on.
        # It's only safe to query 'have_run' for a command class that has
        # been instantiated -- a false value will be inserted when the
        # command object is created, and replaced with a true value when
        # the command is successfully run.  Thus it's probably best to use
        # '.get()' rather than a straight lookup.
        self.have_run = {}

        # Now we'll use the attrs dictionary (ultimately, keyword args from
        # the setup script) to possibly override any or all of these
        # distribution options.

        if attrs:
            # Pull out the set of command options and work on them
            # specifically.  Note that this order guarantees that aliased
            # command options will override any supplied redundantly
            # through the general options dictionary.
            options = attrs.get('options')
            if options is not None:
                del attrs['options']
                for (command, cmd_options) in options.items():
                    opt_dict = self.get_option_dict(command)
                    for (opt, val) in cmd_options.items():
                        opt_dict[opt] = ("setup script", val)

            if 'licence' in attrs:
                attrs['license'] = attrs['licence']
                del attrs['licence']
                msg = "'licence' distribution option is deprecated; use 'license'"
                if warnings is not None:
                    warnings.warn(msg)
                else:
                    sys.stderr.write(msg + "\n")

            # Now work on the rest of the attributes.  Any attribute that's
            # not already defined is invalid!
            for (key, val) in attrs.items():
                if hasattr(self.metadata, "set_" + key):
                    getattr(self.metadata, "set_" + key)(val)
                elif hasattr(self.metadata, key):
                    setattr(self.metadata, key, val)
                elif hasattr(self, key):
                    setattr(self, key, val)
                else:
                    msg = "Unknown distribution option: %s" % repr(key)
                    warnings.warn(msg)

        # no-user-cfg is handled before other command line args
        # because other args override the config files, and this
        # one is needed before we can load the config files.
        # If attrs['script_args'] wasn't passed, assume false.
        #
        # This also make sure we just look at the global options
        self.want_user_cfg = True

        if self.script_args is not None:
            for arg in self.script_args:
                if not arg.startswith('-'):
                    break
                if arg == '--no-user-cfg':
                    self.want_user_cfg = False
                    break

        self.finalize_options()

    def get_option_dict(self, command):
        """Get the option dictionary for a given command.  If that
        command's option dictionary hasn't been created yet, then create it
        and return the new dictionary; otherwise, return the existing
        option dictionary.
        """
        dict = self.command_options.get(command)
        if dict is None:
            dict = self.command_options[command] = {}
        return dict

    def dump_option_dicts(self, header=None, commands=None, indent=""):
        from pprint import pformat

        if commands is None:             # dump all command option dicts
            commands = sorted(self.command_options.keys())

        if header is not None:
            self.announce(indent + header)
            indent = indent + "  "

        if not commands:
            self.announce(indent + "no commands known yet")
            return

        for cmd_name in commands:
            opt_dict = self.command_options.get(cmd_name)
            if opt_dict is None:
                self.announce(indent +
                              "no option dict for '%s' command" % cmd_name)
            else:
                self.announce(indent +
                              "option dict for '%s' command:" % cmd_name)
                out = pformat(opt_dict)
                for line in out.split('\n'):
                    self.announce(indent + "  " + line)

    # -- Config file finding/parsing methods ---------------------------

    def find_config_files(self):
        """Find as many configuration files as should be processed for this
        platform, and return a list of filenames in the order in which they
        should be parsed.  The filenames returned are guaranteed to exist
        (modulo nasty race conditions).

        There are three possible config files: distutils.cfg in the
        Distutils installation directory (ie. where the top-level
        Distutils __inst__.py file lives), a file in the user's home
        directory named .pydistutils.cfg on Unix and pydistutils.cfg
        on Windows/Mac; and setup.cfg in the current directory.

        The file in the user's home directory can be disabled with the
        --no-user-cfg option.
        """
        files = []
        check_environ()

        # Where to look for the system-wide Distutils config file
        sys_dir = os.path.dirname(sys.modules['distutils'].__file__)

        # Look for the system config file
        sys_file = os.path.join(sys_dir, "distutils.cfg")
        if os.path.isfile(sys_file):
            files.append(sys_file)

        # What to call the per-user config file
        if os.name == 'posix':
            user_filename = ".pydistutils.cfg"
        else:
            user_filename = "pydistutils.cfg"

        # And look for the user config file
        if self.want_user_cfg:
            user_file = os.path.join(os.path.expanduser('~'), user_filename)
            if os.path.isfile(user_file):
                files.append(user_file)

        # All platforms support local setup.cfg
        local_file = "setup.cfg"
        if os.path.isfile(local_file):
            files.append(local_file)

        if DEBUG:
            self.announce("using config files: %s" % ', '.join(files))

        return files

    def parse_config_files(self, filenames=None):
        from configparser import ConfigParser

        # Ignore install directory options if we have a venv
        if sys.prefix != sys.base_prefix:
            ignore_options = [
                'install-base', 'install-platbase', 'install-lib',
                'install-platlib', 'install-purelib', 'install-headers',
                'install-scripts', 'install-data', 'prefix', 'exec-prefix',
                'home', 'user', 'root']
        else:
            ignore_options = []

        ignore_options = frozenset(ignore_options)

        if filenames is None:
            filenames = self.find_config_files()

        if DEBUG:
            self.announce("Distribution.parse_config_files():")

        parser = ConfigParser()
        for filename in filenames:
            if DEBUG:
                self.announce("  reading %s" % filename)
            parser.read(filename)
            for section in parser.sections():
                options = parser.options(section)
                opt_dict = self.get_option_dict(section)

                for opt in options:
                    if opt != '__name__' and opt not in ignore_options:
                        val = parser.get(section,opt)
                        opt = opt.replace('-', '_')
                        opt_dict[opt] = (filename, val)

            # Make the ConfigParser forget everything (so we retain
            # the original filenames that options come from)
            parser.__init__()

        # If there was a "global" section in the config file, use it
        # to set Distribution options.

        if 'global' in self.command_options:
            for (opt, (src, val)) in self.command_options['global'].items():
                alias = self.negative_opt.get(opt)
                try:
                    if alias:
                        setattr(self, alias, not strtobool(val))
                    elif opt in ('verbose', 'dry_run'): # ugh!
                        setattr(self, opt, strtobool(val))
                    else:
                        setattr(self, opt, val)
                except ValueError as msg:
                    raise DistutilsOptionError(msg)

    # -- Command-line parsing methods ----------------------------------

    def parse_command_line(self):
        """Parse the setup script's command line, taken from the
        'script_args' instance attribute (which defaults to 'sys.argv[1:]'
        -- see 'setup()' in core.py).  This list is first processed for
        "global options" -- options that set attributes of the Distribution
        instance.  Then, it is alternately scanned for Distutils commands
        and options for that command.  Each new command terminates the
        options for the previous command.  The allowed options for a
        command are determined by the 'user_options' attribute of the
        command class -- thus, we have to be able to load command classes
        in order to parse the command line.  Any error in that 'options'
        attribute raises DistutilsGetoptError; any error on the
        command-line raises DistutilsArgError.  If no Distutils commands
        were found on the command line, raises DistutilsArgError.  Return
        true if command-line was successfully parsed and we should carry
        on with executing commands; false if no errors but we shouldn't
        execute commands (currently, this only happens if user asks for
        help).
        """
        #
        # We now have enough information to show the Macintosh dialog
        # that allows the user to interactively specify the "command line".
        #
        toplevel_options = self._get_toplevel_options()

        # We have to parse the command line a bit at a time -- global
        # options, then the first command, then its options, and so on --
        # because each command will be handled by a different class, and
        # the options that are valid for a particular class aren't known
        # until we have loaded the command class, which doesn't happen
        # until we know what the command is.

        self.commands = []
        parser = FancyGetopt(toplevel_options + self.display_options)
        parser.set_negative_aliases(self.negative_opt)
        parser.set_aliases({'licence': 'license'})
        args = parser.getopt(args=self.script_args, object=self)
        option_order = parser.get_option_order()
        log.set_verbosity(self.verbose)

        # for display options we return immediately
        if self.handle_display_options(option_order):
            return
        while args:
            args = self._parse_command_opts(parser, args)
            if args is None:            # user asked for help (and got it)
                return

        # Handle the cases of --help as a "global" option, ie.
        # "setup.py --help" and "setup.py --help command ...".  For the
        # former, we show global options (--verbose, --dry-run, etc.)
        # and display-only options (--name, --version, etc.); for the
        # latter, we omit the display-only options and show help for
        # each command listed on the command line.
        if self.help:
            self._show_help(parser,
                            display_options=len(self.commands) == 0,
                            commands=self.commands)
            return

        # Oops, no commands found -- an end-user error
        if not self.commands:
            raise DistutilsArgError("no commands supplied")

        # All is well: return true
        return True

    def _get_toplevel_options(self):
        """Return the non-display options recognized at the top level.

        This includes options that are recognized *only* at the top
        level as well as options recognized for commands.
        """
        return self.global_options + [
            ("command-packages=", None,
             "list of packages that provide distutils commands"),
            ]

    def _parse_command_opts(self, parser, args):
        """Parse the command-line options for a single command.
        'parser' must be a FancyGetopt instance; 'args' must be the list
        of arguments, starting with the current command (whose options
        we are about to parse).  Returns a new version of 'args' with
        the next command at the front of the list; will be the empty
        list if there are no more commands on the command line.  Returns
        None if the user asked for help on this command.
        """
        # late import because of mutual dependence between these modules
        from distutils.cmd import Command

        # Pull the current command from the head of the command line
        command = args[0]
        if not command_re.match(command):
            raise SystemExit("invalid command name '%s'" % command)
        self.commands.append(command)

        # Dig up the command class that implements this command, so we
        # 1) know that it's a valid command, and 2) know which options
        # it takes.
        try:
            cmd_class = self.get_command_class(command)
        except DistutilsModuleError as msg:
            raise DistutilsArgError(msg)

        # Require that the command class be derived from Command -- want
        # to be sure that the basic "command" interface is implemented.
        if not issubclass(cmd_class, Command):
            raise DistutilsClassError(
                "command class %s must subclass Command" % cmd_class)

        # Also make sure that the command object provides a list of its
        # known options.
        if not (hasattr(cmd_class, 'user_options') and
                isinstance(cmd_class.user_options, list)):
            msg = ("command class %s must provide "
                "'user_options' attribute (a list of tuples)")
            raise DistutilsClassError(msg % cmd_class)

        # If the command class has a list of negative alias options,
        # merge it in with the global negative aliases.
        negative_opt = self.negative_opt
        if hasattr(cmd_class, 'negative_opt'):
            negative_opt = negative_opt.copy()
            negative_opt.update(cmd_class.negative_opt)

        # Check for help_options in command class.  They have a different
        # format (tuple of four) so we need to preprocess them here.
        if (hasattr(cmd_class, 'help_options') and
                isinstance(cmd_class.help_options, list)):
            help_options = fix_help_options(cmd_class.help_options)
        else:
            help_options = []

        # All commands support the global options too, just by adding
        # in 'global_options'.
        parser.set_option_table(self.global_options +
                                cmd_class.user_options +
                                help_options)
        parser.set_negative_aliases(negative_opt)
        (args, opts) = parser.getopt(args[1:])
        if hasattr(opts, 'help') and opts.help:
            self._show_help(parser, display_options=0, commands=[cmd_class])
            return

        if (hasattr(cmd_class, 'help_options') and
                isinstance(cmd_class.help_options, list)):
            help_option_found=0
            for (help_option, short, desc, func) in cmd_class.help_options:
                if hasattr(opts, parser.get_attr_name(help_option)):
                    help_option_found=1
                    if callable(func):
                        func()
                    else:
                        raise DistutilsClassError(
                            "invalid help function %r for help option '%s': "
                            "must be a callable object (function, etc.)"
                            % (func, help_option))

            if help_option_found:
                return

        # Put the options from the command-line into their official
        # holding pen, the 'command_options' dictionary.
        opt_dict = self.get_option_dict(command)
        for (name, value) in vars(opts).items():
            opt_dict[name] = ("command line", value)

        return args

    def finalize_options(self):
        """Set final values for all the options on the Distribution
        instance, analogous to the .finalize_options() method of Command
        objects.
        """
        for attr in ('keywords', 'platforms'):
            value = getattr(self.metadata, attr)
            if value is None:
                continue
            if isinstance(value, str):
                value = [elm.strip() for elm in value.split(',')]
                setattr(self.metadata, attr, value)

    def _show_help(self, parser, global_options=1, display_options=1,
                   commands=[]):
        """Show help for the setup script command-line in the form of
        several lists of command-line options.  'parser' should be a
        FancyGetopt instance; do not expect it to be returned in the
        same state, as its option table will be reset to make it
        generate the correct help text.

        If 'global_options' is true, lists the global options:
        --verbose, --dry-run, etc.  If 'display_options' is true, lists
        the "display-only" options: --name, --version, etc.  Finally,
        lists per-command help for every command name or command class
        in 'commands'.
        """
        # late import because of mutual dependence between these modules
        from distutils.core import gen_usage
        from distutils.cmd import Command

        if global_options:
            if display_options:
                options = self._get_toplevel_options()
            else:
                options = self.global_options
            parser.set_option_table(options)
            parser.print_help(self.common_usage + "\nGlobal options:")
            print('')

        if display_options:
            parser.set_option_table(self.display_options)
            parser.print_help(
                "Information display options (just display " +
                "information, ignore any commands)")
            print('')

        for command in self.commands:
            if isinstance(command, type) and issubclass(command, Command):
                klass = command
            else:
                klass = self.get_command_class(command)
            if (hasattr(klass, 'help_options') and
                    isinstance(klass.help_options, list)):
                parser.set_option_table(klass.user_options +
                                        fix_help_options(klass.help_options))
            else:
                parser.set_option_table(klass.user_options)
            parser.print_help("Options for '%s' command:" % klass.__name__)
            print('')

        print(gen_usage(self.script_name))

    def handle_display_options(self, option_order):
        """If there were any non-global "display-only" options
        (--help-commands or the metadata display options) on the command
        line, display the requested info and return true; else return
        false.
        """
        from distutils.core import gen_usage

        # User just wants a list of commands -- we'll print it out and stop
        # processing now (ie. if they ran "setup --help-commands foo bar",
        # we ignore "foo bar").
        if self.help_commands:
            self.print_commands()
            print('')
            print(gen_usage(self.script_name))
            return 1

        # If user supplied any of the "display metadata" options, then
        # display that metadata in the order in which the user supplied the
        # metadata options.
        any_display_options = 0
        is_display_option = {}
        for option in self.display_options:
            is_display_option[option[0]] = 1

        for (opt, val) in option_order:
            if val and is_display_option.get(opt):
                opt = translate_longopt(opt)
                value = getattr(self.metadata, "get_"+opt)()
                if opt in ['keywords', 'platforms']:
                    print(','.join(value))
                elif opt in ('classifiers', 'provides', 'requires',
                             'obsoletes'):
                    print('\n'.join(value))
                else:
                    print(value)
                any_display_options = 1

        return any_display_options

    def print_command_list(self, commands, header, max_length):
        """Print a subset of the list of all commands -- used by
        'print_commands()'.
        """
        print(header + ":")

        for cmd in commands:
            klass = self.cmdclass.get(cmd)
            if not klass:
                klass = self.get_command_class(cmd)
            try:
                description = klass.description
            except AttributeError:
                description = "(no description available)"

            print("  %-*s  %s" % (max_length, cmd, description))

    def print_commands(self):
        """Print out a help message listing all available commands with a
        description of each.  The list is divided into "standard commands"
        (listed in distutils.command.__all__) and "extra commands"
        (mentioned in self.cmdclass, but not a standard command).  The
        descriptions come from the command class attribute
        'description'.
        """
        import distutils.command
        std_commands = distutils.command.__all__
        is_std = {}
        for cmd in std_commands:
            is_std[cmd] = 1

        extra_commands = []
        for cmd in self.cmdclass.keys():
            if not is_std.get(cmd):
                extra_commands.append(cmd)

        max_length = 0
        for cmd in (std_commands + extra_commands):
            if len(cmd) > max_length:
                max_length = len(cmd)

        self.print_command_list(std_commands,
                                "Standard commands",
                                max_length)
        if extra_commands:
            print()
            self.print_command_list(extra_commands,
                                    "Extra commands",
                                    max_length)

    def get_command_list(self):
        """Get a list of (command, description) tuples.
        The list is divided into "standard commands" (listed in
        distutils.command.__all__) and "extra commands" (mentioned in
        self.cmdclass, but not a standard command).  The descriptions come
        from the command class attribute 'description'.
        """
        # Currently this is only used on Mac OS, for the Mac-only GUI
        # Distutils interface (by Jack Jansen)
        import distutils.command
        std_commands = distutils.command.__all__
        is_std = {}
        for cmd in std_commands:
            is_std[cmd] = 1

        extra_commands = []
        for cmd in self.cmdclass.keys():
            if not is_std.get(cmd):
                extra_commands.append(cmd)

        rv = []
        for cmd in (std_commands + extra_commands):
            klass = self.cmdclass.get(cmd)
            if not klass:
                klass = self.get_command_class(cmd)
            try:
                description = klass.description
            except AttributeError:
                description = "(no description available)"
            rv.append((cmd, description))
        return rv

    # -- Command class/object methods ----------------------------------

    def get_command_packages(self):
        """Return a list of packages from which commands are loaded."""
        pkgs = self.command_packages
        if not isinstance(pkgs, list):
            if pkgs is None:
                pkgs = ''
            pkgs = [pkg.strip() for pkg in pkgs.split(',') if pkg != '']
            if "distutils.command" not in pkgs:
                pkgs.insert(0, "distutils.command")
            self.command_packages = pkgs
        return pkgs

    def get_command_class(self, command):
        """Return the class that implements the Distutils command named by
        'command'.  First we check the 'cmdclass' dictionary; if the
        command is mentioned there, we fetch the class object from the
        dictionary and return it.  Otherwise we load the command module
        ("distutils.command." + command) and fetch the command class from
        the module.  The loaded class is also stored in 'cmdclass'
        to speed future calls to 'get_command_class()'.

        Raises DistutilsModuleError if the expected module could not be
        found, or if that module does not define the expected class.
        """
        klass = self.cmdclass.get(command)
        if klass:
            return klass

        for pkgname in self.get_command_packages():
            module_name = "%s.%s" % (pkgname, command)
            klass_name = command

            try:
                __import__(module_name)
                module = sys.modules[module_name]
            except ImportError:
                continue

            try:
                klass = getattr(module, klass_name)
            except AttributeError:
                raise DistutilsModuleError(
                    "invalid command '%s' (no class '%s' in module '%s')"
                    % (command, klass_name, module_name))

            self.cmdclass[command] = klass
            return klass

        raise DistutilsModuleError("invalid command '%s'" % command)

    def get_command_obj(self, command, create=1):
        """Return the command object for 'command'.  Normally this object
        is cached on a previous call to 'get_command_obj()'; if no command
        object for 'command' is in the cache, then we either create and
        return it (if 'create' is true) or return None.
        """
        cmd_obj = self.command_obj.get(command)
        if not cmd_obj and create:
            if DEBUG:
                self.announce("Distribution.get_command_obj(): "
                              "creating '%s' command object" % command)

            klass = self.get_command_class(command)
            cmd_obj = self.command_obj[command] = klass(self)
            self.have_run[command] = 0

            # Set any options that were supplied in config files
            # or on the command line.  (NB. support for error
            # reporting is lame here: any errors aren't reported
            # until 'finalize_options()' is called, which means
            # we won't report the source of the error.)
            options = self.command_options.get(command)
            if options:
                self._set_command_options(cmd_obj, options)

        return cmd_obj

    def _set_command_options(self, command_obj, option_dict=None):
        """Set the options for 'command_obj' from 'option_dict'.  Basically
        this means copying elements of a dictionary ('option_dict') to
        attributes of an instance ('command').

        'command_obj' must be a Command instance.  If 'option_dict' is not
        supplied, uses the standard option dictionary for this command
        (from 'self.command_options').
        """
        command_name = command_obj.get_command_name()
        if option_dict is None:
            option_dict = self.get_option_dict(command_name)

        if DEBUG:
            self.announce("  setting options for '%s' command:" % command_name)
        for (option, (source, value)) in option_dict.items():
            if DEBUG:
                self.announce("    %s = %s (from %s)" % (option, value,
                                                         source))
            try:
                bool_opts = [translate_longopt(o)
                             for o in command_obj.boolean_options]
            except AttributeError:
                bool_opts = []
            try:
                neg_opt = command_obj.negative_opt
            except AttributeError:
                neg_opt = {}

            try:
                is_string = isinstance(value, str)
                if option in neg_opt and is_string:
                    setattr(command_obj, neg_opt[option], not strtobool(value))
                elif option in bool_opts and is_string:
                    setattr(command_obj, option, strtobool(value))
                elif hasattr(command_obj, option):
                    setattr(command_obj, option, value)
                else:
                    raise DistutilsOptionError(
                        "error in %s: command '%s' has no such option '%s'"
                        % (source, command_name, option))
            except ValueError as msg:
                raise DistutilsOptionError(msg)

    def reinitialize_command(self, command, reinit_subcommands=0):
        """Reinitializes a command to the state it was in when first
        returned by 'get_command_obj()': ie., initialized but not yet
        finalized.  This provides the opportunity to sneak option
        values in programmatically, overriding or supplementing
        user-supplied values from the config files and command line.
        You'll have to re-finalize the command object (by calling
        'finalize_options()' or 'ensure_finalized()') before using it for
        real.

        'command' should be a command name (string) or command object.  If
        'reinit_subcommands' is true, also reinitializes the command's
        sub-commands, as declared by the 'sub_commands' class attribute (if
        it has one).  See the "install" command for an example.  Only
        reinitializes the sub-commands that actually matter, ie. those
        whose test predicates return true.

        Returns the reinitialized command object.
        """
        from distutils.cmd import Command
        if not isinstance(command, Command):
            command_name = command
            command = self.get_command_obj(command_name)
        else:
            command_name = command.get_command_name()

        if not command.finalized:
            return command
        command.initialize_options()
        command.finalized = 0
        self.have_run[command_name] = 0
        self._set_command_options(command)

        if reinit_subcommands:
            for sub in command.get_sub_commands():
                self.reinitialize_command(sub, reinit_subcommands)

        return command

    # -- Methods that operate on the Distribution ----------------------

    def announce(self, msg, level=log.INFO):
        log.log(level, msg)

    def run_commands(self):
        """Run each command that was seen on the setup script command line.
        Uses the list of commands found and cache of command objects
        created by 'get_command_obj()'.
        """
        for cmd in self.commands:
            self.run_command(cmd)

    # -- Methods that operate on its Commands --------------------------

    def run_command(self, command):
        """Do whatever it takes to run a command (including nothing at all,
        if the command has already been run).  Specifically: if we have
        already created and run the command named by 'command', return
        silently without doing anything.  If the command named by 'command'
        doesn't even have a command object yet, create one.  Then invoke
        'run()' on that command object (or an existing one).
        """
        # Already been here, done that? then return silently.
        if self.have_run.get(command):
            return

        log.info("running %s", command)
        cmd_obj = self.get_command_obj(command)
        cmd_obj.ensure_finalized()
        cmd_obj.run()
        self.have_run[command] = 1

    # -- Distribution query methods ------------------------------------

    def has_pure_modules(self):
        return len(self.packages or self.py_modules or []) > 0

    def has_ext_modules(self):
        return self.ext_modules and len(self.ext_modules) > 0

    def has_c_libraries(self):
        return self.libraries and len(self.libraries) > 0

    def has_modules(self):
        return self.has_pure_modules() or self.has_ext_modules()

    def has_headers(self):
        return self.headers and len(self.headers) > 0

    def has_scripts(self):
        return self.scripts and len(self.scripts) > 0

    def has_data_files(self):
        return self.data_files and len(self.data_files) > 0

    def is_pure(self):
        return (self.has_pure_modules() and
                not self.has_ext_modules() and
                not self.has_c_libraries())

    # -- Metadata query methods ----------------------------------------

    # If you're looking for 'get_name()', 'get_version()', and so forth,
    # they are defined in a sneaky way: the constructor binds self.get_XXX
    # to self.metadata.get_XXX.  The actual code is in the
    # DistributionMetadata class, below.

class DistributionMetadata:
    """Dummy class to hold the distribution meta-data: name, version,
    author, and so forth.
    """

    _METHOD_BASENAMES = ("name", "version", "author", "author_email",
                         "maintainer", "maintainer_email", "url",
                         "license", "description", "long_description",
                         "keywords", "platforms", "fullname", "contact",
                         "contact_email", "classifiers", "download_url",
                         # PEP 314
                         "provides", "requires", "obsoletes",
                         )

    def __init__(self, path=None):
        if path is not None:
            self.read_pkg_file(open(path))
        else:
            self.name = None
            self.version = None
            self.author = None
            self.author_email = None
            self.maintainer = None
            self.maintainer_email = None
            self.url = None
            self.license = None
            self.description = None
            self.long_description = None
            self.keywords = None
            self.platforms = None
            self.classifiers = None
            self.download_url = None
            # PEP 314
            self.provides = None
            self.requires = None
            self.obsoletes = None

    def read_pkg_file(self, file):
        """Reads the metadata values from a file object."""
        msg = message_from_file(file)

        def _read_field(name):
            value = msg[name]
            if value == 'UNKNOWN':
                return None
            return value

        def _read_list(name):
            values = msg.get_all(name, None)
            if values == []:
                return None
            return values

        metadata_version = msg['metadata-version']
        self.name = _read_field('name')
        self.version = _read_field('version')
        self.description = _read_field('summary')
        # we are filling author only.
        self.author = _read_field('author')
        self.maintainer = None
        self.author_email = _read_field('author-email')
        self.maintainer_email = None
        self.url = _read_field('home-page')
        self.license = _read_field('license')

        if 'download-url' in msg:
            self.download_url = _read_field('download-url')
        else:
            self.download_url = None

        self.long_description = _read_field('description')
        self.description = _read_field('summary')

        if 'keywords' in msg:
            self.keywords = _read_field('keywords').split(',')

        self.platforms = _read_list('platform')
        self.classifiers = _read_list('classifier')

        # PEP 314 - these fields only exist in 1.1
        if metadata_version == '1.1':
            self.requires = _read_list('requires')
            self.provides = _read_list('provides')
            self.obsoletes = _read_list('obsoletes')
        else:
            self.requires = None
            self.provides = None
            self.obsoletes = None

    def write_pkg_info(self, base_dir):
        """Write the PKG-INFO file into the release tree.
        """
        with open(os.path.join(base_dir, 'PKG-INFO'), 'w',
                  encoding='UTF-8') as pkg_info:
            self.write_pkg_file(pkg_info)

    def write_pkg_file(self, file):
        """Write the PKG-INFO format data to a file object.
        """
        version = '1.0'
        if (self.provides or self.requires or self.obsoletes or
                self.classifiers or self.download_url):
            version = '1.1'

        file.write('Metadata-Version: %s\n' % version)
        file.write('Name: %s\n' % self.get_name())
        file.write('Version: %s\n' % self.get_version())
        file.write('Summary: %s\n' % self.get_description())
        file.write('Home-page: %s\n' % self.get_url())
        file.write('Author: %s\n' % self.get_contact())
        file.write('Author-email: %s\n' % self.get_contact_email())
        file.write('License: %s\n' % self.get_license())
        if self.download_url:
            file.write('Download-URL: %s\n' % self.download_url)

        long_desc = rfc822_escape(self.get_long_description())
        file.write('Description: %s\n' % long_desc)

        keywords = ','.join(self.get_keywords())
        if keywords:
            file.write('Keywords: %s\n' % keywords)

        self._write_list(file, 'Platform', self.get_platforms())
        self._write_list(file, 'Classifier', self.get_classifiers())

        # PEP 314
        self._write_list(file, 'Requires', self.get_requires())
        self._write_list(file, 'Provides', self.get_provides())
        self._write_list(file, 'Obsoletes', self.get_obsoletes())

    def _write_list(self, file, name, values):
        for value in values:
            file.write('%s: %s\n' % (name, value))

    # -- Metadata query methods ----------------------------------------

    def get_name(self):
        return self.name or "UNKNOWN"

    def get_version(self):
        return self.version or "0.0.0"

    def get_fullname(self):
        return "%s-%s" % (self.get_name(), self.get_version())

    def get_author(self):
        return self.author or "UNKNOWN"

    def get_author_email(self):
        return self.author_email or "UNKNOWN"

    def get_maintainer(self):
        return self.maintainer or "UNKNOWN"

    def get_maintainer_email(self):
        return self.maintainer_email or "UNKNOWN"

    def get_contact(self):
        return self.maintainer or self.author or "UNKNOWN"

    def get_contact_email(self):
        return self.maintainer_email or self.author_email or "UNKNOWN"

    def get_url(self):
        return self.url or "UNKNOWN"

    def get_license(self):
        return self.license or "UNKNOWN"
    get_licence = get_license

    def get_description(self):
        return self.description or "UNKNOWN"

    def get_long_description(self):
        return self.long_description or "UNKNOWN"

    def get_keywords(self):
        return self.keywords or []

    def set_keywords(self, value):
        self.keywords = _ensure_list(value, 'keywords')

    def get_platforms(self):
        return self.platforms or ["UNKNOWN"]

    def set_platforms(self, value):
        self.platforms = _ensure_list(value, 'platforms')

    def get_classifiers(self):
        return self.classifiers or []

    def set_classifiers(self, value):
        self.classifiers = _ensure_list(value, 'classifiers')

    def get_download_url(self):
        return self.download_url or "UNKNOWN"

    # PEP 314
    def get_requires(self):
        return self.requires or []

    def set_requires(self, value):
        import distutils.versionpredicate
        for v in value:
            distutils.versionpredicate.VersionPredicate(v)
        self.requires = list(value)

    def get_provides(self):
        return self.provides or []

    def set_provides(self, value):
        value = [v.strip() for v in value]
        for v in value:
            import distutils.versionpredicate
            distutils.versionpredicate.split_provision(v)
        self.provides = value

    def get_obsoletes(self):
        return self.obsoletes or []

    def set_obsoletes(self, value):
        import distutils.versionpredicate
        for v in value:
            distutils.versionpredicate.VersionPredicate(v)
        self.obsoletes = list(value)

def fix_help_options(options):
    """Convert a 4-tuple 'help_options' list as found in various command
    classes to the 3-tuple form required by FancyGetopt.
    """
    new_options = []
    for help_tuple in options:
        new_options.append(help_tuple[0:3])
    return new_options
site-packages/setuptools/_distutils/config.py000064400000011333150564745010015517 0ustar00"""distutils.pypirc

Provides the PyPIRCCommand class, the base class for the command classes
that uses .pypirc in the distutils.command package.
"""
import os
from configparser import RawConfigParser

from distutils.cmd import Command

DEFAULT_PYPIRC = """\
[distutils]
index-servers =
    pypi

[pypi]
username:%s
password:%s
"""

class PyPIRCCommand(Command):
    """Base command that knows how to handle the .pypirc file
    """
    DEFAULT_REPOSITORY = 'https://upload.pypi.org/legacy/'
    DEFAULT_REALM = 'pypi'
    repository = None
    realm = None

    user_options = [
        ('repository=', 'r',
         "url of repository [default: %s]" % \
            DEFAULT_REPOSITORY),
        ('show-response', None,
         'display full response text from server')]

    boolean_options = ['show-response']

    def _get_rc_file(self):
        """Returns rc file path."""
        return os.path.join(os.path.expanduser('~'), '.pypirc')

    def _store_pypirc(self, username, password):
        """Creates a default .pypirc file."""
        rc = self._get_rc_file()
        with os.fdopen(os.open(rc, os.O_CREAT | os.O_WRONLY, 0o600), 'w') as f:
            f.write(DEFAULT_PYPIRC % (username, password))

    def _read_pypirc(self):
        """Reads the .pypirc file."""
        rc = self._get_rc_file()
        if os.path.exists(rc):
            self.announce('Using PyPI login from %s' % rc)
            repository = self.repository or self.DEFAULT_REPOSITORY

            config = RawConfigParser()
            config.read(rc)
            sections = config.sections()
            if 'distutils' in sections:
                # let's get the list of servers
                index_servers = config.get('distutils', 'index-servers')
                _servers = [server.strip() for server in
                            index_servers.split('\n')
                            if server.strip() != '']
                if _servers == []:
                    # nothing set, let's try to get the default pypi
                    if 'pypi' in sections:
                        _servers = ['pypi']
                    else:
                        # the file is not properly defined, returning
                        # an empty dict
                        return {}
                for server in _servers:
                    current = {'server': server}
                    current['username'] = config.get(server, 'username')

                    # optional params
                    for key, default in (('repository',
                                          self.DEFAULT_REPOSITORY),
                                         ('realm', self.DEFAULT_REALM),
                                         ('password', None)):
                        if config.has_option(server, key):
                            current[key] = config.get(server, key)
                        else:
                            current[key] = default

                    # work around people having "repository" for the "pypi"
                    # section of their config set to the HTTP (rather than
                    # HTTPS) URL
                    if (server == 'pypi' and
                        repository in (self.DEFAULT_REPOSITORY, 'pypi')):
                        current['repository'] = self.DEFAULT_REPOSITORY
                        return current

                    if (current['server'] == repository or
                        current['repository'] == repository):
                        return current
            elif 'server-login' in sections:
                # old format
                server = 'server-login'
                if config.has_option(server, 'repository'):
                    repository = config.get(server, 'repository')
                else:
                    repository = self.DEFAULT_REPOSITORY
                return {'username': config.get(server, 'username'),
                        'password': config.get(server, 'password'),
                        'repository': repository,
                        'server': server,
                        'realm': self.DEFAULT_REALM}

        return {}

    def _read_pypi_response(self, response):
        """Read and decode a PyPI HTTP response."""
        import cgi
        content_type = response.getheader('content-type', 'text/plain')
        encoding = cgi.parse_header(content_type)[1].get('charset', 'ascii')
        return response.read().decode(encoding)

    def initialize_options(self):
        """Initialize options."""
        self.repository = None
        self.realm = None
        self.show_response = 0

    def finalize_options(self):
        """Finalizes options."""
        if self.repository is None:
            self.repository = self.DEFAULT_REPOSITORY
        if self.realm is None:
            self.realm = self.DEFAULT_REALM
site-packages/setuptools/_distutils/cygwinccompiler.py000064400000041052150564745010017451 0ustar00"""distutils.cygwinccompiler

Provides the CygwinCCompiler class, a subclass of UnixCCompiler that
handles the Cygwin port of the GNU C compiler to Windows.  It also contains
the Mingw32CCompiler class which handles the mingw32 port of GCC (same as
cygwin in no-cygwin mode).
"""

# problems:
#
# * if you use a msvc compiled python version (1.5.2)
#   1. you have to insert a __GNUC__ section in its config.h
#   2. you have to generate an import library for its dll
#      - create a def-file for python??.dll
#      - create an import library using
#             dlltool --dllname python15.dll --def python15.def \
#                       --output-lib libpython15.a
#
#   see also http://starship.python.net/crew/kernr/mingw32/Notes.html
#
# * We put export_symbols in a def-file, and don't use
#   --export-all-symbols because it doesn't worked reliable in some
#   tested configurations. And because other windows compilers also
#   need their symbols specified this no serious problem.
#
# tested configurations:
#
# * cygwin gcc 2.91.57/ld 2.9.4/dllwrap 0.2.4 works
#   (after patching python's config.h and for C++ some other include files)
#   see also http://starship.python.net/crew/kernr/mingw32/Notes.html
# * mingw32 gcc 2.95.2/ld 2.9.4/dllwrap 0.2.4 works
#   (ld doesn't support -shared, so we use dllwrap)
# * cygwin gcc 2.95.2/ld 2.10.90/dllwrap 2.10.90 works now
#   - its dllwrap doesn't work, there is a bug in binutils 2.10.90
#     see also http://sources.redhat.com/ml/cygwin/2000-06/msg01274.html
#   - using gcc -mdll instead dllwrap doesn't work without -static because
#     it tries to link against dlls instead their import libraries. (If
#     it finds the dll first.)
#     By specifying -static we force ld to link against the import libraries,
#     this is windows standard and there are normally not the necessary symbols
#     in the dlls.
#   *** only the version of June 2000 shows these problems
# * cygwin gcc 3.2/ld 2.13.90 works
#   (ld supports -shared)
# * mingw gcc 3.2/ld 2.13 works
#   (ld supports -shared)
# * llvm-mingw with Clang 11 works
#   (lld supports -shared)

import os
import sys
import copy
from subprocess import Popen, PIPE, check_output
import re

from distutils.unixccompiler import UnixCCompiler
from distutils.file_util import write_file
from distutils.errors import (DistutilsExecError, CCompilerError,
        CompileError, UnknownFileError)
from distutils.version import LooseVersion
from distutils.spawn import find_executable

def get_msvcr():
    """Include the appropriate MSVC runtime library if Python was built
    with MSVC 7.0 or later.
    """
    msc_pos = sys.version.find('MSC v.')
    if msc_pos != -1:
        msc_ver = sys.version[msc_pos+6:msc_pos+10]
        if msc_ver == '1300':
            # MSVC 7.0
            return ['msvcr70']
        elif msc_ver == '1310':
            # MSVC 7.1
            return ['msvcr71']
        elif msc_ver == '1400':
            # VS2005 / MSVC 8.0
            return ['msvcr80']
        elif msc_ver == '1500':
            # VS2008 / MSVC 9.0
            return ['msvcr90']
        elif msc_ver == '1600':
            # VS2010 / MSVC 10.0
            return ['msvcr100']
        else:
            raise ValueError("Unknown MS Compiler version %s " % msc_ver)


class CygwinCCompiler(UnixCCompiler):
    """ Handles the Cygwin port of the GNU C compiler to Windows.
    """
    compiler_type = 'cygwin'
    obj_extension = ".o"
    static_lib_extension = ".a"
    shared_lib_extension = ".dll"
    static_lib_format = "lib%s%s"
    shared_lib_format = "%s%s"
    exe_extension = ".exe"

    def __init__(self, verbose=0, dry_run=0, force=0):

        UnixCCompiler.__init__(self, verbose, dry_run, force)

        status, details = check_config_h()
        self.debug_print("Python's GCC status: %s (details: %s)" %
                         (status, details))
        if status is not CONFIG_H_OK:
            self.warn(
                "Python's pyconfig.h doesn't seem to support your compiler. "
                "Reason: %s. "
                "Compiling may fail because of undefined preprocessor macros."
                % details)

        self.cc = os.environ.get('CC', 'gcc')
        self.cxx = os.environ.get('CXX', 'g++')

        if ('gcc' in self.cc): # Start gcc workaround
            self.gcc_version, self.ld_version, self.dllwrap_version = \
                get_versions()
            self.debug_print(self.compiler_type + ": gcc %s, ld %s, dllwrap %s\n" %
                             (self.gcc_version,
                              self.ld_version,
                              self.dllwrap_version) )

            # ld_version >= "2.10.90" and < "2.13" should also be able to use
            # gcc -mdll instead of dllwrap
            # Older dllwraps had own version numbers, newer ones use the
            # same as the rest of binutils ( also ld )
            # dllwrap 2.10.90 is buggy
            if self.ld_version >= "2.10.90":
                self.linker_dll = self.cc
            else:
                self.linker_dll = "dllwrap"

            # ld_version >= "2.13" support -shared so use it instead of
            # -mdll -static
            if self.ld_version >= "2.13":
                shared_option = "-shared"
            else:
                shared_option = "-mdll -static"
        else: # Assume linker is up to date
            self.linker_dll = self.cc
            shared_option = "-shared"

        self.set_executables(compiler='%s -mcygwin -O -Wall' % self.cc,
                             compiler_so='%s -mcygwin -mdll -O -Wall' % self.cc,
                             compiler_cxx='%s -mcygwin -O -Wall' % self.cxx,
                             linker_exe='%s -mcygwin' % self.cc,
                             linker_so=('%s -mcygwin %s' %
                                        (self.linker_dll, shared_option)))

        # cygwin and mingw32 need different sets of libraries
        if ('gcc' in self.cc and self.gcc_version == "2.91.57"):
            # cygwin shouldn't need msvcrt, but without the dlls will crash
            # (gcc version 2.91.57) -- perhaps something about initialization
            self.dll_libraries=["msvcrt"]
            self.warn(
                "Consider upgrading to a newer version of gcc")
        else:
            # Include the appropriate MSVC runtime library if Python was built
            # with MSVC 7.0 or later.
            self.dll_libraries = get_msvcr()

    def _compile(self, obj, src, ext, cc_args, extra_postargs, pp_opts):
        """Compiles the source by spawning GCC and windres if needed."""
        if ext == '.rc' or ext == '.res':
            # gcc needs '.res' and '.rc' compiled to object files !!!
            try:
                self.spawn(["windres", "-i", src, "-o", obj])
            except DistutilsExecError as msg:
                raise CompileError(msg)
        else: # for other files use the C-compiler
            try:
                self.spawn(self.compiler_so + cc_args + [src, '-o', obj] +
                           extra_postargs)
            except DistutilsExecError as msg:
                raise CompileError(msg)

    def link(self, target_desc, objects, output_filename, output_dir=None,
             libraries=None, library_dirs=None, runtime_library_dirs=None,
             export_symbols=None, debug=0, extra_preargs=None,
             extra_postargs=None, build_temp=None, target_lang=None):
        """Link the objects."""
        # use separate copies, so we can modify the lists
        extra_preargs = copy.copy(extra_preargs or [])
        libraries = copy.copy(libraries or [])
        objects = copy.copy(objects or [])

        # Additional libraries
        libraries.extend(self.dll_libraries)

        # handle export symbols by creating a def-file
        # with executables this only works with gcc/ld as linker
        if ((export_symbols is not None) and
            (target_desc != self.EXECUTABLE or self.linker_dll == "gcc")):
            # (The linker doesn't do anything if output is up-to-date.
            # So it would probably better to check if we really need this,
            # but for this we had to insert some unchanged parts of
            # UnixCCompiler, and this is not what we want.)

            # we want to put some files in the same directory as the
            # object files are, build_temp doesn't help much
            # where are the object files
            temp_dir = os.path.dirname(objects[0])
            # name of dll to give the helper files the same base name
            (dll_name, dll_extension) = os.path.splitext(
                os.path.basename(output_filename))

            # generate the filenames for these files
            def_file = os.path.join(temp_dir, dll_name + ".def")
            lib_file = os.path.join(temp_dir, 'lib' + dll_name + ".a")

            # Generate .def file
            contents = [
                "LIBRARY %s" % os.path.basename(output_filename),
                "EXPORTS"]
            for sym in export_symbols:
                contents.append(sym)
            self.execute(write_file, (def_file, contents),
                         "writing %s" % def_file)

            # next add options for def-file and to creating import libraries

            # dllwrap uses different options than gcc/ld
            if self.linker_dll == "dllwrap":
                extra_preargs.extend(["--output-lib", lib_file])
                # for dllwrap we have to use a special option
                extra_preargs.extend(["--def", def_file])
            # we use gcc/ld here and can be sure ld is >= 2.9.10
            else:
                # doesn't work: bfd_close build\...\libfoo.a: Invalid operation
                #extra_preargs.extend(["-Wl,--out-implib,%s" % lib_file])
                # for gcc/ld the def-file is specified as any object files
                objects.append(def_file)

        #end: if ((export_symbols is not None) and
        #        (target_desc != self.EXECUTABLE or self.linker_dll == "gcc")):

        # who wants symbols and a many times larger output file
        # should explicitly switch the debug mode on
        # otherwise we let dllwrap/ld strip the output file
        # (On my machine: 10KiB < stripped_file < ??100KiB
        #   unstripped_file = stripped_file + XXX KiB
        #  ( XXX=254 for a typical python extension))
        if not debug:
            extra_preargs.append("-s")

        UnixCCompiler.link(self, target_desc, objects, output_filename,
                           output_dir, libraries, library_dirs,
                           runtime_library_dirs,
                           None, # export_symbols, we do this in our def-file
                           debug, extra_preargs, extra_postargs, build_temp,
                           target_lang)

    # -- Miscellaneous methods -----------------------------------------

    def object_filenames(self, source_filenames, strip_dir=0, output_dir=''):
        """Adds supports for rc and res files."""
        if output_dir is None:
            output_dir = ''
        obj_names = []
        for src_name in source_filenames:
            # use normcase to make sure '.rc' is really '.rc' and not '.RC'
            base, ext = os.path.splitext(os.path.normcase(src_name))
            if ext not in (self.src_extensions + ['.rc','.res']):
                raise UnknownFileError("unknown file type '%s' (from '%s')" % \
                      (ext, src_name))
            if strip_dir:
                base = os.path.basename (base)
            if ext in ('.res', '.rc'):
                # these need to be compiled to object files
                obj_names.append (os.path.join(output_dir,
                                              base + ext + self.obj_extension))
            else:
                obj_names.append (os.path.join(output_dir,
                                               base + self.obj_extension))
        return obj_names

# the same as cygwin plus some additional parameters
class Mingw32CCompiler(CygwinCCompiler):
    """ Handles the Mingw32 port of the GNU C compiler to Windows.
    """
    compiler_type = 'mingw32'

    def __init__(self, verbose=0, dry_run=0, force=0):

        CygwinCCompiler.__init__ (self, verbose, dry_run, force)

        # ld_version >= "2.13" support -shared so use it instead of
        # -mdll -static
        if ('gcc' in self.cc and self.ld_version < "2.13"):
            shared_option = "-mdll -static"
        else:
            shared_option = "-shared"

        # A real mingw32 doesn't need to specify a different entry point,
        # but cygwin 2.91.57 in no-cygwin-mode needs it.
        if ('gcc' in self.cc and self.gcc_version <= "2.91.57"):
            entry_point = '--entry _DllMain@12'
        else:
            entry_point = ''

        if is_cygwincc(self.cc):
            raise CCompilerError(
                'Cygwin gcc cannot be used with --compiler=mingw32')

        self.set_executables(compiler='%s -O -Wall' % self.cc,
                             compiler_so='%s -mdll -O -Wall' % self.cc,
                             compiler_cxx='%s -O -Wall' % self.cxx,
                             linker_exe='%s' % self.cc,
                             linker_so='%s %s %s'
                                        % (self.linker_dll, shared_option,
                                           entry_point))
        # Maybe we should also append -mthreads, but then the finished
        # dlls need another dll (mingwm10.dll see Mingw32 docs)
        # (-mthreads: Support thread-safe exception handling on `Mingw32')

        # no additional libraries needed
        self.dll_libraries=[]

        # Include the appropriate MSVC runtime library if Python was built
        # with MSVC 7.0 or later.
        self.dll_libraries = get_msvcr()

# Because these compilers aren't configured in Python's pyconfig.h file by
# default, we should at least warn the user if he is using an unmodified
# version.

CONFIG_H_OK = "ok"
CONFIG_H_NOTOK = "not ok"
CONFIG_H_UNCERTAIN = "uncertain"

def check_config_h():
    """Check if the current Python installation appears amenable to building
    extensions with GCC.

    Returns a tuple (status, details), where 'status' is one of the following
    constants:

    - CONFIG_H_OK: all is well, go ahead and compile
    - CONFIG_H_NOTOK: doesn't look good
    - CONFIG_H_UNCERTAIN: not sure -- unable to read pyconfig.h

    'details' is a human-readable string explaining the situation.

    Note there are two ways to conclude "OK": either 'sys.version' contains
    the string "GCC" (implying that this Python was built with GCC), or the
    installed "pyconfig.h" contains the string "__GNUC__".
    """

    # XXX since this function also checks sys.version, it's not strictly a
    # "pyconfig.h" check -- should probably be renamed...

    from distutils import sysconfig

    # if sys.version contains GCC then python was compiled with GCC, and the
    # pyconfig.h file should be OK
    if "GCC" in sys.version:
        return CONFIG_H_OK, "sys.version mentions 'GCC'"

    # Clang would also work
    if "Clang" in sys.version:
        return CONFIG_H_OK, "sys.version mentions 'Clang'"

    # let's see if __GNUC__ is mentioned in python.h
    fn = sysconfig.get_config_h_filename()
    try:
        config_h = open(fn)
        try:
            if "__GNUC__" in config_h.read():
                return CONFIG_H_OK, "'%s' mentions '__GNUC__'" % fn
            else:
                return CONFIG_H_NOTOK, "'%s' does not mention '__GNUC__'" % fn
        finally:
            config_h.close()
    except OSError as exc:
        return (CONFIG_H_UNCERTAIN,
                "couldn't read '%s': %s" % (fn, exc.strerror))

RE_VERSION = re.compile(br'(\d+\.\d+(\.\d+)*)')

def _find_exe_version(cmd):
    """Find the version of an executable by running `cmd` in the shell.

    If the command is not found, or the output does not match
    `RE_VERSION`, returns None.
    """
    executable = cmd.split()[0]
    if find_executable(executable) is None:
        return None
    out = Popen(cmd, shell=True, stdout=PIPE).stdout
    try:
        out_string = out.read()
    finally:
        out.close()
    result = RE_VERSION.search(out_string)
    if result is None:
        return None
    # LooseVersion works with strings
    # so we need to decode our bytes
    return LooseVersion(result.group(1).decode())

def get_versions():
    """ Try to find out the versions of gcc, ld and dllwrap.

    If not possible it returns None for it.
    """
    commands = ['gcc -dumpversion', 'ld -v', 'dllwrap --version']
    return tuple([_find_exe_version(cmd) for cmd in commands])

def is_cygwincc(cc):
    '''Try to determine if the compiler that would be used is from cygwin.'''
    out_string = check_output([cc, '-dumpmachine'])
    return out_string.strip().endswith(b'cygwin')
site-packages/setuptools/_distutils/sysconfig.py000064400000052176150564745010016270 0ustar00"""Provide access to Python's configuration information.  The specific
configuration variables available depend heavily on the platform and
configuration.  The values may be retrieved using
get_config_var(name), and the list of variables is available via
get_config_vars().keys().  Additional convenience functions are also
available.

Written by:   Fred L. Drake, Jr.
Email:        <fdrake@acm.org>
"""

import _imp
import os
import re
import sys

from .errors import DistutilsPlatformError

IS_PYPY = '__pypy__' in sys.builtin_module_names

# These are needed in a couple of spots, so just compute them once.
PREFIX = os.path.normpath(sys.prefix)
EXEC_PREFIX = os.path.normpath(sys.exec_prefix)
BASE_PREFIX = os.path.normpath(sys.base_prefix)
BASE_EXEC_PREFIX = os.path.normpath(sys.base_exec_prefix)

# Path to the base directory of the project. On Windows the binary may
# live in project/PCbuild/win32 or project/PCbuild/amd64.
# set for cross builds
if "_PYTHON_PROJECT_BASE" in os.environ:
    project_base = os.path.abspath(os.environ["_PYTHON_PROJECT_BASE"])
else:
    if sys.executable:
        project_base = os.path.dirname(os.path.abspath(sys.executable))
    else:
        # sys.executable can be empty if argv[0] has been changed and Python is
        # unable to retrieve the real program name
        project_base = os.getcwd()


# python_build: (Boolean) if true, we're either building Python or
# building an extension with an un-installed Python, so we use
# different (hard-wired) directories.
def _is_python_source_dir(d):
    for fn in ("Setup", "Setup.local"):
        if os.path.isfile(os.path.join(d, "Modules", fn)):
            return True
    return False

_sys_home = getattr(sys, '_home', None)

if os.name == 'nt':
    def _fix_pcbuild(d):
        if d and os.path.normcase(d).startswith(
                os.path.normcase(os.path.join(PREFIX, "PCbuild"))):
            return PREFIX
        return d
    project_base = _fix_pcbuild(project_base)
    _sys_home = _fix_pcbuild(_sys_home)

def _python_build():
    if _sys_home:
        return _is_python_source_dir(_sys_home)
    return _is_python_source_dir(project_base)

python_build = _python_build()


# Calculate the build qualifier flags if they are defined.  Adding the flags
# to the include and lib directories only makes sense for an installation, not
# an in-source build.
build_flags = ''
try:
    if not python_build:
        build_flags = sys.abiflags
except AttributeError:
    # It's not a configure-based build, so the sys module doesn't have
    # this attribute, which is fine.
    pass

def get_python_version():
    """Return a string containing the major and minor Python version,
    leaving off the patchlevel.  Sample return values could be '1.5'
    or '2.2'.
    """
    return '%d.%d' % sys.version_info[:2]


def get_python_inc(plat_specific=0, prefix=None):
    """Return the directory containing installed Python header files.

    If 'plat_specific' is false (the default), this is the path to the
    non-platform-specific header files, i.e. Python.h and so on;
    otherwise, this is the path to platform-specific header files
    (namely pyconfig.h).

    If 'prefix' is supplied, use it instead of sys.base_prefix or
    sys.base_exec_prefix -- i.e., ignore 'plat_specific'.
    """
    if prefix is None:
        prefix = plat_specific and BASE_EXEC_PREFIX or BASE_PREFIX
    if os.name == "posix":
        if IS_PYPY and sys.version_info < (3, 8):
            return os.path.join(prefix, 'include')
        if python_build:
            # Assume the executable is in the build directory.  The
            # pyconfig.h file should be in the same directory.  Since
            # the build directory may not be the source directory, we
            # must use "srcdir" from the makefile to find the "Include"
            # directory.
            if plat_specific:
                return _sys_home or project_base
            else:
                incdir = os.path.join(get_config_var('srcdir'), 'Include')
                return os.path.normpath(incdir)
        implementation = 'pypy' if IS_PYPY else 'python'
        python_dir = implementation + get_python_version() + build_flags
        return os.path.join(prefix, "include", python_dir)
    elif os.name == "nt":
        if python_build:
            # Include both the include and PC dir to ensure we can find
            # pyconfig.h
            return (os.path.join(prefix, "include") + os.path.pathsep +
                    os.path.join(prefix, "PC"))
        return os.path.join(prefix, "include")
    else:
        raise DistutilsPlatformError(
            "I don't know where Python installs its C header files "
            "on platform '%s'" % os.name)


def get_python_lib(plat_specific=0, standard_lib=0, prefix=None):
    """Return the directory containing the Python library (standard or
    site additions).

    If 'plat_specific' is true, return the directory containing
    platform-specific modules, i.e. any module from a non-pure-Python
    module distribution; otherwise, return the platform-shared library
    directory.  If 'standard_lib' is true, return the directory
    containing standard Python library modules; otherwise, return the
    directory for site-specific modules.

    If 'prefix' is supplied, use it instead of sys.base_prefix or
    sys.base_exec_prefix -- i.e., ignore 'plat_specific'.
    """

    if IS_PYPY and sys.version_info < (3, 8):
        # PyPy-specific schema
        if prefix is None:
            prefix = PREFIX
        if standard_lib:
            return os.path.join(prefix, "lib-python", sys.version[0])
        return os.path.join(prefix, 'site-packages')

    if prefix is None:
        if standard_lib:
            prefix = plat_specific and BASE_EXEC_PREFIX or BASE_PREFIX
        else:
            prefix = plat_specific and EXEC_PREFIX or PREFIX

    if os.name == "posix":
        if plat_specific or standard_lib:
            # Platform-specific modules (any module from a non-pure-Python
            # module distribution) or standard Python library modules.
            libdir = getattr(sys, "platlibdir", "lib")
        else:
            # Pure Python
            libdir = "lib"
        implementation = 'pypy' if IS_PYPY else 'python'
        libpython = os.path.join(prefix, libdir,
                                 implementation + get_python_version())
        if standard_lib:
            return libpython
        else:
            return os.path.join(libpython, "site-packages")
    elif os.name == "nt":
        if standard_lib:
            return os.path.join(prefix, "Lib")
        else:
            return os.path.join(prefix, "Lib", "site-packages")
    else:
        raise DistutilsPlatformError(
            "I don't know where Python installs its library "
            "on platform '%s'" % os.name)



def customize_compiler(compiler):
    """Do any platform-specific customization of a CCompiler instance.

    Mainly needed on Unix, so we can plug in the information that
    varies across Unices and is stored in Python's Makefile.
    """
    if compiler.compiler_type == "unix":
        if sys.platform == "darwin":
            # Perform first-time customization of compiler-related
            # config vars on OS X now that we know we need a compiler.
            # This is primarily to support Pythons from binary
            # installers.  The kind and paths to build tools on
            # the user system may vary significantly from the system
            # that Python itself was built on.  Also the user OS
            # version and build tools may not support the same set
            # of CPU architectures for universal builds.
            global _config_vars
            # Use get_config_var() to ensure _config_vars is initialized.
            if not get_config_var('CUSTOMIZED_OSX_COMPILER'):
                import _osx_support
                _osx_support.customize_compiler(_config_vars)
                _config_vars['CUSTOMIZED_OSX_COMPILER'] = 'True'

        (cc, cxx, cflags, ccshared, ldshared, shlib_suffix, ar, ar_flags) = \
            get_config_vars('CC', 'CXX', 'CFLAGS',
                            'CCSHARED', 'LDSHARED', 'SHLIB_SUFFIX', 'AR', 'ARFLAGS')

        if 'CC' in os.environ:
            newcc = os.environ['CC']
            if('LDSHARED' not in os.environ
                    and ldshared.startswith(cc)):
                # If CC is overridden, use that as the default
                #       command for LDSHARED as well
                ldshared = newcc + ldshared[len(cc):]
            cc = newcc
        if 'CXX' in os.environ:
            cxx = os.environ['CXX']
        if 'LDSHARED' in os.environ:
            ldshared = os.environ['LDSHARED']
        if 'CPP' in os.environ:
            cpp = os.environ['CPP']
        else:
            cpp = cc + " -E"           # not always
        if 'LDFLAGS' in os.environ:
            ldshared = ldshared + ' ' + os.environ['LDFLAGS']
        if 'CFLAGS' in os.environ:
            cflags = cflags + ' ' + os.environ['CFLAGS']
            ldshared = ldshared + ' ' + os.environ['CFLAGS']
        if 'CPPFLAGS' in os.environ:
            cpp = cpp + ' ' + os.environ['CPPFLAGS']
            cflags = cflags + ' ' + os.environ['CPPFLAGS']
            ldshared = ldshared + ' ' + os.environ['CPPFLAGS']
        if 'AR' in os.environ:
            ar = os.environ['AR']
        if 'ARFLAGS' in os.environ:
            archiver = ar + ' ' + os.environ['ARFLAGS']
        else:
            archiver = ar + ' ' + ar_flags

        cc_cmd = cc + ' ' + cflags
        compiler.set_executables(
            preprocessor=cpp,
            compiler=cc_cmd,
            compiler_so=cc_cmd + ' ' + ccshared,
            compiler_cxx=cxx,
            linker_so=ldshared,
            linker_exe=cc,
            archiver=archiver)

        if 'RANLIB' in os.environ and compiler.executables.get('ranlib', None):
            compiler.set_executables(ranlib=os.environ['RANLIB'])

        compiler.shared_lib_extension = shlib_suffix


def get_config_h_filename():
    """Return full pathname of installed pyconfig.h file."""
    if python_build:
        if os.name == "nt":
            inc_dir = os.path.join(_sys_home or project_base, "PC")
        else:
            inc_dir = _sys_home or project_base
    else:
        inc_dir = get_python_inc(plat_specific=1)

    return os.path.join(inc_dir, 'pyconfig.h')


def get_makefile_filename():
    """Return full pathname of installed Makefile from the Python build."""
    if python_build:
        return os.path.join(_sys_home or project_base, "Makefile")
    lib_dir = get_python_lib(plat_specific=0, standard_lib=1)
    config_file = 'config-{}{}'.format(get_python_version(), build_flags)
    if hasattr(sys.implementation, '_multiarch'):
        config_file += '-%s' % sys.implementation._multiarch
    return os.path.join(lib_dir, config_file, 'Makefile')


def parse_config_h(fp, g=None):
    """Parse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    if g is None:
        g = {}
    define_rx = re.compile("#define ([A-Z][A-Za-z0-9_]+) (.*)\n")
    undef_rx = re.compile("/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\n")
    #
    while True:
        line = fp.readline()
        if not line:
            break
        m = define_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            try: v = int(v)
            except ValueError: pass
            g[n] = v
        else:
            m = undef_rx.match(line)
            if m:
                g[m.group(1)] = 0
    return g


# Regexes needed for parsing Makefile (and similar syntaxes,
# like old-style Setup files).
_variable_rx = re.compile(r"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)")
_findvar1_rx = re.compile(r"\$\(([A-Za-z][A-Za-z0-9_]*)\)")
_findvar2_rx = re.compile(r"\${([A-Za-z][A-Za-z0-9_]*)}")

def parse_makefile(fn, g=None):
    """Parse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    from distutils.text_file import TextFile
    fp = TextFile(fn, strip_comments=1, skip_blanks=1, join_lines=1, errors="surrogateescape")

    if g is None:
        g = {}
    done = {}
    notdone = {}

    while True:
        line = fp.readline()
        if line is None: # eof
            break
        m = _variable_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            v = v.strip()
            # `$$' is a literal `$' in make
            tmpv = v.replace('$$', '')

            if "$" in tmpv:
                notdone[n] = v
            else:
                try:
                    v = int(v)
                except ValueError:
                    # insert literal `$'
                    done[n] = v.replace('$$', '$')
                else:
                    done[n] = v

    # Variables with a 'PY_' prefix in the makefile. These need to
    # be made available without that prefix through sysconfig.
    # Special care is needed to ensure that variable expansion works, even
    # if the expansion uses the name without a prefix.
    renamed_variables = ('CFLAGS', 'LDFLAGS', 'CPPFLAGS')

    # do variable interpolation here
    while notdone:
        for name in list(notdone):
            value = notdone[name]
            m = _findvar1_rx.search(value) or _findvar2_rx.search(value)
            if m:
                n = m.group(1)
                found = True
                if n in done:
                    item = str(done[n])
                elif n in notdone:
                    # get it on a subsequent round
                    found = False
                elif n in os.environ:
                    # do it like make: fall back to environment
                    item = os.environ[n]

                elif n in renamed_variables:
                    if name.startswith('PY_') and name[3:] in renamed_variables:
                        item = ""

                    elif 'PY_' + n in notdone:
                        found = False

                    else:
                        item = str(done['PY_' + n])
                else:
                    done[n] = item = ""
                if found:
                    after = value[m.end():]
                    value = value[:m.start()] + item + after
                    if "$" in after:
                        notdone[name] = value
                    else:
                        try: value = int(value)
                        except ValueError:
                            done[name] = value.strip()
                        else:
                            done[name] = value
                        del notdone[name]

                        if name.startswith('PY_') \
                            and name[3:] in renamed_variables:

                            name = name[3:]
                            if name not in done:
                                done[name] = value
            else:
                # bogus variable reference; just drop it since we can't deal
                del notdone[name]

    fp.close()

    # strip spurious spaces
    for k, v in done.items():
        if isinstance(v, str):
            done[k] = v.strip()

    # save the results in the global dictionary
    g.update(done)
    return g


def expand_makefile_vars(s, vars):
    """Expand Makefile-style variables -- "${foo}" or "$(foo)" -- in
    'string' according to 'vars' (a dictionary mapping variable names to
    values).  Variables not present in 'vars' are silently expanded to the
    empty string.  The variable values in 'vars' should not contain further
    variable expansions; if 'vars' is the output of 'parse_makefile()',
    you're fine.  Returns a variable-expanded version of 's'.
    """

    # This algorithm does multiple expansion, so if vars['foo'] contains
    # "${bar}", it will expand ${foo} to ${bar}, and then expand
    # ${bar}... and so forth.  This is fine as long as 'vars' comes from
    # 'parse_makefile()', which takes care of such expansions eagerly,
    # according to make's variable expansion semantics.

    while True:
        m = _findvar1_rx.search(s) or _findvar2_rx.search(s)
        if m:
            (beg, end) = m.span()
            s = s[0:beg] + vars.get(m.group(1)) + s[end:]
        else:
            break
    return s


_config_vars = None

def _init_posix():
    """Initialize the module as appropriate for POSIX systems."""
    # _sysconfigdata is generated at build time, see the sysconfig module
    name = os.environ.get('_PYTHON_SYSCONFIGDATA_NAME',
        '_sysconfigdata_{abi}_{platform}_{multiarch}'.format(
        abi=sys.abiflags,
        platform=sys.platform,
        multiarch=getattr(sys.implementation, '_multiarch', ''),
    ))
    try:
        _temp = __import__(name, globals(), locals(), ['build_time_vars'], 0)
    except ImportError:
        # Python 3.5 and pypy 7.3.1
        _temp = __import__(
            '_sysconfigdata', globals(), locals(), ['build_time_vars'], 0)
    build_time_vars = _temp.build_time_vars
    global _config_vars
    _config_vars = {}
    _config_vars.update(build_time_vars)


def _init_nt():
    """Initialize the module as appropriate for NT"""
    g = {}
    # set basic install directories
    g['LIBDEST'] = get_python_lib(plat_specific=0, standard_lib=1)
    g['BINLIBDEST'] = get_python_lib(plat_specific=1, standard_lib=1)

    # XXX hmmm.. a normal install puts include files here
    g['INCLUDEPY'] = get_python_inc(plat_specific=0)

    g['EXT_SUFFIX'] = _imp.extension_suffixes()[0]
    g['EXE'] = ".exe"
    g['VERSION'] = get_python_version().replace(".", "")
    g['BINDIR'] = os.path.dirname(os.path.abspath(sys.executable))

    global _config_vars
    _config_vars = g


def get_config_vars(*args):
    """With no arguments, return a dictionary of all configuration
    variables relevant for the current platform.  Generally this includes
    everything needed to build extensions and install both pure modules and
    extensions.  On Unix, this means every variable defined in Python's
    installed Makefile; on Windows it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    """
    global _config_vars
    if _config_vars is None:
        func = globals().get("_init_" + os.name)
        if func:
            func()
        else:
            _config_vars = {}

        # Normalized versions of prefix and exec_prefix are handy to have;
        # in fact, these are the standard versions used most places in the
        # Distutils.
        _config_vars['prefix'] = PREFIX
        _config_vars['exec_prefix'] = EXEC_PREFIX

        if not IS_PYPY:
            # For backward compatibility, see issue19555
            SO = _config_vars.get('EXT_SUFFIX')
            if SO is not None:
                _config_vars['SO'] = SO

            # Always convert srcdir to an absolute path
            srcdir = _config_vars.get('srcdir', project_base)
            if os.name == 'posix':
                if python_build:
                    # If srcdir is a relative path (typically '.' or '..')
                    # then it should be interpreted relative to the directory
                    # containing Makefile.
                    base = os.path.dirname(get_makefile_filename())
                    srcdir = os.path.join(base, srcdir)
                else:
                    # srcdir is not meaningful since the installation is
                    # spread about the filesystem.  We choose the
                    # directory containing the Makefile since we know it
                    # exists.
                    srcdir = os.path.dirname(get_makefile_filename())
            _config_vars['srcdir'] = os.path.abspath(os.path.normpath(srcdir))

            # Convert srcdir into an absolute path if it appears necessary.
            # Normally it is relative to the build directory.  However, during
            # testing, for example, we might be running a non-installed python
            # from a different directory.
            if python_build and os.name == "posix":
                base = project_base
                if (not os.path.isabs(_config_vars['srcdir']) and
                    base != os.getcwd()):
                    # srcdir is relative and we are not in the same directory
                    # as the executable. Assume executable is in the build
                    # directory and make srcdir absolute.
                    srcdir = os.path.join(base, _config_vars['srcdir'])
                    _config_vars['srcdir'] = os.path.normpath(srcdir)

        # OS X platforms require special customization to handle
        # multi-architecture, multi-os-version installers
        if sys.platform == 'darwin':
            import _osx_support
            _osx_support.customize_config_vars(_config_vars)

    if args:
        vals = []
        for name in args:
            vals.append(_config_vars.get(name))
        return vals
    else:
        return _config_vars

def get_config_var(name):
    """Return the value of a single variable using the dictionary
    returned by 'get_config_vars()'.  Equivalent to
    get_config_vars().get(name)
    """
    if name == 'SO':
        import warnings
        warnings.warn('SO is deprecated, use EXT_SUFFIX', DeprecationWarning, 2)
    return get_config_vars().get(name)
site-packages/setuptools/_distutils/unixccompiler.py000064400000034312150564745010017135 0ustar00"""distutils.unixccompiler

Contains the UnixCCompiler class, a subclass of CCompiler that handles
the "typical" Unix-style command-line C compiler:
  * macros defined with -Dname[=value]
  * macros undefined with -Uname
  * include search directories specified with -Idir
  * libraries specified with -lllib
  * library search directories specified with -Ldir
  * compile handled by 'cc' (or similar) executable with -c option:
    compiles .c to .o
  * link static library handled by 'ar' command (possibly with 'ranlib')
  * link shared library handled by 'cc -shared'
"""

import os, sys, re, shlex

from distutils import sysconfig
from distutils.dep_util import newer
from distutils.ccompiler import \
     CCompiler, gen_preprocess_options, gen_lib_options
from distutils.errors import \
     DistutilsExecError, CompileError, LibError, LinkError
from distutils import log

if sys.platform == 'darwin':
    import _osx_support

# XXX Things not currently handled:
#   * optimization/debug/warning flags; we just use whatever's in Python's
#     Makefile and live with it.  Is this adequate?  If not, we might
#     have to have a bunch of subclasses GNUCCompiler, SGICCompiler,
#     SunCCompiler, and I suspect down that road lies madness.
#   * even if we don't know a warning flag from an optimization flag,
#     we need some way for outsiders to feed preprocessor/compiler/linker
#     flags in to us -- eg. a sysadmin might want to mandate certain flags
#     via a site config file, or a user might want to set something for
#     compiling this module distribution only via the setup.py command
#     line, whatever.  As long as these options come from something on the
#     current system, they can be as system-dependent as they like, and we
#     should just happily stuff them into the preprocessor/compiler/linker
#     options and carry on.


class UnixCCompiler(CCompiler):

    compiler_type = 'unix'

    # These are used by CCompiler in two places: the constructor sets
    # instance attributes 'preprocessor', 'compiler', etc. from them, and
    # 'set_executable()' allows any of these to be set.  The defaults here
    # are pretty generic; they will probably have to be set by an outsider
    # (eg. using information discovered by the sysconfig about building
    # Python extensions).
    executables = {'preprocessor' : None,
                   'compiler'     : ["cc"],
                   'compiler_so'  : ["cc"],
                   'compiler_cxx' : ["cc"],
                   'linker_so'    : ["cc", "-shared"],
                   'linker_exe'   : ["cc"],
                   'archiver'     : ["ar", "-cr"],
                   'ranlib'       : None,
                  }

    if sys.platform[:6] == "darwin":
        executables['ranlib'] = ["ranlib"]

    # Needed for the filename generation methods provided by the base
    # class, CCompiler.  NB. whoever instantiates/uses a particular
    # UnixCCompiler instance should set 'shared_lib_ext' -- we set a
    # reasonable common default here, but it's not necessarily used on all
    # Unices!

    src_extensions = [".c",".C",".cc",".cxx",".cpp",".m"]
    obj_extension = ".o"
    static_lib_extension = ".a"
    shared_lib_extension = ".so"
    dylib_lib_extension = ".dylib"
    xcode_stub_lib_extension = ".tbd"
    static_lib_format = shared_lib_format = dylib_lib_format = "lib%s%s"
    xcode_stub_lib_format = dylib_lib_format
    if sys.platform == "cygwin":
        exe_extension = ".exe"

    def preprocess(self, source, output_file=None, macros=None,
                   include_dirs=None, extra_preargs=None, extra_postargs=None):
        fixed_args = self._fix_compile_args(None, macros, include_dirs)
        ignore, macros, include_dirs = fixed_args
        pp_opts = gen_preprocess_options(macros, include_dirs)
        pp_args = self.preprocessor + pp_opts
        if output_file:
            pp_args.extend(['-o', output_file])
        if extra_preargs:
            pp_args[:0] = extra_preargs
        if extra_postargs:
            pp_args.extend(extra_postargs)
        pp_args.append(source)

        # We need to preprocess: either we're being forced to, or we're
        # generating output to stdout, or there's a target output file and
        # the source file is newer than the target (or the target doesn't
        # exist).
        if self.force or output_file is None or newer(source, output_file):
            if output_file:
                self.mkpath(os.path.dirname(output_file))
            try:
                self.spawn(pp_args)
            except DistutilsExecError as msg:
                raise CompileError(msg)

    def _compile(self, obj, src, ext, cc_args, extra_postargs, pp_opts):
        compiler_so = self.compiler_so
        if sys.platform == 'darwin':
            compiler_so = _osx_support.compiler_fixup(compiler_so,
                                                    cc_args + extra_postargs)
        try:
            self.spawn(compiler_so + cc_args + [src, '-o', obj] +
                       extra_postargs)
        except DistutilsExecError as msg:
            raise CompileError(msg)

    def create_static_lib(self, objects, output_libname,
                          output_dir=None, debug=0, target_lang=None):
        objects, output_dir = self._fix_object_args(objects, output_dir)

        output_filename = \
            self.library_filename(output_libname, output_dir=output_dir)

        if self._need_link(objects, output_filename):
            self.mkpath(os.path.dirname(output_filename))
            self.spawn(self.archiver +
                       [output_filename] +
                       objects + self.objects)

            # Not many Unices required ranlib anymore -- SunOS 4.x is, I
            # think the only major Unix that does.  Maybe we need some
            # platform intelligence here to skip ranlib if it's not
            # needed -- or maybe Python's configure script took care of
            # it for us, hence the check for leading colon.
            if self.ranlib:
                try:
                    self.spawn(self.ranlib + [output_filename])
                except DistutilsExecError as msg:
                    raise LibError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    def link(self, target_desc, objects,
             output_filename, output_dir=None, libraries=None,
             library_dirs=None, runtime_library_dirs=None,
             export_symbols=None, debug=0, extra_preargs=None,
             extra_postargs=None, build_temp=None, target_lang=None):
        objects, output_dir = self._fix_object_args(objects, output_dir)
        fixed_args = self._fix_lib_args(libraries, library_dirs,
                                        runtime_library_dirs)
        libraries, library_dirs, runtime_library_dirs = fixed_args

        lib_opts = gen_lib_options(self, library_dirs, runtime_library_dirs,
                                   libraries)
        if not isinstance(output_dir, (str, type(None))):
            raise TypeError("'output_dir' must be a string or None")
        if output_dir is not None:
            output_filename = os.path.join(output_dir, output_filename)

        if self._need_link(objects, output_filename):
            ld_args = (objects + self.objects +
                       lib_opts + ['-o', output_filename])
            if debug:
                ld_args[:0] = ['-g']
            if extra_preargs:
                ld_args[:0] = extra_preargs
            if extra_postargs:
                ld_args.extend(extra_postargs)
            self.mkpath(os.path.dirname(output_filename))
            try:
                if target_desc == CCompiler.EXECUTABLE:
                    linker = self.linker_exe[:]
                else:
                    linker = self.linker_so[:]
                if target_lang == "c++" and self.compiler_cxx:
                    # skip over environment variable settings if /usr/bin/env
                    # is used to set up the linker's environment.
                    # This is needed on OSX. Note: this assumes that the
                    # normal and C++ compiler have the same environment
                    # settings.
                    i = 0
                    if os.path.basename(linker[0]) == "env":
                        i = 1
                        while '=' in linker[i]:
                            i += 1

                    if os.path.basename(linker[i]) == 'ld_so_aix':
                        # AIX platforms prefix the compiler with the ld_so_aix
                        # script, so we need to adjust our linker index
                        offset = 1
                    else:
                        offset = 0

                    linker[i+offset] = self.compiler_cxx[i]

                if sys.platform == 'darwin':
                    linker = _osx_support.compiler_fixup(linker, ld_args)

                self.spawn(linker + ld_args)
            except DistutilsExecError as msg:
                raise LinkError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    # -- Miscellaneous methods -----------------------------------------
    # These are all used by the 'gen_lib_options() function, in
    # ccompiler.py.

    def library_dir_option(self, dir):
        return "-L" + dir

    def _is_gcc(self, compiler_name):
        return "gcc" in compiler_name or "g++" in compiler_name

    def runtime_library_dir_option(self, dir):
        # XXX Hackish, at the very least.  See Python bug #445902:
        # http://sourceforge.net/tracker/index.php
        #   ?func=detail&aid=445902&group_id=5470&atid=105470
        # Linkers on different platforms need different options to
        # specify that directories need to be added to the list of
        # directories searched for dependencies when a dynamic library
        # is sought.  GCC on GNU systems (Linux, FreeBSD, ...) has to
        # be told to pass the -R option through to the linker, whereas
        # other compilers and gcc on other systems just know this.
        # Other compilers may need something slightly different.  At
        # this time, there's no way to determine this information from
        # the configuration data stored in the Python installation, so
        # we use this hack.
        compiler = os.path.basename(shlex.split(sysconfig.get_config_var("CC"))[0])
        if sys.platform[:6] == "darwin":
            from distutils.util import get_macosx_target_ver, split_version
            macosx_target_ver = get_macosx_target_ver()
            if macosx_target_ver and split_version(macosx_target_ver) >= [10, 5]:
                return "-Wl,-rpath," + dir
            else: # no support for -rpath on earlier macOS versions
                return "-L" + dir
        elif sys.platform[:7] == "freebsd":
            return "-Wl,-rpath=" + dir
        elif sys.platform[:5] == "hp-ux":
            if self._is_gcc(compiler):
                return ["-Wl,+s", "-L" + dir]
            return ["+s", "-L" + dir]

        # For all compilers, `-Wl` is the presumed way to
        # pass a compiler option to the linker and `-R` is
        # the way to pass an RPATH.
        if sysconfig.get_config_var("GNULD") == "yes":
            # GNU ld needs an extra option to get a RUNPATH
            # instead of just an RPATH.
            return "-Wl,--enable-new-dtags,-R" + dir
        else:
            return "-Wl,-R" + dir

    def library_option(self, lib):
        return "-l" + lib

    def find_library_file(self, dirs, lib, debug=0):
        shared_f = self.library_filename(lib, lib_type='shared')
        dylib_f = self.library_filename(lib, lib_type='dylib')
        xcode_stub_f = self.library_filename(lib, lib_type='xcode_stub')
        static_f = self.library_filename(lib, lib_type='static')

        if sys.platform == 'darwin':
            # On OSX users can specify an alternate SDK using
            # '-isysroot', calculate the SDK root if it is specified
            # (and use it further on)
            #
            # Note that, as of Xcode 7, Apple SDKs may contain textual stub
            # libraries with .tbd extensions rather than the normal .dylib
            # shared libraries installed in /.  The Apple compiler tool
            # chain handles this transparently but it can cause problems
            # for programs that are being built with an SDK and searching
            # for specific libraries.  Callers of find_library_file need to
            # keep in mind that the base filename of the returned SDK library
            # file might have a different extension from that of the library
            # file installed on the running system, for example:
            #   /Applications/Xcode.app/Contents/Developer/Platforms/
            #       MacOSX.platform/Developer/SDKs/MacOSX10.11.sdk/
            #       usr/lib/libedit.tbd
            # vs
            #   /usr/lib/libedit.dylib
            cflags = sysconfig.get_config_var('CFLAGS')
            m = re.search(r'-isysroot\s*(\S+)', cflags)
            if m is None:
                sysroot = '/'
            else:
                sysroot = m.group(1)



        for dir in dirs:
            shared = os.path.join(dir, shared_f)
            dylib = os.path.join(dir, dylib_f)
            static = os.path.join(dir, static_f)
            xcode_stub = os.path.join(dir, xcode_stub_f)

            if sys.platform == 'darwin' and (
                dir.startswith('/System/') or (
                dir.startswith('/usr/') and not dir.startswith('/usr/local/'))):

                shared = os.path.join(sysroot, dir[1:], shared_f)
                dylib = os.path.join(sysroot, dir[1:], dylib_f)
                static = os.path.join(sysroot, dir[1:], static_f)
                xcode_stub = os.path.join(sysroot, dir[1:], xcode_stub_f)

            # We're second-guessing the linker here, with not much hard
            # data to go on: GCC seems to prefer the shared library, so I'm
            # assuming that *all* Unix C compilers do.  And of course I'm
            # ignoring even GCC's "-static" option.  So sue me.
            if os.path.exists(dylib):
                return dylib
            elif os.path.exists(xcode_stub):
                return xcode_stub
            elif os.path.exists(shared):
                return shared
            elif os.path.exists(static):
                return static

        # Oops, didn't find it in *any* of 'dirs'
        return None
site-packages/setuptools/_distutils/extension.py000064400000024423150564745010016272 0ustar00"""distutils.extension

Provides the Extension class, used to describe C/C++ extension
modules in setup scripts."""

import os
import warnings

# This class is really only used by the "build_ext" command, so it might
# make sense to put it in distutils.command.build_ext.  However, that
# module is already big enough, and I want to make this class a bit more
# complex to simplify some common cases ("foo" module in "foo.c") and do
# better error-checking ("foo.c" actually exists).
#
# Also, putting this in build_ext.py means every setup script would have to
# import that large-ish module (indirectly, through distutils.core) in
# order to do anything.

class Extension:
    """Just a collection of attributes that describes an extension
    module and everything needed to build it (hopefully in a portable
    way, but there are hooks that let you be as unportable as you need).

    Instance attributes:
      name : string
        the full name of the extension, including any packages -- ie.
        *not* a filename or pathname, but Python dotted name
      sources : [string]
        list of source filenames, relative to the distribution root
        (where the setup script lives), in Unix form (slash-separated)
        for portability.  Source files may be C, C++, SWIG (.i),
        platform-specific resource files, or whatever else is recognized
        by the "build_ext" command as source for a Python extension.
      include_dirs : [string]
        list of directories to search for C/C++ header files (in Unix
        form for portability)
      define_macros : [(name : string, value : string|None)]
        list of macros to define; each macro is defined using a 2-tuple,
        where 'value' is either the string to define it to or None to
        define it without a particular value (equivalent of "#define
        FOO" in source or -DFOO on Unix C compiler command line)
      undef_macros : [string]
        list of macros to undefine explicitly
      library_dirs : [string]
        list of directories to search for C/C++ libraries at link time
      libraries : [string]
        list of library names (not filenames or paths) to link against
      runtime_library_dirs : [string]
        list of directories to search for C/C++ libraries at run time
        (for shared extensions, this is when the extension is loaded)
      extra_objects : [string]
        list of extra files to link with (eg. object files not implied
        by 'sources', static library that must be explicitly specified,
        binary resource files, etc.)
      extra_compile_args : [string]
        any extra platform- and compiler-specific information to use
        when compiling the source files in 'sources'.  For platforms and
        compilers where "command line" makes sense, this is typically a
        list of command-line arguments, but for other platforms it could
        be anything.
      extra_link_args : [string]
        any extra platform- and compiler-specific information to use
        when linking object files together to create the extension (or
        to create a new static Python interpreter).  Similar
        interpretation as for 'extra_compile_args'.
      export_symbols : [string]
        list of symbols to be exported from a shared extension.  Not
        used on all platforms, and not generally necessary for Python
        extensions, which typically export exactly one symbol: "init" +
        extension_name.
      swig_opts : [string]
        any extra options to pass to SWIG if a source file has the .i
        extension.
      depends : [string]
        list of files that the extension depends on
      language : string
        extension language (i.e. "c", "c++", "objc"). Will be detected
        from the source extensions if not provided.
      optional : boolean
        specifies that a build failure in the extension should not abort the
        build process, but simply not install the failing extension.
    """

    # When adding arguments to this constructor, be sure to update
    # setup_keywords in core.py.
    def __init__(self, name, sources,
                  include_dirs=None,
                  define_macros=None,
                  undef_macros=None,
                  library_dirs=None,
                  libraries=None,
                  runtime_library_dirs=None,
                  extra_objects=None,
                  extra_compile_args=None,
                  extra_link_args=None,
                  export_symbols=None,
                  swig_opts = None,
                  depends=None,
                  language=None,
                  optional=None,
                  **kw                      # To catch unknown keywords
                 ):
        if not isinstance(name, str):
            raise AssertionError("'name' must be a string")
        if not (isinstance(sources, list) and
                all(isinstance(v, str) for v in sources)):
            raise AssertionError("'sources' must be a list of strings")

        self.name = name
        self.sources = sources
        self.include_dirs = include_dirs or []
        self.define_macros = define_macros or []
        self.undef_macros = undef_macros or []
        self.library_dirs = library_dirs or []
        self.libraries = libraries or []
        self.runtime_library_dirs = runtime_library_dirs or []
        self.extra_objects = extra_objects or []
        self.extra_compile_args = extra_compile_args or []
        self.extra_link_args = extra_link_args or []
        self.export_symbols = export_symbols or []
        self.swig_opts = swig_opts or []
        self.depends = depends or []
        self.language = language
        self.optional = optional

        # If there are unknown keyword options, warn about them
        if len(kw) > 0:
            options = [repr(option) for option in kw]
            options = ', '.join(sorted(options))
            msg = "Unknown Extension options: %s" % options
            warnings.warn(msg)

    def __repr__(self):
        return '<%s.%s(%r) at %#x>' % (
            self.__class__.__module__,
            self.__class__.__qualname__,
            self.name,
            id(self))


def read_setup_file(filename):
    """Reads a Setup file and returns Extension instances."""
    from distutils.sysconfig import (parse_makefile, expand_makefile_vars,
                                     _variable_rx)

    from distutils.text_file import TextFile
    from distutils.util import split_quoted

    # First pass over the file to gather "VAR = VALUE" assignments.
    vars = parse_makefile(filename)

    # Second pass to gobble up the real content: lines of the form
    #   <module> ... [<sourcefile> ...] [<cpparg> ...] [<library> ...]
    file = TextFile(filename,
                    strip_comments=1, skip_blanks=1, join_lines=1,
                    lstrip_ws=1, rstrip_ws=1)
    try:
        extensions = []

        while True:
            line = file.readline()
            if line is None:                # eof
                break
            if _variable_rx.match(line):    # VAR=VALUE, handled in first pass
                continue

            if line[0] == line[-1] == "*":
                file.warn("'%s' lines not handled yet" % line)
                continue

            line = expand_makefile_vars(line, vars)
            words = split_quoted(line)

            # NB. this parses a slightly different syntax than the old
            # makesetup script: here, there must be exactly one extension per
            # line, and it must be the first word of the line.  I have no idea
            # why the old syntax supported multiple extensions per line, as
            # they all wind up being the same.

            module = words[0]
            ext = Extension(module, [])
            append_next_word = None

            for word in words[1:]:
                if append_next_word is not None:
                    append_next_word.append(word)
                    append_next_word = None
                    continue

                suffix = os.path.splitext(word)[1]
                switch = word[0:2] ; value = word[2:]

                if suffix in (".c", ".cc", ".cpp", ".cxx", ".c++", ".m", ".mm"):
                    # hmm, should we do something about C vs. C++ sources?
                    # or leave it up to the CCompiler implementation to
                    # worry about?
                    ext.sources.append(word)
                elif switch == "-I":
                    ext.include_dirs.append(value)
                elif switch == "-D":
                    equals = value.find("=")
                    if equals == -1:        # bare "-DFOO" -- no value
                        ext.define_macros.append((value, None))
                    else:                   # "-DFOO=blah"
                        ext.define_macros.append((value[0:equals],
                                                  value[equals+2:]))
                elif switch == "-U":
                    ext.undef_macros.append(value)
                elif switch == "-C":        # only here 'cause makesetup has it!
                    ext.extra_compile_args.append(word)
                elif switch == "-l":
                    ext.libraries.append(value)
                elif switch == "-L":
                    ext.library_dirs.append(value)
                elif switch == "-R":
                    ext.runtime_library_dirs.append(value)
                elif word == "-rpath":
                    append_next_word = ext.runtime_library_dirs
                elif word == "-Xlinker":
                    append_next_word = ext.extra_link_args
                elif word == "-Xcompiler":
                    append_next_word = ext.extra_compile_args
                elif switch == "-u":
                    ext.extra_link_args.append(word)
                    if not value:
                        append_next_word = ext.extra_link_args
                elif suffix in (".a", ".so", ".sl", ".o", ".dylib"):
                    # NB. a really faithful emulation of makesetup would
                    # append a .o file to extra_objects only if it
                    # had a slash in it; otherwise, it would s/.o/.c/
                    # and append it to sources.  Hmmmm.
                    ext.extra_objects.append(word)
                else:
                    file.warn("unrecognized argument '%s'" % word)

            extensions.append(ext)
    finally:
        file.close()

    return extensions
site-packages/setuptools/_distutils/errors.py000064400000006771150564745010015600 0ustar00"""distutils.errors

Provides exceptions used by the Distutils modules.  Note that Distutils
modules may raise standard exceptions; in particular, SystemExit is
usually raised for errors that are obviously the end-user's fault
(eg. bad command-line arguments).

This module is safe to use in "from ... import *" mode; it only exports
symbols whose names start with "Distutils" and end with "Error"."""

class DistutilsError (Exception):
    """The root of all Distutils evil."""
    pass

class DistutilsModuleError (DistutilsError):
    """Unable to load an expected module, or to find an expected class
    within some module (in particular, command modules and classes)."""
    pass

class DistutilsClassError (DistutilsError):
    """Some command class (or possibly distribution class, if anyone
    feels a need to subclass Distribution) is found not to be holding
    up its end of the bargain, ie. implementing some part of the
    "command "interface."""
    pass

class DistutilsGetoptError (DistutilsError):
    """The option table provided to 'fancy_getopt()' is bogus."""
    pass

class DistutilsArgError (DistutilsError):
    """Raised by fancy_getopt in response to getopt.error -- ie. an
    error in the command line usage."""
    pass

class DistutilsFileError (DistutilsError):
    """Any problems in the filesystem: expected file not found, etc.
    Typically this is for problems that we detect before OSError
    could be raised."""
    pass

class DistutilsOptionError (DistutilsError):
    """Syntactic/semantic errors in command options, such as use of
    mutually conflicting options, or inconsistent options,
    badly-spelled values, etc.  No distinction is made between option
    values originating in the setup script, the command line, config
    files, or what-have-you -- but if we *know* something originated in
    the setup script, we'll raise DistutilsSetupError instead."""
    pass

class DistutilsSetupError (DistutilsError):
    """For errors that can be definitely blamed on the setup script,
    such as invalid keyword arguments to 'setup()'."""
    pass

class DistutilsPlatformError (DistutilsError):
    """We don't know how to do something on the current platform (but
    we do know how to do it on some platform) -- eg. trying to compile
    C files on a platform not supported by a CCompiler subclass."""
    pass

class DistutilsExecError (DistutilsError):
    """Any problems executing an external program (such as the C
    compiler, when compiling C files)."""
    pass

class DistutilsInternalError (DistutilsError):
    """Internal inconsistencies or impossibilities (obviously, this
    should never be seen if the code is working!)."""
    pass

class DistutilsTemplateError (DistutilsError):
    """Syntax error in a file list template."""

class DistutilsByteCompileError(DistutilsError):
    """Byte compile error."""

# Exception classes used by the CCompiler implementation classes
class CCompilerError (Exception):
    """Some compile/link operation failed."""

class PreprocessError (CCompilerError):
    """Failure to preprocess one or more C/C++ files."""

class CompileError (CCompilerError):
    """Failure to compile one or more C/C++ source files."""

class LibError (CCompilerError):
    """Failure to create a static library from one or more C/C++ object
    files."""

class LinkError (CCompilerError):
    """Failure to link one or more C/C++ object files into an executable
    or shared library file."""

class UnknownFileError (CCompilerError):
    """Attempt to process an unknown file type."""
site-packages/setuptools/_distutils/core.py000064400000021254150564745010015205 0ustar00"""distutils.core

The only module that needs to be imported to use the Distutils; provides
the 'setup' function (which is to be called from the setup script).  Also
indirectly provides the Distribution and Command classes, although they are
really defined in distutils.dist and distutils.cmd.
"""

import os
import sys

from distutils.debug import DEBUG
from distutils.errors import *

# Mainly import these so setup scripts can "from distutils.core import" them.
from distutils.dist import Distribution
from distutils.cmd import Command
from distutils.config import PyPIRCCommand
from distutils.extension import Extension

# This is a barebones help message generated displayed when the user
# runs the setup script with no arguments at all.  More useful help
# is generated with various --help options: global help, list commands,
# and per-command help.
USAGE = """\
usage: %(script)s [global_opts] cmd1 [cmd1_opts] [cmd2 [cmd2_opts] ...]
   or: %(script)s --help [cmd1 cmd2 ...]
   or: %(script)s --help-commands
   or: %(script)s cmd --help
"""

def gen_usage (script_name):
    script = os.path.basename(script_name)
    return USAGE % vars()


# Some mild magic to control the behaviour of 'setup()' from 'run_setup()'.
_setup_stop_after = None
_setup_distribution = None

# Legal keyword arguments for the setup() function
setup_keywords = ('distclass', 'script_name', 'script_args', 'options',
                  'name', 'version', 'author', 'author_email',
                  'maintainer', 'maintainer_email', 'url', 'license',
                  'description', 'long_description', 'keywords',
                  'platforms', 'classifiers', 'download_url',
                  'requires', 'provides', 'obsoletes',
                  )

# Legal keyword arguments for the Extension constructor
extension_keywords = ('name', 'sources', 'include_dirs',
                      'define_macros', 'undef_macros',
                      'library_dirs', 'libraries', 'runtime_library_dirs',
                      'extra_objects', 'extra_compile_args', 'extra_link_args',
                      'swig_opts', 'export_symbols', 'depends', 'language')

def setup (**attrs):
    """The gateway to the Distutils: do everything your setup script needs
    to do, in a highly flexible and user-driven way.  Briefly: create a
    Distribution instance; find and parse config files; parse the command
    line; run each Distutils command found there, customized by the options
    supplied to 'setup()' (as keyword arguments), in config files, and on
    the command line.

    The Distribution instance might be an instance of a class supplied via
    the 'distclass' keyword argument to 'setup'; if no such class is
    supplied, then the Distribution class (in dist.py) is instantiated.
    All other arguments to 'setup' (except for 'cmdclass') are used to set
    attributes of the Distribution instance.

    The 'cmdclass' argument, if supplied, is a dictionary mapping command
    names to command classes.  Each command encountered on the command line
    will be turned into a command class, which is in turn instantiated; any
    class found in 'cmdclass' is used in place of the default, which is
    (for command 'foo_bar') class 'foo_bar' in module
    'distutils.command.foo_bar'.  The command class must provide a
    'user_options' attribute which is a list of option specifiers for
    'distutils.fancy_getopt'.  Any command-line options between the current
    and the next command are used to set attributes of the current command
    object.

    When the entire command-line has been successfully parsed, calls the
    'run()' method on each command object in turn.  This method will be
    driven entirely by the Distribution object (which each command object
    has a reference to, thanks to its constructor), and the
    command-specific options that became attributes of each command
    object.
    """

    global _setup_stop_after, _setup_distribution

    # Determine the distribution class -- either caller-supplied or
    # our Distribution (see below).
    klass = attrs.get('distclass')
    if klass:
        del attrs['distclass']
    else:
        klass = Distribution

    if 'script_name' not in attrs:
        attrs['script_name'] = os.path.basename(sys.argv[0])
    if 'script_args'  not in attrs:
        attrs['script_args'] = sys.argv[1:]

    # Create the Distribution instance, using the remaining arguments
    # (ie. everything except distclass) to initialize it
    try:
        _setup_distribution = dist = klass(attrs)
    except DistutilsSetupError as msg:
        if 'name' not in attrs:
            raise SystemExit("error in setup command: %s" % msg)
        else:
            raise SystemExit("error in %s setup command: %s" % \
                  (attrs['name'], msg))

    if _setup_stop_after == "init":
        return dist

    # Find and parse the config file(s): they will override options from
    # the setup script, but be overridden by the command line.
    dist.parse_config_files()

    if DEBUG:
        print("options (after parsing config files):")
        dist.dump_option_dicts()

    if _setup_stop_after == "config":
        return dist

    # Parse the command line and override config files; any
    # command-line errors are the end user's fault, so turn them into
    # SystemExit to suppress tracebacks.
    try:
        ok = dist.parse_command_line()
    except DistutilsArgError as msg:
        raise SystemExit(gen_usage(dist.script_name) + "\nerror: %s" % msg)

    if DEBUG:
        print("options (after parsing command line):")
        dist.dump_option_dicts()

    if _setup_stop_after == "commandline":
        return dist

    # And finally, run all the commands found on the command line.
    if ok:
        try:
            dist.run_commands()
        except KeyboardInterrupt:
            raise SystemExit("interrupted")
        except OSError as exc:
            if DEBUG:
                sys.stderr.write("error: %s\n" % (exc,))
                raise
            else:
                raise SystemExit("error: %s" % (exc,))

        except (DistutilsError,
                CCompilerError) as msg:
            if DEBUG:
                raise
            else:
                raise SystemExit("error: " + str(msg))

    return dist

# setup ()


def run_setup (script_name, script_args=None, stop_after="run"):
    """Run a setup script in a somewhat controlled environment, and
    return the Distribution instance that drives things.  This is useful
    if you need to find out the distribution meta-data (passed as
    keyword args from 'script' to 'setup()', or the contents of the
    config files or command-line.

    'script_name' is a file that will be read and run with 'exec()';
    'sys.argv[0]' will be replaced with 'script' for the duration of the
    call.  'script_args' is a list of strings; if supplied,
    'sys.argv[1:]' will be replaced by 'script_args' for the duration of
    the call.

    'stop_after' tells 'setup()' when to stop processing; possible
    values:
      init
        stop after the Distribution instance has been created and
        populated with the keyword arguments to 'setup()'
      config
        stop after config files have been parsed (and their data
        stored in the Distribution instance)
      commandline
        stop after the command-line ('sys.argv[1:]' or 'script_args')
        have been parsed (and the data stored in the Distribution)
      run [default]
        stop after all commands have been run (the same as if 'setup()'
        had been called in the usual way

    Returns the Distribution instance, which provides all information
    used to drive the Distutils.
    """
    if stop_after not in ('init', 'config', 'commandline', 'run'):
        raise ValueError("invalid value for 'stop_after': %r" % (stop_after,))

    global _setup_stop_after, _setup_distribution
    _setup_stop_after = stop_after

    save_argv = sys.argv.copy()
    g = {'__file__': script_name}
    try:
        try:
            sys.argv[0] = script_name
            if script_args is not None:
                sys.argv[1:] = script_args
            with open(script_name, 'rb') as f:
                exec(f.read(), g)
        finally:
            sys.argv = save_argv
            _setup_stop_after = None
    except SystemExit:
        # Hmm, should we do something if exiting with a non-zero code
        # (ie. error)?
        pass

    if _setup_distribution is None:
        raise RuntimeError(("'distutils.core.setup()' was never called -- "
               "perhaps '%s' is not a Distutils setup script?") % \
              script_name)

    # I wonder if the setup script's namespace -- g and l -- would be of
    # any interest to callers?
    #print "_setup_distribution:", _setup_distribution
    return _setup_distribution

# run_setup ()
site-packages/setuptools/_distutils/msvc9compiler.py000064400000073365150564745010017063 0ustar00"""distutils.msvc9compiler

Contains MSVCCompiler, an implementation of the abstract CCompiler class
for the Microsoft Visual Studio 2008.

The module is compatible with VS 2005 and VS 2008. You can find legacy support
for older versions of VS in distutils.msvccompiler.
"""

# Written by Perry Stoll
# hacked by Robin Becker and Thomas Heller to do a better job of
#   finding DevStudio (through the registry)
# ported to VS2005 and VS 2008 by Christian Heimes

import os
import subprocess
import sys
import re

from distutils.errors import DistutilsExecError, DistutilsPlatformError, \
                             CompileError, LibError, LinkError
from distutils.ccompiler import CCompiler, gen_lib_options
from distutils import log
from distutils.util import get_platform

import winreg

RegOpenKeyEx = winreg.OpenKeyEx
RegEnumKey = winreg.EnumKey
RegEnumValue = winreg.EnumValue
RegError = winreg.error

HKEYS = (winreg.HKEY_USERS,
         winreg.HKEY_CURRENT_USER,
         winreg.HKEY_LOCAL_MACHINE,
         winreg.HKEY_CLASSES_ROOT)

NATIVE_WIN64 = (sys.platform == 'win32' and sys.maxsize > 2**32)
if NATIVE_WIN64:
    # Visual C++ is a 32-bit application, so we need to look in
    # the corresponding registry branch, if we're running a
    # 64-bit Python on Win64
    VS_BASE = r"Software\Wow6432Node\Microsoft\VisualStudio\%0.1f"
    WINSDK_BASE = r"Software\Wow6432Node\Microsoft\Microsoft SDKs\Windows"
    NET_BASE = r"Software\Wow6432Node\Microsoft\.NETFramework"
else:
    VS_BASE = r"Software\Microsoft\VisualStudio\%0.1f"
    WINSDK_BASE = r"Software\Microsoft\Microsoft SDKs\Windows"
    NET_BASE = r"Software\Microsoft\.NETFramework"

# A map keyed by get_platform() return values to values accepted by
# 'vcvarsall.bat'.  Note a cross-compile may combine these (eg, 'x86_amd64' is
# the param to cross-compile on x86 targeting amd64.)
PLAT_TO_VCVARS = {
    'win32' : 'x86',
    'win-amd64' : 'amd64',
}

class Reg:
    """Helper class to read values from the registry
    """

    def get_value(cls, path, key):
        for base in HKEYS:
            d = cls.read_values(base, path)
            if d and key in d:
                return d[key]
        raise KeyError(key)
    get_value = classmethod(get_value)

    def read_keys(cls, base, key):
        """Return list of registry keys."""
        try:
            handle = RegOpenKeyEx(base, key)
        except RegError:
            return None
        L = []
        i = 0
        while True:
            try:
                k = RegEnumKey(handle, i)
            except RegError:
                break
            L.append(k)
            i += 1
        return L
    read_keys = classmethod(read_keys)

    def read_values(cls, base, key):
        """Return dict of registry keys and values.

        All names are converted to lowercase.
        """
        try:
            handle = RegOpenKeyEx(base, key)
        except RegError:
            return None
        d = {}
        i = 0
        while True:
            try:
                name, value, type = RegEnumValue(handle, i)
            except RegError:
                break
            name = name.lower()
            d[cls.convert_mbcs(name)] = cls.convert_mbcs(value)
            i += 1
        return d
    read_values = classmethod(read_values)

    def convert_mbcs(s):
        dec = getattr(s, "decode", None)
        if dec is not None:
            try:
                s = dec("mbcs")
            except UnicodeError:
                pass
        return s
    convert_mbcs = staticmethod(convert_mbcs)

class MacroExpander:

    def __init__(self, version):
        self.macros = {}
        self.vsbase = VS_BASE % version
        self.load_macros(version)

    def set_macro(self, macro, path, key):
        self.macros["$(%s)" % macro] = Reg.get_value(path, key)

    def load_macros(self, version):
        self.set_macro("VCInstallDir", self.vsbase + r"\Setup\VC", "productdir")
        self.set_macro("VSInstallDir", self.vsbase + r"\Setup\VS", "productdir")
        self.set_macro("FrameworkDir", NET_BASE, "installroot")
        try:
            if version >= 8.0:
                self.set_macro("FrameworkSDKDir", NET_BASE,
                               "sdkinstallrootv2.0")
            else:
                raise KeyError("sdkinstallrootv2.0")
        except KeyError:
            raise DistutilsPlatformError(
            """Python was built with Visual Studio 2008;
extensions must be built with a compiler than can generate compatible binaries.
Visual Studio 2008 was not found on this system. If you have Cygwin installed,
you can try compiling with MingW32, by passing "-c mingw32" to setup.py.""")

        if version >= 9.0:
            self.set_macro("FrameworkVersion", self.vsbase, "clr version")
            self.set_macro("WindowsSdkDir", WINSDK_BASE, "currentinstallfolder")
        else:
            p = r"Software\Microsoft\NET Framework Setup\Product"
            for base in HKEYS:
                try:
                    h = RegOpenKeyEx(base, p)
                except RegError:
                    continue
                key = RegEnumKey(h, 0)
                d = Reg.get_value(base, r"%s\%s" % (p, key))
                self.macros["$(FrameworkVersion)"] = d["version"]

    def sub(self, s):
        for k, v in self.macros.items():
            s = s.replace(k, v)
        return s

def get_build_version():
    """Return the version of MSVC that was used to build Python.

    For Python 2.3 and up, the version number is included in
    sys.version.  For earlier versions, assume the compiler is MSVC 6.
    """
    prefix = "MSC v."
    i = sys.version.find(prefix)
    if i == -1:
        return 6
    i = i + len(prefix)
    s, rest = sys.version[i:].split(" ", 1)
    majorVersion = int(s[:-2]) - 6
    if majorVersion >= 13:
        # v13 was skipped and should be v14
        majorVersion += 1
    minorVersion = int(s[2:3]) / 10.0
    # I don't think paths are affected by minor version in version 6
    if majorVersion == 6:
        minorVersion = 0
    if majorVersion >= 6:
        return majorVersion + minorVersion
    # else we don't know what version of the compiler this is
    return None

def normalize_and_reduce_paths(paths):
    """Return a list of normalized paths with duplicates removed.

    The current order of paths is maintained.
    """
    # Paths are normalized so things like:  /a and /a/ aren't both preserved.
    reduced_paths = []
    for p in paths:
        np = os.path.normpath(p)
        # XXX(nnorwitz): O(n**2), if reduced_paths gets long perhaps use a set.
        if np not in reduced_paths:
            reduced_paths.append(np)
    return reduced_paths

def removeDuplicates(variable):
    """Remove duplicate values of an environment variable.
    """
    oldList = variable.split(os.pathsep)
    newList = []
    for i in oldList:
        if i not in newList:
            newList.append(i)
    newVariable = os.pathsep.join(newList)
    return newVariable

def find_vcvarsall(version):
    """Find the vcvarsall.bat file

    At first it tries to find the productdir of VS 2008 in the registry. If
    that fails it falls back to the VS90COMNTOOLS env var.
    """
    vsbase = VS_BASE % version
    try:
        productdir = Reg.get_value(r"%s\Setup\VC" % vsbase,
                                   "productdir")
    except KeyError:
        log.debug("Unable to find productdir in registry")
        productdir = None

    if not productdir or not os.path.isdir(productdir):
        toolskey = "VS%0.f0COMNTOOLS" % version
        toolsdir = os.environ.get(toolskey, None)

        if toolsdir and os.path.isdir(toolsdir):
            productdir = os.path.join(toolsdir, os.pardir, os.pardir, "VC")
            productdir = os.path.abspath(productdir)
            if not os.path.isdir(productdir):
                log.debug("%s is not a valid directory" % productdir)
                return None
        else:
            log.debug("Env var %s is not set or invalid" % toolskey)
    if not productdir:
        log.debug("No productdir found")
        return None
    vcvarsall = os.path.join(productdir, "vcvarsall.bat")
    if os.path.isfile(vcvarsall):
        return vcvarsall
    log.debug("Unable to find vcvarsall.bat")
    return None

def query_vcvarsall(version, arch="x86"):
    """Launch vcvarsall.bat and read the settings from its environment
    """
    vcvarsall = find_vcvarsall(version)
    interesting = {"include", "lib", "libpath", "path"}
    result = {}

    if vcvarsall is None:
        raise DistutilsPlatformError("Unable to find vcvarsall.bat")
    log.debug("Calling 'vcvarsall.bat %s' (version=%s)", arch, version)
    popen = subprocess.Popen('"%s" %s & set' % (vcvarsall, arch),
                             stdout=subprocess.PIPE,
                             stderr=subprocess.PIPE)
    try:
        stdout, stderr = popen.communicate()
        if popen.wait() != 0:
            raise DistutilsPlatformError(stderr.decode("mbcs"))

        stdout = stdout.decode("mbcs")
        for line in stdout.split("\n"):
            line = Reg.convert_mbcs(line)
            if '=' not in line:
                continue
            line = line.strip()
            key, value = line.split('=', 1)
            key = key.lower()
            if key in interesting:
                if value.endswith(os.pathsep):
                    value = value[:-1]
                result[key] = removeDuplicates(value)

    finally:
        popen.stdout.close()
        popen.stderr.close()

    if len(result) != len(interesting):
        raise ValueError(str(list(result.keys())))

    return result

# More globals
VERSION = get_build_version()
if VERSION < 8.0:
    raise DistutilsPlatformError("VC %0.1f is not supported by this module" % VERSION)
# MACROS = MacroExpander(VERSION)

class MSVCCompiler(CCompiler) :
    """Concrete class that implements an interface to Microsoft Visual C++,
       as defined by the CCompiler abstract class."""

    compiler_type = 'msvc'

    # Just set this so CCompiler's constructor doesn't barf.  We currently
    # don't use the 'set_executables()' bureaucracy provided by CCompiler,
    # as it really isn't necessary for this sort of single-compiler class.
    # Would be nice to have a consistent interface with UnixCCompiler,
    # though, so it's worth thinking about.
    executables = {}

    # Private class data (need to distinguish C from C++ source for compiler)
    _c_extensions = ['.c']
    _cpp_extensions = ['.cc', '.cpp', '.cxx']
    _rc_extensions = ['.rc']
    _mc_extensions = ['.mc']

    # Needed for the filename generation methods provided by the
    # base class, CCompiler.
    src_extensions = (_c_extensions + _cpp_extensions +
                      _rc_extensions + _mc_extensions)
    res_extension = '.res'
    obj_extension = '.obj'
    static_lib_extension = '.lib'
    shared_lib_extension = '.dll'
    static_lib_format = shared_lib_format = '%s%s'
    exe_extension = '.exe'

    def __init__(self, verbose=0, dry_run=0, force=0):
        CCompiler.__init__ (self, verbose, dry_run, force)
        self.__version = VERSION
        self.__root = r"Software\Microsoft\VisualStudio"
        # self.__macros = MACROS
        self.__paths = []
        # target platform (.plat_name is consistent with 'bdist')
        self.plat_name = None
        self.__arch = None # deprecated name
        self.initialized = False

    def initialize(self, plat_name=None):
        # multi-init means we would need to check platform same each time...
        assert not self.initialized, "don't init multiple times"
        if plat_name is None:
            plat_name = get_platform()
        # sanity check for platforms to prevent obscure errors later.
        ok_plats = 'win32', 'win-amd64'
        if plat_name not in ok_plats:
            raise DistutilsPlatformError("--plat-name must be one of %s" %
                                         (ok_plats,))

        if "DISTUTILS_USE_SDK" in os.environ and "MSSdk" in os.environ and self.find_exe("cl.exe"):
            # Assume that the SDK set up everything alright; don't try to be
            # smarter
            self.cc = "cl.exe"
            self.linker = "link.exe"
            self.lib = "lib.exe"
            self.rc = "rc.exe"
            self.mc = "mc.exe"
        else:
            # On x86, 'vcvars32.bat amd64' creates an env that doesn't work;
            # to cross compile, you use 'x86_amd64'.
            # On AMD64, 'vcvars32.bat amd64' is a native build env; to cross
            # compile use 'x86' (ie, it runs the x86 compiler directly)
            if plat_name == get_platform() or plat_name == 'win32':
                # native build or cross-compile to win32
                plat_spec = PLAT_TO_VCVARS[plat_name]
            else:
                # cross compile from win32 -> some 64bit
                plat_spec = PLAT_TO_VCVARS[get_platform()] + '_' + \
                            PLAT_TO_VCVARS[plat_name]

            vc_env = query_vcvarsall(VERSION, plat_spec)

            self.__paths = vc_env['path'].split(os.pathsep)
            os.environ['lib'] = vc_env['lib']
            os.environ['include'] = vc_env['include']

            if len(self.__paths) == 0:
                raise DistutilsPlatformError("Python was built with %s, "
                       "and extensions need to be built with the same "
                       "version of the compiler, but it isn't installed."
                       % self.__product)

            self.cc = self.find_exe("cl.exe")
            self.linker = self.find_exe("link.exe")
            self.lib = self.find_exe("lib.exe")
            self.rc = self.find_exe("rc.exe")   # resource compiler
            self.mc = self.find_exe("mc.exe")   # message compiler
            #self.set_path_env_var('lib')
            #self.set_path_env_var('include')

        # extend the MSVC path with the current path
        try:
            for p in os.environ['path'].split(';'):
                self.__paths.append(p)
        except KeyError:
            pass
        self.__paths = normalize_and_reduce_paths(self.__paths)
        os.environ['path'] = ";".join(self.__paths)

        self.preprocess_options = None
        if self.__arch == "x86":
            self.compile_options = [ '/nologo', '/O2', '/MD', '/W3',
                                     '/DNDEBUG']
            self.compile_options_debug = ['/nologo', '/Od', '/MDd', '/W3',
                                          '/Z7', '/D_DEBUG']
        else:
            # Win64
            self.compile_options = [ '/nologo', '/O2', '/MD', '/W3', '/GS-' ,
                                     '/DNDEBUG']
            self.compile_options_debug = ['/nologo', '/Od', '/MDd', '/W3', '/GS-',
                                          '/Z7', '/D_DEBUG']

        self.ldflags_shared = ['/DLL', '/nologo', '/INCREMENTAL:NO']
        if self.__version >= 7:
            self.ldflags_shared_debug = [
                '/DLL', '/nologo', '/INCREMENTAL:no', '/DEBUG'
                ]
        self.ldflags_static = [ '/nologo']

        self.initialized = True

    # -- Worker methods ------------------------------------------------

    def object_filenames(self,
                         source_filenames,
                         strip_dir=0,
                         output_dir=''):
        # Copied from ccompiler.py, extended to return .res as 'object'-file
        # for .rc input file
        if output_dir is None: output_dir = ''
        obj_names = []
        for src_name in source_filenames:
            (base, ext) = os.path.splitext (src_name)
            base = os.path.splitdrive(base)[1] # Chop off the drive
            base = base[os.path.isabs(base):]  # If abs, chop off leading /
            if ext not in self.src_extensions:
                # Better to raise an exception instead of silently continuing
                # and later complain about sources and targets having
                # different lengths
                raise CompileError ("Don't know how to compile %s" % src_name)
            if strip_dir:
                base = os.path.basename (base)
            if ext in self._rc_extensions:
                obj_names.append (os.path.join (output_dir,
                                                base + self.res_extension))
            elif ext in self._mc_extensions:
                obj_names.append (os.path.join (output_dir,
                                                base + self.res_extension))
            else:
                obj_names.append (os.path.join (output_dir,
                                                base + self.obj_extension))
        return obj_names


    def compile(self, sources,
                output_dir=None, macros=None, include_dirs=None, debug=0,
                extra_preargs=None, extra_postargs=None, depends=None):

        if not self.initialized:
            self.initialize()
        compile_info = self._setup_compile(output_dir, macros, include_dirs,
                                           sources, depends, extra_postargs)
        macros, objects, extra_postargs, pp_opts, build = compile_info

        compile_opts = extra_preargs or []
        compile_opts.append ('/c')
        if debug:
            compile_opts.extend(self.compile_options_debug)
        else:
            compile_opts.extend(self.compile_options)

        for obj in objects:
            try:
                src, ext = build[obj]
            except KeyError:
                continue
            if debug:
                # pass the full pathname to MSVC in debug mode,
                # this allows the debugger to find the source file
                # without asking the user to browse for it
                src = os.path.abspath(src)

            if ext in self._c_extensions:
                input_opt = "/Tc" + src
            elif ext in self._cpp_extensions:
                input_opt = "/Tp" + src
            elif ext in self._rc_extensions:
                # compile .RC to .RES file
                input_opt = src
                output_opt = "/fo" + obj
                try:
                    self.spawn([self.rc] + pp_opts +
                               [output_opt] + [input_opt])
                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            elif ext in self._mc_extensions:
                # Compile .MC to .RC file to .RES file.
                #   * '-h dir' specifies the directory for the
                #     generated include file
                #   * '-r dir' specifies the target directory of the
                #     generated RC file and the binary message resource
                #     it includes
                #
                # For now (since there are no options to change this),
                # we use the source-directory for the include file and
                # the build directory for the RC file and message
                # resources. This works at least for win32all.
                h_dir = os.path.dirname(src)
                rc_dir = os.path.dirname(obj)
                try:
                    # first compile .MC to .RC and .H file
                    self.spawn([self.mc] +
                               ['-h', h_dir, '-r', rc_dir] + [src])
                    base, _ = os.path.splitext (os.path.basename (src))
                    rc_file = os.path.join (rc_dir, base + '.rc')
                    # then compile .RC to .RES file
                    self.spawn([self.rc] +
                               ["/fo" + obj] + [rc_file])

                except DistutilsExecError as msg:
                    raise CompileError(msg)
                continue
            else:
                # how to handle this file?
                raise CompileError("Don't know how to compile %s to %s"
                                   % (src, obj))

            output_opt = "/Fo" + obj
            try:
                self.spawn([self.cc] + compile_opts + pp_opts +
                           [input_opt, output_opt] +
                           extra_postargs)
            except DistutilsExecError as msg:
                raise CompileError(msg)

        return objects


    def create_static_lib(self,
                          objects,
                          output_libname,
                          output_dir=None,
                          debug=0,
                          target_lang=None):

        if not self.initialized:
            self.initialize()
        (objects, output_dir) = self._fix_object_args(objects, output_dir)
        output_filename = self.library_filename(output_libname,
                                                output_dir=output_dir)

        if self._need_link(objects, output_filename):
            lib_args = objects + ['/OUT:' + output_filename]
            if debug:
                pass # XXX what goes here?
            try:
                self.spawn([self.lib] + lib_args)
            except DistutilsExecError as msg:
                raise LibError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)


    def link(self,
             target_desc,
             objects,
             output_filename,
             output_dir=None,
             libraries=None,
             library_dirs=None,
             runtime_library_dirs=None,
             export_symbols=None,
             debug=0,
             extra_preargs=None,
             extra_postargs=None,
             build_temp=None,
             target_lang=None):

        if not self.initialized:
            self.initialize()
        (objects, output_dir) = self._fix_object_args(objects, output_dir)
        fixed_args = self._fix_lib_args(libraries, library_dirs,
                                        runtime_library_dirs)
        (libraries, library_dirs, runtime_library_dirs) = fixed_args

        if runtime_library_dirs:
            self.warn ("I don't know what to do with 'runtime_library_dirs': "
                       + str (runtime_library_dirs))

        lib_opts = gen_lib_options(self,
                                   library_dirs, runtime_library_dirs,
                                   libraries)
        if output_dir is not None:
            output_filename = os.path.join(output_dir, output_filename)

        if self._need_link(objects, output_filename):
            if target_desc == CCompiler.EXECUTABLE:
                if debug:
                    ldflags = self.ldflags_shared_debug[1:]
                else:
                    ldflags = self.ldflags_shared[1:]
            else:
                if debug:
                    ldflags = self.ldflags_shared_debug
                else:
                    ldflags = self.ldflags_shared

            export_opts = []
            for sym in (export_symbols or []):
                export_opts.append("/EXPORT:" + sym)

            ld_args = (ldflags + lib_opts + export_opts +
                       objects + ['/OUT:' + output_filename])

            # The MSVC linker generates .lib and .exp files, which cannot be
            # suppressed by any linker switches. The .lib files may even be
            # needed! Make sure they are generated in the temporary build
            # directory. Since they have different names for debug and release
            # builds, they can go into the same directory.
            build_temp = os.path.dirname(objects[0])
            if export_symbols is not None:
                (dll_name, dll_ext) = os.path.splitext(
                    os.path.basename(output_filename))
                implib_file = os.path.join(
                    build_temp,
                    self.library_filename(dll_name))
                ld_args.append ('/IMPLIB:' + implib_file)

            self.manifest_setup_ldargs(output_filename, build_temp, ld_args)

            if extra_preargs:
                ld_args[:0] = extra_preargs
            if extra_postargs:
                ld_args.extend(extra_postargs)

            self.mkpath(os.path.dirname(output_filename))
            try:
                self.spawn([self.linker] + ld_args)
            except DistutilsExecError as msg:
                raise LinkError(msg)

            # embed the manifest
            # XXX - this is somewhat fragile - if mt.exe fails, distutils
            # will still consider the DLL up-to-date, but it will not have a
            # manifest.  Maybe we should link to a temp file?  OTOH, that
            # implies a build environment error that shouldn't go undetected.
            mfinfo = self.manifest_get_embed_info(target_desc, ld_args)
            if mfinfo is not None:
                mffilename, mfid = mfinfo
                out_arg = '-outputresource:%s;%s' % (output_filename, mfid)
                try:
                    self.spawn(['mt.exe', '-nologo', '-manifest',
                                mffilename, out_arg])
                except DistutilsExecError as msg:
                    raise LinkError(msg)
        else:
            log.debug("skipping %s (up-to-date)", output_filename)

    def manifest_setup_ldargs(self, output_filename, build_temp, ld_args):
        # If we need a manifest at all, an embedded manifest is recommended.
        # See MSDN article titled
        # "How to: Embed a Manifest Inside a C/C++ Application"
        # (currently at http://msdn2.microsoft.com/en-us/library/ms235591(VS.80).aspx)
        # Ask the linker to generate the manifest in the temp dir, so
        # we can check it, and possibly embed it, later.
        temp_manifest = os.path.join(
                build_temp,
                os.path.basename(output_filename) + ".manifest")
        ld_args.append('/MANIFESTFILE:' + temp_manifest)

    def manifest_get_embed_info(self, target_desc, ld_args):
        # If a manifest should be embedded, return a tuple of
        # (manifest_filename, resource_id).  Returns None if no manifest
        # should be embedded.  See http://bugs.python.org/issue7833 for why
        # we want to avoid any manifest for extension modules if we can)
        for arg in ld_args:
            if arg.startswith("/MANIFESTFILE:"):
                temp_manifest = arg.split(":", 1)[1]
                break
        else:
            # no /MANIFESTFILE so nothing to do.
            return None
        if target_desc == CCompiler.EXECUTABLE:
            # by default, executables always get the manifest with the
            # CRT referenced.
            mfid = 1
        else:
            # Extension modules try and avoid any manifest if possible.
            mfid = 2
            temp_manifest = self._remove_visual_c_ref(temp_manifest)
        if temp_manifest is None:
            return None
        return temp_manifest, mfid

    def _remove_visual_c_ref(self, manifest_file):
        try:
            # Remove references to the Visual C runtime, so they will
            # fall through to the Visual C dependency of Python.exe.
            # This way, when installed for a restricted user (e.g.
            # runtimes are not in WinSxS folder, but in Python's own
            # folder), the runtimes do not need to be in every folder
            # with .pyd's.
            # Returns either the filename of the modified manifest or
            # None if no manifest should be embedded.
            manifest_f = open(manifest_file)
            try:
                manifest_buf = manifest_f.read()
            finally:
                manifest_f.close()
            pattern = re.compile(
                r"""<assemblyIdentity.*?name=("|')Microsoft\."""\
                r"""VC\d{2}\.CRT("|').*?(/>|</assemblyIdentity>)""",
                re.DOTALL)
            manifest_buf = re.sub(pattern, "", manifest_buf)
            pattern = r"<dependentAssembly>\s*</dependentAssembly>"
            manifest_buf = re.sub(pattern, "", manifest_buf)
            # Now see if any other assemblies are referenced - if not, we
            # don't want a manifest embedded.
            pattern = re.compile(
                r"""<assemblyIdentity.*?name=(?:"|')(.+?)(?:"|')"""
                r""".*?(?:/>|</assemblyIdentity>)""", re.DOTALL)
            if re.search(pattern, manifest_buf) is None:
                return None

            manifest_f = open(manifest_file, 'w')
            try:
                manifest_f.write(manifest_buf)
                return manifest_file
            finally:
                manifest_f.close()
        except OSError:
            pass

    # -- Miscellaneous methods -----------------------------------------
    # These are all used by the 'gen_lib_options() function, in
    # ccompiler.py.

    def library_dir_option(self, dir):
        return "/LIBPATH:" + dir

    def runtime_library_dir_option(self, dir):
        raise DistutilsPlatformError(
              "don't know how to set runtime library search path for MSVC++")

    def library_option(self, lib):
        return self.library_filename(lib)


    def find_library_file(self, dirs, lib, debug=0):
        # Prefer a debugging library if found (and requested), but deal
        # with it if we don't have one.
        if debug:
            try_names = [lib + "_d", lib]
        else:
            try_names = [lib]
        for dir in dirs:
            for name in try_names:
                libfile = os.path.join(dir, self.library_filename (name))
                if os.path.exists(libfile):
                    return libfile
        else:
            # Oops, didn't find it in *any* of 'dirs'
            return None

    # Helper methods for using the MSVC registry settings

    def find_exe(self, exe):
        """Return path to an MSVC executable program.

        Tries to find the program in several places: first, one of the
        MSVC program search paths from the registry; next, the directories
        in the PATH environment variable.  If any of those work, return an
        absolute path that is known to exist.  If none of them work, just
        return the original program name, 'exe'.
        """
        for p in self.__paths:
            fn = os.path.join(os.path.abspath(p), exe)
            if os.path.isfile(fn):
                return fn

        # didn't find it; try existing path
        for p in os.environ['Path'].split(';'):
            fn = os.path.join(os.path.abspath(p),exe)
            if os.path.isfile(fn):
                return fn

        return exe
site-packages/setuptools/_distutils/dir_util.py000064400000017142150564745010016071 0ustar00"""distutils.dir_util

Utility functions for manipulating directories and directory trees."""

import os
import errno
from distutils.errors import DistutilsFileError, DistutilsInternalError
from distutils import log

# cache for by mkpath() -- in addition to cheapening redundant calls,
# eliminates redundant "creating /foo/bar/baz" messages in dry-run mode
_path_created = {}

# I don't use os.makedirs because a) it's new to Python 1.5.2, and
# b) it blows up if the directory already exists (I want to silently
# succeed in that case).
def mkpath(name, mode=0o777, verbose=1, dry_run=0):
    """Create a directory and any missing ancestor directories.

    If the directory already exists (or if 'name' is the empty string, which
    means the current directory, which of course exists), then do nothing.
    Raise DistutilsFileError if unable to create some directory along the way
    (eg. some sub-path exists, but is a file rather than a directory).
    If 'verbose' is true, print a one-line summary of each mkdir to stdout.
    Return the list of directories actually created.
    """

    global _path_created

    # Detect a common bug -- name is None
    if not isinstance(name, str):
        raise DistutilsInternalError(
              "mkpath: 'name' must be a string (got %r)" % (name,))

    # XXX what's the better way to handle verbosity? print as we create
    # each directory in the path (the current behaviour), or only announce
    # the creation of the whole path? (quite easy to do the latter since
    # we're not using a recursive algorithm)

    name = os.path.normpath(name)
    created_dirs = []
    if os.path.isdir(name) or name == '':
        return created_dirs
    if _path_created.get(os.path.abspath(name)):
        return created_dirs

    (head, tail) = os.path.split(name)
    tails = [tail]                      # stack of lone dirs to create

    while head and tail and not os.path.isdir(head):
        (head, tail) = os.path.split(head)
        tails.insert(0, tail)          # push next higher dir onto stack

    # now 'head' contains the deepest directory that already exists
    # (that is, the child of 'head' in 'name' is the highest directory
    # that does *not* exist)
    for d in tails:
        #print "head = %s, d = %s: " % (head, d),
        head = os.path.join(head, d)
        abs_head = os.path.abspath(head)

        if _path_created.get(abs_head):
            continue

        if verbose >= 1:
            log.info("creating %s", head)

        if not dry_run:
            try:
                os.mkdir(head, mode)
            except OSError as exc:
                if not (exc.errno == errno.EEXIST and os.path.isdir(head)):
                    raise DistutilsFileError(
                          "could not create '%s': %s" % (head, exc.args[-1]))
            created_dirs.append(head)

        _path_created[abs_head] = 1
    return created_dirs

def create_tree(base_dir, files, mode=0o777, verbose=1, dry_run=0):
    """Create all the empty directories under 'base_dir' needed to put 'files'
    there.

    'base_dir' is just the name of a directory which doesn't necessarily
    exist yet; 'files' is a list of filenames to be interpreted relative to
    'base_dir'.  'base_dir' + the directory portion of every file in 'files'
    will be created if it doesn't already exist.  'mode', 'verbose' and
    'dry_run' flags are as for 'mkpath()'.
    """
    # First get the list of directories to create
    need_dir = set()
    for file in files:
        need_dir.add(os.path.join(base_dir, os.path.dirname(file)))

    # Now create them
    for dir in sorted(need_dir):
        mkpath(dir, mode, verbose=verbose, dry_run=dry_run)

def copy_tree(src, dst, preserve_mode=1, preserve_times=1,
              preserve_symlinks=0, update=0, verbose=1, dry_run=0):
    """Copy an entire directory tree 'src' to a new location 'dst'.

    Both 'src' and 'dst' must be directory names.  If 'src' is not a
    directory, raise DistutilsFileError.  If 'dst' does not exist, it is
    created with 'mkpath()'.  The end result of the copy is that every
    file in 'src' is copied to 'dst', and directories under 'src' are
    recursively copied to 'dst'.  Return the list of files that were
    copied or might have been copied, using their output name.  The
    return value is unaffected by 'update' or 'dry_run': it is simply
    the list of all files under 'src', with the names changed to be
    under 'dst'.

    'preserve_mode' and 'preserve_times' are the same as for
    'copy_file'; note that they only apply to regular files, not to
    directories.  If 'preserve_symlinks' is true, symlinks will be
    copied as symlinks (on platforms that support them!); otherwise
    (the default), the destination of the symlink will be copied.
    'update' and 'verbose' are the same as for 'copy_file'.
    """
    from distutils.file_util import copy_file

    if not dry_run and not os.path.isdir(src):
        raise DistutilsFileError(
              "cannot copy tree '%s': not a directory" % src)
    try:
        names = os.listdir(src)
    except OSError as e:
        if dry_run:
            names = []
        else:
            raise DistutilsFileError(
                  "error listing files in '%s': %s" % (src, e.strerror))

    if not dry_run:
        mkpath(dst, verbose=verbose)

    outputs = []

    for n in names:
        src_name = os.path.join(src, n)
        dst_name = os.path.join(dst, n)

        if n.startswith('.nfs'):
            # skip NFS rename files
            continue

        if preserve_symlinks and os.path.islink(src_name):
            link_dest = os.readlink(src_name)
            if verbose >= 1:
                log.info("linking %s -> %s", dst_name, link_dest)
            if not dry_run:
                os.symlink(link_dest, dst_name)
            outputs.append(dst_name)

        elif os.path.isdir(src_name):
            outputs.extend(
                copy_tree(src_name, dst_name, preserve_mode,
                          preserve_times, preserve_symlinks, update,
                          verbose=verbose, dry_run=dry_run))
        else:
            copy_file(src_name, dst_name, preserve_mode,
                      preserve_times, update, verbose=verbose,
                      dry_run=dry_run)
            outputs.append(dst_name)

    return outputs

def _build_cmdtuple(path, cmdtuples):
    """Helper for remove_tree()."""
    for f in os.listdir(path):
        real_f = os.path.join(path,f)
        if os.path.isdir(real_f) and not os.path.islink(real_f):
            _build_cmdtuple(real_f, cmdtuples)
        else:
            cmdtuples.append((os.remove, real_f))
    cmdtuples.append((os.rmdir, path))

def remove_tree(directory, verbose=1, dry_run=0):
    """Recursively remove an entire directory tree.

    Any errors are ignored (apart from being reported to stdout if 'verbose'
    is true).
    """
    global _path_created

    if verbose >= 1:
        log.info("removing '%s' (and everything under it)", directory)
    if dry_run:
        return
    cmdtuples = []
    _build_cmdtuple(directory, cmdtuples)
    for cmd in cmdtuples:
        try:
            cmd[0](cmd[1])
            # remove dir from cache if it's already there
            abspath = os.path.abspath(cmd[1])
            if abspath in _path_created:
                del _path_created[abspath]
        except OSError as exc:
            log.warn("error removing %s: %s", directory, exc)

def ensure_relative(path):
    """Take the full path 'path', and make it a relative path.

    This is useful to make 'path' the second argument to os.path.join().
    """
    drive, path = os.path.splitdrive(path)
    if path[0:1] == os.sep:
        path = drive + path[1:]
    return path
site-packages/setuptools/_distutils/file_util.py000064400000017724150564745010016240 0ustar00"""distutils.file_util

Utility functions for operating on single files.
"""

import os
from distutils.errors import DistutilsFileError
from distutils import log

# for generating verbose output in 'copy_file()'
_copy_action = { None:   'copying',
                 'hard': 'hard linking',
                 'sym':  'symbolically linking' }


def _copy_file_contents(src, dst, buffer_size=16*1024):
    """Copy the file 'src' to 'dst'; both must be filenames.  Any error
    opening either file, reading from 'src', or writing to 'dst', raises
    DistutilsFileError.  Data is read/written in chunks of 'buffer_size'
    bytes (default 16k).  No attempt is made to handle anything apart from
    regular files.
    """
    # Stolen from shutil module in the standard library, but with
    # custom error-handling added.
    fsrc = None
    fdst = None
    try:
        try:
            fsrc = open(src, 'rb')
        except OSError as e:
            raise DistutilsFileError("could not open '%s': %s" % (src, e.strerror))

        if os.path.exists(dst):
            try:
                os.unlink(dst)
            except OSError as e:
                raise DistutilsFileError(
                      "could not delete '%s': %s" % (dst, e.strerror))

        try:
            fdst = open(dst, 'wb')
        except OSError as e:
            raise DistutilsFileError(
                  "could not create '%s': %s" % (dst, e.strerror))

        while True:
            try:
                buf = fsrc.read(buffer_size)
            except OSError as e:
                raise DistutilsFileError(
                      "could not read from '%s': %s" % (src, e.strerror))

            if not buf:
                break

            try:
                fdst.write(buf)
            except OSError as e:
                raise DistutilsFileError(
                      "could not write to '%s': %s" % (dst, e.strerror))
    finally:
        if fdst:
            fdst.close()
        if fsrc:
            fsrc.close()

def copy_file(src, dst, preserve_mode=1, preserve_times=1, update=0,
              link=None, verbose=1, dry_run=0):
    """Copy a file 'src' to 'dst'.  If 'dst' is a directory, then 'src' is
    copied there with the same name; otherwise, it must be a filename.  (If
    the file exists, it will be ruthlessly clobbered.)  If 'preserve_mode'
    is true (the default), the file's mode (type and permission bits, or
    whatever is analogous on the current platform) is copied.  If
    'preserve_times' is true (the default), the last-modified and
    last-access times are copied as well.  If 'update' is true, 'src' will
    only be copied if 'dst' does not exist, or if 'dst' does exist but is
    older than 'src'.

    'link' allows you to make hard links (os.link) or symbolic links
    (os.symlink) instead of copying: set it to "hard" or "sym"; if it is
    None (the default), files are copied.  Don't set 'link' on systems that
    don't support it: 'copy_file()' doesn't check if hard or symbolic
    linking is available. If hardlink fails, falls back to
    _copy_file_contents().

    Under Mac OS, uses the native file copy function in macostools; on
    other systems, uses '_copy_file_contents()' to copy file contents.

    Return a tuple (dest_name, copied): 'dest_name' is the actual name of
    the output file, and 'copied' is true if the file was copied (or would
    have been copied, if 'dry_run' true).
    """
    # XXX if the destination file already exists, we clobber it if
    # copying, but blow up if linking.  Hmmm.  And I don't know what
    # macostools.copyfile() does.  Should definitely be consistent, and
    # should probably blow up if destination exists and we would be
    # changing it (ie. it's not already a hard/soft link to src OR
    # (not update) and (src newer than dst).

    from distutils.dep_util import newer
    from stat import ST_ATIME, ST_MTIME, ST_MODE, S_IMODE

    if not os.path.isfile(src):
        raise DistutilsFileError(
              "can't copy '%s': doesn't exist or not a regular file" % src)

    if os.path.isdir(dst):
        dir = dst
        dst = os.path.join(dst, os.path.basename(src))
    else:
        dir = os.path.dirname(dst)

    if update and not newer(src, dst):
        if verbose >= 1:
            log.debug("not copying %s (output up-to-date)", src)
        return (dst, 0)

    try:
        action = _copy_action[link]
    except KeyError:
        raise ValueError("invalid value '%s' for 'link' argument" % link)

    if verbose >= 1:
        if os.path.basename(dst) == os.path.basename(src):
            log.info("%s %s -> %s", action, src, dir)
        else:
            log.info("%s %s -> %s", action, src, dst)

    if dry_run:
        return (dst, 1)

    # If linking (hard or symbolic), use the appropriate system call
    # (Unix only, of course, but that's the caller's responsibility)
    elif link == 'hard':
        if not (os.path.exists(dst) and os.path.samefile(src, dst)):
            try:
                os.link(src, dst)
                return (dst, 1)
            except OSError:
                # If hard linking fails, fall back on copying file
                # (some special filesystems don't support hard linking
                #  even under Unix, see issue #8876).
                pass
    elif link == 'sym':
        if not (os.path.exists(dst) and os.path.samefile(src, dst)):
            os.symlink(src, dst)
            return (dst, 1)

    # Otherwise (non-Mac, not linking), copy the file contents and
    # (optionally) copy the times and mode.
    _copy_file_contents(src, dst)
    if preserve_mode or preserve_times:
        st = os.stat(src)

        # According to David Ascher <da@ski.org>, utime() should be done
        # before chmod() (at least under NT).
        if preserve_times:
            os.utime(dst, (st[ST_ATIME], st[ST_MTIME]))
        if preserve_mode:
            os.chmod(dst, S_IMODE(st[ST_MODE]))

    return (dst, 1)


# XXX I suspect this is Unix-specific -- need porting help!
def move_file (src, dst,
               verbose=1,
               dry_run=0):

    """Move a file 'src' to 'dst'.  If 'dst' is a directory, the file will
    be moved into it with the same name; otherwise, 'src' is just renamed
    to 'dst'.  Return the new full name of the file.

    Handles cross-device moves on Unix using 'copy_file()'.  What about
    other systems???
    """
    from os.path import exists, isfile, isdir, basename, dirname
    import errno

    if verbose >= 1:
        log.info("moving %s -> %s", src, dst)

    if dry_run:
        return dst

    if not isfile(src):
        raise DistutilsFileError("can't move '%s': not a regular file" % src)

    if isdir(dst):
        dst = os.path.join(dst, basename(src))
    elif exists(dst):
        raise DistutilsFileError(
              "can't move '%s': destination '%s' already exists" %
              (src, dst))

    if not isdir(dirname(dst)):
        raise DistutilsFileError(
              "can't move '%s': destination '%s' not a valid path" %
              (src, dst))

    copy_it = False
    try:
        os.rename(src, dst)
    except OSError as e:
        (num, msg) = e.args
        if num == errno.EXDEV:
            copy_it = True
        else:
            raise DistutilsFileError(
                  "couldn't move '%s' to '%s': %s" % (src, dst, msg))

    if copy_it:
        copy_file(src, dst, verbose=verbose)
        try:
            os.unlink(src)
        except OSError as e:
            (num, msg) = e.args
            try:
                os.unlink(dst)
            except OSError:
                pass
            raise DistutilsFileError(
                  "couldn't move '%s' to '%s' by copy/delete: "
                  "delete '%s' failed: %s"
                  % (src, dst, src, msg))
    return dst


def write_file (filename, contents):
    """Create a file with the specified name and write 'contents' (a
    sequence of strings without line terminators) to it.
    """
    f = open(filename, "w")
    try:
        for line in contents:
            f.write(line + "\n")
    finally:
        f.close()
site-packages/setuptools/_vendor/__pycache__/ordered_set.cpython-39.pyc000064400000040056150564745010022315 0ustar00a

(�Re;�@szdZddlZddlmZzddlmZmZWn"eyNddlmZmZYn0e	d�Z
dZdd�ZGdd	�d	ee�Z
dS)
z�
An OrderedSet is a custom MutableSet that remembers its order, so that every
entry has an index that can be looked up.

Based on a recipe originally posted to ActiveState Recipes by Raymond Hettiger,
and released under the MIT license.
�N)�deque)�
MutableSet�Sequencez3.1cCs"t|d�o t|t�o t|t�S)a

    Are we being asked to look up a list of things, instead of a single thing?
    We check for the `__iter__` attribute so that this can cover types that
    don't have to be known by this module, such as NumPy arrays.

    Strings, however, should be considered as atomic values to look up, not
    iterables. The same goes for tuples, since they are immutable and therefore
    valid entries.

    We don't need to check for the Python 2 `unicode` type, because it doesn't
    have an `__iter__` attribute anyway.
    �__iter__)�hasattr�
isinstance�str�tuple)�obj�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/ordered_set.py�is_iterables


�
�r
c@s�eZdZdZd;dd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZeZdd�Z
dd�ZeZeZdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�Zd)d*�Zd+d,�Zd-d.�Zd/d0�Zd1d2�Zd3d4�Zd5d6�Z d7d8�Z!d9d:�Z"dS)<�
OrderedSetz�
    An OrderedSet is a custom MutableSet that remembers its order, so that
    every entry has an index that can be looked up.

    Example:
        >>> OrderedSet([1, 1, 2, 3, 2])
        OrderedSet([1, 2, 3])
    NcCs g|_i|_|dur||O}dS�N)�items�map)�self�iterablerrr�__init__4szOrderedSet.__init__cCs
t|j�S)z�
        Returns the number of unique elements in the ordered set

        Example:
            >>> len(OrderedSet([]))
            0
            >>> len(OrderedSet([1, 2]))
            2
        )�lenr�rrrr�__len__:s
zOrderedSet.__len__cs|t|t�r|tkr���St|�r4�fdd�|D�St|d�sHt|t�rl�j|}t|t�rf��|�S|Snt	d|��dS)aQ
        Get the item at a given index.

        If `index` is a slice, you will get back that slice of items, as a
        new OrderedSet.

        If `index` is a list or a similar iterable, you'll get a list of
        items corresponding to those indices. This is similar to NumPy's
        "fancy indexing". The result is not an OrderedSet because you may ask
        for duplicate indices, and the number of elements returned should be
        the number of elements asked for.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset[1]
            2
        csg|]}�j|�qSr)r)�.0�irrr�
<listcomp>[�z*OrderedSet.__getitem__.<locals>.<listcomp>�	__index__z+Don't know how to index an OrderedSet by %rN)
r�slice�	SLICE_ALL�copyr
rr�list�	__class__�	TypeError)r�index�resultrrr�__getitem__Fs


zOrderedSet.__getitem__cCs
|�|�S)z�
        Return a shallow copy of this object.

        Example:
            >>> this = OrderedSet([1, 2, 3])
            >>> other = this.copy()
            >>> this == other
            True
            >>> this is other
            False
        )r!rrrrreszOrderedSet.copycCst|�dkrdSt|�SdS)Nrr)rr rrrr�__getstate__sszOrderedSet.__getstate__cCs"|dkr|�g�n
|�|�dS)Nr)r)r�staterrr�__setstate__szOrderedSet.__setstate__cCs
||jvS)z�
        Test if the item is in this ordered set

        Example:
            >>> 1 in OrderedSet([1, 3, 2])
            True
            >>> 5 in OrderedSet([1, 3, 2])
            False
        )r�r�keyrrr�__contains__�s
zOrderedSet.__contains__cCs0||jvr&t|j�|j|<|j�|�|j|S)aE
        Add `key` as an item to this OrderedSet, then return its index.

        If `key` is already in the OrderedSet, return the index it already
        had.

        Example:
            >>> oset = OrderedSet()
            >>> oset.append(3)
            0
            >>> print(oset)
            OrderedSet([3])
        )rrr�appendr)rrr�add�s
zOrderedSet.addcCsDd}z|D]}|�|�}q
Wn"ty>tdt|���Yn0|S)a<
        Update the set with the given iterable sequence, then return the index
        of the last element inserted.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.update([3, 1, 5, 1, 4])
            4
            >>> print(oset)
            OrderedSet([1, 2, 3, 5, 4])
        Nz(Argument needs to be an iterable, got %s)r-r"�
ValueError�type)r�sequenceZ
item_index�itemrrr�update�s
�
zOrderedSet.updatecs$t|�r�fdd�|D�S�j|S)aH
        Get the index of a given entry, raising an IndexError if it's not
        present.

        `key` can be an iterable of entries that is not a string, in which case
        this returns a list of indices.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.index(2)
            1
        csg|]}��|��qSr)r#)r�subkeyrrrr�rz$OrderedSet.index.<locals>.<listcomp>)r
rr)rrrr#�s
zOrderedSet.indexcCs,|jstd��|jd}|jd=|j|=|S)z�
        Remove and return the last element from the set.

        Raises KeyError if the set is empty.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.pop()
            3
        zSet is empty���)r�KeyErrorr)r�elemrrr�pop�s
zOrderedSet.popcCsP||vrL|j|}|j|=|j|=|j��D]\}}||kr,|d|j|<q,dS)a�
        Remove an element.  Do not raise an exception if absent.

        The MutableSet mixin uses this to implement the .remove() method, which
        *does* raise an error when asked to remove a non-existent item.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.discard(2)
            >>> print(oset)
            OrderedSet([1, 3])
            >>> oset.discard(2)
            >>> print(oset)
            OrderedSet([1, 3])
        �N)rr)rr*r�k�vrrr�discard�s
zOrderedSet.discardcCs|jdd�=|j��dS)z8
        Remove all items from this OrderedSet.
        N)rr�clearrrrrr<�szOrderedSet.clearcCs
t|j�S)zb
        Example:
            >>> list(iter(OrderedSet([1, 2, 3])))
            [1, 2, 3]
        )�iterrrrrrrszOrderedSet.__iter__cCs
t|j�S)zf
        Example:
            >>> list(reversed(OrderedSet([1, 2, 3])))
            [3, 2, 1]
        )�reversedrrrrr�__reversed__szOrderedSet.__reversed__cCs&|sd|jjfSd|jjt|�fS)Nz%s()z%s(%r))r!�__name__r rrrr�__repr__szOrderedSet.__repr__cCsPt|ttf�rt|�t|�kSzt|�}Wnty>YdS0t|�|kSdS)a�
        Returns true if the containers have the same items. If `other` is a
        Sequence, then order is checked, otherwise it is ignored.

        Example:
            >>> oset = OrderedSet([1, 3, 2])
            >>> oset == [1, 3, 2]
            True
            >>> oset == [1, 2, 3]
            False
            >>> oset == [2, 3]
            False
            >>> oset == OrderedSet([3, 2, 1])
            False
        FN)rrrr �setr")r�otherZother_as_setrrr�__eq__szOrderedSet.__eq__cGs<t|t�r|jnt}ttt�|g|��}tj�|�}||�S)a�
        Combines all unique items.
        Each items order is defined by its first appearance.

        Example:
            >>> oset = OrderedSet.union(OrderedSet([3, 1, 4, 1, 5]), [1, 3], [2, 0])
            >>> print(oset)
            OrderedSet([3, 1, 4, 5, 2, 0])
            >>> oset.union([8, 9])
            OrderedSet([3, 1, 4, 5, 2, 0, 8, 9])
            >>> oset | {10}
            OrderedSet([3, 1, 4, 5, 2, 0, 10])
        )rrr!rr �it�chain�
from_iterable)r�sets�clsZ
containersrrrr�union6szOrderedSet.unioncCs
|�|�Sr)�intersection�rrCrrr�__and__IszOrderedSet.__and__csHt|t�r|jnt}|r<tjtt|����fdd�|D�}n|}||�S)a�
        Returns elements in common between all sets. Order is defined only
        by the first set.

        Example:
            >>> oset = OrderedSet.intersection(OrderedSet([0, 1, 2, 3]), [1, 2, 3])
            >>> print(oset)
            OrderedSet([1, 2, 3])
            >>> oset.intersection([2, 4, 5], [1, 2, 3, 4])
            OrderedSet([2])
            >>> oset.intersection()
            OrderedSet([1, 2, 3])
        c3s|]}|�vr|VqdSrr�rr1��commonrr�	<genexpr>^rz*OrderedSet.intersection.<locals>.<genexpr>)rrr!rBrKr�rrHrIrrrOrrKMszOrderedSet.intersectioncs:|j}|r.tjtt|����fdd�|D�}n|}||�S)a�
        Returns all elements that are in this set but not the others.

        Example:
            >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]))
            OrderedSet([1, 3])
            >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]), OrderedSet([3]))
            OrderedSet([1])
            >>> OrderedSet([1, 2, 3]) - OrderedSet([2])
            OrderedSet([1, 3])
            >>> OrderedSet([1, 2, 3]).difference()
            OrderedSet([1, 2, 3])
        c3s|]}|�vr|VqdSrrrN�rCrrrQtrz(OrderedSet.difference.<locals>.<genexpr>)r!rBrJrrRrrSr�
differencecszOrderedSet.differencecs*t|�t��krdSt�fdd�|D��S)a7
        Report whether another set contains this set.

        Example:
            >>> OrderedSet([1, 2, 3]).issubset({1, 2})
            False
            >>> OrderedSet([1, 2, 3]).issubset({1, 2, 3, 4})
            True
            >>> OrderedSet([1, 2, 3]).issubset({1, 4, 3, 5})
            False
        Fc3s|]}|�vVqdSrrrNrSrrrQ�rz&OrderedSet.issubset.<locals>.<genexpr>�r�allrLrrSr�issubsetyszOrderedSet.issubsetcs*t��t|�krdSt�fdd�|D��S)a=
        Report whether this set contains another set.

        Example:
            >>> OrderedSet([1, 2]).issuperset([1, 2, 3])
            False
            >>> OrderedSet([1, 2, 3, 4]).issuperset({1, 2, 3})
            True
            >>> OrderedSet([1, 4, 3, 5]).issuperset({1, 2, 3})
            False
        Fc3s|]}|�vVqdSrrrNrrrrQ�rz(OrderedSet.issuperset.<locals>.<genexpr>rUrLrrr�
issuperset�szOrderedSet.issupersetcCs:t|t�r|jnt}||��|�}||��|�}|�|�S)a�
        Return the symmetric difference of two OrderedSets as a new set.
        That is, the new set will contain all elements that are in exactly
        one of the sets.

        Their order will be preserved, with elements from `self` preceding
        elements from `other`.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.symmetric_difference(other)
            OrderedSet([4, 5, 9, 2])
        )rrr!rTrJ)rrCrIZdiff1Zdiff2rrr�symmetric_difference�szOrderedSet.symmetric_differencecCs||_dd�t|�D�|_dS)zt
        Replace the 'items' list of this OrderedSet with a new one, updating
        self.map accordingly.
        cSsi|]\}}||�qSrr)r�idxr1rrr�
<dictcomp>�rz,OrderedSet._update_items.<locals>.<dictcomp>N)r�	enumerater)rrrrr�
_update_items�szOrderedSet._update_itemscs:t��|D]}�t|�O�q
|��fdd�|jD��dS)a�
        Update this OrderedSet to remove items from one or more other sets.

        Example:
            >>> this = OrderedSet([1, 2, 3])
            >>> this.difference_update(OrderedSet([2, 4]))
            >>> print(this)
            OrderedSet([1, 3])

            >>> this = OrderedSet([1, 2, 3, 4, 5])
            >>> this.difference_update(OrderedSet([2, 4]), OrderedSet([1, 4, 6]))
            >>> print(this)
            OrderedSet([3, 5])
        csg|]}|�vr|�qSrrrN��items_to_removerrr�rz0OrderedSet.difference_update.<locals>.<listcomp>N�rBr]r)rrHrCrr^r�difference_update�szOrderedSet.difference_updatecs&t���|��fdd�|jD��dS)a^
        Update this OrderedSet to keep only items in another set, preserving
        their order in this set.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.intersection_update(other)
            >>> print(this)
            OrderedSet([1, 3, 7])
        csg|]}|�vr|�qSrrrNrSrrr�rz2OrderedSet.intersection_update.<locals>.<listcomp>Nr`rLrrSr�intersection_update�szOrderedSet.intersection_updatecs<�fdd�|D�}t|�����fdd��jD�|�dS)a�
        Update this OrderedSet to remove items from another set, then
        add items from the other set that were not present in this set.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.symmetric_difference_update(other)
            >>> print(this)
            OrderedSet([4, 5, 9, 2])
        csg|]}|�vr|�qSrrrNrrrr�rz:OrderedSet.symmetric_difference_update.<locals>.<listcomp>csg|]}|�vr|�qSrrrNr^rrr�rNr`)rrCZitems_to_addr)r_rr�symmetric_difference_update�s
�z&OrderedSet.symmetric_difference_update)N)#r@�
__module__�__qualname__�__doc__rrr%rr&r(r+r-r,r2r#Zget_locZget_indexerr7r;r<rr?rArDrJrMrKrTrWrXrYr]rarbrcrrrrr*s@	
r)rf�	itertoolsrE�collectionsr�collections.abcrr�ImportErrorrr�__version__r
rrrrr�<module>ssite-packages/setuptools/_vendor/__pycache__/__init__.cpython-39.pyc000064400000000354150564745010021552 0ustar00a

(�Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/__init__.py�<module>�site-packages/setuptools/_vendor/__pycache__/pyparsing.cpython-39.pyc000064400000611256150564745010022040 0ustar00a

(�Rex��@s�
dZdZdZdZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZzddlmZWney�ddlmZYn0zdd	lmZdd
lmZWn*ey�dd	l
mZdd
l
mZYn0zddl
mZWn>e�y>zddlmZWne�y8dZYn0Yn0gd�Zee	j�dd
�Zedd
kZ e �r�e	j!Z"e#Z$e%Z&e#Z'e(e)e*e+e,ee-e.e/e0e1gZ2n^e	j3Z"e4Z5dd�Z'gZ2ddl6Z6d�7�D]6Z8ze2�9e:e6e8��Wne;�y�Y�q�Yn0�q�e<dd�e5d�D��Z=dd�Z>Gdd�de?�Z@ejAejBZCdZDeDdZEeCeDZFe%d�ZGd�Hdd�ejID��ZJGdd�deK�ZLGdd �d eL�ZMGd!d"�d"eL�ZNGd#d$�d$eN�ZOGd%d&�d&eK�ZPGd'd(�d(e?�ZQGd)d*�d*e?�ZRe�SeR�d+d,�ZTd-d.�ZUd/d0�ZVd1d2�ZWd3d4�ZXd5d6�ZYd7d8�ZZ�dd:d;�Z[Gd<d=�d=e?�Z\Gd>d?�d?e\�Z]Gd@dA�dAe]�Z^GdBdC�dCe]�Z_GdDdE�dEe]�Z`e`Zae`e\_bGdFdG�dGe]�ZcGdHdI�dIe`�ZdGdJdK�dKec�ZeGdLdM�dMe]�ZfGdNdO�dOe]�ZgGdPdQ�dQe]�ZhGdRdS�dSe]�ZiGdTdU�dUe]�ZjGdVdW�dWe]�ZkGdXdY�dYe]�ZlGdZd[�d[el�ZmGd\d]�d]el�ZnGd^d_�d_el�ZoGd`da�dael�ZpGdbdc�dcel�ZqGddde�deel�ZrGdfdg�dgel�ZsGdhdi�die\�ZtGdjdk�dket�ZuGdldm�dmet�ZvGdndo�doet�ZwGdpdq�dqet�ZxGdrds�dse\�ZyGdtdu�duey�ZzGdvdw�dwey�Z{Gdxdy�dyey�Z|Gdzd{�d{e|�Z}Gd|d}�d}e|�Z~Gd~d�de?�Ze�Z�Gd�d��d�ey�Z�Gd�d��d�ey�Z�Gd�d��d�ey�Z�Gd�d��d�e��Z�Gd�d��d�ey�Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e?�Z�d�d��Z��dd�d��Z��dd�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z��dd�d��Z�d�d��Z��dd�d��Z�d�d��Z�d�d��Z�e^���d��Z�en���d��Z�eo���d��Z�ep���d��Z�eq���d��Z�egeGd�d9d����d�d���Z�ehd����d�d���Z�ehd����d�d���Z�e�e�Be�Bejd�dd��BZ�e�e�e�d��e��Z�e`d��e�d����d��e�e}e�e�B����d��d�Z�d�dĄZ�d�dƄZ�d�dȄZ�d�dʄZ�d�d̄Z�e�d�d���Z�e�d�d���Z�d�dЄZ�d�d҄Z�d�dԄZ�d�dքZ�e?�e�_��dd�d؄Z�e@�Z�e?�e�_�e?�e�_�e�dكe�dڃfd�d܄Z�e�Z�e�ehd݃d����dߡZ�e�ehd�d����d�Z�e�ehd݃d�ehd�d�B���d�Z�e�ead�e������d�Z�d�d�de���fd�d�Z��dd�d�Z�e�d�Z�e�d�Z�e�egeCeFd����d��\Z�Z�e�e�d�7�d��Z�ehd�d�Heàġ�d����d�Z�d�d��Z�e�ehd��d����d��Z�ehd����d��Z�ehd���ɡ��d��Z�ehd����d��Z�e�ehd��d�e�B���d��Z�e�Z�eh�d����d�Z�e�e}egeJd��d�e�eg�d�e`d��eo�����ϡ���d�Z�e�e�e���e�Bd�d�����d�Z�G�d�d��d�Z�eӐd	k�
r�ed�d
�Z�ed�d�Z�egeCeF�d�Z�e�e֐d
d��d���e��Z�e�e�e׃����d�Zؐde�BZ�e�e֐d
d��d���e��Z�e�e�eڃ����d�Z�eԐd�eِd�e�eېd�Z�eܠݐd�e�jޠݐd�e�jߠݐd�e�j�ݐd�ddl�Z�e�j᠝e�e�j��e�j�ݐd�dS(a�	
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and executing simple grammars,
vs. the traditional lex/yacc approach, or the use of regular expressions.  With pyparsing, you
don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
provides a library of classes that you use to construct the grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form 
C{"<salutation>, <addressee>!"}), built up using L{Word}, L{Literal}, and L{And} elements 
(L{'+'<ParserElement.__add__>} operator gives L{And} expressions, strings are auto-converted to
L{Literal} expressions)::

    from pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the self-explanatory
class names, and the use of '+', '|' and '^' operators.

The L{ParseResults} object returned from L{ParserElement.parseString<ParserElement.parseString>} can be accessed as a nested list, a dictionary, or an
object with named attributes.

The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
 - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello  ,  World  !", etc.)
 - quoted strings
 - embedded comments


Getting Started -
-----------------
Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:
 - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes
 - construct character word-group expressions using the L{Word} class
 - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes
 - use L{'+'<And>}, L{'|'<MatchFirst>}, L{'^'<Or>}, and L{'&'<Each>} operators to combine simple expressions into more complex ones
 - associate names with your parsed results using L{ParserElement.setResultsName}
 - find some helpful expression short-cuts like L{delimitedList} and L{oneOf}
 - find more useful common expressions in the L{pyparsing_common} namespace class
z2.2.1z18 Sep 2018 00:49 UTCz*Paul McGuire <ptmcg@users.sourceforge.net>�N)�ref)�datetime)�RLock)�Iterable)�MutableMapping)�OrderedDict)i�And�CaselessKeyword�CaselessLiteral�
CharsNotIn�Combine�Dict�Each�Empty�
FollowedBy�Forward�
GoToColumn�Group�Keyword�LineEnd�	LineStart�Literal�
MatchFirst�NoMatch�NotAny�	OneOrMore�OnlyOnce�Optional�Or�ParseBaseException�ParseElementEnhance�ParseException�ParseExpression�ParseFatalException�ParseResults�ParseSyntaxException�
ParserElement�QuotedString�RecursiveGrammarException�Regex�SkipTo�	StringEnd�StringStart�Suppress�Token�TokenConverter�White�Word�WordEnd�	WordStart�
ZeroOrMore�	alphanums�alphas�
alphas8bit�anyCloseTag�
anyOpenTag�
cStyleComment�col�commaSeparatedList�commonHTMLEntity�countedArray�cppStyleComment�dblQuotedString�dblSlashComment�
delimitedList�dictOf�downcaseTokens�empty�hexnums�htmlComment�javaStyleComment�line�lineEnd�	lineStart�lineno�makeHTMLTags�makeXMLTags�matchOnlyAtCol�matchPreviousExpr�matchPreviousLiteral�
nestedExpr�nullDebugAction�nums�oneOf�opAssoc�operatorPrecedence�
printables�punc8bit�pythonStyleComment�quotedString�removeQuotes�replaceHTMLEntity�replaceWith�
restOfLine�sglQuotedString�srange�	stringEnd�stringStart�traceParseAction�
unicodeString�upcaseTokens�
withAttribute�
indentedBlock�originalTextFor�ungroup�
infixNotation�locatedExpr�	withClass�
CloseMatch�tokenMap�pyparsing_common�cCsdt|t�r|Sz
t|�WSty^t|��t��d�}td�}|�dd��|�	|�YS0dS)aDrop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
           str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
           then < returns the unicode object | encodes it with the default encoding | ... >.
        �xmlcharrefreplacez&#\d+;cSs$dtt|ddd���dd�S)Nz\ur����)�hex�int��t�ry��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/pyparsing.py�<lambda>��z_ustr.<locals>.<lambda>N)
�
isinstance�unicode�str�UnicodeEncodeError�encode�sys�getdefaultencodingr)�setParseAction�transformString)�obj�retZ
xmlcharrefryryrz�_ustr�s

r�z6sum len sorted reversed list tuple set any all min maxccs|]
}|VqdS�Nry)�.0�yryryrz�	<genexpr>�r|r��cCs:d}dd�d��D�}t||�D]\}}|�||�}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)�&�;Nry)r��sryryrzr��r|z_xml_escape.<locals>.<genexpr>zamp gt lt quot apos)�split�zip�replace)�data�from_symbols�
to_symbols�from_�to_ryryrz�_xml_escape�s
r�c@seZdZdS)�
_ConstantsN)�__name__�
__module__�__qualname__ryryryrzr��sr��
0123456789ZABCDEFabcdef�\�ccs|]}|tjvr|VqdSr�)�string�
whitespace�r��cryryrzr��r|c@sPeZdZdZddd�Zedd��Zdd	�Zd
d�Zdd
�Z	ddd�Z
dd�ZdS)rz7base exception class for all parsing runtime exceptionsrNcCs>||_|dur||_d|_n||_||_||_|||f|_dS�Nr�)�loc�msg�pstr�
parserElement�args)�selfr�r�r��elemryryrz�__init__�szParseBaseException.__init__cCs||j|j|j|j�S)z�
        internal factory method to simplify creating one type of ParseException 
        from another - avoids having __init__ signature conflicts among subclasses
        )r�r�r�r�)�cls�peryryrz�_from_exception�sz"ParseBaseException._from_exceptioncCsN|dkrt|j|j�S|dvr,t|j|j�S|dkrBt|j|j�St|��dS)z�supported attributes by name are:
            - lineno - returns the line number of the exception text
            - col - returns the column number of the exception text
            - line - returns the line containing the exception text
        rL)r;�columnrIN)rLr�r�r;rI�AttributeError)r��anameryryrz�__getattr__�szParseBaseException.__getattr__cCsd|j|j|j|jfS)Nz"%s (at char %d), (line:%d, col:%d))r�r�rLr��r�ryryrz�__str__�s�zParseBaseException.__str__cCst|�Sr��r�r�ryryrz�__repr__�szParseBaseException.__repr__�>!<cCs<|j}|jd}|r4d�|d|�|||d�f�}|��S)z�Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        r�r�N)rIr��join�strip)r��markerString�line_str�line_columnryryrz�
markInputline�s
�z ParseBaseException.markInputlinecCsd��tt|��S)Nzlineno col line)r��dir�typer�ryryrz�__dir__szParseBaseException.__dir__)rNN)r�)r�r�r��__doc__r��classmethodr�r�r�r�r�r�ryryryrzr�s



rc@seZdZdZdS)r!aN
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
     - lineno - returns the line number of the exception text
     - col - returns the column number of the exception text
     - line - returns the line containing the exception text
        
    Example::
        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))
            
    prints::
       Expected integer (at char 0), (line:1, col:1)
        column: 1
    N�r�r�r�r�ryryryrzr!sr!c@seZdZdZdS)r#znuser-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediatelyNr�ryryryrzr#sr#c@seZdZdZdS)r%z�just like L{ParseFatalException}, but thrown internally when an
       L{ErrorStop<And._ErrorStop>} ('-' operator) indicates that parsing is to stop 
       immediately because an unbacktrackable syntax error has been foundNr�ryryryrzr%sr%c@s eZdZdZdd�Zdd�ZdS)r(zZexception thrown by L{ParserElement.validate} if the grammar could be improperly recursivecCs
||_dSr���parseElementTrace�r��parseElementListryryrzr�4sz"RecursiveGrammarException.__init__cCs
d|jS)NzRecursiveGrammarException: %sr�r�ryryrzr�7sz!RecursiveGrammarException.__str__N)r�r�r�r�r�r�ryryryrzr(2sr(c@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�_ParseResultsWithOffsetcCs||f|_dSr���tup)r��p1�p2ryryrzr�;sz _ParseResultsWithOffset.__init__cCs
|j|Sr�r��r��iryryrz�__getitem__=sz#_ParseResultsWithOffset.__getitem__cCst|jd�S�Nr)�reprr�r�ryryrzr�?sz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSr�r�r�ryryrz�	setOffsetAsz!_ParseResultsWithOffset.setOffsetN)r�r�r�r�r�r�r�ryryryrzr�:sr�c@s�eZdZdZd[dd�Zddddefdd�Zdd	�Zefd
d�Zdd
�Z	dd�Z
dd�Zdd�ZeZ
dd�Zdd�Zdd�Zdd�Zdd�Zer�eZeZeZn$eZeZeZdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd\d(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Z d2d3�Z!d4d5�Z"d6d7�Z#d8d9�Z$d:d;�Z%d<d=�Z&d]d?d@�Z'dAdB�Z(dCdD�Z)dEdF�Z*d^dHdI�Z+dJdK�Z,dLdM�Z-d_dOdP�Z.dQdR�Z/dSdT�Z0dUdV�Z1dWdX�Z2dYdZ�Z3dS)`r$aI
    Structured parse results, to provide multiple means of access to the parsed data:
       - as a list (C{len(results)})
       - by list index (C{results[0], results[1]}, etc.)
       - by attribute (C{results.<resultsName>} - see L{ParserElement.setResultsName})

    Example::
        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)
    prints::
        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    NTcCs"t||�r|St�|�}d|_|S�NT)r}�object�__new__�_ParseResults__doinit)r��toklist�name�asList�modal�retobjryryrzr�ks


zParseResults.__new__c
Cs`|jrvd|_d|_d|_i|_||_||_|dur6g}||t�rP|dd�|_n||t�rft|�|_n|g|_t	�|_
|du�r\|�r\|s�d|j|<||t�r�t|�}||_||t
d�ttf�r�|ddgfv�s\||t�r�|g}|�r(||t��rt|��d�||<ntt|d�d�||<|||_n4z|d||<Wn"tttf�yZ|||<Yn0dS)NFrr�)r��_ParseResults__name�_ParseResults__parent�_ParseResults__accumNames�_ParseResults__asList�_ParseResults__modal�list�_ParseResults__toklist�_generatorType�dict�_ParseResults__tokdictrvr�r��
basestringr$r��copy�KeyError�	TypeError�
IndexError)r�r�r�r�r�r}ryryrzr�tsB



$
zParseResults.__init__cCsPt|ttf�r|j|S||jvr4|j|ddStdd�|j|D��SdS)NrtrcSsg|]}|d�qS�rry�r��vryryrz�
<listcomp>�r|z,ParseResults.__getitem__.<locals>.<listcomp>)r}rv�slicer�r�r�r$r�ryryrzr��s


zParseResults.__getitem__cCs�||t�r0|j�|t��|g|j|<|d}nD||ttf�rN||j|<|}n&|j�|t��t|d�g|j|<|}||t�r�t|�|_	dSr�)
r�r��getr�rvr�r�r$�wkrefr�)r��kr�r}�subryryrz�__setitem__�s


"
zParseResults.__setitem__c
Cs�t|ttf�r�t|j�}|j|=t|t�rH|dkr:||7}t||d�}tt|�|���}|��|j	�
�D]>\}}|D]0}t|�D]"\}\}}	t||	|	|k�||<q�qxqln|j	|=dS�Nrr�)
r}rvr��lenr�r��range�indices�reverser��items�	enumerater�)
r�r��mylen�removedr��occurrences�jr��value�positionryryrz�__delitem__�s

zParseResults.__delitem__cCs
||jvSr�)r�)r�r�ryryrz�__contains__�szParseResults.__contains__cCs
t|j�Sr�)r�r�r�ryryrz�__len__�r|zParseResults.__len__cCs
|jSr��r�r�ryryrz�__bool__�r|zParseResults.__bool__cCs
t|j�Sr���iterr�r�ryryrz�__iter__�r|zParseResults.__iter__cCst|jddd��S�Nrtr	r�ryryrz�__reversed__�r|zParseResults.__reversed__cCs$t|jd�r|j��St|j�SdS)N�iterkeys)�hasattrr�rr
r�ryryrz�	_iterkeys�s
zParseResults._iterkeyscs�fdd����D�S)Nc3s|]}�|VqdSr�ry�r�r�r�ryrzr��r|z+ParseResults._itervalues.<locals>.<genexpr>�rr�ryr�rz�_itervalues�szParseResults._itervaluescs�fdd����D�S)Nc3s|]}|�|fVqdSr�ryrr�ryrzr��r|z*ParseResults._iteritems.<locals>.<genexpr>rr�ryr�rz�
_iteritems�szParseResults._iteritemscCst|���S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)r�rr�ryryrz�keys�szParseResults.keyscCst|���S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r��
itervaluesr�ryryrz�values�szParseResults.valuescCst|���S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r��	iteritemsr�ryryrzr��szParseResults.itemscCs
t|j�S)z�Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names.)�boolr�r�ryryrz�haskeys�szParseResults.haskeyscOs�|s
dg}|��D]*\}}|dkr0|d|f}qtd|��qt|dt�sdt|�dksd|d|vr~|d}||}||=|S|d}|SdS)a�
        Removes and returns item at specified index (default=C{last}).
        Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
        argument or an integer argument, it will use C{list} semantics
        and pop tokens from the list of parsed tokens. If passed a 
        non-integer argument (most likely a string), it will use C{dict}
        semantics and pop the corresponding value from any defined 
        results names. A second default return value argument is 
        supported, just as in C{dict.pop()}.

        Example::
            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())
        prints::
            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        rt�defaultrz-pop() got an unexpected keyword argument '%s'r�N)r�r�r}rvr�)r�r��kwargsr�r��indexr��defaultvalueryryrz�pop�s""
�
�zParseResults.popcCs||vr||S|SdS)ai
        Returns named result matching the given key, or if there is no
        such name, then returns the given C{defaultValue} or C{None} if no
        C{defaultValue} is specified.

        Similar to C{dict.get()}.
        
        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        Nry)r��key�defaultValueryryrzr�3szParseResults.getcCsR|j�||�|j��D]4\}}t|�D]"\}\}}t||||k�||<q(qdS)a
        Inserts new element at location index in the list of parsed tokens.
        
        Similar to C{list.insert()}.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        N)r��insertr�r�r�r�)r�r�insStrr�rr�rrryryrzr"IszParseResults.insertcCs|j�|�dS)a�
        Add single element to end of ParseResults list of elements.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            
            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        N)r��append)r��itemryryrzr$]szParseResults.appendcCs$t|t�r||7}n|j�|�dS)a
        Add sequence of elements to end of ParseResults list of elements.

        Example::
            patt = OneOrMore(Word(alphas))
            
            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        N)r}r$r��extend)r��itemseqryryrzr&ks

zParseResults.extendcCs|jdd�=|j��dS)z7
        Clear all elements and results names.
        N)r�r��clearr�ryryrzr(}szParseResults.clearcCshz
||WStyYdS0||jvr`||jvrF|j|ddStdd�|j|D��SndSdS)Nr�rtrcSsg|]}|d�qSr�ryr�ryryrzr��r|z,ParseResults.__getattr__.<locals>.<listcomp>)r�r�r�r$�r�r�ryryrzr��s


zParseResults.__getattr__cCs|��}||7}|Sr��r�)r��otherr�ryryrz�__add__�szParseResults.__add__cs�|jrjt|j���fdd��|j��}�fdd�|D�}|D],\}}|||<t|dt�r<t|�|d_q<|j|j7_|j�	|j�|S)Ncs|dkr�S|�Sr�ry)�a)�offsetryrzr{�r|z'ParseResults.__iadd__.<locals>.<lambda>c	s4g|],\}}|D]}|t|d�|d��f�qqS�rr�)r��r�r��vlistr�)�	addoffsetryrzr��s�z)ParseResults.__iadd__.<locals>.<listcomp>r)
r�r�r�r�r}r$r�r�r��update)r�r+�
otheritems�otherdictitemsr�r�ry)r2r.rz�__iadd__�s


�zParseResults.__iadd__cCs&t|t�r|dkr|��S||SdSr�)r}rvr��r�r+ryryrz�__radd__�szParseResults.__radd__cCsdt|j�t|j�fS)Nz(%s, %s))r�r�r�r�ryryrzr��szParseResults.__repr__cCsdd�dd�|jD��dS)N�[�, css(|] }t|t�rt|�nt|�VqdSr�)r}r$r�r��r�r�ryryrzr��r|z'ParseResults.__str__.<locals>.<genexpr>�])r�r�r�ryryrzr��szParseResults.__str__r�cCsLg}|jD]<}|r |r |�|�t|t�r8||��7}q
|�t|��q
|Sr�)r�r$r}r$�
_asStringListr�)r��sep�outr%ryryrzr=�s


zParseResults._asStringListcCsdd�|jD�S)a�
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::
            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']
            
            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        cSs"g|]}t|t�r|��n|�qSry)r}r$r�)r��resryryrzr��r|z'ParseResults.asList.<locals>.<listcomp>rr�ryryrzr��szParseResults.asListcs6tr|j}n|j}�fdd��t�fdd�|�D��S)a�
        Returns the named parse results as a nested dictionary.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
            
            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        cs6t|t�r.|��r|��S�fdd�|D�Sn|SdS)Ncsg|]}�|��qSryryr���toItemryrzr��r|z7ParseResults.asDict.<locals>.toItem.<locals>.<listcomp>)r}r$r�asDict)r�rAryrzrB�s

z#ParseResults.asDict.<locals>.toItemc3s|]\}}|�|�fVqdSr�ry�r�r�r�rAryrzr��r|z&ParseResults.asDict.<locals>.<genexpr>)�PY_3r�rr�)r��item_fnryrArzrC�s
	zParseResults.asDictcCs8t|j�}|j��|_|j|_|j�|j�|j|_|S)zA
        Returns a new copy of a C{ParseResults} object.
        )r$r�r�r�r�r�r3r��r�r�ryryrzr��s
zParseResults.copyFcCsLd}g}tdd�|j��D��}|d}|s8d}d}d}d}	|durJ|}	n|jrV|j}	|	sf|rbdSd}	|||d|	d	g7}t|j�D]�\}
}t|t�r�|
|vr�||�||
|o�|du||�g7}n||�d|o�|du||�g7}q�d}|
|vr�||
}|�s|�rq�nd}t	t
|��}
|||d|d	|
d
|d	g	7}q�|||d
|	d	g7}d�|�S)z�
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        �
css(|] \}}|D]}|d|fVqqdS�r�Nryr0ryryrzr�s
�z%ParseResults.asXML.<locals>.<genexpr>�  r�N�ITEM�<�>�</)r�r�r�r�r�r�r}r$�asXMLr�r�r�)r��doctag�namedItemsOnly�indent�	formatted�nlr?�
namedItems�nextLevelIndent�selfTagr�r@�resTag�xmlBodyTextryryrzrO�sZ

�

�
�zParseResults.asXMLcCs:|j��D]*\}}|D]\}}||ur|Sqq
dSr�)r�r�)r�r�r�r1r�r�ryryrz�__lookup;s
zParseResults.__lookupcCs�|jr|jS|jr.|��}|r(|�|�SdSnNt|�dkrxt|j�dkrxtt|j����dddvrxtt|j����SdSdS)a(
        Returns the results name for this token expression. Useful when several 
        different expressions might match at a particular location.

        Example::
            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number") 
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)
            
            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])
        prints::
            age : 22
            ssn : 111-22-3333
            house_number : 221B
        Nr�r)rrt)	r�r��_ParseResults__lookupr�r��nextr
rr)r��parryryrz�getNameBs
��zParseResults.getNamercCsZg}d}|�|t|����|�rP|��r�tdd�|��D��}|D]r\}}|r\|�|�|�d|d||f�t|t�r�|r�|�|�||d��q�|�t|��qF|�t	|��qFn�t
dd�|D���rP|}t|�D]r\}	}
t|
t��r$|�d|d||	|d|d|
�||d�f�q�|�d|d||	|d|dt|
�f�q�d	�|�S)
aH
        Diagnostic method for listing out the contents of a C{ParseResults}.
        Accepts an optional C{indent} argument so that this string can be embedded
        in a nested display of other data.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(result.dump())
        prints::
            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        rHcss|]\}}t|�|fVqdSr�)rrDryryrzr�~r|z$ParseResults.dump.<locals>.<genexpr>z
%s%s- %s: rJr�css|]}t|t�VqdSr�)r}r$)r��vvryryrzr��r|z
%s%s[%d]:
%s%s%sr�)
r$r�r�r�sortedr�r}r$�dumpr��anyr�r�)r�rR�depth�fullr?�NLr�r�r�r�r_ryryrzrags,

4,zParseResults.dumpcOs tj|��g|�Ri|��dS)a�
        Pretty-printer for parsed results as a list, using the C{pprint} module.
        Accepts additional positional or keyword args as defined for the 
        C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})

        Example::
            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)
        prints::
            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        N)�pprintr��r�r�rryryrzrf�szParseResults.pprintcCs.|j|j��|jdur|��p d|j|jffSr�)r�r�r�r�r�r�r�ryryrz�__getstate__�s��zParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j�|�|durDt|�|_nd|_dSr�)r�r�r�r�r3r�r�)r��stater]�inAccumNamesryryrz�__setstate__�s
�zParseResults.__setstate__cCs|j|j|j|jfSr�)r�r�r�r�r�ryryrz�__getnewargs__�szParseResults.__getnewargs__cCstt|��t|���Sr�)r�r�r�rr�ryryrzr��szParseResults.__dir__)NNTT)N)r�)NFr�T)r�rT)4r�r�r�r�r�r}r�r�r�rrrr�__nonzero__rr
rrrrErrr�rrrrrr�r"r$r&r(r�r,r6r8r�r�r=r�rCr�rOr[r^rarfrhrkrlr�ryryryrzr$Dsh&
	'	
4

#
=%
-
r$cCsF|}d|krt|�kr4nn||ddkr4dS||�dd|�S)aReturns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   rr�rH)r��rfind)r��strgr�ryryrzr;�s
r;cCs|�dd|�dS)aReturns current line number within a string, counting newlines as line separators.
   The first line is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   rHrr�)�count)r�roryryrzrL�s
rLcCsF|�dd|�}|�d|�}|dkr2||d|�S||dd�SdS)zfReturns the line of text containing loc within a string, counting newlines as line separators.
       rHrr�N)rn�find)r�ro�lastCR�nextCRryryrzrI�s
rIcCs8tdt|�dt|�dt||�t||�f�dS)NzMatch z at loc z(%d,%d))�printr�rLr;)�instringr��exprryryrz�_defaultStartDebugAction�srwcCs$tdt|�dt|����dS)NzMatched z -> )rtr�rr�)ru�startloc�endlocrv�toksryryrz�_defaultSuccessDebugAction�sr{cCstdt|��dS)NzException raised:)rtr�)rur�rv�excryryrz�_defaultExceptionDebugAction�sr}cGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nry)r�ryryrzrS�srSrscs��tvr�fdd�Sdg�dg�tdd�dkrFddd�}dd	d
��ntj}tj�d}|dd�d
}|d|d|f�������fdd�}d}zt�dt�d�j�}Wnty�t��}Yn0||_|S)Ncs�|�Sr�ry�r��lrx)�funcryrzr{r|z_trim_arity.<locals>.<lambda>rFrs)rq�cSs8tdkrdnd}tj||dd�|}|dd�gS)N)rqr�r������r���limitrs)�system_version�	traceback�
extract_stack)r�r.�
frame_summaryryryrzr�sz"_trim_arity.<locals>.extract_stackcSs$tj||d�}|d}|dd�gS)Nr�rtrs)r��
extract_tb)�tbr��framesr�ryryrzr�sz_trim_arity.<locals>.extract_tb�r�rtr�c	s�z"�|�dd��}d�d<|WSty��dr<�n6z0t��d}�|dd�ddd��ksh�W~n~0�d�kr��dd7<Yq�Yq0qdS)NrTrtrsr�r�)r�r��exc_info)r�r�r��r��
foundArityr�r��maxargs�pa_call_line_synthryrz�wrapper-s z_trim_arity.<locals>.wrapperz<parse action>r��	__class__)r)r)	�singleArgBuiltinsr�r�r�r��getattrr��	Exceptionr)r�r�r��	LINE_DIFF�	this_liner��	func_nameryr�rz�_trim_aritys,

�r�cs�eZdZdZdZdZedd��Zedd��Zd�dd	�Z	d
d�Z
dd
�Zd�dd�Zd�dd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd�dd �Zd!d"�Zd�d#d$�Zd%d&�Zd'd(�ZGd)d*�d*e�Zed+ur�Gd,d-�d-e�ZnGd.d-�d-e�ZiZe�Zd/d/gZ d�d0d1�Z!eZ"ed2d3��Z#dZ$ed�d5d6��Z%d�d7d8�Z&e'dfd9d:�Z(d;d<�Z)e'fd=d>�Z*e'dfd?d@�Z+dAdB�Z,dCdD�Z-dEdF�Z.dGdH�Z/dIdJ�Z0dKdL�Z1dMdN�Z2dOdP�Z3dQdR�Z4dSdT�Z5dUdV�Z6dWdX�Z7dYdZ�Z8d�d[d\�Z9d]d^�Z:d_d`�Z;dadb�Z<dcdd�Z=dedf�Z>dgdh�Z?d�didj�Z@dkdl�ZAdmdn�ZBdodp�ZCdqdr�ZDgfdsdt�ZEd�dudv�ZF�fdwdx�ZGdydz�ZHd{d|�ZId}d~�ZJdd��ZKd�d�d��ZLd�d�d��ZM�ZNS)�r&z)Abstract base level parser element class.z 
	
FcCs
|t_dS)a�
        Overrides the default whitespace chars

        Example::
            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']
            
            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        N)r&�DEFAULT_WHITE_CHARS��charsryryrz�setDefaultWhitespaceCharsTs
z'ParserElement.setDefaultWhitespaceCharscCs
|t_dS)a�
        Set class to be used for inclusion of string literals into a parser.
        
        Example::
            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        N)r&�_literalStringClass)r�ryryrz�inlineLiteralsUsingcsz!ParserElement.inlineLiteralsUsingcCs�t�|_d|_d|_d|_||_d|_tj|_	d|_
d|_d|_t�|_
d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr�)NNN)r��parseAction�
failAction�strRepr�resultsName�
saveAsList�skipWhitespacer&r��
whiteChars�copyDefaultWhiteChars�mayReturnEmpty�keepTabs�ignoreExprs�debug�streamlined�
mayIndexError�errmsg�modalResults�debugActions�re�callPreparse�
callDuringTry)r��savelistryryrzr�xs(zParserElement.__init__cCs<t�|�}|jdd�|_|jdd�|_|jr8tj|_|S)a$
        Make a copy of this C{ParserElement}.  Useful for defining different parse actions
        for the same parsing pattern, using copies of the original parse element.
        
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
            
            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
        prints::
            [5120, 100, 655360, 268435456]
        Equivalent form of C{expr.copy()} is just C{expr()}::
            integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
        N)r�r�r�r�r&r�r�)r��cpyryryrzr��s
zParserElement.copycCs*||_d|j|_t|d�r&|j|j_|S)af
        Define name for this expression, makes debugging and exception messages clearer.
        
        Example::
            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        �	Expected �	exception)r�r�rr�r�r)ryryrz�setName�s


zParserElement.setNamecCs4|��}|�d�r"|dd�}d}||_||_|S)aP
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original C{ParserElement} object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        C{expr("name")} in place of C{expr.setResultsName("name")} - 
        see L{I{__call__}<__call__>}.

        Example::
            date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        �*NrtT)r��endswithr�r�)r�r��listAllMatches�newselfryryrz�setResultsName�s
zParserElement.setResultsNameTcs@|r&|j�d�fdd�	}�|_||_nt|jd�r<|jj|_|S)z�Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set C{breakFlag} to True to enable, False to
           disable.
        Tcsddl}|���||||�Sr�)�pdb�	set_trace)rur��	doActions�callPreParser���_parseMethodryrz�breaker�sz'ParserElement.setBreak.<locals>.breaker�_originalParseMethod)TT)�_parser�r)r��	breakFlagr�ryr�rz�setBreak�s
zParserElement.setBreakcOs&tttt|���|_|�dd�|_|S)a
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
        C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
         - s   = the original string being parsed (see note below)
         - loc = the location of the matching substring
         - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        Optional keyword arguments:
         - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See L{I{parseString}<parseString>} for more information
        on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
        consistent view of the parsed string, the parse location, and line and column
        positions within the parsed string.
        
        Example::
            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        r�F)r��mapr�r�r�r��r��fnsrryryrzr��s"zParserElement.setParseActioncOs4|jtttt|���7_|jp,|�dd�|_|S)z�
        Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}<setParseAction>}.
        
        See examples in L{I{copy}<copy>}.
        r�F)r�r�r�r�r�r�r�ryryrz�addParseActionszParserElement.addParseActioncs^|�dd��|�dd�rtnt�|D] ����fdd�}|j�|�q$|jpV|�dd�|_|S)a�Add a boolean predicate function to expression's list of parse actions. See 
        L{I{setParseAction}<setParseAction>} for function call signatures. Unlike C{setParseAction}, 
        functions passed to C{addCondition} need to return boolean success/fail of the condition.

        Optional keyword arguments:
         - message = define a custom message to be used in the raised exception
         - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
         
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        �messagezfailed user-defined condition�fatalFcs$tt��|||��s �||���dSr�)rr�r~��exc_type�fnr�ryrz�pa&sz&ParserElement.addCondition.<locals>.par�)r�r#r!r�r$r�)r�r�rr�ryr�rz�addConditionszParserElement.addConditioncCs
||_|S)aDefine action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           C{fn(s,loc,expr,err)} where:
            - s = string being parsed
            - loc = location where expression match was attempted and failed
            - expr = the parse expression that failed
            - err = the exception thrown
           The function returns no value.  It may throw C{L{ParseFatalException}}
           if it is desired to stop parsing immediately.)r�)r�r�ryryrz�
setFailAction-s
zParserElement.setFailActionc	CsLd}|rHd}|jD]2}z|�||�\}}d}qWqtyBYq0qq|S�NTF)r�r�r!)r�rur��
exprsFound�e�dummyryryrz�_skipIgnorables:s


zParserElement._skipIgnorablescCsH|jr|�||�}|jrD|j}t|�}||krD|||vrD|d7}q&|S�Nr�)r�r�r�r�r�)r�rur��wt�instrlenryryrz�preParseGs
zParserElement.preParsecCs|gfSr�ry�r�rur�r�ryryrz�	parseImplSszParserElement.parseImplcCs|Sr�ry�r�rur��	tokenlistryryrz�	postParseVszParserElement.postParsec
Cs�|j}|s|jr�|jdr,|jd|||�|rD|jrD|�||�}n|}|}zBz|�|||�\}}Wn&ty�t|t|�|j	|��Yn0WnZt
y�}	zB|jdr�|jd||||	�|jr�|�||||	��WYd}	~	n
d}	~	00n�|�r|j�r|�||�}n|}|}|j�s&|t|�k�rhz|�|||�\}}Wn(t�ydt|t|�|j	|��Yn0n|�|||�\}}|�|||�}t
||j|j|jd�}
|j�r�|�s�|j�r�|�rTzN|jD]B}||||
�}|du�r�t
||j|j�o�t|t
tf�|jd�}
�q�WnHt
�yP}	z.|jd�r:|jd||||	��WYd}	~	n
d}	~	00nJ|jD]B}||||
�}|du�rZt
||j|j�o�t|t
tf�|jd�}
�qZ|�r�|jd�r�|jd|||||
�||
fS)Nrrs)r�r�r�)r�r�r�r�r�r�r�r!r�r�rr�r�r$r�r�r�r�r�r}r�)r�rur�r�r��	debugging�preloc�tokensStart�tokens�err�	retTokensr�ryryrz�
_parseNoCacheZst





�

�
zParserElement._parseNoCachecCs>z|j||dd�dWSty8t|||j|��Yn0dS)NF)r�r)r�r#r!r��r�rur�ryryrz�tryParse�szParserElement.tryParsec	Cs2z|�||�Wnttfy(YdS0dSdS)NFT)r�r!r�r�ryryrz�canParseNext�s
zParserElement.canParseNextc@seZdZdd�ZdS)zParserElement._UnboundedCachecs~i�t�|_���fdd�}�fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr��r��r�r ��cache�not_in_cacheryrzr��sz3ParserElement._UnboundedCache.__init__.<locals>.getcs|�|<dSr�ry�r�r r�r�ryrz�set�sz3ParserElement._UnboundedCache.__init__.<locals>.setcs���dSr��r(r�r�ryrzr(�sz5ParserElement._UnboundedCache.__init__.<locals>.clearcst��Sr��r�r�r�ryrz�	cache_len�sz9ParserElement._UnboundedCache.__init__.<locals>.cache_len)r�r��types�
MethodTyper�r�r(r)r�r�r�r(r�ryr�rzr��sz&ParserElement._UnboundedCache.__init__N�r�r�r�r�ryryryrz�_UnboundedCache�sr�Nc@seZdZdd�ZdS)�ParserElement._FifoCachecs�t�|_�t����fdd�}��fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr�r�r�r�ryrzr��s�.ParserElement._FifoCache.__init__.<locals>.getcs<|�|<t���kr8z��d�Wqty4Yq0qdS�NF)r��popitemr�r�)r��sizeryrzr��s�.ParserElement._FifoCache.__init__.<locals>.setcs���dSr�r�r�r�ryrzr(�s�0ParserElement._FifoCache.__init__.<locals>.clearcst��Sr�r�r�r�ryrzr��s�4ParserElement._FifoCache.__init__.<locals>.cache_len)	r�r��_OrderedDictr�r�r�r�r(r�r�rr�r�r(r�ry)r�r�rrzr��s�!ParserElement._FifoCache.__init__Nr�ryryryrz�
_FifoCache�sr	c@seZdZdd�ZdS)r�cs�t�|_�i�t�g�����fdd�}���fdd�}��fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_	dS)	Ncs��|��Sr�r�r�r�ryrzr��sr�cs4|�|<t���kr&�����d�q��|�dSr�)r�r�popleftr$r�)r��key_fiforryrzr��srcs������dSr�r�r�)r�rryrzr(�srcst��Sr�r�r�r�ryrzr��sr)
r�r��collections�dequer�r�r�r�r(rrry)r�rr�rrzr��srNr�ryryryrzr	�srcCs0d\}}|||||f}tj��tj}|�|�}	|	|jur�tj|d7<z|�||||�}	Wn:ty�}
z"|�||
j	|
j
���WYd}
~
n8d}
~
00|�||	d|	d��f�|	Wd�SnBtj|d7<t|	t
�r�|	�|	d|	d��fWd�SWd�n1�s"0YdS)Nr/r�r)r&�packrat_cache_lock�
packrat_cacher�r��packrat_cache_statsr�rr�r�r�r�r}r�)r�rur�r�r��HIT�MISS�lookupr�rr�ryryrz�_parseCaches$


zParserElement._parseCachecCs(tj��dgttj�tjdd�<dSr�)r&rr(r�rryryryrz�
resetCaches
zParserElement.resetCache�cCs8tjs4dt_|dur t��t_nt�|�t_tjt_dS)a�Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.
           
           Parameters:
            - cache_size_limit - (default=C{128}) - if an integer value is provided
              will limit the size of the packrat cache; if None is passed, then
              the cache size will be unbounded; if 0 is passed, the cache will
              be effectively disabled.
            
           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method C{ParserElement.enablePackrat()}.  If
           your program uses C{psyco} to "compile as you go", you must call
           C{enablePackrat} before calling C{psyco.full()}.  If you do not do this,
           Python will crash.  For best results, call C{enablePackrat()} immediately
           after importing pyparsing.
           
           Example::
               import pyparsing
               pyparsing.ParserElement.enablePackrat()
        TN)r&�_packratEnabledr�rr	rr�)�cache_size_limitryryrz�
enablePackrat%szParserElement.enablePackratc
Cs�t��|js|��|jD]}|��q|js8|��}z<|�|d�\}}|rr|�||�}t	�t
�}|�||�Wn2ty�}ztjr��n|�WYd}~nd}~00|SdS)aC
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set C{parseAll} to True (equivalent to ending
        the grammar with C{L{StringEnd()}}).

        Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the C{loc} argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:
         - calling C{parseWithTabs} on your grammar before calling C{parseString}
           (see L{I{parseWithTabs}<parseWithTabs>})
         - define your parse action using the full C{(s,loc,toks)} signature, and
           reference the input string using the parse action's C{s} argument
         - explicitly expand the tabs in your input string before calling
           C{parseString}
        
        Example::
            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        rN)
r&rr��
streamliner�r��
expandtabsr�r�rr+r�verbose_stacktrace)r�ru�parseAllr�r�r��ser|ryryrz�parseStringHs$

zParserElement.parseStringc
cs6|js|��|jD]}|��q|js4t|���}t|�}d}|j}|j}t	�
�d}	z�||kr�|	|kr�z |||�}
|||
dd�\}}Wnty�|
d}YqZ0||kr�|	d7}	||
|fV|r�|||�}
|
|kr�|}q�|d7}q�|}qZ|
d}qZWn6t�y0}zt	j
�r�n|�WYd}~n
d}~00dS)a�
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        C{maxMatches} argument, to clip scanning after 'n' matches are found.  If
        C{overlap} is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See L{I{parseString}<parseString>} for more information on parsing
        strings with embedded tabs.

        Example::
            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens,start,end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])
        
        prints::
        
            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        rF�r�r�N)r�rr�r�r�rr�r�r�r&rr!rr)r�ru�
maxMatches�overlapr�r�r��
preparseFn�parseFn�matchesr��nextLocr��nextlocr|ryryrz�
scanStringzsB




zParserElement.scanStringc
Cs�g}d}d|_z�|�|�D]Z\}}}|�|||��|rpt|t�rR||��7}nt|t�rf||7}n
|�|�|}q|�||d��dd�|D�}d�tt	t
|���WSty�}ztj
rĂn|�WYd}~n
d}~00dS)af
        Extension to C{L{scanString}}, to modify matching text with modified tokens that may
        be returned from a parse action.  To use C{transformString}, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking C{transformString()} on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  C{transformString()} returns the resulting transformed string.
        
        Example::
            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())
            
            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
        Prints::
            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        rTNcSsg|]}|r|�qSryry)r��oryryrzr��r|z1ParserElement.transformString.<locals>.<listcomp>r�)r�r(r$r}r$r�r�r�r�r��_flattenrr&r)r�rur?�lastErxr�r�r|ryryrzr��s(



zParserElement.transformStringc
CsTztdd�|�||�D��WStyN}ztjr6�n|�WYd}~n
d}~00dS)a�
        Another extension to C{L{scanString}}, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        C{maxMatches} argument, to clip searching after 'n' matches are found.
        
        Example::
            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())
            
            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))
        prints::
            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        cSsg|]\}}}|�qSryry)r�rxr�r�ryryrzr��r|z.ParserElement.searchString.<locals>.<listcomp>N)r$r(rr&r)r�rur!r|ryryrz�searchString�szParserElement.searchStringc	csTd}d}|j||d�D]*\}}}|||�V|r<|dV|}q||d�VdS)a[
        Generator method to split a string using the given expression as a separator.
        May be called with optional C{maxsplit} argument, to limit the number of splits;
        and the optional C{includeSeparators} argument (default=C{False}), if the separating
        matching text should be included in the split results.
        
        Example::        
            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
        prints::
            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        r)r!N)r()	r�ru�maxsplit�includeSeparators�splits�lastrxr�r�ryryrzr�s

zParserElement.splitcCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)a�
        Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
        converts them to L{Literal}s by default.
        
        Example::
            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))
        Prints::
            Hello, World! -> ['Hello', ',', 'World', '!']
        �4Cannot combine element of type %s with ParserElementrs��
stacklevelN)	r}r�r&r��warnings�warnr��
SyntaxWarningrr7ryryrzr,s


�zParserElement.__add__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||S)z]
        Implementation of + operator when left operand is not a C{L{ParserElement}}
        r1rsr2N�r}r�r&r�r4r5r�r6r7ryryrzr81s


�zParserElement.__radd__cCsJt|t�rt�|�}t|t�s:tjdt|�tdd�dS|t�	�|S)zQ
        Implementation of - operator, returns C{L{And}} with error stop
        r1rsr2N)
r}r�r&r�r4r5r�r6r�
_ErrorStopr7ryryrz�__sub__=s


�zParserElement.__sub__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||S)z]
        Implementation of - operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rsub__Is


�zParserElement.__rsub__cs�t|t�r|d}}n�t|t�r�|ddd�}|ddurHd|df}t|dt�r�|ddur�|ddkrvt��S|ddkr�t��S�|dt��Sq�t|dt�r�t|dt�r�|\}}||8}q�tdt|d�t|d���ntdt|���|dk�rtd��|dk�rtd	��||k�r6dk�rBnntd
��|�r���fdd��|�r�|dk�rt��|�}nt�g|��|�}n�|�}n|dk�r��}nt�g|�}|S)
a�
        Implementation of * operator, allows use of C{expr * 3} in place of
        C{expr + expr + expr}.  Expressions may also me multiplied by a 2-integer
        tuple, similar to C{{min,max}} multipliers in regular expressions.  Tuples
        may also include C{None} as in:
         - C{expr*(n,None)} or C{expr*(n,)} is equivalent
              to C{expr*n + L{ZeroOrMore}(expr)}
              (read as "at least n instances of C{expr}")
         - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
              (read as "0 to n instances of C{expr}")
         - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
         - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}

        Note that C{expr*(None,n)} does not raise an exception if
        more than n exprs exist in the input stream; that is,
        C{expr*(None,n)} does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        C{expr*(None,n) + ~expr}
        r)NNNrsr�z7cannot multiply 'ParserElement' and ('%s','%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez+cannot multiply ParserElement by 0 or (0,0)cs(|dkrt��|d��St��SdSr�)r��n��makeOptionalListr�ryrzr>�sz/ParserElement.__mul__.<locals>.makeOptionalList)	r}rv�tupler4rr�r��
ValueErrorr)r�r+�minElements�optElementsr�ryr=rz�__mul__UsD







zParserElement.__mul__cCs
|�|�Sr�)rCr7ryryrz�__rmul__�szParserElement.__rmul__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zI
        Implementation of | operator - returns C{L{MatchFirst}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__or__�s


�zParserElement.__or__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||BS)z]
        Implementation of | operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__ror__�s


�zParserElement.__ror__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zA
        Implementation of ^ operator - returns C{L{Or}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__xor__�s


�zParserElement.__xor__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||AS)z]
        Implementation of ^ operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rxor__�s


�zParserElement.__rxor__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zC
        Implementation of & operator - returns C{L{Each}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__and__�s


�zParserElement.__and__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||@S)z]
        Implementation of & operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rand__�s


�zParserElement.__rand__cCst|�S)zE
        Implementation of ~ operator - returns C{L{NotAny}}
        )rr�ryryrz�
__invert__�szParserElement.__invert__cCs|dur|�|�S|��SdS)a

        Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
        
        If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
        passed as C{True}.
           
        If C{name} is omitted, same as calling C{L{copy}}.

        Example::
            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")             
        N)r�r�r)ryryrz�__call__�s
zParserElement.__call__cCst|�S)z�
        Suppresses the output of this C{ParserElement}; useful to keep punctuation from
        cluttering up returned output.
        )r-r�ryryrz�suppress�szParserElement.suppresscCs
d|_|S)a
        Disables the skipping of whitespace before matching the characters in the
        C{ParserElement}'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        F�r�r�ryryrz�leaveWhitespaceszParserElement.leaveWhitespacecCsd|_||_d|_|S)z8
        Overrides the default whitespace chars
        TF)r�r�r�)r�r�ryryrz�setWhitespaceChars
sz ParserElement.setWhitespaceCharscCs
d|_|S)z�
        Overrides default behavior to expand C{<TAB>}s to spaces before parsing the input string.
        Must be called before C{parseString} when the input grammar contains elements that
        match C{<TAB>} characters.
        T)r�r�ryryrz�
parseWithTabsszParserElement.parseWithTabscCsLt|t�rt|�}t|t�r4||jvrH|j�|�n|j�t|����|S)a�
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.
        
        Example::
            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
            
            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        )r}r�r-r�r$r�r7ryryrz�ignores


zParserElement.ignorecCs"|pt|pt|ptf|_d|_|S)zT
        Enable display of debugging messages while doing pattern matching.
        T)rwr{r}r�r�)r��startAction�
successAction�exceptionActionryryrz�setDebugActions6s�zParserElement.setDebugActionscCs|r|�ttt�nd|_|S)a�
        Enable display of debugging messages while doing pattern matching.
        Set C{flag} to True to enable, False to disable.

        Example::
            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer
            
            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")
        
        prints::
            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using L{setDebugActions}. Prior to attempting
        to match the C{wd} expression, the debugging message C{"Match <exprname> at loc <n>(<line>,<col>)"}
        is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
        message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
        F)rVrwr{r}r�)r��flagryryrz�setDebug@s#zParserElement.setDebugcCs|jSr�)r�r�ryryrzr�iszParserElement.__str__cCst|�Sr�r�r�ryryrzr�lszParserElement.__repr__cCsd|_d|_|Sr�)r�r�r�ryryrzroszParserElement.streamlinecCsdSr�ryr�ryryrz�checkRecursiontszParserElement.checkRecursioncCs|�g�dS)zj
        Check defined expressions for valid structure, check for infinite recursive definitions.
        N)rY)r��
validateTraceryryrz�validatewszParserElement.validatecCs�z|��}WnDtyPt|d��}|��}Wd�n1sB0YYn0z|�||�WSty�}ztjrz�n|�WYd}~n
d}~00dS)z�
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        �rN)�readr��openrrr&r)r��file_or_filenamer�
file_contents�fr|ryryrz�	parseFile}s,zParserElement.parseFilecsHt|t�r"||up t|�t|�kSt|t�r6|�|�Stt|�|kSdSr�)r}r&�varsr�r%�superr7�r�ryrz�__eq__�s



zParserElement.__eq__cCs
||kSr�ryr7ryryrz�__ne__�szParserElement.__ne__cCstt|��Sr�)�hash�idr�ryryrz�__hash__�szParserElement.__hash__cCs||kSr�ryr7ryryrz�__req__�szParserElement.__req__cCs
||kSr�ryr7ryryrz�__rne__�szParserElement.__rne__cCs2z|jt|�|d�WdSty,YdS0dS)a�
        Method for quick testing of a parser against a test string. Good for simple 
        inline microtests of sub expressions while building up larger parser.
           
        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
            
        Example::
            expr = Word(nums)
            assert expr.matches("100")
        �rTFN)rr�r)r��
testStringrryryrzr%�s

zParserElement.matches�#cCs�t|t�r"tttj|������}t|t�r4t|�}g}g}d}	|D�]�}
|dur^|�	|
d�sf|rr|
sr|�
|
�qD|
sxqDd�|�|
g}g}z:|
�dd�}
|j
|
|d�}|�
|j|d��|	o�|}	Wn�t�yt}
z�t|
t�r�dnd	}d|
v�r(|�
t|
j|
��|�
d
t|
j|
�dd|�n|�
d
|
jd|�|�
d
t|
��|	�oZ|}	|
}WYd}
~
nNd}
~
0t�y�}z,|�
dt|��|	�o�|}	|}WYd}~n
d}~00|�r�|�r�|�
d	�td�|��|�
|
|f�qD|	|fS)a3
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.
           
        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests           
         - comment - (default=C{'#'}) - expression for indicating embedded comments in the test 
              string; pass None to disable comment filtering
         - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default=C{True}) prints test output to stdout
         - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if C{failureTests} is True), and the results contain a list of lines of each 
        test's output
        
        Example::
            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")
        prints::
            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success
            
            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\n of strings that spans \n 3 lines")
        
        (Note that this is a raw string literal, you must include the leading 'r'.)
        TNFrH�\nrm)rdz(FATAL)r�� r��^zFAIL: zFAIL-EXCEPTION: )r}r�r�r�rr��rstrip�
splitlinesrr%r$r�r�rrarr#rIr�r;r�rt)r��testsr�comment�fullDump�printResults�failureTests�
allResults�comments�successrxr?�resultr�r�r|ryryrz�runTests�sNW




$


zParserElement.runTests)F)F)T)T)TT)TT)r)F)N)T)F)T)TroTTF)Or�r�r�r�r�r�staticmethodr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rr	rrrrrr�rrrr�_MAX_INTr(r�r,r�r,r8r9r:rCrDrErFrGrHrIrJrKrLrMrOrPrQrRrVrXr�r�rrYr[rbrfrgrjrkrlr%r~�
__classcell__ryryrerzr&Os�




&




G

"
2G+D
			

)

r&cs eZdZdZ�fdd�Z�ZS)r.zT
    Abstract C{ParserElement} subclass, for defining atomic matching patterns.
    cstt|�jdd�dS�NF�r�)rdr.r�r�reryrzr�@	szToken.__init__�r�r�r�r�r�r�ryryrerzr.<	sr.cs eZdZdZ�fdd�Z�ZS)rz,
    An empty token, will always match.
    cs$tt|���d|_d|_d|_dS)NrTF)rdrr�r�r�r�r�reryrzr�H	szEmpty.__init__r�ryryrerzrD	srcs*eZdZdZ�fdd�Zddd�Z�ZS)rz(
    A token that will never match.
    cs*tt|���d|_d|_d|_d|_dS)NrTFzUnmatchable token)rdrr�r�r�r�r�r�reryrzr�S	s
zNoMatch.__init__TcCst|||j|��dSr�)r!r�r�ryryrzr�Z	szNoMatch.parseImpl)T�r�r�r�r�r�r�r�ryryrerzrO	srcs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Token to exactly match a specified string.
    
    Example::
        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"
    
    For case-insensitive matching, use L{CaselessLiteral}.
    
    For keyword matching (force word break before and after the matched string),
    use L{Keyword} or L{CaselessKeyword}.
    cs�tt|���||_t|�|_z|d|_Wn(tyTtj	dt
dd�t|_Yn0dt
|j�|_d|j|_d|_d|_dS)Nrz2null string passed to Literal; use Empty() insteadrsr2�"%s"r�F)rdrr��matchr��matchLen�firstMatchCharr�r4r5r6rr�r�r�r�r�r��r��matchStringreryrzr�l	s
�zLiteral.__init__TcCsJ|||jkr6|jdks&|�|j|�r6||j|jfSt|||j|��dSr�)r�r��
startswithr�r!r�r�ryryrzr�	s��zLiteral.parseImpl)Tr�ryryrerzr^	s
rcsLeZdZdZedZd�fdd�	Zddd	�Z�fd
d�Ze	dd
��Z
�ZS)ra\
    Token to exactly match a specified string as a keyword, that is, it must be
    immediately followed by a non-keyword character.  Compare with C{L{Literal}}:
     - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
     - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
    Accepts two optional constructor arguments in addition to the keyword string:
     - C{identChars} is a string of characters that would be valid identifier characters,
          defaulting to all alphanumerics + "_" and "$"
     - C{caseless} allows case-insensitive matching, default is C{False}.
       
    Example::
        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use L{CaselessKeyword}.
    �_$NFcs�tt|���|durtj}||_t|�|_z|d|_Wn"ty\t	j
dtdd�Yn0d|j|_d|j|_
d|_d|_||_|r�|��|_|��}t|�|_dS)Nrz2null string passed to Keyword; use Empty() insteadrsr2r�r�F)rdrr��DEFAULT_KEYWORD_CHARSr�r�r�r�r�r4r5r6r�r�r�r��caseless�upper�
caselessmatchr��
identChars)r�r�r�r�reryrzr��	s(
�
zKeyword.__init__TcCs|jr|||||j���|jkr�|t|�|jksL|||j��|jvr�|dksj||d��|jvr�||j|jfSnv|||jkr�|jdks�|�|j|�r�|t|�|jks�|||j|jvr�|dks�||d|jvr�||j|jfSt	|||j
|��dSr�)r�r�r�r�r�r�r�r�r�r!r�r�ryryrzr��	s4����������zKeyword.parseImplcstt|���}tj|_|Sr�)rdrr�r�r�)r�r�reryrzr��	szKeyword.copycCs
|t_dS)z,Overrides the default Keyword chars
        N)rr�r�ryryrz�setDefaultKeywordChars�	szKeyword.setDefaultKeywordChars)NF)T)r�r�r�r�r5r�r�r�r�rr�r�ryryrerzr�	s
rcs*eZdZdZ�fdd�Zddd�Z�ZS)r
al
    Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::
        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
        
    (Contrast with example for L{CaselessKeyword}.)
    cs6tt|��|���||_d|j|_d|j|_dS)Nz'%s'r�)rdr
r�r��returnStringr�r�r�reryrzr��	szCaselessLiteral.__init__TcCs@||||j���|jkr,||j|jfSt|||j|��dSr�)r�r�r�r�r!r�r�ryryrzr��	szCaselessLiteral.parseImpl)Tr�ryryrerzr
�	s
r
cs,eZdZdZd�fdd�	Zd	dd�Z�ZS)
r	z�
    Caseless version of L{Keyword}.

    Example::
        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
        
    (Contrast with example for L{CaselessLiteral}.)
    Ncstt|�j||dd�dS)NT�r�)rdr	r�)r�r�r�reryrzr��	szCaselessKeyword.__init__TcCsj||||j���|jkrV|t|�|jksF|||j��|jvrV||j|jfSt|||j|��dSr�)r�r�r�r�r�r�r!r�r�ryryrzr��	s��zCaselessKeyword.parseImpl)N)Tr�ryryrerzr	�	sr	cs,eZdZdZd�fdd�	Zd	dd�Z�ZS)
rnax
    A variation on L{Literal} which matches "close" matches, that is, 
    strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
     - C{match_string} - string to be matched
     - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
    
    The results from a successful parse will contain the matched text from the input string and the following named results:
     - C{mismatches} - a list of the positions within the match_string where mismatches were found
     - C{original} - the original match_string used to compare against the input string
    
    If C{mismatches} is an empty list, then the match was an exact match.
    
    Example::
        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    r�csBtt|���||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F)	rdrnr�r��match_string�
maxMismatchesr�r�r�)r�r�r�reryrzr�

szCloseMatch.__init__TcCs�|}t|�}|t|j�}||kr�|j}d}g}	|j}
tt|||�|j��D]2\}}|\}}
||
krN|	�|�t|	�|
krNq�qN|d}t|||�g�}|j|d<|	|d<||fSt|||j|��dS)Nrr��original�
mismatches)	r�r�r�r�r�r$r$r!r�)r�rur�r��startr��maxlocr��match_stringlocr�r��s_m�src�mat�resultsryryrzr�
s( 

zCloseMatch.parseImpl)r�)Tr�ryryrerzrn�	s	rncs8eZdZdZd
�fdd�	Zdd	d
�Z�fdd�Z�ZS)r1a	
    Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters,
    an optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction. An optional
    C{excludeChars} parameter can list characters that might be found in 
    the input C{bodyChars} string; useful to define a word of all printables
    except for one or two characters, for instance.
    
    L{srange} is useful for defining custom character set strings for defining 
    C{Word} expressions, using range notation from regular expression character sets.
    
    A common mistake is to use C{Word} to match a specific literal string, as in 
    C{Word("Address")}. Remember that C{Word} uses the string argument to define
    I{sets} of matchable characters. This expression would match "Add", "AAA",
    "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
    To match an exact literal string, use L{Literal} or L{Keyword}.

    pyparsing includes helper strings for building Words:
     - L{alphas}
     - L{nums}
     - L{alphanums}
     - L{hexnums}
     - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
     - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - L{printables} (any non-whitespace character)

    Example::
        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
        
        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums+'-')
        
        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")
        
        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    Nr�rFcs�tt|����rFd��fdd�|D��}|rFd��fdd�|D��}||_t|�|_|rl||_t|�|_n||_t|�|_|dk|_	|dkr�t
d��||_|dkr�||_nt
|_|dkr�||_||_t|�|_d|j|_d	|_||_d
|j|jv�r�|dk�r�|dk�r�|dk�r�|j|jk�r8dt|j�|_nHt|j�dk�rfdt�|j�t|j�f|_nd
t|j�t|j�f|_|j�r�d|jd|_zt�|j�|_Wnt�y�d|_Yn0dS)Nr�c3s|]}|�vr|VqdSr�ryr���excludeCharsryrzr�`
r|z Word.__init__.<locals>.<genexpr>c3s|]}|�vr|VqdSr�ryr�r�ryrzr�b
r|rr�zZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedr�Frqz[%s]+z%s[%s]*z	[%s][%s]*z\b)rdr1r�r��
initCharsOrigr��	initChars�
bodyCharsOrig�	bodyChars�maxSpecifiedr@�minLen�maxLenr�r�r�r�r��	asKeyword�_escapeRegexRangeChars�reStringr�r��escape�compiler�)r�r�r��min�max�exactr�r�rer�rzr�]
s\



0
����z
Word.__init__Tc
Cs>|jr<|j�||�}|s(t|||j|��|��}||��fS|||jvrZt|||j|��|}|d7}t|�}|j}||j	}t
||�}||kr�|||vr�|d7}q�d}	|||jkr�d}	|jr�||kr�|||vr�d}	|j
�r|dkr�||d|v�s||k�r|||v�rd}	|	�r.t|||j|��||||�fS)Nr�FTr)r�r�r!r��end�groupr�r�r�r�r�r�r�r�)
r�rur�r�r}r�r��	bodycharsr��throwExceptionryryrzr��
s6


2zWord.parseImplcstztt|���WSty"Yn0|jdurndd�}|j|jkr^d||j�||j�f|_nd||j�|_|jS)NcSs$t|�dkr|dd�dS|SdS)N��...r��r�ryryrz�
charsAsStr�
sz Word.__str__.<locals>.charsAsStrz	W:(%s,%s)zW:(%s))rdr1r�r�r�r�r�)r�r�reryrzr��
s
zWord.__str__)Nr�rrFN)T�r�r�r�r�r�r�r�r�ryryrerzr1.
s.6
#r1csFeZdZdZee�d��Zd�fdd�	Zddd�Z	�fd	d
�Z
�ZS)
r)a�
    Token for matching strings that match a given regular expression.
    Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
    If the given regex contains named groups (defined using C{(?P<name>...)}), these will be preserved as 
    named parse results.

    Example::
        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    z[A-Z]rcs�tt|���t|t�r�|s,tjdtdd�||_||_	zt
�|j|j	�|_
|j|_Wq�t
jy�tjd|tdd��Yq�0n2t|tj�r�||_
t|�|_|_||_	ntd��t|�|_d|j|_d|_d|_d	S)
z�The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags.z0null string passed to Regex; use Empty() insteadrsr2�$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectr�FTN)rdr)r�r}r�r4r5r6�pattern�flagsr�r�r��
sre_constants�error�compiledREtyperr@r�r�r�r�r�)r�r�r�reryrzr��
s6
�
�
�
zRegex.__init__TcCs`|j�||�}|s"t|||j|��|��}|��}t|���}|rX|D]}||||<qF||fSr�)r�r�r!r�r��	groupdictr$r�)r�rur�r�r}�dr�r�ryryrzr��
szRegex.parseImplcsDztt|���WSty"Yn0|jdur>dt|j�|_|jS)NzRe:(%s))rdr)r�r�r�r�r�r�reryrzr�
s
z
Regex.__str__)r)T)r�r�r�r�r�r�r�r�r�r�r�r�ryryrerzr)�
s
"

r)cs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
r'a�
    Token for matching strings that are delimited by quoting characters.
    
    Defined with the following parameters:
        - quoteChar - string of one or more characters defining the quote delimiting string
        - escChar - character to escape quotes, typically backslash (default=C{None})
        - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
        - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
        - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
        - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})

    Example::
        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
    prints::
        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    NFTc
sLtt����|��}|s0tjdtdd�t��|dur>|}n"|��}|s`tjdtdd�t��|�_t	|��_
|d�_|�_t	|��_
|�_|�_|�_|�_|r�tjtjB�_dt��j�t�jd�|dur�t|�p�df�_n<d�_dt��j�t�jd�|du�rt|��pdf�_t	�j�d	k�rp�jd
d��fdd
�tt	�j�d	dd�D��d7_|�r��jdt�|�7_|�r��jdt�|�7_t��j�d�_�jdt��j�7_zt��j�j��_�j�_Wn.tj�y$tjd�jtdd��Yn0t ���_!d�j!�_"d�_#d�_$dS)Nz$quoteChar cannot be the empty stringrsr2z'endQuoteChar cannot be the empty stringrz%s(?:[^%s%s]r�z%s(?:[^%s\n\r%s]r�z|(?:z)|(?:c3s4|],}dt��jd|��t�j|�fVqdS)z%s[^%s]N)r�r��endQuoteCharr�r;r�ryrzr�Xs��z(QuotedString.__init__.<locals>.<genexpr>rt�)z|(?:%s)z|(?:%s.)z(.)z)*%sr�r�FT)%rdr'r�r�r4r5r6�SyntaxError�	quoteCharr��quoteCharLen�firstQuoteCharr��endQuoteCharLen�escChar�escQuote�unquoteResults�convertWhitespaceEscapesr��	MULTILINE�DOTALLr�r�r�r�r�r��escCharReplacePatternr�r�r�r�r�r�r�r�r�)r�r�r�r��	multiliner�r�r�rer�rzr�/sz



��
������
zQuotedString.__init__c	Cs�|||jkr|j�||�pd}|s4t|||j|��|��}|��}|jr�||j|j	�}t
|t�r�d|vr�|jr�ddddd�}|�
�D]\}}|�||�}q�|jr�t�|jd|�}|jr�|�|j|j�}||fS)N�\�	rH��
)�\trpz\fz\rz\g<1>)r�r�r�r!r�r�r�r�r�r�r}r�r�r�r�r�r�r�r�r�)	r�rur�r�r}r��ws_map�wslit�wscharryryrzr�ps* 
�zQuotedString.parseImplcsFztt|���WSty"Yn0|jdur@d|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rdr'r�r�r�r�r�r�reryrzr��s
zQuotedString.__str__)NNFTNT)Tr�ryryrerzr'sA
#r'cs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
ra�
    Token for matching words composed of characters I{not} in a given set (will
    include whitespace in matched characters if not listed in the provided exclusion set - see example).
    Defined with string containing all disallowed characters, and an optional
    minimum, maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction.

    Example::
        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
    prints::
        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    r�rcs�tt|���d|_||_|dkr*td��||_|dkr@||_nt|_|dkrZ||_||_t	|�|_
d|j
|_|jdk|_d|_
dS)NFr�zfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr�)rdrr�r��notCharsr@r�r�r�r�r�r�r�r�)r�r�r�r�r�reryrzr��s 
zCharsNotIn.__init__TcCs�|||jvrt|||j|��|}|d7}|j}t||jt|��}||krb|||vrb|d7}qD|||jkr�t|||j|��||||�fSr�)r�r!r�r�r�r�r�)r�rur�r�r��notchars�maxlenryryrzr��s
�
zCharsNotIn.parseImplcsdztt|���WSty"Yn0|jdur^t|j�dkrRd|jdd�|_nd|j|_|jS)Nr�z
!W:(%s...)z!W:(%s))rdrr�r�r�r�r�r�reryrzr��s
zCharsNotIn.__str__)r�rr)Tr�ryryrerzr�s
rcs<eZdZdZdddddd�Zd�fdd�	Zddd�Z�ZS)r0a�
    Special matching class for matching whitespace.  Normally, whitespace is ignored
    by pyparsing grammars.  This class is included when some whitespace structures
    are significant.  Define with a string containing the whitespace characters to be
    matched; default is C{" \t\r\n"}.  Also takes optional C{min}, C{max}, and C{exact} arguments,
    as defined for the C{L{Word}} class.
    z<SPC>z<TAB>z<LF>z<CR>z<FF>)rqr�rHr�r�� 	
r�rcs�tt����|�_��d��fdd��jD���d�dd��jD���_d�_d�j�_	|�_
|dkrt|�_nt�_|dkr�|�_|�_
dS)Nr�c3s|]}|�jvr|VqdSr�)�
matchWhiter�r�ryrzr��r|z!White.__init__.<locals>.<genexpr>css|]}tj|VqdSr�)r0�	whiteStrsr�ryryrzr��r|Tr�r)
rdr0r�r�rPr�r�r�r�r�r�r�r�)r��wsr�r�r�rer�rzr��s zWhite.__init__TcCs�|||jvrt|||j|��|}|d7}||j}t|t|��}||krb|||jvrb|d7}qB|||jkr�t|||j|��||||�fSr�)r�r!r�r�r�r�r�)r�rur�r�r�r�ryryrzr�	s

zWhite.parseImpl)r�r�rr)T)r�r�r�r�r�r�r�r�ryryrerzr0�s�r0cseZdZ�fdd�Z�ZS)�_PositionTokencs(tt|���|jj|_d|_d|_dSr�)rdr�r�r�r�r�r�r�r�reryrzr�s
z_PositionToken.__init__�r�r�r�r�r�ryryrerzr�sr�cs2eZdZdZ�fdd�Zdd�Zd	dd�Z�ZS)
rzb
    Token to advance to a specific column of input text; useful for tabular report scraping.
    cstt|���||_dSr�)rdrr�r;)r��colnoreryrzr�$szGoToColumn.__init__cCs\t||�|jkrXt|�}|jr*|�||�}||krX||��rXt||�|jkrX|d7}q*|Sr�)r;r�r�r��isspace)r�rur�r�ryryrzr�(s$
zGoToColumn.preParseTcCsDt||�}||jkr"t||d|��||j|}|||�}||fS)NzText not in expected column�r;r!)r�rur�r��thiscol�newlocr�ryryrzr�1s

zGoToColumn.parseImpl)T)r�r�r�r�r�r�r�r�ryryrerzr s	rcs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Matches if current position is at the beginning of a line within the parse string
    
    Example::
    
        test = '''        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)
    
    Prints::
        ['AAA', ' this line']
        ['AAA', ' and this line']    

    cstt|���d|_dS)NzExpected start of line)rdrr�r�r�reryrzr�OszLineStart.__init__TcCs*t||�dkr|gfSt|||j|��dSr�)r;r!r�r�ryryrzr�SszLineStart.parseImpl)Tr�ryryrerzr:srcs*eZdZdZ�fdd�Zddd�Z�ZS)rzU
    Matches if current position is at the end of a line within the parse string
    cs,tt|���|�tj�dd��d|_dS)NrHr�zExpected end of line)rdrr�rPr&r�r�r�r�reryrzr�\szLineEnd.__init__TcCsb|t|�kr6||dkr$|ddfSt|||j|��n(|t|�krN|dgfSt|||j|��dS)NrHr��r�r!r�r�ryryrzr�aszLineEnd.parseImpl)Tr�ryryrerzrXsrcs*eZdZdZ�fdd�Zddd�Z�ZS)r,zM
    Matches if current position is at the beginning of the parse string
    cstt|���d|_dS)NzExpected start of text)rdr,r�r�r�reryrzr�pszStringStart.__init__TcCs0|dkr(||�|d�kr(t|||j|��|gfSr�)r�r!r�r�ryryrzr�tszStringStart.parseImpl)Tr�ryryrerzr,lsr,cs*eZdZdZ�fdd�Zddd�Z�ZS)r+zG
    Matches if current position is at the end of the parse string
    cstt|���d|_dS)NzExpected end of text)rdr+r�r�r�reryrzr�szStringEnd.__init__TcCs^|t|�krt|||j|��n<|t|�kr6|dgfS|t|�krJ|gfSt|||j|��dSr�r�r�ryryrzr��szStringEnd.parseImpl)Tr�ryryrerzr+{sr+cs.eZdZdZef�fdd�	Zddd�Z�ZS)r3ap
    Matches if the current position is at the beginning of a Word, and
    is not preceded by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
    the string being parsed, or at the beginning of a line.
    cs"tt|���t|�|_d|_dS)NzNot at the start of a word)rdr3r�r��	wordCharsr��r�r�reryrzr��s
zWordStart.__init__TcCs@|dkr8||d|jvs(|||jvr8t|||j|��|gfSr�)r�r!r�r�ryryrzr��s�zWordStart.parseImpl)T�r�r�r�r�rXr�r�r�ryryrerzr3�sr3cs.eZdZdZef�fdd�	Zddd�Z�ZS)r2aZ
    Matches if the current position is at the end of a Word, and
    is not followed by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
    the string being parsed, or at the end of a line.
    cs(tt|���t|�|_d|_d|_dS)NFzNot at the end of a word)rdr2r�r�r�r�r�r�reryrzr��s
zWordEnd.__init__TcCsPt|�}|dkrH||krH|||jvs8||d|jvrHt|||j|��|gfSr�)r�r�r!r�)r�rur�r�r�ryryrzr��s�zWordEnd.parseImpl)Tr�ryryrerzr2�sr2cs�eZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Z�fdd�Z�fd
d�Z	�fdd�Z
d�fdd�	Zgfdd�Z�fdd�Z
�ZS)r"z^
    Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
    Fcs�tt|��|�t|t�r"t|�}t|t�r<t�|�g|_	nft|t
�rxt|�}tdd�|D��rlttj|�}t|�|_	n*zt|�|_	Wnt
y�|g|_	Yn0d|_dS)Ncss|]}t|t�VqdSr�)r}r�)r�rvryryrzr��r|z+ParseExpression.__init__.<locals>.<genexpr>F)rdr"r�r}r�r�r�r&r��exprsr�allr�r�r��r�r�r�reryrzr��s


zParseExpression.__init__cCs
|j|Sr�)r�r�ryryrzr��szParseExpression.__getitem__cCs|j�|�d|_|Sr�)r�r$r�r7ryryrzr$�szParseExpression.appendcCs0d|_dd�|jD�|_|jD]}|��q|S)z~Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
           all contained expressions.FcSsg|]}|���qSryr*�r�r�ryryrzr��r|z3ParseExpression.leaveWhitespace.<locals>.<listcomp>)r�r�rO)r�r�ryryrzrO�s


zParseExpression.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jD]}|�|jd�q*n,tt|��|�|jD]}|�|jd�qX|Sr)r}r-r�rdr"rRr�)r�r+r�reryrzrR�s



zParseExpression.ignorecsLztt|���WSty"Yn0|jdurFd|jjt|j�f|_|jS�Nz%s:(%s))	rdr"r�r�r�r�r�r�r�r�reryrzr��s
zParseExpression.__str__cs*tt|���|jD]}|��qt|j�dk�r|jd}t||j�r�|js�|jdur�|j	s�|jdd�|jdg|_d|_
|j|jO_|j|jO_|jd}t||j��r|j�s|jdu�r|j	�s|jdd�|jdd�|_d|_
|j|jO_|j|jO_dt
|�|_|S)Nrsrr�rtr�)rdr"rr�r�r}r�r�r�r�r�r�r�r�r�)r�r�r+reryrzr�s<


���
���zParseExpression.streamlinecstt|��||�}|Sr�)rdr"r�)r�r�r�r�reryrzr�
szParseExpression.setResultsNamecCs6|dd�|g}|jD]}|�|�q|�g�dSr�)r�r[rY)r�rZ�tmpr�ryryrzr[
s
zParseExpression.validatecs$tt|���}dd�|jD�|_|S)NcSsg|]}|���qSryr*r�ryryrzr�%
r|z(ParseExpression.copy.<locals>.<listcomp>)rdr"r�r�rGreryrzr�#
szParseExpression.copy)F)F)r�r�r�r�r�r�r$rOrRr�rr�r[r�r�ryryrerzr"�s	
"r"csTeZdZdZGdd�de�Zd�fdd�	Zddd�Zd	d
�Zdd�Z	d
d�Z
�ZS)ra

    Requires all given C{ParseExpression}s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'+'} operator.
    May also be constructed using the C{'-'} operator, which will suppress backtracking.

    Example::
        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"),name_expr("name"),integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    cseZdZ�fdd�Z�ZS)zAnd._ErrorStopcs*ttj|�j|i|��d|_|��dS)N�-)rdrr8r�r�rOrgreryrzr�9
szAnd._ErrorStop.__init__r�ryryrerzr88
sr8TcsRtt|��||�tdd�|jD��|_|�|jdj�|jdj|_d|_	dS)Ncss|]}|jVqdSr��r�r�ryryrzr�@
r|zAnd.__init__.<locals>.<genexpr>rT)
rdrr�r�r�r�rPr�r�r�r�reryrzr�>
s
zAnd.__init__c	Cs�|jdj|||dd�\}}d}|jdd�D]�}t|tj�rDd}q.|r�z|�|||�\}}Wq�tyr�Yq�ty�}zd|_t�|��WYd}~q�d}~0t	y�t|t
|�|j|��Yq�0n|�|||�\}}|s�|��r.||7}q.||fS)NrFr r�T)
r�r�r}rr8r%r�
__traceback__r�r�r�r�r)	r�rur�r��
resultlist�	errorStopr��
exprtokensr�ryryrzr�E
s(
z
And.parseImplcCst|t�rt�|�}|�|�Sr��r}r�r&r�r$r7ryryrzr6^
s

zAnd.__iadd__cCs6|dd�|g}|jD]}|�|�|jsq2qdSr�)r�rYr��r�r��subRecCheckListr�ryryrzrYc
s


zAnd.checkRecursioncCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr��{rqcss|]}t|�VqdSr�r�r�ryryrzr�o
r|zAnd.__str__.<locals>.<genexpr>�}�rr�r�r�r�r�ryryrzr�j
s


 zAnd.__str__)T)T)r�r�r�r�rr8r�r�r6rYr�r�ryryrerzr(
s
rcsDeZdZdZd�fdd�	Zddd�Zdd	�Zd
d�Zdd
�Z�Z	S)ra�
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the expression that matches the longest string will be used.
    May be constructed using the C{'^'} operator.

    Example::
        # construct Or using '^' operator
        
        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))
    prints::
        [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr��
r|zOr.__init__.<locals>.<genexpr>T)rdrr�r�rbr�r�reryrzr��
szOr.__init__TcCsTd}d}g}|jD]�}z|�||�}Wnvtyd}	z&d|	_|	j|krP|	}|	j}WYd}	~	qd}	~	0ty�t|�|kr�t|t|�|j|�}t|�}Yq0|�||f�q|�r*|j	dd�d�|D]`\}
}z|�
|||�WSt�y&}	z(d|	_|	j|k�r|	}|	j}WYd}	~	q�d}	~	00q�|du�rB|j|_|�nt||d|��dS)NrtcSs
|dSr�ry)�xryryrzr{�
r|zOr.parseImpl.<locals>.<lambda>)r � no defined alternatives to match)r�r�r!r�r�r�r�r�r$�sortr�r�)r�rur�r��	maxExcLoc�maxExceptionr%r��loc2r��_ryryrzr��
s<


zOr.parseImplcCst|t�rt�|�}|�|�Sr�r�r7ryryrz�__ixor__�
s

zOr.__ixor__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�z ^ css|]}t|�VqdSr�r�r�ryryrzr��
r|zOr.__str__.<locals>.<genexpr>r�r�r�ryryrzr��
s


 z
Or.__str__cCs,|dd�|g}|jD]}|�|�qdSr��r�rYr�ryryrzrY�
s
zOr.checkRecursion)F)T)
r�r�r�r�r�r�rr�rYr�ryryrerzrt
s

&	rcsDeZdZdZd�fdd�	Zddd�Zdd	�Zd
d�Zdd
�Z�Z	S)ra�
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the first one listed is the one that will match.
    May be constructed using the C{'|'} operator.

    Example::
        # construct MatchFirst using '|' operator
        
        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr��
r|z&MatchFirst.__init__.<locals>.<genexpr>T)rdrr�r�rbr�r�reryrzr��
szMatchFirst.__init__Tc	Cs�d}d}|jD]�}z|�|||�}|WStyb}z |j|krN|}|j}WYd}~qd}~0ty�t|�|kr�t|t|�|j|�}t|�}Yq0q|dur�|j|_|�nt||d|��dS)Nrtr�)r�r�r!r�r�r�r�r�)	r�rur�r�rrr�r�r�ryryrzr��
s$


zMatchFirst.parseImplcCst|t�rt�|�}|�|�Sr�r�r7ryryrz�__ior__�
s

zMatchFirst.__ior__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�� | css|]}t|�VqdSr�r�r�ryryrzr��
r|z%MatchFirst.__str__.<locals>.<genexpr>r�r�r�ryryrzr��
s


 zMatchFirst.__str__cCs,|dd�|g}|jD]}|�|�qdSr�rr�ryryrzrYs
zMatchFirst.checkRecursion)F)T)
r�r�r�r�r�r�rr�rYr�ryryrerzr�
s
	rcs<eZdZdZd�fdd�	Zddd�Zdd�Zd	d
�Z�ZS)
ram
    Requires all given C{ParseExpression}s to be found, but in any order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'&'} operator.

    Example::
        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order 
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )
    prints::
        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    Tcs8tt|��||�tdd�|jD��|_d|_d|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr�?r|z Each.__init__.<locals>.<genexpr>T)rdrr�r�r�r�r��initExprGroupsr�reryrzr�=sz
Each.__init__c	s�|jr�tdd�|jD��|_dd�|jD�}dd�|jD�}|||_dd�|jD�|_dd�|jD�|_dd�|jD�|_|j|j7_d	|_|}|jdd�}|jdd��g}d
}	|	�rh|�|j|j}
g}|
D]t}z|�||�}Wnt	�y|�
|�Yq�0|�
|j�t|�|��||v�r>|�
|�q�|�vr܈�
|�q�t|�t|
�kr�d	}	q�|�r�d�dd�|D��}
t	||d
|
��|�fdd�|jD�7}g}|D]"}|�|||�\}}|�
|��q�t|tg��}||fS)Ncss&|]}t|t�rt|j�|fVqdSr�)r}rrirvr�ryryrzr�Er|z!Each.parseImpl.<locals>.<genexpr>cSsg|]}t|t�r|j�qSry�r}rrvr�ryryrzr�Fr|z"Each.parseImpl.<locals>.<listcomp>cSs g|]}|jrt|t�s|�qSry)r�r}rr�ryryrzr�Gr|cSsg|]}t|t�r|j�qSry)r}r4rvr�ryryrzr�Ir|cSsg|]}t|t�r|j�qSry)r}rrvr�ryryrzr�Jr|cSs g|]}t|tttf�s|�qSry)r}rr4rr�ryryrzr�Kr|FTr:css|]}t|�VqdSr�r�r�ryryrzr�fr|z*Missing one or more required elements (%s)cs$g|]}t|t�r|j�vr|�qSryr
r���tmpOptryrzr�jr|)r	r�r��opt1map�	optionals�multioptionals�
multirequired�requiredr�r!r$r�ri�remover�r�r��sumr$)r�rur�r��opt1�opt2�tmpLoc�tmpReqd�
matchOrder�keepMatching�tmpExprs�failedr��missingr�r��finalResultsryrrzr�CsP

zEach.parseImplcCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�z & css|]}t|�VqdSr�r�r�ryryrzr�yr|zEach.__str__.<locals>.<genexpr>r�r�r�ryryrzr�ts


 zEach.__str__cCs,|dd�|g}|jD]}|�|�qdSr�rr�ryryrzrY}s
zEach.checkRecursion)T)T)	r�r�r�r�r�r�r�rYr�ryryrerzrs
5
1	rcsleZdZdZd�fdd�	Zddd�Zdd	�Z�fd
d�Z�fdd
�Zdd�Z	gfdd�Z
�fdd�Z�ZS)r za
    Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
    Fcs�tt|��|�t|t�r@ttjt�r2t�|�}nt�t	|��}||_
d|_|dur�|j|_|j
|_
|�|j�|j|_|j|_|j|_|j�|j�dSr�)rdr r�r}r��
issubclassr&r�r.rrvr�r�r�rPr�r�r�r�r�r&�r�rvr�reryrzr��s
zParseElementEnhance.__init__TcCs2|jdur|jj|||dd�Std||j|��dS)NFr r�)rvr�r!r�r�ryryrzr��s
zParseElementEnhance.parseImplcCs*d|_|j��|_|jdur&|j��|Sr)r�rvr�rOr�ryryrzrO�s


z#ParseElementEnhance.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jdurn|j�|jd�n,tt|��|�|jdurn|j�|jd�|Sr)r}r-r�rdr rRrvr7reryrzrR�s



zParseElementEnhance.ignorecs&tt|���|jdur"|j��|Sr�)rdr rrvr�reryrzr�s

zParseElementEnhance.streamlinecCsB||vrt||g��|dd�|g}|jdur>|j�|�dSr�)r(rvrY)r�r�r�ryryrzrY�s

z"ParseElementEnhance.checkRecursioncCs6|dd�|g}|jdur(|j�|�|�g�dSr��rvr[rY�r�rZr�ryryrzr[�s
zParseElementEnhance.validatecsVztt|���WSty"Yn0|jdurP|jdurPd|jjt|j�f|_|jSr�)	rdr r�r�r�rvr�r�r�r�reryrzr��szParseElementEnhance.__str__)F)T)
r�r�r�r�r�r�rOrRrrYr[r�r�ryryrerzr �s
r cs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Lookahead matching of the given parse expression.  C{FollowedBy}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression matches at the current
    position.  C{FollowedBy} always returns a null token list.

    Example::
        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
    prints::
        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    cstt|��|�d|_dSr�)rdrr�r��r�rvreryrzr��szFollowedBy.__init__TcCs|j�||�|gfSr�)rvr�r�ryryrzr��szFollowedBy.parseImpl)Tr�ryryrerzr�srcs2eZdZdZ�fdd�Zd	dd�Zdd�Z�ZS)
ra�
    Lookahead to disallow matching with the given parse expression.  C{NotAny}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression does I{not} match at the current
    position.  Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
    always returns a null token list.  May be constructed using the '~' operator.

    Example::
        
    cs0tt|��|�d|_d|_dt|j�|_dS)NFTzFound unwanted token, )rdrr�r�r�r�rvr�r"reryrzr��szNotAny.__init__TcCs&|j�||�rt|||j|��|gfSr�)rvr�r!r�r�ryryrzr��szNotAny.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�z~{r��rr�r�r�rvr�ryryrzr�s


zNotAny.__str__)Tr�ryryrerzr�s

rcs(eZdZd�fdd�	Zddd�Z�ZS)	�_MultipleMatchNcsFtt|��|�d|_|}t|t�r.t�|�}|dur<|nd|_dSr�)	rdr$r�r�r}r�r&r��	not_ender)r�rv�stopOn�enderreryrzr�s

z_MultipleMatch.__init__Tc	Cs�|jj}|j}|jdu}|r$|jj}|r2|||�||||dd�\}}zV|j}	|r`|||�|	rp|||�}
n|}
|||
|�\}}|s�|��rR||7}qRWnttfy�Yn0||fS�NFr )	rvr�r�r%r�r�rr!r�)r�rur�r��self_expr_parse�self_skip_ignorables�check_ender�
try_not_enderr��hasIgnoreExprsr��	tmptokensryryrzr�s*



z_MultipleMatch.parseImpl)N)T)r�r�r�r�r�r�ryryrerzr$
sr$c@seZdZdZdd�ZdS)ra�
    Repetition of one or more of the given expression.
    
    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
        
        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    cCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r�z}...r#r�ryryrzr�Js


zOneOrMore.__str__N)r�r�r�r�r�ryryryrzr0srcs8eZdZdZd
�fdd�	Zd�fdd�	Zdd	�Z�ZS)r4aw
    Optional repetition of zero or more of the given expression.
    
    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example: similar to L{OneOrMore}
    Ncstt|�j||d�d|_dS)N)r&T)rdr4r�r�)r�rvr&reryrzr�_szZeroOrMore.__init__Tc	s:ztt|��|||�WSttfy4|gfYS0dSr�)rdr4r�r!r�r�reryrzr�cszZeroOrMore.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r9�]...r#r�ryryrzr�is


zZeroOrMore.__str__)N)Tr�ryryrerzr4Ssr4c@s eZdZdd�ZeZdd�ZdS)�
_NullTokencCsdSrryr�ryryrzrssz_NullToken.__bool__cCsdSr�ryr�ryryrzr�vsz_NullToken.__str__N)r�r�r�rrmr�ryryryrzr0rsr0cs6eZdZdZef�fdd�	Zd	dd�Zdd�Z�ZS)
raa
    Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::
        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345
            
            # ZIP+4 form
            12101-0001
            
            # invalid ZIP
            98765-
            ''')
    prints::
        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    cs.tt|�j|dd�|jj|_||_d|_dS)NFr�T)rdrr�rvr�r!r�)r�rvrreryrzr��s
zOptional.__init__Tc	Csxz|jj|||dd�\}}WnRttfyn|jturf|jjr\t|jg�}|j||jj<qj|jg}ng}Yn0||fSr()rvr�r!r�r!�_optionalNotMatchedr�r$)r�rur�r�r�ryryrzr��s


zOptional.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r9r<r#r�ryryrzr��s


zOptional.__str__)T)	r�r�r�r�r1r�r�r�r�ryryrerzrzs"
rcs,eZdZdZd	�fdd�	Zd
dd�Z�ZS)r*a�	
    Token for skipping over all undefined text until the matched expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default=C{False}) if True, the target expression is also parsed 
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default=C{None}) used to define grammars (typically quoted strings and 
          comments) that might contain false matches to the target expression
     - failOn - (default=C{None}) define expressions that are not allowed to be 
          included in the skipped test; if found before the target expression is found, 
          the SkipTo is not a match

    Example::
        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP 
                      + string_data("sev") + SEP 
                      + string_data("desc") + SEP 
                      + integer("days_open"))
        
        for tkt in ticket_expr.searchString(report):
            print tkt.dump()
    prints::
        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    FNcs`tt|��|�||_d|_d|_||_d|_t|t	�rFt
�|�|_n||_dt
|j�|_dS)NTFzNo match found for )rdr*r��
ignoreExprr�r��includeMatchr�r}r�r&r��failOnr�rvr�)r�r+�includerRr4reryrzr��s
zSkipTo.__init__Tc	Cs"|}t|�}|j}|jj}|jdur,|jjnd}|jdurB|jjnd}	|}
|
|kr�|durf|||
�rfq�|	dur�z|	||
�}
Wqnty�Yq�Yqn0qnz|||
ddd�Wq�tt	fy�|
d7}
YqJ0q�qJt|||j
|��|
}|||�}t|�}|j�r||||dd�\}}
||
7}||fS)NF)r�r�r�r )
r�rvr�r4r�r2r�rr!r�r�r$r3)r�rur�r�rxr�rv�
expr_parse�self_failOn_canParseNext�self_ignoreExpr_tryParse�tmploc�skiptext�
skipresultr�ryryrzr��s:
zSkipTo.parseImpl)FNN)Tr�ryryrerzr*�s6
r*csbeZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Zdd�Zgfd
d�Z	dd�Z
�fdd�Z�ZS)raK
    Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.

    Note: take care when assigning to C{Forward} not to overlook precedence of operators.
    Specifically, '|' has a lower precedence than '<<', so that::
        fwdExpr << a | b | c
    will actually be evaluated as::
        (fwdExpr << a) | b | c
    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the C{Forward}::
        fwdExpr << (a | b | c)
    Converting to use the '<<=' operator instead will avoid this problem.

    See L{ParseResults.pprint} for an example of a recursive parser created using
    C{Forward}.
    Ncstt|�j|dd�dSr�)rdrr�r7reryrzr�@szForward.__init__cCsjt|t�rt�|�}||_d|_|jj|_|jj|_|�|jj	�|jj
|_
|jj|_|j�
|jj�|Sr�)r}r�r&r�rvr�r�r�rPr�r�r�r�r&r7ryryrz�
__lshift__Cs





zForward.__lshift__cCs||>Sr�ryr7ryryrz�__ilshift__PszForward.__ilshift__cCs
d|_|SrrNr�ryryrzrOSszForward.leaveWhitespacecCs$|js d|_|jdur |j��|Sr�)r�rvrr�ryryrzrWs


zForward.streamlinecCs>||vr0|dd�|g}|jdur0|j�|�|�g�dSr�r r!ryryrzr[^s

zForward.validatecCs^t|d�r|jS|jjdSz&|jdur4t|j�}nd}W|j|_n
|j|_0|jjd|S)Nr�z: ...�Nonez: )rr�r�r�Z_revertClass�_ForwardNoRecurservr�)r��	retStringryryrzr�es

zForward.__str__cs.|jdurtt|���St�}||K}|SdSr�)rvrdrr�rGreryrzr�vs

zForward.copy)N)
r�r�r�r�r�r<r=rOrr[r�r�r�ryryrerzr-s
rc@seZdZdd�ZdS)r?cCsdS)Nr�ryr�ryryrzr�sz_ForwardNoRecurse.__str__N)r�r�r�r�ryryryrzr?~sr?cs"eZdZdZd�fdd�	Z�ZS)r/zQ
    Abstract subclass of C{ParseExpression}, for converting parsed results.
    Fcstt|��|�d|_dSr)rdr/r�r�rreryrzr��szTokenConverter.__init__)Fr�ryryrerzr/�sr/cs6eZdZdZd
�fdd�	Z�fdd�Zdd	�Z�ZS)ra�
    Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the input string;
    this can be disabled by specifying C{'adjacent=False'} in the constructor.

    Example::
        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    r�Tcs8tt|��|�|r|��||_d|_||_d|_dSr�)rdrr�rO�adjacentr��
joinStringr�)r�rvrBrAreryrzr��szCombine.__init__cs(|jrt�||�ntt|��|�|Sr�)rAr&rRrdrr7reryrzrR�szCombine.ignorecCsP|��}|dd�=|td�|�|j��g|jd�7}|jrH|��rH|gS|SdS)Nr�)r�)r�r$r�r=rBr�r�r)r�rur�r��retToksryryrzr��s
"zCombine.postParse)r�T)r�r�r�r�r�rRr�r�ryryrerzr�s
rcs(eZdZdZ�fdd�Zdd�Z�ZS)ra�
    Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.

    Example::
        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a,b,100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a,b,100"))  # -> ['fn', ['a', 'b', '100']]
    cstt|��|�d|_dSr�)rdrr�r�r"reryrzr��szGroup.__init__cCs|gSr�ryr�ryryrzr��szGroup.postParse�r�r�r�r�r�r�r�ryryrerzr�s
rcs(eZdZdZ�fdd�Zdd�Z�ZS)r
aW
    Converter to return a repetitive expression as a list, but also as a dictionary.
    Each element can also be referenced using the first token in the expression as its key.
    Useful for tabular report scraping when the first column can be used as a item key.

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())
        
        # access named fields as dict entries, or output as dict
        print(result['shape'])        
        print(result.asDict())
    prints::
        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
    See more examples at L{ParseResults} of accessing fields by results name.
    cstt|��|�d|_dSr�)rdr
r�r�r"reryrzr��sz
Dict.__init__cCs�t|�D]�\}}t|�dkrq|d}t|t�r@t|d���}t|�dkr\td|�||<qt|�dkr�t|dt�s�t|d|�||<q|��}|d=t|�dks�t|t�r�|�	�r�t||�||<qt|d|�||<q|j
r�|gS|SdS)Nrr�r�rs)r�r�r}rvr�r�r�r$r�rr�)r�rur�r�r��tok�ikey�	dictvalueryryrzr��s$
zDict.postParserDryryrerzr
�s#r
c@s eZdZdZdd�Zdd�ZdS)r-aV
    Converter for ignoring the results of a parsed expression.

    Example::
        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))
    prints::
        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']
    (See also L{delimitedList}.)
    cCsgSr�ryr�ryryrzr�szSuppress.postParsecCs|Sr�ryr�ryryrzrM"szSuppress.suppressN)r�r�r�r�r�rMryryryrzr-sr-c@s(eZdZdZdd�Zdd�Zdd�ZdS)	rzI
    Wrapper for parse actions, to ensure they are only called once.
    cCst|�|_d|_dSr)r��callable�called)r��
methodCallryryrzr�*s
zOnlyOnce.__init__cCs.|js|�|||�}d|_|St||d��dS)NTr�)rIrHr!)r�r�rrxr�ryryrzrL-s
zOnlyOnce.__call__cCs
d|_dSr)rIr�ryryrz�reset3szOnlyOnce.resetN)r�r�r�r�r�rLrKryryryrzr&srcs8t����fdd�}z�j|_Wnty2Yn0|S)at
    Decorator for debugging parse actions. 
    
    When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
    When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.

    Example::
        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
    prints::
        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    c
s��j}|dd�\}}}t|�dkr8|djjd|}tj�d|t||�||f�z�|�}Wn:ty�}z"tj�d||f��WYd}~n
d}~00tj�d||f�|S)Nr�rqr�.z">>entering %s(line: '%s', %d, %r)
z<<leaving %s (exception: %s)
z<<leaving %s (ret: %r)
)r�r�r�r��stderr�writerIr�)�paArgs�thisFuncr�rrxr�r|�raryrz�zLsztraceParseAction.<locals>.z)r�r�r�)rarRryrQrzrd6s
rd�,FcCs`t|�dt|�dt|�d}|rBt|t||���|�S|tt|�|��|�SdS)a�
    Helper to define a delimited list of expressions - the delimiter defaults to ','.
    By default, the list elements and delimiters can have intervening whitespace, and
    comments, but this can be overridden by passing C{combine=True} in the constructor.
    If C{combine} is set to C{True}, the matching tokens are returned as a single token
    string, with the delimiters included; otherwise, the matching tokens are returned
    as a list of tokens, with the delimiters suppressed.

    Example::
        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    z [rqr/N)r�rr4r�r-)rv�delim�combine�dlNameryryrzrBbs
$rBcsjt����fdd�}|dur0tt��dd��}n|��}|�d�|j|dd�|��d	t��d
�S)a:
    Helper to define a counted list of expressions.
    This helper defines a pattern of the form::
        integer expr expr expr...
    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
    
    If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.

    Example::
        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    cs.|d}�|r tt�g|��p&tt�>gSr�)rrrE)r�rrxr<��	arrayExprrvryrz�countFieldParseAction�s"z+countedArray.<locals>.countFieldParseActionNcSst|d�Sr�)rvrwryryrzr{�r|zcountedArray.<locals>.<lambda>�arrayLenT�r�z(len) r�)rr1rTr�r�r�r�r�)rv�intExprrYryrWrzr>us
r>cCs6g}|D](}t|t�r&|�t|��q|�|�q|Sr�)r}r�r&r*r$)�Lr�r�ryryrzr*�s
r*cs6t���fdd�}|j|dd���dt|���S)a*
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches a
    previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
    If this is not desired, use C{matchPreviousExpr}.
    Do I{not} use with packrat parsing enabled.
    csP|rBt|�dkr�|d>qLt|���}�tdd�|D��>n
�t�>dS)Nr�rcss|]}t|�VqdSr�)r�r��ttryryrzr��r|zDmatchPreviousLiteral.<locals>.copyTokenToRepeater.<locals>.<genexpr>)r�r*r�rr)r�rrx�tflat��repryrz�copyTokenToRepeater�sz1matchPreviousLiteral.<locals>.copyTokenToRepeaterTr[�(prev) )rr�r�r�)rvrcryrarzrQ�s


rQcsFt��|��}�|K��fdd�}|j|dd���dt|���S)aS
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches by
    expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
    the expressions are evaluated first, and then compared, so
    C{"1"} is compared with C{"10"}.
    Do I{not} use with packrat parsing enabled.
    cs*t|�����fdd�}�j|dd�dS)Ncs$t|���}|�kr tddd��dS)Nr�r)r*r�r!)r�rrx�theseTokens��matchTokensryrz�mustMatchTheseTokens�szLmatchPreviousExpr.<locals>.copyTokenToRepeater.<locals>.mustMatchTheseTokensTr[)r*r�r�)r�rrxrhrarfrzrc�sz.matchPreviousExpr.<locals>.copyTokenToRepeaterTr[rd)rr�r�r�r�)rv�e2rcryrarzrP�srPcCs:dD]}|�|t|�}q|�dd�}|�dd�}t|�S)Nz\^-]rHrpr�r�)r��_bslashr�)r�r�ryryrzr��s
r�Tc
s�|rdd�}dd�}t�ndd�}dd�}t�g}t|t�rF|��}n$t|t�rZt|�}ntjdt	dd�|stt
�Sd	}|t|�d
k�r||}t||d
d��D]R\}}	||	|�r�|||d
=qxq�|||	�r�|||d
=|�
||	�|	}qxq�|d
7}qx|�s�|�r�zlt|�td�|��k�rTtd
d�dd�|D����d�|��WStd�dd�|D����d�|��WSWn$t�y�tjdt	dd�Yn0t�fdd�|D���d�|��S)a�
    Helper to quickly define a set of alternative Literals, and makes sure to do
    longest-first testing when there is a conflict, regardless of the input order,
    but returns a C{L{MatchFirst}} for best performance.

    Parameters:
     - strs - a string of space-delimited literals, or a collection of string literals
     - caseless - (default=C{False}) - treat all literals as caseless
     - useRegex - (default=C{True}) - as an optimization, will generate a Regex
          object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
          if creating a C{Regex} raises an exception)

    Example::
        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))
    prints::
        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    cSs|��|��kSr�)r��r-�bryryrzr{�r|zoneOf.<locals>.<lambda>cSs|���|���Sr�)r�r�rkryryrzr{�r|cSs||kSr�ryrkryryrzr{�r|cSs
|�|�Sr�)r�rkryryrzr{�r|z6Invalid argument to oneOf, expected string or iterablersr2rr�Nr�z[%s]css|]}t|�VqdSr�)r��r��symryryrzr�r|zoneOf.<locals>.<genexpr>r�|css|]}t�|�VqdSr�)r�r�rmryryrzr�r|z7Exception creating Regex for oneOf, building MatchFirstc3s|]}�|�VqdSr�ryrm��parseElementClassryrzr�$r|)r
rr}r�r�rr�r4r5r6rr�r�r"r�r)r�r�r)
�strsr��useRegex�isequal�masks�symbolsr��currr+ryrprzrU�sP



�


**�rUcCsttt||���S)a�
    Helper to easily and clearly define a dictionary by specifying the respective patterns
    for the key and value.  Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
    in the proper order.  The key pattern can include delimiting markers or punctuation,
    as long as they are suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the C{Dict} results can include named token
    fields.

    Example::
        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())
    prints::
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    )r
r4r)r rryryrzrC&s!rCcCs^t��dd��}|��}d|_|d�||d�}|r@dd�}ndd�}|�|�|j|_|S)	a�
    Helper to return the original, untokenized text for a given expression.  Useful to
    restore the parsed fields of an HTML start tag into the raw tag text itself, or to
    revert separate tokens with intervening whitespace back to the original matching
    input text. By default, returns astring containing the original parsed text.  
       
    If the optional C{asString} argument is passed as C{False}, then the return value is a 
    C{L{ParseResults}} containing any results names that were originally matched, and a 
    single token containing the original matched text from the input string.  So if 
    the expression passed to C{L{originalTextFor}} contains expressions with defined
    results names, you must set C{asString} to C{False} if you want to preserve those
    results name values.

    Example::
        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b","i"):
            opener,closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])
    prints::
        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    cSs|Sr�ry)r�r�rxryryrzr{ar|z!originalTextFor.<locals>.<lambda>F�_original_start�
_original_endcSs||j|j�Sr�)rxryr~ryryrzr{fr|cSs&||�d�|�d��g|dd�<dS)Nrxry)rr~ryryrz�extractTexthsz$originalTextFor.<locals>.extractText)rr�r�r�r�)rv�asString�	locMarker�endlocMarker�	matchExprrzryryrzriIs

ricCst|��dd��S)zp
    Helper to undo pyparsing's default grouping of And expressions, even
    if all but one are non-empty.
    cSs|dSr�ryrwryryrzr{sr|zungroup.<locals>.<lambda>)r/r�)rvryryrzrjnsrjcCs4t��dd��}t|d�|d�|����d��S)a�
    Helper to decorate a returned token with its starting and ending locations in the input string.
    This helper adds the following results names:
     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains C{<TAB>} characters, you may want to call
    C{L{ParserElement.parseWithTabs}}

    Example::
        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)
    prints::
        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    cSs|Sr�ryr~ryryrzr{�r|zlocatedExpr.<locals>.<lambda>�
locn_startr�locn_end)rr�rr�rO)rv�locatorryryrzrlusrlrErKrJrcrbz\[]-*.$+^?()~ �r�cCs|ddSr�ryr~ryryrzr{�r|r{z\\0?[xX][0-9a-fA-F]+cCstt|d�d�d��S)Nrz\0x�)�unichrrv�lstripr~ryryrzr{�r|z	\\0[0-7]+cCstt|ddd�d��S)Nrr��)r�rvr~ryryrzr{�r|z\]r�r9rr�negate�bodyr<csDdd��z"d��fdd�t�|�jD��WSty>YdS0dS)a�
    Helper to easily define string ranges for use in Word construction.  Borrows
    syntax from regexp '[]' string range definitions::
        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
    The input string must be enclosed in []'s, and the returned string is the expanded
    character set joined into a single string.
    The values enclosed in the []'s may be:
     - a single character
     - an escaped character with a leading backslash (such as C{\-} or C{\]})
     - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) 
         (C{\0x##} is also supported for backwards compatibility) 
     - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
     - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
     - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
    cSs<t|t�s|Sd�dd�tt|d�t|d�d�D��S)Nr�css|]}t|�VqdSr�)r�r�ryryrzr��r|z+srange.<locals>.<lambda>.<locals>.<genexpr>rr�)r}r$r�r��ord)�pryryrzr{�r|zsrange.<locals>.<lambda>r�c3s|]}�|�VqdSr�ry)r��part��	_expandedryrzr��r|zsrange.<locals>.<genexpr>N)r��_reBracketExprrr�r�r�ryr�rzra�s
"racs�fdd�}|S)zt
    Helper method for defining parse actions that require matching at a specific
    column in the input text.
    cs"t||��krt||d���dS)Nzmatched token not at column %dr�)ro�locnrzr;ryrz�	verifyCol�sz!matchOnlyAtCol.<locals>.verifyColry)r<r�ryr;rzrO�srOcs�fdd�S)a�
    Helper method for common parse actions that simply return a literal value.  Especially
    useful when used with C{L{transformString<ParserElement.transformString>}()}.

    Example::
        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num
        
        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    cs�gSr�ryr~��replStrryrzr{�r|zreplaceWith.<locals>.<lambda>ryr�ryr�rzr^�sr^cCs|ddd�S)a
    Helper parse action for removing quotation marks from parsed quoted strings.

    Example::
        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    rr�rtryr~ryryrzr\�sr\csL��fdd�}zt�dt�d�j�}Wnty@t��}Yn0||_|S)aG
    Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional 
    args are passed, they are forwarded to the given function as additional arguments after
    the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
    parsed data to an integer using base 16.

    Example (compare the last to example in L{ParserElement.transformString}::
        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')
        
        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')
    prints::
        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    cs��fdd�|D�S)Ncsg|]}�|g��R��qSryry)r��tokn�r�r�ryrzr��r|z(tokenMap.<locals>.pa.<locals>.<listcomp>ryr~r�ryrzr��sztokenMap.<locals>.par�r�)r�r�r�r)r�r�r�r�ryr�rzro�s 
�rocCst|���Sr��r�r�rwryryrzr{r|cCst|���Sr��r��lowerrwryryrzr{r|c	Cs�t|t�r|}t||d�}n|j}tttd�}|r�t���	t
�}td�|d�tt
t|td�|���tddgd��d	��	d
d��td�}n�d
�dd�tD��}t���	t
�t|�B}td�|d�tt
t|�	t�ttd�|����tddgd��d	��	dd��td�}ttd�|d�}|�dd
�|�dd��������d|�}|�dd
�|�dd��������d|�}||_||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namer�z_-:rL�tag�=�/F�rrEcSs|ddkS�Nrr�ryr~ryryrzr{r|z_makeTags.<locals>.<lambda>rMr�css|]}|dvr|VqdS)rMNryr�ryryrzr�r|z_makeTags.<locals>.<genexpr>cSs|ddkSr�ryr~ryryrzr{r|rNr��:rqz<%s>r�z</%s>)r}r�rr�r1r6r5r@r�r�r\r-r
r4rrr�r�rXr[rDr�_Lr��titler�r�r�)�tagStr�xml�resname�tagAttrName�tagAttrValue�openTagZprintablesLessRAbrack�closeTagryryrz�	_makeTagss>
�������..r�cCs
t|d�S)a 
    Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
    tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.

    Example::
        text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
        a,a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end
        
        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are also accessible as named results
            print(link.link_text, '->', link.href)
    prints::
        pyparsing -> http://pyparsing.wikispaces.com
    F�r��r�ryryrzrM(srMcCs
t|d�S)z�
    Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
    tags only in the given upper/lower case.

    Example: similar to L{makeHTMLTags}
    Tr�r�ryryrzrN;srNcs8|r|dd��n|���dd��D���fdd�}|S)a<
    Helper to create a validating parse action to be used with start tags created
    with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
    with a required attribute value, to avoid false matches on common tags such as
    C{<TD>} or C{<DIV>}.

    Call C{withAttribute} with a series of attribute names and values. Specify the list
    of filter attributes names and values as:
     - keyword arguments, as in C{(align="right")}, or
     - as an explicit dict with C{**} operator, when an attribute name is also a Python
          reserved word, as in C{**{"class":"Customer", "align":"right"}}
     - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") )
    For attribute names with a namespace prefix, you must use the second form.  Attribute
    names are matched insensitive to upper/lower case.
       
    If just testing for C{class} (with or without a namespace), use C{L{withClass}}.

    To verify that the attribute exists, but without specifying a value, pass
    C{withAttribute.ANY_VALUE} as the value.

    Example::
        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    NcSsg|]\}}||f�qSryryrDryryrzr�zr|z!withAttribute.<locals>.<listcomp>csZ�D]P\}}||vr$t||d|��|tjkr|||krt||d||||f��qdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r!rg�	ANY_VALUE)r�rr��attrName�	attrValue��attrsryrzr�{s�zwithAttribute.<locals>.pa)r�)r��attrDictr�ryr�rzrgDs2rgcCs"|rd|nd}tfi||i��S)a�
    Simplified version of C{L{withAttribute}} when matching on a div class - made
    difficult because C{class} is a reserved word in Python.

    Example::
        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))
        
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    z%s:class�class)rg)�	classname�	namespace�	classattrryryrzrm�s rm�(r�cCs�t�}||||B}t|�D�]l\}}|ddd�\}}	}
}|	dkrPd|nd|}|	dkr�|dustt|�dkr|td��|\}
}t��|�}|
tjk�r^|	d	kr�t||�t|t	|��}n�|	dk�r|dur�t|||�t|t	||��}nt||�t|t	|��}nD|	dk�rTt||
|||�t||
|||�}ntd
��n�|
tj
k�rB|	d	k�r�t|t��s�t|�}t|j
|�t||�}n�|	dk�r�|du�r�t|||�t|t	||��}nt||�t|t	|��}nD|	dk�r8t||
|||�t||
|||�}ntd
��ntd��|�rvt|ttf��rl|j|�n
|�|�||�|�|BK}|}q||K}|S)aD

    Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary or
    binary, left- or right-associative.  Parse actions can also be attached
    to operator expressions. The generated parser will also recognize the use 
    of parentheses to override operator precedences (see example below).
    
    Note: if you define a deep operator list, you may see performance issues
    when using infixNotation. See L{ParserElement.enablePackrat} for a
    mechanism to potentially improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the nested
     - opList - list of tuples, one for each operator precedence level in the
      expression grammar; each tuple is of the form
      (opExpr, numTerms, rightLeftAssoc, parseAction), where:
       - opExpr is the pyparsing expression for the operator;
          may also be a string, which will be converted to a Literal;
          if numTerms is 3, opExpr is a tuple of two expressions, for the
          two operators separating the 3 terms
       - numTerms is the number of terms for this operator (must
          be 1, 2, or 3)
       - rightLeftAssoc is the indicator whether the operator is
          right or left associative, using the pyparsing-defined
          constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}.
       - parseAction is the parse action to be associated with
          expressions matching this operator expression (the
          parse action tuple member may be omitted); if the parse action
          is passed a tuple or list of functions, this is equivalent to
          calling C{setParseAction(*fn)} (L{ParserElement.setParseAction})
     - lpar - expression for matching left-parentheses (default=C{Suppress('(')})
     - rpar - expression for matching right-parentheses (default=C{Suppress(')')})

    Example::
        # simple example of four-function arithmetic with ints and variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier 
        
        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])
        
        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)
    prints::
        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    r�Nr�rqz%s termz	%s%s termrsz@if numterms=3, opExpr must be a tuple or list of two expressionsr�z6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)rr�r�r@r�rV�LEFTrrr�RIGHTr}rrvr?r�r�)�baseExpr�opList�lpar�rparr��lastExprr��operDef�opExpr�arity�rightLeftAssocr��termName�opExpr1�opExpr2�thisExprr~ryryrzrk�sZ=
&
�



&
�

rkz4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�'z string enclosed in single quotesz*quotedString using single or double quotes�uzunicode string literalcCs�||krtd��|du�r*t|t��r"t|t��r"t|�dkr�t|�dkr�|dur�tt|t||tjdd����	dd��}n$t
��t||tj��	dd��}nx|dur�tt|t|�t|�ttjdd����	dd��}n4ttt|�t|�ttjdd����	d	d��}ntd
��t
�}|du�rd|tt|�t||B|B�t|��K}n$|tt|�t||B�t|��K}|�d||f�|S)a~	
    Helper method for defining nested lists enclosed in opening and closing
    delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression
     - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression
     - content - expression for items within the nested lists (default=C{None})
     - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString})

    If an expression is not provided for the content argument, the nested
    expression will capture all whitespace-delimited content between delimiters
    as a list of separate values.

    Use the C{ignoreExpr} argument to define expressions that may contain
    opening or closing characters that should not be treated as opening
    or closing characters for nesting, such as quotedString or a comment
    expression.  Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}.
    The default is L{quotedString}, but if no expressions are to be ignored,
    then pass C{None} for this argument.

    Example::
        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR,RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type") 
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR 
                      + code_body("body"))
        c_function.ignore(cStyleComment)
        
        source_code = '''
            int is_odd(int x) { 
                return (x%2); 
            }
                
            int dec_to_hex(char hchar) { 
                if (hchar >= '0' && hchar <= '9') { 
                    return (ord(hchar)-ord('0')); 
                } else { 
                    return (10+ord(hchar)-ord('A'));
                } 
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)

    prints::
        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    z.opening and closing strings cannot be the sameNr�r�cSs|d��Sr��r�rwryryrzr{gr|znestedExpr.<locals>.<lambda>cSs|d��Sr�r�rwryryrzr{jr|cSs|d��Sr�r�rwryryrzr{pr|cSs|d��Sr�r�rwryryrzr{tr|zOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)r@r}r�r�rrrr&r�r�rEr�rrrr-r4r�)�opener�closer�contentr2r�ryryrzrR%sH:
���������
*$rRcs��fdd�}�fdd�}�fdd�}tt��d����}t�t��|��d�}t��|��d	�}t��|��d
�}	|r�tt|�|t|t|�t|��|	�}
n$tt|�t|t|�t|���}
|�	t
t��|
�d�S)a
	
    Helper method for defining space-delimited indentation blocks, such as
    those used to define block statements in Python source code.

    Parameters:
     - blockStatementExpr - expression defining syntax of statement that
            is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
            (multiple statementWithIndentedBlock expressions within a single grammar
            should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond the
            the current level; set to False for block of left-most statements
            (default=C{True})

    A valid block must contain at least one C{blockStatement}.

    Example::
        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group( funcDecl + func_body )

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << ( funcDef | assignment | identifier )

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()
    prints::
        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] 
    csN|t|�krdSt||�}|�dkrJ|�dkr>t||d��t||d��dS)Nrtzillegal nestingznot a peer entry)r�r;r#r!�r�rrx�curCol��indentStackryrz�checkPeerIndent�s
z&indentedBlock.<locals>.checkPeerIndentcs2t||�}|�dkr"��|�nt||d��dS)Nrtznot a subentry)r;r$r!r�r�ryrz�checkSubIndent�s
z%indentedBlock.<locals>.checkSubIndentcsN|t|�krdSt||�}�r6|�dkr6|�dksBt||d�����dS)Nrtr�znot an unindent)r�r;r!rr�r�ryrz�
checkUnindent�s

z$indentedBlock.<locals>.checkUnindentz	 �INDENTr��UNINDENTzindented block)rrrPrMrr�r�rrrRrj)�blockStatementExprr�rRr�r�r�rer��PEER�UNDENT�smExprryr�rzrhs(N����rhz#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z&(?P<entity>roz);zcommon HTML entitycCst�|j�S)zRHelper parser action to replace common HTML entities with their special characters)�_htmlEntityMapr��entityrwryryrzr]�sr]z/\*(?:[^*]|\*(?!/))*z*/zC style commentz<!--[\s\S]*?-->zHTML commentz.*zrest of linez//(?:\\\n|[^\n])*z
// commentzC++ style commentz#.*zPython style commentr�� 	�	commaItemr�r<c@s�eZdZdZee�Zee�Ze	e
��d��e�Z
e	e��d��eed��Zed��d��e�Ze��e�de��e��d�Ze�d	d
��eeeed���e�B�d�Ze�e�ed
��d��e�Zed��d��e�ZeeBeB��Zed��d��e�Ze	eded��d�Zed��d�Z ed��d�Z!e!de!d�d�Z"ee!de!d�dee!de!d��d�Z#e#�$dd
��d e �d!�Z%e&e"e%Be#B�d"���d"�Z'ed#��d$�Z(e)d=d&d'��Z*e)d>d)d*��Z+ed+��d,�Z,ed-��d.�Z-ed/��d0�Z.e/��e0��BZ1e)d1d2��Z2e&e3e4d3�e5�e	e6d3d4�ee7d5�������d6�Z8e9ee:�;�e8Bd7d8���d9�Z<e)ed:d
���Z=e)ed;d
���Z>d<S)?rpa�

    Here are some common low-level expressions that may be useful in jump-starting parser development:
     - numeric forms (L{integers<integer>}, L{reals<real>}, L{scientific notation<sci_real>})
     - common L{programming identifiers<identifier>}
     - network addresses (L{MAC<mac_address>}, L{IPv4<ipv4_address>}, L{IPv6<ipv6_address>})
     - ISO8601 L{dates<iso8601_date>} and L{datetime<iso8601_datetime>}
     - L{UUID<uuid>}
     - L{comma-separated list<comma_separated_list>}
    Parse actions:
     - C{L{convertToInteger}}
     - C{L{convertToFloat}}
     - C{L{convertToDate}}
     - C{L{convertToDatetime}}
     - C{L{stripHTMLTags}}
     - C{L{upcaseTokens}}
     - C{L{downcaseTokens}}

    Example::
        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')
    prints::
        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    �integerzhex integerr�z[+-]?\d+zsigned integerr��fractioncCs|d|dS)Nrrtryrwryryrzr{�r|zpyparsing_common.<lambda>r�z"fraction or mixed integer-fractionz
[+-]?\d+\.\d*zreal numberz+[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?�fnumberr�
identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}zIPv4 addressz[0-9a-fA-F]{1,4}�hex_integerr��zfull IPv6 address)rr�z::zshort IPv6 addresscCstdd�|D��dkS)Ncss|]}tj�|�rdVqdSrI)rp�
_ipv6_partr%r^ryryrzr��r|z,pyparsing_common.<lambda>.<locals>.<genexpr>r�)rrwryryrzr{�r|z::ffff:zmixed IPv6 addresszIPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}zMAC address�%Y-%m-%dcs�fdd�}|S)a�
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"})

        Example::
            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))
        prints::
            [datetime.date(1999, 12, 31)]
        c
sPzt�|d����WStyJ}zt||t|���WYd}~n
d}~00dSr�)r�strptime�dater@r!r�r�rrx�ve��fmtryrz�cvt_fn�sz.pyparsing_common.convertToDate.<locals>.cvt_fnry�r�r�ryr�rz�
convertToDate�szpyparsing_common.convertToDate�%Y-%m-%dT%H:%M:%S.%fcs�fdd�}|S)a
        Helper to create a parse action for converting parsed datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"})

        Example::
            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))
        prints::
            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        c
sLzt�|d��WStyF}zt||t|���WYd}~n
d}~00dSr�)rr�r@r!rr�r�ryrzr��sz2pyparsing_common.convertToDatetime.<locals>.cvt_fnryr�ryr�rz�convertToDatetime�sz"pyparsing_common.convertToDatetimez7(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?zISO8601 datez�(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}�UUIDcCstj�|d�S)a
        Parse action to remove HTML tags from web page HTML source

        Example::
            # strip HTML links from normal text 
            text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
            td,td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            
            print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page'
        r)rp�_html_stripperr�)r�rr�ryryrz�
stripHTMLTagss
zpyparsing_common.stripHTMLTagsrSr�r�r�r�r�zcomma separated listcCst|���Sr�r�rwryryrzr{"r|cCst|���Sr�r�rwryryrzr{%r|N)r�)r�)?r�r�r�r�rorv�convertToInteger�float�convertToFloatr1rTr�r�r�rFr�r)�signed_integerr�r�rrM�
mixed_integerr�real�sci_realr�numberr�r6r5r��ipv4_addressr��_full_ipv6_address�_short_ipv6_addressr��_mixed_ipv6_addressr�ipv6_address�mac_addressrr�r��iso8601_date�iso8601_datetime�uuidr9r8r�r�rrrrXr0�
_commasepitemrBr[r��comma_separated_listrfrDryryryrzrpsV""
2
 
��rp�__main__�select�fromr�rL)rU�columnsr�Ztables�commandaK
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        z]
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        z 
        100
        FF
        z6
        12345678-1234-5678-1234-567812345678
        )rs)rSF)N)FT)T)r�)T)�r��__version__�__versionTime__�
__author__r��weakrefrr�r�r�r4r�r�rrfr�r�r�_threadr�ImportError�	threading�collections.abcrrrrZordereddict�__all__r?�version_infor�rE�maxsizer�rr��chrr�r�rr�r`�reversedr�r�rbr�r�r�r�Zmaxint�xranger��__builtin__r��fnamer$r�r�r�r�r�r�r��ascii_uppercase�ascii_lowercaser6rTrFr5rjr��	printablerXr�rr!r#r%r(r�r$�registerr;rLrIrwr{r}rSr�r&r.rrrr�r�rr
r	rnr1r)r'rr0r�rrrr,r+r3r2r"rrrrr rrr$rr4r0r1rr*rr?r/rrr
r-rrdrBr>r*rQrPr�rUrCrirjrlr�rErKrJrcrbr��_escapedPunc�_escapedHexChar�_escapedOctChar�_singleChar�
_charRanger�r�rarOr^r\rorfrDr�rMrNrgr�rmrVr�r�rkrWr@r`r[rerRrhr7rYr9r8r�r�r�rr=r]r:rGrOr_rAr?rHrZrrr<rpr�ZselectTokenZ	fromToken�identZ
columnNameZcolumnNameListZ
columnSpecZ	tableNameZ
tableNameListZ	simpleSQLr~r�r�r�rr�ryryryrz�<module>s��4
8



@v&A= I
G3pLOD|M &#@sQ,A,	I#%0
,	?#p
��Zr 
����"


"
	site-packages/setuptools/_vendor/packaging/_typing.py000064400000003424150564745010017122 0ustar00"""For neatly implementing static typing in packaging.

`mypy` - the static type analysis tool we use - uses the `typing` module, which
provides core functionality fundamental to mypy's functioning.

Generally, `typing` would be imported at runtime and used in that fashion -
it acts as a no-op at runtime and does not have any run-time overhead by
design.

As it turns out, `typing` is not vendorable - it uses separate sources for
Python 2/Python 3. Thus, this codebase can not expect it to be present.
To work around this, mypy allows the typing import to be behind a False-y
optional to prevent it from running at runtime and type-comments can be used
to remove the need for the types to be accessible directly during runtime.

This module provides the False-y guard in a nicely named fashion so that a
curious maintainer can reach here to read this.

In packaging, all static-typing related imports should be guarded as follows:

    from packaging._typing import TYPE_CHECKING

    if TYPE_CHECKING:
        from typing import ...

Ref: https://github.com/python/mypy/issues/3216
"""

__all__ = ["TYPE_CHECKING", "cast"]

# The TYPE_CHECKING constant defined by the typing module is False at runtime
# but True while type checking.
if False:  # pragma: no cover
    from typing import TYPE_CHECKING
else:
    TYPE_CHECKING = False

# typing's cast syntax requires calling typing.cast at runtime, but we don't
# want to import typing at runtime. Here, we inform the type checkers that
# we're importing `typing.cast` as `cast` and re-implement typing.cast's
# runtime behavior in a block that is ignored by type checkers.
if TYPE_CHECKING:  # pragma: no cover
    # not executed at runtime
    from typing import cast
else:
    # executed at runtime
    def cast(type_, value):  # noqa
        return value
site-packages/setuptools/_vendor/packaging/__pycache__/utils.cpython-39.pyc000064400000003247150564745010023103 0ustar00a

(�Re�@sxddlmZmZmZddlZddlmZmZddlm	Z	m
Z
erZddlmZm
Z
ede�Ze�d�Zd	d
�Zdd�ZdS)
�)�absolute_import�division�print_functionN�)�
TYPE_CHECKING�cast)�InvalidVersion�Version)�NewType�Union�NormalizedNamez[-_.]+cCst�d|���}td|�S)N�-r)�_canonicalize_regex�sub�lowerr)�name�value�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/utils.py�canonicalize_namesrc
Cs�zt|�}Wnty"|YS0g}|jdkrD|�d�|j��|�t�ddd�dd�|jD����|j	dur�|�d�d	d�|j	D���|j
dur�|�d
�|j
��|jdur�|�d�|j��|jdur�|�d�|j��d�|�S)
z�
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    rz{0}!z(\.0)+$��.css|]}t|�VqdS�N��str��.0�xrrr�	<genexpr>/�z'canonicalize_version.<locals>.<genexpr>Ncss|]}t|�VqdSrrrrrrr3rz.post{0}z.dev{0}z+{0})
r	r�epoch�append�format�rer�join�release�pre�post�dev�local)�_version�version�partsrrr�canonicalize_versions"

&



r-)�
__future__rrrr#Z_typingrrr+rr	�typingr
rrr�compilerrr-rrrr�<module>s

site-packages/setuptools/_vendor/packaging/__pycache__/tags.cpython-39.pyc000064400000041640150564745010022700 0ustar00a

(�Re^�@sLddlmZddlZzddlmZWn.eyRddlZdd�e��D�Z[Yn0ddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZmZer�ddlmZmZmZmZmZmZmZmZmZmZeeZ eeefZ!eeefZ"e	�#e$�Z%d	d
ddd
d�Z&ej'dkZ(Gdd�de)�Z*dd�Z+dd�Z,dSdd�Z-dd�Z.dd�Z/dTdd�Z0dUdd �Z1d!d"�Z2dVd#d$�Z3d%d&�Z4dWd'd(�Z5e(fd)d*�Z6d+d,�Z7dXd-d.�Z8d/d0�Z9d1d2�Z:d3d4�Z;d5d6�Z<d7d8�Z=d9d:�Z>Gd;d<�d<e)�Z?d=d>�Z@d?d@�ZAdAdB�ZBdCdD�ZCe(fdEdF�ZDdGdH�ZEdIdJ�ZFdKdL�ZGdMdN�ZHdOdP�ZIdQdR�ZJdS)Y�)�absolute_importN)�EXTENSION_SUFFIXEScCsg|]}|d�qS)r�)�.0�xrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/tags.py�
<listcomp>�r�)�
TYPE_CHECKING�cast)
�Dict�	FrozenSet�IO�Iterable�Iterator�List�Optional�Sequence�Tuple�Union�py�cp�pp�ip�jy)�python�cpython�pypy�
ironpython�jythonlc@sdeZdZdZgd�Zdd�Zedd��Zedd��Zed	d
��Z	dd�Z
d
d�Zdd�Zdd�Z
dS)�Tagz�
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    )�_interpreter�_abi�	_platformcCs"|��|_|��|_|��|_dS�N)�lowerr"r#r$)�self�interpreter�abi�platformrrr�__init__Fs

zTag.__init__cCs|jSr%)r"�r'rrrr(LszTag.interpretercCs|jSr%)r#r,rrrr)QszTag.abicCs|jSr%)r$r,rrrr*VszTag.platformcCs2t|t�stS|j|jko0|j|jko0|j|jkSr%)�
isinstancer!�NotImplementedr*r)r()r'�otherrrr�__eq__[s

�
�z
Tag.__eq__cCst|j|j|jf�Sr%)�hashr"r#r$r,rrr�__hash__fszTag.__hash__cCsd�|j|j|j�S)Nz{}-{}-{})�formatr"r#r$r,rrr�__str__jszTag.__str__cCsdj|t|�d�S)Nz<{self} @ {self_id}>)r'�self_id)r3�idr,rrr�__repr__nszTag.__repr__N)�__name__�
__module__�__qualname__�__doc__�	__slots__r+�propertyr(r)r*r0r2r4r7rrrrr!<s


r!c	Cs`t�}|�d�\}}}|�d�D]6}|�d�D]&}|�d�D]}|�t|||��q<q.q t|�S)z�
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    �-�.)�set�split�addr!�	frozenset)�tag�tags�interpreters�abis�	platformsr(r)�	platform_rrr�	parse_tagssrJcCsP|sdSt|�dksd|vrH|�dd�tt|����}td�||���|dS)z[
    Backwards-compatibility with Python 2.7 to allow treating 'warn' as keyword-only.
    Fr
�warnNz,{}() got an unexpected keyword argument {!r})�len�pop�next�iter�keys�	TypeErrorr3)�	func_name�kwargs�argrrr�_warn_keyword_parameter�s
�rUFcCs&t�|�}|dur"|r"t�d|�|S)Nz>Config variable '%s' is unset, Python ABI tag may be incorrect)�	sysconfig�get_config_var�logger�debug)�namerK�valuerrr�_get_config_var�s
�r\cCs|�dd��dd�S)Nr?�_r>)�replace)�stringrrr�_normalize_string�sr`cCst|�dkot|�dkS)zj
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    r
)��)rL�tuple)�python_versionrrr�
_abi3_applies�srec	Cs�t|�}g}t|dd��}d}}}td|�}ttd�}dtv}	|sX|dur\|sX|	r\d}|dkr�td|�}
|
sz|
dur~d	}|d
kr�td|�}|dks�|dur�tjd
kr�d}n|r�|�dj|d��|�	ddj||||d��|S)Nrb��Py_DEBUG�gettotalrefcountz_d.pyd�d)ra��
WITH_PYMALLOC�m)rara�Py_UNICODE_SIZE�i���u�cp{version}��versionrz"cp{version}{debug}{pymalloc}{ucs4})rrrY�pymalloc�ucs4)
rc�_version_nodotr\�hasattr�sysr�
maxunicode�appendr3�insert)�
py_versionrKrGrrrYrsrt�
with_debug�has_refcount�has_ext�
with_pymalloc�unicode_sizerrr�
_cpython_abis�s<



����r�c
	+sZtd|�}|stjdd�}d�t|dd����|durVt|�dkrRt||�}ng}t|�}dD]&}z|�|�Wqbt	y�Yqb0qbt|p�t
��}|D]}|D]}t�||�Vq�q�t|�r�fdd�|D�D]
}|Vqԇfd	d�|D�D]
}|Vq�t|��rVt
|dddd
�D]8}	|D],}djt|d|	f�d
��t�d|�V�q$�qdS)a
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    �cpython_tagsNrbzcp{}r
)�abi3�nonec3s|]}t�d|�VqdS)r�N�r!�rrI�r(rr�	<genexpr>r	zcpython_tags.<locals>.<genexpr>c3s|]}t�d|�VqdS)r�Nr�r�r�rrr�r	���rprrqr�)rUrw�version_infor3rurLr��list�remove�
ValueError�_platform_tagsr!re�range)
rdrGrHrSrK�explicit_abir)rIrD�
minor_versionrr�rr��s<

�r�ccst�d�}|rt|�VdS)N�SOABI)rVrWr`)r)rrr�_generic_abis
r�c	ks�td|�}|s,t�}t|d�}d�||g�}|dur:t�}t|pDt��}t|�}d|vrb|�d�|D]}|D]}t|||�VqnqfdS)z�
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    �generic_tags�rKrfNr�)	rU�interpreter_name�interpreter_version�joinr�r�r�ryr!)	r(rGrHrSrK�interp_name�interp_versionr)rIrrrr�s


r�ccs|t|�dkr&djt|dd��d�Vdj|dd�Vt|�dkrxt|ddd	d	�D]}djt|d|f�d�VqXdS)
z�
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    r
zpy{version}Nrbrqz	py{major}r)�majorr�)rLr3rur�)r{�minorrrr�_py_interpreter_range4sr�ccsx|stjdd�}t|pt��}t|�D]}|D]}t|d|�Vq0q(|rXt|dd�Vt|�D]}t|dd�Vq`dS)z�
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    Nrbr��any)rwr�r�r�r�r!)rdr(rHrrrIrrr�compatible_tagsDsr�cCs|s|S|�d�rdSdS)N�ppc�i386)�
startswith)�arch�is_32bitrrr�	_mac_arch^s

r�cCs�|g}|dkr*|dkrgS|�gd��nn|dkrN|dkr>gS|�gd��nJ|dkrv|dksf|dkrjgS|�d�n"|d	kr�|d
kr�gS|�ddg�|�d
�|S)N�x86_64)�
rn)�intel�fat64�fat32r�)r�r��fat�ppc64)r��r�r�)r��r�r��	universal)�extendry)rr�cpu_arch�formatsrrr�_mac_binary_formatsis&
r�c	cs�t��\}}}|dur:tdttt|�d�dd����}n|}|durPt|�}n|}t|ddd�D]>}|d|f}t	||�}|D]}dj
|d|d|d	�Vq�qddS)
aD
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    N�
MacVersionr?rbr
r�rz&macosx_{major}_{minor}_{binary_format})r�r��
binary_format)r*�mac_verrrc�map�intrAr�r�r�r3)	rrr��version_strr]r�r��compat_version�binary_formatsr�rrr�
mac_platforms�s 
$

�r�c	Cs<zddl}tt||d��WSttfy2Yn0t|�S)NrZ_compatible)�
_manylinux�bool�getattr�ImportError�AttributeError�_have_compatible_glibc)rZ�
glibc_versionr�rrr�_is_manylinux_compatible�sr�cCst�p
t�Sr%)�_glibc_version_string_confstr�_glibc_version_string_ctypesrrrr�_glibc_version_string�sr�cCsHz&t�d�}|dusJ�|��\}}WnttttfyBYdS0|S)zJ
    Primary implementation of glibc_version_string using os.confstr.
    �CS_GNU_LIBC_VERSIONN)�os�confstrrA�AssertionErrorr��OSErrorr�)�version_stringr]rrrrrr��s	�r�cCsrzddl}Wnty YdS0|�d�}z
|j}WntyJYdS0|j|_|�}t|t�sn|�	d�}|S)zG
    Fallback implementation of glibc_version_string using ctypes.
    rN�ascii)
�ctypesr��CDLL�gnu_get_libc_versionr��c_char_p�restyper-�str�decode)r��process_namespacer�r�rrrr��s



r�cCsHt�d|�}|s$t�d|t�dSt|�d��|koFt|�d��|kS)Nz$(?P<major>[0-9]+)\.(?P<minor>[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFr�r�)�re�match�warningsrK�RuntimeWarningr��group)r��required_major�
minimum_minorrlrrr�_check_glibc_version�s���r�cCst�}|durdSt|||�S�NF)r�r�)r�r�r�rrrr�sr�c@sTeZdZGdd�de�ZdZdZdZdZdZ	dZ
dZdZd	Z
d
ZdZdZd
d�ZdS)�_ELFFileHeaderc@seZdZdZdS)z$_ELFFileHeader._InvalidELFFileHeaderz7
        An invalid ELF file header was found.
        N)r8r9r:r;rrrr�_InvalidELFFileHeadersr�iFLEr
rbra��(�>l~iicsr�fdd�}|d�|_|j|jkr*t���|d�|_|j|j|jhvrNt���|d�|_|j|j|j	hvrrt���|d�|_
|d�|_|d�|_��
d�|_|j|jkr�dnd}|j|jkr�dnd}|j|jkr�d	nd
}|j|jkr�|n|}||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_dS)Ncs@zt�|��t�|���\}Wntjy:t���Yn0|Sr%)�struct�unpack�read�calcsize�errorr�r�)�fmt�result��filerrr�)s�
z'_ELFFileHeader.__init__.<locals>.unpackz>I�B�z<Hz>Hz<Iz<Qz>Q)�
e_ident_magic�ELF_MAGIC_NUMBERr�r��
e_ident_class�
ELFCLASS32�
ELFCLASS64�e_ident_data�ELFDATA2LSB�ELFDATA2MSB�e_ident_version�
e_ident_osabi�e_ident_abiversionr��e_ident_pad�e_type�	e_machine�	e_version�e_entry�e_phoff�e_shoff�e_flags�e_ehsize�e_phentsize�e_phnum�e_shentsize�e_shnum�
e_shstrndx)r'r�r��format_h�format_i�format_q�format_prr�rr+'s>


















z_ELFFileHeader.__init__N)r8r9r:r�r�r�r�r�r�r��EM_386�EM_S390�EM_ARM�	EM_X86_64�EF_ARM_ABIMASK�EF_ARM_ABI_VER5�EF_ARM_ABI_FLOAT_HARDr+rrrrr�sr�cCs\z8ttjd��}t|�}Wd�n1s,0YWnttttjfyVYdS0|S)N�rb)�openrw�
executabler��IOErrorr�rQr�)�f�
elf_headerrrr�_get_elf_headerSs*rcCsnt�}|durdS|j|jk}||j|jkM}||j|jkM}||j|j@|j	kM}||j|j
@|j
kM}|Sr�)rr�r�r�r�r�rrrrr�rr�rrr�_is_linux_armhf]s
�
�rcCsBt�}|durdS|j|jk}||j|jkM}||j|jkM}|Sr�)rr�r�r�r�r�r
rrrr�_is_linux_i686qsrcCs |dkrt�S|dkrt�SdS)N�armv7l�i686T)rr)r�rrr�_have_compatible_manylinux_abi|s
r ccs�ttj���}|r,|dkr d}n|dkr,d}g}|�dd�\}}t|�rv|dvrZ|�d�|d	vrv|�d
�|�d�t|�}|D]$\}}t||�r�|�	d|�Vq�q�|D]\}}|�	d|�Vq�|VdS)
N�linux_x86_64�
linux_i686�
linux_aarch64�linux_armv7lr]r
>�ppc64le�aarch64r�s390xrr�r�)�
manylinux2014)rb�>r�r)�
manylinux2010)rb�)�
manylinux1)rbr��linux)
r`�	distutils�util�get_platformrAr ryrOr�r^)r�r-Zmanylinux_supportr]r�Zmanylinux_support_iterrZr�rrr�_linux_platforms�s8���
r1ccsttj���VdSr%)r`r.r/r0rrrr�_generic_platforms�sr2cCs.t��dkrt�St��dkr$t�St�SdS)z;
    Provides the platform tags for this installation.
    �Darwin�LinuxN)r*�systemr�r1r2rrrrr��s
r�cCs:ztjj}Wnty*t����}Yn0t�|�p8|S)z6
    Returns the name of the running interpreter.
    )	rw�implementationrZr�r*�python_implementationr&�INTERPRETER_SHORT_NAMES�get)rZrrrr��s
r�cKs:td|�}td|d�}|r$t|�}nttjdd��}|S)z9
    Returns the version of the running interpreter.
    r��py_version_nodotr�Nrb)rUr\r�rurwr�)rSrKrrrrrr��s

r�cCs,tdd�|D��rd}nd}|�tt|��S)Ncss|]}|dkVqdS)r�Nr)r�vrrrr��r	z!_version_nodot.<locals>.<genexpr>r]rf)r�r�r�r�)rr�seprrrru�srucksXtd|�}t�}|dkr0t|d�D]
}|Vq"nt�D]
}|Vq6t�D]
}|VqHdS)z�
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    �sys_tagsrr�N)rUr�r�r�r�)rSrKr�rDrrrr=�s



r=)F)F)NNN)NNN)NNN)NN)K�
__future__r�distutils.utilr.�importlib.machineryrr��impZget_suffixes�loggingr�r*r�r�rwrVr�Z_typingrr�typingr
rrrrrrrrrr��
PythonVersionr�ZGlibcVersion�	getLoggerr8rXr8�maxsize�_32_BIT_INTERPRETER�objectr!rJrUr\r`rer�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrr r1r2r�r�r�rur=rrrr�<module>s�0

�	
7



&�
<�
�

#@
	!

	site-packages/setuptools/_vendor/packaging/__pycache__/requirements.cpython-39.pyc000064400000010043150564745010024456 0ustar00a

(�Re5�@s�ddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZdd	lmZmZdd
lmZmZmZer�ddlmZGdd
�d
e �Z!eej"ej#�Z$ed��%�Z&ed��%�Z'ed��%�Z(ed��%�Z)ed��%�Z*ed��%�Z+ed��%�Z,ed�Z-e$ee-�e$BZ.ee$ee.��Z/e/d�Z0e/Z1ed�d�Z2e,e2Z3e1ee*e1�Z4e&e
e4�e'd�Z5eej6ej7ej8B�Z9eej6ej7ej8B�Z:e9e:AZ;ee;ee*e;�ddd�d�Z<e
e(e<e)e<B�Z=e=�>dd��e	e=�d�Z?e?�>d d��e	e��d!�Ze�>d"d��e+Z@e@eZAe?e
eA�ZBe3e
eA�ZCe0e
e5�eCeBBZDeeDeZEeE�Fd#�Gd$d%�d%eG�ZHdS)&�)�absolute_import�division�print_functionN)�stringStart�	stringEnd�originalTextFor�ParseException)�
ZeroOrMore�Word�Optional�Regex�Combine)�Literal)�parse�)�
TYPE_CHECKING)�MARKER_EXPR�Marker)�LegacySpecifier�	Specifier�SpecifierSet)�Listc@seZdZdZdS)�InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/requirements.pyrsr�[�]�(�)�,�;�@z-_.�namez[^ ]+�url�extrasF)�
joinString�adjacent�	_raw_speccCs
|jpdS)N�)r+��s�l�trrr�<lambda>;�r1�	specifiercCs|dS)Nrrr-rrrr1>r2�markercCst||j|j��S)N)r�_original_start�
_original_endr-rrrr1Br2zx[]c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�Requirementz�Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt�|�}WnFtyT}z.td�||j|jd�|j���WYd}~n
d}~00|j|_|jr�t	�	|j�}|j
dkr�t	�|�|jkr�td��n(|j
r�|jr�|j
s�|js�td�|j���|j|_nd|_t
|jr�|j��ng�|_t|j�|_|jr�|jnd|_dS)NzParse error at "{0!r}": {1}��filezInvalid URL givenzInvalid URL: {0})�REQUIREMENT�parseStringrr�format�loc�msgr&r'�urlparse�scheme�
urlunparse�netloc�setr(�asListrr3r4)�self�requirement_string�req�e�
parsed_urlrrr�__init___s2��

��
zRequirement.__init__cCs�|jg}|jr*|�d�d�t|j����|jr@|�t|j��|jrh|�d�|j��|j	rh|�d�|j	r�|�d�|j	��d�|�S)Nz[{0}]r#z@ {0}� z; {0}r,)
r&r(�appendr<�join�sortedr3�strr'r4)rE�partsrrr�__str__{s
zRequirement.__str__cCsd�t|��S)Nz<Requirement({0!r})>)r<rO)rErrr�__repr__�szRequirement.__repr__N)rrrrrJrQrRrrrrr7Rsr7)I�
__future__rrr�string�reZsetuptools.extern.pyparsingrrrrr	r
rrr
r�L�urllibrr?Z_typingr�markersrr�
specifiersrrr�typingr�
ValueErrorr�
ascii_letters�digits�ALPHANUM�suppress�LBRACKET�RBRACKET�LPAREN�RPAREN�COMMA�	SEMICOLON�AT�PUNCTUATION�IDENTIFIER_END�
IDENTIFIER�NAME�EXTRA�URI�URL�EXTRAS_LIST�EXTRAS�
_regex_str�VERBOSE�
IGNORECASE�VERSION_PEP440�VERSION_LEGACY�VERSION_ONE�VERSION_MANY�
_VERSION_SPEC�setParseAction�VERSION_SPEC�MARKER_SEPARATOR�MARKER�VERSION_AND_MARKER�URL_AND_MARKER�NAMED_REQUIREMENTr:r;�objectr7rrrr�<module>sj���
site-packages/setuptools/_vendor/packaging/__pycache__/_structures.cpython-39.pyc000064400000005606150564745010024326 0ustar00a

(�Re��@sDddlmZmZmZGdd�de�Ze�ZGdd�de�Ze�ZdS)�)�absolute_import�division�print_functionc@sTeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZdS)�InfinityTypecCsdS)N�Infinity���selfrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/_structures.py�__repr__szInfinityType.__repr__cCstt|��S�N��hash�reprrrrr
�__hash__szInfinityType.__hash__cCsdS�NFr�r	�otherrrr
�__lt__szInfinityType.__lt__cCsdSrrrrrr
�__le__szInfinityType.__le__cCst||j�Sr��
isinstance�	__class__rrrr
�__eq__szInfinityType.__eq__cCst||j�Srrrrrr
�__ne__szInfinityType.__ne__cCsdS�NTrrrrr
�__gt__ szInfinityType.__gt__cCsdSrrrrrr
�__ge__$szInfinityType.__ge__cCstSr)�NegativeInfinityrrrr
�__neg__(szInfinityType.__neg__N��__name__�
__module__�__qualname__rrrrrrrrrrrrr
rsrc@sTeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZdS)�NegativeInfinityTypecCsdS)Nz	-Infinityrrrrr
r1szNegativeInfinityType.__repr__cCstt|��Srr
rrrr
r5szNegativeInfinityType.__hash__cCsdSrrrrrr
r9szNegativeInfinityType.__lt__cCsdSrrrrrr
r=szNegativeInfinityType.__le__cCst||j�Srrrrrr
rAszNegativeInfinityType.__eq__cCst||j�Srrrrrr
rEszNegativeInfinityType.__ne__cCsdSrrrrrr
rIszNegativeInfinityType.__gt__cCsdSrrrrrr
rMszNegativeInfinityType.__ge__cCstSr)rrrrr
rQszNegativeInfinityType.__neg__Nr rrrr
r$0sr$N)	�
__future__rrr�objectrrr$rrrrr
�<module>s&&site-packages/setuptools/_vendor/packaging/__pycache__/version.cpython-39.pyc000064400000032072150564745010023426 0ustar00a

(�Ren<�	@s�ddlmZmZmZddlZddlZddlZddlmZm	Z	ddl
mZe�r.ddlm
Z
mZmZmZmZmZmZddlmZmZeeefZeeeeeffZeeeefZeeeeeeeefeeeffdffZeeeedfeeeefZeeeedffZe
eeefeeefgefZgd	�Z e�!d
gd��Z"dd
�Z#Gdd�de$�Z%Gdd�de&�Z'Gdd�de'�Z(e�)dej*�Z+dddddd�Z,dd�Z-dd�Z.dZ/Gdd�de'�Z0d d!�Z1e�)d"�Z2d#d$�Z3d%d&�Z4dS)'�)�absolute_import�division�print_functionN�)�Infinity�NegativeInfinity)�
TYPE_CHECKING)�Callable�Iterator�List�Optional�SupportsInt�Tuple�Union)�InfinityType�NegativeInfinityType.)�parse�Version�
LegacyVersion�InvalidVersion�VERSION_PATTERN�_Version)�epoch�release�dev�pre�post�localcCs*z
t|�WSty$t|�YS0dS)z�
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rrr)�version�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/version.pyr0s
rc@seZdZdZdS)rzF
    An invalid version was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__rrrr r=src@sPeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�ZdS)�_BaseVersionNcCs
t|j�S�N)�hash�_key��selfrrr �__hash__Fsz_BaseVersion.__hash__cCs|�|dd��S)NcSs||kSr&r��s�orrr �<lambda>L�z%_BaseVersion.__lt__.<locals>.<lambda>��_compare�r*�otherrrr �__lt__Jsz_BaseVersion.__lt__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Pr0z%_BaseVersion.__le__.<locals>.<lambda>r1r3rrr �__le__Nsz_BaseVersion.__le__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Tr0z%_BaseVersion.__eq__.<locals>.<lambda>r1r3rrr �__eq__Rsz_BaseVersion.__eq__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Xr0z%_BaseVersion.__ge__.<locals>.<lambda>r1r3rrr �__ge__Vsz_BaseVersion.__ge__cCs|�|dd��S)NcSs||kSr&rr,rrr r/\r0z%_BaseVersion.__gt__.<locals>.<lambda>r1r3rrr �__gt__Zsz_BaseVersion.__gt__cCs|�|dd��S)NcSs||kSr&rr,rrr r/`r0z%_BaseVersion.__ne__.<locals>.<lambda>r1r3rrr �__ne__^sz_BaseVersion.__ne__cCst|t�stS||j|j�Sr&)�
isinstancer%�NotImplementedr()r*r4�methodrrr r2bs
z_BaseVersion._compare)r!r"r#r(r+r5r6r7r8r9r:r2rrrr r%Csr%c@s�eZdZdd�Zdd�Zdd�Zedd��Zed	d
��Zedd��Z	ed
d��Z
edd��Zedd��Zedd��Z
edd��Zedd��Zedd��Zedd��ZdS)rcCst|�|_t|j�|_dSr&)�str�_version�_legacy_cmpkeyr()r*rrrr �__init__ks
zLegacyVersion.__init__cCs|jSr&�r?r)rrr �__str__pszLegacyVersion.__str__cCsd�tt|���S)Nz<LegacyVersion({0})>��format�reprr>r)rrr �__repr__tszLegacyVersion.__repr__cCs|jSr&rBr)rrr �publicxszLegacyVersion.publiccCs|jSr&rBr)rrr �base_version}szLegacyVersion.base_versioncCsdS)N���rr)rrr r�szLegacyVersion.epochcCsdSr&rr)rrr r�szLegacyVersion.releasecCsdSr&rr)rrr r�szLegacyVersion.precCsdSr&rr)rrr r�szLegacyVersion.postcCsdSr&rr)rrr r�szLegacyVersion.devcCsdSr&rr)rrr r�szLegacyVersion.localcCsdS�NFrr)rrr �
is_prerelease�szLegacyVersion.is_prereleasecCsdSrKrr)rrr �is_postrelease�szLegacyVersion.is_postreleasecCsdSrKrr)rrr �
is_devrelease�szLegacyVersion.is_devreleaseN)r!r"r#rArCrG�propertyrHrIrrrrrrrLrMrNrrrr rjs2









rz(\d+ | [a-z]+ | \.| -)�czfinal-�@)r�preview�-�rcrccs\t�|�D]F}t�||�}|r
|dkr(q
|dd�dvrF|�d�Vq
d|Vq
dVdS)N�.r�
0123456789��*�*final)�_legacy_version_component_re�split�_legacy_version_replacement_map�get�zfill)r-�partrrr �_parse_version_parts�sr`cCsvd}g}t|���D]T}|�d�r^|dkrD|rD|ddkrD|��q*|r^|ddkr^|��qD|�|�q|t|�fS)NrJrXrYz*final-�00000000)r`�lower�
startswith�pop�append�tuple)rr�partsr_rrr r@�s


r@a�
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@s�eZdZe�dedejejB�Zdd�Z	dd�Z
dd�Zed	d
��Z
edd��Zed
d��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd ��Zed!d"��Zed#d$��Zd%S)&rz^\s*z\s*$c
Cs�|j�|�}|std�|���t|�d�r8t|�d��ndtdd�|�d��d�D��t	|�d�|�d	��t	|�d
�|�d�p�|�d��t	|�d
�|�d��t
|�d��d�|_t|jj
|jj|jj|jj|jj|jj�|_dS)NzInvalid version: '{0}'rrcss|]}t|�VqdSr&)�int��.0�irrr �	<genexpr>r0z#Version.__init__.<locals>.<genexpr>rrU�pre_l�pre_n�post_l�post_n1�post_n2�dev_l�dev_nr)rrrrrr)�_regex�searchrrEr�grouprhrfr[�_parse_letter_version�_parse_local_versionr?�_cmpkeyrrrrrrr()r*r�matchrrr rAs*���zVersion.__init__cCsd�tt|���S)Nz<Version({0})>rDr)rrr rG-szVersion.__repr__cCs�g}|jdkr |�d�|j��|�d�dd�|jD���|jdurb|�d�dd�|jD���|jdur~|�d�|j��|jdur�|�d	�|j��|jdur�|�d
�|j��d�|�S)Nr�{0}!rUcss|]}t|�VqdSr&�r>�rj�xrrr rl:r0z"Version.__str__.<locals>.<genexpr>�css|]}t|�VqdSr&r|r}rrr rl>r0z.post{0}z.dev{0}z+{0})	rrerE�joinrrrrr�r*rgrrr rC1s




zVersion.__str__cCs|jj}|Sr&)r?r)r*�_epochrrr rNsz
Version.epochcCs|jj}|Sr&)r?r)r*�_releaserrr rTszVersion.releasecCs|jj}|Sr&)r?r)r*�_prerrr rZszVersion.precCs|jjr|jjdSdS�Nr)r?rr)rrr r`szVersion.postcCs|jjr|jjdSdSr�)r?rr)rrr reszVersion.devcCs(|jjr d�dd�|jjD��SdSdS)NrUcss|]}t|�VqdSr&r|r}rrr rlnr0z Version.local.<locals>.<genexpr>)r?rr�r)rrr rjsz
Version.localcCst|��dd�dS)N�+rr)r>r[r)rrr rHrszVersion.publiccCsFg}|jdkr |�d�|j��|�d�dd�|jD���d�|�S)Nrr{rUcss|]}t|�VqdSr&r|r}rrr rl�r0z'Version.base_version.<locals>.<genexpr>r)rrerEr�rr�rrr rIws

zVersion.base_versioncCs|jdup|jduSr&)rrr)rrr rL�szVersion.is_prereleasecCs
|jduSr&)rr)rrr rM�szVersion.is_postreleasecCs
|jduSr&)rr)rrr rN�szVersion.is_devreleasecCst|j�dkr|jdSdS)Nrr��lenrr)rrr �major�sz
Version.majorcCst|j�dkr|jdSdS)N�rrr�r)rrr �minor�sz
Version.minorcCst|j�dkr|jdSdS)N�r�rr�r)rrr �micro�sz
Version.microN)r!r"r#�re�compiler�VERBOSE�
IGNORECASErtrArGrCrOrrrrrrrHrIrLrMrNr�r�r�rrrr rs@













rcCsv|rZ|durd}|��}|dkr&d}n(|dkr4d}n|dvrBd}n|dvrNd	}|t|�fS|sr|rrd	}|t|�fSdS)
Nr�alpha�a�beta�b)rPrrRrT)�rev�rr)rbrh)�letter�numberrrr rw�s"rwz[\._-]cCs$|dur tdd�t�|�D��SdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|��s|��nt|�VqdSr&)�isdigitrbrh)rjr_rrr rl�s�z'_parse_local_version.<locals>.<genexpr>)rf�_local_version_separatorsr[)rrrr rx�s
�rxcCs�tttt�dd�t|�����}|dur>|dur>|dur>t}n|durLt}n|}|dur^t}n|}|durpt}	n|}	|dur�t}
ntdd�|D��}
|||||	|
fS)NcSs|dkS)Nrr)r~rrr r/�r0z_cmpkey.<locals>.<lambda>css(|] }t|t�r|dfnt|fVqdS)rN)r;rhrrirrr rlsz_cmpkey.<locals>.<genexpr>)rf�reversed�list�	itertools�	dropwhilerr)rrrrrrr�r��_post�_dev�_localrrr ry�s(�	�ry)5�
__future__rrr�collectionsr�r��_structuresrrZ_typingr�typingr	r
rrr
rrrr�
InfiniteTypesr>rh�PrePostDevType�SubLocalType�	LocalType�CmpKey�LegacyCmpKey�bool�VersionComparisonMethod�__all__�
namedtuplerr�
ValueErrorr�objectr%rr�r�rZr\r`r@rrrwr�rxryrrrr �<module>sp$

���������
'F�	 &

site-packages/setuptools/_vendor/packaging/__pycache__/__about__.cpython-39.pyc000064400000001360150564745010023643 0ustar00a

(�Re��@sDddlmZmZmZgd�ZdZdZdZdZdZ	dZ
d	Zd
e	ZdS)�)�absolute_import�division�print_function)�	__title__�__summary__�__uri__�__version__�
__author__�	__email__�__license__�
__copyright__�	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz20.4z)Donald Stufft and individual contributorszdonald@stufft.iozBSD-2-Clause or Apache-2.0zCopyright 2014-2019 %sN)
�
__future__rrr�__all__rrrrr	r
rr�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/__about__.py�<module>ssite-packages/setuptools/_vendor/packaging/__pycache__/markers.cpython-39.pyc000064400000022212150564745010023400 0ustar00a

(�Re%%�	@s�ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZdd	lmZmZer�dd
lmZmZmZmZmZm Z m!Z!ee"e"ge#fZ$gd�Z%Gdd
�d
e&�Z'Gdd�de&�Z(Gdd�de&�Z)Gdd�de*�Z+Gdd�de+�Z,Gdd�de+�Z-Gdd�de+�Z.ed�ed�Bed�Bed�Bed�Bed�Bed �Bed!�Bed"�Bed#�Bed$�Bed%�Bed&�Bed'�Bed(�Bed)�Bed*�Bed+�BZ/d$d#dd ddd,�Z0e/�1d-d.��ed/�ed0�Bed1�Bed2�Bed3�Bed4�Bed5�Bed6�BZ2e2ed7�Bed8�BZ3e3�1d9d.��ed:�ed;�BZ4e4�1d<d.��ed=�ed>�BZ5e/e4BZ6ee6e3e6�Z7e7�1d?d.��ed@��8�Z9edA��8�Z:e�Z;e7ee9e;e:�BZ<e;e<e
e5e;�>ee;eZ=dBdC�Z>dXdEdF�Z?dGd.�dHd.�ej@ejAejBejCejDejEdI�ZFdJdK�ZGGdLdM�dMe*�ZHeH�ZIdNdO�ZJdPdQ�ZKdRdS�ZLdTdU�ZMGdVdW�dWe*�ZNdS)Y�)�absolute_import�division�print_functionN)�ParseException�ParseResults�stringStart�	stringEnd)�
ZeroOrMore�Group�Forward�QuotedString)�Literal�)�string_types)�
TYPE_CHECKING)�	Specifier�InvalidSpecifier)�Any�Callable�Dict�List�Optional�Tuple�Union)�
InvalidMarker�UndefinedComparison�UndefinedEnvironmentName�Marker�default_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N��__name__�
__module__�__qualname__�__doc__�r$r$��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/markers.pyr"src@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrr$r$r$r%r(src@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrr$r$r$r%r.src@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�NodecCs
||_dS�N)�value)�selfr(r$r$r%�__init__6sz
Node.__init__cCs
t|j�Sr')�strr(�r)r$r$r%�__str__:szNode.__str__cCsd�|jjt|��S)Nz<{0}({1!r})>)�format�	__class__r r+r,r$r$r%�__repr__>sz
Node.__repr__cCst�dSr')�NotImplementedErrorr,r$r$r%�	serializeBszNode.serializeN)r r!r"r*r-r0r2r$r$r$r%r&5sr&c@seZdZdd�ZdS)�VariablecCst|�Sr'�r+r,r$r$r%r2HszVariable.serializeN�r r!r"r2r$r$r$r%r3Gsr3c@seZdZdd�ZdS)�ValuecCs
d�|�S)Nz"{0}")r.r,r$r$r%r2NszValue.serializeNr5r$r$r$r%r6Msr6c@seZdZdd�ZdS)�OpcCst|�Sr'r4r,r$r$r%r2TszOp.serializeNr5r$r$r$r%r7Ssr7�implementation_version�platform_python_implementation�implementation_name�python_full_version�platform_release�platform_version�platform_machine�platform_system�python_version�sys_platform�os_name�os.name�sys.platform�platform.version�platform.machine�platform.python_implementation�python_implementation�extra)rCrDrErFrGrHcCstt�|d|d��S�Nr)r3�ALIASES�get��s�l�tr$r$r%�<lambda>u�rQz===�==�>=�<=�!=z~=�>�<�not in�incCst|d�SrJ)r7rMr$r$r%rQ|rR�'�"cCst|d�SrJ)r6rMr$r$r%rQrR�and�orcCst|d�SrJ)�tuplerMr$r$r%rQ�rR�(�)cCs t|t�rdd�|D�S|SdS)NcSsg|]}t|��qSr$)�_coerce_parse_result)�.0�ir$r$r%�
<listcomp>�rRz(_coerce_parse_result.<locals>.<listcomp>)�
isinstancer)�resultsr$r$r%rb�s
rbTcCs�t|tttf�sJ�t|t�rHt|�dkrHt|dttf�rHt|d�St|t�r�dd�|D�}|rnd�|�Sdd�|�dSn"t|t�r�d�dd	�|D��S|SdS)
Nrrcss|]}t|dd�VqdS)F)�firstN)�_format_marker�rc�mr$r$r%�	<genexpr>�rRz!_format_marker.<locals>.<genexpr>� r`racSsg|]}|���qSr$)r2rjr$r$r%re�rRz"_format_marker.<locals>.<listcomp>)rf�listr_r�lenri�join)�markerrh�innerr$r$r%ri�s �
��


ricCs||vSr'r$��lhs�rhsr$r$r%rQ�rRcCs||vSr'r$rsr$r$r%rQ�rR)rZrYrXrUrSrVrTrWcCsjztd�|��|g��}Wnty,Yn0|�|�St�|���}|dur`td�|||���|||�S)N�z#Undefined {0!r} on {1!r} and {2!r}.)	rrpr2r�contains�
_operatorsrLrr.)rt�opru�spec�operr$r$r%�_eval_op�s
�r|c@seZdZdS)�	UndefinedN)r r!r"r$r$r$r%r}�sr}cCs(|�|t�}t|t�r$td�|���|S)Nz/{0!r} does not exist in evaluation environment.)rL�
_undefinedrfr}rr.)�environment�namer(r$r$r%�_get_env�s
�r�c	Cs�gg}|D]�}t|tttf�s"J�t|t�rB|d�t||��q
t|t�r�|\}}}t|t�rtt||j�}|j}n|j}t||j�}|d�t	|||��q
|dvs�J�|dkr
|�g�q
t
dd�|D��S)N���)r]r^r^css|]}t|�VqdSr')�all)rc�itemr$r$r%rlrRz$_evaluate_markers.<locals>.<genexpr>)rfrnr_r�append�_evaluate_markersr3r�r(r|�any)	�markersr�groupsrqrtryru�	lhs_value�	rhs_valuer$r$r%r��s"



r�cCs2d�|�}|j}|dkr.||dt|j�7}|S)Nz{0.major}.{0.minor}.{0.micro}�finalr)r.�releaselevelr+�serial)�info�version�kindr$r$r%�format_full_versions

r�cCsrttd�r ttjj�}tjj}nd}d}||tjt��t�	�t�
�t��t��t��d�
t��dd��tjd�S)N�implementation�0rv�.�)r:r8rBr>r<r?r=r;r9r@rA)�hasattr�sysr�r�r�r��os�platform�machine�release�systemr@rHrp�python_version_tuple)�iverr:r$r$r%rs"

�rc@s.eZdZdd�Zdd�Zdd�Zd
dd	�ZdS)rc
Csbztt�|��|_WnHty\}z0d�|||j|jd��}t|��WYd}~n
d}~00dS)Nz+Invalid marker: {0!r}, parse error at {1!r}�)rb�MARKER�parseString�_markersrr.�locr)r)rq�eZerr_strr$r$r%r*(s�zMarker.__init__cCs
t|j�Sr')rir�r,r$r$r%r-2szMarker.__str__cCsd�t|��S)Nz<Marker({0!r})>)r.r+r,r$r$r%r06szMarker.__repr__NcCs$t�}|dur|�|�t|j|�S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)r�updater�r�)r)r�current_environmentr$r$r%�evaluate:s

zMarker.evaluate)N)r r!r"r*r-r0r�r$r$r$r%r's
r)T)O�
__future__rrr�operatorr�r�r�Zsetuptools.extern.pyparsingrrrrr	r
rrr
�LZ_compatrZ_typingr�
specifiersrr�typingrrrrrrrr+�bool�Operator�__all__�
ValueErrorrrr�objectr&r3r6r7�VARIABLErK�setParseAction�VERSION_CMP�	MARKER_OP�MARKER_VALUE�BOOLOP�
MARKER_VAR�MARKER_ITEM�suppress�LPAREN�RPAREN�MARKER_EXPR�MARKER_ATOMr�rbri�lt�le�eq�ne�ge�gtrxr|r}r~r�r�r�rrr$r$r$r%�<module>s�$	��������	�
���
�������>�
�	site-packages/setuptools/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc000064400000050231150564745010024072 0ustar00a

(�Re�|�@sXddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZddl
mZddlmZmZmZer�ddlmZmZmZmZmZmZmZmZmZeeefZeeeefZeeege fZ!Gd	d
�d
e"�Z#Gdd�de
ej$e%��Z&Gd
d�de&�Z'Gdd�de'�Z(dd�Z)Gdd�de'�Z*e�+d�Z,dd�Z-dd�Z.Gdd�de&�Z/dS)�)�absolute_import�division�print_functionN�)�string_types�with_metaclass)�
TYPE_CHECKING)�canonicalize_version)�Version�
LegacyVersion�parse)	�List�Dict�Union�Iterable�Iterator�Optional�Callable�Tuple�	FrozenSetc@seZdZdZdS)�InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/specifiers.pyr"src@s�eZdZejdd��Zejdd��Zejdd��Zejdd��Zej	d	d
��Z
e
jdd
��Z
ejdd
d��Zejddd��Z
dS)�
BaseSpecifiercCsdS)z�
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nr��selfrrr�__str__)szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr�__hash__1szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nr�r�otherrrr�__eq__8szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrr"rrr�__ne__@szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrr�prereleasesHszBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nr�r�valuerrrr&PsNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nr�r�itemr&rrr�containsXszBaseSpecifier.containscCsdS)z�
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)r�iterabler&rrr�filter_szBaseSpecifier.filter)N)N)rrr�abc�abstractmethodr r!r$r%�abstractpropertyr&�setterr+r-rrrrr(s 





rc@s�eZdZiZd"dd�Zdd�Zdd�Zed	d
��Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
edd��Zedd��Zedd��Zejdd��Zdd�Zd#dd�Zd$d d!�ZdS)%�_IndividualSpecifier�NcCsF|j�|�}|std�|���|�d���|�d���f|_||_dS)NzInvalid specifier: '{0}'�operator�version)�_regex�searchr�format�group�strip�_spec�_prereleases)r�specr&�matchrrr�__init__ls�z_IndividualSpecifier.__init__cCs0|jdurd�|j�nd}d�|jjt|�|�S)N�, prereleases={0!r}r3z<{0}({1!r}{2})>)r<r8r&�	__class__r�str�r�prerrr�__repr__zs
��z_IndividualSpecifier.__repr__cCsdj|j�S)Nz{0}{1})r8r;rrrrr �sz_IndividualSpecifier.__str__cCs|jdt|jd�fS)Nrr)r;r	rrrr�_canonical_spec�sz$_IndividualSpecifier._canonical_speccCs
t|j�S�N)�hashrFrrrrr!�sz_IndividualSpecifier.__hash__cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrG)�
isinstancerrArBr�NotImplementedrFr"rrrr$�s
z_IndividualSpecifier.__eq__cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrG)rIrrArBrrJr;r"rrrr%�s
z_IndividualSpecifier.__ne__cCst|d�|j|��}|S)Nz_compare_{0})�getattrr8�
_operators)r�op�operator_callablerrr�
_get_operator�s�z"_IndividualSpecifier._get_operatorcCst|ttf�st|�}|SrG)rIrr
r�rr5rrr�_coerce_version�sz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nr�r;rrrrr4�sz_IndividualSpecifier.operatorcCs
|jdS)NrrRrrrrr5�sz_IndividualSpecifier.versioncCs|jSrG�r<rrrrr&�sz _IndividualSpecifier.prereleasescCs
||_dSrGrSr'rrrr&�scCs
|�|�SrG�r+�rr*rrr�__contains__�sz!_IndividualSpecifier.__contains__cCs>|dur|j}|�|�}|jr&|s&dS|�|j�}|||j�S�NF)r&rQ�
is_prereleaserOr4r5)rr*r&�normalized_itemrNrrrr+�s

z_IndividualSpecifier.containsccs�d}g}d|dur|ndi}|D]F}|�|�}|j|fi|��r |jr\|s\|js\|�|�q d}|Vq |s�|r�|D]
}|VqtdS)NFr&T)rQr+rXr&�append)rr,r&�yielded�found_prereleases�kwr5�parsed_versionrrrr-�s"
��z_IndividualSpecifier.filter)r3N)N)N)rrrrLr?rEr �propertyrFr!r$r%rOrQr4r5r&r1rVr+r-rrrrr2hs,







r2c@sveZdZdZe�dedejejB�Zdddddd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�ZdS)�LegacySpecifiera�
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        �^\s*�\s*$�equal�	not_equal�less_than_equal�greater_than_equal�	less_than�greater_than)�==�!=�<=�>=�<�>cCst|t�stt|��}|SrG)rIrrBrPrrrrQ s
zLegacySpecifier._coerce_versioncCs||�|�kSrG�rQ�r�prospectiver=rrr�_compare_equal&szLegacySpecifier._compare_equalcCs||�|�kSrGrorprrr�_compare_not_equal*sz"LegacySpecifier._compare_not_equalcCs||�|�kSrGrorprrr�_compare_less_than_equal.sz(LegacySpecifier._compare_less_than_equalcCs||�|�kSrGrorprrr�_compare_greater_than_equal2sz+LegacySpecifier._compare_greater_than_equalcCs||�|�kSrGrorprrr�_compare_less_than6sz"LegacySpecifier._compare_less_thancCs||�|�kSrGrorprrr�_compare_greater_than:sz%LegacySpecifier._compare_greater_thanN)rrr�
_regex_str�re�compile�VERBOSE�
IGNORECASEr6rLrQrrrsrtrurvrwrrrrr`s �	r`cst����fdd��}|S)Ncst|t�sdS�|||�SrW)rIr
rp��fnrr�wrappedCs
z)_require_version_compare.<locals>.wrapped)�	functools�wraps)r~rrr}r�_require_version_compare?sr�c	@s�eZdZdZe�dedejejB�Zdddddd	d
dd�Z	e
d
d��Ze
dd��Ze
dd��Z
e
dd��Ze
dd��Ze
dd��Ze
dd��Zdd�Zedd��Zejdd��Zd S)!�	Specifiera
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        rarb�
compatiblercrdrerfrgrh�	arbitrary)�~=rirjrkrlrmrn�===cCsNd�tt�dd�t|���dd��}|d7}|�d�||�oL|�d�||�S)N�.cSs|�d�o|�d�S)N�post�dev)�
startswith��xrrr�<lambda>��z/Specifier._compare_compatible.<locals>.<lambda>����.*rlri)�join�list�	itertools�	takewhile�_version_splitrO)rrqr=�prefixrrr�_compare_compatible�s
����
�zSpecifier._compare_compatiblec	Csz|�d�rVt|j�}t|dd��}tt|��}|dt|��}t||�\}}||kSt|�}|jsnt|j�}||kSdS)Nr����)�endswithr
�publicr�rB�len�_pad_version�local)	rrqr=�
split_spec�split_prospective�shortened_prospective�padded_spec�padded_prospective�spec_versionrrrrr�s

�
zSpecifier._compare_equalcCs|�||�SrG)rrrprrrrs�szSpecifier._compare_not_equalcCst|j�t|�kSrG�r
r�rprrrrtsz"Specifier._compare_less_than_equalcCst|j�t|�kSrGr�rprrrru
sz%Specifier._compare_greater_than_equalcCs<t|�}||ksdS|js8|jr8t|j�t|j�kr8dSdS�NFT)r
rX�base_version�rrq�spec_strr=rrrrvszSpecifier._compare_less_thancCs^t|�}||ksdS|js8|jr8t|j�t|j�kr8dS|jdurZt|j�t|j�krZdSdSr�)r
�is_postreleaser�r�r�rrrrw1s
zSpecifier._compare_greater_thancCst|���t|���kSrG)rB�lowerrprrr�_compare_arbitraryRszSpecifier._compare_arbitrarycCsR|jdur|jS|j\}}|dvrN|dkr@|�d�r@|dd�}t|�jrNdSdS)N)rirlrkr�r�rir�r�TF)r<r;r�rrX)rr4r5rrrr&Vs


zSpecifier.prereleasescCs
||_dSrGrSr'rrrr&psN)rrrrxryrzr{r|r6rLr�r�rrrsrtrurvrwr�r_r&r1rrrrr�Ms<]�

(




 
r�z^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|�d�D],}t�|�}|r0|�|���q|�|�q|S)Nr�)�split�
_prefix_regexr7�extend�groupsrZ)r5�resultr*r>rrrr�ys
r�c
Cs�gg}}|�tt�dd�|���|�tt�dd�|���|�|t|d�d��|�|t|d�d��|�ddgtdt|d�t|d���|�ddgtdt|d�t|d���ttj|��ttj|��fS)NcSs|��SrG��isdigitr�rrrr��r�z_pad_version.<locals>.<lambda>cSs|��SrGr�r�rrrr��r�rr�0)rZr�r�r�r��insert�max�chain)�left�right�
left_split�right_splitrrrr��s
,,r�c@s�eZdZddd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Zdd�Z	dd�Z
dd�Zedd��Z
e
jdd��Z
dd�Zddd�Zd dd�ZdS)!�SpecifierSetr3Nc	Csldd�|�d�D�}t�}|D]8}z|�t|��WqtyT|�t|��Yq0qt|�|_||_dS)NcSsg|]}|��r|���qSr)r:��.0�srrr�
<listcomp>�r�z)SpecifierSet.__init__.<locals>.<listcomp>�,)	r��set�addr�rr`�	frozenset�_specsr<)r�
specifiersr&�split_specifiers�parsed�	specifierrrrr?�s
zSpecifierSet.__init__cCs*|jdurd�|j�nd}d�t|�|�S)Nr@r3z<SpecifierSet({0!r}{1})>)r<r8r&rBrCrrrrE�s
��zSpecifierSet.__repr__cCsd�tdd�|jD���S)Nr�css|]}t|�VqdSrG)rBr�rrr�	<genexpr>�r�z'SpecifierSet.__str__.<locals>.<genexpr>)r��sortedr�rrrrr �szSpecifierSet.__str__cCs
t|j�SrG)rHr�rrrrr!�szSpecifierSet.__hash__cCs�t|t�rt|�}nt|t�s"tSt�}t|j|jB�|_|jdurX|jdurX|j|_n<|jdurv|jdurv|j|_n|j|jkr�|j|_ntd��|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)rIrr�rJr�r�r<�
ValueError)rr#r�rrr�__and__�s 





�zSpecifierSet.__and__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrG�rIrr2r�rBrJr�r"rrrr$�s

zSpecifierSet.__eq__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrGr�r"rrrr%�s

zSpecifierSet.__ne__cCs
t|j�SrG)r�r�rrrr�__len__�szSpecifierSet.__len__cCs
t|j�SrG)�iterr�rrrr�__iter__�szSpecifierSet.__iter__cCs.|jdur|jS|jsdStdd�|jD��S)Ncss|]}|jVqdSrG�r&r�rrrr�r�z+SpecifierSet.prereleases.<locals>.<genexpr>)r<r��anyrrrrr&�s

zSpecifierSet.prereleasescCs
||_dSrGrSr'rrrr&scCs
|�|�SrGrTrUrrrrVszSpecifierSet.__contains__csLt�ttf�st����dur$|j��s2�jr2dSt��fdd�|jD��S)NFc3s|]}|j��d�VqdS)r�NrTr��r*r&rrr�*r�z(SpecifierSet.contains.<locals>.<genexpr>)rIrr
rr&rX�allr�r)rr�rr+s
zSpecifierSet.containscCs�|dur|j}|jr6|jD]}|j|t|�d�}q|Sg}g}|D]P}t|ttf�s^t|�}n|}t|t�rnqB|jr�|s�|s�|�	|�qB|�	|�qB|s�|r�|dur�|S|SdS)Nr�)
r&r�r-�boolrIrr
rrXrZ)rr,r&r=�filteredr\r*r^rrrr-,s*




zSpecifierSet.filter)r3N)N)N)rrrr?rEr r!r�r$r%r�r�r_r&r1rVr+r-rrrrr��s"

		


�r�)0�
__future__rrrr.r�r�ryZ_compatrrZ_typingr�utilsr	r5r
rr�typingr
rrrrrrrr�
ParsedVersionrB�UnparsedVersionr��CallableOperatorr�r�ABCMeta�objectrr2r`r�r�rzr�r�r�r�rrrr�<module>s4,@ 8+
site-packages/setuptools/_vendor/packaging/__pycache__/_typing.cpython-39.pyc000064400000003005150564745010023404 0ustar00a

(�Re�@s.dZddgZdZer"ddlmZndd�ZdS)a;For neatly implementing static typing in packaging.

`mypy` - the static type analysis tool we use - uses the `typing` module, which
provides core functionality fundamental to mypy's functioning.

Generally, `typing` would be imported at runtime and used in that fashion -
it acts as a no-op at runtime and does not have any run-time overhead by
design.

As it turns out, `typing` is not vendorable - it uses separate sources for
Python 2/Python 3. Thus, this codebase can not expect it to be present.
To work around this, mypy allows the typing import to be behind a False-y
optional to prevent it from running at runtime and type-comments can be used
to remove the need for the types to be accessible directly during runtime.

This module provides the False-y guard in a nicely named fashion so that a
curious maintainer can reach here to read this.

In packaging, all static-typing related imports should be guarded as follows:

    from packaging._typing import TYPE_CHECKING

    if TYPE_CHECKING:
        from typing import ...

Ref: https://github.com/python/mypy/issues/3216
�
TYPE_CHECKING�castF�)rcCs|S)N�)�type_�valuerr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/_typing.pyr/sN)�__doc__�__all__r�typingrrrrr�<module>s
site-packages/setuptools/_vendor/packaging/__pycache__/_compat.cpython-39.pyc000064400000002254150564745010023362 0ustar00a

(�Reh�@s~ddlmZmZmZddlZddlmZerDddlmZm	Z	m
Z
mZejddkZ
ejddkZerlefZnefZdd	�ZdS)
�)�absolute_import�division�print_functionN�)�
TYPE_CHECKING)�Any�Dict�Tuple�Type��cs&G��fdd�d��}t�|ddi�S)z/
    Create a base class with a metaclass.
    cseZdZ��fdd�ZdS)z!with_metaclass.<locals>.metaclasscs�|�|�S)N�)�cls�name�
this_bases�d��bases�metar
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/_compat.py�__new__"sz)with_metaclass.<locals>.metaclass.__new__N)�__name__�
__module__�__qualname__rr
rr
r�	metaclass!sr�temporary_classr
)�typer)rrrr
rr�with_metaclasssr)�
__future__rrr�sysZ_typingr�typingrrr	r
�version_info�PY2�PY3�str�string_types�
basestringrr
r
r
r�<module>ssite-packages/setuptools/_vendor/packaging/__pycache__/__init__.cpython-39.pyc000064400000001126150564745010023474 0ustar00a

(�Re2�@sHddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZgd�Z
dS)�)�absolute_import�division�print_function�)�
__author__�
__copyright__�	__email__�__license__�__summary__�	__title__�__uri__�__version__)rr
rr
rrr	rN)�
__future__rrr�	__about__rrrr	r
rrr
�__all__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/packaging/__init__.py�<module>s(site-packages/setuptools/_vendor/packaging/markers.py000064400000022445150564745010017121 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import operator
import os
import platform
import sys

from setuptools.extern.pyparsing import ParseException, ParseResults, stringStart, stringEnd
from setuptools.extern.pyparsing import ZeroOrMore, Group, Forward, QuotedString
from setuptools.extern.pyparsing import Literal as L  # noqa

from ._compat import string_types
from ._typing import TYPE_CHECKING
from .specifiers import Specifier, InvalidSpecifier

if TYPE_CHECKING:  # pragma: no cover
    from typing import Any, Callable, Dict, List, Optional, Tuple, Union

    Operator = Callable[[str, str], bool]


__all__ = [
    "InvalidMarker",
    "UndefinedComparison",
    "UndefinedEnvironmentName",
    "Marker",
    "default_environment",
]


class InvalidMarker(ValueError):
    """
    An invalid marker was found, users should refer to PEP 508.
    """


class UndefinedComparison(ValueError):
    """
    An invalid operation was attempted on a value that doesn't support it.
    """


class UndefinedEnvironmentName(ValueError):
    """
    A name was attempted to be used that does not exist inside of the
    environment.
    """


class Node(object):
    def __init__(self, value):
        # type: (Any) -> None
        self.value = value

    def __str__(self):
        # type: () -> str
        return str(self.value)

    def __repr__(self):
        # type: () -> str
        return "<{0}({1!r})>".format(self.__class__.__name__, str(self))

    def serialize(self):
        # type: () -> str
        raise NotImplementedError


class Variable(Node):
    def serialize(self):
        # type: () -> str
        return str(self)


class Value(Node):
    def serialize(self):
        # type: () -> str
        return '"{0}"'.format(self)


class Op(Node):
    def serialize(self):
        # type: () -> str
        return str(self)


VARIABLE = (
    L("implementation_version")
    | L("platform_python_implementation")
    | L("implementation_name")
    | L("python_full_version")
    | L("platform_release")
    | L("platform_version")
    | L("platform_machine")
    | L("platform_system")
    | L("python_version")
    | L("sys_platform")
    | L("os_name")
    | L("os.name")  # PEP-345
    | L("sys.platform")  # PEP-345
    | L("platform.version")  # PEP-345
    | L("platform.machine")  # PEP-345
    | L("platform.python_implementation")  # PEP-345
    | L("python_implementation")  # undocumented setuptools legacy
    | L("extra")  # PEP-508
)
ALIASES = {
    "os.name": "os_name",
    "sys.platform": "sys_platform",
    "platform.version": "platform_version",
    "platform.machine": "platform_machine",
    "platform.python_implementation": "platform_python_implementation",
    "python_implementation": "platform_python_implementation",
}
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))

VERSION_CMP = (
    L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<")
)

MARKER_OP = VERSION_CMP | L("not in") | L("in")
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))

MARKER_VALUE = QuotedString("'") | QuotedString('"')
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))

BOOLOP = L("and") | L("or")

MARKER_VAR = VARIABLE | MARKER_VALUE

MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))

LPAREN = L("(").suppress()
RPAREN = L(")").suppress()

MARKER_EXPR = Forward()
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)

MARKER = stringStart + MARKER_EXPR + stringEnd


def _coerce_parse_result(results):
    # type: (Union[ParseResults, List[Any]]) -> List[Any]
    if isinstance(results, ParseResults):
        return [_coerce_parse_result(i) for i in results]
    else:
        return results


def _format_marker(marker, first=True):
    # type: (Union[List[str], Tuple[Node, ...], str], Optional[bool]) -> str

    assert isinstance(marker, (list, tuple, string_types))

    # Sometimes we have a structure like [[...]] which is a single item list
    # where the single item is itself it's own list. In that case we want skip
    # the rest of this function so that we don't get extraneous () on the
    # outside.
    if (
        isinstance(marker, list)
        and len(marker) == 1
        and isinstance(marker[0], (list, tuple))
    ):
        return _format_marker(marker[0])

    if isinstance(marker, list):
        inner = (_format_marker(m, first=False) for m in marker)
        if first:
            return " ".join(inner)
        else:
            return "(" + " ".join(inner) + ")"
    elif isinstance(marker, tuple):
        return " ".join([m.serialize() for m in marker])
    else:
        return marker


_operators = {
    "in": lambda lhs, rhs: lhs in rhs,
    "not in": lambda lhs, rhs: lhs not in rhs,
    "<": operator.lt,
    "<=": operator.le,
    "==": operator.eq,
    "!=": operator.ne,
    ">=": operator.ge,
    ">": operator.gt,
}  # type: Dict[str, Operator]


def _eval_op(lhs, op, rhs):
    # type: (str, Op, str) -> bool
    try:
        spec = Specifier("".join([op.serialize(), rhs]))
    except InvalidSpecifier:
        pass
    else:
        return spec.contains(lhs)

    oper = _operators.get(op.serialize())  # type: Optional[Operator]
    if oper is None:
        raise UndefinedComparison(
            "Undefined {0!r} on {1!r} and {2!r}.".format(op, lhs, rhs)
        )

    return oper(lhs, rhs)


class Undefined(object):
    pass


_undefined = Undefined()


def _get_env(environment, name):
    # type: (Dict[str, str], str) -> str
    value = environment.get(name, _undefined)  # type: Union[str, Undefined]

    if isinstance(value, Undefined):
        raise UndefinedEnvironmentName(
            "{0!r} does not exist in evaluation environment.".format(name)
        )

    return value


def _evaluate_markers(markers, environment):
    # type: (List[Any], Dict[str, str]) -> bool
    groups = [[]]  # type: List[List[bool]]

    for marker in markers:
        assert isinstance(marker, (list, tuple, string_types))

        if isinstance(marker, list):
            groups[-1].append(_evaluate_markers(marker, environment))
        elif isinstance(marker, tuple):
            lhs, op, rhs = marker

            if isinstance(lhs, Variable):
                lhs_value = _get_env(environment, lhs.value)
                rhs_value = rhs.value
            else:
                lhs_value = lhs.value
                rhs_value = _get_env(environment, rhs.value)

            groups[-1].append(_eval_op(lhs_value, op, rhs_value))
        else:
            assert marker in ["and", "or"]
            if marker == "or":
                groups.append([])

    return any(all(item) for item in groups)


def format_full_version(info):
    # type: (sys._version_info) -> str
    version = "{0.major}.{0.minor}.{0.micro}".format(info)
    kind = info.releaselevel
    if kind != "final":
        version += kind[0] + str(info.serial)
    return version


def default_environment():
    # type: () -> Dict[str, str]
    if hasattr(sys, "implementation"):
        # Ignoring the `sys.implementation` reference for type checking due to
        # mypy not liking that the attribute doesn't exist in Python 2.7 when
        # run with the `--py27` flag.
        iver = format_full_version(sys.implementation.version)  # type: ignore
        implementation_name = sys.implementation.name  # type: ignore
    else:
        iver = "0"
        implementation_name = ""

    return {
        "implementation_name": implementation_name,
        "implementation_version": iver,
        "os_name": os.name,
        "platform_machine": platform.machine(),
        "platform_release": platform.release(),
        "platform_system": platform.system(),
        "platform_version": platform.version(),
        "python_full_version": platform.python_version(),
        "platform_python_implementation": platform.python_implementation(),
        "python_version": ".".join(platform.python_version_tuple()[:2]),
        "sys_platform": sys.platform,
    }


class Marker(object):
    def __init__(self, marker):
        # type: (str) -> None
        try:
            self._markers = _coerce_parse_result(MARKER.parseString(marker))
        except ParseException as e:
            err_str = "Invalid marker: {0!r}, parse error at {1!r}".format(
                marker, marker[e.loc : e.loc + 8]
            )
            raise InvalidMarker(err_str)

    def __str__(self):
        # type: () -> str
        return _format_marker(self._markers)

    def __repr__(self):
        # type: () -> str
        return "<Marker({0!r})>".format(str(self))

    def evaluate(self, environment=None):
        # type: (Optional[Dict[str, str]]) -> bool
        """Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        """
        current_environment = default_environment()
        if environment is not None:
            current_environment.update(environment)

        return _evaluate_markers(self._markers, current_environment)
site-packages/setuptools/_vendor/packaging/__about__.py000064400000001340150564745010017352 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]

__title__ = "packaging"
__summary__ = "Core utilities for Python packages"
__uri__ = "https://github.com/pypa/packaging"

__version__ = "20.4"

__author__ = "Donald Stufft and individual contributors"
__email__ = "donald@stufft.io"

__license__ = "BSD-2-Clause or Apache-2.0"
__copyright__ = "Copyright 2014-2019 %s" % __author__
site-packages/setuptools/_vendor/packaging/utils.py000064400000003423150564745010016610 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import re

from ._typing import TYPE_CHECKING, cast
from .version import InvalidVersion, Version

if TYPE_CHECKING:  # pragma: no cover
    from typing import NewType, Union

    NormalizedName = NewType("NormalizedName", str)

_canonicalize_regex = re.compile(r"[-_.]+")


def canonicalize_name(name):
    # type: (str) -> NormalizedName
    # This is taken from PEP 503.
    value = _canonicalize_regex.sub("-", name).lower()
    return cast("NormalizedName", value)


def canonicalize_version(_version):
    # type: (str) -> Union[Version, str]
    """
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    """

    try:
        version = Version(_version)
    except InvalidVersion:
        # Legacy versions cannot be normalized
        return _version

    parts = []

    # Epoch
    if version.epoch != 0:
        parts.append("{0}!".format(version.epoch))

    # Release segment
    # NB: This strips trailing '.0's to normalize
    parts.append(re.sub(r"(\.0)+$", "", ".".join(str(x) for x in version.release)))

    # Pre-release
    if version.pre is not None:
        parts.append("".join(str(x) for x in version.pre))

    # Post-release
    if version.post is not None:
        parts.append(".post{0}".format(version.post))

    # Development release
    if version.dev is not None:
        parts.append(".dev{0}".format(version.dev))

    # Local version segment
    if version.local is not None:
        parts.append("+{0}".format(version.local))

    return "".join(parts)
site-packages/setuptools/_vendor/packaging/_compat.py000064400000002150150564745010017066 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import sys

from ._typing import TYPE_CHECKING

if TYPE_CHECKING:  # pragma: no cover
    from typing import Any, Dict, Tuple, Type


PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3

# flake8: noqa

if PY3:
    string_types = (str,)
else:
    string_types = (basestring,)


def with_metaclass(meta, *bases):
    # type: (Type[Any], Tuple[Type[Any], ...]) -> Any
    """
    Create a base class with a metaclass.
    """
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(meta):  # type: ignore
        def __new__(cls, name, this_bases, d):
            # type: (Type[Any], str, Tuple[Any], Dict[Any, Any]) -> Any
            return meta(name, bases, d)

    return type.__new__(metaclass, "temporary_class", (), {})
site-packages/setuptools/_vendor/packaging/version.py000064400000036156150564745010017146 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import collections
import itertools
import re

from ._structures import Infinity, NegativeInfinity
from ._typing import TYPE_CHECKING

if TYPE_CHECKING:  # pragma: no cover
    from typing import Callable, Iterator, List, Optional, SupportsInt, Tuple, Union

    from ._structures import InfinityType, NegativeInfinityType

    InfiniteTypes = Union[InfinityType, NegativeInfinityType]
    PrePostDevType = Union[InfiniteTypes, Tuple[str, int]]
    SubLocalType = Union[InfiniteTypes, int, str]
    LocalType = Union[
        NegativeInfinityType,
        Tuple[
            Union[
                SubLocalType,
                Tuple[SubLocalType, str],
                Tuple[NegativeInfinityType, SubLocalType],
            ],
            ...,
        ],
    ]
    CmpKey = Tuple[
        int, Tuple[int, ...], PrePostDevType, PrePostDevType, PrePostDevType, LocalType
    ]
    LegacyCmpKey = Tuple[int, Tuple[str, ...]]
    VersionComparisonMethod = Callable[
        [Union[CmpKey, LegacyCmpKey], Union[CmpKey, LegacyCmpKey]], bool
    ]

__all__ = ["parse", "Version", "LegacyVersion", "InvalidVersion", "VERSION_PATTERN"]


_Version = collections.namedtuple(
    "_Version", ["epoch", "release", "dev", "pre", "post", "local"]
)


def parse(version):
    # type: (str) -> Union[LegacyVersion, Version]
    """
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    """
    try:
        return Version(version)
    except InvalidVersion:
        return LegacyVersion(version)


class InvalidVersion(ValueError):
    """
    An invalid version was found, users should refer to PEP 440.
    """


class _BaseVersion(object):
    _key = None  # type: Union[CmpKey, LegacyCmpKey]

    def __hash__(self):
        # type: () -> int
        return hash(self._key)

    def __lt__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s < o)

    def __le__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s <= o)

    def __eq__(self, other):
        # type: (object) -> bool
        return self._compare(other, lambda s, o: s == o)

    def __ge__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s >= o)

    def __gt__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s > o)

    def __ne__(self, other):
        # type: (object) -> bool
        return self._compare(other, lambda s, o: s != o)

    def _compare(self, other, method):
        # type: (object, VersionComparisonMethod) -> Union[bool, NotImplemented]
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return method(self._key, other._key)


class LegacyVersion(_BaseVersion):
    def __init__(self, version):
        # type: (str) -> None
        self._version = str(version)
        self._key = _legacy_cmpkey(self._version)

    def __str__(self):
        # type: () -> str
        return self._version

    def __repr__(self):
        # type: () -> str
        return "<LegacyVersion({0})>".format(repr(str(self)))

    @property
    def public(self):
        # type: () -> str
        return self._version

    @property
    def base_version(self):
        # type: () -> str
        return self._version

    @property
    def epoch(self):
        # type: () -> int
        return -1

    @property
    def release(self):
        # type: () -> None
        return None

    @property
    def pre(self):
        # type: () -> None
        return None

    @property
    def post(self):
        # type: () -> None
        return None

    @property
    def dev(self):
        # type: () -> None
        return None

    @property
    def local(self):
        # type: () -> None
        return None

    @property
    def is_prerelease(self):
        # type: () -> bool
        return False

    @property
    def is_postrelease(self):
        # type: () -> bool
        return False

    @property
    def is_devrelease(self):
        # type: () -> bool
        return False


_legacy_version_component_re = re.compile(r"(\d+ | [a-z]+ | \.| -)", re.VERBOSE)

_legacy_version_replacement_map = {
    "pre": "c",
    "preview": "c",
    "-": "final-",
    "rc": "c",
    "dev": "@",
}


def _parse_version_parts(s):
    # type: (str) -> Iterator[str]
    for part in _legacy_version_component_re.split(s):
        part = _legacy_version_replacement_map.get(part, part)

        if not part or part == ".":
            continue

        if part[:1] in "0123456789":
            # pad for numeric comparison
            yield part.zfill(8)
        else:
            yield "*" + part

    # ensure that alpha/beta/candidate are before final
    yield "*final"


def _legacy_cmpkey(version):
    # type: (str) -> LegacyCmpKey

    # We hardcode an epoch of -1 here. A PEP 440 version can only have a epoch
    # greater than or equal to 0. This will effectively put the LegacyVersion,
    # which uses the defacto standard originally implemented by setuptools,
    # as before all PEP 440 versions.
    epoch = -1

    # This scheme is taken from pkg_resources.parse_version setuptools prior to
    # it's adoption of the packaging library.
    parts = []  # type: List[str]
    for part in _parse_version_parts(version.lower()):
        if part.startswith("*"):
            # remove "-" before a prerelease tag
            if part < "*final":
                while parts and parts[-1] == "*final-":
                    parts.pop()

            # remove trailing zeros from each series of numeric parts
            while parts and parts[-1] == "00000000":
                parts.pop()

        parts.append(part)

    return epoch, tuple(parts)


# Deliberately not anchored to the start and end of the string, to make it
# easier for 3rd party code to reuse
VERSION_PATTERN = r"""
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
"""


class Version(_BaseVersion):

    _regex = re.compile(r"^\s*" + VERSION_PATTERN + r"\s*$", re.VERBOSE | re.IGNORECASE)

    def __init__(self, version):
        # type: (str) -> None

        # Validate the version and parse it into pieces
        match = self._regex.search(version)
        if not match:
            raise InvalidVersion("Invalid version: '{0}'".format(version))

        # Store the parsed out pieces of the version
        self._version = _Version(
            epoch=int(match.group("epoch")) if match.group("epoch") else 0,
            release=tuple(int(i) for i in match.group("release").split(".")),
            pre=_parse_letter_version(match.group("pre_l"), match.group("pre_n")),
            post=_parse_letter_version(
                match.group("post_l"), match.group("post_n1") or match.group("post_n2")
            ),
            dev=_parse_letter_version(match.group("dev_l"), match.group("dev_n")),
            local=_parse_local_version(match.group("local")),
        )

        # Generate a key which will be used for sorting
        self._key = _cmpkey(
            self._version.epoch,
            self._version.release,
            self._version.pre,
            self._version.post,
            self._version.dev,
            self._version.local,
        )

    def __repr__(self):
        # type: () -> str
        return "<Version({0})>".format(repr(str(self)))

    def __str__(self):
        # type: () -> str
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        # Pre-release
        if self.pre is not None:
            parts.append("".join(str(x) for x in self.pre))

        # Post-release
        if self.post is not None:
            parts.append(".post{0}".format(self.post))

        # Development release
        if self.dev is not None:
            parts.append(".dev{0}".format(self.dev))

        # Local version segment
        if self.local is not None:
            parts.append("+{0}".format(self.local))

        return "".join(parts)

    @property
    def epoch(self):
        # type: () -> int
        _epoch = self._version.epoch  # type: int
        return _epoch

    @property
    def release(self):
        # type: () -> Tuple[int, ...]
        _release = self._version.release  # type: Tuple[int, ...]
        return _release

    @property
    def pre(self):
        # type: () -> Optional[Tuple[str, int]]
        _pre = self._version.pre  # type: Optional[Tuple[str, int]]
        return _pre

    @property
    def post(self):
        # type: () -> Optional[Tuple[str, int]]
        return self._version.post[1] if self._version.post else None

    @property
    def dev(self):
        # type: () -> Optional[Tuple[str, int]]
        return self._version.dev[1] if self._version.dev else None

    @property
    def local(self):
        # type: () -> Optional[str]
        if self._version.local:
            return ".".join(str(x) for x in self._version.local)
        else:
            return None

    @property
    def public(self):
        # type: () -> str
        return str(self).split("+", 1)[0]

    @property
    def base_version(self):
        # type: () -> str
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        return "".join(parts)

    @property
    def is_prerelease(self):
        # type: () -> bool
        return self.dev is not None or self.pre is not None

    @property
    def is_postrelease(self):
        # type: () -> bool
        return self.post is not None

    @property
    def is_devrelease(self):
        # type: () -> bool
        return self.dev is not None

    @property
    def major(self):
        # type: () -> int
        return self.release[0] if len(self.release) >= 1 else 0

    @property
    def minor(self):
        # type: () -> int
        return self.release[1] if len(self.release) >= 2 else 0

    @property
    def micro(self):
        # type: () -> int
        return self.release[2] if len(self.release) >= 3 else 0


def _parse_letter_version(
    letter,  # type: str
    number,  # type: Union[str, bytes, SupportsInt]
):
    # type: (...) -> Optional[Tuple[str, int]]

    if letter:
        # We consider there to be an implicit 0 in a pre-release if there is
        # not a numeral associated with it.
        if number is None:
            number = 0

        # We normalize any letters to their lower case form
        letter = letter.lower()

        # We consider some words to be alternate spellings of other words and
        # in those cases we want to normalize the spellings to our preferred
        # spelling.
        if letter == "alpha":
            letter = "a"
        elif letter == "beta":
            letter = "b"
        elif letter in ["c", "pre", "preview"]:
            letter = "rc"
        elif letter in ["rev", "r"]:
            letter = "post"

        return letter, int(number)
    if not letter and number:
        # We assume if we are given a number, but we are not given a letter
        # then this is using the implicit post release syntax (e.g. 1.0-1)
        letter = "post"

        return letter, int(number)

    return None


_local_version_separators = re.compile(r"[\._-]")


def _parse_local_version(local):
    # type: (str) -> Optional[LocalType]
    """
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    """
    if local is not None:
        return tuple(
            part.lower() if not part.isdigit() else int(part)
            for part in _local_version_separators.split(local)
        )
    return None


def _cmpkey(
    epoch,  # type: int
    release,  # type: Tuple[int, ...]
    pre,  # type: Optional[Tuple[str, int]]
    post,  # type: Optional[Tuple[str, int]]
    dev,  # type: Optional[Tuple[str, int]]
    local,  # type: Optional[Tuple[SubLocalType]]
):
    # type: (...) -> CmpKey

    # When we compare a release version, we want to compare it with all of the
    # trailing zeros removed. So we'll use a reverse the list, drop all the now
    # leading zeros until we come to something non zero, then take the rest
    # re-reverse it back into the correct order and make it a tuple and use
    # that for our sorting key.
    _release = tuple(
        reversed(list(itertools.dropwhile(lambda x: x == 0, reversed(release))))
    )

    # We need to "trick" the sorting algorithm to put 1.0.dev0 before 1.0a0.
    # We'll do this by abusing the pre segment, but we _only_ want to do this
    # if there is not a pre or a post segment. If we have one of those then
    # the normal sorting rules will handle this case correctly.
    if pre is None and post is None and dev is not None:
        _pre = NegativeInfinity  # type: PrePostDevType
    # Versions without a pre-release (except as noted above) should sort after
    # those with one.
    elif pre is None:
        _pre = Infinity
    else:
        _pre = pre

    # Versions without a post segment should sort before those with one.
    if post is None:
        _post = NegativeInfinity  # type: PrePostDevType

    else:
        _post = post

    # Versions without a development segment should sort after those with one.
    if dev is None:
        _dev = Infinity  # type: PrePostDevType

    else:
        _dev = dev

    if local is None:
        # Versions without a local segment should sort before those with one.
        _local = NegativeInfinity  # type: LocalType
    else:
        # Versions with a local segment need that segment parsed to implement
        # the sorting rules in PEP440.
        # - Alpha numeric segments sort before numeric segments
        # - Alpha numeric segments sort lexicographically
        # - Numeric segments sort numerically
        # - Shorter versions sort before longer versions when the prefixes
        #   match exactly
        _local = tuple(
            (i, "") if isinstance(i, int) else (NegativeInfinity, i) for i in local
        )

    return epoch, _release, _pre, _post, _dev, _local
site-packages/setuptools/_vendor/packaging/__init__.py000064400000001062150564745010017204 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

from .__about__ import (
    __author__,
    __copyright__,
    __email__,
    __license__,
    __summary__,
    __title__,
    __uri__,
    __version__,
)

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]
site-packages/setuptools/_vendor/packaging/specifiers.py000064400000076310150564745010017611 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import abc
import functools
import itertools
import re

from ._compat import string_types, with_metaclass
from ._typing import TYPE_CHECKING
from .utils import canonicalize_version
from .version import Version, LegacyVersion, parse

if TYPE_CHECKING:  # pragma: no cover
    from typing import (
        List,
        Dict,
        Union,
        Iterable,
        Iterator,
        Optional,
        Callable,
        Tuple,
        FrozenSet,
    )

    ParsedVersion = Union[Version, LegacyVersion]
    UnparsedVersion = Union[Version, LegacyVersion, str]
    CallableOperator = Callable[[ParsedVersion, str], bool]


class InvalidSpecifier(ValueError):
    """
    An invalid specifier was found, users should refer to PEP 440.
    """


class BaseSpecifier(with_metaclass(abc.ABCMeta, object)):  # type: ignore
    @abc.abstractmethod
    def __str__(self):
        # type: () -> str
        """
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        """

    @abc.abstractmethod
    def __hash__(self):
        # type: () -> int
        """
        Returns a hash value for this Specifier like object.
        """

    @abc.abstractmethod
    def __eq__(self, other):
        # type: (object) -> bool
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        """

    @abc.abstractmethod
    def __ne__(self, other):
        # type: (object) -> bool
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        """

    @abc.abstractproperty
    def prereleases(self):
        # type: () -> Optional[bool]
        """
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        """
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @abc.abstractmethod
    def contains(self, item, prereleases=None):
        # type: (str, Optional[bool]) -> bool
        """
        Determines if the given item is contained within this specifier.
        """

    @abc.abstractmethod
    def filter(self, iterable, prereleases=None):
        # type: (Iterable[UnparsedVersion], Optional[bool]) -> Iterable[UnparsedVersion]
        """
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        """


class _IndividualSpecifier(BaseSpecifier):

    _operators = {}  # type: Dict[str, str]

    def __init__(self, spec="", prereleases=None):
        # type: (str, Optional[bool]) -> None
        match = self._regex.search(spec)
        if not match:
            raise InvalidSpecifier("Invalid specifier: '{0}'".format(spec))

        self._spec = (
            match.group("operator").strip(),
            match.group("version").strip(),
        )  # type: Tuple[str, str]

        # Store whether or not this Specifier should accept prereleases
        self._prereleases = prereleases

    def __repr__(self):
        # type: () -> str
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<{0}({1!r}{2})>".format(self.__class__.__name__, str(self), pre)

    def __str__(self):
        # type: () -> str
        return "{0}{1}".format(*self._spec)

    @property
    def _canonical_spec(self):
        # type: () -> Tuple[str, Union[Version, str]]
        return self._spec[0], canonicalize_version(self._spec[1])

    def __hash__(self):
        # type: () -> int
        return hash(self._canonical_spec)

    def __eq__(self, other):
        # type: (object) -> bool
        if isinstance(other, string_types):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._canonical_spec == other._canonical_spec

    def __ne__(self, other):
        # type: (object) -> bool
        if isinstance(other, string_types):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec != other._spec

    def _get_operator(self, op):
        # type: (str) -> CallableOperator
        operator_callable = getattr(
            self, "_compare_{0}".format(self._operators[op])
        )  # type: CallableOperator
        return operator_callable

    def _coerce_version(self, version):
        # type: (UnparsedVersion) -> ParsedVersion
        if not isinstance(version, (LegacyVersion, Version)):
            version = parse(version)
        return version

    @property
    def operator(self):
        # type: () -> str
        return self._spec[0]

    @property
    def version(self):
        # type: () -> str
        return self._spec[1]

    @property
    def prereleases(self):
        # type: () -> Optional[bool]
        return self._prereleases

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value

    def __contains__(self, item):
        # type: (str) -> bool
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # type: (UnparsedVersion, Optional[bool]) -> bool

        # Determine if prereleases are to be allowed or not.
        if prereleases is None:
            prereleases = self.prereleases

        # Normalize item to a Version or LegacyVersion, this allows us to have
        # a shortcut for ``"2.0" in Specifier(">=2")
        normalized_item = self._coerce_version(item)

        # Determine if we should be supporting prereleases in this specifier
        # or not, if we do not support prereleases than we can short circuit
        # logic if this version is a prereleases.
        if normalized_item.is_prerelease and not prereleases:
            return False

        # Actually do the comparison to determine if this item is contained
        # within this Specifier or not.
        operator_callable = self._get_operator(self.operator)  # type: CallableOperator
        return operator_callable(normalized_item, self.version)

    def filter(self, iterable, prereleases=None):
        # type: (Iterable[UnparsedVersion], Optional[bool]) -> Iterable[UnparsedVersion]

        yielded = False
        found_prereleases = []

        kw = {"prereleases": prereleases if prereleases is not None else True}

        # Attempt to iterate over all the values in the iterable and if any of
        # them match, yield them.
        for version in iterable:
            parsed_version = self._coerce_version(version)

            if self.contains(parsed_version, **kw):
                # If our version is a prerelease, and we were not set to allow
                # prereleases, then we'll store it for later incase nothing
                # else matches this specifier.
                if parsed_version.is_prerelease and not (
                    prereleases or self.prereleases
                ):
                    found_prereleases.append(version)
                # Either this is not a prerelease, or we should have been
                # accepting prereleases from the beginning.
                else:
                    yielded = True
                    yield version

        # Now that we've iterated over everything, determine if we've yielded
        # any values, and if we have not and we have any prereleases stored up
        # then we will go ahead and yield the prereleases.
        if not yielded and found_prereleases:
            for version in found_prereleases:
                yield version


class LegacySpecifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
    }

    def _coerce_version(self, version):
        # type: (Union[ParsedVersion, str]) -> LegacyVersion
        if not isinstance(version, LegacyVersion):
            version = LegacyVersion(str(version))
        return version

    def _compare_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective == self._coerce_version(spec)

    def _compare_not_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective != self._coerce_version(spec)

    def _compare_less_than_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective <= self._coerce_version(spec)

    def _compare_greater_than_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective >= self._coerce_version(spec)

    def _compare_less_than(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective < self._coerce_version(spec)

    def _compare_greater_than(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective > self._coerce_version(spec)


def _require_version_compare(
    fn  # type: (Callable[[Specifier, ParsedVersion, str], bool])
):
    # type: (...) -> Callable[[Specifier, ParsedVersion, str], bool]
    @functools.wraps(fn)
    def wrapped(self, prospective, spec):
        # type: (Specifier, ParsedVersion, str) -> bool
        if not isinstance(prospective, Version):
            return False
        return fn(self, prospective, spec)

    return wrapped


class Specifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "~=": "compatible",
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
        "===": "arbitrary",
    }

    @_require_version_compare
    def _compare_compatible(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # Compatible releases have an equivalent combination of >= and ==. That
        # is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
        # implement this in terms of the other specifiers instead of
        # implementing it ourselves. The only thing we need to do is construct
        # the other specifiers.

        # We want everything but the last item in the version, but we want to
        # ignore post and dev releases and we want to treat the pre-release as
        # it's own separate segment.
        prefix = ".".join(
            list(
                itertools.takewhile(
                    lambda x: (not x.startswith("post") and not x.startswith("dev")),
                    _version_split(spec),
                )
            )[:-1]
        )

        # Add the prefix notation to the end of our string
        prefix += ".*"

        return self._get_operator(">=")(prospective, spec) and self._get_operator("==")(
            prospective, prefix
        )

    @_require_version_compare
    def _compare_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # We need special logic to handle prefix matching
        if spec.endswith(".*"):
            # In the case of prefix matching we want to ignore local segment.
            prospective = Version(prospective.public)
            # Split the spec out by dots, and pretend that there is an implicit
            # dot in between a release segment and a pre-release segment.
            split_spec = _version_split(spec[:-2])  # Remove the trailing .*

            # Split the prospective version out by dots, and pretend that there
            # is an implicit dot in between a release segment and a pre-release
            # segment.
            split_prospective = _version_split(str(prospective))

            # Shorten the prospective version to be the same length as the spec
            # so that we can determine if the specifier is a prefix of the
            # prospective version or not.
            shortened_prospective = split_prospective[: len(split_spec)]

            # Pad out our two sides with zeros so that they both equal the same
            # length.
            padded_spec, padded_prospective = _pad_version(
                split_spec, shortened_prospective
            )

            return padded_prospective == padded_spec
        else:
            # Convert our spec string into a Version
            spec_version = Version(spec)

            # If the specifier does not have a local segment, then we want to
            # act as if the prospective version also does not have a local
            # segment.
            if not spec_version.local:
                prospective = Version(prospective.public)

            return prospective == spec_version

    @_require_version_compare
    def _compare_not_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool
        return not self._compare_equal(prospective, spec)

    @_require_version_compare
    def _compare_less_than_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) <= Version(spec)

    @_require_version_compare
    def _compare_greater_than_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) >= Version(spec)

    @_require_version_compare
    def _compare_less_than(self, prospective, spec_str):
        # type: (ParsedVersion, str) -> bool

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is less than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective < spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a pre-release version, that we do not accept pre-release
        # versions for the version mentioned in the specifier (e.g. <3.1 should
        # not match 3.1.dev0, but should match 3.0.dev0).
        if not spec.is_prerelease and prospective.is_prerelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # less than the spec version *and* it's not a pre-release of the same
        # version in the spec.
        return True

    @_require_version_compare
    def _compare_greater_than(self, prospective, spec_str):
        # type: (ParsedVersion, str) -> bool

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is greater than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective > spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a post-release version, that we do not accept
        # post-release versions for the version mentioned in the specifier
        # (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
        if not spec.is_postrelease and prospective.is_postrelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # Ensure that we do not allow a local version of the version mentioned
        # in the specifier, which is technically greater than, to match.
        if prospective.local is not None:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # greater than the spec version *and* it's not a pre-release of the
        # same version in the spec.
        return True

    def _compare_arbitrary(self, prospective, spec):
        # type: (Version, str) -> bool
        return str(prospective).lower() == str(spec).lower()

    @property
    def prereleases(self):
        # type: () -> bool

        # If there is an explicit prereleases set for this, then we'll just
        # blindly use that.
        if self._prereleases is not None:
            return self._prereleases

        # Look at all of our specifiers and determine if they are inclusive
        # operators, and if they are if they are including an explicit
        # prerelease.
        operator, version = self._spec
        if operator in ["==", ">=", "<=", "~=", "==="]:
            # The == specifier can include a trailing .*, if it does we
            # want to remove before parsing.
            if operator == "==" and version.endswith(".*"):
                version = version[:-2]

            # Parse the version, and if it is a pre-release than this
            # specifier allows pre-releases.
            if parse(version).is_prerelease:
                return True

        return False

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value


_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")


def _version_split(version):
    # type: (str) -> List[str]
    result = []  # type: List[str]
    for item in version.split("."):
        match = _prefix_regex.search(item)
        if match:
            result.extend(match.groups())
        else:
            result.append(item)
    return result


def _pad_version(left, right):
    # type: (List[str], List[str]) -> Tuple[List[str], List[str]]
    left_split, right_split = [], []

    # Get the release segment of our versions
    left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
    right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))

    # Get the rest of our versions
    left_split.append(left[len(left_split[0]) :])
    right_split.append(right[len(right_split[0]) :])

    # Insert our padding
    left_split.insert(1, ["0"] * max(0, len(right_split[0]) - len(left_split[0])))
    right_split.insert(1, ["0"] * max(0, len(left_split[0]) - len(right_split[0])))

    return (list(itertools.chain(*left_split)), list(itertools.chain(*right_split)))


class SpecifierSet(BaseSpecifier):
    def __init__(self, specifiers="", prereleases=None):
        # type: (str, Optional[bool]) -> None

        # Split on , to break each individual specifier into it's own item, and
        # strip each item to remove leading/trailing whitespace.
        split_specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]

        # Parsed each individual specifier, attempting first to make it a
        # Specifier and falling back to a LegacySpecifier.
        parsed = set()
        for specifier in split_specifiers:
            try:
                parsed.add(Specifier(specifier))
            except InvalidSpecifier:
                parsed.add(LegacySpecifier(specifier))

        # Turn our parsed specifiers into a frozen set and save them for later.
        self._specs = frozenset(parsed)

        # Store our prereleases value so we can use it later to determine if
        # we accept prereleases or not.
        self._prereleases = prereleases

    def __repr__(self):
        # type: () -> str
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<SpecifierSet({0!r}{1})>".format(str(self), pre)

    def __str__(self):
        # type: () -> str
        return ",".join(sorted(str(s) for s in self._specs))

    def __hash__(self):
        # type: () -> int
        return hash(self._specs)

    def __and__(self, other):
        # type: (Union[SpecifierSet, str]) -> SpecifierSet
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        specifier = SpecifierSet()
        specifier._specs = frozenset(self._specs | other._specs)

        if self._prereleases is None and other._prereleases is not None:
            specifier._prereleases = other._prereleases
        elif self._prereleases is not None and other._prereleases is None:
            specifier._prereleases = self._prereleases
        elif self._prereleases == other._prereleases:
            specifier._prereleases = self._prereleases
        else:
            raise ValueError(
                "Cannot combine SpecifierSets with True and False prerelease "
                "overrides."
            )

        return specifier

    def __eq__(self, other):
        # type: (object) -> bool
        if isinstance(other, (string_types, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs == other._specs

    def __ne__(self, other):
        # type: (object) -> bool
        if isinstance(other, (string_types, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs != other._specs

    def __len__(self):
        # type: () -> int
        return len(self._specs)

    def __iter__(self):
        # type: () -> Iterator[FrozenSet[_IndividualSpecifier]]
        return iter(self._specs)

    @property
    def prereleases(self):
        # type: () -> Optional[bool]

        # If we have been given an explicit prerelease modifier, then we'll
        # pass that through here.
        if self._prereleases is not None:
            return self._prereleases

        # If we don't have any specifiers, and we don't have a forced value,
        # then we'll just return None since we don't know if this should have
        # pre-releases or not.
        if not self._specs:
            return None

        # Otherwise we'll see if any of the given specifiers accept
        # prereleases, if any of them do we'll return True, otherwise False.
        return any(s.prereleases for s in self._specs)

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value

    def __contains__(self, item):
        # type: (Union[ParsedVersion, str]) -> bool
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # type: (Union[ParsedVersion, str], Optional[bool]) -> bool

        # Ensure that our item is a Version or LegacyVersion instance.
        if not isinstance(item, (LegacyVersion, Version)):
            item = parse(item)

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # We can determine if we're going to allow pre-releases by looking to
        # see if any of the underlying items supports them. If none of them do
        # and this item is a pre-release then we do not allow it and we can
        # short circuit that here.
        # Note: This means that 1.0.dev1 would not be contained in something
        #       like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
        if not prereleases and item.is_prerelease:
            return False

        # We simply dispatch to the underlying specs here to make sure that the
        # given version is contained within all of them.
        # Note: This use of all() here means that an empty set of specifiers
        #       will always return True, this is an explicit design decision.
        return all(s.contains(item, prereleases=prereleases) for s in self._specs)

    def filter(
        self,
        iterable,  # type: Iterable[Union[ParsedVersion, str]]
        prereleases=None,  # type: Optional[bool]
    ):
        # type: (...) -> Iterable[Union[ParsedVersion, str]]

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # If we have any specifiers, then we want to wrap our iterable in the
        # filter method for each one, this will act as a logical AND amongst
        # each specifier.
        if self._specs:
            for spec in self._specs:
                iterable = spec.filter(iterable, prereleases=bool(prereleases))
            return iterable
        # If we do not have any specifiers, then we need to have a rough filter
        # which will filter out any pre-releases, unless there are no final
        # releases, and which will filter out LegacyVersion in general.
        else:
            filtered = []  # type: List[Union[ParsedVersion, str]]
            found_prereleases = []  # type: List[Union[ParsedVersion, str]]

            for item in iterable:
                # Ensure that we some kind of Version class for this item.
                if not isinstance(item, (LegacyVersion, Version)):
                    parsed_version = parse(item)
                else:
                    parsed_version = item

                # Filter out any item which is parsed as a LegacyVersion
                if isinstance(parsed_version, LegacyVersion):
                    continue

                # Store any item which is a pre-release for later unless we've
                # already found a final version or we are accepting prereleases
                if parsed_version.is_prerelease and not prereleases:
                    if not filtered:
                        found_prereleases.append(item)
                else:
                    filtered.append(item)

            # If we've found no items except for pre-releases, then we'll go
            # ahead and use the pre-releases
            if not filtered and found_prereleases and prereleases is None:
                return found_prereleases

            return filtered
site-packages/setuptools/_vendor/packaging/_structures.py000064400000003746150564745010020042 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function


class InfinityType(object):
    def __repr__(self):
        # type: () -> str
        return "Infinity"

    def __hash__(self):
        # type: () -> int
        return hash(repr(self))

    def __lt__(self, other):
        # type: (object) -> bool
        return False

    def __le__(self, other):
        # type: (object) -> bool
        return False

    def __eq__(self, other):
        # type: (object) -> bool
        return isinstance(other, self.__class__)

    def __ne__(self, other):
        # type: (object) -> bool
        return not isinstance(other, self.__class__)

    def __gt__(self, other):
        # type: (object) -> bool
        return True

    def __ge__(self, other):
        # type: (object) -> bool
        return True

    def __neg__(self):
        # type: (object) -> NegativeInfinityType
        return NegativeInfinity


Infinity = InfinityType()


class NegativeInfinityType(object):
    def __repr__(self):
        # type: () -> str
        return "-Infinity"

    def __hash__(self):
        # type: () -> int
        return hash(repr(self))

    def __lt__(self, other):
        # type: (object) -> bool
        return True

    def __le__(self, other):
        # type: (object) -> bool
        return True

    def __eq__(self, other):
        # type: (object) -> bool
        return isinstance(other, self.__class__)

    def __ne__(self, other):
        # type: (object) -> bool
        return not isinstance(other, self.__class__)

    def __gt__(self, other):
        # type: (object) -> bool
        return False

    def __ge__(self, other):
        # type: (object) -> bool
        return False

    def __neg__(self):
        # type: (object) -> InfinityType
        return Infinity


NegativeInfinity = NegativeInfinityType()
site-packages/setuptools/_vendor/packaging/tags.py000064400000057003150564745010016411 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

from __future__ import absolute_import

import distutils.util

try:
    from importlib.machinery import EXTENSION_SUFFIXES
except ImportError:  # pragma: no cover
    import imp

    EXTENSION_SUFFIXES = [x[0] for x in imp.get_suffixes()]
    del imp
import logging
import os
import platform
import re
import struct
import sys
import sysconfig
import warnings

from ._typing import TYPE_CHECKING, cast

if TYPE_CHECKING:  # pragma: no cover
    from typing import (
        Dict,
        FrozenSet,
        IO,
        Iterable,
        Iterator,
        List,
        Optional,
        Sequence,
        Tuple,
        Union,
    )

    PythonVersion = Sequence[int]
    MacVersion = Tuple[int, int]
    GlibcVersion = Tuple[int, int]


logger = logging.getLogger(__name__)

INTERPRETER_SHORT_NAMES = {
    "python": "py",  # Generic.
    "cpython": "cp",
    "pypy": "pp",
    "ironpython": "ip",
    "jython": "jy",
}  # type: Dict[str, str]


_32_BIT_INTERPRETER = sys.maxsize <= 2 ** 32


class Tag(object):
    """
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    """

    __slots__ = ["_interpreter", "_abi", "_platform"]

    def __init__(self, interpreter, abi, platform):
        # type: (str, str, str) -> None
        self._interpreter = interpreter.lower()
        self._abi = abi.lower()
        self._platform = platform.lower()

    @property
    def interpreter(self):
        # type: () -> str
        return self._interpreter

    @property
    def abi(self):
        # type: () -> str
        return self._abi

    @property
    def platform(self):
        # type: () -> str
        return self._platform

    def __eq__(self, other):
        # type: (object) -> bool
        if not isinstance(other, Tag):
            return NotImplemented

        return (
            (self.platform == other.platform)
            and (self.abi == other.abi)
            and (self.interpreter == other.interpreter)
        )

    def __hash__(self):
        # type: () -> int
        return hash((self._interpreter, self._abi, self._platform))

    def __str__(self):
        # type: () -> str
        return "{}-{}-{}".format(self._interpreter, self._abi, self._platform)

    def __repr__(self):
        # type: () -> str
        return "<{self} @ {self_id}>".format(self=self, self_id=id(self))


def parse_tag(tag):
    # type: (str) -> FrozenSet[Tag]
    """
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    """
    tags = set()
    interpreters, abis, platforms = tag.split("-")
    for interpreter in interpreters.split("."):
        for abi in abis.split("."):
            for platform_ in platforms.split("."):
                tags.add(Tag(interpreter, abi, platform_))
    return frozenset(tags)


def _warn_keyword_parameter(func_name, kwargs):
    # type: (str, Dict[str, bool]) -> bool
    """
    Backwards-compatibility with Python 2.7 to allow treating 'warn' as keyword-only.
    """
    if not kwargs:
        return False
    elif len(kwargs) > 1 or "warn" not in kwargs:
        kwargs.pop("warn", None)
        arg = next(iter(kwargs.keys()))
        raise TypeError(
            "{}() got an unexpected keyword argument {!r}".format(func_name, arg)
        )
    return kwargs["warn"]


def _get_config_var(name, warn=False):
    # type: (str, bool) -> Union[int, str, None]
    value = sysconfig.get_config_var(name)
    if value is None and warn:
        logger.debug(
            "Config variable '%s' is unset, Python ABI tag may be incorrect", name
        )
    return value


def _normalize_string(string):
    # type: (str) -> str
    return string.replace(".", "_").replace("-", "_")


def _abi3_applies(python_version):
    # type: (PythonVersion) -> bool
    """
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    """
    return len(python_version) > 1 and tuple(python_version) >= (3, 2)


def _cpython_abis(py_version, warn=False):
    # type: (PythonVersion, bool) -> List[str]
    py_version = tuple(py_version)  # To allow for version comparison.
    abis = []
    version = _version_nodot(py_version[:2])
    debug = pymalloc = ucs4 = ""
    with_debug = _get_config_var("Py_DEBUG", warn)
    has_refcount = hasattr(sys, "gettotalrefcount")
    # Windows doesn't set Py_DEBUG, so checking for support of debug-compiled
    # extension modules is the best option.
    # https://github.com/pypa/pip/issues/3383#issuecomment-173267692
    has_ext = "_d.pyd" in EXTENSION_SUFFIXES
    if with_debug or (with_debug is None and (has_refcount or has_ext)):
        debug = "d"
    if py_version < (3, 8):
        with_pymalloc = _get_config_var("WITH_PYMALLOC", warn)
        if with_pymalloc or with_pymalloc is None:
            pymalloc = "m"
        if py_version < (3, 3):
            unicode_size = _get_config_var("Py_UNICODE_SIZE", warn)
            if unicode_size == 4 or (
                unicode_size is None and sys.maxunicode == 0x10FFFF
            ):
                ucs4 = "u"
    elif debug:
        # Debug builds can also load "normal" extension modules.
        # We can also assume no UCS-4 or pymalloc requirement.
        abis.append("cp{version}".format(version=version))
    abis.insert(
        0,
        "cp{version}{debug}{pymalloc}{ucs4}".format(
            version=version, debug=debug, pymalloc=pymalloc, ucs4=ucs4
        ),
    )
    return abis


def cpython_tags(
    python_version=None,  # type: Optional[PythonVersion]
    abis=None,  # type: Optional[Iterable[str]]
    platforms=None,  # type: Optional[Iterable[str]]
    **kwargs  # type: bool
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    """
    warn = _warn_keyword_parameter("cpython_tags", kwargs)
    if not python_version:
        python_version = sys.version_info[:2]

    interpreter = "cp{}".format(_version_nodot(python_version[:2]))

    if abis is None:
        if len(python_version) > 1:
            abis = _cpython_abis(python_version, warn)
        else:
            abis = []
    abis = list(abis)
    # 'abi3' and 'none' are explicitly handled later.
    for explicit_abi in ("abi3", "none"):
        try:
            abis.remove(explicit_abi)
        except ValueError:
            pass

    platforms = list(platforms or _platform_tags())
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)
    if _abi3_applies(python_version):
        for tag in (Tag(interpreter, "abi3", platform_) for platform_ in platforms):
            yield tag
    for tag in (Tag(interpreter, "none", platform_) for platform_ in platforms):
        yield tag

    if _abi3_applies(python_version):
        for minor_version in range(python_version[1] - 1, 1, -1):
            for platform_ in platforms:
                interpreter = "cp{version}".format(
                    version=_version_nodot((python_version[0], minor_version))
                )
                yield Tag(interpreter, "abi3", platform_)


def _generic_abi():
    # type: () -> Iterator[str]
    abi = sysconfig.get_config_var("SOABI")
    if abi:
        yield _normalize_string(abi)


def generic_tags(
    interpreter=None,  # type: Optional[str]
    abis=None,  # type: Optional[Iterable[str]]
    platforms=None,  # type: Optional[Iterable[str]]
    **kwargs  # type: bool
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    """
    warn = _warn_keyword_parameter("generic_tags", kwargs)
    if not interpreter:
        interp_name = interpreter_name()
        interp_version = interpreter_version(warn=warn)
        interpreter = "".join([interp_name, interp_version])
    if abis is None:
        abis = _generic_abi()
    platforms = list(platforms or _platform_tags())
    abis = list(abis)
    if "none" not in abis:
        abis.append("none")
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)


def _py_interpreter_range(py_version):
    # type: (PythonVersion) -> Iterator[str]
    """
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    """
    if len(py_version) > 1:
        yield "py{version}".format(version=_version_nodot(py_version[:2]))
    yield "py{major}".format(major=py_version[0])
    if len(py_version) > 1:
        for minor in range(py_version[1] - 1, -1, -1):
            yield "py{version}".format(version=_version_nodot((py_version[0], minor)))


def compatible_tags(
    python_version=None,  # type: Optional[PythonVersion]
    interpreter=None,  # type: Optional[str]
    platforms=None,  # type: Optional[Iterable[str]]
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    """
    if not python_version:
        python_version = sys.version_info[:2]
    platforms = list(platforms or _platform_tags())
    for version in _py_interpreter_range(python_version):
        for platform_ in platforms:
            yield Tag(version, "none", platform_)
    if interpreter:
        yield Tag(interpreter, "none", "any")
    for version in _py_interpreter_range(python_version):
        yield Tag(version, "none", "any")


def _mac_arch(arch, is_32bit=_32_BIT_INTERPRETER):
    # type: (str, bool) -> str
    if not is_32bit:
        return arch

    if arch.startswith("ppc"):
        return "ppc"

    return "i386"


def _mac_binary_formats(version, cpu_arch):
    # type: (MacVersion, str) -> List[str]
    formats = [cpu_arch]
    if cpu_arch == "x86_64":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat64", "fat32"])

    elif cpu_arch == "i386":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat32", "fat"])

    elif cpu_arch == "ppc64":
        # TODO: Need to care about 32-bit PPC for ppc64 through 10.2?
        if version > (10, 5) or version < (10, 4):
            return []
        formats.append("fat64")

    elif cpu_arch == "ppc":
        if version > (10, 6):
            return []
        formats.extend(["fat32", "fat"])

    formats.append("universal")
    return formats


def mac_platforms(version=None, arch=None):
    # type: (Optional[MacVersion], Optional[str]) -> Iterator[str]
    """
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    """
    version_str, _, cpu_arch = platform.mac_ver()  # type: ignore
    if version is None:
        version = cast("MacVersion", tuple(map(int, version_str.split(".")[:2])))
    else:
        version = version
    if arch is None:
        arch = _mac_arch(cpu_arch)
    else:
        arch = arch
    for minor_version in range(version[1], -1, -1):
        compat_version = version[0], minor_version
        binary_formats = _mac_binary_formats(compat_version, arch)
        for binary_format in binary_formats:
            yield "macosx_{major}_{minor}_{binary_format}".format(
                major=compat_version[0],
                minor=compat_version[1],
                binary_format=binary_format,
            )


# From PEP 513.
def _is_manylinux_compatible(name, glibc_version):
    # type: (str, GlibcVersion) -> bool
    # Check for presence of _manylinux module.
    try:
        import _manylinux  # noqa

        return bool(getattr(_manylinux, name + "_compatible"))
    except (ImportError, AttributeError):
        # Fall through to heuristic check below.
        pass

    return _have_compatible_glibc(*glibc_version)


def _glibc_version_string():
    # type: () -> Optional[str]
    # Returns glibc version string, or None if not using glibc.
    return _glibc_version_string_confstr() or _glibc_version_string_ctypes()


def _glibc_version_string_confstr():
    # type: () -> Optional[str]
    """
    Primary implementation of glibc_version_string using os.confstr.
    """
    # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely
    # to be broken or missing. This strategy is used in the standard library
    # platform module.
    # https://github.com/python/cpython/blob/fcf1d003bf4f0100c9d0921ff3d70e1127ca1b71/Lib/platform.py#L175-L183
    try:
        # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17".
        version_string = os.confstr(  # type: ignore[attr-defined] # noqa: F821
            "CS_GNU_LIBC_VERSION"
        )
        assert version_string is not None
        _, version = version_string.split()  # type: Tuple[str, str]
    except (AssertionError, AttributeError, OSError, ValueError):
        # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)...
        return None
    return version


def _glibc_version_string_ctypes():
    # type: () -> Optional[str]
    """
    Fallback implementation of glibc_version_string using ctypes.
    """
    try:
        import ctypes
    except ImportError:
        return None

    # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen
    # manpage says, "If filename is NULL, then the returned handle is for the
    # main program". This way we can let the linker do the work to figure out
    # which libc our process is actually using.
    #
    # Note: typeshed is wrong here so we are ignoring this line.
    process_namespace = ctypes.CDLL(None)  # type: ignore
    try:
        gnu_get_libc_version = process_namespace.gnu_get_libc_version
    except AttributeError:
        # Symbol doesn't exist -> therefore, we are not linked to
        # glibc.
        return None

    # Call gnu_get_libc_version, which returns a string like "2.5"
    gnu_get_libc_version.restype = ctypes.c_char_p
    version_str = gnu_get_libc_version()  # type: str
    # py2 / py3 compatibility:
    if not isinstance(version_str, str):
        version_str = version_str.decode("ascii")

    return version_str


# Separated out from have_compatible_glibc for easier unit testing.
def _check_glibc_version(version_str, required_major, minimum_minor):
    # type: (str, int, int) -> bool
    # Parse string and check against requested version.
    #
    # We use a regexp instead of str.split because we want to discard any
    # random junk that might come after the minor version -- this might happen
    # in patched/forked versions of glibc (e.g. Linaro's version of glibc
    # uses version strings like "2.20-2014.11"). See gh-3588.
    m = re.match(r"(?P<major>[0-9]+)\.(?P<minor>[0-9]+)", version_str)
    if not m:
        warnings.warn(
            "Expected glibc version with 2 components major.minor,"
            " got: %s" % version_str,
            RuntimeWarning,
        )
        return False
    return (
        int(m.group("major")) == required_major
        and int(m.group("minor")) >= minimum_minor
    )


def _have_compatible_glibc(required_major, minimum_minor):
    # type: (int, int) -> bool
    version_str = _glibc_version_string()
    if version_str is None:
        return False
    return _check_glibc_version(version_str, required_major, minimum_minor)


# Python does not provide platform information at sufficient granularity to
# identify the architecture of the running executable in some cases, so we
# determine it dynamically by reading the information from the running
# process. This only applies on Linux, which uses the ELF format.
class _ELFFileHeader(object):
    # https://en.wikipedia.org/wiki/Executable_and_Linkable_Format#File_header
    class _InvalidELFFileHeader(ValueError):
        """
        An invalid ELF file header was found.
        """

    ELF_MAGIC_NUMBER = 0x7F454C46
    ELFCLASS32 = 1
    ELFCLASS64 = 2
    ELFDATA2LSB = 1
    ELFDATA2MSB = 2
    EM_386 = 3
    EM_S390 = 22
    EM_ARM = 40
    EM_X86_64 = 62
    EF_ARM_ABIMASK = 0xFF000000
    EF_ARM_ABI_VER5 = 0x05000000
    EF_ARM_ABI_FLOAT_HARD = 0x00000400

    def __init__(self, file):
        # type: (IO[bytes]) -> None
        def unpack(fmt):
            # type: (str) -> int
            try:
                (result,) = struct.unpack(
                    fmt, file.read(struct.calcsize(fmt))
                )  # type: (int, )
            except struct.error:
                raise _ELFFileHeader._InvalidELFFileHeader()
            return result

        self.e_ident_magic = unpack(">I")
        if self.e_ident_magic != self.ELF_MAGIC_NUMBER:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_class = unpack("B")
        if self.e_ident_class not in {self.ELFCLASS32, self.ELFCLASS64}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_data = unpack("B")
        if self.e_ident_data not in {self.ELFDATA2LSB, self.ELFDATA2MSB}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_version = unpack("B")
        self.e_ident_osabi = unpack("B")
        self.e_ident_abiversion = unpack("B")
        self.e_ident_pad = file.read(7)
        format_h = "<H" if self.e_ident_data == self.ELFDATA2LSB else ">H"
        format_i = "<I" if self.e_ident_data == self.ELFDATA2LSB else ">I"
        format_q = "<Q" if self.e_ident_data == self.ELFDATA2LSB else ">Q"
        format_p = format_i if self.e_ident_class == self.ELFCLASS32 else format_q
        self.e_type = unpack(format_h)
        self.e_machine = unpack(format_h)
        self.e_version = unpack(format_i)
        self.e_entry = unpack(format_p)
        self.e_phoff = unpack(format_p)
        self.e_shoff = unpack(format_p)
        self.e_flags = unpack(format_i)
        self.e_ehsize = unpack(format_h)
        self.e_phentsize = unpack(format_h)
        self.e_phnum = unpack(format_h)
        self.e_shentsize = unpack(format_h)
        self.e_shnum = unpack(format_h)
        self.e_shstrndx = unpack(format_h)


def _get_elf_header():
    # type: () -> Optional[_ELFFileHeader]
    try:
        with open(sys.executable, "rb") as f:
            elf_header = _ELFFileHeader(f)
    except (IOError, OSError, TypeError, _ELFFileHeader._InvalidELFFileHeader):
        return None
    return elf_header


def _is_linux_armhf():
    # type: () -> bool
    # hard-float ABI can be detected from the ELF header of the running
    # process
    # https://static.docs.arm.com/ihi0044/g/aaelf32.pdf
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_ARM
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABIMASK
    ) == elf_header.EF_ARM_ABI_VER5
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABI_FLOAT_HARD
    ) == elf_header.EF_ARM_ABI_FLOAT_HARD
    return result


def _is_linux_i686():
    # type: () -> bool
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_386
    return result


def _have_compatible_manylinux_abi(arch):
    # type: (str) -> bool
    if arch == "armv7l":
        return _is_linux_armhf()
    if arch == "i686":
        return _is_linux_i686()
    return True


def _linux_platforms(is_32bit=_32_BIT_INTERPRETER):
    # type: (bool) -> Iterator[str]
    linux = _normalize_string(distutils.util.get_platform())
    if is_32bit:
        if linux == "linux_x86_64":
            linux = "linux_i686"
        elif linux == "linux_aarch64":
            linux = "linux_armv7l"
    manylinux_support = []
    _, arch = linux.split("_", 1)
    if _have_compatible_manylinux_abi(arch):
        if arch in {"x86_64", "i686", "aarch64", "armv7l", "ppc64", "ppc64le", "s390x"}:
            manylinux_support.append(
                ("manylinux2014", (2, 17))
            )  # CentOS 7 w/ glibc 2.17 (PEP 599)
        if arch in {"x86_64", "i686"}:
            manylinux_support.append(
                ("manylinux2010", (2, 12))
            )  # CentOS 6 w/ glibc 2.12 (PEP 571)
            manylinux_support.append(
                ("manylinux1", (2, 5))
            )  # CentOS 5 w/ glibc 2.5 (PEP 513)
    manylinux_support_iter = iter(manylinux_support)
    for name, glibc_version in manylinux_support_iter:
        if _is_manylinux_compatible(name, glibc_version):
            yield linux.replace("linux", name)
            break
    # Support for a later manylinux implies support for an earlier version.
    for name, _ in manylinux_support_iter:
        yield linux.replace("linux", name)
    yield linux


def _generic_platforms():
    # type: () -> Iterator[str]
    yield _normalize_string(distutils.util.get_platform())


def _platform_tags():
    # type: () -> Iterator[str]
    """
    Provides the platform tags for this installation.
    """
    if platform.system() == "Darwin":
        return mac_platforms()
    elif platform.system() == "Linux":
        return _linux_platforms()
    else:
        return _generic_platforms()


def interpreter_name():
    # type: () -> str
    """
    Returns the name of the running interpreter.
    """
    try:
        name = sys.implementation.name  # type: ignore
    except AttributeError:  # pragma: no cover
        # Python 2.7 compatibility.
        name = platform.python_implementation().lower()
    return INTERPRETER_SHORT_NAMES.get(name) or name


def interpreter_version(**kwargs):
    # type: (bool) -> str
    """
    Returns the version of the running interpreter.
    """
    warn = _warn_keyword_parameter("interpreter_version", kwargs)
    version = _get_config_var("py_version_nodot", warn=warn)
    if version:
        version = str(version)
    else:
        version = _version_nodot(sys.version_info[:2])
    return version


def _version_nodot(version):
    # type: (PythonVersion) -> str
    if any(v >= 10 for v in version):
        sep = "_"
    else:
        sep = ""
    return sep.join(map(str, version))


def sys_tags(**kwargs):
    # type: (bool) -> Iterator[Tag]
    """
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    """
    warn = _warn_keyword_parameter("sys_tags", kwargs)

    interp_name = interpreter_name()
    if interp_name == "cp":
        for tag in cpython_tags(warn=warn):
            yield tag
    else:
        for tag in generic_tags():
            yield tag

    for tag in compatible_tags():
        yield tag
site-packages/setuptools/_vendor/packaging/requirements.py000064400000011465150564745010020200 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import string
import re

from setuptools.extern.pyparsing import stringStart, stringEnd, originalTextFor, ParseException
from setuptools.extern.pyparsing import ZeroOrMore, Word, Optional, Regex, Combine
from setuptools.extern.pyparsing import Literal as L  # noqa
from urllib import parse as urlparse

from ._typing import TYPE_CHECKING
from .markers import MARKER_EXPR, Marker
from .specifiers import LegacySpecifier, Specifier, SpecifierSet

if TYPE_CHECKING:  # pragma: no cover
    from typing import List


class InvalidRequirement(ValueError):
    """
    An invalid requirement was found, users should refer to PEP 508.
    """


ALPHANUM = Word(string.ascii_letters + string.digits)

LBRACKET = L("[").suppress()
RBRACKET = L("]").suppress()
LPAREN = L("(").suppress()
RPAREN = L(")").suppress()
COMMA = L(",").suppress()
SEMICOLON = L(";").suppress()
AT = L("@").suppress()

PUNCTUATION = Word("-_.")
IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM)
IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END))

NAME = IDENTIFIER("name")
EXTRA = IDENTIFIER

URI = Regex(r"[^ ]+")("url")
URL = AT + URI

EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA)
EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras")

VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE)
VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE)

VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY
VERSION_MANY = Combine(
    VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False
)("_raw_spec")
_VERSION_SPEC = Optional(((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY))
_VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or "")

VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier")
VERSION_SPEC.setParseAction(lambda s, l, t: t[1])

MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker")
MARKER_EXPR.setParseAction(
    lambda s, l, t: Marker(s[t._original_start : t._original_end])
)
MARKER_SEPARATOR = SEMICOLON
MARKER = MARKER_SEPARATOR + MARKER_EXPR

VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER)
URL_AND_MARKER = URL + Optional(MARKER)

NAMED_REQUIREMENT = NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER)

REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd
# setuptools.extern.pyparsing isn't thread safe during initialization, so we do it eagerly, see
# issue #104
REQUIREMENT.parseString("x[]")


class Requirement(object):
    """Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    """

    # TODO: Can we test whether something is contained within a requirement?
    #       If so how do we do that? Do we need to test against the _name_ of
    #       the thing as well as the version? What about the markers?
    # TODO: Can we normalize the name and extra name?

    def __init__(self, requirement_string):
        # type: (str) -> None
        try:
            req = REQUIREMENT.parseString(requirement_string)
        except ParseException as e:
            raise InvalidRequirement(
                'Parse error at "{0!r}": {1}'.format(
                    requirement_string[e.loc : e.loc + 8], e.msg
                )
            )

        self.name = req.name
        if req.url:
            parsed_url = urlparse.urlparse(req.url)
            if parsed_url.scheme == "file":
                if urlparse.urlunparse(parsed_url) != req.url:
                    raise InvalidRequirement("Invalid URL given")
            elif not (parsed_url.scheme and parsed_url.netloc) or (
                not parsed_url.scheme and not parsed_url.netloc
            ):
                raise InvalidRequirement("Invalid URL: {0}".format(req.url))
            self.url = req.url
        else:
            self.url = None
        self.extras = set(req.extras.asList() if req.extras else [])
        self.specifier = SpecifierSet(req.specifier)
        self.marker = req.marker if req.marker else None

    def __str__(self):
        # type: () -> str
        parts = [self.name]  # type: List[str]

        if self.extras:
            parts.append("[{0}]".format(",".join(sorted(self.extras))))

        if self.specifier:
            parts.append(str(self.specifier))

        if self.url:
            parts.append("@ {0}".format(self.url))
            if self.marker:
                parts.append(" ")

        if self.marker:
            parts.append("; {0}".format(self.marker))

        return "".join(parts)

    def __repr__(self):
        # type: () -> str
        return "<Requirement({0!r})>".format(str(self))
site-packages/setuptools/_vendor/more_itertools/__pycache__/more.cpython-39.pyc000064400000326756150564745010024044 0ustar00a

(�Re���@s�ddlZddlmZmZmZmZddlmZddlm	Z	ddl
mZmZm
Z
ddlmZmZmZmZddlmZmZmZmZmZmZmZmZmZmZmZmZddl m!Z!m"Z"m#Z#m$Z$dd	l%m&Z&m'Z'dd
l(m(Z(m)Z)m*Z*ddl+m,Z,m-Z-m.Z.m/Z/m0Z0ddl1m2Z2m3Z3dd
l4m5Z5ddl6m7Z7m8Z8m9Z9m:Z:m;Z;m<Z<gd�Z=e>�Z?d�dd�Z@e?fdd�ZAe?fdd�ZBe?fdd�ZCGdd�d�ZDdd�ZEdd�ZFd d!�ZGd"d#�ZHd$d%�ZId�d&d'�ZJd�d(d)�ZKd�d*d+�ZLd,d-�ZMd�d.d/�ZNd0d1�ZOd�d2d3�ZPGd4d5�d5�ZQd�d6d7�ZRd8d9�ZSd:d;�ZTd�d<d=�ZUd�d>d?�ZVd�d@dA�ZWd�dCdD�ZXd�dEdF�ZYd�dGdH�ZZd�dIdJ�Z[dKdL�Z\d�dMdN�Z]d�dOdP�Z^dQdR�Z_d�dTdU�Z`GdVdW�dWea�ZbdXdY�ZcdZd[�Zdddd\�d]d^�Zed�d`da�Zfdbdc�Zgddde�Zheiejffdfdg�Zkd�dhdi�Zld�djdk�ZmGdldm�dmejejn�Zod�dndo�Zpdpdq�Zqerdfdrds�Zsdtdu�Ztdvdw�Zudxdy�ZvGdzd{�d{�Zwd|d}�Zxd~d�Zyd�d��fd�d��Zze.fdd��d�d��Z{Gd�d��d�e�Z|Gd�d��d��Z}Gd�d��d��Z~erfd�d��Zd�d��Z�d�d�d��Z�d�d�d��Z�erdfd�d��Z�d�d�d��Z�d�d��Z�d�d�d��Z�Gd�d��d��Z�d�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d�d��Z�d�d�d��Z�Gd�d��d�e��Z�Gd�d��d��Z�d�d��Z�d�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d„Z�d�dĄZ�Gd�dƄdƃZ�dS)��N)�Counter�defaultdict�deque�abc)�Sequence)�ThreadPoolExecutor)�partial�reduce�wraps)�merge�heapify�heapreplace�heappop)�chain�compress�count�cycle�	dropwhile�groupby�islice�repeat�starmap�	takewhile�tee�zip_longest)�exp�	factorial�floor�log)�Empty�Queue)�random�	randrange�uniform)�
itemgetter�mul�sub�gt�lt)�
hexversion�maxsize)�	monotonic�)�consume�flatten�pairwise�powerset�take�unique_everseen)S�AbortThread�adjacent�always_iterable�always_reversible�bucket�
callback_iter�chunked�circular_shifts�collapse�collate�consecutive_groups�consumer�	countable�count_cycle�	mark_ends�
difference�distinct_combinations�distinct_permutations�
distribute�divide�	exactly_n�
filter_except�first�groupby_transform�ilen�interleave_longest�
interleave�intersperse�islice_extended�iterate�ichunked�	is_sorted�last�locate�lstrip�make_decorator�
map_except�
map_reduce�nth_or_last�nth_permutation�nth_product�
numeric_range�one�only�padded�
partitions�set_partitions�peekable�repeat_last�replace�rlocate�rstrip�
run_length�sample�seekable�SequenceView�side_effect�sliced�
sort_together�split_at�split_after�split_before�
split_when�
split_into�spy�stagger�strip�
substrings�substrings_indexes�time_limited�unique_to_each�unzip�windowed�	with_iter�UnequalIterablesError�	zip_equal�
zip_offset�windowed_complete�
all_unique�value_chain�
product_index�combination_index�permutation_indexFcs:ttt�t|��g��|r2��fdd�}t|��S�SdS)aJBreak *iterable* into lists of length *n*:

        >>> list(chunked([1, 2, 3, 4, 5, 6], 3))
        [[1, 2, 3], [4, 5, 6]]

    By the default, the last yielded list will have fewer than *n* elements
    if the length of *iterable* is not divisible by *n*:

        >>> list(chunked([1, 2, 3, 4, 5, 6, 7, 8], 3))
        [[1, 2, 3], [4, 5, 6], [7, 8]]

    To use a fill-in value instead, see the :func:`grouper` recipe.

    If the length of *iterable* is not divisible by *n* and *strict* is
    ``True``, then ``ValueError`` will be raised before the last
    list is yielded.

    c3s(�D]}t|��krtd��|VqdS)Nziterable is not divisible by n.��len�
ValueError)�chunk��iterator�n���/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/more_itertools/more.py�ret�szchunked.<locals>.retN)�iterrr1)�iterabler��strictr�r�r�r�r9s

r9c
CsNztt|��WStyH}z"|tur0td�|�|WYd}~Sd}~00dS)a�Return the first item of *iterable*, or *default* if *iterable* is
    empty.

        >>> first([0, 1, 2, 3])
        0
        >>> first([], 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.

    :func:`first` is useful when you have a generator of expensive-to-retrieve
    values and want any arbitrary one. It is marginally shorter than
    ``next(iter(iterable), default)``.

    zKfirst() was called on an empty iterable, and no default value was provided.N)�nextr��
StopIteration�_markerr�)r��default�er�r�r�rI�s��rIc
Cs|zJt|t�r|dWSt|d�r6tdkr6tt|��WSt|dd�dWSWn,ttt	fyv|t
urntd��|YS0dS)aReturn the last item of *iterable*, or *default* if *iterable* is
    empty.

        >>> last([0, 1, 2, 3])
        3
        >>> last([], 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.
    ����__reversed__i�r,��maxlenzDlast() was called on an empty iterable, and no default was provided.N)�
isinstancer�hasattrr)r��reversedr�
IndexError�	TypeErrorr�r�r�)r�r�r�r�r�rS�s

�rScCstt||d�|d�S)agReturn the nth or the last item of *iterable*,
    or *default* if *iterable* is empty.

        >>> nth_or_last([0, 1, 2, 3], 2)
        2
        >>> nth_or_last([0, 1], 2)
        1
        >>> nth_or_last([], 0, 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.
    r,�r�)rSr)r�r�r�r�r�r�rY�srYc@sTeZdZdZdd�Zdd�Zdd�Zefdd	�Zd
d�Z	dd
�Z
dd�Zdd�ZdS)rbaWrap an iterator to allow lookahead and prepending elements.

    Call :meth:`peek` on the result to get the value that will be returned
    by :func:`next`. This won't advance the iterator:

        >>> p = peekable(['a', 'b'])
        >>> p.peek()
        'a'
        >>> next(p)
        'a'

    Pass :meth:`peek` a default value to return that instead of raising
    ``StopIteration`` when the iterator is exhausted.

        >>> p = peekable([])
        >>> p.peek('hi')
        'hi'

    peekables also offer a :meth:`prepend` method, which "inserts" items
    at the head of the iterable:

        >>> p = peekable([1, 2, 3])
        >>> p.prepend(10, 11, 12)
        >>> next(p)
        10
        >>> p.peek()
        11
        >>> list(p)
        [11, 12, 1, 2, 3]

    peekables can be indexed. Index 0 is the item that will be returned by
    :func:`next`, index 1 is the item after that, and so on:
    The values up to the given index will be cached.

        >>> p = peekable(['a', 'b', 'c', 'd'])
        >>> p[0]
        'a'
        >>> p[1]
        'b'
        >>> next(p)
        'a'

    Negative indexes are supported, but be aware that they will cache the
    remaining items in the source iterator, which may require significant
    storage.

    To check whether a peekable is exhausted, check its truth value:

        >>> p = peekable(['a', 'b'])
        >>> if p:  # peekable has items
        ...     list(p)
        ['a', 'b']
        >>> if not p:  # peekable is exhausted
        ...     list(p)
        []

    cCst|�|_t�|_dS�N)r��_itr�_cache��selfr�r�r�r��__init__%s
zpeekable.__init__cCs|Sr�r��r�r�r�r��__iter__)szpeekable.__iter__cCs&z|��Wnty YdS0dS�NFT��peekr�r�r�r�r��__bool__,s
zpeekable.__bool__cCsH|js>z|j�t|j��Wn ty<|tur4�|YS0|jdS)z�Return the item that will be next returned from ``next()``.

        Return ``default`` if there are no items left. If ``default`` is not
        provided, raise ``StopIteration``.

        r)r��appendr�r�r�r�)r�r�r�r�r�r�3s
z
peekable.peekcGs|j�t|��dS)aStack up items to be the next ones returned from ``next()`` or
        ``self.peek()``. The items will be returned in
        first in, first out order::

            >>> p = peekable([1, 2, 3])
            >>> p.prepend(10, 11, 12)
            >>> next(p)
            10
            >>> list(p)
            [11, 12, 1, 2, 3]

        It is possible, by prepending items, to "resurrect" a peekable that
        previously raised ``StopIteration``.

            >>> p = peekable([])
            >>> next(p)
            Traceback (most recent call last):
              ...
            StopIteration
            >>> p.prepend(1)
            >>> next(p)
            1
            >>> next(p)
            Traceback (most recent call last):
              ...
            StopIteration

        N)r��
extendleftr�)r��itemsr�r�r��prependCszpeekable.prependcCs|jr|j��St|j�Sr�)r��popleftr�r�r�r�r�r��__next__bs
zpeekable.__next__cCs�|jdurdn|j}|dkrF|jdur*dn|j}|jdur>tn|j}n@|dkr~|jdur\dn|j}|jdurvtdn|j}ntd��|dks�|dkr�|j�|j�n>tt	||�dt�}t
|j�}||kr�|j�t|j||��t|j�|S)Nr,rr�zslice step cannot be zero)
�step�start�stopr*r�r��extendr��min�maxr�r�list)r��indexr�r�r�r��	cache_lenr�r�r��
_get_slicehs
zpeekable._get_slicecCsdt|t�r|�|�St|j�}|dkr6|j�|j�n$||krZ|j�t|j|d|��|j|S�Nrr,)r��slicer�r�r�r�r�r)r�r�r�r�r�r��__getitem__�s


zpeekable.__getitem__N)
�__name__�
__module__�__qualname__�__doc__r�r�r�r�r�r�r�r�r�r�r�r�r�rb�s:rbcOst�dt�t|i|��S)a�Return a sorted merge of the items from each of several already-sorted
    *iterables*.

        >>> list(collate('ACDZ', 'AZ', 'JKL'))
        ['A', 'A', 'C', 'D', 'J', 'K', 'L', 'Z', 'Z']

    Works lazily, keeping only the next value from each iterable in memory. Use
    :func:`collate` to, for example, perform a n-way mergesort of items that
    don't fit in memory.

    If a *key* function is specified, the iterables will be sorted according
    to its result:

        >>> key = lambda s: int(s)  # Sort by numeric value, not by string
        >>> list(collate(['1', '10'], ['2', '11'], key=key))
        ['1', '2', '10', '11']


    If the *iterables* are sorted in descending order, set *reverse* to
    ``True``:

        >>> list(collate([5, 3, 1], [4, 2, 0], reverse=True))
        [5, 4, 3, 2, 1, 0]

    If the elements of the passed-in iterables are out of order, you might get
    unexpected results.

    On Python 3.5+, this function is an alias for :func:`heapq.merge`.

    z<collate is no longer part of more_itertools, use heapq.merge)�warnings�warn�DeprecationWarningr)�	iterables�kwargsr�r�r�r<�s
�r<cst���fdd��}|S)abDecorator that automatically advances a PEP-342-style "reverse iterator"
    to its first yield point so you don't have to call ``next()`` on it
    manually.

        >>> @consumer
        ... def tally():
        ...     i = 0
        ...     while True:
        ...         print('Thing number %s is %s.' % (i, (yield)))
        ...         i += 1
        ...
        >>> t = tally()
        >>> t.send('red')
        Thing number 0 is red.
        >>> t.send('fish')
        Thing number 1 is fish.

    Without the decorator, you would have to call ``next(t)`` before
    ``t.send()`` could be used.

    cs�|i|��}t|�|Sr�)r�)�argsr��gen��funcr�r��wrapper�szconsumer.<locals>.wrapper)r
)r�r�r�r�r�r>�sr>cCs t�}tt||�dd�t|�S)z�Return the number of items in *iterable*.

        >>> ilen(x for x in range(1000000) if x % 3 == 0)
        333334

    This consumes the iterable, so handle with care.

    rr�)rr�zipr�)r��counterr�r�r�rK�srKccs|V||�}qdS)z�Return ``start``, ``func(start)``, ``func(func(start))``, ...

    >>> from itertools import islice
    >>> list(islice(iterate(lambda x: 2*x, 1), 10))
    [1, 2, 4, 8, 16, 32, 64, 128, 256, 512]

    Nr�)r�r�r�r�r�rP�s	rPccs2|�}|EdHWd�n1s$0YdS)a:Wrap an iterable in a ``with`` statement, so it closes once exhausted.

    For example, this will close the file when the iterator is exhausted::

        upper_lines = (line.upper() for line in with_iter(open('foo')))

    Any context manager which returns an iterable is a candidate for
    ``with_iter``.

    Nr�)Zcontext_managerr�r�r�r�r|�sr|c
Cs�t|�}zt|�}Wn2tyF}z|p.td�|�WYd}~n
d}~00zt|�}WntyfYn0d�||�}|p~t|��|S)a�Return the first item from *iterable*, which is expected to contain only
    that item. Raise an exception if *iterable* is empty or has more than one
    item.

    :func:`one` is useful for ensuring that an iterable contains only one item.
    For example, it can be used to retrieve the result of a database query
    that is expected to return a single row.

    If *iterable* is empty, ``ValueError`` will be raised. You may specify a
    different exception with the *too_short* keyword:

        >>> it = []
        >>> one(it)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        ValueError: too many items in iterable (expected 1)'
        >>> too_short = IndexError('too few items')
        >>> one(it, too_short=too_short)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        IndexError: too few items

    Similarly, if *iterable* contains more than one item, ``ValueError`` will
    be raised. You may specify a different exception with the *too_long*
    keyword:

        >>> it = ['too', 'many']
        >>> one(it)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        ValueError: Expected exactly one item in iterable, but got 'too',
        'many', and perhaps more.
        >>> too_long = RuntimeError
        >>> one(it, too_long=too_long)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        RuntimeError

    Note that :func:`one` attempts to advance *iterable* twice to ensure there
    is only one item. See :func:`spy` or :func:`peekable` to check iterable
    contents less destructively.

    z&too few items in iterable (expected 1)N�LExpected exactly one item in iterable, but got {!r}, {!r}, and perhaps more.)r�r�r�r��format)r�Z	too_short�too_long�it�first_valuer��second_value�msgr�r�r�r]s",
���r]csr�fdd�}dd�}t|�}t|��|dur0�}d|krD�krbnn|�krX||�S|||�St|rldnd�S)	a�Yield successive distinct permutations of the elements in *iterable*.

        >>> sorted(distinct_permutations([1, 0, 1]))
        [(0, 1, 1), (1, 0, 1), (1, 1, 0)]

    Equivalent to ``set(permutations(iterable))``, except duplicates are not
    generated and thrown away. For larger input sequences this is much more
    efficient.

    Duplicate permutations arise when there are duplicated elements in the
    input iterable. The number of items returned is
    `n! / (x_1! * x_2! * ... * x_n!)`, where `n` is the total number of
    items input, and each `x_i` is the count of a distinct item in the input
    sequence.

    If *r* is given, only the *r*-length permutations are yielded.

        >>> sorted(distinct_permutations([1, 0, 1], r=2))
        [(0, 1), (1, 0), (1, 1)]
        >>> sorted(distinct_permutations(range(3), r=2))
        [(0, 1), (0, 2), (1, 0), (1, 2), (2, 0), (2, 1)]

    c3s�t|�Vt�ddd�D]}||||dkrq<qdSt�d|d�D]}||||krLqfqL||||||<||<|d|�d�||dd�<qdS)N�r�r,)�tuple�range)�A�i�j��sizer�r��_full^s
z$distinct_permutations.<locals>._fullc	ss4|d|�||d�}}t|ddd�}tt|��}t|�V|d}|D]}|||kr`qn||}qLdS|D]2}||||krr||||||<||<q�qr|D]2}||||kr�||||||<||<q�q�||d||d�7}|d7}|d||�|||d�||d�<|dd�<q6dS)Nr,r�)r�r�r�)	r��r�head�tailZright_head_indexesZleft_tail_indexesZpivotr�r�r�r�r��_partialvs*

z'distinct_permutations.<locals>._partialNrr�)r�)�sortedr�r�)r�r�r�r�r�r�r�r�rDEs'rDcCs^|dkrtd��nH|dkr0ttt|�|�dd�St|g�}t||�}ttt||�dd��SdS)a6Intersperse filler element *e* among the items in *iterable*, leaving
    *n* items between each filler element.

        >>> list(intersperse('!', [1, 2, 3, 4, 5]))
        [1, '!', 2, '!', 3, '!', 4, '!', 5]

        >>> list(intersperse(None, [1, 2, 3, 4, 5], n=2))
        [1, 2, None, 3, 4, None, 5]

    rz
n must be > 0r,N)r�rrMrr9r.)r�r�r�Zfiller�chunksr�r�r�rN�s


rNcsFdd�|D�}tt�tt|�����fdd��D���fdd�|D�S)a�Return the elements from each of the input iterables that aren't in the
    other input iterables.

    For example, suppose you have a set of packages, each with a set of
    dependencies::

        {'pkg_1': {'A', 'B'}, 'pkg_2': {'B', 'C'}, 'pkg_3': {'B', 'D'}}

    If you remove one package, which dependencies can also be removed?

    If ``pkg_1`` is removed, then ``A`` is no longer necessary - it is not
    associated with ``pkg_2`` or ``pkg_3``. Similarly, ``C`` is only needed for
    ``pkg_2``, and ``D`` is only needed for ``pkg_3``::

        >>> unique_to_each({'A', 'B'}, {'B', 'C'}, {'B', 'D'})
        [['A'], ['C'], ['D']]

    If there are duplicates in one input iterable that aren't in the others
    they will be duplicated in the output. Input order is preserved::

        >>> unique_to_each("mississippi", "missouri")
        [['p', 'p'], ['o', 'u', 'r']]

    It is assumed that the elements of each iterable are hashable.

    cSsg|]}t|��qSr�)r���.0r�r�r�r��
<listcomp>��z"unique_to_each.<locals>.<listcomp>csh|]}�|dkr|�qS�r,r�)r��element)�countsr�r��	<setcomp>�r�z!unique_to_each.<locals>.<setcomp>csg|]}tt�j|���qSr�)r��filter�__contains__r�)�uniquesr�r�r��r�)rr�
from_iterable�map�set)r��poolr�)r�r�r�ry�sryccs�|dkrtd��|dkr$t�VdS|dkr4td��t|d�}|}t|j|�D]}|d8}|sN|}t|�VqNt|�}||kr�tt|t|||���Vn6d|kr�t||�kr�nn||f|7}t|�VdS)aMReturn a sliding window of width *n* over the given iterable.

        >>> all_windows = windowed([1, 2, 3, 4, 5], 3)
        >>> list(all_windows)
        [(1, 2, 3), (2, 3, 4), (3, 4, 5)]

    When the window is larger than the iterable, *fillvalue* is used in place
    of missing values:

        >>> list(windowed([1, 2, 3], 4))
        [(1, 2, 3, None)]

    Each window will advance in increments of *step*:

        >>> list(windowed([1, 2, 3, 4, 5, 6], 3, fillvalue='!', step=2))
        [(1, 2, 3), (3, 4, 5), (5, 6, '!')]

    To slide into the iterable's items, use :func:`chain` to add filler items
    to the left:

        >>> iterable = [1, 2, 3, 4]
        >>> n = 3
        >>> padding = [None] * (n - 1)
        >>> list(windowed(chain(padding, iterable), 3))
        [(None, None, 1), (None, 1, 2), (1, 2, 3), (2, 3, 4)]
    r�n must be >= 0Nr,zstep must be >= 1r�)	r�r�rr�r�r�rrr�)�seqr��	fillvaluer�Zwindowr��_r�r�r�r�r{�s(
r{ccstg}t|�D]}|�|�|fVqt|�}t|�}td|d�D],}t||d�D]}||||�VqVqBdS)aFYield all of the substrings of *iterable*.

        >>> [''.join(s) for s in substrings('more')]
        ['m', 'o', 'r', 'e', 'mo', 'or', 're', 'mor', 'ore', 'more']

    Note that non-string iterables can also be subdivided.

        >>> list(substrings([0, 1, 2]))
        [(0,), (1,), (2,), (0, 1), (1, 2), (0, 1, 2)]

    r�r,N)r�r�r�r�r�)r�r��item�
item_countr�r�r�r�r�rvs


rvcs0tdt��d�}|rt|�}�fdd�|D�S)a@Yield all substrings and their positions in *seq*

    The items yielded will be a tuple of the form ``(substr, i, j)``, where
    ``substr == seq[i:j]``.

    This function only works for iterables that support slicing, such as
    ``str`` objects.

    >>> for item in substrings_indexes('more'):
    ...    print(item)
    ('m', 0, 1)
    ('o', 1, 2)
    ('r', 2, 3)
    ('e', 3, 4)
    ('mo', 0, 2)
    ('or', 1, 3)
    ('re', 2, 4)
    ('mor', 0, 3)
    ('ore', 1, 4)
    ('more', 0, 4)

    Set *reverse* to ``True`` to yield the same items in the opposite order.


    r,c3sB|]:}tt��|d�D] }�|||�|||fVqqdS�r,N)r�r�)r��Lr��r�r�r��	<genexpr>Nsz%substrings_indexes.<locals>.<genexpr>)r�r�r�)r��reverser�r�r�r�rw1s
�rwc@s:eZdZdZd
dd�Zdd�Zdd�Zd	d
�Zdd�ZdS)r7a�Wrap *iterable* and return an object that buckets it iterable into
    child iterables based on a *key* function.

        >>> iterable = ['a1', 'b1', 'c1', 'a2', 'b2', 'c2', 'b3']
        >>> s = bucket(iterable, key=lambda x: x[0])  # Bucket by 1st character
        >>> sorted(list(s))  # Get the keys
        ['a', 'b', 'c']
        >>> a_iterable = s['a']
        >>> next(a_iterable)
        'a1'
        >>> next(a_iterable)
        'a2'
        >>> list(s['b'])
        ['b1', 'b2', 'b3']

    The original iterable will be advanced and its items will be cached until
    they are used by the child iterables. This may require significant storage.

    By default, attempting to select a bucket to which no items belong  will
    exhaust the iterable and cache all values.
    If you specify a *validator* function, selected buckets will instead be
    checked against it.

        >>> from itertools import count
        >>> it = count(1, 2)  # Infinite sequence of odd numbers
        >>> key = lambda x: x % 10  # Bucket by last digit
        >>> validator = lambda x: x in {1, 3, 5, 7, 9}  # Odd digits only
        >>> s = bucket(it, key=key, validator=validator)
        >>> 2 in s
        False
        >>> list(s[2])
        []

    NcCs,t|�|_||_tt�|_|p$dd�|_dS)NcSsdS�NTr���xr�r�r��<lambda>{r�z!bucket.__init__.<locals>.<lambda>)r�r��_keyrrr��
_validator)r�r��key�	validatorr�r�r�r�ws

zbucket.__init__cCsH|�|�sdSzt||�}Wnty2YdS0|j|�|�dSr�)rr�r�r��
appendleft)r��valuer�r�r�r�r�}s
zbucket.__contains__ccs~|j|r|j|��Vqzt|j�}Wnty>YdS0|�|�}||kr\|Vqq|�|�r|j|�|�qqdS)z�
        Helper to yield items from the parent iterator that match *value*.
        Items that don't match are stored in the local cache as they
        are encountered.
        N)r�r�r�r�r�rrr�)r�r
r��
item_valuer�r�r��_get_values�s	


zbucket._get_valuesccsD|jD](}|�|�}|�|�r|j|�|�q|j��EdHdSr�)r�rrr�r��keys)r�r�rr�r�r�r��s



zbucket.__iter__cCs|�|�std�S|�|�S)Nr�)rr�r�r�r
r�r�r�r��s
zbucket.__getitem__)N)	r�r�r�r�r�r�rr�r�r�r�r�r�r7Ss#

r7cCs$t|�}t||�}|��t||�fS)a�Return a 2-tuple with a list containing the first *n* elements of
    *iterable*, and an iterator with the same items as *iterable*.
    This allows you to "look ahead" at the items in the iterable without
    advancing it.

    There is one item in the list by default:

        >>> iterable = 'abcdefg'
        >>> head, iterable = spy(iterable)
        >>> head
        ['a']
        >>> list(iterable)
        ['a', 'b', 'c', 'd', 'e', 'f', 'g']

    You may use unpacking to retrieve items instead of lists:

        >>> (head,), iterable = spy('abcdefg')
        >>> head
        'a'
        >>> (first, second), iterable = spy('abcdefg', 2)
        >>> first
        'a'
        >>> second
        'b'

    The number of items requested can be larger than the number of items in
    the iterable:

        >>> iterable = [1, 2, 3, 4, 5]
        >>> head, iterable = spy(iterable, 10)
        >>> head
        [1, 2, 3, 4, 5]
        >>> list(iterable)
        [1, 2, 3, 4, 5]

    )r�r1�copyr)r�r�r�r�r�r�r�rs�s%
rscGst�t|��S)a4Return a new iterable yielding from each iterable in turn,
    until the shortest is exhausted.

        >>> list(interleave([1, 2, 3], [4, 5], [6, 7, 8]))
        [1, 4, 6, 2, 5, 7]

    For a version that doesn't terminate after the shortest iterable is
    exhausted, see :func:`interleave_longest`.

    )rr�r�)r�r�r�r�rM�srMcGs"t�t|dti��}dd�|D�S)asReturn a new iterable yielding from each iterable in turn,
    skipping any that are exhausted.

        >>> list(interleave_longest([1, 2, 3], [4, 5], [6, 7, 8]))
        [1, 4, 6, 2, 5, 7, 3, 8]

    This function produces the same output as :func:`roundrobin`, but may
    perform better for some inputs (in particular when the number of iterables
    is large).

    r�css|]}|tur|VqdSr�)r�)r�rr�r�r�r��r�z%interleave_longest.<locals>.<genexpr>)rr�rr�)r�r�r�r�r�rL�srLc#s$���fdd���|d�EdHdS)a>Flatten an iterable with multiple levels of nesting (e.g., a list of
    lists of tuples) into non-iterable types.

        >>> iterable = [(1, 2), ([3, 4], [[5], [6]])]
        >>> list(collapse(iterable))
        [1, 2, 3, 4, 5, 6]

    Binary and text strings are not considered iterable and
    will not be collapsed.

    To avoid collapsing other types, specify *base_type*:

        >>> iterable = ['ab', ('cd', 'ef'), ['gh', 'ij']]
        >>> list(collapse(iterable, base_type=tuple))
        ['ab', ('cd', 'ef'), 'gh', 'ij']

    Specify *levels* to stop flattening after a certain level:

    >>> iterable = [('a', ['b']), ('c', ['d'])]
    >>> list(collapse(iterable))  # Fully flattened
    ['a', 'b', 'c', 'd']
    >>> list(collapse(iterable, levels=1))  # Only one level flattened
    ['a', ['b'], 'c', ['d']]

    c3s��dur|�ks0t|ttf�s0�dur:t|��r:|VdSzt|�}Wnty`|VYdS0|D]}�||d�EdHqfdS�Nr,)r��str�bytesr�r�)�node�level�tree�child��	base_type�levels�walkr�r�rs&�����zcollapse.<locals>.walkrNr�)r�rrr�rr�r;�sr;ccszzd|dur|�|dur2|D]}||�|Vqn"t||�D]}||�|EdHq<W|durv|�n|durt|�0dS)auInvoke *func* on each item in *iterable* (or on each *chunk_size* group
    of items) before yielding the item.

    `func` must be a function that takes a single argument. Its return value
    will be discarded.

    *before* and *after* are optional functions that take no arguments. They
    will be executed before iteration starts and after it ends, respectively.

    `side_effect` can be used for logging, updating progress bars, or anything
    that is not functionally "pure."

    Emitting a status message:

        >>> from more_itertools import consume
        >>> func = lambda item: print('Received {}'.format(item))
        >>> consume(side_effect(func, range(2)))
        Received 0
        Received 1

    Operating on chunks of items:

        >>> pair_sums = []
        >>> func = lambda chunk: pair_sums.append(sum(chunk))
        >>> list(side_effect(func, [0, 1, 2, 3, 4, 5], 2))
        [0, 1, 2, 3, 4, 5]
        >>> list(pair_sums)
        [1, 5, 9]

    Writing to a file-like object:

        >>> from io import StringIO
        >>> from more_itertools import consume
        >>> f = StringIO()
        >>> func = lambda x: print(x, file=f)
        >>> before = lambda: print(u'HEADER', file=f)
        >>> after = f.close
        >>> it = [u'a', u'b', u'c']
        >>> consume(side_effect(func, it, before=before, after=after))
        >>> f.closed
        True

    N)r9)r�r��
chunk_size�before�afterr�r�r�r�r�rk,s,
�rkcsDtt��fdd�td��D���|r<��fdd�}t|��S�SdS)apYield slices of length *n* from the sequence *seq*.

    >>> list(sliced((1, 2, 3, 4, 5, 6), 3))
    [(1, 2, 3), (4, 5, 6)]

    By the default, the last yielded slice will have fewer than *n* elements
    if the length of *seq* is not divisible by *n*:

    >>> list(sliced((1, 2, 3, 4, 5, 6, 7, 8), 3))
    [(1, 2, 3), (4, 5, 6), (7, 8)]

    If the length of *seq* is not divisible by *n* and *strict* is
    ``True``, then ``ValueError`` will be raised before the last
    slice is yielded.

    This function will only work for iterables that support slicing.
    For non-sliceable iterables, see :func:`chunked`.

    c3s|]}�||��VqdSr�r��r�r�)r�r�r�r�r�}r�zsliced.<locals>.<genexpr>rc3s(�D]}t|��krtd��|VqdS)Nzseq is not divisible by n.r�)Z_slicer�r�r�r��szsliced.<locals>.retN)rr�rr�)r�r�r�r�r�)r�r�r�r�rlis
 
rlr�ccs�|dkrt|�VdSg}t|�}|D]N}||�rj|V|rD|gV|dkr\t|�VdSg}|d8}q&|�|�q&|VdS)a<Yield lists of items from *iterable*, where each list is delimited by
    an item where callable *pred* returns ``True``.

        >>> list(split_at('abcdcba', lambda x: x == 'b'))
        [['a'], ['c', 'd', 'c'], ['a']]

        >>> list(split_at(range(10), lambda n: n % 2 == 1))
        [[0], [2], [4], [6], [8], []]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_at(range(10), lambda n: n % 2 == 1, maxsplit=2))
        [[0], [2], [4, 5, 6, 7, 8, 9]]

    By default, the delimiting items are not included in the output.
    The include them, set *keep_separator* to ``True``.

        >>> list(split_at('abcdcba', lambda x: x == 'b', keep_separator=True))
        [['a'], ['b'], ['c', 'd', 'c'], ['b'], ['a']]

    rNr,�r�r�r�)r��pred�maxsplitZkeep_separator�bufr�r�r�r�r�rn�s"


rnccs�|dkrt|�VdSg}t|�}|D]J}||�rf|rf|V|dkrZ|gt|�VdSg}|d8}|�|�q&|r||VdS)a\Yield lists of items from *iterable*, where each list ends just before
    an item for which callable *pred* returns ``True``:

        >>> list(split_before('OneTwo', lambda s: s.isupper()))
        [['O', 'n', 'e'], ['T', 'w', 'o']]

        >>> list(split_before(range(10), lambda n: n % 3 == 0))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8], [9]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_before(range(10), lambda n: n % 3 == 0, maxsplit=2))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8, 9]]
    rNr,r�r�r r!r"r�r�r�r�r�rp�s 
rpccsz|dkrt|�VdSg}t|�}|D]D}|�|�||�r&|r&|V|dkr^t|�VdSg}|d8}q&|rv|VdS)a[Yield lists of items from *iterable*, where each list ends with an
    item where callable *pred* returns ``True``:

        >>> list(split_after('one1two2', lambda s: s.isdigit()))
        [['o', 'n', 'e', '1'], ['t', 'w', 'o', '2']]

        >>> list(split_after(range(10), lambda n: n % 3 == 0))
        [[0], [1, 2, 3], [4, 5, 6], [7, 8, 9]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_after(range(10), lambda n: n % 3 == 0, maxsplit=2))
        [[0], [1, 2, 3], [4, 5, 6, 7, 8, 9]]

    rNr,rr#r�r�r�ro�s 



roccs�|dkrt|�VdSt|�}zt|�}Wnty>YdS0|g}|D]L}|||�r�|V|dkr||gt|�VdSg}|d8}|�|�|}qJ|VdS)a�Split *iterable* into pieces based on the output of *pred*.
    *pred* should be a function that takes successive pairs of items and
    returns ``True`` if the iterable should be split in between them.

    For example, to find runs of increasing numbers, split the iterable when
    element ``i`` is larger than element ``i + 1``:

        >>> list(split_when([1, 2, 3, 3, 2, 5, 2, 4, 2], lambda x, y: x > y))
        [[1, 2, 3, 3], [2, 5], [2, 4], [2]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_when([1, 2, 3, 3, 2, 5, 2, 4, 2],
        ...                 lambda x, y: x > y, maxsplit=2))
        [[1, 2, 3, 3], [2, 5], [2, 4, 2]]

    rNr,)r�r�r�r�r�)r�r r!r�Zcur_itemr"Z	next_itemr�r�r�rq�s(


rqccs>t|�}|D],}|dur(t|�VdStt||��VqdS)a�Yield a list of sequential items from *iterable* of length 'n' for each
    integer 'n' in *sizes*.

        >>> list(split_into([1,2,3,4,5,6], [1,2,3]))
        [[1], [2, 3], [4, 5, 6]]

    If the sum of *sizes* is smaller than the length of *iterable*, then the
    remaining items of *iterable* will not be returned.

        >>> list(split_into([1,2,3,4,5,6], [2,3]))
        [[1, 2], [3, 4, 5]]

    If the sum of *sizes* is larger than the length of *iterable*, fewer items
    will be returned in the iteration that overruns *iterable* and further
    lists will be empty:

        >>> list(split_into([1,2,3,4], [1,2,3,4]))
        [[1], [2, 3], [4], []]

    When a ``None`` object is encountered in *sizes*, the returned list will
    contain items up to the end of *iterable* the same way that itertools.slice
    does:

        >>> list(split_into([1,2,3,4,5,6,7,8,9,0], [2,3,None]))
        [[1, 2], [3, 4, 5], [6, 7, 8, 9, 0]]

    :func:`split_into` can be useful for grouping a series of items where the
    sizes of the groups are not uniform. An example would be where in a row
    from a table, multiple columns represent elements of the same feature
    (e.g. a point represented by x,y,z) but, the format is not the same for
    all columns.
    N)r�r�r)r��sizesr�r�r�r�r�rr+s#
rrc	cs�t|�}|dur&t|t|��EdHnZ|dkr8td��nHd}|D]}|V|d7}q@|rd|||n||}t|�D]
}|VqtdS)a�Yield the elements from *iterable*, followed by *fillvalue*, such that
    at least *n* items are emitted.

        >>> list(padded([1, 2, 3], '?', 5))
        [1, 2, 3, '?', '?']

    If *next_multiple* is ``True``, *fillvalue* will be emitted until the
    number of items emitted is a multiple of *n*::

        >>> list(padded([1, 2, 3, 4], n=3, next_multiple=True))
        [1, 2, 3, 4, None, None]

    If *n* is ``None``, *fillvalue* will be emitted indefinitely.

    Nr,�n must be at least 1r)r�rrr�r�)	r�r�r�Z
next_multipler�r�r��	remainingr�r�r�r�r_Xs

r_ccs6t}|D]
}|Vq|tur |n|}t|�EdHdS)a"After the *iterable* is exhausted, keep yielding its last element.

        >>> list(islice(repeat_last(range(3)), 5))
        [0, 1, 2, 2, 2]

    If the iterable is empty, yield *default* forever::

        >>> list(islice(repeat_last(range(0), 42), 5))
        [42, 42, 42, 42, 42]

    N)r�r)r�r�r��finalr�r�r�rcxs
rccs0�dkrtd��t|��}�fdd�t|�D�S)a�Distribute the items from *iterable* among *n* smaller iterables.

        >>> group_1, group_2 = distribute(2, [1, 2, 3, 4, 5, 6])
        >>> list(group_1)
        [1, 3, 5]
        >>> list(group_2)
        [2, 4, 6]

    If the length of *iterable* is not evenly divisible by *n*, then the
    length of the returned iterables will not be identical:

        >>> children = distribute(3, [1, 2, 3, 4, 5, 6, 7])
        >>> [list(c) for c in children]
        [[1, 4, 7], [2, 5], [3, 6]]

    If the length of *iterable* is smaller than *n*, then the last returned
    iterables will be empty:

        >>> children = distribute(5, [1, 2, 3])
        >>> [list(c) for c in children]
        [[1], [2], [3], [], []]

    This function uses :func:`itertools.tee` and may require significant
    storage. If you need the order items in the smaller iterables to match the
    original iterable, see :func:`divide`.

    r,r%csg|]\}}t||d���qSr�)r)r�r�r��r�r�r�r��r�zdistribute.<locals>.<listcomp>)r�r�	enumerate)r�r��childrenr�r(r�rE�s
rE�r�rr,cCs t|t|��}t||||d��S)a[Yield tuples whose elements are offset from *iterable*.
    The amount by which the `i`-th item in each tuple is offset is given by
    the `i`-th item in *offsets*.

        >>> list(stagger([0, 1, 2, 3]))
        [(None, 0, 1), (0, 1, 2), (1, 2, 3)]
        >>> list(stagger(range(8), offsets=(0, 2, 4)))
        [(0, 2, 4), (1, 3, 5), (2, 4, 6), (3, 5, 7)]

    By default, the sequence will end when the final element of a tuple is the
    last item in the iterable. To continue until the first element of a tuple
    is the last item in the iterable, set *longest* to ``True``::

        >>> list(stagger([0, 1, 2, 3], longest=True))
        [(None, 0, 1), (0, 1, 2), (1, 2, 3), (2, 3, None), (3, None, None)]

    By default, ``None`` will be used to replace offsets beyond the end of the
    sequence. Specify *fillvalue* to use some other value.

    )�offsets�longestr�)rr�r)r�r,r-r�r*r�r�r�rt�s�rtcseZdZd�fdd�	Z�ZS)r}Ncs*d}|dur|dj|�7}t��|�dS)Nz Iterables have different lengthsz/: index 0 has length {}; index {} has length {})r��superr�)r��detailsr���	__class__r�r�r��s�zUnequalIterablesError.__init__)N)r�r�r�r��
__classcell__r�r�r0r�r}�sr}ccs6t|dti�D]"}|D]}|turt��q|VqdS)Nr�)rr�r})r�Zcombo�valr�r�r��_zip_equal_generator�s
r4cGs�tdkrt�dt�zZt|d�}t|dd�d�D]\}}t|�}||kr4q\q4t|�WSt|||fd��Wnty�t	|�YS0dS)a ``zip`` the input *iterables* together, but raise
    ``UnequalIterablesError`` if they aren't all the same length.

        >>> it_1 = range(3)
        >>> it_2 = iter('abc')
        >>> list(zip_equal(it_1, it_2))
        [(0, 'a'), (1, 'b'), (2, 'c')]

        >>> it_1 = range(3)
        >>> it_2 = iter('abcd')
        >>> list(zip_equal(it_1, it_2)) # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        more_itertools.more.UnequalIterablesError: Iterables have different
        lengths

    i�
zwzip_equal will be removed in a future version of more-itertools. Use the builtin zip function with strict=True instead.rr,N)r/)
r)r�r�r�r�r)r�r}r�r4)r�Z
first_sizer�r�r�r�r�r�r~�s�	
r~)r-r�cGs�t|�t|�krtd��g}t||�D]P\}}|dkrP|�tt||�|��q&|dkrl|�t||d��q&|�|�q&|r�t|d|i�St|�S)aF``zip`` the input *iterables* together, but offset the `i`-th iterable
    by the `i`-th item in *offsets*.

        >>> list(zip_offset('0123', 'abcdef', offsets=(0, 1)))
        [('0', 'b'), ('1', 'c'), ('2', 'd'), ('3', 'e')]

    This can be used as a lightweight alternative to SciPy or pandas to analyze
    data sets in which some series have a lead or lag relationship.

    By default, the sequence will end when the shortest iterable is exhausted.
    To continue until the longest iterable is exhausted, set *longest* to
    ``True``.

        >>> list(zip_offset('0123', 'abcdef', offsets=(0, 1), longest=True))
        [('0', 'b'), ('1', 'c'), ('2', 'd'), ('3', 'e'), (None, 'f')]

    By default, ``None`` will be used to replace offsets beyond the end of the
    sequence. Specify *fillvalue* to use some other value.

    z,Number of iterables and offsets didn't matchrNr�)r�r�r�r�rrrr)r,r-r�r��	staggeredr�r�r�r�r�rsr�rcsn�durt|�}nBt|�}t|�dkr>|d���fdd�}nt|����fdd�}tttt|�||d���S)a�Return the input iterables sorted together, with *key_list* as the
    priority for sorting. All iterables are trimmed to the length of the
    shortest one.

    This can be used like the sorting function in a spreadsheet. If each
    iterable represents a column of data, the key list determines which
    columns are used for sorting.

    By default, all iterables are sorted using the ``0``-th iterable::

        >>> iterables = [(4, 3, 2, 1), ('a', 'b', 'c', 'd')]
        >>> sort_together(iterables)
        [(1, 2, 3, 4), ('d', 'c', 'b', 'a')]

    Set a different key list to sort according to another iterable.
    Specifying multiple keys dictates how ties are broken::

        >>> iterables = [(3, 1, 2), (0, 1, 0), ('c', 'b', 'a')]
        >>> sort_together(iterables, key_list=(1, 2))
        [(2, 3, 1), (0, 0, 1), ('a', 'c', 'b')]

    To sort by a function of the elements of the iterable, pass a *key*
    function. Its arguments are the elements of the iterables corresponding to
    the key list::

        >>> names = ('a', 'b', 'c')
        >>> lengths = (1, 2, 3)
        >>> widths = (5, 2, 1)
        >>> def area(length, width):
        ...     return length * width
        >>> sort_together([names, lengths, widths], key_list=(1, 2), key=area)
        [('c', 'b', 'a'), (3, 2, 1), (1, 2, 5)]

    Set *reverse* to ``True`` to sort in descending order.

        >>> sort_together([(1, 2, 3), ('c', 'b', 'a')], reverse=True)
        [(3, 2, 1), ('a', 'b', 'c')]

    Nr,rcs�|��Sr�r��Zzipped_items)r�
key_offsetr�r�rfr�zsort_together.<locals>.<lambda>cs��|��Sr�r�r7)�
get_key_itemsrr�r�rks�)rr)r$r�r�r�r�)r�Zkey_listrrZkey_argumentr�)r9rr8r�rm2s(
�rmcsPtt|��\}}|sdS|d}t|t|��}dd��t�fdd�t|�D��S)a�The inverse of :func:`zip`, this function disaggregates the elements
    of the zipped *iterable*.

    The ``i``-th iterable contains the ``i``-th element from each element
    of the zipped iterable. The first element is used to to determine the
    length of the remaining elements.

        >>> iterable = [('a', 1), ('b', 2), ('c', 3), ('d', 4)]
        >>> letters, numbers = unzip(iterable)
        >>> list(letters)
        ['a', 'b', 'c', 'd']
        >>> list(numbers)
        [1, 2, 3, 4]

    This is similar to using ``zip(*iterable)``, but it avoids reading
    *iterable* into memory. Note, however, that this function uses
    :func:`itertools.tee` and thus may require significant storage.

    r�rcs�fdd�}|S)Ncs&z
|�WSty t�Yn0dSr�)r�r�)�obj�r�r�r��getter�s

z)unzip.<locals>.itemgetter.<locals>.getterr�)r�r<r�r;r�r$�szunzip.<locals>.itemgetterc3s |]\}}t�|�|�VqdSr��r�)r�r�r��r$r�r�r��r�zunzip.<locals>.<genexpr>)rsr�rr�r�r))r�r�r�r�r>r�rztsrzc	Cs�|dkrtd��z|dd�Wnty:t|�}Yn0|}tt|�|�\}}g}d}td|d�D]6}|}|||kr�|dn|7}|�t|||���qh|S)a�Divide the elements from *iterable* into *n* parts, maintaining
    order.

        >>> group_1, group_2 = divide(2, [1, 2, 3, 4, 5, 6])
        >>> list(group_1)
        [1, 2, 3]
        >>> list(group_2)
        [4, 5, 6]

    If the length of *iterable* is not evenly divisible by *n*, then the
    length of the returned iterables will not be identical:

        >>> children = divide(3, [1, 2, 3, 4, 5, 6, 7])
        >>> [list(c) for c in children]
        [[1, 2, 3], [4, 5], [6, 7]]

    If the length of the iterable is smaller than n, then the last returned
    iterables will be empty:

        >>> children = divide(5, [1, 2, 3])
        >>> [list(c) for c in children]
        [[1], [2], [3], [], []]

    This function will exhaust the iterable before returning and may require
    significant storage. If order is not important, see :func:`distribute`,
    which does not first pull the iterable into memory.

    r,r%Nr)r�r�r��divmodr�r�r�r�)	r�r�r��qr�r�r�r�r�r�r�r�rF�srFcCsX|durtd�S|dur,t||�r,t|f�Sz
t|�WStyRt|f�YS0dS)axIf *obj* is iterable, return an iterator over its items::

        >>> obj = (1, 2, 3)
        >>> list(always_iterable(obj))
        [1, 2, 3]

    If *obj* is not iterable, return a one-item iterable containing *obj*::

        >>> obj = 1
        >>> list(always_iterable(obj))
        [1]

    If *obj* is ``None``, return an empty iterable:

        >>> obj = None
        >>> list(always_iterable(None))
        []

    By default, binary and text strings are not considered iterable::

        >>> obj = 'foo'
        >>> list(always_iterable(obj))
        ['foo']

    If *base_type* is set, objects for which ``isinstance(obj, base_type)``
    returns ``True`` won't be considered iterable.

        >>> obj = {'a': 1}
        >>> list(always_iterable(obj))  # Iterate over the dict's keys
        ['a']
        >>> list(always_iterable(obj, base_type=dict))  # Treat dicts as a unit
        [{'a': 1}]

    Set *base_type* to ``None`` to avoid any special handling and treat objects
    Python considers iterable as iterable:

        >>> obj = 'foo'
        >>> list(always_iterable(obj, base_type=None))
        ['f', 'o', 'o']
    Nr�)r�r�r�)r:rr�r�r�r5�s)

r5cCsZ|dkrtd��t|�\}}dg|}t|t||�|�}ttt|d|d��}t||�S)asReturn an iterable over `(bool, item)` tuples where the `item` is
    drawn from *iterable* and the `bool` indicates whether
    that item satisfies the *predicate* or is adjacent to an item that does.

    For example, to find whether items are adjacent to a ``3``::

        >>> list(adjacent(lambda x: x == 3, range(6)))
        [(False, 0), (False, 1), (True, 2), (True, 3), (True, 4), (False, 5)]

    Set *distance* to change what counts as adjacent. For example, to find
    whether items are two places away from a ``3``:

        >>> list(adjacent(lambda x: x == 3, range(6), distance=2))
        [(False, 0), (True, 1), (True, 2), (True, 3), (True, 4), (True, 5)]

    This is useful for contextualizing the results of a search function.
    For example, a code comparison tool might want to identify lines that
    have changed, but also surrounding lines to give the viewer of the diff
    context.

    The predicate function will only be called once for each item in the
    iterable.

    See also :func:`groupby_transform`, which can be used with this function
    to group ranges of items with the same `bool` value.

    rzdistance must be at least 0Fr�r,)r�rrr��anyr{r�)�	predicater��distance�i1�i2�padding�selectedZadjacent_to_selectedr�r�r�r4
s
r4cs:t||�}�r �fdd�|D�}�r6�fdd�|D�}|S)aAn extension of :func:`itertools.groupby` that can apply transformations
    to the grouped data.

    * *keyfunc* is a function computing a key value for each item in *iterable*
    * *valuefunc* is a function that transforms the individual items from
      *iterable* after grouping
    * *reducefunc* is a function that transforms each group of items

    >>> iterable = 'aAAbBBcCC'
    >>> keyfunc = lambda k: k.upper()
    >>> valuefunc = lambda v: v.lower()
    >>> reducefunc = lambda g: ''.join(g)
    >>> list(groupby_transform(iterable, keyfunc, valuefunc, reducefunc))
    [('A', 'aaa'), ('B', 'bbb'), ('C', 'ccc')]

    Each optional argument defaults to an identity function if not specified.

    :func:`groupby_transform` is useful when grouping elements of an iterable
    using a separate iterable as the key. To do this, :func:`zip` the iterables
    and pass a *keyfunc* that extracts the first element and a *valuefunc*
    that extracts the second element::

        >>> from operator import itemgetter
        >>> keys = [0, 0, 1, 1, 1, 2, 2, 2, 3]
        >>> values = 'abcdefghi'
        >>> iterable = zip(keys, values)
        >>> grouper = groupby_transform(iterable, itemgetter(0), itemgetter(1))
        >>> [(k, ''.join(g)) for k, g in grouper]
        [(0, 'ab'), (1, 'cde'), (2, 'fgh'), (3, 'i')]

    Note that the order of items in the iterable is significant.
    Only adjacent items are grouped together, so if you don't want any
    duplicate groups, you should sort the iterable by the key function.

    c3s |]\}}|t�|�fVqdSr�r=�r��k�g)�	valuefuncr�r�r�Zr�z$groupby_transform.<locals>.<genexpr>c3s|]\}}|�|�fVqdSr�r�rH)�
reducefuncr�r�r�\r��r)r��keyfuncrKrLr�r�)rLrKr�rJ4s$
rJc@s�eZdZdZeedd��Zdd�Zdd�Zdd�Z	d	d
�Z
dd�Zd
d�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!S)"r\a<An extension of the built-in ``range()`` function whose arguments can
    be any orderable numeric type.

    With only *stop* specified, *start* defaults to ``0`` and *step*
    defaults to ``1``. The output items will match the type of *stop*:

        >>> list(numeric_range(3.5))
        [0.0, 1.0, 2.0, 3.0]

    With only *start* and *stop* specified, *step* defaults to ``1``. The
    output items will match the type of *start*:

        >>> from decimal import Decimal
        >>> start = Decimal('2.1')
        >>> stop = Decimal('5.1')
        >>> list(numeric_range(start, stop))
        [Decimal('2.1'), Decimal('3.1'), Decimal('4.1')]

    With *start*, *stop*, and *step*  specified the output items will match
    the type of ``start + step``:

        >>> from fractions import Fraction
        >>> start = Fraction(1, 2)  # Start at 1/2
        >>> stop = Fraction(5, 2)  # End at 5/2
        >>> step = Fraction(1, 2)  # Count by 1/2
        >>> list(numeric_range(start, stop, step))
        [Fraction(1, 2), Fraction(1, 1), Fraction(3, 2), Fraction(2, 1)]

    If *step* is zero, ``ValueError`` is raised. Negative steps are supported:

        >>> list(numeric_range(3, -1, -1.0))
        [3.0, 2.0, 1.0, 0.0]

    Be aware of the limitations of floating point numbers; the representation
    of the yielded numbers may be surprising.

    ``datetime.datetime`` objects can be used for *start* and *stop*, if *step*
    is a ``datetime.timedelta`` object:

        >>> import datetime
        >>> start = datetime.datetime(2019, 1, 1)
        >>> stop = datetime.datetime(2019, 1, 3)
        >>> step = datetime.timedelta(days=1)
        >>> items = iter(numeric_range(start, stop, step))
        >>> next(items)
        datetime.datetime(2019, 1, 1, 0, 0)
        >>> next(items)
        datetime.datetime(2019, 1, 2, 0, 0)

    rcGs�t|�}|dkr@|\|_t|j�d�|_t|j|j�d�|_nl|dkrl|\|_|_t|j|j�d�|_n@|dkr�|\|_|_|_n&|dkr�td�|���ntd�|���t|j�d�|_|j|jkr�td��|j|jk|_	|�
�dS)Nr,rr��z2numeric_range expected at least 1 argument, got {}z2numeric_range expected at most 3 arguments, got {}z&numeric_range() arg 3 must not be zero)r��_stop�type�_start�_stepr�r��_zeror��_growing�	_init_len)r�r�Zargcr�r�r�r��s4����znumeric_range.__init__cCs"|jr|j|jkS|j|jkSdSr�)rUrRrPr�r�r�r�r��sznumeric_range.__bool__cCsr|jr:|j|kr|jkrnnqn||j|j|jkSn4|j|krR|jkrnnn|j||j|jkSdS�NF)rUrRrPrSrT)r��elemr�r�r�r��sznumeric_range.__contains__cCsdt|t�r\t|�}t|�}|s&|r.|o,|S|j|jkoX|j|jkoX|�d�|�d�kSndSdS)Nr�F)r�r\�boolrRrS�
_get_by_index)r��otherZ
empty_selfZempty_otherr�r�r��__eq__�s



��znumeric_range.__eq__cCs�t|t�r|�|�St|t�r�|jdur.|jn
|j|j}|jdusR|j|jkrZ|j}n |j|jkrn|j	}n|�|j�}|j
dus�|j
|jkr�|j	}n"|j
|jkr�|j}n|�|j
�}t|||�Std�
t|�j���dS)Nz8numeric range indices must be integers or slices, not {})r��intrZr�r�rSr��_lenrRrPr�r\r�r�rQr�)r�rr�r�r�r�r�r�r��s(


��znumeric_range.__getitem__cCs&|rt|j|�d�|jf�S|jSdS�Nr�)�hashrRrZrS�_EMPTY_HASHr�r�r�r��__hash__�sznumeric_range.__hash__csB�fdd�t�D�}�jr,ttt�j�|�Sttt�j�|�SdS)Nc3s|]}�j|�jVqdSr�)rRrS)r�r�r�r�r�r��r�z)numeric_range.__iter__.<locals>.<genexpr>)rrUrrr'rPr()r��valuesr�r�r�r��sznumeric_range.__iter__cCs|jSr�)r^r�r�r�r��__len__�sznumeric_range.__len__cCsr|jr|j}|j}|j}n|j}|j}|j}||}||jkrHd|_n&t||�\}}t|�t||jk�|_dS�Nr)rUrRrPrSrTr^r?r])r�r�r�r�rCr@r�r�r�r�rV�s
znumeric_range._init_lencCst|j|j|jffSr�)r\rRrPrSr�r�r�r��
__reduce__
sznumeric_range.__reduce__cCsF|jdkr"d�t|j�t|j��Sd�t|j�t|j�t|j��SdS)Nr,znumeric_range({}, {})znumeric_range({}, {}, {}))rSr��reprrRrPr�r�r�r��__repr__s
��znumeric_range.__repr__cCs"tt|�d�|j|j|j��Sr_)r�r\rZrRrSr�r�r�r�r�s
��znumeric_range.__reversed__cCst||v�Sr�)r]rr�r�r�r!sznumeric_range.countcCs�|jrL|j|kr|jkr�nq�t||j|j�\}}||jkr�t|�SnF|j|krd|jkr�nn*t|j||j�\}}||jkr�t|�Std�|���dS)Nz{} is not in numeric range)	rUrRrPr?rSrTr]r�r�)r�r
r@r�r�r�r�r�$s


znumeric_range.indexcCs<|dkr||j7}|dks$||jkr,td��|j||jS)Nrz'numeric range object index out of range)r^r�rRrS)r�r�r�r�r�rZ2s

znumeric_range._get_by_indexN)r�r�r�r�r`r�rar�r�r�r\r�rbr�rdrVrfrhr�rr�rZr�r�r�r�r\as"3

r\cs<t����std�S|dur"t�nt|�}�fdd�|D�S)aCycle through the items from *iterable* up to *n* times, yielding
    the number of completed cycles along with each item. If *n* is omitted the
    process repeats indefinitely.

    >>> list(count_cycle('AB', 3))
    [(0, 'A'), (0, 'B'), (1, 'A'), (1, 'B'), (2, 'A'), (2, 'B')]

    r�Nc3s |]}�D]}||fVq
qdSr�r�)r�r�r��r�r�r�r�Gr�zcount_cycle.<locals>.<genexpr>)r�r�rr�)r�r�r�r�rir�r@:s
	r@ccs~t|�}zt|�}Wnty(YdS0z,t�D] }|}t|�}|dkd|fVq2Wn"tyx|dkd|fVYn0dS)aHYield 3-tuples of the form ``(is_first, is_last, item)``.

    >>> list(mark_ends('ABC'))
    [(True, False, 'A'), (False, False, 'B'), (False, True, 'C')]

    Use this when looping over an iterable to take special action on its first
    and/or last items:

    >>> iterable = ['Header', 100, 200, 'Footer']
    >>> total = 0
    >>> for is_first, is_last, item in mark_ends(iterable):
    ...     if is_first:
    ...         continue  # Skip the header
    ...     if is_last:
    ...         continue  # Skip the footer
    ...     total += item
    >>> print(total)
    300
    NrFT)r�r�r�r)r�r��br��ar�r�r�rAJs
rAcCsJ|durtt�t||��S|dkr*td��t||td�}tt�t||��S)a�Yield the index of each item in *iterable* for which *pred* returns
    ``True``.

    *pred* defaults to :func:`bool`, which will select truthy items:

        >>> list(locate([0, 1, 1, 0, 1, 0, 0]))
        [1, 2, 4]

    Set *pred* to a custom function to, e.g., find the indexes for a particular
    item.

        >>> list(locate(['a', 'b', 'c', 'b'], lambda x: x == 'b'))
        [1, 3]

    If *window_size* is given, then the *pred* function will be called with
    that many items. This enables searching for sub-sequences:

        >>> iterable = [0, 1, 2, 3, 0, 1, 2, 3, 0, 1, 2, 3]
        >>> pred = lambda *args: args == (1, 2, 3)
        >>> list(locate(iterable, pred=pred, window_size=3))
        [1, 5, 9]

    Use with :func:`seekable` to find indexes and then retrieve the associated
    items:

        >>> from itertools import count
        >>> from more_itertools import seekable
        >>> source = (3 * n + 1 if (n % 2) else n // 2 for n in count())
        >>> it = seekable(source)
        >>> pred = lambda x: x > 100
        >>> indexes = locate(it, pred=pred)
        >>> i = next(indexes)
        >>> it.seek(i)
        >>> next(it)
        106

    Nr,zwindow size must be at least 1�r�)rrr�r�r{r�r)r�r �window_sizer�r�r�r�rTos&rTcCs
t||�S)a�Yield the items from *iterable*, but strip any from the beginning
    for which *pred* returns ``True``.

    For example, to remove a set of items from the start of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(lstrip(iterable, pred))
        [1, 2, None, 3, False, None]

    This function is analogous to to :func:`str.lstrip`, and is essentially
    an wrapper for :func:`itertools.dropwhile`.

    )r�r�r r�r�r�rU�srUccsFg}|j}|j}|D],}||�r*||�q|EdH|�|VqdS)a�Yield the items from *iterable*, but strip any from the end
    for which *pred* returns ``True``.

    For example, to remove a set of items from the end of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(rstrip(iterable, pred))
        [None, False, None, 1, 2, None, 3]

    This function is analogous to :func:`str.rstrip`.

    N)r��clear)r�r �cacheZcache_append�cache_clearrr�r�r�rf�s

rfcCstt||�|�S)a�Yield the items from *iterable*, but strip any from the
    beginning and end for which *pred* returns ``True``.

    For example, to remove a set of items from both ends of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(strip(iterable, pred))
        [1, 2, None, 3]

    This function is analogous to :func:`str.strip`.

    )rfrUrnr�r�r�ru�sruc@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)rOaAn extension of :func:`itertools.islice` that supports negative values
    for *stop*, *start*, and *step*.

        >>> iterable = iter('abcdefgh')
        >>> list(islice_extended(iterable, -4, -1))
        ['e', 'f', 'g']

    Slices with negative values require some caching of *iterable*, but this
    function takes care to minimize the amount of memory required.

    For example, you can use a negative step with an infinite iterator:

        >>> from itertools import count
        >>> list(islice_extended(count(), 110, 99, -2))
        [110, 108, 106, 104, 102, 100]

    You can also use slice notation directly:

        >>> iterable = map(str, count())
        >>> it = islice_extended(iterable)[10:20:2]
        >>> list(it)
        ['10', '12', '14', '16', '18']

    cGs(t|�}|rt|t|��|_n||_dSr�)r��_islice_helperr��	_iterable)r�r�r�r�r�r�r�r��szislice_extended.__init__cCs|Sr�r�r�r�r�r�r��szislice_extended.__iter__cCs
t|j�Sr�)r�rsr�r�r�r�r�	szislice_extended.__next__cCs&t|t�rtt|j|��Std��dS)Nz4islice_extended.__getitem__ argument must be a slice)r�r�rOrrrsr�)r�rr�r�r�r�	s
zislice_extended.__getitem__N)r�r�r�r�r�r�r�r�r�r�r�r�rO�s
rOccs�|j}|j}|jdkrtd��|jp&d}|dk�rt|dur>dn|}|dkr�tt|d�|d�}|rn|ddnd}t||d�}|dur�|}n"|dkr�t||�}nt||d�}||}	|	dkr�dSt|d|	|�D]\}
}|Vq�n�|du�r\|dk�r\t	t|||�d�tt||�|d�}t|�D]0\}
}|�
�}|
|dk�rL|V|�|��q(nt||||�EdH�n4|du�r�dn|}|du�r|dk�r|d}	tt|d�|	d�}|�r�|ddnd}|dk�r�||}}nt||d�d}}t|�|||�D]\}
}|V�qn�|du�r@|d}
t	t||
|
�d�|dk�rT|}d}	n2|du�rld}|d}	nd}||}	|	dk�r�dStt||	��}||d|�EdHdS)Nrz1step argument must be a non-zero integer or None.r,r�r�)
r�r�r�r�rr)r�r�rr�r�r�r�)r��sr�r�r�rp�len_iterr�r�r�r�r�Zcached_item�mr�r�r�rr
	sn










rrcCs.z
t|�WSty(tt|��YS0dS)a�An extension of :func:`reversed` that supports all iterables, not
    just those which implement the ``Reversible`` or ``Sequence`` protocols.

        >>> print(*always_reversible(x for x in range(3)))
        2 1 0

    If the iterable is already reversible, this function returns the
    result of :func:`reversed()`. If the iterable is not reversible,
    this function will cache the remaining items in the iterable and
    yield them in reverse order, which may require significant storage.
    N)r�r�r�rir�r�r�r6j	s
r6cCs|Sr�r�rr�r�r�r|	r�rc#s6tt|��fdd�d�D]\}}ttd�|�VqdS)a�Yield groups of consecutive items using :func:`itertools.groupby`.
    The *ordering* function determines whether two items are adjacent by
    returning their position.

    By default, the ordering function is the identity function. This is
    suitable for finding runs of numbers:

        >>> iterable = [1, 10, 11, 12, 20, 30, 31, 32, 33, 40]
        >>> for group in consecutive_groups(iterable):
        ...     print(list(group))
        [1]
        [10, 11, 12]
        [20]
        [30, 31, 32, 33]
        [40]

    For finding runs of adjacent letters, try using the :meth:`index` method
    of a string of letters:

        >>> from string import ascii_lowercase
        >>> iterable = 'abcdfgilmnop'
        >>> ordering = ascii_lowercase.index
        >>> for group in consecutive_groups(iterable, ordering):
        ...     print(list(group))
        ['a', 'b', 'c', 'd']
        ['f', 'g']
        ['i']
        ['l', 'm', 'n', 'o', 'p']

    Each group of consecutive items is an iterator that shares it source with
    *iterable*. When an an output group is advanced, the previous group is
    no longer available unless its elements are copied (e.g., into a ``list``).

        >>> iterable = [1, 2, 11, 12, 21, 22]
        >>> saved_groups = []
        >>> for group in consecutive_groups(iterable):
        ...     saved_groups.append(list(group))  # Copy group elements
        >>> saved_groups
        [[1, 2], [11, 12], [21, 22]]

    cs|d�|d�Sr�r�r��orderingr�r�r�	r�z$consecutive_groups.<locals>.<lambda>�rr,N)rr)r�r$)r�rxrIrJr�rwr�r=|	s*�r=)�initialcCsXt|�\}}zt|�g}Wnty4tg�YS0|durBg}t|t|t||���S)a�This function is the inverse of :func:`itertools.accumulate`. By default
    it will compute the first difference of *iterable* using
    :func:`operator.sub`:

        >>> from itertools import accumulate
        >>> iterable = accumulate([0, 1, 2, 3, 4])  # produces 0, 1, 3, 6, 10
        >>> list(difference(iterable))
        [0, 1, 2, 3, 4]

    *func* defaults to :func:`operator.sub`, but other functions can be
    specified. They will be applied as follows::

        A, B, C, D, ... --> A, func(B, A), func(C, B), func(D, C), ...

    For example, to do progressive division:

        >>> iterable = [1, 2, 6, 24, 120]
        >>> func = lambda x, y: x // y
        >>> list(difference(iterable, func))
        [1, 2, 3, 4, 5]

    If the *initial* keyword is set, the first element will be skipped when
    computing successive differences.

        >>> it = [10, 11, 13, 16]  # from accumulate([1, 2, 3], initial=10)
        >>> list(difference(it, initial=10))
        [1, 2, 3]

    N)rr�r�r�rrr�)r�r�rzrkrjrIr�r�r�rB�	srBc@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)rjaSReturn a read-only view of the sequence object *target*.

    :class:`SequenceView` objects are analogous to Python's built-in
    "dictionary view" types. They provide a dynamic view of a sequence's items,
    meaning that when the sequence updates, so does the view.

        >>> seq = ['0', '1', '2']
        >>> view = SequenceView(seq)
        >>> view
        SequenceView(['0', '1', '2'])
        >>> seq.append('3')
        >>> view
        SequenceView(['0', '1', '2', '3'])

    Sequence views support indexing, slicing, and length queries. They act
    like the underlying sequence, except they don't allow assignment:

        >>> view[1]
        '1'
        >>> view[1:-1]
        ['1', '2']
        >>> len(view)
        4

    Sequence views are useful as an alternative to copying, as they don't
    require (much) extra storage.

    cCst|t�st�||_dSr�)r�rr��_target)r��targetr�r�r�r��	s
zSequenceView.__init__cCs
|j|Sr�)r{)r�r�r�r�r�r��	szSequenceView.__getitem__cCs
t|j�Sr�)r�r{r�r�r�r�rd�	szSequenceView.__len__cCsd�|jjt|j��S)Nz{}({}))r�r1r�rgr{r�r�r�r�rh�	szSequenceView.__repr__N)r�r�r�r�r�r�rdrhr�r�r�r�rj�	s
rjc@sNeZdZdZddd�Zdd�Zdd�Zd	d
�Zefdd�Z	d
d�Z
dd�ZdS)ria
Wrap an iterator to allow for seeking backward and forward. This
    progressively caches the items in the source iterable so they can be
    re-visited.

    Call :meth:`seek` with an index to seek to that position in the source
    iterable.

    To "reset" an iterator, seek to ``0``:

        >>> from itertools import count
        >>> it = seekable((str(n) for n in count()))
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> it.seek(0)
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> next(it)
        '3'

    You can also seek forward:

        >>> it = seekable((str(n) for n in range(20)))
        >>> it.seek(10)
        >>> next(it)
        '10'
        >>> it.seek(20)  # Seeking past the end of the source isn't a problem
        >>> list(it)
        []
        >>> it.seek(0)  # Resetting works even after hitting the end
        >>> next(it), next(it), next(it)
        ('0', '1', '2')

    Call :meth:`peek` to look ahead one item without advancing the iterator:

        >>> it = seekable('1234')
        >>> it.peek()
        '1'
        >>> list(it)
        ['1', '2', '3', '4']
        >>> it.peek(default='empty')
        'empty'

    Before the iterator is at its end, calling :func:`bool` on it will return
    ``True``. After it will return ``False``:

        >>> it = seekable('5678')
        >>> bool(it)
        True
        >>> list(it)
        ['5', '6', '7', '8']
        >>> bool(it)
        False

    You may view the contents of the cache with the :meth:`elements` method.
    That returns a :class:`SequenceView`, a view that updates automatically:

        >>> it = seekable((str(n) for n in range(10)))
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> elements = it.elements()
        >>> elements
        SequenceView(['0', '1', '2'])
        >>> next(it)
        '3'
        >>> elements
        SequenceView(['0', '1', '2', '3'])

    By default, the cache grows as the source iterable progresses, so beware of
    wrapping very large or infinite iterables. Supply *maxlen* to limit the
    size of the cache (this of course limits how far back you can seek).

        >>> from itertools import count
        >>> it = seekable((str(n) for n in count()), maxlen=2)
        >>> next(it), next(it), next(it), next(it)
        ('0', '1', '2', '3')
        >>> list(it.elements())
        ['2', '3']
        >>> it.seek(0)
        >>> next(it), next(it), next(it), next(it)
        ('2', '3', '4', '5')
        >>> next(it)
        '6'

    NcCs0t|�|_|durg|_ntg|�|_d|_dSr�)r��_sourcer�r�_index)r�r�r�r�r�r�r�Y
s

zseekable.__init__cCs|Sr�r�r�r�r�r�r�a
szseekable.__iter__cCs`|jdurFz|j|j}Wnty2d|_Yn0|jd7_|St|j�}|j�|�|Sr)r~r�r�r�r}r��r�r�r�r�r�r�d
s

zseekable.__next__cCs&z|��Wnty YdS0dSr�r�r�r�r�r�r�r
s
zseekable.__bool__cCsVzt|�}Wn ty,|tur$�|YS0|jdurDt|j�|_|jd8_|Sr)r�r�r�r~r�r�)r�r�Zpeekedr�r�r�r�y
s

z
seekable.peekcCs
t|j�Sr�)rjr�r�r�r�r��elements�
szseekable.elementscCs*||_|t|j�}|dkr&t||�dSre)r~r�r�r-)r�r��	remainderr�r�r��seek�
sz
seekable.seek)N)r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�ri
sU
ric@s(eZdZdZedd��Zedd��ZdS)rga�
    :func:`run_length.encode` compresses an iterable with run-length encoding.
    It yields groups of repeated items with the count of how many times they
    were repeated:

        >>> uncompressed = 'abbcccdddd'
        >>> list(run_length.encode(uncompressed))
        [('a', 1), ('b', 2), ('c', 3), ('d', 4)]

    :func:`run_length.decode` decompresses an iterable that was previously
    compressed with run-length encoding. It yields the items of the
    decompressed iterable:

        >>> compressed = [('a', 1), ('b', 2), ('c', 3), ('d', 4)]
        >>> list(run_length.decode(compressed))
        ['a', 'b', 'b', 'c', 'c', 'c', 'd', 'd', 'd', 'd']

    cCsdd�t|�D�S)Ncss|]\}}|t|�fVqdSr�)rKrHr�r�r�r��
r�z$run_length.encode.<locals>.<genexpr>rMrir�r�r��encode�
szrun_length.encodecCst�dd�|D��S)Ncss|]\}}t||�VqdSr�)r)r�rIr�r�r�r�r��
r�z$run_length.decode.<locals>.<genexpr>)rr�rir�r�r��decode�
szrun_length.decodeN)r�r�r�r��staticmethodr�r�r�r�r�r�rg�
s

rgcCstt|dt||���|kS)a�Return ``True`` if exactly ``n`` items in the iterable are ``True``
    according to the *predicate* function.

        >>> exactly_n([True, True, False], 2)
        True
        >>> exactly_n([True, True, False], 1)
        False
        >>> exactly_n([0, 1, 2, 3, 4, 5], 3, lambda x: x < 3)
        True

    The iterable will be advanced until ``n + 1`` truthy items are encountered,
    so avoid calling it on infinite iterables.

    r,)r�r1r�)r�r�rBr�r�r�rG�
srGcCs$t|�}tt|�tt|�t|���S)z�Return a list of circular shifts of *iterable*.

    >>> circular_shifts(range(4))
    [(0, 1, 2, 3), (1, 2, 3, 0), (2, 3, 0, 1), (3, 0, 1, 2)]
    )r�r1r�r{r)r��lstr�r�r�r:�
sr:cs��fdd�}|S)a�Return a decorator version of *wrapping_func*, which is a function that
    modifies an iterable. *result_index* is the position in that function's
    signature where the iterable goes.

    This lets you use itertools on the "production end," i.e. at function
    definition. This can augment what the function returns without changing the
    function's code.

    For example, to produce a decorator version of :func:`chunked`:

        >>> from more_itertools import chunked
        >>> chunker = make_decorator(chunked, result_index=0)
        >>> @chunker(3)
        ... def iter_range(n):
        ...     return iter(range(n))
        ...
        >>> list(iter_range(9))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8]]

    To only allow truthy items to be returned:

        >>> truth_serum = make_decorator(filter, result_index=1)
        >>> @truth_serum(bool)
        ... def boolean_test():
        ...     return [0, 1, '', ' ', False, True]
        ...
        >>> list(boolean_test())
        [1, ' ', True]

    The :func:`peekable` and :func:`seekable` wrappers make for practical
    decorators:

        >>> from more_itertools import peekable
        >>> peekable_function = make_decorator(peekable)
        >>> @peekable_function()
        ... def str_range(*args):
        ...     return (str(x) for x in range(*args))
        ...
        >>> it = str_range(1, 20, 2)
        >>> next(it), next(it), next(it)
        ('1', '3', '5')
        >>> it.peek()
        '7'
        >>> next(it)
        '7'

    cs����fdd�}|S)Ncs�����fdd�}|S)Ncs0�|i|��}t��}|��|��|i���Sr�)r��insert)r�r��resultZwrapping_args_)�f�result_index�
wrapping_args�
wrapping_func�wrapping_kwargsr�r��
inner_wrapper�
szOmake_decorator.<locals>.decorator.<locals>.outer_wrapper.<locals>.inner_wrapperr�)r�r�)r�r�r�r�)r�r��
outer_wrapper�
sz8make_decorator.<locals>.decorator.<locals>.outer_wrapperr�)r�r�r��r�r�)r�r�r��	decorator�
s	z!make_decorator.<locals>.decoratorr�)r�r�r�r�r�r�rV�
s2rVc	Cst|durdd�n|}tt�}|D]"}||�}||�}||�|�q |durj|��D]\}}||�||<qTd|_|S)a�Return a dictionary that maps the items in *iterable* to categories
    defined by *keyfunc*, transforms them with *valuefunc*, and
    then summarizes them by category with *reducefunc*.

    *valuefunc* defaults to the identity function if it is unspecified.
    If *reducefunc* is unspecified, no summarization takes place:

        >>> keyfunc = lambda x: x.upper()
        >>> result = map_reduce('abbccc', keyfunc)
        >>> sorted(result.items())
        [('A', ['a']), ('B', ['b', 'b']), ('C', ['c', 'c', 'c'])]

    Specifying *valuefunc* transforms the categorized items:

        >>> keyfunc = lambda x: x.upper()
        >>> valuefunc = lambda x: 1
        >>> result = map_reduce('abbccc', keyfunc, valuefunc)
        >>> sorted(result.items())
        [('A', [1]), ('B', [1, 1]), ('C', [1, 1, 1])]

    Specifying *reducefunc* summarizes the categorized items:

        >>> keyfunc = lambda x: x.upper()
        >>> valuefunc = lambda x: 1
        >>> reducefunc = sum
        >>> result = map_reduce('abbccc', keyfunc, valuefunc, reducefunc)
        >>> sorted(result.items())
        [('A', 1), ('B', 2), ('C', 3)]

    You may want to filter the input iterable before applying the map/reduce
    procedure:

        >>> all_items = range(30)
        >>> items = [x for x in all_items if 10 <= x <= 20]  # Filter
        >>> keyfunc = lambda x: x % 2  # Evens map to 0; odds to 1
        >>> categories = map_reduce(items, keyfunc=keyfunc)
        >>> sorted(categories.items())
        [(0, [10, 12, 14, 16, 18, 20]), (1, [11, 13, 15, 17, 19])]
        >>> summaries = map_reduce(items, keyfunc=keyfunc, reducefunc=sum)
        >>> sorted(summaries.items())
        [(0, 90), (1, 75)]

    Note that all items in the iterable are gathered into a list before the
    summarization step, which may require significant storage.

    The returned object is a :obj:`collections.defaultdict` with the
    ``default_factory`` set to ``None``, such that it behaves like a normal
    dictionary.

    NcSs|Sr�r�rr�r�r�r<r�zmap_reduce.<locals>.<lambda>)rr�r�r��default_factory)	r�rNrKrLr�r�rr
Z
value_listr�r�r�rX	s3rXcsV|durBz&t|���fdd�tt|�|�D�WSty@Yn0ttt|||���S)a�Yield the index of each item in *iterable* for which *pred* returns
    ``True``, starting from the right and moving left.

    *pred* defaults to :func:`bool`, which will select truthy items:

        >>> list(rlocate([0, 1, 1, 0, 1, 0, 0]))  # Truthy at 1, 2, and 4
        [4, 2, 1]

    Set *pred* to a custom function to, e.g., find the indexes for a particular
    item:

        >>> iterable = iter('abcb')
        >>> pred = lambda x: x == 'b'
        >>> list(rlocate(iterable, pred))
        [3, 1]

    If *window_size* is given, then the *pred* function will be called with
    that many items. This enables searching for sub-sequences:

        >>> iterable = [0, 1, 2, 3, 0, 1, 2, 3, 0, 1, 2, 3]
        >>> pred = lambda *args: args == (1, 2, 3)
        >>> list(rlocate(iterable, pred=pred, window_size=3))
        [9, 5, 1]

    Beware, this function won't return anything for infinite iterables.
    If *iterable* is reversible, ``rlocate`` will reverse it and search from
    the right. Otherwise, it will search from the left and return the results
    in reverse order.

    See :func:`locate` to for other example applications.

    Nc3s|]}�|dVqdSr�r�r�rur�r�r�pr�zrlocate.<locals>.<genexpr>)r�rTr�r�r�)r�r rmr�r�r�reLs!rec	cs�|dkrtd��t|�}t|tg|d�}t||�}d}|D]X}||�r||dusZ||kr||d7}|EdHt||d�q>|r>|dtur>|dVq>dS)aYYield the items from *iterable*, replacing the items for which *pred*
    returns ``True`` with the items from the iterable *substitutes*.

        >>> iterable = [1, 1, 0, 1, 1, 0, 1, 1]
        >>> pred = lambda x: x == 0
        >>> substitutes = (2, 3)
        >>> list(replace(iterable, pred, substitutes))
        [1, 1, 2, 3, 1, 1, 2, 3, 1, 1]

    If *count* is given, the number of replacements will be limited:

        >>> iterable = [1, 1, 0, 1, 1, 0, 1, 1, 0]
        >>> pred = lambda x: x == 0
        >>> substitutes = [None]
        >>> list(replace(iterable, pred, substitutes, count=2))
        [1, 1, None, 1, 1, None, 1, 1, 0]

    Use *window_size* to control the number of items passed as arguments to
    *pred*. This allows for locating and replacing subsequences.

        >>> iterable = [0, 1, 2, 5, 0, 1, 2, 5]
        >>> window_size = 3
        >>> pred = lambda *args: args == (0, 1, 2)  # 3 items passed to pred
        >>> substitutes = [3, 4] # Splice in these items
        >>> list(replace(iterable, pred, substitutes, window_size=window_size))
        [3, 4, 5, 3, 4, 5]

    r,zwindow_size must be at least 1rN)r�r�rr�r{r-)	r�r Zsubstitutesrrmr�Zwindowsr��wr�r�r�rdws

rdc#sLt|��t��}ttd|��D](}�fdd�td|||f�D�VqdS)a"Yield all possible order-preserving partitions of *iterable*.

    >>> iterable = 'abc'
    >>> for part in partitions(iterable):
    ...     print([''.join(p) for p in part])
    ['abc']
    ['a', 'bc']
    ['ab', 'c']
    ['a', 'b', 'c']

    This is unrelated to :func:`partition`.

    r,csg|]\}}�||��qSr�r�)r�r�r���sequencer�r�r��r�zpartitions.<locals>.<listcomp>r6N)r�r�r0r�r�)r�r�r�r�r�r�r`�sr`c#s�t|�}t|�}|dur6|dkr*td��n||kr6dS�fdd��|durptd|d�D]}�||�EdHqXn�||�EdHdS)a
    Yield the set partitions of *iterable* into *k* parts. Set partitions are
    not order-preserving.

    >>> iterable = 'abc'
    >>> for part in set_partitions(iterable, 2):
    ...     print([''.join(p) for p in part])
    ['a', 'bc']
    ['ab', 'c']
    ['b', 'ac']


    If *k* is not given, every set partition is generated.

    >>> iterable = 'abc'
    >>> for part in set_partitions(iterable):
    ...     print([''.join(p) for p in part])
    ['abc']
    ['a', 'bc']
    ['ab', 'c']
    ['b', 'ac']
    ['a', 'b', 'c']

    Nr,z6Can't partition in a negative or zero number of groupsc3s�t|�}|dkr|gVn�||kr4dd�|D�Vnz|^}}�||d�D]}|gg|�VqJ�||�D]D}tt|��D]2}|d|�|g||g||dd�VqxqhdS)Nr,cSsg|]
}|g�qSr�r�)r�rtr�r�r�r��r�zAset_partitions.<locals>.set_partitions_helper.<locals>.<listcomp>)r�r�)r�rIr�r��M�pr���set_partitions_helperr�r�r��s
z-set_partitions.<locals>.set_partitions_helper)r�r�r�r�)r�rIr�r�r�r�r�ra�s�rac@s(eZdZdZdd�Zdd�Zdd�ZdS)	rxa�
    Yield items from *iterable* until *limit_seconds* have passed.
    If the time limit expires before all items have been yielded, the
    ``timed_out`` parameter will be set to ``True``.

    >>> from time import sleep
    >>> def generator():
    ...     yield 1
    ...     yield 2
    ...     sleep(0.2)
    ...     yield 3
    >>> iterable = time_limited(0.1, generator())
    >>> list(iterable)
    [1, 2]
    >>> iterable.timed_out
    True

    Note that the time is checked before each item is yielded, and iteration
    stops if  the time elapsed is greater than *limit_seconds*. If your time
    limit is 1 second, but it takes 2 seconds to generate the first item from
    the iterable, the function will run for 2 seconds and not yield anything.

    cCs2|dkrtd��||_t|�|_t�|_d|_dS)Nrzlimit_seconds must be positiveF)r��
limit_secondsr�rsr+�_start_time�	timed_out)r�r�r�r�r�r�r�s
ztime_limited.__init__cCs|Sr�r�r�r�r�r�r�!sztime_limited.__iter__cCs*t|j�}t�|j|jkr&d|_t�|Sr)r�rsr+r�r�r�r�rr�r�r�r�$s

ztime_limited.__next__N�r�r�r�r�r�r�r�r�r�r�r�rxsrxcCsNt|�}t||�}zt|�}Wnty0Yn0d�||�}|pHt|��|S)a*If *iterable* has only one item, return it.
    If it has zero items, return *default*.
    If it has more than one item, raise the exception given by *too_long*,
    which is ``ValueError`` by default.

    >>> only([], default='missing')
    'missing'
    >>> only([1])
    1
    >>> only([1, 2])  # doctest: +IGNORE_EXCEPTION_DETAIL
    Traceback (most recent call last):
    ...
    ValueError: Expected exactly one item in iterable, but got 1, 2,
     and perhaps more.'
    >>> only([1, 2], too_long=TypeError)  # doctest: +IGNORE_EXCEPTION_DETAIL
    Traceback (most recent call last):
    ...
    TypeError

    Note that :func:`only` attempts to advance *iterable* twice to ensure there
    is only one item.  See :func:`spy` or :func:`peekable` to check
    iterable contents less destructively.
    r�)r�r�r�r�r�)r�r�r�r�r�r�r�r�r�r�r^-s
��r^ccsNt|�}t|t�}|turdStt|g|��\}}t||�Vt||�qdS)a�Break *iterable* into sub-iterables with *n* elements each.
    :func:`ichunked` is like :func:`chunked`, but it yields iterables
    instead of lists.

    If the sub-iterables are read in order, the elements of *iterable*
    won't be stored in memory.
    If they are read out of order, :func:`itertools.tee` is used to cache
    elements as necessary.

    >>> from itertools import count
    >>> all_chunks = ichunked(count(), 4)
    >>> c_1, c_2, c_3 = next(all_chunks), next(all_chunks), next(all_chunks)
    >>> list(c_2)  # c_1's elements have been cached; c_3's haven't been
    [4, 5, 6, 7]
    >>> list(c_1)
    [0, 1, 2, 3]
    >>> list(c_3)
    [8, 9, 10, 11]

    N)r�r�r�rrrr-)r�r��sourcer�r�r�r�r�rQVs
rQccs�|dkrtd��n|dkr$dVdSt|�}tt|�td�d�g}dg|}d}|r�zt|d�\}}Wn&ty�|��|d8}YqPYn0|||<|d|kr�t|�VqP|�tt||dd�|d�td�d��|d7}qPdS)aBYield the distinct combinations of *r* items taken from *iterable*.

        >>> list(distinct_combinations([0, 0, 1], 2))
        [(0, 0), (0, 1)]

    Equivalent to ``set(combinations(iterable))``, except duplicates are not
    generated and thrown away. For larger input sequences this is much more
    efficient.

    rzr must be non-negativer�Nr,ryr�)	r�r�r2r)r$r�r��popr�)r�r�r��
generatorsZ
current_comborZcur_idxr�r�r�r�rC{s4


��rCc	gs4|D]*}z||�Wn|y&Yq0|VqdS)aYield the items from *iterable* for which the *validator* function does
    not raise one of the specified *exceptions*.

    *validator* is called for each item in *iterable*.
    It should be a function that accepts one argument and raises an exception
    if that item is not valid.

    >>> iterable = ['1', '2', 'three', '4', None]
    >>> list(filter_except(int, iterable, ValueError, TypeError))
    ['1', '2', '4']

    If an exception other than one given by *exceptions* is raised by
    *validator*, it is raised like normal.
    Nr�)rr��
exceptionsr�r�r�r�rH�srHc	gs0|D]&}z||�VWq|y(Yq0qdS)a�Transform each item from *iterable* with *function* and yield the
    result, unless *function* raises one of the specified *exceptions*.

    *function* is called to transform each item in *iterable*.
    It should be a accept one argument.

    >>> iterable = ['1', '2', 'three', '4', None]
    >>> list(map_except(int, iterable, ValueError, TypeError))
    [1, 2, 4]

    If an exception other than one given by *exceptions* is raised by
    *function*, it is raised like normal.
    Nr�)�functionr�r�r�r�r�r�rW�s
rWcCs�t||�}ttt��|�}|ttt��td|��}t||�D]T\}}||krD||t|�<|ttt��|�9}|ttt��td|��d7}qD|Sr)r1rrr!rr)r")r�rI�	reservoir�WZ
next_indexr�r�r�r�r��_sample_unweighted�s
$r�cs�dd�|D�}t|t||���t���d\}}tt��|}t||�D]p\}}||kr��d\}}t||�}	t|	d�}
t|
�|}t�||f��d\}}tt��|}qJ||8}qJ�fdd�t|�D�S)Ncss|]}tt��|VqdSr�)rr!)r��weightr�r�r�r��r�z#_sample_weighted.<locals>.<genexpr>rr,csg|]}t��d�qSr�)r)r�r��r�r�r�r�

r�z$_sample_weighted.<locals>.<listcomp>)	r1r�rrr!rr#r
r�)r�rI�weightsZweight_keysZsmallest_weight_keyr�Zweights_to_skipr�r�Zt_wZr_2Z
weight_keyr�r�r��_sample_weighted�s 

r�cCs>|dkrgSt|�}|dur&t||�St|�}t|||�SdS)afReturn a *k*-length list of elements chosen (without replacement)
    from the *iterable*. Like :func:`random.sample`, but works on iterables
    of unknown length.

    >>> iterable = range(100)
    >>> sample(iterable, 5)  # doctest: +SKIP
    [81, 60, 96, 16, 4]

    An iterable with *weights* may also be given:

    >>> iterable = range(100)
    >>> weights = (i * i + 1 for i in range(100))
    >>> sampled = sample(iterable, 5, weights=weights)  # doctest: +SKIP
    [79, 67, 74, 66, 78]

    The algorithm can also be used to generate weighted random permutations.
    The relative weight of each item determines the probability that it
    appears late in the permutation.

    >>> data = "abcdefgh"
    >>> weights = range(1, len(data) + 1)
    >>> sample(data, k=len(data), weights=weights)  # doctest: +SKIP
    ['c', 'a', 'b', 'e', 'g', 'd', 'h', 'f']
    rN)r�r�r�)r�rIr�r�r�r�rh

s
rhcCs6|rtnt}|dur|nt||�}tt|t|���S)a�Returns ``True`` if the items of iterable are in sorted order, and
    ``False`` otherwise. *key* and *reverse* have the same meaning that they do
    in the built-in :func:`sorted` function.

    >>> is_sorted(['1', '2', '3', '4', '5'], key=int)
    True
    >>> is_sorted([5, 4, 3, 1, 2], reverse=True)
    False

    The function returns ``False`` after encountering the first out-of-order
    item. If there are no out-of-order items, the iterable is exhausted.
    N)r(r'r�rArr/)r�rr�comparer�r�r�r�rR1
srRc@seZdZdS)r3N)r�r�r�r�r�r�r�r3D
sr3c@sZeZdZdZddd�Zdd�Zdd	�Zd
d�Zdd
�Ze	dd��Z
e	dd��Zdd�ZdS)r8a�Convert a function that uses callbacks to an iterator.

    Let *func* be a function that takes a `callback` keyword argument.
    For example:

    >>> def func(callback=None):
    ...     for i, c in [(1, 'a'), (2, 'b'), (3, 'c')]:
    ...         if callback:
    ...             callback(i, c)
    ...     return 4


    Use ``with callback_iter(func)`` to get an iterator over the parameters
    that are delivered to the callback.

    >>> with callback_iter(func) as it:
    ...     for args, kwargs in it:
    ...         print(args)
    (1, 'a')
    (2, 'b')
    (3, 'c')

    The function will be called in a background thread. The ``done`` property
    indicates whether it has completed execution.

    >>> it.done
    True

    If it completes successfully, its return value will be available
    in the ``result`` property.

    >>> it.result
    4

    Notes:

    * If the function uses some keyword argument besides ``callback``, supply
      *callback_kwd*.
    * If it finished executing, but raised an exception, accessing the
      ``result`` property will raise the same exception.
    * If it hasn't finished executing, accessing the ``result``
      property from within the ``with`` block will raise ``RuntimeError``.
    * If it hasn't finished executing, accessing the ``result`` property from
      outside the ``with`` block will raise a
      ``more_itertools.AbortThread`` exception.
    * Provide *wait_seconds* to adjust how frequently the it is polled for
      output.

    �callback皙�����?cCs8||_||_d|_d|_||_tdd�|_|��|_dS)NFr,)�max_workers)	�_func�
_callback_kwd�_aborted�_future�
_wait_secondsr�	_executor�_reader�	_iterator)r�r�Zcallback_kwdZwait_secondsr�r�r�r�{
szcallback_iter.__init__cCs|Sr�r�r�r�r�r��	__enter__�
szcallback_iter.__enter__cCsd|_|j��dSr)r�r��shutdown)r��exc_type�	exc_value�	tracebackr�r�r��__exit__�
szcallback_iter.__exit__cCs|Sr�r�r�r�r�r�r��
szcallback_iter.__iter__cCs
t|j�Sr�)r�r�r�r�r�r�r��
szcallback_iter.__next__cCs|jdurdS|j��SrW)r��doner�r�r�r�r��
s
zcallback_iter.donecCs|jstd��|j��S)NzFunction has not yet completed)r��RuntimeErrorr�r�r�r�r�r�r��
szcallback_iter.resultc#s�t����fdd�}�jj�jfi�j|i���_z�j�jd�}WntyVYn0��	�|V�j�
�r2qtq2g}z���}Wnty�Yq�Yqx0��	�|�|�qx��
�|EdHdS)Ncs �jrtd����||f�dS)Nzcanceled by user)r�r3�put)r�r��r@r�r�r�r��
sz'callback_iter._reader.<locals>.callback)�timeout)r r��submitr�r�r��getr�r�	task_doner��
get_nowaitr��join)r�r�r�r&r�r�r�r��
s0��

zcallback_iter._readerN)r�r�)
r�r�r�r�r�r�r�r�r��propertyr�r�r�r�r�r�r�r8H
s2
	

r8ccs�|dkrtd��t|�}t|�}||kr0td��t||d�D]<}|d|�}||||�}|||d�}|||fVq@dS)a�
    Yield ``(beginning, middle, end)`` tuples, where:

    * Each ``middle`` has *n* items from *iterable*
    * Each ``beginning`` has the items before the ones in ``middle``
    * Each ``end`` has the items after the ones in ``middle``

    >>> iterable = range(7)
    >>> n = 3
    >>> for beginning, middle, end in windowed_complete(iterable, n):
    ...     print(beginning, middle, end)
    () (0, 1, 2) (3, 4, 5, 6)
    (0,) (1, 2, 3) (4, 5, 6)
    (0, 1) (2, 3, 4) (5, 6)
    (0, 1, 2) (3, 4, 5) (6,)
    (0, 1, 2, 3) (4, 5, 6) ()

    Note that *n* must be at least 0 and most equal to the length of
    *iterable*.

    This function will exhaust the iterable and may require significant
    storage.
    rr�zn must be <= len(seq)r,N)r�r�r�r�)r�r�r�r�r�Z	beginningZmiddle�endr�r�r�r��
sr�c	Cszt�}|j}g}|j}|r$t||�n|D]L}z||vr>WdS||�Wq(tyr||vrfYdS||�Yq(0q(dS)a
    Returns ``True`` if all the elements of *iterable* are unique (no two
    elements are equal).

        >>> all_unique('ABCB')
        False

    If a *key* function is specified, it will be used to make comparisons.

        >>> all_unique('ABCb')
        True
        >>> all_unique('ABCb', str.lower)
        False

    The function returns as soon as the first non-unique element is
    encountered. Iterables with a mix of hashable and unhashable items can
    be used, but the function will be slower for unhashable items.
    FT)r��addr�r�r�)r�rZseensetZseenset_addZseenlistZseenlist_addr�r�r�r�r��
sr�cGs�tttt|���}ttt|��}tt|�}|dkr:||7}d|krN|ksTnt�g}t||�D]"\}}|�	|||�||}qbtt|��S)a�Equivalent to ``list(product(*args))[index]``.

    The products of *args* can be ordered lexicographically.
    :func:`nth_product` computes the product at sort position *index* without
    computing the previous products.

        >>> nth_product(8, range(2), range(2), range(2), range(2))
        (1, 0, 0, 0)

    ``IndexError`` will be raised if the given *index* is invalid.
    r)
r�r�r�r�r�r	r%r�r�r�)r�r��pools�ns�cr�r�r�r�r�r�r[s

r[c
Cs(t|�}t|�}|dus ||kr0|t|�}}n0d|krD|ksLnt�nt|�t||�}|dkrp||7}d|kr�|ks�nt�|dkr�t�Sdg|}||kr�|t|�|n|}td|d�D]J}t||�\}}	d||kr�|k�rnn|	|||<|dkr��qq�tt|j	|��S)a'Equivalent to ``list(permutations(iterable, r))[index]```

    The subsequences of *iterable* that are of length *r* where order is
    important can be ordered lexicographically. :func:`nth_permutation`
    computes the subsequence at sort position *index* directly, without
    computing the previous subsequences.

        >>> nth_permutation('ghijk', 2, 5)
        ('h', 'i')

    ``ValueError`` will be raised If *r* is negative or greater than the length
    of *iterable*.
    ``IndexError`` will be raised if the given *index* is invalid.
    Nrr,)
r�r�rr�r�r�r�r?r�r�)
r�r�r�r�r�r�r�r@�dr�r�r�r�rZ.s,
rZc	gsL|D]B}t|ttf�r|Vqz|EdHWqtyD|VYq0qdS)a�Yield all arguments passed to the function in the same order in which
    they were passed. If an argument itself is iterable then iterate over its
    values.

        >>> list(value_chain(1, 2, 3, [4, 5, 6]))
        [1, 2, 3, 4, 5, 6]

    Binary and text strings are not considered iterable and are emitted
    as-is:

        >>> list(value_chain('12', '34', ['56', '78']))
        ['12', '34', '56', '78']


    Multiple levels of nesting are not flattened.

    N)r�rrr�)r�r
r�r�r�r�\sr�cGsVd}t||td�D]>\}}|tus*|tur2td��t|�}|t|�|�|�}q|S)a�Equivalent to ``list(product(*args)).index(element)``

    The products of *args* can be ordered lexicographically.
    :func:`product_index` computes the first index of *element* without
    computing the previous products.

        >>> product_index([8, 2], range(10), range(5))
        42

    ``ValueError`` will be raised if the given *element* isn't in the product
    of *args*.
    rrlz element is not a product of args)rr�r�r�r�r�)r�r�r�rr�r�r�r�r�xs
r�c
Cst|�}t|d�\}}|dur"dSg}t|�}|D]:\}}||kr2|�|�t|d�\}}|durhqvq2|}q2td��t||dfd�\}}	d}
tt|�dd�D]8\}}||}||kr�|
t|�t|�t||�7}
q�t|d�t|d�t||�|
S)a�Equivalent to ``list(combinations(iterable, r)).index(element)``

    The subsequences of *iterable* that are of length *r* can be ordered
    lexicographically. :func:`combination_index` computes the index of the
    first *element*, without computing the previous combinations.

        >>> combination_index('adf', 'abcdefg')
        10

    ``ValueError`` will be raised if the given *element* isn't one of the
    combinations of *iterable*.
    )NNNrz(element is not a combination of iterabler�r,)r�)r)r�r�r�rSr�r)
r�r�rI�yZindexesr�r�r�tmpr�r�r�r�r�r�r�r��s*

"r�cCsLd}t|�}ttt|�dd�|�D]$\}}|�|�}|||}||=q"|S)a�Equivalent to ``list(permutations(iterable, r)).index(element)```

    The subsequences of *iterable* that are of length *r* where order is
    important can be ordered lexicographically. :func:`permutation_index`
    computes the index of the first *element* directly, without computing
    the previous permutations.

        >>> permutation_index([1, 3, 2], range(5))
        19

    ``ValueError`` will be raised if the given *element* isn't one of the
    permutations of *iterable*.
    rr�)r�r�r�r�r�)r�r�r�r�r�rr�r�r�r�r��s
r�c@s(eZdZdZdd�Zdd�Zdd�ZdS)	r?a�Wrap *iterable* and keep a count of how many items have been consumed.

    The ``items_seen`` attribute starts at ``0`` and increments as the iterable
    is consumed:

        >>> iterable = map(str, range(10))
        >>> it = countable(iterable)
        >>> it.items_seen
        0
        >>> next(it), next(it)
        ('0', '1')
        >>> list(it)
        ['2', '3', '4', '5', '6', '7', '8', '9']
        >>> it.items_seen
        10
    cCst|�|_d|_dSre)r�r��
items_seenr�r�r�r�r��s
zcountable.__init__cCs|Sr�r�r�r�r�r�r��szcountable.__iter__cCst|j�}|jd7_|Sr)r�r�r�rr�r�r�r��s
zcountable.__next__Nr�r�r�r�r�r?�sr?)F)NN)N)r,)Nr,)F)r,)NN)NNN)F)r�F)r�)r�)r�)NNF)N)r+FN)r6NF)r,)NNN)N)r)NN)Nr,)N)NN)N)NF)N)�r��collectionsrrrr�collections.abcr�concurrent.futuresr�	functoolsrr	r
�heapqrrr
r�	itertoolsrrrrrrrrrrrr�mathrrrr�queuerr r!r"r#�operatorr$r%r&r'r(�sysr)r*�timer+Zrecipesr-r.r/r0r1r2�__all__�objectr�r9rIrSrYrbr<r>rKrPr|r]rDrNryr{rvrwr7rsrMrLr;rkrlrnrprorqrrr_rcrErtr�r}r4r~rrmrzrFrrr5r4rJ�Hashabler\r@rArYrTrUrfrurOrrr6r=rBrjrirgrGr:rVrXrerdr`rarxr^rQrCrHrWr�r�rhrR�
BaseExceptionr3r8r�r�r[rZr�r�r�r�r?r�r�r�r��<module>s�8 	V
!&& 

C
d
!
3
"`
+
0
=
"
,
#
$
--
 
#
.'
B135
'
-Z
%0.`0*-


A
C+
=
8-
)%(#
$
|(
#.+site-packages/setuptools/_vendor/more_itertools/__pycache__/__init__.cpython-39.pyc000064400000000475150564745010024624 0ustar00a

(�ReR�@sddlTddlTdZdS)�)�*z8.8.0N)�moreZrecipes�__version__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/more_itertools/__init__.py�<module>ssite-packages/setuptools/_vendor/more_itertools/__pycache__/recipes.cpython-39.pyc000064400000043076150564745010024523 0ustar00a

(�Re�?�@s�dZddlZddlmZddlmZmZmZmZm	Z	m
Z
mZmZm
Z
mZddlZddlmZmZmZgd�Zdd�ZdDd	d
�Zdd�ZdEd
d�ZdFdd�Zdd�Zefdd�Zdd�ZeZdd�Zdd�Z dd�Z!dGdd�Z"dd �Z#zdd!lm$Z%Wne&�ye#Z$Yn0d"d#�Z$e#je$_dHd$d%�Z'd&d'�Z(d(d)�Z)d*d+�Z*dId,d-�Z+dJd.d/�Z,dKd0d1�Z-dLd2d3�Z.d4d5�d6d7�Z/dMd8d9�Z0d:d;�Z1d<d=�Z2d>d?�Z3d@dA�Z4dBdC�Z5dS)NaImported from the recipes section of the itertools documentation.

All functions taken from the recipes section of the itertools library docs
[1]_.
Some backward-compatible usability improvements have been made.

.. [1] http://docs.python.org/library/itertools.html#recipes

�N)�deque)
�chain�combinations�count�cycle�groupby�islice�repeat�starmap�tee�zip_longest)�	randrange�sample�choice)�	all_equal�consume�convolve�
dotproduct�
first_true�flatten�grouper�iter_except�ncycles�nth�nth_combination�padnone�pad_none�pairwise�	partition�powerset�prepend�quantify�#random_combination_with_replacement�random_combination�random_permutation�random_product�
repeatfunc�
roundrobin�tabulate�tail�take�unique_everseen�unique_justseencCstt||��S)z�Return first *n* items of the iterable as a list.

        >>> take(3, range(10))
        [0, 1, 2]

    If there are fewer than *n* items in the iterable, all of them are
    returned.

        >>> take(10, range(3))
        [0, 1, 2]

    )�listr��n�iterable�r1��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/_vendor/more_itertools/recipes.pyr*<s
r*cCst|t|��S)a�Return an iterator over the results of ``func(start)``,
    ``func(start + 1)``, ``func(start + 2)``...

    *func* should be a function that accepts one integer argument.

    If *start* is not specified it defaults to 0. It will be incremented each
    time the iterator is advanced.

        >>> square = lambda x: x ** 2
        >>> iterator = tabulate(square, -3)
        >>> take(4, iterator)
        [9, 4, 1, 0]

    )�mapr)�function�startr1r1r2r(Lsr(cCstt||d��S)z�Return an iterator over the last *n* items of *iterable*.

    >>> t = tail(3, 'ABCDEFG')
    >>> list(t)
    ['E', 'F', 'G']

    ��maxlen)�iterrr.r1r1r2r)^sr)cCs,|durt|dd�ntt|||�d�dS)aXAdvance *iterable* by *n* steps. If *n* is ``None``, consume it
    entirely.

    Efficiently exhausts an iterator without returning values. Defaults to
    consuming the whole iterator, but an optional second argument may be
    provided to limit consumption.

        >>> i = (x for x in range(10))
        >>> next(i)
        0
        >>> consume(i, 3)
        >>> next(i)
        4
        >>> consume(i)
        >>> next(i)
        Traceback (most recent call last):
          File "<stdin>", line 1, in <module>
        StopIteration

    If the iterator has fewer items remaining than the provided limit, the
    whole iterator will be consumed.

        >>> i = (x for x in range(3))
        >>> consume(i, 5)
        >>> next(i)
        Traceback (most recent call last):
          File "<stdin>", line 1, in <module>
        StopIteration

    Nrr6)r�nextr)�iteratorr/r1r1r2ris rcCstt||d�|�S)z�Returns the nth item or a default value.

    >>> l = range(10)
    >>> nth(l, 3)
    3
    >>> nth(l, 20, "zebra")
    'zebra'

    N)r9r)r0r/�defaultr1r1r2r�s
rcCst|�}t|d�ot|d�S)z�
    Returns ``True`` if all the elements are equal to each other.

        >>> all_equal('aaaa')
        True
        >>> all_equal('aaab')
        False

    TF)rr9)r0�gr1r1r2r�s
rcCstt||��S)zcReturn the how many times the predicate is true.

    >>> quantify([True, False, True])
    2

    )�sumr3)r0�predr1r1r2r!�sr!cCst|td��S)aReturns the sequence of elements and then returns ``None`` indefinitely.

        >>> take(5, pad_none(range(3)))
        [0, 1, 2, None, None]

    Useful for emulating the behavior of the built-in :func:`map` function.

    See also :func:`padded`.

    N)rr	�r0r1r1r2r�srcCst�tt|�|��S)zvReturns the sequence elements *n* times

    >>> list(ncycles(["a", "b"], 3))
    ['a', 'b', 'a', 'b', 'a', 'b']

    )r�
from_iterabler	�tuple)r0r/r1r1r2r�srcCstttj||��S)zcReturns the dot product of the two iterables.

    >>> dotproduct([10, 10], [20, 20])
    400

    )r=r3�operator�mul)Zvec1Zvec2r1r1r2r�srcCs
t�|�S)z�Return an iterator flattening one level of nesting in a list of lists.

        >>> list(flatten([[0, 1], [2, 3]]))
        [0, 1, 2, 3]

    See also :func:`collapse`, which can flatten multiple levels of nesting.

    )rr@)ZlistOfListsr1r1r2r�s	rcGs&|durt|t|��St|t||��S)aGCall *func* with *args* repeatedly, returning an iterable over the
    results.

    If *times* is specified, the iterable will terminate after that many
    repetitions:

        >>> from operator import add
        >>> times = 4
        >>> args = 3, 5
        >>> list(repeatfunc(add, times, *args))
        [8, 8, 8, 8]

    If *times* is ``None`` the iterable will not terminate:

        >>> from random import randrange
        >>> times = None
        >>> args = 1, 11
        >>> take(6, repeatfunc(randrange, times, *args))  # doctest:+SKIP
        [2, 4, 8, 1, 8, 4]

    N)r
r	)�func�times�argsr1r1r2r&�sr&ccs*t|�\}}t|d�t||�EdHdS)z�Returns an iterator of paired items, overlapping, from the original

    >>> take(4, pairwise(count()))
    [(0, 1), (1, 2), (2, 3), (3, 4)]

    On Python 3.10 and above, this is an alias for :func:`itertools.pairwise`.

    N)rr9�zip)r0�a�br1r1r2�	_pairwises	
rJ)rccst|�EdHdS�N)�itertools_pairwiser?r1r1r2rsrcCs<t|t�r t�dt�||}}t|�g|}t|d|i�S)z�Collect data into fixed-length chunks or blocks.

    >>> list(grouper('ABCDEFG', 3, 'x'))
    [('A', 'B', 'C'), ('D', 'E', 'F'), ('G', 'x', 'x')]

    z+grouper expects iterable as first parameter�	fillvalue)�
isinstance�int�warnings�warn�DeprecationWarningr8r)r0r/rMrFr1r1r2rs
�
rcgsdt|�}tdd�|D��}|r`z|D]}|�Vq$Wqty\|d8}tt||��}Yq0qdS)aJYields an item from each iterable, alternating between them.

        >>> list(roundrobin('ABC', 'D', 'EF'))
        ['A', 'D', 'E', 'B', 'F', 'C']

    This function produces the same output as :func:`interleave_longest`, but
    may perform better for some inputs (in particular when the number of
    iterables is small).

    css|]}t|�jVqdSrK)r8�__next__)�.0�itr1r1r2�	<genexpr>9�zroundrobin.<locals>.<genexpr>�N)�lenr�
StopIterationr)�	iterables�pendingZnextsr9r1r1r2r',sr'csF�durt��fdd�|D�}t|�\}}dd�|D�dd�|D�fS)a�
    Returns a 2-tuple of iterables derived from the input iterable.
    The first yields the items that have ``pred(item) == False``.
    The second yields the items that have ``pred(item) == True``.

        >>> is_odd = lambda x: x % 2 != 0
        >>> iterable = range(10)
        >>> even_items, odd_items = partition(is_odd, iterable)
        >>> list(even_items), list(odd_items)
        ([0, 2, 4, 6, 8], [1, 3, 5, 7, 9])

    If *pred* is None, :func:`bool` is used.

        >>> iterable = [0, 1, False, True, '', ' ']
        >>> false_items, true_items = partition(None, iterable)
        >>> list(false_items), list(true_items)
        ([0, False, ''], [1, True, ' '])

    Nc3s|]}�|�|fVqdSrKr1)rT�x�r>r1r2rVZrWzpartition.<locals>.<genexpr>css|]\}}|s|VqdSrKr1�rTZcondr]r1r1r2rV]rWcss|]\}}|r|VqdSrKr1r_r1r1r2rV^rW)�boolr)r>r0Zevaluations�t1�t2r1r^r2rCs�rcs,t|��t��fdd�tt��d�D��S)a�Yields all possible subsets of the iterable.

        >>> list(powerset([1, 2, 3]))
        [(), (1,), (2,), (3,), (1, 2), (1, 3), (2, 3), (1, 2, 3)]

    :func:`powerset` will operate on iterables that aren't :class:`set`
    instances, so repeated elements in the input will produce repeated elements
    in the output. Use :func:`unique_everseen` on the input to avoid generating
    duplicates:

        >>> seq = [1, 1, 0]
        >>> list(powerset(seq))
        [(), (1,), (1,), (0,), (1, 1), (1, 0), (1, 0), (1, 1, 0)]
        >>> from more_itertools import unique_everseen
        >>> list(powerset(unique_everseen(seq)))
        [(), (1,), (0,), (1, 0)]

    c3s|]}t�|�VqdSrK)r)rT�r��sr1r2rVvrWzpowerset.<locals>.<genexpr>rX)r-rr@�rangerYr?r1rdr2rbsrc		cs�t�}|j}g}|j}|du}|D]X}|r2||�n|}z||vrN||�|VWq"tyx||vrt||�|VYq"0q"dS)a�
    Yield unique elements, preserving order.

        >>> list(unique_everseen('AAAABBBCCDAABBB'))
        ['A', 'B', 'C', 'D']
        >>> list(unique_everseen('ABBCcAD', str.lower))
        ['A', 'B', 'C', 'D']

    Sequences with a mix of hashable and unhashable items can be used.
    The function will be slower (i.e., `O(n^2)`) for unhashable items.

    Remember that ``list`` objects are unhashable - you can use the *key*
    parameter to transform the list to a tuple (which is hashable) to
    avoid a slowdown.

        >>> iterable = ([1, 2], [2, 3], [1, 2])
        >>> list(unique_everseen(iterable))  # Slow
        [[1, 2], [2, 3]]
        >>> list(unique_everseen(iterable, key=tuple))  # Faster
        [[1, 2], [2, 3]]

    Similary, you may want to convert unhashable ``set`` objects with
    ``key=frozenset``. For ``dict`` objects,
    ``key=lambda x: frozenset(x.items())`` can be used.

    N)�set�add�append�	TypeError)	r0�keyZseensetZseenset_addZseenlistZseenlist_addZuse_key�element�kr1r1r2r+ys
r+cCstttt�d�t||���S)z�Yields elements in order, ignoring serial duplicates

    >>> list(unique_justseen('AAAABBBCCDAABBB'))
    ['A', 'B', 'C', 'D', 'A', 'B']
    >>> list(unique_justseen('ABBCcAD', str.lower))
    ['A', 'B', 'C', 'A', 'D']

    rX)r3r9rB�
itemgetterr)r0rkr1r1r2r,�s	r,ccs6z|dur|�V|�VqWn|y0Yn0dS)aXYields results from a function repeatedly until an exception is raised.

    Converts a call-until-exception interface to an iterator interface.
    Like ``iter(func, sentinel)``, but uses an exception instead of a sentinel
    to end the loop.

        >>> l = [0, 1, 2]
        >>> list(iter_except(l.pop, IndexError))
        [2, 1, 0]

    Nr1)rD�	exception�firstr1r1r2r�srcCstt||�|�S)a�
    Returns the first true value in the iterable.

    If no true value is found, returns *default*

    If *pred* is not None, returns the first item for which
    ``pred(item) == True`` .

        >>> first_true(range(10))
        1
        >>> first_true(range(10), pred=lambda x: x > 5)
        6
        >>> first_true(range(10), default='missing', pred=lambda x: x > 9)
        'missing'

    )r9�filter)r0r;r>r1r1r2r�srrX)r	cGs$dd�|D�|}tdd�|D��S)a�Draw an item at random from each of the input iterables.

        >>> random_product('abc', range(4), 'XYZ')  # doctest:+SKIP
        ('c', 3, 'Z')

    If *repeat* is provided as a keyword argument, that many items will be
    drawn from each iterable.

        >>> random_product('abcd', range(4), repeat=2)  # doctest:+SKIP
        ('a', 2, 'd', 3)

    This equivalent to taking a random selection from
    ``itertools.product(*args, **kwarg)``.

    cSsg|]}t|��qSr1�rA�rT�poolr1r1r2�
<listcomp>�rWz"random_product.<locals>.<listcomp>css|]}t|�VqdSrK)rrsr1r1r2rV�rWz!random_product.<locals>.<genexpr>rr)r	rF�poolsr1r1r2r%�sr%cCs*t|�}|durt|�n|}tt||��S)abReturn a random *r* length permutation of the elements in *iterable*.

    If *r* is not specified or is ``None``, then *r* defaults to the length of
    *iterable*.

        >>> random_permutation(range(5))  # doctest:+SKIP
        (3, 4, 0, 1, 2)

    This equivalent to taking a random selection from
    ``itertools.permutations(iterable, r)``.

    N)rArYr)r0rcrtr1r1r2r$�s
r$cs8t|��t��}ttt|�|��}t�fdd�|D��S)z�Return a random *r* length subsequence of the elements in *iterable*.

        >>> random_combination(range(5), 3)  # doctest:+SKIP
        (2, 3, 4)

    This equivalent to taking a random selection from
    ``itertools.combinations(iterable, r)``.

    c3s|]}�|VqdSrKr1�rT�i�rtr1r2rVrWz%random_combination.<locals>.<genexpr>)rArY�sortedrrf)r0rcr/�indicesr1ryr2r#s
r#cs@t|��t���t�fdd�t|�D��}t�fdd�|D��S)aSReturn a random *r* length subsequence of elements in *iterable*,
    allowing individual elements to be repeated.

        >>> random_combination_with_replacement(range(3), 5) # doctest:+SKIP
        (0, 0, 1, 2, 2)

    This equivalent to taking a random selection from
    ``itertools.combinations_with_replacement(iterable, r)``.

    c3s|]}t��VqdSrK)r
rw)r/r1r2rVrWz6random_combination_with_replacement.<locals>.<genexpr>c3s|]}�|VqdSrKr1rwryr1r2rVrW)rArYrzrf)r0rcr{r1)r/rtr2r"sr"c	Cs�t|�}t|�}|dks ||kr$t�d}t|||�}td|d�D]}|||||}qD|dkrn||7}|dks~||kr�t�g}|r�||||d|d}}}||kr�||8}|||||d}}q�|�|d|�q�t|�S)aEquivalent to ``list(combinations(iterable, r))[index]``.

    The subsequences of *iterable* that are of length *r* can be ordered
    lexicographically. :func:`nth_combination` computes the subsequence at
    sort position *index* directly, without computing the previous
    subsequences.

        >>> nth_combination(range(5), 3, 5)
        (0, 3, 4)

    ``ValueError`` will be raised If *r* is negative or greater than the length
    of *iterable*.
    ``IndexError`` will be raised if the given *index* is invalid.
    rrX���)rArY�
ValueError�minrf�
IndexErrorri)	r0rc�indexrtr/�crmrx�resultr1r1r2r"s( rcCst|g|�S)aYield *value*, followed by the elements in *iterator*.

        >>> value = '0'
        >>> iterator = ['1', '2', '3']
        >>> list(prepend(value, iterator))
        ['0', '1', '2', '3']

    To prepend multiple values, see :func:`itertools.chain`
    or :func:`value_chain`.

    )r)�valuer:r1r1r2r Lsr ccsht|�ddd�}t|�}tdg|d�|}t|td|d��D]"}|�|�tttj	||��Vq@dS)aBConvolve the iterable *signal* with the iterable *kernel*.

        >>> signal = (1, 2, 3, 4, 5)
        >>> kernel = [3, 2, 1]
        >>> list(convolve(signal, kernel))
        [3, 8, 14, 20, 26, 14, 5]

    Note: the input arguments are not interchangeable, as the *kernel*
    is immediately consumed and stored.

    Nr|rr6rX)
rArYrrr	rir=r3rBrC)�signalZkernelr/Zwindowr]r1r1r2r[s
r)r)N)N)N)N)N)N)N)NN)N)6�__doc__rP�collectionsr�	itertoolsrrrrrrr	r
rrrB�randomr
rr�__all__r*r(r)rrrr`r!rrrrrr&rJrrL�ImportErrorrr'rrr+r,rrr%r$r#r"rr rr1r1r1r2�<module>sR	0!

(








-



*site-packages/setuptools/_vendor/more_itertools/__init__.py000064400000000122150564745010020322 0ustar00from .more import *  # noqa
from .recipes import *  # noqa

__version__ = '8.8.0'
site-packages/setuptools/_vendor/more_itertools/more.py000064400000346320150564745010017542 0ustar00import warnings

from collections import Counter, defaultdict, deque, abc
from collections.abc import Sequence
from concurrent.futures import ThreadPoolExecutor
from functools import partial, reduce, wraps
from heapq import merge, heapify, heapreplace, heappop
from itertools import (
    chain,
    compress,
    count,
    cycle,
    dropwhile,
    groupby,
    islice,
    repeat,
    starmap,
    takewhile,
    tee,
    zip_longest,
)
from math import exp, factorial, floor, log
from queue import Empty, Queue
from random import random, randrange, uniform
from operator import itemgetter, mul, sub, gt, lt
from sys import hexversion, maxsize
from time import monotonic

from .recipes import (
    consume,
    flatten,
    pairwise,
    powerset,
    take,
    unique_everseen,
)

__all__ = [
    'AbortThread',
    'adjacent',
    'always_iterable',
    'always_reversible',
    'bucket',
    'callback_iter',
    'chunked',
    'circular_shifts',
    'collapse',
    'collate',
    'consecutive_groups',
    'consumer',
    'countable',
    'count_cycle',
    'mark_ends',
    'difference',
    'distinct_combinations',
    'distinct_permutations',
    'distribute',
    'divide',
    'exactly_n',
    'filter_except',
    'first',
    'groupby_transform',
    'ilen',
    'interleave_longest',
    'interleave',
    'intersperse',
    'islice_extended',
    'iterate',
    'ichunked',
    'is_sorted',
    'last',
    'locate',
    'lstrip',
    'make_decorator',
    'map_except',
    'map_reduce',
    'nth_or_last',
    'nth_permutation',
    'nth_product',
    'numeric_range',
    'one',
    'only',
    'padded',
    'partitions',
    'set_partitions',
    'peekable',
    'repeat_last',
    'replace',
    'rlocate',
    'rstrip',
    'run_length',
    'sample',
    'seekable',
    'SequenceView',
    'side_effect',
    'sliced',
    'sort_together',
    'split_at',
    'split_after',
    'split_before',
    'split_when',
    'split_into',
    'spy',
    'stagger',
    'strip',
    'substrings',
    'substrings_indexes',
    'time_limited',
    'unique_to_each',
    'unzip',
    'windowed',
    'with_iter',
    'UnequalIterablesError',
    'zip_equal',
    'zip_offset',
    'windowed_complete',
    'all_unique',
    'value_chain',
    'product_index',
    'combination_index',
    'permutation_index',
]

_marker = object()


def chunked(iterable, n, strict=False):
    """Break *iterable* into lists of length *n*:

        >>> list(chunked([1, 2, 3, 4, 5, 6], 3))
        [[1, 2, 3], [4, 5, 6]]

    By the default, the last yielded list will have fewer than *n* elements
    if the length of *iterable* is not divisible by *n*:

        >>> list(chunked([1, 2, 3, 4, 5, 6, 7, 8], 3))
        [[1, 2, 3], [4, 5, 6], [7, 8]]

    To use a fill-in value instead, see the :func:`grouper` recipe.

    If the length of *iterable* is not divisible by *n* and *strict* is
    ``True``, then ``ValueError`` will be raised before the last
    list is yielded.

    """
    iterator = iter(partial(take, n, iter(iterable)), [])
    if strict:

        def ret():
            for chunk in iterator:
                if len(chunk) != n:
                    raise ValueError('iterable is not divisible by n.')
                yield chunk

        return iter(ret())
    else:
        return iterator


def first(iterable, default=_marker):
    """Return the first item of *iterable*, or *default* if *iterable* is
    empty.

        >>> first([0, 1, 2, 3])
        0
        >>> first([], 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.

    :func:`first` is useful when you have a generator of expensive-to-retrieve
    values and want any arbitrary one. It is marginally shorter than
    ``next(iter(iterable), default)``.

    """
    try:
        return next(iter(iterable))
    except StopIteration as e:
        if default is _marker:
            raise ValueError(
                'first() was called on an empty iterable, and no '
                'default value was provided.'
            ) from e
        return default


def last(iterable, default=_marker):
    """Return the last item of *iterable*, or *default* if *iterable* is
    empty.

        >>> last([0, 1, 2, 3])
        3
        >>> last([], 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.
    """
    try:
        if isinstance(iterable, Sequence):
            return iterable[-1]
        # Work around https://bugs.python.org/issue38525
        elif hasattr(iterable, '__reversed__') and (hexversion != 0x030800F0):
            return next(reversed(iterable))
        else:
            return deque(iterable, maxlen=1)[-1]
    except (IndexError, TypeError, StopIteration):
        if default is _marker:
            raise ValueError(
                'last() was called on an empty iterable, and no default was '
                'provided.'
            )
        return default


def nth_or_last(iterable, n, default=_marker):
    """Return the nth or the last item of *iterable*,
    or *default* if *iterable* is empty.

        >>> nth_or_last([0, 1, 2, 3], 2)
        2
        >>> nth_or_last([0, 1], 2)
        1
        >>> nth_or_last([], 0, 'some default')
        'some default'

    If *default* is not provided and there are no items in the iterable,
    raise ``ValueError``.
    """
    return last(islice(iterable, n + 1), default=default)


class peekable:
    """Wrap an iterator to allow lookahead and prepending elements.

    Call :meth:`peek` on the result to get the value that will be returned
    by :func:`next`. This won't advance the iterator:

        >>> p = peekable(['a', 'b'])
        >>> p.peek()
        'a'
        >>> next(p)
        'a'

    Pass :meth:`peek` a default value to return that instead of raising
    ``StopIteration`` when the iterator is exhausted.

        >>> p = peekable([])
        >>> p.peek('hi')
        'hi'

    peekables also offer a :meth:`prepend` method, which "inserts" items
    at the head of the iterable:

        >>> p = peekable([1, 2, 3])
        >>> p.prepend(10, 11, 12)
        >>> next(p)
        10
        >>> p.peek()
        11
        >>> list(p)
        [11, 12, 1, 2, 3]

    peekables can be indexed. Index 0 is the item that will be returned by
    :func:`next`, index 1 is the item after that, and so on:
    The values up to the given index will be cached.

        >>> p = peekable(['a', 'b', 'c', 'd'])
        >>> p[0]
        'a'
        >>> p[1]
        'b'
        >>> next(p)
        'a'

    Negative indexes are supported, but be aware that they will cache the
    remaining items in the source iterator, which may require significant
    storage.

    To check whether a peekable is exhausted, check its truth value:

        >>> p = peekable(['a', 'b'])
        >>> if p:  # peekable has items
        ...     list(p)
        ['a', 'b']
        >>> if not p:  # peekable is exhausted
        ...     list(p)
        []

    """

    def __init__(self, iterable):
        self._it = iter(iterable)
        self._cache = deque()

    def __iter__(self):
        return self

    def __bool__(self):
        try:
            self.peek()
        except StopIteration:
            return False
        return True

    def peek(self, default=_marker):
        """Return the item that will be next returned from ``next()``.

        Return ``default`` if there are no items left. If ``default`` is not
        provided, raise ``StopIteration``.

        """
        if not self._cache:
            try:
                self._cache.append(next(self._it))
            except StopIteration:
                if default is _marker:
                    raise
                return default
        return self._cache[0]

    def prepend(self, *items):
        """Stack up items to be the next ones returned from ``next()`` or
        ``self.peek()``. The items will be returned in
        first in, first out order::

            >>> p = peekable([1, 2, 3])
            >>> p.prepend(10, 11, 12)
            >>> next(p)
            10
            >>> list(p)
            [11, 12, 1, 2, 3]

        It is possible, by prepending items, to "resurrect" a peekable that
        previously raised ``StopIteration``.

            >>> p = peekable([])
            >>> next(p)
            Traceback (most recent call last):
              ...
            StopIteration
            >>> p.prepend(1)
            >>> next(p)
            1
            >>> next(p)
            Traceback (most recent call last):
              ...
            StopIteration

        """
        self._cache.extendleft(reversed(items))

    def __next__(self):
        if self._cache:
            return self._cache.popleft()

        return next(self._it)

    def _get_slice(self, index):
        # Normalize the slice's arguments
        step = 1 if (index.step is None) else index.step
        if step > 0:
            start = 0 if (index.start is None) else index.start
            stop = maxsize if (index.stop is None) else index.stop
        elif step < 0:
            start = -1 if (index.start is None) else index.start
            stop = (-maxsize - 1) if (index.stop is None) else index.stop
        else:
            raise ValueError('slice step cannot be zero')

        # If either the start or stop index is negative, we'll need to cache
        # the rest of the iterable in order to slice from the right side.
        if (start < 0) or (stop < 0):
            self._cache.extend(self._it)
        # Otherwise we'll need to find the rightmost index and cache to that
        # point.
        else:
            n = min(max(start, stop) + 1, maxsize)
            cache_len = len(self._cache)
            if n >= cache_len:
                self._cache.extend(islice(self._it, n - cache_len))

        return list(self._cache)[index]

    def __getitem__(self, index):
        if isinstance(index, slice):
            return self._get_slice(index)

        cache_len = len(self._cache)
        if index < 0:
            self._cache.extend(self._it)
        elif index >= cache_len:
            self._cache.extend(islice(self._it, index + 1 - cache_len))

        return self._cache[index]


def collate(*iterables, **kwargs):
    """Return a sorted merge of the items from each of several already-sorted
    *iterables*.

        >>> list(collate('ACDZ', 'AZ', 'JKL'))
        ['A', 'A', 'C', 'D', 'J', 'K', 'L', 'Z', 'Z']

    Works lazily, keeping only the next value from each iterable in memory. Use
    :func:`collate` to, for example, perform a n-way mergesort of items that
    don't fit in memory.

    If a *key* function is specified, the iterables will be sorted according
    to its result:

        >>> key = lambda s: int(s)  # Sort by numeric value, not by string
        >>> list(collate(['1', '10'], ['2', '11'], key=key))
        ['1', '2', '10', '11']


    If the *iterables* are sorted in descending order, set *reverse* to
    ``True``:

        >>> list(collate([5, 3, 1], [4, 2, 0], reverse=True))
        [5, 4, 3, 2, 1, 0]

    If the elements of the passed-in iterables are out of order, you might get
    unexpected results.

    On Python 3.5+, this function is an alias for :func:`heapq.merge`.

    """
    warnings.warn(
        "collate is no longer part of more_itertools, use heapq.merge",
        DeprecationWarning,
    )
    return merge(*iterables, **kwargs)


def consumer(func):
    """Decorator that automatically advances a PEP-342-style "reverse iterator"
    to its first yield point so you don't have to call ``next()`` on it
    manually.

        >>> @consumer
        ... def tally():
        ...     i = 0
        ...     while True:
        ...         print('Thing number %s is %s.' % (i, (yield)))
        ...         i += 1
        ...
        >>> t = tally()
        >>> t.send('red')
        Thing number 0 is red.
        >>> t.send('fish')
        Thing number 1 is fish.

    Without the decorator, you would have to call ``next(t)`` before
    ``t.send()`` could be used.

    """

    @wraps(func)
    def wrapper(*args, **kwargs):
        gen = func(*args, **kwargs)
        next(gen)
        return gen

    return wrapper


def ilen(iterable):
    """Return the number of items in *iterable*.

        >>> ilen(x for x in range(1000000) if x % 3 == 0)
        333334

    This consumes the iterable, so handle with care.

    """
    # This approach was selected because benchmarks showed it's likely the
    # fastest of the known implementations at the time of writing.
    # See GitHub tracker: #236, #230.
    counter = count()
    deque(zip(iterable, counter), maxlen=0)
    return next(counter)


def iterate(func, start):
    """Return ``start``, ``func(start)``, ``func(func(start))``, ...

    >>> from itertools import islice
    >>> list(islice(iterate(lambda x: 2*x, 1), 10))
    [1, 2, 4, 8, 16, 32, 64, 128, 256, 512]

    """
    while True:
        yield start
        start = func(start)


def with_iter(context_manager):
    """Wrap an iterable in a ``with`` statement, so it closes once exhausted.

    For example, this will close the file when the iterator is exhausted::

        upper_lines = (line.upper() for line in with_iter(open('foo')))

    Any context manager which returns an iterable is a candidate for
    ``with_iter``.

    """
    with context_manager as iterable:
        yield from iterable


def one(iterable, too_short=None, too_long=None):
    """Return the first item from *iterable*, which is expected to contain only
    that item. Raise an exception if *iterable* is empty or has more than one
    item.

    :func:`one` is useful for ensuring that an iterable contains only one item.
    For example, it can be used to retrieve the result of a database query
    that is expected to return a single row.

    If *iterable* is empty, ``ValueError`` will be raised. You may specify a
    different exception with the *too_short* keyword:

        >>> it = []
        >>> one(it)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        ValueError: too many items in iterable (expected 1)'
        >>> too_short = IndexError('too few items')
        >>> one(it, too_short=too_short)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        IndexError: too few items

    Similarly, if *iterable* contains more than one item, ``ValueError`` will
    be raised. You may specify a different exception with the *too_long*
    keyword:

        >>> it = ['too', 'many']
        >>> one(it)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        ValueError: Expected exactly one item in iterable, but got 'too',
        'many', and perhaps more.
        >>> too_long = RuntimeError
        >>> one(it, too_long=too_long)  # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        RuntimeError

    Note that :func:`one` attempts to advance *iterable* twice to ensure there
    is only one item. See :func:`spy` or :func:`peekable` to check iterable
    contents less destructively.

    """
    it = iter(iterable)

    try:
        first_value = next(it)
    except StopIteration as e:
        raise (
            too_short or ValueError('too few items in iterable (expected 1)')
        ) from e

    try:
        second_value = next(it)
    except StopIteration:
        pass
    else:
        msg = (
            'Expected exactly one item in iterable, but got {!r}, {!r}, '
            'and perhaps more.'.format(first_value, second_value)
        )
        raise too_long or ValueError(msg)

    return first_value


def distinct_permutations(iterable, r=None):
    """Yield successive distinct permutations of the elements in *iterable*.

        >>> sorted(distinct_permutations([1, 0, 1]))
        [(0, 1, 1), (1, 0, 1), (1, 1, 0)]

    Equivalent to ``set(permutations(iterable))``, except duplicates are not
    generated and thrown away. For larger input sequences this is much more
    efficient.

    Duplicate permutations arise when there are duplicated elements in the
    input iterable. The number of items returned is
    `n! / (x_1! * x_2! * ... * x_n!)`, where `n` is the total number of
    items input, and each `x_i` is the count of a distinct item in the input
    sequence.

    If *r* is given, only the *r*-length permutations are yielded.

        >>> sorted(distinct_permutations([1, 0, 1], r=2))
        [(0, 1), (1, 0), (1, 1)]
        >>> sorted(distinct_permutations(range(3), r=2))
        [(0, 1), (0, 2), (1, 0), (1, 2), (2, 0), (2, 1)]

    """
    # Algorithm: https://w.wiki/Qai
    def _full(A):
        while True:
            # Yield the permutation we have
            yield tuple(A)

            # Find the largest index i such that A[i] < A[i + 1]
            for i in range(size - 2, -1, -1):
                if A[i] < A[i + 1]:
                    break
            #  If no such index exists, this permutation is the last one
            else:
                return

            # Find the largest index j greater than j such that A[i] < A[j]
            for j in range(size - 1, i, -1):
                if A[i] < A[j]:
                    break

            # Swap the value of A[i] with that of A[j], then reverse the
            # sequence from A[i + 1] to form the new permutation
            A[i], A[j] = A[j], A[i]
            A[i + 1 :] = A[: i - size : -1]  # A[i + 1:][::-1]

    # Algorithm: modified from the above
    def _partial(A, r):
        # Split A into the first r items and the last r items
        head, tail = A[:r], A[r:]
        right_head_indexes = range(r - 1, -1, -1)
        left_tail_indexes = range(len(tail))

        while True:
            # Yield the permutation we have
            yield tuple(head)

            # Starting from the right, find the first index of the head with
            # value smaller than the maximum value of the tail - call it i.
            pivot = tail[-1]
            for i in right_head_indexes:
                if head[i] < pivot:
                    break
                pivot = head[i]
            else:
                return

            # Starting from the left, find the first value of the tail
            # with a value greater than head[i] and swap.
            for j in left_tail_indexes:
                if tail[j] > head[i]:
                    head[i], tail[j] = tail[j], head[i]
                    break
            # If we didn't find one, start from the right and find the first
            # index of the head with a value greater than head[i] and swap.
            else:
                for j in right_head_indexes:
                    if head[j] > head[i]:
                        head[i], head[j] = head[j], head[i]
                        break

            # Reverse head[i + 1:] and swap it with tail[:r - (i + 1)]
            tail += head[: i - r : -1]  # head[i + 1:][::-1]
            i += 1
            head[i:], tail[:] = tail[: r - i], tail[r - i :]

    items = sorted(iterable)

    size = len(items)
    if r is None:
        r = size

    if 0 < r <= size:
        return _full(items) if (r == size) else _partial(items, r)

    return iter(() if r else ((),))


def intersperse(e, iterable, n=1):
    """Intersperse filler element *e* among the items in *iterable*, leaving
    *n* items between each filler element.

        >>> list(intersperse('!', [1, 2, 3, 4, 5]))
        [1, '!', 2, '!', 3, '!', 4, '!', 5]

        >>> list(intersperse(None, [1, 2, 3, 4, 5], n=2))
        [1, 2, None, 3, 4, None, 5]

    """
    if n == 0:
        raise ValueError('n must be > 0')
    elif n == 1:
        # interleave(repeat(e), iterable) -> e, x_0, e, e, x_1, e, x_2...
        # islice(..., 1, None) -> x_0, e, e, x_1, e, x_2...
        return islice(interleave(repeat(e), iterable), 1, None)
    else:
        # interleave(filler, chunks) -> [e], [x_0, x_1], [e], [x_2, x_3]...
        # islice(..., 1, None) -> [x_0, x_1], [e], [x_2, x_3]...
        # flatten(...) -> x_0, x_1, e, x_2, x_3...
        filler = repeat([e])
        chunks = chunked(iterable, n)
        return flatten(islice(interleave(filler, chunks), 1, None))


def unique_to_each(*iterables):
    """Return the elements from each of the input iterables that aren't in the
    other input iterables.

    For example, suppose you have a set of packages, each with a set of
    dependencies::

        {'pkg_1': {'A', 'B'}, 'pkg_2': {'B', 'C'}, 'pkg_3': {'B', 'D'}}

    If you remove one package, which dependencies can also be removed?

    If ``pkg_1`` is removed, then ``A`` is no longer necessary - it is not
    associated with ``pkg_2`` or ``pkg_3``. Similarly, ``C`` is only needed for
    ``pkg_2``, and ``D`` is only needed for ``pkg_3``::

        >>> unique_to_each({'A', 'B'}, {'B', 'C'}, {'B', 'D'})
        [['A'], ['C'], ['D']]

    If there are duplicates in one input iterable that aren't in the others
    they will be duplicated in the output. Input order is preserved::

        >>> unique_to_each("mississippi", "missouri")
        [['p', 'p'], ['o', 'u', 'r']]

    It is assumed that the elements of each iterable are hashable.

    """
    pool = [list(it) for it in iterables]
    counts = Counter(chain.from_iterable(map(set, pool)))
    uniques = {element for element in counts if counts[element] == 1}
    return [list(filter(uniques.__contains__, it)) for it in pool]


def windowed(seq, n, fillvalue=None, step=1):
    """Return a sliding window of width *n* over the given iterable.

        >>> all_windows = windowed([1, 2, 3, 4, 5], 3)
        >>> list(all_windows)
        [(1, 2, 3), (2, 3, 4), (3, 4, 5)]

    When the window is larger than the iterable, *fillvalue* is used in place
    of missing values:

        >>> list(windowed([1, 2, 3], 4))
        [(1, 2, 3, None)]

    Each window will advance in increments of *step*:

        >>> list(windowed([1, 2, 3, 4, 5, 6], 3, fillvalue='!', step=2))
        [(1, 2, 3), (3, 4, 5), (5, 6, '!')]

    To slide into the iterable's items, use :func:`chain` to add filler items
    to the left:

        >>> iterable = [1, 2, 3, 4]
        >>> n = 3
        >>> padding = [None] * (n - 1)
        >>> list(windowed(chain(padding, iterable), 3))
        [(None, None, 1), (None, 1, 2), (1, 2, 3), (2, 3, 4)]
    """
    if n < 0:
        raise ValueError('n must be >= 0')
    if n == 0:
        yield tuple()
        return
    if step < 1:
        raise ValueError('step must be >= 1')

    window = deque(maxlen=n)
    i = n
    for _ in map(window.append, seq):
        i -= 1
        if not i:
            i = step
            yield tuple(window)

    size = len(window)
    if size < n:
        yield tuple(chain(window, repeat(fillvalue, n - size)))
    elif 0 < i < min(step, n):
        window += (fillvalue,) * i
        yield tuple(window)


def substrings(iterable):
    """Yield all of the substrings of *iterable*.

        >>> [''.join(s) for s in substrings('more')]
        ['m', 'o', 'r', 'e', 'mo', 'or', 're', 'mor', 'ore', 'more']

    Note that non-string iterables can also be subdivided.

        >>> list(substrings([0, 1, 2]))
        [(0,), (1,), (2,), (0, 1), (1, 2), (0, 1, 2)]

    """
    # The length-1 substrings
    seq = []
    for item in iter(iterable):
        seq.append(item)
        yield (item,)
    seq = tuple(seq)
    item_count = len(seq)

    # And the rest
    for n in range(2, item_count + 1):
        for i in range(item_count - n + 1):
            yield seq[i : i + n]


def substrings_indexes(seq, reverse=False):
    """Yield all substrings and their positions in *seq*

    The items yielded will be a tuple of the form ``(substr, i, j)``, where
    ``substr == seq[i:j]``.

    This function only works for iterables that support slicing, such as
    ``str`` objects.

    >>> for item in substrings_indexes('more'):
    ...    print(item)
    ('m', 0, 1)
    ('o', 1, 2)
    ('r', 2, 3)
    ('e', 3, 4)
    ('mo', 0, 2)
    ('or', 1, 3)
    ('re', 2, 4)
    ('mor', 0, 3)
    ('ore', 1, 4)
    ('more', 0, 4)

    Set *reverse* to ``True`` to yield the same items in the opposite order.


    """
    r = range(1, len(seq) + 1)
    if reverse:
        r = reversed(r)
    return (
        (seq[i : i + L], i, i + L) for L in r for i in range(len(seq) - L + 1)
    )


class bucket:
    """Wrap *iterable* and return an object that buckets it iterable into
    child iterables based on a *key* function.

        >>> iterable = ['a1', 'b1', 'c1', 'a2', 'b2', 'c2', 'b3']
        >>> s = bucket(iterable, key=lambda x: x[0])  # Bucket by 1st character
        >>> sorted(list(s))  # Get the keys
        ['a', 'b', 'c']
        >>> a_iterable = s['a']
        >>> next(a_iterable)
        'a1'
        >>> next(a_iterable)
        'a2'
        >>> list(s['b'])
        ['b1', 'b2', 'b3']

    The original iterable will be advanced and its items will be cached until
    they are used by the child iterables. This may require significant storage.

    By default, attempting to select a bucket to which no items belong  will
    exhaust the iterable and cache all values.
    If you specify a *validator* function, selected buckets will instead be
    checked against it.

        >>> from itertools import count
        >>> it = count(1, 2)  # Infinite sequence of odd numbers
        >>> key = lambda x: x % 10  # Bucket by last digit
        >>> validator = lambda x: x in {1, 3, 5, 7, 9}  # Odd digits only
        >>> s = bucket(it, key=key, validator=validator)
        >>> 2 in s
        False
        >>> list(s[2])
        []

    """

    def __init__(self, iterable, key, validator=None):
        self._it = iter(iterable)
        self._key = key
        self._cache = defaultdict(deque)
        self._validator = validator or (lambda x: True)

    def __contains__(self, value):
        if not self._validator(value):
            return False

        try:
            item = next(self[value])
        except StopIteration:
            return False
        else:
            self._cache[value].appendleft(item)

        return True

    def _get_values(self, value):
        """
        Helper to yield items from the parent iterator that match *value*.
        Items that don't match are stored in the local cache as they
        are encountered.
        """
        while True:
            # If we've cached some items that match the target value, emit
            # the first one and evict it from the cache.
            if self._cache[value]:
                yield self._cache[value].popleft()
            # Otherwise we need to advance the parent iterator to search for
            # a matching item, caching the rest.
            else:
                while True:
                    try:
                        item = next(self._it)
                    except StopIteration:
                        return
                    item_value = self._key(item)
                    if item_value == value:
                        yield item
                        break
                    elif self._validator(item_value):
                        self._cache[item_value].append(item)

    def __iter__(self):
        for item in self._it:
            item_value = self._key(item)
            if self._validator(item_value):
                self._cache[item_value].append(item)

        yield from self._cache.keys()

    def __getitem__(self, value):
        if not self._validator(value):
            return iter(())

        return self._get_values(value)


def spy(iterable, n=1):
    """Return a 2-tuple with a list containing the first *n* elements of
    *iterable*, and an iterator with the same items as *iterable*.
    This allows you to "look ahead" at the items in the iterable without
    advancing it.

    There is one item in the list by default:

        >>> iterable = 'abcdefg'
        >>> head, iterable = spy(iterable)
        >>> head
        ['a']
        >>> list(iterable)
        ['a', 'b', 'c', 'd', 'e', 'f', 'g']

    You may use unpacking to retrieve items instead of lists:

        >>> (head,), iterable = spy('abcdefg')
        >>> head
        'a'
        >>> (first, second), iterable = spy('abcdefg', 2)
        >>> first
        'a'
        >>> second
        'b'

    The number of items requested can be larger than the number of items in
    the iterable:

        >>> iterable = [1, 2, 3, 4, 5]
        >>> head, iterable = spy(iterable, 10)
        >>> head
        [1, 2, 3, 4, 5]
        >>> list(iterable)
        [1, 2, 3, 4, 5]

    """
    it = iter(iterable)
    head = take(n, it)

    return head.copy(), chain(head, it)


def interleave(*iterables):
    """Return a new iterable yielding from each iterable in turn,
    until the shortest is exhausted.

        >>> list(interleave([1, 2, 3], [4, 5], [6, 7, 8]))
        [1, 4, 6, 2, 5, 7]

    For a version that doesn't terminate after the shortest iterable is
    exhausted, see :func:`interleave_longest`.

    """
    return chain.from_iterable(zip(*iterables))


def interleave_longest(*iterables):
    """Return a new iterable yielding from each iterable in turn,
    skipping any that are exhausted.

        >>> list(interleave_longest([1, 2, 3], [4, 5], [6, 7, 8]))
        [1, 4, 6, 2, 5, 7, 3, 8]

    This function produces the same output as :func:`roundrobin`, but may
    perform better for some inputs (in particular when the number of iterables
    is large).

    """
    i = chain.from_iterable(zip_longest(*iterables, fillvalue=_marker))
    return (x for x in i if x is not _marker)


def collapse(iterable, base_type=None, levels=None):
    """Flatten an iterable with multiple levels of nesting (e.g., a list of
    lists of tuples) into non-iterable types.

        >>> iterable = [(1, 2), ([3, 4], [[5], [6]])]
        >>> list(collapse(iterable))
        [1, 2, 3, 4, 5, 6]

    Binary and text strings are not considered iterable and
    will not be collapsed.

    To avoid collapsing other types, specify *base_type*:

        >>> iterable = ['ab', ('cd', 'ef'), ['gh', 'ij']]
        >>> list(collapse(iterable, base_type=tuple))
        ['ab', ('cd', 'ef'), 'gh', 'ij']

    Specify *levels* to stop flattening after a certain level:

    >>> iterable = [('a', ['b']), ('c', ['d'])]
    >>> list(collapse(iterable))  # Fully flattened
    ['a', 'b', 'c', 'd']
    >>> list(collapse(iterable, levels=1))  # Only one level flattened
    ['a', ['b'], 'c', ['d']]

    """

    def walk(node, level):
        if (
            ((levels is not None) and (level > levels))
            or isinstance(node, (str, bytes))
            or ((base_type is not None) and isinstance(node, base_type))
        ):
            yield node
            return

        try:
            tree = iter(node)
        except TypeError:
            yield node
            return
        else:
            for child in tree:
                yield from walk(child, level + 1)

    yield from walk(iterable, 0)


def side_effect(func, iterable, chunk_size=None, before=None, after=None):
    """Invoke *func* on each item in *iterable* (or on each *chunk_size* group
    of items) before yielding the item.

    `func` must be a function that takes a single argument. Its return value
    will be discarded.

    *before* and *after* are optional functions that take no arguments. They
    will be executed before iteration starts and after it ends, respectively.

    `side_effect` can be used for logging, updating progress bars, or anything
    that is not functionally "pure."

    Emitting a status message:

        >>> from more_itertools import consume
        >>> func = lambda item: print('Received {}'.format(item))
        >>> consume(side_effect(func, range(2)))
        Received 0
        Received 1

    Operating on chunks of items:

        >>> pair_sums = []
        >>> func = lambda chunk: pair_sums.append(sum(chunk))
        >>> list(side_effect(func, [0, 1, 2, 3, 4, 5], 2))
        [0, 1, 2, 3, 4, 5]
        >>> list(pair_sums)
        [1, 5, 9]

    Writing to a file-like object:

        >>> from io import StringIO
        >>> from more_itertools import consume
        >>> f = StringIO()
        >>> func = lambda x: print(x, file=f)
        >>> before = lambda: print(u'HEADER', file=f)
        >>> after = f.close
        >>> it = [u'a', u'b', u'c']
        >>> consume(side_effect(func, it, before=before, after=after))
        >>> f.closed
        True

    """
    try:
        if before is not None:
            before()

        if chunk_size is None:
            for item in iterable:
                func(item)
                yield item
        else:
            for chunk in chunked(iterable, chunk_size):
                func(chunk)
                yield from chunk
    finally:
        if after is not None:
            after()


def sliced(seq, n, strict=False):
    """Yield slices of length *n* from the sequence *seq*.

    >>> list(sliced((1, 2, 3, 4, 5, 6), 3))
    [(1, 2, 3), (4, 5, 6)]

    By the default, the last yielded slice will have fewer than *n* elements
    if the length of *seq* is not divisible by *n*:

    >>> list(sliced((1, 2, 3, 4, 5, 6, 7, 8), 3))
    [(1, 2, 3), (4, 5, 6), (7, 8)]

    If the length of *seq* is not divisible by *n* and *strict* is
    ``True``, then ``ValueError`` will be raised before the last
    slice is yielded.

    This function will only work for iterables that support slicing.
    For non-sliceable iterables, see :func:`chunked`.

    """
    iterator = takewhile(len, (seq[i : i + n] for i in count(0, n)))
    if strict:

        def ret():
            for _slice in iterator:
                if len(_slice) != n:
                    raise ValueError("seq is not divisible by n.")
                yield _slice

        return iter(ret())
    else:
        return iterator


def split_at(iterable, pred, maxsplit=-1, keep_separator=False):
    """Yield lists of items from *iterable*, where each list is delimited by
    an item where callable *pred* returns ``True``.

        >>> list(split_at('abcdcba', lambda x: x == 'b'))
        [['a'], ['c', 'd', 'c'], ['a']]

        >>> list(split_at(range(10), lambda n: n % 2 == 1))
        [[0], [2], [4], [6], [8], []]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_at(range(10), lambda n: n % 2 == 1, maxsplit=2))
        [[0], [2], [4, 5, 6, 7, 8, 9]]

    By default, the delimiting items are not included in the output.
    The include them, set *keep_separator* to ``True``.

        >>> list(split_at('abcdcba', lambda x: x == 'b', keep_separator=True))
        [['a'], ['b'], ['c', 'd', 'c'], ['b'], ['a']]

    """
    if maxsplit == 0:
        yield list(iterable)
        return

    buf = []
    it = iter(iterable)
    for item in it:
        if pred(item):
            yield buf
            if keep_separator:
                yield [item]
            if maxsplit == 1:
                yield list(it)
                return
            buf = []
            maxsplit -= 1
        else:
            buf.append(item)
    yield buf


def split_before(iterable, pred, maxsplit=-1):
    """Yield lists of items from *iterable*, where each list ends just before
    an item for which callable *pred* returns ``True``:

        >>> list(split_before('OneTwo', lambda s: s.isupper()))
        [['O', 'n', 'e'], ['T', 'w', 'o']]

        >>> list(split_before(range(10), lambda n: n % 3 == 0))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8], [9]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_before(range(10), lambda n: n % 3 == 0, maxsplit=2))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8, 9]]
    """
    if maxsplit == 0:
        yield list(iterable)
        return

    buf = []
    it = iter(iterable)
    for item in it:
        if pred(item) and buf:
            yield buf
            if maxsplit == 1:
                yield [item] + list(it)
                return
            buf = []
            maxsplit -= 1
        buf.append(item)
    if buf:
        yield buf


def split_after(iterable, pred, maxsplit=-1):
    """Yield lists of items from *iterable*, where each list ends with an
    item where callable *pred* returns ``True``:

        >>> list(split_after('one1two2', lambda s: s.isdigit()))
        [['o', 'n', 'e', '1'], ['t', 'w', 'o', '2']]

        >>> list(split_after(range(10), lambda n: n % 3 == 0))
        [[0], [1, 2, 3], [4, 5, 6], [7, 8, 9]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_after(range(10), lambda n: n % 3 == 0, maxsplit=2))
        [[0], [1, 2, 3], [4, 5, 6, 7, 8, 9]]

    """
    if maxsplit == 0:
        yield list(iterable)
        return

    buf = []
    it = iter(iterable)
    for item in it:
        buf.append(item)
        if pred(item) and buf:
            yield buf
            if maxsplit == 1:
                yield list(it)
                return
            buf = []
            maxsplit -= 1
    if buf:
        yield buf


def split_when(iterable, pred, maxsplit=-1):
    """Split *iterable* into pieces based on the output of *pred*.
    *pred* should be a function that takes successive pairs of items and
    returns ``True`` if the iterable should be split in between them.

    For example, to find runs of increasing numbers, split the iterable when
    element ``i`` is larger than element ``i + 1``:

        >>> list(split_when([1, 2, 3, 3, 2, 5, 2, 4, 2], lambda x, y: x > y))
        [[1, 2, 3, 3], [2, 5], [2, 4], [2]]

    At most *maxsplit* splits are done. If *maxsplit* is not specified or -1,
    then there is no limit on the number of splits:

        >>> list(split_when([1, 2, 3, 3, 2, 5, 2, 4, 2],
        ...                 lambda x, y: x > y, maxsplit=2))
        [[1, 2, 3, 3], [2, 5], [2, 4, 2]]

    """
    if maxsplit == 0:
        yield list(iterable)
        return

    it = iter(iterable)
    try:
        cur_item = next(it)
    except StopIteration:
        return

    buf = [cur_item]
    for next_item in it:
        if pred(cur_item, next_item):
            yield buf
            if maxsplit == 1:
                yield [next_item] + list(it)
                return
            buf = []
            maxsplit -= 1

        buf.append(next_item)
        cur_item = next_item

    yield buf


def split_into(iterable, sizes):
    """Yield a list of sequential items from *iterable* of length 'n' for each
    integer 'n' in *sizes*.

        >>> list(split_into([1,2,3,4,5,6], [1,2,3]))
        [[1], [2, 3], [4, 5, 6]]

    If the sum of *sizes* is smaller than the length of *iterable*, then the
    remaining items of *iterable* will not be returned.

        >>> list(split_into([1,2,3,4,5,6], [2,3]))
        [[1, 2], [3, 4, 5]]

    If the sum of *sizes* is larger than the length of *iterable*, fewer items
    will be returned in the iteration that overruns *iterable* and further
    lists will be empty:

        >>> list(split_into([1,2,3,4], [1,2,3,4]))
        [[1], [2, 3], [4], []]

    When a ``None`` object is encountered in *sizes*, the returned list will
    contain items up to the end of *iterable* the same way that itertools.slice
    does:

        >>> list(split_into([1,2,3,4,5,6,7,8,9,0], [2,3,None]))
        [[1, 2], [3, 4, 5], [6, 7, 8, 9, 0]]

    :func:`split_into` can be useful for grouping a series of items where the
    sizes of the groups are not uniform. An example would be where in a row
    from a table, multiple columns represent elements of the same feature
    (e.g. a point represented by x,y,z) but, the format is not the same for
    all columns.
    """
    # convert the iterable argument into an iterator so its contents can
    # be consumed by islice in case it is a generator
    it = iter(iterable)

    for size in sizes:
        if size is None:
            yield list(it)
            return
        else:
            yield list(islice(it, size))


def padded(iterable, fillvalue=None, n=None, next_multiple=False):
    """Yield the elements from *iterable*, followed by *fillvalue*, such that
    at least *n* items are emitted.

        >>> list(padded([1, 2, 3], '?', 5))
        [1, 2, 3, '?', '?']

    If *next_multiple* is ``True``, *fillvalue* will be emitted until the
    number of items emitted is a multiple of *n*::

        >>> list(padded([1, 2, 3, 4], n=3, next_multiple=True))
        [1, 2, 3, 4, None, None]

    If *n* is ``None``, *fillvalue* will be emitted indefinitely.

    """
    it = iter(iterable)
    if n is None:
        yield from chain(it, repeat(fillvalue))
    elif n < 1:
        raise ValueError('n must be at least 1')
    else:
        item_count = 0
        for item in it:
            yield item
            item_count += 1

        remaining = (n - item_count) % n if next_multiple else n - item_count
        for _ in range(remaining):
            yield fillvalue


def repeat_last(iterable, default=None):
    """After the *iterable* is exhausted, keep yielding its last element.

        >>> list(islice(repeat_last(range(3)), 5))
        [0, 1, 2, 2, 2]

    If the iterable is empty, yield *default* forever::

        >>> list(islice(repeat_last(range(0), 42), 5))
        [42, 42, 42, 42, 42]

    """
    item = _marker
    for item in iterable:
        yield item
    final = default if item is _marker else item
    yield from repeat(final)


def distribute(n, iterable):
    """Distribute the items from *iterable* among *n* smaller iterables.

        >>> group_1, group_2 = distribute(2, [1, 2, 3, 4, 5, 6])
        >>> list(group_1)
        [1, 3, 5]
        >>> list(group_2)
        [2, 4, 6]

    If the length of *iterable* is not evenly divisible by *n*, then the
    length of the returned iterables will not be identical:

        >>> children = distribute(3, [1, 2, 3, 4, 5, 6, 7])
        >>> [list(c) for c in children]
        [[1, 4, 7], [2, 5], [3, 6]]

    If the length of *iterable* is smaller than *n*, then the last returned
    iterables will be empty:

        >>> children = distribute(5, [1, 2, 3])
        >>> [list(c) for c in children]
        [[1], [2], [3], [], []]

    This function uses :func:`itertools.tee` and may require significant
    storage. If you need the order items in the smaller iterables to match the
    original iterable, see :func:`divide`.

    """
    if n < 1:
        raise ValueError('n must be at least 1')

    children = tee(iterable, n)
    return [islice(it, index, None, n) for index, it in enumerate(children)]


def stagger(iterable, offsets=(-1, 0, 1), longest=False, fillvalue=None):
    """Yield tuples whose elements are offset from *iterable*.
    The amount by which the `i`-th item in each tuple is offset is given by
    the `i`-th item in *offsets*.

        >>> list(stagger([0, 1, 2, 3]))
        [(None, 0, 1), (0, 1, 2), (1, 2, 3)]
        >>> list(stagger(range(8), offsets=(0, 2, 4)))
        [(0, 2, 4), (1, 3, 5), (2, 4, 6), (3, 5, 7)]

    By default, the sequence will end when the final element of a tuple is the
    last item in the iterable. To continue until the first element of a tuple
    is the last item in the iterable, set *longest* to ``True``::

        >>> list(stagger([0, 1, 2, 3], longest=True))
        [(None, 0, 1), (0, 1, 2), (1, 2, 3), (2, 3, None), (3, None, None)]

    By default, ``None`` will be used to replace offsets beyond the end of the
    sequence. Specify *fillvalue* to use some other value.

    """
    children = tee(iterable, len(offsets))

    return zip_offset(
        *children, offsets=offsets, longest=longest, fillvalue=fillvalue
    )


class UnequalIterablesError(ValueError):
    def __init__(self, details=None):
        msg = 'Iterables have different lengths'
        if details is not None:
            msg += (': index 0 has length {}; index {} has length {}').format(
                *details
            )

        super().__init__(msg)


def _zip_equal_generator(iterables):
    for combo in zip_longest(*iterables, fillvalue=_marker):
        for val in combo:
            if val is _marker:
                raise UnequalIterablesError()
        yield combo


def zip_equal(*iterables):
    """``zip`` the input *iterables* together, but raise
    ``UnequalIterablesError`` if they aren't all the same length.

        >>> it_1 = range(3)
        >>> it_2 = iter('abc')
        >>> list(zip_equal(it_1, it_2))
        [(0, 'a'), (1, 'b'), (2, 'c')]

        >>> it_1 = range(3)
        >>> it_2 = iter('abcd')
        >>> list(zip_equal(it_1, it_2)) # doctest: +IGNORE_EXCEPTION_DETAIL
        Traceback (most recent call last):
        ...
        more_itertools.more.UnequalIterablesError: Iterables have different
        lengths

    """
    if hexversion >= 0x30A00A6:
        warnings.warn(
            (
                'zip_equal will be removed in a future version of '
                'more-itertools. Use the builtin zip function with '
                'strict=True instead.'
            ),
            DeprecationWarning,
        )
    # Check whether the iterables are all the same size.
    try:
        first_size = len(iterables[0])
        for i, it in enumerate(iterables[1:], 1):
            size = len(it)
            if size != first_size:
                break
        else:
            # If we didn't break out, we can use the built-in zip.
            return zip(*iterables)

        # If we did break out, there was a mismatch.
        raise UnequalIterablesError(details=(first_size, i, size))
    # If any one of the iterables didn't have a length, start reading
    # them until one runs out.
    except TypeError:
        return _zip_equal_generator(iterables)


def zip_offset(*iterables, offsets, longest=False, fillvalue=None):
    """``zip`` the input *iterables* together, but offset the `i`-th iterable
    by the `i`-th item in *offsets*.

        >>> list(zip_offset('0123', 'abcdef', offsets=(0, 1)))
        [('0', 'b'), ('1', 'c'), ('2', 'd'), ('3', 'e')]

    This can be used as a lightweight alternative to SciPy or pandas to analyze
    data sets in which some series have a lead or lag relationship.

    By default, the sequence will end when the shortest iterable is exhausted.
    To continue until the longest iterable is exhausted, set *longest* to
    ``True``.

        >>> list(zip_offset('0123', 'abcdef', offsets=(0, 1), longest=True))
        [('0', 'b'), ('1', 'c'), ('2', 'd'), ('3', 'e'), (None, 'f')]

    By default, ``None`` will be used to replace offsets beyond the end of the
    sequence. Specify *fillvalue* to use some other value.

    """
    if len(iterables) != len(offsets):
        raise ValueError("Number of iterables and offsets didn't match")

    staggered = []
    for it, n in zip(iterables, offsets):
        if n < 0:
            staggered.append(chain(repeat(fillvalue, -n), it))
        elif n > 0:
            staggered.append(islice(it, n, None))
        else:
            staggered.append(it)

    if longest:
        return zip_longest(*staggered, fillvalue=fillvalue)

    return zip(*staggered)


def sort_together(iterables, key_list=(0,), key=None, reverse=False):
    """Return the input iterables sorted together, with *key_list* as the
    priority for sorting. All iterables are trimmed to the length of the
    shortest one.

    This can be used like the sorting function in a spreadsheet. If each
    iterable represents a column of data, the key list determines which
    columns are used for sorting.

    By default, all iterables are sorted using the ``0``-th iterable::

        >>> iterables = [(4, 3, 2, 1), ('a', 'b', 'c', 'd')]
        >>> sort_together(iterables)
        [(1, 2, 3, 4), ('d', 'c', 'b', 'a')]

    Set a different key list to sort according to another iterable.
    Specifying multiple keys dictates how ties are broken::

        >>> iterables = [(3, 1, 2), (0, 1, 0), ('c', 'b', 'a')]
        >>> sort_together(iterables, key_list=(1, 2))
        [(2, 3, 1), (0, 0, 1), ('a', 'c', 'b')]

    To sort by a function of the elements of the iterable, pass a *key*
    function. Its arguments are the elements of the iterables corresponding to
    the key list::

        >>> names = ('a', 'b', 'c')
        >>> lengths = (1, 2, 3)
        >>> widths = (5, 2, 1)
        >>> def area(length, width):
        ...     return length * width
        >>> sort_together([names, lengths, widths], key_list=(1, 2), key=area)
        [('c', 'b', 'a'), (3, 2, 1), (1, 2, 5)]

    Set *reverse* to ``True`` to sort in descending order.

        >>> sort_together([(1, 2, 3), ('c', 'b', 'a')], reverse=True)
        [(3, 2, 1), ('a', 'b', 'c')]

    """
    if key is None:
        # if there is no key function, the key argument to sorted is an
        # itemgetter
        key_argument = itemgetter(*key_list)
    else:
        # if there is a key function, call it with the items at the offsets
        # specified by the key function as arguments
        key_list = list(key_list)
        if len(key_list) == 1:
            # if key_list contains a single item, pass the item at that offset
            # as the only argument to the key function
            key_offset = key_list[0]
            key_argument = lambda zipped_items: key(zipped_items[key_offset])
        else:
            # if key_list contains multiple items, use itemgetter to return a
            # tuple of items, which we pass as *args to the key function
            get_key_items = itemgetter(*key_list)
            key_argument = lambda zipped_items: key(
                *get_key_items(zipped_items)
            )

    return list(
        zip(*sorted(zip(*iterables), key=key_argument, reverse=reverse))
    )


def unzip(iterable):
    """The inverse of :func:`zip`, this function disaggregates the elements
    of the zipped *iterable*.

    The ``i``-th iterable contains the ``i``-th element from each element
    of the zipped iterable. The first element is used to to determine the
    length of the remaining elements.

        >>> iterable = [('a', 1), ('b', 2), ('c', 3), ('d', 4)]
        >>> letters, numbers = unzip(iterable)
        >>> list(letters)
        ['a', 'b', 'c', 'd']
        >>> list(numbers)
        [1, 2, 3, 4]

    This is similar to using ``zip(*iterable)``, but it avoids reading
    *iterable* into memory. Note, however, that this function uses
    :func:`itertools.tee` and thus may require significant storage.

    """
    head, iterable = spy(iter(iterable))
    if not head:
        # empty iterable, e.g. zip([], [], [])
        return ()
    # spy returns a one-length iterable as head
    head = head[0]
    iterables = tee(iterable, len(head))

    def itemgetter(i):
        def getter(obj):
            try:
                return obj[i]
            except IndexError:
                # basically if we have an iterable like
                # iter([(1, 2, 3), (4, 5), (6,)])
                # the second unzipped iterable would fail at the third tuple
                # since it would try to access tup[1]
                # same with the third unzipped iterable and the second tuple
                # to support these "improperly zipped" iterables,
                # we create a custom itemgetter
                # which just stops the unzipped iterables
                # at first length mismatch
                raise StopIteration

        return getter

    return tuple(map(itemgetter(i), it) for i, it in enumerate(iterables))


def divide(n, iterable):
    """Divide the elements from *iterable* into *n* parts, maintaining
    order.

        >>> group_1, group_2 = divide(2, [1, 2, 3, 4, 5, 6])
        >>> list(group_1)
        [1, 2, 3]
        >>> list(group_2)
        [4, 5, 6]

    If the length of *iterable* is not evenly divisible by *n*, then the
    length of the returned iterables will not be identical:

        >>> children = divide(3, [1, 2, 3, 4, 5, 6, 7])
        >>> [list(c) for c in children]
        [[1, 2, 3], [4, 5], [6, 7]]

    If the length of the iterable is smaller than n, then the last returned
    iterables will be empty:

        >>> children = divide(5, [1, 2, 3])
        >>> [list(c) for c in children]
        [[1], [2], [3], [], []]

    This function will exhaust the iterable before returning and may require
    significant storage. If order is not important, see :func:`distribute`,
    which does not first pull the iterable into memory.

    """
    if n < 1:
        raise ValueError('n must be at least 1')

    try:
        iterable[:0]
    except TypeError:
        seq = tuple(iterable)
    else:
        seq = iterable

    q, r = divmod(len(seq), n)

    ret = []
    stop = 0
    for i in range(1, n + 1):
        start = stop
        stop += q + 1 if i <= r else q
        ret.append(iter(seq[start:stop]))

    return ret


def always_iterable(obj, base_type=(str, bytes)):
    """If *obj* is iterable, return an iterator over its items::

        >>> obj = (1, 2, 3)
        >>> list(always_iterable(obj))
        [1, 2, 3]

    If *obj* is not iterable, return a one-item iterable containing *obj*::

        >>> obj = 1
        >>> list(always_iterable(obj))
        [1]

    If *obj* is ``None``, return an empty iterable:

        >>> obj = None
        >>> list(always_iterable(None))
        []

    By default, binary and text strings are not considered iterable::

        >>> obj = 'foo'
        >>> list(always_iterable(obj))
        ['foo']

    If *base_type* is set, objects for which ``isinstance(obj, base_type)``
    returns ``True`` won't be considered iterable.

        >>> obj = {'a': 1}
        >>> list(always_iterable(obj))  # Iterate over the dict's keys
        ['a']
        >>> list(always_iterable(obj, base_type=dict))  # Treat dicts as a unit
        [{'a': 1}]

    Set *base_type* to ``None`` to avoid any special handling and treat objects
    Python considers iterable as iterable:

        >>> obj = 'foo'
        >>> list(always_iterable(obj, base_type=None))
        ['f', 'o', 'o']
    """
    if obj is None:
        return iter(())

    if (base_type is not None) and isinstance(obj, base_type):
        return iter((obj,))

    try:
        return iter(obj)
    except TypeError:
        return iter((obj,))


def adjacent(predicate, iterable, distance=1):
    """Return an iterable over `(bool, item)` tuples where the `item` is
    drawn from *iterable* and the `bool` indicates whether
    that item satisfies the *predicate* or is adjacent to an item that does.

    For example, to find whether items are adjacent to a ``3``::

        >>> list(adjacent(lambda x: x == 3, range(6)))
        [(False, 0), (False, 1), (True, 2), (True, 3), (True, 4), (False, 5)]

    Set *distance* to change what counts as adjacent. For example, to find
    whether items are two places away from a ``3``:

        >>> list(adjacent(lambda x: x == 3, range(6), distance=2))
        [(False, 0), (True, 1), (True, 2), (True, 3), (True, 4), (True, 5)]

    This is useful for contextualizing the results of a search function.
    For example, a code comparison tool might want to identify lines that
    have changed, but also surrounding lines to give the viewer of the diff
    context.

    The predicate function will only be called once for each item in the
    iterable.

    See also :func:`groupby_transform`, which can be used with this function
    to group ranges of items with the same `bool` value.

    """
    # Allow distance=0 mainly for testing that it reproduces results with map()
    if distance < 0:
        raise ValueError('distance must be at least 0')

    i1, i2 = tee(iterable)
    padding = [False] * distance
    selected = chain(padding, map(predicate, i1), padding)
    adjacent_to_selected = map(any, windowed(selected, 2 * distance + 1))
    return zip(adjacent_to_selected, i2)


def groupby_transform(iterable, keyfunc=None, valuefunc=None, reducefunc=None):
    """An extension of :func:`itertools.groupby` that can apply transformations
    to the grouped data.

    * *keyfunc* is a function computing a key value for each item in *iterable*
    * *valuefunc* is a function that transforms the individual items from
      *iterable* after grouping
    * *reducefunc* is a function that transforms each group of items

    >>> iterable = 'aAAbBBcCC'
    >>> keyfunc = lambda k: k.upper()
    >>> valuefunc = lambda v: v.lower()
    >>> reducefunc = lambda g: ''.join(g)
    >>> list(groupby_transform(iterable, keyfunc, valuefunc, reducefunc))
    [('A', 'aaa'), ('B', 'bbb'), ('C', 'ccc')]

    Each optional argument defaults to an identity function if not specified.

    :func:`groupby_transform` is useful when grouping elements of an iterable
    using a separate iterable as the key. To do this, :func:`zip` the iterables
    and pass a *keyfunc* that extracts the first element and a *valuefunc*
    that extracts the second element::

        >>> from operator import itemgetter
        >>> keys = [0, 0, 1, 1, 1, 2, 2, 2, 3]
        >>> values = 'abcdefghi'
        >>> iterable = zip(keys, values)
        >>> grouper = groupby_transform(iterable, itemgetter(0), itemgetter(1))
        >>> [(k, ''.join(g)) for k, g in grouper]
        [(0, 'ab'), (1, 'cde'), (2, 'fgh'), (3, 'i')]

    Note that the order of items in the iterable is significant.
    Only adjacent items are grouped together, so if you don't want any
    duplicate groups, you should sort the iterable by the key function.

    """
    ret = groupby(iterable, keyfunc)
    if valuefunc:
        ret = ((k, map(valuefunc, g)) for k, g in ret)
    if reducefunc:
        ret = ((k, reducefunc(g)) for k, g in ret)

    return ret


class numeric_range(abc.Sequence, abc.Hashable):
    """An extension of the built-in ``range()`` function whose arguments can
    be any orderable numeric type.

    With only *stop* specified, *start* defaults to ``0`` and *step*
    defaults to ``1``. The output items will match the type of *stop*:

        >>> list(numeric_range(3.5))
        [0.0, 1.0, 2.0, 3.0]

    With only *start* and *stop* specified, *step* defaults to ``1``. The
    output items will match the type of *start*:

        >>> from decimal import Decimal
        >>> start = Decimal('2.1')
        >>> stop = Decimal('5.1')
        >>> list(numeric_range(start, stop))
        [Decimal('2.1'), Decimal('3.1'), Decimal('4.1')]

    With *start*, *stop*, and *step*  specified the output items will match
    the type of ``start + step``:

        >>> from fractions import Fraction
        >>> start = Fraction(1, 2)  # Start at 1/2
        >>> stop = Fraction(5, 2)  # End at 5/2
        >>> step = Fraction(1, 2)  # Count by 1/2
        >>> list(numeric_range(start, stop, step))
        [Fraction(1, 2), Fraction(1, 1), Fraction(3, 2), Fraction(2, 1)]

    If *step* is zero, ``ValueError`` is raised. Negative steps are supported:

        >>> list(numeric_range(3, -1, -1.0))
        [3.0, 2.0, 1.0, 0.0]

    Be aware of the limitations of floating point numbers; the representation
    of the yielded numbers may be surprising.

    ``datetime.datetime`` objects can be used for *start* and *stop*, if *step*
    is a ``datetime.timedelta`` object:

        >>> import datetime
        >>> start = datetime.datetime(2019, 1, 1)
        >>> stop = datetime.datetime(2019, 1, 3)
        >>> step = datetime.timedelta(days=1)
        >>> items = iter(numeric_range(start, stop, step))
        >>> next(items)
        datetime.datetime(2019, 1, 1, 0, 0)
        >>> next(items)
        datetime.datetime(2019, 1, 2, 0, 0)

    """

    _EMPTY_HASH = hash(range(0, 0))

    def __init__(self, *args):
        argc = len(args)
        if argc == 1:
            (self._stop,) = args
            self._start = type(self._stop)(0)
            self._step = type(self._stop - self._start)(1)
        elif argc == 2:
            self._start, self._stop = args
            self._step = type(self._stop - self._start)(1)
        elif argc == 3:
            self._start, self._stop, self._step = args
        elif argc == 0:
            raise TypeError(
                'numeric_range expected at least '
                '1 argument, got {}'.format(argc)
            )
        else:
            raise TypeError(
                'numeric_range expected at most '
                '3 arguments, got {}'.format(argc)
            )

        self._zero = type(self._step)(0)
        if self._step == self._zero:
            raise ValueError('numeric_range() arg 3 must not be zero')
        self._growing = self._step > self._zero
        self._init_len()

    def __bool__(self):
        if self._growing:
            return self._start < self._stop
        else:
            return self._start > self._stop

    def __contains__(self, elem):
        if self._growing:
            if self._start <= elem < self._stop:
                return (elem - self._start) % self._step == self._zero
        else:
            if self._start >= elem > self._stop:
                return (self._start - elem) % (-self._step) == self._zero

        return False

    def __eq__(self, other):
        if isinstance(other, numeric_range):
            empty_self = not bool(self)
            empty_other = not bool(other)
            if empty_self or empty_other:
                return empty_self and empty_other  # True if both empty
            else:
                return (
                    self._start == other._start
                    and self._step == other._step
                    and self._get_by_index(-1) == other._get_by_index(-1)
                )
        else:
            return False

    def __getitem__(self, key):
        if isinstance(key, int):
            return self._get_by_index(key)
        elif isinstance(key, slice):
            step = self._step if key.step is None else key.step * self._step

            if key.start is None or key.start <= -self._len:
                start = self._start
            elif key.start >= self._len:
                start = self._stop
            else:  # -self._len < key.start < self._len
                start = self._get_by_index(key.start)

            if key.stop is None or key.stop >= self._len:
                stop = self._stop
            elif key.stop <= -self._len:
                stop = self._start
            else:  # -self._len < key.stop < self._len
                stop = self._get_by_index(key.stop)

            return numeric_range(start, stop, step)
        else:
            raise TypeError(
                'numeric range indices must be '
                'integers or slices, not {}'.format(type(key).__name__)
            )

    def __hash__(self):
        if self:
            return hash((self._start, self._get_by_index(-1), self._step))
        else:
            return self._EMPTY_HASH

    def __iter__(self):
        values = (self._start + (n * self._step) for n in count())
        if self._growing:
            return takewhile(partial(gt, self._stop), values)
        else:
            return takewhile(partial(lt, self._stop), values)

    def __len__(self):
        return self._len

    def _init_len(self):
        if self._growing:
            start = self._start
            stop = self._stop
            step = self._step
        else:
            start = self._stop
            stop = self._start
            step = -self._step
        distance = stop - start
        if distance <= self._zero:
            self._len = 0
        else:  # distance > 0 and step > 0: regular euclidean division
            q, r = divmod(distance, step)
            self._len = int(q) + int(r != self._zero)

    def __reduce__(self):
        return numeric_range, (self._start, self._stop, self._step)

    def __repr__(self):
        if self._step == 1:
            return "numeric_range({}, {})".format(
                repr(self._start), repr(self._stop)
            )
        else:
            return "numeric_range({}, {}, {})".format(
                repr(self._start), repr(self._stop), repr(self._step)
            )

    def __reversed__(self):
        return iter(
            numeric_range(
                self._get_by_index(-1), self._start - self._step, -self._step
            )
        )

    def count(self, value):
        return int(value in self)

    def index(self, value):
        if self._growing:
            if self._start <= value < self._stop:
                q, r = divmod(value - self._start, self._step)
                if r == self._zero:
                    return int(q)
        else:
            if self._start >= value > self._stop:
                q, r = divmod(self._start - value, -self._step)
                if r == self._zero:
                    return int(q)

        raise ValueError("{} is not in numeric range".format(value))

    def _get_by_index(self, i):
        if i < 0:
            i += self._len
        if i < 0 or i >= self._len:
            raise IndexError("numeric range object index out of range")
        return self._start + i * self._step


def count_cycle(iterable, n=None):
    """Cycle through the items from *iterable* up to *n* times, yielding
    the number of completed cycles along with each item. If *n* is omitted the
    process repeats indefinitely.

    >>> list(count_cycle('AB', 3))
    [(0, 'A'), (0, 'B'), (1, 'A'), (1, 'B'), (2, 'A'), (2, 'B')]

    """
    iterable = tuple(iterable)
    if not iterable:
        return iter(())
    counter = count() if n is None else range(n)
    return ((i, item) for i in counter for item in iterable)


def mark_ends(iterable):
    """Yield 3-tuples of the form ``(is_first, is_last, item)``.

    >>> list(mark_ends('ABC'))
    [(True, False, 'A'), (False, False, 'B'), (False, True, 'C')]

    Use this when looping over an iterable to take special action on its first
    and/or last items:

    >>> iterable = ['Header', 100, 200, 'Footer']
    >>> total = 0
    >>> for is_first, is_last, item in mark_ends(iterable):
    ...     if is_first:
    ...         continue  # Skip the header
    ...     if is_last:
    ...         continue  # Skip the footer
    ...     total += item
    >>> print(total)
    300
    """
    it = iter(iterable)

    try:
        b = next(it)
    except StopIteration:
        return

    try:
        for i in count():
            a = b
            b = next(it)
            yield i == 0, False, a

    except StopIteration:
        yield i == 0, True, a


def locate(iterable, pred=bool, window_size=None):
    """Yield the index of each item in *iterable* for which *pred* returns
    ``True``.

    *pred* defaults to :func:`bool`, which will select truthy items:

        >>> list(locate([0, 1, 1, 0, 1, 0, 0]))
        [1, 2, 4]

    Set *pred* to a custom function to, e.g., find the indexes for a particular
    item.

        >>> list(locate(['a', 'b', 'c', 'b'], lambda x: x == 'b'))
        [1, 3]

    If *window_size* is given, then the *pred* function will be called with
    that many items. This enables searching for sub-sequences:

        >>> iterable = [0, 1, 2, 3, 0, 1, 2, 3, 0, 1, 2, 3]
        >>> pred = lambda *args: args == (1, 2, 3)
        >>> list(locate(iterable, pred=pred, window_size=3))
        [1, 5, 9]

    Use with :func:`seekable` to find indexes and then retrieve the associated
    items:

        >>> from itertools import count
        >>> from more_itertools import seekable
        >>> source = (3 * n + 1 if (n % 2) else n // 2 for n in count())
        >>> it = seekable(source)
        >>> pred = lambda x: x > 100
        >>> indexes = locate(it, pred=pred)
        >>> i = next(indexes)
        >>> it.seek(i)
        >>> next(it)
        106

    """
    if window_size is None:
        return compress(count(), map(pred, iterable))

    if window_size < 1:
        raise ValueError('window size must be at least 1')

    it = windowed(iterable, window_size, fillvalue=_marker)
    return compress(count(), starmap(pred, it))


def lstrip(iterable, pred):
    """Yield the items from *iterable*, but strip any from the beginning
    for which *pred* returns ``True``.

    For example, to remove a set of items from the start of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(lstrip(iterable, pred))
        [1, 2, None, 3, False, None]

    This function is analogous to to :func:`str.lstrip`, and is essentially
    an wrapper for :func:`itertools.dropwhile`.

    """
    return dropwhile(pred, iterable)


def rstrip(iterable, pred):
    """Yield the items from *iterable*, but strip any from the end
    for which *pred* returns ``True``.

    For example, to remove a set of items from the end of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(rstrip(iterable, pred))
        [None, False, None, 1, 2, None, 3]

    This function is analogous to :func:`str.rstrip`.

    """
    cache = []
    cache_append = cache.append
    cache_clear = cache.clear
    for x in iterable:
        if pred(x):
            cache_append(x)
        else:
            yield from cache
            cache_clear()
            yield x


def strip(iterable, pred):
    """Yield the items from *iterable*, but strip any from the
    beginning and end for which *pred* returns ``True``.

    For example, to remove a set of items from both ends of an iterable:

        >>> iterable = (None, False, None, 1, 2, None, 3, False, None)
        >>> pred = lambda x: x in {None, False, ''}
        >>> list(strip(iterable, pred))
        [1, 2, None, 3]

    This function is analogous to :func:`str.strip`.

    """
    return rstrip(lstrip(iterable, pred), pred)


class islice_extended:
    """An extension of :func:`itertools.islice` that supports negative values
    for *stop*, *start*, and *step*.

        >>> iterable = iter('abcdefgh')
        >>> list(islice_extended(iterable, -4, -1))
        ['e', 'f', 'g']

    Slices with negative values require some caching of *iterable*, but this
    function takes care to minimize the amount of memory required.

    For example, you can use a negative step with an infinite iterator:

        >>> from itertools import count
        >>> list(islice_extended(count(), 110, 99, -2))
        [110, 108, 106, 104, 102, 100]

    You can also use slice notation directly:

        >>> iterable = map(str, count())
        >>> it = islice_extended(iterable)[10:20:2]
        >>> list(it)
        ['10', '12', '14', '16', '18']

    """

    def __init__(self, iterable, *args):
        it = iter(iterable)
        if args:
            self._iterable = _islice_helper(it, slice(*args))
        else:
            self._iterable = it

    def __iter__(self):
        return self

    def __next__(self):
        return next(self._iterable)

    def __getitem__(self, key):
        if isinstance(key, slice):
            return islice_extended(_islice_helper(self._iterable, key))

        raise TypeError('islice_extended.__getitem__ argument must be a slice')


def _islice_helper(it, s):
    start = s.start
    stop = s.stop
    if s.step == 0:
        raise ValueError('step argument must be a non-zero integer or None.')
    step = s.step or 1

    if step > 0:
        start = 0 if (start is None) else start

        if start < 0:
            # Consume all but the last -start items
            cache = deque(enumerate(it, 1), maxlen=-start)
            len_iter = cache[-1][0] if cache else 0

            # Adjust start to be positive
            i = max(len_iter + start, 0)

            # Adjust stop to be positive
            if stop is None:
                j = len_iter
            elif stop >= 0:
                j = min(stop, len_iter)
            else:
                j = max(len_iter + stop, 0)

            # Slice the cache
            n = j - i
            if n <= 0:
                return

            for index, item in islice(cache, 0, n, step):
                yield item
        elif (stop is not None) and (stop < 0):
            # Advance to the start position
            next(islice(it, start, start), None)

            # When stop is negative, we have to carry -stop items while
            # iterating
            cache = deque(islice(it, -stop), maxlen=-stop)

            for index, item in enumerate(it):
                cached_item = cache.popleft()
                if index % step == 0:
                    yield cached_item
                cache.append(item)
        else:
            # When both start and stop are positive we have the normal case
            yield from islice(it, start, stop, step)
    else:
        start = -1 if (start is None) else start

        if (stop is not None) and (stop < 0):
            # Consume all but the last items
            n = -stop - 1
            cache = deque(enumerate(it, 1), maxlen=n)
            len_iter = cache[-1][0] if cache else 0

            # If start and stop are both negative they are comparable and
            # we can just slice. Otherwise we can adjust start to be negative
            # and then slice.
            if start < 0:
                i, j = start, stop
            else:
                i, j = min(start - len_iter, -1), None

            for index, item in list(cache)[i:j:step]:
                yield item
        else:
            # Advance to the stop position
            if stop is not None:
                m = stop + 1
                next(islice(it, m, m), None)

            # stop is positive, so if start is negative they are not comparable
            # and we need the rest of the items.
            if start < 0:
                i = start
                n = None
            # stop is None and start is positive, so we just need items up to
            # the start index.
            elif stop is None:
                i = None
                n = start + 1
            # Both stop and start are positive, so they are comparable.
            else:
                i = None
                n = start - stop
                if n <= 0:
                    return

            cache = list(islice(it, n))

            yield from cache[i::step]


def always_reversible(iterable):
    """An extension of :func:`reversed` that supports all iterables, not
    just those which implement the ``Reversible`` or ``Sequence`` protocols.

        >>> print(*always_reversible(x for x in range(3)))
        2 1 0

    If the iterable is already reversible, this function returns the
    result of :func:`reversed()`. If the iterable is not reversible,
    this function will cache the remaining items in the iterable and
    yield them in reverse order, which may require significant storage.
    """
    try:
        return reversed(iterable)
    except TypeError:
        return reversed(list(iterable))


def consecutive_groups(iterable, ordering=lambda x: x):
    """Yield groups of consecutive items using :func:`itertools.groupby`.
    The *ordering* function determines whether two items are adjacent by
    returning their position.

    By default, the ordering function is the identity function. This is
    suitable for finding runs of numbers:

        >>> iterable = [1, 10, 11, 12, 20, 30, 31, 32, 33, 40]
        >>> for group in consecutive_groups(iterable):
        ...     print(list(group))
        [1]
        [10, 11, 12]
        [20]
        [30, 31, 32, 33]
        [40]

    For finding runs of adjacent letters, try using the :meth:`index` method
    of a string of letters:

        >>> from string import ascii_lowercase
        >>> iterable = 'abcdfgilmnop'
        >>> ordering = ascii_lowercase.index
        >>> for group in consecutive_groups(iterable, ordering):
        ...     print(list(group))
        ['a', 'b', 'c', 'd']
        ['f', 'g']
        ['i']
        ['l', 'm', 'n', 'o', 'p']

    Each group of consecutive items is an iterator that shares it source with
    *iterable*. When an an output group is advanced, the previous group is
    no longer available unless its elements are copied (e.g., into a ``list``).

        >>> iterable = [1, 2, 11, 12, 21, 22]
        >>> saved_groups = []
        >>> for group in consecutive_groups(iterable):
        ...     saved_groups.append(list(group))  # Copy group elements
        >>> saved_groups
        [[1, 2], [11, 12], [21, 22]]

    """
    for k, g in groupby(
        enumerate(iterable), key=lambda x: x[0] - ordering(x[1])
    ):
        yield map(itemgetter(1), g)


def difference(iterable, func=sub, *, initial=None):
    """This function is the inverse of :func:`itertools.accumulate`. By default
    it will compute the first difference of *iterable* using
    :func:`operator.sub`:

        >>> from itertools import accumulate
        >>> iterable = accumulate([0, 1, 2, 3, 4])  # produces 0, 1, 3, 6, 10
        >>> list(difference(iterable))
        [0, 1, 2, 3, 4]

    *func* defaults to :func:`operator.sub`, but other functions can be
    specified. They will be applied as follows::

        A, B, C, D, ... --> A, func(B, A), func(C, B), func(D, C), ...

    For example, to do progressive division:

        >>> iterable = [1, 2, 6, 24, 120]
        >>> func = lambda x, y: x // y
        >>> list(difference(iterable, func))
        [1, 2, 3, 4, 5]

    If the *initial* keyword is set, the first element will be skipped when
    computing successive differences.

        >>> it = [10, 11, 13, 16]  # from accumulate([1, 2, 3], initial=10)
        >>> list(difference(it, initial=10))
        [1, 2, 3]

    """
    a, b = tee(iterable)
    try:
        first = [next(b)]
    except StopIteration:
        return iter([])

    if initial is not None:
        first = []

    return chain(first, starmap(func, zip(b, a)))


class SequenceView(Sequence):
    """Return a read-only view of the sequence object *target*.

    :class:`SequenceView` objects are analogous to Python's built-in
    "dictionary view" types. They provide a dynamic view of a sequence's items,
    meaning that when the sequence updates, so does the view.

        >>> seq = ['0', '1', '2']
        >>> view = SequenceView(seq)
        >>> view
        SequenceView(['0', '1', '2'])
        >>> seq.append('3')
        >>> view
        SequenceView(['0', '1', '2', '3'])

    Sequence views support indexing, slicing, and length queries. They act
    like the underlying sequence, except they don't allow assignment:

        >>> view[1]
        '1'
        >>> view[1:-1]
        ['1', '2']
        >>> len(view)
        4

    Sequence views are useful as an alternative to copying, as they don't
    require (much) extra storage.

    """

    def __init__(self, target):
        if not isinstance(target, Sequence):
            raise TypeError
        self._target = target

    def __getitem__(self, index):
        return self._target[index]

    def __len__(self):
        return len(self._target)

    def __repr__(self):
        return '{}({})'.format(self.__class__.__name__, repr(self._target))


class seekable:
    """Wrap an iterator to allow for seeking backward and forward. This
    progressively caches the items in the source iterable so they can be
    re-visited.

    Call :meth:`seek` with an index to seek to that position in the source
    iterable.

    To "reset" an iterator, seek to ``0``:

        >>> from itertools import count
        >>> it = seekable((str(n) for n in count()))
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> it.seek(0)
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> next(it)
        '3'

    You can also seek forward:

        >>> it = seekable((str(n) for n in range(20)))
        >>> it.seek(10)
        >>> next(it)
        '10'
        >>> it.seek(20)  # Seeking past the end of the source isn't a problem
        >>> list(it)
        []
        >>> it.seek(0)  # Resetting works even after hitting the end
        >>> next(it), next(it), next(it)
        ('0', '1', '2')

    Call :meth:`peek` to look ahead one item without advancing the iterator:

        >>> it = seekable('1234')
        >>> it.peek()
        '1'
        >>> list(it)
        ['1', '2', '3', '4']
        >>> it.peek(default='empty')
        'empty'

    Before the iterator is at its end, calling :func:`bool` on it will return
    ``True``. After it will return ``False``:

        >>> it = seekable('5678')
        >>> bool(it)
        True
        >>> list(it)
        ['5', '6', '7', '8']
        >>> bool(it)
        False

    You may view the contents of the cache with the :meth:`elements` method.
    That returns a :class:`SequenceView`, a view that updates automatically:

        >>> it = seekable((str(n) for n in range(10)))
        >>> next(it), next(it), next(it)
        ('0', '1', '2')
        >>> elements = it.elements()
        >>> elements
        SequenceView(['0', '1', '2'])
        >>> next(it)
        '3'
        >>> elements
        SequenceView(['0', '1', '2', '3'])

    By default, the cache grows as the source iterable progresses, so beware of
    wrapping very large or infinite iterables. Supply *maxlen* to limit the
    size of the cache (this of course limits how far back you can seek).

        >>> from itertools import count
        >>> it = seekable((str(n) for n in count()), maxlen=2)
        >>> next(it), next(it), next(it), next(it)
        ('0', '1', '2', '3')
        >>> list(it.elements())
        ['2', '3']
        >>> it.seek(0)
        >>> next(it), next(it), next(it), next(it)
        ('2', '3', '4', '5')
        >>> next(it)
        '6'

    """

    def __init__(self, iterable, maxlen=None):
        self._source = iter(iterable)
        if maxlen is None:
            self._cache = []
        else:
            self._cache = deque([], maxlen)
        self._index = None

    def __iter__(self):
        return self

    def __next__(self):
        if self._index is not None:
            try:
                item = self._cache[self._index]
            except IndexError:
                self._index = None
            else:
                self._index += 1
                return item

        item = next(self._source)
        self._cache.append(item)
        return item

    def __bool__(self):
        try:
            self.peek()
        except StopIteration:
            return False
        return True

    def peek(self, default=_marker):
        try:
            peeked = next(self)
        except StopIteration:
            if default is _marker:
                raise
            return default
        if self._index is None:
            self._index = len(self._cache)
        self._index -= 1
        return peeked

    def elements(self):
        return SequenceView(self._cache)

    def seek(self, index):
        self._index = index
        remainder = index - len(self._cache)
        if remainder > 0:
            consume(self, remainder)


class run_length:
    """
    :func:`run_length.encode` compresses an iterable with run-length encoding.
    It yields groups of repeated items with the count of how many times they
    were repeated:

        >>> uncompressed = 'abbcccdddd'
        >>> list(run_length.encode(uncompressed))
        [('a', 1), ('b', 2), ('c', 3), ('d', 4)]

    :func:`run_length.decode` decompresses an iterable that was previously
    compressed with run-length encoding. It yields the items of the
    decompressed iterable:

        >>> compressed = [('a', 1), ('b', 2), ('c', 3), ('d', 4)]
        >>> list(run_length.decode(compressed))
        ['a', 'b', 'b', 'c', 'c', 'c', 'd', 'd', 'd', 'd']

    """

    @staticmethod
    def encode(iterable):
        return ((k, ilen(g)) for k, g in groupby(iterable))

    @staticmethod
    def decode(iterable):
        return chain.from_iterable(repeat(k, n) for k, n in iterable)


def exactly_n(iterable, n, predicate=bool):
    """Return ``True`` if exactly ``n`` items in the iterable are ``True``
    according to the *predicate* function.

        >>> exactly_n([True, True, False], 2)
        True
        >>> exactly_n([True, True, False], 1)
        False
        >>> exactly_n([0, 1, 2, 3, 4, 5], 3, lambda x: x < 3)
        True

    The iterable will be advanced until ``n + 1`` truthy items are encountered,
    so avoid calling it on infinite iterables.

    """
    return len(take(n + 1, filter(predicate, iterable))) == n


def circular_shifts(iterable):
    """Return a list of circular shifts of *iterable*.

    >>> circular_shifts(range(4))
    [(0, 1, 2, 3), (1, 2, 3, 0), (2, 3, 0, 1), (3, 0, 1, 2)]
    """
    lst = list(iterable)
    return take(len(lst), windowed(cycle(lst), len(lst)))


def make_decorator(wrapping_func, result_index=0):
    """Return a decorator version of *wrapping_func*, which is a function that
    modifies an iterable. *result_index* is the position in that function's
    signature where the iterable goes.

    This lets you use itertools on the "production end," i.e. at function
    definition. This can augment what the function returns without changing the
    function's code.

    For example, to produce a decorator version of :func:`chunked`:

        >>> from more_itertools import chunked
        >>> chunker = make_decorator(chunked, result_index=0)
        >>> @chunker(3)
        ... def iter_range(n):
        ...     return iter(range(n))
        ...
        >>> list(iter_range(9))
        [[0, 1, 2], [3, 4, 5], [6, 7, 8]]

    To only allow truthy items to be returned:

        >>> truth_serum = make_decorator(filter, result_index=1)
        >>> @truth_serum(bool)
        ... def boolean_test():
        ...     return [0, 1, '', ' ', False, True]
        ...
        >>> list(boolean_test())
        [1, ' ', True]

    The :func:`peekable` and :func:`seekable` wrappers make for practical
    decorators:

        >>> from more_itertools import peekable
        >>> peekable_function = make_decorator(peekable)
        >>> @peekable_function()
        ... def str_range(*args):
        ...     return (str(x) for x in range(*args))
        ...
        >>> it = str_range(1, 20, 2)
        >>> next(it), next(it), next(it)
        ('1', '3', '5')
        >>> it.peek()
        '7'
        >>> next(it)
        '7'

    """
    # See https://sites.google.com/site/bbayles/index/decorator_factory for
    # notes on how this works.
    def decorator(*wrapping_args, **wrapping_kwargs):
        def outer_wrapper(f):
            def inner_wrapper(*args, **kwargs):
                result = f(*args, **kwargs)
                wrapping_args_ = list(wrapping_args)
                wrapping_args_.insert(result_index, result)
                return wrapping_func(*wrapping_args_, **wrapping_kwargs)

            return inner_wrapper

        return outer_wrapper

    return decorator


def map_reduce(iterable, keyfunc, valuefunc=None, reducefunc=None):
    """Return a dictionary that maps the items in *iterable* to categories
    defined by *keyfunc*, transforms them with *valuefunc*, and
    then summarizes them by category with *reducefunc*.

    *valuefunc* defaults to the identity function if it is unspecified.
    If *reducefunc* is unspecified, no summarization takes place:

        >>> keyfunc = lambda x: x.upper()
        >>> result = map_reduce('abbccc', keyfunc)
        >>> sorted(result.items())
        [('A', ['a']), ('B', ['b', 'b']), ('C', ['c', 'c', 'c'])]

    Specifying *valuefunc* transforms the categorized items:

        >>> keyfunc = lambda x: x.upper()
        >>> valuefunc = lambda x: 1
        >>> result = map_reduce('abbccc', keyfunc, valuefunc)
        >>> sorted(result.items())
        [('A', [1]), ('B', [1, 1]), ('C', [1, 1, 1])]

    Specifying *reducefunc* summarizes the categorized items:

        >>> keyfunc = lambda x: x.upper()
        >>> valuefunc = lambda x: 1
        >>> reducefunc = sum
        >>> result = map_reduce('abbccc', keyfunc, valuefunc, reducefunc)
        >>> sorted(result.items())
        [('A', 1), ('B', 2), ('C', 3)]

    You may want to filter the input iterable before applying the map/reduce
    procedure:

        >>> all_items = range(30)
        >>> items = [x for x in all_items if 10 <= x <= 20]  # Filter
        >>> keyfunc = lambda x: x % 2  # Evens map to 0; odds to 1
        >>> categories = map_reduce(items, keyfunc=keyfunc)
        >>> sorted(categories.items())
        [(0, [10, 12, 14, 16, 18, 20]), (1, [11, 13, 15, 17, 19])]
        >>> summaries = map_reduce(items, keyfunc=keyfunc, reducefunc=sum)
        >>> sorted(summaries.items())
        [(0, 90), (1, 75)]

    Note that all items in the iterable are gathered into a list before the
    summarization step, which may require significant storage.

    The returned object is a :obj:`collections.defaultdict` with the
    ``default_factory`` set to ``None``, such that it behaves like a normal
    dictionary.

    """
    valuefunc = (lambda x: x) if (valuefunc is None) else valuefunc

    ret = defaultdict(list)
    for item in iterable:
        key = keyfunc(item)
        value = valuefunc(item)
        ret[key].append(value)

    if reducefunc is not None:
        for key, value_list in ret.items():
            ret[key] = reducefunc(value_list)

    ret.default_factory = None
    return ret


def rlocate(iterable, pred=bool, window_size=None):
    """Yield the index of each item in *iterable* for which *pred* returns
    ``True``, starting from the right and moving left.

    *pred* defaults to :func:`bool`, which will select truthy items:

        >>> list(rlocate([0, 1, 1, 0, 1, 0, 0]))  # Truthy at 1, 2, and 4
        [4, 2, 1]

    Set *pred* to a custom function to, e.g., find the indexes for a particular
    item:

        >>> iterable = iter('abcb')
        >>> pred = lambda x: x == 'b'
        >>> list(rlocate(iterable, pred))
        [3, 1]

    If *window_size* is given, then the *pred* function will be called with
    that many items. This enables searching for sub-sequences:

        >>> iterable = [0, 1, 2, 3, 0, 1, 2, 3, 0, 1, 2, 3]
        >>> pred = lambda *args: args == (1, 2, 3)
        >>> list(rlocate(iterable, pred=pred, window_size=3))
        [9, 5, 1]

    Beware, this function won't return anything for infinite iterables.
    If *iterable* is reversible, ``rlocate`` will reverse it and search from
    the right. Otherwise, it will search from the left and return the results
    in reverse order.

    See :func:`locate` to for other example applications.

    """
    if window_size is None:
        try:
            len_iter = len(iterable)
            return (len_iter - i - 1 for i in locate(reversed(iterable), pred))
        except TypeError:
            pass

    return reversed(list(locate(iterable, pred, window_size)))


def replace(iterable, pred, substitutes, count=None, window_size=1):
    """Yield the items from *iterable*, replacing the items for which *pred*
    returns ``True`` with the items from the iterable *substitutes*.

        >>> iterable = [1, 1, 0, 1, 1, 0, 1, 1]
        >>> pred = lambda x: x == 0
        >>> substitutes = (2, 3)
        >>> list(replace(iterable, pred, substitutes))
        [1, 1, 2, 3, 1, 1, 2, 3, 1, 1]

    If *count* is given, the number of replacements will be limited:

        >>> iterable = [1, 1, 0, 1, 1, 0, 1, 1, 0]
        >>> pred = lambda x: x == 0
        >>> substitutes = [None]
        >>> list(replace(iterable, pred, substitutes, count=2))
        [1, 1, None, 1, 1, None, 1, 1, 0]

    Use *window_size* to control the number of items passed as arguments to
    *pred*. This allows for locating and replacing subsequences.

        >>> iterable = [0, 1, 2, 5, 0, 1, 2, 5]
        >>> window_size = 3
        >>> pred = lambda *args: args == (0, 1, 2)  # 3 items passed to pred
        >>> substitutes = [3, 4] # Splice in these items
        >>> list(replace(iterable, pred, substitutes, window_size=window_size))
        [3, 4, 5, 3, 4, 5]

    """
    if window_size < 1:
        raise ValueError('window_size must be at least 1')

    # Save the substitutes iterable, since it's used more than once
    substitutes = tuple(substitutes)

    # Add padding such that the number of windows matches the length of the
    # iterable
    it = chain(iterable, [_marker] * (window_size - 1))
    windows = windowed(it, window_size)

    n = 0
    for w in windows:
        # If the current window matches our predicate (and we haven't hit
        # our maximum number of replacements), splice in the substitutes
        # and then consume the following windows that overlap with this one.
        # For example, if the iterable is (0, 1, 2, 3, 4...)
        # and the window size is 2, we have (0, 1), (1, 2), (2, 3)...
        # If the predicate matches on (0, 1), we need to zap (0, 1) and (1, 2)
        if pred(*w):
            if (count is None) or (n < count):
                n += 1
                yield from substitutes
                consume(windows, window_size - 1)
                continue

        # If there was no match (or we've reached the replacement limit),
        # yield the first item from the window.
        if w and (w[0] is not _marker):
            yield w[0]


def partitions(iterable):
    """Yield all possible order-preserving partitions of *iterable*.

    >>> iterable = 'abc'
    >>> for part in partitions(iterable):
    ...     print([''.join(p) for p in part])
    ['abc']
    ['a', 'bc']
    ['ab', 'c']
    ['a', 'b', 'c']

    This is unrelated to :func:`partition`.

    """
    sequence = list(iterable)
    n = len(sequence)
    for i in powerset(range(1, n)):
        yield [sequence[i:j] for i, j in zip((0,) + i, i + (n,))]


def set_partitions(iterable, k=None):
    """
    Yield the set partitions of *iterable* into *k* parts. Set partitions are
    not order-preserving.

    >>> iterable = 'abc'
    >>> for part in set_partitions(iterable, 2):
    ...     print([''.join(p) for p in part])
    ['a', 'bc']
    ['ab', 'c']
    ['b', 'ac']


    If *k* is not given, every set partition is generated.

    >>> iterable = 'abc'
    >>> for part in set_partitions(iterable):
    ...     print([''.join(p) for p in part])
    ['abc']
    ['a', 'bc']
    ['ab', 'c']
    ['b', 'ac']
    ['a', 'b', 'c']

    """
    L = list(iterable)
    n = len(L)
    if k is not None:
        if k < 1:
            raise ValueError(
                "Can't partition in a negative or zero number of groups"
            )
        elif k > n:
            return

    def set_partitions_helper(L, k):
        n = len(L)
        if k == 1:
            yield [L]
        elif n == k:
            yield [[s] for s in L]
        else:
            e, *M = L
            for p in set_partitions_helper(M, k - 1):
                yield [[e], *p]
            for p in set_partitions_helper(M, k):
                for i in range(len(p)):
                    yield p[:i] + [[e] + p[i]] + p[i + 1 :]

    if k is None:
        for k in range(1, n + 1):
            yield from set_partitions_helper(L, k)
    else:
        yield from set_partitions_helper(L, k)


class time_limited:
    """
    Yield items from *iterable* until *limit_seconds* have passed.
    If the time limit expires before all items have been yielded, the
    ``timed_out`` parameter will be set to ``True``.

    >>> from time import sleep
    >>> def generator():
    ...     yield 1
    ...     yield 2
    ...     sleep(0.2)
    ...     yield 3
    >>> iterable = time_limited(0.1, generator())
    >>> list(iterable)
    [1, 2]
    >>> iterable.timed_out
    True

    Note that the time is checked before each item is yielded, and iteration
    stops if  the time elapsed is greater than *limit_seconds*. If your time
    limit is 1 second, but it takes 2 seconds to generate the first item from
    the iterable, the function will run for 2 seconds and not yield anything.

    """

    def __init__(self, limit_seconds, iterable):
        if limit_seconds < 0:
            raise ValueError('limit_seconds must be positive')
        self.limit_seconds = limit_seconds
        self._iterable = iter(iterable)
        self._start_time = monotonic()
        self.timed_out = False

    def __iter__(self):
        return self

    def __next__(self):
        item = next(self._iterable)
        if monotonic() - self._start_time > self.limit_seconds:
            self.timed_out = True
            raise StopIteration

        return item


def only(iterable, default=None, too_long=None):
    """If *iterable* has only one item, return it.
    If it has zero items, return *default*.
    If it has more than one item, raise the exception given by *too_long*,
    which is ``ValueError`` by default.

    >>> only([], default='missing')
    'missing'
    >>> only([1])
    1
    >>> only([1, 2])  # doctest: +IGNORE_EXCEPTION_DETAIL
    Traceback (most recent call last):
    ...
    ValueError: Expected exactly one item in iterable, but got 1, 2,
     and perhaps more.'
    >>> only([1, 2], too_long=TypeError)  # doctest: +IGNORE_EXCEPTION_DETAIL
    Traceback (most recent call last):
    ...
    TypeError

    Note that :func:`only` attempts to advance *iterable* twice to ensure there
    is only one item.  See :func:`spy` or :func:`peekable` to check
    iterable contents less destructively.
    """
    it = iter(iterable)
    first_value = next(it, default)

    try:
        second_value = next(it)
    except StopIteration:
        pass
    else:
        msg = (
            'Expected exactly one item in iterable, but got {!r}, {!r}, '
            'and perhaps more.'.format(first_value, second_value)
        )
        raise too_long or ValueError(msg)

    return first_value


def ichunked(iterable, n):
    """Break *iterable* into sub-iterables with *n* elements each.
    :func:`ichunked` is like :func:`chunked`, but it yields iterables
    instead of lists.

    If the sub-iterables are read in order, the elements of *iterable*
    won't be stored in memory.
    If they are read out of order, :func:`itertools.tee` is used to cache
    elements as necessary.

    >>> from itertools import count
    >>> all_chunks = ichunked(count(), 4)
    >>> c_1, c_2, c_3 = next(all_chunks), next(all_chunks), next(all_chunks)
    >>> list(c_2)  # c_1's elements have been cached; c_3's haven't been
    [4, 5, 6, 7]
    >>> list(c_1)
    [0, 1, 2, 3]
    >>> list(c_3)
    [8, 9, 10, 11]

    """
    source = iter(iterable)

    while True:
        # Check to see whether we're at the end of the source iterable
        item = next(source, _marker)
        if item is _marker:
            return

        # Clone the source and yield an n-length slice
        source, it = tee(chain([item], source))
        yield islice(it, n)

        # Advance the source iterable
        consume(source, n)


def distinct_combinations(iterable, r):
    """Yield the distinct combinations of *r* items taken from *iterable*.

        >>> list(distinct_combinations([0, 0, 1], 2))
        [(0, 0), (0, 1)]

    Equivalent to ``set(combinations(iterable))``, except duplicates are not
    generated and thrown away. For larger input sequences this is much more
    efficient.

    """
    if r < 0:
        raise ValueError('r must be non-negative')
    elif r == 0:
        yield ()
        return
    pool = tuple(iterable)
    generators = [unique_everseen(enumerate(pool), key=itemgetter(1))]
    current_combo = [None] * r
    level = 0
    while generators:
        try:
            cur_idx, p = next(generators[-1])
        except StopIteration:
            generators.pop()
            level -= 1
            continue
        current_combo[level] = p
        if level + 1 == r:
            yield tuple(current_combo)
        else:
            generators.append(
                unique_everseen(
                    enumerate(pool[cur_idx + 1 :], cur_idx + 1),
                    key=itemgetter(1),
                )
            )
            level += 1


def filter_except(validator, iterable, *exceptions):
    """Yield the items from *iterable* for which the *validator* function does
    not raise one of the specified *exceptions*.

    *validator* is called for each item in *iterable*.
    It should be a function that accepts one argument and raises an exception
    if that item is not valid.

    >>> iterable = ['1', '2', 'three', '4', None]
    >>> list(filter_except(int, iterable, ValueError, TypeError))
    ['1', '2', '4']

    If an exception other than one given by *exceptions* is raised by
    *validator*, it is raised like normal.
    """
    for item in iterable:
        try:
            validator(item)
        except exceptions:
            pass
        else:
            yield item


def map_except(function, iterable, *exceptions):
    """Transform each item from *iterable* with *function* and yield the
    result, unless *function* raises one of the specified *exceptions*.

    *function* is called to transform each item in *iterable*.
    It should be a accept one argument.

    >>> iterable = ['1', '2', 'three', '4', None]
    >>> list(map_except(int, iterable, ValueError, TypeError))
    [1, 2, 4]

    If an exception other than one given by *exceptions* is raised by
    *function*, it is raised like normal.
    """
    for item in iterable:
        try:
            yield function(item)
        except exceptions:
            pass


def _sample_unweighted(iterable, k):
    # Implementation of "Algorithm L" from the 1994 paper by Kim-Hung Li:
    # "Reservoir-Sampling Algorithms of Time Complexity O(n(1+log(N/n)))".

    # Fill up the reservoir (collection of samples) with the first `k` samples
    reservoir = take(k, iterable)

    # Generate random number that's the largest in a sample of k U(0,1) numbers
    # Largest order statistic: https://en.wikipedia.org/wiki/Order_statistic
    W = exp(log(random()) / k)

    # The number of elements to skip before changing the reservoir is a random
    # number with a geometric distribution. Sample it using random() and logs.
    next_index = k + floor(log(random()) / log(1 - W))

    for index, element in enumerate(iterable, k):

        if index == next_index:
            reservoir[randrange(k)] = element
            # The new W is the largest in a sample of k U(0, `old_W`) numbers
            W *= exp(log(random()) / k)
            next_index += floor(log(random()) / log(1 - W)) + 1

    return reservoir


def _sample_weighted(iterable, k, weights):
    # Implementation of "A-ExpJ" from the 2006 paper by Efraimidis et al. :
    # "Weighted random sampling with a reservoir".

    # Log-transform for numerical stability for weights that are small/large
    weight_keys = (log(random()) / weight for weight in weights)

    # Fill up the reservoir (collection of samples) with the first `k`
    # weight-keys and elements, then heapify the list.
    reservoir = take(k, zip(weight_keys, iterable))
    heapify(reservoir)

    # The number of jumps before changing the reservoir is a random variable
    # with an exponential distribution. Sample it using random() and logs.
    smallest_weight_key, _ = reservoir[0]
    weights_to_skip = log(random()) / smallest_weight_key

    for weight, element in zip(weights, iterable):
        if weight >= weights_to_skip:
            # The notation here is consistent with the paper, but we store
            # the weight-keys in log-space for better numerical stability.
            smallest_weight_key, _ = reservoir[0]
            t_w = exp(weight * smallest_weight_key)
            r_2 = uniform(t_w, 1)  # generate U(t_w, 1)
            weight_key = log(r_2) / weight
            heapreplace(reservoir, (weight_key, element))
            smallest_weight_key, _ = reservoir[0]
            weights_to_skip = log(random()) / smallest_weight_key
        else:
            weights_to_skip -= weight

    # Equivalent to [element for weight_key, element in sorted(reservoir)]
    return [heappop(reservoir)[1] for _ in range(k)]


def sample(iterable, k, weights=None):
    """Return a *k*-length list of elements chosen (without replacement)
    from the *iterable*. Like :func:`random.sample`, but works on iterables
    of unknown length.

    >>> iterable = range(100)
    >>> sample(iterable, 5)  # doctest: +SKIP
    [81, 60, 96, 16, 4]

    An iterable with *weights* may also be given:

    >>> iterable = range(100)
    >>> weights = (i * i + 1 for i in range(100))
    >>> sampled = sample(iterable, 5, weights=weights)  # doctest: +SKIP
    [79, 67, 74, 66, 78]

    The algorithm can also be used to generate weighted random permutations.
    The relative weight of each item determines the probability that it
    appears late in the permutation.

    >>> data = "abcdefgh"
    >>> weights = range(1, len(data) + 1)
    >>> sample(data, k=len(data), weights=weights)  # doctest: +SKIP
    ['c', 'a', 'b', 'e', 'g', 'd', 'h', 'f']
    """
    if k == 0:
        return []

    iterable = iter(iterable)
    if weights is None:
        return _sample_unweighted(iterable, k)
    else:
        weights = iter(weights)
        return _sample_weighted(iterable, k, weights)


def is_sorted(iterable, key=None, reverse=False):
    """Returns ``True`` if the items of iterable are in sorted order, and
    ``False`` otherwise. *key* and *reverse* have the same meaning that they do
    in the built-in :func:`sorted` function.

    >>> is_sorted(['1', '2', '3', '4', '5'], key=int)
    True
    >>> is_sorted([5, 4, 3, 1, 2], reverse=True)
    False

    The function returns ``False`` after encountering the first out-of-order
    item. If there are no out-of-order items, the iterable is exhausted.
    """

    compare = lt if reverse else gt
    it = iterable if (key is None) else map(key, iterable)
    return not any(starmap(compare, pairwise(it)))


class AbortThread(BaseException):
    pass


class callback_iter:
    """Convert a function that uses callbacks to an iterator.

    Let *func* be a function that takes a `callback` keyword argument.
    For example:

    >>> def func(callback=None):
    ...     for i, c in [(1, 'a'), (2, 'b'), (3, 'c')]:
    ...         if callback:
    ...             callback(i, c)
    ...     return 4


    Use ``with callback_iter(func)`` to get an iterator over the parameters
    that are delivered to the callback.

    >>> with callback_iter(func) as it:
    ...     for args, kwargs in it:
    ...         print(args)
    (1, 'a')
    (2, 'b')
    (3, 'c')

    The function will be called in a background thread. The ``done`` property
    indicates whether it has completed execution.

    >>> it.done
    True

    If it completes successfully, its return value will be available
    in the ``result`` property.

    >>> it.result
    4

    Notes:

    * If the function uses some keyword argument besides ``callback``, supply
      *callback_kwd*.
    * If it finished executing, but raised an exception, accessing the
      ``result`` property will raise the same exception.
    * If it hasn't finished executing, accessing the ``result``
      property from within the ``with`` block will raise ``RuntimeError``.
    * If it hasn't finished executing, accessing the ``result`` property from
      outside the ``with`` block will raise a
      ``more_itertools.AbortThread`` exception.
    * Provide *wait_seconds* to adjust how frequently the it is polled for
      output.

    """

    def __init__(self, func, callback_kwd='callback', wait_seconds=0.1):
        self._func = func
        self._callback_kwd = callback_kwd
        self._aborted = False
        self._future = None
        self._wait_seconds = wait_seconds
        self._executor = ThreadPoolExecutor(max_workers=1)
        self._iterator = self._reader()

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_value, traceback):
        self._aborted = True
        self._executor.shutdown()

    def __iter__(self):
        return self

    def __next__(self):
        return next(self._iterator)

    @property
    def done(self):
        if self._future is None:
            return False
        return self._future.done()

    @property
    def result(self):
        if not self.done:
            raise RuntimeError('Function has not yet completed')

        return self._future.result()

    def _reader(self):
        q = Queue()

        def callback(*args, **kwargs):
            if self._aborted:
                raise AbortThread('canceled by user')

            q.put((args, kwargs))

        self._future = self._executor.submit(
            self._func, **{self._callback_kwd: callback}
        )

        while True:
            try:
                item = q.get(timeout=self._wait_seconds)
            except Empty:
                pass
            else:
                q.task_done()
                yield item

            if self._future.done():
                break

        remaining = []
        while True:
            try:
                item = q.get_nowait()
            except Empty:
                break
            else:
                q.task_done()
                remaining.append(item)
        q.join()
        yield from remaining


def windowed_complete(iterable, n):
    """
    Yield ``(beginning, middle, end)`` tuples, where:

    * Each ``middle`` has *n* items from *iterable*
    * Each ``beginning`` has the items before the ones in ``middle``
    * Each ``end`` has the items after the ones in ``middle``

    >>> iterable = range(7)
    >>> n = 3
    >>> for beginning, middle, end in windowed_complete(iterable, n):
    ...     print(beginning, middle, end)
    () (0, 1, 2) (3, 4, 5, 6)
    (0,) (1, 2, 3) (4, 5, 6)
    (0, 1) (2, 3, 4) (5, 6)
    (0, 1, 2) (3, 4, 5) (6,)
    (0, 1, 2, 3) (4, 5, 6) ()

    Note that *n* must be at least 0 and most equal to the length of
    *iterable*.

    This function will exhaust the iterable and may require significant
    storage.
    """
    if n < 0:
        raise ValueError('n must be >= 0')

    seq = tuple(iterable)
    size = len(seq)

    if n > size:
        raise ValueError('n must be <= len(seq)')

    for i in range(size - n + 1):
        beginning = seq[:i]
        middle = seq[i : i + n]
        end = seq[i + n :]
        yield beginning, middle, end


def all_unique(iterable, key=None):
    """
    Returns ``True`` if all the elements of *iterable* are unique (no two
    elements are equal).

        >>> all_unique('ABCB')
        False

    If a *key* function is specified, it will be used to make comparisons.

        >>> all_unique('ABCb')
        True
        >>> all_unique('ABCb', str.lower)
        False

    The function returns as soon as the first non-unique element is
    encountered. Iterables with a mix of hashable and unhashable items can
    be used, but the function will be slower for unhashable items.
    """
    seenset = set()
    seenset_add = seenset.add
    seenlist = []
    seenlist_add = seenlist.append
    for element in map(key, iterable) if key else iterable:
        try:
            if element in seenset:
                return False
            seenset_add(element)
        except TypeError:
            if element in seenlist:
                return False
            seenlist_add(element)
    return True


def nth_product(index, *args):
    """Equivalent to ``list(product(*args))[index]``.

    The products of *args* can be ordered lexicographically.
    :func:`nth_product` computes the product at sort position *index* without
    computing the previous products.

        >>> nth_product(8, range(2), range(2), range(2), range(2))
        (1, 0, 0, 0)

    ``IndexError`` will be raised if the given *index* is invalid.
    """
    pools = list(map(tuple, reversed(args)))
    ns = list(map(len, pools))

    c = reduce(mul, ns)

    if index < 0:
        index += c

    if not 0 <= index < c:
        raise IndexError

    result = []
    for pool, n in zip(pools, ns):
        result.append(pool[index % n])
        index //= n

    return tuple(reversed(result))


def nth_permutation(iterable, r, index):
    """Equivalent to ``list(permutations(iterable, r))[index]```

    The subsequences of *iterable* that are of length *r* where order is
    important can be ordered lexicographically. :func:`nth_permutation`
    computes the subsequence at sort position *index* directly, without
    computing the previous subsequences.

        >>> nth_permutation('ghijk', 2, 5)
        ('h', 'i')

    ``ValueError`` will be raised If *r* is negative or greater than the length
    of *iterable*.
    ``IndexError`` will be raised if the given *index* is invalid.
    """
    pool = list(iterable)
    n = len(pool)

    if r is None or r == n:
        r, c = n, factorial(n)
    elif not 0 <= r < n:
        raise ValueError
    else:
        c = factorial(n) // factorial(n - r)

    if index < 0:
        index += c

    if not 0 <= index < c:
        raise IndexError

    if c == 0:
        return tuple()

    result = [0] * r
    q = index * factorial(n) // c if r < n else index
    for d in range(1, n + 1):
        q, i = divmod(q, d)
        if 0 <= n - d < r:
            result[n - d] = i
        if q == 0:
            break

    return tuple(map(pool.pop, result))


def value_chain(*args):
    """Yield all arguments passed to the function in the same order in which
    they were passed. If an argument itself is iterable then iterate over its
    values.

        >>> list(value_chain(1, 2, 3, [4, 5, 6]))
        [1, 2, 3, 4, 5, 6]

    Binary and text strings are not considered iterable and are emitted
    as-is:

        >>> list(value_chain('12', '34', ['56', '78']))
        ['12', '34', '56', '78']


    Multiple levels of nesting are not flattened.

    """
    for value in args:
        if isinstance(value, (str, bytes)):
            yield value
            continue
        try:
            yield from value
        except TypeError:
            yield value


def product_index(element, *args):
    """Equivalent to ``list(product(*args)).index(element)``

    The products of *args* can be ordered lexicographically.
    :func:`product_index` computes the first index of *element* without
    computing the previous products.

        >>> product_index([8, 2], range(10), range(5))
        42

    ``ValueError`` will be raised if the given *element* isn't in the product
    of *args*.
    """
    index = 0

    for x, pool in zip_longest(element, args, fillvalue=_marker):
        if x is _marker or pool is _marker:
            raise ValueError('element is not a product of args')

        pool = tuple(pool)
        index = index * len(pool) + pool.index(x)

    return index


def combination_index(element, iterable):
    """Equivalent to ``list(combinations(iterable, r)).index(element)``

    The subsequences of *iterable* that are of length *r* can be ordered
    lexicographically. :func:`combination_index` computes the index of the
    first *element*, without computing the previous combinations.

        >>> combination_index('adf', 'abcdefg')
        10

    ``ValueError`` will be raised if the given *element* isn't one of the
    combinations of *iterable*.
    """
    element = enumerate(element)
    k, y = next(element, (None, None))
    if k is None:
        return 0

    indexes = []
    pool = enumerate(iterable)
    for n, x in pool:
        if x == y:
            indexes.append(n)
            tmp, y = next(element, (None, None))
            if tmp is None:
                break
            else:
                k = tmp
    else:
        raise ValueError('element is not a combination of iterable')

    n, _ = last(pool, default=(n, None))

    # Python versiosn below 3.8 don't have math.comb
    index = 1
    for i, j in enumerate(reversed(indexes), start=1):
        j = n - j
        if i <= j:
            index += factorial(j) // (factorial(i) * factorial(j - i))

    return factorial(n + 1) // (factorial(k + 1) * factorial(n - k)) - index


def permutation_index(element, iterable):
    """Equivalent to ``list(permutations(iterable, r)).index(element)```

    The subsequences of *iterable* that are of length *r* where order is
    important can be ordered lexicographically. :func:`permutation_index`
    computes the index of the first *element* directly, without computing
    the previous permutations.

        >>> permutation_index([1, 3, 2], range(5))
        19

    ``ValueError`` will be raised if the given *element* isn't one of the
    permutations of *iterable*.
    """
    index = 0
    pool = list(iterable)
    for i, x in zip(range(len(pool), -1, -1), element):
        r = pool.index(x)
        index = index * i + r
        del pool[r]

    return index


class countable:
    """Wrap *iterable* and keep a count of how many items have been consumed.

    The ``items_seen`` attribute starts at ``0`` and increments as the iterable
    is consumed:

        >>> iterable = map(str, range(10))
        >>> it = countable(iterable)
        >>> it.items_seen
        0
        >>> next(it), next(it)
        ('0', '1')
        >>> list(it)
        ['2', '3', '4', '5', '6', '7', '8', '9']
        >>> it.items_seen
        10
    """

    def __init__(self, iterable):
        self._it = iter(iterable)
        self.items_seen = 0

    def __iter__(self):
        return self

    def __next__(self):
        item = next(self._it)
        self.items_seen += 1

        return item
site-packages/setuptools/_vendor/more_itertools/recipes.py000064400000037600150564745010020230 0ustar00"""Imported from the recipes section of the itertools documentation.

All functions taken from the recipes section of the itertools library docs
[1]_.
Some backward-compatible usability improvements have been made.

.. [1] http://docs.python.org/library/itertools.html#recipes

"""
import warnings
from collections import deque
from itertools import (
    chain,
    combinations,
    count,
    cycle,
    groupby,
    islice,
    repeat,
    starmap,
    tee,
    zip_longest,
)
import operator
from random import randrange, sample, choice

__all__ = [
    'all_equal',
    'consume',
    'convolve',
    'dotproduct',
    'first_true',
    'flatten',
    'grouper',
    'iter_except',
    'ncycles',
    'nth',
    'nth_combination',
    'padnone',
    'pad_none',
    'pairwise',
    'partition',
    'powerset',
    'prepend',
    'quantify',
    'random_combination_with_replacement',
    'random_combination',
    'random_permutation',
    'random_product',
    'repeatfunc',
    'roundrobin',
    'tabulate',
    'tail',
    'take',
    'unique_everseen',
    'unique_justseen',
]


def take(n, iterable):
    """Return first *n* items of the iterable as a list.

        >>> take(3, range(10))
        [0, 1, 2]

    If there are fewer than *n* items in the iterable, all of them are
    returned.

        >>> take(10, range(3))
        [0, 1, 2]

    """
    return list(islice(iterable, n))


def tabulate(function, start=0):
    """Return an iterator over the results of ``func(start)``,
    ``func(start + 1)``, ``func(start + 2)``...

    *func* should be a function that accepts one integer argument.

    If *start* is not specified it defaults to 0. It will be incremented each
    time the iterator is advanced.

        >>> square = lambda x: x ** 2
        >>> iterator = tabulate(square, -3)
        >>> take(4, iterator)
        [9, 4, 1, 0]

    """
    return map(function, count(start))


def tail(n, iterable):
    """Return an iterator over the last *n* items of *iterable*.

    >>> t = tail(3, 'ABCDEFG')
    >>> list(t)
    ['E', 'F', 'G']

    """
    return iter(deque(iterable, maxlen=n))


def consume(iterator, n=None):
    """Advance *iterable* by *n* steps. If *n* is ``None``, consume it
    entirely.

    Efficiently exhausts an iterator without returning values. Defaults to
    consuming the whole iterator, but an optional second argument may be
    provided to limit consumption.

        >>> i = (x for x in range(10))
        >>> next(i)
        0
        >>> consume(i, 3)
        >>> next(i)
        4
        >>> consume(i)
        >>> next(i)
        Traceback (most recent call last):
          File "<stdin>", line 1, in <module>
        StopIteration

    If the iterator has fewer items remaining than the provided limit, the
    whole iterator will be consumed.

        >>> i = (x for x in range(3))
        >>> consume(i, 5)
        >>> next(i)
        Traceback (most recent call last):
          File "<stdin>", line 1, in <module>
        StopIteration

    """
    # Use functions that consume iterators at C speed.
    if n is None:
        # feed the entire iterator into a zero-length deque
        deque(iterator, maxlen=0)
    else:
        # advance to the empty slice starting at position n
        next(islice(iterator, n, n), None)


def nth(iterable, n, default=None):
    """Returns the nth item or a default value.

    >>> l = range(10)
    >>> nth(l, 3)
    3
    >>> nth(l, 20, "zebra")
    'zebra'

    """
    return next(islice(iterable, n, None), default)


def all_equal(iterable):
    """
    Returns ``True`` if all the elements are equal to each other.

        >>> all_equal('aaaa')
        True
        >>> all_equal('aaab')
        False

    """
    g = groupby(iterable)
    return next(g, True) and not next(g, False)


def quantify(iterable, pred=bool):
    """Return the how many times the predicate is true.

    >>> quantify([True, False, True])
    2

    """
    return sum(map(pred, iterable))


def pad_none(iterable):
    """Returns the sequence of elements and then returns ``None`` indefinitely.

        >>> take(5, pad_none(range(3)))
        [0, 1, 2, None, None]

    Useful for emulating the behavior of the built-in :func:`map` function.

    See also :func:`padded`.

    """
    return chain(iterable, repeat(None))


padnone = pad_none


def ncycles(iterable, n):
    """Returns the sequence elements *n* times

    >>> list(ncycles(["a", "b"], 3))
    ['a', 'b', 'a', 'b', 'a', 'b']

    """
    return chain.from_iterable(repeat(tuple(iterable), n))


def dotproduct(vec1, vec2):
    """Returns the dot product of the two iterables.

    >>> dotproduct([10, 10], [20, 20])
    400

    """
    return sum(map(operator.mul, vec1, vec2))


def flatten(listOfLists):
    """Return an iterator flattening one level of nesting in a list of lists.

        >>> list(flatten([[0, 1], [2, 3]]))
        [0, 1, 2, 3]

    See also :func:`collapse`, which can flatten multiple levels of nesting.

    """
    return chain.from_iterable(listOfLists)


def repeatfunc(func, times=None, *args):
    """Call *func* with *args* repeatedly, returning an iterable over the
    results.

    If *times* is specified, the iterable will terminate after that many
    repetitions:

        >>> from operator import add
        >>> times = 4
        >>> args = 3, 5
        >>> list(repeatfunc(add, times, *args))
        [8, 8, 8, 8]

    If *times* is ``None`` the iterable will not terminate:

        >>> from random import randrange
        >>> times = None
        >>> args = 1, 11
        >>> take(6, repeatfunc(randrange, times, *args))  # doctest:+SKIP
        [2, 4, 8, 1, 8, 4]

    """
    if times is None:
        return starmap(func, repeat(args))
    return starmap(func, repeat(args, times))


def _pairwise(iterable):
    """Returns an iterator of paired items, overlapping, from the original

    >>> take(4, pairwise(count()))
    [(0, 1), (1, 2), (2, 3), (3, 4)]

    On Python 3.10 and above, this is an alias for :func:`itertools.pairwise`.

    """
    a, b = tee(iterable)
    next(b, None)
    yield from zip(a, b)


try:
    from itertools import pairwise as itertools_pairwise
except ImportError:
    pairwise = _pairwise
else:

    def pairwise(iterable):
        yield from itertools_pairwise(iterable)

    pairwise.__doc__ = _pairwise.__doc__


def grouper(iterable, n, fillvalue=None):
    """Collect data into fixed-length chunks or blocks.

    >>> list(grouper('ABCDEFG', 3, 'x'))
    [('A', 'B', 'C'), ('D', 'E', 'F'), ('G', 'x', 'x')]

    """
    if isinstance(iterable, int):
        warnings.warn(
            "grouper expects iterable as first parameter", DeprecationWarning
        )
        n, iterable = iterable, n
    args = [iter(iterable)] * n
    return zip_longest(fillvalue=fillvalue, *args)


def roundrobin(*iterables):
    """Yields an item from each iterable, alternating between them.

        >>> list(roundrobin('ABC', 'D', 'EF'))
        ['A', 'D', 'E', 'B', 'F', 'C']

    This function produces the same output as :func:`interleave_longest`, but
    may perform better for some inputs (in particular when the number of
    iterables is small).

    """
    # Recipe credited to George Sakkis
    pending = len(iterables)
    nexts = cycle(iter(it).__next__ for it in iterables)
    while pending:
        try:
            for next in nexts:
                yield next()
        except StopIteration:
            pending -= 1
            nexts = cycle(islice(nexts, pending))


def partition(pred, iterable):
    """
    Returns a 2-tuple of iterables derived from the input iterable.
    The first yields the items that have ``pred(item) == False``.
    The second yields the items that have ``pred(item) == True``.

        >>> is_odd = lambda x: x % 2 != 0
        >>> iterable = range(10)
        >>> even_items, odd_items = partition(is_odd, iterable)
        >>> list(even_items), list(odd_items)
        ([0, 2, 4, 6, 8], [1, 3, 5, 7, 9])

    If *pred* is None, :func:`bool` is used.

        >>> iterable = [0, 1, False, True, '', ' ']
        >>> false_items, true_items = partition(None, iterable)
        >>> list(false_items), list(true_items)
        ([0, False, ''], [1, True, ' '])

    """
    if pred is None:
        pred = bool

    evaluations = ((pred(x), x) for x in iterable)
    t1, t2 = tee(evaluations)
    return (
        (x for (cond, x) in t1 if not cond),
        (x for (cond, x) in t2 if cond),
    )


def powerset(iterable):
    """Yields all possible subsets of the iterable.

        >>> list(powerset([1, 2, 3]))
        [(), (1,), (2,), (3,), (1, 2), (1, 3), (2, 3), (1, 2, 3)]

    :func:`powerset` will operate on iterables that aren't :class:`set`
    instances, so repeated elements in the input will produce repeated elements
    in the output. Use :func:`unique_everseen` on the input to avoid generating
    duplicates:

        >>> seq = [1, 1, 0]
        >>> list(powerset(seq))
        [(), (1,), (1,), (0,), (1, 1), (1, 0), (1, 0), (1, 1, 0)]
        >>> from more_itertools import unique_everseen
        >>> list(powerset(unique_everseen(seq)))
        [(), (1,), (0,), (1, 0)]

    """
    s = list(iterable)
    return chain.from_iterable(combinations(s, r) for r in range(len(s) + 1))


def unique_everseen(iterable, key=None):
    """
    Yield unique elements, preserving order.

        >>> list(unique_everseen('AAAABBBCCDAABBB'))
        ['A', 'B', 'C', 'D']
        >>> list(unique_everseen('ABBCcAD', str.lower))
        ['A', 'B', 'C', 'D']

    Sequences with a mix of hashable and unhashable items can be used.
    The function will be slower (i.e., `O(n^2)`) for unhashable items.

    Remember that ``list`` objects are unhashable - you can use the *key*
    parameter to transform the list to a tuple (which is hashable) to
    avoid a slowdown.

        >>> iterable = ([1, 2], [2, 3], [1, 2])
        >>> list(unique_everseen(iterable))  # Slow
        [[1, 2], [2, 3]]
        >>> list(unique_everseen(iterable, key=tuple))  # Faster
        [[1, 2], [2, 3]]

    Similary, you may want to convert unhashable ``set`` objects with
    ``key=frozenset``. For ``dict`` objects,
    ``key=lambda x: frozenset(x.items())`` can be used.

    """
    seenset = set()
    seenset_add = seenset.add
    seenlist = []
    seenlist_add = seenlist.append
    use_key = key is not None

    for element in iterable:
        k = key(element) if use_key else element
        try:
            if k not in seenset:
                seenset_add(k)
                yield element
        except TypeError:
            if k not in seenlist:
                seenlist_add(k)
                yield element


def unique_justseen(iterable, key=None):
    """Yields elements in order, ignoring serial duplicates

    >>> list(unique_justseen('AAAABBBCCDAABBB'))
    ['A', 'B', 'C', 'D', 'A', 'B']
    >>> list(unique_justseen('ABBCcAD', str.lower))
    ['A', 'B', 'C', 'A', 'D']

    """
    return map(next, map(operator.itemgetter(1), groupby(iterable, key)))


def iter_except(func, exception, first=None):
    """Yields results from a function repeatedly until an exception is raised.

    Converts a call-until-exception interface to an iterator interface.
    Like ``iter(func, sentinel)``, but uses an exception instead of a sentinel
    to end the loop.

        >>> l = [0, 1, 2]
        >>> list(iter_except(l.pop, IndexError))
        [2, 1, 0]

    """
    try:
        if first is not None:
            yield first()
        while 1:
            yield func()
    except exception:
        pass


def first_true(iterable, default=None, pred=None):
    """
    Returns the first true value in the iterable.

    If no true value is found, returns *default*

    If *pred* is not None, returns the first item for which
    ``pred(item) == True`` .

        >>> first_true(range(10))
        1
        >>> first_true(range(10), pred=lambda x: x > 5)
        6
        >>> first_true(range(10), default='missing', pred=lambda x: x > 9)
        'missing'

    """
    return next(filter(pred, iterable), default)


def random_product(*args, repeat=1):
    """Draw an item at random from each of the input iterables.

        >>> random_product('abc', range(4), 'XYZ')  # doctest:+SKIP
        ('c', 3, 'Z')

    If *repeat* is provided as a keyword argument, that many items will be
    drawn from each iterable.

        >>> random_product('abcd', range(4), repeat=2)  # doctest:+SKIP
        ('a', 2, 'd', 3)

    This equivalent to taking a random selection from
    ``itertools.product(*args, **kwarg)``.

    """
    pools = [tuple(pool) for pool in args] * repeat
    return tuple(choice(pool) for pool in pools)


def random_permutation(iterable, r=None):
    """Return a random *r* length permutation of the elements in *iterable*.

    If *r* is not specified or is ``None``, then *r* defaults to the length of
    *iterable*.

        >>> random_permutation(range(5))  # doctest:+SKIP
        (3, 4, 0, 1, 2)

    This equivalent to taking a random selection from
    ``itertools.permutations(iterable, r)``.

    """
    pool = tuple(iterable)
    r = len(pool) if r is None else r
    return tuple(sample(pool, r))


def random_combination(iterable, r):
    """Return a random *r* length subsequence of the elements in *iterable*.

        >>> random_combination(range(5), 3)  # doctest:+SKIP
        (2, 3, 4)

    This equivalent to taking a random selection from
    ``itertools.combinations(iterable, r)``.

    """
    pool = tuple(iterable)
    n = len(pool)
    indices = sorted(sample(range(n), r))
    return tuple(pool[i] for i in indices)


def random_combination_with_replacement(iterable, r):
    """Return a random *r* length subsequence of elements in *iterable*,
    allowing individual elements to be repeated.

        >>> random_combination_with_replacement(range(3), 5) # doctest:+SKIP
        (0, 0, 1, 2, 2)

    This equivalent to taking a random selection from
    ``itertools.combinations_with_replacement(iterable, r)``.

    """
    pool = tuple(iterable)
    n = len(pool)
    indices = sorted(randrange(n) for i in range(r))
    return tuple(pool[i] for i in indices)


def nth_combination(iterable, r, index):
    """Equivalent to ``list(combinations(iterable, r))[index]``.

    The subsequences of *iterable* that are of length *r* can be ordered
    lexicographically. :func:`nth_combination` computes the subsequence at
    sort position *index* directly, without computing the previous
    subsequences.

        >>> nth_combination(range(5), 3, 5)
        (0, 3, 4)

    ``ValueError`` will be raised If *r* is negative or greater than the length
    of *iterable*.
    ``IndexError`` will be raised if the given *index* is invalid.
    """
    pool = tuple(iterable)
    n = len(pool)
    if (r < 0) or (r > n):
        raise ValueError

    c = 1
    k = min(r, n - r)
    for i in range(1, k + 1):
        c = c * (n - k + i) // i

    if index < 0:
        index += c

    if (index < 0) or (index >= c):
        raise IndexError

    result = []
    while r:
        c, n, r = c * r // n, n - 1, r - 1
        while index >= c:
            index -= c
            c, n = c * (n - r) // n, n - 1
        result.append(pool[-1 - n])

    return tuple(result)


def prepend(value, iterator):
    """Yield *value*, followed by the elements in *iterator*.

        >>> value = '0'
        >>> iterator = ['1', '2', '3']
        >>> list(prepend(value, iterator))
        ['0', '1', '2', '3']

    To prepend multiple values, see :func:`itertools.chain`
    or :func:`value_chain`.

    """
    return chain([value], iterator)


def convolve(signal, kernel):
    """Convolve the iterable *signal* with the iterable *kernel*.

        >>> signal = (1, 2, 3, 4, 5)
        >>> kernel = [3, 2, 1]
        >>> list(convolve(signal, kernel))
        [3, 8, 14, 20, 26, 14, 5]

    Note: the input arguments are not interchangeable, as the *kernel*
    is immediately consumed and stored.

    """
    kernel = tuple(kernel)[::-1]
    n = len(kernel)
    window = deque([0], maxlen=n) * n
    for x in chain(signal, repeat(0, n - 1)):
        window.append(x)
        yield sum(map(operator.mul, kernel, window))
site-packages/setuptools/_vendor/ordered_set.py000064400000035432150564745010016030 0ustar00"""
An OrderedSet is a custom MutableSet that remembers its order, so that every
entry has an index that can be looked up.

Based on a recipe originally posted to ActiveState Recipes by Raymond Hettiger,
and released under the MIT license.
"""
import itertools as it
from collections import deque

try:
    # Python 3
    from collections.abc import MutableSet, Sequence
except ImportError:
    # Python 2.7
    from collections import MutableSet, Sequence

SLICE_ALL = slice(None)
__version__ = "3.1"


def is_iterable(obj):
    """
    Are we being asked to look up a list of things, instead of a single thing?
    We check for the `__iter__` attribute so that this can cover types that
    don't have to be known by this module, such as NumPy arrays.

    Strings, however, should be considered as atomic values to look up, not
    iterables. The same goes for tuples, since they are immutable and therefore
    valid entries.

    We don't need to check for the Python 2 `unicode` type, because it doesn't
    have an `__iter__` attribute anyway.
    """
    return (
        hasattr(obj, "__iter__")
        and not isinstance(obj, str)
        and not isinstance(obj, tuple)
    )


class OrderedSet(MutableSet, Sequence):
    """
    An OrderedSet is a custom MutableSet that remembers its order, so that
    every entry has an index that can be looked up.

    Example:
        >>> OrderedSet([1, 1, 2, 3, 2])
        OrderedSet([1, 2, 3])
    """

    def __init__(self, iterable=None):
        self.items = []
        self.map = {}
        if iterable is not None:
            self |= iterable

    def __len__(self):
        """
        Returns the number of unique elements in the ordered set

        Example:
            >>> len(OrderedSet([]))
            0
            >>> len(OrderedSet([1, 2]))
            2
        """
        return len(self.items)

    def __getitem__(self, index):
        """
        Get the item at a given index.

        If `index` is a slice, you will get back that slice of items, as a
        new OrderedSet.

        If `index` is a list or a similar iterable, you'll get a list of
        items corresponding to those indices. This is similar to NumPy's
        "fancy indexing". The result is not an OrderedSet because you may ask
        for duplicate indices, and the number of elements returned should be
        the number of elements asked for.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset[1]
            2
        """
        if isinstance(index, slice) and index == SLICE_ALL:
            return self.copy()
        elif is_iterable(index):
            return [self.items[i] for i in index]
        elif hasattr(index, "__index__") or isinstance(index, slice):
            result = self.items[index]
            if isinstance(result, list):
                return self.__class__(result)
            else:
                return result
        else:
            raise TypeError("Don't know how to index an OrderedSet by %r" % index)

    def copy(self):
        """
        Return a shallow copy of this object.

        Example:
            >>> this = OrderedSet([1, 2, 3])
            >>> other = this.copy()
            >>> this == other
            True
            >>> this is other
            False
        """
        return self.__class__(self)

    def __getstate__(self):
        if len(self) == 0:
            # The state can't be an empty list.
            # We need to return a truthy value, or else __setstate__ won't be run.
            #
            # This could have been done more gracefully by always putting the state
            # in a tuple, but this way is backwards- and forwards- compatible with
            # previous versions of OrderedSet.
            return (None,)
        else:
            return list(self)

    def __setstate__(self, state):
        if state == (None,):
            self.__init__([])
        else:
            self.__init__(state)

    def __contains__(self, key):
        """
        Test if the item is in this ordered set

        Example:
            >>> 1 in OrderedSet([1, 3, 2])
            True
            >>> 5 in OrderedSet([1, 3, 2])
            False
        """
        return key in self.map

    def add(self, key):
        """
        Add `key` as an item to this OrderedSet, then return its index.

        If `key` is already in the OrderedSet, return the index it already
        had.

        Example:
            >>> oset = OrderedSet()
            >>> oset.append(3)
            0
            >>> print(oset)
            OrderedSet([3])
        """
        if key not in self.map:
            self.map[key] = len(self.items)
            self.items.append(key)
        return self.map[key]

    append = add

    def update(self, sequence):
        """
        Update the set with the given iterable sequence, then return the index
        of the last element inserted.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.update([3, 1, 5, 1, 4])
            4
            >>> print(oset)
            OrderedSet([1, 2, 3, 5, 4])
        """
        item_index = None
        try:
            for item in sequence:
                item_index = self.add(item)
        except TypeError:
            raise ValueError(
                "Argument needs to be an iterable, got %s" % type(sequence)
            )
        return item_index

    def index(self, key):
        """
        Get the index of a given entry, raising an IndexError if it's not
        present.

        `key` can be an iterable of entries that is not a string, in which case
        this returns a list of indices.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.index(2)
            1
        """
        if is_iterable(key):
            return [self.index(subkey) for subkey in key]
        return self.map[key]

    # Provide some compatibility with pd.Index
    get_loc = index
    get_indexer = index

    def pop(self):
        """
        Remove and return the last element from the set.

        Raises KeyError if the set is empty.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.pop()
            3
        """
        if not self.items:
            raise KeyError("Set is empty")

        elem = self.items[-1]
        del self.items[-1]
        del self.map[elem]
        return elem

    def discard(self, key):
        """
        Remove an element.  Do not raise an exception if absent.

        The MutableSet mixin uses this to implement the .remove() method, which
        *does* raise an error when asked to remove a non-existent item.

        Example:
            >>> oset = OrderedSet([1, 2, 3])
            >>> oset.discard(2)
            >>> print(oset)
            OrderedSet([1, 3])
            >>> oset.discard(2)
            >>> print(oset)
            OrderedSet([1, 3])
        """
        if key in self:
            i = self.map[key]
            del self.items[i]
            del self.map[key]
            for k, v in self.map.items():
                if v >= i:
                    self.map[k] = v - 1

    def clear(self):
        """
        Remove all items from this OrderedSet.
        """
        del self.items[:]
        self.map.clear()

    def __iter__(self):
        """
        Example:
            >>> list(iter(OrderedSet([1, 2, 3])))
            [1, 2, 3]
        """
        return iter(self.items)

    def __reversed__(self):
        """
        Example:
            >>> list(reversed(OrderedSet([1, 2, 3])))
            [3, 2, 1]
        """
        return reversed(self.items)

    def __repr__(self):
        if not self:
            return "%s()" % (self.__class__.__name__,)
        return "%s(%r)" % (self.__class__.__name__, list(self))

    def __eq__(self, other):
        """
        Returns true if the containers have the same items. If `other` is a
        Sequence, then order is checked, otherwise it is ignored.

        Example:
            >>> oset = OrderedSet([1, 3, 2])
            >>> oset == [1, 3, 2]
            True
            >>> oset == [1, 2, 3]
            False
            >>> oset == [2, 3]
            False
            >>> oset == OrderedSet([3, 2, 1])
            False
        """
        # In Python 2 deque is not a Sequence, so treat it as one for
        # consistent behavior with Python 3.
        if isinstance(other, (Sequence, deque)):
            # Check that this OrderedSet contains the same elements, in the
            # same order, as the other object.
            return list(self) == list(other)
        try:
            other_as_set = set(other)
        except TypeError:
            # If `other` can't be converted into a set, it's not equal.
            return False
        else:
            return set(self) == other_as_set

    def union(self, *sets):
        """
        Combines all unique items.
        Each items order is defined by its first appearance.

        Example:
            >>> oset = OrderedSet.union(OrderedSet([3, 1, 4, 1, 5]), [1, 3], [2, 0])
            >>> print(oset)
            OrderedSet([3, 1, 4, 5, 2, 0])
            >>> oset.union([8, 9])
            OrderedSet([3, 1, 4, 5, 2, 0, 8, 9])
            >>> oset | {10}
            OrderedSet([3, 1, 4, 5, 2, 0, 10])
        """
        cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet
        containers = map(list, it.chain([self], sets))
        items = it.chain.from_iterable(containers)
        return cls(items)

    def __and__(self, other):
        # the parent implementation of this is backwards
        return self.intersection(other)

    def intersection(self, *sets):
        """
        Returns elements in common between all sets. Order is defined only
        by the first set.

        Example:
            >>> oset = OrderedSet.intersection(OrderedSet([0, 1, 2, 3]), [1, 2, 3])
            >>> print(oset)
            OrderedSet([1, 2, 3])
            >>> oset.intersection([2, 4, 5], [1, 2, 3, 4])
            OrderedSet([2])
            >>> oset.intersection()
            OrderedSet([1, 2, 3])
        """
        cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet
        if sets:
            common = set.intersection(*map(set, sets))
            items = (item for item in self if item in common)
        else:
            items = self
        return cls(items)

    def difference(self, *sets):
        """
        Returns all elements that are in this set but not the others.

        Example:
            >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]))
            OrderedSet([1, 3])
            >>> OrderedSet([1, 2, 3]).difference(OrderedSet([2]), OrderedSet([3]))
            OrderedSet([1])
            >>> OrderedSet([1, 2, 3]) - OrderedSet([2])
            OrderedSet([1, 3])
            >>> OrderedSet([1, 2, 3]).difference()
            OrderedSet([1, 2, 3])
        """
        cls = self.__class__
        if sets:
            other = set.union(*map(set, sets))
            items = (item for item in self if item not in other)
        else:
            items = self
        return cls(items)

    def issubset(self, other):
        """
        Report whether another set contains this set.

        Example:
            >>> OrderedSet([1, 2, 3]).issubset({1, 2})
            False
            >>> OrderedSet([1, 2, 3]).issubset({1, 2, 3, 4})
            True
            >>> OrderedSet([1, 2, 3]).issubset({1, 4, 3, 5})
            False
        """
        if len(self) > len(other):  # Fast check for obvious cases
            return False
        return all(item in other for item in self)

    def issuperset(self, other):
        """
        Report whether this set contains another set.

        Example:
            >>> OrderedSet([1, 2]).issuperset([1, 2, 3])
            False
            >>> OrderedSet([1, 2, 3, 4]).issuperset({1, 2, 3})
            True
            >>> OrderedSet([1, 4, 3, 5]).issuperset({1, 2, 3})
            False
        """
        if len(self) < len(other):  # Fast check for obvious cases
            return False
        return all(item in self for item in other)

    def symmetric_difference(self, other):
        """
        Return the symmetric difference of two OrderedSets as a new set.
        That is, the new set will contain all elements that are in exactly
        one of the sets.

        Their order will be preserved, with elements from `self` preceding
        elements from `other`.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.symmetric_difference(other)
            OrderedSet([4, 5, 9, 2])
        """
        cls = self.__class__ if isinstance(self, OrderedSet) else OrderedSet
        diff1 = cls(self).difference(other)
        diff2 = cls(other).difference(self)
        return diff1.union(diff2)

    def _update_items(self, items):
        """
        Replace the 'items' list of this OrderedSet with a new one, updating
        self.map accordingly.
        """
        self.items = items
        self.map = {item: idx for (idx, item) in enumerate(items)}

    def difference_update(self, *sets):
        """
        Update this OrderedSet to remove items from one or more other sets.

        Example:
            >>> this = OrderedSet([1, 2, 3])
            >>> this.difference_update(OrderedSet([2, 4]))
            >>> print(this)
            OrderedSet([1, 3])

            >>> this = OrderedSet([1, 2, 3, 4, 5])
            >>> this.difference_update(OrderedSet([2, 4]), OrderedSet([1, 4, 6]))
            >>> print(this)
            OrderedSet([3, 5])
        """
        items_to_remove = set()
        for other in sets:
            items_to_remove |= set(other)
        self._update_items([item for item in self.items if item not in items_to_remove])

    def intersection_update(self, other):
        """
        Update this OrderedSet to keep only items in another set, preserving
        their order in this set.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.intersection_update(other)
            >>> print(this)
            OrderedSet([1, 3, 7])
        """
        other = set(other)
        self._update_items([item for item in self.items if item in other])

    def symmetric_difference_update(self, other):
        """
        Update this OrderedSet to remove items from another set, then
        add items from the other set that were not present in this set.

        Example:
            >>> this = OrderedSet([1, 4, 3, 5, 7])
            >>> other = OrderedSet([9, 7, 1, 3, 2])
            >>> this.symmetric_difference_update(other)
            >>> print(this)
            OrderedSet([4, 5, 9, 2])
        """
        items_to_add = [item for item in other if item not in self]
        items_to_remove = set(other)
        self._update_items(
            [item for item in self.items if item not in items_to_remove] + items_to_add
        )
site-packages/setuptools/_vendor/pyparsing.py000064400000705170150564745010015550 0ustar00# module pyparsing.py
#
# Copyright (c) 2003-2018  Paul T. McGuire
#
# Permission is hereby granted, free of charge, to any person obtaining
# a copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish,
# distribute, sublicense, and/or sell copies of the Software, and to
# permit persons to whom the Software is furnished to do so, subject to
# the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
# IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
# CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
#

__doc__ = \
"""
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and executing simple grammars,
vs. the traditional lex/yacc approach, or the use of regular expressions.  With pyparsing, you
don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
provides a library of classes that you use to construct the grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form 
C{"<salutation>, <addressee>!"}), built up using L{Word}, L{Literal}, and L{And} elements 
(L{'+'<ParserElement.__add__>} operator gives L{And} expressions, strings are auto-converted to
L{Literal} expressions)::

    from pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the self-explanatory
class names, and the use of '+', '|' and '^' operators.

The L{ParseResults} object returned from L{ParserElement.parseString<ParserElement.parseString>} can be accessed as a nested list, a dictionary, or an
object with named attributes.

The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
 - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello  ,  World  !", etc.)
 - quoted strings
 - embedded comments


Getting Started -
-----------------
Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:
 - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes
 - construct character word-group expressions using the L{Word} class
 - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes
 - use L{'+'<And>}, L{'|'<MatchFirst>}, L{'^'<Or>}, and L{'&'<Each>} operators to combine simple expressions into more complex ones
 - associate names with your parsed results using L{ParserElement.setResultsName}
 - find some helpful expression short-cuts like L{delimitedList} and L{oneOf}
 - find more useful common expressions in the L{pyparsing_common} namespace class
"""

__version__ = "2.2.1"
__versionTime__ = "18 Sep 2018 00:49 UTC"
__author__ = "Paul McGuire <ptmcg@users.sourceforge.net>"

import string
from weakref import ref as wkref
import copy
import sys
import warnings
import re
import sre_constants
import collections
import pprint
import traceback
import types
from datetime import datetime

try:
    from _thread import RLock
except ImportError:
    from threading import RLock

try:
    # Python 3
    from collections.abc import Iterable
    from collections.abc import MutableMapping
except ImportError:
    # Python 2.7
    from collections import Iterable
    from collections import MutableMapping

try:
    from collections import OrderedDict as _OrderedDict
except ImportError:
    try:
        from ordereddict import OrderedDict as _OrderedDict
    except ImportError:
        _OrderedDict = None

#~ sys.stderr.write( "testing pyparsing module, version %s, %s\n" % (__version__,__versionTime__ ) )

__all__ = [
'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty',
'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal',
'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or',
'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException',
'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException',
'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter', 
'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore',
'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col',
'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString',
'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums',
'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno',
'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral',
'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables',
'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity', 
'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd',
'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute',
'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation','locatedExpr', 'withClass',
'CloseMatch', 'tokenMap', 'pyparsing_common',
]

system_version = tuple(sys.version_info)[:3]
PY_3 = system_version[0] == 3
if PY_3:
    _MAX_INT = sys.maxsize
    basestring = str
    unichr = chr
    _ustr = str

    # build list of single arg builtins, that can be used as parse actions
    singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max]

else:
    _MAX_INT = sys.maxint
    range = xrange

    def _ustr(obj):
        """Drop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
           str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
           then < returns the unicode object | encodes it with the default encoding | ... >.
        """
        if isinstance(obj,unicode):
            return obj

        try:
            # If this works, then _ustr(obj) has the same behaviour as str(obj), so
            # it won't break any existing code.
            return str(obj)

        except UnicodeEncodeError:
            # Else encode it
            ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace')
            xmlcharref = Regex(r'&#\d+;')
            xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:])
            return xmlcharref.transformString(ret)

    # build list of single arg builtins, tolerant of Python version, that can be used as parse actions
    singleArgBuiltins = []
    import __builtin__
    for fname in "sum len sorted reversed list tuple set any all min max".split():
        try:
            singleArgBuiltins.append(getattr(__builtin__,fname))
        except AttributeError:
            continue
            
_generatorType = type((y for y in range(1)))
 
def _xml_escape(data):
    """Escape &, <, >, ", ', etc. in a string of data."""

    # ampersand must be replaced first
    from_symbols = '&><"\''
    to_symbols = ('&'+s+';' for s in "amp gt lt quot apos".split())
    for from_,to_ in zip(from_symbols, to_symbols):
        data = data.replace(from_, to_)
    return data

class _Constants(object):
    pass

alphas     = string.ascii_uppercase + string.ascii_lowercase
nums       = "0123456789"
hexnums    = nums + "ABCDEFabcdef"
alphanums  = alphas + nums
_bslash    = chr(92)
printables = "".join(c for c in string.printable if c not in string.whitespace)

class ParseBaseException(Exception):
    """base exception class for all parsing runtime exceptions"""
    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__( self, pstr, loc=0, msg=None, elem=None ):
        self.loc = loc
        if msg is None:
            self.msg = pstr
            self.pstr = ""
        else:
            self.msg = msg
            self.pstr = pstr
        self.parserElement = elem
        self.args = (pstr, loc, msg)

    @classmethod
    def _from_exception(cls, pe):
        """
        internal factory method to simplify creating one type of ParseException 
        from another - avoids having __init__ signature conflicts among subclasses
        """
        return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement)

    def __getattr__( self, aname ):
        """supported attributes by name are:
            - lineno - returns the line number of the exception text
            - col - returns the column number of the exception text
            - line - returns the line containing the exception text
        """
        if( aname == "lineno" ):
            return lineno( self.loc, self.pstr )
        elif( aname in ("col", "column") ):
            return col( self.loc, self.pstr )
        elif( aname == "line" ):
            return line( self.loc, self.pstr )
        else:
            raise AttributeError(aname)

    def __str__( self ):
        return "%s (at char %d), (line:%d, col:%d)" % \
                ( self.msg, self.loc, self.lineno, self.column )
    def __repr__( self ):
        return _ustr(self)
    def markInputline( self, markerString = ">!<" ):
        """Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        """
        line_str = self.line
        line_column = self.column - 1
        if markerString:
            line_str = "".join((line_str[:line_column],
                                markerString, line_str[line_column:]))
        return line_str.strip()
    def __dir__(self):
        return "lineno col line".split() + dir(type(self))

class ParseException(ParseBaseException):
    """
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
     - lineno - returns the line number of the exception text
     - col - returns the column number of the exception text
     - line - returns the line containing the exception text
        
    Example::
        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))
            
    prints::
       Expected integer (at char 0), (line:1, col:1)
        column: 1
    """
    pass

class ParseFatalException(ParseBaseException):
    """user-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediately"""
    pass

class ParseSyntaxException(ParseFatalException):
    """just like L{ParseFatalException}, but thrown internally when an
       L{ErrorStop<And._ErrorStop>} ('-' operator) indicates that parsing is to stop 
       immediately because an unbacktrackable syntax error has been found"""
    pass

#~ class ReparseException(ParseBaseException):
    #~ """Experimental class - parse actions can raise this exception to cause
       #~ pyparsing to reparse the input string:
        #~ - with a modified input string, and/or
        #~ - with a modified start location
       #~ Set the values of the ReparseException in the constructor, and raise the
       #~ exception in a parse action to cause pyparsing to use the new string/location.
       #~ Setting the values as None causes no change to be made.
       #~ """
    #~ def __init_( self, newstring, restartLoc ):
        #~ self.newParseText = newstring
        #~ self.reparseLoc = restartLoc

class RecursiveGrammarException(Exception):
    """exception thrown by L{ParserElement.validate} if the grammar could be improperly recursive"""
    def __init__( self, parseElementList ):
        self.parseElementTrace = parseElementList

    def __str__( self ):
        return "RecursiveGrammarException: %s" % self.parseElementTrace

class _ParseResultsWithOffset(object):
    def __init__(self,p1,p2):
        self.tup = (p1,p2)
    def __getitem__(self,i):
        return self.tup[i]
    def __repr__(self):
        return repr(self.tup[0])
    def setOffset(self,i):
        self.tup = (self.tup[0],i)

class ParseResults(object):
    """
    Structured parse results, to provide multiple means of access to the parsed data:
       - as a list (C{len(results)})
       - by list index (C{results[0], results[1]}, etc.)
       - by attribute (C{results.<resultsName>} - see L{ParserElement.setResultsName})

    Example::
        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)
    prints::
        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    """
    def __new__(cls, toklist=None, name=None, asList=True, modal=True ):
        if isinstance(toklist, cls):
            return toklist
        retobj = object.__new__(cls)
        retobj.__doinit = True
        return retobj

    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__( self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance ):
        if self.__doinit:
            self.__doinit = False
            self.__name = None
            self.__parent = None
            self.__accumNames = {}
            self.__asList = asList
            self.__modal = modal
            if toklist is None:
                toklist = []
            if isinstance(toklist, list):
                self.__toklist = toklist[:]
            elif isinstance(toklist, _generatorType):
                self.__toklist = list(toklist)
            else:
                self.__toklist = [toklist]
            self.__tokdict = dict()

        if name is not None and name:
            if not modal:
                self.__accumNames[name] = 0
            if isinstance(name,int):
                name = _ustr(name) # will always return a str, but use _ustr for consistency
            self.__name = name
            if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None,'',[])):
                if isinstance(toklist,basestring):
                    toklist = [ toklist ]
                if asList:
                    if isinstance(toklist,ParseResults):
                        self[name] = _ParseResultsWithOffset(toklist.copy(),0)
                    else:
                        self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]),0)
                    self[name].__name = name
                else:
                    try:
                        self[name] = toklist[0]
                    except (KeyError,TypeError,IndexError):
                        self[name] = toklist

    def __getitem__( self, i ):
        if isinstance( i, (int,slice) ):
            return self.__toklist[i]
        else:
            if i not in self.__accumNames:
                return self.__tokdict[i][-1][0]
            else:
                return ParseResults([ v[0] for v in self.__tokdict[i] ])

    def __setitem__( self, k, v, isinstance=isinstance ):
        if isinstance(v,_ParseResultsWithOffset):
            self.__tokdict[k] = self.__tokdict.get(k,list()) + [v]
            sub = v[0]
        elif isinstance(k,(int,slice)):
            self.__toklist[k] = v
            sub = v
        else:
            self.__tokdict[k] = self.__tokdict.get(k,list()) + [_ParseResultsWithOffset(v,0)]
            sub = v
        if isinstance(sub,ParseResults):
            sub.__parent = wkref(self)

    def __delitem__( self, i ):
        if isinstance(i,(int,slice)):
            mylen = len( self.__toklist )
            del self.__toklist[i]

            # convert int to slice
            if isinstance(i, int):
                if i < 0:
                    i += mylen
                i = slice(i, i+1)
            # get removed indices
            removed = list(range(*i.indices(mylen)))
            removed.reverse()
            # fixup indices in token dictionary
            for name,occurrences in self.__tokdict.items():
                for j in removed:
                    for k, (value, position) in enumerate(occurrences):
                        occurrences[k] = _ParseResultsWithOffset(value, position - (position > j))
        else:
            del self.__tokdict[i]

    def __contains__( self, k ):
        return k in self.__tokdict

    def __len__( self ): return len( self.__toklist )
    def __bool__(self): return ( not not self.__toklist )
    __nonzero__ = __bool__
    def __iter__( self ): return iter( self.__toklist )
    def __reversed__( self ): return iter( self.__toklist[::-1] )
    def _iterkeys( self ):
        if hasattr(self.__tokdict, "iterkeys"):
            return self.__tokdict.iterkeys()
        else:
            return iter(self.__tokdict)

    def _itervalues( self ):
        return (self[k] for k in self._iterkeys())
            
    def _iteritems( self ):
        return ((k, self[k]) for k in self._iterkeys())

    if PY_3:
        keys = _iterkeys       
        """Returns an iterator of all named result keys (Python 3.x only)."""

        values = _itervalues
        """Returns an iterator of all named result values (Python 3.x only)."""

        items = _iteritems
        """Returns an iterator of all named result key-value tuples (Python 3.x only)."""

    else:
        iterkeys = _iterkeys
        """Returns an iterator of all named result keys (Python 2.x only)."""

        itervalues = _itervalues
        """Returns an iterator of all named result values (Python 2.x only)."""

        iteritems = _iteritems
        """Returns an iterator of all named result key-value tuples (Python 2.x only)."""

        def keys( self ):
            """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iterkeys())

        def values( self ):
            """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.itervalues())
                
        def items( self ):
            """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iteritems())

    def haskeys( self ):
        """Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names."""
        return bool(self.__tokdict)
        
    def pop( self, *args, **kwargs):
        """
        Removes and returns item at specified index (default=C{last}).
        Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
        argument or an integer argument, it will use C{list} semantics
        and pop tokens from the list of parsed tokens. If passed a 
        non-integer argument (most likely a string), it will use C{dict}
        semantics and pop the corresponding value from any defined 
        results names. A second default return value argument is 
        supported, just as in C{dict.pop()}.

        Example::
            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())
        prints::
            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        """
        if not args:
            args = [-1]
        for k,v in kwargs.items():
            if k == 'default':
                args = (args[0], v)
            else:
                raise TypeError("pop() got an unexpected keyword argument '%s'" % k)
        if (isinstance(args[0], int) or 
                        len(args) == 1 or 
                        args[0] in self):
            index = args[0]
            ret = self[index]
            del self[index]
            return ret
        else:
            defaultvalue = args[1]
            return defaultvalue

    def get(self, key, defaultValue=None):
        """
        Returns named result matching the given key, or if there is no
        such name, then returns the given C{defaultValue} or C{None} if no
        C{defaultValue} is specified.

        Similar to C{dict.get()}.
        
        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        """
        if key in self:
            return self[key]
        else:
            return defaultValue

    def insert( self, index, insStr ):
        """
        Inserts new element at location index in the list of parsed tokens.
        
        Similar to C{list.insert()}.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        """
        self.__toklist.insert(index, insStr)
        # fixup indices in token dictionary
        for name,occurrences in self.__tokdict.items():
            for k, (value, position) in enumerate(occurrences):
                occurrences[k] = _ParseResultsWithOffset(value, position + (position > index))

    def append( self, item ):
        """
        Add single element to end of ParseResults list of elements.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            
            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        """
        self.__toklist.append(item)

    def extend( self, itemseq ):
        """
        Add sequence of elements to end of ParseResults list of elements.

        Example::
            patt = OneOrMore(Word(alphas))
            
            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        """
        if isinstance(itemseq, ParseResults):
            self += itemseq
        else:
            self.__toklist.extend(itemseq)

    def clear( self ):
        """
        Clear all elements and results names.
        """
        del self.__toklist[:]
        self.__tokdict.clear()

    def __getattr__( self, name ):
        try:
            return self[name]
        except KeyError:
            return ""
            
        if name in self.__tokdict:
            if name not in self.__accumNames:
                return self.__tokdict[name][-1][0]
            else:
                return ParseResults([ v[0] for v in self.__tokdict[name] ])
        else:
            return ""

    def __add__( self, other ):
        ret = self.copy()
        ret += other
        return ret

    def __iadd__( self, other ):
        if other.__tokdict:
            offset = len(self.__toklist)
            addoffset = lambda a: offset if a<0 else a+offset
            otheritems = other.__tokdict.items()
            otherdictitems = [(k, _ParseResultsWithOffset(v[0],addoffset(v[1])) )
                                for (k,vlist) in otheritems for v in vlist]
            for k,v in otherdictitems:
                self[k] = v
                if isinstance(v[0],ParseResults):
                    v[0].__parent = wkref(self)
            
        self.__toklist += other.__toklist
        self.__accumNames.update( other.__accumNames )
        return self

    def __radd__(self, other):
        if isinstance(other,int) and other == 0:
            # useful for merging many ParseResults using sum() builtin
            return self.copy()
        else:
            # this may raise a TypeError - so be it
            return other + self
        
    def __repr__( self ):
        return "(%s, %s)" % ( repr( self.__toklist ), repr( self.__tokdict ) )

    def __str__( self ):
        return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']'

    def _asStringList( self, sep='' ):
        out = []
        for item in self.__toklist:
            if out and sep:
                out.append(sep)
            if isinstance( item, ParseResults ):
                out += item._asStringList()
            else:
                out.append( _ustr(item) )
        return out

    def asList( self ):
        """
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::
            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']
            
            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        """
        return [res.asList() if isinstance(res,ParseResults) else res for res in self.__toklist]

    def asDict( self ):
        """
        Returns the named parse results as a nested dictionary.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
            
            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        """
        if PY_3:
            item_fn = self.items
        else:
            item_fn = self.iteritems
            
        def toItem(obj):
            if isinstance(obj, ParseResults):
                if obj.haskeys():
                    return obj.asDict()
                else:
                    return [toItem(v) for v in obj]
            else:
                return obj
                
        return dict((k,toItem(v)) for k,v in item_fn())

    def copy( self ):
        """
        Returns a new copy of a C{ParseResults} object.
        """
        ret = ParseResults( self.__toklist )
        ret.__tokdict = self.__tokdict.copy()
        ret.__parent = self.__parent
        ret.__accumNames.update( self.__accumNames )
        ret.__name = self.__name
        return ret

    def asXML( self, doctag=None, namedItemsOnly=False, indent="", formatted=True ):
        """
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        """
        nl = "\n"
        out = []
        namedItems = dict((v[1],k) for (k,vlist) in self.__tokdict.items()
                                                            for v in vlist)
        nextLevelIndent = indent + "  "

        # collapse out indents if formatting is not desired
        if not formatted:
            indent = ""
            nextLevelIndent = ""
            nl = ""

        selfTag = None
        if doctag is not None:
            selfTag = doctag
        else:
            if self.__name:
                selfTag = self.__name

        if not selfTag:
            if namedItemsOnly:
                return ""
            else:
                selfTag = "ITEM"

        out += [ nl, indent, "<", selfTag, ">" ]

        for i,res in enumerate(self.__toklist):
            if isinstance(res,ParseResults):
                if i in namedItems:
                    out += [ res.asXML(namedItems[i],
                                        namedItemsOnly and doctag is None,
                                        nextLevelIndent,
                                        formatted)]
                else:
                    out += [ res.asXML(None,
                                        namedItemsOnly and doctag is None,
                                        nextLevelIndent,
                                        formatted)]
            else:
                # individual token, see if there is a name for it
                resTag = None
                if i in namedItems:
                    resTag = namedItems[i]
                if not resTag:
                    if namedItemsOnly:
                        continue
                    else:
                        resTag = "ITEM"
                xmlBodyText = _xml_escape(_ustr(res))
                out += [ nl, nextLevelIndent, "<", resTag, ">",
                                                xmlBodyText,
                                                "</", resTag, ">" ]

        out += [ nl, indent, "</", selfTag, ">" ]
        return "".join(out)

    def __lookup(self,sub):
        for k,vlist in self.__tokdict.items():
            for v,loc in vlist:
                if sub is v:
                    return k
        return None

    def getName(self):
        r"""
        Returns the results name for this token expression. Useful when several 
        different expressions might match at a particular location.

        Example::
            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number") 
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)
            
            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])
        prints::
            age : 22
            ssn : 111-22-3333
            house_number : 221B
        """
        if self.__name:
            return self.__name
        elif self.__parent:
            par = self.__parent()
            if par:
                return par.__lookup(self)
            else:
                return None
        elif (len(self) == 1 and
               len(self.__tokdict) == 1 and
               next(iter(self.__tokdict.values()))[0][1] in (0,-1)):
            return next(iter(self.__tokdict.keys()))
        else:
            return None

    def dump(self, indent='', depth=0, full=True):
        """
        Diagnostic method for listing out the contents of a C{ParseResults}.
        Accepts an optional C{indent} argument so that this string can be embedded
        in a nested display of other data.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(result.dump())
        prints::
            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        """
        out = []
        NL = '\n'
        out.append( indent+_ustr(self.asList()) )
        if full:
            if self.haskeys():
                items = sorted((str(k), v) for k,v in self.items())
                for k,v in items:
                    if out:
                        out.append(NL)
                    out.append( "%s%s- %s: " % (indent,('  '*depth), k) )
                    if isinstance(v,ParseResults):
                        if v:
                            out.append( v.dump(indent,depth+1) )
                        else:
                            out.append(_ustr(v))
                    else:
                        out.append(repr(v))
            elif any(isinstance(vv,ParseResults) for vv in self):
                v = self
                for i,vv in enumerate(v):
                    if isinstance(vv,ParseResults):
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,('  '*(depth)),i,indent,('  '*(depth+1)),vv.dump(indent,depth+1) ))
                    else:
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,('  '*(depth)),i,indent,('  '*(depth+1)),_ustr(vv)))
            
        return "".join(out)

    def pprint(self, *args, **kwargs):
        """
        Pretty-printer for parsed results as a list, using the C{pprint} module.
        Accepts additional positional or keyword args as defined for the 
        C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})

        Example::
            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)
        prints::
            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        """
        pprint.pprint(self.asList(), *args, **kwargs)

    # add support for pickle protocol
    def __getstate__(self):
        return ( self.__toklist,
                 ( self.__tokdict.copy(),
                   self.__parent is not None and self.__parent() or None,
                   self.__accumNames,
                   self.__name ) )

    def __setstate__(self,state):
        self.__toklist = state[0]
        (self.__tokdict,
         par,
         inAccumNames,
         self.__name) = state[1]
        self.__accumNames = {}
        self.__accumNames.update(inAccumNames)
        if par is not None:
            self.__parent = wkref(par)
        else:
            self.__parent = None

    def __getnewargs__(self):
        return self.__toklist, self.__name, self.__asList, self.__modal

    def __dir__(self):
        return (dir(type(self)) + list(self.keys()))

MutableMapping.register(ParseResults)

def col (loc,strg):
    """Returns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   """
    s = strg
    return 1 if 0<loc<len(s) and s[loc-1] == '\n' else loc - s.rfind("\n", 0, loc)

def lineno(loc,strg):
    """Returns current line number within a string, counting newlines as line separators.
   The first line is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   """
    return strg.count("\n",0,loc) + 1

def line( loc, strg ):
    """Returns the line of text containing loc within a string, counting newlines as line separators.
       """
    lastCR = strg.rfind("\n", 0, loc)
    nextCR = strg.find("\n", loc)
    if nextCR >= 0:
        return strg[lastCR+1:nextCR]
    else:
        return strg[lastCR+1:]

def _defaultStartDebugAction( instring, loc, expr ):
    print (("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % ( lineno(loc,instring), col(loc,instring) )))

def _defaultSuccessDebugAction( instring, startloc, endloc, expr, toks ):
    print ("Matched " + _ustr(expr) + " -> " + str(toks.asList()))

def _defaultExceptionDebugAction( instring, loc, expr, exc ):
    print ("Exception raised:" + _ustr(exc))

def nullDebugAction(*args):
    """'Do-nothing' debug action, to suppress debugging output during parsing."""
    pass

# Only works on Python 3.x - nonlocal is toxic to Python 2 installs
#~ 'decorator to trim function calls to match the arity of the target'
#~ def _trim_arity(func, maxargs=3):
    #~ if func in singleArgBuiltins:
        #~ return lambda s,l,t: func(t)
    #~ limit = 0
    #~ foundArity = False
    #~ def wrapper(*args):
        #~ nonlocal limit,foundArity
        #~ while 1:
            #~ try:
                #~ ret = func(*args[limit:])
                #~ foundArity = True
                #~ return ret
            #~ except TypeError:
                #~ if limit == maxargs or foundArity:
                    #~ raise
                #~ limit += 1
                #~ continue
    #~ return wrapper

# this version is Python 2.x-3.x cross-compatible
'decorator to trim function calls to match the arity of the target'
def _trim_arity(func, maxargs=2):
    if func in singleArgBuiltins:
        return lambda s,l,t: func(t)
    limit = [0]
    foundArity = [False]
    
    # traceback return data structure changed in Py3.5 - normalize back to plain tuples
    if system_version[:2] >= (3,5):
        def extract_stack(limit=0):
            # special handling for Python 3.5.0 - extra deep call stack by 1
            offset = -3 if system_version == (3,5,0) else -2
            frame_summary = traceback.extract_stack(limit=-offset+limit-1)[offset]
            return [frame_summary[:2]]
        def extract_tb(tb, limit=0):
            frames = traceback.extract_tb(tb, limit=limit)
            frame_summary = frames[-1]
            return [frame_summary[:2]]
    else:
        extract_stack = traceback.extract_stack
        extract_tb = traceback.extract_tb
    
    # synthesize what would be returned by traceback.extract_stack at the call to 
    # user's parse action 'func', so that we don't incur call penalty at parse time
    
    LINE_DIFF = 6
    # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND 
    # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!!
    this_line = extract_stack(limit=2)[-1]
    pa_call_line_synth = (this_line[0], this_line[1]+LINE_DIFF)

    def wrapper(*args):
        while 1:
            try:
                ret = func(*args[limit[0]:])
                foundArity[0] = True
                return ret
            except TypeError:
                # re-raise TypeErrors if they did not come from our arity testing
                if foundArity[0]:
                    raise
                else:
                    try:
                        tb = sys.exc_info()[-1]
                        if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth:
                            raise
                    finally:
                        del tb

                if limit[0] <= maxargs:
                    limit[0] += 1
                    continue
                raise

    # copy func name to wrapper for sensible debug output
    func_name = "<parse action>"
    try:
        func_name = getattr(func, '__name__', 
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    wrapper.__name__ = func_name

    return wrapper

class ParserElement(object):
    """Abstract base level parser element class."""
    DEFAULT_WHITE_CHARS = " \n\t\r"
    verbose_stacktrace = False

    @staticmethod
    def setDefaultWhitespaceChars( chars ):
        r"""
        Overrides the default whitespace chars

        Example::
            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']
            
            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        """
        ParserElement.DEFAULT_WHITE_CHARS = chars

    @staticmethod
    def inlineLiteralsUsing(cls):
        """
        Set class to be used for inclusion of string literals into a parser.
        
        Example::
            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        """
        ParserElement._literalStringClass = cls

    def __init__( self, savelist=False ):
        self.parseAction = list()
        self.failAction = None
        #~ self.name = "<unknown>"  # don't define self.name, let subclasses try/except upcall
        self.strRepr = None
        self.resultsName = None
        self.saveAsList = savelist
        self.skipWhitespace = True
        self.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
        self.copyDefaultWhiteChars = True
        self.mayReturnEmpty = False # used when checking for left-recursion
        self.keepTabs = False
        self.ignoreExprs = list()
        self.debug = False
        self.streamlined = False
        self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index
        self.errmsg = ""
        self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all)
        self.debugActions = ( None, None, None ) #custom debug actions
        self.re = None
        self.callPreparse = True # used to avoid redundant calls to preParse
        self.callDuringTry = False

    def copy( self ):
        """
        Make a copy of this C{ParserElement}.  Useful for defining different parse actions
        for the same parsing pattern, using copies of the original parse element.
        
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
            
            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
        prints::
            [5120, 100, 655360, 268435456]
        Equivalent form of C{expr.copy()} is just C{expr()}::
            integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
        """
        cpy = copy.copy( self )
        cpy.parseAction = self.parseAction[:]
        cpy.ignoreExprs = self.ignoreExprs[:]
        if self.copyDefaultWhiteChars:
            cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
        return cpy

    def setName( self, name ):
        """
        Define name for this expression, makes debugging and exception messages clearer.
        
        Example::
            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        """
        self.name = name
        self.errmsg = "Expected " + self.name
        if hasattr(self,"exception"):
            self.exception.msg = self.errmsg
        return self

    def setResultsName( self, name, listAllMatches=False ):
        """
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original C{ParserElement} object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        C{expr("name")} in place of C{expr.setResultsName("name")} - 
        see L{I{__call__}<__call__>}.

        Example::
            date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        """
        newself = self.copy()
        if name.endswith("*"):
            name = name[:-1]
            listAllMatches=True
        newself.resultsName = name
        newself.modalResults = not listAllMatches
        return newself

    def setBreak(self,breakFlag = True):
        """Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set C{breakFlag} to True to enable, False to
           disable.
        """
        if breakFlag:
            _parseMethod = self._parse
            def breaker(instring, loc, doActions=True, callPreParse=True):
                import pdb
                pdb.set_trace()
                return _parseMethod( instring, loc, doActions, callPreParse )
            breaker._originalParseMethod = _parseMethod
            self._parse = breaker
        else:
            if hasattr(self._parse,"_originalParseMethod"):
                self._parse = self._parse._originalParseMethod
        return self

    def setParseAction( self, *fns, **kwargs ):
        """
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
        C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
         - s   = the original string being parsed (see note below)
         - loc = the location of the matching substring
         - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        Optional keyword arguments:
         - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See L{I{parseString}<parseString>} for more information
        on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
        consistent view of the parsed string, the parse location, and line and column
        positions within the parsed string.
        
        Example::
            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        """
        self.parseAction = list(map(_trim_arity, list(fns)))
        self.callDuringTry = kwargs.get("callDuringTry", False)
        return self

    def addParseAction( self, *fns, **kwargs ):
        """
        Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}<setParseAction>}.
        
        See examples in L{I{copy}<copy>}.
        """
        self.parseAction += list(map(_trim_arity, list(fns)))
        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def addCondition(self, *fns, **kwargs):
        """Add a boolean predicate function to expression's list of parse actions. See 
        L{I{setParseAction}<setParseAction>} for function call signatures. Unlike C{setParseAction}, 
        functions passed to C{addCondition} need to return boolean success/fail of the condition.

        Optional keyword arguments:
         - message = define a custom message to be used in the raised exception
         - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
         
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        """
        msg = kwargs.get("message", "failed user-defined condition")
        exc_type = ParseFatalException if kwargs.get("fatal", False) else ParseException
        for fn in fns:
            def pa(s,l,t):
                if not bool(_trim_arity(fn)(s,l,t)):
                    raise exc_type(s,l,msg)
            self.parseAction.append(pa)
        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def setFailAction( self, fn ):
        """Define action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           C{fn(s,loc,expr,err)} where:
            - s = string being parsed
            - loc = location where expression match was attempted and failed
            - expr = the parse expression that failed
            - err = the exception thrown
           The function returns no value.  It may throw C{L{ParseFatalException}}
           if it is desired to stop parsing immediately."""
        self.failAction = fn
        return self

    def _skipIgnorables( self, instring, loc ):
        exprsFound = True
        while exprsFound:
            exprsFound = False
            for e in self.ignoreExprs:
                try:
                    while 1:
                        loc,dummy = e._parse( instring, loc )
                        exprsFound = True
                except ParseException:
                    pass
        return loc

    def preParse( self, instring, loc ):
        if self.ignoreExprs:
            loc = self._skipIgnorables( instring, loc )

        if self.skipWhitespace:
            wt = self.whiteChars
            instrlen = len(instring)
            while loc < instrlen and instring[loc] in wt:
                loc += 1

        return loc

    def parseImpl( self, instring, loc, doActions=True ):
        return loc, []

    def postParse( self, instring, loc, tokenlist ):
        return tokenlist

    #~ @profile
    def _parseNoCache( self, instring, loc, doActions=True, callPreParse=True ):
        debugging = ( self.debug ) #and doActions )

        if debugging or self.failAction:
            #~ print ("Match",self,"at loc",loc,"(%d,%d)" % ( lineno(loc,instring), col(loc,instring) ))
            if (self.debugActions[0] ):
                self.debugActions[0]( instring, loc, self )
            if callPreParse and self.callPreparse:
                preloc = self.preParse( instring, loc )
            else:
                preloc = loc
            tokensStart = preloc
            try:
                try:
                    loc,tokens = self.parseImpl( instring, preloc, doActions )
                except IndexError:
                    raise ParseException( instring, len(instring), self.errmsg, self )
            except ParseBaseException as err:
                #~ print ("Exception raised:", err)
                if self.debugActions[2]:
                    self.debugActions[2]( instring, tokensStart, self, err )
                if self.failAction:
                    self.failAction( instring, tokensStart, self, err )
                raise
        else:
            if callPreParse and self.callPreparse:
                preloc = self.preParse( instring, loc )
            else:
                preloc = loc
            tokensStart = preloc
            if self.mayIndexError or preloc >= len(instring):
                try:
                    loc,tokens = self.parseImpl( instring, preloc, doActions )
                except IndexError:
                    raise ParseException( instring, len(instring), self.errmsg, self )
            else:
                loc,tokens = self.parseImpl( instring, preloc, doActions )

        tokens = self.postParse( instring, loc, tokens )

        retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults )
        if self.parseAction and (doActions or self.callDuringTry):
            if debugging:
                try:
                    for fn in self.parseAction:
                        tokens = fn( instring, tokensStart, retTokens )
                        if tokens is not None:
                            retTokens = ParseResults( tokens,
                                                      self.resultsName,
                                                      asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
                                                      modal=self.modalResults )
                except ParseBaseException as err:
                    #~ print "Exception raised in user parse action:", err
                    if (self.debugActions[2] ):
                        self.debugActions[2]( instring, tokensStart, self, err )
                    raise
            else:
                for fn in self.parseAction:
                    tokens = fn( instring, tokensStart, retTokens )
                    if tokens is not None:
                        retTokens = ParseResults( tokens,
                                                  self.resultsName,
                                                  asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
                                                  modal=self.modalResults )
        if debugging:
            #~ print ("Matched",self,"->",retTokens.asList())
            if (self.debugActions[1] ):
                self.debugActions[1]( instring, tokensStart, loc, self, retTokens )

        return loc, retTokens

    def tryParse( self, instring, loc ):
        try:
            return self._parse( instring, loc, doActions=False )[0]
        except ParseFatalException:
            raise ParseException( instring, loc, self.errmsg, self)
    
    def canParseNext(self, instring, loc):
        try:
            self.tryParse(instring, loc)
        except (ParseException, IndexError):
            return False
        else:
            return True

    class _UnboundedCache(object):
        def __init__(self):
            cache = {}
            self.not_in_cache = not_in_cache = object()

            def get(self, key):
                return cache.get(key, not_in_cache)

            def set(self, key, value):
                cache[key] = value

            def clear(self):
                cache.clear()
                
            def cache_len(self):
                return len(cache)

            self.get = types.MethodType(get, self)
            self.set = types.MethodType(set, self)
            self.clear = types.MethodType(clear, self)
            self.__len__ = types.MethodType(cache_len, self)

    if _OrderedDict is not None:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = _OrderedDict()

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(cache) > size:
                        try:
                            cache.popitem(False)
                        except KeyError:
                            pass

                def clear(self):
                    cache.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    else:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = {}
                key_fifo = collections.deque([], size)

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(key_fifo) > size:
                        cache.pop(key_fifo.popleft(), None)
                    key_fifo.append(key)

                def clear(self):
                    cache.clear()
                    key_fifo.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    # argument cache for optimizing repeated calls when backtracking through recursive expressions
    packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail
    packrat_cache_lock = RLock()
    packrat_cache_stats = [0, 0]

    # this method gets repeatedly called during backtracking with the same arguments -
    # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression
    def _parseCache( self, instring, loc, doActions=True, callPreParse=True ):
        HIT, MISS = 0, 1
        lookup = (self, instring, loc, callPreParse, doActions)
        with ParserElement.packrat_cache_lock:
            cache = ParserElement.packrat_cache
            value = cache.get(lookup)
            if value is cache.not_in_cache:
                ParserElement.packrat_cache_stats[MISS] += 1
                try:
                    value = self._parseNoCache(instring, loc, doActions, callPreParse)
                except ParseBaseException as pe:
                    # cache a copy of the exception, without the traceback
                    cache.set(lookup, pe.__class__(*pe.args))
                    raise
                else:
                    cache.set(lookup, (value[0], value[1].copy()))
                    return value
            else:
                ParserElement.packrat_cache_stats[HIT] += 1
                if isinstance(value, Exception):
                    raise value
                return (value[0], value[1].copy())

    _parse = _parseNoCache

    @staticmethod
    def resetCache():
        ParserElement.packrat_cache.clear()
        ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats)

    _packratEnabled = False
    @staticmethod
    def enablePackrat(cache_size_limit=128):
        """Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.
           
           Parameters:
            - cache_size_limit - (default=C{128}) - if an integer value is provided
              will limit the size of the packrat cache; if None is passed, then
              the cache size will be unbounded; if 0 is passed, the cache will
              be effectively disabled.
            
           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method C{ParserElement.enablePackrat()}.  If
           your program uses C{psyco} to "compile as you go", you must call
           C{enablePackrat} before calling C{psyco.full()}.  If you do not do this,
           Python will crash.  For best results, call C{enablePackrat()} immediately
           after importing pyparsing.
           
           Example::
               import pyparsing
               pyparsing.ParserElement.enablePackrat()
        """
        if not ParserElement._packratEnabled:
            ParserElement._packratEnabled = True
            if cache_size_limit is None:
                ParserElement.packrat_cache = ParserElement._UnboundedCache()
            else:
                ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit)
            ParserElement._parse = ParserElement._parseCache

    def parseString( self, instring, parseAll=False ):
        """
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set C{parseAll} to True (equivalent to ending
        the grammar with C{L{StringEnd()}}).

        Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the C{loc} argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:
         - calling C{parseWithTabs} on your grammar before calling C{parseString}
           (see L{I{parseWithTabs}<parseWithTabs>})
         - define your parse action using the full C{(s,loc,toks)} signature, and
           reference the input string using the parse action's C{s} argument
         - explicitly expand the tabs in your input string before calling
           C{parseString}
        
        Example::
            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        """
        ParserElement.resetCache()
        if not self.streamlined:
            self.streamline()
            #~ self.saveAsList = True
        for e in self.ignoreExprs:
            e.streamline()
        if not self.keepTabs:
            instring = instring.expandtabs()
        try:
            loc, tokens = self._parse( instring, 0 )
            if parseAll:
                loc = self.preParse( instring, loc )
                se = Empty() + StringEnd()
                se._parse( instring, loc )
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc
        else:
            return tokens

    def scanString( self, instring, maxMatches=_MAX_INT, overlap=False ):
        """
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        C{maxMatches} argument, to clip scanning after 'n' matches are found.  If
        C{overlap} is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See L{I{parseString}<parseString>} for more information on parsing
        strings with embedded tabs.

        Example::
            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens,start,end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])
        
        prints::
        
            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        """
        if not self.streamlined:
            self.streamline()
        for e in self.ignoreExprs:
            e.streamline()

        if not self.keepTabs:
            instring = _ustr(instring).expandtabs()
        instrlen = len(instring)
        loc = 0
        preparseFn = self.preParse
        parseFn = self._parse
        ParserElement.resetCache()
        matches = 0
        try:
            while loc <= instrlen and matches < maxMatches:
                try:
                    preloc = preparseFn( instring, loc )
                    nextLoc,tokens = parseFn( instring, preloc, callPreParse=False )
                except ParseException:
                    loc = preloc+1
                else:
                    if nextLoc > loc:
                        matches += 1
                        yield tokens, preloc, nextLoc
                        if overlap:
                            nextloc = preparseFn( instring, loc )
                            if nextloc > loc:
                                loc = nextLoc
                            else:
                                loc += 1
                        else:
                            loc = nextLoc
                    else:
                        loc = preloc+1
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def transformString( self, instring ):
        """
        Extension to C{L{scanString}}, to modify matching text with modified tokens that may
        be returned from a parse action.  To use C{transformString}, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking C{transformString()} on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  C{transformString()} returns the resulting transformed string.
        
        Example::
            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())
            
            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
        Prints::
            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        """
        out = []
        lastE = 0
        # force preservation of <TAB>s, to minimize unwanted transformation of string, and to
        # keep string locs straight between transformString and scanString
        self.keepTabs = True
        try:
            for t,s,e in self.scanString( instring ):
                out.append( instring[lastE:s] )
                if t:
                    if isinstance(t,ParseResults):
                        out += t.asList()
                    elif isinstance(t,list):
                        out += t
                    else:
                        out.append(t)
                lastE = e
            out.append(instring[lastE:])
            out = [o for o in out if o]
            return "".join(map(_ustr,_flatten(out)))
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def searchString( self, instring, maxMatches=_MAX_INT ):
        """
        Another extension to C{L{scanString}}, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        C{maxMatches} argument, to clip searching after 'n' matches are found.
        
        Example::
            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())
            
            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))
        prints::
            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        """
        try:
            return ParseResults([ t for t,s,e in self.scanString( instring, maxMatches ) ])
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False):
        """
        Generator method to split a string using the given expression as a separator.
        May be called with optional C{maxsplit} argument, to limit the number of splits;
        and the optional C{includeSeparators} argument (default=C{False}), if the separating
        matching text should be included in the split results.
        
        Example::        
            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
        prints::
            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        """
        splits = 0
        last = 0
        for t,s,e in self.scanString(instring, maxMatches=maxsplit):
            yield instring[last:s]
            if includeSeparators:
                yield t[0]
            last = e
        yield instring[last:]

    def __add__(self, other ):
        """
        Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
        converts them to L{Literal}s by default.
        
        Example::
            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))
        Prints::
            Hello, World! -> ['Hello', ',', 'World', '!']
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return And( [ self, other ] )

    def __radd__(self, other ):
        """
        Implementation of + operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other + self

    def __sub__(self, other):
        """
        Implementation of - operator, returns C{L{And}} with error stop
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return self + And._ErrorStop() + other

    def __rsub__(self, other ):
        """
        Implementation of - operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other - self

    def __mul__(self,other):
        """
        Implementation of * operator, allows use of C{expr * 3} in place of
        C{expr + expr + expr}.  Expressions may also me multiplied by a 2-integer
        tuple, similar to C{{min,max}} multipliers in regular expressions.  Tuples
        may also include C{None} as in:
         - C{expr*(n,None)} or C{expr*(n,)} is equivalent
              to C{expr*n + L{ZeroOrMore}(expr)}
              (read as "at least n instances of C{expr}")
         - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
              (read as "0 to n instances of C{expr}")
         - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
         - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}

        Note that C{expr*(None,n)} does not raise an exception if
        more than n exprs exist in the input stream; that is,
        C{expr*(None,n)} does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        C{expr*(None,n) + ~expr}
        """
        if isinstance(other,int):
            minElements, optElements = other,0
        elif isinstance(other,tuple):
            other = (other + (None, None))[:2]
            if other[0] is None:
                other = (0, other[1])
            if isinstance(other[0],int) and other[1] is None:
                if other[0] == 0:
                    return ZeroOrMore(self)
                if other[0] == 1:
                    return OneOrMore(self)
                else:
                    return self*other[0] + ZeroOrMore(self)
            elif isinstance(other[0],int) and isinstance(other[1],int):
                minElements, optElements = other
                optElements -= minElements
            else:
                raise TypeError("cannot multiply 'ParserElement' and ('%s','%s') objects", type(other[0]),type(other[1]))
        else:
            raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other))

        if minElements < 0:
            raise ValueError("cannot multiply ParserElement by negative value")
        if optElements < 0:
            raise ValueError("second tuple value must be greater or equal to first tuple value")
        if minElements == optElements == 0:
            raise ValueError("cannot multiply ParserElement by 0 or (0,0)")

        if (optElements):
            def makeOptionalList(n):
                if n>1:
                    return Optional(self + makeOptionalList(n-1))
                else:
                    return Optional(self)
            if minElements:
                if minElements == 1:
                    ret = self + makeOptionalList(optElements)
                else:
                    ret = And([self]*minElements) + makeOptionalList(optElements)
            else:
                ret = makeOptionalList(optElements)
        else:
            if minElements == 1:
                ret = self
            else:
                ret = And([self]*minElements)
        return ret

    def __rmul__(self, other):
        return self.__mul__(other)

    def __or__(self, other ):
        """
        Implementation of | operator - returns C{L{MatchFirst}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return MatchFirst( [ self, other ] )

    def __ror__(self, other ):
        """
        Implementation of | operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other | self

    def __xor__(self, other ):
        """
        Implementation of ^ operator - returns C{L{Or}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return Or( [ self, other ] )

    def __rxor__(self, other ):
        """
        Implementation of ^ operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other ^ self

    def __and__(self, other ):
        """
        Implementation of & operator - returns C{L{Each}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return Each( [ self, other ] )

    def __rand__(self, other ):
        """
        Implementation of & operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other & self

    def __invert__( self ):
        """
        Implementation of ~ operator - returns C{L{NotAny}}
        """
        return NotAny( self )

    def __call__(self, name=None):
        """
        Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
        
        If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
        passed as C{True}.
           
        If C{name} is omitted, same as calling C{L{copy}}.

        Example::
            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")             
        """
        if name is not None:
            return self.setResultsName(name)
        else:
            return self.copy()

    def suppress( self ):
        """
        Suppresses the output of this C{ParserElement}; useful to keep punctuation from
        cluttering up returned output.
        """
        return Suppress( self )

    def leaveWhitespace( self ):
        """
        Disables the skipping of whitespace before matching the characters in the
        C{ParserElement}'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        """
        self.skipWhitespace = False
        return self

    def setWhitespaceChars( self, chars ):
        """
        Overrides the default whitespace chars
        """
        self.skipWhitespace = True
        self.whiteChars = chars
        self.copyDefaultWhiteChars = False
        return self

    def parseWithTabs( self ):
        """
        Overrides default behavior to expand C{<TAB>}s to spaces before parsing the input string.
        Must be called before C{parseString} when the input grammar contains elements that
        match C{<TAB>} characters.
        """
        self.keepTabs = True
        return self

    def ignore( self, other ):
        """
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.
        
        Example::
            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
            
            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        """
        if isinstance(other, basestring):
            other = Suppress(other)

        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                self.ignoreExprs.append(other)
        else:
            self.ignoreExprs.append( Suppress( other.copy() ) )
        return self

    def setDebugActions( self, startAction, successAction, exceptionAction ):
        """
        Enable display of debugging messages while doing pattern matching.
        """
        self.debugActions = (startAction or _defaultStartDebugAction,
                             successAction or _defaultSuccessDebugAction,
                             exceptionAction or _defaultExceptionDebugAction)
        self.debug = True
        return self

    def setDebug( self, flag=True ):
        """
        Enable display of debugging messages while doing pattern matching.
        Set C{flag} to True to enable, False to disable.

        Example::
            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer
            
            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")
        
        prints::
            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using L{setDebugActions}. Prior to attempting
        to match the C{wd} expression, the debugging message C{"Match <exprname> at loc <n>(<line>,<col>)"}
        is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
        message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
        """
        if flag:
            self.setDebugActions( _defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction )
        else:
            self.debug = False
        return self

    def __str__( self ):
        return self.name

    def __repr__( self ):
        return _ustr(self)

    def streamline( self ):
        self.streamlined = True
        self.strRepr = None
        return self

    def checkRecursion( self, parseElementList ):
        pass

    def validate( self, validateTrace=[] ):
        """
        Check defined expressions for valid structure, check for infinite recursive definitions.
        """
        self.checkRecursion( [] )

    def parseFile( self, file_or_filename, parseAll=False ):
        """
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        """
        try:
            file_contents = file_or_filename.read()
        except AttributeError:
            with open(file_or_filename, "r") as f:
                file_contents = f.read()
        try:
            return self.parseString(file_contents, parseAll)
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def __eq__(self,other):
        if isinstance(other, ParserElement):
            return self is other or vars(self) == vars(other)
        elif isinstance(other, basestring):
            return self.matches(other)
        else:
            return super(ParserElement,self)==other

    def __ne__(self,other):
        return not (self == other)

    def __hash__(self):
        return hash(id(self))

    def __req__(self,other):
        return self == other

    def __rne__(self,other):
        return not (self == other)

    def matches(self, testString, parseAll=True):
        """
        Method for quick testing of a parser against a test string. Good for simple 
        inline microtests of sub expressions while building up larger parser.
           
        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
            
        Example::
            expr = Word(nums)
            assert expr.matches("100")
        """
        try:
            self.parseString(_ustr(testString), parseAll=parseAll)
            return True
        except ParseBaseException:
            return False
                
    def runTests(self, tests, parseAll=True, comment='#', fullDump=True, printResults=True, failureTests=False):
        """
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.
           
        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests           
         - comment - (default=C{'#'}) - expression for indicating embedded comments in the test 
              string; pass None to disable comment filtering
         - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default=C{True}) prints test output to stdout
         - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if C{failureTests} is True), and the results contain a list of lines of each 
        test's output
        
        Example::
            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")
        prints::
            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success
            
            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines")
        
        (Note that this is a raw string literal, you must include the leading 'r'.)
        """
        if isinstance(tests, basestring):
            tests = list(map(str.strip, tests.rstrip().splitlines()))
        if isinstance(comment, basestring):
            comment = Literal(comment)
        allResults = []
        comments = []
        success = True
        for t in tests:
            if comment is not None and comment.matches(t, False) or comments and not t:
                comments.append(t)
                continue
            if not t:
                continue
            out = ['\n'.join(comments), t]
            comments = []
            try:
                t = t.replace(r'\n','\n')
                result = self.parseString(t, parseAll=parseAll)
                out.append(result.dump(full=fullDump))
                success = success and not failureTests
            except ParseBaseException as pe:
                fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else ""
                if '\n' in t:
                    out.append(line(pe.loc, t))
                    out.append(' '*(col(pe.loc,t)-1) + '^' + fatal)
                else:
                    out.append(' '*pe.loc + '^' + fatal)
                out.append("FAIL: " + str(pe))
                success = success and failureTests
                result = pe
            except Exception as exc:
                out.append("FAIL-EXCEPTION: " + str(exc))
                success = success and failureTests
                result = exc

            if printResults:
                if fullDump:
                    out.append('')
                print('\n'.join(out))

            allResults.append((t, result))
        
        return success, allResults

        
class Token(ParserElement):
    """
    Abstract C{ParserElement} subclass, for defining atomic matching patterns.
    """
    def __init__( self ):
        super(Token,self).__init__( savelist=False )


class Empty(Token):
    """
    An empty token, will always match.
    """
    def __init__( self ):
        super(Empty,self).__init__()
        self.name = "Empty"
        self.mayReturnEmpty = True
        self.mayIndexError = False


class NoMatch(Token):
    """
    A token that will never match.
    """
    def __init__( self ):
        super(NoMatch,self).__init__()
        self.name = "NoMatch"
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.errmsg = "Unmatchable token"

    def parseImpl( self, instring, loc, doActions=True ):
        raise ParseException(instring, loc, self.errmsg, self)


class Literal(Token):
    """
    Token to exactly match a specified string.
    
    Example::
        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"
    
    For case-insensitive matching, use L{CaselessLiteral}.
    
    For keyword matching (force word break before and after the matched string),
    use L{Keyword} or L{CaselessKeyword}.
    """
    def __init__( self, matchString ):
        super(Literal,self).__init__()
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Literal; use Empty() instead",
                            SyntaxWarning, stacklevel=2)
            self.__class__ = Empty
        self.name = '"%s"' % _ustr(self.match)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False

    # Performance tuning: this routine gets called a *lot*
    # if this is a single character match string  and the first character matches,
    # short-circuit as quickly as possible, and avoid calling startswith
    #~ @profile
    def parseImpl( self, instring, loc, doActions=True ):
        if (instring[loc] == self.firstMatchChar and
            (self.matchLen==1 or instring.startswith(self.match,loc)) ):
            return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)
_L = Literal
ParserElement._literalStringClass = Literal

class Keyword(Token):
    """
    Token to exactly match a specified string as a keyword, that is, it must be
    immediately followed by a non-keyword character.  Compare with C{L{Literal}}:
     - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
     - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
    Accepts two optional constructor arguments in addition to the keyword string:
     - C{identChars} is a string of characters that would be valid identifier characters,
          defaulting to all alphanumerics + "_" and "$"
     - C{caseless} allows case-insensitive matching, default is C{False}.
       
    Example::
        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use L{CaselessKeyword}.
    """
    DEFAULT_KEYWORD_CHARS = alphanums+"_$"

    def __init__( self, matchString, identChars=None, caseless=False ):
        super(Keyword,self).__init__()
        if identChars is None:
            identChars = Keyword.DEFAULT_KEYWORD_CHARS
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Keyword; use Empty() instead",
                            SyntaxWarning, stacklevel=2)
        self.name = '"%s"' % self.match
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False
        self.caseless = caseless
        if caseless:
            self.caselessmatch = matchString.upper()
            identChars = identChars.upper()
        self.identChars = set(identChars)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.caseless:
            if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
                 (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) and
                 (loc == 0 or instring[loc-1].upper() not in self.identChars) ):
                return loc+self.matchLen, self.match
        else:
            if (instring[loc] == self.firstMatchChar and
                (self.matchLen==1 or instring.startswith(self.match,loc)) and
                (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen] not in self.identChars) and
                (loc == 0 or instring[loc-1] not in self.identChars) ):
                return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)

    def copy(self):
        c = super(Keyword,self).copy()
        c.identChars = Keyword.DEFAULT_KEYWORD_CHARS
        return c

    @staticmethod
    def setDefaultKeywordChars( chars ):
        """Overrides the default Keyword chars
        """
        Keyword.DEFAULT_KEYWORD_CHARS = chars

class CaselessLiteral(Literal):
    """
    Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::
        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
        
    (Contrast with example for L{CaselessKeyword}.)
    """
    def __init__( self, matchString ):
        super(CaselessLiteral,self).__init__( matchString.upper() )
        # Preserve the defining literal.
        self.returnString = matchString
        self.name = "'%s'" % self.returnString
        self.errmsg = "Expected " + self.name

    def parseImpl( self, instring, loc, doActions=True ):
        if instring[ loc:loc+self.matchLen ].upper() == self.match:
            return loc+self.matchLen, self.returnString
        raise ParseException(instring, loc, self.errmsg, self)

class CaselessKeyword(Keyword):
    """
    Caseless version of L{Keyword}.

    Example::
        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
        
    (Contrast with example for L{CaselessLiteral}.)
    """
    def __init__( self, matchString, identChars=None ):
        super(CaselessKeyword,self).__init__( matchString, identChars, caseless=True )

    def parseImpl( self, instring, loc, doActions=True ):
        if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
             (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) ):
            return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)

class CloseMatch(Token):
    """
    A variation on L{Literal} which matches "close" matches, that is, 
    strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
     - C{match_string} - string to be matched
     - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
    
    The results from a successful parse will contain the matched text from the input string and the following named results:
     - C{mismatches} - a list of the positions within the match_string where mismatches were found
     - C{original} - the original match_string used to compare against the input string
    
    If C{mismatches} is an empty list, then the match was an exact match.
    
    Example::
        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    """
    def __init__(self, match_string, maxMismatches=1):
        super(CloseMatch,self).__init__()
        self.name = match_string
        self.match_string = match_string
        self.maxMismatches = maxMismatches
        self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches)
        self.mayIndexError = False
        self.mayReturnEmpty = False

    def parseImpl( self, instring, loc, doActions=True ):
        start = loc
        instrlen = len(instring)
        maxloc = start + len(self.match_string)

        if maxloc <= instrlen:
            match_string = self.match_string
            match_stringloc = 0
            mismatches = []
            maxMismatches = self.maxMismatches

            for match_stringloc,s_m in enumerate(zip(instring[loc:maxloc], self.match_string)):
                src,mat = s_m
                if src != mat:
                    mismatches.append(match_stringloc)
                    if len(mismatches) > maxMismatches:
                        break
            else:
                loc = match_stringloc + 1
                results = ParseResults([instring[start:loc]])
                results['original'] = self.match_string
                results['mismatches'] = mismatches
                return loc, results

        raise ParseException(instring, loc, self.errmsg, self)


class Word(Token):
    """
    Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters,
    an optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction. An optional
    C{excludeChars} parameter can list characters that might be found in 
    the input C{bodyChars} string; useful to define a word of all printables
    except for one or two characters, for instance.
    
    L{srange} is useful for defining custom character set strings for defining 
    C{Word} expressions, using range notation from regular expression character sets.
    
    A common mistake is to use C{Word} to match a specific literal string, as in 
    C{Word("Address")}. Remember that C{Word} uses the string argument to define
    I{sets} of matchable characters. This expression would match "Add", "AAA",
    "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
    To match an exact literal string, use L{Literal} or L{Keyword}.

    pyparsing includes helper strings for building Words:
     - L{alphas}
     - L{nums}
     - L{alphanums}
     - L{hexnums}
     - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
     - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - L{printables} (any non-whitespace character)

    Example::
        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
        
        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums+'-')
        
        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")
        
        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    """
    def __init__( self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None ):
        super(Word,self).__init__()
        if excludeChars:
            initChars = ''.join(c for c in initChars if c not in excludeChars)
            if bodyChars:
                bodyChars = ''.join(c for c in bodyChars if c not in excludeChars)
        self.initCharsOrig = initChars
        self.initChars = set(initChars)
        if bodyChars :
            self.bodyCharsOrig = bodyChars
            self.bodyChars = set(bodyChars)
        else:
            self.bodyCharsOrig = initChars
            self.bodyChars = set(initChars)

        self.maxSpecified = max > 0

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.asKeyword = asKeyword

        if ' ' not in self.initCharsOrig+self.bodyCharsOrig and (min==1 and max==0 and exact==0):
            if self.bodyCharsOrig == self.initCharsOrig:
                self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig)
            elif len(self.initCharsOrig) == 1:
                self.reString = "%s[%s]*" % \
                                      (re.escape(self.initCharsOrig),
                                      _escapeRegexRangeChars(self.bodyCharsOrig),)
            else:
                self.reString = "[%s][%s]*" % \
                                      (_escapeRegexRangeChars(self.initCharsOrig),
                                      _escapeRegexRangeChars(self.bodyCharsOrig),)
            if self.asKeyword:
                self.reString = r"\b"+self.reString+r"\b"
            try:
                self.re = re.compile( self.reString )
            except Exception:
                self.re = None

    def parseImpl( self, instring, loc, doActions=True ):
        if self.re:
            result = self.re.match(instring,loc)
            if not result:
                raise ParseException(instring, loc, self.errmsg, self)

            loc = result.end()
            return loc, result.group()

        if not(instring[ loc ] in self.initChars):
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        instrlen = len(instring)
        bodychars = self.bodyChars
        maxloc = start + self.maxLen
        maxloc = min( maxloc, instrlen )
        while loc < maxloc and instring[loc] in bodychars:
            loc += 1

        throwException = False
        if loc - start < self.minLen:
            throwException = True
        if self.maxSpecified and loc < instrlen and instring[loc] in bodychars:
            throwException = True
        if self.asKeyword:
            if (start>0 and instring[start-1] in bodychars) or (loc<instrlen and instring[loc] in bodychars):
                throwException = True

        if throwException:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__( self ):
        try:
            return super(Word,self).__str__()
        except Exception:
            pass


        if self.strRepr is None:

            def charsAsStr(s):
                if len(s)>4:
                    return s[:4]+"..."
                else:
                    return s

            if ( self.initCharsOrig != self.bodyCharsOrig ):
                self.strRepr = "W:(%s,%s)" % ( charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig) )
            else:
                self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig)

        return self.strRepr


class Regex(Token):
    r"""
    Token for matching strings that match a given regular expression.
    Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
    If the given regex contains named groups (defined using C{(?P<name>...)}), these will be preserved as 
    named parse results.

    Example::
        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    """
    compiledREtype = type(re.compile("[A-Z]"))
    def __init__( self, pattern, flags=0):
        """The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags."""
        super(Regex,self).__init__()

        if isinstance(pattern, basestring):
            if not pattern:
                warnings.warn("null string passed to Regex; use Empty() instead",
                        SyntaxWarning, stacklevel=2)

            self.pattern = pattern
            self.flags = flags

            try:
                self.re = re.compile(self.pattern, self.flags)
                self.reString = self.pattern
            except sre_constants.error:
                warnings.warn("invalid pattern (%s) passed to Regex" % pattern,
                    SyntaxWarning, stacklevel=2)
                raise

        elif isinstance(pattern, Regex.compiledREtype):
            self.re = pattern
            self.pattern = \
            self.reString = str(pattern)
            self.flags = flags
            
        else:
            raise ValueError("Regex may only be constructed with a string or a compiled RE object")

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        result = self.re.match(instring,loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        d = result.groupdict()
        ret = ParseResults(result.group())
        if d:
            for k in d:
                ret[k] = d[k]
        return loc,ret

    def __str__( self ):
        try:
            return super(Regex,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "Re:(%s)" % repr(self.pattern)

        return self.strRepr


class QuotedString(Token):
    r"""
    Token for matching strings that are delimited by quoting characters.
    
    Defined with the following parameters:
        - quoteChar - string of one or more characters defining the quote delimiting string
        - escChar - character to escape quotes, typically backslash (default=C{None})
        - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
        - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
        - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
        - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})

    Example::
        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
    prints::
        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    """
    def __init__( self, quoteChar, escChar=None, escQuote=None, multiline=False, unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True):
        super(QuotedString,self).__init__()

        # remove white space from quote chars - wont work anyway
        quoteChar = quoteChar.strip()
        if not quoteChar:
            warnings.warn("quoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
            raise SyntaxError()

        if endQuoteChar is None:
            endQuoteChar = quoteChar
        else:
            endQuoteChar = endQuoteChar.strip()
            if not endQuoteChar:
                warnings.warn("endQuoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
                raise SyntaxError()

        self.quoteChar = quoteChar
        self.quoteCharLen = len(quoteChar)
        self.firstQuoteChar = quoteChar[0]
        self.endQuoteChar = endQuoteChar
        self.endQuoteCharLen = len(endQuoteChar)
        self.escChar = escChar
        self.escQuote = escQuote
        self.unquoteResults = unquoteResults
        self.convertWhitespaceEscapes = convertWhitespaceEscapes

        if multiline:
            self.flags = re.MULTILINE | re.DOTALL
            self.pattern = r'%s(?:[^%s%s]' % \
                ( re.escape(self.quoteChar),
                  _escapeRegexRangeChars(self.endQuoteChar[0]),
                  (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
        else:
            self.flags = 0
            self.pattern = r'%s(?:[^%s\n\r%s]' % \
                ( re.escape(self.quoteChar),
                  _escapeRegexRangeChars(self.endQuoteChar[0]),
                  (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
        if len(self.endQuoteChar) > 1:
            self.pattern += (
                '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]),
                                               _escapeRegexRangeChars(self.endQuoteChar[i]))
                                    for i in range(len(self.endQuoteChar)-1,0,-1)) + ')'
                )
        if escQuote:
            self.pattern += (r'|(?:%s)' % re.escape(escQuote))
        if escChar:
            self.pattern += (r'|(?:%s.)' % re.escape(escChar))
            self.escCharReplacePattern = re.escape(self.escChar)+"(.)"
        self.pattern += (r')*%s' % re.escape(self.endQuoteChar))

        try:
            self.re = re.compile(self.pattern, self.flags)
            self.reString = self.pattern
        except sre_constants.error:
            warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern,
                SyntaxWarning, stacklevel=2)
            raise

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        result = instring[loc] == self.firstQuoteChar and self.re.match(instring,loc) or None
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = result.group()

        if self.unquoteResults:

            # strip off quotes
            ret = ret[self.quoteCharLen:-self.endQuoteCharLen]

            if isinstance(ret,basestring):
                # replace escaped whitespace
                if '\\' in ret and self.convertWhitespaceEscapes:
                    ws_map = {
                        r'\t' : '\t',
                        r'\n' : '\n',
                        r'\f' : '\f',
                        r'\r' : '\r',
                    }
                    for wslit,wschar in ws_map.items():
                        ret = ret.replace(wslit, wschar)

                # replace escaped characters
                if self.escChar:
                    ret = re.sub(self.escCharReplacePattern, r"\g<1>", ret)

                # replace escaped quotes
                if self.escQuote:
                    ret = ret.replace(self.escQuote, self.endQuoteChar)

        return loc, ret

    def __str__( self ):
        try:
            return super(QuotedString,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar)

        return self.strRepr


class CharsNotIn(Token):
    """
    Token for matching words composed of characters I{not} in a given set (will
    include whitespace in matched characters if not listed in the provided exclusion set - see example).
    Defined with string containing all disallowed characters, and an optional
    minimum, maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction.

    Example::
        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
    prints::
        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    """
    def __init__( self, notChars, min=1, max=0, exact=0 ):
        super(CharsNotIn,self).__init__()
        self.skipWhitespace = False
        self.notChars = notChars

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = ( self.minLen == 0 )
        self.mayIndexError = False

    def parseImpl( self, instring, loc, doActions=True ):
        if instring[loc] in self.notChars:
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        notchars = self.notChars
        maxlen = min( start+self.maxLen, len(instring) )
        while loc < maxlen and \
              (instring[loc] not in notchars):
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__( self ):
        try:
            return super(CharsNotIn, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            if len(self.notChars) > 4:
                self.strRepr = "!W:(%s...)" % self.notChars[:4]
            else:
                self.strRepr = "!W:(%s)" % self.notChars

        return self.strRepr

class White(Token):
    """
    Special matching class for matching whitespace.  Normally, whitespace is ignored
    by pyparsing grammars.  This class is included when some whitespace structures
    are significant.  Define with a string containing the whitespace characters to be
    matched; default is C{" \\t\\r\\n"}.  Also takes optional C{min}, C{max}, and C{exact} arguments,
    as defined for the C{L{Word}} class.
    """
    whiteStrs = {
        " " : "<SPC>",
        "\t": "<TAB>",
        "\n": "<LF>",
        "\r": "<CR>",
        "\f": "<FF>",
        }
    def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0):
        super(White,self).__init__()
        self.matchWhite = ws
        self.setWhitespaceChars( "".join(c for c in self.whiteChars if c not in self.matchWhite) )
        #~ self.leaveWhitespace()
        self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite))
        self.mayReturnEmpty = True
        self.errmsg = "Expected " + self.name

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

    def parseImpl( self, instring, loc, doActions=True ):
        if not(instring[ loc ] in self.matchWhite):
            raise ParseException(instring, loc, self.errmsg, self)
        start = loc
        loc += 1
        maxloc = start + self.maxLen
        maxloc = min( maxloc, len(instring) )
        while loc < maxloc and instring[loc] in self.matchWhite:
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]


class _PositionToken(Token):
    def __init__( self ):
        super(_PositionToken,self).__init__()
        self.name=self.__class__.__name__
        self.mayReturnEmpty = True
        self.mayIndexError = False

class GoToColumn(_PositionToken):
    """
    Token to advance to a specific column of input text; useful for tabular report scraping.
    """
    def __init__( self, colno ):
        super(GoToColumn,self).__init__()
        self.col = colno

    def preParse( self, instring, loc ):
        if col(loc,instring) != self.col:
            instrlen = len(instring)
            if self.ignoreExprs:
                loc = self._skipIgnorables( instring, loc )
            while loc < instrlen and instring[loc].isspace() and col( loc, instring ) != self.col :
                loc += 1
        return loc

    def parseImpl( self, instring, loc, doActions=True ):
        thiscol = col( loc, instring )
        if thiscol > self.col:
            raise ParseException( instring, loc, "Text not in expected column", self )
        newloc = loc + self.col - thiscol
        ret = instring[ loc: newloc ]
        return newloc, ret


class LineStart(_PositionToken):
    """
    Matches if current position is at the beginning of a line within the parse string
    
    Example::
    
        test = '''\
        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)
    
    Prints::
        ['AAA', ' this line']
        ['AAA', ' and this line']    

    """
    def __init__( self ):
        super(LineStart,self).__init__()
        self.errmsg = "Expected start of line"

    def parseImpl( self, instring, loc, doActions=True ):
        if col(loc, instring) == 1:
            return loc, []
        raise ParseException(instring, loc, self.errmsg, self)

class LineEnd(_PositionToken):
    """
    Matches if current position is at the end of a line within the parse string
    """
    def __init__( self ):
        super(LineEnd,self).__init__()
        self.setWhitespaceChars( ParserElement.DEFAULT_WHITE_CHARS.replace("\n","") )
        self.errmsg = "Expected end of line"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc<len(instring):
            if instring[loc] == "\n":
                return loc+1, "\n"
            else:
                raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc+1, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class StringStart(_PositionToken):
    """
    Matches if current position is at the beginning of the parse string
    """
    def __init__( self ):
        super(StringStart,self).__init__()
        self.errmsg = "Expected start of text"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc != 0:
            # see if entire string up to here is just whitespace and ignoreables
            if loc != self.preParse( instring, 0 ):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class StringEnd(_PositionToken):
    """
    Matches if current position is at the end of the parse string
    """
    def __init__( self ):
        super(StringEnd,self).__init__()
        self.errmsg = "Expected end of text"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc < len(instring):
            raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc+1, []
        elif loc > len(instring):
            return loc, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class WordStart(_PositionToken):
    """
    Matches if the current position is at the beginning of a Word, and
    is not preceded by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
    use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
    the string being parsed, or at the beginning of a line.
    """
    def __init__(self, wordChars = printables):
        super(WordStart,self).__init__()
        self.wordChars = set(wordChars)
        self.errmsg = "Not at the start of a word"

    def parseImpl(self, instring, loc, doActions=True ):
        if loc != 0:
            if (instring[loc-1] in self.wordChars or
                instring[loc] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class WordEnd(_PositionToken):
    """
    Matches if the current position is at the end of a Word, and
    is not followed by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
    use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
    the string being parsed, or at the end of a line.
    """
    def __init__(self, wordChars = printables):
        super(WordEnd,self).__init__()
        self.wordChars = set(wordChars)
        self.skipWhitespace = False
        self.errmsg = "Not at the end of a word"

    def parseImpl(self, instring, loc, doActions=True ):
        instrlen = len(instring)
        if instrlen>0 and loc<instrlen:
            if (instring[loc] in self.wordChars or
                instring[loc-1] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []


class ParseExpression(ParserElement):
    """
    Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
    """
    def __init__( self, exprs, savelist = False ):
        super(ParseExpression,self).__init__(savelist)
        if isinstance( exprs, _generatorType ):
            exprs = list(exprs)

        if isinstance( exprs, basestring ):
            self.exprs = [ ParserElement._literalStringClass( exprs ) ]
        elif isinstance( exprs, Iterable ):
            exprs = list(exprs)
            # if sequence of strings provided, wrap with Literal
            if all(isinstance(expr, basestring) for expr in exprs):
                exprs = map(ParserElement._literalStringClass, exprs)
            self.exprs = list(exprs)
        else:
            try:
                self.exprs = list( exprs )
            except TypeError:
                self.exprs = [ exprs ]
        self.callPreparse = False

    def __getitem__( self, i ):
        return self.exprs[i]

    def append( self, other ):
        self.exprs.append( other )
        self.strRepr = None
        return self

    def leaveWhitespace( self ):
        """Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
           all contained expressions."""
        self.skipWhitespace = False
        self.exprs = [ e.copy() for e in self.exprs ]
        for e in self.exprs:
            e.leaveWhitespace()
        return self

    def ignore( self, other ):
        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                super( ParseExpression, self).ignore( other )
                for e in self.exprs:
                    e.ignore( self.ignoreExprs[-1] )
        else:
            super( ParseExpression, self).ignore( other )
            for e in self.exprs:
                e.ignore( self.ignoreExprs[-1] )
        return self

    def __str__( self ):
        try:
            return super(ParseExpression,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.exprs) )
        return self.strRepr

    def streamline( self ):
        super(ParseExpression,self).streamline()

        for e in self.exprs:
            e.streamline()

        # collapse nested And's of the form And( And( And( a,b), c), d) to And( a,b,c,d )
        # but only if there are no parse actions or resultsNames on the nested And's
        # (likewise for Or's and MatchFirst's)
        if ( len(self.exprs) == 2 ):
            other = self.exprs[0]
            if ( isinstance( other, self.__class__ ) and
                  not(other.parseAction) and
                  other.resultsName is None and
                  not other.debug ):
                self.exprs = other.exprs[:] + [ self.exprs[1] ]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

            other = self.exprs[-1]
            if ( isinstance( other, self.__class__ ) and
                  not(other.parseAction) and
                  other.resultsName is None and
                  not other.debug ):
                self.exprs = self.exprs[:-1] + other.exprs[:]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

        self.errmsg = "Expected " + _ustr(self)
        
        return self

    def setResultsName( self, name, listAllMatches=False ):
        ret = super(ParseExpression,self).setResultsName(name,listAllMatches)
        return ret

    def validate( self, validateTrace=[] ):
        tmp = validateTrace[:]+[self]
        for e in self.exprs:
            e.validate(tmp)
        self.checkRecursion( [] )
        
    def copy(self):
        ret = super(ParseExpression,self).copy()
        ret.exprs = [e.copy() for e in self.exprs]
        return ret

class And(ParseExpression):
    """
    Requires all given C{ParseExpression}s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'+'} operator.
    May also be constructed using the C{'-'} operator, which will suppress backtracking.

    Example::
        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"),name_expr("name"),integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    """

    class _ErrorStop(Empty):
        def __init__(self, *args, **kwargs):
            super(And._ErrorStop,self).__init__(*args, **kwargs)
            self.name = '-'
            self.leaveWhitespace()

    def __init__( self, exprs, savelist = True ):
        super(And,self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.setWhitespaceChars( self.exprs[0].whiteChars )
        self.skipWhitespace = self.exprs[0].skipWhitespace
        self.callPreparse = True

    def parseImpl( self, instring, loc, doActions=True ):
        # pass False as last arg to _parse for first element, since we already
        # pre-parsed the string as part of our And pre-parsing
        loc, resultlist = self.exprs[0]._parse( instring, loc, doActions, callPreParse=False )
        errorStop = False
        for e in self.exprs[1:]:
            if isinstance(e, And._ErrorStop):
                errorStop = True
                continue
            if errorStop:
                try:
                    loc, exprtokens = e._parse( instring, loc, doActions )
                except ParseSyntaxException:
                    raise
                except ParseBaseException as pe:
                    pe.__traceback__ = None
                    raise ParseSyntaxException._from_exception(pe)
                except IndexError:
                    raise ParseSyntaxException(instring, len(instring), self.errmsg, self)
            else:
                loc, exprtokens = e._parse( instring, loc, doActions )
            if exprtokens or exprtokens.haskeys():
                resultlist += exprtokens
        return loc, resultlist

    def __iadd__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #And( [ self, other ] )

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )
            if not e.mayReturnEmpty:
                break

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr


class Or(ParseExpression):
    """
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the expression that matches the longest string will be used.
    May be constructed using the C{'^'} operator.

    Example::
        # construct Or using '^' operator
        
        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))
    prints::
        [['123'], ['3.1416'], ['789']]
    """
    def __init__( self, exprs, savelist = False ):
        super(Or,self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        maxExcLoc = -1
        maxException = None
        matches = []
        for e in self.exprs:
            try:
                loc2 = e.tryParse( instring, loc )
            except ParseException as err:
                err.__traceback__ = None
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring,len(instring),e.errmsg,self)
                    maxExcLoc = len(instring)
            else:
                # save match among all matches, to retry longest to shortest
                matches.append((loc2, e))

        if matches:
            matches.sort(key=lambda x: -x[0])
            for _,e in matches:
                try:
                    return e._parse( instring, loc, doActions )
                except ParseException as err:
                    err.__traceback__ = None
                    if err.loc > maxExcLoc:
                        maxException = err
                        maxExcLoc = err.loc

        if maxException is not None:
            maxException.msg = self.errmsg
            raise maxException
        else:
            raise ParseException(instring, loc, "no defined alternatives to match", self)


    def __ixor__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #Or( [ self, other ] )

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class MatchFirst(ParseExpression):
    """
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the first one listed is the one that will match.
    May be constructed using the C{'|'} operator.

    Example::
        # construct MatchFirst using '|' operator
        
        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    """
    def __init__( self, exprs, savelist = False ):
        super(MatchFirst,self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        maxExcLoc = -1
        maxException = None
        for e in self.exprs:
            try:
                ret = e._parse( instring, loc, doActions )
                return ret
            except ParseException as err:
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring,len(instring),e.errmsg,self)
                    maxExcLoc = len(instring)

        # only got here if no expression matched, raise exception for match that made it the furthest
        else:
            if maxException is not None:
                maxException.msg = self.errmsg
                raise maxException
            else:
                raise ParseException(instring, loc, "no defined alternatives to match", self)

    def __ior__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #MatchFirst( [ self, other ] )

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class Each(ParseExpression):
    """
    Requires all given C{ParseExpression}s to be found, but in any order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'&'} operator.

    Example::
        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order 
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )
    prints::
        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    """
    def __init__( self, exprs, savelist = True ):
        super(Each,self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.skipWhitespace = True
        self.initExprGroups = True

    def parseImpl( self, instring, loc, doActions=True ):
        if self.initExprGroups:
            self.opt1map = dict((id(e.expr),e) for e in self.exprs if isinstance(e,Optional))
            opt1 = [ e.expr for e in self.exprs if isinstance(e,Optional) ]
            opt2 = [ e for e in self.exprs if e.mayReturnEmpty and not isinstance(e,Optional)]
            self.optionals = opt1 + opt2
            self.multioptionals = [ e.expr for e in self.exprs if isinstance(e,ZeroOrMore) ]
            self.multirequired = [ e.expr for e in self.exprs if isinstance(e,OneOrMore) ]
            self.required = [ e for e in self.exprs if not isinstance(e,(Optional,ZeroOrMore,OneOrMore)) ]
            self.required += self.multirequired
            self.initExprGroups = False
        tmpLoc = loc
        tmpReqd = self.required[:]
        tmpOpt  = self.optionals[:]
        matchOrder = []

        keepMatching = True
        while keepMatching:
            tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired
            failed = []
            for e in tmpExprs:
                try:
                    tmpLoc = e.tryParse( instring, tmpLoc )
                except ParseException:
                    failed.append(e)
                else:
                    matchOrder.append(self.opt1map.get(id(e),e))
                    if e in tmpReqd:
                        tmpReqd.remove(e)
                    elif e in tmpOpt:
                        tmpOpt.remove(e)
            if len(failed) == len(tmpExprs):
                keepMatching = False

        if tmpReqd:
            missing = ", ".join(_ustr(e) for e in tmpReqd)
            raise ParseException(instring,loc,"Missing one or more required elements (%s)" % missing )

        # add any unmatched Optionals, in case they have default values defined
        matchOrder += [e for e in self.exprs if isinstance(e,Optional) and e.expr in tmpOpt]

        resultlist = []
        for e in matchOrder:
            loc,results = e._parse(instring,loc,doActions)
            resultlist.append(results)

        finalResults = sum(resultlist, ParseResults([]))
        return loc, finalResults

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class ParseElementEnhance(ParserElement):
    """
    Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
    """
    def __init__( self, expr, savelist=False ):
        super(ParseElementEnhance,self).__init__(savelist)
        if isinstance( expr, basestring ):
            if issubclass(ParserElement._literalStringClass, Token):
                expr = ParserElement._literalStringClass(expr)
            else:
                expr = ParserElement._literalStringClass(Literal(expr))
        self.expr = expr
        self.strRepr = None
        if expr is not None:
            self.mayIndexError = expr.mayIndexError
            self.mayReturnEmpty = expr.mayReturnEmpty
            self.setWhitespaceChars( expr.whiteChars )
            self.skipWhitespace = expr.skipWhitespace
            self.saveAsList = expr.saveAsList
            self.callPreparse = expr.callPreparse
            self.ignoreExprs.extend(expr.ignoreExprs)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.expr is not None:
            return self.expr._parse( instring, loc, doActions, callPreParse=False )
        else:
            raise ParseException("",loc,self.errmsg,self)

    def leaveWhitespace( self ):
        self.skipWhitespace = False
        self.expr = self.expr.copy()
        if self.expr is not None:
            self.expr.leaveWhitespace()
        return self

    def ignore( self, other ):
        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                super( ParseElementEnhance, self).ignore( other )
                if self.expr is not None:
                    self.expr.ignore( self.ignoreExprs[-1] )
        else:
            super( ParseElementEnhance, self).ignore( other )
            if self.expr is not None:
                self.expr.ignore( self.ignoreExprs[-1] )
        return self

    def streamline( self ):
        super(ParseElementEnhance,self).streamline()
        if self.expr is not None:
            self.expr.streamline()
        return self

    def checkRecursion( self, parseElementList ):
        if self in parseElementList:
            raise RecursiveGrammarException( parseElementList+[self] )
        subRecCheckList = parseElementList[:] + [ self ]
        if self.expr is not None:
            self.expr.checkRecursion( subRecCheckList )

    def validate( self, validateTrace=[] ):
        tmp = validateTrace[:]+[self]
        if self.expr is not None:
            self.expr.validate(tmp)
        self.checkRecursion( [] )

    def __str__( self ):
        try:
            return super(ParseElementEnhance,self).__str__()
        except Exception:
            pass

        if self.strRepr is None and self.expr is not None:
            self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.expr) )
        return self.strRepr


class FollowedBy(ParseElementEnhance):
    """
    Lookahead matching of the given parse expression.  C{FollowedBy}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression matches at the current
    position.  C{FollowedBy} always returns a null token list.

    Example::
        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
    prints::
        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    """
    def __init__( self, expr ):
        super(FollowedBy,self).__init__(expr)
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        self.expr.tryParse( instring, loc )
        return loc, []


class NotAny(ParseElementEnhance):
    """
    Lookahead to disallow matching with the given parse expression.  C{NotAny}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression does I{not} match at the current
    position.  Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
    always returns a null token list.  May be constructed using the '~' operator.

    Example::
        
    """
    def __init__( self, expr ):
        super(NotAny,self).__init__(expr)
        #~ self.leaveWhitespace()
        self.skipWhitespace = False  # do NOT use self.leaveWhitespace(), don't want to propagate to exprs
        self.mayReturnEmpty = True
        self.errmsg = "Found unwanted token, "+_ustr(self.expr)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.expr.canParseNext(instring, loc):
            raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "~{" + _ustr(self.expr) + "}"

        return self.strRepr

class _MultipleMatch(ParseElementEnhance):
    def __init__( self, expr, stopOn=None):
        super(_MultipleMatch, self).__init__(expr)
        self.saveAsList = True
        ender = stopOn
        if isinstance(ender, basestring):
            ender = ParserElement._literalStringClass(ender)
        self.not_ender = ~ender if ender is not None else None

    def parseImpl( self, instring, loc, doActions=True ):
        self_expr_parse = self.expr._parse
        self_skip_ignorables = self._skipIgnorables
        check_ender = self.not_ender is not None
        if check_ender:
            try_not_ender = self.not_ender.tryParse
        
        # must be at least one (but first see if we are the stopOn sentinel;
        # if so, fail)
        if check_ender:
            try_not_ender(instring, loc)
        loc, tokens = self_expr_parse( instring, loc, doActions, callPreParse=False )
        try:
            hasIgnoreExprs = (not not self.ignoreExprs)
            while 1:
                if check_ender:
                    try_not_ender(instring, loc)
                if hasIgnoreExprs:
                    preloc = self_skip_ignorables( instring, loc )
                else:
                    preloc = loc
                loc, tmptokens = self_expr_parse( instring, preloc, doActions )
                if tmptokens or tmptokens.haskeys():
                    tokens += tmptokens
        except (ParseException,IndexError):
            pass

        return loc, tokens
        
class OneOrMore(_MultipleMatch):
    """
    Repetition of one or more of the given expression.
    
    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
        
        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    """

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + _ustr(self.expr) + "}..."

        return self.strRepr

class ZeroOrMore(_MultipleMatch):
    """
    Optional repetition of zero or more of the given expression.
    
    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example: similar to L{OneOrMore}
    """
    def __init__( self, expr, stopOn=None):
        super(ZeroOrMore,self).__init__(expr, stopOn=stopOn)
        self.mayReturnEmpty = True
        
    def parseImpl( self, instring, loc, doActions=True ):
        try:
            return super(ZeroOrMore, self).parseImpl(instring, loc, doActions)
        except (ParseException,IndexError):
            return loc, []

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]..."

        return self.strRepr

class _NullToken(object):
    def __bool__(self):
        return False
    __nonzero__ = __bool__
    def __str__(self):
        return ""

_optionalNotMatched = _NullToken()
class Optional(ParseElementEnhance):
    """
    Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::
        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345
            
            # ZIP+4 form
            12101-0001
            
            # invalid ZIP
            98765-
            ''')
    prints::
        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    """
    def __init__( self, expr, default=_optionalNotMatched ):
        super(Optional,self).__init__( expr, savelist=False )
        self.saveAsList = self.expr.saveAsList
        self.defaultValue = default
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        try:
            loc, tokens = self.expr._parse( instring, loc, doActions, callPreParse=False )
        except (ParseException,IndexError):
            if self.defaultValue is not _optionalNotMatched:
                if self.expr.resultsName:
                    tokens = ParseResults([ self.defaultValue ])
                    tokens[self.expr.resultsName] = self.defaultValue
                else:
                    tokens = [ self.defaultValue ]
            else:
                tokens = []
        return loc, tokens

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]"

        return self.strRepr

class SkipTo(ParseElementEnhance):
    """
    Token for skipping over all undefined text until the matched expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default=C{False}) if True, the target expression is also parsed 
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default=C{None}) used to define grammars (typically quoted strings and 
          comments) that might contain false matches to the target expression
     - failOn - (default=C{None}) define expressions that are not allowed to be 
          included in the skipped test; if found before the target expression is found, 
          the SkipTo is not a match

    Example::
        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP 
                      + string_data("sev") + SEP 
                      + string_data("desc") + SEP 
                      + integer("days_open"))
        
        for tkt in ticket_expr.searchString(report):
            print tkt.dump()
    prints::
        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    """
    def __init__( self, other, include=False, ignore=None, failOn=None ):
        super( SkipTo, self ).__init__( other )
        self.ignoreExpr = ignore
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.includeMatch = include
        self.asList = False
        if isinstance(failOn, basestring):
            self.failOn = ParserElement._literalStringClass(failOn)
        else:
            self.failOn = failOn
        self.errmsg = "No match found for "+_ustr(self.expr)

    def parseImpl( self, instring, loc, doActions=True ):
        startloc = loc
        instrlen = len(instring)
        expr = self.expr
        expr_parse = self.expr._parse
        self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None
        self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None
        
        tmploc = loc
        while tmploc <= instrlen:
            if self_failOn_canParseNext is not None:
                # break if failOn expression matches
                if self_failOn_canParseNext(instring, tmploc):
                    break
                    
            if self_ignoreExpr_tryParse is not None:
                # advance past ignore expressions
                while 1:
                    try:
                        tmploc = self_ignoreExpr_tryParse(instring, tmploc)
                    except ParseBaseException:
                        break
            
            try:
                expr_parse(instring, tmploc, doActions=False, callPreParse=False)
            except (ParseException, IndexError):
                # no match, advance loc in string
                tmploc += 1
            else:
                # matched skipto expr, done
                break

        else:
            # ran off the end of the input string without matching skipto expr, fail
            raise ParseException(instring, loc, self.errmsg, self)

        # build up return values
        loc = tmploc
        skiptext = instring[startloc:loc]
        skipresult = ParseResults(skiptext)
        
        if self.includeMatch:
            loc, mat = expr_parse(instring,loc,doActions,callPreParse=False)
            skipresult += mat

        return loc, skipresult

class Forward(ParseElementEnhance):
    """
    Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.

    Note: take care when assigning to C{Forward} not to overlook precedence of operators.
    Specifically, '|' has a lower precedence than '<<', so that::
        fwdExpr << a | b | c
    will actually be evaluated as::
        (fwdExpr << a) | b | c
    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the C{Forward}::
        fwdExpr << (a | b | c)
    Converting to use the '<<=' operator instead will avoid this problem.

    See L{ParseResults.pprint} for an example of a recursive parser created using
    C{Forward}.
    """
    def __init__( self, other=None ):
        super(Forward,self).__init__( other, savelist=False )

    def __lshift__( self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass(other)
        self.expr = other
        self.strRepr = None
        self.mayIndexError = self.expr.mayIndexError
        self.mayReturnEmpty = self.expr.mayReturnEmpty
        self.setWhitespaceChars( self.expr.whiteChars )
        self.skipWhitespace = self.expr.skipWhitespace
        self.saveAsList = self.expr.saveAsList
        self.ignoreExprs.extend(self.expr.ignoreExprs)
        return self
        
    def __ilshift__(self, other):
        return self << other
    
    def leaveWhitespace( self ):
        self.skipWhitespace = False
        return self

    def streamline( self ):
        if not self.streamlined:
            self.streamlined = True
            if self.expr is not None:
                self.expr.streamline()
        return self

    def validate( self, validateTrace=[] ):
        if self not in validateTrace:
            tmp = validateTrace[:]+[self]
            if self.expr is not None:
                self.expr.validate(tmp)
        self.checkRecursion([])

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name
        return self.__class__.__name__ + ": ..."

        # stubbed out for now - creates awful memory and perf issues
        self._revertClass = self.__class__
        self.__class__ = _ForwardNoRecurse
        try:
            if self.expr is not None:
                retString = _ustr(self.expr)
            else:
                retString = "None"
        finally:
            self.__class__ = self._revertClass
        return self.__class__.__name__ + ": " + retString

    def copy(self):
        if self.expr is not None:
            return super(Forward,self).copy()
        else:
            ret = Forward()
            ret <<= self
            return ret

class _ForwardNoRecurse(Forward):
    def __str__( self ):
        return "..."

class TokenConverter(ParseElementEnhance):
    """
    Abstract subclass of C{ParseExpression}, for converting parsed results.
    """
    def __init__( self, expr, savelist=False ):
        super(TokenConverter,self).__init__( expr )#, savelist )
        self.saveAsList = False

class Combine(TokenConverter):
    """
    Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the input string;
    this can be disabled by specifying C{'adjacent=False'} in the constructor.

    Example::
        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    """
    def __init__( self, expr, joinString="", adjacent=True ):
        super(Combine,self).__init__( expr )
        # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself
        if adjacent:
            self.leaveWhitespace()
        self.adjacent = adjacent
        self.skipWhitespace = True
        self.joinString = joinString
        self.callPreparse = True

    def ignore( self, other ):
        if self.adjacent:
            ParserElement.ignore(self, other)
        else:
            super( Combine, self).ignore( other )
        return self

    def postParse( self, instring, loc, tokenlist ):
        retToks = tokenlist.copy()
        del retToks[:]
        retToks += ParseResults([ "".join(tokenlist._asStringList(self.joinString)) ], modal=self.modalResults)

        if self.resultsName and retToks.haskeys():
            return [ retToks ]
        else:
            return retToks

class Group(TokenConverter):
    """
    Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.

    Example::
        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a,b,100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a,b,100"))  # -> ['fn', ['a', 'b', '100']]
    """
    def __init__( self, expr ):
        super(Group,self).__init__( expr )
        self.saveAsList = True

    def postParse( self, instring, loc, tokenlist ):
        return [ tokenlist ]

class Dict(TokenConverter):
    """
    Converter to return a repetitive expression as a list, but also as a dictionary.
    Each element can also be referenced using the first token in the expression as its key.
    Useful for tabular report scraping when the first column can be used as a item key.

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())
        
        # access named fields as dict entries, or output as dict
        print(result['shape'])        
        print(result.asDict())
    prints::
        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
    See more examples at L{ParseResults} of accessing fields by results name.
    """
    def __init__( self, expr ):
        super(Dict,self).__init__( expr )
        self.saveAsList = True

    def postParse( self, instring, loc, tokenlist ):
        for i,tok in enumerate(tokenlist):
            if len(tok) == 0:
                continue
            ikey = tok[0]
            if isinstance(ikey,int):
                ikey = _ustr(tok[0]).strip()
            if len(tok)==1:
                tokenlist[ikey] = _ParseResultsWithOffset("",i)
            elif len(tok)==2 and not isinstance(tok[1],ParseResults):
                tokenlist[ikey] = _ParseResultsWithOffset(tok[1],i)
            else:
                dictvalue = tok.copy() #ParseResults(i)
                del dictvalue[0]
                if len(dictvalue)!= 1 or (isinstance(dictvalue,ParseResults) and dictvalue.haskeys()):
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue,i)
                else:
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0],i)

        if self.resultsName:
            return [ tokenlist ]
        else:
            return tokenlist


class Suppress(TokenConverter):
    """
    Converter for ignoring the results of a parsed expression.

    Example::
        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))
    prints::
        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']
    (See also L{delimitedList}.)
    """
    def postParse( self, instring, loc, tokenlist ):
        return []

    def suppress( self ):
        return self


class OnlyOnce(object):
    """
    Wrapper for parse actions, to ensure they are only called once.
    """
    def __init__(self, methodCall):
        self.callable = _trim_arity(methodCall)
        self.called = False
    def __call__(self,s,l,t):
        if not self.called:
            results = self.callable(s,l,t)
            self.called = True
            return results
        raise ParseException(s,l,"")
    def reset(self):
        self.called = False

def traceParseAction(f):
    """
    Decorator for debugging parse actions. 
    
    When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
    When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.

    Example::
        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
    prints::
        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    """
    f = _trim_arity(f)
    def z(*paArgs):
        thisFunc = f.__name__
        s,l,t = paArgs[-3:]
        if len(paArgs)>3:
            thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc
        sys.stderr.write( ">>entering %s(line: '%s', %d, %r)\n" % (thisFunc,line(l,s),l,t) )
        try:
            ret = f(*paArgs)
        except Exception as exc:
            sys.stderr.write( "<<leaving %s (exception: %s)\n" % (thisFunc,exc) )
            raise
        sys.stderr.write( "<<leaving %s (ret: %r)\n" % (thisFunc,ret) )
        return ret
    try:
        z.__name__ = f.__name__
    except AttributeError:
        pass
    return z

#
# global helpers
#
def delimitedList( expr, delim=",", combine=False ):
    """
    Helper to define a delimited list of expressions - the delimiter defaults to ','.
    By default, the list elements and delimiters can have intervening whitespace, and
    comments, but this can be overridden by passing C{combine=True} in the constructor.
    If C{combine} is set to C{True}, the matching tokens are returned as a single token
    string, with the delimiters included; otherwise, the matching tokens are returned
    as a list of tokens, with the delimiters suppressed.

    Example::
        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    """
    dlName = _ustr(expr)+" ["+_ustr(delim)+" "+_ustr(expr)+"]..."
    if combine:
        return Combine( expr + ZeroOrMore( delim + expr ) ).setName(dlName)
    else:
        return ( expr + ZeroOrMore( Suppress( delim ) + expr ) ).setName(dlName)

def countedArray( expr, intExpr=None ):
    """
    Helper to define a counted list of expressions.
    This helper defines a pattern of the form::
        integer expr expr expr...
    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
    
    If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.

    Example::
        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    """
    arrayExpr = Forward()
    def countFieldParseAction(s,l,t):
        n = t[0]
        arrayExpr << (n and Group(And([expr]*n)) or Group(empty))
        return []
    if intExpr is None:
        intExpr = Word(nums).setParseAction(lambda t:int(t[0]))
    else:
        intExpr = intExpr.copy()
    intExpr.setName("arrayLen")
    intExpr.addParseAction(countFieldParseAction, callDuringTry=True)
    return ( intExpr + arrayExpr ).setName('(len) ' + _ustr(expr) + '...')

def _flatten(L):
    ret = []
    for i in L:
        if isinstance(i,list):
            ret.extend(_flatten(i))
        else:
            ret.append(i)
    return ret

def matchPreviousLiteral(expr):
    """
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches a
    previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
    If this is not desired, use C{matchPreviousExpr}.
    Do I{not} use with packrat parsing enabled.
    """
    rep = Forward()
    def copyTokenToRepeater(s,l,t):
        if t:
            if len(t) == 1:
                rep << t[0]
            else:
                # flatten t tokens
                tflat = _flatten(t.asList())
                rep << And(Literal(tt) for tt in tflat)
        else:
            rep << Empty()
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def matchPreviousExpr(expr):
    """
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches by
    expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
    the expressions are evaluated first, and then compared, so
    C{"1"} is compared with C{"10"}.
    Do I{not} use with packrat parsing enabled.
    """
    rep = Forward()
    e2 = expr.copy()
    rep <<= e2
    def copyTokenToRepeater(s,l,t):
        matchTokens = _flatten(t.asList())
        def mustMatchTheseTokens(s,l,t):
            theseTokens = _flatten(t.asList())
            if  theseTokens != matchTokens:
                raise ParseException("",0,"")
        rep.setParseAction( mustMatchTheseTokens, callDuringTry=True )
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def _escapeRegexRangeChars(s):
    #~  escape these chars: ^-]
    for c in r"\^-]":
        s = s.replace(c,_bslash+c)
    s = s.replace("\n",r"\n")
    s = s.replace("\t",r"\t")
    return _ustr(s)

def oneOf( strs, caseless=False, useRegex=True ):
    """
    Helper to quickly define a set of alternative Literals, and makes sure to do
    longest-first testing when there is a conflict, regardless of the input order,
    but returns a C{L{MatchFirst}} for best performance.

    Parameters:
     - strs - a string of space-delimited literals, or a collection of string literals
     - caseless - (default=C{False}) - treat all literals as caseless
     - useRegex - (default=C{True}) - as an optimization, will generate a Regex
          object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
          if creating a C{Regex} raises an exception)

    Example::
        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))
    prints::
        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    """
    if caseless:
        isequal = ( lambda a,b: a.upper() == b.upper() )
        masks = ( lambda a,b: b.upper().startswith(a.upper()) )
        parseElementClass = CaselessLiteral
    else:
        isequal = ( lambda a,b: a == b )
        masks = ( lambda a,b: b.startswith(a) )
        parseElementClass = Literal

    symbols = []
    if isinstance(strs,basestring):
        symbols = strs.split()
    elif isinstance(strs, Iterable):
        symbols = list(strs)
    else:
        warnings.warn("Invalid argument to oneOf, expected string or iterable",
                SyntaxWarning, stacklevel=2)
    if not symbols:
        return NoMatch()

    i = 0
    while i < len(symbols)-1:
        cur = symbols[i]
        for j,other in enumerate(symbols[i+1:]):
            if ( isequal(other, cur) ):
                del symbols[i+j+1]
                break
            elif ( masks(cur, other) ):
                del symbols[i+j+1]
                symbols.insert(i,other)
                cur = other
                break
        else:
            i += 1

    if not caseless and useRegex:
        #~ print (strs,"->", "|".join( [ _escapeRegexChars(sym) for sym in symbols] ))
        try:
            if len(symbols)==len("".join(symbols)):
                return Regex( "[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols) ).setName(' | '.join(symbols))
            else:
                return Regex( "|".join(re.escape(sym) for sym in symbols) ).setName(' | '.join(symbols))
        except Exception:
            warnings.warn("Exception creating Regex for oneOf, building MatchFirst",
                    SyntaxWarning, stacklevel=2)


    # last resort, just use MatchFirst
    return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols))

def dictOf( key, value ):
    """
    Helper to easily and clearly define a dictionary by specifying the respective patterns
    for the key and value.  Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
    in the proper order.  The key pattern can include delimiting markers or punctuation,
    as long as they are suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the C{Dict} results can include named token
    fields.

    Example::
        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())
    prints::
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    """
    return Dict( ZeroOrMore( Group ( key + value ) ) )

def originalTextFor(expr, asString=True):
    """
    Helper to return the original, untokenized text for a given expression.  Useful to
    restore the parsed fields of an HTML start tag into the raw tag text itself, or to
    revert separate tokens with intervening whitespace back to the original matching
    input text. By default, returns astring containing the original parsed text.  
       
    If the optional C{asString} argument is passed as C{False}, then the return value is a 
    C{L{ParseResults}} containing any results names that were originally matched, and a 
    single token containing the original matched text from the input string.  So if 
    the expression passed to C{L{originalTextFor}} contains expressions with defined
    results names, you must set C{asString} to C{False} if you want to preserve those
    results name values.

    Example::
        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b","i"):
            opener,closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])
    prints::
        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    """
    locMarker = Empty().setParseAction(lambda s,loc,t: loc)
    endlocMarker = locMarker.copy()
    endlocMarker.callPreparse = False
    matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end")
    if asString:
        extractText = lambda s,l,t: s[t._original_start:t._original_end]
    else:
        def extractText(s,l,t):
            t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]]
    matchExpr.setParseAction(extractText)
    matchExpr.ignoreExprs = expr.ignoreExprs
    return matchExpr

def ungroup(expr): 
    """
    Helper to undo pyparsing's default grouping of And expressions, even
    if all but one are non-empty.
    """
    return TokenConverter(expr).setParseAction(lambda t:t[0])

def locatedExpr(expr):
    """
    Helper to decorate a returned token with its starting and ending locations in the input string.
    This helper adds the following results names:
     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains C{<TAB>} characters, you may want to call
    C{L{ParserElement.parseWithTabs}}

    Example::
        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)
    prints::
        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    """
    locator = Empty().setParseAction(lambda s,l,t: l)
    return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end"))


# convenience constants for positional expressions
empty       = Empty().setName("empty")
lineStart   = LineStart().setName("lineStart")
lineEnd     = LineEnd().setName("lineEnd")
stringStart = StringStart().setName("stringStart")
stringEnd   = StringEnd().setName("stringEnd")

_escapedPunc = Word( _bslash, r"\[]-*.$+^?()~ ", exact=2 ).setParseAction(lambda s,l,t:t[0][1])
_escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s,l,t:unichr(int(t[0].lstrip(r'\0x'),16)))
_escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s,l,t:unichr(int(t[0][1:],8)))
_singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | CharsNotIn(r'\]', exact=1)
_charRange = Group(_singleChar + Suppress("-") + _singleChar)
_reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group( OneOrMore( _charRange | _singleChar ) ).setResultsName("body") + "]"

def srange(s):
    r"""
    Helper to easily define string ranges for use in Word construction.  Borrows
    syntax from regexp '[]' string range definitions::
        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
    The input string must be enclosed in []'s, and the returned string is the expanded
    character set joined into a single string.
    The values enclosed in the []'s may be:
     - a single character
     - an escaped character with a leading backslash (such as C{\-} or C{\]})
     - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) 
         (C{\0x##} is also supported for backwards compatibility) 
     - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
     - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
     - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
    """
    _expanded = lambda p: p if not isinstance(p,ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]),ord(p[1])+1))
    try:
        return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body)
    except Exception:
        return ""

def matchOnlyAtCol(n):
    """
    Helper method for defining parse actions that require matching at a specific
    column in the input text.
    """
    def verifyCol(strg,locn,toks):
        if col(locn,strg) != n:
            raise ParseException(strg,locn,"matched token not at column %d" % n)
    return verifyCol

def replaceWith(replStr):
    """
    Helper method for common parse actions that simply return a literal value.  Especially
    useful when used with C{L{transformString<ParserElement.transformString>}()}.

    Example::
        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num
        
        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    """
    return lambda s,l,t: [replStr]

def removeQuotes(s,l,t):
    """
    Helper parse action for removing quotation marks from parsed quoted strings.

    Example::
        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    """
    return t[0][1:-1]

def tokenMap(func, *args):
    """
    Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional 
    args are passed, they are forwarded to the given function as additional arguments after
    the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
    parsed data to an integer using base 16.

    Example (compare the last to example in L{ParserElement.transformString}::
        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')
        
        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')
    prints::
        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    """
    def pa(s,l,t):
        return [func(tokn, *args) for tokn in t]

    try:
        func_name = getattr(func, '__name__', 
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    pa.__name__ = func_name

    return pa

upcaseTokens = tokenMap(lambda t: _ustr(t).upper())
"""(Deprecated) Helper parse action to convert tokens to upper case. Deprecated in favor of L{pyparsing_common.upcaseTokens}"""

downcaseTokens = tokenMap(lambda t: _ustr(t).lower())
"""(Deprecated) Helper parse action to convert tokens to lower case. Deprecated in favor of L{pyparsing_common.downcaseTokens}"""
    
def _makeTags(tagStr, xml):
    """Internal helper to construct opening and closing tag expressions, given a tag name"""
    if isinstance(tagStr,basestring):
        resname = tagStr
        tagStr = Keyword(tagStr, caseless=not xml)
    else:
        resname = tagStr.name

    tagAttrName = Word(alphas,alphanums+"_-:")
    if (xml):
        tagAttrValue = dblQuotedString.copy().setParseAction( removeQuotes )
        openTag = Suppress("<") + tagStr("tag") + \
                Dict(ZeroOrMore(Group( tagAttrName + Suppress("=") + tagAttrValue ))) + \
                Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
    else:
        printablesLessRAbrack = "".join(c for c in printables if c not in ">")
        tagAttrValue = quotedString.copy().setParseAction( removeQuotes ) | Word(printablesLessRAbrack)
        openTag = Suppress("<") + tagStr("tag") + \
                Dict(ZeroOrMore(Group( tagAttrName.setParseAction(downcaseTokens) + \
                Optional( Suppress("=") + tagAttrValue ) ))) + \
                Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
    closeTag = Combine(_L("</") + tagStr + ">")

    openTag = openTag.setResultsName("start"+"".join(resname.replace(":"," ").title().split())).setName("<%s>" % resname)
    closeTag = closeTag.setResultsName("end"+"".join(resname.replace(":"," ").title().split())).setName("</%s>" % resname)
    openTag.tag = resname
    closeTag.tag = resname
    return openTag, closeTag

def makeHTMLTags(tagStr):
    """
    Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
    tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.

    Example::
        text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
        a,a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end
        
        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are also accessible as named results
            print(link.link_text, '->', link.href)
    prints::
        pyparsing -> http://pyparsing.wikispaces.com
    """
    return _makeTags( tagStr, False )

def makeXMLTags(tagStr):
    """
    Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
    tags only in the given upper/lower case.

    Example: similar to L{makeHTMLTags}
    """
    return _makeTags( tagStr, True )

def withAttribute(*args,**attrDict):
    """
    Helper to create a validating parse action to be used with start tags created
    with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
    with a required attribute value, to avoid false matches on common tags such as
    C{<TD>} or C{<DIV>}.

    Call C{withAttribute} with a series of attribute names and values. Specify the list
    of filter attributes names and values as:
     - keyword arguments, as in C{(align="right")}, or
     - as an explicit dict with C{**} operator, when an attribute name is also a Python
          reserved word, as in C{**{"class":"Customer", "align":"right"}}
     - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") )
    For attribute names with a namespace prefix, you must use the second form.  Attribute
    names are matched insensitive to upper/lower case.
       
    If just testing for C{class} (with or without a namespace), use C{L{withClass}}.

    To verify that the attribute exists, but without specifying a value, pass
    C{withAttribute.ANY_VALUE} as the value.

    Example::
        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    if args:
        attrs = args[:]
    else:
        attrs = attrDict.items()
    attrs = [(k,v) for k,v in attrs]
    def pa(s,l,tokens):
        for attrName,attrValue in attrs:
            if attrName not in tokens:
                raise ParseException(s,l,"no matching attribute " + attrName)
            if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue:
                raise ParseException(s,l,"attribute '%s' has value '%s', must be '%s'" %
                                            (attrName, tokens[attrName], attrValue))
    return pa
withAttribute.ANY_VALUE = object()

def withClass(classname, namespace=''):
    """
    Simplified version of C{L{withAttribute}} when matching on a div class - made
    difficult because C{class} is a reserved word in Python.

    Example::
        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))
        
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    classattr = "%s:class" % namespace if namespace else "class"
    return withAttribute(**{classattr : classname})        

opAssoc = _Constants()
opAssoc.LEFT = object()
opAssoc.RIGHT = object()

def infixNotation( baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')') ):
    """
    Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary or
    binary, left- or right-associative.  Parse actions can also be attached
    to operator expressions. The generated parser will also recognize the use 
    of parentheses to override operator precedences (see example below).
    
    Note: if you define a deep operator list, you may see performance issues
    when using infixNotation. See L{ParserElement.enablePackrat} for a
    mechanism to potentially improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the nested
     - opList - list of tuples, one for each operator precedence level in the
      expression grammar; each tuple is of the form
      (opExpr, numTerms, rightLeftAssoc, parseAction), where:
       - opExpr is the pyparsing expression for the operator;
          may also be a string, which will be converted to a Literal;
          if numTerms is 3, opExpr is a tuple of two expressions, for the
          two operators separating the 3 terms
       - numTerms is the number of terms for this operator (must
          be 1, 2, or 3)
       - rightLeftAssoc is the indicator whether the operator is
          right or left associative, using the pyparsing-defined
          constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}.
       - parseAction is the parse action to be associated with
          expressions matching this operator expression (the
          parse action tuple member may be omitted); if the parse action
          is passed a tuple or list of functions, this is equivalent to
          calling C{setParseAction(*fn)} (L{ParserElement.setParseAction})
     - lpar - expression for matching left-parentheses (default=C{Suppress('(')})
     - rpar - expression for matching right-parentheses (default=C{Suppress(')')})

    Example::
        # simple example of four-function arithmetic with ints and variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier 
        
        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])
        
        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)
    prints::
        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    """
    ret = Forward()
    lastExpr = baseExpr | ( lpar + ret + rpar )
    for i,operDef in enumerate(opList):
        opExpr,arity,rightLeftAssoc,pa = (operDef + (None,))[:4]
        termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr
        if arity == 3:
            if opExpr is None or len(opExpr) != 2:
                raise ValueError("if numterms=3, opExpr must be a tuple or list of two expressions")
            opExpr1, opExpr2 = opExpr
        thisExpr = Forward().setName(termName)
        if rightLeftAssoc == opAssoc.LEFT:
            if arity == 1:
                matchExpr = FollowedBy(lastExpr + opExpr) + Group( lastExpr + OneOrMore( opExpr ) )
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = FollowedBy(lastExpr + opExpr + lastExpr) + Group( lastExpr + OneOrMore( opExpr + lastExpr ) )
                else:
                    matchExpr = FollowedBy(lastExpr+lastExpr) + Group( lastExpr + OneOrMore(lastExpr) )
            elif arity == 3:
                matchExpr = FollowedBy(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr) + \
                            Group( lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr )
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        elif rightLeftAssoc == opAssoc.RIGHT:
            if arity == 1:
                # try to avoid LR with this extra test
                if not isinstance(opExpr, Optional):
                    opExpr = Optional(opExpr)
                matchExpr = FollowedBy(opExpr.expr + thisExpr) + Group( opExpr + thisExpr )
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = FollowedBy(lastExpr + opExpr + thisExpr) + Group( lastExpr + OneOrMore( opExpr + thisExpr ) )
                else:
                    matchExpr = FollowedBy(lastExpr + thisExpr) + Group( lastExpr + OneOrMore( thisExpr ) )
            elif arity == 3:
                matchExpr = FollowedBy(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr) + \
                            Group( lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr )
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        else:
            raise ValueError("operator must indicate right or left associativity")
        if pa:
            if isinstance(pa, (tuple, list)):
                matchExpr.setParseAction(*pa)
            else:
                matchExpr.setParseAction(pa)
        thisExpr <<= ( matchExpr.setName(termName) | lastExpr )
        lastExpr = thisExpr
    ret <<= lastExpr
    return ret

operatorPrecedence = infixNotation
"""(Deprecated) Former name of C{L{infixNotation}}, will be dropped in a future release."""

dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"').setName("string enclosed in double quotes")
sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("string enclosed in single quotes")
quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"'|
                       Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("quotedString using single or double quotes")
unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal")

def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()):
    """
    Helper method for defining nested lists enclosed in opening and closing
    delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression
     - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression
     - content - expression for items within the nested lists (default=C{None})
     - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString})

    If an expression is not provided for the content argument, the nested
    expression will capture all whitespace-delimited content between delimiters
    as a list of separate values.

    Use the C{ignoreExpr} argument to define expressions that may contain
    opening or closing characters that should not be treated as opening
    or closing characters for nesting, such as quotedString or a comment
    expression.  Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}.
    The default is L{quotedString}, but if no expressions are to be ignored,
    then pass C{None} for this argument.

    Example::
        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR,RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type") 
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR 
                      + code_body("body"))
        c_function.ignore(cStyleComment)
        
        source_code = '''
            int is_odd(int x) { 
                return (x%2); 
            }
                
            int dec_to_hex(char hchar) { 
                if (hchar >= '0' && hchar <= '9') { 
                    return (ord(hchar)-ord('0')); 
                } else { 
                    return (10+ord(hchar)-ord('A'));
                } 
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)

    prints::
        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    """
    if opener == closer:
        raise ValueError("opening and closing strings cannot be the same")
    if content is None:
        if isinstance(opener,basestring) and isinstance(closer,basestring):
            if len(opener) == 1 and len(closer)==1:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr +
                                    CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
                else:
                    content = (empty.copy()+CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS
                                ).setParseAction(lambda t:t[0].strip()))
            else:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr + 
                                    ~Literal(opener) + ~Literal(closer) +
                                    CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
                else:
                    content = (Combine(OneOrMore(~Literal(opener) + ~Literal(closer) +
                                    CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
        else:
            raise ValueError("opening and closing arguments must be strings if no content expression is given")
    ret = Forward()
    if ignoreExpr is not None:
        ret <<= Group( Suppress(opener) + ZeroOrMore( ignoreExpr | ret | content ) + Suppress(closer) )
    else:
        ret <<= Group( Suppress(opener) + ZeroOrMore( ret | content )  + Suppress(closer) )
    ret.setName('nested %s%s expression' % (opener,closer))
    return ret

def indentedBlock(blockStatementExpr, indentStack, indent=True):
    """
    Helper method for defining space-delimited indentation blocks, such as
    those used to define block statements in Python source code.

    Parameters:
     - blockStatementExpr - expression defining syntax of statement that
            is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
            (multiple statementWithIndentedBlock expressions within a single grammar
            should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond the
            the current level; set to False for block of left-most statements
            (default=C{True})

    A valid block must contain at least one C{blockStatement}.

    Example::
        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group( funcDecl + func_body )

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << ( funcDef | assignment | identifier )

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()
    prints::
        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] 
    """
    def checkPeerIndent(s,l,t):
        if l >= len(s): return
        curCol = col(l,s)
        if curCol != indentStack[-1]:
            if curCol > indentStack[-1]:
                raise ParseFatalException(s,l,"illegal nesting")
            raise ParseException(s,l,"not a peer entry")

    def checkSubIndent(s,l,t):
        curCol = col(l,s)
        if curCol > indentStack[-1]:
            indentStack.append( curCol )
        else:
            raise ParseException(s,l,"not a subentry")

    def checkUnindent(s,l,t):
        if l >= len(s): return
        curCol = col(l,s)
        if not(indentStack and curCol < indentStack[-1] and curCol <= indentStack[-2]):
            raise ParseException(s,l,"not an unindent")
        indentStack.pop()

    NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress())
    INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT')
    PEER   = Empty().setParseAction(checkPeerIndent).setName('')
    UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT')
    if indent:
        smExpr = Group( Optional(NL) +
            #~ FollowedBy(blockStatementExpr) +
            INDENT + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) + UNDENT)
    else:
        smExpr = Group( Optional(NL) +
            (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) )
    blockStatementExpr.ignore(_bslash + LineEnd())
    return smExpr.setName('indented block')

alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]")
punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]")

anyOpenTag,anyCloseTag = makeHTMLTags(Word(alphas,alphanums+"_:").setName('any tag'))
_htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(),'><& "\''))
commonHTMLEntity = Regex('&(?P<entity>' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity")
def replaceHTMLEntity(t):
    """Helper parser action to replace common HTML entities with their special characters"""
    return _htmlEntityMap.get(t.entity)

# it's easy to get these comment structures wrong - they're very common, so may as well make them available
cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment")
"Comment of the form C{/* ... */}"

htmlComment = Regex(r"<!--[\s\S]*?-->").setName("HTML comment")
"Comment of the form C{<!-- ... -->}"

restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line")
dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment")
"Comment of the form C{// ... (to end of line)}"

cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/'| dblSlashComment).setName("C++ style comment")
"Comment of either form C{L{cStyleComment}} or C{L{dblSlashComment}}"

javaStyleComment = cppStyleComment
"Same as C{L{cppStyleComment}}"

pythonStyleComment = Regex(r"#.*").setName("Python style comment")
"Comment of the form C{# ... (to end of line)}"

_commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',') +
                                  Optional( Word(" \t") +
                                            ~Literal(",") + ~LineEnd() ) ) ).streamline().setName("commaItem")
commaSeparatedList = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("commaSeparatedList")
"""(Deprecated) Predefined expression of 1 or more printable words or quoted strings, separated by commas.
   This expression is deprecated in favor of L{pyparsing_common.comma_separated_list}."""

# some other useful expressions - using lower-case class name since we are really using this as a namespace
class pyparsing_common:
    """
    Here are some common low-level expressions that may be useful in jump-starting parser development:
     - numeric forms (L{integers<integer>}, L{reals<real>}, L{scientific notation<sci_real>})
     - common L{programming identifiers<identifier>}
     - network addresses (L{MAC<mac_address>}, L{IPv4<ipv4_address>}, L{IPv6<ipv6_address>})
     - ISO8601 L{dates<iso8601_date>} and L{datetime<iso8601_datetime>}
     - L{UUID<uuid>}
     - L{comma-separated list<comma_separated_list>}
    Parse actions:
     - C{L{convertToInteger}}
     - C{L{convertToFloat}}
     - C{L{convertToDate}}
     - C{L{convertToDatetime}}
     - C{L{stripHTMLTags}}
     - C{L{upcaseTokens}}
     - C{L{downcaseTokens}}

    Example::
        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')
    prints::
        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    """

    convertToInteger = tokenMap(int)
    """
    Parse action for converting parsed integers to Python int
    """

    convertToFloat = tokenMap(float)
    """
    Parse action for converting parsed numbers to Python float
    """

    integer = Word(nums).setName("integer").setParseAction(convertToInteger)
    """expression that parses an unsigned integer, returns an int"""

    hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int,16))
    """expression that parses a hexadecimal integer, returns an int"""

    signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger)
    """expression that parses an integer with optional leading sign, returns an int"""

    fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction")
    """fractional expression of an integer divided by an integer, returns a float"""
    fraction.addParseAction(lambda t: t[0]/t[-1])

    mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction")
    """mixed integer of the form 'integer - fraction', with optional leading integer, returns float"""
    mixed_integer.addParseAction(sum)

    real = Regex(r'[+-]?\d+\.\d*').setName("real number").setParseAction(convertToFloat)
    """expression that parses a floating point number and returns a float"""

    sci_real = Regex(r'[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat)
    """expression that parses a floating point number with optional scientific notation and returns a float"""

    # streamlining this expression makes the docs nicer-looking
    number = (sci_real | real | signed_integer).streamline()
    """any numeric expression, returns the corresponding Python type"""

    fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat)
    """any int or real number, returned as float"""
    
    identifier = Word(alphas+'_', alphanums+'_').setName("identifier")
    """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')"""
    
    ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address")
    "IPv4 address (C{0.0.0.0 - 255.255.255.255})"

    _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer")
    _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part)*7).setName("full IPv6 address")
    _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part)*(0,6)) + "::" + Optional(_ipv6_part + (':' + _ipv6_part)*(0,6))).setName("short IPv6 address")
    _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8)
    _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address")
    ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address")
    "IPv6 address (long, short, or mixed form)"
    
    mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address")
    "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)"

    @staticmethod
    def convertToDate(fmt="%Y-%m-%d"):
        """
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"})

        Example::
            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))
        prints::
            [datetime.date(1999, 12, 31)]
        """
        def cvt_fn(s,l,t):
            try:
                return datetime.strptime(t[0], fmt).date()
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    @staticmethod
    def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"):
        """
        Helper to create a parse action for converting parsed datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"})

        Example::
            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))
        prints::
            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        """
        def cvt_fn(s,l,t):
            try:
                return datetime.strptime(t[0], fmt)
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    iso8601_date = Regex(r'(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?').setName("ISO8601 date")
    "ISO8601 date (C{yyyy-mm-dd})"

    iso8601_datetime = Regex(r'(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime")
    "ISO8601 datetime (C{yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)}) - trailing seconds, milliseconds, and timezone optional; accepts separating C{'T'} or C{' '}"

    uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID")
    "UUID (C{xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx})"

    _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress()
    @staticmethod
    def stripHTMLTags(s, l, tokens):
        """
        Parse action to remove HTML tags from web page HTML source

        Example::
            # strip HTML links from normal text 
            text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
            td,td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            
            print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page'
        """
        return pyparsing_common._html_stripper.transformString(tokens[0])

    _commasepitem = Combine(OneOrMore(~Literal(",") + ~LineEnd() + Word(printables, excludeChars=',') 
                                        + Optional( White(" \t") ) ) ).streamline().setName("commaItem")
    comma_separated_list = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("comma separated list")
    """Predefined expression of 1 or more printable words or quoted strings, separated by commas."""

    upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper()))
    """Parse action to convert tokens to upper case."""

    downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower()))
    """Parse action to convert tokens to lower case."""


if __name__ == "__main__":

    selectToken    = CaselessLiteral("select")
    fromToken      = CaselessLiteral("from")

    ident          = Word(alphas, alphanums + "_$")

    columnName     = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    columnNameList = Group(delimitedList(columnName)).setName("columns")
    columnSpec     = ('*' | columnNameList)

    tableName      = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    tableNameList  = Group(delimitedList(tableName)).setName("tables")
    
    simpleSQL      = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables")

    # demo runTests method, including embedded comments in test string
    simpleSQL.runTests("""
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        """)

    pyparsing_common.number.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    # any int or real number, returned as float
    pyparsing_common.fnumber.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    pyparsing_common.hex_integer.runTests("""
        100
        FF
        """)

    import uuid
    pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
    pyparsing_common.uuid.runTests("""
        12345678-1234-5678-1234-567812345678
        """)
site-packages/setuptools/_vendor/__init__.py000064400000000000150564745010015247 0ustar00site-packages/setuptools/extern/__pycache__/__init__.cpython-39.pyc000064400000005627150564745010021433 0ustar00a

(�Reg	�@s6ddlZddlZGdd�d�ZdZeeed���dS)�Nc@sXeZdZdZddd�Zedd��Zdd	�Zd
d�Zdd
�Z	dd�Z
ddd�Zdd�ZdS)�VendorImporterz�
    A PEP 302 meta path importer for finding optionally-vendored
    or otherwise naturally-installed packages from root_name.
    �NcCs&||_t|�|_|p|�dd�|_dS)NZextern�_vendor)�	root_name�set�vendored_names�replace�
vendor_pkg)�selfrrr	rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/extern/__init__.py�__init__s
zVendorImporter.__init__ccs|jdVdVdS)zL
        Search first the vendor package then as a natural package.
        �.�N)r	�r
rrr�search_pathszVendorImporter.search_pathcCs.|�|jd�\}}}|o,tt|j|j��S)z,Figure out if the target module is vendored.r
)�	partitionr�any�map�
startswithr)r
�fullname�root�base�targetrrr�_module_matches_namespacesz(VendorImporter._module_matches_namespacec	Cs~|�|jd�\}}}|jD]F}z.||}t|�tj|}|tj|<|WSty`Yq0qtdjfit�����dS)zK
        Iterate over the search path to locate and load fullname.
        r
z�The '{target}' package is required; normally this is bundled with this package so if you get this warning, consult the packager of your distribution.N)	rrr�
__import__�sys�modules�ImportError�format�locals)r
rrrr�prefixZextant�modrrr�load_modules



��zVendorImporter.load_modulecCs|�|j�S�N)r"�name)r
�specrrr�
create_module3szVendorImporter.create_modulecCsdSr#r)r
�modulerrr�exec_module6szVendorImporter.exec_modulecCs|�|�rtj�||�SdS)z(Return a module spec for vendored names.N)r�	importlib�util�spec_from_loader)r
r�pathrrrr�	find_spec9s��zVendorImporter.find_speccCs|tjvrtj�|�dS)zR
        Install this importer into sys.meta_path if not already present.
        N)r�	meta_path�appendrrrr�install@s
zVendorImporter.install)rN)NN)
�__name__�
__module__�__qualname__�__doc__r�propertyrrr"r&r(r-r0rrrrrs


r)�	packaging�	pyparsingZordered_setZmore_itertoolszsetuptools._vendor)�importlib.utilr)rr�namesr1r0rrrr�<module>sCsite-packages/setuptools/extern/__init__.py000064400000004547150564745010015144 0ustar00import importlib.util
import sys


class VendorImporter:
    """
    A PEP 302 meta path importer for finding optionally-vendored
    or otherwise naturally-installed packages from root_name.
    """

    def __init__(self, root_name, vendored_names=(), vendor_pkg=None):
        self.root_name = root_name
        self.vendored_names = set(vendored_names)
        self.vendor_pkg = vendor_pkg or root_name.replace('extern', '_vendor')

    @property
    def search_path(self):
        """
        Search first the vendor package then as a natural package.
        """
        yield self.vendor_pkg + '.'
        yield ''

    def _module_matches_namespace(self, fullname):
        """Figure out if the target module is vendored."""
        root, base, target = fullname.partition(self.root_name + '.')
        return not root and any(map(target.startswith, self.vendored_names))

    def load_module(self, fullname):
        """
        Iterate over the search path to locate and load fullname.
        """
        root, base, target = fullname.partition(self.root_name + '.')
        for prefix in self.search_path:
            try:
                extant = prefix + target
                __import__(extant)
                mod = sys.modules[extant]
                sys.modules[fullname] = mod
                return mod
            except ImportError:
                pass
        else:
            raise ImportError(
                "The '{target}' package is required; "
                "normally this is bundled with this package so if you get "
                "this warning, consult the packager of your "
                "distribution.".format(**locals())
            )

    def create_module(self, spec):
        return self.load_module(spec.name)

    def exec_module(self, module):
        pass

    def find_spec(self, fullname, path=None, target=None):
        """Return a module spec for vendored names."""
        return (
            importlib.util.spec_from_loader(fullname, self)
            if self._module_matches_namespace(fullname) else None
        )

    def install(self):
        """
        Install this importer into sys.meta_path if not already present.
        """
        if self not in sys.meta_path:
            sys.meta_path.append(self)


names = 'packaging', 'pyparsing', 'ordered_set', 'more_itertools',
VendorImporter(__name__, names, 'setuptools._vendor').install()
site-packages/setuptools/command/easy_install.py000064400000247234150564745010016207 0ustar00"""
Easy Install
------------

A tool for doing automatic download/extract/build of distutils-based Python
packages.  For detailed documentation, see the accompanying EasyInstall.txt
file, or visit the `EasyInstall home page`__.

__ https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html

"""

from glob import glob
from distutils.util import get_platform
from distutils.util import convert_path, subst_vars
from distutils.errors import (
    DistutilsArgError, DistutilsOptionError,
    DistutilsError, DistutilsPlatformError,
)
from distutils.command.install import INSTALL_SCHEMES, SCHEME_KEYS
from distutils import log, dir_util
from distutils.command.build_scripts import first_line_re
from distutils.spawn import find_executable
import sys
import os
import zipimport
import shutil
import tempfile
import zipfile
import re
import stat
import random
import textwrap
import warnings
import site
import struct
import contextlib
import subprocess
import shlex
import io
import configparser


from sysconfig import get_config_vars, get_path

from setuptools import SetuptoolsDeprecationWarning

from setuptools import Command
from setuptools.sandbox import run_setup
from setuptools.command import setopt
from setuptools.archive_util import unpack_archive
from setuptools.package_index import (
    PackageIndex, parse_requirement_arg, URL_SCHEME,
)
from setuptools.command import bdist_egg, egg_info
from setuptools.wheel import Wheel
from pkg_resources import (
    yield_lines, normalize_path, resource_string, ensure_directory,
    get_distribution, find_distributions, Environment, Requirement,
    Distribution, PathMetadata, EggMetadata, WorkingSet, DistributionNotFound,
    VersionConflict, DEVELOP_DIST,
)
import pkg_resources

# Turn on PEP440Warnings
warnings.filterwarnings("default", category=pkg_resources.PEP440Warning)

__all__ = [
    'samefile', 'easy_install', 'PthDistributions', 'extract_wininst_cfg',
    'get_exe_prefixes',
]


def is_64bit():
    return struct.calcsize("P") == 8


def samefile(p1, p2):
    """
    Determine if two paths reference the same file.

    Augments os.path.samefile to work on Windows and
    suppresses errors if the path doesn't exist.
    """
    both_exist = os.path.exists(p1) and os.path.exists(p2)
    use_samefile = hasattr(os.path, 'samefile') and both_exist
    if use_samefile:
        return os.path.samefile(p1, p2)
    norm_p1 = os.path.normpath(os.path.normcase(p1))
    norm_p2 = os.path.normpath(os.path.normcase(p2))
    return norm_p1 == norm_p2


def _to_bytes(s):
    return s.encode('utf8')


def isascii(s):
    try:
        s.encode('ascii')
        return True
    except UnicodeError:
        return False


def _one_liner(text):
    return textwrap.dedent(text).strip().replace('\n', '; ')


class easy_install(Command):
    """Manage a download/build/install process"""
    description = "Find/get/install Python packages"
    command_consumes_arguments = True

    user_options = [
        ('prefix=', None, "installation prefix"),
        ("zip-ok", "z", "install package as a zipfile"),
        ("multi-version", "m", "make apps have to require() a version"),
        ("upgrade", "U", "force upgrade (searches PyPI for latest versions)"),
        ("install-dir=", "d", "install package to DIR"),
        ("script-dir=", "s", "install scripts to DIR"),
        ("exclude-scripts", "x", "Don't install scripts"),
        ("always-copy", "a", "Copy all needed packages to install dir"),
        ("index-url=", "i", "base URL of Python Package Index"),
        ("find-links=", "f", "additional URL(s) to search for packages"),
        ("build-directory=", "b",
         "download/extract/build in DIR; keep the results"),
        ('optimize=', 'O',
         "also compile with optimization: -O1 for \"python -O\", "
         "-O2 for \"python -OO\", and -O0 to disable [default: -O0]"),
        ('record=', None,
         "filename in which to record list of installed files"),
        ('always-unzip', 'Z', "don't install as a zipfile, no matter what"),
        ('site-dirs=', 'S', "list of directories where .pth files work"),
        ('editable', 'e', "Install specified packages in editable form"),
        ('no-deps', 'N', "don't install dependencies"),
        ('allow-hosts=', 'H', "pattern(s) that hostnames must match"),
        ('local-snapshots-ok', 'l',
         "allow building eggs from local checkouts"),
        ('version', None, "print version information and exit"),
        ('no-find-links', None,
         "Don't load find-links defined in packages being installed"),
        ('user', None, "install in user site-package '%s'" % site.USER_SITE)
    ]
    boolean_options = [
        'zip-ok', 'multi-version', 'exclude-scripts', 'upgrade', 'always-copy',
        'editable',
        'no-deps', 'local-snapshots-ok', 'version',
        'user'
    ]

    negative_opt = {'always-unzip': 'zip-ok'}
    create_index = PackageIndex

    def initialize_options(self):
        warnings.warn(
            "easy_install command is deprecated. "
            "Use build and pip and other standards-based tools.",
            EasyInstallDeprecationWarning,
        )

        # the --user option seems to be an opt-in one,
        # so the default should be False.
        self.user = 0
        self.zip_ok = self.local_snapshots_ok = None
        self.install_dir = self.script_dir = self.exclude_scripts = None
        self.index_url = None
        self.find_links = None
        self.build_directory = None
        self.args = None
        self.optimize = self.record = None
        self.upgrade = self.always_copy = self.multi_version = None
        self.editable = self.no_deps = self.allow_hosts = None
        self.root = self.prefix = self.no_report = None
        self.version = None
        self.install_purelib = None  # for pure module distributions
        self.install_platlib = None  # non-pure (dists w/ extensions)
        self.install_headers = None  # for C/C++ headers
        self.install_lib = None  # set to either purelib or platlib
        self.install_scripts = None
        self.install_data = None
        self.install_base = None
        self.install_platbase = None
        if site.ENABLE_USER_SITE:
            self.install_userbase = site.USER_BASE
            self.install_usersite = site.USER_SITE
        else:
            self.install_userbase = None
            self.install_usersite = None
        self.no_find_links = None

        # Options not specifiable via command line
        self.package_index = None
        self.pth_file = self.always_copy_from = None
        self.site_dirs = None
        self.installed_projects = {}
        # Always read easy_install options, even if we are subclassed, or have
        # an independent instance created.  This ensures that defaults will
        # always come from the standard configuration file(s)' "easy_install"
        # section, even if this is a "develop" or "install" command, or some
        # other embedding.
        self._dry_run = None
        self.verbose = self.distribution.verbose
        self.distribution._set_command_options(
            self, self.distribution.get_option_dict('easy_install')
        )

    def delete_blockers(self, blockers):
        extant_blockers = (
            filename for filename in blockers
            if os.path.exists(filename) or os.path.islink(filename)
        )
        list(map(self._delete_path, extant_blockers))

    def _delete_path(self, path):
        log.info("Deleting %s", path)
        if self.dry_run:
            return

        is_tree = os.path.isdir(path) and not os.path.islink(path)
        remover = rmtree if is_tree else os.unlink
        remover(path)

    @staticmethod
    def _render_version():
        """
        Render the Setuptools version and installation details, then exit.
        """
        ver = '{}.{}'.format(*sys.version_info)
        dist = get_distribution('setuptools')
        tmpl = 'setuptools {dist.version} from {dist.location} (Python {ver})'
        print(tmpl.format(**locals()))
        raise SystemExit()

    def finalize_options(self):  # noqa: C901  # is too complex (25)  # FIXME
        self.version and self._render_version()

        py_version = sys.version.split()[0]
        prefix, exec_prefix = get_config_vars('prefix', 'exec_prefix')

        self.config_vars = {
            'dist_name': self.distribution.get_name(),
            'dist_version': self.distribution.get_version(),
            'dist_fullname': self.distribution.get_fullname(),
            'py_version': py_version,
            'py_version_short': py_version[0:3],
            'py_version_nodot': py_version[0] + py_version[2],
            'sys_prefix': prefix,
            'prefix': prefix,
            'sys_exec_prefix': exec_prefix,
            'exec_prefix': exec_prefix,
            # Only python 3.2+ has abiflags
            'abiflags': getattr(sys, 'abiflags', ''),
        }

        if site.ENABLE_USER_SITE:
            self.config_vars['userbase'] = self.install_userbase
            self.config_vars['usersite'] = self.install_usersite

        elif self.user:
            log.warn("WARNING: The user site-packages directory is disabled.")

        self._fix_install_dir_for_user_site()

        self.expand_basedirs()
        self.expand_dirs()

        self._expand(
            'install_dir', 'script_dir', 'build_directory',
            'site_dirs',
        )
        # If a non-default installation directory was specified, default the
        # script directory to match it.
        if self.script_dir is None:
            self.script_dir = self.install_dir

        if self.no_find_links is None:
            self.no_find_links = False

        # Let install_dir get set by install_lib command, which in turn
        # gets its info from the install command, and takes into account
        # --prefix and --home and all that other crud.
        self.set_undefined_options(
            'install_lib', ('install_dir', 'install_dir')
        )
        # Likewise, set default script_dir from 'install_scripts.install_dir'
        self.set_undefined_options(
            'install_scripts', ('install_dir', 'script_dir')
        )

        if self.user and self.install_purelib:
            self.install_dir = self.install_purelib
            self.script_dir = self.install_scripts
        # default --record from the install command
        self.set_undefined_options('install', ('record', 'record'))
        # Should this be moved to the if statement below? It's not used
        # elsewhere
        normpath = map(normalize_path, sys.path)
        self.all_site_dirs = get_site_dirs()
        if self.site_dirs is not None:
            site_dirs = [
                os.path.expanduser(s.strip()) for s in
                self.site_dirs.split(',')
            ]
            for d in site_dirs:
                if not os.path.isdir(d):
                    log.warn("%s (in --site-dirs) does not exist", d)
                elif normalize_path(d) not in normpath:
                    raise DistutilsOptionError(
                        d + " (in --site-dirs) is not on sys.path"
                    )
                else:
                    self.all_site_dirs.append(normalize_path(d))
        if not self.editable:
            self.check_site_dir()
        self.index_url = self.index_url or "https://pypi.org/simple/"
        self.shadow_path = self.all_site_dirs[:]
        for path_item in self.install_dir, normalize_path(self.script_dir):
            if path_item not in self.shadow_path:
                self.shadow_path.insert(0, path_item)

        if self.allow_hosts is not None:
            hosts = [s.strip() for s in self.allow_hosts.split(',')]
        else:
            hosts = ['*']
        if self.package_index is None:
            self.package_index = self.create_index(
                self.index_url, search_path=self.shadow_path, hosts=hosts,
            )
        self.local_index = Environment(self.shadow_path + sys.path)

        if self.find_links is not None:
            if isinstance(self.find_links, str):
                self.find_links = self.find_links.split()
        else:
            self.find_links = []
        if self.local_snapshots_ok:
            self.package_index.scan_egg_links(self.shadow_path + sys.path)
        if not self.no_find_links:
            self.package_index.add_find_links(self.find_links)
        self.set_undefined_options('install_lib', ('optimize', 'optimize'))
        if not isinstance(self.optimize, int):
            try:
                self.optimize = int(self.optimize)
                if not (0 <= self.optimize <= 2):
                    raise ValueError
            except ValueError as e:
                raise DistutilsOptionError(
                    "--optimize must be 0, 1, or 2"
                ) from e

        if self.editable and not self.build_directory:
            raise DistutilsArgError(
                "Must specify a build directory (-b) when using --editable"
            )
        if not self.args:
            raise DistutilsArgError(
                "No urls, filenames, or requirements specified (see --help)")

        self.outputs = []

    def _fix_install_dir_for_user_site(self):
        """
        Fix the install_dir if "--user" was used.
        """
        if not self.user or not site.ENABLE_USER_SITE:
            return

        self.create_home_path()
        if self.install_userbase is None:
            msg = "User base directory is not specified"
            raise DistutilsPlatformError(msg)
        self.install_base = self.install_platbase = self.install_userbase
        scheme_name = os.name.replace('posix', 'unix') + '_user'
        self.select_scheme(scheme_name)

    def _expand_attrs(self, attrs):
        for attr in attrs:
            val = getattr(self, attr)
            if val is not None:
                if os.name == 'posix' or os.name == 'nt':
                    val = os.path.expanduser(val)
                val = subst_vars(val, self.config_vars)
                setattr(self, attr, val)

    def expand_basedirs(self):
        """Calls `os.path.expanduser` on install_base, install_platbase and
        root."""
        self._expand_attrs(['install_base', 'install_platbase', 'root'])

    def expand_dirs(self):
        """Calls `os.path.expanduser` on install dirs."""
        dirs = [
            'install_purelib',
            'install_platlib',
            'install_lib',
            'install_headers',
            'install_scripts',
            'install_data',
        ]
        self._expand_attrs(dirs)

    def run(self, show_deprecation=True):
        if show_deprecation:
            self.announce(
                "WARNING: The easy_install command is deprecated "
                "and will be removed in a future version.",
                log.WARN,
            )
        if self.verbose != self.distribution.verbose:
            log.set_verbosity(self.verbose)
        try:
            for spec in self.args:
                self.easy_install(spec, not self.no_deps)
            if self.record:
                outputs = self.outputs
                if self.root:  # strip any package prefix
                    root_len = len(self.root)
                    for counter in range(len(outputs)):
                        outputs[counter] = outputs[counter][root_len:]
                from distutils import file_util

                self.execute(
                    file_util.write_file, (self.record, outputs),
                    "writing list of installed files to '%s'" %
                    self.record
                )
            self.warn_deprecated_options()
        finally:
            log.set_verbosity(self.distribution.verbose)

    def pseudo_tempname(self):
        """Return a pseudo-tempname base in the install directory.
        This code is intentionally naive; if a malicious party can write to
        the target directory you're already in deep doodoo.
        """
        try:
            pid = os.getpid()
        except Exception:
            pid = random.randint(0, sys.maxsize)
        return os.path.join(self.install_dir, "test-easy-install-%s" % pid)

    def warn_deprecated_options(self):
        pass

    def check_site_dir(self):  # noqa: C901  # is too complex (12)  # FIXME
        """Verify that self.install_dir is .pth-capable dir, if needed"""

        instdir = normalize_path(self.install_dir)
        pth_file = os.path.join(instdir, 'easy-install.pth')

        if not os.path.exists(instdir):
            try:
                os.makedirs(instdir)
            except (OSError, IOError):
                self.cant_write_to_target()

        # Is it a configured, PYTHONPATH, implicit, or explicit site dir?
        is_site_dir = instdir in self.all_site_dirs

        if not is_site_dir and not self.multi_version:
            # No?  Then directly test whether it does .pth file processing
            is_site_dir = self.check_pth_processing()
        else:
            # make sure we can write to target dir
            testfile = self.pseudo_tempname() + '.write-test'
            test_exists = os.path.exists(testfile)
            try:
                if test_exists:
                    os.unlink(testfile)
                open(testfile, 'w').close()
                os.unlink(testfile)
            except (OSError, IOError):
                self.cant_write_to_target()

        if not is_site_dir and not self.multi_version:
            # Can't install non-multi to non-site dir with easy_install
            pythonpath = os.environ.get('PYTHONPATH', '')
            log.warn(self.__no_default_msg, self.install_dir, pythonpath)

        if is_site_dir:
            if self.pth_file is None:
                self.pth_file = PthDistributions(pth_file, self.all_site_dirs)
        else:
            self.pth_file = None

        if self.multi_version and not os.path.exists(pth_file):
            self.pth_file = None  # don't create a .pth file
        self.install_dir = instdir

    __cant_write_msg = textwrap.dedent("""
        can't create or remove files in install directory

        The following error occurred while trying to add or remove files in the
        installation directory:

            %s

        The installation directory you specified (via --install-dir, --prefix, or
        the distutils default setting) was:

            %s
        """).lstrip()  # noqa

    __not_exists_id = textwrap.dedent("""
        This directory does not currently exist.  Please create it and try again, or
        choose a different installation directory (using the -d or --install-dir
        option).
        """).lstrip()  # noqa

    __access_msg = textwrap.dedent("""
        Perhaps your account does not have write access to this directory?  If the
        installation directory is a system-owned directory, you may need to sign in
        as the administrator or "root" account.  If you do not have administrative
        access to this machine, you may wish to choose a different installation
        directory, preferably one that is listed in your PYTHONPATH environment
        variable.

        For information on other options, you may wish to consult the
        documentation at:

          https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html

        Please make the appropriate changes for your system and try again.
        """).lstrip()  # noqa

    def cant_write_to_target(self):
        msg = self.__cant_write_msg % (sys.exc_info()[1], self.install_dir,)

        if not os.path.exists(self.install_dir):
            msg += '\n' + self.__not_exists_id
        else:
            msg += '\n' + self.__access_msg
        raise DistutilsError(msg)

    def check_pth_processing(self):
        """Empirically verify whether .pth files are supported in inst. dir"""
        instdir = self.install_dir
        log.info("Checking .pth file support in %s", instdir)
        pth_file = self.pseudo_tempname() + ".pth"
        ok_file = pth_file + '.ok'
        ok_exists = os.path.exists(ok_file)
        tmpl = _one_liner("""
            import os
            f = open({ok_file!r}, 'w')
            f.write('OK')
            f.close()
            """) + '\n'
        try:
            if ok_exists:
                os.unlink(ok_file)
            dirname = os.path.dirname(ok_file)
            os.makedirs(dirname, exist_ok=True)
            f = open(pth_file, 'w')
        except (OSError, IOError):
            self.cant_write_to_target()
        else:
            try:
                f.write(tmpl.format(**locals()))
                f.close()
                f = None
                executable = sys.executable
                if os.name == 'nt':
                    dirname, basename = os.path.split(executable)
                    alt = os.path.join(dirname, 'pythonw.exe')
                    use_alt = (
                        basename.lower() == 'python.exe' and
                        os.path.exists(alt)
                    )
                    if use_alt:
                        # use pythonw.exe to avoid opening a console window
                        executable = alt

                from distutils.spawn import spawn

                spawn([executable, '-E', '-c', 'pass'], 0)

                if os.path.exists(ok_file):
                    log.info(
                        "TEST PASSED: %s appears to support .pth files",
                        instdir
                    )
                    return True
            finally:
                if f:
                    f.close()
                if os.path.exists(ok_file):
                    os.unlink(ok_file)
                if os.path.exists(pth_file):
                    os.unlink(pth_file)
        if not self.multi_version:
            log.warn("TEST FAILED: %s does NOT support .pth files", instdir)
        return False

    def install_egg_scripts(self, dist):
        """Write all the scripts for `dist`, unless scripts are excluded"""
        if not self.exclude_scripts and dist.metadata_isdir('scripts'):
            for script_name in dist.metadata_listdir('scripts'):
                if dist.metadata_isdir('scripts/' + script_name):
                    # The "script" is a directory, likely a Python 3
                    # __pycache__ directory, so skip it.
                    continue
                self.install_script(
                    dist, script_name,
                    dist.get_metadata('scripts/' + script_name)
                )
        self.install_wrapper_scripts(dist)

    def add_output(self, path):
        if os.path.isdir(path):
            for base, dirs, files in os.walk(path):
                for filename in files:
                    self.outputs.append(os.path.join(base, filename))
        else:
            self.outputs.append(path)

    def not_editable(self, spec):
        if self.editable:
            raise DistutilsArgError(
                "Invalid argument %r: you can't use filenames or URLs "
                "with --editable (except via the --find-links option)."
                % (spec,)
            )

    def check_editable(self, spec):
        if not self.editable:
            return

        if os.path.exists(os.path.join(self.build_directory, spec.key)):
            raise DistutilsArgError(
                "%r already exists in %s; can't do a checkout there" %
                (spec.key, self.build_directory)
            )

    @contextlib.contextmanager
    def _tmpdir(self):
        tmpdir = tempfile.mkdtemp(prefix=u"easy_install-")
        try:
            # cast to str as workaround for #709 and #710 and #712
            yield str(tmpdir)
        finally:
            os.path.exists(tmpdir) and rmtree(tmpdir)

    def easy_install(self, spec, deps=False):
        with self._tmpdir() as tmpdir:
            if not isinstance(spec, Requirement):
                if URL_SCHEME(spec):
                    # It's a url, download it to tmpdir and process
                    self.not_editable(spec)
                    dl = self.package_index.download(spec, tmpdir)
                    return self.install_item(None, dl, tmpdir, deps, True)

                elif os.path.exists(spec):
                    # Existing file or directory, just process it directly
                    self.not_editable(spec)
                    return self.install_item(None, spec, tmpdir, deps, True)
                else:
                    spec = parse_requirement_arg(spec)

            self.check_editable(spec)
            dist = self.package_index.fetch_distribution(
                spec, tmpdir, self.upgrade, self.editable,
                not self.always_copy, self.local_index
            )
            if dist is None:
                msg = "Could not find suitable distribution for %r" % spec
                if self.always_copy:
                    msg += " (--always-copy skips system and development eggs)"
                raise DistutilsError(msg)
            elif dist.precedence == DEVELOP_DIST:
                # .egg-info dists don't need installing, just process deps
                self.process_distribution(spec, dist, deps, "Using")
                return dist
            else:
                return self.install_item(spec, dist.location, tmpdir, deps)

    def install_item(self, spec, download, tmpdir, deps, install_needed=False):

        # Installation is also needed if file in tmpdir or is not an egg
        install_needed = install_needed or self.always_copy
        install_needed = install_needed or os.path.dirname(download) == tmpdir
        install_needed = install_needed or not download.endswith('.egg')
        install_needed = install_needed or (
            self.always_copy_from is not None and
            os.path.dirname(normalize_path(download)) ==
            normalize_path(self.always_copy_from)
        )

        if spec and not install_needed:
            # at this point, we know it's a local .egg, we just don't know if
            # it's already installed.
            for dist in self.local_index[spec.project_name]:
                if dist.location == download:
                    break
            else:
                install_needed = True  # it's not in the local index

        log.info("Processing %s", os.path.basename(download))

        if install_needed:
            dists = self.install_eggs(spec, download, tmpdir)
            for dist in dists:
                self.process_distribution(spec, dist, deps)
        else:
            dists = [self.egg_distribution(download)]
            self.process_distribution(spec, dists[0], deps, "Using")

        if spec is not None:
            for dist in dists:
                if dist in spec:
                    return dist

    def select_scheme(self, name):
        """Sets the install directories by applying the install schemes."""
        # it's the caller's problem if they supply a bad name!
        scheme = INSTALL_SCHEMES[name]
        for key in SCHEME_KEYS:
            attrname = 'install_' + key
            if getattr(self, attrname) is None:
                setattr(self, attrname, scheme[key])

    # FIXME: 'easy_install.process_distribution' is too complex (12)
    def process_distribution(  # noqa: C901
            self, requirement, dist, deps=True, *info,
    ):
        self.update_pth(dist)
        self.package_index.add(dist)
        if dist in self.local_index[dist.key]:
            self.local_index.remove(dist)
        self.local_index.add(dist)
        self.install_egg_scripts(dist)
        self.installed_projects[dist.key] = dist
        log.info(self.installation_report(requirement, dist, *info))
        if (dist.has_metadata('dependency_links.txt') and
                not self.no_find_links):
            self.package_index.add_find_links(
                dist.get_metadata_lines('dependency_links.txt')
            )
        if not deps and not self.always_copy:
            return
        elif requirement is not None and dist.key != requirement.key:
            log.warn("Skipping dependencies for %s", dist)
            return  # XXX this is not the distribution we were looking for
        elif requirement is None or dist not in requirement:
            # if we wound up with a different version, resolve what we've got
            distreq = dist.as_requirement()
            requirement = Requirement(str(distreq))
        log.info("Processing dependencies for %s", requirement)
        try:
            distros = WorkingSet([]).resolve(
                [requirement], self.local_index, self.easy_install
            )
        except DistributionNotFound as e:
            raise DistutilsError(str(e)) from e
        except VersionConflict as e:
            raise DistutilsError(e.report()) from e
        if self.always_copy or self.always_copy_from:
            # Force all the relevant distros to be copied or activated
            for dist in distros:
                if dist.key not in self.installed_projects:
                    self.easy_install(dist.as_requirement())
        log.info("Finished processing dependencies for %s", requirement)

    def should_unzip(self, dist):
        if self.zip_ok is not None:
            return not self.zip_ok
        if dist.has_metadata('not-zip-safe'):
            return True
        if not dist.has_metadata('zip-safe'):
            return True
        return False

    def maybe_move(self, spec, dist_filename, setup_base):
        dst = os.path.join(self.build_directory, spec.key)
        if os.path.exists(dst):
            msg = (
                "%r already exists in %s; build directory %s will not be kept"
            )
            log.warn(msg, spec.key, self.build_directory, setup_base)
            return setup_base
        if os.path.isdir(dist_filename):
            setup_base = dist_filename
        else:
            if os.path.dirname(dist_filename) == setup_base:
                os.unlink(dist_filename)  # get it out of the tmp dir
            contents = os.listdir(setup_base)
            if len(contents) == 1:
                dist_filename = os.path.join(setup_base, contents[0])
                if os.path.isdir(dist_filename):
                    # if the only thing there is a directory, move it instead
                    setup_base = dist_filename
        ensure_directory(dst)
        shutil.move(setup_base, dst)
        return dst

    def install_wrapper_scripts(self, dist):
        if self.exclude_scripts:
            return
        for args in ScriptWriter.best().get_args(dist):
            self.write_script(*args)

    def install_script(self, dist, script_name, script_text, dev_path=None):
        """Generate a legacy script wrapper and install it"""
        spec = str(dist.as_requirement())
        is_script = is_python_script(script_text, script_name)

        if is_script:
            body = self._load_template(dev_path) % locals()
            script_text = ScriptWriter.get_header(script_text) + body
        self.write_script(script_name, _to_bytes(script_text), 'b')

    @staticmethod
    def _load_template(dev_path):
        """
        There are a couple of template scripts in the package. This
        function loads one of them and prepares it for use.
        """
        # See https://github.com/pypa/setuptools/issues/134 for info
        # on script file naming and downstream issues with SVR4
        name = 'script.tmpl'
        if dev_path:
            name = name.replace('.tmpl', ' (dev).tmpl')

        raw_bytes = resource_string('setuptools', name)
        return raw_bytes.decode('utf-8')

    def write_script(self, script_name, contents, mode="t", blockers=()):
        """Write an executable file to the scripts directory"""
        self.delete_blockers(  # clean up old .py/.pyw w/o a script
            [os.path.join(self.script_dir, x) for x in blockers]
        )
        log.info("Installing %s script to %s", script_name, self.script_dir)
        target = os.path.join(self.script_dir, script_name)
        self.add_output(target)

        if self.dry_run:
            return

        mask = current_umask()
        ensure_directory(target)
        if os.path.exists(target):
            os.unlink(target)
        with open(target, "w" + mode) as f:
            f.write(contents)
        chmod(target, 0o777 - mask)

    def install_eggs(self, spec, dist_filename, tmpdir):
        # .egg dirs or files are already built, so just return them
        installer_map = {
            '.egg': self.install_egg,
            '.exe': self.install_exe,
            '.whl': self.install_wheel,
        }
        try:
            install_dist = installer_map[
                dist_filename.lower()[-4:]
            ]
        except KeyError:
            pass
        else:
            return [install_dist(dist_filename, tmpdir)]

        # Anything else, try to extract and build
        setup_base = tmpdir
        if os.path.isfile(dist_filename) and not dist_filename.endswith('.py'):
            unpack_archive(dist_filename, tmpdir, self.unpack_progress)
        elif os.path.isdir(dist_filename):
            setup_base = os.path.abspath(dist_filename)

        if (setup_base.startswith(tmpdir)  # something we downloaded
                and self.build_directory and spec is not None):
            setup_base = self.maybe_move(spec, dist_filename, setup_base)

        # Find the setup.py file
        setup_script = os.path.join(setup_base, 'setup.py')

        if not os.path.exists(setup_script):
            setups = glob(os.path.join(setup_base, '*', 'setup.py'))
            if not setups:
                raise DistutilsError(
                    "Couldn't find a setup script in %s" %
                    os.path.abspath(dist_filename)
                )
            if len(setups) > 1:
                raise DistutilsError(
                    "Multiple setup scripts in %s" %
                    os.path.abspath(dist_filename)
                )
            setup_script = setups[0]

        # Now run it, and return the result
        if self.editable:
            log.info(self.report_editable(spec, setup_script))
            return []
        else:
            return self.build_and_install(setup_script, setup_base)

    def egg_distribution(self, egg_path):
        if os.path.isdir(egg_path):
            metadata = PathMetadata(egg_path, os.path.join(egg_path,
                                                           'EGG-INFO'))
        else:
            metadata = EggMetadata(zipimport.zipimporter(egg_path))
        return Distribution.from_filename(egg_path, metadata=metadata)

    # FIXME: 'easy_install.install_egg' is too complex (11)
    def install_egg(self, egg_path, tmpdir):  # noqa: C901
        destination = os.path.join(
            self.install_dir,
            os.path.basename(egg_path),
        )
        destination = os.path.abspath(destination)
        if not self.dry_run:
            ensure_directory(destination)

        dist = self.egg_distribution(egg_path)
        if not samefile(egg_path, destination):
            if os.path.isdir(destination) and not os.path.islink(destination):
                dir_util.remove_tree(destination, dry_run=self.dry_run)
            elif os.path.exists(destination):
                self.execute(
                    os.unlink,
                    (destination,),
                    "Removing " + destination,
                )
            try:
                new_dist_is_zipped = False
                if os.path.isdir(egg_path):
                    if egg_path.startswith(tmpdir):
                        f, m = shutil.move, "Moving"
                    else:
                        f, m = shutil.copytree, "Copying"
                elif self.should_unzip(dist):
                    self.mkpath(destination)
                    f, m = self.unpack_and_compile, "Extracting"
                else:
                    new_dist_is_zipped = True
                    if egg_path.startswith(tmpdir):
                        f, m = shutil.move, "Moving"
                    else:
                        f, m = shutil.copy2, "Copying"
                self.execute(
                    f,
                    (egg_path, destination),
                    (m + " %s to %s") % (
                        os.path.basename(egg_path),
                        os.path.dirname(destination)
                    ),
                )
                update_dist_caches(
                    destination,
                    fix_zipimporter_caches=new_dist_is_zipped,
                )
            except Exception:
                update_dist_caches(destination, fix_zipimporter_caches=False)
                raise

        self.add_output(destination)
        return self.egg_distribution(destination)

    def install_exe(self, dist_filename, tmpdir):
        # See if it's valid, get data
        cfg = extract_wininst_cfg(dist_filename)
        if cfg is None:
            raise DistutilsError(
                "%s is not a valid distutils Windows .exe" % dist_filename
            )
        # Create a dummy distribution object until we build the real distro
        dist = Distribution(
            None,
            project_name=cfg.get('metadata', 'name'),
            version=cfg.get('metadata', 'version'), platform=get_platform(),
        )

        # Convert the .exe to an unpacked egg
        egg_path = os.path.join(tmpdir, dist.egg_name() + '.egg')
        dist.location = egg_path
        egg_tmp = egg_path + '.tmp'
        _egg_info = os.path.join(egg_tmp, 'EGG-INFO')
        pkg_inf = os.path.join(_egg_info, 'PKG-INFO')
        ensure_directory(pkg_inf)  # make sure EGG-INFO dir exists
        dist._provider = PathMetadata(egg_tmp, _egg_info)  # XXX
        self.exe_to_egg(dist_filename, egg_tmp)

        # Write EGG-INFO/PKG-INFO
        if not os.path.exists(pkg_inf):
            f = open(pkg_inf, 'w')
            f.write('Metadata-Version: 1.0\n')
            for k, v in cfg.items('metadata'):
                if k != 'target_version':
                    f.write('%s: %s\n' % (k.replace('_', '-').title(), v))
            f.close()
        script_dir = os.path.join(_egg_info, 'scripts')
        # delete entry-point scripts to avoid duping
        self.delete_blockers([
            os.path.join(script_dir, args[0])
            for args in ScriptWriter.get_args(dist)
        ])
        # Build .egg file from tmpdir
        bdist_egg.make_zipfile(
            egg_path, egg_tmp, verbose=self.verbose, dry_run=self.dry_run,
        )
        # install the .egg
        return self.install_egg(egg_path, tmpdir)

    # FIXME: 'easy_install.exe_to_egg' is too complex (12)
    def exe_to_egg(self, dist_filename, egg_tmp):  # noqa: C901
        """Extract a bdist_wininst to the directories an egg would use"""
        # Check for .pth file and set up prefix translations
        prefixes = get_exe_prefixes(dist_filename)
        to_compile = []
        native_libs = []
        top_level = {}

        def process(src, dst):
            s = src.lower()
            for old, new in prefixes:
                if s.startswith(old):
                    src = new + src[len(old):]
                    parts = src.split('/')
                    dst = os.path.join(egg_tmp, *parts)
                    dl = dst.lower()
                    if dl.endswith('.pyd') or dl.endswith('.dll'):
                        parts[-1] = bdist_egg.strip_module(parts[-1])
                        top_level[os.path.splitext(parts[0])[0]] = 1
                        native_libs.append(src)
                    elif dl.endswith('.py') and old != 'SCRIPTS/':
                        top_level[os.path.splitext(parts[0])[0]] = 1
                        to_compile.append(dst)
                    return dst
            if not src.endswith('.pth'):
                log.warn("WARNING: can't process %s", src)
            return None

        # extract, tracking .pyd/.dll->native_libs and .py -> to_compile
        unpack_archive(dist_filename, egg_tmp, process)
        stubs = []
        for res in native_libs:
            if res.lower().endswith('.pyd'):  # create stubs for .pyd's
                parts = res.split('/')
                resource = parts[-1]
                parts[-1] = bdist_egg.strip_module(parts[-1]) + '.py'
                pyfile = os.path.join(egg_tmp, *parts)
                to_compile.append(pyfile)
                stubs.append(pyfile)
                bdist_egg.write_stub(resource, pyfile)
        self.byte_compile(to_compile)  # compile .py's
        bdist_egg.write_safety_flag(
            os.path.join(egg_tmp, 'EGG-INFO'),
            bdist_egg.analyze_egg(egg_tmp, stubs))  # write zip-safety flag

        for name in 'top_level', 'native_libs':
            if locals()[name]:
                txt = os.path.join(egg_tmp, 'EGG-INFO', name + '.txt')
                if not os.path.exists(txt):
                    f = open(txt, 'w')
                    f.write('\n'.join(locals()[name]) + '\n')
                    f.close()

    def install_wheel(self, wheel_path, tmpdir):
        wheel = Wheel(wheel_path)
        assert wheel.is_compatible()
        destination = os.path.join(self.install_dir, wheel.egg_name())
        destination = os.path.abspath(destination)
        if not self.dry_run:
            ensure_directory(destination)
        if os.path.isdir(destination) and not os.path.islink(destination):
            dir_util.remove_tree(destination, dry_run=self.dry_run)
        elif os.path.exists(destination):
            self.execute(
                os.unlink,
                (destination,),
                "Removing " + destination,
            )
        try:
            self.execute(
                wheel.install_as_egg,
                (destination,),
                ("Installing %s to %s") % (
                    os.path.basename(wheel_path),
                    os.path.dirname(destination)
                ),
            )
        finally:
            update_dist_caches(destination, fix_zipimporter_caches=False)
        self.add_output(destination)
        return self.egg_distribution(destination)

    __mv_warning = textwrap.dedent("""
        Because this distribution was installed --multi-version, before you can
        import modules from this package in an application, you will need to
        'import pkg_resources' and then use a 'require()' call similar to one of
        these examples, in order to select the desired version:

            pkg_resources.require("%(name)s")  # latest installed version
            pkg_resources.require("%(name)s==%(version)s")  # this exact version
            pkg_resources.require("%(name)s>=%(version)s")  # this version or higher
        """).lstrip()  # noqa

    __id_warning = textwrap.dedent("""
        Note also that the installation directory must be on sys.path at runtime for
        this to work.  (e.g. by being the application's script directory, by being on
        PYTHONPATH, or by being added to sys.path by your code.)
        """)  # noqa

    def installation_report(self, req, dist, what="Installed"):
        """Helpful installation message for display to package users"""
        msg = "\n%(what)s %(eggloc)s%(extras)s"
        if self.multi_version and not self.no_report:
            msg += '\n' + self.__mv_warning
            if self.install_dir not in map(normalize_path, sys.path):
                msg += '\n' + self.__id_warning

        eggloc = dist.location
        name = dist.project_name
        version = dist.version
        extras = ''  # TODO: self.report_extras(req, dist)
        return msg % locals()

    __editable_msg = textwrap.dedent("""
        Extracted editable version of %(spec)s to %(dirname)s

        If it uses setuptools in its setup script, you can activate it in
        "development" mode by going to that directory and running::

            %(python)s setup.py develop

        See the setuptools documentation for the "develop" command for more info.
        """).lstrip()  # noqa

    def report_editable(self, spec, setup_script):
        dirname = os.path.dirname(setup_script)
        python = sys.executable
        return '\n' + self.__editable_msg % locals()

    def run_setup(self, setup_script, setup_base, args):
        sys.modules.setdefault('distutils.command.bdist_egg', bdist_egg)
        sys.modules.setdefault('distutils.command.egg_info', egg_info)

        args = list(args)
        if self.verbose > 2:
            v = 'v' * (self.verbose - 1)
            args.insert(0, '-' + v)
        elif self.verbose < 2:
            args.insert(0, '-q')
        if self.dry_run:
            args.insert(0, '-n')
        log.info(
            "Running %s %s", setup_script[len(setup_base) + 1:], ' '.join(args)
        )
        try:
            run_setup(setup_script, args)
        except SystemExit as v:
            raise DistutilsError(
                "Setup script exited with %s" % (v.args[0],)
            ) from v

    def build_and_install(self, setup_script, setup_base):
        args = ['bdist_egg', '--dist-dir']

        dist_dir = tempfile.mkdtemp(
            prefix='egg-dist-tmp-', dir=os.path.dirname(setup_script)
        )
        try:
            self._set_fetcher_options(os.path.dirname(setup_script))
            args.append(dist_dir)

            self.run_setup(setup_script, setup_base, args)
            all_eggs = Environment([dist_dir])
            eggs = []
            for key in all_eggs:
                for dist in all_eggs[key]:
                    eggs.append(self.install_egg(dist.location, setup_base))
            if not eggs and not self.dry_run:
                log.warn("No eggs found in %s (setup script problem?)",
                         dist_dir)
            return eggs
        finally:
            rmtree(dist_dir)
            log.set_verbosity(self.verbose)  # restore our log verbosity

    def _set_fetcher_options(self, base):
        """
        When easy_install is about to run bdist_egg on a source dist, that
        source dist might have 'setup_requires' directives, requiring
        additional fetching. Ensure the fetcher options given to easy_install
        are available to that command as well.
        """
        # find the fetch options from easy_install and write them out
        # to the setup.cfg file.
        ei_opts = self.distribution.get_option_dict('easy_install').copy()
        fetch_directives = (
            'find_links', 'site_dirs', 'index_url', 'optimize', 'allow_hosts',
        )
        fetch_options = {}
        for key, val in ei_opts.items():
            if key not in fetch_directives:
                continue
            fetch_options[key] = val[1]
        # create a settings dictionary suitable for `edit_config`
        settings = dict(easy_install=fetch_options)
        cfg_filename = os.path.join(base, 'setup.cfg')
        setopt.edit_config(cfg_filename, settings)

    def update_pth(self, dist):  # noqa: C901  # is too complex (11)  # FIXME
        if self.pth_file is None:
            return

        for d in self.pth_file[dist.key]:  # drop old entries
            if not self.multi_version and d.location == dist.location:
                continue

            log.info("Removing %s from easy-install.pth file", d)
            self.pth_file.remove(d)
            if d.location in self.shadow_path:
                self.shadow_path.remove(d.location)

        if not self.multi_version:
            if dist.location in self.pth_file.paths:
                log.info(
                    "%s is already the active version in easy-install.pth",
                    dist,
                )
            else:
                log.info("Adding %s to easy-install.pth file", dist)
                self.pth_file.add(dist)  # add new entry
                if dist.location not in self.shadow_path:
                    self.shadow_path.append(dist.location)

        if self.dry_run:
            return

        self.pth_file.save()

        if dist.key != 'setuptools':
            return

        # Ensure that setuptools itself never becomes unavailable!
        # XXX should this check for latest version?
        filename = os.path.join(self.install_dir, 'setuptools.pth')
        if os.path.islink(filename):
            os.unlink(filename)
        with open(filename, 'wt') as f:
            f.write(self.pth_file.make_relative(dist.location) + '\n')

    def unpack_progress(self, src, dst):
        # Progress filter for unpacking
        log.debug("Unpacking %s to %s", src, dst)
        return dst  # only unpack-and-compile skips files for dry run

    def unpack_and_compile(self, egg_path, destination):
        to_compile = []
        to_chmod = []

        def pf(src, dst):
            if dst.endswith('.py') and not src.startswith('EGG-INFO/'):
                to_compile.append(dst)
            elif dst.endswith('.dll') or dst.endswith('.so'):
                to_chmod.append(dst)
            self.unpack_progress(src, dst)
            return not self.dry_run and dst or None

        unpack_archive(egg_path, destination, pf)
        self.byte_compile(to_compile)
        if not self.dry_run:
            for f in to_chmod:
                mode = ((os.stat(f)[stat.ST_MODE]) | 0o555) & 0o7755
                chmod(f, mode)

    def byte_compile(self, to_compile):
        if sys.dont_write_bytecode:
            return

        from distutils.util import byte_compile

        try:
            # try to make the byte compile messages quieter
            log.set_verbosity(self.verbose - 1)

            byte_compile(to_compile, optimize=0, force=1, dry_run=self.dry_run)
            if self.optimize:
                byte_compile(
                    to_compile, optimize=self.optimize, force=1,
                    dry_run=self.dry_run,
                )
        finally:
            log.set_verbosity(self.verbose)  # restore original verbosity

    __no_default_msg = textwrap.dedent("""
        bad install directory or PYTHONPATH

        You are attempting to install a package to a directory that is not
        on PYTHONPATH and which Python does not read ".pth" files from.  The
        installation directory you specified (via --install-dir, --prefix, or
        the distutils default setting) was:

            %s

        and your PYTHONPATH environment variable currently contains:

            %r

        Here are some of your options for correcting the problem:

        * You can choose a different installation directory, i.e., one that is
          on PYTHONPATH or supports .pth files

        * You can add the installation directory to the PYTHONPATH environment
          variable.  (It must then also be on PYTHONPATH whenever you run
          Python and want to use the package(s) you are installing.)

        * You can set up the installation directory to support ".pth" files by
          using one of the approaches described here:

          https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html#custom-installation-locations


        Please make the appropriate changes for your system and try again.
        """).strip()

    def create_home_path(self):
        """Create directories under ~."""
        if not self.user:
            return
        home = convert_path(os.path.expanduser("~"))
        for name, path in self.config_vars.items():
            if path.startswith(home) and not os.path.isdir(path):
                self.debug_print("os.makedirs('%s', 0o700)" % path)
                os.makedirs(path, 0o700)

    INSTALL_SCHEMES = dict(
        posix=dict(
            install_dir='$base/lib/python$py_version_short/site-packages',
            script_dir='$base/bin',
        ),
    )

    DEFAULT_SCHEME = dict(
        install_dir='$base/Lib/site-packages',
        script_dir='$base/Scripts',
    )

    def _expand(self, *attrs):
        config_vars = self.get_finalized_command('install').config_vars

        if self.prefix:
            # Set default install_dir/scripts from --prefix
            config_vars = config_vars.copy()
            config_vars['base'] = self.prefix
            scheme = self.INSTALL_SCHEMES.get(os.name, self.DEFAULT_SCHEME)
            for attr, val in scheme.items():
                if getattr(self, attr, None) is None:
                    setattr(self, attr, val)

        from distutils.util import subst_vars

        for attr in attrs:
            val = getattr(self, attr)
            if val is not None:
                val = subst_vars(val, config_vars)
                if os.name == 'posix':
                    val = os.path.expanduser(val)
                setattr(self, attr, val)


def _pythonpath():
    items = os.environ.get('PYTHONPATH', '').split(os.pathsep)
    return filter(None, items)


def get_site_dirs():
    """
    Return a list of 'site' dirs
    """

    sitedirs = []

    # start with PYTHONPATH
    sitedirs.extend(_pythonpath())

    prefixes = [sys.prefix]
    if sys.exec_prefix != sys.prefix:
        prefixes.append(sys.exec_prefix)
    for prefix in prefixes:
        if not prefix:
            continue

        if sys.platform in ('os2emx', 'riscos'):
            sitedirs.append(os.path.join(prefix, "Lib", "site-packages"))
        elif os.sep == '/':
            sitedirs.extend([
                os.path.join(
                    prefix,
                    "lib",
                    "python{}.{}".format(*sys.version_info),
                    "site-packages",
                ),
                os.path.join(prefix, "lib", "site-python"),
            ])
        else:
            sitedirs.extend([
                prefix,
                os.path.join(prefix, "lib", "site-packages"),
            ])
        if sys.platform != 'darwin':
            continue

        # for framework builds *only* we add the standard Apple
        # locations. Currently only per-user, but /Library and
        # /Network/Library could be added too
        if 'Python.framework' not in prefix:
            continue

        home = os.environ.get('HOME')
        if not home:
            continue

        home_sp = os.path.join(
            home,
            'Library',
            'Python',
            '{}.{}'.format(*sys.version_info),
            'site-packages',
        )
        sitedirs.append(home_sp)
    lib_paths = get_path('purelib'), get_path('platlib')

    sitedirs.extend(s for s in lib_paths if s not in sitedirs)

    if site.ENABLE_USER_SITE:
        sitedirs.append(site.USER_SITE)

    with contextlib.suppress(AttributeError):
        sitedirs.extend(site.getsitepackages())

    sitedirs = list(map(normalize_path, sitedirs))

    return sitedirs


def expand_paths(inputs):  # noqa: C901  # is too complex (11)  # FIXME
    """Yield sys.path directories that might contain "old-style" packages"""

    seen = {}

    for dirname in inputs:
        dirname = normalize_path(dirname)
        if dirname in seen:
            continue

        seen[dirname] = 1
        if not os.path.isdir(dirname):
            continue

        files = os.listdir(dirname)
        yield dirname, files

        for name in files:
            if not name.endswith('.pth'):
                # We only care about the .pth files
                continue
            if name in ('easy-install.pth', 'setuptools.pth'):
                # Ignore .pth files that we control
                continue

            # Read the .pth file
            f = open(os.path.join(dirname, name))
            lines = list(yield_lines(f))
            f.close()

            # Yield existing non-dupe, non-import directory lines from it
            for line in lines:
                if line.startswith("import"):
                    continue

                line = normalize_path(line.rstrip())
                if line in seen:
                    continue

                seen[line] = 1
                if not os.path.isdir(line):
                    continue

                yield line, os.listdir(line)


def extract_wininst_cfg(dist_filename):
    """Extract configuration data from a bdist_wininst .exe

    Returns a configparser.RawConfigParser, or None
    """
    f = open(dist_filename, 'rb')
    try:
        endrec = zipfile._EndRecData(f)
        if endrec is None:
            return None

        prepended = (endrec[9] - endrec[5]) - endrec[6]
        if prepended < 12:  # no wininst data here
            return None
        f.seek(prepended - 12)

        tag, cfglen, bmlen = struct.unpack("<iii", f.read(12))
        if tag not in (0x1234567A, 0x1234567B):
            return None  # not a valid tag

        f.seek(prepended - (12 + cfglen))
        init = {'version': '', 'target_version': ''}
        cfg = configparser.RawConfigParser(init)
        try:
            part = f.read(cfglen)
            # Read up to the first null byte.
            config = part.split(b'\0', 1)[0]
            # Now the config is in bytes, but for RawConfigParser, it should
            #  be text, so decode it.
            config = config.decode(sys.getfilesystemencoding())
            cfg.read_file(io.StringIO(config))
        except configparser.Error:
            return None
        if not cfg.has_section('metadata') or not cfg.has_section('Setup'):
            return None
        return cfg

    finally:
        f.close()


def get_exe_prefixes(exe_filename):
    """Get exe->egg path translations for a given .exe file"""

    prefixes = [
        ('PURELIB/', ''),
        ('PLATLIB/pywin32_system32', ''),
        ('PLATLIB/', ''),
        ('SCRIPTS/', 'EGG-INFO/scripts/'),
        ('DATA/lib/site-packages', ''),
    ]
    z = zipfile.ZipFile(exe_filename)
    try:
        for info in z.infolist():
            name = info.filename
            parts = name.split('/')
            if len(parts) == 3 and parts[2] == 'PKG-INFO':
                if parts[1].endswith('.egg-info'):
                    prefixes.insert(0, ('/'.join(parts[:2]), 'EGG-INFO/'))
                    break
            if len(parts) != 2 or not name.endswith('.pth'):
                continue
            if name.endswith('-nspkg.pth'):
                continue
            if parts[0].upper() in ('PURELIB', 'PLATLIB'):
                contents = z.read(name).decode()
                for pth in yield_lines(contents):
                    pth = pth.strip().replace('\\', '/')
                    if not pth.startswith('import'):
                        prefixes.append((('%s/%s/' % (parts[0], pth)), ''))
    finally:
        z.close()
    prefixes = [(x.lower(), y) for x, y in prefixes]
    prefixes.sort()
    prefixes.reverse()
    return prefixes


class PthDistributions(Environment):
    """A .pth file with Distribution paths in it"""

    dirty = False

    def __init__(self, filename, sitedirs=()):
        self.filename = filename
        self.sitedirs = list(map(normalize_path, sitedirs))
        self.basedir = normalize_path(os.path.dirname(self.filename))
        self._load()
        Environment.__init__(self, [], None, None)
        for path in yield_lines(self.paths):
            list(map(self.add, find_distributions(path, True)))

    def _load(self):
        self.paths = []
        saw_import = False
        seen = dict.fromkeys(self.sitedirs)
        if os.path.isfile(self.filename):
            f = open(self.filename, 'rt')
            for line in f:
                if line.startswith('import'):
                    saw_import = True
                    continue
                path = line.rstrip()
                self.paths.append(path)
                if not path.strip() or path.strip().startswith('#'):
                    continue
                # skip non-existent paths, in case somebody deleted a package
                # manually, and duplicate paths as well
                path = self.paths[-1] = normalize_path(
                    os.path.join(self.basedir, path)
                )
                if not os.path.exists(path) or path in seen:
                    self.paths.pop()  # skip it
                    self.dirty = True  # we cleaned up, so we're dirty now :)
                    continue
                seen[path] = 1
            f.close()

        if self.paths and not saw_import:
            self.dirty = True  # ensure anything we touch has import wrappers
        while self.paths and not self.paths[-1].strip():
            self.paths.pop()

    def save(self):
        """Write changed .pth file back to disk"""
        if not self.dirty:
            return

        rel_paths = list(map(self.make_relative, self.paths))
        if rel_paths:
            log.debug("Saving %s", self.filename)
            lines = self._wrap_lines(rel_paths)
            data = '\n'.join(lines) + '\n'

            if os.path.islink(self.filename):
                os.unlink(self.filename)
            with open(self.filename, 'wt') as f:
                f.write(data)

        elif os.path.exists(self.filename):
            log.debug("Deleting empty %s", self.filename)
            os.unlink(self.filename)

        self.dirty = False

    @staticmethod
    def _wrap_lines(lines):
        return lines

    def add(self, dist):
        """Add `dist` to the distribution map"""
        new_path = (
            dist.location not in self.paths and (
                dist.location not in self.sitedirs or
                # account for '.' being in PYTHONPATH
                dist.location == os.getcwd()
            )
        )
        if new_path:
            self.paths.append(dist.location)
            self.dirty = True
        Environment.add(self, dist)

    def remove(self, dist):
        """Remove `dist` from the distribution map"""
        while dist.location in self.paths:
            self.paths.remove(dist.location)
            self.dirty = True
        Environment.remove(self, dist)

    def make_relative(self, path):
        npath, last = os.path.split(normalize_path(path))
        baselen = len(self.basedir)
        parts = [last]
        sep = os.altsep == '/' and '/' or os.sep
        while len(npath) >= baselen:
            if npath == self.basedir:
                parts.append(os.curdir)
                parts.reverse()
                return sep.join(parts)
            npath, last = os.path.split(npath)
            parts.append(last)
        else:
            return path


class RewritePthDistributions(PthDistributions):
    @classmethod
    def _wrap_lines(cls, lines):
        yield cls.prelude
        for line in lines:
            yield line
        yield cls.postlude

    prelude = _one_liner("""
        import sys
        sys.__plen = len(sys.path)
        """)
    postlude = _one_liner("""
        import sys
        new = sys.path[sys.__plen:]
        del sys.path[sys.__plen:]
        p = getattr(sys, '__egginsert', 0)
        sys.path[p:p] = new
        sys.__egginsert = p + len(new)
        """)


if os.environ.get('SETUPTOOLS_SYS_PATH_TECHNIQUE', 'raw') == 'rewrite':
    PthDistributions = RewritePthDistributions


def _first_line_re():
    """
    Return a regular expression based on first_line_re suitable for matching
    strings.
    """
    if isinstance(first_line_re.pattern, str):
        return first_line_re

    # first_line_re in Python >=3.1.4 and >=3.2.1 is a bytes pattern.
    return re.compile(first_line_re.pattern.decode())


def auto_chmod(func, arg, exc):
    if func in [os.unlink, os.remove] and os.name == 'nt':
        chmod(arg, stat.S_IWRITE)
        return func(arg)
    et, ev, _ = sys.exc_info()
    # TODO: This code doesn't make sense. What is it trying to do?
    raise (ev[0], ev[1] + (" %s %s" % (func, arg)))


def update_dist_caches(dist_path, fix_zipimporter_caches):
    """
    Fix any globally cached `dist_path` related data

    `dist_path` should be a path of a newly installed egg distribution (zipped
    or unzipped).

    sys.path_importer_cache contains finder objects that have been cached when
    importing data from the original distribution. Any such finders need to be
    cleared since the replacement distribution might be packaged differently,
    e.g. a zipped egg distribution might get replaced with an unzipped egg
    folder or vice versa. Having the old finders cached may then cause Python
    to attempt loading modules from the replacement distribution using an
    incorrect loader.

    zipimport.zipimporter objects are Python loaders charged with importing
    data packaged inside zip archives. If stale loaders referencing the
    original distribution, are left behind, they can fail to load modules from
    the replacement distribution. E.g. if an old zipimport.zipimporter instance
    is used to load data from a new zipped egg archive, it may cause the
    operation to attempt to locate the requested data in the wrong location -
    one indicated by the original distribution's zip archive directory
    information. Such an operation may then fail outright, e.g. report having
    read a 'bad local file header', or even worse, it may fail silently &
    return invalid data.

    zipimport._zip_directory_cache contains cached zip archive directory
    information for all existing zipimport.zipimporter instances and all such
    instances connected to the same archive share the same cached directory
    information.

    If asked, and the underlying Python implementation allows it, we can fix
    all existing zipimport.zipimporter instances instead of having to track
    them down and remove them one by one, by updating their shared cached zip
    archive directory information. This, of course, assumes that the
    replacement distribution is packaged as a zipped egg.

    If not asked to fix existing zipimport.zipimporter instances, we still do
    our best to clear any remaining zipimport.zipimporter related cached data
    that might somehow later get used when attempting to load data from the new
    distribution and thus cause such load operations to fail. Note that when
    tracking down such remaining stale data, we can not catch every conceivable
    usage from here, and we clear only those that we know of and have found to
    cause problems if left alive. Any remaining caches should be updated by
    whomever is in charge of maintaining them, i.e. they should be ready to
    handle us replacing their zip archives with new distributions at runtime.

    """
    # There are several other known sources of stale zipimport.zipimporter
    # instances that we do not clear here, but might if ever given a reason to
    # do so:
    # * Global setuptools pkg_resources.working_set (a.k.a. 'master working
    # set') may contain distributions which may in turn contain their
    #   zipimport.zipimporter loaders.
    # * Several zipimport.zipimporter loaders held by local variables further
    #   up the function call stack when running the setuptools installation.
    # * Already loaded modules may have their __loader__ attribute set to the
    #   exact loader instance used when importing them. Python 3.4 docs state
    #   that this information is intended mostly for introspection and so is
    #   not expected to cause us problems.
    normalized_path = normalize_path(dist_path)
    _uncache(normalized_path, sys.path_importer_cache)
    if fix_zipimporter_caches:
        _replace_zip_directory_cache_data(normalized_path)
    else:
        # Here, even though we do not want to fix existing and now stale
        # zipimporter cache information, we still want to remove it. Related to
        # Python's zip archive directory information cache, we clear each of
        # its stale entries in two phases:
        #   1. Clear the entry so attempting to access zip archive information
        #      via any existing stale zipimport.zipimporter instances fails.
        #   2. Remove the entry from the cache so any newly constructed
        #      zipimport.zipimporter instances do not end up using old stale
        #      zip archive directory information.
        # This whole stale data removal step does not seem strictly necessary,
        # but has been left in because it was done before we started replacing
        # the zip archive directory information cache content if possible, and
        # there are no relevant unit tests that we can depend on to tell us if
        # this is really needed.
        _remove_and_clear_zip_directory_cache_data(normalized_path)


def _collect_zipimporter_cache_entries(normalized_path, cache):
    """
    Return zipimporter cache entry keys related to a given normalized path.

    Alternative path spellings (e.g. those using different character case or
    those using alternative path separators) related to the same path are
    included. Any sub-path entries are included as well, i.e. those
    corresponding to zip archives embedded in other zip archives.

    """
    result = []
    prefix_len = len(normalized_path)
    for p in cache:
        np = normalize_path(p)
        if (np.startswith(normalized_path) and
                np[prefix_len:prefix_len + 1] in (os.sep, '')):
            result.append(p)
    return result


def _update_zipimporter_cache(normalized_path, cache, updater=None):
    """
    Update zipimporter cache data for a given normalized path.

    Any sub-path entries are processed as well, i.e. those corresponding to zip
    archives embedded in other zip archives.

    Given updater is a callable taking a cache entry key and the original entry
    (after already removing the entry from the cache), and expected to update
    the entry and possibly return a new one to be inserted in its place.
    Returning None indicates that the entry should not be replaced with a new
    one. If no updater is given, the cache entries are simply removed without
    any additional processing, the same as if the updater simply returned None.

    """
    for p in _collect_zipimporter_cache_entries(normalized_path, cache):
        # N.B. pypy's custom zipimport._zip_directory_cache implementation does
        # not support the complete dict interface:
        # * Does not support item assignment, thus not allowing this function
        #    to be used only for removing existing cache entries.
        #  * Does not support the dict.pop() method, forcing us to use the
        #    get/del patterns instead. For more detailed information see the
        #    following links:
        #      https://github.com/pypa/setuptools/issues/202#issuecomment-202913420
        #      http://bit.ly/2h9itJX
        old_entry = cache[p]
        del cache[p]
        new_entry = updater and updater(p, old_entry)
        if new_entry is not None:
            cache[p] = new_entry


def _uncache(normalized_path, cache):
    _update_zipimporter_cache(normalized_path, cache)


def _remove_and_clear_zip_directory_cache_data(normalized_path):
    def clear_and_remove_cached_zip_archive_directory_data(path, old_entry):
        old_entry.clear()

    _update_zipimporter_cache(
        normalized_path, zipimport._zip_directory_cache,
        updater=clear_and_remove_cached_zip_archive_directory_data)


# PyPy Python implementation does not allow directly writing to the
# zipimport._zip_directory_cache and so prevents us from attempting to correct
# its content. The best we can do there is clear the problematic cache content
# and have PyPy repopulate it as needed. The downside is that if there are any
# stale zipimport.zipimporter instances laying around, attempting to use them
# will fail due to not having its zip archive directory information available
# instead of being automatically corrected to use the new correct zip archive
# directory information.
if '__pypy__' in sys.builtin_module_names:
    _replace_zip_directory_cache_data = \
        _remove_and_clear_zip_directory_cache_data
else:

    def _replace_zip_directory_cache_data(normalized_path):
        def replace_cached_zip_archive_directory_data(path, old_entry):
            # N.B. In theory, we could load the zip directory information just
            # once for all updated path spellings, and then copy it locally and
            # update its contained path strings to contain the correct
            # spelling, but that seems like a way too invasive move (this cache
            # structure is not officially documented anywhere and could in
            # theory change with new Python releases) for no significant
            # benefit.
            old_entry.clear()
            zipimport.zipimporter(path)
            old_entry.update(zipimport._zip_directory_cache[path])
            return old_entry

        _update_zipimporter_cache(
            normalized_path, zipimport._zip_directory_cache,
            updater=replace_cached_zip_archive_directory_data)


def is_python(text, filename='<string>'):
    "Is this string a valid Python script?"
    try:
        compile(text, filename, 'exec')
    except (SyntaxError, TypeError):
        return False
    else:
        return True


def is_sh(executable):
    """Determine if the specified executable is a .sh (contains a #! line)"""
    try:
        with io.open(executable, encoding='latin-1') as fp:
            magic = fp.read(2)
    except (OSError, IOError):
        return executable
    return magic == '#!'


def nt_quote_arg(arg):
    """Quote a command line argument according to Windows parsing rules"""
    return subprocess.list2cmdline([arg])


def is_python_script(script_text, filename):
    """Is this text, as a whole, a Python script? (as opposed to shell/bat/etc.
    """
    if filename.endswith('.py') or filename.endswith('.pyw'):
        return True  # extension says it's Python
    if is_python(script_text, filename):
        return True  # it's syntactically valid Python
    if script_text.startswith('#!'):
        # It begins with a '#!' line, so check if 'python' is in it somewhere
        return 'python' in script_text.splitlines()[0].lower()

    return False  # Not any Python I can recognize


try:
    from os import chmod as _chmod
except ImportError:
    # Jython compatibility
    def _chmod(*args):
        pass


def chmod(path, mode):
    log.debug("changing mode of %s to %o", path, mode)
    try:
        _chmod(path, mode)
    except os.error as e:
        log.debug("chmod failed: %s", e)


class CommandSpec(list):
    """
    A command spec for a #! header, specified as a list of arguments akin to
    those passed to Popen.
    """

    options = []
    split_args = dict()

    @classmethod
    def best(cls):
        """
        Choose the best CommandSpec class based on environmental conditions.
        """
        return cls

    @classmethod
    def _sys_executable(cls):
        _default = os.path.normpath(sys.executable)
        return os.environ.get('__PYVENV_LAUNCHER__', _default)

    @classmethod
    def from_param(cls, param):
        """
        Construct a CommandSpec from a parameter to build_scripts, which may
        be None.
        """
        if isinstance(param, cls):
            return param
        if isinstance(param, list):
            return cls(param)
        if param is None:
            return cls.from_environment()
        # otherwise, assume it's a string.
        return cls.from_string(param)

    @classmethod
    def from_environment(cls):
        return cls([cls._sys_executable()])

    @classmethod
    def from_string(cls, string):
        """
        Construct a command spec from a simple string representing a command
        line parseable by shlex.split.
        """
        items = shlex.split(string, **cls.split_args)
        return cls(items)

    def install_options(self, script_text):
        self.options = shlex.split(self._extract_options(script_text))
        cmdline = subprocess.list2cmdline(self)
        if not isascii(cmdline):
            self.options[:0] = ['-x']

    @staticmethod
    def _extract_options(orig_script):
        """
        Extract any options from the first line of the script.
        """
        first = (orig_script + '\n').splitlines()[0]
        match = _first_line_re().match(first)
        options = match.group(1) or '' if match else ''
        return options.strip()

    def as_header(self):
        return self._render(self + list(self.options))

    @staticmethod
    def _strip_quotes(item):
        _QUOTES = '"\''
        for q in _QUOTES:
            if item.startswith(q) and item.endswith(q):
                return item[1:-1]
        return item

    @staticmethod
    def _render(items):
        cmdline = subprocess.list2cmdline(
            CommandSpec._strip_quotes(item.strip()) for item in items)
        return '#!' + cmdline + '\n'


# For pbr compat; will be removed in a future version.
sys_executable = CommandSpec._sys_executable()


class WindowsCommandSpec(CommandSpec):
    split_args = dict(posix=False)


class ScriptWriter:
    """
    Encapsulates behavior around writing entry point scripts for console and
    gui apps.
    """

    template = textwrap.dedent(r"""
        # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r
        import re
        import sys

        # for compatibility with easy_install; see #2198
        __requires__ = %(spec)r

        try:
            from importlib.metadata import distribution
        except ImportError:
            try:
                from importlib_metadata import distribution
            except ImportError:
                from pkg_resources import load_entry_point


        def importlib_load_entry_point(spec, group, name):
            dist_name, _, _ = spec.partition('==')
            matches = (
                entry_point
                for entry_point in distribution(dist_name).entry_points
                if entry_point.group == group and entry_point.name == name
            )
            return next(matches).load()


        globals().setdefault('load_entry_point', importlib_load_entry_point)


        if __name__ == '__main__':
            sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
            sys.exit(load_entry_point(%(spec)r, %(group)r, %(name)r)())
        """).lstrip()

    command_spec_class = CommandSpec

    @classmethod
    def get_script_args(cls, dist, executable=None, wininst=False):
        # for backward compatibility
        warnings.warn("Use get_args", EasyInstallDeprecationWarning)
        writer = (WindowsScriptWriter if wininst else ScriptWriter).best()
        header = cls.get_script_header("", executable, wininst)
        return writer.get_args(dist, header)

    @classmethod
    def get_script_header(cls, script_text, executable=None, wininst=False):
        # for backward compatibility
        warnings.warn(
            "Use get_header", EasyInstallDeprecationWarning, stacklevel=2)
        if wininst:
            executable = "python.exe"
        return cls.get_header(script_text, executable)

    @classmethod
    def get_args(cls, dist, header=None):
        """
        Yield write_script() argument tuples for a distribution's
        console_scripts and gui_scripts entry points.
        """
        if header is None:
            header = cls.get_header()
        spec = str(dist.as_requirement())
        for type_ in 'console', 'gui':
            group = type_ + '_scripts'
            for name, ep in dist.get_entry_map(group).items():
                cls._ensure_safe_name(name)
                script_text = cls.template % locals()
                args = cls._get_script_args(type_, name, header, script_text)
                for res in args:
                    yield res

    @staticmethod
    def _ensure_safe_name(name):
        """
        Prevent paths in *_scripts entry point names.
        """
        has_path_sep = re.search(r'[\\/]', name)
        if has_path_sep:
            raise ValueError("Path separators not allowed in script names")

    @classmethod
    def get_writer(cls, force_windows):
        # for backward compatibility
        warnings.warn("Use best", EasyInstallDeprecationWarning)
        return WindowsScriptWriter.best() if force_windows else cls.best()

    @classmethod
    def best(cls):
        """
        Select the best ScriptWriter for this environment.
        """
        if sys.platform == 'win32' or (os.name == 'java' and os._name == 'nt'):
            return WindowsScriptWriter.best()
        else:
            return cls

    @classmethod
    def _get_script_args(cls, type_, name, header, script_text):
        # Simply write the stub with no extension.
        yield (name, header + script_text)

    @classmethod
    def get_header(cls, script_text="", executable=None):
        """Create a #! line, getting options (if any) from script_text"""
        cmd = cls.command_spec_class.best().from_param(executable)
        cmd.install_options(script_text)
        return cmd.as_header()


class WindowsScriptWriter(ScriptWriter):
    command_spec_class = WindowsCommandSpec

    @classmethod
    def get_writer(cls):
        # for backward compatibility
        warnings.warn("Use best", EasyInstallDeprecationWarning)
        return cls.best()

    @classmethod
    def best(cls):
        """
        Select the best ScriptWriter suitable for Windows
        """
        writer_lookup = dict(
            executable=WindowsExecutableLauncherWriter,
            natural=cls,
        )
        # for compatibility, use the executable launcher by default
        launcher = os.environ.get('SETUPTOOLS_LAUNCHER', 'executable')
        return writer_lookup[launcher]

    @classmethod
    def _get_script_args(cls, type_, name, header, script_text):
        "For Windows, add a .py extension"
        ext = dict(console='.pya', gui='.pyw')[type_]
        if ext not in os.environ['PATHEXT'].lower().split(';'):
            msg = (
                "{ext} not listed in PATHEXT; scripts will not be "
                "recognized as executables."
            ).format(**locals())
            warnings.warn(msg, UserWarning)
        old = ['.pya', '.py', '-script.py', '.pyc', '.pyo', '.pyw', '.exe']
        old.remove(ext)
        header = cls._adjust_header(type_, header)
        blockers = [name + x for x in old]
        yield name + ext, header + script_text, 't', blockers

    @classmethod
    def _adjust_header(cls, type_, orig_header):
        """
        Make sure 'pythonw' is used for gui and 'python' is used for
        console (regardless of what sys.executable is).
        """
        pattern = 'pythonw.exe'
        repl = 'python.exe'
        if type_ == 'gui':
            pattern, repl = repl, pattern
        pattern_ob = re.compile(re.escape(pattern), re.IGNORECASE)
        new_header = pattern_ob.sub(string=orig_header, repl=repl)
        return new_header if cls._use_header(new_header) else orig_header

    @staticmethod
    def _use_header(new_header):
        """
        Should _adjust_header use the replaced header?

        On non-windows systems, always use. On
        Windows systems, only use the replaced header if it resolves
        to an executable on the system.
        """
        clean_header = new_header[2:-1].strip('"')
        return sys.platform != 'win32' or find_executable(clean_header)


class WindowsExecutableLauncherWriter(WindowsScriptWriter):
    @classmethod
    def _get_script_args(cls, type_, name, header, script_text):
        """
        For Windows, add a .py extension and an .exe launcher
        """
        if type_ == 'gui':
            launcher_type = 'gui'
            ext = '-script.pyw'
            old = ['.pyw']
        else:
            launcher_type = 'cli'
            ext = '-script.py'
            old = ['.py', '.pyc', '.pyo']
        hdr = cls._adjust_header(type_, header)
        blockers = [name + x for x in old]
        yield (name + ext, hdr + script_text, 't', blockers)
        yield (
            name + '.exe', get_win_launcher(launcher_type),
            'b'  # write in binary mode
        )
        if not is_64bit():
            # install a manifest for the launcher to prevent Windows
            # from detecting it as an installer (which it will for
            #  launchers like easy_install.exe). Consider only
            #  adding a manifest for launchers detected as installers.
            #  See Distribute #143 for details.
            m_name = name + '.exe.manifest'
            yield (m_name, load_launcher_manifest(name), 't')


# for backward-compatibility
get_script_args = ScriptWriter.get_script_args
get_script_header = ScriptWriter.get_script_header


def get_win_launcher(type):
    """
    Load the Windows launcher (executable) suitable for launching a script.

    `type` should be either 'cli' or 'gui'

    Returns the executable as a byte string.
    """
    launcher_fn = '%s.exe' % type
    if is_64bit():
        if get_platform() == "win-arm64":
            launcher_fn = launcher_fn.replace(".", "-arm64.")
        else:
            launcher_fn = launcher_fn.replace(".", "-64.")
    else:
        launcher_fn = launcher_fn.replace(".", "-32.")
    return resource_string('setuptools', launcher_fn)


def load_launcher_manifest(name):
    manifest = pkg_resources.resource_string(__name__, 'launcher manifest.xml')
    return manifest.decode('utf-8') % vars()


def rmtree(path, ignore_errors=False, onerror=auto_chmod):
    return shutil.rmtree(path, ignore_errors, onerror)


def current_umask():
    tmp = os.umask(0o022)
    os.umask(tmp)
    return tmp


class EasyInstallDeprecationWarning(SetuptoolsDeprecationWarning):
    """
    Warning for EasyInstall deprecations, bypassing suppression.
    """
site-packages/setuptools/command/rotate.py000064400000004120150564745010014777 0ustar00from distutils.util import convert_path
from distutils import log
from distutils.errors import DistutilsOptionError
import os
import shutil

from setuptools import Command


class rotate(Command):
    """Delete older distributions"""

    description = "delete older distributions, keeping N newest files"
    user_options = [
        ('match=', 'm', "patterns to match (required)"),
        ('dist-dir=', 'd', "directory where the distributions are"),
        ('keep=', 'k', "number of matching distributions to keep"),
    ]

    boolean_options = []

    def initialize_options(self):
        self.match = None
        self.dist_dir = None
        self.keep = None

    def finalize_options(self):
        if self.match is None:
            raise DistutilsOptionError(
                "Must specify one or more (comma-separated) match patterns "
                "(e.g. '.zip' or '.egg')"
            )
        if self.keep is None:
            raise DistutilsOptionError("Must specify number of files to keep")
        try:
            self.keep = int(self.keep)
        except ValueError as e:
            raise DistutilsOptionError("--keep must be an integer") from e
        if isinstance(self.match, str):
            self.match = [
                convert_path(p.strip()) for p in self.match.split(',')
            ]
        self.set_undefined_options('bdist', ('dist_dir', 'dist_dir'))

    def run(self):
        self.run_command("egg_info")
        from glob import glob

        for pattern in self.match:
            pattern = self.distribution.get_name() + '*' + pattern
            files = glob(os.path.join(self.dist_dir, pattern))
            files = [(os.path.getmtime(f), f) for f in files]
            files.sort()
            files.reverse()

            log.info("%d file(s) matching %s", len(files), pattern)
            files = files[self.keep:]
            for (t, f) in files:
                log.info("Deleting %s", f)
                if not self.dry_run:
                    if os.path.isdir(f):
                        shutil.rmtree(f)
                    else:
                        os.unlink(f)
site-packages/setuptools/command/upload.py000064400000000716150564745010014774 0ustar00from distutils import log
from distutils.command import upload as orig

from setuptools.errors import RemovedCommandError


class upload(orig.upload):
    """Formerly used to upload packages to PyPI."""

    def run(self):
        msg = (
            "The upload command has been removed, use twine to upload "
            + "instead (https://pypi.org/p/twine)"
        )

        self.announce("ERROR: " + msg, log.ERROR)
        raise RemovedCommandError(msg)
site-packages/setuptools/command/launcher manifest.xml000064400000001164150564745010017246 0ustar00<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
    <assemblyIdentity version="1.0.0.0"
                      processorArchitecture="X86"
                      name="%(name)s"
                      type="win32"/>
    <!-- Identify the application security requirements. -->
    <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
        <security>
            <requestedPrivileges>
                <requestedExecutionLevel level="asInvoker" uiAccess="false"/>
            </requestedPrivileges>
        </security>
    </trustInfo>
</assembly>
site-packages/setuptools/command/build_ext.py000064400000031634150564745010015472 0ustar00import os
import sys
import itertools
from importlib.machinery import EXTENSION_SUFFIXES
from distutils.command.build_ext import build_ext as _du_build_ext
from distutils.file_util import copy_file
from distutils.ccompiler import new_compiler
from distutils.sysconfig import customize_compiler, get_config_var
from distutils.errors import DistutilsError
from distutils import log

from setuptools.extension import Library

try:
    # Attempt to use Cython for building extensions, if available
    from Cython.Distutils.build_ext import build_ext as _build_ext
    # Additionally, assert that the compiler module will load
    # also. Ref #1229.
    __import__('Cython.Compiler.Main')
except ImportError:
    _build_ext = _du_build_ext

# make sure _config_vars is initialized
get_config_var("LDSHARED")
from distutils.sysconfig import _config_vars as _CONFIG_VARS  # noqa


def _customize_compiler_for_shlib(compiler):
    if sys.platform == "darwin":
        # building .dylib requires additional compiler flags on OSX; here we
        # temporarily substitute the pyconfig.h variables so that distutils'
        # 'customize_compiler' uses them before we build the shared libraries.
        tmp = _CONFIG_VARS.copy()
        try:
            # XXX Help!  I don't have any idea whether these are right...
            _CONFIG_VARS['LDSHARED'] = (
                "gcc -Wl,-x -dynamiclib -undefined dynamic_lookup")
            _CONFIG_VARS['CCSHARED'] = " -dynamiclib"
            _CONFIG_VARS['SO'] = ".dylib"
            customize_compiler(compiler)
        finally:
            _CONFIG_VARS.clear()
            _CONFIG_VARS.update(tmp)
    else:
        customize_compiler(compiler)


have_rtld = False
use_stubs = False
libtype = 'shared'

if sys.platform == "darwin":
    use_stubs = True
elif os.name != 'nt':
    try:
        import dl
        use_stubs = have_rtld = hasattr(dl, 'RTLD_NOW')
    except ImportError:
        pass


def if_dl(s):
    return s if have_rtld else ''


def get_abi3_suffix():
    """Return the file extension for an abi3-compliant Extension()"""
    for suffix in EXTENSION_SUFFIXES:
        if '.abi3' in suffix:  # Unix
            return suffix
        elif suffix == '.pyd':  # Windows
            return suffix


class build_ext(_build_ext):
    def run(self):
        """Build extensions in build directory, then copy if --inplace"""
        old_inplace, self.inplace = self.inplace, 0
        _build_ext.run(self)
        self.inplace = old_inplace
        if old_inplace:
            self.copy_extensions_to_source()

    def copy_extensions_to_source(self):
        build_py = self.get_finalized_command('build_py')
        for ext in self.extensions:
            fullname = self.get_ext_fullname(ext.name)
            filename = self.get_ext_filename(fullname)
            modpath = fullname.split('.')
            package = '.'.join(modpath[:-1])
            package_dir = build_py.get_package_dir(package)
            dest_filename = os.path.join(package_dir,
                                         os.path.basename(filename))
            src_filename = os.path.join(self.build_lib, filename)

            # Always copy, even if source is older than destination, to ensure
            # that the right extensions for the current Python/platform are
            # used.
            copy_file(
                src_filename, dest_filename, verbose=self.verbose,
                dry_run=self.dry_run
            )
            if ext._needs_stub:
                self.write_stub(package_dir or os.curdir, ext, True)

    def get_ext_filename(self, fullname):
        so_ext = os.getenv('SETUPTOOLS_EXT_SUFFIX')
        if so_ext:
            filename = os.path.join(*fullname.split('.')) + so_ext
        else:
            filename = _build_ext.get_ext_filename(self, fullname)
            so_ext = get_config_var('EXT_SUFFIX')

        if fullname in self.ext_map:
            ext = self.ext_map[fullname]
            use_abi3 = getattr(ext, 'py_limited_api') and get_abi3_suffix()
            if use_abi3:
                filename = filename[:-len(so_ext)]
                so_ext = get_abi3_suffix()
                filename = filename + so_ext
            if isinstance(ext, Library):
                fn, ext = os.path.splitext(filename)
                return self.shlib_compiler.library_filename(fn, libtype)
            elif use_stubs and ext._links_to_dynamic:
                d, fn = os.path.split(filename)
                return os.path.join(d, 'dl-' + fn)
        return filename

    def initialize_options(self):
        _build_ext.initialize_options(self)
        self.shlib_compiler = None
        self.shlibs = []
        self.ext_map = {}

    def finalize_options(self):
        _build_ext.finalize_options(self)
        self.extensions = self.extensions or []
        self.check_extensions_list(self.extensions)
        self.shlibs = [ext for ext in self.extensions
                       if isinstance(ext, Library)]
        if self.shlibs:
            self.setup_shlib_compiler()
        for ext in self.extensions:
            ext._full_name = self.get_ext_fullname(ext.name)
        for ext in self.extensions:
            fullname = ext._full_name
            self.ext_map[fullname] = ext

            # distutils 3.1 will also ask for module names
            # XXX what to do with conflicts?
            self.ext_map[fullname.split('.')[-1]] = ext

            ltd = self.shlibs and self.links_to_dynamic(ext) or False
            ns = ltd and use_stubs and not isinstance(ext, Library)
            ext._links_to_dynamic = ltd
            ext._needs_stub = ns
            filename = ext._file_name = self.get_ext_filename(fullname)
            libdir = os.path.dirname(os.path.join(self.build_lib, filename))
            if ltd and libdir not in ext.library_dirs:
                ext.library_dirs.append(libdir)
            if ltd and use_stubs and os.curdir not in ext.runtime_library_dirs:
                ext.runtime_library_dirs.append(os.curdir)

    def setup_shlib_compiler(self):
        compiler = self.shlib_compiler = new_compiler(
            compiler=self.compiler, dry_run=self.dry_run, force=self.force
        )
        _customize_compiler_for_shlib(compiler)

        if self.include_dirs is not None:
            compiler.set_include_dirs(self.include_dirs)
        if self.define is not None:
            # 'define' option is a list of (name,value) tuples
            for (name, value) in self.define:
                compiler.define_macro(name, value)
        if self.undef is not None:
            for macro in self.undef:
                compiler.undefine_macro(macro)
        if self.libraries is not None:
            compiler.set_libraries(self.libraries)
        if self.library_dirs is not None:
            compiler.set_library_dirs(self.library_dirs)
        if self.rpath is not None:
            compiler.set_runtime_library_dirs(self.rpath)
        if self.link_objects is not None:
            compiler.set_link_objects(self.link_objects)

        # hack so distutils' build_extension() builds a library instead
        compiler.link_shared_object = link_shared_object.__get__(compiler)

    def get_export_symbols(self, ext):
        if isinstance(ext, Library):
            return ext.export_symbols
        return _build_ext.get_export_symbols(self, ext)

    def build_extension(self, ext):
        ext._convert_pyx_sources_to_lang()
        _compiler = self.compiler
        try:
            if isinstance(ext, Library):
                self.compiler = self.shlib_compiler
            _build_ext.build_extension(self, ext)
            if ext._needs_stub:
                cmd = self.get_finalized_command('build_py').build_lib
                self.write_stub(cmd, ext)
        finally:
            self.compiler = _compiler

    def links_to_dynamic(self, ext):
        """Return true if 'ext' links to a dynamic lib in the same package"""
        # XXX this should check to ensure the lib is actually being built
        # XXX as dynamic, and not just using a locally-found version or a
        # XXX static-compiled version
        libnames = dict.fromkeys([lib._full_name for lib in self.shlibs])
        pkg = '.'.join(ext._full_name.split('.')[:-1] + [''])
        return any(pkg + libname in libnames for libname in ext.libraries)

    def get_outputs(self):
        return _build_ext.get_outputs(self) + self.__get_stubs_outputs()

    def __get_stubs_outputs(self):
        # assemble the base name for each extension that needs a stub
        ns_ext_bases = (
            os.path.join(self.build_lib, *ext._full_name.split('.'))
            for ext in self.extensions
            if ext._needs_stub
        )
        # pair each base with the extension
        pairs = itertools.product(ns_ext_bases, self.__get_output_extensions())
        return list(base + fnext for base, fnext in pairs)

    def __get_output_extensions(self):
        yield '.py'
        yield '.pyc'
        if self.get_finalized_command('build_py').optimize:
            yield '.pyo'

    def write_stub(self, output_dir, ext, compile=False):
        log.info("writing stub loader for %s to %s", ext._full_name,
                 output_dir)
        stub_file = (os.path.join(output_dir, *ext._full_name.split('.')) +
                     '.py')
        if compile and os.path.exists(stub_file):
            raise DistutilsError(stub_file + " already exists! Please delete.")
        if not self.dry_run:
            f = open(stub_file, 'w')
            f.write(
                '\n'.join([
                    "def __bootstrap__():",
                    "   global __bootstrap__, __file__, __loader__",
                    "   import sys, os, pkg_resources, importlib.util" +
                    if_dl(", dl"),
                    "   __file__ = pkg_resources.resource_filename"
                    "(__name__,%r)"
                    % os.path.basename(ext._file_name),
                    "   del __bootstrap__",
                    "   if '__loader__' in globals():",
                    "       del __loader__",
                    if_dl("   old_flags = sys.getdlopenflags()"),
                    "   old_dir = os.getcwd()",
                    "   try:",
                    "     os.chdir(os.path.dirname(__file__))",
                    if_dl("     sys.setdlopenflags(dl.RTLD_NOW)"),
                    "     spec = importlib.util.spec_from_file_location(",
                    "                __name__, __file__)",
                    "     mod = importlib.util.module_from_spec(spec)",
                    "     spec.loader.exec_module(mod)",
                    "   finally:",
                    if_dl("     sys.setdlopenflags(old_flags)"),
                    "     os.chdir(old_dir)",
                    "__bootstrap__()",
                    ""  # terminal \n
                ])
            )
            f.close()
        if compile:
            from distutils.util import byte_compile

            byte_compile([stub_file], optimize=0,
                         force=True, dry_run=self.dry_run)
            optimize = self.get_finalized_command('install_lib').optimize
            if optimize > 0:
                byte_compile([stub_file], optimize=optimize,
                             force=True, dry_run=self.dry_run)
            if os.path.exists(stub_file) and not self.dry_run:
                os.unlink(stub_file)


if use_stubs or os.name == 'nt':
    # Build shared libraries
    #
    def link_shared_object(
            self, objects, output_libname, output_dir=None, libraries=None,
            library_dirs=None, runtime_library_dirs=None, export_symbols=None,
            debug=0, extra_preargs=None, extra_postargs=None, build_temp=None,
            target_lang=None):
        self.link(
            self.SHARED_LIBRARY, objects, output_libname,
            output_dir, libraries, library_dirs, runtime_library_dirs,
            export_symbols, debug, extra_preargs, extra_postargs,
            build_temp, target_lang
        )
else:
    # Build static libraries everywhere else
    libtype = 'static'

    def link_shared_object(
            self, objects, output_libname, output_dir=None, libraries=None,
            library_dirs=None, runtime_library_dirs=None, export_symbols=None,
            debug=0, extra_preargs=None, extra_postargs=None, build_temp=None,
            target_lang=None):
        # XXX we need to either disallow these attrs on Library instances,
        # or warn/abort here if set, or something...
        # libraries=None, library_dirs=None, runtime_library_dirs=None,
        # export_symbols=None, extra_preargs=None, extra_postargs=None,
        # build_temp=None

        assert output_dir is None  # distutils build_ext doesn't pass this
        output_dir, filename = os.path.split(output_libname)
        basename, ext = os.path.splitext(filename)
        if self.library_filename("x").startswith('lib'):
            # strip 'lib' prefix; this is kludgy if some platform uses
            # a different prefix
            basename = basename[3:]

        self.create_static_lib(
            objects, basename, output_dir, debug, target_lang
        )
site-packages/setuptools/command/__pycache__/upload_docs.cpython-39.pyc000064400000014101150564745010022264 0ustar00a

(�Re2�@s�dZddlmZddlmZddlmZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlmZddlmZd	d
�ZGdd�de�ZdS)
z|upload_docs

Implements a Distutils 'upload_docs' subcommand (upload documentation to
sites other than PyPi such as devpi).
�)�standard_b64encode)�log)�DistutilsOptionErrorN)�iter_entry_points�)�uploadcCs|�dd�S)Nzutf-8�surrogateescape)�encode)�s�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/upload_docs.py�_encodesr
c@s�eZdZdZdZdddejfddgZejZdd	�Zd
efgZ	dd�Z
d
d�Zdd�Zdd�Z
edd��Zedd��Zdd�ZdS)�upload_docszhttps://pypi.python.org/pypi/z;Upload documentation to sites other than PyPi such as devpizrepository=�rzurl of repository [default: %s])z
show-responseNz&display full response text from server)zupload-dir=Nzdirectory to uploadcCs"|jdurtdd�D]}dSdS)Nzdistutils.commands�build_sphinxT)�
upload_dirr)�self�eprrr�
has_sphinx-s
zupload_docs.has_sphinxrcCst�|�d|_d|_dS)N)r�initialize_optionsr�
target_dir)rrrrr4s
zupload_docs.initialize_optionscCs�t�|�|jdurV|��r8|�d�}t|j�d|_qh|�d�}tj	�
|jd�|_n|�d�|j|_d|j
vr|t�d�|�d|j�dS)	Nr�html�buildZdocsrzpypi.python.orgz<Upload_docs command is deprecated for PyPi. Use RTD instead.zUsing upload directory %s)r�finalize_optionsrr�get_finalized_command�dictZbuilder_target_dirsr�os�path�join�
build_base�ensure_dirname�
repositoryr�warn�announce)rrrrrrr9s






zupload_docs.finalize_optionscCs�t�|d�}z�|�|j�t�|j�D]x\}}}||jkrP|sPd}t||j��|D]H}tj�||�}|t	|j�d��
tjj�}	tj�|	|�}
|�||
�qTq&W|�
�n
|�
�0dS)N�wz'no files found in upload directory '%s')�zipfile�ZipFile�mkpathrr�walkrrr�len�lstrip�sep�write�close)r�filename�zip_file�root�dirs�files�tmpl�name�full�relative�destrrr�create_zipfileIszupload_docs.create_zipfilec	Csr|��D]}|�|�qt��}|jj��}tj�	|d|�}z"|�
|�|�|�Wt�
|�nt�
|�0dS)Nz%s.zip)�get_sub_commands�run_command�tempfile�mkdtemp�distribution�metadata�get_namerrrr8�upload_file�shutil�rmtree)r�cmd_name�tmp_dirr4r/rrr�runYs
zupload_docs.runccs�|\}}d|}t|t�s |g}|D]f}t|t�rL|d|d7}|d}nt|�}|Vt|�VdV|V|r$|dd�dkr$dVq$dS)	Nz*
Content-Disposition: form-data; name="%s"z; filename="%s"rrs

����
�
)�
isinstance�list�tupler
)�item�sep_boundary�key�values�title�valuerrr�_build_partgs



zupload_docs._build_partcCsnd}d|�d�}|d}|df}tj|j|d�}t||���}tj�|�}t�||�}	d|}
d�	|	�|
fS)	z=
        Build up the MIME payload for the POST data
        z3--------------GHSKFJDLGDS7543FJKLFHRE75642756743254s
--�asciis--rH)rMz multipart/form-data; boundary=%s�)
r	�	functools�partialrR�map�items�	itertools�chain�
from_iterabler)�cls�data�boundaryrMZend_boundaryZ	end_itemsZbuilderZpart_groups�partsZ
body_items�content_typerrr�_build_multipart{s�zupload_docs._build_multipartc
CsVt|d��}|��}Wd�n1s(0Y|jj}d|��tj�|�|fd�}t|j	d|j
�}t|��d�}d|}|�
|�\}}	d|j}
|�|
tj�tj�|j�\}}}
}}}|s�|s�|r�J�|dkr�tj�|�}n"|d	kr�tj�|�}ntd
|��d}zZ|��|�d|
�|	}|�d
|�|�dtt|���|�d|�|��|� |�Wn<t!j"�y�}z |�t|�tj#�WYd}~dSd}~00|�$�}|j%dk�r�d|j%|j&f}
|�|
tj�nb|j%dk�r|�'d�}|du�rd|��}d|}
|�|
tj�nd|j%|j&f}
|�|
tj#�|j(�rRt)d|��d�dS)N�rbZ
doc_upload)z:actionr4�content�:rSzBasic zSubmitting documentation to %s�http�httpszunsupported schema ��POSTzContent-typezContent-length�
Authorization��zServer response (%s): %si-ZLocationzhttps://pythonhosted.org/%s/zUpload successful. Visit %szUpload failed (%s): %szK---------------------------------------------------------------------------)*�open�readr=r>r?rr�basenamer
�username�passwordr�decoderar!r#r�INFO�urllib�parse�urlparsere�client�HTTPConnection�HTTPSConnection�AssertionError�connect�
putrequest�	putheader�strr)�
endheaders�send�socket�error�ERROR�getresponse�status�reason�	getheader�
show_response�print)rr.�frc�metar]�credentials�auth�body�ct�msgZschema�netloc�url�params�query�	fragments�connr`�er�locationrrrr@�s`&�
�

zupload_docs.upload_fileN)�__name__�
__module__�__qualname__�DEFAULT_REPOSITORY�descriptionr�user_options�boolean_optionsr�sub_commandsrrr8rE�staticmethodrR�classmethodrar@rrrrrs(��


r)�__doc__�base64r�	distutilsr�distutils.errorsrrrr%r;rArYrU�http.clientre�urllib.parserr�
pkg_resourcesrrr
rrrrr�<module>s site-packages/setuptools/command/__pycache__/egg_info.cpython-39.pyc000064400000053012150564745010021551 0ustar00a

(�Re�b�@s�dZddlmZddlmZddlmZddlm	Z	ddlZddlZddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlmZddlmZdd	lmZdd
lmZddlmZddlmZmZmZmZm Z m!Z!m"Z"m#Z#ddl$m%Z%dd
l&m'Z'ddl(m)Z)ddlm*Z*dd�Z+Gdd�d�Z,Gdd�de,e�Z-Gdd�de�ZGdd�de�Z.dd�Z/dd�Z0dd�Z1d d!�Z2d"d#�Z3d$d%�Z4d&d'�Z5d(d)�Z6d3d+d,�Z7d-d.�Z8d/d0�Z9Gd1d2�d2e*�Z:dS)4zUsetuptools.command.egg_info

Create a distribution's .egg-info directory and contents�)�FileList)�DistutilsInternalError)�convert_path)�logN)�Command)�sdist)�walk_revctrl)�edit_config)�	bdist_egg)�parse_requirements�	safe_name�
parse_version�safe_version�yield_lines�
EntryPoint�iter_entry_points�to_filename)�glob)�	packaging)�SetuptoolsDeprecationWarningcCs�d}|�tjj�}t�tj�}d|f}t|�D�]�\}}|t|�dk}|dkrr|r`|d7}q0|d||f7}q0d}t|�}	||	k�r�||}
|
dkr�||d7}�n|
d	kr�||7}n�|
d
k�r�|d}||	kr�||dkr�|d}||	k�r||dk�r|d}||	k�r,||dk�r,|d}�q||	k�rF|t�|
�7}nR||d|�}d}
|ddk�rxd
}
|dd�}|
t�|�7}
|d|
f7}|}n|t�|
�7}|d7}q~|s0||7}q0|d7}tj|tj	tj
Bd�S)z�
    Translate a file path glob like '*.txt' in to a regular expression.
    This differs from fnmatch.translate which allows wildcards to match
    directory separators. It also knows about '**/' which matches any number of
    directories.
    �z[^%s]��**z.*z
(?:%s+%s)*r�*�?�[�!�]�^Nz[%s]z\Z)�flags)�split�os�path�sep�re�escape�	enumerate�len�compile�	MULTILINE�DOTALL)r�pat�chunksr#Z
valid_char�c�chunk�
last_chunk�iZ	chunk_len�charZinner_i�inner�
char_class�r4��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/egg_info.py�translate_pattern#sV







r6c@s@eZdZdZdZedd��Zdd�Zdd�Zdd	�Z	ee	�Z
dS)
�
InfoCommonNcCst|j���S�N)r�distribution�get_name��selfr4r4r5�namezszInfoCommon.namecCst|�|j����Sr8)r�
_maybe_tagr9�get_versionr;r4r4r5�tagged_version~szInfoCommon.tagged_versioncCs |jr|�|j�r|S||jS)z�
        egg_info may be called more than once for a distribution,
        in which case the version string already contains all tags.
        )�vtags�endswith�r<�versionr4r4r5r>�s�zInfoCommon._maybe_tagcCs,d}|jr||j7}|jr(|t�d�7}|S)Nrz-%Y%m%d)�	tag_build�tag_date�time�strftimerCr4r4r5�tags�s
zInfoCommon.tags)�__name__�
__module__�__qualname__rErF�propertyr=r@r>rIrAr4r4r4r5r7vs

r7c@s�eZdZdZgd�ZdgZddiZdd�Zedd��Z	e	j
d	d��Z	d
d�Zdd
�Zddd�Z
dd�Zdd�Zdd�Zdd�Zdd�ZdS)�egg_infoz+create a distribution's .egg-info directory))z	egg-base=�ezLdirectory containing .egg-info directories (default: top of the source tree))�tag-date�dz0Add date stamp (e.g. 20050528) to version number)z
tag-build=�bz-Specify explicit tag to add to version number)�no-date�Dz"Don't include date stamp [default]rPrScCs"d|_d|_d|_d|_d|_dS)NF)�egg_base�egg_namerN�egg_version�broken_egg_infor;r4r4r5�initialize_options�s
zegg_info.initialize_optionscCsdSr8r4r;r4r4r5�tag_svn_revision�szegg_info.tag_svn_revisioncCsdSr8r4)r<�valuer4r4r5rZ�scCs0t��}|��|d<d|d<t|t|d��dS)z�
        Materialize the value of date into the
        build tag. Install build keys in a deterministic order
        to avoid arbitrary reordering on subsequent builds.
        rErrF)rNN)�collections�OrderedDictrIr	�dict)r<�filenamerNr4r4r5�save_version_info�szegg_info.save_version_infoc
CsV|j|_|��|_t|j�}z6t|tjj�}|r4dnd}t	t
||j|jf��Wn>ty�}z&tj
�d|j|jf�|�WYd}~n
d}~00|jdur�|jj}|p�i�dtj�|_|�d�t|j�d|_|jtjkr�tj�|j|j�|_d|jv�r|��|j|jj_|jj}|du�rR|j|j��k�rR|j|_t|j�|_ d|j_dS)Nz%s==%sz%s===%sz2Invalid distribution name or version syntax: %s-%srrU�	.egg-info�-)!r=rVr@rWr
�
isinstancerrD�Version�listr�
ValueError�	distutils�errors�DistutilsOptionErrorrUr9�package_dir�getr!�curdir�ensure_dirnamerrNr"�join�check_broken_egg_info�metadataZ
_patched_dist�key�lower�_version�_parsed_version)r<�parsed_versionZ
is_version�specrO�dirs�pdr4r4r5�finalize_options�sB


��
���

zegg_info.finalize_optionsFcCsL|r|�|||�n4tj�|�rH|dur>|s>t�d||�dS|�|�dS)a�Write `data` to `filename` or delete if empty

        If `data` is non-empty, this routine is the same as ``write_file()``.
        If `data` is empty but not ``None``, this is the same as calling
        ``delete_file(filename)`.  If `data` is ``None``, then this is a no-op
        unless `filename` exists, in which case a warning is issued about the
        orphaned file (if `force` is false), or deleted (if `force` is true).
        Nz$%s not set in setup(), but %s exists)�
write_filer!r"�existsr�warn�delete_file)r<�whatr_�data�forcer4r4r5�write_or_delete_file�s	�zegg_info.write_or_delete_filecCs>t�d||�|�d�}|js:t|d�}|�|�|��dS)z�Write `data` to `filename` (if not a dry run) after announcing it

        `what` is used in a log message to identify what is being written
        to the file.
        zwriting %s to %s�utf-8�wbN)r�info�encode�dry_run�open�write�close)r<r~r_r�fr4r4r5rz
s


zegg_info.write_filecCs t�d|�|jst�|�dS)z8Delete `filename` (if not a dry run) after announcing itzdeleting %sN)rr�r�r!�unlink)r<r_r4r4r5r}szegg_info.delete_filecCs�|�|j�t�|jd�|jj}td�D]4}|j|d�|��}|||j	tj
�|j|j	��q*tj
�|jd�}tj
�|�r�|�
|�|��dS)Nzegg_info.writers)�	installerznative_libs.txt)�mkpathrNr!�utimer9Zfetch_build_eggr�require�resolver=r"rnr{r}�find_sources)r<r��ep�writer�nlr4r4r5�runs
zegg_info.runcCs4tj�|jd�}t|j�}||_|��|j|_dS)z"Generate SOURCES.txt manifest filezSOURCES.txtN)	r!r"rnrN�manifest_makerr9�manifestr��filelist)r<Zmanifest_filename�mmr4r4r5r�-s

zegg_info.find_sourcescCsT|jd}|jtjkr&tj�|j|�}tj�|�rPt�d||j	�|j	|_
||_	dS)NraaB------------------------------------------------------------------------------
Note: Your current .egg-info directory has a '-' in its name;
this will not work correctly with "setup.py develop".

Please rename %s to %s to correct this problem.
------------------------------------------------------------------------------)rVrUr!rlr"rnr{rr|rNrX)r<Zbeir4r4r5ro5s
�zegg_info.check_broken_egg_infoN)F)rJrKrL�description�user_options�boolean_options�negative_optrYrMrZ�setterr`ryr�rzr}r�r�ror4r4r4r5rN�s$�



1

rNc@s|eZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�ZdS)rc
	Cs
|�|�\}}}}|j|j|j|jt�|j|�t�|j|�|j	|j
d�}dddddddd	d�}z||}Wn"ty�td
j
|d���Yn0|�d�}	|d
vr�|g}|	r�|fnd}
||}|�d�|g|	r�|gng|��|D] }||�s�tj||g|
�R�q�dS)N)�include�excludezglobal-includezglobal-excludezrecursive-includezrecursive-exclude�graft�prunez%warning: no files found matching '%s'z9warning: no previously-included files found matching '%s'z>warning: no files found matching '%s' anywhere in distributionzRwarning: no previously-included files matching '%s' found anywhere in distributionz:warning: no files found matching '%s' under directory '%s'zNwarning: no previously-included files matching '%s' found under directory '%s'z+warning: no directories found matching '%s'z6no previously-included directories found matching '%s'z/this cannot happen: invalid action '{action!s}')�actionz
recursive->r�r�r4� )Z_parse_template_liner�r��global_include�global_exclude�	functools�partial�recursive_include�recursive_excluder�r��KeyErrorr�format�
startswith�debug_printrnrr|)
r<�liner��patterns�dirZdir_patternZ
action_mapZlog_mapZprocess_actionZaction_is_recursiveZextra_log_argsZlog_tmpl�patternr4r4r5�process_template_lineHs`������
	
����zFileList.process_template_linecCsRd}tt|j�ddd�D]2}||j|�r|�d|j|�|j|=d}q|S)z�
        Remove all files from the file list that match the predicate.
        Return True if any matching files were removed
        Fr���z
 removing T)�ranger'�filesr�)r<�	predicate�foundr0r4r4r5�
_remove_files�szFileList._remove_filescCs$dd�t|�D�}|�|�t|�S)z#Include files that match 'pattern'.cSsg|]}tj�|�s|�qSr4�r!r"�isdir��.0r�r4r4r5�
<listcomp>��z$FileList.include.<locals>.<listcomp>�r�extend�bool�r<r�r�r4r4r5r��s
zFileList.includecCst|�}|�|j�S)z#Exclude files that match 'pattern'.)r6r��match�r<r�r�r4r4r5r��szFileList.excludecCs8tj�|d|�}dd�t|dd�D�}|�|�t|�S)zN
        Include all files anywhere in 'dir/' that match the pattern.
        rcSsg|]}tj�|�s|�qSr4r�r�r4r4r5r��s�z.FileList.recursive_include.<locals>.<listcomp>T)�	recursive)r!r"rnrr�r�)r<r�r�Zfull_patternr�r4r4r5r��s
zFileList.recursive_includecCs ttj�|d|��}|�|j�S)zM
        Exclude any file anywhere in 'dir/' that match the pattern.
        r�r6r!r"rnr�r�)r<r�r�r�r4r4r5r��szFileList.recursive_excludecCs$dd�t|�D�}|�|�t|�S)zInclude all files from 'dir/'.cSs"g|]}tj�|�D]}|�qqSr4)rgr��findall)r�Z	match_dir�itemr4r4r5r��s�z"FileList.graft.<locals>.<listcomp>r�)r<r�r�r4r4r5r��s
�
zFileList.graftcCsttj�|d��}|�|j�S)zFilter out files from 'dir/'.rr�)r<r�r�r4r4r5r��szFileList.prunecsJ|jdur|��ttj�d|����fdd�|jD�}|�|�t|�S)z�
        Include all files anywhere in the current directory that match the
        pattern. This is very inefficient on large file trees.
        Nrcsg|]}��|�r|�qSr4�r�r�r�r4r5r��r�z+FileList.global_include.<locals>.<listcomp>)�allfilesr�r6r!r"rnr�r�r�r4r�r5r��s

zFileList.global_includecCsttj�d|��}|�|j�S)zD
        Exclude all files anywhere that match the pattern.
        rr�r�r4r4r5r��szFileList.global_excludecCs8|�d�r|dd�}t|�}|�|�r4|j�|�dS)N�
r�)rBr�
_safe_pathr��append)r<r�r"r4r4r5r��s


zFileList.appendcCs|j�t|j|��dSr8)r�r��filterr�)r<�pathsr4r4r5r��szFileList.extendcCstt|j|j��|_dS)z�
        Replace self.files with only safe paths

        Because some owners of FileList manipulate the underlying
        ``files`` attribute directly, this method must be called to
        repair those paths.
        N)rer�r�r�r;r4r4r5�_repair�szFileList._repairc	Cs�d}t�|�}|dur(t�d|�dSt�|d�}|durNt�||d�dSz"tj�|�shtj�|�rnWdSWn$ty�t�||t	�
��Yn0dS)Nz!'%s' not %s encodable -- skippingz''%s' in unexpected encoding -- skippingFr�T)�
unicode_utils�filesys_decoderr|Z
try_encoder!r"r{�UnicodeEncodeError�sys�getfilesystemencoding)r<r"Zenc_warnZu_pathZ	utf8_pathr4r4r5r��s

zFileList._safe_pathN)rJrKrLr�r�r�r�r�r�r�r�r�r�r�r�r�r�r4r4r4r5rEsM



rc@sdeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	e
dd��Zdd�Zdd�Z
dd�ZdS)r�zMANIFEST.incCsd|_d|_d|_d|_dS)Nr)Zuse_defaultsr�Z
manifest_onlyZforce_manifestr;r4r4r5rYsz!manifest_maker.initialize_optionscCsdSr8r4r;r4r4r5ryszmanifest_maker.finalize_optionscCslt�|_tj�|j�s|��|��tj�|j�r<|�	�|�
�|��|j��|j�
�|��dSr8)rr�r!r"r{r��write_manifest�add_defaults�templateZ
read_template�add_license_files�prune_file_list�sortZremove_duplicatesr;r4r4r5r�s

zmanifest_maker.runcCst�|�}|�tjd�S)N�/)r�r��replacer!r#)r<r"r4r4r5�_manifest_normalize&s
z"manifest_maker._manifest_normalizecsB�j���fdd��jjD�}d�j}��t�j|f|�dS)zo
        Write the file list in 'self.filelist' to the manifest file
        named by 'self.manifest'.
        csg|]}��|��qSr4)r�r�r;r4r5r�2r�z1manifest_maker.write_manifest.<locals>.<listcomp>zwriting manifest file '%s'N)r�r�r�r��executerz)r<r��msgr4r;r5r�*s

zmanifest_maker.write_manifestcCs|�|�st�||�dSr8)�_should_suppress_warningrr|)r<r�r4r4r5r|6s
zmanifest_maker.warncCst�d|�S)z;
        suppress missing-file warnings from sdist
        zstandard file .*not found)r$r�)r�r4r4r5r�:sz'manifest_maker._should_suppress_warningcCs�t�|�|j�|j�|j�|j�tt��}|rB|j�|�nt	j
�|j�rX|��t	j
�d�rp|j�d�|�
d�}|j�|j�dS)Nzsetup.pyrN)rr�r�r�r�r�rerr�r!r"r{Z
read_manifest�get_finalized_commandr�rN)r<ZrcfilesZei_cmdr4r4r5r�As


zmanifest_maker.add_defaultscCs4|jjjpg}|D]}t�d|�q|j�|�dS)Nzadding license file '%s')r9rp�
license_filesrr�r�r�)r<r��lfr4r4r5r�Ss
z manifest_maker.add_license_filescCsZ|�d�}|j��}|j�|j�|j�|�t�tj	�}|jj
d|d|dd�dS)N�buildz(^|z)(RCS|CVS|\.svn)r)Zis_regex)r�r9�get_fullnamer�r��
build_baser$r%r!r#Zexclude_pattern)r<r��base_dirr#r4r4r5r�Zs

�zmanifest_maker.prune_file_listN)rJrKrLr�rYryr�r�r�r|�staticmethodr�r�r�r�r4r4r4r5r�
s

r�cCsLd�|�}|�d�}t|d��}|�|�Wd�n1s>0YdS)z{Create a file with the specified name and write 'contents' (a
    sequence of strings without line terminators) to it.
    �
r�r�N)rnr�r�r�)r_�contentsr�r4r4r5rzds

rzc	Cs�t�d|�|js�|jj}|j|j|_}|j|j|_}z|�	|j
�W|||_|_n|||_|_0t|jdd�}t�
|j
|�dS)Nz
writing %sZzip_safe)rr�r�r9rprWrDrVr=�write_pkg_inforN�getattrr
Zwrite_safety_flag)�cmd�basenamer_rpZoldverZoldname�safer4r4r5r�qs r�cCstj�|�rt�d�dS)NzsWARNING: 'depends.txt' is not used by setuptools 0.6!
Use the install_requires/extras_require setup() args instead.)r!r"r{rr|�r�r�r_r4r4r5�warn_depends_obsolete�s�r�cCs,t|pd�}dd�}t||�}|�|�dS)Nr4cSs|dS)Nr�r4)r�r4r4r5�	append_cr�sz&_write_requirements.<locals>.append_cr)r�map�
writelines)�stream�reqs�linesr�r4r4r5�_write_requirements�s
r�cCsn|j}t��}t||j�|jp"i}t|�D]*}|�djfit	����t|||�q,|�
d||���dS)Nz
[{extra}]
�requirements)r9�io�StringIOr�Zinstall_requires�extras_require�sortedr�r��varsr��getvalue)r�r�r_�distrr��extrar4r4r5�write_requirements�s
rcCs,t��}t||jj�|�d||���dS)Nzsetup-requirements)r�r�r�r9Zsetup_requiresr�r)r�r�r_rr4r4r5�write_setup_requirements�srcCs:t�dd�|j��D��}|�d|d�t|��d�dS)NcSsg|]}|�dd�d�qS)�.rr)r )r��kr4r4r5r��s�z(write_toplevel_names.<locals>.<listcomp>ztop-level namesr�)r^�fromkeysr9Ziter_distribution_namesrzrnr�)r�r�r_�pkgsr4r4r5�write_toplevel_names�s��r
cCst|||d�dS)NT)�	write_argr�r4r4r5�
overwrite_arg�srFcCsHtj�|�d}t|j|d�}|dur4d�|�d}|�||||�dS)Nrr�)r!r"�splitextr�r9rnr�)r�r�r_r��argnamer[r4r4r5r�s
rcCs�|jj}t|t�s|dur |}nl|dur�g}t|���D]H\}}t|t�snt�||�}d�tt	t|�
����}|�d||f�q8d�|�}|�d||d�dS)Nr�z	[%s]
%s

rzentry pointsT)
r9Zentry_pointsrc�strr��itemsr�parse_grouprnr��valuesr�r�)r�r�r_r�r�sectionr�r4r4r5�
write_entries�s

rcCs�t�dt�tj�d�r|t�d��J}|D]4}t�	d|�}|r(t
|�d��Wd�Sq(Wd�n1sr0YdS)zd
    Get a -r### off of PKG-INFO Version in case this is an sdist of
    a subversion revision.
    z$get_pkg_info_revision is deprecated.zPKG-INFOzVersion:.*-r(\d+)\s*$rNr)�warningsr|�EggInfoDeprecationWarningr!r"r{r�r�r$r��int�group)r�r�r�r4r4r5�get_pkg_info_revision�s�@rc@seZdZdZdS)rz?Deprecated behavior warning for EggInfo, bypassing suppression.N)rJrKrL�__doc__r4r4r4r5r�sr)F);r�distutils.filelistrZ	_FileList�distutils.errorsr�distutils.utilrrgrr�r!r$r�r�rrGr\�
setuptoolsrZsetuptools.command.sdistrrZsetuptools.command.setoptr	Zsetuptools.commandr
�
pkg_resourcesrrr
rrrrrZsetuptools.unicode_utilsr�Zsetuptools.globrZsetuptools.externrrr6r7rNr�rzr�r�r�rrr
rrrrrr4r4r4r5�<module>sV(S1IW
	

site-packages/setuptools/command/__pycache__/develop.cpython-39.pyc000064400000014051150564745010021432 0ustar00a

(�Red�@s�ddlmZddlmZddlmZmZddlZddlZddl	Z	ddl
Z
ddlmZddl
mZddl
Z
Gdd�deje�ZGd	d
�d
�ZdS)�)�convert_path)�log)�DistutilsError�DistutilsOptionErrorN)�easy_install)�
namespacesc@sveZdZdZdZejddgZejdgZdZdd�Z	d	d
�Z
dd�Zed
d��Z
dd�Zdd�Zdd�Zdd�ZdS)�developzSet up package for developmentz%install package in 'development mode')�	uninstall�uzUninstall this source package)z	egg-path=Nz-Set the path to be used in the .egg-link filer	FcCs2|jrd|_|��|��n|��|��dS)NT)r	Z
multi_version�uninstall_linkZuninstall_namespaces�install_for_developmentZwarn_deprecated_options��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/develop.py�runs
zdevelop.runcCs&d|_d|_t�|�d|_d|_dS)N�.)r	�egg_pathr�initialize_options�
setup_pathZalways_copy_fromr
rrrr%s

zdevelop.initialize_optionscCs|�d�}|jr,d}|j|jf}t||��|jg|_t�|�|��|�	�|j
�t�d��|jd}t
j�|j|�|_|j|_|jdur�t
j�|j�|_t�|j�}t�t
j�|j|j��}||kr�td|��tj|t�|t
j�|j��|jd�|_|�|j|j|j�|_dS)N�egg_infoz-Please rename %r to %r before using 'develop'z*.eggz	.egg-linkzA--egg-path must be a relative path from the install directory to ��project_name)�get_finalized_commandZbroken_egg_inforr�egg_name�argsr�finalize_options�expand_basedirs�expand_dirsZ
package_index�scan�glob�os�path�join�install_dir�egg_link�egg_baser�abspath�
pkg_resources�normalize_pathr�Distribution�PathMetadata�dist�_resolve_setup_pathr)r�ei�templaterZegg_link_fn�targetrrrrr,sF




�����zdevelop.finalize_optionscCsn|�tjd��d�}|tjkr0d|�d�d}t�tj�	|||��}|t�tj�krjt
d|t�tj���|S)z�
        Generate a path from egg_base back to '.' where the
        setup script resides and ensure that path points to the
        setup path from $install_dir/$egg_path.
        �/z../�zGCan't get a consistent path to setup script from installation directory)�replacer!�sep�rstrip�curdir�countr(r)r"r#r)r&r$rZ
path_to_setup�resolvedrrrr-Ws
�
�zdevelop._resolve_setup_pathcCs�|�d�|jddd�|�d�tjr:|�tj�dt_|��t�d|j|j	�|j
s�t|jd��&}|�|j
d|j�Wd�n1s�0Y|�d|j|j�dS)Nr�	build_extr2)ZinplacezCreating %s (link to %s)�w�
)�run_command�reinitialize_command�
setuptoolsZbootstrap_install_fromrZinstall_namespacesr�infor%r&�dry_run�open�writerrZprocess_distributionr,�no_deps)r�frrrrms

4zdevelop.install_for_developmentcCs�tj�|j�rzt�d|j|j�t|j�}dd�|D�}|��||j	g|j	|j
gfvrht�d|�dS|jszt�
|j�|js�|�|j�|jjr�t�d�dS)NzRemoving %s (link to %s)cSsg|]}|���qSr)r5)�.0�linerrr�
<listcomp>��z*develop.uninstall_link.<locals>.<listcomp>z$Link points to %s: uninstall abortedz5Note: you must uninstall or replace scripts manually!)r!r"�existsr%rr?r&rA�closerr�warnr@�unlinkZ
update_pthr,�distribution�scripts)rZ
egg_link_file�contentsrrrr�s
zdevelop.uninstall_linkc	Cs�||jurt�||�S|�|�|jjp*gD]b}tj�t	|��}tj�
|�}t�|��}|�
�}Wd�n1st0Y|�||||�q,dS�N)r,r�install_egg_scripts�install_wrapper_scriptsrMrNr!r"r'r�basename�iorA�readZinstall_script)rr,�script_nameZscript_path�strm�script_textrrrrQ�s

&zdevelop.install_egg_scriptscCst|�}t�||�SrP)�VersionlessRequirementrrR�rr,rrrrR�szdevelop.install_wrapper_scriptsN)�__name__�
__module__�__qualname__�__doc__�descriptionr�user_options�boolean_optionsZcommand_consumes_argumentsrrr�staticmethodr-rrrQrRrrrrrs"�	+
rc@s(eZdZdZdd�Zdd�Zdd�ZdS)	rYa�
    Adapt a pkg_resources.Distribution to simply return the project
    name as the 'requirement' so that scripts will work across
    multiple versions.

    >>> from pkg_resources import Distribution
    >>> dist = Distribution(project_name='foo', version='1.0')
    >>> str(dist.as_requirement())
    'foo==1.0'
    >>> adapted_dist = VersionlessRequirement(dist)
    >>> str(adapted_dist.as_requirement())
    'foo'
    cCs
||_dSrP)�_VersionlessRequirement__distrZrrr�__init__�szVersionlessRequirement.__init__cCst|j|�SrP)�getattrrc)r�namerrr�__getattr__�sz"VersionlessRequirement.__getattr__cCs|jSrPrr
rrr�as_requirement�sz%VersionlessRequirement.as_requirementN)r[r\r]r^rdrgrhrrrrrY�srY)�distutils.utilr�	distutilsr�distutils.errorsrrr!r rTr(Zsetuptools.command.easy_installrr>rZDevelopInstallerrrYrrrr�<module>ssite-packages/setuptools/command/__pycache__/setopt.cpython-39.pyc000064400000011162150564745010021312 0ustar00a

(�Re��@s�ddlmZddlmZddlmZddlZddlZddlZddlm	Z	gd�Z
ddd	�Zddd�ZGd
d�de	�Z
Gdd�de
�ZdS)�)�convert_path)�log)�DistutilsOptionErrorN)�Command)�config_file�edit_config�option_base�setopt�localcCsh|dkrdS|dkr,tj�tj�tj�d�S|dkrZtjdkrBdpDd}tj�td	|��St	d
|��dS)z�Get the filename of the distutils, local, global, or per-user config

    `kind` must be one of "local", "global", or "user"
    r
z	setup.cfg�globalz
distutils.cfg�user�posix�.�z~/%spydistutils.cfgz7config_file() type must be 'local', 'global', or 'user'N)
�os�path�join�dirname�	distutils�__file__�name�
expanduserr�
ValueError)�kind�dot�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/setopt.pyr
s��rFc		CsFt�d|�t��}dd�|_|�|g�|��D]�\}}|dur\t�d||�|�|�q2|�	|�s~t�d||�|�
|�|��D]p\}}|dur�t�d|||�|�||�|�|�s�t�d||�|�|�q�t�d	||||�|�
|||�q�q2t�d
|�|�sBt|d��}|�|�Wd�n1�s80YdS)aYEdit a configuration file to include `settings`

    `settings` is a dictionary of dictionaries or ``None`` values, keyed by
    command/section name.  A ``None`` value means to delete the entire section,
    while a dictionary lists settings to be changed or deleted in that section.
    A setting of ``None`` means to delete that setting.
    zReading configuration from %scSs|S�Nr)�xrrr�<lambda>*�zedit_config.<locals>.<lambda>NzDeleting section [%s] from %szAdding new section [%s] to %szDeleting %s.%s from %sz#Deleting empty [%s] section from %szSetting %s.%s to %r in %sz
Writing %s�w)r�debug�configparser�RawConfigParser�optionxform�read�items�info�remove_section�has_section�add_section�
remove_option�options�set�open�write)	�filename�settings�dry_run�opts�sectionr-�option�value�frrrr s@


�
��rc@s0eZdZdZgd�ZddgZdd�Zdd�Zd	S)
rz<Abstract base class for commands that mess with config files))�
global-config�gz0save options to the site-wide distutils.cfg file)�user-config�uz7save options to the current user's pydistutils.cfg file)z	filename=r8z-configuration file to use (default=setup.cfg)r9r;cCsd|_d|_d|_dSr)�
global_config�user_configr1��selfrrr�initialize_options\szoption_base.initialize_optionscCsvg}|jr|�td��|jr,|�td��|jdurB|�|j�|sT|�td��t|�dkrjtd|��|\|_dS)Nrrr
�z/Must specify only one configuration file option)r=�appendrr>r1�lenr)r@�	filenamesrrr�finalize_optionsas
�zoption_base.finalize_optionsN)�__name__�
__module__�__qualname__�__doc__�user_options�boolean_optionsrArFrrrrrLs
�rc@sFeZdZdZdZgd�ejZejdgZdd�Zdd�Z	d	d
�Z
dS)r	z#Save command-line options to a filez1set an option in setup.cfg or another config file))zcommand=�czcommand to set an option for)zoption=�oz
option to set)z
set-value=�szvalue of the option)�remove�rzremove (unset) the valuerPcCs&t�|�d|_d|_d|_d|_dSr)rrA�commandr6�	set_valuerPr?rrrrA�s

zsetopt.initialize_optionscCsBt�|�|jdus|jdur&td��|jdur>|js>td��dS)Nz%Must specify --command *and* --optionz$Must specify --set-value or --remove)rrFrRr6rrSrPr?rrrrF�s

zsetopt.finalize_optionscCs*t|j|j|j�dd�|jii|j�dS)N�-�_)rr1rRr6�replacerSr3r?rrr�run�s��z
setopt.runN)rGrHrIrJ�descriptionrrKrLrArFrWrrrrr	ss�r	)r
)F)�distutils.utilrrr�distutils.errorsrrr#�
setuptoolsr�__all__rrrr	rrrr�<module>s

,'site-packages/setuptools/command/__pycache__/build_py.cpython-39.pyc000064400000017272150564745010021613 0ustar00a

(�ReT �@s�ddlmZddlmZddlmmZddlZddlZddl	Z	ddl
Z
ddlZddl
Z
ddlZddlmZdd�ZGdd�dej�Zd	d
�ZdS)�)�glob)�convert_pathN)�unique_everseencCst�|t�|�jtjB�dS�N)�os�chmod�stat�st_mode�S_IWRITE)�target�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/build_py.py�
make_writablesrc@s�eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zedd��Zd S)!�build_pyaXEnhanced 'build_py' command that includes data files with packages

    The data files are specified via a 'package_data' argument to 'setup()'.
    See 'setuptools.dist.Distribution' for more details.

    Also, this version of the 'build_py' command allows you to specify both
    'py_modules' and 'packages' in the same setup operation.
    cCs@tj�|�|jj|_|jjp i|_d|jvr6|jd=g|_dS)N�
data_files)�origr�finalize_options�distribution�package_data�exclude_package_data�__dict__�_build_py__updated_files��selfrrr
rs

zbuild_py.finalize_optionscCsN|js|jsdS|jr|��|jr4|��|��|�tjj|dd��dS)z?Build modules, packages, and copy data files to build directoryNr)Zinclude_bytecode)	�
py_modules�packagesZ
build_modulesZbuild_packages�build_package_data�byte_compilerr�get_outputsrrrr
�run$szbuild_py.runcCs&|dkr|��|_|jStj�||�S)zlazily compute data filesr)�_get_data_filesrrr�__getattr__)r�attrrrr
r!4s
zbuild_py.__getattr__cCs.tj�||||�\}}|r&|j�|�||fSr)rr�build_moduler�append)r�moduleZmodule_file�package�outfile�copiedrrr
r#;szbuild_py.build_modulecCs|��tt|j|jpd��S)z?Generate list of '(package,src_dir,build_dir,filenames)' tuplesr)�analyze_manifest�list�map�_get_pkg_data_filesrrrrr
r Aszbuild_py._get_data_filescsJ|�|��tjj|jg|�d��}�fdd�|�|��D�}|�||fS)N�.csg|]}tj�|���qSr)r�path�relpath)�.0�file��src_dirrr
�
<listcomp>Ns�z0build_py._get_pkg_data_files.<locals>.<listcomp>)�get_package_dirrr.�join�	build_lib�split�find_data_files)rr&�	build_dir�	filenamesrr2r
r,Fs


�zbuild_py._get_pkg_data_filescCsX|�|j||�}tt|�}tj�|�}ttj	j
|�}t�|j�|g�|�}|�
|||�S)z6Return filenames for package's data files in 'src_dir')�_get_platform_patternsrr+r�	itertools�chain�
from_iterable�filterrr.�isfile�manifest_files�get�exclude_data_files)rr&r3�patternsZglobs_expandedZ
globs_matchesZ
glob_files�filesrrr
r9Ts�
�zbuild_py.find_data_filesc
Cst|jD]h\}}}}|D]V}tj�||�}|�tj�|��tj�||�}|�||�\}}	t|�tj�|�}qqdS)z$Copy data files into build directoryN)	rrr.r6�mkpath�dirname�	copy_filer�abspath)
rr&r3r:r;�filenamer�srcfile�outfr(rrr
reszbuild_py.build_package_datacCs�i|_}|jjsdSi}|jp"dD]}||t|�|��<q$|�d�|�d�}|jj	D]�}t
j�t|��\}}d}|}	|r�||kr�||vr�|}t
j�|�\}}
t
j�
|
|�}qx||vrX|�d�r�||	kr�qX|�||g��|�qXdS)Nr�egg_infoz.py)rBrZinclude_package_datar�assert_relativer5�run_command�get_finalized_command�filelistrFrr.r8r6�endswith�
setdefaultr$)rZmfZsrc_dirsr&Zei_cmdr.�d�f�prevZoldfZdfrrr
r)ps(


zbuild_py.analyze_manifestcCsdSrrrrrr
�get_data_files�szbuild_py.get_data_filescCs�z|j|WStyYn0tj�|||�}||j|<|rF|jjsJ|S|jjD]}||ksl|�|d�rRqvqR|St�	|d��}|�
�}Wd�n1s�0Yd|vr�tj�
d|f��|S)z8Check namespace packages' __init__ for declare_namespacer-�rbNsdeclare_namespacez�Namespace package problem: %s is a namespace package, but its
__init__.py does not call declare_namespace()! Please fix it.
(See the setuptools manual under "Namespace Packages" for details.)
")�packages_checked�KeyErrorrr�
check_packagerZnamespace_packages�
startswith�io�open�read�	distutils�errors�DistutilsError)rr&�package_dirZinit_py�pkgrV�contentsrrr
r\�s*
&��zbuild_py.check_packagecCsi|_tj�|�dSr)rZrr�initialize_optionsrrrr
rg�szbuild_py.initialize_optionscCs0tj�||�}|jjdur,tj�|jj|�S|Sr)rrr5rZsrc_rootrr.r6)rr&�resrrr
r5�szbuild_py.get_package_dircs\t���|�|j||�}�fdd�|D�}tj�|�}t|���fdd��D�}tt|��S)z6Filter filenames for package's data files in 'src_dir'c3s|]}t��|�VqdSr)�fnmatchr@�r0�pattern)rFrr
�	<genexpr>��z.build_py.exclude_data_files.<locals>.<genexpr>c3s|]}|�vr|VqdSrr)r0�fn)�badrr
rl�rm)r*r<rr=r>r?�setr)rr&r3rFrEZmatch_groups�matchesZkeepersr)rorFr
rD�s�zbuild_py.exclude_data_filescs.t�|�dg�|�|g��}�fdd�|D�S)z�
        yield platform-specific path patterns (suitable for glob
        or fn_match) from a glob-based spec (such as
        self.package_data or self.exclude_package_data)
        matching package in src_dir.
        �c3s |]}tj��t|��VqdSr)rr.r6rrjr2rr
rl�s�z2build_py._get_platform_patterns.<locals>.<genexpr>)r=r>rC)�specr&r3Zraw_patternsrr2r
r<�s

�
�zbuild_py._get_platform_patternsN)�__name__�
__module__�__qualname__�__doc__rrr!r#r r,r9rr)rXr\rgr5rD�staticmethodr<rrrr
rs"	rcCs:tj�|�s|Sddlm}t�d���|}||��dS)Nr)�DistutilsSetupErrorz�
        Error: setup script specifies an absolute path:

            %s

        setup() arguments must *always* be /-separated paths relative to the
        setup.py directory, *never* absolute paths.
        )rr.�isabs�distutils.errorsry�textwrap�dedent�lstrip)r.ry�msgrrr
rO�s�
��
rO)r�distutils.utilrZdistutils.command.build_py�commandrrrrir|r^r{rar=rZ setuptools.extern.more_itertoolsrrrOrrrr
�<module>sEsite-packages/setuptools/command/__pycache__/install.cpython-39.pyc000064400000010211150564745010021434 0ustar00a

(�Re*�@s|ddlmZddlZddlZddlZddlZddlmmZ	ddl
Z
e	jZGdd�de	j�Zdd�e	jjD�ej
e_dS)�)�DistutilsArgErrorNc@s�eZdZdZejjddgZejjddgZddd�fd	d
d�fgZe	e�Z
dd�Zd
d�Zdd�Z
dd�Zedd��Zdd�ZdS)�installz7Use easy_install to install the package, w/dependencies)�old-and-unmanageableNzTry not to use this!)�!single-version-externally-managedNz5used by system package builders to create 'flat' eggsrr�install_egg_infocCsdS�NT���selfrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/install.py�<lambda>�zinstall.<lambda>�install_scriptscCsdSrrr	rrrrr
cCs*t�dtj�tj�|�d|_d|_dS)NzRsetup.py install is deprecated. Use build and pip and other standards-based tools.)	�warnings�warn�
setuptoolsZSetuptoolsDeprecationWarning�origr�initialize_options�old_and_unmanageable�!single_version_externally_managedr	rrrr s�zinstall.initialize_optionscCs8tj�|�|jrd|_n|jr4|js4|js4td��dS)NTzAYou must specify --record or --root when building system packages)rr�finalize_options�rootr�recordrr	rrrr,s�zinstall.finalize_optionscCs(|js|jrtj�|�Sd|_d|_dS)N�)rrrr�handle_extra_path�	path_file�
extra_dirsr	rrrr7szinstall.handle_extra_pathcCs@|js|jrtj�|�S|�t���s4tj�|�n|��dS)N)	rrrr�run�_called_from_setup�inspect�currentframe�do_egg_installr	rrrrAs
zinstall.runcCsz|dur4d}t�|�t��dkr0d}t�|�dSt�|�d}|dd�\}t�|�}|j�dd	�}|d
kox|j	dkS)a�
        Attempt to detect whether run() was called from setup() or by another
        command.  If called by setup(), the parent caller will be the
        'run_command' method in 'distutils.dist', and *its* caller will be
        the 'run_commands' method.  If called any other way, the
        immediate caller *might* be 'run_command', but it won't have been
        called by 'run_commands'. Return True in that case or if a call stack
        is unavailable. Return False otherwise.
        Nz4Call stack not available. bdist_* commands may fail.�
IronPythonz6For best results, pass -X:Frames to enable call stack.T���__name__rzdistutils.dist�run_commands)
rr�platform�python_implementationr�getouterframes�getframeinfo�	f_globals�get�function)Z	run_frame�msg�resZcaller�infoZ
caller_modulerrrrLs


�zinstall._called_from_setupcCs�|j�d�}||jd|j|jd�}|��d|_|j�t�d��|�	d�|j�
d�jg}tj
rp|�dtj
�||_|jdd	�dt_
dS)
N�easy_install�x)�argsrr�.z*.eggZ	bdist_eggrF)Zshow_deprecation)�distribution�get_command_classrr�ensure_finalizedZalways_copy_fromZ
package_index�scan�glob�run_command�get_command_objZ
egg_outputrZbootstrap_install_from�insertr3r)r
r1�cmdr3rrrr!gs�
zinstall.do_egg_installN)r%�
__module__�__qualname__�__doc__rr�user_options�boolean_options�new_commands�dict�_ncrrrr�staticmethodrr!rrrrrs&��

�

rcCsg|]}|dtjvr|�qS)r)rrE)�.0r=rrr�
<listcomp>�r
rH)�distutils.errorsrrr9rr'�distutils.command.install�commandrrr�_install�sub_commandsrCrrrr�<module>ss��site-packages/setuptools/command/__pycache__/build_clib.cpython-39.pyc000064400000004724150564745010022072 0ustar00a

(�Re?�@sLddlmmZddlmZddlmZddlm	Z	Gdd�dej�ZdS)�N)�DistutilsSetupError)�log)�newer_pairwise_groupc@seZdZdZdd�ZdS)�
build_clibav
    Override the default build_clib behaviour to do the following:

    1. Implement a rudimentary timestamp-based dependency system
       so 'compile()' doesn't run every time.
    2. Add more keys to the 'build_info' dictionary:
        * obj_deps - specify dependencies for each object compiled.
                     this should be a dictionary mapping a key
                     with the source filename to a list of
                     dependencies. Use an empty string for global
                     dependencies.
        * cflags   - specify a list of additional flags to pass to
                     the compiler.
    c	Csr|D�]f\}}|�d�}|dus.t|ttf�s:td|��t|�}t�d|�|�dt��}t|t�srtd|��g}|�dt��}t|ttf�s�td|��|D]P}|g}	|	�|�|�|t��}
t|
ttf�s�td|��|	�|
�|�	|	�q�|j
j||jd�}t
||�ggfk�rT|�d�}|�d	�}
|�d
�}|j
j||j||
||jd�|j
j|||j|jd�qdS)
N�sourceszfin 'libraries' option (library '%s'), 'sources' must be present and must be a list of source filenameszbuilding '%s' library�obj_depsz\in 'libraries' option (library '%s'), 'obj_deps' must be a dictionary of type 'source: list'�)�
output_dir�macros�include_dirs�cflags)r	r
rZextra_postargs�debug)r	r
)�get�
isinstance�list�tuplerr�info�dict�extend�append�compilerZobject_filenames�
build_tempr�compiler
Zcreate_static_libr)�self�	librariesZlib_nameZ
build_inforr�dependenciesZglobal_deps�sourceZsrc_depsZ
extra_depsZexpected_objectsr
rr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/build_clib.py�build_librariess|
��
����
��
���


��zbuild_clib.build_librariesN)�__name__�
__module__�__qualname__�__doc__rrrrrrsr)
Zdistutils.command.build_clib�commandr�orig�distutils.errorsr�	distutilsrZsetuptools.dep_utilrrrrr�<module>ssite-packages/setuptools/command/__pycache__/install_scripts.cpython-39.pyc000064400000004646150564745010023222 0ustar00a

(�Re!
�@sdddlmZddlmmZddlmZddlZddl	Z	ddl
mZmZm
Z
Gdd�dej�ZdS)�)�logN)�DistutilsModuleError)�Distribution�PathMetadata�ensure_directoryc@s*eZdZdZdd�Zdd�Zd
dd�Zd	S)�install_scriptsz;Do normal script install, plus any egg_info wrapper scriptscCstj�|�d|_dS)NF)�origr�initialize_options�no_ep)�self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/install_scripts.pyr	
sz"install_scripts.initialize_optionsc	Csddlmm}|�d�|jjr2tj�|�ng|_	|j
rBdS|�d�}t|j
t|j
|j�|j|j�}|�d�}t|dd�}z|�d�}t|dd�}Wnttfy�d}Yn0|j}|r�d}|j}|tjkr�|g}|��}|j���|�}	|�||	���D]}
|j|
��qdS)	Nr�egg_info�
build_scripts�
executable�
bdist_wininstZ_is_runningFz
python.exe)�setuptools.command.easy_install�command�easy_install�run_command�distribution�scriptsrr�run�outfilesr
�get_finalized_commandrZegg_baserr�egg_nameZegg_version�getattr�ImportErrorrZScriptWriterZWindowsScriptWriter�sysr�bestZcommand_spec_class�
from_param�get_argsZ	as_header�write_script)r�eiZei_cmd�distZbs_cmdZ
exec_paramZbw_cmdZ
is_wininst�writer�cmd�argsrrr
rs:

�



zinstall_scripts.run�tc
Gs�ddlm}m}t�d||j�tj�|j|�}|j	�
|�|�}|js~t|�t
|d|�}	|	�|�|	��||d|�dS)z1Write an executable file to the scripts directoryr)�chmod�
current_umaskzInstalling %s script to %s�wi�N)rr)r*r�infoZinstall_dir�os�path�joinr�append�dry_runr�open�write�close)
r�script_name�contents�modeZignoredr)r*�target�mask�frrr
r"7s
zinstall_scripts.write_scriptN)r()�__name__�
__module__�__qualname__�__doc__r	rr"rrrr
r
s&r)�	distutilsrZ!distutils.command.install_scriptsrrr�distutils.errorsrr-r�
pkg_resourcesrrrrrrr
�<module>ssite-packages/setuptools/command/__pycache__/alias.cpython-39.pyc000064400000004563150564745010021074 0ustar00a

(�ReM	�@sDddlmZddlmZmZmZdd�ZGdd�de�Zdd�Zd	S)
�)�DistutilsOptionError)�edit_config�option_base�config_filecCs8dD]}||vrt|�Sq|��|gkr4t|�S|S)z4Quote an argument for later parsing by shlex.split())�"�'�\�#)�repr�split)�arg�c�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/alias.py�shquotesrc@sHeZdZdZdZdZdgejZejdgZdd�Z	dd	�Z
d
d�ZdS)
�aliasz3Define a shortcut that invokes one or more commandsz0define a shortcut to invoke one or more commandsT)�remove�rzremove (unset) the aliasrcCst�|�d|_d|_dS)N)r�initialize_options�argsr��selfrrrrs
zalias.initialize_optionscCs*t�|�|jr&t|j�dkr&td��dS)N�zFMust specify exactly one argument (the alias name) when using --remove)r�finalize_optionsr�lenrrrrrrr!s

�zalias.finalize_optionscCs�|j�d�}|js@td�td�|D]}tdt||��q&dSt|j�dkr�|j\}|jrbd}q�||vr~tdt||��dStd|�dSn$|jd}d�tt	|jdd���}t
|jd||ii|j�dS)	N�aliaseszCommand Aliasesz---------------zsetup.py aliasrz No alias definition found for %rr� )
�distribution�get_option_dictr�print�format_aliasrr�join�maprr�filename�dry_run)rrr�commandrrr�run)s&
z	alias.runN)�__name__�
__module__�__qualname__�__doc__�descriptionZcommand_consumes_argumentsr�user_options�boolean_optionsrrr&rrrrrs��rcCsZ||\}}|td�krd}n,|td�kr0d}n|td�krBd}nd|}||d|S)	N�globalz--global-config �userz--user-config �local�z
--filename=%rr)r)�namer�sourcer%rrrr Dsr N)	�distutils.errorsrZsetuptools.command.setoptrrrrrr rrrr�<module>s
4site-packages/setuptools/command/__pycache__/rotate.cpython-39.pyc000064400000004770150564745010021301 0ustar00a

(�ReP�@sTddlmZddlmZddlmZddlZddlZddlm	Z	Gdd�de	�Z
dS)�)�convert_path)�log)�DistutilsOptionErrorN)�Commandc@s8eZdZdZdZgd�ZgZdd�Zdd�Zdd	�Z	d
S)�rotatezDelete older distributionsz2delete older distributions, keeping N newest files))zmatch=�mzpatterns to match (required))z	dist-dir=�dz%directory where the distributions are)zkeep=�kz(number of matching distributions to keepcCsd|_d|_d|_dS)N)�match�dist_dir�keep)�self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/rotate.py�initialize_optionsszrotate.initialize_optionsc
Cs�|jdurtd��|jdur$td��zt|j�|_Wn.tyb}ztd�|�WYd}~n
d}~00t|jt�r�dd�|j�d�D�|_|�dd�dS)	NzQMust specify one or more (comma-separated) match patterns (e.g. '.zip' or '.egg')z$Must specify number of files to keepz--keep must be an integercSsg|]}t|����qSr)r�strip)�.0�prrr�
<listcomp>(sz+rotate.finalize_options.<locals>.<listcomp>�,�bdist)rr)	r
rr�int�
ValueError�
isinstance�str�split�set_undefined_options)r
�errr�finalize_optionss
�
 
�zrotate.finalize_optionscCs�|�d�ddlm}|jD]�}|j��d|}|tj�|j|��}dd�|D�}|�	�|�
�t�dt
|�|�||jd�}|D]<\}}t�d|�|js�tj�|�r�t�|�q�t�|�q�qdS)	N�egg_infor)�glob�*cSsg|]}tj�|�|f�qSr)�os�path�getmtime)r�frrrr4�zrotate.run.<locals>.<listcomp>z%d file(s) matching %szDeleting %s)�run_commandr r
�distribution�get_namer"r#�joinr�sort�reverser�info�lenr�dry_run�isdir�shutil�rmtree�unlink)r
r �pattern�files�tr%rrr�run-s 

z
rotate.runN)
�__name__�
__module__�__qualname__�__doc__�description�user_options�boolean_optionsrrr7rrrrr
sr)�distutils.utilr�	distutilsr�distutils.errorsrr"r1�
setuptoolsrrrrrr�<module>ssite-packages/setuptools/command/__pycache__/saveopts.cpython-39.pyc000064400000001713150564745010021641 0ustar00a

(�Re��@s$ddlmZmZGdd�de�ZdS)�)�edit_config�option_basec@seZdZdZdZdd�ZdS)�saveoptsz#Save command-line options to a filez7save supplied options to setup.cfg or other config filecCsh|j}i}|jD]B}|dkrq|�|���D]$\}\}}|dkr,||�|i�|<q,qt|j||j�dS)Nrzcommand line)�distribution�command_options�get_option_dict�items�
setdefaultr�filename�dry_run)�self�dist�settings�cmd�opt�src�val�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/saveopts.py�run	s
zsaveopts.runN)�__name__�
__module__�__qualname__�__doc__�descriptionrrrrrrsrN)Zsetuptools.command.setoptrrrrrrr�<module>ssite-packages/setuptools/command/__pycache__/install_egg_info.cpython-39.pyc000064400000004653150564745010023306 0ustar00a

(�Re��@s\ddlmZmZddlZddlmZddlmZddlmZddl	Z	Gdd�dej
e�ZdS)�)�log�dir_utilN)�Command)�
namespaces)�unpack_archivec@sBeZdZdZdZdgZdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
S)�install_egg_infoz.Install an .egg-info directory for the package)zinstall-dir=�dzdirectory to install tocCs
d|_dS�N)�install_dir��self�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/install_egg_info.py�initialize_optionssz#install_egg_info.initialize_optionscCsV|�dd�|�d�}t�dd|j|j���d}|j|_tj	�
|j|�|_g|_
dS)N�install_lib)r
r
�egg_infoz	.egg-info)�set_undefined_options�get_finalized_command�
pkg_resources�Distribution�egg_nameZegg_versionr�source�os�path�joinr
�target�outputs)rZei_cmd�basenamer
r
r�finalize_optionss�
��z!install_egg_info.finalize_optionscCs�|�d�tj�|j�r:tj�|j�s:tj|j|jd�n(tj�	|j�rb|�
tj|jfd|j�|jstt�
|j�|�
|jdd|j|jf�|��dS)Nr)�dry_runz	Removing r
�Copying %s to %s)�run_commandrr�isdirr�islinkr�remove_treer�exists�execute�unlinkr�ensure_directory�copytreerZinstall_namespacesrr
r
r�run!s
�zinstall_egg_info.runcCs|jSr	)rrr
r
r�get_outputs.szinstall_egg_info.get_outputscs �fdd�}t�j�j|�dS)NcsDdD] }|�|�sd||vrdSq�j�|�t�d||�|S)N)z.svn/zCVS/�/r )�
startswithr�appendr�debug)�src�dst�skiprr
r�skimmer3sz*install_egg_info.copytree.<locals>.skimmer)rrr)rr3r
rrr)1szinstall_egg_info.copytreeN)�__name__�
__module__�__qualname__�__doc__�description�user_optionsrrr*r+r)r
r
r
rr
s�
r)�	distutilsrrr�
setuptoolsrrZsetuptools.archive_utilrrZ	Installerrr
r
r
r�<module>ssite-packages/setuptools/command/__pycache__/upload.cpython-39.pyc000064400000001542150564745010021261 0ustar00a

(�Re��@s:ddlmZddlmZddlmZGdd�dej�ZdS)�)�log)�upload)�RemovedCommandErrorc@seZdZdZdd�ZdS)rz)Formerly used to upload packages to PyPI.cCs"d}|�d|tj�t|��dS)Nz[The upload command has been removed, use twine to upload instead (https://pypi.org/p/twine)zERROR: )�announcer�ERRORr)�self�msg�r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/upload.py�run
s�z
upload.runN)�__name__�
__module__�__qualname__�__doc__rr	r	r	r
rsrN)�	distutilsr�distutils.commandr�origZsetuptools.errorsrr	r	r	r
�<module>ssite-packages/setuptools/command/__pycache__/register.cpython-39.pyc000064400000001575150564745010021627 0ustar00a

(�Re��@s@ddlmZddlmmZddlmZGdd�dej�ZdS)�)�logN)�RemovedCommandErrorc@seZdZdZdd�ZdS)�registerz+Formerly used to register packages on PyPI.cCs"d}|�d|tj�t|��dS)Nz]The register command has been removed, use twine to upload instead (https://pypi.org/p/twine)zERROR: )�announcer�ERRORr)�self�msg�r	��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/register.py�run
s�zregister.runN)�__name__�
__module__�__qualname__�__doc__rr	r	r	r
rsr)�	distutilsrZdistutils.command.register�commandr�origZsetuptools.errorsrr	r	r	r
�<module>ssite-packages/setuptools/command/__pycache__/test.cpython-39.pyc000064400000017615150564745010020764 0ustar00a

(�Re��@s�ddlZddlZddlZddlZddlZddlZddlmZmZddl	m
Z
ddlmZddlm
Z
mZmZmZmZmZmZmZddlmZddlmZGdd	�d	e�ZGd
d�d�ZGdd
�d
e�ZdS)�N)�DistutilsError�DistutilsOptionError)�log)�
TestLoader)�resource_listdir�resource_exists�normalize_path�working_set�evaluate_marker�add_activation_listener�require�
EntryPoint)�Command)�unique_everseenc@seZdZdd�Zddd�ZdS)�ScanningLoadercCst�|�t�|_dS�N)r�__init__�set�_visited��self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/test.pyrs
zScanningLoader.__init__NcCs�||jvrdS|j�|�g}|�t�||��t|d�rH|�|���t|d�r�t|jd�D]`}|�	d�r�|dkr�|jd|dd�}n"t
|j|d	�r^|jd|}nq^|�|�|��q^t|�d
kr�|�
|�S|dSdS)aReturn a suite of all tests cases contained in the given module

        If the module is a package, load tests from all the modules in it.
        If the module has an ``additional_tests`` function, call it and add
        the return value to the tests.
        N�additional_tests�__path__�z.pyz__init__.py�.���z/__init__.py�r)r�add�appendr�loadTestsFromModule�hasattrrr�__name__�endswithrZloadTestsFromName�lenZ
suiteClass)r�module�pattern�tests�file�	submodulerrrr!s$



z"ScanningLoader.loadTestsFromModule)N)r#�
__module__�__qualname__rr!rrrrrsrc@seZdZdd�Zddd�ZdS)�NonDataPropertycCs
||_dSr��fget)rr/rrrrBszNonDataProperty.__init__NcCs|dur|S|�|�Srr.)r�objZobjtyperrr�__get__EszNonDataProperty.__get__)N)r#r+r,rr1rrrrr-Asr-c@s�eZdZdZdZgd�Zdd�Zdd�Zedd	��Z	d
d�Z
dd
�Zej
gfdd��Zeej
dd���Zedd��Zdd�Zdd�Zedd��Zedd��ZdS)�testz.Command to run unit tests after in-place buildz0run unit tests after in-place build (deprecated)))ztest-module=�mz$Run 'test_suite' in specified module)ztest-suite=�sz9Run single test, case or suite (e.g. 'module.test_suite'))ztest-runner=�rzTest runner to usecCsd|_d|_d|_d|_dSr)�
test_suite�test_module�test_loader�test_runnerrrrr�initialize_optionsZsztest.initialize_optionscCs�|jr|jrd}t|��|jdurD|jdur8|jj|_n|jd|_|jdur^t|jdd�|_|jdurnd|_|jdur�t|jdd�|_dS)Nz1You may specify a module or a suite, but not bothz.test_suiter8z&setuptools.command.test:ScanningLoaderr9)r6r7r�distributionr8�getattrr9)r�msgrrr�finalize_options`s




ztest.finalize_optionscCst|���Sr)�list�
_test_argsrrrr�	test_argsssztest.test_argsccs4|jstjdkrdV|jr"dV|jr0|jVdS)N)��Zdiscoverz	--verbose)r6�sys�version_info�verboserrrrr@wsztest._test_argscCs2|���|�Wd�n1s$0YdS)zI
        Backward compatibility for project_on_sys_path context.
        N)�project_on_sys_path)r�funcrrr�with_project_on_sys_paths
ztest.with_project_on_sys_pathc
cs|�d�|jddd�|�d�|�d�}tjdd�}tj��}z�t|j�}tj�	d|�t
��tdd��t
d|j|jf�|�|g��dVWd�n1s�0YW|tjdd�<tj��tj�|�t
��n.|tjdd�<tj��tj�|�t
��0dS)	N�egg_info�	build_extr)ZinplacercSs|��Sr)�activate)�distrrr�<lambda>��z*test.project_on_sys_path.<locals>.<lambda>z%s==%s)�run_command�reinitialize_command�get_finalized_commandrD�path�modules�copyrZegg_base�insertr	rrr�egg_nameZegg_version�paths_on_pythonpath�clear�update)rZ
include_distsZei_cmd�old_pathZold_modulesZproject_pathrrrrG�s,




&

�
ztest.project_on_sys_pathc
cs�t�}tj�d|�}tj�dd�}zdtj�t|��}td||g�}tj�|�}|r\|tjd<dVW||ur|tj�dd�q�|tjd<n$||ur�tj�dd�n
|tjd<0dS)z�
        Add the indicated paths to the head of the PYTHONPATH environment
        variable so that subprocesses will also see the packages at
        these paths.

        Do this in a context that restores the value on exit.
        �
PYTHONPATHrN)	�object�os�environ�get�pathsep�joinr�filter�pop)�pathsZnothingZorig_pythonpathZcurrent_pythonpath�prefixZto_join�new_pathrrrrX�s 

�ztest.paths_on_pythonpathcCsD|�|j�}|�|jpg�}|�dd�|j��D��}t�|||�S)z�
        Install the requirements indicated by self.distribution and
        return an iterable of the dists that were built.
        css0|](\}}|�d�rt|dd��r|VqdS)�:rN)�
startswithr
)�.0�k�vrrr�	<genexpr>�s�z%test.install_dists.<locals>.<genexpr>)Zfetch_build_eggsZinstall_requiresZ
tests_requireZextras_require�items�	itertools�chain)rMZir_dZtr_dZer_drrr�
install_dists�s
�ztest.install_distsc	Cs�|�dtj�|�|j�}d�|j�}|jr>|�d|�dS|�d|�tt	�
d�|�}|�|��@|���|�
�Wd�n1s�0YWd�n1s�0YdS)Nz�WARNING: Testing via this command is deprecated and will be removed in a future version. Users looking for a generic test entry point independent of test runner are encouraged to use tox.� zskipping "%s" (dry run)zrunning "%s"�location)�announcer�WARNrqr;rb�_argv�dry_run�map�operator�
attrgetterrXrG�	run_tests)rZinstalled_dists�cmdrerrr�run�s�
ztest.runcCsVtjdd|j|�|j�|�|j�dd�}|j��sRd|j}|�|t	j
�t|��dS)NF)Z
testLoaderZ
testRunner�exitzTest failed: %s)�unittest�mainrv�_resolve_as_epr8r9�resultZ
wasSuccessfulrtr�ERRORr)rr2r=rrrr{�s

�

ztest.run_testscCsdg|jS)Nr)rArrrrrv�sz
test._argvcCs$|durdSt�d|�}|���S)zu
        Load the indicated attribute value, called, as a as if it were
        specified as an entry point.
        Nzx=)r
�parse�resolve)�val�parsedrrrr��sztest._resolve_as_epN)r#r+r,�__doc__�description�user_optionsr:r>r-rAr@rI�
contextlib�contextmanagerrG�staticmethodrXrqr}r{�propertyrvr�rrrrr2Ks,



r2)r^ryrDr�ror�distutils.errorsrr�	distutilsrr�
pkg_resourcesrrrr	r
rrr
�
setuptoolsrZ setuptools.extern.more_itertoolsrrr-r2rrrr�<module>s(
(
site-packages/setuptools/command/__pycache__/easy_install.cpython-39.pyc000064400000174504150564745010022475 0ustar00a

(�Re�N�@s�dZddlmZddlmZddlmZmZddlmZmZm	Z	m
Z
ddlmZm
Z
ddlmZmZddlmZdd	lmZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
lZdd
l Z dd
l!Z!dd
l"Z"dd
l#Z#dd
l$Z$dd
l%Z%dd
l&Z&ddl'm(Z(m)Z)ddl*m+Z+dd
l*m,Z,ddl-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4m5Z5m6Z6ddl/m7Z7m8Z8ddl9m:Z:ddl;m<Z<m=Z=m>Z>m?Z?m@Z@mAZAmBZBmCZCmDZDmEZEmFZFmGZGmHZHmIZImJZJdd
l;Z;ejKde;jLd�gd�ZMdd�ZNdd�ZOdd�ZPdd�ZQd d!�ZRGd"d#�d#e,�ZSd$d%�ZTd&d'�ZUd(d)�ZVd*d+�ZWd,d-�ZXGd.d/�d/eB�ZYGd0d1�d1eY�ZZej[�\d2d3�d4k�r�eZZYd5d6�Z]d7d8�Z^d9d:�Z_d;d<�Z`did=d>�Zad?d@�ZbdAdB�ZcdCejdv�r�ecZendDdE�ZedjdGdH�ZfdIdJ�ZgdKdL�ZhdMdN�ZizddOlmjZkWnel�y&dPdQ�ZkYn0dRdS�ZjGdTdU�dUem�Znen�o�ZpGdVdW�dWen�ZqGdXdY�dY�ZrGdZd[�d[er�ZsGd\d]�d]es�ZterjuZuerjvZvd^d_�Zwd`da�Zxdbe^fdcdd�Zydedf�ZzGdgdh�dhe+�Z{d
S)ka0
Easy Install
------------

A tool for doing automatic download/extract/build of distutils-based Python
packages.  For detailed documentation, see the accompanying EasyInstall.txt
file, or visit the `EasyInstall home page`__.

__ https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html

�)�glob)�get_platform)�convert_path�
subst_vars)�DistutilsArgError�DistutilsOptionError�DistutilsError�DistutilsPlatformError)�INSTALL_SCHEMES�SCHEME_KEYS)�log�dir_util)�
first_line_re)�find_executableN)�get_config_vars�get_path)�SetuptoolsDeprecationWarning)�Command)�	run_setup)�setopt)�unpack_archive)�PackageIndex�parse_requirement_arg�
URL_SCHEME)�	bdist_egg�egg_info)�Wheel)�yield_lines�normalize_path�resource_string�ensure_directory�get_distribution�find_distributions�Environment�Requirement�Distribution�PathMetadata�EggMetadata�
WorkingSet�DistributionNotFound�VersionConflict�DEVELOP_DIST�default)�category)�samefile�easy_install�PthDistributions�extract_wininst_cfg�get_exe_prefixescCst�d�dkS)N�P�)�struct�calcsize�r7r7��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/easy_install.py�is_64bitJsr9cCsjtj�|�otj�|�}ttjd�o&|}|r:tj�||�Stj�tj�|��}tj�tj�|��}||kS)z�
    Determine if two paths reference the same file.

    Augments os.path.samefile to work on Windows and
    suppresses errors if the path doesn't exist.
    r.)�os�path�exists�hasattrr.�normpath�normcase)�p1�p2Z
both_existZuse_samefileZnorm_p1Znorm_p2r7r7r8r.Nsr.cCs
|�d�S)N�utf8)�encode��sr7r7r8�	_to_bytes^srFcCs*z|�d�WdSty$YdS0dS)N�asciiTF)rC�UnicodeErrorrDr7r7r8�isasciibs

rIcCst�|����dd�S)N�
z; )�textwrap�dedent�strip�replace)�textr7r7r8�
_one_linerjsrPc@sxeZdZdZdZdZdddddd	d
ddd
ddddddddddddddejfgZgd�Z	ddiZ
eZdd �Z
d!d"�Zd#d$�Zed%d&��Zd'd(�Zd)d*�Zd+d,�Zd-d.�Zd/d0�Zd�d1d2�Zd3d4�Zd5d6�Zd7d8�Ze�d9���Ze�d:���Ze�d;���Z d<d=�Z!d>d?�Z"d@dA�Z#dBdC�Z$dDdE�Z%dFdG�Z&e'j(dHdI��Z)d�dKdL�Z*d�dMdN�Z+dOdP�Z,d�dQdR�Z-dSdT�Z.dUdV�Z/dWdX�Z0d�dYdZ�Z1ed[d\��Z2d�d_d`�Z3dadb�Z4dcdd�Z5dedf�Z6dgdh�Z7didj�Z8dkdl�Z9e�dm���Z:e�dn�Z;d�dpdq�Z<e�dr���Z=dsdt�Z>dudv�Z?dwdx�Z@dydz�ZAd{d|�ZBd}d~�ZCdd��ZDd�d��ZEe�d���F�ZGd�d��ZHeIeId�d�d��d��ZJeId�d�d��ZKd�d��ZLdS)�r/z'Manage a download/build/install processz Find/get/install Python packagesT)zprefix=Nzinstallation prefix)�zip-ok�zzinstall package as a zipfile)�
multi-version�mz%make apps have to require() a version)�upgrade�Uz1force upgrade (searches PyPI for latest versions))zinstall-dir=�dzinstall package to DIR)zscript-dir=rEzinstall scripts to DIR)�exclude-scripts�xzDon't install scripts)�always-copy�az'Copy all needed packages to install dir)z
index-url=�iz base URL of Python Package Index)zfind-links=�fz(additional URL(s) to search for packages)zbuild-directory=�bz/download/extract/build in DIR; keep the results)z	optimize=�Ozlalso compile with optimization: -O1 for "python -O", -O2 for "python -OO", and -O0 to disable [default: -O0])zrecord=Nz3filename in which to record list of installed files)�always-unzip�Zz*don't install as a zipfile, no matter what)z
site-dirs=�Sz)list of directories where .pth files work)�editable�ez+Install specified packages in editable form)�no-deps�Nzdon't install dependencies)zallow-hosts=�Hz$pattern(s) that hostnames must match)�local-snapshots-ok�lz(allow building eggs from local checkouts)�versionNz"print version information and exit)z
no-find-linksNz9Don't load find-links defined in packages being installed�userNz!install in user site-package '%s')
rQrSrXrUrZrcrerhrjrkr`rQcCs2t�dt�d|_d|_|_d|_|_|_d|_	d|_
d|_d|_d|_
|_d|_|_|_d|_|_|_d|_|_|_d|_d|_d|_d|_d|_d|_d|_d|_d|_ t!j"r�t!j#|_$t!j%|_&nd|_$d|_&d|_'d|_(d|_)|_*d|_+i|_,d|_-|j.j/|_/|j.�0||j.�1d��dS)NzVeasy_install command is deprecated. Use build and pip and other standards-based tools.rr/)2�warnings�warn�EasyInstallDeprecationWarningrk�zip_ok�local_snapshots_ok�install_dir�
script_dir�exclude_scripts�	index_url�
find_links�build_directory�args�optimize�recordrU�always_copy�
multi_versionrc�no_deps�allow_hosts�root�prefix�	no_reportrj�install_purelib�install_platlib�install_headers�install_lib�install_scripts�install_data�install_base�install_platbase�site�ENABLE_USER_SITE�	USER_BASE�install_userbase�	USER_SITE�install_usersite�
no_find_links�
package_index�pth_file�always_copy_from�	site_dirs�installed_projects�_dry_run�distribution�verbose�_set_command_options�get_option_dict��selfr7r7r8�initialize_options�sN�

�zeasy_install.initialize_optionscCs"dd�|D�}tt|j|��dS)Ncss*|]"}tj�|�stj�|�r|VqdS�N)r:r;r<�islink)�.0�filenamer7r7r8�	<genexpr>�s�z/easy_install.delete_blockers.<locals>.<genexpr>)�list�map�_delete_path)r��blockersZextant_blockersr7r7r8�delete_blockers�s�zeasy_install.delete_blockerscCsJt�d|�|jrdStj�|�o.tj�|�}|r8tntj}||�dS)NzDeleting %s)	r�info�dry_runr:r;�isdirr��rmtree�unlink)r�r;Zis_treeZremoverr7r7r8r��szeasy_install._delete_pathcCs8djtj�}td�}d}t|jfit����t��dS)zT
        Render the Setuptools version and installation details, then exit.
        �{}.{}�
setuptoolsz=setuptools {dist.version} from {dist.location} (Python {ver})N)�format�sys�version_infor!�print�locals�
SystemExit)�ver�dist�tmplr7r7r8�_render_version�s
zeasy_install._render_versionc
Cs�|jo|��tj��d}tdd�\}}|j��|j��|j��||dd�|d|d||||t	tdd�d�|_
tjr�|j
|j
d	<|j|j
d
<n|jr�t�d�|��|��|��|�dd
dd�|jdur�|j|_|jdur�d|_|�dd�|�dd�|j�r(|j�r(|j|_|j|_|�dd�tttj�}t�|_ |j!du�r�dd�|j!�d�D�}|D]N}t"j�#|��s�t�d|�n,t|�|v�r�t$|d��n|j �%t|���qn|j&�s�|�'�|j(�p�d|_(|j dd�|_)|jt|j�fD] }||j)v�r�|j)�*d|��q�|j+du�rBdd�|j+�d�D�}ndg}|j,du�rj|j-|j(|j)|d�|_,t.|j)tj�|_/|j0du�r�t1|j0t2��r�|j0��|_0ng|_0|j3�r�|j,�4|j)tj�|j�s�|j,�5|j0�|�dd �t1|j6t7��sXz0t7|j6�|_6d|j6k�rdk�s$nt8�Wn0t8�yV}	zt$d!�|	�WYd}	~	n
d}	~	00|j&�rp|j9�spt:d"��|j;�s�t:d#��g|_<dS)$Nrr�exec_prefix���abiflags�)�	dist_name�dist_version�
dist_fullname�
py_version�py_version_short�py_version_nodot�
sys_prefixr�sys_exec_prefixr�r��userbase�usersitez6WARNING: The user site-packages directory is disabled.rqrrrvr�Fr�)rqrqr��rqrr�install)ryrycSsg|]}tj�|����qSr7)r:r;�
expanduserrM�r�rEr7r7r8�
<listcomp>-sz1easy_install.finalize_options.<locals>.<listcomp>�,z"%s (in --site-dirs) does not existz$ (in --site-dirs) is not on sys.pathzhttps://pypi.org/simple/cSsg|]}|���qSr7)rMr�r7r7r8r�C��*)�search_path�hosts)rxrxz--optimize must be 0, 1, or 2z9Must specify a build directory (-b) when using --editablez:No urls, filenames, or requirements specified (see --help))=rjr�r��splitrr��get_name�get_version�get_fullname�getattr�config_varsr�r�r�r�rkrrm�_fix_install_dir_for_user_site�expand_basedirs�expand_dirs�_expandrrrqr��set_undefined_optionsr�r�r�rr;�
get_site_dirs�
all_site_dirsr�r:r�r�appendrc�check_site_dirrt�shadow_path�insertr}r��create_indexr#�local_indexru�
isinstance�strrpZscan_egg_links�add_find_linksrx�int�
ValueErrorrvrrw�outputs)
r�r�rr�r>r�rW�	path_itemr�rdr7r7r8�finalize_options�s�

�
�

��
��
�����zeasy_install.finalize_optionscCs\|jrtjsdS|��|jdur.d}t|��|j|_|_tj	�
dd�d}|�|�dS)z;
        Fix the install_dir if "--user" was used.
        Nz$User base directory is not specified�posix�unix�_user)rkr�r��create_home_pathr�r	r�r�r:�namerN�
select_scheme)r��msg�scheme_namer7r7r8r�js
z+easy_install._fix_install_dir_for_user_sitecCsX|D]N}t||�}|durtjdks.tjdkr:tj�|�}t||j�}t|||�qdS)Nr��nt)r�r:r�r;r�rr��setattr)r��attrs�attr�valr7r7r8�
_expand_attrsys
zeasy_install._expand_attrscCs|�gd��dS)zNCalls `os.path.expanduser` on install_base, install_platbase and
        root.)r�r�r~N�r�r�r7r7r8r��szeasy_install.expand_basedirscCsgd�}|�|�dS)z+Calls `os.path.expanduser` on install dirs.)r�r�r�r�r�r�Nr�)r��dirsr7r7r8r��szeasy_install.expand_dirsc	Cs�|r|�dtj�|j|jjkr,t�|j�z�|jD]}|�||j�q4|j	r�|j
}|jr�t|j�}t
t|��D]}|||d�||<qrddlm}|�|j|j	|fd|j	�|��Wt�|jj�nt�|jj�0dS)NzXWARNING: The easy_install command is deprecated and will be removed in a future version.r)�	file_utilz'writing list of installed files to '%s')�announcer�WARNr�r��
set_verbosityrwr/r|ryr�r~�len�range�	distutilsr��execute�
write_file�warn_deprecated_options)r�Zshow_deprecation�specr��root_len�counterr�r7r7r8�run�s2�

��
zeasy_install.runcCsBzt��}Wn ty,t�dtj�}Yn0tj�|j	d|�S)z�Return a pseudo-tempname base in the install directory.
        This code is intentionally naive; if a malicious party can write to
        the target directory you're already in deep doodoo.
        rztest-easy-install-%s)
r:�getpid�	Exception�random�randintr��maxsizer;�joinrq)r��pidr7r7r8�pseudo_tempname�s
zeasy_install.pseudo_tempnamecCsdSr�r7r�r7r7r8r�sz$easy_install.warn_deprecated_optionsc	CsJt|j�}tj�|d�}tj�|�sRzt�|�WnttfyP|�	�Yn0||j
v}|sp|jsp|��}nb|�
�d}tj�|�}z*|r�t�|�t|d���t�|�Wnttfy�|�	�Yn0|s�|js�tj�dd�}t�|j|j|�|�r|jdu�r$t||j
�|_nd|_|j�r@tj�|��s@d|_||_dS)z;Verify that self.install_dir is .pth-capable dir, if needed�easy-install.pthz.write-test�w�
PYTHONPATHr�N)rrqr:r;rr<�makedirs�OSError�IOError�cant_write_to_targetr�r{�check_pth_processingrr��open�close�environ�getrrm�_easy_install__no_default_msgr�r0)r��instdirr�Zis_site_dirZtestfileZtest_exists�
pythonpathr7r7r8r��s:





zeasy_install.check_site_diraS
        can't create or remove files in install directory

        The following error occurred while trying to add or remove files in the
        installation directory:

            %s

        The installation directory you specified (via --install-dir, --prefix, or
        the distutils default setting) was:

            %s
        z�
        This directory does not currently exist.  Please create it and try again, or
        choose a different installation directory (using the -d or --install-dir
        option).
        a�
        Perhaps your account does not have write access to this directory?  If the
        installation directory is a system-owned directory, you may need to sign in
        as the administrator or "root" account.  If you do not have administrative
        access to this machine, you may wish to choose a different installation
        directory, preferably one that is listed in your PYTHONPATH environment
        variable.

        For information on other options, you may wish to consult the
        documentation at:

          https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html

        Please make the appropriate changes for your system and try again.
        cCsP|jt��d|jf}tj�|j�s6|d|j7}n|d|j7}t	|��dS)N�rJ)
�_easy_install__cant_write_msgr��exc_inforqr:r;r<�_easy_install__not_exists_id�_easy_install__access_msgr)r�r�r7r7r8rs
z!easy_install.cant_write_to_targetc
	Cs(|j}t�d|�|��d}|d}tj�|�}td�d}z6|rNt�|�tj�	|�}tj
|dd�t|d�}Wn tt
fy�|��Y�n|0�z6|�|jfit����|��d	}tj}tjd
k�rtj�|�\}}	tj�|d�}
|	��dk�otj�|
�}|�r|
}d
dlm}||dddgd
�tj�|��r�t�d|�W|�rZ|��tj�|��rrt�|�tj�|��r�t�|�dSW|�r�|��tj�|��r�t�|�tj�|��rt�|�n@|�r�|��tj�|��r�t�|�tj�|��rt�|�0|j�s$t�d|�dS)z@Empirically verify whether .pth files are supported in inst. dirz Checking .pth file support in %s�.pthz.okzz
            import os
            f = open({ok_file!r}, 'w')
            f.write('OK')
            f.close()
            rJT)�exist_okrNr��pythonw.exe�
python.exer)�spawnz-Ez-c�passz-TEST PASSED: %s appears to support .pth filesz+TEST FAILED: %s does NOT support .pth filesF)rqrr�rr:r;r<rPr��dirnamerrrrr�writer�r�rr��
executabler�r�r�lower�distutils.spawnr*r{rm)
r�rr�Zok_fileZ	ok_existsr�r,r]r.�basenameZaltZuse_altr*r7r7r8rsv�

��

�
�
z!easy_install.check_pth_processingc	CsV|jsH|�d�rH|�d�D],}|�d|�r.q|�|||�d|��q|�|�dS)z=Write all the scripts for `dist`, unless scripts are excluded�scriptszscripts/N)rs�metadata_isdir�metadata_listdir�install_script�get_metadata�install_wrapper_scripts)r�r��script_namer7r7r8�install_egg_scriptsSs�z easy_install.install_egg_scriptscCsTtj�|�rDt�|�D]*\}}}|D]}|j�tj�||��q$qn|j�|�dSr�)r:r;r��walkr�r�r)r�r;�baser��filesr�r7r7r8�
add_outputas
zeasy_install.add_outputcCs|jrtd|f��dS)NzjInvalid argument %r: you can't use filenames or URLs with --editable (except via the --find-links option).)rcr�r�rr7r7r8�not_editableis��zeasy_install.not_editablecCs<|js
dStj�tj�|j|j��r8td|j|jf��dS)Nz2%r already exists in %s; can't do a checkout there)rcr:r;r<rrv�keyrr>r7r7r8�check_editableqs
��zeasy_install.check_editablec	csJtjdd�}z"t|�VWtj�|�o,t|�ntj�|�oBt|�0dS)Nz
easy_install-)r)�tempfile�mkdtempr�r:r;r<r�)r��tmpdirr7r7r8�_tmpdir{szeasy_install._tmpdirFc	CsX|����8}t|t�s�t|�rV|�|�|j�||�}|�d|||d�Wd�Stj	�
|�r�|�|�|�d|||d�Wd�St|�}|�|�|j�
|||j|j|j|j�}|dur�d|}|jr�|d7}t|��nN|jtk�r|�|||d�|Wd�S|�||j||�Wd�SWd�n1�sJ0YdS)NTz+Could not find suitable distribution for %rz2 (--always-copy skips system and development eggs)�Using)rEr�r$rr?r��download�install_itemr:r;r<rrAZfetch_distributionrUrcrzr�r�
precedencer+�process_distribution�location)r�r�depsrD�dlr�r�r7r7r8r/�s0

 
 

�
zeasy_install.easy_installcCs
|p|j}|ptj�|�|k}|p,|�d�}|pT|jduoTtj�t|��t|j�k}|r�|s�|j|jD]}|j	|krjq�qjd}t
�dtj�|��|r�|�
|||�}|D]}|�|||�q�n |�|�g}|�||d|d�|du�r|D]}||vr�|Sq�dS)N�.eggTz
Processing %srrF)rzr:r;r,�endswithr�rr��project_namerKrr�r1�install_eggsrJ�egg_distribution)r�rrGrDrLZinstall_neededr��distsr7r7r8rH�s2

��

zeasy_install.install_itemcCs<t|}tD]*}d|}t||�durt||||�qdS)z=Sets the install directories by applying the install schemes.�install_N)r
rr�r�)r�r��schemer@�attrnamer7r7r8r��s
zeasy_install.select_schemec
Gs�|�|�|j�|�||j|jvr2|j�|�|j�|�|�|�||j|j<t�	|j
||g|�R��|�d�r�|js�|j�
|�d��|s�|js�dS|dur�|j|jkr�t�d|�dS|dus�||vr�|��}tt|��}t�	d|�ztg��|g|j|j�}Wnft�yB}ztt|��|�WYd}~n<d}~0t�yt}zt|���|�WYd}~n
d}~00|j�s�|j�r�|D]"}|j|jv�r�|�|����q�t�	d|�dS)Nzdependency_links.txtzSkipping dependencies for %szProcessing dependencies for %sz'Finished processing dependencies for %s)�
update_pthr��addr�r@�remover9r�rr��installation_report�has_metadatar�r��get_metadata_linesrzrm�as_requirementr$r�r(�resolver/r)rr*�reportr�)r��requirementr�rLr�ZdistreqZdistrosrdr7r7r8rJ�sH


��
�"$z!easy_install.process_distributioncCs2|jdur|jS|�d�r dS|�d�s.dSdS)Nznot-zip-safeTzzip-safeF)ror[�r�r�r7r7r8�should_unzip�s


zeasy_install.should_unzipcCs�tj�|j|j�}tj�|�r:d}t�||j|j|�|Stj�|�rL|}nRtj�	|�|krft�
|�t�|�}t|�dkr�tj�||d�}tj�|�r�|}t
|�t�||�|S)Nz<%r already exists in %s; build directory %s will not be keptr!r)r:r;rrvr@r<rrmr�r,r��listdirrr �shutil�move)r�r�
dist_filename�
setup_base�dstr��contentsr7r7r8�
maybe_moves$�

zeasy_install.maybe_movecCs,|jr
dSt���|�D]}|j|�qdSr�)rs�ScriptWriter�best�get_args�write_script)r�r�rwr7r7r8r7sz$easy_install.install_wrapper_scriptscCsNt|���}t||�}|r8|�|�t�}t�|�|}|�|t|�d�dS)z/Generate a legacy script wrapper and install itr^N)	r�r]�is_python_script�_load_templater�rk�
get_headerrnrF)r�r�r8�script_text�dev_pathrZ	is_script�bodyr7r7r8r5"s
zeasy_install.install_scriptcCs(d}|r|�dd�}td|�}|�d�S)z�
        There are a couple of template scripts in the package. This
        function loads one of them and prepares it for use.
        zscript.tmplz.tmplz (dev).tmplr��utf-8)rNr�decode)rsr�Z	raw_bytesr7r7r8rp,s

zeasy_install._load_template�tr7cs����fdd�|D��t�d|�j�tj��j|�}��|��jrLdSt	�}t
|�tj�|�rpt�|�t
|d|��}|�|�Wd�n1s�0Yt|d|�dS)z1Write an executable file to the scripts directorycsg|]}tj��j|��qSr7)r:r;rrr�r�rYr�r7r8r�>r�z-easy_install.write_script.<locals>.<listcomp>zInstalling %s script to %sNri�)r�rr�rrr:r;rr=r��
current_umaskr r<r�rr-�chmod)r�r8ri�moder��target�maskr]r7r�r8rn;s�

(zeasy_install.write_scriptc	CsR|j|j|jd�}z||��dd�}Wnty<Yn0|||�gS|}tj�|�rt|�d�stt	|||j
�ntj�|�r�tj�|�}|�
|�r�|jr�|dur�|�|||�}tj�|d�}tj�|��s$ttj�|dd��}|s�tdtj�|���t|�dk�rtdtj�|���|d	}|j�rBt�|�||��gS|�||�SdS)
N)rN�.exez.whl����.pyzsetup.pyr�z"Couldn't find a setup script in %sr!zMultiple setup scripts in %sr)�install_egg�install_exe�
install_wheelr/�KeyErrorr:r;�isfilerOr�unpack_progressr��abspath�
startswithrvrjrr<rrrrcrr��report_editable�build_and_install)	r�rrfrDZ
installer_mapZinstall_distrg�setup_scriptZsetupsr7r7r8rQOsT��
��
��
��zeasy_install.install_eggscCs>tj�|�r"t|tj�|d��}ntt�|��}tj	||d�S)N�EGG-INFO)�metadata)
r:r;r�r&rr'�	zipimport�zipimporterr%�
from_filename)r��egg_pathr�r7r7r8rR�s�zeasy_install.egg_distributionc	Cs�tj�|jtj�|��}tj�|�}|js2t|�|�|�}t	||��sxtj�
|�rrtj�|�srtj
||jd�n"tj�|�r�|�tj|fd|�z�d}tj�
|�r�|�|�r�tjd}}ntjd}}nL|�|�r�|�|�|jd}}n*d}|�|��rtjd}}ntjd}}|�|||f|dtj�|�tj�|�f�t||d	�Wn"t�yvt|dd	��Yn0|�|�|�|�S)
N�r��	Removing FZMovingZCopyingZ
ExtractingTz	 %s to %s��fix_zipimporter_caches)r:r;rrqr1r�r�r rRr.r�r�r
�remove_treer<rr�r�rdre�copytreerb�mkpath�unpack_and_compile�copy2r,�update_dist_cachesrr=)r�r�rD�destinationr�Znew_dist_is_zippedr]rTr7r7r8r��s^
�
�




���

zeasy_install.install_eggcsPt|�}|durtd|��td|�dd�|�dd�t�d�}tj�||��d�}||_	|d}tj�|d�}tj�|d	�}t
|�t||�|_|�
||�tj�|��st|d
�}	|	�d�|�d�D].\}
}|
dkr�|	�d
|
�dd���|f�q�|	��tj�|d��|��fdd�t�|�D��tj|||j|jd�|�||�S)Nz(%s is not a valid distutils Windows .exer�r�rj)rPrj�platformrNz.tmpr��PKG-INFOrzMetadata-Version: 1.0
�target_versionz%s: %s
�_�-r2csg|]}tj��|d��qS)r)r:r;r)r�rw�rrr7r8r��s�z,easy_install.install_exe.<locals>.<listcomp>)r�r�)r1rr%rrr:r;r�egg_namerKr r&�	_provider�
exe_to_eggr<rr-�itemsrN�titlerr�rkrmr�make_zipfiler�r�r�)r�rfrD�cfgr�r��egg_tmpZ	_egg_infoZpkg_infr]�k�vr7r�r8r��sB�
�

 ��zeasy_install.install_execs8t|��g�g�i������fdd�}t|�|�g}�D]n}|���d�r<|�d�}|d}t�|d�d|d<tjj	�g|�R�}��
|�|�
|�t�||�q<|���t�
tj�	�d�t��|��dD]Z}	t�|	r�tj�	�d|	d	�}
tj�|
�s�t|
d
�}|�d�	t�|	�d�|��q�dS)
z;Extract a bdist_wininst to the directories an egg would usecs�|��}�D]�\}}|�|�r||t|�d�}|�d�}tjj�g|�R�}|��}|�d�sl|�d�r�t�	|d�|d<d�tj�
|d�d<��|�n4|�d�r�|dkr�d�tj�
|d�d<��|�|Sq|�d	�s�t�
d
|�dS)N�/�.pyd�.dll���r!rr��SCRIPTS/r&zWARNING: can't process %s)r/r�rr�r:r;rrOr�strip_module�splitextr�rrm)�srcrhrE�old�new�partsrM�r��native_libs�prefixes�
to_compile�	top_levelr7r8�process�s$




z(easy_install.exe_to_egg.<locals>.processr�r�r�r�r�)r�r�z.txtrrJN)r2rr/rOr�rr�r:r;rr�Z
write_stub�byte_compileZwrite_safety_flagZanalyze_eggr�r<rr-r)r�rfr�r�Zstubs�resr��resourceZpyfiler��txtr]r7r�r8r��s8




�

zeasy_install.exe_to_eggc
Cs�t|�}|��sJ�tj�|j|���}tj�|�}|jsBt	|�tj�
|�rltj�|�sltj
||jd�n"tj�|�r�|�tj|fd|�z:|�|j|fdtj�|�tj�|�f�Wt|dd�nt|dd�0|�|�|�|�S)Nr�r�zInstalling %s to %sFr�)r�
is_compatibler:r;rrqr�r�r�r r�r�r
r�r<rr�Zinstall_as_eggr1r,r�r=rR)r��
wheel_pathrD�wheelr�r7r7r8r�$s4�

��	
zeasy_install.install_wheela(
        Because this distribution was installed --multi-version, before you can
        import modules from this package in an application, you will need to
        'import pkg_resources' and then use a 'require()' call similar to one of
        these examples, in order to select the desired version:

            pkg_resources.require("%(name)s")  # latest installed version
            pkg_resources.require("%(name)s==%(version)s")  # this exact version
            pkg_resources.require("%(name)s>=%(version)s")  # this version or higher
        z�
        Note also that the installation directory must be on sys.path at runtime for
        this to work.  (e.g. by being the application's script directory, by being on
        PYTHONPATH, or by being added to sys.path by your code.)
        �	Installedc	Cs^d}|jr>|js>|d|j7}|jtttj�vr>|d|j7}|j	}|j
}|j}d}|t�S)z9Helpful installation message for display to package usersz
%(what)s %(eggloc)s%(extras)srJr�)
r{r��_easy_install__mv_warningrqr�rr�r;�_easy_install__id_warningrKrPrjr�)	r��reqr��whatr�Zegglocr�rj�extrasr7r7r8rZRsz easy_install.installation_reportaR
        Extracted editable version of %(spec)s to %(dirname)s

        If it uses setuptools in its setup script, you can activate it in
        "development" mode by going to that directory and running::

            %(python)s setup.py develop

        See the setuptools documentation for the "develop" command for more info.
        cCs"tj�|�}tj}d|jt�S)NrJ)r:r;r,r�r.�_easy_install__editable_msgr�)r�rr�r,�pythonr7r7r8r�kszeasy_install.report_editablec
Cs�tj�dt�tj�dt�t|�}|jdkrNd|jd}|�dd|�n|jdkrd|�dd�|jrv|�dd	�t	�
d
|t|�dd�d�|��zt
||�Wn:ty�}z"td|jdf�|�WYd}~n
d}~00dS)
Nzdistutils.command.bdist_eggzdistutils.command.egg_infor�r�r!rr�z-qz-nz
Running %s %s� zSetup script exited with %s)r��modules�
setdefaultrrr�r�r�r�rr�rrrr�rrw)r�r�rgrwr�r7r7r8rps*

���zeasy_install.run_setupc		Cs�ddg}tjdtj�|�d�}z�|�tj�|��|�|�|�|||�t|g�}g}|D]&}||D]}|�|�	|j
|��qhq\|s�|js�t�
d|�|Wt|�t�|j�St|�t�|j�0dS)Nrz
--dist-dirz
egg-dist-tmp-)r�dirz+No eggs found in %s (setup script problem?))rBrCr:r;r,�_set_fetcher_optionsr�rr#r�rKr�rrmr�r�r�)	r�r�rgrw�dist_dirZall_eggs�eggsr@r�r7r7r8r��s2�


���zeasy_install.build_and_installc	Csh|j�d���}d}i}|��D]\}}||vr2q |d||<q t|d�}tj�|d�}t�	||�dS)a
        When easy_install is about to run bdist_egg on a source dist, that
        source dist might have 'setup_requires' directives, requiring
        additional fetching. Ensure the fetcher options given to easy_install
        are available to that command as well.
        r/)rur�rtrxr}r!)r/z	setup.cfgN)
r�r��copyr��dictr:r;rrZedit_config)	r�r;Zei_optsZfetch_directivesZ
fetch_optionsr@r��settingsZcfg_filenamer7r7r8r��s	
z!easy_install._set_fetcher_optionscCsL|jdurdS|j|jD]J}|js2|j|jkr2qt�d|�|j�|�|j|jvr|j�|j�q|js�|j|jjvr�t�d|�n2t�d|�|j�	|�|j|jvr�|j�
|j�|jr�dS|j��|jdkr�dSt
j�|jd�}t
j�|��rt
�|�t|d��(}|�|j�|j�d�Wd�n1�s>0YdS)Nz&Removing %s from easy-install.pth filez4%s is already the active version in easy-install.pthz"Adding %s to easy-install.pth filer��setuptools.pth�wtrJ)r�r@r{rKrr�rYr��pathsrXr�r��saver:r;rrqr�r�rr-�
make_relative)r�r�rWr�r]r7r7r8rW�s:
�


zeasy_install.update_pthcCst�d||�|S)NzUnpacking %s to %s)r�debug)r�r�rhr7r7r8r��szeasy_install.unpack_progresscsdg�g����fdd�}t|||������js`�D]&}t�|�tjdBd@}t||�q8dS)NcsZ|�d�r |�d�s ��|�n|�d�s4|�d�r>��|���||��jrV|pXdS)Nr��	EGG-INFO/r�z.so)rOr�r�r�r�)r�rh�r�Zto_chmodr�r7r8�pf�s
z+easy_install.unpack_and_compile.<locals>.pfimi�)rr�r�r:�stat�ST_MODErz)r�r�r�r�r]r{r7r�r8r��s
zeasy_install.unpack_and_compilec	Csvtjr
dSddlm}zLt�|jd�||dd|jd�|jrT|||jd|jd�Wt�|j�nt�|j�0dS)Nr)r�r!)rx�forcer�)	r��dont_write_bytecode�distutils.utilr�rr�r�r�rx)r�r�r�r7r7r8r��s�zeasy_install.byte_compilea�
        bad install directory or PYTHONPATH

        You are attempting to install a package to a directory that is not
        on PYTHONPATH and which Python does not read ".pth" files from.  The
        installation directory you specified (via --install-dir, --prefix, or
        the distutils default setting) was:

            %s

        and your PYTHONPATH environment variable currently contains:

            %r

        Here are some of your options for correcting the problem:

        * You can choose a different installation directory, i.e., one that is
          on PYTHONPATH or supports .pth files

        * You can add the installation directory to the PYTHONPATH environment
          variable.  (It must then also be on PYTHONPATH whenever you run
          Python and want to use the package(s) you are installing.)

        * You can set up the installation directory to support ".pth" files by
          using one of the approaches described here:

          https://setuptools.readthedocs.io/en/latest/deprecated/easy_install.html#custom-installation-locations


        Please make the appropriate changes for your system and try again.
        cCsb|js
dSttj�d��}|j��D]8\}}|�|�r$tj�|�s$|�	d|�t�
|d�q$dS)zCreate directories under ~.N�~zos.makedirs('%s', 0o700)i�)rkrr:r;r�r�r�r�r��debug_printr)r��homer�r;r7r7r8r�)szeasy_install.create_home_pathz/$base/lib/python$py_version_short/site-packagesz	$base/binr��r�z$base/Lib/site-packagesz
$base/ScriptscGs�|�d�j}|jrd|��}|j|d<|j�tj|j�}|�	�D]$\}}t
||d�dur>t|||�q>ddlm
}|D]B}t
||�}|durt|||�}tjdkr�tj�|�}t|||�qtdS)Nr�r;r)rr�)�get_finalized_commandr�rr�r
rr:r��DEFAULT_SCHEMEr�r�r�r�rr;r�)r�r�r�rUr�r�rr7r7r8r�?s 



zeasy_install._expand)T)F)F)T)N)rwr7)r�)M�__name__�
__module__�__qualname__�__doc__�descriptionZcommand_consumes_argumentsr�r��user_options�boolean_options�negative_optrr�r�r�r��staticmethodr�r�r�r�r�r�r	rrr�rKrL�lstripr"r$r%rrr9r=r?rA�
contextlib�contextmanagerrEr/rHr�rJrbrjr7r5rprnrQrRr�r�r�r�r�r�rZr�r�rr�r�rWr�r�r�rMrr�r�r
r�r�r7r7r7r8r/ns��5	

	
-	;


!
$�
)	



3	6.5

) 
���r/cCs tj�dd��tj�}td|�S)Nrr�)r:rrr��pathsep�filter)r�r7r7r8�_pythonpathVsr�c	s�g���t��tjg}tjtjkr0|�tj�|D]�}|s>q4tjdvr`��tj�	|dd��nVtj
dkr���tj�	|ddjtj�d�tj�	|dd�g�n��|tj�	|dd�g�tjdkr�q4d	|vr�q4tj
�d
�}|s�q4tj�	|ddd
jtj�d�}��|�q4td�td�f}���fdd�|D��tj�rB��tj�t�t����t���Wd�n1�sr0Yttt�����S)z&
    Return a list of 'site' dirs
    )Zos2emxZriscos�Libz
site-packagesr��libzpython{}.{}zsite-python�darwinzPython.framework�HOME�Library�Pythonr��purelib�platlibc3s|]}|�vr|VqdSr�r7r���sitedirsr7r8r��r�z get_site_dirs.<locals>.<genexpr>N)�extendr�r�rr�r�r�r:r;r�sepr�r�rrrr�r�r�r��suppress�AttributeError�getsitepackagesr�r�r)r�rr�Zhome_spZ	lib_pathsr7rr8r�[s^


��
�

�.r�ccs�i}|D]�}t|�}||vrqd||<tj�|�s4qt�|�}||fV|D]�}|�d�s\qL|dvrfqLttj�||��}tt	|��}|�
�|D]L}|�d�r�q�t|���}||vr�q�d||<tj�|�s�q�|t�|�fVq�qLqdS)zBYield sys.path directories that might contain "old-style" packagesr!r&)rr��importN)
rr:r;r�rcrOrrr�rrr��rstrip)�inputs�seenr,r<r�r]�lines�liner7r7r8�expand_paths�s8



rcCs`t|d�}�zDt�|�}|dur.W|��dS|d|d|d}|dkr\W|��dS|�|d�t�d|�d��\}}}|dvr�W|��dS|�|d|�d	d	d
�}t�	|�}z<|�|�}	|	�
dd�d
}
|
�t�
��}
|�t�|
��Wn"tj�yYW|��dS0|�d��r6|�d��sDW|��dS|W|��S|��0dS)znExtract configuration data from a bdist_wininst .exe

    Returns a configparser.RawConfigParser, or None
    �rbN�	���z<iii)izV4i{V4r�)rjr��r!rr��Setup)r�zipfile�_EndRecDatar�seekr5�unpack�read�configparser�RawConfigParserr�rvr��getfilesystemencoding�	read_file�io�StringIO�Error�has_section)rfr]�endrecZ	prepended�tagZcfglenZbmlen�initr��part�configr7r7r8r1�sL

���


���r1c	Cs6gd�}t�|�}z�|��D]�}|j}|�d�}t|�dkrx|ddkrx|d�d�rx|�dd�|d	d��d
f�q�t|�dks|�d�s�q|�d�r�q|d�	�d
vr|�
|���}t|�D]8}|�
��dd�}|�d�s�|�d|d|fdf�q�qW|��n
|��0dd�|D�}|��|��|S)z4Get exe->egg path translations for a given .exe file))zPURELIB/r�)zPLATLIB/pywin32_system32r�)zPLATLIB/r�)r�zEGG-INFO/scripts/)zDATA/lib/site-packagesr�r�r�r�r�r!z	.egg-inforNr�r&z
-nspkg.pth)ZPURELIBZPLATLIB�\r	z%s/%s/r�cSsg|]\}}|��|f�qSr7)r/)r�rY�yr7r7r8r�r�z$get_exe_prefixes.<locals>.<listcomp>)r�ZipFile�infolistr�r�rrOr�r�upperrrvrrMrNr�r�r�sort�reverse)Zexe_filenamer�rRr�r�r�ri�pthr7r7r8r2�s2



 r2c@sReZdZdZdZddd�Zdd�Zdd	�Zed
d��Z	dd
�Z
dd�Zdd�ZdS)r0z)A .pth file with Distribution paths in itFr7cCsl||_ttt|��|_ttj�|j��|_|�	�t
�|gdd�t|j
�D]}tt|jt|d���qLdS)NT)r�r�r�rrr:r;r,�basedir�_loadr#�__init__rr�rXr")r�r�rr;r7r7r8r3"szPthDistributions.__init__cCs�g|_d}t�|j�}tj�|j�r�t|jd�}|D]�}|�	d�rHd}q4|�
�}|j�|�|��r4|���	d�rtq4t
tj�|j|��}|jd<tj�|�r�||vr�|j��d|_q4d||<q4|��|jr�|s�d|_|jr�|jd��s�|j��q�dS)NF�rtr	T�#r�r!)r�r��fromkeysrr:r;r�r�rr�r
r�rMrrr1r<�pop�dirtyr)r�Z
saw_importrr]rr;r7r7r8r2+s4
�


zPthDistributions._loadcCs�|js
dStt|j|j��}|r�t�d|j�|�|�}d�	|�d}t
j�|j�r`t
�
|j�t|jd��}|�|�Wd�q�1s�0Yn(t
j�|j�r�t�d|j�t
�
|j�d|_dS)z$Write changed .pth file back to diskNz	Saving %srJr�zDeleting empty %sF)r8r�r�r�r�rr�r��_wrap_linesrr:r;r�r�rr-r<)r�Z	rel_pathsr
�datar]r7r7r8r�Js
*zPthDistributions.savecCs|Sr�r7)r
r7r7r8r9`szPthDistributions._wrap_linescCsN|j|jvo$|j|jvp$|jt��k}|r>|j�|j�d|_t�||�dS)z"Add `dist` to the distribution mapTN)	rKr�rr:�getcwdr�r8r#rX)r�r��new_pathr7r7r8rXds�zPthDistributions.addcCs2|j|jvr"|j�|j�d|_qt�||�dS)z'Remove `dist` from the distribution mapTN)rKr�rYr8r#rar7r7r8rYrszPthDistributions.removecCs�tj�t|��\}}t|j�}|g}tjdkr2dp6tj}t|�|kr�||jkrl|�tj	�|�
�|�|�Stj�|�\}}|�|�q8|S)Nr�)r:r;r�rrr1�altseprr��curdirr/r)r�r;�npath�lastZbaselenr�rr7r7r8r�ys


zPthDistributions.make_relativeN)r7)
r�r�r�r�r8r3r2r�r�r9rXrYr�r7r7r7r8r0s
	
r0c@s(eZdZedd��Zed�Zed�ZdS)�RewritePthDistributionsccs$|jV|D]
}|Vq|jVdSr�)�prelude�postlude)�clsr
rr7r7r8r9�sz#RewritePthDistributions._wrap_linesz?
        import sys
        sys.__plen = len(sys.path)
        z�
        import sys
        new = sys.path[sys.__plen:]
        del sys.path[sys.__plen:]
        p = getattr(sys, '__egginsert', 0)
        sys.path[p:p] = new
        sys.__egginsert = p + len(new)
        N)r�r�r��classmethodr9rPrBrCr7r7r7r8rA�s
rAZSETUPTOOLS_SYS_PATH_TECHNIQUE�rawZrewritecCs ttjt�rtSt�tj���S)z_
    Return a regular expression based on first_line_re suitable for matching
    strings.
    )r�r�patternr��re�compilervr7r7r7r8�_first_line_re�srJcCs\|tjtjfvr.tjdkr.t|tj�||�St��\}}}|d|dd||ff�dS)Nr�rr!z %s %s)	r:r�rYr�rzr��S_IWRITEr�r#)�func�arg�exc�etZevr�r7r7r8�
auto_chmod�s
rPcCs.t|�}t|tj�|r"t|�nt|�dS)aa

    Fix any globally cached `dist_path` related data

    `dist_path` should be a path of a newly installed egg distribution (zipped
    or unzipped).

    sys.path_importer_cache contains finder objects that have been cached when
    importing data from the original distribution. Any such finders need to be
    cleared since the replacement distribution might be packaged differently,
    e.g. a zipped egg distribution might get replaced with an unzipped egg
    folder or vice versa. Having the old finders cached may then cause Python
    to attempt loading modules from the replacement distribution using an
    incorrect loader.

    zipimport.zipimporter objects are Python loaders charged with importing
    data packaged inside zip archives. If stale loaders referencing the
    original distribution, are left behind, they can fail to load modules from
    the replacement distribution. E.g. if an old zipimport.zipimporter instance
    is used to load data from a new zipped egg archive, it may cause the
    operation to attempt to locate the requested data in the wrong location -
    one indicated by the original distribution's zip archive directory
    information. Such an operation may then fail outright, e.g. report having
    read a 'bad local file header', or even worse, it may fail silently &
    return invalid data.

    zipimport._zip_directory_cache contains cached zip archive directory
    information for all existing zipimport.zipimporter instances and all such
    instances connected to the same archive share the same cached directory
    information.

    If asked, and the underlying Python implementation allows it, we can fix
    all existing zipimport.zipimporter instances instead of having to track
    them down and remove them one by one, by updating their shared cached zip
    archive directory information. This, of course, assumes that the
    replacement distribution is packaged as a zipped egg.

    If not asked to fix existing zipimport.zipimporter instances, we still do
    our best to clear any remaining zipimport.zipimporter related cached data
    that might somehow later get used when attempting to load data from the new
    distribution and thus cause such load operations to fail. Note that when
    tracking down such remaining stale data, we can not catch every conceivable
    usage from here, and we clear only those that we know of and have found to
    cause problems if left alive. Any remaining caches should be updated by
    whomever is in charge of maintaining them, i.e. they should be ready to
    handle us replacing their zip archives with new distributions at runtime.

    N)r�_uncacher��path_importer_cache�!_replace_zip_directory_cache_data�*_remove_and_clear_zip_directory_cache_data)�	dist_pathr��normalized_pathr7r7r8r��s
<
r�cCsPg}t|�}|D]:}t|�}|�|�r|||d�tjdfvr|�|�q|S)ap
    Return zipimporter cache entry keys related to a given normalized path.

    Alternative path spellings (e.g. those using different character case or
    those using alternative path separators) related to the same path are
    included. Any sub-path entries are included as well, i.e. those
    corresponding to zip archives embedded in other zip archives.

    r!r�)rrr�r:rr�)rV�cache�result�
prefix_len�p�npr7r7r8�"_collect_zipimporter_cache_entries
s

�r\cCs@t||�D]0}||}||=|o(|||�}|dur
|||<q
dS)a�
    Update zipimporter cache data for a given normalized path.

    Any sub-path entries are processed as well, i.e. those corresponding to zip
    archives embedded in other zip archives.

    Given updater is a callable taking a cache entry key and the original entry
    (after already removing the entry from the cache), and expected to update
    the entry and possibly return a new one to be inserted in its place.
    Returning None indicates that the entry should not be replaced with a new
    one. If no updater is given, the cache entries are simply removed without
    any additional processing, the same as if the updater simply returned None.

    N)r\)rVrW�updaterrZ�	old_entryZ	new_entryr7r7r8�_update_zipimporter_caches
r_cCst||�dSr�)r_)rVrWr7r7r8rQ>srQcCsdd�}t|tj|d�dS)NcSs|��dSr�)�clear�r;r^r7r7r8�2clear_and_remove_cached_zip_archive_directory_dataCszf_remove_and_clear_zip_directory_cache_data.<locals>.clear_and_remove_cached_zip_archive_directory_data�r]�r_r��_zip_directory_cache)rVrbr7r7r8rTBs
�rTZ__pypy__cCsdd�}t|tj|d�dS)NcSs&|��t�|�|�tj|�|Sr�)r`r�r��updatererar7r7r8�)replace_cached_zip_archive_directory_dataYs
zT_replace_zip_directory_cache_data.<locals>.replace_cached_zip_archive_directory_datarcrd)rVrgr7r7r8rSXs

�rS�<string>c	Cs2zt||d�Wnttfy(YdS0dSdS)z%Is this string a valid Python script?�execFTN)rI�SyntaxError�	TypeError)rOr�r7r7r8�	is_pythonks
rlc	Cs`z<tj|dd��}|�d�}Wd�n1s00YWnttfyV|YS0|dkS)zCDetermine if the specified executable is a .sh (contains a #! line)zlatin-1)�encodingr�N�#!)r rrrr)r.�fp�magicr7r7r8�is_shus,
rqcCst�|g�S)z@Quote a command line argument according to Windows parsing rules��
subprocess�list2cmdline)rMr7r7r8�nt_quote_argsrucCsH|�d�s|�d�rdSt||�r&dS|�d�rDd|��d��vSdS)zMIs this text, as a whole, a Python script? (as opposed to shell/bat/etc.
    r��.pywTrnr�rF)rOrlr��
splitlinesr/)rrr�r7r7r8ro�s

ro)rzcGsdSr�r7)rwr7r7r8�_chmod�srxc
CsTt�d||�zt||�Wn2tjyN}zt�d|�WYd}~n
d}~00dS)Nzchanging mode of %s to %ozchmod failed: %s)rr�rxr:�error)r;r{rdr7r7r8rz�s
rzc@s�eZdZdZgZe�Zedd��Zedd��Z	edd��Z
edd	��Zed
d��Zdd
�Z
edd��Zdd�Zedd��Zedd��ZdS)�CommandSpeczm
    A command spec for a #! header, specified as a list of arguments akin to
    those passed to Popen.
    cCs|S)zV
        Choose the best CommandSpec class based on environmental conditions.
        r7�rDr7r7r8rl�szCommandSpec.bestcCstj�tj�}tj�d|�S)N�__PYVENV_LAUNCHER__)r:r;r>r�r.rr)rD�_defaultr7r7r8�_sys_executable�szCommandSpec._sys_executablecCs:t||�r|St|t�r ||�S|dur0|��S|�|�S)zg
        Construct a CommandSpec from a parameter to build_scripts, which may
        be None.
        N)r�r��from_environment�from_string)rD�paramr7r7r8�
from_param�s

zCommandSpec.from_paramcCs||��g�Sr�)r~r{r7r7r8r�szCommandSpec.from_environmentcCstj|fi|j��}||�S)z}
        Construct a command spec from a simple string representing a command
        line parseable by shlex.split.
        )�shlexr��
split_args)rD�stringr�r7r7r8r��szCommandSpec.from_stringcCs8t�|�|��|_t�|�}t|�s4dg|jdd�<dS)Nz-xr)r�r��_extract_options�optionsrsrtrI)r�rr�cmdliner7r7r8�install_options�s
zCommandSpec.install_optionscCs:|d��d}t��|�}|r.|�d�p0dnd}|��S)zH
        Extract any options from the first line of the script.
        rJrr!r�)rwrJ�match�grouprM)Zorig_script�firstr�r�r7r7r8r��szCommandSpec._extract_optionscCs|�|t|j��Sr�)�_renderr�r�r�r7r7r8�	as_header�szCommandSpec.as_headercCs6d}|D](}|�|�r|�|�r|dd�Sq|S)Nz"'r!r�)r�rO)�itemZ_QUOTES�qr7r7r8�
_strip_quotes�s
zCommandSpec._strip_quotescCs t�dd�|D��}d|dS)Ncss|]}t�|���VqdSr�)rzr�rM)r�r�r7r7r8r��sz&CommandSpec._render.<locals>.<genexpr>rnrJrr)r�r�r7r7r8r��s
�zCommandSpec._renderN)r�r�r�r�r�r�r�rErlr~r�rr�r�r�r�r�r�r�r7r7r7r8rz�s*





	
rzc@seZdZedd�ZdS)�WindowsCommandSpecFr�N)r�r�r�r�r�r7r7r7r8r��sr�c@s�eZdZdZe�d���ZeZ	e
ddd��Ze
ddd��Ze
dd	d
��Z
edd��Ze
d
d��Ze
dd��Ze
dd��Ze
ddd��ZdS)rkz`
    Encapsulates behavior around writing entry point scripts for console and
    gui apps.
    aJ
        # EASY-INSTALL-ENTRY-SCRIPT: %(spec)r,%(group)r,%(name)r
        import re
        import sys

        # for compatibility with easy_install; see #2198
        __requires__ = %(spec)r

        try:
            from importlib.metadata import distribution
        except ImportError:
            try:
                from importlib_metadata import distribution
            except ImportError:
                from pkg_resources import load_entry_point


        def importlib_load_entry_point(spec, group, name):
            dist_name, _, _ = spec.partition('==')
            matches = (
                entry_point
                for entry_point in distribution(dist_name).entry_points
                if entry_point.group == group and entry_point.name == name
            )
            return next(matches).load()


        globals().setdefault('load_entry_point', importlib_load_entry_point)


        if __name__ == '__main__':
            sys.argv[0] = re.sub(r'(-script\.pyw?|\.exe)?$', '', sys.argv[0])
            sys.exit(load_entry_point(%(spec)r, %(group)r, %(name)r)())
        NFcCs6t�dt�|rtnt��}|�d||�}|�||�S)NzUse get_argsr�)rlrmrn�WindowsScriptWriterrkrl�get_script_headerrm)rDr�r.�wininst�writer�headerr7r7r8�get_script_args(szScriptWriter.get_script_argscCs$tjdtdd�|rd}|�||�S)NzUse get_headerr�)�
stacklevelr))rlrmrnrq)rDrrr.r�r7r7r8r�0s�zScriptWriter.get_script_headerccs�|dur|��}t|���}dD]Z}|d}|�|���D]>\}}|�|�|jt�}|�||||�}	|	D]
}
|
Vqlq:q dS)z�
        Yield write_script() argument tuples for a distribution's
        console_scripts and gui_scripts entry points.
        N��console�guiZ_scripts)	rqr�r]�
get_entry_mapr��_ensure_safe_name�templater��_get_script_args)rDr�r�r�type_r�r��eprrrwr�r7r7r8rm9s
zScriptWriter.get_argscCst�d|�}|rtd��dS)z?
        Prevent paths in *_scripts entry point names.
        z[\\/]z+Path separators not allowed in script namesN)rH�searchr�)r�Zhas_path_sepr7r7r8r�KszScriptWriter._ensure_safe_namecCs t�dt�|rt��S|��S�NzUse best)rlrmrnr�rl)rDZ
force_windowsr7r7r8�
get_writerTszScriptWriter.get_writercCs.tjdkstjdkr&tjdkr&t��S|SdS)zD
        Select the best ScriptWriter for this environment.
        �win32�javar�N)r�r�r:r��_namer�rlr{r7r7r8rlZszScriptWriter.bestccs|||fVdSr�r7)rDr�r�r�rrr7r7r8r�dszScriptWriter._get_script_argsr�cCs"|j���|�}|�|�|��S)z;Create a #! line, getting options (if any) from script_text)�command_spec_classrlr�r�r�)rDrrr.�cmdr7r7r8rqis
zScriptWriter.get_header)NF)NF)N)r�N)r�r�r�r�rKrLr�r�rzr�rEr�r�rmr�r�r�rlr�rqr7r7r7r8rk�s&#


	
rkc@sLeZdZeZedd��Zedd��Zedd��Zedd��Z	e
d	d
��ZdS)r�cCst�dt�|��Sr�)rlrmrnrlr{r7r7r8r�tszWindowsScriptWriter.get_writercCs"tt|d�}tj�dd�}||S)zC
        Select the best ScriptWriter suitable for Windows
        )r.ZnaturalZSETUPTOOLS_LAUNCHERr.)r��WindowsExecutableLauncherWriterr:rr)rDZ
writer_lookup�launcherr7r7r8rlzs�zWindowsScriptWriter.bestc	#s�tddd�|}|tjd���d�vrFdjfit���}t�|t	�gd�}|�
|�|�||�}�fdd	�|D�}�|||d
|fVdS)z For Windows, add a .py extension�.pyarvr��PATHEXT�;zK{ext} not listed in PATHEXT; scripts will not be recognized as executables.)r�r��
-script.py�.pyc�.pyorvr~csg|]}�|�qSr7r7rx�r�r7r8r��r�z8WindowsScriptWriter._get_script_args.<locals>.<listcomp>rwN)r�r:rr/r�r�r�rlrm�UserWarningrY�_adjust_header)	rDr�r�r�rr�extr�r�r�r7r�r8r��s��
z$WindowsScriptWriter._get_script_argscCsNd}d}|dkr||}}t�t�|�tj�}|j||d�}|�|�rJ|S|S)z�
        Make sure 'pythonw' is used for gui and 'python' is used for
        console (regardless of what sys.executable is).
        r(r)r�)r��repl)rHrI�escape�
IGNORECASE�sub�_use_header)rDr�Zorig_headerrGr�Z
pattern_ob�
new_headerr7r7r8r��s
z"WindowsScriptWriter._adjust_headercCs$|dd��d�}tjdkp"t|�S)z�
        Should _adjust_header use the replaced header?

        On non-windows systems, always use. On
        Windows systems, only use the replaced header if it resolves
        to an executable on the system.
        r�r��"r�)rMr�r�r)r�Zclean_headerr7r7r8r��s	zWindowsScriptWriter._use_headerN)r�r�r�r�r�rEr�rlr�r�r�r�r7r7r7r8r�qs




r�c@seZdZedd��ZdS)r�c#s�|dkrd}d}dg}nd}d}gd�}|�||�}�fdd�|D�}	�|||d	|	fV�d
t|�dfVt�s��d}
|
t��d	fVd
S)zG
        For Windows, add a .py extension and an .exe launcher
        r�z-script.pywrv�clir�)r�r�r�csg|]}�|�qSr7r7rxr�r7r8r��r�zDWindowsExecutableLauncherWriter._get_script_args.<locals>.<listcomp>rwr~r^z
.exe.manifestN)r��get_win_launcherr9�load_launcher_manifest)rDr�r�r�rrZ
launcher_typer�r��hdrr�Zm_namer7r�r8r��s �z0WindowsExecutableLauncherWriter._get_script_argsN)r�r�r�rEr�r7r7r7r8r��sr�cCsJd|}t�r4t�dkr&|�dd�}q@|�dd�}n|�dd�}td|�S)z�
    Load the Windows launcher (executable) suitable for launching a script.

    `type` should be either 'cli' or 'gui'

    Returns the executable as a byte string.
    z%s.exez	win-arm64�.z-arm64.z-64.z-32.r�)r9rrNr)�typeZlauncher_fnr7r7r8r��s
r�cCst�td�}|�d�t�S)Nzlauncher manifest.xmlru)�
pkg_resourcesrr�rv�vars)r��manifestr7r7r8r��sr�FcCst�|||�Sr�)rdr�)r;�
ignore_errors�onerrorr7r7r8r��sr�cCst�d�}t�|�|S)N�)r:�umask)�tmpr7r7r8ry�s

ryc@seZdZdZdS)rnzF
    Warning for EasyInstall deprecations, bypassing suppression.
    N)r�r�r�r�r7r7r7r8rn�srn)N)rh)|r�rr�rrr�distutils.errorsrrrr	�distutils.command.installr
rrrr
Zdistutils.command.build_scriptsrr0rr�r:r�rdrBrrHr�rrKrlr�r5r�rsr�r r�	sysconfigrrr�rrZsetuptools.sandboxrZsetuptools.commandrZsetuptools.archive_utilrZsetuptools.package_indexrrrrrZsetuptools.wheelrr�rrrr r!r"r#r$r%r&r'r(r)r*r+�filterwarnings�
PEP440Warning�__all__r9r.rFrIrPr/r�r�rr1r2r0rArrrJrPr�r\r_rQrT�builtin_module_namesrSrlrqrurorzrx�ImportErrorr�rzr~Zsys_executabler�rkr�r�r�r�r�r�r�ryrnr7r7r7r8�<module>s�DqF.)%l	R
 �


TtA site-packages/setuptools/command/__pycache__/build_ext.cpython-39.pyc000064400000023223150564745010021754 0ustar00a

(�Re�3�
@spddlZddlZddlZddlmZddlmZddlm	Z	ddl
mZddlm
Z
mZddlmZddlmZdd	lmZzddlmZed
�Wney�eZYn0ed�ddlmZd
d�ZdZdZdZejdkr�dZn<ej dk�r zddl!Z!e"e!d�ZZWne�yYn0dd�Z#dd�Z$Gdd�de�Ze�sRej dk�r^ddd�Z%ndZd dd�Z%dS)!�N��EXTENSION_SUFFIXES)�	build_ext)�	copy_file)�new_compiler)�customize_compiler�get_config_var)�DistutilsError)�log)�LibraryzCython.Compiler.Main�LDSHARED)�_config_varsc	Csltjdkr`t��}z6dtd<dtd<dtd<t|�Wt��t�|�qht��t�|�0nt|�dS)N�darwinz0gcc -Wl,-x -dynamiclib -undefined dynamic_lookuprz -dynamiclib�CCSHAREDz.dylib�SO)�sys�platform�_CONFIG_VARS�copyr�clear�update)�compiler�tmp�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/build_ext.py�_customize_compiler_for_shlibs
�
�rFZsharedrT�nt�RTLD_NOWcCstr|SdS)N�)�	have_rtld)�srrr�if_dl>sr!cCs.tD]$}d|vr|S|dkr|SqdS)z;Return the file extension for an abi3-compliant Extension()z.abi3z.pydNr)�suffixrrr�get_abi3_suffixBs
r#c@sveZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zddd�ZdS)rcCs.|jd}|_t�|�||_|r*|��dS)z;Build extensions in build directory, then copy if --inplacerN)Zinplace�
_build_ext�run�copy_extensions_to_source)�selfZold_inplacerrrr%Ls

z
build_ext.runc
Cs�|�d�}|jD]�}|�|j�}|�|�}|�d�}d�|dd��}|�|�}tj	�|tj	�
|��}tj	�|j|�}	t|	||j
|jd�|jr|�|p�tj|d�qdS)N�build_py�.���)�verbose�dry_runT)�get_finalized_command�
extensions�get_ext_fullname�name�get_ext_filename�split�joinZget_package_dir�os�path�basename�	build_librr+r,�_needs_stub�
write_stub�curdir)
r'r(�ext�fullname�filename�modpath�package�package_dirZ
dest_filenameZsrc_filenamerrrr&Ts"





��z#build_ext.copy_extensions_to_sourcecCs�t�d�}|r&tjj|�d��|}nt�||�}td�}||jvr�|j|}t	|d�o\t
�}|r�|dt|��}t
�}||}t|t
�r�tj�|�\}}|j�|t�Str�|jr�tj�|�\}}tj�|d|�S|S)NZSETUPTOOLS_EXT_SUFFIXr)�
EXT_SUFFIXZpy_limited_apizdl-)r4�getenvr5r3r2r$r1r�ext_map�getattrr#�len�
isinstancer�splitext�shlib_compiler�library_filename�libtype�	use_stubs�_links_to_dynamic)r'r<Zso_extr=r;Zuse_abi3�fn�drrrr1js&




zbuild_ext.get_ext_filenamecCs t�|�d|_g|_i|_dS�N)r$�initialize_optionsrH�shlibsrC�r'rrrrP�s
zbuild_ext.initialize_optionscCs,t�|�|jpg|_|�|j�dd�|jD�|_|jrB|��|jD]}|�|j�|_qH|jD]�}|j}||j	|<||j	|�
d�d<|jr�|�|�p�d}|o�to�t
|t�}||_||_|�|�}|_tj�tj�|j|��}|�r||jv�r|j�|�|rbtrbtj|jvrb|j�tj�qbdS)NcSsg|]}t|t�r|�qSr)rFr��.0r;rrr�
<listcomp>�s
�z.build_ext.finalize_options.<locals>.<listcomp>r)r*F)r$�finalize_optionsr.Zcheck_extensions_listrQ�setup_shlib_compilerr/r0�
_full_namerCr2�links_to_dynamicrKrFrrLr8r1�
_file_namer4r5�dirnamer3r7�library_dirs�appendr:�runtime_library_dirs)r'r;r<Zltd�nsr=�libdirrrrrV�s,



zbuild_ext.finalize_optionscCs�t|j|j|jd�}|_t|�|jdur8|�|j�|jdur^|jD]\}}|�	||�qH|j
dur~|j
D]}|�|�qn|jdur�|�
|j�|jdur�|�|j�|jdur�|�|j�|jdur�|�|j�t�|�|_dS)N)rr,�force)rrr,rarHr�include_dirsZset_include_dirsZdefineZdefine_macroZundefZundefine_macro�	librariesZ
set_librariesr\Zset_library_dirsZrpathZset_runtime_library_dirsZlink_objectsZset_link_objects�link_shared_object�__get__)r'rr0�valueZmacrorrrrW�s*�







zbuild_ext.setup_shlib_compilercCst|t�r|jSt�||�SrO)rFr�export_symbolsr$�get_export_symbols�r'r;rrrrh�s
zbuild_ext.get_export_symbolscCsb|��|j}zFt|t�r"|j|_t�||�|jrL|�d�j	}|�
||�W||_n||_0dS)Nr()Z_convert_pyx_sources_to_langrrFrrHr$�build_extensionr8r-r7r9)r'r;Z	_compiler�cmdrrrrj�s
zbuild_ext.build_extensioncsPt�dd�|jD���d�|j�d�dd�dg��t��fdd�|jD��S)	z?Return true if 'ext' links to a dynamic lib in the same packagecSsg|]
}|j�qSr)rX)rT�librrrrU��z.build_ext.links_to_dynamic.<locals>.<listcomp>r)Nr*rc3s|]}�|�vVqdSrOr)rTZlibname�Zlibnames�pkgrr�	<genexpr>�rmz-build_ext.links_to_dynamic.<locals>.<genexpr>)�dict�fromkeysrQr3rXr2�anyrcrirrnrrY�s zbuild_ext.links_to_dynamiccCst�|�|��SrO)r$�get_outputs�_build_ext__get_stubs_outputsrRrrrrt�szbuild_ext.get_outputscs6�fdd��jD�}t�|����}tdd�|D��S)Nc3s2|]*}|jrtjj�jg|j�d��R�VqdS)r)N)r8r4r5r3r7rXr2rSrRrrrp�s�z0build_ext.__get_stubs_outputs.<locals>.<genexpr>css|]\}}||VqdSrOr)rT�baseZfnextrrrrp�rm)r.�	itertools�product�!_build_ext__get_output_extensions�list)r'Zns_ext_bases�pairsrrRrZ__get_stubs_outputs�s

�zbuild_ext.__get_stubs_outputsccs"dVdV|�d�jrdVdS)N�.pyz.pycr(z.pyo)r-�optimizerRrrrZ__get_output_extensions�sz!build_ext.__get_output_extensionsFcCs4t�d|j|�tjj|g|j�d��R�d}|rLtj�|�rLt|d��|j	s�t
|d�}|�d�ddd	td
�dtj�
|j�dd
dtd�dddtd�dddddtd�dddg��|��|�r0ddlm}||gdd|j	d �|�d!�j}|dk�r||g|d|j	d �tj�|��r0|j	�s0t�|�dS)"Nz writing stub loader for %s to %sr)r|z already exists! Please delete.�w�
zdef __bootstrap__():z-   global __bootstrap__, __file__, __loader__z0   import sys, os, pkg_resources, importlib.utilz, dlz:   __file__ = pkg_resources.resource_filename(__name__,%r)z   del __bootstrap__z    if '__loader__' in globals():z       del __loader__z#   old_flags = sys.getdlopenflags()z   old_dir = os.getcwd()z   try:z(     os.chdir(os.path.dirname(__file__))z$     sys.setdlopenflags(dl.RTLD_NOW)z3     spec = importlib.util.spec_from_file_location(z#                __name__, __file__)z0     mod = importlib.util.module_from_spec(spec)z!     spec.loader.exec_module(mod)z   finally:z"     sys.setdlopenflags(old_flags)z     os.chdir(old_dir)z__bootstrap__()rr)�byte_compileT)r}rar,�install_lib)r
�inforXr4r5r3r2�existsr	r,�open�writer!r6rZ�close�distutils.utilr�r-r}�unlink)r'�
output_dirr;�compileZ	stub_file�fr�r}rrrr9�sh
��
�����
�zbuild_ext.write_stubN)F)�__name__�
__module__�__qualname__r%r&r1rPrVrWrhrjrYrtruryr9rrrrrKs
	rc

Cs(|�|j|||||||||	|
||�
dSrO)�linkZSHARED_LIBRARY)
r'�objects�output_libnamer�rcr\r^rg�debug�
extra_preargs�extra_postargs�
build_temp�target_langrrrrd$s�rdZstaticc
Cs^|dusJ�tj�|�\}}
tj�|
�\}}|�d��d�rH|dd�}|�|||||�dS)N�xrl�)r4r5r2rGrI�
startswithZcreate_static_lib)r'r�r�r�rcr\r^rgr�r�r�r�r�r=r6r;rrrrd3s
�)
NNNNNrNNNN)
NNNNNrNNNN)&r4rrw�importlib.machineryrZdistutils.command.build_extrZ
_du_build_ext�distutils.file_utilr�distutils.ccompilerr�distutils.sysconfigrr�distutils.errorsr	�	distutilsr
Zsetuptools.extensionrZCython.Distutils.build_extr$�
__import__�ImportErrorr
rrrrKrJrr0�dl�hasattrr!r#rdrrrr�<module>sZ

	W�
�site-packages/setuptools/command/__pycache__/bdist_rpm.cpython-39.pyc000064400000003140150564745010021754 0ustar00a

(�Re��@s<ddlmmZddlZddlmZGdd�dej�ZdS)�N)�SetuptoolsDeprecationWarningc@s eZdZdZdd�Zdd�ZdS)�	bdist_rpma
    Override the default bdist_rpm behavior to do the following:

    1. Run egg_info to ensure the name and version are properly calculated.
    2. Always run 'install' using --single-version-externally-managed to
       disable eggs in RPM distributions.
    cCs&t�dt�|�d�tj�|�dS)Nzjbdist_rpm is deprecated and will be removed in a future version. Use bdist_wheel (wheel packages) instead.�egg_info)�warnings�warnr�run_command�origr�run)�self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/bdist_rpm.pyr	s�
z
bdist_rpm.runcCstj�|�}dd�|D�}|S)NcSs g|]}|�dd��dd��qS)zsetup.py install z5setup.py install --single-version-externally-managed z%setupz&%setup -n %{name}-%{unmangled_version})�replace)�.0�linerrr�
<listcomp>s���z-bdist_rpm._make_spec_file.<locals>.<listcomp>)rr�_make_spec_file)r
�specrrrrs
�
zbdist_rpm._make_spec_fileN)�__name__�
__module__�__qualname__�__doc__r	rrrrrrsr)Zdistutils.command.bdist_rpm�commandrrr�
setuptoolsrrrrr�<module>ssite-packages/setuptools/command/__pycache__/sdist.cpython-39.pyc000064400000015017150564745010021125 0ustar00a

(�Re�@sxddlmZddlmmZddlZddlZddlZddl	Z	ddl
mZddlZe
Zd
dd�ZGdd	�d	eej�ZdS)�)�logN�)�sdist_add_defaults�ccs,t�d�D]}|��|�D]
}|Vqq
dS)z%Find all files under revision controlzsetuptools.file_findersN)�
pkg_resources�iter_entry_points�load)�dirname�ep�item�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/sdist.py�walk_revctrlsrcs�eZdZdZgd�ZiZgd�Zedd�eD��Zdd�Z	dd	�Z
d
d�Zdd
�Ze
ejdd���Z�fdd�Zdd�Zdd�Zdd�Z�fdd�Zdd�Zdd�Zdd�Zd d!�Z�ZS)"�sdistz=Smart sdist that finds anything supported by revision control))zformats=Nz6formats for source distribution (comma-separated list))z	keep-temp�kz@keep the distribution tree around after creating archive file(s))z	dist-dir=�dzFdirectory to put the source distribution archive(s) in [default: dist])zowner=�uz@Owner name used when creating a tar file [default: current user])zgroup=�gzAGroup name used when creating a tar file [default: current group])rz.rstz.txtz.mdccs|]}d�|�VqdS)z	README{0}N)�format)�.0�extrrr
�	<genexpr>+�zsdist.<genexpr>cCs�|�d�|�d�}|j|_|j�tj�|jd��|��|�	�D]}|�|�qD|�
�t|jdg�}|j
D] }dd|f}||vrp|�|�qpdS)N�egg_infozSOURCES.txt�
dist_filesrr)�run_command�get_finalized_command�filelist�append�os�path�joinr�check_readme�get_sub_commands�make_distribution�getattr�distributionZ
archive_files)�selfZei_cmd�cmd_namer�file�datarrr
�run-s



z	sdist.runcCstj�|�|��dS�N)�origr�initialize_options�_default_to_gztar�r'rrr
r.@szsdist.initialize_optionscCstjdkrdSdg|_dS)N)��r�betar�gztar)�sys�version_info�formatsr0rrr
r/Es
zsdist._default_to_gztarcCs8|���tj�|�Wd�n1s*0YdS)z%
        Workaround for #516
        N)�_remove_os_linkr-rr$r0rrr
r$Ks
zsdist.make_distributionc
cspGdd�d�}ttd|�}zt`Wnty4Yn0zdVW||urlttd|�n||urjttd|�0dS)zG
        In a context, remove and restore os.link if it exists
        c@seZdZdS)z&sdist._remove_os_link.<locals>.NoValueN)�__name__�
__module__�__qualname__rrrr
�NoValueYsr<�linkN)r%rr=�	Exception�setattr)r<Zorig_valrrr
r8Rs�zsdist._remove_os_linkcs&t���tj�d�r"|j�d�dS)Nzpyproject.toml)�super�_add_defaults_optionalrr �isfilerrr0��	__class__rr
rAgs
zsdist._add_defaults_optionalcCs8|j��r4|�d�}|j�|���|�|�|��dS)zgetting python files�build_pyN)r&�has_pure_modulesrr�extendZget_source_files�_add_data_files�_safe_data_files�r'rErrr
�_add_defaults_pythonls

zsdist._add_defaults_pythoncCs|jjrdS|jS)z�
        Extracting data_files from build_py is known to cause
        infinite recursion errors when `include_package_data`
        is enabled, so suppress it in that case.
        r)r&Zinclude_package_data�
data_filesrJrrr
rIsszsdist._safe_data_filescCs|j�dd�|D��dS)zA
        Add data files as found in build_py.data_files.
        css.|]&\}}}}|D]}tj�||�VqqdSr,)rr r!)r�_�src_dir�	filenames�namerrr
r�s
�z(sdist._add_data_files.<locals>.<genexpr>N)rrG)r'rLrrr
rH}s�zsdist._add_data_filescs0zt���Wnty*t�d�Yn0dS)Nz&data_files contains unexpected objects)r@�_add_defaults_data_files�	TypeErrorr�warnr0rCrr
rQ�szsdist._add_defaults_data_filescCs8|jD]}tj�|�rdSq|�dd�|j��dS)Nz,standard file not found: should have one of z, )�READMESrr �existsrSr!)r'�frrr
r"�s

��zsdist.check_readmecCs^tj�|||�tj�|d�}ttd�rJtj�|�rJt�|�|�	d|�|�
d��|�dS)Nz	setup.cfgr=r)r-r�make_release_treerr r!�hasattrrU�unlink�	copy_filerZsave_version_info)r'�base_dir�files�destrrr
rW�s
zsdist.make_release_treecCsTtj�|j�sdSt�|jd��}|��}Wd�n1s>0Y|d��kS)NF�rbz+# file GENERATED by distutils, do NOT edit
)rr rB�manifest�io�open�readline�encode)r'�fp�
first_linerrr
�_manifest_is_not_generated�s&�z sdist._manifest_is_not_generatedc	Cs�t�d|j�t|jd�}|D]\}z|�d�}Wn$tyTt�d|�YqYn0|��}|�d�s|snq|j	�
|�q|��dS)z�Read the manifest file (named by 'self.manifest') and use it to
        fill in 'self.filelist', the list of files to include in the source
        distribution.
        zreading manifest file '%s'r^zUTF-8z"%r not UTF-8 decodable -- skipping�#N)r�infor_ra�decode�UnicodeDecodeErrorrS�strip�
startswithrr�close)r'r_�linerrr
�
read_manifest�s
zsdist.read_manifest)r9r:r;�__doc__�user_options�negative_optZREADME_EXTENSIONS�tuplerTr+r.r/r$�staticmethod�
contextlib�contextmanagerr8rArKrIrHrQr"rWrfro�
__classcell__rrrCr
rs*




r)r)�	distutilsrZdistutils.command.sdist�commandrr-rr5r`ruZ
py36compatrr�listZ_default_revctrlrrrrr
�<module>s
site-packages/setuptools/command/__pycache__/dist_info.cpython-39.pyc000064400000002643150564745010021756 0ustar00a

(�Re��@s8dZddlZddlmZddlmZGdd�de�ZdS)zD
Create a dist_info directory
As defined in the wheel specification
�N)�Command)�logc@s.eZdZdZdgZdd�Zdd�Zdd�Zd	S)
�	dist_infozcreate a .dist-info directory)z	egg-base=�ezLdirectory containing .egg-info directories (default: top of the source tree)cCs
d|_dS�N)�egg_base��self�r
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/dist_info.py�initialize_optionsszdist_info.initialize_optionscCsdSrr
rr
r
r�finalize_optionsszdist_info.finalize_optionscCsn|�d�}|j|_|��|��|jdtd��d}t�d�t	j
�|���|�d�}|�|j|�dS)N�egg_infoz	.egg-infoz
.dist-infoz
creating '{}'�bdist_wheel)
�get_finalized_commandrr
�runr�lenr�info�format�os�path�abspathZegg2dist)r	r�
dist_info_dirrr
r
rrs

z
dist_info.runN)�__name__�
__module__�__qualname__�description�user_optionsrr
rr
r
r
rrs�r)�__doc__r�distutils.corer�	distutilsrrr
r
r
r�<module>ssite-packages/setuptools/command/__pycache__/__init__.cpython-39.pyc000064400000000651150564745010021534 0ustar00a

(�Re��@s<ddlmZddlZdejvr4dejd<ej�d�[[dS)�)�bdistN�egg)Z	bdist_eggzPython .egg file)Zdistutils.command.bdistr�sysZformat_commandsZformat_command�append�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/__init__.py�<module>s


site-packages/setuptools/command/__pycache__/bdist_egg.cpython-39.pyc000064400000031435150564745010021730 0ustar00a

(�Re�@�@sdZddlmZmZddlmZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
mZmZmZddlmZddlmZdd	lmZmZd
d�Zdd
�Zdd�Zdd�ZGdd�de�Ze�d���Z dd�Z!dd�Z"dd�Z#ddd�Z$dd�Z%d d!�Z&d"d#�Z'gd$�Z(d)d'd(�Z)dS)*z6setuptools.command.bdist_egg

Build .egg distributions�)�remove_tree�mkpath)�log)�CodeTypeN)�get_build_platform�Distribution�ensure_directory)�Library)�Command)�get_path�get_python_versioncCstd�S)N�purelib)r�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/bdist_egg.py�_get_purelibsrcCs2d|vrtj�|�d}|�d�r.|dd�}|S)N�.r�modulei����)�os�path�splitext�endswith)�filenamerrr�strip_modules

rccs6t�|�D]&\}}}|��|��|||fVq
dS)zbDo os.walk in a reproducible way,
    independent of indeterministic filesystem readdir order
    N)r�walk�sort)�dir�base�dirs�filesrrr�sorted_walk!srcCsJt�d���}t|d��}|�||�Wd�n1s<0YdS)Na�
        def __bootstrap__():
            global __bootstrap__, __loader__, __file__
            import sys, pkg_resources, importlib.util
            __file__ = pkg_resources.resource_filename(__name__, %r)
            __loader__ = None; del __bootstrap__, __loader__
            spec = importlib.util.spec_from_file_location(__name__,__file__)
            mod = importlib.util.module_from_spec(spec)
            spec.loader.exec_module(mod)
        __bootstrap__()
        �w)�textwrap�dedent�lstrip�open�write)�resource�pyfileZ_stub_template�frrr�
write_stub+sr)c@s�eZdZdZdddde�fdddd	gZgd
�Zdd�Zd
d�Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd �Zd!S)"�	bdist_eggzcreate an "egg" distribution)z
bdist-dir=�bz1temporary directory for creating the distributionz
plat-name=�pz;platform name to embed in generated filenames (default: %s))�exclude-source-filesNz+remove all .py files from the generated egg)�	keep-temp�kzPkeep the pseudo-installation tree around after creating the distribution archive)z	dist-dir=�dz-directory to put final built distributions in)�
skip-buildNz2skip rebuilding everything (for testing/debugging))r.r1r-cCs.d|_d|_d|_d|_d|_d|_d|_dS)Nr)�	bdist_dir�	plat_name�	keep_temp�dist_dir�
skip_build�
egg_output�exclude_source_files��selfrrr�initialize_optionsRszbdist_egg.initialize_optionscCs�|�d�}|_|j|_|jdur>|�d�j}tj�|d�|_|jdurPt	�|_|�
dd�|jdur�tdd|j
|jt�|j��o�|j��
�}tj�|j|d�|_dS)N�egg_info�bdist�egg)r5r5z.egg)�get_finalized_command�ei_cmdr<r2�
bdist_baserr�joinr3r�set_undefined_optionsr7r�egg_nameZegg_versionr�distribution�has_ext_modulesr5)r:r@rA�basenamerrr�finalize_options[s


�zbdist_egg.finalize_optionscCs�|j|�d�_tj�tj�t���}|jj	g}|j_	|D]�}t
|t�r�t|�dkr�tj�
|d�r�tj�|d�}tj�|�}||ks�|�|tj�r�|t|�dd�|df}|jj	�|�q:z*t�d|j�|jdddd�W||j_	n
||j_	0dS)N�install�r�zinstalling package data to %s�install_data)�force�root)r2r?�install_librr�normcase�realpathrrE�
data_files�
isinstance�tuple�len�isabs�
startswith�sep�appendr�info�call_command)r:�
site_packages�old�itemrQ�
normalizedrrr�do_install_datass"�zbdist_egg.do_install_datacCs|jgS)N)r7r9rrr�get_outputs�szbdist_egg.get_outputscKsTtD]}|�||j�q|�d|j�|�d|j�|j|fi|��}|�|�|S)z8Invoke reinitialized command `cmdname` with keyword argsr6�dry_run)�INSTALL_DIRECTORY_ATTRS�
setdefaultr2r6rb�reinitialize_command�run_command)r:Zcmdname�kw�dirname�cmdrrrr[�s
zbdist_egg.call_commandcCs�|�d�t�d|j�|�d�}|j}d|_|j��rH|jsH|�d�|j	ddd�}||_|�
�\}}g|_g}t|�D]|\}}t
j�|�\}	}
t
j�|jt|	�d�}|j�|�t�d	|�|js�tt
j�|�|�|�|�|�t
jd
�||<qz|�r|�|�|jj�r|��|j}t
j�|d�}
|�|
�|jj�rlt
j�|
d�}t�d
|�|j	d|dd�|�|
�t
j�|
d�}|�r�t�d|�|j�st|�t|d�}|� d�|��|� d�|�!�n,t
j�"|��rt�d|�|j�st
�#|�t$t
j�|d�|�%��t
j�&t
j�|j'd���r<t�(d�|j)�rL|�*�t+|j,||j-|j|�.�d�|j/�s�t0|j|jd�t1|jdg��dt2�|j,f�dS)Nr<zinstalling library code to %srI�
build_clibrOr)�warn_dir�.pyzcreating stub loader for %s�/�EGG-INFO�scriptszinstalling scripts to %s�install_scriptsrK)�install_dirZno_epznative_libs.txtz
writing %s�wt�
zremoving %szdepends.txtzxWARNING: 'depends.txt' will not be used by setuptools 0.6!
Use the install_requires/extras_require setup() args instead.)�verboserb�mode�rb�
dist_filesr*)3rfrrZr2r?rNrE�has_c_librariesr6r[�get_ext_outputs�stubs�	enumeraterrrrBrrYrbr)rG�replacerX�byte_compilerRr`rro�copy_metadata_torr$r%�close�isfile�unlink�write_safety_flag�zip_safe�existsr<�warnr8�zap_pyfiles�make_zipfiler7rt�
gen_headerr4r�getattrr)r:ZinstcmdZold_rootri�all_outputs�ext_outputsZ
to_compiler,Zext_namer�extr'Zarchive_rootr<Z
script_dirZnative_libsZ	libs_filerrr�run�s�


�




�




��
��z
bdist_egg.runc

Cs�t�d�t|j�D]�\}}}|D]�}tj�||�}|�d�rTt�d|�t�	|�|�d�r"|}d}t
�||�}tj�|tj|�
d�d�}	t�d||	f�zt�|	�Wnty�Yn0t�||	�q"qdS)	Nz+Removing .py files from temporary directoryrlzDeleting %s�__pycache__z#(?P<name>.+)\.(?P<magic>[^.]+)\.pyc�name�.pyczRenaming file from [%s] to [%s])rrZ�walk_eggr2rrrBr�debugr��re�match�pardir�group�remove�OSError�rename)
r:rrrr�rZpath_old�pattern�mZpath_newrrrr��s0



���zbdist_egg.zap_pyfilescCs2t|jdd�}|dur|St�d�t|j|j�S)Nr�z4zip_safe flag not set; analyzing archive contents...)r�rErr��analyze_eggr2rz)r:�saferrrr�s

zbdist_egg.zip_safecCsdS)Nr rr9rrrr�szbdist_egg.gen_headercCshtj�|j�}tj�|d�}|jjjD]<}|�|�r&tj�||t	|�d��}t
|�|�||�q&dS)z*Copy metadata (egg info) to the target_dir�N)rr�normpathr<rBr@�filelistrrWrUr�	copy_file)r:�
target_dirZ
norm_egg_info�prefixr�targetrrrr~s
zbdist_egg.copy_metadata_tocCsg}g}|jdi}t|j�D]f\}}}|D].}tj�|�d��tvr*|�|||�q*|D]"}|||d|tj�||�<q^q|j	�
�r�|�d�}|jD]Z}	t
|	t�r�q�|�|	j�}
|�|
�}tj�|��d�s�tj�tj�|j|��r�|�|�q�||fS)zAGet a list of relative paths to C extensions in the output distror�rKrm�	build_extzdl-)r2rrrr�lower�NATIVE_EXTENSIONSrYrBrErFr?�
extensionsrSr	Zget_ext_fullnamer�Zget_ext_filenamerGrWr�)r:r�r��pathsrrrrZ	build_cmdr��fullnamerrrrys0
��




zbdist_egg.get_ext_outputsN)�__name__�
__module__�__qualname__�descriptionr�user_options�boolean_optionsr;rHr`rar[r�r�r�r�r~ryrrrrr*;s,��	
Qr*z.dll .so .dylib .pydccsHt|�}t|�\}}}d|vr(|�d�|||fV|D]
}|Vq8dS)z@Walk an unpacked egg's contents, skipping the metadata directoryrnN)r�nextr�)�egg_dir�walkerrrrZbdfrrrr�:s
r�c	Cs�t��D](\}}tj�tj�|d|��r|Sqt�s<dSd}t|�D]R\}}}|D]B}|�d�sV|�d�rrqVqV|�d�s�|�d�rVt	||||�o�|}qVqH|S)NrnFTrlz.pywr�z.pyo)
�safety_flags�itemsrrr�rB�can_scanr�r�scan_module)	r�rz�flag�fnr�rrrr�rrrr�Es
r�cCs�t��D]r\}}tj�||�}tj�|�rJ|dus>t|�|krzt�|�q|durt|�|krt|d�}|�	d�|�
�qdS)Nrrrs)r�r�rrrBr��boolr�r$r%r)r�r�r�r�r(rrrr�Ws

r�zzip-safeznot-zip-safe)TFc
Cstj�||�}|dd�|vr"dS|t|�dd��tjd�}||rJdpLdtj�|�d}tjdkrpd	}nd
}t	|d�}|�
|�t�|�}	|�
�d}
t�t|	��}dD]}||vr�t�d
||�d}
q�d|vr�dD]}||vr�t�d||�d}
q�|
S)z;Check whether module possibly uses unsafe-for-zipfile stuffN���TrKrr�r)�����rb)�__file__�__path__z%s: module references %sF�inspect)�	getsource�
getabsfile�
getsourcefileZgetfilegetsourcelines�
findsource�getcomments�getframeinfo�getinnerframes�getouterframes�stack�tracez"%s: module MAY be using inspect.%s)rrrBrUr|rXr�sys�version_infor$�read�marshal�loadr�dict�fromkeys�iter_symbolsrr�)
r�rr�rzr�pkgr�skipr(�coder��symbols�badrrrr�js0 



r�ccsR|jD]
}|Vq|jD]4}t|t�r.|Vqt|t�rt|�D]
}|Vq@qdS)zBYield names and strings used by `code` and its nested code objectsN)�co_names�	co_constsrS�strrr�)r�r��constrrrr��s



r�cCs2tj�d�stjdkrdSt�d�t�d�dS)N�java�cliTz1Unable to analyze compiled code on this platform.zfPlease ask the author to include a 'zip_safe' setting (either True or False) in the package's setup.py)r��platformrWrr�rrrrr��s
r�)rOrqrL�install_baseTr c
s�ddl}ttj�|��d�t�d|����fdd�}|rB|jn|j}�s�|j	|||d�}	t
��D]\}
}}||	|
|�qd|	��n t
��D]\}
}}|d|
|�q�|S)aqCreate a zip file from all the files under 'base_dir'.  The output
    zip file will be named 'base_dir' + ".zip".  Uses either the "zipfile"
    Python module (if available) or the InfoZIP "zip" utility (if installed
    and found on the default search path).  If neither tool is available,
    raises DistutilsExecError.  Returns the name of the output zip file.
    rNrvz#creating '%s' and adding '%s' to itcs`|D]V}tj�tj�||��}tj�|�r|t��dd�}�sN|�||�t�d|�qdS)NrKzadding '%s')	rrr�rBr�rUr%rr�)�zrh�namesr�rr,��base_dirrbrr�visit�szmake_zipfile.<locals>.visit)�compression)�zipfilerrrrhrrZ�ZIP_DEFLATED�
ZIP_STORED�ZipFilerr)
�zip_filenamer�rtrb�compressrur�r�r�r�rhrrrr�rr��s	
r�)rrTr )*�__doc__�distutils.dir_utilrr�	distutilsr�typesrr�rr�r!r��
pkg_resourcesrrrZsetuptools.extensionr	�
setuptoolsr
�	sysconfigrrrrrr)r*r�r��splitr�r�r�r�r�r�r�r�rcr�rrrr�<module>s@
}�"�site-packages/setuptools/command/__pycache__/install_lib.cpython-39.pyc000064400000010126150564745010022267 0ustar00a

(�Re#�@sHddlZddlZddlmZmZddlmmZGdd�dej�ZdS)�N)�product�starmapc@sZeZdZdZdd�Zdd�Zdd�Zedd	��Zd
d�Z	edd
��Z
ddd�Zdd�ZdS)�install_libz9Don't add compiled flags to filenames of non-Python filescCs&|��|��}|dur"|�|�dS�N)�build�install�byte_compile)�self�outfiles�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/install_lib.py�run
szinstall_lib.runcs4�fdd����D�}t|����}tt�j|��S)z�
        Return a collections.Sized collections.Container of paths to be
        excluded for single_version_externally_managed installations.
        c3s"|]}��|�D]
}|VqqdSr)�
_all_packages)�.0Zns_pkg�pkg�r	rr�	<genexpr>s�z-install_lib.get_exclusions.<locals>.<genexpr>)�_get_SVEM_NSPsr�_gen_exclusion_paths�setr�_exclude_pkg_path)r	Zall_packagesZ
excl_specsrrr�get_exclusionss

�zinstall_lib.get_exclusionscCs&|�d�|g}tjj|jg|�R�S)zw
        Given a package name and exclusion path within that package,
        compute the full exclusion path.
        �.)�split�os�path�joinZinstall_dir)r	rZexclusion_path�partsrrrrszinstall_lib._exclude_pkg_pathccs |r|V|�d�\}}}qdS)zn
        >>> list(install_lib._all_packages('foo.bar.baz'))
        ['foo.bar.baz', 'foo.bar', 'foo']
        rN)�
rpartition)�pkg_name�sep�childrrrr'szinstall_lib._all_packagescCs,|jjsgS|�d�}|j}|r(|jjSgS)z�
        Get namespace packages (list) but only for
        single_version_externally_managed installations and empty otherwise.
        r)�distributionZnamespace_packages�get_finalized_commandZ!single_version_externally_managed)r	Zinstall_cmdZsvemrrrr1s

zinstall_lib._get_SVEM_NSPsccsbdVdVdVttd�s dStj�ddtjj�}|dV|d	V|d
V|dVdS)zk
        Generate file paths to be excluded for namespace packages (bytecode
        cache files).
        z__init__.pyz__init__.pycz__init__.pyo�implementationN�__pycache__z	__init__.z.pycz.pyoz
.opt-1.pycz
.opt-2.pyc)�hasattr�sysrrrr$�	cache_tag)�baserrrrAs
�


z install_lib._gen_exclusion_paths�rc	sh|r|r|rJ�|����s,tj�|||�Sddlm}ddlm�g����fdd�}||||��S)Nr)�unpack_directory)�logcs<|�vr��d|�dS��d|tj�|����|�|S)Nz/Skipping installation of %s (namespace package)Fzcopying %s -> %s)�warn�inforr�dirname�append)�src�dst��excluder,r
rr�pfhs�
z!install_lib.copy_tree.<locals>.pf)r�origr�	copy_treeZsetuptools.archive_utilr+�	distutilsr,)	r	�infile�outfile�
preserve_mode�preserve_times�preserve_symlinks�levelr+r5rr3rr7Ws
zinstall_lib.copy_treecs.tj�|�}|����r*�fdd�|D�S|S)Ncsg|]}|�vr|�qSrr)r�f�r4rr�
<listcomp>y�z+install_lib.get_outputs.<locals>.<listcomp>)r6r�get_outputsr)r	�outputsrr@rrCus
zinstall_lib.get_outputsN)r*r*rr*)
�__name__�
__module__�__qualname__�__doc__r
rr�staticmethodrrrr7rCrrrrrs
	
�
r)	rr'�	itertoolsrrZdistutils.command.install_lib�commandrr6rrrr�<module>ssite-packages/setuptools/command/__pycache__/py36compat.cpython-39.pyc000064400000011035150564745010022000 0ustar00a

(�ReR�@sXddlZddlmZddlmZddlmZGdd�d�Zeejd�rTGdd�d�ZdS)	�N)�glob)�convert_path)�sdistc@s\eZdZdZdd�Zedd��Zdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dS)�sdist_add_defaultsz�
    Mix-in providing forward-compatibility for functionality as found in
    distutils on Python 3.7.

    Do not edit the code in this class except to update functionality
    as implemented in distutils. Instead, override in the subclass.
    cCs<|��|��|��|��|��|��|��dS)a9Add all the default files to self.filelist:
          - README or README.txt
          - setup.py
          - test/test*.py
          - all pure Python modules mentioned in setup script
          - all files pointed by package_data (build_py)
          - all files defined in data_files.
          - all files defined as scripts.
          - all C sources listed as part of extensions or C libraries
            in the setup script (doesn't catch C headers!)
        Warns if (README or README.txt) or setup.py are missing; everything
        else is optional.
        N)�_add_defaults_standards�_add_defaults_optional�_add_defaults_python�_add_defaults_data_files�_add_defaults_ext�_add_defaults_c_libs�_add_defaults_scripts)�self�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/setuptools/command/py36compat.py�add_defaultsszsdist_add_defaults.add_defaultscCs:tj�|�sdStj�|�}tj�|�\}}|t�|�vS)z�
        Case-sensitive path existence check

        >>> sdist_add_defaults._cs_path_exists(__file__)
        True
        >>> sdist_add_defaults._cs_path_exists(__file__.upper())
        False
        F)�os�path�exists�abspath�split�listdir)�fspathr�	directory�filenamerrr�_cs_path_exists&s

z"sdist_add_defaults._cs_path_existscCs�|j|jjg}|D]~}t|t�rj|}d}|D]"}|�|�r,d}|j�|�qPq,|s�|�dd�	|��q|�|�r�|j�|�q|�d|�qdS)NFTz,standard file not found: should have one of z, zstandard file '%s' not found)
ZREADMES�distribution�script_name�
isinstance�tupler�filelist�append�warn�join)r
Z	standards�fnZaltsZgot_itrrrr7s"

�
z*sdist_add_defaults._add_defaults_standardscCs4ddg}|D]"}ttjjt|��}|j�|�qdS)Nz
test/test*.pyz	setup.cfg)�filterrr�isfilerr�extend)r
�optional�pattern�filesrrrrLsz)sdist_add_defaults._add_defaults_optionalcCs\|�d�}|j��r$|j�|���|jD],\}}}}|D]}|j�tj	�
||��q:q*dS)N�build_py)�get_finalized_commandr�has_pure_modulesrr&�get_source_files�
data_filesr rrr")r
r*�pkg�src_dir�	build_dir�	filenamesrrrrrRs

z'sdist_add_defaults._add_defaults_pythoncCsz|j��rv|jjD]b}t|t�rBt|�}tj�|�rt|j	�
|�q|\}}|D]$}t|�}tj�|�rN|j	�
|�qNqdS)N)r�has_data_filesr.r�strrrrr%rr )r
�item�dirnamer2�frrrr	bs

z+sdist_add_defaults._add_defaults_data_filescCs(|j��r$|�d�}|j�|���dS)N�	build_ext)r�has_ext_modulesr+rr&r-)r
r8rrrr
ss

z$sdist_add_defaults._add_defaults_extcCs(|j��r$|�d�}|j�|���dS)N�
build_clib)r�has_c_librariesr+rr&r-)r
r:rrrrxs

z'sdist_add_defaults._add_defaults_c_libscCs(|j��r$|�d�}|j�|���dS)N�
build_scripts)r�has_scriptsr+rr&r-)r
r<rrrr}s

z(sdist_add_defaults._add_defaults_scriptsN)�__name__�
__module__�__qualname__�__doc__r�staticmethodrrrrr	r
rrrrrrrs
rrc@seZdZdS)rN)r>r?r@rrrrr�s)rr�distutils.utilr�distutils.commandrr�hasattrrrrr�<module>s|site-packages/setuptools/command/py36compat.py000064400000011522150564745010015512 0ustar00import os
from glob import glob
from distutils.util import convert_path
from distutils.command import sdist


class sdist_add_defaults:
    """
    Mix-in providing forward-compatibility for functionality as found in
    distutils on Python 3.7.

    Do not edit the code in this class except to update functionality
    as implemented in distutils. Instead, override in the subclass.
    """

    def add_defaults(self):
        """Add all the default files to self.filelist:
          - README or README.txt
          - setup.py
          - test/test*.py
          - all pure Python modules mentioned in setup script
          - all files pointed by package_data (build_py)
          - all files defined in data_files.
          - all files defined as scripts.
          - all C sources listed as part of extensions or C libraries
            in the setup script (doesn't catch C headers!)
        Warns if (README or README.txt) or setup.py are missing; everything
        else is optional.
        """
        self._add_defaults_standards()
        self._add_defaults_optional()
        self._add_defaults_python()
        self._add_defaults_data_files()
        self._add_defaults_ext()
        self._add_defaults_c_libs()
        self._add_defaults_scripts()

    @staticmethod
    def _cs_path_exists(fspath):
        """
        Case-sensitive path existence check

        >>> sdist_add_defaults._cs_path_exists(__file__)
        True
        >>> sdist_add_defaults._cs_path_exists(__file__.upper())
        False
        """
        if not os.path.exists(fspath):
            return False
        # make absolute so we always have a directory
        abspath = os.path.abspath(fspath)
        directory, filename = os.path.split(abspath)
        return filename in os.listdir(directory)

    def _add_defaults_standards(self):
        standards = [self.READMES, self.distribution.script_name]
        for fn in standards:
            if isinstance(fn, tuple):
                alts = fn
                got_it = False
                for fn in alts:
                    if self._cs_path_exists(fn):
                        got_it = True
                        self.filelist.append(fn)
                        break

                if not got_it:
                    self.warn("standard file not found: should have one of " +
                              ', '.join(alts))
            else:
                if self._cs_path_exists(fn):
                    self.filelist.append(fn)
                else:
                    self.warn("standard file '%s' not found" % fn)

    def _add_defaults_optional(self):
        optional = ['test/test*.py', 'setup.cfg']
        for pattern in optional:
            files = filter(os.path.isfile, glob(pattern))
            self.filelist.extend(files)

    def _add_defaults_python(self):
        # build_py is used to get:
        #  - python modules
        #  - files defined in package_data
        build_py = self.get_finalized_command('build_py')

        # getting python files
        if self.distribution.has_pure_modules():
            self.filelist.extend(build_py.get_source_files())

        # getting package_data files
        # (computed in build_py.data_files by build_py.finalize_options)
        for pkg, src_dir, build_dir, filenames in build_py.data_files:
            for filename in filenames:
                self.filelist.append(os.path.join(src_dir, filename))

    def _add_defaults_data_files(self):
        # getting distribution.data_files
        if self.distribution.has_data_files():
            for item in self.distribution.data_files:
                if isinstance(item, str):
                    # plain file
                    item = convert_path(item)
                    if os.path.isfile(item):
                        self.filelist.append(item)
                else:
                    # a (dirname, filenames) tuple
                    dirname, filenames = item
                    for f in filenames:
                        f = convert_path(f)
                        if os.path.isfile(f):
                            self.filelist.append(f)

    def _add_defaults_ext(self):
        if self.distribution.has_ext_modules():
            build_ext = self.get_finalized_command('build_ext')
            self.filelist.extend(build_ext.get_source_files())

    def _add_defaults_c_libs(self):
        if self.distribution.has_c_libraries():
            build_clib = self.get_finalized_command('build_clib')
            self.filelist.extend(build_clib.get_source_files())

    def _add_defaults_scripts(self):
        if self.distribution.has_scripts():
            build_scripts = self.get_finalized_command('build_scripts')
            self.filelist.extend(build_scripts.get_source_files())


if hasattr(sdist.sdist, '_add_defaults_standards'):
    # disable the functionality already available upstream
    class sdist_add_defaults:  # noqa
        pass
site-packages/setuptools/command/install_lib.py000064400000007443150564745010016010 0ustar00import os
import sys
from itertools import product, starmap
import distutils.command.install_lib as orig


class install_lib(orig.install_lib):
    """Don't add compiled flags to filenames of non-Python files"""

    def run(self):
        self.build()
        outfiles = self.install()
        if outfiles is not None:
            # always compile, in case we have any extension stubs to deal with
            self.byte_compile(outfiles)

    def get_exclusions(self):
        """
        Return a collections.Sized collections.Container of paths to be
        excluded for single_version_externally_managed installations.
        """
        all_packages = (
            pkg
            for ns_pkg in self._get_SVEM_NSPs()
            for pkg in self._all_packages(ns_pkg)
        )

        excl_specs = product(all_packages, self._gen_exclusion_paths())
        return set(starmap(self._exclude_pkg_path, excl_specs))

    def _exclude_pkg_path(self, pkg, exclusion_path):
        """
        Given a package name and exclusion path within that package,
        compute the full exclusion path.
        """
        parts = pkg.split('.') + [exclusion_path]
        return os.path.join(self.install_dir, *parts)

    @staticmethod
    def _all_packages(pkg_name):
        """
        >>> list(install_lib._all_packages('foo.bar.baz'))
        ['foo.bar.baz', 'foo.bar', 'foo']
        """
        while pkg_name:
            yield pkg_name
            pkg_name, sep, child = pkg_name.rpartition('.')

    def _get_SVEM_NSPs(self):
        """
        Get namespace packages (list) but only for
        single_version_externally_managed installations and empty otherwise.
        """
        # TODO: is it necessary to short-circuit here? i.e. what's the cost
        # if get_finalized_command is called even when namespace_packages is
        # False?
        if not self.distribution.namespace_packages:
            return []

        install_cmd = self.get_finalized_command('install')
        svem = install_cmd.single_version_externally_managed

        return self.distribution.namespace_packages if svem else []

    @staticmethod
    def _gen_exclusion_paths():
        """
        Generate file paths to be excluded for namespace packages (bytecode
        cache files).
        """
        # always exclude the package module itself
        yield '__init__.py'

        yield '__init__.pyc'
        yield '__init__.pyo'

        if not hasattr(sys, 'implementation'):
            return

        base = os.path.join(
            '__pycache__', '__init__.' + sys.implementation.cache_tag)
        yield base + '.pyc'
        yield base + '.pyo'
        yield base + '.opt-1.pyc'
        yield base + '.opt-2.pyc'

    def copy_tree(
            self, infile, outfile,
            preserve_mode=1, preserve_times=1, preserve_symlinks=0, level=1
    ):
        assert preserve_mode and preserve_times and not preserve_symlinks
        exclude = self.get_exclusions()

        if not exclude:
            return orig.install_lib.copy_tree(self, infile, outfile)

        # Exclude namespace package __init__.py* files from the output

        from setuptools.archive_util import unpack_directory
        from distutils import log

        outfiles = []

        def pf(src, dst):
            if dst in exclude:
                log.warn("Skipping installation of %s (namespace package)",
                         dst)
                return False

            log.info("copying %s -> %s", src, os.path.dirname(dst))
            outfiles.append(dst)
            return dst

        unpack_directory(infile, outfile, pf)
        return outfiles

    def get_outputs(self):
        outputs = orig.install_lib.get_outputs(self)
        exclude = self.get_exclusions()
        if exclude:
            return [f for f in outputs if f not in exclude]
        return outputs
site-packages/setuptools/command/build_py.py000064400000020124150564745010015312 0ustar00from glob import glob
from distutils.util import convert_path
import distutils.command.build_py as orig
import os
import fnmatch
import textwrap
import io
import distutils.errors
import itertools
import stat
from setuptools.extern.more_itertools import unique_everseen


def make_writable(target):
    os.chmod(target, os.stat(target).st_mode | stat.S_IWRITE)


class build_py(orig.build_py):
    """Enhanced 'build_py' command that includes data files with packages

    The data files are specified via a 'package_data' argument to 'setup()'.
    See 'setuptools.dist.Distribution' for more details.

    Also, this version of the 'build_py' command allows you to specify both
    'py_modules' and 'packages' in the same setup operation.
    """

    def finalize_options(self):
        orig.build_py.finalize_options(self)
        self.package_data = self.distribution.package_data
        self.exclude_package_data = self.distribution.exclude_package_data or {}
        if 'data_files' in self.__dict__:
            del self.__dict__['data_files']
        self.__updated_files = []

    def run(self):
        """Build modules, packages, and copy data files to build directory"""
        if not self.py_modules and not self.packages:
            return

        if self.py_modules:
            self.build_modules()

        if self.packages:
            self.build_packages()
            self.build_package_data()

        # Only compile actual .py files, using our base class' idea of what our
        # output files are.
        self.byte_compile(orig.build_py.get_outputs(self, include_bytecode=0))

    def __getattr__(self, attr):
        "lazily compute data files"
        if attr == 'data_files':
            self.data_files = self._get_data_files()
            return self.data_files
        return orig.build_py.__getattr__(self, attr)

    def build_module(self, module, module_file, package):
        outfile, copied = orig.build_py.build_module(self, module, module_file, package)
        if copied:
            self.__updated_files.append(outfile)
        return outfile, copied

    def _get_data_files(self):
        """Generate list of '(package,src_dir,build_dir,filenames)' tuples"""
        self.analyze_manifest()
        return list(map(self._get_pkg_data_files, self.packages or ()))

    def _get_pkg_data_files(self, package):
        # Locate package source directory
        src_dir = self.get_package_dir(package)

        # Compute package build directory
        build_dir = os.path.join(*([self.build_lib] + package.split('.')))

        # Strip directory from globbed filenames
        filenames = [
            os.path.relpath(file, src_dir)
            for file in self.find_data_files(package, src_dir)
        ]
        return package, src_dir, build_dir, filenames

    def find_data_files(self, package, src_dir):
        """Return filenames for package's data files in 'src_dir'"""
        patterns = self._get_platform_patterns(
            self.package_data,
            package,
            src_dir,
        )
        globs_expanded = map(glob, patterns)
        # flatten the expanded globs into an iterable of matches
        globs_matches = itertools.chain.from_iterable(globs_expanded)
        glob_files = filter(os.path.isfile, globs_matches)
        files = itertools.chain(
            self.manifest_files.get(package, []),
            glob_files,
        )
        return self.exclude_data_files(package, src_dir, files)

    def build_package_data(self):
        """Copy data files into build directory"""
        for package, src_dir, build_dir, filenames in self.data_files:
            for filename in filenames:
                target = os.path.join(build_dir, filename)
                self.mkpath(os.path.dirname(target))
                srcfile = os.path.join(src_dir, filename)
                outf, copied = self.copy_file(srcfile, target)
                make_writable(target)
                srcfile = os.path.abspath(srcfile)

    def analyze_manifest(self):
        self.manifest_files = mf = {}
        if not self.distribution.include_package_data:
            return
        src_dirs = {}
        for package in self.packages or ():
            # Locate package source directory
            src_dirs[assert_relative(self.get_package_dir(package))] = package

        self.run_command('egg_info')
        ei_cmd = self.get_finalized_command('egg_info')
        for path in ei_cmd.filelist.files:
            d, f = os.path.split(assert_relative(path))
            prev = None
            oldf = f
            while d and d != prev and d not in src_dirs:
                prev = d
                d, df = os.path.split(d)
                f = os.path.join(df, f)
            if d in src_dirs:
                if path.endswith('.py') and f == oldf:
                    continue  # it's a module, not data
                mf.setdefault(src_dirs[d], []).append(path)

    def get_data_files(self):
        pass  # Lazily compute data files in _get_data_files() function.

    def check_package(self, package, package_dir):
        """Check namespace packages' __init__ for declare_namespace"""
        try:
            return self.packages_checked[package]
        except KeyError:
            pass

        init_py = orig.build_py.check_package(self, package, package_dir)
        self.packages_checked[package] = init_py

        if not init_py or not self.distribution.namespace_packages:
            return init_py

        for pkg in self.distribution.namespace_packages:
            if pkg == package or pkg.startswith(package + '.'):
                break
        else:
            return init_py

        with io.open(init_py, 'rb') as f:
            contents = f.read()
        if b'declare_namespace' not in contents:
            raise distutils.errors.DistutilsError(
                "Namespace package problem: %s is a namespace package, but "
                "its\n__init__.py does not call declare_namespace()! Please "
                'fix it.\n(See the setuptools manual under '
                '"Namespace Packages" for details.)\n"' % (package,)
            )
        return init_py

    def initialize_options(self):
        self.packages_checked = {}
        orig.build_py.initialize_options(self)

    def get_package_dir(self, package):
        res = orig.build_py.get_package_dir(self, package)
        if self.distribution.src_root is not None:
            return os.path.join(self.distribution.src_root, res)
        return res

    def exclude_data_files(self, package, src_dir, files):
        """Filter filenames for package's data files in 'src_dir'"""
        files = list(files)
        patterns = self._get_platform_patterns(
            self.exclude_package_data,
            package,
            src_dir,
        )
        match_groups = (fnmatch.filter(files, pattern) for pattern in patterns)
        # flatten the groups of matches into an iterable of matches
        matches = itertools.chain.from_iterable(match_groups)
        bad = set(matches)
        keepers = (fn for fn in files if fn not in bad)
        # ditch dupes
        return list(unique_everseen(keepers))

    @staticmethod
    def _get_platform_patterns(spec, package, src_dir):
        """
        yield platform-specific path patterns (suitable for glob
        or fn_match) from a glob-based spec (such as
        self.package_data or self.exclude_package_data)
        matching package in src_dir.
        """
        raw_patterns = itertools.chain(
            spec.get('', []),
            spec.get(package, []),
        )
        return (
            # Each pattern has to be converted to a platform-specific path
            os.path.join(src_dir, convert_path(pattern))
            for pattern in raw_patterns
        )


def assert_relative(path):
    if not os.path.isabs(path):
        return path
    from distutils.errors import DistutilsSetupError

    msg = (
        textwrap.dedent(
            """
        Error: setup script specifies an absolute path:

            %s

        setup() arguments must *always* be /-separated paths relative to the
        setup.py directory, *never* absolute paths.
        """
        ).lstrip()
        % path
    )
    raise DistutilsSetupError(msg)
site-packages/setuptools/command/install_scripts.py000064400000005041150564745010016721 0ustar00from distutils import log
import distutils.command.install_scripts as orig
from distutils.errors import DistutilsModuleError
import os
import sys

from pkg_resources import Distribution, PathMetadata, ensure_directory


class install_scripts(orig.install_scripts):
    """Do normal script install, plus any egg_info wrapper scripts"""

    def initialize_options(self):
        orig.install_scripts.initialize_options(self)
        self.no_ep = False

    def run(self):
        import setuptools.command.easy_install as ei

        self.run_command("egg_info")
        if self.distribution.scripts:
            orig.install_scripts.run(self)  # run first to set up self.outfiles
        else:
            self.outfiles = []
        if self.no_ep:
            # don't install entry point scripts into .egg file!
            return

        ei_cmd = self.get_finalized_command("egg_info")
        dist = Distribution(
            ei_cmd.egg_base, PathMetadata(ei_cmd.egg_base, ei_cmd.egg_info),
            ei_cmd.egg_name, ei_cmd.egg_version,
        )
        bs_cmd = self.get_finalized_command('build_scripts')
        exec_param = getattr(bs_cmd, 'executable', None)
        try:
            bw_cmd = self.get_finalized_command("bdist_wininst")
            is_wininst = getattr(bw_cmd, '_is_running', False)
        except (ImportError, DistutilsModuleError):
            is_wininst = False
        writer = ei.ScriptWriter
        if is_wininst:
            exec_param = "python.exe"
            writer = ei.WindowsScriptWriter
        if exec_param == sys.executable:
            # In case the path to the Python executable contains a space, wrap
            # it so it's not split up.
            exec_param = [exec_param]
        # resolve the writer to the environment
        writer = writer.best()
        cmd = writer.command_spec_class.best().from_param(exec_param)
        for args in writer.get_args(dist, cmd.as_header()):
            self.write_script(*args)

    def write_script(self, script_name, contents, mode="t", *ignored):
        """Write an executable file to the scripts directory"""
        from setuptools.command.easy_install import chmod, current_umask

        log.info("Installing %s script to %s", script_name, self.install_dir)
        target = os.path.join(self.install_dir, script_name)
        self.outfiles.append(target)

        mask = current_umask()
        if not self.dry_run:
            ensure_directory(target)
            f = open(target, "w" + mode)
            f.write(contents)
            f.close()
            chmod(target, 0o777 - mask)
site-packages/setuptools/command/install.py000064400000011452150564745010015155 0ustar00from distutils.errors import DistutilsArgError
import inspect
import glob
import warnings
import platform
import distutils.command.install as orig

import setuptools

# Prior to numpy 1.9, NumPy relies on the '_install' name, so provide it for
# now. See https://github.com/pypa/setuptools/issues/199/
_install = orig.install


class install(orig.install):
    """Use easy_install to install the package, w/dependencies"""

    user_options = orig.install.user_options + [
        ('old-and-unmanageable', None, "Try not to use this!"),
        ('single-version-externally-managed', None,
         "used by system package builders to create 'flat' eggs"),
    ]
    boolean_options = orig.install.boolean_options + [
        'old-and-unmanageable', 'single-version-externally-managed',
    ]
    new_commands = [
        ('install_egg_info', lambda self: True),
        ('install_scripts', lambda self: True),
    ]
    _nc = dict(new_commands)

    def initialize_options(self):

        warnings.warn(
            "setup.py install is deprecated. "
            "Use build and pip and other standards-based tools.",
            setuptools.SetuptoolsDeprecationWarning,
        )

        orig.install.initialize_options(self)
        self.old_and_unmanageable = None
        self.single_version_externally_managed = None

    def finalize_options(self):
        orig.install.finalize_options(self)
        if self.root:
            self.single_version_externally_managed = True
        elif self.single_version_externally_managed:
            if not self.root and not self.record:
                raise DistutilsArgError(
                    "You must specify --record or --root when building system"
                    " packages"
                )

    def handle_extra_path(self):
        if self.root or self.single_version_externally_managed:
            # explicit backward-compatibility mode, allow extra_path to work
            return orig.install.handle_extra_path(self)

        # Ignore extra_path when installing an egg (or being run by another
        # command without --root or --single-version-externally-managed
        self.path_file = None
        self.extra_dirs = ''

    def run(self):
        # Explicit request for old-style install?  Just do it
        if self.old_and_unmanageable or self.single_version_externally_managed:
            return orig.install.run(self)

        if not self._called_from_setup(inspect.currentframe()):
            # Run in backward-compatibility mode to support bdist_* commands.
            orig.install.run(self)
        else:
            self.do_egg_install()

    @staticmethod
    def _called_from_setup(run_frame):
        """
        Attempt to detect whether run() was called from setup() or by another
        command.  If called by setup(), the parent caller will be the
        'run_command' method in 'distutils.dist', and *its* caller will be
        the 'run_commands' method.  If called any other way, the
        immediate caller *might* be 'run_command', but it won't have been
        called by 'run_commands'. Return True in that case or if a call stack
        is unavailable. Return False otherwise.
        """
        if run_frame is None:
            msg = "Call stack not available. bdist_* commands may fail."
            warnings.warn(msg)
            if platform.python_implementation() == 'IronPython':
                msg = "For best results, pass -X:Frames to enable call stack."
                warnings.warn(msg)
            return True
        res = inspect.getouterframes(run_frame)[2]
        caller, = res[:1]
        info = inspect.getframeinfo(caller)
        caller_module = caller.f_globals.get('__name__', '')
        return (
            caller_module == 'distutils.dist'
            and info.function == 'run_commands'
        )

    def do_egg_install(self):

        easy_install = self.distribution.get_command_class('easy_install')

        cmd = easy_install(
            self.distribution, args="x", root=self.root, record=self.record,
        )
        cmd.ensure_finalized()  # finalize before bdist_egg munges install cmd
        cmd.always_copy_from = '.'  # make sure local-dir eggs get installed

        # pick up setup-dir .egg files only: no .egg-info
        cmd.package_index.scan(glob.glob('*.egg'))

        self.run_command('bdist_egg')
        args = [self.distribution.get_command_obj('bdist_egg').egg_output]

        if setuptools.bootstrap_install_from:
            # Bootstrap self-installation of setuptools
            args.insert(0, setuptools.bootstrap_install_from)

        cmd.args = args
        cmd.run(show_deprecation=False)
        setuptools.bootstrap_install_from = None


# XXX Python 3.1 doesn't see _nc if this is inside the class
install.sub_commands = (
    [cmd for cmd in orig.install.sub_commands if cmd[0] not in install._nc] +
    install.new_commands
)
site-packages/setuptools/command/bdist_egg.py000064400000040334150564745010015437 0ustar00"""setuptools.command.bdist_egg

Build .egg distributions"""

from distutils.dir_util import remove_tree, mkpath
from distutils import log
from types import CodeType
import sys
import os
import re
import textwrap
import marshal

from pkg_resources import get_build_platform, Distribution, ensure_directory
from setuptools.extension import Library
from setuptools import Command

from sysconfig import get_path, get_python_version


def _get_purelib():
    return get_path("purelib")


def strip_module(filename):
    if '.' in filename:
        filename = os.path.splitext(filename)[0]
    if filename.endswith('module'):
        filename = filename[:-6]
    return filename


def sorted_walk(dir):
    """Do os.walk in a reproducible way,
    independent of indeterministic filesystem readdir order
    """
    for base, dirs, files in os.walk(dir):
        dirs.sort()
        files.sort()
        yield base, dirs, files


def write_stub(resource, pyfile):
    _stub_template = textwrap.dedent("""
        def __bootstrap__():
            global __bootstrap__, __loader__, __file__
            import sys, pkg_resources, importlib.util
            __file__ = pkg_resources.resource_filename(__name__, %r)
            __loader__ = None; del __bootstrap__, __loader__
            spec = importlib.util.spec_from_file_location(__name__,__file__)
            mod = importlib.util.module_from_spec(spec)
            spec.loader.exec_module(mod)
        __bootstrap__()
        """).lstrip()
    with open(pyfile, 'w') as f:
        f.write(_stub_template % resource)


class bdist_egg(Command):
    description = "create an \"egg\" distribution"

    user_options = [
        ('bdist-dir=', 'b',
         "temporary directory for creating the distribution"),
        ('plat-name=', 'p', "platform name to embed in generated filenames "
                            "(default: %s)" % get_build_platform()),
        ('exclude-source-files', None,
         "remove all .py files from the generated egg"),
        ('keep-temp', 'k',
         "keep the pseudo-installation tree around after " +
         "creating the distribution archive"),
        ('dist-dir=', 'd',
         "directory to put final built distributions in"),
        ('skip-build', None,
         "skip rebuilding everything (for testing/debugging)"),
    ]

    boolean_options = [
        'keep-temp', 'skip-build', 'exclude-source-files'
    ]

    def initialize_options(self):
        self.bdist_dir = None
        self.plat_name = None
        self.keep_temp = 0
        self.dist_dir = None
        self.skip_build = 0
        self.egg_output = None
        self.exclude_source_files = None

    def finalize_options(self):
        ei_cmd = self.ei_cmd = self.get_finalized_command("egg_info")
        self.egg_info = ei_cmd.egg_info

        if self.bdist_dir is None:
            bdist_base = self.get_finalized_command('bdist').bdist_base
            self.bdist_dir = os.path.join(bdist_base, 'egg')

        if self.plat_name is None:
            self.plat_name = get_build_platform()

        self.set_undefined_options('bdist', ('dist_dir', 'dist_dir'))

        if self.egg_output is None:

            # Compute filename of the output egg
            basename = Distribution(
                None, None, ei_cmd.egg_name, ei_cmd.egg_version,
                get_python_version(),
                self.distribution.has_ext_modules() and self.plat_name
            ).egg_name()

            self.egg_output = os.path.join(self.dist_dir, basename + '.egg')

    def do_install_data(self):
        # Hack for packages that install data to install's --install-lib
        self.get_finalized_command('install').install_lib = self.bdist_dir

        site_packages = os.path.normcase(os.path.realpath(_get_purelib()))
        old, self.distribution.data_files = self.distribution.data_files, []

        for item in old:
            if isinstance(item, tuple) and len(item) == 2:
                if os.path.isabs(item[0]):
                    realpath = os.path.realpath(item[0])
                    normalized = os.path.normcase(realpath)
                    if normalized == site_packages or normalized.startswith(
                        site_packages + os.sep
                    ):
                        item = realpath[len(site_packages) + 1:], item[1]
                        # XXX else: raise ???
            self.distribution.data_files.append(item)

        try:
            log.info("installing package data to %s", self.bdist_dir)
            self.call_command('install_data', force=0, root=None)
        finally:
            self.distribution.data_files = old

    def get_outputs(self):
        return [self.egg_output]

    def call_command(self, cmdname, **kw):
        """Invoke reinitialized command `cmdname` with keyword args"""
        for dirname in INSTALL_DIRECTORY_ATTRS:
            kw.setdefault(dirname, self.bdist_dir)
        kw.setdefault('skip_build', self.skip_build)
        kw.setdefault('dry_run', self.dry_run)
        cmd = self.reinitialize_command(cmdname, **kw)
        self.run_command(cmdname)
        return cmd

    def run(self):  # noqa: C901  # is too complex (14)  # FIXME
        # Generate metadata first
        self.run_command("egg_info")
        # We run install_lib before install_data, because some data hacks
        # pull their data path from the install_lib command.
        log.info("installing library code to %s", self.bdist_dir)
        instcmd = self.get_finalized_command('install')
        old_root = instcmd.root
        instcmd.root = None
        if self.distribution.has_c_libraries() and not self.skip_build:
            self.run_command('build_clib')
        cmd = self.call_command('install_lib', warn_dir=0)
        instcmd.root = old_root

        all_outputs, ext_outputs = self.get_ext_outputs()
        self.stubs = []
        to_compile = []
        for (p, ext_name) in enumerate(ext_outputs):
            filename, ext = os.path.splitext(ext_name)
            pyfile = os.path.join(self.bdist_dir, strip_module(filename) +
                                  '.py')
            self.stubs.append(pyfile)
            log.info("creating stub loader for %s", ext_name)
            if not self.dry_run:
                write_stub(os.path.basename(ext_name), pyfile)
            to_compile.append(pyfile)
            ext_outputs[p] = ext_name.replace(os.sep, '/')

        if to_compile:
            cmd.byte_compile(to_compile)
        if self.distribution.data_files:
            self.do_install_data()

        # Make the EGG-INFO directory
        archive_root = self.bdist_dir
        egg_info = os.path.join(archive_root, 'EGG-INFO')
        self.mkpath(egg_info)
        if self.distribution.scripts:
            script_dir = os.path.join(egg_info, 'scripts')
            log.info("installing scripts to %s", script_dir)
            self.call_command('install_scripts', install_dir=script_dir,
                              no_ep=1)

        self.copy_metadata_to(egg_info)
        native_libs = os.path.join(egg_info, "native_libs.txt")
        if all_outputs:
            log.info("writing %s", native_libs)
            if not self.dry_run:
                ensure_directory(native_libs)
                libs_file = open(native_libs, 'wt')
                libs_file.write('\n'.join(all_outputs))
                libs_file.write('\n')
                libs_file.close()
        elif os.path.isfile(native_libs):
            log.info("removing %s", native_libs)
            if not self.dry_run:
                os.unlink(native_libs)

        write_safety_flag(
            os.path.join(archive_root, 'EGG-INFO'), self.zip_safe()
        )

        if os.path.exists(os.path.join(self.egg_info, 'depends.txt')):
            log.warn(
                "WARNING: 'depends.txt' will not be used by setuptools 0.6!\n"
                "Use the install_requires/extras_require setup() args instead."
            )

        if self.exclude_source_files:
            self.zap_pyfiles()

        # Make the archive
        make_zipfile(self.egg_output, archive_root, verbose=self.verbose,
                     dry_run=self.dry_run, mode=self.gen_header())
        if not self.keep_temp:
            remove_tree(self.bdist_dir, dry_run=self.dry_run)

        # Add to 'Distribution.dist_files' so that the "upload" command works
        getattr(self.distribution, 'dist_files', []).append(
            ('bdist_egg', get_python_version(), self.egg_output))

    def zap_pyfiles(self):
        log.info("Removing .py files from temporary directory")
        for base, dirs, files in walk_egg(self.bdist_dir):
            for name in files:
                path = os.path.join(base, name)

                if name.endswith('.py'):
                    log.debug("Deleting %s", path)
                    os.unlink(path)

                if base.endswith('__pycache__'):
                    path_old = path

                    pattern = r'(?P<name>.+)\.(?P<magic>[^.]+)\.pyc'
                    m = re.match(pattern, name)
                    path_new = os.path.join(
                        base, os.pardir, m.group('name') + '.pyc')
                    log.info(
                        "Renaming file from [%s] to [%s]"
                        % (path_old, path_new))
                    try:
                        os.remove(path_new)
                    except OSError:
                        pass
                    os.rename(path_old, path_new)

    def zip_safe(self):
        safe = getattr(self.distribution, 'zip_safe', None)
        if safe is not None:
            return safe
        log.warn("zip_safe flag not set; analyzing archive contents...")
        return analyze_egg(self.bdist_dir, self.stubs)

    def gen_header(self):
        return 'w'

    def copy_metadata_to(self, target_dir):
        "Copy metadata (egg info) to the target_dir"
        # normalize the path (so that a forward-slash in egg_info will
        # match using startswith below)
        norm_egg_info = os.path.normpath(self.egg_info)
        prefix = os.path.join(norm_egg_info, '')
        for path in self.ei_cmd.filelist.files:
            if path.startswith(prefix):
                target = os.path.join(target_dir, path[len(prefix):])
                ensure_directory(target)
                self.copy_file(path, target)

    def get_ext_outputs(self):
        """Get a list of relative paths to C extensions in the output distro"""

        all_outputs = []
        ext_outputs = []

        paths = {self.bdist_dir: ''}
        for base, dirs, files in sorted_walk(self.bdist_dir):
            for filename in files:
                if os.path.splitext(filename)[1].lower() in NATIVE_EXTENSIONS:
                    all_outputs.append(paths[base] + filename)
            for filename in dirs:
                paths[os.path.join(base, filename)] = (paths[base] +
                                                       filename + '/')

        if self.distribution.has_ext_modules():
            build_cmd = self.get_finalized_command('build_ext')
            for ext in build_cmd.extensions:
                if isinstance(ext, Library):
                    continue
                fullname = build_cmd.get_ext_fullname(ext.name)
                filename = build_cmd.get_ext_filename(fullname)
                if not os.path.basename(filename).startswith('dl-'):
                    if os.path.exists(os.path.join(self.bdist_dir, filename)):
                        ext_outputs.append(filename)

        return all_outputs, ext_outputs


NATIVE_EXTENSIONS = dict.fromkeys('.dll .so .dylib .pyd'.split())


def walk_egg(egg_dir):
    """Walk an unpacked egg's contents, skipping the metadata directory"""
    walker = sorted_walk(egg_dir)
    base, dirs, files = next(walker)
    if 'EGG-INFO' in dirs:
        dirs.remove('EGG-INFO')
    yield base, dirs, files
    for bdf in walker:
        yield bdf


def analyze_egg(egg_dir, stubs):
    # check for existing flag in EGG-INFO
    for flag, fn in safety_flags.items():
        if os.path.exists(os.path.join(egg_dir, 'EGG-INFO', fn)):
            return flag
    if not can_scan():
        return False
    safe = True
    for base, dirs, files in walk_egg(egg_dir):
        for name in files:
            if name.endswith('.py') or name.endswith('.pyw'):
                continue
            elif name.endswith('.pyc') or name.endswith('.pyo'):
                # always scan, even if we already know we're not safe
                safe = scan_module(egg_dir, base, name, stubs) and safe
    return safe


def write_safety_flag(egg_dir, safe):
    # Write or remove zip safety flag file(s)
    for flag, fn in safety_flags.items():
        fn = os.path.join(egg_dir, fn)
        if os.path.exists(fn):
            if safe is None or bool(safe) != flag:
                os.unlink(fn)
        elif safe is not None and bool(safe) == flag:
            f = open(fn, 'wt')
            f.write('\n')
            f.close()


safety_flags = {
    True: 'zip-safe',
    False: 'not-zip-safe',
}


def scan_module(egg_dir, base, name, stubs):
    """Check whether module possibly uses unsafe-for-zipfile stuff"""

    filename = os.path.join(base, name)
    if filename[:-1] in stubs:
        return True  # Extension module
    pkg = base[len(egg_dir) + 1:].replace(os.sep, '.')
    module = pkg + (pkg and '.' or '') + os.path.splitext(name)[0]
    if sys.version_info < (3, 7):
        skip = 12  # skip magic & date & file size
    else:
        skip = 16  # skip magic & reserved? & date & file size
    f = open(filename, 'rb')
    f.read(skip)
    code = marshal.load(f)
    f.close()
    safe = True
    symbols = dict.fromkeys(iter_symbols(code))
    for bad in ['__file__', '__path__']:
        if bad in symbols:
            log.warn("%s: module references %s", module, bad)
            safe = False
    if 'inspect' in symbols:
        for bad in [
            'getsource', 'getabsfile', 'getsourcefile', 'getfile'
            'getsourcelines', 'findsource', 'getcomments', 'getframeinfo',
            'getinnerframes', 'getouterframes', 'stack', 'trace'
        ]:
            if bad in symbols:
                log.warn("%s: module MAY be using inspect.%s", module, bad)
                safe = False
    return safe


def iter_symbols(code):
    """Yield names and strings used by `code` and its nested code objects"""
    for name in code.co_names:
        yield name
    for const in code.co_consts:
        if isinstance(const, str):
            yield const
        elif isinstance(const, CodeType):
            for name in iter_symbols(const):
                yield name


def can_scan():
    if not sys.platform.startswith('java') and sys.platform != 'cli':
        # CPython, PyPy, etc.
        return True
    log.warn("Unable to analyze compiled code on this platform.")
    log.warn("Please ask the author to include a 'zip_safe'"
             " setting (either True or False) in the package's setup.py")


# Attribute names of options for commands that might need to be convinced to
# install to the egg build directory

INSTALL_DIRECTORY_ATTRS = [
    'install_lib', 'install_dir', 'install_data', 'install_base'
]


def make_zipfile(zip_filename, base_dir, verbose=0, dry_run=0, compress=True,
                 mode='w'):
    """Create a zip file from all the files under 'base_dir'.  The output
    zip file will be named 'base_dir' + ".zip".  Uses either the "zipfile"
    Python module (if available) or the InfoZIP "zip" utility (if installed
    and found on the default search path).  If neither tool is available,
    raises DistutilsExecError.  Returns the name of the output zip file.
    """
    import zipfile

    mkpath(os.path.dirname(zip_filename), dry_run=dry_run)
    log.info("creating '%s' and adding '%s' to it", zip_filename, base_dir)

    def visit(z, dirname, names):
        for name in names:
            path = os.path.normpath(os.path.join(dirname, name))
            if os.path.isfile(path):
                p = path[len(base_dir) + 1:]
                if not dry_run:
                    z.write(path, p)
                log.debug("adding '%s'", p)

    compression = zipfile.ZIP_DEFLATED if compress else zipfile.ZIP_STORED
    if not dry_run:
        z = zipfile.ZipFile(zip_filename, mode, compression=compression)
        for dirname, dirs, files in sorted_walk(base_dir):
            visit(z, dirname, files)
        z.close()
    else:
        for dirname, dirs, files in sorted_walk(base_dir):
            visit(None, dirname, files)
    return zip_filename
site-packages/setuptools/command/build_clib.py000064400000010477150564745010015605 0ustar00import distutils.command.build_clib as orig
from distutils.errors import DistutilsSetupError
from distutils import log
from setuptools.dep_util import newer_pairwise_group


class build_clib(orig.build_clib):
    """
    Override the default build_clib behaviour to do the following:

    1. Implement a rudimentary timestamp-based dependency system
       so 'compile()' doesn't run every time.
    2. Add more keys to the 'build_info' dictionary:
        * obj_deps - specify dependencies for each object compiled.
                     this should be a dictionary mapping a key
                     with the source filename to a list of
                     dependencies. Use an empty string for global
                     dependencies.
        * cflags   - specify a list of additional flags to pass to
                     the compiler.
    """

    def build_libraries(self, libraries):
        for (lib_name, build_info) in libraries:
            sources = build_info.get('sources')
            if sources is None or not isinstance(sources, (list, tuple)):
                raise DistutilsSetupError(
                    "in 'libraries' option (library '%s'), "
                    "'sources' must be present and must be "
                    "a list of source filenames" % lib_name)
            sources = list(sources)

            log.info("building '%s' library", lib_name)

            # Make sure everything is the correct type.
            # obj_deps should be a dictionary of keys as sources
            # and a list/tuple of files that are its dependencies.
            obj_deps = build_info.get('obj_deps', dict())
            if not isinstance(obj_deps, dict):
                raise DistutilsSetupError(
                    "in 'libraries' option (library '%s'), "
                    "'obj_deps' must be a dictionary of "
                    "type 'source: list'" % lib_name)
            dependencies = []

            # Get the global dependencies that are specified by the '' key.
            # These will go into every source's dependency list.
            global_deps = obj_deps.get('', list())
            if not isinstance(global_deps, (list, tuple)):
                raise DistutilsSetupError(
                    "in 'libraries' option (library '%s'), "
                    "'obj_deps' must be a dictionary of "
                    "type 'source: list'" % lib_name)

            # Build the list to be used by newer_pairwise_group
            # each source will be auto-added to its dependencies.
            for source in sources:
                src_deps = [source]
                src_deps.extend(global_deps)
                extra_deps = obj_deps.get(source, list())
                if not isinstance(extra_deps, (list, tuple)):
                    raise DistutilsSetupError(
                        "in 'libraries' option (library '%s'), "
                        "'obj_deps' must be a dictionary of "
                        "type 'source: list'" % lib_name)
                src_deps.extend(extra_deps)
                dependencies.append(src_deps)

            expected_objects = self.compiler.object_filenames(
                sources,
                output_dir=self.build_temp,
            )

            if (
                newer_pairwise_group(dependencies, expected_objects)
                != ([], [])
            ):
                # First, compile the source code to object files in the library
                # directory.  (This should probably change to putting object
                # files in a temporary build directory.)
                macros = build_info.get('macros')
                include_dirs = build_info.get('include_dirs')
                cflags = build_info.get('cflags')
                self.compiler.compile(
                    sources,
                    output_dir=self.build_temp,
                    macros=macros,
                    include_dirs=include_dirs,
                    extra_postargs=cflags,
                    debug=self.debug
                )

            # Now "link" the object files together into a static library.
            # (On Unix at least, this isn't really linking -- it just
            # builds an archive.  Whatever.)
            self.compiler.create_static_lib(
                expected_objects,
                lib_name,
                output_dir=self.build_clib,
                debug=self.debug
            )
site-packages/setuptools/command/__init__.py000064400000000331150564745010015240 0ustar00from distutils.command.bdist import bdist
import sys

if 'egg' not in bdist.format_commands:
    bdist.format_command['egg'] = ('bdist_egg', "Python .egg file")
    bdist.format_commands.append('egg')

del bdist, sys
site-packages/setuptools/command/upload_docs.py000064400000016062150564745010016005 0ustar00# -*- coding: utf-8 -*-
"""upload_docs

Implements a Distutils 'upload_docs' subcommand (upload documentation to
sites other than PyPi such as devpi).
"""

from base64 import standard_b64encode
from distutils import log
from distutils.errors import DistutilsOptionError
import os
import socket
import zipfile
import tempfile
import shutil
import itertools
import functools
import http.client
import urllib.parse

from pkg_resources import iter_entry_points
from .upload import upload


def _encode(s):
    return s.encode('utf-8', 'surrogateescape')


class upload_docs(upload):
    # override the default repository as upload_docs isn't
    # supported by Warehouse (and won't be).
    DEFAULT_REPOSITORY = 'https://pypi.python.org/pypi/'

    description = 'Upload documentation to sites other than PyPi such as devpi'

    user_options = [
        ('repository=', 'r',
         "url of repository [default: %s]" % upload.DEFAULT_REPOSITORY),
        ('show-response', None,
         'display full response text from server'),
        ('upload-dir=', None, 'directory to upload'),
    ]
    boolean_options = upload.boolean_options

    def has_sphinx(self):
        if self.upload_dir is None:
            for ep in iter_entry_points('distutils.commands', 'build_sphinx'):
                return True

    sub_commands = [('build_sphinx', has_sphinx)]

    def initialize_options(self):
        upload.initialize_options(self)
        self.upload_dir = None
        self.target_dir = None

    def finalize_options(self):
        upload.finalize_options(self)
        if self.upload_dir is None:
            if self.has_sphinx():
                build_sphinx = self.get_finalized_command('build_sphinx')
                self.target_dir = dict(build_sphinx.builder_target_dirs)['html']
            else:
                build = self.get_finalized_command('build')
                self.target_dir = os.path.join(build.build_base, 'docs')
        else:
            self.ensure_dirname('upload_dir')
            self.target_dir = self.upload_dir
        if 'pypi.python.org' in self.repository:
            log.warn("Upload_docs command is deprecated for PyPi. Use RTD instead.")
        self.announce('Using upload directory %s' % self.target_dir)

    def create_zipfile(self, filename):
        zip_file = zipfile.ZipFile(filename, "w")
        try:
            self.mkpath(self.target_dir)  # just in case
            for root, dirs, files in os.walk(self.target_dir):
                if root == self.target_dir and not files:
                    tmpl = "no files found in upload directory '%s'"
                    raise DistutilsOptionError(tmpl % self.target_dir)
                for name in files:
                    full = os.path.join(root, name)
                    relative = root[len(self.target_dir):].lstrip(os.path.sep)
                    dest = os.path.join(relative, name)
                    zip_file.write(full, dest)
        finally:
            zip_file.close()

    def run(self):
        # Run sub commands
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)

        tmp_dir = tempfile.mkdtemp()
        name = self.distribution.metadata.get_name()
        zip_file = os.path.join(tmp_dir, "%s.zip" % name)
        try:
            self.create_zipfile(zip_file)
            self.upload_file(zip_file)
        finally:
            shutil.rmtree(tmp_dir)

    @staticmethod
    def _build_part(item, sep_boundary):
        key, values = item
        title = '\nContent-Disposition: form-data; name="%s"' % key
        # handle multiple entries for the same name
        if not isinstance(values, list):
            values = [values]
        for value in values:
            if isinstance(value, tuple):
                title += '; filename="%s"' % value[0]
                value = value[1]
            else:
                value = _encode(value)
            yield sep_boundary
            yield _encode(title)
            yield b"\n\n"
            yield value
            if value and value[-1:] == b'\r':
                yield b'\n'  # write an extra newline (lurve Macs)

    @classmethod
    def _build_multipart(cls, data):
        """
        Build up the MIME payload for the POST data
        """
        boundary = '--------------GHSKFJDLGDS7543FJKLFHRE75642756743254'
        sep_boundary = b'\n--' + boundary.encode('ascii')
        end_boundary = sep_boundary + b'--'
        end_items = end_boundary, b"\n",
        builder = functools.partial(
            cls._build_part,
            sep_boundary=sep_boundary,
        )
        part_groups = map(builder, data.items())
        parts = itertools.chain.from_iterable(part_groups)
        body_items = itertools.chain(parts, end_items)
        content_type = 'multipart/form-data; boundary=%s' % boundary
        return b''.join(body_items), content_type

    def upload_file(self, filename):
        with open(filename, 'rb') as f:
            content = f.read()
        meta = self.distribution.metadata
        data = {
            ':action': 'doc_upload',
            'name': meta.get_name(),
            'content': (os.path.basename(filename), content),
        }
        # set up the authentication
        credentials = _encode(self.username + ':' + self.password)
        credentials = standard_b64encode(credentials).decode('ascii')
        auth = "Basic " + credentials

        body, ct = self._build_multipart(data)

        msg = "Submitting documentation to %s" % (self.repository)
        self.announce(msg, log.INFO)

        # build the Request
        # We can't use urllib2 since we need to send the Basic
        # auth right with the first request
        schema, netloc, url, params, query, fragments = \
            urllib.parse.urlparse(self.repository)
        assert not params and not query and not fragments
        if schema == 'http':
            conn = http.client.HTTPConnection(netloc)
        elif schema == 'https':
            conn = http.client.HTTPSConnection(netloc)
        else:
            raise AssertionError("unsupported schema " + schema)

        data = ''
        try:
            conn.connect()
            conn.putrequest("POST", url)
            content_type = ct
            conn.putheader('Content-type', content_type)
            conn.putheader('Content-length', str(len(body)))
            conn.putheader('Authorization', auth)
            conn.endheaders()
            conn.send(body)
        except socket.error as e:
            self.announce(str(e), log.ERROR)
            return

        r = conn.getresponse()
        if r.status == 200:
            msg = 'Server response (%s): %s' % (r.status, r.reason)
            self.announce(msg, log.INFO)
        elif r.status == 301:
            location = r.getheader('Location')
            if location is None:
                location = 'https://pythonhosted.org/%s/' % meta.get_name()
            msg = 'Upload successful. Visit %s' % location
            self.announce(msg, log.INFO)
        else:
            msg = 'Upload failed (%s): %s' % (r.status, r.reason)
            self.announce(msg, log.ERROR)
        if self.show_response:
            print('-' * 75, r.read(), '-' * 75)
site-packages/setuptools/command/develop.py000064400000015544150564745010015153 0ustar00from distutils.util import convert_path
from distutils import log
from distutils.errors import DistutilsError, DistutilsOptionError
import os
import glob
import io

import pkg_resources
from setuptools.command.easy_install import easy_install
from setuptools import namespaces
import setuptools


class develop(namespaces.DevelopInstaller, easy_install):
    """Set up package for development"""

    description = "install package in 'development mode'"

    user_options = easy_install.user_options + [
        ("uninstall", "u", "Uninstall this source package"),
        ("egg-path=", None, "Set the path to be used in the .egg-link file"),
    ]

    boolean_options = easy_install.boolean_options + ['uninstall']

    command_consumes_arguments = False  # override base

    def run(self):
        if self.uninstall:
            self.multi_version = True
            self.uninstall_link()
            self.uninstall_namespaces()
        else:
            self.install_for_development()
        self.warn_deprecated_options()

    def initialize_options(self):
        self.uninstall = None
        self.egg_path = None
        easy_install.initialize_options(self)
        self.setup_path = None
        self.always_copy_from = '.'  # always copy eggs installed in curdir

    def finalize_options(self):
        ei = self.get_finalized_command("egg_info")
        if ei.broken_egg_info:
            template = "Please rename %r to %r before using 'develop'"
            args = ei.egg_info, ei.broken_egg_info
            raise DistutilsError(template % args)
        self.args = [ei.egg_name]

        easy_install.finalize_options(self)
        self.expand_basedirs()
        self.expand_dirs()
        # pick up setup-dir .egg files only: no .egg-info
        self.package_index.scan(glob.glob('*.egg'))

        egg_link_fn = ei.egg_name + '.egg-link'
        self.egg_link = os.path.join(self.install_dir, egg_link_fn)
        self.egg_base = ei.egg_base
        if self.egg_path is None:
            self.egg_path = os.path.abspath(ei.egg_base)

        target = pkg_resources.normalize_path(self.egg_base)
        egg_path = pkg_resources.normalize_path(
            os.path.join(self.install_dir, self.egg_path)
        )
        if egg_path != target:
            raise DistutilsOptionError(
                "--egg-path must be a relative path from the install"
                " directory to " + target
            )

        # Make a distribution for the package's source
        self.dist = pkg_resources.Distribution(
            target,
            pkg_resources.PathMetadata(target, os.path.abspath(ei.egg_info)),
            project_name=ei.egg_name,
        )

        self.setup_path = self._resolve_setup_path(
            self.egg_base,
            self.install_dir,
            self.egg_path,
        )

    @staticmethod
    def _resolve_setup_path(egg_base, install_dir, egg_path):
        """
        Generate a path from egg_base back to '.' where the
        setup script resides and ensure that path points to the
        setup path from $install_dir/$egg_path.
        """
        path_to_setup = egg_base.replace(os.sep, '/').rstrip('/')
        if path_to_setup != os.curdir:
            path_to_setup = '../' * (path_to_setup.count('/') + 1)
        resolved = pkg_resources.normalize_path(
            os.path.join(install_dir, egg_path, path_to_setup)
        )
        if resolved != pkg_resources.normalize_path(os.curdir):
            raise DistutilsOptionError(
                "Can't get a consistent path to setup script from"
                " installation directory",
                resolved,
                pkg_resources.normalize_path(os.curdir),
            )
        return path_to_setup

    def install_for_development(self):
        self.run_command('egg_info')

        # Build extensions in-place
        self.reinitialize_command('build_ext', inplace=1)
        self.run_command('build_ext')

        if setuptools.bootstrap_install_from:
            self.easy_install(setuptools.bootstrap_install_from)
            setuptools.bootstrap_install_from = None

        self.install_namespaces()

        # create an .egg-link in the installation dir, pointing to our egg
        log.info("Creating %s (link to %s)", self.egg_link, self.egg_base)
        if not self.dry_run:
            with open(self.egg_link, "w") as f:
                f.write(self.egg_path + "\n" + self.setup_path)
        # postprocess the installed distro, fixing up .pth, installing scripts,
        # and handling requirements
        self.process_distribution(None, self.dist, not self.no_deps)

    def uninstall_link(self):
        if os.path.exists(self.egg_link):
            log.info("Removing %s (link to %s)", self.egg_link, self.egg_base)
            egg_link_file = open(self.egg_link)
            contents = [line.rstrip() for line in egg_link_file]
            egg_link_file.close()
            if contents not in ([self.egg_path], [self.egg_path, self.setup_path]):
                log.warn("Link points to %s: uninstall aborted", contents)
                return
            if not self.dry_run:
                os.unlink(self.egg_link)
        if not self.dry_run:
            self.update_pth(self.dist)  # remove any .pth link to us
        if self.distribution.scripts:
            # XXX should also check for entry point scripts!
            log.warn("Note: you must uninstall or replace scripts manually!")

    def install_egg_scripts(self, dist):
        if dist is not self.dist:
            # Installing a dependency, so fall back to normal behavior
            return easy_install.install_egg_scripts(self, dist)

        # create wrapper scripts in the script dir, pointing to dist.scripts

        # new-style...
        self.install_wrapper_scripts(dist)

        # ...and old-style
        for script_name in self.distribution.scripts or []:
            script_path = os.path.abspath(convert_path(script_name))
            script_name = os.path.basename(script_path)
            with io.open(script_path) as strm:
                script_text = strm.read()
            self.install_script(dist, script_name, script_text, script_path)

    def install_wrapper_scripts(self, dist):
        dist = VersionlessRequirement(dist)
        return easy_install.install_wrapper_scripts(self, dist)


class VersionlessRequirement:
    """
    Adapt a pkg_resources.Distribution to simply return the project
    name as the 'requirement' so that scripts will work across
    multiple versions.

    >>> from pkg_resources import Distribution
    >>> dist = Distribution(project_name='foo', version='1.0')
    >>> str(dist.as_requirement())
    'foo==1.0'
    >>> adapted_dist = VersionlessRequirement(dist)
    >>> str(adapted_dist.as_requirement())
    'foo'
    """

    def __init__(self, dist):
        self.__dist = dist

    def __getattr__(self, name):
        return getattr(self.__dist, name)

    def as_requirement(self):
        return self.project_name
site-packages/setuptools/command/egg_info.py000064400000061367150564745010015276 0ustar00"""setuptools.command.egg_info

Create a distribution's .egg-info directory and contents"""

from distutils.filelist import FileList as _FileList
from distutils.errors import DistutilsInternalError
from distutils.util import convert_path
from distutils import log
import distutils.errors
import distutils.filelist
import functools
import os
import re
import sys
import io
import warnings
import time
import collections

from setuptools import Command
from setuptools.command.sdist import sdist
from setuptools.command.sdist import walk_revctrl
from setuptools.command.setopt import edit_config
from setuptools.command import bdist_egg
from pkg_resources import (
    parse_requirements, safe_name, parse_version,
    safe_version, yield_lines, EntryPoint, iter_entry_points, to_filename)
import setuptools.unicode_utils as unicode_utils
from setuptools.glob import glob

from setuptools.extern import packaging
from setuptools import SetuptoolsDeprecationWarning


def translate_pattern(glob):  # noqa: C901  # is too complex (14)  # FIXME
    """
    Translate a file path glob like '*.txt' in to a regular expression.
    This differs from fnmatch.translate which allows wildcards to match
    directory separators. It also knows about '**/' which matches any number of
    directories.
    """
    pat = ''

    # This will split on '/' within [character classes]. This is deliberate.
    chunks = glob.split(os.path.sep)

    sep = re.escape(os.sep)
    valid_char = '[^%s]' % (sep,)

    for c, chunk in enumerate(chunks):
        last_chunk = c == len(chunks) - 1

        # Chunks that are a literal ** are globstars. They match anything.
        if chunk == '**':
            if last_chunk:
                # Match anything if this is the last component
                pat += '.*'
            else:
                # Match '(name/)*'
                pat += '(?:%s+%s)*' % (valid_char, sep)
            continue  # Break here as the whole path component has been handled

        # Find any special characters in the remainder
        i = 0
        chunk_len = len(chunk)
        while i < chunk_len:
            char = chunk[i]
            if char == '*':
                # Match any number of name characters
                pat += valid_char + '*'
            elif char == '?':
                # Match a name character
                pat += valid_char
            elif char == '[':
                # Character class
                inner_i = i + 1
                # Skip initial !/] chars
                if inner_i < chunk_len and chunk[inner_i] == '!':
                    inner_i = inner_i + 1
                if inner_i < chunk_len and chunk[inner_i] == ']':
                    inner_i = inner_i + 1

                # Loop till the closing ] is found
                while inner_i < chunk_len and chunk[inner_i] != ']':
                    inner_i = inner_i + 1

                if inner_i >= chunk_len:
                    # Got to the end of the string without finding a closing ]
                    # Do not treat this as a matching group, but as a literal [
                    pat += re.escape(char)
                else:
                    # Grab the insides of the [brackets]
                    inner = chunk[i + 1:inner_i]
                    char_class = ''

                    # Class negation
                    if inner[0] == '!':
                        char_class = '^'
                        inner = inner[1:]

                    char_class += re.escape(inner)
                    pat += '[%s]' % (char_class,)

                    # Skip to the end ]
                    i = inner_i
            else:
                pat += re.escape(char)
            i += 1

        # Join each chunk with the dir separator
        if not last_chunk:
            pat += sep

    pat += r'\Z'
    return re.compile(pat, flags=re.MULTILINE | re.DOTALL)


class InfoCommon:
    tag_build = None
    tag_date = None

    @property
    def name(self):
        return safe_name(self.distribution.get_name())

    def tagged_version(self):
        return safe_version(self._maybe_tag(self.distribution.get_version()))

    def _maybe_tag(self, version):
        """
        egg_info may be called more than once for a distribution,
        in which case the version string already contains all tags.
        """
        return (
            version if self.vtags and version.endswith(self.vtags)
            else version + self.vtags
        )

    def tags(self):
        version = ''
        if self.tag_build:
            version += self.tag_build
        if self.tag_date:
            version += time.strftime("-%Y%m%d")
        return version
    vtags = property(tags)


class egg_info(InfoCommon, Command):
    description = "create a distribution's .egg-info directory"

    user_options = [
        ('egg-base=', 'e', "directory containing .egg-info directories"
                           " (default: top of the source tree)"),
        ('tag-date', 'd', "Add date stamp (e.g. 20050528) to version number"),
        ('tag-build=', 'b', "Specify explicit tag to add to version number"),
        ('no-date', 'D', "Don't include date stamp [default]"),
    ]

    boolean_options = ['tag-date']
    negative_opt = {
        'no-date': 'tag-date',
    }

    def initialize_options(self):
        self.egg_base = None
        self.egg_name = None
        self.egg_info = None
        self.egg_version = None
        self.broken_egg_info = False

    ####################################
    # allow the 'tag_svn_revision' to be detected and
    # set, supporting sdists built on older Setuptools.
    @property
    def tag_svn_revision(self):
        pass

    @tag_svn_revision.setter
    def tag_svn_revision(self, value):
        pass
    ####################################

    def save_version_info(self, filename):
        """
        Materialize the value of date into the
        build tag. Install build keys in a deterministic order
        to avoid arbitrary reordering on subsequent builds.
        """
        egg_info = collections.OrderedDict()
        # follow the order these keys would have been added
        # when PYTHONHASHSEED=0
        egg_info['tag_build'] = self.tags()
        egg_info['tag_date'] = 0
        edit_config(filename, dict(egg_info=egg_info))

    def finalize_options(self):
        # Note: we need to capture the current value returned
        # by `self.tagged_version()`, so we can later update
        # `self.distribution.metadata.version` without
        # repercussions.
        self.egg_name = self.name
        self.egg_version = self.tagged_version()
        parsed_version = parse_version(self.egg_version)

        try:
            is_version = isinstance(parsed_version, packaging.version.Version)
            spec = (
                "%s==%s" if is_version else "%s===%s"
            )
            list(
                parse_requirements(spec % (self.egg_name, self.egg_version))
            )
        except ValueError as e:
            raise distutils.errors.DistutilsOptionError(
                "Invalid distribution name or version syntax: %s-%s" %
                (self.egg_name, self.egg_version)
            ) from e

        if self.egg_base is None:
            dirs = self.distribution.package_dir
            self.egg_base = (dirs or {}).get('', os.curdir)

        self.ensure_dirname('egg_base')
        self.egg_info = to_filename(self.egg_name) + '.egg-info'
        if self.egg_base != os.curdir:
            self.egg_info = os.path.join(self.egg_base, self.egg_info)
        if '-' in self.egg_name:
            self.check_broken_egg_info()

        # Set package version for the benefit of dumber commands
        # (e.g. sdist, bdist_wininst, etc.)
        #
        self.distribution.metadata.version = self.egg_version

        # If we bootstrapped around the lack of a PKG-INFO, as might be the
        # case in a fresh checkout, make sure that any special tags get added
        # to the version info
        #
        pd = self.distribution._patched_dist
        if pd is not None and pd.key == self.egg_name.lower():
            pd._version = self.egg_version
            pd._parsed_version = parse_version(self.egg_version)
            self.distribution._patched_dist = None

    def write_or_delete_file(self, what, filename, data, force=False):
        """Write `data` to `filename` or delete if empty

        If `data` is non-empty, this routine is the same as ``write_file()``.
        If `data` is empty but not ``None``, this is the same as calling
        ``delete_file(filename)`.  If `data` is ``None``, then this is a no-op
        unless `filename` exists, in which case a warning is issued about the
        orphaned file (if `force` is false), or deleted (if `force` is true).
        """
        if data:
            self.write_file(what, filename, data)
        elif os.path.exists(filename):
            if data is None and not force:
                log.warn(
                    "%s not set in setup(), but %s exists", what, filename
                )
                return
            else:
                self.delete_file(filename)

    def write_file(self, what, filename, data):
        """Write `data` to `filename` (if not a dry run) after announcing it

        `what` is used in a log message to identify what is being written
        to the file.
        """
        log.info("writing %s to %s", what, filename)
        data = data.encode("utf-8")
        if not self.dry_run:
            f = open(filename, 'wb')
            f.write(data)
            f.close()

    def delete_file(self, filename):
        """Delete `filename` (if not a dry run) after announcing it"""
        log.info("deleting %s", filename)
        if not self.dry_run:
            os.unlink(filename)

    def run(self):
        self.mkpath(self.egg_info)
        os.utime(self.egg_info, None)
        installer = self.distribution.fetch_build_egg
        for ep in iter_entry_points('egg_info.writers'):
            ep.require(installer=installer)
            writer = ep.resolve()
            writer(self, ep.name, os.path.join(self.egg_info, ep.name))

        # Get rid of native_libs.txt if it was put there by older bdist_egg
        nl = os.path.join(self.egg_info, "native_libs.txt")
        if os.path.exists(nl):
            self.delete_file(nl)

        self.find_sources()

    def find_sources(self):
        """Generate SOURCES.txt manifest file"""
        manifest_filename = os.path.join(self.egg_info, "SOURCES.txt")
        mm = manifest_maker(self.distribution)
        mm.manifest = manifest_filename
        mm.run()
        self.filelist = mm.filelist

    def check_broken_egg_info(self):
        bei = self.egg_name + '.egg-info'
        if self.egg_base != os.curdir:
            bei = os.path.join(self.egg_base, bei)
        if os.path.exists(bei):
            log.warn(
                "-" * 78 + '\n'
                "Note: Your current .egg-info directory has a '-' in its name;"
                '\nthis will not work correctly with "setup.py develop".\n\n'
                'Please rename %s to %s to correct this problem.\n' + '-' * 78,
                bei, self.egg_info
            )
            self.broken_egg_info = self.egg_info
            self.egg_info = bei  # make it work for now


class FileList(_FileList):
    # Implementations of the various MANIFEST.in commands

    def process_template_line(self, line):
        # Parse the line: split it up, make sure the right number of words
        # is there, and return the relevant words.  'action' is always
        # defined: it's the first word of the line.  Which of the other
        # three are defined depends on the action; it'll be either
        # patterns, (dir and patterns), or (dir_pattern).
        (action, patterns, dir, dir_pattern) = self._parse_template_line(line)

        action_map = {
            'include': self.include,
            'exclude': self.exclude,
            'global-include': self.global_include,
            'global-exclude': self.global_exclude,
            'recursive-include': functools.partial(
                self.recursive_include, dir,
            ),
            'recursive-exclude': functools.partial(
                self.recursive_exclude, dir,
            ),
            'graft': self.graft,
            'prune': self.prune,
        }
        log_map = {
            'include': "warning: no files found matching '%s'",
            'exclude': (
                "warning: no previously-included files found "
                "matching '%s'"
            ),
            'global-include': (
                "warning: no files found matching '%s' "
                "anywhere in distribution"
            ),
            'global-exclude': (
                "warning: no previously-included files matching "
                "'%s' found anywhere in distribution"
            ),
            'recursive-include': (
                "warning: no files found matching '%s' "
                "under directory '%s'"
            ),
            'recursive-exclude': (
                "warning: no previously-included files matching "
                "'%s' found under directory '%s'"
            ),
            'graft': "warning: no directories found matching '%s'",
            'prune': "no previously-included directories found matching '%s'",
        }

        try:
            process_action = action_map[action]
        except KeyError:
            raise DistutilsInternalError(
                "this cannot happen: invalid action '{action!s}'".
                format(action=action),
            )

        # OK, now we know that the action is valid and we have the
        # right number of words on the line for that action -- so we
        # can proceed with minimal error-checking.

        action_is_recursive = action.startswith('recursive-')
        if action in {'graft', 'prune'}:
            patterns = [dir_pattern]
        extra_log_args = (dir, ) if action_is_recursive else ()
        log_tmpl = log_map[action]

        self.debug_print(
            ' '.join(
                [action] +
                ([dir] if action_is_recursive else []) +
                patterns,
            )
        )
        for pattern in patterns:
            if not process_action(pattern):
                log.warn(log_tmpl, pattern, *extra_log_args)

    def _remove_files(self, predicate):
        """
        Remove all files from the file list that match the predicate.
        Return True if any matching files were removed
        """
        found = False
        for i in range(len(self.files) - 1, -1, -1):
            if predicate(self.files[i]):
                self.debug_print(" removing " + self.files[i])
                del self.files[i]
                found = True
        return found

    def include(self, pattern):
        """Include files that match 'pattern'."""
        found = [f for f in glob(pattern) if not os.path.isdir(f)]
        self.extend(found)
        return bool(found)

    def exclude(self, pattern):
        """Exclude files that match 'pattern'."""
        match = translate_pattern(pattern)
        return self._remove_files(match.match)

    def recursive_include(self, dir, pattern):
        """
        Include all files anywhere in 'dir/' that match the pattern.
        """
        full_pattern = os.path.join(dir, '**', pattern)
        found = [f for f in glob(full_pattern, recursive=True)
                 if not os.path.isdir(f)]
        self.extend(found)
        return bool(found)

    def recursive_exclude(self, dir, pattern):
        """
        Exclude any file anywhere in 'dir/' that match the pattern.
        """
        match = translate_pattern(os.path.join(dir, '**', pattern))
        return self._remove_files(match.match)

    def graft(self, dir):
        """Include all files from 'dir/'."""
        found = [
            item
            for match_dir in glob(dir)
            for item in distutils.filelist.findall(match_dir)
        ]
        self.extend(found)
        return bool(found)

    def prune(self, dir):
        """Filter out files from 'dir/'."""
        match = translate_pattern(os.path.join(dir, '**'))
        return self._remove_files(match.match)

    def global_include(self, pattern):
        """
        Include all files anywhere in the current directory that match the
        pattern. This is very inefficient on large file trees.
        """
        if self.allfiles is None:
            self.findall()
        match = translate_pattern(os.path.join('**', pattern))
        found = [f for f in self.allfiles if match.match(f)]
        self.extend(found)
        return bool(found)

    def global_exclude(self, pattern):
        """
        Exclude all files anywhere that match the pattern.
        """
        match = translate_pattern(os.path.join('**', pattern))
        return self._remove_files(match.match)

    def append(self, item):
        if item.endswith('\r'):  # Fix older sdists built on Windows
            item = item[:-1]
        path = convert_path(item)

        if self._safe_path(path):
            self.files.append(path)

    def extend(self, paths):
        self.files.extend(filter(self._safe_path, paths))

    def _repair(self):
        """
        Replace self.files with only safe paths

        Because some owners of FileList manipulate the underlying
        ``files`` attribute directly, this method must be called to
        repair those paths.
        """
        self.files = list(filter(self._safe_path, self.files))

    def _safe_path(self, path):
        enc_warn = "'%s' not %s encodable -- skipping"

        # To avoid accidental trans-codings errors, first to unicode
        u_path = unicode_utils.filesys_decode(path)
        if u_path is None:
            log.warn("'%s' in unexpected encoding -- skipping" % path)
            return False

        # Must ensure utf-8 encodability
        utf8_path = unicode_utils.try_encode(u_path, "utf-8")
        if utf8_path is None:
            log.warn(enc_warn, path, 'utf-8')
            return False

        try:
            # accept is either way checks out
            if os.path.exists(u_path) or os.path.exists(utf8_path):
                return True
        # this will catch any encode errors decoding u_path
        except UnicodeEncodeError:
            log.warn(enc_warn, path, sys.getfilesystemencoding())


class manifest_maker(sdist):
    template = "MANIFEST.in"

    def initialize_options(self):
        self.use_defaults = 1
        self.prune = 1
        self.manifest_only = 1
        self.force_manifest = 1

    def finalize_options(self):
        pass

    def run(self):
        self.filelist = FileList()
        if not os.path.exists(self.manifest):
            self.write_manifest()  # it must exist so it'll get in the list
        self.add_defaults()
        if os.path.exists(self.template):
            self.read_template()
        self.add_license_files()
        self.prune_file_list()
        self.filelist.sort()
        self.filelist.remove_duplicates()
        self.write_manifest()

    def _manifest_normalize(self, path):
        path = unicode_utils.filesys_decode(path)
        return path.replace(os.sep, '/')

    def write_manifest(self):
        """
        Write the file list in 'self.filelist' to the manifest file
        named by 'self.manifest'.
        """
        self.filelist._repair()

        # Now _repairs should encodability, but not unicode
        files = [self._manifest_normalize(f) for f in self.filelist.files]
        msg = "writing manifest file '%s'" % self.manifest
        self.execute(write_file, (self.manifest, files), msg)

    def warn(self, msg):
        if not self._should_suppress_warning(msg):
            sdist.warn(self, msg)

    @staticmethod
    def _should_suppress_warning(msg):
        """
        suppress missing-file warnings from sdist
        """
        return re.match(r"standard file .*not found", msg)

    def add_defaults(self):
        sdist.add_defaults(self)
        self.filelist.append(self.template)
        self.filelist.append(self.manifest)
        rcfiles = list(walk_revctrl())
        if rcfiles:
            self.filelist.extend(rcfiles)
        elif os.path.exists(self.manifest):
            self.read_manifest()

        if os.path.exists("setup.py"):
            # setup.py should be included by default, even if it's not
            # the script called to create the sdist
            self.filelist.append("setup.py")

        ei_cmd = self.get_finalized_command('egg_info')
        self.filelist.graft(ei_cmd.egg_info)

    def add_license_files(self):
        license_files = self.distribution.metadata.license_files or []
        for lf in license_files:
            log.info("adding license file '%s'", lf)
            pass
        self.filelist.extend(license_files)

    def prune_file_list(self):
        build = self.get_finalized_command('build')
        base_dir = self.distribution.get_fullname()
        self.filelist.prune(build.build_base)
        self.filelist.prune(base_dir)
        sep = re.escape(os.sep)
        self.filelist.exclude_pattern(r'(^|' + sep + r')(RCS|CVS|\.svn)' + sep,
                                      is_regex=1)


def write_file(filename, contents):
    """Create a file with the specified name and write 'contents' (a
    sequence of strings without line terminators) to it.
    """
    contents = "\n".join(contents)

    # assuming the contents has been vetted for utf-8 encoding
    contents = contents.encode("utf-8")

    with open(filename, "wb") as f:  # always write POSIX-style manifest
        f.write(contents)


def write_pkg_info(cmd, basename, filename):
    log.info("writing %s", filename)
    if not cmd.dry_run:
        metadata = cmd.distribution.metadata
        metadata.version, oldver = cmd.egg_version, metadata.version
        metadata.name, oldname = cmd.egg_name, metadata.name

        try:
            # write unescaped data to PKG-INFO, so older pkg_resources
            # can still parse it
            metadata.write_pkg_info(cmd.egg_info)
        finally:
            metadata.name, metadata.version = oldname, oldver

        safe = getattr(cmd.distribution, 'zip_safe', None)

        bdist_egg.write_safety_flag(cmd.egg_info, safe)


def warn_depends_obsolete(cmd, basename, filename):
    if os.path.exists(filename):
        log.warn(
            "WARNING: 'depends.txt' is not used by setuptools 0.6!\n"
            "Use the install_requires/extras_require setup() args instead."
        )


def _write_requirements(stream, reqs):
    lines = yield_lines(reqs or ())

    def append_cr(line):
        return line + '\n'
    lines = map(append_cr, lines)
    stream.writelines(lines)


def write_requirements(cmd, basename, filename):
    dist = cmd.distribution
    data = io.StringIO()
    _write_requirements(data, dist.install_requires)
    extras_require = dist.extras_require or {}
    for extra in sorted(extras_require):
        data.write('\n[{extra}]\n'.format(**vars()))
        _write_requirements(data, extras_require[extra])
    cmd.write_or_delete_file("requirements", filename, data.getvalue())


def write_setup_requirements(cmd, basename, filename):
    data = io.StringIO()
    _write_requirements(data, cmd.distribution.setup_requires)
    cmd.write_or_delete_file("setup-requirements", filename, data.getvalue())


def write_toplevel_names(cmd, basename, filename):
    pkgs = dict.fromkeys(
        [
            k.split('.', 1)[0]
            for k in cmd.distribution.iter_distribution_names()
        ]
    )
    cmd.write_file("top-level names", filename, '\n'.join(sorted(pkgs)) + '\n')


def overwrite_arg(cmd, basename, filename):
    write_arg(cmd, basename, filename, True)


def write_arg(cmd, basename, filename, force=False):
    argname = os.path.splitext(basename)[0]
    value = getattr(cmd.distribution, argname, None)
    if value is not None:
        value = '\n'.join(value) + '\n'
    cmd.write_or_delete_file(argname, filename, value, force)


def write_entries(cmd, basename, filename):
    ep = cmd.distribution.entry_points

    if isinstance(ep, str) or ep is None:
        data = ep
    elif ep is not None:
        data = []
        for section, contents in sorted(ep.items()):
            if not isinstance(contents, str):
                contents = EntryPoint.parse_group(section, contents)
                contents = '\n'.join(sorted(map(str, contents.values())))
            data.append('[%s]\n%s\n\n' % (section, contents))
        data = ''.join(data)

    cmd.write_or_delete_file('entry points', filename, data, True)


def get_pkg_info_revision():
    """
    Get a -r### off of PKG-INFO Version in case this is an sdist of
    a subversion revision.
    """
    warnings.warn(
        "get_pkg_info_revision is deprecated.", EggInfoDeprecationWarning)
    if os.path.exists('PKG-INFO'):
        with io.open('PKG-INFO') as f:
            for line in f:
                match = re.match(r"Version:.*-r(\d+)\s*$", line)
                if match:
                    return int(match.group(1))
    return 0


class EggInfoDeprecationWarning(SetuptoolsDeprecationWarning):
    """Deprecated behavior warning for EggInfo, bypassing suppression."""
site-packages/setuptools/command/alias.py000064400000004515150564745010014602 0ustar00from distutils.errors import DistutilsOptionError

from setuptools.command.setopt import edit_config, option_base, config_file


def shquote(arg):
    """Quote an argument for later parsing by shlex.split()"""
    for c in '"', "'", "\\", "#":
        if c in arg:
            return repr(arg)
    if arg.split() != [arg]:
        return repr(arg)
    return arg


class alias(option_base):
    """Define a shortcut that invokes one or more commands"""

    description = "define a shortcut to invoke one or more commands"
    command_consumes_arguments = True

    user_options = [
        ('remove', 'r', 'remove (unset) the alias'),
    ] + option_base.user_options

    boolean_options = option_base.boolean_options + ['remove']

    def initialize_options(self):
        option_base.initialize_options(self)
        self.args = None
        self.remove = None

    def finalize_options(self):
        option_base.finalize_options(self)
        if self.remove and len(self.args) != 1:
            raise DistutilsOptionError(
                "Must specify exactly one argument (the alias name) when "
                "using --remove"
            )

    def run(self):
        aliases = self.distribution.get_option_dict('aliases')

        if not self.args:
            print("Command Aliases")
            print("---------------")
            for alias in aliases:
                print("setup.py alias", format_alias(alias, aliases))
            return

        elif len(self.args) == 1:
            alias, = self.args
            if self.remove:
                command = None
            elif alias in aliases:
                print("setup.py alias", format_alias(alias, aliases))
                return
            else:
                print("No alias definition found for %r" % alias)
                return
        else:
            alias = self.args[0]
            command = ' '.join(map(shquote, self.args[1:]))

        edit_config(self.filename, {'aliases': {alias: command}}, self.dry_run)


def format_alias(name, aliases):
    source, command = aliases[name]
    if source == config_file('global'):
        source = '--global-config '
    elif source == config_file('user'):
        source = '--user-config '
    elif source == config_file('local'):
        source = ''
    else:
        source = '--filename=%r' % source
    return source + name + ' ' + command
site-packages/setuptools/command/bdist_rpm.py000064400000002236150564745010015472 0ustar00import distutils.command.bdist_rpm as orig
import warnings

from setuptools import SetuptoolsDeprecationWarning


class bdist_rpm(orig.bdist_rpm):
    """
    Override the default bdist_rpm behavior to do the following:

    1. Run egg_info to ensure the name and version are properly calculated.
    2. Always run 'install' using --single-version-externally-managed to
       disable eggs in RPM distributions.
    """

    def run(self):
        warnings.warn(
            "bdist_rpm is deprecated and will be removed in a future "
            "version. Use bdist_wheel (wheel packages) instead.",
            SetuptoolsDeprecationWarning,
        )

        # ensure distro name is up-to-date
        self.run_command('egg_info')

        orig.bdist_rpm.run(self)

    def _make_spec_file(self):
        spec = orig.bdist_rpm._make_spec_file(self)
        spec = [
            line.replace(
                "setup.py install ",
                "setup.py install --single-version-externally-managed "
            ).replace(
                "%setup",
                "%setup -n %{name}-%{unmangled_version}"
            )
            for line in spec
        ]
        return spec
site-packages/setuptools/command/test.py000064400000017630150564745010014472 0ustar00import os
import operator
import sys
import contextlib
import itertools
import unittest
from distutils.errors import DistutilsError, DistutilsOptionError
from distutils import log
from unittest import TestLoader

from pkg_resources import (
    resource_listdir,
    resource_exists,
    normalize_path,
    working_set,
    evaluate_marker,
    add_activation_listener,
    require,
    EntryPoint,
)
from setuptools import Command
from setuptools.extern.more_itertools import unique_everseen


class ScanningLoader(TestLoader):
    def __init__(self):
        TestLoader.__init__(self)
        self._visited = set()

    def loadTestsFromModule(self, module, pattern=None):
        """Return a suite of all tests cases contained in the given module

        If the module is a package, load tests from all the modules in it.
        If the module has an ``additional_tests`` function, call it and add
        the return value to the tests.
        """
        if module in self._visited:
            return None
        self._visited.add(module)

        tests = []
        tests.append(TestLoader.loadTestsFromModule(self, module))

        if hasattr(module, "additional_tests"):
            tests.append(module.additional_tests())

        if hasattr(module, '__path__'):
            for file in resource_listdir(module.__name__, ''):
                if file.endswith('.py') and file != '__init__.py':
                    submodule = module.__name__ + '.' + file[:-3]
                else:
                    if resource_exists(module.__name__, file + '/__init__.py'):
                        submodule = module.__name__ + '.' + file
                    else:
                        continue
                tests.append(self.loadTestsFromName(submodule))

        if len(tests) != 1:
            return self.suiteClass(tests)
        else:
            return tests[0]  # don't create a nested suite for only one return


# adapted from jaraco.classes.properties:NonDataProperty
class NonDataProperty:
    def __init__(self, fget):
        self.fget = fget

    def __get__(self, obj, objtype=None):
        if obj is None:
            return self
        return self.fget(obj)


class test(Command):
    """Command to run unit tests after in-place build"""

    description = "run unit tests after in-place build (deprecated)"

    user_options = [
        ('test-module=', 'm', "Run 'test_suite' in specified module"),
        (
            'test-suite=',
            's',
            "Run single test, case or suite (e.g. 'module.test_suite')",
        ),
        ('test-runner=', 'r', "Test runner to use"),
    ]

    def initialize_options(self):
        self.test_suite = None
        self.test_module = None
        self.test_loader = None
        self.test_runner = None

    def finalize_options(self):

        if self.test_suite and self.test_module:
            msg = "You may specify a module or a suite, but not both"
            raise DistutilsOptionError(msg)

        if self.test_suite is None:
            if self.test_module is None:
                self.test_suite = self.distribution.test_suite
            else:
                self.test_suite = self.test_module + ".test_suite"

        if self.test_loader is None:
            self.test_loader = getattr(self.distribution, 'test_loader', None)
        if self.test_loader is None:
            self.test_loader = "setuptools.command.test:ScanningLoader"
        if self.test_runner is None:
            self.test_runner = getattr(self.distribution, 'test_runner', None)

    @NonDataProperty
    def test_args(self):
        return list(self._test_args())

    def _test_args(self):
        if not self.test_suite and sys.version_info >= (2, 7):
            yield 'discover'
        if self.verbose:
            yield '--verbose'
        if self.test_suite:
            yield self.test_suite

    def with_project_on_sys_path(self, func):
        """
        Backward compatibility for project_on_sys_path context.
        """
        with self.project_on_sys_path():
            func()

    @contextlib.contextmanager
    def project_on_sys_path(self, include_dists=[]):
        self.run_command('egg_info')

        # Build extensions in-place
        self.reinitialize_command('build_ext', inplace=1)
        self.run_command('build_ext')

        ei_cmd = self.get_finalized_command("egg_info")

        old_path = sys.path[:]
        old_modules = sys.modules.copy()

        try:
            project_path = normalize_path(ei_cmd.egg_base)
            sys.path.insert(0, project_path)
            working_set.__init__()
            add_activation_listener(lambda dist: dist.activate())
            require('%s==%s' % (ei_cmd.egg_name, ei_cmd.egg_version))
            with self.paths_on_pythonpath([project_path]):
                yield
        finally:
            sys.path[:] = old_path
            sys.modules.clear()
            sys.modules.update(old_modules)
            working_set.__init__()

    @staticmethod
    @contextlib.contextmanager
    def paths_on_pythonpath(paths):
        """
        Add the indicated paths to the head of the PYTHONPATH environment
        variable so that subprocesses will also see the packages at
        these paths.

        Do this in a context that restores the value on exit.
        """
        nothing = object()
        orig_pythonpath = os.environ.get('PYTHONPATH', nothing)
        current_pythonpath = os.environ.get('PYTHONPATH', '')
        try:
            prefix = os.pathsep.join(unique_everseen(paths))
            to_join = filter(None, [prefix, current_pythonpath])
            new_path = os.pathsep.join(to_join)
            if new_path:
                os.environ['PYTHONPATH'] = new_path
            yield
        finally:
            if orig_pythonpath is nothing:
                os.environ.pop('PYTHONPATH', None)
            else:
                os.environ['PYTHONPATH'] = orig_pythonpath

    @staticmethod
    def install_dists(dist):
        """
        Install the requirements indicated by self.distribution and
        return an iterable of the dists that were built.
        """
        ir_d = dist.fetch_build_eggs(dist.install_requires)
        tr_d = dist.fetch_build_eggs(dist.tests_require or [])
        er_d = dist.fetch_build_eggs(
            v
            for k, v in dist.extras_require.items()
            if k.startswith(':') and evaluate_marker(k[1:])
        )
        return itertools.chain(ir_d, tr_d, er_d)

    def run(self):
        self.announce(
            "WARNING: Testing via this command is deprecated and will be "
            "removed in a future version. Users looking for a generic test "
            "entry point independent of test runner are encouraged to use "
            "tox.",
            log.WARN,
        )

        installed_dists = self.install_dists(self.distribution)

        cmd = ' '.join(self._argv)
        if self.dry_run:
            self.announce('skipping "%s" (dry run)' % cmd)
            return

        self.announce('running "%s"' % cmd)

        paths = map(operator.attrgetter('location'), installed_dists)
        with self.paths_on_pythonpath(paths):
            with self.project_on_sys_path():
                self.run_tests()

    def run_tests(self):
        test = unittest.main(
            None,
            None,
            self._argv,
            testLoader=self._resolve_as_ep(self.test_loader),
            testRunner=self._resolve_as_ep(self.test_runner),
            exit=False,
        )
        if not test.result.wasSuccessful():
            msg = 'Test failed: %s' % test.result
            self.announce(msg, log.ERROR)
            raise DistutilsError(msg)

    @property
    def _argv(self):
        return ['unittest'] + self.test_args

    @staticmethod
    def _resolve_as_ep(val):
        """
        Load the indicated attribute value, called, as a as if it were
        specified as an entry point.
        """
        if val is None:
            return
        parsed = EntryPoint.parse("x=" + val)
        return parsed.resolve()()
site-packages/setuptools/command/install_egg_info.py000064400000004233150564745010017011 0ustar00from distutils import log, dir_util
import os

from setuptools import Command
from setuptools import namespaces
from setuptools.archive_util import unpack_archive
import pkg_resources


class install_egg_info(namespaces.Installer, Command):
    """Install an .egg-info directory for the package"""

    description = "Install an .egg-info directory for the package"

    user_options = [
        ('install-dir=', 'd', "directory to install to"),
    ]

    def initialize_options(self):
        self.install_dir = None

    def finalize_options(self):
        self.set_undefined_options('install_lib',
                                   ('install_dir', 'install_dir'))
        ei_cmd = self.get_finalized_command("egg_info")
        basename = pkg_resources.Distribution(
            None, None, ei_cmd.egg_name, ei_cmd.egg_version
        ).egg_name() + '.egg-info'
        self.source = ei_cmd.egg_info
        self.target = os.path.join(self.install_dir, basename)
        self.outputs = []

    def run(self):
        self.run_command('egg_info')
        if os.path.isdir(self.target) and not os.path.islink(self.target):
            dir_util.remove_tree(self.target, dry_run=self.dry_run)
        elif os.path.exists(self.target):
            self.execute(os.unlink, (self.target,), "Removing " + self.target)
        if not self.dry_run:
            pkg_resources.ensure_directory(self.target)
        self.execute(
            self.copytree, (), "Copying %s to %s" % (self.source, self.target)
        )
        self.install_namespaces()

    def get_outputs(self):
        return self.outputs

    def copytree(self):
        # Copy the .egg-info tree to site-packages
        def skimmer(src, dst):
            # filter out source-control directories; note that 'src' is always
            # a '/'-separated path, regardless of platform.  'dst' is a
            # platform-specific path.
            for skip in '.svn/', 'CVS/':
                if src.startswith(skip) or '/' + skip in src:
                    return None
            self.outputs.append(dst)
            log.debug("Copying %s to %s", src, dst)
            return dst

        unpack_archive(self.source, self.target, skimmer)
site-packages/setuptools/command/dist_info.py000064400000001700150564745010015460 0ustar00"""
Create a dist_info directory
As defined in the wheel specification
"""

import os

from distutils.core import Command
from distutils import log


class dist_info(Command):

    description = 'create a .dist-info directory'

    user_options = [
        ('egg-base=', 'e', "directory containing .egg-info directories"
                           " (default: top of the source tree)"),
    ]

    def initialize_options(self):
        self.egg_base = None

    def finalize_options(self):
        pass

    def run(self):
        egg_info = self.get_finalized_command('egg_info')
        egg_info.egg_base = self.egg_base
        egg_info.finalize_options()
        egg_info.run()
        dist_info_dir = egg_info.egg_info[:-len('.egg-info')] + '.dist-info'
        log.info("creating '{}'".format(os.path.abspath(dist_info_dir)))

        bdist_wheel = self.get_finalized_command('bdist_wheel')
        bdist_wheel.egg2dist(egg_info.egg_info, dist_info_dir)
site-packages/setuptools/command/setopt.py000064400000011736150564745010015032 0ustar00from distutils.util import convert_path
from distutils import log
from distutils.errors import DistutilsOptionError
import distutils
import os
import configparser

from setuptools import Command

__all__ = ['config_file', 'edit_config', 'option_base', 'setopt']


def config_file(kind="local"):
    """Get the filename of the distutils, local, global, or per-user config

    `kind` must be one of "local", "global", or "user"
    """
    if kind == 'local':
        return 'setup.cfg'
    if kind == 'global':
        return os.path.join(
            os.path.dirname(distutils.__file__), 'distutils.cfg'
        )
    if kind == 'user':
        dot = os.name == 'posix' and '.' or ''
        return os.path.expanduser(convert_path("~/%spydistutils.cfg" % dot))
    raise ValueError(
        "config_file() type must be 'local', 'global', or 'user'", kind
    )


def edit_config(filename, settings, dry_run=False):
    """Edit a configuration file to include `settings`

    `settings` is a dictionary of dictionaries or ``None`` values, keyed by
    command/section name.  A ``None`` value means to delete the entire section,
    while a dictionary lists settings to be changed or deleted in that section.
    A setting of ``None`` means to delete that setting.
    """
    log.debug("Reading configuration from %s", filename)
    opts = configparser.RawConfigParser()
    opts.optionxform = lambda x: x
    opts.read([filename])
    for section, options in settings.items():
        if options is None:
            log.info("Deleting section [%s] from %s", section, filename)
            opts.remove_section(section)
        else:
            if not opts.has_section(section):
                log.debug("Adding new section [%s] to %s", section, filename)
                opts.add_section(section)
            for option, value in options.items():
                if value is None:
                    log.debug(
                        "Deleting %s.%s from %s",
                        section, option, filename
                    )
                    opts.remove_option(section, option)
                    if not opts.options(section):
                        log.info("Deleting empty [%s] section from %s",
                                 section, filename)
                        opts.remove_section(section)
                else:
                    log.debug(
                        "Setting %s.%s to %r in %s",
                        section, option, value, filename
                    )
                    opts.set(section, option, value)

    log.info("Writing %s", filename)
    if not dry_run:
        with open(filename, 'w') as f:
            opts.write(f)


class option_base(Command):
    """Abstract base class for commands that mess with config files"""

    user_options = [
        ('global-config', 'g',
         "save options to the site-wide distutils.cfg file"),
        ('user-config', 'u',
         "save options to the current user's pydistutils.cfg file"),
        ('filename=', 'f',
         "configuration file to use (default=setup.cfg)"),
    ]

    boolean_options = [
        'global-config', 'user-config',
    ]

    def initialize_options(self):
        self.global_config = None
        self.user_config = None
        self.filename = None

    def finalize_options(self):
        filenames = []
        if self.global_config:
            filenames.append(config_file('global'))
        if self.user_config:
            filenames.append(config_file('user'))
        if self.filename is not None:
            filenames.append(self.filename)
        if not filenames:
            filenames.append(config_file('local'))
        if len(filenames) > 1:
            raise DistutilsOptionError(
                "Must specify only one configuration file option",
                filenames
            )
        self.filename, = filenames


class setopt(option_base):
    """Save command-line options to a file"""

    description = "set an option in setup.cfg or another config file"

    user_options = [
        ('command=', 'c', 'command to set an option for'),
        ('option=', 'o', 'option to set'),
        ('set-value=', 's', 'value of the option'),
        ('remove', 'r', 'remove (unset) the value'),
    ] + option_base.user_options

    boolean_options = option_base.boolean_options + ['remove']

    def initialize_options(self):
        option_base.initialize_options(self)
        self.command = None
        self.option = None
        self.set_value = None
        self.remove = None

    def finalize_options(self):
        option_base.finalize_options(self)
        if self.command is None or self.option is None:
            raise DistutilsOptionError("Must specify --command *and* --option")
        if self.set_value is None and not self.remove:
            raise DistutilsOptionError("Must specify --set-value or --remove")

    def run(self):
        edit_config(
            self.filename, {
                self.command: {self.option.replace('-', '_'): self.set_value}
            },
            self.dry_run
        )
site-packages/setuptools/command/saveopts.py000064400000001222150564745010015345 0ustar00from setuptools.command.setopt import edit_config, option_base


class saveopts(option_base):
    """Save command-line options to a file"""

    description = "save supplied options to setup.cfg or other config file"

    def run(self):
        dist = self.distribution
        settings = {}

        for cmd in dist.command_options:

            if cmd == 'saveopts':
                continue  # don't save our own options!

            for opt, (src, val) in dist.get_option_dict(cmd).items():
                if src == "command line":
                    settings.setdefault(cmd, {})[opt] = val

        edit_config(self.filename, settings, self.dry_run)
site-packages/setuptools/command/register.py000064400000000724150564745010015333 0ustar00from distutils import log
import distutils.command.register as orig

from setuptools.errors import RemovedCommandError


class register(orig.register):
    """Formerly used to register packages on PyPI."""

    def run(self):
        msg = (
            "The register command has been removed, use twine to upload "
            + "instead (https://pypi.org/p/twine)"
        )

        self.announce("ERROR: " + msg, log.ERROR)

        raise RemovedCommandError(msg)
site-packages/setuptools/command/sdist.py000064400000014034150564745010014634 0ustar00from distutils import log
import distutils.command.sdist as orig
import os
import sys
import io
import contextlib

from .py36compat import sdist_add_defaults

import pkg_resources

_default_revctrl = list


def walk_revctrl(dirname=''):
    """Find all files under revision control"""
    for ep in pkg_resources.iter_entry_points('setuptools.file_finders'):
        for item in ep.load()(dirname):
            yield item


class sdist(sdist_add_defaults, orig.sdist):
    """Smart sdist that finds anything supported by revision control"""

    user_options = [
        ('formats=', None,
         "formats for source distribution (comma-separated list)"),
        ('keep-temp', 'k',
         "keep the distribution tree around after creating " +
         "archive file(s)"),
        ('dist-dir=', 'd',
         "directory to put the source distribution archive(s) in "
         "[default: dist]"),
        ('owner=', 'u',
         "Owner name used when creating a tar file [default: current user]"),
        ('group=', 'g',
         "Group name used when creating a tar file [default: current group]"),
    ]

    negative_opt = {}

    README_EXTENSIONS = ['', '.rst', '.txt', '.md']
    READMES = tuple('README{0}'.format(ext) for ext in README_EXTENSIONS)

    def run(self):
        self.run_command('egg_info')
        ei_cmd = self.get_finalized_command('egg_info')
        self.filelist = ei_cmd.filelist
        self.filelist.append(os.path.join(ei_cmd.egg_info, 'SOURCES.txt'))
        self.check_readme()

        # Run sub commands
        for cmd_name in self.get_sub_commands():
            self.run_command(cmd_name)

        self.make_distribution()

        dist_files = getattr(self.distribution, 'dist_files', [])
        for file in self.archive_files:
            data = ('sdist', '', file)
            if data not in dist_files:
                dist_files.append(data)

    def initialize_options(self):
        orig.sdist.initialize_options(self)

        self._default_to_gztar()

    def _default_to_gztar(self):
        # only needed on Python prior to 3.6.
        if sys.version_info >= (3, 6, 0, 'beta', 1):
            return
        self.formats = ['gztar']

    def make_distribution(self):
        """
        Workaround for #516
        """
        with self._remove_os_link():
            orig.sdist.make_distribution(self)

    @staticmethod
    @contextlib.contextmanager
    def _remove_os_link():
        """
        In a context, remove and restore os.link if it exists
        """

        class NoValue:
            pass

        orig_val = getattr(os, 'link', NoValue)
        try:
            del os.link
        except Exception:
            pass
        try:
            yield
        finally:
            if orig_val is not NoValue:
                setattr(os, 'link', orig_val)

    def _add_defaults_optional(self):
        super()._add_defaults_optional()
        if os.path.isfile('pyproject.toml'):
            self.filelist.append('pyproject.toml')

    def _add_defaults_python(self):
        """getting python files"""
        if self.distribution.has_pure_modules():
            build_py = self.get_finalized_command('build_py')
            self.filelist.extend(build_py.get_source_files())
            self._add_data_files(self._safe_data_files(build_py))

    def _safe_data_files(self, build_py):
        """
        Extracting data_files from build_py is known to cause
        infinite recursion errors when `include_package_data`
        is enabled, so suppress it in that case.
        """
        if self.distribution.include_package_data:
            return ()
        return build_py.data_files

    def _add_data_files(self, data_files):
        """
        Add data files as found in build_py.data_files.
        """
        self.filelist.extend(
            os.path.join(src_dir, name)
            for _, src_dir, _, filenames in data_files
            for name in filenames
        )

    def _add_defaults_data_files(self):
        try:
            super()._add_defaults_data_files()
        except TypeError:
            log.warn("data_files contains unexpected objects")

    def check_readme(self):
        for f in self.READMES:
            if os.path.exists(f):
                return
        else:
            self.warn(
                "standard file not found: should have one of " +
                ', '.join(self.READMES)
            )

    def make_release_tree(self, base_dir, files):
        orig.sdist.make_release_tree(self, base_dir, files)

        # Save any egg_info command line options used to create this sdist
        dest = os.path.join(base_dir, 'setup.cfg')
        if hasattr(os, 'link') and os.path.exists(dest):
            # unlink and re-copy, since it might be hard-linked, and
            # we don't want to change the source version
            os.unlink(dest)
            self.copy_file('setup.cfg', dest)

        self.get_finalized_command('egg_info').save_version_info(dest)

    def _manifest_is_not_generated(self):
        # check for special comment used in 2.7.1 and higher
        if not os.path.isfile(self.manifest):
            return False

        with io.open(self.manifest, 'rb') as fp:
            first_line = fp.readline()
        return (first_line !=
                '# file GENERATED by distutils, do NOT edit\n'.encode())

    def read_manifest(self):
        """Read the manifest file (named by 'self.manifest') and use it to
        fill in 'self.filelist', the list of files to include in the source
        distribution.
        """
        log.info("reading manifest file '%s'", self.manifest)
        manifest = open(self.manifest, 'rb')
        for line in manifest:
            # The manifest must contain UTF-8. See #303.
            try:
                line = line.decode('UTF-8')
            except UnicodeDecodeError:
                log.warn("%r not UTF-8 decodable -- skipping" % line)
                continue
            # ignore comments and blank lines
            line = line.strip()
            if line.startswith('#') or not line:
                continue
            self.filelist.append(line)
        manifest.close()
site-packages/setuptools/depends.py000064400000012542150564745010013514 0ustar00import sys
import marshal
import contextlib
import dis
from distutils.version import StrictVersion

from ._imp import find_module, PY_COMPILED, PY_FROZEN, PY_SOURCE
from . import _imp


__all__ = [
    'Require', 'find_module', 'get_module_constant', 'extract_constant'
]


class Require:
    """A prerequisite to building or installing a distribution"""

    def __init__(
            self, name, requested_version, module, homepage='',
            attribute=None, format=None):

        if format is None and requested_version is not None:
            format = StrictVersion

        if format is not None:
            requested_version = format(requested_version)
            if attribute is None:
                attribute = '__version__'

        self.__dict__.update(locals())
        del self.self

    def full_name(self):
        """Return full package/distribution name, w/version"""
        if self.requested_version is not None:
            return '%s-%s' % (self.name, self.requested_version)
        return self.name

    def version_ok(self, version):
        """Is 'version' sufficiently up-to-date?"""
        return self.attribute is None or self.format is None or \
            str(version) != "unknown" and version >= self.requested_version

    def get_version(self, paths=None, default="unknown"):
        """Get version number of installed module, 'None', or 'default'

        Search 'paths' for module.  If not found, return 'None'.  If found,
        return the extracted version attribute, or 'default' if no version
        attribute was specified, or the value cannot be determined without
        importing the module.  The version is formatted according to the
        requirement's version format (if any), unless it is 'None' or the
        supplied 'default'.
        """

        if self.attribute is None:
            try:
                f, p, i = find_module(self.module, paths)
                if f:
                    f.close()
                return default
            except ImportError:
                return None

        v = get_module_constant(self.module, self.attribute, default, paths)

        if v is not None and v is not default and self.format is not None:
            return self.format(v)

        return v

    def is_present(self, paths=None):
        """Return true if dependency is present on 'paths'"""
        return self.get_version(paths) is not None

    def is_current(self, paths=None):
        """Return true if dependency is present and up-to-date on 'paths'"""
        version = self.get_version(paths)
        if version is None:
            return False
        return self.version_ok(version)


def maybe_close(f):
    @contextlib.contextmanager
    def empty():
        yield
        return
    if not f:
        return empty()

    return contextlib.closing(f)


def get_module_constant(module, symbol, default=-1, paths=None):
    """Find 'module' by searching 'paths', and extract 'symbol'

    Return 'None' if 'module' does not exist on 'paths', or it does not define
    'symbol'.  If the module defines 'symbol' as a constant, return the
    constant.  Otherwise, return 'default'."""

    try:
        f, path, (suffix, mode, kind) = info = find_module(module, paths)
    except ImportError:
        # Module doesn't exist
        return None

    with maybe_close(f):
        if kind == PY_COMPILED:
            f.read(8)  # skip magic & date
            code = marshal.load(f)
        elif kind == PY_FROZEN:
            code = _imp.get_frozen_object(module, paths)
        elif kind == PY_SOURCE:
            code = compile(f.read(), path, 'exec')
        else:
            # Not something we can parse; we'll have to import it.  :(
            imported = _imp.get_module(module, paths, info)
            return getattr(imported, symbol, None)

    return extract_constant(code, symbol, default)


def extract_constant(code, symbol, default=-1):
    """Extract the constant value of 'symbol' from 'code'

    If the name 'symbol' is bound to a constant value by the Python code
    object 'code', return that value.  If 'symbol' is bound to an expression,
    return 'default'.  Otherwise, return 'None'.

    Return value is based on the first assignment to 'symbol'.  'symbol' must
    be a global, or at least a non-"fast" local in the code block.  That is,
    only 'STORE_NAME' and 'STORE_GLOBAL' opcodes are checked, and 'symbol'
    must be present in 'code.co_names'.
    """
    if symbol not in code.co_names:
        # name's not there, can't possibly be an assignment
        return None

    name_idx = list(code.co_names).index(symbol)

    STORE_NAME = 90
    STORE_GLOBAL = 97
    LOAD_CONST = 100

    const = default

    for byte_code in dis.Bytecode(code):
        op = byte_code.opcode
        arg = byte_code.arg

        if op == LOAD_CONST:
            const = code.co_consts[arg]
        elif arg == name_idx and (op == STORE_NAME or op == STORE_GLOBAL):
            return const
        else:
            const = default


def _update_globals():
    """
    Patch the globals to remove the objects not available on some platforms.

    XXX it'd be better to test assertions about bytecode instead.
    """

    if not sys.platform.startswith('java') and sys.platform != 'cli':
        return
    incompatible = 'extract_constant', 'get_module_constant'
    for name in incompatible:
        del globals()[name]
        __all__.remove(name)


_update_globals()
site-packages/setuptools/wheel.py000064400000020140150564745010013167 0ustar00"""Wheels support."""

from distutils.util import get_platform
from distutils import log
import email
import itertools
import os
import posixpath
import re
import zipfile

import pkg_resources
import setuptools
from pkg_resources import parse_version
from setuptools.extern.packaging.tags import sys_tags
from setuptools.extern.packaging.utils import canonicalize_name
from setuptools.command.egg_info import write_requirements


WHEEL_NAME = re.compile(
    r"""^(?P<project_name>.+?)-(?P<version>\d.*?)
    ((-(?P<build>\d.*?))?-(?P<py_version>.+?)-(?P<abi>.+?)-(?P<platform>.+?)
    )\.whl$""",
    re.VERBOSE).match

NAMESPACE_PACKAGE_INIT = \
    "__import__('pkg_resources').declare_namespace(__name__)\n"


def unpack(src_dir, dst_dir):
    '''Move everything under `src_dir` to `dst_dir`, and delete the former.'''
    for dirpath, dirnames, filenames in os.walk(src_dir):
        subdir = os.path.relpath(dirpath, src_dir)
        for f in filenames:
            src = os.path.join(dirpath, f)
            dst = os.path.join(dst_dir, subdir, f)
            os.renames(src, dst)
        for n, d in reversed(list(enumerate(dirnames))):
            src = os.path.join(dirpath, d)
            dst = os.path.join(dst_dir, subdir, d)
            if not os.path.exists(dst):
                # Directory does not exist in destination,
                # rename it and prune it from os.walk list.
                os.renames(src, dst)
                del dirnames[n]
    # Cleanup.
    for dirpath, dirnames, filenames in os.walk(src_dir, topdown=True):
        assert not filenames
        os.rmdir(dirpath)


class Wheel:

    def __init__(self, filename):
        match = WHEEL_NAME(os.path.basename(filename))
        if match is None:
            raise ValueError('invalid wheel name: %r' % filename)
        self.filename = filename
        for k, v in match.groupdict().items():
            setattr(self, k, v)

    def tags(self):
        '''List tags (py_version, abi, platform) supported by this wheel.'''
        return itertools.product(
            self.py_version.split('.'),
            self.abi.split('.'),
            self.platform.split('.'),
        )

    def is_compatible(self):
        '''Is the wheel is compatible with the current platform?'''
        supported_tags = set(
            (t.interpreter, t.abi, t.platform) for t in sys_tags())
        return next((True for t in self.tags() if t in supported_tags), False)

    def egg_name(self):
        return pkg_resources.Distribution(
            project_name=self.project_name, version=self.version,
            platform=(None if self.platform == 'any' else get_platform()),
        ).egg_name() + '.egg'

    def get_dist_info(self, zf):
        # find the correct name of the .dist-info dir in the wheel file
        for member in zf.namelist():
            dirname = posixpath.dirname(member)
            if (dirname.endswith('.dist-info') and
                    canonicalize_name(dirname).startswith(
                        canonicalize_name(self.project_name))):
                return dirname
        raise ValueError("unsupported wheel format. .dist-info not found")

    def install_as_egg(self, destination_eggdir):
        '''Install wheel as an egg directory.'''
        with zipfile.ZipFile(self.filename) as zf:
            self._install_as_egg(destination_eggdir, zf)

    def _install_as_egg(self, destination_eggdir, zf):
        dist_basename = '%s-%s' % (self.project_name, self.version)
        dist_info = self.get_dist_info(zf)
        dist_data = '%s.data' % dist_basename
        egg_info = os.path.join(destination_eggdir, 'EGG-INFO')

        self._convert_metadata(zf, destination_eggdir, dist_info, egg_info)
        self._move_data_entries(destination_eggdir, dist_data)
        self._fix_namespace_packages(egg_info, destination_eggdir)

    @staticmethod
    def _convert_metadata(zf, destination_eggdir, dist_info, egg_info):
        def get_metadata(name):
            with zf.open(posixpath.join(dist_info, name)) as fp:
                value = fp.read().decode('utf-8')
                return email.parser.Parser().parsestr(value)

        wheel_metadata = get_metadata('WHEEL')
        # Check wheel format version is supported.
        wheel_version = parse_version(wheel_metadata.get('Wheel-Version'))
        wheel_v1 = (
            parse_version('1.0') <= wheel_version < parse_version('2.0dev0')
        )
        if not wheel_v1:
            raise ValueError(
                'unsupported wheel format version: %s' % wheel_version)
        # Extract to target directory.
        os.mkdir(destination_eggdir)
        zf.extractall(destination_eggdir)
        # Convert metadata.
        dist_info = os.path.join(destination_eggdir, dist_info)
        dist = pkg_resources.Distribution.from_location(
            destination_eggdir, dist_info,
            metadata=pkg_resources.PathMetadata(destination_eggdir, dist_info),
        )

        # Note: Evaluate and strip markers now,
        # as it's difficult to convert back from the syntax:
        # foobar; "linux" in sys_platform and extra == 'test'
        def raw_req(req):
            req.marker = None
            return str(req)
        install_requires = list(sorted(map(raw_req, dist.requires())))
        extras_require = {
            extra: sorted(
                req
                for req in map(raw_req, dist.requires((extra,)))
                if req not in install_requires
            )
            for extra in dist.extras
        }
        os.rename(dist_info, egg_info)
        os.rename(
            os.path.join(egg_info, 'METADATA'),
            os.path.join(egg_info, 'PKG-INFO'),
        )
        setup_dist = setuptools.Distribution(
            attrs=dict(
                install_requires=install_requires,
                extras_require=extras_require,
            ),
        )
        # Temporarily disable info traces.
        log_threshold = log._global_log.threshold
        log.set_threshold(log.WARN)
        try:
            write_requirements(
                setup_dist.get_command_obj('egg_info'),
                None,
                os.path.join(egg_info, 'requires.txt'),
            )
        finally:
            log.set_threshold(log_threshold)

    @staticmethod
    def _move_data_entries(destination_eggdir, dist_data):
        """Move data entries to their correct location."""
        dist_data = os.path.join(destination_eggdir, dist_data)
        dist_data_scripts = os.path.join(dist_data, 'scripts')
        if os.path.exists(dist_data_scripts):
            egg_info_scripts = os.path.join(
                destination_eggdir, 'EGG-INFO', 'scripts')
            os.mkdir(egg_info_scripts)
            for entry in os.listdir(dist_data_scripts):
                # Remove bytecode, as it's not properly handled
                # during easy_install scripts install phase.
                if entry.endswith('.pyc'):
                    os.unlink(os.path.join(dist_data_scripts, entry))
                else:
                    os.rename(
                        os.path.join(dist_data_scripts, entry),
                        os.path.join(egg_info_scripts, entry),
                    )
            os.rmdir(dist_data_scripts)
        for subdir in filter(os.path.exists, (
            os.path.join(dist_data, d)
            for d in ('data', 'headers', 'purelib', 'platlib')
        )):
            unpack(subdir, destination_eggdir)
        if os.path.exists(dist_data):
            os.rmdir(dist_data)

    @staticmethod
    def _fix_namespace_packages(egg_info, destination_eggdir):
        namespace_packages = os.path.join(
            egg_info, 'namespace_packages.txt')
        if os.path.exists(namespace_packages):
            with open(namespace_packages) as fp:
                namespace_packages = fp.read().split()
            for mod in namespace_packages:
                mod_dir = os.path.join(destination_eggdir, *mod.split('.'))
                mod_init = os.path.join(mod_dir, '__init__.py')
                if not os.path.exists(mod_dir):
                    os.mkdir(mod_dir)
                if not os.path.exists(mod_init):
                    with open(mod_init, 'w') as fp:
                        fp.write(NAMESPACE_PACKAGE_INIT)
site-packages/setuptools/msvc.py000064400000142601150564745010013042 0ustar00"""
Improved support for Microsoft Visual C++ compilers.

Known supported compilers:
--------------------------
Microsoft Visual C++ 9.0:
    Microsoft Visual C++ Compiler for Python 2.7 (x86, amd64)
    Microsoft Windows SDK 6.1 (x86, x64, ia64)
    Microsoft Windows SDK 7.0 (x86, x64, ia64)

Microsoft Visual C++ 10.0:
    Microsoft Windows SDK 7.1 (x86, x64, ia64)

Microsoft Visual C++ 14.X:
    Microsoft Visual C++ Build Tools 2015 (x86, x64, arm)
    Microsoft Visual Studio Build Tools 2017 (x86, x64, arm, arm64)
    Microsoft Visual Studio Build Tools 2019 (x86, x64, arm, arm64)

This may also support compilers shipped with compatible Visual Studio versions.
"""

import json
from io import open
from os import listdir, pathsep
from os.path import join, isfile, isdir, dirname
import sys
import contextlib
import platform
import itertools
import subprocess
import distutils.errors
from setuptools.extern.packaging.version import LegacyVersion
from setuptools.extern.more_itertools import unique_everseen

from .monkey import get_unpatched

if platform.system() == 'Windows':
    import winreg
    from os import environ
else:
    # Mock winreg and environ so the module can be imported on this platform.

    class winreg:
        HKEY_USERS = None
        HKEY_CURRENT_USER = None
        HKEY_LOCAL_MACHINE = None
        HKEY_CLASSES_ROOT = None

    environ = dict()

_msvc9_suppress_errors = (
    # msvc9compiler isn't available on some platforms
    ImportError,

    # msvc9compiler raises DistutilsPlatformError in some
    # environments. See #1118.
    distutils.errors.DistutilsPlatformError,
)

try:
    from distutils.msvc9compiler import Reg
except _msvc9_suppress_errors:
    pass


def msvc9_find_vcvarsall(version):
    """
    Patched "distutils.msvc9compiler.find_vcvarsall" to use the standalone
    compiler build for Python
    (VCForPython / Microsoft Visual C++ Compiler for Python 2.7).

    Fall back to original behavior when the standalone compiler is not
    available.

    Redirect the path of "vcvarsall.bat".

    Parameters
    ----------
    version: float
        Required Microsoft Visual C++ version.

    Return
    ------
    str
        vcvarsall.bat path
    """
    vc_base = r'Software\%sMicrosoft\DevDiv\VCForPython\%0.1f'
    key = vc_base % ('', version)
    try:
        # Per-user installs register the compiler path here
        productdir = Reg.get_value(key, "installdir")
    except KeyError:
        try:
            # All-user installs on a 64-bit system register here
            key = vc_base % ('Wow6432Node\\', version)
            productdir = Reg.get_value(key, "installdir")
        except KeyError:
            productdir = None

    if productdir:
        vcvarsall = join(productdir, "vcvarsall.bat")
        if isfile(vcvarsall):
            return vcvarsall

    return get_unpatched(msvc9_find_vcvarsall)(version)


def msvc9_query_vcvarsall(ver, arch='x86', *args, **kwargs):
    """
    Patched "distutils.msvc9compiler.query_vcvarsall" for support extra
    Microsoft Visual C++ 9.0 and 10.0 compilers.

    Set environment without use of "vcvarsall.bat".

    Parameters
    ----------
    ver: float
        Required Microsoft Visual C++ version.
    arch: str
        Target architecture.

    Return
    ------
    dict
        environment
    """
    # Try to get environment from vcvarsall.bat (Classical way)
    try:
        orig = get_unpatched(msvc9_query_vcvarsall)
        return orig(ver, arch, *args, **kwargs)
    except distutils.errors.DistutilsPlatformError:
        # Pass error if Vcvarsall.bat is missing
        pass
    except ValueError:
        # Pass error if environment not set after executing vcvarsall.bat
        pass

    # If error, try to set environment directly
    try:
        return EnvironmentInfo(arch, ver).return_env()
    except distutils.errors.DistutilsPlatformError as exc:
        _augment_exception(exc, ver, arch)
        raise


def _msvc14_find_vc2015():
    """Python 3.8 "distutils/_msvccompiler.py" backport"""
    try:
        key = winreg.OpenKey(
            winreg.HKEY_LOCAL_MACHINE,
            r"Software\Microsoft\VisualStudio\SxS\VC7",
            0,
            winreg.KEY_READ | winreg.KEY_WOW64_32KEY
        )
    except OSError:
        return None, None

    best_version = 0
    best_dir = None
    with key:
        for i in itertools.count():
            try:
                v, vc_dir, vt = winreg.EnumValue(key, i)
            except OSError:
                break
            if v and vt == winreg.REG_SZ and isdir(vc_dir):
                try:
                    version = int(float(v))
                except (ValueError, TypeError):
                    continue
                if version >= 14 and version > best_version:
                    best_version, best_dir = version, vc_dir
    return best_version, best_dir


def _msvc14_find_vc2017():
    """Python 3.8 "distutils/_msvccompiler.py" backport

    Returns "15, path" based on the result of invoking vswhere.exe
    If no install is found, returns "None, None"

    The version is returned to avoid unnecessarily changing the function
    result. It may be ignored when the path is not None.

    If vswhere.exe is not available, by definition, VS 2017 is not
    installed.
    """
    root = environ.get("ProgramFiles(x86)") or environ.get("ProgramFiles")
    if not root:
        return None, None

    try:
        path = subprocess.check_output([
            join(root, "Microsoft Visual Studio", "Installer", "vswhere.exe"),
            "-latest",
            "-prerelease",
            "-requiresAny",
            "-requires", "Microsoft.VisualStudio.Component.VC.Tools.x86.x64",
            "-requires", "Microsoft.VisualStudio.Workload.WDExpress",
            "-property", "installationPath",
            "-products", "*",
        ]).decode(encoding="mbcs", errors="strict").strip()
    except (subprocess.CalledProcessError, OSError, UnicodeDecodeError):
        return None, None

    path = join(path, "VC", "Auxiliary", "Build")
    if isdir(path):
        return 15, path

    return None, None


PLAT_SPEC_TO_RUNTIME = {
    'x86': 'x86',
    'x86_amd64': 'x64',
    'x86_arm': 'arm',
    'x86_arm64': 'arm64'
}


def _msvc14_find_vcvarsall(plat_spec):
    """Python 3.8 "distutils/_msvccompiler.py" backport"""
    _, best_dir = _msvc14_find_vc2017()
    vcruntime = None

    if plat_spec in PLAT_SPEC_TO_RUNTIME:
        vcruntime_plat = PLAT_SPEC_TO_RUNTIME[plat_spec]
    else:
        vcruntime_plat = 'x64' if 'amd64' in plat_spec else 'x86'

    if best_dir:
        vcredist = join(best_dir, "..", "..", "redist", "MSVC", "**",
                        vcruntime_plat, "Microsoft.VC14*.CRT",
                        "vcruntime140.dll")
        try:
            import glob
            vcruntime = glob.glob(vcredist, recursive=True)[-1]
        except (ImportError, OSError, LookupError):
            vcruntime = None

    if not best_dir:
        best_version, best_dir = _msvc14_find_vc2015()
        if best_version:
            vcruntime = join(best_dir, 'redist', vcruntime_plat,
                             "Microsoft.VC140.CRT", "vcruntime140.dll")

    if not best_dir:
        return None, None

    vcvarsall = join(best_dir, "vcvarsall.bat")
    if not isfile(vcvarsall):
        return None, None

    if not vcruntime or not isfile(vcruntime):
        vcruntime = None

    return vcvarsall, vcruntime


def _msvc14_get_vc_env(plat_spec):
    """Python 3.8 "distutils/_msvccompiler.py" backport"""
    if "DISTUTILS_USE_SDK" in environ:
        return {
            key.lower(): value
            for key, value in environ.items()
        }

    vcvarsall, vcruntime = _msvc14_find_vcvarsall(plat_spec)
    if not vcvarsall:
        raise distutils.errors.DistutilsPlatformError(
            "Unable to find vcvarsall.bat"
        )

    try:
        out = subprocess.check_output(
            'cmd /u /c "{}" {} && set'.format(vcvarsall, plat_spec),
            stderr=subprocess.STDOUT,
        ).decode('utf-16le', errors='replace')
    except subprocess.CalledProcessError as exc:
        raise distutils.errors.DistutilsPlatformError(
            "Error executing {}".format(exc.cmd)
        ) from exc

    env = {
        key.lower(): value
        for key, _, value in
        (line.partition('=') for line in out.splitlines())
        if key and value
    }

    if vcruntime:
        env['py_vcruntime_redist'] = vcruntime
    return env


def msvc14_get_vc_env(plat_spec):
    """
    Patched "distutils._msvccompiler._get_vc_env" for support extra
    Microsoft Visual C++ 14.X compilers.

    Set environment without use of "vcvarsall.bat".

    Parameters
    ----------
    plat_spec: str
        Target architecture.

    Return
    ------
    dict
        environment
    """

    # Always use backport from CPython 3.8
    try:
        return _msvc14_get_vc_env(plat_spec)
    except distutils.errors.DistutilsPlatformError as exc:
        _augment_exception(exc, 14.0)
        raise


def msvc14_gen_lib_options(*args, **kwargs):
    """
    Patched "distutils._msvccompiler.gen_lib_options" for fix
    compatibility between "numpy.distutils" and "distutils._msvccompiler"
    (for Numpy < 1.11.2)
    """
    if "numpy.distutils" in sys.modules:
        import numpy as np
        if LegacyVersion(np.__version__) < LegacyVersion('1.11.2'):
            return np.distutils.ccompiler.gen_lib_options(*args, **kwargs)
    return get_unpatched(msvc14_gen_lib_options)(*args, **kwargs)


def _augment_exception(exc, version, arch=''):
    """
    Add details to the exception message to help guide the user
    as to what action will resolve it.
    """
    # Error if MSVC++ directory not found or environment not set
    message = exc.args[0]

    if "vcvarsall" in message.lower() or "visual c" in message.lower():
        # Special error message if MSVC++ not installed
        tmpl = 'Microsoft Visual C++ {version:0.1f} or greater is required.'
        message = tmpl.format(**locals())
        msdownload = 'www.microsoft.com/download/details.aspx?id=%d'
        if version == 9.0:
            if arch.lower().find('ia64') > -1:
                # For VC++ 9.0, if IA64 support is needed, redirect user
                # to Windows SDK 7.0.
                # Note: No download link available from Microsoft.
                message += ' Get it with "Microsoft Windows SDK 7.0"'
            else:
                # For VC++ 9.0 redirect user to Vc++ for Python 2.7 :
                # This redirection link is maintained by Microsoft.
                # Contact vspython@microsoft.com if it needs updating.
                message += ' Get it from http://aka.ms/vcpython27'
        elif version == 10.0:
            # For VC++ 10.0 Redirect user to Windows SDK 7.1
            message += ' Get it with "Microsoft Windows SDK 7.1": '
            message += msdownload % 8279
        elif version >= 14.0:
            # For VC++ 14.X Redirect user to latest Visual C++ Build Tools
            message += (' Get it with "Microsoft C++ Build Tools": '
                        r'https://visualstudio.microsoft.com'
                        r'/visual-cpp-build-tools/')

    exc.args = (message, )


class PlatformInfo:
    """
    Current and Target Architectures information.

    Parameters
    ----------
    arch: str
        Target architecture.
    """
    current_cpu = environ.get('processor_architecture', '').lower()

    def __init__(self, arch):
        self.arch = arch.lower().replace('x64', 'amd64')

    @property
    def target_cpu(self):
        """
        Return Target CPU architecture.

        Return
        ------
        str
            Target CPU
        """
        return self.arch[self.arch.find('_') + 1:]

    def target_is_x86(self):
        """
        Return True if target CPU is x86 32 bits..

        Return
        ------
        bool
            CPU is x86 32 bits
        """
        return self.target_cpu == 'x86'

    def current_is_x86(self):
        """
        Return True if current CPU is x86 32 bits..

        Return
        ------
        bool
            CPU is x86 32 bits
        """
        return self.current_cpu == 'x86'

    def current_dir(self, hidex86=False, x64=False):
        """
        Current platform specific subfolder.

        Parameters
        ----------
        hidex86: bool
            return '' and not '\x86' if architecture is x86.
        x64: bool
            return '\x64' and not '\amd64' if architecture is amd64.

        Return
        ------
        str
            subfolder: '\target', or '' (see hidex86 parameter)
        """
        return (
            '' if (self.current_cpu == 'x86' and hidex86) else
            r'\x64' if (self.current_cpu == 'amd64' and x64) else
            r'\%s' % self.current_cpu
        )

    def target_dir(self, hidex86=False, x64=False):
        r"""
        Target platform specific subfolder.

        Parameters
        ----------
        hidex86: bool
            return '' and not '\x86' if architecture is x86.
        x64: bool
            return '\x64' and not '\amd64' if architecture is amd64.

        Return
        ------
        str
            subfolder: '\current', or '' (see hidex86 parameter)
        """
        return (
            '' if (self.target_cpu == 'x86' and hidex86) else
            r'\x64' if (self.target_cpu == 'amd64' and x64) else
            r'\%s' % self.target_cpu
        )

    def cross_dir(self, forcex86=False):
        r"""
        Cross platform specific subfolder.

        Parameters
        ----------
        forcex86: bool
            Use 'x86' as current architecture even if current architecture is
            not x86.

        Return
        ------
        str
            subfolder: '' if target architecture is current architecture,
            '\current_target' if not.
        """
        current = 'x86' if forcex86 else self.current_cpu
        return (
            '' if self.target_cpu == current else
            self.target_dir().replace('\\', '\\%s_' % current)
        )


class RegistryInfo:
    """
    Microsoft Visual Studio related registry information.

    Parameters
    ----------
    platform_info: PlatformInfo
        "PlatformInfo" instance.
    """
    HKEYS = (winreg.HKEY_USERS,
             winreg.HKEY_CURRENT_USER,
             winreg.HKEY_LOCAL_MACHINE,
             winreg.HKEY_CLASSES_ROOT)

    def __init__(self, platform_info):
        self.pi = platform_info

    @property
    def visualstudio(self):
        """
        Microsoft Visual Studio root registry key.

        Return
        ------
        str
            Registry key
        """
        return 'VisualStudio'

    @property
    def sxs(self):
        """
        Microsoft Visual Studio SxS registry key.

        Return
        ------
        str
            Registry key
        """
        return join(self.visualstudio, 'SxS')

    @property
    def vc(self):
        """
        Microsoft Visual C++ VC7 registry key.

        Return
        ------
        str
            Registry key
        """
        return join(self.sxs, 'VC7')

    @property
    def vs(self):
        """
        Microsoft Visual Studio VS7 registry key.

        Return
        ------
        str
            Registry key
        """
        return join(self.sxs, 'VS7')

    @property
    def vc_for_python(self):
        """
        Microsoft Visual C++ for Python registry key.

        Return
        ------
        str
            Registry key
        """
        return r'DevDiv\VCForPython'

    @property
    def microsoft_sdk(self):
        """
        Microsoft SDK registry key.

        Return
        ------
        str
            Registry key
        """
        return 'Microsoft SDKs'

    @property
    def windows_sdk(self):
        """
        Microsoft Windows/Platform SDK registry key.

        Return
        ------
        str
            Registry key
        """
        return join(self.microsoft_sdk, 'Windows')

    @property
    def netfx_sdk(self):
        """
        Microsoft .NET Framework SDK registry key.

        Return
        ------
        str
            Registry key
        """
        return join(self.microsoft_sdk, 'NETFXSDK')

    @property
    def windows_kits_roots(self):
        """
        Microsoft Windows Kits Roots registry key.

        Return
        ------
        str
            Registry key
        """
        return r'Windows Kits\Installed Roots'

    def microsoft(self, key, x86=False):
        """
        Return key in Microsoft software registry.

        Parameters
        ----------
        key: str
            Registry key path where look.
        x86: str
            Force x86 software registry.

        Return
        ------
        str
            Registry key
        """
        node64 = '' if self.pi.current_is_x86() or x86 else 'Wow6432Node'
        return join('Software', node64, 'Microsoft', key)

    def lookup(self, key, name):
        """
        Look for values in registry in Microsoft software registry.

        Parameters
        ----------
        key: str
            Registry key path where look.
        name: str
            Value name to find.

        Return
        ------
        str
            value
        """
        key_read = winreg.KEY_READ
        openkey = winreg.OpenKey
        closekey = winreg.CloseKey
        ms = self.microsoft
        for hkey in self.HKEYS:
            bkey = None
            try:
                bkey = openkey(hkey, ms(key), 0, key_read)
            except (OSError, IOError):
                if not self.pi.current_is_x86():
                    try:
                        bkey = openkey(hkey, ms(key, True), 0, key_read)
                    except (OSError, IOError):
                        continue
                else:
                    continue
            try:
                return winreg.QueryValueEx(bkey, name)[0]
            except (OSError, IOError):
                pass
            finally:
                if bkey:
                    closekey(bkey)


class SystemInfo:
    """
    Microsoft Windows and Visual Studio related system information.

    Parameters
    ----------
    registry_info: RegistryInfo
        "RegistryInfo" instance.
    vc_ver: float
        Required Microsoft Visual C++ version.
    """

    # Variables and properties in this class use originals CamelCase variables
    # names from Microsoft source files for more easy comparison.
    WinDir = environ.get('WinDir', '')
    ProgramFiles = environ.get('ProgramFiles', '')
    ProgramFilesx86 = environ.get('ProgramFiles(x86)', ProgramFiles)

    def __init__(self, registry_info, vc_ver=None):
        self.ri = registry_info
        self.pi = self.ri.pi

        self.known_vs_paths = self.find_programdata_vs_vers()

        # Except for VS15+, VC version is aligned with VS version
        self.vs_ver = self.vc_ver = (
            vc_ver or self._find_latest_available_vs_ver())

    def _find_latest_available_vs_ver(self):
        """
        Find the latest VC version

        Return
        ------
        float
            version
        """
        reg_vc_vers = self.find_reg_vs_vers()

        if not (reg_vc_vers or self.known_vs_paths):
            raise distutils.errors.DistutilsPlatformError(
                'No Microsoft Visual C++ version found')

        vc_vers = set(reg_vc_vers)
        vc_vers.update(self.known_vs_paths)
        return sorted(vc_vers)[-1]

    def find_reg_vs_vers(self):
        """
        Find Microsoft Visual Studio versions available in registry.

        Return
        ------
        list of float
            Versions
        """
        ms = self.ri.microsoft
        vckeys = (self.ri.vc, self.ri.vc_for_python, self.ri.vs)
        vs_vers = []
        for hkey, key in itertools.product(self.ri.HKEYS, vckeys):
            try:
                bkey = winreg.OpenKey(hkey, ms(key), 0, winreg.KEY_READ)
            except (OSError, IOError):
                continue
            with bkey:
                subkeys, values, _ = winreg.QueryInfoKey(bkey)
                for i in range(values):
                    with contextlib.suppress(ValueError):
                        ver = float(winreg.EnumValue(bkey, i)[0])
                        if ver not in vs_vers:
                            vs_vers.append(ver)
                for i in range(subkeys):
                    with contextlib.suppress(ValueError):
                        ver = float(winreg.EnumKey(bkey, i))
                        if ver not in vs_vers:
                            vs_vers.append(ver)
        return sorted(vs_vers)

    def find_programdata_vs_vers(self):
        r"""
        Find Visual studio 2017+ versions from information in
        "C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances".

        Return
        ------
        dict
            float version as key, path as value.
        """
        vs_versions = {}
        instances_dir = \
            r'C:\ProgramData\Microsoft\VisualStudio\Packages\_Instances'

        try:
            hashed_names = listdir(instances_dir)

        except (OSError, IOError):
            # Directory not exists with all Visual Studio versions
            return vs_versions

        for name in hashed_names:
            try:
                # Get VS installation path from "state.json" file
                state_path = join(instances_dir, name, 'state.json')
                with open(state_path, 'rt', encoding='utf-8') as state_file:
                    state = json.load(state_file)
                vs_path = state['installationPath']

                # Raises OSError if this VS installation does not contain VC
                listdir(join(vs_path, r'VC\Tools\MSVC'))

                # Store version and path
                vs_versions[self._as_float_version(
                    state['installationVersion'])] = vs_path

            except (OSError, IOError, KeyError):
                # Skip if "state.json" file is missing or bad format
                continue

        return vs_versions

    @staticmethod
    def _as_float_version(version):
        """
        Return a string version as a simplified float version (major.minor)

        Parameters
        ----------
        version: str
            Version.

        Return
        ------
        float
            version
        """
        return float('.'.join(version.split('.')[:2]))

    @property
    def VSInstallDir(self):
        """
        Microsoft Visual Studio directory.

        Return
        ------
        str
            path
        """
        # Default path
        default = join(self.ProgramFilesx86,
                       'Microsoft Visual Studio %0.1f' % self.vs_ver)

        # Try to get path from registry, if fail use default path
        return self.ri.lookup(self.ri.vs, '%0.1f' % self.vs_ver) or default

    @property
    def VCInstallDir(self):
        """
        Microsoft Visual C++ directory.

        Return
        ------
        str
            path
        """
        path = self._guess_vc() or self._guess_vc_legacy()

        if not isdir(path):
            msg = 'Microsoft Visual C++ directory not found'
            raise distutils.errors.DistutilsPlatformError(msg)

        return path

    def _guess_vc(self):
        """
        Locate Visual C++ for VS2017+.

        Return
        ------
        str
            path
        """
        if self.vs_ver <= 14.0:
            return ''

        try:
            # First search in known VS paths
            vs_dir = self.known_vs_paths[self.vs_ver]
        except KeyError:
            # Else, search with path from registry
            vs_dir = self.VSInstallDir

        guess_vc = join(vs_dir, r'VC\Tools\MSVC')

        # Subdir with VC exact version as name
        try:
            # Update the VC version with real one instead of VS version
            vc_ver = listdir(guess_vc)[-1]
            self.vc_ver = self._as_float_version(vc_ver)
            return join(guess_vc, vc_ver)
        except (OSError, IOError, IndexError):
            return ''

    def _guess_vc_legacy(self):
        """
        Locate Visual C++ for versions prior to 2017.

        Return
        ------
        str
            path
        """
        default = join(self.ProgramFilesx86,
                       r'Microsoft Visual Studio %0.1f\VC' % self.vs_ver)

        # Try to get "VC++ for Python" path from registry as default path
        reg_path = join(self.ri.vc_for_python, '%0.1f' % self.vs_ver)
        python_vc = self.ri.lookup(reg_path, 'installdir')
        default_vc = join(python_vc, 'VC') if python_vc else default

        # Try to get path from registry, if fail use default path
        return self.ri.lookup(self.ri.vc, '%0.1f' % self.vs_ver) or default_vc

    @property
    def WindowsSdkVersion(self):
        """
        Microsoft Windows SDK versions for specified MSVC++ version.

        Return
        ------
        tuple of str
            versions
        """
        if self.vs_ver <= 9.0:
            return '7.0', '6.1', '6.0a'
        elif self.vs_ver == 10.0:
            return '7.1', '7.0a'
        elif self.vs_ver == 11.0:
            return '8.0', '8.0a'
        elif self.vs_ver == 12.0:
            return '8.1', '8.1a'
        elif self.vs_ver >= 14.0:
            return '10.0', '8.1'

    @property
    def WindowsSdkLastVersion(self):
        """
        Microsoft Windows SDK last version.

        Return
        ------
        str
            version
        """
        return self._use_last_dir_name(join(self.WindowsSdkDir, 'lib'))

    @property  # noqa: C901
    def WindowsSdkDir(self):  # noqa: C901  # is too complex (12)  # FIXME
        """
        Microsoft Windows SDK directory.

        Return
        ------
        str
            path
        """
        sdkdir = ''
        for ver in self.WindowsSdkVersion:
            # Try to get it from registry
            loc = join(self.ri.windows_sdk, 'v%s' % ver)
            sdkdir = self.ri.lookup(loc, 'installationfolder')
            if sdkdir:
                break
        if not sdkdir or not isdir(sdkdir):
            # Try to get "VC++ for Python" version from registry
            path = join(self.ri.vc_for_python, '%0.1f' % self.vc_ver)
            install_base = self.ri.lookup(path, 'installdir')
            if install_base:
                sdkdir = join(install_base, 'WinSDK')
        if not sdkdir or not isdir(sdkdir):
            # If fail, use default new path
            for ver in self.WindowsSdkVersion:
                intver = ver[:ver.rfind('.')]
                path = r'Microsoft SDKs\Windows Kits\%s' % intver
                d = join(self.ProgramFiles, path)
                if isdir(d):
                    sdkdir = d
        if not sdkdir or not isdir(sdkdir):
            # If fail, use default old path
            for ver in self.WindowsSdkVersion:
                path = r'Microsoft SDKs\Windows\v%s' % ver
                d = join(self.ProgramFiles, path)
                if isdir(d):
                    sdkdir = d
        if not sdkdir:
            # If fail, use Platform SDK
            sdkdir = join(self.VCInstallDir, 'PlatformSDK')
        return sdkdir

    @property
    def WindowsSDKExecutablePath(self):
        """
        Microsoft Windows SDK executable directory.

        Return
        ------
        str
            path
        """
        # Find WinSDK NetFx Tools registry dir name
        if self.vs_ver <= 11.0:
            netfxver = 35
            arch = ''
        else:
            netfxver = 40
            hidex86 = True if self.vs_ver <= 12.0 else False
            arch = self.pi.current_dir(x64=True, hidex86=hidex86)
        fx = 'WinSDK-NetFx%dTools%s' % (netfxver, arch.replace('\\', '-'))

        # list all possibles registry paths
        regpaths = []
        if self.vs_ver >= 14.0:
            for ver in self.NetFxSdkVersion:
                regpaths += [join(self.ri.netfx_sdk, ver, fx)]

        for ver in self.WindowsSdkVersion:
            regpaths += [join(self.ri.windows_sdk, 'v%sA' % ver, fx)]

        # Return installation folder from the more recent path
        for path in regpaths:
            execpath = self.ri.lookup(path, 'installationfolder')
            if execpath:
                return execpath

    @property
    def FSharpInstallDir(self):
        """
        Microsoft Visual F# directory.

        Return
        ------
        str
            path
        """
        path = join(self.ri.visualstudio, r'%0.1f\Setup\F#' % self.vs_ver)
        return self.ri.lookup(path, 'productdir') or ''

    @property
    def UniversalCRTSdkDir(self):
        """
        Microsoft Universal CRT SDK directory.

        Return
        ------
        str
            path
        """
        # Set Kit Roots versions for specified MSVC++ version
        vers = ('10', '81') if self.vs_ver >= 14.0 else ()

        # Find path of the more recent Kit
        for ver in vers:
            sdkdir = self.ri.lookup(self.ri.windows_kits_roots,
                                    'kitsroot%s' % ver)
            if sdkdir:
                return sdkdir or ''

    @property
    def UniversalCRTSdkLastVersion(self):
        """
        Microsoft Universal C Runtime SDK last version.

        Return
        ------
        str
            version
        """
        return self._use_last_dir_name(join(self.UniversalCRTSdkDir, 'lib'))

    @property
    def NetFxSdkVersion(self):
        """
        Microsoft .NET Framework SDK versions.

        Return
        ------
        tuple of str
            versions
        """
        # Set FxSdk versions for specified VS version
        return (('4.7.2', '4.7.1', '4.7',
                 '4.6.2', '4.6.1', '4.6',
                 '4.5.2', '4.5.1', '4.5')
                if self.vs_ver >= 14.0 else ())

    @property
    def NetFxSdkDir(self):
        """
        Microsoft .NET Framework SDK directory.

        Return
        ------
        str
            path
        """
        sdkdir = ''
        for ver in self.NetFxSdkVersion:
            loc = join(self.ri.netfx_sdk, ver)
            sdkdir = self.ri.lookup(loc, 'kitsinstallationfolder')
            if sdkdir:
                break
        return sdkdir

    @property
    def FrameworkDir32(self):
        """
        Microsoft .NET Framework 32bit directory.

        Return
        ------
        str
            path
        """
        # Default path
        guess_fw = join(self.WinDir, r'Microsoft.NET\Framework')

        # Try to get path from registry, if fail use default path
        return self.ri.lookup(self.ri.vc, 'frameworkdir32') or guess_fw

    @property
    def FrameworkDir64(self):
        """
        Microsoft .NET Framework 64bit directory.

        Return
        ------
        str
            path
        """
        # Default path
        guess_fw = join(self.WinDir, r'Microsoft.NET\Framework64')

        # Try to get path from registry, if fail use default path
        return self.ri.lookup(self.ri.vc, 'frameworkdir64') or guess_fw

    @property
    def FrameworkVersion32(self):
        """
        Microsoft .NET Framework 32bit versions.

        Return
        ------
        tuple of str
            versions
        """
        return self._find_dot_net_versions(32)

    @property
    def FrameworkVersion64(self):
        """
        Microsoft .NET Framework 64bit versions.

        Return
        ------
        tuple of str
            versions
        """
        return self._find_dot_net_versions(64)

    def _find_dot_net_versions(self, bits):
        """
        Find Microsoft .NET Framework versions.

        Parameters
        ----------
        bits: int
            Platform number of bits: 32 or 64.

        Return
        ------
        tuple of str
            versions
        """
        # Find actual .NET version in registry
        reg_ver = self.ri.lookup(self.ri.vc, 'frameworkver%d' % bits)
        dot_net_dir = getattr(self, 'FrameworkDir%d' % bits)
        ver = reg_ver or self._use_last_dir_name(dot_net_dir, 'v') or ''

        # Set .NET versions for specified MSVC++ version
        if self.vs_ver >= 12.0:
            return ver, 'v4.0'
        elif self.vs_ver >= 10.0:
            return 'v4.0.30319' if ver.lower()[:2] != 'v4' else ver, 'v3.5'
        elif self.vs_ver == 9.0:
            return 'v3.5', 'v2.0.50727'
        elif self.vs_ver == 8.0:
            return 'v3.0', 'v2.0.50727'

    @staticmethod
    def _use_last_dir_name(path, prefix=''):
        """
        Return name of the last dir in path or '' if no dir found.

        Parameters
        ----------
        path: str
            Use dirs in this path
        prefix: str
            Use only dirs starting by this prefix

        Return
        ------
        str
            name
        """
        matching_dirs = (
            dir_name
            for dir_name in reversed(listdir(path))
            if isdir(join(path, dir_name)) and
            dir_name.startswith(prefix)
        )
        return next(matching_dirs, None) or ''


class EnvironmentInfo:
    """
    Return environment variables for specified Microsoft Visual C++ version
    and platform : Lib, Include, Path and libpath.

    This function is compatible with Microsoft Visual C++ 9.0 to 14.X.

    Script created by analysing Microsoft environment configuration files like
    "vcvars[...].bat", "SetEnv.Cmd", "vcbuildtools.bat", ...

    Parameters
    ----------
    arch: str
        Target architecture.
    vc_ver: float
        Required Microsoft Visual C++ version. If not set, autodetect the last
        version.
    vc_min_ver: float
        Minimum Microsoft Visual C++ version.
    """

    # Variables and properties in this class use originals CamelCase variables
    # names from Microsoft source files for more easy comparison.

    def __init__(self, arch, vc_ver=None, vc_min_ver=0):
        self.pi = PlatformInfo(arch)
        self.ri = RegistryInfo(self.pi)
        self.si = SystemInfo(self.ri, vc_ver)

        if self.vc_ver < vc_min_ver:
            err = 'No suitable Microsoft Visual C++ version found'
            raise distutils.errors.DistutilsPlatformError(err)

    @property
    def vs_ver(self):
        """
        Microsoft Visual Studio.

        Return
        ------
        float
            version
        """
        return self.si.vs_ver

    @property
    def vc_ver(self):
        """
        Microsoft Visual C++ version.

        Return
        ------
        float
            version
        """
        return self.si.vc_ver

    @property
    def VSTools(self):
        """
        Microsoft Visual Studio Tools.

        Return
        ------
        list of str
            paths
        """
        paths = [r'Common7\IDE', r'Common7\Tools']

        if self.vs_ver >= 14.0:
            arch_subdir = self.pi.current_dir(hidex86=True, x64=True)
            paths += [r'Common7\IDE\CommonExtensions\Microsoft\TestWindow']
            paths += [r'Team Tools\Performance Tools']
            paths += [r'Team Tools\Performance Tools%s' % arch_subdir]

        return [join(self.si.VSInstallDir, path) for path in paths]

    @property
    def VCIncludes(self):
        """
        Microsoft Visual C++ & Microsoft Foundation Class Includes.

        Return
        ------
        list of str
            paths
        """
        return [join(self.si.VCInstallDir, 'Include'),
                join(self.si.VCInstallDir, r'ATLMFC\Include')]

    @property
    def VCLibraries(self):
        """
        Microsoft Visual C++ & Microsoft Foundation Class Libraries.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver >= 15.0:
            arch_subdir = self.pi.target_dir(x64=True)
        else:
            arch_subdir = self.pi.target_dir(hidex86=True)
        paths = ['Lib%s' % arch_subdir, r'ATLMFC\Lib%s' % arch_subdir]

        if self.vs_ver >= 14.0:
            paths += [r'Lib\store%s' % arch_subdir]

        return [join(self.si.VCInstallDir, path) for path in paths]

    @property
    def VCStoreRefs(self):
        """
        Microsoft Visual C++ store references Libraries.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 14.0:
            return []
        return [join(self.si.VCInstallDir, r'Lib\store\references')]

    @property
    def VCTools(self):
        """
        Microsoft Visual C++ Tools.

        Return
        ------
        list of str
            paths
        """
        si = self.si
        tools = [join(si.VCInstallDir, 'VCPackages')]

        forcex86 = True if self.vs_ver <= 10.0 else False
        arch_subdir = self.pi.cross_dir(forcex86)
        if arch_subdir:
            tools += [join(si.VCInstallDir, 'Bin%s' % arch_subdir)]

        if self.vs_ver == 14.0:
            path = 'Bin%s' % self.pi.current_dir(hidex86=True)
            tools += [join(si.VCInstallDir, path)]

        elif self.vs_ver >= 15.0:
            host_dir = (r'bin\HostX86%s' if self.pi.current_is_x86() else
                        r'bin\HostX64%s')
            tools += [join(
                si.VCInstallDir, host_dir % self.pi.target_dir(x64=True))]

            if self.pi.current_cpu != self.pi.target_cpu:
                tools += [join(
                    si.VCInstallDir, host_dir % self.pi.current_dir(x64=True))]

        else:
            tools += [join(si.VCInstallDir, 'Bin')]

        return tools

    @property
    def OSLibraries(self):
        """
        Microsoft Windows SDK Libraries.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver <= 10.0:
            arch_subdir = self.pi.target_dir(hidex86=True, x64=True)
            return [join(self.si.WindowsSdkDir, 'Lib%s' % arch_subdir)]

        else:
            arch_subdir = self.pi.target_dir(x64=True)
            lib = join(self.si.WindowsSdkDir, 'lib')
            libver = self._sdk_subdir
            return [join(lib, '%sum%s' % (libver, arch_subdir))]

    @property
    def OSIncludes(self):
        """
        Microsoft Windows SDK Include.

        Return
        ------
        list of str
            paths
        """
        include = join(self.si.WindowsSdkDir, 'include')

        if self.vs_ver <= 10.0:
            return [include, join(include, 'gl')]

        else:
            if self.vs_ver >= 14.0:
                sdkver = self._sdk_subdir
            else:
                sdkver = ''
            return [join(include, '%sshared' % sdkver),
                    join(include, '%sum' % sdkver),
                    join(include, '%swinrt' % sdkver)]

    @property
    def OSLibpath(self):
        """
        Microsoft Windows SDK Libraries Paths.

        Return
        ------
        list of str
            paths
        """
        ref = join(self.si.WindowsSdkDir, 'References')
        libpath = []

        if self.vs_ver <= 9.0:
            libpath += self.OSLibraries

        if self.vs_ver >= 11.0:
            libpath += [join(ref, r'CommonConfiguration\Neutral')]

        if self.vs_ver >= 14.0:
            libpath += [
                ref,
                join(self.si.WindowsSdkDir, 'UnionMetadata'),
                join(
                    ref, 'Windows.Foundation.UniversalApiContract', '1.0.0.0'),
                join(ref, 'Windows.Foundation.FoundationContract', '1.0.0.0'),
                join(
                    ref, 'Windows.Networking.Connectivity.WwanContract',
                    '1.0.0.0'),
                join(
                    self.si.WindowsSdkDir, 'ExtensionSDKs', 'Microsoft.VCLibs',
                    '%0.1f' % self.vs_ver, 'References', 'CommonConfiguration',
                    'neutral'),
            ]
        return libpath

    @property
    def SdkTools(self):
        """
        Microsoft Windows SDK Tools.

        Return
        ------
        list of str
            paths
        """
        return list(self._sdk_tools())

    def _sdk_tools(self):
        """
        Microsoft Windows SDK Tools paths generator.

        Return
        ------
        generator of str
            paths
        """
        if self.vs_ver < 15.0:
            bin_dir = 'Bin' if self.vs_ver <= 11.0 else r'Bin\x86'
            yield join(self.si.WindowsSdkDir, bin_dir)

        if not self.pi.current_is_x86():
            arch_subdir = self.pi.current_dir(x64=True)
            path = 'Bin%s' % arch_subdir
            yield join(self.si.WindowsSdkDir, path)

        if self.vs_ver in (10.0, 11.0):
            if self.pi.target_is_x86():
                arch_subdir = ''
            else:
                arch_subdir = self.pi.current_dir(hidex86=True, x64=True)
            path = r'Bin\NETFX 4.0 Tools%s' % arch_subdir
            yield join(self.si.WindowsSdkDir, path)

        elif self.vs_ver >= 15.0:
            path = join(self.si.WindowsSdkDir, 'Bin')
            arch_subdir = self.pi.current_dir(x64=True)
            sdkver = self.si.WindowsSdkLastVersion
            yield join(path, '%s%s' % (sdkver, arch_subdir))

        if self.si.WindowsSDKExecutablePath:
            yield self.si.WindowsSDKExecutablePath

    @property
    def _sdk_subdir(self):
        """
        Microsoft Windows SDK version subdir.

        Return
        ------
        str
            subdir
        """
        ucrtver = self.si.WindowsSdkLastVersion
        return ('%s\\' % ucrtver) if ucrtver else ''

    @property
    def SdkSetup(self):
        """
        Microsoft Windows SDK Setup.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver > 9.0:
            return []

        return [join(self.si.WindowsSdkDir, 'Setup')]

    @property
    def FxTools(self):
        """
        Microsoft .NET Framework Tools.

        Return
        ------
        list of str
            paths
        """
        pi = self.pi
        si = self.si

        if self.vs_ver <= 10.0:
            include32 = True
            include64 = not pi.target_is_x86() and not pi.current_is_x86()
        else:
            include32 = pi.target_is_x86() or pi.current_is_x86()
            include64 = pi.current_cpu == 'amd64' or pi.target_cpu == 'amd64'

        tools = []
        if include32:
            tools += [join(si.FrameworkDir32, ver)
                      for ver in si.FrameworkVersion32]
        if include64:
            tools += [join(si.FrameworkDir64, ver)
                      for ver in si.FrameworkVersion64]
        return tools

    @property
    def NetFxSDKLibraries(self):
        """
        Microsoft .Net Framework SDK Libraries.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 14.0 or not self.si.NetFxSdkDir:
            return []

        arch_subdir = self.pi.target_dir(x64=True)
        return [join(self.si.NetFxSdkDir, r'lib\um%s' % arch_subdir)]

    @property
    def NetFxSDKIncludes(self):
        """
        Microsoft .Net Framework SDK Includes.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 14.0 or not self.si.NetFxSdkDir:
            return []

        return [join(self.si.NetFxSdkDir, r'include\um')]

    @property
    def VsTDb(self):
        """
        Microsoft Visual Studio Team System Database.

        Return
        ------
        list of str
            paths
        """
        return [join(self.si.VSInstallDir, r'VSTSDB\Deploy')]

    @property
    def MSBuild(self):
        """
        Microsoft Build Engine.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 12.0:
            return []
        elif self.vs_ver < 15.0:
            base_path = self.si.ProgramFilesx86
            arch_subdir = self.pi.current_dir(hidex86=True)
        else:
            base_path = self.si.VSInstallDir
            arch_subdir = ''

        path = r'MSBuild\%0.1f\bin%s' % (self.vs_ver, arch_subdir)
        build = [join(base_path, path)]

        if self.vs_ver >= 15.0:
            # Add Roslyn C# & Visual Basic Compiler
            build += [join(base_path, path, 'Roslyn')]

        return build

    @property
    def HTMLHelpWorkshop(self):
        """
        Microsoft HTML Help Workshop.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 11.0:
            return []

        return [join(self.si.ProgramFilesx86, 'HTML Help Workshop')]

    @property
    def UCRTLibraries(self):
        """
        Microsoft Universal C Runtime SDK Libraries.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 14.0:
            return []

        arch_subdir = self.pi.target_dir(x64=True)
        lib = join(self.si.UniversalCRTSdkDir, 'lib')
        ucrtver = self._ucrt_subdir
        return [join(lib, '%sucrt%s' % (ucrtver, arch_subdir))]

    @property
    def UCRTIncludes(self):
        """
        Microsoft Universal C Runtime SDK Include.

        Return
        ------
        list of str
            paths
        """
        if self.vs_ver < 14.0:
            return []

        include = join(self.si.UniversalCRTSdkDir, 'include')
        return [join(include, '%sucrt' % self._ucrt_subdir)]

    @property
    def _ucrt_subdir(self):
        """
        Microsoft Universal C Runtime SDK version subdir.

        Return
        ------
        str
            subdir
        """
        ucrtver = self.si.UniversalCRTSdkLastVersion
        return ('%s\\' % ucrtver) if ucrtver else ''

    @property
    def FSharp(self):
        """
        Microsoft Visual F#.

        Return
        ------
        list of str
            paths
        """
        if 11.0 > self.vs_ver > 12.0:
            return []

        return [self.si.FSharpInstallDir]

    @property
    def VCRuntimeRedist(self):
        """
        Microsoft Visual C++ runtime redistributable dll.

        Return
        ------
        str
            path
        """
        vcruntime = 'vcruntime%d0.dll' % self.vc_ver
        arch_subdir = self.pi.target_dir(x64=True).strip('\\')

        # Installation prefixes candidates
        prefixes = []
        tools_path = self.si.VCInstallDir
        redist_path = dirname(tools_path.replace(r'\Tools', r'\Redist'))
        if isdir(redist_path):
            # Redist version may not be exactly the same as tools
            redist_path = join(redist_path, listdir(redist_path)[-1])
            prefixes += [redist_path, join(redist_path, 'onecore')]

        prefixes += [join(tools_path, 'redist')]  # VS14 legacy path

        # CRT directory
        crt_dirs = ('Microsoft.VC%d.CRT' % (self.vc_ver * 10),
                    # Sometime store in directory with VS version instead of VC
                    'Microsoft.VC%d.CRT' % (int(self.vs_ver) * 10))

        # vcruntime path
        for prefix, crt_dir in itertools.product(prefixes, crt_dirs):
            path = join(prefix, arch_subdir, crt_dir, vcruntime)
            if isfile(path):
                return path

    def return_env(self, exists=True):
        """
        Return environment dict.

        Parameters
        ----------
        exists: bool
            It True, only return existing paths.

        Return
        ------
        dict
            environment
        """
        env = dict(
            include=self._build_paths('include',
                                      [self.VCIncludes,
                                       self.OSIncludes,
                                       self.UCRTIncludes,
                                       self.NetFxSDKIncludes],
                                      exists),
            lib=self._build_paths('lib',
                                  [self.VCLibraries,
                                   self.OSLibraries,
                                   self.FxTools,
                                   self.UCRTLibraries,
                                   self.NetFxSDKLibraries],
                                  exists),
            libpath=self._build_paths('libpath',
                                      [self.VCLibraries,
                                       self.FxTools,
                                       self.VCStoreRefs,
                                       self.OSLibpath],
                                      exists),
            path=self._build_paths('path',
                                   [self.VCTools,
                                    self.VSTools,
                                    self.VsTDb,
                                    self.SdkTools,
                                    self.SdkSetup,
                                    self.FxTools,
                                    self.MSBuild,
                                    self.HTMLHelpWorkshop,
                                    self.FSharp],
                                   exists),
        )
        if self.vs_ver >= 14 and isfile(self.VCRuntimeRedist):
            env['py_vcruntime_redist'] = self.VCRuntimeRedist
        return env

    def _build_paths(self, name, spec_path_lists, exists):
        """
        Given an environment variable name and specified paths,
        return a pathsep-separated string of paths containing
        unique, extant, directories from those paths and from
        the environment variable. Raise an error if no paths
        are resolved.

        Parameters
        ----------
        name: str
            Environment variable name
        spec_path_lists: list of str
            Paths
        exists: bool
            It True, only return existing paths.

        Return
        ------
        str
            Pathsep-separated paths
        """
        # flatten spec_path_lists
        spec_paths = itertools.chain.from_iterable(spec_path_lists)
        env_paths = environ.get(name, '').split(pathsep)
        paths = itertools.chain(spec_paths, env_paths)
        extant_paths = list(filter(isdir, paths)) if exists else paths
        if not extant_paths:
            msg = "%s environment variable is empty" % name.upper()
            raise distutils.errors.DistutilsPlatformError(msg)
        unique_paths = unique_everseen(extant_paths)
        return pathsep.join(unique_paths)
site-packages/setuptools/script.tmpl000064400000000212150564745010013711 0ustar00# EASY-INSTALL-SCRIPT: %(spec)r,%(script_name)r
__requires__ = %(spec)r
__import__('pkg_resources').run_script(%(spec)r, %(script_name)r)
site-packages/setuptools/launch.py000064400000001454150564745010013344 0ustar00"""
Launch the Python script on the command line after
setuptools is bootstrapped via import.
"""

# Note that setuptools gets imported implicitly by the
# invocation of this script using python -m setuptools.launch

import tokenize
import sys


def run():
    """
    Run the script in sys.argv[1] as if it had
    been invoked naturally.
    """
    __builtins__
    script_name = sys.argv[1]
    namespace = dict(
        __file__=script_name,
        __name__='__main__',
        __doc__=None,
    )
    sys.argv[:] = sys.argv[1:]

    open_ = getattr(tokenize, 'open', open)
    with open_(script_name) as fid:
        script = fid.read()
    norm_script = script.replace('\\r\\n', '\\n')
    code = compile(norm_script, script_name, 'exec')
    exec(code, namespace)


if __name__ == '__main__':
    run()
site-packages/setuptools/dep_util.py000064400000001665150564745010013703 0ustar00from distutils.dep_util import newer_group


# yes, this is was almost entirely copy-pasted from
# 'newer_pairwise()', this is just another convenience
# function.
def newer_pairwise_group(sources_groups, targets):
    """Walk both arguments in parallel, testing if each source group is newer
    than its corresponding target. Returns a pair of lists (sources_groups,
    targets) where sources is newer than target, according to the semantics
    of 'newer_group()'.
    """
    if len(sources_groups) != len(targets):
        raise ValueError(
            "'sources_group' and 'targets' must be the same length")

    # build a pair of lists (sources_groups, targets) where source is newer
    n_sources = []
    n_targets = []
    for i in range(len(sources_groups)):
        if newer_group(sources_groups[i], targets[i]):
            n_sources.append(sources_groups[i])
            n_targets.append(targets[i])

    return n_sources, n_targets
site-packages/setuptools/version.py000064400000000220150564745010013545 0ustar00import pkg_resources

try:
    __version__ = pkg_resources.get_distribution('setuptools').version
except Exception:
    __version__ = 'unknown'
site-packages/setuptools/monkey.py000064400000012141150564745010013367 0ustar00"""
Monkey patching of distutils.
"""

import sys
import distutils.filelist
import platform
import types
import functools
from importlib import import_module
import inspect

import setuptools

__all__ = []
"""
Everything is private. Contact the project team
if you think you need this functionality.
"""


def _get_mro(cls):
    """
    Returns the bases classes for cls sorted by the MRO.

    Works around an issue on Jython where inspect.getmro will not return all
    base classes if multiple classes share the same name. Instead, this
    function will return a tuple containing the class itself, and the contents
    of cls.__bases__. See https://github.com/pypa/setuptools/issues/1024.
    """
    if platform.python_implementation() == "Jython":
        return (cls,) + cls.__bases__
    return inspect.getmro(cls)


def get_unpatched(item):
    lookup = (
        get_unpatched_class if isinstance(item, type) else
        get_unpatched_function if isinstance(item, types.FunctionType) else
        lambda item: None
    )
    return lookup(item)


def get_unpatched_class(cls):
    """Protect against re-patching the distutils if reloaded

    Also ensures that no other distutils extension monkeypatched the distutils
    first.
    """
    external_bases = (
        cls
        for cls in _get_mro(cls)
        if not cls.__module__.startswith('setuptools')
    )
    base = next(external_bases)
    if not base.__module__.startswith('distutils'):
        msg = "distutils has already been patched by %r" % cls
        raise AssertionError(msg)
    return base


def patch_all():
    # we can't patch distutils.cmd, alas
    distutils.core.Command = setuptools.Command

    has_issue_12885 = sys.version_info <= (3, 5, 3)

    if has_issue_12885:
        # fix findall bug in distutils (http://bugs.python.org/issue12885)
        distutils.filelist.findall = setuptools.findall

    needs_warehouse = (
        sys.version_info < (2, 7, 13)
        or
        (3, 4) < sys.version_info < (3, 4, 6)
        or
        (3, 5) < sys.version_info <= (3, 5, 3)
    )

    if needs_warehouse:
        warehouse = 'https://upload.pypi.org/legacy/'
        distutils.config.PyPIRCCommand.DEFAULT_REPOSITORY = warehouse

    _patch_distribution_metadata()

    # Install Distribution throughout the distutils
    for module in distutils.dist, distutils.core, distutils.cmd:
        module.Distribution = setuptools.dist.Distribution

    # Install the patched Extension
    distutils.core.Extension = setuptools.extension.Extension
    distutils.extension.Extension = setuptools.extension.Extension
    if 'distutils.command.build_ext' in sys.modules:
        sys.modules['distutils.command.build_ext'].Extension = (
            setuptools.extension.Extension
        )

    patch_for_msvc_specialized_compiler()


def _patch_distribution_metadata():
    """Patch write_pkg_file and read_pkg_file for higher metadata standards"""
    for attr in ('write_pkg_file', 'read_pkg_file', 'get_metadata_version'):
        new_val = getattr(setuptools.dist, attr)
        setattr(distutils.dist.DistributionMetadata, attr, new_val)


def patch_func(replacement, target_mod, func_name):
    """
    Patch func_name in target_mod with replacement

    Important - original must be resolved by name to avoid
    patching an already patched function.
    """
    original = getattr(target_mod, func_name)

    # set the 'unpatched' attribute on the replacement to
    # point to the original.
    vars(replacement).setdefault('unpatched', original)

    # replace the function in the original module
    setattr(target_mod, func_name, replacement)


def get_unpatched_function(candidate):
    return getattr(candidate, 'unpatched')


def patch_for_msvc_specialized_compiler():
    """
    Patch functions in distutils to use standalone Microsoft Visual C++
    compilers.
    """
    # import late to avoid circular imports on Python < 3.5
    msvc = import_module('setuptools.msvc')

    if platform.system() != 'Windows':
        # Compilers only available on Microsoft Windows
        return

    def patch_params(mod_name, func_name):
        """
        Prepare the parameters for patch_func to patch indicated function.
        """
        repl_prefix = 'msvc9_' if 'msvc9' in mod_name else 'msvc14_'
        repl_name = repl_prefix + func_name.lstrip('_')
        repl = getattr(msvc, repl_name)
        mod = import_module(mod_name)
        if not hasattr(mod, func_name):
            raise ImportError(func_name)
        return repl, mod, func_name

    # Python 2.7 to 3.4
    msvc9 = functools.partial(patch_params, 'distutils.msvc9compiler')

    # Python 3.5+
    msvc14 = functools.partial(patch_params, 'distutils._msvccompiler')

    try:
        # Patch distutils.msvc9compiler
        patch_func(*msvc9('find_vcvarsall'))
        patch_func(*msvc9('query_vcvarsall'))
    except ImportError:
        pass

    try:
        # Patch distutils._msvccompiler._get_vc_env
        patch_func(*msvc14('_get_vc_env'))
    except ImportError:
        pass

    try:
        # Patch distutils._msvccompiler.gen_lib_options for Numpy
        patch_func(*msvc14('gen_lib_options'))
    except ImportError:
        pass
site-packages/setuptools/script (dev).tmpl000064400000000332150564745010014574 0ustar00# EASY-INSTALL-DEV-SCRIPT: %(spec)r,%(script_name)r
__requires__ = %(spec)r
__import__('pkg_resources').require(%(spec)r)
__file__ = %(dev_path)r
with open(__file__) as f:
    exec(compile(f.read(), __file__, 'exec'))
site-packages/setuptools/installer.py000064400000006757150564745010014102 0ustar00import glob
import os
import subprocess
import sys
import tempfile
from distutils import log
from distutils.errors import DistutilsError

import pkg_resources
from setuptools.wheel import Wheel


def _fixup_find_links(find_links):
    """Ensure find-links option end-up being a list of strings."""
    if isinstance(find_links, str):
        return find_links.split()
    assert isinstance(find_links, (tuple, list))
    return find_links


def fetch_build_egg(dist, req):  # noqa: C901  # is too complex (16)  # FIXME
    """Fetch an egg needed for building.

    Use pip/wheel to fetch/build a wheel."""
    # Warn if wheel is not available
    try:
        pkg_resources.get_distribution('wheel')
    except pkg_resources.DistributionNotFound:
        dist.announce('WARNING: The wheel package is not available.', log.WARN)
    # Ignore environment markers; if supplied, it is required.
    req = strip_marker(req)
    # Take easy_install options into account, but do not override relevant
    # pip environment variables (like PIP_INDEX_URL or PIP_QUIET); they'll
    # take precedence.
    opts = dist.get_option_dict('easy_install')
    if 'allow_hosts' in opts:
        raise DistutilsError('the `allow-hosts` option is not supported '
                             'when using pip to install requirements.')
    quiet = 'PIP_QUIET' not in os.environ and 'PIP_VERBOSE' not in os.environ
    if 'PIP_INDEX_URL' in os.environ:
        index_url = None
    elif 'index_url' in opts:
        index_url = opts['index_url'][1]
    else:
        index_url = None
    find_links = (
        _fixup_find_links(opts['find_links'][1])[:] if 'find_links' in opts
        else []
    )
    if dist.dependency_links:
        find_links.extend(dist.dependency_links)
    eggs_dir = os.path.realpath(dist.get_egg_cache_dir())
    environment = pkg_resources.Environment()
    for egg_dist in pkg_resources.find_distributions(eggs_dir):
        if egg_dist in req and environment.can_add(egg_dist):
            return egg_dist
    with tempfile.TemporaryDirectory() as tmpdir:
        cmd = [
            sys.executable, '-m', 'pip',
            '--disable-pip-version-check',
            'wheel', '--no-deps',
            '-w', tmpdir,
        ]
        if quiet:
            cmd.append('--quiet')
        if index_url is not None:
            cmd.extend(('--index-url', index_url))
        for link in find_links or []:
            cmd.extend(('--find-links', link))
        # If requirement is a PEP 508 direct URL, directly pass
        # the URL to pip, as `req @ url` does not work on the
        # command line.
        cmd.append(req.url or str(req))
        try:
            subprocess.check_call(cmd)
        except subprocess.CalledProcessError as e:
            raise DistutilsError(str(e)) from e
        wheel = Wheel(glob.glob(os.path.join(tmpdir, '*.whl'))[0])
        dist_location = os.path.join(eggs_dir, wheel.egg_name())
        wheel.install_as_egg(dist_location)
        dist_metadata = pkg_resources.PathMetadata(
            dist_location, os.path.join(dist_location, 'EGG-INFO'))
        dist = pkg_resources.Distribution.from_filename(
            dist_location, metadata=dist_metadata)
        return dist


def strip_marker(req):
    """
    Return a new requirement without the environment marker to avoid
    calling pip with something like `babel; extra == "i18n"`, which
    would always be ignored.
    """
    # create a copy to avoid mutating the input
    req = pkg_resources.Requirement.parse(str(req))
    req.marker = None
    return req
site-packages/setuptools/__init__.py000064400000016766150564745010013645 0ustar00"""Extensions to the 'distutils' for large or complex distributions"""

from fnmatch import fnmatchcase
import functools
import os
import re
import warnings

import _distutils_hack.override  # noqa: F401

import distutils.core
from distutils.errors import DistutilsOptionError
from distutils.util import convert_path

from ._deprecation_warning import SetuptoolsDeprecationWarning

import setuptools.version
from setuptools.extension import Extension
from setuptools.dist import Distribution
from setuptools.depends import Require
from . import monkey


__all__ = [
    'setup',
    'Distribution',
    'Command',
    'Extension',
    'Require',
    'SetuptoolsDeprecationWarning',
    'find_packages',
    'find_namespace_packages',
]

__version__ = setuptools.version.__version__

bootstrap_install_from = None


class PackageFinder:
    """
    Generate a list of all Python packages found within a directory
    """

    @classmethod
    def find(cls, where='.', exclude=(), include=('*',)):
        """Return a list all Python packages found within directory 'where'

        'where' is the root directory which will be searched for packages.  It
        should be supplied as a "cross-platform" (i.e. URL-style) path; it will
        be converted to the appropriate local path syntax.

        'exclude' is a sequence of package names to exclude; '*' can be used
        as a wildcard in the names, such that 'foo.*' will exclude all
        subpackages of 'foo' (but not 'foo' itself).

        'include' is a sequence of package names to include.  If it's
        specified, only the named packages will be included.  If it's not
        specified, all found packages will be included.  'include' can contain
        shell style wildcard patterns just like 'exclude'.
        """

        return list(
            cls._find_packages_iter(
                convert_path(where),
                cls._build_filter('ez_setup', '*__pycache__', *exclude),
                cls._build_filter(*include),
            )
        )

    @classmethod
    def _find_packages_iter(cls, where, exclude, include):
        """
        All the packages found in 'where' that pass the 'include' filter, but
        not the 'exclude' filter.
        """
        for root, dirs, files in os.walk(where, followlinks=True):
            # Copy dirs to iterate over it, then empty dirs.
            all_dirs = dirs[:]
            dirs[:] = []

            for dir in all_dirs:
                full_path = os.path.join(root, dir)
                rel_path = os.path.relpath(full_path, where)
                package = rel_path.replace(os.path.sep, '.')

                # Skip directory trees that are not valid packages
                if '.' in dir or not cls._looks_like_package(full_path):
                    continue

                # Should this package be included?
                if include(package) and not exclude(package):
                    yield package

                # Keep searching subdirectories, as there may be more packages
                # down there, even if the parent was excluded.
                dirs.append(dir)

    @staticmethod
    def _looks_like_package(path):
        """Does a directory look like a package?"""
        return os.path.isfile(os.path.join(path, '__init__.py'))

    @staticmethod
    def _build_filter(*patterns):
        """
        Given a list of patterns, return a callable that will be true only if
        the input matches at least one of the patterns.
        """
        return lambda name: any(fnmatchcase(name, pat=pat) for pat in patterns)


class PEP420PackageFinder(PackageFinder):
    @staticmethod
    def _looks_like_package(path):
        return True


find_packages = PackageFinder.find
find_namespace_packages = PEP420PackageFinder.find


def _install_setup_requires(attrs):
    # Note: do not use `setuptools.Distribution` directly, as
    # our PEP 517 backend patch `distutils.core.Distribution`.
    class MinimalDistribution(distutils.core.Distribution):
        """
        A minimal version of a distribution for supporting the
        fetch_build_eggs interface.
        """

        def __init__(self, attrs):
            _incl = 'dependency_links', 'setup_requires'
            filtered = {k: attrs[k] for k in set(_incl) & set(attrs)}
            distutils.core.Distribution.__init__(self, filtered)

        def finalize_options(self):
            """
            Disable finalize_options to avoid building the working set.
            Ref #2158.
            """

    dist = MinimalDistribution(attrs)

    # Honor setup.cfg's options.
    dist.parse_config_files(ignore_option_errors=True)
    if dist.setup_requires:
        warnings.warn(
            "setup_requires is deprecated. Supply build "
            "dependencies using PEP 517 pyproject.toml build-requires.",
            SetuptoolsDeprecationWarning,
        )
        dist.fetch_build_eggs(dist.setup_requires)


def setup(**attrs):
    # Make sure we have any requirements needed to interpret 'attrs'.
    _install_setup_requires(attrs)
    return distutils.core.setup(**attrs)


setup.__doc__ = distutils.core.setup.__doc__


_Command = monkey.get_unpatched(distutils.core.Command)


class Command(_Command):
    __doc__ = _Command.__doc__

    command_consumes_arguments = False

    def __init__(self, dist, **kw):
        """
        Construct the command for dist, updating
        vars(self) with any keyword parameters.
        """
        _Command.__init__(self, dist)
        vars(self).update(kw)

    def _ensure_stringlike(self, option, what, default=None):
        val = getattr(self, option)
        if val is None:
            setattr(self, option, default)
            return default
        elif not isinstance(val, str):
            raise DistutilsOptionError(
                "'%s' must be a %s (got `%s`)" % (option, what, val)
            )
        return val

    def ensure_string_list(self, option):
        r"""Ensure that 'option' is a list of strings.  If 'option' is
        currently a string, we split it either on /,\s*/ or /\s+/, so
        "foo bar baz", "foo,bar,baz", and "foo,   bar baz" all become
        ["foo", "bar", "baz"].
        """
        val = getattr(self, option)
        if val is None:
            return
        elif isinstance(val, str):
            setattr(self, option, re.split(r',\s*|\s+', val))
        else:
            if isinstance(val, list):
                ok = all(isinstance(v, str) for v in val)
            else:
                ok = False
            if not ok:
                raise DistutilsOptionError(
                    "'%s' must be a list of strings (got %r)" % (option, val)
                )

    def reinitialize_command(self, command, reinit_subcommands=0, **kw):
        cmd = _Command.reinitialize_command(self, command, reinit_subcommands)
        vars(cmd).update(kw)
        return cmd


def _find_all_simple(path):
    """
    Find all files under 'path'
    """
    results = (
        os.path.join(base, file)
        for base, dirs, files in os.walk(path, followlinks=True)
        for file in files
    )
    return filter(os.path.isfile, results)


def findall(dir=os.curdir):
    """
    Find all files under 'dir' and return the list of full filenames.
    Unless dir is '.', return full filenames with dir prepended.
    """
    files = _find_all_simple(dir)
    if dir == os.curdir:
        make_rel = functools.partial(os.path.relpath, start=dir)
        files = map(make_rel, files)
    return list(files)


class sic(str):
    """Treat this string as-is (https://en.wikipedia.org/wiki/Sic)"""


# Apply monkey patches
monkey.patch_all()
site-packages/setuptools/dist.py000064400000124117150564745010013037 0ustar00# -*- coding: utf-8 -*-
__all__ = ['Distribution']

import io
import sys
import re
import os
import warnings
import numbers
import distutils.log
import distutils.core
import distutils.cmd
import distutils.dist
import distutils.command
from distutils.util import strtobool
from distutils.debug import DEBUG
from distutils.fancy_getopt import translate_longopt
from glob import iglob
import itertools
import textwrap
from typing import List, Optional, TYPE_CHECKING

from collections import defaultdict
from email import message_from_file

from distutils.errors import DistutilsOptionError, DistutilsSetupError
from distutils.util import rfc822_escape
from distutils.version import StrictVersion

from setuptools.extern import packaging
from setuptools.extern import ordered_set
from setuptools.extern.more_itertools import unique_everseen

from . import SetuptoolsDeprecationWarning

import setuptools
import setuptools.command
from setuptools import windows_support
from setuptools.monkey import get_unpatched
from setuptools.config import parse_configuration
import pkg_resources

if TYPE_CHECKING:
    from email.message import Message

__import__('setuptools.extern.packaging.specifiers')
__import__('setuptools.extern.packaging.version')


def _get_unpatched(cls):
    warnings.warn("Do not call this function", DistDeprecationWarning)
    return get_unpatched(cls)


def get_metadata_version(self):
    mv = getattr(self, 'metadata_version', None)
    if mv is None:
        mv = StrictVersion('2.1')
        self.metadata_version = mv
    return mv


def rfc822_unescape(content: str) -> str:
    """Reverse RFC-822 escaping by removing leading whitespaces from content."""
    lines = content.splitlines()
    if len(lines) == 1:
        return lines[0].lstrip()
    return '\n'.join((lines[0].lstrip(), textwrap.dedent('\n'.join(lines[1:]))))


def _read_field_from_msg(msg: "Message", field: str) -> Optional[str]:
    """Read Message header field."""
    value = msg[field]
    if value == 'UNKNOWN':
        return None
    return value


def _read_field_unescaped_from_msg(msg: "Message", field: str) -> Optional[str]:
    """Read Message header field and apply rfc822_unescape."""
    value = _read_field_from_msg(msg, field)
    if value is None:
        return value
    return rfc822_unescape(value)


def _read_list_from_msg(msg: "Message", field: str) -> Optional[List[str]]:
    """Read Message header field and return all results as list."""
    values = msg.get_all(field, None)
    if values == []:
        return None
    return values


def _read_payload_from_msg(msg: "Message") -> Optional[str]:
    value = msg.get_payload().strip()
    if value == 'UNKNOWN':
        return None
    return value


def read_pkg_file(self, file):
    """Reads the metadata values from a file object."""
    msg = message_from_file(file)

    self.metadata_version = StrictVersion(msg['metadata-version'])
    self.name = _read_field_from_msg(msg, 'name')
    self.version = _read_field_from_msg(msg, 'version')
    self.description = _read_field_from_msg(msg, 'summary')
    # we are filling author only.
    self.author = _read_field_from_msg(msg, 'author')
    self.maintainer = None
    self.author_email = _read_field_from_msg(msg, 'author-email')
    self.maintainer_email = None
    self.url = _read_field_from_msg(msg, 'home-page')
    self.license = _read_field_unescaped_from_msg(msg, 'license')

    if 'download-url' in msg:
        self.download_url = _read_field_from_msg(msg, 'download-url')
    else:
        self.download_url = None

    self.long_description = _read_field_unescaped_from_msg(msg, 'description')
    if self.long_description is None and self.metadata_version >= StrictVersion('2.1'):
        self.long_description = _read_payload_from_msg(msg)
    self.description = _read_field_from_msg(msg, 'summary')

    if 'keywords' in msg:
        self.keywords = _read_field_from_msg(msg, 'keywords').split(',')

    self.platforms = _read_list_from_msg(msg, 'platform')
    self.classifiers = _read_list_from_msg(msg, 'classifier')

    # PEP 314 - these fields only exist in 1.1
    if self.metadata_version == StrictVersion('1.1'):
        self.requires = _read_list_from_msg(msg, 'requires')
        self.provides = _read_list_from_msg(msg, 'provides')
        self.obsoletes = _read_list_from_msg(msg, 'obsoletes')
    else:
        self.requires = None
        self.provides = None
        self.obsoletes = None

    self.license_files = _read_list_from_msg(msg, 'license-file')


def single_line(val):
    # quick and dirty validation for description pypa/setuptools#1390
    if '\n' in val:
        # TODO after 2021-07-31: Replace with `raise ValueError("newlines not allowed")`
        warnings.warn("newlines not allowed and will break in the future")
        val = val.replace('\n', ' ')
    return val


# Based on Python 3.5 version
def write_pkg_file(self, file):  # noqa: C901  # is too complex (14)  # FIXME
    """Write the PKG-INFO format data to a file object."""
    version = self.get_metadata_version()

    def write_field(key, value):
        file.write("%s: %s\n" % (key, value))

    write_field('Metadata-Version', str(version))
    write_field('Name', self.get_name())
    write_field('Version', self.get_version())
    write_field('Summary', single_line(self.get_description()))
    write_field('Home-page', self.get_url())

    optional_fields = (
        ('Author', 'author'),
        ('Author-email', 'author_email'),
        ('Maintainer', 'maintainer'),
        ('Maintainer-email', 'maintainer_email'),
    )

    for field, attr in optional_fields:
        attr_val = getattr(self, attr, None)
        if attr_val is not None:
            write_field(field, attr_val)

    license = rfc822_escape(self.get_license())
    write_field('License', license)
    if self.download_url:
        write_field('Download-URL', self.download_url)
    for project_url in self.project_urls.items():
        write_field('Project-URL', '%s, %s' % project_url)

    keywords = ','.join(self.get_keywords())
    if keywords:
        write_field('Keywords', keywords)

    for platform in self.get_platforms():
        write_field('Platform', platform)

    self._write_list(file, 'Classifier', self.get_classifiers())

    # PEP 314
    self._write_list(file, 'Requires', self.get_requires())
    self._write_list(file, 'Provides', self.get_provides())
    self._write_list(file, 'Obsoletes', self.get_obsoletes())

    # Setuptools specific for PEP 345
    if hasattr(self, 'python_requires'):
        write_field('Requires-Python', self.python_requires)

    # PEP 566
    if self.long_description_content_type:
        write_field('Description-Content-Type', self.long_description_content_type)
    if self.provides_extras:
        for extra in self.provides_extras:
            write_field('Provides-Extra', extra)

    self._write_list(file, 'License-File', self.license_files or [])

    file.write("\n%s\n\n" % self.get_long_description())


sequence = tuple, list


def check_importable(dist, attr, value):
    try:
        ep = pkg_resources.EntryPoint.parse('x=' + value)
        assert not ep.extras
    except (TypeError, ValueError, AttributeError, AssertionError) as e:
        raise DistutilsSetupError(
            "%r must be importable 'module:attrs' string (got %r)" % (attr, value)
        ) from e


def assert_string_list(dist, attr, value):
    """Verify that value is a string list"""
    try:
        # verify that value is a list or tuple to exclude unordered
        # or single-use iterables
        assert isinstance(value, (list, tuple))
        # verify that elements of value are strings
        assert ''.join(value) != value
    except (TypeError, ValueError, AttributeError, AssertionError) as e:
        raise DistutilsSetupError(
            "%r must be a list of strings (got %r)" % (attr, value)
        ) from e


def check_nsp(dist, attr, value):
    """Verify that namespace packages are valid"""
    ns_packages = value
    assert_string_list(dist, attr, ns_packages)
    for nsp in ns_packages:
        if not dist.has_contents_for(nsp):
            raise DistutilsSetupError(
                "Distribution contains no modules or packages for "
                + "namespace package %r" % nsp
            )
        parent, sep, child = nsp.rpartition('.')
        if parent and parent not in ns_packages:
            distutils.log.warn(
                "WARNING: %r is declared as a package namespace, but %r"
                " is not: please correct this in setup.py",
                nsp,
                parent,
            )


def check_extras(dist, attr, value):
    """Verify that extras_require mapping is valid"""
    try:
        list(itertools.starmap(_check_extra, value.items()))
    except (TypeError, ValueError, AttributeError) as e:
        raise DistutilsSetupError(
            "'extras_require' must be a dictionary whose values are "
            "strings or lists of strings containing valid project/version "
            "requirement specifiers."
        ) from e


def _check_extra(extra, reqs):
    name, sep, marker = extra.partition(':')
    if marker and pkg_resources.invalid_marker(marker):
        raise DistutilsSetupError("Invalid environment marker: " + marker)
    list(pkg_resources.parse_requirements(reqs))


def assert_bool(dist, attr, value):
    """Verify that value is True, False, 0, or 1"""
    if bool(value) != value:
        tmpl = "{attr!r} must be a boolean value (got {value!r})"
        raise DistutilsSetupError(tmpl.format(attr=attr, value=value))


def invalid_unless_false(dist, attr, value):
    if not value:
        warnings.warn(f"{attr} is ignored.", DistDeprecationWarning)
        return
    raise DistutilsSetupError(f"{attr} is invalid.")


def check_requirements(dist, attr, value):
    """Verify that install_requires is a valid requirements list"""
    try:
        list(pkg_resources.parse_requirements(value))
        if isinstance(value, (dict, set)):
            raise TypeError("Unordered types are not allowed")
    except (TypeError, ValueError) as error:
        tmpl = (
            "{attr!r} must be a string or list of strings "
            "containing valid project/version requirement specifiers; {error}"
        )
        raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) from error


def check_specifier(dist, attr, value):
    """Verify that value is a valid version specifier"""
    try:
        packaging.specifiers.SpecifierSet(value)
    except (packaging.specifiers.InvalidSpecifier, AttributeError) as error:
        tmpl = (
            "{attr!r} must be a string " "containing valid version specifiers; {error}"
        )
        raise DistutilsSetupError(tmpl.format(attr=attr, error=error)) from error


def check_entry_points(dist, attr, value):
    """Verify that entry_points map is parseable"""
    try:
        pkg_resources.EntryPoint.parse_map(value)
    except ValueError as e:
        raise DistutilsSetupError(e) from e


def check_test_suite(dist, attr, value):
    if not isinstance(value, str):
        raise DistutilsSetupError("test_suite must be a string")


def check_package_data(dist, attr, value):
    """Verify that value is a dictionary of package names to glob lists"""
    if not isinstance(value, dict):
        raise DistutilsSetupError(
            "{!r} must be a dictionary mapping package names to lists of "
            "string wildcard patterns".format(attr)
        )
    for k, v in value.items():
        if not isinstance(k, str):
            raise DistutilsSetupError(
                "keys of {!r} dict must be strings (got {!r})".format(attr, k)
            )
        assert_string_list(dist, 'values of {!r} dict'.format(attr), v)


def check_packages(dist, attr, value):
    for pkgname in value:
        if not re.match(r'\w+(\.\w+)*', pkgname):
            distutils.log.warn(
                "WARNING: %r not a valid package name; please use only "
                ".-separated package names in setup.py",
                pkgname,
            )


_Distribution = get_unpatched(distutils.core.Distribution)


class Distribution(_Distribution):
    """Distribution with support for tests and package data

    This is an enhanced version of 'distutils.dist.Distribution' that
    effectively adds the following new optional keyword arguments to 'setup()':

     'install_requires' -- a string or sequence of strings specifying project
        versions that the distribution requires when installed, in the format
        used by 'pkg_resources.require()'.  They will be installed
        automatically when the package is installed.  If you wish to use
        packages that are not available in PyPI, or want to give your users an
        alternate download location, you can add a 'find_links' option to the
        '[easy_install]' section of your project's 'setup.cfg' file, and then
        setuptools will scan the listed web pages for links that satisfy the
        requirements.

     'extras_require' -- a dictionary mapping names of optional "extras" to the
        additional requirement(s) that using those extras incurs. For example,
        this::

            extras_require = dict(reST = ["docutils>=0.3", "reSTedit"])

        indicates that the distribution can optionally provide an extra
        capability called "reST", but it can only be used if docutils and
        reSTedit are installed.  If the user installs your package using
        EasyInstall and requests one of your extras, the corresponding
        additional requirements will be installed if needed.

     'test_suite' -- the name of a test suite to run for the 'test' command.
        If the user runs 'python setup.py test', the package will be installed,
        and the named test suite will be run.  The format is the same as
        would be used on a 'unittest.py' command line.  That is, it is the
        dotted name of an object to import and call to generate a test suite.

     'package_data' -- a dictionary mapping package names to lists of filenames
        or globs to use to find data files contained in the named packages.
        If the dictionary has filenames or globs listed under '""' (the empty
        string), those names will be searched for in every package, in addition
        to any names for the specific package.  Data files found using these
        names/globs will be installed along with the package, in the same
        location as the package.  Note that globs are allowed to reference
        the contents of non-package subdirectories, as long as you use '/' as
        a path separator.  (Globs are automatically converted to
        platform-specific paths at runtime.)

    In addition to these new keywords, this class also has several new methods
    for manipulating the distribution's contents.  For example, the 'include()'
    and 'exclude()' methods can be thought of as in-place add and subtract
    commands that add or remove packages, modules, extensions, and so on from
    the distribution.
    """

    _DISTUTILS_UNSUPPORTED_METADATA = {
        'long_description_content_type': lambda: None,
        'project_urls': dict,
        'provides_extras': ordered_set.OrderedSet,
        'license_file': lambda: None,
        'license_files': lambda: None,
    }

    _patched_dist = None

    def patch_missing_pkg_info(self, attrs):
        # Fake up a replacement for the data that would normally come from
        # PKG-INFO, but which might not yet be built if this is a fresh
        # checkout.
        #
        if not attrs or 'name' not in attrs or 'version' not in attrs:
            return
        key = pkg_resources.safe_name(str(attrs['name'])).lower()
        dist = pkg_resources.working_set.by_key.get(key)
        if dist is not None and not dist.has_metadata('PKG-INFO'):
            dist._version = pkg_resources.safe_version(str(attrs['version']))
            self._patched_dist = dist

    def __init__(self, attrs=None):
        have_package_data = hasattr(self, "package_data")
        if not have_package_data:
            self.package_data = {}
        attrs = attrs or {}
        self.dist_files = []
        # Filter-out setuptools' specific options.
        self.src_root = attrs.pop("src_root", None)
        self.patch_missing_pkg_info(attrs)
        self.dependency_links = attrs.pop('dependency_links', [])
        self.setup_requires = attrs.pop('setup_requires', [])
        for ep in pkg_resources.iter_entry_points('distutils.setup_keywords'):
            vars(self).setdefault(ep.name, None)
        _Distribution.__init__(
            self,
            {
                k: v
                for k, v in attrs.items()
                if k not in self._DISTUTILS_UNSUPPORTED_METADATA
            },
        )

        self._set_metadata_defaults(attrs)

        self.metadata.version = self._normalize_version(
            self._validate_version(self.metadata.version)
        )
        self._finalize_requires()

    def _set_metadata_defaults(self, attrs):
        """
        Fill-in missing metadata fields not supported by distutils.
        Some fields may have been set by other tools (e.g. pbr).
        Those fields (vars(self.metadata)) take precedence to
        supplied attrs.
        """
        for option, default in self._DISTUTILS_UNSUPPORTED_METADATA.items():
            vars(self.metadata).setdefault(option, attrs.get(option, default()))

    @staticmethod
    def _normalize_version(version):
        if isinstance(version, setuptools.sic) or version is None:
            return version

        normalized = str(packaging.version.Version(version))
        if version != normalized:
            tmpl = "Normalizing '{version}' to '{normalized}'"
            warnings.warn(tmpl.format(**locals()))
            return normalized
        return version

    @staticmethod
    def _validate_version(version):
        if isinstance(version, numbers.Number):
            # Some people apparently take "version number" too literally :)
            version = str(version)

        if version is not None:
            try:
                packaging.version.Version(version)
            except (packaging.version.InvalidVersion, TypeError):
                warnings.warn(
                    "The version specified (%r) is an invalid version, this "
                    "may not work as expected with newer versions of "
                    "setuptools, pip, and PyPI. Please see PEP 440 for more "
                    "details." % version
                )
                return setuptools.sic(version)
        return version

    def _finalize_requires(self):
        """
        Set `metadata.python_requires` and fix environment markers
        in `install_requires` and `extras_require`.
        """
        if getattr(self, 'python_requires', None):
            self.metadata.python_requires = self.python_requires

        if getattr(self, 'extras_require', None):
            for extra in self.extras_require.keys():
                # Since this gets called multiple times at points where the
                # keys have become 'converted' extras, ensure that we are only
                # truly adding extras we haven't seen before here.
                extra = extra.split(':')[0]
                if extra:
                    self.metadata.provides_extras.add(extra)

        self._convert_extras_requirements()
        self._move_install_requirements_markers()

    def _convert_extras_requirements(self):
        """
        Convert requirements in `extras_require` of the form
        `"extra": ["barbazquux; {marker}"]` to
        `"extra:{marker}": ["barbazquux"]`.
        """
        spec_ext_reqs = getattr(self, 'extras_require', None) or {}
        self._tmp_extras_require = defaultdict(list)
        for section, v in spec_ext_reqs.items():
            # Do not strip empty sections.
            self._tmp_extras_require[section]
            for r in pkg_resources.parse_requirements(v):
                suffix = self._suffix_for(r)
                self._tmp_extras_require[section + suffix].append(r)

    @staticmethod
    def _suffix_for(req):
        """
        For a requirement, return the 'extras_require' suffix for
        that requirement.
        """
        return ':' + str(req.marker) if req.marker else ''

    def _move_install_requirements_markers(self):
        """
        Move requirements in `install_requires` that are using environment
        markers `extras_require`.
        """

        # divide the install_requires into two sets, simple ones still
        # handled by install_requires and more complex ones handled
        # by extras_require.

        def is_simple_req(req):
            return not req.marker

        spec_inst_reqs = getattr(self, 'install_requires', None) or ()
        inst_reqs = list(pkg_resources.parse_requirements(spec_inst_reqs))
        simple_reqs = filter(is_simple_req, inst_reqs)
        complex_reqs = itertools.filterfalse(is_simple_req, inst_reqs)
        self.install_requires = list(map(str, simple_reqs))

        for r in complex_reqs:
            self._tmp_extras_require[':' + str(r.marker)].append(r)
        self.extras_require = dict(
            (k, [str(r) for r in map(self._clean_req, v)])
            for k, v in self._tmp_extras_require.items()
        )

    def _clean_req(self, req):
        """
        Given a Requirement, remove environment markers and return it.
        """
        req.marker = None
        return req

    def _finalize_license_files(self):
        """Compute names of all license files which should be included."""
        license_files: Optional[List[str]] = self.metadata.license_files
        patterns: List[str] = license_files if license_files else []

        license_file: Optional[str] = self.metadata.license_file
        if license_file and license_file not in patterns:
            patterns.append(license_file)

        if license_files is None and license_file is None:
            # Default patterns match the ones wheel uses
            # See https://wheel.readthedocs.io/en/stable/user_guide.html
            # -> 'Including license files in the generated wheel file'
            patterns = ('LICEN[CS]E*', 'COPYING*', 'NOTICE*', 'AUTHORS*')

        self.metadata.license_files = list(
            unique_everseen(self._expand_patterns(patterns))
        )

    @staticmethod
    def _expand_patterns(patterns):
        """
        >>> list(Distribution._expand_patterns(['LICENSE']))
        ['LICENSE']
        >>> list(Distribution._expand_patterns(['setup.cfg', 'LIC*']))
        ['setup.cfg', 'LICENSE']
        """
        return (
            path
            for pattern in patterns
            for path in sorted(iglob(pattern))
            if not path.endswith('~') and os.path.isfile(path)
        )

    # FIXME: 'Distribution._parse_config_files' is too complex (14)
    def _parse_config_files(self, filenames=None):  # noqa: C901
        """
        Adapted from distutils.dist.Distribution.parse_config_files,
        this method provides the same functionality in subtly-improved
        ways.
        """
        from configparser import ConfigParser

        # Ignore install directory options if we have a venv
        ignore_options = (
            []
            if sys.prefix == sys.base_prefix
            else [
                'install-base',
                'install-platbase',
                'install-lib',
                'install-platlib',
                'install-purelib',
                'install-headers',
                'install-scripts',
                'install-data',
                'prefix',
                'exec-prefix',
                'home',
                'user',
                'root',
            ]
        )

        ignore_options = frozenset(ignore_options)

        if filenames is None:
            filenames = self.find_config_files()

        if DEBUG:
            self.announce("Distribution.parse_config_files():")

        parser = ConfigParser()
        parser.optionxform = str
        for filename in filenames:
            with io.open(filename, encoding='utf-8') as reader:
                if DEBUG:
                    self.announce("  reading {filename}".format(**locals()))
                parser.read_file(reader)
            for section in parser.sections():
                options = parser.options(section)
                opt_dict = self.get_option_dict(section)

                for opt in options:
                    if opt == '__name__' or opt in ignore_options:
                        continue

                    val = parser.get(section, opt)
                    opt = self.warn_dash_deprecation(opt, section)
                    opt = self.make_option_lowercase(opt, section)
                    opt_dict[opt] = (filename, val)

            # Make the ConfigParser forget everything (so we retain
            # the original filenames that options come from)
            parser.__init__()

        if 'global' not in self.command_options:
            return

        # If there was a "global" section in the config file, use it
        # to set Distribution options.

        for (opt, (src, val)) in self.command_options['global'].items():
            alias = self.negative_opt.get(opt)
            if alias:
                val = not strtobool(val)
            elif opt in ('verbose', 'dry_run'):  # ugh!
                val = strtobool(val)

            try:
                setattr(self, alias or opt, val)
            except ValueError as e:
                raise DistutilsOptionError(e) from e

    def warn_dash_deprecation(self, opt, section):
        if section in (
            'options.extras_require',
            'options.data_files',
        ):
            return opt

        underscore_opt = opt.replace('-', '_')
        commands = distutils.command.__all__ + self._setuptools_commands()
        if (
            not section.startswith('options')
            and section != 'metadata'
            and section not in commands
        ):
            return underscore_opt

        if '-' in opt:
            warnings.warn(
                "Usage of dash-separated '%s' will not be supported in future "
                "versions. Please use the underscore name '%s' instead"
                % (opt, underscore_opt)
            )
        return underscore_opt

    def _setuptools_commands(self):
        try:
            dist = pkg_resources.get_distribution('setuptools')
            return list(dist.get_entry_map('distutils.commands'))
        except pkg_resources.DistributionNotFound:
            # during bootstrapping, distribution doesn't exist
            return []

    def make_option_lowercase(self, opt, section):
        if section != 'metadata' or opt.islower():
            return opt

        lowercase_opt = opt.lower()
        warnings.warn(
            "Usage of uppercase key '%s' in '%s' will be deprecated in future "
            "versions. Please use lowercase '%s' instead"
            % (opt, section, lowercase_opt)
        )
        return lowercase_opt

    # FIXME: 'Distribution._set_command_options' is too complex (14)
    def _set_command_options(self, command_obj, option_dict=None):  # noqa: C901
        """
        Set the options for 'command_obj' from 'option_dict'.  Basically
        this means copying elements of a dictionary ('option_dict') to
        attributes of an instance ('command').

        'command_obj' must be a Command instance.  If 'option_dict' is not
        supplied, uses the standard option dictionary for this command
        (from 'self.command_options').

        (Adopted from distutils.dist.Distribution._set_command_options)
        """
        command_name = command_obj.get_command_name()
        if option_dict is None:
            option_dict = self.get_option_dict(command_name)

        if DEBUG:
            self.announce("  setting options for '%s' command:" % command_name)
        for (option, (source, value)) in option_dict.items():
            if DEBUG:
                self.announce("    %s = %s (from %s)" % (option, value, source))
            try:
                bool_opts = [translate_longopt(o) for o in command_obj.boolean_options]
            except AttributeError:
                bool_opts = []
            try:
                neg_opt = command_obj.negative_opt
            except AttributeError:
                neg_opt = {}

            try:
                is_string = isinstance(value, str)
                if option in neg_opt and is_string:
                    setattr(command_obj, neg_opt[option], not strtobool(value))
                elif option in bool_opts and is_string:
                    setattr(command_obj, option, strtobool(value))
                elif hasattr(command_obj, option):
                    setattr(command_obj, option, value)
                else:
                    raise DistutilsOptionError(
                        "error in %s: command '%s' has no such option '%s'"
                        % (source, command_name, option)
                    )
            except ValueError as e:
                raise DistutilsOptionError(e) from e

    def parse_config_files(self, filenames=None, ignore_option_errors=False):
        """Parses configuration files from various levels
        and loads configuration.

        """
        self._parse_config_files(filenames=filenames)

        parse_configuration(
            self, self.command_options, ignore_option_errors=ignore_option_errors
        )
        self._finalize_requires()
        self._finalize_license_files()

    def fetch_build_eggs(self, requires):
        """Resolve pre-setup requirements"""
        resolved_dists = pkg_resources.working_set.resolve(
            pkg_resources.parse_requirements(requires),
            installer=self.fetch_build_egg,
            replace_conflicting=True,
        )
        for dist in resolved_dists:
            pkg_resources.working_set.add(dist, replace=True)
        return resolved_dists

    def finalize_options(self):
        """
        Allow plugins to apply arbitrary operations to the
        distribution. Each hook may optionally define a 'order'
        to influence the order of execution. Smaller numbers
        go first and the default is 0.
        """
        group = 'setuptools.finalize_distribution_options'

        def by_order(hook):
            return getattr(hook, 'order', 0)

        defined = pkg_resources.iter_entry_points(group)
        filtered = itertools.filterfalse(self._removed, defined)
        loaded = map(lambda e: e.load(), filtered)
        for ep in sorted(loaded, key=by_order):
            ep(self)

    @staticmethod
    def _removed(ep):
        """
        When removing an entry point, if metadata is loaded
        from an older version of Setuptools, that removed
        entry point will attempt to be loaded and will fail.
        See #2765 for more details.
        """
        removed = {
            # removed 2021-09-05
            '2to3_doctests',
        }
        return ep.name in removed

    def _finalize_setup_keywords(self):
        for ep in pkg_resources.iter_entry_points('distutils.setup_keywords'):
            value = getattr(self, ep.name, None)
            if value is not None:
                ep.require(installer=self.fetch_build_egg)
                ep.load()(self, ep.name, value)

    def get_egg_cache_dir(self):
        egg_cache_dir = os.path.join(os.curdir, '.eggs')
        if not os.path.exists(egg_cache_dir):
            os.mkdir(egg_cache_dir)
            windows_support.hide_file(egg_cache_dir)
            readme_txt_filename = os.path.join(egg_cache_dir, 'README.txt')
            with open(readme_txt_filename, 'w') as f:
                f.write(
                    'This directory contains eggs that were downloaded '
                    'by setuptools to build, test, and run plug-ins.\n\n'
                )
                f.write(
                    'This directory caches those eggs to prevent '
                    'repeated downloads.\n\n'
                )
                f.write('However, it is safe to delete this directory.\n\n')

        return egg_cache_dir

    def fetch_build_egg(self, req):
        """Fetch an egg needed for building"""
        from setuptools.installer import fetch_build_egg

        return fetch_build_egg(self, req)

    def get_command_class(self, command):
        """Pluggable version of get_command_class()"""
        if command in self.cmdclass:
            return self.cmdclass[command]

        eps = pkg_resources.iter_entry_points('distutils.commands', command)
        for ep in eps:
            ep.require(installer=self.fetch_build_egg)
            self.cmdclass[command] = cmdclass = ep.load()
            return cmdclass
        else:
            return _Distribution.get_command_class(self, command)

    def print_commands(self):
        for ep in pkg_resources.iter_entry_points('distutils.commands'):
            if ep.name not in self.cmdclass:
                # don't require extras as the commands won't be invoked
                cmdclass = ep.resolve()
                self.cmdclass[ep.name] = cmdclass
        return _Distribution.print_commands(self)

    def get_command_list(self):
        for ep in pkg_resources.iter_entry_points('distutils.commands'):
            if ep.name not in self.cmdclass:
                # don't require extras as the commands won't be invoked
                cmdclass = ep.resolve()
                self.cmdclass[ep.name] = cmdclass
        return _Distribution.get_command_list(self)

    def include(self, **attrs):
        """Add items to distribution that are named in keyword arguments

        For example, 'dist.include(py_modules=["x"])' would add 'x' to
        the distribution's 'py_modules' attribute, if it was not already
        there.

        Currently, this method only supports inclusion for attributes that are
        lists or tuples.  If you need to add support for adding to other
        attributes in this or a subclass, you can add an '_include_X' method,
        where 'X' is the name of the attribute.  The method will be called with
        the value passed to 'include()'.  So, 'dist.include(foo={"bar":"baz"})'
        will try to call 'dist._include_foo({"bar":"baz"})', which can then
        handle whatever special inclusion logic is needed.
        """
        for k, v in attrs.items():
            include = getattr(self, '_include_' + k, None)
            if include:
                include(v)
            else:
                self._include_misc(k, v)

    def exclude_package(self, package):
        """Remove packages, modules, and extensions in named package"""

        pfx = package + '.'
        if self.packages:
            self.packages = [
                p for p in self.packages if p != package and not p.startswith(pfx)
            ]

        if self.py_modules:
            self.py_modules = [
                p for p in self.py_modules if p != package and not p.startswith(pfx)
            ]

        if self.ext_modules:
            self.ext_modules = [
                p
                for p in self.ext_modules
                if p.name != package and not p.name.startswith(pfx)
            ]

    def has_contents_for(self, package):
        """Return true if 'exclude_package(package)' would do something"""

        pfx = package + '.'

        for p in self.iter_distribution_names():
            if p == package or p.startswith(pfx):
                return True

    def _exclude_misc(self, name, value):
        """Handle 'exclude()' for list/tuple attrs without a special handler"""
        if not isinstance(value, sequence):
            raise DistutilsSetupError(
                "%s: setting must be a list or tuple (%r)" % (name, value)
            )
        try:
            old = getattr(self, name)
        except AttributeError as e:
            raise DistutilsSetupError("%s: No such distribution setting" % name) from e
        if old is not None and not isinstance(old, sequence):
            raise DistutilsSetupError(
                name + ": this setting cannot be changed via include/exclude"
            )
        elif old:
            setattr(self, name, [item for item in old if item not in value])

    def _include_misc(self, name, value):
        """Handle 'include()' for list/tuple attrs without a special handler"""

        if not isinstance(value, sequence):
            raise DistutilsSetupError("%s: setting must be a list (%r)" % (name, value))
        try:
            old = getattr(self, name)
        except AttributeError as e:
            raise DistutilsSetupError("%s: No such distribution setting" % name) from e
        if old is None:
            setattr(self, name, value)
        elif not isinstance(old, sequence):
            raise DistutilsSetupError(
                name + ": this setting cannot be changed via include/exclude"
            )
        else:
            new = [item for item in value if item not in old]
            setattr(self, name, old + new)

    def exclude(self, **attrs):
        """Remove items from distribution that are named in keyword arguments

        For example, 'dist.exclude(py_modules=["x"])' would remove 'x' from
        the distribution's 'py_modules' attribute.  Excluding packages uses
        the 'exclude_package()' method, so all of the package's contained
        packages, modules, and extensions are also excluded.

        Currently, this method only supports exclusion from attributes that are
        lists or tuples.  If you need to add support for excluding from other
        attributes in this or a subclass, you can add an '_exclude_X' method,
        where 'X' is the name of the attribute.  The method will be called with
        the value passed to 'exclude()'.  So, 'dist.exclude(foo={"bar":"baz"})'
        will try to call 'dist._exclude_foo({"bar":"baz"})', which can then
        handle whatever special exclusion logic is needed.
        """
        for k, v in attrs.items():
            exclude = getattr(self, '_exclude_' + k, None)
            if exclude:
                exclude(v)
            else:
                self._exclude_misc(k, v)

    def _exclude_packages(self, packages):
        if not isinstance(packages, sequence):
            raise DistutilsSetupError(
                "packages: setting must be a list or tuple (%r)" % (packages,)
            )
        list(map(self.exclude_package, packages))

    def _parse_command_opts(self, parser, args):
        # Remove --with-X/--without-X options when processing command args
        self.global_options = self.__class__.global_options
        self.negative_opt = self.__class__.negative_opt

        # First, expand any aliases
        command = args[0]
        aliases = self.get_option_dict('aliases')
        while command in aliases:
            src, alias = aliases[command]
            del aliases[command]  # ensure each alias can expand only once!
            import shlex

            args[:1] = shlex.split(alias, True)
            command = args[0]

        nargs = _Distribution._parse_command_opts(self, parser, args)

        # Handle commands that want to consume all remaining arguments
        cmd_class = self.get_command_class(command)
        if getattr(cmd_class, 'command_consumes_arguments', None):
            self.get_option_dict(command)['args'] = ("command line", nargs)
            if nargs is not None:
                return []

        return nargs

    def get_cmdline_options(self):
        """Return a '{cmd: {opt:val}}' map of all command-line options

        Option names are all long, but do not include the leading '--', and
        contain dashes rather than underscores.  If the option doesn't take
        an argument (e.g. '--quiet'), the 'val' is 'None'.

        Note that options provided by config files are intentionally excluded.
        """

        d = {}

        for cmd, opts in self.command_options.items():

            for opt, (src, val) in opts.items():

                if src != "command line":
                    continue

                opt = opt.replace('_', '-')

                if val == 0:
                    cmdobj = self.get_command_obj(cmd)
                    neg_opt = self.negative_opt.copy()
                    neg_opt.update(getattr(cmdobj, 'negative_opt', {}))
                    for neg, pos in neg_opt.items():
                        if pos == opt:
                            opt = neg
                            val = None
                            break
                    else:
                        raise AssertionError("Shouldn't be able to get here")

                elif val == 1:
                    val = None

                d.setdefault(cmd, {})[opt] = val

        return d

    def iter_distribution_names(self):
        """Yield all packages, modules, and extension names in distribution"""

        for pkg in self.packages or ():
            yield pkg

        for module in self.py_modules or ():
            yield module

        for ext in self.ext_modules or ():
            if isinstance(ext, tuple):
                name, buildinfo = ext
            else:
                name = ext.name
            if name.endswith('module'):
                name = name[:-6]
            yield name

    def handle_display_options(self, option_order):
        """If there were any non-global "display-only" options
        (--help-commands or the metadata display options) on the command
        line, display the requested info and return true; else return
        false.
        """
        import sys

        if self.help_commands:
            return _Distribution.handle_display_options(self, option_order)

        # Stdout may be StringIO (e.g. in tests)
        if not isinstance(sys.stdout, io.TextIOWrapper):
            return _Distribution.handle_display_options(self, option_order)

        # Don't wrap stdout if utf-8 is already the encoding. Provides
        #  workaround for #334.
        if sys.stdout.encoding.lower() in ('utf-8', 'utf8'):
            return _Distribution.handle_display_options(self, option_order)

        # Print metadata in UTF-8 no matter the platform
        encoding = sys.stdout.encoding
        errors = sys.stdout.errors
        newline = sys.platform != 'win32' and '\n' or None
        line_buffering = sys.stdout.line_buffering

        sys.stdout = io.TextIOWrapper(
            sys.stdout.detach(), 'utf-8', errors, newline, line_buffering
        )
        try:
            return _Distribution.handle_display_options(self, option_order)
        finally:
            sys.stdout = io.TextIOWrapper(
                sys.stdout.detach(), encoding, errors, newline, line_buffering
            )


class DistDeprecationWarning(SetuptoolsDeprecationWarning):
    """Class for warning about deprecations in dist in
    setuptools. Not ignored by default, unlike DeprecationWarning."""
site-packages/setuptools/windows_support.py000064400000001312150564745010015351 0ustar00import platform
import ctypes


def windows_only(func):
    if platform.system() != 'Windows':
        return lambda *args, **kwargs: None
    return func


@windows_only
def hide_file(path):
    """
    Set the hidden attribute on a file or directory.

    From http://stackoverflow.com/questions/19622133/

    `path` must be text.
    """
    __import__('ctypes.wintypes')
    SetFileAttributes = ctypes.windll.kernel32.SetFileAttributesW
    SetFileAttributes.argtypes = ctypes.wintypes.LPWSTR, ctypes.wintypes.DWORD
    SetFileAttributes.restype = ctypes.wintypes.BOOL

    FILE_ATTRIBUTE_HIDDEN = 0x02

    ret = SetFileAttributes(path, FILE_ATTRIBUTE_HIDDEN)
    if not ret:
        raise ctypes.WinError()
site-packages/setuptools/config.py000064400000055123150564745010013341 0ustar00import ast
import io
import os
import sys

import warnings
import functools
import importlib
from collections import defaultdict
from functools import partial
from functools import wraps
from glob import iglob
import contextlib

from distutils.errors import DistutilsOptionError, DistutilsFileError
from setuptools.extern.packaging.version import LegacyVersion, parse
from setuptools.extern.packaging.specifiers import SpecifierSet


class StaticModule:
    """
    Attempt to load the module by the name
    """

    def __init__(self, name):
        spec = importlib.util.find_spec(name)
        with open(spec.origin) as strm:
            src = strm.read()
        module = ast.parse(src)
        vars(self).update(locals())
        del self.self

    def __getattr__(self, attr):
        try:
            return next(
                ast.literal_eval(statement.value)
                for statement in self.module.body
                if isinstance(statement, ast.Assign)
                for target in statement.targets
                if isinstance(target, ast.Name) and target.id == attr
            )
        except Exception as e:
            raise AttributeError(
                "{self.name} has no attribute {attr}".format(**locals())
            ) from e


@contextlib.contextmanager
def patch_path(path):
    """
    Add path to front of sys.path for the duration of the context.
    """
    try:
        sys.path.insert(0, path)
        yield
    finally:
        sys.path.remove(path)


def read_configuration(filepath, find_others=False, ignore_option_errors=False):
    """Read given configuration file and returns options from it as a dict.

    :param str|unicode filepath: Path to configuration file
        to get options from.

    :param bool find_others: Whether to search for other configuration files
        which could be on in various places.

    :param bool ignore_option_errors: Whether to silently ignore
        options, values of which could not be resolved (e.g. due to exceptions
        in directives such as file:, attr:, etc.).
        If False exceptions are propagated as expected.

    :rtype: dict
    """
    from setuptools.dist import Distribution, _Distribution

    filepath = os.path.abspath(filepath)

    if not os.path.isfile(filepath):
        raise DistutilsFileError('Configuration file %s does not exist.' % filepath)

    current_directory = os.getcwd()
    os.chdir(os.path.dirname(filepath))

    try:
        dist = Distribution()

        filenames = dist.find_config_files() if find_others else []
        if filepath not in filenames:
            filenames.append(filepath)

        _Distribution.parse_config_files(dist, filenames=filenames)

        handlers = parse_configuration(
            dist, dist.command_options, ignore_option_errors=ignore_option_errors
        )

    finally:
        os.chdir(current_directory)

    return configuration_to_dict(handlers)


def _get_option(target_obj, key):
    """
    Given a target object and option key, get that option from
    the target object, either through a get_{key} method or
    from an attribute directly.
    """
    getter_name = 'get_{key}'.format(**locals())
    by_attribute = functools.partial(getattr, target_obj, key)
    getter = getattr(target_obj, getter_name, by_attribute)
    return getter()


def configuration_to_dict(handlers):
    """Returns configuration data gathered by given handlers as a dict.

    :param list[ConfigHandler] handlers: Handlers list,
        usually from parse_configuration()

    :rtype: dict
    """
    config_dict = defaultdict(dict)

    for handler in handlers:
        for option in handler.set_options:
            value = _get_option(handler.target_obj, option)
            config_dict[handler.section_prefix][option] = value

    return config_dict


def parse_configuration(distribution, command_options, ignore_option_errors=False):
    """Performs additional parsing of configuration options
    for a distribution.

    Returns a list of used option handlers.

    :param Distribution distribution:
    :param dict command_options:
    :param bool ignore_option_errors: Whether to silently ignore
        options, values of which could not be resolved (e.g. due to exceptions
        in directives such as file:, attr:, etc.).
        If False exceptions are propagated as expected.
    :rtype: list
    """
    options = ConfigOptionsHandler(distribution, command_options, ignore_option_errors)
    options.parse()

    meta = ConfigMetadataHandler(
        distribution.metadata,
        command_options,
        ignore_option_errors,
        distribution.package_dir,
    )
    meta.parse()

    return meta, options


class ConfigHandler:
    """Handles metadata supplied in configuration files."""

    section_prefix = None
    """Prefix for config sections handled by this handler.
    Must be provided by class heirs.

    """

    aliases = {}
    """Options aliases.
    For compatibility with various packages. E.g.: d2to1 and pbr.
    Note: `-` in keys is replaced with `_` by config parser.

    """

    def __init__(self, target_obj, options, ignore_option_errors=False):
        sections = {}

        section_prefix = self.section_prefix
        for section_name, section_options in options.items():
            if not section_name.startswith(section_prefix):
                continue

            section_name = section_name.replace(section_prefix, '').strip('.')
            sections[section_name] = section_options

        self.ignore_option_errors = ignore_option_errors
        self.target_obj = target_obj
        self.sections = sections
        self.set_options = []

    @property
    def parsers(self):
        """Metadata item name to parser function mapping."""
        raise NotImplementedError(
            '%s must provide .parsers property' % self.__class__.__name__
        )

    def __setitem__(self, option_name, value):
        unknown = tuple()
        target_obj = self.target_obj

        # Translate alias into real name.
        option_name = self.aliases.get(option_name, option_name)

        current_value = getattr(target_obj, option_name, unknown)

        if current_value is unknown:
            raise KeyError(option_name)

        if current_value:
            # Already inhabited. Skipping.
            return

        skip_option = False
        parser = self.parsers.get(option_name)
        if parser:
            try:
                value = parser(value)

            except Exception:
                skip_option = True
                if not self.ignore_option_errors:
                    raise

        if skip_option:
            return

        setter = getattr(target_obj, 'set_%s' % option_name, None)
        if setter is None:
            setattr(target_obj, option_name, value)
        else:
            setter(value)

        self.set_options.append(option_name)

    @classmethod
    def _parse_list(cls, value, separator=','):
        """Represents value as a list.

        Value is split either by separator (defaults to comma) or by lines.

        :param value:
        :param separator: List items separator character.
        :rtype: list
        """
        if isinstance(value, list):  # _get_parser_compound case
            return value

        if '\n' in value:
            value = value.splitlines()
        else:
            value = value.split(separator)

        return [chunk.strip() for chunk in value if chunk.strip()]

    @classmethod
    def _parse_list_glob(cls, value, separator=','):
        """Equivalent to _parse_list() but expands any glob patterns using glob().

        However, unlike with glob() calls, the results remain relative paths.

        :param value:
        :param separator: List items separator character.
        :rtype: list
        """
        glob_characters = ('*', '?', '[', ']', '{', '}')
        values = cls._parse_list(value, separator=separator)
        expanded_values = []
        for value in values:

            # Has globby characters?
            if any(char in value for char in glob_characters):
                # then expand the glob pattern while keeping paths *relative*:
                expanded_values.extend(sorted(
                    os.path.relpath(path, os.getcwd())
                    for path in iglob(os.path.abspath(value))))

            else:
                # take the value as-is:
                expanded_values.append(value)

        return expanded_values

    @classmethod
    def _parse_dict(cls, value):
        """Represents value as a dict.

        :param value:
        :rtype: dict
        """
        separator = '='
        result = {}
        for line in cls._parse_list(value):
            key, sep, val = line.partition(separator)
            if sep != separator:
                raise DistutilsOptionError(
                    'Unable to parse option value to dict: %s' % value
                )
            result[key.strip()] = val.strip()

        return result

    @classmethod
    def _parse_bool(cls, value):
        """Represents value as boolean.

        :param value:
        :rtype: bool
        """
        value = value.lower()
        return value in ('1', 'true', 'yes')

    @classmethod
    def _exclude_files_parser(cls, key):
        """Returns a parser function to make sure field inputs
        are not files.

        Parses a value after getting the key so error messages are
        more informative.

        :param key:
        :rtype: callable
        """

        def parser(value):
            exclude_directive = 'file:'
            if value.startswith(exclude_directive):
                raise ValueError(
                    'Only strings are accepted for the {0} field, '
                    'files are not accepted'.format(key)
                )
            return value

        return parser

    @classmethod
    def _parse_file(cls, value):
        """Represents value as a string, allowing including text
        from nearest files using `file:` directive.

        Directive is sandboxed and won't reach anything outside
        directory with setup.py.

        Examples:
            file: README.rst, CHANGELOG.md, src/file.txt

        :param str value:
        :rtype: str
        """
        include_directive = 'file:'

        if not isinstance(value, str):
            return value

        if not value.startswith(include_directive):
            return value

        spec = value[len(include_directive) :]
        filepaths = (os.path.abspath(path.strip()) for path in spec.split(','))
        return '\n'.join(
            cls._read_file(path)
            for path in filepaths
            if (cls._assert_local(path) or True) and os.path.isfile(path)
        )

    @staticmethod
    def _assert_local(filepath):
        if not filepath.startswith(os.getcwd()):
            raise DistutilsOptionError('`file:` directive can not access %s' % filepath)

    @staticmethod
    def _read_file(filepath):
        with io.open(filepath, encoding='utf-8') as f:
            return f.read()

    @classmethod
    def _parse_attr(cls, value, package_dir=None):
        """Represents value as a module attribute.

        Examples:
            attr: package.attr
            attr: package.module.attr

        :param str value:
        :rtype: str
        """
        attr_directive = 'attr:'
        if not value.startswith(attr_directive):
            return value

        attrs_path = value.replace(attr_directive, '').strip().split('.')
        attr_name = attrs_path.pop()

        module_name = '.'.join(attrs_path)
        module_name = module_name or '__init__'

        parent_path = os.getcwd()
        if package_dir:
            if attrs_path[0] in package_dir:
                # A custom path was specified for the module we want to import
                custom_path = package_dir[attrs_path[0]]
                parts = custom_path.rsplit('/', 1)
                if len(parts) > 1:
                    parent_path = os.path.join(os.getcwd(), parts[0])
                    module_name = parts[1]
                else:
                    module_name = custom_path
            elif '' in package_dir:
                # A custom parent directory was specified for all root modules
                parent_path = os.path.join(os.getcwd(), package_dir[''])

        with patch_path(parent_path):
            try:
                # attempt to load value statically
                return getattr(StaticModule(module_name), attr_name)
            except Exception:
                # fallback to simple import
                module = importlib.import_module(module_name)

        return getattr(module, attr_name)

    @classmethod
    def _get_parser_compound(cls, *parse_methods):
        """Returns parser function to represents value as a list.

        Parses a value applying given methods one after another.

        :param parse_methods:
        :rtype: callable
        """

        def parse(value):
            parsed = value

            for method in parse_methods:
                parsed = method(parsed)

            return parsed

        return parse

    @classmethod
    def _parse_section_to_dict(cls, section_options, values_parser=None):
        """Parses section options into a dictionary.

        Optionally applies a given parser to values.

        :param dict section_options:
        :param callable values_parser:
        :rtype: dict
        """
        value = {}
        values_parser = values_parser or (lambda val: val)
        for key, (_, val) in section_options.items():
            value[key] = values_parser(val)
        return value

    def parse_section(self, section_options):
        """Parses configuration file section.

        :param dict section_options:
        """
        for (name, (_, value)) in section_options.items():
            try:
                self[name] = value

            except KeyError:
                pass  # Keep silent for a new option may appear anytime.

    def parse(self):
        """Parses configuration file items from one
        or more related sections.

        """
        for section_name, section_options in self.sections.items():

            method_postfix = ''
            if section_name:  # [section.option] variant
                method_postfix = '_%s' % section_name

            section_parser_method = getattr(
                self,
                # Dots in section names are translated into dunderscores.
                ('parse_section%s' % method_postfix).replace('.', '__'),
                None,
            )

            if section_parser_method is None:
                raise DistutilsOptionError(
                    'Unsupported distribution option section: [%s.%s]'
                    % (self.section_prefix, section_name)
                )

            section_parser_method(section_options)

    def _deprecated_config_handler(self, func, msg, warning_class):
        """this function will wrap around parameters that are deprecated

        :param msg: deprecation message
        :param warning_class: class of warning exception to be raised
        :param func: function to be wrapped around
        """

        @wraps(func)
        def config_handler(*args, **kwargs):
            warnings.warn(msg, warning_class)
            return func(*args, **kwargs)

        return config_handler


class ConfigMetadataHandler(ConfigHandler):

    section_prefix = 'metadata'

    aliases = {
        'home_page': 'url',
        'summary': 'description',
        'classifier': 'classifiers',
        'platform': 'platforms',
    }

    strict_mode = False
    """We need to keep it loose, to be partially compatible with
    `pbr` and `d2to1` packages which also uses `metadata` section.

    """

    def __init__(
        self, target_obj, options, ignore_option_errors=False, package_dir=None
    ):
        super(ConfigMetadataHandler, self).__init__(
            target_obj, options, ignore_option_errors
        )
        self.package_dir = package_dir

    @property
    def parsers(self):
        """Metadata item name to parser function mapping."""
        parse_list = self._parse_list
        parse_file = self._parse_file
        parse_dict = self._parse_dict
        exclude_files_parser = self._exclude_files_parser

        return {
            'platforms': parse_list,
            'keywords': parse_list,
            'provides': parse_list,
            'requires': self._deprecated_config_handler(
                parse_list,
                "The requires parameter is deprecated, please use "
                "install_requires for runtime dependencies.",
                DeprecationWarning,
            ),
            'obsoletes': parse_list,
            'classifiers': self._get_parser_compound(parse_file, parse_list),
            'license': exclude_files_parser('license'),
            'license_file': self._deprecated_config_handler(
                exclude_files_parser('license_file'),
                "The license_file parameter is deprecated, "
                "use license_files instead.",
                DeprecationWarning,
            ),
            'license_files': parse_list,
            'description': parse_file,
            'long_description': parse_file,
            'version': self._parse_version,
            'project_urls': parse_dict,
        }

    def _parse_version(self, value):
        """Parses `version` option value.

        :param value:
        :rtype: str

        """
        version = self._parse_file(value)

        if version != value:
            version = version.strip()
            # Be strict about versions loaded from file because it's easy to
            # accidentally include newlines and other unintended content
            if isinstance(parse(version), LegacyVersion):
                tmpl = (
                    'Version loaded from {value} does not '
                    'comply with PEP 440: {version}'
                )
                raise DistutilsOptionError(tmpl.format(**locals()))

            return version

        version = self._parse_attr(value, self.package_dir)

        if callable(version):
            version = version()

        if not isinstance(version, str):
            if hasattr(version, '__iter__'):
                version = '.'.join(map(str, version))
            else:
                version = '%s' % version

        return version


class ConfigOptionsHandler(ConfigHandler):

    section_prefix = 'options'

    @property
    def parsers(self):
        """Metadata item name to parser function mapping."""
        parse_list = self._parse_list
        parse_list_semicolon = partial(self._parse_list, separator=';')
        parse_bool = self._parse_bool
        parse_dict = self._parse_dict
        parse_cmdclass = self._parse_cmdclass

        return {
            'zip_safe': parse_bool,
            'include_package_data': parse_bool,
            'package_dir': parse_dict,
            'scripts': parse_list,
            'eager_resources': parse_list,
            'dependency_links': parse_list,
            'namespace_packages': parse_list,
            'install_requires': parse_list_semicolon,
            'setup_requires': parse_list_semicolon,
            'tests_require': parse_list_semicolon,
            'packages': self._parse_packages,
            'entry_points': self._parse_file,
            'py_modules': parse_list,
            'python_requires': SpecifierSet,
            'cmdclass': parse_cmdclass,
        }

    def _parse_cmdclass(self, value):
        def resolve_class(qualified_class_name):
            idx = qualified_class_name.rfind('.')
            class_name = qualified_class_name[idx + 1 :]
            pkg_name = qualified_class_name[:idx]

            module = __import__(pkg_name)

            return getattr(module, class_name)

        return {k: resolve_class(v) for k, v in self._parse_dict(value).items()}

    def _parse_packages(self, value):
        """Parses `packages` option value.

        :param value:
        :rtype: list
        """
        find_directives = ['find:', 'find_namespace:']
        trimmed_value = value.strip()

        if trimmed_value not in find_directives:
            return self._parse_list(value)

        findns = trimmed_value == find_directives[1]

        # Read function arguments from a dedicated section.
        find_kwargs = self.parse_section_packages__find(
            self.sections.get('packages.find', {})
        )

        if findns:
            from setuptools import find_namespace_packages as find_packages
        else:
            from setuptools import find_packages

        return find_packages(**find_kwargs)

    def parse_section_packages__find(self, section_options):
        """Parses `packages.find` configuration file section.

        To be used in conjunction with _parse_packages().

        :param dict section_options:
        """
        section_data = self._parse_section_to_dict(section_options, self._parse_list)

        valid_keys = ['where', 'include', 'exclude']

        find_kwargs = dict(
            [(k, v) for k, v in section_data.items() if k in valid_keys and v]
        )

        where = find_kwargs.get('where')
        if where is not None:
            find_kwargs['where'] = where[0]  # cast list to single val

        return find_kwargs

    def parse_section_entry_points(self, section_options):
        """Parses `entry_points` configuration file section.

        :param dict section_options:
        """
        parsed = self._parse_section_to_dict(section_options, self._parse_list)
        self['entry_points'] = parsed

    def _parse_package_data(self, section_options):
        parsed = self._parse_section_to_dict(section_options, self._parse_list)

        root = parsed.get('*')
        if root:
            parsed[''] = root
            del parsed['*']

        return parsed

    def parse_section_package_data(self, section_options):
        """Parses `package_data` configuration file section.

        :param dict section_options:
        """
        self['package_data'] = self._parse_package_data(section_options)

    def parse_section_exclude_package_data(self, section_options):
        """Parses `exclude_package_data` configuration file section.

        :param dict section_options:
        """
        self['exclude_package_data'] = self._parse_package_data(section_options)

    def parse_section_extras_require(self, section_options):
        """Parses `extras_require` configuration file section.

        :param dict section_options:
        """
        parse_list = partial(self._parse_list, separator=';')
        self['extras_require'] = self._parse_section_to_dict(
            section_options, parse_list
        )

    def parse_section_data_files(self, section_options):
        """Parses `data_files` configuration file section.

        :param dict section_options:
        """
        parsed = self._parse_section_to_dict(section_options, self._parse_list_glob)
        self['data_files'] = [(k, v) for k, v in parsed.items()]
site-packages/setuptools/py34compat.py000064400000000365150564745010014075 0ustar00import importlib

try:
    import importlib.util
except ImportError:
    pass


try:
    module_from_spec = importlib.util.module_from_spec
except AttributeError:
    def module_from_spec(spec):
        return spec.loader.load_module(spec.name)
site-packages/setuptools/namespaces.py000064400000006025150564745010014210 0ustar00import os
from distutils import log
import itertools


flatten = itertools.chain.from_iterable


class Installer:

    nspkg_ext = '-nspkg.pth'

    def install_namespaces(self):
        nsp = self._get_all_ns_packages()
        if not nsp:
            return
        filename, ext = os.path.splitext(self._get_target())
        filename += self.nspkg_ext
        self.outputs.append(filename)
        log.info("Installing %s", filename)
        lines = map(self._gen_nspkg_line, nsp)

        if self.dry_run:
            # always generate the lines, even in dry run
            list(lines)
            return

        with open(filename, 'wt') as f:
            f.writelines(lines)

    def uninstall_namespaces(self):
        filename, ext = os.path.splitext(self._get_target())
        filename += self.nspkg_ext
        if not os.path.exists(filename):
            return
        log.info("Removing %s", filename)
        os.remove(filename)

    def _get_target(self):
        return self.target

    _nspkg_tmpl = (
        "import sys, types, os",
        "has_mfs = sys.version_info > (3, 5)",
        "p = os.path.join(%(root)s, *%(pth)r)",
        "importlib = has_mfs and __import__('importlib.util')",
        "has_mfs and __import__('importlib.machinery')",
        (
            "m = has_mfs and "
            "sys.modules.setdefault(%(pkg)r, "
            "importlib.util.module_from_spec("
            "importlib.machinery.PathFinder.find_spec(%(pkg)r, "
            "[os.path.dirname(p)])))"
        ),
        (
            "m = m or "
            "sys.modules.setdefault(%(pkg)r, types.ModuleType(%(pkg)r))"
        ),
        "mp = (m or []) and m.__dict__.setdefault('__path__',[])",
        "(p not in mp) and mp.append(p)",
    )
    "lines for the namespace installer"

    _nspkg_tmpl_multi = (
        'm and setattr(sys.modules[%(parent)r], %(child)r, m)',
    )
    "additional line(s) when a parent package is indicated"

    def _get_root(self):
        return "sys._getframe(1).f_locals['sitedir']"

    def _gen_nspkg_line(self, pkg):
        pth = tuple(pkg.split('.'))
        root = self._get_root()
        tmpl_lines = self._nspkg_tmpl
        parent, sep, child = pkg.rpartition('.')
        if parent:
            tmpl_lines += self._nspkg_tmpl_multi
        return ';'.join(tmpl_lines) % locals() + '\n'

    def _get_all_ns_packages(self):
        """Return sorted list of all package namespaces"""
        pkgs = self.distribution.namespace_packages or []
        return sorted(flatten(map(self._pkg_names, pkgs)))

    @staticmethod
    def _pkg_names(pkg):
        """
        Given a namespace package, yield the components of that
        package.

        >>> names = Installer._pkg_names('a.b.c')
        >>> set(names) == set(['a', 'a.b', 'a.b.c'])
        True
        """
        parts = pkg.split('.')
        while parts:
            yield '.'.join(parts)
            parts.pop()


class DevelopInstaller(Installer):
    def _get_root(self):
        return repr(str(self.egg_path))

    def _get_target(self):
        return self.egg_link
site-packages/setuptools/extension.py000064400000003224150564745010014103 0ustar00import re
import functools
import distutils.core
import distutils.errors
import distutils.extension

from .monkey import get_unpatched


def _have_cython():
    """
    Return True if Cython can be imported.
    """
    cython_impl = 'Cython.Distutils.build_ext'
    try:
        # from (cython_impl) import build_ext
        __import__(cython_impl, fromlist=['build_ext']).build_ext
        return True
    except Exception:
        pass
    return False


# for compatibility
have_pyrex = _have_cython

_Extension = get_unpatched(distutils.core.Extension)


class Extension(_Extension):
    """Extension that uses '.c' files in place of '.pyx' files"""

    def __init__(self, name, sources, *args, **kw):
        # The *args is needed for compatibility as calls may use positional
        # arguments. py_limited_api may be set only via keyword.
        self.py_limited_api = kw.pop("py_limited_api", False)
        _Extension.__init__(self, name, sources, *args, **kw)

    def _convert_pyx_sources_to_lang(self):
        """
        Replace sources with .pyx extensions to sources with the target
        language extension. This mechanism allows language authors to supply
        pre-converted sources but to prefer the .pyx sources.
        """
        if _have_cython():
            # the build has Cython, so allow it to compile the .pyx files
            return
        lang = self.language or ''
        target_ext = '.cpp' if lang.lower() == 'c++' else '.c'
        sub = functools.partial(re.sub, '.pyx$', target_ext)
        self.sources = list(map(sub, self.sources))


class Library(Extension):
    """Just like a regular Extension, but built as a library instead"""
site-packages/setuptools/errors.py000064400000001014150564745010013376 0ustar00"""setuptools.errors

Provides exceptions used by setuptools modules.
"""

from distutils.errors import DistutilsError


class RemovedCommandError(DistutilsError, RuntimeError):
    """Error used for commands that have been removed in setuptools.

    Since ``setuptools`` is built on ``distutils``, simply removing a command
    from ``setuptools`` will make the behavior fall back to ``distutils``; this
    error is raised if a command exists in ``distutils`` but has been actively
    removed in ``setuptools``.
    """
site-packages/setuptools/package_index.py000064400000115716150564745010014663 0ustar00"""PyPI and direct package downloading"""
import sys
import os
import re
import io
import shutil
import socket
import base64
import hashlib
import itertools
import warnings
import configparser
import html
import http.client
import urllib.parse
import urllib.request
import urllib.error
from functools import wraps

import setuptools
from pkg_resources import (
    CHECKOUT_DIST, Distribution, BINARY_DIST, normalize_path, SOURCE_DIST,
    Environment, find_distributions, safe_name, safe_version,
    to_filename, Requirement, DEVELOP_DIST, EGG_DIST,
)
from distutils import log
from distutils.errors import DistutilsError
from fnmatch import translate
from setuptools.wheel import Wheel
from setuptools.extern.more_itertools import unique_everseen


EGG_FRAGMENT = re.compile(r'^egg=([-A-Za-z0-9_.+!]+)$')
HREF = re.compile(r"""href\s*=\s*['"]?([^'"> ]+)""", re.I)
PYPI_MD5 = re.compile(
    r'<a href="([^"#]+)">([^<]+)</a>\n\s+\(<a (?:title="MD5 hash"\n\s+)'
    r'href="[^?]+\?:action=show_md5&amp;digest=([0-9a-f]{32})">md5</a>\)'
)
URL_SCHEME = re.compile('([-+.a-z0-9]{2,}):', re.I).match
EXTENSIONS = ".tar.gz .tar.bz2 .tar .zip .tgz".split()

__all__ = [
    'PackageIndex', 'distros_for_url', 'parse_bdist_wininst',
    'interpret_distro_name',
]

_SOCKET_TIMEOUT = 15

_tmpl = "setuptools/{setuptools.__version__} Python-urllib/{py_major}"
user_agent = _tmpl.format(
    py_major='{}.{}'.format(*sys.version_info), setuptools=setuptools)


def parse_requirement_arg(spec):
    try:
        return Requirement.parse(spec)
    except ValueError as e:
        raise DistutilsError(
            "Not a URL, existing file, or requirement spec: %r" % (spec,)
        ) from e


def parse_bdist_wininst(name):
    """Return (base,pyversion) or (None,None) for possible .exe name"""

    lower = name.lower()
    base, py_ver, plat = None, None, None

    if lower.endswith('.exe'):
        if lower.endswith('.win32.exe'):
            base = name[:-10]
            plat = 'win32'
        elif lower.startswith('.win32-py', -16):
            py_ver = name[-7:-4]
            base = name[:-16]
            plat = 'win32'
        elif lower.endswith('.win-amd64.exe'):
            base = name[:-14]
            plat = 'win-amd64'
        elif lower.startswith('.win-amd64-py', -20):
            py_ver = name[-7:-4]
            base = name[:-20]
            plat = 'win-amd64'
    return base, py_ver, plat


def egg_info_for_url(url):
    parts = urllib.parse.urlparse(url)
    scheme, server, path, parameters, query, fragment = parts
    base = urllib.parse.unquote(path.split('/')[-1])
    if server == 'sourceforge.net' and base == 'download':  # XXX Yuck
        base = urllib.parse.unquote(path.split('/')[-2])
    if '#' in base:
        base, fragment = base.split('#', 1)
    return base, fragment


def distros_for_url(url, metadata=None):
    """Yield egg or source distribution objects that might be found at a URL"""
    base, fragment = egg_info_for_url(url)
    for dist in distros_for_location(url, base, metadata):
        yield dist
    if fragment:
        match = EGG_FRAGMENT.match(fragment)
        if match:
            for dist in interpret_distro_name(
                url, match.group(1), metadata, precedence=CHECKOUT_DIST
            ):
                yield dist


def distros_for_location(location, basename, metadata=None):
    """Yield egg or source distribution objects based on basename"""
    if basename.endswith('.egg.zip'):
        basename = basename[:-4]  # strip the .zip
    if basename.endswith('.egg') and '-' in basename:
        # only one, unambiguous interpretation
        return [Distribution.from_location(location, basename, metadata)]
    if basename.endswith('.whl') and '-' in basename:
        wheel = Wheel(basename)
        if not wheel.is_compatible():
            return []
        return [Distribution(
            location=location,
            project_name=wheel.project_name,
            version=wheel.version,
            # Increase priority over eggs.
            precedence=EGG_DIST + 1,
        )]
    if basename.endswith('.exe'):
        win_base, py_ver, platform = parse_bdist_wininst(basename)
        if win_base is not None:
            return interpret_distro_name(
                location, win_base, metadata, py_ver, BINARY_DIST, platform
            )
    # Try source distro extensions (.zip, .tgz, etc.)
    #
    for ext in EXTENSIONS:
        if basename.endswith(ext):
            basename = basename[:-len(ext)]
            return interpret_distro_name(location, basename, metadata)
    return []  # no extension matched


def distros_for_filename(filename, metadata=None):
    """Yield possible egg or source distribution objects based on a filename"""
    return distros_for_location(
        normalize_path(filename), os.path.basename(filename), metadata
    )


def interpret_distro_name(
        location, basename, metadata, py_version=None, precedence=SOURCE_DIST,
        platform=None
):
    """Generate alternative interpretations of a source distro name

    Note: if `location` is a filesystem filename, you should call
    ``pkg_resources.normalize_path()`` on it before passing it to this
    routine!
    """
    # Generate alternative interpretations of a source distro name
    # Because some packages are ambiguous as to name/versions split
    # e.g. "adns-python-1.1.0", "egenix-mx-commercial", etc.
    # So, we generate each possible interpretation (e.g. "adns, python-1.1.0"
    # "adns-python, 1.1.0", and "adns-python-1.1.0, no version").  In practice,
    # the spurious interpretations should be ignored, because in the event
    # there's also an "adns" package, the spurious "python-1.1.0" version will
    # compare lower than any numeric version number, and is therefore unlikely
    # to match a request for it.  It's still a potential problem, though, and
    # in the long run PyPI and the distutils should go for "safe" names and
    # versions in distribution archive names (sdist and bdist).

    parts = basename.split('-')
    if not py_version and any(re.match(r'py\d\.\d$', p) for p in parts[2:]):
        # it is a bdist_dumb, not an sdist -- bail out
        return

    for p in range(1, len(parts) + 1):
        yield Distribution(
            location, metadata, '-'.join(parts[:p]), '-'.join(parts[p:]),
            py_version=py_version, precedence=precedence,
            platform=platform
        )


def unique_values(func):
    """
    Wrap a function returning an iterable such that the resulting iterable
    only ever yields unique items.
    """

    @wraps(func)
    def wrapper(*args, **kwargs):
        return unique_everseen(func(*args, **kwargs))

    return wrapper


REL = re.compile(r"""<([^>]*\srel\s*=\s*['"]?([^'">]+)[^>]*)>""", re.I)
# this line is here to fix emacs' cruddy broken syntax highlighting


@unique_values
def find_external_links(url, page):
    """Find rel="homepage" and rel="download" links in `page`, yielding URLs"""

    for match in REL.finditer(page):
        tag, rel = match.groups()
        rels = set(map(str.strip, rel.lower().split(',')))
        if 'homepage' in rels or 'download' in rels:
            for match in HREF.finditer(tag):
                yield urllib.parse.urljoin(url, htmldecode(match.group(1)))

    for tag in ("<th>Home Page", "<th>Download URL"):
        pos = page.find(tag)
        if pos != -1:
            match = HREF.search(page, pos)
            if match:
                yield urllib.parse.urljoin(url, htmldecode(match.group(1)))


class ContentChecker:
    """
    A null content checker that defines the interface for checking content
    """

    def feed(self, block):
        """
        Feed a block of data to the hash.
        """
        return

    def is_valid(self):
        """
        Check the hash. Return False if validation fails.
        """
        return True

    def report(self, reporter, template):
        """
        Call reporter with information about the checker (hash name)
        substituted into the template.
        """
        return


class HashChecker(ContentChecker):
    pattern = re.compile(
        r'(?P<hash_name>sha1|sha224|sha384|sha256|sha512|md5)='
        r'(?P<expected>[a-f0-9]+)'
    )

    def __init__(self, hash_name, expected):
        self.hash_name = hash_name
        self.hash = hashlib.new(hash_name)
        self.expected = expected

    @classmethod
    def from_url(cls, url):
        "Construct a (possibly null) ContentChecker from a URL"
        fragment = urllib.parse.urlparse(url)[-1]
        if not fragment:
            return ContentChecker()
        match = cls.pattern.search(fragment)
        if not match:
            return ContentChecker()
        return cls(**match.groupdict())

    def feed(self, block):
        self.hash.update(block)

    def is_valid(self):
        return self.hash.hexdigest() == self.expected

    def report(self, reporter, template):
        msg = template % self.hash_name
        return reporter(msg)


class PackageIndex(Environment):
    """A distribution index that scans web pages for download URLs"""

    def __init__(
            self, index_url="https://pypi.org/simple/", hosts=('*',),
            ca_bundle=None, verify_ssl=True, *args, **kw
    ):
        Environment.__init__(self, *args, **kw)
        self.index_url = index_url + "/" [:not index_url.endswith('/')]
        self.scanned_urls = {}
        self.fetched_urls = {}
        self.package_pages = {}
        self.allows = re.compile('|'.join(map(translate, hosts))).match
        self.to_scan = []
        self.opener = urllib.request.urlopen

    # FIXME: 'PackageIndex.process_url' is too complex (14)
    def process_url(self, url, retrieve=False):  # noqa: C901
        """Evaluate a URL as a possible download, and maybe retrieve it"""
        if url in self.scanned_urls and not retrieve:
            return
        self.scanned_urls[url] = True
        if not URL_SCHEME(url):
            self.process_filename(url)
            return
        else:
            dists = list(distros_for_url(url))
            if dists:
                if not self.url_ok(url):
                    return
                self.debug("Found link: %s", url)

        if dists or not retrieve or url in self.fetched_urls:
            list(map(self.add, dists))
            return  # don't need the actual page

        if not self.url_ok(url):
            self.fetched_urls[url] = True
            return

        self.info("Reading %s", url)
        self.fetched_urls[url] = True  # prevent multiple fetch attempts
        tmpl = "Download error on %s: %%s -- Some packages may not be found!"
        f = self.open_url(url, tmpl % url)
        if f is None:
            return
        if isinstance(f, urllib.error.HTTPError) and f.code == 401:
            self.info("Authentication error: %s" % f.msg)
        self.fetched_urls[f.url] = True
        if 'html' not in f.headers.get('content-type', '').lower():
            f.close()  # not html, we can't process it
            return

        base = f.url  # handle redirects
        page = f.read()
        if not isinstance(page, str):
            # In Python 3 and got bytes but want str.
            if isinstance(f, urllib.error.HTTPError):
                # Errors have no charset, assume latin1:
                charset = 'latin-1'
            else:
                charset = f.headers.get_param('charset') or 'latin-1'
            page = page.decode(charset, "ignore")
        f.close()
        for match in HREF.finditer(page):
            link = urllib.parse.urljoin(base, htmldecode(match.group(1)))
            self.process_url(link)
        if url.startswith(self.index_url) and getattr(f, 'code', None) != 404:
            page = self.process_index(url, page)

    def process_filename(self, fn, nested=False):
        # process filenames or directories
        if not os.path.exists(fn):
            self.warn("Not found: %s", fn)
            return

        if os.path.isdir(fn) and not nested:
            path = os.path.realpath(fn)
            for item in os.listdir(path):
                self.process_filename(os.path.join(path, item), True)

        dists = distros_for_filename(fn)
        if dists:
            self.debug("Found: %s", fn)
            list(map(self.add, dists))

    def url_ok(self, url, fatal=False):
        s = URL_SCHEME(url)
        is_file = s and s.group(1).lower() == 'file'
        if is_file or self.allows(urllib.parse.urlparse(url)[1]):
            return True
        msg = (
            "\nNote: Bypassing %s (disallowed host; see "
            "http://bit.ly/2hrImnY for details).\n")
        if fatal:
            raise DistutilsError(msg % url)
        else:
            self.warn(msg, url)

    def scan_egg_links(self, search_path):
        dirs = filter(os.path.isdir, search_path)
        egg_links = (
            (path, entry)
            for path in dirs
            for entry in os.listdir(path)
            if entry.endswith('.egg-link')
        )
        list(itertools.starmap(self.scan_egg_link, egg_links))

    def scan_egg_link(self, path, entry):
        with open(os.path.join(path, entry)) as raw_lines:
            # filter non-empty lines
            lines = list(filter(None, map(str.strip, raw_lines)))

        if len(lines) != 2:
            # format is not recognized; punt
            return

        egg_path, setup_path = lines

        for dist in find_distributions(os.path.join(path, egg_path)):
            dist.location = os.path.join(path, *lines)
            dist.precedence = SOURCE_DIST
            self.add(dist)

    def _scan(self, link):
        # Process a URL to see if it's for a package page
        NO_MATCH_SENTINEL = None, None
        if not link.startswith(self.index_url):
            return NO_MATCH_SENTINEL

        parts = list(map(
            urllib.parse.unquote, link[len(self.index_url):].split('/')
        ))
        if len(parts) != 2 or '#' in parts[1]:
            return NO_MATCH_SENTINEL

        # it's a package page, sanitize and index it
        pkg = safe_name(parts[0])
        ver = safe_version(parts[1])
        self.package_pages.setdefault(pkg.lower(), {})[link] = True
        return to_filename(pkg), to_filename(ver)

    def process_index(self, url, page):
        """Process the contents of a PyPI page"""

        # process an index page into the package-page index
        for match in HREF.finditer(page):
            try:
                self._scan(urllib.parse.urljoin(url, htmldecode(match.group(1))))
            except ValueError:
                pass

        pkg, ver = self._scan(url)  # ensure this page is in the page index
        if not pkg:
            return ""  # no sense double-scanning non-package pages

        # process individual package page
        for new_url in find_external_links(url, page):
            # Process the found URL
            base, frag = egg_info_for_url(new_url)
            if base.endswith('.py') and not frag:
                if ver:
                    new_url += '#egg=%s-%s' % (pkg, ver)
                else:
                    self.need_version_info(url)
            self.scan_url(new_url)

        return PYPI_MD5.sub(
            lambda m: '<a href="%s#md5=%s">%s</a>' % m.group(1, 3, 2), page
        )

    def need_version_info(self, url):
        self.scan_all(
            "Page at %s links to .py file(s) without version info; an index "
            "scan is required.", url
        )

    def scan_all(self, msg=None, *args):
        if self.index_url not in self.fetched_urls:
            if msg:
                self.warn(msg, *args)
            self.info(
                "Scanning index of all packages (this may take a while)"
            )
        self.scan_url(self.index_url)

    def find_packages(self, requirement):
        self.scan_url(self.index_url + requirement.unsafe_name + '/')

        if not self.package_pages.get(requirement.key):
            # Fall back to safe version of the name
            self.scan_url(self.index_url + requirement.project_name + '/')

        if not self.package_pages.get(requirement.key):
            # We couldn't find the target package, so search the index page too
            self.not_found_in_index(requirement)

        for url in list(self.package_pages.get(requirement.key, ())):
            # scan each page that might be related to the desired package
            self.scan_url(url)

    def obtain(self, requirement, installer=None):
        self.prescan()
        self.find_packages(requirement)
        for dist in self[requirement.key]:
            if dist in requirement:
                return dist
            self.debug("%s does not match %s", requirement, dist)
        return super(PackageIndex, self).obtain(requirement, installer)

    def check_hash(self, checker, filename, tfp):
        """
        checker is a ContentChecker
        """
        checker.report(
            self.debug,
            "Validating %%s checksum for %s" % filename)
        if not checker.is_valid():
            tfp.close()
            os.unlink(filename)
            raise DistutilsError(
                "%s validation failed for %s; "
                "possible download problem?"
                % (checker.hash.name, os.path.basename(filename))
            )

    def add_find_links(self, urls):
        """Add `urls` to the list that will be prescanned for searches"""
        for url in urls:
            if (
                self.to_scan is None  # if we have already "gone online"
                or not URL_SCHEME(url)  # or it's a local file/directory
                or url.startswith('file:')
                or list(distros_for_url(url))  # or a direct package link
            ):
                # then go ahead and process it now
                self.scan_url(url)
            else:
                # otherwise, defer retrieval till later
                self.to_scan.append(url)

    def prescan(self):
        """Scan urls scheduled for prescanning (e.g. --find-links)"""
        if self.to_scan:
            list(map(self.scan_url, self.to_scan))
        self.to_scan = None  # from now on, go ahead and process immediately

    def not_found_in_index(self, requirement):
        if self[requirement.key]:  # we've seen at least one distro
            meth, msg = self.info, "Couldn't retrieve index page for %r"
        else:  # no distros seen for this name, might be misspelled
            meth, msg = (
                self.warn,
                "Couldn't find index page for %r (maybe misspelled?)")
        meth(msg, requirement.unsafe_name)
        self.scan_all()

    def download(self, spec, tmpdir):
        """Locate and/or download `spec` to `tmpdir`, returning a local path

        `spec` may be a ``Requirement`` object, or a string containing a URL,
        an existing local filename, or a project/version requirement spec
        (i.e. the string form of a ``Requirement`` object).  If it is the URL
        of a .py file with an unambiguous ``#egg=name-version`` tag (i.e., one
        that escapes ``-`` as ``_`` throughout), a trivial ``setup.py`` is
        automatically created alongside the downloaded file.

        If `spec` is a ``Requirement`` object or a string containing a
        project/version requirement spec, this method returns the location of
        a matching distribution (possibly after downloading it to `tmpdir`).
        If `spec` is a locally existing file or directory name, it is simply
        returned unchanged.  If `spec` is a URL, it is downloaded to a subpath
        of `tmpdir`, and the local filename is returned.  Various errors may be
        raised if a problem occurs during downloading.
        """
        if not isinstance(spec, Requirement):
            scheme = URL_SCHEME(spec)
            if scheme:
                # It's a url, download it to tmpdir
                found = self._download_url(scheme.group(1), spec, tmpdir)
                base, fragment = egg_info_for_url(spec)
                if base.endswith('.py'):
                    found = self.gen_setup(found, fragment, tmpdir)
                return found
            elif os.path.exists(spec):
                # Existing file or directory, just return it
                return spec
            else:
                spec = parse_requirement_arg(spec)
        return getattr(self.fetch_distribution(spec, tmpdir), 'location', None)

    def fetch_distribution(  # noqa: C901  # is too complex (14)  # FIXME
            self, requirement, tmpdir, force_scan=False, source=False,
            develop_ok=False, local_index=None):
        """Obtain a distribution suitable for fulfilling `requirement`

        `requirement` must be a ``pkg_resources.Requirement`` instance.
        If necessary, or if the `force_scan` flag is set, the requirement is
        searched for in the (online) package index as well as the locally
        installed packages.  If a distribution matching `requirement` is found,
        the returned distribution's ``location`` is the value you would have
        gotten from calling the ``download()`` method with the matching
        distribution's URL or filename.  If no matching distribution is found,
        ``None`` is returned.

        If the `source` flag is set, only source distributions and source
        checkout links will be considered.  Unless the `develop_ok` flag is
        set, development and system eggs (i.e., those using the ``.egg-info``
        format) will be ignored.
        """
        # process a Requirement
        self.info("Searching for %s", requirement)
        skipped = {}
        dist = None

        def find(req, env=None):
            if env is None:
                env = self
            # Find a matching distribution; may be called more than once

            for dist in env[req.key]:

                if dist.precedence == DEVELOP_DIST and not develop_ok:
                    if dist not in skipped:
                        self.warn(
                            "Skipping development or system egg: %s", dist,
                        )
                        skipped[dist] = 1
                    continue

                test = (
                    dist in req
                    and (dist.precedence <= SOURCE_DIST or not source)
                )
                if test:
                    loc = self.download(dist.location, tmpdir)
                    dist.download_location = loc
                    if os.path.exists(dist.download_location):
                        return dist

        if force_scan:
            self.prescan()
            self.find_packages(requirement)
            dist = find(requirement)

        if not dist and local_index is not None:
            dist = find(requirement, local_index)

        if dist is None:
            if self.to_scan is not None:
                self.prescan()
            dist = find(requirement)

        if dist is None and not force_scan:
            self.find_packages(requirement)
            dist = find(requirement)

        if dist is None:
            self.warn(
                "No local packages or working download links found for %s%s",
                (source and "a source distribution of " or ""),
                requirement,
            )
        else:
            self.info("Best match: %s", dist)
            return dist.clone(location=dist.download_location)

    def fetch(self, requirement, tmpdir, force_scan=False, source=False):
        """Obtain a file suitable for fulfilling `requirement`

        DEPRECATED; use the ``fetch_distribution()`` method now instead.  For
        backward compatibility, this routine is identical but returns the
        ``location`` of the downloaded distribution instead of a distribution
        object.
        """
        dist = self.fetch_distribution(requirement, tmpdir, force_scan, source)
        if dist is not None:
            return dist.location
        return None

    def gen_setup(self, filename, fragment, tmpdir):
        match = EGG_FRAGMENT.match(fragment)
        dists = match and [
            d for d in
            interpret_distro_name(filename, match.group(1), None) if d.version
        ] or []

        if len(dists) == 1:  # unambiguous ``#egg`` fragment
            basename = os.path.basename(filename)

            # Make sure the file has been downloaded to the temp dir.
            if os.path.dirname(filename) != tmpdir:
                dst = os.path.join(tmpdir, basename)
                from setuptools.command.easy_install import samefile
                if not samefile(filename, dst):
                    shutil.copy2(filename, dst)
                    filename = dst

            with open(os.path.join(tmpdir, 'setup.py'), 'w') as file:
                file.write(
                    "from setuptools import setup\n"
                    "setup(name=%r, version=%r, py_modules=[%r])\n"
                    % (
                        dists[0].project_name, dists[0].version,
                        os.path.splitext(basename)[0]
                    )
                )
            return filename

        elif match:
            raise DistutilsError(
                "Can't unambiguously interpret project/version identifier %r; "
                "any dashes in the name or version should be escaped using "
                "underscores. %r" % (fragment, dists)
            )
        else:
            raise DistutilsError(
                "Can't process plain .py files without an '#egg=name-version'"
                " suffix to enable automatic setup script generation."
            )

    dl_blocksize = 8192

    def _download_to(self, url, filename):
        self.info("Downloading %s", url)
        # Download the file
        fp = None
        try:
            checker = HashChecker.from_url(url)
            fp = self.open_url(url)
            if isinstance(fp, urllib.error.HTTPError):
                raise DistutilsError(
                    "Can't download %s: %s %s" % (url, fp.code, fp.msg)
                )
            headers = fp.info()
            blocknum = 0
            bs = self.dl_blocksize
            size = -1
            if "content-length" in headers:
                # Some servers return multiple Content-Length headers :(
                sizes = headers.get_all('Content-Length')
                size = max(map(int, sizes))
                self.reporthook(url, filename, blocknum, bs, size)
            with open(filename, 'wb') as tfp:
                while True:
                    block = fp.read(bs)
                    if block:
                        checker.feed(block)
                        tfp.write(block)
                        blocknum += 1
                        self.reporthook(url, filename, blocknum, bs, size)
                    else:
                        break
                self.check_hash(checker, filename, tfp)
            return headers
        finally:
            if fp:
                fp.close()

    def reporthook(self, url, filename, blocknum, blksize, size):
        pass  # no-op

    # FIXME:
    def open_url(self, url, warning=None):  # noqa: C901  # is too complex (12)
        if url.startswith('file:'):
            return local_open(url)
        try:
            return open_with_auth(url, self.opener)
        except (ValueError, http.client.InvalidURL) as v:
            msg = ' '.join([str(arg) for arg in v.args])
            if warning:
                self.warn(warning, msg)
            else:
                raise DistutilsError('%s %s' % (url, msg)) from v
        except urllib.error.HTTPError as v:
            return v
        except urllib.error.URLError as v:
            if warning:
                self.warn(warning, v.reason)
            else:
                raise DistutilsError("Download error for %s: %s"
                                     % (url, v.reason)) from v
        except http.client.BadStatusLine as v:
            if warning:
                self.warn(warning, v.line)
            else:
                raise DistutilsError(
                    '%s returned a bad status line. The server might be '
                    'down, %s' %
                    (url, v.line)
                ) from v
        except (http.client.HTTPException, socket.error) as v:
            if warning:
                self.warn(warning, v)
            else:
                raise DistutilsError("Download error for %s: %s"
                                     % (url, v)) from v

    def _download_url(self, scheme, url, tmpdir):
        # Determine download filename
        #
        name, fragment = egg_info_for_url(url)
        if name:
            while '..' in name:
                name = name.replace('..', '.').replace('\\', '_')
        else:
            name = "__downloaded__"  # default if URL has no path contents

        if name.endswith('.egg.zip'):
            name = name[:-4]  # strip the extra .zip before download

        filename = os.path.join(tmpdir, name)

        # Download the file
        #
        if scheme == 'svn' or scheme.startswith('svn+'):
            return self._download_svn(url, filename)
        elif scheme == 'git' or scheme.startswith('git+'):
            return self._download_git(url, filename)
        elif scheme.startswith('hg+'):
            return self._download_hg(url, filename)
        elif scheme == 'file':
            return urllib.request.url2pathname(urllib.parse.urlparse(url)[2])
        else:
            self.url_ok(url, True)  # raises error if not allowed
            return self._attempt_download(url, filename)

    def scan_url(self, url):
        self.process_url(url, True)

    def _attempt_download(self, url, filename):
        headers = self._download_to(url, filename)
        if 'html' in headers.get('content-type', '').lower():
            return self._download_html(url, headers, filename)
        else:
            return filename

    def _download_html(self, url, headers, filename):
        file = open(filename)
        for line in file:
            if line.strip():
                # Check for a subversion index page
                if re.search(r'<title>([^- ]+ - )?Revision \d+:', line):
                    # it's a subversion index page:
                    file.close()
                    os.unlink(filename)
                    return self._download_svn(url, filename)
                break  # not an index page
        file.close()
        os.unlink(filename)
        raise DistutilsError("Unexpected HTML page found at " + url)

    def _download_svn(self, url, filename):
        warnings.warn("SVN download support is deprecated", UserWarning)
        url = url.split('#', 1)[0]  # remove any fragment for svn's sake
        creds = ''
        if url.lower().startswith('svn:') and '@' in url:
            scheme, netloc, path, p, q, f = urllib.parse.urlparse(url)
            if not netloc and path.startswith('//') and '/' in path[2:]:
                netloc, path = path[2:].split('/', 1)
                auth, host = _splituser(netloc)
                if auth:
                    if ':' in auth:
                        user, pw = auth.split(':', 1)
                        creds = " --username=%s --password=%s" % (user, pw)
                    else:
                        creds = " --username=" + auth
                    netloc = host
                    parts = scheme, netloc, url, p, q, f
                    url = urllib.parse.urlunparse(parts)
        self.info("Doing subversion checkout from %s to %s", url, filename)
        os.system("svn checkout%s -q %s %s" % (creds, url, filename))
        return filename

    @staticmethod
    def _vcs_split_rev_from_url(url, pop_prefix=False):
        scheme, netloc, path, query, frag = urllib.parse.urlsplit(url)

        scheme = scheme.split('+', 1)[-1]

        # Some fragment identification fails
        path = path.split('#', 1)[0]

        rev = None
        if '@' in path:
            path, rev = path.rsplit('@', 1)

        # Also, discard fragment
        url = urllib.parse.urlunsplit((scheme, netloc, path, query, ''))

        return url, rev

    def _download_git(self, url, filename):
        filename = filename.split('#', 1)[0]
        url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True)

        self.info("Doing git clone from %s to %s", url, filename)
        os.system("git clone --quiet %s %s" % (url, filename))

        if rev is not None:
            self.info("Checking out %s", rev)
            os.system("git -C %s checkout --quiet %s" % (
                filename,
                rev,
            ))

        return filename

    def _download_hg(self, url, filename):
        filename = filename.split('#', 1)[0]
        url, rev = self._vcs_split_rev_from_url(url, pop_prefix=True)

        self.info("Doing hg clone from %s to %s", url, filename)
        os.system("hg clone --quiet %s %s" % (url, filename))

        if rev is not None:
            self.info("Updating to %s", rev)
            os.system("hg --cwd %s up -C -r %s -q" % (
                filename,
                rev,
            ))

        return filename

    def debug(self, msg, *args):
        log.debug(msg, *args)

    def info(self, msg, *args):
        log.info(msg, *args)

    def warn(self, msg, *args):
        log.warn(msg, *args)


# This pattern matches a character entity reference (a decimal numeric
# references, a hexadecimal numeric reference, or a named reference).
entity_sub = re.compile(r'&(#(\d+|x[\da-fA-F]+)|[\w.:-]+);?').sub


def decode_entity(match):
    what = match.group(0)
    return html.unescape(what)


def htmldecode(text):
    """
    Decode HTML entities in the given text.

    >>> htmldecode(
    ...     'https://../package_name-0.1.2.tar.gz'
    ...     '?tokena=A&amp;tokenb=B">package_name-0.1.2.tar.gz')
    'https://../package_name-0.1.2.tar.gz?tokena=A&tokenb=B">package_name-0.1.2.tar.gz'
    """
    return entity_sub(decode_entity, text)


def socket_timeout(timeout=15):
    def _socket_timeout(func):
        def _socket_timeout(*args, **kwargs):
            old_timeout = socket.getdefaulttimeout()
            socket.setdefaulttimeout(timeout)
            try:
                return func(*args, **kwargs)
            finally:
                socket.setdefaulttimeout(old_timeout)

        return _socket_timeout

    return _socket_timeout


def _encode_auth(auth):
    """
    Encode auth from a URL suitable for an HTTP header.
    >>> str(_encode_auth('username%3Apassword'))
    'dXNlcm5hbWU6cGFzc3dvcmQ='

    Long auth strings should not cause a newline to be inserted.
    >>> long_auth = 'username:' + 'password'*10
    >>> chr(10) in str(_encode_auth(long_auth))
    False
    """
    auth_s = urllib.parse.unquote(auth)
    # convert to bytes
    auth_bytes = auth_s.encode()
    encoded_bytes = base64.b64encode(auth_bytes)
    # convert back to a string
    encoded = encoded_bytes.decode()
    # strip the trailing carriage return
    return encoded.replace('\n', '')


class Credential:
    """
    A username/password pair. Use like a namedtuple.
    """

    def __init__(self, username, password):
        self.username = username
        self.password = password

    def __iter__(self):
        yield self.username
        yield self.password

    def __str__(self):
        return '%(username)s:%(password)s' % vars(self)


class PyPIConfig(configparser.RawConfigParser):
    def __init__(self):
        """
        Load from ~/.pypirc
        """
        defaults = dict.fromkeys(['username', 'password', 'repository'], '')
        configparser.RawConfigParser.__init__(self, defaults)

        rc = os.path.join(os.path.expanduser('~'), '.pypirc')
        if os.path.exists(rc):
            self.read(rc)

    @property
    def creds_by_repository(self):
        sections_with_repositories = [
            section for section in self.sections()
            if self.get(section, 'repository').strip()
        ]

        return dict(map(self._get_repo_cred, sections_with_repositories))

    def _get_repo_cred(self, section):
        repo = self.get(section, 'repository').strip()
        return repo, Credential(
            self.get(section, 'username').strip(),
            self.get(section, 'password').strip(),
        )

    def find_credential(self, url):
        """
        If the URL indicated appears to be a repository defined in this
        config, return the credential for that repository.
        """
        for repository, cred in self.creds_by_repository.items():
            if url.startswith(repository):
                return cred


def open_with_auth(url, opener=urllib.request.urlopen):
    """Open a urllib2 request, handling HTTP authentication"""

    parsed = urllib.parse.urlparse(url)
    scheme, netloc, path, params, query, frag = parsed

    # Double scheme does not raise on macOS as revealed by a
    # failing test. We would expect "nonnumeric port". Refs #20.
    if netloc.endswith(':'):
        raise http.client.InvalidURL("nonnumeric port: ''")

    if scheme in ('http', 'https'):
        auth, address = _splituser(netloc)
    else:
        auth = None

    if not auth:
        cred = PyPIConfig().find_credential(url)
        if cred:
            auth = str(cred)
            info = cred.username, url
            log.info('Authenticating as %s for %s (from .pypirc)', *info)

    if auth:
        auth = "Basic " + _encode_auth(auth)
        parts = scheme, address, path, params, query, frag
        new_url = urllib.parse.urlunparse(parts)
        request = urllib.request.Request(new_url)
        request.add_header("Authorization", auth)
    else:
        request = urllib.request.Request(url)

    request.add_header('User-Agent', user_agent)
    fp = opener(request)

    if auth:
        # Put authentication info back into request URL if same host,
        # so that links found on the page will work
        s2, h2, path2, param2, query2, frag2 = urllib.parse.urlparse(fp.url)
        if s2 == scheme and h2 == address:
            parts = s2, netloc, path2, param2, query2, frag2
            fp.url = urllib.parse.urlunparse(parts)

    return fp


# copy of urllib.parse._splituser from Python 3.8
def _splituser(host):
    """splituser('user[:passwd]@host[:port]')
    --> 'user[:passwd]', 'host[:port]'."""
    user, delim, host = host.rpartition('@')
    return (user if delim else None), host


# adding a timeout to avoid freezing package_index
open_with_auth = socket_timeout(_SOCKET_TIMEOUT)(open_with_auth)


def fix_sf_url(url):
    return url  # backward compatibility


def local_open(url):
    """Read a local path, with special support for directories"""
    scheme, server, path, param, query, frag = urllib.parse.urlparse(url)
    filename = urllib.request.url2pathname(path)
    if os.path.isfile(filename):
        return urllib.request.urlopen(url)
    elif path.endswith('/') and os.path.isdir(filename):
        files = []
        for f in os.listdir(filename):
            filepath = os.path.join(filename, f)
            if f == 'index.html':
                with open(filepath, 'r') as fp:
                    body = fp.read()
                break
            elif os.path.isdir(filepath):
                f += '/'
            files.append('<a href="{name}">{name}</a>'.format(name=f))
        else:
            tmpl = (
                "<html><head><title>{url}</title>"
                "</head><body>{files}</body></html>")
            body = tmpl.format(url=url, files='\n'.join(files))
        status, message = 200, "OK"
    else:
        status, message, body = 404, "Path not found", "Not found"

    headers = {'content-type': 'text/html'}
    body_stream = io.StringIO(body)
    return urllib.error.HTTPError(url, status, message, headers, body_stream)
site-packages/setuptools/sandbox.py000064400000034014150564745010013526 0ustar00import os
import sys
import tempfile
import operator
import functools
import itertools
import re
import contextlib
import pickle
import textwrap
import builtins

import pkg_resources
from distutils.errors import DistutilsError
from pkg_resources import working_set

if sys.platform.startswith('java'):
    import org.python.modules.posix.PosixModule as _os
else:
    _os = sys.modules[os.name]
try:
    _file = file
except NameError:
    _file = None
_open = open


__all__ = [
    "AbstractSandbox",
    "DirectorySandbox",
    "SandboxViolation",
    "run_setup",
]


def _execfile(filename, globals, locals=None):
    """
    Python 3 implementation of execfile.
    """
    mode = 'rb'
    with open(filename, mode) as stream:
        script = stream.read()
    if locals is None:
        locals = globals
    code = compile(script, filename, 'exec')
    exec(code, globals, locals)


@contextlib.contextmanager
def save_argv(repl=None):
    saved = sys.argv[:]
    if repl is not None:
        sys.argv[:] = repl
    try:
        yield saved
    finally:
        sys.argv[:] = saved


@contextlib.contextmanager
def save_path():
    saved = sys.path[:]
    try:
        yield saved
    finally:
        sys.path[:] = saved


@contextlib.contextmanager
def override_temp(replacement):
    """
    Monkey-patch tempfile.tempdir with replacement, ensuring it exists
    """
    os.makedirs(replacement, exist_ok=True)

    saved = tempfile.tempdir

    tempfile.tempdir = replacement

    try:
        yield
    finally:
        tempfile.tempdir = saved


@contextlib.contextmanager
def pushd(target):
    saved = os.getcwd()
    os.chdir(target)
    try:
        yield saved
    finally:
        os.chdir(saved)


class UnpickleableException(Exception):
    """
    An exception representing another Exception that could not be pickled.
    """

    @staticmethod
    def dump(type, exc):
        """
        Always return a dumped (pickled) type and exc. If exc can't be pickled,
        wrap it in UnpickleableException first.
        """
        try:
            return pickle.dumps(type), pickle.dumps(exc)
        except Exception:
            # get UnpickleableException inside the sandbox
            from setuptools.sandbox import UnpickleableException as cls

            return cls.dump(cls, cls(repr(exc)))


class ExceptionSaver:
    """
    A Context Manager that will save an exception, serialized, and restore it
    later.
    """

    def __enter__(self):
        return self

    def __exit__(self, type, exc, tb):
        if not exc:
            return

        # dump the exception
        self._saved = UnpickleableException.dump(type, exc)
        self._tb = tb

        # suppress the exception
        return True

    def resume(self):
        "restore and re-raise any exception"

        if '_saved' not in vars(self):
            return

        type, exc = map(pickle.loads, self._saved)
        raise exc.with_traceback(self._tb)


@contextlib.contextmanager
def save_modules():
    """
    Context in which imported modules are saved.

    Translates exceptions internal to the context into the equivalent exception
    outside the context.
    """
    saved = sys.modules.copy()
    with ExceptionSaver() as saved_exc:
        yield saved

    sys.modules.update(saved)
    # remove any modules imported since
    del_modules = (
        mod_name
        for mod_name in sys.modules
        if mod_name not in saved
        # exclude any encodings modules. See #285
        and not mod_name.startswith('encodings.')
    )
    _clear_modules(del_modules)

    saved_exc.resume()


def _clear_modules(module_names):
    for mod_name in list(module_names):
        del sys.modules[mod_name]


@contextlib.contextmanager
def save_pkg_resources_state():
    saved = pkg_resources.__getstate__()
    try:
        yield saved
    finally:
        pkg_resources.__setstate__(saved)


@contextlib.contextmanager
def setup_context(setup_dir):
    temp_dir = os.path.join(setup_dir, 'temp')
    with save_pkg_resources_state():
        with save_modules():
            with save_path():
                hide_setuptools()
                with save_argv():
                    with override_temp(temp_dir):
                        with pushd(setup_dir):
                            # ensure setuptools commands are available
                            __import__('setuptools')
                            yield


_MODULES_TO_HIDE = {
    'setuptools',
    'distutils',
    'pkg_resources',
    'Cython',
    '_distutils_hack',
}


def _needs_hiding(mod_name):
    """
    >>> _needs_hiding('setuptools')
    True
    >>> _needs_hiding('pkg_resources')
    True
    >>> _needs_hiding('setuptools_plugin')
    False
    >>> _needs_hiding('setuptools.__init__')
    True
    >>> _needs_hiding('distutils')
    True
    >>> _needs_hiding('os')
    False
    >>> _needs_hiding('Cython')
    True
    """
    base_module = mod_name.split('.', 1)[0]
    return base_module in _MODULES_TO_HIDE


def hide_setuptools():
    """
    Remove references to setuptools' modules from sys.modules to allow the
    invocation to import the most appropriate setuptools. This technique is
    necessary to avoid issues such as #315 where setuptools upgrading itself
    would fail to find a function declared in the metadata.
    """
    _distutils_hack = sys.modules.get('_distutils_hack', None)
    if _distutils_hack is not None:
        _distutils_hack.remove_shim()

    modules = filter(_needs_hiding, sys.modules)
    _clear_modules(modules)


def run_setup(setup_script, args):
    """Run a distutils setup script, sandboxed in its directory"""
    setup_dir = os.path.abspath(os.path.dirname(setup_script))
    with setup_context(setup_dir):
        try:
            sys.argv[:] = [setup_script] + list(args)
            sys.path.insert(0, setup_dir)
            # reset to include setup dir, w/clean callback list
            working_set.__init__()
            working_set.callbacks.append(lambda dist: dist.activate())

            with DirectorySandbox(setup_dir):
                ns = dict(__file__=setup_script, __name__='__main__')
                _execfile(setup_script, ns)
        except SystemExit as v:
            if v.args and v.args[0]:
                raise
            # Normal exit, just return


class AbstractSandbox:
    """Wrap 'os' module and 'open()' builtin for virtualizing setup scripts"""

    _active = False

    def __init__(self):
        self._attrs = [
            name
            for name in dir(_os)
            if not name.startswith('_') and hasattr(self, name)
        ]

    def _copy(self, source):
        for name in self._attrs:
            setattr(os, name, getattr(source, name))

    def __enter__(self):
        self._copy(self)
        if _file:
            builtins.file = self._file
        builtins.open = self._open
        self._active = True

    def __exit__(self, exc_type, exc_value, traceback):
        self._active = False
        if _file:
            builtins.file = _file
        builtins.open = _open
        self._copy(_os)

    def run(self, func):
        """Run 'func' under os sandboxing"""
        with self:
            return func()

    def _mk_dual_path_wrapper(name):
        original = getattr(_os, name)

        def wrap(self, src, dst, *args, **kw):
            if self._active:
                src, dst = self._remap_pair(name, src, dst, *args, **kw)
            return original(src, dst, *args, **kw)

        return wrap

    for name in ["rename", "link", "symlink"]:
        if hasattr(_os, name):
            locals()[name] = _mk_dual_path_wrapper(name)

    def _mk_single_path_wrapper(name, original=None):
        original = original or getattr(_os, name)

        def wrap(self, path, *args, **kw):
            if self._active:
                path = self._remap_input(name, path, *args, **kw)
            return original(path, *args, **kw)

        return wrap

    if _file:
        _file = _mk_single_path_wrapper('file', _file)
    _open = _mk_single_path_wrapper('open', _open)
    for name in [
        "stat",
        "listdir",
        "chdir",
        "open",
        "chmod",
        "chown",
        "mkdir",
        "remove",
        "unlink",
        "rmdir",
        "utime",
        "lchown",
        "chroot",
        "lstat",
        "startfile",
        "mkfifo",
        "mknod",
        "pathconf",
        "access",
    ]:
        if hasattr(_os, name):
            locals()[name] = _mk_single_path_wrapper(name)

    def _mk_single_with_return(name):
        original = getattr(_os, name)

        def wrap(self, path, *args, **kw):
            if self._active:
                path = self._remap_input(name, path, *args, **kw)
                return self._remap_output(name, original(path, *args, **kw))
            return original(path, *args, **kw)

        return wrap

    for name in ['readlink', 'tempnam']:
        if hasattr(_os, name):
            locals()[name] = _mk_single_with_return(name)

    def _mk_query(name):
        original = getattr(_os, name)

        def wrap(self, *args, **kw):
            retval = original(*args, **kw)
            if self._active:
                return self._remap_output(name, retval)
            return retval

        return wrap

    for name in ['getcwd', 'tmpnam']:
        if hasattr(_os, name):
            locals()[name] = _mk_query(name)

    def _validate_path(self, path):
        """Called to remap or validate any path, whether input or output"""
        return path

    def _remap_input(self, operation, path, *args, **kw):
        """Called for path inputs"""
        return self._validate_path(path)

    def _remap_output(self, operation, path):
        """Called for path outputs"""
        return self._validate_path(path)

    def _remap_pair(self, operation, src, dst, *args, **kw):
        """Called for path pairs like rename, link, and symlink operations"""
        return (
            self._remap_input(operation + '-from', src, *args, **kw),
            self._remap_input(operation + '-to', dst, *args, **kw),
        )


if hasattr(os, 'devnull'):
    _EXCEPTIONS = [os.devnull]
else:
    _EXCEPTIONS = []


class DirectorySandbox(AbstractSandbox):
    """Restrict operations to a single subdirectory - pseudo-chroot"""

    write_ops = dict.fromkeys(
        [
            "open",
            "chmod",
            "chown",
            "mkdir",
            "remove",
            "unlink",
            "rmdir",
            "utime",
            "lchown",
            "chroot",
            "mkfifo",
            "mknod",
            "tempnam",
        ]
    )

    _exception_patterns = []
    "exempt writing to paths that match the pattern"

    def __init__(self, sandbox, exceptions=_EXCEPTIONS):
        self._sandbox = os.path.normcase(os.path.realpath(sandbox))
        self._prefix = os.path.join(self._sandbox, '')
        self._exceptions = [
            os.path.normcase(os.path.realpath(path)) for path in exceptions
        ]
        AbstractSandbox.__init__(self)

    def _violation(self, operation, *args, **kw):
        from setuptools.sandbox import SandboxViolation

        raise SandboxViolation(operation, args, kw)

    if _file:

        def _file(self, path, mode='r', *args, **kw):
            if mode not in ('r', 'rt', 'rb', 'rU', 'U') and not self._ok(path):
                self._violation("file", path, mode, *args, **kw)
            return _file(path, mode, *args, **kw)

    def _open(self, path, mode='r', *args, **kw):
        if mode not in ('r', 'rt', 'rb', 'rU', 'U') and not self._ok(path):
            self._violation("open", path, mode, *args, **kw)
        return _open(path, mode, *args, **kw)

    def tmpnam(self):
        self._violation("tmpnam")

    def _ok(self, path):
        active = self._active
        try:
            self._active = False
            realpath = os.path.normcase(os.path.realpath(path))
            return (
                self._exempted(realpath)
                or realpath == self._sandbox
                or realpath.startswith(self._prefix)
            )
        finally:
            self._active = active

    def _exempted(self, filepath):
        start_matches = (
            filepath.startswith(exception) for exception in self._exceptions
        )
        pattern_matches = (
            re.match(pattern, filepath) for pattern in self._exception_patterns
        )
        candidates = itertools.chain(start_matches, pattern_matches)
        return any(candidates)

    def _remap_input(self, operation, path, *args, **kw):
        """Called for path inputs"""
        if operation in self.write_ops and not self._ok(path):
            self._violation(operation, os.path.realpath(path), *args, **kw)
        return path

    def _remap_pair(self, operation, src, dst, *args, **kw):
        """Called for path pairs like rename, link, and symlink operations"""
        if not self._ok(src) or not self._ok(dst):
            self._violation(operation, src, dst, *args, **kw)
        return (src, dst)

    def open(self, file, flags, mode=0o777, *args, **kw):
        """Called for low-level os.open()"""
        if flags & WRITE_FLAGS and not self._ok(file):
            self._violation("os.open", file, flags, mode, *args, **kw)
        return _os.open(file, flags, mode, *args, **kw)


WRITE_FLAGS = functools.reduce(
    operator.or_,
    [
        getattr(_os, a, 0)
        for a in "O_WRONLY O_RDWR O_APPEND O_CREAT O_TRUNC O_TEMPORARY".split()
    ],
)


class SandboxViolation(DistutilsError):
    """A setup script attempted to modify the filesystem outside the sandbox"""

    tmpl = textwrap.dedent(
        """
        SandboxViolation: {cmd}{args!r} {kwargs}

        The package setup script has attempted to modify files on your system
        that are not within the EasyInstall build area, and has been aborted.

        This package cannot be safely installed by EasyInstall, and may not
        support alternate installation locations even if you run its setup
        script by hand.  Please inform the package's author and the EasyInstall
        maintainers to find out if a fix or workaround is available.
        """
    ).lstrip()

    def __str__(self):
        cmd, args, kwargs = self.args
        return self.tmpl.format(**locals())
site-packages/setuptools/_imp.py000064400000004530150564745010013014 0ustar00"""
Re-implementation of find_module and get_frozen_object
from the deprecated imp module.
"""

import os
import importlib.util
import importlib.machinery

from .py34compat import module_from_spec


PY_SOURCE = 1
PY_COMPILED = 2
C_EXTENSION = 3
C_BUILTIN = 6
PY_FROZEN = 7


def find_spec(module, paths):
    finder = (
        importlib.machinery.PathFinder().find_spec
        if isinstance(paths, list) else
        importlib.util.find_spec
    )
    return finder(module, paths)


def find_module(module, paths=None):
    """Just like 'imp.find_module()', but with package support"""
    spec = find_spec(module, paths)
    if spec is None:
        raise ImportError("Can't find %s" % module)
    if not spec.has_location and hasattr(spec, 'submodule_search_locations'):
        spec = importlib.util.spec_from_loader('__init__.py', spec.loader)

    kind = -1
    file = None
    static = isinstance(spec.loader, type)
    if spec.origin == 'frozen' or static and issubclass(
            spec.loader, importlib.machinery.FrozenImporter):
        kind = PY_FROZEN
        path = None  # imp compabilty
        suffix = mode = ''  # imp compatibility
    elif spec.origin == 'built-in' or static and issubclass(
            spec.loader, importlib.machinery.BuiltinImporter):
        kind = C_BUILTIN
        path = None  # imp compabilty
        suffix = mode = ''  # imp compatibility
    elif spec.has_location:
        path = spec.origin
        suffix = os.path.splitext(path)[1]
        mode = 'r' if suffix in importlib.machinery.SOURCE_SUFFIXES else 'rb'

        if suffix in importlib.machinery.SOURCE_SUFFIXES:
            kind = PY_SOURCE
        elif suffix in importlib.machinery.BYTECODE_SUFFIXES:
            kind = PY_COMPILED
        elif suffix in importlib.machinery.EXTENSION_SUFFIXES:
            kind = C_EXTENSION

        if kind in {PY_SOURCE, PY_COMPILED}:
            file = open(path, mode)
    else:
        path = None
        suffix = mode = ''

    return file, path, (suffix, mode, kind)


def get_frozen_object(module, paths=None):
    spec = find_spec(module, paths)
    if not spec:
        raise ImportError("Can't find %s" % module)
    return spec.loader.get_code(module)


def get_module(module, paths, info):
    spec = find_spec(module, paths)
    if not spec:
        raise ImportError("Can't find %s" % module)
    return module_from_spec(spec)
site-packages/setuptools/unicode_utils.py000064400000001655150564745010014743 0ustar00import unicodedata
import sys


# HFS Plus uses decomposed UTF-8
def decompose(path):
    if isinstance(path, str):
        return unicodedata.normalize('NFD', path)
    try:
        path = path.decode('utf-8')
        path = unicodedata.normalize('NFD', path)
        path = path.encode('utf-8')
    except UnicodeError:
        pass  # Not UTF-8
    return path


def filesys_decode(path):
    """
    Ensure that the given path is decoded,
    NONE when no expected encoding works
    """

    if isinstance(path, str):
        return path

    fs_enc = sys.getfilesystemencoding() or 'utf-8'
    candidates = fs_enc, 'utf-8'

    for enc in candidates:
        try:
            return path.decode(enc)
        except UnicodeDecodeError:
            continue


def try_encode(string, enc):
    "turn unicode encoding into a functional routine"
    try:
        return string.encode(enc)
    except UnicodeEncodeError:
        return None
site-packages/pkg_resources/__pycache__/__init__.cpython-39.pyc000064400000304205150564745010020552 0ustar00a

(�Re���@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlmZzddlZWney�ddlZYn0zeWne �y e!ZYn0ddlm"Z"zddlm#Z#m$Z$m%Z%dZ&Wne�ybdZ&Yn0ddlm'Z(dd	l)m*Z*m+Z+zddl,m-Z.e.j/Wne�y�dZ.Yn0dd
l0m1Z1ddl0m2Z2e3d�e3d
�e3d�e3d�ej4dk�r�e5d��dZ6dZ7dZ8dZ9dZ:dZ;dZ<dZ=dZ>dZ?dZ@dZAdZBdZCdZDdZEdZFGdd�deG�ZHdd�ZIiZJdd�ZKdd�ZLdd�ZMdd�ZNdd�ZOd d!�ZPd"d#�ZQd$d%�ZRZSd&d'�ZTgd(�ZUGd)d*�d*eV�ZWGd+d,�d,eW�ZXGd-d.�d.eX�ZYGd/d0�d0eW�ZZGd1d2�d2eW�Z[iZ\d3j]ej4�Z^d4Z_d5Z`d6ZadZbd7Zcd8d9�Zdd:d;�Zegfd<d=�Zfd>d?�Zgd@dA�Zhe�idB�Zje�idC�ZkehZldDdE�ZmdFdG�ZnenZodHdI�ZpdJdK�Zqd�dLdM�ZrdNdO�ZsGdPdQ�dQ�ZtGdRdS�dSet�ZuGdTdU�dU�ZvGdVdW�dWew�ZxGdXdY�dY�ZyeyZzGdZd[�d[e5�Z{Gd\d]�d]�Z|d^d_�Z}d`da�Z~dbdc�Zddde�Z�dfdg�Z�dhdi�Z�d�djdk�Z�Gdldm�dm�Z�ede�e��dndo�Z�Gdpdq�dqe��Z�Gdrds�dse��Z�e����Gdtdu�due��Z�e��Z�Gdvdw�dwew�Z�Gdxdy�dye��Z�Gdzd{�d{e��Z�edej�e��Gd|d}�d}e��Z�Gd~d�de��Z�Gd�d��d�e��Z�eKd�id��d�d��Z�d�d�d��Z�d�d�d��Z�e�ej�e��d�d�d��Z�e�e�e��d�d��Z�d�d�d��Z�d�d��Z�Gd�d��d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�e�ej�e��e�e.d���r�e�e.j�e��eKd�id��eKd�id��d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d�d��Z�d�d��Z�e�ej�e��e�ej�e��e�e.d���r<e�e.j�e��d�d��Z�e�e�e��d�d��Z�d�d��Z�ifd�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�e�id��j�Z�e�id�ej�ej�B�j�Z�Gd�d��d��Z�d�d„Z�Gd�dĄdăZ�Gd�dƄd�e��Z�Gd�dȄd�e��Z�e�e�e�dɜZ�d�d˄Z�d�d̈́Z�Gd�dτd�e2j�j��Z�Gd�dфd�e2j�j��Z�d�dӄZ�d�dՄZ�d�dׄZ�d�dلZ�d�dۄZ�d�d݄Z�e	j�d�eHddߍd�d�Z�e�eʃfd�d��Z�e�d�d��Z�Gd�d�d�e̓Z�dS)�aZ
Package resource API
--------------------

A resource is a logical file contained within a package, or a logical
subdirectory thereof.  The package resource API expects resource names
to have their path parts separated with ``/``, *not* whatever the local
path separator is.  Do not use os.path operations to manipulate resource
names being passed into the API.

The package resource API is designed to work with normal filesystem packages,
.egg files, and unpacked .egg files.  It can also work in a limited way with
.zip files and with custom PEP 302 loaders that support the ``get_data()``
method.
�N)�get_importer)�utime)�mkdir�rename�unlinkTF)�open)�isdir�split)�appdirs)�	packagingz&pkg_resources.extern.packaging.versionz)pkg_resources.extern.packaging.specifiersz+pkg_resources.extern.packaging.requirementsz&pkg_resources.extern.packaging.markers)��zPython 3.5 or later is requiredc@seZdZdZdS)�
PEP440Warningza
    Used when there is an issue with a version or specifier not complying with
    PEP 440.
    N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/__init__.pyrisrcCs6ztj�|�WStjjy0tj�|�YS0dS�N)r�version�Version�InvalidVersion�
LegacyVersion)�vrrr�
parse_versionpsrcKs"t��|�t�t�||��dSr)�globals�update�_state_vars�dict�fromkeys)�vartype�kwrrr�_declare_statezsr$cCs8i}t�}t��D] \}}|d|||�||<q|S)N�_sget_)rr�items��state�g�krrrr�__getstate__s
r+cCs8t�}|��D]$\}}|dt|||||�q|S)N�_sset_)rr&rr'rrr�__setstate__�sr-cCs|��Sr)�copy��valrrr�
_sget_dict�sr1cCs|��|�|�dSr)�clearr��key�obr(rrr�
_sset_dict�sr6cCs|��Sr)r+r/rrr�_sget_object�sr7cCs|�|�dSr)r-r3rrr�_sset_object�sr8cGsdSrr��argsrrr�<lambda>��r;cCs`t�}t�|�}|dur\tjdkr\z&dd�t�dd��|�d�f}WntyZYn0|S)aQReturn this platform's maximum compatible version.

    distutils.util.get_platform() normally reports the minimum version
    of macOS that would be required to *use* extensions produced by
    distutils.  But what we want when checking compatibility is to know the
    version of macOS that we are *running*.  To allow usage of packages that
    explicitly require a newer version of macOS, we must also know the
    current version of the OS.

    If this condition occurs for any other platform with a version in its
    platform strings, this function should be extended accordingly.
    N�darwinzmacosx-%s-%s�.�r)	�get_build_platform�macosVersionString�match�sys�platform�join�_macos_vers�group�
ValueError)�plat�mrrr�get_supported_platform�s

&rK)G�require�
run_script�get_provider�get_distribution�load_entry_point�
get_entry_map�get_entry_info�iter_entry_points�resource_string�resource_stream�resource_filename�resource_listdir�resource_exists�resource_isdir�declare_namespace�working_set�add_activation_listener�find_distributions�set_extraction_path�cleanup_resources�get_default_cache�Environment�
WorkingSet�ResourceManager�Distribution�Requirement�
EntryPoint�ResolutionError�VersionConflict�DistributionNotFound�UnknownExtra�ExtractionErrorr�parse_requirementsr�	safe_name�safe_version�get_platform�compatible_platforms�yield_lines�split_sections�
safe_extra�to_filename�invalid_marker�evaluate_marker�ensure_directory�normalize_path�EGG_DIST�BINARY_DIST�SOURCE_DIST�
CHECKOUT_DIST�DEVELOP_DIST�IMetadataProvider�IResourceProvider�FileMetadata�PathMetadata�EggMetadata�
EmptyProvider�empty_provider�NullProvider�EggProvider�DefaultProvider�ZipProvider�register_finder�register_namespace_handler�register_loader_type�fixup_namespace_packagesr�PkgResourcesDeprecationWarning�run_main�AvailableDistributionsc@seZdZdZdd�ZdS)rgz.Abstract base for dependency resolution errorscCs|jjt|j�Sr)�	__class__r�reprr:��selfrrr�__repr__�szResolutionError.__repr__N)rrrrr�rrrrrg�srgc@s<eZdZdZdZedd��Zedd��Zdd�Zd	d
�Z	dS)rhz�
    An already-installed version conflicts with the requested version.

    Should be initialized with the installed Distribution and the requested
    Requirement.
    z3{self.dist} is installed but {self.req} is requiredcCs
|jdS�Nrr9r�rrr�dist�szVersionConflict.distcCs
|jdS�N�r9r�rrr�reqszVersionConflict.reqcCs|jjfit���Sr��	_template�format�localsr�rrr�reportszVersionConflict.reportcCs|s|S|j|f}t|�S)zt
        If required_by is non-empty, return a version of self that is a
        ContextualVersionConflict.
        )r:�ContextualVersionConflict)r��required_byr:rrr�with_contextszVersionConflict.with_contextN)
rrrrr��propertyr�r�r�r�rrrrrh�s

rhc@s&eZdZdZejdZedd��ZdS)r�z�
    A VersionConflict that accepts a third parameter, the set of the
    requirements that required the installed Distribution.
    z by {self.required_by}cCs
|jdS)Nr?r9r�rrrr�sz%ContextualVersionConflict.required_byN)rrrrrhr�r�r�rrrrr�s
r�c@sHeZdZdZdZedd��Zedd��Zedd��Zd	d
�Z	dd�Z
d
S)riz&A requested distribution was not foundzSThe '{self.req}' distribution was not found and is required by {self.requirers_str}cCs
|jdSr�r9r�rrrr�&szDistributionNotFound.reqcCs
|jdSr�r9r�rrr�	requirers*szDistributionNotFound.requirerscCs|js
dSd�|j�S)Nzthe applicationz, )r�rEr�rrr�
requirers_str.sz"DistributionNotFound.requirers_strcCs|jjfit���Srr�r�rrrr�4szDistributionNotFound.reportcCs|��Sr)r�r�rrr�__str__7szDistributionNotFound.__str__N)rrrrr�r�r�r�r�r�r�rrrrri s


ric@seZdZdZdS)rjz>Distribution doesn't have an "extra feature" of the given nameNrrrrrrj;srjz{}.{}rr?r����cCs|t|<dS)aRegister `provider_factory` to make providers for `loader_type`

    `loader_type` is the type or class of a PEP 302 ``module.__loader__``,
    and `provider_factory` is a function that, passed a *module* object,
    returns an ``IResourceProvider`` for that module.
    N)�_provider_factories)�loader_type�provider_factoryrrrr�Isr�cCsrt|t�r$t�|�p"tt|��dSztj|}Wn$tyVt	|�tj|}Yn0t
|dd�}tt|�|�S)z?Return an IResourceProvider for the named module or requirementr�
__loader__N)
�
isinstancerer[�findrL�strrC�modules�KeyError�
__import__�getattr�
_find_adapterr�)�moduleOrReq�module�loaderrrrrNSs
rNcCsd|s\t��d}|dkrLd}tj�|�rLttd�rLt�|�}d|vrL|d}|�|�	d��|dS)Nr�z0/System/Library/CoreServices/SystemVersion.plist�	readPlist�ProductVersionr>)
rD�mac_ver�os�path�exists�hasattr�plistlibr��appendr	)�_cacher�plist�
plist_contentrrrrF`s

rFcCsddd��||�S)N�ppc)�PowerPC�Power_Macintosh)�get)�machinerrr�_macos_archpsr�cCs|ddlm}|�}tjdkrx|�d�sxz>t�}t��d�dd�}dt	|d�t	|d	�t
|�fWStyvYn0|S)
z�Return this platform's string for platform-specific distributions

    XXX Currently this is the same as ``distutils.util.get_platform()``, but it
    needs some hacks for Linux and macOS.
    r)ror=zmacosx-�� �_zmacosx-%d.%d-%sr�)�	sysconfigrorCrD�
startswithrFr��uname�replace�intr�rH)rorIrr�rrrr@ts�r@zmacosx-(\d+)\.(\d+)-(.*)zdarwin-(\d+)\.(\d+)\.(\d+)-(.*)cCs�|dus|dus||krdSt�|�}|r�t�|�}|s�t�|�}|r�t|�d��}d|�d�|�d�f}|dkr||dks�|dkr�|d	kr�dSd
S|�d�|�d�ks�|�d�|�d�kr�d
St|�d��t|�d��kr�d
SdSd
S)z�Can code for the `provided` platform run on the `required` platform?

    Returns true if either platform is ``None``, or the platforms are equal.

    XXX Needs compatibility checks for Linux and other unixy OSes.
    NTr�z%s.%sr?�z10.3�z10.4Fr)rArB�darwinVersionStringr�rG)�provided�required�reqMac�provMac�
provDarwin�dversion�macosversionrrrrp�s2


���rpcCs<t�d�j}|d}|��||d<t|�d�||�dS)z@Locate distribution `dist_spec` and run its `script_name` scriptr�rrN�rC�	_getframe�	f_globalsr2rLrM)Z	dist_spec�script_name�ns�namerrrrM�s
rMcCs>t|t�rt�|�}t|t�r&t|�}t|t�s:td|��|S)z@Return a current distribution object for a Requirement or stringz-Expected string, Requirement, or Distribution)r�r�re�parserNrd�	TypeError�r�rrrrO�s




rOcCst|��||�S)zDReturn `name` entry point of `group` for `dist` or raise ImportError)rOrP�r�rGr�rrrrP�srPcCst|��|�S)�=Return the entry point map for `group`, or the full entry map)rOrQ)r�rGrrrrQ�srQcCst|��||�S�z<Return the EntryPoint object for `group`+`name`, or ``None``)rOrRr�rrrrR�srRc@s<eZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
S)r~cCsdS)z;Does the package's distribution contain the named metadata?Nr�r�rrr�has_metadata�szIMetadataProvider.has_metadatacCsdS)z'The named metadata resource as a stringNrr�rrr�get_metadata�szIMetadataProvider.get_metadatacCsdS)z�Yield named metadata resource as list of non-blank non-comment lines

       Leading and trailing whitespace is stripped from each line, and lines
       with ``#`` as the first non-blank character are omitted.Nrr�rrr�get_metadata_lines�sz$IMetadataProvider.get_metadata_linescCsdS)z>Is the named metadata a directory?  (like ``os.path.isdir()``)Nrr�rrr�metadata_isdir�sz IMetadataProvider.metadata_isdircCsdS)z?List of metadata names in the directory (like ``os.listdir()``)Nrr�rrr�metadata_listdir�sz"IMetadataProvider.metadata_listdircCsdS)z=Execute the named script in the supplied namespace dictionaryNr)r��	namespacerrrrM�szIMetadataProvider.run_scriptN)	rrrr�r�r�r�r�rMrrrrr~�sr~c@s@eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dS)rz3An object that provides access to package resourcescCsdS)zdReturn a true filesystem path for `resource_name`

        `manager` must be an ``IResourceManager``Nr��manager�
resource_namerrr�get_resource_filenamesz'IResourceProvider.get_resource_filenamecCsdS)ziReturn a readable file-like object for `resource_name`

        `manager` must be an ``IResourceManager``Nrr�rrr�get_resource_streamsz%IResourceProvider.get_resource_streamcCsdS)zmReturn a string containing the contents of `resource_name`

        `manager` must be an ``IResourceManager``Nrr�rrr�get_resource_stringsz%IResourceProvider.get_resource_stringcCsdS)z,Does the package contain the named resource?Nr�r�rrr�has_resourceszIResourceProvider.has_resourcecCsdS)z>Is the named resource a directory?  (like ``os.path.isdir()``)Nrr�rrrrYsz IResourceProvider.resource_isdircCsdS)z?List of resource names in the directory (like ``os.listdir()``)Nrr�rrrrWsz"IResourceProvider.resource_listdirN)
rrrrr�r�r�r�rYrWrrrrr�src@s�eZdZdZd'dd�Zedd��Zedd��Zd	d
�Zdd�Z	d
d�Z
d(dd�Zdd�Zdd�Z
d)dd�Zd*dd�Zd+dd�Zdd�Zd,dd �Zd!d"�Zd#d$�Zd%d&�ZdS)-rbzDA collection of active distributions on sys.path (or a similar list)NcCs>g|_i|_i|_g|_|dur&tj}|D]}|�|�q*dS)z?Create working set from list of path entries (default=sys.path)N)�entries�
entry_keys�by_key�	callbacksrCr��	add_entry)r�r��entryrrr�__init__szWorkingSet.__init__cCs^|�}zddlm}Wnty,|YS0z|�|�WntyX|�|�YS0|S)z1
        Prepare the master working set.
        r)�__requires__)�__main__r��ImportErrorrLrh�_build_from_requirements)�cls�wsr�rrr�
_build_master*s
zWorkingSet._build_mastercCsf|g�}t|�}|�|t��}|D]}|�|�q"tjD]}||jvr8|�|�q8|jtjdd�<|S)zQ
        Build a working set from a requirement spec. Rewrites sys.path.
        N)rl�resolvera�addrCr�r�r�)r�req_specr�reqs�distsr�r�rrrr>s

z#WorkingSet._build_from_requirementscCs<|j�|g�|j�|�t|d�D]}|�||d�q$dS)a�Add a path item to ``.entries``, finding any distributions on it

        ``find_distributions(entry, True)`` is used to find distributions
        corresponding to the path entry, and they are added.  `entry` is
        always appended to ``.entries``, even if it is already present.
        (This is because ``sys.path`` can contain the same value more than
        once, and the ``.entries`` of the ``sys.path`` WorkingSet should always
        equal ``sys.path``.)
        TFN)r��
setdefaultr�r�r]r)r�r�r�rrrr�Ts
zWorkingSet.add_entrycCs|j�|j�|kS)z9True if `dist` is the active distribution for its project)r�r�r4�r�r�rrr�__contains__cszWorkingSet.__contains__cCs,|j�|j�}|dur(||vr(t||��|S)a�Find a distribution matching requirement `req`

        If there is an active distribution for the requested project, this
        returns it as long as it meets the version requirement specified by
        `req`.  But, if there is an active distribution for the project and it
        does *not* meet the `req` requirement, ``VersionConflict`` is raised.
        If there is no active distribution for the requested project, ``None``
        is returned.
        N)r�r�r4rh)r�r�r�rrrr�gs

zWorkingSet.findcs��fdd�|D�S)aYield entry point objects from `group` matching `name`

        If `name` is None, yields all entry points in `group` from all
        distributions in the working set, otherwise only ones matching
        both `group` and `name` are yielded (in distribution order).
        c3s8|]0}|�����D]}�dus*�|jkr|VqqdSr)rQ�valuesr�)�.0r�r��rGr�rr�	<genexpr>~s�z/WorkingSet.iter_entry_points.<locals>.<genexpr>r�r�rGr�rrrrSws�zWorkingSet.iter_entry_pointscCs>t�d�j}|d}|��||d<|�|�d�||�dS)z?Locate distribution for `requires` and run `script_name` scriptr�rrNr�)r��requiresr�r�r�rrrrM�s
zWorkingSet.run_scriptccsLi}|jD]<}||jvrq
|j|D] }||vr$d||<|j|Vq$q
dS)z�Yield distributions for non-duplicate projects in the working set

        The yield order is the order in which the items' path entries were
        added to the working set.
        r�N)r�r�r�)r��seen�itemr4rrr�__iter__�s

zWorkingSet.__iter__TFcCs�|r|j|j||d�|dur$|j}|j�|g�}|j�|jg�}|sV|j|jvrVdS||j|j<|j|vrx|�|j�|j|vr�|�|j�|�|�dS)aAdd `dist` to working set, associated with `entry`

        If `entry` is unspecified, it defaults to the ``.location`` of `dist`.
        On exit from this routine, `entry` is added to the end of the working
        set's ``.entries`` (if it wasn't already present).

        `dist` is only added to the working set if it's for a project that
        doesn't already have a distribution in the set, unless `replace=True`.
        If it's added, any callbacks registered with the ``subscribe()`` method
        will be called.
        �r�N)	�	insert_onr��locationr�r	r4r�r��
_added_new)r�r�r��insertr��keys�keys2rrrr�s

zWorkingSet.addcCsxt|�ddd�}i}i}g}t�}	t�t�}
|�rt|�d�}||vrHq.|	�||�sVq.|�|j�}|du�r|j	�|j�}|dus�||vr�|r�|}
|dur�|dur�t
|j�}nt
g�}tg�}
|j
||
||d�}||j<|dur�|
�|d�}t||��|�|�||v�r$|
|}t||��|��|�|j�ddd�}|�|�|D] }|
|�|j�|j|	|<�qHd||<q.|S)a�List all distributions needed to (recursively) meet `requirements`

        `requirements` must be a sequence of ``Requirement`` objects.  `env`,
        if supplied, should be an ``Environment`` instance.  If
        not supplied, it defaults to all distributions available within any
        entry or distribution in the working set.  `installer`, if supplied,
        will be invoked with each requirement that cannot be met by an
        already-installed distribution; it should return a ``Distribution`` or
        ``None``.

        Unless `replace_conflicting=True`, raises a VersionConflict exception
        if
        any requirements are found on the path that have the correct name but
        the wrong version.  Otherwise, if an `installer` is supplied it will be
        invoked to obtain the correct version of the requirement and activate
        it.

        `extras` is a list of the extras to be used with these requirements.
        This is important because extra requirements may look like `my_req;
        extra = "my_extra"`, which would otherwise be interpreted as a purely
        optional requirement.  Instead, we want to be able to assert that these
        requirements are truly required.
        Nr�r)�replace_conflictingT)�list�
_ReqExtras�collections�defaultdict�set�pop�markers_passr�r4r�rar�rb�
best_matchrir�rhr�r�extras�extendr�project_name)r��requirements�env�	installerrr%�	processed�best�to_activate�
req_extrasr�r�r�rr��
dependent_req�new_requirements�new_requirementrrrr�sP


�




zWorkingSet.resolvecCst|�}|��i}i}|dur4t|j�}||7}n||}|�g�}	tt|	j|��|D]�}
||
D]�}|��g}z|	�|||�}
WnLt	y�}z4|||<|r�WYd}~qfnWYd}~qZWYd}~qfd}~00tt|	j|
��|�
t�|
��qZqfqZt|�}|��||fS)asFind all activatable distributions in `plugin_env`

        Example usage::

            distributions, errors = working_set.find_plugins(
                Environment(plugin_dirlist)
            )
            # add plugins+libs to sys.path
            map(working_set.add, distributions)
            # display errors
            print('Could not load', errors)

        The `plugin_env` should be an ``Environment`` instance that contains
        only distributions that are in the project's "plugin directory" or
        directories. The `full_env`, if supplied, should be an ``Environment``
        contains all currently-available distributions.  If `full_env` is not
        supplied, one is created automatically from the ``WorkingSet`` this
        method is called on, which will typically mean that every directory on
        ``sys.path`` will be scanned for distributions.

        `installer` is a standard installer callback as used by the
        ``resolve()`` method. The `fallback` flag indicates whether we should
        attempt to resolve older versions of a plugin if the newest version
        cannot be resolved.

        This method returns a 2-tuple: (`distributions`, `error_info`), where
        `distributions` is a list of the distributions found in `plugin_env`
        that were loadable, along with any other distributions that are needed
        to resolve their dependencies.  `error_info` is a dictionary mapping
        unloadable plugin distributions to an exception instance describing the
        error that occurred. Usually this will be a ``DistributionNotFound`` or
        ``VersionConflict`` instance.
        N)
r�sortrar�r��mapr�as_requirementrrgrr r!)r��
plugin_env�full_envr*�fallback�plugin_projects�
error_info�
distributionsr)�
shadow_setr'r�r��	resolveesrrrr�find_pluginss4$



$zWorkingSet.find_pluginscGs&|�t|��}|D]}|�|�q|S)a�Ensure that distributions matching `requirements` are activated

        `requirements` must be a string or a (possibly-nested) sequence
        thereof, specifying the distributions and versions required.  The
        return value is a sequence of the distributions that needed to be
        activated to fulfill the requirements; all relevant distributions are
        included, even if they were already activated in this working set.
        )rrlr)r�r(�neededr�rrrrLms	zWorkingSet.requirecCs8||jvrdS|j�|�|s"dS|D]}||�q&dS)z�Invoke `callback` for all distributions

        If `existing=True` (default),
        call on all existing ones, as well.
        N)r�r�)r��callback�existingr�rrr�	subscribe}s
zWorkingSet.subscribecCs|jD]}||�qdSr)r�)r�r�r?rrrr�s
zWorkingSet._added_newcCs,|jdd�|j��|j��|jdd�fSr)r�r�r.r�r�r�rrrr+�s�zWorkingSet.__getstate__cCs@|\}}}}|dd�|_|��|_|��|_|dd�|_dSr)r�r.r�r�r�)r��e_k_b_cr�rr�r�rrrr-�s


zWorkingSet.__setstate__)N)N)NTF)NNFN)NNT)T)rrrrr��classmethodrrr�rr�rSrMrrrr=rLrArr+r-rrrrrbs.





�
]�
T
rbc@seZdZdZddd�ZdS)rz>
    Map each requirement to the extras that demanded it.
    Ncs2�fdd�|��d�|pdD�}�jp0t|�S)z�
        Evaluate markers for req against each extra that
        demanded it.

        Return False if the req has a marker and fails
        evaluation. Otherwise, return True.
        c3s|]}�j�d|i�VqdS)�extraN��marker�evaluate)r
rD�r�rrr�s�z*_ReqExtras.markers_pass.<locals>.<genexpr>rr)r�rF�any)r�r�r%�extra_evalsrrHrr#�s
�z_ReqExtras.markers_pass)N)rrrrr#rrrrr�src@sxeZdZdZde�efdd�Zdd�Zdd�Zdd	d
�Z	dd�Z
d
d�Zddd�Zddd�Z
dd�Zdd�Zdd�ZdS)raz5Searchable snapshot of distributions on a search pathNcCs i|_||_||_|�|�dS)a!Snapshot distributions available on a search path

        Any distributions found on `search_path` are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.

        `platform` is an optional string specifying the name of the platform
        that platform-specific distributions must be compatible with.  If
        unspecified, it defaults to the current platform.  `python` is an
        optional string naming the desired version of Python (e.g. ``'3.6'``);
        it defaults to the current version.

        You may explicitly set `platform` (and/or `python`) to ``None`` if you
        wish to map *all* distributions, not just those compatible with the
        running platform or Python version.
        N)�_distmaprD�python�scan)r��search_pathrDrLrrrr��szEnvironment.__init__cCs2|jdup|jdup|j|jk}|o0t|j|j�S)z�Is distribution `dist` acceptable for this environment?

        The distribution must match the platform and python version
        requirements specified when this environment was created, or False
        is returned.
        N)rL�
py_versionrprD)r�r��	py_compatrrr�can_add�s
�
�zEnvironment.can_addcCs|j|j�|�dS)z"Remove `dist` from the environmentN)rKr4�remover
rrrrR�szEnvironment.removecCs4|durtj}|D]}t|�D]}|�|�qqdS)adScan `search_path` for distributions usable in this environment

        Any distributions found are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.  Only distributions conforming to
        the platform/python version defined at initialization are added.
        N)rCr�r]r)r�rNrr�rrrrM�s
zEnvironment.scancCs|��}|j�|g�S)aReturn a newest-to-oldest list of distributions for `project_name`

        Uses case-insensitive `project_name` comparison, assuming all the
        project's distributions use their project's name converted to all
        lowercase as their key.

        )�lowerrKr�)r�r'�distribution_keyrrr�__getitem__�szEnvironment.__getitem__cCsL|�|�rH|��rH|j�|jg�}||vrH|�|�|jt�d�dd�dS)zLAdd `dist` if we ``can_add()`` it and it has not already been added
        �hashcmpT�r4�reverseN)	rQ�has_versionrKr	r4r�r2�operator�
attrgetter)r�r�rrrrr�s

zEnvironment.addFcCsdz|�|�}Wnty*|s"�d}Yn0|dur8|S||jD]}||vrB|SqB|�||�S)a�Find distribution best matching `req` and usable on `working_set`

        This calls the ``find(req)`` method of the `working_set` to see if a
        suitable distribution is already active.  (This may raise
        ``VersionConflict`` if an unsuitable version of the project is already
        active in the specified `working_set`.)  If a suitable distribution
        isn't active, this method returns the newest distribution in the
        environment that meets the ``Requirement`` in `req`.  If no suitable
        distribution is found, and `installer` is supplied, then the result of
        calling the environment's ``obtain(req, installer)`` method will be
        returned.
        N)r�rhr4�obtain)r�r�r[r*rr�rrrr$s

zEnvironment.best_matchcCs|dur||�SdS)a�Obtain a distribution matching `requirement` (e.g. via download)

        Obtain a distro that matches requirement (e.g. via download).  In the
        base ``Environment`` class, this routine just returns
        ``installer(requirement)``, unless `installer` is None, in which case
        None is returned instead.  This method is a hook that allows subclasses
        to attempt other ways of obtaining a distribution before falling back
        to the `installer` argument.Nr)r��requirementr*rrrr\s	zEnvironment.obtainccs"|j��D]}||r
|Vq
dS)z=Yield the unique project names of the available distributionsN)rKr�r�r4rrrr)szEnvironment.__iter__cCsVt|t�r|�|�n<t|t�rD|D]}||D]}|�|�q0q$ntd|f��|S)z2In-place addition of a distribution or environmentzCan't add %r to environment)r�rdrrar�)r��other�projectr�rrr�__iadd__/s

zEnvironment.__iadd__cCs*|jgddd�}||fD]}||7}q|S)z4Add an environment or distribution to an environmentN)rDrL�r�)r�r_�newr)rrr�__add__;s
zEnvironment.__add__)N)NF)N)rrrrrK�PY_MAJORr�rQrRrMrUrr$r\rrardrrrrra�s�


�

rac@seZdZdZdS)rkaTAn error occurred extracting a resource

    The following attributes are available from instances of this exception:

    manager
        The resource manager that raised this exception

    cache_path
        The base directory for resource extraction

    original_error
        The exception instance that caused extraction to fail
    NrrrrrrkGsrkc@s�eZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zddd�Z
edd��Zdd�Zdd�Zd dd�ZdS)!rcz'Manage resource extraction and packagesNcCs
i|_dSr)�cached_filesr�rrrr�[szResourceManager.__init__cCst|��|�S)zDoes the named resource exist?)rNr��r��package_or_requirementr�rrrrX^szResourceManager.resource_existscCst|��|�S)z,Is the named resource an existing directory?)rNrYrgrrrrYbs�zResourceManager.resource_isdircCst|��||�S)z4Return a true filesystem path for specified resource)rNr�rgrrrrVhs�z!ResourceManager.resource_filenamecCst|��||�S)z9Return a readable file-like object for specified resource)rNr�rgrrrrUns�zResourceManager.resource_streamcCst|��||�S)z%Return specified resource as a string)rNr�rgrrrrTts�zResourceManager.resource_stringcCst|��|�S)z1List the contents of the named resource directory)rNrWrgrrrrWzs�z ResourceManager.resource_listdircCsVt��d}|jpt�}t�d���}t|jfit	����}||_
||_||_|�dS)z5Give an error message for problems extracting file(s)r�a
            Can't extract file(s) to egg cache

            The following error occurred while trying to extract file(s)
            to the Python egg cache:

              {old_exc}

            The Python egg cache directory is currently set to:

              {cache_path}

            Perhaps your account does not have write access to this directory?
            You can change the cache directory by setting the PYTHON_EGG_CACHE
            environment variable to point to an accessible directory.
            N)
rC�exc_info�extraction_pathr`�textwrap�dedent�lstriprkr�r�r��
cache_path�original_error)r��old_excrn�tmpl�errrrr�extraction_error�sz ResourceManager.extraction_errorrcCsf|jp
t�}tjj||dg|�R�}zt|�WntyL|��Yn0|�|�d|j	|<|S)a�Return absolute location in cache for `archive_name` and `names`

        The parent directory of the resulting path will be created if it does
        not already exist.  `archive_name` should be the base filename of the
        enclosing egg (which may not be the name of the enclosing zipfile!),
        including its ".egg" extension.  `names`, if provided, should be a
        sequence of path name parts "under" the egg's extraction location.

        This method should only be called by resource providers that need to
        obtain an extraction location, and only for names they intend to
        extract, as it tracks the generated names for possible cleanup later.
        z-tmpr�)
rjr`r�r�rE�_bypass_ensure_directory�	Exceptionrs�_warn_unsafe_extraction_pathrf)r��archive_name�names�extract_path�target_pathrrr�get_cache_path�s


zResourceManager.get_cache_pathcCs`tjdkr|�tjd�sdSt�|�j}|tj@s>|tj@r\djfit	���}t
�|t�dS)aN
        If the default extraction path is overridden and set to an insecure
        location, such as /tmp, it opens up an opportunity for an attacker to
        replace an extracted file with an unauthorized payload. Warn the user
        if a known insecure location is used.

        See Distribute #375 for more details.
        �nt�windirNz�Extraction path is writable by group/others and vulnerable to attack when used with get_resource_filename ({path}). Consider a more secure location (set with .set_extraction_path or the PYTHON_EGG_CACHE environment variable).)
r�r�r��environ�stat�st_mode�S_IWOTH�S_IWGRPr�r��warnings�warn�UserWarning)r��mode�msgrrrrv�s
��z,ResourceManager._warn_unsafe_extraction_pathcCs.tjdkr*t�|�jdBd@}t�||�dS)a4Perform any platform-specific postprocessing of `tempname`

        This is where Mac header rewrites should be done; other platforms don't
        have anything special they should do.

        Resource providers should call this method ONLY after successfully
        extracting a compressed resource.  They must NOT call it on resources
        that are already in the filesystem.

        `tempname` is the current (temporary) name of the file, and `filename`
        is the name it will be renamed to by the caller after this routine
        returns.
        �posiximi�N)r�r�rr��chmod)r��tempname�filenamer�rrr�postprocess�s
zResourceManager.postprocesscCs|jrtd��||_dS)a�Set the base path where resources will be extracted to, if needed.

        If you do not call this routine before any extractions take place, the
        path defaults to the return value of ``get_default_cache()``.  (Which
        is based on the ``PYTHON_EGG_CACHE`` environment variable, with various
        platform-specific fallbacks.  See that routine's documentation for more
        details.)

        Resources are extracted to subdirectories of this path based upon
        information given by the ``IResourceProvider``.  You may set this to a
        temporary directory, but then you must call ``cleanup_resources()`` to
        delete the extracted files when done.  There is no guarantee that
        ``cleanup_resources()`` will be able to remove all extracted files.

        (Note: you may not change the extraction path for a given resource
        manager once resources have been extracted, unless you first call
        ``cleanup_resources()``.)
        z5Can't change extraction path, files already extractedN)rfrHrj�r�r�rrrr^�s
�z#ResourceManager.set_extraction_pathFcCsdS)aB
        Delete all extracted resource files and directories, returning a list
        of the file and directory names that could not be successfully removed.
        This function does not have any concurrency protection, so it should
        generally only be called when the extraction path is a temporary
        directory exclusive to a single process.  This method is not
        automatically called; you must call it explicitly or register it as an
        ``atexit`` function if you wish to ensure cleanup of a temporary
        directory used for extractions.
        Nr)r��forcerrrr_�sz!ResourceManager.cleanup_resources)r)F)rrrrrjr�rXrYrVrUrTrWrsr{�staticmethodrvr�r^r_rrrrrcWs 

rccCstj�d�ptjdd�S)z�
    Return the ``PYTHON_EGG_CACHE`` environment variable
    or a platform-relevant user cache dir for an app
    named "Python-Eggs".
    �PYTHON_EGG_CACHEzPython-Eggs)�appname)r�r~r�r
�user_cache_dirrrrrr`s
�r`cCst�dd|�S)z�Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    �[^A-Za-z0-9.]+�-)�re�subr�rrrrmsrmcCsHzttj�|��WStjjyB|�dd�}t�dd|�YS0dS)zB
    Convert an arbitrary string to a standard version string
    r�r>r�r�N)r�rrrrr�r�r�)rrrrrn s
rncCst�dd|���S)z�Convert an arbitrary string to a standard 'extra' name

    Any runs of non-alphanumeric characters are replaced with a single '_',
    and the result is always lowercased.
    z[^A-Za-z0-9.-]+r�)r�r�rS)rDrrrrs,srscCs|�dd�S)z|Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    r�r�rr�rrrrt5srtc
CsFzt|�Wn4ty@}zd|_d|_|WYd}~Sd}~00dS)zo
    Validate text as a PEP 508 environment marker; return an exception
    if invalid or False otherwise.
    NF)rv�SyntaxErrorr��lineno)�text�errrru=sruc
CsNztj�|�}|��WStjjyH}zt|�|�WYd}~n
d}~00dS)z�
    Evaluate a PEP 508 environment marker.
    Return a boolean indicating the marker result in this environment.
    Raise SyntaxError if marker is invalid.

    This implementation uses the 'pyparsing' module.
    N)r�markers�MarkerrG�
InvalidMarkerr�)r�rDrFr�rrrrvKs

rvc@s�eZdZdZdZdZdZdd�Zdd�Zdd�Z	d	d
�Z
dd�Zd
d�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zed'd(��Zd)d*�ZdS)+r�zETry to implement resources and metadata for arbitrary PEP 302 loadersNcCs(t|dd�|_tj�t|dd��|_dS)Nr��__file__r�)r�r�r�r��dirname�module_path�r�r�rrrr�aszNullProvider.__init__cCs|�|j|�Sr)�_fnr��r�r�r�rrrr�esz"NullProvider.get_resource_filenamecCst�|�||��Sr)�io�BytesIOr�r�rrrr�hsz NullProvider.get_resource_streamcCs|�|�|j|��Sr)�_getr�r�r�rrrr�ksz NullProvider.get_resource_stringcCs|�|�|j|��Sr)�_hasr�r��r�r�rrrr�nszNullProvider.has_resourcecCs|�|j|�Sr)r��egg_info�r�r�rrr�_get_metadata_pathqszNullProvider._get_metadata_pathcCs |js|jS|�|�}|�|�Sr)r�r�r��r�r�r�rrrr�ts
zNullProvider.has_metadatac
Csl|js
dS|�|�}|�|�}z|�d�WStyf}z$|jd�||�7_�WYd}~n
d}~00dS)Nr��utf-8z in {} file at path: {})r�r�r��decode�UnicodeDecodeError�reasonr�)r�r�r��value�excrrrr�{s

zNullProvider.get_metadatacCst|�|��Sr�rqr�r�rrrr��szNullProvider.get_metadata_linescCs|�|�|j|��Sr)�_isdirr�r�r�rrrrY�szNullProvider.resource_isdircCs|jo|�|�|j|��Sr)r�r�r�r�rrrr��szNullProvider.metadata_isdircCs|�|�|j|��Sr)�_listdirr�r�r�rrrrW�szNullProvider.resource_listdircCs|jr|�|�|j|��SgSr)r�r�r�r�rrrr��szNullProvider.metadata_listdircCs�d|}|�|�s(tdjfit�����|�|��dd�}|�dd�}|�|j|�}||d<tj	�
|�r�t|��}|��}Wd�n1s�0Yt
||d�}t|||�n>dd	lm}	t|�d|�d�|f|	|<t
||d�}
t|
||�dS)
Nzscripts/z<Script {script!r} not found in metadata at {self.egg_info!r}z
�
�
r��execr)�cache)r�rgr�r�r�r�r�r�r�r�r�r�read�compiler��	linecacher��lenr	)r�r�r��script�script_text�script_filenameZfid�source�coder��script_coderrrrM�s*
��
&�zNullProvider.run_scriptcCstd��dS�Nz9Can't perform this operation for unregistered loader type��NotImplementedErrorr�rrrr��s�zNullProvider._hascCstd��dSr�r�r�rrrr��s�zNullProvider._isdircCstd��dSr�r�r�rrrr��s�zNullProvider._listdircCs,|�|�|r(tjj|g|�d��R�S|S)N�/)�_validate_resource_pathr�r�rEr	)r��baser�rrrr��s
zNullProvider._fncCsptjj|�tj�vp&t�|�p&t�|�}|s0dSd}t�|�rPt�|�sPt|��t	j
|dd�dtdd�dS)aO
        Validate the resource paths according to the docs.
        https://setuptools.readthedocs.io/en/latest/pkg_resources.html#basic-resource-access

        >>> warned = getfixture('recwarn')
        >>> warnings.simplefilter('always')
        >>> vrp = NullProvider._validate_resource_path
        >>> vrp('foo/bar.txt')
        >>> bool(warned)
        False
        >>> vrp('../foo/bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('/foo/bar.txt')
        >>> bool(warned)
        True
        >>> vrp('foo/../../bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('foo/f../bar.txt')
        >>> bool(warned)
        False

        Windows path separators are straight-up disallowed.
        >>> vrp(r'\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path is not allowed.

        >>> vrp(r'C:\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path is not allowed.

        Blank values are allowed

        >>> vrp('')
        >>> bool(warned)
        False

        Non-string values are not.

        >>> vrp(None)
        Traceback (most recent call last):
        ...
        AttributeError: ...
        Nz=Use of .. or absolute path in a resource path is not allowed.r�z/ and will raise exceptions in a future release.r���
stacklevel)r�r��pardirr	�	posixpath�sep�isabs�ntpathrHr�r��DeprecationWarning)r��invalidr�rrrr��s6���z$NullProvider._validate_resource_pathcCs$t|jd�r|j�|�Std��dS)N�get_dataz=Can't perform this operation for loaders without 'get_data()')r�r�r�r�r�rrrr�s
�zNullProvider._get)rrrr�egg_namer�r�r�r�r�r�r�r�r�r�r�rYr�rWr�rMr�r�r�r�r�r�r�rrrrr�Zs2

Jr�ccs,d}||kr(|V|}tj�|�\}}qdS)z2
    yield all parents of path including path
    N)r�r�r	)r��lastr�rrr�_parentss
r�c@s(eZdZdZdd�Zdd�Zdd�ZdS)	r�z&Provider based on a virtual filesystemcCst�||�|��dSr)r�r��
_setup_prefixr�rrrr�*szEggProvider.__init__cCs,ttt|j��}t|d�}|o&|�|�dSr)�filter�_is_egg_pathr�r��next�_set_egg)r��eggs�eggrrrr�.s
zEggProvider._setup_prefixcCs(tj�|�|_tj�|d�|_||_dS)N�EGG-INFO)r�r��basenamer�rEr��egg_rootr�rrrr�5szEggProvider._set_eggN)rrrrr�r�r�rrrrr�'sr�c@sDeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Ze	dd
��Z
dS)r�z6Provides access to package resources in the filesystemcCstj�|�Sr)r�r�r�r�rrrr�>szDefaultProvider._hascCstj�|�Sr)r�r�rr�rrrr�AszDefaultProvider._isdircCs
t�|�Sr)r��listdirr�rrrr�DszDefaultProvider._listdircCst|�|j|�d�S�N�rb)rr�r�r�rrrr�Gsz#DefaultProvider.get_resource_streamcCs6t|d��}|��Wd�S1s(0YdSr�)rr�)r�r��streamrrrr�JszDefaultProvider._getcCs,d}|D]}tt|td��}t||�qdS)N)�SourceFileLoader�SourcelessFileLoader)r��importlib_machinery�typer�)r�loader_namesr��
loader_clsrrr�	_registerNszDefaultProvider._registerN)rrrrr�r�r�r�r�rCr�rrrrr�;sr�c@s8eZdZdZdZdd�ZZdd�Zdd�Zd	d
�Z	dS)r�z.Provider that returns nothing for all requestsNcCsdS�NFrr�rrrr;^r<zEmptyProvider.<lambda>cCsdS�Nr�rr�rrrr�`szEmptyProvider._getcCsgSrrr�rrrr�cszEmptyProvider._listdircCsdSrrr�rrrr�fszEmptyProvider.__init__)
rrrrr�r�r�r�r�r�rrrrr�Ysr�c@s eZdZdZedd��ZeZdS)�ZipManifestsz
    zip manifest builder
    csLt�|��.��fdd����D�}t|�Wd�S1s>0YdS)a
        Build a dictionary similar to the zipimport directory
        caches, except instead of tuples, store ZipInfo objects.

        Use a platform-specific path separator (os.sep) for the path keys
        for compatibility with pypy on Windows.
        c3s&|]}|�dtj���|�fVqdS)r�N)r�r�r��getinfo�r
r���zfilerrr|s��z%ZipManifests.build.<locals>.<genexpr>N)�zipfile�ZipFile�namelistr )rr�r&rr�r�buildrs
	
�zZipManifests.buildN)rrrrrCr��loadrrrrr�ms
r�c@s$eZdZdZe�dd�Zdd�ZdS)�MemoizedZipManifestsz%
    Memoized zipfile manifests.
    �manifest_modzmanifest mtimecCsRtj�|�}t�|�j}||vs.||j|krH|�|�}|�||�||<||jS)zW
        Load a manifest at path or return a suitable manifest already loaded.
        )	r�r��normpathr�st_mtime�mtimer�r��manifest)r�r�r�r�rrrr��s
zMemoizedZipManifests.loadN)rrrrr�
namedtupler�r�rrrrr��sr�c@s�eZdZdZdZe�Zdd�Zdd�Zdd�Z	e
d	d
��Zdd�Ze
d
d��Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �ZdS)!r�z"Resource support for zips and eggsNcCs t�||�|jjtj|_dSr)r�r�r��archiver�r��zip_prer�rrrr��szZipProvider.__init__cCsP|�tj�}||jjkrdS|�|j�r:|t|j�d�Std||jf��dS)Nr��%s is not a subpath of %s)	�rstripr�r�r�r�r�rr��AssertionError�r��fspathrrr�
_zipinfo_name�s�zZipProvider._zipinfo_namecCsP|j|}|�|jtj�r:|t|j�dd��tj�Std||jf��dS)Nr�r)rr�r�r�r�r�r	r)r��zip_pathrrrr�_parts�s
�zZipProvider._partscCs|j�|jj�Sr)�_zip_manifestsr�r�r�r�rrr�zipinfo�szZipProvider.zipinfocCs\|jstd��|�|�}|��}d�|�|��|vrP|D]}|�||�|��q8|�||�S)Nz5resource_filename() only supported for .egg, not .zipr�)r�r��_resource_to_zip�_get_eager_resourcesrEr�_extract_resource�
_eager_to_zip)r�r�r�r�eagersr�rrrr��s�
z!ZipProvider.get_resource_filenamecCs"|j}|jd}t�|�}||fS)N)rrr�)�	file_size�	date_time�time�mktime)�zip_stat�sizer�	timestamprrr�_get_date_and_size�s

zZipProvider._get_date_and_sizec
Cst||��vr@|��|D]}|�|tj�||��}qtj�|�S|�|j|�\}}ts`t	d��z�|�
|j|�|��}|�
||�r�|WStdtj�|�d�\}}	t�||j�|��t�|�t|	||f�|�|	|�zt|	|�Wnftj�yLtj�|��rF|�
||��r|YWStjdk�rFt|�t|	|�|YWS�Yn0Wntj�yn|��Yn0|S)Nz>"os.rename" and "os.unlink" are not supported on this platformz	.$extract)�dirr|)�_indexr
r�r�rEr�rr
�
WRITE_SUPPORT�IOErrorr{r�r�_is_current�_mkstemp�writer�r��closerr�r�error�isfiler�rrs)
r�r�rr�r�rr�	real_path�outf�tmpnamrrrr
�sJ��
�




zZipProvider._extract_resourcec	Cs�|�|j|�\}}tj�|�s$dSt�|�}|j|ksB|j|krFdS|j�	|�}t
|d��}|��}Wd�n1sz0Y||kS)zK
        Return True if the file_path is current for this zip_path
        Fr�N)rr
r�r�r!r�st_sizer�r�r�rr�)	r��	file_pathrrrr�zip_contents�f�
file_contentsrrrrs
&zZipProvider._is_currentcCs>|jdur8g}dD]}|�|�r|�|�|��q||_|jS)N)znative_libs.txtzeager_resources.txt)rr�r&r�)r�rr�rrrrs

z ZipProvider._get_eager_resourcesc	Cs�z|jWSty�i}|jD]V}|�tj�}|r tj�|dd��}||vrf||�|d�q q0|��g||<q0q ||_|YS0dS)Nr�)	�	_dirindex�AttributeErrorr
r	r�r�rEr�r")r��indr��parts�parentrrrr's
zZipProvider._indexcCs |�|�}||jvp||��vSr)rr
r)r�rrrrrr�8s
zZipProvider._hascCs|�|�|��vSr)rrrrrrr�<szZipProvider._isdircCst|���|�|�d��S�Nr)rrr�rrrrrr�?szZipProvider._listdircCs|�|�|j|��Sr)rr�r�r�rrrrBszZipProvider._eager_to_zipcCs|�|�|j|��Sr)rr�r�r�rrrrEszZipProvider._resource_to_zip)rrrrrr�r	r�rrr�r
r�r�rr
rrrr�r�r�rrrrrrr��s(



	7	r�c@s@eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dS)r�a*Metadata handler for standalone PKG-INFO files

    Usage::

        metadata = FileMetadata("/path/to/PKG-INFO")

    This provider rejects all data and metadata requests except for PKG-INFO,
    which is treated as existing, and will be the contents of the file at
    the provided location.
    cCs
||_dSr�r�r�rrrr�XszFileMetadata.__init__cCs|jSrr0r�rrrr�[szFileMetadata._get_metadata_pathcCs|dkotj�|j�S)N�PKG-INFO)r�r�r!r�rrrr�^szFileMetadata.has_metadatacCsX|dkrtd��tj|jddd��}|��}Wd�n1s@0Y|�|�|S)Nr1z(No metadata except PKG-INFO is availabler�r�)�encoding�errors)r�r�rr�r��_warn_on_replacement)r�r�r(�metadatarrrr�as&
zFileMetadata.get_metadatacCs0d}||vr,d}|jfit���}t�|�dS)Nu�z2{self.path} could not be properly decoded in UTF-8)r�r�r�r�)r�r5�replacement_charrqr�rrrr4js
z!FileMetadata._warn_on_replacementcCst|�|��Srr�r�rrrr�qszFileMetadata.get_metadata_linesN)
rrrrr�r�r�r�r4r�rrrrr�Ls	r�c@seZdZdZdd�ZdS)r�asMetadata provider for egg directories

    Usage::

        # Development eggs:

        egg_info = "/path/to/PackageName.egg-info"
        base_dir = os.path.dirname(egg_info)
        metadata = PathMetadata(base_dir, egg_info)
        dist_name = os.path.splitext(os.path.basename(egg_info))[0]
        dist = Distribution(basedir, project_name=dist_name, metadata=metadata)

        # Unpacked egg directories:

        egg_path = "/path/to/PackageName-ver-pyver-etc.egg"
        metadata = PathMetadata(egg_path, os.path.join(egg_path,'EGG-INFO'))
        dist = Distribution.from_filename(egg_path, metadata=metadata)
    cCs||_||_dSr)r�r�)r�r�r�rrrr��szPathMetadata.__init__N�rrrrr�rrrrr�usr�c@seZdZdZdd�ZdS)r�z Metadata provider for .egg filescCsD|jtj|_||_|jr0tj�|j|j�|_n|j|_|�	�dS)z-Create a metadata provider from a zipimporterN)
r�r�r�rr��prefixr�rEr�r�)r��importerrrrr��szEggMetadata.__init__Nr7rrrrr��sr�r ��_distribution_finderscCs|t|<dS)axRegister `distribution_finder` to find distributions in sys.path items

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `distribution_finder` is a callable that, passed a path
    item and the importer instance, yields ``Distribution`` instances found on
    that path item.  See ``pkg_resources.find_on_path`` for an example.Nr:)�
importer_type�distribution_finderrrrr��sr�cCst|�}tt|�}||||�S)z.Yield distributions accessible via `path_item`)rr�r;)�	path_item�onlyr9�finderrrrr]�s
r]c	cs�|j�d�rdSt|�}|�d�r2tj||d�V|r:dS|�d�D]|}t|�r�tj	�
||�}tt�
|�|�}|D]
}|VqrqD|���d�rDtj	�
||�}tt�
|��}||_t�|||�VqDdS)z@
    Find eggs in zip files; possibly multiple nested eggs.
    z.whlNr1�r5r�)�
.dist-info�	.egg-info)r��endswithr�r�rd�
from_filenamerWr�r�r�rE�find_eggs_in_zip�	zipimport�zipimporterrSr��
from_location)	r9r>r?r5�subitem�subpathrr��submetarrrrF�s$

rFcCsdSr/r)r9r>r?rrr�find_nothing�srMcCsdd�}t||dd�S)aL
    Given a list of filenames, return them in descending order
    by version number.

    >>> names = 'bar', 'foo', 'Python-2.7.10.egg', 'Python-2.7.2.egg'
    >>> _by_version_descending(names)
    ['Python-2.7.10.egg', 'Python-2.7.2.egg', 'foo', 'bar']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.egg', 'Setuptools-1.2.3b1.egg']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.post1.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.post1.egg', 'Setuptools-1.2.3b1.egg']
    cSs2tj�|�\}}t�|�d�|g�}dd�|D�S)z6
        Parse each component of the filename
        r�cSsg|]}tj�|��qSr)rrr�)r
�partrrr�
<listcomp>�r<z?_by_version_descending.<locals>._by_version.<locals>.<listcomp>)r�r��splitext�	itertools�chainr	)r��extr-rrr�_by_version�sz+_by_version_descending.<locals>._by_versionTrW)�sorted)rxrTrrr�_by_version_descending�srVc
#s�t���t��r4tj�t�tj��d��d�VdS�fdd�t��D�}��fdd�|D�}t	|�}|D]2}tj��|�}t
�|��}||�D]
}	|	Vq�qjdS)z6Yield distributions accessible on a sys.path directoryr�rANc3s|]}tj��|�VqdSr)r�r�rE)r
�child)r>rrr�s�zfind_on_path.<locals>.<genexpr>c3s|]}t�|��r|VqdSr)�dist_factory)r
r��r?r>rrrs�)�_normalize_cached�_is_unpacked_eggrdrEr�r�r�rE�safe_listdirrVrX)
r9r>r?r��filtered�path_item_entriesr��fullpath�factoryr�rrYr�find_on_path�s(��
��racCsj|��}|�d�}|�d�o0tj�tj�||��}|p8|}|rBtS|sRt|�rRtS|sd|�d�rdt	St
�S)z*Return a dist_factory for the given entry.rCrBz	.egg-link)rSrDr�r�rrE�distributions_from_metadatar�r]�resolve_egg_link�NoDists)r>r�r?rSZis_egg_infoZis_dist_info�is_metarrrrXs$

�������rXc@s eZdZdZdd�Zdd�ZdS)rdzS
    >>> bool(NoDists())
    False

    >>> list(NoDists()('anything'))
    []
    cCsdSr�rr�rrr�__bool__/szNoDists.__bool__cCstd�Sr/)�iter)r�r_rrr�__call__2szNoDists.__call__N)rrrrrfrhrrrrrd'srdc
Csbzt�|�WSttfy"Yn<ty\}z$|jtjtjtjfvrH�WYd}~n
d}~00dS)zI
    Attempt to list contents of path, but suppress some exceptions.
    Nr)	r�r��PermissionError�NotADirectoryError�OSError�errno�ENOTDIR�EACCES�ENOENT)r�r�rrrr\6sr\ccsftj�|�}tj�|�r:tt�|��dkr.dSt||�}nt|�}tj�|�}t	j
|||td�VdS)Nr)�
precedence)r�r�r�rr�r�r�r�r�rdrIr})r��rootr5r�rrrrbFs�rbccsHt|��,}|D]}|��}|r|VqWd�n1s:0YdS)z1
    Yield non-empty lines from file at path
    N)r�strip)r�r(�linerrr�non_empty_linesUs

rtcs.t��}�fdd�|D�}tt|�}t|d�S)za
    Given a path to an .egg-link, resolve distributions
    present in the referenced path.
    c3s$|]}tj�tj���|�VqdSr)r�r�rEr�)r
�refr0rrrfs�z#resolve_egg_link.<locals>.<genexpr>r)rtr3r]r�)r��referenced_paths�resolved_paths�dist_groupsrr0rrc`s
�
rc�
FileFinder��_namespace_handlers)�_namespace_packagescCs|t|<dS)a�Register `namespace_handler` to declare namespace packages

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `namespace_handler` is a callable like this::

        def namespace_handler(importer, path_entry, moduleName, module):
            # return a path_entry to use for child packages

    Namespace handlers are only called if the importer object has already
    agreed that it can handle the relevant path item, and they should only
    return a subpath if the module __path__ does not already contain an
    equivalent subpath.  For an example namespace handler, see
    ``pkg_resources.file_ns_handler``.
    Nrz)r<�namespace_handlerrrrr�wsr�cCst|�}|durdSz|�|�j}WnNtyrt���$t�d�|�|�}Wd�n1sd0YYn0|dur�dStj	�
|�}|dur�t�|�}tj	|<g|_
t|�nt|d�s�td|��tt|�}|||||�}|du�r|j
}|�|�t�|�t|||�|S)zEEnsure that named package includes a subpath of path_item (if needed)N�ignore�__path__�Not a package:)r�	find_specr�r+r��catch_warnings�simplefilter�find_modulerCr�r��types�
ModuleTyper�_set_parent_nsr�r�r�r{r��	importlib�
import_module�_rebuild_mod_path)�packageNamer>r9r�r��handlerrKr�rrr�
_handle_ns�s4

.






r�csjdd�tjD���fdd����fdd�}t||d�}dd�|D�}t|jt�r`||jd	d	�<n||_d	S)
zq
    Rebuild module.__path__ ensuring that all entries are ordered
    corresponding to their sys.path order
    cSsg|]}t|��qSr�rZ�r
�prrrrO�r<z%_rebuild_mod_path.<locals>.<listcomp>cs,z��|�WSty&td�YS0dS)z/
        Workaround for #520 and #513.
        �infN)�indexrH�float)r�)�sys_pathrr�safe_sys_path_index�sz._rebuild_mod_path.<locals>.safe_sys_path_indexcs<|�tj�}��d�d}|d|�}�ttj�|���S)zR
        Return the ordinal of the path based on its position in sys.path
        r>r�N)r	r�r��countrZrE)r��
path_parts�module_partsr-)�package_namer�rr�position_in_sys_path�sz/_rebuild_mod_path.<locals>.position_in_sys_path)r4cSsg|]}t|��qSrr�r�rrrrO�r<N)rCr�rUr�rr)�	orig_pathr�r�r��new_pathr)r�r�r�rr��s		r�c
Cs�t��z�|tvr Wt��dStj}|�d�\}}}|r�t|�|tvrRt|�ztj	|j
}Wn0ty�}ztd|�|�WYd}~n
d}~00t�
|p�dg��|�t�
|g�|D]}t||�q�Wt��n
t��0dS)z9Declare that package 'packageName' is a namespace packageNr>r�)�_imp�acquire_lockr|�release_lockrCr��
rpartitionrZr�r�rr+r�r	r�r�)r�r�r.r�r�r>rrrrZ�s*�"rZcCsNt��z6t�|d�D]}t||�}|rt||�qWt��n
t��0dS)zDEnsure that previously-declared namespace packages include path_itemrN)r�r�r|r�r�r�r�)r>r.�packagerKrrrr��s
r�cCsDtj�||�d�d�}t|�}|jD]}t|�|kr&q@q&|SdS)zBCompute an ns-package subpath for a filesystem or zipfile importerr>r�N)r�r�rEr	rZr)r9r>r�r�rK�
normalizedrrrr�file_ns_handler�s
r�cCsdSrr)r9r>r�r�rrr�null_ns_handler	sr�cCs tj�tj�tj�t|����S)z1Normalize a file/dir name for comparison purposes)r�r��normcase�realpathr��
_cygwin_patch�r�rrrrx	s�rxcCstjdkrtj�|�S|S)a
    Contrary to POSIX 2008, on Cygwin, getcwd (3) contains
    symlink components. Using
    os.path.abspath() works around this limitation. A fix in os.getcwd()
    would probably better, in Cygwin even more so, except
    that this seems to be by design...
    �cygwin)rCrDr�r��abspathr�rrrr�	sr�cCs6z
||WSty0t|�||<}|YS0dSr)r�rx)r�r��resultrrrrZ(	s

rZcCst|�pt|�S)z7
    Determine if given path appears to be an egg.
    )�_is_zip_eggr[r0rrrr�0	sr�cCs$|���d�o"tj�|�o"t�|�S)N�.egg)rSrDr�r�r!r��
is_zipfiler0rrrr�7	s

��r�cCs&|���d�o$tj�tj�|dd��S)z@
    Determine if given path appears to be an unpacked egg.
    r�r�r1)rSrDr�r�r!rEr0rrrr[?	s�r[cCs<|�d�}|��}|r8d�|�}ttj||tj|�dS)Nr>)r	r"rE�setattrrCr�)r�r-r�r.rrrr�I	s


r�ccsXt|t�r6|��D] }|��}|r|�d�s|Vqn|D]}t|�D]
}|VqFq:dS)z9Yield non-empty/non-comment lines of a string or sequence�#N)r�r��
splitlinesrrr�rq)�strs�s�ssrrrrqQ	s

rqz\w+(\.\w+)*$z�
    (?P<name>[^-]+) (
        -(?P<ver>[^-]+) (
            -py(?P<pyver>[^-]+) (
                -(?P<plat>.+)
            )?
        )?
    )?
    c@s�eZdZdZddd�Zdd�Zdd	�Zddd�Zd
d�Zddd�Z	e
�d�Ze
ddd��Ze
dd��Ze
ddd��Ze
ddd��ZdS) rfz3Object representing an advertised importable objectrNcCs<t|�std|��||_||_t|�|_t|�|_||_dS)NzInvalid module name)�MODULErHr��module_name�tuple�attrsr%r�)r�r�r�r�r%r�rrrr�q	s


zEntryPoint.__init__cCsHd|j|jf}|jr*|dd�|j�7}|jrD|dd�|j�7}|S)Nz%s = %s�:r>z [%s]�,)r�r�r�rEr%)r�r�rrrr�z	szEntryPoint.__str__cCsdt|�S)NzEntryPoint.parse(%r)�r�r�rrrr��	szEntryPoint.__repr__TcOs8|r|s|rtjdtdd�|r0|j|i|��|��S)zH
        Require packages for this EntryPoint, then resolve it.
        zJParameters to load are deprecated.  Call .resolve and .require separately.r?r�)r�r�r�rLr)r�rLr:�kwargsrrrr��	s�zEntryPoint.loadc
Cs\t|jdgdd�}zt�t|j|�WStyV}ztt|��|�WYd}~n
d}~00dS)zD
        Resolve the entry point from its module and attrs.
        rr)�fromlist�levelN)	r�r��	functools�reducer�r�r+r�r�)r�r�r�rrrr�	s
zEntryPoint.resolvecCsL|jr|jstd|��|j�|j�}tj||||jd�}tttj|��dS)Nz&Can't require() without a distribution)r%)	r%r�rjrr[rrr3r)r�r)r*rr&rrrrL�	s

zEntryPoint.requirez]\s*(?P<name>.+?)\s*=\s*(?P<module>[\w.]+)\s*(:\s*(?P<attr>[\w.]+))?\s*(?P<extras>\[.*\])?\s*$cCsf|j�|�}|sd}t||��|��}|�|d�}|drJ|d�d�nd}||d|d|||�S)aParse a single entry point from string `src`

        Entry point syntax follows the form::

            name = some.module:some.attr [extra1, extra2]

        The entry name and module name are required, but the ``:attrs`` and
        ``[extras]`` parts are optional
        z9EntryPoint must be in 'name=module:attrs [extras]' formatr%�attrr>rr�r�)�patternrBrH�	groupdict�
_parse_extrasr	)r�srcr�rJr��resr%r�rrrr��	s
zEntryPoint.parsecCs(|sdSt�d|�}|jr"t��|jS)Nr�x)rer��specsrHr%)r�extras_specr�rrrr��	szEntryPoint._parse_extrascCsVt|�std|��i}t|�D]2}|�||�}|j|vrFtd||j��|||j<q|S)zParse an entry point groupzInvalid group namezDuplicate entry point)r�rHrqr�r�)rrG�linesr��thisrs�eprrr�parse_group�	s

zEntryPoint.parse_groupcCstt|t�r|��}nt|�}i}|D]J\}}|durB|s:q$td��|��}||vr\td|��|�|||�||<q$|S)z!Parse a map of entry point groupsNz%Entry points must be listed in groupszDuplicate group name)r�r r&rrrHrrr�)r�datar��mapsrGr�rrr�	parse_map�	s


zEntryPoint.parse_map)rrN)T)NN)N)N)N)rrrrr�r�r�r�rrLr�r�r�rCr�r�r�r�rrrrrfn	s$
	



�	
rfcCs@dd�}t||�}tt|�d�}|�d�\}}}t|���p>dS)z�
    Given an iterable of lines from a Metadata file, return
    the value of the Version field, if present, or None otherwise.
    cSs|���d�S)Nzversion:)rSr�)rsrrr�is_version_line�	sz+_version_from_file.<locals>.is_version_liner�r�N)r�r�rg�	partitionrnrr)r�r��
version_linesrsr�r�rrr�_version_from_file�	s

r�cs�eZdZdZdZddddedefdd�ZedRdd��Z	dd	�Z
ed
d��Zdd
�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zedd��Zedd��Zdd�Zed d!��Zed"d#��Zed$d%��Zd&d'�ZdSd)d*�Zd+d,�Zd-d.�Zd/d0�ZdTd2d3�Z d4d5�Z!d6d7�Z"d8d9�Z#d:d;�Z$�fd<d=�Z%edUd>d?��Z&d@dA�Z'dBdC�Z(dVdDdE�Z)dFdG�Z*dWdHdI�Z+dJdK�Z,dLdM�Z-dNdO�Z.edPdQ��Z/�Z0S)Xrdz5Wrap an actual or potential sys.path entry w/metadatar1NcCsFt|pd�|_|dur t|�|_||_||_||_||_|p>t|_	dS)N�Unknown)
rmr'rn�_versionrOrDrrpr��	_provider)r�rr5r'rrOrDrprrrr�
s
zDistribution.__init__cKs~dgd\}}}}tj�|�\}}	|	��tvr^t|	��}t|�}
|
r^|
�dddd�\}}}}|||f||||d�|����S)Nr�r��ver�pyverrI)r'rrOrD)r�r�rPrS�_distributionImpl�EGG_NAMErG�_reload_version)rrr�r5r#r'rrOrDrSrBrrrrI
s"����zDistribution.from_locationcCs|Srrr�rrrr�!
szDistribution._reload_versioncCs$|j|j|j|j|jpd|jp dfSr�)�parsed_versionrpr4rrOrDr�rrrrV$
s�zDistribution.hashcmpcCs
t|j�Sr)�hashrVr�rrr�__hash__/
szDistribution.__hash__cCs|j|jkSr�rV�r�r_rrr�__lt__2
szDistribution.__lt__cCs|j|jkSrr�r�rrr�__le__5
szDistribution.__le__cCs|j|jkSrr�r�rrr�__gt__8
szDistribution.__gt__cCs|j|jkSrr�r�rrr�__ge__;
szDistribution.__ge__cCst||j�sdS|j|jkSr�)r�r�rVr�rrr�__eq__>
szDistribution.__eq__cCs
||kSrrr�rrr�__ne__D
szDistribution.__ne__cCs4z|jWSty.|j��|_}|YS0dSr)�_keyr+r'rSr^rrrr4K
s
zDistribution.keycCst|d�st|j�|_|jS)N�_parsed_version)r�rrr�r�rrrr�S
s
zDistribution.parsed_versioncCs\tjj}t|j|�}|sdS|js&dSt�d����dd�}t	�
|jfit|���t
�dS)Na>
            '{project_name} ({version})' is being parsed as a legacy,
            non PEP 440,
            version. You may find odd behavior and sort order.
            In particular it will be sorted as less than 0.0. It
            is recommended to migrate to PEP 440 compatible
            versions.
            r�r�)rrrr�r�rkrlrrr�r�r�r��varsr)r��LV�	is_legacyrqrrr�_warn_legacy_versionZ
s�	z!Distribution._warn_legacy_versionc
Cslz|jWStyf}zF|��}|durN|�|j�}d�|j|�}t||�|�|WYd}~Sd}~00dS)Nz4Missing 'Version:' header and/or {} file at path: {})r�r+�_get_version�_get_metadata_path_for_display�PKG_INFOr�rH)r�r�rr�r�rrrrt
s��zDistribution.versioncCs2z|jWSty*|�|���|_Yn0|jS)z~
        A map of extra to its list of (direct) requirements
        for this distribution, including the null extra.
        )�_Distribution__dep_mapr+�_filter_extras�_build_dep_mapr�rrr�_dep_map�
s
zDistribution._dep_mapcCsrttd|��D]^}|}|�|�}|�d�\}}}|oDt|�pDt|�}|rNg}t|�pXd}|�|g��|�q|S)z�
        Given a mapping of extras to dependencies, strip off
        environment markers and filter out any dependencies
        not matching the markers.
        Nr�)	rr�r"r�rurvrsr	r&)�dmrD�	new_extrarr�rF�fails_markerrrrr��
s
�zDistribution._filter_extrascCs@i}dD]2}t|�|��D]\}}|�|g��t|��qq|S)N)zrequires.txtzdepends.txt)rr�
_get_metadatar	r&rl)r�r�r�rDrrrrr��
s
zDistribution._build_dep_maprcCsx|j}g}|�|�dd��|D]R}z|�|t|��Wq typ}ztd||f�|�WYd}~q d}~00q |S)z@List of Requirements needed for this distro if `extras` are usedNrz%s has no such extra feature %r)r�r&r�rsr�rj)r�r%r��depsrSr�rrrr�
s
��zDistribution.requirescCs*z|j�|�}Wnty$YdS0|S)zK
        Return the path to the given metadata file, if available.
        z[could not detect])r�r�rur�rrrr��
s
z+Distribution._get_metadata_path_for_displayccs$|�|�r |�|�D]
}|VqdSr)r�r�)r�r�rsrrrr��
s
zDistribution._get_metadatacCs|�|j�}t|�}|Sr)r�r�r�)r�r�rrrrr��
szDistribution._get_versionFcCsV|durtj}|j||d�|tjurRt|j�|�d�D]}|tjvr:t|�q:dS)z>Ensure distribution is importable on `path` (default=sys.path)Nr�namespace_packages.txt)rCr�rr�rr�r�rZ)r�r�r��pkgrrr�activate�
s


zDistribution.activatecCs8dt|j�t|j�|jptf}|jr4|d|j7}|S)z@Return what this distribution's standard .egg filename should bez
%s-%s-py%sr�)rtr'rrOrerD)r�r�rrrr��
s�zDistribution.egg_namecCs |jrd||jfSt|�SdS)Nz%s (%s))rr�r�rrrr��
szDistribution.__repr__cCs>zt|dd�}Wnty&d}Yn0|p.d}d|j|fS)Nrz[unknown version]z%s %s)r�rHr')r�rrrrr��
s
zDistribution.__str__cCs|�d�rt|��t|j|�S)zADelegate all unrecognized public attributes to .metadata providerr�)r�r+r�r�)r�r�rrr�__getattr__�
s
zDistribution.__getattr__cs.tttt|����tdd�|j��D��B�S)Ncss|]}|�d�s|VqdS�r�N)r�)r
r�rrrrs
�z'Distribution.__dir__.<locals>.<genexpr>)rr!�superrd�__dir__r�r�rbrrrs���zDistribution.__dir__cKs"|jt|�tj�|�|fi|��Sr)rIrZr�r�r�)rr�r5r#rrrrE	s
��zDistribution.from_filenamecCs<t|jtjj�r"d|j|jf}nd|j|jf}t�|�S)z?Return a ``Requirement`` that matches this distribution exactlyz%s==%sz%s===%s)r�r�rrrr'rer�)r��specrrrr4szDistribution.as_requirementcCs.|�||�}|dur&td||ff��|��S)z=Return the `name` entry point of `group` or raise ImportErrorNzEntry point %r not found)rRr�r�)r�rGr�r�rrrrPszDistribution.load_entry_pointcCsNz
|j}Wn*ty4t�|�d�|�}|_Yn0|durJ|�|i�S|S)r�zentry_points.txtN)�_ep_mapr+rfr�r�r�)r�rG�ep_maprrrrQ s

�zDistribution.get_entry_mapcCs|�|��|�Sr�)rQr�rrrrrR,szDistribution.get_entry_infoc
Cs2|p|j}|sdSt|�}tj�|�}dd�|D�}t|�D]|\}}||kr^|rVq�q�dSq<||kr<|jtkr<|s�|||d�vr�dS|tjur�|�	�|�
||�|�
||�q�q<|tjur�|�	�|r�|�
d|�n
|�|�dSz|�||d�}	Wnt
�yY�q.Yq�0||	=||	=|	}q�dS)a�Ensure self.location is on path

        If replace=False (default):
            - If location is already in path anywhere, do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent.
              - Else: add to the end of path.
        If replace=True:
            - If location is already on path anywhere (not eggs)
              or higher priority than its parent (eggs)
              do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent,
                removing any lower-priority entries.
              - Else: add it to the front of path.
        NcSsg|]}|rt|�p|�qSrr�r�rrrrOKr<z*Distribution.insert_on.<locals>.<listcomp>rr�)rrZr�r�r��	enumeraterpryrC�check_version_conflictrr�r�rH)
r�r��locr��nloc�bdir�npathr�r�nprrrr1s@



zDistribution.insert_oncCs�|jdkrdSt�|�d��}t|j�}|�d�D]p}|tjvs2||vs2|tvrRq2|dvr\q2t	tj|dd�}|r�t|��
|�s2|�
|j�r�q2td|||jf�q2dS)N�
setuptoolsr�z
top_level.txt)�
pkg_resourcesr�siter�zIModule %s was already imported from %s, but %s is being added to sys.path)r4r r!r�rxrrCr�r|r�r��
issue_warning)r��nspr�modname�fnrrrrus*

�
�
��z#Distribution.check_version_conflictcCs4z
|jWn$ty.tdt|��YdS0dS)NzUnbuilt egg for FT)rrHrr�r�rrrrY�s
zDistribution.has_versioncKsDd}|��D]}|�|t||d��q|�d|j�|jfi|��S)z@Copy this distribution, substituting in any changed keyword argsz<project_name version py_version platform location precedenceNr5)r	r	r�r�r�)r�r#rxr�rrr�clone�s
zDistribution.clonecCsdd�|jD�S)NcSsg|]}|r|�qSrr)r
�deprrrrO�r<z'Distribution.extras.<locals>.<listcomp>)r�r�rrrr%�szDistribution.extras)N)r)NF)N)N)NF)1rrrrr�reryr�rCrIr�r�rVr�r�r�r�r�r�r�r4r�r�rr�r�r�r�rr�r�r�r�r�r�r�r�rrEr4rPrQrRrrrYrr%�
__classcell__rrrbrrd�	sh�










		

Drdc@seZdZdd�ZdS)�EggInfoDistributioncCs|��}|r||_|S)a�
        Packages installed by distutils (e.g. numpy or scipy),
        which uses an old safe_version, and so
        their version numbers can get mangled when
        converted to filenames (e.g., 1.11.0.dev0+2329eae to
        1.11.0.dev0_2329eae). These distributions will not be
        parsed properly
        downstream by Distribution and safe_version, so
        take an extra step and try to get the version number from
        the metadata file itself instead of the filename.
        )r�r�)r��
md_versionrrrr��sz#EggInfoDistribution._reload_versionN)rrrr�rrrrr�src@s>eZdZdZdZe�d�Zedd��Z	edd��Z
dd	�Zd
S)�DistInfoDistributionzV
    Wrap an actual or potential sys.path entry
    w/metadata, .dist-info style.
    �METADATAz([\(,])\s*(\d.*?)\s*([,\)])cCsDz|jWSty>|�|j�}tj���|�|_|jYS0dS)zParse and cache metadataN)�	_pkg_infor+r�r��email�parser�Parser�parsestr)r�r5rrr�_parsed_pkg_info�sz%DistInfoDistribution._parsed_pkg_infocCs0z|jWSty*|��|_|jYS0dSr)�_DistInfoDistribution__dep_mapr+�_compute_dependenciesr�rrrr��s

zDistInfoDistribution._dep_mapcs�dgi}|_g�|j�d�p gD]}��t|��q"�fdd�}t|d��}|d�|�|j�d�pjgD](}t|���}tt||��|�||<ql|S)z+Recompute this distribution's dependencies.Nz
Requires-Distc3s*�D] }|jr|j�d|i�r|VqdS)NrDrE)rDr��rrr�reqs_for_extra�szBDistInfoDistribution._compute_dependencies.<locals>.reqs_for_extrazProvides-Extra)	rr�get_allr&rl�	frozensetrsrrr)r�r�r�r"�commonrD�s_extrarr!rr �sz*DistInfoDistribution._compute_dependenciesN)rrrrr�r�r��EQEQr�rr�r rrrrr�s

	
r)r�rCrBcOsXd}t�}zt�|�j|ur&|d7}qWnty:Yn0tj|d|di|��dS)Nr�r�)rrCr�r�rHr�r�)r:r#r�r)rrrr�src	cs�tt|��}|D]j}d|vr.|d|�d��}|�d�rp|dd���}z|t|�7}WntynYdS0t|�VqdS)z�Yield ``Requirement`` objects for each specification in `strs`

    `strs` must be a string, or a (possibly-nested) iterable thereof.
    z #N�\���)rgrqr�rDrrr��
StopIterationre)r�r�rsrrrrl�s

rlc@seZdZdZdS)�RequirementParseErrorz,Compatibility wrapper for InvalidRequirementNrrrrrr+sr+csPeZdZ�fdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Ze	d
d��Z
�ZS)recs�tt|��|�|j|_t|j�}||��|_|_dd�|j	D�|_
ttt
|j��|_|j|j|j	t|j�|jr|t|j�ndf|_t|j�|_dS)z>DO NOT CALL THIS UNDOCUMENTED METHOD; use Requirement.parse()!cSsg|]}|j|jf�qSr)rZr)r
rrrrrOsz(Requirement.__init__.<locals>.<listcomp>N)r�rer�r��unsafe_namermrSr'r4�	specifierr�r�r3rsr%�urlr$rFr��hashCmpr��_Requirement__hash)r��requirement_stringr'rbrrr�s
��zRequirement.__init__cCst|t�o|j|jkSr)r�rer/r�rrrr�&s

�zRequirement.__eq__cCs
||kSrrr�rrrr�,szRequirement.__ne__cCs0t|t�r |j|jkrdS|j}|jj|dd�S)NFT)�prereleases)r�rdr4rr-�contains)r�rrrrr/s

zRequirement.__contains__cCs|jSr)r0r�rrrr�;szRequirement.__hash__cCsdt|�S)NzRequirement.parse(%r)r�r�rrrr�>szRequirement.__repr__cCst|�\}|Sr)rl)r�r�rrrr�As
zRequirement.parse)rrrr�r�r�rr�r�r�r�rrrrbrresrecCst|vr|tfS|S)zJ
    Ensure object appears in the mro even
    for old-style classes.
    )�object)�classesrrr�_always_objectGs
r6cCs<tt�t|dt|����}|D]}||vr||SqdS)z2Return an adapter factory for `ob` from `registry`r�N)r6�inspect�getmror�r�)�registryr5r��trrrr�Qsr�cCstj�|�}tj|dd�dS)z1Ensure that the parent directory of `path` existsT)�exist_okN)r�r�r��makedirs)r�r�rrrrwYsrwcCsVtstd��t|�\}}|rR|rRt|�sRt|�zt|d�WntyPYn0dS)z/Sandbox-bypassing version of ensure_directory()z*"os.mkdir" not supported on this platform.i�N)rrr	rrtr�FileExistsError)r�r�r�rrrrt_srtccsvd}g}t|�D]V}|�d�r\|�d�rP|s0|r:||fV|dd���}g}qftd|��q|�|�q||fVdS)asSplit a string or iterable thereof into (section, content) pairs

    Each ``section`` is a stripped version of the section header ("[section]")
    and each ``content`` is a list of stripped lines excluding blank lines and
    comment-only lines.  If there are any such lines before the first section
    header, they're returned in a first ``section`` of ``None``.
    N�[�]r�r�zInvalid section heading)rqr�rDrrrHr�)r��section�contentrsrrrrrls


rrcOs2tj}ztt_tj|i|��W|t_S|t_0dSr)r�r�os_open�tempfile�mkstemp)r:r#�old_openrrrr�s�rr~)�categoryr�cOs||i|��|Srr)r(r:r�rrr�_call_aside�srGcs.t���|d<|��fdd�t��D��dS)z=Set up global resource manager (deliberately not state-saved)�_managerc3s&|]}|�d�s|t�|�fVqdSr�)r�r�r��r�rrr�s
�z_initialize.<locals>.<genexpr>N)rcrr)r)rrIr�_initialize�s
�rJcCs|t��}td|d�|j}|j}|j}|j}|}tdd�|D��|dd�dd�g|_t	t
|jtj
��t��t��d	S)
aE
    Prepare the master working set and make the ``require()``
    API available.

    This function has explicit effects on the global state
    of pkg_resources. It is intended to be invoked once at
    the initialization of this module.

    Invocation by other packages is unsupported and done
    at their own risk.
    r4)r[css|]}|jdd�VqdS)FrN�r�)r
r�rrrr�s�z1_initialize_master_working_set.<locals>.<genexpr>cSs|jdd�S)NTrrKr�rrrr;�r<z0_initialize_master_working_set.<locals>.<lambda>F)r@N)rbrr$rLrSrArMr�r�rr3r�rCr�rrr�)r[rLrSr\rMr�rrr�_initialize_master_working_set�s"
��rLc@seZdZdZdS)r�z�
    Base class for warning about deprecations in ``pkg_resources``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    Nrrrrrr��sr�)N)N)F)F)F)F)N)�rrCr�r�rr�r�r�rGr�rr��pkgutilrZrDrr��email.parserrrlrCrkrQr7r�r�r�rr�r��impr=�	NameErrorrkrrrrrrrB�os.pathrr	�importlib.machinery�	machineryr�rZpkg_resources.externr
rr��version_info�RuntimeErrorrLr[r\�resources_streamr_�resource_dirrUr^rYrTrSrWrVrXr;r{r|�RuntimeWarningrrrr$r+r-r1r6r7r8�
_sget_none�
_sset_nonerK�__all__rurgrhr�rirjr�r�reryrzr{r|r}r�rNrFr�r@r�rAr�rorprMr�rOrPrQrRr~rrbr rrar�rkrcr`rmrnrsrtrurvr�r4r�r�r�r�r�r�r�r�r�rHr�r�r�r�r]rFrMrVrarXrdr\rbrtrc�ImpImporterr�ryr�r�r�rZr�r�r�rxr�rZr�r�r[r�rqrBr��VERBOSE�
IGNORECASEr�rfr�rdrrr�rrlr(�InvalidRequirementr+rer6r�rwrtrrr�filterwarningsrGrrJrL�Warningr�rrrr�<module>s�



2 




.

6	
@
.)





##""



	
�
$3�
4


'site-packages/pkg_resources/_vendor/__pycache__/appdirs.cpython-39.pyc000064400000050114150564745010022106 0ustar00a

(�Re}`�@s�dZdZd�eee��ZddlZddlZejddkZ	e	r>eZ
ej�d�r�ddlZe�
�ddZe�d�rrdZq�e�d	�r�d
Zq�dZnejZd4d
d�Zd5dd�Zd6dd�Zd7dd�Zd8dd�Zd9dd�Zd:dd�ZGdd�de�Zdd�Zd d!�Zd"d#�Zd$d%�Zedk�r�zddlZeZWnhe �y�zdd&l!m"Z"eZWn>e �y�zddl#Z$eZWne �yzeZYn0Yn0Yn0e%d'k�r~d(Z&d)Z'd*Z(e)d+e�e)d,�ee&e'd-d.�Z*e(D]Z+e)d/e+e,e*e+�f��q�e)d0�ee&e'�Z*e(D]Z+e)d/e+e,e*e+�f��q�e)d1�ee&�Z*e(D]Z+e)d/e+e,e*e+�f��q*e)d2�ee&dd3�Z*e(D]Z+e)d/e+e,e*e+�f��q`dS);zyUtilities for determining application-specific dirs.

See <http://github.com/ActiveState/appdirs> for details and usage.
)����.�Nr�java�Windows�win32ZMac�darwinZlinux2FcCs�tdkr^|dur|}|rdpd}tj�t|��}|r�|durNtj�|||�}q�tj�||�}nNtdkr�tj�d�}|r�tj�||�}n&t�dtj�d	��}|r�tj�||�}|r�|r�tj�||�}|S)
aJReturn full path to the user-specific data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user data directories are:
        Mac OS X:               ~/Library/Application Support/<AppName>
        Unix:                   ~/.local/share/<AppName>    # or in $XDG_DATA_HOME, if defined
        Win XP (not roaming):   C:\Documents and Settings\<username>\Application Data\<AppAuthor>\<AppName>
        Win XP (roaming):       C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>
        Win 7  (not roaming):   C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>
        Win 7  (roaming):       C:\Users\<username>\AppData\Roaming\<AppAuthor>\<AppName>

    For Unix, we follow the XDG spec and support $XDG_DATA_HOME.
    That means, by default "~/.local/share/<AppName>".
    rN�
CSIDL_APPDATA�CSIDL_LOCAL_APPDATAFr	z~/Library/Application Support/�
XDG_DATA_HOMEz~/.local/share��system�os�path�normpath�_get_win_folder�join�
expanduser�getenv)�appname�	appauthor�version�roaming�constr�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/appdirs.py�
user_data_dir-s& rcstdkrR|dur�}tj�td��}�r�|durBtj�||��}q�tj�|��}n�tdkrztj�d�}�r�tj�|��}ntt�dtj�dd	g��}d
d�|�	tj�D�}�r�|r�tj��|���fdd�|D�}|r�tj�|�}n|d
}|S��r|�rtj�||�}|S)aiReturn full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of data dirs should be
            returned. By default, the first item from XDG_DATA_DIRS is
            returned, or '/usr/local/share/<AppName>',
            if XDG_DATA_DIRS is not set

    Typical site data directories are:
        Mac OS X:   /Library/Application Support/<AppName>
        Unix:       /usr/local/share/<AppName> or /usr/share/<AppName>
        Win XP:     C:\Documents and Settings\All Users\Application Data\<AppAuthor>\<AppName>
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)
        Win 7:      C:\ProgramData\<AppAuthor>\<AppName>   # Hidden, but writeable on Win 7.

    For Unix, this is using the $XDG_DATA_DIRS[0] default.

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    rN�CSIDL_COMMON_APPDATAFr	z/Library/Application Support�
XDG_DATA_DIRSz/usr/local/sharez
/usr/sharecSs g|]}tj�|�tj���qSr�rrr�rstrip�sep��.0�xrrr�
<listcomp>��z!site_data_dir.<locals>.<listcomp>csg|]}tj�|�g��qSr�rr"rr#�rrrr&�r'r)
rrrrrrrr�pathsep�split�rrr�	multipathrZpathlistrr)r�
site_data_dirds6�r.cCsXtdvrt||d|�}n&t�dtj�d��}|r>tj�||�}|rT|rTtj�||�}|S)a�Return full path to the user-specific config dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user config directories are:
        Mac OS X:               same as user_data_dir
        Unix:                   ~/.config/<AppName>     # or in $XDG_CONFIG_HOME, if defined
        Win *:                  same as user_data_dir

    For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME.
    That means, by default "~/.config/<AppName>".
    �rr	N�XDG_CONFIG_HOMEz	~/.config�rrrrrrr�rrrrrrrr�user_config_dir�sr3cs�tdvr*t�|�}�r�|r�tj�||�}ndt�dd�}dd�|�tj�D�}�rt|rbtj��|���fdd�|D�}|r�tj�|�}n|d}|S)aReturn full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of config dirs should be
            returned. By default, the first item from XDG_CONFIG_DIRS is
            returned, or '/etc/xdg/<AppName>', if XDG_CONFIG_DIRS is not set

    Typical site config directories are:
        Mac OS X:   same as site_data_dir
        Unix:       /etc/xdg/<AppName> or $XDG_CONFIG_DIRS[i]/<AppName> for each value in
                    $XDG_CONFIG_DIRS
        Win *:      same as site_data_dir
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)

    For Unix, this is using the $XDG_CONFIG_DIRS[0] default, if multipath=False

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    r/�XDG_CONFIG_DIRSz/etc/xdgcSs g|]}tj�|�tj���qSrr r#rrrr&�r'z#site_config_dir.<locals>.<listcomp>csg|]}tj�|�g��qSrr(r#r)rrr&�r'r)rr.rrrrr+r*r,rr)r�site_config_dir�s
r5TcCs�tdkrd|dur|}tj�td��}|r�|durBtj�|||�}ntj�||�}|r�tj�|d�}nNtdkr�tj�d�}|r�tj�||�}n&t�dtj�d	��}|r�tj�||�}|r�|r�tj�||�}|S)
aReturn full path to the user-specific cache dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Cache" to the base app data dir for Windows. See
            discussion below.

    Typical user cache directories are:
        Mac OS X:   ~/Library/Caches/<AppName>
        Unix:       ~/.cache/<AppName> (XDG default)
        Win XP:     C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>\Cache
        Vista:      C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>\Cache

    On Windows the only suggestion in the MSDN docs is that local settings go in
    the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming
    app data dir (the default returned by `user_data_dir` above). Apps typically
    put cache data somewhere *under* the given dir here. Some examples:
        ...\Mozilla\Firefox\Profiles\<ProfileName>\Cache
        ...\Acme\SuperApp\Cache\1.0
    OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value.
    This can be disabled with the `opinion=False` option.
    rNrF�Cacher	z~/Library/Caches�XDG_CACHE_HOMEz~/.cacher
�rrr�opinionrrrr�user_cache_dirs(!r:cCsXtdvrt||d|�}n&t�dtj�d��}|r>tj�||�}|rT|rTtj�||�}|S)aReturn full path to the user-specific state dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user state directories are:
        Mac OS X:  same as user_data_dir
        Unix:      ~/.local/state/<AppName>   # or in $XDG_STATE_HOME, if defined
        Win *:     same as user_data_dir

    For Unix, we follow this Debian proposal <https://wiki.debian.org/XDGBaseDirectorySpecification#state>
    to extend the XDG spec and support $XDG_STATE_HOME.

    That means, by default "~/.local/state/<AppName>".
    r/N�XDG_STATE_HOMEz~/.local/stater1r2rrr�user_state_dir:sr<cCs�tdkr tj�tj�d�|�}nNtdkrLt|||�}d}|rntj�|d�}n"t|||�}d}|rntj�|d�}|r�|r�tj�||�}|S)a�Return full path to the user-specific log dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Logs" to the base app data dir for Windows, and "log" to the
            base cache dir for Unix. See discussion below.

    Typical user log directories are:
        Mac OS X:   ~/Library/Logs/<AppName>
        Unix:       ~/.cache/<AppName>/log  # or under $XDG_CACHE_HOME if defined
        Win XP:     C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>\Logs
        Vista:      C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>\Logs

    On Windows the only suggestion in the MSDN docs is that local settings
    go in the `CSIDL_LOCAL_APPDATA` directory. (Note: I'm interested in
    examples of what some windows apps use for a logs dir.)

    OPINION: This function appends "Logs" to the `CSIDL_LOCAL_APPDATA`
    value for Windows and appends "log" to the user cache dir for Unix.
    This can be disabled with the `opinion=False` option.
    r	z~/Library/LogsrFZLogs�log)rrrrrrr:r8rrr�user_log_dirds" 
�r>c@sneZdZdZddd�Zedd��Zedd	��Zed
d��Zedd
��Z	edd��Z
edd��Zedd��ZdS)�AppDirsz1Convenience wrapper for getting application dirs.NFcCs"||_||_||_||_||_dS)N)rrrrr-)�selfrrrrr-rrr�__init__�s
zAppDirs.__init__cCst|j|j|j|jd�S�N)rr)rrrrr�r@rrrr�s
�zAppDirs.user_data_dircCst|j|j|j|jd�S�N)rr-)r.rrrr-rCrrrr.�s
�zAppDirs.site_data_dircCst|j|j|j|jd�SrB)r3rrrrrCrrrr3�s
�zAppDirs.user_config_dircCst|j|j|j|jd�SrD)r5rrrr-rCrrrr5�s
�zAppDirs.site_config_dircCst|j|j|jd�S�N�r)r:rrrrCrrrr:�s
�zAppDirs.user_cache_dircCst|j|j|jd�SrE)r<rrrrCrrrr<�s
�zAppDirs.user_state_dircCst|j|j|jd�SrE)r>rrrrCrrrr>�s
�zAppDirs.user_log_dir)NNNFF)
�__name__�
__module__�__qualname__�__doc__rA�propertyrr.r3r5r:r<r>rrrrr?�s$�






r?cCsHtrddl}nddl}dddd�|}|�|jd�}|�||�\}}|S)z�This is a fallback technique at best. I'm not sure if using the
    registry for this guarantees us the correct answer for all CSIDL_*
    names.
    rNZAppDatazCommon AppDataz
Local AppData�r
rrz@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)�PY3�winreg�_winreg�OpenKey�HKEY_CURRENT_USER�QueryValueEx)�
csidl_namerOZshell_folder_name�key�dir�typerrr�_get_win_folder_from_registry�s
���rWcCs�ddlm}m}|�dt||�dd�}z\t|�}d}|D]}t|�dkr8d}qRq8|r�zddl}|�|�}Wnt	y~Yn0Wnt
y�Yn0|S)Nr)�shellcon�shellF�T)�win32com.shellrXrY�SHGetFolderPath�getattr�unicode�ord�win32api�GetShortPathName�ImportError�UnicodeError)rSrXrYrU�
has_high_char�cr`rrr�_get_win_folder_with_pywin32�s$
rfcCs�ddl}dddd�|}|�d�}|jj�d|dd|�d}|D]}t|�dkr@d	}qZq@|r�|�d�}|jj�|j|d�r�|}|jS)
Nr��#�rLiFrZT)	�ctypes�create_unicode_buffer�windllZshell32ZSHGetFolderPathWr_Zkernel32ZGetShortPathNameW�value)rSrjZcsidl_const�bufrdreZbuf2rrr�_get_win_folder_with_ctypes�s&��

rocCs�ddl}ddlm}ddlm}|jjd}|�d|�}|jj	}|�
dt|j|�d|jj
|�|j�|����d�}d}|D]}	t|	�dkr|d	}q�q||r�|�d|�}|jj	}
|
�|||�r�|j�|����d�}|S)
Nr)�jna)r�re�FrZT)�arrayZcom.sunrpZcom.sun.jna.platformrZWinDefZMAX_PATHZzerosZShell32ZINSTANCEr\r]ZShlObjZSHGFP_TYPE_CURRENTZNativeZtoString�tostringr!r_ZKernel32ra)rSrsrprZbuf_sizernrYrUrdreZkernelrrr�_get_win_folder_with_jnas&ru)rl�__main__ZMyAppZ	MyCompany)rr3r:r<r>r.r5z-- app dirs %s --z%-- app dirs (with optional 'version')z1.0rFz%s: %sz)
-- app dirs (without optional 'version')z+
-- app dirs (without optional 'appauthor')z(
-- app dirs (with disabled 'appauthor'))r)NNNF)NNNF)NNNF)NNNF)NNNT)NNNF)NNNT)-rJ�__version_info__r�map�str�__version__�sysr�version_inforMr^�platform�
startswith�java_ver�os_namerrr.r3r5r:r<r>�objectr?rWrfrorur[Zwin32comrrbrjrlZcom.sun.jna�comrGrr�props�print�dirs�propr]rrrr�<module>s~



7
B
(
3
9
*
30


site-packages/pkg_resources/_vendor/__pycache__/__init__.cpython-39.pyc000064400000000357150564745010022207 0ustar00a

(�Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/__init__.py�<module>�site-packages/pkg_resources/_vendor/__pycache__/pyparsing.cpython-39.pyc000064400000611261150564745010022466 0ustar00a

(�Rex��@s�
dZdZdZdZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZzddlmZWney�ddlmZYn0zdd	lmZdd
lmZWn*ey�dd	l
mZdd
l
mZYn0zddl
mZWn>e�y>zddlmZWne�y8dZYn0Yn0gd�Zee	j�dd
�Zedd
kZ e �r�e	j!Z"e#Z$e%Z&e#Z'e(e)e*e+e,ee-e.e/e0e1gZ2n^e	j3Z"e4Z5dd�Z'gZ2ddl6Z6d�7�D]6Z8ze2�9e:e6e8��Wne;�y�Y�q�Yn0�q�e<dd�e5d�D��Z=dd�Z>Gdd�de?�Z@ejAejBZCdZDeDdZEeCeDZFe%d�ZGd�Hdd�ejID��ZJGdd�deK�ZLGdd �d eL�ZMGd!d"�d"eL�ZNGd#d$�d$eN�ZOGd%d&�d&eK�ZPGd'd(�d(e?�ZQGd)d*�d*e?�ZRe�SeR�d+d,�ZTd-d.�ZUd/d0�ZVd1d2�ZWd3d4�ZXd5d6�ZYd7d8�ZZ�dd:d;�Z[Gd<d=�d=e?�Z\Gd>d?�d?e\�Z]Gd@dA�dAe]�Z^GdBdC�dCe]�Z_GdDdE�dEe]�Z`e`Zae`e\_bGdFdG�dGe]�ZcGdHdI�dIe`�ZdGdJdK�dKec�ZeGdLdM�dMe]�ZfGdNdO�dOe]�ZgGdPdQ�dQe]�ZhGdRdS�dSe]�ZiGdTdU�dUe]�ZjGdVdW�dWe]�ZkGdXdY�dYe]�ZlGdZd[�d[el�ZmGd\d]�d]el�ZnGd^d_�d_el�ZoGd`da�dael�ZpGdbdc�dcel�ZqGddde�deel�ZrGdfdg�dgel�ZsGdhdi�die\�ZtGdjdk�dket�ZuGdldm�dmet�ZvGdndo�doet�ZwGdpdq�dqet�ZxGdrds�dse\�ZyGdtdu�duey�ZzGdvdw�dwey�Z{Gdxdy�dyey�Z|Gdzd{�d{e|�Z}Gd|d}�d}e|�Z~Gd~d�de?�Ze�Z�Gd�d��d�ey�Z�Gd�d��d�ey�Z�Gd�d��d�ey�Z�Gd�d��d�e��Z�Gd�d��d�ey�Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e?�Z�d�d��Z��dd�d��Z��dd�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z��dd�d��Z�d�d��Z��dd�d��Z�d�d��Z�d�d��Z�e^���d��Z�en���d��Z�eo���d��Z�ep���d��Z�eq���d��Z�egeGd�d9d����d�d���Z�ehd����d�d���Z�ehd����d�d���Z�e�e�Be�Bejd�dd��BZ�e�e�e�d��e��Z�e`d��e�d����d��e�e}e�e�B����d��d�Z�d�dĄZ�d�dƄZ�d�dȄZ�d�dʄZ�d�d̄Z�e�d�d���Z�e�d�d���Z�d�dЄZ�d�d҄Z�d�dԄZ�d�dքZ�e?�e�_��dd�d؄Z�e@�Z�e?�e�_�e?�e�_�e�dكe�dڃfd�d܄Z�e�Z�e�ehd݃d����dߡZ�e�ehd�d����d�Z�e�ehd݃d�ehd�d�B���d�Z�e�ead�e������d�Z�d�d�de���fd�d�Z��dd�d�Z�e�d�Z�e�d�Z�e�egeCeFd����d��\Z�Z�e�e�d�7�d��Z�ehd�d�Heàġ�d����d�Z�d�d��Z�e�ehd��d����d��Z�ehd����d��Z�ehd���ɡ��d��Z�ehd����d��Z�e�ehd��d�e�B���d��Z�e�Z�eh�d����d�Z�e�e}egeJd��d�e�eg�d�e`d��eo�����ϡ���d�Z�e�e�e���e�Bd�d�����d�Z�G�d�d��d�Z�eӐd	k�
r�ed�d
�Z�ed�d�Z�egeCeF�d�Z�e�e֐d
d��d���e��Z�e�e�e׃����d�Zؐde�BZ�e�e֐d
d��d���e��Z�e�e�eڃ����d�Z�eԐd�eِd�e�eېd�Z�eܠݐd�e�jޠݐd�e�jߠݐd�e�j�ݐd�ddl�Z�e�j᠝e�e�j��e�j�ݐd�dS(a�	
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and executing simple grammars,
vs. the traditional lex/yacc approach, or the use of regular expressions.  With pyparsing, you
don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
provides a library of classes that you use to construct the grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form 
C{"<salutation>, <addressee>!"}), built up using L{Word}, L{Literal}, and L{And} elements 
(L{'+'<ParserElement.__add__>} operator gives L{And} expressions, strings are auto-converted to
L{Literal} expressions)::

    from pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the self-explanatory
class names, and the use of '+', '|' and '^' operators.

The L{ParseResults} object returned from L{ParserElement.parseString<ParserElement.parseString>} can be accessed as a nested list, a dictionary, or an
object with named attributes.

The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
 - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello  ,  World  !", etc.)
 - quoted strings
 - embedded comments


Getting Started -
-----------------
Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:
 - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes
 - construct character word-group expressions using the L{Word} class
 - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes
 - use L{'+'<And>}, L{'|'<MatchFirst>}, L{'^'<Or>}, and L{'&'<Each>} operators to combine simple expressions into more complex ones
 - associate names with your parsed results using L{ParserElement.setResultsName}
 - find some helpful expression short-cuts like L{delimitedList} and L{oneOf}
 - find more useful common expressions in the L{pyparsing_common} namespace class
z2.2.1z18 Sep 2018 00:49 UTCz*Paul McGuire <ptmcg@users.sourceforge.net>�N)�ref)�datetime)�RLock)�Iterable)�MutableMapping)�OrderedDict)i�And�CaselessKeyword�CaselessLiteral�
CharsNotIn�Combine�Dict�Each�Empty�
FollowedBy�Forward�
GoToColumn�Group�Keyword�LineEnd�	LineStart�Literal�
MatchFirst�NoMatch�NotAny�	OneOrMore�OnlyOnce�Optional�Or�ParseBaseException�ParseElementEnhance�ParseException�ParseExpression�ParseFatalException�ParseResults�ParseSyntaxException�
ParserElement�QuotedString�RecursiveGrammarException�Regex�SkipTo�	StringEnd�StringStart�Suppress�Token�TokenConverter�White�Word�WordEnd�	WordStart�
ZeroOrMore�	alphanums�alphas�
alphas8bit�anyCloseTag�
anyOpenTag�
cStyleComment�col�commaSeparatedList�commonHTMLEntity�countedArray�cppStyleComment�dblQuotedString�dblSlashComment�
delimitedList�dictOf�downcaseTokens�empty�hexnums�htmlComment�javaStyleComment�line�lineEnd�	lineStart�lineno�makeHTMLTags�makeXMLTags�matchOnlyAtCol�matchPreviousExpr�matchPreviousLiteral�
nestedExpr�nullDebugAction�nums�oneOf�opAssoc�operatorPrecedence�
printables�punc8bit�pythonStyleComment�quotedString�removeQuotes�replaceHTMLEntity�replaceWith�
restOfLine�sglQuotedString�srange�	stringEnd�stringStart�traceParseAction�
unicodeString�upcaseTokens�
withAttribute�
indentedBlock�originalTextFor�ungroup�
infixNotation�locatedExpr�	withClass�
CloseMatch�tokenMap�pyparsing_common�cCsdt|t�r|Sz
t|�WSty^t|��t��d�}td�}|�dd��|�	|�YS0dS)aDrop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
           str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
           then < returns the unicode object | encodes it with the default encoding | ... >.
        �xmlcharrefreplacez&#\d+;cSs$dtt|ddd���dd�S)Nz\ur����)�hex�int��t�ry��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/pyparsing.py�<lambda>��z_ustr.<locals>.<lambda>N)
�
isinstance�unicode�str�UnicodeEncodeError�encode�sys�getdefaultencodingr)�setParseAction�transformString)�obj�retZ
xmlcharrefryryrz�_ustr�s

r�z6sum len sorted reversed list tuple set any all min maxccs|]
}|VqdS�Nry)�.0�yryryrz�	<genexpr>�r|r��cCs:d}dd�d��D�}t||�D]\}}|�||�}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)�&�;Nry)r��sryryrzr��r|z_xml_escape.<locals>.<genexpr>zamp gt lt quot apos)�split�zip�replace)�data�from_symbols�
to_symbols�from_�to_ryryrz�_xml_escape�s
r�c@seZdZdS)�
_ConstantsN)�__name__�
__module__�__qualname__ryryryrzr��sr��
0123456789ZABCDEFabcdef�\�ccs|]}|tjvr|VqdSr�)�string�
whitespace�r��cryryrzr��r|c@sPeZdZdZddd�Zedd��Zdd	�Zd
d�Zdd
�Z	ddd�Z
dd�ZdS)rz7base exception class for all parsing runtime exceptionsrNcCs>||_|dur||_d|_n||_||_||_|||f|_dS�Nr�)�loc�msg�pstr�
parserElement�args)�selfr�r�r��elemryryrz�__init__�szParseBaseException.__init__cCs||j|j|j|j�S)z�
        internal factory method to simplify creating one type of ParseException 
        from another - avoids having __init__ signature conflicts among subclasses
        )r�r�r�r�)�cls�peryryrz�_from_exception�sz"ParseBaseException._from_exceptioncCsN|dkrt|j|j�S|dvr,t|j|j�S|dkrBt|j|j�St|��dS)z�supported attributes by name are:
            - lineno - returns the line number of the exception text
            - col - returns the column number of the exception text
            - line - returns the line containing the exception text
        rL)r;�columnrIN)rLr�r�r;rI�AttributeError)r��anameryryrz�__getattr__�szParseBaseException.__getattr__cCsd|j|j|j|jfS)Nz"%s (at char %d), (line:%d, col:%d))r�r�rLr��r�ryryrz�__str__�s�zParseBaseException.__str__cCst|�Sr��r�r�ryryrz�__repr__�szParseBaseException.__repr__�>!<cCs<|j}|jd}|r4d�|d|�|||d�f�}|��S)z�Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        r�r�N)rIr��join�strip)r��markerString�line_str�line_columnryryrz�
markInputline�s
�z ParseBaseException.markInputlinecCsd��tt|��S)Nzlineno col line)r��dir�typer�ryryrz�__dir__szParseBaseException.__dir__)rNN)r�)r�r�r��__doc__r��classmethodr�r�r�r�r�r�ryryryrzr�s



rc@seZdZdZdS)r!aN
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
     - lineno - returns the line number of the exception text
     - col - returns the column number of the exception text
     - line - returns the line containing the exception text
        
    Example::
        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))
            
    prints::
       Expected integer (at char 0), (line:1, col:1)
        column: 1
    N�r�r�r�r�ryryryrzr!sr!c@seZdZdZdS)r#znuser-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediatelyNr�ryryryrzr#sr#c@seZdZdZdS)r%z�just like L{ParseFatalException}, but thrown internally when an
       L{ErrorStop<And._ErrorStop>} ('-' operator) indicates that parsing is to stop 
       immediately because an unbacktrackable syntax error has been foundNr�ryryryrzr%sr%c@s eZdZdZdd�Zdd�ZdS)r(zZexception thrown by L{ParserElement.validate} if the grammar could be improperly recursivecCs
||_dSr���parseElementTrace�r��parseElementListryryrzr�4sz"RecursiveGrammarException.__init__cCs
d|jS)NzRecursiveGrammarException: %sr�r�ryryrzr�7sz!RecursiveGrammarException.__str__N)r�r�r�r�r�r�ryryryrzr(2sr(c@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�_ParseResultsWithOffsetcCs||f|_dSr���tup)r��p1�p2ryryrzr�;sz _ParseResultsWithOffset.__init__cCs
|j|Sr�r��r��iryryrz�__getitem__=sz#_ParseResultsWithOffset.__getitem__cCst|jd�S�Nr)�reprr�r�ryryrzr�?sz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSr�r�r�ryryrz�	setOffsetAsz!_ParseResultsWithOffset.setOffsetN)r�r�r�r�r�r�r�ryryryrzr�:sr�c@s�eZdZdZd[dd�Zddddefdd�Zdd	�Zefd
d�Zdd
�Z	dd�Z
dd�Zdd�ZeZ
dd�Zdd�Zdd�Zdd�Zdd�Zer�eZeZeZn$eZeZeZdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd\d(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Z d2d3�Z!d4d5�Z"d6d7�Z#d8d9�Z$d:d;�Z%d<d=�Z&d]d?d@�Z'dAdB�Z(dCdD�Z)dEdF�Z*d^dHdI�Z+dJdK�Z,dLdM�Z-d_dOdP�Z.dQdR�Z/dSdT�Z0dUdV�Z1dWdX�Z2dYdZ�Z3dS)`r$aI
    Structured parse results, to provide multiple means of access to the parsed data:
       - as a list (C{len(results)})
       - by list index (C{results[0], results[1]}, etc.)
       - by attribute (C{results.<resultsName>} - see L{ParserElement.setResultsName})

    Example::
        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)
    prints::
        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    NTcCs"t||�r|St�|�}d|_|S�NT)r}�object�__new__�_ParseResults__doinit)r��toklist�name�asList�modal�retobjryryrzr�ks


zParseResults.__new__c
Cs`|jrvd|_d|_d|_i|_||_||_|dur6g}||t�rP|dd�|_n||t�rft|�|_n|g|_t	�|_
|du�r\|�r\|s�d|j|<||t�r�t|�}||_||t
d�ttf�r�|ddgfv�s\||t�r�|g}|�r(||t��rt|��d�||<ntt|d�d�||<|||_n4z|d||<Wn"tttf�yZ|||<Yn0dS)NFrr�)r��_ParseResults__name�_ParseResults__parent�_ParseResults__accumNames�_ParseResults__asList�_ParseResults__modal�list�_ParseResults__toklist�_generatorType�dict�_ParseResults__tokdictrvr�r��
basestringr$r��copy�KeyError�	TypeError�
IndexError)r�r�r�r�r�r}ryryrzr�tsB



$
zParseResults.__init__cCsPt|ttf�r|j|S||jvr4|j|ddStdd�|j|D��SdS)NrtrcSsg|]}|d�qS�rry�r��vryryrz�
<listcomp>�r|z,ParseResults.__getitem__.<locals>.<listcomp>)r}rv�slicer�r�r�r$r�ryryrzr��s


zParseResults.__getitem__cCs�||t�r0|j�|t��|g|j|<|d}nD||ttf�rN||j|<|}n&|j�|t��t|d�g|j|<|}||t�r�t|�|_	dSr�)
r�r��getr�rvr�r�r$�wkrefr�)r��kr�r}�subryryrz�__setitem__�s


"
zParseResults.__setitem__c
Cs�t|ttf�r�t|j�}|j|=t|t�rH|dkr:||7}t||d�}tt|�|���}|��|j	�
�D]>\}}|D]0}t|�D]"\}\}}	t||	|	|k�||<q�qxqln|j	|=dS�Nrr�)
r}rvr��lenr�r��range�indices�reverser��items�	enumerater�)
r�r��mylen�removedr��occurrences�jr��value�positionryryrz�__delitem__�s

zParseResults.__delitem__cCs
||jvSr�)r�)r�r�ryryrz�__contains__�szParseResults.__contains__cCs
t|j�Sr�)r�r�r�ryryrz�__len__�r|zParseResults.__len__cCs
|jSr��r�r�ryryrz�__bool__�r|zParseResults.__bool__cCs
t|j�Sr���iterr�r�ryryrz�__iter__�r|zParseResults.__iter__cCst|jddd��S�Nrtr	r�ryryrz�__reversed__�r|zParseResults.__reversed__cCs$t|jd�r|j��St|j�SdS)N�iterkeys)�hasattrr�rr
r�ryryrz�	_iterkeys�s
zParseResults._iterkeyscs�fdd����D�S)Nc3s|]}�|VqdSr�ry�r�r�r�ryrzr��r|z+ParseResults._itervalues.<locals>.<genexpr>�rr�ryr�rz�_itervalues�szParseResults._itervaluescs�fdd����D�S)Nc3s|]}|�|fVqdSr�ryrr�ryrzr��r|z*ParseResults._iteritems.<locals>.<genexpr>rr�ryr�rz�
_iteritems�szParseResults._iteritemscCst|���S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)r�rr�ryryrz�keys�szParseResults.keyscCst|���S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r��
itervaluesr�ryryrz�values�szParseResults.valuescCst|���S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r��	iteritemsr�ryryrzr��szParseResults.itemscCs
t|j�S)z�Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names.)�boolr�r�ryryrz�haskeys�szParseResults.haskeyscOs�|s
dg}|��D]*\}}|dkr0|d|f}qtd|��qt|dt�sdt|�dksd|d|vr~|d}||}||=|S|d}|SdS)a�
        Removes and returns item at specified index (default=C{last}).
        Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
        argument or an integer argument, it will use C{list} semantics
        and pop tokens from the list of parsed tokens. If passed a 
        non-integer argument (most likely a string), it will use C{dict}
        semantics and pop the corresponding value from any defined 
        results names. A second default return value argument is 
        supported, just as in C{dict.pop()}.

        Example::
            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())
        prints::
            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        rt�defaultrz-pop() got an unexpected keyword argument '%s'r�N)r�r�r}rvr�)r�r��kwargsr�r��indexr��defaultvalueryryrz�pop�s""
�
�zParseResults.popcCs||vr||S|SdS)ai
        Returns named result matching the given key, or if there is no
        such name, then returns the given C{defaultValue} or C{None} if no
        C{defaultValue} is specified.

        Similar to C{dict.get()}.
        
        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        Nry)r��key�defaultValueryryrzr�3szParseResults.getcCsR|j�||�|j��D]4\}}t|�D]"\}\}}t||||k�||<q(qdS)a
        Inserts new element at location index in the list of parsed tokens.
        
        Similar to C{list.insert()}.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        N)r��insertr�r�r�r�)r�r�insStrr�rr�rrryryrzr"IszParseResults.insertcCs|j�|�dS)a�
        Add single element to end of ParseResults list of elements.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            
            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        N)r��append)r��itemryryrzr$]szParseResults.appendcCs$t|t�r||7}n|j�|�dS)a
        Add sequence of elements to end of ParseResults list of elements.

        Example::
            patt = OneOrMore(Word(alphas))
            
            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        N)r}r$r��extend)r��itemseqryryrzr&ks

zParseResults.extendcCs|jdd�=|j��dS)z7
        Clear all elements and results names.
        N)r�r��clearr�ryryrzr(}szParseResults.clearcCshz
||WStyYdS0||jvr`||jvrF|j|ddStdd�|j|D��SndSdS)Nr�rtrcSsg|]}|d�qSr�ryr�ryryrzr��r|z,ParseResults.__getattr__.<locals>.<listcomp>)r�r�r�r$�r�r�ryryrzr��s


zParseResults.__getattr__cCs|��}||7}|Sr��r�)r��otherr�ryryrz�__add__�szParseResults.__add__cs�|jrjt|j���fdd��|j��}�fdd�|D�}|D],\}}|||<t|dt�r<t|�|d_q<|j|j7_|j�	|j�|S)Ncs|dkr�S|�Sr�ry)�a)�offsetryrzr{�r|z'ParseResults.__iadd__.<locals>.<lambda>c	s4g|],\}}|D]}|t|d�|d��f�qqS�rr�)r��r�r��vlistr�)�	addoffsetryrzr��s�z)ParseResults.__iadd__.<locals>.<listcomp>r)
r�r�r�r�r}r$r�r�r��update)r�r+�
otheritems�otherdictitemsr�r�ry)r2r.rz�__iadd__�s


�zParseResults.__iadd__cCs&t|t�r|dkr|��S||SdSr�)r}rvr��r�r+ryryrz�__radd__�szParseResults.__radd__cCsdt|j�t|j�fS)Nz(%s, %s))r�r�r�r�ryryrzr��szParseResults.__repr__cCsdd�dd�|jD��dS)N�[�, css(|] }t|t�rt|�nt|�VqdSr�)r}r$r�r��r�r�ryryrzr��r|z'ParseResults.__str__.<locals>.<genexpr>�])r�r�r�ryryrzr��szParseResults.__str__r�cCsLg}|jD]<}|r |r |�|�t|t�r8||��7}q
|�t|��q
|Sr�)r�r$r}r$�
_asStringListr�)r��sep�outr%ryryrzr=�s


zParseResults._asStringListcCsdd�|jD�S)a�
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::
            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']
            
            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        cSs"g|]}t|t�r|��n|�qSry)r}r$r�)r��resryryrzr��r|z'ParseResults.asList.<locals>.<listcomp>rr�ryryrzr��szParseResults.asListcs6tr|j}n|j}�fdd��t�fdd�|�D��S)a�
        Returns the named parse results as a nested dictionary.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
            
            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        cs6t|t�r.|��r|��S�fdd�|D�Sn|SdS)Ncsg|]}�|��qSryryr���toItemryrzr��r|z7ParseResults.asDict.<locals>.toItem.<locals>.<listcomp>)r}r$r�asDict)r�rAryrzrB�s

z#ParseResults.asDict.<locals>.toItemc3s|]\}}|�|�fVqdSr�ry�r�r�r�rAryrzr��r|z&ParseResults.asDict.<locals>.<genexpr>)�PY_3r�rr�)r��item_fnryrArzrC�s
	zParseResults.asDictcCs8t|j�}|j��|_|j|_|j�|j�|j|_|S)zA
        Returns a new copy of a C{ParseResults} object.
        )r$r�r�r�r�r�r3r��r�r�ryryrzr��s
zParseResults.copyFcCsLd}g}tdd�|j��D��}|d}|s8d}d}d}d}	|durJ|}	n|jrV|j}	|	sf|rbdSd}	|||d|	d	g7}t|j�D]�\}
}t|t�r�|
|vr�||�||
|o�|du||�g7}n||�d|o�|du||�g7}q�d}|
|vr�||
}|�s|�rq�nd}t	t
|��}
|||d|d	|
d
|d	g	7}q�|||d
|	d	g7}d�|�S)z�
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        �
css(|] \}}|D]}|d|fVqqdS�r�Nryr0ryryrzr�s
�z%ParseResults.asXML.<locals>.<genexpr>�  r�N�ITEM�<�>�</)r�r�r�r�r�r�r}r$�asXMLr�r�r�)r��doctag�namedItemsOnly�indent�	formatted�nlr?�
namedItems�nextLevelIndent�selfTagr�r@�resTag�xmlBodyTextryryrzrO�sZ

�

�
�zParseResults.asXMLcCs:|j��D]*\}}|D]\}}||ur|Sqq
dSr�)r�r�)r�r�r�r1r�r�ryryrz�__lookup;s
zParseResults.__lookupcCs�|jr|jS|jr.|��}|r(|�|�SdSnNt|�dkrxt|j�dkrxtt|j����dddvrxtt|j����SdSdS)a(
        Returns the results name for this token expression. Useful when several 
        different expressions might match at a particular location.

        Example::
            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number") 
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)
            
            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])
        prints::
            age : 22
            ssn : 111-22-3333
            house_number : 221B
        Nr�r)rrt)	r�r��_ParseResults__lookupr�r��nextr
rr)r��parryryrz�getNameBs
��zParseResults.getNamercCsZg}d}|�|t|����|�rP|��r�tdd�|��D��}|D]r\}}|r\|�|�|�d|d||f�t|t�r�|r�|�|�||d��q�|�t|��qF|�t	|��qFn�t
dd�|D���rP|}t|�D]r\}	}
t|
t��r$|�d|d||	|d|d|
�||d�f�q�|�d|d||	|d|dt|
�f�q�d	�|�S)
aH
        Diagnostic method for listing out the contents of a C{ParseResults}.
        Accepts an optional C{indent} argument so that this string can be embedded
        in a nested display of other data.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(result.dump())
        prints::
            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        rHcss|]\}}t|�|fVqdSr�)rrDryryrzr�~r|z$ParseResults.dump.<locals>.<genexpr>z
%s%s- %s: rJr�css|]}t|t�VqdSr�)r}r$)r��vvryryrzr��r|z
%s%s[%d]:
%s%s%sr�)
r$r�r�r�sortedr�r}r$�dumpr��anyr�r�)r�rR�depth�fullr?�NLr�r�r�r�r_ryryrzrags,

4,zParseResults.dumpcOs tj|��g|�Ri|��dS)a�
        Pretty-printer for parsed results as a list, using the C{pprint} module.
        Accepts additional positional or keyword args as defined for the 
        C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})

        Example::
            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)
        prints::
            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        N)�pprintr��r�r�rryryrzrf�szParseResults.pprintcCs.|j|j��|jdur|��p d|j|jffSr�)r�r�r�r�r�r�r�ryryrz�__getstate__�s��zParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j�|�|durDt|�|_nd|_dSr�)r�r�r�r�r3r�r�)r��stater]�inAccumNamesryryrz�__setstate__�s
�zParseResults.__setstate__cCs|j|j|j|jfSr�)r�r�r�r�r�ryryrz�__getnewargs__�szParseResults.__getnewargs__cCstt|��t|���Sr�)r�r�r�rr�ryryrzr��szParseResults.__dir__)NNTT)N)r�)NFr�T)r�rT)4r�r�r�r�r�r}r�r�r�rrrr�__nonzero__rr
rrrrErrr�rrrrrr�r"r$r&r(r�r,r6r8r�r�r=r�rCr�rOr[r^rarfrhrkrlr�ryryryrzr$Dsh&
	'	
4

#
=%
-
r$cCsF|}d|krt|�kr4nn||ddkr4dS||�dd|�S)aReturns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   rr�rH)r��rfind)r��strgr�ryryrzr;�s
r;cCs|�dd|�dS)aReturns current line number within a string, counting newlines as line separators.
   The first line is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   rHrr�)�count)r�roryryrzrL�s
rLcCsF|�dd|�}|�d|�}|dkr2||d|�S||dd�SdS)zfReturns the line of text containing loc within a string, counting newlines as line separators.
       rHrr�N)rn�find)r�ro�lastCR�nextCRryryrzrI�s
rIcCs8tdt|�dt|�dt||�t||�f�dS)NzMatch z at loc z(%d,%d))�printr�rLr;)�instringr��exprryryrz�_defaultStartDebugAction�srwcCs$tdt|�dt|����dS)NzMatched z -> )rtr�rr�)ru�startloc�endlocrv�toksryryrz�_defaultSuccessDebugAction�sr{cCstdt|��dS)NzException raised:)rtr�)rur�rv�excryryrz�_defaultExceptionDebugAction�sr}cGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nry)r�ryryrzrS�srSrscs��tvr�fdd�Sdg�dg�tdd�dkrFddd�}dd	d
��ntj}tj�d}|dd�d
}|d|d|f�������fdd�}d}zt�dt�d�j�}Wnty�t��}Yn0||_|S)Ncs�|�Sr�ry�r��lrx)�funcryrzr{r|z_trim_arity.<locals>.<lambda>rFrs)rq�cSs8tdkrdnd}tj||dd�|}|dd�gS)N)rqr�r������r���limitrs)�system_version�	traceback�
extract_stack)r�r.�
frame_summaryryryrzr�sz"_trim_arity.<locals>.extract_stackcSs$tj||d�}|d}|dd�gS)Nr�rtrs)r��
extract_tb)�tbr��framesr�ryryrzr�sz_trim_arity.<locals>.extract_tb�r�rtr�c	s�z"�|�dd��}d�d<|WSty��dr<�n6z0t��d}�|dd�ddd��ksh�W~n~0�d�kr��dd7<Yq�Yq0qdS)NrTrtrsr�r�)r�r��exc_info)r�r�r��r��
foundArityr�r��maxargs�pa_call_line_synthryrz�wrapper-s z_trim_arity.<locals>.wrapperz<parse action>r��	__class__)r)r)	�singleArgBuiltinsr�r�r�r��getattrr��	Exceptionr)r�r�r��	LINE_DIFF�	this_liner��	func_nameryr�rz�_trim_aritys,

�r�cs�eZdZdZdZdZedd��Zedd��Zd�dd	�Z	d
d�Z
dd
�Zd�dd�Zd�dd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd�dd �Zd!d"�Zd�d#d$�Zd%d&�Zd'd(�ZGd)d*�d*e�Zed+ur�Gd,d-�d-e�ZnGd.d-�d-e�ZiZe�Zd/d/gZ d�d0d1�Z!eZ"ed2d3��Z#dZ$ed�d5d6��Z%d�d7d8�Z&e'dfd9d:�Z(d;d<�Z)e'fd=d>�Z*e'dfd?d@�Z+dAdB�Z,dCdD�Z-dEdF�Z.dGdH�Z/dIdJ�Z0dKdL�Z1dMdN�Z2dOdP�Z3dQdR�Z4dSdT�Z5dUdV�Z6dWdX�Z7dYdZ�Z8d�d[d\�Z9d]d^�Z:d_d`�Z;dadb�Z<dcdd�Z=dedf�Z>dgdh�Z?d�didj�Z@dkdl�ZAdmdn�ZBdodp�ZCdqdr�ZDgfdsdt�ZEd�dudv�ZF�fdwdx�ZGdydz�ZHd{d|�ZId}d~�ZJdd��ZKd�d�d��ZLd�d�d��ZM�ZNS)�r&z)Abstract base level parser element class.z 
	
FcCs
|t_dS)a�
        Overrides the default whitespace chars

        Example::
            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']
            
            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        N)r&�DEFAULT_WHITE_CHARS��charsryryrz�setDefaultWhitespaceCharsTs
z'ParserElement.setDefaultWhitespaceCharscCs
|t_dS)a�
        Set class to be used for inclusion of string literals into a parser.
        
        Example::
            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        N)r&�_literalStringClass)r�ryryrz�inlineLiteralsUsingcsz!ParserElement.inlineLiteralsUsingcCs�t�|_d|_d|_d|_||_d|_tj|_	d|_
d|_d|_t�|_
d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr�)NNN)r��parseAction�
failAction�strRepr�resultsName�
saveAsList�skipWhitespacer&r��
whiteChars�copyDefaultWhiteChars�mayReturnEmpty�keepTabs�ignoreExprs�debug�streamlined�
mayIndexError�errmsg�modalResults�debugActions�re�callPreparse�
callDuringTry)r��savelistryryrzr�xs(zParserElement.__init__cCs<t�|�}|jdd�|_|jdd�|_|jr8tj|_|S)a$
        Make a copy of this C{ParserElement}.  Useful for defining different parse actions
        for the same parsing pattern, using copies of the original parse element.
        
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
            
            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
        prints::
            [5120, 100, 655360, 268435456]
        Equivalent form of C{expr.copy()} is just C{expr()}::
            integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
        N)r�r�r�r�r&r�r�)r��cpyryryrzr��s
zParserElement.copycCs*||_d|j|_t|d�r&|j|j_|S)af
        Define name for this expression, makes debugging and exception messages clearer.
        
        Example::
            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        �	Expected �	exception)r�r�rr�r�r)ryryrz�setName�s


zParserElement.setNamecCs4|��}|�d�r"|dd�}d}||_||_|S)aP
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original C{ParserElement} object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        C{expr("name")} in place of C{expr.setResultsName("name")} - 
        see L{I{__call__}<__call__>}.

        Example::
            date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        �*NrtT)r��endswithr�r�)r�r��listAllMatches�newselfryryrz�setResultsName�s
zParserElement.setResultsNameTcs@|r&|j�d�fdd�	}�|_||_nt|jd�r<|jj|_|S)z�Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set C{breakFlag} to True to enable, False to
           disable.
        Tcsddl}|���||||�Sr�)�pdb�	set_trace)rur��	doActions�callPreParser���_parseMethodryrz�breaker�sz'ParserElement.setBreak.<locals>.breaker�_originalParseMethod)TT)�_parser�r)r��	breakFlagr�ryr�rz�setBreak�s
zParserElement.setBreakcOs&tttt|���|_|�dd�|_|S)a
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
        C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
         - s   = the original string being parsed (see note below)
         - loc = the location of the matching substring
         - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        Optional keyword arguments:
         - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See L{I{parseString}<parseString>} for more information
        on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
        consistent view of the parsed string, the parse location, and line and column
        positions within the parsed string.
        
        Example::
            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        r�F)r��mapr�r�r�r��r��fnsrryryrzr��s"zParserElement.setParseActioncOs4|jtttt|���7_|jp,|�dd�|_|S)z�
        Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}<setParseAction>}.
        
        See examples in L{I{copy}<copy>}.
        r�F)r�r�r�r�r�r�r�ryryrz�addParseActionszParserElement.addParseActioncs^|�dd��|�dd�rtnt�|D] ����fdd�}|j�|�q$|jpV|�dd�|_|S)a�Add a boolean predicate function to expression's list of parse actions. See 
        L{I{setParseAction}<setParseAction>} for function call signatures. Unlike C{setParseAction}, 
        functions passed to C{addCondition} need to return boolean success/fail of the condition.

        Optional keyword arguments:
         - message = define a custom message to be used in the raised exception
         - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
         
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        �messagezfailed user-defined condition�fatalFcs$tt��|||��s �||���dSr�)rr�r~��exc_type�fnr�ryrz�pa&sz&ParserElement.addCondition.<locals>.par�)r�r#r!r�r$r�)r�r�rr�ryr�rz�addConditionszParserElement.addConditioncCs
||_|S)aDefine action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           C{fn(s,loc,expr,err)} where:
            - s = string being parsed
            - loc = location where expression match was attempted and failed
            - expr = the parse expression that failed
            - err = the exception thrown
           The function returns no value.  It may throw C{L{ParseFatalException}}
           if it is desired to stop parsing immediately.)r�)r�r�ryryrz�
setFailAction-s
zParserElement.setFailActionc	CsLd}|rHd}|jD]2}z|�||�\}}d}qWqtyBYq0qq|S�NTF)r�r�r!)r�rur��
exprsFound�e�dummyryryrz�_skipIgnorables:s


zParserElement._skipIgnorablescCsH|jr|�||�}|jrD|j}t|�}||krD|||vrD|d7}q&|S�Nr�)r�r�r�r�r�)r�rur��wt�instrlenryryrz�preParseGs
zParserElement.preParsecCs|gfSr�ry�r�rur�r�ryryrz�	parseImplSszParserElement.parseImplcCs|Sr�ry�r�rur��	tokenlistryryrz�	postParseVszParserElement.postParsec
Cs�|j}|s|jr�|jdr,|jd|||�|rD|jrD|�||�}n|}|}zBz|�|||�\}}Wn&ty�t|t|�|j	|��Yn0WnZt
y�}	zB|jdr�|jd||||	�|jr�|�||||	��WYd}	~	n
d}	~	00n�|�r|j�r|�||�}n|}|}|j�s&|t|�k�rhz|�|||�\}}Wn(t�ydt|t|�|j	|��Yn0n|�|||�\}}|�|||�}t
||j|j|jd�}
|j�r�|�s�|j�r�|�rTzN|jD]B}||||
�}|du�r�t
||j|j�o�t|t
tf�|jd�}
�q�WnHt
�yP}	z.|jd�r:|jd||||	��WYd}	~	n
d}	~	00nJ|jD]B}||||
�}|du�rZt
||j|j�o�t|t
tf�|jd�}
�qZ|�r�|jd�r�|jd|||||
�||
fS)Nrrs)r�r�r�)r�r�r�r�r�r�r�r!r�r�rr�r�r$r�r�r�r�r�r}r�)r�rur�r�r��	debugging�preloc�tokensStart�tokens�err�	retTokensr�ryryrz�
_parseNoCacheZst





�

�
zParserElement._parseNoCachecCs>z|j||dd�dWSty8t|||j|��Yn0dS)NF)r�r)r�r#r!r��r�rur�ryryrz�tryParse�szParserElement.tryParsec	Cs2z|�||�Wnttfy(YdS0dSdS)NFT)r�r!r�r�ryryrz�canParseNext�s
zParserElement.canParseNextc@seZdZdd�ZdS)zParserElement._UnboundedCachecs~i�t�|_���fdd�}�fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr��r��r�r ��cache�not_in_cacheryrzr��sz3ParserElement._UnboundedCache.__init__.<locals>.getcs|�|<dSr�ry�r�r r�r�ryrz�set�sz3ParserElement._UnboundedCache.__init__.<locals>.setcs���dSr��r(r�r�ryrzr(�sz5ParserElement._UnboundedCache.__init__.<locals>.clearcst��Sr��r�r�r�ryrz�	cache_len�sz9ParserElement._UnboundedCache.__init__.<locals>.cache_len)r�r��types�
MethodTyper�r�r(r)r�r�r�r(r�ryr�rzr��sz&ParserElement._UnboundedCache.__init__N�r�r�r�r�ryryryrz�_UnboundedCache�sr�Nc@seZdZdd�ZdS)�ParserElement._FifoCachecs�t�|_�t����fdd�}��fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr�r�r�r�ryrzr��s�.ParserElement._FifoCache.__init__.<locals>.getcs<|�|<t���kr8z��d�Wqty4Yq0qdS�NF)r��popitemr�r�)r��sizeryrzr��s�.ParserElement._FifoCache.__init__.<locals>.setcs���dSr�r�r�r�ryrzr(�s�0ParserElement._FifoCache.__init__.<locals>.clearcst��Sr�r�r�r�ryrzr��s�4ParserElement._FifoCache.__init__.<locals>.cache_len)	r�r��_OrderedDictr�r�r�r�r(r�r�rr�r�r(r�ry)r�r�rrzr��s�!ParserElement._FifoCache.__init__Nr�ryryryrz�
_FifoCache�sr	c@seZdZdd�ZdS)r�cs�t�|_�i�t�g�����fdd�}���fdd�}��fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_	dS)	Ncs��|��Sr�r�r�r�ryrzr��sr�cs4|�|<t���kr&�����d�q��|�dSr�)r�r�popleftr$r�)r��key_fiforryrzr��srcs������dSr�r�r�)r�rryrzr(�srcst��Sr�r�r�r�ryrzr��sr)
r�r��collections�dequer�r�r�r�r(rrry)r�rr�rrzr��srNr�ryryryrzr	�srcCs0d\}}|||||f}tj��tj}|�|�}	|	|jur�tj|d7<z|�||||�}	Wn:ty�}
z"|�||
j	|
j
���WYd}
~
n8d}
~
00|�||	d|	d��f�|	Wd�SnBtj|d7<t|	t
�r�|	�|	d|	d��fWd�SWd�n1�s"0YdS)Nr/r�r)r&�packrat_cache_lock�
packrat_cacher�r��packrat_cache_statsr�rr�r�r�r�r}r�)r�rur�r�r��HIT�MISS�lookupr�rr�ryryrz�_parseCaches$


zParserElement._parseCachecCs(tj��dgttj�tjdd�<dSr�)r&rr(r�rryryryrz�
resetCaches
zParserElement.resetCache�cCs8tjs4dt_|dur t��t_nt�|�t_tjt_dS)a�Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.
           
           Parameters:
            - cache_size_limit - (default=C{128}) - if an integer value is provided
              will limit the size of the packrat cache; if None is passed, then
              the cache size will be unbounded; if 0 is passed, the cache will
              be effectively disabled.
            
           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method C{ParserElement.enablePackrat()}.  If
           your program uses C{psyco} to "compile as you go", you must call
           C{enablePackrat} before calling C{psyco.full()}.  If you do not do this,
           Python will crash.  For best results, call C{enablePackrat()} immediately
           after importing pyparsing.
           
           Example::
               import pyparsing
               pyparsing.ParserElement.enablePackrat()
        TN)r&�_packratEnabledr�rr	rr�)�cache_size_limitryryrz�
enablePackrat%szParserElement.enablePackratc
Cs�t��|js|��|jD]}|��q|js8|��}z<|�|d�\}}|rr|�||�}t	�t
�}|�||�Wn2ty�}ztjr��n|�WYd}~nd}~00|SdS)aC
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set C{parseAll} to True (equivalent to ending
        the grammar with C{L{StringEnd()}}).

        Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the C{loc} argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:
         - calling C{parseWithTabs} on your grammar before calling C{parseString}
           (see L{I{parseWithTabs}<parseWithTabs>})
         - define your parse action using the full C{(s,loc,toks)} signature, and
           reference the input string using the parse action's C{s} argument
         - explicitly expand the tabs in your input string before calling
           C{parseString}
        
        Example::
            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        rN)
r&rr��
streamliner�r��
expandtabsr�r�rr+r�verbose_stacktrace)r�ru�parseAllr�r�r��ser|ryryrz�parseStringHs$

zParserElement.parseStringc
cs6|js|��|jD]}|��q|js4t|���}t|�}d}|j}|j}t	�
�d}	z�||kr�|	|kr�z |||�}
|||
dd�\}}Wnty�|
d}YqZ0||kr�|	d7}	||
|fV|r�|||�}
|
|kr�|}q�|d7}q�|}qZ|
d}qZWn6t�y0}zt	j
�r�n|�WYd}~n
d}~00dS)a�
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        C{maxMatches} argument, to clip scanning after 'n' matches are found.  If
        C{overlap} is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See L{I{parseString}<parseString>} for more information on parsing
        strings with embedded tabs.

        Example::
            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens,start,end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])
        
        prints::
        
            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        rF�r�r�N)r�rr�r�r�rr�r�r�r&rr!rr)r�ru�
maxMatches�overlapr�r�r��
preparseFn�parseFn�matchesr��nextLocr��nextlocr|ryryrz�
scanStringzsB




zParserElement.scanStringc
Cs�g}d}d|_z�|�|�D]Z\}}}|�|||��|rpt|t�rR||��7}nt|t�rf||7}n
|�|�|}q|�||d��dd�|D�}d�tt	t
|���WSty�}ztj
rĂn|�WYd}~n
d}~00dS)af
        Extension to C{L{scanString}}, to modify matching text with modified tokens that may
        be returned from a parse action.  To use C{transformString}, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking C{transformString()} on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  C{transformString()} returns the resulting transformed string.
        
        Example::
            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())
            
            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
        Prints::
            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        rTNcSsg|]}|r|�qSryry)r��oryryrzr��r|z1ParserElement.transformString.<locals>.<listcomp>r�)r�r(r$r}r$r�r�r�r�r��_flattenrr&r)r�rur?�lastErxr�r�r|ryryrzr��s(



zParserElement.transformStringc
CsTztdd�|�||�D��WStyN}ztjr6�n|�WYd}~n
d}~00dS)a�
        Another extension to C{L{scanString}}, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        C{maxMatches} argument, to clip searching after 'n' matches are found.
        
        Example::
            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())
            
            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))
        prints::
            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        cSsg|]\}}}|�qSryry)r�rxr�r�ryryrzr��r|z.ParserElement.searchString.<locals>.<listcomp>N)r$r(rr&r)r�rur!r|ryryrz�searchString�szParserElement.searchStringc	csTd}d}|j||d�D]*\}}}|||�V|r<|dV|}q||d�VdS)a[
        Generator method to split a string using the given expression as a separator.
        May be called with optional C{maxsplit} argument, to limit the number of splits;
        and the optional C{includeSeparators} argument (default=C{False}), if the separating
        matching text should be included in the split results.
        
        Example::        
            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
        prints::
            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        r)r!N)r()	r�ru�maxsplit�includeSeparators�splits�lastrxr�r�ryryrzr�s

zParserElement.splitcCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)a�
        Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
        converts them to L{Literal}s by default.
        
        Example::
            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))
        Prints::
            Hello, World! -> ['Hello', ',', 'World', '!']
        �4Cannot combine element of type %s with ParserElementrs��
stacklevelN)	r}r�r&r��warnings�warnr��
SyntaxWarningrr7ryryrzr,s


�zParserElement.__add__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||S)z]
        Implementation of + operator when left operand is not a C{L{ParserElement}}
        r1rsr2N�r}r�r&r�r4r5r�r6r7ryryrzr81s


�zParserElement.__radd__cCsJt|t�rt�|�}t|t�s:tjdt|�tdd�dS|t�	�|S)zQ
        Implementation of - operator, returns C{L{And}} with error stop
        r1rsr2N)
r}r�r&r�r4r5r�r6r�
_ErrorStopr7ryryrz�__sub__=s


�zParserElement.__sub__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||S)z]
        Implementation of - operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rsub__Is


�zParserElement.__rsub__cs�t|t�r|d}}n�t|t�r�|ddd�}|ddurHd|df}t|dt�r�|ddur�|ddkrvt��S|ddkr�t��S�|dt��Sq�t|dt�r�t|dt�r�|\}}||8}q�tdt|d�t|d���ntdt|���|dk�rtd��|dk�rtd	��||k�r6dk�rBnntd
��|�r���fdd��|�r�|dk�rt��|�}nt�g|��|�}n�|�}n|dk�r��}nt�g|�}|S)
a�
        Implementation of * operator, allows use of C{expr * 3} in place of
        C{expr + expr + expr}.  Expressions may also me multiplied by a 2-integer
        tuple, similar to C{{min,max}} multipliers in regular expressions.  Tuples
        may also include C{None} as in:
         - C{expr*(n,None)} or C{expr*(n,)} is equivalent
              to C{expr*n + L{ZeroOrMore}(expr)}
              (read as "at least n instances of C{expr}")
         - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
              (read as "0 to n instances of C{expr}")
         - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
         - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}

        Note that C{expr*(None,n)} does not raise an exception if
        more than n exprs exist in the input stream; that is,
        C{expr*(None,n)} does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        C{expr*(None,n) + ~expr}
        r)NNNrsr�z7cannot multiply 'ParserElement' and ('%s','%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez+cannot multiply ParserElement by 0 or (0,0)cs(|dkrt��|d��St��SdSr�)r��n��makeOptionalListr�ryrzr>�sz/ParserElement.__mul__.<locals>.makeOptionalList)	r}rv�tupler4rr�r��
ValueErrorr)r�r+�minElements�optElementsr�ryr=rz�__mul__UsD







zParserElement.__mul__cCs
|�|�Sr�)rCr7ryryrz�__rmul__�szParserElement.__rmul__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zI
        Implementation of | operator - returns C{L{MatchFirst}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__or__�s


�zParserElement.__or__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||BS)z]
        Implementation of | operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__ror__�s


�zParserElement.__ror__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zA
        Implementation of ^ operator - returns C{L{Or}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__xor__�s


�zParserElement.__xor__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||AS)z]
        Implementation of ^ operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rxor__�s


�zParserElement.__rxor__cCsFt|t�rt�|�}t|t�s:tjdt|�tdd�dSt||g�S)zC
        Implementation of & operator - returns C{L{Each}}
        r1rsr2N)	r}r�r&r�r4r5r�r6rr7ryryrz�__and__�s


�zParserElement.__and__cCsBt|t�rt�|�}t|t�s:tjdt|�tdd�dS||@S)z]
        Implementation of & operator when left operand is not a C{L{ParserElement}}
        r1rsr2Nr7r7ryryrz�__rand__�s


�zParserElement.__rand__cCst|�S)zE
        Implementation of ~ operator - returns C{L{NotAny}}
        )rr�ryryrz�
__invert__�szParserElement.__invert__cCs|dur|�|�S|��SdS)a

        Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
        
        If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
        passed as C{True}.
           
        If C{name} is omitted, same as calling C{L{copy}}.

        Example::
            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")             
        N)r�r�r)ryryrz�__call__�s
zParserElement.__call__cCst|�S)z�
        Suppresses the output of this C{ParserElement}; useful to keep punctuation from
        cluttering up returned output.
        )r-r�ryryrz�suppress�szParserElement.suppresscCs
d|_|S)a
        Disables the skipping of whitespace before matching the characters in the
        C{ParserElement}'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        F�r�r�ryryrz�leaveWhitespaceszParserElement.leaveWhitespacecCsd|_||_d|_|S)z8
        Overrides the default whitespace chars
        TF)r�r�r�)r�r�ryryrz�setWhitespaceChars
sz ParserElement.setWhitespaceCharscCs
d|_|S)z�
        Overrides default behavior to expand C{<TAB>}s to spaces before parsing the input string.
        Must be called before C{parseString} when the input grammar contains elements that
        match C{<TAB>} characters.
        T)r�r�ryryrz�
parseWithTabsszParserElement.parseWithTabscCsLt|t�rt|�}t|t�r4||jvrH|j�|�n|j�t|����|S)a�
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.
        
        Example::
            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
            
            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        )r}r�r-r�r$r�r7ryryrz�ignores


zParserElement.ignorecCs"|pt|pt|ptf|_d|_|S)zT
        Enable display of debugging messages while doing pattern matching.
        T)rwr{r}r�r�)r��startAction�
successAction�exceptionActionryryrz�setDebugActions6s�zParserElement.setDebugActionscCs|r|�ttt�nd|_|S)a�
        Enable display of debugging messages while doing pattern matching.
        Set C{flag} to True to enable, False to disable.

        Example::
            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer
            
            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")
        
        prints::
            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using L{setDebugActions}. Prior to attempting
        to match the C{wd} expression, the debugging message C{"Match <exprname> at loc <n>(<line>,<col>)"}
        is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
        message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
        F)rVrwr{r}r�)r��flagryryrz�setDebug@s#zParserElement.setDebugcCs|jSr�)r�r�ryryrzr�iszParserElement.__str__cCst|�Sr�r�r�ryryrzr�lszParserElement.__repr__cCsd|_d|_|Sr�)r�r�r�ryryrzroszParserElement.streamlinecCsdSr�ryr�ryryrz�checkRecursiontszParserElement.checkRecursioncCs|�g�dS)zj
        Check defined expressions for valid structure, check for infinite recursive definitions.
        N)rY)r��
validateTraceryryrz�validatewszParserElement.validatecCs�z|��}WnDtyPt|d��}|��}Wd�n1sB0YYn0z|�||�WSty�}ztjrz�n|�WYd}~n
d}~00dS)z�
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        �rN)�readr��openrrr&r)r��file_or_filenamer�
file_contents�fr|ryryrz�	parseFile}s,zParserElement.parseFilecsHt|t�r"||up t|�t|�kSt|t�r6|�|�Stt|�|kSdSr�)r}r&�varsr�r%�superr7�r�ryrz�__eq__�s



zParserElement.__eq__cCs
||kSr�ryr7ryryrz�__ne__�szParserElement.__ne__cCstt|��Sr�)�hash�idr�ryryrz�__hash__�szParserElement.__hash__cCs||kSr�ryr7ryryrz�__req__�szParserElement.__req__cCs
||kSr�ryr7ryryrz�__rne__�szParserElement.__rne__cCs2z|jt|�|d�WdSty,YdS0dS)a�
        Method for quick testing of a parser against a test string. Good for simple 
        inline microtests of sub expressions while building up larger parser.
           
        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
            
        Example::
            expr = Word(nums)
            assert expr.matches("100")
        �rTFN)rr�r)r��
testStringrryryrzr%�s

zParserElement.matches�#cCs�t|t�r"tttj|������}t|t�r4t|�}g}g}d}	|D�]�}
|dur^|�	|
d�sf|rr|
sr|�
|
�qD|
sxqDd�|�|
g}g}z:|
�dd�}
|j
|
|d�}|�
|j|d��|	o�|}	Wn�t�yt}
z�t|
t�r�dnd	}d|
v�r(|�
t|
j|
��|�
d
t|
j|
�dd|�n|�
d
|
jd|�|�
d
t|
��|	�oZ|}	|
}WYd}
~
nNd}
~
0t�y�}z,|�
dt|��|	�o�|}	|}WYd}~n
d}~00|�r�|�r�|�
d	�td�|��|�
|
|f�qD|	|fS)a3
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.
           
        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests           
         - comment - (default=C{'#'}) - expression for indicating embedded comments in the test 
              string; pass None to disable comment filtering
         - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default=C{True}) prints test output to stdout
         - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if C{failureTests} is True), and the results contain a list of lines of each 
        test's output
        
        Example::
            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")
        prints::
            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success
            
            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\n of strings that spans \n 3 lines")
        
        (Note that this is a raw string literal, you must include the leading 'r'.)
        TNFrH�\nrm)rdz(FATAL)r�� r��^zFAIL: zFAIL-EXCEPTION: )r}r�r�r�rr��rstrip�
splitlinesrr%r$r�r�rrarr#rIr�r;r�rt)r��testsr�comment�fullDump�printResults�failureTests�
allResults�comments�successrxr?�resultr�r�r|ryryrz�runTests�sNW




$


zParserElement.runTests)F)F)T)T)TT)TT)r)F)N)T)F)T)TroTTF)Or�r�r�r�r�r�staticmethodr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rr	rrrrrr�rrrr�_MAX_INTr(r�r,r�r,r8r9r:rCrDrErFrGrHrIrJrKrLrMrOrPrQrRrVrXr�r�rrYr[rbrfrgrjrkrlr%r~�
__classcell__ryryrerzr&Os�




&




G

"
2G+D
			

)

r&cs eZdZdZ�fdd�Z�ZS)r.zT
    Abstract C{ParserElement} subclass, for defining atomic matching patterns.
    cstt|�jdd�dS�NF�r�)rdr.r�r�reryrzr�@	szToken.__init__�r�r�r�r�r�r�ryryrerzr.<	sr.cs eZdZdZ�fdd�Z�ZS)rz,
    An empty token, will always match.
    cs$tt|���d|_d|_d|_dS)NrTF)rdrr�r�r�r�r�reryrzr�H	szEmpty.__init__r�ryryrerzrD	srcs*eZdZdZ�fdd�Zddd�Z�ZS)rz(
    A token that will never match.
    cs*tt|���d|_d|_d|_d|_dS)NrTFzUnmatchable token)rdrr�r�r�r�r�r�reryrzr�S	s
zNoMatch.__init__TcCst|||j|��dSr�)r!r�r�ryryrzr�Z	szNoMatch.parseImpl)T�r�r�r�r�r�r�r�ryryrerzrO	srcs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Token to exactly match a specified string.
    
    Example::
        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"
    
    For case-insensitive matching, use L{CaselessLiteral}.
    
    For keyword matching (force word break before and after the matched string),
    use L{Keyword} or L{CaselessKeyword}.
    cs�tt|���||_t|�|_z|d|_Wn(tyTtj	dt
dd�t|_Yn0dt
|j�|_d|j|_d|_d|_dS)Nrz2null string passed to Literal; use Empty() insteadrsr2�"%s"r�F)rdrr��matchr��matchLen�firstMatchCharr�r4r5r6rr�r�r�r�r�r��r��matchStringreryrzr�l	s
�zLiteral.__init__TcCsJ|||jkr6|jdks&|�|j|�r6||j|jfSt|||j|��dSr�)r�r��
startswithr�r!r�r�ryryrzr�	s��zLiteral.parseImpl)Tr�ryryrerzr^	s
rcsLeZdZdZedZd�fdd�	Zddd	�Z�fd
d�Ze	dd
��Z
�ZS)ra\
    Token to exactly match a specified string as a keyword, that is, it must be
    immediately followed by a non-keyword character.  Compare with C{L{Literal}}:
     - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
     - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
    Accepts two optional constructor arguments in addition to the keyword string:
     - C{identChars} is a string of characters that would be valid identifier characters,
          defaulting to all alphanumerics + "_" and "$"
     - C{caseless} allows case-insensitive matching, default is C{False}.
       
    Example::
        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use L{CaselessKeyword}.
    �_$NFcs�tt|���|durtj}||_t|�|_z|d|_Wn"ty\t	j
dtdd�Yn0d|j|_d|j|_
d|_d|_||_|r�|��|_|��}t|�|_dS)Nrz2null string passed to Keyword; use Empty() insteadrsr2r�r�F)rdrr��DEFAULT_KEYWORD_CHARSr�r�r�r�r�r4r5r6r�r�r�r��caseless�upper�
caselessmatchr��
identChars)r�r�r�r�reryrzr��	s(
�
zKeyword.__init__TcCs|jr|||||j���|jkr�|t|�|jksL|||j��|jvr�|dksj||d��|jvr�||j|jfSnv|||jkr�|jdks�|�|j|�r�|t|�|jks�|||j|jvr�|dks�||d|jvr�||j|jfSt	|||j
|��dSr�)r�r�r�r�r�r�r�r�r�r!r�r�ryryrzr��	s4����������zKeyword.parseImplcstt|���}tj|_|Sr�)rdrr�r�r�)r�r�reryrzr��	szKeyword.copycCs
|t_dS)z,Overrides the default Keyword chars
        N)rr�r�ryryrz�setDefaultKeywordChars�	szKeyword.setDefaultKeywordChars)NF)T)r�r�r�r�r5r�r�r�r�rr�r�ryryrerzr�	s
rcs*eZdZdZ�fdd�Zddd�Z�ZS)r
al
    Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::
        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
        
    (Contrast with example for L{CaselessKeyword}.)
    cs6tt|��|���||_d|j|_d|j|_dS)Nz'%s'r�)rdr
r�r��returnStringr�r�r�reryrzr��	szCaselessLiteral.__init__TcCs@||||j���|jkr,||j|jfSt|||j|��dSr�)r�r�r�r�r!r�r�ryryrzr��	szCaselessLiteral.parseImpl)Tr�ryryrerzr
�	s
r
cs,eZdZdZd�fdd�	Zd	dd�Z�ZS)
r	z�
    Caseless version of L{Keyword}.

    Example::
        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
        
    (Contrast with example for L{CaselessLiteral}.)
    Ncstt|�j||dd�dS)NT�r�)rdr	r�)r�r�r�reryrzr��	szCaselessKeyword.__init__TcCsj||||j���|jkrV|t|�|jksF|||j��|jvrV||j|jfSt|||j|��dSr�)r�r�r�r�r�r�r!r�r�ryryrzr��	s��zCaselessKeyword.parseImpl)N)Tr�ryryrerzr	�	sr	cs,eZdZdZd�fdd�	Zd	dd�Z�ZS)
rnax
    A variation on L{Literal} which matches "close" matches, that is, 
    strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
     - C{match_string} - string to be matched
     - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
    
    The results from a successful parse will contain the matched text from the input string and the following named results:
     - C{mismatches} - a list of the positions within the match_string where mismatches were found
     - C{original} - the original match_string used to compare against the input string
    
    If C{mismatches} is an empty list, then the match was an exact match.
    
    Example::
        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    r�csBtt|���||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F)	rdrnr�r��match_string�
maxMismatchesr�r�r�)r�r�r�reryrzr�

szCloseMatch.__init__TcCs�|}t|�}|t|j�}||kr�|j}d}g}	|j}
tt|||�|j��D]2\}}|\}}
||
krN|	�|�t|	�|
krNq�qN|d}t|||�g�}|j|d<|	|d<||fSt|||j|��dS)Nrr��original�
mismatches)	r�r�r�r�r�r$r$r!r�)r�rur�r��startr��maxlocr��match_stringlocr�r��s_m�src�mat�resultsryryrzr�
s( 

zCloseMatch.parseImpl)r�)Tr�ryryrerzrn�	s	rncs8eZdZdZd
�fdd�	Zdd	d
�Z�fdd�Z�ZS)r1a	
    Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters,
    an optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction. An optional
    C{excludeChars} parameter can list characters that might be found in 
    the input C{bodyChars} string; useful to define a word of all printables
    except for one or two characters, for instance.
    
    L{srange} is useful for defining custom character set strings for defining 
    C{Word} expressions, using range notation from regular expression character sets.
    
    A common mistake is to use C{Word} to match a specific literal string, as in 
    C{Word("Address")}. Remember that C{Word} uses the string argument to define
    I{sets} of matchable characters. This expression would match "Add", "AAA",
    "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
    To match an exact literal string, use L{Literal} or L{Keyword}.

    pyparsing includes helper strings for building Words:
     - L{alphas}
     - L{nums}
     - L{alphanums}
     - L{hexnums}
     - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
     - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - L{printables} (any non-whitespace character)

    Example::
        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
        
        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums+'-')
        
        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")
        
        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    Nr�rFcs�tt|����rFd��fdd�|D��}|rFd��fdd�|D��}||_t|�|_|rl||_t|�|_n||_t|�|_|dk|_	|dkr�t
d��||_|dkr�||_nt
|_|dkr�||_||_t|�|_d|j|_d	|_||_d
|j|jv�r�|dk�r�|dk�r�|dk�r�|j|jk�r8dt|j�|_nHt|j�dk�rfdt�|j�t|j�f|_nd
t|j�t|j�f|_|j�r�d|jd|_zt�|j�|_Wnt�y�d|_Yn0dS)Nr�c3s|]}|�vr|VqdSr�ryr���excludeCharsryrzr�`
r|z Word.__init__.<locals>.<genexpr>c3s|]}|�vr|VqdSr�ryr�r�ryrzr�b
r|rr�zZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedr�Frqz[%s]+z%s[%s]*z	[%s][%s]*z\b)rdr1r�r��
initCharsOrigr��	initChars�
bodyCharsOrig�	bodyChars�maxSpecifiedr@�minLen�maxLenr�r�r�r�r��	asKeyword�_escapeRegexRangeChars�reStringr�r��escape�compiler�)r�r�r��min�max�exactr�r�rer�rzr�]
s\



0
����z
Word.__init__Tc
Cs>|jr<|j�||�}|s(t|||j|��|��}||��fS|||jvrZt|||j|��|}|d7}t|�}|j}||j	}t
||�}||kr�|||vr�|d7}q�d}	|||jkr�d}	|jr�||kr�|||vr�d}	|j
�r|dkr�||d|v�s||k�r|||v�rd}	|	�r.t|||j|��||||�fS)Nr�FTr)r�r�r!r��end�groupr�r�r�r�r�r�r�r�)
r�rur�r�r}r�r��	bodycharsr��throwExceptionryryrzr��
s6


2zWord.parseImplcstztt|���WSty"Yn0|jdurndd�}|j|jkr^d||j�||j�f|_nd||j�|_|jS)NcSs$t|�dkr|dd�dS|SdS)N��...r��r�ryryrz�
charsAsStr�
sz Word.__str__.<locals>.charsAsStrz	W:(%s,%s)zW:(%s))rdr1r�r�r�r�r�)r�r�reryrzr��
s
zWord.__str__)Nr�rrFN)T�r�r�r�r�r�r�r�r�ryryrerzr1.
s.6
#r1csFeZdZdZee�d��Zd�fdd�	Zddd�Z	�fd	d
�Z
�ZS)
r)a�
    Token for matching strings that match a given regular expression.
    Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
    If the given regex contains named groups (defined using C{(?P<name>...)}), these will be preserved as 
    named parse results.

    Example::
        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    z[A-Z]rcs�tt|���t|t�r�|s,tjdtdd�||_||_	zt
�|j|j	�|_
|j|_Wq�t
jy�tjd|tdd��Yq�0n2t|tj�r�||_
t|�|_|_||_	ntd��t|�|_d|j|_d|_d|_d	S)
z�The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags.z0null string passed to Regex; use Empty() insteadrsr2�$invalid pattern (%s) passed to RegexzCRegex may only be constructed with a string or a compiled RE objectr�FTN)rdr)r�r}r�r4r5r6�pattern�flagsr�r�r��
sre_constants�error�compiledREtyperr@r�r�r�r�r�)r�r�r�reryrzr��
s6
�
�
�
zRegex.__init__TcCs`|j�||�}|s"t|||j|��|��}|��}t|���}|rX|D]}||||<qF||fSr�)r�r�r!r�r��	groupdictr$r�)r�rur�r�r}�dr�r�ryryrzr��
szRegex.parseImplcsDztt|���WSty"Yn0|jdur>dt|j�|_|jS)NzRe:(%s))rdr)r�r�r�r�r�r�reryrzr�
s
z
Regex.__str__)r)T)r�r�r�r�r�r�r�r�r�r�r�r�ryryrerzr)�
s
"

r)cs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
r'a�
    Token for matching strings that are delimited by quoting characters.
    
    Defined with the following parameters:
        - quoteChar - string of one or more characters defining the quote delimiting string
        - escChar - character to escape quotes, typically backslash (default=C{None})
        - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
        - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
        - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
        - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})

    Example::
        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
    prints::
        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    NFTc
sLtt����|��}|s0tjdtdd�t��|dur>|}n"|��}|s`tjdtdd�t��|�_t	|��_
|d�_|�_t	|��_
|�_|�_|�_|�_|r�tjtjB�_dt��j�t�jd�|dur�t|�p�df�_n<d�_dt��j�t�jd�|du�rt|��pdf�_t	�j�d	k�rp�jd
d��fdd
�tt	�j�d	dd�D��d7_|�r��jdt�|�7_|�r��jdt�|�7_t��j�d�_�jdt��j�7_zt��j�j��_�j�_Wn.tj�y$tjd�jtdd��Yn0t ���_!d�j!�_"d�_#d�_$dS)Nz$quoteChar cannot be the empty stringrsr2z'endQuoteChar cannot be the empty stringrz%s(?:[^%s%s]r�z%s(?:[^%s\n\r%s]r�z|(?:z)|(?:c3s4|],}dt��jd|��t�j|�fVqdS)z%s[^%s]N)r�r��endQuoteCharr�r;r�ryrzr�Xs��z(QuotedString.__init__.<locals>.<genexpr>rt�)z|(?:%s)z|(?:%s.)z(.)z)*%sr�r�FT)%rdr'r�r�r4r5r6�SyntaxError�	quoteCharr��quoteCharLen�firstQuoteCharr��endQuoteCharLen�escChar�escQuote�unquoteResults�convertWhitespaceEscapesr��	MULTILINE�DOTALLr�r�r�r�r�r��escCharReplacePatternr�r�r�r�r�r�r�r�r�)r�r�r�r��	multiliner�r�r�rer�rzr�/sz



��
������
zQuotedString.__init__c	Cs�|||jkr|j�||�pd}|s4t|||j|��|��}|��}|jr�||j|j	�}t
|t�r�d|vr�|jr�ddddd�}|�
�D]\}}|�||�}q�|jr�t�|jd|�}|jr�|�|j|j�}||fS)N�\�	rH��
)�\trpz\fz\rz\g<1>)r�r�r�r!r�r�r�r�r�r�r}r�r�r�r�r�r�r�r�r�)	r�rur�r�r}r��ws_map�wslit�wscharryryrzr�ps* 
�zQuotedString.parseImplcsFztt|���WSty"Yn0|jdur@d|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)rdr'r�r�r�r�r�r�reryrzr��s
zQuotedString.__str__)NNFTNT)Tr�ryryrerzr'sA
#r'cs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
ra�
    Token for matching words composed of characters I{not} in a given set (will
    include whitespace in matched characters if not listed in the provided exclusion set - see example).
    Defined with string containing all disallowed characters, and an optional
    minimum, maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction.

    Example::
        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
    prints::
        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    r�rcs�tt|���d|_||_|dkr*td��||_|dkr@||_nt|_|dkrZ||_||_t	|�|_
d|j
|_|jdk|_d|_
dS)NFr�zfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr�)rdrr�r��notCharsr@r�r�r�r�r�r�r�r�)r�r�r�r�r�reryrzr��s 
zCharsNotIn.__init__TcCs�|||jvrt|||j|��|}|d7}|j}t||jt|��}||krb|||vrb|d7}qD|||jkr�t|||j|��||||�fSr�)r�r!r�r�r�r�r�)r�rur�r�r��notchars�maxlenryryrzr��s
�
zCharsNotIn.parseImplcsdztt|���WSty"Yn0|jdur^t|j�dkrRd|jdd�|_nd|j|_|jS)Nr�z
!W:(%s...)z!W:(%s))rdrr�r�r�r�r�r�reryrzr��s
zCharsNotIn.__str__)r�rr)Tr�ryryrerzr�s
rcs<eZdZdZdddddd�Zd�fdd�	Zddd�Z�ZS)r0a�
    Special matching class for matching whitespace.  Normally, whitespace is ignored
    by pyparsing grammars.  This class is included when some whitespace structures
    are significant.  Define with a string containing the whitespace characters to be
    matched; default is C{" \t\r\n"}.  Also takes optional C{min}, C{max}, and C{exact} arguments,
    as defined for the C{L{Word}} class.
    z<SPC>z<TAB>z<LF>z<CR>z<FF>)rqr�rHr�r�� 	
r�rcs�tt����|�_��d��fdd��jD���d�dd��jD���_d�_d�j�_	|�_
|dkrt|�_nt�_|dkr�|�_|�_
dS)Nr�c3s|]}|�jvr|VqdSr�)�
matchWhiter�r�ryrzr��r|z!White.__init__.<locals>.<genexpr>css|]}tj|VqdSr�)r0�	whiteStrsr�ryryrzr��r|Tr�r)
rdr0r�r�rPr�r�r�r�r�r�r�r�)r��wsr�r�r�rer�rzr��s zWhite.__init__TcCs�|||jvrt|||j|��|}|d7}||j}t|t|��}||krb|||jvrb|d7}qB|||jkr�t|||j|��||||�fSr�)r�r!r�r�r�r�r�)r�rur�r�r�r�ryryrzr�	s

zWhite.parseImpl)r�r�rr)T)r�r�r�r�r�r�r�r�ryryrerzr0�s�r0cseZdZ�fdd�Z�ZS)�_PositionTokencs(tt|���|jj|_d|_d|_dSr�)rdr�r�r�r�r�r�r�r�reryrzr�s
z_PositionToken.__init__�r�r�r�r�r�ryryrerzr�sr�cs2eZdZdZ�fdd�Zdd�Zd	dd�Z�ZS)
rzb
    Token to advance to a specific column of input text; useful for tabular report scraping.
    cstt|���||_dSr�)rdrr�r;)r��colnoreryrzr�$szGoToColumn.__init__cCs\t||�|jkrXt|�}|jr*|�||�}||krX||��rXt||�|jkrX|d7}q*|Sr�)r;r�r�r��isspace)r�rur�r�ryryrzr�(s$
zGoToColumn.preParseTcCsDt||�}||jkr"t||d|��||j|}|||�}||fS)NzText not in expected column�r;r!)r�rur�r��thiscol�newlocr�ryryrzr�1s

zGoToColumn.parseImpl)T)r�r�r�r�r�r�r�r�ryryrerzr s	rcs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Matches if current position is at the beginning of a line within the parse string
    
    Example::
    
        test = '''        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)
    
    Prints::
        ['AAA', ' this line']
        ['AAA', ' and this line']    

    cstt|���d|_dS)NzExpected start of line)rdrr�r�r�reryrzr�OszLineStart.__init__TcCs*t||�dkr|gfSt|||j|��dSr�)r;r!r�r�ryryrzr�SszLineStart.parseImpl)Tr�ryryrerzr:srcs*eZdZdZ�fdd�Zddd�Z�ZS)rzU
    Matches if current position is at the end of a line within the parse string
    cs,tt|���|�tj�dd��d|_dS)NrHr�zExpected end of line)rdrr�rPr&r�r�r�r�reryrzr�\szLineEnd.__init__TcCsb|t|�kr6||dkr$|ddfSt|||j|��n(|t|�krN|dgfSt|||j|��dS)NrHr��r�r!r�r�ryryrzr�aszLineEnd.parseImpl)Tr�ryryrerzrXsrcs*eZdZdZ�fdd�Zddd�Z�ZS)r,zM
    Matches if current position is at the beginning of the parse string
    cstt|���d|_dS)NzExpected start of text)rdr,r�r�r�reryrzr�pszStringStart.__init__TcCs0|dkr(||�|d�kr(t|||j|��|gfSr�)r�r!r�r�ryryrzr�tszStringStart.parseImpl)Tr�ryryrerzr,lsr,cs*eZdZdZ�fdd�Zddd�Z�ZS)r+zG
    Matches if current position is at the end of the parse string
    cstt|���d|_dS)NzExpected end of text)rdr+r�r�r�reryrzr�szStringEnd.__init__TcCs^|t|�krt|||j|��n<|t|�kr6|dgfS|t|�krJ|gfSt|||j|��dSr�r�r�ryryrzr��szStringEnd.parseImpl)Tr�ryryrerzr+{sr+cs.eZdZdZef�fdd�	Zddd�Z�ZS)r3ap
    Matches if the current position is at the beginning of a Word, and
    is not preceded by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
    the string being parsed, or at the beginning of a line.
    cs"tt|���t|�|_d|_dS)NzNot at the start of a word)rdr3r�r��	wordCharsr��r�r�reryrzr��s
zWordStart.__init__TcCs@|dkr8||d|jvs(|||jvr8t|||j|��|gfSr�)r�r!r�r�ryryrzr��s�zWordStart.parseImpl)T�r�r�r�r�rXr�r�r�ryryrerzr3�sr3cs.eZdZdZef�fdd�	Zddd�Z�ZS)r2aZ
    Matches if the current position is at the end of a Word, and
    is not followed by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{} behavior of regular expressions,
    use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
    the string being parsed, or at the end of a line.
    cs(tt|���t|�|_d|_d|_dS)NFzNot at the end of a word)rdr2r�r�r�r�r�r�reryrzr��s
zWordEnd.__init__TcCsPt|�}|dkrH||krH|||jvs8||d|jvrHt|||j|��|gfSr�)r�r�r!r�)r�rur�r�r�ryryrzr��s�zWordEnd.parseImpl)Tr�ryryrerzr2�sr2cs�eZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Z�fdd�Z�fd
d�Z	�fdd�Z
d�fdd�	Zgfdd�Z�fdd�Z
�ZS)r"z^
    Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
    Fcs�tt|��|�t|t�r"t|�}t|t�r<t�|�g|_	nft|t
�rxt|�}tdd�|D��rlttj|�}t|�|_	n*zt|�|_	Wnt
y�|g|_	Yn0d|_dS)Ncss|]}t|t�VqdSr�)r}r�)r�rvryryrzr��r|z+ParseExpression.__init__.<locals>.<genexpr>F)rdr"r�r}r�r�r�r&r��exprsr�allr�r�r��r�r�r�reryrzr��s


zParseExpression.__init__cCs
|j|Sr�)r�r�ryryrzr��szParseExpression.__getitem__cCs|j�|�d|_|Sr�)r�r$r�r7ryryrzr$�szParseExpression.appendcCs0d|_dd�|jD�|_|jD]}|��q|S)z~Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
           all contained expressions.FcSsg|]}|���qSryr*�r�r�ryryrzr��r|z3ParseExpression.leaveWhitespace.<locals>.<listcomp>)r�r�rO)r�r�ryryrzrO�s


zParseExpression.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jD]}|�|jd�q*n,tt|��|�|jD]}|�|jd�qX|Sr)r}r-r�rdr"rRr�)r�r+r�reryrzrR�s



zParseExpression.ignorecsLztt|���WSty"Yn0|jdurFd|jjt|j�f|_|jS�Nz%s:(%s))	rdr"r�r�r�r�r�r�r�r�reryrzr��s
zParseExpression.__str__cs*tt|���|jD]}|��qt|j�dk�r|jd}t||j�r�|js�|jdur�|j	s�|jdd�|jdg|_d|_
|j|jO_|j|jO_|jd}t||j��r|j�s|jdu�r|j	�s|jdd�|jdd�|_d|_
|j|jO_|j|jO_dt
|�|_|S)Nrsrr�rtr�)rdr"rr�r�r}r�r�r�r�r�r�r�r�r�)r�r�r+reryrzr�s<


���
���zParseExpression.streamlinecstt|��||�}|Sr�)rdr"r�)r�r�r�r�reryrzr�
szParseExpression.setResultsNamecCs6|dd�|g}|jD]}|�|�q|�g�dSr�)r�r[rY)r�rZ�tmpr�ryryrzr[
s
zParseExpression.validatecs$tt|���}dd�|jD�|_|S)NcSsg|]}|���qSryr*r�ryryrzr�%
r|z(ParseExpression.copy.<locals>.<listcomp>)rdr"r�r�rGreryrzr�#
szParseExpression.copy)F)F)r�r�r�r�r�r�r$rOrRr�rr�r[r�r�ryryrerzr"�s	
"r"csTeZdZdZGdd�de�Zd�fdd�	Zddd�Zd	d
�Zdd�Z	d
d�Z
�ZS)ra

    Requires all given C{ParseExpression}s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'+'} operator.
    May also be constructed using the C{'-'} operator, which will suppress backtracking.

    Example::
        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"),name_expr("name"),integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    cseZdZ�fdd�Z�ZS)zAnd._ErrorStopcs*ttj|�j|i|��d|_|��dS)N�-)rdrr8r�r�rOrgreryrzr�9
szAnd._ErrorStop.__init__r�ryryrerzr88
sr8TcsRtt|��||�tdd�|jD��|_|�|jdj�|jdj|_d|_	dS)Ncss|]}|jVqdSr��r�r�ryryrzr�@
r|zAnd.__init__.<locals>.<genexpr>rT)
rdrr�r�r�r�rPr�r�r�r�reryrzr�>
s
zAnd.__init__c	Cs�|jdj|||dd�\}}d}|jdd�D]�}t|tj�rDd}q.|r�z|�|||�\}}Wq�tyr�Yq�ty�}zd|_t�|��WYd}~q�d}~0t	y�t|t
|�|j|��Yq�0n|�|||�\}}|s�|��r.||7}q.||fS)NrFr r�T)
r�r�r}rr8r%r�
__traceback__r�r�r�r�r)	r�rur�r��
resultlist�	errorStopr��
exprtokensr�ryryrzr�E
s(
z
And.parseImplcCst|t�rt�|�}|�|�Sr��r}r�r&r�r$r7ryryrzr6^
s

zAnd.__iadd__cCs6|dd�|g}|jD]}|�|�|jsq2qdSr�)r�rYr��r�r��subRecCheckListr�ryryrzrYc
s


zAnd.checkRecursioncCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr��{rqcss|]}t|�VqdSr�r�r�ryryrzr�o
r|zAnd.__str__.<locals>.<genexpr>�}�rr�r�r�r�r�ryryrzr�j
s


 zAnd.__str__)T)T)r�r�r�r�rr8r�r�r6rYr�r�ryryrerzr(
s
rcsDeZdZdZd�fdd�	Zddd�Zdd	�Zd
d�Zdd
�Z�Z	S)ra�
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the expression that matches the longest string will be used.
    May be constructed using the C{'^'} operator.

    Example::
        # construct Or using '^' operator
        
        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))
    prints::
        [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr��
r|zOr.__init__.<locals>.<genexpr>T)rdrr�r�rbr�r�reryrzr��
szOr.__init__TcCsTd}d}g}|jD]�}z|�||�}Wnvtyd}	z&d|	_|	j|krP|	}|	j}WYd}	~	qd}	~	0ty�t|�|kr�t|t|�|j|�}t|�}Yq0|�||f�q|�r*|j	dd�d�|D]`\}
}z|�
|||�WSt�y&}	z(d|	_|	j|k�r|	}|	j}WYd}	~	q�d}	~	00q�|du�rB|j|_|�nt||d|��dS)NrtcSs
|dSr�ry)�xryryrzr{�
r|zOr.parseImpl.<locals>.<lambda>)r � no defined alternatives to match)r�r�r!r�r�r�r�r�r$�sortr�r�)r�rur�r��	maxExcLoc�maxExceptionr%r��loc2r��_ryryrzr��
s<


zOr.parseImplcCst|t�rt�|�}|�|�Sr�r�r7ryryrz�__ixor__�
s

zOr.__ixor__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�z ^ css|]}t|�VqdSr�r�r�ryryrzr��
r|zOr.__str__.<locals>.<genexpr>r�r�r�ryryrzr��
s


 z
Or.__str__cCs,|dd�|g}|jD]}|�|�qdSr��r�rYr�ryryrzrY�
s
zOr.checkRecursion)F)T)
r�r�r�r�r�r�rr�rYr�ryryrerzrt
s

&	rcsDeZdZdZd�fdd�	Zddd�Zdd	�Zd
d�Zdd
�Z�Z	S)ra�
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the first one listed is the one that will match.
    May be constructed using the C{'|'} operator.

    Example::
        # construct MatchFirst using '|' operator
        
        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr��
r|z&MatchFirst.__init__.<locals>.<genexpr>T)rdrr�r�rbr�r�reryrzr��
szMatchFirst.__init__Tc	Cs�d}d}|jD]�}z|�|||�}|WStyb}z |j|krN|}|j}WYd}~qd}~0ty�t|�|kr�t|t|�|j|�}t|�}Yq0q|dur�|j|_|�nt||d|��dS)Nrtr�)r�r�r!r�r�r�r�r�)	r�rur�r�rrr�r�r�ryryrzr��
s$


zMatchFirst.parseImplcCst|t�rt�|�}|�|�Sr�r�r7ryryrz�__ior__�
s

zMatchFirst.__ior__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�� | css|]}t|�VqdSr�r�r�ryryrzr��
r|z%MatchFirst.__str__.<locals>.<genexpr>r�r�r�ryryrzr��
s


 zMatchFirst.__str__cCs,|dd�|g}|jD]}|�|�qdSr�rr�ryryrzrYs
zMatchFirst.checkRecursion)F)T)
r�r�r�r�r�r�rr�rYr�ryryrerzr�
s
	rcs<eZdZdZd�fdd�	Zddd�Zdd�Zd	d
�Z�ZS)
ram
    Requires all given C{ParseExpression}s to be found, but in any order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'&'} operator.

    Example::
        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order 
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )
    prints::
        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    Tcs8tt|��||�tdd�|jD��|_d|_d|_dS)Ncss|]}|jVqdSr�r�r�ryryrzr�?r|z Each.__init__.<locals>.<genexpr>T)rdrr�r�r�r�r��initExprGroupsr�reryrzr�=sz
Each.__init__c	s�|jr�tdd�|jD��|_dd�|jD�}dd�|jD�}|||_dd�|jD�|_dd�|jD�|_dd�|jD�|_|j|j7_d	|_|}|jdd�}|jdd��g}d
}	|	�rh|�|j|j}
g}|
D]t}z|�||�}Wnt	�y|�
|�Yq�0|�
|j�t|�|��||v�r>|�
|�q�|�vr܈�
|�q�t|�t|
�kr�d	}	q�|�r�d�dd�|D��}
t	||d
|
��|�fdd�|jD�7}g}|D]"}|�|||�\}}|�
|��q�t|tg��}||fS)Ncss&|]}t|t�rt|j�|fVqdSr�)r}rrirvr�ryryrzr�Er|z!Each.parseImpl.<locals>.<genexpr>cSsg|]}t|t�r|j�qSry�r}rrvr�ryryrzr�Fr|z"Each.parseImpl.<locals>.<listcomp>cSs g|]}|jrt|t�s|�qSry)r�r}rr�ryryrzr�Gr|cSsg|]}t|t�r|j�qSry)r}r4rvr�ryryrzr�Ir|cSsg|]}t|t�r|j�qSry)r}rrvr�ryryrzr�Jr|cSs g|]}t|tttf�s|�qSry)r}rr4rr�ryryrzr�Kr|FTr:css|]}t|�VqdSr�r�r�ryryrzr�fr|z*Missing one or more required elements (%s)cs$g|]}t|t�r|j�vr|�qSryr
r���tmpOptryrzr�jr|)r	r�r��opt1map�	optionals�multioptionals�
multirequired�requiredr�r!r$r�ri�remover�r�r��sumr$)r�rur�r��opt1�opt2�tmpLoc�tmpReqd�
matchOrder�keepMatching�tmpExprs�failedr��missingr�r��finalResultsryrrzr�CsP

zEach.parseImplcCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr�r�z & css|]}t|�VqdSr�r�r�ryryrzr�yr|zEach.__str__.<locals>.<genexpr>r�r�r�ryryrzr�ts


 zEach.__str__cCs,|dd�|g}|jD]}|�|�qdSr�rr�ryryrzrY}s
zEach.checkRecursion)T)T)	r�r�r�r�r�r�r�rYr�ryryrerzrs
5
1	rcsleZdZdZd�fdd�	Zddd�Zdd	�Z�fd
d�Z�fdd
�Zdd�Z	gfdd�Z
�fdd�Z�ZS)r za
    Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
    Fcs�tt|��|�t|t�r@ttjt�r2t�|�}nt�t	|��}||_
d|_|dur�|j|_|j
|_
|�|j�|j|_|j|_|j|_|j�|j�dSr�)rdr r�r}r��
issubclassr&r�r.rrvr�r�r�rPr�r�r�r�r�r&�r�rvr�reryrzr��s
zParseElementEnhance.__init__TcCs2|jdur|jj|||dd�Std||j|��dS)NFr r�)rvr�r!r�r�ryryrzr��s
zParseElementEnhance.parseImplcCs*d|_|j��|_|jdur&|j��|Sr)r�rvr�rOr�ryryrzrO�s


z#ParseElementEnhance.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jdurn|j�|jd�n,tt|��|�|jdurn|j�|jd�|Sr)r}r-r�rdr rRrvr7reryrzrR�s



zParseElementEnhance.ignorecs&tt|���|jdur"|j��|Sr�)rdr rrvr�reryrzr�s

zParseElementEnhance.streamlinecCsB||vrt||g��|dd�|g}|jdur>|j�|�dSr�)r(rvrY)r�r�r�ryryrzrY�s

z"ParseElementEnhance.checkRecursioncCs6|dd�|g}|jdur(|j�|�|�g�dSr��rvr[rY�r�rZr�ryryrzr[�s
zParseElementEnhance.validatecsVztt|���WSty"Yn0|jdurP|jdurPd|jjt|j�f|_|jSr�)	rdr r�r�r�rvr�r�r�r�reryrzr��szParseElementEnhance.__str__)F)T)
r�r�r�r�r�r�rOrRrrYr[r�r�ryryrerzr �s
r cs*eZdZdZ�fdd�Zddd�Z�ZS)ra�
    Lookahead matching of the given parse expression.  C{FollowedBy}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression matches at the current
    position.  C{FollowedBy} always returns a null token list.

    Example::
        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
    prints::
        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    cstt|��|�d|_dSr�)rdrr�r��r�rvreryrzr��szFollowedBy.__init__TcCs|j�||�|gfSr�)rvr�r�ryryrzr��szFollowedBy.parseImpl)Tr�ryryrerzr�srcs2eZdZdZ�fdd�Zd	dd�Zdd�Z�ZS)
ra�
    Lookahead to disallow matching with the given parse expression.  C{NotAny}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression does I{not} match at the current
    position.  Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
    always returns a null token list.  May be constructed using the '~' operator.

    Example::
        
    cs0tt|��|�d|_d|_dt|j�|_dS)NFTzFound unwanted token, )rdrr�r�r�r�rvr�r"reryrzr��szNotAny.__init__TcCs&|j�||�rt|||j|��|gfSr�)rvr�r!r�r�ryryrzr��szNotAny.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�z~{r��rr�r�r�rvr�ryryrzr�s


zNotAny.__str__)Tr�ryryrerzr�s

rcs(eZdZd�fdd�	Zddd�Z�ZS)	�_MultipleMatchNcsFtt|��|�d|_|}t|t�r.t�|�}|dur<|nd|_dSr�)	rdr$r�r�r}r�r&r��	not_ender)r�rv�stopOn�enderreryrzr�s

z_MultipleMatch.__init__Tc	Cs�|jj}|j}|jdu}|r$|jj}|r2|||�||||dd�\}}zV|j}	|r`|||�|	rp|||�}
n|}
|||
|�\}}|s�|��rR||7}qRWnttfy�Yn0||fS�NFr )	rvr�r�r%r�r�rr!r�)r�rur�r��self_expr_parse�self_skip_ignorables�check_ender�
try_not_enderr��hasIgnoreExprsr��	tmptokensryryrzr�s*



z_MultipleMatch.parseImpl)N)T)r�r�r�r�r�r�ryryrerzr$
sr$c@seZdZdZdd�ZdS)ra�
    Repetition of one or more of the given expression.
    
    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
        
        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    cCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r�z}...r#r�ryryrzr�Js


zOneOrMore.__str__N)r�r�r�r�r�ryryryrzr0srcs8eZdZdZd
�fdd�	Zd�fdd�	Zdd	�Z�ZS)r4aw
    Optional repetition of zero or more of the given expression.
    
    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example: similar to L{OneOrMore}
    Ncstt|�j||d�d|_dS)N)r&T)rdr4r�r�)r�rvr&reryrzr�_szZeroOrMore.__init__Tc	s:ztt|��|||�WSttfy4|gfYS0dSr�)rdr4r�r!r�r�reryrzr�cszZeroOrMore.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r9�]...r#r�ryryrzr�is


zZeroOrMore.__str__)N)Tr�ryryrerzr4Ssr4c@s eZdZdd�ZeZdd�ZdS)�
_NullTokencCsdSrryr�ryryrzrssz_NullToken.__bool__cCsdSr�ryr�ryryrzr�vsz_NullToken.__str__N)r�r�r�rrmr�ryryryrzr0rsr0cs6eZdZdZef�fdd�	Zd	dd�Zdd�Z�ZS)
raa
    Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::
        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345
            
            # ZIP+4 form
            12101-0001
            
            # invalid ZIP
            98765-
            ''')
    prints::
        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    cs.tt|�j|dd�|jj|_||_d|_dS)NFr�T)rdrr�rvr�r!r�)r�rvrreryrzr��s
zOptional.__init__Tc	Csxz|jj|||dd�\}}WnRttfyn|jturf|jjr\t|jg�}|j||jj<qj|jg}ng}Yn0||fSr()rvr�r!r�r!�_optionalNotMatchedr�r$)r�rur�r�r�ryryrzr��s


zOptional.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr�r9r<r#r�ryryrzr��s


zOptional.__str__)T)	r�r�r�r�r1r�r�r�r�ryryrerzrzs"
rcs,eZdZdZd	�fdd�	Zd
dd�Z�ZS)r*a�	
    Token for skipping over all undefined text until the matched expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default=C{False}) if True, the target expression is also parsed 
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default=C{None}) used to define grammars (typically quoted strings and 
          comments) that might contain false matches to the target expression
     - failOn - (default=C{None}) define expressions that are not allowed to be 
          included in the skipped test; if found before the target expression is found, 
          the SkipTo is not a match

    Example::
        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP 
                      + string_data("sev") + SEP 
                      + string_data("desc") + SEP 
                      + integer("days_open"))
        
        for tkt in ticket_expr.searchString(report):
            print tkt.dump()
    prints::
        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    FNcs`tt|��|�||_d|_d|_||_d|_t|t	�rFt
�|�|_n||_dt
|j�|_dS)NTFzNo match found for )rdr*r��
ignoreExprr�r��includeMatchr�r}r�r&r��failOnr�rvr�)r�r+�includerRr4reryrzr��s
zSkipTo.__init__Tc	Cs"|}t|�}|j}|jj}|jdur,|jjnd}|jdurB|jjnd}	|}
|
|kr�|durf|||
�rfq�|	dur�z|	||
�}
Wqnty�Yq�Yqn0qnz|||
ddd�Wq�tt	fy�|
d7}
YqJ0q�qJt|||j
|��|
}|||�}t|�}|j�r||||dd�\}}
||
7}||fS)NF)r�r�r�r )
r�rvr�r4r�r2r�rr!r�r�r$r3)r�rur�r�rxr�rv�
expr_parse�self_failOn_canParseNext�self_ignoreExpr_tryParse�tmploc�skiptext�
skipresultr�ryryrzr��s:
zSkipTo.parseImpl)FNN)Tr�ryryrerzr*�s6
r*csbeZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Zdd�Zgfd
d�Z	dd�Z
�fdd�Z�ZS)raK
    Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.

    Note: take care when assigning to C{Forward} not to overlook precedence of operators.
    Specifically, '|' has a lower precedence than '<<', so that::
        fwdExpr << a | b | c
    will actually be evaluated as::
        (fwdExpr << a) | b | c
    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the C{Forward}::
        fwdExpr << (a | b | c)
    Converting to use the '<<=' operator instead will avoid this problem.

    See L{ParseResults.pprint} for an example of a recursive parser created using
    C{Forward}.
    Ncstt|�j|dd�dSr�)rdrr�r7reryrzr�@szForward.__init__cCsjt|t�rt�|�}||_d|_|jj|_|jj|_|�|jj	�|jj
|_
|jj|_|j�
|jj�|Sr�)r}r�r&r�rvr�r�r�rPr�r�r�r�r&r7ryryrz�
__lshift__Cs





zForward.__lshift__cCs||>Sr�ryr7ryryrz�__ilshift__PszForward.__ilshift__cCs
d|_|SrrNr�ryryrzrOSszForward.leaveWhitespacecCs$|js d|_|jdur |j��|Sr�)r�rvrr�ryryrzrWs


zForward.streamlinecCs>||vr0|dd�|g}|jdur0|j�|�|�g�dSr�r r!ryryrzr[^s

zForward.validatecCs^t|d�r|jS|jjdSz&|jdur4t|j�}nd}W|j|_n
|j|_0|jjd|S)Nr�z: ...�Nonez: )rr�r�r�Z_revertClass�_ForwardNoRecurservr�)r��	retStringryryrzr�es

zForward.__str__cs.|jdurtt|���St�}||K}|SdSr�)rvrdrr�rGreryrzr�vs

zForward.copy)N)
r�r�r�r�r�r<r=rOrr[r�r�r�ryryrerzr-s
rc@seZdZdd�ZdS)r?cCsdS)Nr�ryr�ryryrzr�sz_ForwardNoRecurse.__str__N)r�r�r�r�ryryryrzr?~sr?cs"eZdZdZd�fdd�	Z�ZS)r/zQ
    Abstract subclass of C{ParseExpression}, for converting parsed results.
    Fcstt|��|�d|_dSr)rdr/r�r�rreryrzr��szTokenConverter.__init__)Fr�ryryrerzr/�sr/cs6eZdZdZd
�fdd�	Z�fdd�Zdd	�Z�ZS)ra�
    Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the input string;
    this can be disabled by specifying C{'adjacent=False'} in the constructor.

    Example::
        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    r�Tcs8tt|��|�|r|��||_d|_||_d|_dSr�)rdrr�rO�adjacentr��
joinStringr�)r�rvrBrAreryrzr��szCombine.__init__cs(|jrt�||�ntt|��|�|Sr�)rAr&rRrdrr7reryrzrR�szCombine.ignorecCsP|��}|dd�=|td�|�|j��g|jd�7}|jrH|��rH|gS|SdS)Nr�)r�)r�r$r�r=rBr�r�r)r�rur�r��retToksryryrzr��s
"zCombine.postParse)r�T)r�r�r�r�r�rRr�r�ryryrerzr�s
rcs(eZdZdZ�fdd�Zdd�Z�ZS)ra�
    Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.

    Example::
        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a,b,100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a,b,100"))  # -> ['fn', ['a', 'b', '100']]
    cstt|��|�d|_dSr�)rdrr�r�r"reryrzr��szGroup.__init__cCs|gSr�ryr�ryryrzr��szGroup.postParse�r�r�r�r�r�r�r�ryryrerzr�s
rcs(eZdZdZ�fdd�Zdd�Z�ZS)r
aW
    Converter to return a repetitive expression as a list, but also as a dictionary.
    Each element can also be referenced using the first token in the expression as its key.
    Useful for tabular report scraping when the first column can be used as a item key.

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())
        
        # access named fields as dict entries, or output as dict
        print(result['shape'])        
        print(result.asDict())
    prints::
        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
    See more examples at L{ParseResults} of accessing fields by results name.
    cstt|��|�d|_dSr�)rdr
r�r�r"reryrzr��sz
Dict.__init__cCs�t|�D]�\}}t|�dkrq|d}t|t�r@t|d���}t|�dkr\td|�||<qt|�dkr�t|dt�s�t|d|�||<q|��}|d=t|�dks�t|t�r�|�	�r�t||�||<qt|d|�||<q|j
r�|gS|SdS)Nrr�r�rs)r�r�r}rvr�r�r�r$r�rr�)r�rur�r�r��tok�ikey�	dictvalueryryrzr��s$
zDict.postParserDryryrerzr
�s#r
c@s eZdZdZdd�Zdd�ZdS)r-aV
    Converter for ignoring the results of a parsed expression.

    Example::
        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))
    prints::
        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']
    (See also L{delimitedList}.)
    cCsgSr�ryr�ryryrzr�szSuppress.postParsecCs|Sr�ryr�ryryrzrM"szSuppress.suppressN)r�r�r�r�r�rMryryryrzr-sr-c@s(eZdZdZdd�Zdd�Zdd�ZdS)	rzI
    Wrapper for parse actions, to ensure they are only called once.
    cCst|�|_d|_dSr)r��callable�called)r��
methodCallryryrzr�*s
zOnlyOnce.__init__cCs.|js|�|||�}d|_|St||d��dS)NTr�)rIrHr!)r�r�rrxr�ryryrzrL-s
zOnlyOnce.__call__cCs
d|_dSr)rIr�ryryrz�reset3szOnlyOnce.resetN)r�r�r�r�r�rLrKryryryrzr&srcs8t����fdd�}z�j|_Wnty2Yn0|S)at
    Decorator for debugging parse actions. 
    
    When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
    When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.

    Example::
        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
    prints::
        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    c
s��j}|dd�\}}}t|�dkr8|djjd|}tj�d|t||�||f�z�|�}Wn:ty�}z"tj�d||f��WYd}~n
d}~00tj�d||f�|S)Nr�rqr�.z">>entering %s(line: '%s', %d, %r)
z<<leaving %s (exception: %s)
z<<leaving %s (ret: %r)
)r�r�r�r��stderr�writerIr�)�paArgs�thisFuncr�rrxr�r|�raryrz�zLsztraceParseAction.<locals>.z)r�r�r�)rarRryrQrzrd6s
rd�,FcCs`t|�dt|�dt|�d}|rBt|t||���|�S|tt|�|��|�SdS)a�
    Helper to define a delimited list of expressions - the delimiter defaults to ','.
    By default, the list elements and delimiters can have intervening whitespace, and
    comments, but this can be overridden by passing C{combine=True} in the constructor.
    If C{combine} is set to C{True}, the matching tokens are returned as a single token
    string, with the delimiters included; otherwise, the matching tokens are returned
    as a list of tokens, with the delimiters suppressed.

    Example::
        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    z [rqr/N)r�rr4r�r-)rv�delim�combine�dlNameryryrzrBbs
$rBcsjt����fdd�}|dur0tt��dd��}n|��}|�d�|j|dd�|��d	t��d
�S)a:
    Helper to define a counted list of expressions.
    This helper defines a pattern of the form::
        integer expr expr expr...
    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
    
    If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.

    Example::
        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    cs.|d}�|r tt�g|��p&tt�>gSr�)rrrE)r�rrxr<��	arrayExprrvryrz�countFieldParseAction�s"z+countedArray.<locals>.countFieldParseActionNcSst|d�Sr�)rvrwryryrzr{�r|zcountedArray.<locals>.<lambda>�arrayLenT�r�z(len) r�)rr1rTr�r�r�r�r�)rv�intExprrYryrWrzr>us
r>cCs6g}|D](}t|t�r&|�t|��q|�|�q|Sr�)r}r�r&r*r$)�Lr�r�ryryrzr*�s
r*cs6t���fdd�}|j|dd���dt|���S)a*
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches a
    previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
    If this is not desired, use C{matchPreviousExpr}.
    Do I{not} use with packrat parsing enabled.
    csP|rBt|�dkr�|d>qLt|���}�tdd�|D��>n
�t�>dS)Nr�rcss|]}t|�VqdSr�)r�r��ttryryrzr��r|zDmatchPreviousLiteral.<locals>.copyTokenToRepeater.<locals>.<genexpr>)r�r*r�rr)r�rrx�tflat��repryrz�copyTokenToRepeater�sz1matchPreviousLiteral.<locals>.copyTokenToRepeaterTr[�(prev) )rr�r�r�)rvrcryrarzrQ�s


rQcsFt��|��}�|K��fdd�}|j|dd���dt|���S)aS
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches by
    expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
    the expressions are evaluated first, and then compared, so
    C{"1"} is compared with C{"10"}.
    Do I{not} use with packrat parsing enabled.
    cs*t|�����fdd�}�j|dd�dS)Ncs$t|���}|�kr tddd��dS)Nr�r)r*r�r!)r�rrx�theseTokens��matchTokensryrz�mustMatchTheseTokens�szLmatchPreviousExpr.<locals>.copyTokenToRepeater.<locals>.mustMatchTheseTokensTr[)r*r�r�)r�rrxrhrarfrzrc�sz.matchPreviousExpr.<locals>.copyTokenToRepeaterTr[rd)rr�r�r�r�)rv�e2rcryrarzrP�srPcCs:dD]}|�|t|�}q|�dd�}|�dd�}t|�S)Nz\^-]rHrpr�r�)r��_bslashr�)r�r�ryryrzr��s
r�Tc
s�|rdd�}dd�}t�ndd�}dd�}t�g}t|t�rF|��}n$t|t�rZt|�}ntjdt	dd�|stt
�Sd	}|t|�d
k�r||}t||d
d��D]R\}}	||	|�r�|||d
=qxq�|||	�r�|||d
=|�
||	�|	}qxq�|d
7}qx|�s�|�r�zlt|�td�|��k�rTtd
d�dd�|D����d�|��WStd�dd�|D����d�|��WSWn$t�y�tjdt	dd�Yn0t�fdd�|D���d�|��S)a�
    Helper to quickly define a set of alternative Literals, and makes sure to do
    longest-first testing when there is a conflict, regardless of the input order,
    but returns a C{L{MatchFirst}} for best performance.

    Parameters:
     - strs - a string of space-delimited literals, or a collection of string literals
     - caseless - (default=C{False}) - treat all literals as caseless
     - useRegex - (default=C{True}) - as an optimization, will generate a Regex
          object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
          if creating a C{Regex} raises an exception)

    Example::
        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))
    prints::
        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    cSs|��|��kSr�)r��r-�bryryrzr{�r|zoneOf.<locals>.<lambda>cSs|���|���Sr�)r�r�rkryryrzr{�r|cSs||kSr�ryrkryryrzr{�r|cSs
|�|�Sr�)r�rkryryrzr{�r|z6Invalid argument to oneOf, expected string or iterablersr2rr�Nr�z[%s]css|]}t|�VqdSr�)r��r��symryryrzr�r|zoneOf.<locals>.<genexpr>r�|css|]}t�|�VqdSr�)r�r�rmryryrzr�r|z7Exception creating Regex for oneOf, building MatchFirstc3s|]}�|�VqdSr�ryrm��parseElementClassryrzr�$r|)r
rr}r�r�rr�r4r5r6rr�r�r"r�r)r�r�r)
�strsr��useRegex�isequal�masks�symbolsr��currr+ryrprzrU�sP



�


**�rUcCsttt||���S)a�
    Helper to easily and clearly define a dictionary by specifying the respective patterns
    for the key and value.  Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
    in the proper order.  The key pattern can include delimiting markers or punctuation,
    as long as they are suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the C{Dict} results can include named token
    fields.

    Example::
        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())
    prints::
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    )r
r4r)r rryryrzrC&s!rCcCs^t��dd��}|��}d|_|d�||d�}|r@dd�}ndd�}|�|�|j|_|S)	a�
    Helper to return the original, untokenized text for a given expression.  Useful to
    restore the parsed fields of an HTML start tag into the raw tag text itself, or to
    revert separate tokens with intervening whitespace back to the original matching
    input text. By default, returns astring containing the original parsed text.  
       
    If the optional C{asString} argument is passed as C{False}, then the return value is a 
    C{L{ParseResults}} containing any results names that were originally matched, and a 
    single token containing the original matched text from the input string.  So if 
    the expression passed to C{L{originalTextFor}} contains expressions with defined
    results names, you must set C{asString} to C{False} if you want to preserve those
    results name values.

    Example::
        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b","i"):
            opener,closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])
    prints::
        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    cSs|Sr�ry)r�r�rxryryrzr{ar|z!originalTextFor.<locals>.<lambda>F�_original_start�
_original_endcSs||j|j�Sr�)rxryr~ryryrzr{fr|cSs&||�d�|�d��g|dd�<dS)Nrxry)rr~ryryrz�extractTexthsz$originalTextFor.<locals>.extractText)rr�r�r�r�)rv�asString�	locMarker�endlocMarker�	matchExprrzryryrzriIs

ricCst|��dd��S)zp
    Helper to undo pyparsing's default grouping of And expressions, even
    if all but one are non-empty.
    cSs|dSr�ryrwryryrzr{sr|zungroup.<locals>.<lambda>)r/r�)rvryryrzrjnsrjcCs4t��dd��}t|d�|d�|����d��S)a�
    Helper to decorate a returned token with its starting and ending locations in the input string.
    This helper adds the following results names:
     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains C{<TAB>} characters, you may want to call
    C{L{ParserElement.parseWithTabs}}

    Example::
        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)
    prints::
        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    cSs|Sr�ryr~ryryrzr{�r|zlocatedExpr.<locals>.<lambda>�
locn_startr�locn_end)rr�rr�rO)rv�locatorryryrzrlusrlrErKrJrcrbz\[]-*.$+^?()~ �r�cCs|ddSr�ryr~ryryrzr{�r|r{z\\0?[xX][0-9a-fA-F]+cCstt|d�d�d��S)Nrz\0x�)�unichrrv�lstripr~ryryrzr{�r|z	\\0[0-7]+cCstt|ddd�d��S)Nrr��)r�rvr~ryryrzr{�r|z\]r�r9rr�negate�bodyr<csDdd��z"d��fdd�t�|�jD��WSty>YdS0dS)a�
    Helper to easily define string ranges for use in Word construction.  Borrows
    syntax from regexp '[]' string range definitions::
        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
    The input string must be enclosed in []'s, and the returned string is the expanded
    character set joined into a single string.
    The values enclosed in the []'s may be:
     - a single character
     - an escaped character with a leading backslash (such as C{\-} or C{\]})
     - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) 
         (C{\0x##} is also supported for backwards compatibility) 
     - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
     - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
     - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
    cSs<t|t�s|Sd�dd�tt|d�t|d�d�D��S)Nr�css|]}t|�VqdSr�)r�r�ryryrzr��r|z+srange.<locals>.<lambda>.<locals>.<genexpr>rr�)r}r$r�r��ord)�pryryrzr{�r|zsrange.<locals>.<lambda>r�c3s|]}�|�VqdSr�ry)r��part��	_expandedryrzr��r|zsrange.<locals>.<genexpr>N)r��_reBracketExprrr�r�r�ryr�rzra�s
"racs�fdd�}|S)zt
    Helper method for defining parse actions that require matching at a specific
    column in the input text.
    cs"t||��krt||d���dS)Nzmatched token not at column %dr�)ro�locnrzr;ryrz�	verifyCol�sz!matchOnlyAtCol.<locals>.verifyColry)r<r�ryr;rzrO�srOcs�fdd�S)a�
    Helper method for common parse actions that simply return a literal value.  Especially
    useful when used with C{L{transformString<ParserElement.transformString>}()}.

    Example::
        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num
        
        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    cs�gSr�ryr~��replStrryrzr{�r|zreplaceWith.<locals>.<lambda>ryr�ryr�rzr^�sr^cCs|ddd�S)a
    Helper parse action for removing quotation marks from parsed quoted strings.

    Example::
        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    rr�rtryr~ryryrzr\�sr\csL��fdd�}zt�dt�d�j�}Wnty@t��}Yn0||_|S)aG
    Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional 
    args are passed, they are forwarded to the given function as additional arguments after
    the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
    parsed data to an integer using base 16.

    Example (compare the last to example in L{ParserElement.transformString}::
        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')
        
        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')
    prints::
        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    cs��fdd�|D�S)Ncsg|]}�|g��R��qSryry)r��tokn�r�r�ryrzr��r|z(tokenMap.<locals>.pa.<locals>.<listcomp>ryr~r�ryrzr��sztokenMap.<locals>.par�r�)r�r�r�r)r�r�r�r�ryr�rzro�s 
�rocCst|���Sr��r�r�rwryryrzr{r|cCst|���Sr��r��lowerrwryryrzr{r|c	Cs�t|t�r|}t||d�}n|j}tttd�}|r�t���	t
�}td�|d�tt
t|td�|���tddgd��d	��	d
d��td�}n�d
�dd�tD��}t���	t
�t|�B}td�|d�tt
t|�	t�ttd�|����tddgd��d	��	dd��td�}ttd�|d�}|�dd
�|�dd��������d|�}|�dd
�|�dd��������d|�}||_||_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namer�z_-:rL�tag�=�/F�rrEcSs|ddkS�Nrr�ryr~ryryrzr{r|z_makeTags.<locals>.<lambda>rMr�css|]}|dvr|VqdS)rMNryr�ryryrzr�r|z_makeTags.<locals>.<genexpr>cSs|ddkSr�ryr~ryryrzr{r|rNr��:rqz<%s>r�z</%s>)r}r�rr�r1r6r5r@r�r�r\r-r
r4rrr�r�rXr[rDr�_Lr��titler�r�r�)�tagStr�xml�resname�tagAttrName�tagAttrValue�openTagZprintablesLessRAbrack�closeTagryryrz�	_makeTagss>
�������..r�cCs
t|d�S)a 
    Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
    tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.

    Example::
        text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
        a,a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end
        
        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are also accessible as named results
            print(link.link_text, '->', link.href)
    prints::
        pyparsing -> http://pyparsing.wikispaces.com
    F�r��r�ryryrzrM(srMcCs
t|d�S)z�
    Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
    tags only in the given upper/lower case.

    Example: similar to L{makeHTMLTags}
    Tr�r�ryryrzrN;srNcs8|r|dd��n|���dd��D���fdd�}|S)a<
    Helper to create a validating parse action to be used with start tags created
    with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
    with a required attribute value, to avoid false matches on common tags such as
    C{<TD>} or C{<DIV>}.

    Call C{withAttribute} with a series of attribute names and values. Specify the list
    of filter attributes names and values as:
     - keyword arguments, as in C{(align="right")}, or
     - as an explicit dict with C{**} operator, when an attribute name is also a Python
          reserved word, as in C{**{"class":"Customer", "align":"right"}}
     - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") )
    For attribute names with a namespace prefix, you must use the second form.  Attribute
    names are matched insensitive to upper/lower case.
       
    If just testing for C{class} (with or without a namespace), use C{L{withClass}}.

    To verify that the attribute exists, but without specifying a value, pass
    C{withAttribute.ANY_VALUE} as the value.

    Example::
        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    NcSsg|]\}}||f�qSryryrDryryrzr�zr|z!withAttribute.<locals>.<listcomp>csZ�D]P\}}||vr$t||d|��|tjkr|||krt||d||||f��qdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r!rg�	ANY_VALUE)r�rr��attrName�	attrValue��attrsryrzr�{s�zwithAttribute.<locals>.pa)r�)r��attrDictr�ryr�rzrgDs2rgcCs"|rd|nd}tfi||i��S)a�
    Simplified version of C{L{withAttribute}} when matching on a div class - made
    difficult because C{class} is a reserved word in Python.

    Example::
        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))
        
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    z%s:class�class)rg)�	classname�	namespace�	classattrryryrzrm�s rm�(r�cCs�t�}||||B}t|�D�]l\}}|ddd�\}}	}
}|	dkrPd|nd|}|	dkr�|dustt|�dkr|td��|\}
}t��|�}|
tjk�r^|	d	kr�t||�t|t	|��}n�|	dk�r|dur�t|||�t|t	||��}nt||�t|t	|��}nD|	dk�rTt||
|||�t||
|||�}ntd
��n�|
tj
k�rB|	d	k�r�t|t��s�t|�}t|j
|�t||�}n�|	dk�r�|du�r�t|||�t|t	||��}nt||�t|t	|��}nD|	dk�r8t||
|||�t||
|||�}ntd
��ntd��|�rvt|ttf��rl|j|�n
|�|�||�|�|BK}|}q||K}|S)aD

    Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary or
    binary, left- or right-associative.  Parse actions can also be attached
    to operator expressions. The generated parser will also recognize the use 
    of parentheses to override operator precedences (see example below).
    
    Note: if you define a deep operator list, you may see performance issues
    when using infixNotation. See L{ParserElement.enablePackrat} for a
    mechanism to potentially improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the nested
     - opList - list of tuples, one for each operator precedence level in the
      expression grammar; each tuple is of the form
      (opExpr, numTerms, rightLeftAssoc, parseAction), where:
       - opExpr is the pyparsing expression for the operator;
          may also be a string, which will be converted to a Literal;
          if numTerms is 3, opExpr is a tuple of two expressions, for the
          two operators separating the 3 terms
       - numTerms is the number of terms for this operator (must
          be 1, 2, or 3)
       - rightLeftAssoc is the indicator whether the operator is
          right or left associative, using the pyparsing-defined
          constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}.
       - parseAction is the parse action to be associated with
          expressions matching this operator expression (the
          parse action tuple member may be omitted); if the parse action
          is passed a tuple or list of functions, this is equivalent to
          calling C{setParseAction(*fn)} (L{ParserElement.setParseAction})
     - lpar - expression for matching left-parentheses (default=C{Suppress('(')})
     - rpar - expression for matching right-parentheses (default=C{Suppress(')')})

    Example::
        # simple example of four-function arithmetic with ints and variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier 
        
        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])
        
        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)
    prints::
        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    r�Nr�rqz%s termz	%s%s termrsz@if numterms=3, opExpr must be a tuple or list of two expressionsr�z6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)rr�r�r@r�rV�LEFTrrr�RIGHTr}rrvr?r�r�)�baseExpr�opList�lpar�rparr��lastExprr��operDef�opExpr�arity�rightLeftAssocr��termName�opExpr1�opExpr2�thisExprr~ryryrzrk�sZ=
&
�



&
�

rkz4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�'z string enclosed in single quotesz*quotedString using single or double quotes�uzunicode string literalcCs�||krtd��|du�r*t|t��r"t|t��r"t|�dkr�t|�dkr�|dur�tt|t||tjdd����	dd��}n$t
��t||tj��	dd��}nx|dur�tt|t|�t|�ttjdd����	dd��}n4ttt|�t|�ttjdd����	d	d��}ntd
��t
�}|du�rd|tt|�t||B|B�t|��K}n$|tt|�t||B�t|��K}|�d||f�|S)a~	
    Helper method for defining nested lists enclosed in opening and closing
    delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression
     - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression
     - content - expression for items within the nested lists (default=C{None})
     - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString})

    If an expression is not provided for the content argument, the nested
    expression will capture all whitespace-delimited content between delimiters
    as a list of separate values.

    Use the C{ignoreExpr} argument to define expressions that may contain
    opening or closing characters that should not be treated as opening
    or closing characters for nesting, such as quotedString or a comment
    expression.  Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}.
    The default is L{quotedString}, but if no expressions are to be ignored,
    then pass C{None} for this argument.

    Example::
        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR,RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type") 
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR 
                      + code_body("body"))
        c_function.ignore(cStyleComment)
        
        source_code = '''
            int is_odd(int x) { 
                return (x%2); 
            }
                
            int dec_to_hex(char hchar) { 
                if (hchar >= '0' && hchar <= '9') { 
                    return (ord(hchar)-ord('0')); 
                } else { 
                    return (10+ord(hchar)-ord('A'));
                } 
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)

    prints::
        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    z.opening and closing strings cannot be the sameNr�r�cSs|d��Sr��r�rwryryrzr{gr|znestedExpr.<locals>.<lambda>cSs|d��Sr�r�rwryryrzr{jr|cSs|d��Sr�r�rwryryrzr{pr|cSs|d��Sr�r�rwryryrzr{tr|zOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)r@r}r�r�rrrr&r�r�rEr�rrrr-r4r�)�opener�closer�contentr2r�ryryrzrR%sH:
���������
*$rRcs��fdd�}�fdd�}�fdd�}tt��d����}t�t��|��d�}t��|��d	�}t��|��d
�}	|r�tt|�|t|t|�t|��|	�}
n$tt|�t|t|�t|���}
|�	t
t��|
�d�S)a
	
    Helper method for defining space-delimited indentation blocks, such as
    those used to define block statements in Python source code.

    Parameters:
     - blockStatementExpr - expression defining syntax of statement that
            is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
            (multiple statementWithIndentedBlock expressions within a single grammar
            should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond the
            the current level; set to False for block of left-most statements
            (default=C{True})

    A valid block must contain at least one C{blockStatement}.

    Example::
        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group( funcDecl + func_body )

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << ( funcDef | assignment | identifier )

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()
    prints::
        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] 
    csN|t|�krdSt||�}|�dkrJ|�dkr>t||d��t||d��dS)Nrtzillegal nestingznot a peer entry)r�r;r#r!�r�rrx�curCol��indentStackryrz�checkPeerIndent�s
z&indentedBlock.<locals>.checkPeerIndentcs2t||�}|�dkr"��|�nt||d��dS)Nrtznot a subentry)r;r$r!r�r�ryrz�checkSubIndent�s
z%indentedBlock.<locals>.checkSubIndentcsN|t|�krdSt||�}�r6|�dkr6|�dksBt||d�����dS)Nrtr�znot an unindent)r�r;r!rr�r�ryrz�
checkUnindent�s

z$indentedBlock.<locals>.checkUnindentz	 �INDENTr��UNINDENTzindented block)rrrPrMrr�r�rrrRrj)�blockStatementExprr�rRr�r�r�rer��PEER�UNDENT�smExprryr�rzrhs(N����rhz#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z&(?P<entity>roz);zcommon HTML entitycCst�|j�S)zRHelper parser action to replace common HTML entities with their special characters)�_htmlEntityMapr��entityrwryryrzr]�sr]z/\*(?:[^*]|\*(?!/))*z*/zC style commentz<!--[\s\S]*?-->zHTML commentz.*zrest of linez//(?:\\\n|[^\n])*z
// commentzC++ style commentz#.*zPython style commentr�� 	�	commaItemr�r<c@s�eZdZdZee�Zee�Ze	e
��d��e�Z
e	e��d��eed��Zed��d��e�Ze��e�de��e��d�Ze�d	d
��eeeed���e�B�d�Ze�e�ed
��d��e�Zed��d��e�ZeeBeB��Zed��d��e�Ze	eded��d�Zed��d�Z ed��d�Z!e!de!d�d�Z"ee!de!d�dee!de!d��d�Z#e#�$dd
��d e �d!�Z%e&e"e%Be#B�d"���d"�Z'ed#��d$�Z(e)d=d&d'��Z*e)d>d)d*��Z+ed+��d,�Z,ed-��d.�Z-ed/��d0�Z.e/��e0��BZ1e)d1d2��Z2e&e3e4d3�e5�e	e6d3d4�ee7d5�������d6�Z8e9ee:�;�e8Bd7d8���d9�Z<e)ed:d
���Z=e)ed;d
���Z>d<S)?rpa�

    Here are some common low-level expressions that may be useful in jump-starting parser development:
     - numeric forms (L{integers<integer>}, L{reals<real>}, L{scientific notation<sci_real>})
     - common L{programming identifiers<identifier>}
     - network addresses (L{MAC<mac_address>}, L{IPv4<ipv4_address>}, L{IPv6<ipv6_address>})
     - ISO8601 L{dates<iso8601_date>} and L{datetime<iso8601_datetime>}
     - L{UUID<uuid>}
     - L{comma-separated list<comma_separated_list>}
    Parse actions:
     - C{L{convertToInteger}}
     - C{L{convertToFloat}}
     - C{L{convertToDate}}
     - C{L{convertToDatetime}}
     - C{L{stripHTMLTags}}
     - C{L{upcaseTokens}}
     - C{L{downcaseTokens}}

    Example::
        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')
    prints::
        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    �integerzhex integerr�z[+-]?\d+zsigned integerr��fractioncCs|d|dS)Nrrtryrwryryrzr{�r|zpyparsing_common.<lambda>r�z"fraction or mixed integer-fractionz
[+-]?\d+\.\d*zreal numberz+[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?�fnumberr�
identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}zIPv4 addressz[0-9a-fA-F]{1,4}�hex_integerr��zfull IPv6 address)rr�z::zshort IPv6 addresscCstdd�|D��dkS)Ncss|]}tj�|�rdVqdSrI)rp�
_ipv6_partr%r^ryryrzr��r|z,pyparsing_common.<lambda>.<locals>.<genexpr>r�)rrwryryrzr{�r|z::ffff:zmixed IPv6 addresszIPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}zMAC address�%Y-%m-%dcs�fdd�}|S)a�
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"})

        Example::
            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))
        prints::
            [datetime.date(1999, 12, 31)]
        c
sPzt�|d����WStyJ}zt||t|���WYd}~n
d}~00dSr�)r�strptime�dater@r!r�r�rrx�ve��fmtryrz�cvt_fn�sz.pyparsing_common.convertToDate.<locals>.cvt_fnry�r�r�ryr�rz�
convertToDate�szpyparsing_common.convertToDate�%Y-%m-%dT%H:%M:%S.%fcs�fdd�}|S)a
        Helper to create a parse action for converting parsed datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"})

        Example::
            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))
        prints::
            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        c
sLzt�|d��WStyF}zt||t|���WYd}~n
d}~00dSr�)rr�r@r!rr�r�ryrzr��sz2pyparsing_common.convertToDatetime.<locals>.cvt_fnryr�ryr�rz�convertToDatetime�sz"pyparsing_common.convertToDatetimez7(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?zISO8601 datez�(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}�UUIDcCstj�|d�S)a
        Parse action to remove HTML tags from web page HTML source

        Example::
            # strip HTML links from normal text 
            text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
            td,td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            
            print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page'
        r)rp�_html_stripperr�)r�rr�ryryrz�
stripHTMLTagss
zpyparsing_common.stripHTMLTagsrSr�r�r�r�r�zcomma separated listcCst|���Sr�r�rwryryrzr{"r|cCst|���Sr�r�rwryryrzr{%r|N)r�)r�)?r�r�r�r�rorv�convertToInteger�float�convertToFloatr1rTr�r�r�rFr�r)�signed_integerr�r�rrM�
mixed_integerr�real�sci_realr�numberr�r6r5r��ipv4_addressr��_full_ipv6_address�_short_ipv6_addressr��_mixed_ipv6_addressr�ipv6_address�mac_addressrr�r��iso8601_date�iso8601_datetime�uuidr9r8r�r�rrrrXr0�
_commasepitemrBr[r��comma_separated_listrfrDryryryrzrpsV""
2
 
��rp�__main__�select�fromr�rL)rU�columnsr�Ztables�commandaK
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        z]
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        z 
        100
        FF
        z6
        12345678-1234-5678-1234-567812345678
        )rs)rSF)N)FT)T)r�)T)�r��__version__�__versionTime__�
__author__r��weakrefrr�r�r�r4r�r�rrfr�r�r�_threadr�ImportError�	threading�collections.abcrrrrZordereddict�__all__r?�version_infor�rE�maxsizer�rr��chrr�r�rr�r`�reversedr�r�rbr�r�r�r�Zmaxint�xranger��__builtin__r��fnamer$r�r�r�r�r�r�r��ascii_uppercase�ascii_lowercaser6rTrFr5rjr��	printablerXr�rr!r#r%r(r�r$�registerr;rLrIrwr{r}rSr�r&r.rrrr�r�rr
r	rnr1r)r'rr0r�rrrr,r+r3r2r"rrrrr rrr$rr4r0r1rr*rr?r/rrr
r-rrdrBr>r*rQrPr�rUrCrirjrlr�rErKrJrcrbr��_escapedPunc�_escapedHexChar�_escapedOctChar�_singleChar�
_charRanger�r�rarOr^r\rorfrDr�rMrNrgr�rmrVr�r�rkrWr@r`r[rerRrhr7rYr9r8r�r�r�rr=r]r:rGrOr_rAr?rHrZrrr<rpr�ZselectTokenZ	fromToken�identZ
columnNameZcolumnNameListZ
columnSpecZ	tableNameZ
tableNameListZ	simpleSQLr~r�r�r�rr�ryryryrz�<module>s��4
8



@v&A= I
G3pLOD|M &#@sQ,A,	I#%0
,	?#p
��Zr 
����"


"
	site-packages/pkg_resources/_vendor/packaging/_typing.py000064400000003424150564745010017554 0ustar00"""For neatly implementing static typing in packaging.

`mypy` - the static type analysis tool we use - uses the `typing` module, which
provides core functionality fundamental to mypy's functioning.

Generally, `typing` would be imported at runtime and used in that fashion -
it acts as a no-op at runtime and does not have any run-time overhead by
design.

As it turns out, `typing` is not vendorable - it uses separate sources for
Python 2/Python 3. Thus, this codebase can not expect it to be present.
To work around this, mypy allows the typing import to be behind a False-y
optional to prevent it from running at runtime and type-comments can be used
to remove the need for the types to be accessible directly during runtime.

This module provides the False-y guard in a nicely named fashion so that a
curious maintainer can reach here to read this.

In packaging, all static-typing related imports should be guarded as follows:

    from packaging._typing import TYPE_CHECKING

    if TYPE_CHECKING:
        from typing import ...

Ref: https://github.com/python/mypy/issues/3216
"""

__all__ = ["TYPE_CHECKING", "cast"]

# The TYPE_CHECKING constant defined by the typing module is False at runtime
# but True while type checking.
if False:  # pragma: no cover
    from typing import TYPE_CHECKING
else:
    TYPE_CHECKING = False

# typing's cast syntax requires calling typing.cast at runtime, but we don't
# want to import typing at runtime. Here, we inform the type checkers that
# we're importing `typing.cast` as `cast` and re-implement typing.cast's
# runtime behavior in a block that is ignored by type checkers.
if TYPE_CHECKING:  # pragma: no cover
    # not executed at runtime
    from typing import cast
else:
    # executed at runtime
    def cast(type_, value):  # noqa
        return value
site-packages/pkg_resources/_vendor/packaging/__pycache__/utils.cpython-39.pyc000064400000003252150564745010023531 0ustar00a

(�Re�@sxddlmZmZmZddlZddlmZmZddlm	Z	m
Z
erZddlmZm
Z
ede�Ze�d�Zd	d
�Zdd�ZdS)
�)�absolute_import�division�print_functionN�)�
TYPE_CHECKING�cast)�InvalidVersion�Version)�NewType�Union�NormalizedNamez[-_.]+cCst�d|���}td|�S)N�-r)�_canonicalize_regex�sub�lowerr)�name�value�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/utils.py�canonicalize_namesrc
Cs�zt|�}Wnty"|YS0g}|jdkrD|�d�|j��|�t�ddd�dd�|jD����|j	dur�|�d�d	d�|j	D���|j
dur�|�d
�|j
��|jdur�|�d�|j��|jdur�|�d�|j��d�|�S)
z�
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    rz{0}!z(\.0)+$��.css|]}t|�VqdS�N��str��.0�xrrr�	<genexpr>/�z'canonicalize_version.<locals>.<genexpr>Ncss|]}t|�VqdSrrrrrrr3rz.post{0}z.dev{0}z+{0})
r	r�epoch�append�format�rer�join�release�pre�post�dev�local)�_version�version�partsrrr�canonicalize_versions"

&



r-)�
__future__rrrr#Z_typingrrr+rr	�typingr
rrr�compilerrr-rrrr�<module>s

site-packages/pkg_resources/_vendor/packaging/__pycache__/tags.cpython-39.pyc000064400000041643150564745010023335 0ustar00a

(�Re^�@sLddlmZddlZzddlmZWn.eyRddlZdd�e��D�Z[Yn0ddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZmZer�ddlmZmZmZmZmZmZmZmZmZmZeeZ eeefZ!eeefZ"e	�#e$�Z%d	d
ddd
d�Z&ej'dkZ(Gdd�de)�Z*dd�Z+dd�Z,dSdd�Z-dd�Z.dd�Z/dTdd�Z0dUdd �Z1d!d"�Z2dVd#d$�Z3d%d&�Z4dWd'd(�Z5e(fd)d*�Z6d+d,�Z7dXd-d.�Z8d/d0�Z9d1d2�Z:d3d4�Z;d5d6�Z<d7d8�Z=d9d:�Z>Gd;d<�d<e)�Z?d=d>�Z@d?d@�ZAdAdB�ZBdCdD�ZCe(fdEdF�ZDdGdH�ZEdIdJ�ZFdKdL�ZGdMdN�ZHdOdP�ZIdQdR�ZJdS)Y�)�absolute_importN)�EXTENSION_SUFFIXEScCsg|]}|d�qS)r�)�.0�xrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/tags.py�
<listcomp>�r�)�
TYPE_CHECKING�cast)
�Dict�	FrozenSet�IO�Iterable�Iterator�List�Optional�Sequence�Tuple�Union�py�cp�pp�ip�jy)�python�cpython�pypy�
ironpython�jythonlc@sdeZdZdZgd�Zdd�Zedd��Zedd��Zed	d
��Z	dd�Z
d
d�Zdd�Zdd�Z
dS)�Tagz�
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    )�_interpreter�_abi�	_platformcCs"|��|_|��|_|��|_dS�N)�lowerr"r#r$)�self�interpreter�abi�platformrrr�__init__Fs

zTag.__init__cCs|jSr%)r"�r'rrrr(LszTag.interpretercCs|jSr%)r#r,rrrr)QszTag.abicCs|jSr%)r$r,rrrr*VszTag.platformcCs2t|t�stS|j|jko0|j|jko0|j|jkSr%)�
isinstancer!�NotImplementedr*r)r()r'�otherrrr�__eq__[s

�
�z
Tag.__eq__cCst|j|j|jf�Sr%)�hashr"r#r$r,rrr�__hash__fszTag.__hash__cCsd�|j|j|j�S)Nz{}-{}-{})�formatr"r#r$r,rrr�__str__jszTag.__str__cCsdj|t|�d�S)Nz<{self} @ {self_id}>)r'�self_id)r3�idr,rrr�__repr__nszTag.__repr__N)�__name__�
__module__�__qualname__�__doc__�	__slots__r+�propertyr(r)r*r0r2r4r7rrrrr!<s


r!c	Cs`t�}|�d�\}}}|�d�D]6}|�d�D]&}|�d�D]}|�t|||��q<q.q t|�S)z�
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    �-�.)�set�split�addr!�	frozenset)�tag�tags�interpreters�abis�	platformsr(r)�	platform_rrr�	parse_tagssrJcCsP|sdSt|�dksd|vrH|�dd�tt|����}td�||���|dS)z[
    Backwards-compatibility with Python 2.7 to allow treating 'warn' as keyword-only.
    Fr
�warnNz,{}() got an unexpected keyword argument {!r})�len�pop�next�iter�keys�	TypeErrorr3)�	func_name�kwargs�argrrr�_warn_keyword_parameter�s
�rUFcCs&t�|�}|dur"|r"t�d|�|S)Nz>Config variable '%s' is unset, Python ABI tag may be incorrect)�	sysconfig�get_config_var�logger�debug)�namerK�valuerrr�_get_config_var�s
�r\cCs|�dd��dd�S)Nr?�_r>)�replace)�stringrrr�_normalize_string�sr`cCst|�dkot|�dkS)zj
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    r
)��)rL�tuple)�python_versionrrr�
_abi3_applies�srec	Cs�t|�}g}t|dd��}d}}}td|�}ttd�}dtv}	|sX|dur\|sX|	r\d}|dkr�td|�}
|
sz|
dur~d	}|d
kr�td|�}|dks�|dur�tjd
kr�d}n|r�|�dj|d��|�	ddj||||d��|S)Nrb��Py_DEBUG�gettotalrefcountz_d.pyd�d)ra��
WITH_PYMALLOC�m)rara�Py_UNICODE_SIZE�i���u�cp{version}��versionrz"cp{version}{debug}{pymalloc}{ucs4})rrrY�pymalloc�ucs4)
rc�_version_nodotr\�hasattr�sysr�
maxunicode�appendr3�insert)�
py_versionrKrGrrrYrsrt�
with_debug�has_refcount�has_ext�
with_pymalloc�unicode_sizerrr�
_cpython_abis�s<



����r�c
	+sZtd|�}|stjdd�}d�t|dd����|durVt|�dkrRt||�}ng}t|�}dD]&}z|�|�Wqbt	y�Yqb0qbt|p�t
��}|D]}|D]}t�||�Vq�q�t|�r�fdd�|D�D]
}|Vqԇfd	d�|D�D]
}|Vq�t|��rVt
|dddd
�D]8}	|D],}djt|d|	f�d
��t�d|�V�q$�qdS)a
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    �cpython_tagsNrbzcp{}r
)�abi3�nonec3s|]}t�d|�VqdS)r�N�r!�rrI�r(rr�	<genexpr>r	zcpython_tags.<locals>.<genexpr>c3s|]}t�d|�VqdS)r�Nr�r�r�rrr�r	���rprrqr�)rUrw�version_infor3rurLr��list�remove�
ValueError�_platform_tagsr!re�range)
rdrGrHrSrK�explicit_abir)rIrD�
minor_versionrr�rr��s<

�r�ccst�d�}|rt|�VdS)N�SOABI)rVrWr`)r)rrr�_generic_abis
r�c	ks�td|�}|s,t�}t|d�}d�||g�}|dur:t�}t|pDt��}t|�}d|vrb|�d�|D]}|D]}t|||�VqnqfdS)z�
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    �generic_tags�rKrfNr�)	rU�interpreter_name�interpreter_version�joinr�r�r�ryr!)	r(rGrHrSrK�interp_name�interp_versionr)rIrrrr�s


r�ccs|t|�dkr&djt|dd��d�Vdj|dd�Vt|�dkrxt|ddd	d	�D]}djt|d|f�d�VqXdS)
z�
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    r
zpy{version}Nrbrqz	py{major}r)�majorr�)rLr3rur�)r{�minorrrr�_py_interpreter_range4sr�ccsx|stjdd�}t|pt��}t|�D]}|D]}t|d|�Vq0q(|rXt|dd�Vt|�D]}t|dd�Vq`dS)z�
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    Nrbr��any)rwr�r�r�r�r!)rdr(rHrrrIrrr�compatible_tagsDsr�cCs|s|S|�d�rdSdS)N�ppc�i386)�
startswith)�arch�is_32bitrrr�	_mac_arch^s

r�cCs�|g}|dkr*|dkrgS|�gd��nn|dkrN|dkr>gS|�gd��nJ|dkrv|dksf|dkrjgS|�d�n"|d	kr�|d
kr�gS|�ddg�|�d
�|S)N�x86_64)�
rn)�intel�fat64�fat32r�)r�r��fat�ppc64)r��r�r�)r��r�r��	universal)�extendry)rr�cpu_arch�formatsrrr�_mac_binary_formatsis&
r�c	cs�t��\}}}|dur:tdttt|�d�dd����}n|}|durPt|�}n|}t|ddd�D]>}|d|f}t	||�}|D]}dj
|d|d|d	�Vq�qddS)
aD
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    N�
MacVersionr?rbr
r�rz&macosx_{major}_{minor}_{binary_format})r�r��
binary_format)r*�mac_verrrc�map�intrAr�r�r�r3)	rrr��version_strr]r�r��compat_version�binary_formatsr�rrr�
mac_platforms�s 
$

�r�c	Cs<zddl}tt||d��WSttfy2Yn0t|�S)NrZ_compatible)�
_manylinux�bool�getattr�ImportError�AttributeError�_have_compatible_glibc)rZ�
glibc_versionr�rrr�_is_manylinux_compatible�sr�cCst�p
t�Sr%)�_glibc_version_string_confstr�_glibc_version_string_ctypesrrrr�_glibc_version_string�sr�cCsHz&t�d�}|dusJ�|��\}}WnttttfyBYdS0|S)zJ
    Primary implementation of glibc_version_string using os.confstr.
    �CS_GNU_LIBC_VERSIONN)�os�confstrrA�AssertionErrorr��OSErrorr�)�version_stringr]rrrrrr��s	�r�cCsrzddl}Wnty YdS0|�d�}z
|j}WntyJYdS0|j|_|�}t|t�sn|�	d�}|S)zG
    Fallback implementation of glibc_version_string using ctypes.
    rN�ascii)
�ctypesr��CDLL�gnu_get_libc_versionr��c_char_p�restyper-�str�decode)r��process_namespacer�r�rrrr��s



r�cCsHt�d|�}|s$t�d|t�dSt|�d��|koFt|�d��|kS)Nz$(?P<major>[0-9]+)\.(?P<minor>[0-9]+)z=Expected glibc version with 2 components major.minor, got: %sFr�r�)�re�match�warningsrK�RuntimeWarningr��group)r��required_major�
minimum_minorrlrrr�_check_glibc_version�s���r�cCst�}|durdSt|||�S�NF)r�r�)r�r�r�rrrr�sr�c@sTeZdZGdd�de�ZdZdZdZdZdZ	dZ
dZdZd	Z
d
ZdZdZd
d�ZdS)�_ELFFileHeaderc@seZdZdZdS)z$_ELFFileHeader._InvalidELFFileHeaderz7
        An invalid ELF file header was found.
        N)r8r9r:r;rrrr�_InvalidELFFileHeadersr�iFLEr
rbra��(�>l~iicsr�fdd�}|d�|_|j|jkr*t���|d�|_|j|j|jhvrNt���|d�|_|j|j|j	hvrrt���|d�|_
|d�|_|d�|_��
d�|_|j|jkr�dnd}|j|jkr�dnd}|j|jkr�d	nd
}|j|jkr�|n|}||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_dS)Ncs@zt�|��t�|���\}Wntjy:t���Yn0|Sr%)�struct�unpack�read�calcsize�errorr�r�)�fmt�result��filerrr�)s�
z'_ELFFileHeader.__init__.<locals>.unpackz>I�B�z<Hz>Hz<Iz<Qz>Q)�
e_ident_magic�ELF_MAGIC_NUMBERr�r��
e_ident_class�
ELFCLASS32�
ELFCLASS64�e_ident_data�ELFDATA2LSB�ELFDATA2MSB�e_ident_version�
e_ident_osabi�e_ident_abiversionr��e_ident_pad�e_type�	e_machine�	e_version�e_entry�e_phoff�e_shoff�e_flags�e_ehsize�e_phentsize�e_phnum�e_shentsize�e_shnum�
e_shstrndx)r'r�r��format_h�format_i�format_q�format_prr�rr+'s>


















z_ELFFileHeader.__init__N)r8r9r:r�r�r�r�r�r�r��EM_386�EM_S390�EM_ARM�	EM_X86_64�EF_ARM_ABIMASK�EF_ARM_ABI_VER5�EF_ARM_ABI_FLOAT_HARDr+rrrrr�sr�cCs\z8ttjd��}t|�}Wd�n1s,0YWnttttjfyVYdS0|S)N�rb)�openrw�
executabler��IOErrorr�rQr�)�f�
elf_headerrrr�_get_elf_headerSs*rcCsnt�}|durdS|j|jk}||j|jkM}||j|jkM}||j|j@|j	kM}||j|j
@|j
kM}|Sr�)rr�r�r�r�r�rrrrr�rr�rrr�_is_linux_armhf]s
�
�rcCsBt�}|durdS|j|jk}||j|jkM}||j|jkM}|Sr�)rr�r�r�r�r�r
rrrr�_is_linux_i686qsrcCs |dkrt�S|dkrt�SdS)N�armv7l�i686T)rr)r�rrr�_have_compatible_manylinux_abi|s
r ccs�ttj���}|r,|dkr d}n|dkr,d}g}|�dd�\}}t|�rv|dvrZ|�d�|d	vrv|�d
�|�d�t|�}|D]$\}}t||�r�|�	d|�Vq�q�|D]\}}|�	d|�Vq�|VdS)
N�linux_x86_64�
linux_i686�
linux_aarch64�linux_armv7lr]r
>�ppc64le�aarch64r�s390xrr�r�)�
manylinux2014)rb�>r�r)�
manylinux2010)rb�)�
manylinux1)rbr��linux)
r`�	distutils�util�get_platformrAr ryrOr�r^)r�r-Zmanylinux_supportr]r�Zmanylinux_support_iterrZr�rrr�_linux_platforms�s8���
r1ccsttj���VdSr%)r`r.r/r0rrrr�_generic_platforms�sr2cCs.t��dkrt�St��dkr$t�St�SdS)z;
    Provides the platform tags for this installation.
    �Darwin�LinuxN)r*�systemr�r1r2rrrrr��s
r�cCs:ztjj}Wnty*t����}Yn0t�|�p8|S)z6
    Returns the name of the running interpreter.
    )	rw�implementationrZr�r*�python_implementationr&�INTERPRETER_SHORT_NAMES�get)rZrrrr��s
r�cKs:td|�}td|d�}|r$t|�}nttjdd��}|S)z9
    Returns the version of the running interpreter.
    r��py_version_nodotr�Nrb)rUr\r�rurwr�)rSrKrrrrrr��s

r�cCs,tdd�|D��rd}nd}|�tt|��S)Ncss|]}|dkVqdS)r�Nr)r�vrrrr��r	z!_version_nodot.<locals>.<genexpr>r]rf)r�r�r�r�)rr�seprrrru�srucksXtd|�}t�}|dkr0t|d�D]
}|Vq"nt�D]
}|Vq6t�D]
}|VqHdS)z�
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    �sys_tagsrr�N)rUr�r�r�r�)rSrKr�rDrrrr=�s



r=)F)F)NNN)NNN)NNN)NN)K�
__future__r�distutils.utilr.�importlib.machineryrr��impZget_suffixes�loggingr�r*r�r�rwrVr�Z_typingrr�typingr
rrrrrrrrrr��
PythonVersionr�ZGlibcVersion�	getLoggerr8rXr8�maxsize�_32_BIT_INTERPRETER�objectr!rJrUr\r`rer�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrr r1r2r�r�r�rur=rrrr�<module>s�0

�	
7



&�
<�
�

#@
	!

	site-packages/pkg_resources/_vendor/packaging/__pycache__/requirements.cpython-39.pyc000064400000010051150564745010025107 0ustar00a

(�ReA�@s�ddlmZmZmZddlZddlZddlmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZddlmZddlmZddlmZdd	lmZmZdd
lmZmZmZer�ddlmZGdd
�d
e �Z!eej"ej#�Z$ed��%�Z&ed��%�Z'ed��%�Z(ed��%�Z)ed��%�Z*ed��%�Z+ed��%�Z,ed�Z-e$ee-�e$BZ.ee$ee.��Z/e/d�Z0e/Z1ed�d�Z2e,e2Z3e1ee*e1�Z4e&e
e4�e'd�Z5eej6ej7ej8B�Z9eej6ej7ej8B�Z:e9e:AZ;ee;ee*e;�ddd�d�Z<e
e(e<e)e<B�Z=e=�>dd��e	e=�d�Z?e?�>d d��e	e��d!�Ze�>d"d��e+Z@e@eZAe?e
eA�ZBe3e
eA�ZCe0e
e5�eCeBBZDeeDeZEeE�Fd#�Gd$d%�d%eG�ZHdS)&�)�absolute_import�division�print_functionN)�stringStart�	stringEnd�originalTextFor�ParseException)�
ZeroOrMore�Word�Optional�Regex�Combine)�Literal)�parse�)�
TYPE_CHECKING)�MARKER_EXPR�Marker)�LegacySpecifier�	Specifier�SpecifierSet)�Listc@seZdZdZdS)�InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/requirements.pyrsr�[�]�(�)�,�;�@z-_.�namez[^ ]+�url�extrasF)�
joinString�adjacent�	_raw_speccCs
|jpdS)N�)r+��s�l�trrr�<lambda>;�r1�	specifiercCs|dS)Nrrr-rrrr1>r2�markercCst||j|j��S)N)r�_original_start�
_original_endr-rrrr1Br2zx[]c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�Requirementz�Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    c
Cszt�|�}WnFtyT}z.td�||j|jd�|j���WYd}~n
d}~00|j|_|jr�t	�	|j�}|j
dkr�t	�|�|jkr�td��n(|j
r�|jr�|j
s�|js�td�|j���|j|_nd|_t
|jr�|j��ng�|_t|j�|_|jr�|jnd|_dS)NzParse error at "{0!r}": {1}��filezInvalid URL givenzInvalid URL: {0})�REQUIREMENT�parseStringrr�format�loc�msgr&r'�urlparse�scheme�
urlunparse�netloc�setr(�asListrr3r4)�self�requirement_string�req�e�
parsed_urlrrr�__init___s2��

��
zRequirement.__init__cCs�|jg}|jr*|�d�d�t|j����|jr@|�t|j��|jrh|�d�|j��|j	rh|�d�|j	r�|�d�|j	��d�|�S)Nz[{0}]r#z@ {0}� z; {0}r,)
r&r(�appendr<�join�sortedr3�strr'r4)rE�partsrrr�__str__{s
zRequirement.__str__cCsd�t|��S)Nz<Requirement({0!r})>)r<rO)rErrr�__repr__�szRequirement.__repr__N)rrrrrJrQrRrrrrr7Rsr7)I�
__future__rrr�string�reZpkg_resources.extern.pyparsingrrrrr	r
rrr
r�L�urllibrr?Z_typingr�markersrr�
specifiersrrr�typingr�
ValueErrorr�
ascii_letters�digits�ALPHANUM�suppress�LBRACKET�RBRACKET�LPAREN�RPAREN�COMMA�	SEMICOLON�AT�PUNCTUATION�IDENTIFIER_END�
IDENTIFIER�NAME�EXTRA�URI�URL�EXTRAS_LIST�EXTRAS�
_regex_str�VERBOSE�
IGNORECASE�VERSION_PEP440�VERSION_LEGACY�VERSION_ONE�VERSION_MANY�
_VERSION_SPEC�setParseAction�VERSION_SPEC�MARKER_SEPARATOR�MARKER�VERSION_AND_MARKER�URL_AND_MARKER�NAMED_REQUIREMENTr:r;�objectr7rrrr�<module>sj���
site-packages/pkg_resources/_vendor/packaging/__pycache__/_structures.cpython-39.pyc000064400000005611150564745010024754 0ustar00a

(�Re��@sDddlmZmZmZGdd�de�Ze�ZGdd�de�Ze�ZdS)�)�absolute_import�division�print_functionc@sTeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZdS)�InfinityTypecCsdS)N�Infinity���selfrr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/_structures.py�__repr__szInfinityType.__repr__cCstt|��S�N��hash�reprrrrr
�__hash__szInfinityType.__hash__cCsdS�NFr�r	�otherrrr
�__lt__szInfinityType.__lt__cCsdSrrrrrr
�__le__szInfinityType.__le__cCst||j�Sr��
isinstance�	__class__rrrr
�__eq__szInfinityType.__eq__cCst||j�Srrrrrr
�__ne__szInfinityType.__ne__cCsdS�NTrrrrr
�__gt__ szInfinityType.__gt__cCsdSrrrrrr
�__ge__$szInfinityType.__ge__cCstSr)�NegativeInfinityrrrr
�__neg__(szInfinityType.__neg__N��__name__�
__module__�__qualname__rrrrrrrrrrrrr
rsrc@sTeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZdS)�NegativeInfinityTypecCsdS)Nz	-Infinityrrrrr
r1szNegativeInfinityType.__repr__cCstt|��Srr
rrrr
r5szNegativeInfinityType.__hash__cCsdSrrrrrr
r9szNegativeInfinityType.__lt__cCsdSrrrrrr
r=szNegativeInfinityType.__le__cCst||j�Srrrrrr
rAszNegativeInfinityType.__eq__cCst||j�Srrrrrr
rEszNegativeInfinityType.__ne__cCsdSrrrrrr
rIszNegativeInfinityType.__gt__cCsdSrrrrrr
rMszNegativeInfinityType.__ge__cCstSr)rrrrr
rQszNegativeInfinityType.__neg__Nr rrrr
r$0sr$N)	�
__future__rrr�objectrrr$rrrrr
�<module>s&&site-packages/pkg_resources/_vendor/packaging/__pycache__/version.cpython-39.pyc000064400000032075150564745010024063 0ustar00a

(�Ren<�	@s�ddlmZmZmZddlZddlZddlZddlmZm	Z	ddl
mZe�r.ddlm
Z
mZmZmZmZmZmZddlmZmZeeefZeeeeeffZeeeefZeeeeeeeefeeeffdffZeeeedfeeeefZeeeedffZe
eeefeeefgefZgd	�Z e�!d
gd��Z"dd
�Z#Gdd�de$�Z%Gdd�de&�Z'Gdd�de'�Z(e�)dej*�Z+dddddd�Z,dd�Z-dd�Z.dZ/Gdd�de'�Z0d d!�Z1e�)d"�Z2d#d$�Z3d%d&�Z4dS)'�)�absolute_import�division�print_functionN�)�Infinity�NegativeInfinity)�
TYPE_CHECKING)�Callable�Iterator�List�Optional�SupportsInt�Tuple�Union)�InfinityType�NegativeInfinityType.)�parse�Version�
LegacyVersion�InvalidVersion�VERSION_PATTERN�_Version)�epoch�release�dev�pre�post�localcCs*z
t|�WSty$t|�YS0dS)z�
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rrr)�version�r��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/version.pyr0s
rc@seZdZdZdS)rzF
    An invalid version was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__rrrr r=src@sPeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�ZdS)�_BaseVersionNcCs
t|j�S�N)�hash�_key��selfrrr �__hash__Fsz_BaseVersion.__hash__cCs|�|dd��S)NcSs||kSr&r��s�orrr �<lambda>L�z%_BaseVersion.__lt__.<locals>.<lambda>��_compare�r*�otherrrr �__lt__Jsz_BaseVersion.__lt__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Pr0z%_BaseVersion.__le__.<locals>.<lambda>r1r3rrr �__le__Nsz_BaseVersion.__le__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Tr0z%_BaseVersion.__eq__.<locals>.<lambda>r1r3rrr �__eq__Rsz_BaseVersion.__eq__cCs|�|dd��S)NcSs||kSr&rr,rrr r/Xr0z%_BaseVersion.__ge__.<locals>.<lambda>r1r3rrr �__ge__Vsz_BaseVersion.__ge__cCs|�|dd��S)NcSs||kSr&rr,rrr r/\r0z%_BaseVersion.__gt__.<locals>.<lambda>r1r3rrr �__gt__Zsz_BaseVersion.__gt__cCs|�|dd��S)NcSs||kSr&rr,rrr r/`r0z%_BaseVersion.__ne__.<locals>.<lambda>r1r3rrr �__ne__^sz_BaseVersion.__ne__cCst|t�stS||j|j�Sr&)�
isinstancer%�NotImplementedr()r*r4�methodrrr r2bs
z_BaseVersion._compare)r!r"r#r(r+r5r6r7r8r9r:r2rrrr r%Csr%c@s�eZdZdd�Zdd�Zdd�Zedd��Zed	d
��Zedd��Z	ed
d��Z
edd��Zedd��Zedd��Z
edd��Zedd��Zedd��Zedd��ZdS)rcCst|�|_t|j�|_dSr&)�str�_version�_legacy_cmpkeyr()r*rrrr �__init__ks
zLegacyVersion.__init__cCs|jSr&�r?r)rrr �__str__pszLegacyVersion.__str__cCsd�tt|���S)Nz<LegacyVersion({0})>��format�reprr>r)rrr �__repr__tszLegacyVersion.__repr__cCs|jSr&rBr)rrr �publicxszLegacyVersion.publiccCs|jSr&rBr)rrr �base_version}szLegacyVersion.base_versioncCsdS)N���rr)rrr r�szLegacyVersion.epochcCsdSr&rr)rrr r�szLegacyVersion.releasecCsdSr&rr)rrr r�szLegacyVersion.precCsdSr&rr)rrr r�szLegacyVersion.postcCsdSr&rr)rrr r�szLegacyVersion.devcCsdSr&rr)rrr r�szLegacyVersion.localcCsdS�NFrr)rrr �
is_prerelease�szLegacyVersion.is_prereleasecCsdSrKrr)rrr �is_postrelease�szLegacyVersion.is_postreleasecCsdSrKrr)rrr �
is_devrelease�szLegacyVersion.is_devreleaseN)r!r"r#rArCrG�propertyrHrIrrrrrrrLrMrNrrrr rjs2









rz(\d+ | [a-z]+ | \.| -)�czfinal-�@)r�preview�-�rcrccs\t�|�D]F}t�||�}|r
|dkr(q
|dd�dvrF|�d�Vq
d|Vq
dVdS)N�.r�
0123456789��*�*final)�_legacy_version_component_re�split�_legacy_version_replacement_map�get�zfill)r-�partrrr �_parse_version_parts�sr`cCsvd}g}t|���D]T}|�d�r^|dkrD|rD|ddkrD|��q*|r^|ddkr^|��qD|�|�q|t|�fS)NrJrXrYz*final-�00000000)r`�lower�
startswith�pop�append�tuple)rr�partsr_rrr r@�s


r@a�
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@s�eZdZe�dedejejB�Zdd�Z	dd�Z
dd�Zed	d
��Z
edd��Zed
d��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd��Zedd ��Zed!d"��Zed#d$��Zd%S)&rz^\s*z\s*$c
Cs�|j�|�}|std�|���t|�d�r8t|�d��ndtdd�|�d��d�D��t	|�d�|�d	��t	|�d
�|�d�p�|�d��t	|�d
�|�d��t
|�d��d�|_t|jj
|jj|jj|jj|jj|jj�|_dS)NzInvalid version: '{0}'rrcss|]}t|�VqdSr&)�int��.0�irrr �	<genexpr>r0z#Version.__init__.<locals>.<genexpr>rrU�pre_l�pre_n�post_l�post_n1�post_n2�dev_l�dev_nr)rrrrrr)�_regex�searchrrEr�grouprhrfr[�_parse_letter_version�_parse_local_versionr?�_cmpkeyrrrrrrr()r*r�matchrrr rAs*���zVersion.__init__cCsd�tt|���S)Nz<Version({0})>rDr)rrr rG-szVersion.__repr__cCs�g}|jdkr |�d�|j��|�d�dd�|jD���|jdurb|�d�dd�|jD���|jdur~|�d�|j��|jdur�|�d	�|j��|jdur�|�d
�|j��d�|�S)Nr�{0}!rUcss|]}t|�VqdSr&�r>�rj�xrrr rl:r0z"Version.__str__.<locals>.<genexpr>�css|]}t|�VqdSr&r|r}rrr rl>r0z.post{0}z.dev{0}z+{0})	rrerE�joinrrrrr�r*rgrrr rC1s




zVersion.__str__cCs|jj}|Sr&)r?r)r*�_epochrrr rNsz
Version.epochcCs|jj}|Sr&)r?r)r*�_releaserrr rTszVersion.releasecCs|jj}|Sr&)r?r)r*�_prerrr rZszVersion.precCs|jjr|jjdSdS�Nr)r?rr)rrr r`szVersion.postcCs|jjr|jjdSdSr�)r?rr)rrr reszVersion.devcCs(|jjr d�dd�|jjD��SdSdS)NrUcss|]}t|�VqdSr&r|r}rrr rlnr0z Version.local.<locals>.<genexpr>)r?rr�r)rrr rjsz
Version.localcCst|��dd�dS)N�+rr)r>r[r)rrr rHrszVersion.publiccCsFg}|jdkr |�d�|j��|�d�dd�|jD���d�|�S)Nrr{rUcss|]}t|�VqdSr&r|r}rrr rl�r0z'Version.base_version.<locals>.<genexpr>r)rrerEr�rr�rrr rIws

zVersion.base_versioncCs|jdup|jduSr&)rrr)rrr rL�szVersion.is_prereleasecCs
|jduSr&)rr)rrr rM�szVersion.is_postreleasecCs
|jduSr&)rr)rrr rN�szVersion.is_devreleasecCst|j�dkr|jdSdS)Nrr��lenrr)rrr �major�sz
Version.majorcCst|j�dkr|jdSdS)N�rrr�r)rrr �minor�sz
Version.minorcCst|j�dkr|jdSdS)N�r�rr�r)rrr �micro�sz
Version.microN)r!r"r#�re�compiler�VERBOSE�
IGNORECASErtrArGrCrOrrrrrrrHrIrLrMrNr�r�r�rrrr rs@













rcCsv|rZ|durd}|��}|dkr&d}n(|dkr4d}n|dvrBd}n|dvrNd	}|t|�fS|sr|rrd	}|t|�fSdS)
Nr�alpha�a�beta�b)rPrrRrT)�rev�rr)rbrh)�letter�numberrrr rw�s"rwz[\._-]cCs$|dur tdd�t�|�D��SdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|��s|��nt|�VqdSr&)�isdigitrbrh)rjr_rrr rl�s�z'_parse_local_version.<locals>.<genexpr>)rf�_local_version_separatorsr[)rrrr rx�s
�rxcCs�tttt�dd�t|�����}|dur>|dur>|dur>t}n|durLt}n|}|dur^t}n|}|durpt}	n|}	|dur�t}
ntdd�|D��}
|||||	|
fS)NcSs|dkS)Nrr)r~rrr r/�r0z_cmpkey.<locals>.<lambda>css(|] }t|t�r|dfnt|fVqdS)rN)r;rhrrirrr rlsz_cmpkey.<locals>.<genexpr>)rf�reversed�list�	itertools�	dropwhilerr)rrrrrrr�r��_post�_dev�_localrrr ry�s(�	�ry)5�
__future__rrr�collectionsr�r��_structuresrrZ_typingr�typingr	r
rrr
rrrr�
InfiniteTypesr>rh�PrePostDevType�SubLocalType�	LocalType�CmpKey�LegacyCmpKey�bool�VersionComparisonMethod�__all__�
namedtuplerr�
ValueErrorr�objectr%rr�r�rZr\r`r@rrrwr�rxryrrrr �<module>sp$

���������
'F�	 &

site-packages/pkg_resources/_vendor/packaging/__pycache__/__about__.cpython-39.pyc000064400000001363150564745010024300 0ustar00a

(�Re��@sDddlmZmZmZgd�ZdZdZdZdZdZ	dZ
d	Zd
e	ZdS)�)�absolute_import�division�print_function)�	__title__�__summary__�__uri__�__version__�
__author__�	__email__�__license__�
__copyright__�	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz20.4z)Donald Stufft and individual contributorszdonald@stufft.iozBSD-2-Clause or Apache-2.0zCopyright 2014-2019 %sN)
�
__future__rrr�__all__rrrrr	r
rr�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/__about__.py�<module>ssite-packages/pkg_resources/_vendor/packaging/__pycache__/markers.cpython-39.pyc000064400000022220150564745010024031 0ustar00a

(�Re.%�	@s�ddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
mZmZddlm
Z
mZmZmZddlmZddlmZddlmZdd	lmZmZer�dd
lmZmZmZmZmZm Z m!Z!ee"e"ge#fZ$gd�Z%Gdd
�d
e&�Z'Gdd�de&�Z(Gdd�de&�Z)Gdd�de*�Z+Gdd�de+�Z,Gdd�de+�Z-Gdd�de+�Z.ed�ed�Bed�Bed�Bed�Bed�Bed �Bed!�Bed"�Bed#�Bed$�Bed%�Bed&�Bed'�Bed(�Bed)�Bed*�Bed+�BZ/d$d#dd ddd,�Z0e/�1d-d.��ed/�ed0�Bed1�Bed2�Bed3�Bed4�Bed5�Bed6�BZ2e2ed7�Bed8�BZ3e3�1d9d.��ed:�ed;�BZ4e4�1d<d.��ed=�ed>�BZ5e/e4BZ6ee6e3e6�Z7e7�1d?d.��ed@��8�Z9edA��8�Z:e�Z;e7ee9e;e:�BZ<e;e<e
e5e;�>ee;eZ=dBdC�Z>dXdEdF�Z?dGd.�dHd.�ej@ejAejBejCejDejEdI�ZFdJdK�ZGGdLdM�dMe*�ZHeH�ZIdNdO�ZJdPdQ�ZKdRdS�ZLdTdU�ZMGdVdW�dWe*�ZNdS)Y�)�absolute_import�division�print_functionN)�ParseException�ParseResults�stringStart�	stringEnd)�
ZeroOrMore�Group�Forward�QuotedString)�Literal�)�string_types)�
TYPE_CHECKING)�	Specifier�InvalidSpecifier)�Any�Callable�Dict�List�Optional�Tuple�Union)�
InvalidMarker�UndefinedComparison�UndefinedEnvironmentName�Marker�default_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N��__name__�
__module__�__qualname__�__doc__�r$r$��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/markers.pyr"src@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrr$r$r$r%r(src@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrr$r$r$r%r.src@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�NodecCs
||_dS�N)�value)�selfr(r$r$r%�__init__6sz
Node.__init__cCs
t|j�Sr')�strr(�r)r$r$r%�__str__:szNode.__str__cCsd�|jjt|��S)Nz<{0}({1!r})>)�format�	__class__r r+r,r$r$r%�__repr__>sz
Node.__repr__cCst�dSr')�NotImplementedErrorr,r$r$r%�	serializeBszNode.serializeN)r r!r"r*r-r0r2r$r$r$r%r&5sr&c@seZdZdd�ZdS)�VariablecCst|�Sr'�r+r,r$r$r%r2HszVariable.serializeN�r r!r"r2r$r$r$r%r3Gsr3c@seZdZdd�ZdS)�ValuecCs
d�|�S)Nz"{0}")r.r,r$r$r%r2NszValue.serializeNr5r$r$r$r%r6Msr6c@seZdZdd�ZdS)�OpcCst|�Sr'r4r,r$r$r%r2TszOp.serializeNr5r$r$r$r%r7Ssr7�implementation_version�platform_python_implementation�implementation_name�python_full_version�platform_release�platform_version�platform_machine�platform_system�python_version�sys_platform�os_name�os.name�sys.platform�platform.version�platform.machine�platform.python_implementation�python_implementation�extra)rCrDrErFrGrHcCstt�|d|d��S�Nr)r3�ALIASES�get��s�l�tr$r$r%�<lambda>u�rQz===�==�>=�<=�!=z~=�>�<�not in�incCst|d�SrJ)r7rMr$r$r%rQ|rR�'�"cCst|d�SrJ)r6rMr$r$r%rQrR�and�orcCst|d�SrJ)�tuplerMr$r$r%rQ�rR�(�)cCs t|t�rdd�|D�S|SdS)NcSsg|]}t|��qSr$)�_coerce_parse_result)�.0�ir$r$r%�
<listcomp>�rRz(_coerce_parse_result.<locals>.<listcomp>)�
isinstancer)�resultsr$r$r%rb�s
rbTcCs�t|tttf�sJ�t|t�rHt|�dkrHt|dttf�rHt|d�St|t�r�dd�|D�}|rnd�|�Sdd�|�dSn"t|t�r�d�dd	�|D��S|SdS)
Nrrcss|]}t|dd�VqdS)F)�firstN)�_format_marker�rc�mr$r$r%�	<genexpr>�rRz!_format_marker.<locals>.<genexpr>� r`racSsg|]}|���qSr$)r2rjr$r$r%re�rRz"_format_marker.<locals>.<listcomp>)rf�listr_r�lenri�join)�markerrh�innerr$r$r%ri�s �
��


ricCs||vSr'r$��lhs�rhsr$r$r%rQ�rRcCs||vSr'r$rsr$r$r%rQ�rR)rZrYrXrUrSrVrTrWcCsjztd�|��|g��}Wnty,Yn0|�|�St�|���}|dur`td�|||���|||�S)N�z#Undefined {0!r} on {1!r} and {2!r}.)	rrpr2r�contains�
_operatorsrLrr.)rt�opru�spec�operr$r$r%�_eval_op�s
�r|c@seZdZdS)�	UndefinedN)r r!r"r$r$r$r%r}�sr}cCs(|�|t�}t|t�r$td�|���|S)Nz/{0!r} does not exist in evaluation environment.)rL�
_undefinedrfr}rr.)�environment�namer(r$r$r%�_get_env�s
�r�c	Cs�gg}|D]�}t|tttf�s"J�t|t�rB|d�t||��q
t|t�r�|\}}}t|t�rtt||j�}|j}n|j}t||j�}|d�t	|||��q
|dvs�J�|dkr
|�g�q
t
dd�|D��S)N���)r]r^r^css|]}t|�VqdSr')�all)rc�itemr$r$r%rlrRz$_evaluate_markers.<locals>.<genexpr>)rfrnr_r�append�_evaluate_markersr3r�r(r|�any)	�markersr�groupsrqrtryru�	lhs_value�	rhs_valuer$r$r%r��s"



r�cCs2d�|�}|j}|dkr.||dt|j�7}|S)Nz{0.major}.{0.minor}.{0.micro}�finalr)r.�releaselevelr+�serial)�info�version�kindr$r$r%�format_full_versions

r�cCsrttd�r ttjj�}tjj}nd}d}||tjt��t�	�t�
�t��t��t��d�
t��dd��tjd�S)N�implementation�0rv�.�)r:r8rBr>r<r?r=r;r9r@rA)�hasattr�sysr�r�r�r��os�platform�machine�release�systemr@rHrp�python_version_tuple)�iverr:r$r$r%rs"

�rc@s.eZdZdd�Zdd�Zdd�Zd
dd	�ZdS)rc
Csbztt�|��|_WnHty\}z0d�|||j|jd��}t|��WYd}~n
d}~00dS)Nz+Invalid marker: {0!r}, parse error at {1!r}�)rb�MARKER�parseString�_markersrr.�locr)r)rq�eZerr_strr$r$r%r*(s�zMarker.__init__cCs
t|j�Sr')rir�r,r$r$r%r-2szMarker.__str__cCsd�t|��S)Nz<Marker({0!r})>)r.r+r,r$r$r%r06szMarker.__repr__NcCs$t�}|dur|�|�t|j|�S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)r�updater�r�)r)r�current_environmentr$r$r%�evaluate:s

zMarker.evaluate)N)r r!r"r*r-r0r�r$r$r$r%r's
r)T)O�
__future__rrr�operatorr�r�r�Zpkg_resources.extern.pyparsingrrrrr	r
rrr
�LZ_compatrZ_typingr�
specifiersrr�typingrrrrrrrr+�bool�Operator�__all__�
ValueErrorrrr�objectr&r3r6r7�VARIABLErK�setParseAction�VERSION_CMP�	MARKER_OP�MARKER_VALUE�BOOLOP�
MARKER_VAR�MARKER_ITEM�suppress�LPAREN�RPAREN�MARKER_EXPR�MARKER_ATOMr�rbri�lt�le�eq�ne�ge�gtrxr|r}r~r�r�r�rrr$r$r$r%�<module>s�$	��������	�
���
�������>�
�	site-packages/pkg_resources/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc000064400000050234150564745010024527 0ustar00a

(�Re�|�@sXddlmZmZmZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZddl
mZddlmZmZmZer�ddlmZmZmZmZmZmZmZmZmZeeefZeeeefZeeege fZ!Gd	d
�d
e"�Z#Gdd�de
ej$e%��Z&Gd
d�de&�Z'Gdd�de'�Z(dd�Z)Gdd�de'�Z*e�+d�Z,dd�Z-dd�Z.Gdd�de&�Z/dS)�)�absolute_import�division�print_functionN�)�string_types�with_metaclass)�
TYPE_CHECKING)�canonicalize_version)�Version�
LegacyVersion�parse)	�List�Dict�Union�Iterable�Iterator�Optional�Callable�Tuple�	FrozenSetc@seZdZdZdS)�InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/specifiers.pyr"src@s�eZdZejdd��Zejdd��Zejdd��Zejdd��Zej	d	d
��Z
e
jdd
��Z
ejdd
d��Zejddd��Z
dS)�
BaseSpecifiercCsdS)z�
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nr��selfrrr�__str__)szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr�__hash__1szBaseSpecifier.__hash__cCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nr�r�otherrrr�__eq__8szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrr"rrr�__ne__@szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrr�prereleasesHszBaseSpecifier.prereleasescCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nr�r�valuerrrr&PsNcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nr�r�itemr&rrr�containsXszBaseSpecifier.containscCsdS)z�
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)r�iterabler&rrr�filter_szBaseSpecifier.filter)N)N)rrr�abc�abstractmethodr r!r$r%�abstractpropertyr&�setterr+r-rrrrr(s 





rc@s�eZdZiZd"dd�Zdd�Zdd�Zed	d
��Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
edd��Zedd��Zedd��Zejdd��Zdd�Zd#dd�Zd$d d!�ZdS)%�_IndividualSpecifier�NcCsF|j�|�}|std�|���|�d���|�d���f|_||_dS)NzInvalid specifier: '{0}'�operator�version)�_regex�searchr�format�group�strip�_spec�_prereleases)r�specr&�matchrrr�__init__ls�z_IndividualSpecifier.__init__cCs0|jdurd�|j�nd}d�|jjt|�|�S)N�, prereleases={0!r}r3z<{0}({1!r}{2})>)r<r8r&�	__class__r�str�r�prerrr�__repr__zs
��z_IndividualSpecifier.__repr__cCsdj|j�S)Nz{0}{1})r8r;rrrrr �sz_IndividualSpecifier.__str__cCs|jdt|jd�fS)Nrr)r;r	rrrr�_canonical_spec�sz$_IndividualSpecifier._canonical_speccCs
t|j�S�N)�hashrFrrrrr!�sz_IndividualSpecifier.__hash__cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrG)�
isinstancerrArBr�NotImplementedrFr"rrrr$�s
z_IndividualSpecifier.__eq__cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrG)rIrrArBrrJr;r"rrrr%�s
z_IndividualSpecifier.__ne__cCst|d�|j|��}|S)Nz_compare_{0})�getattrr8�
_operators)r�op�operator_callablerrr�
_get_operator�s�z"_IndividualSpecifier._get_operatorcCst|ttf�st|�}|SrG)rIrr
r�rr5rrr�_coerce_version�sz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nr�r;rrrrr4�sz_IndividualSpecifier.operatorcCs
|jdS)NrrRrrrrr5�sz_IndividualSpecifier.versioncCs|jSrG�r<rrrrr&�sz _IndividualSpecifier.prereleasescCs
||_dSrGrSr'rrrr&�scCs
|�|�SrG�r+�rr*rrr�__contains__�sz!_IndividualSpecifier.__contains__cCs>|dur|j}|�|�}|jr&|s&dS|�|j�}|||j�S�NF)r&rQ�
is_prereleaserOr4r5)rr*r&�normalized_itemrNrrrr+�s

z_IndividualSpecifier.containsccs�d}g}d|dur|ndi}|D]F}|�|�}|j|fi|��r |jr\|s\|js\|�|�q d}|Vq |s�|r�|D]
}|VqtdS)NFr&T)rQr+rXr&�append)rr,r&�yielded�found_prereleases�kwr5�parsed_versionrrrr-�s"
��z_IndividualSpecifier.filter)r3N)N)N)rrrrLr?rEr �propertyrFr!r$r%rOrQr4r5r&r1rVr+r-rrrrr2hs,







r2c@sveZdZdZe�dedejejB�Zdddddd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�ZdS)�LegacySpecifiera�
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        �^\s*�\s*$�equal�	not_equal�less_than_equal�greater_than_equal�	less_than�greater_than)�==�!=�<=�>=�<�>cCst|t�stt|��}|SrG)rIrrBrPrrrrQ s
zLegacySpecifier._coerce_versioncCs||�|�kSrG�rQ�r�prospectiver=rrr�_compare_equal&szLegacySpecifier._compare_equalcCs||�|�kSrGrorprrr�_compare_not_equal*sz"LegacySpecifier._compare_not_equalcCs||�|�kSrGrorprrr�_compare_less_than_equal.sz(LegacySpecifier._compare_less_than_equalcCs||�|�kSrGrorprrr�_compare_greater_than_equal2sz+LegacySpecifier._compare_greater_than_equalcCs||�|�kSrGrorprrr�_compare_less_than6sz"LegacySpecifier._compare_less_thancCs||�|�kSrGrorprrr�_compare_greater_than:sz%LegacySpecifier._compare_greater_thanN)rrr�
_regex_str�re�compile�VERBOSE�
IGNORECASEr6rLrQrrrsrtrurvrwrrrrr`s �	r`cst����fdd��}|S)Ncst|t�sdS�|||�SrW)rIr
rp��fnrr�wrappedCs
z)_require_version_compare.<locals>.wrapped)�	functools�wraps)r~rrr}r�_require_version_compare?sr�c	@s�eZdZdZe�dedejejB�Zdddddd	d
dd�Z	e
d
d��Ze
dd��Ze
dd��Z
e
dd��Ze
dd��Ze
dd��Ze
dd��Zdd�Zedd��Zejdd��Zd S)!�	Specifiera
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        rarb�
compatiblercrdrerfrgrh�	arbitrary)�~=rirjrkrlrmrn�===cCsNd�tt�dd�t|���dd��}|d7}|�d�||�oL|�d�||�S)N�.cSs|�d�o|�d�S)N�post�dev)�
startswith��xrrr�<lambda>��z/Specifier._compare_compatible.<locals>.<lambda>����.*rlri)�join�list�	itertools�	takewhile�_version_splitrO)rrqr=�prefixrrr�_compare_compatible�s
����
�zSpecifier._compare_compatiblec	Csz|�d�rVt|j�}t|dd��}tt|��}|dt|��}t||�\}}||kSt|�}|jsnt|j�}||kSdS)Nr����)�endswithr
�publicr�rB�len�_pad_version�local)	rrqr=�
split_spec�split_prospective�shortened_prospective�padded_spec�padded_prospective�spec_versionrrrrr�s

�
zSpecifier._compare_equalcCs|�||�SrG)rrrprrrrs�szSpecifier._compare_not_equalcCst|j�t|�kSrG�r
r�rprrrrtsz"Specifier._compare_less_than_equalcCst|j�t|�kSrGr�rprrrru
sz%Specifier._compare_greater_than_equalcCs<t|�}||ksdS|js8|jr8t|j�t|j�kr8dSdS�NFT)r
rX�base_version�rrq�spec_strr=rrrrvszSpecifier._compare_less_thancCs^t|�}||ksdS|js8|jr8t|j�t|j�kr8dS|jdurZt|j�t|j�krZdSdSr�)r
�is_postreleaser�r�r�rrrrw1s
zSpecifier._compare_greater_thancCst|���t|���kSrG)rB�lowerrprrr�_compare_arbitraryRszSpecifier._compare_arbitrarycCsR|jdur|jS|j\}}|dvrN|dkr@|�d�r@|dd�}t|�jrNdSdS)N)rirlrkr�r�rir�r�TF)r<r;r�rrX)rr4r5rrrr&Vs


zSpecifier.prereleasescCs
||_dSrGrSr'rrrr&psN)rrrrxryrzr{r|r6rLr�r�rrrsrtrurvrwr�r_r&r1rrrrr�Ms<]�

(




 
r�z^([0-9]+)((?:a|b|c|rc)[0-9]+)$cCs@g}|�d�D],}t�|�}|r0|�|���q|�|�q|S)Nr�)�split�
_prefix_regexr7�extend�groupsrZ)r5�resultr*r>rrrr�ys
r�c
Cs�gg}}|�tt�dd�|���|�tt�dd�|���|�|t|d�d��|�|t|d�d��|�ddgtdt|d�t|d���|�ddgtdt|d�t|d���ttj|��ttj|��fS)NcSs|��SrG��isdigitr�rrrr��r�z_pad_version.<locals>.<lambda>cSs|��SrGr�r�rrrr��r�rr�0)rZr�r�r�r��insert�max�chain)�left�right�
left_split�right_splitrrrr��s
,,r�c@s�eZdZddd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Zdd�Z	dd�Z
dd�Zedd��Z
e
jdd��Z
dd�Zddd�Zd dd�ZdS)!�SpecifierSetr3Nc	Csldd�|�d�D�}t�}|D]8}z|�t|��WqtyT|�t|��Yq0qt|�|_||_dS)NcSsg|]}|��r|���qSr)r:��.0�srrr�
<listcomp>�r�z)SpecifierSet.__init__.<locals>.<listcomp>�,)	r��set�addr�rr`�	frozenset�_specsr<)r�
specifiersr&�split_specifiers�parsed�	specifierrrrr?�s
zSpecifierSet.__init__cCs*|jdurd�|j�nd}d�t|�|�S)Nr@r3z<SpecifierSet({0!r}{1})>)r<r8r&rBrCrrrrE�s
��zSpecifierSet.__repr__cCsd�tdd�|jD���S)Nr�css|]}t|�VqdSrG)rBr�rrr�	<genexpr>�r�z'SpecifierSet.__str__.<locals>.<genexpr>)r��sortedr�rrrrr �szSpecifierSet.__str__cCs
t|j�SrG)rHr�rrrrr!�szSpecifierSet.__hash__cCs�t|t�rt|�}nt|t�s"tSt�}t|j|jB�|_|jdurX|jdurX|j|_n<|jdurv|jdurv|j|_n|j|jkr�|j|_ntd��|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)rIrr�rJr�r�r<�
ValueError)rr#r�rrr�__and__�s 





�zSpecifierSet.__and__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrG�rIrr2r�rBrJr�r"rrrr$�s

zSpecifierSet.__eq__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrGr�r"rrrr%�s

zSpecifierSet.__ne__cCs
t|j�SrG)r�r�rrrr�__len__�szSpecifierSet.__len__cCs
t|j�SrG)�iterr�rrrr�__iter__�szSpecifierSet.__iter__cCs.|jdur|jS|jsdStdd�|jD��S)Ncss|]}|jVqdSrG�r&r�rrrr�r�z+SpecifierSet.prereleases.<locals>.<genexpr>)r<r��anyrrrrr&�s

zSpecifierSet.prereleasescCs
||_dSrGrSr'rrrr&scCs
|�|�SrGrTrUrrrrVszSpecifierSet.__contains__csLt�ttf�st����dur$|j��s2�jr2dSt��fdd�|jD��S)NFc3s|]}|j��d�VqdS)r�NrTr��r*r&rrr�*r�z(SpecifierSet.contains.<locals>.<genexpr>)rIrr
rr&rX�allr�r)rr�rr+s
zSpecifierSet.containscCs�|dur|j}|jr6|jD]}|j|t|�d�}q|Sg}g}|D]P}t|ttf�s^t|�}n|}t|t�rnqB|jr�|s�|s�|�	|�qB|�	|�qB|s�|r�|dur�|S|SdS)Nr�)
r&r�r-�boolrIrr
rrXrZ)rr,r&r=�filteredr\r*r^rrrr-,s*




zSpecifierSet.filter)r3N)N)N)rrrr?rEr r!r�r$r%r�r�r_r&r1rVr+r-rrrrr��s"

		


�r�)0�
__future__rrrr.r�r�ryZ_compatrrZ_typingr�utilsr	r5r
rr�typingr
rrrrrrrr�
ParsedVersionrB�UnparsedVersionr��CallableOperatorr�r�ABCMeta�objectrr2r`r�r�rzr�r�r�r�rrrr�<module>s4,@ 8+
site-packages/pkg_resources/_vendor/packaging/__pycache__/_typing.cpython-39.pyc000064400000003010150564745010024032 0ustar00a

(�Re�@s.dZddgZdZer"ddlmZndd�ZdS)a;For neatly implementing static typing in packaging.

`mypy` - the static type analysis tool we use - uses the `typing` module, which
provides core functionality fundamental to mypy's functioning.

Generally, `typing` would be imported at runtime and used in that fashion -
it acts as a no-op at runtime and does not have any run-time overhead by
design.

As it turns out, `typing` is not vendorable - it uses separate sources for
Python 2/Python 3. Thus, this codebase can not expect it to be present.
To work around this, mypy allows the typing import to be behind a False-y
optional to prevent it from running at runtime and type-comments can be used
to remove the need for the types to be accessible directly during runtime.

This module provides the False-y guard in a nicely named fashion so that a
curious maintainer can reach here to read this.

In packaging, all static-typing related imports should be guarded as follows:

    from packaging._typing import TYPE_CHECKING

    if TYPE_CHECKING:
        from typing import ...

Ref: https://github.com/python/mypy/issues/3216
�
TYPE_CHECKING�castF�)rcCs|S)N�)�type_�valuerr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/_typing.pyr/sN)�__doc__�__all__r�typingrrrrr�<module>s
site-packages/pkg_resources/_vendor/packaging/__pycache__/_compat.cpython-39.pyc000064400000002257150564745010024017 0ustar00a

(�Reh�@s~ddlmZmZmZddlZddlmZerDddlmZm	Z	m
Z
mZejddkZ
ejddkZerlefZnefZdd	�ZdS)
�)�absolute_import�division�print_functionN�)�
TYPE_CHECKING)�Any�Dict�Tuple�Type��cs&G��fdd�d��}t�|ddi�S)z/
    Create a base class with a metaclass.
    cseZdZ��fdd�ZdS)z!with_metaclass.<locals>.metaclasscs�|�|�S)N�)�cls�name�
this_bases�d��bases�metar
��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/_compat.py�__new__"sz)with_metaclass.<locals>.metaclass.__new__N)�__name__�
__module__�__qualname__rr
rr
r�	metaclass!sr�temporary_classr
)�typer)rrrr
rr�with_metaclasssr)�
__future__rrr�sysZ_typingr�typingrrr	r
�version_info�PY2�PY3�str�string_types�
basestringrr
r
r
r�<module>ssite-packages/pkg_resources/_vendor/packaging/__pycache__/__init__.cpython-39.pyc000064400000001131150564745010024122 0ustar00a

(�Re2�@sHddlmZmZmZddlmZmZmZmZm	Z	m
Z
mZmZgd�Z
dS)�)�absolute_import�division�print_function�)�
__author__�
__copyright__�	__email__�__license__�__summary__�	__title__�__uri__�__version__)rr
rr
rrr	rN)�
__future__rrr�	__about__rrrr	r
rrr
�__all__�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/_vendor/packaging/__init__.py�<module>s(site-packages/pkg_resources/_vendor/packaging/markers.py000064400000022456150564745010017555 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import operator
import os
import platform
import sys

from pkg_resources.extern.pyparsing import ParseException, ParseResults, stringStart, stringEnd
from pkg_resources.extern.pyparsing import ZeroOrMore, Group, Forward, QuotedString
from pkg_resources.extern.pyparsing import Literal as L  # noqa

from ._compat import string_types
from ._typing import TYPE_CHECKING
from .specifiers import Specifier, InvalidSpecifier

if TYPE_CHECKING:  # pragma: no cover
    from typing import Any, Callable, Dict, List, Optional, Tuple, Union

    Operator = Callable[[str, str], bool]


__all__ = [
    "InvalidMarker",
    "UndefinedComparison",
    "UndefinedEnvironmentName",
    "Marker",
    "default_environment",
]


class InvalidMarker(ValueError):
    """
    An invalid marker was found, users should refer to PEP 508.
    """


class UndefinedComparison(ValueError):
    """
    An invalid operation was attempted on a value that doesn't support it.
    """


class UndefinedEnvironmentName(ValueError):
    """
    A name was attempted to be used that does not exist inside of the
    environment.
    """


class Node(object):
    def __init__(self, value):
        # type: (Any) -> None
        self.value = value

    def __str__(self):
        # type: () -> str
        return str(self.value)

    def __repr__(self):
        # type: () -> str
        return "<{0}({1!r})>".format(self.__class__.__name__, str(self))

    def serialize(self):
        # type: () -> str
        raise NotImplementedError


class Variable(Node):
    def serialize(self):
        # type: () -> str
        return str(self)


class Value(Node):
    def serialize(self):
        # type: () -> str
        return '"{0}"'.format(self)


class Op(Node):
    def serialize(self):
        # type: () -> str
        return str(self)


VARIABLE = (
    L("implementation_version")
    | L("platform_python_implementation")
    | L("implementation_name")
    | L("python_full_version")
    | L("platform_release")
    | L("platform_version")
    | L("platform_machine")
    | L("platform_system")
    | L("python_version")
    | L("sys_platform")
    | L("os_name")
    | L("os.name")  # PEP-345
    | L("sys.platform")  # PEP-345
    | L("platform.version")  # PEP-345
    | L("platform.machine")  # PEP-345
    | L("platform.python_implementation")  # PEP-345
    | L("python_implementation")  # undocumented setuptools legacy
    | L("extra")  # PEP-508
)
ALIASES = {
    "os.name": "os_name",
    "sys.platform": "sys_platform",
    "platform.version": "platform_version",
    "platform.machine": "platform_machine",
    "platform.python_implementation": "platform_python_implementation",
    "python_implementation": "platform_python_implementation",
}
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))

VERSION_CMP = (
    L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<")
)

MARKER_OP = VERSION_CMP | L("not in") | L("in")
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))

MARKER_VALUE = QuotedString("'") | QuotedString('"')
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))

BOOLOP = L("and") | L("or")

MARKER_VAR = VARIABLE | MARKER_VALUE

MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))

LPAREN = L("(").suppress()
RPAREN = L(")").suppress()

MARKER_EXPR = Forward()
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)

MARKER = stringStart + MARKER_EXPR + stringEnd


def _coerce_parse_result(results):
    # type: (Union[ParseResults, List[Any]]) -> List[Any]
    if isinstance(results, ParseResults):
        return [_coerce_parse_result(i) for i in results]
    else:
        return results


def _format_marker(marker, first=True):
    # type: (Union[List[str], Tuple[Node, ...], str], Optional[bool]) -> str

    assert isinstance(marker, (list, tuple, string_types))

    # Sometimes we have a structure like [[...]] which is a single item list
    # where the single item is itself it's own list. In that case we want skip
    # the rest of this function so that we don't get extraneous () on the
    # outside.
    if (
        isinstance(marker, list)
        and len(marker) == 1
        and isinstance(marker[0], (list, tuple))
    ):
        return _format_marker(marker[0])

    if isinstance(marker, list):
        inner = (_format_marker(m, first=False) for m in marker)
        if first:
            return " ".join(inner)
        else:
            return "(" + " ".join(inner) + ")"
    elif isinstance(marker, tuple):
        return " ".join([m.serialize() for m in marker])
    else:
        return marker


_operators = {
    "in": lambda lhs, rhs: lhs in rhs,
    "not in": lambda lhs, rhs: lhs not in rhs,
    "<": operator.lt,
    "<=": operator.le,
    "==": operator.eq,
    "!=": operator.ne,
    ">=": operator.ge,
    ">": operator.gt,
}  # type: Dict[str, Operator]


def _eval_op(lhs, op, rhs):
    # type: (str, Op, str) -> bool
    try:
        spec = Specifier("".join([op.serialize(), rhs]))
    except InvalidSpecifier:
        pass
    else:
        return spec.contains(lhs)

    oper = _operators.get(op.serialize())  # type: Optional[Operator]
    if oper is None:
        raise UndefinedComparison(
            "Undefined {0!r} on {1!r} and {2!r}.".format(op, lhs, rhs)
        )

    return oper(lhs, rhs)


class Undefined(object):
    pass


_undefined = Undefined()


def _get_env(environment, name):
    # type: (Dict[str, str], str) -> str
    value = environment.get(name, _undefined)  # type: Union[str, Undefined]

    if isinstance(value, Undefined):
        raise UndefinedEnvironmentName(
            "{0!r} does not exist in evaluation environment.".format(name)
        )

    return value


def _evaluate_markers(markers, environment):
    # type: (List[Any], Dict[str, str]) -> bool
    groups = [[]]  # type: List[List[bool]]

    for marker in markers:
        assert isinstance(marker, (list, tuple, string_types))

        if isinstance(marker, list):
            groups[-1].append(_evaluate_markers(marker, environment))
        elif isinstance(marker, tuple):
            lhs, op, rhs = marker

            if isinstance(lhs, Variable):
                lhs_value = _get_env(environment, lhs.value)
                rhs_value = rhs.value
            else:
                lhs_value = lhs.value
                rhs_value = _get_env(environment, rhs.value)

            groups[-1].append(_eval_op(lhs_value, op, rhs_value))
        else:
            assert marker in ["and", "or"]
            if marker == "or":
                groups.append([])

    return any(all(item) for item in groups)


def format_full_version(info):
    # type: (sys._version_info) -> str
    version = "{0.major}.{0.minor}.{0.micro}".format(info)
    kind = info.releaselevel
    if kind != "final":
        version += kind[0] + str(info.serial)
    return version


def default_environment():
    # type: () -> Dict[str, str]
    if hasattr(sys, "implementation"):
        # Ignoring the `sys.implementation` reference for type checking due to
        # mypy not liking that the attribute doesn't exist in Python 2.7 when
        # run with the `--py27` flag.
        iver = format_full_version(sys.implementation.version)  # type: ignore
        implementation_name = sys.implementation.name  # type: ignore
    else:
        iver = "0"
        implementation_name = ""

    return {
        "implementation_name": implementation_name,
        "implementation_version": iver,
        "os_name": os.name,
        "platform_machine": platform.machine(),
        "platform_release": platform.release(),
        "platform_system": platform.system(),
        "platform_version": platform.version(),
        "python_full_version": platform.python_version(),
        "platform_python_implementation": platform.python_implementation(),
        "python_version": ".".join(platform.python_version_tuple()[:2]),
        "sys_platform": sys.platform,
    }


class Marker(object):
    def __init__(self, marker):
        # type: (str) -> None
        try:
            self._markers = _coerce_parse_result(MARKER.parseString(marker))
        except ParseException as e:
            err_str = "Invalid marker: {0!r}, parse error at {1!r}".format(
                marker, marker[e.loc : e.loc + 8]
            )
            raise InvalidMarker(err_str)

    def __str__(self):
        # type: () -> str
        return _format_marker(self._markers)

    def __repr__(self):
        # type: () -> str
        return "<Marker({0!r})>".format(str(self))

    def evaluate(self, environment=None):
        # type: (Optional[Dict[str, str]]) -> bool
        """Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        """
        current_environment = default_environment()
        if environment is not None:
            current_environment.update(environment)

        return _evaluate_markers(self._markers, current_environment)
site-packages/pkg_resources/_vendor/packaging/__about__.py000064400000001340150564745010020004 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]

__title__ = "packaging"
__summary__ = "Core utilities for Python packages"
__uri__ = "https://github.com/pypa/packaging"

__version__ = "20.4"

__author__ = "Donald Stufft and individual contributors"
__email__ = "donald@stufft.io"

__license__ = "BSD-2-Clause or Apache-2.0"
__copyright__ = "Copyright 2014-2019 %s" % __author__
site-packages/pkg_resources/_vendor/packaging/utils.py000064400000003423150564745010017242 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import re

from ._typing import TYPE_CHECKING, cast
from .version import InvalidVersion, Version

if TYPE_CHECKING:  # pragma: no cover
    from typing import NewType, Union

    NormalizedName = NewType("NormalizedName", str)

_canonicalize_regex = re.compile(r"[-_.]+")


def canonicalize_name(name):
    # type: (str) -> NormalizedName
    # This is taken from PEP 503.
    value = _canonicalize_regex.sub("-", name).lower()
    return cast("NormalizedName", value)


def canonicalize_version(_version):
    # type: (str) -> Union[Version, str]
    """
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    """

    try:
        version = Version(_version)
    except InvalidVersion:
        # Legacy versions cannot be normalized
        return _version

    parts = []

    # Epoch
    if version.epoch != 0:
        parts.append("{0}!".format(version.epoch))

    # Release segment
    # NB: This strips trailing '.0's to normalize
    parts.append(re.sub(r"(\.0)+$", "", ".".join(str(x) for x in version.release)))

    # Pre-release
    if version.pre is not None:
        parts.append("".join(str(x) for x in version.pre))

    # Post-release
    if version.post is not None:
        parts.append(".post{0}".format(version.post))

    # Development release
    if version.dev is not None:
        parts.append(".dev{0}".format(version.dev))

    # Local version segment
    if version.local is not None:
        parts.append("+{0}".format(version.local))

    return "".join(parts)
site-packages/pkg_resources/_vendor/packaging/_compat.py000064400000002150150564745010017520 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import sys

from ._typing import TYPE_CHECKING

if TYPE_CHECKING:  # pragma: no cover
    from typing import Any, Dict, Tuple, Type


PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3

# flake8: noqa

if PY3:
    string_types = (str,)
else:
    string_types = (basestring,)


def with_metaclass(meta, *bases):
    # type: (Type[Any], Tuple[Type[Any], ...]) -> Any
    """
    Create a base class with a metaclass.
    """
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(meta):  # type: ignore
        def __new__(cls, name, this_bases, d):
            # type: (Type[Any], str, Tuple[Any], Dict[Any, Any]) -> Any
            return meta(name, bases, d)

    return type.__new__(metaclass, "temporary_class", (), {})
site-packages/pkg_resources/_vendor/packaging/version.py000064400000036156150564745010017600 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import collections
import itertools
import re

from ._structures import Infinity, NegativeInfinity
from ._typing import TYPE_CHECKING

if TYPE_CHECKING:  # pragma: no cover
    from typing import Callable, Iterator, List, Optional, SupportsInt, Tuple, Union

    from ._structures import InfinityType, NegativeInfinityType

    InfiniteTypes = Union[InfinityType, NegativeInfinityType]
    PrePostDevType = Union[InfiniteTypes, Tuple[str, int]]
    SubLocalType = Union[InfiniteTypes, int, str]
    LocalType = Union[
        NegativeInfinityType,
        Tuple[
            Union[
                SubLocalType,
                Tuple[SubLocalType, str],
                Tuple[NegativeInfinityType, SubLocalType],
            ],
            ...,
        ],
    ]
    CmpKey = Tuple[
        int, Tuple[int, ...], PrePostDevType, PrePostDevType, PrePostDevType, LocalType
    ]
    LegacyCmpKey = Tuple[int, Tuple[str, ...]]
    VersionComparisonMethod = Callable[
        [Union[CmpKey, LegacyCmpKey], Union[CmpKey, LegacyCmpKey]], bool
    ]

__all__ = ["parse", "Version", "LegacyVersion", "InvalidVersion", "VERSION_PATTERN"]


_Version = collections.namedtuple(
    "_Version", ["epoch", "release", "dev", "pre", "post", "local"]
)


def parse(version):
    # type: (str) -> Union[LegacyVersion, Version]
    """
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    """
    try:
        return Version(version)
    except InvalidVersion:
        return LegacyVersion(version)


class InvalidVersion(ValueError):
    """
    An invalid version was found, users should refer to PEP 440.
    """


class _BaseVersion(object):
    _key = None  # type: Union[CmpKey, LegacyCmpKey]

    def __hash__(self):
        # type: () -> int
        return hash(self._key)

    def __lt__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s < o)

    def __le__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s <= o)

    def __eq__(self, other):
        # type: (object) -> bool
        return self._compare(other, lambda s, o: s == o)

    def __ge__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s >= o)

    def __gt__(self, other):
        # type: (_BaseVersion) -> bool
        return self._compare(other, lambda s, o: s > o)

    def __ne__(self, other):
        # type: (object) -> bool
        return self._compare(other, lambda s, o: s != o)

    def _compare(self, other, method):
        # type: (object, VersionComparisonMethod) -> Union[bool, NotImplemented]
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return method(self._key, other._key)


class LegacyVersion(_BaseVersion):
    def __init__(self, version):
        # type: (str) -> None
        self._version = str(version)
        self._key = _legacy_cmpkey(self._version)

    def __str__(self):
        # type: () -> str
        return self._version

    def __repr__(self):
        # type: () -> str
        return "<LegacyVersion({0})>".format(repr(str(self)))

    @property
    def public(self):
        # type: () -> str
        return self._version

    @property
    def base_version(self):
        # type: () -> str
        return self._version

    @property
    def epoch(self):
        # type: () -> int
        return -1

    @property
    def release(self):
        # type: () -> None
        return None

    @property
    def pre(self):
        # type: () -> None
        return None

    @property
    def post(self):
        # type: () -> None
        return None

    @property
    def dev(self):
        # type: () -> None
        return None

    @property
    def local(self):
        # type: () -> None
        return None

    @property
    def is_prerelease(self):
        # type: () -> bool
        return False

    @property
    def is_postrelease(self):
        # type: () -> bool
        return False

    @property
    def is_devrelease(self):
        # type: () -> bool
        return False


_legacy_version_component_re = re.compile(r"(\d+ | [a-z]+ | \.| -)", re.VERBOSE)

_legacy_version_replacement_map = {
    "pre": "c",
    "preview": "c",
    "-": "final-",
    "rc": "c",
    "dev": "@",
}


def _parse_version_parts(s):
    # type: (str) -> Iterator[str]
    for part in _legacy_version_component_re.split(s):
        part = _legacy_version_replacement_map.get(part, part)

        if not part or part == ".":
            continue

        if part[:1] in "0123456789":
            # pad for numeric comparison
            yield part.zfill(8)
        else:
            yield "*" + part

    # ensure that alpha/beta/candidate are before final
    yield "*final"


def _legacy_cmpkey(version):
    # type: (str) -> LegacyCmpKey

    # We hardcode an epoch of -1 here. A PEP 440 version can only have a epoch
    # greater than or equal to 0. This will effectively put the LegacyVersion,
    # which uses the defacto standard originally implemented by setuptools,
    # as before all PEP 440 versions.
    epoch = -1

    # This scheme is taken from pkg_resources.parse_version setuptools prior to
    # it's adoption of the packaging library.
    parts = []  # type: List[str]
    for part in _parse_version_parts(version.lower()):
        if part.startswith("*"):
            # remove "-" before a prerelease tag
            if part < "*final":
                while parts and parts[-1] == "*final-":
                    parts.pop()

            # remove trailing zeros from each series of numeric parts
            while parts and parts[-1] == "00000000":
                parts.pop()

        parts.append(part)

    return epoch, tuple(parts)


# Deliberately not anchored to the start and end of the string, to make it
# easier for 3rd party code to reuse
VERSION_PATTERN = r"""
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
"""


class Version(_BaseVersion):

    _regex = re.compile(r"^\s*" + VERSION_PATTERN + r"\s*$", re.VERBOSE | re.IGNORECASE)

    def __init__(self, version):
        # type: (str) -> None

        # Validate the version and parse it into pieces
        match = self._regex.search(version)
        if not match:
            raise InvalidVersion("Invalid version: '{0}'".format(version))

        # Store the parsed out pieces of the version
        self._version = _Version(
            epoch=int(match.group("epoch")) if match.group("epoch") else 0,
            release=tuple(int(i) for i in match.group("release").split(".")),
            pre=_parse_letter_version(match.group("pre_l"), match.group("pre_n")),
            post=_parse_letter_version(
                match.group("post_l"), match.group("post_n1") or match.group("post_n2")
            ),
            dev=_parse_letter_version(match.group("dev_l"), match.group("dev_n")),
            local=_parse_local_version(match.group("local")),
        )

        # Generate a key which will be used for sorting
        self._key = _cmpkey(
            self._version.epoch,
            self._version.release,
            self._version.pre,
            self._version.post,
            self._version.dev,
            self._version.local,
        )

    def __repr__(self):
        # type: () -> str
        return "<Version({0})>".format(repr(str(self)))

    def __str__(self):
        # type: () -> str
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        # Pre-release
        if self.pre is not None:
            parts.append("".join(str(x) for x in self.pre))

        # Post-release
        if self.post is not None:
            parts.append(".post{0}".format(self.post))

        # Development release
        if self.dev is not None:
            parts.append(".dev{0}".format(self.dev))

        # Local version segment
        if self.local is not None:
            parts.append("+{0}".format(self.local))

        return "".join(parts)

    @property
    def epoch(self):
        # type: () -> int
        _epoch = self._version.epoch  # type: int
        return _epoch

    @property
    def release(self):
        # type: () -> Tuple[int, ...]
        _release = self._version.release  # type: Tuple[int, ...]
        return _release

    @property
    def pre(self):
        # type: () -> Optional[Tuple[str, int]]
        _pre = self._version.pre  # type: Optional[Tuple[str, int]]
        return _pre

    @property
    def post(self):
        # type: () -> Optional[Tuple[str, int]]
        return self._version.post[1] if self._version.post else None

    @property
    def dev(self):
        # type: () -> Optional[Tuple[str, int]]
        return self._version.dev[1] if self._version.dev else None

    @property
    def local(self):
        # type: () -> Optional[str]
        if self._version.local:
            return ".".join(str(x) for x in self._version.local)
        else:
            return None

    @property
    def public(self):
        # type: () -> str
        return str(self).split("+", 1)[0]

    @property
    def base_version(self):
        # type: () -> str
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append("{0}!".format(self.epoch))

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        return "".join(parts)

    @property
    def is_prerelease(self):
        # type: () -> bool
        return self.dev is not None or self.pre is not None

    @property
    def is_postrelease(self):
        # type: () -> bool
        return self.post is not None

    @property
    def is_devrelease(self):
        # type: () -> bool
        return self.dev is not None

    @property
    def major(self):
        # type: () -> int
        return self.release[0] if len(self.release) >= 1 else 0

    @property
    def minor(self):
        # type: () -> int
        return self.release[1] if len(self.release) >= 2 else 0

    @property
    def micro(self):
        # type: () -> int
        return self.release[2] if len(self.release) >= 3 else 0


def _parse_letter_version(
    letter,  # type: str
    number,  # type: Union[str, bytes, SupportsInt]
):
    # type: (...) -> Optional[Tuple[str, int]]

    if letter:
        # We consider there to be an implicit 0 in a pre-release if there is
        # not a numeral associated with it.
        if number is None:
            number = 0

        # We normalize any letters to their lower case form
        letter = letter.lower()

        # We consider some words to be alternate spellings of other words and
        # in those cases we want to normalize the spellings to our preferred
        # spelling.
        if letter == "alpha":
            letter = "a"
        elif letter == "beta":
            letter = "b"
        elif letter in ["c", "pre", "preview"]:
            letter = "rc"
        elif letter in ["rev", "r"]:
            letter = "post"

        return letter, int(number)
    if not letter and number:
        # We assume if we are given a number, but we are not given a letter
        # then this is using the implicit post release syntax (e.g. 1.0-1)
        letter = "post"

        return letter, int(number)

    return None


_local_version_separators = re.compile(r"[\._-]")


def _parse_local_version(local):
    # type: (str) -> Optional[LocalType]
    """
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    """
    if local is not None:
        return tuple(
            part.lower() if not part.isdigit() else int(part)
            for part in _local_version_separators.split(local)
        )
    return None


def _cmpkey(
    epoch,  # type: int
    release,  # type: Tuple[int, ...]
    pre,  # type: Optional[Tuple[str, int]]
    post,  # type: Optional[Tuple[str, int]]
    dev,  # type: Optional[Tuple[str, int]]
    local,  # type: Optional[Tuple[SubLocalType]]
):
    # type: (...) -> CmpKey

    # When we compare a release version, we want to compare it with all of the
    # trailing zeros removed. So we'll use a reverse the list, drop all the now
    # leading zeros until we come to something non zero, then take the rest
    # re-reverse it back into the correct order and make it a tuple and use
    # that for our sorting key.
    _release = tuple(
        reversed(list(itertools.dropwhile(lambda x: x == 0, reversed(release))))
    )

    # We need to "trick" the sorting algorithm to put 1.0.dev0 before 1.0a0.
    # We'll do this by abusing the pre segment, but we _only_ want to do this
    # if there is not a pre or a post segment. If we have one of those then
    # the normal sorting rules will handle this case correctly.
    if pre is None and post is None and dev is not None:
        _pre = NegativeInfinity  # type: PrePostDevType
    # Versions without a pre-release (except as noted above) should sort after
    # those with one.
    elif pre is None:
        _pre = Infinity
    else:
        _pre = pre

    # Versions without a post segment should sort before those with one.
    if post is None:
        _post = NegativeInfinity  # type: PrePostDevType

    else:
        _post = post

    # Versions without a development segment should sort after those with one.
    if dev is None:
        _dev = Infinity  # type: PrePostDevType

    else:
        _dev = dev

    if local is None:
        # Versions without a local segment should sort before those with one.
        _local = NegativeInfinity  # type: LocalType
    else:
        # Versions with a local segment need that segment parsed to implement
        # the sorting rules in PEP440.
        # - Alpha numeric segments sort before numeric segments
        # - Alpha numeric segments sort lexicographically
        # - Numeric segments sort numerically
        # - Shorter versions sort before longer versions when the prefixes
        #   match exactly
        _local = tuple(
            (i, "") if isinstance(i, int) else (NegativeInfinity, i) for i in local
        )

    return epoch, _release, _pre, _post, _dev, _local
site-packages/pkg_resources/_vendor/packaging/__init__.py000064400000001062150564745010017636 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

from .__about__ import (
    __author__,
    __copyright__,
    __email__,
    __license__,
    __summary__,
    __title__,
    __uri__,
    __version__,
)

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]
site-packages/pkg_resources/_vendor/packaging/specifiers.py000064400000076310150564745010020243 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import abc
import functools
import itertools
import re

from ._compat import string_types, with_metaclass
from ._typing import TYPE_CHECKING
from .utils import canonicalize_version
from .version import Version, LegacyVersion, parse

if TYPE_CHECKING:  # pragma: no cover
    from typing import (
        List,
        Dict,
        Union,
        Iterable,
        Iterator,
        Optional,
        Callable,
        Tuple,
        FrozenSet,
    )

    ParsedVersion = Union[Version, LegacyVersion]
    UnparsedVersion = Union[Version, LegacyVersion, str]
    CallableOperator = Callable[[ParsedVersion, str], bool]


class InvalidSpecifier(ValueError):
    """
    An invalid specifier was found, users should refer to PEP 440.
    """


class BaseSpecifier(with_metaclass(abc.ABCMeta, object)):  # type: ignore
    @abc.abstractmethod
    def __str__(self):
        # type: () -> str
        """
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        """

    @abc.abstractmethod
    def __hash__(self):
        # type: () -> int
        """
        Returns a hash value for this Specifier like object.
        """

    @abc.abstractmethod
    def __eq__(self, other):
        # type: (object) -> bool
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        """

    @abc.abstractmethod
    def __ne__(self, other):
        # type: (object) -> bool
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        """

    @abc.abstractproperty
    def prereleases(self):
        # type: () -> Optional[bool]
        """
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        """
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @abc.abstractmethod
    def contains(self, item, prereleases=None):
        # type: (str, Optional[bool]) -> bool
        """
        Determines if the given item is contained within this specifier.
        """

    @abc.abstractmethod
    def filter(self, iterable, prereleases=None):
        # type: (Iterable[UnparsedVersion], Optional[bool]) -> Iterable[UnparsedVersion]
        """
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        """


class _IndividualSpecifier(BaseSpecifier):

    _operators = {}  # type: Dict[str, str]

    def __init__(self, spec="", prereleases=None):
        # type: (str, Optional[bool]) -> None
        match = self._regex.search(spec)
        if not match:
            raise InvalidSpecifier("Invalid specifier: '{0}'".format(spec))

        self._spec = (
            match.group("operator").strip(),
            match.group("version").strip(),
        )  # type: Tuple[str, str]

        # Store whether or not this Specifier should accept prereleases
        self._prereleases = prereleases

    def __repr__(self):
        # type: () -> str
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<{0}({1!r}{2})>".format(self.__class__.__name__, str(self), pre)

    def __str__(self):
        # type: () -> str
        return "{0}{1}".format(*self._spec)

    @property
    def _canonical_spec(self):
        # type: () -> Tuple[str, Union[Version, str]]
        return self._spec[0], canonicalize_version(self._spec[1])

    def __hash__(self):
        # type: () -> int
        return hash(self._canonical_spec)

    def __eq__(self, other):
        # type: (object) -> bool
        if isinstance(other, string_types):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._canonical_spec == other._canonical_spec

    def __ne__(self, other):
        # type: (object) -> bool
        if isinstance(other, string_types):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec != other._spec

    def _get_operator(self, op):
        # type: (str) -> CallableOperator
        operator_callable = getattr(
            self, "_compare_{0}".format(self._operators[op])
        )  # type: CallableOperator
        return operator_callable

    def _coerce_version(self, version):
        # type: (UnparsedVersion) -> ParsedVersion
        if not isinstance(version, (LegacyVersion, Version)):
            version = parse(version)
        return version

    @property
    def operator(self):
        # type: () -> str
        return self._spec[0]

    @property
    def version(self):
        # type: () -> str
        return self._spec[1]

    @property
    def prereleases(self):
        # type: () -> Optional[bool]
        return self._prereleases

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value

    def __contains__(self, item):
        # type: (str) -> bool
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # type: (UnparsedVersion, Optional[bool]) -> bool

        # Determine if prereleases are to be allowed or not.
        if prereleases is None:
            prereleases = self.prereleases

        # Normalize item to a Version or LegacyVersion, this allows us to have
        # a shortcut for ``"2.0" in Specifier(">=2")
        normalized_item = self._coerce_version(item)

        # Determine if we should be supporting prereleases in this specifier
        # or not, if we do not support prereleases than we can short circuit
        # logic if this version is a prereleases.
        if normalized_item.is_prerelease and not prereleases:
            return False

        # Actually do the comparison to determine if this item is contained
        # within this Specifier or not.
        operator_callable = self._get_operator(self.operator)  # type: CallableOperator
        return operator_callable(normalized_item, self.version)

    def filter(self, iterable, prereleases=None):
        # type: (Iterable[UnparsedVersion], Optional[bool]) -> Iterable[UnparsedVersion]

        yielded = False
        found_prereleases = []

        kw = {"prereleases": prereleases if prereleases is not None else True}

        # Attempt to iterate over all the values in the iterable and if any of
        # them match, yield them.
        for version in iterable:
            parsed_version = self._coerce_version(version)

            if self.contains(parsed_version, **kw):
                # If our version is a prerelease, and we were not set to allow
                # prereleases, then we'll store it for later incase nothing
                # else matches this specifier.
                if parsed_version.is_prerelease and not (
                    prereleases or self.prereleases
                ):
                    found_prereleases.append(version)
                # Either this is not a prerelease, or we should have been
                # accepting prereleases from the beginning.
                else:
                    yielded = True
                    yield version

        # Now that we've iterated over everything, determine if we've yielded
        # any values, and if we have not and we have any prereleases stored up
        # then we will go ahead and yield the prereleases.
        if not yielded and found_prereleases:
            for version in found_prereleases:
                yield version


class LegacySpecifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
    }

    def _coerce_version(self, version):
        # type: (Union[ParsedVersion, str]) -> LegacyVersion
        if not isinstance(version, LegacyVersion):
            version = LegacyVersion(str(version))
        return version

    def _compare_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective == self._coerce_version(spec)

    def _compare_not_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective != self._coerce_version(spec)

    def _compare_less_than_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective <= self._coerce_version(spec)

    def _compare_greater_than_equal(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective >= self._coerce_version(spec)

    def _compare_less_than(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective < self._coerce_version(spec)

    def _compare_greater_than(self, prospective, spec):
        # type: (LegacyVersion, str) -> bool
        return prospective > self._coerce_version(spec)


def _require_version_compare(
    fn  # type: (Callable[[Specifier, ParsedVersion, str], bool])
):
    # type: (...) -> Callable[[Specifier, ParsedVersion, str], bool]
    @functools.wraps(fn)
    def wrapped(self, prospective, spec):
        # type: (Specifier, ParsedVersion, str) -> bool
        if not isinstance(prospective, Version):
            return False
        return fn(self, prospective, spec)

    return wrapped


class Specifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "~=": "compatible",
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
        "===": "arbitrary",
    }

    @_require_version_compare
    def _compare_compatible(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # Compatible releases have an equivalent combination of >= and ==. That
        # is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
        # implement this in terms of the other specifiers instead of
        # implementing it ourselves. The only thing we need to do is construct
        # the other specifiers.

        # We want everything but the last item in the version, but we want to
        # ignore post and dev releases and we want to treat the pre-release as
        # it's own separate segment.
        prefix = ".".join(
            list(
                itertools.takewhile(
                    lambda x: (not x.startswith("post") and not x.startswith("dev")),
                    _version_split(spec),
                )
            )[:-1]
        )

        # Add the prefix notation to the end of our string
        prefix += ".*"

        return self._get_operator(">=")(prospective, spec) and self._get_operator("==")(
            prospective, prefix
        )

    @_require_version_compare
    def _compare_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # We need special logic to handle prefix matching
        if spec.endswith(".*"):
            # In the case of prefix matching we want to ignore local segment.
            prospective = Version(prospective.public)
            # Split the spec out by dots, and pretend that there is an implicit
            # dot in between a release segment and a pre-release segment.
            split_spec = _version_split(spec[:-2])  # Remove the trailing .*

            # Split the prospective version out by dots, and pretend that there
            # is an implicit dot in between a release segment and a pre-release
            # segment.
            split_prospective = _version_split(str(prospective))

            # Shorten the prospective version to be the same length as the spec
            # so that we can determine if the specifier is a prefix of the
            # prospective version or not.
            shortened_prospective = split_prospective[: len(split_spec)]

            # Pad out our two sides with zeros so that they both equal the same
            # length.
            padded_spec, padded_prospective = _pad_version(
                split_spec, shortened_prospective
            )

            return padded_prospective == padded_spec
        else:
            # Convert our spec string into a Version
            spec_version = Version(spec)

            # If the specifier does not have a local segment, then we want to
            # act as if the prospective version also does not have a local
            # segment.
            if not spec_version.local:
                prospective = Version(prospective.public)

            return prospective == spec_version

    @_require_version_compare
    def _compare_not_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool
        return not self._compare_equal(prospective, spec)

    @_require_version_compare
    def _compare_less_than_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) <= Version(spec)

    @_require_version_compare
    def _compare_greater_than_equal(self, prospective, spec):
        # type: (ParsedVersion, str) -> bool

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) >= Version(spec)

    @_require_version_compare
    def _compare_less_than(self, prospective, spec_str):
        # type: (ParsedVersion, str) -> bool

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is less than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective < spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a pre-release version, that we do not accept pre-release
        # versions for the version mentioned in the specifier (e.g. <3.1 should
        # not match 3.1.dev0, but should match 3.0.dev0).
        if not spec.is_prerelease and prospective.is_prerelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # less than the spec version *and* it's not a pre-release of the same
        # version in the spec.
        return True

    @_require_version_compare
    def _compare_greater_than(self, prospective, spec_str):
        # type: (ParsedVersion, str) -> bool

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is greater than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective > spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a post-release version, that we do not accept
        # post-release versions for the version mentioned in the specifier
        # (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
        if not spec.is_postrelease and prospective.is_postrelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # Ensure that we do not allow a local version of the version mentioned
        # in the specifier, which is technically greater than, to match.
        if prospective.local is not None:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # greater than the spec version *and* it's not a pre-release of the
        # same version in the spec.
        return True

    def _compare_arbitrary(self, prospective, spec):
        # type: (Version, str) -> bool
        return str(prospective).lower() == str(spec).lower()

    @property
    def prereleases(self):
        # type: () -> bool

        # If there is an explicit prereleases set for this, then we'll just
        # blindly use that.
        if self._prereleases is not None:
            return self._prereleases

        # Look at all of our specifiers and determine if they are inclusive
        # operators, and if they are if they are including an explicit
        # prerelease.
        operator, version = self._spec
        if operator in ["==", ">=", "<=", "~=", "==="]:
            # The == specifier can include a trailing .*, if it does we
            # want to remove before parsing.
            if operator == "==" and version.endswith(".*"):
                version = version[:-2]

            # Parse the version, and if it is a pre-release than this
            # specifier allows pre-releases.
            if parse(version).is_prerelease:
                return True

        return False

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value


_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")


def _version_split(version):
    # type: (str) -> List[str]
    result = []  # type: List[str]
    for item in version.split("."):
        match = _prefix_regex.search(item)
        if match:
            result.extend(match.groups())
        else:
            result.append(item)
    return result


def _pad_version(left, right):
    # type: (List[str], List[str]) -> Tuple[List[str], List[str]]
    left_split, right_split = [], []

    # Get the release segment of our versions
    left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
    right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))

    # Get the rest of our versions
    left_split.append(left[len(left_split[0]) :])
    right_split.append(right[len(right_split[0]) :])

    # Insert our padding
    left_split.insert(1, ["0"] * max(0, len(right_split[0]) - len(left_split[0])))
    right_split.insert(1, ["0"] * max(0, len(left_split[0]) - len(right_split[0])))

    return (list(itertools.chain(*left_split)), list(itertools.chain(*right_split)))


class SpecifierSet(BaseSpecifier):
    def __init__(self, specifiers="", prereleases=None):
        # type: (str, Optional[bool]) -> None

        # Split on , to break each individual specifier into it's own item, and
        # strip each item to remove leading/trailing whitespace.
        split_specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]

        # Parsed each individual specifier, attempting first to make it a
        # Specifier and falling back to a LegacySpecifier.
        parsed = set()
        for specifier in split_specifiers:
            try:
                parsed.add(Specifier(specifier))
            except InvalidSpecifier:
                parsed.add(LegacySpecifier(specifier))

        # Turn our parsed specifiers into a frozen set and save them for later.
        self._specs = frozenset(parsed)

        # Store our prereleases value so we can use it later to determine if
        # we accept prereleases or not.
        self._prereleases = prereleases

    def __repr__(self):
        # type: () -> str
        pre = (
            ", prereleases={0!r}".format(self.prereleases)
            if self._prereleases is not None
            else ""
        )

        return "<SpecifierSet({0!r}{1})>".format(str(self), pre)

    def __str__(self):
        # type: () -> str
        return ",".join(sorted(str(s) for s in self._specs))

    def __hash__(self):
        # type: () -> int
        return hash(self._specs)

    def __and__(self, other):
        # type: (Union[SpecifierSet, str]) -> SpecifierSet
        if isinstance(other, string_types):
            other = SpecifierSet(other)
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        specifier = SpecifierSet()
        specifier._specs = frozenset(self._specs | other._specs)

        if self._prereleases is None and other._prereleases is not None:
            specifier._prereleases = other._prereleases
        elif self._prereleases is not None and other._prereleases is None:
            specifier._prereleases = self._prereleases
        elif self._prereleases == other._prereleases:
            specifier._prereleases = self._prereleases
        else:
            raise ValueError(
                "Cannot combine SpecifierSets with True and False prerelease "
                "overrides."
            )

        return specifier

    def __eq__(self, other):
        # type: (object) -> bool
        if isinstance(other, (string_types, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs == other._specs

    def __ne__(self, other):
        # type: (object) -> bool
        if isinstance(other, (string_types, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs != other._specs

    def __len__(self):
        # type: () -> int
        return len(self._specs)

    def __iter__(self):
        # type: () -> Iterator[FrozenSet[_IndividualSpecifier]]
        return iter(self._specs)

    @property
    def prereleases(self):
        # type: () -> Optional[bool]

        # If we have been given an explicit prerelease modifier, then we'll
        # pass that through here.
        if self._prereleases is not None:
            return self._prereleases

        # If we don't have any specifiers, and we don't have a forced value,
        # then we'll just return None since we don't know if this should have
        # pre-releases or not.
        if not self._specs:
            return None

        # Otherwise we'll see if any of the given specifiers accept
        # prereleases, if any of them do we'll return True, otherwise False.
        return any(s.prereleases for s in self._specs)

    @prereleases.setter
    def prereleases(self, value):
        # type: (bool) -> None
        self._prereleases = value

    def __contains__(self, item):
        # type: (Union[ParsedVersion, str]) -> bool
        return self.contains(item)

    def contains(self, item, prereleases=None):
        # type: (Union[ParsedVersion, str], Optional[bool]) -> bool

        # Ensure that our item is a Version or LegacyVersion instance.
        if not isinstance(item, (LegacyVersion, Version)):
            item = parse(item)

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # We can determine if we're going to allow pre-releases by looking to
        # see if any of the underlying items supports them. If none of them do
        # and this item is a pre-release then we do not allow it and we can
        # short circuit that here.
        # Note: This means that 1.0.dev1 would not be contained in something
        #       like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
        if not prereleases and item.is_prerelease:
            return False

        # We simply dispatch to the underlying specs here to make sure that the
        # given version is contained within all of them.
        # Note: This use of all() here means that an empty set of specifiers
        #       will always return True, this is an explicit design decision.
        return all(s.contains(item, prereleases=prereleases) for s in self._specs)

    def filter(
        self,
        iterable,  # type: Iterable[Union[ParsedVersion, str]]
        prereleases=None,  # type: Optional[bool]
    ):
        # type: (...) -> Iterable[Union[ParsedVersion, str]]

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # If we have any specifiers, then we want to wrap our iterable in the
        # filter method for each one, this will act as a logical AND amongst
        # each specifier.
        if self._specs:
            for spec in self._specs:
                iterable = spec.filter(iterable, prereleases=bool(prereleases))
            return iterable
        # If we do not have any specifiers, then we need to have a rough filter
        # which will filter out any pre-releases, unless there are no final
        # releases, and which will filter out LegacyVersion in general.
        else:
            filtered = []  # type: List[Union[ParsedVersion, str]]
            found_prereleases = []  # type: List[Union[ParsedVersion, str]]

            for item in iterable:
                # Ensure that we some kind of Version class for this item.
                if not isinstance(item, (LegacyVersion, Version)):
                    parsed_version = parse(item)
                else:
                    parsed_version = item

                # Filter out any item which is parsed as a LegacyVersion
                if isinstance(parsed_version, LegacyVersion):
                    continue

                # Store any item which is a pre-release for later unless we've
                # already found a final version or we are accepting prereleases
                if parsed_version.is_prerelease and not prereleases:
                    if not filtered:
                        found_prereleases.append(item)
                else:
                    filtered.append(item)

            # If we've found no items except for pre-releases, then we'll go
            # ahead and use the pre-releases
            if not filtered and found_prereleases and prereleases is None:
                return found_prereleases

            return filtered
site-packages/pkg_resources/_vendor/packaging/_structures.py000064400000003746150564745010020474 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function


class InfinityType(object):
    def __repr__(self):
        # type: () -> str
        return "Infinity"

    def __hash__(self):
        # type: () -> int
        return hash(repr(self))

    def __lt__(self, other):
        # type: (object) -> bool
        return False

    def __le__(self, other):
        # type: (object) -> bool
        return False

    def __eq__(self, other):
        # type: (object) -> bool
        return isinstance(other, self.__class__)

    def __ne__(self, other):
        # type: (object) -> bool
        return not isinstance(other, self.__class__)

    def __gt__(self, other):
        # type: (object) -> bool
        return True

    def __ge__(self, other):
        # type: (object) -> bool
        return True

    def __neg__(self):
        # type: (object) -> NegativeInfinityType
        return NegativeInfinity


Infinity = InfinityType()


class NegativeInfinityType(object):
    def __repr__(self):
        # type: () -> str
        return "-Infinity"

    def __hash__(self):
        # type: () -> int
        return hash(repr(self))

    def __lt__(self, other):
        # type: (object) -> bool
        return True

    def __le__(self, other):
        # type: (object) -> bool
        return True

    def __eq__(self, other):
        # type: (object) -> bool
        return isinstance(other, self.__class__)

    def __ne__(self, other):
        # type: (object) -> bool
        return not isinstance(other, self.__class__)

    def __gt__(self, other):
        # type: (object) -> bool
        return False

    def __ge__(self, other):
        # type: (object) -> bool
        return False

    def __neg__(self):
        # type: (object) -> InfinityType
        return Infinity


NegativeInfinity = NegativeInfinityType()
site-packages/pkg_resources/_vendor/packaging/tags.py000064400000057003150564745010017043 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

from __future__ import absolute_import

import distutils.util

try:
    from importlib.machinery import EXTENSION_SUFFIXES
except ImportError:  # pragma: no cover
    import imp

    EXTENSION_SUFFIXES = [x[0] for x in imp.get_suffixes()]
    del imp
import logging
import os
import platform
import re
import struct
import sys
import sysconfig
import warnings

from ._typing import TYPE_CHECKING, cast

if TYPE_CHECKING:  # pragma: no cover
    from typing import (
        Dict,
        FrozenSet,
        IO,
        Iterable,
        Iterator,
        List,
        Optional,
        Sequence,
        Tuple,
        Union,
    )

    PythonVersion = Sequence[int]
    MacVersion = Tuple[int, int]
    GlibcVersion = Tuple[int, int]


logger = logging.getLogger(__name__)

INTERPRETER_SHORT_NAMES = {
    "python": "py",  # Generic.
    "cpython": "cp",
    "pypy": "pp",
    "ironpython": "ip",
    "jython": "jy",
}  # type: Dict[str, str]


_32_BIT_INTERPRETER = sys.maxsize <= 2 ** 32


class Tag(object):
    """
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    """

    __slots__ = ["_interpreter", "_abi", "_platform"]

    def __init__(self, interpreter, abi, platform):
        # type: (str, str, str) -> None
        self._interpreter = interpreter.lower()
        self._abi = abi.lower()
        self._platform = platform.lower()

    @property
    def interpreter(self):
        # type: () -> str
        return self._interpreter

    @property
    def abi(self):
        # type: () -> str
        return self._abi

    @property
    def platform(self):
        # type: () -> str
        return self._platform

    def __eq__(self, other):
        # type: (object) -> bool
        if not isinstance(other, Tag):
            return NotImplemented

        return (
            (self.platform == other.platform)
            and (self.abi == other.abi)
            and (self.interpreter == other.interpreter)
        )

    def __hash__(self):
        # type: () -> int
        return hash((self._interpreter, self._abi, self._platform))

    def __str__(self):
        # type: () -> str
        return "{}-{}-{}".format(self._interpreter, self._abi, self._platform)

    def __repr__(self):
        # type: () -> str
        return "<{self} @ {self_id}>".format(self=self, self_id=id(self))


def parse_tag(tag):
    # type: (str) -> FrozenSet[Tag]
    """
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    """
    tags = set()
    interpreters, abis, platforms = tag.split("-")
    for interpreter in interpreters.split("."):
        for abi in abis.split("."):
            for platform_ in platforms.split("."):
                tags.add(Tag(interpreter, abi, platform_))
    return frozenset(tags)


def _warn_keyword_parameter(func_name, kwargs):
    # type: (str, Dict[str, bool]) -> bool
    """
    Backwards-compatibility with Python 2.7 to allow treating 'warn' as keyword-only.
    """
    if not kwargs:
        return False
    elif len(kwargs) > 1 or "warn" not in kwargs:
        kwargs.pop("warn", None)
        arg = next(iter(kwargs.keys()))
        raise TypeError(
            "{}() got an unexpected keyword argument {!r}".format(func_name, arg)
        )
    return kwargs["warn"]


def _get_config_var(name, warn=False):
    # type: (str, bool) -> Union[int, str, None]
    value = sysconfig.get_config_var(name)
    if value is None and warn:
        logger.debug(
            "Config variable '%s' is unset, Python ABI tag may be incorrect", name
        )
    return value


def _normalize_string(string):
    # type: (str) -> str
    return string.replace(".", "_").replace("-", "_")


def _abi3_applies(python_version):
    # type: (PythonVersion) -> bool
    """
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    """
    return len(python_version) > 1 and tuple(python_version) >= (3, 2)


def _cpython_abis(py_version, warn=False):
    # type: (PythonVersion, bool) -> List[str]
    py_version = tuple(py_version)  # To allow for version comparison.
    abis = []
    version = _version_nodot(py_version[:2])
    debug = pymalloc = ucs4 = ""
    with_debug = _get_config_var("Py_DEBUG", warn)
    has_refcount = hasattr(sys, "gettotalrefcount")
    # Windows doesn't set Py_DEBUG, so checking for support of debug-compiled
    # extension modules is the best option.
    # https://github.com/pypa/pip/issues/3383#issuecomment-173267692
    has_ext = "_d.pyd" in EXTENSION_SUFFIXES
    if with_debug or (with_debug is None and (has_refcount or has_ext)):
        debug = "d"
    if py_version < (3, 8):
        with_pymalloc = _get_config_var("WITH_PYMALLOC", warn)
        if with_pymalloc or with_pymalloc is None:
            pymalloc = "m"
        if py_version < (3, 3):
            unicode_size = _get_config_var("Py_UNICODE_SIZE", warn)
            if unicode_size == 4 or (
                unicode_size is None and sys.maxunicode == 0x10FFFF
            ):
                ucs4 = "u"
    elif debug:
        # Debug builds can also load "normal" extension modules.
        # We can also assume no UCS-4 or pymalloc requirement.
        abis.append("cp{version}".format(version=version))
    abis.insert(
        0,
        "cp{version}{debug}{pymalloc}{ucs4}".format(
            version=version, debug=debug, pymalloc=pymalloc, ucs4=ucs4
        ),
    )
    return abis


def cpython_tags(
    python_version=None,  # type: Optional[PythonVersion]
    abis=None,  # type: Optional[Iterable[str]]
    platforms=None,  # type: Optional[Iterable[str]]
    **kwargs  # type: bool
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    """
    warn = _warn_keyword_parameter("cpython_tags", kwargs)
    if not python_version:
        python_version = sys.version_info[:2]

    interpreter = "cp{}".format(_version_nodot(python_version[:2]))

    if abis is None:
        if len(python_version) > 1:
            abis = _cpython_abis(python_version, warn)
        else:
            abis = []
    abis = list(abis)
    # 'abi3' and 'none' are explicitly handled later.
    for explicit_abi in ("abi3", "none"):
        try:
            abis.remove(explicit_abi)
        except ValueError:
            pass

    platforms = list(platforms or _platform_tags())
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)
    if _abi3_applies(python_version):
        for tag in (Tag(interpreter, "abi3", platform_) for platform_ in platforms):
            yield tag
    for tag in (Tag(interpreter, "none", platform_) for platform_ in platforms):
        yield tag

    if _abi3_applies(python_version):
        for minor_version in range(python_version[1] - 1, 1, -1):
            for platform_ in platforms:
                interpreter = "cp{version}".format(
                    version=_version_nodot((python_version[0], minor_version))
                )
                yield Tag(interpreter, "abi3", platform_)


def _generic_abi():
    # type: () -> Iterator[str]
    abi = sysconfig.get_config_var("SOABI")
    if abi:
        yield _normalize_string(abi)


def generic_tags(
    interpreter=None,  # type: Optional[str]
    abis=None,  # type: Optional[Iterable[str]]
    platforms=None,  # type: Optional[Iterable[str]]
    **kwargs  # type: bool
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    """
    warn = _warn_keyword_parameter("generic_tags", kwargs)
    if not interpreter:
        interp_name = interpreter_name()
        interp_version = interpreter_version(warn=warn)
        interpreter = "".join([interp_name, interp_version])
    if abis is None:
        abis = _generic_abi()
    platforms = list(platforms or _platform_tags())
    abis = list(abis)
    if "none" not in abis:
        abis.append("none")
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)


def _py_interpreter_range(py_version):
    # type: (PythonVersion) -> Iterator[str]
    """
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    """
    if len(py_version) > 1:
        yield "py{version}".format(version=_version_nodot(py_version[:2]))
    yield "py{major}".format(major=py_version[0])
    if len(py_version) > 1:
        for minor in range(py_version[1] - 1, -1, -1):
            yield "py{version}".format(version=_version_nodot((py_version[0], minor)))


def compatible_tags(
    python_version=None,  # type: Optional[PythonVersion]
    interpreter=None,  # type: Optional[str]
    platforms=None,  # type: Optional[Iterable[str]]
):
    # type: (...) -> Iterator[Tag]
    """
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    """
    if not python_version:
        python_version = sys.version_info[:2]
    platforms = list(platforms or _platform_tags())
    for version in _py_interpreter_range(python_version):
        for platform_ in platforms:
            yield Tag(version, "none", platform_)
    if interpreter:
        yield Tag(interpreter, "none", "any")
    for version in _py_interpreter_range(python_version):
        yield Tag(version, "none", "any")


def _mac_arch(arch, is_32bit=_32_BIT_INTERPRETER):
    # type: (str, bool) -> str
    if not is_32bit:
        return arch

    if arch.startswith("ppc"):
        return "ppc"

    return "i386"


def _mac_binary_formats(version, cpu_arch):
    # type: (MacVersion, str) -> List[str]
    formats = [cpu_arch]
    if cpu_arch == "x86_64":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat64", "fat32"])

    elif cpu_arch == "i386":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat32", "fat"])

    elif cpu_arch == "ppc64":
        # TODO: Need to care about 32-bit PPC for ppc64 through 10.2?
        if version > (10, 5) or version < (10, 4):
            return []
        formats.append("fat64")

    elif cpu_arch == "ppc":
        if version > (10, 6):
            return []
        formats.extend(["fat32", "fat"])

    formats.append("universal")
    return formats


def mac_platforms(version=None, arch=None):
    # type: (Optional[MacVersion], Optional[str]) -> Iterator[str]
    """
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    """
    version_str, _, cpu_arch = platform.mac_ver()  # type: ignore
    if version is None:
        version = cast("MacVersion", tuple(map(int, version_str.split(".")[:2])))
    else:
        version = version
    if arch is None:
        arch = _mac_arch(cpu_arch)
    else:
        arch = arch
    for minor_version in range(version[1], -1, -1):
        compat_version = version[0], minor_version
        binary_formats = _mac_binary_formats(compat_version, arch)
        for binary_format in binary_formats:
            yield "macosx_{major}_{minor}_{binary_format}".format(
                major=compat_version[0],
                minor=compat_version[1],
                binary_format=binary_format,
            )


# From PEP 513.
def _is_manylinux_compatible(name, glibc_version):
    # type: (str, GlibcVersion) -> bool
    # Check for presence of _manylinux module.
    try:
        import _manylinux  # noqa

        return bool(getattr(_manylinux, name + "_compatible"))
    except (ImportError, AttributeError):
        # Fall through to heuristic check below.
        pass

    return _have_compatible_glibc(*glibc_version)


def _glibc_version_string():
    # type: () -> Optional[str]
    # Returns glibc version string, or None if not using glibc.
    return _glibc_version_string_confstr() or _glibc_version_string_ctypes()


def _glibc_version_string_confstr():
    # type: () -> Optional[str]
    """
    Primary implementation of glibc_version_string using os.confstr.
    """
    # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely
    # to be broken or missing. This strategy is used in the standard library
    # platform module.
    # https://github.com/python/cpython/blob/fcf1d003bf4f0100c9d0921ff3d70e1127ca1b71/Lib/platform.py#L175-L183
    try:
        # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17".
        version_string = os.confstr(  # type: ignore[attr-defined] # noqa: F821
            "CS_GNU_LIBC_VERSION"
        )
        assert version_string is not None
        _, version = version_string.split()  # type: Tuple[str, str]
    except (AssertionError, AttributeError, OSError, ValueError):
        # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)...
        return None
    return version


def _glibc_version_string_ctypes():
    # type: () -> Optional[str]
    """
    Fallback implementation of glibc_version_string using ctypes.
    """
    try:
        import ctypes
    except ImportError:
        return None

    # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen
    # manpage says, "If filename is NULL, then the returned handle is for the
    # main program". This way we can let the linker do the work to figure out
    # which libc our process is actually using.
    #
    # Note: typeshed is wrong here so we are ignoring this line.
    process_namespace = ctypes.CDLL(None)  # type: ignore
    try:
        gnu_get_libc_version = process_namespace.gnu_get_libc_version
    except AttributeError:
        # Symbol doesn't exist -> therefore, we are not linked to
        # glibc.
        return None

    # Call gnu_get_libc_version, which returns a string like "2.5"
    gnu_get_libc_version.restype = ctypes.c_char_p
    version_str = gnu_get_libc_version()  # type: str
    # py2 / py3 compatibility:
    if not isinstance(version_str, str):
        version_str = version_str.decode("ascii")

    return version_str


# Separated out from have_compatible_glibc for easier unit testing.
def _check_glibc_version(version_str, required_major, minimum_minor):
    # type: (str, int, int) -> bool
    # Parse string and check against requested version.
    #
    # We use a regexp instead of str.split because we want to discard any
    # random junk that might come after the minor version -- this might happen
    # in patched/forked versions of glibc (e.g. Linaro's version of glibc
    # uses version strings like "2.20-2014.11"). See gh-3588.
    m = re.match(r"(?P<major>[0-9]+)\.(?P<minor>[0-9]+)", version_str)
    if not m:
        warnings.warn(
            "Expected glibc version with 2 components major.minor,"
            " got: %s" % version_str,
            RuntimeWarning,
        )
        return False
    return (
        int(m.group("major")) == required_major
        and int(m.group("minor")) >= minimum_minor
    )


def _have_compatible_glibc(required_major, minimum_minor):
    # type: (int, int) -> bool
    version_str = _glibc_version_string()
    if version_str is None:
        return False
    return _check_glibc_version(version_str, required_major, minimum_minor)


# Python does not provide platform information at sufficient granularity to
# identify the architecture of the running executable in some cases, so we
# determine it dynamically by reading the information from the running
# process. This only applies on Linux, which uses the ELF format.
class _ELFFileHeader(object):
    # https://en.wikipedia.org/wiki/Executable_and_Linkable_Format#File_header
    class _InvalidELFFileHeader(ValueError):
        """
        An invalid ELF file header was found.
        """

    ELF_MAGIC_NUMBER = 0x7F454C46
    ELFCLASS32 = 1
    ELFCLASS64 = 2
    ELFDATA2LSB = 1
    ELFDATA2MSB = 2
    EM_386 = 3
    EM_S390 = 22
    EM_ARM = 40
    EM_X86_64 = 62
    EF_ARM_ABIMASK = 0xFF000000
    EF_ARM_ABI_VER5 = 0x05000000
    EF_ARM_ABI_FLOAT_HARD = 0x00000400

    def __init__(self, file):
        # type: (IO[bytes]) -> None
        def unpack(fmt):
            # type: (str) -> int
            try:
                (result,) = struct.unpack(
                    fmt, file.read(struct.calcsize(fmt))
                )  # type: (int, )
            except struct.error:
                raise _ELFFileHeader._InvalidELFFileHeader()
            return result

        self.e_ident_magic = unpack(">I")
        if self.e_ident_magic != self.ELF_MAGIC_NUMBER:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_class = unpack("B")
        if self.e_ident_class not in {self.ELFCLASS32, self.ELFCLASS64}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_data = unpack("B")
        if self.e_ident_data not in {self.ELFDATA2LSB, self.ELFDATA2MSB}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_version = unpack("B")
        self.e_ident_osabi = unpack("B")
        self.e_ident_abiversion = unpack("B")
        self.e_ident_pad = file.read(7)
        format_h = "<H" if self.e_ident_data == self.ELFDATA2LSB else ">H"
        format_i = "<I" if self.e_ident_data == self.ELFDATA2LSB else ">I"
        format_q = "<Q" if self.e_ident_data == self.ELFDATA2LSB else ">Q"
        format_p = format_i if self.e_ident_class == self.ELFCLASS32 else format_q
        self.e_type = unpack(format_h)
        self.e_machine = unpack(format_h)
        self.e_version = unpack(format_i)
        self.e_entry = unpack(format_p)
        self.e_phoff = unpack(format_p)
        self.e_shoff = unpack(format_p)
        self.e_flags = unpack(format_i)
        self.e_ehsize = unpack(format_h)
        self.e_phentsize = unpack(format_h)
        self.e_phnum = unpack(format_h)
        self.e_shentsize = unpack(format_h)
        self.e_shnum = unpack(format_h)
        self.e_shstrndx = unpack(format_h)


def _get_elf_header():
    # type: () -> Optional[_ELFFileHeader]
    try:
        with open(sys.executable, "rb") as f:
            elf_header = _ELFFileHeader(f)
    except (IOError, OSError, TypeError, _ELFFileHeader._InvalidELFFileHeader):
        return None
    return elf_header


def _is_linux_armhf():
    # type: () -> bool
    # hard-float ABI can be detected from the ELF header of the running
    # process
    # https://static.docs.arm.com/ihi0044/g/aaelf32.pdf
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_ARM
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABIMASK
    ) == elf_header.EF_ARM_ABI_VER5
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABI_FLOAT_HARD
    ) == elf_header.EF_ARM_ABI_FLOAT_HARD
    return result


def _is_linux_i686():
    # type: () -> bool
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_386
    return result


def _have_compatible_manylinux_abi(arch):
    # type: (str) -> bool
    if arch == "armv7l":
        return _is_linux_armhf()
    if arch == "i686":
        return _is_linux_i686()
    return True


def _linux_platforms(is_32bit=_32_BIT_INTERPRETER):
    # type: (bool) -> Iterator[str]
    linux = _normalize_string(distutils.util.get_platform())
    if is_32bit:
        if linux == "linux_x86_64":
            linux = "linux_i686"
        elif linux == "linux_aarch64":
            linux = "linux_armv7l"
    manylinux_support = []
    _, arch = linux.split("_", 1)
    if _have_compatible_manylinux_abi(arch):
        if arch in {"x86_64", "i686", "aarch64", "armv7l", "ppc64", "ppc64le", "s390x"}:
            manylinux_support.append(
                ("manylinux2014", (2, 17))
            )  # CentOS 7 w/ glibc 2.17 (PEP 599)
        if arch in {"x86_64", "i686"}:
            manylinux_support.append(
                ("manylinux2010", (2, 12))
            )  # CentOS 6 w/ glibc 2.12 (PEP 571)
            manylinux_support.append(
                ("manylinux1", (2, 5))
            )  # CentOS 5 w/ glibc 2.5 (PEP 513)
    manylinux_support_iter = iter(manylinux_support)
    for name, glibc_version in manylinux_support_iter:
        if _is_manylinux_compatible(name, glibc_version):
            yield linux.replace("linux", name)
            break
    # Support for a later manylinux implies support for an earlier version.
    for name, _ in manylinux_support_iter:
        yield linux.replace("linux", name)
    yield linux


def _generic_platforms():
    # type: () -> Iterator[str]
    yield _normalize_string(distutils.util.get_platform())


def _platform_tags():
    # type: () -> Iterator[str]
    """
    Provides the platform tags for this installation.
    """
    if platform.system() == "Darwin":
        return mac_platforms()
    elif platform.system() == "Linux":
        return _linux_platforms()
    else:
        return _generic_platforms()


def interpreter_name():
    # type: () -> str
    """
    Returns the name of the running interpreter.
    """
    try:
        name = sys.implementation.name  # type: ignore
    except AttributeError:  # pragma: no cover
        # Python 2.7 compatibility.
        name = platform.python_implementation().lower()
    return INTERPRETER_SHORT_NAMES.get(name) or name


def interpreter_version(**kwargs):
    # type: (bool) -> str
    """
    Returns the version of the running interpreter.
    """
    warn = _warn_keyword_parameter("interpreter_version", kwargs)
    version = _get_config_var("py_version_nodot", warn=warn)
    if version:
        version = str(version)
    else:
        version = _version_nodot(sys.version_info[:2])
    return version


def _version_nodot(version):
    # type: (PythonVersion) -> str
    if any(v >= 10 for v in version):
        sep = "_"
    else:
        sep = ""
    return sep.join(map(str, version))


def sys_tags(**kwargs):
    # type: (bool) -> Iterator[Tag]
    """
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    """
    warn = _warn_keyword_parameter("sys_tags", kwargs)

    interp_name = interpreter_name()
    if interp_name == "cp":
        for tag in cpython_tags(warn=warn):
            yield tag
    else:
        for tag in generic_tags():
            yield tag

    for tag in compatible_tags():
        yield tag
site-packages/pkg_resources/_vendor/packaging/requirements.py000064400000011501150564745010020621 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.
from __future__ import absolute_import, division, print_function

import string
import re

from pkg_resources.extern.pyparsing import stringStart, stringEnd, originalTextFor, ParseException
from pkg_resources.extern.pyparsing import ZeroOrMore, Word, Optional, Regex, Combine
from pkg_resources.extern.pyparsing import Literal as L  # noqa
from urllib import parse as urlparse

from ._typing import TYPE_CHECKING
from .markers import MARKER_EXPR, Marker
from .specifiers import LegacySpecifier, Specifier, SpecifierSet

if TYPE_CHECKING:  # pragma: no cover
    from typing import List


class InvalidRequirement(ValueError):
    """
    An invalid requirement was found, users should refer to PEP 508.
    """


ALPHANUM = Word(string.ascii_letters + string.digits)

LBRACKET = L("[").suppress()
RBRACKET = L("]").suppress()
LPAREN = L("(").suppress()
RPAREN = L(")").suppress()
COMMA = L(",").suppress()
SEMICOLON = L(";").suppress()
AT = L("@").suppress()

PUNCTUATION = Word("-_.")
IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM)
IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END))

NAME = IDENTIFIER("name")
EXTRA = IDENTIFIER

URI = Regex(r"[^ ]+")("url")
URL = AT + URI

EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA)
EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras")

VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE)
VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE)

VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY
VERSION_MANY = Combine(
    VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False
)("_raw_spec")
_VERSION_SPEC = Optional(((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY))
_VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or "")

VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier")
VERSION_SPEC.setParseAction(lambda s, l, t: t[1])

MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker")
MARKER_EXPR.setParseAction(
    lambda s, l, t: Marker(s[t._original_start : t._original_end])
)
MARKER_SEPARATOR = SEMICOLON
MARKER = MARKER_SEPARATOR + MARKER_EXPR

VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER)
URL_AND_MARKER = URL + Optional(MARKER)

NAMED_REQUIREMENT = NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER)

REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd
# pkg_resources.extern.pyparsing isn't thread safe during initialization, so we do it eagerly, see
# issue #104
REQUIREMENT.parseString("x[]")


class Requirement(object):
    """Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    """

    # TODO: Can we test whether something is contained within a requirement?
    #       If so how do we do that? Do we need to test against the _name_ of
    #       the thing as well as the version? What about the markers?
    # TODO: Can we normalize the name and extra name?

    def __init__(self, requirement_string):
        # type: (str) -> None
        try:
            req = REQUIREMENT.parseString(requirement_string)
        except ParseException as e:
            raise InvalidRequirement(
                'Parse error at "{0!r}": {1}'.format(
                    requirement_string[e.loc : e.loc + 8], e.msg
                )
            )

        self.name = req.name
        if req.url:
            parsed_url = urlparse.urlparse(req.url)
            if parsed_url.scheme == "file":
                if urlparse.urlunparse(parsed_url) != req.url:
                    raise InvalidRequirement("Invalid URL given")
            elif not (parsed_url.scheme and parsed_url.netloc) or (
                not parsed_url.scheme and not parsed_url.netloc
            ):
                raise InvalidRequirement("Invalid URL: {0}".format(req.url))
            self.url = req.url
        else:
            self.url = None
        self.extras = set(req.extras.asList() if req.extras else [])
        self.specifier = SpecifierSet(req.specifier)
        self.marker = req.marker if req.marker else None

    def __str__(self):
        # type: () -> str
        parts = [self.name]  # type: List[str]

        if self.extras:
            parts.append("[{0}]".format(",".join(sorted(self.extras))))

        if self.specifier:
            parts.append(str(self.specifier))

        if self.url:
            parts.append("@ {0}".format(self.url))
            if self.marker:
                parts.append(" ")

        if self.marker:
            parts.append("; {0}".format(self.marker))

        return "".join(parts)

    def __repr__(self):
        # type: () -> str
        return "<Requirement({0!r})>".format(str(self))
site-packages/pkg_resources/_vendor/pyparsing.py000064400000705170150564745010016202 0ustar00# module pyparsing.py
#
# Copyright (c) 2003-2018  Paul T. McGuire
#
# Permission is hereby granted, free of charge, to any person obtaining
# a copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish,
# distribute, sublicense, and/or sell copies of the Software, and to
# permit persons to whom the Software is furnished to do so, subject to
# the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
# IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
# CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
#

__doc__ = \
"""
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and executing simple grammars,
vs. the traditional lex/yacc approach, or the use of regular expressions.  With pyparsing, you
don't need to learn a new syntax for defining grammars or matching expressions - the parsing module
provides a library of classes that you use to construct the grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form 
C{"<salutation>, <addressee>!"}), built up using L{Word}, L{Literal}, and L{And} elements 
(L{'+'<ParserElement.__add__>} operator gives L{And} expressions, strings are auto-converted to
L{Literal} expressions)::

    from pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the self-explanatory
class names, and the use of '+', '|' and '^' operators.

The L{ParseResults} object returned from L{ParserElement.parseString<ParserElement.parseString>} can be accessed as a nested list, a dictionary, or an
object with named attributes.

The pyparsing module handles some of the problems that are typically vexing when writing text parsers:
 - extra or missing whitespace (the above program will also handle "Hello,World!", "Hello  ,  World  !", etc.)
 - quoted strings
 - embedded comments


Getting Started -
-----------------
Visit the classes L{ParserElement} and L{ParseResults} to see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:
 - construct literal match expressions from L{Literal} and L{CaselessLiteral} classes
 - construct character word-group expressions using the L{Word} class
 - see how to create repetitive expressions using L{ZeroOrMore} and L{OneOrMore} classes
 - use L{'+'<And>}, L{'|'<MatchFirst>}, L{'^'<Or>}, and L{'&'<Each>} operators to combine simple expressions into more complex ones
 - associate names with your parsed results using L{ParserElement.setResultsName}
 - find some helpful expression short-cuts like L{delimitedList} and L{oneOf}
 - find more useful common expressions in the L{pyparsing_common} namespace class
"""

__version__ = "2.2.1"
__versionTime__ = "18 Sep 2018 00:49 UTC"
__author__ = "Paul McGuire <ptmcg@users.sourceforge.net>"

import string
from weakref import ref as wkref
import copy
import sys
import warnings
import re
import sre_constants
import collections
import pprint
import traceback
import types
from datetime import datetime

try:
    from _thread import RLock
except ImportError:
    from threading import RLock

try:
    # Python 3
    from collections.abc import Iterable
    from collections.abc import MutableMapping
except ImportError:
    # Python 2.7
    from collections import Iterable
    from collections import MutableMapping

try:
    from collections import OrderedDict as _OrderedDict
except ImportError:
    try:
        from ordereddict import OrderedDict as _OrderedDict
    except ImportError:
        _OrderedDict = None

#~ sys.stderr.write( "testing pyparsing module, version %s, %s\n" % (__version__,__versionTime__ ) )

__all__ = [
'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty',
'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal',
'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or',
'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException',
'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException',
'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter', 
'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore',
'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col',
'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString',
'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums',
'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno',
'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral',
'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables',
'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity', 
'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd',
'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute',
'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation','locatedExpr', 'withClass',
'CloseMatch', 'tokenMap', 'pyparsing_common',
]

system_version = tuple(sys.version_info)[:3]
PY_3 = system_version[0] == 3
if PY_3:
    _MAX_INT = sys.maxsize
    basestring = str
    unichr = chr
    _ustr = str

    # build list of single arg builtins, that can be used as parse actions
    singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max]

else:
    _MAX_INT = sys.maxint
    range = xrange

    def _ustr(obj):
        """Drop-in replacement for str(obj) that tries to be Unicode friendly. It first tries
           str(obj). If that fails with a UnicodeEncodeError, then it tries unicode(obj). It
           then < returns the unicode object | encodes it with the default encoding | ... >.
        """
        if isinstance(obj,unicode):
            return obj

        try:
            # If this works, then _ustr(obj) has the same behaviour as str(obj), so
            # it won't break any existing code.
            return str(obj)

        except UnicodeEncodeError:
            # Else encode it
            ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace')
            xmlcharref = Regex(r'&#\d+;')
            xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:])
            return xmlcharref.transformString(ret)

    # build list of single arg builtins, tolerant of Python version, that can be used as parse actions
    singleArgBuiltins = []
    import __builtin__
    for fname in "sum len sorted reversed list tuple set any all min max".split():
        try:
            singleArgBuiltins.append(getattr(__builtin__,fname))
        except AttributeError:
            continue
            
_generatorType = type((y for y in range(1)))
 
def _xml_escape(data):
    """Escape &, <, >, ", ', etc. in a string of data."""

    # ampersand must be replaced first
    from_symbols = '&><"\''
    to_symbols = ('&'+s+';' for s in "amp gt lt quot apos".split())
    for from_,to_ in zip(from_symbols, to_symbols):
        data = data.replace(from_, to_)
    return data

class _Constants(object):
    pass

alphas     = string.ascii_uppercase + string.ascii_lowercase
nums       = "0123456789"
hexnums    = nums + "ABCDEFabcdef"
alphanums  = alphas + nums
_bslash    = chr(92)
printables = "".join(c for c in string.printable if c not in string.whitespace)

class ParseBaseException(Exception):
    """base exception class for all parsing runtime exceptions"""
    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__( self, pstr, loc=0, msg=None, elem=None ):
        self.loc = loc
        if msg is None:
            self.msg = pstr
            self.pstr = ""
        else:
            self.msg = msg
            self.pstr = pstr
        self.parserElement = elem
        self.args = (pstr, loc, msg)

    @classmethod
    def _from_exception(cls, pe):
        """
        internal factory method to simplify creating one type of ParseException 
        from another - avoids having __init__ signature conflicts among subclasses
        """
        return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement)

    def __getattr__( self, aname ):
        """supported attributes by name are:
            - lineno - returns the line number of the exception text
            - col - returns the column number of the exception text
            - line - returns the line containing the exception text
        """
        if( aname == "lineno" ):
            return lineno( self.loc, self.pstr )
        elif( aname in ("col", "column") ):
            return col( self.loc, self.pstr )
        elif( aname == "line" ):
            return line( self.loc, self.pstr )
        else:
            raise AttributeError(aname)

    def __str__( self ):
        return "%s (at char %d), (line:%d, col:%d)" % \
                ( self.msg, self.loc, self.lineno, self.column )
    def __repr__( self ):
        return _ustr(self)
    def markInputline( self, markerString = ">!<" ):
        """Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        """
        line_str = self.line
        line_column = self.column - 1
        if markerString:
            line_str = "".join((line_str[:line_column],
                                markerString, line_str[line_column:]))
        return line_str.strip()
    def __dir__(self):
        return "lineno col line".split() + dir(type(self))

class ParseException(ParseBaseException):
    """
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
     - lineno - returns the line number of the exception text
     - col - returns the column number of the exception text
     - line - returns the line containing the exception text
        
    Example::
        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))
            
    prints::
       Expected integer (at char 0), (line:1, col:1)
        column: 1
    """
    pass

class ParseFatalException(ParseBaseException):
    """user-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediately"""
    pass

class ParseSyntaxException(ParseFatalException):
    """just like L{ParseFatalException}, but thrown internally when an
       L{ErrorStop<And._ErrorStop>} ('-' operator) indicates that parsing is to stop 
       immediately because an unbacktrackable syntax error has been found"""
    pass

#~ class ReparseException(ParseBaseException):
    #~ """Experimental class - parse actions can raise this exception to cause
       #~ pyparsing to reparse the input string:
        #~ - with a modified input string, and/or
        #~ - with a modified start location
       #~ Set the values of the ReparseException in the constructor, and raise the
       #~ exception in a parse action to cause pyparsing to use the new string/location.
       #~ Setting the values as None causes no change to be made.
       #~ """
    #~ def __init_( self, newstring, restartLoc ):
        #~ self.newParseText = newstring
        #~ self.reparseLoc = restartLoc

class RecursiveGrammarException(Exception):
    """exception thrown by L{ParserElement.validate} if the grammar could be improperly recursive"""
    def __init__( self, parseElementList ):
        self.parseElementTrace = parseElementList

    def __str__( self ):
        return "RecursiveGrammarException: %s" % self.parseElementTrace

class _ParseResultsWithOffset(object):
    def __init__(self,p1,p2):
        self.tup = (p1,p2)
    def __getitem__(self,i):
        return self.tup[i]
    def __repr__(self):
        return repr(self.tup[0])
    def setOffset(self,i):
        self.tup = (self.tup[0],i)

class ParseResults(object):
    """
    Structured parse results, to provide multiple means of access to the parsed data:
       - as a list (C{len(results)})
       - by list index (C{results[0], results[1]}, etc.)
       - by attribute (C{results.<resultsName>} - see L{ParserElement.setResultsName})

    Example::
        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)
    prints::
        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    """
    def __new__(cls, toklist=None, name=None, asList=True, modal=True ):
        if isinstance(toklist, cls):
            return toklist
        retobj = object.__new__(cls)
        retobj.__doinit = True
        return retobj

    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__( self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance ):
        if self.__doinit:
            self.__doinit = False
            self.__name = None
            self.__parent = None
            self.__accumNames = {}
            self.__asList = asList
            self.__modal = modal
            if toklist is None:
                toklist = []
            if isinstance(toklist, list):
                self.__toklist = toklist[:]
            elif isinstance(toklist, _generatorType):
                self.__toklist = list(toklist)
            else:
                self.__toklist = [toklist]
            self.__tokdict = dict()

        if name is not None and name:
            if not modal:
                self.__accumNames[name] = 0
            if isinstance(name,int):
                name = _ustr(name) # will always return a str, but use _ustr for consistency
            self.__name = name
            if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None,'',[])):
                if isinstance(toklist,basestring):
                    toklist = [ toklist ]
                if asList:
                    if isinstance(toklist,ParseResults):
                        self[name] = _ParseResultsWithOffset(toklist.copy(),0)
                    else:
                        self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]),0)
                    self[name].__name = name
                else:
                    try:
                        self[name] = toklist[0]
                    except (KeyError,TypeError,IndexError):
                        self[name] = toklist

    def __getitem__( self, i ):
        if isinstance( i, (int,slice) ):
            return self.__toklist[i]
        else:
            if i not in self.__accumNames:
                return self.__tokdict[i][-1][0]
            else:
                return ParseResults([ v[0] for v in self.__tokdict[i] ])

    def __setitem__( self, k, v, isinstance=isinstance ):
        if isinstance(v,_ParseResultsWithOffset):
            self.__tokdict[k] = self.__tokdict.get(k,list()) + [v]
            sub = v[0]
        elif isinstance(k,(int,slice)):
            self.__toklist[k] = v
            sub = v
        else:
            self.__tokdict[k] = self.__tokdict.get(k,list()) + [_ParseResultsWithOffset(v,0)]
            sub = v
        if isinstance(sub,ParseResults):
            sub.__parent = wkref(self)

    def __delitem__( self, i ):
        if isinstance(i,(int,slice)):
            mylen = len( self.__toklist )
            del self.__toklist[i]

            # convert int to slice
            if isinstance(i, int):
                if i < 0:
                    i += mylen
                i = slice(i, i+1)
            # get removed indices
            removed = list(range(*i.indices(mylen)))
            removed.reverse()
            # fixup indices in token dictionary
            for name,occurrences in self.__tokdict.items():
                for j in removed:
                    for k, (value, position) in enumerate(occurrences):
                        occurrences[k] = _ParseResultsWithOffset(value, position - (position > j))
        else:
            del self.__tokdict[i]

    def __contains__( self, k ):
        return k in self.__tokdict

    def __len__( self ): return len( self.__toklist )
    def __bool__(self): return ( not not self.__toklist )
    __nonzero__ = __bool__
    def __iter__( self ): return iter( self.__toklist )
    def __reversed__( self ): return iter( self.__toklist[::-1] )
    def _iterkeys( self ):
        if hasattr(self.__tokdict, "iterkeys"):
            return self.__tokdict.iterkeys()
        else:
            return iter(self.__tokdict)

    def _itervalues( self ):
        return (self[k] for k in self._iterkeys())
            
    def _iteritems( self ):
        return ((k, self[k]) for k in self._iterkeys())

    if PY_3:
        keys = _iterkeys       
        """Returns an iterator of all named result keys (Python 3.x only)."""

        values = _itervalues
        """Returns an iterator of all named result values (Python 3.x only)."""

        items = _iteritems
        """Returns an iterator of all named result key-value tuples (Python 3.x only)."""

    else:
        iterkeys = _iterkeys
        """Returns an iterator of all named result keys (Python 2.x only)."""

        itervalues = _itervalues
        """Returns an iterator of all named result values (Python 2.x only)."""

        iteritems = _iteritems
        """Returns an iterator of all named result key-value tuples (Python 2.x only)."""

        def keys( self ):
            """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iterkeys())

        def values( self ):
            """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.itervalues())
                
        def items( self ):
            """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iteritems())

    def haskeys( self ):
        """Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names."""
        return bool(self.__tokdict)
        
    def pop( self, *args, **kwargs):
        """
        Removes and returns item at specified index (default=C{last}).
        Supports both C{list} and C{dict} semantics for C{pop()}. If passed no
        argument or an integer argument, it will use C{list} semantics
        and pop tokens from the list of parsed tokens. If passed a 
        non-integer argument (most likely a string), it will use C{dict}
        semantics and pop the corresponding value from any defined 
        results names. A second default return value argument is 
        supported, just as in C{dict.pop()}.

        Example::
            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())
        prints::
            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        """
        if not args:
            args = [-1]
        for k,v in kwargs.items():
            if k == 'default':
                args = (args[0], v)
            else:
                raise TypeError("pop() got an unexpected keyword argument '%s'" % k)
        if (isinstance(args[0], int) or 
                        len(args) == 1 or 
                        args[0] in self):
            index = args[0]
            ret = self[index]
            del self[index]
            return ret
        else:
            defaultvalue = args[1]
            return defaultvalue

    def get(self, key, defaultValue=None):
        """
        Returns named result matching the given key, or if there is no
        such name, then returns the given C{defaultValue} or C{None} if no
        C{defaultValue} is specified.

        Similar to C{dict.get()}.
        
        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        """
        if key in self:
            return self[key]
        else:
            return defaultValue

    def insert( self, index, insStr ):
        """
        Inserts new element at location index in the list of parsed tokens.
        
        Similar to C{list.insert()}.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        """
        self.__toklist.insert(index, insStr)
        # fixup indices in token dictionary
        for name,occurrences in self.__tokdict.items():
            for k, (value, position) in enumerate(occurrences):
                occurrences[k] = _ParseResultsWithOffset(value, position + (position > index))

    def append( self, item ):
        """
        Add single element to end of ParseResults list of elements.

        Example::
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            
            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        """
        self.__toklist.append(item)

    def extend( self, itemseq ):
        """
        Add sequence of elements to end of ParseResults list of elements.

        Example::
            patt = OneOrMore(Word(alphas))
            
            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        """
        if isinstance(itemseq, ParseResults):
            self += itemseq
        else:
            self.__toklist.extend(itemseq)

    def clear( self ):
        """
        Clear all elements and results names.
        """
        del self.__toklist[:]
        self.__tokdict.clear()

    def __getattr__( self, name ):
        try:
            return self[name]
        except KeyError:
            return ""
            
        if name in self.__tokdict:
            if name not in self.__accumNames:
                return self.__tokdict[name][-1][0]
            else:
                return ParseResults([ v[0] for v in self.__tokdict[name] ])
        else:
            return ""

    def __add__( self, other ):
        ret = self.copy()
        ret += other
        return ret

    def __iadd__( self, other ):
        if other.__tokdict:
            offset = len(self.__toklist)
            addoffset = lambda a: offset if a<0 else a+offset
            otheritems = other.__tokdict.items()
            otherdictitems = [(k, _ParseResultsWithOffset(v[0],addoffset(v[1])) )
                                for (k,vlist) in otheritems for v in vlist]
            for k,v in otherdictitems:
                self[k] = v
                if isinstance(v[0],ParseResults):
                    v[0].__parent = wkref(self)
            
        self.__toklist += other.__toklist
        self.__accumNames.update( other.__accumNames )
        return self

    def __radd__(self, other):
        if isinstance(other,int) and other == 0:
            # useful for merging many ParseResults using sum() builtin
            return self.copy()
        else:
            # this may raise a TypeError - so be it
            return other + self
        
    def __repr__( self ):
        return "(%s, %s)" % ( repr( self.__toklist ), repr( self.__tokdict ) )

    def __str__( self ):
        return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']'

    def _asStringList( self, sep='' ):
        out = []
        for item in self.__toklist:
            if out and sep:
                out.append(sep)
            if isinstance( item, ParseResults ):
                out += item._asStringList()
            else:
                out.append( _ustr(item) )
        return out

    def asList( self ):
        """
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::
            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']
            
            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        """
        return [res.asList() if isinstance(res,ParseResults) else res for res in self.__toklist]

    def asDict( self ):
        """
        Returns the named parse results as a nested dictionary.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})
            
            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        """
        if PY_3:
            item_fn = self.items
        else:
            item_fn = self.iteritems
            
        def toItem(obj):
            if isinstance(obj, ParseResults):
                if obj.haskeys():
                    return obj.asDict()
                else:
                    return [toItem(v) for v in obj]
            else:
                return obj
                
        return dict((k,toItem(v)) for k,v in item_fn())

    def copy( self ):
        """
        Returns a new copy of a C{ParseResults} object.
        """
        ret = ParseResults( self.__toklist )
        ret.__tokdict = self.__tokdict.copy()
        ret.__parent = self.__parent
        ret.__accumNames.update( self.__accumNames )
        ret.__name = self.__name
        return ret

    def asXML( self, doctag=None, namedItemsOnly=False, indent="", formatted=True ):
        """
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        """
        nl = "\n"
        out = []
        namedItems = dict((v[1],k) for (k,vlist) in self.__tokdict.items()
                                                            for v in vlist)
        nextLevelIndent = indent + "  "

        # collapse out indents if formatting is not desired
        if not formatted:
            indent = ""
            nextLevelIndent = ""
            nl = ""

        selfTag = None
        if doctag is not None:
            selfTag = doctag
        else:
            if self.__name:
                selfTag = self.__name

        if not selfTag:
            if namedItemsOnly:
                return ""
            else:
                selfTag = "ITEM"

        out += [ nl, indent, "<", selfTag, ">" ]

        for i,res in enumerate(self.__toklist):
            if isinstance(res,ParseResults):
                if i in namedItems:
                    out += [ res.asXML(namedItems[i],
                                        namedItemsOnly and doctag is None,
                                        nextLevelIndent,
                                        formatted)]
                else:
                    out += [ res.asXML(None,
                                        namedItemsOnly and doctag is None,
                                        nextLevelIndent,
                                        formatted)]
            else:
                # individual token, see if there is a name for it
                resTag = None
                if i in namedItems:
                    resTag = namedItems[i]
                if not resTag:
                    if namedItemsOnly:
                        continue
                    else:
                        resTag = "ITEM"
                xmlBodyText = _xml_escape(_ustr(res))
                out += [ nl, nextLevelIndent, "<", resTag, ">",
                                                xmlBodyText,
                                                "</", resTag, ">" ]

        out += [ nl, indent, "</", selfTag, ">" ]
        return "".join(out)

    def __lookup(self,sub):
        for k,vlist in self.__tokdict.items():
            for v,loc in vlist:
                if sub is v:
                    return k
        return None

    def getName(self):
        r"""
        Returns the results name for this token expression. Useful when several 
        different expressions might match at a particular location.

        Example::
            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number") 
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)
            
            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])
        prints::
            age : 22
            ssn : 111-22-3333
            house_number : 221B
        """
        if self.__name:
            return self.__name
        elif self.__parent:
            par = self.__parent()
            if par:
                return par.__lookup(self)
            else:
                return None
        elif (len(self) == 1 and
               len(self.__tokdict) == 1 and
               next(iter(self.__tokdict.values()))[0][1] in (0,-1)):
            return next(iter(self.__tokdict.keys()))
        else:
            return None

    def dump(self, indent='', depth=0, full=True):
        """
        Diagnostic method for listing out the contents of a C{ParseResults}.
        Accepts an optional C{indent} argument so that this string can be embedded
        in a nested display of other data.

        Example::
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
            
            result = date_str.parseString('12/31/1999')
            print(result.dump())
        prints::
            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        """
        out = []
        NL = '\n'
        out.append( indent+_ustr(self.asList()) )
        if full:
            if self.haskeys():
                items = sorted((str(k), v) for k,v in self.items())
                for k,v in items:
                    if out:
                        out.append(NL)
                    out.append( "%s%s- %s: " % (indent,('  '*depth), k) )
                    if isinstance(v,ParseResults):
                        if v:
                            out.append( v.dump(indent,depth+1) )
                        else:
                            out.append(_ustr(v))
                    else:
                        out.append(repr(v))
            elif any(isinstance(vv,ParseResults) for vv in self):
                v = self
                for i,vv in enumerate(v):
                    if isinstance(vv,ParseResults):
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,('  '*(depth)),i,indent,('  '*(depth+1)),vv.dump(indent,depth+1) ))
                    else:
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,('  '*(depth)),i,indent,('  '*(depth+1)),_ustr(vv)))
            
        return "".join(out)

    def pprint(self, *args, **kwargs):
        """
        Pretty-printer for parsed results as a list, using the C{pprint} module.
        Accepts additional positional or keyword args as defined for the 
        C{pprint.pprint} method. (U{http://docs.python.org/3/library/pprint.html#pprint.pprint})

        Example::
            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)
        prints::
            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        """
        pprint.pprint(self.asList(), *args, **kwargs)

    # add support for pickle protocol
    def __getstate__(self):
        return ( self.__toklist,
                 ( self.__tokdict.copy(),
                   self.__parent is not None and self.__parent() or None,
                   self.__accumNames,
                   self.__name ) )

    def __setstate__(self,state):
        self.__toklist = state[0]
        (self.__tokdict,
         par,
         inAccumNames,
         self.__name) = state[1]
        self.__accumNames = {}
        self.__accumNames.update(inAccumNames)
        if par is not None:
            self.__parent = wkref(par)
        else:
            self.__parent = None

    def __getnewargs__(self):
        return self.__toklist, self.__name, self.__asList, self.__modal

    def __dir__(self):
        return (dir(type(self)) + list(self.keys()))

MutableMapping.register(ParseResults)

def col (loc,strg):
    """Returns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   """
    s = strg
    return 1 if 0<loc<len(s) and s[loc-1] == '\n' else loc - s.rfind("\n", 0, loc)

def lineno(loc,strg):
    """Returns current line number within a string, counting newlines as line separators.
   The first line is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See L{I{ParserElement.parseString}<ParserElement.parseString>} for more information
   on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
   consistent view of the parsed string, the parse location, and line and column
   positions within the parsed string.
   """
    return strg.count("\n",0,loc) + 1

def line( loc, strg ):
    """Returns the line of text containing loc within a string, counting newlines as line separators.
       """
    lastCR = strg.rfind("\n", 0, loc)
    nextCR = strg.find("\n", loc)
    if nextCR >= 0:
        return strg[lastCR+1:nextCR]
    else:
        return strg[lastCR+1:]

def _defaultStartDebugAction( instring, loc, expr ):
    print (("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % ( lineno(loc,instring), col(loc,instring) )))

def _defaultSuccessDebugAction( instring, startloc, endloc, expr, toks ):
    print ("Matched " + _ustr(expr) + " -> " + str(toks.asList()))

def _defaultExceptionDebugAction( instring, loc, expr, exc ):
    print ("Exception raised:" + _ustr(exc))

def nullDebugAction(*args):
    """'Do-nothing' debug action, to suppress debugging output during parsing."""
    pass

# Only works on Python 3.x - nonlocal is toxic to Python 2 installs
#~ 'decorator to trim function calls to match the arity of the target'
#~ def _trim_arity(func, maxargs=3):
    #~ if func in singleArgBuiltins:
        #~ return lambda s,l,t: func(t)
    #~ limit = 0
    #~ foundArity = False
    #~ def wrapper(*args):
        #~ nonlocal limit,foundArity
        #~ while 1:
            #~ try:
                #~ ret = func(*args[limit:])
                #~ foundArity = True
                #~ return ret
            #~ except TypeError:
                #~ if limit == maxargs or foundArity:
                    #~ raise
                #~ limit += 1
                #~ continue
    #~ return wrapper

# this version is Python 2.x-3.x cross-compatible
'decorator to trim function calls to match the arity of the target'
def _trim_arity(func, maxargs=2):
    if func in singleArgBuiltins:
        return lambda s,l,t: func(t)
    limit = [0]
    foundArity = [False]
    
    # traceback return data structure changed in Py3.5 - normalize back to plain tuples
    if system_version[:2] >= (3,5):
        def extract_stack(limit=0):
            # special handling for Python 3.5.0 - extra deep call stack by 1
            offset = -3 if system_version == (3,5,0) else -2
            frame_summary = traceback.extract_stack(limit=-offset+limit-1)[offset]
            return [frame_summary[:2]]
        def extract_tb(tb, limit=0):
            frames = traceback.extract_tb(tb, limit=limit)
            frame_summary = frames[-1]
            return [frame_summary[:2]]
    else:
        extract_stack = traceback.extract_stack
        extract_tb = traceback.extract_tb
    
    # synthesize what would be returned by traceback.extract_stack at the call to 
    # user's parse action 'func', so that we don't incur call penalty at parse time
    
    LINE_DIFF = 6
    # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND 
    # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!!
    this_line = extract_stack(limit=2)[-1]
    pa_call_line_synth = (this_line[0], this_line[1]+LINE_DIFF)

    def wrapper(*args):
        while 1:
            try:
                ret = func(*args[limit[0]:])
                foundArity[0] = True
                return ret
            except TypeError:
                # re-raise TypeErrors if they did not come from our arity testing
                if foundArity[0]:
                    raise
                else:
                    try:
                        tb = sys.exc_info()[-1]
                        if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth:
                            raise
                    finally:
                        del tb

                if limit[0] <= maxargs:
                    limit[0] += 1
                    continue
                raise

    # copy func name to wrapper for sensible debug output
    func_name = "<parse action>"
    try:
        func_name = getattr(func, '__name__', 
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    wrapper.__name__ = func_name

    return wrapper

class ParserElement(object):
    """Abstract base level parser element class."""
    DEFAULT_WHITE_CHARS = " \n\t\r"
    verbose_stacktrace = False

    @staticmethod
    def setDefaultWhitespaceChars( chars ):
        r"""
        Overrides the default whitespace chars

        Example::
            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']
            
            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        """
        ParserElement.DEFAULT_WHITE_CHARS = chars

    @staticmethod
    def inlineLiteralsUsing(cls):
        """
        Set class to be used for inclusion of string literals into a parser.
        
        Example::
            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")           

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        """
        ParserElement._literalStringClass = cls

    def __init__( self, savelist=False ):
        self.parseAction = list()
        self.failAction = None
        #~ self.name = "<unknown>"  # don't define self.name, let subclasses try/except upcall
        self.strRepr = None
        self.resultsName = None
        self.saveAsList = savelist
        self.skipWhitespace = True
        self.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
        self.copyDefaultWhiteChars = True
        self.mayReturnEmpty = False # used when checking for left-recursion
        self.keepTabs = False
        self.ignoreExprs = list()
        self.debug = False
        self.streamlined = False
        self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index
        self.errmsg = ""
        self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all)
        self.debugActions = ( None, None, None ) #custom debug actions
        self.re = None
        self.callPreparse = True # used to avoid redundant calls to preParse
        self.callDuringTry = False

    def copy( self ):
        """
        Make a copy of this C{ParserElement}.  Useful for defining different parse actions
        for the same parsing pattern, using copies of the original parse element.
        
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0]*1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
            
            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))
        prints::
            [5120, 100, 655360, 268435456]
        Equivalent form of C{expr.copy()} is just C{expr()}::
            integerM = integer().addParseAction(lambda toks: toks[0]*1024*1024) + Suppress("M")
        """
        cpy = copy.copy( self )
        cpy.parseAction = self.parseAction[:]
        cpy.ignoreExprs = self.ignoreExprs[:]
        if self.copyDefaultWhiteChars:
            cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
        return cpy

    def setName( self, name ):
        """
        Define name for this expression, makes debugging and exception messages clearer.
        
        Example::
            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        """
        self.name = name
        self.errmsg = "Expected " + self.name
        if hasattr(self,"exception"):
            self.exception.msg = self.errmsg
        return self

    def setResultsName( self, name, listAllMatches=False ):
        """
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original C{ParserElement} object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        C{expr("name")} in place of C{expr.setResultsName("name")} - 
        see L{I{__call__}<__call__>}.

        Example::
            date_str = (integer.setResultsName("year") + '/' 
                        + integer.setResultsName("month") + '/' 
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        """
        newself = self.copy()
        if name.endswith("*"):
            name = name[:-1]
            listAllMatches=True
        newself.resultsName = name
        newself.modalResults = not listAllMatches
        return newself

    def setBreak(self,breakFlag = True):
        """Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set C{breakFlag} to True to enable, False to
           disable.
        """
        if breakFlag:
            _parseMethod = self._parse
            def breaker(instring, loc, doActions=True, callPreParse=True):
                import pdb
                pdb.set_trace()
                return _parseMethod( instring, loc, doActions, callPreParse )
            breaker._originalParseMethod = _parseMethod
            self._parse = breaker
        else:
            if hasattr(self._parse,"_originalParseMethod"):
                self._parse = self._parse._originalParseMethod
        return self

    def setParseAction( self, *fns, **kwargs ):
        """
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as C{fn(s,loc,toks)},
        C{fn(loc,toks)}, C{fn(toks)}, or just C{fn()}, where:
         - s   = the original string being parsed (see note below)
         - loc = the location of the matching substring
         - toks = a list of the matched tokens, packaged as a C{L{ParseResults}} object
        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        Optional keyword arguments:
         - callDuringTry = (default=C{False}) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See L{I{parseString}<parseString>} for more information
        on parsing strings containing C{<TAB>}s, and suggested methods to maintain a
        consistent view of the parsed string, the parse location, and line and column
        positions within the parsed string.
        
        Example::
            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        """
        self.parseAction = list(map(_trim_arity, list(fns)))
        self.callDuringTry = kwargs.get("callDuringTry", False)
        return self

    def addParseAction( self, *fns, **kwargs ):
        """
        Add one or more parse actions to expression's list of parse actions. See L{I{setParseAction}<setParseAction>}.
        
        See examples in L{I{copy}<copy>}.
        """
        self.parseAction += list(map(_trim_arity, list(fns)))
        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def addCondition(self, *fns, **kwargs):
        """Add a boolean predicate function to expression's list of parse actions. See 
        L{I{setParseAction}<setParseAction>} for function call signatures. Unlike C{setParseAction}, 
        functions passed to C{addCondition} need to return boolean success/fail of the condition.

        Optional keyword arguments:
         - message = define a custom message to be used in the raised exception
         - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException
         
        Example::
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        """
        msg = kwargs.get("message", "failed user-defined condition")
        exc_type = ParseFatalException if kwargs.get("fatal", False) else ParseException
        for fn in fns:
            def pa(s,l,t):
                if not bool(_trim_arity(fn)(s,l,t)):
                    raise exc_type(s,l,msg)
            self.parseAction.append(pa)
        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def setFailAction( self, fn ):
        """Define action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           C{fn(s,loc,expr,err)} where:
            - s = string being parsed
            - loc = location where expression match was attempted and failed
            - expr = the parse expression that failed
            - err = the exception thrown
           The function returns no value.  It may throw C{L{ParseFatalException}}
           if it is desired to stop parsing immediately."""
        self.failAction = fn
        return self

    def _skipIgnorables( self, instring, loc ):
        exprsFound = True
        while exprsFound:
            exprsFound = False
            for e in self.ignoreExprs:
                try:
                    while 1:
                        loc,dummy = e._parse( instring, loc )
                        exprsFound = True
                except ParseException:
                    pass
        return loc

    def preParse( self, instring, loc ):
        if self.ignoreExprs:
            loc = self._skipIgnorables( instring, loc )

        if self.skipWhitespace:
            wt = self.whiteChars
            instrlen = len(instring)
            while loc < instrlen and instring[loc] in wt:
                loc += 1

        return loc

    def parseImpl( self, instring, loc, doActions=True ):
        return loc, []

    def postParse( self, instring, loc, tokenlist ):
        return tokenlist

    #~ @profile
    def _parseNoCache( self, instring, loc, doActions=True, callPreParse=True ):
        debugging = ( self.debug ) #and doActions )

        if debugging or self.failAction:
            #~ print ("Match",self,"at loc",loc,"(%d,%d)" % ( lineno(loc,instring), col(loc,instring) ))
            if (self.debugActions[0] ):
                self.debugActions[0]( instring, loc, self )
            if callPreParse and self.callPreparse:
                preloc = self.preParse( instring, loc )
            else:
                preloc = loc
            tokensStart = preloc
            try:
                try:
                    loc,tokens = self.parseImpl( instring, preloc, doActions )
                except IndexError:
                    raise ParseException( instring, len(instring), self.errmsg, self )
            except ParseBaseException as err:
                #~ print ("Exception raised:", err)
                if self.debugActions[2]:
                    self.debugActions[2]( instring, tokensStart, self, err )
                if self.failAction:
                    self.failAction( instring, tokensStart, self, err )
                raise
        else:
            if callPreParse and self.callPreparse:
                preloc = self.preParse( instring, loc )
            else:
                preloc = loc
            tokensStart = preloc
            if self.mayIndexError or preloc >= len(instring):
                try:
                    loc,tokens = self.parseImpl( instring, preloc, doActions )
                except IndexError:
                    raise ParseException( instring, len(instring), self.errmsg, self )
            else:
                loc,tokens = self.parseImpl( instring, preloc, doActions )

        tokens = self.postParse( instring, loc, tokens )

        retTokens = ParseResults( tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults )
        if self.parseAction and (doActions or self.callDuringTry):
            if debugging:
                try:
                    for fn in self.parseAction:
                        tokens = fn( instring, tokensStart, retTokens )
                        if tokens is not None:
                            retTokens = ParseResults( tokens,
                                                      self.resultsName,
                                                      asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
                                                      modal=self.modalResults )
                except ParseBaseException as err:
                    #~ print "Exception raised in user parse action:", err
                    if (self.debugActions[2] ):
                        self.debugActions[2]( instring, tokensStart, self, err )
                    raise
            else:
                for fn in self.parseAction:
                    tokens = fn( instring, tokensStart, retTokens )
                    if tokens is not None:
                        retTokens = ParseResults( tokens,
                                                  self.resultsName,
                                                  asList=self.saveAsList and isinstance(tokens,(ParseResults,list)),
                                                  modal=self.modalResults )
        if debugging:
            #~ print ("Matched",self,"->",retTokens.asList())
            if (self.debugActions[1] ):
                self.debugActions[1]( instring, tokensStart, loc, self, retTokens )

        return loc, retTokens

    def tryParse( self, instring, loc ):
        try:
            return self._parse( instring, loc, doActions=False )[0]
        except ParseFatalException:
            raise ParseException( instring, loc, self.errmsg, self)
    
    def canParseNext(self, instring, loc):
        try:
            self.tryParse(instring, loc)
        except (ParseException, IndexError):
            return False
        else:
            return True

    class _UnboundedCache(object):
        def __init__(self):
            cache = {}
            self.not_in_cache = not_in_cache = object()

            def get(self, key):
                return cache.get(key, not_in_cache)

            def set(self, key, value):
                cache[key] = value

            def clear(self):
                cache.clear()
                
            def cache_len(self):
                return len(cache)

            self.get = types.MethodType(get, self)
            self.set = types.MethodType(set, self)
            self.clear = types.MethodType(clear, self)
            self.__len__ = types.MethodType(cache_len, self)

    if _OrderedDict is not None:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = _OrderedDict()

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(cache) > size:
                        try:
                            cache.popitem(False)
                        except KeyError:
                            pass

                def clear(self):
                    cache.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    else:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = {}
                key_fifo = collections.deque([], size)

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(key_fifo) > size:
                        cache.pop(key_fifo.popleft(), None)
                    key_fifo.append(key)

                def clear(self):
                    cache.clear()
                    key_fifo.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    # argument cache for optimizing repeated calls when backtracking through recursive expressions
    packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail
    packrat_cache_lock = RLock()
    packrat_cache_stats = [0, 0]

    # this method gets repeatedly called during backtracking with the same arguments -
    # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression
    def _parseCache( self, instring, loc, doActions=True, callPreParse=True ):
        HIT, MISS = 0, 1
        lookup = (self, instring, loc, callPreParse, doActions)
        with ParserElement.packrat_cache_lock:
            cache = ParserElement.packrat_cache
            value = cache.get(lookup)
            if value is cache.not_in_cache:
                ParserElement.packrat_cache_stats[MISS] += 1
                try:
                    value = self._parseNoCache(instring, loc, doActions, callPreParse)
                except ParseBaseException as pe:
                    # cache a copy of the exception, without the traceback
                    cache.set(lookup, pe.__class__(*pe.args))
                    raise
                else:
                    cache.set(lookup, (value[0], value[1].copy()))
                    return value
            else:
                ParserElement.packrat_cache_stats[HIT] += 1
                if isinstance(value, Exception):
                    raise value
                return (value[0], value[1].copy())

    _parse = _parseNoCache

    @staticmethod
    def resetCache():
        ParserElement.packrat_cache.clear()
        ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats)

    _packratEnabled = False
    @staticmethod
    def enablePackrat(cache_size_limit=128):
        """Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.
           
           Parameters:
            - cache_size_limit - (default=C{128}) - if an integer value is provided
              will limit the size of the packrat cache; if None is passed, then
              the cache size will be unbounded; if 0 is passed, the cache will
              be effectively disabled.
            
           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method C{ParserElement.enablePackrat()}.  If
           your program uses C{psyco} to "compile as you go", you must call
           C{enablePackrat} before calling C{psyco.full()}.  If you do not do this,
           Python will crash.  For best results, call C{enablePackrat()} immediately
           after importing pyparsing.
           
           Example::
               import pyparsing
               pyparsing.ParserElement.enablePackrat()
        """
        if not ParserElement._packratEnabled:
            ParserElement._packratEnabled = True
            if cache_size_limit is None:
                ParserElement.packrat_cache = ParserElement._UnboundedCache()
            else:
                ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit)
            ParserElement._parse = ParserElement._parseCache

    def parseString( self, instring, parseAll=False ):
        """
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set C{parseAll} to True (equivalent to ending
        the grammar with C{L{StringEnd()}}).

        Note: C{parseString} implicitly calls C{expandtabs()} on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the C{loc} argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:
         - calling C{parseWithTabs} on your grammar before calling C{parseString}
           (see L{I{parseWithTabs}<parseWithTabs>})
         - define your parse action using the full C{(s,loc,toks)} signature, and
           reference the input string using the parse action's C{s} argument
         - explicitly expand the tabs in your input string before calling
           C{parseString}
        
        Example::
            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        """
        ParserElement.resetCache()
        if not self.streamlined:
            self.streamline()
            #~ self.saveAsList = True
        for e in self.ignoreExprs:
            e.streamline()
        if not self.keepTabs:
            instring = instring.expandtabs()
        try:
            loc, tokens = self._parse( instring, 0 )
            if parseAll:
                loc = self.preParse( instring, loc )
                se = Empty() + StringEnd()
                se._parse( instring, loc )
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc
        else:
            return tokens

    def scanString( self, instring, maxMatches=_MAX_INT, overlap=False ):
        """
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        C{maxMatches} argument, to clip scanning after 'n' matches are found.  If
        C{overlap} is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See L{I{parseString}<parseString>} for more information on parsing
        strings with embedded tabs.

        Example::
            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens,start,end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])
        
        prints::
        
            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        """
        if not self.streamlined:
            self.streamline()
        for e in self.ignoreExprs:
            e.streamline()

        if not self.keepTabs:
            instring = _ustr(instring).expandtabs()
        instrlen = len(instring)
        loc = 0
        preparseFn = self.preParse
        parseFn = self._parse
        ParserElement.resetCache()
        matches = 0
        try:
            while loc <= instrlen and matches < maxMatches:
                try:
                    preloc = preparseFn( instring, loc )
                    nextLoc,tokens = parseFn( instring, preloc, callPreParse=False )
                except ParseException:
                    loc = preloc+1
                else:
                    if nextLoc > loc:
                        matches += 1
                        yield tokens, preloc, nextLoc
                        if overlap:
                            nextloc = preparseFn( instring, loc )
                            if nextloc > loc:
                                loc = nextLoc
                            else:
                                loc += 1
                        else:
                            loc = nextLoc
                    else:
                        loc = preloc+1
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def transformString( self, instring ):
        """
        Extension to C{L{scanString}}, to modify matching text with modified tokens that may
        be returned from a parse action.  To use C{transformString}, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking C{transformString()} on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  C{transformString()} returns the resulting transformed string.
        
        Example::
            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())
            
            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))
        Prints::
            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        """
        out = []
        lastE = 0
        # force preservation of <TAB>s, to minimize unwanted transformation of string, and to
        # keep string locs straight between transformString and scanString
        self.keepTabs = True
        try:
            for t,s,e in self.scanString( instring ):
                out.append( instring[lastE:s] )
                if t:
                    if isinstance(t,ParseResults):
                        out += t.asList()
                    elif isinstance(t,list):
                        out += t
                    else:
                        out.append(t)
                lastE = e
            out.append(instring[lastE:])
            out = [o for o in out if o]
            return "".join(map(_ustr,_flatten(out)))
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def searchString( self, instring, maxMatches=_MAX_INT ):
        """
        Another extension to C{L{scanString}}, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        C{maxMatches} argument, to clip searching after 'n' matches are found.
        
        Example::
            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())
            
            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))
        prints::
            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        """
        try:
            return ParseResults([ t for t,s,e in self.scanString( instring, maxMatches ) ])
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False):
        """
        Generator method to split a string using the given expression as a separator.
        May be called with optional C{maxsplit} argument, to limit the number of splits;
        and the optional C{includeSeparators} argument (default=C{False}), if the separating
        matching text should be included in the split results.
        
        Example::        
            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))
        prints::
            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        """
        splits = 0
        last = 0
        for t,s,e in self.scanString(instring, maxMatches=maxsplit):
            yield instring[last:s]
            if includeSeparators:
                yield t[0]
            last = e
        yield instring[last:]

    def __add__(self, other ):
        """
        Implementation of + operator - returns C{L{And}}. Adding strings to a ParserElement
        converts them to L{Literal}s by default.
        
        Example::
            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))
        Prints::
            Hello, World! -> ['Hello', ',', 'World', '!']
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return And( [ self, other ] )

    def __radd__(self, other ):
        """
        Implementation of + operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other + self

    def __sub__(self, other):
        """
        Implementation of - operator, returns C{L{And}} with error stop
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return self + And._ErrorStop() + other

    def __rsub__(self, other ):
        """
        Implementation of - operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other - self

    def __mul__(self,other):
        """
        Implementation of * operator, allows use of C{expr * 3} in place of
        C{expr + expr + expr}.  Expressions may also me multiplied by a 2-integer
        tuple, similar to C{{min,max}} multipliers in regular expressions.  Tuples
        may also include C{None} as in:
         - C{expr*(n,None)} or C{expr*(n,)} is equivalent
              to C{expr*n + L{ZeroOrMore}(expr)}
              (read as "at least n instances of C{expr}")
         - C{expr*(None,n)} is equivalent to C{expr*(0,n)}
              (read as "0 to n instances of C{expr}")
         - C{expr*(None,None)} is equivalent to C{L{ZeroOrMore}(expr)}
         - C{expr*(1,None)} is equivalent to C{L{OneOrMore}(expr)}

        Note that C{expr*(None,n)} does not raise an exception if
        more than n exprs exist in the input stream; that is,
        C{expr*(None,n)} does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        C{expr*(None,n) + ~expr}
        """
        if isinstance(other,int):
            minElements, optElements = other,0
        elif isinstance(other,tuple):
            other = (other + (None, None))[:2]
            if other[0] is None:
                other = (0, other[1])
            if isinstance(other[0],int) and other[1] is None:
                if other[0] == 0:
                    return ZeroOrMore(self)
                if other[0] == 1:
                    return OneOrMore(self)
                else:
                    return self*other[0] + ZeroOrMore(self)
            elif isinstance(other[0],int) and isinstance(other[1],int):
                minElements, optElements = other
                optElements -= minElements
            else:
                raise TypeError("cannot multiply 'ParserElement' and ('%s','%s') objects", type(other[0]),type(other[1]))
        else:
            raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other))

        if minElements < 0:
            raise ValueError("cannot multiply ParserElement by negative value")
        if optElements < 0:
            raise ValueError("second tuple value must be greater or equal to first tuple value")
        if minElements == optElements == 0:
            raise ValueError("cannot multiply ParserElement by 0 or (0,0)")

        if (optElements):
            def makeOptionalList(n):
                if n>1:
                    return Optional(self + makeOptionalList(n-1))
                else:
                    return Optional(self)
            if minElements:
                if minElements == 1:
                    ret = self + makeOptionalList(optElements)
                else:
                    ret = And([self]*minElements) + makeOptionalList(optElements)
            else:
                ret = makeOptionalList(optElements)
        else:
            if minElements == 1:
                ret = self
            else:
                ret = And([self]*minElements)
        return ret

    def __rmul__(self, other):
        return self.__mul__(other)

    def __or__(self, other ):
        """
        Implementation of | operator - returns C{L{MatchFirst}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return MatchFirst( [ self, other ] )

    def __ror__(self, other ):
        """
        Implementation of | operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other | self

    def __xor__(self, other ):
        """
        Implementation of ^ operator - returns C{L{Or}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return Or( [ self, other ] )

    def __rxor__(self, other ):
        """
        Implementation of ^ operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other ^ self

    def __and__(self, other ):
        """
        Implementation of & operator - returns C{L{Each}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return Each( [ self, other ] )

    def __rand__(self, other ):
        """
        Implementation of & operator when left operand is not a C{L{ParserElement}}
        """
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        if not isinstance( other, ParserElement ):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                    SyntaxWarning, stacklevel=2)
            return None
        return other & self

    def __invert__( self ):
        """
        Implementation of ~ operator - returns C{L{NotAny}}
        """
        return NotAny( self )

    def __call__(self, name=None):
        """
        Shortcut for C{L{setResultsName}}, with C{listAllMatches=False}.
        
        If C{name} is given with a trailing C{'*'} character, then C{listAllMatches} will be
        passed as C{True}.
           
        If C{name} is omitted, same as calling C{L{copy}}.

        Example::
            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums+"-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums+"-")("socsecno")             
        """
        if name is not None:
            return self.setResultsName(name)
        else:
            return self.copy()

    def suppress( self ):
        """
        Suppresses the output of this C{ParserElement}; useful to keep punctuation from
        cluttering up returned output.
        """
        return Suppress( self )

    def leaveWhitespace( self ):
        """
        Disables the skipping of whitespace before matching the characters in the
        C{ParserElement}'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        """
        self.skipWhitespace = False
        return self

    def setWhitespaceChars( self, chars ):
        """
        Overrides the default whitespace chars
        """
        self.skipWhitespace = True
        self.whiteChars = chars
        self.copyDefaultWhiteChars = False
        return self

    def parseWithTabs( self ):
        """
        Overrides default behavior to expand C{<TAB>}s to spaces before parsing the input string.
        Must be called before C{parseString} when the input grammar contains elements that
        match C{<TAB>} characters.
        """
        self.keepTabs = True
        return self

    def ignore( self, other ):
        """
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.
        
        Example::
            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']
            
            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        """
        if isinstance(other, basestring):
            other = Suppress(other)

        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                self.ignoreExprs.append(other)
        else:
            self.ignoreExprs.append( Suppress( other.copy() ) )
        return self

    def setDebugActions( self, startAction, successAction, exceptionAction ):
        """
        Enable display of debugging messages while doing pattern matching.
        """
        self.debugActions = (startAction or _defaultStartDebugAction,
                             successAction or _defaultSuccessDebugAction,
                             exceptionAction or _defaultExceptionDebugAction)
        self.debug = True
        return self

    def setDebug( self, flag=True ):
        """
        Enable display of debugging messages while doing pattern matching.
        Set C{flag} to True to enable, False to disable.

        Example::
            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer
            
            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")
        
        prints::
            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using L{setDebugActions}. Prior to attempting
        to match the C{wd} expression, the debugging message C{"Match <exprname> at loc <n>(<line>,<col>)"}
        is shown. Then if the parse succeeds, a C{"Matched"} message is shown, or an C{"Exception raised"}
        message is shown. Also note the use of L{setName} to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the C{Word} expression without calling C{setName} is C{"W:(ABCD...)"}.
        """
        if flag:
            self.setDebugActions( _defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction )
        else:
            self.debug = False
        return self

    def __str__( self ):
        return self.name

    def __repr__( self ):
        return _ustr(self)

    def streamline( self ):
        self.streamlined = True
        self.strRepr = None
        return self

    def checkRecursion( self, parseElementList ):
        pass

    def validate( self, validateTrace=[] ):
        """
        Check defined expressions for valid structure, check for infinite recursive definitions.
        """
        self.checkRecursion( [] )

    def parseFile( self, file_or_filename, parseAll=False ):
        """
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        """
        try:
            file_contents = file_or_filename.read()
        except AttributeError:
            with open(file_or_filename, "r") as f:
                file_contents = f.read()
        try:
            return self.parseString(file_contents, parseAll)
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clears out pyparsing internal stack trace
                raise exc

    def __eq__(self,other):
        if isinstance(other, ParserElement):
            return self is other or vars(self) == vars(other)
        elif isinstance(other, basestring):
            return self.matches(other)
        else:
            return super(ParserElement,self)==other

    def __ne__(self,other):
        return not (self == other)

    def __hash__(self):
        return hash(id(self))

    def __req__(self,other):
        return self == other

    def __rne__(self,other):
        return not (self == other)

    def matches(self, testString, parseAll=True):
        """
        Method for quick testing of a parser against a test string. Good for simple 
        inline microtests of sub expressions while building up larger parser.
           
        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests
            
        Example::
            expr = Word(nums)
            assert expr.matches("100")
        """
        try:
            self.parseString(_ustr(testString), parseAll=parseAll)
            return True
        except ParseBaseException:
            return False
                
    def runTests(self, tests, parseAll=True, comment='#', fullDump=True, printResults=True, failureTests=False):
        """
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.
           
        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default=C{True}) - flag to pass to C{L{parseString}} when running tests           
         - comment - (default=C{'#'}) - expression for indicating embedded comments in the test 
              string; pass None to disable comment filtering
         - fullDump - (default=C{True}) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default=C{True}) prints test output to stdout
         - failureTests - (default=C{False}) indicates if these tests are expected to fail parsing

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if C{failureTests} is True), and the results contain a list of lines of each 
        test's output
        
        Example::
            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")
        prints::
            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success
            
            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines")
        
        (Note that this is a raw string literal, you must include the leading 'r'.)
        """
        if isinstance(tests, basestring):
            tests = list(map(str.strip, tests.rstrip().splitlines()))
        if isinstance(comment, basestring):
            comment = Literal(comment)
        allResults = []
        comments = []
        success = True
        for t in tests:
            if comment is not None and comment.matches(t, False) or comments and not t:
                comments.append(t)
                continue
            if not t:
                continue
            out = ['\n'.join(comments), t]
            comments = []
            try:
                t = t.replace(r'\n','\n')
                result = self.parseString(t, parseAll=parseAll)
                out.append(result.dump(full=fullDump))
                success = success and not failureTests
            except ParseBaseException as pe:
                fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else ""
                if '\n' in t:
                    out.append(line(pe.loc, t))
                    out.append(' '*(col(pe.loc,t)-1) + '^' + fatal)
                else:
                    out.append(' '*pe.loc + '^' + fatal)
                out.append("FAIL: " + str(pe))
                success = success and failureTests
                result = pe
            except Exception as exc:
                out.append("FAIL-EXCEPTION: " + str(exc))
                success = success and failureTests
                result = exc

            if printResults:
                if fullDump:
                    out.append('')
                print('\n'.join(out))

            allResults.append((t, result))
        
        return success, allResults

        
class Token(ParserElement):
    """
    Abstract C{ParserElement} subclass, for defining atomic matching patterns.
    """
    def __init__( self ):
        super(Token,self).__init__( savelist=False )


class Empty(Token):
    """
    An empty token, will always match.
    """
    def __init__( self ):
        super(Empty,self).__init__()
        self.name = "Empty"
        self.mayReturnEmpty = True
        self.mayIndexError = False


class NoMatch(Token):
    """
    A token that will never match.
    """
    def __init__( self ):
        super(NoMatch,self).__init__()
        self.name = "NoMatch"
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.errmsg = "Unmatchable token"

    def parseImpl( self, instring, loc, doActions=True ):
        raise ParseException(instring, loc, self.errmsg, self)


class Literal(Token):
    """
    Token to exactly match a specified string.
    
    Example::
        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"
    
    For case-insensitive matching, use L{CaselessLiteral}.
    
    For keyword matching (force word break before and after the matched string),
    use L{Keyword} or L{CaselessKeyword}.
    """
    def __init__( self, matchString ):
        super(Literal,self).__init__()
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Literal; use Empty() instead",
                            SyntaxWarning, stacklevel=2)
            self.__class__ = Empty
        self.name = '"%s"' % _ustr(self.match)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False

    # Performance tuning: this routine gets called a *lot*
    # if this is a single character match string  and the first character matches,
    # short-circuit as quickly as possible, and avoid calling startswith
    #~ @profile
    def parseImpl( self, instring, loc, doActions=True ):
        if (instring[loc] == self.firstMatchChar and
            (self.matchLen==1 or instring.startswith(self.match,loc)) ):
            return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)
_L = Literal
ParserElement._literalStringClass = Literal

class Keyword(Token):
    """
    Token to exactly match a specified string as a keyword, that is, it must be
    immediately followed by a non-keyword character.  Compare with C{L{Literal}}:
     - C{Literal("if")} will match the leading C{'if'} in C{'ifAndOnlyIf'}.
     - C{Keyword("if")} will not; it will only match the leading C{'if'} in C{'if x=1'}, or C{'if(y==2)'}
    Accepts two optional constructor arguments in addition to the keyword string:
     - C{identChars} is a string of characters that would be valid identifier characters,
          defaulting to all alphanumerics + "_" and "$"
     - C{caseless} allows case-insensitive matching, default is C{False}.
       
    Example::
        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use L{CaselessKeyword}.
    """
    DEFAULT_KEYWORD_CHARS = alphanums+"_$"

    def __init__( self, matchString, identChars=None, caseless=False ):
        super(Keyword,self).__init__()
        if identChars is None:
            identChars = Keyword.DEFAULT_KEYWORD_CHARS
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Keyword; use Empty() instead",
                            SyntaxWarning, stacklevel=2)
        self.name = '"%s"' % self.match
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False
        self.caseless = caseless
        if caseless:
            self.caselessmatch = matchString.upper()
            identChars = identChars.upper()
        self.identChars = set(identChars)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.caseless:
            if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
                 (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) and
                 (loc == 0 or instring[loc-1].upper() not in self.identChars) ):
                return loc+self.matchLen, self.match
        else:
            if (instring[loc] == self.firstMatchChar and
                (self.matchLen==1 or instring.startswith(self.match,loc)) and
                (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen] not in self.identChars) and
                (loc == 0 or instring[loc-1] not in self.identChars) ):
                return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)

    def copy(self):
        c = super(Keyword,self).copy()
        c.identChars = Keyword.DEFAULT_KEYWORD_CHARS
        return c

    @staticmethod
    def setDefaultKeywordChars( chars ):
        """Overrides the default Keyword chars
        """
        Keyword.DEFAULT_KEYWORD_CHARS = chars

class CaselessLiteral(Literal):
    """
    Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::
        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']
        
    (Contrast with example for L{CaselessKeyword}.)
    """
    def __init__( self, matchString ):
        super(CaselessLiteral,self).__init__( matchString.upper() )
        # Preserve the defining literal.
        self.returnString = matchString
        self.name = "'%s'" % self.returnString
        self.errmsg = "Expected " + self.name

    def parseImpl( self, instring, loc, doActions=True ):
        if instring[ loc:loc+self.matchLen ].upper() == self.match:
            return loc+self.matchLen, self.returnString
        raise ParseException(instring, loc, self.errmsg, self)

class CaselessKeyword(Keyword):
    """
    Caseless version of L{Keyword}.

    Example::
        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']
        
    (Contrast with example for L{CaselessLiteral}.)
    """
    def __init__( self, matchString, identChars=None ):
        super(CaselessKeyword,self).__init__( matchString, identChars, caseless=True )

    def parseImpl( self, instring, loc, doActions=True ):
        if ( (instring[ loc:loc+self.matchLen ].upper() == self.caselessmatch) and
             (loc >= len(instring)-self.matchLen or instring[loc+self.matchLen].upper() not in self.identChars) ):
            return loc+self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)

class CloseMatch(Token):
    """
    A variation on L{Literal} which matches "close" matches, that is, 
    strings with at most 'n' mismatching characters. C{CloseMatch} takes parameters:
     - C{match_string} - string to be matched
     - C{maxMismatches} - (C{default=1}) maximum number of mismatches allowed to count as a match
    
    The results from a successful parse will contain the matched text from the input string and the following named results:
     - C{mismatches} - a list of the positions within the match_string where mismatches were found
     - C{original} - the original match_string used to compare against the input string
    
    If C{mismatches} is an empty list, then the match was an exact match.
    
    Example::
        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    """
    def __init__(self, match_string, maxMismatches=1):
        super(CloseMatch,self).__init__()
        self.name = match_string
        self.match_string = match_string
        self.maxMismatches = maxMismatches
        self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches)
        self.mayIndexError = False
        self.mayReturnEmpty = False

    def parseImpl( self, instring, loc, doActions=True ):
        start = loc
        instrlen = len(instring)
        maxloc = start + len(self.match_string)

        if maxloc <= instrlen:
            match_string = self.match_string
            match_stringloc = 0
            mismatches = []
            maxMismatches = self.maxMismatches

            for match_stringloc,s_m in enumerate(zip(instring[loc:maxloc], self.match_string)):
                src,mat = s_m
                if src != mat:
                    mismatches.append(match_stringloc)
                    if len(mismatches) > maxMismatches:
                        break
            else:
                loc = match_stringloc + 1
                results = ParseResults([instring[start:loc]])
                results['original'] = self.match_string
                results['mismatches'] = mismatches
                return loc, results

        raise ParseException(instring, loc, self.errmsg, self)


class Word(Token):
    """
    Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters,
    an optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction. An optional
    C{excludeChars} parameter can list characters that might be found in 
    the input C{bodyChars} string; useful to define a word of all printables
    except for one or two characters, for instance.
    
    L{srange} is useful for defining custom character set strings for defining 
    C{Word} expressions, using range notation from regular expression character sets.
    
    A common mistake is to use C{Word} to match a specific literal string, as in 
    C{Word("Address")}. Remember that C{Word} uses the string argument to define
    I{sets} of matchable characters. This expression would match "Add", "AAA",
    "dAred", or any other word made up of the characters 'A', 'd', 'r', 'e', and 's'.
    To match an exact literal string, use L{Literal} or L{Keyword}.

    pyparsing includes helper strings for building Words:
     - L{alphas}
     - L{nums}
     - L{alphanums}
     - L{hexnums}
     - L{alphas8bit} (alphabetic characters in ASCII range 128-255 - accented, tilded, umlauted, etc.)
     - L{punc8bit} (non-alphabetic characters in ASCII range 128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - L{printables} (any non-whitespace character)

    Example::
        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))
        
        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums+'-')
        
        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")
        
        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    """
    def __init__( self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None ):
        super(Word,self).__init__()
        if excludeChars:
            initChars = ''.join(c for c in initChars if c not in excludeChars)
            if bodyChars:
                bodyChars = ''.join(c for c in bodyChars if c not in excludeChars)
        self.initCharsOrig = initChars
        self.initChars = set(initChars)
        if bodyChars :
            self.bodyCharsOrig = bodyChars
            self.bodyChars = set(bodyChars)
        else:
            self.bodyCharsOrig = initChars
            self.bodyChars = set(initChars)

        self.maxSpecified = max > 0

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.asKeyword = asKeyword

        if ' ' not in self.initCharsOrig+self.bodyCharsOrig and (min==1 and max==0 and exact==0):
            if self.bodyCharsOrig == self.initCharsOrig:
                self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig)
            elif len(self.initCharsOrig) == 1:
                self.reString = "%s[%s]*" % \
                                      (re.escape(self.initCharsOrig),
                                      _escapeRegexRangeChars(self.bodyCharsOrig),)
            else:
                self.reString = "[%s][%s]*" % \
                                      (_escapeRegexRangeChars(self.initCharsOrig),
                                      _escapeRegexRangeChars(self.bodyCharsOrig),)
            if self.asKeyword:
                self.reString = r"\b"+self.reString+r"\b"
            try:
                self.re = re.compile( self.reString )
            except Exception:
                self.re = None

    def parseImpl( self, instring, loc, doActions=True ):
        if self.re:
            result = self.re.match(instring,loc)
            if not result:
                raise ParseException(instring, loc, self.errmsg, self)

            loc = result.end()
            return loc, result.group()

        if not(instring[ loc ] in self.initChars):
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        instrlen = len(instring)
        bodychars = self.bodyChars
        maxloc = start + self.maxLen
        maxloc = min( maxloc, instrlen )
        while loc < maxloc and instring[loc] in bodychars:
            loc += 1

        throwException = False
        if loc - start < self.minLen:
            throwException = True
        if self.maxSpecified and loc < instrlen and instring[loc] in bodychars:
            throwException = True
        if self.asKeyword:
            if (start>0 and instring[start-1] in bodychars) or (loc<instrlen and instring[loc] in bodychars):
                throwException = True

        if throwException:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__( self ):
        try:
            return super(Word,self).__str__()
        except Exception:
            pass


        if self.strRepr is None:

            def charsAsStr(s):
                if len(s)>4:
                    return s[:4]+"..."
                else:
                    return s

            if ( self.initCharsOrig != self.bodyCharsOrig ):
                self.strRepr = "W:(%s,%s)" % ( charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig) )
            else:
                self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig)

        return self.strRepr


class Regex(Token):
    r"""
    Token for matching strings that match a given regular expression.
    Defined with string specifying the regular expression in a form recognized by the inbuilt Python re module.
    If the given regex contains named groups (defined using C{(?P<name>...)}), these will be preserved as 
    named parse results.

    Example::
        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: http://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?C{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")
    """
    compiledREtype = type(re.compile("[A-Z]"))
    def __init__( self, pattern, flags=0):
        """The parameters C{pattern} and C{flags} are passed to the C{re.compile()} function as-is. See the Python C{re} module for an explanation of the acceptable patterns and flags."""
        super(Regex,self).__init__()

        if isinstance(pattern, basestring):
            if not pattern:
                warnings.warn("null string passed to Regex; use Empty() instead",
                        SyntaxWarning, stacklevel=2)

            self.pattern = pattern
            self.flags = flags

            try:
                self.re = re.compile(self.pattern, self.flags)
                self.reString = self.pattern
            except sre_constants.error:
                warnings.warn("invalid pattern (%s) passed to Regex" % pattern,
                    SyntaxWarning, stacklevel=2)
                raise

        elif isinstance(pattern, Regex.compiledREtype):
            self.re = pattern
            self.pattern = \
            self.reString = str(pattern)
            self.flags = flags
            
        else:
            raise ValueError("Regex may only be constructed with a string or a compiled RE object")

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        result = self.re.match(instring,loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        d = result.groupdict()
        ret = ParseResults(result.group())
        if d:
            for k in d:
                ret[k] = d[k]
        return loc,ret

    def __str__( self ):
        try:
            return super(Regex,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "Re:(%s)" % repr(self.pattern)

        return self.strRepr


class QuotedString(Token):
    r"""
    Token for matching strings that are delimited by quoting characters.
    
    Defined with the following parameters:
        - quoteChar - string of one or more characters defining the quote delimiting string
        - escChar - character to escape quotes, typically backslash (default=C{None})
        - escQuote - special quote sequence to escape an embedded quote string (such as SQL's "" to escape an embedded ") (default=C{None})
        - multiline - boolean indicating whether quotes can span multiple lines (default=C{False})
        - unquoteResults - boolean indicating whether the matched text should be unquoted (default=C{True})
        - endQuoteChar - string of one or more characters defining the end of the quote delimited string (default=C{None} => same as quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace (C{'\t'}, C{'\n'}, etc.) to actual whitespace (default=C{True})

    Example::
        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))
    prints::
        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    """
    def __init__( self, quoteChar, escChar=None, escQuote=None, multiline=False, unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True):
        super(QuotedString,self).__init__()

        # remove white space from quote chars - wont work anyway
        quoteChar = quoteChar.strip()
        if not quoteChar:
            warnings.warn("quoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
            raise SyntaxError()

        if endQuoteChar is None:
            endQuoteChar = quoteChar
        else:
            endQuoteChar = endQuoteChar.strip()
            if not endQuoteChar:
                warnings.warn("endQuoteChar cannot be the empty string",SyntaxWarning,stacklevel=2)
                raise SyntaxError()

        self.quoteChar = quoteChar
        self.quoteCharLen = len(quoteChar)
        self.firstQuoteChar = quoteChar[0]
        self.endQuoteChar = endQuoteChar
        self.endQuoteCharLen = len(endQuoteChar)
        self.escChar = escChar
        self.escQuote = escQuote
        self.unquoteResults = unquoteResults
        self.convertWhitespaceEscapes = convertWhitespaceEscapes

        if multiline:
            self.flags = re.MULTILINE | re.DOTALL
            self.pattern = r'%s(?:[^%s%s]' % \
                ( re.escape(self.quoteChar),
                  _escapeRegexRangeChars(self.endQuoteChar[0]),
                  (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
        else:
            self.flags = 0
            self.pattern = r'%s(?:[^%s\n\r%s]' % \
                ( re.escape(self.quoteChar),
                  _escapeRegexRangeChars(self.endQuoteChar[0]),
                  (escChar is not None and _escapeRegexRangeChars(escChar) or '') )
        if len(self.endQuoteChar) > 1:
            self.pattern += (
                '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]),
                                               _escapeRegexRangeChars(self.endQuoteChar[i]))
                                    for i in range(len(self.endQuoteChar)-1,0,-1)) + ')'
                )
        if escQuote:
            self.pattern += (r'|(?:%s)' % re.escape(escQuote))
        if escChar:
            self.pattern += (r'|(?:%s.)' % re.escape(escChar))
            self.escCharReplacePattern = re.escape(self.escChar)+"(.)"
        self.pattern += (r')*%s' % re.escape(self.endQuoteChar))

        try:
            self.re = re.compile(self.pattern, self.flags)
            self.reString = self.pattern
        except sre_constants.error:
            warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern,
                SyntaxWarning, stacklevel=2)
            raise

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        result = instring[loc] == self.firstQuoteChar and self.re.match(instring,loc) or None
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = result.group()

        if self.unquoteResults:

            # strip off quotes
            ret = ret[self.quoteCharLen:-self.endQuoteCharLen]

            if isinstance(ret,basestring):
                # replace escaped whitespace
                if '\\' in ret and self.convertWhitespaceEscapes:
                    ws_map = {
                        r'\t' : '\t',
                        r'\n' : '\n',
                        r'\f' : '\f',
                        r'\r' : '\r',
                    }
                    for wslit,wschar in ws_map.items():
                        ret = ret.replace(wslit, wschar)

                # replace escaped characters
                if self.escChar:
                    ret = re.sub(self.escCharReplacePattern, r"\g<1>", ret)

                # replace escaped quotes
                if self.escQuote:
                    ret = ret.replace(self.escQuote, self.endQuoteChar)

        return loc, ret

    def __str__( self ):
        try:
            return super(QuotedString,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar)

        return self.strRepr


class CharsNotIn(Token):
    """
    Token for matching words composed of characters I{not} in a given set (will
    include whitespace in matched characters if not listed in the provided exclusion set - see example).
    Defined with string containing all disallowed characters, and an optional
    minimum, maximum, and/or exact length.  The default value for C{min} is 1 (a
    minimum value < 1 is not valid); the default values for C{max} and C{exact}
    are 0, meaning no maximum or exact length restriction.

    Example::
        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))
    prints::
        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    """
    def __init__( self, notChars, min=1, max=0, exact=0 ):
        super(CharsNotIn,self).__init__()
        self.skipWhitespace = False
        self.notChars = notChars

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = ( self.minLen == 0 )
        self.mayIndexError = False

    def parseImpl( self, instring, loc, doActions=True ):
        if instring[loc] in self.notChars:
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        notchars = self.notChars
        maxlen = min( start+self.maxLen, len(instring) )
        while loc < maxlen and \
              (instring[loc] not in notchars):
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__( self ):
        try:
            return super(CharsNotIn, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            if len(self.notChars) > 4:
                self.strRepr = "!W:(%s...)" % self.notChars[:4]
            else:
                self.strRepr = "!W:(%s)" % self.notChars

        return self.strRepr

class White(Token):
    """
    Special matching class for matching whitespace.  Normally, whitespace is ignored
    by pyparsing grammars.  This class is included when some whitespace structures
    are significant.  Define with a string containing the whitespace characters to be
    matched; default is C{" \\t\\r\\n"}.  Also takes optional C{min}, C{max}, and C{exact} arguments,
    as defined for the C{L{Word}} class.
    """
    whiteStrs = {
        " " : "<SPC>",
        "\t": "<TAB>",
        "\n": "<LF>",
        "\r": "<CR>",
        "\f": "<FF>",
        }
    def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0):
        super(White,self).__init__()
        self.matchWhite = ws
        self.setWhitespaceChars( "".join(c for c in self.whiteChars if c not in self.matchWhite) )
        #~ self.leaveWhitespace()
        self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite))
        self.mayReturnEmpty = True
        self.errmsg = "Expected " + self.name

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

    def parseImpl( self, instring, loc, doActions=True ):
        if not(instring[ loc ] in self.matchWhite):
            raise ParseException(instring, loc, self.errmsg, self)
        start = loc
        loc += 1
        maxloc = start + self.maxLen
        maxloc = min( maxloc, len(instring) )
        while loc < maxloc and instring[loc] in self.matchWhite:
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]


class _PositionToken(Token):
    def __init__( self ):
        super(_PositionToken,self).__init__()
        self.name=self.__class__.__name__
        self.mayReturnEmpty = True
        self.mayIndexError = False

class GoToColumn(_PositionToken):
    """
    Token to advance to a specific column of input text; useful for tabular report scraping.
    """
    def __init__( self, colno ):
        super(GoToColumn,self).__init__()
        self.col = colno

    def preParse( self, instring, loc ):
        if col(loc,instring) != self.col:
            instrlen = len(instring)
            if self.ignoreExprs:
                loc = self._skipIgnorables( instring, loc )
            while loc < instrlen and instring[loc].isspace() and col( loc, instring ) != self.col :
                loc += 1
        return loc

    def parseImpl( self, instring, loc, doActions=True ):
        thiscol = col( loc, instring )
        if thiscol > self.col:
            raise ParseException( instring, loc, "Text not in expected column", self )
        newloc = loc + self.col - thiscol
        ret = instring[ loc: newloc ]
        return newloc, ret


class LineStart(_PositionToken):
    """
    Matches if current position is at the beginning of a line within the parse string
    
    Example::
    
        test = '''\
        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)
    
    Prints::
        ['AAA', ' this line']
        ['AAA', ' and this line']    

    """
    def __init__( self ):
        super(LineStart,self).__init__()
        self.errmsg = "Expected start of line"

    def parseImpl( self, instring, loc, doActions=True ):
        if col(loc, instring) == 1:
            return loc, []
        raise ParseException(instring, loc, self.errmsg, self)

class LineEnd(_PositionToken):
    """
    Matches if current position is at the end of a line within the parse string
    """
    def __init__( self ):
        super(LineEnd,self).__init__()
        self.setWhitespaceChars( ParserElement.DEFAULT_WHITE_CHARS.replace("\n","") )
        self.errmsg = "Expected end of line"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc<len(instring):
            if instring[loc] == "\n":
                return loc+1, "\n"
            else:
                raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc+1, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class StringStart(_PositionToken):
    """
    Matches if current position is at the beginning of the parse string
    """
    def __init__( self ):
        super(StringStart,self).__init__()
        self.errmsg = "Expected start of text"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc != 0:
            # see if entire string up to here is just whitespace and ignoreables
            if loc != self.preParse( instring, 0 ):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class StringEnd(_PositionToken):
    """
    Matches if current position is at the end of the parse string
    """
    def __init__( self ):
        super(StringEnd,self).__init__()
        self.errmsg = "Expected end of text"

    def parseImpl( self, instring, loc, doActions=True ):
        if loc < len(instring):
            raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc+1, []
        elif loc > len(instring):
            return loc, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class WordStart(_PositionToken):
    """
    Matches if the current position is at the beginning of a Word, and
    is not preceded by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
    use C{WordStart(alphanums)}. C{WordStart} will also match at the beginning of
    the string being parsed, or at the beginning of a line.
    """
    def __init__(self, wordChars = printables):
        super(WordStart,self).__init__()
        self.wordChars = set(wordChars)
        self.errmsg = "Not at the start of a word"

    def parseImpl(self, instring, loc, doActions=True ):
        if loc != 0:
            if (instring[loc-1] in self.wordChars or
                instring[loc] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class WordEnd(_PositionToken):
    """
    Matches if the current position is at the end of a Word, and
    is not followed by any character in a given set of C{wordChars}
    (default=C{printables}). To emulate the C{\b} behavior of regular expressions,
    use C{WordEnd(alphanums)}. C{WordEnd} will also match at the end of
    the string being parsed, or at the end of a line.
    """
    def __init__(self, wordChars = printables):
        super(WordEnd,self).__init__()
        self.wordChars = set(wordChars)
        self.skipWhitespace = False
        self.errmsg = "Not at the end of a word"

    def parseImpl(self, instring, loc, doActions=True ):
        instrlen = len(instring)
        if instrlen>0 and loc<instrlen:
            if (instring[loc] in self.wordChars or
                instring[loc-1] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []


class ParseExpression(ParserElement):
    """
    Abstract subclass of ParserElement, for combining and post-processing parsed tokens.
    """
    def __init__( self, exprs, savelist = False ):
        super(ParseExpression,self).__init__(savelist)
        if isinstance( exprs, _generatorType ):
            exprs = list(exprs)

        if isinstance( exprs, basestring ):
            self.exprs = [ ParserElement._literalStringClass( exprs ) ]
        elif isinstance( exprs, Iterable ):
            exprs = list(exprs)
            # if sequence of strings provided, wrap with Literal
            if all(isinstance(expr, basestring) for expr in exprs):
                exprs = map(ParserElement._literalStringClass, exprs)
            self.exprs = list(exprs)
        else:
            try:
                self.exprs = list( exprs )
            except TypeError:
                self.exprs = [ exprs ]
        self.callPreparse = False

    def __getitem__( self, i ):
        return self.exprs[i]

    def append( self, other ):
        self.exprs.append( other )
        self.strRepr = None
        return self

    def leaveWhitespace( self ):
        """Extends C{leaveWhitespace} defined in base class, and also invokes C{leaveWhitespace} on
           all contained expressions."""
        self.skipWhitespace = False
        self.exprs = [ e.copy() for e in self.exprs ]
        for e in self.exprs:
            e.leaveWhitespace()
        return self

    def ignore( self, other ):
        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                super( ParseExpression, self).ignore( other )
                for e in self.exprs:
                    e.ignore( self.ignoreExprs[-1] )
        else:
            super( ParseExpression, self).ignore( other )
            for e in self.exprs:
                e.ignore( self.ignoreExprs[-1] )
        return self

    def __str__( self ):
        try:
            return super(ParseExpression,self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.exprs) )
        return self.strRepr

    def streamline( self ):
        super(ParseExpression,self).streamline()

        for e in self.exprs:
            e.streamline()

        # collapse nested And's of the form And( And( And( a,b), c), d) to And( a,b,c,d )
        # but only if there are no parse actions or resultsNames on the nested And's
        # (likewise for Or's and MatchFirst's)
        if ( len(self.exprs) == 2 ):
            other = self.exprs[0]
            if ( isinstance( other, self.__class__ ) and
                  not(other.parseAction) and
                  other.resultsName is None and
                  not other.debug ):
                self.exprs = other.exprs[:] + [ self.exprs[1] ]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

            other = self.exprs[-1]
            if ( isinstance( other, self.__class__ ) and
                  not(other.parseAction) and
                  other.resultsName is None and
                  not other.debug ):
                self.exprs = self.exprs[:-1] + other.exprs[:]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

        self.errmsg = "Expected " + _ustr(self)
        
        return self

    def setResultsName( self, name, listAllMatches=False ):
        ret = super(ParseExpression,self).setResultsName(name,listAllMatches)
        return ret

    def validate( self, validateTrace=[] ):
        tmp = validateTrace[:]+[self]
        for e in self.exprs:
            e.validate(tmp)
        self.checkRecursion( [] )
        
    def copy(self):
        ret = super(ParseExpression,self).copy()
        ret.exprs = [e.copy() for e in self.exprs]
        return ret

class And(ParseExpression):
    """
    Requires all given C{ParseExpression}s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'+'} operator.
    May also be constructed using the C{'-'} operator, which will suppress backtracking.

    Example::
        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"),name_expr("name"),integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    """

    class _ErrorStop(Empty):
        def __init__(self, *args, **kwargs):
            super(And._ErrorStop,self).__init__(*args, **kwargs)
            self.name = '-'
            self.leaveWhitespace()

    def __init__( self, exprs, savelist = True ):
        super(And,self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.setWhitespaceChars( self.exprs[0].whiteChars )
        self.skipWhitespace = self.exprs[0].skipWhitespace
        self.callPreparse = True

    def parseImpl( self, instring, loc, doActions=True ):
        # pass False as last arg to _parse for first element, since we already
        # pre-parsed the string as part of our And pre-parsing
        loc, resultlist = self.exprs[0]._parse( instring, loc, doActions, callPreParse=False )
        errorStop = False
        for e in self.exprs[1:]:
            if isinstance(e, And._ErrorStop):
                errorStop = True
                continue
            if errorStop:
                try:
                    loc, exprtokens = e._parse( instring, loc, doActions )
                except ParseSyntaxException:
                    raise
                except ParseBaseException as pe:
                    pe.__traceback__ = None
                    raise ParseSyntaxException._from_exception(pe)
                except IndexError:
                    raise ParseSyntaxException(instring, len(instring), self.errmsg, self)
            else:
                loc, exprtokens = e._parse( instring, loc, doActions )
            if exprtokens or exprtokens.haskeys():
                resultlist += exprtokens
        return loc, resultlist

    def __iadd__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #And( [ self, other ] )

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )
            if not e.mayReturnEmpty:
                break

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr


class Or(ParseExpression):
    """
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the expression that matches the longest string will be used.
    May be constructed using the C{'^'} operator.

    Example::
        # construct Or using '^' operator
        
        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))
    prints::
        [['123'], ['3.1416'], ['789']]
    """
    def __init__( self, exprs, savelist = False ):
        super(Or,self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        maxExcLoc = -1
        maxException = None
        matches = []
        for e in self.exprs:
            try:
                loc2 = e.tryParse( instring, loc )
            except ParseException as err:
                err.__traceback__ = None
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring,len(instring),e.errmsg,self)
                    maxExcLoc = len(instring)
            else:
                # save match among all matches, to retry longest to shortest
                matches.append((loc2, e))

        if matches:
            matches.sort(key=lambda x: -x[0])
            for _,e in matches:
                try:
                    return e._parse( instring, loc, doActions )
                except ParseException as err:
                    err.__traceback__ = None
                    if err.loc > maxExcLoc:
                        maxException = err
                        maxExcLoc = err.loc

        if maxException is not None:
            maxException.msg = self.errmsg
            raise maxException
        else:
            raise ParseException(instring, loc, "no defined alternatives to match", self)


    def __ixor__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #Or( [ self, other ] )

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class MatchFirst(ParseExpression):
    """
    Requires that at least one C{ParseExpression} is found.
    If two expressions match, the first one listed is the one that will match.
    May be constructed using the C{'|'} operator.

    Example::
        # construct MatchFirst using '|' operator
        
        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    """
    def __init__( self, exprs, savelist = False ):
        super(MatchFirst,self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        maxExcLoc = -1
        maxException = None
        for e in self.exprs:
            try:
                ret = e._parse( instring, loc, doActions )
                return ret
            except ParseException as err:
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring,len(instring),e.errmsg,self)
                    maxExcLoc = len(instring)

        # only got here if no expression matched, raise exception for match that made it the furthest
        else:
            if maxException is not None:
                maxException.msg = self.errmsg
                raise maxException
            else:
                raise ParseException(instring, loc, "no defined alternatives to match", self)

    def __ior__(self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass( other )
        return self.append( other ) #MatchFirst( [ self, other ] )

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class Each(ParseExpression):
    """
    Requires all given C{ParseExpression}s to be found, but in any order.
    Expressions may be separated by whitespace.
    May be constructed using the C{'&'} operator.

    Example::
        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order 
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )
    prints::
        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    """
    def __init__( self, exprs, savelist = True ):
        super(Each,self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.skipWhitespace = True
        self.initExprGroups = True

    def parseImpl( self, instring, loc, doActions=True ):
        if self.initExprGroups:
            self.opt1map = dict((id(e.expr),e) for e in self.exprs if isinstance(e,Optional))
            opt1 = [ e.expr for e in self.exprs if isinstance(e,Optional) ]
            opt2 = [ e for e in self.exprs if e.mayReturnEmpty and not isinstance(e,Optional)]
            self.optionals = opt1 + opt2
            self.multioptionals = [ e.expr for e in self.exprs if isinstance(e,ZeroOrMore) ]
            self.multirequired = [ e.expr for e in self.exprs if isinstance(e,OneOrMore) ]
            self.required = [ e for e in self.exprs if not isinstance(e,(Optional,ZeroOrMore,OneOrMore)) ]
            self.required += self.multirequired
            self.initExprGroups = False
        tmpLoc = loc
        tmpReqd = self.required[:]
        tmpOpt  = self.optionals[:]
        matchOrder = []

        keepMatching = True
        while keepMatching:
            tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired
            failed = []
            for e in tmpExprs:
                try:
                    tmpLoc = e.tryParse( instring, tmpLoc )
                except ParseException:
                    failed.append(e)
                else:
                    matchOrder.append(self.opt1map.get(id(e),e))
                    if e in tmpReqd:
                        tmpReqd.remove(e)
                    elif e in tmpOpt:
                        tmpOpt.remove(e)
            if len(failed) == len(tmpExprs):
                keepMatching = False

        if tmpReqd:
            missing = ", ".join(_ustr(e) for e in tmpReqd)
            raise ParseException(instring,loc,"Missing one or more required elements (%s)" % missing )

        # add any unmatched Optionals, in case they have default values defined
        matchOrder += [e for e in self.exprs if isinstance(e,Optional) and e.expr in tmpOpt]

        resultlist = []
        for e in matchOrder:
            loc,results = e._parse(instring,loc,doActions)
            resultlist.append(results)

        finalResults = sum(resultlist, ParseResults([]))
        return loc, finalResults

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion( self, parseElementList ):
        subRecCheckList = parseElementList[:] + [ self ]
        for e in self.exprs:
            e.checkRecursion( subRecCheckList )


class ParseElementEnhance(ParserElement):
    """
    Abstract subclass of C{ParserElement}, for combining and post-processing parsed tokens.
    """
    def __init__( self, expr, savelist=False ):
        super(ParseElementEnhance,self).__init__(savelist)
        if isinstance( expr, basestring ):
            if issubclass(ParserElement._literalStringClass, Token):
                expr = ParserElement._literalStringClass(expr)
            else:
                expr = ParserElement._literalStringClass(Literal(expr))
        self.expr = expr
        self.strRepr = None
        if expr is not None:
            self.mayIndexError = expr.mayIndexError
            self.mayReturnEmpty = expr.mayReturnEmpty
            self.setWhitespaceChars( expr.whiteChars )
            self.skipWhitespace = expr.skipWhitespace
            self.saveAsList = expr.saveAsList
            self.callPreparse = expr.callPreparse
            self.ignoreExprs.extend(expr.ignoreExprs)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.expr is not None:
            return self.expr._parse( instring, loc, doActions, callPreParse=False )
        else:
            raise ParseException("",loc,self.errmsg,self)

    def leaveWhitespace( self ):
        self.skipWhitespace = False
        self.expr = self.expr.copy()
        if self.expr is not None:
            self.expr.leaveWhitespace()
        return self

    def ignore( self, other ):
        if isinstance( other, Suppress ):
            if other not in self.ignoreExprs:
                super( ParseElementEnhance, self).ignore( other )
                if self.expr is not None:
                    self.expr.ignore( self.ignoreExprs[-1] )
        else:
            super( ParseElementEnhance, self).ignore( other )
            if self.expr is not None:
                self.expr.ignore( self.ignoreExprs[-1] )
        return self

    def streamline( self ):
        super(ParseElementEnhance,self).streamline()
        if self.expr is not None:
            self.expr.streamline()
        return self

    def checkRecursion( self, parseElementList ):
        if self in parseElementList:
            raise RecursiveGrammarException( parseElementList+[self] )
        subRecCheckList = parseElementList[:] + [ self ]
        if self.expr is not None:
            self.expr.checkRecursion( subRecCheckList )

    def validate( self, validateTrace=[] ):
        tmp = validateTrace[:]+[self]
        if self.expr is not None:
            self.expr.validate(tmp)
        self.checkRecursion( [] )

    def __str__( self ):
        try:
            return super(ParseElementEnhance,self).__str__()
        except Exception:
            pass

        if self.strRepr is None and self.expr is not None:
            self.strRepr = "%s:(%s)" % ( self.__class__.__name__, _ustr(self.expr) )
        return self.strRepr


class FollowedBy(ParseElementEnhance):
    """
    Lookahead matching of the given parse expression.  C{FollowedBy}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression matches at the current
    position.  C{FollowedBy} always returns a null token list.

    Example::
        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()
    prints::
        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    """
    def __init__( self, expr ):
        super(FollowedBy,self).__init__(expr)
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        self.expr.tryParse( instring, loc )
        return loc, []


class NotAny(ParseElementEnhance):
    """
    Lookahead to disallow matching with the given parse expression.  C{NotAny}
    does I{not} advance the parsing position within the input string, it only
    verifies that the specified parse expression does I{not} match at the current
    position.  Also, C{NotAny} does I{not} skip over leading whitespace. C{NotAny}
    always returns a null token list.  May be constructed using the '~' operator.

    Example::
        
    """
    def __init__( self, expr ):
        super(NotAny,self).__init__(expr)
        #~ self.leaveWhitespace()
        self.skipWhitespace = False  # do NOT use self.leaveWhitespace(), don't want to propagate to exprs
        self.mayReturnEmpty = True
        self.errmsg = "Found unwanted token, "+_ustr(self.expr)

    def parseImpl( self, instring, loc, doActions=True ):
        if self.expr.canParseNext(instring, loc):
            raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "~{" + _ustr(self.expr) + "}"

        return self.strRepr

class _MultipleMatch(ParseElementEnhance):
    def __init__( self, expr, stopOn=None):
        super(_MultipleMatch, self).__init__(expr)
        self.saveAsList = True
        ender = stopOn
        if isinstance(ender, basestring):
            ender = ParserElement._literalStringClass(ender)
        self.not_ender = ~ender if ender is not None else None

    def parseImpl( self, instring, loc, doActions=True ):
        self_expr_parse = self.expr._parse
        self_skip_ignorables = self._skipIgnorables
        check_ender = self.not_ender is not None
        if check_ender:
            try_not_ender = self.not_ender.tryParse
        
        # must be at least one (but first see if we are the stopOn sentinel;
        # if so, fail)
        if check_ender:
            try_not_ender(instring, loc)
        loc, tokens = self_expr_parse( instring, loc, doActions, callPreParse=False )
        try:
            hasIgnoreExprs = (not not self.ignoreExprs)
            while 1:
                if check_ender:
                    try_not_ender(instring, loc)
                if hasIgnoreExprs:
                    preloc = self_skip_ignorables( instring, loc )
                else:
                    preloc = loc
                loc, tmptokens = self_expr_parse( instring, preloc, doActions )
                if tmptokens or tmptokens.haskeys():
                    tokens += tmptokens
        except (ParseException,IndexError):
            pass

        return loc, tokens
        
class OneOrMore(_MultipleMatch):
    """
    Repetition of one or more of the given expression.
    
    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]
        
        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    """

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + _ustr(self.expr) + "}..."

        return self.strRepr

class ZeroOrMore(_MultipleMatch):
    """
    Optional repetition of zero or more of the given expression.
    
    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default=C{None}) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition 
          expression)          

    Example: similar to L{OneOrMore}
    """
    def __init__( self, expr, stopOn=None):
        super(ZeroOrMore,self).__init__(expr, stopOn=stopOn)
        self.mayReturnEmpty = True
        
    def parseImpl( self, instring, loc, doActions=True ):
        try:
            return super(ZeroOrMore, self).parseImpl(instring, loc, doActions)
        except (ParseException,IndexError):
            return loc, []

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]..."

        return self.strRepr

class _NullToken(object):
    def __bool__(self):
        return False
    __nonzero__ = __bool__
    def __str__(self):
        return ""

_optionalNotMatched = _NullToken()
class Optional(ParseElementEnhance):
    """
    Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::
        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345
            
            # ZIP+4 form
            12101-0001
            
            # invalid ZIP
            98765-
            ''')
    prints::
        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    """
    def __init__( self, expr, default=_optionalNotMatched ):
        super(Optional,self).__init__( expr, savelist=False )
        self.saveAsList = self.expr.saveAsList
        self.defaultValue = default
        self.mayReturnEmpty = True

    def parseImpl( self, instring, loc, doActions=True ):
        try:
            loc, tokens = self.expr._parse( instring, loc, doActions, callPreParse=False )
        except (ParseException,IndexError):
            if self.defaultValue is not _optionalNotMatched:
                if self.expr.resultsName:
                    tokens = ParseResults([ self.defaultValue ])
                    tokens[self.expr.resultsName] = self.defaultValue
                else:
                    tokens = [ self.defaultValue ]
            else:
                tokens = []
        return loc, tokens

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]"

        return self.strRepr

class SkipTo(ParseElementEnhance):
    """
    Token for skipping over all undefined text until the matched expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default=C{False}) if True, the target expression is also parsed 
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default=C{None}) used to define grammars (typically quoted strings and 
          comments) that might contain false matches to the target expression
     - failOn - (default=C{None}) define expressions that are not allowed to be 
          included in the skipped test; if found before the target expression is found, 
          the SkipTo is not a match

    Example::
        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP 
                      + string_data("sev") + SEP 
                      + string_data("desc") + SEP 
                      + integer("days_open"))
        
        for tkt in ticket_expr.searchString(report):
            print tkt.dump()
    prints::
        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    """
    def __init__( self, other, include=False, ignore=None, failOn=None ):
        super( SkipTo, self ).__init__( other )
        self.ignoreExpr = ignore
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.includeMatch = include
        self.asList = False
        if isinstance(failOn, basestring):
            self.failOn = ParserElement._literalStringClass(failOn)
        else:
            self.failOn = failOn
        self.errmsg = "No match found for "+_ustr(self.expr)

    def parseImpl( self, instring, loc, doActions=True ):
        startloc = loc
        instrlen = len(instring)
        expr = self.expr
        expr_parse = self.expr._parse
        self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None
        self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None
        
        tmploc = loc
        while tmploc <= instrlen:
            if self_failOn_canParseNext is not None:
                # break if failOn expression matches
                if self_failOn_canParseNext(instring, tmploc):
                    break
                    
            if self_ignoreExpr_tryParse is not None:
                # advance past ignore expressions
                while 1:
                    try:
                        tmploc = self_ignoreExpr_tryParse(instring, tmploc)
                    except ParseBaseException:
                        break
            
            try:
                expr_parse(instring, tmploc, doActions=False, callPreParse=False)
            except (ParseException, IndexError):
                # no match, advance loc in string
                tmploc += 1
            else:
                # matched skipto expr, done
                break

        else:
            # ran off the end of the input string without matching skipto expr, fail
            raise ParseException(instring, loc, self.errmsg, self)

        # build up return values
        loc = tmploc
        skiptext = instring[startloc:loc]
        skipresult = ParseResults(skiptext)
        
        if self.includeMatch:
            loc, mat = expr_parse(instring,loc,doActions,callPreParse=False)
            skipresult += mat

        return loc, skipresult

class Forward(ParseElementEnhance):
    """
    Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the C{Forward} variable using the '<<' operator.

    Note: take care when assigning to C{Forward} not to overlook precedence of operators.
    Specifically, '|' has a lower precedence than '<<', so that::
        fwdExpr << a | b | c
    will actually be evaluated as::
        (fwdExpr << a) | b | c
    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the C{Forward}::
        fwdExpr << (a | b | c)
    Converting to use the '<<=' operator instead will avoid this problem.

    See L{ParseResults.pprint} for an example of a recursive parser created using
    C{Forward}.
    """
    def __init__( self, other=None ):
        super(Forward,self).__init__( other, savelist=False )

    def __lshift__( self, other ):
        if isinstance( other, basestring ):
            other = ParserElement._literalStringClass(other)
        self.expr = other
        self.strRepr = None
        self.mayIndexError = self.expr.mayIndexError
        self.mayReturnEmpty = self.expr.mayReturnEmpty
        self.setWhitespaceChars( self.expr.whiteChars )
        self.skipWhitespace = self.expr.skipWhitespace
        self.saveAsList = self.expr.saveAsList
        self.ignoreExprs.extend(self.expr.ignoreExprs)
        return self
        
    def __ilshift__(self, other):
        return self << other
    
    def leaveWhitespace( self ):
        self.skipWhitespace = False
        return self

    def streamline( self ):
        if not self.streamlined:
            self.streamlined = True
            if self.expr is not None:
                self.expr.streamline()
        return self

    def validate( self, validateTrace=[] ):
        if self not in validateTrace:
            tmp = validateTrace[:]+[self]
            if self.expr is not None:
                self.expr.validate(tmp)
        self.checkRecursion([])

    def __str__( self ):
        if hasattr(self,"name"):
            return self.name
        return self.__class__.__name__ + ": ..."

        # stubbed out for now - creates awful memory and perf issues
        self._revertClass = self.__class__
        self.__class__ = _ForwardNoRecurse
        try:
            if self.expr is not None:
                retString = _ustr(self.expr)
            else:
                retString = "None"
        finally:
            self.__class__ = self._revertClass
        return self.__class__.__name__ + ": " + retString

    def copy(self):
        if self.expr is not None:
            return super(Forward,self).copy()
        else:
            ret = Forward()
            ret <<= self
            return ret

class _ForwardNoRecurse(Forward):
    def __str__( self ):
        return "..."

class TokenConverter(ParseElementEnhance):
    """
    Abstract subclass of C{ParseExpression}, for converting parsed results.
    """
    def __init__( self, expr, savelist=False ):
        super(TokenConverter,self).__init__( expr )#, savelist )
        self.saveAsList = False

class Combine(TokenConverter):
    """
    Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the input string;
    this can be disabled by specifying C{'adjacent=False'} in the constructor.

    Example::
        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    """
    def __init__( self, expr, joinString="", adjacent=True ):
        super(Combine,self).__init__( expr )
        # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself
        if adjacent:
            self.leaveWhitespace()
        self.adjacent = adjacent
        self.skipWhitespace = True
        self.joinString = joinString
        self.callPreparse = True

    def ignore( self, other ):
        if self.adjacent:
            ParserElement.ignore(self, other)
        else:
            super( Combine, self).ignore( other )
        return self

    def postParse( self, instring, loc, tokenlist ):
        retToks = tokenlist.copy()
        del retToks[:]
        retToks += ParseResults([ "".join(tokenlist._asStringList(self.joinString)) ], modal=self.modalResults)

        if self.resultsName and retToks.haskeys():
            return [ retToks ]
        else:
            return retToks

class Group(TokenConverter):
    """
    Converter to return the matched tokens as a list - useful for returning tokens of C{L{ZeroOrMore}} and C{L{OneOrMore}} expressions.

    Example::
        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a,b,100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a,b,100"))  # -> ['fn', ['a', 'b', '100']]
    """
    def __init__( self, expr ):
        super(Group,self).__init__( expr )
        self.saveAsList = True

    def postParse( self, instring, loc, tokenlist ):
        return [ tokenlist ]

class Dict(TokenConverter):
    """
    Converter to return a repetitive expression as a list, but also as a dictionary.
    Each element can also be referenced using the first token in the expression as its key.
    Useful for tabular report scraping when the first column can be used as a item key.

    Example::
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        
        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())
        
        # access named fields as dict entries, or output as dict
        print(result['shape'])        
        print(result.asDict())
    prints::
        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}
    See more examples at L{ParseResults} of accessing fields by results name.
    """
    def __init__( self, expr ):
        super(Dict,self).__init__( expr )
        self.saveAsList = True

    def postParse( self, instring, loc, tokenlist ):
        for i,tok in enumerate(tokenlist):
            if len(tok) == 0:
                continue
            ikey = tok[0]
            if isinstance(ikey,int):
                ikey = _ustr(tok[0]).strip()
            if len(tok)==1:
                tokenlist[ikey] = _ParseResultsWithOffset("",i)
            elif len(tok)==2 and not isinstance(tok[1],ParseResults):
                tokenlist[ikey] = _ParseResultsWithOffset(tok[1],i)
            else:
                dictvalue = tok.copy() #ParseResults(i)
                del dictvalue[0]
                if len(dictvalue)!= 1 or (isinstance(dictvalue,ParseResults) and dictvalue.haskeys()):
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue,i)
                else:
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0],i)

        if self.resultsName:
            return [ tokenlist ]
        else:
            return tokenlist


class Suppress(TokenConverter):
    """
    Converter for ignoring the results of a parsed expression.

    Example::
        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))
    prints::
        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']
    (See also L{delimitedList}.)
    """
    def postParse( self, instring, loc, tokenlist ):
        return []

    def suppress( self ):
        return self


class OnlyOnce(object):
    """
    Wrapper for parse actions, to ensure they are only called once.
    """
    def __init__(self, methodCall):
        self.callable = _trim_arity(methodCall)
        self.called = False
    def __call__(self,s,l,t):
        if not self.called:
            results = self.callable(s,l,t)
            self.called = True
            return results
        raise ParseException(s,l,"")
    def reset(self):
        self.called = False

def traceParseAction(f):
    """
    Decorator for debugging parse actions. 
    
    When the parse action is called, this decorator will print C{">> entering I{method-name}(line:I{current_source_line}, I{parse_location}, I{matched_tokens})".}
    When the parse action completes, the decorator will print C{"<<"} followed by the returned value, or any exception that the parse action raised.

    Example::
        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))
    prints::
        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    """
    f = _trim_arity(f)
    def z(*paArgs):
        thisFunc = f.__name__
        s,l,t = paArgs[-3:]
        if len(paArgs)>3:
            thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc
        sys.stderr.write( ">>entering %s(line: '%s', %d, %r)\n" % (thisFunc,line(l,s),l,t) )
        try:
            ret = f(*paArgs)
        except Exception as exc:
            sys.stderr.write( "<<leaving %s (exception: %s)\n" % (thisFunc,exc) )
            raise
        sys.stderr.write( "<<leaving %s (ret: %r)\n" % (thisFunc,ret) )
        return ret
    try:
        z.__name__ = f.__name__
    except AttributeError:
        pass
    return z

#
# global helpers
#
def delimitedList( expr, delim=",", combine=False ):
    """
    Helper to define a delimited list of expressions - the delimiter defaults to ','.
    By default, the list elements and delimiters can have intervening whitespace, and
    comments, but this can be overridden by passing C{combine=True} in the constructor.
    If C{combine} is set to C{True}, the matching tokens are returned as a single token
    string, with the delimiters included; otherwise, the matching tokens are returned
    as a list of tokens, with the delimiters suppressed.

    Example::
        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    """
    dlName = _ustr(expr)+" ["+_ustr(delim)+" "+_ustr(expr)+"]..."
    if combine:
        return Combine( expr + ZeroOrMore( delim + expr ) ).setName(dlName)
    else:
        return ( expr + ZeroOrMore( Suppress( delim ) + expr ) ).setName(dlName)

def countedArray( expr, intExpr=None ):
    """
    Helper to define a counted list of expressions.
    This helper defines a pattern of the form::
        integer expr expr expr...
    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the leading count token is suppressed.
    
    If C{intExpr} is specified, it should be a pyparsing expression that produces an integer value.

    Example::
        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    """
    arrayExpr = Forward()
    def countFieldParseAction(s,l,t):
        n = t[0]
        arrayExpr << (n and Group(And([expr]*n)) or Group(empty))
        return []
    if intExpr is None:
        intExpr = Word(nums).setParseAction(lambda t:int(t[0]))
    else:
        intExpr = intExpr.copy()
    intExpr.setName("arrayLen")
    intExpr.addParseAction(countFieldParseAction, callDuringTry=True)
    return ( intExpr + arrayExpr ).setName('(len) ' + _ustr(expr) + '...')

def _flatten(L):
    ret = []
    for i in L:
        if isinstance(i,list):
            ret.extend(_flatten(i))
        else:
            ret.append(i)
    return ret

def matchPreviousLiteral(expr):
    """
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches a
    previous literal, will also match the leading C{"1:1"} in C{"1:10"}.
    If this is not desired, use C{matchPreviousExpr}.
    Do I{not} use with packrat parsing enabled.
    """
    rep = Forward()
    def copyTokenToRepeater(s,l,t):
        if t:
            if len(t) == 1:
                rep << t[0]
            else:
                # flatten t tokens
                tflat = _flatten(t.asList())
                rep << And(Literal(tt) for tt in tflat)
        else:
            rep << Empty()
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def matchPreviousExpr(expr):
    """
    Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks
    for a 'repeat' of a previous expression.  For example::
        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second
    will match C{"1:1"}, but not C{"1:2"}.  Because this matches by
    expressions, will I{not} match the leading C{"1:1"} in C{"1:10"};
    the expressions are evaluated first, and then compared, so
    C{"1"} is compared with C{"10"}.
    Do I{not} use with packrat parsing enabled.
    """
    rep = Forward()
    e2 = expr.copy()
    rep <<= e2
    def copyTokenToRepeater(s,l,t):
        matchTokens = _flatten(t.asList())
        def mustMatchTheseTokens(s,l,t):
            theseTokens = _flatten(t.asList())
            if  theseTokens != matchTokens:
                raise ParseException("",0,"")
        rep.setParseAction( mustMatchTheseTokens, callDuringTry=True )
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def _escapeRegexRangeChars(s):
    #~  escape these chars: ^-]
    for c in r"\^-]":
        s = s.replace(c,_bslash+c)
    s = s.replace("\n",r"\n")
    s = s.replace("\t",r"\t")
    return _ustr(s)

def oneOf( strs, caseless=False, useRegex=True ):
    """
    Helper to quickly define a set of alternative Literals, and makes sure to do
    longest-first testing when there is a conflict, regardless of the input order,
    but returns a C{L{MatchFirst}} for best performance.

    Parameters:
     - strs - a string of space-delimited literals, or a collection of string literals
     - caseless - (default=C{False}) - treat all literals as caseless
     - useRegex - (default=C{True}) - as an optimization, will generate a Regex
          object; otherwise, will generate a C{MatchFirst} object (if C{caseless=True}, or
          if creating a C{Regex} raises an exception)

    Example::
        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))
    prints::
        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    """
    if caseless:
        isequal = ( lambda a,b: a.upper() == b.upper() )
        masks = ( lambda a,b: b.upper().startswith(a.upper()) )
        parseElementClass = CaselessLiteral
    else:
        isequal = ( lambda a,b: a == b )
        masks = ( lambda a,b: b.startswith(a) )
        parseElementClass = Literal

    symbols = []
    if isinstance(strs,basestring):
        symbols = strs.split()
    elif isinstance(strs, Iterable):
        symbols = list(strs)
    else:
        warnings.warn("Invalid argument to oneOf, expected string or iterable",
                SyntaxWarning, stacklevel=2)
    if not symbols:
        return NoMatch()

    i = 0
    while i < len(symbols)-1:
        cur = symbols[i]
        for j,other in enumerate(symbols[i+1:]):
            if ( isequal(other, cur) ):
                del symbols[i+j+1]
                break
            elif ( masks(cur, other) ):
                del symbols[i+j+1]
                symbols.insert(i,other)
                cur = other
                break
        else:
            i += 1

    if not caseless and useRegex:
        #~ print (strs,"->", "|".join( [ _escapeRegexChars(sym) for sym in symbols] ))
        try:
            if len(symbols)==len("".join(symbols)):
                return Regex( "[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols) ).setName(' | '.join(symbols))
            else:
                return Regex( "|".join(re.escape(sym) for sym in symbols) ).setName(' | '.join(symbols))
        except Exception:
            warnings.warn("Exception creating Regex for oneOf, building MatchFirst",
                    SyntaxWarning, stacklevel=2)


    # last resort, just use MatchFirst
    return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols))

def dictOf( key, value ):
    """
    Helper to easily and clearly define a dictionary by specifying the respective patterns
    for the key and value.  Takes care of defining the C{L{Dict}}, C{L{ZeroOrMore}}, and C{L{Group}} tokens
    in the proper order.  The key pattern can include delimiting markers or punctuation,
    as long as they are suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the C{Dict} results can include named token
    fields.

    Example::
        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())
        
        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())
    prints::
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    """
    return Dict( ZeroOrMore( Group ( key + value ) ) )

def originalTextFor(expr, asString=True):
    """
    Helper to return the original, untokenized text for a given expression.  Useful to
    restore the parsed fields of an HTML start tag into the raw tag text itself, or to
    revert separate tokens with intervening whitespace back to the original matching
    input text. By default, returns astring containing the original parsed text.  
       
    If the optional C{asString} argument is passed as C{False}, then the return value is a 
    C{L{ParseResults}} containing any results names that were originally matched, and a 
    single token containing the original matched text from the input string.  So if 
    the expression passed to C{L{originalTextFor}} contains expressions with defined
    results names, you must set C{asString} to C{False} if you want to preserve those
    results name values.

    Example::
        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b","i"):
            opener,closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])
    prints::
        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    """
    locMarker = Empty().setParseAction(lambda s,loc,t: loc)
    endlocMarker = locMarker.copy()
    endlocMarker.callPreparse = False
    matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end")
    if asString:
        extractText = lambda s,l,t: s[t._original_start:t._original_end]
    else:
        def extractText(s,l,t):
            t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]]
    matchExpr.setParseAction(extractText)
    matchExpr.ignoreExprs = expr.ignoreExprs
    return matchExpr

def ungroup(expr): 
    """
    Helper to undo pyparsing's default grouping of And expressions, even
    if all but one are non-empty.
    """
    return TokenConverter(expr).setParseAction(lambda t:t[0])

def locatedExpr(expr):
    """
    Helper to decorate a returned token with its starting and ending locations in the input string.
    This helper adds the following results names:
     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains C{<TAB>} characters, you may want to call
    C{L{ParserElement.parseWithTabs}}

    Example::
        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)
    prints::
        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    """
    locator = Empty().setParseAction(lambda s,l,t: l)
    return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end"))


# convenience constants for positional expressions
empty       = Empty().setName("empty")
lineStart   = LineStart().setName("lineStart")
lineEnd     = LineEnd().setName("lineEnd")
stringStart = StringStart().setName("stringStart")
stringEnd   = StringEnd().setName("stringEnd")

_escapedPunc = Word( _bslash, r"\[]-*.$+^?()~ ", exact=2 ).setParseAction(lambda s,l,t:t[0][1])
_escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s,l,t:unichr(int(t[0].lstrip(r'\0x'),16)))
_escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s,l,t:unichr(int(t[0][1:],8)))
_singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | CharsNotIn(r'\]', exact=1)
_charRange = Group(_singleChar + Suppress("-") + _singleChar)
_reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group( OneOrMore( _charRange | _singleChar ) ).setResultsName("body") + "]"

def srange(s):
    r"""
    Helper to easily define string ranges for use in Word construction.  Borrows
    syntax from regexp '[]' string range definitions::
        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"
    The input string must be enclosed in []'s, and the returned string is the expanded
    character set joined into a single string.
    The values enclosed in the []'s may be:
     - a single character
     - an escaped character with a leading backslash (such as C{\-} or C{\]})
     - an escaped hex character with a leading C{'\x'} (C{\x21}, which is a C{'!'} character) 
         (C{\0x##} is also supported for backwards compatibility) 
     - an escaped octal character with a leading C{'\0'} (C{\041}, which is a C{'!'} character)
     - a range of any of the above, separated by a dash (C{'a-z'}, etc.)
     - any combination of the above (C{'aeiouy'}, C{'a-zA-Z0-9_$'}, etc.)
    """
    _expanded = lambda p: p if not isinstance(p,ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]),ord(p[1])+1))
    try:
        return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body)
    except Exception:
        return ""

def matchOnlyAtCol(n):
    """
    Helper method for defining parse actions that require matching at a specific
    column in the input text.
    """
    def verifyCol(strg,locn,toks):
        if col(locn,strg) != n:
            raise ParseException(strg,locn,"matched token not at column %d" % n)
    return verifyCol

def replaceWith(replStr):
    """
    Helper method for common parse actions that simply return a literal value.  Especially
    useful when used with C{L{transformString<ParserElement.transformString>}()}.

    Example::
        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num
        
        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    """
    return lambda s,l,t: [replStr]

def removeQuotes(s,l,t):
    """
    Helper parse action for removing quotation marks from parsed quoted strings.

    Example::
        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    """
    return t[0][1:-1]

def tokenMap(func, *args):
    """
    Helper to define a parse action by mapping a function to all elements of a ParseResults list.If any additional 
    args are passed, they are forwarded to the given function as additional arguments after
    the token, as in C{hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))}, which will convert the
    parsed data to an integer using base 16.

    Example (compare the last to example in L{ParserElement.transformString}::
        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')
        
        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')
    prints::
        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    """
    def pa(s,l,t):
        return [func(tokn, *args) for tokn in t]

    try:
        func_name = getattr(func, '__name__', 
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    pa.__name__ = func_name

    return pa

upcaseTokens = tokenMap(lambda t: _ustr(t).upper())
"""(Deprecated) Helper parse action to convert tokens to upper case. Deprecated in favor of L{pyparsing_common.upcaseTokens}"""

downcaseTokens = tokenMap(lambda t: _ustr(t).lower())
"""(Deprecated) Helper parse action to convert tokens to lower case. Deprecated in favor of L{pyparsing_common.downcaseTokens}"""
    
def _makeTags(tagStr, xml):
    """Internal helper to construct opening and closing tag expressions, given a tag name"""
    if isinstance(tagStr,basestring):
        resname = tagStr
        tagStr = Keyword(tagStr, caseless=not xml)
    else:
        resname = tagStr.name

    tagAttrName = Word(alphas,alphanums+"_-:")
    if (xml):
        tagAttrValue = dblQuotedString.copy().setParseAction( removeQuotes )
        openTag = Suppress("<") + tagStr("tag") + \
                Dict(ZeroOrMore(Group( tagAttrName + Suppress("=") + tagAttrValue ))) + \
                Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
    else:
        printablesLessRAbrack = "".join(c for c in printables if c not in ">")
        tagAttrValue = quotedString.copy().setParseAction( removeQuotes ) | Word(printablesLessRAbrack)
        openTag = Suppress("<") + tagStr("tag") + \
                Dict(ZeroOrMore(Group( tagAttrName.setParseAction(downcaseTokens) + \
                Optional( Suppress("=") + tagAttrValue ) ))) + \
                Optional("/",default=[False]).setResultsName("empty").setParseAction(lambda s,l,t:t[0]=='/') + Suppress(">")
    closeTag = Combine(_L("</") + tagStr + ">")

    openTag = openTag.setResultsName("start"+"".join(resname.replace(":"," ").title().split())).setName("<%s>" % resname)
    closeTag = closeTag.setResultsName("end"+"".join(resname.replace(":"," ").title().split())).setName("</%s>" % resname)
    openTag.tag = resname
    closeTag.tag = resname
    return openTag, closeTag

def makeHTMLTags(tagStr):
    """
    Helper to construct opening and closing tag expressions for HTML, given a tag name. Matches
    tags in either upper or lower case, attributes with namespaces and with quoted or unquoted values.

    Example::
        text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and closing tags as a 2-tuple
        a,a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end
        
        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are also accessible as named results
            print(link.link_text, '->', link.href)
    prints::
        pyparsing -> http://pyparsing.wikispaces.com
    """
    return _makeTags( tagStr, False )

def makeXMLTags(tagStr):
    """
    Helper to construct opening and closing tag expressions for XML, given a tag name. Matches
    tags only in the given upper/lower case.

    Example: similar to L{makeHTMLTags}
    """
    return _makeTags( tagStr, True )

def withAttribute(*args,**attrDict):
    """
    Helper to create a validating parse action to be used with start tags created
    with C{L{makeXMLTags}} or C{L{makeHTMLTags}}. Use C{withAttribute} to qualify a starting tag
    with a required attribute value, to avoid false matches on common tags such as
    C{<TD>} or C{<DIV>}.

    Call C{withAttribute} with a series of attribute names and values. Specify the list
    of filter attributes names and values as:
     - keyword arguments, as in C{(align="right")}, or
     - as an explicit dict with C{**} operator, when an attribute name is also a Python
          reserved word, as in C{**{"class":"Customer", "align":"right"}}
     - a list of name-value tuples, as in ( ("ns1:class", "Customer"), ("ns2:align","right") )
    For attribute names with a namespace prefix, you must use the second form.  Attribute
    names are matched insensitive to upper/lower case.
       
    If just testing for C{class} (with or without a namespace), use C{L{withClass}}.

    To verify that the attribute exists, but without specifying a value, pass
    C{withAttribute.ANY_VALUE} as the value.

    Example::
        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    if args:
        attrs = args[:]
    else:
        attrs = attrDict.items()
    attrs = [(k,v) for k,v in attrs]
    def pa(s,l,tokens):
        for attrName,attrValue in attrs:
            if attrName not in tokens:
                raise ParseException(s,l,"no matching attribute " + attrName)
            if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue:
                raise ParseException(s,l,"attribute '%s' has value '%s', must be '%s'" %
                                            (attrName, tokens[attrName], attrValue))
    return pa
withAttribute.ANY_VALUE = object()

def withClass(classname, namespace=''):
    """
    Simplified version of C{L{withAttribute}} when matching on a div class - made
    difficult because C{class} is a reserved word in Python.

    Example::
        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>
                
        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))
        
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)
        
        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)
    prints::
        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    classattr = "%s:class" % namespace if namespace else "class"
    return withAttribute(**{classattr : classname})        

opAssoc = _Constants()
opAssoc.LEFT = object()
opAssoc.RIGHT = object()

def infixNotation( baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')') ):
    """
    Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary or
    binary, left- or right-associative.  Parse actions can also be attached
    to operator expressions. The generated parser will also recognize the use 
    of parentheses to override operator precedences (see example below).
    
    Note: if you define a deep operator list, you may see performance issues
    when using infixNotation. See L{ParserElement.enablePackrat} for a
    mechanism to potentially improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the nested
     - opList - list of tuples, one for each operator precedence level in the
      expression grammar; each tuple is of the form
      (opExpr, numTerms, rightLeftAssoc, parseAction), where:
       - opExpr is the pyparsing expression for the operator;
          may also be a string, which will be converted to a Literal;
          if numTerms is 3, opExpr is a tuple of two expressions, for the
          two operators separating the 3 terms
       - numTerms is the number of terms for this operator (must
          be 1, 2, or 3)
       - rightLeftAssoc is the indicator whether the operator is
          right or left associative, using the pyparsing-defined
          constants C{opAssoc.RIGHT} and C{opAssoc.LEFT}.
       - parseAction is the parse action to be associated with
          expressions matching this operator expression (the
          parse action tuple member may be omitted); if the parse action
          is passed a tuple or list of functions, this is equivalent to
          calling C{setParseAction(*fn)} (L{ParserElement.setParseAction})
     - lpar - expression for matching left-parentheses (default=C{Suppress('(')})
     - rpar - expression for matching right-parentheses (default=C{Suppress(')')})

    Example::
        # simple example of four-function arithmetic with ints and variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier 
        
        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])
        
        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)
    prints::
        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    """
    ret = Forward()
    lastExpr = baseExpr | ( lpar + ret + rpar )
    for i,operDef in enumerate(opList):
        opExpr,arity,rightLeftAssoc,pa = (operDef + (None,))[:4]
        termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr
        if arity == 3:
            if opExpr is None or len(opExpr) != 2:
                raise ValueError("if numterms=3, opExpr must be a tuple or list of two expressions")
            opExpr1, opExpr2 = opExpr
        thisExpr = Forward().setName(termName)
        if rightLeftAssoc == opAssoc.LEFT:
            if arity == 1:
                matchExpr = FollowedBy(lastExpr + opExpr) + Group( lastExpr + OneOrMore( opExpr ) )
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = FollowedBy(lastExpr + opExpr + lastExpr) + Group( lastExpr + OneOrMore( opExpr + lastExpr ) )
                else:
                    matchExpr = FollowedBy(lastExpr+lastExpr) + Group( lastExpr + OneOrMore(lastExpr) )
            elif arity == 3:
                matchExpr = FollowedBy(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr) + \
                            Group( lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr )
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        elif rightLeftAssoc == opAssoc.RIGHT:
            if arity == 1:
                # try to avoid LR with this extra test
                if not isinstance(opExpr, Optional):
                    opExpr = Optional(opExpr)
                matchExpr = FollowedBy(opExpr.expr + thisExpr) + Group( opExpr + thisExpr )
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = FollowedBy(lastExpr + opExpr + thisExpr) + Group( lastExpr + OneOrMore( opExpr + thisExpr ) )
                else:
                    matchExpr = FollowedBy(lastExpr + thisExpr) + Group( lastExpr + OneOrMore( thisExpr ) )
            elif arity == 3:
                matchExpr = FollowedBy(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr) + \
                            Group( lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr )
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        else:
            raise ValueError("operator must indicate right or left associativity")
        if pa:
            if isinstance(pa, (tuple, list)):
                matchExpr.setParseAction(*pa)
            else:
                matchExpr.setParseAction(pa)
        thisExpr <<= ( matchExpr.setName(termName) | lastExpr )
        lastExpr = thisExpr
    ret <<= lastExpr
    return ret

operatorPrecedence = infixNotation
"""(Deprecated) Former name of C{L{infixNotation}}, will be dropped in a future release."""

dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"').setName("string enclosed in double quotes")
sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("string enclosed in single quotes")
quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*')+'"'|
                       Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*")+"'").setName("quotedString using single or double quotes")
unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal")

def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()):
    """
    Helper method for defining nested lists enclosed in opening and closing
    delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list (default=C{"("}); can also be a pyparsing expression
     - closer - closing character for a nested list (default=C{")"}); can also be a pyparsing expression
     - content - expression for items within the nested lists (default=C{None})
     - ignoreExpr - expression for ignoring opening and closing delimiters (default=C{quotedString})

    If an expression is not provided for the content argument, the nested
    expression will capture all whitespace-delimited content between delimiters
    as a list of separate values.

    Use the C{ignoreExpr} argument to define expressions that may contain
    opening or closing characters that should not be treated as opening
    or closing characters for nesting, such as quotedString or a comment
    expression.  Specify multiple expressions using an C{L{Or}} or C{L{MatchFirst}}.
    The default is L{quotedString}, but if no expressions are to be ignored,
    then pass C{None} for this argument.

    Example::
        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR,RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type") 
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR 
                      + code_body("body"))
        c_function.ignore(cStyleComment)
        
        source_code = '''
            int is_odd(int x) { 
                return (x%2); 
            }
                
            int dec_to_hex(char hchar) { 
                if (hchar >= '0' && hchar <= '9') { 
                    return (ord(hchar)-ord('0')); 
                } else { 
                    return (10+ord(hchar)-ord('A'));
                } 
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)

    prints::
        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    """
    if opener == closer:
        raise ValueError("opening and closing strings cannot be the same")
    if content is None:
        if isinstance(opener,basestring) and isinstance(closer,basestring):
            if len(opener) == 1 and len(closer)==1:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr +
                                    CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
                else:
                    content = (empty.copy()+CharsNotIn(opener+closer+ParserElement.DEFAULT_WHITE_CHARS
                                ).setParseAction(lambda t:t[0].strip()))
            else:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr + 
                                    ~Literal(opener) + ~Literal(closer) +
                                    CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
                else:
                    content = (Combine(OneOrMore(~Literal(opener) + ~Literal(closer) +
                                    CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS,exact=1))
                                ).setParseAction(lambda t:t[0].strip()))
        else:
            raise ValueError("opening and closing arguments must be strings if no content expression is given")
    ret = Forward()
    if ignoreExpr is not None:
        ret <<= Group( Suppress(opener) + ZeroOrMore( ignoreExpr | ret | content ) + Suppress(closer) )
    else:
        ret <<= Group( Suppress(opener) + ZeroOrMore( ret | content )  + Suppress(closer) )
    ret.setName('nested %s%s expression' % (opener,closer))
    return ret

def indentedBlock(blockStatementExpr, indentStack, indent=True):
    """
    Helper method for defining space-delimited indentation blocks, such as
    those used to define block statements in Python source code.

    Parameters:
     - blockStatementExpr - expression defining syntax of statement that
            is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
            (multiple statementWithIndentedBlock expressions within a single grammar
            should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond the
            the current level; set to False for block of left-most statements
            (default=C{True})

    A valid block must contain at least one C{blockStatement}.

    Example::
        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group( "(" + Optional( delimitedList(identifier) ) + ")" ) + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group( funcDecl + func_body )

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << ( funcDef | assignment | identifier )

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()
    prints::
        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]] 
    """
    def checkPeerIndent(s,l,t):
        if l >= len(s): return
        curCol = col(l,s)
        if curCol != indentStack[-1]:
            if curCol > indentStack[-1]:
                raise ParseFatalException(s,l,"illegal nesting")
            raise ParseException(s,l,"not a peer entry")

    def checkSubIndent(s,l,t):
        curCol = col(l,s)
        if curCol > indentStack[-1]:
            indentStack.append( curCol )
        else:
            raise ParseException(s,l,"not a subentry")

    def checkUnindent(s,l,t):
        if l >= len(s): return
        curCol = col(l,s)
        if not(indentStack and curCol < indentStack[-1] and curCol <= indentStack[-2]):
            raise ParseException(s,l,"not an unindent")
        indentStack.pop()

    NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress())
    INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT')
    PEER   = Empty().setParseAction(checkPeerIndent).setName('')
    UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT')
    if indent:
        smExpr = Group( Optional(NL) +
            #~ FollowedBy(blockStatementExpr) +
            INDENT + (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) + UNDENT)
    else:
        smExpr = Group( Optional(NL) +
            (OneOrMore( PEER + Group(blockStatementExpr) + Optional(NL) )) )
    blockStatementExpr.ignore(_bslash + LineEnd())
    return smExpr.setName('indented block')

alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]")
punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]")

anyOpenTag,anyCloseTag = makeHTMLTags(Word(alphas,alphanums+"_:").setName('any tag'))
_htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(),'><& "\''))
commonHTMLEntity = Regex('&(?P<entity>' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity")
def replaceHTMLEntity(t):
    """Helper parser action to replace common HTML entities with their special characters"""
    return _htmlEntityMap.get(t.entity)

# it's easy to get these comment structures wrong - they're very common, so may as well make them available
cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment")
"Comment of the form C{/* ... */}"

htmlComment = Regex(r"<!--[\s\S]*?-->").setName("HTML comment")
"Comment of the form C{<!-- ... -->}"

restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line")
dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment")
"Comment of the form C{// ... (to end of line)}"

cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/'| dblSlashComment).setName("C++ style comment")
"Comment of either form C{L{cStyleComment}} or C{L{dblSlashComment}}"

javaStyleComment = cppStyleComment
"Same as C{L{cppStyleComment}}"

pythonStyleComment = Regex(r"#.*").setName("Python style comment")
"Comment of the form C{# ... (to end of line)}"

_commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',') +
                                  Optional( Word(" \t") +
                                            ~Literal(",") + ~LineEnd() ) ) ).streamline().setName("commaItem")
commaSeparatedList = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("commaSeparatedList")
"""(Deprecated) Predefined expression of 1 or more printable words or quoted strings, separated by commas.
   This expression is deprecated in favor of L{pyparsing_common.comma_separated_list}."""

# some other useful expressions - using lower-case class name since we are really using this as a namespace
class pyparsing_common:
    """
    Here are some common low-level expressions that may be useful in jump-starting parser development:
     - numeric forms (L{integers<integer>}, L{reals<real>}, L{scientific notation<sci_real>})
     - common L{programming identifiers<identifier>}
     - network addresses (L{MAC<mac_address>}, L{IPv4<ipv4_address>}, L{IPv6<ipv6_address>})
     - ISO8601 L{dates<iso8601_date>} and L{datetime<iso8601_datetime>}
     - L{UUID<uuid>}
     - L{comma-separated list<comma_separated_list>}
    Parse actions:
     - C{L{convertToInteger}}
     - C{L{convertToFloat}}
     - C{L{convertToDate}}
     - C{L{convertToDatetime}}
     - C{L{stripHTMLTags}}
     - C{L{upcaseTokens}}
     - C{L{downcaseTokens}}

    Example::
        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')
    prints::
        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    """

    convertToInteger = tokenMap(int)
    """
    Parse action for converting parsed integers to Python int
    """

    convertToFloat = tokenMap(float)
    """
    Parse action for converting parsed numbers to Python float
    """

    integer = Word(nums).setName("integer").setParseAction(convertToInteger)
    """expression that parses an unsigned integer, returns an int"""

    hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int,16))
    """expression that parses a hexadecimal integer, returns an int"""

    signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger)
    """expression that parses an integer with optional leading sign, returns an int"""

    fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction")
    """fractional expression of an integer divided by an integer, returns a float"""
    fraction.addParseAction(lambda t: t[0]/t[-1])

    mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction")
    """mixed integer of the form 'integer - fraction', with optional leading integer, returns float"""
    mixed_integer.addParseAction(sum)

    real = Regex(r'[+-]?\d+\.\d*').setName("real number").setParseAction(convertToFloat)
    """expression that parses a floating point number and returns a float"""

    sci_real = Regex(r'[+-]?\d+([eE][+-]?\d+|\.\d*([eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat)
    """expression that parses a floating point number with optional scientific notation and returns a float"""

    # streamlining this expression makes the docs nicer-looking
    number = (sci_real | real | signed_integer).streamline()
    """any numeric expression, returns the corresponding Python type"""

    fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat)
    """any int or real number, returned as float"""
    
    identifier = Word(alphas+'_', alphanums+'_').setName("identifier")
    """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')"""
    
    ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address")
    "IPv4 address (C{0.0.0.0 - 255.255.255.255})"

    _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer")
    _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part)*7).setName("full IPv6 address")
    _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part)*(0,6)) + "::" + Optional(_ipv6_part + (':' + _ipv6_part)*(0,6))).setName("short IPv6 address")
    _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8)
    _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address")
    ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address")
    "IPv6 address (long, short, or mixed form)"
    
    mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address")
    "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)"

    @staticmethod
    def convertToDate(fmt="%Y-%m-%d"):
        """
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%d"})

        Example::
            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))
        prints::
            [datetime.date(1999, 12, 31)]
        """
        def cvt_fn(s,l,t):
            try:
                return datetime.strptime(t[0], fmt).date()
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    @staticmethod
    def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"):
        """
        Helper to create a parse action for converting parsed datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default=C{"%Y-%m-%dT%H:%M:%S.%f"})

        Example::
            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))
        prints::
            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        """
        def cvt_fn(s,l,t):
            try:
                return datetime.strptime(t[0], fmt)
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    iso8601_date = Regex(r'(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?').setName("ISO8601 date")
    "ISO8601 date (C{yyyy-mm-dd})"

    iso8601_datetime = Regex(r'(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime")
    "ISO8601 datetime (C{yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)}) - trailing seconds, milliseconds, and timezone optional; accepts separating C{'T'} or C{' '}"

    uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID")
    "UUID (C{xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx})"

    _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress()
    @staticmethod
    def stripHTMLTags(s, l, tokens):
        """
        Parse action to remove HTML tags from web page HTML source

        Example::
            # strip HTML links from normal text 
            text = '<td>More info at the <a href="http://pyparsing.wikispaces.com">pyparsing</a> wiki page</td>'
            td,td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            
            print(table_text.parseString(text).body) # -> 'More info at the pyparsing wiki page'
        """
        return pyparsing_common._html_stripper.transformString(tokens[0])

    _commasepitem = Combine(OneOrMore(~Literal(",") + ~LineEnd() + Word(printables, excludeChars=',') 
                                        + Optional( White(" \t") ) ) ).streamline().setName("commaItem")
    comma_separated_list = delimitedList( Optional( quotedString.copy() | _commasepitem, default="") ).setName("comma separated list")
    """Predefined expression of 1 or more printable words or quoted strings, separated by commas."""

    upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper()))
    """Parse action to convert tokens to upper case."""

    downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower()))
    """Parse action to convert tokens to lower case."""


if __name__ == "__main__":

    selectToken    = CaselessLiteral("select")
    fromToken      = CaselessLiteral("from")

    ident          = Word(alphas, alphanums + "_$")

    columnName     = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    columnNameList = Group(delimitedList(columnName)).setName("columns")
    columnSpec     = ('*' | columnNameList)

    tableName      = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    tableNameList  = Group(delimitedList(tableName)).setName("tables")
    
    simpleSQL      = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables")

    # demo runTests method, including embedded comments in test string
    simpleSQL.runTests("""
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        """)

    pyparsing_common.number.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    # any int or real number, returned as float
    pyparsing_common.fnumber.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    pyparsing_common.hex_integer.runTests("""
        100
        FF
        """)

    import uuid
    pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
    pyparsing_common.uuid.runTests("""
        12345678-1234-5678-1234-567812345678
        """)
site-packages/pkg_resources/_vendor/__init__.py000064400000000000150564745010015701 0ustar00site-packages/pkg_resources/_vendor/appdirs.py000064400000060175150564745010015627 0ustar00#!/usr/bin/env python
# -*- coding: utf-8 -*-
# Copyright (c) 2005-2010 ActiveState Software Inc.
# Copyright (c) 2013 Eddy Petrișor

"""Utilities for determining application-specific dirs.

See <http://github.com/ActiveState/appdirs> for details and usage.
"""
# Dev Notes:
# - MSDN on where to store app data files:
#   http://support.microsoft.com/default.aspx?scid=kb;en-us;310294#XSLTH3194121123120121120120
# - Mac OS X: http://developer.apple.com/documentation/MacOSX/Conceptual/BPFileSystem/index.html
# - XDG spec for Un*x: http://standards.freedesktop.org/basedir-spec/basedir-spec-latest.html

__version_info__ = (1, 4, 3)
__version__ = '.'.join(map(str, __version_info__))


import sys
import os

PY3 = sys.version_info[0] == 3

if PY3:
    unicode = str

if sys.platform.startswith('java'):
    import platform
    os_name = platform.java_ver()[3][0]
    if os_name.startswith('Windows'): # "Windows XP", "Windows 7", etc.
        system = 'win32'
    elif os_name.startswith('Mac'): # "Mac OS X", etc.
        system = 'darwin'
    else: # "Linux", "SunOS", "FreeBSD", etc.
        # Setting this to "linux2" is not ideal, but only Windows or Mac
        # are actually checked for and the rest of the module expects
        # *sys.platform* style strings.
        system = 'linux2'
else:
    system = sys.platform



def user_data_dir(appname=None, appauthor=None, version=None, roaming=False):
    r"""Return full path to the user-specific data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user data directories are:
        Mac OS X:               ~/Library/Application Support/<AppName>
        Unix:                   ~/.local/share/<AppName>    # or in $XDG_DATA_HOME, if defined
        Win XP (not roaming):   C:\Documents and Settings\<username>\Application Data\<AppAuthor>\<AppName>
        Win XP (roaming):       C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>
        Win 7  (not roaming):   C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>
        Win 7  (roaming):       C:\Users\<username>\AppData\Roaming\<AppAuthor>\<AppName>

    For Unix, we follow the XDG spec and support $XDG_DATA_HOME.
    That means, by default "~/.local/share/<AppName>".
    """
    if system == "win32":
        if appauthor is None:
            appauthor = appname
        const = roaming and "CSIDL_APPDATA" or "CSIDL_LOCAL_APPDATA"
        path = os.path.normpath(_get_win_folder(const))
        if appname:
            if appauthor is not False:
                path = os.path.join(path, appauthor, appname)
            else:
                path = os.path.join(path, appname)
    elif system == 'darwin':
        path = os.path.expanduser('~/Library/Application Support/')
        if appname:
            path = os.path.join(path, appname)
    else:
        path = os.getenv('XDG_DATA_HOME', os.path.expanduser("~/.local/share"))
        if appname:
            path = os.path.join(path, appname)
    if appname and version:
        path = os.path.join(path, version)
    return path


def site_data_dir(appname=None, appauthor=None, version=None, multipath=False):
    r"""Return full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of data dirs should be
            returned. By default, the first item from XDG_DATA_DIRS is
            returned, or '/usr/local/share/<AppName>',
            if XDG_DATA_DIRS is not set

    Typical site data directories are:
        Mac OS X:   /Library/Application Support/<AppName>
        Unix:       /usr/local/share/<AppName> or /usr/share/<AppName>
        Win XP:     C:\Documents and Settings\All Users\Application Data\<AppAuthor>\<AppName>
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)
        Win 7:      C:\ProgramData\<AppAuthor>\<AppName>   # Hidden, but writeable on Win 7.

    For Unix, this is using the $XDG_DATA_DIRS[0] default.

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    """
    if system == "win32":
        if appauthor is None:
            appauthor = appname
        path = os.path.normpath(_get_win_folder("CSIDL_COMMON_APPDATA"))
        if appname:
            if appauthor is not False:
                path = os.path.join(path, appauthor, appname)
            else:
                path = os.path.join(path, appname)
    elif system == 'darwin':
        path = os.path.expanduser('/Library/Application Support')
        if appname:
            path = os.path.join(path, appname)
    else:
        # XDG default for $XDG_DATA_DIRS
        # only first, if multipath is False
        path = os.getenv('XDG_DATA_DIRS',
                         os.pathsep.join(['/usr/local/share', '/usr/share']))
        pathlist = [os.path.expanduser(x.rstrip(os.sep)) for x in path.split(os.pathsep)]
        if appname:
            if version:
                appname = os.path.join(appname, version)
            pathlist = [os.sep.join([x, appname]) for x in pathlist]

        if multipath:
            path = os.pathsep.join(pathlist)
        else:
            path = pathlist[0]
        return path

    if appname and version:
        path = os.path.join(path, version)
    return path


def user_config_dir(appname=None, appauthor=None, version=None, roaming=False):
    r"""Return full path to the user-specific config dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user config directories are:
        Mac OS X:               same as user_data_dir
        Unix:                   ~/.config/<AppName>     # or in $XDG_CONFIG_HOME, if defined
        Win *:                  same as user_data_dir

    For Unix, we follow the XDG spec and support $XDG_CONFIG_HOME.
    That means, by default "~/.config/<AppName>".
    """
    if system in ["win32", "darwin"]:
        path = user_data_dir(appname, appauthor, None, roaming)
    else:
        path = os.getenv('XDG_CONFIG_HOME', os.path.expanduser("~/.config"))
        if appname:
            path = os.path.join(path, appname)
    if appname and version:
        path = os.path.join(path, version)
    return path


def site_config_dir(appname=None, appauthor=None, version=None, multipath=False):
    r"""Return full path to the user-shared data dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "multipath" is an optional parameter only applicable to *nix
            which indicates that the entire list of config dirs should be
            returned. By default, the first item from XDG_CONFIG_DIRS is
            returned, or '/etc/xdg/<AppName>', if XDG_CONFIG_DIRS is not set

    Typical site config directories are:
        Mac OS X:   same as site_data_dir
        Unix:       /etc/xdg/<AppName> or $XDG_CONFIG_DIRS[i]/<AppName> for each value in
                    $XDG_CONFIG_DIRS
        Win *:      same as site_data_dir
        Vista:      (Fail! "C:\ProgramData" is a hidden *system* directory on Vista.)

    For Unix, this is using the $XDG_CONFIG_DIRS[0] default, if multipath=False

    WARNING: Do not use this on Windows. See the Vista-Fail note above for why.
    """
    if system in ["win32", "darwin"]:
        path = site_data_dir(appname, appauthor)
        if appname and version:
            path = os.path.join(path, version)
    else:
        # XDG default for $XDG_CONFIG_DIRS
        # only first, if multipath is False
        path = os.getenv('XDG_CONFIG_DIRS', '/etc/xdg')
        pathlist = [os.path.expanduser(x.rstrip(os.sep)) for x in path.split(os.pathsep)]
        if appname:
            if version:
                appname = os.path.join(appname, version)
            pathlist = [os.sep.join([x, appname]) for x in pathlist]

        if multipath:
            path = os.pathsep.join(pathlist)
        else:
            path = pathlist[0]
    return path


def user_cache_dir(appname=None, appauthor=None, version=None, opinion=True):
    r"""Return full path to the user-specific cache dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Cache" to the base app data dir for Windows. See
            discussion below.

    Typical user cache directories are:
        Mac OS X:   ~/Library/Caches/<AppName>
        Unix:       ~/.cache/<AppName> (XDG default)
        Win XP:     C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>\Cache
        Vista:      C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>\Cache

    On Windows the only suggestion in the MSDN docs is that local settings go in
    the `CSIDL_LOCAL_APPDATA` directory. This is identical to the non-roaming
    app data dir (the default returned by `user_data_dir` above). Apps typically
    put cache data somewhere *under* the given dir here. Some examples:
        ...\Mozilla\Firefox\Profiles\<ProfileName>\Cache
        ...\Acme\SuperApp\Cache\1.0
    OPINION: This function appends "Cache" to the `CSIDL_LOCAL_APPDATA` value.
    This can be disabled with the `opinion=False` option.
    """
    if system == "win32":
        if appauthor is None:
            appauthor = appname
        path = os.path.normpath(_get_win_folder("CSIDL_LOCAL_APPDATA"))
        if appname:
            if appauthor is not False:
                path = os.path.join(path, appauthor, appname)
            else:
                path = os.path.join(path, appname)
            if opinion:
                path = os.path.join(path, "Cache")
    elif system == 'darwin':
        path = os.path.expanduser('~/Library/Caches')
        if appname:
            path = os.path.join(path, appname)
    else:
        path = os.getenv('XDG_CACHE_HOME', os.path.expanduser('~/.cache'))
        if appname:
            path = os.path.join(path, appname)
    if appname and version:
        path = os.path.join(path, version)
    return path


def user_state_dir(appname=None, appauthor=None, version=None, roaming=False):
    r"""Return full path to the user-specific state dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "roaming" (boolean, default False) can be set True to use the Windows
            roaming appdata directory. That means that for users on a Windows
            network setup for roaming profiles, this user data will be
            sync'd on login. See
            <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>
            for a discussion of issues.

    Typical user state directories are:
        Mac OS X:  same as user_data_dir
        Unix:      ~/.local/state/<AppName>   # or in $XDG_STATE_HOME, if defined
        Win *:     same as user_data_dir

    For Unix, we follow this Debian proposal <https://wiki.debian.org/XDGBaseDirectorySpecification#state>
    to extend the XDG spec and support $XDG_STATE_HOME.

    That means, by default "~/.local/state/<AppName>".
    """
    if system in ["win32", "darwin"]:
        path = user_data_dir(appname, appauthor, None, roaming)
    else:
        path = os.getenv('XDG_STATE_HOME', os.path.expanduser("~/.local/state"))
        if appname:
            path = os.path.join(path, appname)
    if appname and version:
        path = os.path.join(path, version)
    return path


def user_log_dir(appname=None, appauthor=None, version=None, opinion=True):
    r"""Return full path to the user-specific log dir for this application.

        "appname" is the name of application.
            If None, just the system directory is returned.
        "appauthor" (only used on Windows) is the name of the
            appauthor or distributing body for this application. Typically
            it is the owning company name. This falls back to appname. You may
            pass False to disable it.
        "version" is an optional version path element to append to the
            path. You might want to use this if you want multiple versions
            of your app to be able to run independently. If used, this
            would typically be "<major>.<minor>".
            Only applied when appname is present.
        "opinion" (boolean) can be False to disable the appending of
            "Logs" to the base app data dir for Windows, and "log" to the
            base cache dir for Unix. See discussion below.

    Typical user log directories are:
        Mac OS X:   ~/Library/Logs/<AppName>
        Unix:       ~/.cache/<AppName>/log  # or under $XDG_CACHE_HOME if defined
        Win XP:     C:\Documents and Settings\<username>\Local Settings\Application Data\<AppAuthor>\<AppName>\Logs
        Vista:      C:\Users\<username>\AppData\Local\<AppAuthor>\<AppName>\Logs

    On Windows the only suggestion in the MSDN docs is that local settings
    go in the `CSIDL_LOCAL_APPDATA` directory. (Note: I'm interested in
    examples of what some windows apps use for a logs dir.)

    OPINION: This function appends "Logs" to the `CSIDL_LOCAL_APPDATA`
    value for Windows and appends "log" to the user cache dir for Unix.
    This can be disabled with the `opinion=False` option.
    """
    if system == "darwin":
        path = os.path.join(
            os.path.expanduser('~/Library/Logs'),
            appname)
    elif system == "win32":
        path = user_data_dir(appname, appauthor, version)
        version = False
        if opinion:
            path = os.path.join(path, "Logs")
    else:
        path = user_cache_dir(appname, appauthor, version)
        version = False
        if opinion:
            path = os.path.join(path, "log")
    if appname and version:
        path = os.path.join(path, version)
    return path


class AppDirs(object):
    """Convenience wrapper for getting application dirs."""
    def __init__(self, appname=None, appauthor=None, version=None,
            roaming=False, multipath=False):
        self.appname = appname
        self.appauthor = appauthor
        self.version = version
        self.roaming = roaming
        self.multipath = multipath

    @property
    def user_data_dir(self):
        return user_data_dir(self.appname, self.appauthor,
                             version=self.version, roaming=self.roaming)

    @property
    def site_data_dir(self):
        return site_data_dir(self.appname, self.appauthor,
                             version=self.version, multipath=self.multipath)

    @property
    def user_config_dir(self):
        return user_config_dir(self.appname, self.appauthor,
                               version=self.version, roaming=self.roaming)

    @property
    def site_config_dir(self):
        return site_config_dir(self.appname, self.appauthor,
                             version=self.version, multipath=self.multipath)

    @property
    def user_cache_dir(self):
        return user_cache_dir(self.appname, self.appauthor,
                              version=self.version)

    @property
    def user_state_dir(self):
        return user_state_dir(self.appname, self.appauthor,
                              version=self.version)

    @property
    def user_log_dir(self):
        return user_log_dir(self.appname, self.appauthor,
                            version=self.version)


#---- internal support stuff

def _get_win_folder_from_registry(csidl_name):
    """This is a fallback technique at best. I'm not sure if using the
    registry for this guarantees us the correct answer for all CSIDL_*
    names.
    """
    if PY3:
      import winreg as _winreg
    else:
      import _winreg

    shell_folder_name = {
        "CSIDL_APPDATA": "AppData",
        "CSIDL_COMMON_APPDATA": "Common AppData",
        "CSIDL_LOCAL_APPDATA": "Local AppData",
    }[csidl_name]

    key = _winreg.OpenKey(
        _winreg.HKEY_CURRENT_USER,
        r"Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders"
    )
    dir, type = _winreg.QueryValueEx(key, shell_folder_name)
    return dir


def _get_win_folder_with_pywin32(csidl_name):
    from win32com.shell import shellcon, shell
    dir = shell.SHGetFolderPath(0, getattr(shellcon, csidl_name), 0, 0)
    # Try to make this a unicode path because SHGetFolderPath does
    # not return unicode strings when there is unicode data in the
    # path.
    try:
        dir = unicode(dir)

        # Downgrade to short path name if have highbit chars. See
        # <http://bugs.activestate.com/show_bug.cgi?id=85099>.
        has_high_char = False
        for c in dir:
            if ord(c) > 255:
                has_high_char = True
                break
        if has_high_char:
            try:
                import win32api
                dir = win32api.GetShortPathName(dir)
            except ImportError:
                pass
    except UnicodeError:
        pass
    return dir


def _get_win_folder_with_ctypes(csidl_name):
    import ctypes

    csidl_const = {
        "CSIDL_APPDATA": 26,
        "CSIDL_COMMON_APPDATA": 35,
        "CSIDL_LOCAL_APPDATA": 28,
    }[csidl_name]

    buf = ctypes.create_unicode_buffer(1024)
    ctypes.windll.shell32.SHGetFolderPathW(None, csidl_const, None, 0, buf)

    # Downgrade to short path name if have highbit chars. See
    # <http://bugs.activestate.com/show_bug.cgi?id=85099>.
    has_high_char = False
    for c in buf:
        if ord(c) > 255:
            has_high_char = True
            break
    if has_high_char:
        buf2 = ctypes.create_unicode_buffer(1024)
        if ctypes.windll.kernel32.GetShortPathNameW(buf.value, buf2, 1024):
            buf = buf2

    return buf.value

def _get_win_folder_with_jna(csidl_name):
    import array
    from com.sun import jna
    from com.sun.jna.platform import win32

    buf_size = win32.WinDef.MAX_PATH * 2
    buf = array.zeros('c', buf_size)
    shell = win32.Shell32.INSTANCE
    shell.SHGetFolderPath(None, getattr(win32.ShlObj, csidl_name), None, win32.ShlObj.SHGFP_TYPE_CURRENT, buf)
    dir = jna.Native.toString(buf.tostring()).rstrip("\0")

    # Downgrade to short path name if have highbit chars. See
    # <http://bugs.activestate.com/show_bug.cgi?id=85099>.
    has_high_char = False
    for c in dir:
        if ord(c) > 255:
            has_high_char = True
            break
    if has_high_char:
        buf = array.zeros('c', buf_size)
        kernel = win32.Kernel32.INSTANCE
        if kernel.GetShortPathName(dir, buf, buf_size):
            dir = jna.Native.toString(buf.tostring()).rstrip("\0")

    return dir

if system == "win32":
    try:
        import win32com.shell
        _get_win_folder = _get_win_folder_with_pywin32
    except ImportError:
        try:
            from ctypes import windll
            _get_win_folder = _get_win_folder_with_ctypes
        except ImportError:
            try:
                import com.sun.jna
                _get_win_folder = _get_win_folder_with_jna
            except ImportError:
                _get_win_folder = _get_win_folder_from_registry


#---- self test code

if __name__ == "__main__":
    appname = "MyApp"
    appauthor = "MyCompany"

    props = ("user_data_dir",
             "user_config_dir",
             "user_cache_dir",
             "user_state_dir",
             "user_log_dir",
             "site_data_dir",
             "site_config_dir")

    print("-- app dirs %s --" % __version__)

    print("-- app dirs (with optional 'version')")
    dirs = AppDirs(appname, appauthor, version="1.0")
    for prop in props:
        print("%s: %s" % (prop, getattr(dirs, prop)))

    print("\n-- app dirs (without optional 'version')")
    dirs = AppDirs(appname, appauthor)
    for prop in props:
        print("%s: %s" % (prop, getattr(dirs, prop)))

    print("\n-- app dirs (without optional 'appauthor')")
    dirs = AppDirs(appname)
    for prop in props:
        print("%s: %s" % (prop, getattr(dirs, prop)))

    print("\n-- app dirs (with disabled 'appauthor')")
    dirs = AppDirs(appname, appauthor=False)
    for prop in props:
        print("%s: %s" % (prop, getattr(dirs, prop)))
site-packages/pkg_resources/extern/__pycache__/__init__.cpython-39.pyc000064400000005560150564745010022061 0ustar00a

(�Re:	�@s4ddlZddlZGdd�d�ZdZeee���dS)�Nc@sXeZdZdZddd�Zedd��Zdd	�Zd
d�Zdd
�Z	dd�Z
ddd�Zdd�ZdS)�VendorImporterz�
    A PEP 302 meta path importer for finding optionally-vendored
    or otherwise naturally-installed packages from root_name.
    �NcCs&||_t|�|_|p|�dd�|_dS)NZextern�_vendor)�	root_name�set�vendored_names�replace�
vendor_pkg)�selfrrr	rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/extern/__init__.py�__init__s
zVendorImporter.__init__ccs|jdVdVdS)zL
        Search first the vendor package then as a natural package.
        �.�N)r	�r
rrr�search_pathszVendorImporter.search_pathcCs.|�|jd�\}}}|o,tt|j|j��S)z,Figure out if the target module is vendored.r
)�	partitionr�any�map�
startswithr)r
�fullname�root�base�targetrrr�_module_matches_namespacesz(VendorImporter._module_matches_namespacec	Cs~|�|jd�\}}}|jD]F}z.||}t|�tj|}|tj|<|WSty`Yq0qtdjfit�����dS)zK
        Iterate over the search path to locate and load fullname.
        r
z�The '{target}' package is required; normally this is bundled with this package so if you get this warning, consult the packager of your distribution.N)	rrr�
__import__�sys�modules�ImportError�format�locals)r
rrrr�prefixZextant�modrrr�load_modules



��zVendorImporter.load_modulecCs|�|j�S�N)r"�name)r
�specrrr�
create_module3szVendorImporter.create_modulecCsdSr#r)r
�modulerrr�exec_module6szVendorImporter.exec_modulecCs|�|�rtj�||�SdS)z(Return a module spec for vendored names.N)r�	importlib�util�spec_from_loader)r
r�pathrrrr�	find_spec9s��zVendorImporter.find_speccCs|tjvrtj�|�dS)zR
        Install this importer into sys.meta_path if not already present.
        N)r�	meta_path�appendrrrr�install@s
zVendorImporter.install)rN)NN)
�__name__�
__module__�__qualname__�__doc__r�propertyrrr"r&r(r-r0rrrrrs


r)�	packaging�	pyparsing�appdirs)�importlib.utilr)rr�namesr1r0rrrr�<module>sCsite-packages/pkg_resources/extern/__init__.py000064400000004472150564745010015573 0ustar00import importlib.util
import sys


class VendorImporter:
    """
    A PEP 302 meta path importer for finding optionally-vendored
    or otherwise naturally-installed packages from root_name.
    """

    def __init__(self, root_name, vendored_names=(), vendor_pkg=None):
        self.root_name = root_name
        self.vendored_names = set(vendored_names)
        self.vendor_pkg = vendor_pkg or root_name.replace('extern', '_vendor')

    @property
    def search_path(self):
        """
        Search first the vendor package then as a natural package.
        """
        yield self.vendor_pkg + '.'
        yield ''

    def _module_matches_namespace(self, fullname):
        """Figure out if the target module is vendored."""
        root, base, target = fullname.partition(self.root_name + '.')
        return not root and any(map(target.startswith, self.vendored_names))

    def load_module(self, fullname):
        """
        Iterate over the search path to locate and load fullname.
        """
        root, base, target = fullname.partition(self.root_name + '.')
        for prefix in self.search_path:
            try:
                extant = prefix + target
                __import__(extant)
                mod = sys.modules[extant]
                sys.modules[fullname] = mod
                return mod
            except ImportError:
                pass
        else:
            raise ImportError(
                "The '{target}' package is required; "
                "normally this is bundled with this package so if you get "
                "this warning, consult the packager of your "
                "distribution.".format(**locals())
            )

    def create_module(self, spec):
        return self.load_module(spec.name)

    def exec_module(self, module):
        pass

    def find_spec(self, fullname, path=None, target=None):
        """Return a module spec for vendored names."""
        return (
            importlib.util.spec_from_loader(fullname, self)
            if self._module_matches_namespace(fullname) else None
        )

    def install(self):
        """
        Install this importer into sys.meta_path if not already present.
        """
        if self not in sys.meta_path:
            sys.meta_path.append(self)


names = 'packaging', 'pyparsing', 'appdirs'
VendorImporter(__name__, names).install()
site-packages/pkg_resources/tests/data/my-test-package-source/__pycache__/setup.cpython-39.pyc000064400000000557150564745010026522 0ustar00a

(�Reh�@sddlZejdddd�dS)�Nzmy-test-packagez1.0T)�name�versionZzip_safe)�
setuptools�setup�rr��/builddir/build/BUILDROOT/alt-python39-setuptools-58.3.0-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pkg_resources/tests/data/my-test-package-source/setup.py�<module>s
�site-packages/pkg_resources/tests/data/my-test-package-source/setup.py000064400000000150150564745010022220 0ustar00import setuptools
setuptools.setup(
    name="my-test-package",
    version="1.0",
    zip_safe=True,
)
site-packages/pkg_resources/__init__.py000064400000323252150564745010014266 0ustar00"""
Package resource API
--------------------

A resource is a logical file contained within a package, or a logical
subdirectory thereof.  The package resource API expects resource names
to have their path parts separated with ``/``, *not* whatever the local
path separator is.  Do not use os.path operations to manipulate resource
names being passed into the API.

The package resource API is designed to work with normal filesystem packages,
.egg files, and unpacked .egg files.  It can also work in a limited way with
.zip files and with custom PEP 302 loaders that support the ``get_data()``
method.
"""

import sys
import os
import io
import time
import re
import types
import zipfile
import zipimport
import warnings
import stat
import functools
import pkgutil
import operator
import platform
import collections
import plistlib
import email.parser
import errno
import tempfile
import textwrap
import itertools
import inspect
import ntpath
import posixpath
import importlib
from pkgutil import get_importer

try:
    import _imp
except ImportError:
    # Python 3.2 compatibility
    import imp as _imp

try:
    FileExistsError
except NameError:
    FileExistsError = OSError

# capture these to bypass sandboxing
from os import utime
try:
    from os import mkdir, rename, unlink
    WRITE_SUPPORT = True
except ImportError:
    # no write support, probably under GAE
    WRITE_SUPPORT = False

from os import open as os_open
from os.path import isdir, split

try:
    import importlib.machinery as importlib_machinery
    # access attribute to force import under delayed import mechanisms.
    importlib_machinery.__name__
except ImportError:
    importlib_machinery = None

from pkg_resources.extern import appdirs
from pkg_resources.extern import packaging
__import__('pkg_resources.extern.packaging.version')
__import__('pkg_resources.extern.packaging.specifiers')
__import__('pkg_resources.extern.packaging.requirements')
__import__('pkg_resources.extern.packaging.markers')

if sys.version_info < (3, 5):
    raise RuntimeError("Python 3.5 or later is required")

# declare some globals that will be defined later to
# satisfy the linters.
require = None
working_set = None
add_activation_listener = None
resources_stream = None
cleanup_resources = None
resource_dir = None
resource_stream = None
set_extraction_path = None
resource_isdir = None
resource_string = None
iter_entry_points = None
resource_listdir = None
resource_filename = None
resource_exists = None
_distribution_finders = None
_namespace_handlers = None
_namespace_packages = None


class PEP440Warning(RuntimeWarning):
    """
    Used when there is an issue with a version or specifier not complying with
    PEP 440.
    """


def parse_version(v):
    try:
        return packaging.version.Version(v)
    except packaging.version.InvalidVersion:
        return packaging.version.LegacyVersion(v)


_state_vars = {}


def _declare_state(vartype, **kw):
    globals().update(kw)
    _state_vars.update(dict.fromkeys(kw, vartype))


def __getstate__():
    state = {}
    g = globals()
    for k, v in _state_vars.items():
        state[k] = g['_sget_' + v](g[k])
    return state


def __setstate__(state):
    g = globals()
    for k, v in state.items():
        g['_sset_' + _state_vars[k]](k, g[k], v)
    return state


def _sget_dict(val):
    return val.copy()


def _sset_dict(key, ob, state):
    ob.clear()
    ob.update(state)


def _sget_object(val):
    return val.__getstate__()


def _sset_object(key, ob, state):
    ob.__setstate__(state)


_sget_none = _sset_none = lambda *args: None


def get_supported_platform():
    """Return this platform's maximum compatible version.

    distutils.util.get_platform() normally reports the minimum version
    of macOS that would be required to *use* extensions produced by
    distutils.  But what we want when checking compatibility is to know the
    version of macOS that we are *running*.  To allow usage of packages that
    explicitly require a newer version of macOS, we must also know the
    current version of the OS.

    If this condition occurs for any other platform with a version in its
    platform strings, this function should be extended accordingly.
    """
    plat = get_build_platform()
    m = macosVersionString.match(plat)
    if m is not None and sys.platform == "darwin":
        try:
            plat = 'macosx-%s-%s' % ('.'.join(_macos_vers()[:2]), m.group(3))
        except ValueError:
            # not macOS
            pass
    return plat


__all__ = [
    # Basic resource access and distribution/entry point discovery
    'require', 'run_script', 'get_provider', 'get_distribution',
    'load_entry_point', 'get_entry_map', 'get_entry_info',
    'iter_entry_points',
    'resource_string', 'resource_stream', 'resource_filename',
    'resource_listdir', 'resource_exists', 'resource_isdir',

    # Environmental control
    'declare_namespace', 'working_set', 'add_activation_listener',
    'find_distributions', 'set_extraction_path', 'cleanup_resources',
    'get_default_cache',

    # Primary implementation classes
    'Environment', 'WorkingSet', 'ResourceManager',
    'Distribution', 'Requirement', 'EntryPoint',

    # Exceptions
    'ResolutionError', 'VersionConflict', 'DistributionNotFound',
    'UnknownExtra', 'ExtractionError',

    # Warnings
    'PEP440Warning',

    # Parsing functions and string utilities
    'parse_requirements', 'parse_version', 'safe_name', 'safe_version',
    'get_platform', 'compatible_platforms', 'yield_lines', 'split_sections',
    'safe_extra', 'to_filename', 'invalid_marker', 'evaluate_marker',

    # filesystem utilities
    'ensure_directory', 'normalize_path',

    # Distribution "precedence" constants
    'EGG_DIST', 'BINARY_DIST', 'SOURCE_DIST', 'CHECKOUT_DIST', 'DEVELOP_DIST',

    # "Provider" interfaces, implementations, and registration/lookup APIs
    'IMetadataProvider', 'IResourceProvider', 'FileMetadata',
    'PathMetadata', 'EggMetadata', 'EmptyProvider', 'empty_provider',
    'NullProvider', 'EggProvider', 'DefaultProvider', 'ZipProvider',
    'register_finder', 'register_namespace_handler', 'register_loader_type',
    'fixup_namespace_packages', 'get_importer',

    # Warnings
    'PkgResourcesDeprecationWarning',

    # Deprecated/backward compatibility only
    'run_main', 'AvailableDistributions',
]


class ResolutionError(Exception):
    """Abstract base for dependency resolution errors"""

    def __repr__(self):
        return self.__class__.__name__ + repr(self.args)


class VersionConflict(ResolutionError):
    """
    An already-installed version conflicts with the requested version.

    Should be initialized with the installed Distribution and the requested
    Requirement.
    """

    _template = "{self.dist} is installed but {self.req} is required"

    @property
    def dist(self):
        return self.args[0]

    @property
    def req(self):
        return self.args[1]

    def report(self):
        return self._template.format(**locals())

    def with_context(self, required_by):
        """
        If required_by is non-empty, return a version of self that is a
        ContextualVersionConflict.
        """
        if not required_by:
            return self
        args = self.args + (required_by,)
        return ContextualVersionConflict(*args)


class ContextualVersionConflict(VersionConflict):
    """
    A VersionConflict that accepts a third parameter, the set of the
    requirements that required the installed Distribution.
    """

    _template = VersionConflict._template + ' by {self.required_by}'

    @property
    def required_by(self):
        return self.args[2]


class DistributionNotFound(ResolutionError):
    """A requested distribution was not found"""

    _template = ("The '{self.req}' distribution was not found "
                 "and is required by {self.requirers_str}")

    @property
    def req(self):
        return self.args[0]

    @property
    def requirers(self):
        return self.args[1]

    @property
    def requirers_str(self):
        if not self.requirers:
            return 'the application'
        return ', '.join(self.requirers)

    def report(self):
        return self._template.format(**locals())

    def __str__(self):
        return self.report()


class UnknownExtra(ResolutionError):
    """Distribution doesn't have an "extra feature" of the given name"""


_provider_factories = {}

PY_MAJOR = '{}.{}'.format(*sys.version_info)
EGG_DIST = 3
BINARY_DIST = 2
SOURCE_DIST = 1
CHECKOUT_DIST = 0
DEVELOP_DIST = -1


def register_loader_type(loader_type, provider_factory):
    """Register `provider_factory` to make providers for `loader_type`

    `loader_type` is the type or class of a PEP 302 ``module.__loader__``,
    and `provider_factory` is a function that, passed a *module* object,
    returns an ``IResourceProvider`` for that module.
    """
    _provider_factories[loader_type] = provider_factory


def get_provider(moduleOrReq):
    """Return an IResourceProvider for the named module or requirement"""
    if isinstance(moduleOrReq, Requirement):
        return working_set.find(moduleOrReq) or require(str(moduleOrReq))[0]
    try:
        module = sys.modules[moduleOrReq]
    except KeyError:
        __import__(moduleOrReq)
        module = sys.modules[moduleOrReq]
    loader = getattr(module, '__loader__', None)
    return _find_adapter(_provider_factories, loader)(module)


def _macos_vers(_cache=[]):
    if not _cache:
        version = platform.mac_ver()[0]
        # fallback for MacPorts
        if version == '':
            plist = '/System/Library/CoreServices/SystemVersion.plist'
            if os.path.exists(plist):
                if hasattr(plistlib, 'readPlist'):
                    plist_content = plistlib.readPlist(plist)
                    if 'ProductVersion' in plist_content:
                        version = plist_content['ProductVersion']

        _cache.append(version.split('.'))
    return _cache[0]


def _macos_arch(machine):
    return {'PowerPC': 'ppc', 'Power_Macintosh': 'ppc'}.get(machine, machine)


def get_build_platform():
    """Return this platform's string for platform-specific distributions

    XXX Currently this is the same as ``distutils.util.get_platform()``, but it
    needs some hacks for Linux and macOS.
    """
    from sysconfig import get_platform

    plat = get_platform()
    if sys.platform == "darwin" and not plat.startswith('macosx-'):
        try:
            version = _macos_vers()
            machine = os.uname()[4].replace(" ", "_")
            return "macosx-%d.%d-%s" % (
                int(version[0]), int(version[1]),
                _macos_arch(machine),
            )
        except ValueError:
            # if someone is running a non-Mac darwin system, this will fall
            # through to the default implementation
            pass
    return plat


macosVersionString = re.compile(r"macosx-(\d+)\.(\d+)-(.*)")
darwinVersionString = re.compile(r"darwin-(\d+)\.(\d+)\.(\d+)-(.*)")
# XXX backward compat
get_platform = get_build_platform


def compatible_platforms(provided, required):
    """Can code for the `provided` platform run on the `required` platform?

    Returns true if either platform is ``None``, or the platforms are equal.

    XXX Needs compatibility checks for Linux and other unixy OSes.
    """
    if provided is None or required is None or provided == required:
        # easy case
        return True

    # macOS special cases
    reqMac = macosVersionString.match(required)
    if reqMac:
        provMac = macosVersionString.match(provided)

        # is this a Mac package?
        if not provMac:
            # this is backwards compatibility for packages built before
            # setuptools 0.6. All packages built after this point will
            # use the new macOS designation.
            provDarwin = darwinVersionString.match(provided)
            if provDarwin:
                dversion = int(provDarwin.group(1))
                macosversion = "%s.%s" % (reqMac.group(1), reqMac.group(2))
                if dversion == 7 and macosversion >= "10.3" or \
                        dversion == 8 and macosversion >= "10.4":
                    return True
            # egg isn't macOS or legacy darwin
            return False

        # are they the same major version and machine type?
        if provMac.group(1) != reqMac.group(1) or \
                provMac.group(3) != reqMac.group(3):
            return False

        # is the required OS major update >= the provided one?
        if int(provMac.group(2)) > int(reqMac.group(2)):
            return False

        return True

    # XXX Linux and other platforms' special cases should go here
    return False


def run_script(dist_spec, script_name):
    """Locate distribution `dist_spec` and run its `script_name` script"""
    ns = sys._getframe(1).f_globals
    name = ns['__name__']
    ns.clear()
    ns['__name__'] = name
    require(dist_spec)[0].run_script(script_name, ns)


# backward compatibility
run_main = run_script


def get_distribution(dist):
    """Return a current distribution object for a Requirement or string"""
    if isinstance(dist, str):
        dist = Requirement.parse(dist)
    if isinstance(dist, Requirement):
        dist = get_provider(dist)
    if not isinstance(dist, Distribution):
        raise TypeError("Expected string, Requirement, or Distribution", dist)
    return dist


def load_entry_point(dist, group, name):
    """Return `name` entry point of `group` for `dist` or raise ImportError"""
    return get_distribution(dist).load_entry_point(group, name)


def get_entry_map(dist, group=None):
    """Return the entry point map for `group`, or the full entry map"""
    return get_distribution(dist).get_entry_map(group)


def get_entry_info(dist, group, name):
    """Return the EntryPoint object for `group`+`name`, or ``None``"""
    return get_distribution(dist).get_entry_info(group, name)


class IMetadataProvider:
    def has_metadata(name):
        """Does the package's distribution contain the named metadata?"""

    def get_metadata(name):
        """The named metadata resource as a string"""

    def get_metadata_lines(name):
        """Yield named metadata resource as list of non-blank non-comment lines

       Leading and trailing whitespace is stripped from each line, and lines
       with ``#`` as the first non-blank character are omitted."""

    def metadata_isdir(name):
        """Is the named metadata a directory?  (like ``os.path.isdir()``)"""

    def metadata_listdir(name):
        """List of metadata names in the directory (like ``os.listdir()``)"""

    def run_script(script_name, namespace):
        """Execute the named script in the supplied namespace dictionary"""


class IResourceProvider(IMetadataProvider):
    """An object that provides access to package resources"""

    def get_resource_filename(manager, resource_name):
        """Return a true filesystem path for `resource_name`

        `manager` must be an ``IResourceManager``"""

    def get_resource_stream(manager, resource_name):
        """Return a readable file-like object for `resource_name`

        `manager` must be an ``IResourceManager``"""

    def get_resource_string(manager, resource_name):
        """Return a string containing the contents of `resource_name`

        `manager` must be an ``IResourceManager``"""

    def has_resource(resource_name):
        """Does the package contain the named resource?"""

    def resource_isdir(resource_name):
        """Is the named resource a directory?  (like ``os.path.isdir()``)"""

    def resource_listdir(resource_name):
        """List of resource names in the directory (like ``os.listdir()``)"""


class WorkingSet:
    """A collection of active distributions on sys.path (or a similar list)"""

    def __init__(self, entries=None):
        """Create working set from list of path entries (default=sys.path)"""
        self.entries = []
        self.entry_keys = {}
        self.by_key = {}
        self.callbacks = []

        if entries is None:
            entries = sys.path

        for entry in entries:
            self.add_entry(entry)

    @classmethod
    def _build_master(cls):
        """
        Prepare the master working set.
        """
        ws = cls()
        try:
            from __main__ import __requires__
        except ImportError:
            # The main program does not list any requirements
            return ws

        # ensure the requirements are met
        try:
            ws.require(__requires__)
        except VersionConflict:
            return cls._build_from_requirements(__requires__)

        return ws

    @classmethod
    def _build_from_requirements(cls, req_spec):
        """
        Build a working set from a requirement spec. Rewrites sys.path.
        """
        # try it without defaults already on sys.path
        # by starting with an empty path
        ws = cls([])
        reqs = parse_requirements(req_spec)
        dists = ws.resolve(reqs, Environment())
        for dist in dists:
            ws.add(dist)

        # add any missing entries from sys.path
        for entry in sys.path:
            if entry not in ws.entries:
                ws.add_entry(entry)

        # then copy back to sys.path
        sys.path[:] = ws.entries
        return ws

    def add_entry(self, entry):
        """Add a path item to ``.entries``, finding any distributions on it

        ``find_distributions(entry, True)`` is used to find distributions
        corresponding to the path entry, and they are added.  `entry` is
        always appended to ``.entries``, even if it is already present.
        (This is because ``sys.path`` can contain the same value more than
        once, and the ``.entries`` of the ``sys.path`` WorkingSet should always
        equal ``sys.path``.)
        """
        self.entry_keys.setdefault(entry, [])
        self.entries.append(entry)
        for dist in find_distributions(entry, True):
            self.add(dist, entry, False)

    def __contains__(self, dist):
        """True if `dist` is the active distribution for its project"""
        return self.by_key.get(dist.key) == dist

    def find(self, req):
        """Find a distribution matching requirement `req`

        If there is an active distribution for the requested project, this
        returns it as long as it meets the version requirement specified by
        `req`.  But, if there is an active distribution for the project and it
        does *not* meet the `req` requirement, ``VersionConflict`` is raised.
        If there is no active distribution for the requested project, ``None``
        is returned.
        """
        dist = self.by_key.get(req.key)
        if dist is not None and dist not in req:
            # XXX add more info
            raise VersionConflict(dist, req)
        return dist

    def iter_entry_points(self, group, name=None):
        """Yield entry point objects from `group` matching `name`

        If `name` is None, yields all entry points in `group` from all
        distributions in the working set, otherwise only ones matching
        both `group` and `name` are yielded (in distribution order).
        """
        return (
            entry
            for dist in self
            for entry in dist.get_entry_map(group).values()
            if name is None or name == entry.name
        )

    def run_script(self, requires, script_name):
        """Locate distribution for `requires` and run `script_name` script"""
        ns = sys._getframe(1).f_globals
        name = ns['__name__']
        ns.clear()
        ns['__name__'] = name
        self.require(requires)[0].run_script(script_name, ns)

    def __iter__(self):
        """Yield distributions for non-duplicate projects in the working set

        The yield order is the order in which the items' path entries were
        added to the working set.
        """
        seen = {}
        for item in self.entries:
            if item not in self.entry_keys:
                # workaround a cache issue
                continue

            for key in self.entry_keys[item]:
                if key not in seen:
                    seen[key] = 1
                    yield self.by_key[key]

    def add(self, dist, entry=None, insert=True, replace=False):
        """Add `dist` to working set, associated with `entry`

        If `entry` is unspecified, it defaults to the ``.location`` of `dist`.
        On exit from this routine, `entry` is added to the end of the working
        set's ``.entries`` (if it wasn't already present).

        `dist` is only added to the working set if it's for a project that
        doesn't already have a distribution in the set, unless `replace=True`.
        If it's added, any callbacks registered with the ``subscribe()`` method
        will be called.
        """
        if insert:
            dist.insert_on(self.entries, entry, replace=replace)

        if entry is None:
            entry = dist.location
        keys = self.entry_keys.setdefault(entry, [])
        keys2 = self.entry_keys.setdefault(dist.location, [])
        if not replace and dist.key in self.by_key:
            # ignore hidden distros
            return

        self.by_key[dist.key] = dist
        if dist.key not in keys:
            keys.append(dist.key)
        if dist.key not in keys2:
            keys2.append(dist.key)
        self._added_new(dist)

    # FIXME: 'WorkingSet.resolve' is too complex (11)
    def resolve(self, requirements, env=None, installer=None,  # noqa: C901
                replace_conflicting=False, extras=None):
        """List all distributions needed to (recursively) meet `requirements`

        `requirements` must be a sequence of ``Requirement`` objects.  `env`,
        if supplied, should be an ``Environment`` instance.  If
        not supplied, it defaults to all distributions available within any
        entry or distribution in the working set.  `installer`, if supplied,
        will be invoked with each requirement that cannot be met by an
        already-installed distribution; it should return a ``Distribution`` or
        ``None``.

        Unless `replace_conflicting=True`, raises a VersionConflict exception
        if
        any requirements are found on the path that have the correct name but
        the wrong version.  Otherwise, if an `installer` is supplied it will be
        invoked to obtain the correct version of the requirement and activate
        it.

        `extras` is a list of the extras to be used with these requirements.
        This is important because extra requirements may look like `my_req;
        extra = "my_extra"`, which would otherwise be interpreted as a purely
        optional requirement.  Instead, we want to be able to assert that these
        requirements are truly required.
        """

        # set up the stack
        requirements = list(requirements)[::-1]
        # set of processed requirements
        processed = {}
        # key -> dist
        best = {}
        to_activate = []

        req_extras = _ReqExtras()

        # Mapping of requirement to set of distributions that required it;
        # useful for reporting info about conflicts.
        required_by = collections.defaultdict(set)

        while requirements:
            # process dependencies breadth-first
            req = requirements.pop(0)
            if req in processed:
                # Ignore cyclic or redundant dependencies
                continue

            if not req_extras.markers_pass(req, extras):
                continue

            dist = best.get(req.key)
            if dist is None:
                # Find the best distribution and add it to the map
                dist = self.by_key.get(req.key)
                if dist is None or (dist not in req and replace_conflicting):
                    ws = self
                    if env is None:
                        if dist is None:
                            env = Environment(self.entries)
                        else:
                            # Use an empty environment and workingset to avoid
                            # any further conflicts with the conflicting
                            # distribution
                            env = Environment([])
                            ws = WorkingSet([])
                    dist = best[req.key] = env.best_match(
                        req, ws, installer,
                        replace_conflicting=replace_conflicting
                    )
                    if dist is None:
                        requirers = required_by.get(req, None)
                        raise DistributionNotFound(req, requirers)
                to_activate.append(dist)
            if dist not in req:
                # Oops, the "best" so far conflicts with a dependency
                dependent_req = required_by[req]
                raise VersionConflict(dist, req).with_context(dependent_req)

            # push the new requirements onto the stack
            new_requirements = dist.requires(req.extras)[::-1]
            requirements.extend(new_requirements)

            # Register the new requirements needed by req
            for new_requirement in new_requirements:
                required_by[new_requirement].add(req.project_name)
                req_extras[new_requirement] = req.extras

            processed[req] = True

        # return list of distros to activate
        return to_activate

    def find_plugins(
            self, plugin_env, full_env=None, installer=None, fallback=True):
        """Find all activatable distributions in `plugin_env`

        Example usage::

            distributions, errors = working_set.find_plugins(
                Environment(plugin_dirlist)
            )
            # add plugins+libs to sys.path
            map(working_set.add, distributions)
            # display errors
            print('Could not load', errors)

        The `plugin_env` should be an ``Environment`` instance that contains
        only distributions that are in the project's "plugin directory" or
        directories. The `full_env`, if supplied, should be an ``Environment``
        contains all currently-available distributions.  If `full_env` is not
        supplied, one is created automatically from the ``WorkingSet`` this
        method is called on, which will typically mean that every directory on
        ``sys.path`` will be scanned for distributions.

        `installer` is a standard installer callback as used by the
        ``resolve()`` method. The `fallback` flag indicates whether we should
        attempt to resolve older versions of a plugin if the newest version
        cannot be resolved.

        This method returns a 2-tuple: (`distributions`, `error_info`), where
        `distributions` is a list of the distributions found in `plugin_env`
        that were loadable, along with any other distributions that are needed
        to resolve their dependencies.  `error_info` is a dictionary mapping
        unloadable plugin distributions to an exception instance describing the
        error that occurred. Usually this will be a ``DistributionNotFound`` or
        ``VersionConflict`` instance.
        """

        plugin_projects = list(plugin_env)
        # scan project names in alphabetic order
        plugin_projects.sort()

        error_info = {}
        distributions = {}

        if full_env is None:
            env = Environment(self.entries)
            env += plugin_env
        else:
            env = full_env + plugin_env

        shadow_set = self.__class__([])
        # put all our entries in shadow_set
        list(map(shadow_set.add, self))

        for project_name in plugin_projects:

            for dist in plugin_env[project_name]:

                req = [dist.as_requirement()]

                try:
                    resolvees = shadow_set.resolve(req, env, installer)

                except ResolutionError as v:
                    # save error info
                    error_info[dist] = v
                    if fallback:
                        # try the next older version of project
                        continue
                    else:
                        # give up on this project, keep going
                        break

                else:
                    list(map(shadow_set.add, resolvees))
                    distributions.update(dict.fromkeys(resolvees))

                    # success, no need to try any more versions of this project
                    break

        distributions = list(distributions)
        distributions.sort()

        return distributions, error_info

    def require(self, *requirements):
        """Ensure that distributions matching `requirements` are activated

        `requirements` must be a string or a (possibly-nested) sequence
        thereof, specifying the distributions and versions required.  The
        return value is a sequence of the distributions that needed to be
        activated to fulfill the requirements; all relevant distributions are
        included, even if they were already activated in this working set.
        """
        needed = self.resolve(parse_requirements(requirements))

        for dist in needed:
            self.add(dist)

        return needed

    def subscribe(self, callback, existing=True):
        """Invoke `callback` for all distributions

        If `existing=True` (default),
        call on all existing ones, as well.
        """
        if callback in self.callbacks:
            return
        self.callbacks.append(callback)
        if not existing:
            return
        for dist in self:
            callback(dist)

    def _added_new(self, dist):
        for callback in self.callbacks:
            callback(dist)

    def __getstate__(self):
        return (
            self.entries[:], self.entry_keys.copy(), self.by_key.copy(),
            self.callbacks[:]
        )

    def __setstate__(self, e_k_b_c):
        entries, keys, by_key, callbacks = e_k_b_c
        self.entries = entries[:]
        self.entry_keys = keys.copy()
        self.by_key = by_key.copy()
        self.callbacks = callbacks[:]


class _ReqExtras(dict):
    """
    Map each requirement to the extras that demanded it.
    """

    def markers_pass(self, req, extras=None):
        """
        Evaluate markers for req against each extra that
        demanded it.

        Return False if the req has a marker and fails
        evaluation. Otherwise, return True.
        """
        extra_evals = (
            req.marker.evaluate({'extra': extra})
            for extra in self.get(req, ()) + (extras or (None,))
        )
        return not req.marker or any(extra_evals)


class Environment:
    """Searchable snapshot of distributions on a search path"""

    def __init__(
            self, search_path=None, platform=get_supported_platform(),
            python=PY_MAJOR):
        """Snapshot distributions available on a search path

        Any distributions found on `search_path` are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.

        `platform` is an optional string specifying the name of the platform
        that platform-specific distributions must be compatible with.  If
        unspecified, it defaults to the current platform.  `python` is an
        optional string naming the desired version of Python (e.g. ``'3.6'``);
        it defaults to the current version.

        You may explicitly set `platform` (and/or `python`) to ``None`` if you
        wish to map *all* distributions, not just those compatible with the
        running platform or Python version.
        """
        self._distmap = {}
        self.platform = platform
        self.python = python
        self.scan(search_path)

    def can_add(self, dist):
        """Is distribution `dist` acceptable for this environment?

        The distribution must match the platform and python version
        requirements specified when this environment was created, or False
        is returned.
        """
        py_compat = (
            self.python is None
            or dist.py_version is None
            or dist.py_version == self.python
        )
        return py_compat and compatible_platforms(dist.platform, self.platform)

    def remove(self, dist):
        """Remove `dist` from the environment"""
        self._distmap[dist.key].remove(dist)

    def scan(self, search_path=None):
        """Scan `search_path` for distributions usable in this environment

        Any distributions found are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.  Only distributions conforming to
        the platform/python version defined at initialization are added.
        """
        if search_path is None:
            search_path = sys.path

        for item in search_path:
            for dist in find_distributions(item):
                self.add(dist)

    def __getitem__(self, project_name):
        """Return a newest-to-oldest list of distributions for `project_name`

        Uses case-insensitive `project_name` comparison, assuming all the
        project's distributions use their project's name converted to all
        lowercase as their key.

        """
        distribution_key = project_name.lower()
        return self._distmap.get(distribution_key, [])

    def add(self, dist):
        """Add `dist` if we ``can_add()`` it and it has not already been added
        """
        if self.can_add(dist) and dist.has_version():
            dists = self._distmap.setdefault(dist.key, [])
            if dist not in dists:
                dists.append(dist)
                dists.sort(key=operator.attrgetter('hashcmp'), reverse=True)

    def best_match(
            self, req, working_set, installer=None, replace_conflicting=False):
        """Find distribution best matching `req` and usable on `working_set`

        This calls the ``find(req)`` method of the `working_set` to see if a
        suitable distribution is already active.  (This may raise
        ``VersionConflict`` if an unsuitable version of the project is already
        active in the specified `working_set`.)  If a suitable distribution
        isn't active, this method returns the newest distribution in the
        environment that meets the ``Requirement`` in `req`.  If no suitable
        distribution is found, and `installer` is supplied, then the result of
        calling the environment's ``obtain(req, installer)`` method will be
        returned.
        """
        try:
            dist = working_set.find(req)
        except VersionConflict:
            if not replace_conflicting:
                raise
            dist = None
        if dist is not None:
            return dist
        for dist in self[req.key]:
            if dist in req:
                return dist
        # try to download/install
        return self.obtain(req, installer)

    def obtain(self, requirement, installer=None):
        """Obtain a distribution matching `requirement` (e.g. via download)

        Obtain a distro that matches requirement (e.g. via download).  In the
        base ``Environment`` class, this routine just returns
        ``installer(requirement)``, unless `installer` is None, in which case
        None is returned instead.  This method is a hook that allows subclasses
        to attempt other ways of obtaining a distribution before falling back
        to the `installer` argument."""
        if installer is not None:
            return installer(requirement)

    def __iter__(self):
        """Yield the unique project names of the available distributions"""
        for key in self._distmap.keys():
            if self[key]:
                yield key

    def __iadd__(self, other):
        """In-place addition of a distribution or environment"""
        if isinstance(other, Distribution):
            self.add(other)
        elif isinstance(other, Environment):
            for project in other:
                for dist in other[project]:
                    self.add(dist)
        else:
            raise TypeError("Can't add %r to environment" % (other,))
        return self

    def __add__(self, other):
        """Add an environment or distribution to an environment"""
        new = self.__class__([], platform=None, python=None)
        for env in self, other:
            new += env
        return new


# XXX backward compatibility
AvailableDistributions = Environment


class ExtractionError(RuntimeError):
    """An error occurred extracting a resource

    The following attributes are available from instances of this exception:

    manager
        The resource manager that raised this exception

    cache_path
        The base directory for resource extraction

    original_error
        The exception instance that caused extraction to fail
    """


class ResourceManager:
    """Manage resource extraction and packages"""
    extraction_path = None

    def __init__(self):
        self.cached_files = {}

    def resource_exists(self, package_or_requirement, resource_name):
        """Does the named resource exist?"""
        return get_provider(package_or_requirement).has_resource(resource_name)

    def resource_isdir(self, package_or_requirement, resource_name):
        """Is the named resource an existing directory?"""
        return get_provider(package_or_requirement).resource_isdir(
            resource_name
        )

    def resource_filename(self, package_or_requirement, resource_name):
        """Return a true filesystem path for specified resource"""
        return get_provider(package_or_requirement).get_resource_filename(
            self, resource_name
        )

    def resource_stream(self, package_or_requirement, resource_name):
        """Return a readable file-like object for specified resource"""
        return get_provider(package_or_requirement).get_resource_stream(
            self, resource_name
        )

    def resource_string(self, package_or_requirement, resource_name):
        """Return specified resource as a string"""
        return get_provider(package_or_requirement).get_resource_string(
            self, resource_name
        )

    def resource_listdir(self, package_or_requirement, resource_name):
        """List the contents of the named resource directory"""
        return get_provider(package_or_requirement).resource_listdir(
            resource_name
        )

    def extraction_error(self):
        """Give an error message for problems extracting file(s)"""

        old_exc = sys.exc_info()[1]
        cache_path = self.extraction_path or get_default_cache()

        tmpl = textwrap.dedent("""
            Can't extract file(s) to egg cache

            The following error occurred while trying to extract file(s)
            to the Python egg cache:

              {old_exc}

            The Python egg cache directory is currently set to:

              {cache_path}

            Perhaps your account does not have write access to this directory?
            You can change the cache directory by setting the PYTHON_EGG_CACHE
            environment variable to point to an accessible directory.
            """).lstrip()
        err = ExtractionError(tmpl.format(**locals()))
        err.manager = self
        err.cache_path = cache_path
        err.original_error = old_exc
        raise err

    def get_cache_path(self, archive_name, names=()):
        """Return absolute location in cache for `archive_name` and `names`

        The parent directory of the resulting path will be created if it does
        not already exist.  `archive_name` should be the base filename of the
        enclosing egg (which may not be the name of the enclosing zipfile!),
        including its ".egg" extension.  `names`, if provided, should be a
        sequence of path name parts "under" the egg's extraction location.

        This method should only be called by resource providers that need to
        obtain an extraction location, and only for names they intend to
        extract, as it tracks the generated names for possible cleanup later.
        """
        extract_path = self.extraction_path or get_default_cache()
        target_path = os.path.join(extract_path, archive_name + '-tmp', *names)
        try:
            _bypass_ensure_directory(target_path)
        except Exception:
            self.extraction_error()

        self._warn_unsafe_extraction_path(extract_path)

        self.cached_files[target_path] = 1
        return target_path

    @staticmethod
    def _warn_unsafe_extraction_path(path):
        """
        If the default extraction path is overridden and set to an insecure
        location, such as /tmp, it opens up an opportunity for an attacker to
        replace an extracted file with an unauthorized payload. Warn the user
        if a known insecure location is used.

        See Distribute #375 for more details.
        """
        if os.name == 'nt' and not path.startswith(os.environ['windir']):
            # On Windows, permissions are generally restrictive by default
            #  and temp directories are not writable by other users, so
            #  bypass the warning.
            return
        mode = os.stat(path).st_mode
        if mode & stat.S_IWOTH or mode & stat.S_IWGRP:
            msg = (
                "Extraction path is writable by group/others "
                "and vulnerable to attack when "
                "used with get_resource_filename ({path}). "
                "Consider a more secure "
                "location (set with .set_extraction_path or the "
                "PYTHON_EGG_CACHE environment variable)."
            ).format(**locals())
            warnings.warn(msg, UserWarning)

    def postprocess(self, tempname, filename):
        """Perform any platform-specific postprocessing of `tempname`

        This is where Mac header rewrites should be done; other platforms don't
        have anything special they should do.

        Resource providers should call this method ONLY after successfully
        extracting a compressed resource.  They must NOT call it on resources
        that are already in the filesystem.

        `tempname` is the current (temporary) name of the file, and `filename`
        is the name it will be renamed to by the caller after this routine
        returns.
        """

        if os.name == 'posix':
            # Make the resource executable
            mode = ((os.stat(tempname).st_mode) | 0o555) & 0o7777
            os.chmod(tempname, mode)

    def set_extraction_path(self, path):
        """Set the base path where resources will be extracted to, if needed.

        If you do not call this routine before any extractions take place, the
        path defaults to the return value of ``get_default_cache()``.  (Which
        is based on the ``PYTHON_EGG_CACHE`` environment variable, with various
        platform-specific fallbacks.  See that routine's documentation for more
        details.)

        Resources are extracted to subdirectories of this path based upon
        information given by the ``IResourceProvider``.  You may set this to a
        temporary directory, but then you must call ``cleanup_resources()`` to
        delete the extracted files when done.  There is no guarantee that
        ``cleanup_resources()`` will be able to remove all extracted files.

        (Note: you may not change the extraction path for a given resource
        manager once resources have been extracted, unless you first call
        ``cleanup_resources()``.)
        """
        if self.cached_files:
            raise ValueError(
                "Can't change extraction path, files already extracted"
            )

        self.extraction_path = path

    def cleanup_resources(self, force=False):
        """
        Delete all extracted resource files and directories, returning a list
        of the file and directory names that could not be successfully removed.
        This function does not have any concurrency protection, so it should
        generally only be called when the extraction path is a temporary
        directory exclusive to a single process.  This method is not
        automatically called; you must call it explicitly or register it as an
        ``atexit`` function if you wish to ensure cleanup of a temporary
        directory used for extractions.
        """
        # XXX


def get_default_cache():
    """
    Return the ``PYTHON_EGG_CACHE`` environment variable
    or a platform-relevant user cache dir for an app
    named "Python-Eggs".
    """
    return (
        os.environ.get('PYTHON_EGG_CACHE')
        or appdirs.user_cache_dir(appname='Python-Eggs')
    )


def safe_name(name):
    """Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    """
    return re.sub('[^A-Za-z0-9.]+', '-', name)


def safe_version(version):
    """
    Convert an arbitrary string to a standard version string
    """
    try:
        # normalize the version
        return str(packaging.version.Version(version))
    except packaging.version.InvalidVersion:
        version = version.replace(' ', '.')
        return re.sub('[^A-Za-z0-9.]+', '-', version)


def safe_extra(extra):
    """Convert an arbitrary string to a standard 'extra' name

    Any runs of non-alphanumeric characters are replaced with a single '_',
    and the result is always lowercased.
    """
    return re.sub('[^A-Za-z0-9.-]+', '_', extra).lower()


def to_filename(name):
    """Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    """
    return name.replace('-', '_')


def invalid_marker(text):
    """
    Validate text as a PEP 508 environment marker; return an exception
    if invalid or False otherwise.
    """
    try:
        evaluate_marker(text)
    except SyntaxError as e:
        e.filename = None
        e.lineno = None
        return e
    return False


def evaluate_marker(text, extra=None):
    """
    Evaluate a PEP 508 environment marker.
    Return a boolean indicating the marker result in this environment.
    Raise SyntaxError if marker is invalid.

    This implementation uses the 'pyparsing' module.
    """
    try:
        marker = packaging.markers.Marker(text)
        return marker.evaluate()
    except packaging.markers.InvalidMarker as e:
        raise SyntaxError(e) from e


class NullProvider:
    """Try to implement resources and metadata for arbitrary PEP 302 loaders"""

    egg_name = None
    egg_info = None
    loader = None

    def __init__(self, module):
        self.loader = getattr(module, '__loader__', None)
        self.module_path = os.path.dirname(getattr(module, '__file__', ''))

    def get_resource_filename(self, manager, resource_name):
        return self._fn(self.module_path, resource_name)

    def get_resource_stream(self, manager, resource_name):
        return io.BytesIO(self.get_resource_string(manager, resource_name))

    def get_resource_string(self, manager, resource_name):
        return self._get(self._fn(self.module_path, resource_name))

    def has_resource(self, resource_name):
        return self._has(self._fn(self.module_path, resource_name))

    def _get_metadata_path(self, name):
        return self._fn(self.egg_info, name)

    def has_metadata(self, name):
        if not self.egg_info:
            return self.egg_info

        path = self._get_metadata_path(name)
        return self._has(path)

    def get_metadata(self, name):
        if not self.egg_info:
            return ""
        path = self._get_metadata_path(name)
        value = self._get(path)
        try:
            return value.decode('utf-8')
        except UnicodeDecodeError as exc:
            # Include the path in the error message to simplify
            # troubleshooting, and without changing the exception type.
            exc.reason += ' in {} file at path: {}'.format(name, path)
            raise

    def get_metadata_lines(self, name):
        return yield_lines(self.get_metadata(name))

    def resource_isdir(self, resource_name):
        return self._isdir(self._fn(self.module_path, resource_name))

    def metadata_isdir(self, name):
        return self.egg_info and self._isdir(self._fn(self.egg_info, name))

    def resource_listdir(self, resource_name):
        return self._listdir(self._fn(self.module_path, resource_name))

    def metadata_listdir(self, name):
        if self.egg_info:
            return self._listdir(self._fn(self.egg_info, name))
        return []

    def run_script(self, script_name, namespace):
        script = 'scripts/' + script_name
        if not self.has_metadata(script):
            raise ResolutionError(
                "Script {script!r} not found in metadata at {self.egg_info!r}"
                .format(**locals()),
            )
        script_text = self.get_metadata(script).replace('\r\n', '\n')
        script_text = script_text.replace('\r', '\n')
        script_filename = self._fn(self.egg_info, script)
        namespace['__file__'] = script_filename
        if os.path.exists(script_filename):
            with open(script_filename) as fid:
                source = fid.read()
            code = compile(source, script_filename, 'exec')
            exec(code, namespace, namespace)
        else:
            from linecache import cache
            cache[script_filename] = (
                len(script_text), 0, script_text.split('\n'), script_filename
            )
            script_code = compile(script_text, script_filename, 'exec')
            exec(script_code, namespace, namespace)

    def _has(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _isdir(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _listdir(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _fn(self, base, resource_name):
        self._validate_resource_path(resource_name)
        if resource_name:
            return os.path.join(base, *resource_name.split('/'))
        return base

    @staticmethod
    def _validate_resource_path(path):
        """
        Validate the resource paths according to the docs.
        https://setuptools.readthedocs.io/en/latest/pkg_resources.html#basic-resource-access

        >>> warned = getfixture('recwarn')
        >>> warnings.simplefilter('always')
        >>> vrp = NullProvider._validate_resource_path
        >>> vrp('foo/bar.txt')
        >>> bool(warned)
        False
        >>> vrp('../foo/bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('/foo/bar.txt')
        >>> bool(warned)
        True
        >>> vrp('foo/../../bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('foo/f../bar.txt')
        >>> bool(warned)
        False

        Windows path separators are straight-up disallowed.
        >>> vrp(r'\\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path \
is not allowed.

        >>> vrp(r'C:\\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path \
is not allowed.

        Blank values are allowed

        >>> vrp('')
        >>> bool(warned)
        False

        Non-string values are not.

        >>> vrp(None)
        Traceback (most recent call last):
        ...
        AttributeError: ...
        """
        invalid = (
            os.path.pardir in path.split(posixpath.sep) or
            posixpath.isabs(path) or
            ntpath.isabs(path)
        )
        if not invalid:
            return

        msg = "Use of .. or absolute path in a resource path is not allowed."

        # Aggressively disallow Windows absolute paths
        if ntpath.isabs(path) and not posixpath.isabs(path):
            raise ValueError(msg)

        # for compatibility, warn; in future
        # raise ValueError(msg)
        warnings.warn(
            msg[:-1] + " and will raise exceptions in a future release.",
            DeprecationWarning,
            stacklevel=4,
        )

    def _get(self, path):
        if hasattr(self.loader, 'get_data'):
            return self.loader.get_data(path)
        raise NotImplementedError(
            "Can't perform this operation for loaders without 'get_data()'"
        )


register_loader_type(object, NullProvider)


def _parents(path):
    """
    yield all parents of path including path
    """
    last = None
    while path != last:
        yield path
        last = path
        path, _ = os.path.split(path)


class EggProvider(NullProvider):
    """Provider based on a virtual filesystem"""

    def __init__(self, module):
        NullProvider.__init__(self, module)
        self._setup_prefix()

    def _setup_prefix(self):
        # Assume that metadata may be nested inside a "basket"
        # of multiple eggs and use module_path instead of .archive.
        eggs = filter(_is_egg_path, _parents(self.module_path))
        egg = next(eggs, None)
        egg and self._set_egg(egg)

    def _set_egg(self, path):
        self.egg_name = os.path.basename(path)
        self.egg_info = os.path.join(path, 'EGG-INFO')
        self.egg_root = path


class DefaultProvider(EggProvider):
    """Provides access to package resources in the filesystem"""

    def _has(self, path):
        return os.path.exists(path)

    def _isdir(self, path):
        return os.path.isdir(path)

    def _listdir(self, path):
        return os.listdir(path)

    def get_resource_stream(self, manager, resource_name):
        return open(self._fn(self.module_path, resource_name), 'rb')

    def _get(self, path):
        with open(path, 'rb') as stream:
            return stream.read()

    @classmethod
    def _register(cls):
        loader_names = 'SourceFileLoader', 'SourcelessFileLoader',
        for name in loader_names:
            loader_cls = getattr(importlib_machinery, name, type(None))
            register_loader_type(loader_cls, cls)


DefaultProvider._register()


class EmptyProvider(NullProvider):
    """Provider that returns nothing for all requests"""

    module_path = None

    _isdir = _has = lambda self, path: False

    def _get(self, path):
        return ''

    def _listdir(self, path):
        return []

    def __init__(self):
        pass


empty_provider = EmptyProvider()


class ZipManifests(dict):
    """
    zip manifest builder
    """

    @classmethod
    def build(cls, path):
        """
        Build a dictionary similar to the zipimport directory
        caches, except instead of tuples, store ZipInfo objects.

        Use a platform-specific path separator (os.sep) for the path keys
        for compatibility with pypy on Windows.
        """
        with zipfile.ZipFile(path) as zfile:
            items = (
                (
                    name.replace('/', os.sep),
                    zfile.getinfo(name),
                )
                for name in zfile.namelist()
            )
            return dict(items)

    load = build


class MemoizedZipManifests(ZipManifests):
    """
    Memoized zipfile manifests.
    """
    manifest_mod = collections.namedtuple('manifest_mod', 'manifest mtime')

    def load(self, path):
        """
        Load a manifest at path or return a suitable manifest already loaded.
        """
        path = os.path.normpath(path)
        mtime = os.stat(path).st_mtime

        if path not in self or self[path].mtime != mtime:
            manifest = self.build(path)
            self[path] = self.manifest_mod(manifest, mtime)

        return self[path].manifest


class ZipProvider(EggProvider):
    """Resource support for zips and eggs"""

    eagers = None
    _zip_manifests = MemoizedZipManifests()

    def __init__(self, module):
        EggProvider.__init__(self, module)
        self.zip_pre = self.loader.archive + os.sep

    def _zipinfo_name(self, fspath):
        # Convert a virtual filename (full path to file) into a zipfile subpath
        # usable with the zipimport directory cache for our target archive
        fspath = fspath.rstrip(os.sep)
        if fspath == self.loader.archive:
            return ''
        if fspath.startswith(self.zip_pre):
            return fspath[len(self.zip_pre):]
        raise AssertionError(
            "%s is not a subpath of %s" % (fspath, self.zip_pre)
        )

    def _parts(self, zip_path):
        # Convert a zipfile subpath into an egg-relative path part list.
        # pseudo-fs path
        fspath = self.zip_pre + zip_path
        if fspath.startswith(self.egg_root + os.sep):
            return fspath[len(self.egg_root) + 1:].split(os.sep)
        raise AssertionError(
            "%s is not a subpath of %s" % (fspath, self.egg_root)
        )

    @property
    def zipinfo(self):
        return self._zip_manifests.load(self.loader.archive)

    def get_resource_filename(self, manager, resource_name):
        if not self.egg_name:
            raise NotImplementedError(
                "resource_filename() only supported for .egg, not .zip"
            )
        # no need to lock for extraction, since we use temp names
        zip_path = self._resource_to_zip(resource_name)
        eagers = self._get_eager_resources()
        if '/'.join(self._parts(zip_path)) in eagers:
            for name in eagers:
                self._extract_resource(manager, self._eager_to_zip(name))
        return self._extract_resource(manager, zip_path)

    @staticmethod
    def _get_date_and_size(zip_stat):
        size = zip_stat.file_size
        # ymdhms+wday, yday, dst
        date_time = zip_stat.date_time + (0, 0, -1)
        # 1980 offset already done
        timestamp = time.mktime(date_time)
        return timestamp, size

    # FIXME: 'ZipProvider._extract_resource' is too complex (12)
    def _extract_resource(self, manager, zip_path):  # noqa: C901

        if zip_path in self._index():
            for name in self._index()[zip_path]:
                last = self._extract_resource(
                    manager, os.path.join(zip_path, name)
                )
            # return the extracted directory name
            return os.path.dirname(last)

        timestamp, size = self._get_date_and_size(self.zipinfo[zip_path])

        if not WRITE_SUPPORT:
            raise IOError('"os.rename" and "os.unlink" are not supported '
                          'on this platform')
        try:

            real_path = manager.get_cache_path(
                self.egg_name, self._parts(zip_path)
            )

            if self._is_current(real_path, zip_path):
                return real_path

            outf, tmpnam = _mkstemp(
                ".$extract",
                dir=os.path.dirname(real_path),
            )
            os.write(outf, self.loader.get_data(zip_path))
            os.close(outf)
            utime(tmpnam, (timestamp, timestamp))
            manager.postprocess(tmpnam, real_path)

            try:
                rename(tmpnam, real_path)

            except os.error:
                if os.path.isfile(real_path):
                    if self._is_current(real_path, zip_path):
                        # the file became current since it was checked above,
                        #  so proceed.
                        return real_path
                    # Windows, del old file and retry
                    elif os.name == 'nt':
                        unlink(real_path)
                        rename(tmpnam, real_path)
                        return real_path
                raise

        except os.error:
            # report a user-friendly error
            manager.extraction_error()

        return real_path

    def _is_current(self, file_path, zip_path):
        """
        Return True if the file_path is current for this zip_path
        """
        timestamp, size = self._get_date_and_size(self.zipinfo[zip_path])
        if not os.path.isfile(file_path):
            return False
        stat = os.stat(file_path)
        if stat.st_size != size or stat.st_mtime != timestamp:
            return False
        # check that the contents match
        zip_contents = self.loader.get_data(zip_path)
        with open(file_path, 'rb') as f:
            file_contents = f.read()
        return zip_contents == file_contents

    def _get_eager_resources(self):
        if self.eagers is None:
            eagers = []
            for name in ('native_libs.txt', 'eager_resources.txt'):
                if self.has_metadata(name):
                    eagers.extend(self.get_metadata_lines(name))
            self.eagers = eagers
        return self.eagers

    def _index(self):
        try:
            return self._dirindex
        except AttributeError:
            ind = {}
            for path in self.zipinfo:
                parts = path.split(os.sep)
                while parts:
                    parent = os.sep.join(parts[:-1])
                    if parent in ind:
                        ind[parent].append(parts[-1])
                        break
                    else:
                        ind[parent] = [parts.pop()]
            self._dirindex = ind
            return ind

    def _has(self, fspath):
        zip_path = self._zipinfo_name(fspath)
        return zip_path in self.zipinfo or zip_path in self._index()

    def _isdir(self, fspath):
        return self._zipinfo_name(fspath) in self._index()

    def _listdir(self, fspath):
        return list(self._index().get(self._zipinfo_name(fspath), ()))

    def _eager_to_zip(self, resource_name):
        return self._zipinfo_name(self._fn(self.egg_root, resource_name))

    def _resource_to_zip(self, resource_name):
        return self._zipinfo_name(self._fn(self.module_path, resource_name))


register_loader_type(zipimport.zipimporter, ZipProvider)


class FileMetadata(EmptyProvider):
    """Metadata handler for standalone PKG-INFO files

    Usage::

        metadata = FileMetadata("/path/to/PKG-INFO")

    This provider rejects all data and metadata requests except for PKG-INFO,
    which is treated as existing, and will be the contents of the file at
    the provided location.
    """

    def __init__(self, path):
        self.path = path

    def _get_metadata_path(self, name):
        return self.path

    def has_metadata(self, name):
        return name == 'PKG-INFO' and os.path.isfile(self.path)

    def get_metadata(self, name):
        if name != 'PKG-INFO':
            raise KeyError("No metadata except PKG-INFO is available")

        with io.open(self.path, encoding='utf-8', errors="replace") as f:
            metadata = f.read()
        self._warn_on_replacement(metadata)
        return metadata

    def _warn_on_replacement(self, metadata):
        replacement_char = '�'
        if replacement_char in metadata:
            tmpl = "{self.path} could not be properly decoded in UTF-8"
            msg = tmpl.format(**locals())
            warnings.warn(msg)

    def get_metadata_lines(self, name):
        return yield_lines(self.get_metadata(name))


class PathMetadata(DefaultProvider):
    """Metadata provider for egg directories

    Usage::

        # Development eggs:

        egg_info = "/path/to/PackageName.egg-info"
        base_dir = os.path.dirname(egg_info)
        metadata = PathMetadata(base_dir, egg_info)
        dist_name = os.path.splitext(os.path.basename(egg_info))[0]
        dist = Distribution(basedir, project_name=dist_name, metadata=metadata)

        # Unpacked egg directories:

        egg_path = "/path/to/PackageName-ver-pyver-etc.egg"
        metadata = PathMetadata(egg_path, os.path.join(egg_path,'EGG-INFO'))
        dist = Distribution.from_filename(egg_path, metadata=metadata)
    """

    def __init__(self, path, egg_info):
        self.module_path = path
        self.egg_info = egg_info


class EggMetadata(ZipProvider):
    """Metadata provider for .egg files"""

    def __init__(self, importer):
        """Create a metadata provider from a zipimporter"""

        self.zip_pre = importer.archive + os.sep
        self.loader = importer
        if importer.prefix:
            self.module_path = os.path.join(importer.archive, importer.prefix)
        else:
            self.module_path = importer.archive
        self._setup_prefix()


_declare_state('dict', _distribution_finders={})


def register_finder(importer_type, distribution_finder):
    """Register `distribution_finder` to find distributions in sys.path items

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `distribution_finder` is a callable that, passed a path
    item and the importer instance, yields ``Distribution`` instances found on
    that path item.  See ``pkg_resources.find_on_path`` for an example."""
    _distribution_finders[importer_type] = distribution_finder


def find_distributions(path_item, only=False):
    """Yield distributions accessible via `path_item`"""
    importer = get_importer(path_item)
    finder = _find_adapter(_distribution_finders, importer)
    return finder(importer, path_item, only)


def find_eggs_in_zip(importer, path_item, only=False):
    """
    Find eggs in zip files; possibly multiple nested eggs.
    """
    if importer.archive.endswith('.whl'):
        # wheels are not supported with this finder
        # they don't have PKG-INFO metadata, and won't ever contain eggs
        return
    metadata = EggMetadata(importer)
    if metadata.has_metadata('PKG-INFO'):
        yield Distribution.from_filename(path_item, metadata=metadata)
    if only:
        # don't yield nested distros
        return
    for subitem in metadata.resource_listdir(''):
        if _is_egg_path(subitem):
            subpath = os.path.join(path_item, subitem)
            dists = find_eggs_in_zip(zipimport.zipimporter(subpath), subpath)
            for dist in dists:
                yield dist
        elif subitem.lower().endswith(('.dist-info', '.egg-info')):
            subpath = os.path.join(path_item, subitem)
            submeta = EggMetadata(zipimport.zipimporter(subpath))
            submeta.egg_info = subpath
            yield Distribution.from_location(path_item, subitem, submeta)


register_finder(zipimport.zipimporter, find_eggs_in_zip)


def find_nothing(importer, path_item, only=False):
    return ()


register_finder(object, find_nothing)


def _by_version_descending(names):
    """
    Given a list of filenames, return them in descending order
    by version number.

    >>> names = 'bar', 'foo', 'Python-2.7.10.egg', 'Python-2.7.2.egg'
    >>> _by_version_descending(names)
    ['Python-2.7.10.egg', 'Python-2.7.2.egg', 'foo', 'bar']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.egg', 'Setuptools-1.2.3b1.egg']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.post1.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.post1.egg', 'Setuptools-1.2.3b1.egg']
    """
    def _by_version(name):
        """
        Parse each component of the filename
        """
        name, ext = os.path.splitext(name)
        parts = itertools.chain(name.split('-'), [ext])
        return [packaging.version.parse(part) for part in parts]

    return sorted(names, key=_by_version, reverse=True)


def find_on_path(importer, path_item, only=False):
    """Yield distributions accessible on a sys.path directory"""
    path_item = _normalize_cached(path_item)

    if _is_unpacked_egg(path_item):
        yield Distribution.from_filename(
            path_item, metadata=PathMetadata(
                path_item, os.path.join(path_item, 'EGG-INFO')
            )
        )
        return

    entries = (
        os.path.join(path_item, child)
        for child in safe_listdir(path_item)
    )

    # for performance, before sorting by version,
    # screen entries for only those that will yield
    # distributions
    filtered = (
        entry
        for entry in entries
        if dist_factory(path_item, entry, only)
    )

    # scan for .egg and .egg-info in directory
    path_item_entries = _by_version_descending(filtered)
    for entry in path_item_entries:
        fullpath = os.path.join(path_item, entry)
        factory = dist_factory(path_item, entry, only)
        for dist in factory(fullpath):
            yield dist


def dist_factory(path_item, entry, only):
    """Return a dist_factory for the given entry."""
    lower = entry.lower()
    is_egg_info = lower.endswith('.egg-info')
    is_dist_info = (
        lower.endswith('.dist-info') and
        os.path.isdir(os.path.join(path_item, entry))
    )
    is_meta = is_egg_info or is_dist_info
    return (
        distributions_from_metadata
        if is_meta else
        find_distributions
        if not only and _is_egg_path(entry) else
        resolve_egg_link
        if not only and lower.endswith('.egg-link') else
        NoDists()
    )


class NoDists:
    """
    >>> bool(NoDists())
    False

    >>> list(NoDists()('anything'))
    []
    """
    def __bool__(self):
        return False

    def __call__(self, fullpath):
        return iter(())


def safe_listdir(path):
    """
    Attempt to list contents of path, but suppress some exceptions.
    """
    try:
        return os.listdir(path)
    except (PermissionError, NotADirectoryError):
        pass
    except OSError as e:
        # Ignore the directory if does not exist, not a directory or
        # permission denied
        if e.errno not in (errno.ENOTDIR, errno.EACCES, errno.ENOENT):
            raise
    return ()


def distributions_from_metadata(path):
    root = os.path.dirname(path)
    if os.path.isdir(path):
        if len(os.listdir(path)) == 0:
            # empty metadata dir; skip
            return
        metadata = PathMetadata(root, path)
    else:
        metadata = FileMetadata(path)
    entry = os.path.basename(path)
    yield Distribution.from_location(
        root, entry, metadata, precedence=DEVELOP_DIST,
    )


def non_empty_lines(path):
    """
    Yield non-empty lines from file at path
    """
    with open(path) as f:
        for line in f:
            line = line.strip()
            if line:
                yield line


def resolve_egg_link(path):
    """
    Given a path to an .egg-link, resolve distributions
    present in the referenced path.
    """
    referenced_paths = non_empty_lines(path)
    resolved_paths = (
        os.path.join(os.path.dirname(path), ref)
        for ref in referenced_paths
    )
    dist_groups = map(find_distributions, resolved_paths)
    return next(dist_groups, ())


register_finder(pkgutil.ImpImporter, find_on_path)

if hasattr(importlib_machinery, 'FileFinder'):
    register_finder(importlib_machinery.FileFinder, find_on_path)

_declare_state('dict', _namespace_handlers={})
_declare_state('dict', _namespace_packages={})


def register_namespace_handler(importer_type, namespace_handler):
    """Register `namespace_handler` to declare namespace packages

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `namespace_handler` is a callable like this::

        def namespace_handler(importer, path_entry, moduleName, module):
            # return a path_entry to use for child packages

    Namespace handlers are only called if the importer object has already
    agreed that it can handle the relevant path item, and they should only
    return a subpath if the module __path__ does not already contain an
    equivalent subpath.  For an example namespace handler, see
    ``pkg_resources.file_ns_handler``.
    """
    _namespace_handlers[importer_type] = namespace_handler


def _handle_ns(packageName, path_item):
    """Ensure that named package includes a subpath of path_item (if needed)"""

    importer = get_importer(path_item)
    if importer is None:
        return None

    # use find_spec (PEP 451) and fall-back to find_module (PEP 302)
    try:
        loader = importer.find_spec(packageName).loader
    except AttributeError:
        # capture warnings due to #1111
        with warnings.catch_warnings():
            warnings.simplefilter("ignore")
            loader = importer.find_module(packageName)

    if loader is None:
        return None
    module = sys.modules.get(packageName)
    if module is None:
        module = sys.modules[packageName] = types.ModuleType(packageName)
        module.__path__ = []
        _set_parent_ns(packageName)
    elif not hasattr(module, '__path__'):
        raise TypeError("Not a package:", packageName)
    handler = _find_adapter(_namespace_handlers, importer)
    subpath = handler(importer, path_item, packageName, module)
    if subpath is not None:
        path = module.__path__
        path.append(subpath)
        importlib.import_module(packageName)
        _rebuild_mod_path(path, packageName, module)
    return subpath


def _rebuild_mod_path(orig_path, package_name, module):
    """
    Rebuild module.__path__ ensuring that all entries are ordered
    corresponding to their sys.path order
    """
    sys_path = [_normalize_cached(p) for p in sys.path]

    def safe_sys_path_index(entry):
        """
        Workaround for #520 and #513.
        """
        try:
            return sys_path.index(entry)
        except ValueError:
            return float('inf')

    def position_in_sys_path(path):
        """
        Return the ordinal of the path based on its position in sys.path
        """
        path_parts = path.split(os.sep)
        module_parts = package_name.count('.') + 1
        parts = path_parts[:-module_parts]
        return safe_sys_path_index(_normalize_cached(os.sep.join(parts)))

    new_path = sorted(orig_path, key=position_in_sys_path)
    new_path = [_normalize_cached(p) for p in new_path]

    if isinstance(module.__path__, list):
        module.__path__[:] = new_path
    else:
        module.__path__ = new_path


def declare_namespace(packageName):
    """Declare that package 'packageName' is a namespace package"""

    _imp.acquire_lock()
    try:
        if packageName in _namespace_packages:
            return

        path = sys.path
        parent, _, _ = packageName.rpartition('.')

        if parent:
            declare_namespace(parent)
            if parent not in _namespace_packages:
                __import__(parent)
            try:
                path = sys.modules[parent].__path__
            except AttributeError as e:
                raise TypeError("Not a package:", parent) from e

        # Track what packages are namespaces, so when new path items are added,
        # they can be updated
        _namespace_packages.setdefault(parent or None, []).append(packageName)
        _namespace_packages.setdefault(packageName, [])

        for path_item in path:
            # Ensure all the parent's path items are reflected in the child,
            # if they apply
            _handle_ns(packageName, path_item)

    finally:
        _imp.release_lock()


def fixup_namespace_packages(path_item, parent=None):
    """Ensure that previously-declared namespace packages include path_item"""
    _imp.acquire_lock()
    try:
        for package in _namespace_packages.get(parent, ()):
            subpath = _handle_ns(package, path_item)
            if subpath:
                fixup_namespace_packages(subpath, package)
    finally:
        _imp.release_lock()


def file_ns_handler(importer, path_item, packageName, module):
    """Compute an ns-package subpath for a filesystem or zipfile importer"""

    subpath = os.path.join(path_item, packageName.split('.')[-1])
    normalized = _normalize_cached(subpath)
    for item in module.__path__:
        if _normalize_cached(item) == normalized:
            break
    else:
        # Only return the path if it's not already there
        return subpath


register_namespace_handler(pkgutil.ImpImporter, file_ns_handler)
register_namespace_handler(zipimport.zipimporter, file_ns_handler)

if hasattr(importlib_machinery, 'FileFinder'):
    register_namespace_handler(importlib_machinery.FileFinder, file_ns_handler)


def null_ns_handler(importer, path_item, packageName, module):
    return None


register_namespace_handler(object, null_ns_handler)


def normalize_path(filename):
    """Normalize a file/dir name for comparison purposes"""
    return os.path.normcase(os.path.realpath(os.path.normpath(
        _cygwin_patch(filename))))


def _cygwin_patch(filename):  # pragma: nocover
    """
    Contrary to POSIX 2008, on Cygwin, getcwd (3) contains
    symlink components. Using
    os.path.abspath() works around this limitation. A fix in os.getcwd()
    would probably better, in Cygwin even more so, except
    that this seems to be by design...
    """
    return os.path.abspath(filename) if sys.platform == 'cygwin' else filename


def _normalize_cached(filename, _cache={}):
    try:
        return _cache[filename]
    except KeyError:
        _cache[filename] = result = normalize_path(filename)
        return result


def _is_egg_path(path):
    """
    Determine if given path appears to be an egg.
    """
    return _is_zip_egg(path) or _is_unpacked_egg(path)


def _is_zip_egg(path):
    return (
        path.lower().endswith('.egg') and
        os.path.isfile(path) and
        zipfile.is_zipfile(path)
    )


def _is_unpacked_egg(path):
    """
    Determine if given path appears to be an unpacked egg.
    """
    return (
        path.lower().endswith('.egg') and
        os.path.isfile(os.path.join(path, 'EGG-INFO', 'PKG-INFO'))
    )


def _set_parent_ns(packageName):
    parts = packageName.split('.')
    name = parts.pop()
    if parts:
        parent = '.'.join(parts)
        setattr(sys.modules[parent], name, sys.modules[packageName])


def yield_lines(strs):
    """Yield non-empty/non-comment lines of a string or sequence"""
    if isinstance(strs, str):
        for s in strs.splitlines():
            s = s.strip()
            # skip blank lines/comments
            if s and not s.startswith('#'):
                yield s
    else:
        for ss in strs:
            for s in yield_lines(ss):
                yield s


MODULE = re.compile(r"\w+(\.\w+)*$").match
EGG_NAME = re.compile(
    r"""
    (?P<name>[^-]+) (
        -(?P<ver>[^-]+) (
            -py(?P<pyver>[^-]+) (
                -(?P<plat>.+)
            )?
        )?
    )?
    """,
    re.VERBOSE | re.IGNORECASE,
).match


class EntryPoint:
    """Object representing an advertised importable object"""

    def __init__(self, name, module_name, attrs=(), extras=(), dist=None):
        if not MODULE(module_name):
            raise ValueError("Invalid module name", module_name)
        self.name = name
        self.module_name = module_name
        self.attrs = tuple(attrs)
        self.extras = tuple(extras)
        self.dist = dist

    def __str__(self):
        s = "%s = %s" % (self.name, self.module_name)
        if self.attrs:
            s += ':' + '.'.join(self.attrs)
        if self.extras:
            s += ' [%s]' % ','.join(self.extras)
        return s

    def __repr__(self):
        return "EntryPoint.parse(%r)" % str(self)

    def load(self, require=True, *args, **kwargs):
        """
        Require packages for this EntryPoint, then resolve it.
        """
        if not require or args or kwargs:
            warnings.warn(
                "Parameters to load are deprecated.  Call .resolve and "
                ".require separately.",
                PkgResourcesDeprecationWarning,
                stacklevel=2,
            )
        if require:
            self.require(*args, **kwargs)
        return self.resolve()

    def resolve(self):
        """
        Resolve the entry point from its module and attrs.
        """
        module = __import__(self.module_name, fromlist=['__name__'], level=0)
        try:
            return functools.reduce(getattr, self.attrs, module)
        except AttributeError as exc:
            raise ImportError(str(exc)) from exc

    def require(self, env=None, installer=None):
        if self.extras and not self.dist:
            raise UnknownExtra("Can't require() without a distribution", self)

        # Get the requirements for this entry point with all its extras and
        # then resolve them. We have to pass `extras` along when resolving so
        # that the working set knows what extras we want. Otherwise, for
        # dist-info distributions, the working set will assume that the
        # requirements for that extra are purely optional and skip over them.
        reqs = self.dist.requires(self.extras)
        items = working_set.resolve(reqs, env, installer, extras=self.extras)
        list(map(working_set.add, items))

    pattern = re.compile(
        r'\s*'
        r'(?P<name>.+?)\s*'
        r'=\s*'
        r'(?P<module>[\w.]+)\s*'
        r'(:\s*(?P<attr>[\w.]+))?\s*'
        r'(?P<extras>\[.*\])?\s*$'
    )

    @classmethod
    def parse(cls, src, dist=None):
        """Parse a single entry point from string `src`

        Entry point syntax follows the form::

            name = some.module:some.attr [extra1, extra2]

        The entry name and module name are required, but the ``:attrs`` and
        ``[extras]`` parts are optional
        """
        m = cls.pattern.match(src)
        if not m:
            msg = "EntryPoint must be in 'name=module:attrs [extras]' format"
            raise ValueError(msg, src)
        res = m.groupdict()
        extras = cls._parse_extras(res['extras'])
        attrs = res['attr'].split('.') if res['attr'] else ()
        return cls(res['name'], res['module'], attrs, extras, dist)

    @classmethod
    def _parse_extras(cls, extras_spec):
        if not extras_spec:
            return ()
        req = Requirement.parse('x' + extras_spec)
        if req.specs:
            raise ValueError()
        return req.extras

    @classmethod
    def parse_group(cls, group, lines, dist=None):
        """Parse an entry point group"""
        if not MODULE(group):
            raise ValueError("Invalid group name", group)
        this = {}
        for line in yield_lines(lines):
            ep = cls.parse(line, dist)
            if ep.name in this:
                raise ValueError("Duplicate entry point", group, ep.name)
            this[ep.name] = ep
        return this

    @classmethod
    def parse_map(cls, data, dist=None):
        """Parse a map of entry point groups"""
        if isinstance(data, dict):
            data = data.items()
        else:
            data = split_sections(data)
        maps = {}
        for group, lines in data:
            if group is None:
                if not lines:
                    continue
                raise ValueError("Entry points must be listed in groups")
            group = group.strip()
            if group in maps:
                raise ValueError("Duplicate group name", group)
            maps[group] = cls.parse_group(group, lines, dist)
        return maps


def _version_from_file(lines):
    """
    Given an iterable of lines from a Metadata file, return
    the value of the Version field, if present, or None otherwise.
    """
    def is_version_line(line):
        return line.lower().startswith('version:')
    version_lines = filter(is_version_line, lines)
    line = next(iter(version_lines), '')
    _, _, value = line.partition(':')
    return safe_version(value.strip()) or None


class Distribution:
    """Wrap an actual or potential sys.path entry w/metadata"""
    PKG_INFO = 'PKG-INFO'

    def __init__(
            self, location=None, metadata=None, project_name=None,
            version=None, py_version=PY_MAJOR, platform=None,
            precedence=EGG_DIST):
        self.project_name = safe_name(project_name or 'Unknown')
        if version is not None:
            self._version = safe_version(version)
        self.py_version = py_version
        self.platform = platform
        self.location = location
        self.precedence = precedence
        self._provider = metadata or empty_provider

    @classmethod
    def from_location(cls, location, basename, metadata=None, **kw):
        project_name, version, py_version, platform = [None] * 4
        basename, ext = os.path.splitext(basename)
        if ext.lower() in _distributionImpl:
            cls = _distributionImpl[ext.lower()]

            match = EGG_NAME(basename)
            if match:
                project_name, version, py_version, platform = match.group(
                    'name', 'ver', 'pyver', 'plat'
                )
        return cls(
            location, metadata, project_name=project_name, version=version,
            py_version=py_version, platform=platform, **kw
        )._reload_version()

    def _reload_version(self):
        return self

    @property
    def hashcmp(self):
        return (
            self.parsed_version,
            self.precedence,
            self.key,
            self.location,
            self.py_version or '',
            self.platform or '',
        )

    def __hash__(self):
        return hash(self.hashcmp)

    def __lt__(self, other):
        return self.hashcmp < other.hashcmp

    def __le__(self, other):
        return self.hashcmp <= other.hashcmp

    def __gt__(self, other):
        return self.hashcmp > other.hashcmp

    def __ge__(self, other):
        return self.hashcmp >= other.hashcmp

    def __eq__(self, other):
        if not isinstance(other, self.__class__):
            # It's not a Distribution, so they are not equal
            return False
        return self.hashcmp == other.hashcmp

    def __ne__(self, other):
        return not self == other

    # These properties have to be lazy so that we don't have to load any
    # metadata until/unless it's actually needed.  (i.e., some distributions
    # may not know their name or version without loading PKG-INFO)

    @property
    def key(self):
        try:
            return self._key
        except AttributeError:
            self._key = key = self.project_name.lower()
            return key

    @property
    def parsed_version(self):
        if not hasattr(self, "_parsed_version"):
            self._parsed_version = parse_version(self.version)

        return self._parsed_version

    def _warn_legacy_version(self):
        LV = packaging.version.LegacyVersion
        is_legacy = isinstance(self._parsed_version, LV)
        if not is_legacy:
            return

        # While an empty version is technically a legacy version and
        # is not a valid PEP 440 version, it's also unlikely to
        # actually come from someone and instead it is more likely that
        # it comes from setuptools attempting to parse a filename and
        # including it in the list. So for that we'll gate this warning
        # on if the version is anything at all or not.
        if not self.version:
            return

        tmpl = textwrap.dedent("""
            '{project_name} ({version})' is being parsed as a legacy,
            non PEP 440,
            version. You may find odd behavior and sort order.
            In particular it will be sorted as less than 0.0. It
            is recommended to migrate to PEP 440 compatible
            versions.
            """).strip().replace('\n', ' ')

        warnings.warn(tmpl.format(**vars(self)), PEP440Warning)

    @property
    def version(self):
        try:
            return self._version
        except AttributeError as e:
            version = self._get_version()
            if version is None:
                path = self._get_metadata_path_for_display(self.PKG_INFO)
                msg = (
                    "Missing 'Version:' header and/or {} file at path: {}"
                ).format(self.PKG_INFO, path)
                raise ValueError(msg, self) from e

            return version

    @property
    def _dep_map(self):
        """
        A map of extra to its list of (direct) requirements
        for this distribution, including the null extra.
        """
        try:
            return self.__dep_map
        except AttributeError:
            self.__dep_map = self._filter_extras(self._build_dep_map())
        return self.__dep_map

    @staticmethod
    def _filter_extras(dm):
        """
        Given a mapping of extras to dependencies, strip off
        environment markers and filter out any dependencies
        not matching the markers.
        """
        for extra in list(filter(None, dm)):
            new_extra = extra
            reqs = dm.pop(extra)
            new_extra, _, marker = extra.partition(':')
            fails_marker = marker and (
                invalid_marker(marker)
                or not evaluate_marker(marker)
            )
            if fails_marker:
                reqs = []
            new_extra = safe_extra(new_extra) or None

            dm.setdefault(new_extra, []).extend(reqs)
        return dm

    def _build_dep_map(self):
        dm = {}
        for name in 'requires.txt', 'depends.txt':
            for extra, reqs in split_sections(self._get_metadata(name)):
                dm.setdefault(extra, []).extend(parse_requirements(reqs))
        return dm

    def requires(self, extras=()):
        """List of Requirements needed for this distro if `extras` are used"""
        dm = self._dep_map
        deps = []
        deps.extend(dm.get(None, ()))
        for ext in extras:
            try:
                deps.extend(dm[safe_extra(ext)])
            except KeyError as e:
                raise UnknownExtra(
                    "%s has no such extra feature %r" % (self, ext)
                ) from e
        return deps

    def _get_metadata_path_for_display(self, name):
        """
        Return the path to the given metadata file, if available.
        """
        try:
            # We need to access _get_metadata_path() on the provider object
            # directly rather than through this class's __getattr__()
            # since _get_metadata_path() is marked private.
            path = self._provider._get_metadata_path(name)

        # Handle exceptions e.g. in case the distribution's metadata
        # provider doesn't support _get_metadata_path().
        except Exception:
            return '[could not detect]'

        return path

    def _get_metadata(self, name):
        if self.has_metadata(name):
            for line in self.get_metadata_lines(name):
                yield line

    def _get_version(self):
        lines = self._get_metadata(self.PKG_INFO)
        version = _version_from_file(lines)

        return version

    def activate(self, path=None, replace=False):
        """Ensure distribution is importable on `path` (default=sys.path)"""
        if path is None:
            path = sys.path
        self.insert_on(path, replace=replace)
        if path is sys.path:
            fixup_namespace_packages(self.location)
            for pkg in self._get_metadata('namespace_packages.txt'):
                if pkg in sys.modules:
                    declare_namespace(pkg)

    def egg_name(self):
        """Return what this distribution's standard .egg filename should be"""
        filename = "%s-%s-py%s" % (
            to_filename(self.project_name), to_filename(self.version),
            self.py_version or PY_MAJOR
        )

        if self.platform:
            filename += '-' + self.platform
        return filename

    def __repr__(self):
        if self.location:
            return "%s (%s)" % (self, self.location)
        else:
            return str(self)

    def __str__(self):
        try:
            version = getattr(self, 'version', None)
        except ValueError:
            version = None
        version = version or "[unknown version]"
        return "%s %s" % (self.project_name, version)

    def __getattr__(self, attr):
        """Delegate all unrecognized public attributes to .metadata provider"""
        if attr.startswith('_'):
            raise AttributeError(attr)
        return getattr(self._provider, attr)

    def __dir__(self):
        return list(
            set(super(Distribution, self).__dir__())
            | set(
                attr for attr in self._provider.__dir__()
                if not attr.startswith('_')
            )
        )

    @classmethod
    def from_filename(cls, filename, metadata=None, **kw):
        return cls.from_location(
            _normalize_cached(filename), os.path.basename(filename), metadata,
            **kw
        )

    def as_requirement(self):
        """Return a ``Requirement`` that matches this distribution exactly"""
        if isinstance(self.parsed_version, packaging.version.Version):
            spec = "%s==%s" % (self.project_name, self.parsed_version)
        else:
            spec = "%s===%s" % (self.project_name, self.parsed_version)

        return Requirement.parse(spec)

    def load_entry_point(self, group, name):
        """Return the `name` entry point of `group` or raise ImportError"""
        ep = self.get_entry_info(group, name)
        if ep is None:
            raise ImportError("Entry point %r not found" % ((group, name),))
        return ep.load()

    def get_entry_map(self, group=None):
        """Return the entry point map for `group`, or the full entry map"""
        try:
            ep_map = self._ep_map
        except AttributeError:
            ep_map = self._ep_map = EntryPoint.parse_map(
                self._get_metadata('entry_points.txt'), self
            )
        if group is not None:
            return ep_map.get(group, {})
        return ep_map

    def get_entry_info(self, group, name):
        """Return the EntryPoint object for `group`+`name`, or ``None``"""
        return self.get_entry_map(group).get(name)

    # FIXME: 'Distribution.insert_on' is too complex (13)
    def insert_on(self, path, loc=None, replace=False):  # noqa: C901
        """Ensure self.location is on path

        If replace=False (default):
            - If location is already in path anywhere, do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent.
              - Else: add to the end of path.
        If replace=True:
            - If location is already on path anywhere (not eggs)
              or higher priority than its parent (eggs)
              do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent,
                removing any lower-priority entries.
              - Else: add it to the front of path.
        """

        loc = loc or self.location
        if not loc:
            return

        nloc = _normalize_cached(loc)
        bdir = os.path.dirname(nloc)
        npath = [(p and _normalize_cached(p) or p) for p in path]

        for p, item in enumerate(npath):
            if item == nloc:
                if replace:
                    break
                else:
                    # don't modify path (even removing duplicates) if
                    # found and not replace
                    return
            elif item == bdir and self.precedence == EGG_DIST:
                # if it's an .egg, give it precedence over its directory
                # UNLESS it's already been added to sys.path and replace=False
                if (not replace) and nloc in npath[p:]:
                    return
                if path is sys.path:
                    self.check_version_conflict()
                path.insert(p, loc)
                npath.insert(p, nloc)
                break
        else:
            if path is sys.path:
                self.check_version_conflict()
            if replace:
                path.insert(0, loc)
            else:
                path.append(loc)
            return

        # p is the spot where we found or inserted loc; now remove duplicates
        while True:
            try:
                np = npath.index(nloc, p + 1)
            except ValueError:
                break
            else:
                del npath[np], path[np]
                # ha!
                p = np

        return

    def check_version_conflict(self):
        if self.key == 'setuptools':
            # ignore the inevitable setuptools self-conflicts  :(
            return

        nsp = dict.fromkeys(self._get_metadata('namespace_packages.txt'))
        loc = normalize_path(self.location)
        for modname in self._get_metadata('top_level.txt'):
            if (modname not in sys.modules or modname in nsp
                    or modname in _namespace_packages):
                continue
            if modname in ('pkg_resources', 'setuptools', 'site'):
                continue
            fn = getattr(sys.modules[modname], '__file__', None)
            if fn and (normalize_path(fn).startswith(loc) or
                       fn.startswith(self.location)):
                continue
            issue_warning(
                "Module %s was already imported from %s, but %s is being added"
                " to sys.path" % (modname, fn, self.location),
            )

    def has_version(self):
        try:
            self.version
        except ValueError:
            issue_warning("Unbuilt egg for " + repr(self))
            return False
        return True

    def clone(self, **kw):
        """Copy this distribution, substituting in any changed keyword args"""
        names = 'project_name version py_version platform location precedence'
        for attr in names.split():
            kw.setdefault(attr, getattr(self, attr, None))
        kw.setdefault('metadata', self._provider)
        return self.__class__(**kw)

    @property
    def extras(self):
        return [dep for dep in self._dep_map if dep]


class EggInfoDistribution(Distribution):
    def _reload_version(self):
        """
        Packages installed by distutils (e.g. numpy or scipy),
        which uses an old safe_version, and so
        their version numbers can get mangled when
        converted to filenames (e.g., 1.11.0.dev0+2329eae to
        1.11.0.dev0_2329eae). These distributions will not be
        parsed properly
        downstream by Distribution and safe_version, so
        take an extra step and try to get the version number from
        the metadata file itself instead of the filename.
        """
        md_version = self._get_version()
        if md_version:
            self._version = md_version
        return self


class DistInfoDistribution(Distribution):
    """
    Wrap an actual or potential sys.path entry
    w/metadata, .dist-info style.
    """
    PKG_INFO = 'METADATA'
    EQEQ = re.compile(r"([\(,])\s*(\d.*?)\s*([,\)])")

    @property
    def _parsed_pkg_info(self):
        """Parse and cache metadata"""
        try:
            return self._pkg_info
        except AttributeError:
            metadata = self.get_metadata(self.PKG_INFO)
            self._pkg_info = email.parser.Parser().parsestr(metadata)
            return self._pkg_info

    @property
    def _dep_map(self):
        try:
            return self.__dep_map
        except AttributeError:
            self.__dep_map = self._compute_dependencies()
            return self.__dep_map

    def _compute_dependencies(self):
        """Recompute this distribution's dependencies."""
        dm = self.__dep_map = {None: []}

        reqs = []
        # Including any condition expressions
        for req in self._parsed_pkg_info.get_all('Requires-Dist') or []:
            reqs.extend(parse_requirements(req))

        def reqs_for_extra(extra):
            for req in reqs:
                if not req.marker or req.marker.evaluate({'extra': extra}):
                    yield req

        common = frozenset(reqs_for_extra(None))
        dm[None].extend(common)

        for extra in self._parsed_pkg_info.get_all('Provides-Extra') or []:
            s_extra = safe_extra(extra.strip())
            dm[s_extra] = list(frozenset(reqs_for_extra(extra)) - common)

        return dm


_distributionImpl = {
    '.egg': Distribution,
    '.egg-info': EggInfoDistribution,
    '.dist-info': DistInfoDistribution,
}


def issue_warning(*args, **kw):
    level = 1
    g = globals()
    try:
        # find the first stack frame that is *not* code in
        # the pkg_resources module, to use for the warning
        while sys._getframe(level).f_globals is g:
            level += 1
    except ValueError:
        pass
    warnings.warn(stacklevel=level + 1, *args, **kw)


def parse_requirements(strs):
    """Yield ``Requirement`` objects for each specification in `strs`

    `strs` must be a string, or a (possibly-nested) iterable thereof.
    """
    # create a steppable iterator, so we can handle \-continuations
    lines = iter(yield_lines(strs))

    for line in lines:
        # Drop comments -- a hash without a space may be in a URL.
        if ' #' in line:
            line = line[:line.find(' #')]
        # If there is a line continuation, drop it, and append the next line.
        if line.endswith('\\'):
            line = line[:-2].strip()
            try:
                line += next(lines)
            except StopIteration:
                return
        yield Requirement(line)


class RequirementParseError(packaging.requirements.InvalidRequirement):
    "Compatibility wrapper for InvalidRequirement"


class Requirement(packaging.requirements.Requirement):
    def __init__(self, requirement_string):
        """DO NOT CALL THIS UNDOCUMENTED METHOD; use Requirement.parse()!"""
        super(Requirement, self).__init__(requirement_string)
        self.unsafe_name = self.name
        project_name = safe_name(self.name)
        self.project_name, self.key = project_name, project_name.lower()
        self.specs = [
            (spec.operator, spec.version) for spec in self.specifier]
        self.extras = tuple(map(safe_extra, self.extras))
        self.hashCmp = (
            self.key,
            self.url,
            self.specifier,
            frozenset(self.extras),
            str(self.marker) if self.marker else None,
        )
        self.__hash = hash(self.hashCmp)

    def __eq__(self, other):
        return (
            isinstance(other, Requirement) and
            self.hashCmp == other.hashCmp
        )

    def __ne__(self, other):
        return not self == other

    def __contains__(self, item):
        if isinstance(item, Distribution):
            if item.key != self.key:
                return False

            item = item.version

        # Allow prereleases always in order to match the previous behavior of
        # this method. In the future this should be smarter and follow PEP 440
        # more accurately.
        return self.specifier.contains(item, prereleases=True)

    def __hash__(self):
        return self.__hash

    def __repr__(self):
        return "Requirement.parse(%r)" % str(self)

    @staticmethod
    def parse(s):
        req, = parse_requirements(s)
        return req


def _always_object(classes):
    """
    Ensure object appears in the mro even
    for old-style classes.
    """
    if object not in classes:
        return classes + (object,)
    return classes


def _find_adapter(registry, ob):
    """Return an adapter factory for `ob` from `registry`"""
    types = _always_object(inspect.getmro(getattr(ob, '__class__', type(ob))))
    for t in types:
        if t in registry:
            return registry[t]


def ensure_directory(path):
    """Ensure that the parent directory of `path` exists"""
    dirname = os.path.dirname(path)
    os.makedirs(dirname, exist_ok=True)


def _bypass_ensure_directory(path):
    """Sandbox-bypassing version of ensure_directory()"""
    if not WRITE_SUPPORT:
        raise IOError('"os.mkdir" not supported on this platform.')
    dirname, filename = split(path)
    if dirname and filename and not isdir(dirname):
        _bypass_ensure_directory(dirname)
        try:
            mkdir(dirname, 0o755)
        except FileExistsError:
            pass


def split_sections(s):
    """Split a string or iterable thereof into (section, content) pairs

    Each ``section`` is a stripped version of the section header ("[section]")
    and each ``content`` is a list of stripped lines excluding blank lines and
    comment-only lines.  If there are any such lines before the first section
    header, they're returned in a first ``section`` of ``None``.
    """
    section = None
    content = []
    for line in yield_lines(s):
        if line.startswith("["):
            if line.endswith("]"):
                if section or content:
                    yield section, content
                section = line[1:-1].strip()
                content = []
            else:
                raise ValueError("Invalid section heading", line)
        else:
            content.append(line)

    # wrap up last segment
    yield section, content


def _mkstemp(*args, **kw):
    old_open = os.open
    try:
        # temporarily bypass sandboxing
        os.open = os_open
        return tempfile.mkstemp(*args, **kw)
    finally:
        # and then put it back
        os.open = old_open


# Silence the PEP440Warning by default, so that end users don't get hit by it
# randomly just because they use pkg_resources. We want to append the rule
# because we want earlier uses of filterwarnings to take precedence over this
# one.
warnings.filterwarnings("ignore", category=PEP440Warning, append=True)


# from jaraco.functools 1.3
def _call_aside(f, *args, **kwargs):
    f(*args, **kwargs)
    return f


@_call_aside
def _initialize(g=globals()):
    "Set up global resource manager (deliberately not state-saved)"
    manager = ResourceManager()
    g['_manager'] = manager
    g.update(
        (name, getattr(manager, name))
        for name in dir(manager)
        if not name.startswith('_')
    )


@_call_aside
def _initialize_master_working_set():
    """
    Prepare the master working set and make the ``require()``
    API available.

    This function has explicit effects on the global state
    of pkg_resources. It is intended to be invoked once at
    the initialization of this module.

    Invocation by other packages is unsupported and done
    at their own risk.
    """
    working_set = WorkingSet._build_master()
    _declare_state('object', working_set=working_set)

    require = working_set.require
    iter_entry_points = working_set.iter_entry_points
    add_activation_listener = working_set.subscribe
    run_script = working_set.run_script
    # backward compatibility
    run_main = run_script
    # Activate all distributions already on sys.path with replace=False and
    # ensure that all distributions added to the working set in the future
    # (e.g. by calling ``require()``) will get activated as well,
    # with higher priority (replace=True).
    tuple(
        dist.activate(replace=False)
        for dist in working_set
    )
    add_activation_listener(
        lambda dist: dist.activate(replace=True),
        existing=False,
    )
    working_set.entries = []
    # match order
    list(map(working_set.add_entry, sys.path))
    globals().update(locals())


class PkgResourcesDeprecationWarning(Warning):
    """
    Base class for warning about deprecations in ``pkg_resources``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    """
site-packages/setuptools-58.3.0.dist-info/WHEEL000064400000000134150564745010014725 0ustar00Wheel-Version: 1.0
Generator: bdist_wheel (0.37.0)
Root-Is-Purelib: true
Tag: py3-none-any

site-packages/setuptools-58.3.0.dist-info/LICENSE000064400000002032150564745010015142 0ustar00Copyright Jason R. Coombs

Permission is hereby granted, free of charge, to any person obtaining a copy
of this software and associated documentation files (the "Software"), to
deal in the Software without restriction, including without limitation the
rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
sell copies of the Software, and to permit persons to whom the Software is
furnished to do so, subject to the following conditions:

The above copyright notice and this permission notice shall be included in
all copies or substantial portions of the Software.

THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
IN THE SOFTWARE.
site-packages/setuptools-58.3.0.dist-info/direct_url.json000064400000000165150564745010017171 0ustar00{"archive_info": {}, "url": "file:///builddir/build/BUILD/setuptools-58.3.0/dist/setuptools-58.3.0-py3-none-any.whl"}site-packages/setuptools-58.3.0.dist-info/RECORD000064400000054047150564745010015053 0ustar00_distutils_hack/__init__.py,sha256=X3RUiA6KBPoEmco_CjACyltyQbFRGVUpZRAbSkPGwMs,3688
_distutils_hack/__pycache__/__init__.cpython-39.pyc,,
_distutils_hack/__pycache__/override.cpython-39.pyc,,
_distutils_hack/override.py,sha256=Eu_s-NF6VIZ4Cqd0tbbA5wtWky2IZPNd8et6GLt1mzo,44
distutils-precedence.pth,sha256=fqf_7z_ioRfuEsaO1lU2F_DX_S8FkCV8JcSElZo7c3M,152
pkg_resources/__init__.py,sha256=P3PNN3_m8JJrYMp-i-Sq-3rhK5vuViqqjn1UXKHfe7Q,108202
pkg_resources/__pycache__/__init__.cpython-39.pyc,,
pkg_resources/_vendor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
pkg_resources/_vendor/__pycache__/__init__.cpython-39.pyc,,
pkg_resources/_vendor/__pycache__/appdirs.cpython-39.pyc,,
pkg_resources/_vendor/__pycache__/pyparsing.cpython-39.pyc,,
pkg_resources/_vendor/appdirs.py,sha256=MievUEuv3l_mQISH5SF0shDk_BNhHHzYiAPrT3ITN4I,24701
pkg_resources/_vendor/packaging/__about__.py,sha256=PNMsaZn4UcCHyubgROH1bl6CluduPjI5kFrSp_Zgklo,736
pkg_resources/_vendor/packaging/__init__.py,sha256=6enbp5XgRfjBjsI9-bn00HjHf5TH21PDMOKkJW8xw-w,562
pkg_resources/_vendor/packaging/__pycache__/__about__.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/__init__.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/_compat.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/_structures.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/_typing.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/markers.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/requirements.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/tags.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/utils.cpython-39.pyc,,
pkg_resources/_vendor/packaging/__pycache__/version.cpython-39.pyc,,
pkg_resources/_vendor/packaging/_compat.py,sha256=MXdsGpSE_W-ZrHoC87andI4LV2FAwU7HLL-eHe_CjhU,1128
pkg_resources/_vendor/packaging/_structures.py,sha256=ozkCX8Q8f2qE1Eic3YiQ4buDVfgz2iYevY9e7R2y3iY,2022
pkg_resources/_vendor/packaging/_typing.py,sha256=x59EhQ57TMT-kTRyLZV25HZvYGGwbucTo6iKh_O0tMw,1812
pkg_resources/_vendor/packaging/markers.py,sha256=YSntQkMnKyw1_FG6oRNNnGxLL6bAxcGXOtuFE-YTS3k,9518
pkg_resources/_vendor/packaging/requirements.py,sha256=R8K4H4xX_iD4LvpGw1U3ouuPbGN-wzsFgD7brhAM71Y,4929
pkg_resources/_vendor/packaging/specifiers.py,sha256=uYp9l13F0LcknS6d4N60ytiBgFmIhKideOq9AnsxTco,31944
pkg_resources/_vendor/packaging/tags.py,sha256=NKMS37Zo_nWrZxgsD6zbXsXgc9edn9m160cBiLmHJdE,24067
pkg_resources/_vendor/packaging/utils.py,sha256=RShlvnjO2CtYSD8uri32frMMFMTmB-3ihsq1-ghzLEw,1811
pkg_resources/_vendor/packaging/version.py,sha256=Cnbm-OO9D_qd8ZTFxzFcjSavexSYFZmyeaoPvMsjgPc,15470
pkg_resources/_vendor/pyparsing.py,sha256=mahtkgcp3grNAD0re_9R0DLvBnvjzpeLwgJqT-3H1CE,232056
pkg_resources/extern/__init__.py,sha256=3PixaT9Tzzd4NoyV6CVhGd7S_9Z-U5yvMWAftZKvC6k,2362
pkg_resources/extern/__pycache__/__init__.cpython-39.pyc,,
pkg_resources/tests/data/my-test-package-source/__pycache__/setup.cpython-39.pyc,,
pkg_resources/tests/data/my-test-package-source/setup.py,sha256=Mrezl3nqxkYkjCYpIxmjhhg4AR8hgi4QZdEYmk-I7R8,104
setuptools-58.3.0.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4
setuptools-58.3.0.dist-info/LICENSE,sha256=2z8CRrH5J48VhFuZ_sR4uLUG63ZIeZNyL4xuJUKF-vg,1050
setuptools-58.3.0.dist-info/METADATA,sha256=uI89jlY7w7UjYJQzF3Bc4W3jK7VofkekiCtuLgOnNQY,4852
setuptools-58.3.0.dist-info/RECORD,,
setuptools-58.3.0.dist-info/REQUESTED,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
setuptools-58.3.0.dist-info/WHEEL,sha256=ewwEueio1C2XeHTvT17n8dZUJgOvyCWCt0WVNLClP9o,92
setuptools-58.3.0.dist-info/direct_url.json,sha256=vNpU8J6cPCxDEDob5K4oYQk3R6QhUvQT1f4Uikr2KuE,117
setuptools-58.3.0.dist-info/entry_points.txt,sha256=wpnhLrbtyk4hZ1qCCw48cCSxoQPzULMhIuaFqsB7GxQ,2636
setuptools-58.3.0.dist-info/top_level.txt,sha256=d9yL39v_W7qmKDDSH6sT4bE0j_Ls1M3P161OGgdsm4g,41
setuptools/__init__.py,sha256=h06afcaxHUhFjHFiEIoJhv6Sr0bTy9Aku4Rd8_82iys,7670
setuptools/__pycache__/__init__.cpython-39.pyc,,
setuptools/__pycache__/_deprecation_warning.cpython-39.pyc,,
setuptools/__pycache__/_imp.cpython-39.pyc,,
setuptools/__pycache__/archive_util.cpython-39.pyc,,
setuptools/__pycache__/build_meta.cpython-39.pyc,,
setuptools/__pycache__/config.cpython-39.pyc,,
setuptools/__pycache__/dep_util.cpython-39.pyc,,
setuptools/__pycache__/depends.cpython-39.pyc,,
setuptools/__pycache__/dist.cpython-39.pyc,,
setuptools/__pycache__/errors.cpython-39.pyc,,
setuptools/__pycache__/extension.cpython-39.pyc,,
setuptools/__pycache__/glob.cpython-39.pyc,,
setuptools/__pycache__/installer.cpython-39.pyc,,
setuptools/__pycache__/launch.cpython-39.pyc,,
setuptools/__pycache__/monkey.cpython-39.pyc,,
setuptools/__pycache__/msvc.cpython-39.pyc,,
setuptools/__pycache__/namespaces.cpython-39.pyc,,
setuptools/__pycache__/package_index.cpython-39.pyc,,
setuptools/__pycache__/py34compat.cpython-39.pyc,,
setuptools/__pycache__/sandbox.cpython-39.pyc,,
setuptools/__pycache__/unicode_utils.cpython-39.pyc,,
setuptools/__pycache__/version.cpython-39.pyc,,
setuptools/__pycache__/wheel.cpython-39.pyc,,
setuptools/__pycache__/windows_support.cpython-39.pyc,,
setuptools/_deprecation_warning.py,sha256=jU9-dtfv6cKmtQJOXN8nP1mm7gONw5kKEtiPtbwnZyI,218
setuptools/_distutils/__init__.py,sha256=lpQAphR_7uhWC2fbSEps4Ja9W4YwezN_IX_LJEt3khU,250
setuptools/_distutils/__pycache__/__init__.cpython-39.pyc,,
setuptools/_distutils/__pycache__/_msvccompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/archive_util.cpython-39.pyc,,
setuptools/_distutils/__pycache__/bcppcompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/ccompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/cmd.cpython-39.pyc,,
setuptools/_distutils/__pycache__/config.cpython-39.pyc,,
setuptools/_distutils/__pycache__/core.cpython-39.pyc,,
setuptools/_distutils/__pycache__/cygwinccompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/debug.cpython-39.pyc,,
setuptools/_distutils/__pycache__/dep_util.cpython-39.pyc,,
setuptools/_distutils/__pycache__/dir_util.cpython-39.pyc,,
setuptools/_distutils/__pycache__/dist.cpython-39.pyc,,
setuptools/_distutils/__pycache__/errors.cpython-39.pyc,,
setuptools/_distutils/__pycache__/extension.cpython-39.pyc,,
setuptools/_distutils/__pycache__/fancy_getopt.cpython-39.pyc,,
setuptools/_distutils/__pycache__/file_util.cpython-39.pyc,,
setuptools/_distutils/__pycache__/filelist.cpython-39.pyc,,
setuptools/_distutils/__pycache__/log.cpython-39.pyc,,
setuptools/_distutils/__pycache__/msvc9compiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/msvccompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/py35compat.cpython-39.pyc,,
setuptools/_distutils/__pycache__/py38compat.cpython-39.pyc,,
setuptools/_distutils/__pycache__/spawn.cpython-39.pyc,,
setuptools/_distutils/__pycache__/sysconfig.cpython-39.pyc,,
setuptools/_distutils/__pycache__/text_file.cpython-39.pyc,,
setuptools/_distutils/__pycache__/unixccompiler.cpython-39.pyc,,
setuptools/_distutils/__pycache__/util.cpython-39.pyc,,
setuptools/_distutils/__pycache__/version.cpython-39.pyc,,
setuptools/_distutils/__pycache__/versionpredicate.cpython-39.pyc,,
setuptools/_distutils/_msvccompiler.py,sha256=jR0JM5A1JMnZ6xMDicQzhXWgXTVXs1lWAeUexC1z198,20813
setuptools/_distutils/archive_util.py,sha256=qW-uiGwYexTvK5e-iSel_31Dshx-CqTanNPK6snwf98,8572
setuptools/_distutils/bcppcompiler.py,sha256=OJDVpCUmX6H8v_7lV1zifV1fcx92Cr2dhiUh6989UJI,14894
setuptools/_distutils/ccompiler.py,sha256=G2tn9Q3zQ0VUNfW1LM-nrnLt_6OhtiUunugCv85D1PQ,47607
setuptools/_distutils/cmd.py,sha256=eco6LAGUtobLuPafuhmgKgkwRRL_WY8KJ4YeDCHpcls,18079
setuptools/_distutils/command/__init__.py,sha256=2TA-rlNDlzeI-csbWHXFjGD8uOYqALMfyWOhT49nC6g,799
setuptools/_distutils/command/__pycache__/__init__.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/bdist.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/bdist_dumb.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/bdist_msi.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/bdist_rpm.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/bdist_wininst.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/build.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/build_clib.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/build_ext.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/build_py.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/build_scripts.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/check.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/clean.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/config.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install_data.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install_egg_info.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install_headers.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install_lib.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/install_scripts.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/py37compat.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/register.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/sdist.cpython-39.pyc,,
setuptools/_distutils/command/__pycache__/upload.cpython-39.pyc,,
setuptools/_distutils/command/bdist.py,sha256=2z4eudRl_n7m3lG9leL0IYqes4bsm8c0fxfZuiafjMg,5562
setuptools/_distutils/command/bdist_dumb.py,sha256=BTur9jcIppyP7Piavjfsk7YjElqvxeYO2npUyPPOekc,4913
setuptools/_distutils/command/bdist_msi.py,sha256=EVFQYN_X-ExeeP8gmdV9JcINsuUGsLJUz9afMU0Rt8c,35579
setuptools/_distutils/command/bdist_rpm.py,sha256=gjOw22GhDSbcq0bdq25cTb-n6HWWm0bShLQad_mkJ4k,21537
setuptools/_distutils/command/bdist_wininst.py,sha256=iGlaI-VfElHOneeczKHWnSN5a10-7IMcJaXuR1mdS3c,16030
setuptools/_distutils/command/build.py,sha256=1AF-dxN_NlOEyoydBz19AwpeWYPSYCZvOLJSN_PdatY,5773
setuptools/_distutils/command/build_clib.py,sha256=bgVTHh28eLQA2Gkw68amApd_j7qQBX4MTI-zTvAK_J4,8022
setuptools/_distutils/command/build_ext.py,sha256=hX4ksYXRC-Q3fEvLrANIG97mq3twO6ZwkMhfANNK3Yg,31683
setuptools/_distutils/command/build_py.py,sha256=hXesMrH_epNj6K8SUtJdipgEis3EdICKeZ8VWe_ndck,16495
setuptools/_distutils/command/build_scripts.py,sha256=urdn6wPxPMW5dLqpqFkZ8dqaFG1tf9TiAao6U9LCoEI,5963
setuptools/_distutils/command/check.py,sha256=5qDtI75ccZg3sAItQWeaIu8y3FR314O4rr9Smz4HsEo,5637
setuptools/_distutils/command/clean.py,sha256=2TCt47ru4hZZM0RfVfUYj5bbpicpGLP4Qhw5jBtvp9k,2776
setuptools/_distutils/command/config.py,sha256=2aTjww3PwjMB8-ZibCe4P7B-qG1hM1gn_rJXYyxRz6c,13117
setuptools/_distutils/command/install.py,sha256=oaYyzj2vAGb_HKqdFts7rY0gx80W9MrqPQCZpfvGj2k,27534
setuptools/_distutils/command/install_data.py,sha256=YhGOAwh3gJPqF7em5XA0rmpR42z1bLh80ooElzDyUvk,2822
setuptools/_distutils/command/install_egg_info.py,sha256=0kW0liVMeadkjX0ZcRfMptKFen07Gw6gyw1VHT5KIwc,2603
setuptools/_distutils/command/install_headers.py,sha256=XQ6idkbIDfr1ljXCOznuVUMvOFpHBn6cK0Wz9gIM2b4,1298
setuptools/_distutils/command/install_lib.py,sha256=9AofR-MO9lAtjwwuukCptepOaJEKMZW2VHiyR5hU7HA,8397
setuptools/_distutils/command/install_scripts.py,sha256=_CLUeQwGJRcY2kik7azPMn5IdtDCrjWdUvZ1khlG6ck,2017
setuptools/_distutils/command/py37compat.py,sha256=qzRhhvTihqx_PZZt2ZYECxh1X3Oj255VqatzelYFAKw,671
setuptools/_distutils/command/register.py,sha256=2jaq9968rt2puRVDBx1HbNiXv27uOk8idE_4lPf_3VM,11712
setuptools/_distutils/command/sdist.py,sha256=qotJjAOzyhJjq2-oDImjNFrOtaSneEFDJTB-sEk1wnU,19005
setuptools/_distutils/command/upload.py,sha256=BLO1w7eSAqsCjCLXtf_CRVSjwF1WmyOByGVGNdcQ8oY,7597
setuptools/_distutils/config.py,sha256=dtHgblx9JhfyrKx1-J7Jlxw_f7s8ZbPFQii2UWMTZpY,4827
setuptools/_distutils/core.py,sha256=jbdOkpOK09xi-56vhhwvn3fYdhLb5DJO8q3K1fnQz0Q,8876
setuptools/_distutils/cygwinccompiler.py,sha256=QpmRAopZOYEKww_iCWTu3KLjs9gggyl90E0fagAxqCM,16938
setuptools/_distutils/debug.py,sha256=N6MrTAqK6l9SVk6tWweR108PM8Ol7qNlfyV-nHcLhsY,139
setuptools/_distutils/dep_util.py,sha256=GuR9Iw_jzZRkyemJ5HX8rB_wRGxkIBcBm1qh54r7zhk,3491
setuptools/_distutils/dir_util.py,sha256=UwhBOUTcV65GTwce4SPuTXR8Z8q3LYEcmttqcGb0bYo,7778
setuptools/_distutils/dist.py,sha256=Biuf6ca8uiFfMScRFsYUKtb5neMPtxKxRtXn50_1f3U,50421
setuptools/_distutils/errors.py,sha256=Yr6tKZGdzBoNi53vBtiq0UJ__X05CmxSdQJqOWaw6SY,3577
setuptools/_distutils/extension.py,sha256=bTb3Q0CoevGKYv5dX1ls--Ln8tlB0-UEOsi9BwzlZ-s,10515
setuptools/_distutils/fancy_getopt.py,sha256=OPxp2CxHi1Yp_d1D8JxW4Ueq9fC71tegQFaafh58GGU,17784
setuptools/_distutils/file_util.py,sha256=0hUqfItN_x2DVihR0MHdA4KCMVCOO8VoByaFp_a6MDg,8148
setuptools/_distutils/filelist.py,sha256=Z9f5hvepZnpniZ2IFmCnWIjdviWozs8sbARBhWajwoM,13407
setuptools/_distutils/log.py,sha256=hWBmdUC2K927QcVv3REMW3HMPclxccPQngxLSuUXQl0,1969
setuptools/_distutils/msvc9compiler.py,sha256=X623B92g0v8A3BEM9qpRf396AEd_hfjkfDUVTKu0hcE,30453
setuptools/_distutils/msvccompiler.py,sha256=qruALeGRq8-CjtjE2tLQ8W26QnchcYedWzFme8AxZ4Q,23540
setuptools/_distutils/py35compat.py,sha256=-sk1vBIsOgH-AobjIYbK_OEjdJF_54Ul_D1EiE9XM_c,455
setuptools/_distutils/py38compat.py,sha256=II7ddBxOijC7uNN4z_46HYUjwYTJYMNiLJoGTormZm0,212
setuptools/_distutils/spawn.py,sha256=4uE9k3VZWijxy7E_Rlcmh1MoamaPJ8rajdNBagKxjgU,3498
setuptools/_distutils/sysconfig.py,sha256=mrtbAa9QXYXrNEe2HGKFyes2oJTNqImcgJGWiXnxOtQ,21630
setuptools/_distutils/text_file.py,sha256=PsuAJeWdKJoLSV_6N6IpB5-0Pa84KzLUucJMFRazw3I,12483
setuptools/_distutils/unixccompiler.py,sha256=u2Sfs6LRmqQux4nZW08GwDtoFMded6wYnkiaO2TvKC4,14538
setuptools/_distutils/util.py,sha256=QJLa8Xkzuw_7-o7fK5TnVc_zp_UHBXbTjiGmrY8nQ_4,20375
setuptools/_distutils/version.py,sha256=8NogP6NPPQpp3EUMZcT9czEHia-ehqPo8spo_e7AgUU,12514
setuptools/_distutils/versionpredicate.py,sha256=ZxpEA-TQv88mUWc6hetUO4qSqA2sa7ipjZ3QEK5evDk,5133
setuptools/_imp.py,sha256=HmF91IbitRfsD5z-g4_wmcuH-RahyIONbPgiCOFgtzA,2392
setuptools/_vendor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
setuptools/_vendor/__pycache__/__init__.cpython-39.pyc,,
setuptools/_vendor/__pycache__/ordered_set.cpython-39.pyc,,
setuptools/_vendor/__pycache__/pyparsing.cpython-39.pyc,,
setuptools/_vendor/more_itertools/__init__.py,sha256=C7sXffHTXM3P-iaLPPfqfmDoxOflQMJLcM7ed9p3jak,82
setuptools/_vendor/more_itertools/__pycache__/__init__.cpython-39.pyc,,
setuptools/_vendor/more_itertools/__pycache__/more.cpython-39.pyc,,
setuptools/_vendor/more_itertools/__pycache__/recipes.cpython-39.pyc,,
setuptools/_vendor/more_itertools/more.py,sha256=DlZa8v6JihVwfQ5zHidOA-xDE0orcQIUyxVnCaUoDKE,117968
setuptools/_vendor/more_itertools/recipes.py,sha256=UkNkrsZyqiwgLHANBTmvMhCvaNSvSNYhyOpz_Jc55DY,16256
setuptools/_vendor/ordered_set.py,sha256=dbaCcs27dyN9gnMWGF5nA_BrVn6Q-NrjKYJpV9_fgBs,15130
setuptools/_vendor/packaging/__about__.py,sha256=PNMsaZn4UcCHyubgROH1bl6CluduPjI5kFrSp_Zgklo,736
setuptools/_vendor/packaging/__init__.py,sha256=6enbp5XgRfjBjsI9-bn00HjHf5TH21PDMOKkJW8xw-w,562
setuptools/_vendor/packaging/__pycache__/__about__.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/__init__.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/_compat.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/_structures.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/_typing.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/markers.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/requirements.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/tags.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/utils.cpython-39.pyc,,
setuptools/_vendor/packaging/__pycache__/version.cpython-39.pyc,,
setuptools/_vendor/packaging/_compat.py,sha256=MXdsGpSE_W-ZrHoC87andI4LV2FAwU7HLL-eHe_CjhU,1128
setuptools/_vendor/packaging/_structures.py,sha256=ozkCX8Q8f2qE1Eic3YiQ4buDVfgz2iYevY9e7R2y3iY,2022
setuptools/_vendor/packaging/_typing.py,sha256=x59EhQ57TMT-kTRyLZV25HZvYGGwbucTo6iKh_O0tMw,1812
setuptools/_vendor/packaging/markers.py,sha256=BCCxZbt8xgysH8v5pqbLkdtQnRZHIGkJQqlNBGek4nQ,9509
setuptools/_vendor/packaging/requirements.py,sha256=VHydZdk8m3qFxReomNwKr71cmpjantEV_xOhkEyyINI,4917
setuptools/_vendor/packaging/specifiers.py,sha256=uYp9l13F0LcknS6d4N60ytiBgFmIhKideOq9AnsxTco,31944
setuptools/_vendor/packaging/tags.py,sha256=NKMS37Zo_nWrZxgsD6zbXsXgc9edn9m160cBiLmHJdE,24067
setuptools/_vendor/packaging/utils.py,sha256=RShlvnjO2CtYSD8uri32frMMFMTmB-3ihsq1-ghzLEw,1811
setuptools/_vendor/packaging/version.py,sha256=Cnbm-OO9D_qd8ZTFxzFcjSavexSYFZmyeaoPvMsjgPc,15470
setuptools/_vendor/pyparsing.py,sha256=mahtkgcp3grNAD0re_9R0DLvBnvjzpeLwgJqT-3H1CE,232056
setuptools/archive_util.py,sha256=maJDbozRbDeSPw53VT0cb_IS3W0Ap73lJR8tX8RZDx0,7077
setuptools/build_meta.py,sha256=x7FI1UPKCKxBBSopXocfGDnJa98rQO8atKXSwJtdid8,10280
setuptools/command/__init__.py,sha256=e-8TJOikUe3St0fw2b2p9u5EDdSxl5zHUBJJKifbcQ8,217
setuptools/command/__pycache__/__init__.cpython-39.pyc,,
setuptools/command/__pycache__/alias.cpython-39.pyc,,
setuptools/command/__pycache__/bdist_egg.cpython-39.pyc,,
setuptools/command/__pycache__/bdist_rpm.cpython-39.pyc,,
setuptools/command/__pycache__/build_clib.cpython-39.pyc,,
setuptools/command/__pycache__/build_ext.cpython-39.pyc,,
setuptools/command/__pycache__/build_py.cpython-39.pyc,,
setuptools/command/__pycache__/develop.cpython-39.pyc,,
setuptools/command/__pycache__/dist_info.cpython-39.pyc,,
setuptools/command/__pycache__/easy_install.cpython-39.pyc,,
setuptools/command/__pycache__/egg_info.cpython-39.pyc,,
setuptools/command/__pycache__/install.cpython-39.pyc,,
setuptools/command/__pycache__/install_egg_info.cpython-39.pyc,,
setuptools/command/__pycache__/install_lib.cpython-39.pyc,,
setuptools/command/__pycache__/install_scripts.cpython-39.pyc,,
setuptools/command/__pycache__/py36compat.cpython-39.pyc,,
setuptools/command/__pycache__/register.cpython-39.pyc,,
setuptools/command/__pycache__/rotate.cpython-39.pyc,,
setuptools/command/__pycache__/saveopts.cpython-39.pyc,,
setuptools/command/__pycache__/sdist.cpython-39.pyc,,
setuptools/command/__pycache__/setopt.cpython-39.pyc,,
setuptools/command/__pycache__/test.cpython-39.pyc,,
setuptools/command/__pycache__/upload.cpython-39.pyc,,
setuptools/command/__pycache__/upload_docs.cpython-39.pyc,,
setuptools/command/alias.py,sha256=1sLQxZcNh6dDQpDmm4G7UGGTol83nY1NTPmNBbm2siI,2381
setuptools/command/bdist_egg.py,sha256=-upiB6fFtm8cQSQj1LRDVpG1-T143DsXCvV0fh03u7U,16604
setuptools/command/bdist_rpm.py,sha256=PxrgoHPNaw2Pw2qNjjHDPC-Ay_IaDbCqP3d_5N-cj2A,1182
setuptools/command/build_clib.py,sha256=fWHSFGkk10VCddBWCszvNhowbG9Z9CZXVjQ2uSInoOs,4415
setuptools/command/build_ext.py,sha256=SNK042HfB2ezlDQbSVRGFqI1IM5A4AsjU1wpV3fgskE,13212
setuptools/command/build_py.py,sha256=UydjclXl6FSyrPjXOOwZD-gHby0tIKoP-qu5itvyP0g,8276
setuptools/command/develop.py,sha256=5_Ss7ENd1_B_jVMY1tF5UV_y1Xu6jbVzAPG8oKeluGA,7012
setuptools/command/dist_info.py,sha256=5t6kOfrdgALT-P3ogss6PF9k-Leyesueycuk3dUyZnI,960
setuptools/command/easy_install.py,sha256=C8Ppz7A2js_6Mv0KO8rqy_YPWCZzJX34e0KtVZRJxNo,85660
setuptools/command/egg_info.py,sha256=se-FhYI1sZMzKd6lndV_-vNkJ31hX4HY4ZcMUu71l9k,25335
setuptools/command/install.py,sha256=UynjFBgRyyHrDZRVAmXrXG0vChJAMx-sxnOO3JoAzVo,4906
setuptools/command/install_egg_info.py,sha256=bMgeIeRiXzQ4DAGPV1328kcjwQjHjOWU4FngAWLV78Q,2203
setuptools/command/install_lib.py,sha256=Uz42McsyHZAjrB6cw9E7Bz0xsaTbzxnM1PI9CBhiPtE,3875
setuptools/command/install_scripts.py,sha256=o0jN_ex7yYYk8W5clymTFOXwkFMKzW9q_zd9Npcex7M,2593
setuptools/command/launcher manifest.xml,sha256=xlLbjWrB01tKC0-hlVkOKkiSPbzMml2eOPtJ_ucCnbE,628
setuptools/command/py36compat.py,sha256=7yLWzQj179Enx3pJ8V1cDDCzeLMFMd9XJXlK-iZTq5Y,4946
setuptools/command/register.py,sha256=kk3DxXCb5lXTvqnhfwx2g6q7iwbUmgTyXUCaBooBOUk,468
setuptools/command/rotate.py,sha256=SvsQPasezIojPjvMnfkqzh8P0U0tCj0daczF8uc3NQM,2128
setuptools/command/saveopts.py,sha256=za7QCBcQimKKriWcoCcbhxPjUz30gSB74zuTL47xpP4,658
setuptools/command/sdist.py,sha256=2wJds5JaCDpDZmxyN3vo4BqqrTkL-wRmDDLZPeYEGGE,6172
setuptools/command/setopt.py,sha256=okxhqD1NM1nQlbSVDCNv6P7Y7g680sc2r-tUW7wPH1Y,5086
setuptools/command/test.py,sha256=qGY-Hx1RPCndlVh2rsrEs5479CgmxRsrEflVLr98jVA,8088
setuptools/command/upload.py,sha256=XT3YFVfYPAmA5qhGg0euluU98ftxRUW-PzKcODMLxUs,462
setuptools/command/upload_docs.py,sha256=ba5kOyedD_u62weinrxqqnvpuQvBIuamXehJG6tAvO0,7218
setuptools/config.py,sha256=sm9ZbziX9DlOugcVlIbhqttMJwxwznGEsk82D8MVaDM,23123
setuptools/dep_util.py,sha256=BDx1BkzNQntvAB4alypHbW5UVBzjqths000PrUL4Zqc,949
setuptools/depends.py,sha256=iHfZdLdlCu2BllSF9bRg7NU0oqbPWMH8ljm4BuwQDY0,5474
setuptools/dist.py,sha256=cZtPPzGEhSomPH_vXH_DeCFetjJ9B8Hv8VUCG0KbZh8,43087
setuptools/errors.py,sha256=MVOcv381HNSajDgEUWzOQ4J6B5BHCBMSjHfaWcEwA1o,524
setuptools/extension.py,sha256=NMM46XjNdVelWemc0x8CyVKA5Ks6Zm3xTWSA2SS6xZM,1684
setuptools/extern/__init__.py,sha256=Hhf9W73WAitw9TdRJfDIb6YFjmK56CF61afds1Mg0HY,2407
setuptools/extern/__pycache__/__init__.cpython-39.pyc,,
setuptools/glob.py,sha256=1oZjbfjAHSXbgdhSuR6YGU8jKob9L8NtEmBYqcPTLYk,4873
setuptools/installer.py,sha256=jbhb7ZVkNV_bSUMgfnLcZw0IHr6REFnKF4o7_1Jqxm0,3567
setuptools/launch.py,sha256=TyPT-Ic1T2EnYvGO26gfNRP4ysBlrhpbRjQxWsiO414,812
setuptools/monkey.py,sha256=0e3HdVKXHL415O7np-AUqhEFXPPuDdJKbI47chQ_DE4,5217
setuptools/msvc.py,sha256=3LLt938e6OR7wWPzIvCQu7LCWZSIKqoKV6w3r8jV3kY,50561
setuptools/namespaces.py,sha256=PMqGVPXPYQgjUTvEg9bGccRAkIODrQ6NmsDg_fwErwI,3093
setuptools/package_index.py,sha256=2A1O7fpTXcfeD5IV4HWrIoEXXkgq5k8t9aWrjx90Vnw,39886
setuptools/py34compat.py,sha256=KYOd6ybRxjBW8NJmYD8t_UyyVmysppFXqHpFLdslGXU,245
setuptools/sandbox.py,sha256=mR83i-mu-ZUU_7TaMgYCeRSyzkqv8loJ_GR9xhS2DDw,14348
setuptools/script (dev).tmpl,sha256=RUzQzCQUaXtwdLtYHWYbIQmOaES5Brqq1FvUA_tu-5I,218
setuptools/script.tmpl,sha256=WGTt5piezO27c-Dbx6l5Q4T3Ff20A5z7872hv3aAhYY,138
setuptools/unicode_utils.py,sha256=aOOFo4JGwAsiBttGYDsqFS7YqWQeZ2j6DWiCuctR_00,941
setuptools/version.py,sha256=og_cuZQb0QI6ukKZFfZWPlr1HgJBPPn2vO2m_bI9ZTE,144
setuptools/wheel.py,sha256=0P8tSk105uF_Ub-30N2HU2X2v7MKDSdjpeQlRRW3SkI,8288
setuptools/windows_support.py,sha256=5GrfqSP2-dLGJoZTq2g6dCKkyQxxa2n5IQiXlJCoYEE,714
site-packages/setuptools-58.3.0.dist-info/entry_points.txt000064400000005114150564745010017437 0ustar00[distutils.commands]
alias = setuptools.command.alias:alias
bdist_egg = setuptools.command.bdist_egg:bdist_egg
bdist_rpm = setuptools.command.bdist_rpm:bdist_rpm
build_clib = setuptools.command.build_clib:build_clib
build_ext = setuptools.command.build_ext:build_ext
build_py = setuptools.command.build_py:build_py
develop = setuptools.command.develop:develop
dist_info = setuptools.command.dist_info:dist_info
easy_install = setuptools.command.easy_install:easy_install
egg_info = setuptools.command.egg_info:egg_info
install = setuptools.command.install:install
install_egg_info = setuptools.command.install_egg_info:install_egg_info
install_lib = setuptools.command.install_lib:install_lib
install_scripts = setuptools.command.install_scripts:install_scripts
rotate = setuptools.command.rotate:rotate
saveopts = setuptools.command.saveopts:saveopts
sdist = setuptools.command.sdist:sdist
setopt = setuptools.command.setopt:setopt
test = setuptools.command.test:test
upload_docs = setuptools.command.upload_docs:upload_docs

[distutils.setup_keywords]
dependency_links = setuptools.dist:assert_string_list
eager_resources = setuptools.dist:assert_string_list
entry_points = setuptools.dist:check_entry_points
exclude_package_data = setuptools.dist:check_package_data
extras_require = setuptools.dist:check_extras
include_package_data = setuptools.dist:assert_bool
install_requires = setuptools.dist:check_requirements
namespace_packages = setuptools.dist:check_nsp
package_data = setuptools.dist:check_package_data
packages = setuptools.dist:check_packages
python_requires = setuptools.dist:check_specifier
setup_requires = setuptools.dist:check_requirements
test_loader = setuptools.dist:check_importable
test_runner = setuptools.dist:check_importable
test_suite = setuptools.dist:check_test_suite
tests_require = setuptools.dist:check_requirements
use_2to3 = setuptools.dist:invalid_unless_false
zip_safe = setuptools.dist:assert_bool

[egg_info.writers]
PKG-INFO = setuptools.command.egg_info:write_pkg_info
dependency_links.txt = setuptools.command.egg_info:overwrite_arg
depends.txt = setuptools.command.egg_info:warn_depends_obsolete
eager_resources.txt = setuptools.command.egg_info:overwrite_arg
entry_points.txt = setuptools.command.egg_info:write_entries
namespace_packages.txt = setuptools.command.egg_info:overwrite_arg
requires.txt = setuptools.command.egg_info:write_requirements
top_level.txt = setuptools.command.egg_info:write_toplevel_names

[setuptools.finalize_distribution_options]
keywords = setuptools.dist:Distribution._finalize_setup_keywords
parent_finalize = setuptools.dist:_Distribution.finalize_options

site-packages/setuptools-58.3.0.dist-info/INSTALLER000064400000000004150564745010015412 0ustar00pip
site-packages/setuptools-58.3.0.dist-info/REQUESTED000064400000000000150564745010015412 0ustar00site-packages/setuptools-58.3.0.dist-info/METADATA000064400000011364150564745010015250 0ustar00Metadata-Version: 2.1
Name: setuptools
Version: 58.3.0
Summary: Easily download, build, install, upgrade, and uninstall Python packages
Home-page: https://github.com/pypa/setuptools
Author: Python Packaging Authority
Author-email: distutils-sig@python.org
License: UNKNOWN
Project-URL: Documentation, https://setuptools.readthedocs.io/
Keywords: CPAN PyPI distutils eggs package management
Platform: UNKNOWN
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: MIT License
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3 :: Only
Classifier: Topic :: Software Development :: Libraries :: Python Modules
Classifier: Topic :: System :: Archiving :: Packaging
Classifier: Topic :: System :: Systems Administration
Classifier: Topic :: Utilities
Requires-Python: >=3.6
License-File: LICENSE
Provides-Extra: certs
Provides-Extra: docs
Requires-Dist: sphinx ; extra == 'docs'
Requires-Dist: jaraco.packaging (>=8.2) ; extra == 'docs'
Requires-Dist: rst.linker (>=1.9) ; extra == 'docs'
Requires-Dist: jaraco.tidelift (>=1.4) ; extra == 'docs'
Requires-Dist: pygments-github-lexers (==0.0.5) ; extra == 'docs'
Requires-Dist: sphinx-inline-tabs ; extra == 'docs'
Requires-Dist: sphinxcontrib-towncrier ; extra == 'docs'
Requires-Dist: furo ; extra == 'docs'
Provides-Extra: ssl
Provides-Extra: testing
Requires-Dist: pytest (>=4.6) ; extra == 'testing'
Requires-Dist: pytest-checkdocs (>=2.4) ; extra == 'testing'
Requires-Dist: pytest-flake8 ; extra == 'testing'
Requires-Dist: pytest-cov ; extra == 'testing'
Requires-Dist: pytest-enabler (>=1.0.1) ; extra == 'testing'
Requires-Dist: mock ; extra == 'testing'
Requires-Dist: flake8-2020 ; extra == 'testing'
Requires-Dist: virtualenv (>=13.0.0) ; extra == 'testing'
Requires-Dist: pytest-virtualenv (>=1.2.7) ; extra == 'testing'
Requires-Dist: wheel ; extra == 'testing'
Requires-Dist: paver ; extra == 'testing'
Requires-Dist: pip (>=19.1) ; extra == 'testing'
Requires-Dist: jaraco.envs ; extra == 'testing'
Requires-Dist: pytest-xdist ; extra == 'testing'
Requires-Dist: sphinx ; extra == 'testing'
Requires-Dist: jaraco.path (>=3.2.0) ; extra == 'testing'
Requires-Dist: pytest-black (>=0.3.7) ; (platform_python_implementation != "PyPy") and extra == 'testing'
Requires-Dist: pytest-mypy ; (platform_python_implementation != "PyPy") and extra == 'testing'

.. image:: https://img.shields.io/pypi/v/setuptools.svg
   :target: `PyPI link`_

.. image:: https://img.shields.io/pypi/pyversions/setuptools.svg
   :target: `PyPI link`_

.. _PyPI link: https://pypi.org/project/setuptools

.. image:: https://github.com/pypa/setuptools/workflows/tests/badge.svg
   :target: https://github.com/pypa/setuptools/actions?query=workflow%3A%22tests%22
   :alt: tests

.. image:: https://img.shields.io/badge/code%20style-black-000000.svg
   :target: https://github.com/psf/black
   :alt: Code style: Black

.. image:: https://img.shields.io/readthedocs/setuptools/latest.svg
    :target: https://setuptools.readthedocs.io

.. image:: https://img.shields.io/badge/skeleton-2021-informational
   :target: https://blog.jaraco.com/skeleton

.. image:: https://img.shields.io/codecov/c/github/pypa/setuptools/master.svg?logo=codecov&logoColor=white
   :target: https://codecov.io/gh/pypa/setuptools

.. image:: https://tidelift.com/badges/github/pypa/setuptools?style=flat
   :target: https://tidelift.com/subscription/pkg/pypi-setuptools?utm_source=pypi-setuptools&utm_medium=readme

See the `Installation Instructions
<https://packaging.python.org/installing/>`_ in the Python Packaging
User's Guide for instructions on installing, upgrading, and uninstalling
Setuptools.

Questions and comments should be directed to the `distutils-sig
mailing list <http://mail.python.org/pipermail/distutils-sig/>`_.
Bug reports and especially tested patches may be
submitted directly to the `bug tracker
<https://github.com/pypa/setuptools/issues>`_.


Code of Conduct
===============

Everyone interacting in the setuptools project's codebases, issue trackers,
chat rooms, and mailing lists is expected to follow the
`PSF Code of Conduct <https://github.com/pypa/.github/blob/main/CODE_OF_CONDUCT.md>`_.


For Enterprise
==============

Available as part of the Tidelift Subscription.

Setuptools and the maintainers of thousands of other packages are working with Tidelift to deliver one enterprise subscription that covers all of the open source you use.

`Learn more <https://tidelift.com/subscription/pkg/pypi-setuptools?utm_source=pypi-setuptools&utm_medium=referral&utm_campaign=github>`_.


Security Contact
================

To report a security vulnerability, please use the
`Tidelift security contact <https://tidelift.com/security>`_.
Tidelift will coordinate the fix and disclosure.


site-packages/setuptools-58.3.0.dist-info/top_level.txt000064400000000051150564745010016666 0ustar00_distutils_hack
pkg_resources
setuptools
site-packages/pip/__pycache__/__main__.cpython-39.pyc000064400000001162150564745010016444 0ustar00a

��Re��@s�ddlZddlZddlZejdde��fvr:ej�d�edkrdej�ej�e��Zej�	de�e
dkr�ejdedd�ddl
mZe�e��dS)�N��__main__�ignorez.*packaging\.version)�category�module)�main)�os�sys�warnings�path�getcwd�pop�__package__�dirname�__file__�insert�__name__�filterwarnings�DeprecationWarning�pip._internal.cli.mainr�_main�exit�rr�{/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/__main__.py�<module>s�site-packages/pip/__pycache__/__init__.cpython-39.pyc000064400000001236150564745010016465 0ustar00a

��Ree�@s2ddlmZmZdZdeeeed�dd�ZdS)�)�List�Optionalz21.3.1N)�args�returncCsddlm}||�S)z�This is an internal API only meant for use by pip's own console scripts.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    r)�_wrapper)�pip._internal.utils.entrypointsr)rr�r�{/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/__init__.py�mainsr
)N)�typingrr�__version__�str�intr
rrrr	�<module>ssite-packages/pip/_vendor/tomli/__pycache__/_parser.cpython-39.pyc000064400000040011150564745010021133 0ustar00a

��Re�W�	@s�ddlZddlmZddlmZmZmZmZmZm	Z	m
Z
mZmZddl
mZmZmZmZmZmZmZmZmZer|ddlmZedd�ed�D��eed	��BZeed
�Zeed�ZeZeed�Z eZ!ed
�Z"e"ed�BZ#eej$ej%d�Z&e&ed�BZ'edd
dddddd��Z(ee)gefZ*ee)dfZ+e,Z-Gdd�de.�Z/e0d�ee*ee)efd�dd�Z1e0d�e)e*ee)efd�dd �Z2Gd!d"�d"�Z3Gd#d$�d$�Z4Gd%d&�d&�Z5e)e-e	e)e-d'�d(d)�Z6e)e-e)ee)e7e-d*�d+d,�Z8e)e-e-d-�d.d/�Z9e)e-e-d-�d0d1�Z:e)e-e3e-d2�d3d4�Z;e)e-e3e-d2�d5d6�Z<e)e-e3e*e-d7�d8d9�Z=e)e-e*ee-e+efd:�d;d<�Z>e)e-ee-e+fd-�d=d>�Z?e)e-ee-e)fd-�d?d@�Z@e)e-ee-e)fd-�dAdB�ZAe)e-e*ee-eBfd:�dCdD�ZCe)e-e*ee-eDfd:�dEdF�ZEdGdH�e)e-e7ee-e)fdI�dJdK�ZFe)e-ee-e)fd-�dLdM�ZGe)e-e,ee-e)fdN�dOdP�ZHe)e-ee-e)fd-�dQdR�ZIe)e-e7ee-e)fdS�dTdU�ZJe)e-e7ee-e)fdI�dVdW�ZKe)e-dXee-e)fdY�dZd[�ZLe)e-e*ee-efd:�d\d]�ZMe)e-e)e/d^�d_d`�ZNe,e7da�dbdc�ZOdS)d�N)�MappingProxyType)	�
TYPE_CHECKING�Any�Callable�Dict�	FrozenSet�Iterable�Optional�TextIO�Tuple)	�RE_BIN�RE_DATETIME�RE_HEX�RE_LOCALTIME�	RE_NUMBER�RE_OCT�match_to_datetime�match_to_localtime�match_to_number)�Patternccs|]}t|�VqdS�N)�chr)�.0�i�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tomli/_parser.py�	<genexpr>�r� ��	z	

z	
z 	�
z-_z"'���
�"�\)z\bz\tz\nz\fz\rz\"z\\.c@seZdZdZdS)�TOMLDecodeErrorz0An error raised if a document is not valid TOML.N)�__name__�
__module__�__qualname__�__doc__rrrrr'Bsr'��parse_float)�fpr-�returncCs|��}t||d�S)zParse TOML from a file object.r,)�read�loads)r.r-�srrr�loadFsr3)r2r-r/cCsL|�dd�}d}t�}t||t�}z||}WntyFY�qDYn0|dkrZ|d7}q|tvr~t||||�}t||t�}nt|dkr�z||d}Wnty�d}Yn0|dkr�t|||�}nt|||�}t||t�}n|dkr�t	||d��t
||�}z||}Wnt�y"Y�qDYn0|dk�r:t	||d	��|d7}q|jjS)
zParse TOML from a string.z
r!r��[N�#zInvalid statementz5Expected newline or end of document after a statement)
�replace�State�
skip_chars�TOML_WS�
IndexError�KEY_INITIAL_CHARS�key_value_rule�create_list_rule�create_dict_rule�suffixed_err�skip_comment�out�dict)r2r-�src�pos�state�char�second_charrrrr1LsH



�
r1c@seZdZdd�dd�ZdS)r8N�r/cCst�|_t�|_d|_dS)Nr)�
NestedDictrB�Flags�flags�header_namespace��selfrrr�__init__�szState.__init__)r(r)r*rPrrrrr8�sr8c@speZdZdZdZdZdd�dd�Zedd�d	d
�Zeee	dd�dd
�Z
ee	edd�dd�Zee	ed�dd�Z
dS)rKz)Flags that map to parsed keys/namespaces.rr4NrIcCs
i|_dSr��_flagsrNrrrrP�szFlags.__init__��keyr/cCsF|j}|dd�D]}||vr$dS||d}q|�|dd�dS)N����nested)rR�pop)rOrT�cont�krrr�	unset_all�szFlags.unset_all)�head_key�rel_key�flagr/cCs�|j}|D],}||vr*t�t�id�||<||d}q
|D]@}||vr\||d�|�n|ht�id�||<||d}q<dS)N�rL�recursive_flagsrVrVrL�rR�set�add)rOr[r\r]rXrYrrr�set_for_relative_key�szFlags.set_for_relative_key)rTr]�	recursiver/cCs�|j}|dd�|d}}|D],}||vr@t�t�id�||<||d}q ||vrjt�t�id�||<|||rxdnd�|�dS)NrUr^rVr_rLr`)rOrTr]rdrX�
key_parent�key_stemrYrrrra�sz	Flags.set)rTr]r/cCs�|sdS|j}|dd�D]4}||vr,dS||}||dvrFdS|d}q|d}||vr�||}||dvp~||dvSdS)NFrUr_TrVrLrQ)rOrTr]rXrY�
inner_contrfrrr�is_�s
z	Flags.is_)r(r)r*r+�FROZEN�
EXPLICIT_NESTrP�KeyrZ�intrc�boolrarhrrrrrK�s
rKc@sBeZdZdd�dd�Zdd�eeed�dd	�Zedd
�dd�ZdS)
rJNrIcCs
i|_dSr)rCrNrrrrP�szNestedDict.__init__T��access_lists)rTror/cCsT|j}|D]D}||vri||<||}|r<t|t�r<|d}t|t�s
td��q
|S)NrUz There is no nest behind this key)rC�
isinstance�list�KeyError)rOrTrorXrYrrr�get_or_create_nest�s

zNestedDict.get_or_create_nestrScCsV|�|dd��}|d}||vrH||}t|t�s<td��|�i�n
ig||<dS)NrUz/An object other than list found behind this key)rsrprqrr�append)rOrTrX�last_key�list_rrr�append_nest_to_list�s
zNestedDict.append_nest_to_list)	r(r)r*rPrkrmrCrsrwrrrrrJ�s��rJ)rDrE�charsr/cCs2z|||vr|d7}qWnty,Yn0|S)Nr4)r;)rDrErxrrrr9�sr9)rDrE�expect�error_on�error_on_eofr/c		Cs�z|�||�}Wn2tyBt|�}|r>t||d|�d���Yn0|�|||��}|r�tt|��}|�||�}t||d|�d���|S)Nz
Expected "r%zFound invalid character ")�index�
ValueError�lenr@�intersection�next�iter)	rDrEryrzr{�new_pos�	bad_chars�bad_char�bad_posrrr�
skip_until�sr��rDrEr/cCsFz||}Wnty"d}Yn0|dkrBt||ddtdd�S|S)Nr6r4r!F�rzr{)r;r��ILLEGAL_COMMENT_CHARS)rDrErGrrrrAs
�rAcCs,|}t||t�}t||�}||kr|SqdSr)r9�TOML_WS_AND_NEWLINErA)rDrE�pos_before_skiprrr�skip_comments_and_array_ws!s

r�)rDrErFr/cCs�|d7}t||t�}t||�\}}|j�|tj�sB|j�|tj�rVt||d|�d���|jj	|tjdd�z|j
�|�Wnty�t||d��Yn0||_
|||d�dkr�t||d��|dS)	Nr4zCan not declare z twiceF�rd�Can not overwrite a value�]z.Expected "]" at the end of a table declaration)r9r:�	parse_keyrLrhrKrjrir@rarBrsrrrM)rDrErFrTrrrr?*s r?cCs�|d7}t||t�}t||�\}}|j�|tj�rDt||d|����|j�|�|jj	|tj
dd�z|j�|�Wnt
y�t||d��Yn0||_|||d�}|dkr�t||d|�d���|dS)	N��#Can not mutate immutable namespace Fr�r�z]]zFound "z3" at the end of an array declaration. Expected "]]")r9r:r�rLrhrKrir@rZrarjrBrwrrrM)rDrErFrT�
end_markerrrrr>=s(
�r>)rDrErFr-r/cCs�t|||�\}}}|dd�|d}}|j|}|j�|tj�rTt||d|����|j�|j|tj�z|j	�
|�}	Wnty�t||d��Yn0||	vr�t||d��t|t
tf�r�|j|}
|jj|
tjdd�||	|<|S)NrUr�r�Tr�)�parse_key_value_pairrMrLrhrKrir@rcrjrBrsrrrprCrqra)rDrErFr-rT�valuererf�abs_key_parent�nest�abs_keyrrrr=Ys&
�
r=)rDrEr-r/cCstt||�\}}z||}Wnty0d}Yn0|dkrFt||d��|d7}t||t�}t|||�\}}|||fS)N�=z,Expected "=" after a key in a key/value pairr4)r�r;r@r9r:�parse_value)rDrEr-rTrGr�rrrr�rs
r�cCs�t||�\}}|g}t||t�}z||}WntyBd}Yn0|dkrX|t|�fS|d7}t||t�}t||�\}}|�|�t||t�}q dS)N�.r4)�parse_key_partr9r:r;�tuplert)rDrE�key_partrTrGrrrr��s

r�cCs�z||}Wnty"d}Yn0|tvrL|}t||t�}||||�fS|dkr^t||�S|dkrpt||�St||d��dS)N�'r%z(Invalid initial character for a key part)r;�BARE_KEY_CHARSr9�parse_literal_str�parse_one_line_basic_strr@)rDrErG�	start_posrrrr��s


r�cCs|d7}t||dd�S)Nr4F��	multiline)�parse_basic_str�rDrErrrr��sr�cCs�|d7}g}t||�}|||d�dkr6|d|fSt|||�\}}|�|�t||�}|||d�}|dkr~|d|fS|dkr�t||d��|d7}t||�}|||d�dkr6|d|fSq6dS)Nr4r��,zUnclosed array)r�r�rtr@)rDrEr-�array�val�crrr�parse_array�s"



r�cCsf|d7}t�}t�}t||t�}|||d�dkrB|d|jfSt|||�\}}}|dd�|d}}|�|tj�r�t||d|����z|j	|dd�}	Wnt
y�t||d��Yn0||	vr�t||d|�d	���||	|<t||t�}|||d�}
|
dk�r|d|jfS|
d
k�r*t||d��t|ttf��rL|j
|tjdd
�|d7}t||t�}qBdS)Nr4�}rUr�Frnr�zDuplicate inline table key "r%r�zUnclosed inline tableTr�)rJrKr9r:rCr�rhrir@rsrrrprqra)rDrEr-�nested_dictrLrTr�rerfr�r�rrr�parse_inline_table�s6

r�Fr�)rDrEr�r/cCs�|||d�}|d7}|r�|dvr�|dkrpt||t�}|||d�}|sT|dfS|dkrht||d��|d7}t||t�}|dfS|dkr�t||d	�S|d
kr�t||d�Sz|t|fWSty�t|�dkr�t||d��t||d��Yn0dS)
Nr�>z\	�\
z\ r�r4�r!zUnescaped "\" in a stringz\u�z\U��Unterminated string)r9r:r@r��parse_hex_char�BASIC_STR_ESCAPE_REPLACEMENTSrrr~)rDrEr��	escape_idrGrrr�parse_basic_str_escape�s.r�cCst||dd�S)NTr�)r�r�rrr� parse_basic_str_escape_multiline�sr�)rDrE�hex_lenr/cCsl||||�}t|�|ks.tdd�|D��r:t||d��||7}t|d�}t|�s`t||d��|t|�fS)Ncss|]}|tjvVqdSr)�string�	hexdigits)rr�rrrrrz!parse_hex_char.<locals>.<genexpr>zInvalid hex value�z/Escaped character is not a Unicode scalar value)r~�anyr@rl�is_unicode_scalar_valuer)rDrEr��hex_str�hex_intrrrr�s
r�cCs2|d7}|}t||dtdd�}|d|||�fS)Nr4r�Tr�)r��ILLEGAL_LITERAL_STR_CHARS)rDrEr�rrrr�s
�r�)rDrE�literalr/cCs�|d7}|||d�dkr$|d7}|rTd}t||dtdd�}|||�}|d}nd}t||dd	�\}}|||d�|kr�||fS|d7}|||d�|kr�|||fS|d7}|||d
fS)N�r4r!r�z'''Tr�r%r�r�)r��#ILLEGAL_MULTILINE_LITERAL_STR_CHARSr�)rDrEr��delim�end_pos�resultrrr�parse_multiline_strs.�
r�c	Cs|rt}t}nt}t}d}|}z||}WntyHt||d��Yn0|dkr�|sn|d||||�fS||d|d�dkr�|d||||�fS|d7}q|dkr�||||�7}|||�\}}||7}|}q||vr�t||d|�d���|d7}qdS)	Nr�r�r%r4r��""r&zIllegal character ")�!ILLEGAL_MULTILINE_BASIC_STR_CHARSr��ILLEGAL_BASIC_STR_CHARSr�r;r@)	rDrEr�rz�
parse_escapesr�r�rG�
parsed_escaperrrr�6s6r�r)rDrE�regexr/cCs,|�||�}|st||d��|��|��fS)NzUnexpected sequence)�matchr@�end�group)rDrEr�r�rrr�parse_regexVsr�cCs�z||}Wnty"d}Yn0|dkr\||d|d�dkrRt||dd�St||�S|dkr�||d|d�dkr�t||d	d�St||�S|d
kr�||d|d�dkr�|dd	fS|d
kr�||d|d�dkr�|ddfSt�||�}|�r8zt|�}Wn t�y*t||d��Yn0|�	�|fSt
�||�}|�rZ|�	�t|�fS|dk�r�||d|d�}|dk�r�t||dt
�\}}|t|d�fS|dk�r�t||dt�\}}	|t|	d�fS|dk�r�t||dt�\}}
|t|
d�fSt�||�}|�r |�	�t||�fS|dk�r6t|||�S|dk�rLt|||�S|||d�}|dv�rv|d||�fS|||d�}
|
dv�r�|d||
�fSt||d��dS)Nr%r4r�r�F)r�r�z''T�tr��rue�f��alsezInvalid date or datetime�0r��xr��or��br5�{>�nan�inf>z+nanz+infz-infz-nanz
Invalid value)r;r�r�r�r
r�rr}r@r�rrr�rrlrrrrr�r�)rDrEr-rG�datetime_match�datetime_obj�localtime_matchrHr��oct_str�bin_str�number_match�first_three�
first_fourrrrr�]sj










r�)rDrE�msgr/cCs,tttd�dd�}t|�d|||��d��S)zZReturn a `TOMLDecodeError` where error message is suffixed with
    coordinates in source.r�cSsV|t|�krdS|�dd|�d}|dkr4|d}n||�dd|�}d|�d|��S)Nzend of documentr!rr4zline z	, column )r~�count�rindex)rDrE�line�columnrrr�
coord_repr�s
z suffixed_err.<locals>.coord_reprz (at �))�str�Posr')rDrEr�r�rrrr@�s
r@)�	codepointr/cCs0d|kodknp.d|ko*dkSS)Nri��i�i��r)r�rrrr��sr�)Pr��typesr�typingrrrrrrr	r
r�pip._vendor.tomli._rerr
rrrrrrr�rer�	frozenset�ranger�
ASCII_CTRLr�r�r�r�r�r:r��
ascii_letters�digitsr�r<r�r��
ParseFloatrkrlr�r}r'�floatr3r1r8rKrJr9rmr�rAr�r?r>r=r�r�r�r�rqr�rCr�r�r�r�r�r�r�r�r�r@r�rrrr�<module>s�,,"��
  ?
>"
�	�!�
�	 
�Psite-packages/pip/_vendor/tomli/__pycache__/_re.cpython-39.pyc000064400000004644150564745010020261 0ustar00a

��Rey
�@s�ddlmZmZmZmZmZmZddlZddlmZm	Z	m
Z
mZer\ddlmZddl
mZdZe�d�Ze�d�Ze�d	�Ze�d
�Ze�e�Ze�dedd
�Zdeeefd�dd�Zded�dd�Zdde	d�dd�ZdS)�)�date�datetime�time�	timedelta�timezone�tzinfoN)�
TYPE_CHECKING�Any�Optional�Union)�Match)�
ParseFloatz7([01][0-9]|2[0-3]):([0-5][0-9]):([0-5][0-9])(\.[0-9]+)?z[0-9A-Fa-f](?:_?[0-9A-Fa-f])*z[01](?:_?[01])*z[0-7](?:_?[0-7])*zS[+-]?(?:0|[1-9](?:_?[0-9])*)(?:\.[0-9](?:_?[0-9])*)?(?:[eE][+-]?[0-9](?:_?[0-9])*)?z>([0-9]{4})-(0[1-9]|1[0-2])-(0[1-9]|1[0-9]|2[0-9]|3[01])(?:[T ]z.(?:(Z)|([+-])([01][0-9]|2[0-3]):([0-5][0-9]))?z)?r)�match�returncCs�|��\}}}}}}}}}	}
}t|�t|�t|�}}
}|durNt||
|�St|�t|�t|�}}}|r�t|dd��dd�dd��nd}|	r�|	dkr�dnd}tt|t|
�|t|�d��}n|r�tj}nd}t||
||||||d	�S)
z�Convert a `RE_DATETIME` match to `datetime.datetime` or `datetime.date`.

    Raises ValueError if the match does not correspond to a valid date
    or datetime.
    N���0r�+���)�hours�minutes)r)�groups�intr�ljustrr�utcr)r�year_str�	month_str�day_str�hour_str�
minute_str�sec_str�
micros_str�	zulu_time�offset_dir_str�offset_hour_str�offset_minute_str�year�month�day�hour�minute�sec�micros�
offset_dir�tz�r/��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tomli/_re.py�match_to_datetime"s<�(

��r1cCsR|��\}}}}|r4t|dd��dd�dd��nd}tt|�t|�t|�|�S)Nrrrr)rrrr)rrrr r!r,r/r/r0�match_to_localtimeIs(r2r
)r�parse_floatrcCs0|��}d|vs d|vs d|vr(||�St|�S)N�.�e�E)�groupr)rr3�	match_strr/r/r0�match_to_numberOsr9)rrrrrr�re�typingrr	r
rr�pip._vendor.tomli._parserr
�_TIME_RE_STR�compile�RE_HEX�RE_BIN�RE_OCT�	RE_NUMBER�RE_LOCALTIME�RE_DATETIMEr1r2r9r/r/r/r0�<module>s2 


�
����
'site-packages/pip/_vendor/tomli/__pycache__/__init__.cpython-39.pyc000064400000000642150564745010021245 0ustar00a

��Re��@s$dZdZdZddlmZmZmZdS)zA lil' TOML parser.)�loads�load�TOMLDecodeErrorz1.0.3�)rrrN)�__doc__�__all__�__version__�pip._vendor.tomli._parserrrr�r	r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tomli/__init__.py�<module>ssite-packages/pip/_vendor/tomli/_re.py000064400000005171150564745010013766 0ustar00from datetime import date, datetime, time, timedelta, timezone, tzinfo
import re
from typing import TYPE_CHECKING, Any, Optional, Union

if TYPE_CHECKING:
    from re import Match

    from pip._vendor.tomli._parser import ParseFloat

# E.g.
# - 00:32:00.999999
# - 00:32:00
_TIME_RE_STR = r"([01][0-9]|2[0-3]):([0-5][0-9]):([0-5][0-9])(\.[0-9]+)?"

RE_HEX = re.compile(r"[0-9A-Fa-f](?:_?[0-9A-Fa-f])*")
RE_BIN = re.compile(r"[01](?:_?[01])*")
RE_OCT = re.compile(r"[0-7](?:_?[0-7])*")
RE_NUMBER = re.compile(
    r"[+-]?(?:0|[1-9](?:_?[0-9])*)"  # integer
    + r"(?:\.[0-9](?:_?[0-9])*)?"  # optional fractional part
    + r"(?:[eE][+-]?[0-9](?:_?[0-9])*)?"  # optional exponent part
)
RE_LOCALTIME = re.compile(_TIME_RE_STR)
RE_DATETIME = re.compile(
    r"([0-9]{4})-(0[1-9]|1[0-2])-(0[1-9]|1[0-9]|2[0-9]|3[01])"  # date, e.g. 1988-10-27
    + r"(?:"
    + r"[T ]"
    + _TIME_RE_STR
    + r"(?:(Z)|([+-])([01][0-9]|2[0-3]):([0-5][0-9]))?"  # time offset
    + r")?"
)


def match_to_datetime(match: "Match") -> Union[datetime, date]:
    """Convert a `RE_DATETIME` match to `datetime.datetime` or `datetime.date`.

    Raises ValueError if the match does not correspond to a valid date
    or datetime.
    """
    (
        year_str,
        month_str,
        day_str,
        hour_str,
        minute_str,
        sec_str,
        micros_str,
        zulu_time,
        offset_dir_str,
        offset_hour_str,
        offset_minute_str,
    ) = match.groups()
    year, month, day = int(year_str), int(month_str), int(day_str)
    if hour_str is None:
        return date(year, month, day)
    hour, minute, sec = int(hour_str), int(minute_str), int(sec_str)
    micros = int(micros_str[1:].ljust(6, "0")[:6]) if micros_str else 0
    if offset_dir_str:
        offset_dir = 1 if offset_dir_str == "+" else -1
        tz: Optional[tzinfo] = timezone(
            timedelta(
                hours=offset_dir * int(offset_hour_str),
                minutes=offset_dir * int(offset_minute_str),
            )
        )
    elif zulu_time:
        tz = timezone.utc
    else:  # local date-time
        tz = None
    return datetime(year, month, day, hour, minute, sec, micros, tzinfo=tz)


def match_to_localtime(match: "Match") -> time:
    hour_str, minute_str, sec_str, micros_str = match.groups()
    micros = int(micros_str[1:].ljust(6, "0")[:6]) if micros_str else 0
    return time(int(hour_str), int(minute_str), int(sec_str), micros)


def match_to_number(match: "Match", parse_float: "ParseFloat") -> Any:
    match_str = match.group()
    if "." in match_str or "e" in match_str or "E" in match_str:
        return parse_float(match_str)
    return int(match_str)
site-packages/pip/_vendor/tomli/__init__.py000064400000000346150564745010014757 0ustar00"""A lil' TOML parser."""

__all__ = ("loads", "load", "TOMLDecodeError")
__version__ = "1.0.3"  # DO NOT EDIT THIS LINE MANUALLY. LET bump2version UTILITY DO IT

from pip._vendor.tomli._parser import TOMLDecodeError, load, loads
site-packages/pip/_vendor/tomli/_parser.py000064400000053617150564745010014664 0ustar00import string
from types import MappingProxyType
from typing import (
    TYPE_CHECKING,
    Any,
    Callable,
    Dict,
    FrozenSet,
    Iterable,
    Optional,
    TextIO,
    Tuple,
)

from pip._vendor.tomli._re import (
    RE_BIN,
    RE_DATETIME,
    RE_HEX,
    RE_LOCALTIME,
    RE_NUMBER,
    RE_OCT,
    match_to_datetime,
    match_to_localtime,
    match_to_number,
)

if TYPE_CHECKING:
    from re import Pattern


ASCII_CTRL = frozenset(chr(i) for i in range(32)) | frozenset(chr(127))

# Neither of these sets include quotation mark or backslash. They are
# currently handled as separate cases in the parser functions.
ILLEGAL_BASIC_STR_CHARS = ASCII_CTRL - frozenset("\t")
ILLEGAL_MULTILINE_BASIC_STR_CHARS = ASCII_CTRL - frozenset("\t\n\r")

ILLEGAL_LITERAL_STR_CHARS = ILLEGAL_BASIC_STR_CHARS
ILLEGAL_MULTILINE_LITERAL_STR_CHARS = ASCII_CTRL - frozenset("\t\n")

ILLEGAL_COMMENT_CHARS = ILLEGAL_BASIC_STR_CHARS

TOML_WS = frozenset(" \t")
TOML_WS_AND_NEWLINE = TOML_WS | frozenset("\n")
BARE_KEY_CHARS = frozenset(string.ascii_letters + string.digits + "-_")
KEY_INITIAL_CHARS = BARE_KEY_CHARS | frozenset("\"'")

BASIC_STR_ESCAPE_REPLACEMENTS = MappingProxyType(
    {
        "\\b": "\u0008",  # backspace
        "\\t": "\u0009",  # tab
        "\\n": "\u000A",  # linefeed
        "\\f": "\u000C",  # form feed
        "\\r": "\u000D",  # carriage return
        '\\"': "\u0022",  # quote
        "\\\\": "\u005C",  # backslash
    }
)

# Type annotations
ParseFloat = Callable[[str], Any]
Key = Tuple[str, ...]
Pos = int


class TOMLDecodeError(ValueError):
    """An error raised if a document is not valid TOML."""


def load(fp: TextIO, *, parse_float: ParseFloat = float) -> Dict[str, Any]:
    """Parse TOML from a file object."""
    s = fp.read()
    return loads(s, parse_float=parse_float)


def loads(s: str, *, parse_float: ParseFloat = float) -> Dict[str, Any]:  # noqa: C901
    """Parse TOML from a string."""

    # The spec allows converting "\r\n" to "\n", even in string
    # literals. Let's do so to simplify parsing.
    src = s.replace("\r\n", "\n")
    pos = 0
    state = State()

    # Parse one statement at a time
    # (typically means one line in TOML source)
    while True:
        # 1. Skip line leading whitespace
        pos = skip_chars(src, pos, TOML_WS)

        # 2. Parse rules. Expect one of the following:
        #    - end of file
        #    - end of line
        #    - comment
        #    - key/value pair
        #    - append dict to list (and move to its namespace)
        #    - create dict (and move to its namespace)
        # Skip trailing whitespace when applicable.
        try:
            char = src[pos]
        except IndexError:
            break
        if char == "\n":
            pos += 1
            continue
        if char in KEY_INITIAL_CHARS:
            pos = key_value_rule(src, pos, state, parse_float)
            pos = skip_chars(src, pos, TOML_WS)
        elif char == "[":
            try:
                second_char: Optional[str] = src[pos + 1]
            except IndexError:
                second_char = None
            if second_char == "[":
                pos = create_list_rule(src, pos, state)
            else:
                pos = create_dict_rule(src, pos, state)
            pos = skip_chars(src, pos, TOML_WS)
        elif char != "#":
            raise suffixed_err(src, pos, "Invalid statement")

        # 3. Skip comment
        pos = skip_comment(src, pos)

        # 4. Expect end of line or end of file
        try:
            char = src[pos]
        except IndexError:
            break
        if char != "\n":
            raise suffixed_err(
                src, pos, "Expected newline or end of document after a statement"
            )
        pos += 1

    return state.out.dict


class State:
    def __init__(self) -> None:
        # Mutable, read-only
        self.out = NestedDict()
        self.flags = Flags()

        # Immutable, read and write
        self.header_namespace: Key = ()


class Flags:
    """Flags that map to parsed keys/namespaces."""

    # Marks an immutable namespace (inline array or inline table).
    FROZEN = 0
    # Marks a nest that has been explicitly created and can no longer
    # be opened using the "[table]" syntax.
    EXPLICIT_NEST = 1

    def __init__(self) -> None:
        self._flags: Dict[str, dict] = {}

    def unset_all(self, key: Key) -> None:
        cont = self._flags
        for k in key[:-1]:
            if k not in cont:
                return
            cont = cont[k]["nested"]
        cont.pop(key[-1], None)

    def set_for_relative_key(self, head_key: Key, rel_key: Key, flag: int) -> None:
        cont = self._flags
        for k in head_key:
            if k not in cont:
                cont[k] = {"flags": set(), "recursive_flags": set(), "nested": {}}
            cont = cont[k]["nested"]
        for k in rel_key:
            if k in cont:
                cont[k]["flags"].add(flag)
            else:
                cont[k] = {"flags": {flag}, "recursive_flags": set(), "nested": {}}
            cont = cont[k]["nested"]

    def set(self, key: Key, flag: int, *, recursive: bool) -> None:  # noqa: A003
        cont = self._flags
        key_parent, key_stem = key[:-1], key[-1]
        for k in key_parent:
            if k not in cont:
                cont[k] = {"flags": set(), "recursive_flags": set(), "nested": {}}
            cont = cont[k]["nested"]
        if key_stem not in cont:
            cont[key_stem] = {"flags": set(), "recursive_flags": set(), "nested": {}}
        cont[key_stem]["recursive_flags" if recursive else "flags"].add(flag)

    def is_(self, key: Key, flag: int) -> bool:
        if not key:
            return False  # document root has no flags
        cont = self._flags
        for k in key[:-1]:
            if k not in cont:
                return False
            inner_cont = cont[k]
            if flag in inner_cont["recursive_flags"]:
                return True
            cont = inner_cont["nested"]
        key_stem = key[-1]
        if key_stem in cont:
            cont = cont[key_stem]
            return flag in cont["flags"] or flag in cont["recursive_flags"]
        return False


class NestedDict:
    def __init__(self) -> None:
        # The parsed content of the TOML document
        self.dict: Dict[str, Any] = {}

    def get_or_create_nest(
        self,
        key: Key,
        *,
        access_lists: bool = True,
    ) -> dict:
        cont: Any = self.dict
        for k in key:
            if k not in cont:
                cont[k] = {}
            cont = cont[k]
            if access_lists and isinstance(cont, list):
                cont = cont[-1]
            if not isinstance(cont, dict):
                raise KeyError("There is no nest behind this key")
        return cont

    def append_nest_to_list(self, key: Key) -> None:
        cont = self.get_or_create_nest(key[:-1])
        last_key = key[-1]
        if last_key in cont:
            list_ = cont[last_key]
            if not isinstance(list_, list):
                raise KeyError("An object other than list found behind this key")
            list_.append({})
        else:
            cont[last_key] = [{}]


def skip_chars(src: str, pos: Pos, chars: Iterable[str]) -> Pos:
    try:
        while src[pos] in chars:
            pos += 1
    except IndexError:
        pass
    return pos


def skip_until(
    src: str,
    pos: Pos,
    expect: str,
    *,
    error_on: FrozenSet[str],
    error_on_eof: bool,
) -> Pos:
    try:
        new_pos = src.index(expect, pos)
    except ValueError:
        new_pos = len(src)
        if error_on_eof:
            raise suffixed_err(src, new_pos, f'Expected "{expect!r}"')

    bad_chars = error_on.intersection(src[pos:new_pos])
    if bad_chars:
        bad_char = next(iter(bad_chars))
        bad_pos = src.index(bad_char, pos)
        raise suffixed_err(src, bad_pos, f'Found invalid character "{bad_char!r}"')
    return new_pos


def skip_comment(src: str, pos: Pos) -> Pos:
    try:
        char: Optional[str] = src[pos]
    except IndexError:
        char = None
    if char == "#":
        return skip_until(
            src, pos + 1, "\n", error_on=ILLEGAL_COMMENT_CHARS, error_on_eof=False
        )
    return pos


def skip_comments_and_array_ws(src: str, pos: Pos) -> Pos:
    while True:
        pos_before_skip = pos
        pos = skip_chars(src, pos, TOML_WS_AND_NEWLINE)
        pos = skip_comment(src, pos)
        if pos == pos_before_skip:
            return pos


def create_dict_rule(src: str, pos: Pos, state: State) -> Pos:
    pos += 1  # Skip "["
    pos = skip_chars(src, pos, TOML_WS)
    pos, key = parse_key(src, pos)

    if state.flags.is_(key, Flags.EXPLICIT_NEST) or state.flags.is_(key, Flags.FROZEN):
        raise suffixed_err(src, pos, f"Can not declare {key} twice")
    state.flags.set(key, Flags.EXPLICIT_NEST, recursive=False)
    try:
        state.out.get_or_create_nest(key)
    except KeyError:
        raise suffixed_err(src, pos, "Can not overwrite a value")
    state.header_namespace = key

    if src[pos : pos + 1] != "]":
        raise suffixed_err(src, pos, 'Expected "]" at the end of a table declaration')
    return pos + 1


def create_list_rule(src: str, pos: Pos, state: State) -> Pos:
    pos += 2  # Skip "[["
    pos = skip_chars(src, pos, TOML_WS)
    pos, key = parse_key(src, pos)

    if state.flags.is_(key, Flags.FROZEN):
        raise suffixed_err(src, pos, f"Can not mutate immutable namespace {key}")
    # Free the namespace now that it points to another empty list item...
    state.flags.unset_all(key)
    # ...but this key precisely is still prohibited from table declaration
    state.flags.set(key, Flags.EXPLICIT_NEST, recursive=False)
    try:
        state.out.append_nest_to_list(key)
    except KeyError:
        raise suffixed_err(src, pos, "Can not overwrite a value")
    state.header_namespace = key

    end_marker = src[pos : pos + 2]
    if end_marker != "]]":
        raise suffixed_err(
            src,
            pos,
            f'Found "{end_marker!r}" at the end of an array declaration.'
            ' Expected "]]"',
        )
    return pos + 2


def key_value_rule(src: str, pos: Pos, state: State, parse_float: ParseFloat) -> Pos:
    pos, key, value = parse_key_value_pair(src, pos, parse_float)
    key_parent, key_stem = key[:-1], key[-1]
    abs_key_parent = state.header_namespace + key_parent

    if state.flags.is_(abs_key_parent, Flags.FROZEN):
        raise suffixed_err(
            src, pos, f"Can not mutate immutable namespace {abs_key_parent}"
        )
    # Containers in the relative path can't be opened with the table syntax after this
    state.flags.set_for_relative_key(state.header_namespace, key, Flags.EXPLICIT_NEST)
    try:
        nest = state.out.get_or_create_nest(abs_key_parent)
    except KeyError:
        raise suffixed_err(src, pos, "Can not overwrite a value")
    if key_stem in nest:
        raise suffixed_err(src, pos, "Can not overwrite a value")
    # Mark inline table and array namespaces recursively immutable
    if isinstance(value, (dict, list)):
        abs_key = state.header_namespace + key
        state.flags.set(abs_key, Flags.FROZEN, recursive=True)
    nest[key_stem] = value
    return pos


def parse_key_value_pair(
    src: str, pos: Pos, parse_float: ParseFloat
) -> Tuple[Pos, Key, Any]:
    pos, key = parse_key(src, pos)
    try:
        char: Optional[str] = src[pos]
    except IndexError:
        char = None
    if char != "=":
        raise suffixed_err(src, pos, 'Expected "=" after a key in a key/value pair')
    pos += 1
    pos = skip_chars(src, pos, TOML_WS)
    pos, value = parse_value(src, pos, parse_float)
    return pos, key, value


def parse_key(src: str, pos: Pos) -> Tuple[Pos, Key]:
    pos, key_part = parse_key_part(src, pos)
    key = [key_part]
    pos = skip_chars(src, pos, TOML_WS)
    while True:
        try:
            char: Optional[str] = src[pos]
        except IndexError:
            char = None
        if char != ".":
            return pos, tuple(key)
        pos += 1
        pos = skip_chars(src, pos, TOML_WS)
        pos, key_part = parse_key_part(src, pos)
        key.append(key_part)
        pos = skip_chars(src, pos, TOML_WS)


def parse_key_part(src: str, pos: Pos) -> Tuple[Pos, str]:
    try:
        char: Optional[str] = src[pos]
    except IndexError:
        char = None
    if char in BARE_KEY_CHARS:
        start_pos = pos
        pos = skip_chars(src, pos, BARE_KEY_CHARS)
        return pos, src[start_pos:pos]
    if char == "'":
        return parse_literal_str(src, pos)
    if char == '"':
        return parse_one_line_basic_str(src, pos)
    raise suffixed_err(src, pos, "Invalid initial character for a key part")


def parse_one_line_basic_str(src: str, pos: Pos) -> Tuple[Pos, str]:
    pos += 1
    return parse_basic_str(src, pos, multiline=False)


def parse_array(src: str, pos: Pos, parse_float: ParseFloat) -> Tuple[Pos, list]:
    pos += 1
    array: list = []

    pos = skip_comments_and_array_ws(src, pos)
    if src[pos : pos + 1] == "]":
        return pos + 1, array
    while True:
        pos, val = parse_value(src, pos, parse_float)
        array.append(val)
        pos = skip_comments_and_array_ws(src, pos)

        c = src[pos : pos + 1]
        if c == "]":
            return pos + 1, array
        if c != ",":
            raise suffixed_err(src, pos, "Unclosed array")
        pos += 1

        pos = skip_comments_and_array_ws(src, pos)
        if src[pos : pos + 1] == "]":
            return pos + 1, array


def parse_inline_table(src: str, pos: Pos, parse_float: ParseFloat) -> Tuple[Pos, dict]:
    pos += 1
    nested_dict = NestedDict()
    flags = Flags()

    pos = skip_chars(src, pos, TOML_WS)
    if src[pos : pos + 1] == "}":
        return pos + 1, nested_dict.dict
    while True:
        pos, key, value = parse_key_value_pair(src, pos, parse_float)
        key_parent, key_stem = key[:-1], key[-1]
        if flags.is_(key, Flags.FROZEN):
            raise suffixed_err(src, pos, f"Can not mutate immutable namespace {key}")
        try:
            nest = nested_dict.get_or_create_nest(key_parent, access_lists=False)
        except KeyError:
            raise suffixed_err(src, pos, "Can not overwrite a value")
        if key_stem in nest:
            raise suffixed_err(src, pos, f'Duplicate inline table key "{key_stem}"')
        nest[key_stem] = value
        pos = skip_chars(src, pos, TOML_WS)
        c = src[pos : pos + 1]
        if c == "}":
            return pos + 1, nested_dict.dict
        if c != ",":
            raise suffixed_err(src, pos, "Unclosed inline table")
        if isinstance(value, (dict, list)):
            flags.set(key, Flags.FROZEN, recursive=True)
        pos += 1
        pos = skip_chars(src, pos, TOML_WS)


def parse_basic_str_escape(
    src: str, pos: Pos, *, multiline: bool = False
) -> Tuple[Pos, str]:
    escape_id = src[pos : pos + 2]
    pos += 2
    if multiline and escape_id in {"\\ ", "\\\t", "\\\n"}:
        # Skip whitespace until next non-whitespace character or end of
        # the doc. Error if non-whitespace is found before newline.
        if escape_id != "\\\n":
            pos = skip_chars(src, pos, TOML_WS)
            char = src[pos : pos + 1]
            if not char:
                return pos, ""
            if char != "\n":
                raise suffixed_err(src, pos, 'Unescaped "\\" in a string')
            pos += 1
        pos = skip_chars(src, pos, TOML_WS_AND_NEWLINE)
        return pos, ""
    if escape_id == "\\u":
        return parse_hex_char(src, pos, 4)
    if escape_id == "\\U":
        return parse_hex_char(src, pos, 8)
    try:
        return pos, BASIC_STR_ESCAPE_REPLACEMENTS[escape_id]
    except KeyError:
        if len(escape_id) != 2:
            raise suffixed_err(src, pos, "Unterminated string")
        raise suffixed_err(src, pos, 'Unescaped "\\" in a string')


def parse_basic_str_escape_multiline(src: str, pos: Pos) -> Tuple[Pos, str]:
    return parse_basic_str_escape(src, pos, multiline=True)


def parse_hex_char(src: str, pos: Pos, hex_len: int) -> Tuple[Pos, str]:
    hex_str = src[pos : pos + hex_len]
    if len(hex_str) != hex_len or any(c not in string.hexdigits for c in hex_str):
        raise suffixed_err(src, pos, "Invalid hex value")
    pos += hex_len
    hex_int = int(hex_str, 16)
    if not is_unicode_scalar_value(hex_int):
        raise suffixed_err(src, pos, "Escaped character is not a Unicode scalar value")
    return pos, chr(hex_int)


def parse_literal_str(src: str, pos: Pos) -> Tuple[Pos, str]:
    pos += 1  # Skip starting apostrophe
    start_pos = pos
    pos = skip_until(
        src, pos, "'", error_on=ILLEGAL_LITERAL_STR_CHARS, error_on_eof=True
    )
    return pos + 1, src[start_pos:pos]  # Skip ending apostrophe


def parse_multiline_str(src: str, pos: Pos, *, literal: bool) -> Tuple[Pos, str]:
    pos += 3
    if src[pos : pos + 1] == "\n":
        pos += 1

    if literal:
        delim = "'"
        end_pos = skip_until(
            src,
            pos,
            "'''",
            error_on=ILLEGAL_MULTILINE_LITERAL_STR_CHARS,
            error_on_eof=True,
        )
        result = src[pos:end_pos]
        pos = end_pos + 3
    else:
        delim = '"'
        pos, result = parse_basic_str(src, pos, multiline=True)

    # Add at maximum two extra apostrophes/quotes if the end sequence
    # is 4 or 5 chars long instead of just 3.
    if src[pos : pos + 1] != delim:
        return pos, result
    pos += 1
    if src[pos : pos + 1] != delim:
        return pos, result + delim
    pos += 1
    return pos, result + (delim * 2)


def parse_basic_str(src: str, pos: Pos, *, multiline: bool) -> Tuple[Pos, str]:
    if multiline:
        error_on = ILLEGAL_MULTILINE_BASIC_STR_CHARS
        parse_escapes = parse_basic_str_escape_multiline
    else:
        error_on = ILLEGAL_BASIC_STR_CHARS
        parse_escapes = parse_basic_str_escape
    result = ""
    start_pos = pos
    while True:
        try:
            char = src[pos]
        except IndexError:
            raise suffixed_err(src, pos, "Unterminated string")
        if char == '"':
            if not multiline:
                return pos + 1, result + src[start_pos:pos]
            if src[pos + 1 : pos + 3] == '""':
                return pos + 3, result + src[start_pos:pos]
            pos += 1
            continue
        if char == "\\":
            result += src[start_pos:pos]
            pos, parsed_escape = parse_escapes(src, pos)
            result += parsed_escape
            start_pos = pos
            continue
        if char in error_on:
            raise suffixed_err(src, pos, f'Illegal character "{char!r}"')
        pos += 1


def parse_regex(src: str, pos: Pos, regex: "Pattern") -> Tuple[Pos, str]:
    match = regex.match(src, pos)
    if not match:
        raise suffixed_err(src, pos, "Unexpected sequence")
    return match.end(), match.group()


def parse_value(  # noqa: C901
    src: str, pos: Pos, parse_float: ParseFloat
) -> Tuple[Pos, Any]:
    try:
        char: Optional[str] = src[pos]
    except IndexError:
        char = None

    # Basic strings
    if char == '"':
        if src[pos + 1 : pos + 3] == '""':
            return parse_multiline_str(src, pos, literal=False)
        return parse_one_line_basic_str(src, pos)

    # Literal strings
    if char == "'":
        if src[pos + 1 : pos + 3] == "''":
            return parse_multiline_str(src, pos, literal=True)
        return parse_literal_str(src, pos)

    # Booleans
    if char == "t":
        if src[pos + 1 : pos + 4] == "rue":
            return pos + 4, True
    if char == "f":
        if src[pos + 1 : pos + 5] == "alse":
            return pos + 5, False

    # Dates and times
    datetime_match = RE_DATETIME.match(src, pos)
    if datetime_match:
        try:
            datetime_obj = match_to_datetime(datetime_match)
        except ValueError:
            raise suffixed_err(src, pos, "Invalid date or datetime")
        return datetime_match.end(), datetime_obj
    localtime_match = RE_LOCALTIME.match(src, pos)
    if localtime_match:
        return localtime_match.end(), match_to_localtime(localtime_match)

    # Non-decimal integers
    if char == "0":
        second_char = src[pos + 1 : pos + 2]
        if second_char == "x":
            pos, hex_str = parse_regex(src, pos + 2, RE_HEX)
            return pos, int(hex_str, 16)
        if second_char == "o":
            pos, oct_str = parse_regex(src, pos + 2, RE_OCT)
            return pos, int(oct_str, 8)
        if second_char == "b":
            pos, bin_str = parse_regex(src, pos + 2, RE_BIN)
            return pos, int(bin_str, 2)

    # Decimal integers and "normal" floats.
    # The regex will greedily match any type starting with a decimal
    # char, so needs to be located after handling of non-decimal ints,
    # and dates and times.
    number_match = RE_NUMBER.match(src, pos)
    if number_match:
        return number_match.end(), match_to_number(number_match, parse_float)

    # Arrays
    if char == "[":
        return parse_array(src, pos, parse_float)

    # Inline tables
    if char == "{":
        return parse_inline_table(src, pos, parse_float)

    # Special floats
    first_three = src[pos : pos + 3]
    if first_three in {"inf", "nan"}:
        return pos + 3, parse_float(first_three)
    first_four = src[pos : pos + 4]
    if first_four in {"-inf", "+inf", "-nan", "+nan"}:
        return pos + 4, parse_float(first_four)

    raise suffixed_err(src, pos, "Invalid value")


def suffixed_err(src: str, pos: Pos, msg: str) -> TOMLDecodeError:
    """Return a `TOMLDecodeError` where error message is suffixed with
    coordinates in source."""

    def coord_repr(src: str, pos: Pos) -> str:
        if pos >= len(src):
            return "end of document"
        line = src.count("\n", 0, pos) + 1
        if line == 1:
            column = pos + 1
        else:
            column = pos - src.rindex("\n", 0, pos)
        return f"line {line}, column {column}"

    return TOMLDecodeError(f"{msg} (at {coord_repr(src, pos)})")


def is_unicode_scalar_value(codepoint: int) -> bool:
    return (0 <= codepoint <= 55295) or (57344 <= codepoint <= 1114111)
site-packages/pip/_vendor/__pycache__/six.cpython-39.pyc000064400000065633150564745010017200 0ustar00a

��Re���E@sRdZddlmZddlZddlZddlZddlZddlZdZdZ	ej
ddkZej
ddkZej
dd�dkZ
er�efZefZefZeZeZejZn|efZeefZeejfZeZeZej�d	�r�ed
�ZnFGdd�de�Z ze!e ��Wne"�yed
�ZYn
0ed
�Z[ e
�r$ddl#m$Z$ndZ$dd�Z%dd�Z&Gdd�de�Z'Gdd�de'�Z(Gdd�dej)�Z*Gdd�de'�Z+Gdd�de�Z,e,e-�Z.Gdd�de*�Z/e+ddd d!�e+d"d#d$d%d"�e+d&d#d#d'd&�e+d(d)d$d*d(�e+d+d)d,�e+d-d#d$d.d-�e+d/d0d0d1d/�e+d2d0d0d/d2�e+d3d4d5�e+d6d)d$d7d6�e+d8d)e
�r4d9nd:d;�e+d<d)d=�e+d>d?d@dA�e+d!d!d �e+dBdBdC�e+dDdDdC�e+dEdEdC�e+d7d)d$d7d6�e+dFd#d$dGdF�e+dHd#d#dIdH�e(d$d)�e(dJdK�e(dLdCej
dMk�r�dNndC�e(dOdP�e(dQdRdS�e(dTdUdV�e(dWdXej
dYk�r�dWndZ�e(d[d\d]�e(d^d_d`�e(dadbdc�e(dddedf�e(dgdhdi�e(djdkdl�e(dmdndo�e(dpdqdr�e(dsdtdu�e(dvdwdx�e(dydydz�e(d{d{dz�e(d|d|dz�e(d}d}d~�e(dd��e(d�d��e(d�d��e(dZd�dZ�e(d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�e-d�d��e(d�e-d�d��e(d�e-d�e-d��e(d�d�d��e(d�d�d��e(d�d�d��gBZ0ejd�k�r�e0e(d�dÃg7Z0e0D]2Z1e2e/e1j3e1�e4e1e(��r�e.�5e1d�e1j3��q�[1e0e/_0e/e-d��Z6e.�5e6dơGd�dȄd�e*�Z7e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+dAd�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d�d�dփe+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��gZ8e8D]Z1e2e7e1j3e1��q[1e8e7_0e.�5e7e-d��d�d�Gd�d�d�e*�Z9e+d�d�d��e+d�d�d��e+d�d�d��gZ:e:D]Z1e2e9e1j3e1��q�[1e:e9_0e.�5e9e-d��d�d�Gd�d�d�e*�Z;e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�d	d�d�e+�d
d�d�e+�dd�d�e+�dd�d�e+�d
d�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�g#Z<e<D]Z1e2e;e1j3e1��qV[1e<e;_0e.�5e;e-�d��d�d�G�d�d��de*�Z=e+�dd��d�e+�dd��d�e+�dd��d�e+�dd��d�gZ>e>D]Z1e2e=e1j3e1��q�[1e>e=_0e.�5e=e-�d��d�d�G�d�d ��d e*�Z?e+�d!d�d��gZ@e@D]Z1e2e?e1j3e1��	qJ[1e@e?_0e.�5e?e-�d"��d#�d$�G�d%�d&��d&ej)�ZAe.�5eAe-d���d'��d(�d)�ZB�d*�d+�ZCe�	r��d,ZD�d-ZE�d.ZF�d/ZG�d0ZH�d1ZIn$�d2ZD�d3ZE�d4ZF�d5ZG�d6ZH�d7ZIzeJZKWn eL�
yB�d8�d9�ZKYn0eKZJzeMZMWn eL�
yp�d:�d;�ZMYn0e�
r��d<�d=�ZNejOZP�d>�d?�ZQeZRn>�d@�d=�ZN�dA�dB�ZP�dC�d?�ZQG�dD�dE��dEe�ZReMZMe%eN�dF�e�SeD�ZTe�SeE�ZUe�SeF�ZVe�SeG�ZWe�SeH�ZXe�SeI�ZYe�r~�dG�dH�ZZ�dI�dJ�Z[�dK�dL�Z\�dM�dN�Z]e�^�dO�Z_e�^�dP�Z`e�^�dQ�ZanT�dR�dH�ZZ�dS�dJ�Z[�dT�dL�Z\�dU�dN�Z]e�^�dV�Z_e�^�dW�Z`e�^�dX�Zae%eZ�dY�e%e[�dZ�e%e\�d[�e%e]�d\�e�r��d]�d^�Zb�d_�d`�ZcedZeddlfZfef�g�da�jhZi[fe�jd�ZkejlZmenZoddlpZpepjqZqepjrZr[p�dbZsej
�dc�dck�r��ddZt�deZu�dfZvn�dgZt�dhZu�diZvnp�dj�d^�Zb�dk�d`�ZceeZeedZi�dl�dm�Zk�dn�do�Zme�wejxey�ZoddlqZqeqjqZqZr�dpZs�ddZt�deZu�dfZve%eb�dq�e%ec�dr��ds�db�Zz�dt�dg�Z{�du�dh�Z|�dv�di�Z}e�
r�e~e6j�dw�Z��d��dx�dy�Z�n�d��dz�d{�Z�e��d|�ej
dd��d}k�
r�e��d~�n�d�d��Z�e~e6j�d�d�Z�e�du�
r�d��d��Z�ej
dd�dMk�re�Z��d��d��Z�e%e��d��ej
dd�dk�rzej�ej�f�d��d��Z�ej�je�_ej�ej�f�d��d��Z�ej�je�_nej�Z��d��d��Z��d��d��Z��d��d��d��Z��d��d��d��Z��d��d��d��Z��d��d��Z�gZ�e-Z�e�����d��du�r�ge�_�ej��rBe�ej��D]4\Z�Z�ee��j-dk�re�j3e-k�rej�e�=�q>�q[�[�ej���e.�dS(�z6Utilities for writing code that runs on Python 2 and 3�)�absolute_importNz'Benjamin Peterson <benjamin@python.org>z1.16.0��)r��javai���c@seZdZdd�ZdS)�XcCsdS)Nl���selfrr�~/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/six.py�__len__>sz	X.__len__N)�__name__�
__module__�__qualname__rrrrrr<srl����)�spec_from_loadercCs
||_dS)z Add documentation to a function.N)�__doc__)�func�docrrr�_add_docPsrcCst|�tj|S)z7Import module, returning the module after the last dot.)�
__import__�sys�modules��namerrr�_import_moduleUsrc@seZdZdd�Zdd�ZdS)�
_LazyDescrcCs
||_dS�Nr�r
rrrr�__init__]sz_LazyDescr.__init__cCs@|��}t||j|�zt|j|j�Wnty:Yn0|Sr)�_resolve�setattrr�delattr�	__class__�AttributeError)r
�obj�tp�resultrrr�__get__`sz_LazyDescr.__get__N)r
rrrr'rrrrrrcs.eZdZd�fdd�	Zdd�Zdd�Z�ZS)	�MovedModuleNcs2tt|��|�tr(|dur |}||_n||_dSr)�superr(r�PY3�mod)r
r�old�new�r"rrrnszMovedModule.__init__cCs
t|j�Sr)rr+r	rrrrwszMovedModule._resolvecCs"|��}t||�}t|||�|Sr)r�getattrr )r
�attr�_module�valuerrr�__getattr__zs
zMovedModule.__getattr__)N)r
rrrrr3�
__classcell__rrr.rr(ls	r(cs(eZdZ�fdd�Zdd�ZgZ�ZS)�_LazyModulecstt|��|�|jj|_dSr)r)r5rr"rrr.rrr�sz_LazyModule.__init__cCs ddg}|dd�|jD�7}|S)Nrr
cSsg|]
}|j�qSrr)�.0r0rrr�
<listcomp>��z'_LazyModule.__dir__.<locals>.<listcomp>)�_moved_attributes)r
�attrsrrr�__dir__�sz_LazyModule.__dir__)r
rrrr;r9r4rrr.rr5�sr5cs&eZdZd�fdd�	Zdd�Z�ZS)�MovedAttributeNcsdtt|��|�trH|dur |}||_|dur@|dur<|}n|}||_n||_|durZ|}||_dSr)r)r<rr*r+r0)r
r�old_mod�new_mod�old_attr�new_attrr.rrr�szMovedAttribute.__init__cCst|j�}t||j�Sr)rr+r/r0�r
�modulerrrr�s
zMovedAttribute._resolve)NN)r
rrrrr4rrr.rr<�sr<c@speZdZdZdd�Zdd�Zdd�Zdd	d
�Zddd�Zd
d�Z	dd�Z
dd�Zdd�ZeZ
dd�Zdd�ZdS)�_SixMetaPathImporterz�
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    cCs||_i|_dSr)r�
known_modules)r
�six_module_namerrrr�sz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdS�N�.�rDr)r
r+�	fullnames�fullnamerrr�_add_module�sz _SixMetaPathImporter._add_modulecCs|j|jd|SrFrH�r
rJrrr�_get_module�sz _SixMetaPathImporter._get_moduleNcCs||jvr|SdSr)rD)r
rJ�pathrrr�find_module�s
z _SixMetaPathImporter.find_modulecCs||jvrt||�SdSr)rDr)r
rJrN�targetrrr�	find_spec�s

z_SixMetaPathImporter.find_speccCs0z|j|WSty*td|��Yn0dS)Nz!This loader does not know module )rD�KeyError�ImportErrorrLrrr�__get_module�sz!_SixMetaPathImporter.__get_modulecCsRztj|WStyYn0|�|�}t|t�r>|��}n||_|tj|<|Sr)rrrR� _SixMetaPathImporter__get_module�
isinstancer(r�
__loader__)r
rJr+rrr�load_module�s



z _SixMetaPathImporter.load_modulecCst|�|�d�S)z�
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        �__path__)�hasattrrUrLrrr�
is_package�sz_SixMetaPathImporter.is_packagecCs|�|�dS)z;Return None

        Required, if is_package is implementedN)rUrLrrr�get_code�s
z_SixMetaPathImporter.get_codecCs|�|j�Sr)rXr)r
�specrrr�
create_module�sz"_SixMetaPathImporter.create_modulecCsdSrrrArrr�exec_module�sz _SixMetaPathImporter.exec_module)N)N)r
rrrrrKrMrOrQrUrXr[r\�
get_sourcer^r_rrrrrC�s

	rCc@seZdZdZgZdS)�_MovedItemszLazy loading of moved objectsN)r
rrrrYrrrrra�sra�	cStringIO�io�StringIO�filter�	itertools�builtins�ifilter�filterfalse�ifilterfalse�input�__builtin__�	raw_input�internr�map�imap�getcwd�osZgetcwdu�getcwdb�	getoutput�commands�
subprocess�range�xrange�
reload_module�	importlib�imp�reload�reduce�	functools�shlex_quoteZpipes�shlex�quote�UserDict�collections�UserList�
UserString�zip�izip�zip_longest�izip_longest�configparser�ConfigParser�collections_abc)rrzcollections.abc�copyreg�copy_reg�dbm_gnu�gdbmzdbm.gnu�dbm_ndbm�dbmzdbm.ndbm�
_dummy_thread�dummy_thread)r�	�_thread�http_cookiejar�	cookielibzhttp.cookiejar�http_cookies�Cookiezhttp.cookies�
html_entities�htmlentitydefsz
html.entities�html_parser�
HTMLParserzhtml.parser�http_client�httplibzhttp.client�email_mime_basezemail.MIMEBasezemail.mime.base�email_mime_imagezemail.MIMEImagezemail.mime.image�email_mime_multipartzemail.MIMEMultipartzemail.mime.multipart�email_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipart�email_mime_textzemail.MIMETextzemail.mime.text�BaseHTTPServerzhttp.server�
CGIHTTPServer�SimpleHTTPServer�cPickle�pickle�queue�Queue�reprlib�repr�socketserver�SocketServer�thread�tkinter�Tkinter�tkinter_dialog�Dialogztkinter.dialog�tkinter_filedialog�
FileDialogztkinter.filedialog�tkinter_scrolledtext�ScrolledTextztkinter.scrolledtext�tkinter_simpledialog�SimpleDialogztkinter.simpledialog�tkinter_tix�Tixztkinter.tix�tkinter_ttk�ttkztkinter.ttk�tkinter_constants�Tkconstantsztkinter.constants�tkinter_dnd�Tkdndztkinter.dnd�tkinter_colorchooser�tkColorChooserztkinter.colorchooser�tkinter_commondialog�tkCommonDialogztkinter.commondialog�tkinter_tkfiledialog�tkFileDialog�tkinter_font�tkFontztkinter.font�tkinter_messagebox�tkMessageBoxztkinter.messagebox�tkinter_tksimpledialog�tkSimpleDialog�urllib_parsez.moves.urllib_parsezurllib.parse�urllib_errorz.moves.urllib_errorzurllib.error�urllibz
.moves.urllib�urllib_robotparser�robotparserzurllib.robotparser�
xmlrpc_client�	xmlrpclibz
xmlrpc.client�
xmlrpc_server�SimpleXMLRPCServerz
xmlrpc.server�win32�winreg�_winregzmoves.z.moves�movesc@seZdZdZdS)�Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseN�r
rrrrrrrr�Tsr��ParseResult�urlparse�SplitResult�parse_qs�	parse_qsl�	urldefrag�urljoin�urlsplit�
urlunparse�
urlunsplit�
quote_plus�unquote�unquote_plus�unquote_to_bytes�	urlencode�
splitquery�splittag�	splituser�
splitvalue�
uses_fragment�uses_netloc�uses_params�
uses_query�
uses_relative�moves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)�Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNr�rrrrr�~sr��URLError�urllib2�	HTTPError�ContentTooShortErrorz.moves.urllib.error�moves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)�Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNr�rrrrr��sr��urlopenzurllib.request�install_opener�build_opener�pathname2url�url2pathname�
getproxies�Request�OpenerDirector�HTTPDefaultErrorHandler�HTTPRedirectHandler�HTTPCookieProcessor�ProxyHandler�BaseHandler�HTTPPasswordMgr�HTTPPasswordMgrWithDefaultRealm�AbstractBasicAuthHandler�HTTPBasicAuthHandler�ProxyBasicAuthHandler�AbstractDigestAuthHandler�HTTPDigestAuthHandler�ProxyDigestAuthHandler�HTTPHandler�HTTPSHandler�FileHandler�
FTPHandler�CacheFTPHandler�UnknownHandler�HTTPErrorProcessor�urlretrieve�
urlcleanup�	URLopener�FancyURLopener�proxy_bypass�parse_http_list�parse_keqv_listz.moves.urllib.request�moves.urllib_requestzmoves.urllib.requestc@seZdZdZdS)� Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNr�rrrrr#�sr#�addbasezurllib.response�addclosehook�addinfo�
addinfourlz.moves.urllib.response�moves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)�#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNr�rrrrr)�sr)�RobotFileParserz.moves.urllib.robotparser�moves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZe�d�Ze�d�Ze�d�Z	e�d�Z
e�d�Zdd�Zd	S)
�Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacer�r�r"r(r+cCsgd�S)N)�parse�error�request�responser�rr	rrrr;�szModule_six_moves_urllib.__dir__N)
r
rrrrY�	_importerrMr-r.r/r0r�r;rrrrr,�s




r,zmoves.urllibcCstt|j|�dS)zAdd an item to six.moves.N)r rar)�moverrr�add_move�sr3cCsTztt|�Wn@tyNztj|=Wn tyHtd|f��Yn0Yn0dS)zRemove item from six.moves.zno such move, %rN)r!rar#r��__dict__rRrrrr�remove_movesr5�__func__�__self__�__closure__�__code__�__defaults__�__globals__�im_funcZim_selfZfunc_closure�	func_codeZ
func_defaultsZfunc_globalscCs|��Sr)�next)�itrrr�advance_iterator#sr@cCstdd�t|�jD��S)Ncss|]}d|jvVqdS)�__call__N)r4)r6�klassrrr�	<genexpr>,r8zcallable.<locals>.<genexpr>)�any�type�__mro__)r$rrr�callable+srGcCs|Srr��unboundrrr�get_unbound_function0srJcCs|Srr�r�clsrrr�create_unbound_method5srMcCs|jSr)r<rHrrrrJ:scCst�|||j�Sr)�types�
MethodTyper")rr$rrr�create_bound_method=srPcCst�|d|�Sr)rNrOrKrrrrM@sc@seZdZdd�ZdS)�IteratorcCst|��|�Sr)rE�__next__r	rrrr>Esz
Iterator.nextN)r
rrr>rrrrrQCsrQz3Get the function out of a possibly unbound functioncKst|jfi|���Sr)�iter�keys��d�kwrrr�iterkeysVsrXcKst|jfi|���Sr)rS�valuesrUrrr�
itervaluesYsrZcKst|jfi|���Sr)rS�itemsrUrrr�	iteritems\sr\cKst|jfi|���Sr)rS�listsrUrrr�	iterlists_sr^rTrYr[cKs|jfi|��Sr)rXrUrrrrXhscKs|jfi|��Sr)rZrUrrrrZkscKs|jfi|��Sr)r\rUrrrr\nscKs|jfi|��Sr)r^rUrrrr^qs�viewkeys�
viewvalues�	viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs
|�d�S)Nzlatin-1)�encode��srrr�b�srecCs|Srrrcrrr�u�srfz>B�assertCountEqual�ZassertRaisesRegexpZassertRegexpMatchesZassertNotRegexpMatches�assertRaisesRegex�assertRegex�assertNotRegexcCs|Srrrcrrrre�scCst|�dd�d�S)Nz\\z\\\\�unicode_escape)�unicode�replacercrrrrf�scCst|d�S)Nr��ord)�bsrrr�byte2int�srrcCst||�Srro)�buf�irrr�
indexbytes�sruZassertItemsEqualzByte literalzText literalcOst|t�|i|��Sr)r/�_assertCountEqual�r
�args�kwargsrrrrg�scOst|t�|i|��Sr)r/�_assertRaisesRegexrwrrrri�scOst|t�|i|��Sr)r/�_assertRegexrwrrrrj�scOst|t�|i|��Sr)r/�_assertNotRegexrwrrrrk�s�execcCsBz2|dur|�}|j|ur$|�|��|�Wd}d}n
d}d}0dSr)�
__traceback__�with_traceback)r%r2�tbrrr�reraise�s

�r�cCsB|dur*t�d�}|j}|dur&|j}~n|dur6|}td�dS)zExecute code in a namespace.Nrhzexec _code_ in _globs_, _locs_)r�	_getframe�	f_globals�f_localsr})Z_code_Z_globs_Z_locs_�framerrr�exec_�s
r�zedef reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
)rzrdef raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
cCs|�dSrr)r2�
from_valuerrr�
raise_from�sr��printc
s.|�dtj���durdS�fdd�}d}|�dd�}|dur`t|t�rNd}nt|t�s`td��|�d	d�}|dur�t|t�r�d}nt|t�s�td
��|r�td��|s�|D]}t|t�r�d}q�q�|r�td�}td
�}nd}d
}|dur�|}|dur�|}t|�D] \}	}|	�r||�||��q||�dS)z4The new-style print function for Python 2.4 and 2.5.�fileNcsdt|t�st|�}t�t�rVt|t�rV�jdurVt�dd�}|durHd}|��j|�}��|�dS)N�errors�strict)	rV�
basestring�strr�rm�encodingr/rb�write)�datar���fprrr��s

��zprint_.<locals>.writeF�sepTzsep must be None or a string�endzend must be None or a stringz$invalid keyword arguments to print()�
� )�popr�stdoutrVrmr��	TypeError�	enumerate)
rxryr�Zwant_unicoder�r��arg�newline�spacertrr�r�print_�sL





r�cOs@|�dtj�}|�dd�}t|i|��|r<|dur<|��dS)Nr��flushF)�getrr�r��_printr�)rxryr�r�rrrr�/s
zReraise an exception.c	Csh|D]6}zt||�}Wnty,YqYq0t|||�q|D]}t||��t||i��q@||_|Sr)r/r#r �update�__wrapped__)�wrapper�wrapped�assigned�updatedr0r2rrr�_update_wrapper>s
r�cCstjt|||d�S)N�r�r�r�)r~�partialr�r�rrr�wrapsNs�r�cs&G��fdd�dt�}t�|ddi�S)z%Create a base class with a metaclass.cs,eZdZ��fdd�Ze��fdd��ZdS)z!with_metaclass.<locals>.metaclasscs>tjdd�dkr.t���}|�ur2�|d<n�}�|||�S)Nr)r��__orig_bases__)r�version_inforN�
resolve_bases)rLr�
this_basesrV�resolved_bases��bases�metarr�__new___s

z)with_metaclass.<locals>.metaclass.__new__cs��|��Sr)�__prepare__)rLrr�r�rrr�jsz-with_metaclass.<locals>.metaclass.__prepare__N)r
rrr��classmethodr�rr�rr�	metaclass]sr��temporary_classr)rEr�)r�r�r�rr�r�with_metaclassXsr�cs�fdd�}|S)z6Class decorator for creating a class with a metaclass.cs||j��}|�d�}|dur@t|t�r,|g}|D]}|�|�q0|�dd�|�dd�t|d�rl|j|d<�|j|j	|�S)N�	__slots__r4�__weakref__r)
r4�copyr�rVr�r�rZrr
�	__bases__)rL�	orig_vars�slots�	slots_var�r�rrr�rs




zadd_metaclass.<locals>.wrapperr)r�r�rr�r�
add_metaclassps
r��utf-8r�cCs8t|t�r|St|t�r$|�||�Stdt|���dS)z�Coerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    �not expecting type '%s'N)rV�binary_type�	text_typerbr�rE�rdr�r�rrr�
ensure_binary�s


r�cCsft|�tur|Str*t|t�r*|�||�StrDt|t�rD|�||�St|ttf�sbt	dt|���|S)z�Coerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r�)
rEr��PY2rVr�rbr*r��decoder�r�rrr�
ensure_str�sr�cCs8t|t�r|�||�St|t�r$|Stdt|���dS)z�Coerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r�N)rVr�r�r�r�rEr�rrr�ensure_text�s


r�cCs2tr.d|jvrtd|j��|j|_dd�|_|S)a
    A class decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    �__str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|���d�S)Nr�)�__unicode__rbr	rrr�<lambda>�r8z-python_2_unicode_compatible.<locals>.<lambda>)r�r4�
ValueErrorr
r�r�)rBrrr�python_2_unicode_compatible�s
�
r��__spec__)N)NN)r�r�)r�r�)r�r�)�r�
__future__rr~rf�operatorrrN�
__author__�__version__r�r�r*�PY34r��string_types�int�
integer_typesrE�class_typesr��bytesr��maxsize�MAXSIZEr��longZ	ClassTyperm�platform�
startswith�objectr�len�
OverflowError�importlib.utilrrrrr(�
ModuleTyper5r<rCr
r1rar9r0r rrVrKr�r��_urllib_parse_moved_attributesr��_urllib_error_moved_attributesr�� _urllib_request_moved_attributesr#�!_urllib_response_moved_attributesr)�$_urllib_robotparser_moved_attributesr,r3r5�
_meth_func�
_meth_self�
_func_closure�
_func_code�_func_defaults�
_func_globalsr>r@�	NameErrorrGrJrOrPrMrQ�
attrgetter�get_method_function�get_method_self�get_function_closure�get_function_code�get_function_defaults�get_function_globalsrXrZr\r^�methodcallerr_r`rarerf�chr�unichr�struct�Struct�pack�int2byte�
itemgetterrr�getitemrurS�	iterbytesrcrd�BytesIOrvrzr{r|r�rprprgrirjrkr/rgr�r�r�r�r��WRAPPER_ASSIGNMENTS�WRAPPER_UPDATESr��update_wrapperr�r�r�r�r�r�r�rY�__package__�globalsr�r��submodule_search_locations�	meta_pathr�rt�importer�appendrrrr�<module>s

I































��


�


�H�























��


��
















�%�����
��





��



5�
��
site-packages/pip/_vendor/__pycache__/distro.cpython-39.pyc000064400000113005150564745010017664 0ustar00a

��Re��@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	dZ
ej�dd�Zej�dd�ZdZd	d
iZd
d
dddd�Zd
diZ e�!d�Z"e�!d�Z#dddedddfZ$dFdd�Z%dd�Z&dGdd�Z'dHdd�Z(dId d!�Z)dJd"d#�Z*dKd$d%�Z+dLd&d'�Z,d(d)�Z-d*d+�Z.dMd,d-�Z/d.d/�Z0d0d1�Z1d2d3�Z2d4d5�Z3d6d7�Z4d8d9�Z5d:d;�Z6d<d=�Z7zdd>l8m9Z9Wn$e:�y�Gd?d@�d@e;�Z9Yn0GdAdB�dBe;�Z<e<�Z=dCdD�Z>e?dEk�r�e>�dS)Na�
The ``distro`` package (``distro`` stands for Linux Distribution) provides
information about the Linux distribution it runs on, such as a reliable
machine-readable distro ID, or version information.

It is the recommended replacement for Python's original
:py:func:`platform.linux_distribution` function, but it provides much more
functionality. An alternative implementation became necessary because Python
3.5 deprecated this function, and Python 3.8 removed it altogether. Its
predecessor function :py:func:`platform.dist` was already deprecated since
Python 2.6 and removed in Python 3.8. Still, there are many cases in which
access to OS distribution information is needed. See `Python issue 1322
<https://bugs.python.org/issue1322>`_ for more information.
�Nz1.6.0ZUNIXCONFDIRz/etcZ
UNIXUSRLIBDIRz/usr/libz
os-release�ol�oracle�rhel)�enterpriseenterpriseas�enterpriseenterpriseserver�redhatenterpriseworkstation�redhatenterpriseserver�redhatenterprisecomputenode�redhatzA(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)z(\w+)[-_](release|version)$�debian_versionzlsb-releasezoem-releasezsystem-releasez
plesk-releaseziredmail-releaseTcCstjdtdd�t�|�S)aG
    .. deprecated:: 1.6.0

        :func:`distro.linux_distribution()` is deprecated. It should only be
        used as a compatibility shim with Python's
        :py:func:`platform.linux_distribution()`. Please use :func:`distro.id`,
        :func:`distro.version` and :func:`distro.name` instead.

    Return information about the current OS distribution as a tuple
    ``(id_name, version, codename)`` with items as follows:

    * ``id_name``:  If *full_distribution_name* is false, the result of
      :func:`distro.id`. Otherwise, the result of :func:`distro.name`.

    * ``version``:  The result of :func:`distro.version`.

    * ``codename``:  The result of :func:`distro.codename`.

    The interface of this function is compatible with the original
    :py:func:`platform.linux_distribution` function, supporting a subset of
    its parameters.

    The data it returns may not exactly be the same, because it uses more data
    sources than the original function, and that may lead to different data if
    the OS distribution is not consistent across multiple data sources it
    provides (there are indeed such distributions ...).

    Another reason for differences is the fact that the :func:`distro.id`
    method normalizes the distro ID string to a reliable machine-readable value
    for a number of popular OS distributions.
    z�distro.linux_distribution() is deprecated. It should only be used as a compatibility shim with Python's platform.linux_distribution(). Please use distro.id(), distro.version() and distro.name() instead.�)�
stacklevel)�warnings�warn�DeprecationWarning�_distro�linux_distribution)�full_distribution_name�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distro.pyr�s!�rcCst��S)aD
    Return the distro ID of the current distribution, as a
    machine-readable string.

    For a number of OS distributions, the returned distro ID value is
    *reliable*, in the sense that it is documented and that it does not change
    across releases of the distribution.

    This package maintains the following reliable distro ID values:

    ==============  =========================================
    Distro ID       Distribution
    ==============  =========================================
    "ubuntu"        Ubuntu
    "debian"        Debian
    "rhel"          RedHat Enterprise Linux
    "centos"        CentOS
    "fedora"        Fedora
    "sles"          SUSE Linux Enterprise Server
    "opensuse"      openSUSE
    "amazon"        Amazon Linux
    "arch"          Arch Linux
    "cloudlinux"    CloudLinux OS
    "exherbo"       Exherbo Linux
    "gentoo"        GenToo Linux
    "ibm_powerkvm"  IBM PowerKVM
    "kvmibm"        KVM for IBM z Systems
    "linuxmint"     Linux Mint
    "mageia"        Mageia
    "mandriva"      Mandriva Linux
    "parallels"     Parallels
    "pidora"        Pidora
    "raspbian"      Raspbian
    "oracle"        Oracle Linux (and Oracle Enterprise Linux)
    "scientific"    Scientific Linux
    "slackware"     Slackware
    "xenserver"     XenServer
    "openbsd"       OpenBSD
    "netbsd"        NetBSD
    "freebsd"       FreeBSD
    "midnightbsd"   MidnightBSD
    ==============  =========================================

    If you have a need to get distros for reliable IDs added into this set,
    or if you find that the :func:`distro.id` function returns a different
    distro ID for one of the listed distros, please create an issue in the
    `distro issue tracker`_.

    **Lookup hierarchy and transformations:**

    First, the ID is obtained from the following sources, in the specified
    order. The first available and non-empty value is used:

    * the value of the "ID" attribute of the os-release file,

    * the value of the "Distributor ID" attribute returned by the lsb_release
      command,

    * the first part of the file name of the distro release file,

    The so determined ID value then passes the following transformations,
    before it is returned by this method:

    * it is translated to lower case,

    * blanks (which should not be there anyway) are translated to underscores,

    * a normalization of the ID is performed, based upon
      `normalization tables`_. The purpose of this normalization is to ensure
      that the ID is as reliable as possible, even across incompatible changes
      in the OS distributions. A common reason for an incompatible change is
      the addition of an os-release file, or the addition of the lsb_release
      command, with ID values that differ from what was previously determined
      from the distro release file name.
    )r�idrrrrr�sMrFcCs
t�|�S)ak
    Return the name of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the name is returned without version or codename.
    (e.g. "CentOS Linux")

    If *pretty* is true, the version and codename are appended.
    (e.g. "CentOS Linux 7.1.1503 (Core)")

    **Lookup hierarchy:**

    The name is obtained from the following sources, in the specified order.
    The first available and non-empty value is used:

    * If *pretty* is false:

      - the value of the "NAME" attribute of the os-release file,

      - the value of the "Distributor ID" attribute returned by the lsb_release
        command,

      - the value of the "<name>" field of the distro release file.

    * If *pretty* is true:

      - the value of the "PRETTY_NAME" attribute of the os-release file,

      - the value of the "Description" attribute returned by the lsb_release
        command,

      - the value of the "<name>" field of the distro release file, appended
        with the value of the pretty version ("<version_id>" and "<codename>"
        fields) of the distro release file, if available.
    )r�name��prettyrrrrs%rcCst�||�S)av
    Return the version of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the version is returned without codename (e.g.
    "7.0").

    If *pretty* is true, the codename in parenthesis is appended, if the
    codename is non-empty (e.g. "7.0 (Maipo)").

    Some distributions provide version numbers with different precisions in
    the different sources of distribution information. Examining the different
    sources in a fixed priority order does not always yield the most precise
    version (e.g. for Debian 8.2, or CentOS 7.1).

    The *best* parameter can be used to control the approach for the returned
    version:

    If *best* is false, the first non-empty version number in priority order of
    the examined sources is returned.

    If *best* is true, the most precise version number out of all examined
    sources is returned.

    **Lookup hierarchy:**

    In all cases, the version number is obtained from the following sources.
    If *best* is false, this order represents the priority order:

    * the value of the "VERSION_ID" attribute of the os-release file,
    * the value of the "Release" attribute returned by the lsb_release
      command,
    * the version number parsed from the "<version_id>" field of the first line
      of the distro release file,
    * the version number parsed from the "PRETTY_NAME" attribute of the
      os-release file, if it follows the format of the distro release files.
    * the version number parsed from the "Description" attribute returned by
      the lsb_release command, if it follows the format of the distro release
      files.
    )r�version�r�bestrrrr+s*rcCs
t�|�S)a�
    Return the version of the current OS distribution as a tuple
    ``(major, minor, build_number)`` with items as follows:

    * ``major``:  The result of :func:`distro.major_version`.

    * ``minor``:  The result of :func:`distro.minor_version`.

    * ``build_number``:  The result of :func:`distro.build_number`.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r�
version_parts�rrrrrXsrcCs
t�|�S)a5
    Return the major version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The major version is the first
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r�
major_versionrrrrrjsrcCs
t�|�S)a6
    Return the minor version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The minor version is the second
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r�
minor_versionrrrrr xsr cCs
t�|�S)a3
    Return the build number of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The build number is the third part
    of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    )r�build_numberrrrrr!�sr!cCst��S)a
    Return a space-separated list of distro IDs of distributions that are
    closely related to the current OS distribution in regards to packaging
    and programming interfaces, for example distributions the current
    distribution is a derivative from.

    **Lookup hierarchy:**

    This information item is only provided by the os-release file.
    For details, see the description of the "ID_LIKE" attribute in the
    `os-release man page
    <http://www.freedesktop.org/software/systemd/man/os-release.html>`_.
    )r�likerrrrr"�sr"cCst��S)a�
    Return the codename for the release of the current OS distribution,
    as a string.

    If the distribution does not have a codename, an empty string is returned.

    Note that the returned codename is not always really a codename. For
    example, openSUSE returns "x86_64". This function does not handle such
    cases in any special way and just returns the string it finds, if any.

    **Lookup hierarchy:**

    * the codename within the "VERSION" attribute of the os-release file, if
      provided,

    * the value of the "Codename" attribute returned by the lsb_release
      command,

    * the value of the "<codename>" field of the distro release file.
    )r�codenamerrrrr#�sr#cCst�||�S)a�
    Return certain machine-readable information items about the current OS
    distribution in a dictionary, as shown in the following example:

    .. sourcecode:: python

        {
            'id': 'rhel',
            'version': '7.0',
            'version_parts': {
                'major': '7',
                'minor': '0',
                'build_number': ''
            },
            'like': 'fedora',
            'codename': 'Maipo'
        }

    The dictionary structure and keys are always the same, regardless of which
    information items are available in the underlying data sources. The values
    for the various keys are as follows:

    * ``id``:  The result of :func:`distro.id`.

    * ``version``:  The result of :func:`distro.version`.

    * ``version_parts -> major``:  The result of :func:`distro.major_version`.

    * ``version_parts -> minor``:  The result of :func:`distro.minor_version`.

    * ``version_parts -> build_number``:  The result of
      :func:`distro.build_number`.

    * ``like``:  The result of :func:`distro.like`.

    * ``codename``:  The result of :func:`distro.codename`.

    For a description of the *pretty* and *best* parameters, see the
    :func:`distro.version` method.
    )r�inforrrrr$�s*r$cCst��S)z�
    Return a dictionary containing key-value pairs for the information items
    from the os-release file data source of the current OS distribution.

    See `os-release file`_ for details about these information items.
    )r�os_release_inforrrrr%�sr%cCst��S)z�
    Return a dictionary containing key-value pairs for the information items
    from the lsb_release command data source of the current OS distribution.

    See `lsb_release command output`_ for details about these information
    items.
    )r�lsb_release_inforrrrr&�s	r&cCst��S)z�
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.

    See `distro release file`_ for details about these information items.
    )r�distro_release_inforrrrr'sr'cCst��S)z�
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.
    )r�
uname_inforrrrr(sr(cCs
t�|�S)a�
    Return a single named information item from the os-release file data source
    of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `os-release file`_ for details about these information items.
    )r�os_release_attr��	attributerrrr)sr)cCs
t�|�S)a�
    Return a single named information item from the lsb_release command output
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `lsb_release command output`_ for details about these information
    items.
    )r�lsb_release_attrr*rrrr,+sr,cCs
t�|�S)a�
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `distro release file`_ for details about these information items.
    )r�distro_release_attrr*rrrr-@sr-cCs
t�|�S)aZ
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
                The empty string, if the item does not exist.
    )r�
uname_attrr*rrrr.Tsr.)�cached_propertyc@s eZdZdZdd�Zdd�ZdS)r/z�A version of @property which caches the value.  On access, it calls the
        underlying function and sets the value in `__dict__` so future accesses
        will not re-call the property.
        cCs|j|_||_dS)N)�__name__�_fname�_f)�self�frrr�__init__pszcached_property.__init__cCs2|dusJd�|j���|�|�}|j|j<|S)Nzcall {} on an instance)�formatr1r2�__dict__)r3�obj�owner�retrrr�__get__uszcached_property.__get__N)r0�
__module__�__qualname__�__doc__r5r;rrrrr/jsr/c@s>eZdZdZdDdd�Zdd�ZdEd	d
�Zdd�ZdFdd�ZdGdd�Z	dHdd�Z
dIdd�ZdJdd�ZdKdd�Z
dd�Zdd�ZdLdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zed0d1��Zed2d3��Zed4d5��Zed6d7��Zed8d9��Zed:d;��Z ed<d=��Z!ed>d?��Z"d@dA�Z#edBdC��Z$dS)M�LinuxDistributiona
    Provides information about a OS distribution.

    This package creates a private module-global instance of this class with
    default initialization arguments, that is used by the
    `consolidated accessor functions`_ and `single source accessor functions`_.
    By using default initialization arguments, that module-global instance
    returns data about the current OS distribution (i.e. the distro this
    package runs on).

    Normally, it is not necessary to create additional instances of this class.
    However, in situations where control is needed over the exact data sources
    that are used, instances of this class can be created with a specific
    distro release file, or a specific os-release file, or without invoking the
    lsb_release command.
    T�NcCs�||_|rtj�|d�nt|_|r0tj�|d�nt|_|rB||_nFtj�|jt	�}tj�|jt	�}tj�
|�sztj�
|�s�||_n||_|p�d|_||_||_
dS)a�
        The initialization method of this class gathers information from the
        available data sources, and stores that in private instance attributes.
        Subsequent access to the information items uses these private instance
        attributes, so that the data sources are read only once.

        Parameters:

        * ``include_lsb`` (bool): Controls whether the
          `lsb_release command output`_ is included as a data source.

          If the lsb_release command is not available in the program execution
          path, the data source for the lsb_release command will be empty.

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is to be used as a data source.

          An empty string (the default) will cause the default path name to
          be used (see `os-release file`_ for details).

          If the specified or defaulted os-release file does not exist, the
          data source for the os-release file will be empty.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is to be used as a data source.

          An empty string (the default) will cause a default search algorithm
          to be used (see `distro release file`_ for details).

          If the specified distro release file does not exist, or if no default
          distro release file can be found, the data source for the distro
          release file will be empty.

        * ``include_uname`` (bool): Controls whether uname command output is
          included as a data source. If the uname command is not available in
          the program execution path the data source for the uname command will
          be empty.

        * ``root_dir`` (string): The absolute path to the root directory to use
          to find distro-related information files.

        Public instance attributes:

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``include_lsb`` (bool): The result of the ``include_lsb`` parameter.
          This controls whether the lsb information will be loaded.

        * ``include_uname`` (bool): The result of the ``include_uname``
          parameter. This controls whether the uname information will
          be loaded.

        Raises:

        * :py:exc:`IOError`: Some I/O issue with an os-release file or distro
          release file.

        * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had
          some issue (other than not being available in the program execution
          path).

        * :py:exc:`UnicodeError`: A data source has unexpected characters or
          uses an unexpected encoding.
        �etczusr/libr@N)�root_dir�os�path�join�_UNIXCONFDIR�etc_dir�_UNIXUSRLIBDIR�usr_lib_dir�os_release_file�_OS_RELEASE_BASENAME�isfile�distro_release_file�include_lsb�
include_uname)r3rNrJrMrOrB�etc_dir_os_release_file�usr_lib_os_release_filerrrr5�s$O���
zLinuxDistribution.__init__cCsdj|d�S)zReturn repr of all infoaeLinuxDistribution(os_release_file={self.os_release_file!r}, distro_release_file={self.distro_release_file!r}, include_lsb={self.include_lsb!r}, include_uname={self.include_uname!r}, _os_release_info={self._os_release_info!r}, _lsb_release_info={self._lsb_release_info!r}, _distro_release_info={self._distro_release_info!r}, _uname_info={self._uname_info!r})�r3)r6rRrrr�__repr__�s��zLinuxDistribution.__repr__cCs"|r|��n|��|��|��fS)z�
        Return information about the OS distribution that is compatible
        with Python's :func:`platform.linux_distribution`, supporting a subset
        of its parameters.

        For details, see :func:`distro.linux_distribution`.
        )rrrr#)r3rrrrrs
�z$LinuxDistribution.linux_distributioncCsldd�}|�d�}|r ||t�S|�d�}|r8||t�S|�d�}|rP||t�S|�d�}|rh||t�SdS)zoReturn the distro ID of the OS distribution, as a string.

        For details, see :func:`distro.id`.
        cSs|���dd�}|�||�S)N� �_)�lower�replace�get)�	distro_id�tablerrr�	normalizesz'LinuxDistribution.id.<locals>.normalizer�distributor_idr@)r)�NORMALIZED_OS_IDr,�NORMALIZED_LSB_IDr-�NORMALIZED_DISTRO_IDr.)r3r[rYrrrrs







zLinuxDistribution.idFcCs||�d�p&|�d�p&|�d�p&|�d�}|rt|�d�p>|�d�}|st|�d�pV|�d�}|jdd�}|rt|d|}|pzdS)	zu
        Return the name of the OS distribution, as a string.

        For details, see :func:`distro.name`.
        rr\�pretty_name�descriptionTrrTr@)r)r,r-r.r)r3rrrrrrr4s"
����zLinuxDistribution.namec	Cs�|�d�|�d�|�d�|�|�d���dd�|�|�d���dd�|�d�g}d}|r�|D]$}|�d�|�d�ks||dkr\|}q\n|D]}|dkr�|}q�q�|r�|r�|��r�d�||���}|S)z{
        Return the version of the OS distribution, as a string.

        For details, see :func:`distro.version`.
        �
version_id�releaser`r@ra�.z	{0} ({1}))	r)r,r-�_parse_distro_release_contentrXr.�countr#r6)r3rr�versionsr�vrrrrLs2����zLinuxDistribution.versioncCsL|j|d�}|rHt�d�}|�|�}|rH|��\}}}||p>d|pDdfSdS)z�
        Return the version of the OS distribution, as a tuple of version
        numbers.

        For details, see :func:`distro.version_parts`.
        rz(\d+)\.?(\d+)?\.?(\d+)?r@)r@r@r@)r�re�compile�match�groups)r3r�version_str�
version_regex�matches�major�minorr!rrrrqs

zLinuxDistribution.version_partscCs|�|�dS)z�
        Return the major version number of the current distribution.

        For details, see :func:`distro.major_version`.
        r�r�r3rrrrr�szLinuxDistribution.major_versioncCs|�|�dS)z�
        Return the minor version number of the current distribution.

        For details, see :func:`distro.minor_version`.
        �rrrsrrrr �szLinuxDistribution.minor_versioncCs|�|�dS)z}
        Return the build number of the current distribution.

        For details, see :func:`distro.build_number`.
        rrrrsrrrr!�szLinuxDistribution.build_numbercCs|�d�pdS)z�
        Return the IDs of distributions that are like the OS distribution.

        For details, see :func:`distro.like`.
        �id_liker@)r)rRrrrr"�szLinuxDistribution.likecCs<z|jdWSty6|�d�p0|�d�p0dYS0dS)zp
        Return the codename of the OS distribution.

        For details, see :func:`distro.codename`.
        r#r@N)�_os_release_info�KeyErrorr,r-rRrrrr#�s
��zLinuxDistribution.codenamec	CsBt|��|�||�t|�|�|�|�|�|�d�|��|��d�S)z�
        Return certain machine-readable information about the OS
        distribution.

        For details, see :func:`distro.info`.
        )rprqr!)rrrr"r#)�dictrrrr r!r"r#)r3rrrrrr$�s
��zLinuxDistribution.infocCs|jS)z�
        Return a dictionary containing key-value pairs for the information
        items from the os-release file data source of the OS distribution.

        For details, see :func:`distro.os_release_info`.
        )rvrRrrrr%�sz!LinuxDistribution.os_release_infocCs|jS)z�
        Return a dictionary containing key-value pairs for the information
        items from the lsb_release command data source of the OS
        distribution.

        For details, see :func:`distro.lsb_release_info`.
        )�_lsb_release_inforRrrrr&�s	z"LinuxDistribution.lsb_release_infocCs|jS)z�
        Return a dictionary containing key-value pairs for the information
        items from the distro release file data source of the OS
        distribution.

        For details, see :func:`distro.distro_release_info`.
        )�_distro_release_inforRrrrr'�s	z%LinuxDistribution.distro_release_infocCs|jS)z�
        Return a dictionary containing key-value pairs for the information
        items from the uname command data source of the OS distribution.

        For details, see :func:`distro.uname_info`.
        )�_uname_inforRrrrr(�szLinuxDistribution.uname_infocCs|j�|d�S)z�
        Return a single named information item from the os-release file data
        source of the OS distribution.

        For details, see :func:`distro.os_release_attr`.
        r@)rvrX�r3r+rrrr)�sz!LinuxDistribution.os_release_attrcCs|j�|d�S)z�
        Return a single named information item from the lsb_release command
        output data source of the OS distribution.

        For details, see :func:`distro.lsb_release_attr`.
        r@)ryrXr|rrrr,sz"LinuxDistribution.lsb_release_attrcCs|j�|d�S)z�
        Return a single named information item from the distro release file
        data source of the OS distribution.

        For details, see :func:`distro.distro_release_attr`.
        r@)rzrXr|rrrr-
sz%LinuxDistribution.distro_release_attrcCs|j�|d�S)z�
        Return a single named information item from the uname command
        output data source of the OS distribution.

        For details, see :func:`distro.uname_attr`.
        r@)r{rXr|rrrr.szLinuxDistribution.uname_attrcCsFtj�|j�rBt|j��}|�|�Wd�S1s80YiS)z�
        Get the information items from the specified os-release file.

        Returns:
            A dictionary containing all information items.
        N)rCrDrLrJ�open�_parse_os_release_content)r3�release_filerrrrvs	(z"LinuxDistribution._os_release_infoc	Cs�i}tj|dd�}d|_tjddkr@t|jt�r@|j�d�|_t|�}|D]*}d|vrL|�	dd�\}}|||�
�<qLqLd|vr�|d|d	<n^d
|vr�|d
|d	<nHd|vr�t�d|d�}|r�|�
�}|�d
�}|�d�}|��}||d	<|S)aD
        Parse the lines of an os-release file.

        Parameters:

        * lines: Iterable through the lines in the os-release file.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        T)�posixrrz
iso-8859-1�=rt�version_codenamer#�ubuntu_codenamerz(\(\D+\))|,(\s+)?\D+z()�,)�shlex�whitespace_split�sys�version_info�
isinstance�	wordchars�bytes�decode�list�splitrVri�search�group�strip)	�lines�props�lexer�tokens�token�krhrkr#rrrr~,s0	

z+LinuxDistribution._parse_os_release_contentc
Cs�|js
iSttjd��R}zd}tj||d�}Wn*ttjfyXiYWd�S0Wd�n1sn0Y|�|��	�}|�
|�S)z�
        Get the information items from the lsb_release command output.

        Returns:
            A dictionary containing all information items.
        �wb)�lsb_releasez-a��stderrN)rNr}rC�devnull�
subprocess�check_output�OSError�CalledProcessError�_to_str�
splitlines�_parse_lsb_release_content�r3r��cmd�stdout�contentrrrryns	6z#LinuxDistribution._lsb_release_infocCsXi}|D]J}|�d��dd�}t|�dkr,q|\}}|�|�dd���|��i�q|S)aM
        Parse the output of the lsb_release command.

        Parameters:

        * lines: Iterable through the lines of the lsb_release output.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        �
�:rtrrTrU)r�r��len�updaterWrV)r�r��line�kvr�rhrrrr��s z,LinuxDistribution._parse_lsb_release_contentc	Cs�ttjd��L}zd}tj||d�}Wn$tyHiYWd�S0Wd�n1s^0Y|�|���}|�|�S)Nr�)�unamez-rsr�)	r}rCr�r�r�r�r�r��_parse_uname_contentr�rrrr{�s6zLinuxDistribution._uname_infocCsTi}t�d|d���}|rP|��\}}|dkr4iS|��|d<||d<||d<|S)Nz^([^\s]+)\s+([\d\.]+)r�Linuxrrrc)rir�r�rlrV)r�r�rkrrrrrr��sz&LinuxDistribution._parse_uname_contentcCsTt��}|dkrdn|}tjddkr<t|t�rP|�|�Snt|t�rP|�|�S|S)N�asciizutf-8r�)r��getfilesystemencodingr�r�r�r��unicode�encode)�text�encodingrrrr��s


zLinuxDistribution._to_strcCs|jrb|�|j�}tj�|j�}t�|�}d|vrLd|d��vrLd|d<n|r^|�d�|d<|Szt�	|j
�}|��Wnty�gd�}Yn0|D]t}|t
vr�q�t�|�}|r�tj�|j
|�}|�|�}d|vr�||_|�d�|d<d|d��v�rd|d<|Sq�iSdS)z�
        Get the information items from the specified distro release file.

        Returns:
            A dictionary containing all information items.
        r�
cloudlinuxrrt)zSuSE-releasezarch-releasezbase-releasezcentos-releasezfedora-releasezgentoo-releasezmageia-releasezmandrake-releasezmandriva-releasezmandrivalinux-releasezmanjaro-releasezoracle-releasezredhat-releasez
sl-releasezslackware-versionN)rM�_parse_distro_release_filerCrD�basename� _DISTRO_RELEASE_BASENAME_PATTERNrkrVr��listdirrG�sortr�� _DISTRO_RELEASE_IGNORE_BASENAMESrE)r3�distro_infor�rk�	basenames�filepathrrrrz�s8	




z&LinuxDistribution._distro_release_infoc	Cs\z<t|�� }|�|���Wd�WS1s00YWnttfyViYS0dS)z�
        Parse a distro release file.

        Parameters:

        * filepath: Path name of the distro release file.

        Returns:
            A dictionary containing all information items.
        N)r}re�readliner��IOError)r3r��fprrrr�s

2z,LinuxDistribution._parse_distro_release_filecCs�t�|��ddd��}i}|r~|�d�ddd�|d<|�d�rZ|�d�ddd�|d<|�d�r�|�d�ddd�|d<n|r�|��|d<|S)	a
        Parse a line from a distro release file.

        Parameters:
        * line: Line from the distro release file. Must be a unicode string
                or a UTF-8 encoded byte string.

        Returns:
            A dictionary containing all information items.
        N���r�rrrbrtr#)�(_DISTRO_RELEASE_CONTENT_REVERSED_PATTERNrkr�r�)r�ror�rrrre%s


z/LinuxDistribution._parse_distro_release_content)Tr@r@TN)T)F)FF)F)F)F)F)FF)%r0r<r=r>r5rSrrrrrrr r!r"r#r$r%r&r'r(r)r,r-r.r/rv�staticmethodr~ryr�r{r�r�rzr�rerrrrr?|s\�
j


%

	
	
		









A





Dr?cCs�t�t�}|�tj�|�t�tj��t	j
dd�}|jddddd�|jdd	td
dd�|�
�}|jrztd
d
|jd�}nt}|jr�|�tj|��ddd��n@|�d|jdd��|jdd�}|�d|�|��}|�d|�dS)NzOS distro info tool)raz--jsonz-jz!Output in machine readable format�
store_true)�help�actionz
--root-dirz-rrBz5Path to the root filesystem directory (defaults to /))�type�destr�F)rNrOrB�T)�indent�	sort_keyszName: %srzVersion: %szCodename: %s)�logging�	getLoggerr0�setLevel�DEBUG�
addHandler�
StreamHandlerr�r��argparse�ArgumentParser�add_argument�str�
parse_argsrBr?r�jsonr$�dumpsrrr#)�logger�parser�args�dist�distribution_version�distribution_codenamerrr�mainCs6
���r��__main__)T)F)FF)F)F)F)F)FF)@r>r�r�r�rCrir�r�r�r�__version__�Any�Callable�Dict�Iterable�Optional�Sequence�TextIO�Tuple�Type�	TypedDict�Unionr�ZVersionDictZInfoDict�environrXrFrHrKr]r^r_rjr�r�r�rrrrrrr r!r"r#r$r%r&r'r(r)r,r-r.�	functoolsr/�ImportError�objectr?rr�r0rrrr�<module>s�#
����
�
+P
(
-




-	I&
site-packages/pip/_vendor/__pycache__/__init__.cpython-39.pyc000064400000005577150564745010020135 0ustar00a

��Red�@sdZddlmZddlZddlZddlZdZej�	ej�
e��Zdd�Z
e�re�ej�ed��ejejdd�<e
d�e
d	�e
d
�e
d�e
d�e
d
�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d�e
d �e
d!�e
d"�e
d#�e
d$�e
d%�e
d&�e
d'�e
d(�e
d)�e
d*�e
d+�e
d,�e
d-�e
d.�e
d/�e
d0�e
d1�e
d2�e
d3�e
d4�e
d5�e
d6�e
d7�e
d8�e
d9�dS):z�
pip._vendor is for vendoring dependencies of pip to prevent needing pip to
depend on something external.

Files inside of pip._vendor should be considered immutable and should only be
updated to versions from upstream.
�)�absolute_importNFcCstd�t|�}zt|t�t�dd�Wnty6Yn:0tj|tj|<|�dd�\}}t	tj||tj|�dS)Nz{0}.{1}r)�level�.�)
�format�__name__�
__import__�globals�locals�ImportError�sys�modules�rsplit�setattr)�
modulename�
vendored_name�base�head�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/__init__.py�vendoreds	rz*.whl�cachecontrol�certifi�colorama�distlib�distro�html5lib�sixz	six.moveszsix.moves.urllibzsix.moves.urllib.parse�	packagingzpackaging.versionzpackaging.specifiers�pep517�
pkg_resources�platformdirs�progress�requestszrequests.exceptionszrequests.packageszrequests.packages.urllib3z&requests.packages.urllib3._collectionsz$requests.packages.urllib3.connectionz(requests.packages.urllib3.connectionpoolz!requests.packages.urllib3.contribz*requests.packages.urllib3.contrib.ntlmpoolz+requests.packages.urllib3.contrib.pyopensslz$requests.packages.urllib3.exceptionsz requests.packages.urllib3.fieldsz"requests.packages.urllib3.filepostz"requests.packages.urllib3.packagesz/requests.packages.urllib3.packages.ordered_dictz&requests.packages.urllib3.packages.sixz5requests.packages.urllib3.packages.ssl_match_hostnamezErequests.packages.urllib3.packages.ssl_match_hostname._implementationz%requests.packages.urllib3.poolmanagerz!requests.packages.urllib3.requestz"requests.packages.urllib3.responsezrequests.packages.urllib3.utilz)requests.packages.urllib3.util.connectionz&requests.packages.urllib3.util.requestz'requests.packages.urllib3.util.responsez$requests.packages.urllib3.util.retryz#requests.packages.urllib3.util.ssl_z&requests.packages.urllib3.util.timeoutz"requests.packages.urllib3.util.url�
resolvelib�tenacity�tomli�urllib3)�__doc__�
__future__r�glob�os.path�osr�	DEBUNDLED�path�abspath�dirname�__file__�	WHEEL_DIRr�joinrrrr�<module>sv$site-packages/pip/_vendor/__pycache__/pyparsing.cpython-39.pyc000064400000725532150564745010020412 0ustar00a

��Re�+�@s�
dZdZdZdZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlmZddlmZddlZdd	lmZdd
lmZzddlmZWney�ddlmZYn0zdd
lmZWn e�ydd
lmZYn0z ddlm Z ddlm!Z!m"Z"Wn0e�yVddl
m Z ddl
m!Z!m"Z"Yn0zddl
m#Z$Wn>e�y�zddl%m#Z$Wne�y�dZ$Yn0Yn0zddlm&Z&Wn"e�y�Gdd�d�Z&Yn0e&�Z'de'_de'_(e&�Z)de)_de)_*de)_+de)_,de)_-de)_.dd�e/e)�D�e)_0dd�Z1e1e)_2gd�Z3e4e	j5�dd�Z6e6ddkZ7e7�r�e	j8Z9e:Z;e<Z=e:Z>e:Z?e@eAeBeCeDe4eEeFeGeHeIgZJn^e	jKZ9eLZMdd�Z?gZJddlNZNd �O�D]6ZPzeJ�QeReNeP��WneS�y�Y�q�Yn0�q�eTd!d"�eMd#�D��ZUd$d%�ZVejWejXZYd&ZZeZd'Z[eYeZZ\e<d(�Z]d)�^d*d"�ej_D��Z`�dCd+d,�ZaGd-d.�d.eb�ZcGd/d0�d0ec�ZdGd1d2�d2ec�ZeGd3d4�d4ee�ZfGd5d6�d6eb�ZgGd7d8�d8eh�ZiGd9d:�d:eh�Zje!�kej�d;d<�Zld=d>�Zmd?d@�ZndAdB�ZodCdD�ZpdEdF�ZqdGdH�Zr�dDdJdK�ZsGdLdM�dMeh�ZtGdNdO�dOet�ZuGdPdQ�dQet�ZvGdRdS�dSev�ZwGdTdU�dUev�ZxGdVdW�dWev�ZyGdXdY�dYey�ZzeyZ{eyet_|GdZd[�d[ev�Z}Gd\d]�d]ey�Z~Gd^d_�d_e}�ZGd`da�daev�Z�Gdbdc�dcev�Z�Gddde�dee��Z�Gdfdg�dge��Z�Gdhdi�diev�Z�Gdjdk�dkev�Z�Gdldm�dmev�Z�Gdndo�doev�Z�Gdpdq�dqev�Z�Gdrds�dse��Z�Gdtdu�due��Z�Gdvdw�dwe��Z�Gdxdy�dye��Z�Gdzd{�d{e��Z�Gd|d}�d}e��Z�Gd~d�de��Z�Gd�d��d�et�Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�et�Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�eh�Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�Gd�d��d�eh�Z�d�d��Z��dEd�d��Z��dFd�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z��dGd�d��Z�d�d��Z��dHd�d��Z�d�d„Z�d�dĄZ�ew���dšZ�e����dơZ�e����dǡZ�e����dȡZ�e����dɡZ�e�e]d�dIdˍ��d�d̈́�Z�e�d΃��d�d̈́�Z�e�dЃ��d�d̈́�Z�e�e�Be�Be�d�d#dˍBZ�e�e�e�dӃe��Z�eydԃe�dՃ��d֡e�e�e�e�B����dסd�Z�d�dڄZ�d�d܄Z�d�dބZ�d�d�Z�d�d�Z�e�d�d̈́�Z�e�d�d̈́�Z�e�d�e�d�fd�d�Z�d�d�Z�d�d�Z�d�d�Z�eh�e�_ːdId�d�Z�e&�Z�eh�e�_�eh�e�_�e�d�e�d�fd�d�Z�e�Z�e�e�d��d����d��Z�e�e�d��d����d��Z�e�e�d��d�e�d��d�B���d��Z�e�e{d��eԠ����d��Z�d�d�deԠ�fd�d��Z֐dJ�d�d�Z�e��d�Z�e��d�Z�e�e�eYe\�d����d��\Z�Z�e�eݐd�O��d��Z�e��d�d	�^eޠߡ��d
����d�Z�d�d
�Z�e�e��d��d����d�Z�e��d����d�Z�e��d������d�Z�e��d����d�Z�e�e��d��de�B����d�Z�e�Z�e��d����d�Z�e�e�e�e`d��d�e�e��d�eyd��e����������d�Z�e�e�eԠ�e�Bd)�d�����d�Z�G�d�d ��d �Z�G�d!�d"��d"eh�Z�G�d#�d$��d$eh�Z�G�d%�d&��d&e�Z�e�j�j�j�e�j�j�j�e�j�j�j�e�j�_�e7�rbe�e�d'e�j��e�e�d(e�j��e�e�d)e�j��e�e�d*e�j��e�e�d+e�j��e�e�d,e�j�e�e�j�d-e�j�j�e�e�j�d.e�j�j��e�e�j�d/e�j�j�e�e�d0e�j��e�e�d1e�j��e�e�d2e�j��G�d3�d4��d4�Z��e�d5k�
r�e~�d6��Ze~�d7��Ze�eYe\�d8��Ze��e�d9d�d:���eš�Ze�e��e�����d;��Z�d<�eB�Ze��e�d9d�d:���eš�Ze�e��e�����d=��Z�e�d>��e�d;��e�e�d=��Z	�e	��
�d?�e�j��
�d@�e�j��
�d@�e�j
��
�dA�dd�l�Ze�j��eĐe�j��e�j��
�dB�dS(Ka�

pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and
executing simple grammars, vs. the traditional lex/yacc approach, or the
use of regular expressions.  With pyparsing, you don't need to learn
a new syntax for defining grammars or matching expressions - the parsing
module provides a library of classes that you use to construct the
grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form
``"<salutation>, <addressee>!"``), built up using :class:`Word`,
:class:`Literal`, and :class:`And` elements
(the :class:`'+'<ParserElement.__add__>` operators create :class:`And` expressions,
and the strings are auto-converted to :class:`Literal` expressions)::

    from pip._vendor.pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the
self-explanatory class names, and the use of '+', '|' and '^' operators.

The :class:`ParseResults` object returned from
:class:`ParserElement.parseString` can be
accessed as a nested list, a dictionary, or an object with named
attributes.

The pyparsing module handles some of the problems that are typically
vexing when writing text parsers:

  - extra or missing whitespace (the above program will also handle
    "Hello,World!", "Hello  ,  World  !", etc.)
  - quoted strings
  - embedded comments


Getting Started -
-----------------
Visit the classes :class:`ParserElement` and :class:`ParseResults` to
see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:

 - construct literal match expressions from :class:`Literal` and
   :class:`CaselessLiteral` classes
 - construct character word-group expressions using the :class:`Word`
   class
 - see how to create repetitive expressions using :class:`ZeroOrMore`
   and :class:`OneOrMore` classes
 - use :class:`'+'<And>`, :class:`'|'<MatchFirst>`, :class:`'^'<Or>`,
   and :class:`'&'<Each>` operators to combine simple expressions into
   more complex ones
 - associate names with your parsed results using
   :class:`ParserElement.setResultsName`
 - access the parsed data, which is returned as a :class:`ParseResults`
   object
 - find some helpful expression short-cuts like :class:`delimitedList`
   and :class:`oneOf`
 - find more useful common expressions in the :class:`pyparsing_common`
   namespace class
z2.4.7z30 Mar 2020 00:43 UTCz*Paul McGuire <ptmcg@users.sourceforge.net>�N)�ref)�datetime)�
itemgetter)�wraps)�contextmanager)�filterfalse)�ifilterfalse)�RLock)�Iterable)�MutableMapping�Mapping)�OrderedDict)�SimpleNamespacec@seZdZdS)rN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pyparsing.pyr��raA
    A cross-version compatibility configuration for pyparsing features that will be
    released in a future version. By setting values in this configuration to True,
    those features can be enabled in prior versions for compatibility development
    and testing.

     - collect_all_And_tokens - flag to enable fix for Issue #63 that fixes erroneous grouping
       of results names when an And expression is nested within an Or or MatchFirst; set to
       True to enable bugfix released in pyparsing 2.3.0, or False to preserve
       pre-2.3.0 handling of named results
Ta�
Diagnostic configuration (all default to False)
     - warn_multiple_tokens_in_named_alternation - flag to enable warnings when a results
       name is defined on a MatchFirst or Or expression with one or more And subexpressions
       (only warns if __compat__.collect_all_And_tokens is False)
     - warn_ungrouped_named_tokens_in_collection - flag to enable warnings when a results
       name is defined on a containing expression with ungrouped subexpressions that also
       have results names
     - warn_name_set_on_empty_Forward - flag to enable warnings whan a Forward is defined
       with a results name, but has no contents defined
     - warn_on_multiple_string_args_to_oneof - flag to enable warnings whan oneOf is
       incorrectly called with multiple str arguments
     - enable_debug_on_named_expressions - flag to auto-enable debug on all subsequent
       calls to ParserElement.setName()
FcCs$g|]}|�d�s|�d�r|�qS)Zenable_Zwarn_��
startswith)�.0�nmrrr�
<listcomp>�rrcCsdt_dt_dt_dt_dS�NT)�__diag__�)warn_multiple_tokens_in_named_alternation�)warn_ungrouped_named_tokens_in_collection�warn_name_set_on_empty_Forward�%warn_on_multiple_string_args_to_oneofrrrr�_enable_all_warnings�sr )t�__version__�__versionTime__�
__author__�
__compat__r�And�CaselessKeyword�CaselessLiteral�
CharsNotIn�Combine�Dict�Each�Empty�
FollowedBy�Forward�
GoToColumn�Group�Keyword�LineEnd�	LineStart�Literal�
PrecededBy�
MatchFirst�NoMatch�NotAny�	OneOrMore�OnlyOnce�Optional�Or�ParseBaseException�ParseElementEnhance�ParseException�ParseExpression�ParseFatalException�ParseResults�ParseSyntaxException�
ParserElement�QuotedString�RecursiveGrammarException�Regex�SkipTo�	StringEnd�StringStart�Suppress�Token�TokenConverter�White�Word�WordEnd�	WordStart�
ZeroOrMore�Char�	alphanums�alphas�
alphas8bit�anyCloseTag�
anyOpenTag�
cStyleComment�col�commaSeparatedList�commonHTMLEntity�countedArray�cppStyleComment�dblQuotedString�dblSlashComment�
delimitedList�dictOf�downcaseTokens�empty�hexnums�htmlComment�javaStyleComment�line�lineEnd�	lineStart�lineno�makeHTMLTags�makeXMLTags�matchOnlyAtCol�matchPreviousExpr�matchPreviousLiteral�
nestedExpr�nullDebugAction�nums�oneOf�opAssoc�operatorPrecedence�
printables�punc8bit�pythonStyleComment�quotedString�removeQuotes�replaceHTMLEntity�replaceWith�
restOfLine�sglQuotedString�srange�	stringEnd�stringStart�traceParseAction�
unicodeString�upcaseTokens�
withAttribute�
indentedBlock�originalTextFor�ungroup�
infixNotation�locatedExpr�	withClass�
CloseMatch�tokenMap�pyparsing_common�pyparsing_unicode�unicode_set�conditionAsParseAction�re�cCsdt|t�r|Sz
t|�WSty^t|��t��d�}td�}|�dd��|�	|�YS0dS)aDrop-in replacement for str(obj) that tries to be Unicode
        friendly. It first tries str(obj). If that fails with
        a UnicodeEncodeError, then it tries unicode(obj). It then
        < returns the unicode object | encodes it with the default
        encoding | ... >.
        �xmlcharrefreplacez&#\d+;cSs$dtt|ddd���dd�S)Nz\ur����)�hex�int��trrr�<lambda>�rz_ustr.<locals>.<lambda>N)
�
isinstance�unicode�str�UnicodeEncodeError�encode�sys�getdefaultencodingrG�setParseAction�transformString)�obj�retZ
xmlcharrefrrr�_ustr�s

r�z6sum len sorted reversed list tuple set any all min maxccs|]
}|VqdS�Nr)r�yrrr�	<genexpr>rr��cCs:d}dd�d��D�}t||�D]\}}|�||�}q |S)z/Escape &, <, >, ", ', etc. in a string of data.z&><"'css|]}d|dVqdS)�&�;Nr)r�srrrr�rz_xml_escape.<locals>.<genexpr>zamp gt lt quot apos)�split�zip�replace)�data�from_symbols�
to_symbols�from_�to_rrr�_xml_escapes
r��
0123456789ZABCDEFabcdef�\�ccs|]}|tjvr|VqdSr�)�string�
whitespace�r�crrrr�rcs@|dur|nd�|rtnt�t���t�����fdd��}|S)Nzfailed user-defined conditioncs t�|||��s�||���dSr�)�bool�r��lr���exc_type�fn�msgrr�pa%sz"conditionAsParseAction.<locals>.pa)rAr?�_trim_arityr)r��message�fatalr�rr�rr� sr�c@sPeZdZdZddd�Zedd��Zdd	�Zd
d�Zdd
�Z	ddd�Z
dd�ZdS)r=z7base exception class for all parsing runtime exceptionsrNcCs>||_|dur||_d|_n||_||_||_|||f|_dS�Nr�)�locr��pstr�
parserElement�args)�selfr�r�r��elemrrr�__init__0szParseBaseException.__init__cCs||j|j|j|j�S)z�
        internal factory method to simplify creating one type of ParseException
        from another - avoids having __init__ signature conflicts among subclasses
        )r�r�r�r�)�cls�perrr�_from_exception;sz"ParseBaseException._from_exceptioncCsN|dkrt|j|j�S|dvr,t|j|j�S|dkrBt|j|j�St|��dS)z�supported attributes by name are:
           - lineno - returns the line number of the exception text
           - col - returns the column number of the exception text
           - line - returns the line containing the exception text
        rk)rZ�columnrhN)rkr�r�rZrh�AttributeError)r��anamerrr�__getattr__CszParseBaseException.__getattr__cCs^|jr@|jt|j�krd}qDd|j|j|jd��dd�}nd}d|j||j|j|jfS)Nz, found end of textz
, found %rr�z\\�\r�z%%s%s  (at char %d), (line:%d, col:%d))r�r��lenr�r�rkr�)r��foundstrrrr�__str__Rs$�zParseBaseException.__str__cCst|�Sr��r��r�rrr�__repr__\szParseBaseException.__repr__�>!<cCs<|j}|jd}|r4d�|d|�|||d�f�}|��S)z�Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        r�r�N)rhr��join�strip)r��markerString�line_str�line_columnrrr�
markInputline^s
�z ParseBaseException.markInputlinecCsd��tt|��S)Nzlineno col line)r��dir�typer�rrr�__dir__hszParseBaseException.__dir__)rNN)r�)rrr�__doc__r��classmethodr�r�r�r�r�r�rrrrr=,s




r=c@seZdZdZeddd��ZdS)r?a:
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
    - lineno - returns the line number of the exception text
    - col - returns the column number of the exception text
    - line - returns the line containing the exception text

    Example::

        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))

    prints::

       Expected integer (at char 0), (line:1, col:1)
        column: 1

    �cCspddl}|durt��}g}t|t�rJ|�|j�|�d|jdd�|�d�t	|�j
|��|dk�rf|j|j|d�}t
�}t||d��D]�\}}|d}|j�dd�}	t|	t��r|jjd	vr�q�|	|vr�q�|�|	�t	|	�}
|�d
�|
j|
j
|	��nP|	du�r,t	|	�}
|�d�|
j|
j
��n&|j}|jdv�r@q�|�d
�|j��|d8}|s��qfq�d�|�S)ap
        Method to take an exception and translate the Python internal traceback into a list
        of the pyparsing expressions that caused the exception to be raised.

        Parameters:

         - exc - exception raised during parsing (need not be a ParseException, in support
           of Python exceptions that might be raised in a parse action)
         - depth (default=16) - number of levels back in the stack trace to list expression
           and function names; if None, the full stack trace names will be listed; if 0, only
           the failing input line, marker, and exception string will be shown

        Returns a multi-line string listing the ParserElements and/or function names in the
        exception's stack trace.

        Note: the diagnostic output will include string representations of the expressions
        that failed to parse. These representations will be more helpful if you use `setName` to
        give identifiable names to your expressions. Otherwise they will use the default string
        forms, which may be cryptic to read.

        explain() is only supported under Python 3.
        rN� r��^z{0}: {1})�contextr�)�	parseImpl�
_parseNoCachez
{0}.{1} - {2}z{0}.{1})�wrapperz<module>z{0}�
)�inspectr��getrecursionlimitr�r=�appendrhrZ�formatr�r�getinnerframes�
__traceback__�set�	enumerate�f_locals�getrD�f_code�co_name�addrr�)�exc�depthr�r��callers�seen�i�ff�frm�f_self�	self_type�coderrr�explain�sL


�
�zParseException.explainN)r�)rrrr��staticmethodrrrrrr?ksr?c@seZdZdZdS)rAznuser-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediatelyN�rrrr�rrrrrA�srAc@seZdZdZdS)rCz�just like :class:`ParseFatalException`, but thrown internally
    when an :class:`ErrorStop<And._ErrorStop>` ('-' operator) indicates
    that parsing is to stop immediately because an unbacktrackable
    syntax error has been found.
    NrrrrrrC�srCc@s eZdZdZdd�Zdd�ZdS)rFziexception thrown by :class:`ParserElement.validate` if the
    grammar could be improperly recursive
    cCs
||_dSr���parseElementTrace�r��parseElementListrrrr��sz"RecursiveGrammarException.__init__cCs
d|jS)NzRecursiveGrammarException: %srr�rrrr��sz!RecursiveGrammarException.__str__N)rrrr�r�r�rrrrrF�srFc@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�_ParseResultsWithOffsetcCs||f|_dSr���tup)r��p1�p2rrrr��sz _ParseResultsWithOffset.__init__cCs
|j|Sr�r�r�rrrr�__getitem__�sz#_ParseResultsWithOffset.__getitem__cCst|jd�S�Nr)�reprrr�rrrr��sz _ParseResultsWithOffset.__repr__cCs|jd|f|_dSrrrrrr�	setOffset�sz!_ParseResultsWithOffset.setOffsetN)rrrr�rr�rrrrrr�src@s�eZdZdZd]dd�Zddddefdd�Zdd	�Zefd
d�Zdd
�Z	dd�Z
dd�Zdd�ZeZ
dd�Zdd�Zdd�Zdd�Zdd�Zer�eZeZeZn$eZeZeZdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd^d(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Z d2d3�Z!d4d5�Z"d6d7�Z#d8d9�Z$d:d;�Z%d<d=�Z&d_d?d@�Z'dAdB�Z(dCdD�Z)dEdF�Z*d`dHdI�Z+dJdK�Z,dLdM�Z-dadOdP�Z.dQdR�Z/dSdT�Z0dUdV�Z1dWdX�Z2dYdZ�Z3e4dbd[d\��Z5dS)crBaSStructured parse results, to provide multiple means of access to
    the parsed data:

       - as a list (``len(results)``)
       - by list index (``results[0], results[1]``, etc.)
       - by attribute (``results.<resultsName>`` - see :class:`ParserElement.setResultsName`)

    Example::

        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)

    prints::

        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    NTcCs"t||�r|St�|�}d|_|Sr)r��object�__new__�_ParseResults__doinit)r��toklist�name�asList�modal�retobjrrrr!s


zParseResults.__new__c
Csb|jrvd|_d|_d|_i|_||_||_|dur6g}||t�rP|dd�|_n||t�rft|�|_n|g|_t	�|_
|du�r^|�r^|s�d|j|<||t�r�t|�}||_||t
d�ttf�r�|ddgfv�s^||t�r�|g}|�r*||t��rtt|j�d�||<ntt|d�d�||<|||_n4z|d||<Wn"tttf�y\|||<Yn0dS)NFrr�)r�_ParseResults__name�_ParseResults__parent�_ParseResults__accumNames�_ParseResults__asList�_ParseResults__modal�list�_ParseResults__toklist�_generatorType�dict�_ParseResults__tokdictr�r�r��
basestringrBr�KeyError�	TypeError�
IndexError)r�r r!r"r#r�rrrr�*sB



$
zParseResults.__init__cCsPt|ttf�r|j|S||jvr4|j|ddStdd�|j|D��SdS)Nr�rcSsg|]}|d�qS�rr�r�vrrrrXrz,ParseResults.__getitem__.<locals>.<listcomp>)r�r��slicer+r'r.rBrrrrrQs


zParseResults.__getitem__cCs�||t�r0|j�|t��|g|j|<|d}nD||ttf�rN||j|<|}n&|j�|t��t|d�g|j|<|}||t�r�t|�|_	dSr)
rr.r�r*r�r6r+rB�wkrefr&)r��kr5r��subrrr�__setitem__Zs


"
zParseResults.__setitem__c
Cs�t|ttf�r�t|j�}|j|=t|t�rH|dkr:||7}t||d�}tt|�|���}|��|j	�
�D]>\}}|D]0}t|�D]"\}\}}	t||	|	|k�||<q�qxqln|j	|=dS�Nrr�)
r�r�r6r�r+r*�range�indices�reverser.�itemsr�r)
r�r�mylen�removedr!�occurrences�jr8�value�positionrrr�__delitem__gs

zParseResults.__delitem__cCs
||jvSr�)r.)r�r8rrr�__contains__|szParseResults.__contains__cCs
t|j�Sr�)r�r+r�rrr�__len__szParseResults.__len__cCs
|jSr��r+r�rrr�__bool__�szParseResults.__bool__cCs
t|j�Sr���iterr+r�rrr�__iter__�szParseResults.__iter__cCst|jddd��S�Nr�rKr�rrr�__reversed__�szParseResults.__reversed__cCs$t|jd�r|j��St|j�SdS)N�iterkeys)�hasattrr.rPrLr�rrr�	_iterkeys�s
zParseResults._iterkeyscs�fdd����D�S)Nc3s|]}�|VqdSr�r�rr8r�rrr��rz+ParseResults._itervalues.<locals>.<genexpr>�rRr�rr�r�_itervalues�szParseResults._itervaluescs�fdd����D�S)Nc3s|]}|�|fVqdSr�rrSr�rrr��rz*ParseResults._iteritems.<locals>.<genexpr>rTr�rr�r�
_iteritems�szParseResults._iteritemscCst|���S)zVReturns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x).)r*rPr�rrr�keys�szParseResults.keyscCst|���S)zXReturns all named result values (as a list in Python 2.x, as an iterator in Python 3.x).)r*�
itervaluesr�rrr�values�szParseResults.valuescCst|���S)zfReturns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x).)r*�	iteritemsr�rrrr?�szParseResults.itemscCs
t|j�S)z�Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names.)r�r.r�rrr�haskeys�szParseResults.haskeyscOs�|s
dg}|��D]*\}}|dkr0|d|f}qtd|��qt|dt�sdt|�dksd|d|vr~|d}||}||=|S|d}|SdS)a�
        Removes and returns item at specified index (default= ``last``).
        Supports both ``list`` and ``dict`` semantics for ``pop()``. If
        passed no argument or an integer argument, it will use ``list``
        semantics and pop tokens from the list of parsed tokens. If passed
        a non-integer argument (most likely a string), it will use ``dict``
        semantics and pop the corresponding value from any defined results
        names. A second default return value argument is supported, just as in
        ``dict.pop()``.

        Example::

            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())

        prints::

            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        r��defaultrz-pop() got an unexpected keyword argument '%s'r�N)r?r1r�r�r�)r�r��kwargsr8r5�indexr��defaultvaluerrr�pop�s"%
�
�zParseResults.popcCs||vr||S|SdS)a[
        Returns named result matching the given key, or if there is no
        such name, then returns the given ``defaultValue`` or ``None`` if no
        ``defaultValue`` is specified.

        Similar to ``dict.get()``.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        Nr)r��key�defaultValuerrrr��szParseResults.getcCsR|j�||�|j��D]4\}}t|�D]"\}\}}t||||k�||<q(qdS)a
        Inserts new element at location index in the list of parsed tokens.

        Similar to ``list.insert()``.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        N)r+�insertr.r?r�r)r�r^�insStrr!rBr8rDrErrrrcszParseResults.insertcCs|j�|�dS)a�
        Add single element to end of ParseResults list of elements.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        N)r+r�)r��itemrrrr� s
zParseResults.appendcCs&t|t�r|�|�n|j�|�dS)a	
        Add sequence of elements to end of ParseResults list of elements.

        Example::

            patt = OneOrMore(Word(alphas))

            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        N)r�rB�__iadd__r+�extend)r��itemseqrrrrg/s
zParseResults.extendcCs|jdd�=|j��dS)z7
        Clear all elements and results names.
        N)r+r.�clearr�rrrriBszParseResults.clearcCs$z
||WStyYdS0dSr�)r0�r�r!rrrr�Is
zParseResults.__getattr__cCs|��}||7}|Sr���copy)r��otherr�rrr�__add__OszParseResults.__add__cs�|jrjt|j���fdd��|j��}�fdd�|D�}|D],\}}|||<t|dt�r<t|�|d_q<|j|j7_|j�	|j�|S)Ncs|dkr�S|�Srr)�a)�offsetrrr�Wrz'ParseResults.__iadd__.<locals>.<lambda>c	s4g|],\}}|D]}|t|d�|d��f�qqS�rr�)r�rr8�vlistr5)�	addoffsetrrrYs�z)ParseResults.__iadd__.<locals>.<listcomp>r)
r.r�r+r?r�rBr7r&r'�update)r�rm�
otheritems�otherdictitemsr8r5r)rtrprrfTs


�zParseResults.__iadd__cCs&t|t�r|dkr|��S||SdSr)r�r�rl�r�rmrrr�__radd__dszParseResults.__radd__cCsdt|j�t|j�fS)Nz(%s, %s))rr+r.r�rrrr�lszParseResults.__repr__cCsdd�dd�|jD��dS)N�[�, css(|] }t|t�rt|�nt|�VqdSr�)r�rBr�r�rrrrrr�prz'ParseResults.__str__.<locals>.<genexpr>�])r�r+r�rrrr�oszParseResults.__str__r�cCsLg}|jD]<}|r |r |�|�t|t�r8||��7}q
|�t|��q
|Sr�)r+r�r�rB�
_asStringListr�)r��sep�outrerrrr~rs


zParseResults._asStringListcCsdd�|jD�S)ax
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::

            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']

            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        cSs"g|]}t|t�r|��n|�qSr)r�rBr")r�resrrrr�rz'ParseResults.asList.<locals>.<listcomp>rIr�rrrr"}szParseResults.asListcs6tr|j}n|j}�fdd��t�fdd�|�D��S)a�
        Returns the named parse results as a nested dictionary.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})

            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        cs6t|t�r.|��r|��S�fdd�|D�Sn|SdS)Ncsg|]}�|��qSrrr4��toItemrrr�rz7ParseResults.asDict.<locals>.toItem.<locals>.<listcomp>)r�rBr[�asDict�r�r�rrr��s

z#ParseResults.asDict.<locals>.toItemc3s|]\}}|�|�fVqdSr�r�rr8r5r�rrr��rz&ParseResults.asDict.<locals>.<genexpr>)�PY_3r?rZr-)r��item_fnrr�rr��s
	zParseResults.asDictcCs<t|j�}t|j���|_|j|_|j�|j�|j|_|S)zG
        Returns a new copy of a :class:`ParseResults` object.
        )	rBr+r-r.r?r&r'rur%�r�r�rrrrl�s
zParseResults.copyFcCsLd}g}tdd�|j��D��}|d}|s8d}d}d}d}	|durJ|}	n|jrV|j}	|	sf|rbdSd}	|||d|	d	g7}t|j�D]�\}
}t|t�r�|
|vr�||�||
|o�|du||�g7}n||�d|o�|du||�g7}q�d}|
|vr�||
}|�s|�rq�nd}t	t
|��}
|||d|d	|
d
|d	g	7}q�|||d
|	d	g7}d�|�S)z�
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        r�css(|] \}}|D]}|d|fVqqdS�r�Nrrrrrrr��s
�z%ParseResults.asXML.<locals>.<genexpr>�  r�N�ITEM�<�>�</)r-r.r?r%r�r+r�rB�asXMLr�r�r�)r��doctag�namedItemsOnly�indent�	formatted�nlr��
namedItems�nextLevelIndent�selfTagrr��resTag�xmlBodyTextrrrr��sZ

�

�
�zParseResults.asXMLcCs:|j��D]*\}}|D]\}}||ur|Sqq
dSr�)r.r?)r�r9r8rsr5r�rrr�__lookup�s
zParseResults.__lookupcCs�|jr|jS|jr.|��}|r(|�|�SdSnNt|�dkrxt|j�dkrxtt|j����dddvrxtt|j����SdSdS)a
        Returns the results name for this token expression. Useful when several
        different expressions might match at a particular location.

        Example::

            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number")
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)

            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])

        prints::

            age : 22
            ssn : 111-22-3333
            house_number : 221B
        Nr�r)rr�)	r%r&�_ParseResults__lookupr�r.�nextrLrYrW)r��parrrr�getNames
��zParseResults.getNamercCsvg}d}|r$|�|t|����n
|�d�|�rl|��r�tdd�|��D��}|D]x\}}	|rl|�|�|�d|d||f�t|	t�r�|	r�|�|	j||||dd��q�|�t|	��qV|�t	|	��qVn�t
d	d�|D���rl|}	t|	�D]x\}
}t|t��r@|�d
|d||
|d|d|j||||dd�f�q�|�d
|d||
|d|dt|�f�q�d�|�S)aF
        Diagnostic method for listing out the contents of
        a :class:`ParseResults`. Accepts an optional ``indent`` argument so
        that this string can be embedded in a nested display of other data.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString('12/31/1999')
            print(result.dump())

        prints::

            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        r�r�css|]\}}t|�|fVqdSr�)r�r�rrrr�Grz$ParseResults.dump.<locals>.<genexpr>z
%s%s- %s: r�r�)r��full�include_list�_depthcss|]}t|t�VqdSr�)r�rB)r�vvrrrr�Srz
%s%s[%d]:
%s%s%s)
r�r�r"r[�sortedr?r�rB�dumpr�anyr�r�)r�r�r�r�r�r��NLr?r8r5rr�rrrr�)sP



��


�
zParseResults.dumpcOs tj|��g|�Ri|��dS)a#
        Pretty-printer for parsed results as a list, using the
        `pprint <https://docs.python.org/3/library/pprint.html>`_ module.
        Accepts additional positional or keyword args as defined for
        `pprint.pprint <https://docs.python.org/3/library/pprint.html#pprint.pprint>`_ .

        Example::

            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)

        prints::

            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        N)�pprintr"�r�r�r]rrrr�jszParseResults.pprintcCs.|j|j��|jdur|��p d|j|jffSr�)r+r.rlr&r'r%r�rrr�__getstate__�s��zParseResults.__getstate__cCsN|d|_|d\|_}}|_i|_|j�|�|durDt|�|_nd|_dSr;)r+r.r%r'rur7r&)r��stater��inAccumNamesrrr�__setstate__�s
zParseResults.__setstate__cCs|j|j|j|jfSr�)r+r%r(r)r�rrr�__getnewargs__�szParseResults.__getnewargs__cCstt|��t|���Sr�)r�r�r*rWr�rrrr��szParseResults.__dir__cCsrdd�}|g�}|��D]>\}}t|t�r>||j||d�7}q|||g|||�d�7}q|durn||g|d�}|S)z�
        Helper classmethod to construct a ParseResults from a dict, preserving the
        name-value relations as results names. If an optional 'name' argument is
        given, a nested ParseResults will be returned
        cSsFzt|�Wnty YdS0tr6t|ttf�St|t�SdS�NF)rL�	Exceptionr�r�r��bytesr/r�rrr�is_iterable�sz+ParseResults.from_dict.<locals>.is_iterable�r!)r!r"N)r?r�r�	from_dict)r�rmr!r�r�r8r5rrrr��s
zParseResults.from_dict)NNTT)N)r�)NFr�T)r�TTr)N)6rrrr�rr�r�rr:rFrGrHrJ�__nonzero__rMrOrRrUrVr�rWrYr?rPrXrZr[r`r�rcr�rgrir�rnrfryr�r�r~r"r�rlr�r�r�r�r�r�r�r�r�r�r�rrrrrB�sl*
	'	
7

$
=(
A
rBcCsF|}d|krt|�kr4nn||ddkr4dS||�dd|�S)a�Returns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See
   :class:`ParserElement.parseString` for more
   information on parsing strings containing ``<TAB>`` s, and suggested
   methods to maintain a consistent view of the parsed string, the parse
   location, and line and column positions within the parsed string.
   rr�r�)r��rfind)r��strgr�rrrrZ�srZcCs|�dd|�dS)a�Returns current line number within a string, counting newlines as line separators.
    The first line is number 1.

    Note - the default parsing behavior is to expand tabs in the input string
    before starting the parsing process.  See :class:`ParserElement.parseString`
    for more information on parsing strings containing ``<TAB>`` s, and
    suggested methods to maintain a consistent view of the parsed string, the
    parse location, and line and column positions within the parsed string.
    r�rr�)�count)r�r�rrrrk�s
rkcCsF|�dd|�}|�d|�}|dkr2||d|�S||dd�SdS)zfReturns the line of text containing loc within a string, counting newlines as line separators.
       r�rr�N)r��find)r�r��lastCR�nextCRrrrrh�s
rhcCs8tdt|�dt|�dt||�t||�f�dS)NzMatch z at loc z(%d,%d))�printr�rkrZ)�instringr��exprrrr�_defaultStartDebugAction�sr�cCs$tdt|�dt|����dS)NzMatched z -> )r�r�r�r")r��startloc�endlocr��toksrrr�_defaultSuccessDebugAction�sr�cCstdt|��dS)NzException raised:)r�r�)r�r�r�rrrr�_defaultExceptionDebugAction�sr�cGsdS)zG'Do-nothing' debug action, to suppress debugging output during parsing.Nr)r�rrrrr�srrr�cs��tvr�fdd�Sdg�dg�tdd�dkrFddd�}dd	d
��ntj}tj�d}|dd�d
}|d|d|f�������fdd�}d}zt�dt�d�j�}Wnty�t��}Yn0||_|S)Ncs�|�Sr�rr�)�funcrrr�rz_trim_arity.<locals>.<lambda>rFr�)r��cSs8tdkrdnd}tj||dd�|}|dd�gS)N)r�r�r������r���limitr�)�system_version�	traceback�
extract_stack)r�rp�
frame_summaryrrrr�sz"_trim_arity.<locals>.extract_stackcSs$tj||d�}|d}|dd�gS)Nr�r�r�)r��
extract_tb)�tbr��framesr�rrrr�sz_trim_arity.<locals>.extract_tb�r�r�r�cs�z"�|�dd��}d�d<|WSty��dr<�nfzHt��d}�|dd�ddd��ksh�Wz~Wq�ty�Yq�0nz~Wnty�Yn00�d�krˆdd7<Yq�Yq0qdS)NrTr�r�r�r�)r1r��exc_info�	NameError)r�r�r��r��
foundArityr�r��maxargs�pa_call_line_synthrrr�!s.�z_trim_arity.<locals>.wrapperz<parse action>r�	__class__)r)r)	�singleArgBuiltinsr�r�r�r��getattrrr�r�)r�r�r��	LINE_DIFF�	this_liner��	func_namerr�rr�s,

�r�c@s�eZdZdZdZdZedd��Zedd��Ze	dd	��Z
d�d
d�Zdd
�Zdd�Z
d�dd�Zd�dd�Zd�dd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd�d#d$�Zd%d&�Zd�d'd(�Zd)d*�Zd+d,�ZGd-d.�d.e�Zed/u�r�Gd0d1�d1e�ZnGd2d1�d1e�ZiZ e!�Z"d3d3gZ#d�d4d5�Z$eZ%ed6d7��Z&dZ'ed�d9d:��Z(d�d;d<�Z)e*dfd=d>�Z+d?d@�Z,e*fdAdB�Z-e*dfdCdD�Z.dEdF�Z/dGdH�Z0dIdJ�Z1dKdL�Z2dMdN�Z3dOdP�Z4dQdR�Z5dSdT�Z6dUdV�Z7dWdX�Z8dYdZ�Z9d[d\�Z:d]d^�Z;d_d`�Z<dadb�Z=d�dcdd�Z>dedf�Z?dgdh�Z@didj�ZAdkdl�ZBdmdn�ZCdodp�ZDd�dqdr�ZEdsdt�ZFdudv�ZGdwdx�ZHdydz�ZId�d{d|�ZJd�d}d~�ZKdd��ZLd�d��ZMd�d��ZNd�d��ZOd�d��ZPd�d�d��ZQd�d�d��ZRd/S)�rDz)Abstract base level parser element class.z 
	
FcCs
|t_dS)a�
        Overrides the default whitespace chars

        Example::

            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']

            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        N)rD�DEFAULT_WHITE_CHARS��charsrrr�setDefaultWhitespaceCharsLsz'ParserElement.setDefaultWhitespaceCharscCs
|t_dS)ah
        Set class to be used for inclusion of string literals into a parser.

        Example::

            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        N)rD�_literalStringClass�r�rrr�inlineLiteralsUsing\sz!ParserElement.inlineLiteralsUsingcCs|jr|j}q|Sr�)�tb_next)r�r�rrr�_trim_tracebackrszParserElement._trim_tracebackcCs�t�|_d|_d|_d|_||_d|_ttj	�|_
d|_d|_d|_
t�|_d|_d|_d|_d|_d|_d|_d|_d|_d|_dS)NTFr�)NNN)r*�parseAction�
failAction�strRepr�resultsName�
saveAsList�skipWhitespacer�rDr��
whiteChars�copyDefaultWhiteChars�mayReturnEmpty�keepTabs�ignoreExprs�debug�streamlined�
mayIndexError�errmsg�modalResults�debugActionsr��callPreparse�
callDuringTry)r��savelistrrrr�xs(zParserElement.__init__cCs<t�|�}|jdd�|_|jdd�|_|jr8tj|_|S)a/
        Make a copy of this :class:`ParserElement`.  Useful for defining
        different parse actions for the same parsing pattern, using copies of
        the original parse element.

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0] * 1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")

            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))

        prints::

            [5120, 100, 655360, 268435456]

        Equivalent form of ``expr.copy()`` is just ``expr()``::

            integerM = integer().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")
        N)rlr�r�r�rDr�r�)r��cpyrrrrl�s
zParserElement.copycCs$||_d|j|_tjr |��|S)a_
        Define name for this expression, makes debugging and exception messages clearer.

        Example::

            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        �	Expected )r!r�r�!enable_debug_on_named_expressions�setDebugrjrrr�setName�s
	zParserElement.setNamecCs|�||�S)aO
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original :class:`ParserElement` object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        ``expr("name")`` in place of ``expr.setResultsName("name")``
        - see :class:`__call__`.

        Example::

            date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        )�_setResultsName�r�r!�listAllMatchesrrr�setResultsName�szParserElement.setResultsNamecCs4|��}|�d�r"|dd�}d}||_||_|S)N�*r�T)rl�endswithr�r�)r�r!r�newselfrrrr�s
zParserElement._setResultsNameTcs@|r&|j�d�fdd�	}�|_||_nt|jd�r<|jj|_|S)z�Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set ``breakFlag`` to True to enable, False to
           disable.
        Tcsddl}|���||||�Sr)�pdb�	set_trace)r�r��	doActions�callPreParser��_parseMethodrr�breaker�sz'ParserElement.setBreak.<locals>.breaker�_originalParseMethod)TT)�_parserrQ)r��	breakFlagrrrr�setBreak�s
zParserElement.setBreakcOsVt|�dgkrg|_n<tdd�|D��s0td��tttt|���|_|�dd�|_|S)a�
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as ``fn(s, loc, toks)`` ,
        ``fn(loc, toks)`` , ``fn(toks)`` , or just ``fn()`` , where:

        - s   = the original string being parsed (see note below)
        - loc = the location of the matching substring
        - toks = a list of the matched tokens, packaged as a :class:`ParseResults` object

        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        If None is passed as the parse action, all previously added parse actions for this
        expression are cleared.

        Optional keyword arguments:
        - callDuringTry = (default= ``False``) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See :class:`parseString for more
        information on parsing strings containing ``<TAB>`` s, and suggested
        methods to maintain a consistent view of the parsed string, the parse
        location, and line and column positions within the parsed string.

        Example::

            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        Ncss|]}t|�VqdSr�)�callable)rr�rrrr�rz/ParserElement.setParseAction.<locals>.<genexpr>zparse actions must be callabler�F)r*r��allr1�mapr�r�r��r��fnsr]rrrr��s(zParserElement.setParseActioncOs4|jtttt|���7_|jp,|�dd�|_|S)z�
        Add one or more parse actions to expression's list of parse actions. See :class:`setParseAction`.

        See examples in :class:`copy`.
        r�F)r�r*rr�r�r�rrrr�addParseActionszParserElement.addParseActionc
OsF|D](}|j�t||�d�|�dd�d��q|jp>|�dd�|_|S)a�Add a boolean predicate function to expression's list of parse actions. See
        :class:`setParseAction` for function call signatures. Unlike ``setParseAction``,
        functions passed to ``addCondition`` need to return boolean success/fail of the condition.

        Optional keyword arguments:
        - message = define a custom message to be used in the raised exception
        - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        r�r�F)r�r�r�)r�r�r�r�r�)r�rr]r�rrr�addCondition)s
�
zParserElement.addConditioncCs
||_|S)aDefine action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           ``fn(s, loc, expr, err)`` where:
           - s = string being parsed
           - loc = location where expression match was attempted and failed
           - expr = the parse expression that failed
           - err = the exception thrown
           The function returns no value.  It may throw :class:`ParseFatalException`
           if it is desired to stop parsing immediately.)r��r�r�rrr�
setFailActionBs
zParserElement.setFailActionc	CsLd}|rHd}|jD]2}z|�||�\}}d}qWqtyBYq0qq|S�NTF)r�rr?)r�r�r��
exprsFound�e�dummyrrr�_skipIgnorablesOs


zParserElement._skipIgnorablescCsH|jr|�||�}|jrD|j}t|�}||krD|||vrD|d7}q&|S�Nr�)r�r r�r�r�)r�r�r��wt�instrlenrrr�preParse\s
zParserElement.preParsecCs|gfSr�r�r�r�r�r
rrrr�hszParserElement.parseImplcCs|Sr�r�r�r�r��	tokenlistrrr�	postParsekszParserElement.postParsecCs�d\}}}|j}|s|j�r"|j|r8|j||||�z�|rR|jrR|�||�}	n|}	|	}
|jsl|	t|�kr�z|�||	|�\}}Wq�ty�t	|t|�|j
|��Yq�0n|�||	|�\}}Wn^t�y}zD|j|r�|j|||
||�|j�r|�||
||��WYd}~n
d}~00n�|�r>|j�r>|�||�}	n|}	|	}
|j�s\|	t|�k�r�z|�||	|�\}}Wn(t�y�t	|t|�|j
|��Yn0n|�||	|�\}}|�|||�}t
||j|j|jd�}
|j�rd|�s�|j�rd|�r�z�|jD]�}z|||
|
�}Wn8t�yD}zt	d�}||_|�WYd}~n
d}~00|du�r�||
u�r�t
||j|j�ovt|t
tf�|jd�}
�q�WnHt�y�}z.|j|�r�|j|||
||��WYd}~n
d}~00n�|jD]�}z|||
|
�}Wn8t�y$}zt	d�}||_|�WYd}~n
d}~00|du�r�||
u�r�t
||j|j�oVt|t
tf�|jd�}
�q�|�r�|j|�r�|j|||
|||
�||
fS)N)rr�r�)r"r#z exception raised in parse action)r�r�r�r�r$r�r�r�r2r?r�r�r(rBr�r�r�r�r��	__cause__r�r*)r�r�r�r
r�TRY�MATCH�FAIL�	debugging�preloc�tokensStart�tokens�err�	retTokensr��parse_action_excrrrrr�os�




�
�
zParserElement._parseNoCachecCs>z|j||dd�dWSty8t|||j|��Yn0dS)NF�r
r)rrAr?r��r�r�r�rrr�tryParse�szParserElement.tryParsec	Cs2z|�||�Wnttfy(YdS0dSdS)NFT)r6r?r2r5rrr�canParseNext�s
zParserElement.canParseNextc@seZdZdd�ZdS)zParserElement._UnboundedCachecs~i�t�|_���fdd�}�fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr��r��r�ra��cache�not_in_cacherrr��sz3ParserElement._UnboundedCache.__init__.<locals>.getcs|�|<dSr�r�r�rarD�r;rrr��sz3ParserElement._UnboundedCache.__init__.<locals>.setcs���dSr��rir�r>rrri�sz5ParserElement._UnboundedCache.__init__.<locals>.clearcst��Sr��r�r�r>rr�	cache_len�sz9ParserElement._UnboundedCache.__init__.<locals>.cache_len)rr<�types�
MethodTyper�r�rirH)r�r�r�rirArr:rr��sz&ParserElement._UnboundedCache.__init__N�rrrr�rrrr�_UnboundedCache�srENc@seZdZdd�ZdS)�ParserElement._FifoCachecs�t�|_�t����fdd�}��fdd�}�fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_dS)	Ncs��|��Sr�r8r9r:rrr��s�.ParserElement._FifoCache.__init__.<locals>.getcs<|�|<t���kr8z��d�Wqty4Yq0qdSr�)r��popitemr0r=)r;�sizerrr��s�.ParserElement._FifoCache.__init__.<locals>.setcs���dSr�r?r�r>rrri�s�0ParserElement._FifoCache.__init__.<locals>.clearcst��Sr�r@r�r>rrrA�s�4ParserElement._FifoCache.__init__.<locals>.cache_len)	rr<�_OrderedDictrBrCr�r�rirH�r�rIr�r�rirAr)r;r<rIrr��s�!ParserElement._FifoCache.__init__NrDrrrr�
_FifoCache�srPc@seZdZdd�ZdS)rFcs�t�|_�i�t�g�����fdd�}���fdd�}��fdd�}�fdd�}t�||�|_t�||�|_t�||�|_t�||�|_	dS)	Ncs��|��Sr�r8r9r:rrr�srGcs4|�|<t���kr&�����d�q��|�dSr�)r�r`�popleftr�r=)r;�key_fiforIrrr�srJcs������dSr�r?r�)r;rRrrrisrKcst��Sr�r@r�r>rrrAsrL)
rr<�collections�dequerBrCr�r�rirHrNr)r;rRr<rIrr�	srONrDrrrrrPsrcCs0d\}}|||||f}tj��tj}|�|�}	|	|jur�tj|d7<z|�||||�}	Wn:ty�}
z"|�||
j	|
j
���WYd}
~
n8d}
~
00|�||	d|	d��f�|	Wd�SnBtj|d7<t|	t
�r�|	�|	d|	d��fWd�SWd�n1�s"0YdS)Nrqr�r)rD�packrat_cache_lock�
packrat_cacher�r<�packrat_cache_statsr�r=r�r�r�rlr�r�)r�r�r�r
r�HIT�MISS�lookupr;rDr�rrr�_parseCache+s$


zParserElement._parseCachecCs(tj��dgttj�tjdd�<dSr)rDrVrir�rWrrrr�
resetCacheDs
zParserElement.resetCache�cCs8tjs4dt_|dur t��t_nt�|�t_tjt_dS)a�Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.

           Parameters:

           - cache_size_limit - (default= ``128``) - if an integer value is provided
             will limit the size of the packrat cache; if None is passed, then
             the cache size will be unbounded; if 0 is passed, the cache will
             be effectively disabled.

           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method :class:`ParserElement.enablePackrat`.
           For best results, call ``enablePackrat()`` immediately after
           importing pyparsing.

           Example::

               from pip._vendor import pyparsing
               pyparsing.ParserElement.enablePackrat()
        TN)rD�_packratEnabledrErVrPr[r)�cache_size_limitrrr�
enablePackratJszParserElement.enablePackratc
Cs�t��|js|��|jD]}|��q|js8|��}z<|�|d�\}}|rr|�||�}t	�t
�}|�||�WnPty�}z8tjr��n"t
|dd�dur�|�|j�|_|�WYd}~nd}~00|SdS)a�
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        Returns the parsed data as a :class:`ParseResults` object, which may be
        accessed as a list, or as a dict or object with attributes if the given parser
        includes results names.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set ``parseAll`` to True (equivalent to ending
        the grammar with ``StringEnd()``).

        Note: ``parseString`` implicitly calls ``expandtabs()`` on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the ``loc`` argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:

        - calling ``parseWithTabs`` on your grammar before calling ``parseString``
          (see :class:`parseWithTabs`)
        - define your parse action using the full ``(s, loc, toks)`` signature, and
          reference the input string using the parse action's ``s`` argument
        - explictly expand the tabs in your input string before calling
          ``parseString``

        Example::

            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        rr�N)rDr\r��
streamliner�r��
expandtabsrr$r,rIr=�verbose_stacktracer�r�r�)r�r��parseAllrr�r0�serrrr�parseStringms(!

zParserElement.parseStringc
csV|js|��|jD]}|��q|js4t|���}t|�}d}|j}|j}t	�
�d}	z�||kr�|	|kr�z |||�}
|||
dd�\}}Wnty�|
d}YqZ0||kr�|	d7}	||
|fV|r�|||�}
|
|kr�|}q�|d7}q�|}qZ|
d}qZWnVt�yP}z<t	j
�r�n$t|dd�du�r8|�|j�|_|�WYd}~n
d}~00dS)aq
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        ``maxMatches`` argument, to clip scanning after 'n' matches are found.  If
        ``overlap`` is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See :class:`parseString` for more information on parsing
        strings with embedded tabs.

        Example::

            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens, start, end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])

        prints::

            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        rF�rr�r�N)r�rar�r�r�rbr�r$rrDr\r?r=rcr�r�r�)r�r��
maxMatches�overlaprr#r��
preparseFn�parseFn�matchesr.�nextLocr0�nextlocrrrr�
scanString�sF




zParserElement.scanStringc
Csg}d}d|_z�|�|�D]Z\}}}|�|||��|rpt|t�rR||��7}nt|t�rf||7}n
|�|�|}q|�||d��dd�|D�}d�tt	t
|���WSty�}z8tj
rĂn"t|dd�dur�|�|j�|_|�WYd}~n
d}~00dS)a[
        Extension to :class:`scanString`, to modify matching text with modified tokens that may
        be returned from a parse action.  To use ``transformString``, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking ``transformString()`` on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  ``transformString()`` returns the resulting transformed string.

        Example::

            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())

            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))

        prints::

            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        rTNcSsg|]}|r|�qSrr�r�orrrrrz1ParserElement.transformString.<locals>.<listcomp>r�r�)r�ror�r�rBr"r*r�rr��_flattenr=rDrcr�r�r�)r�r�r��lastEr�r�rrrrrr��s,



zParserElement.transformStringc
Csrztdd�|�||�D��WStyl}z8tjr6�n"t|dd�durT|�|j�|_|�WYd}~n
d}~00dS)a�
        Another extension to :class:`scanString`, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        ``maxMatches`` argument, to clip searching after 'n' matches are found.

        Example::

            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())

            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))

        prints::

            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        cSsg|]\}}}|�qSrr)rr�r�rrrrr7rz.ParserElement.searchString.<locals>.<listcomp>r�N)rBror=rDrcr�r�r�)r�r�rhrrrr�searchString!szParserElement.searchStringc	csTd}d}|j||d�D]*\}}}|||�V|r<|dV|}q||d�VdS)aR
        Generator method to split a string using the given expression as a separator.
        May be called with optional ``maxsplit`` argument, to limit the number of splits;
        and the optional ``includeSeparators`` argument (default= ``False``), if the separating
        matching text should be included in the split results.

        Example::

            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))

        prints::

            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        r)rhN)ro)	r�r��maxsplit�includeSeparators�splits�lastr�r�rrrrr�As
zParserElement.splitcCsV|turt|�St|t�r$|�|�}t|t�sJtjdt|�t	dd�dSt
||g�S)a[
        Implementation of + operator - returns :class:`And`. Adding strings to a ParserElement
        converts them to :class:`Literal`s by default.

        Example::

            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))

        prints::

            Hello, World! -> ['Hello', ',', 'World', '!']

        ``...`` may be used as a parse expression as a short form of :class:`SkipTo`.

            Literal('start') + ... + Literal('end')

        is equivalent to:

            Literal('start') + SkipTo('end')("_skipped*") + Literal('end')

        Note that the skipped text is returned with '_skipped' as a results name,
        and to support having multiple skips in the same parser, the value returned is
        a list of all skipped text.
        �4Cannot combine element of type %s with ParserElementr���
stacklevelN)�Ellipsis�_PendingSkipr�r/r�rD�warnings�warnr��
SyntaxWarningr%rxrrrrnZs


�zParserElement.__add__cCsZ|turt|�d�|St|t�r,|�|�}t|t�sRtjdt|�t	dd�dS||S)z`
        Implementation of + operator when left operand is not a :class:`ParserElement`
        �	_skipped*ryr�rzN)
r|rHr�r/r�rDr~rr�r�rxrrrry�s


�zParserElement.__radd__cCsJt|t�r|�|�}t|t�s:tjdt|�tdd�dS|t�	�|S)zT
        Implementation of - operator, returns :class:`And` with error stop
        ryr�rzN)
r�r/r�rDr~rr�r�r%�
_ErrorStoprxrrr�__sub__�s


�zParserElement.__sub__cCsBt|t�r|�|�}t|t�s:tjdt|�tdd�dS||S)z`
        Implementation of - operator when left operand is not a :class:`ParserElement`
        ryr�rzN�r�r/r�rDr~rr�r�rxrrr�__rsub__�s


�zParserElement.__rsub__cs|turd}n8t|t�rF|dd�tfkrFd|dd�ddd�}t|t�r\|d}}n�t|t��rJtdd	�|D��}|d
dd�}|ddur�d|df}t|dt�r�|ddur�|ddkr�t��S|ddkr�t��S�|dt��SnNt|dt��r,t|dt��r,|\}}||8}ntdt|d�t|d���ntdt|���|dk�rjtd
��|dk�r|td��||k�r�dk�r�nntd��|�r��fdd��|�r�|dk�r҈�|�}nt	�g|��|�}n�|�}n|dk�r�}nt	�g|�}|S)a�
        Implementation of * operator, allows use of ``expr * 3`` in place of
        ``expr + expr + expr``.  Expressions may also me multiplied by a 2-integer
        tuple, similar to ``{min, max}`` multipliers in regular expressions.  Tuples
        may also include ``None`` as in:
         - ``expr*(n, None)`` or ``expr*(n, )`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr*(None, n)`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr*(None, None)`` is equivalent to ``ZeroOrMore(expr)``
         - ``expr*(1, None)`` is equivalent to ``OneOrMore(expr)``

        Note that ``expr*(None, n)`` does not raise an exception if
        more than n exprs exist in the input stream; that is,
        ``expr*(None, n)`` does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        ``expr*(None, n) + ~expr``
        )rNNr�r3r�r�rcss|]}|tur|ndVqdSr�)r|rprrrr��rz(ParserElement.__mul__.<locals>.<genexpr>�NNz8cannot multiply 'ParserElement' and ('%s', '%s') objectsz0cannot multiply 'ParserElement' and '%s' objectsz/cannot multiply ParserElement by negative valuez@second tuple value must be greater or equal to first tuple valuez,cannot multiply ParserElement by 0 or (0, 0)cs(|dkrt��|d��St��SdSr!)r;��n��makeOptionalListr�rrr��sz/ParserElement.__mul__.<locals>.makeOptionalList)
r|r��tupler�rRr9r1r��
ValueErrorr%)r�rm�minElements�optElementsr�rr�r�__mul__�sN
 





zParserElement.__mul__cCs
|�|�Sr�)r�rxrrr�__rmul__�szParserElement.__rmul__cCsZ|turt|dd�St|t�r(|�|�}t|t�sNtjdt|�t	dd�dSt
||g�S)zL
        Implementation of | operator - returns :class:`MatchFirst`
        T)�	must_skipryr�rzN)r|r}r�r/r�rDr~rr�r�r6rxrrr�__or__�s


�zParserElement.__or__cCsBt|t�r|�|�}t|t�s:tjdt|�tdd�dS||BS)z`
        Implementation of | operator when left operand is not a :class:`ParserElement`
        ryr�rzNr�rxrrr�__ror__	s


�zParserElement.__ror__cCsFt|t�r|�|�}t|t�s:tjdt|�tdd�dSt||g�S)zD
        Implementation of ^ operator - returns :class:`Or`
        ryr�rzN)	r�r/r�rDr~rr�r�r<rxrrr�__xor__	s


�zParserElement.__xor__cCsBt|t�r|�|�}t|t�s:tjdt|�tdd�dS||AS)z`
        Implementation of ^ operator when left operand is not a :class:`ParserElement`
        ryr�rzNr�rxrrr�__rxor__	s


�zParserElement.__rxor__cCsFt|t�r|�|�}t|t�s:tjdt|�tdd�dSt||g�S)zF
        Implementation of & operator - returns :class:`Each`
        ryr�rzN)	r�r/r�rDr~rr�r�r+rxrrr�__and__'	s


�zParserElement.__and__cCsBt|t�r|�|�}t|t�s:tjdt|�tdd�dS||@S)z`
        Implementation of & operator when left operand is not a :class:`ParserElement`
        ryr�rzNr�rxrrr�__rand__3	s


�zParserElement.__rand__cCst|�S)zH
        Implementation of ~ operator - returns :class:`NotAny`
        )r8r�rrr�
__invert__?	szParserElement.__invert__cCstd|jj��dS)Nz%r object is not iterable)r1r�rr�rrrrME	szParserElement.__iter__c	Cs�zt|t�r|f}t|�Wnty6||f}Yn0t|�dkrxt�d�|dd�t|�dkrpd�t|��nd��|t|dd��}|S)a�
        use ``[]`` indexing notation as a short form for expression repetition:
         - ``expr[n]`` is equivalent to ``expr*n``
         - ``expr[m, n]`` is equivalent to ``expr*(m, n)``
         - ``expr[n, ...]`` or ``expr[n,]`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr[..., n]`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr[...]`` and ``expr[0, ...]`` are equivalent to ``ZeroOrMore(expr)``
         - ``expr[1, ...]`` is equivalent to ``OneOrMore(expr)``
         ``None`` may be used in place of ``...``.

        Note that ``expr[..., n]`` and ``expr[m, n]``do not raise an exception
        if more than ``n`` ``expr``s exist in the input stream.  If this behavior is
        desired, then write ``expr[..., n] + ~expr``.
       r�z.only 1 or 2 index arguments supported ({0}{1})Nr�z	... [{0}]r�)	r�r�rLr1r�r~rr�r�)r�rar�rrrrJ	s

��zParserElement.__getitem__cCs|dur|�|�S|��SdS)a�
        Shortcut for :class:`setResultsName`, with ``listAllMatches=False``.

        If ``name`` is given with a trailing ``'*'`` character, then ``listAllMatches`` will be
        passed as ``True``.

        If ``name` is omitted, same as calling :class:`copy`.

        Example::

            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums + "-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums + "-")("socsecno")
        N)rrlrjrrr�__call__n	s
zParserElement.__call__cCst|�S)z�
        Suppresses the output of this :class:`ParserElement`; useful to keep punctuation from
        cluttering up returned output.
        )rKr�rrr�suppress�	szParserElement.suppresscCs
d|_|S)a
        Disables the skipping of whitespace before matching the characters in the
        :class:`ParserElement`'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        F�r�r�rrr�leaveWhitespace�	szParserElement.leaveWhitespacecCsd|_||_d|_|S)z8
        Overrides the default whitespace chars
        TF)r�r�r�)r�r�rrr�setWhitespaceChars�	sz ParserElement.setWhitespaceCharscCs
d|_|S)z�
        Overrides default behavior to expand ``<TAB>``s to spaces before parsing the input string.
        Must be called before ``parseString`` when the input grammar contains elements that
        match ``<TAB>`` characters.
        T)r�r�rrr�
parseWithTabs�	szParserElement.parseWithTabscCsLt|t�rt|�}t|t�r4||jvrH|j�|�n|j�t|����|S)a�
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.

        Example::

            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']

            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        )r�r/rKr�r�rlrxrrr�ignore�	s


zParserElement.ignorecCs"|pt|pt|ptf|_d|_|S)zT
        Enable display of debugging messages while doing pattern matching.
        T)r�r�r�r�r�)r��startAction�
successAction�exceptionActionrrr�setDebugActions�	s�zParserElement.setDebugActionscCs|r|�ttt�nd|_|S)a�
        Enable display of debugging messages while doing pattern matching.
        Set ``flag`` to True to enable, False to disable.

        Example::

            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer

            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")

        prints::

            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using :class:`setDebugActions`. Prior to attempting
        to match the ``wd`` expression, the debugging message ``"Match <exprname> at loc <n>(<line>,<col>)"``
        is shown. Then if the parse succeeds, a ``"Matched"`` message is shown, or an ``"Exception raised"``
        message is shown. Also note the use of :class:`setName` to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the :class:`Word` expression without calling ``setName`` is ``"W:(ABCD...)"``.
        F)r�r�r�r�r�)r��flagrrrr��	s%zParserElement.setDebugcCs|jSr�r�r�rrrr��	szParserElement.__str__cCst|�Sr�r�r�rrrr��	szParserElement.__repr__cCsd|_d|_|Sr)r�r�r�rrrra�	szParserElement.streamlinecCsdSr�rrrrr�checkRecursion�	szParserElement.checkRecursioncCs|�g�dS)zj
        Check defined expressions for valid structure, check for infinite recursive definitions.
        N)r�)r��
validateTracerrr�validate�	szParserElement.validatecCs�z|��}WnDtyPt|d��}|��}Wd�n1sB0YYn0z|�||�WSty�}z8tjrz�n"t|dd�dur�|�|j	�|_	|�WYd}~n
d}~00dS)z�
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        �rNr�)
�readr��openrfr=rDrcr�r�r�)r��file_or_filenamerd�
file_contents�frrrr�	parseFile
s,zParserElement.parseFilecCs>||urdSt|t�r |�|�St|t�r:t|�t|�kSdSr)r�r/rlrD�varsrxrrr�__eq__
s


zParserElement.__eq__cCs
||kSr�rrxrrr�__ne__$
szParserElement.__ne__cCst|�Sr�)�idr�rrr�__hash__'
szParserElement.__hash__cCs||kSr�rrxrrr�__req__*
szParserElement.__req__cCs
||kSr�rrxrrr�__rne__-
szParserElement.__rne__cCs2z|jt|�|d�WdSty,YdS0dS)a�
        Method for quick testing of a parser against a test string. Good for simple
        inline microtests of sub expressions while building up larger parser.

        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests

        Example::

            expr = Word(nums)
            assert expr.matches("100")
        �rdTFN)rfr�r=)r��
testStringrdrrrrl0
s
zParserElement.matches�#c	Cs�t|t�r"tttj|������}t|t�r4t|�}|durBt	j
}|j}	g}
g}d}td��t
d���t�}
d}|D�]p}|dur�|�|d�s�|r�|s�|�|�qt|s�qt|r�dd�|�nd|g}g}z"|
�|�|��}|j||d�}Wn�t�y�}z�t|t��r
d	nd}d|v�rN|�t|j|��|�d
t|j|�dd|�n|�d
|jd|�|�d
t|��|�o�|}|}WYd}~�n d}~0t�y�}z,|�dt|��|�o�|}|}WYd}~n�d}~00|�o�|}|du�r�zR|||�}|du�r:t|t��r*|�|���n|�t|��n|�|���WnTt�y�}z:|�|j|d��|�d�|jt |�j|��WYd}~n
d}~00n|�|j|d��|�r�|�r�|�d�|	d�|��|
�||f�qt||
fS)as

        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.

        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests
         - comment - (default= ``'#'``) - expression for indicating embedded comments in the test
              string; pass None to disable comment filtering
         - fullDump - (default= ``True``) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default= ``True``) prints test output to stdout
         - failureTests - (default= ``False``) indicates if these tests are expected to fail parsing
         - postParse - (default= ``None``) optional callback for successful parse results; called as
              `fn(test_string, parse_results)` and returns a string to be added to the test output
         - file - (default=``None``) optional file-like object to which test output will be written;
              if None, will default to ``sys.stdout``

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if ``failureTests`` is True), and the results contain a list of lines of each
        test's output

        Example::

            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")

        prints::

            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success

            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\n of strings that spans \n 3 lines")

        (Note that this is a raw string literal, you must include the leading 'r'.)
        NT�\nr�uFr�r�z(FATAL)r�r�r�zFAIL: zFAIL-EXCEPTION: )r�z{0} failed: {1}: {2})!r�r/r*rr�r��rstrip�
splitlinesr4r��stdout�writerr}r�rzrlr�r�r��lstriprfr=rArhr�rZr�rBr�r�rr�)r��testsrd�comment�fullDump�printResults�failureTestsr(�file�print_�
allResults�comments�successr��BOMr�r��resultr�r�r�pp_valuerrrr�runTestsD
sn`




$




4
zParserElement.runTests)F)F)F)T)T)TT)TT)r])F)N)T)N)F)T)Tr�TTFNN)Srrrr�r�rcr
r�r�r�r�r�rlrrrrr�rrrr r$r�r(r�r6r7rrErMrPrVr	rUrWr[rr\r^r`rf�_MAX_INTror�rtr�rnryr�r�r�r�r�r�r�r�r�r�r�rMrr�r�r�r�r�r�r�r�r�r�rar�r�r�r�r�r�r�r�rlr�rrrrrDGs�





	
1




W


"
:J0 &J$
			

+

	
�rDcs6eZdZd
�fdd�	Zdd�Zdd�Zdd	�Z�ZS)r}Fcs>tt|���t|t���dd�|_|j|_||_||_	dS)Nr,�...)
�superr}r�r�r,r�r�r!�anchorr�)r�r�r��r�rrr��
s
z_PendingSkip.__init__cs\t|��d�d�}�jrNdd�}�fdd�}�j|��|�|��|�B|S�j||S)Nr�r�cSs,|jr|j��dgkr(|d=|�dd�dS)Nr�r�_skipped)r�r"r`r�rrrr��
sz'_PendingSkip.__add__.<locals>.must_skipcs<|j��dd�dgkr8|�d�}dt�j�d|d<dS)Nr�r�r�z	missing <r�)r�r"r`rr�)r��skippedr�rr�	show_skip�
s
z'_PendingSkip.__add__.<locals>.show_skip)rHrr�r�r)r�rm�skipperr�r�rr�rrn�
s
��z_PendingSkip.__add__cCs|jSr�)r�r�rrrr�sz_PendingSkip.__repr__cGstd��dS)NzBuse of `...` expression without following SkipTo target expression)r��r�r�rrrr�sz_PendingSkip.parseImpl)F)rrrr�rnr�r��
__classcell__rrr�rr}�
sr}cs eZdZdZ�fdd�Z�ZS)rLzYAbstract :class:`ParserElement` subclass, for defining atomic
    matching patterns.
    cstt|�jdd�dS�NF�r�)r�rLr�r�r�rrr�szToken.__init__�rrrr�r�r�rrr�rrLsrLcs eZdZdZ�fdd�Z�ZS)r,z'An empty token, will always match.
    cs$tt|���d|_d|_d|_dS)Nr,TF)r�r,r�r!r�r�r�r�rrr�szEmpty.__init__r�rrr�rr,sr,cs*eZdZdZ�fdd�Zddd�Z�ZS)r7z#A token that will never match.
    cs*tt|���d|_d|_d|_d|_dS)Nr7TFzUnmatchable token)r�r7r�r!r�r�r�r�r�rrr�s
zNoMatch.__init__TcCst|||j|��dSr�)r?r�r%rrrr�$szNoMatch.parseImpl)T�rrrr�r�r�r�rrr�rr7sr7cs*eZdZdZ�fdd�Zddd�Z�ZS)r4a�Token to exactly match a specified string.

    Example::

        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"

    For case-insensitive matching, use :class:`CaselessLiteral`.

    For keyword matching (force word break before and after the matched string),
    use :class:`Keyword` or :class:`CaselessKeyword`.
    cs�tt|���||_t|�|_z|d|_Wn(tyTtj	dt
dd�t|_Yn0dt
|j�|_d|j|_d|_d|_|jdkr�t|�tur�t|_dS)	Nrz2null string passed to Literal; use Empty() insteadr�rz�"%s"r�Fr�)r�r4r��matchr��matchLen�firstMatchCharr2r~rr�r,r�r�r!r�r�r�r��_SingleCharLiteral�r��matchStringr�rrr�6s 
�zLiteral.__init__TcCs@|||jkr,|�|j|�r,||j|jfSt|||j|��dSr�)r�rr�r�r?r�r%rrrr�JszLiteral.parseImpl)Tr�rrr�rr4(s
r4c@seZdZddd�ZdS)r�TcCs0|||jkr|d|jfSt|||j|��dSr!)r�r�r?r�r%rrrr�Psz_SingleCharLiteral.parseImplN)T�rrrr�rrrrr�Osr�csLeZdZdZedZd�fdd�	Zddd	�Z�fd
d�Ze	dd
��Z
�ZS)r1a�Token to exactly match a specified string as a keyword, that is,
    it must be immediately followed by a non-keyword character.  Compare
    with :class:`Literal`:

     - ``Literal("if")`` will match the leading ``'if'`` in
       ``'ifAndOnlyIf'``.
     - ``Keyword("if")`` will not; it will only match the leading
       ``'if'`` in ``'if x=1'``, or ``'if(y==2)'``

    Accepts two optional constructor arguments in addition to the
    keyword string:

     - ``identChars`` is a string of characters that would be valid
       identifier characters, defaulting to all alphanumerics + "_" and
       "$"
     - ``caseless`` allows case-insensitive matching, default is ``False``.

    Example::

        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use :class:`CaselessKeyword`.
    �_$NFcs�tt|���|durtj}||_t|�|_z|d|_Wn"ty\t	j
dtdd�Yn0d|j|_d|j|_
d|_d|_||_|r�|��|_|��}t|�|_dS)Nrz2null string passed to Keyword; use Empty() insteadr�rzr�r�F)r�r1r��DEFAULT_KEYWORD_CHARSr�r�r�r�r2r~rr�r!r�r�r��caseless�upper�
caselessmatchr��
identChars)r�r�r�r�r�rrr�ss(
�
zKeyword.__init__TcCs|jr|||||j���|jkr�|t|�|jksL|||j��|jvr�|dksj||d��|jvr�||j|jfSnv|||jkr�|jdks�|�|j|�r�|t|�|jks�|||j|jvr�|dks�||d|jvr�||j|jfSt	|||j
|��dSr;)r�r�r�r�r�r�r�r�rr?r�r%rrrr��s.��������zKeyword.parseImplcstt|���}tj|_|Sr�)r�r1rlr�r�)r�r�r�rrrl�szKeyword.copycCs
|t_dS)z,Overrides the default Keyword chars
        N)r1r�r�rrr�setDefaultKeywordChars�szKeyword.setDefaultKeywordChars)NF)T)rrrr�rTr�r�r�rlr
r�r�rrr�rr1Xs
r1cs*eZdZdZ�fdd�Zddd�Z�ZS)r'afToken to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::

        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']

    (Contrast with example for :class:`CaselessKeyword`.)
    cs6tt|��|���||_d|j|_d|j|_dS)Nz'%s'r�)r�r'r�r��returnStringr!r�r�r�rrr��szCaselessLiteral.__init__TcCs@||||j���|jkr,||j|jfSt|||j|��dSr�)r�r�r�r�r?r�r%rrrr��szCaselessLiteral.parseImpl)Tr�rrr�rr'�s
r'cs"eZdZdZd�fdd�	Z�ZS)r&z�
    Caseless version of :class:`Keyword`.

    Example::

        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']

    (Contrast with example for :class:`CaselessLiteral`.)
    Ncstt|�j||dd�dS)NT�r�)r�r&r�)r�r�r�r�rrr��szCaselessKeyword.__init__)Nr�rrr�rr&�s	r&cs,eZdZdZd�fdd�	Zd	dd�Z�ZS)
r�a�A variation on :class:`Literal` which matches "close" matches,
    that is, strings with at most 'n' mismatching characters.
    :class:`CloseMatch` takes parameters:

     - ``match_string`` - string to be matched
     - ``maxMismatches`` - (``default=1``) maximum number of
       mismatches allowed to count as a match

    The results from a successful parse will contain the matched text
    from the input string and the following named results:

     - ``mismatches`` - a list of the positions within the
       match_string where mismatches were found
     - ``original`` - the original match_string used to compare
       against the input string

    If ``mismatches`` is an empty list, then the match was an exact
    match.

    Example::

        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    r�csBtt|���||_||_||_d|j|jf|_d|_d|_dS)Nz&Expected %r (with up to %d mismatches)F)	r�r�r�r!�match_string�
maxMismatchesr�r�r�)r�r�r�r�rrr��szCloseMatch.__init__TcCs�|}t|�}|t|j�}||kr�|j}d}g}	|j}
tt|||�|��D]2\}}|\}}
||
krL|	�|�t|	�|
krLq�qL|d}t|||�g�}||d<|	|d<||fSt|||j|��dS)Nrr��original�
mismatches)	r�r�r�r�r�r�rBr?r�)r�r�r�r
�startr#�maxlocr��match_stringlocr�r��s_m�src�mat�resultsrrrr��s(
zCloseMatch.parseImpl)r�)Tr�rrr�rr��s 	r�cs8eZdZdZd
�fdd�	Zdd	d
�Z�fdd�Z�ZS)rOaX	Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters, an
    optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for ``min`` is
    1 (a minimum value < 1 is not valid); the default values for
    ``max`` and ``exact`` are 0, meaning no maximum or exact
    length restriction. An optional ``excludeChars`` parameter can
    list characters that might be found in the input ``bodyChars``
    string; useful to define a word of all printables except for one or
    two characters, for instance.

    :class:`srange` is useful for defining custom character set strings
    for defining ``Word`` expressions, using range notation from
    regular expression character sets.

    A common mistake is to use :class:`Word` to match a specific literal
    string, as in ``Word("Address")``. Remember that :class:`Word`
    uses the string argument to define *sets* of matchable characters.
    This expression would match "Add", "AAA", "dAred", or any other word
    made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an
    exact literal string, use :class:`Literal` or :class:`Keyword`.

    pyparsing includes helper strings for building Words:

     - :class:`alphas`
     - :class:`nums`
     - :class:`alphanums`
     - :class:`hexnums`
     - :class:`alphas8bit` (alphabetic characters in ASCII range 128-255
       - accented, tilded, umlauted, etc.)
     - :class:`punc8bit` (non-alphabetic characters in ASCII range
       128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - :class:`printables` (any non-whitespace character)

    Example::

        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))

        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums + '-')

        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")

        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    Nr�rFcs�tt|����rNt���d��fdd�|D��}|rNd��fdd�|D��}||_t|�|_|rt||_t|�|_n||_t|�|_|dk|_	|dkr�t
d��||_|dkr�||_nt
|_|dkr�||_||_t|�|_d|j|_d	|_||_d
|j|jv�r�|dk�r�|dk�r�|dk�r�|j|jk�r@dt|j�|_nHt|j�dk�rndt�|j�t|j�f|_nd
t|j�t|j�f|_|j�r�d|jd|_zt�|j�|_Wnt�y�d|_Yn0|jj|_t|_dS)Nr�c3s|]}|�vr|VqdSr�rr���excludeCharsrrr�Hrz Word.__init__.<locals>.<genexpr>c3s|]}|�vr|VqdSr�rr�r�rrr�Jrrr�zZcannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permittedr�Fr�z[%s]+z%s[%s]*z	[%s][%s]*z\b)r�rOr�r�r��
initCharsOrig�	initChars�
bodyCharsOrig�	bodyChars�maxSpecifiedr��minLen�maxLenr�r�r!r�r��	asKeyword�_escapeRegexRangeChars�reStringr�r��escape�compiler�r��re_match�
_WordRegexr�)r�r�r��min�max�exactrr�r�r�rr�DsZ



0�

�
z
Word.__init__Tc	Cs�|||jvrt|||j|��|}|d7}t|�}|j}||j}t||�}||krj|||vrj|d7}qLd}|||jkr�d}nV|jr�||kr�|||vr�d}n6|j	r�|dkr�||d|vs�||kr�|||vr�d}|r�t|||j|��||||�fS)Nr�FTr)
r�r?r�r�r�rrr�r�r)	r�r�r�r
r�r#�	bodycharsr��throwExceptionrrrr�}s2


�
�zWord.parseImplcstztt|���WSty"Yn0|jdurndd�}|j|jkr^d||j�||j�f|_nd||j�|_|jS)NcSs$t|�dkr|dd�dS|SdS)N�r�r@�r�rrr�
charsAsStr�sz Word.__str__.<locals>.charsAsStrz
W:(%s, %s)zW:(%s))r�rOr�r�r�r�r�)r�rr�rrr��s
zWord.__str__)Nr�rrFN)T�rrrr�r�r�r�r�rrr�rrOs49
rOc@seZdZddd�ZdS)rTcCs4|�||�}|s t|||j|��|��}||��fSr�)rr?r��end�group)r�r�r�r
r�rrrr��s
z_WordRegex.parseImplN)Tr�rrrrr�srcs"eZdZdZd�fdd�	Z�ZS)rSz�A short-cut class for defining ``Word(characters, exact=1)``,
    when defining a match of any single character in a string of
    characters.
    FNcsZtt|�j|d||d�dtd�|j��|_|r>d|j|_t�|j�|_|jj	|_
dS)Nr�)r
rr��[%s]r�z\b%s\b)r�rSr�rr�r�rr�rr�r)r��charsetrr�r�rrr��sz
Char.__init__)FNr�rrr�rrS�srScsTeZdZdZd�fdd�	Zddd�Zdd	d
�Zddd�Z�fd
d�Zdd�Z	�Z
S)rGahToken for matching strings that match a given regular
    expression. Defined with string specifying the regular expression in
    a form recognized by the stdlib Python  `re module <https://docs.python.org/3/library/re.html>`_.
    If the given regex contains named groups (defined using ``(?P<name>...)``),
    these will be preserved as named parse results.

    If instead of the Python stdlib re module you wish to use a different RE module
    (such as the `regex` module), you can replace it by either building your
    Regex object with a compiled RE that was compiled using regex:

    Example::

        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: https://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")

        # use regex module instead of stdlib re module to construct a Regex using
        # a compiled regular expression
        import regex
        parser = pp.Regex(regex.compile(r'[0-9]'))

    rFcs"tt|���t|t�r�|s,tjdtdd�||_||_	zt
�|j|j	�|_
|j|_Wq�t
jy�tjd|tdd��Yq�0n8t|d�r�t|d�r�||_
|j|_|_||_	ntd��|j
j|_t|�|_d|j|_d	|_|�d
�du|_||_||_|j�r|j|_|j�r|j|_dS)aThe parameters ``pattern`` and ``flags`` are passed
        to the ``re.compile()`` function as-is. See the Python
        `re module <https://docs.python.org/3/library/re.html>`_ module for an
        explanation of the acceptable patterns and flags.
        z0null string passed to Regex; use Empty() insteadr�rz�$invalid pattern (%s) passed to Regex�patternr�zCRegex may only be constructed with a string or a compiled RE objectr�Fr�N)r�rGr�r�r/r~rr�r�flagsr�rr�
sre_constants�errorrQr1r�rr�r!r�r�r��asGroupList�asMatch�parseImplAsGroupListr��parseImplAsMatch)r�rrrrr�rrr��s@
�
�


zRegex.__init__Tc	Csb|�||�}|s t|||j|��|��}t|���}|��}|rZ|��D]\}}|||<qH||fSr�)rr?r�rrBr�	groupdictr?)	r�r�r�r
r�r��dr8r5rrrr�
s
zRegex.parseImplcCs8|�||�}|s t|||j|��|��}|��}||fSr�)rr?r�r�groups�r�r�r�r
r�r�rrrr
szRegex.parseImplAsGroupListcCs4|�||�}|s t|||j|��|��}|}||fSr�)rr?r�rr!rrrr!
szRegex.parseImplAsMatchcsDztt|���WSty"Yn0|jdur>dt|j�|_|jS)NzRe:(%s))r�rGr�r�r�rrr�r�rrr�*
s
z
Regex.__str__csl�jrtjdtdd�t���jr@t��r@tjdtdd�t���jrT�fdd�}n��fdd�}��|�S)a�
        Return Regex with an attached parse action to transform the parsed
        result as if called using `re.sub(expr, repl, string) <https://docs.python.org/3/library/re.html#re.sub>`_.

        Example::

            make_html = Regex(r"(\w+):(.*?):").sub(r"<\1>\2</\1>")
            print(make_html.transformString("h1:main title:"))
            # prints "<h1>main title</h1>"
        z-cannot use sub() with Regex(asGroupList=True)r�rzz9cannot use sub() with a callable with Regex(asMatch=True)cs|d���Sr)�expand�r0)�replrrr�K
szRegex.sub.<locals>.pacs�j��|d�Sr)r�r9r#�r$r�rrr�N
s)rr~rr��SyntaxErrorrrr)r�r$r�rr%rr95
s��z	Regex.sub)rFF)T)T)T)rrrr�r�r�rrr�r9r�rrr�rrG�s-

	
	rGcs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
rEa&
    Token for matching strings that are delimited by quoting characters.

    Defined with the following parameters:

        - quoteChar - string of one or more characters defining the
          quote delimiting string
        - escChar - character to escape quotes, typically backslash
          (default= ``None``)
        - escQuote - special quote sequence to escape an embedded quote
          string (such as SQL's ``""`` to escape an embedded ``"``)
          (default= ``None``)
        - multiline - boolean indicating whether quotes can span
          multiple lines (default= ``False``)
        - unquoteResults - boolean indicating whether the matched text
          should be unquoted (default= ``True``)
        - endQuoteChar - string of one or more characters defining the
          end of the quote delimited string (default= ``None``  => same as
          quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace
          (``'\t'``, ``'\n'``, etc.) to actual whitespace
          (default= ``True``)

    Example::

        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))

    prints::

        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    NFTc
sVtt����|��}|s0tjdtdd�t��|dur>|}n"|��}|s`tjdtdd�t��|�_t	|��_
|d�_|�_t	|��_
|�_|�_|�_|�_|r�tjtjB�_dt��j�t�jd�|dur�t|�p�df�_n<d�_dt��j�t�jd�|du�rt|��pdf�_t	�j�d	k�rp�jd
d��fdd
�tt	�j�d	dd�D��d7_|�r��jdt�|�7_|�r��jdt�|�7_t��j�d�_�jdt��j�7_z(t��j�j��_�j�_�jj�_Wn.t j!�y.tjd�jtdd��Yn0t"���_#d�j#�_$d�_%d�_&dS)Nz$quoteChar cannot be the empty stringr�rzz'endQuoteChar cannot be the empty stringrz%s(?:[^%s%s]r�z%s(?:[^%s\n\r%s]r�z|(?:z)|(?:c3s4|],}dt��jd|��t�j|�fVqdS)z%s[^%s]N)r�r�endQuoteCharrr|r�rrr��
s��z(QuotedString.__init__.<locals>.<genexpr>r��)z|(?:%s)z|(?:%s.)z(.)z)*%srr�FT)'r�rEr�r�r~rr�r&�	quoteCharr��quoteCharLen�firstQuoteCharr'�endQuoteCharLen�escChar�escQuote�unquoteResults�convertWhitespaceEscapesr��	MULTILINE�DOTALLrrrrr�r<�escCharReplacePatternrrr�rrrr�r!r�r�r�)r�r)r-r.�	multiliner/r'r0r�r�rr�y
st


�
�����
zQuotedString.__init__c	Cs�|||jkr|�||�pd}|s2t|||j|��|��}|��}|jr�||j|j�}t	|t
�r�d|vr�|jr�ddddd�}|��D]\}}|�
||�}q�|jr�t�|jd|�}|jr�|�
|j|j�}||fS)Nr��	r���
)�\tr�z\fz\rz\g<1>)r+rr?r�rrr/r*r,r�r/r0r?r�r-r�r9r3r.r')	r�r�r�r
r�r��ws_map�wslit�wscharrrrr��
s*
�zQuotedString.parseImplcsFztt|���WSty"Yn0|jdur@d|j|jf|_|jS)Nz.quoted string, starting with %s ending with %s)r�rEr�r�r�r)r'r�r�rrr��
s
zQuotedString.__str__)NNFTNT)Trrrr�rrER
s&�A
#rEcs8eZdZdZd�fdd�	Zddd�Z�fd	d
�Z�ZS)
r(a�Token for matching words composed of characters *not* in a given
    set (will include whitespace in matched characters if not listed in
    the provided exclusion set - see example). Defined with string
    containing all disallowed characters, and an optional minimum,
    maximum, and/or exact length.  The default value for ``min`` is
    1 (a minimum value < 1 is not valid); the default values for
    ``max`` and ``exact`` are 0, meaning no maximum or exact
    length restriction.

    Example::

        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))

    prints::

        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    r�rcs�tt|���d|_||_|dkr*td��||_|dkr@||_nt|_|dkrZ||_||_t	|�|_
d|j
|_|jdk|_d|_
dS)NFr�zfcannot specify a minimum length < 1; use Optional(CharsNotIn()) if zero-length char group is permittedrr�)r�r(r�r��notCharsr�r�rr�r�r!r�r�r�)r�r<rr	r
r�rrr��
s 
zCharsNotIn.__init__TcCs�|||jvrt|||j|��|}|d7}|j}t||jt|��}||krb|||vrb|d7}qD|||jkr�t|||j|��||||�fSr!)r<r?r�rrr�r�)r�r�r�r
r��notchars�maxlenrrrr�s
zCharsNotIn.parseImplcsdztt|���WSty"Yn0|jdur^t|j�dkrRd|jdd�|_nd|j|_|jS)Nr
z
!W:(%s...)z!W:(%s))r�r(r�r�r�r�r<r�r�rrr�&s
zCharsNotIn.__str__)r�rr)Trrrr�rr(�
s
r(cs`eZdZdZdddddddd	d
ddd
dddddddddddd�Zd"�fdd�	Zd#d d!�Z�ZS)$rNa�Special matching class for matching whitespace.  Normally,
    whitespace is ignored by pyparsing grammars.  This class is included
    when some whitespace structures are significant.  Define with
    a string containing the whitespace characters to be matched; default
    is ``" \t\r\n"``.  Also takes optional ``min``,
    ``max``, and ``exact`` arguments, as defined for the
    :class:`Word` class.
    z<SP>z<TAB>z<LF>z<CR>z<FF>z<NBSP>z<OGHAM_SPACE_MARK>z<MONGOLIAN_VOWEL_SEPARATOR>z	<EN_QUAD>z	<EM_QUAD>z
<EN_SPACE>z
<EM_SPACE>z<THREE-PER-EM_SPACE>z<FOUR-PER-EM_SPACE>z<SIX-PER-EM_SPACE>z<FIGURE_SPACE>z<PUNCTUATION_SPACE>z<THIN_SPACE>z<HAIR_SPACE>z<ZERO_WIDTH_SPACE>z<NNBSP>z<MMSP>z<IDEOGRAPHIC_SPACE>)r�r5r�r7r6� u u᠎u u u u u u u u u u u u​u u u � 	
r�rcs�tt����|�_��d��fdd��jD���d�dd��jD���_d�_d�j�_	|�_
|dkrt|�_nt�_|dkr�|�_|�_
dS)Nr�c3s|]}|�jvr|VqdSr�)�
matchWhiter�r�rrr�Yrz!White.__init__.<locals>.<genexpr>css|]}tj|VqdSr�)rN�	whiteStrsr�rrrr�[rTr�r)
r�rNr�rAr�r�r�r!r�r�r�rr�)r��wsrr	r
r�r�rr�Vs zWhite.__init__TcCs�|||jvrt|||j|��|}|d7}||j}t|t|��}||krb|||jvrb|d7}qB|||jkr�t|||j|��||||�fSr!)rAr?r�rrr�r�)r�r�r�r
r�r�rrrr�js

zWhite.parseImpl)r@r�rr)T)rrrr�rBr�r�r�rrr�rrN4s6	�rNcseZdZ�fdd�Z�ZS)�_PositionTokencs(tt|���|jj|_d|_d|_dSr)r�rDr�r�rr!r�r�r�r�rrr�{s
z_PositionToken.__init__�rrrr�r�rrr�rrDzsrDcs2eZdZdZ�fdd�Zdd�Zd	dd�Z�ZS)
r/zaToken to advance to a specific column of input text; useful for
    tabular report scraping.
    cstt|���||_dSr�)r�r/r�rZ)r��colnor�rrr��szGoToColumn.__init__cCs\t||�|jkrXt|�}|jr*|�||�}||krX||��rXt||�|jkrX|d7}q*|Sr!)rZr�r�r �isspace)r�r�r�r#rrrr$�s$
zGoToColumn.preParseTcCsDt||�}||jkr"t||d|��||j|}|||�}||fS)NzText not in expected column�rZr?)r�r�r�r
�thiscol�newlocr�rrrr��s

zGoToColumn.parseImpl)T)rrrr�r�r$r�r�rrr�rr/�s	r/cs*eZdZdZ�fdd�Zddd�Z�ZS)r3a�Matches if current position is at the beginning of a line within
    the parse string

    Example::

        test = '''\
        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)

    prints::

        ['AAA', ' this line']
        ['AAA', ' and this line']

    cstt|���d|_dS)NzExpected start of line)r�r3r�r�r�r�rrr��szLineStart.__init__TcCs*t||�dkr|gfSt|||j|��dSr!)rZr?r�r%rrrr��szLineStart.parseImpl)Tr�rrr�rr3�sr3cs*eZdZdZ�fdd�Zddd�Z�ZS)r2zTMatches if current position is at the end of a line within the
    parse string
    cs,tt|���|�tj�dd��d|_dS)Nr�r�zExpected end of line)r�r2r�r�rDr�r�r�r�r�rrr��szLineEnd.__init__TcCsb|t|�kr6||dkr$|ddfSt|||j|��n(|t|�krN|dgfSt|||j|��dS)Nr�r��r�r?r�r%rrrr��szLineEnd.parseImpl)Tr�rrr�rr2�sr2cs*eZdZdZ�fdd�Zddd�Z�ZS)rJzLMatches if current position is at the beginning of the parse
    string
    cstt|���d|_dS)NzExpected start of text)r�rJr�r�r�r�rrr��szStringStart.__init__TcCs0|dkr(||�|d�kr(t|||j|��|gfSr)r$r?r�r%rrrr��szStringStart.parseImpl)Tr�rrr�rrJ�srJcs*eZdZdZ�fdd�Zddd�Z�ZS)rIzBMatches if current position is at the end of the parse string
    cstt|���d|_dS)NzExpected end of text)r�rIr�r�r�r�rrr��szStringEnd.__init__TcCs^|t|�krt|||j|��n<|t|�kr6|dgfS|t|�krJ|gfSt|||j|��dSr!rKr%rrrr��szStringEnd.parseImpl)Tr�rrr�rrI�srIcs.eZdZdZef�fdd�	Zddd�Z�ZS)rQayMatches if the current position is at the beginning of a Word,
    and is not preceded by any character in a given set of
    ``wordChars`` (default= ``printables``). To emulate the
    ```` behavior of regular expressions, use
    ``WordStart(alphanums)``. ``WordStart`` will also match at
    the beginning of the string being parsed, or at the beginning of
    a line.
    cs"tt|���t|�|_d|_dS)NzNot at the start of a word)r�rQr�r��	wordCharsr��r�rLr�rrr��s
zWordStart.__init__TcCs@|dkr8||d|jvs(|||jvr8t|||j|��|gfSr;)rLr?r�r%rrrr��s�zWordStart.parseImpl)T�rrrr�rwr�r�r�rrr�rrQ�srQcs.eZdZdZef�fdd�	Zddd�Z�ZS)rPa_Matches if the current position is at the end of a Word, and is
    not followed by any character in a given set of ``wordChars``
    (default= ``printables``). To emulate the ```` behavior of
    regular expressions, use ``WordEnd(alphanums)``. ``WordEnd``
    will also match at the end of the string being parsed, or at the end
    of a line.
    cs(tt|���t|�|_d|_d|_dS)NFzNot at the end of a word)r�rPr�r�rLr�r�rMr�rrr�s
zWordEnd.__init__TcCsPt|�}|dkrH||krH|||jvs8||d|jvrHt|||j|��|gfSr;)r�rLr?r�)r�r�r�r
r#rrrr�s�zWordEnd.parseImpl)TrNrrr�rrPsrPcszeZdZdZd�fdd�	Zdd�Zdd�Z�fd	d
�Z�fdd�Z�fd
d�Z	ddd�Z
�fdd�Zd�fdd�	Z�Z
S)r@z]Abstract subclass of ParserElement, for combining and
    post-processing parsed tokens.
    Fcs�tt���|�t|t�r"t|�}t|t�r<��|�g�_n�t|t	�rP|g�_nlt|t
�r�t|�}tdd�|D��r��fdd�|D�}t|��_n*zt|��_Wnty�|g�_Yn0d�_
dS)Ncss|]}t|t�VqdSr�)r�r/)rr�rrrr�*rz+ParseExpression.__init__.<locals>.<genexpr>c3s&|]}t|t�r��|�n|VqdSr�)r�r/r��rrr�rrr�+rF)r�r@r�r�r,r*r/r��exprsrDr
r�r1r��r�rPr�r�r�rr�s"




zParseExpression.__init__cCs|j�|�d|_|Sr�)rPr�r�rxrrrr�4szParseExpression.appendcCs0d|_dd�|jD�|_|jD]}|��q|S)z�Extends ``leaveWhitespace`` defined in base class, and also invokes ``leaveWhitespace`` on
           all contained expressions.FcSsg|]}|���qSrrkrOrrrr=rz3ParseExpression.leaveWhitespace.<locals>.<listcomp>)r�rPr�)r�rrrrr�9s


zParseExpression.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jD]}|�|jd�q*n,tt|��|�|jD]}|�|jd�qX|SrN)r�rKr�r�r@r�rP)r�rmrr�rrr�Bs



zParseExpression.ignorecsLztt|���WSty"Yn0|jdurFd|jjt|j�f|_|jS�Nz%s:(%s))	r�r@r�r�r�r�rr�rPr�r�rrr�Ns
zParseExpression.__str__cs*tt|���|jD]}|��qt|j�dk�r|jd}t||j�r�|js�|jdur�|j	s�|jdd�|jdg|_d|_
|j|jO_|j|jO_|jd}t||j��r|j�s|jdu�r|j	�s|jdd�|jdd�|_d|_
|j|jO_|j|jO_dt
|�|_|S)Nr�rr�r�r�)r�r@rarPr�r�r�r�r�r�r�r�r�r�r�)r�rrmr�rrraXs<


���
���zParseExpression.streamlineNcCsB|dur|ngdd�|g}|jD]}|�|�q$|�g�dSr�)rPr�r�)r�r��tmprrrrr�zs
zParseExpression.validatecs$tt|���}dd�|jD�|_|S)NcSsg|]}|���qSrrkrOrrrr�rz(ParseExpression.copy.<locals>.<listcomp>)r�r@rlrPr�r�rrrl�szParseExpression.copycsVtjrD|jD]6}t|t�r|jrtjd�d|t	|�j
|j�dd�qtt|��
||�S)N�]{0}: setting results name {1!r} on {2} expression collides with {3!r} on contained expressionrr�rz)rrrPr�rDr�r~rr�r�rr�r@r�r�r!rrr�rrr�s
��zParseExpression._setResultsName)F)N)F)rrrr�r�r�r�r�r�rar�rlrr�rrr�rr@s	
"
r@cs`eZdZdZGdd�de�Zd�fdd�	Z�fdd�Zdd	d
�Zdd�Z	d
d�Z
dd�Z�ZS)r%a
    Requires all given :class:`ParseExpression` s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the ``'+'`` operator.
    May also be constructed using the ``'-'`` operator, which will
    suppress backtracking.

    Example::

        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"), name_expr("name"), integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    cseZdZ�fdd�Z�ZS)zAnd._ErrorStopcs*ttj|�j|i|��d|_|��dS)N�-)r�r%r�r�r!r�r�r�rrr��szAnd._ErrorStop.__init__rErrr�rr��sr�Tcs�t|�}|r�t|vr�g}t|�D]`\}}|turv|t|�dkrlt�||djd}|�t|�d��q�td��q |�|�q ||dd�<t	t
|��||�tdd�|jD��|_
|�|jdj�|jdj|_d|_dS)	Nr�r�r�z0cannot construct And with sequence ending in ...css|]}|jVqdSr��r�rOrrrr��rzAnd.__init__.<locals>.<genexpr>rT)r*r|r�r�r,rPr�rHr�r�r%r�rr�r�r�r�r�)r�rPr�rSrr��
skipto_argr�rrr��s 
zAnd.__init__cs�|jr�tdd�|jdd�D��r�t|jdd��D]^\}}|durFq4t|t�r4|jr4t|jdt�r4|jd|j|d|jd<d|j|d<q4dd�|jD�|_tt|���t	dd�|jD��|_
|S)Ncss.|]&}t|t�o$|jo$t|jdt�VqdS�r�N)r�r@rPr}rOrrrr��s�z!And.streamline.<locals>.<genexpr>r�r�cSsg|]}|dur|�qSr�rrOrrrr�rz"And.streamline.<locals>.<listcomp>css|]}|jVqdSr�rWrOrrrr��r)rPr�r�r�r@r}r�r%rarr�)r�rrr�rrra�s$�
��zAnd.streamlinec	Cs�|jdj|||dd�\}}d}|jdd�D]�}t|tj�rDd}q.|r�z|�|||�\}}Wq�tyr�Yq�ty�}zd|_t�|��WYd}~q�d}~0t	y�t|t
|�|j|��Yq�0n|�|||�\}}|s�|��r.||7}q.||fS)NrFrgr�T)
rPrr�r%r�rCr=r�r�r2r�r�r[)	r�r�r�r
�
resultlist�	errorStopr�
exprtokensr�rrrr��s(
z
And.parseImplcCst|t�r|�|�}|�|�Sr��r�r/r�r�rxrrrrf�s

zAnd.__iadd__cCs6|dd�|g}|jD]}|�|�|jsq2qdSr�)rPr�r��r�r�subRecCheckListrrrrr��s


zAnd.checkRecursioncCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr!�{r�css|]}t|�VqdSr�r�rOrrrr��rzAnd.__str__.<locals>.<genexpr>�}�rQr!r�r�rPr�rrrr��s


 zAnd.__str__)T)T)
rrrr�r,r�r�rar�rfr�r�r�rrr�rr%�s
r%cs^eZdZdZd�fdd�	Z�fdd�Zddd	�Zd
d�Zdd
�Zdd�Z	d�fdd�	Z
�ZS)r<a�Requires that at least one :class:`ParseExpression` is found. If
    two expressions match, the expression that matches the longest
    string will be used. May be constructed using the ``'^'``
    operator.

    Example::

        # construct Or using '^' operator

        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))

    prints::

        [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�rWrOrrrr�rzOr.__init__.<locals>.<genexpr>T)r�r<r�rPr�r�rQr�rrr�szOr.__init__cs.tt|���tjr*tdd�|jD��|_|S)Ncss|]}|jVqdSr��r�rOrrrr�rz Or.streamline.<locals>.<genexpr>)r�r<rar$�collect_all_And_tokensr�rPr�r�r�rrrasz
Or.streamlineTcCs�d}d}g}|jD]�}z|�||�}Wnvtyd}	z&d|	_|	j|krP|	}|	j}WYd}	~	qd}	~	0ty�t|�|kr�t|t|�|j|�}t|�}Yq0|�||f�q|�r�|j	t
d�dd�|s�|dd}
|
�|||�Sd}|D]�\}}
||dk�r
|Sz|
�|||�\}}WnBt�yb}	z(d|	_|	j|k�rN|	}|	j}WYd}	~	q�d}	~	00||k�rz||fS||dkr�||f}q�|dk�r�|S|du�r�|j|_|�nt||d|��dS)Nr�rT)rar>r�rY� no defined alternatives to match)
rPr6r?r�r�r2r�r�r��sortrrr�)r�r�r�r
�	maxExcLoc�maxExceptionrlr�loc2r1�	best_expr�longest�loc1�expr1r�rrrr�sT





zOr.parseImplcCst|t�r|�|�}|�|�Sr�r]rxrrr�__ixor__[s

zOr.__ixor__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr!r`z ^ css|]}t|�VqdSr�r�rOrrrr�erzOr.__str__.<locals>.<genexpr>rarbr�rrrr�`s


 z
Or.__str__cCs,|dd�|g}|jD]}|�|�qdSr��rPr�r^rrrr�is
zOr.checkRecursioncsPtjs>tjr>tdd�|jD��r>tjd�d|t	|�j
�dd�tt|��
||�S)Ncss|]}t|t�VqdSr��r�r%rOrrrr�qrz%Or._setResultsName.<locals>.<genexpr>��{0}: setting results name {1!r} on {2} expression may only return a single token for an And alternative, in future will return the full list of tokensrr�rz)r$rdrrr�rPr~rr�r�rr�r<rrr�rrrns���zOr._setResultsName)F)T)F)rrrr�r�rar�rnr�r�rr�rrr�rr<s
=	r<cs^eZdZdZd�fdd�	Z�fdd�Zddd	�Zd
d�Zdd
�Zdd�Z	d�fdd�	Z
�ZS)r6a�Requires that at least one :class:`ParseExpression` is found. If
    two expressions match, the first one listed is the one that will
    match. May be constructed using the ``'|'`` operator.

    Example::

        # construct MatchFirst using '|' operator

        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    Fcs:tt|��||�|jr0tdd�|jD��|_nd|_dS)Ncss|]}|jVqdSr�rWrOrrrr��rz&MatchFirst.__init__.<locals>.<genexpr>T)r�r6r�rPr�r�rQr�rrr��szMatchFirst.__init__cs.tt|���tjr*tdd�|jD��|_|S)Ncss|]}|jVqdSr�rcrOrrrr��rz(MatchFirst.streamline.<locals>.<genexpr>)r�r6rar$rdr�rPr�r�r�rrra�szMatchFirst.streamlineTc	Cs�d}d}|jD]�}z|�|||�}|WStyb}z |j|krN|}|j}WYd}~qd}~0ty�t|�|kr�t|t|�|j|�}t|�}Yq0q|dur�|j|_|�nt||d|��dS)Nr�re)rPrr?r�r2r�r�r�)	r�r�r�r
rgrhrr�r1rrrr��s$


zMatchFirst.parseImplcCst|t�r|�|�}|�|�Sr�r]rxrrr�__ior__�s

zMatchFirst.__ior__cCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr!r`� | css|]}t|�VqdSr�r�rOrrrr��rz%MatchFirst.__str__.<locals>.<genexpr>rarbr�rrrr��s


 zMatchFirst.__str__cCs,|dd�|g}|jD]}|�|�qdSr�ror^rrrr��s
zMatchFirst.checkRecursioncsPtjs>tjr>tdd�|jD��r>tjd�d|t	|�j
�dd�tt|��
||�S)Ncss|]}t|t�VqdSr�rprOrrrr��rz-MatchFirst._setResultsName.<locals>.<genexpr>rqrr�rz)r$rdrrr�rPr~rr�r�rr�r6rrr�rrr�s���zMatchFirst._setResultsName)F)T)F)rrrr�r�rar�rrr�r�rr�rrr�rr6{s
	r6csHeZdZdZd
�fdd�	Z�fdd�Zddd�Zd	d
�Zdd�Z�Z	S)r+asRequires all given :class:`ParseExpression` s to be found, but in
    any order. Expressions may be separated by whitespace.

    May be constructed using the ``'&'`` operator.

    Example::

        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )

    prints::

        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    Tcs>tt|��||�tdd�|jD��|_d|_d|_d|_dS)Ncss|]}|jVqdSr�rWrOrrrr�rz Each.__init__.<locals>.<genexpr>T)	r�r+r�rrPr�r��initExprGroupsr�rQr�rrr�
s
z
Each.__init__cs(tt|���tdd�|jD��|_|S)Ncss|]}|jVqdSr�rWrOrrrr�rz"Each.streamline.<locals>.<genexpr>)r�r+rarrPr�r�r�rrraszEach.streamlinec	s�|jr�tdd�|jD��|_dd�|jD�}dd�|jD�}|||_dd�|jD�|_dd�|jD�|_dd�|jD�|_|j|j7_d	|_|}|jdd�}|jdd��g}d
}	|	�rh|�|j|j}
g}|
D]t}z|�||�}Wnt	�y|�
|�Yq�0|�
|j�t|�|��||v�r>|�
|�q�|�vr܈�
|�q�t|�t|
�kr�d	}	q�|�r�d�dd�|D��}
t	||d
|
��|�fdd�|jD�7}g}|D]"}|�|||�\}}|�
|��q�t|tg��}||fS)Ncss&|]}t|t�rt|j�|fVqdSr�)r�r;r�r�rOrrrr�rz!Each.parseImpl.<locals>.<genexpr>cSsg|]}t|t�r|j�qSr�r�r;r�rOrrrrrz"Each.parseImpl.<locals>.<listcomp>cSs$g|]}|jrt|ttf�s|�qSr)r�r�r;rGrOrrrrrcSsg|]}t|t�r|j�qSr)r�rRr�rOrrrrrcSsg|]}t|t�r|j�qSr)r�r9r�rOrrrrrcSs g|]}t|tttf�s|�qSr)r�r;rRr9rOrrrrrFTr{css|]}t|�VqdSr�r�rOrrrr�9rz*Missing one or more required elements (%s)cs$g|]}t|t�r|j�vr|�qSrrurO��tmpOptrrr=r)rtr-rP�opt1map�	optionals�multioptionals�
multirequired�requiredr6r?r�r�r��remover�r�r�sumrB)r�r�r�r
�opt1�opt2�tmpLoc�tmpReqd�
matchOrder�keepMatching�tmpExprs�failedr�missingrZr��finalResultsrrvrr�sP

zEach.parseImplcCs@t|d�r|jS|jdur:dd�dd�|jD��d|_|jS)Nr!r`z & css|]}t|�VqdSr�r�rOrrrr�LrzEach.__str__.<locals>.<genexpr>rarbr�rrrr�Gs


 zEach.__str__cCs,|dd�|g}|jD]}|�|�qdSr�ror^rrrr�Ps
zEach.checkRecursion)T)T)
rrrr�r�rar�r�r�r�rrr�rr+�s8
1	r+csjeZdZdZd�fdd�	Zddd�Zdd	�Z�fd
d�Z�fdd
�Zdd�Z	ddd�Z
�fdd�Z�ZS)r>zfAbstract subclass of :class:`ParserElement`, for combining and
    post-processing parsed tokens.
    Fcs�tt|��|�t|t�r@t|jt�r2|�|�}n|�t|��}||_	d|_
|dur�|j|_|j|_|�
|j�|j|_|j|_|j|_|j�|j�dSr�)r�r>r�r�r/�
issubclassr�rLr4r�r�r�r�r�r�r�r�r�r�rg�r�r�r�r�rrr�Zs
zParseElementEnhance.__init__TcCs2|jdur|jj|||dd�Std||j|��dS)NFrgr�)r�rr?r�r%rrrr�ls
zParseElementEnhance.parseImplcCs*d|_|j��|_|jdur&|j��|Sr�)r�r�rlr�r�rrrr�rs


z#ParseElementEnhance.leaveWhitespacecsrt|t�rB||jvrntt|��|�|jdurn|j�|jd�n,tt|��|�|jdurn|j�|jd�|SrN)r�rKr�r�r>r�r�rxr�rrr�ys



zParseElementEnhance.ignorecs&tt|���|jdur"|j��|Sr�)r�r>rar�r�r�rrra�s

zParseElementEnhance.streamlinecCsB||vrt||g��|dd�|g}|jdur>|j�|�dSr�)rFr�r�)r�rr_rrrr��s

z"ParseElementEnhance.checkRecursionNcCsB|durg}|dd�|g}|jdur4|j�|�|�g�dSr��r�r�r��r�r�rSrrrr��s
zParseElementEnhance.validatecsVztt|���WSty"Yn0|jdurP|jdurPd|jjt|j�f|_|jSrR)	r�r>r�r�r�r�r�rr�r�r�rrr��szParseElementEnhance.__str__)F)T)N)
rrrr�r�r�r�r�rar�r�r�r�rrr�rr>Vs

r>cs*eZdZdZ�fdd�Zddd�Z�ZS)r-abLookahead matching of the given parse expression.
    ``FollowedBy`` does *not* advance the parsing position within
    the input string, it only verifies that the specified parse
    expression matches at the current position.  ``FollowedBy``
    always returns a null token list. If any results names are defined
    in the lookahead expression, those *will* be returned for access by
    name.

    Example::

        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))

        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()

    prints::

        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    cstt|��|�d|_dSr)r�r-r�r��r�r�r�rrr��szFollowedBy.__init__TcCs(|jj|||d�\}}|dd�=||fS)Nr4)r�r)r�r�r�r
�_r�rrrr��s
zFollowedBy.parseImpl)Tr�rrr�rr-�sr-cs,eZdZdZd	�fdd�	Zd
dd�Z�ZS)r5apLookbehind matching of the given parse expression.
    ``PrecededBy`` does not advance the parsing position within the
    input string, it only verifies that the specified parse expression
    matches prior to the current position.  ``PrecededBy`` always
    returns a null token list, but if a results name is defined on the
    given expression, it is returned.

    Parameters:

     - expr - expression that must match prior to the current parse
       location
     - retreat - (default= ``None``) - (int) maximum number of characters
       to lookbehind prior to the current parse location

    If the lookbehind expression is a string, Literal, Keyword, or
    a Word or CharsNotIn with a specified exact or maximum length, then
    the retreat parameter is not required. Otherwise, retreat must be
    specified to give a maximum number of characters to look back from
    the current parse position for a lookbehind match.

    Example::

        # VB-style variable names with type prefixes
        int_var = PrecededBy("#") + pyparsing_common.identifier
        str_var = PrecededBy("$") + pyparsing_common.identifier

    Ncs�tt|��|�|����|_d|_d|_d|_t|t	�rJt
|�}d|_nVt|ttf�rf|j
}d|_n:t|ttf�r�|jtkr�|j}d|_nt|t�r�d}d|_||_dt	|�|_d|_|j�dd��dS)NTFrznot preceded by cSs|�tdd��Sr�)rFr6r�rrrr��rz%PrecededBy.__init__.<locals>.<lambda>)r�r5r�r�r�r�r�r
r�r�r�r4r1r�rOr(rr�rD�retreatr�r�r�r�)r�r�r�r�rrr��s*

zPrecededBy.__init__rTcCs�|jr<||jkrt|||j��||j}|j�||�\}}n�|jt�}|td||j�|�}t|||j�}	tdt	||jd�d�D]N}
z|�|t
|�|
�\}}Wn(ty�}z|}	WYd}~q�d}~00q�q�|	�||fSr;)r
r�r?r�r�rrIr	r<rr�r=)r�r�r�r
r�r�r��	test_expr�instring_slice�	last_exprrp�pberrrr��s 

zPrecededBy.parseImpl)N)rTr�rrr�rr5�sr5cs2eZdZdZ�fdd�Zd	dd�Zdd�Z�ZS)
r8a�Lookahead to disallow matching with the given parse expression.
    ``NotAny`` does *not* advance the parsing position within the
    input string, it only verifies that the specified parse expression
    does *not* match at the current position.  Also, ``NotAny`` does
    *not* skip over leading whitespace. ``NotAny`` always returns
    a null token list.  May be constructed using the '~' operator.

    Example::

        AND, OR, NOT = map(CaselessKeyword, "AND OR NOT".split())

        # take care not to mistake keywords for identifiers
        ident = ~(AND | OR | NOT) + Word(alphas)
        boolean_term = Optional(NOT) + ident

        # very crude boolean expression - to support parenthesis groups and
        # operation hierarchy, use infixNotation
        boolean_expr = boolean_term + ZeroOrMore((AND | OR) + boolean_term)

        # integers that are followed by "." are actually floats
        integer = Word(nums) + ~Char(".")
    cs0tt|��|�d|_d|_dt|j�|_dS)NFTzFound unwanted token, )r�r8r�r�r�r�r�r�r�r�rrr�*szNotAny.__init__TcCs&|j�||�rt|||j|��|gfSr�)r�r7r?r�r%rrrr�1szNotAny.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr!z~{ra�rQr!r�r�r�r�rrrr�6s


zNotAny.__str__)Trrrr�rr8s
r8cs>eZdZd�fdd�	Zdd�Zd
dd�Zd�fd
d�	Z�ZS)�_MultipleMatchNcs<tt|��|�d|_|}t|t�r.|�|�}|�|�dSr)r�r�r�r�r�r/r��stopOn)r�r�r��enderr�rrr�@s

z_MultipleMatch.__init__cCs,t|t�r|�|�}|dur"|nd|_|Sr�)r�r/r��	not_ender)r�r�rrrr�Hs

z_MultipleMatch.stopOnTc	Cs�|jj}|j}|jdu}|r$|jj}|r2|||�||||dd�\}}zV|j}	|r`|||�|	rp|||�}
n|}
|||
|�\}}|s�|��rR||7}qRWnttfy�Yn0||fS�NFrg)	r�rr r�r6r�r[r?r2)r�r�r�r
�self_expr_parse�self_skip_ignorables�check_ender�
try_not_enderr0�hasIgnoreExprsr.�	tmptokensrrrr�Ns*



z_MultipleMatch.parseImplFcsftjrT|jgt|jdg�D]6}t|t�r|jrtjd�	d|t
|�j|j�dd�qtt
|��||�S)NrPrTrr�rz)rrr�r�r�rDr�r~rr�r�rr�r�rrUr�rrrks��z_MultipleMatch._setResultsName)N)T)F)rrrr�r�r�rr�rrr�rr�?s
r�c@seZdZdZdd�ZdS)r9ajRepetition of one or more of the given expression.

    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default= ``None``) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition
          expression)

    Example::

        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]

        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    cCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr!r`z}...r�r�rrrr��s


zOneOrMore.__str__N)rrrr�r�rrrrr9ysr9cs8eZdZdZd
�fdd�	Zd�fdd�	Zdd	�Z�ZS)rRakOptional repetition of zero or more of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default= ``None``) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition
          expression)

    Example: similar to :class:`OneOrMore`
    Ncstt|�j||d�d|_dS)N�r�T)r�rRr�r�)r�r�r�r�rrr��szZeroOrMore.__init__Tc	s:ztt|��|||�WSttfy4|gfYS0dSr�)r�rRr�r?r2r%r�rrr��szZeroOrMore.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr!rz�]...r�r�rrrr��s


zZeroOrMore.__str__)N)Trrrr�rrR�s
rRc@s eZdZdd�ZeZdd�ZdS)�
_NullTokencCsdSr�rr�rrrrJ�sz_NullToken.__bool__cCsdSr�rr�rrrr��sz_NullToken.__str__N)rrrrJr�r�rrrrr��sr�cs<eZdZdZe�Zef�fdd�	Zd	dd�Zdd�Z�Z	S)
r;aGOptional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::

        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345

            # ZIP+4 form
            12101-0001

            # invalid ZIP
            98765-
            ''')

    prints::

        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    cs.tt|�j|dd�|jj|_||_d|_dS)NFr�T)r�r;r�r�r�rbr�)r�r�r\r�rrr��s
zOptional.__init__Tc	Cszz|jj|||dd�\}}WnTttfyp|j|jurh|jjr^t|jg�}|j||jj<ql|jg}ng}Yn0||fSr�)r�rr?r2rb�_Optional__optionalNotMatchedr�rB)r�r�r�r
r0rrrr��s

zOptional.parseImplcCs4t|d�r|jS|jdur.dt|j�d|_|jS)Nr!rzr}r�r�rrrr��s


zOptional.__str__)T)
rrrr�r�r�r�r�r�r�rrr�rr;�s
$
r;cs,eZdZdZd	�fdd�	Zd
dd�Z�ZS)rHa�	Token for skipping over all undefined text until the matched
    expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default= ``False``) if True, the target expression is also parsed
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default= ``None``) used to define grammars (typically quoted strings and
          comments) that might contain false matches to the target expression
     - failOn - (default= ``None``) define expressions that are not allowed to be
          included in the skipped test; if found before the target expression is found,
          the SkipTo is not a match

    Example::

        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP
                      + string_data("sev") + SEP
                      + string_data("desc") + SEP
                      + integer("days_open"))

        for tkt in ticket_expr.searchString(report):
            print tkt.dump()

    prints::

        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    FNcs`tt|��|�||_d|_d|_||_d|_t|t	�rF|�
|�|_n||_dt|j
�|_dS)NTFzNo match found for )r�rHr��
ignoreExprr�r��includeMatchr�r�r/r��failOnr�r�r�)r�rm�includer�r�r�rrr�@s
zSkipTo.__init__Tc	Cs"|}t|�}|j}|jj}|jdur,|jjnd}|jdurB|jjnd}	|}
|
|kr�|durf|||
�rfq�|	dur�z|	||
�}
Wqnty�Yq�Yqn0qnz|||
ddd�Wq�tt	fy�|
d7}
YqJ0q�qJt|||j
|��|
}|||�}t|�}|j�r||||dd�\}}
||
7}||fS)NF)r
rr�rg)
r�r�rr�r7r�r6r=r?r2r�rBr�)r�r�r�r
r�r#r��
expr_parse�self_failOn_canParseNext�self_ignoreExpr_tryParse�tmploc�skiptext�
skipresultr�rrrr�Ms:
zSkipTo.parseImpl)FNN)Tr�rrr�rrHs9
rHcsneZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Zdd�Zdd
d�Z	dd�Z
�fdd�Zd�fdd�	Z�Z
S)r.a_Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the ``Forward``
    variable using the '<<' operator.

    Note: take care when assigning to ``Forward`` not to overlook
    precedence of operators.

    Specifically, '|' has a lower precedence than '<<', so that::

        fwdExpr << a | b | c

    will actually be evaluated as::

        (fwdExpr << a) | b | c

    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the ``Forward``::

        fwdExpr << (a | b | c)

    Converting to use the '<<=' operator instead will avoid this problem.

    See :class:`ParseResults.pprint` for an example of a recursive
    parser created using ``Forward``.
    Ncstt|�j|dd�dSr�)r�r.r�rxr�rrr��szForward.__init__cCsjt|t�r|�|�}||_d|_|jj|_|jj|_|�|jj�|jj	|_	|jj
|_
|j�|jj�|Sr�)
r�r/r�r�r�r�r�r�r�r�r�r�rgrxrrr�
__lshift__�s





zForward.__lshift__cCs||>Sr�rrxrrr�__ilshift__�szForward.__ilshift__cCs
d|_|Sr�r�r�rrrr��szForward.leaveWhitespacecCs$|js d|_|jdur |j��|Sr)r�r�rar�rrrra�s


zForward.streamlinecCsJ|durg}||vr<|dd�|g}|jdur<|j�|�|�g�dSr�r�r�rrrr��s
zForward.validatecCs~t|d�r|jS|jdur |jSd|_d}z8|jdurJt|j�dd�}nd}W|jjd||_n|jjd||_0|jS)Nr!z: ...r�i��Nonez: )rQr!r�r�r�r�r)r��	retStringrrrr��s


(zForward.__str__cs.|jdurtt|���St�}||K}|SdSr�)r�r�r.rlr�r�rrrl�s

zForward.copyFcs@tjr.|jdur.tjd�d|t|�j�dd�tt	|��
||�S)NzR{0}: setting results name {0!r} on {1} expression that has no contained expressionrr�rz)rrr�r~rr�r�rr�r.rrr�rrr�s
��zForward._setResultsName)N)N)F)rrrr�r�r�r�r�rar�r�rlrr�rrr�rr.|s


r.cs"eZdZdZd�fdd�	Z�ZS)rMzW
    Abstract subclass of :class:`ParseExpression`, for converting parsed results.
    Fcstt|��|�d|_dSr�)r�rMr�r�r�r�rrr��szTokenConverter.__init__)Fr�rrr�rrM�srMcs6eZdZdZd
�fdd�	Z�fdd�Zdd	�Z�ZS)r)a�Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the
    input string; this can be disabled by specifying
    ``'adjacent=False'`` in the constructor.

    Example::

        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    r�Tcs8tt|��|�|r|��||_d|_||_d|_dSr)r�r)r�r��adjacentr��
joinStringr�)r�r�r�r�r�rrr�szCombine.__init__cs(|jrt�||�ntt|��|�|Sr�)r�rDr�r�r)rxr�rrr�
szCombine.ignorecCsP|��}|dd�=|td�|�|j��g|jd�7}|jrH|��rH|gS|SdS)Nr�)r#)rlrBr�r~r�r�r�r[)r�r�r�r'�retToksrrrr(s
"zCombine.postParse)r�T)rrrr�r�r�r(r�rrr�rr)�s
r)cs(eZdZdZ�fdd�Zdd�Z�ZS)r0aConverter to return the matched tokens as a list - useful for
    returning tokens of :class:`ZeroOrMore` and :class:`OneOrMore` expressions.

    Example::

        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a, b, 100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a, b, 100"))  # -> ['fn', ['a', 'b', '100']]
    cstt|��|�d|_dSr)r�r0r�r�r�r�rrr�*szGroup.__init__cCs|gSr�rr&rrrr(.szGroup.postParse�rrrr�r�r(r�rrr�rr0sr0cs(eZdZdZ�fdd�Zdd�Z�ZS)r*a?Converter to return a repetitive expression as a list, but also
    as a dictionary. Each element can also be referenced using the first
    token in the expression as its key. Useful for tabular report
    scraping when the first column can be used as a item key.

    Example::

        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))

        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())

        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())

        # access named fields as dict entries, or output as dict
        print(result['shape'])
        print(result.asDict())

    prints::

        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}

    See more examples at :class:`ParseResults` of accessing fields by results name.
    cstt|��|�d|_dSr)r�r*r�r�r�r�rrr�Xsz
Dict.__init__cCs�t|�D]�\}}t|�dkrq|d}t|t�r@t|d���}t|�dkr\td|�||<qt|�dkr�t|dt�s�t|d|�||<q|��}|d=t|�dks�t|t�r�|�	�r�t||�||<qt|d|�||<q|j
r�|gS|SdS)Nrr�r�r�)r�r�r�r�r�r�rrBrlr[r�)r�r�r�r'r�tok�ikey�	dictvaluerrrr(\s$
zDict.postParser�rrr�rr*1s&r*c@s eZdZdZdd�Zdd�ZdS)rKa[Converter for ignoring the results of a parsed expression.

    Example::

        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))

    prints::

        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']

    (See also :class:`delimitedList`.)
    cCsgSr�rr&rrrr(�szSuppress.postParsecCs|Sr�rr�rrrr��szSuppress.suppressN)rrrr�r(r�rrrrrKusrKc@s(eZdZdZdd�Zdd�Zdd�ZdS)	r:zDWrapper for parse actions, to ensure they are only called once.
    cCst|�|_d|_dSr�)r�r�called)r��
methodCallrrrr��s
zOnlyOnce.__init__cCs.|js|�|||�}d|_|St||d��dS)NTr�)r�rr?)r�r�r�r�r�rrrr��s
zOnlyOnce.__call__cCs
d|_dSr�)r�r�rrr�reset�szOnlyOnce.resetN)rrrr�r�r�r�rrrrr:�sr:cs8t����fdd�}z�j|_Wnty2Yn0|S)aqDecorator for debugging parse actions.

    When the parse action is called, this decorator will print
    ``">> entering method-name(line:<current_source_line>, <parse_location>, <matched_tokens>)"``.
    When the parse action completes, the decorator will print
    ``"<<"`` followed by the returned value, or any exception that the parse action raised.

    Example::

        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))

    prints::

        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    c
s��j}|dd�\}}}t|�dkr8|djjd|}tj�d|t||�||f�z�|�}Wn:ty�}z"tj�d||f��WYd}~n
d}~00tj�d||f�|S)Nr�r�r�.z">>entering %s(line: '%s', %d, %r)
z<<leaving %s (exception: %s)
z<<leaving %s (ret: %r)
)rr�r�r��stderrr�rhr�)�paArgs�thisFuncr�r�r�r�r�r�rr�z�sztraceParseAction.<locals>.z)r�rr�)r�r�rr�rr��s
r��,cCs`t|�dt|�dt|�d}|rBt|t||���|�S|tt|�|��|�SdS)a�Helper to define a delimited list of expressions - the delimiter
    defaults to ','. By default, the list elements and delimiters can
    have intervening whitespace, and comments, but this can be
    overridden by passing ``combine=True`` in the constructor. If
    ``combine`` is set to ``True``, the matching tokens are
    returned as a single token string, with the delimiters included;
    otherwise, the matching tokens are returned as a list of tokens,
    with the delimiters suppressed.

    Example::

        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    z [r�r�N)r�r)rRrrK)r��delim�combine�dlNamerrrra�s$racsjt����fdd�}|dur0tt��dd��}n|��}|�d�|j|dd�|��d	t��d
�S)a>Helper to define a counted list of expressions.

    This helper defines a pattern of the form::

        integer expr expr expr...

    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the
    leading count token is suppressed.

    If ``intExpr`` is specified, it should be a pyparsing expression
    that produces an integer value.

    Example::

        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    cs.|d}�|r tt�g|��p&tt�>gSr)r0r%rd)r�r�r�r���	arrayExprr�rr�countFieldParseAction�s"z+countedArray.<locals>.countFieldParseActionNcSst|d�Sr)r�r�rrrr�rzcountedArray.<locals>.<lambda>�arrayLenT�r�z(len) r�)r.rOrsr�rlrrr�)r��intExprr�rr�rr]�s
r]cCs6g}|D](}t|t�r&|�t|��q|�|�q|Sr�)r�r*rgrrr�)�Lr�rrrrrr
s
rrcs6t���fdd�}|j|dd���dt|���S)a4Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks for
    a 'repeat' of a previous expression.  For example::

        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second

    will match ``"1:1"``, but not ``"1:2"``.  Because this
    matches a previous literal, will also match the leading
    ``"1:1"`` in ``"1:10"``. If this is not desired, use
    :class:`matchPreviousExpr`. Do *not* use with packrat parsing
    enabled.
    csP|rBt|�dkr�|d>qLt|���}�tdd�|D��>n
�t�>dS)Nr�rcss|]}t|�VqdSr�)r4�r�ttrrrr�*rzDmatchPreviousLiteral.<locals>.copyTokenToRepeater.<locals>.<genexpr>)r�rrr"r%r,)r�r�r��tflat��reprr�copyTokenToRepeater#sz1matchPreviousLiteral.<locals>.copyTokenToRepeaterTr��(prev) )r.rrr�)r�r�rr�rrps

rpcsFt��|��}�|K��fdd�}|j|dd���dt|���S)aTHelper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks for
    a 'repeat' of a previous expression.  For example::

        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second

    will match ``"1:1"``, but not ``"1:2"``.  Because this
    matches by expressions, will *not* match the leading ``"1:1"``
    in ``"1:10"``; the expressions are evaluated first, and then
    compared, so ``"1"`` is compared with ``"10"``. Do *not* use
    with packrat parsing enabled.
    cs*t|�����fdd�}�j|dd�dS)Ncs$t|���}|�kr tddd��dS)Nr�r)rrr"r?)r�r�r��theseTokens��matchTokensrr�mustMatchTheseTokensEszLmatchPreviousExpr.<locals>.copyTokenToRepeater.<locals>.mustMatchTheseTokensTr�)rrr"r�)r�r�r�r�r�r�rr�Csz.matchPreviousExpr.<locals>.copyTokenToRepeaterTr�r�)r.rlrrr�)r��e2r�rr�rro1srocCs:dD]}|�|t|�}q|�dd�}|�dd�}t|�S)Nz\^-[]r�r�r5r8)r��_bslashr�)r�r�rrrrNs
rcs�t|t�rtjddd�|r:dd�}dd�}|r4tnt�ndd�}dd�}|rRtnt�g}t|t�rn|��}n$t|t	�r�t
|�}ntjd	tdd�|s�t�S|�s.d
}|t
|�dk�r.||}t||dd��D]N\}	}
||
|�r�|||	d=q�q�|||
�r�|||	d=|�||
�q�q�|d7}q�|�s�|�s�|�r�zlt
|�t
d
�|��k�r�tdd
�dd�|D����d�|��WStd�dd�|D����d�|��WSWn$t�y�tjdtdd�Yn0t�fdd�|D���d�|��S)a�Helper to quickly define a set of alternative Literals, and makes
    sure to do longest-first testing when there is a conflict,
    regardless of the input order, but returns
    a :class:`MatchFirst` for best performance.

    Parameters:

     - strs - a string of space-delimited literals, or a collection of
       string literals
     - caseless - (default= ``False``) - treat all literals as
       caseless
     - useRegex - (default= ``True``) - as an optimization, will
       generate a Regex object; otherwise, will generate
       a :class:`MatchFirst` object (if ``caseless=True`` or ``asKeyword=True``, or if
       creating a :class:`Regex` raises an exception)
     - asKeyword - (default=``False``) - enforce Keyword-style matching on the
       generated expressions

    Example::

        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))

    prints::

        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    z_More than one string argument passed to oneOf, pass choices as a list or space-delimited stringr�rzcSs|��|��kSr�)r��ro�brrrr�{rzoneOf.<locals>.<lambda>cSs|���|���Sr�)r�rr�rrrr�|rcSs||kSr�rr�rrrr�rcSs
|�|�Sr�rr�rrrr��rz6Invalid argument to oneOf, expected string or iterablerr�Nr�rcss|]}t|�VqdSr�)r�r�symrrrr��rzoneOf.<locals>.<genexpr>rs�|css|]}t�|�VqdSr�)r�rr�rrrr��rz7Exception creating Regex for oneOf, building MatchFirstc3s|]}�|�VqdSr�rr���parseElementClassrrr��r)r�r/r~rr&r'r1r4r�r
r*r�r7r�r�rcr�rGrr�r6)�strsr��useRegexr�isequal�masks�symbolsr�currCrmrr�rrtVsX 
�



�


**�rtcCsttt||���S)a�Helper to easily and clearly define a dictionary by specifying
    the respective patterns for the key and value.  Takes care of
    defining the :class:`Dict`, :class:`ZeroOrMore`, and
    :class:`Group` tokens in the proper order.  The key pattern
    can include delimiting markers or punctuation, as long as they are
    suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the :class:`Dict` results
    can include named token fields.

    Example::

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())

        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())

    prints::

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    )r*r9r0)rarDrrrrb�s%rbcCs^t��dd��}|��}d|_|d�||d�}|r@dd�}ndd�}|�|�|j|_|S)	a�Helper to return the original, untokenized text for a given
    expression.  Useful to restore the parsed fields of an HTML start
    tag into the raw tag text itself, or to revert separate tokens with
    intervening whitespace back to the original matching input text. By
    default, returns astring containing the original parsed text.

    If the optional ``asString`` argument is passed as
    ``False``, then the return value is
    a :class:`ParseResults` containing any results names that
    were originally matched, and a single token containing the original
    matched text from the input string.  So if the expression passed to
    :class:`originalTextFor` contains expressions with defined
    results names, you must set ``asString`` to ``False`` if you
    want to preserve those results name values.

    Example::

        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b", "i"):
            opener, closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])

    prints::

        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    cSs|Sr�r)r�r�r�rrrr��rz!originalTextFor.<locals>.<lambda>F�_original_start�
_original_endcSs||j|j�Sr�)r�r�r�rrrr��rcSs&||�d�|�d��g|dd�<dS)Nr�r�)r`r�rrr�extractText�sz$originalTextFor.<locals>.extractText)r,r�rlr�r�)r��asString�	locMarker�endlocMarker�	matchExprr�rrrr��s

r�cCst|��dd��S)zkHelper to undo pyparsing's default grouping of And expressions,
    even if all but one are non-empty.
    cSs|dSrrr�rrrr�rzungroup.<locals>.<lambda>)rMr)r�rrrr��sr�cCs4t��dd��}t|d�|d�|����d��S)a�Helper to decorate a returned token with its starting and ending
    locations in the input string.

    This helper adds the following results names:

     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains ``<TAB>`` characters, you
    may want to call :class:`ParserElement.parseWithTabs`

    Example::

        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)

    prints::

        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    cSs|Sr�rr�rrrr�rzlocatedExpr.<locals>.<lambda>�
locn_startrD�locn_end)r,r�r0rlr�)r��locatorrrrr�sr�rdrjrir�r�z\[]-*.$+^?()~ �r
cCs|ddSr;rr�rrrr�(rr�z\\0?[xX][0-9a-fA-F]+cCstt|d�d�d��S)Nrz\0xr�)�unichrr�r�r�rrrr�)rz	\\0[0-7]+cCstt|ddd�d��S)Nrr��)r�r�r�rrrr�*rz\]rVrzr��negate�bodyr}csDdd��z"d��fdd�t�|�jD��WSty>YdS0dS)aHelper to easily define string ranges for use in Word
    construction. Borrows syntax from regexp '[]' string range
    definitions::

        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"

    The input string must be enclosed in []'s, and the returned string
    is the expanded character set joined into a single string. The
    values enclosed in the []'s may be:

     - a single character
     - an escaped character with a leading backslash (such as ``\-``
       or ``\]``)
     - an escaped hex character with a leading ``'\x'``
       (``\x21``, which is a ``'!'`` character) (``\0x##``
       is also supported for backwards compatibility)
     - an escaped octal character with a leading ``'\0'``
       (``\041``, which is a ``'!'`` character)
     - a range of any of the above, separated by a dash (``'a-z'``,
       etc.)
     - any combination of the above (``'aeiouy'``,
       ``'a-zA-Z0-9_$'``, etc.)
    cSs<t|t�s|Sd�dd�tt|d�t|d�d�D��S)Nr�css|]}t|�VqdSr��r�r�rrrr�Irz+srange.<locals>.<lambda>.<locals>.<genexpr>rr�)r�rBr�r<�ord)�prrrr�Irzsrange.<locals>.<lambda>r�c3s|]}�|�VqdSr�r)r�part��	_expandedrrr�Krzsrange.<locals>.<genexpr>N)r��_reBracketExprrfr�r�rrr�rr�/s
"r�cs�fdd�}|S)zoHelper method for defining parse actions that require matching at
    a specific column in the input text.
    cs"t||��krt||d���dS)Nzmatched token not at column %drH)r��locnr�r�rr�	verifyColSsz!matchOnlyAtCol.<locals>.verifyColr)r�r�rr�rrnOsrncs�fdd�S)a�Helper method for common parse actions that simply return
    a literal value.  Especially useful when used with
    :class:`transformString<ParserElement.transformString>` ().

    Example::

        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num

        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    cs�gSr�rr���replStrrrr�erzreplaceWith.<locals>.<lambda>rr�rr�rr}Xs
r}cCs|ddd�S)aHelper parse action for removing quotation marks from parsed
    quoted strings.

    Example::

        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    rr�r�rr�rrrr{gs
r{csL��fdd�}zt�dt�d�j�}Wnty@t��}Yn0||_|S)aLHelper to define a parse action by mapping a function to all
    elements of a ParseResults list. If any additional args are passed,
    they are forwarded to the given function as additional arguments
    after the token, as in
    ``hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))``,
    which will convert the parsed data to an integer using base 16.

    Example (compare the last to example in :class:`ParserElement.transformString`::

        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')

        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')

    prints::

        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    cs��fdd�|D�S)Ncsg|]}�|g��R��qSrr)r�tokn�r�r�rrr�rz(tokenMap.<locals>.pa.<locals>.<listcomp>rr�rrrr��sztokenMap.<locals>.parr�)r�rr�r�)r�r�r�r�rrrr�vs$
�r�cCst|���Sr��r�r�r�rrrr��rcCst|���Sr��r��lowerr�rrrr��rr�r�cs~t|t�r|�t||d�}n|j�tttd�}|r�t���	t
�}||d�ttt
|td�|���tddgd�d��	d	d
��|}nlt���	t
�ttdd�B}||d�ttt
|�	t�ttd�|����tddgd�d��	d
d
��|}ttd�|ddd�}|�d��|��fdd
��|dd���dd��������d��}�|_�|_t|��|_||fS)zRInternal helper to construct opening and closing tag expressions, given a tag namer�z_-:�tag�=�/F�r\rdcSs|ddkS�Nrrrr�rrrr��rz_makeTags.<locals>.<lambda>r�r�cSs|ddkSr
rr�rrrr��rr�)r�z<%s>c	s*|�dd���dd������|���S)Nr�r��:r�)r:r�r��titler�rlr���resnamerrr��rrr�rr�z</%s>)r�r/r1r!rOrUrTr_rlr�r{r*rRr0rKr;rzrwrcr)�_Lrrr�r�rr�rrH�tag_body)�tagStr�xml�suppress_LT�suppress_GT�tagAttrName�tagAttrValue�openTag�closeTagrr
r�	_makeTags�sH
���������,rcCs
t|d�S)aKHelper to construct opening and closing tag expressions for HTML,
    given a tag name. Matches tags in either upper or lower case,
    attributes with namespaces and with quoted or unquoted values.

    Example::

        text = '<td>More info at the <a href="https://github.com/pyparsing/pyparsing/wiki">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and
        # closing tags as a 2-tuple
        a, a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end

        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are
            # also accessible as named results
            print(link.link_text, '->', link.href)

    prints::

        pyparsing -> https://github.com/pyparsing/pyparsing/wiki
    F�r�rrrrrl�srlcCs
t|d�S)z�Helper to construct opening and closing tag expressions for XML,
    given a tag name. Matches tags only in the given upper/lower case.

    Example: similar to :class:`makeHTMLTags`
    Trrrrrrm�srmcs8|r|dd��n|���dd��D���fdd�}|S)a7Helper to create a validating parse action to be used with start
    tags created with :class:`makeXMLTags` or
    :class:`makeHTMLTags`. Use ``withAttribute`` to qualify
    a starting tag with a required attribute value, to avoid false
    matches on common tags such as ``<TD>`` or ``<DIV>``.

    Call ``withAttribute`` with a series of attribute names and
    values. Specify the list of filter attributes names and values as:

     - keyword arguments, as in ``(align="right")``, or
     - as an explicit dict with ``**`` operator, when an attribute
       name is also a Python reserved word, as in ``**{"class":"Customer", "align":"right"}``
     - a list of name-value tuples, as in ``(("ns1:class", "Customer"), ("ns2:align", "right"))``

    For attribute names with a namespace prefix, you must use the second
    form.  Attribute names are matched insensitive to upper/lower case.

    If just testing for ``class`` (with or without a namespace), use
    :class:`withClass`.

    To verify that the attribute exists, but without specifying a value,
    pass ``withAttribute.ANY_VALUE`` as the value.

    Example::

        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>

        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)

        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)

    prints::

        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    NcSsg|]\}}||f�qSrrr�rrrr/rz!withAttribute.<locals>.<listcomp>csZ�D]P\}}||vr$t||d|��|tjkr|||krt||d||||f��qdS)Nzno matching attribute z+attribute '%s' has value '%s', must be '%s')r?r��	ANY_VALUE)r�r�r0�attrName�	attrValue��attrsrrr�0s�zwithAttribute.<locals>.pa)r?)r��attrDictr�rrrr��s8r�cCs"|rd|nd}tfi||i��S)a�Simplified version of :class:`withAttribute` when
    matching on a div class - made difficult because ``class`` is
    a reserved word in Python.

    Example::

        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>

        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))

        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)

        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)

    prints::

        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    z%s:class�class)r�)�	classname�	namespace�	classattrrrrr�:s#r��(r(cCs�Gdd�dt�}t�}||||B}t|�D�]r\}}|ddd�\}	}
}}|
dkr`d|	nd|	}
|
dkr�|	dus�t|	�d	kr�td
��|	\}}t��|
�}|tjk�rt|
dkr�|||	�t|t	|	��}n�|
d	k�r*|	du�r|||	|�t|t	|	|��}n|||�t|t	|��}nH|
dk�rj||||||�t|t	||||��}ntd��n�|tj
k�rX|
dk�r�t|	t��s�t|	�}	||	j
|�t|	|�}n�|
d	k�r|	du�r�|||	|�t|t	|	|��}n|||�t|t	|��}nD|
dk�rN||||||�t|||||�}ntd��ntd
��|�r�t|ttf��r�|j|�n
|�|�||�|
�|BK}|}q.||K}|S)al
Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary
    or binary, left- or right-associative.  Parse actions can also be
    attached to operator expressions. The generated parser will also
    recognize the use of parentheses to override operator precedences
    (see example below).

    Note: if you define a deep operator list, you may see performance
    issues when using infixNotation. See
    :class:`ParserElement.enablePackrat` for a mechanism to potentially
    improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the
       nested
     - opList - list of tuples, one for each operator precedence level
       in the expression grammar; each tuple is of the form ``(opExpr,
       numTerms, rightLeftAssoc, parseAction)``, where:

       - opExpr is the pyparsing expression for the operator; may also
         be a string, which will be converted to a Literal; if numTerms
         is 3, opExpr is a tuple of two expressions, for the two
         operators separating the 3 terms
       - numTerms is the number of terms for this operator (must be 1,
         2, or 3)
       - rightLeftAssoc is the indicator whether the operator is right
         or left associative, using the pyparsing-defined constants
         ``opAssoc.RIGHT`` and ``opAssoc.LEFT``.
       - parseAction is the parse action to be associated with
         expressions matching this operator expression (the parse action
         tuple member may be omitted); if the parse action is passed
         a tuple or list of functions, this is equivalent to calling
         ``setParseAction(*fn)``
         (:class:`ParserElement.setParseAction`)
     - lpar - expression for matching left-parentheses
       (default= ``Suppress('(')``)
     - rpar - expression for matching right-parentheses
       (default= ``Suppress(')')``)

    Example::

        # simple example of four-function arithmetic with ints and
        # variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier

        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])

        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)

    prints::

        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    c@seZdZddd�ZdS)zinfixNotation.<locals>._FBTcSs|j�||�|gfSr�)r�r6r%rrrr��sz$infixNotation.<locals>._FB.parseImplN)Tr�rrrr�_FB�sr'r�Nr
r�z%s termz	%s%s termr�z@if numterms=3, opExpr must be a tuple or list of two expressionsr�z6operator must be unary (1), binary (2), or ternary (3)z2operator must indicate right or left associativity)r-r.r�r�r�rru�LEFTr0r9�RIGHTr�r;r�r�r*r�)�baseExpr�opList�lpar�rparr'r��lastExprr�operDef�opExpr�arity�rightLeftAssocr��termName�opExpr1�opExpr2�thisExprr�rrrr�ds`H�

&
�



&
�

r�z4"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�"z string enclosed in double quotesz4'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*�'z string enclosed in single quotesz*quotedString using single or double quotes�uzunicode string literalcCs�||krtd��|du�r*t|t��r"t|t��r"t|�dkr�t|�dkr�|dur�tt|t||tjdd����	dd��}n$t
��t||tj��	dd��}nx|dur�tt|t|�t|�ttjdd����	dd��}n4ttt|�t|�ttjdd����	d	d��}ntd
��t
�}|du�rd|tt|�t||B|B�t|��K}n$|tt|�t||B�t|��K}|�d||f�|S)a�	Helper method for defining nested lists enclosed in opening and
    closing delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list
       (default= ``"("``); can also be a pyparsing expression
     - closer - closing character for a nested list
       (default= ``")"``); can also be a pyparsing expression
     - content - expression for items within the nested lists
       (default= ``None``)
     - ignoreExpr - expression for ignoring opening and closing
       delimiters (default= :class:`quotedString`)

    If an expression is not provided for the content argument, the
    nested expression will capture all whitespace-delimited content
    between delimiters as a list of separate values.

    Use the ``ignoreExpr`` argument to define expressions that may
    contain opening or closing characters that should not be treated as
    opening or closing characters for nesting, such as quotedString or
    a comment expression.  Specify multiple expressions using an
    :class:`Or` or :class:`MatchFirst`. The default is
    :class:`quotedString`, but if no expressions are to be ignored, then
    pass ``None`` for this argument.

    Example::

        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR, RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type")
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR
                      + code_body("body"))
        c_function.ignore(cStyleComment)

        source_code = '''
            int is_odd(int x) {
                return (x%2);
            }

            int dec_to_hex(char hchar) {
                if (hchar >= '0' && hchar <= '9') {
                    return (ord(hchar)-ord('0'));
                } else {
                    return (10+ord(hchar)-ord('A'));
                }
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)


    prints::

        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    z.opening and closing strings cannot be the sameNr�r�cSs|d��Sr�r�r�rrrr�;rznestedExpr.<locals>.<lambda>cSs|d��Srr:r�rrrr�@rcSs|d��Srr:r�rrrr�GrcSs|d��Srr:r�rrrr�LrzOopening and closing arguments must be strings if no content expression is givenznested %s%s expression)r�r�r/r�r)r9r(rDr�r�rdrlr4r.r0rKrRr)�opener�closer�contentr�r�rrrrq�s`A
�����
����������
*$rqcs&�dd����fdd���fdd�}�fdd�}�fdd	�}tt��d
���t�d�}t�t��|��d�}t��|��d
�}t��|��d�}	|r�tt	|�|t|t|�t	|�t�d�|	�}
n.tt	|�t|t|�t	|�t�d�|	�}
|
�
�fdd��|�tt��|
�d�S)a�Helper method for defining space-delimited indentation blocks,
    such as those used to define block statements in Python source code.

    Parameters:

     - blockStatementExpr - expression defining syntax of statement that
       is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
       (multiple statementWithIndentedBlock expressions within a single
       grammar should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond
       the current level; set to False for block of left-most
       statements (default= ``True``)

    A valid block must contain at least one ``blockStatement``.

    Example::

        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group("(" + Optional(delimitedList(identifier)) + ")") + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group(funcDecl + func_body)

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << (funcDef | assignment | identifier)

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()

    prints::

        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]]
    Ncs��dd�<dSr�rr)�backup_stack�indentStackrr�reset_stack�sz"indentedBlock.<locals>.reset_stackcsN|t|�krdSt||�}|�dkrJ|�dkr>t||d��t||d��dS)Nr�zillegal nestingznot a peer entry)r�rZr?�r�r�r��curCol�r?rr�checkPeerIndent�s
z&indentedBlock.<locals>.checkPeerIndentcs2t||�}|�dkr"��|�nt||d��dS)Nr�znot a subentry)rZr�r?rArCrr�checkSubIndent�s
z%indentedBlock.<locals>.checkSubIndentcsJ|t|�krdSt||�}�r&|�vs2t||d��|�dkrF���dS)Nznot an unindentr�)r�rZr?r`rArCrr�
checkUnindent�s
z$indentedBlock.<locals>.checkUnindentz	 r��INDENTr��UNINDENTcs��Sr�r)ror�r�r)r@rrr��rzindentedBlock.<locals>.<lambda>zindented block)
r9r2r�r�rIr,r�rr0r;rr�r�)�blockStatementExprr?r�rDrErFr�rG�PEER�UNDENT�smExprr)r>r?r@rr�Ws2Q�����r�z#[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]z[\0xa1-\0xbf\0xd7\0xf7]z_:zany tagzgt lt amp nbsp quot aposz><& "'z&(?P<entity>r�z);zcommon HTML entitycCst�|j�S)zRHelper parser action to replace common HTML entities with their special characters)�_htmlEntityMapr��entityr�rrrr|�sr|z/\*(?:[^*]|\*(?!/))*z*/zC style commentz<!--[\s\S]*?-->zHTML commentz.*zrest of linez//(?:\\\n|[^\n])*z
// commentzC++ style commentz#.*zPython style commentr�� 	�	commaItemr	r[c@s�eZdZdZee�Zee�Ze	e
��d��e�Z
e	e��d��eed��Zed��d��e�Ze��e�de��e��d�Ze�d	d
��eeeed���e�B�d�Ze�e�ed
��d��e�Zed��d��e�ZeeBeB��Zed��d��e�Ze	eded��d�Zed��d�Z ed��d�Z!e!de!d�d�Z"ee!de!d�dee!de!d��d�Z#e#�$dd
��d e �d!�Z%e&e"e%Be#B�d"���d"�Z'ed#��d$�Z(e)d=d&d'��Z*e)d>d)d*��Z+ed+��d,�Z,ed-��d.�Z-ed/��d0�Z.e/��e0��BZ1e)d1d2��Z2e&e3e4d3�e5�e	e6d3d4�ee7d5�������d6�Z8e9ee:�;�e8Bd7d8���d9�Z<e)ed:d
���Z=e)ed;d
���Z>d<S)?r�aHere are some common low-level expressions that may be useful in
    jump-starting parser development:

     - numeric forms (:class:`integers<integer>`, :class:`reals<real>`,
       :class:`scientific notation<sci_real>`)
     - common :class:`programming identifiers<identifier>`
     - network addresses (:class:`MAC<mac_address>`,
       :class:`IPv4<ipv4_address>`, :class:`IPv6<ipv6_address>`)
     - ISO8601 :class:`dates<iso8601_date>` and
       :class:`datetime<iso8601_datetime>`
     - :class:`UUID<uuid>`
     - :class:`comma-separated list<comma_separated_list>`

    Parse actions:

     - :class:`convertToInteger`
     - :class:`convertToFloat`
     - :class:`convertToDate`
     - :class:`convertToDatetime`
     - :class:`stripHTMLTags`
     - :class:`upcaseTokens`
     - :class:`downcaseTokens`

    Example::

        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')

    prints::

        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    �integerzhex integerr�z[+-]?\d+zsigned integerr�fractioncCs|d|dS)Nrr�rr�rrrr��rzpyparsing_common.<lambda>rVz"fraction or mixed integer-fractionz[+-]?(?:\d+\.\d*|\.\d+)zreal numberz@[+-]?(?:\d+(?:[eE][+-]?\d+)|(?:\d+\.\d*|\.\d+)(?:[eE][+-]?\d+)?)z$real number with scientific notationz[+-]?\d+\.?\d*([eE][+-]?\d+)?�fnumberr��
identifierzK(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}zIPv4 addressz[0-9a-fA-F]{1,4}�hex_integerr�zfull IPv6 address)rr�z::zshort IPv6 addresscCstdd�|D��dkS)Ncss|]}tj�|�rdVqdSr�)r��
_ipv6_partrlr�rrrr��rz,pyparsing_common.<lambda>.<locals>.<genexpr>r�)r~r�rrrr��rz::ffff:zmixed IPv6 addresszIPv6 addressz:[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}zMAC address�%Y-%m-%dcs�fdd�}|S)a�
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%d"``)

        Example::

            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))

        prints::

            [datetime.date(1999, 12, 31)]
        c
sPzt�|d����WStyJ}zt||t|���WYd}~n
d}~00dSr)r�strptime�dater�r?r��r�r�r��ve��fmtrr�cvt_fn�sz.pyparsing_common.convertToDate.<locals>.cvt_fnr�r^r_rr]r�
convertToDate�szpyparsing_common.convertToDate�%Y-%m-%dT%H:%M:%S.%fcs�fdd�}|S)aHelper to create a parse action for converting parsed
        datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%dT%H:%M:%S.%f"``)

        Example::

            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))

        prints::

            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        c
sLzt�|d��WStyF}zt||t|���WYd}~n
d}~00dSr)rrYr�r?r�r[r]rrr_�sz2pyparsing_common.convertToDatetime.<locals>.cvt_fnrr`rr]r�convertToDatetime�sz"pyparsing_common.convertToDatetimez7(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?zISO8601 datez�(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?zISO8601 datetimez2[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}�UUIDcCstj�|d�S)aParse action to remove HTML tags from web page HTML source

        Example::

            # strip HTML links from normal text
            text = '<td>More info at the <a href="https://github.com/pyparsing/pyparsing/wiki">pyparsing</a> wiki page</td>'
            td, td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            print(table_text.parseString(text).body)

        Prints::

            More info at the pyparsing wiki page
        r)r��_html_stripperr�)r�r�r0rrr�
stripHTMLTagsszpyparsing_common.stripHTMLTagsr�r�rOrPr�r	zcomma separated listcCst|���Sr�rr�rrrr�#rcCst|���Sr�rr�rrrr�&rN)rX)rb)?rrrr�r�r��convertToInteger�float�convertToFloatrOrsrr�rQrerUrG�signed_integerrRrr;r��
mixed_integerr~�real�sci_realra�numberrSrUrTrT�ipv4_addressrW�_full_ipv6_address�_short_ipv6_addressr�_mixed_ipv6_addressr)�ipv6_address�mac_addressr
rarc�iso8601_date�iso8601_datetime�uuidrXrWrerfr9r4r2rwrN�
_commasepitemrarzrl�comma_separated_listr�rcrrrrr��sv""
���
�
�
��
���r�c@seZdZdd�Zdd�ZdS)�_lazyclasspropertycCs||_|j|_|j|_dSr�)r�r�rrrrrr�+sz_lazyclassproperty.__init__csl�durt|��t�d�r:t�fdd��jdd�D��r@i�_|jj}|�jvrb|����j|<�j|S)N�_internc3s |]}�jt|dg�uVqdS)r{N)r{r�)r�
superclassr�rrr�3s�z-_lazyclassproperty.__get__.<locals>.<genexpr>r�)r�rQr��__mro__r{r�r)r�r�r��attrnamerr�r�__get__0s�
z_lazyclassproperty.__get__N)rrrr�rrrrrrz*srzc@sPeZdZdZgZedd��Zedd��Zedd��Z	edd	��Z
ed
d��ZdS)
r�a�
    A set of Unicode characters, for language-specific strings for
    ``alphas``, ``nums``, ``alphanums``, and ``printables``.
    A unicode_set is defined by a list of ranges in the Unicode character
    set, in a class attribute ``_ranges``, such as::

        _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),]

    A unicode set can also be defined using multiple inheritance of other unicode sets::

        class CJK(Chinese, Japanese, Korean):
            pass
    cCsZg}|jD]8}|turqD|jD] }|�t|d|dd��q q
dd�tt|��D�S)Nrr�r�cSsg|]}t|��qSrr�r�rrrrTrz5unicode_set._get_chars_for_ranges.<locals>.<listcomp>)r}r��_rangesrgr<r�r�)r�r��cc�rrrrr�_get_chars_for_rangesLs

 z!unicode_set._get_chars_for_rangescCsd�ttj|����S)z+all non-whitespace characters in this ranger�)r�rr�rGr�r�rrrrwVszunicode_set.printablescCsd�ttj|����S)z'all alphabetic characters in this ranger�)r��filterr��isalphar�r�rrrrU[szunicode_set.alphascCsd�ttj|����S)z*all numeric digit characters in this ranger�)r�r�r��isdigitr�r�rrrrs`szunicode_set.numscCs|j|jS)z)all alphanumeric characters in this range)rUrsr�rrrrTeszunicode_set.alphanumsN)rrrr�r�r�r�rzrwrUrsrTrrrrr�<s

	


r�c@s�eZdZdZdejfgZGdd�de�ZGdd�de�Z	Gdd�de�Z
Gd	d
�d
e�ZGdd�de�ZGd
d�de�Z
Gdd�de�ZGdd�de�ZGdd�de
ee�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZdS)r�zF
    A namespace class for defining common language unicode_sets.
    � c@seZdZdZddgZdS)zpyparsing_unicode.Latin1z/Unicode set for Latin-1 Unicode Character Range)r��~)��N�rrrr�r�rrrr�Latin1qsr�c@seZdZdZdgZdS)zpyparsing_unicode.LatinAz/Unicode set for Latin-A Unicode Character Range)�iNr�rrrr�LatinAusr�c@seZdZdZdgZdS)zpyparsing_unicode.LatinBz/Unicode set for Latin-B Unicode Character Range)i�iONr�rrrr�LatinBysr�c@seZdZdZgd�ZdS)zpyparsing_unicode.Greekz.Unicode set for Greek Unicode Character Ranges))ipi�)ii)ii)i iE)iHiM)iPiW)iY)i[)i])i_i})i�i�)i�i�)i�i�)i�i�)i�i�)i�i�)i�i�Nr�rrrr�Greek}sr�c@seZdZdZdgZdS)zpyparsing_unicode.Cyrillicz0Unicode set for Cyrillic Unicode Character Range)ii�Nr�rrrr�Cyrillic�sr�c@seZdZdZddgZdS)zpyparsing_unicode.Chinesez/Unicode set for Chinese Unicode Character Range)�Ni���i0i?0Nr�rrrr�Chinese�sr�c@sDeZdZdZgZGdd�de�ZGdd�de�ZGdd�de�ZdS)	zpyparsing_unicode.Japanesez`Unicode set for Japanese Unicode Character Range, combining Kanji, Hiragana, and Katakana rangesc@seZdZdZddgZdS)z pyparsing_unicode.Japanese.Kanjiz-Unicode set for Kanji Unicode Character Range)r�i��r�Nr�rrrr�Kanji�sr�c@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Hiraganaz0Unicode set for Hiragana Unicode Character Range)i@0i�0Nr�rrrr�Hiragana�sr�c@seZdZdZdgZdS)z#pyparsing_unicode.Japanese.Katakanaz1Unicode set for Katakana  Unicode Character Range)i�0i�0Nr�rrrr�Katakana�sr�N)	rrrr�r�r�r�r�r�rrrr�Japanese�s
r�c@seZdZdZgd�ZdS)zpyparsing_unicode.Koreanz.Unicode set for Korean Unicode Character Range))i�i��)ii�)i01i�1)i`�i�)i��i��r�Nr�rrrr�Korean�sr�c@seZdZdZdS)zpyparsing_unicode.CJKzTUnicode set for combined Chinese, Japanese, and Korean (CJK) Unicode Character RangeNrrrrr�CJK�sr�c@seZdZdZddgZdS)zpyparsing_unicode.Thaiz,Unicode set for Thai Unicode Character Range)ii:)i?i[Nr�rrrr�Thai�sr�c@seZdZdZgd�ZdS)zpyparsing_unicode.Arabicz.Unicode set for Arabic Unicode Character Range))ii)ii�)iiNr�rrrr�Arabic�sr�c@seZdZdZdgZdS)zpyparsing_unicode.Hebrewz.Unicode set for Hebrew Unicode Character Range)i�i�Nr�rrrr�Hebrew�sr�c@seZdZdZddgZdS)zpyparsing_unicode.Devanagariz2Unicode set for Devanagari Unicode Character Range)i	i	)i�i��Nr�rrrr�
Devanagari�sr�N)rrrr�r��
maxunicoder�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr�ksr�uالعربيةu中文uкириллицаuΕλληνικάuעִברִיתu	日本語u漢字uカタカナuひらがなu	한국어u	ไทยuदेवनागरीc@s,eZdZdZGdd�d�ZGdd�d�ZdS)�pyparsing_testzB
    namespace class for classes useful in writing unit tests
    c@s8eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�ZdS)
z&pyparsing_test.reset_pyparsing_contextax
        Context manager to be used when writing unit tests that modify pyparsing config values:
         - packrat parsing
         - default whitespace characters.
         - default keyword characters
         - literal string auto-conversion class
         - __diag__ settings

        Example:
            with reset_pyparsing_context():
                # test that literals used to construct a grammar are automatically suppressed
                ParserElement.inlineLiteralsUsing(Suppress)

                term = Word(alphas) | Word(nums)
                group = Group('(' + term[...] + ')')

                # assert that the '()' characters are not included in the parsed tokens
                self.assertParseAndCheckLisst(group, "(abc 123 def)", ['abc', '123', 'def'])

            # after exiting context manager, literals are converted to Literal expressions again
        cCs
i|_dSr�)�
_save_contextr�rrrr��sz/pyparsing_test.reset_pyparsing_context.__init__cCsftj|jd<tj|jd<tj|jd<tj|jd<tj|jd<dd�tj	D�|jd<d	t
ji|jd
<|S)N�default_whitespace�default_keyword_chars�literal_string_class�packrat_enabled�
packrat_parsecSsi|]}|tt|��qSr)r�r)rr!rrr�
<dictcomp>�sz?pyparsing_test.reset_pyparsing_context.save.<locals>.<dictcomp>rrdr$)rDr�r�r1r�r�r^rr�
_all_namesr$rdr�rrr�save�s����
z+pyparsing_test.reset_pyparsing_context.savecCs�tj|jdkr t�|jd�|jdt_t�|jd�|jd��D]\}}tt	||�qJ|jdt_
|jdt_|jdt_
dS)Nr�r�r�rr�r�r$)rDr�r�r�r1r�r�r?�setattrrr^rr$rd)r�r!rDrrr�restore�s ����z.pyparsing_test.reset_pyparsing_context.restorecCs|��Sr�)r�r�rrr�	__enter__sz0pyparsing_test.reset_pyparsing_context.__enter__cGs|��Sr�)r�r�rrr�__exit__sz/pyparsing_test.reset_pyparsing_context.__exit__N)	rrrr�r�r�r�r�r�rrrr�reset_pyparsing_context�sr�c@sJeZdZdZddd�Zddd�Zddd	�Zdd
d�Zee	dfdd
��Z
dS)z&pyparsing_test.TestParseResultsAssertszk
        A mixin class to add parse results assertion methods to normal unittest.TestCase classes.
        NcCs<|dur|j||��|d�|dur8|j||��|d�dS)z�
            Unit test assertion to compare a ParseResults object with an optional expected_list,
            and compare any defined results names with an optional expected_dict.
            N�r�)�assertEqualr"r�)r�r��
expected_list�
expected_dictr�rrr�assertParseResultsEqualssz?pyparsing_test.TestParseResultsAsserts.assertParseResultsEqualsTcCs2|j|dd�}|rt|���|j|||d�dS)z�
            Convenience wrapper assert to test a parser element and input string, and assert that
            the resulting ParseResults.asList() is equal to the expected_list.
            Tr�)r�r�N�rfr�r�r�)r�r��test_stringr�r��verboser�rrr�assertParseAndCheckList!sz>pyparsing_test.TestParseResultsAsserts.assertParseAndCheckListcCs2|j|dd�}|rt|���|j|||d�dS)z�
            Convenience wrapper assert to test a parser element and input string, and assert that
            the resulting ParseResults.asDict() is equal to the expected_dict.
            Tr�)r�r�Nr�)r�r�r�r�r�r�r�rrr�assertParseAndCheckDict-sz>pyparsing_test.TestParseResultsAsserts.assertParseAndCheckDictc	Cs"|\}}|du�rdd�t||�D�}|D]�\}}}	tdd�|	D�d�}
tdd�|	D�d�}|dur�|j||
pp|d��t|t�r�|�Wd�n1s�0Yq*tdd�|	D�d�}td	d�|	D�d�}
||
fd
kr�|j|||
|
p�|d�q*td�|��q*|j||du�r|nd
d�dS)aP
            Unit test assertion to evaluate output of ParserElement.runTests(). If a list of
            list-dict tuples is given as the expected_parse_results argument, then these are zipped
            with the report tuples returned by runTests and evaluated using assertParseResultsEquals.
            Finally, asserts that the overall runTests() success value is True.

            :param run_tests_report: tuple(bool, [tuple(str, ParseResults or Exception)]) returned from runTests
            :param expected_parse_results (optional): [tuple(str, list, dict, Exception)]
            NcSs"g|]\}}|d|d|f�qSrqr)r�rpt�expectedrrrrHs�zOpyparsing_test.TestParseResultsAsserts.assertRunTestResults.<locals>.<listcomp>css|]}t|t�r|VqdSr�)r�r��r�exprrrr�QrzNpyparsing_test.TestParseResultsAsserts.assertRunTestResults.<locals>.<genexpr>css&|]}t|t�rt|t�r|VqdSr�)r�r�r�r�r�rrrr�Ts�)�expected_exceptionr�css|]}t|t�r|VqdSr�)r�r*r�rrrr�crcss|]}t|t�r|VqdSr�)r�r-r�rrrr�frr�)r�r�r�zno validation for {!r}zfailed runTestsr�)	r�r��assertRaisesr�r�r�r�r��
assertTrue)r��run_tests_report�expected_parse_resultsr��run_test_success�run_test_results�mergedr�r�r��fail_msgr�r�r�rrr�assertRunTestResults9sL
�����
$����z;pyparsing_test.TestParseResultsAsserts.assertRunTestResultsccs8|j||d��dVWd�n1s*0YdS)Nr�)r�)r�r�r�rrr�assertRaisesParseExceptionxszApyparsing_test.TestParseResultsAsserts.assertRaisesParseException)NNN)NT)NT)NN)rrrr�r�r�r�r�rr?r�rrrr�TestParseResultsAssertss�

�

�

�
?r�N)rrrr�r�r�rrrrr��sCr��__main__�select�fromr�r�)r��columnsrZtables�commandaK
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        z]
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        z 
        100
        FF
        z6
        12345678-1234-5678-1234-567812345678
        )NF)r�)r�F)N)FTF)T)r�)T(r�r!r"r#r��weakrefrr7rlr�r~r�rrSr�r�rBr�operatorr�	itertools�	functoolsr�
contextlibrr�ImportErrorr�_threadr	�	threading�collections.abcr
rrr
rMZordereddictrr$rdrrrrrr�r�r�r �enable_all_warnings�__all__r��version_infor�r��maxsizer�r�r/�chrr�r�r�r~r�r��reversedr*r�r�rrr	r�Zmaxint�xranger<�__builtin__r��fnamer�r�r�r�r,r��ascii_uppercase�ascii_lowercaserUrsrerTr�r��	printablerwr�r�r=r?rArCrFrrrB�registerrZrkrhr�r�r�rrr�rDr}rLr,r7r4r�rr�r1r'r&r�rOrrSrGrEr(rNrDr/r3r2rJrIrQrPr@r%r<r6r+r>r-r5r8r�r9rRr�r;rHr.rMr)r0r*rKr:r�rar]rrrprorrtrbr�r�r�rrdrjrir�r�r��_escapedPunc�_escapedHexChar�_escapedOctChar�_singleChar�
_charRangerr�r�rnr}r{r�r�rcrrlrmr�rr�rur(r)r�rvr_rrzr�rqr�rVrxrXrWr-r�rMrWr\r|rYrfr�r~r`r^rgryrarxr[r�rzr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rZselectTokenZ	fromToken�identZ
columnNameZcolumnNameListZ
columnSpecZ	tableNameZ
tableNameListZ	simpleSQLr�rnrSrUrwrdrrrr�<module>sJ�H
?]

H

D+!
'N
E 

KFym{VO#K,:#Dvj-D0$	W'*0 	0�
%E&
��h~

 ( 
����"./J��6,
	site-packages/pip/_vendor/packaging/__pycache__/utils.cpython-39.pyc000064400000007102150564745010021444 0ustar00a

��Reh�@s�ddlZddlmZmZmZmZmZddlmZm	Z	ddl
mZmZeedee
effZede�ZGdd	�d	e�ZGd
d�de�Ze�d�Ze�d
�Zeed�dd�Zeeefed�dd�Zeeeeeeefd�dd�Zeeeefd�dd�ZdS)�N)�	FrozenSet�NewType�Tuple�Union�cast�)�Tag�	parse_tag)�InvalidVersion�Version��NormalizedNamec@seZdZdZdS)�InvalidWheelFilenamezM
    An invalid wheel filename was found, users should refer to PEP 427.
    N��__name__�
__module__�__qualname__�__doc__rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/utils.pyrsrc@seZdZdZdS)�InvalidSdistFilenamez^
    An invalid sdist filename was found, users should refer to the packaging user guide.
    Nrrrrrrsrz[-_.]+z	(\d+)(.*))�name�returncCst�d|���}tt|�S)N�-)�_canonicalize_regex�sub�lowerrr
)r�valuerrr�canonicalize_name sr)�versionrc
Cs�t|t�r0zt|�}Wq4ty,|YS0n|}g}|jdkrT|�|j�d��|�t�ddd�dd�|j	D����|j
dur�|�d�d	d�|j
D���|jdur�|�d
|j���|jdur�|�d|j���|j
dur�|�d|j
���d�|�S)
z�
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    r�!z(\.0)+$��.css|]}t|�VqdS�N��str��.0�xrrr�	<genexpr><�z'canonicalize_version.<locals>.<genexpr>Ncss|]}t|�VqdSr"r#r%rrrr(@r)z.postz.dev�+)�
isinstancer$rr
�epoch�append�rer�join�release�pre�post�dev�local)r�parsed�partsrrr�canonicalize_version&s&

&



r7)�filenamerc
Cs|�d�std|����|dd�}|�d�}|dvrDtd|����|�d|d�}|d}d	|vsxt�d
|tj�dur�td|����t|�}t|d�}|d
kr�|d}t	�|�}|dur�td|�d|�d���t
tt|�
d��|�
d�f�}nd}t|d�}	||||	fS)Nz.whlz3Invalid wheel filename (extension must be '.whl'): ���r)��z0Invalid wheel filename (wrong number of parts): �r�__z^[\w\d._]*$zInvalid project name: rr;zInvalid build number: z in '�'r���)�endswithr�count�splitr.�match�UNICODErr�_build_tag_regexr�BuildTag�int�groupr	)
r8�dashesr6�	name_partrr�
build_part�build_match�build�tagsrrr�parse_wheel_filenameQs6
�
�
� rOcCs�|�d�r|dtd��}n,|�d�r<|dtd��}ntd|����|�d�\}}}|sltd|����t|�}t|�}||fS)Nz.tar.gzz.zipz@Invalid sdist filename (extension must be '.tar.gz' or '.zip'): rzInvalid sdist filename: )r@�lenr�
rpartitionrr)r8�	file_stemrJ�sep�version_partrrrrr�parse_sdist_filenameus

��rU)r.�typingrrrrrrNrr	rr
rrGr$rFr
�
ValueErrorrr�compilerrErr7rOrUrrrr�<module>s


,�$site-packages/pip/_vendor/packaging/__pycache__/tags.cpython-39.pyc000064400000030050150564745010021240 0ustar00a

��Reb=�@s�UddlZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
mZmZm
Z
mZmZmZddlmZmZe�e�Ze
eZeeefZdddd	d
d�Zeeefed<ejd
kZGdd�d�Zeeed�dd�Z dJee!eeedfd�dd�Z"eed�dd�Z#ee!d�dd�Z$dKee!eed�dd�Z%dLdd �eeee	eee	ee!e
ed!�d"d#�Z&e
ed$�d%d&�Z'dMdd �eeee	eee	ee!e
ed'�d(d)�Z(ee
ed*�d+d,�Z)dNeeeeee	ee
ed-�d.d/�Z*efee!ed0�d1d2�Z+eeeed3�d4d5�Z,dOeeeee
ed6�d7d8�Z-efe!e
ed9�d:d;�Z.e
ed$�d<d=�Z/e
ed$�d>d?�Z0ed$�d@dA�Z1dd �e!edB�dCdD�Z2eedE�dFdG�Z3dd �e!e
edB�dHdI�Z4dS)P�N)�EXTENSION_SUFFIXES)
�Dict�	FrozenSet�Iterable�Iterator�List�Optional�Sequence�Tuple�Union�cast�)�
_manylinux�
_musllinux�py�cp�pp�ip�jy)�python�cpython�pypy�
ironpython�jython�INTERPRETER_SHORT_NAMESlc@s�eZdZdZgd�Zeeedd�dd�Zeed�dd	��Zeed�d
d��Z	eed�dd
��Z
eed�dd�Z
ed�dd�Zed�dd�Zed�dd�ZdS)�Tagz�
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    )�_interpreter�_abi�	_platform�_hashN)�interpreter�abi�platform�returncCs8|��|_|��|_|��|_t|j|j|jf�|_dS�N)�lowerrrr�hashr)�selfr r!r"�r(��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/tags.py�__init__4s


zTag.__init__�r#cCs|jSr$)r�r'r(r(r)r ?szTag.interpretercCs|jSr$)rr,r(r(r)r!CszTag.abicCs|jSr$)rr,r(r(r)r"GszTag.platform)�otherr#cCs>t|t�stS|j|jko<|j|jko<|j|jko<|j|jkSr$)�
isinstancer�NotImplementedrrrr)r'r-r(r(r)�__eq__Ks

�
�
�z
Tag.__eq__cCs|jSr$)rr,r(r(r)�__hash__VszTag.__hash__cCs|j�d|j�d|j��S)N�-)rrrr,r(r(r)�__str__YszTag.__str__cCsdj|t|�d�S)Nz<{self} @ {self_id}>)r'�self_id)�format�idr,r(r(r)�__repr__\szTag.__repr__)�__name__�
__module__�__qualname__�__doc__�	__slots__�strr*�propertyr r!r"�object�boolr0�intr1r3r7r(r(r(r)r*sr)�tagr#c	Cs`t�}|�d�\}}}|�d�D]6}|�d�D]&}|�d�D]}|�t|||��q<q.q t|�S)z�
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    r2�.)�set�split�addr�	frozenset)rB�tags�interpreters�abis�	platformsr r!�	platform_r(r(r)�	parse_tag`srMF)�name�warnr#cCs&t�|�}|dur"|r"t�d|�|S)Nz>Config variable '%s' is unset, Python ABI tag may be incorrect)�	sysconfig�get_config_var�logger�debug)rNrO�valuer(r(r)�_get_config_varps
�rU)�stringr#cCs|�dd��dd�S)NrC�_r2)�replace)rVr(r(r)�_normalize_stringysrY)�python_versionr#cCst|�dkot|�dkS)zj
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    r
)��)�len�tuple)rZr(r(r)�
_abi3_applies}sr_)�
py_versionrOr#c	Cs�t|�}g}t|dd��}d}}}td|�}ttd�}dtv}	|sX|dur\|sX|	r\d}|dkr�td|�}
|
sz|
dur~d	}|d
kr�td|�}|dks�|dur�tjd
kr�d}n|r�|�d|���|�ddj	||||d��|S)Nr\��Py_DEBUG�gettotalrefcountz_d.pyd�d)r[��
WITH_PYMALLOC�m)r[r[�Py_UNICODE_SIZE�i���urrz"cp{version}{debug}{pymalloc}{ucs4})�versionrS�pymalloc�ucs4)
r^�_version_nodotrU�hasattr�sysr�
maxunicode�append�insertr5)r`rOrJrkrSrlrm�
with_debug�has_refcount�has_ext�
with_pymalloc�unicode_sizer(r(r)�
_cpython_abis�s<



����ry�rO)rZrJrKrOr#c	#sD|stjdd�}d�t|dd����|durLt|�dkrHt||�}ng}t|�}dD]&}z|�|�WqXty|YqX0qXt|p�t	��}|D]}|D]}t
�||�Vq�q�t|�rЇfdd�|D�EdH�fdd�|D�EdHt|��r@t|dddd	�D]8}|D],}d
jt|d|f�d��t
�d
|�V�q�qdS)a
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    Nr\zcp{}r
)�abi3�nonec3s|]}t�d|�VqdS)r{N�r��.0rL�r r(r)�	<genexpr>��zcpython_tags.<locals>.<genexpr>c3s|]}t�d|�VqdS)r|Nr}r~r�r(r)r��r����zcp{version}r�rkr{)
rp�version_infor5rnr]ry�list�remove�
ValueError�_platform_tagsrr_�range)rZrJrKrO�explicit_abir!rL�
minor_versionr(r�r)�cpython_tags�s6
�r�r+ccst�d�}|rt|�VdS)N�SOABI)rPrQrY)r!r(r(r)�_generic_abi�s
r�)r rJrKrOr#ccs~|s"t�}t|d�}d�||g�}|dur0t�}t|p:t��}t|�}d|vrX|�d�|D]}|D]}t|||�Vqdq\dS)z�
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    rzraNr|)�interpreter_name�interpreter_version�joinr�r�r�rrr)r rJrKrO�interp_name�interp_versionr!rLr(r(r)�generic_tags�s

r�)r`r#ccs|t|�dkr&djt|dd��d�Vdj|dd�Vt|�dkrxt|ddd	d	�D]}djt|d|f�d�VqXdS)
z�
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    r
zpy{version}Nr\r�z	py{major}r)�majorr�)r]r5rnr�)r`�minorr(r(r)�_py_interpreter_rangesr�)rZr rKr#ccsx|stjdd�}t|pt��}t|�D]}|D]}t|d|�Vq0q(|rXt|dd�Vt|�D]}t|dd�Vq`dS)z�
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    Nr\r|�any)rpr�r�r�r�r)rZr rKrkrLr(r(r)�compatible_tagss
r�)�arch�is_32bitr#cCs|s|S|�d�rdSdS)N�ppc�i386)�
startswith)r�r�r(r(r)�	_mac_arch/s

r�)rk�cpu_archr#cCs�|g}|dkr*|dkrgS|�gd��nn|dkrN|dkr>gS|�gd��nJ|dkrv|dksf|dkrjgS|�d�n"|d	kr�|d
kr�gS|�ddg�|d
vr�|�d�|dvr�|�d�|S)N�x86_64)�
ri)�intel�fat64�fat32r�)r�r��fat�ppc64)r��r�r�)r��r�r�>r��arm64�
universal2>r�r�r�r�r��	universal)�extendrr)rkr��formatsr(r(r)�_mac_binary_formats9s,

r�)rkr�r#c
cs�t��\}}}|dur:tdttt|�d�dd����}n|}|durPt|�}n|}d|kr�|dkr�t|ddd�D]2}d	|f}t	||�}|D]}d
j
d	||d�Vq�qt|dkr�t|dd	d�D]2}	|	df}t	||�}|D]}d
j
|	d|d�Vq�q�|dk�r�|d
k�rVtddd�D]>}d	|f}t	||�}|D] }d
j
|d|d|d�V�q.�qn:tddd�D],}d	|f}d}d
j
|d|d|d�V�qbdS)aD
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    N�
MacVersionrCr\)r�r)�rr
r�r�z&macosx_{major}_{minor}_{binary_format})r�r��
binary_formatrr��r[r�)r"�mac_verrr^�maprArEr�r�r�r5)
rkr��version_strrWr�r��compat_version�binary_formatsr��
major_versionr(r(r)�
mac_platformsYsT$

�
�


��r�)r�r#ccsftt���}|r*|dkrd}n|dkr*d}|�dd�\}}t�||�EdHt�|�EdH|VdS)N�linux_x86_64�
linux_i686�
linux_aarch64�linux_armv7lrWr
)rYrP�get_platformrEr�
platform_tagsr)r��linuxrWr�r(r(r)�_linux_platforms�sr�ccstt���VdSr$)rYrPr�r(r(r(r)�_generic_platforms�sr�cCs.t��dkrt�St��dkr$t�St�SdS)z;
    Provides the platform tags for this installation.
    �Darwin�LinuxN)r"�systemr�r�r�r(r(r(r)r��s
r�cCstjj}t�|�p|S)z6
    Returns the name of the running interpreter.
    )rp�implementationrNr�get)rNr(r(r)r��sr�)rOr#cCs0td|d�}|rt|�}nttjdd��}|S)z9
    Returns the version of the running interpreter.
    �py_version_nodotrzNr\)rUr=rnrpr�)rOrkr(r(r)r��s

r�)rkr#cCsd�tt|��S)Nra)r�r�r=r�r(r(r)rn�srnccs<t�}|dkr t|d�EdHnt�EdHt�EdHdS)z�
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    rrzN)r�r�r�r�)rOr�r(r(r)�sys_tags�s
r�)F)F)NNN)NNN)NNN)NN)5�loggingr"rprP�importlib.machineryr�typingrrrrrrr	r
rrrarr�	getLoggerr8rRrA�
PythonVersionr�rr=�__annotations__�maxsize�_32_BIT_INTERPRETERrrMr@rUrYr_ryr�r�r�r�r�r�r�r�r�r�r�r�r�rnr�r(r(r(r)�<module>s�
0

�	
6		%��

�9��

��
�
!��H
site-packages/pip/_vendor/packaging/__pycache__/requirements.cpython-39.pyc000064400000007655150564745010023044 0ustar00a

��ReD�@sfddlZddlZddlZddlmZmZmZddl	m
Z
mZmZm
Z
mZmZmZmZmZmZddlmZmZddlmZmZmZGdd�de�Zeejej�Zed	�� �Z!ed
�� �Z"ed�� �Z#ed�� �Z$ed
�� �Z%ed�� �Z&ed�� �Z'ed�Z(eee(�eBZ)e
eee)��Z*e*d�Z+e*Z,ed�d�Z-e'e-Z.e,ee%e,�Z/e!ee/�e"d�Z0eej1ej2ej3B�Z4eej1ej2ej3B�Z5e4e5AZ6e
e6ee%e6�d
dd�d�Z7ee#e7e$e7B�Z8e8�9dd��ee8�d�Z:e:�9dd��ee��d�Ze�9dd��e&Z;e;eZ<e:ee<�Z=e.ee<�Z>e+ee0�e>e=BZ?ee?eZ@e@�Ad�Gdd �d �ZBdS)!�N)�List�Optional�Set)
�Combine�Literalr�ParseException�Regex�Word�
ZeroOrMore�originalTextFor�	stringEnd�stringStart�)�MARKER_EXPR�Marker)�LegacySpecifier�	Specifier�SpecifierSetc@seZdZdZdS)�InvalidRequirementzJ
    An invalid requirement was found, users should refer to PEP 508.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/requirements.pyrsr�[�]�(�)�,�;�@z-_.�namez[^ ]+�url�extrasF)�
joinString�adjacent�	_raw_speccCs
|jpdS)N�)r'��s�l�trrr�<lambda>@�r-�	specifiercCs|dS)Nrrr)rrrr-Cr.�markercCst||j|j��S)N)r�_original_start�
_original_endr)rrrr-Gr.zx[]c@s<eZdZdZedd�dd�Zed�dd�Zed�d	d
�ZdS)�Requirementz�Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    N)�requirement_string�returnc
Cszt�|�}WnJtyX}z2td||j|jd��d|j����WYd}~n
d}~00|j|_|jr�tj	�
|j�}|jdkr�tj	�|�|jkr�td��n(|jr�|j
r�|js�|j
s�td|j����|j|_nd|_t|jr�|j��ng�|_t|j�|_|j�r|jnd|_dS)NzParse error at "�z": �filezInvalid URL givenz
Invalid URL: )�REQUIREMENT�parseStringrr�loc�msgr"r#�urllib�parse�urlparse�scheme�
urlunparse�netloc�setr$�asListrr/r0)�selfr4�req�e�
parsed_urlrrr�__init__ds. �

��
zRequirement.__init__)r5cCs�|jg}|jr0d�t|j��}|�d|�d��|jrF|�t|j��|jrn|�d|j���|jrn|�d�|jr�|�d|j���d�|�S)Nrrrz@ � z; r()	r"r$�join�sorted�appendr/�strr#r0)rD�parts�formatted_extrasrrr�__str__}s
zRequirement.__str__cCsd|�d�S)Nz<Requirement('z')>r)rDrrr�__repr__�szRequirement.__repr__)rrrrrMrHrPrQrrrrr3Wsr3)C�re�string�urllib.parser<�typingrr�	TOptionalr�pip._vendor.pyparsingrr�Lrrr	r
rrr
�markersrr�
specifiersrrr�
ValueErrorr�
ascii_letters�digits�ALPHANUM�suppress�LBRACKET�RBRACKET�LPAREN�RPAREN�COMMA�	SEMICOLON�AT�PUNCTUATION�IDENTIFIER_END�
IDENTIFIER�NAME�EXTRA�URI�URL�EXTRAS_LIST�EXTRAS�
_regex_str�VERBOSE�
IGNORECASE�VERSION_PEP440�VERSION_LEGACY�VERSION_ONE�VERSION_MANY�
_VERSION_SPEC�setParseAction�VERSION_SPEC�MARKER_SEPARATOR�MARKER�VERSION_AND_MARKER�URL_AND_MARKER�NAMED_REQUIREMENTr8r9r3rrrr�<module>s`0
���
site-packages/pip/_vendor/packaging/__pycache__/_structures.cpython-39.pyc000064400000006061150564745010022671 0ustar00a

��Re]�@s,Gdd�d�Ze�ZGdd�d�Ze�ZdS)c@s�eZdZed�dd�Zed�dd�Zeed�dd�Z	eed�d	d
�Z
eed�dd�Zeed�d
d�Zeed�dd�Z
eed�dd�Zedd�dd�ZdS)�InfinityType��returncCsdS)N�Infinity���selfrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/_structures.py�__repr__szInfinityType.__repr__cCstt|��S�N��hash�reprrrrr�__hash__
szInfinityType.__hash__��otherrcCsdS�NFr�rrrrr�__lt__
szInfinityType.__lt__cCsdSrrrrrr�__le__szInfinityType.__le__cCst||j�Sr
��
isinstance�	__class__rrrr�__eq__szInfinityType.__eq__cCst||j�Sr
rrrrr�__ne__szInfinityType.__ne__cCsdS�NTrrrrr�__gt__szInfinityType.__gt__cCsdSrrrrrr�__ge__szInfinityType.__ge__�NegativeInfinityType�rrcCstSr
)�NegativeInfinityrrrr�__neg__szInfinityType.__neg__N)�__name__�
__module__�__qualname__�strr	�intr�object�boolrrrrrrr rrrrrsrc@s�eZdZed�dd�Zed�dd�Zeed�dd�Z	eed�d	d
�Z
eed�dd�Zeed�d
d�Zeed�dd�Z
eed�dd�Zeed�dd�ZdS)rrcCsdS)Nz	-Infinityrrrrrr	'szNegativeInfinityType.__repr__cCstt|��Sr
rrrrrr*szNegativeInfinityType.__hash__rcCsdSrrrrrrr-szNegativeInfinityType.__lt__cCsdSrrrrrrr0szNegativeInfinityType.__le__cCst||j�Sr
rrrrrr3szNegativeInfinityType.__eq__cCst||j�Sr
rrrrrr6szNegativeInfinityType.__ne__cCsdSrrrrrrr9szNegativeInfinityType.__gt__cCsdSrrrrrrr<szNegativeInfinityType.__ge__rcCstSr
)rrrrrr ?szNegativeInfinityType.__neg__N)r!r"r#r$r	r%rr&r'rrrrrrrr rrrrr&srN)rrrrrrrr�<module>ssite-packages/pip/_vendor/packaging/__pycache__/version.cpython-39.pyc000064400000031372150564745010021777 0ustar00a

��Re�8�	@sPddlZddlZddlZddlZddlmZmZmZmZm	Z	m
Z
mZddlm
Z
mZmZmZgd�ZeeefZeee
eeffZeeeefZeee
eee
eefe
eeffdffZe
ee
edfeeeefZe
ee
edffZeeeefeeefgefZe�dgd��Zeed	d
�dd�ZGd
d�de�Z Gdd�d�Z!Gdd�de!�Z"e�#dej$�Z%dddddd�Z&eeed�dd�Z'eed
�dd�Z(dZ)Gdd�de!�Z*eeee+e	fee
eefd �d!d"�Z,e�#d#�Z-eeed$�d%d&�Z.ee
edfee
eefee
eefee
eefee
eed'�d(d)�Z/dS)*�N)�Callable�Iterator�List�Optional�SupportsInt�Tuple�Union�)�Infinity�InfinityType�NegativeInfinity�NegativeInfinityType)�parse�Version�
LegacyVersion�InvalidVersion�VERSION_PATTERN.�_Version)�epoch�release�dev�pre�post�local)rr��version�returncCs*z
t|�WSty$t|�YS0dS)z�
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    N)rrr)r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/version.pyr*s
rc@seZdZdZdS)rzF
    An invalid version was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__rrrrr6src@s�eZdZUeeefed<ed�dd�Zde	d�dd�Z
de	d�dd	�Zee	d�d
d�Z
de	d�dd
�Zde	d�dd�Zee	d�dd�ZdS)�_BaseVersion�_key�rcCs
t|j�S�N)�hashr$��selfrrr�__hash__?sz_BaseVersion.__hash__)�otherrcCst|t�stS|j|jkSr&��
isinstancer#�NotImplementedr$�r)r+rrr�__lt__Es
z_BaseVersion.__lt__cCst|t�stS|j|jkSr&r,r/rrr�__le__Ks
z_BaseVersion.__le__cCst|t�stS|j|jkSr&r,r/rrr�__eq__Qs
z_BaseVersion.__eq__cCst|t�stS|j|jkSr&r,r/rrr�__ge__Ws
z_BaseVersion.__ge__cCst|t�stS|j|jkSr&r,r/rrr�__gt__]s
z_BaseVersion.__gt__cCst|t�stS|j|jkSr&r,r/rrr�__ne__cs
z_BaseVersion.__ne__N)rr r!r�CmpKey�LegacyCmpKey�__annotations__�intr*�boolr0r1�objectr2r3r4r5rrrrr#<s
r#c@s�eZdZedd�dd�Zed�dd�Zed�dd	�Zeed�d
d��Zeed�dd
��Z	ee
d�dd��Zedd�dd��Zedd�dd��Z
edd�dd��Zedd�dd��Zedd�dd��Zeed�dd��Zeed�dd��Zeed�dd��ZdS) rNrcCst|�|_t|j�|_dSr&)�str�_version�_legacy_cmpkeyr$)r)rrrr�__init__ks
zLegacyVersion.__init__r%cCs|jSr&�r=r(rrr�__str__pszLegacyVersion.__str__cCsd|�d�S)Nz<LegacyVersion('�')>rr(rrr�__repr__sszLegacyVersion.__repr__cCs|jSr&r@r(rrr�publicvszLegacyVersion.publiccCs|jSr&r@r(rrr�base_versionzszLegacyVersion.base_versioncCsdS)N���rr(rrrr~szLegacyVersion.epochcCsdSr&rr(rrrr�szLegacyVersion.releasecCsdSr&rr(rrrr�szLegacyVersion.precCsdSr&rr(rrrr�szLegacyVersion.postcCsdSr&rr(rrrr�szLegacyVersion.devcCsdSr&rr(rrrr�szLegacyVersion.localcCsdS�NFrr(rrr�
is_prerelease�szLegacyVersion.is_prereleasecCsdSrGrr(rrr�is_postrelease�szLegacyVersion.is_postreleasecCsdSrGrr(rrr�
is_devrelease�szLegacyVersion.is_devrelease)rr r!r<r?rArC�propertyrDrEr9rrrrrrr:rHrIrJrrrrrjs2rz(\d+ | [a-z]+ | \.| -)�czfinal-�@)r�preview�-�rcr)�srccs\t�|�D]F}t�||�}|r
|dkr(q
|dd�dvrF|�d�Vq
d|Vq
dVdS)N�.r	�
0123456789��*�*final)�_legacy_version_component_re�split�_legacy_version_replacement_map�get�zfill)rQ�partrrr�_parse_version_parts�sr]cCsvd}g}t|���D]T}|�d�r^|dkrD|rD|ddkrD|��q*|r^|ddkr^|��qD|�|�q|t|�fS)NrFrUrVz*final-�00000000)r]�lower�
startswith�pop�append�tuple)rr�partsr\rrrr>�s


r>a�
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
c@speZdZe�dedejejB�Ze	dd�dd�Z
e	d�dd	�Ze	d�d
d�Ze
ed�dd
��Ze
eedfd�dd��Ze
eee	efd�dd��Ze
eed�dd��Ze
eed�dd��Ze
ee	d�dd��Ze
e	d�dd��Ze
e	d�dd��Ze
ed�dd��Ze
ed�dd ��Ze
ed�d!d"��Ze
ed�d#d$��Ze
ed�d%d&��Ze
ed�d'd(��ZdS))rz^\s*z\s*$Nrc
Cs�|j�|�}|s td|�d���t|�d�r:t|�d��ndtdd�|�d��d�D��t|�d	�|�d
��t|�d�|�d�p�|�d
��t|�d�|�d��t	|�d��d�|_
t|j
j|j
j
|j
j|j
j|j
j|j
j�|_dS)NzInvalid version: '�'rrcss|]}t|�VqdSr&)r9��.0�irrr�	<genexpr>
�z#Version.__init__.<locals>.<genexpr>rrR�pre_l�pre_n�post_l�post_n1�post_n2�dev_l�dev_nr)rrrrrr)�_regex�searchrr�groupr9rcrX�_parse_letter_version�_parse_local_versionr=�_cmpkeyrrrrrrr$)r)r�matchrrrr?s*���zVersion.__init__r%cCsd|�d�S)Nz
<Version('rBrr(rrrrCszVersion.__repr__cCs�g}|jdkr |�|j�d��|�d�dd�|jD���|jdurb|�d�dd�|jD���|jdur~|�d|j���|jdur�|�d	|j���|jdur�|�d
|j���d�|�S)Nr�!rRcss|]}t|�VqdSr&�r<�rg�xrrrri(rjz"Version.__str__.<locals>.<genexpr>�css|]}t|�VqdSr&rzr{rrrri,rjz.postz.dev�+)rrb�joinrrrrr�r)rdrrrrA s




zVersion.__str__cCs|jj}|Sr&)r=r)r)�_epochrrrr<sz
Version.epoch.cCs|jj}|Sr&)r=r)r)�_releaserrrrAszVersion.releasecCs|jj}|Sr&)r=r)r)�_prerrrrFszVersion.precCs|jjr|jjdSdS�Nr	)r=rr(rrrrKszVersion.postcCs|jjr|jjdSdSr�)r=rr(rrrrOszVersion.devcCs(|jjr d�dd�|jjD��SdSdS)NrRcss|]}t|�VqdSr&rzr{rrrriVrjz Version.local.<locals>.<genexpr>)r=rrr(rrrrSsz
Version.localcCst|��dd�dS)Nr~r	r)r<rXr(rrrrDZszVersion.publiccCsFg}|jdkr |�|j�d��|�d�dd�|jD���d�|�S)NrryrRcss|]}t|�VqdSr&rzr{rrrrigrjz'Version.base_version.<locals>.<genexpr>r})rrbrrr�rrrrE^s

zVersion.base_versioncCs|jdup|jduSr&)rrr(rrrrHkszVersion.is_prereleasecCs
|jduSr&)rr(rrrrIoszVersion.is_postreleasecCs
|jduSr&)rr(rrrrJsszVersion.is_devreleasecCst|j�dkr|jdSdS)Nr	r��lenrr(rrr�majorwsz
Version.majorcCst|j�dkr|jdSdS)N�r	rr�r(rrr�minor{sz
Version.minorcCst|j�dkr|jdSdS)N�r�rr�r(rrr�microsz
Version.micro) rr r!�re�compiler�VERBOSE�
IGNORECASErrr<r?rCrArKr9rrrrrrrrrDrEr:rHrIrJr�r�r�rrrrr�s@r)�letter�numberrcCsv|rZ|durd}|��}|dkr&d}n(|dkr4d}n|dvrBd}n|dvrNd	}|t|�fS|sr|rrd	}|t|�fSdS)
Nr�alpha�a�beta�b)rLrrNrP)�rev�rr)r_r9)r�r�rrrru�s"ruz[\._-])rrcCs$|dur tdd�t�|�D��SdS)zR
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    Ncss&|]}|��s|��nt|�VqdSr&)�isdigitr_r9)rgr\rrrri�s�z'_parse_local_version.<locals>.<genexpr>)rc�_local_version_separatorsrX)rrrrrv�s
�rv)rrrrrrrcCs�tttt�dd�t|�����}|dur>|dur>|dur>t}n|durLt}n|}|dur^t}n|}|durpt}	n|}	|dur�t}
ntdd�|D��}
|||||	|
fS)NcSs|dkS)Nrr)r|rrr�<lambda>�rjz_cmpkey.<locals>.<lambda>css(|] }t|t�r|dfnt|fVqdS)r}N)r-r9rrfrrrri�sz_cmpkey.<locals>.<genexpr>)rc�reversed�list�	itertools�	dropwhilerr
)rrrrrrr�r��_post�_dev�_localrrrrw�s(�	�rw)0�collectionsr�r��warnings�typingrrrrrrr�_structuresr
rrr
�__all__�
InfiniteTypesr<r9�PrePostDevType�SubLocalType�	LocalTyper6r7r:�VersionComparisonMethod�
namedtuplerr�
ValueErrorrr#rr�r�rWrYr]r>rr�bytesrur�rvrwrrrr�<module>s|$

���������.9�	 
�$



�site-packages/pip/_vendor/packaging/__pycache__/__about__.cpython-39.pyc000064400000001165150564745010022215 0ustar00a

��Re��@s0gd�ZdZdZdZdZdZdZdZdeZd	S)
)�	__title__�__summary__�__uri__�__version__�
__author__�	__email__�__license__�
__copyright__�	packagingz"Core utilities for Python packagesz!https://github.com/pypa/packagingz21.0z)Donald Stufft and individual contributorszdonald@stufft.iozBSD-2-Clause or Apache-2.0z2014-2019 %sN)	�__all__rrrrrrrr�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/__about__.py�<module>ssite-packages/pip/_vendor/packaging/__pycache__/markers.cpython-39.pyc000064400000022424150564745010021754 0ustar00a

��Re'!�	@s�UddlZddlZddlZddlZddlmZmZmZmZm	Z	m
Z
mZddlm
Z
mZmZmZmZmZmZmZmZddlmZmZgd�ZeeegefZGdd�de�ZGd	d
�d
e�Z Gdd�de�Z!Gd
d�d�Z"Gdd�de"�Z#Gdd�de"�Z$Gdd�de"�Z%ed�ed�Bed�Bed�Bed�Bed�Bed�Bed�Bed�Bed�Bed�Bed �Bed!�Bed"�Bed#�Bed$�Bed%�Bed&�BZ&ddddddd'�Z'e&�(d(d)��ed*�ed+�Bed,�Bed-�Bed.�Bed/�Bed0�Bed1�BZ)e)ed2�Bed3�BZ*e*�(d4d)��ed5�ed6�BZ+e+�(d7d)��ed8�ed9�BZ,e&e+BZ-ee-e*e-�Z.e.�(d:d)��ed;��/�Z0ed<��/�Z1e
�Z2e.ee0e2e1�BZ3e2e3ee,e2�>ee2eZ4eeeefeed=�d>d?�Z5d]eeee
e"dAfefe	eedB�dCdD�Z6dEd)�dFd)�ej7ej8ej9ej:ej;ej<dG�Z=eeefe>dH<ee%eedI�dJdK�Z?GdLdM�dM�Z@e@�ZAeeefeedN�dOdP�ZBeeeeefedQ�dRdS�ZCdTedU�dVdW�ZDeeefdX�dYdZ�ZEGd[d\�d\�ZFdS)^�N)�Any�Callable�Dict�List�Optional�Tuple�Union)	�Forward�Group�Literal�ParseException�ParseResults�QuotedString�
ZeroOrMore�	stringEnd�stringStart�)�InvalidSpecifier�	Specifier)�
InvalidMarker�UndefinedComparison�UndefinedEnvironmentName�Marker�default_environmentc@seZdZdZdS)rzE
    An invalid marker was found, users should refer to PEP 508.
    N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/markers.pyr$src@seZdZdZdS)rzP
    An invalid operation was attempted on a value that doesn't support it.
    Nrrrrr r*src@seZdZdZdS)rz\
    A name was attempted to be used that does not exist inside of the
    environment.
    Nrrrrr r0src@sFeZdZedd�dd�Zed�dd�Zed�dd	�Zed�d
d�ZdS)�NodeN)�value�returncCs
||_dS�N)r")�selfr"rrr �__init__8sz
Node.__init__�r#cCs
t|j�Sr$)�strr"�r%rrr �__str__;szNode.__str__cCsd|jj�d|�d�S)N�<z('�')>)�	__class__rr)rrr �__repr__>sz
Node.__repr__cCst�dSr$)�NotImplementedErrorr)rrr �	serializeAszNode.serialize)	rrrrr&r(r*r.r0rrrr r!7sr!c@seZdZed�dd�ZdS)�Variabler'cCst|�Sr$�r(r)rrr r0FszVariable.serializeN�rrrr(r0rrrr r1Esr1c@seZdZed�dd�ZdS)�Valuer'cCsd|�d�S)N�"rr)rrr r0KszValue.serializeNr3rrrr r4Jsr4c@seZdZed�dd�ZdS)�Opr'cCst|�Sr$r2r)rrr r0PszOp.serializeNr3rrrr r6Osr6�implementation_version�platform_python_implementation�implementation_name�python_full_version�platform_release�platform_version�platform_machine�platform_system�python_version�sys_platform�os_name�os.name�sys.platform�platform.version�platform.machine�platform.python_implementation�python_implementation�extra)rBrCrDrErFrGcCstt�|d|d��S�Nr)r1�ALIASES�get��s�l�trrr �<lambda>p�rPz===�==�>=�<=�!=z~=�>r+�not in�incCst|d�SrI)r6rLrrr rPwrQ�'r5cCst|d�SrI)r4rLrrr rPzrQ�and�orcCst|d�SrI)�tuplerLrrr rP�rQ�(�))�resultsr#cCs t|t�rdd�|D�S|SdS)NcSsg|]}t|��qSr)�_coerce_parse_result)�.0�irrr �
<listcomp>�rQz(_coerce_parse_result.<locals>.<listcomp>)�
isinstancer
)r_rrr r`�s
r`T.)�marker�firstr#cCs�t|tttf�sJ�t|t�rHt|�dkrHt|dttf�rHt|d�St|t�r�dd�|D�}|rnd�|�Sdd�|�dSn"t|t�r�d�dd	�|D��S|SdS)
Nrrcss|]}t|dd�VqdS)F)rfN)�_format_marker�ra�mrrr �	<genexpr>�rQz!_format_marker.<locals>.<genexpr>� r]r^cSsg|]}|���qSr)r0rhrrr rc�rQz"_format_marker.<locals>.<listcomp>)rd�listr\r(�lenrg�join)rerf�innerrrr rg�s �
��


rgcCs||vSr$r��lhs�rhsrrr rP�rQcCs||vSr$rrprrr rP�rQ)rXrWr+rTrRrUrSrV�
_operators)rq�oprrr#cCstztd�|��|g��}Wnty,Yn0|�|�St�|���}|durjtd|�d|�d|�d���|||�S)N�z
Undefined z on z and �.)rrnr0r�containsrsrKr)rqrtrr�spec�operrrr �_eval_op�s
rzc@seZdZdS)�	UndefinedN)rrrrrrr r{�sr{)�environment�namer#cCs(|�|t�}t|t�r$t|�d���|S)Nz* does not exist in evaluation environment.)rK�
_undefinedrdr{r)r|r}r"rrr �_get_env�s
�r)�markersr|r#c	Cs�gg}|D]�}t|tttf�s"J�t|t�rB|d�t||��q
t|t�r�|\}}}t|t�rtt||j�}|j}n|j}t||j�}|d�t	|||��q
|dvs�J�|dkr
|�g�q
t
dd�|D��S)N���)rZr[r[css|]}t|�VqdSr$)�all)ra�itemrrr rj�rQz$_evaluate_markers.<locals>.<genexpr>)rdrlr\r(�append�_evaluate_markersr1rr"rz�any)	r�r|�groupsrerqrtrr�	lhs_value�	rhs_valuerrr r��s"



r�zsys._version_info)�infor#cCs2d�|�}|j}|dkr.||dt|j�7}|S)Nz{0.major}.{0.minor}.{0.micro}�finalr)�format�releaselevelr(�serial)r��version�kindrrr �format_full_version�s

r�r'cCs^ttjj�}tjj}||tjt��t��t�	�t��t�
�t��d�t�
�dd��tjd�S)Nrv�)r9r7rAr=r;r>r<r:r8r?r@)r��sys�implementationr�r}�os�platform�machine�release�systemr?rGrn�python_version_tuple)�iverr9rrr rs�rc@sVeZdZedd�dd�Zed�dd�Zed�dd	�Zd
eeeefe	d
�dd�Z
dS)rN)rer#cCsbztt�|��|_WnHty\}z0td|�d||j|jd�����WYd}~n
d}~00dS)NzInvalid marker: z, parse error at �)r`�MARKER�parseString�_markersrr�loc)r%re�errr r&s��zMarker.__init__r'cCs
t|j�Sr$)rgr�r)rrr r*szMarker.__str__cCsd|�d�S)Nz	<Marker('r,rr)rrr r. szMarker.__repr__)r|r#cCs$t�}|dur|�|�t|j|�S)a$Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        N)r�updater�r�)r%r|�current_environmentrrr �evaluate#s	
zMarker.evaluate)N)rrrr(r&r*r.rr�boolr�rrrr rs	r)T)G�operatorr�r�r��typingrrrrrrr�pip._vendor.pyparsingr	r
r�Lrr
rrrr�
specifiersrr�__all__r(r��Operator�
ValueErrorrrrr!r1r4r6�VARIABLErJ�setParseAction�VERSION_CMP�	MARKER_OP�MARKER_VALUE�BOOLOP�
MARKER_VAR�MARKER_ITEM�suppress�LPAREN�RPAREN�MARKER_EXPR�MARKER_ATOMr�r`rg�lt�le�eq�ne�ge�gtrs�__annotations__rzr{r~rr�r�rrrrrr �<module>s�
$,��������	�
���
�������>� ���site-packages/pip/_vendor/packaging/__pycache__/specifiers.cpython-39.pyc000064400000053351150564745010022447 0ustar00a

��Re�x�@s�ddlZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
mZmZm
Z
mZmZmZddlmZddlmZmZmZeeefZeeeefZeded�ZeeegefZGdd	�d	e�ZGd
d�dejd�Z Gd
d�de �Z!Gdd�de!�Z"edeegefedeegefd�dd�Z#Gdd�de!�Z$e�%d�Z&ee
ed�dd�Z'eed�dd�Z(e
ee
eee
ee
efd�dd�Z)Gd d!�d!e �Z*dS)"�N)�Callable�Dict�Iterable�Iterator�List�Optional�Pattern�Set�Tuple�TypeVar�Union�)�canonicalize_version)�
LegacyVersion�Version�parse�VersionTypeVar)�boundc@seZdZdZdS)�InvalidSpecifierzH
    An invalid specifier was found, users should refer to PEP 440.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/specifiers.pyr!src@s�eZdZejed�dd��Zejed�dd��Zeje	e
d�dd��Zeje	e
d�d	d
��Zej
ee
d�dd��Zeje
d
d�dd��Zejdeee
e
d�dd��Zejdeeee
eed�dd��Zd
S)�
BaseSpecifier��returncCsdS)z�
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        Nr��selfrrr�__str__(szBaseSpecifier.__str__cCsdS)zF
        Returns a hash value for this Specifier like object.
        Nrrrrr�__hash__/szBaseSpecifier.__hash__��otherrcCsdS)zq
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        Nr�rr#rrr�__eq__5szBaseSpecifier.__eq__cCsdS)zu
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        Nrr$rrr�__ne__<szBaseSpecifier.__ne__cCsdS)zg
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        Nrrrrr�prereleasesCszBaseSpecifier.prereleasesN��valuercCsdS)zd
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        Nr�rr)rrrr'Js��itemr'rcCsdS)zR
        Determines if the given item is contained within this specifier.
        Nr�rr,r'rrr�containsQszBaseSpecifier.contains��iterabler'rcCsdS)z�
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        Nr)rr0r'rrr�filterWszBaseSpecifier.filter)N)N)rrr�abc�abstractmethod�strr �intr!�object�boolr%r&�abstractpropertyrr'�setterr.rrr1rrrrr's(��r)�	metaclassc@sbeZdZUiZeeefed<eeed<d-eee	dd�dd�Z
ed�d	d
�Zed�dd�Ze
eeefd�d
d��Zed�dd�Zee	d�dd�Zee	d�dd�Zeed�dd�Zeed�dd�Ze
ed�dd��Ze
ed�dd��Ze
ee	d�d d!��Zeje	dd"�d#d!��Zee	d$�d%d&�Zd.eee	e	d'�d(d)�Zd/e e!ee	e e!d*�d+d,�Z"dS)0�_IndividualSpecifier�
_operators�_regex�N��specr'rcCsH|j�|�}|s td|�d���|�d���|�d���f|_||_dS)NzInvalid specifier: '�'�operator�version)r=�searchr�group�strip�_spec�_prereleases)rr@r'�matchrrr�__init__fs�z_IndividualSpecifier.__init__rcCs0|jdurd|j��nd}d�|jjt|�|�S)N�, prereleases=r>z<{}({!r}{})>)rHr'�format�	__class__rr4�r�prerrr�__repr__ss
��z_IndividualSpecifier.__repr__cCsdj|j�S)Nz{}{})rLrGrrrrr |sz_IndividualSpecifier.__str__cCs|jdt|jd�fS)Nrr
)rGrrrrr�_canonical_specsz$_IndividualSpecifier._canonical_speccCs
t|j�S�N)�hashrQrrrrr!�sz_IndividualSpecifier.__hash__r"cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrR)�
isinstancer4rMr�NotImplementedrQr$rrrr%�s
z_IndividualSpecifier.__eq__cCsRt|t�r6z|�t|��}WqFty2tYS0nt||j�sFtS|j|jkSrR)rTr4rMrrUrGr$rrrr&�s
z_IndividualSpecifier.__ne__)�oprcCst|d|j|���}|S)N�	_compare_)�getattrr<)rrV�operator_callablerrr�
_get_operator�s�z"_IndividualSpecifier._get_operator�rCrcCst|ttf�st|�}|SrR)rTrrr�rrCrrr�_coerce_version�sz$_IndividualSpecifier._coerce_versioncCs
|jdS)Nr�rGrrrrrB�sz_IndividualSpecifier.operatorcCs
|jdS)Nr
r^rrrrrC�sz_IndividualSpecifier.versioncCs|jSrR�rHrrrrr'�sz _IndividualSpecifier.prereleasesr(cCs
||_dSrRr_r*rrrr'�s�r,rcCs
|�|�SrR�r.�rr,rrr�__contains__�sz!_IndividualSpecifier.__contains__r+cCs>|dur|j}|�|�}|jr&|s&dS|�|j�}|||j�S�NF)r'r]�
is_prereleaserZrBrC)rr,r'�normalized_itemrYrrrr.�s

z_IndividualSpecifier.containsr/ccs�d}g}d|dur|ndi}|D]F}|�|�}|j|fi|��r |jr\|s\|js\|�|�q d}|Vq |s�|r�|D]
}|VqtdS)NFr'T)r]r.rer'�append)rr0r'�yielded�found_prereleases�kwrC�parsed_versionrrrr1�s"
��z_IndividualSpecifier.filter)r>N)N)N)#rrrr<rr4�__annotations__rrr7rJrPr �propertyr
rQr5r!r6r%r&�CallableOperatorrZ�UnparsedVersion�
ParsedVersionr]rBrCr'r9rcr.rrr1rrrrr;as>

	����r;cs�eZdZdZe�dedejejB�Zdddddd	d
�Z	d e
eedd
��fdd�
Z
eed�dd�Zee
ed�dd�Zee
ed�dd�Zee
ed�dd�Zee
ed�dd�Zee
ed�dd�Zee
ed�dd�Z�ZS)!�LegacySpecifiera�
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        �^\s*�\s*$�equal�	not_equal�less_than_equal�greater_than_equal�	less_than�greater_than)�==�!=�<=�>=�<�>r>Nr?cst��||�t�dt�dS)NzZCreating a LegacyVersion has been deprecated and will be removed in the next major release)�superrJ�warnings�warn�DeprecationWarning)rr@r'�rMrrrJs
�zLegacySpecifier.__init__r[cCst|t�stt|��}|SrR)rTrr4r\rrrr]s
zLegacySpecifier._coerce_version��prospectiver@rcCs||�|�kSrR�r]�rr�r@rrr�_compare_equalszLegacySpecifier._compare_equalcCs||�|�kSrRr�r�rrr�_compare_not_equalsz"LegacySpecifier._compare_not_equalcCs||�|�kSrRr�r�rrr�_compare_less_than_equal"sz(LegacySpecifier._compare_less_than_equalcCs||�|�kSrRr�r�rrr�_compare_greater_than_equal%sz+LegacySpecifier._compare_greater_than_equalcCs||�|�kSrRr�r�rrr�_compare_less_than*sz"LegacySpecifier._compare_less_thancCs||�|�kSrRr�r�rrr�_compare_greater_than-sz%LegacySpecifier._compare_greater_than)r>N)rrr�
_regex_str�re�compile�VERBOSE�
IGNORECASEr=r<r4rr7rJrorr]r�r�r�r�r�r��
__classcell__rrr�rrq�s&�		�rq�	Specifier)�fnrcs&t���dtttd��fdd��}|S)Nr�)rr�r@rcst|t�sdS�|||�Srd)rTrr��r�rr�wrapped4s
z)_require_version_compare.<locals>.wrapped)�	functools�wrapsrpr4r7)r�r�rr�r�_require_version_compare1sr�c	@seZdZdZe�dedejejB�Zdddddd	d
dd�Z	e
eee
d
�dd��Ze
eee
d
�dd��Ze
eee
d
�dd��Ze
eee
d
�dd��Ze
eee
d
�dd��Ze
eee
d�dd��Ze
eee
d�dd��Zeee
d
�dd�Zee
d�d d!��Zeje
d"d#�d$d!��Zd"S)%r�a
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        rrrs�
compatiblertrurvrwrxry�	arbitrary)�~=rzr{r|r}r~r�===r�cCsJd�tt�tt|���dd��}|d7}|�d�||�oH|�d�||�S)N�.����.*r}rz)�join�list�	itertools�	takewhile�_is_not_suffix�_version_splitrZ)rr�r@�prefixrrr�_compare_compatible�s��zSpecifier._compare_compatiblec	Csz|�d�rVt|j�}t|dd��}tt|��}|dt|��}t||�\}}||kSt|�}|jsnt|j�}||kSdS)Nr����)�endswithr�publicr�r4�len�_pad_version�local)	rr�r@�
split_spec�split_prospective�shortened_prospective�padded_spec�padded_prospective�spec_versionrrrr��s

�
zSpecifier._compare_equalcCs|�||�SrR)r�r�rrrr��szSpecifier._compare_not_equalcCst|j�t|�kSrR�rr�r�rrrr��sz"Specifier._compare_less_than_equalcCst|j�t|�kSrRr�r�rrrr��sz%Specifier._compare_greater_than_equal)r��spec_strrcCs<t|�}||ksdS|js8|jr8t|j�t|j�kr8dSdS�NFT)rre�base_version�rr�r�r@rrrr��szSpecifier._compare_less_thancCs^t|�}||ksdS|js8|jr8t|j�t|j�kr8dS|jdurZt|j�t|j�krZdSdSr�)r�is_postreleaser�r�r�rrrr�s
zSpecifier._compare_greater_thancCst|���t|���kSrR)r4�lowerr�rrr�_compare_arbitrary7szSpecifier._compare_arbitraryrcCsR|jdur|jS|j\}}|dvrN|dkr@|�d�r@|dd�}t|�jrNdSdS)N)rzr}r|r�r�rzr�r�TF)rHrGr�rre)rrBrCrrrr':s


zSpecifier.prereleasesNr(cCs
||_dSrRr_r*rrrr'Ss)rrrr�r�r�r�r�r=r<r�rpr4r7r�r�r�r�r�r�r�rr�rmr'r9rrrrr�=s@]�'�	z^([0-9]+)((?:a|b|c|rc)[0-9]+)$r[cCs@g}|�d�D],}t�|�}|r0|�|���q|�|�q|S)Nr�)�split�
_prefix_regexrD�extend�groupsrg)rC�resultr,rIrrrr�[s
r�)�segmentrcst�fdd�dD��S)Nc3s|]}��|�VqdSrR)�
startswith)�.0r��r�rr�	<genexpr>gsz!_is_not_suffix.<locals>.<genexpr>)�dev�a�b�rc�post)�anyr�rr�rr�fs�r�)�left�rightrc
Cs�gg}}|�tt�dd�|���|�tt�dd�|���|�|t|d�d��|�|t|d�d��|�ddgtdt|d�t|d���|�ddgtdt|d�t|d���ttj|��ttj|��fS)NcSs|��SrR��isdigit��xrrr�<lambda>p�z_pad_version.<locals>.<lambda>cSs|��SrRr�r�rrrr�qr�rr
�0)rgr�r�r�r��insert�max�chain)r�r��
left_split�right_splitrrrr�ls
,,r�c@seZdZd%eeedd�dd�Zed�dd�Zed�d	d
�Ze	d�dd�Z
edefdd
�dd�Ze
ed
�dd�Ze
ed
�dd�Ze	d�dd�Zeed�dd�Zeeed�dd��Zejedd�dd��Zeed�dd�Zd&eeeed�d d!�Zd'eeeeeed"�d#d$�ZdS)(�SpecifierSetr>N)�
specifiersr'rc	Csldd�|�d�D�}t�}|D]8}z|�t|��WqtyT|�t|��Yq0qt|�|_||_dS)NcSsg|]}|��r|���qSr)rF�r��srrr�
<listcomp>�r�z)SpecifierSet.__init__.<locals>.<listcomp>�,)	r��set�addr�rrq�	frozenset�_specsrH)rr�r'�split_specifiers�parsed�	specifierrrrrJs
zSpecifierSet.__init__rcCs*|jdurd|j��nd}d�t|�|�S)NrKr>z<SpecifierSet({!r}{})>)rHr'rLr4rNrrrrP�s
��zSpecifierSet.__repr__cCsd�tdd�|jD���S)Nr�css|]}t|�VqdSrR)r4r�rrrr��r�z'SpecifierSet.__str__.<locals>.<genexpr>)r��sortedr�rrrrr �szSpecifierSet.__str__cCs
t|j�SrR)rSr�rrrrr!�szSpecifierSet.__hash__r"cCs�t|t�rt|�}nt|t�s"tSt�}t|j|jB�|_|jdurX|jdurX|j|_n<|jdurv|jdurv|j|_n|j|jkr�|j|_ntd��|S)NzFCannot combine SpecifierSets with True and False prerelease overrides.)rTr4r�rUr�r�rH�
ValueError)rr#r�rrr�__and__�s 





�zSpecifierSet.__and__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrR�rTr4r;r�rUr�r$rrrr%�s

zSpecifierSet.__eq__cCs6t|ttf�rtt|��}nt|t�s*tS|j|jkSrRr�r$rrrr&�s

zSpecifierSet.__ne__cCs
t|j�SrR)r�r�rrrr�__len__�szSpecifierSet.__len__cCs
t|j�SrR)�iterr�rrrr�__iter__�szSpecifierSet.__iter__cCs.|jdur|jS|jsdStdd�|jD��S)Ncss|]}|jVqdSrR�r'r�rrrr��r�z+SpecifierSet.prereleases.<locals>.<genexpr>)rHr�r�rrrrr'�s

zSpecifierSet.prereleasesr(cCs
||_dSrRr_r*rrrr'�sr`cCs
|�|�SrRrarbrrrrc�szSpecifierSet.__contains__r+csLt�ttf�st����dur$|j��s2�jr2dSt��fdd�|jD��S)NFc3s|]}|j��d�VqdS)r�Nrar��r,r'rrr�r�z(SpecifierSet.contains.<locals>.<genexpr>)rTrrrr're�allr�r-rr�rr.�s
zSpecifierSet.containsr/cCs�|dur|j}|jr6|jD]}|j|t|�d�}q|Sg}g}|D]P}t|ttf�s^t|�}n|}t|t�rnqB|jr�|s�|s�|�	|�qB|�	|�qB|s�|r�|dur�|S|SdS)Nr�)
r'r�r1r7rTrrrrerg)rr0r'r@�filteredrir,rkrrrr1	s*



zSpecifierSet.filter)r>N)N)N)rrrr4rr7rJrPr r5r!rr�r6r%r&r�rr;r�rmr'r9rorcr.rrr1rrrrr�~s8��	����r�)+r2r�r�r�r��typingrrrrrrrr	r
rr�utilsrrCrrrrpr4rorr7rnr�r�ABCMetarr;rqr�r�r�r�r�r�r�r�rrrr�<module>s64:=�
*site-packages/pip/_vendor/packaging/__pycache__/__init__.cpython-39.pyc000064400000000745150564745010022051 0ustar00a

��Re��@s4ddlmZmZmZmZmZmZmZmZgd�Z	dS)�)�
__author__�
__copyright__�	__email__�__license__�__summary__�	__title__�__uri__�__version__)rrrr	rrrrN)
�	__about__rrrrrrrr	�__all__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/__init__.py�<module>s(site-packages/pip/_vendor/packaging/__pycache__/_musllinux.cpython-39.pyc000064400000011050150564745010022500 0ustar00a

��Re�	@stdZddlZddlZddlZddlZddlZddlZddlZddlZddl	m
Z
mZmZm
Z
mZe
eeeedfd�dd�Ze
ee
ed�d	d
�ZGdd�de�Zee
ed
�dd�Ze��ee
ed�dd��Zeeed�dd�Zedk�rpddlZe��Ze�d��sJd��ede�edeej��eddd�ee� dde�!d d!�d"��D]Z"ee"d#d��q\dS)$z�PEP 656 support.

This module implements logic to detect if the currently running Python is
linked against musl, and what musl version is used.
�N)�IO�Iterator�
NamedTuple�Optional�Tuple.)�f�fmt�returncCst�||�t�|���S�N)�struct�unpack�read�calcsize)rr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/_musllinux.py�_read_unpackedsr)rr	c	Csh|�d�zt|d�}Wntjy.YdS0|dd�td�krHdS|�t�d�d�zdd	d
�|d\}}}Wnty�YdS0tj|�}zt||�\}}}}}}}	Wntjy�YdS0t	|	d�D]�}
|�|||
�z|t||��\}}}
Wntj�yYdS0|dk�r,q�|�|�t
�|�|
���
d�}d
|v�r\dS|SdS)z�Detect musl libc location by parsing the Python executable.

    Based on: https://gist.github.com/lyssdod/f51579ae8d93c8657a5564aefc2ffbca
    ELF header: https://refspecs.linuxfoundation.org/elf/gabi4+/ch4.eheader.html
    r�16BN�sELF�HHI�)�IIIIHHH�IIIIIIII)rrr)�QQQIHHH�IIQQQQQQ)r��)rr���musl)�seekrr�error�tupler�KeyError�operator�
itemgetter�range�os�fsdecoder
�strip)r�ident�e_fmt�p_fmt�p_idx�p_get�_�e_phoff�e_phentsize�e_phnum�i�p_type�p_offset�p_filesz�interpreterrrr�_parse_ld_musl_from_elfsF
��




r7c@seZdZUeed<eed<dS)�_MuslVersion�major�minorN)�__name__�
__module__�__qualname__�int�__annotations__rrrrr8Gs
r8)�outputr	cCsxdd�dd�|��D�D�}t|�dks<|ddd�dkr@dSt�d	|d
�}|sXdStt|�d
��t|�d��d�S)NcSsg|]}|r|�qSrr��.0�nrrr�
<listcomp>M�z'_parse_musl_version.<locals>.<listcomp>css|]}|��VqdSr
)r(rArrr�	<genexpr>MrEz&_parse_musl_version.<locals>.<genexpr>rrrrzVersion (\d+)\.(\d+)r)r9r:)�
splitlines�len�re�matchr8r>�group)r@�lines�mrrr�_parse_musl_versionLs rN)�
executabler	c	Cs�t���N}z|�t|d��}Wn ty>YWd�dS0t|�}Wd�n1s\0Y|sndStj|gtjdd�}t	|j
�S)a`Detect currently-running musl runtime version.

    This is done by checking the specified executable's dynamic linking
    information, and invoking the loader to parse its output for a version
    string. If the loader is musl, the output would be something like::

        musl libc (x86_64)
        Version 1.2.2
        Dynamic Program Loader
    �rbNT)�stderr�universal_newlines)�
contextlib�	ExitStack�
enter_context�open�IOErrorr7�
subprocess�run�PIPErNrQ)rO�stackr�ld�procrrr�_get_musl_versionVs
&r^)�archr	ccsHttj�}|durdSt|jdd�D]}d|j�d|�d|��Vq$dS)aTGenerate musllinux tags compatible to the current platform.

    :param arch: Should be the part of platform tag after the ``linux_``
        prefix, e.g. ``x86_64``. The ``linux_`` prefix is assumed as a
        prerequisite for the current platform to be musllinux-compatible.

    :returns: An iterator of compatible musllinux tags.
    N����
musllinux_r.)r^�sysrOr%r:r9)r_�sys_muslr:rrr�
platform_tagsns
	
rd�__main__zlinux-z	not linuxzplat:zmusl:ztags:� )�endz[.-]r.�-rr`z
      )#�__doc__rS�	functoolsr#r&rIrrXrb�typingrrrrr�bytes�strr>rr7r8rN�	lru_cacher^rdr;�	sysconfig�get_platform�plat�
startswith�printrO�sub�split�trrrr�<module>s21


"site-packages/pip/_vendor/packaging/__pycache__/_manylinux.cpython-39.pyc000064400000016245150564745010022477 0ustar00a

��Re�,�@szUddlZddlZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
mZmZm
Z
Gdd�d�Zeed�dd�Zed�dd	�Zed�d
d�Zeed�d
d�Ze�dd��Ze	eefed<Gdd�de�Zeed�dd�Zeed�dd�Zeed�dd�Zee
eefd�dd�Ze��e
eefd�dd��Zeeeed�d d!�Z d"d#d$d%�Z!eee
ed&�d'd(�Z"dS))�N)�IO�Dict�Iterator�
NamedTuple�Optional�Tuplec@s`eZdZGdd�de�ZdZdZdZdZdZ	dZ
dZdZd	Z
d
ZdZdZeed
d�dd�Zd
S)�_ELFFileHeaderc@seZdZdZdS)z$_ELFFileHeader._InvalidELFFileHeaderz7
        An invalid ELF file header was found.
        N)�__name__�
__module__�__qualname__�__doc__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/packaging/_manylinux.py�_InvalidELFFileHeadersriFLE�����(�>l~iiN)�file�returncszttd��fdd�}|d�|_|j|jkr2t���|d�|_|j|j|jhvrVt���|d�|_	|j	|j
|jhvrzt���|d�|_|d�|_
|d�|_��d�|_|j	|j
kr�dnd}|j	|j
kr�d	nd}|j	|j
kr�d
nd}|j|jkr�|n|}||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_||�|_dS)N)�fmtrcsFz ��t�|��}t�||�}Wntjy<t���Yn0|dS)Nr)�read�struct�calcsize�unpack�errorrr)r�data�result�rr
rr$sz'_ELFFileHeader.__init__.<locals>.unpackz>I�B�z<Hz>Hz<Iz<Qz>Q)�str�int�
e_ident_magic�ELF_MAGIC_NUMBERrr�
e_ident_class�
ELFCLASS32�
ELFCLASS64�e_ident_data�ELFDATA2LSB�ELFDATA2MSB�e_ident_version�
e_ident_osabi�e_ident_abiversionr�e_ident_pad�e_type�	e_machine�	e_version�e_entry�e_phoff�e_shoff�e_flags�e_ehsize�e_phentsize�e_phnum�e_shentsize�e_shnum�
e_shstrndx)�selfrr�format_h�format_i�format_q�format_pr
r r�__init__#s>

















z_ELFFileHeader.__init__)r	r
r�
ValueErrorrr&r(r)r+r,�EM_386�EM_S390�EM_ARM�	EM_X86_64�EF_ARM_ABIMASK�EF_ARM_ABI_VER5�EF_ARM_ABI_FLOAT_HARDr�bytesrCr
r
r
rrsr)rc
CsZz8ttjd��}t|�}Wd�n1s,0YWntttjfyTYdS0|S)N�rb)�open�sys�
executabler�OSError�	TypeErrorr)�f�
elf_headerr
r
r�_get_elf_headerLs*rUcCsnt�}|durdS|j|jk}||j|jkM}||j|jkM}||j|j@|j	kM}||j|j
@|j
kM}|S�NF)rUr'r(r*r+r2rGr7rIrJrK�rTrr
r
r�_is_linux_armhfUs
�
�rXcCsBt�}|durdS|j|jk}||j|jkM}||j|jkM}|SrV)rUr'r(r*r+r2rErWr
r
r�_is_linux_i686hsrY)�archrcCs$|dkrt�S|dkrt�S|dvS)N�armv7l�i686>�ppc64le�x86_64�aarch64�ppc64�s390x)rXrY)rZr
r
r�_have_compatible_abirs
rbcCsdS)N�2r
r
r
r
r�<lambda>�rd�_LAST_GLIBC_MINORc@seZdZUeed<eed<dS)�
_GLibCVersion�major�minorN)r	r
rr$�__annotations__r
r
r
rrg�s
rgcCsHz&t�d�}|dusJ�|��\}}WnttttfyBYdS0|S)zJ
    Primary implementation of glibc_version_string using os.confstr.
    �CS_GNU_LIBC_VERSIONN)�os�confstr�split�AssertionError�AttributeErrorrQrD)�version_string�_�versionr
r
r�_glibc_version_string_confstr�s
rtcCs�zddl}Wnty YdS0z|�d�}WntyDYdS0z
|j}WntydYdS0|j|_|�}t|t	�s�|�
d�}|S)zG
    Fallback implementation of glibc_version_string using ctypes.
    rN�ascii)�ctypes�ImportError�CDLLrQ�gnu_get_libc_versionrp�c_char_p�restype�
isinstancer#�decode)rv�process_namespacery�version_strr
r
r�_glibc_version_string_ctypes�s"


r�cCst�p
t�S)z9Returns glibc version string, or None if not using glibc.)rtr�r
r
r
r�_glibc_version_string�sr�)rrcCs@t�d|�}|s$t�d|t�dSt|�d��t|�d��fS)a3Parse glibc version.

    We use a regexp instead of str.split because we want to discard any
    random junk that might come after the minor version -- this might happen
    in patched/forked versions of glibc (e.g. Linaro's version of glibc
    uses version strings like "2.20-2014.11"). See gh-3588.
    z$(?P<major>[0-9]+)\.(?P<minor>[0-9]+)z=Expected glibc version with 2 components major.minor, got: %s����r�rhri)�re�match�warnings�warn�RuntimeWarningr$�group)r�mr
r
r�_parse_glibc_version�s��r�cCst�}|durdSt|�S)Nr�)r�r�)rr
r
r�_get_glibc_version�sr�)�namerZrsrcCs�t�}||krdSzddl}Wnty2YdS0t|d�rh|�|d|d|�}|durdt|�SdS|tdd�kr�t|d�r�t|j�S|tdd	�kr�t|d
�r�t|j�S|tdd�kr�t|d�r�t|j	�SdS)
NFrT�manylinux_compatiblerr��manylinux1_compatible��manylinux2010_compatible��manylinux2014_compatible)
r��
_manylinuxrw�hasattrr��boolrgr�r�r�)r�rZrs�	sys_glibcr�rr
r
r�_is_compatible�s,






r��
manylinux2014�
manylinux2010�
manylinux1))rr�)rr�)rr�)�linuxrZrccs�t|�sdStdd�}|dvr(tdd�}tt��}|g}t|jddd�D]}t|}|�t||��qJ|D]�}|j|jkr�|j}nd}t|j|d�D]^}t|j|�}	dj|	�}
t	|
||	�r�|�
d|
�V|	tvr�t|	}t	|||	�r�|�
d|�Vq�qldS)	Nr�>r\r^�rr�zmanylinux_{}_{}r�)rbrgr��rangerhrf�appendri�formatr��replace�_LEGACY_MANYLINUX_MAP)r�rZ�too_old_glibc2�
current_glibc�glibc_max_list�glibc_major�glibc_minor�	glibc_max�	min_minor�
glibc_version�tag�
legacy_tagr
r
r�
platform_tagss.



r�)#�collections�	functoolsrlr�rrOr��typingrrrrrrrrUr�rXrYr#rb�defaultdictrfr$rjrgrtr�r�r��	lru_cacher�r�r�r�r
r
r
r�<module>s4
 =	

,�
site-packages/pip/_vendor/packaging/_manylinux.py000064400000026340150564745010016205 0ustar00import collections
import functools
import os
import re
import struct
import sys
import warnings
from typing import IO, Dict, Iterator, NamedTuple, Optional, Tuple


# Python does not provide platform information at sufficient granularity to
# identify the architecture of the running executable in some cases, so we
# determine it dynamically by reading the information from the running
# process. This only applies on Linux, which uses the ELF format.
class _ELFFileHeader:
    # https://en.wikipedia.org/wiki/Executable_and_Linkable_Format#File_header
    class _InvalidELFFileHeader(ValueError):
        """
        An invalid ELF file header was found.
        """

    ELF_MAGIC_NUMBER = 0x7F454C46
    ELFCLASS32 = 1
    ELFCLASS64 = 2
    ELFDATA2LSB = 1
    ELFDATA2MSB = 2
    EM_386 = 3
    EM_S390 = 22
    EM_ARM = 40
    EM_X86_64 = 62
    EF_ARM_ABIMASK = 0xFF000000
    EF_ARM_ABI_VER5 = 0x05000000
    EF_ARM_ABI_FLOAT_HARD = 0x00000400

    def __init__(self, file: IO[bytes]) -> None:
        def unpack(fmt: str) -> int:
            try:
                data = file.read(struct.calcsize(fmt))
                result: Tuple[int, ...] = struct.unpack(fmt, data)
            except struct.error:
                raise _ELFFileHeader._InvalidELFFileHeader()
            return result[0]

        self.e_ident_magic = unpack(">I")
        if self.e_ident_magic != self.ELF_MAGIC_NUMBER:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_class = unpack("B")
        if self.e_ident_class not in {self.ELFCLASS32, self.ELFCLASS64}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_data = unpack("B")
        if self.e_ident_data not in {self.ELFDATA2LSB, self.ELFDATA2MSB}:
            raise _ELFFileHeader._InvalidELFFileHeader()
        self.e_ident_version = unpack("B")
        self.e_ident_osabi = unpack("B")
        self.e_ident_abiversion = unpack("B")
        self.e_ident_pad = file.read(7)
        format_h = "<H" if self.e_ident_data == self.ELFDATA2LSB else ">H"
        format_i = "<I" if self.e_ident_data == self.ELFDATA2LSB else ">I"
        format_q = "<Q" if self.e_ident_data == self.ELFDATA2LSB else ">Q"
        format_p = format_i if self.e_ident_class == self.ELFCLASS32 else format_q
        self.e_type = unpack(format_h)
        self.e_machine = unpack(format_h)
        self.e_version = unpack(format_i)
        self.e_entry = unpack(format_p)
        self.e_phoff = unpack(format_p)
        self.e_shoff = unpack(format_p)
        self.e_flags = unpack(format_i)
        self.e_ehsize = unpack(format_h)
        self.e_phentsize = unpack(format_h)
        self.e_phnum = unpack(format_h)
        self.e_shentsize = unpack(format_h)
        self.e_shnum = unpack(format_h)
        self.e_shstrndx = unpack(format_h)


def _get_elf_header() -> Optional[_ELFFileHeader]:
    try:
        with open(sys.executable, "rb") as f:
            elf_header = _ELFFileHeader(f)
    except (OSError, TypeError, _ELFFileHeader._InvalidELFFileHeader):
        return None
    return elf_header


def _is_linux_armhf() -> bool:
    # hard-float ABI can be detected from the ELF header of the running
    # process
    # https://static.docs.arm.com/ihi0044/g/aaelf32.pdf
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_ARM
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABIMASK
    ) == elf_header.EF_ARM_ABI_VER5
    result &= (
        elf_header.e_flags & elf_header.EF_ARM_ABI_FLOAT_HARD
    ) == elf_header.EF_ARM_ABI_FLOAT_HARD
    return result


def _is_linux_i686() -> bool:
    elf_header = _get_elf_header()
    if elf_header is None:
        return False
    result = elf_header.e_ident_class == elf_header.ELFCLASS32
    result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
    result &= elf_header.e_machine == elf_header.EM_386
    return result


def _have_compatible_abi(arch: str) -> bool:
    if arch == "armv7l":
        return _is_linux_armhf()
    if arch == "i686":
        return _is_linux_i686()
    return arch in {"x86_64", "aarch64", "ppc64", "ppc64le", "s390x"}


# If glibc ever changes its major version, we need to know what the last
# minor version was, so we can build the complete list of all versions.
# For now, guess what the highest minor version might be, assume it will
# be 50 for testing. Once this actually happens, update the dictionary
# with the actual value.
_LAST_GLIBC_MINOR: Dict[int, int] = collections.defaultdict(lambda: 50)


class _GLibCVersion(NamedTuple):
    major: int
    minor: int


def _glibc_version_string_confstr() -> Optional[str]:
    """
    Primary implementation of glibc_version_string using os.confstr.
    """
    # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely
    # to be broken or missing. This strategy is used in the standard library
    # platform module.
    # https://github.com/python/cpython/blob/fcf1d003bf4f0100c/Lib/platform.py#L175-L183
    try:
        # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17".
        version_string = os.confstr("CS_GNU_LIBC_VERSION")
        assert version_string is not None
        _, version = version_string.split()
    except (AssertionError, AttributeError, OSError, ValueError):
        # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)...
        return None
    return version


def _glibc_version_string_ctypes() -> Optional[str]:
    """
    Fallback implementation of glibc_version_string using ctypes.
    """
    try:
        import ctypes
    except ImportError:
        return None

    # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen
    # manpage says, "If filename is NULL, then the returned handle is for the
    # main program". This way we can let the linker do the work to figure out
    # which libc our process is actually using.
    #
    # We must also handle the special case where the executable is not a
    # dynamically linked executable. This can occur when using musl libc,
    # for example. In this situation, dlopen() will error, leading to an
    # OSError. Interestingly, at least in the case of musl, there is no
    # errno set on the OSError. The single string argument used to construct
    # OSError comes from libc itself and is therefore not portable to
    # hard code here. In any case, failure to call dlopen() means we
    # can proceed, so we bail on our attempt.
    try:
        process_namespace = ctypes.CDLL(None)
    except OSError:
        return None

    try:
        gnu_get_libc_version = process_namespace.gnu_get_libc_version
    except AttributeError:
        # Symbol doesn't exist -> therefore, we are not linked to
        # glibc.
        return None

    # Call gnu_get_libc_version, which returns a string like "2.5"
    gnu_get_libc_version.restype = ctypes.c_char_p
    version_str: str = gnu_get_libc_version()
    # py2 / py3 compatibility:
    if not isinstance(version_str, str):
        version_str = version_str.decode("ascii")

    return version_str


def _glibc_version_string() -> Optional[str]:
    """Returns glibc version string, or None if not using glibc."""
    return _glibc_version_string_confstr() or _glibc_version_string_ctypes()


def _parse_glibc_version(version_str: str) -> Tuple[int, int]:
    """Parse glibc version.

    We use a regexp instead of str.split because we want to discard any
    random junk that might come after the minor version -- this might happen
    in patched/forked versions of glibc (e.g. Linaro's version of glibc
    uses version strings like "2.20-2014.11"). See gh-3588.
    """
    m = re.match(r"(?P<major>[0-9]+)\.(?P<minor>[0-9]+)", version_str)
    if not m:
        warnings.warn(
            "Expected glibc version with 2 components major.minor,"
            " got: %s" % version_str,
            RuntimeWarning,
        )
        return -1, -1
    return int(m.group("major")), int(m.group("minor"))


@functools.lru_cache()
def _get_glibc_version() -> Tuple[int, int]:
    version_str = _glibc_version_string()
    if version_str is None:
        return (-1, -1)
    return _parse_glibc_version(version_str)


# From PEP 513, PEP 600
def _is_compatible(name: str, arch: str, version: _GLibCVersion) -> bool:
    sys_glibc = _get_glibc_version()
    if sys_glibc < version:
        return False
    # Check for presence of _manylinux module.
    try:
        import _manylinux  # noqa
    except ImportError:
        return True
    if hasattr(_manylinux, "manylinux_compatible"):
        result = _manylinux.manylinux_compatible(version[0], version[1], arch)
        if result is not None:
            return bool(result)
        return True
    if version == _GLibCVersion(2, 5):
        if hasattr(_manylinux, "manylinux1_compatible"):
            return bool(_manylinux.manylinux1_compatible)
    if version == _GLibCVersion(2, 12):
        if hasattr(_manylinux, "manylinux2010_compatible"):
            return bool(_manylinux.manylinux2010_compatible)
    if version == _GLibCVersion(2, 17):
        if hasattr(_manylinux, "manylinux2014_compatible"):
            return bool(_manylinux.manylinux2014_compatible)
    return True


_LEGACY_MANYLINUX_MAP = {
    # CentOS 7 w/ glibc 2.17 (PEP 599)
    (2, 17): "manylinux2014",
    # CentOS 6 w/ glibc 2.12 (PEP 571)
    (2, 12): "manylinux2010",
    # CentOS 5 w/ glibc 2.5 (PEP 513)
    (2, 5): "manylinux1",
}


def platform_tags(linux: str, arch: str) -> Iterator[str]:
    if not _have_compatible_abi(arch):
        return
    # Oldest glibc to be supported regardless of architecture is (2, 17).
    too_old_glibc2 = _GLibCVersion(2, 16)
    if arch in {"x86_64", "i686"}:
        # On x86/i686 also oldest glibc to be supported is (2, 5).
        too_old_glibc2 = _GLibCVersion(2, 4)
    current_glibc = _GLibCVersion(*_get_glibc_version())
    glibc_max_list = [current_glibc]
    # We can assume compatibility across glibc major versions.
    # https://sourceware.org/bugzilla/show_bug.cgi?id=24636
    #
    # Build a list of maximum glibc versions so that we can
    # output the canonical list of all glibc from current_glibc
    # down to too_old_glibc2, including all intermediary versions.
    for glibc_major in range(current_glibc.major - 1, 1, -1):
        glibc_minor = _LAST_GLIBC_MINOR[glibc_major]
        glibc_max_list.append(_GLibCVersion(glibc_major, glibc_minor))
    for glibc_max in glibc_max_list:
        if glibc_max.major == too_old_glibc2.major:
            min_minor = too_old_glibc2.minor
        else:
            # For other glibc major versions oldest supported is (x, 0).
            min_minor = -1
        for glibc_minor in range(glibc_max.minor, min_minor, -1):
            glibc_version = _GLibCVersion(glibc_max.major, glibc_minor)
            tag = "manylinux_{}_{}".format(*glibc_version)
            if _is_compatible(tag, arch, glibc_version):
                yield linux.replace("linux", tag)
            # Handle the legacy manylinux1, manylinux2010, manylinux2014 tags.
            if glibc_version in _LEGACY_MANYLINUX_MAP:
                legacy_tag = _LEGACY_MANYLINUX_MAP[glibc_version]
                if _is_compatible(legacy_tag, arch, glibc_version):
                    yield linux.replace("linux", legacy_tag)
site-packages/pip/_vendor/packaging/markers.py000064400000020447150564745010015470 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import operator
import os
import platform
import sys
from typing import Any, Callable, Dict, List, Optional, Tuple, Union

from pip._vendor.pyparsing import (  # noqa: N817
    Forward,
    Group,
    Literal as L,
    ParseException,
    ParseResults,
    QuotedString,
    ZeroOrMore,
    stringEnd,
    stringStart,
)

from .specifiers import InvalidSpecifier, Specifier

__all__ = [
    "InvalidMarker",
    "UndefinedComparison",
    "UndefinedEnvironmentName",
    "Marker",
    "default_environment",
]

Operator = Callable[[str, str], bool]


class InvalidMarker(ValueError):
    """
    An invalid marker was found, users should refer to PEP 508.
    """


class UndefinedComparison(ValueError):
    """
    An invalid operation was attempted on a value that doesn't support it.
    """


class UndefinedEnvironmentName(ValueError):
    """
    A name was attempted to be used that does not exist inside of the
    environment.
    """


class Node:
    def __init__(self, value: Any) -> None:
        self.value = value

    def __str__(self) -> str:
        return str(self.value)

    def __repr__(self) -> str:
        return f"<{self.__class__.__name__}('{self}')>"

    def serialize(self) -> str:
        raise NotImplementedError


class Variable(Node):
    def serialize(self) -> str:
        return str(self)


class Value(Node):
    def serialize(self) -> str:
        return f'"{self}"'


class Op(Node):
    def serialize(self) -> str:
        return str(self)


VARIABLE = (
    L("implementation_version")
    | L("platform_python_implementation")
    | L("implementation_name")
    | L("python_full_version")
    | L("platform_release")
    | L("platform_version")
    | L("platform_machine")
    | L("platform_system")
    | L("python_version")
    | L("sys_platform")
    | L("os_name")
    | L("os.name")  # PEP-345
    | L("sys.platform")  # PEP-345
    | L("platform.version")  # PEP-345
    | L("platform.machine")  # PEP-345
    | L("platform.python_implementation")  # PEP-345
    | L("python_implementation")  # undocumented setuptools legacy
    | L("extra")  # PEP-508
)
ALIASES = {
    "os.name": "os_name",
    "sys.platform": "sys_platform",
    "platform.version": "platform_version",
    "platform.machine": "platform_machine",
    "platform.python_implementation": "platform_python_implementation",
    "python_implementation": "platform_python_implementation",
}
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))

VERSION_CMP = (
    L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<")
)

MARKER_OP = VERSION_CMP | L("not in") | L("in")
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))

MARKER_VALUE = QuotedString("'") | QuotedString('"')
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))

BOOLOP = L("and") | L("or")

MARKER_VAR = VARIABLE | MARKER_VALUE

MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))

LPAREN = L("(").suppress()
RPAREN = L(")").suppress()

MARKER_EXPR = Forward()
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)

MARKER = stringStart + MARKER_EXPR + stringEnd


def _coerce_parse_result(results: Union[ParseResults, List[Any]]) -> List[Any]:
    if isinstance(results, ParseResults):
        return [_coerce_parse_result(i) for i in results]
    else:
        return results


def _format_marker(
    marker: Union[List[str], Tuple[Node, ...], str], first: Optional[bool] = True
) -> str:

    assert isinstance(marker, (list, tuple, str))

    # Sometimes we have a structure like [[...]] which is a single item list
    # where the single item is itself it's own list. In that case we want skip
    # the rest of this function so that we don't get extraneous () on the
    # outside.
    if (
        isinstance(marker, list)
        and len(marker) == 1
        and isinstance(marker[0], (list, tuple))
    ):
        return _format_marker(marker[0])

    if isinstance(marker, list):
        inner = (_format_marker(m, first=False) for m in marker)
        if first:
            return " ".join(inner)
        else:
            return "(" + " ".join(inner) + ")"
    elif isinstance(marker, tuple):
        return " ".join([m.serialize() for m in marker])
    else:
        return marker


_operators: Dict[str, Operator] = {
    "in": lambda lhs, rhs: lhs in rhs,
    "not in": lambda lhs, rhs: lhs not in rhs,
    "<": operator.lt,
    "<=": operator.le,
    "==": operator.eq,
    "!=": operator.ne,
    ">=": operator.ge,
    ">": operator.gt,
}


def _eval_op(lhs: str, op: Op, rhs: str) -> bool:
    try:
        spec = Specifier("".join([op.serialize(), rhs]))
    except InvalidSpecifier:
        pass
    else:
        return spec.contains(lhs)

    oper: Optional[Operator] = _operators.get(op.serialize())
    if oper is None:
        raise UndefinedComparison(f"Undefined {op!r} on {lhs!r} and {rhs!r}.")

    return oper(lhs, rhs)


class Undefined:
    pass


_undefined = Undefined()


def _get_env(environment: Dict[str, str], name: str) -> str:
    value: Union[str, Undefined] = environment.get(name, _undefined)

    if isinstance(value, Undefined):
        raise UndefinedEnvironmentName(
            f"{name!r} does not exist in evaluation environment."
        )

    return value


def _evaluate_markers(markers: List[Any], environment: Dict[str, str]) -> bool:
    groups: List[List[bool]] = [[]]

    for marker in markers:
        assert isinstance(marker, (list, tuple, str))

        if isinstance(marker, list):
            groups[-1].append(_evaluate_markers(marker, environment))
        elif isinstance(marker, tuple):
            lhs, op, rhs = marker

            if isinstance(lhs, Variable):
                lhs_value = _get_env(environment, lhs.value)
                rhs_value = rhs.value
            else:
                lhs_value = lhs.value
                rhs_value = _get_env(environment, rhs.value)

            groups[-1].append(_eval_op(lhs_value, op, rhs_value))
        else:
            assert marker in ["and", "or"]
            if marker == "or":
                groups.append([])

    return any(all(item) for item in groups)


def format_full_version(info: "sys._version_info") -> str:
    version = "{0.major}.{0.minor}.{0.micro}".format(info)
    kind = info.releaselevel
    if kind != "final":
        version += kind[0] + str(info.serial)
    return version


def default_environment() -> Dict[str, str]:
    iver = format_full_version(sys.implementation.version)
    implementation_name = sys.implementation.name
    return {
        "implementation_name": implementation_name,
        "implementation_version": iver,
        "os_name": os.name,
        "platform_machine": platform.machine(),
        "platform_release": platform.release(),
        "platform_system": platform.system(),
        "platform_version": platform.version(),
        "python_full_version": platform.python_version(),
        "platform_python_implementation": platform.python_implementation(),
        "python_version": ".".join(platform.python_version_tuple()[:2]),
        "sys_platform": sys.platform,
    }


class Marker:
    def __init__(self, marker: str) -> None:
        try:
            self._markers = _coerce_parse_result(MARKER.parseString(marker))
        except ParseException as e:
            raise InvalidMarker(
                f"Invalid marker: {marker!r}, parse error at "
                f"{marker[e.loc : e.loc + 8]!r}"
            )

    def __str__(self) -> str:
        return _format_marker(self._markers)

    def __repr__(self) -> str:
        return f"<Marker('{self}')>"

    def evaluate(self, environment: Optional[Dict[str, str]] = None) -> bool:
        """Evaluate a marker.

        Return the boolean from evaluating the given marker against the
        environment. environment is an optional argument to override all or
        part of the determined environment.

        The environment is determined from the current Python process.
        """
        current_environment = default_environment()
        if environment is not None:
            current_environment.update(environment)

        return _evaluate_markers(self._markers, current_environment)
site-packages/pip/_vendor/packaging/__about__.py000064400000001225150564745010015723 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]

__title__ = "packaging"
__summary__ = "Core utilities for Python packages"
__uri__ = "https://github.com/pypa/packaging"

__version__ = "21.0"

__author__ = "Donald Stufft and individual contributors"
__email__ = "donald@stufft.io"

__license__ = "BSD-2-Clause or Apache-2.0"
__copyright__ = "2014-2019 %s" % __author__
site-packages/pip/_vendor/packaging/utils.py000064400000010150150564745010015152 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import re
from typing import FrozenSet, NewType, Tuple, Union, cast

from .tags import Tag, parse_tag
from .version import InvalidVersion, Version

BuildTag = Union[Tuple[()], Tuple[int, str]]
NormalizedName = NewType("NormalizedName", str)


class InvalidWheelFilename(ValueError):
    """
    An invalid wheel filename was found, users should refer to PEP 427.
    """


class InvalidSdistFilename(ValueError):
    """
    An invalid sdist filename was found, users should refer to the packaging user guide.
    """


_canonicalize_regex = re.compile(r"[-_.]+")
# PEP 427: The build number must start with a digit.
_build_tag_regex = re.compile(r"(\d+)(.*)")


def canonicalize_name(name: str) -> NormalizedName:
    # This is taken from PEP 503.
    value = _canonicalize_regex.sub("-", name).lower()
    return cast(NormalizedName, value)


def canonicalize_version(version: Union[Version, str]) -> str:
    """
    This is very similar to Version.__str__, but has one subtle difference
    with the way it handles the release segment.
    """
    if isinstance(version, str):
        try:
            parsed = Version(version)
        except InvalidVersion:
            # Legacy versions cannot be normalized
            return version
    else:
        parsed = version

    parts = []

    # Epoch
    if parsed.epoch != 0:
        parts.append(f"{parsed.epoch}!")

    # Release segment
    # NB: This strips trailing '.0's to normalize
    parts.append(re.sub(r"(\.0)+$", "", ".".join(str(x) for x in parsed.release)))

    # Pre-release
    if parsed.pre is not None:
        parts.append("".join(str(x) for x in parsed.pre))

    # Post-release
    if parsed.post is not None:
        parts.append(f".post{parsed.post}")

    # Development release
    if parsed.dev is not None:
        parts.append(f".dev{parsed.dev}")

    # Local version segment
    if parsed.local is not None:
        parts.append(f"+{parsed.local}")

    return "".join(parts)


def parse_wheel_filename(
    filename: str,
) -> Tuple[NormalizedName, Version, BuildTag, FrozenSet[Tag]]:
    if not filename.endswith(".whl"):
        raise InvalidWheelFilename(
            f"Invalid wheel filename (extension must be '.whl'): {filename}"
        )

    filename = filename[:-4]
    dashes = filename.count("-")
    if dashes not in (4, 5):
        raise InvalidWheelFilename(
            f"Invalid wheel filename (wrong number of parts): {filename}"
        )

    parts = filename.split("-", dashes - 2)
    name_part = parts[0]
    # See PEP 427 for the rules on escaping the project name
    if "__" in name_part or re.match(r"^[\w\d._]*$", name_part, re.UNICODE) is None:
        raise InvalidWheelFilename(f"Invalid project name: {filename}")
    name = canonicalize_name(name_part)
    version = Version(parts[1])
    if dashes == 5:
        build_part = parts[2]
        build_match = _build_tag_regex.match(build_part)
        if build_match is None:
            raise InvalidWheelFilename(
                f"Invalid build number: {build_part} in '{filename}'"
            )
        build = cast(BuildTag, (int(build_match.group(1)), build_match.group(2)))
    else:
        build = ()
    tags = parse_tag(parts[-1])
    return (name, version, build, tags)


def parse_sdist_filename(filename: str) -> Tuple[NormalizedName, Version]:
    if filename.endswith(".tar.gz"):
        file_stem = filename[: -len(".tar.gz")]
    elif filename.endswith(".zip"):
        file_stem = filename[: -len(".zip")]
    else:
        raise InvalidSdistFilename(
            f"Invalid sdist filename (extension must be '.tar.gz' or '.zip'):"
            f" {filename}"
        )

    # We are requiring a PEP 440 version, which cannot contain dashes,
    # so we split on the last dash.
    name_part, sep, version_part = file_stem.rpartition("-")
    if not sep:
        raise InvalidSdistFilename(f"Invalid sdist filename: {filename}")

    name = canonicalize_name(name_part)
    version = Version(version_part)
    return (name, version)
site-packages/pip/_vendor/packaging/_musllinux.py000064400000010432150564745010016214 0ustar00"""PEP 656 support.

This module implements logic to detect if the currently running Python is
linked against musl, and what musl version is used.
"""

import contextlib
import functools
import operator
import os
import re
import struct
import subprocess
import sys
from typing import IO, Iterator, NamedTuple, Optional, Tuple


def _read_unpacked(f: IO[bytes], fmt: str) -> Tuple[int, ...]:
    return struct.unpack(fmt, f.read(struct.calcsize(fmt)))


def _parse_ld_musl_from_elf(f: IO[bytes]) -> Optional[str]:
    """Detect musl libc location by parsing the Python executable.

    Based on: https://gist.github.com/lyssdod/f51579ae8d93c8657a5564aefc2ffbca
    ELF header: https://refspecs.linuxfoundation.org/elf/gabi4+/ch4.eheader.html
    """
    f.seek(0)
    try:
        ident = _read_unpacked(f, "16B")
    except struct.error:
        return None
    if ident[:4] != tuple(b"\x7fELF"):  # Invalid magic, not ELF.
        return None
    f.seek(struct.calcsize("HHI"), 1)  # Skip file type, machine, and version.

    try:
        # e_fmt: Format for program header.
        # p_fmt: Format for section header.
        # p_idx: Indexes to find p_type, p_offset, and p_filesz.
        e_fmt, p_fmt, p_idx = {
            1: ("IIIIHHH", "IIIIIIII", (0, 1, 4)),  # 32-bit.
            2: ("QQQIHHH", "IIQQQQQQ", (0, 2, 5)),  # 64-bit.
        }[ident[4]]
    except KeyError:
        return None
    else:
        p_get = operator.itemgetter(*p_idx)

    # Find the interpreter section and return its content.
    try:
        _, e_phoff, _, _, _, e_phentsize, e_phnum = _read_unpacked(f, e_fmt)
    except struct.error:
        return None
    for i in range(e_phnum + 1):
        f.seek(e_phoff + e_phentsize * i)
        try:
            p_type, p_offset, p_filesz = p_get(_read_unpacked(f, p_fmt))
        except struct.error:
            return None
        if p_type != 3:  # Not PT_INTERP.
            continue
        f.seek(p_offset)
        interpreter = os.fsdecode(f.read(p_filesz)).strip("\0")
        if "musl" not in interpreter:
            return None
        return interpreter
    return None


class _MuslVersion(NamedTuple):
    major: int
    minor: int


def _parse_musl_version(output: str) -> Optional[_MuslVersion]:
    lines = [n for n in (n.strip() for n in output.splitlines()) if n]
    if len(lines) < 2 or lines[0][:4] != "musl":
        return None
    m = re.match(r"Version (\d+)\.(\d+)", lines[1])
    if not m:
        return None
    return _MuslVersion(major=int(m.group(1)), minor=int(m.group(2)))


@functools.lru_cache()
def _get_musl_version(executable: str) -> Optional[_MuslVersion]:
    """Detect currently-running musl runtime version.

    This is done by checking the specified executable's dynamic linking
    information, and invoking the loader to parse its output for a version
    string. If the loader is musl, the output would be something like::

        musl libc (x86_64)
        Version 1.2.2
        Dynamic Program Loader
    """
    with contextlib.ExitStack() as stack:
        try:
            f = stack.enter_context(open(executable, "rb"))
        except IOError:
            return None
        ld = _parse_ld_musl_from_elf(f)
    if not ld:
        return None
    proc = subprocess.run([ld], stderr=subprocess.PIPE, universal_newlines=True)
    return _parse_musl_version(proc.stderr)


def platform_tags(arch: str) -> Iterator[str]:
    """Generate musllinux tags compatible to the current platform.

    :param arch: Should be the part of platform tag after the ``linux_``
        prefix, e.g. ``x86_64``. The ``linux_`` prefix is assumed as a
        prerequisite for the current platform to be musllinux-compatible.

    :returns: An iterator of compatible musllinux tags.
    """
    sys_musl = _get_musl_version(sys.executable)
    if sys_musl is None:  # Python not dynamically linked against musl.
        return
    for minor in range(sys_musl.minor, -1, -1):
        yield f"musllinux_{sys_musl.major}_{minor}_{arch}"


if __name__ == "__main__":  # pragma: no cover
    import sysconfig

    plat = sysconfig.get_platform()
    assert plat.startswith("linux-"), "not linux"

    print("plat:", plat)
    print("musl:", _get_musl_version(sys.executable))
    print("tags:", end=" ")
    for t in platform_tags(re.sub(r"[.-]", "_", plat.split("-", 1)[-1])):
        print(t, end="\n      ")
site-packages/pip/_vendor/packaging/version.py000064400000034217150564745010015511 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import collections
import itertools
import re
import warnings
from typing import Callable, Iterator, List, Optional, SupportsInt, Tuple, Union

from ._structures import Infinity, InfinityType, NegativeInfinity, NegativeInfinityType

__all__ = ["parse", "Version", "LegacyVersion", "InvalidVersion", "VERSION_PATTERN"]

InfiniteTypes = Union[InfinityType, NegativeInfinityType]
PrePostDevType = Union[InfiniteTypes, Tuple[str, int]]
SubLocalType = Union[InfiniteTypes, int, str]
LocalType = Union[
    NegativeInfinityType,
    Tuple[
        Union[
            SubLocalType,
            Tuple[SubLocalType, str],
            Tuple[NegativeInfinityType, SubLocalType],
        ],
        ...,
    ],
]
CmpKey = Tuple[
    int, Tuple[int, ...], PrePostDevType, PrePostDevType, PrePostDevType, LocalType
]
LegacyCmpKey = Tuple[int, Tuple[str, ...]]
VersionComparisonMethod = Callable[
    [Union[CmpKey, LegacyCmpKey], Union[CmpKey, LegacyCmpKey]], bool
]

_Version = collections.namedtuple(
    "_Version", ["epoch", "release", "dev", "pre", "post", "local"]
)


def parse(version: str) -> Union["LegacyVersion", "Version"]:
    """
    Parse the given version string and return either a :class:`Version` object
    or a :class:`LegacyVersion` object depending on if the given version is
    a valid PEP 440 version or a legacy version.
    """
    try:
        return Version(version)
    except InvalidVersion:
        return LegacyVersion(version)


class InvalidVersion(ValueError):
    """
    An invalid version was found, users should refer to PEP 440.
    """


class _BaseVersion:
    _key: Union[CmpKey, LegacyCmpKey]

    def __hash__(self) -> int:
        return hash(self._key)

    # Please keep the duplicated `isinstance` check
    # in the six comparisons hereunder
    # unless you find a way to avoid adding overhead function calls.
    def __lt__(self, other: "_BaseVersion") -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key < other._key

    def __le__(self, other: "_BaseVersion") -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key <= other._key

    def __eq__(self, other: object) -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key == other._key

    def __ge__(self, other: "_BaseVersion") -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key >= other._key

    def __gt__(self, other: "_BaseVersion") -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key > other._key

    def __ne__(self, other: object) -> bool:
        if not isinstance(other, _BaseVersion):
            return NotImplemented

        return self._key != other._key


class LegacyVersion(_BaseVersion):
    def __init__(self, version: str) -> None:
        self._version = str(version)
        self._key = _legacy_cmpkey(self._version)


    def __str__(self) -> str:
        return self._version

    def __repr__(self) -> str:
        return f"<LegacyVersion('{self}')>"

    @property
    def public(self) -> str:
        return self._version

    @property
    def base_version(self) -> str:
        return self._version

    @property
    def epoch(self) -> int:
        return -1

    @property
    def release(self) -> None:
        return None

    @property
    def pre(self) -> None:
        return None

    @property
    def post(self) -> None:
        return None

    @property
    def dev(self) -> None:
        return None

    @property
    def local(self) -> None:
        return None

    @property
    def is_prerelease(self) -> bool:
        return False

    @property
    def is_postrelease(self) -> bool:
        return False

    @property
    def is_devrelease(self) -> bool:
        return False


_legacy_version_component_re = re.compile(r"(\d+ | [a-z]+ | \.| -)", re.VERBOSE)

_legacy_version_replacement_map = {
    "pre": "c",
    "preview": "c",
    "-": "final-",
    "rc": "c",
    "dev": "@",
}


def _parse_version_parts(s: str) -> Iterator[str]:
    for part in _legacy_version_component_re.split(s):
        part = _legacy_version_replacement_map.get(part, part)

        if not part or part == ".":
            continue

        if part[:1] in "0123456789":
            # pad for numeric comparison
            yield part.zfill(8)
        else:
            yield "*" + part

    # ensure that alpha/beta/candidate are before final
    yield "*final"


def _legacy_cmpkey(version: str) -> LegacyCmpKey:

    # We hardcode an epoch of -1 here. A PEP 440 version can only have a epoch
    # greater than or equal to 0. This will effectively put the LegacyVersion,
    # which uses the defacto standard originally implemented by setuptools,
    # as before all PEP 440 versions.
    epoch = -1

    # This scheme is taken from pkg_resources.parse_version setuptools prior to
    # it's adoption of the packaging library.
    parts: List[str] = []
    for part in _parse_version_parts(version.lower()):
        if part.startswith("*"):
            # remove "-" before a prerelease tag
            if part < "*final":
                while parts and parts[-1] == "*final-":
                    parts.pop()

            # remove trailing zeros from each series of numeric parts
            while parts and parts[-1] == "00000000":
                parts.pop()

        parts.append(part)

    return epoch, tuple(parts)


# Deliberately not anchored to the start and end of the string, to make it
# easier for 3rd party code to reuse
VERSION_PATTERN = r"""
    v?
    (?:
        (?:(?P<epoch>[0-9]+)!)?                           # epoch
        (?P<release>[0-9]+(?:\.[0-9]+)*)                  # release segment
        (?P<pre>                                          # pre-release
            [-_\.]?
            (?P<pre_l>(a|b|c|rc|alpha|beta|pre|preview))
            [-_\.]?
            (?P<pre_n>[0-9]+)?
        )?
        (?P<post>                                         # post release
            (?:-(?P<post_n1>[0-9]+))
            |
            (?:
                [-_\.]?
                (?P<post_l>post|rev|r)
                [-_\.]?
                (?P<post_n2>[0-9]+)?
            )
        )?
        (?P<dev>                                          # dev release
            [-_\.]?
            (?P<dev_l>dev)
            [-_\.]?
            (?P<dev_n>[0-9]+)?
        )?
    )
    (?:\+(?P<local>[a-z0-9]+(?:[-_\.][a-z0-9]+)*))?       # local version
"""


class Version(_BaseVersion):

    _regex = re.compile(r"^\s*" + VERSION_PATTERN + r"\s*$", re.VERBOSE | re.IGNORECASE)

    def __init__(self, version: str) -> None:

        # Validate the version and parse it into pieces
        match = self._regex.search(version)
        if not match:
            raise InvalidVersion(f"Invalid version: '{version}'")

        # Store the parsed out pieces of the version
        self._version = _Version(
            epoch=int(match.group("epoch")) if match.group("epoch") else 0,
            release=tuple(int(i) for i in match.group("release").split(".")),
            pre=_parse_letter_version(match.group("pre_l"), match.group("pre_n")),
            post=_parse_letter_version(
                match.group("post_l"), match.group("post_n1") or match.group("post_n2")
            ),
            dev=_parse_letter_version(match.group("dev_l"), match.group("dev_n")),
            local=_parse_local_version(match.group("local")),
        )

        # Generate a key which will be used for sorting
        self._key = _cmpkey(
            self._version.epoch,
            self._version.release,
            self._version.pre,
            self._version.post,
            self._version.dev,
            self._version.local,
        )

    def __repr__(self) -> str:
        return f"<Version('{self}')>"

    def __str__(self) -> str:
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append(f"{self.epoch}!")

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        # Pre-release
        if self.pre is not None:
            parts.append("".join(str(x) for x in self.pre))

        # Post-release
        if self.post is not None:
            parts.append(f".post{self.post}")

        # Development release
        if self.dev is not None:
            parts.append(f".dev{self.dev}")

        # Local version segment
        if self.local is not None:
            parts.append(f"+{self.local}")

        return "".join(parts)

    @property
    def epoch(self) -> int:
        _epoch: int = self._version.epoch
        return _epoch

    @property
    def release(self) -> Tuple[int, ...]:
        _release: Tuple[int, ...] = self._version.release
        return _release

    @property
    def pre(self) -> Optional[Tuple[str, int]]:
        _pre: Optional[Tuple[str, int]] = self._version.pre
        return _pre

    @property
    def post(self) -> Optional[int]:
        return self._version.post[1] if self._version.post else None

    @property
    def dev(self) -> Optional[int]:
        return self._version.dev[1] if self._version.dev else None

    @property
    def local(self) -> Optional[str]:
        if self._version.local:
            return ".".join(str(x) for x in self._version.local)
        else:
            return None

    @property
    def public(self) -> str:
        return str(self).split("+", 1)[0]

    @property
    def base_version(self) -> str:
        parts = []

        # Epoch
        if self.epoch != 0:
            parts.append(f"{self.epoch}!")

        # Release segment
        parts.append(".".join(str(x) for x in self.release))

        return "".join(parts)

    @property
    def is_prerelease(self) -> bool:
        return self.dev is not None or self.pre is not None

    @property
    def is_postrelease(self) -> bool:
        return self.post is not None

    @property
    def is_devrelease(self) -> bool:
        return self.dev is not None

    @property
    def major(self) -> int:
        return self.release[0] if len(self.release) >= 1 else 0

    @property
    def minor(self) -> int:
        return self.release[1] if len(self.release) >= 2 else 0

    @property
    def micro(self) -> int:
        return self.release[2] if len(self.release) >= 3 else 0


def _parse_letter_version(
    letter: str, number: Union[str, bytes, SupportsInt]
) -> Optional[Tuple[str, int]]:

    if letter:
        # We consider there to be an implicit 0 in a pre-release if there is
        # not a numeral associated with it.
        if number is None:
            number = 0

        # We normalize any letters to their lower case form
        letter = letter.lower()

        # We consider some words to be alternate spellings of other words and
        # in those cases we want to normalize the spellings to our preferred
        # spelling.
        if letter == "alpha":
            letter = "a"
        elif letter == "beta":
            letter = "b"
        elif letter in ["c", "pre", "preview"]:
            letter = "rc"
        elif letter in ["rev", "r"]:
            letter = "post"

        return letter, int(number)
    if not letter and number:
        # We assume if we are given a number, but we are not given a letter
        # then this is using the implicit post release syntax (e.g. 1.0-1)
        letter = "post"

        return letter, int(number)

    return None


_local_version_separators = re.compile(r"[\._-]")


def _parse_local_version(local: str) -> Optional[LocalType]:
    """
    Takes a string like abc.1.twelve and turns it into ("abc", 1, "twelve").
    """
    if local is not None:
        return tuple(
            part.lower() if not part.isdigit() else int(part)
            for part in _local_version_separators.split(local)
        )
    return None


def _cmpkey(
    epoch: int,
    release: Tuple[int, ...],
    pre: Optional[Tuple[str, int]],
    post: Optional[Tuple[str, int]],
    dev: Optional[Tuple[str, int]],
    local: Optional[Tuple[SubLocalType]],
) -> CmpKey:

    # When we compare a release version, we want to compare it with all of the
    # trailing zeros removed. So we'll use a reverse the list, drop all the now
    # leading zeros until we come to something non zero, then take the rest
    # re-reverse it back into the correct order and make it a tuple and use
    # that for our sorting key.
    _release = tuple(
        reversed(list(itertools.dropwhile(lambda x: x == 0, reversed(release))))
    )

    # We need to "trick" the sorting algorithm to put 1.0.dev0 before 1.0a0.
    # We'll do this by abusing the pre segment, but we _only_ want to do this
    # if there is not a pre or a post segment. If we have one of those then
    # the normal sorting rules will handle this case correctly.
    if pre is None and post is None and dev is not None:
        _pre: PrePostDevType = NegativeInfinity
    # Versions without a pre-release (except as noted above) should sort after
    # those with one.
    elif pre is None:
        _pre = Infinity
    else:
        _pre = pre

    # Versions without a post segment should sort before those with one.
    if post is None:
        _post: PrePostDevType = NegativeInfinity

    else:
        _post = post

    # Versions without a development segment should sort after those with one.
    if dev is None:
        _dev: PrePostDevType = Infinity

    else:
        _dev = dev

    if local is None:
        # Versions without a local segment should sort before those with one.
        _local: LocalType = NegativeInfinity
    else:
        # Versions with a local segment need that segment parsed to implement
        # the sorting rules in PEP440.
        # - Alpha numeric segments sort before numeric segments
        # - Alpha numeric segments sort lexicographically
        # - Numeric segments sort numerically
        # - Shorter versions sort before longer versions when the prefixes
        #   match exactly
        _local = tuple(
            (i, "") if isinstance(i, int) else (NegativeInfinity, i) for i in local
        )

    return epoch, _release, _pre, _post, _dev, _local
site-packages/pip/_vendor/packaging/__init__.py000064400000000761150564745010015560 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

from .__about__ import (
    __author__,
    __copyright__,
    __email__,
    __license__,
    __summary__,
    __title__,
    __uri__,
    __version__,
)

__all__ = [
    "__title__",
    "__summary__",
    "__uri__",
    "__version__",
    "__author__",
    "__email__",
    "__license__",
    "__copyright__",
]
site-packages/pip/_vendor/packaging/specifiers.py000064400000074364150564745010016167 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import abc
import functools
import itertools
import re
import warnings
from typing import (
    Callable,
    Dict,
    Iterable,
    Iterator,
    List,
    Optional,
    Pattern,
    Set,
    Tuple,
    TypeVar,
    Union,
)

from .utils import canonicalize_version
from .version import LegacyVersion, Version, parse

ParsedVersion = Union[Version, LegacyVersion]
UnparsedVersion = Union[Version, LegacyVersion, str]
VersionTypeVar = TypeVar("VersionTypeVar", bound=UnparsedVersion)
CallableOperator = Callable[[ParsedVersion, str], bool]


class InvalidSpecifier(ValueError):
    """
    An invalid specifier was found, users should refer to PEP 440.
    """


class BaseSpecifier(metaclass=abc.ABCMeta):
    @abc.abstractmethod
    def __str__(self) -> str:
        """
        Returns the str representation of this Specifier like object. This
        should be representative of the Specifier itself.
        """

    @abc.abstractmethod
    def __hash__(self) -> int:
        """
        Returns a hash value for this Specifier like object.
        """

    @abc.abstractmethod
    def __eq__(self, other: object) -> bool:
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are equal.
        """

    @abc.abstractmethod
    def __ne__(self, other: object) -> bool:
        """
        Returns a boolean representing whether or not the two Specifier like
        objects are not equal.
        """

    @abc.abstractproperty
    def prereleases(self) -> Optional[bool]:
        """
        Returns whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @prereleases.setter
    def prereleases(self, value: bool) -> None:
        """
        Sets whether or not pre-releases as a whole are allowed by this
        specifier.
        """

    @abc.abstractmethod
    def contains(self, item: str, prereleases: Optional[bool] = None) -> bool:
        """
        Determines if the given item is contained within this specifier.
        """

    @abc.abstractmethod
    def filter(
        self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
    ) -> Iterable[VersionTypeVar]:
        """
        Takes an iterable of items and filters them so that only items which
        are contained within this specifier are allowed in it.
        """


class _IndividualSpecifier(BaseSpecifier):

    _operators: Dict[str, str] = {}
    _regex: Pattern[str]

    def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None:
        match = self._regex.search(spec)
        if not match:
            raise InvalidSpecifier(f"Invalid specifier: '{spec}'")

        self._spec: Tuple[str, str] = (
            match.group("operator").strip(),
            match.group("version").strip(),
        )

        # Store whether or not this Specifier should accept prereleases
        self._prereleases = prereleases

    def __repr__(self) -> str:
        pre = (
            f", prereleases={self.prereleases!r}"
            if self._prereleases is not None
            else ""
        )

        return "<{}({!r}{})>".format(self.__class__.__name__, str(self), pre)

    def __str__(self) -> str:
        return "{}{}".format(*self._spec)

    @property
    def _canonical_spec(self) -> Tuple[str, str]:
        return self._spec[0], canonicalize_version(self._spec[1])

    def __hash__(self) -> int:
        return hash(self._canonical_spec)

    def __eq__(self, other: object) -> bool:
        if isinstance(other, str):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._canonical_spec == other._canonical_spec

    def __ne__(self, other: object) -> bool:
        if isinstance(other, str):
            try:
                other = self.__class__(str(other))
            except InvalidSpecifier:
                return NotImplemented
        elif not isinstance(other, self.__class__):
            return NotImplemented

        return self._spec != other._spec

    def _get_operator(self, op: str) -> CallableOperator:
        operator_callable: CallableOperator = getattr(
            self, f"_compare_{self._operators[op]}"
        )
        return operator_callable

    def _coerce_version(self, version: UnparsedVersion) -> ParsedVersion:
        if not isinstance(version, (LegacyVersion, Version)):
            version = parse(version)
        return version

    @property
    def operator(self) -> str:
        return self._spec[0]

    @property
    def version(self) -> str:
        return self._spec[1]

    @property
    def prereleases(self) -> Optional[bool]:
        return self._prereleases

    @prereleases.setter
    def prereleases(self, value: bool) -> None:
        self._prereleases = value

    def __contains__(self, item: str) -> bool:
        return self.contains(item)

    def contains(
        self, item: UnparsedVersion, prereleases: Optional[bool] = None
    ) -> bool:

        # Determine if prereleases are to be allowed or not.
        if prereleases is None:
            prereleases = self.prereleases

        # Normalize item to a Version or LegacyVersion, this allows us to have
        # a shortcut for ``"2.0" in Specifier(">=2")
        normalized_item = self._coerce_version(item)

        # Determine if we should be supporting prereleases in this specifier
        # or not, if we do not support prereleases than we can short circuit
        # logic if this version is a prereleases.
        if normalized_item.is_prerelease and not prereleases:
            return False

        # Actually do the comparison to determine if this item is contained
        # within this Specifier or not.
        operator_callable: CallableOperator = self._get_operator(self.operator)
        return operator_callable(normalized_item, self.version)

    def filter(
        self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
    ) -> Iterable[VersionTypeVar]:

        yielded = False
        found_prereleases = []

        kw = {"prereleases": prereleases if prereleases is not None else True}

        # Attempt to iterate over all the values in the iterable and if any of
        # them match, yield them.
        for version in iterable:
            parsed_version = self._coerce_version(version)

            if self.contains(parsed_version, **kw):
                # If our version is a prerelease, and we were not set to allow
                # prereleases, then we'll store it for later in case nothing
                # else matches this specifier.
                if parsed_version.is_prerelease and not (
                    prereleases or self.prereleases
                ):
                    found_prereleases.append(version)
                # Either this is not a prerelease, or we should have been
                # accepting prereleases from the beginning.
                else:
                    yielded = True
                    yield version

        # Now that we've iterated over everything, determine if we've yielded
        # any values, and if we have not and we have any prereleases stored up
        # then we will go ahead and yield the prereleases.
        if not yielded and found_prereleases:
            for version in found_prereleases:
                yield version


class LegacySpecifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(==|!=|<=|>=|<|>))
        \s*
        (?P<version>
            [^,;\s)]* # Since this is a "legacy" specifier, and the version
                      # string can be just about anything, we match everything
                      # except for whitespace, a semi-colon for marker support,
                      # a closing paren since versions can be enclosed in
                      # them, and a comma since it's a version separator.
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
    }

    def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None:
        super().__init__(spec, prereleases)

        warnings.warn(
            "Creating a LegacyVersion has been deprecated and will be "
            "removed in the next major release",
            DeprecationWarning,
        )

    def _coerce_version(self, version: UnparsedVersion) -> LegacyVersion:
        if not isinstance(version, LegacyVersion):
            version = LegacyVersion(str(version))
        return version

    def _compare_equal(self, prospective: LegacyVersion, spec: str) -> bool:
        return prospective == self._coerce_version(spec)

    def _compare_not_equal(self, prospective: LegacyVersion, spec: str) -> bool:
        return prospective != self._coerce_version(spec)

    def _compare_less_than_equal(self, prospective: LegacyVersion, spec: str) -> bool:
        return prospective <= self._coerce_version(spec)

    def _compare_greater_than_equal(
        self, prospective: LegacyVersion, spec: str
    ) -> bool:
        return prospective >= self._coerce_version(spec)

    def _compare_less_than(self, prospective: LegacyVersion, spec: str) -> bool:
        return prospective < self._coerce_version(spec)

    def _compare_greater_than(self, prospective: LegacyVersion, spec: str) -> bool:
        return prospective > self._coerce_version(spec)


def _require_version_compare(
    fn: Callable[["Specifier", ParsedVersion, str], bool]
) -> Callable[["Specifier", ParsedVersion, str], bool]:
    @functools.wraps(fn)
    def wrapped(self: "Specifier", prospective: ParsedVersion, spec: str) -> bool:
        if not isinstance(prospective, Version):
            return False
        return fn(self, prospective, spec)

    return wrapped


class Specifier(_IndividualSpecifier):

    _regex_str = r"""
        (?P<operator>(~=|==|!=|<=|>=|<|>|===))
        (?P<version>
            (?:
                # The identity operators allow for an escape hatch that will
                # do an exact string match of the version you wish to install.
                # This will not be parsed by PEP 440 and we cannot determine
                # any semantic meaning from it. This operator is discouraged
                # but included entirely as an escape hatch.
                (?<====)  # Only match for the identity operator
                \s*
                [^\s]*    # We just match everything, except for whitespace
                          # since we are only testing for strict identity.
            )
            |
            (?:
                # The (non)equality operators allow for wild card and local
                # versions to be specified so we have to define these two
                # operators separately to enable that.
                (?<===|!=)            # Only match for equals and not equals

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?

                # You cannot use a wild card and a dev or local version
                # together so group them with a | and make them optional.
                (?:
                    (?:[-_\.]?dev[-_\.]?[0-9]*)?         # dev release
                    (?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
                    |
                    \.\*  # Wild card syntax of .*
                )?
            )
            |
            (?:
                # The compatible operator requires at least two digits in the
                # release segment.
                (?<=~=)               # Only match for the compatible operator

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)+   # release  (We have a + instead of a *)
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
            |
            (?:
                # All other operators only allow a sub set of what the
                # (non)equality operators do. Specifically they do not allow
                # local versions to be specified nor do they allow the prefix
                # matching wild cards.
                (?<!==|!=|~=)         # We have special cases for these
                                      # operators so we want to make sure they
                                      # don't match here.

                \s*
                v?
                (?:[0-9]+!)?          # epoch
                [0-9]+(?:\.[0-9]+)*   # release
                (?:                   # pre release
                    [-_\.]?
                    (a|b|c|rc|alpha|beta|pre|preview)
                    [-_\.]?
                    [0-9]*
                )?
                (?:                                   # post release
                    (?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
                )?
                (?:[-_\.]?dev[-_\.]?[0-9]*)?          # dev release
            )
        )
        """

    _regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)

    _operators = {
        "~=": "compatible",
        "==": "equal",
        "!=": "not_equal",
        "<=": "less_than_equal",
        ">=": "greater_than_equal",
        "<": "less_than",
        ">": "greater_than",
        "===": "arbitrary",
    }

    @_require_version_compare
    def _compare_compatible(self, prospective: ParsedVersion, spec: str) -> bool:

        # Compatible releases have an equivalent combination of >= and ==. That
        # is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
        # implement this in terms of the other specifiers instead of
        # implementing it ourselves. The only thing we need to do is construct
        # the other specifiers.

        # We want everything but the last item in the version, but we want to
        # ignore suffix segments.
        prefix = ".".join(
            list(itertools.takewhile(_is_not_suffix, _version_split(spec)))[:-1]
        )

        # Add the prefix notation to the end of our string
        prefix += ".*"

        return self._get_operator(">=")(prospective, spec) and self._get_operator("==")(
            prospective, prefix
        )

    @_require_version_compare
    def _compare_equal(self, prospective: ParsedVersion, spec: str) -> bool:

        # We need special logic to handle prefix matching
        if spec.endswith(".*"):
            # In the case of prefix matching we want to ignore local segment.
            prospective = Version(prospective.public)
            # Split the spec out by dots, and pretend that there is an implicit
            # dot in between a release segment and a pre-release segment.
            split_spec = _version_split(spec[:-2])  # Remove the trailing .*

            # Split the prospective version out by dots, and pretend that there
            # is an implicit dot in between a release segment and a pre-release
            # segment.
            split_prospective = _version_split(str(prospective))

            # Shorten the prospective version to be the same length as the spec
            # so that we can determine if the specifier is a prefix of the
            # prospective version or not.
            shortened_prospective = split_prospective[: len(split_spec)]

            # Pad out our two sides with zeros so that they both equal the same
            # length.
            padded_spec, padded_prospective = _pad_version(
                split_spec, shortened_prospective
            )

            return padded_prospective == padded_spec
        else:
            # Convert our spec string into a Version
            spec_version = Version(spec)

            # If the specifier does not have a local segment, then we want to
            # act as if the prospective version also does not have a local
            # segment.
            if not spec_version.local:
                prospective = Version(prospective.public)

            return prospective == spec_version

    @_require_version_compare
    def _compare_not_equal(self, prospective: ParsedVersion, spec: str) -> bool:
        return not self._compare_equal(prospective, spec)

    @_require_version_compare
    def _compare_less_than_equal(self, prospective: ParsedVersion, spec: str) -> bool:

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) <= Version(spec)

    @_require_version_compare
    def _compare_greater_than_equal(
        self, prospective: ParsedVersion, spec: str
    ) -> bool:

        # NB: Local version identifiers are NOT permitted in the version
        # specifier, so local version labels can be universally removed from
        # the prospective version.
        return Version(prospective.public) >= Version(spec)

    @_require_version_compare
    def _compare_less_than(self, prospective: ParsedVersion, spec_str: str) -> bool:

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is less than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective < spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a pre-release version, that we do not accept pre-release
        # versions for the version mentioned in the specifier (e.g. <3.1 should
        # not match 3.1.dev0, but should match 3.0.dev0).
        if not spec.is_prerelease and prospective.is_prerelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # less than the spec version *and* it's not a pre-release of the same
        # version in the spec.
        return True

    @_require_version_compare
    def _compare_greater_than(self, prospective: ParsedVersion, spec_str: str) -> bool:

        # Convert our spec to a Version instance, since we'll want to work with
        # it as a version.
        spec = Version(spec_str)

        # Check to see if the prospective version is greater than the spec
        # version. If it's not we can short circuit and just return False now
        # instead of doing extra unneeded work.
        if not prospective > spec:
            return False

        # This special case is here so that, unless the specifier itself
        # includes is a post-release version, that we do not accept
        # post-release versions for the version mentioned in the specifier
        # (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
        if not spec.is_postrelease and prospective.is_postrelease:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # Ensure that we do not allow a local version of the version mentioned
        # in the specifier, which is technically greater than, to match.
        if prospective.local is not None:
            if Version(prospective.base_version) == Version(spec.base_version):
                return False

        # If we've gotten to here, it means that prospective version is both
        # greater than the spec version *and* it's not a pre-release of the
        # same version in the spec.
        return True

    def _compare_arbitrary(self, prospective: Version, spec: str) -> bool:
        return str(prospective).lower() == str(spec).lower()

    @property
    def prereleases(self) -> bool:

        # If there is an explicit prereleases set for this, then we'll just
        # blindly use that.
        if self._prereleases is not None:
            return self._prereleases

        # Look at all of our specifiers and determine if they are inclusive
        # operators, and if they are if they are including an explicit
        # prerelease.
        operator, version = self._spec
        if operator in ["==", ">=", "<=", "~=", "==="]:
            # The == specifier can include a trailing .*, if it does we
            # want to remove before parsing.
            if operator == "==" and version.endswith(".*"):
                version = version[:-2]

            # Parse the version, and if it is a pre-release than this
            # specifier allows pre-releases.
            if parse(version).is_prerelease:
                return True

        return False

    @prereleases.setter
    def prereleases(self, value: bool) -> None:
        self._prereleases = value


_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")


def _version_split(version: str) -> List[str]:
    result: List[str] = []
    for item in version.split("."):
        match = _prefix_regex.search(item)
        if match:
            result.extend(match.groups())
        else:
            result.append(item)
    return result


def _is_not_suffix(segment: str) -> bool:
    return not any(
        segment.startswith(prefix) for prefix in ("dev", "a", "b", "rc", "post")
    )


def _pad_version(left: List[str], right: List[str]) -> Tuple[List[str], List[str]]:
    left_split, right_split = [], []

    # Get the release segment of our versions
    left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
    right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))

    # Get the rest of our versions
    left_split.append(left[len(left_split[0]) :])
    right_split.append(right[len(right_split[0]) :])

    # Insert our padding
    left_split.insert(1, ["0"] * max(0, len(right_split[0]) - len(left_split[0])))
    right_split.insert(1, ["0"] * max(0, len(left_split[0]) - len(right_split[0])))

    return (list(itertools.chain(*left_split)), list(itertools.chain(*right_split)))


class SpecifierSet(BaseSpecifier):
    def __init__(
        self, specifiers: str = "", prereleases: Optional[bool] = None
    ) -> None:

        # Split on , to break each individual specifier into it's own item, and
        # strip each item to remove leading/trailing whitespace.
        split_specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]

        # Parsed each individual specifier, attempting first to make it a
        # Specifier and falling back to a LegacySpecifier.
        parsed: Set[_IndividualSpecifier] = set()
        for specifier in split_specifiers:
            try:
                parsed.add(Specifier(specifier))
            except InvalidSpecifier:
                parsed.add(LegacySpecifier(specifier))

        # Turn our parsed specifiers into a frozen set and save them for later.
        self._specs = frozenset(parsed)

        # Store our prereleases value so we can use it later to determine if
        # we accept prereleases or not.
        self._prereleases = prereleases

    def __repr__(self) -> str:
        pre = (
            f", prereleases={self.prereleases!r}"
            if self._prereleases is not None
            else ""
        )

        return "<SpecifierSet({!r}{})>".format(str(self), pre)

    def __str__(self) -> str:
        return ",".join(sorted(str(s) for s in self._specs))

    def __hash__(self) -> int:
        return hash(self._specs)

    def __and__(self, other: Union["SpecifierSet", str]) -> "SpecifierSet":
        if isinstance(other, str):
            other = SpecifierSet(other)
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        specifier = SpecifierSet()
        specifier._specs = frozenset(self._specs | other._specs)

        if self._prereleases is None and other._prereleases is not None:
            specifier._prereleases = other._prereleases
        elif self._prereleases is not None and other._prereleases is None:
            specifier._prereleases = self._prereleases
        elif self._prereleases == other._prereleases:
            specifier._prereleases = self._prereleases
        else:
            raise ValueError(
                "Cannot combine SpecifierSets with True and False prerelease "
                "overrides."
            )

        return specifier

    def __eq__(self, other: object) -> bool:
        if isinstance(other, (str, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs == other._specs

    def __ne__(self, other: object) -> bool:
        if isinstance(other, (str, _IndividualSpecifier)):
            other = SpecifierSet(str(other))
        elif not isinstance(other, SpecifierSet):
            return NotImplemented

        return self._specs != other._specs

    def __len__(self) -> int:
        return len(self._specs)

    def __iter__(self) -> Iterator[_IndividualSpecifier]:
        return iter(self._specs)

    @property
    def prereleases(self) -> Optional[bool]:

        # If we have been given an explicit prerelease modifier, then we'll
        # pass that through here.
        if self._prereleases is not None:
            return self._prereleases

        # If we don't have any specifiers, and we don't have a forced value,
        # then we'll just return None since we don't know if this should have
        # pre-releases or not.
        if not self._specs:
            return None

        # Otherwise we'll see if any of the given specifiers accept
        # prereleases, if any of them do we'll return True, otherwise False.
        return any(s.prereleases for s in self._specs)

    @prereleases.setter
    def prereleases(self, value: bool) -> None:
        self._prereleases = value

    def __contains__(self, item: UnparsedVersion) -> bool:
        return self.contains(item)

    def contains(
        self, item: UnparsedVersion, prereleases: Optional[bool] = None
    ) -> bool:

        # Ensure that our item is a Version or LegacyVersion instance.
        if not isinstance(item, (LegacyVersion, Version)):
            item = parse(item)

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # We can determine if we're going to allow pre-releases by looking to
        # see if any of the underlying items supports them. If none of them do
        # and this item is a pre-release then we do not allow it and we can
        # short circuit that here.
        # Note: This means that 1.0.dev1 would not be contained in something
        #       like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
        if not prereleases and item.is_prerelease:
            return False

        # We simply dispatch to the underlying specs here to make sure that the
        # given version is contained within all of them.
        # Note: This use of all() here means that an empty set of specifiers
        #       will always return True, this is an explicit design decision.
        return all(s.contains(item, prereleases=prereleases) for s in self._specs)

    def filter(
        self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
    ) -> Iterable[VersionTypeVar]:

        # Determine if we're forcing a prerelease or not, if we're not forcing
        # one for this particular filter call, then we'll use whatever the
        # SpecifierSet thinks for whether or not we should support prereleases.
        if prereleases is None:
            prereleases = self.prereleases

        # If we have any specifiers, then we want to wrap our iterable in the
        # filter method for each one, this will act as a logical AND amongst
        # each specifier.
        if self._specs:
            for spec in self._specs:
                iterable = spec.filter(iterable, prereleases=bool(prereleases))
            return iterable
        # If we do not have any specifiers, then we need to have a rough filter
        # which will filter out any pre-releases, unless there are no final
        # releases, and which will filter out LegacyVersion in general.
        else:
            filtered: List[VersionTypeVar] = []
            found_prereleases: List[VersionTypeVar] = []

            item: UnparsedVersion
            parsed_version: Union[Version, LegacyVersion]

            for item in iterable:
                # Ensure that we some kind of Version class for this item.
                if not isinstance(item, (LegacyVersion, Version)):
                    parsed_version = parse(item)
                else:
                    parsed_version = item

                # Filter out any item which is parsed as a LegacyVersion
                if isinstance(parsed_version, LegacyVersion):
                    continue

                # Store any item which is a pre-release for later unless we've
                # already found a final version or we are accepting prereleases
                if parsed_version.is_prerelease and not prereleases:
                    if not filtered:
                        found_prereleases.append(item)
                else:
                    filtered.append(item)

            # If we've found no items except for pre-releases, then we'll go
            # ahead and use the pre-releases
            if not filtered and found_prereleases and prereleases is None:
                return found_prereleases

            return filtered
site-packages/pip/_vendor/packaging/_structures.py000064400000003135150564745010016401 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.


class InfinityType:
    def __repr__(self) -> str:
        return "Infinity"

    def __hash__(self) -> int:
        return hash(repr(self))

    def __lt__(self, other: object) -> bool:
        return False

    def __le__(self, other: object) -> bool:
        return False

    def __eq__(self, other: object) -> bool:
        return isinstance(other, self.__class__)

    def __ne__(self, other: object) -> bool:
        return not isinstance(other, self.__class__)

    def __gt__(self, other: object) -> bool:
        return True

    def __ge__(self, other: object) -> bool:
        return True

    def __neg__(self: object) -> "NegativeInfinityType":
        return NegativeInfinity


Infinity = InfinityType()


class NegativeInfinityType:
    def __repr__(self) -> str:
        return "-Infinity"

    def __hash__(self) -> int:
        return hash(repr(self))

    def __lt__(self, other: object) -> bool:
        return True

    def __le__(self, other: object) -> bool:
        return True

    def __eq__(self, other: object) -> bool:
        return isinstance(other, self.__class__)

    def __ne__(self, other: object) -> bool:
        return not isinstance(other, self.__class__)

    def __gt__(self, other: object) -> bool:
        return False

    def __ge__(self, other: object) -> bool:
        return False

    def __neg__(self: object) -> InfinityType:
        return Infinity


NegativeInfinity = NegativeInfinityType()
site-packages/pip/_vendor/packaging/tags.py000064400000036542150564745010014765 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import logging
import platform
import sys
import sysconfig
from importlib.machinery import EXTENSION_SUFFIXES
from typing import (
    Dict,
    FrozenSet,
    Iterable,
    Iterator,
    List,
    Optional,
    Sequence,
    Tuple,
    Union,
    cast,
)

from . import _manylinux, _musllinux

logger = logging.getLogger(__name__)

PythonVersion = Sequence[int]
MacVersion = Tuple[int, int]

INTERPRETER_SHORT_NAMES: Dict[str, str] = {
    "python": "py",  # Generic.
    "cpython": "cp",
    "pypy": "pp",
    "ironpython": "ip",
    "jython": "jy",
}


_32_BIT_INTERPRETER = sys.maxsize <= 2 ** 32


class Tag:
    """
    A representation of the tag triple for a wheel.

    Instances are considered immutable and thus are hashable. Equality checking
    is also supported.
    """

    __slots__ = ["_interpreter", "_abi", "_platform", "_hash"]

    def __init__(self, interpreter: str, abi: str, platform: str) -> None:
        self._interpreter = interpreter.lower()
        self._abi = abi.lower()
        self._platform = platform.lower()
        # The __hash__ of every single element in a Set[Tag] will be evaluated each time
        # that a set calls its `.disjoint()` method, which may be called hundreds of
        # times when scanning a page of links for packages with tags matching that
        # Set[Tag]. Pre-computing the value here produces significant speedups for
        # downstream consumers.
        self._hash = hash((self._interpreter, self._abi, self._platform))

    @property
    def interpreter(self) -> str:
        return self._interpreter

    @property
    def abi(self) -> str:
        return self._abi

    @property
    def platform(self) -> str:
        return self._platform

    def __eq__(self, other: object) -> bool:
        if not isinstance(other, Tag):
            return NotImplemented

        return (
            (self._hash == other._hash)  # Short-circuit ASAP for perf reasons.
            and (self._platform == other._platform)
            and (self._abi == other._abi)
            and (self._interpreter == other._interpreter)
        )

    def __hash__(self) -> int:
        return self._hash

    def __str__(self) -> str:
        return f"{self._interpreter}-{self._abi}-{self._platform}"

    def __repr__(self) -> str:
        return "<{self} @ {self_id}>".format(self=self, self_id=id(self))


def parse_tag(tag: str) -> FrozenSet[Tag]:
    """
    Parses the provided tag (e.g. `py3-none-any`) into a frozenset of Tag instances.

    Returning a set is required due to the possibility that the tag is a
    compressed tag set.
    """
    tags = set()
    interpreters, abis, platforms = tag.split("-")
    for interpreter in interpreters.split("."):
        for abi in abis.split("."):
            for platform_ in platforms.split("."):
                tags.add(Tag(interpreter, abi, platform_))
    return frozenset(tags)


def _get_config_var(name: str, warn: bool = False) -> Union[int, str, None]:
    value = sysconfig.get_config_var(name)
    if value is None and warn:
        logger.debug(
            "Config variable '%s' is unset, Python ABI tag may be incorrect", name
        )
    return value


def _normalize_string(string: str) -> str:
    return string.replace(".", "_").replace("-", "_")


def _abi3_applies(python_version: PythonVersion) -> bool:
    """
    Determine if the Python version supports abi3.

    PEP 384 was first implemented in Python 3.2.
    """
    return len(python_version) > 1 and tuple(python_version) >= (3, 2)


def _cpython_abis(py_version: PythonVersion, warn: bool = False) -> List[str]:
    py_version = tuple(py_version)  # To allow for version comparison.
    abis = []
    version = _version_nodot(py_version[:2])
    debug = pymalloc = ucs4 = ""
    with_debug = _get_config_var("Py_DEBUG", warn)
    has_refcount = hasattr(sys, "gettotalrefcount")
    # Windows doesn't set Py_DEBUG, so checking for support of debug-compiled
    # extension modules is the best option.
    # https://github.com/pypa/pip/issues/3383#issuecomment-173267692
    has_ext = "_d.pyd" in EXTENSION_SUFFIXES
    if with_debug or (with_debug is None and (has_refcount or has_ext)):
        debug = "d"
    if py_version < (3, 8):
        with_pymalloc = _get_config_var("WITH_PYMALLOC", warn)
        if with_pymalloc or with_pymalloc is None:
            pymalloc = "m"
        if py_version < (3, 3):
            unicode_size = _get_config_var("Py_UNICODE_SIZE", warn)
            if unicode_size == 4 or (
                unicode_size is None and sys.maxunicode == 0x10FFFF
            ):
                ucs4 = "u"
    elif debug:
        # Debug builds can also load "normal" extension modules.
        # We can also assume no UCS-4 or pymalloc requirement.
        abis.append(f"cp{version}")
    abis.insert(
        0,
        "cp{version}{debug}{pymalloc}{ucs4}".format(
            version=version, debug=debug, pymalloc=pymalloc, ucs4=ucs4
        ),
    )
    return abis


def cpython_tags(
    python_version: Optional[PythonVersion] = None,
    abis: Optional[Iterable[str]] = None,
    platforms: Optional[Iterable[str]] = None,
    *,
    warn: bool = False,
) -> Iterator[Tag]:
    """
    Yields the tags for a CPython interpreter.

    The tags consist of:
    - cp<python_version>-<abi>-<platform>
    - cp<python_version>-abi3-<platform>
    - cp<python_version>-none-<platform>
    - cp<less than python_version>-abi3-<platform>  # Older Python versions down to 3.2.

    If python_version only specifies a major version then user-provided ABIs and
    the 'none' ABItag will be used.

    If 'abi3' or 'none' are specified in 'abis' then they will be yielded at
    their normal position and not at the beginning.
    """
    if not python_version:
        python_version = sys.version_info[:2]

    interpreter = "cp{}".format(_version_nodot(python_version[:2]))

    if abis is None:
        if len(python_version) > 1:
            abis = _cpython_abis(python_version, warn)
        else:
            abis = []
    abis = list(abis)
    # 'abi3' and 'none' are explicitly handled later.
    for explicit_abi in ("abi3", "none"):
        try:
            abis.remove(explicit_abi)
        except ValueError:
            pass

    platforms = list(platforms or _platform_tags())
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)
    if _abi3_applies(python_version):
        yield from (Tag(interpreter, "abi3", platform_) for platform_ in platforms)
    yield from (Tag(interpreter, "none", platform_) for platform_ in platforms)

    if _abi3_applies(python_version):
        for minor_version in range(python_version[1] - 1, 1, -1):
            for platform_ in platforms:
                interpreter = "cp{version}".format(
                    version=_version_nodot((python_version[0], minor_version))
                )
                yield Tag(interpreter, "abi3", platform_)


def _generic_abi() -> Iterator[str]:
    abi = sysconfig.get_config_var("SOABI")
    if abi:
        yield _normalize_string(abi)


def generic_tags(
    interpreter: Optional[str] = None,
    abis: Optional[Iterable[str]] = None,
    platforms: Optional[Iterable[str]] = None,
    *,
    warn: bool = False,
) -> Iterator[Tag]:
    """
    Yields the tags for a generic interpreter.

    The tags consist of:
    - <interpreter>-<abi>-<platform>

    The "none" ABI will be added if it was not explicitly provided.
    """
    if not interpreter:
        interp_name = interpreter_name()
        interp_version = interpreter_version(warn=warn)
        interpreter = "".join([interp_name, interp_version])
    if abis is None:
        abis = _generic_abi()
    platforms = list(platforms or _platform_tags())
    abis = list(abis)
    if "none" not in abis:
        abis.append("none")
    for abi in abis:
        for platform_ in platforms:
            yield Tag(interpreter, abi, platform_)


def _py_interpreter_range(py_version: PythonVersion) -> Iterator[str]:
    """
    Yields Python versions in descending order.

    After the latest version, the major-only version will be yielded, and then
    all previous versions of that major version.
    """
    if len(py_version) > 1:
        yield "py{version}".format(version=_version_nodot(py_version[:2]))
    yield "py{major}".format(major=py_version[0])
    if len(py_version) > 1:
        for minor in range(py_version[1] - 1, -1, -1):
            yield "py{version}".format(version=_version_nodot((py_version[0], minor)))


def compatible_tags(
    python_version: Optional[PythonVersion] = None,
    interpreter: Optional[str] = None,
    platforms: Optional[Iterable[str]] = None,
) -> Iterator[Tag]:
    """
    Yields the sequence of tags that are compatible with a specific version of Python.

    The tags consist of:
    - py*-none-<platform>
    - <interpreter>-none-any  # ... if `interpreter` is provided.
    - py*-none-any
    """
    if not python_version:
        python_version = sys.version_info[:2]
    platforms = list(platforms or _platform_tags())
    for version in _py_interpreter_range(python_version):
        for platform_ in platforms:
            yield Tag(version, "none", platform_)
    if interpreter:
        yield Tag(interpreter, "none", "any")
    for version in _py_interpreter_range(python_version):
        yield Tag(version, "none", "any")


def _mac_arch(arch: str, is_32bit: bool = _32_BIT_INTERPRETER) -> str:
    if not is_32bit:
        return arch

    if arch.startswith("ppc"):
        return "ppc"

    return "i386"


def _mac_binary_formats(version: MacVersion, cpu_arch: str) -> List[str]:
    formats = [cpu_arch]
    if cpu_arch == "x86_64":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat64", "fat32"])

    elif cpu_arch == "i386":
        if version < (10, 4):
            return []
        formats.extend(["intel", "fat32", "fat"])

    elif cpu_arch == "ppc64":
        # TODO: Need to care about 32-bit PPC for ppc64 through 10.2?
        if version > (10, 5) or version < (10, 4):
            return []
        formats.append("fat64")

    elif cpu_arch == "ppc":
        if version > (10, 6):
            return []
        formats.extend(["fat32", "fat"])

    if cpu_arch in {"arm64", "x86_64"}:
        formats.append("universal2")

    if cpu_arch in {"x86_64", "i386", "ppc64", "ppc", "intel"}:
        formats.append("universal")

    return formats


def mac_platforms(
    version: Optional[MacVersion] = None, arch: Optional[str] = None
) -> Iterator[str]:
    """
    Yields the platform tags for a macOS system.

    The `version` parameter is a two-item tuple specifying the macOS version to
    generate platform tags for. The `arch` parameter is the CPU architecture to
    generate platform tags for. Both parameters default to the appropriate value
    for the current system.
    """
    version_str, _, cpu_arch = platform.mac_ver()
    if version is None:
        version = cast("MacVersion", tuple(map(int, version_str.split(".")[:2])))
    else:
        version = version
    if arch is None:
        arch = _mac_arch(cpu_arch)
    else:
        arch = arch

    if (10, 0) <= version and version < (11, 0):
        # Prior to Mac OS 11, each yearly release of Mac OS bumped the
        # "minor" version number.  The major version was always 10.
        for minor_version in range(version[1], -1, -1):
            compat_version = 10, minor_version
            binary_formats = _mac_binary_formats(compat_version, arch)
            for binary_format in binary_formats:
                yield "macosx_{major}_{minor}_{binary_format}".format(
                    major=10, minor=minor_version, binary_format=binary_format
                )

    if version >= (11, 0):
        # Starting with Mac OS 11, each yearly release bumps the major version
        # number.   The minor versions are now the midyear updates.
        for major_version in range(version[0], 10, -1):
            compat_version = major_version, 0
            binary_formats = _mac_binary_formats(compat_version, arch)
            for binary_format in binary_formats:
                yield "macosx_{major}_{minor}_{binary_format}".format(
                    major=major_version, minor=0, binary_format=binary_format
                )

    if version >= (11, 0):
        # Mac OS 11 on x86_64 is compatible with binaries from previous releases.
        # Arm64 support was introduced in 11.0, so no Arm binaries from previous
        # releases exist.
        #
        # However, the "universal2" binary format can have a
        # macOS version earlier than 11.0 when the x86_64 part of the binary supports
        # that version of macOS.
        if arch == "x86_64":
            for minor_version in range(16, 3, -1):
                compat_version = 10, minor_version
                binary_formats = _mac_binary_formats(compat_version, arch)
                for binary_format in binary_formats:
                    yield "macosx_{major}_{minor}_{binary_format}".format(
                        major=compat_version[0],
                        minor=compat_version[1],
                        binary_format=binary_format,
                    )
        else:
            for minor_version in range(16, 3, -1):
                compat_version = 10, minor_version
                binary_format = "universal2"
                yield "macosx_{major}_{minor}_{binary_format}".format(
                    major=compat_version[0],
                    minor=compat_version[1],
                    binary_format=binary_format,
                )


def _linux_platforms(is_32bit: bool = _32_BIT_INTERPRETER) -> Iterator[str]:
    linux = _normalize_string(sysconfig.get_platform())
    if is_32bit:
        if linux == "linux_x86_64":
            linux = "linux_i686"
        elif linux == "linux_aarch64":
            linux = "linux_armv7l"
    _, arch = linux.split("_", 1)
    yield from _manylinux.platform_tags(linux, arch)
    yield from _musllinux.platform_tags(arch)
    yield linux


def _generic_platforms() -> Iterator[str]:
    yield _normalize_string(sysconfig.get_platform())


def _platform_tags() -> Iterator[str]:
    """
    Provides the platform tags for this installation.
    """
    if platform.system() == "Darwin":
        return mac_platforms()
    elif platform.system() == "Linux":
        return _linux_platforms()
    else:
        return _generic_platforms()


def interpreter_name() -> str:
    """
    Returns the name of the running interpreter.
    """
    name = sys.implementation.name
    return INTERPRETER_SHORT_NAMES.get(name) or name


def interpreter_version(*, warn: bool = False) -> str:
    """
    Returns the version of the running interpreter.
    """
    version = _get_config_var("py_version_nodot", warn=warn)
    if version:
        version = str(version)
    else:
        version = _version_nodot(sys.version_info[:2])
    return version


def _version_nodot(version: PythonVersion) -> str:
    return "".join(map(str, version))


def sys_tags(*, warn: bool = False) -> Iterator[Tag]:
    """
    Returns the sequence of tag triples for the running interpreter.

    The order of the sequence corresponds to priority order for the
    interpreter, from most to least important.
    """

    interp_name = interpreter_name()
    if interp_name == "cp":
        yield from cpython_tags(warn=warn)
    else:
        yield from generic_tags()

    yield from compatible_tags()
site-packages/pip/_vendor/packaging/requirements.py000064400000011104150564745010016535 0ustar00# This file is dual licensed under the terms of the Apache License, Version
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
# for complete details.

import re
import string
import urllib.parse
from typing import List, Optional as TOptional, Set

from pip._vendor.pyparsing import (  # noqa
    Combine,
    Literal as L,
    Optional,
    ParseException,
    Regex,
    Word,
    ZeroOrMore,
    originalTextFor,
    stringEnd,
    stringStart,
)

from .markers import MARKER_EXPR, Marker
from .specifiers import LegacySpecifier, Specifier, SpecifierSet


class InvalidRequirement(ValueError):
    """
    An invalid requirement was found, users should refer to PEP 508.
    """


ALPHANUM = Word(string.ascii_letters + string.digits)

LBRACKET = L("[").suppress()
RBRACKET = L("]").suppress()
LPAREN = L("(").suppress()
RPAREN = L(")").suppress()
COMMA = L(",").suppress()
SEMICOLON = L(";").suppress()
AT = L("@").suppress()

PUNCTUATION = Word("-_.")
IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM)
IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END))

NAME = IDENTIFIER("name")
EXTRA = IDENTIFIER

URI = Regex(r"[^ ]+")("url")
URL = AT + URI

EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA)
EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras")

VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE)
VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE)

VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY
VERSION_MANY = Combine(
    VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False
)("_raw_spec")
_VERSION_SPEC = Optional((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY)
_VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or "")

VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier")
VERSION_SPEC.setParseAction(lambda s, l, t: t[1])

MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker")
MARKER_EXPR.setParseAction(
    lambda s, l, t: Marker(s[t._original_start : t._original_end])
)
MARKER_SEPARATOR = SEMICOLON
MARKER = MARKER_SEPARATOR + MARKER_EXPR

VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER)
URL_AND_MARKER = URL + Optional(MARKER)

NAMED_REQUIREMENT = NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER)

REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd
# pyparsing isn't thread safe during initialization, so we do it eagerly, see
# issue #104
REQUIREMENT.parseString("x[]")


class Requirement:
    """Parse a requirement.

    Parse a given requirement string into its parts, such as name, specifier,
    URL, and extras. Raises InvalidRequirement on a badly-formed requirement
    string.
    """

    # TODO: Can we test whether something is contained within a requirement?
    #       If so how do we do that? Do we need to test against the _name_ of
    #       the thing as well as the version? What about the markers?
    # TODO: Can we normalize the name and extra name?

    def __init__(self, requirement_string: str) -> None:
        try:
            req = REQUIREMENT.parseString(requirement_string)
        except ParseException as e:
            raise InvalidRequirement(
                f'Parse error at "{ requirement_string[e.loc : e.loc + 8]!r}": {e.msg}'
            )

        self.name: str = req.name
        if req.url:
            parsed_url = urllib.parse.urlparse(req.url)
            if parsed_url.scheme == "file":
                if urllib.parse.urlunparse(parsed_url) != req.url:
                    raise InvalidRequirement("Invalid URL given")
            elif not (parsed_url.scheme and parsed_url.netloc) or (
                not parsed_url.scheme and not parsed_url.netloc
            ):
                raise InvalidRequirement(f"Invalid URL: {req.url}")
            self.url: TOptional[str] = req.url
        else:
            self.url = None
        self.extras: Set[str] = set(req.extras.asList() if req.extras else [])
        self.specifier: SpecifierSet = SpecifierSet(req.specifier)
        self.marker: TOptional[Marker] = req.marker if req.marker else None

    def __str__(self) -> str:
        parts: List[str] = [self.name]

        if self.extras:
            formatted_extras = ",".join(sorted(self.extras))
            parts.append(f"[{formatted_extras}]")

        if self.specifier:
            parts.append(str(self.specifier))

        if self.url:
            parts.append(f"@ {self.url}")
            if self.marker:
                parts.append(" ")

        if self.marker:
            parts.append(f"; {self.marker}")

        return "".join(parts)

    def __repr__(self) -> str:
        return f"<Requirement('{self}')>"
site-packages/pip/_vendor/html5lib/__pycache__/_ihatexml.cpython-39.pyc000064400000032762150564745010022064 0ustar00a

��ReXA�	@s�ddlmZmZmZddlZddlZddlmZdZdZ	dZ
dZd	Zd
�
ee	g�Zd
�
eeddd
e
eg�Zd
�
ed
g�Ze�d�Ze�d�Zdd�Zdd�Zedd�Zdd�Zdd�Zdd�Zdd�Ze�d�Ze�d�Ze�d �ZGd!d"�d"e�Z dS)#�)�absolute_import�division�unicode_literalsN�)�DataLossWarninga^
[#x0041-#x005A] | [#x0061-#x007A] | [#x00C0-#x00D6] | [#x00D8-#x00F6] |
[#x00F8-#x00FF] | [#x0100-#x0131] | [#x0134-#x013E] | [#x0141-#x0148] |
[#x014A-#x017E] | [#x0180-#x01C3] | [#x01CD-#x01F0] | [#x01F4-#x01F5] |
[#x01FA-#x0217] | [#x0250-#x02A8] | [#x02BB-#x02C1] | #x0386 |
[#x0388-#x038A] | #x038C | [#x038E-#x03A1] | [#x03A3-#x03CE] |
[#x03D0-#x03D6] | #x03DA | #x03DC | #x03DE | #x03E0 | [#x03E2-#x03F3] |
[#x0401-#x040C] | [#x040E-#x044F] | [#x0451-#x045C] | [#x045E-#x0481] |
[#x0490-#x04C4] | [#x04C7-#x04C8] | [#x04CB-#x04CC] | [#x04D0-#x04EB] |
[#x04EE-#x04F5] | [#x04F8-#x04F9] | [#x0531-#x0556] | #x0559 |
[#x0561-#x0586] | [#x05D0-#x05EA] | [#x05F0-#x05F2] | [#x0621-#x063A] |
[#x0641-#x064A] | [#x0671-#x06B7] | [#x06BA-#x06BE] | [#x06C0-#x06CE] |
[#x06D0-#x06D3] | #x06D5 | [#x06E5-#x06E6] | [#x0905-#x0939] | #x093D |
[#x0958-#x0961] | [#x0985-#x098C] | [#x098F-#x0990] | [#x0993-#x09A8] |
[#x09AA-#x09B0] | #x09B2 | [#x09B6-#x09B9] | [#x09DC-#x09DD] |
[#x09DF-#x09E1] | [#x09F0-#x09F1] | [#x0A05-#x0A0A] | [#x0A0F-#x0A10] |
[#x0A13-#x0A28] | [#x0A2A-#x0A30] | [#x0A32-#x0A33] | [#x0A35-#x0A36] |
[#x0A38-#x0A39] | [#x0A59-#x0A5C] | #x0A5E | [#x0A72-#x0A74] |
[#x0A85-#x0A8B] | #x0A8D | [#x0A8F-#x0A91] | [#x0A93-#x0AA8] |
[#x0AAA-#x0AB0] | [#x0AB2-#x0AB3] | [#x0AB5-#x0AB9] | #x0ABD | #x0AE0 |
[#x0B05-#x0B0C] | [#x0B0F-#x0B10] | [#x0B13-#x0B28] | [#x0B2A-#x0B30] |
[#x0B32-#x0B33] | [#x0B36-#x0B39] | #x0B3D | [#x0B5C-#x0B5D] |
[#x0B5F-#x0B61] | [#x0B85-#x0B8A] | [#x0B8E-#x0B90] | [#x0B92-#x0B95] |
[#x0B99-#x0B9A] | #x0B9C | [#x0B9E-#x0B9F] | [#x0BA3-#x0BA4] |
[#x0BA8-#x0BAA] | [#x0BAE-#x0BB5] | [#x0BB7-#x0BB9] | [#x0C05-#x0C0C] |
[#x0C0E-#x0C10] | [#x0C12-#x0C28] | [#x0C2A-#x0C33] | [#x0C35-#x0C39] |
[#x0C60-#x0C61] | [#x0C85-#x0C8C] | [#x0C8E-#x0C90] | [#x0C92-#x0CA8] |
[#x0CAA-#x0CB3] | [#x0CB5-#x0CB9] | #x0CDE | [#x0CE0-#x0CE1] |
[#x0D05-#x0D0C] | [#x0D0E-#x0D10] | [#x0D12-#x0D28] | [#x0D2A-#x0D39] |
[#x0D60-#x0D61] | [#x0E01-#x0E2E] | #x0E30 | [#x0E32-#x0E33] |
[#x0E40-#x0E45] | [#x0E81-#x0E82] | #x0E84 | [#x0E87-#x0E88] | #x0E8A |
#x0E8D | [#x0E94-#x0E97] | [#x0E99-#x0E9F] | [#x0EA1-#x0EA3] | #x0EA5 |
#x0EA7 | [#x0EAA-#x0EAB] | [#x0EAD-#x0EAE] | #x0EB0 | [#x0EB2-#x0EB3] |
#x0EBD | [#x0EC0-#x0EC4] | [#x0F40-#x0F47] | [#x0F49-#x0F69] |
[#x10A0-#x10C5] | [#x10D0-#x10F6] | #x1100 | [#x1102-#x1103] |
[#x1105-#x1107] | #x1109 | [#x110B-#x110C] | [#x110E-#x1112] | #x113C |
#x113E | #x1140 | #x114C | #x114E | #x1150 | [#x1154-#x1155] | #x1159 |
[#x115F-#x1161] | #x1163 | #x1165 | #x1167 | #x1169 | [#x116D-#x116E] |
[#x1172-#x1173] | #x1175 | #x119E | #x11A8 | #x11AB | [#x11AE-#x11AF] |
[#x11B7-#x11B8] | #x11BA | [#x11BC-#x11C2] | #x11EB | #x11F0 | #x11F9 |
[#x1E00-#x1E9B] | [#x1EA0-#x1EF9] | [#x1F00-#x1F15] | [#x1F18-#x1F1D] |
[#x1F20-#x1F45] | [#x1F48-#x1F4D] | [#x1F50-#x1F57] | #x1F59 | #x1F5B |
#x1F5D | [#x1F5F-#x1F7D] | [#x1F80-#x1FB4] | [#x1FB6-#x1FBC] | #x1FBE |
[#x1FC2-#x1FC4] | [#x1FC6-#x1FCC] | [#x1FD0-#x1FD3] | [#x1FD6-#x1FDB] |
[#x1FE0-#x1FEC] | [#x1FF2-#x1FF4] | [#x1FF6-#x1FFC] | #x2126 |
[#x212A-#x212B] | #x212E | [#x2180-#x2182] | [#x3041-#x3094] |
[#x30A1-#x30FA] | [#x3105-#x312C] | [#xAC00-#xD7A3]z*[#x4E00-#x9FA5] | #x3007 | [#x3021-#x3029]a�
[#x0300-#x0345] | [#x0360-#x0361] | [#x0483-#x0486] | [#x0591-#x05A1] |
[#x05A3-#x05B9] | [#x05BB-#x05BD] | #x05BF | [#x05C1-#x05C2] | #x05C4 |
[#x064B-#x0652] | #x0670 | [#x06D6-#x06DC] | [#x06DD-#x06DF] |
[#x06E0-#x06E4] | [#x06E7-#x06E8] | [#x06EA-#x06ED] | [#x0901-#x0903] |
#x093C | [#x093E-#x094C] | #x094D | [#x0951-#x0954] | [#x0962-#x0963] |
[#x0981-#x0983] | #x09BC | #x09BE | #x09BF | [#x09C0-#x09C4] |
[#x09C7-#x09C8] | [#x09CB-#x09CD] | #x09D7 | [#x09E2-#x09E3] | #x0A02 |
#x0A3C | #x0A3E | #x0A3F | [#x0A40-#x0A42] | [#x0A47-#x0A48] |
[#x0A4B-#x0A4D] | [#x0A70-#x0A71] | [#x0A81-#x0A83] | #x0ABC |
[#x0ABE-#x0AC5] | [#x0AC7-#x0AC9] | [#x0ACB-#x0ACD] | [#x0B01-#x0B03] |
#x0B3C | [#x0B3E-#x0B43] | [#x0B47-#x0B48] | [#x0B4B-#x0B4D] |
[#x0B56-#x0B57] | [#x0B82-#x0B83] | [#x0BBE-#x0BC2] | [#x0BC6-#x0BC8] |
[#x0BCA-#x0BCD] | #x0BD7 | [#x0C01-#x0C03] | [#x0C3E-#x0C44] |
[#x0C46-#x0C48] | [#x0C4A-#x0C4D] | [#x0C55-#x0C56] | [#x0C82-#x0C83] |
[#x0CBE-#x0CC4] | [#x0CC6-#x0CC8] | [#x0CCA-#x0CCD] | [#x0CD5-#x0CD6] |
[#x0D02-#x0D03] | [#x0D3E-#x0D43] | [#x0D46-#x0D48] | [#x0D4A-#x0D4D] |
#x0D57 | #x0E31 | [#x0E34-#x0E3A] | [#x0E47-#x0E4E] | #x0EB1 |
[#x0EB4-#x0EB9] | [#x0EBB-#x0EBC] | [#x0EC8-#x0ECD] | [#x0F18-#x0F19] |
#x0F35 | #x0F37 | #x0F39 | #x0F3E | #x0F3F | [#x0F71-#x0F84] |
[#x0F86-#x0F8B] | [#x0F90-#x0F95] | #x0F97 | [#x0F99-#x0FAD] |
[#x0FB1-#x0FB7] | #x0FB9 | [#x20D0-#x20DC] | #x20E1 | [#x302A-#x302F] |
#x3099 | #x309Aa
[#x0030-#x0039] | [#x0660-#x0669] | [#x06F0-#x06F9] | [#x0966-#x096F] |
[#x09E6-#x09EF] | [#x0A66-#x0A6F] | [#x0AE6-#x0AEF] | [#x0B66-#x0B6F] |
[#x0BE7-#x0BEF] | [#x0C66-#x0C6F] | [#x0CE6-#x0CEF] | [#x0D66-#x0D6F] |
[#x0E50-#x0E59] | [#x0ED0-#x0ED9] | [#x0F20-#x0F29]z}
#x00B7 | #x02D0 | #x02D1 | #x0387 | #x0640 | #x0E46 | #x0EC6 | #x3005 |
#[#x3031-#x3035] | [#x309D-#x309E] | [#x30FC-#x30FE]� | �.�-�_z#x([\d|A-F]{4,4})z'\[#x([\d|A-F]{4,4})-#x([\d|A-F]{4,4})\]cCs�dd�|�d�D�}g}|D]�}d}ttfD]V}|�|�}|dur,|�dd�|��D��t|d�dkrz|dd|d<d	}q�q,|st|�dks�J�|�t|�gd�qt|�}|S)
NcSsg|]}|���qS�)�strip��.0�itemrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_ihatexml.py�
<listcomp>h�z$charStringToList.<locals>.<listcomp>rFcSsg|]}t|��qSr)�hexToIntr
rrrror���r�T)	�split�reChar�reCharRange�match�append�groups�len�ord�normaliseCharList)�charsZ
charRanges�rvrZ
foundMatch�regexprrrr�charStringToListgs"
r"cCs�t|�}|D]}|d|dksJ�qg}d}|t|�kr�d}|�||�||t|�kr�|||d|dddkr�|||d|dd<|d7}qL||7}q.|S)Nrrr)�sortedrr)�charListrr �i�jrrrr|s0

rZFFFF�cCs�g}|ddkr*|�d|dddg�t|dd��D].\}}|�|dd||dddg�q:|ddtkr�|�|dddtg�|S)Nrrr)r�	enumerate�max_unicode)r$r r%rrrr�
missingRanges�s(r*cCsng}|D]V}|d|dkr4|�tt|d���q|�tt|d��dtt|d���qdd�|�S)Nrrr	z[%s]�)r�escapeRegexp�chr�join)r$r rrrr�listToRegexpStr�s�r/cCs
t|d�S)Nr')�int)�hex_strrrrr�srcCs"d}|D]}|�|d|�}q|S)N)r�^�$�*�+�?�{�}�[�]�|�(�)r	�\)�replace)�stringZspecialCharacters�charrrrr,�sr,u�[-,/:-@\[-\^`\{-¶¸-¿×÷IJ-ijĿ-ŀʼnſDŽ-njDZ-dzǶ-ǹȘ-ɏʩ-ʺ˂-ˏ˒-˿͆-͟͢-΅΋΍΢Ϗϗ-ϙϛϝϟϡϴ-ЀЍѐѝ҂҇-ҏӅ-ӆӉ-ӊӍ-ӏӬ-ӭӶ-ӷӺ-԰՗-՘՚-ՠև-֐ֺ֢־׀׃ׅ-׏׫-ׯ׳-ؠػ-ؿٓ-ٟ٪-ٯڸ-ڹڿۏ۔۩ۮ-ۯۺ-ऀऄऺ-ऻॎ-ॐॕ-ॗ।-॥॰-ঀ঄঍-঎঑-঒঩঱঳-঵঺-঻ঽ৅-৆৉-৊ৎ-৖৘-৛৞৤-৥৲-ਁਃ-਄਋-਎਑-਒਩਱਴਷਺-਻਽੃-੆੉-੊੎-੘੝੟-੥ੵ-઀઄ઌ઎઒઩઱઴઺-઻૆૊૎-૟ૡ-૥૰-଀଄଍-଎଑-଒଩଱଴-ଵ଺-଻ୄ-୆୉-୊୎-୕୘-୛୞ୢ-୥୰-஁஄஋-஍஑஖-஘஛஝஠-஢஥-஧஫-஭ஶ஺-஽௃-௅௉௎-௖௘-௦௰-ఀఄ఍఑఩ఴ఺-ఽ౅౉౎-౔౗-౟ౢ-౥౰-ಁ಄಍಑಩಴಺-ಽ೅೉೎-೔೗-ೝ೟ೢ-೥೰-ഁഄ഍഑ഩഺ-ഽൄ-൅൉ൎ-ൖ൘-ൟൢ-൥൰-฀ฯ฻-฿๏๚-຀຃຅-ຆຉ຋-ຌຎ-ຓຘຠ຤຦ຨ-ຩຬຯ຺຾-຿໅໇໎-໏໚-༗༚-༟༪-༴༶༸༺-༽཈ཪ-཰྅ྌ-ྏྖ྘ྮ-ྰྸྺ-႟჆-჏ჷ-ჿᄁᄄᄈᄊᄍᄓ-ᄻᄽᄿᅁ-ᅋᅍᅏᅑ-ᅓᅖ-ᅘᅚ-ᅞᅢᅤᅦᅨᅪ-ᅬᅯ-ᅱᅴᅶ-ᆝᆟ-ᆧᆩ-ᆪᆬ-ᆭᆰ-ᆶᆹᆻᇃ-ᇪᇬ-ᇯᇱ-ᇸᇺ-᷿ẜ-ẟỺ-ỿ἖-἗἞-἟὆-὇὎-὏὘὚὜὞὾-὿᾵᾽᾿-῁῅῍-῏῔-῕῜-῟῭-῱῵´-⃏⃝-⃠⃢-℥℧-℩ℬ-ℭℯ-ⅿↃ-〄〆〈-〠〰〶-぀ゕ-゘゛-゜ゟ-゠・ヿ-㄄ㄭ-䷿龦-꯿힤-￿]u�[-@\[-\^`\{-¿×÷IJ-ijĿ-ŀʼnſDŽ-njDZ-dzǶ-ǹȘ-ɏʩ-ʺ˂-΅·΋΍΢Ϗϗ-ϙϛϝϟϡϴ-ЀЍѐѝ҂-ҏӅ-ӆӉ-ӊӍ-ӏӬ-ӭӶ-ӷӺ-԰՗-՘՚-ՠև-׏׫-ׯ׳-ؠػ-ـً-ٰڸ-ڹڿۏ۔ۖ-ۤۧ-ऄऺ-़ा-ॗॢ-঄঍-঎঑-঒঩঱঳-঵঺-৛৞ৢ-৯৲-਄਋-਎਑-਒਩਱਴਷਺-੘੝੟-ੱੵ-઄ઌ઎઒઩઱઴઺-઼ા-૟ૡ-଄଍-଎଑-଒଩଱଴-ଵ଺-଼ା-୛୞ୢ-஄஋-஍஑஖-஘஛஝஠-஢஥-஧஫-஭ஶ஺-ఄ఍఑఩ఴ఺-౟ౢ-಄಍಑಩಴಺-ೝ೟ೢ-ഄ഍഑ഩഺ-ൟൢ-฀ฯัิ-฿ๆ-຀຃຅-ຆຉ຋-ຌຎ-ຓຘຠ຤຦ຨ-ຩຬຯັິ-ຼ຾-຿໅-༿཈ཪ-႟჆-჏ჷ-ჿᄁᄄᄈᄊᄍᄓ-ᄻᄽᄿᅁ-ᅋᅍᅏᅑ-ᅓᅖ-ᅘᅚ-ᅞᅢᅤᅦᅨᅪ-ᅬᅯ-ᅱᅴᅶ-ᆝᆟ-ᆧᆩ-ᆪᆬ-ᆭᆰ-ᆶᆹᆻᇃ-ᇪᇬ-ᇯᇱ-ᇸᇺ-᷿ẜ-ẟỺ-ỿ἖-἗἞-἟὆-὇὎-὏὘὚὜὞὾-὿᾵᾽᾿-῁῅῍-῏῔-῕῜-῟῭-῱῵´-℥℧-℩ℬ-ℭℯ-ⅿↃ-〆〈-〠〪-぀ゕ-゠・-㄄ㄭ-䷿龦-꯿힤-￿]z#[^ 
a-zA-Z0-9\-'()+,./:=?;!*#@$_%]c@sreZdZe�d�Zddd�Zddd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�ZdS)�
InfosetFilterz
U[\dA-F]{5,5}FTcCs.||_||_||_||_||_||_i|_dS�N)�dropXmlnsLocalName�dropXmlnsAttrNs�preventDoubleDashComments�preventDashAtCommentEnd�replaceFormFeedCharacters�preventSingleQuotePubid�replaceCache)�selfrDrErFrGrHrIrrr�__init__�szInfosetFilter.__init__NcCsL|jr |�d�r t�dt�dS|jr>|dkr>t�dt�dS|�|�SdS)Nzxmlns:z"Attributes cannot begin with xmlnszhttp://www.w3.org/2000/xmlns/z)Attributes cannot be in the xml namespace)rD�
startswith�warnings�warnrrE�	toXmlName)rK�name�	namespacerrr�coerceAttribute�s�zInfosetFilter.coerceAttributecCs
|�|�SrC)rP)rKrQrrr�
coerceElement�szInfosetFilter.coerceElementcCsJ|jrFd|vr(t�dt�|�dd�}q|�d�rFt�dt�|d7}|S)Nz--z'Comments cannot contain adjacent dashesz- -r	zComments cannot end in a dash� )rFrNrOrr?�endswith)rK�datarrr�
coerceComment�s
zInfosetFilter.coerceCommentcCs6|jr2t|�d��D]}t�dt�q|�dd�}|S)N�zText cannot contain U+000CrU)rH�range�countrNrOrr?)rKrWr
rrr�coerceCharacters�s
zInfosetFilter.coerceCharacterscCsl|}t�|�D]&}t�dt�|�|�}|�||�}q|jrh|�d�dkrht�dt�|�d|�d��}|S)NzCoercing non-XML pubid�'rz!Pubid cannot contain single quote)	�nonPubidCharRegexp�findallrNrOr�getReplacementCharacterr?rI�find)rKrWZ
dataOutputrA�replacementrrr�coercePubid�s
zInfosetFilter.coercePubidc
Cs�|d}|dd�}t�|�}|r>t�d|t�|�|�}n|}|}tt�|��}|D]*}t�d|t�|�|�}	|�	||	�}qX||S)NrrzCoercing non-XML name: %s)
�nonXmlNameFirstBMPRegexprrNrOrr`�set�nonXmlNameBMPRegexpr_r?)
rKrQ�	nameFirstZnameRest�mZnameFirstOutputZnameRestOutputZreplaceCharsrArbrrrrP�s

zInfosetFilter.toXmlNamecCs$||jvr|j|}n
|�|�}|SrC)rJ�
escapeChar�rKrArbrrrr`s

z%InfosetFilter.getReplacementCharactercCs,t|j�|��D]}|�||�|��}q|SrC)re�replacementRegexpr_r?�unescapeChar)rKrQrrrr�fromXmlNameszInfosetFilter.fromXmlNamecCsdt|�}||j|<|S)NzU%05X)rrJrjrrrris
zInfosetFilter.escapeCharcCstt|dd�d��S)Nrr')r-r0)rKZcharcoderrrrl szInfosetFilter.unescapeChar)FFFFTF)N)�__name__�
__module__�__qualname__�re�compilerkrLrSrTrXr\rcrPr`rmrirlrrrrrB�s$
�


rB)!�
__future__rrrrqrN�	constantsrZbaseCharZideographicZcombiningCharacter�digitZextenderr.�letterrQrgrrrrr"rr0r)r*r/rr,rfrdr^�objectrBrrrr�<module>s40�


	


site-packages/pip/_vendor/html5lib/__pycache__/html5parser.cpython-39.pyc000064400000261652150564745010022362 0ustar00a

��Re���@sddlmZmZmZddlmZmZddlZddlm	Z	ddlm
Z
ddlmZddlm
Z
dd	lmZdd
lmZmZmZmZmZmZmZmZmZmZmZmZmZmZmZm Z d d
d�Z!d!dd�Z"dd�Z#Gdd�de$�Z%ej&dd��Z'dd�Z(d"dd�Z)Gdd�de*�Z+dS)#�)�absolute_import�division�unicode_literals)�with_metaclass�viewkeysN�)�_inputstream)�
_tokenizer)�treebuilders)�Marker)�_utils)�spaceCharacters�asciiUpper2Lower�specialElements�headingElements�
cdataElements�rcdataElements�
tokenTypes�
tagTokenTypes�
namespaces�htmlIntegrationPointElements�"mathmlTextIntegrationPointElements�adjustForeignAttributes�adjustMathMLAttributes�adjustSVGAttributes�E�_ReparseException�etreeTcKs(t�|�}t||d�}|j|fi|��S)a�Parse an HTML document as a string or file-like object into a tree

    :arg doc: the document to parse as a string or file-like object

    :arg treebuilder: the treebuilder to use when parsing

    :arg namespaceHTMLElements: whether or not to namespace HTML elements

    :returns: parsed tree

    Example:

    >>> from html5lib.html5parser import parse
    >>> parse('<html><body><p>This is a doc</p></body></html>')
    <Element u'{http://www.w3.org/1999/xhtml}html' at 0x7feac4909db0>

    ��namespaceHTMLElements)r
�getTreeBuilder�
HTMLParser�parse)�doc�treebuilderr�kwargs�tb�p�r(��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/html5parser.pyr"s
r"�divcKs,t�|�}t||d�}|j|fd|i|��S)a#Parse an HTML fragment as a string or file-like object into a tree

    :arg doc: the fragment to parse as a string or file-like object

    :arg container: the container context to parse the fragment in

    :arg treebuilder: the treebuilder to use when parsing

    :arg namespaceHTMLElements: whether or not to namespace HTML elements

    :returns: parsed tree

    Example:

    >>> from html5lib.html5libparser import parseFragment
    >>> parseFragment('<b>this is a fragment</b>')
    <Element u'DOCUMENT_FRAGMENT' at 0x7feac484b090>

    r�	container)r
r r!�
parseFragment)r#r+r$rr%r&r'r(r(r)r,1s
r,csG�fdd�dt�}|S)NcseZdZ�fdd�ZdS)z-method_decorator_metaclass.<locals>.Decoratedcs>|��D]$\}}t|tj�r$�|�}|||<qt�||||�S�N)�items�
isinstance�types�FunctionType�type�__new__)�meta�	classname�bases�	classDict�
attributeName�	attribute��functionr(r)r3Ls

z5method_decorator_metaclass.<locals>.Decorated.__new__N)�__name__�
__module__�__qualname__r3r(r:r(r)�	DecoratedKsr?)r2)r;r?r(r:r)�method_decorator_metaclassJsr@c@s�eZdZdZd'dd�Zd(dd	�Zd
d�Zedd
��Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
d)dd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�ZdS)*r!z]HTML parser

    Generates a tree structure from a stream of (possibly malformed) HTML.

    NFTcsH|�_|durt�d�}||��_g�_�fdd�t|���D��_dS)a�
        :arg tree: a treebuilder class controlling the type of tree that will be
            returned. Built in treebuilders can be accessed through
            html5lib.treebuilders.getTreeBuilder(treeType)

        :arg strict: raise an exception when a parse error is encountered

        :arg namespaceHTMLElements: whether or not to namespace HTML elements

        :arg debug: whether or not to enable debug mode which logs things

        Example:

        >>> from html5lib.html5parser import HTMLParser
        >>> parser = HTMLParser()                     # generates parser with etree builder
        >>> parser = HTMLParser('lxml', strict=True)  # generates parser with lxml builder which is strict

        Nrcsi|]\}}||��j��qSr()�tree)�.0�name�cls��selfr(r)�
<dictcomp>y�z'HTMLParser.__init__.<locals>.<dictcomp>)�strictr
r rA�errors�	getPhasesr.�phases)rFrArIr�debugr(rEr)�__init__]s



�zHTMLParser.__init__r*cKsf||_||_||_tj|fd|i|��|_|��z|��Wn"ty`|��|��Yn0dS)N�parser)	�
innerHTMLModer+�	scriptingr	�
HTMLTokenizer�	tokenizer�reset�mainLoopr)rF�stream�	innerHTMLr+rQr%r(r(r)�_parse|szHTMLParser._parsecCs�|j��d|_g|_g|_d|_|jr�|j��|_	|j	t
vrL|jj|j_
n0|j	tvrd|jj|j_
n|j	dkr||jj|j_
n|jd|_|j��|��nd|_	|jd|_d|_d|_d|_dS)NFz	no quirks�	plaintext�
beforeHtml�initialT)rArT�
firstStartTagrJ�log�
compatModerPr+�lowerrWrrS�rcdataState�stater�rawtextState�plaintextStaterL�phase�insertHtmlElement�resetInsertionMode�	lastPhase�beforeRCDataPhase�
framesetOKrEr(r(r)rT�s*





zHTMLParser.resetcCst|d�sdS|jjjdjS)z�Name of the character encoding that was used to decode the input stream, or
        :obj:`None` if that is not determined yet

        rSNr)�hasattrrSrV�charEncodingrCrEr(r(r)�documentEncoding�s
zHTMLParser.documentEncodingcCsJ|jdkr6|jtdkr6d|jvo4|jd�t�dvS|j|jftvSdS)N�annotation-xml�mathml�encoding)z	text/htmlzapplication/xhtml+xml)rC�	namespacer�
attributes�	translaterr�rF�elementr(r(r)�isHTMLIntegrationPoint�s
�

���z!HTMLParser.isHTMLIntegrationPointcCs|j|jftvSr-)rprCrrsr(r(r)�isMathMLTextIntegrationPoint�sz'HTMLParser.isMathMLTextIntegrationPointcCsftd}td}td}td}td}td}td}|jD�]�}d}	|}
|
du�r�|
}	|jjrn|jjdnd}|r||jnd}|r�|jnd}
|
d	}||kr�|�|
d
|
�di��d}
qLt|jj�dk�sb||jj	k�sb|�
|��r||k�r|d
tddg�v�sb|||fv�sb|tdk�rF|
dk�rF||k�rF|d
dk�sb|�
|��rj||||fv�rj|j}n
|jd}||k�r�|�|
�}
qL||k�r�|�|
�}
qL||k�r�|�|
�}
qL||k�r�|�|
�}
qL||k�r�|�|
�}
qL||krL|�|
�}
qL||kr>|	dr>|	ds>|�dd
|	d
i�q>d}g}|�rb|�|j�|j��}|�r,|j|v�s,J��q,dS)N�
Characters�SpaceCharacters�StartTag�EndTag�Comment�Doctype�
ParseError���r2�data�datavarsrrC�mglyph�
malignmarkrnrm�svg�inForeignContent�selfClosing�selfClosingAcknowledgedz&non-void-element-with-trailing-solidusT)rrSrA�openElementsrprC�
parseError�get�len�defaultNamespacerv�	frozensetrrurdrL�processCharacters�processSpaceCharacters�processStartTag�
processEndTag�processComment�processDoctype�append�
processEOF)rF�CharactersToken�SpaceCharactersToken�
StartTagToken�EndTagToken�CommentToken�DoctypeToken�ParseErrorToken�token�
prev_token�	new_token�currentNode�currentNodeNamespace�currentNodeNamer2rd�	reprocessrLr(r(r)rU�s�

����
�
���	
�
��





�
�
zHTMLParser.mainLoopcOs&|j|ddg|�Ri|��|j��S)a�Parse a HTML document into a well-formed tree

        :arg stream: a file-like object or string containing the HTML to be parsed

            The optional encoding parameter must be a string that indicates
            the encoding.  If specified, that encoding will be used,
            regardless of any BOM or later declaration (such as in a meta
            element).

        :arg scripting: treat noscript elements as if JavaScript was turned on

        :returns: parsed tree

        Example:

        >>> from html5lib.html5parser import HTMLParser
        >>> parser = HTMLParser()
        >>> parser.parse('<html><body><p>This is a doc</p></body></html>')
        <Element u'{http://www.w3.org/1999/xhtml}html' at 0x7feac4909db0>

        FN)rXrA�getDocument�rFrV�argsr%r(r(r)r"szHTMLParser.parsecOs$|j|dg|�Ri|��|j��S)aZParse a HTML fragment into a well-formed tree fragment

        :arg container: name of the element we're setting the innerHTML
            property if set to None, default to 'div'

        :arg stream: a file-like object or string containing the HTML to be parsed

            The optional encoding parameter must be a string that indicates
            the encoding.  If specified, that encoding will be used,
            regardless of any BOM or later declaration (such as in a meta
            element)

        :arg scripting: treat noscript elements as if JavaScript was turned on

        :returns: parsed tree

        Example:

        >>> from html5lib.html5libparser import HTMLParser
        >>> parser = HTMLParser()
        >>> parser.parseFragment('<b>this is a fragment</b>')
        <Element u'DOCUMENT_FRAGMENT' at 0x7feac484b090>

        T)rXrA�getFragmentr�r(r(r)r,szHTMLParser.parseFragment�XXX-undefined-errorcCs@|duri}|j�|jj��||f�|jr<tt||��dSr-)rJr�rSrV�positionrIr}r)rF�	errorcoder�r(r(r)r�;s
zHTMLParser.parseErrorcCst|t�dSr-)�adjust_attributesr�rFr�r(r(r)rCsz!HTMLParser.adjustMathMLAttributescCst|t�dSr-)r�rr�r(r(r)rFszHTMLParser.adjustSVGAttributescCst|t�dSr-)r��adjustForeignAttributesMapr�r(r(r)rIsz"HTMLParser.adjustForeignAttributescCs|j��dSr-)rOrdr�r(r(r)�reparseTokenNormalLszHTMLParser.reparseTokenNormalcCs�d}ddddddddddd	d	d
dd�}|jjddd
�D]�}|j}d}||jjdkrj|js`J�d}|j}|dvr||js|J�|s�|j|jjkr�q8||vr�|j||}q�q8|r8|jd	}q�q8||_dS)NF�inSelect�inCell�inRow�inTableBody�	inCaption�
inColumnGroup�inTable�inBody�
inFrameset�
beforeHead)�select�td�th�tr�tbody�thead�tfoot�caption�colgroup�table�head�body�frameset�htmlr~rT)r�r�r�r�)rAr�rCrWrpr�rLrd)rF�last�newModes�node�nodeName�	new_phaser(r(r)rfPsD�


zHTMLParser.resetInsertionModecCsR|dvsJ�|j�|�|dkr.|jj|j_n|jj|j_|j|_|jd|_dS)N)�RAWTEXT�RCDATAr��text)	rA�
insertElementrSrbrar`rd�
originalPhaserL)rFr��contentTyper(r(r)�parseRCDataRawtext|szHTMLParser.parseRCDataRawtext)NFTF)Fr*F)r�N)r<r=r>�__doc__rNrXrT�propertyrlrurvrUr"r,r�rrrr�rfr�r(r(r(r)r!Vs$

"
	
C
,r!csdd�}dd�}Gdd�dt|||����Gdd�d��}Gd	d
�d
��}Gdd�d��}Gd
d�d��}Gdd�d��}Gdd�d��}G��fdd�d���Gdd�d��}	G�fdd�d��}
G�fdd�d���G�fdd�d��}G�fdd�d��}G�fdd �d ��}
G�fd!d"�d"��}G�fd#d$�d$��}G�fd%d&�d&��}Gd'd(�d(��}G�fd)d*�d*��}Gd+d,�d,��}G�fd-d.�d.��}G�fd/d0�d0��}Gd1d2�d2��}Gd3d4�d4��}||||||�|	|
�|||
||||||||||d5�S)6Ncs$dd�t��D����fdd�}|S)z4Logger that records which phase processes each tokencSsi|]\}}||�qSr(r()rB�key�valuer(r(r)rG�rHz*getPhases.<locals>.log.<locals>.<dictcomp>cs��j�d�r�t|�dkr�|d}d�|di}|dtvrH|d|d<|jj�|jjjj|jj	j
j|j
j�j|f��|g|�Ri|��S�|g|�Ri|��SdS)N�processrr2rC)r<�
startswithr�rrOr]r�rSrard�	__class__)rFr�r%r��info�r;�
type_namesr(r)�wrapped�s
�z'getPhases.<locals>.log.<locals>.wrapped)rr.)r;r�r(r�r)r]�szgetPhases.<locals>.logcSs|rt|�StSdSr-)r@r2)�
use_metaclass�metaclass_funcr(r(r)�getMetaclass�szgetPhases.<locals>.getMetaclassc@s\eZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
dS)zgetPhases.<locals>.PhasezNBase class for helper object that implements each phase of processing
        )rOrA�__startTagCache�
__endTagCachecSs||_||_i|_i|_dSr-)rOrA�_Phase__startTagCache�_Phase__endTagCache)rFrOrAr(r(r)rN�sz!getPhases.<locals>.Phase.__init__cSst�dSr-)�NotImplementedErrorrEr(r(r)r��sz#getPhases.<locals>.Phase.processEOFcSs|j�||jjd�dS)Nr~�rA�
insertCommentr�r�r(r(r)r��sz'getPhases.<locals>.Phase.processCommentcSs|j�d�dS)Nzunexpected-doctype�rOr�r�r(r(r)r��sz'getPhases.<locals>.Phase.processDoctypecSs|j�|d�dS�Nr�rA�
insertTextr�r(r(r)r��sz*getPhases.<locals>.Phase.processCharacterscSs|j�|d�dSr�r�r�r(r(r)r��sz/getPhases.<locals>.Phase.processSpaceCharacterscSsj|d}||jvr|j|}nD|j|}|j|<t|j�t|j�dkrb|j�tt|j���q2||�S�NrCg�������?)r��startTagHandlerr��pop�next�iter�rFr�rC�funcr(r(r)r��s
z(getPhases.<locals>.Phase.processStartTagcSsf|jjs |ddkr |j�d�|d��D],\}}||jjdjvr,||jjdj|<q,d|j_dS)NrCr�z
non-html-rootrrF)rOr\r�r.rAr�rq�rFr��attrr�r(r(r)�startTagHtml�sz%getPhases.<locals>.Phase.startTagHtmlcSsj|d}||jvr|j|}nD|j|}|j|<t|j�t|j�dkrb|j�tt|j���q2||�Sr�)r��
endTagHandlerr�r�r�r�r�r(r(r)r��s
z&getPhases.<locals>.Phase.processEndTagN)r<r=r>r��	__slots__rNr�r�r�r�r�r�r�r�r(r(r(r)�Phase�s
r�c@sReZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�ZdS)zgetPhases.<locals>.InitialPhasecSsdSr-r(r�r(r(r)r��sz6getPhases.<locals>.InitialPhase.processSpaceCharacterscSs|j�||jj�dSr-�rAr��documentr�r(r(r)r��sz.getPhases.<locals>.InitialPhase.processCommentcSs�|d}|d}|d}|d}|dks@|dus@|durL|dkrL|j�d�|durXd}|j�|�|dkrv|�t�}|r�|ddks�|�d	�s�|d
vs�|�d�r�|dus�|r�|��dkr�d
|j_n$|�d�s�|�d�r�|dur�d|j_|jj	d|j_
dS)NrC�publicId�systemId�correctr�zabout:legacy-compatzunknown-doctype�)7z*+//silmaril//dtd html pro v0r11 19970101//z4-//advasoft ltd//dtd html 3.0 aswedit + extensions//z*-//as//dtd html 3.0 aswedit + extensions//z-//ietf//dtd html 2.0 level 1//z-//ietf//dtd html 2.0 level 2//z&-//ietf//dtd html 2.0 strict level 1//z&-//ietf//dtd html 2.0 strict level 2//z-//ietf//dtd html 2.0 strict//z-//ietf//dtd html 2.0//z-//ietf//dtd html 2.1e//z-//ietf//dtd html 3.0//z-//ietf//dtd html 3.2 final//z-//ietf//dtd html 3.2//z-//ietf//dtd html 3//z-//ietf//dtd html level 0//z-//ietf//dtd html level 1//z-//ietf//dtd html level 2//z-//ietf//dtd html level 3//z"-//ietf//dtd html strict level 0//z"-//ietf//dtd html strict level 1//z"-//ietf//dtd html strict level 2//z"-//ietf//dtd html strict level 3//z-//ietf//dtd html strict//z-//ietf//dtd html//z(-//metrius//dtd metrius presentational//z5-//microsoft//dtd internet explorer 2.0 html strict//z.-//microsoft//dtd internet explorer 2.0 html//z0-//microsoft//dtd internet explorer 2.0 tables//z5-//microsoft//dtd internet explorer 3.0 html strict//z.-//microsoft//dtd internet explorer 3.0 html//z0-//microsoft//dtd internet explorer 3.0 tables//z#-//netscape comm. corp.//dtd html//z*-//netscape comm. corp.//dtd strict html//z*-//o'reilly and associates//dtd html 2.0//z3-//o'reilly and associates//dtd html extended 1.0//z;-//o'reilly and associates//dtd html extended relaxed 1.0//zN-//softquad software//dtd hotmetal pro 6.0::19990601::extensions to html 4.0//zE-//softquad//dtd hotmetal pro 4.0::19971010::extensions to html 4.0//z$-//spyglass//dtd html 2.0 extended//z+-//sq//dtd html 2.0 hotmetal + extensions//z--//sun microsystems corp.//dtd hotjava html//z4-//sun microsystems corp.//dtd hotjava strict html//z-//w3c//dtd html 3 1995-03-24//z-//w3c//dtd html 3.2 draft//z-//w3c//dtd html 3.2 final//z-//w3c//dtd html 3.2//z-//w3c//dtd html 3.2s draft//z-//w3c//dtd html 4.0 frameset//z#-//w3c//dtd html 4.0 transitional//z(-//w3c//dtd html experimental 19960712//z&-//w3c//dtd html experimental 970421//z-//w3c//dtd w3 html//z-//w3o//dtd w3 html 3.0//z#-//webtechs//dtd mozilla html 2.0//z-//webtechs//dtd mozilla html//)z$-//w3o//dtd w3 html strict 3.0//en//z"-/w3c/dtd html 4.0 transitional/enr�)z -//w3c//dtd html 4.01 frameset//z$-//w3c//dtd html 4.01 transitional//z:http://www.ibm.com/data/dtd/v11/ibmxhtml1-transitional.dtd�quirks)z -//w3c//dtd xhtml 1.0 frameset//z$-//w3c//dtd xhtml 1.0 transitional//zlimited quirksrZ)rOr�rA�
insertDoctyperrrr�r_r^rLrd)rFr�rCr�r�r�r(r(r)r��sX��
��9�<��?�@�@
�A
����z.getPhases.<locals>.InitialPhase.processDoctypecSsd|j_|jjd|j_dS)Nr�rZ)rOr^rLrdrEr(r(r)�anythingElseZsz,getPhases.<locals>.InitialPhase.anythingElsecSs|j�d�|��|S)Nzexpected-doctype-but-got-chars�rOr�rr�r(r(r)r�^sz1getPhases.<locals>.InitialPhase.processCharacterscSs"|j�dd|di�|��|S)Nz"expected-doctype-but-got-start-tagrCrr�r(r(r)r�cs

�z/getPhases.<locals>.InitialPhase.processStartTagcSs"|j�dd|di�|��|S)Nz expected-doctype-but-got-end-tagrCrr�r(r(r)r�is

�z-getPhases.<locals>.InitialPhase.processEndTagcSs|j�d�|��dS)Nzexpected-doctype-but-got-eofTrrEr(r(r)r�osz*getPhases.<locals>.InitialPhase.processEOFN)
r<r=r>�tupler�r�r�r�rr�r�r�r�r(r(r(r)�InitialPhase�s_rc@sJeZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�ZdS)z"getPhases.<locals>.BeforeHtmlPhasecSs&|j�tdd��|jjd|j_dS)Nr�ryr�)rA�
insertRoot�impliedTagTokenrOrLrdrEr(r(r)rexsz4getPhases.<locals>.BeforeHtmlPhase.insertHtmlElementcSs|��dS�NT�rerEr(r(r)r�}sz-getPhases.<locals>.BeforeHtmlPhase.processEOFcSs|j�||jj�dSr-r�r�r(r(r)r��sz1getPhases.<locals>.BeforeHtmlPhase.processCommentcSsdSr-r(r�r(r(r)r��sz9getPhases.<locals>.BeforeHtmlPhase.processSpaceCharacterscSs|��|Sr-rr�r(r(r)r��sz4getPhases.<locals>.BeforeHtmlPhase.processCharacterscSs |ddkrd|j_|��|S)NrCr�T)rOr\rer�r(r(r)r��sz2getPhases.<locals>.BeforeHtmlPhase.processStartTagcSs4|ddvr$|j�dd|di�n|��|SdS)NrC�r�r�r��brzunexpected-end-tag-before-html)rOr�rer�r(r(r)r��s
�z0getPhases.<locals>.BeforeHtmlPhase.processEndTagN)r<r=r>rr�rer�r�r�r�r�r�r(r(r(r)�BeforeHtmlPhasetsr
c@s�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Ze
�defde	fg�Ze
e_e
�defg�Zee_dS)z"getPhases.<locals>.BeforeHeadPhasecSs|�tdd��dS)Nr�ryT��startTagHeadrrEr(r(r)r��sz-getPhases.<locals>.BeforeHeadPhase.processEOFcSsdSr-r(r�r(r(r)r��sz9getPhases.<locals>.BeforeHeadPhase.processSpaceCharacterscSs|�tdd��|S�Nr�ryrr�r(r(r)r��sz4getPhases.<locals>.BeforeHeadPhase.processCharacterscSs|jjd�|�S�Nr��rOrLr�r�r(r(r)r��sz/getPhases.<locals>.BeforeHeadPhase.startTagHtmlcSs0|j�|�|jjd|j_|jjd|j_dS)Nr~�inHead)rAr�r��headPointerrOrLrdr�r(r(r)r�sz/getPhases.<locals>.BeforeHeadPhase.startTagHeadcSs|�tdd��|Sr
rr�r(r(r)�
startTagOther�sz0getPhases.<locals>.BeforeHeadPhase.startTagOthercSs|�tdd��|Sr
rr�r(r(r)�endTagImplyHead�sz2getPhases.<locals>.BeforeHeadPhase.endTagImplyHeadcSs|j�dd|di�dS)Nzend-tag-after-implied-rootrCr�r�r(r(r)�endTagOther�s
�z.getPhases.<locals>.BeforeHeadPhase.endTagOtherr�r�rN)r<r=r>rr�r�r�r�r�rrrrr�MethodDispatcherr��defaultr�r(r(r(r)�BeforeHeadPhase�s$��rc@s�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Ze�defd efd!efd"e
fd#efd$e	fd%e
fd&efg�Zee_e�d&efd'efg�Zee_d(S))zgetPhases.<locals>.InHeadPhasecSs|��dSr�rrEr(r(r)r��sz)getPhases.<locals>.InHeadPhase.processEOFcSs|��|Sr-rr�r(r(r)r��sz0getPhases.<locals>.InHeadPhase.processCharacterscSs|jjd�|�Srrr�r(r(r)r��sz+getPhases.<locals>.InHeadPhase.startTagHtmlcSs|j�d�dS)Nz!two-heads-are-not-better-than-oner�r�r(r(r)r�sz+getPhases.<locals>.InHeadPhase.startTagHeadcSs$|j�|�|jj��d|d<dS�NTr��rAr�r�r�r�r(r(r)�startTagBaseLinkCommand�sz6getPhases.<locals>.InHeadPhase.startTagBaseLinkCommandcSs�|j�|�|jj��d|d<|d}|jjjjddkr�d|vrZ|jjj�|d�nVd|vr�d|vr�|d�	�d	kr�t
�|d�d
��}t
�
|�}|��}|jjj�|�dS)NTr�rr�	tentative�charset�contentz
http-equivzcontent-typezutf-8)rAr�r�r�rOrSrVrk�changeEncodingr_r�
EncodingBytes�encode�ContentAttrParserr")rFr�rqrrO�codecr(r(r)�startTagMeta�s ��
z+getPhases.<locals>.InHeadPhase.startTagMetacSs|j�|d�dS)Nr��rOr�r�r(r(r)�
startTagTitle�sz,getPhases.<locals>.InHeadPhase.startTagTitlecSs|j�|d�dS)Nr�r%r�r(r(r)�startTagNoFramesStyle�sz4getPhases.<locals>.InHeadPhase.startTagNoFramesStylecSs8|jjr|j�|d�n|j�|�|jjd|j_dS)Nr��inHeadNoscript)rOrQr�rAr�rLrdr�r(r(r)�startTagNoscript�sz/getPhases.<locals>.InHeadPhase.startTagNoscriptcSs<|j�|�|jjj|jj_|jj|j_|jjd|j_dS)Nr�)	rAr�rOrS�scriptDataStaterardr�rLr�r(r(r)�startTagScriptsz-getPhases.<locals>.InHeadPhase.startTagScriptcSs|��|Sr-rr�r(r(r)rsz,getPhases.<locals>.InHeadPhase.startTagOthercSs:|jjj��}|jdks&Jd|j��|jjd|j_dS)Nr�zExpected head got %s�	afterHead�rOrAr�r�rCrLrd�rFr�r�r(r(r)�
endTagHead
sz)getPhases.<locals>.InHeadPhase.endTagHeadcSs|��|Sr-rr�r(r(r)�endTagHtmlBodyBrsz/getPhases.<locals>.InHeadPhase.endTagHtmlBodyBrcSs|j�dd|di�dS�N�unexpected-end-tagrCr�r�r(r(r)rsz*getPhases.<locals>.InHeadPhase.endTagOthercSs|�td��dS)Nr�)r/rrEr(r(r)rsz+getPhases.<locals>.InHeadPhase.anythingElser��title)�noframes�style�noscript�script)�base�basefont�bgsound�command�linkr4r�)r	r�r�N)r<r=r>rr�r�r�r�rrr$r&r'r)r+rr/r0rrrrr�rr�r(r(r(r)�InHeadPhase�sD���r=c@s�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Ze�de	fde
fdefg�Zee_e�de
fdefg�Zee_dS)z&getPhases.<locals>.InHeadNoscriptPhasecSs|j�d�|��dS)Nzeof-in-head-noscriptTrrEr(r(r)r�/sz1getPhases.<locals>.InHeadNoscriptPhase.processEOFcSs|jjd�|�S�Nr)rOrLr�r�r(r(r)r�4sz5getPhases.<locals>.InHeadNoscriptPhase.processCommentcSs|j�d�|��|S)Nzchar-in-head-noscriptrr�r(r(r)r�7sz8getPhases.<locals>.InHeadNoscriptPhase.processCharacterscSs|jjd�|�Sr>�rOrLr�r�r(r(r)r�<sz=getPhases.<locals>.InHeadNoscriptPhase.processSpaceCharacterscSs|jjd�|�Srrr�r(r(r)r�?sz3getPhases.<locals>.InHeadNoscriptPhase.startTagHtmlcSs|jjd�|�Sr>rr�r(r(r)rBsz>getPhases.<locals>.InHeadNoscriptPhase.startTagBaseLinkCommandcSs|j�dd|di�dS�N�unexpected-start-tagrCr�r�r(r(r)�startTagHeadNoscriptEsz;getPhases.<locals>.InHeadNoscriptPhase.startTagHeadNoscriptcSs"|j�dd|di�|��|S�Nzunexpected-inhead-noscript-tagrCrr�r(r(r)rHsz4getPhases.<locals>.InHeadNoscriptPhase.startTagOthercSs:|jjj��}|jdks&Jd|j��|jjd|j_dS)Nr6zExpected noscript got %srr-r.r(r(r)�endTagNoscriptMsz5getPhases.<locals>.InHeadNoscriptPhase.endTagNoscriptcSs"|j�dd|di�|��|SrCrr�r(r(r)�endTagBrRsz/getPhases.<locals>.InHeadNoscriptPhase.endTagBrcSs|j�dd|di�dSr1r�r�r(r(r)rWsz2getPhases.<locals>.InHeadNoscriptPhase.endTagOthercSs|�td��dS)Nr6)rDrrEr(r(r)rZsz3getPhases.<locals>.InHeadNoscriptPhase.anythingElser�)r9r:r<r4r4r5)r�r6r6r	N)r<r=r>rr�r�r�r�r�r�rrBrrDrErrrrr�rr�r(r(r(r)�InHeadNoscriptPhase,s0��rFc@s�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Ze�defdefde	fde
fdefg�Zee_e�de
fg�Zee_dS)z!getPhases.<locals>.AfterHeadPhasecSs|��dSrrrEr(r(r)r�nsz,getPhases.<locals>.AfterHeadPhase.processEOFcSs|��|Sr-rr�r(r(r)r�rsz3getPhases.<locals>.AfterHeadPhase.processCharacterscSs|jjd�|�Srrr�r(r(r)r�vsz.getPhases.<locals>.AfterHeadPhase.startTagHtmlcSs(d|j_|j�|�|jjd|j_dS)NFr�)rOrirAr�rLrdr�r(r(r)�startTagBodyysz.getPhases.<locals>.AfterHeadPhase.startTagBodycSs |j�|�|jjd|j_dS)Nr�)rAr�rOrLrdr�r(r(r)�startTagFrameset~sz2getPhases.<locals>.AfterHeadPhase.startTagFramesetcSsr|j�dd|di�|jj�|jj�|jjd�|�|jjddd�D] }|jdkrL|jj�	|�qnqLdS)Nz#unexpected-start-tag-out-of-my-headrCrr~r�)
rOr�rAr�r�rrLr�rC�remover.r(r(r)�startTagFromHead�s
�
z2getPhases.<locals>.AfterHeadPhase.startTagFromHeadcSs|j�dd|di�dSr@r�r�r(r(r)r�sz.getPhases.<locals>.AfterHeadPhase.startTagHeadcSs|��|Sr-rr�r(r(r)r�sz/getPhases.<locals>.AfterHeadPhase.startTagOthercSs|��|Sr-rr�r(r(r)r0�sz2getPhases.<locals>.AfterHeadPhase.endTagHtmlBodyBrcSs|j�dd|di�dSr1r�r�r(r(r)r�sz-getPhases.<locals>.AfterHeadPhase.endTagOthercSs.|j�tdd��|jjd|j_d|j_dS)Nr�ryr�T)rAr�rrOrLrdrirEr(r(r)r�sz.getPhases.<locals>.AfterHeadPhase.anythingElser�r�r�)	r8r9r:r<r4r4r7r5r3r�)r�r�r	N)r<r=r>rr�r�r�r�rGrHrJrrr0rrrrr�rr�r(r(r(r)�AfterHeadPhaseks4
��	�rKc$s�eZdZdZ��fdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zd2d3�Zd4d5�Zd6d7�Zd8d9�Zd:d;�Z d<d=�Z!d>d?�Z"d@dA�Z#dBdC�Z$dDdE�Z%dFdG�Z&dHdI�Z'dJdK�Z(dLdM�Z)dNdO�Z*dPdQ�Z+dRdS�Z,dTdU�Z-dVdW�Z.dXdY�Z/dZd[�Z0d\d]�Z1d^d_�Z2d`da�Z3dbdc�Z4ddde�Z5dfdg�Z6e7�8dh�j9fdiefdjefdke
fdlefe:efdmefdnefdoefdpefdqefdrefdsefdtefduefdvefdwefdxefdyefdzefd{efd|efd}e fd~e!fde"fd�e#fd�e$fd�e&fd�e'fd�e%fd�e(fd�e)fd�e*fg!�Z;e+e;_<e7�8dje-fdhe.fd�e/fdne0fd�e,fd�e1fe:e2fd�e3fdue4fd�e5fg
�Z=e6e=_<�Z>S)�zgetPhases.<locals>.InBodyPhase)r�cs"t�|�j|i|��|j|_dSr-)�superrN�processSpaceCharactersNonPrer��rFr�r%)�InBodyPhaser�r(r)rN�sz'getPhases.<locals>.InBodyPhase.__init__cSs$|j|jko"|j|jko"|j|jkSr-)rCrprq)rF�node1�node2r(r(r)�isMatchingFormattingElement�s

�
�z:getPhases.<locals>.InBodyPhase.isMatchingFormattingElementcSs�|j�|�|jjd}g}|jjddd�D](}|tur@qXq.|�||�r.|�|�q.t|�dkshJ�t|�dkr�|jj�|d�|jj�|�dS)Nr~�)	rAr�r��activeFormattingElementsrrRr�r�rI)rFr�rt�matchingElementsr�r(r(r)�addFormattingElement�sz3getPhases.<locals>.InBodyPhase.addFormattingElementcSs>td�}|jjddd�D]}|j|vr|j�d�q:qdS)N)�dd�dt�lir'r�r�r�r�r�r�r�r�r~z expected-closing-tag-but-got-eof)r�rAr�rCrOr�)rF�allowed_elementsr�r(r(r)r��s

z)getPhases.<locals>.InBodyPhase.processEOFcSsf|d}|j|_|�d�rH|jjdjdvrH|jjd��sH|dd�}|rb|j��|j�|�dS)Nr�
r~)�pre�listing�textarear)	rMr�r�rAr�rC�
hasContent�#reconstructActiveFormattingElementsr�)rFr�rr(r(r)�!processSpaceCharactersDropNewline�s
��
z@getPhases.<locals>.InBodyPhase.processSpaceCharactersDropNewlinecSsT|ddkrdS|j��|j�|d�|jjrPtdd�|dD��rPd|j_dS)Nr�cSsg|]}|tv�qSr(�r
�rB�charr(r(r)�
<listcomp>�s�zDgetPhases.<locals>.InBodyPhase.processCharacters.<locals>.<listcomp>F)rAr`r�rOri�anyr�r(r(r)r��s
��z0getPhases.<locals>.InBodyPhase.processCharacterscSs|j��|j�|d�dSr�)rAr`r�r�r(r(r)rM�s
z;getPhases.<locals>.InBodyPhase.processSpaceCharactersNonPrecSs|jjd�|�Sr>rr�r(r(r)�startTagProcessInHead�sz4getPhases.<locals>.InBodyPhase.startTagProcessInHeadcSs�|j�dddi�t|jj�dks4|jjdjdkrB|jjs�J�nBd|j_|d��D],\}}||jjdj	vrV||jjdj	|<qVdS)NrArCr�rFr)
rOr�r�rAr�rCrWrir.rqr�r(r(r)rG�s�z+getPhases.<locals>.InBodyPhase.startTagBodycSs�|j�dddi�t|jj�dks4|jjdjdkrB|jjs�J�np|jjsLnf|jjdjrv|jjdj�	|jjd�|jjdjdkr�|jj�
�qv|j�|�|jjd|j_
dS)	NrArCr�rr�r~r�r�)rOr�r�rAr�rCrWri�parent�removeChildr�r�rLrdr�r(r(r)rHs"z/getPhases.<locals>.InBodyPhase.startTagFramesetcSs.|jjddd�r|�td��|j�|�dS�Nr'�button��variant)rA�elementInScope�endTagPrr�r�r(r(r)�startTagClosePsz-getPhases.<locals>.InBodyPhase.startTagClosePcSs>|jjddd�r|�td��|j�|�d|j_|j|_dS)Nr'rlrmF)	rArorprr�rOrirar�r�r(r(r)�startTagPreListings
z1getPhases.<locals>.InBodyPhase.startTagPreListingcSsZ|jjr|j�dddi�n:|jjddd�r:|�td��|j�|�|jjd|j_dS)NrArC�formr'rlrmr~)	rA�formPointerrOr�rorprr�r�r�r(r(r)�startTagFormsz+getPhases.<locals>.InBodyPhase.startTagFormcSs�d|j_dgddgddgd�}||d}t|jj�D]@}|j|vr^|jj�t|jd��qx|j	t
vr6|jdvr6qxq6|jjd	d
d�r�|jj�td	d��|j�|�dS)NFrYrXrW)rYrXrWrCrz)�addressr*r'r'rlrm)
rOri�reversedrAr�rCrdr�r�	nameTuplerror�)rFr��stopNamesMap�	stopNamesr�r(r(r)�startTagListItem&s*�

�
��z/getPhases.<locals>.InBodyPhase.startTagListItemcSs>|jjddd�r|�td��|j�|�|jjj|jj_dSrk)	rArorprr�rOrSrcrar�r(r(r)�startTagPlaintext<sz0getPhases.<locals>.InBodyPhase.startTagPlaintextcSsb|jjddd�r|�td��|jjdjtvrR|j�dd|di�|jj�	�|j�
|�dS)Nr'rlrmr~rArC)rArorprr�rCrrOr�r�r�r�r(r(r)�startTagHeadingBsz.getPhases.<locals>.InBodyPhase.startTagHeadingcSs~|j�d�}|rf|j�dddd��|�td��||jjvrL|jj�|�||jjvrf|jj�|�|j�	�|�
|�dS)N�a�$unexpected-start-tag-implies-end-tag��	startName�endName)rA�!elementInActiveFormattingElementsrOr��endTagFormattingrr�rIrTr`rV)rFr��afeAElementr(r(r)�	startTagAJs�
z(getPhases.<locals>.InBodyPhase.startTagAcSs|j��|�|�dSr-)rAr`rVr�r(r(r)�startTagFormattingWs
z1getPhases.<locals>.InBodyPhase.startTagFormattingcSsP|j��|j�d�rB|j�dddd��|�td��|j��|�|�dS)N�nobrrr�)rAr`rorOr�r�rrVr�r(r(r)�startTagNobr[s
�
z+getPhases.<locals>.InBodyPhase.startTagNobrcSsT|j�d�r2|j�dddd��|�td��|S|j��|j�|�d|j_dS)Nrlrr�F)	rArorOr�r�rr`r�rir�r(r(r)�startTagButtones�
z-getPhases.<locals>.InBodyPhase.startTagButtoncSs0|j��|j�|�|jj�t�d|j_dS�NF)rAr`r�rTr�rrOrir�r(r(r)�startTagAppletMarqueeObjectps
z:getPhases.<locals>.InBodyPhase.startTagAppletMarqueeObjectcSsB|jjddd�r|�td��|j��d|j_|j�|d�dS)Nr'rlrmFr�)rArorprr`rOrir�r�r(r(r)�startTagXmpvs

z*getPhases.<locals>.InBodyPhase.startTagXmpcSsR|jjdkr*|jjddd�r*|�td��|j�|�d|j_|jjd|j_	dS)Nr�r'rlrmFr�)
rOr^rAror�rr�rirLrdr�r(r(r)�
startTagTable}sz,getPhases.<locals>.InBodyPhase.startTagTablecSs6|j��|j�|�|jj��d|d<d|j_dS)NTr�F)rAr`r�r�r�rOrir�r(r(r)�startTagVoidFormatting�s

z5getPhases.<locals>.InBodyPhase.startTagVoidFormattingcSs@|jj}|�|�d|dvr<|dd�t�dkr<||j_dS)Nr2r�hidden)rOrir�rrr)rFr�rir(r(r)�
startTagInput�s
�z,getPhases.<locals>.InBodyPhase.startTagInputcSs$|j�|�|jj��d|d<dSrrr�r(r(r)�startTagParamSource�sz2getPhases.<locals>.InBodyPhase.startTagParamSourcecSsJ|jjddd�r|�td��|j�|�|jj��d|d<d|j_dS)Nr'rlrmTr�F)	rArorprr�r�r�rOrir�r(r(r)�
startTagHr�sz)getPhases.<locals>.InBodyPhase.startTagHrcSs6|j�dddd��|�tdd|d|dd��dS)	Nzunexpected-start-tag-treated-as�image�img��originalName�newNameryrr��rqr�)rOr�r�rr�r(r(r)�
startTagImage�s�
�z,getPhases.<locals>.InBodyPhase.startTagImagecSs|j�dddi�|jjrdSi}d|dvr>|dd|d<|�tdd|d��|�td	d��|�td
d��d|dvr�|dd}nd}|�td
|d��|d��}d|vr�|d=d|vr�|d=d|d<|�tdd||dd��|�	td
��|�td	d��|�	td��dS)Nzdeprecated-tagrC�isindex�actionrrsry)rq�hr�label�promptz3This is a searchable index. Enter search keywords: rw�r2r�inputr�r�)
rOr�rArtr�rr�r�copyr�)rFr��
form_attrsr�rqr(r(r)�startTagIsIndex�s<
��
�z.getPhases.<locals>.InBodyPhase.startTagIsIndexcSs0|j�|�|jjj|jj_|j|_d|j_dSr�)	rAr�rOrSr`rarar�rir�r(r(r)�startTagTextarea�sz/getPhases.<locals>.InBodyPhase.startTagTextareacSsd|j_|�|�dSr�)rOri�startTagRawtextr�r(r(r)�startTagIFrame�sz-getPhases.<locals>.InBodyPhase.startTagIFramecSs"|jjr|�|�n
|�|�dSr-)rOrQr�rr�r(r(r)r)�sz/getPhases.<locals>.InBodyPhase.startTagNoscriptcSs|j�|d�dS)z8iframe, noembed noframes, noscript(if scripting enabled)r�Nr%r�r(r(r)r��sz.getPhases.<locals>.InBodyPhase.startTagRawtextcSs@|jjdjdkr$|jj�td��|j��|jj�|�dS�Nr~�option)	rAr�rCrOrdr�rr`r�r�r(r(r)�startTagOpt�s
z*getPhases.<locals>.InBodyPhase.startTagOptcSs�|j��|j�|�d|j_|jj|jjd|jjd|jjd|jjd|jjd|jjdfvrx|jjd|j_n|jjd	|j_dS)
NFr�r�r�r�r�r��inSelectInTabler�)rAr`r�rOrirdrLr�r(r(r)�startTagSelect�s





�z-getPhases.<locals>.InBodyPhase.startTagSelectcSsB|j�d�r2|j��|jjdjdkr2|j��|j�|�dS)N�rubyr~)rAro�generateImpliedEndTagsr�rCrOr�r�r�r(r(r)�startTagRpRt�s


z+getPhases.<locals>.InBodyPhase.startTagRpRtcSsZ|j��|j�|�|j�|�td|d<|j�|�|drV|jj��d|d<dS)Nrnrpr�Tr�)	rAr`rOrrrr�r�r�r�r(r(r)�startTagMath�s
z+getPhases.<locals>.InBodyPhase.startTagMathcSsZ|j��|j�|�|j�|�td|d<|j�|�|drV|jj��d|d<dS)Nr�rpr�Tr�)	rAr`rOrrrr�r�r�r�r(r(r)�startTagSvgs
z*getPhases.<locals>.InBodyPhase.startTagSvgcSs|j�dd|di�dS)a5 Elements that should be children of other elements that have a
            different insertion mode; here they are ignored
            "caption", "col", "colgroup", "frame", "frameset", "head",
            "option", "optgroup", "tbody", "td", "tfoot", "th", "thead",
            "tr", "noscript"
            zunexpected-start-tag-ignoredrCNr�r�r(r(r)�startTagMisplacedsz0getPhases.<locals>.InBodyPhase.startTagMisplacedcSs|j��|j�|�dSr-)rAr`r�r�r(r(r)rs
z,getPhases.<locals>.InBodyPhase.startTagOthercSs�|jjddd�sD|�tdd��|j�dddi�|�tdd��nT|j�d�|jjdj	dkrt|j�dddi�|jj�
�}|j	dkr�|jj�
�}q�dS)	Nr'rlrmryr2rCrzr~)rArorqrrOr�rpr�r�rCr�r.r(r(r)rps
z&getPhases.<locals>.InBodyPhase.endTagPcSs~|j�d�s|j��dS|jjdjdkrj|jjdd�D],}|jtd�vr<|j�dd|jd��qjq<|jjd|j_dS)Nr�r~�)rWrXrY�optgroupr�r'�rp�rtr�r�r�r�r�r�r�r��$expected-one-end-tag-but-got-another��gotName�expectedName�	afterBody)	rArorOr�r�rCr�rLrdr.r(r(r)�
endTagBody)s

�z)getPhases.<locals>.InBodyPhase.endTagBodycSs"|j�d�r|�td��|SdS)Nr�)rAror�rr�r(r(r)�
endTagHtml;sz)getPhases.<locals>.InBodyPhase.endTagHtmlcSs�|ddkr|j|_|j�|d�}|r2|j��|jjdj|dkr^|j�dd|di�|r�|jj�	�}|j|dkr�|jj�	�}qndS)NrCr\r~�end-tag-too-early)
rMr�rAror�r�rCrOr�r�)rFr��inScoper�r(r(r)�endTagBlockAs
z*getPhases.<locals>.InBodyPhase.endTagBlockcSsv|jj}d|j_|dus$|j�|�s8|j�dddi�n:|j��|jjd|krd|j�dddi�|jj�|�dS)Nr2rCrsr~zend-tag-too-early-ignored)rArtrorOr�r�r�rIr.r(r(r)�
endTagFormOs�
�z)getPhases.<locals>.InBodyPhase.endTagFormcSs�|ddkrd}nd}|jj|d|d�sB|j�dd|di�nf|jj|dd�|jjdj|dkr�|j�dd|di�|jj��}|j|dkr�|jj��}q�dS)	NrCrY�listrmr2��excluder~r�)rArorOr�r�r�rCr�)rFr�rnr�r(r(r)�endTagListItem\s
�z-getPhases.<locals>.InBodyPhase.endTagListItemcSs�tD]}|j�|�r|j��q$q|jjdj|dkrP|j�dd|di�tD]8}|j�|�rT|jj��}|jtvr�|jj��}qpq�qTdS)Nr~rCr�)	rrAror�r�rCrOr�r�)rFr��itemr(r(r)�
endTagHeadingms

z,getPhases.<locals>.InBodyPhase.endTagHeadingcSsd}|dk�r
|d7}|j�|d�}|rD||jjvrR|j�|j�sR|�|�dS||jjvr�|j�dd|di�|jj�	|�dS|j�|j�s�|j�dd|di�dS||jjdkr�|j�d	d|di�|jj�
|�}d}|jj|d�D]}|jtvr�|}�qq�|du�rR|jj�
�}||k�r@|jj�
�}�q&|jj�	|�dS|jj|d}|jj�
|�}|}	}
d}|jj�
|
�}|d
k�rX|d7}|d8}|jj|}
|
|jjv�r�|jj�	|
��q�|
|k�rސqX|	|k�r�|jj�
|
�d}|
��}
|
|jj|jj�
|
�<|
|jj|jj�
|
�<|
}
|	j�rF|	j�|	�|
�|	�|
}	�q�|	j�rl|	j�|	�|jtd�v�r�|j��\}}|�|	|�n
|�|	�|��}
|�|
�|�|
�|jj�	|�|jj�||
�|jj�	|�|jj�|jj�
|�d|
�qdS)z)The much-feared adoption agency algorithmr�rrCNzadoption-agency-1.2zadoption-agency-4.4r~zadoption-agency-1.3rS�r�r�r�r�r�)rAr�r�rorCrrOr�rTrI�indexrxrr��	cloneNoderirj�appendChildr��getTableMisnestedNodePosition�insertBefore�reparentChildren�insert)rFr��outerLoopCounter�formattingElement�afeIndex�
furthestBlockrt�commonAncestor�bookmark�lastNoder��innerLoopCounterr��clonerir�r(r(r)r�|s�

�
��







����



�z/getPhases.<locals>.InBodyPhase.endTagFormattingcSs�|j�|d�r|j��|jjdj|dkrF|j�dd|di�|j�|d�r�|jj��}|j|dkr~|jj��}qb|j��dS)NrCr~r�)	rAror�r�rCrOr�r��clearActiveFormattingElements)rFr�rtr(r(r)�endTagAppletMarqueeObjects
z8getPhases.<locals>.InBodyPhase.endTagAppletMarqueeObjectcSs@|j�dddd��|j��|j�tdd��|jj��dS)Nzunexpected-end-tag-treated-asr	z
br elementr�ry)rOr�rAr`r�rr�r�r�r(r(r)rE+s�
z'getPhases.<locals>.InBodyPhase.endTagBrcSs�|jjddd�D]�}|j|dkrz|jj|dd�|jjdj|dkrb|j�dd|di�|jj��|krtqbq�q|jtvr|j�dd|di�q�qdS)Nr~rCr�r2)	rAr�rCr�rOr�r�rxrr.r(r(r)r2s
z*getPhases.<locals>.InBodyPhase.endTagOtherr�)	r8r9r:r;r<r4r7r5r3r�r�)rv�article�aside�
blockquote�center�details�dirr*�dl�fieldset�
figcaption�figure�footer�header�hgroup�main�menu�nav�olr'�section�summary�ul)r\r]rs)rYrWrXrYr~)�b�big�code�em�font�i�s�small�strike�strong�tt�ur�rl)�applet�marquee�object�xmpr�)�arear	�embedr��keygen�wbr)�param�source�trackr�r�r�r�r^�iframer6)�noembedr4r�)r�r�)r�r��mathr�)r��colr��framer�r�r�r�r�r�r�)rvr�r�r�rlr�r��dialogr�r*r�r�r�r�r�r�r�r]r�r�r�r�r\r�r�r�r')rWrXrY)r~r�r�r�r�r�r�r�r�r�r�r�r�r�r	)?r<r=r>r�rNrRrVr�rar�rMrhrGrHrqrrrur{r|r}r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r)r�r�r�r�r�r�r�rrpr�r�r�r�r�r�r�r�rErrrr�rr�rr��
__classcell__r()rOr��r�r)rO�s�

	

	

$������-���rOc@s`eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	e
�g�Zee_
e
�defg�Ze	e_
dS)
zgetPhases.<locals>.TextPhasecSs|j�|d�dSr�r�r�r(r(r)r��sz.getPhases.<locals>.TextPhase.processCharacterscSs8|j�dd|jjdji�|jj��|jj|j_dS)Nz&expected-named-closing-tag-but-got-eofrCr~T)rOr�rAr�rCr�r�rdrEr(r(r)r��s�z'getPhases.<locals>.TextPhase.processEOFcSsdsJd|d��dS)NFz4Tried to process start tag %s in RCDATA/RAWTEXT moderCr(r�r(r(r)r�sz*getPhases.<locals>.TextPhase.startTagOthercSs*|jj��}|jdksJ�|jj|j_dS)Nr7)rAr�r�rCrOr�rdr.r(r(r)�endTagScript�sz)getPhases.<locals>.TextPhase.endTagScriptcSs|jj��|jj|j_dSr-)rAr�r�rOr�rdr�r(r(r)r�sz(getPhases.<locals>.TextPhase.endTagOtherr7N)r<r=r>rr�r�r�rrrrrr�rr�r(r(r(r)�	TextPhase�s
�rc
seZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Ze�d%�jfd&e
fd'efd(efd)e
fd*efd+efd,efd-efd.efg
�Zee_e�d+efd/efg�Zee_d0S)1zgetPhases.<locals>.InTablePhasecSs$|jjdjdvr |jj��qdS)Nr~)r�r�)rAr�rCr�rEr(r(r)�clearStackToTableContext�sz8getPhases.<locals>.InTablePhase.clearStackToTableContextcSs0|jjdjdkr |j�d�n|jjs,J�dS)Nr~r�zeof-in-table�rAr�rCrOr�rWrEr(r(r)r��sz*getPhases.<locals>.InTablePhase.processEOFcSs4|jj}|jjd|j_||jj_|jj�|�dS�N�inTableText)rOrdrLr�r��rFr�r�r(r(r)r��s
z6getPhases.<locals>.InTablePhase.processSpaceCharacterscSs4|jj}|jjd|j_||jj_|jj�|�dSr)rOrdrLr�r�rr(r(r)r��s
z1getPhases.<locals>.InTablePhase.processCharacterscSs&d|j_|jjd�|�d|j_dS)NTr�F)rA�insertFromTablerOrLr�r�r(r(r)r��sz*getPhases.<locals>.InTablePhase.insertTextcSs6|��|jj�t�|j�|�|jjd|j_dS)Nr�)	rrArTr�rr�rOrLrdr�r(r(r)�startTagCaption�sz/getPhases.<locals>.InTablePhase.startTagCaptioncSs(|��|j�|�|jjd|j_dS)Nr��rrAr�rOrLrdr�r(r(r)�startTagColgroup�sz0getPhases.<locals>.InTablePhase.startTagColgroupcSs|�tdd��|S)Nr�ry)rrr�r(r(r)�startTagCol�sz+getPhases.<locals>.InTablePhase.startTagColcSs(|��|j�|�|jjd|j_dS�Nr�rr�r(r(r)�startTagRowGroup�sz0getPhases.<locals>.InTablePhase.startTagRowGroupcSs|�tdd��|S)Nr�ry)rrr�r(r(r)�startTagImplyTbody�sz2getPhases.<locals>.InTablePhase.startTagImplyTbodycSs6|j�dddd��|jj�td��|jjs2|SdS)Nrr�r�)rOr�rdr�rrWr�r(r(r)r��s�z-getPhases.<locals>.InTablePhase.startTagTablecSs|jjd�|�Sr>rr�r(r(r)�startTagStyleScript�sz3getPhases.<locals>.InTablePhase.startTagStyleScriptcSsVd|dvrH|dd�t�dkrH|j�d�|j�|�|jj��n
|�|�dS)Nr2rr�z unexpected-hidden-input-in-table)	rrrrOr�rAr�r�r�rr�r(r(r)r��s�z-getPhases.<locals>.InTablePhase.startTagInputcSsD|j�d�|jjdur@|j�|�|jjd|j_|jj��dS)Nzunexpected-form-in-tabler~)rOr�rArtr�r�r�r�r(r(r)ru�s
z,getPhases.<locals>.InTablePhase.startTagFormcSs<|j�dd|di�d|j_|jjd�|�d|j_dS)Nz)unexpected-start-tag-implies-table-voodoorCTr�F)rOr�rArrLr�r�r(r(r)r�sz-getPhases.<locals>.InTablePhase.startTagOthercSs�|jjddd�r�|j��|jjdjdkrJ|j�dd|jjdjd��|jjdjdkrj|jj��qJ|jj��|j��n|jj	s�J�|j��dS)Nr�rmr~zend-tag-too-early-namedr�)
rAror�r�rCrOr�r�rfrWr�r(r(r)�endTagTables
��z+getPhases.<locals>.InTablePhase.endTagTablecSs|j�dd|di�dSr1r�r�r(r(r)�endTagIgnoresz,getPhases.<locals>.InTablePhase.endTagIgnorecSs<|j�dd|di�d|j_|jjd�|�d|j_dS)Nz'unexpected-end-tag-implies-table-voodoorCTr�F)rOr�rArrLr�r�r(r(r)rsz+getPhases.<locals>.InTablePhase.endTagOtherr�r�r�r
�r�r�r�)r�r�r�r�)r5r7r�rs)r�r�r
r�r�r�r�r�r�r�r�N)r<r=r>rr�rr�r�r�r�rrrrrr�rr�rurrr rrrr�r�rr�r(�r�r(r)�InTablePhase�sN	
���r#csZeZdZdZ��fdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Z�ZS)z#getPhases.<locals>.InTableTextPhase)r��characterTokenscs&t�|�j|i|��d|_g|_dSr-)rLrNr�r$rN)�InTableTextPhaser�r(r)rN6sz,getPhases.<locals>.InTableTextPhase.__init__cSsdd�dd�|jD��}tdd�|D��rJtd|d�}|jjd�|�n|rZ|j�|�g|_dS)Nr�cSsg|]}|d�qS)rr(�rBr�r(r(r)rf<rHzGgetPhases.<locals>.InTableTextPhase.flushCharacters.<locals>.<listcomp>cSsg|]}|tv�qSr(rcr&r(r(r)rf=rHrwr�r�)�joinr$rgrrOrLr�rA)rFrr�r(r(r)�flushCharacters;sz3getPhases.<locals>.InTableTextPhase.flushCharacterscSs|��|j|j_|Sr-�r(r�rOrdr�r(r(r)r�Ds
z2getPhases.<locals>.InTableTextPhase.processCommentcSs|��|j|j_dSrr)rEr(r(r)r�Is
z.getPhases.<locals>.InTableTextPhase.processEOFcSs |ddkrdS|j�|�dS�Nrrb�r$r�r�r(r(r)r�Nsz5getPhases.<locals>.InTableTextPhase.processCharacterscSs|j�|�dSr-r+r�r(r(r)r�Ssz:getPhases.<locals>.InTableTextPhase.processSpaceCharacterscSs|��|j|j_|Sr-r)r�r(r(r)r�Xs
z3getPhases.<locals>.InTableTextPhase.processStartTagcSs|��|j|j_|Sr-r)r�r(r(r)r�]s
z1getPhases.<locals>.InTableTextPhase.processEndTag)
r<r=r>r�rNr(r�r�r�r�r�r�r
r()r%rr)r%3s	r%cs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
e�d�jfdefg�Ze	e_e�de
fdefdefg�Ze
e_dS)z!getPhases.<locals>.InCaptionPhasecSs|jjddd�S)Nr�r�rm�rArorEr(r(r)�ignoreEndTagCaptionfsz5getPhases.<locals>.InCaptionPhase.ignoreEndTagCaptioncSs|jjd��dSr�rOrLr�rEr(r(r)r�isz,getPhases.<locals>.InCaptionPhase.processEOFcSs|jjd�|�Sr�rOrLr�r�r(r(r)r�lsz3getPhases.<locals>.InCaptionPhase.processCharacterscSs0|j��|��}|jj�td��|s,|SdS�Nr��rOr�r-rdr�r�rFr��ignoreEndTagr(r(r)�startTagTableElementos

z6getPhases.<locals>.InCaptionPhase.startTagTableElementcSs|jjd�|�Srrr�r(r(r)rwsz/getPhases.<locals>.InCaptionPhase.startTagOthercSs�|��s�|j��|jjdjdkrB|j�dd|jjdjd��|jjdjdkrb|jj��qB|jj��|j��|jj	d|j_
n|jjs�J�|j��dS)Nr~r�r�r�r�)r-rAr�r�rCrOr�r�r�rLrdrWr�r(r(r)�
endTagCaptionzs
��
z/getPhases.<locals>.InCaptionPhase.endTagCaptioncSs0|j��|��}|jj�td��|s,|SdSr0r1r2r(r(r)r�s

z-getPhases.<locals>.InCaptionPhase.endTagTablecSs|j�dd|di�dSr1r�r�r(r(r)r �sz.getPhases.<locals>.InCaptionPhase.endTagIgnorecSs|jjd�|�Sr�rOrLr�r�r(r(r)r�sz-getPhases.<locals>.InCaptionPhase.endTagOtherr��	r�r
r�r�r�r�r�r�r�r�r�)
r�r
r�r�r�r�r�r�r�r�N)r<r=r>rr�r-r�r�r4rr5rr rrrr�r�rr�r(r"r(r)�InCaptionPhasebs2����r8cs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Ze
�d�jfdefg�Ze	e_e
�de
fdefg�Zee_dS)z%getPhases.<locals>.InColumnGroupPhasecSs|jjdjdkS)Nr~r�)rAr�rCrEr(r(r)�ignoreEndTagColgroup�sz:getPhases.<locals>.InColumnGroupPhase.ignoreEndTagColgroupcSsD|jjdjdkr"|jjsJ�dS|��}|�td��|s@dSdS)Nr~r�r�T)rAr�rCrOrWr9�endTagColgroupr)rFr3r(r(r)r��sz0getPhases.<locals>.InColumnGroupPhase.processEOFcSs"|��}|�td��|s|SdS�Nr��r9r:rr2r(r(r)r��sz7getPhases.<locals>.InColumnGroupPhase.processCharacterscSs$|j�|�|jj��d|d<dSrrr�r(r(r)r�sz1getPhases.<locals>.InColumnGroupPhase.startTagColcSs"|��}|�td��|s|SdSr;r<r2r(r(r)r�sz3getPhases.<locals>.InColumnGroupPhase.startTagOthercSs@|��r |jjsJ�|j��n|jj��|jjd|j_dS�Nr�)	r9rOrWr�rAr�r�rLrdr�r(r(r)r:�s
z4getPhases.<locals>.InColumnGroupPhase.endTagColgroupcSs|j�dddi�dS)Nz
no-end-tagrCr
r�r�r(r(r)�	endTagCol�sz/getPhases.<locals>.InColumnGroupPhase.endTagColcSs"|��}|�td��|s|SdSr;r<r2r(r(r)r�sz1getPhases.<locals>.InColumnGroupPhase.endTagOtherr�r
r�N)r<r=r>rr�r9r�r�rrr:r>rrrr�r�rr�r(r"r(r)�InColumnGroupPhase�s&
	��r?cs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Ze�d�jfde	fde
fdefg�Zee_e�de
fdefdefg�Zee_d S)!z#getPhases.<locals>.InTableBodyPhasecSsB|jjdjdvr |jj��q|jjdjdkr>|jjs>J�dS)Nr~)r�r�r�r�r�)rAr�rCr�rOrWrEr(r(r)�clearStackToTableBodyContext�sz@getPhases.<locals>.InTableBodyPhase.clearStackToTableBodyContextcSs|jjd��dSr=r.rEr(r(r)r��sz.getPhases.<locals>.InTableBodyPhase.processEOFcSs|jjd�|�Sr=r?r�r(r(r)r��sz:getPhases.<locals>.InTableBodyPhase.processSpaceCharacterscSs|jjd�|�Sr=r/r�r(r(r)r��sz5getPhases.<locals>.InTableBodyPhase.processCharacterscSs(|��|j�|�|jjd|j_dS)Nr�)r@rAr�rOrLrdr�r(r(r)�
startTagTrsz.getPhases.<locals>.InTableBodyPhase.startTagTrcSs*|j�dd|di�|�tdd��|S)Nzunexpected-cell-in-table-bodyrCr�ry)rOr�rArr�r(r(r)�startTagTableCells

�z5getPhases.<locals>.InTableBodyPhase.startTagTableCellcSsn|jjddd�s0|jjddd�s0|jjddd�rT|��|�t|jjdj��|S|jjs`J�|j�	�dS�Nr�r�rmr�r�r~�
rAror@�endTagTableRowGrouprr�rCrOrWr�r�r(r(r)�startTagTableOthers���z6getPhases.<locals>.InTableBodyPhase.startTagTableOthercSs|jjd�|�Sr=rr�r(r(r)rsz1getPhases.<locals>.InTableBodyPhase.startTagOthercSsT|jj|ddd�r:|��|jj��|jjd|j_n|j�dd|di�dS)NrCr�rmr�� unexpected-end-tag-in-table-body)	rAror@r�r�rOrLrdr�r�r(r(r)rEs
�z7getPhases.<locals>.InTableBodyPhase.endTagTableRowGroupcSsn|jjddd�s0|jjddd�s0|jjddd�rT|��|�t|jjdj��|S|jjs`J�|j�	�dSrCrDr�r(r(r)r%s���z/getPhases.<locals>.InTableBodyPhase.endTagTablecSs|j�dd|di�dS)NrGrCr�r�r(r(r)r 2s
�z0getPhases.<locals>.InTableBodyPhase.endTagIgnorecSs|jjd�|�Sr=r6r�r(r(r)r6sz/getPhases.<locals>.InTableBodyPhase.endTagOtherr�r��r�r�)r�r
r�r�r�r�r!r�)r�r�r
r�r�r�r�r�N)r<r=r>rr�r@r�r�r�rArBrFrrErr rrrr�r�rr�r(r"r(r)�InTableBodyPhase�s<
	
����rIcs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Ze�d�jfde
fdefg�Zee_e�de
fdefd efd!efg�Zee_d"S)#zgetPhases.<locals>.InRowPhasecSs@|jjdjdvr<|j�dd|jjdji�|jj��qdS)Nr~)r�r�z'unexpected-implied-end-tag-in-table-rowrC)rAr�rCrOr�r�rEr(r(r)�clearStackToTableRowContextOs
�z9getPhases.<locals>.InRowPhase.clearStackToTableRowContextcSs|jjddd�S)Nr�r�rmr,rEr(r(r)�ignoreEndTagTrUsz,getPhases.<locals>.InRowPhase.ignoreEndTagTrcSs|jjd��dSr=r.rEr(r(r)r�Ysz(getPhases.<locals>.InRowPhase.processEOFcSs|jjd�|�Sr=r?r�r(r(r)r�\sz4getPhases.<locals>.InRowPhase.processSpaceCharacterscSs|jjd�|�Sr=r/r�r(r(r)r�_sz/getPhases.<locals>.InRowPhase.processCharacterscSs6|��|j�|�|jjd|j_|jj�t�dS)Nr�)	rJrAr�rOrLrdrTr�rr�r(r(r)rBbsz/getPhases.<locals>.InRowPhase.startTagTableCellcSs"|��}|�td��|s|SdS�Nr��rK�endTagTrrr2r(r(r)rFhsz0getPhases.<locals>.InRowPhase.startTagTableOthercSs|jjd�|�Sr=rr�r(r(r)rosz+getPhases.<locals>.InRowPhase.startTagOthercSsH|��s.|��|jj��|jjd|j_n|jjs:J�|j�	�dSr)
rKrJrAr�r�rOrLrdrWr�r�r(r(r)rNrsz&getPhases.<locals>.InRowPhase.endTagTrcSs"|��}|�td��|s|SdSrLrMr2r(r(r)r|sz)getPhases.<locals>.InRowPhase.endTagTablecSs4|jj|ddd�r&|�td��|S|j��dS)NrCr�rmr�)rArorNrrOr�r�r(r(r)rE�sz1getPhases.<locals>.InRowPhase.endTagTableRowGroupcSs|j�dd|di�dS)Nzunexpected-end-tag-in-table-rowrCr�r�r(r(r)r �s
�z*getPhases.<locals>.InRowPhase.endTagIgnorecSs|jjd�|�Sr=r6r�r(r(r)r�sz)getPhases.<locals>.InRowPhase.endTagOtherr�rH)r�r
r�r�r�r�r�r�r�r!)r�r�r
r�r�r�r�N)r<r=r>rr�rJrKr�r�r�rBrFrrNrrEr rrrr�r�rr�r(r"r(r)�
InRowPhaseJs>
����rOcs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
e�d�jfdefg�Ze	e_e�de
fdefdefg�Ze
e_dS)zgetPhases.<locals>.InCellPhasecSsB|jjddd�r |�td��n|jjddd�r>|�td��dS�Nr�r�rmr�)rAro�endTagTableCellrrEr(r(r)�	closeCell�sz(getPhases.<locals>.InCellPhase.closeCellcSs|jjd��dSrr.rEr(r(r)r��sz)getPhases.<locals>.InCellPhase.processEOFcSs|jjd�|�Srr/r�r(r(r)r��sz0getPhases.<locals>.InCellPhase.processCharacterscSsF|jjddd�s |jjddd�r,|��|S|jjs8J�|j��dSrP)rArorRrOrWr�r�r(r(r)rF�s�z1getPhases.<locals>.InCellPhase.startTagTableOthercSs|jjd�|�Srrr�r(r(r)r�sz,getPhases.<locals>.InCellPhase.startTagOthercSs�|jj|ddd�r�|j�|d�|jjdj|dkrp|j�dd|di�|jj��}|j|dkrPq|qPn|jj��|j��|jj	d|j_
n|j�dd|di�dS)NrCr�rmr~zunexpected-cell-end-tagr�r2)rAror�r�rCrOr�r�r�rLrdr.r(r(r)rQ�s
�
z.getPhases.<locals>.InCellPhase.endTagTableCellcSs|j�dd|di�dSr1r�r�r(r(r)r �sz+getPhases.<locals>.InCellPhase.endTagIgnorecSs.|jj|ddd�r |��|S|j��dS)NrCr�rm)rArorRrOr�r�r(r(r)�endTagImply�sz*getPhases.<locals>.InCellPhase.endTagImplycSs|jjd�|�Srr6r�r(r(r)r�sz*getPhases.<locals>.InCellPhase.endTagOtherr�r7rH)r�r�r
r�r�r�N)r<r=r>rr�rRr�r�rFrrQr rSrrrr�r�rr�r(r"r(r)�InCellPhase�s.
���rTc	s�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zdd�Ze�d�jfdefdefde	fde
fdefg�Zee_e�de
fdefdefg�Zee_dS) z getPhases.<locals>.InSelectPhasecSs0|jjdjdkr |j�d�n|jjs,J�dS)Nr~r�z
eof-in-selectrrEr(r(r)r��sz+getPhases.<locals>.InSelectPhase.processEOFcSs$|ddkrdS|j�|d�dSr*r�r�r(r(r)r��sz2getPhases.<locals>.InSelectPhase.processCharacterscSs.|jjdjdkr|jj��|j�|�dSr��rAr�rCr�r�r�r(r(r)�startTagOption�sz/getPhases.<locals>.InSelectPhase.startTagOptioncSsL|jjdjdkr|jj��|jjdjdkr<|jj��|j�|�dS)Nr~r�r�rUr�r(r(r)�startTagOptgroup	s
z1getPhases.<locals>.InSelectPhase.startTagOptgroupcSs|j�d�|�td��dS)Nzunexpected-select-in-selectr�)rOr��endTagSelectrr�r(r(r)r�	sz/getPhases.<locals>.InSelectPhase.startTagSelectcSs>|j�d�|jjddd�r.|�td��|S|jjs:J�dS)Nzunexpected-input-in-selectr�rm)rOr�rArorXrrWr�r(r(r)r�	s
z.getPhases.<locals>.InSelectPhase.startTagInputcSs|jjd�|�Sr>rr�r(r(r)r+	sz/getPhases.<locals>.InSelectPhase.startTagScriptcSs|j�dd|di�dS)Nzunexpected-start-tag-in-selectrCr�r�r(r(r)r	s
�z.getPhases.<locals>.InSelectPhase.startTagOthercSs6|jjdjdkr |jj��n|j�dddi�dS)Nr~r��unexpected-end-tag-in-selectrC�rAr�rCr�rOr�r�r(r(r)�endTagOption	s
�z-getPhases.<locals>.InSelectPhase.endTagOptioncSsf|jjdjdkr0|jjdjdkr0|jj��|jjdjdkrP|jj��n|j�dddi�dS)Nr~r����r�rYrCrZr�r(r(r)�endTagOptgroup%	s��z/getPhases.<locals>.InSelectPhase.endTagOptgroupcSsZ|jjddd�r@|jj��}|jdkr4|jj��}q|j��n|jjsLJ�|j��dS)Nr�rm)	rAror�r�rCrOrfrWr�r.r(r(r)rX2	s
z-getPhases.<locals>.InSelectPhase.endTagSelectcSs|j�dd|di�dS)NrYrCr�r�r(r(r)r=	s
�z,getPhases.<locals>.InSelectPhase.endTagOtherr�r�r�r�)r�rr^r7N)r<r=r>rr�r�r�rVrWr�r�r+rr[r]rXrrrr�r�rr�r(r"r(r)�
InSelectPhase�s8
��r^c@sneZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
e�d
efg�Z
ee
_e�d
e	fg�Ze
e_dS)z'getPhases.<locals>.InSelectInTablePhasecSs|jjd��dS�Nr�r.rEr(r(r)r�U	sz2getPhases.<locals>.InSelectInTablePhase.processEOFcSs|jjd�|�Sr_r/r�r(r(r)r�X	sz9getPhases.<locals>.InSelectInTablePhase.processCharacterscSs(|j�dd|di�|�td��|S)Nz5unexpected-table-element-start-tag-in-select-in-tablerCr�)rOr�rrr�r(r(r)r�[	sz5getPhases.<locals>.InSelectInTablePhase.startTagTablecSs|jjd�|�Sr_rr�r(r(r)r`	sz5getPhases.<locals>.InSelectInTablePhase.startTagOthercSs@|j�dd|di�|jj|ddd�r<|�td��|SdS)Nz3unexpected-table-element-end-tag-in-select-in-tablerCr�rmr�)rOr�rArorrr�r(r(r)rc	sz3getPhases.<locals>.InSelectInTablePhase.endTagTablecSs|jjd�|�Sr_r6r�r(r(r)ri	sz3getPhases.<locals>.InSelectInTablePhase.endTagOther)r�r�r�r�r�r�r�r�N)r<r=r>rr�r�r�r�rrrrrr�rr�r(r(r(r)�InSelectInTablePhaseR	s&����r`csBeZdZe�Zegd��Zdd�Z�fdd�Zdd�Z	dd	�Z
d
S)z(getPhases.<locals>.InForeignContentPhase),r�r�r�r�r	r�r�rWr*r�rXr�r�h1�h2�h3�h4�h5�h6r�r�r�r�rYr]r�r4r�r�r'r\r�r�r��spanr�r��sub�supr�r�r�r��varc%Ssnddddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%�$}|d&|vrj||d&|d&<dS)'N�altGlyph�altGlyphDef�altGlyphItem�animateColor�
animateMotion�animateTransform�clipPath�feBlend�
feColorMatrix�feComponentTransfer�feComposite�feConvolveMatrix�feDiffuseLighting�feDisplacementMap�feDistantLight�feFlood�feFuncA�feFuncB�feFuncG�feFuncR�feGaussianBlur�feImage�feMerge�feMergeNode�feMorphology�feOffset�fePointLight�feSpecularLighting�feSpotLight�feTile�feTurbulence�
foreignObject�glyphRef�linearGradient�radialGradient�textPath)$�altglyph�altglyphdef�altglyphitem�animatecolor�
animatemotion�animatetransform�clippath�feblend�
fecolormatrix�fecomponenttransfer�fecomposite�feconvolvematrix�fediffuselighting�fedisplacementmap�fedistantlight�feflood�fefunca�fefuncb�fefuncg�fefuncr�fegaussianblur�feimage�femerge�femergenode�femorphology�feoffset�fepointlight�fespecularlighting�fespotlight�fetile�feturbulence�
foreignobject�glyphref�lineargradient�radialgradient�textpathrCr()rFr��replacementsr(r(r)�adjustSVGTagNames�	sN�%z:getPhases.<locals>.InForeignContentPhase.adjustSVGTagNamescsL|ddkrd|d<n&|jjr<tdd�|dD��r<d|j_��||�dS)Nrrbu�css|]}|tvVqdSr-rcrdr(r(r)�	<genexpr>�	rHzMgetPhases.<locals>.InForeignContentPhase.processCharacters.<locals>.<genexpr>F)rOrirgr�r�r"r(r)r��	s
�z:getPhases.<locals>.InForeignContentPhase.processCharacterscSs(|jjd}|d|jvs>|ddkr�t|d���hd�@r�|j�dd|di�|jjdj|jjkr�|j�	|jjd�s�|j�
|jjd�s�|jj��qT|S|jtdkr�|j�
|�n$|jtdkr�|�|�|j�|�|j�|�|j|d	<|j�|�|d
�r$|jj��d|d<dS)
Nr~rCr�r>�size�face�colorz*unexpected-html-element-in-foreign-contentrnr�rpr�Tr�)rAr��breakoutElements�set�keysrOr�rpr�rurvr�rrr�rrr�)rFr�r�r(r(r)r��	s:
��
����


z8getPhases.<locals>.InForeignContentPhase.processStartTagcSs�t|jj�d}|jjd}|j�t�|dkrF|j�dd|di�|j�t�|dkr�|jj|jj	dkr�|jj�
�|jjj|j_|jj��|kr�|jjs�J�q�d}q�|d8}|jj|}|j
|jjkr�qFqF|jj�|�}q�qF|S)Nrr~rCr2r)r�rAr�rCrrrrOr�rdrLr(r�r�rpr�r�)rFr��	nodeIndexr�r�r(r(r)r��	s&z6getPhases.<locals>.InForeignContentPhase.processEndTagN)r<r=r>rr�r�r�r�r�r�r�r(r"r(r)�InForeignContentPhasex	s	)r�c@sveZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Ze�
defg�Ze	e_e�
de
fg�Zee_dS)z!getPhases.<locals>.AfterBodyPhasecSsdSr-r(rEr(r(r)r��	sz,getPhases.<locals>.AfterBodyPhase.processEOFcSs|j�||jjd�dS)Nrr�r�r(r(r)r��	sz0getPhases.<locals>.AfterBodyPhase.processCommentcSs |j�d�|jjd|j_|S)Nzunexpected-char-after-bodyr��rOr�rLrdr�r(r(r)r��	sz3getPhases.<locals>.AfterBodyPhase.processCharacterscSs|jjd�|�Srrr�r(r(r)r��	sz.getPhases.<locals>.AfterBodyPhase.startTagHtmlcSs*|j�dd|di�|jjd|j_|S)Nzunexpected-start-tag-after-bodyrCr�r�r�r(r(r)r�	s

�z/getPhases.<locals>.AfterBodyPhase.startTagOthercSs*|jjr|j�d�n|jjd|j_dS)Nz'unexpected-end-tag-after-body-innerhtml�afterAfterBody)rOrWr�rLrd)rFrCr(r(r)r�
sz,getPhases.<locals>.AfterBodyPhase.endTagHtmlcSs*|j�dd|di�|jjd|j_|S)Nzunexpected-end-tag-after-bodyrCr�r�r�r(r(r)r

s

�z-getPhases.<locals>.AfterBodyPhase.endTagOtherr�N)r<r=r>rr�r�r�r�r�rr�rrrr�rr�r(r(r(r)�AfterBodyPhase�	s�r�cs�eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Ze
�d�jfdefdefde	fg�Ze
e_e
�defg�Zee_dS)z"getPhases.<locals>.InFramesetPhasecSs0|jjdjdkr |j�d�n|jjs,J�dS)Nr~r�zeof-in-framesetrrEr(r(r)r�
sz-getPhases.<locals>.InFramesetPhase.processEOFcSs|j�d�dS)Nzunexpected-char-in-framesetr�r�r(r(r)r�"
sz4getPhases.<locals>.InFramesetPhase.processCharacterscSs|j�|�dSr-)rAr�r�r(r(r)rH%
sz3getPhases.<locals>.InFramesetPhase.startTagFramesetcSs|j�|�|jj��dSr-rr�r(r(r)�
startTagFrame(
sz0getPhases.<locals>.InFramesetPhase.startTagFramecSs|jjd�|�Srrr�r(r(r)�startTagNoframes,
sz3getPhases.<locals>.InFramesetPhase.startTagNoframescSs|j�dd|di�dS)Nz unexpected-start-tag-in-framesetrCr�r�r(r(r)r/
s
�z0getPhases.<locals>.InFramesetPhase.startTagOthercSsZ|jjdjdkr |j�d�n|jj��|jjsV|jjdjdkrV|jjd|j_dS)Nr~r�z)unexpected-frameset-in-frameset-innerhtmlr��
afterFrameset)	rAr�rCrOr�r�rWrLrdr�r(r(r)�endTagFrameset3
s�z1getPhases.<locals>.InFramesetPhase.endTagFramesetcSs|j�dd|di�dS)Nzunexpected-end-tag-in-framesetrCr�r�r(r(r)r?
s
�z.getPhases.<locals>.InFramesetPhase.endTagOtherr�r�rr4N)r<r=r>rr�r�r�rHr�r�rr�rrrr�r�rr�r(r"r(r)�InFramesetPhase
s(��r�csveZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
e�d
�j
fdefg�Zee_e�d
e	fg�Ze
e_dS)z%getPhases.<locals>.AfterFramesetPhasecSsdSr-r(rEr(r(r)r�T
sz0getPhases.<locals>.AfterFramesetPhase.processEOFcSs|j�d�dS)Nzunexpected-char-after-framesetr�r�r(r(r)r�X
sz7getPhases.<locals>.AfterFramesetPhase.processCharacterscSs|jjd�|�Sr>rr�r(r(r)r�[
sz6getPhases.<locals>.AfterFramesetPhase.startTagNoframescSs|j�dd|di�dS)Nz#unexpected-start-tag-after-framesetrCr�r�r(r(r)r^
s
�z3getPhases.<locals>.AfterFramesetPhase.startTagOthercSs|jjd|j_dS)N�afterAfterFrameset)rOrLrdr�r(r(r)r�b
sz0getPhases.<locals>.AfterFramesetPhase.endTagHtmlcSs|j�dd|di�dS)Nz!unexpected-end-tag-after-framesetrCr�r�r(r(r)re
s
�z1getPhases.<locals>.AfterFramesetPhase.endTagOtherr�r4N)r<r=r>rr�r�r�r�rr�rrrr�r�rr�r(r"r(r)�AfterFramesetPhaseP
s ��r�c@s`eZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Ze�
de	fg�Ze
e_dS)z&getPhases.<locals>.AfterAfterBodyPhasecSsdSr-r(rEr(r(r)r�w
sz1getPhases.<locals>.AfterAfterBodyPhase.processEOFcSs|j�||jj�dSr-r�r�r(r(r)r�z
sz5getPhases.<locals>.AfterAfterBodyPhase.processCommentcSs|jjd�|�Srr?r�r(r(r)r�}
sz=getPhases.<locals>.AfterAfterBodyPhase.processSpaceCharacterscSs |j�d�|jjd|j_|S)N�expected-eof-but-got-charr�r�r�r(r(r)r��
sz8getPhases.<locals>.AfterAfterBodyPhase.processCharacterscSs|jjd�|�Srrr�r(r(r)r��
sz3getPhases.<locals>.AfterAfterBodyPhase.startTagHtmlcSs*|j�dd|di�|jjd|j_|S)N�expected-eof-but-got-start-tagrCr�r�r�r(r(r)r�
s

�z4getPhases.<locals>.AfterAfterBodyPhase.startTagOthercSs*|j�dd|di�|jjd|j_|S)N�expected-eof-but-got-end-tagrCr�r�r�r(r(r)r��
s

�z4getPhases.<locals>.AfterAfterBodyPhase.processEndTagr�N)r<r=r>rr�r�r�r�r�r�rr�rrr�rr(r(r(r)�AfterAfterBodyPhaset
s�r�c@sneZdZe�Zdd�Zdd�Zdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�Ze
�de	fde
fg�Zee_dS)z*getPhases.<locals>.AfterAfterFramesetPhasecSsdSr-r(rEr(r(r)r��
sz5getPhases.<locals>.AfterAfterFramesetPhase.processEOFcSs|j�||jj�dSr-r�r�r(r(r)r��
sz9getPhases.<locals>.AfterAfterFramesetPhase.processCommentcSs|jjd�|�Srr?r�r(r(r)r��
szAgetPhases.<locals>.AfterAfterFramesetPhase.processSpaceCharacterscSs|j�d�dS)Nr�r�r�r(r(r)r��
sz<getPhases.<locals>.AfterAfterFramesetPhase.processCharacterscSs|jjd�|�Srrr�r(r(r)r��
sz7getPhases.<locals>.AfterAfterFramesetPhase.startTagHtmlcSs|jjd�|�Sr>rr�r(r(r)�startTagNoFrames�
sz;getPhases.<locals>.AfterAfterFramesetPhase.startTagNoFramescSs|j�dd|di�dS)Nr�rCr�r�r(r(r)r�
s
�z8getPhases.<locals>.AfterAfterFramesetPhase.startTagOthercSs|j�dd|di�dS)Nr�rCr�r�r(r(r)r��
s
�z8getPhases.<locals>.AfterAfterFramesetPhase.processEndTagr�r4N)r<r=r>rr�r�r�r�r�r�r�rr�rrr�rr(r(r(r)�AfterAfterFramesetPhase�
s�r�)r[rZr�rr(r,r�r�r�rr�r�r�r�r�r�r�r�r�r�r�r�r�)r)rMr]r�rr
rr=rFrKrr#r8r?rIrOrTr^r`r�r�r�r�r�r�r()rOr%r�r)rK�srH%-f?BY!/F@bYLc&r.8$%&�rKcsFt|d�t��@}|rBt|d��fdd�|d��D��|d<dS)Nrc3s"|]\}}��||�|fVqdSr-)r�)rB�k�v�r�r(r)r��
s�z$adjust_attributes.<locals>.<genexpr>)rr2r.)r�r��needs_adjustmentr(r�r)r��
s

�r�rzFcCs|duri}t||||d�S)N)r2rCrr�)r)rCr2rqr�r(r(r)r�
s

�rc@seZdZdZdS)r}zError in parsed documentN)r<r=r>r�r(r(r(r)r}�
sr})rT)r*rT)rzNF),�
__future__rrrZpip._vendor.sixrrr0r�rr	r
Ztreebuilders.baserr�	constantsr
rrrrrrrrrrrr�rrrrr"r,r@r�r!�memoizerKr�r�	Exceptionr}r(r(r(r)�<module>sNH

8
_�
site-packages/pip/_vendor/html5lib/__pycache__/_inputstream.cpython-39.pyc000064400000052245150564745010022622 0ustar00a

��Rea~�@s�ddlmZmZmZddlmZddlmZmZddl	Z	ddl
Z
ddlmZm
Z
ddlmZddlmZmZmZmZdd	lmZdd
lmZedd�eD��Zed
d�eD��Zedd�eD��Zeeddg�BZdZej�reddkr�e�d�dks�J�e
� edd�e!d�d�Z"n
e
� e�Z"hd�Z#e
� d�Z$iZ%Gdd�de&�Z'dd�Z(Gdd�de&�Z)Gdd�de)�Z*Gdd �d e+�Z,Gd!d"�d"e&�Z-Gd#d$�d$e&�Z.d%d&�Z/dS)'�)�absolute_import�division�unicode_literals)�	text_type)�http_client�urllibN)�BytesIO�StringIO)�webencodings�)�EOF�spaceCharacters�asciiLetters�asciiUppercase)�_ReparseException)�_utilscCsg|]}|�d��qS��ascii��encode��.0�item�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_inputstream.py�
<listcomp>�rcCsg|]}|�d��qSrrrrrrrrcCsg|]}|�d��qSrrrrrrrr�>�<u�[---Ÿ﷐-﷯￾￿🿾🿿𯿾𯿿𿿾𿿿񏿾񏿿񟿾񟿿񯿾񯿿񿿾񿿿򏿾򏿿򟿾򟿿򯿾򯿿򿿾򿿿󏿾󏿿󟿾󟿿󯿾󯿿󿿾󿿿􏿾􏿿]����]z"\uD800-\uDFFF"> i��
i��i��i��i��i��i��i��i��i��i��i��i��i��	i��i��i��	i��i��i��
i��i��i��i��i��i��i��
i��
i��i��i��i��z[	-
 -/:-@\[-`{-~]c@sHeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dS)�BufferedStreamz�Buffering for streams that do not have buffering of their own

    The buffer is implemented as a list of chunks on the assumption that
    joining many strings will be slow since it is O(n**2)
    cCs||_g|_ddg|_dS)Nrr)�stream�buffer�position)�selfr"rrr�__init__:szBufferedStream.__init__cCs<d}|jd|jd�D]}|t|�7}q||jd7}|S�Nrr)r#r$�len)r%�pos�chunkrrr�tell?s
zBufferedStream.tellcCsT||��ksJ�|}d}t|j|�|krF|t|j|�8}|d7}q||g|_dSr')�_bufferedBytesr(r#r$)r%r)�offset�irrr�seekFs
zBufferedStream.seekcCsT|js|�|�S|jdt|j�krF|jdt|jd�krF|�|�S|�|�SdS)Nrrr)r#�_readStreamr$r(�_readFromBuffer�r%�bytesrrr�readOs
�
zBufferedStream.readcCstdd�|jD��S)NcSsg|]}t|��qSr)r(rrrrrYrz1BufferedStream._bufferedBytes.<locals>.<listcomp>)�sumr#�r%rrrr,XszBufferedStream._bufferedBytescCs<|j�|�}|j�|�|jdd7<t|�|jd<|Sr')r"r4r#�appendr$r()r%r3�datarrrr0[s
zBufferedStream._readStreamcCs�|}g}|jd}|jd}|t|j�kr�|dkr�|dks>J�|j|}|t|�|krl|}|||g|_n"t|�|}|t|�g|_|d7}|�||||��||8}d}q|r�|�|�|��d�|�S)Nrrr)r$r(r#r7r0�join)r%r3�remainingBytes�rv�bufferIndex�bufferOffset�bufferedData�bytesToReadrrrr1bs&


zBufferedStream._readFromBufferN)�__name__�
__module__�__qualname__�__doc__r&r+r/r4r,r0r1rrrrr!3s		r!cKs�t|tj�s(t|tjj�r.t|jtj�r.d}n&t|d�rJt|�d�t	�}n
t|t	�}|r�dd�|D�}|rvt
d|��t|fi|��St|fi|��SdS)NFr4rcSsg|]}|�d�r|�qS)�	_encoding)�endswith)r�xrrrr�rz#HTMLInputStream.<locals>.<listcomp>z3Cannot set an encoding with a unicode input, set %r)
�
isinstancer�HTTPResponser�response�addbase�fp�hasattrr4r�	TypeError�HTMLUnicodeInputStream�HTMLBinaryInputStream)�source�kwargs�	isUnicode�	encodingsrrr�HTMLInputStream}s��

rTc@speZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
ddd�Zdd�Zdd�Z
ddd�Zdd�ZdS)rN��Provides a unicode stream of characters to the HTMLTokenizer.

    This class takes care of character encoding and removing or replacing
    incorrect byte-sequences and also provides column and line tracking.

    i(cCsZtjsd|_ntd�dkr$|j|_n|j|_dg|_td�df|_|�	|�|_
|��dS)�Initialises the HTMLInputStream.

        HTMLInputStream(source, [encoding]) -> Normalized stream from source
        for use by html5lib.

        source can be either a file-object, local filename or a string.

        The optional encoding parameter must be a string that indicates
        the encoding.  If specified, that encoding will be used,
        regardless of any BOM or later declaration (such as in a meta
        element)

        Nu􏿿rr�utf-8�certain)r�supports_lone_surrogates�reportCharacterErrorsr(�characterErrorsUCS4�characterErrorsUCS2�newLines�lookupEncoding�charEncoding�
openStream�
dataStream�reset)r%rPrrrr&�s
zHTMLUnicodeInputStream.__init__cCs.d|_d|_d|_g|_d|_d|_d|_dS)N�r)r*�	chunkSize�chunkOffset�errors�prevNumLines�prevNumCols�_bufferedCharacterr6rrrrb�szHTMLUnicodeInputStream.resetcCst|d�r|}nt|�}|S�zvProduces a file object from source.

        source can be either a file object, local filename or a string.

        r4)rLr	�r%rPr"rrrr`�s
z!HTMLUnicodeInputStream.openStreamcCsT|j}|�dd|�}|j|}|�dd|�}|dkr@|j|}n||d}||fS)N�
rrr)r*�countrg�rfindrh)r%r-r*�nLines�positionLine�lastLinePos�positionColumnrrr�	_position�s
z HTMLUnicodeInputStream._positioncCs|�|j�\}}|d|fS)z:Returns (line, col) of the current position in the stream.r)rsre)r%�line�colrrrr$�szHTMLUnicodeInputStream.positioncCs6|j|jkr|��stS|j}|j|}|d|_|S)zo Read one character from the stream or queue if available. Return
            EOF when EOF is reached.
        r)rerd�	readChunkrr*)r%re�charrrrrw�s

zHTMLUnicodeInputStream.charNcCs�|dur|j}|�|j�\|_|_d|_d|_d|_|j�|�}|j	rX|j	|}d|_	n|s`dSt
|�dkr�t|d�}|dks�d|kr�dkr�nn|d|_	|dd�}|jr�|�|�|�
d	d
�}|�
dd
�}||_t
|�|_dS)
NrcrFrr�
��i��z
rl�
T)�_defaultChunkSizersrdrgrhr*rerar4rir(�ordrZ�replace)r%rdr8�lastvrrrrv�s0
 


z HTMLUnicodeInputStream.readChunkcCs(ttt�|���D]}|j�d�qdS)N�invalid-codepoint)�ranger(�invalid_unicode_re�findallrfr7)r%r8�_rrrr[sz*HTMLUnicodeInputStream.characterErrorsUCS4cCs�d}t�|�D]�}|rqt|���}|��}t�|||d��rrt�|||d��}|tvrl|j	�
d�d}q|dkr�|dkr�|t|�dkr�|j	�
d�qd}|j	�
d�qdS)NF�rTryi��r)r��finditerr|�group�startr�isSurrogatePair�surrogatePairToCodepoint�non_bmp_invalid_codepointsrfr7r()r%r8�skip�match�	codepointr)�char_valrrrr\#s"�z*HTMLUnicodeInputStream.characterErrorsUCS2Fc
Cszt||f}Wnftyv|D]}t|�dks"J�q"d�dd�|D��}|sXd|}t�d|�}t||f<Yn0g}|�|j|j�}|dur�|j|j	kr�q�n0|�
�}||j	kr�|�|j|j|��||_q�|�|j|jd��|��s|q�q|d�|�}	|	S)z� Returns a string of characters from the stream up to but not
        including any character in 'characters' or EOF. 'characters' must be
        a container that supports the 'in' method and iteration over its
        characters.
        �rccSsg|]}dt|��qS)z\x%02x)r|)r�crrrrHrz5HTMLUnicodeInputStream.charsUntil.<locals>.<listcomp>z^%sz[%s]+N)
�charsUntilRegEx�KeyErrorr|r9�re�compiler�r*rerd�endr7rv)
r%�
characters�opposite�charsr��regexr;�mr��rrrr�
charsUntil:s0 

z!HTMLUnicodeInputStream.charsUntilcCsT|turP|jdkr.||j|_|jd7_n"|jd8_|j|j|ksPJ�dSr')rrer*rd)r%rwrrr�ungetis
zHTMLUnicodeInputStream.unget)N)F)r@rArBrCr{r&rbr`rsr$rwrvr[r\r�r�rrrrrN�s 
&
/rNc@sLeZdZdZddd�Zdd�Zd	d
�Zddd�Zd
d�Zdd�Z	dd�Z
dS)rOrUN�windows-1252TcCsn|�|�|_t�||j�d|_d|_||_||_||_||_	||_
|�|�|_|jddusbJ�|�
�dS)rVi�drN)r`�	rawStreamrNr&�numBytesMeta�numBytesChardet�override_encoding�transport_encoding�same_origin_parent_encoding�likely_encoding�default_encoding�determineEncodingr_rb)r%rPr�r�r�r�r��
useChardetrrrr&�szHTMLBinaryInputStream.__init__cCs&|jdj�|jd�|_t�|�dS)Nrr})r_�
codec_info�streamreaderr�rarNrbr6rrrrb�szHTMLBinaryInputStream.resetcCsJt|d�r|}nt|�}z|�|���WntyDt|�}Yn0|Srj)rLrr/r+�	Exceptionr!rkrrrr`�s
z HTMLBinaryInputStream.openStreamcCs�|��df}|ddur|St|j�df}|ddur:|St|j�df}|ddurX|S|��df}|ddurt|St|j�df}|ddur�|dj�d�s�|St|j�df}|ddur�|S|�rnzddl	m
}Wnty�Yn�0g}|�}|j�s:|j
�|j�}t|t��sJ�|�s$�q:|�|�|�|�q�|��t|jd�}|j
�d�|du�rn|dfSt|j�df}|ddu�r�|Std�dfS)NrXr�	tentativezutf-16)�UniversalDetector�encodingr�)�	detectBOMr^r�r��detectEncodingMetar��name�
startswithr��%pip._vendor.chardet.universaldetectorr��ImportError�doner�r4r�rGr3r7�feed�close�resultr/r�)r%�chardetr_r��buffers�detectorr#r�rrrr��sR

z'HTMLBinaryInputStream.determineEncodingcCs�|jddksJ�t|�}|dur&dS|jdvrFtd�}|dus�J�nT||jdkrf|jddf|_n4|j�d�|df|_|��td|jd|f��dS)NrrX��utf-16be�utf-16lerWrzEncoding changed from %s to %s)r_r^r�r�r/rbr)r%�newEncodingrrr�changeEncodings

z$HTMLBinaryInputStream.changeEncodingc
Cs�tjdtjdtjdtjdtjdi}|j�d�}t|t	�s<J�|�
|dd��}d}|s~|�
|�}d}|s~|�
|dd	��}d	}|r�|j�|�t|�S|j�d
�dSdS)z�Attempts to detect at BOM at the start of the stream. If
        an encoding can be determined from the BOM return the name of the
        encoding otherwise return NonerWr�r�zutf-32lezutf-32be�N�r�r)
�codecs�BOM_UTF8�BOM_UTF16_LE�BOM_UTF16_BE�BOM_UTF32_LE�BOM_UTF32_BEr�r4rGr3�getr/r^)r%�bomDict�stringr�r/rrrr�s&�
zHTMLBinaryInputStream.detectBOMcCsV|j�|j�}t|t�sJ�t|�}|j�d�|��}|durR|jdvrRt	d�}|S)z9Report the encoding declared by the meta element
        rNr�rW)
r�r4r�rGr3�EncodingParserr/�getEncodingr�r^)r%r#�parserr�rrrr�3sz(HTMLBinaryInputStream.detectEncodingMeta)NNNNr�T)T)r@rArBrCr&rbr`r�r�r�r�rrrrrOzs�
*
>"rOc@s�eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zeee
�Z
dd�Zee�Zefdd�Zdd�Zdd�Zdd�ZdS)�
EncodingBytesz�String-like object with an associated position and various extra methods
    If the position is ever greater than the string length then an exception is
    raisedcCst|t�sJ�t�||���S�N)rGr3�__new__�lower�r%�valuerrrr�FszEncodingBytes.__new__cCs
d|_dS)Nr)rsr�rrrr&JszEncodingBytes.__init__cCs|Sr�rr6rrr�__iter__NszEncodingBytes.__iter__cCs>|jd}|_|t|�kr"t�n|dkr.t�|||d�S)Nrr�rsr(�
StopIterationrM�r%�prrr�__next__QszEncodingBytes.__next__cCs|��Sr�)r�r6rrr�nextYszEncodingBytes.nextcCsB|j}|t|�krt�n|dkr$t�|d|_}|||d�Sr'r�r�rrr�previous]szEncodingBytes.previouscCs|jt|�krt�||_dSr��rsr(r�)r%r$rrr�setPositionfszEncodingBytes.setPositioncCs*|jt|�krt�|jdkr"|jSdSdS)Nrr�r6rrr�getPositionks

zEncodingBytes.getPositioncCs||j|jd�S�Nr)r$r6rrr�getCurrentByteuszEncodingBytes.getCurrentBytecCsH|j}|t|�kr>|||d�}||vr4||_|S|d7}q||_dS)zSkip past a list of charactersrN�r$r(rs�r%r�r�r�rrrr�zs
zEncodingBytes.skipcCsH|j}|t|�kr>|||d�}||vr4||_|S|d7}q||_dSr�r�r�rrr�	skipUntil�s
zEncodingBytes.skipUntilcCs(|�||j�}|r$|jt|�7_|S)z�Look for a sequence of bytes at the start of a string. If the bytes
        are found return True and advance the position to the byte after the
        match. Otherwise return False and leave the position alone)r�r$r()r%r3r;rrr�
matchBytes�szEncodingBytes.matchBytescCs<z |�||j�t|�d|_Wnty6t�Yn0dS)z�Look for the next sequence of bytes matching a given sequence. If
        a match is found advance the position to the last byte of the matchrT)�indexr$r(rs�
ValueErrorr�r2rrr�jumpTo�s
 
zEncodingBytes.jumpToN)r@rArBrCr�r&r�r�r�r�r�r��propertyr$r��currentByte�spaceCharactersBytesr�r�r�r�rrrrr�Bs 	
	r�c@sXeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�ZdS)r�z?Mini parser for detecting character encoding from meta elementscCst|�|_d|_dS)z3string - the data to work on for encoding detectionN)r�r8r��r%r8rrrr&�s
zEncodingParser.__init__c
Cs�d|jvrdSd|jfd|jfd|jfd|jfd|jfd|jff}|jD]�}d}z|j�d�WntyxYq�Yn0|D]B\}}|j�|�r~z|�}Wq�Wq~ty�d}Yq�Yq~0q~|sHq�qH|j	S)	Ns<metas<!--s</s<!s<?rTF)
r8�
handleComment�
handleMeta�handlePossibleEndTag�handleOther�handlePossibleStartTagr�r�r�r�)r%�methodDispatchr��keepParsing�key�methodrrrr��s4
�

zEncodingParser.getEncodingcCs|j�d�S)zSkip over commentss-->�r8r�r6rrrr��szEncodingParser.handleCommentcCs�|jjtvrdSd}d}|��}|dur,dS|ddkr\|ddk}|r�|dur�||_dSq|ddkr�|d}t|�}|dur�||_dSq|ddkrtt|d��}|��}|durt|�}|dur|r�||_dS|}qdS)	NTFrs
http-equivrscontent-type�charsetscontent)	r8r�r��getAttributer�r^�ContentAttrParserr��parse)r%�	hasPragma�pendingEncoding�attr�tentativeEncoding�codec�
contentParserrrrr��s8zEncodingParser.handleMetacCs
|�d�S)NF)�handlePossibleTagr6rrrr��sz%EncodingParser.handlePossibleStartTagcCst|j�|�d�S)NT)r�r8rr6rrrr��s
z#EncodingParser.handlePossibleEndTagcCsb|j}|jtvr(|r$|��|��dS|�t�}|dkrD|��n|��}|dur^|��}qLdS)NTr)r8r��asciiLettersBytesr�r�r��spacesAngleBracketsr�)r%�endTagr8r�r�rrrr�s



z EncodingParser.handlePossibleTagcCs|j�d�S)Nrr�r6rrrr�szEncodingParser.handleOthercCs�|j}|�ttdg�B�}|dus2t|�dks2J�|dvr>dSg}g}|dkrV|rVq�nX|tvrj|��}q�nD|dvr�d�|�dfS|tvr�|�|���n|dur�dS|�|�t	|�}qF|dkr�|�
�d�|�dfSt	|�|��}|dv�rJ|}t	|�}||k�r"t	|�d�|�d�|�fS|tv�r<|�|���q�|�|�q�nJ|d	k�rbd�|�dfS|tv�r||�|���n|du�r�dS|�|�t	|�}|tv�r�d�|�d�|�fS|tv�r�|�|���n|du�r�dS|�|��q�dS)
z_Return a name,value pair for the next attribute in the stream,
        if one is found, or None�/Nr)rN�=)rrr)�'�"r)r8r�r��	frozensetr(r9�asciiUppercaseBytesr7r�r�r�r)r%r8r��attrName�	attrValue�	quoteCharrrrr�sb











zEncodingParser.getAttributeN)
r@rArBrCr&r�r�r�r�r�rr�r�rrrrr��s$r�c@seZdZdd�Zdd�ZdS)r�cCst|t�sJ�||_dSr�)rGr3r8r�rrrr&aszContentAttrParser.__init__cCs�z�|j�d�|jjd7_|j��|jjdks<WdS|jjd7_|j��|jjdvr�|jj}|jjd7_|jj}|j�|�r�|j||jj�WSWdSnP|jj}z"|j�t�|j||jj�WWSt�y�|j|d�YWS0Wnt�yYdS0dS)Nr�rr)r	r)r8r�r$r�r�r�r�r�)r%�	quoteMark�oldPositionrrrr�es.

zContentAttrParser.parseN)r@rArBr&r�rrrrr�`sr�cCsbt|t�r.z|�d�}Wnty,YdS0|durZzt�|�WStyVYdS0ndSdS)z{Return the python codec name corresponding to an encoding or None if the
    string doesn't correspond to a valid encoding.rN)rGr3�decode�UnicodeDecodeErrorr
�lookup�AttributeError)r�rrrr^�s

r^)0�
__future__rrrZpip._vendor.sixr�pip._vendor.six.movesrrr�r��iorr	�pip._vendorr
�	constantsrr
rrrrcrr
r�rrr�invalid_unicode_no_surrogaterYrmr��evalr�r��ascii_punctuation_rer��objectr!rTrNrOr3r�r�r�r^rrrr�<module>sH��

JgIb='site-packages/pip/_vendor/html5lib/__pycache__/constants.cpython-39.pyc000064400000201513150564745010022116 0ustar00a

��ReF��@s~-ddlmZmZmZddlZdZddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d,d,d0d1d2d3d4d5d6d7d8d9d:d;d<d=d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d���Zd�d�d�d�d�d�d��Zeed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fg�Z	eed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fg�Z
eed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fgN�Zeed�d�fed�d�fed�d�fed�d�fg�Zeed�d�fed�d�fed�d�fed�d�fed�d�fg�Z
d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d�d�d�d�d�d�d�d�d�d	�d
�d�d�d
�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d �d!�d"�d#�d$�d%�d&�d'�d(�d)�d*�d+�d,�d-�d.�>Z�d/�d0iZ�d1�d2e�d1f�d1�d3e�d1f�d1�d4e�d1f�d1�d5e�d1f�d1�d6e�d1f�d1d�e�d1f�d1�d7e�d1f�d8d�e�d8f�d8�d9e�d8f�d8�d:e�d8fd�d;e�d;f�d;�d1e�d;f�d<�Z�d=�d>�e��D�Zeg�d?��Zeg�d@��Zeej�Zeej�Zeej�Zeej�Zeej�Z�dA�d>�ejD�Z�dBZeg�dC��Z ed�d�g�Z!eg�dD��Z"e�dE�dFg�e�dGg�e�dHg�e�dI�dJg�e�dI�dJg�e�dK�dLg�e�dMg�e�dN�dOg�eg�dP��e�dQg�e�dRg�e�dO�dSg�eg�dT��e�dO�dSg�e�dO�dUg�eg�dV��eg�dW��e�dO�dSg�e�dXg��dY�Z#�dZZ$eg�d[��Z%�d\�d\�d]�d]�d^�d^�d_�d`�d`�da�db�dc�dc�dd�de�df�dg�dh�di�dj�dj�dk�dl�dm�dm�dn�dn�do�dp�dq�dr�ds�dt�du�dv�dw�dx�dt�dy�dz�d{�d{�d|�d}�d~�d�d��d��d��d��d��d��d��d��d�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d~�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��ddÐdĐdŐdƐdǐdȐdɐd��dʐdːd̐d͐d͐dΐdΐdϐdАdАdѐdҐdӐdԐdԐdՐd֐dאdؐdِdڐdېdܐdݐdސdߐd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d��d��d��d��d��d��d��d��d��d��d�d�d�d�d�d�d�d�d�d�d	�dy�d
�d�d�d
�d�d�d�d�d�d�d�d�d�d�d�d�d��d�d�d�d�d�d�d�d�d�d�d �d!�d!�d"�d#�d$�d%�d&�d'�d(�d)�d*�d+�d,�d-�d.�d/�d0�d1�d2�d2�d3�d4�d5�d6�d7�d8�d9�d:�d;�d<�d=�d>�d?�d@�dA�dB�dC�dD�dE�dF�dG�dH�dI�dJ�dK�dL�dM�dN�dO�dP�dQ�dR�d��d��dS�dT�dU�dV�dW�dX�dY�dZ�d[�d\�d]�d^�d_�d��d��d��d`�da�db�d6�dc�dd�de�df�dg�dh�di�dj�dk�dl�di�dm�dn�do�dp�dq�dr�ds�ds�ds�ds�d�de�dt�du�dv�dw�dx�dy�dz�d{�d|�d}�d~�d�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��dddÐdĐdŐdƐdǐdȐdɐdʐdːd̐d�d͐dΐdϐdАdѐdҐdӐdԐd��dՐd֐dאdؐdؐdِdڐdېdܐdݐdݐdސdߐd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d��d��d��d��d��d��d��d��d��d��d��d�d�d�d�d�d�d�d�d�d	�d
�d�d�d
�d�d�d��d<�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d �d!�d"�d#�d�d$�d%�d&�d'�d%�d(�d(�d)�d*�d+�d,�d-�d.�d/�d0�d1�d2�d3�d4�d5�d6�d7�d8�d9�d:�d;�d<�d=�d>�d>�d?�d@�dA�dB�dC�dC�dD�dE�dF�dG�dG�dH�dI�dJ�dK�dL�dM�dN�dO�dP�d�dQ�dR�dS�dT�dU�dV�d��d��dW�dX�dY�dZ�d[�d\�d]�d^�d^�d_�d`�da�db�dc�dd�de�de�df�dg�dh�di�dj�dk�dl�dm�dn�do�dp�dq�dr�ds�dt�du�dv�dw�dx�dy�dz�d{�d{�d|�d}�d~�d�d��d��d��d��d��d��d��ds�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��di�d��d��d��d��d��d��d��d��d]�d]�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��dj�d��d��d��d9�d��d��d��d��d��d9�d��d��d��d��d��d9�d��d��d��d��d��d��d��d��d��d��d��d��d��d��dddK�dÐd��dĐdŐds�ds�dƐd��dt�dǐdȐdɐdʐd�dːdM�d̐d͐dΐdϐdАdѐdҐdӐdd�dp�dԐdՐd�d֐dאdؐdِdڐdېdܐdݐdސdߐd�d�d�dU�dU�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d��d��d��d��d��d��d��d�d��d��d��d�d�d�d�d�d�d�d�d�d	�d
�d�d�d
�d�d��dx�d�d�d�d�d��d��d�d�d�d�d�dy�d�d
�d
�d�d�d�d�d�d�d�d�d�d�d �d!�d"�d"�d#�d$�d%�d&�d��d��d'�d(�d(�d��d)�d*�d+�d+�d,�d-�d.�d/�d0�d1�d2�dݐd3�d4�d5�d6�d0�d7�d8�d9�d:�d:�d;�dl�dl�d<�d=�d>�d�d>�d��d8�d?�d��d@�d��d{�d{�dA�dB�dC�dD�dE�dF�dG�dH�dI�dJ�dK�dL�dM�dN�dO�dP�dQ�dR�dS�dT�dU�dV�dW�dX�dL�dM�dY�dZ�d[�d[�dN�dW�dY�dZ�d��d\�d]�d��d^�d_�d`�d��da�dG�db�d��dc�dd�d��d��de�df�dg�dg�dh�di�dj�dk�dB�d�d��d��dl�dl�d��dm�dn�do�do�do�dp�dp�dq�dr�ds�dt�du�d��d��dv�dw�dx�dy�dq�d��de�dB�d�dܐdz�d{�d|�d}�d~�d�d��d��dאd��d�d��d��d��df�dv�d��d��d��d��d��d��d��d��d��d��d�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��dݐd��d��d��d��d��d��d��d��d��d��d��dݐd��d��d��d��d��d��d��d�dߐd�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d�d��d��d��dddÐdĐdĐdŐdƐdǐdȐdɐdʐdʐdːd̐d͐dΐdϐdАdѐdҐdӐd��dԐdՐd֐dm�dאdؐdِdڐdېd��d��d��d��d��d��dܐdݐdސdߐd�d�d�d�d��d�d�d��d�d�d�d�d�d�d�d�d�d�d�d��d�d��d�d�d�d�d�d�d�d�d�dאd��d�d��dԐd��d��d��d��d��dj�dd�d��dE�d��d��d��d��d��d��d��d��d��d�d�d�d�d�d�d�d�d�d��d	�d
�da�d�d�d�d�d�d
�d�d�d�d��d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dՐd�d�d�d�d�d�d��d�d�d �d!�d"�d#�d$�d �d%�d%�d&�dՐd'�d(�d)�d*�dՐd�d+�d,�d-�d-�d.�d/�d0�d1�d2�d3�d4�d5�d6�d7�d8�d9�d:�d;�d<�d=�d>�d?�d[�d��d@�dA�dT�dB�dC�dD�dE�d6�dF�d;�dG�d;�dH�dI�dI�d<�d=�dJ�dK�d�dL�dM�dN�dO�dP�dQ�dR�dS�dT�dU�dV�dW�dX�dY�dZ�d[�d\�d?�dV�d]�d^�d��dŐd_�d`�da�db�d<�dO�dÐdQ�dc�dE�d>�d�d��dd�dS�db�dT�dW�dW�de�df�dg�dh�di�dj�dH�dk�dS�dB�dU�dX�dl�dD�dm�dU�dn�dÐdQ�do�dp�dq�de�dc�dr�dr�ds�dt�du�du�dv�dw�dw�dx�dx�dv�dy�dz�d{�d@�d]�d^�d|�d_�dL�d}�d~�d�d��d��d��dI�d��d��dՐd��d��d>�dz�d�d��d��d��d��d��dc�dX�d��d��dW�d��d��d��d2�d2�d��d��dk�dd�d��d��d��d��d��dL�dؐd��d��d��d��d��d��d��d��d��d��d��d��dɐdH�dV�d��d��d��d��d��d��d��d��d��df�d��d��d��d��d��d�dw�d��d��d��dk�d��d��dk�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��dx�dw�dw�d��d��d��d��d��d��d��d��d��dd~�dÐdĐdX�dX�dŐdz�dƐd�d��d��dǐd��d��d��d��d��d��d��d��d��d��dȐdɐd�dʐdːd�d̐d��d͐dΐdϐd��dΐdȐd��d͐d��d��d��d��d��d��d��d��dАdѐdѐd}�dҐdӐd}�dԐdՐd��d��d֐dאd|�d|�dؐdِdڐd��d��d��d��d��d��dېdܐdݐdېd��d��d��d��d��dސd��d|�d��d��d��d��d|�d��d��dߐd�d��d��d��d�d��d��d�d�d��d��d��d�d��d�d�d��d��d��d��d��d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d��d��dW�dW�d��d3�d��d��d4�d5�d��d��d��d6�d��d��d��d��d��d�d�d�d�d�d�d�d��d��d1�d�d�dÐd�d	�d
�d�d�d��d
�d�d�d��d�d2�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d�d�d�d�d��d�d�d��d �d!�dǐd"�d#�d$�d%�dU�d&�d'�d(�d)�di�d*�d+�d��d,�d��d-�d��d.�d/�d�d0�dx�d1�d2�d̐d̐d3�d4�d̐d5�d6�d7�d7�dϐd8�d9�dѐdАdϐd9�dѐdАd:�d;�d<�dҐd=�d͐d;�d:�d<�dԐd>�d?�d@�dՐdՐdҐdA�dB�dC�dD�dE�d�dF�dG�dH�d�dI�dJ�d��dؐdؐd�d��dK�db�dL�dM�dN�d�dO�d�dP�dQ�d�dR�dR�d�dS�d�dT�dU�dV�d�d}�dW�dX�dY�dZ�d[�d\�dڐdԐd]�d^�d_�d`�da�db�dc�dd�d�d^�de�df�dg�d��d��dh�d�d�d�d�di�dݐdݐdj�d�dk�dƐd��dl�dm�dn�d�dY�dƐd��d�dސdo�dZ�dp�dq�d��d�dސdr�ds�ds�dt�du�dv�d�dw�dx�dy�dz�d{�d|�d}�do�d~�d�d�d_�d��d��dp�d��d��d��dِd��d��d��d��d��d �d��d��d��d"�d!�d��d��d��d��d��d��d#�d��d��dy�d��d��d�db�db�d��d��d��d��do�do�d��d��dҐd��d��d��df�d��d��d��d��d��d��d6�d��d7�d7�d��d��d��d��d��d��d��d<�do�d��d��df�d��d��d��d��d��d��d��d��d��d��d��d��d�d��d�d��d�d�d�d�d�d�d�d�d�d�d�d�d�d��do�d��d�d��dАd��d)�d��d��d��d��d�d��d��d��d��d��d��d��d�d��d��d��d��d��d��d �d��d"�d!�d��d��d��d#�d$�ddÐdÐdĐdĐdŐdŐd&�dƐdǐdȐd'�dɐdʐdːd̐d͐dΐdϐdАd&�d'�dƐdϐdΐdѐdҐdӐdԐd�da�da�dՐd֐d֐dאdؐdŐdِdڐdېd;�dܐdݐd2�d2�dސdߐdߐd9�d6�d5�d9�d6�d�d�d��d�d�d�d�d�d�dƐdȐd�d�d�d�d��d�d)�d�d��d��dL�d�d��d��d�d�d�d�d�d�d�d�d�d�d�dɐd7�d�d��d��d��d��d�d��d��d��d��d��dR�d��dT�d��d��d��d��dO�d��d��d�d�d�d�d�d��d��d�d�d�dS�dO�d��dN�d�dY�d�d�d�d�d	�d
�d�d�d
�d�d�d֐d�d�d�d�d��d�d�d��d�d��d7�d��d)�d,�dՐdS�dn�d��d�d�d�d�dߐdJ�d��d�d�d�d�d�d�dg�dg�d�dJ�d��d��d�dՐd��d�d�d�d�d�d�d �d!�d��d"�d#�d$�d%�d#�di�di�d&�d�dːdM�dѐd'�d��d^�d(�d��d]�d|�d)�d̐d*�d͐dΐd��d_�d+�dϐdӐdҐdd�dp�d,�d,�d-�d.�d/�d0�d0�d1�d2�d3�d4�d5�d6�d6�d7�d8�d9�d:�d��d;�d<�d=�d>�d?�d@�dA�dB�dC���Z&�dD�dE�d��dF�d��d��dŐd��d_�d��d/�d%�d�d��d��dG�d��dH�dI�d��d��d��d��d�dd��d��d)�d��d~�d��dJ�d8�d��dK�"Z'd�dL�dM�dN�dO�dP�dQ�dR�dS�Z(ee(�dTe(�dUe(�dVg�Z)�dW�d>�e��D�Z*�dXe*d�<G�dY�dZ��dZe+�Z,G�d[�d\��d\e-�Z.dS(]�)�absolute_import�division�unicode_literalsNz5Null character in input stream, replaced with U+FFFD.zInvalid codepoint in stream.z&Solidus (/) incorrectly placed in tag.z.Incorrect CR newline entity, replaced with LF.z9Entity used with illegal number (windows-1252 reference).zPNumeric entity couldn't be converted to character (codepoint U+%(charAsInt)08x).zBNumeric entity represents an illegal codepoint: U+%(charAsInt)08x.z#Numeric entity didn't end with ';'.z1Numeric entity expected. Got end of file instead.z'Numeric entity expected but none found.z!Named entity didn't end with ';'.z Named entity expected. Got none.z'End tag contains unexpected attributes.z.End tag contains unexpected self-closing flag.z#Expected tag name. Got '>' instead.zSExpected tag name. Got '?' instead. (HTML doesn't support processing instructions.)z-Expected tag name. Got something else insteadz6Expected closing tag. Got '>' instead. Ignoring '</>'.z-Expected closing tag. Unexpected end of file.z<Expected closing tag. Unexpected character '%(data)s' found.z'Unexpected end of file in the tag name.z8Unexpected end of file. Expected attribute name instead.z)Unexpected end of file in attribute name.z#Invalid character in attribute namez#Dropped duplicate attribute on tag.z1Unexpected end of file. Expected = or end of tag.z1Unexpected end of file. Expected attribute value.z*Expected attribute value. Got '>' instead.z"Unexpected = in unquoted attributez*Unexpected character in unquoted attributez*Unexpected character after attribute name.z+Unexpected character after attribute value.z.Unexpected end of file in attribute value (").z.Unexpected end of file in attribute value (').z*Unexpected end of file in attribute value.z)Unexpected end of file in tag. Expected >z/Unexpected character after / in tag. Expected >z&Expected '--' or 'DOCTYPE'. Not found.z Unexpected ! after -- in commentz$Unexpected space after -- in commentzIncorrect comment.z"Unexpected end of file in comment.z%Unexpected end of file in comment (-)z+Unexpected '-' after '--' found in comment.z'Unexpected end of file in comment (--).z&Unexpected character in comment found.z(No space after literal string 'DOCTYPE'.z.Unexpected > character. Expected DOCTYPE name.z.Unexpected end of file. Expected DOCTYPE name.z'Unexpected end of file in DOCTYPE name.z"Unexpected end of file in DOCTYPE.z%Expected space or '>'. Got '%(data)s'zUnexpected end of DOCTYPE.z Unexpected character in DOCTYPE.zXXX innerHTML EOFzUnexpected DOCTYPE. Ignored.z%html needs to be the first start tag.z)Unexpected End of file. Expected DOCTYPE.zErroneous DOCTYPE.z2Unexpected non-space characters. Expected DOCTYPE.z2Unexpected start tag (%(name)s). Expected DOCTYPE.z0Unexpected end tag (%(name)s). Expected DOCTYPE.z?Unexpected end tag (%(name)s) after the (implied) root element.z4Unexpected end of file. Expected end tag (%(name)s).z4Unexpected start tag head in existing head. Ignored.z'Unexpected end tag (%(name)s). Ignored.z;Unexpected start tag (%(name)s) that can be in head. Moved.z Unexpected start tag (%(name)s).zMissing end tag (%(name)s).zMissing end tags (%(name)s).zCUnexpected start tag (%(startName)s) implies end tag (%(endName)s).z@Unexpected start tag (%(originalName)s). Treated as %(newName)s.z,Unexpected start tag %(name)s. Don't use it!z'Unexpected start tag %(name)s. Ignored.zEUnexpected end tag (%(gotName)s). Missing end tag (%(expectedName)s).z:End tag (%(name)s) seen too early. Expected other end tag.zFUnexpected end tag (%(gotName)s). Expected end tag (%(expectedName)s).z+End tag (%(name)s) seen too early. Ignored.zQEnd tag (%(name)s) violates step 1, paragraph 1 of the adoption agency algorithm.zQEnd tag (%(name)s) violates step 1, paragraph 2 of the adoption agency algorithm.zQEnd tag (%(name)s) violates step 1, paragraph 3 of the adoption agency algorithm.zQEnd tag (%(name)s) violates step 4, paragraph 4 of the adoption agency algorithm.z>Unexpected end tag (%(originalName)s). Treated as %(newName)s.z'This element (%(name)s) has no end tag.z9Unexpected implied end tag (%(name)s) in the table phase.z>Unexpected implied end tag (%(name)s) in the table body phase.zDUnexpected non-space characters in table context caused voodoo mode.z3Unexpected input with type hidden in table context.z!Unexpected form in table context.zDUnexpected start tag (%(name)s) in table context caused voodoo mode.zBUnexpected end tag (%(name)s) in table context caused voodoo mode.zCUnexpected table cell start tag (%(name)s) in the table body phase.zFGot table cell end tag (%(name)s) while required end tags are missing.z?Unexpected end tag (%(name)s) in the table body phase. Ignored.z=Unexpected implied end tag (%(name)s) in the table row phase.z>Unexpected end tag (%(name)s) in the table row phase. Ignored.zJUnexpected select start tag in the select phase treated as select end tag.z/Unexpected input start tag in the select phase.zBUnexpected start tag token (%(name)s in the select phase. Ignored.z;Unexpected end tag (%(name)s) in the select phase. Ignored.zKUnexpected table element start tag (%(name)s) in the select in table phase.zIUnexpected table element end tag (%(name)s) in the select in table phase.z8Unexpected non-space characters in the after body phase.z>Unexpected start tag token (%(name)s) in the after body phase.z<Unexpected end tag token (%(name)s) in the after body phase.z@Unexpected characters in the frameset phase. Characters ignored.zEUnexpected start tag token (%(name)s) in the frameset phase. Ignored.zFUnexpected end tag token (frameset) in the frameset phase (innerHTML).zCUnexpected end tag token (%(name)s) in the frameset phase. Ignored.zEUnexpected non-space characters in the after frameset phase. Ignored.zEUnexpected start tag (%(name)s) in the after frameset phase. Ignored.zCUnexpected end tag (%(name)s) in the after frameset phase. Ignored.z(Unexpected end tag after body(innerHtml)z6Unexpected non-space characters. Expected end of file.z6Unexpected start tag (%(name)s). Expected end of file.z4Unexpected end tag (%(name)s). Expected end of file.z/Unexpected end of file. Expected table content.z0Unexpected end of file. Expected select content.z2Unexpected end of file. Expected frameset content.z0Unexpected end of file. Expected script content.z0Unexpected end of file. Expected foreign contentz0Trailing solidus not allowed on element %(name)sz2Element %(name)s not allowed in a non-html contextz*Unexpected end tag (%(name)s) before html.z9Element %(name)s not allowed in a inhead-noscript contextz8Unexpected end of file. Expected inhead-noscript contentz@Unexpected non-space character. Expected inhead-noscript contentz0Undefined error (this sucks and should be fixed))�znull-characterzinvalid-codepointzincorrectly-placed-soliduszincorrect-cr-newline-entityzillegal-windows-1252-entityzcant-convert-numeric-entityz$illegal-codepoint-for-numeric-entityz numeric-entity-without-semicolonz#expected-numeric-entity-but-got-eofzexpected-numeric-entityznamed-entity-without-semicolonzexpected-named-entityzattributes-in-end-tagzself-closing-flag-on-end-tagz'expected-tag-name-but-got-right-bracketz'expected-tag-name-but-got-question-markzexpected-tag-namez*expected-closing-tag-but-got-right-bracketz expected-closing-tag-but-got-eofz!expected-closing-tag-but-got-charzeof-in-tag-namez#expected-attribute-name-but-got-eofzeof-in-attribute-namez#invalid-character-in-attribute-namezduplicate-attributez$expected-end-of-tag-name-but-got-eofz$expected-attribute-value-but-got-eofz.expected-attribute-value-but-got-right-bracketz"equals-in-unquoted-attribute-valuez0unexpected-character-in-unquoted-attribute-valuez&invalid-character-after-attribute-namez*unexpected-character-after-attribute-valuez#eof-in-attribute-value-double-quotez#eof-in-attribute-value-single-quotez eof-in-attribute-value-no-quotesz#unexpected-EOF-after-solidus-in-tagz)unexpected-character-after-solidus-in-tagzexpected-dashes-or-doctypez,unexpected-bang-after-double-dash-in-commentz-unexpected-space-after-double-dash-in-commentzincorrect-commentzeof-in-commentzeof-in-comment-end-dashz,unexpected-dash-after-double-dash-in-commentzeof-in-comment-double-dashzeof-in-comment-end-space-statezeof-in-comment-end-bang-statezunexpected-char-in-commentzneed-space-after-doctypez+expected-doctype-name-but-got-right-bracketz!expected-doctype-name-but-got-eofzeof-in-doctype-namezeof-in-doctypez*expected-space-or-right-bracket-in-doctypezunexpected-end-of-doctypezunexpected-char-in-doctypezeof-in-innerhtmlzunexpected-doctypez
non-html-rootzexpected-doctype-but-got-eofzunknown-doctypezexpected-doctype-but-got-charsz"expected-doctype-but-got-start-tagz expected-doctype-but-got-end-tagzend-tag-after-implied-rootz&expected-named-closing-tag-but-got-eofz!two-heads-are-not-better-than-onezunexpected-end-tagz#unexpected-start-tag-out-of-my-headzunexpected-start-tagzmissing-end-tagzmissing-end-tagsz$unexpected-start-tag-implies-end-tagzunexpected-start-tag-treated-aszdeprecated-tagzunexpected-start-tag-ignoredz$expected-one-end-tag-but-got-anotherzend-tag-too-earlyzend-tag-too-early-namedzend-tag-too-early-ignoredzadoption-agency-1.1zadoption-agency-1.2zadoption-agency-1.3zadoption-agency-4.4zunexpected-end-tag-treated-asz
no-end-tagz#unexpected-implied-end-tag-in-tablez(unexpected-implied-end-tag-in-table-bodyz$unexpected-char-implies-table-voodooz unexpected-hidden-input-in-tablezunexpected-form-in-tablez)unexpected-start-tag-implies-table-voodooz'unexpected-end-tag-implies-table-voodoozunexpected-cell-in-table-bodyzunexpected-cell-end-tagz unexpected-end-tag-in-table-bodyz'unexpected-implied-end-tag-in-table-rowzunexpected-end-tag-in-table-rowzunexpected-select-in-selectzunexpected-input-in-selectzunexpected-start-tag-in-selectzunexpected-end-tag-in-selectz5unexpected-table-element-start-tag-in-select-in-tablez3unexpected-table-element-end-tag-in-select-in-tablezunexpected-char-after-bodyzunexpected-start-tag-after-bodyzunexpected-end-tag-after-bodyzunexpected-char-in-framesetz unexpected-start-tag-in-framesetz)unexpected-frameset-in-frameset-innerhtmlzunexpected-end-tag-in-framesetzunexpected-char-after-framesetz#unexpected-start-tag-after-framesetz!unexpected-end-tag-after-framesetz'unexpected-end-tag-after-body-innerhtmlzexpected-eof-but-got-charzexpected-eof-but-got-start-tagzexpected-eof-but-got-end-tagzeof-in-tablez
eof-in-selectzeof-in-framesetzeof-in-script-in-scriptzeof-in-foreign-landsz&non-void-element-with-trailing-solidusz*unexpected-html-element-in-foreign-contentzunexpected-end-tag-before-htmlzunexpected-inhead-noscript-tagzeof-in-head-noscriptzchar-in-head-noscriptzXXX-undefined-errorzhttp://www.w3.org/1999/xhtmlz"http://www.w3.org/1998/Math/MathMLzhttp://www.w3.org/2000/svgzhttp://www.w3.org/1999/xlinkz$http://www.w3.org/XML/1998/namespacezhttp://www.w3.org/2000/xmlns/)�html�mathml�svg�xlink�xml�xmlnsr�applet�caption�marquee�object�table�td�thr�mi�mo�mn�ms�mtextzannotation-xmlr�
foreignObject�desc�title�a�b�big�code�em�font�i�nobr�s�small�strike�strong�tt�u�address�area�article�aside�base�basefont�bgsound�
blockquote�body�br�button�center�col�colgroup�command�dd�details�dir�div�dl�dt�embed�fieldset�figure�footer�form�frame�frameset�h1�h2�h3�h4�h5�h6�head�header�hr�iframe�image�img�input�isindex�li�link�listing�menu�meta�nav�noembed�noframes�noscript�ol�p�param�	plaintext�pre�script�section�select�style�tbody�textarea�tfoot�thead�tr�ul�wbr�xmp�
attributeName�
attributeType�
baseFrequency�baseProfile�calcMode�
clipPathUnits�contentScriptType�contentStyleType�diffuseConstant�edgeMode�externalResourcesRequired�	filterRes�filterUnits�glyphRef�gradientTransform�
gradientUnits�kernelMatrix�kernelUnitLength�	keyPoints�
keySplines�keyTimes�lengthAdjust�limitingConeAngle�markerHeight�markerUnits�markerWidth�maskContentUnits�	maskUnits�
numOctaves�
pathLength�patternContentUnits�patternTransform�patternUnits�	pointsAtX�	pointsAtY�	pointsAtZ�
preserveAlpha�preserveAspectRatio�primitiveUnits�refX�refY�repeatCount�	repeatDur�requiredExtensions�requiredFeatures�specularConstant�specularExponent�spreadMethod�startOffset�stdDeviation�stitchTiles�surfaceScale�systemLanguage�tableValues�targetX�targetY�
textLength�viewBox�
viewTarget�xChannelSelector�yChannelSelector�
zoomAndPan)>�
attributename�
attributetype�
basefrequency�baseprofile�calcmode�
clippathunits�contentscripttype�contentstyletype�diffuseconstant�edgemode�externalresourcesrequired�	filterres�filterunits�glyphref�gradienttransform�
gradientunits�kernelmatrix�kernelunitlength�	keypoints�
keysplines�keytimes�lengthadjust�limitingconeangle�markerheight�markerunits�markerwidth�maskcontentunits�	maskunits�
numoctaves�
pathlength�patterncontentunits�patterntransform�patternunits�	pointsatx�	pointsaty�	pointsatz�
preservealpha�preserveaspectratio�primitiveunits�refx�refy�repeatcount�	repeatdur�requiredextensions�requiredfeatures�specularconstant�specularexponent�spreadmethod�startoffset�stddeviation�stitchtiles�surfacescale�systemlanguage�tablevalues�targetx�targety�
textlength�viewbox�
viewtarget�xchannelselector�ychannelselector�
zoomandpan�
definitionurl�
definitionURLr�actuate�arcrole�href�role�show�typer	�lang�spacer
)z
xlink:actuatez
xlink:arcrolez
xlink:hrefz
xlink:rolez
xlink:showzxlink:titlez
xlink:typezxml:basezxml:langz	xml:spacer
zxmlns:xlinkcCs i|]\}\}}}||f|�qS�r�)�.0�qname�prefix�local�nsr�r���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/constants.py�
<dictcomp>
�r�)�	�
�� �
)rrdrfrgrhcCsi|]}t|�t|����qSr�)�ord�lower)r��cr�r�r�r�#r�)rDrErFrGrHrI)r,r6zevent-sourcerSrVrLr1rOr=r]r)r4rP�source�track)rcr`rkrMrXrYrZ�
irrelevant�	itemscope�scoped�ismap�autoplay�controls�defer�async�open�multiple�disabled)�hiddenr�checked�default�noshade�
autosubmit�readonly)rr�selected�	autofocus)rr�requiredrrr)rrrr�seamless)�rcrO�audio�videor`r8�datagridr6rLrUr>�option�optgroupr2rPrb�outputrM) i� �i i�i i& i  i! i�i0 i`i9 iRr!i}r!r!i i i i i" i i i�i"!iai: iSr!i~ix)�lt;�gt;�amp;�apos;�quot;�Æ�&�ÁuĂ�ÂuАu𝔄�ÀuΑuĀu⩓uĄu𝔸u⁡�Åu𝒜u≔�Ã�Äu∖u⫧u⌆uБu∵uℬuΒu𝔅u𝔹u˘u≎uЧ�©uĆu⋒uⅅuℭuČ�ÇuĈu∰uĊ�¸�·uΧu⊙u⊖u⊕u⊗u∲u”u’u∷u⩴u≡u∯u∮uℂu∐u∳u⨯u𝒞u⋓u≍u⤑uЂuЅuЏu‡u↡u⫤uĎuДu∇uΔu𝔇�´u˙u˝�`u˜u⋄uⅆu𝔻�¨u⃜u≐u⇓u⇐u⇔u⟸u⟺u⟹u⇒u⊨u⇑u⇕u∥u↓u⤓u⇵ȗu⥐u⥞u↽u⥖u⥟u⇁u⥗u⊤u↧u𝒟uĐuŊ�Ð�ÉuĚ�ÊuЭuĖu𝔈�Èu∈uĒu◻u▫uĘu𝔼uΕu⩵u≂u⇌uℰu⩳uΗ�Ëu∃uⅇuФu𝔉u◼u▪u𝔽u∀uℱuЃ�>uΓuϜuĞuĢuĜuГuĠu𝔊u⋙u𝔾u≥u⋛u≧u⪢u≷u⩾u≳u𝒢u≫uЪuˇ�^uĤuℌuℋuℍu─uĦu≏uЕuIJuЁ�Í�ÎuИuİuℑ�ÌuĪuⅈu∬u∫u⋂u⁣u⁢uĮu𝕀uΙuℐuĨuІ�ÏuĴuЙu𝔍u𝕁u𝒥uЈuЄuХuЌuΚuĶuКu𝔎u𝕂u𝒦uЉ�<uĹuΛu⟪uℒu↞uĽuĻuЛu⟨u←u⇤u⇆u⌈u⟦u⥡u⇃u⥙u⌊u↔u⥎u⊣u↤u⥚u⊲u⧏u⊴u⥑u⥠u↿u⥘u↼u⥒u⋚u≦u≶u⪡u⩽u≲u𝔏u⋘u⇚uĿu⟵u⟷u⟶u𝕃u↙u↘u↰uŁu≪u⤅uМu uℳu𝔐u∓u𝕄uΜuЊuŃuŇuŅuНu​r�u𝔑u⁠� uℕu⫬u≢u≭u∦u∉u≠u≂̸u∄u≯u≱u≧̸u≫̸u≹u⩾̸u≵u≎̸u≏̸u⋪u⧏̸u⋬u≮u≰u≸u≪̸u⩽̸u≴u⪢̸u⪡̸u⊀u⪯̸u⋠u∌u⋫u⧐̸u⋭u⊏̸u⋢u⊐̸u⋣u⊂⃒u⊈u⊁u⪰̸u⋡u≿̸u⊃⃒u⊉u≁u≄u≇u≉u∤u𝒩�ÑuΝuŒ�Ó�ÔuОuŐu𝔒�ÒuŌuΩuΟu𝕆u“u‘u⩔u𝒪�Ø�Õu⨷�Öu‾u⏞u⎴u⏜u∂uПu𝔓uΦuΠ�±uℙu⪻u≺u⪯u≼u≾u″u∏u∝u𝒫uΨ�"u𝔔uℚu𝒬u⤐�®uŔu⟫u↠u⤖uŘuŖuРuℜu∋u⇋u⥯uΡu⟩u→u⇥u⇄u⌉u⟧u⥝u⇂u⥕u⌋u⊢u↦u⥛u⊳u⧐u⊵u⥏u⥜u↾u⥔u⇀u⥓uℝu⥰u⇛uℛu↱u⧴uЩuШuЬuŚu⪼uŠuŞuŜuСu𝔖u↑uΣu∘u𝕊u√u□u⊓u⊏u⊑u⊐u⊒u⊔u𝒮u⋆u⋐u⊆u≻u⪰u≽u≿u∑u⋑u⊃u⊇�Þu™uЋuЦr�uΤuŤuŢuТu𝔗u∴uΘu  u u∼u≃u≅u≈u𝕋u⃛u𝒯uŦ�Úu↟u⥉uЎuŬ�ÛuУuŰu𝔘�ÙuŪ�_u⏟u⎵u⏝u⋃u⊎uŲu𝕌u⤒u⇅u↕u⥮u⊥u↥u↖u↗uϒuΥuŮu𝒰uŨ�Üu⊫u⫫uВu⊩u⫦u⋁u‖u∣�|u❘u≀u u𝔙u𝕍u𝒱u⊪uŴu⋀u𝔚u𝕎u𝒲u𝔛uΞu𝕏u𝒳uЯuЇuЮ�ÝuŶuЫu𝔜u𝕐u𝒴uŸuЖuŹuŽuЗuŻuΖuℨuℤu𝒵�áuău∾u∾̳u∿�âuа�æu𝔞�àuℵuαuāu⨿u∧u⩕u⩜u⩘u⩚u∠u⦤u∡u⦨u⦩u⦪u⦫u⦬u⦭u⦮u⦯u∟u⊾u⦝u∢u⍼uąu𝕒u⩰u⩯u≊u≋�'�åu𝒶�*�ã�äu⨑u⫭u≌u϶u‵u∽u⋍u⊽u⌅u⎶uбu„u⦰uβuℶu≬u𝔟u◯u⨀u⨁u⨂u⨆u★u▽u△u⨄u⤍u⧫u▴u▾u◂u▸u␣u▒u░u▓u█u=⃥u≡⃥u⌐u𝕓u⋈u╗u╔u╖u╓u═u╦u╩u╤u╧u╝u╚u╜u╙u║u╬u╣u╠u╫u╢u╟u⧉u╕u╒u┐u┌u╥u╨u┬u┴u⊟u⊞u⊠u╛u╘u┘u└u│u╪u╡u╞u┼u┤u├�¦u𝒷u⁏�\u⧅u⟈u•u⪮uću∩u⩄u⩉u⩋u⩇u⩀u∩︀u⁁u⩍uč�çuĉu⩌u⩐uċu⦲�¢u𝔠uчu✓uχu○u⧃uˆu≗u↺u↻uⓈu⊛u⊚u⊝u⨐u⫯u⧂u♣�:�,�@u∁u⩭u𝕔u℗u↵u✗u𝒸u⫏u⫑u⫐u⫒u⋯u⤸u⤵u⋞u⋟u↶u⤽u∪u⩈u⩆u⩊u⊍u⩅u∪︀u↷u⤼u⋎u⋏�¤u∱u⌭u⥥u†uℸu‐u⤏uďuдu⇊u⩷�°uδu⦱u⥿u𝔡u♦uϝu⋲�÷u⋇uђu⌞u⌍�$u𝕕u≑u∸u∔u⊡u⌟u⌌u𝒹uѕu⧶uđu⋱u▿u⦦uџu⟿�éu⩮uěu≖�êu≕uэuėu≒u𝔢u⪚�èu⪖u⪘u⪙u⏧uℓu⪕u⪗uēu∅u u u uŋu uęu𝕖u⋕u⧣u⩱uεuϵ�=u≟u⩸u⧥u≓u⥱uℯuη�ð�ëu€�!uфu♀uffiuffufflu𝔣ufi�fju♭uflu▱uƒu𝕗u⋔u⫙u⨍�½u⅓�¼u⅕u⅙u⅛u⅔u⅖�¾u⅗u⅜u⅘u⅚u⅝u⅞u⁄u⌢u𝒻u⪌uǵuγu⪆uğuĝuгuġu⪩u⪀u⪂u⪄u⋛︀u⪔u𝔤uℷuѓu⪒u⪥u⪤u≩u⪊u⪈u⋧u𝕘uℊu⪎u⪐u⪧u⩺u⋗u⦕u⩼u⥸u≩︀uъu⥈u↭uℏuĥu♥u…u⊹u𝔥u⤥u⤦u⇿u∻u↩u↪u𝕙u―u𝒽uħu⁃�í�îuиuе�¡u𝔦�ìu⨌u∭u⧜u℩uijuīuıu⊷uƵu℅u∞u⧝u⊺u⨗u⨼uёuįu𝕚uι�¿u𝒾u⋹u⋵u⋴u⋳uĩuі�ïuĵuйu𝔧uȷu𝕛u𝒿uјuєuκuϰuķuкu𝔨uĸuхuќu𝕜u𝓀u⤛u⤎u⪋u⥢uĺu⦴uλu⦑u⪅�«u⤟u⤝u↫u⤹u⥳u↢u⪫u⤙u⪭u⪭︀u⤌u❲�{�[u⦋u⦏u⦍uľuļuлu⤶u⥧u⥋u↲u≤u⇇u⋋u⪨u⩿u⪁u⪃u⋚︀u⪓u⋖u⥼u𝔩u⪑u⥪u▄uљu⥫u◺uŀu⎰u≨u⪉u⪇u⋦u⟬u⇽u⟼u↬u⦅u𝕝u⨭u⨴u∗u◊�(u⦓u⥭u‎u⊿u‹u𝓁u⪍u⪏u‚ułu⪦u⩹u⋉u⥶u⩻u⦖u◃u⥊u⥦u≨︀u∺�¯u♂u✠u▮u⨩uмu—u𝔪u℧�µu⫰u−u⨪u⫛u⊧u𝕞u𝓂uμu⊸u⋙̸u≫⃒u⇍u⇎u⋘̸u≪⃒u⇏u⊯u⊮uńu∠⃒u⩰̸u≋̸uʼnu♮u⩃uňuņu⩭̸u⩂uнu–u⇗u⤤u≐̸u⤨u𝔫u↮u⫲u⋼u⋺uњu≦̸u↚u‥u𝕟�¬u⋹̸u⋵̸u⋷u⋶u⋾u⋽u⫽⃥u∂̸u⨔u↛u⤳̸u↝̸u𝓃u⊄u⫅̸u⊅u⫆̸�ñuν�#u№u u⊭u⤄u≍⃒u⊬u≥⃒u>⃒u⧞u⤂u≤⃒u<⃒u⊴⃒u⤃u⊵⃒u∼⃒u⇖u⤣u⤧�ó�ôuоuőu⨸u⦼uœu⦿u𝔬u˛�òu⧁u⦵u⦾u⦻u⧀uōuωuοu⦶u𝕠u⦷u⦹u∨u⩝uℴ�ª�ºu⊶u⩖u⩗u⩛�øu⊘�õu⨶�öu⌽�¶u⫳u⫽uп�%�.u‰u‱u𝔭uφuϕu☎uπuϖuℎ�+u⨣u⨢u⨥u⩲u⨦u⨧u⨕u𝕡�£u⪳u⪷u⪹u⪵u⋨u′u⌮u⌒u⌓u⊰u𝓅uψu u𝔮u𝕢u⁗u𝓆u⨖�?u⤜u⥤u∽̱uŕu⦳u⦒u⦥�»u⥵u⤠u⤳u⤞u⥅u⥴u↣u↝u⤚u∶u❳�}�]u⦌u⦎u⦐uřuŗuрu⤷u⥩u↳u▭u⥽u𝔯u⥬uρuϱu⇉u⋌u˚u‏u⎱u⫮u⟭u⇾u⦆u𝕣u⨮u⨵�)u⦔u⨒u›u𝓇u⋊u▹u⧎u⥨u℞uśu⪴u⪸ušuşuŝu⪶u⪺u⋩u⨓uсu⋅u⩦u⇘�§�;u⤩u✶u𝔰u♯uщuш�­uσuςu⩪u⪞u⪠u⪝u⪟u≆u⨤u⥲u⨳u⧤u⌣u⪪u⪬u⪬︀uь�/u⧄u⌿u𝕤u♠u⊓︀u⊔︀u𝓈u☆u⊂u⫅u⪽u⫃u⫁u⫋u⊊u⪿u⥹u⫇u⫕u⫓u♪�¹�²�³u⫆u⪾u⫘u⫄u⟉u⫗u⥻u⫂u⫌u⊋u⫀u⫈u⫔u⫖u⇙u⤪�ßu⌖uτuťuţuтu⌕u𝔱uθuϑ�þ�×u⨱u⨰u⌶u⫱u𝕥u⫚u‴u▵u≜u◬u⨺u⨹u⧍u⨻u⏢u𝓉uцuћuŧu⥣�úuўuŭ�ûuуuűu⥾u𝔲�ùu▀u⌜u⌏u◸uūuųu𝕦uυu⇈u⌝u⌎uůu◹u𝓊u⋰uũ�üu⦧u⫨u⫩u⦜u⊊︀u⫋︀u⊋︀u⫌︀uвu⊻u≚u⋮u𝔳u𝕧u𝓋u⦚uŵu⩟u≙u℘u𝔴u𝕨u𝓌u𝔵uξu⋻u𝕩u𝓍�ýuяuŷuы�¥u𝔶uїu𝕪u𝓎uю�ÿuźužuзużuζu𝔷uжu⇝u𝕫u𝓏u‍u‌(��AEligzAElig;�AMPzAMP;�AacutezAacute;zAbreve;�AcirczAcirc;zAcy;zAfr;�AgravezAgrave;zAlpha;zAmacr;zAnd;zAogon;zAopf;zApplyFunction;�AringzAring;zAscr;zAssign;�AtildezAtilde;�AumlzAuml;z
Backslash;zBarv;zBarwed;zBcy;zBecause;zBernoullis;zBeta;zBfr;zBopf;zBreve;zBscr;zBumpeq;zCHcy;�COPYzCOPY;zCacute;zCap;zCapitalDifferentialD;zCayleys;zCcaron;�CcedilzCcedil;zCcirc;zCconint;zCdot;zCedilla;z
CenterDot;zCfr;zChi;z
CircleDot;zCircleMinus;zCirclePlus;zCircleTimes;zClockwiseContourIntegral;zCloseCurlyDoubleQuote;zCloseCurlyQuote;zColon;zColone;z
Congruent;zConint;zContourIntegral;zCopf;z
Coproduct;z CounterClockwiseContourIntegral;zCross;zCscr;zCup;zCupCap;zDD;z	DDotrahd;zDJcy;zDScy;zDZcy;zDagger;zDarr;zDashv;zDcaron;zDcy;zDel;zDelta;zDfr;zDiacriticalAcute;zDiacriticalDot;zDiacriticalDoubleAcute;zDiacriticalGrave;zDiacriticalTilde;zDiamond;zDifferentialD;zDopf;zDot;zDotDot;z	DotEqual;zDoubleContourIntegral;z
DoubleDot;zDoubleDownArrow;zDoubleLeftArrow;zDoubleLeftRightArrow;zDoubleLeftTee;zDoubleLongLeftArrow;zDoubleLongLeftRightArrow;zDoubleLongRightArrow;zDoubleRightArrow;zDoubleRightTee;zDoubleUpArrow;zDoubleUpDownArrow;zDoubleVerticalBar;z
DownArrow;z
DownArrowBar;zDownArrowUpArrow;z
DownBreve;zDownLeftRightVector;zDownLeftTeeVector;zDownLeftVector;zDownLeftVectorBar;zDownRightTeeVector;zDownRightVector;zDownRightVectorBar;zDownTee;z
DownTeeArrow;z
Downarrow;zDscr;zDstrok;zENG;�ETHzETH;�EacutezEacute;zEcaron;�EcirczEcirc;zEcy;zEdot;zEfr;�EgravezEgrave;zElement;zEmacr;zEmptySmallSquare;zEmptyVerySmallSquare;zEogon;zEopf;zEpsilon;zEqual;zEqualTilde;zEquilibrium;zEscr;zEsim;zEta;�EumlzEuml;zExists;z
ExponentialE;zFcy;zFfr;zFilledSmallSquare;zFilledVerySmallSquare;zFopf;zForAll;zFouriertrf;zFscr;zGJcy;�GTzGT;zGamma;zGammad;zGbreve;zGcedil;zGcirc;zGcy;zGdot;zGfr;zGg;zGopf;z
GreaterEqual;zGreaterEqualLess;zGreaterFullEqual;zGreaterGreater;zGreaterLess;zGreaterSlantEqual;z
GreaterTilde;zGscr;zGt;zHARDcy;zHacek;zHat;zHcirc;zHfr;z
HilbertSpace;zHopf;zHorizontalLine;zHscr;zHstrok;z
HumpDownHump;z
HumpEqual;zIEcy;zIJlig;zIOcy;�IacutezIacute;�IcirczIcirc;zIcy;zIdot;zIfr;�IgravezIgrave;zIm;zImacr;zImaginaryI;zImplies;zInt;z	Integral;z
Intersection;zInvisibleComma;zInvisibleTimes;zIogon;zIopf;zIota;zIscr;zItilde;zIukcy;�IumlzIuml;zJcirc;zJcy;zJfr;zJopf;zJscr;zJsercy;zJukcy;zKHcy;zKJcy;zKappa;zKcedil;zKcy;zKfr;zKopf;zKscr;zLJcy;�LTzLT;zLacute;zLambda;zLang;zLaplacetrf;zLarr;zLcaron;zLcedil;zLcy;zLeftAngleBracket;z
LeftArrow;z
LeftArrowBar;zLeftArrowRightArrow;zLeftCeiling;zLeftDoubleBracket;zLeftDownTeeVector;zLeftDownVector;zLeftDownVectorBar;z
LeftFloor;zLeftRightArrow;zLeftRightVector;zLeftTee;z
LeftTeeArrow;zLeftTeeVector;z
LeftTriangle;zLeftTriangleBar;zLeftTriangleEqual;zLeftUpDownVector;zLeftUpTeeVector;z
LeftUpVector;zLeftUpVectorBar;zLeftVector;zLeftVectorBar;z
Leftarrow;zLeftrightarrow;zLessEqualGreater;zLessFullEqual;zLessGreater;z	LessLess;zLessSlantEqual;z
LessTilde;zLfr;zLl;zLleftarrow;zLmidot;zLongLeftArrow;zLongLeftRightArrow;zLongRightArrow;zLongleftarrow;zLongleftrightarrow;zLongrightarrow;zLopf;zLowerLeftArrow;zLowerRightArrow;zLscr;zLsh;zLstrok;zLt;zMap;zMcy;zMediumSpace;z
Mellintrf;zMfr;z
MinusPlus;zMopf;zMscr;zMu;zNJcy;zNacute;zNcaron;zNcedil;zNcy;zNegativeMediumSpace;zNegativeThickSpace;zNegativeThinSpace;zNegativeVeryThinSpace;zNestedGreaterGreater;zNestedLessLess;zNewLine;zNfr;zNoBreak;zNonBreakingSpace;zNopf;zNot;z
NotCongruent;z
NotCupCap;zNotDoubleVerticalBar;zNotElement;z	NotEqual;zNotEqualTilde;z
NotExists;zNotGreater;zNotGreaterEqual;zNotGreaterFullEqual;zNotGreaterGreater;zNotGreaterLess;zNotGreaterSlantEqual;zNotGreaterTilde;zNotHumpDownHump;z
NotHumpEqual;zNotLeftTriangle;zNotLeftTriangleBar;zNotLeftTriangleEqual;zNotLess;z
NotLessEqual;zNotLessGreater;zNotLessLess;zNotLessSlantEqual;z
NotLessTilde;zNotNestedGreaterGreater;zNotNestedLessLess;zNotPrecedes;zNotPrecedesEqual;zNotPrecedesSlantEqual;zNotReverseElement;zNotRightTriangle;zNotRightTriangleBar;zNotRightTriangleEqual;zNotSquareSubset;zNotSquareSubsetEqual;zNotSquareSuperset;zNotSquareSupersetEqual;z
NotSubset;zNotSubsetEqual;zNotSucceeds;zNotSucceedsEqual;zNotSucceedsSlantEqual;zNotSucceedsTilde;zNotSuperset;zNotSupersetEqual;z	NotTilde;zNotTildeEqual;zNotTildeFullEqual;zNotTildeTilde;zNotVerticalBar;zNscr;�NtildezNtilde;zNu;zOElig;�OacutezOacute;�OcirczOcirc;zOcy;zOdblac;zOfr;�OgravezOgrave;zOmacr;zOmega;zOmicron;zOopf;zOpenCurlyDoubleQuote;zOpenCurlyQuote;zOr;zOscr;�OslashzOslash;�OtildezOtilde;zOtimes;�OumlzOuml;zOverBar;z
OverBrace;zOverBracket;zOverParenthesis;z	PartialD;zPcy;zPfr;zPhi;zPi;z
PlusMinus;zPoincareplane;zPopf;zPr;z	Precedes;zPrecedesEqual;zPrecedesSlantEqual;zPrecedesTilde;zPrime;zProduct;zProportion;z
Proportional;zPscr;zPsi;�QUOTzQUOT;zQfr;zQopf;zQscr;zRBarr;�REGzREG;zRacute;zRang;zRarr;zRarrtl;zRcaron;zRcedil;zRcy;zRe;zReverseElement;zReverseEquilibrium;zReverseUpEquilibrium;zRfr;zRho;zRightAngleBracket;zRightArrow;zRightArrowBar;zRightArrowLeftArrow;z
RightCeiling;zRightDoubleBracket;zRightDownTeeVector;zRightDownVector;zRightDownVectorBar;zRightFloor;z	RightTee;zRightTeeArrow;zRightTeeVector;zRightTriangle;zRightTriangleBar;zRightTriangleEqual;zRightUpDownVector;zRightUpTeeVector;zRightUpVector;zRightUpVectorBar;zRightVector;zRightVectorBar;zRightarrow;zRopf;z
RoundImplies;zRrightarrow;zRscr;zRsh;zRuleDelayed;zSHCHcy;zSHcy;zSOFTcy;zSacute;zSc;zScaron;zScedil;zScirc;zScy;zSfr;zShortDownArrow;zShortLeftArrow;zShortRightArrow;z
ShortUpArrow;zSigma;zSmallCircle;zSopf;zSqrt;zSquare;zSquareIntersection;z
SquareSubset;zSquareSubsetEqual;zSquareSuperset;zSquareSupersetEqual;zSquareUnion;zSscr;zStar;zSub;zSubset;zSubsetEqual;z	Succeeds;zSucceedsEqual;zSucceedsSlantEqual;zSucceedsTilde;z	SuchThat;zSum;zSup;z	Superset;zSupersetEqual;zSupset;�THORNzTHORN;zTRADE;zTSHcy;zTScy;zTab;zTau;zTcaron;zTcedil;zTcy;zTfr;z
Therefore;zTheta;zThickSpace;z
ThinSpace;zTilde;zTildeEqual;zTildeFullEqual;zTildeTilde;zTopf;z
TripleDot;zTscr;zTstrok;�UacutezUacute;zUarr;z	Uarrocir;zUbrcy;zUbreve;�UcirczUcirc;zUcy;zUdblac;zUfr;�UgravezUgrave;zUmacr;z	UnderBar;zUnderBrace;z
UnderBracket;zUnderParenthesis;zUnion;z
UnionPlus;zUogon;zUopf;zUpArrow;zUpArrowBar;zUpArrowDownArrow;zUpDownArrow;zUpEquilibrium;zUpTee;zUpTeeArrow;zUparrow;zUpdownarrow;zUpperLeftArrow;zUpperRightArrow;zUpsi;zUpsilon;zUring;zUscr;zUtilde;�UumlzUuml;zVDash;zVbar;zVcy;zVdash;zVdashl;zVee;zVerbar;zVert;zVerticalBar;z
VerticalLine;zVerticalSeparator;zVerticalTilde;zVeryThinSpace;zVfr;zVopf;zVscr;zVvdash;zWcirc;zWedge;zWfr;zWopf;zWscr;zXfr;zXi;zXopf;zXscr;zYAcy;zYIcy;zYUcy;�YacutezYacute;zYcirc;zYcy;zYfr;zYopf;zYscr;zYuml;zZHcy;zZacute;zZcaron;zZcy;zZdot;zZeroWidthSpace;zZeta;zZfr;zZopf;zZscr;�aacutezaacute;zabreve;zac;zacE;zacd;�acirczacirc;�acutezacute;zacy;�aeligzaelig;zaf;zafr;�agravezagrave;zalefsym;zaleph;zalpha;zamacr;zamalg;�ampr$zand;zandand;zandd;z	andslope;zandv;zang;zange;zangle;zangmsd;z	angmsdaa;z	angmsdab;z	angmsdac;z	angmsdad;z	angmsdae;z	angmsdaf;z	angmsdag;z	angmsdah;zangrt;zangrtvb;z	angrtvbd;zangsph;zangst;zangzarr;zaogon;zaopf;zap;zapE;zapacir;zape;zapid;r%zapprox;z	approxeq;�aringzaring;zascr;zast;zasymp;zasympeq;�atildezatilde;�aumlzauml;z	awconint;zawint;zbNot;z	backcong;zbackepsilon;z
backprime;zbacksim;z
backsimeq;zbarvee;zbarwed;z	barwedge;zbbrk;z	bbrktbrk;zbcong;zbcy;zbdquo;zbecaus;zbecause;zbemptyv;zbepsi;zbernou;zbeta;zbeth;zbetween;zbfr;zbigcap;zbigcirc;zbigcup;zbigodot;z	bigoplus;z
bigotimes;z	bigsqcup;zbigstar;zbigtriangledown;zbigtriangleup;z	biguplus;zbigvee;z	bigwedge;zbkarow;z
blacklozenge;zblacksquare;zblacktriangle;zblacktriangledown;zblacktriangleleft;zblacktriangleright;zblank;zblk12;zblk14;zblk34;zblock;zbne;zbnequiv;zbnot;zbopf;zbot;zbottom;zbowtie;zboxDL;zboxDR;zboxDl;zboxDr;zboxH;zboxHD;zboxHU;zboxHd;zboxHu;zboxUL;zboxUR;zboxUl;zboxUr;zboxV;zboxVH;zboxVL;zboxVR;zboxVh;zboxVl;zboxVr;zboxbox;zboxdL;zboxdR;zboxdl;zboxdr;zboxh;zboxhD;zboxhU;zboxhd;zboxhu;z	boxminus;zboxplus;z	boxtimes;zboxuL;zboxuR;zboxul;zboxur;zboxv;zboxvH;zboxvL;zboxvR;zboxvh;zboxvl;zboxvr;zbprime;zbreve;�brvbarzbrvbar;zbscr;zbsemi;zbsim;zbsime;zbsol;zbsolb;z	bsolhsub;zbull;zbullet;zbump;zbumpE;zbumpe;zbumpeq;zcacute;zcap;zcapand;z	capbrcup;zcapcap;zcapcup;zcapdot;zcaps;zcaret;zcaron;zccaps;zccaron;�ccedilzccedil;zccirc;zccups;zccupssm;zcdot;�cedilzcedil;zcemptyv;�centzcent;z
centerdot;zcfr;zchcy;zcheck;z
checkmark;zchi;zcir;zcirE;zcirc;zcirceq;zcirclearrowleft;zcirclearrowright;z	circledR;z	circledS;zcircledast;zcircledcirc;zcircleddash;zcire;z	cirfnint;zcirmid;zcirscir;zclubs;z	clubsuit;zcolon;zcolone;zcoloneq;zcomma;zcommat;zcomp;zcompfn;zcomplement;z
complexes;zcong;zcongdot;zconint;zcopf;zcoprod;�copyzcopy;zcopysr;zcrarr;zcross;zcscr;zcsub;zcsube;zcsup;zcsupe;zctdot;zcudarrl;zcudarrr;zcuepr;zcuesc;zcularr;zcularrp;zcup;z	cupbrcap;zcupcap;zcupcup;zcupdot;zcupor;zcups;zcurarr;zcurarrm;zcurlyeqprec;zcurlyeqsucc;z	curlyvee;zcurlywedge;�currenzcurren;zcurvearrowleft;zcurvearrowright;zcuvee;zcuwed;z	cwconint;zcwint;zcylcty;zdArr;zdHar;zdagger;zdaleth;zdarr;zdash;zdashv;zdbkarow;zdblac;zdcaron;zdcy;zdd;zddagger;zddarr;zddotseq;�degzdeg;zdelta;zdemptyv;zdfisht;zdfr;zdharl;zdharr;zdiam;zdiamond;zdiamondsuit;zdiams;zdie;zdigamma;zdisin;zdiv;�dividezdivide;zdivideontimes;zdivonx;zdjcy;zdlcorn;zdlcrop;zdollar;zdopf;zdot;zdoteq;z	doteqdot;z	dotminus;zdotplus;z
dotsquare;zdoublebarwedge;z
downarrow;zdowndownarrows;zdownharpoonleft;zdownharpoonright;z	drbkarow;zdrcorn;zdrcrop;zdscr;zdscy;zdsol;zdstrok;zdtdot;zdtri;zdtrif;zduarr;zduhar;zdwangle;zdzcy;z	dzigrarr;zeDDot;zeDot;�eacutezeacute;zeaster;zecaron;zecir;�ecirczecirc;zecolon;zecy;zedot;zee;zefDot;zefr;zeg;�egravezegrave;zegs;zegsdot;zel;z	elinters;zell;zels;zelsdot;zemacr;zempty;z	emptyset;zemptyv;zemsp13;zemsp14;zemsp;zeng;zensp;zeogon;zeopf;zepar;zeparsl;zeplus;zepsi;zepsilon;zepsiv;zeqcirc;zeqcolon;zeqsim;zeqslantgtr;zeqslantless;zequals;zequest;zequiv;zequivDD;z	eqvparsl;zerDot;zerarr;zescr;zesdot;zesim;zeta;�ethzeth;�eumlzeuml;zeuro;zexcl;zexist;zexpectation;z
exponentiale;zfallingdotseq;zfcy;zfemale;zffilig;zfflig;zffllig;zffr;zfilig;zfjlig;zflat;zfllig;zfltns;zfnof;zfopf;zforall;zfork;zforkv;z	fpartint;�frac12zfrac12;zfrac13;�frac14zfrac14;zfrac15;zfrac16;zfrac18;zfrac23;zfrac25;�frac34zfrac34;zfrac35;zfrac38;zfrac45;zfrac56;zfrac58;zfrac78;zfrasl;zfrown;zfscr;zgE;zgEl;zgacute;zgamma;zgammad;zgap;zgbreve;zgcirc;zgcy;zgdot;zge;zgel;zgeq;zgeqq;z	geqslant;zges;zgescc;zgesdot;zgesdoto;z	gesdotol;zgesl;zgesles;zgfr;zgg;zggg;zgimel;zgjcy;zgl;zglE;zgla;zglj;zgnE;zgnap;z	gnapprox;zgne;zgneq;zgneqq;zgnsim;zgopf;zgrave;zgscr;zgsim;zgsime;zgsiml;�gtr#zgtcc;zgtcir;zgtdot;zgtlPar;zgtquest;z
gtrapprox;zgtrarr;zgtrdot;z
gtreqless;zgtreqqless;zgtrless;zgtrsim;z
gvertneqq;zgvnE;zhArr;zhairsp;zhalf;zhamilt;zhardcy;zharr;zharrcir;zharrw;zhbar;zhcirc;zhearts;z
heartsuit;zhellip;zhercon;zhfr;z	hksearow;z	hkswarow;zhoarr;zhomtht;zhookleftarrow;zhookrightarrow;zhopf;zhorbar;zhscr;zhslash;zhstrok;zhybull;zhyphen;�iacuteziacute;zic;�icirczicirc;zicy;ziecy;�iexclziexcl;ziff;zifr;�igravezigrave;zii;ziiiint;ziiint;ziinfin;ziiota;zijlig;zimacr;zimage;z	imagline;z	imagpart;zimath;zimof;zimped;zin;zincare;zinfin;z	infintie;zinodot;zint;zintcal;z	integers;z	intercal;z	intlarhk;zintprod;ziocy;ziogon;ziopf;ziota;ziprod;�iquestziquest;ziscr;zisin;zisinE;zisindot;zisins;zisinsv;zisinv;zit;zitilde;ziukcy;�iumlziuml;zjcirc;zjcy;zjfr;zjmath;zjopf;zjscr;zjsercy;zjukcy;zkappa;zkappav;zkcedil;zkcy;zkfr;zkgreen;zkhcy;zkjcy;zkopf;zkscr;zlAarr;zlArr;zlAtail;zlBarr;zlE;zlEg;zlHar;zlacute;z	laemptyv;zlagran;zlambda;zlang;zlangd;zlangle;zlap;�laquozlaquo;zlarr;zlarrb;zlarrbfs;zlarrfs;zlarrhk;zlarrlp;zlarrpl;zlarrsim;zlarrtl;zlat;zlatail;zlate;zlates;zlbarr;zlbbrk;zlbrace;zlbrack;zlbrke;zlbrksld;zlbrkslu;zlcaron;zlcedil;zlceil;zlcub;zlcy;zldca;zldquo;zldquor;zldrdhar;z	ldrushar;zldsh;zle;z
leftarrow;zleftarrowtail;zleftharpoondown;zleftharpoonup;zleftleftarrows;zleftrightarrow;zleftrightarrows;zleftrightharpoons;zleftrightsquigarrow;zleftthreetimes;zleg;zleq;zleqq;z	leqslant;zles;zlescc;zlesdot;zlesdoto;z	lesdotor;zlesg;zlesges;zlessapprox;zlessdot;z
lesseqgtr;zlesseqqgtr;zlessgtr;zlesssim;zlfisht;zlfloor;zlfr;zlg;zlgE;zlhard;zlharu;zlharul;zlhblk;zljcy;zll;zllarr;z	llcorner;zllhard;zlltri;zlmidot;zlmoust;zlmoustache;zlnE;zlnap;z	lnapprox;zlne;zlneq;zlneqq;zlnsim;zloang;zloarr;zlobrk;zlongleftarrow;zlongleftrightarrow;zlongmapsto;zlongrightarrow;zlooparrowleft;zlooparrowright;zlopar;zlopf;zloplus;zlotimes;zlowast;zlowbar;zloz;zlozenge;zlozf;zlpar;zlparlt;zlrarr;z	lrcorner;zlrhar;zlrhard;zlrm;zlrtri;zlsaquo;zlscr;zlsh;zlsim;zlsime;zlsimg;zlsqb;zlsquo;zlsquor;zlstrok;�ltr"zltcc;zltcir;zltdot;zlthree;zltimes;zltlarr;zltquest;zltrPar;zltri;zltrie;zltrif;z	lurdshar;zluruhar;z
lvertneqq;zlvnE;zmDDot;�macrzmacr;zmale;zmalt;zmaltese;zmap;zmapsto;zmapstodown;zmapstoleft;z	mapstoup;zmarker;zmcomma;zmcy;zmdash;zmeasuredangle;zmfr;zmho;�microzmicro;zmid;zmidast;zmidcir;�middotzmiddot;zminus;zminusb;zminusd;zminusdu;zmlcp;zmldr;zmnplus;zmodels;zmopf;zmp;zmscr;zmstpos;zmu;z	multimap;zmumap;znGg;znGt;znGtv;znLeftarrow;znLeftrightarrow;znLl;znLt;znLtv;znRightarrow;znVDash;znVdash;znabla;znacute;znang;znap;znapE;znapid;znapos;znapprox;znatur;znatural;z	naturals;�nbspznbsp;znbump;znbumpe;zncap;zncaron;zncedil;zncong;z	ncongdot;zncup;zncy;zndash;zne;zneArr;znearhk;znearr;znearrow;znedot;znequiv;znesear;znesim;znexist;znexists;znfr;zngE;znge;zngeq;zngeqq;z
ngeqslant;znges;zngsim;zngt;zngtr;znhArr;znharr;znhpar;zni;znis;znisd;zniv;znjcy;znlArr;znlE;znlarr;znldr;znle;znleftarrow;znleftrightarrow;znleq;znleqq;z
nleqslant;znles;znless;znlsim;znlt;znltri;znltrie;znmid;znopf;�notznot;znotin;znotinE;z	notindot;znotinva;znotinvb;znotinvc;znotni;znotniva;znotnivb;znotnivc;znpar;z
nparallel;znparsl;znpart;znpolint;znpr;znprcue;znpre;znprec;znpreceq;znrArr;znrarr;znrarrc;znrarrw;znrightarrow;znrtri;znrtrie;znsc;znsccue;znsce;znscr;z
nshortmid;znshortparallel;znsim;znsime;znsimeq;znsmid;znspar;znsqsube;znsqsupe;znsub;znsubE;znsube;znsubset;z
nsubseteq;znsubseteqq;znsucc;znsucceq;znsup;znsupE;znsupe;znsupset;z
nsupseteq;znsupseteqq;zntgl;�ntildezntilde;zntlg;zntriangleleft;zntrianglelefteq;zntriangleright;zntrianglerighteq;znu;znum;znumero;znumsp;znvDash;znvHarr;znvap;znvdash;znvge;znvgt;znvinfin;znvlArr;znvle;znvlt;znvltrie;znvrArr;znvrtrie;znvsim;znwArr;znwarhk;znwarr;znwarrow;znwnear;zoS;�oacutezoacute;zoast;zocir;�ocirczocirc;zocy;zodash;zodblac;zodiv;zodot;zodsold;zoelig;zofcir;zofr;zogon;�ogravezograve;zogt;zohbar;zohm;zoint;zolarr;zolcir;zolcross;zoline;zolt;zomacr;zomega;zomicron;zomid;zominus;zoopf;zopar;zoperp;zoplus;zor;zorarr;zord;zorder;zorderof;�ordfzordf;�ordmzordm;zorigof;zoror;zorslope;zorv;zoscr;�oslashzoslash;zosol;�otildezotilde;zotimes;z	otimesas;�oumlzouml;zovbar;zpar;�parazpara;z	parallel;zparsim;zparsl;zpart;zpcy;zpercnt;zperiod;zpermil;zperp;zpertenk;zpfr;zphi;zphiv;zphmmat;zphone;zpi;z
pitchfork;zpiv;zplanck;zplanckh;zplankv;zplus;z	plusacir;zplusb;zpluscir;zplusdo;zplusdu;zpluse;�plusmnzplusmn;zplussim;zplustwo;zpm;z	pointint;zpopf;�poundzpound;zpr;zprE;zprap;zprcue;zpre;zprec;zprecapprox;zpreccurlyeq;zpreceq;zprecnapprox;z	precneqq;z	precnsim;zprecsim;zprime;zprimes;zprnE;zprnap;zprnsim;zprod;z	profalar;z	profline;z	profsurf;zprop;zpropto;zprsim;zprurel;zpscr;zpsi;zpuncsp;zqfr;zqint;zqopf;zqprime;zqscr;zquaternions;zquatint;zquest;zquesteq;�quotr&zrAarr;zrArr;zrAtail;zrBarr;zrHar;zrace;zracute;zradic;z	raemptyv;zrang;zrangd;zrange;zrangle;�raquozraquo;zrarr;zrarrap;zrarrb;zrarrbfs;zrarrc;zrarrfs;zrarrhk;zrarrlp;zrarrpl;zrarrsim;zrarrtl;zrarrw;zratail;zratio;z
rationals;zrbarr;zrbbrk;zrbrace;zrbrack;zrbrke;zrbrksld;zrbrkslu;zrcaron;zrcedil;zrceil;zrcub;zrcy;zrdca;zrdldhar;zrdquo;zrdquor;zrdsh;zreal;zrealine;z	realpart;zreals;zrect;�regzreg;zrfisht;zrfloor;zrfr;zrhard;zrharu;zrharul;zrho;zrhov;zrightarrow;zrightarrowtail;zrightharpoondown;zrightharpoonup;zrightleftarrows;zrightleftharpoons;zrightrightarrows;zrightsquigarrow;zrightthreetimes;zring;z
risingdotseq;zrlarr;zrlhar;zrlm;zrmoust;zrmoustache;zrnmid;zroang;zroarr;zrobrk;zropar;zropf;zroplus;zrotimes;zrpar;zrpargt;z	rppolint;zrrarr;zrsaquo;zrscr;zrsh;zrsqb;zrsquo;zrsquor;zrthree;zrtimes;zrtri;zrtrie;zrtrif;z	rtriltri;zruluhar;zrx;zsacute;zsbquo;zsc;zscE;zscap;zscaron;zsccue;zsce;zscedil;zscirc;zscnE;zscnap;zscnsim;z	scpolint;zscsim;zscy;zsdot;zsdotb;zsdote;zseArr;zsearhk;zsearr;zsearrow;�sectzsect;zsemi;zseswar;z	setminus;zsetmn;zsext;zsfr;zsfrown;zsharp;zshchcy;zshcy;z	shortmid;zshortparallel;�shyzshy;zsigma;zsigmaf;zsigmav;zsim;zsimdot;zsime;zsimeq;zsimg;zsimgE;zsiml;zsimlE;zsimne;zsimplus;zsimrarr;zslarr;zsmallsetminus;zsmashp;z	smeparsl;zsmid;zsmile;zsmt;zsmte;zsmtes;zsoftcy;zsol;zsolb;zsolbar;zsopf;zspades;z
spadesuit;zspar;zsqcap;zsqcaps;zsqcup;zsqcups;zsqsub;zsqsube;z	sqsubset;zsqsubseteq;zsqsup;zsqsupe;z	sqsupset;zsqsupseteq;zsqu;zsquare;zsquarf;zsquf;zsrarr;zsscr;zssetmn;zssmile;zsstarf;zstar;zstarf;zstraightepsilon;zstraightphi;zstrns;zsub;zsubE;zsubdot;zsube;zsubedot;zsubmult;zsubnE;zsubne;zsubplus;zsubrarr;zsubset;z	subseteq;z
subseteqq;z
subsetneq;zsubsetneqq;zsubsim;zsubsub;zsubsup;zsucc;zsuccapprox;zsucccurlyeq;zsucceq;zsuccnapprox;z	succneqq;z	succnsim;zsuccsim;zsum;zsung;�sup1zsup1;�sup2zsup2;�sup3zsup3;zsup;zsupE;zsupdot;zsupdsub;zsupe;zsupedot;zsuphsol;zsuphsub;zsuplarr;zsupmult;zsupnE;zsupne;zsupplus;zsupset;z	supseteq;z
supseteqq;z
supsetneq;zsupsetneqq;zsupsim;zsupsub;zsupsup;zswArr;zswarhk;zswarr;zswarrow;zswnwar;�szligzszlig;ztarget;ztau;ztbrk;ztcaron;ztcedil;ztcy;ztdot;ztelrec;ztfr;zthere4;z
therefore;ztheta;z	thetasym;zthetav;zthickapprox;z	thicksim;zthinsp;zthkap;zthksim;�thornzthorn;ztilde;�timesztimes;ztimesb;z	timesbar;ztimesd;ztint;ztoea;ztop;ztopbot;ztopcir;ztopf;ztopfork;ztosa;ztprime;ztrade;z	triangle;z
triangledown;z
triangleleft;ztrianglelefteq;z
triangleq;ztriangleright;ztrianglerighteq;ztridot;ztrie;z	triminus;ztriplus;ztrisb;ztritime;z	trpezium;ztscr;ztscy;ztshcy;ztstrok;ztwixt;ztwoheadleftarrow;ztwoheadrightarrow;zuArr;zuHar;�uacutezuacute;zuarr;zubrcy;zubreve;�ucirczucirc;zucy;zudarr;zudblac;zudhar;zufisht;zufr;�ugravezugrave;zuharl;zuharr;zuhblk;zulcorn;z	ulcorner;zulcrop;zultri;zumacr;�umlzuml;zuogon;zuopf;zuparrow;zupdownarrow;zupharpoonleft;zupharpoonright;zuplus;zupsi;zupsih;zupsilon;zupuparrows;zurcorn;z	urcorner;zurcrop;zuring;zurtri;zuscr;zutdot;zutilde;zutri;zutrif;zuuarr;�uumlzuuml;zuwangle;zvArr;zvBar;zvBarv;zvDash;zvangrt;zvarepsilon;z	varkappa;zvarnothing;zvarphi;zvarpi;z
varpropto;zvarr;zvarrho;z	varsigma;z
varsubsetneq;zvarsubsetneqq;z
varsupsetneq;zvarsupsetneqq;z	vartheta;zvartriangleleft;zvartriangleright;zvcy;zvdash;zvee;zveebar;zveeeq;zvellip;zverbar;zvert;zvfr;zvltri;zvnsub;zvnsup;zvopf;zvprop;zvrtri;zvscr;zvsubnE;zvsubne;zvsupnE;zvsupne;zvzigzag;zwcirc;zwedbar;zwedge;zwedgeq;zweierp;zwfr;zwopf;zwp;zwr;zwreath;zwscr;zxcap;zxcirc;zxcup;zxdtri;zxfr;zxhArr;zxharr;zxi;zxlArr;zxlarr;zxmap;zxnis;zxodot;zxopf;zxoplus;zxotime;zxrArr;zxrarr;zxscr;zxsqcup;zxuplus;zxutri;zxvee;zxwedge;�yacutezyacute;zyacy;zycirc;zycy;�yenzyen;zyfr;zyicy;zyopf;zyscr;zyucy;�yumlzyuml;zzacute;zzcaron;zzcy;zzdot;zzeetrf;zzeta;zzfr;zzhcy;zzigrarr;zzopf;zzscr;zzwj;zzwnj;u�r������)"r�
���������������������������������������)�Doctype�
Characters�SpaceCharacters�StartTag�EndTag�EmptyTag�Comment�
ParseErrorr@rArBcCsi|]\}}||�qSr�r�)r��k�vr�r�r�r�xr��mathc@seZdZdZdS)�DataLossWarningzBRaised when the current tree is unable to represent the input dataN)�__name__�
__module__�__qualname__�__doc__r�r�r�r�rH|srHc@seZdZdS)�_ReparseExceptionN)rIrJrKr�r�r�r�rM�srM)/�
__future__rrr�string�EOF�E�
namespaces�	frozenset�scopingElements�formattingElements�specialElements�htmlIntegrationPointElements�"mathmlTextIntegrationPointElements�adjustSVGAttributes�adjustMathMLAttributes�adjustForeignAttributes�items�unadjustForeignAttributes�spaceCharacters�tableInsertModeElements�ascii_lowercase�asciiLowercase�ascii_uppercase�asciiUppercase�
ascii_letters�asciiLetters�digits�	hexdigits�	hexDigits�asciiUpper2Lower�headingElements�voidElements�
cdataElements�rcdataElements�booleanAttributes�entitiesWindows1252�xmlEntities�entities�replacementCharacters�
tokenTypes�
tagTokenTypes�prefixes�UserWarningrH�	ExceptionrMr�r�r�r��<module>s���$�	
















�













�













































































�S



�




�	�A�
�




	





�#������������������
L�&��
site-packages/pip/_vendor/html5lib/__pycache__/serializer.cpython-39.pyc000064400000025144150564745010022257 0ustar00a

��Re�=�@szddlmZmZmZddlmZddlZddlmZm	Z	ddl
mZmZm
Z
ddl
mZmZmZddlmZmZdd	lmZd
�e
�dZe�ded
�Ze�ded�ZiZed�dkZee���D]n\Z Z!er�ee!�dks�es�ee!�dkr�q�e!dkr�ee!�dk�re�"e!�Z!ne#e!�Z!e!ev�s0e �$�r�e ee!<q�dd�Z%ede%�ddd�Z&Gdd�de'�Z(Gdd�de)�Z*dS)�)�absolute_import�division�unicode_literals)�	text_typeN)�register_error�xmlcharrefreplace_errors�)�voidElements�booleanAttributes�spaceCharacters)�rcdataElements�entities�xmlEntities)�treewalkers�_utils)�escape�z"'=<>`�[�]u_	

 /`  ᠎᠏           

   ]u􏿿��&c
	Cst|ttf��rg}g}d}t|j|j|j��D]n\}}|rDd}q2||j}t�|j|t	|j|dg���r�t�
|j||d��}d}nt|�}|�|�q2|D]V}t
�|�}	|	r�|�d�|�|	�|	�d�s�|�d�q�|�dt|�dd��q�d�|�|jfSt|�SdS)NFrTr�;z&#x%s;r)�
isinstance�UnicodeEncodeError�UnicodeTranslateError�	enumerate�object�start�endr�isSurrogatePair�min�surrogatePairToCodepoint�ord�append�_encode_entity_map�get�endswith�hex�joinr)
�exc�res�
codepoints�skip�i�c�index�	codepoint�cp�e�r3��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/serializer.py�htmlentityreplace_errors*s0
"



r5�htmlentityreplace�etreecKs(t�|�}tfi|��}|�||�|�S)a�Serializes the input token stream using the specified treewalker

    :arg input: the token stream to serialize

    :arg tree: the treewalker to use

    :arg encoding: the encoding to use

    :arg serializer_opts: any options to pass to the
        :py:class:`html5lib.serializer.HTMLSerializer` that gets created

    :returns: the tree serialized as a string

    Example:

    >>> from html5lib.html5parser import parse
    >>> from html5lib.serializer import serialize
    >>> token_stream = parse('<html><body><p>Hi!</p></body></html>')
    >>> serialize(token_stream, omit_optional_tags=False)
    '<html><head></head><body><p>Hi!</p></body></html>'

    )r�
getTreeWalker�HTMLSerializer�render)�input�tree�encoding�serializer_opts�walker�sr3r3r4�	serializeKs
rAc@s~eZdZdZdZdZdZdZdZdZ	dZ
dZdZdZ
dZdZdZdZdd�Zdd	�Zd
d�Zdd
d�Zddd�Zddd�ZdS)r9�legacy�"TF)�quote_attr_values�
quote_char�use_best_quote_char�omit_optional_tags�minimize_boolean_attributes�use_trailing_solidus�space_before_trailing_solidus�escape_lt_in_attrs�
escape_rcdata�resolve_entities�alphabetical_attributes�inject_meta_charset�strip_whitespace�sanitizec
Ksvt|�t|j�}t|�dkr2tdtt|����d|vr@d|_|jD]}t|||�|t	||���qFg|_
d|_dS)aB
Initialize HTMLSerializer

        :arg inject_meta_charset: Whether or not to inject the meta charset.

            Defaults to ``True``.

        :arg quote_attr_values: Whether to quote attribute values that don't
            require quoting per legacy browser behavior (``"legacy"``), when
            required by the standard (``"spec"``), or always (``"always"``).

            Defaults to ``"legacy"``.

        :arg quote_char: Use given quote character for attribute quoting.

            Defaults to ``"`` which will use double quotes unless attribute
            value contains a double quote, in which case single quotes are
            used.

        :arg escape_lt_in_attrs: Whether or not to escape ``<`` in attribute
            values.

            Defaults to ``False``.

        :arg escape_rcdata: Whether to escape characters that need to be
            escaped within normal elements within rcdata elements such as
            style.

            Defaults to ``False``.

        :arg resolve_entities: Whether to resolve named character entities that
            appear in the source tree. The XML predefined entities &lt; &gt;
            &amp; &quot; &apos; are unaffected by this setting.

            Defaults to ``True``.

        :arg strip_whitespace: Whether to remove semantically meaningless
            whitespace. (This compresses all whitespace to a single space
            except within ``pre``.)

            Defaults to ``False``.

        :arg minimize_boolean_attributes: Shortens boolean attributes to give
            just the attribute value, for example::

              <input disabled="disabled">

            becomes::

              <input disabled>

            Defaults to ``True``.

        :arg use_trailing_solidus: Includes a close-tag slash at the end of the
            start tag of void elements (empty elements whose end tag is
            forbidden). E.g. ``<hr/>``.

            Defaults to ``False``.

        :arg space_before_trailing_solidus: Places a space immediately before
            the closing slash in a tag using a trailing solidus. E.g.
            ``<hr />``. Requires ``use_trailing_solidus=True``.

            Defaults to ``True``.

        :arg sanitize: Strip all unsafe or unknown constructs from output.
            See :py:class:`html5lib.filters.sanitizer.Filter`.

            Defaults to ``False``.

        :arg omit_optional_tags: Omit start/end tags that are optional.

            Defaults to ``True``.

        :arg alphabetical_attributes: Reorder attributes to be in alphabetical order.

            Defaults to ``False``.

        rz2__init__() got an unexpected keyword argument '%s'rEFN)�	frozenset�options�len�	TypeError�next�iterrF�setattrr%�getattr�errors�strict)�self�kwargs�unexpected_args�attrr3r3r4�__init__�sO
zHTMLSerializer.__init__cCs*t|t�sJ�|jr"|�|jd�S|SdS)Nr6�rrr=�encode�r\�stringr3r3r4rb�szHTMLSerializer.encodecCs*t|t�sJ�|jr"|�|jd�S|SdS)Nr[rarcr3r3r4�encodeStrict�szHTMLSerializer.encodeStrictNccs�||_d}g|_|r0|jr0ddlm}|||�}|jrJddlm}||�}|jrdddlm}||�}|j	r~ddl
m}||�}|jr�ddlm}||�}|D�]>}|d}|dk�r\d|d}|dr�|d	|d7}n|d
r�|d7}|d
�rF|d
�
d�d
k�r,|d
�
d�d
k�r&|�d�d}nd}|d||d
|f7}|d7}|�|�Vq�|dv�r�|dk�sv|�r�|�r�|d�
d�d
k�r�|�d�|�|d�Vn|�t|d��Vq�|dv�r�|d}	|�d|	�V|	tv�r�|j�s�d}n|�r|�d�|d��D�]�\\}
}}|}
|}|�d�V|�|
�V|j�rv|
t�|	t��v�r|
t�dt��v�r|�d�V|jdk�s�t|�d
k�r�d}n@|jdk�r�t�|�du}n$|jd k�r�t�|�du}ntd!��|�d"d#�}|j �r|�d$d%�}|�r�|j!}|j"�rHd|v�r0d|v�r0d}nd|v�rHd|v�rHd}|dk�r`|�dd&�}n|�dd'�}|�|�V|�|�V|�|�Vn|�|�V�q|	t#v�r�|j$�r�|j%�r�|�d(�Vn|�d)�V|�d�Vq�|d*k�r(|d}	|	tv�rd}n|�r|�d�|�d+|	�Vq�|d,k�rj|d}|�
d-�d
k�rT|�d.�|�d/|d�Vq�|d0k�r�|d}	|	d1}|t&v�r�|�d2|	�|j'�r�|t(v�r�t&|}nd3|	}|�|�Vq�|�|d�q�dS)4NFr)�Filter�type�Doctypez<!DOCTYPE %s�name�publicIdz PUBLIC "%s"�systemIdz SYSTEMrCr�'zBSystem identifier contains both single and double quote charactersz %s%s%s�>)�
Characters�SpaceCharactersro�dataz</zUnexpected </ in CDATA)�StartTag�EmptyTagz<%sTz+Unexpected child element of a CDATA element� r�=�always�specrBz?quote_attr_values must be one of: 'always', 'spec', or 'legacy'rz&amp;�<z&lt;z&#39;z&quot;z /�/�EndTagz</%s>�Commentz--zComment contains --z	<!--%s-->�EntityrzEntity %s not recognizedz&%s;))r=rZrO�filters.inject_meta_charsetrfrN�filters.alphabeticalattributesrP�filters.whitespacerQ�filters.sanitizerrG�filters.optionaltags�find�serializeErrorrerbrrrL�itemsrHr
r%�tuplerDrT�_quoteAttributeSpec�search�_quoteAttributeLegacy�
ValueError�replacerKrErFr	rIrJr
rMr)r\�
treewalkerr=�in_cdatarf�tokenrg�doctyperEri�_�	attr_name�
attr_value�k�v�
quote_attrrp�keyr3r3r4rA�s�









��








zHTMLSerializer.serializecCs2|rd�t|�||���Sd�t|�|���SdS)anSerializes the stream from the treewalker into a string

        :arg treewalker: the treewalker to serialize

        :arg encoding: the string encoding to use

        :returns: the serialized tree

        Example:

        >>> from html5lib import parse, getTreeWalker
        >>> from html5lib.serializer import HTMLSerializer
        >>> token_stream = parse('<html><body>Hi!</body></html>')
        >>> walker = getTreeWalker('etree')
        >>> serializer = HTMLSerializer(omit_optional_tags=False)
        >>> serializer.render(walker(token_stream))
        '<html><head></head><body>Hi!</body></html>'

        �rN)r(�listrA)r\r�r=r3r3r4r:wszHTMLSerializer.render�XXX ERROR MESSAGE NEEDEDcCs|j�|�|jrt�dS)N)rZr#r[�SerializeError)r\rpr3r3r4r��szHTMLSerializer.serializeError)N)N)r�)�__name__�
__module__�__qualname__rDrErFrGrHrIrJrKrLrMrNrOrPrQrSr`rbrerAr:r�r3r3r3r4r9hs,Y


r9c@seZdZdZdS)r�zError in serialized treeN)r�r�r��__doc__r3r3r3r4r��sr�)r7N)+�
__future__rrrZpip._vendor.sixr�re�codecsrr�	constantsr	r
rrr
rrrr�xml.sax.saxutilsrr(�_quoteAttributeSpecChars�compiler�r�r$rT�_is_ucs4r�r�r�r�r!r"�islowerr5rArr9�	Exceptionr�r3r3r3r4�<module>sD
�
�
�


1site-packages/pip/_vendor/html5lib/__pycache__/__init__.cpython-39.pyc000064400000002475150564745010021647 0ustar00a

��Re��@s`dZddlmZmZmZddlmZmZmZddl	m
Z
ddlmZddl
mZgd�Zd	Zd
S)a�
HTML parsing library based on the `WHATWG HTML specification
<https://whatwg.org/html>`_. The parser is designed to be compatible with
existing HTML found in the wild and implements well-defined error recovery that
is largely compatible with modern desktop web browsers.

Example usage::

    from pip._vendor import html5lib
    with open("my_document.html", "rb") as f:
        tree = html5lib.parse(f)

For convenience, this module re-exports the following names:

* :func:`~.html5parser.parse`
* :func:`~.html5parser.parseFragment`
* :class:`~.html5parser.HTMLParser`
* :func:`~.treebuilders.getTreeBuilder`
* :func:`~.treewalkers.getTreeWalker`
* :func:`~.serializer.serialize`
�)�absolute_import�division�unicode_literals�)�
HTMLParser�parse�
parseFragment)�getTreeBuilder)�
getTreeWalker)�	serialize)rrrr	r
rz1.1N)�__doc__�
__future__rrr�html5parserrrr�treebuildersr	�treewalkersr
�
serializerr�__all__�__version__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/__init__.py�<module>ssite-packages/pip/_vendor/html5lib/__pycache__/_utils.cpython-39.pyc000064400000011351150564745010021400 0ustar00a

��ReC�@sHddlmZmZmZddlmZzddlmZWneyNddl	mZYn0ddl
mZmZerxddl
mmZn<zddlmmZWn$ey�ddl
mmZYn0gd�Zz,ed�Zeee�s�ed�Zeee�s�J�Wney�d	ZYn0d
ZGdd�de�ZGd
d�de�Zdd�Zdd�Zdd�Zdd�ZdS)�)�absolute_import�division�unicode_literals)�
ModuleType)�Mapping)�	text_type�PY3N)�
default_etree�MethodDispatcher�isSurrogatePair�surrogatePairToCodepoint�moduleFactoryFactory�supports_lone_surrogatesz"\uD800"z	u"\uD800"FTc@s,eZdZdZd
dd�Zdd�Zddd	�ZdS)r
apDict with 2 special properties:

    On initiation, keys that are lists, sets or tuples are converted to
    multiple keys so accessing any one of the items in the original
    list-like object returns the matching value

    md = MethodDispatcher({("foo", "bar"):"baz"})
    md["foo"] == "baz"

    A default value which can be set through the default attribute.
    �cCsvg}|D]B\}}t|ttttf�r<|D]}|�||f�q&q|�||f�qt�||�t|�t|�kslJ�d|_	dS�N)
�
isinstance�list�tuple�	frozenset�set�append�dict�__init__�len�default)�self�items�_dictEntries�name�value�itemrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_utils.pyr9szMethodDispatcher.__init__cCst�|||j�Sr)r�getr�r�keyrrr!�__getitem__EszMethodDispatcher.__getitem__NcCs
t||�Sr)�BoundMethodDispatcher)r�instance�ownerrrr!�__get__HszMethodDispatcher.__get__)r)N)�__name__�
__module__�__qualname__�__doc__rr%r)rrrr!r
,s
r
c@s@eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dS)r&zAWraps a MethodDispatcher, binding its return values to `instance`cCs||_||_dSr)r'�
dispatcher)rr'r.rrr!rNszBoundMethodDispatcher.__init__cCs|j|�|j�Sr)r.r)r'r#rrr!r%Rsz!BoundMethodDispatcher.__getitem__cCs||jvr||S|SdSr�r.)rr$rrrr!r"Ws
zBoundMethodDispatcher.getcCs
t|j�Sr)�iterr.�rrrr!�__iter__]szBoundMethodDispatcher.__iter__cCs
t|j�Sr)rr.r1rrr!�__len__`szBoundMethodDispatcher.__len__cCs
||jvSrr/r#rrr!�__contains__csz"BoundMethodDispatcher.__contains__N)
r*r+r,r-rr%r"r2r3r4rrrr!r&Lsr&cCsLt|�dkoJt|d�dkoJt|d�dkoJt|d�dkoJt|d�dkS)N�r��i�����i��)r�ord)�datarrr!rjs����rcCs,dt|d�ddt|d�d}|S)Nirr6ir7r8)r9)r:�char_valrrr!rps�rcsi���fdd�}|S)Ncs�ttjtd��rd|j}n
d|j}t|���}z�|||WSty�t|�}�|g|�Ri|��}|j�|�d�vr�i�|<d�|vr�i�||<d�||vr�i�|||<|�|||<|YS0dS)N�z_%s_factorys_%s_factoryr�args�kwargs)	rrr*�typerr�KeyError�__dict__�update)�
baseModuler=r>r�kwargs_tuple�mod�objs��factory�moduleCacherr!�
moduleFactory|s$
z+moduleFactoryFactory.<locals>.moduleFactoryr)rHrJrrGr!r
ysr
csi���fdd�}|S)Ncs6t|�t|���f}|�vr.�|i|���|<�|Sr)rr)r=r>r$��cache�funcrr!�wrapped�szmemoize.<locals>.wrappedr)rMrNrrKr!�memoize�srO) �
__future__rrr�typesr�collections.abcr�ImportError�collectionsZpip._vendor.sixrr�xml.etree.ElementTree�etree�ElementTreer	Zxml.etree.cElementTreeZcElementTree�__all__�eval�_xr�	Exceptionrrr
r&rrr
rOrrrr!�<module>s6

 	site-packages/pip/_vendor/html5lib/__pycache__/_tokenizer.cpython-39.pyc000064400000115524150564745010022261 0ustar00a

��Re�,�@s�ddlmZmZmZddlmZddlmZm	Z	ddl
mZddlm
Z
ddlmZddlmZmZdd	lmZmZmZdd
lmZmZddlmZddlmZdd
lmZee�Zedkr�eZne	ZGdd�de�ZdS)�)�absolute_import�division�unicode_literals)�unichr)�deque�OrderedDict)�version_info�)�spaceCharacters)�entities)�asciiLetters�asciiUpper2Lower)�digits�	hexDigits�EOF)�
tokenTypes�
tagTokenTypes)�replacementCharacters)�HTMLInputStream)�Trie)��csdeZdZdZd��fdd�	Zdd�Zdd�Zd�d
d�Zdd
�Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zd2d3�Zd4d5�Zd6d7�Zd8d9�Zd:d;�Zd<d=�Z d>d?�Z!d@dA�Z"dBdC�Z#dDdE�Z$dFdG�Z%dHdI�Z&dJdK�Z'dLdM�Z(dNdO�Z)dPdQ�Z*dRdS�Z+dTdU�Z,dVdW�Z-dXdY�Z.dZd[�Z/d\d]�Z0d^d_�Z1d`da�Z2dbdc�Z3ddde�Z4dfdg�Z5dhdi�Z6djdk�Z7dldm�Z8dndo�Z9dpdq�Z:drds�Z;dtdu�Z<dvdw�Z=dxdy�Z>dzd{�Z?d|d}�Z@d~d�ZAd�d��ZBd�d��ZCd�d��ZDd�d��ZEd�d��ZFd�d��ZGd�d��ZHd�d��ZId�d��ZJd�d��ZKd�d��ZL�ZMS)��
HTMLTokenizera	 This class takes care of tokenizing HTML.

    * self.currentToken
      Holds the token that is currently being processed.

    * self.state
      Holds a reference to the method to be invoked... XXX

    * self.stream
      Points to HTMLInputStream object.
    NcsJt|fi|��|_||_d|_g|_|j|_d|_d|_t	t
|���dS)NF)r�stream�parser�
escapeFlag�
lastFourChars�	dataState�state�escape�currentToken�superr�__init__)�selfrr�kwargs��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_tokenizer.pyr"(szHTMLTokenizer.__init__ccsPtg�|_|��rL|jjr6td|jj�d�d�Vq|jr
|j��Vq6q
dS)z� This is where the magic happens.

        We do our usually processing through the states and when we have a token
        to return we yield the token which pauses processing until the next token
        is requested.
        �
ParseErrorr��type�dataN)r�
tokenQueuerr�errorsr�pop�popleft�r#r'r'r(�__iter__7s
zHTMLTokenizer.__iter__c	Cs�t}d}|rt}d}g}|j��}||vrH|turH|�|�|j��}q"td�|�|�}|tvr�t|}|j	�t
ddd|id���nd|kr�d	ks�n|d
kr�d}|j	�t
ddd|id��n�d|kr�d
k�s>nd|kr�dk�s>nd|k�rdk�s>nd|k�r*dk�s>n|tgd��v�rZ|j	�t
ddd|id��zt|�}Wn<t
�y�|d}td|d?B�td|d@B�}Yn0|dk�r�|j	�t
ddd��|j�|�|S)z�This function returns either U+FFFD or the character based on the
        decimal or hexadecimal representation. It also discards ";" if present.
        If not present self.tokenQueue.append({"type": tokenTypes["ParseError"]}) is invoked.
        �
��r)z$illegal-codepoint-for-numeric-entity�	charAsInt�r+r,�datavarsi�i�����r	�����i�i�)#�i��i��i��i��i��i��i��i��i��i��i��i��i��i��i��i��i��i��i��	i��	i��
i��
i��i��i��i��i��
i��
i��i��i��i��i��r9ii�i��;z numeric-entity-without-semicolonr*)rrr�charr�append�int�joinrr-r�	frozenset�chr�
ValueError�unget)	r#�isHex�allowed�radix�	charStack�cr6rB�vr'r'r(�consumeNumberEntityGsn

���
��
�
�
�
�
�
�&
�z!HTMLTokenizer.consumeNumberEntityFc	Cs�d}|j��g}|dtvsB|dtddfvsB|durV||dkrV|j�|d��n|ddk�rd}|�|j���|ddvr�d}|�|j���|r�|dtvs�|s�|dtvr�|j�|d�|�|�}n4|j	�t
d	d
d��|j�|���dd�|�}�nf|dtu�rDt
�d�|���s0�qD|�|j����qz$t
�d�|dd���}t|�}Wnt�y�d}Yn0|du�r>|dd
k�r�|j	�t
d	dd��|dd
k�r|�r||tv�s�||tv�s�||dk�r|j�|���dd�|�}n.t|}|j�|���|d�||d��7}n4|j	�t
d	dd��|j�|���dd�|�}|�r�|jddd|7<n*|tv�r�d}nd}|j	�t
||d��dS)N�&r�<�#F���)�x�XTr)zexpected-numeric-entityr*r5rAznamed-entity-without-semicolon�=zexpected-named-entityr,r	�SpaceCharacters�
Characters)rrBr
rrIrCrrrPr-rr/rE�entitiesTrie�has_keys_with_prefix�longest_prefix�len�KeyErrorrrr )	r#�allowedChar�
fromAttribute�outputrM�hex�
entityName�entityLength�	tokenTyper'r'r(�
consumeEntity�s~�
��
��

�
�
�
��
zHTMLTokenizer.consumeEntitycCs|j|dd�dS)zIThis method replaces the need for "entityInAttributeValueState".
        T)r_r`N)rf)r#r_r'r'r(�processEntityInAttribute�sz&HTMLTokenizer.processEntityInAttributecCs�|j}|dtvr�|d�t�|d<|dtdkrp|d}t|�}t|�t|�krh|�|ddd��||d<|dtdkr�|dr�|j�	tdd	d
��|dr�|j�	tddd
��|j�	|�|j
|_dS)
z�This method is a generic handler for emitting the tags. It also sets
        the state to "data" because that's what's needed after a token has been
        emitted.
        r+�name�StartTagr,NrT�EndTagr)zattributes-in-end-tagr*�selfClosingzself-closing-flag-on-end-tag)r r�	translater
r�attributeMapr]�updater-rCrr)r#�token�rawr,r'r'r(�emitCurrentToken�s(��zHTMLTokenizer.emitCurrentTokencCs�|j��}|dkr|j|_n�|dkr.|j|_n�|dkrd|j�tddd��|j�tddd��n`|turpdS|t	vr�|j�td	||j�
t	d
�d��n&|j�
d�}|j�td||d��d
S)NrQrR�r)�invalid-codepointr*rYFrXT�rQrRrr)rrB�entityDataStater�tagOpenStater-rCrrr
�
charsUntil�r#r,�charsr'r'r(rs.


��
�
�zHTMLTokenizer.dataStatecCs|��|j|_dS�NT)rfrrr1r'r'r(ruszHTMLTokenizer.entityDataStatecCs�|j��}|dkr|j|_n�|dkr.|j|_n�|tkr:dS|dkrp|j�tddd��|j�tdd	d��nT|t	vr�|j�td
||j�
t	d�d��n&|j�
d�}|j�td||d��dS)
NrQrRFrrr)rsr*rYr:rXTrt)rrB�characterReferenceInRcdatar�rcdataLessThanSignStaterr-rCrr
rwrxr'r'r(�rcdataState"s.


��
�
�zHTMLTokenizer.rcdataStatecCs|��|j|_dSrz)rfr}rr1r'r'r(r{?sz(HTMLTokenizer.characterReferenceInRcdatacCs�|j��}|dkr|j|_nh|dkrR|j�tddd��|j�tddd��n2|tkr^dS|j�d	�}|j�td||d��d
S�NrRrrr)rsr*rYr:F)rRrrT)	rrB�rawtextLessThanSignStaterr-rCrrrwrxr'r'r(�rawtextStateDs"

��
�zHTMLTokenizer.rawtextStatecCs�|j��}|dkr|j|_nh|dkrR|j�tddd��|j�tddd��n2|tkr^dS|j�d	�}|j�td||d��d
Sr~)	rrB�scriptDataLessThanSignStaterr-rCrrrwrxr'r'r(�scriptDataStateVs"

��
�zHTMLTokenizer.scriptDataStatecCsr|j��}|tkrdS|dkrL|j�tddd��|j�tddd��n"|j�td||j�d�d��dS)	NFrrr)rsr*rYr:T)rrBrr-rCrrw�r#r,r'r'r(�plaintextStatehs
��
�zHTMLTokenizer.plaintextStatecCs|j��}|dkr|j|_n�|dkr.|j|_n�|tvrVtd|gddd�|_|j|_n�|dkr�|j	�
tddd	��|j	�
td
dd	��|j|_nt|dkr�|j	�
tdd
d	��|j�|�|j
|_n@|j	�
tddd	��|j	�
td
dd	��|j�|�|j|_dS)N�!�/riF)r+rhr,rk�selfClosingAcknowledged�>r)z'expected-tag-name-but-got-right-bracketr*rYz<>�?z'expected-tag-name-but-got-question-markzexpected-tag-namerRT)rrB�markupDeclarationOpenStater�closeTagOpenStaterrr �tagNameStater-rCrrI�bogusCommentStater�r'r'r(rvws>


�
�
�
�zHTMLTokenizer.tagOpenStatecCs�|j��}|tvr0td|gdd�|_|j|_n�|dkrX|j�tddd��|j	|_nn|t
ur�|j�tddd��|j�td	d
d��|j	|_n0|j�tddd|id
��|j�|�|j|_dS)NrjF�r+rhr,rkr�r)z*expected-closing-tag-but-got-right-bracketr*z expected-closing-tag-but-got-eofrY�</z!expected-closing-tag-but-got-charr,r7T)
rrBrrr r�rr-rCrrrIr�r�r'r'r(r��s0
�
�
�
�zHTMLTokenizer.closeTagOpenStatecCs�|j��}|tvr|j|_n�|dkr.|��n~|turV|j�t	ddd��|j
|_nV|dkrh|j|_nD|dkr�|j�t	ddd��|jdd	7<n|jd|7<d
S)Nr�r)zeof-in-tag-namer*r�rrrsrhr:T)
rrBr
�beforeAttributeNameStaterrqrr-rCrr�selfClosingStartTagStater r�r'r'r(r��s&


�

�zHTMLTokenizer.tagNameStatecCsP|j��}|dkr"d|_|j|_n*|j�tddd��|j�|�|j	|_dS�Nr�r5rYrRr*T)
rrB�temporaryBuffer�rcdataEndTagOpenStaterr-rCrrIr}r�r'r'r(r|�s

z%HTMLTokenizer.rcdataLessThanSignStatecCsX|j��}|tvr*|j|7_|j|_n*|j�tddd��|j�	|�|j
|_dS�NrYr�r*T)rrBrr��rcdataEndTagNameStaterr-rCrrIr}r�r'r'r(r��s

z#HTMLTokenizer.rcdataEndTagOpenStatecCs|jo|jd��|j��k}|j��}|tvrT|rTtd|jgdd�|_|j|_n�|dkr�|r�td|jgdd�|_|j	|_n||dkr�|r�td|jgdd�|_|�
�|j|_nH|tvr�|j|7_n0|j
�tdd|jd	��|j�|�|j|_d
S�NrhrjFr�r�r�rYr�r*T)r �lowerr�rrBr
rr�rr�rqrrr-rCrIr}�r#�appropriater,r'r'r(r��s:
�
�
�
�z#HTMLTokenizer.rcdataEndTagNameStatecCsP|j��}|dkr"d|_|j|_n*|j�tddd��|j�|�|j	|_dSr�)
rrBr��rawtextEndTagOpenStaterr-rCrrIr�r�r'r'r(r�s

z&HTMLTokenizer.rawtextLessThanSignStatecCsX|j��}|tvr*|j|7_|j|_n*|j�tddd��|j�	|�|j
|_dSr�)rrBrr��rawtextEndTagNameStaterr-rCrrIr�r�r'r'r(r�s

z$HTMLTokenizer.rawtextEndTagOpenStatecCs|jo|jd��|j��k}|j��}|tvrT|rTtd|jgdd�|_|j|_n�|dkr�|r�td|jgdd�|_|j	|_n||dkr�|r�td|jgdd�|_|�
�|j|_nH|tvr�|j|7_n0|j
�tdd|jd	��|j�|�|j|_d
Sr�)r r�r�rrBr
rr�rr�rqrrr-rCrIr�r�r'r'r(r�s:
�
�
�
�z$HTMLTokenizer.rawtextEndTagNameStatecCsx|j��}|dkr"d|_|j|_nR|dkrJ|j�tddd��|j|_n*|j�tddd��|j�	|�|j
|_dS)	Nr�r5r�rYz<!r*rRT)rrBr��scriptDataEndTagOpenStaterr-rCr�scriptDataEscapeStartStaterIr�r�r'r'r(r�,s


z)HTMLTokenizer.scriptDataLessThanSignStatecCsX|j��}|tvr*|j|7_|j|_n*|j�tddd��|j�	|�|j
|_dSr�)rrBrr��scriptDataEndTagNameStaterr-rCrrIr�r�r'r'r(r�:s

z'HTMLTokenizer.scriptDataEndTagOpenStatecCs|jo|jd��|j��k}|j��}|tvrT|rTtd|jgdd�|_|j|_n�|dkr�|r�td|jgdd�|_|j	|_n||dkr�|r�td|jgdd�|_|�
�|j|_nH|tvr�|j|7_n0|j
�tdd|jd	��|j�|�|j|_d
Sr�)r r�r�rrBr
rr�rr�rqrrr-rCrIr�r�r'r'r(r�Es:
�
�
�
�z'HTMLTokenizer.scriptDataEndTagNameStatecCsJ|j��}|dkr2|j�tddd��|j|_n|j�|�|j|_dS�N�-rYr*T)	rrBr-rCr�scriptDataEscapeStartDashStaterrIr�r�r'r'r(r�as

z(HTMLTokenizer.scriptDataEscapeStartStatecCsJ|j��}|dkr2|j�tddd��|j|_n|j�|�|j|_dSr�)	rrBr-rCr�scriptDataEscapedDashDashStaterrIr�r�r'r'r(r�ks

z,HTMLTokenizer.scriptDataEscapeStartDashStatecCs�|j��}|dkr2|j�tddd��|j|_n�|dkrD|j|_nn|dkrz|j�tddd��|j�tddd��n8|tkr�|j	|_n&|j�
d	�}|j�td||d��d
S)Nr�rYr*rRrrr)rsr:)rRr�rrT)rrBr-rCr�scriptDataEscapedDashStater�"scriptDataEscapedLessThanSignStaterrrwrxr'r'r(�scriptDataEscapedStateus(


��

�z$HTMLTokenizer.scriptDataEscapedStatecCs�|j��}|dkr2|j�tddd��|j|_n�|dkrD|j|_nn|dkr�|j�tddd��|j�tddd��|j|_n0|t	kr�|j
|_n|j�td|d��|j|_d	S)
Nr�rYr*rRrrr)rsr:T)rrBr-rCrr�rr�r�rrr�r'r'r(r��s&


��

z(HTMLTokenizer.scriptDataEscapedDashStatecCs�|j��}|dkr*|j�tddd��n�|dkr<|j|_n�|dkrd|j�tddd��|j|_nn|dkr�|j�tddd��|j�tdd	d��|j|_n0|t	kr�|j
|_n|j�td|d��|j|_d
S)Nr�rYr*rRr�rrr)rsr:T)rrBr-rCrr�rr�r�rrr�r'r'r(r��s*


��

z,HTMLTokenizer.scriptDataEscapedDashDashStatecCs�|j��}|dkr"d|_|j|_n\|tvrT|j�tdd|d��||_|j	|_n*|j�tddd��|j�
|�|j|_dSr�)rrBr�� scriptDataEscapedEndTagOpenStaterrr-rCr� scriptDataDoubleEscapeStartStaterIr�r�r'r'r(r��s


z0HTMLTokenizer.scriptDataEscapedLessThanSignStatecCsP|j��}|tvr"||_|j|_n*|j�tddd��|j�	|�|j
|_dSr�)rrBrr�� scriptDataEscapedEndTagNameStaterr-rCrrIr�r�r'r'r(r��s

z.HTMLTokenizer.scriptDataEscapedEndTagOpenStatecCs|jo|jd��|j��k}|j��}|tvrT|rTtd|jgdd�|_|j|_n�|dkr�|r�td|jgdd�|_|j	|_n||dkr�|r�td|jgdd�|_|�
�|j|_nH|tvr�|j|7_n0|j
�tdd|jd	��|j�|�|j|_d
Sr�)r r�r�rrBr
rr�rr�rqrrr-rCrIr�r�r'r'r(r��s:
�
�
�
�z.HTMLTokenizer.scriptDataEscapedEndTagNameStatecCs�|j��}|ttd�BvrR|j�td|d��|j��dkrH|j	|_
q�|j|_
nB|tvr�|j�td|d��|j|7_n|j�
|�|j|_
dS�N)r�r�rYr*�scriptT)rrBr
rFr-rCrr�r��scriptDataDoubleEscapedStaterr�rrIr�r'r'r(r��s


z.HTMLTokenizer.scriptDataDoubleEscapeStartStatecCs�|j��}|dkr2|j�tddd��|j|_n�|dkrZ|j�tddd��|j|_nt|dkr�|j�tddd��|j�tddd��n>|tkr�|j�tdd	d��|j	|_n|j�td|d��d
S�Nr�rYr*rRrrr)rsr:�eof-in-script-in-scriptT)
rrBr-rCr� scriptDataDoubleEscapedDashStater�(scriptDataDoubleEscapedLessThanSignStaterrr�r'r'r(r��s*


��
�
z*HTMLTokenizer.scriptDataDoubleEscapedStatecCs�|j��}|dkr2|j�tddd��|j|_n�|dkrZ|j�tddd��|j|_n�|dkr�|j�tddd��|j�tddd��|j|_nF|t	kr�|j�tdd	d��|j
|_n|j�td|d��|j|_d
Sr�)rrBr-rCr�$scriptDataDoubleEscapedDashDashStaterr�r�rrr�r'r'r(r�s.


��
�
z.HTMLTokenizer.scriptDataDoubleEscapedDashStatecCs|j��}|dkr*|j�tddd��n�|dkrR|j�tddd��|j|_n�|dkrz|j�tddd��|j|_n�|dkr�|j�tddd��|j�tdd	d��|j|_nF|t	kr�|j�tdd
d��|j
|_n|j�td|d��|j|_dS)Nr�rYr*rRr�rrr)rsr:r�T)rrBr-rCrr�rr�r�rrr�r'r'r(r�%s2


��
�
z2HTMLTokenizer.scriptDataDoubleEscapedDashDashStatecCsP|j��}|dkr8|j�tddd��d|_|j|_n|j�|�|j	|_dS)Nr�rYr*r5T)
rrBr-rCrr��scriptDataDoubleEscapeEndStaterrIr�r�r'r'r(r�>s

z6HTMLTokenizer.scriptDataDoubleEscapedLessThanSignStatecCs�|j��}|ttd�BvrR|j�td|d��|j��dkrH|j	|_
q�|j|_
nB|tvr�|j�td|d��|j|7_n|j�
|�|j|_
dSr�)rrBr
rFr-rCrr�r�r�rr�rrIr�r'r'r(r�Is


z,HTMLTokenizer.scriptDataDoubleEscapeEndStatecCs0|j��}|tvr$|j�td��n|tvrJ|jd�|dg�|j|_n�|dkr\|�	�n�|dkrn|j
|_n�|dvr�|j�tddd	��|jd�|dg�|j|_n�|d
kr�|j�tddd	��|jd�ddg�|j|_nF|t
u�r|j�tdd
d	��|j|_n|jd�|dg�|j|_dS)NTr,r5r�r�)�'�"rWrRr)�#invalid-character-in-attribute-namer*rrrsr:z#expected-attribute-name-but-got-eof)rrBr
rwrr rC�attributeNameStaterrqr�r-rrrr�r'r'r(r�Ys<



�
�

�
z&HTMLTokenizer.beforeAttributeNameStatecCs�|j��}d}d}|dkr&|j|_�n.|tvr\|jddd||j�td�7<d}n�|dkrjd}n�|tvr||j|_n�|dkr�|j	|_n�|d	kr�|j
�td
dd��|jdddd
7<d}n�|dv�r|j
�td
dd��|jddd|7<d}nH|t
u�r6|j
�td
dd��|j|_n|jddd|7<d}|�r�|jddd�t�|jddd<|jddd�D]>\}}|jddd|k�r�|j
�td
dd���qҐq�|�r�|��dS)NTFrWr,rTrr�r�rrr)rsr*r:�r�r�rRr�zeof-in-attribute-namezduplicate-attribute)rrB�beforeAttributeValueStaterrr rwr
�afterAttributeNameStater�r-rCrrrrlr
rq)r#r,�leavingThisState�	emitTokenrh�_r'r'r(r�ws^
�

�
�
�
��
z HTMLTokenizer.attributeNameStatecCsD|j��}|tvr$|j�td��n|dkr8|j|_�n|dkrJ|��n�|tvrp|jd�	|dg�|j
|_n�|dkr�|j|_n�|dkr�|j�	t
dd	d
��|jd�	ddg�|j
|_n�|dvr�|j�	t
dd
d
��|jd�	|dg�|j
|_nF|tu�r$|j�	t
ddd
��|j|_n|jd�	|dg�|j
|_dS)NTrWr�r,r5r�rrr)rsr*r:r�z&invalid-character-after-attribute-namezexpected-end-of-tag-but-got-eof)rrBr
rwr�rrqrr rCr�r�r-rrrr�r'r'r(r��s@



�
�

�
z%HTMLTokenizer.afterAttributeNameStatecCsh|j��}|tvr$|j�td��n@|dkr8|j|_�n,|dkrX|j|_|j�|��n|dkrj|j|_n�|dkr�|j	�
tddd��|��n�|d	kr�|j	�
tdd
d��|j
ddd
d7<|j|_n�|dv�r|j	�
tddd��|j
ddd
|7<|j|_nL|tu�rB|j	�
tddd��|j|_n"|j
ddd
|7<|j|_dS)NTr�rQr�r�r)z.expected-attribute-value-but-got-right-bracketr*rrrsr,rTr	r:)rWrR�`z"equals-in-unquoted-attribute-valuez$expected-attribute-value-but-got-eof)rrBr
rw�attributeValueDoubleQuotedStater�attributeValueUnQuotedStaterI�attributeValueSingleQuotedStater-rCrrqr rrr�r'r'r(r��sF

�
�

�

�
z'HTMLTokenizer.beforeAttributeValueStatecCs�|j��}|dkr|j|_n�|dkr0|�d�n�|dkrj|j�tddd��|jddd	d
7<nN|t	ur�|j�tddd��|j
|_n&|jddd	||j�d�7<d
S)Nr�rQrrr)rsr*r,rTr	r:z#eof-in-attribute-value-double-quote)r�rQrrT�rrB�afterAttributeValueStaterrgr-rCrr rrrwr�r'r'r(r��s&

��

�z-HTMLTokenizer.attributeValueDoubleQuotedStatecCs�|j��}|dkr|j|_n�|dkr0|�d�n�|dkrj|j�tddd��|jddd	d
7<nN|t	ur�|j�tddd��|j
|_n&|jddd	||j�d�7<d
S)Nr�rQrrr)rsr*r,rTr	r:z#eof-in-attribute-value-single-quote)r�rQrrTr�r�r'r'r(r�s&

��

�z-HTMLTokenizer.attributeValueSingleQuotedStatecCs|j��}|tvr|j|_n�|dkr0|�d�n�|dkrB|��n�|dvr||j�t	ddd��|j
ddd	|7<n�|d
kr�|j�t	ddd��|j
ddd	d7<nV|tur�|j�t	dd
d��|j|_n.|j
ddd	||j�
td�tB�7<dS)NrQr�)r�r�rWrRr�r)z0unexpected-character-in-unquoted-attribute-valuer*r,rTr	rrrsr:z eof-in-attribute-value-no-quotes)rQr�r�r�rWrRr�rrT)rrBr
r�rrgrqr-rCrr rrrwrFr�r'r'r(r�s4


���

�
z)HTMLTokenizer.attributeValueUnQuotedStatecCs�|j��}|tvr|j|_n�|dkr.|��np|dkr@|j|_n^|turt|j�	t
ddd��|j�|�|j|_n*|j�	t
ddd��|j�|�|j|_dS)Nr�r�r)z$unexpected-EOF-after-attribute-valuer*z*unexpected-character-after-attribute-valueT)
rrBr
r�rrqr�rr-rCrrIrr�r'r'r(r�.s&



�
�z&HTMLTokenizer.afterAttributeValueStatecCs�|j��}|dkr&d|jd<|��n^|turZ|j�tddd��|j�|�|j	|_
n*|j�tddd��|j�|�|j|_
dS)Nr�Trkr)z#unexpected-EOF-after-solidus-in-tagr*z)unexpected-character-after-solidus-in-tag)rrBr rqrr-rCrrIrrr�r�r'r'r(r�Bs 


�
�z&HTMLTokenizer.selfClosingStartTagStatecCsD|j�d�}|�dd�}|j�td|d��|j��|j|_dS)Nr�rrr:�Commentr*T)	rrw�replacer-rCrrBrrr�r'r'r(r�Ts�
zHTMLTokenizer.bogusCommentStatecCs�|j��g}|ddkrR|�|j���|ddkrPtddd�|_|j|_dSn�|ddvr�d}dD](}|�|j���|d|vrfd	}q�qf|r�td
ddddd�|_|j|_dSn�|ddk�rD|jdu�rD|jj	j
�rD|jj	j
dj|jj	jk�rDd}d
D].}|�|j���|d|k�rd	}�q2�q|�rD|j
|_dS|j�tddd��|�rt|j�|����qZ|j|_dS)NrTr�r�r5r*T)�d�D))�o�O�rN�C��t�T��y�Y��p�P��e�EF�Doctype)r+rh�publicId�systemId�correct�[)r�r��Ar�r�r�r)zexpected-dashes-or-doctype)rrBrCrr �commentStartStater�doctypeStater�tree�openElements�	namespace�defaultNamespace�cdataSectionStater-rIr/r�)r#rM�matched�expectedr'r'r(r�csZ����
�z(HTMLTokenizer.markupDeclarationOpenStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��|j�|j�|j|_nP|t	ur�|j�tdd
d��|j�|j�|j|_n|jd|7<|j
|_dS)Nr�rrr)rsr*r,r:r��incorrect-comment�eof-in-commentT)rrB�commentStartDashStaterr-rCrr rr�commentStater�r'r'r(r��s.

��
�
zHTMLTokenizer.commentStartStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��|j�|j�|j|_nT|t	ur�|j�tdd
d��|j�|j�|j|_n|jdd|7<|j
|_dS)Nr�rrr)rsr*r,�-�r�r�r�T)rrB�commentEndStaterr-rCrr rrr�r�r'r'r(r��s.

��
�
z#HTMLTokenizer.commentStartDashStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<nT|tur�|j�tddd��|j�|j�|j	|_n|jd||j�
d	�7<d
S)Nr�rrr)rsr*r,r:r�)r�rrT)rrB�commentEndDashStaterr-rCrr rrrwr�r'r'r(r��s$

��

�zHTMLTokenizer.commentStatecCs�|j��}|dkr|j|_n�|dkrV|j�tddd��|jdd7<|j|_nT|t	ur�|j�tddd��|j�|j�|j
|_n|jdd|7<|j|_d	S)
Nr�rrr)rsr*r,r�zeof-in-comment-end-dashT)rrBr�rr-rCrr r�rrr�r'r'r(r��s$

�
�
z!HTMLTokenizer.commentEndDashStatecCs,|j��}|dkr*|j�|j�|j|_n�|dkrd|j�tddd��|jdd7<|j|_n�|dkr�|j�tdd	d��|j	|_n�|d
kr�|j�tddd��|jd|7<nj|t
ur�|j�tddd��|j�|j�|j|_n4|j�tdd
d��|jdd|7<|j|_dS)Nr�rrr)rsr*r,u--�r�z,unexpected-bang-after-double-dash-in-commentr�z,unexpected-dash-after-double-dash-in-commentzeof-in-comment-double-dashzunexpected-char-in-commentz--T)rrBr-rCr rrrr��commentEndBangStaterr�r'r'r(r��s@

�
�
��
�zHTMLTokenizer.commentEndStatecCs�|j��}|dkr*|j�|j�|j|_n�|dkrN|jdd7<|j|_n�|dkr�|j�tddd��|jdd	7<|j	|_nT|t
ur�|j�tdd
d��|j�|j�|j|_n|jdd|7<|j	|_dS)Nr�r�r,z--!rrr)rsr*u--!�zeof-in-comment-end-bang-stateT)rrBr-rCr rrr�rr�rr�r'r'r(r�s,


�
�
z!HTMLTokenizer.commentEndBangStatecCs�|j��}|tvr|j|_nj|tur\|j�tddd��d|j	d<|j�|j	�|j
|_n*|j�tddd��|j�|�|j|_dS)Nr)�!expected-doctype-name-but-got-eofr*Fr�zneed-space-after-doctypeT)rrBr
�beforeDoctypeNameStaterrr-rCrr rrIr�r'r'r(r�s 

�

�zHTMLTokenizer.doctypeStatecCs�|j��}|tvrn�|dkrT|j�tddd��d|jd<|j�|j�|j|_n�|dkr�|j�tddd��d	|jd
<|j	|_nR|t
ur�|j�tddd��d|jd<|j�|j�|j|_n||jd
<|j	|_dS)
Nr�r)z+expected-doctype-name-but-got-right-bracketr*Fr�rrrsr:rhr�T)rrBr
r-rCrr rr�doctypeNameStaterr�r'r'r(r�*s4
�

�

�


z$HTMLTokenizer.beforeDoctypeNameStatecCs|j��}|tvr2|jd�t�|jd<|j|_n�|dkrh|jd�t�|jd<|j�	|j�|j
|_n�|dkr�|j�	tddd��|jdd7<|j|_nh|t
ur�|j�	tddd��d	|jd
<|jd�t�|jd<|j�	|j�|j
|_n|jd|7<dS)Nrhr�rrr)rsr*r:zeof-in-doctype-nameFr�T)rrBr
r rlr
�afterDoctypeNameStaterr-rCrrr�rr�r'r'r(r�Ds0


�
�

zHTMLTokenizer.doctypeNameStatecCsH|j��}|tvr�n.|dkr8|j�|j�|j|_�n|tur�d|jd<|j�	|�|j�t
ddd��|j�|j�|j|_n�|dvr�d}d	D]}|j��}||vr�d}q�q�|r�|j|_dSnD|d
v�r
d}dD]}|j��}||vr�d}q�q�|�r
|j|_dS|j�	|�|j�t
ddd
|id��d|jd<|j
|_dS)Nr�Fr�r)�eof-in-doctyper*r�T))�u�U)�b�B)�l�L)�i�Ir���s�S)r�r�r�r�)�m�Mz*expected-space-or-right-bracket-in-doctyper,r7)rrBr
r-rCr rrrrIr�afterDoctypePublicKeywordState�afterDoctypeSystemKeywordState�bogusDoctypeState)r#r,r�r�r'r'r(r�]sT

�



�
z#HTMLTokenizer.afterDoctypeNameStatecCs�|j��}|tvr|j|_n�|dvrP|j�tddd��|j�|�|j|_nT|t	ur�|j�tddd��d|j
d<|j�|j
�|j|_n|j�|�|j|_dS�	N)r�r�r)�unexpected-char-in-doctyper*r�Fr�T)rrBr
�"beforeDoctypePublicIdentifierStaterr-rCrrIrr rr�r'r'r(r�s&

�
�

z,HTMLTokenizer.afterDoctypePublicKeywordStatecCs�|j��}|tvrn�|dkr0d|jd<|j|_n�|dkrLd|jd<|j|_n�|dkr�|j�t	ddd��d	|jd
<|j�|j�|j
|_nh|tur�|j�t	ddd��d	|jd
<|j�|j�|j
|_n(|j�t	ddd��d	|jd
<|j|_d
S)Nr�r5r�r�r�r)�unexpected-end-of-doctyper*Fr�r�rT)
rrBr
r �(doctypePublicIdentifierDoubleQuotedStater�(doctypePublicIdentifierSingleQuotedStater-rCrrrrr�r'r'r(r�s:




�

�

�
z0HTMLTokenizer.beforeDoctypePublicIdentifierStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��d
|jd<|j�|j�|j|_nR|t	ur�|j�tddd��d
|jd<|j�|j�|j|_n|jd|7<d
S)Nr�rrr)rsr*r�r:r�rFr�r�T�
rrB�!afterDoctypePublicIdentifierStaterr-rCrr rrr�r'r'r(r�s0

��

�

z6HTMLTokenizer.doctypePublicIdentifierDoubleQuotedStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��d
|jd<|j�|j�|j|_nR|t	ur�|j�tddd��d
|jd<|j�|j�|j|_n|jd|7<d
S)Nr�rrr)rsr*r�r:r�rFr�r�Tr	r�r'r'r(r�s0

��

�

z6HTMLTokenizer.doctypePublicIdentifierSingleQuotedStatecCs|j��}|tvr|j|_n�|dkr<|j�|j�|j|_n�|dkrn|j�t	ddd��d|jd<|j
|_n�|dkr�|j�t	ddd��d|jd<|j|_nh|tur�|j�t	dd	d��d
|jd<|j�|j�|j|_n(|j�t	ddd��d
|jd<|j
|_dS)
Nr�r�r)rr*r5r�r�r�Fr�T)rrBr
�-betweenDoctypePublicAndSystemIdentifiersStaterr-rCr rr�(doctypeSystemIdentifierDoubleQuotedState�(doctypeSystemIdentifierSingleQuotedStaterrr�r'r'r(r
�s>


�

�

�

�
z/HTMLTokenizer.afterDoctypePublicIdentifierStatecCs�|j��}|tvrn�|dkr4|j�|j�|j|_n�|dkrPd|jd<|j|_n�|dkrld|jd<|j	|_nh|t
kr�|j�tddd��d	|jd
<|j�|j�|j|_n(|j�tddd��d	|jd
<|j|_dS)
Nr�r�r5r�r�r)r�r*Fr�rT)
rrBr
r-rCr rrrr
rrrr�r'r'r(rs2





�

�
z;HTMLTokenizer.betweenDoctypePublicAndSystemIdentifiersStatecCs�|j��}|tvr|j|_n�|dvrP|j�tddd��|j�|�|j|_nT|t	ur�|j�tddd��d|j
d<|j�|j
�|j|_n|j�|�|j|_dSr)rrBr
�"beforeDoctypeSystemIdentifierStaterr-rCrrIrr rr�r'r'r(r)s&

�
�

z,HTMLTokenizer.afterDoctypeSystemKeywordStatecCs�|j��}|tvrn�|dkr0d|jd<|j|_n�|dkrLd|jd<|j|_n�|dkr�|j�t	ddd��d	|jd
<|j�|j�|j
|_nh|tur�|j�t	ddd��d	|jd
<|j�|j�|j
|_n(|j�t	ddd��d	|jd
<|j|_dS)
Nr�r5r�r�r�r)rr*Fr�r�T)
rrBr
r rrr
r-rCrrrrr�r'r'r(r=s:




�

�

�
z0HTMLTokenizer.beforeDoctypeSystemIdentifierStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��d
|jd<|j�|j�|j|_nR|t	ur�|j�tddd��d
|jd<|j�|j�|j|_n|jd|7<d
S)Nr�rrr)rsr*r�r:r�rFr�r�T�
rrB�!afterDoctypeSystemIdentifierStaterr-rCrr rrr�r'r'r(rZs0

��

�

z6HTMLTokenizer.doctypeSystemIdentifierDoubleQuotedStatecCs�|j��}|dkr|j|_n�|dkrN|j�tddd��|jdd7<n�|dkr�|j�tdd	d��d
|jd<|j�|j�|j|_nR|t	ur�|j�tddd��d
|jd<|j�|j�|j|_n|jd|7<d
S)Nr�rrr)rsr*r�r:r�rFr�r�Trr�r'r'r(r
rs0

��

�

z6HTMLTokenizer.doctypeSystemIdentifierSingleQuotedStatecCs�|j��}|tvrn~|dkr4|j�|j�|j|_n^|turt|j�t	ddd��d|jd<|j�|j�|j|_n|j�t	ddd��|j
|_dS)	Nr�r)r�r*Fr�rT)rrBr
r-rCr rrrrrr�r'r'r(r�s$

�

�z/HTMLTokenizer.afterDoctypeSystemIdentifierStatecCsZ|j��}|dkr*|j�|j�|j|_n,|turV|j�|�|j�|j�|j|_ndS)Nr�T)	rrBr-rCr rrrrIr�r'r'r(r�s


zHTMLTokenizer.bogusDoctypeStatecCs�g}|�|j�d��|�|j�d��|j��}|tkr>q�q|dksJJ�|ddd�dkrv|ddd�|d<q�q|�|�qd�|�}|�d�}|dkr�t|�D]}|j�t	d	d
d��q�|�
dd�}|r�|j�t	d
|d��|j|_dS)N�]r�rT���z]]r5rrrr)rsr*r:rYT)
rCrrwrBrrE�count�ranger-rr�rr)r#r,rB�	nullCountr�r'r'r(r��s2


�
�zHTMLTokenizer.cdataSectionState)N)NF)N�__name__�
__module__�__qualname__�__doc__r"r2rPrfrgrqrrur}r{r�r�r�rvr�r�r|r�r�rr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr
rrrrr
rrr��
__classcell__r'r'r%r(rs�H
P#

6 "-3rN) �
__future__rrrZpip._vendor.sixrrG�collectionsrr�sysr�	constantsr
rrr
rrrrrr�_inputstreamr�_trierrZ�dictrm�objectrr'r'r'r(�<module>s site-packages/pip/_vendor/html5lib/treewalkers/__pycache__/base.cpython-39.pyc000064400000015574150564745010023356 0ustar00a

��Re4�@s�ddlmZmZmZddlmZddlmZmZm	Z	gd�Z
ejZej
ZejZejZejZejZdZd�e	�Z	Gdd	�d	e�ZGd
d�de�ZdS)
�)�absolute_import�division�unicode_literals)�Node�)�
namespaces�voidElements�spaceCharacters)	�DOCUMENT�DOCTYPE�TEXT�ELEMENT�COMMENT�ENTITY�UNKNOWN�
TreeWalker�NonRecursiveTreeWalkerz<#UNKNOWN#>�c@sleZdZdZdd�Zdd�Zdd�Zdd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zddd�Zdd�Z
dd�ZdS)rz}Walks a tree yielding tokens

    Tokens are dicts that all have a ``type`` field specifying the type of the
    token.

    cCs
||_dS)zCCreates a TreeWalker

        :arg tree: the tree to walk

        N)�tree)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/base.py�__init__szTreeWalker.__init__cCst�dS�N��NotImplementedError)rrrr�__iter__#szTreeWalker.__iter__cCs
d|d�S)z�Generates an error token with the given message

        :arg msg: the error message

        :returns: SerializeError token

        �SerializeError��type�datar)r�msgrrr�error&szTreeWalker.errorFccs$d|||d�V|r |�d�VdS)arGenerates an EmptyTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :arg attrs: the attributes of the element as a dict

        :arg hasChildren: whether or not to yield a SerializationError because
            this tag shouldn't have children

        :returns: EmptyTag token

        �EmptyTag�r�name�	namespacer zVoid element has childrenN�r")rr&r%�attrs�hasChildrenrrr�emptyTag0s�zTreeWalker.emptyTagcCsd|||d�S)z�Generates a StartTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :arg attrs: the attributes of the element as a dict

        :returns: StartTag token

        �StartTagr$r)rr&r%r(rrr�startTagEs
�zTreeWalker.startTagcCsd||d�S)z�Generates an EndTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :returns: EndTag token

        �EndTag)rr%r&r)rr&r%rrr�endTagVs
�zTreeWalker.endTagccsx|}|�t�}|dt|�t|��}|r6d|d�V|}|�t�}|t|�d�}|rdd|d�V|rtd|d�VdS)atGenerates SpaceCharacters and Characters tokens

        Depending on what's in the data, this generates one or more
        ``SpaceCharacters`` and ``Characters`` tokens.

        For example:

            >>> from html5lib.treewalkers.base import TreeWalker
            >>> # Give it an empty tree just so it instantiates
            >>> walker = TreeWalker([])
            >>> list(walker.text(''))
            []
            >>> list(walker.text('  '))
            [{u'data': '  ', u'type': u'SpaceCharacters'}]
            >>> list(walker.text(' abc '))  # doctest: +NORMALIZE_WHITESPACE
            [{u'data': ' ', u'type': u'SpaceCharacters'},
            {u'data': u'abc', u'type': u'Characters'},
            {u'data': u' ', u'type': u'SpaceCharacters'}]

        :arg data: the text data

        :returns: one or more ``SpaceCharacters`` and ``Characters`` tokens

        N�SpaceCharactersr�
Characters)�lstripr	�len�rstrip)rr Zmiddle�left�rightrrr�textds

zTreeWalker.textcCs
d|d�S)zdGenerates a Comment token

        :arg data: the comment

        :returns: Comment token

        �Commentrr)rr rrr�comment�szTreeWalker.commentNcCsd|||d�S)z�Generates a Doctype token

        :arg name:

        :arg publicId:

        :arg systemId:

        :returns: the Doctype token

        �Doctype)rr%�publicId�systemIdr)rr%r:r;rrr�doctype�s
�zTreeWalker.doctypecCs
d|d�S)zjGenerates an Entity token

        :arg name: the entity name

        :returns: an Entity token

        �Entity)rr%r)rr%rrr�entity�szTreeWalker.entitycCs|�d|�S)zHandles unknown node typeszUnknown node type: r')rZnodeTyperrr�unknown�szTreeWalker.unknown)F)NN)�__name__�
__module__�__qualname__�__doc__rrr"r*r,r.r6r8r<r>r?rrrrrs

&


rc@s4eZdZdd�Zdd�Zdd�Zdd�Zd	d
�ZdS)rcCst�dSrr�r�noderrr�getNodeDetails�sz%NonRecursiveTreeWalker.getNodeDetailscCst�dSrrrDrrr�
getFirstChild�sz$NonRecursiveTreeWalker.getFirstChildcCst�dSrrrDrrr�getNextSibling�sz%NonRecursiveTreeWalker.getNextSiblingcCst�dSrrrDrrr�
getParentNode�sz$NonRecursiveTreeWalker.getParentNodeccs�|j}|du�r�|�|�}|d|dd�}}d}|tkrJ|j|�Vn�|tkrj|j|�D]
}|Vq\n�|tkr�|\}}}}|r�|tdkr�|tvr�|�	||||�D]
}|Vq�d}n|�
|||�VnV|tkr�|�|d�Vn<|t
k�r|�|d�Vn |tk�rd}n|�|d�V|�r2|�|�}	nd}	|	du�rF|	}q|dur|�|�}|d|dd�}}|tk�r�|\}}}}|�r�|tdk�s�|tv�r�|�||�V|j|u�r�d}q|�|�}
|
du�r�|
}qn
|�|�}�qFqdS)Nr�F�htmlT)rrFrr<rr6r
rrr*r,rr8rr>r
r?rGr.rHrI)r�currentNode�detailsrr)�tokenr&r%�
attributesZ
firstChildZnextSiblingrrrr�s\



�






zNonRecursiveTreeWalker.__iter__N)r@rArBrFrGrHrIrrrrrr�s
rN)�
__future__rrr�xml.domr�	constantsrrr	�__all__Z
DOCUMENT_NODEr
ZDOCUMENT_TYPE_NODErZ	TEXT_NODErZELEMENT_NODEr
ZCOMMENT_NODErZENTITY_NODErr�join�objectrrrrrr�<module>s
!site-packages/pip/_vendor/html5lib/treewalkers/__pycache__/etree_lxml.cpython-39.pyc000064400000015016150564745010024573 0ustar00a

��Re��@s�ddlmZmZmZddlmZddlmZddlm	Z	ddl
mZddlm
Z
dd	lmZd
d�ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de
j�ZdS)�)�absolute_import�division�unicode_literals)�	text_type)�OrderedDict)�etree�)�
tag_regexp�)�base)�	_ihatexmlcCs*|durdSt|t�r|S|�dd�SdS)N�ascii�strict)�
isinstancer�decode)�s�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/etree_lxml.py�
ensure_strs

rc@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�RootcCs�||_g|_z:|jjrD|j�t|t|jj�t|jj�t|jj	���Wnt
yXYn0z|��}Wnt
y||}Yn0|��dur�|��}q~|dur�|j�|�|�
�}q�d|_d|_dS�N)Zelementtree�childrenZdocinfoZinternalDTD�append�DoctyperZ	root_name�	public_idZ
system_url�AttributeError�getrootZgetprevious�getnext�text�tail)�self�et�noderrr�__init__s,



�



z
Root.__init__cCs
|j|Sr)r�r �keyrrr�__getitem__3szRoot.__getitem__cCsdSrr�r rrrr6szRoot.getnextcCsdS�Nr
rr'rrr�__len__9szRoot.__len__N)�__name__�
__module__�__qualname__r#r&rr)rrrrrsrc@seZdZdd�Zdd�ZdS)rcCs(||_||_||_||_d|_d|_dSr)�	root_node�namer�	system_idrr)r r-r.rr/rrrr#>szDoctype.__init__cCs|jjdSr()r-rr'rrrrGszDoctype.getnextN�r*r+r,r#rrrrrr=s	rc@seZdZdd�Zdd�ZdS)�FragmentRootcs$�fdd�|D��_d�_�_dS)Ncsg|]}t�|��qSr)�FragmentWrapper)�.0�childr'rr�
<listcomp>M�z)FragmentRoot.__init__.<locals>.<listcomp>)rrr)r rrr'rr#LszFragmentRoot.__init__cCsdSrrr'rrrrPszFragmentRoot.getnextNr0rrrrr1Ksr1c@sTeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�ZdS)r2cCsT||_||_t|jd�r(t|jj�|_nd|_t|jd�rJt|jj�|_nd|_dS)Nrr)r-�obj�hasattrrrr)r Z
fragment_rootr7rrrr#UszFragmentWrapper.__init__cCst|j|�Sr)�getattrr7)r r.rrr�__getattr__aszFragmentWrapper.__getattr__cCs6|jj}|�|�}|t|�dkr.||dSdSdSr()r-r�index�len)r Zsiblings�idxrrrrds

zFragmentWrapper.getnextcCs
|j|Sr)r7r$rrrr&lszFragmentWrapper.__getitem__cCs
t|j�Sr)�boolr7r'rrr�__bool__oszFragmentWrapper.__bool__cCsdSrrr'rrr�	getparentrszFragmentWrapper.getparentcCs
t|j�Sr��strr7r'rrr�__str__uszFragmentWrapper.__str__cCs
t|j�SrrAr'rrr�__unicode__xszFragmentWrapper.__unicode__cCs
t|j�Sr)r<r7r'rrrr){szFragmentWrapper.__len__N)r*r+r,r#r:rr&r?r@rCrDr)rrrrr2Tsr2c@s4eZdZdd�Zdd�Zdd�Zdd�Zd	d
�ZdS)�
TreeWalkercCsJt|t�rt|�|_t|�}nt�|_t|�}tj�||�t	�
�|_dSr)r�list�set�fragmentChildrenr1rr�NonRecursiveTreeWalkerr#rZ
InfosetFilter�filter)r �treerrrr#�s


zTreeWalker.__init__c	Cs�t|t�r:|\}}|dvs&Jd|��tjtt||��fSt|t�rLtjfSt|t�rjtj	|j
|j|jfSt|t
�r�t|d�s�tjt|j�fS|jtjkr�tjt|j�fS|jtjkr�tjt|j�dd�fSt�t|j��}|�r�|��\}}nd}t|j�}t�}t|j���D]P\}}t|�}t|�}t�|�}|�rV|||�d�|�d�f<n||d|f<�qtj||j �!|�|t"|�dk�p�|jfSdS)N�rr�%Text nodes are text or tail, found %s�tagr
���rr)#r�tuplerZTEXTrr9rZDOCUMENTrZDOCTYPEr.rr/r2r8r7rNr�Comment�COMMENTr�EntityZENTITYr	�match�groupsrrF�attrib�items�groupZELEMENTrJZfromXmlNamer<)	r r"r%rT�	namespacerN�attrsr.�valuerrr�getNodeDetails�s<




�zTreeWalker.getNodeDetailscCsBt|t�rJd��t|�s(|js(Jd��|jr6|dfS|dSdS)NzText nodes have no childrenzNode has no childrenrr)rrPr<r)r r"rrr�
getFirstChild�s
zTreeWalker.getFirstChildcCsbt|t�rL|\}}|dvs&Jd|��|dkrDt|�r>|dSdSn|��S|jrZ|dfS|��S)NrLrMrrr)rrPr<rr�r r"r%rrr�getNextSibling�s
zTreeWalker.getNextSiblingcCsJt|t�r4|\}}|dvs&Jd|��|dkrB|Sn||jvrBdS|��S)NrLrMr)rrPrHr@r^rrr�
getParentNode�s

zTreeWalker.getParentNodeN)r*r+r,r#r\r]r_r`rrrrrEs
)	rEN)�
__future__rrrZpip._vendor.sixr�collectionsr�lxmlrZtreebuilders.etreer	�rrr�objectrrr1r2rIrErrrr�<module>s	&	+site-packages/pip/_vendor/html5lib/treewalkers/__pycache__/genshi.cpython-39.pyc000064400000003607150564745010023713 0ustar00a

��Re	�@s�ddlmZmZmZddlmZddlmZmZmZm	Z	m
Z
ddlmZmZm
Z
mZmZmZddlmZddlmZmZGd	d
�d
ej�ZdS)�)�absolute_import�division�unicode_literals)�QName)�START�END�
XML_NAMESPACE�DOCTYPE�TEXT)�START_NS�END_NS�START_CDATA�	END_CDATA�PI�COMMENT�)�base�)�voidElements�
namespacesc@seZdZdd�Zdd�ZdS)�
TreeWalkerccsXd}|jD](}|dur.|�||�D]
}|Vq"|}q
|durT|�|d�D]
}|VqHdS)N)�tree�tokens)�self�previous�event�token�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/genshi.py�__iter__
s
zTreeWalker.__iter__ccst|\}}}|tkr�|\}}|j}|j}	i}
|D]0\}}t|t�rR||
|j|jf<q.||
d|f<q.|	tdkr�|tvr�|�|	||
|p�|dtkp�|d|k�D]
}
|
Vq�n|�	|	||
�Vn�|tkr�|j}|j}	|	tdks�|tvr�|�
|	|�Vnz|tk�r|�|�Vnb|t
k�r2|�|�D]}
|
V�q"n>|tk�rJ|j|�Vn&|tttttttfv�rdn|�|�VdS)N�htmlrr)r�	localname�	namespace�
isinstancerrrZemptyTagrZstartTag�endTagr�commentr
�textr	�doctyperrrr
rr�unknown)rr�next�kind�data�_�tagZattribs�namer"Zconverted_attribs�k�vrrrrrsD



�




�zTreeWalker.tokensN)�__name__�
__module__�__qualname__rrrrrrrsrN)�
__future__rrrZgenshi.corerrrrr	r
rrr
rrr�r�	constantsrrrrrrr�<module>s site-packages/pip/_vendor/html5lib/treewalkers/__pycache__/etree.cpython-39.pyc000064400000006715150564745010023545 0ustar00a

��Re��@sjddlmZmZmZddlmZddlZddlmZddl	m
Z
ddlmZe�
d	�Zd
d�Zee�ZdS)�)�absolute_import�division�unicode_literals)�OrderedDictN)�string_types�)�base�)�moduleFactoryFactoryz
{([^}]*)}(.*)cs,|}|�d�j�G�fdd�dtj�}t�S)NZasdcs4eZdZdZ�fdd�Zdd�Zdd�Zdd	�Zd
S)z#getETreeBuilder.<locals>.TreeWalkera�Given the particular ElementTree representation, this implementation,
        to avoid using recursion, returns "nodes" as tuples with the following
        content:

        1. The current element

        2. The index of the element relative to its parent

        3. A stack of ancestor elements

        4. A flag "text", "tail" or None to indicate if the current node is a
           text node; either the text or tail of the current element (1)
        csHt|t�r2|\}}}}|dvr.tjt||�fS|}t|d�sD|��}|jdvrVtjfS|jdkr|tj	|j
|�d�|�d�fS|j�kr�tj|j
fSt|jt
�s�Jt|j���t�|j�}|r�|��\}}n
d}|j}t�}t|j���D]>\}	}
t�|	�}|�r|
||�d�|�d�f<q�|
|d|	f<q�tj|||t|��p@|j
fSdS)	N��text�tail�tag)Z
DOCUMENT_ROOTZDOCUMENT_FRAGMENTz
<!DOCTYPE>�publicId�systemIdrr	)�
isinstance�tuplerZTEXT�getattr�hasattr�getrootrZDOCUMENTZDOCTYPEr�get�COMMENTr�type�
tag_regexp�match�groupsr�list�attrib�items�groupZELEMENT�len)�self�node�elt�_�flagr�	namespacer�attrs�name�value�ZElementTreeCommentType���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/etree.py�getNodeDetails s<



�

�z2getETreeBuilder.<locals>.TreeWalker.getNodeDetailscSstt|t�r|\}}}}n|dgdf\}}}}|dvr8dS|jrJ|||dfSt|�rl|�|�|dd|dfSdSdS)Nrrr)rrrr �append�r!r"�element�key�parentsr%r+r+r,�
getFirstChildHs

z1getETreeBuilder.<locals>.TreeWalker.getFirstChildcSs�t|t�r|\}}}}ndS|dkrLt|�rF|�|�|dd|dfSdSnN|jrf|dkrf|||dfS|t|d�dkr�|d|d|d|dfSdSdS)Nrrr
���r)rrr r.r
r/r+r+r,�getNextSiblingYs

z2getETreeBuilder.<locals>.TreeWalker.getNextSiblingcSs�t|t�r|\}}}}ndS|dkr:|s,|S|||dfSnD|��}|sJ|St|d��|�dksdJ�|t|d��|�|dfSdS)Nrr4r)rr�popr�count�index)r!r"r0r1r2r%�parentr+r+r,�
getParentNodems
z1getETreeBuilder.<locals>.TreeWalker.getParentNodeN)�__name__�
__module__�__qualname__�__doc__r-r3r5r:r+r*r+r,�
TreeWalkers

(r?)�CommentrrZNonRecursiveTreeWalker�locals)ZElementTreeImplementation�ElementTreer?r+r*r,�getETreeBuildersnrC)�
__future__rrr�collectionsr�reZpip._vendor.sixr�r�_utilsr
�compilerrC�getETreeModuler+r+r+r,�<module>s
usite-packages/pip/_vendor/html5lib/treewalkers/__pycache__/__init__.cpython-39.pyc000064400000007705150564745010024200 0ustar00a

��ReW�@sZdZddlmZmZmZddlmZddlmZddgZ	iZ
d
d	d�Zd
d�Zdd�Z
dS)a�A collection of modules for iterating through different kinds of
tree, generating tokens identical to those produced by the tokenizer
module.

To create a tree walker for a new type of tree, you need to
implement a tree walker object (called TreeWalker by convention) that
implements a 'serialize' method which takes a tree as sole argument and
returns an iterator which generates tokens.
�)�absolute_import�division�unicode_literals�)�	constants)�
default_etree�
getTreeWalker�pprintNcKs�|��}|tvr�|dkr0ddlm}|jt|<nt|dkrPddlm}|jt|<nT|dkrpddlm}|jt|<n4|dkr�dd	lm}|d
ur�t}|j	|fi|��jSt�
|�S)a;Get a TreeWalker class for various types of tree with built-in support

    :arg str treeType: the name of the tree type required (case-insensitive).
        Supported values are:

        * "dom": The xml.dom.minidom DOM implementation
        * "etree": A generic walker for tree implementations exposing an
          elementtree-like interface (known to work with ElementTree,
          cElementTree and lxml.etree).
        * "lxml": Optimized walker for lxml.etree
        * "genshi": a Genshi stream

    :arg implementation: A module implementing the tree type e.g.
        xml.etree.ElementTree or cElementTree (Currently applies to the "etree"
        tree type only).

    :arg kwargs: keyword arguments passed to the etree walker--for other
        walkers, this has no effect

    :returns: a TreeWalker class

    �dom�)r
�genshi)r�lxml)�
etree_lxml�etree)rN)�lower�treeWalkerCache�r
�
TreeWalkerrrrr�getETreeModule�get)�treeType�implementation�kwargsr
rrr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/__init__.pyrs"ccshg}|D]D}|d}|dvr,|�|d�q|rFdd�|�d�Vg}|Vq|rddd�|�d�VdS)N�type)�
Characters�SpaceCharacters�datarr)rr)�append�join)�tokens�pendingCharacters�tokenrrrr�concatenateCharacterTokensAsr$c	Csbg}d}t|�D�]D}|d}|dv�r|drz|dtjdkrz|dtjvr`tj|d}n|d}d||df}n|d}|�dd	||f�|d
7}|d}t|���D]T\\}}	}
|r�|tjvr�tj|}n|}d||	f}n|	}|�dd	|||
f�q�|d
k�rV|d
8}q|dk�r2|d
8}q|dk�rX|�dd	||df�q|dk�r
|d�r�|d�r�|�dd	||d|d|d�r�|dndf�nF|d�r�|�dd	||d|df�n|�dd	||df�n|�dd	|f�q|dk�r0|�dd	||df�q|dk�rJd�sVJd��qtd|��qd�|�S) z�Pretty printer for tree walkers

    Takes a TreeWalker instance and pretty prints the output of walking the tree.

    :arg walker: a TreeWalker instance

    rr)�StartTag�EmptyTag�	namespace�htmlz%s %s�namez%s<%s>� rrz	%s%s="%s"r&�EndTag�Commentz
%s<!-- %s -->�Doctype�publicIdz%s<!DOCTYPE %s "%s" "%s">�systemIdrz%s<!DOCTYPE %s "" "%s">z%s<!DOCTYPE %s>z
%s<!DOCTYPE >rz%s"%s"rFzBconcatenateCharacterTokens should have got rid of all Space tokenszUnknown token type, %s�
)	r$r�
namespaces�prefixesr�sorted�items�
ValueErrorr )�walker�output�indentr#r�nsr)�attrsr'�	localname�valuerrrr	Psn









��
���


)N)�__doc__�
__future__rrrrr�_utilsr�__all__rrr$r	rrrr�<module>s

,site-packages/pip/_vendor/html5lib/treewalkers/__pycache__/dom.cpython-39.pyc000064400000003353150564745010023213 0ustar00a

��Re��@sBddlmZmZmZddlmZddlmZGdd�dej�Z	dS)�)�absolute_import�division�unicode_literals)�Node�)�basec@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�
TreeWalkercCs�|jtjkr tj|j|j|jfS|jtjtj	fvr>tj
|jfS|jtjkr�i}t
|j���D]8}|�|�}|jr�|j||j|jf<q\|j|d|jf<q\tj|j|j||��fS|jtjkr�tj|jfS|jtjtjfvr�tjfStj|jfSdS�N)ZnodeTyperZDOCUMENT_TYPE_NODErZDOCTYPE�name�publicId�systemIdZ	TEXT_NODEZCDATA_SECTION_NODEZTEXTZ	nodeValueZELEMENT_NODE�list�
attributes�keysZgetAttributeNodeZnamespaceURI�valueZ	localNameZELEMENT�nodeNameZ
hasChildNodesZCOMMENT_NODE�COMMENTZ
DOCUMENT_NODEZDOCUMENT_FRAGMENT_NODEZDOCUMENT�UNKNOWN)�self�node�attrs�attr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treewalkers/dom.py�getNodeDetails	s&
�zTreeWalker.getNodeDetailscCs|jSr	)Z
firstChild�rrrrr�
getFirstChild$szTreeWalker.getFirstChildcCs|jSr	)ZnextSiblingrrrr�getNextSibling'szTreeWalker.getNextSiblingcCs|jSr	)Z
parentNoderrrr�
getParentNode*szTreeWalker.getParentNodeN)�__name__�
__module__�__qualname__rrrrrrrrrsrN)
�
__future__rrr�xml.domr�rZNonRecursiveTreeWalkerrrrrr�<module>ssite-packages/pip/_vendor/html5lib/treewalkers/genshi.py000064400000004405150564745010017421 0ustar00from __future__ import absolute_import, division, unicode_literals

from genshi.core import QName
from genshi.core import START, END, XML_NAMESPACE, DOCTYPE, TEXT
from genshi.core import START_NS, END_NS, START_CDATA, END_CDATA, PI, COMMENT

from . import base

from ..constants import voidElements, namespaces


class TreeWalker(base.TreeWalker):
    def __iter__(self):
        # Buffer the events so we can pass in the following one
        previous = None
        for event in self.tree:
            if previous is not None:
                for token in self.tokens(previous, event):
                    yield token
            previous = event

        # Don't forget the final event!
        if previous is not None:
            for token in self.tokens(previous, None):
                yield token

    def tokens(self, event, next):
        kind, data, _ = event
        if kind == START:
            tag, attribs = data
            name = tag.localname
            namespace = tag.namespace
            converted_attribs = {}
            for k, v in attribs:
                if isinstance(k, QName):
                    converted_attribs[(k.namespace, k.localname)] = v
                else:
                    converted_attribs[(None, k)] = v

            if namespace == namespaces["html"] and name in voidElements:
                for token in self.emptyTag(namespace, name, converted_attribs,
                                           not next or next[0] != END or
                                           next[1] != tag):
                    yield token
            else:
                yield self.startTag(namespace, name, converted_attribs)

        elif kind == END:
            name = data.localname
            namespace = data.namespace
            if namespace != namespaces["html"] or name not in voidElements:
                yield self.endTag(namespace, name)

        elif kind == COMMENT:
            yield self.comment(data)

        elif kind == TEXT:
            for token in self.text(data):
                yield token

        elif kind == DOCTYPE:
            yield self.doctype(*data)

        elif kind in (XML_NAMESPACE, DOCTYPE, START_NS, END_NS,
                      START_CDATA, END_CDATA, PI):
            pass

        else:
            yield self.unknown(kind)
site-packages/pip/_vendor/html5lib/treewalkers/base.py000064400000016464150564745010017066 0ustar00from __future__ import absolute_import, division, unicode_literals

from xml.dom import Node
from ..constants import namespaces, voidElements, spaceCharacters

__all__ = ["DOCUMENT", "DOCTYPE", "TEXT", "ELEMENT", "COMMENT", "ENTITY", "UNKNOWN",
           "TreeWalker", "NonRecursiveTreeWalker"]

DOCUMENT = Node.DOCUMENT_NODE
DOCTYPE = Node.DOCUMENT_TYPE_NODE
TEXT = Node.TEXT_NODE
ELEMENT = Node.ELEMENT_NODE
COMMENT = Node.COMMENT_NODE
ENTITY = Node.ENTITY_NODE
UNKNOWN = "<#UNKNOWN#>"

spaceCharacters = "".join(spaceCharacters)


class TreeWalker(object):
    """Walks a tree yielding tokens

    Tokens are dicts that all have a ``type`` field specifying the type of the
    token.

    """
    def __init__(self, tree):
        """Creates a TreeWalker

        :arg tree: the tree to walk

        """
        self.tree = tree

    def __iter__(self):
        raise NotImplementedError

    def error(self, msg):
        """Generates an error token with the given message

        :arg msg: the error message

        :returns: SerializeError token

        """
        return {"type": "SerializeError", "data": msg}

    def emptyTag(self, namespace, name, attrs, hasChildren=False):
        """Generates an EmptyTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :arg attrs: the attributes of the element as a dict

        :arg hasChildren: whether or not to yield a SerializationError because
            this tag shouldn't have children

        :returns: EmptyTag token

        """
        yield {"type": "EmptyTag", "name": name,
               "namespace": namespace,
               "data": attrs}
        if hasChildren:
            yield self.error("Void element has children")

    def startTag(self, namespace, name, attrs):
        """Generates a StartTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :arg attrs: the attributes of the element as a dict

        :returns: StartTag token

        """
        return {"type": "StartTag",
                "name": name,
                "namespace": namespace,
                "data": attrs}

    def endTag(self, namespace, name):
        """Generates an EndTag token

        :arg namespace: the namespace of the token--can be ``None``

        :arg name: the name of the element

        :returns: EndTag token

        """
        return {"type": "EndTag",
                "name": name,
                "namespace": namespace}

    def text(self, data):
        """Generates SpaceCharacters and Characters tokens

        Depending on what's in the data, this generates one or more
        ``SpaceCharacters`` and ``Characters`` tokens.

        For example:

            >>> from html5lib.treewalkers.base import TreeWalker
            >>> # Give it an empty tree just so it instantiates
            >>> walker = TreeWalker([])
            >>> list(walker.text(''))
            []
            >>> list(walker.text('  '))
            [{u'data': '  ', u'type': u'SpaceCharacters'}]
            >>> list(walker.text(' abc '))  # doctest: +NORMALIZE_WHITESPACE
            [{u'data': ' ', u'type': u'SpaceCharacters'},
            {u'data': u'abc', u'type': u'Characters'},
            {u'data': u' ', u'type': u'SpaceCharacters'}]

        :arg data: the text data

        :returns: one or more ``SpaceCharacters`` and ``Characters`` tokens

        """
        data = data
        middle = data.lstrip(spaceCharacters)
        left = data[:len(data) - len(middle)]
        if left:
            yield {"type": "SpaceCharacters", "data": left}
        data = middle
        middle = data.rstrip(spaceCharacters)
        right = data[len(middle):]
        if middle:
            yield {"type": "Characters", "data": middle}
        if right:
            yield {"type": "SpaceCharacters", "data": right}

    def comment(self, data):
        """Generates a Comment token

        :arg data: the comment

        :returns: Comment token

        """
        return {"type": "Comment", "data": data}

    def doctype(self, name, publicId=None, systemId=None):
        """Generates a Doctype token

        :arg name:

        :arg publicId:

        :arg systemId:

        :returns: the Doctype token

        """
        return {"type": "Doctype",
                "name": name,
                "publicId": publicId,
                "systemId": systemId}

    def entity(self, name):
        """Generates an Entity token

        :arg name: the entity name

        :returns: an Entity token

        """
        return {"type": "Entity", "name": name}

    def unknown(self, nodeType):
        """Handles unknown node types"""
        return self.error("Unknown node type: " + nodeType)


class NonRecursiveTreeWalker(TreeWalker):
    def getNodeDetails(self, node):
        raise NotImplementedError

    def getFirstChild(self, node):
        raise NotImplementedError

    def getNextSibling(self, node):
        raise NotImplementedError

    def getParentNode(self, node):
        raise NotImplementedError

    def __iter__(self):
        currentNode = self.tree
        while currentNode is not None:
            details = self.getNodeDetails(currentNode)
            type, details = details[0], details[1:]
            hasChildren = False

            if type == DOCTYPE:
                yield self.doctype(*details)

            elif type == TEXT:
                for token in self.text(*details):
                    yield token

            elif type == ELEMENT:
                namespace, name, attributes, hasChildren = details
                if (not namespace or namespace == namespaces["html"]) and name in voidElements:
                    for token in self.emptyTag(namespace, name, attributes,
                                               hasChildren):
                        yield token
                    hasChildren = False
                else:
                    yield self.startTag(namespace, name, attributes)

            elif type == COMMENT:
                yield self.comment(details[0])

            elif type == ENTITY:
                yield self.entity(details[0])

            elif type == DOCUMENT:
                hasChildren = True

            else:
                yield self.unknown(details[0])

            if hasChildren:
                firstChild = self.getFirstChild(currentNode)
            else:
                firstChild = None

            if firstChild is not None:
                currentNode = firstChild
            else:
                while currentNode is not None:
                    details = self.getNodeDetails(currentNode)
                    type, details = details[0], details[1:]
                    if type == ELEMENT:
                        namespace, name, attributes, hasChildren = details
                        if (namespace and namespace != namespaces["html"]) or name not in voidElements:
                            yield self.endTag(namespace, name)
                    if self.tree is currentNode:
                        currentNode = None
                        break
                    nextSibling = self.getNextSibling(currentNode)
                    if nextSibling is not None:
                        currentNode = nextSibling
                        break
                    else:
                        currentNode = self.getParentNode(currentNode)
site-packages/pip/_vendor/html5lib/treewalkers/etree.py000064400000010707150564745010017252 0ustar00from __future__ import absolute_import, division, unicode_literals

from collections import OrderedDict
import re

from pip._vendor.six import string_types

from . import base
from .._utils import moduleFactoryFactory

tag_regexp = re.compile("{([^}]*)}(.*)")


def getETreeBuilder(ElementTreeImplementation):
    ElementTree = ElementTreeImplementation
    ElementTreeCommentType = ElementTree.Comment("asd").tag

    class TreeWalker(base.NonRecursiveTreeWalker):  # pylint:disable=unused-variable
        """Given the particular ElementTree representation, this implementation,
        to avoid using recursion, returns "nodes" as tuples with the following
        content:

        1. The current element

        2. The index of the element relative to its parent

        3. A stack of ancestor elements

        4. A flag "text", "tail" or None to indicate if the current node is a
           text node; either the text or tail of the current element (1)
        """
        def getNodeDetails(self, node):
            if isinstance(node, tuple):  # It might be the root Element
                elt, _, _, flag = node
                if flag in ("text", "tail"):
                    return base.TEXT, getattr(elt, flag)
                else:
                    node = elt

            if not(hasattr(node, "tag")):
                node = node.getroot()

            if node.tag in ("DOCUMENT_ROOT", "DOCUMENT_FRAGMENT"):
                return (base.DOCUMENT,)

            elif node.tag == "<!DOCTYPE>":
                return (base.DOCTYPE, node.text,
                        node.get("publicId"), node.get("systemId"))

            elif node.tag == ElementTreeCommentType:
                return base.COMMENT, node.text

            else:
                assert isinstance(node.tag, string_types), type(node.tag)
                # This is assumed to be an ordinary element
                match = tag_regexp.match(node.tag)
                if match:
                    namespace, tag = match.groups()
                else:
                    namespace = None
                    tag = node.tag
                attrs = OrderedDict()
                for name, value in list(node.attrib.items()):
                    match = tag_regexp.match(name)
                    if match:
                        attrs[(match.group(1), match.group(2))] = value
                    else:
                        attrs[(None, name)] = value
                return (base.ELEMENT, namespace, tag,
                        attrs, len(node) or node.text)

        def getFirstChild(self, node):
            if isinstance(node, tuple):
                element, key, parents, flag = node
            else:
                element, key, parents, flag = node, None, [], None

            if flag in ("text", "tail"):
                return None
            else:
                if element.text:
                    return element, key, parents, "text"
                elif len(element):
                    parents.append(element)
                    return element[0], 0, parents, None
                else:
                    return None

        def getNextSibling(self, node):
            if isinstance(node, tuple):
                element, key, parents, flag = node
            else:
                return None

            if flag == "text":
                if len(element):
                    parents.append(element)
                    return element[0], 0, parents, None
                else:
                    return None
            else:
                if element.tail and flag != "tail":
                    return element, key, parents, "tail"
                elif key < len(parents[-1]) - 1:
                    return parents[-1][key + 1], key + 1, parents, None
                else:
                    return None

        def getParentNode(self, node):
            if isinstance(node, tuple):
                element, key, parents, flag = node
            else:
                return None

            if flag == "text":
                if not parents:
                    return element
                else:
                    return element, key, parents, None
            else:
                parent = parents.pop()
                if not parents:
                    return parent
                else:
                    assert list(parents[-1]).count(parent) == 1
                    return parent, list(parents[-1]).index(parent), parents, None

    return locals()


getETreeModule = moduleFactoryFactory(getETreeBuilder)
site-packages/pip/_vendor/html5lib/treewalkers/dom.py000064400000002605150564745010016723 0ustar00from __future__ import absolute_import, division, unicode_literals

from xml.dom import Node

from . import base


class TreeWalker(base.NonRecursiveTreeWalker):
    def getNodeDetails(self, node):
        if node.nodeType == Node.DOCUMENT_TYPE_NODE:
            return base.DOCTYPE, node.name, node.publicId, node.systemId

        elif node.nodeType in (Node.TEXT_NODE, Node.CDATA_SECTION_NODE):
            return base.TEXT, node.nodeValue

        elif node.nodeType == Node.ELEMENT_NODE:
            attrs = {}
            for attr in list(node.attributes.keys()):
                attr = node.getAttributeNode(attr)
                if attr.namespaceURI:
                    attrs[(attr.namespaceURI, attr.localName)] = attr.value
                else:
                    attrs[(None, attr.name)] = attr.value
            return (base.ELEMENT, node.namespaceURI, node.nodeName,
                    attrs, node.hasChildNodes())

        elif node.nodeType == Node.COMMENT_NODE:
            return base.COMMENT, node.nodeValue

        elif node.nodeType in (Node.DOCUMENT_NODE, Node.DOCUMENT_FRAGMENT_NODE):
            return (base.DOCUMENT,)

        else:
            return base.UNKNOWN, node.nodeType

    def getFirstChild(self, node):
        return node.firstChild

    def getNextSibling(self, node):
        return node.nextSibling

    def getParentNode(self, node):
        return node.parentNode
site-packages/pip/_vendor/html5lib/treewalkers/__init__.py000064400000013127150564745010017704 0ustar00"""A collection of modules for iterating through different kinds of
tree, generating tokens identical to those produced by the tokenizer
module.

To create a tree walker for a new type of tree, you need to
implement a tree walker object (called TreeWalker by convention) that
implements a 'serialize' method which takes a tree as sole argument and
returns an iterator which generates tokens.
"""

from __future__ import absolute_import, division, unicode_literals

from .. import constants
from .._utils import default_etree

__all__ = ["getTreeWalker", "pprint"]

treeWalkerCache = {}


def getTreeWalker(treeType, implementation=None, **kwargs):
    """Get a TreeWalker class for various types of tree with built-in support

    :arg str treeType: the name of the tree type required (case-insensitive).
        Supported values are:

        * "dom": The xml.dom.minidom DOM implementation
        * "etree": A generic walker for tree implementations exposing an
          elementtree-like interface (known to work with ElementTree,
          cElementTree and lxml.etree).
        * "lxml": Optimized walker for lxml.etree
        * "genshi": a Genshi stream

    :arg implementation: A module implementing the tree type e.g.
        xml.etree.ElementTree or cElementTree (Currently applies to the "etree"
        tree type only).

    :arg kwargs: keyword arguments passed to the etree walker--for other
        walkers, this has no effect

    :returns: a TreeWalker class

    """

    treeType = treeType.lower()
    if treeType not in treeWalkerCache:
        if treeType == "dom":
            from . import dom
            treeWalkerCache[treeType] = dom.TreeWalker
        elif treeType == "genshi":
            from . import genshi
            treeWalkerCache[treeType] = genshi.TreeWalker
        elif treeType == "lxml":
            from . import etree_lxml
            treeWalkerCache[treeType] = etree_lxml.TreeWalker
        elif treeType == "etree":
            from . import etree
            if implementation is None:
                implementation = default_etree
            # XXX: NEVER cache here, caching is done in the etree submodule
            return etree.getETreeModule(implementation, **kwargs).TreeWalker
    return treeWalkerCache.get(treeType)


def concatenateCharacterTokens(tokens):
    pendingCharacters = []
    for token in tokens:
        type = token["type"]
        if type in ("Characters", "SpaceCharacters"):
            pendingCharacters.append(token["data"])
        else:
            if pendingCharacters:
                yield {"type": "Characters", "data": "".join(pendingCharacters)}
                pendingCharacters = []
            yield token
    if pendingCharacters:
        yield {"type": "Characters", "data": "".join(pendingCharacters)}


def pprint(walker):
    """Pretty printer for tree walkers

    Takes a TreeWalker instance and pretty prints the output of walking the tree.

    :arg walker: a TreeWalker instance

    """
    output = []
    indent = 0
    for token in concatenateCharacterTokens(walker):
        type = token["type"]
        if type in ("StartTag", "EmptyTag"):
            # tag name
            if token["namespace"] and token["namespace"] != constants.namespaces["html"]:
                if token["namespace"] in constants.prefixes:
                    ns = constants.prefixes[token["namespace"]]
                else:
                    ns = token["namespace"]
                name = "%s %s" % (ns, token["name"])
            else:
                name = token["name"]
            output.append("%s<%s>" % (" " * indent, name))
            indent += 2
            # attributes (sorted for consistent ordering)
            attrs = token["data"]
            for (namespace, localname), value in sorted(attrs.items()):
                if namespace:
                    if namespace in constants.prefixes:
                        ns = constants.prefixes[namespace]
                    else:
                        ns = namespace
                    name = "%s %s" % (ns, localname)
                else:
                    name = localname
                output.append("%s%s=\"%s\"" % (" " * indent, name, value))
            # self-closing
            if type == "EmptyTag":
                indent -= 2

        elif type == "EndTag":
            indent -= 2

        elif type == "Comment":
            output.append("%s<!-- %s -->" % (" " * indent, token["data"]))

        elif type == "Doctype":
            if token["name"]:
                if token["publicId"]:
                    output.append("""%s<!DOCTYPE %s "%s" "%s">""" %
                                  (" " * indent,
                                   token["name"],
                                   token["publicId"],
                                   token["systemId"] if token["systemId"] else ""))
                elif token["systemId"]:
                    output.append("""%s<!DOCTYPE %s "" "%s">""" %
                                  (" " * indent,
                                   token["name"],
                                   token["systemId"]))
                else:
                    output.append("%s<!DOCTYPE %s>" % (" " * indent,
                                                       token["name"]))
            else:
                output.append("%s<!DOCTYPE >" % (" " * indent,))

        elif type == "Characters":
            output.append("%s\"%s\"" % (" " * indent, token["data"]))

        elif type == "SpaceCharacters":
            assert False, "concatenateCharacterTokens should have got rid of all Space tokens"

        else:
            raise ValueError("Unknown token type, %s" % type)

    return "\n".join(output)
site-packages/pip/_vendor/html5lib/treewalkers/etree_lxml.py000064400000014325150564745010020306 0ustar00from __future__ import absolute_import, division, unicode_literals
from pip._vendor.six import text_type

from collections import OrderedDict

from lxml import etree
from ..treebuilders.etree import tag_regexp

from . import base

from .. import _ihatexml


def ensure_str(s):
    if s is None:
        return None
    elif isinstance(s, text_type):
        return s
    else:
        return s.decode("ascii", "strict")


class Root(object):
    def __init__(self, et):
        self.elementtree = et
        self.children = []

        try:
            if et.docinfo.internalDTD:
                self.children.append(Doctype(self,
                                             ensure_str(et.docinfo.root_name),
                                             ensure_str(et.docinfo.public_id),
                                             ensure_str(et.docinfo.system_url)))
        except AttributeError:
            pass

        try:
            node = et.getroot()
        except AttributeError:
            node = et

        while node.getprevious() is not None:
            node = node.getprevious()
        while node is not None:
            self.children.append(node)
            node = node.getnext()

        self.text = None
        self.tail = None

    def __getitem__(self, key):
        return self.children[key]

    def getnext(self):
        return None

    def __len__(self):
        return 1


class Doctype(object):
    def __init__(self, root_node, name, public_id, system_id):
        self.root_node = root_node
        self.name = name
        self.public_id = public_id
        self.system_id = system_id

        self.text = None
        self.tail = None

    def getnext(self):
        return self.root_node.children[1]


class FragmentRoot(Root):
    def __init__(self, children):
        self.children = [FragmentWrapper(self, child) for child in children]
        self.text = self.tail = None

    def getnext(self):
        return None


class FragmentWrapper(object):
    def __init__(self, fragment_root, obj):
        self.root_node = fragment_root
        self.obj = obj
        if hasattr(self.obj, 'text'):
            self.text = ensure_str(self.obj.text)
        else:
            self.text = None
        if hasattr(self.obj, 'tail'):
            self.tail = ensure_str(self.obj.tail)
        else:
            self.tail = None

    def __getattr__(self, name):
        return getattr(self.obj, name)

    def getnext(self):
        siblings = self.root_node.children
        idx = siblings.index(self)
        if idx < len(siblings) - 1:
            return siblings[idx + 1]
        else:
            return None

    def __getitem__(self, key):
        return self.obj[key]

    def __bool__(self):
        return bool(self.obj)

    def getparent(self):
        return None

    def __str__(self):
        return str(self.obj)

    def __unicode__(self):
        return str(self.obj)

    def __len__(self):
        return len(self.obj)


class TreeWalker(base.NonRecursiveTreeWalker):
    def __init__(self, tree):
        # pylint:disable=redefined-variable-type
        if isinstance(tree, list):
            self.fragmentChildren = set(tree)
            tree = FragmentRoot(tree)
        else:
            self.fragmentChildren = set()
            tree = Root(tree)
        base.NonRecursiveTreeWalker.__init__(self, tree)
        self.filter = _ihatexml.InfosetFilter()

    def getNodeDetails(self, node):
        if isinstance(node, tuple):  # Text node
            node, key = node
            assert key in ("text", "tail"), "Text nodes are text or tail, found %s" % key
            return base.TEXT, ensure_str(getattr(node, key))

        elif isinstance(node, Root):
            return (base.DOCUMENT,)

        elif isinstance(node, Doctype):
            return base.DOCTYPE, node.name, node.public_id, node.system_id

        elif isinstance(node, FragmentWrapper) and not hasattr(node, "tag"):
            return base.TEXT, ensure_str(node.obj)

        elif node.tag == etree.Comment:
            return base.COMMENT, ensure_str(node.text)

        elif node.tag == etree.Entity:
            return base.ENTITY, ensure_str(node.text)[1:-1]  # strip &;

        else:
            # This is assumed to be an ordinary element
            match = tag_regexp.match(ensure_str(node.tag))
            if match:
                namespace, tag = match.groups()
            else:
                namespace = None
                tag = ensure_str(node.tag)
            attrs = OrderedDict()
            for name, value in list(node.attrib.items()):
                name = ensure_str(name)
                value = ensure_str(value)
                match = tag_regexp.match(name)
                if match:
                    attrs[(match.group(1), match.group(2))] = value
                else:
                    attrs[(None, name)] = value
            return (base.ELEMENT, namespace, self.filter.fromXmlName(tag),
                    attrs, len(node) > 0 or node.text)

    def getFirstChild(self, node):
        assert not isinstance(node, tuple), "Text nodes have no children"

        assert len(node) or node.text, "Node has no children"
        if node.text:
            return (node, "text")
        else:
            return node[0]

    def getNextSibling(self, node):
        if isinstance(node, tuple):  # Text node
            node, key = node
            assert key in ("text", "tail"), "Text nodes are text or tail, found %s" % key
            if key == "text":
                # XXX: we cannot use a "bool(node) and node[0] or None" construct here
                # because node[0] might evaluate to False if it has no child element
                if len(node):
                    return node[0]
                else:
                    return None
            else:  # tail
                return node.getnext()

        return (node, "tail") if node.tail else node.getnext()

    def getParentNode(self, node):
        if isinstance(node, tuple):  # Text node
            node, key = node
            assert key in ("text", "tail"), "Text nodes are text or tail, found %s" % key
            if key == "text":
                return node
            # else: fallback to "normal" processing
        elif node in self.fragmentChildren:
            return None

        return node.getparent()
site-packages/pip/_vendor/html5lib/_inputstream.py000064400000077141150564745010016335 0ustar00from __future__ import absolute_import, division, unicode_literals

from pip._vendor.six import text_type
from pip._vendor.six.moves import http_client, urllib

import codecs
import re
from io import BytesIO, StringIO

from pip._vendor import webencodings

from .constants import EOF, spaceCharacters, asciiLetters, asciiUppercase
from .constants import _ReparseException
from . import _utils

# Non-unicode versions of constants for use in the pre-parser
spaceCharactersBytes = frozenset([item.encode("ascii") for item in spaceCharacters])
asciiLettersBytes = frozenset([item.encode("ascii") for item in asciiLetters])
asciiUppercaseBytes = frozenset([item.encode("ascii") for item in asciiUppercase])
spacesAngleBrackets = spaceCharactersBytes | frozenset([b">", b"<"])


invalid_unicode_no_surrogate = "[\u0001-\u0008\u000B\u000E-\u001F\u007F-\u009F\uFDD0-\uFDEF\uFFFE\uFFFF\U0001FFFE\U0001FFFF\U0002FFFE\U0002FFFF\U0003FFFE\U0003FFFF\U0004FFFE\U0004FFFF\U0005FFFE\U0005FFFF\U0006FFFE\U0006FFFF\U0007FFFE\U0007FFFF\U0008FFFE\U0008FFFF\U0009FFFE\U0009FFFF\U000AFFFE\U000AFFFF\U000BFFFE\U000BFFFF\U000CFFFE\U000CFFFF\U000DFFFE\U000DFFFF\U000EFFFE\U000EFFFF\U000FFFFE\U000FFFFF\U0010FFFE\U0010FFFF]"  # noqa

if _utils.supports_lone_surrogates:
    # Use one extra step of indirection and create surrogates with
    # eval. Not using this indirection would introduce an illegal
    # unicode literal on platforms not supporting such lone
    # surrogates.
    assert invalid_unicode_no_surrogate[-1] == "]" and invalid_unicode_no_surrogate.count("]") == 1
    invalid_unicode_re = re.compile(invalid_unicode_no_surrogate[:-1] +
                                    eval('"\\uD800-\\uDFFF"') +  # pylint:disable=eval-used
                                    "]")
else:
    invalid_unicode_re = re.compile(invalid_unicode_no_surrogate)

non_bmp_invalid_codepoints = {0x1FFFE, 0x1FFFF, 0x2FFFE, 0x2FFFF, 0x3FFFE,
                              0x3FFFF, 0x4FFFE, 0x4FFFF, 0x5FFFE, 0x5FFFF,
                              0x6FFFE, 0x6FFFF, 0x7FFFE, 0x7FFFF, 0x8FFFE,
                              0x8FFFF, 0x9FFFE, 0x9FFFF, 0xAFFFE, 0xAFFFF,
                              0xBFFFE, 0xBFFFF, 0xCFFFE, 0xCFFFF, 0xDFFFE,
                              0xDFFFF, 0xEFFFE, 0xEFFFF, 0xFFFFE, 0xFFFFF,
                              0x10FFFE, 0x10FFFF}

ascii_punctuation_re = re.compile("[\u0009-\u000D\u0020-\u002F\u003A-\u0040\u005C\u005B-\u0060\u007B-\u007E]")

# Cache for charsUntil()
charsUntilRegEx = {}


class BufferedStream(object):
    """Buffering for streams that do not have buffering of their own

    The buffer is implemented as a list of chunks on the assumption that
    joining many strings will be slow since it is O(n**2)
    """

    def __init__(self, stream):
        self.stream = stream
        self.buffer = []
        self.position = [-1, 0]  # chunk number, offset

    def tell(self):
        pos = 0
        for chunk in self.buffer[:self.position[0]]:
            pos += len(chunk)
        pos += self.position[1]
        return pos

    def seek(self, pos):
        assert pos <= self._bufferedBytes()
        offset = pos
        i = 0
        while len(self.buffer[i]) < offset:
            offset -= len(self.buffer[i])
            i += 1
        self.position = [i, offset]

    def read(self, bytes):
        if not self.buffer:
            return self._readStream(bytes)
        elif (self.position[0] == len(self.buffer) and
              self.position[1] == len(self.buffer[-1])):
            return self._readStream(bytes)
        else:
            return self._readFromBuffer(bytes)

    def _bufferedBytes(self):
        return sum([len(item) for item in self.buffer])

    def _readStream(self, bytes):
        data = self.stream.read(bytes)
        self.buffer.append(data)
        self.position[0] += 1
        self.position[1] = len(data)
        return data

    def _readFromBuffer(self, bytes):
        remainingBytes = bytes
        rv = []
        bufferIndex = self.position[0]
        bufferOffset = self.position[1]
        while bufferIndex < len(self.buffer) and remainingBytes != 0:
            assert remainingBytes > 0
            bufferedData = self.buffer[bufferIndex]

            if remainingBytes <= len(bufferedData) - bufferOffset:
                bytesToRead = remainingBytes
                self.position = [bufferIndex, bufferOffset + bytesToRead]
            else:
                bytesToRead = len(bufferedData) - bufferOffset
                self.position = [bufferIndex, len(bufferedData)]
                bufferIndex += 1
            rv.append(bufferedData[bufferOffset:bufferOffset + bytesToRead])
            remainingBytes -= bytesToRead

            bufferOffset = 0

        if remainingBytes:
            rv.append(self._readStream(remainingBytes))

        return b"".join(rv)


def HTMLInputStream(source, **kwargs):
    # Work around Python bug #20007: read(0) closes the connection.
    # http://bugs.python.org/issue20007
    if (isinstance(source, http_client.HTTPResponse) or
        # Also check for addinfourl wrapping HTTPResponse
        (isinstance(source, urllib.response.addbase) and
         isinstance(source.fp, http_client.HTTPResponse))):
        isUnicode = False
    elif hasattr(source, "read"):
        isUnicode = isinstance(source.read(0), text_type)
    else:
        isUnicode = isinstance(source, text_type)

    if isUnicode:
        encodings = [x for x in kwargs if x.endswith("_encoding")]
        if encodings:
            raise TypeError("Cannot set an encoding with a unicode input, set %r" % encodings)

        return HTMLUnicodeInputStream(source, **kwargs)
    else:
        return HTMLBinaryInputStream(source, **kwargs)


class HTMLUnicodeInputStream(object):
    """Provides a unicode stream of characters to the HTMLTokenizer.

    This class takes care of character encoding and removing or replacing
    incorrect byte-sequences and also provides column and line tracking.

    """

    _defaultChunkSize = 10240

    def __init__(self, source):
        """Initialises the HTMLInputStream.

        HTMLInputStream(source, [encoding]) -> Normalized stream from source
        for use by html5lib.

        source can be either a file-object, local filename or a string.

        The optional encoding parameter must be a string that indicates
        the encoding.  If specified, that encoding will be used,
        regardless of any BOM or later declaration (such as in a meta
        element)

        """

        if not _utils.supports_lone_surrogates:
            # Such platforms will have already checked for such
            # surrogate errors, so no need to do this checking.
            self.reportCharacterErrors = None
        elif len("\U0010FFFF") == 1:
            self.reportCharacterErrors = self.characterErrorsUCS4
        else:
            self.reportCharacterErrors = self.characterErrorsUCS2

        # List of where new lines occur
        self.newLines = [0]

        self.charEncoding = (lookupEncoding("utf-8"), "certain")
        self.dataStream = self.openStream(source)

        self.reset()

    def reset(self):
        self.chunk = ""
        self.chunkSize = 0
        self.chunkOffset = 0
        self.errors = []

        # number of (complete) lines in previous chunks
        self.prevNumLines = 0
        # number of columns in the last line of the previous chunk
        self.prevNumCols = 0

        # Deal with CR LF and surrogates split over chunk boundaries
        self._bufferedCharacter = None

    def openStream(self, source):
        """Produces a file object from source.

        source can be either a file object, local filename or a string.

        """
        # Already a file object
        if hasattr(source, 'read'):
            stream = source
        else:
            stream = StringIO(source)

        return stream

    def _position(self, offset):
        chunk = self.chunk
        nLines = chunk.count('\n', 0, offset)
        positionLine = self.prevNumLines + nLines
        lastLinePos = chunk.rfind('\n', 0, offset)
        if lastLinePos == -1:
            positionColumn = self.prevNumCols + offset
        else:
            positionColumn = offset - (lastLinePos + 1)
        return (positionLine, positionColumn)

    def position(self):
        """Returns (line, col) of the current position in the stream."""
        line, col = self._position(self.chunkOffset)
        return (line + 1, col)

    def char(self):
        """ Read one character from the stream or queue if available. Return
            EOF when EOF is reached.
        """
        # Read a new chunk from the input stream if necessary
        if self.chunkOffset >= self.chunkSize:
            if not self.readChunk():
                return EOF

        chunkOffset = self.chunkOffset
        char = self.chunk[chunkOffset]
        self.chunkOffset = chunkOffset + 1

        return char

    def readChunk(self, chunkSize=None):
        if chunkSize is None:
            chunkSize = self._defaultChunkSize

        self.prevNumLines, self.prevNumCols = self._position(self.chunkSize)

        self.chunk = ""
        self.chunkSize = 0
        self.chunkOffset = 0

        data = self.dataStream.read(chunkSize)

        # Deal with CR LF and surrogates broken across chunks
        if self._bufferedCharacter:
            data = self._bufferedCharacter + data
            self._bufferedCharacter = None
        elif not data:
            # We have no more data, bye-bye stream
            return False

        if len(data) > 1:
            lastv = ord(data[-1])
            if lastv == 0x0D or 0xD800 <= lastv <= 0xDBFF:
                self._bufferedCharacter = data[-1]
                data = data[:-1]

        if self.reportCharacterErrors:
            self.reportCharacterErrors(data)

        # Replace invalid characters
        data = data.replace("\r\n", "\n")
        data = data.replace("\r", "\n")

        self.chunk = data
        self.chunkSize = len(data)

        return True

    def characterErrorsUCS4(self, data):
        for _ in range(len(invalid_unicode_re.findall(data))):
            self.errors.append("invalid-codepoint")

    def characterErrorsUCS2(self, data):
        # Someone picked the wrong compile option
        # You lose
        skip = False
        for match in invalid_unicode_re.finditer(data):
            if skip:
                continue
            codepoint = ord(match.group())
            pos = match.start()
            # Pretty sure there should be endianness issues here
            if _utils.isSurrogatePair(data[pos:pos + 2]):
                # We have a surrogate pair!
                char_val = _utils.surrogatePairToCodepoint(data[pos:pos + 2])
                if char_val in non_bmp_invalid_codepoints:
                    self.errors.append("invalid-codepoint")
                skip = True
            elif (codepoint >= 0xD800 and codepoint <= 0xDFFF and
                  pos == len(data) - 1):
                self.errors.append("invalid-codepoint")
            else:
                skip = False
                self.errors.append("invalid-codepoint")

    def charsUntil(self, characters, opposite=False):
        """ Returns a string of characters from the stream up to but not
        including any character in 'characters' or EOF. 'characters' must be
        a container that supports the 'in' method and iteration over its
        characters.
        """

        # Use a cache of regexps to find the required characters
        try:
            chars = charsUntilRegEx[(characters, opposite)]
        except KeyError:
            if __debug__:
                for c in characters:
                    assert(ord(c) < 128)
            regex = "".join(["\\x%02x" % ord(c) for c in characters])
            if not opposite:
                regex = "^%s" % regex
            chars = charsUntilRegEx[(characters, opposite)] = re.compile("[%s]+" % regex)

        rv = []

        while True:
            # Find the longest matching prefix
            m = chars.match(self.chunk, self.chunkOffset)
            if m is None:
                # If nothing matched, and it wasn't because we ran out of chunk,
                # then stop
                if self.chunkOffset != self.chunkSize:
                    break
            else:
                end = m.end()
                # If not the whole chunk matched, return everything
                # up to the part that didn't match
                if end != self.chunkSize:
                    rv.append(self.chunk[self.chunkOffset:end])
                    self.chunkOffset = end
                    break
            # If the whole remainder of the chunk matched,
            # use it all and read the next chunk
            rv.append(self.chunk[self.chunkOffset:])
            if not self.readChunk():
                # Reached EOF
                break

        r = "".join(rv)
        return r

    def unget(self, char):
        # Only one character is allowed to be ungotten at once - it must
        # be consumed again before any further call to unget
        if char is not EOF:
            if self.chunkOffset == 0:
                # unget is called quite rarely, so it's a good idea to do
                # more work here if it saves a bit of work in the frequently
                # called char and charsUntil.
                # So, just prepend the ungotten character onto the current
                # chunk:
                self.chunk = char + self.chunk
                self.chunkSize += 1
            else:
                self.chunkOffset -= 1
                assert self.chunk[self.chunkOffset] == char


class HTMLBinaryInputStream(HTMLUnicodeInputStream):
    """Provides a unicode stream of characters to the HTMLTokenizer.

    This class takes care of character encoding and removing or replacing
    incorrect byte-sequences and also provides column and line tracking.

    """

    def __init__(self, source, override_encoding=None, transport_encoding=None,
                 same_origin_parent_encoding=None, likely_encoding=None,
                 default_encoding="windows-1252", useChardet=True):
        """Initialises the HTMLInputStream.

        HTMLInputStream(source, [encoding]) -> Normalized stream from source
        for use by html5lib.

        source can be either a file-object, local filename or a string.

        The optional encoding parameter must be a string that indicates
        the encoding.  If specified, that encoding will be used,
        regardless of any BOM or later declaration (such as in a meta
        element)

        """
        # Raw Stream - for unicode objects this will encode to utf-8 and set
        #              self.charEncoding as appropriate
        self.rawStream = self.openStream(source)

        HTMLUnicodeInputStream.__init__(self, self.rawStream)

        # Encoding Information
        # Number of bytes to use when looking for a meta element with
        # encoding information
        self.numBytesMeta = 1024
        # Number of bytes to use when using detecting encoding using chardet
        self.numBytesChardet = 100
        # Things from args
        self.override_encoding = override_encoding
        self.transport_encoding = transport_encoding
        self.same_origin_parent_encoding = same_origin_parent_encoding
        self.likely_encoding = likely_encoding
        self.default_encoding = default_encoding

        # Determine encoding
        self.charEncoding = self.determineEncoding(useChardet)
        assert self.charEncoding[0] is not None

        # Call superclass
        self.reset()

    def reset(self):
        self.dataStream = self.charEncoding[0].codec_info.streamreader(self.rawStream, 'replace')
        HTMLUnicodeInputStream.reset(self)

    def openStream(self, source):
        """Produces a file object from source.

        source can be either a file object, local filename or a string.

        """
        # Already a file object
        if hasattr(source, 'read'):
            stream = source
        else:
            stream = BytesIO(source)

        try:
            stream.seek(stream.tell())
        except Exception:
            stream = BufferedStream(stream)

        return stream

    def determineEncoding(self, chardet=True):
        # BOMs take precedence over everything
        # This will also read past the BOM if present
        charEncoding = self.detectBOM(), "certain"
        if charEncoding[0] is not None:
            return charEncoding

        # If we've been overridden, we've been overridden
        charEncoding = lookupEncoding(self.override_encoding), "certain"
        if charEncoding[0] is not None:
            return charEncoding

        # Now check the transport layer
        charEncoding = lookupEncoding(self.transport_encoding), "certain"
        if charEncoding[0] is not None:
            return charEncoding

        # Look for meta elements with encoding information
        charEncoding = self.detectEncodingMeta(), "tentative"
        if charEncoding[0] is not None:
            return charEncoding

        # Parent document encoding
        charEncoding = lookupEncoding(self.same_origin_parent_encoding), "tentative"
        if charEncoding[0] is not None and not charEncoding[0].name.startswith("utf-16"):
            return charEncoding

        # "likely" encoding
        charEncoding = lookupEncoding(self.likely_encoding), "tentative"
        if charEncoding[0] is not None:
            return charEncoding

        # Guess with chardet, if available
        if chardet:
            try:
                from pip._vendor.chardet.universaldetector import UniversalDetector
            except ImportError:
                pass
            else:
                buffers = []
                detector = UniversalDetector()
                while not detector.done:
                    buffer = self.rawStream.read(self.numBytesChardet)
                    assert isinstance(buffer, bytes)
                    if not buffer:
                        break
                    buffers.append(buffer)
                    detector.feed(buffer)
                detector.close()
                encoding = lookupEncoding(detector.result['encoding'])
                self.rawStream.seek(0)
                if encoding is not None:
                    return encoding, "tentative"

        # Try the default encoding
        charEncoding = lookupEncoding(self.default_encoding), "tentative"
        if charEncoding[0] is not None:
            return charEncoding

        # Fallback to html5lib's default if even that hasn't worked
        return lookupEncoding("windows-1252"), "tentative"

    def changeEncoding(self, newEncoding):
        assert self.charEncoding[1] != "certain"
        newEncoding = lookupEncoding(newEncoding)
        if newEncoding is None:
            return
        if newEncoding.name in ("utf-16be", "utf-16le"):
            newEncoding = lookupEncoding("utf-8")
            assert newEncoding is not None
        elif newEncoding == self.charEncoding[0]:
            self.charEncoding = (self.charEncoding[0], "certain")
        else:
            self.rawStream.seek(0)
            self.charEncoding = (newEncoding, "certain")
            self.reset()
            raise _ReparseException("Encoding changed from %s to %s" % (self.charEncoding[0], newEncoding))

    def detectBOM(self):
        """Attempts to detect at BOM at the start of the stream. If
        an encoding can be determined from the BOM return the name of the
        encoding otherwise return None"""
        bomDict = {
            codecs.BOM_UTF8: 'utf-8',
            codecs.BOM_UTF16_LE: 'utf-16le', codecs.BOM_UTF16_BE: 'utf-16be',
            codecs.BOM_UTF32_LE: 'utf-32le', codecs.BOM_UTF32_BE: 'utf-32be'
        }

        # Go to beginning of file and read in 4 bytes
        string = self.rawStream.read(4)
        assert isinstance(string, bytes)

        # Try detecting the BOM using bytes from the string
        encoding = bomDict.get(string[:3])         # UTF-8
        seek = 3
        if not encoding:
            # Need to detect UTF-32 before UTF-16
            encoding = bomDict.get(string)         # UTF-32
            seek = 4
            if not encoding:
                encoding = bomDict.get(string[:2])  # UTF-16
                seek = 2

        # Set the read position past the BOM if one was found, otherwise
        # set it to the start of the stream
        if encoding:
            self.rawStream.seek(seek)
            return lookupEncoding(encoding)
        else:
            self.rawStream.seek(0)
            return None

    def detectEncodingMeta(self):
        """Report the encoding declared by the meta element
        """
        buffer = self.rawStream.read(self.numBytesMeta)
        assert isinstance(buffer, bytes)
        parser = EncodingParser(buffer)
        self.rawStream.seek(0)
        encoding = parser.getEncoding()

        if encoding is not None and encoding.name in ("utf-16be", "utf-16le"):
            encoding = lookupEncoding("utf-8")

        return encoding


class EncodingBytes(bytes):
    """String-like object with an associated position and various extra methods
    If the position is ever greater than the string length then an exception is
    raised"""
    def __new__(self, value):
        assert isinstance(value, bytes)
        return bytes.__new__(self, value.lower())

    def __init__(self, value):
        # pylint:disable=unused-argument
        self._position = -1

    def __iter__(self):
        return self

    def __next__(self):
        p = self._position = self._position + 1
        if p >= len(self):
            raise StopIteration
        elif p < 0:
            raise TypeError
        return self[p:p + 1]

    def next(self):
        # Py2 compat
        return self.__next__()

    def previous(self):
        p = self._position
        if p >= len(self):
            raise StopIteration
        elif p < 0:
            raise TypeError
        self._position = p = p - 1
        return self[p:p + 1]

    def setPosition(self, position):
        if self._position >= len(self):
            raise StopIteration
        self._position = position

    def getPosition(self):
        if self._position >= len(self):
            raise StopIteration
        if self._position >= 0:
            return self._position
        else:
            return None

    position = property(getPosition, setPosition)

    def getCurrentByte(self):
        return self[self.position:self.position + 1]

    currentByte = property(getCurrentByte)

    def skip(self, chars=spaceCharactersBytes):
        """Skip past a list of characters"""
        p = self.position               # use property for the error-checking
        while p < len(self):
            c = self[p:p + 1]
            if c not in chars:
                self._position = p
                return c
            p += 1
        self._position = p
        return None

    def skipUntil(self, chars):
        p = self.position
        while p < len(self):
            c = self[p:p + 1]
            if c in chars:
                self._position = p
                return c
            p += 1
        self._position = p
        return None

    def matchBytes(self, bytes):
        """Look for a sequence of bytes at the start of a string. If the bytes
        are found return True and advance the position to the byte after the
        match. Otherwise return False and leave the position alone"""
        rv = self.startswith(bytes, self.position)
        if rv:
            self.position += len(bytes)
        return rv

    def jumpTo(self, bytes):
        """Look for the next sequence of bytes matching a given sequence. If
        a match is found advance the position to the last byte of the match"""
        try:
            self._position = self.index(bytes, self.position) + len(bytes) - 1
        except ValueError:
            raise StopIteration
        return True


class EncodingParser(object):
    """Mini parser for detecting character encoding from meta elements"""

    def __init__(self, data):
        """string - the data to work on for encoding detection"""
        self.data = EncodingBytes(data)
        self.encoding = None

    def getEncoding(self):
        if b"<meta" not in self.data:
            return None

        methodDispatch = (
            (b"<!--", self.handleComment),
            (b"<meta", self.handleMeta),
            (b"</", self.handlePossibleEndTag),
            (b"<!", self.handleOther),
            (b"<?", self.handleOther),
            (b"<", self.handlePossibleStartTag))
        for _ in self.data:
            keepParsing = True
            try:
                self.data.jumpTo(b"<")
            except StopIteration:
                break
            for key, method in methodDispatch:
                if self.data.matchBytes(key):
                    try:
                        keepParsing = method()
                        break
                    except StopIteration:
                        keepParsing = False
                        break
            if not keepParsing:
                break

        return self.encoding

    def handleComment(self):
        """Skip over comments"""
        return self.data.jumpTo(b"-->")

    def handleMeta(self):
        if self.data.currentByte not in spaceCharactersBytes:
            # if we have <meta not followed by a space so just keep going
            return True
        # We have a valid meta element we want to search for attributes
        hasPragma = False
        pendingEncoding = None
        while True:
            # Try to find the next attribute after the current position
            attr = self.getAttribute()
            if attr is None:
                return True
            else:
                if attr[0] == b"http-equiv":
                    hasPragma = attr[1] == b"content-type"
                    if hasPragma and pendingEncoding is not None:
                        self.encoding = pendingEncoding
                        return False
                elif attr[0] == b"charset":
                    tentativeEncoding = attr[1]
                    codec = lookupEncoding(tentativeEncoding)
                    if codec is not None:
                        self.encoding = codec
                        return False
                elif attr[0] == b"content":
                    contentParser = ContentAttrParser(EncodingBytes(attr[1]))
                    tentativeEncoding = contentParser.parse()
                    if tentativeEncoding is not None:
                        codec = lookupEncoding(tentativeEncoding)
                        if codec is not None:
                            if hasPragma:
                                self.encoding = codec
                                return False
                            else:
                                pendingEncoding = codec

    def handlePossibleStartTag(self):
        return self.handlePossibleTag(False)

    def handlePossibleEndTag(self):
        next(self.data)
        return self.handlePossibleTag(True)

    def handlePossibleTag(self, endTag):
        data = self.data
        if data.currentByte not in asciiLettersBytes:
            # If the next byte is not an ascii letter either ignore this
            # fragment (possible start tag case) or treat it according to
            # handleOther
            if endTag:
                data.previous()
                self.handleOther()
            return True

        c = data.skipUntil(spacesAngleBrackets)
        if c == b"<":
            # return to the first step in the overall "two step" algorithm
            # reprocessing the < byte
            data.previous()
        else:
            # Read all attributes
            attr = self.getAttribute()
            while attr is not None:
                attr = self.getAttribute()
        return True

    def handleOther(self):
        return self.data.jumpTo(b">")

    def getAttribute(self):
        """Return a name,value pair for the next attribute in the stream,
        if one is found, or None"""
        data = self.data
        # Step 1 (skip chars)
        c = data.skip(spaceCharactersBytes | frozenset([b"/"]))
        assert c is None or len(c) == 1
        # Step 2
        if c in (b">", None):
            return None
        # Step 3
        attrName = []
        attrValue = []
        # Step 4 attribute name
        while True:
            if c == b"=" and attrName:
                break
            elif c in spaceCharactersBytes:
                # Step 6!
                c = data.skip()
                break
            elif c in (b"/", b">"):
                return b"".join(attrName), b""
            elif c in asciiUppercaseBytes:
                attrName.append(c.lower())
            elif c is None:
                return None
            else:
                attrName.append(c)
            # Step 5
            c = next(data)
        # Step 7
        if c != b"=":
            data.previous()
            return b"".join(attrName), b""
        # Step 8
        next(data)
        # Step 9
        c = data.skip()
        # Step 10
        if c in (b"'", b'"'):
            # 10.1
            quoteChar = c
            while True:
                # 10.2
                c = next(data)
                # 10.3
                if c == quoteChar:
                    next(data)
                    return b"".join(attrName), b"".join(attrValue)
                # 10.4
                elif c in asciiUppercaseBytes:
                    attrValue.append(c.lower())
                # 10.5
                else:
                    attrValue.append(c)
        elif c == b">":
            return b"".join(attrName), b""
        elif c in asciiUppercaseBytes:
            attrValue.append(c.lower())
        elif c is None:
            return None
        else:
            attrValue.append(c)
        # Step 11
        while True:
            c = next(data)
            if c in spacesAngleBrackets:
                return b"".join(attrName), b"".join(attrValue)
            elif c in asciiUppercaseBytes:
                attrValue.append(c.lower())
            elif c is None:
                return None
            else:
                attrValue.append(c)


class ContentAttrParser(object):
    def __init__(self, data):
        assert isinstance(data, bytes)
        self.data = data

    def parse(self):
        try:
            # Check if the attr name is charset
            # otherwise return
            self.data.jumpTo(b"charset")
            self.data.position += 1
            self.data.skip()
            if not self.data.currentByte == b"=":
                # If there is no = sign keep looking for attrs
                return None
            self.data.position += 1
            self.data.skip()
            # Look for an encoding between matching quote marks
            if self.data.currentByte in (b'"', b"'"):
                quoteMark = self.data.currentByte
                self.data.position += 1
                oldPosition = self.data.position
                if self.data.jumpTo(quoteMark):
                    return self.data[oldPosition:self.data.position]
                else:
                    return None
            else:
                # Unquoted value
                oldPosition = self.data.position
                try:
                    self.data.skipUntil(spaceCharactersBytes)
                    return self.data[oldPosition:self.data.position]
                except StopIteration:
                    # Return the whole remaining value
                    return self.data[oldPosition:]
        except StopIteration:
            return None


def lookupEncoding(encoding):
    """Return the python codec name corresponding to an encoding or None if the
    string doesn't correspond to a valid encoding."""
    if isinstance(encoding, bytes):
        try:
            encoding = encoding.decode("ascii")
        except UnicodeDecodeError:
            return None

    if encoding is not None:
        try:
            return webencodings.lookup(encoding)
        except AttributeError:
            return None
    else:
        return None
site-packages/pip/_vendor/html5lib/serializer.py000064400000036617150564745010015777 0ustar00from __future__ import absolute_import, division, unicode_literals
from pip._vendor.six import text_type

import re

from codecs import register_error, xmlcharrefreplace_errors

from .constants import voidElements, booleanAttributes, spaceCharacters
from .constants import rcdataElements, entities, xmlEntities
from . import treewalkers, _utils
from xml.sax.saxutils import escape

_quoteAttributeSpecChars = "".join(spaceCharacters) + "\"'=<>`"
_quoteAttributeSpec = re.compile("[" + _quoteAttributeSpecChars + "]")
_quoteAttributeLegacy = re.compile("[" + _quoteAttributeSpecChars +
                                   "\x00\x01\x02\x03\x04\x05\x06\x07\x08\t\n"
                                   "\x0b\x0c\r\x0e\x0f\x10\x11\x12\x13\x14\x15"
                                   "\x16\x17\x18\x19\x1a\x1b\x1c\x1d\x1e\x1f"
                                   "\x20\x2f\x60\xa0\u1680\u180e\u180f\u2000"
                                   "\u2001\u2002\u2003\u2004\u2005\u2006\u2007"
                                   "\u2008\u2009\u200a\u2028\u2029\u202f\u205f"
                                   "\u3000]")


_encode_entity_map = {}
_is_ucs4 = len("\U0010FFFF") == 1
for k, v in list(entities.items()):
    # skip multi-character entities
    if ((_is_ucs4 and len(v) > 1) or
            (not _is_ucs4 and len(v) > 2)):
        continue
    if v != "&":
        if len(v) == 2:
            v = _utils.surrogatePairToCodepoint(v)
        else:
            v = ord(v)
        if v not in _encode_entity_map or k.islower():
            # prefer &lt; over &LT; and similarly for &amp;, &gt;, etc.
            _encode_entity_map[v] = k


def htmlentityreplace_errors(exc):
    if isinstance(exc, (UnicodeEncodeError, UnicodeTranslateError)):
        res = []
        codepoints = []
        skip = False
        for i, c in enumerate(exc.object[exc.start:exc.end]):
            if skip:
                skip = False
                continue
            index = i + exc.start
            if _utils.isSurrogatePair(exc.object[index:min([exc.end, index + 2])]):
                codepoint = _utils.surrogatePairToCodepoint(exc.object[index:index + 2])
                skip = True
            else:
                codepoint = ord(c)
            codepoints.append(codepoint)
        for cp in codepoints:
            e = _encode_entity_map.get(cp)
            if e:
                res.append("&")
                res.append(e)
                if not e.endswith(";"):
                    res.append(";")
            else:
                res.append("&#x%s;" % (hex(cp)[2:]))
        return ("".join(res), exc.end)
    else:
        return xmlcharrefreplace_errors(exc)


register_error("htmlentityreplace", htmlentityreplace_errors)


def serialize(input, tree="etree", encoding=None, **serializer_opts):
    """Serializes the input token stream using the specified treewalker

    :arg input: the token stream to serialize

    :arg tree: the treewalker to use

    :arg encoding: the encoding to use

    :arg serializer_opts: any options to pass to the
        :py:class:`html5lib.serializer.HTMLSerializer` that gets created

    :returns: the tree serialized as a string

    Example:

    >>> from html5lib.html5parser import parse
    >>> from html5lib.serializer import serialize
    >>> token_stream = parse('<html><body><p>Hi!</p></body></html>')
    >>> serialize(token_stream, omit_optional_tags=False)
    '<html><head></head><body><p>Hi!</p></body></html>'

    """
    # XXX: Should we cache this?
    walker = treewalkers.getTreeWalker(tree)
    s = HTMLSerializer(**serializer_opts)
    return s.render(walker(input), encoding)


class HTMLSerializer(object):

    # attribute quoting options
    quote_attr_values = "legacy"  # be secure by default
    quote_char = '"'
    use_best_quote_char = True

    # tag syntax options
    omit_optional_tags = True
    minimize_boolean_attributes = True
    use_trailing_solidus = False
    space_before_trailing_solidus = True

    # escaping options
    escape_lt_in_attrs = False
    escape_rcdata = False
    resolve_entities = True

    # miscellaneous options
    alphabetical_attributes = False
    inject_meta_charset = True
    strip_whitespace = False
    sanitize = False

    options = ("quote_attr_values", "quote_char", "use_best_quote_char",
               "omit_optional_tags", "minimize_boolean_attributes",
               "use_trailing_solidus", "space_before_trailing_solidus",
               "escape_lt_in_attrs", "escape_rcdata", "resolve_entities",
               "alphabetical_attributes", "inject_meta_charset",
               "strip_whitespace", "sanitize")

    def __init__(self, **kwargs):
        """Initialize HTMLSerializer

        :arg inject_meta_charset: Whether or not to inject the meta charset.

            Defaults to ``True``.

        :arg quote_attr_values: Whether to quote attribute values that don't
            require quoting per legacy browser behavior (``"legacy"``), when
            required by the standard (``"spec"``), or always (``"always"``).

            Defaults to ``"legacy"``.

        :arg quote_char: Use given quote character for attribute quoting.

            Defaults to ``"`` which will use double quotes unless attribute
            value contains a double quote, in which case single quotes are
            used.

        :arg escape_lt_in_attrs: Whether or not to escape ``<`` in attribute
            values.

            Defaults to ``False``.

        :arg escape_rcdata: Whether to escape characters that need to be
            escaped within normal elements within rcdata elements such as
            style.

            Defaults to ``False``.

        :arg resolve_entities: Whether to resolve named character entities that
            appear in the source tree. The XML predefined entities &lt; &gt;
            &amp; &quot; &apos; are unaffected by this setting.

            Defaults to ``True``.

        :arg strip_whitespace: Whether to remove semantically meaningless
            whitespace. (This compresses all whitespace to a single space
            except within ``pre``.)

            Defaults to ``False``.

        :arg minimize_boolean_attributes: Shortens boolean attributes to give
            just the attribute value, for example::

              <input disabled="disabled">

            becomes::

              <input disabled>

            Defaults to ``True``.

        :arg use_trailing_solidus: Includes a close-tag slash at the end of the
            start tag of void elements (empty elements whose end tag is
            forbidden). E.g. ``<hr/>``.

            Defaults to ``False``.

        :arg space_before_trailing_solidus: Places a space immediately before
            the closing slash in a tag using a trailing solidus. E.g.
            ``<hr />``. Requires ``use_trailing_solidus=True``.

            Defaults to ``True``.

        :arg sanitize: Strip all unsafe or unknown constructs from output.
            See :py:class:`html5lib.filters.sanitizer.Filter`.

            Defaults to ``False``.

        :arg omit_optional_tags: Omit start/end tags that are optional.

            Defaults to ``True``.

        :arg alphabetical_attributes: Reorder attributes to be in alphabetical order.

            Defaults to ``False``.

        """
        unexpected_args = frozenset(kwargs) - frozenset(self.options)
        if len(unexpected_args) > 0:
            raise TypeError("__init__() got an unexpected keyword argument '%s'" % next(iter(unexpected_args)))
        if 'quote_char' in kwargs:
            self.use_best_quote_char = False
        for attr in self.options:
            setattr(self, attr, kwargs.get(attr, getattr(self, attr)))
        self.errors = []
        self.strict = False

    def encode(self, string):
        assert(isinstance(string, text_type))
        if self.encoding:
            return string.encode(self.encoding, "htmlentityreplace")
        else:
            return string

    def encodeStrict(self, string):
        assert(isinstance(string, text_type))
        if self.encoding:
            return string.encode(self.encoding, "strict")
        else:
            return string

    def serialize(self, treewalker, encoding=None):
        # pylint:disable=too-many-nested-blocks
        self.encoding = encoding
        in_cdata = False
        self.errors = []

        if encoding and self.inject_meta_charset:
            from .filters.inject_meta_charset import Filter
            treewalker = Filter(treewalker, encoding)
        # Alphabetical attributes is here under the assumption that none of
        # the later filters add or change order of attributes; it needs to be
        # before the sanitizer so escaped elements come out correctly
        if self.alphabetical_attributes:
            from .filters.alphabeticalattributes import Filter
            treewalker = Filter(treewalker)
        # WhitespaceFilter should be used before OptionalTagFilter
        # for maximum efficiently of this latter filter
        if self.strip_whitespace:
            from .filters.whitespace import Filter
            treewalker = Filter(treewalker)
        if self.sanitize:
            from .filters.sanitizer import Filter
            treewalker = Filter(treewalker)
        if self.omit_optional_tags:
            from .filters.optionaltags import Filter
            treewalker = Filter(treewalker)

        for token in treewalker:
            type = token["type"]
            if type == "Doctype":
                doctype = "<!DOCTYPE %s" % token["name"]

                if token["publicId"]:
                    doctype += ' PUBLIC "%s"' % token["publicId"]
                elif token["systemId"]:
                    doctype += " SYSTEM"
                if token["systemId"]:
                    if token["systemId"].find('"') >= 0:
                        if token["systemId"].find("'") >= 0:
                            self.serializeError("System identifier contains both single and double quote characters")
                        quote_char = "'"
                    else:
                        quote_char = '"'
                    doctype += " %s%s%s" % (quote_char, token["systemId"], quote_char)

                doctype += ">"
                yield self.encodeStrict(doctype)

            elif type in ("Characters", "SpaceCharacters"):
                if type == "SpaceCharacters" or in_cdata:
                    if in_cdata and token["data"].find("</") >= 0:
                        self.serializeError("Unexpected </ in CDATA")
                    yield self.encode(token["data"])
                else:
                    yield self.encode(escape(token["data"]))

            elif type in ("StartTag", "EmptyTag"):
                name = token["name"]
                yield self.encodeStrict("<%s" % name)
                if name in rcdataElements and not self.escape_rcdata:
                    in_cdata = True
                elif in_cdata:
                    self.serializeError("Unexpected child element of a CDATA element")
                for (_, attr_name), attr_value in token["data"].items():
                    # TODO: Add namespace support here
                    k = attr_name
                    v = attr_value
                    yield self.encodeStrict(' ')

                    yield self.encodeStrict(k)
                    if not self.minimize_boolean_attributes or \
                        (k not in booleanAttributes.get(name, tuple()) and
                         k not in booleanAttributes.get("", tuple())):
                        yield self.encodeStrict("=")
                        if self.quote_attr_values == "always" or len(v) == 0:
                            quote_attr = True
                        elif self.quote_attr_values == "spec":
                            quote_attr = _quoteAttributeSpec.search(v) is not None
                        elif self.quote_attr_values == "legacy":
                            quote_attr = _quoteAttributeLegacy.search(v) is not None
                        else:
                            raise ValueError("quote_attr_values must be one of: "
                                             "'always', 'spec', or 'legacy'")
                        v = v.replace("&", "&amp;")
                        if self.escape_lt_in_attrs:
                            v = v.replace("<", "&lt;")
                        if quote_attr:
                            quote_char = self.quote_char
                            if self.use_best_quote_char:
                                if "'" in v and '"' not in v:
                                    quote_char = '"'
                                elif '"' in v and "'" not in v:
                                    quote_char = "'"
                            if quote_char == "'":
                                v = v.replace("'", "&#39;")
                            else:
                                v = v.replace('"', "&quot;")
                            yield self.encodeStrict(quote_char)
                            yield self.encode(v)
                            yield self.encodeStrict(quote_char)
                        else:
                            yield self.encode(v)
                if name in voidElements and self.use_trailing_solidus:
                    if self.space_before_trailing_solidus:
                        yield self.encodeStrict(" /")
                    else:
                        yield self.encodeStrict("/")
                yield self.encode(">")

            elif type == "EndTag":
                name = token["name"]
                if name in rcdataElements:
                    in_cdata = False
                elif in_cdata:
                    self.serializeError("Unexpected child element of a CDATA element")
                yield self.encodeStrict("</%s>" % name)

            elif type == "Comment":
                data = token["data"]
                if data.find("--") >= 0:
                    self.serializeError("Comment contains --")
                yield self.encodeStrict("<!--%s-->" % token["data"])

            elif type == "Entity":
                name = token["name"]
                key = name + ";"
                if key not in entities:
                    self.serializeError("Entity %s not recognized" % name)
                if self.resolve_entities and key not in xmlEntities:
                    data = entities[key]
                else:
                    data = "&%s;" % name
                yield self.encodeStrict(data)

            else:
                self.serializeError(token["data"])

    def render(self, treewalker, encoding=None):
        """Serializes the stream from the treewalker into a string

        :arg treewalker: the treewalker to serialize

        :arg encoding: the string encoding to use

        :returns: the serialized tree

        Example:

        >>> from html5lib import parse, getTreeWalker
        >>> from html5lib.serializer import HTMLSerializer
        >>> token_stream = parse('<html><body>Hi!</body></html>')
        >>> walker = getTreeWalker('etree')
        >>> serializer = HTMLSerializer(omit_optional_tags=False)
        >>> serializer.render(walker(token_stream))
        '<html><head></head><body>Hi!</body></html>'

        """
        if encoding:
            return b"".join(list(self.serialize(treewalker, encoding)))
        else:
            return "".join(list(self.serialize(treewalker)))

    def serializeError(self, data="XXX ERROR MESSAGE NEEDED"):
        # XXX The idea is to make data mandatory.
        self.errors.append(data)
        if self.strict:
            raise SerializeError


class SerializeError(Exception):
    """Error in serialized tree"""
    pass
site-packages/pip/_vendor/html5lib/_trie/__pycache__/py.cpython-39.pyc000064400000004370150564745010021636 0ustar00a

��Re��@sLddlmZmZmZddlmZddlmZddlm	Z
Gdd�de
�Z	dS)	�)�absolute_import�division�unicode_literals)�	text_type)�bisect_left�)�Triec@sFeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zddd
�Zdd�Z	dS)rcCsJtdd�|��D��std��||_t|���|_d|_dt|�f|_dS)Ncss|]}t|t�VqdS�N)�
isinstancer)�.0�x�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_trie/py.py�	<genexpr>�z Trie.__init__.<locals>.<genexpr>zAll keys must be strings�r)	�all�keys�	TypeError�_data�sorted�_keys�	_cachestr�len�_cachepoints)�self�datar
r
r�__init__
sz
Trie.__init__cCs
||jvSr	�r�r�keyr
r
r�__contains__szTrie.__contains__cCs
t|j�Sr	)rr�rr
r
r�__len__szTrie.__len__cCs
t|j�Sr	)�iterrr"r
r
r�__iter__sz
Trie.__iter__cCs
|j|Sr	rrr
r
r�__getitem__szTrie.__getitem__NcCs�|dus|dks|js t|j�S|�|j�rL|j\}}t|j|||�}}nt|j|�}}t�}|t|j�krt|S|j|�|�r�|�|j|�|d7}qt||_||f|_|S)Nrr)r�set�
startswithrrrr�add)r�prefix�lo�hi�start�irr
r
rrs



z	Trie.keyscCsd||jvrdS|�|j�r6|j\}}t|j|||�}nt|j|�}|t|j�krTdS|j|�|�S)NTF)rr(rrrrr)rr*r+r,r.r
r
r�has_keys_with_prefix6s

zTrie.has_keys_with_prefix)N)
�__name__�
__module__�__qualname__rr!r#r%r&rr/r
r
r
rr	s	
rN)�
__future__rrrZpip._vendor.sixr�bisectr�_baser�ABCTrier
r
r
r�<module>ssite-packages/pip/_vendor/html5lib/_trie/__pycache__/_base.cpython-39.pyc000064400000003143150564745010022254 0ustar00a

��Re��@sXddlmZmZmZzddlmZWneyBddlmZYn0Gdd�de�ZdS)�)�absolute_import�division�unicode_literals)�Mappingcs:eZdZdZd�fdd�	Zdd�Zdd�Zd	d
�Z�ZS)�TriezAbstract base class for triesNcs0tt|���}�durt|�S�fdd�|D�S)Ncsh|]}|���r|�qS�)�
startswith)�.0�x��prefixr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_trie/_base.py�	<setcomp>�zTrie.keys.<locals>.<setcomp>)�superr�keys�set)�selfrr��	__class__rr
rsz	Trie.keyscCs"|��D]}|�|�rdSqdS)NTF)rr)rr�keyrrr
�has_keys_with_prefixs
zTrie.has_keys_with_prefixcCsT||vr|Stdt|�d�D](}|d|�|vr|d|�Sqt|��dS)N�)�range�len�KeyError)rr�irrr
�longest_prefixszTrie.longest_prefixcCs|�|�}|||fS)N)r)rr�lprefixrrr
�longest_prefix_item&s
zTrie.longest_prefix_item)N)	�__name__�
__module__�__qualname__�__doc__rrrr�
__classcell__rrrr
r	s
	
rN)	�
__future__rrr�collections.abcr�ImportError�collectionsrrrrr
�<module>s
site-packages/pip/_vendor/html5lib/_trie/__pycache__/__init__.cpython-39.pyc000064400000000607150564745010022744 0ustar00a

��Rem�@s*ddlmZmZmZddlmZdgZdS)�)�absolute_import�division�unicode_literals�)�TrierN)�
__future__rrr�pyr�__all__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/_trie/__init__.py�<module>ssite-packages/pip/_vendor/html5lib/_trie/py.py000064400000003357150564745010015353 0ustar00from __future__ import absolute_import, division, unicode_literals
from pip._vendor.six import text_type

from bisect import bisect_left

from ._base import Trie as ABCTrie


class Trie(ABCTrie):
    def __init__(self, data):
        if not all(isinstance(x, text_type) for x in data.keys()):
            raise TypeError("All keys must be strings")

        self._data = data
        self._keys = sorted(data.keys())
        self._cachestr = ""
        self._cachepoints = (0, len(data))

    def __contains__(self, key):
        return key in self._data

    def __len__(self):
        return len(self._data)

    def __iter__(self):
        return iter(self._data)

    def __getitem__(self, key):
        return self._data[key]

    def keys(self, prefix=None):
        if prefix is None or prefix == "" or not self._keys:
            return set(self._keys)

        if prefix.startswith(self._cachestr):
            lo, hi = self._cachepoints
            start = i = bisect_left(self._keys, prefix, lo, hi)
        else:
            start = i = bisect_left(self._keys, prefix)

        keys = set()
        if start == len(self._keys):
            return keys

        while self._keys[i].startswith(prefix):
            keys.add(self._keys[i])
            i += 1

        self._cachestr = prefix
        self._cachepoints = (start, i)

        return keys

    def has_keys_with_prefix(self, prefix):
        if prefix in self._data:
            return True

        if prefix.startswith(self._cachestr):
            lo, hi = self._cachepoints
            i = bisect_left(self._keys, prefix, lo, hi)
        else:
            i = bisect_left(self._keys, prefix)

        if i == len(self._keys):
            return False

        return self._keys[i].startswith(prefix)
site-packages/pip/_vendor/html5lib/_trie/_base.py000064400000001765150564745010015775 0ustar00from __future__ import absolute_import, division, unicode_literals

try:
    from collections.abc import Mapping
except ImportError:  # Python 2.7
    from collections import Mapping


class Trie(Mapping):
    """Abstract base class for tries"""

    def keys(self, prefix=None):
        # pylint:disable=arguments-differ
        keys = super(Trie, self).keys()

        if prefix is None:
            return set(keys)

        return {x for x in keys if x.startswith(prefix)}

    def has_keys_with_prefix(self, prefix):
        for key in self.keys():
            if key.startswith(prefix):
                return True

        return False

    def longest_prefix(self, prefix):
        if prefix in self:
            return prefix

        for i in range(1, len(prefix) + 1):
            if prefix[:-i] in self:
                return prefix[:-i]

        raise KeyError(prefix)

    def longest_prefix_item(self, prefix):
        lprefix = self.longest_prefix(prefix)
        return (lprefix, self[lprefix])
site-packages/pip/_vendor/html5lib/_trie/__init__.py000064400000000155150564745010016453 0ustar00from __future__ import absolute_import, division, unicode_literals

from .py import Trie

__all__ = ["Trie"]
site-packages/pip/_vendor/html5lib/filters/__pycache__/optionaltags.cpython-39.pyc000064400000005362150564745010024262 0ustar00a

��Re\)�@s6ddlmZmZmZddlmZGdd�dej�ZdS)�)�absolute_import�division�unicode_literals�)�basec@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)�Filterz+Removes optional tags from the token streamccsHd}}|jD] }|dur&|||fV|}|}q|durD||dfVdS)N)�source)�selfZ	previous1Z	previous2�token�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/optionaltags.py�sliders
z
Filter.sliderccsp|��D]b\}}}|d}|dkrD|ds<|�|d||�sj|Vq|dkrd|�|d|�sj|Vq|VqdS)N�type�StartTag�data�name�EndTag)r
�is_optional_start�is_optional_end)r	�previousr
�nextrrrr�__iter__s�zFilter.__iter__cCs�|r|dpd}|dvr |dvS|dkrJ|dvr4dS|dkr�|ddkSn�|d	krx|dvr^d
S|dkrr|ddvSdSnb|d
kr�|dvr�|ddkSd
Sn@|dkr�|dkr�|r�|ddkr�|ddvr�d
S|ddkSd
Sd
S)Nr�html��Comment�SpaceCharacters�head�r�EmptyTagTrr�bodyFr)�script�style�colgroup�col�tbody)r$�thead�tfoot�trr)r	�tagnamerrrrrrrs6
�zFilter.is_optional_startcCs|r|dpd}|dvr |dvS|dvrP|dkr<|d|kS|dkpJ|duS�n�|dvr�|dkrl|ddvS|d	kr�|dkp�|duSd
S�np|dkr�|dvr�|dd
vS|dkp�|duS�n@|dkr�|dkr�|ddvS|dkp�|duS�n|dv�r |dk�r|ddvS|dk�p|duSn�|dk�rT|dv�r8d
S|dk�rN|ddkSdSn�|dv�r�|dk�rt|ddvS|dk�r�|dk�p�|duSd
Snf|dk�r�|dk�r�|ddkS|dk�p�|duSn2|dv�r�|dk�r�|ddvS|dk�p�|duSd
S)Nr)rrrr)�li�optgroupr'rrr)�dt�ddr,F�pr)�address�article�aside�
blockquote�datagrid�dialog�dir�div�dl�fieldset�footer�form�h1�h2�h3�h4�h5�h6�header�hr�menu�nav�olr-�pre�section�table�ul�option)rIr*)�rt�rpr"T)r%r$)r$r&r$r&)�td�thr)r	r(rrrrrrXs\












zFilter.is_optional_endN)�__name__�
__module__�__qualname__�__doc__r
rrrrrrrrs


9rN)�
__future__rrr�rrrrrr�<module>ssite-packages/pip/_vendor/html5lib/filters/__pycache__/whitespace.cpython-39.pyc000064400000002577150564745010023717 0ustar00a

��Re��@snddlmZmZmZddlZddlmZddlmZm	Z	d�
e	�Z	e�de	�ZGd	d
�d
ej
�Z
dd�ZdS)
�)�absolute_import�division�unicode_literalsN�)�base�)�rcdataElements�spaceCharacters�z[%s]+c@s,eZdZdZeddgee��Zdd�ZdS)�FilterzACollapses whitespace except in pre, textarea, and script elements�pre�textareaccs�d}tj�|�D]�}|d}|dkr@|s6|d|jvr@|d7}nP|dkrV|rV|d8}n:|st|dkrt|drtd	|d<n|s�|d
kr�t|d�|d<|VqdS)Nr�type�StartTag�namer�EndTag�SpaceCharacters�data� �
Characters)rr�__iter__�spacePreserveElements�collapse_spaces)�self�preserve�tokenr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/whitespace.pyrs ��


zFilter.__iter__N)	�__name__�
__module__�__qualname__�__doc__�	frozenset�listrrrrrrrrsrcCst�d|�S)Nr)�SPACES_REGEX�sub)�textrrrr%sr)�
__future__rrr�rer
r�	constantsrr	�join�compiler$rrrrrr�<module>s
site-packages/pip/_vendor/html5lib/filters/__pycache__/sanitizer.cpython-39.pyc000064400000041035150564745010023563 0ustar00a

��Rei�D@s^dZddlmZmZmZddlZddlZddlmZm	Z	ddl
mZddl
mZdd	lmZmZd
gZdZe�ee�eedd
feddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfeddfedd fedd!fedd"fedd#fedd$fedd%fedd&fedd'fedd(fedd)fedd*fedd+fedd,fedd-fedd.fedd/fedd0fedd1fedd2fedd3fedd4fedd5fedd6fedd7fedd8fedd9fedd:fedd;fedd<fedd=fedd>fedd?fedd@feddAfeddBfeddCfeddDfeddEfeddFfeddGfeddHfeddIfeddJfeddKfeddLfeddMfeddNfeddOfeddPfeddQfeddRfeddSfeddTfeddUfeddVfeddWfeddXfeddYfeddZfedd[fedd\fedd]fedd^fedd_fedd`feddafeddbfeddcfedddfeddefeddffeddgfeddhfeddifeddjfeddkfeddlfeddmfeddnfeddofedpdqfedpdrfedpdsfedpdtfedpdufedpdvfedpdwfedpdxfedpdyfedpdzfedpd{fedpd|fedpd}fedpd~fedpdfedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fedpd�fed�d
fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�fed�d�ff��Zed�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d��d�d�d�d�d�d�d�d�d�d	�d
�d�d�d
�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d �d!�d"�d#�d$�d%�d&�d'�d(�d)�d*�d+�d,�d-�d.�d/�d0�d1�d2�d3�d4�d5�d6�d7�d8�d9e�d:�d;f�d<d��d=�d=�d=�d>�d?�d@�dA�dB�dC�dD�dE�dF�dG�dHd�d�dI�dJ�dK�dL�dM�dM�dN�dO�dP�dQ�dQ�dQ�dR�dS�d�dT�dU�dV�dW�dX�d8�d8e�dY�dZfe�dY�d[fe�dY�d\f�d]�d^�d_�d`�da�db�dc�dd�de�df�dg�dh�di�djdːdkd͐dl�dm�dn�do�dp�dq�dr�ds�dB�dtd�du�dv�dw�dx�dy�dz�d{�d|�d}�d~�d�d��d��d��d��d��d�d�d��d�d�d��d��d��d��d�d��d��d��d��d��d��d��d�d��d�d�d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d��d*�d��d��d��d.�d��d��d��d��d��ddÐdĐdŐdƐdǐd8�dȐdɐdʐdːd�e�dY�d�fe�dY�d�fe�dY�dZfe�dY�d�fe�dY�d[fe�dYd�fe�dY�d\fe�d:�d�fe�d:�d;fe�d:�d�f�dҐdӐdԐdՐfC�Zed�d#d�d�d��dd�d�d�d��de�dY�dZfe�d:�d�ff
�Ze�dփZe�d׃Ze�d؃Ze�dكZe�dڃZe�dۃZe�d܃Z e�!�d�ej"�Z#G�d�d
�d
ej$�Z$dS(�a2Deprecated from html5lib 1.1.

See `here <https://github.com/html5lib/html5lib-python/issues/443>`_ for
information about its deprecation; `Bleach <https://github.com/mozilla/bleach>`_
is recommended as a replacement. Please let us know in the aforementioned issue
if Bleach is unsuitable for your needs.

�)�absolute_import�division�unicode_literalsN)�escape�unescape)�urllib_parse�)�base�)�
namespaces�prefixes�Filterz�html5lib's sanitizer is deprecated; see https://github.com/html5lib/html5lib-python/issues/443 and please let us know if Bleach is unsuitable for your needs�html�a�abbrZacronym�address�area�article�aside�audio�b�big�
blockquote�br�buttonZcanvas�caption�center�cite�code�col�colgroup�command�datagridZdatalist�dd�del�details�dfn�dialog�dir�div�dl�dt�emzevent-source�fieldset�
figcaption�figure�footer�font�form�header�h1�h2�h3�h4�h5�h6�hr�i�img�inputZins�keygenZkbd�labelZlegend�li�m�map�menuZmeterZmulticol�navZnextid�ol�output�optgroup�option�p�pre�progress�q�sZsamp�section�select�smallZsound�sourceZspacer�span�strike�strong�sub�sup�table�tbody�td�textarea�time�tfoot�th�thead�tr�tt�u�ul�var�video�mathmlZmaction�mathZmerrorZmfrac�miZ
mmultiscripts�mn�moZmoverZmpaddedZmphantomZmprescriptsZmrootZmrowZmspaceZmsqrtZmstyleZmsubZmsubsupZmsupZmtableZmtd�mtextZmtrZmunderZ
munderover�none�svg�animate�animateColor�
animateMotion�animateTransform�clipPathZcircleZdefs�descZellipsez	font-facezfont-face-namez
font-face-src�gZglyphZhkern�linearGradient�line�marker�metadataz
missing-glyphZmpath�pathZpolygonZpolyline�radialGradientZrect�set�stop�switch�text�titleZtspan�use)Nr)N�accept)Nzaccept-charset)NZ	accesskey)N�action)NZalign)NZalt)N�autocomplete)N�	autofocus)NZaxis)N�
background)NZbalance)NZbgcolor)NZbgproperties)N�border)NZbordercolor)NZbordercolordark)NZbordercolorlight)NZ
bottompadding)NZcellpadding)NZcellspacing)N�ch)N�	challenge)N�char)NZcharoff)NZchoff)N�charset)N�checked)Nr)N�class)N�clear)N�color)N�cols)NZcolspan)N�compact)NZcontenteditable)N�controls)NZcoords)N�data)NZdatafld)NZdatapagesize)NZdatasrc)N�datetime)N�default)N�delay)Nr()N�disabled)NZ	draggable)NZdynsrc)NZenctype)N�end)N�face)N�for)Nr2)N�frame)NZ
galleryimg)NZgutter)N�headers)N�height)NZ	hidefocus)N�hidden)N�high)N�href)NZhreflang)NZhspace)NZicon)N�id)NZ	inputmode)N�ismap)NZkeytype)Nr?)NZleftspacing)N�lang)N�list)NZlongdesc)N�loop)NZ	loopcount)NZloopend)NZ	loopstart)N�low)NZlowsrc)N�max)NZ	maxlength)NZmedia)N�method)N�min)N�multiple)N�name)NZnohref)N�noshade)N�nowrap)N�open)NZoptimum�N�pattern)NZping)Nz
point-size)NZposter)NZpqg)NZpreload)N�prompt)NZ
radiogroup)N�readonly)N�rel)Nz
repeat-max)Nz
repeat-min)N�replace)N�required)N�rev)NZrightspacing)N�rows)NZrowspan)N�rules)NZscope)N�selected)N�shape)N�size)NrR)N�src)N�start)N�step�N�style)N�summary)N�suppress)NZtabindex)N�target)N�template)Nr~)NZ
toppadding)N�type)NZunselectable)NZusemap)N�urn)NZvalign)N�value)N�variable)N�volume)NZvspace)NZvrml)N�width)N�wrap�xmlr�)NZ
actiontype)NZcolumnalign)NZcolumnlines)NZ
columnspacing)NZ
columnspan)N�depth)N�display)NZdisplaystyle)NZequalcolumns)NZ	equalrows)NZfence)NZ	fontstyle)NZ
fontweight)NZ
linethickness)NZlspace)NZmathbackground)NZ	mathcolor)NZmathvariant)N�maxsize)NZminsize)N�other)NZrowalign)NZrowlines)NZ
rowspacing)NZrspace)NZscriptlevel)NZ	selection)N�	separator)NZstretchy�xlinkr��showr�)Nz
accent-height)N�
accumulate)NZadditive)NZ
alphabetic)Nzarabic-form)NZascent)N�
attributeName)N�
attributeType)N�baseProfile)NZbbox)N�begin)NZby)N�calcMode)Nz
cap-height�Nz	clip-path)Nzcolor-rendering)N�content)NZcx)N�cy)N�d)NZdx)NZdy)NZdescent)NZdur�N�fill)N�fill-opacity)N�	fill-rule)N�font-family)N�	font-size)Nzfont-stretch)N�
font-style)N�font-variant)N�font-weight)N�from)NZfx)NZfy)NZg1)NZg2)Nz
glyph-name)N�
gradientUnits)NZhanging)Nzhoriz-adv-x)Nzhoriz-origin-x)NZideographic)N�k)N�	keyPoints)N�
keySplines)N�keyTimes�Nz
marker-end�Nz
marker-mid�Nzmarker-start)N�markerHeight)N�markerUnits)N�markerWidth)NZmathematical)N�offset)NZopacity)NZorient)N�origin)Nzoverline-position)Nzoverline-thickness)Nzpanose-1)Nrx)N�
pathLength)NZpoints)N�preserveAspectRatio)N�r)N�refX)N�refY)N�repeatCount)N�	repeatDur)N�requiredExtensions)N�requiredFeatures)NZrestart)N�rotate)N�rx)NZry)NZslope)NZstemh)NZstemv)Nz
stop-color)Nzstop-opacity)Nzstrikethrough-position)Nzstrikethrough-thickness�N�stroke)Nzstroke-dasharray)Nzstroke-dashoffset)N�stroke-linecap)N�stroke-linejoin)Nzstroke-miterlimit)N�stroke-opacity)N�stroke-width)N�systemLanguage)Nztext-anchor)N�to)NZ	transform)N�u1)N�u2)Nzunderline-position)Nzunderline-thickness)N�unicode)Nz
unicode-range)Nzunits-per-em)N�values)N�version)N�viewBox)NZ
visibility)NZwidths)N�x)Nzx-height)N�x1)NZx2�actuate�arcrole�roler	�space)N�y)N�y1)N�y2)N�
zoomAndPan)r�)Nz
color-profile�N�cursorr��N�filter)Nrvr�r�r�)N�maskr�))N�altGlyph)Nrm)Nrn)Nro)Nrpr)N�feImager)Nrtr�)Nry)N�textpath)NZtref)Nrz)Nr).Zazimuthzbackground-colorzborder-bottom-colorzborder-collapsezborder-colorzborder-left-colorzborder-right-colorzborder-top-colorr�r�r�	directionr�Z	elevation�floatr1r�r�r�r�r�r�zletter-spacingzline-heightZoverflow�pausezpause-afterzpause-beforeZpitchzpitch-rangeZrichnessZspeakzspeak-headerz
speak-numeralzspeak-punctuationzspeech-rateZstressz
text-alignztext-decorationztext-indentzunicode-bidizvertical-alignzvoice-familyr�zwhite-spacer�)'�autoZaqua�black�block�blue�boldZbothZbottomZbrownrZcollapseZdashedZdottedZfuchsiaZgray�greenz
!important�italic�leftZlimeZmaroonZmediumrkZnavy�normalr�Zolive�pointerZpurple�red�rightZsolidZsilverZteal�topZtransparent�	underline�white�yellow)r�r�r�r�rrrr)Zed2k�ftp�http�httpsZirc�mailto�news�gopher�nntp�telnetZwebcalZxmppZcallto�feedr�Zaim�rsync�tag�ssh�sftp�rtsp�afsr�)z	image/pngz
image/jpegz	image/gifz
image/webpz	image/bmpz
text/plainaL
                                ^
                                # Match a content type <application>/<type>
                                (?P<content_type>[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+)
                                # Match any character set and encoding
                                (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?)
                                  |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?)
                                # Assume the rest is data
                                ,.*
                                $
                                c
s^eZdZdZeeeeee	e
eee
f
�fdd�	Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z�ZS)r
zISanitizes token stream of XHTML+MathML+SVG and of inline style attributescs\tt|��|�t�tt�||_||_||_	||_
||_||_||_
|	|_|
|_||_dS)aSCreates a Filter

        :arg allowed_elements: set of elements to allow--everything else will
            be escaped

        :arg allowed_attributes: set of attributes to allow in
            elements--everything else will be stripped

        :arg allowed_css_properties: set of CSS properties to allow--everything
            else will be stripped

        :arg allowed_css_keywords: set of CSS keywords to allow--everything
            else will be stripped

        :arg allowed_svg_properties: set of SVG properties to allow--everything
            else will be removed

        :arg allowed_protocols: set of allowed protocols for URIs

        :arg allowed_content_types: set of allowed content types for ``data`` URIs.

        :arg attr_val_is_uri: set of attributes that have URI values--values
            that have a scheme not listed in ``allowed_protocols`` are removed

        :arg svg_attr_val_allows_ref: set of SVG attributes that can have
            references

        :arg svg_allow_local_href: set of SVG elements that can have local
            hrefs--these are removed

        N)�superr
�__init__�warnings�warn�_deprecation_msg�DeprecationWarning�allowed_elements�allowed_attributes�allowed_css_properties�allowed_css_keywords�allowed_svg_properties�allowed_protocols�allowed_content_types�attr_val_is_uri�svg_attr_val_allows_ref�svg_allow_local_href)�selfrQrFrGrHrIrJrKrLrMrNrO��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/sanitizer.pyrA�s+zFilter.__init__ccs*tj�|�D]}|�|�}|r|VqdS)N)r	r
�__iter__�sanitize_token)rP�tokenrSrSrTrUs
zFilter.__iter__cCsp|d}|dvr^|d}|d}||f|jvsH|durRtd|f|jvrR|�|�S|�|�Sn|dkrhn|SdS)Nr�)�StartTag�EndTag�EmptyTagr��	namespacer�Comment)rFr�
allowed_token�disallowed_token)rPrW�
token_typer�r[rSrSrTrV!s��
zFilter.sanitize_tokenc		Cs�d|v�r�|d}t|���}||jD]}|d|=|�|�q(||j@D]�}||vs\J�t�ddt||����}|�	dd�}zt
�
|�}Wnty�d}||=Yn0|rL|jrL|j|j
vr�||=|jdkrLt�|j�}|s�||=qL|�d�|jvrL||=qL|jD]*}||v�r
t�ddt||��||<�q
|d|jv�r�td	d
f|v�r�t�d|td	d
f��r�|td	d
f=d|v�r�|�|d�|d<||d<|S)
Nr�u
[`- - \s]+�u��content_typezurl\s*\(\s*[^#\s][^)]+?\)� r�r�r�z^\s*[^#\s].*r�)rz�keysrG�removerM�rerUr�lowerr��urlparse�
ValueError�schemerK�data_content_type�matchrx�grouprLrNrOr�search�sanitize_css)	rPrW�attrsZ
attr_names�	to_remove�attrZ
val_unescaped�urirArSrSrTr]3sX


�




����
zFilter.allowed_tokencCs�|d}|dkr"d|d|d<n�|dr�|dvs6J�g}|d��D]:\\}}}|�d|durd|ndt||ft|�f�qFd	|dd
�|�f|d<nd|d|d<|�d�r�|ddd
�d|d<d|d<|d=|S)Nr�rYz</%s>r�r�)rXrZz %s="%s"z%s:%sz<%s%s>r`z<%s>�selfClosing���z/>�
Characters)�items�appendrr�join�get)rPrWr_ro�nsr��vrSrSrTr^es0
zFilter.disallowed_tokencCs�t�d��d|�}t�d|�s"dSt�d|�s2dSg}t�d|�D]�\}}|sPqB|��|jvrv|�|d|d�qB|�d	�d
��dvr�|��D]}||j	vr�t�d|�s�q�q�|�|d|d�qB|��|j
vrB|�|d|d�qBd�|�S)
Nzurl\s*\(\s*[^\s)]+?\s*\)\s*rbz@^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$r`z ^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$z([-\w]+)\s*:\s*([^:;]*)z: �;�-r)r�r��margin�paddingz_^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$)re�compilerUrk�findallrfrHrw�splitrIrJrx)rPr��clean�propr��keywordrSrSrTrnys*

�zFilter.sanitize_css)�__name__�
__module__�__qualname__�__doc__rFrGrHrIrJrKrLrMrNrOrArUrVr]r^rn�
__classcell__rSrSrQrTr
�s"�:2)%r��
__future__rrrrerB�xml.sax.saxutilsrr�pip._vendor.six.movesrrgr`r	�	constantsrr�__all__rDrCrE�	frozensetrFrGrMrNrOrHrIrJrKrLr��VERBOSErjr
rSrSrSrT�<module>s&�
































































































































































��%���K�


1
*



�site-packages/pip/_vendor/html5lib/filters/__pycache__/base.cpython-39.pyc000064400000001615150564745010022465 0ustar00a

��Re�@s(ddlmZmZmZGdd�de�ZdS)�)�absolute_import�division�unicode_literalsc@s$eZdZdd�Zdd�Zdd�ZdS)�FiltercCs
||_dS�N)�source)�selfr�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/base.py�__init__szFilter.__init__cCs
t|j�Sr)�iterr)rr	r	r
�__iter__szFilter.__iter__cCst|j|�Sr)�getattrr)r�namer	r	r
�__getattr__szFilter.__getattr__N)�__name__�
__module__�__qualname__rr
rr	r	r	r
rsrN)�
__future__rrr�objectrr	r	r	r
�<module>ssite-packages/pip/_vendor/html5lib/filters/__pycache__/inject_meta_charset.cpython-39.pyc000064400000003573150564745010025553 0ustar00a

��Re��@s6ddlmZmZmZddlmZGdd�dej�ZdS)�)�absolute_import�division�unicode_literals�)�basec@s eZdZdZdd�Zdd�ZdS)�Filterz=Injects ``<meta charset=ENCODING>`` tag into head of documentcCstj�||�||_dS)ztCreates a Filter

        :arg source: the source token stream

        :arg encoding: the encoding to set

        N)rr�__init__�encoding)�self�sourcer	�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/inject_meta_charset.pyrszFilter.__init__c
cs�d}|jdu}g}tj�|�D�]�}|d}|dkrL|d��dkrHd}�nd|dk�rJ|d��dkr�d	}|d
��D]X\\}}}	|dur�qvqv|��dkr�|j|d
||f<d}q�qv|d
krv|	��dkrvd}qv|r�d|d
vr�d|j|d
d<d}nP|d��dk�r�|�s�dd|d
d�Vddd|jid�Vddd�Vd}qnf|dk�r�|d��dk�r�|�r�|�d�V|�s�ddd|jid�V|�r�|�d�V�q�d}d}|dk�r�|�|�q|VqdS)NZpre_head�type�StartTag�name�headZin_head�EmptyTag�metaF�data�charsetTz
http-equivzcontent-type)N�contentztext/html; charset=%s)rrr)Nr�EndTag)rrrZ	post_head)r	rr�__iter__�lower�items�pop�append)
r
�stateZ
meta_found�pending�tokenrZhas_http_equiv_content_type�	namespacer�valuerrr
rs^

��
�
zFilter.__iter__N)�__name__�
__module__�__qualname__�__doc__rrrrrr
rsrN)�
__future__rrr�rrrrrr
�<module>ssite-packages/pip/_vendor/html5lib/filters/__pycache__/lint.cpython-39.pyc000064400000005141150564745010022517 0ustar00a

��Re;�@shddlmZmZmZddlmZddlmZddlm	Z	m
Z
ddlmZd�e�ZGd	d
�d
ej
�Z
dS)�)�absolute_import�division�unicode_literals)�	text_type�)�base�)�
namespaces�voidElements)�spaceCharacters�cs*eZdZdZd�fdd�	Zdd�Z�ZS)�FilterzgLints the token stream for errors

    If it finds any errors, it'll raise an ``AssertionError``.

    Tcstt|��|�||_dS)z�Creates a Filter

        :arg source: the source token stream

        :arg require_matching_tags: whether or not to require matching tags

        N)�superr
�__init__�require_matching_tags)�self�sourcer��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/lint.pyrszFilter.__init__c	csDg}tj�|�D�],}|d}|dv�rF|d}|d}|dusNt|t�sNJ�|dksZJ�t|t�shJ�|dkstJ�t|dt�s�J�|r�|tdkr�|tvr�|dks�J�n|d	ks�J�|d	kr�|jr�|�	||f�|d�
�D]`\\}}}|du�st|t��sJ�|dk�sJ�t|t��s$J�|dk�s2J�t|t�s�J�q�n�|d
k�r|d}|d}|du�szt|t��szJ�|dk�s�J�t|t��s�J�|dk�s�J�|�r�|tdk�r�|tv�r�d�s�Jdd
|i��n"|j�r8|��}|||fk�s8J��n6|dk�r(|d}t|t��s8J��n|dv�rx|d}t|t��sJJ�|dk�sXJ�|dk�r8|�t
�dk�s8J�n�|dk�r�|d}|du�s�t|t��s�J�|ddu�s�t|t��s�J�|ddu�s8t|t��s8J�nV|dk�rt|dt��s8J�n6|dk�r"t|dt��s8J�nd�s8Jdd|i��|VqdS)N�type)�StartTag�EmptyTag�	namespace�namer�data�htmlrr�EndTagFz.Void element reported as EndTag token: %(tag)s�tag�Comment)�
Characters�SpaceCharactersr"�Doctype�publicId�systemId�EntityZSerializerErrorzUnknown token type: %(type)s)rr
�__iter__�
isinstancer�dictr	r
r�append�items�pop�stripr)	rZ
open_elements�tokenrrr�value�startrrrrr'sl





 

zFilter.__iter__)T)�__name__�
__module__�__qualname__�__doc__rr'�
__classcell__rrrrr
sr
N)�
__future__rrrZpip._vendor.sixrrr�	constantsr	r
r�joinr
rrrr�<module>s
site-packages/pip/_vendor/html5lib/filters/__pycache__/__init__.cpython-39.pyc000064400000000357150564745010023314 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/__init__.py�<module>�site-packages/pip/_vendor/html5lib/filters/__pycache__/alphabeticalattributes.cpython-39.pyc000064400000002521150564745010026270 0ustar00a

��Re��@sJddlmZmZmZddlmZddlmZdd�ZGdd�dej	�Z	d	S)
�)�absolute_import�division�unicode_literals�)�base)�OrderedDictcCs|ddpd|ddfS)z�Return an appropriate key for an attribute for sorting

    Attributes have a namespace that can be either ``None`` or a string. We
    can't compare the two because they're different types, so we convert
    ``None`` to an empty string first.

    r�r�)�attrr	r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/filters/alphabeticalattributes.py�	_attr_keysrc@seZdZdZdd�ZdS)�Filterz$Alphabetizes attributes for elementsccs\tj�|�D]J}|ddvrPt�}t|d��td�D]\}}|||<q6||d<|VqdS)N�type)�StartTag�EmptyTag�data)�key)rr
�__iter__r�sorted�itemsr)�self�token�attrs�name�valuer	r	rrs�
zFilter.__iter__N)�__name__�
__module__�__qualname__�__doc__rr	r	r	rr
sr
N)
�
__future__rrrrr�collectionsrrr
r	r	r	r�<module>ssite-packages/pip/_vendor/html5lib/filters/lint.py000064400000007073150564745010016236 0ustar00from __future__ import absolute_import, division, unicode_literals

from pip._vendor.six import text_type

from . import base
from ..constants import namespaces, voidElements

from ..constants import spaceCharacters
spaceCharacters = "".join(spaceCharacters)


class Filter(base.Filter):
    """Lints the token stream for errors

    If it finds any errors, it'll raise an ``AssertionError``.

    """
    def __init__(self, source, require_matching_tags=True):
        """Creates a Filter

        :arg source: the source token stream

        :arg require_matching_tags: whether or not to require matching tags

        """
        super(Filter, self).__init__(source)
        self.require_matching_tags = require_matching_tags

    def __iter__(self):
        open_elements = []
        for token in base.Filter.__iter__(self):
            type = token["type"]
            if type in ("StartTag", "EmptyTag"):
                namespace = token["namespace"]
                name = token["name"]
                assert namespace is None or isinstance(namespace, text_type)
                assert namespace != ""
                assert isinstance(name, text_type)
                assert name != ""
                assert isinstance(token["data"], dict)
                if (not namespace or namespace == namespaces["html"]) and name in voidElements:
                    assert type == "EmptyTag"
                else:
                    assert type == "StartTag"
                if type == "StartTag" and self.require_matching_tags:
                    open_elements.append((namespace, name))
                for (namespace, name), value in token["data"].items():
                    assert namespace is None or isinstance(namespace, text_type)
                    assert namespace != ""
                    assert isinstance(name, text_type)
                    assert name != ""
                    assert isinstance(value, text_type)

            elif type == "EndTag":
                namespace = token["namespace"]
                name = token["name"]
                assert namespace is None or isinstance(namespace, text_type)
                assert namespace != ""
                assert isinstance(name, text_type)
                assert name != ""
                if (not namespace or namespace == namespaces["html"]) and name in voidElements:
                    assert False, "Void element reported as EndTag token: %(tag)s" % {"tag": name}
                elif self.require_matching_tags:
                    start = open_elements.pop()
                    assert start == (namespace, name)

            elif type == "Comment":
                data = token["data"]
                assert isinstance(data, text_type)

            elif type in ("Characters", "SpaceCharacters"):
                data = token["data"]
                assert isinstance(data, text_type)
                assert data != ""
                if type == "SpaceCharacters":
                    assert data.strip(spaceCharacters) == ""

            elif type == "Doctype":
                name = token["name"]
                assert name is None or isinstance(name, text_type)
                assert token["publicId"] is None or isinstance(name, text_type)
                assert token["systemId"] is None or isinstance(name, text_type)

            elif type == "Entity":
                assert isinstance(token["name"], text_type)

            elif type == "SerializerError":
                assert isinstance(token["data"], text_type)

            else:
                assert False, "Unknown token type: %(type)s" % {"type": type}

            yield token
site-packages/pip/_vendor/html5lib/filters/base.py000064400000000436150564745010016176 0ustar00from __future__ import absolute_import, division, unicode_literals


class Filter(object):
    def __init__(self, source):
        self.source = source

    def __iter__(self):
        return iter(self.source)

    def __getattr__(self, name):
        return getattr(self.source, name)
site-packages/pip/_vendor/html5lib/filters/sanitizer.py000064400000064421150564745010017300 0ustar00"""Deprecated from html5lib 1.1.

See `here <https://github.com/html5lib/html5lib-python/issues/443>`_ for
information about its deprecation; `Bleach <https://github.com/mozilla/bleach>`_
is recommended as a replacement. Please let us know in the aforementioned issue
if Bleach is unsuitable for your needs.

"""
from __future__ import absolute_import, division, unicode_literals

import re
import warnings
from xml.sax.saxutils import escape, unescape

from pip._vendor.six.moves import urllib_parse as urlparse

from . import base
from ..constants import namespaces, prefixes

__all__ = ["Filter"]


_deprecation_msg = (
    "html5lib's sanitizer is deprecated; see " +
    "https://github.com/html5lib/html5lib-python/issues/443 and please let " +
    "us know if Bleach is unsuitable for your needs"
)

warnings.warn(_deprecation_msg, DeprecationWarning)

allowed_elements = frozenset((
    (namespaces['html'], 'a'),
    (namespaces['html'], 'abbr'),
    (namespaces['html'], 'acronym'),
    (namespaces['html'], 'address'),
    (namespaces['html'], 'area'),
    (namespaces['html'], 'article'),
    (namespaces['html'], 'aside'),
    (namespaces['html'], 'audio'),
    (namespaces['html'], 'b'),
    (namespaces['html'], 'big'),
    (namespaces['html'], 'blockquote'),
    (namespaces['html'], 'br'),
    (namespaces['html'], 'button'),
    (namespaces['html'], 'canvas'),
    (namespaces['html'], 'caption'),
    (namespaces['html'], 'center'),
    (namespaces['html'], 'cite'),
    (namespaces['html'], 'code'),
    (namespaces['html'], 'col'),
    (namespaces['html'], 'colgroup'),
    (namespaces['html'], 'command'),
    (namespaces['html'], 'datagrid'),
    (namespaces['html'], 'datalist'),
    (namespaces['html'], 'dd'),
    (namespaces['html'], 'del'),
    (namespaces['html'], 'details'),
    (namespaces['html'], 'dfn'),
    (namespaces['html'], 'dialog'),
    (namespaces['html'], 'dir'),
    (namespaces['html'], 'div'),
    (namespaces['html'], 'dl'),
    (namespaces['html'], 'dt'),
    (namespaces['html'], 'em'),
    (namespaces['html'], 'event-source'),
    (namespaces['html'], 'fieldset'),
    (namespaces['html'], 'figcaption'),
    (namespaces['html'], 'figure'),
    (namespaces['html'], 'footer'),
    (namespaces['html'], 'font'),
    (namespaces['html'], 'form'),
    (namespaces['html'], 'header'),
    (namespaces['html'], 'h1'),
    (namespaces['html'], 'h2'),
    (namespaces['html'], 'h3'),
    (namespaces['html'], 'h4'),
    (namespaces['html'], 'h5'),
    (namespaces['html'], 'h6'),
    (namespaces['html'], 'hr'),
    (namespaces['html'], 'i'),
    (namespaces['html'], 'img'),
    (namespaces['html'], 'input'),
    (namespaces['html'], 'ins'),
    (namespaces['html'], 'keygen'),
    (namespaces['html'], 'kbd'),
    (namespaces['html'], 'label'),
    (namespaces['html'], 'legend'),
    (namespaces['html'], 'li'),
    (namespaces['html'], 'm'),
    (namespaces['html'], 'map'),
    (namespaces['html'], 'menu'),
    (namespaces['html'], 'meter'),
    (namespaces['html'], 'multicol'),
    (namespaces['html'], 'nav'),
    (namespaces['html'], 'nextid'),
    (namespaces['html'], 'ol'),
    (namespaces['html'], 'output'),
    (namespaces['html'], 'optgroup'),
    (namespaces['html'], 'option'),
    (namespaces['html'], 'p'),
    (namespaces['html'], 'pre'),
    (namespaces['html'], 'progress'),
    (namespaces['html'], 'q'),
    (namespaces['html'], 's'),
    (namespaces['html'], 'samp'),
    (namespaces['html'], 'section'),
    (namespaces['html'], 'select'),
    (namespaces['html'], 'small'),
    (namespaces['html'], 'sound'),
    (namespaces['html'], 'source'),
    (namespaces['html'], 'spacer'),
    (namespaces['html'], 'span'),
    (namespaces['html'], 'strike'),
    (namespaces['html'], 'strong'),
    (namespaces['html'], 'sub'),
    (namespaces['html'], 'sup'),
    (namespaces['html'], 'table'),
    (namespaces['html'], 'tbody'),
    (namespaces['html'], 'td'),
    (namespaces['html'], 'textarea'),
    (namespaces['html'], 'time'),
    (namespaces['html'], 'tfoot'),
    (namespaces['html'], 'th'),
    (namespaces['html'], 'thead'),
    (namespaces['html'], 'tr'),
    (namespaces['html'], 'tt'),
    (namespaces['html'], 'u'),
    (namespaces['html'], 'ul'),
    (namespaces['html'], 'var'),
    (namespaces['html'], 'video'),
    (namespaces['mathml'], 'maction'),
    (namespaces['mathml'], 'math'),
    (namespaces['mathml'], 'merror'),
    (namespaces['mathml'], 'mfrac'),
    (namespaces['mathml'], 'mi'),
    (namespaces['mathml'], 'mmultiscripts'),
    (namespaces['mathml'], 'mn'),
    (namespaces['mathml'], 'mo'),
    (namespaces['mathml'], 'mover'),
    (namespaces['mathml'], 'mpadded'),
    (namespaces['mathml'], 'mphantom'),
    (namespaces['mathml'], 'mprescripts'),
    (namespaces['mathml'], 'mroot'),
    (namespaces['mathml'], 'mrow'),
    (namespaces['mathml'], 'mspace'),
    (namespaces['mathml'], 'msqrt'),
    (namespaces['mathml'], 'mstyle'),
    (namespaces['mathml'], 'msub'),
    (namespaces['mathml'], 'msubsup'),
    (namespaces['mathml'], 'msup'),
    (namespaces['mathml'], 'mtable'),
    (namespaces['mathml'], 'mtd'),
    (namespaces['mathml'], 'mtext'),
    (namespaces['mathml'], 'mtr'),
    (namespaces['mathml'], 'munder'),
    (namespaces['mathml'], 'munderover'),
    (namespaces['mathml'], 'none'),
    (namespaces['svg'], 'a'),
    (namespaces['svg'], 'animate'),
    (namespaces['svg'], 'animateColor'),
    (namespaces['svg'], 'animateMotion'),
    (namespaces['svg'], 'animateTransform'),
    (namespaces['svg'], 'clipPath'),
    (namespaces['svg'], 'circle'),
    (namespaces['svg'], 'defs'),
    (namespaces['svg'], 'desc'),
    (namespaces['svg'], 'ellipse'),
    (namespaces['svg'], 'font-face'),
    (namespaces['svg'], 'font-face-name'),
    (namespaces['svg'], 'font-face-src'),
    (namespaces['svg'], 'g'),
    (namespaces['svg'], 'glyph'),
    (namespaces['svg'], 'hkern'),
    (namespaces['svg'], 'linearGradient'),
    (namespaces['svg'], 'line'),
    (namespaces['svg'], 'marker'),
    (namespaces['svg'], 'metadata'),
    (namespaces['svg'], 'missing-glyph'),
    (namespaces['svg'], 'mpath'),
    (namespaces['svg'], 'path'),
    (namespaces['svg'], 'polygon'),
    (namespaces['svg'], 'polyline'),
    (namespaces['svg'], 'radialGradient'),
    (namespaces['svg'], 'rect'),
    (namespaces['svg'], 'set'),
    (namespaces['svg'], 'stop'),
    (namespaces['svg'], 'svg'),
    (namespaces['svg'], 'switch'),
    (namespaces['svg'], 'text'),
    (namespaces['svg'], 'title'),
    (namespaces['svg'], 'tspan'),
    (namespaces['svg'], 'use'),
))

allowed_attributes = frozenset((
    # HTML attributes
    (None, 'abbr'),
    (None, 'accept'),
    (None, 'accept-charset'),
    (None, 'accesskey'),
    (None, 'action'),
    (None, 'align'),
    (None, 'alt'),
    (None, 'autocomplete'),
    (None, 'autofocus'),
    (None, 'axis'),
    (None, 'background'),
    (None, 'balance'),
    (None, 'bgcolor'),
    (None, 'bgproperties'),
    (None, 'border'),
    (None, 'bordercolor'),
    (None, 'bordercolordark'),
    (None, 'bordercolorlight'),
    (None, 'bottompadding'),
    (None, 'cellpadding'),
    (None, 'cellspacing'),
    (None, 'ch'),
    (None, 'challenge'),
    (None, 'char'),
    (None, 'charoff'),
    (None, 'choff'),
    (None, 'charset'),
    (None, 'checked'),
    (None, 'cite'),
    (None, 'class'),
    (None, 'clear'),
    (None, 'color'),
    (None, 'cols'),
    (None, 'colspan'),
    (None, 'compact'),
    (None, 'contenteditable'),
    (None, 'controls'),
    (None, 'coords'),
    (None, 'data'),
    (None, 'datafld'),
    (None, 'datapagesize'),
    (None, 'datasrc'),
    (None, 'datetime'),
    (None, 'default'),
    (None, 'delay'),
    (None, 'dir'),
    (None, 'disabled'),
    (None, 'draggable'),
    (None, 'dynsrc'),
    (None, 'enctype'),
    (None, 'end'),
    (None, 'face'),
    (None, 'for'),
    (None, 'form'),
    (None, 'frame'),
    (None, 'galleryimg'),
    (None, 'gutter'),
    (None, 'headers'),
    (None, 'height'),
    (None, 'hidefocus'),
    (None, 'hidden'),
    (None, 'high'),
    (None, 'href'),
    (None, 'hreflang'),
    (None, 'hspace'),
    (None, 'icon'),
    (None, 'id'),
    (None, 'inputmode'),
    (None, 'ismap'),
    (None, 'keytype'),
    (None, 'label'),
    (None, 'leftspacing'),
    (None, 'lang'),
    (None, 'list'),
    (None, 'longdesc'),
    (None, 'loop'),
    (None, 'loopcount'),
    (None, 'loopend'),
    (None, 'loopstart'),
    (None, 'low'),
    (None, 'lowsrc'),
    (None, 'max'),
    (None, 'maxlength'),
    (None, 'media'),
    (None, 'method'),
    (None, 'min'),
    (None, 'multiple'),
    (None, 'name'),
    (None, 'nohref'),
    (None, 'noshade'),
    (None, 'nowrap'),
    (None, 'open'),
    (None, 'optimum'),
    (None, 'pattern'),
    (None, 'ping'),
    (None, 'point-size'),
    (None, 'poster'),
    (None, 'pqg'),
    (None, 'preload'),
    (None, 'prompt'),
    (None, 'radiogroup'),
    (None, 'readonly'),
    (None, 'rel'),
    (None, 'repeat-max'),
    (None, 'repeat-min'),
    (None, 'replace'),
    (None, 'required'),
    (None, 'rev'),
    (None, 'rightspacing'),
    (None, 'rows'),
    (None, 'rowspan'),
    (None, 'rules'),
    (None, 'scope'),
    (None, 'selected'),
    (None, 'shape'),
    (None, 'size'),
    (None, 'span'),
    (None, 'src'),
    (None, 'start'),
    (None, 'step'),
    (None, 'style'),
    (None, 'summary'),
    (None, 'suppress'),
    (None, 'tabindex'),
    (None, 'target'),
    (None, 'template'),
    (None, 'title'),
    (None, 'toppadding'),
    (None, 'type'),
    (None, 'unselectable'),
    (None, 'usemap'),
    (None, 'urn'),
    (None, 'valign'),
    (None, 'value'),
    (None, 'variable'),
    (None, 'volume'),
    (None, 'vspace'),
    (None, 'vrml'),
    (None, 'width'),
    (None, 'wrap'),
    (namespaces['xml'], 'lang'),
    # MathML attributes
    (None, 'actiontype'),
    (None, 'align'),
    (None, 'columnalign'),
    (None, 'columnalign'),
    (None, 'columnalign'),
    (None, 'columnlines'),
    (None, 'columnspacing'),
    (None, 'columnspan'),
    (None, 'depth'),
    (None, 'display'),
    (None, 'displaystyle'),
    (None, 'equalcolumns'),
    (None, 'equalrows'),
    (None, 'fence'),
    (None, 'fontstyle'),
    (None, 'fontweight'),
    (None, 'frame'),
    (None, 'height'),
    (None, 'linethickness'),
    (None, 'lspace'),
    (None, 'mathbackground'),
    (None, 'mathcolor'),
    (None, 'mathvariant'),
    (None, 'mathvariant'),
    (None, 'maxsize'),
    (None, 'minsize'),
    (None, 'other'),
    (None, 'rowalign'),
    (None, 'rowalign'),
    (None, 'rowalign'),
    (None, 'rowlines'),
    (None, 'rowspacing'),
    (None, 'rowspan'),
    (None, 'rspace'),
    (None, 'scriptlevel'),
    (None, 'selection'),
    (None, 'separator'),
    (None, 'stretchy'),
    (None, 'width'),
    (None, 'width'),
    (namespaces['xlink'], 'href'),
    (namespaces['xlink'], 'show'),
    (namespaces['xlink'], 'type'),
    # SVG attributes
    (None, 'accent-height'),
    (None, 'accumulate'),
    (None, 'additive'),
    (None, 'alphabetic'),
    (None, 'arabic-form'),
    (None, 'ascent'),
    (None, 'attributeName'),
    (None, 'attributeType'),
    (None, 'baseProfile'),
    (None, 'bbox'),
    (None, 'begin'),
    (None, 'by'),
    (None, 'calcMode'),
    (None, 'cap-height'),
    (None, 'class'),
    (None, 'clip-path'),
    (None, 'color'),
    (None, 'color-rendering'),
    (None, 'content'),
    (None, 'cx'),
    (None, 'cy'),
    (None, 'd'),
    (None, 'dx'),
    (None, 'dy'),
    (None, 'descent'),
    (None, 'display'),
    (None, 'dur'),
    (None, 'end'),
    (None, 'fill'),
    (None, 'fill-opacity'),
    (None, 'fill-rule'),
    (None, 'font-family'),
    (None, 'font-size'),
    (None, 'font-stretch'),
    (None, 'font-style'),
    (None, 'font-variant'),
    (None, 'font-weight'),
    (None, 'from'),
    (None, 'fx'),
    (None, 'fy'),
    (None, 'g1'),
    (None, 'g2'),
    (None, 'glyph-name'),
    (None, 'gradientUnits'),
    (None, 'hanging'),
    (None, 'height'),
    (None, 'horiz-adv-x'),
    (None, 'horiz-origin-x'),
    (None, 'id'),
    (None, 'ideographic'),
    (None, 'k'),
    (None, 'keyPoints'),
    (None, 'keySplines'),
    (None, 'keyTimes'),
    (None, 'lang'),
    (None, 'marker-end'),
    (None, 'marker-mid'),
    (None, 'marker-start'),
    (None, 'markerHeight'),
    (None, 'markerUnits'),
    (None, 'markerWidth'),
    (None, 'mathematical'),
    (None, 'max'),
    (None, 'min'),
    (None, 'name'),
    (None, 'offset'),
    (None, 'opacity'),
    (None, 'orient'),
    (None, 'origin'),
    (None, 'overline-position'),
    (None, 'overline-thickness'),
    (None, 'panose-1'),
    (None, 'path'),
    (None, 'pathLength'),
    (None, 'points'),
    (None, 'preserveAspectRatio'),
    (None, 'r'),
    (None, 'refX'),
    (None, 'refY'),
    (None, 'repeatCount'),
    (None, 'repeatDur'),
    (None, 'requiredExtensions'),
    (None, 'requiredFeatures'),
    (None, 'restart'),
    (None, 'rotate'),
    (None, 'rx'),
    (None, 'ry'),
    (None, 'slope'),
    (None, 'stemh'),
    (None, 'stemv'),
    (None, 'stop-color'),
    (None, 'stop-opacity'),
    (None, 'strikethrough-position'),
    (None, 'strikethrough-thickness'),
    (None, 'stroke'),
    (None, 'stroke-dasharray'),
    (None, 'stroke-dashoffset'),
    (None, 'stroke-linecap'),
    (None, 'stroke-linejoin'),
    (None, 'stroke-miterlimit'),
    (None, 'stroke-opacity'),
    (None, 'stroke-width'),
    (None, 'systemLanguage'),
    (None, 'target'),
    (None, 'text-anchor'),
    (None, 'to'),
    (None, 'transform'),
    (None, 'type'),
    (None, 'u1'),
    (None, 'u2'),
    (None, 'underline-position'),
    (None, 'underline-thickness'),
    (None, 'unicode'),
    (None, 'unicode-range'),
    (None, 'units-per-em'),
    (None, 'values'),
    (None, 'version'),
    (None, 'viewBox'),
    (None, 'visibility'),
    (None, 'width'),
    (None, 'widths'),
    (None, 'x'),
    (None, 'x-height'),
    (None, 'x1'),
    (None, 'x2'),
    (namespaces['xlink'], 'actuate'),
    (namespaces['xlink'], 'arcrole'),
    (namespaces['xlink'], 'href'),
    (namespaces['xlink'], 'role'),
    (namespaces['xlink'], 'show'),
    (namespaces['xlink'], 'title'),
    (namespaces['xlink'], 'type'),
    (namespaces['xml'], 'base'),
    (namespaces['xml'], 'lang'),
    (namespaces['xml'], 'space'),
    (None, 'y'),
    (None, 'y1'),
    (None, 'y2'),
    (None, 'zoomAndPan'),
))

attr_val_is_uri = frozenset((
    (None, 'href'),
    (None, 'src'),
    (None, 'cite'),
    (None, 'action'),
    (None, 'longdesc'),
    (None, 'poster'),
    (None, 'background'),
    (None, 'datasrc'),
    (None, 'dynsrc'),
    (None, 'lowsrc'),
    (None, 'ping'),
    (namespaces['xlink'], 'href'),
    (namespaces['xml'], 'base'),
))

svg_attr_val_allows_ref = frozenset((
    (None, 'clip-path'),
    (None, 'color-profile'),
    (None, 'cursor'),
    (None, 'fill'),
    (None, 'filter'),
    (None, 'marker'),
    (None, 'marker-start'),
    (None, 'marker-mid'),
    (None, 'marker-end'),
    (None, 'mask'),
    (None, 'stroke'),
))

svg_allow_local_href = frozenset((
    (None, 'altGlyph'),
    (None, 'animate'),
    (None, 'animateColor'),
    (None, 'animateMotion'),
    (None, 'animateTransform'),
    (None, 'cursor'),
    (None, 'feImage'),
    (None, 'filter'),
    (None, 'linearGradient'),
    (None, 'pattern'),
    (None, 'radialGradient'),
    (None, 'textpath'),
    (None, 'tref'),
    (None, 'set'),
    (None, 'use')
))

allowed_css_properties = frozenset((
    'azimuth',
    'background-color',
    'border-bottom-color',
    'border-collapse',
    'border-color',
    'border-left-color',
    'border-right-color',
    'border-top-color',
    'clear',
    'color',
    'cursor',
    'direction',
    'display',
    'elevation',
    'float',
    'font',
    'font-family',
    'font-size',
    'font-style',
    'font-variant',
    'font-weight',
    'height',
    'letter-spacing',
    'line-height',
    'overflow',
    'pause',
    'pause-after',
    'pause-before',
    'pitch',
    'pitch-range',
    'richness',
    'speak',
    'speak-header',
    'speak-numeral',
    'speak-punctuation',
    'speech-rate',
    'stress',
    'text-align',
    'text-decoration',
    'text-indent',
    'unicode-bidi',
    'vertical-align',
    'voice-family',
    'volume',
    'white-space',
    'width',
))

allowed_css_keywords = frozenset((
    'auto',
    'aqua',
    'black',
    'block',
    'blue',
    'bold',
    'both',
    'bottom',
    'brown',
    'center',
    'collapse',
    'dashed',
    'dotted',
    'fuchsia',
    'gray',
    'green',
    '!important',
    'italic',
    'left',
    'lime',
    'maroon',
    'medium',
    'none',
    'navy',
    'normal',
    'nowrap',
    'olive',
    'pointer',
    'purple',
    'red',
    'right',
    'solid',
    'silver',
    'teal',
    'top',
    'transparent',
    'underline',
    'white',
    'yellow',
))

allowed_svg_properties = frozenset((
    'fill',
    'fill-opacity',
    'fill-rule',
    'stroke',
    'stroke-width',
    'stroke-linecap',
    'stroke-linejoin',
    'stroke-opacity',
))

allowed_protocols = frozenset((
    'ed2k',
    'ftp',
    'http',
    'https',
    'irc',
    'mailto',
    'news',
    'gopher',
    'nntp',
    'telnet',
    'webcal',
    'xmpp',
    'callto',
    'feed',
    'urn',
    'aim',
    'rsync',
    'tag',
    'ssh',
    'sftp',
    'rtsp',
    'afs',
    'data',
))

allowed_content_types = frozenset((
    'image/png',
    'image/jpeg',
    'image/gif',
    'image/webp',
    'image/bmp',
    'text/plain',
))


data_content_type = re.compile(r'''
                                ^
                                # Match a content type <application>/<type>
                                (?P<content_type>[-a-zA-Z0-9.]+/[-a-zA-Z0-9.]+)
                                # Match any character set and encoding
                                (?:(?:;charset=(?:[-a-zA-Z0-9]+)(?:;(?:base64))?)
                                  |(?:;(?:base64))?(?:;charset=(?:[-a-zA-Z0-9]+))?)
                                # Assume the rest is data
                                ,.*
                                $
                                ''',
                               re.VERBOSE)


class Filter(base.Filter):
    """Sanitizes token stream of XHTML+MathML+SVG and of inline style attributes"""
    def __init__(self,
                 source,
                 allowed_elements=allowed_elements,
                 allowed_attributes=allowed_attributes,
                 allowed_css_properties=allowed_css_properties,
                 allowed_css_keywords=allowed_css_keywords,
                 allowed_svg_properties=allowed_svg_properties,
                 allowed_protocols=allowed_protocols,
                 allowed_content_types=allowed_content_types,
                 attr_val_is_uri=attr_val_is_uri,
                 svg_attr_val_allows_ref=svg_attr_val_allows_ref,
                 svg_allow_local_href=svg_allow_local_href):
        """Creates a Filter

        :arg allowed_elements: set of elements to allow--everything else will
            be escaped

        :arg allowed_attributes: set of attributes to allow in
            elements--everything else will be stripped

        :arg allowed_css_properties: set of CSS properties to allow--everything
            else will be stripped

        :arg allowed_css_keywords: set of CSS keywords to allow--everything
            else will be stripped

        :arg allowed_svg_properties: set of SVG properties to allow--everything
            else will be removed

        :arg allowed_protocols: set of allowed protocols for URIs

        :arg allowed_content_types: set of allowed content types for ``data`` URIs.

        :arg attr_val_is_uri: set of attributes that have URI values--values
            that have a scheme not listed in ``allowed_protocols`` are removed

        :arg svg_attr_val_allows_ref: set of SVG attributes that can have
            references

        :arg svg_allow_local_href: set of SVG elements that can have local
            hrefs--these are removed

        """
        super(Filter, self).__init__(source)

        warnings.warn(_deprecation_msg, DeprecationWarning)

        self.allowed_elements = allowed_elements
        self.allowed_attributes = allowed_attributes
        self.allowed_css_properties = allowed_css_properties
        self.allowed_css_keywords = allowed_css_keywords
        self.allowed_svg_properties = allowed_svg_properties
        self.allowed_protocols = allowed_protocols
        self.allowed_content_types = allowed_content_types
        self.attr_val_is_uri = attr_val_is_uri
        self.svg_attr_val_allows_ref = svg_attr_val_allows_ref
        self.svg_allow_local_href = svg_allow_local_href

    def __iter__(self):
        for token in base.Filter.__iter__(self):
            token = self.sanitize_token(token)
            if token:
                yield token

    # Sanitize the +html+, escaping all elements not in ALLOWED_ELEMENTS, and
    # stripping out all attributes not in ALLOWED_ATTRIBUTES. Style attributes
    # are parsed, and a restricted set, specified by ALLOWED_CSS_PROPERTIES and
    # ALLOWED_CSS_KEYWORDS, are allowed through. attributes in ATTR_VAL_IS_URI
    # are scanned, and only URI schemes specified in ALLOWED_PROTOCOLS are
    # allowed.
    #
    #   sanitize_html('<script> do_nasty_stuff() </script>')
    #    => &lt;script> do_nasty_stuff() &lt;/script>
    #   sanitize_html('<a href="javascript: sucker();">Click here for $100</a>')
    #    => <a>Click here for $100</a>
    def sanitize_token(self, token):

        # accommodate filters which use token_type differently
        token_type = token["type"]
        if token_type in ("StartTag", "EndTag", "EmptyTag"):
            name = token["name"]
            namespace = token["namespace"]
            if ((namespace, name) in self.allowed_elements or
                (namespace is None and
                 (namespaces["html"], name) in self.allowed_elements)):
                return self.allowed_token(token)
            else:
                return self.disallowed_token(token)
        elif token_type == "Comment":
            pass
        else:
            return token

    def allowed_token(self, token):
        if "data" in token:
            attrs = token["data"]
            attr_names = set(attrs.keys())

            # Remove forbidden attributes
            for to_remove in (attr_names - self.allowed_attributes):
                del token["data"][to_remove]
                attr_names.remove(to_remove)

            # Remove attributes with disallowed URL values
            for attr in (attr_names & self.attr_val_is_uri):
                assert attr in attrs
                # I don't have a clue where this regexp comes from or why it matches those
                # characters, nor why we call unescape. I just know it's always been here.
                # Should you be worried by this comment in a sanitizer? Yes. On the other hand, all
                # this will do is remove *more* than it otherwise would.
                val_unescaped = re.sub("[`\x00-\x20\x7f-\xa0\\s]+", '',
                                       unescape(attrs[attr])).lower()
                # remove replacement characters from unescaped characters
                val_unescaped = val_unescaped.replace("\ufffd", "")
                try:
                    uri = urlparse.urlparse(val_unescaped)
                except ValueError:
                    uri = None
                    del attrs[attr]
                if uri and uri.scheme:
                    if uri.scheme not in self.allowed_protocols:
                        del attrs[attr]
                    if uri.scheme == 'data':
                        m = data_content_type.match(uri.path)
                        if not m:
                            del attrs[attr]
                        elif m.group('content_type') not in self.allowed_content_types:
                            del attrs[attr]

            for attr in self.svg_attr_val_allows_ref:
                if attr in attrs:
                    attrs[attr] = re.sub(r'url\s*\(\s*[^#\s][^)]+?\)',
                                         ' ',
                                         unescape(attrs[attr]))
            if (token["name"] in self.svg_allow_local_href and
                (namespaces['xlink'], 'href') in attrs and re.search(r'^\s*[^#\s].*',
                                                                     attrs[(namespaces['xlink'], 'href')])):
                del attrs[(namespaces['xlink'], 'href')]
            if (None, 'style') in attrs:
                attrs[(None, 'style')] = self.sanitize_css(attrs[(None, 'style')])
            token["data"] = attrs
        return token

    def disallowed_token(self, token):
        token_type = token["type"]
        if token_type == "EndTag":
            token["data"] = "</%s>" % token["name"]
        elif token["data"]:
            assert token_type in ("StartTag", "EmptyTag")
            attrs = []
            for (ns, name), v in token["data"].items():
                attrs.append(' %s="%s"' % (name if ns is None else "%s:%s" % (prefixes[ns], name), escape(v)))
            token["data"] = "<%s%s>" % (token["name"], ''.join(attrs))
        else:
            token["data"] = "<%s>" % token["name"]
        if token.get("selfClosing"):
            token["data"] = token["data"][:-1] + "/>"

        token["type"] = "Characters"

        del token["name"]
        return token

    def sanitize_css(self, style):
        # disallow urls
        style = re.compile(r'url\s*\(\s*[^\s)]+?\s*\)\s*').sub(' ', style)

        # gauntlet
        if not re.match(r"""^([:,;#%.\sa-zA-Z0-9!]|\w-\w|'[\s\w]+'|"[\s\w]+"|\([\d,\s]+\))*$""", style):
            return ''
        if not re.match(r"^\s*([-\w]+\s*:[^:;]*(;\s*|$))*$", style):
            return ''

        clean = []
        for prop, value in re.findall(r"([-\w]+)\s*:\s*([^:;]*)", style):
            if not value:
                continue
            if prop.lower() in self.allowed_css_properties:
                clean.append(prop + ': ' + value + ';')
            elif prop.split('-')[0].lower() in ['background', 'border', 'margin',
                                                'padding']:
                for keyword in value.split():
                    if keyword not in self.allowed_css_keywords and \
                            not re.match(r"^(#[0-9a-fA-F]+|rgb\(\d+%?,\d*%?,?\d*%?\)?|\d{0,2}\.?\d{0,2}(cm|em|ex|in|mm|pc|pt|px|%|,|\))?)$", keyword):  # noqa
                        break
                else:
                    clean.append(prop + ': ' + value + ';')
            elif prop.lower() in self.allowed_svg_properties:
                clean.append(prop + ': ' + value + ';')

        return ' '.join(clean)
site-packages/pip/_vendor/html5lib/filters/whitespace.py000064400000002276150564745010017424 0ustar00from __future__ import absolute_import, division, unicode_literals

import re

from . import base
from ..constants import rcdataElements, spaceCharacters
spaceCharacters = "".join(spaceCharacters)

SPACES_REGEX = re.compile("[%s]+" % spaceCharacters)


class Filter(base.Filter):
    """Collapses whitespace except in pre, textarea, and script elements"""
    spacePreserveElements = frozenset(["pre", "textarea"] + list(rcdataElements))

    def __iter__(self):
        preserve = 0
        for token in base.Filter.__iter__(self):
            type = token["type"]
            if type == "StartTag" \
                    and (preserve or token["name"] in self.spacePreserveElements):
                preserve += 1

            elif type == "EndTag" and preserve:
                preserve -= 1

            elif not preserve and type == "SpaceCharacters" and token["data"]:
                # Test on token["data"] above to not introduce spaces where there were not
                token["data"] = " "

            elif not preserve and type == "Characters":
                token["data"] = collapse_spaces(token["data"])

            yield token


def collapse_spaces(text):
    return SPACES_REGEX.sub(' ', text)
site-packages/pip/_vendor/html5lib/filters/optionaltags.py000064400000024534150564745010017775 0ustar00from __future__ import absolute_import, division, unicode_literals

from . import base


class Filter(base.Filter):
    """Removes optional tags from the token stream"""
    def slider(self):
        previous1 = previous2 = None
        for token in self.source:
            if previous1 is not None:
                yield previous2, previous1, token
            previous2 = previous1
            previous1 = token
        if previous1 is not None:
            yield previous2, previous1, None

    def __iter__(self):
        for previous, token, next in self.slider():
            type = token["type"]
            if type == "StartTag":
                if (token["data"] or
                        not self.is_optional_start(token["name"], previous, next)):
                    yield token
            elif type == "EndTag":
                if not self.is_optional_end(token["name"], next):
                    yield token
            else:
                yield token

    def is_optional_start(self, tagname, previous, next):
        type = next and next["type"] or None
        if tagname in 'html':
            # An html element's start tag may be omitted if the first thing
            # inside the html element is not a space character or a comment.
            return type not in ("Comment", "SpaceCharacters")
        elif tagname == 'head':
            # A head element's start tag may be omitted if the first thing
            # inside the head element is an element.
            # XXX: we also omit the start tag if the head element is empty
            if type in ("StartTag", "EmptyTag"):
                return True
            elif type == "EndTag":
                return next["name"] == "head"
        elif tagname == 'body':
            # A body element's start tag may be omitted if the first thing
            # inside the body element is not a space character or a comment,
            # except if the first thing inside the body element is a script
            # or style element and the node immediately preceding the body
            # element is a head element whose end tag has been omitted.
            if type in ("Comment", "SpaceCharacters"):
                return False
            elif type == "StartTag":
                # XXX: we do not look at the preceding event, so we never omit
                # the body element's start tag if it's followed by a script or
                # a style element.
                return next["name"] not in ('script', 'style')
            else:
                return True
        elif tagname == 'colgroup':
            # A colgroup element's start tag may be omitted if the first thing
            # inside the colgroup element is a col element, and if the element
            # is not immediately preceded by another colgroup element whose
            # end tag has been omitted.
            if type in ("StartTag", "EmptyTag"):
                # XXX: we do not look at the preceding event, so instead we never
                # omit the colgroup element's end tag when it is immediately
                # followed by another colgroup element. See is_optional_end.
                return next["name"] == "col"
            else:
                return False
        elif tagname == 'tbody':
            # A tbody element's start tag may be omitted if the first thing
            # inside the tbody element is a tr element, and if the element is
            # not immediately preceded by a tbody, thead, or tfoot element
            # whose end tag has been omitted.
            if type == "StartTag":
                # omit the thead and tfoot elements' end tag when they are
                # immediately followed by a tbody element. See is_optional_end.
                if previous and previous['type'] == 'EndTag' and \
                        previous['name'] in ('tbody', 'thead', 'tfoot'):
                    return False
                return next["name"] == 'tr'
            else:
                return False
        return False

    def is_optional_end(self, tagname, next):
        type = next and next["type"] or None
        if tagname in ('html', 'head', 'body'):
            # An html element's end tag may be omitted if the html element
            # is not immediately followed by a space character or a comment.
            return type not in ("Comment", "SpaceCharacters")
        elif tagname in ('li', 'optgroup', 'tr'):
            # A li element's end tag may be omitted if the li element is
            # immediately followed by another li element or if there is
            # no more content in the parent element.
            # An optgroup element's end tag may be omitted if the optgroup
            # element is immediately followed by another optgroup element,
            # or if there is no more content in the parent element.
            # A tr element's end tag may be omitted if the tr element is
            # immediately followed by another tr element, or if there is
            # no more content in the parent element.
            if type == "StartTag":
                return next["name"] == tagname
            else:
                return type == "EndTag" or type is None
        elif tagname in ('dt', 'dd'):
            # A dt element's end tag may be omitted if the dt element is
            # immediately followed by another dt element or a dd element.
            # A dd element's end tag may be omitted if the dd element is
            # immediately followed by another dd element or a dt element,
            # or if there is no more content in the parent element.
            if type == "StartTag":
                return next["name"] in ('dt', 'dd')
            elif tagname == 'dd':
                return type == "EndTag" or type is None
            else:
                return False
        elif tagname == 'p':
            # A p element's end tag may be omitted if the p element is
            # immediately followed by an address, article, aside,
            # blockquote, datagrid, dialog, dir, div, dl, fieldset,
            # footer, form, h1, h2, h3, h4, h5, h6, header, hr, menu,
            # nav, ol, p, pre, section, table, or ul, element, or if
            # there is no more content in the parent element.
            if type in ("StartTag", "EmptyTag"):
                return next["name"] in ('address', 'article', 'aside',
                                        'blockquote', 'datagrid', 'dialog',
                                        'dir', 'div', 'dl', 'fieldset', 'footer',
                                        'form', 'h1', 'h2', 'h3', 'h4', 'h5', 'h6',
                                        'header', 'hr', 'menu', 'nav', 'ol',
                                        'p', 'pre', 'section', 'table', 'ul')
            else:
                return type == "EndTag" or type is None
        elif tagname == 'option':
            # An option element's end tag may be omitted if the option
            # element is immediately followed by another option element,
            # or if it is immediately followed by an <code>optgroup</code>
            # element, or if there is no more content in the parent
            # element.
            if type == "StartTag":
                return next["name"] in ('option', 'optgroup')
            else:
                return type == "EndTag" or type is None
        elif tagname in ('rt', 'rp'):
            # An rt element's end tag may be omitted if the rt element is
            # immediately followed by an rt or rp element, or if there is
            # no more content in the parent element.
            # An rp element's end tag may be omitted if the rp element is
            # immediately followed by an rt or rp element, or if there is
            # no more content in the parent element.
            if type == "StartTag":
                return next["name"] in ('rt', 'rp')
            else:
                return type == "EndTag" or type is None
        elif tagname == 'colgroup':
            # A colgroup element's end tag may be omitted if the colgroup
            # element is not immediately followed by a space character or
            # a comment.
            if type in ("Comment", "SpaceCharacters"):
                return False
            elif type == "StartTag":
                # XXX: we also look for an immediately following colgroup
                # element. See is_optional_start.
                return next["name"] != 'colgroup'
            else:
                return True
        elif tagname in ('thead', 'tbody'):
            # A thead element's end tag may be omitted if the thead element
            # is immediately followed by a tbody or tfoot element.
            # A tbody element's end tag may be omitted if the tbody element
            # is immediately followed by a tbody or tfoot element, or if
            # there is no more content in the parent element.
            # A tfoot element's end tag may be omitted if the tfoot element
            # is immediately followed by a tbody element, or if there is no
            # more content in the parent element.
            # XXX: we never omit the end tag when the following element is
            # a tbody. See is_optional_start.
            if type == "StartTag":
                return next["name"] in ['tbody', 'tfoot']
            elif tagname == 'tbody':
                return type == "EndTag" or type is None
            else:
                return False
        elif tagname == 'tfoot':
            # A tfoot element's end tag may be omitted if the tfoot element
            # is immediately followed by a tbody element, or if there is no
            # more content in the parent element.
            # XXX: we never omit the end tag when the following element is
            # a tbody. See is_optional_start.
            if type == "StartTag":
                return next["name"] == 'tbody'
            else:
                return type == "EndTag" or type is None
        elif tagname in ('td', 'th'):
            # A td element's end tag may be omitted if the td element is
            # immediately followed by a td or th element, or if there is
            # no more content in the parent element.
            # A th element's end tag may be omitted if the th element is
            # immediately followed by a td or th element, or if there is
            # no more content in the parent element.
            if type == "StartTag":
                return next["name"] in ('td', 'th')
            else:
                return type == "EndTag" or type is None
        return False
site-packages/pip/_vendor/html5lib/filters/__init__.py000064400000000000150564745010017006 0ustar00site-packages/pip/_vendor/html5lib/filters/alphabeticalattributes.py000064400000001627150564745010022007 0ustar00from __future__ import absolute_import, division, unicode_literals

from . import base

from collections import OrderedDict


def _attr_key(attr):
    """Return an appropriate key for an attribute for sorting

    Attributes have a namespace that can be either ``None`` or a string. We
    can't compare the two because they're different types, so we convert
    ``None`` to an empty string first.

    """
    return (attr[0][0] or ''), attr[0][1]


class Filter(base.Filter):
    """Alphabetizes attributes for elements"""
    def __iter__(self):
        for token in base.Filter.__iter__(self):
            if token["type"] in ("StartTag", "EmptyTag"):
                attrs = OrderedDict()
                for name, value in sorted(token["data"].items(),
                                          key=_attr_key):
                    attrs[name] = value
                token["data"] = attrs
            yield token
site-packages/pip/_vendor/html5lib/filters/inject_meta_charset.py000064400000005601150564745010021256 0ustar00from __future__ import absolute_import, division, unicode_literals

from . import base


class Filter(base.Filter):
    """Injects ``<meta charset=ENCODING>`` tag into head of document"""
    def __init__(self, source, encoding):
        """Creates a Filter

        :arg source: the source token stream

        :arg encoding: the encoding to set

        """
        base.Filter.__init__(self, source)
        self.encoding = encoding

    def __iter__(self):
        state = "pre_head"
        meta_found = (self.encoding is None)
        pending = []

        for token in base.Filter.__iter__(self):
            type = token["type"]
            if type == "StartTag":
                if token["name"].lower() == "head":
                    state = "in_head"

            elif type == "EmptyTag":
                if token["name"].lower() == "meta":
                    # replace charset with actual encoding
                    has_http_equiv_content_type = False
                    for (namespace, name), value in token["data"].items():
                        if namespace is not None:
                            continue
                        elif name.lower() == 'charset':
                            token["data"][(namespace, name)] = self.encoding
                            meta_found = True
                            break
                        elif name == 'http-equiv' and value.lower() == 'content-type':
                            has_http_equiv_content_type = True
                    else:
                        if has_http_equiv_content_type and (None, "content") in token["data"]:
                            token["data"][(None, "content")] = 'text/html; charset=%s' % self.encoding
                            meta_found = True

                elif token["name"].lower() == "head" and not meta_found:
                    # insert meta into empty head
                    yield {"type": "StartTag", "name": "head",
                           "data": token["data"]}
                    yield {"type": "EmptyTag", "name": "meta",
                           "data": {(None, "charset"): self.encoding}}
                    yield {"type": "EndTag", "name": "head"}
                    meta_found = True
                    continue

            elif type == "EndTag":
                if token["name"].lower() == "head" and pending:
                    # insert meta into head (if necessary) and flush pending queue
                    yield pending.pop(0)
                    if not meta_found:
                        yield {"type": "EmptyTag", "name": "meta",
                               "data": {(None, "charset"): self.encoding}}
                    while pending:
                        yield pending.pop(0)
                    meta_found = True
                    state = "post_head"

            if state == "in_head":
                pending.append(token)
            else:
                yield token
site-packages/pip/_vendor/html5lib/constants.py000064400000243010150564745010015625 0ustar00from __future__ import absolute_import, division, unicode_literals

import string

EOF = None

E = {
    "null-character":
        "Null character in input stream, replaced with U+FFFD.",
    "invalid-codepoint":
        "Invalid codepoint in stream.",
    "incorrectly-placed-solidus":
        "Solidus (/) incorrectly placed in tag.",
    "incorrect-cr-newline-entity":
        "Incorrect CR newline entity, replaced with LF.",
    "illegal-windows-1252-entity":
        "Entity used with illegal number (windows-1252 reference).",
    "cant-convert-numeric-entity":
        "Numeric entity couldn't be converted to character "
        "(codepoint U+%(charAsInt)08x).",
    "illegal-codepoint-for-numeric-entity":
        "Numeric entity represents an illegal codepoint: "
        "U+%(charAsInt)08x.",
    "numeric-entity-without-semicolon":
        "Numeric entity didn't end with ';'.",
    "expected-numeric-entity-but-got-eof":
        "Numeric entity expected. Got end of file instead.",
    "expected-numeric-entity":
        "Numeric entity expected but none found.",
    "named-entity-without-semicolon":
        "Named entity didn't end with ';'.",
    "expected-named-entity":
        "Named entity expected. Got none.",
    "attributes-in-end-tag":
        "End tag contains unexpected attributes.",
    'self-closing-flag-on-end-tag':
        "End tag contains unexpected self-closing flag.",
    "expected-tag-name-but-got-right-bracket":
        "Expected tag name. Got '>' instead.",
    "expected-tag-name-but-got-question-mark":
        "Expected tag name. Got '?' instead. (HTML doesn't "
        "support processing instructions.)",
    "expected-tag-name":
        "Expected tag name. Got something else instead",
    "expected-closing-tag-but-got-right-bracket":
        "Expected closing tag. Got '>' instead. Ignoring '</>'.",
    "expected-closing-tag-but-got-eof":
        "Expected closing tag. Unexpected end of file.",
    "expected-closing-tag-but-got-char":
        "Expected closing tag. Unexpected character '%(data)s' found.",
    "eof-in-tag-name":
        "Unexpected end of file in the tag name.",
    "expected-attribute-name-but-got-eof":
        "Unexpected end of file. Expected attribute name instead.",
    "eof-in-attribute-name":
        "Unexpected end of file in attribute name.",
    "invalid-character-in-attribute-name":
        "Invalid character in attribute name",
    "duplicate-attribute":
        "Dropped duplicate attribute on tag.",
    "expected-end-of-tag-name-but-got-eof":
        "Unexpected end of file. Expected = or end of tag.",
    "expected-attribute-value-but-got-eof":
        "Unexpected end of file. Expected attribute value.",
    "expected-attribute-value-but-got-right-bracket":
        "Expected attribute value. Got '>' instead.",
    'equals-in-unquoted-attribute-value':
        "Unexpected = in unquoted attribute",
    'unexpected-character-in-unquoted-attribute-value':
        "Unexpected character in unquoted attribute",
    "invalid-character-after-attribute-name":
        "Unexpected character after attribute name.",
    "unexpected-character-after-attribute-value":
        "Unexpected character after attribute value.",
    "eof-in-attribute-value-double-quote":
        "Unexpected end of file in attribute value (\").",
    "eof-in-attribute-value-single-quote":
        "Unexpected end of file in attribute value (').",
    "eof-in-attribute-value-no-quotes":
        "Unexpected end of file in attribute value.",
    "unexpected-EOF-after-solidus-in-tag":
        "Unexpected end of file in tag. Expected >",
    "unexpected-character-after-solidus-in-tag":
        "Unexpected character after / in tag. Expected >",
    "expected-dashes-or-doctype":
        "Expected '--' or 'DOCTYPE'. Not found.",
    "unexpected-bang-after-double-dash-in-comment":
        "Unexpected ! after -- in comment",
    "unexpected-space-after-double-dash-in-comment":
        "Unexpected space after -- in comment",
    "incorrect-comment":
        "Incorrect comment.",
    "eof-in-comment":
        "Unexpected end of file in comment.",
    "eof-in-comment-end-dash":
        "Unexpected end of file in comment (-)",
    "unexpected-dash-after-double-dash-in-comment":
        "Unexpected '-' after '--' found in comment.",
    "eof-in-comment-double-dash":
        "Unexpected end of file in comment (--).",
    "eof-in-comment-end-space-state":
        "Unexpected end of file in comment.",
    "eof-in-comment-end-bang-state":
        "Unexpected end of file in comment.",
    "unexpected-char-in-comment":
        "Unexpected character in comment found.",
    "need-space-after-doctype":
        "No space after literal string 'DOCTYPE'.",
    "expected-doctype-name-but-got-right-bracket":
        "Unexpected > character. Expected DOCTYPE name.",
    "expected-doctype-name-but-got-eof":
        "Unexpected end of file. Expected DOCTYPE name.",
    "eof-in-doctype-name":
        "Unexpected end of file in DOCTYPE name.",
    "eof-in-doctype":
        "Unexpected end of file in DOCTYPE.",
    "expected-space-or-right-bracket-in-doctype":
        "Expected space or '>'. Got '%(data)s'",
    "unexpected-end-of-doctype":
        "Unexpected end of DOCTYPE.",
    "unexpected-char-in-doctype":
        "Unexpected character in DOCTYPE.",
    "eof-in-innerhtml":
        "XXX innerHTML EOF",
    "unexpected-doctype":
        "Unexpected DOCTYPE. Ignored.",
    "non-html-root":
        "html needs to be the first start tag.",
    "expected-doctype-but-got-eof":
        "Unexpected End of file. Expected DOCTYPE.",
    "unknown-doctype":
        "Erroneous DOCTYPE.",
    "expected-doctype-but-got-chars":
        "Unexpected non-space characters. Expected DOCTYPE.",
    "expected-doctype-but-got-start-tag":
        "Unexpected start tag (%(name)s). Expected DOCTYPE.",
    "expected-doctype-but-got-end-tag":
        "Unexpected end tag (%(name)s). Expected DOCTYPE.",
    "end-tag-after-implied-root":
        "Unexpected end tag (%(name)s) after the (implied) root element.",
    "expected-named-closing-tag-but-got-eof":
        "Unexpected end of file. Expected end tag (%(name)s).",
    "two-heads-are-not-better-than-one":
        "Unexpected start tag head in existing head. Ignored.",
    "unexpected-end-tag":
        "Unexpected end tag (%(name)s). Ignored.",
    "unexpected-start-tag-out-of-my-head":
        "Unexpected start tag (%(name)s) that can be in head. Moved.",
    "unexpected-start-tag":
        "Unexpected start tag (%(name)s).",
    "missing-end-tag":
        "Missing end tag (%(name)s).",
    "missing-end-tags":
        "Missing end tags (%(name)s).",
    "unexpected-start-tag-implies-end-tag":
        "Unexpected start tag (%(startName)s) "
        "implies end tag (%(endName)s).",
    "unexpected-start-tag-treated-as":
        "Unexpected start tag (%(originalName)s). Treated as %(newName)s.",
    "deprecated-tag":
        "Unexpected start tag %(name)s. Don't use it!",
    "unexpected-start-tag-ignored":
        "Unexpected start tag %(name)s. Ignored.",
    "expected-one-end-tag-but-got-another":
        "Unexpected end tag (%(gotName)s). "
        "Missing end tag (%(expectedName)s).",
    "end-tag-too-early":
        "End tag (%(name)s) seen too early. Expected other end tag.",
    "end-tag-too-early-named":
        "Unexpected end tag (%(gotName)s). Expected end tag (%(expectedName)s).",
    "end-tag-too-early-ignored":
        "End tag (%(name)s) seen too early. Ignored.",
    "adoption-agency-1.1":
        "End tag (%(name)s) violates step 1, "
        "paragraph 1 of the adoption agency algorithm.",
    "adoption-agency-1.2":
        "End tag (%(name)s) violates step 1, "
        "paragraph 2 of the adoption agency algorithm.",
    "adoption-agency-1.3":
        "End tag (%(name)s) violates step 1, "
        "paragraph 3 of the adoption agency algorithm.",
    "adoption-agency-4.4":
        "End tag (%(name)s) violates step 4, "
        "paragraph 4 of the adoption agency algorithm.",
    "unexpected-end-tag-treated-as":
        "Unexpected end tag (%(originalName)s). Treated as %(newName)s.",
    "no-end-tag":
        "This element (%(name)s) has no end tag.",
    "unexpected-implied-end-tag-in-table":
        "Unexpected implied end tag (%(name)s) in the table phase.",
    "unexpected-implied-end-tag-in-table-body":
        "Unexpected implied end tag (%(name)s) in the table body phase.",
    "unexpected-char-implies-table-voodoo":
        "Unexpected non-space characters in "
        "table context caused voodoo mode.",
    "unexpected-hidden-input-in-table":
        "Unexpected input with type hidden in table context.",
    "unexpected-form-in-table":
        "Unexpected form in table context.",
    "unexpected-start-tag-implies-table-voodoo":
        "Unexpected start tag (%(name)s) in "
        "table context caused voodoo mode.",
    "unexpected-end-tag-implies-table-voodoo":
        "Unexpected end tag (%(name)s) in "
        "table context caused voodoo mode.",
    "unexpected-cell-in-table-body":
        "Unexpected table cell start tag (%(name)s) "
        "in the table body phase.",
    "unexpected-cell-end-tag":
        "Got table cell end tag (%(name)s) "
        "while required end tags are missing.",
    "unexpected-end-tag-in-table-body":
        "Unexpected end tag (%(name)s) in the table body phase. Ignored.",
    "unexpected-implied-end-tag-in-table-row":
        "Unexpected implied end tag (%(name)s) in the table row phase.",
    "unexpected-end-tag-in-table-row":
        "Unexpected end tag (%(name)s) in the table row phase. Ignored.",
    "unexpected-select-in-select":
        "Unexpected select start tag in the select phase "
        "treated as select end tag.",
    "unexpected-input-in-select":
        "Unexpected input start tag in the select phase.",
    "unexpected-start-tag-in-select":
        "Unexpected start tag token (%(name)s in the select phase. "
        "Ignored.",
    "unexpected-end-tag-in-select":
        "Unexpected end tag (%(name)s) in the select phase. Ignored.",
    "unexpected-table-element-start-tag-in-select-in-table":
        "Unexpected table element start tag (%(name)s) in the select in table phase.",
    "unexpected-table-element-end-tag-in-select-in-table":
        "Unexpected table element end tag (%(name)s) in the select in table phase.",
    "unexpected-char-after-body":
        "Unexpected non-space characters in the after body phase.",
    "unexpected-start-tag-after-body":
        "Unexpected start tag token (%(name)s)"
        " in the after body phase.",
    "unexpected-end-tag-after-body":
        "Unexpected end tag token (%(name)s)"
        " in the after body phase.",
    "unexpected-char-in-frameset":
        "Unexpected characters in the frameset phase. Characters ignored.",
    "unexpected-start-tag-in-frameset":
        "Unexpected start tag token (%(name)s)"
        " in the frameset phase. Ignored.",
    "unexpected-frameset-in-frameset-innerhtml":
        "Unexpected end tag token (frameset) "
        "in the frameset phase (innerHTML).",
    "unexpected-end-tag-in-frameset":
        "Unexpected end tag token (%(name)s)"
        " in the frameset phase. Ignored.",
    "unexpected-char-after-frameset":
        "Unexpected non-space characters in the "
        "after frameset phase. Ignored.",
    "unexpected-start-tag-after-frameset":
        "Unexpected start tag (%(name)s)"
        " in the after frameset phase. Ignored.",
    "unexpected-end-tag-after-frameset":
        "Unexpected end tag (%(name)s)"
        " in the after frameset phase. Ignored.",
    "unexpected-end-tag-after-body-innerhtml":
        "Unexpected end tag after body(innerHtml)",
    "expected-eof-but-got-char":
        "Unexpected non-space characters. Expected end of file.",
    "expected-eof-but-got-start-tag":
        "Unexpected start tag (%(name)s)"
        ". Expected end of file.",
    "expected-eof-but-got-end-tag":
        "Unexpected end tag (%(name)s)"
        ". Expected end of file.",
    "eof-in-table":
        "Unexpected end of file. Expected table content.",
    "eof-in-select":
        "Unexpected end of file. Expected select content.",
    "eof-in-frameset":
        "Unexpected end of file. Expected frameset content.",
    "eof-in-script-in-script":
        "Unexpected end of file. Expected script content.",
    "eof-in-foreign-lands":
        "Unexpected end of file. Expected foreign content",
    "non-void-element-with-trailing-solidus":
        "Trailing solidus not allowed on element %(name)s",
    "unexpected-html-element-in-foreign-content":
        "Element %(name)s not allowed in a non-html context",
    "unexpected-end-tag-before-html":
        "Unexpected end tag (%(name)s) before html.",
    "unexpected-inhead-noscript-tag":
        "Element %(name)s not allowed in a inhead-noscript context",
    "eof-in-head-noscript":
        "Unexpected end of file. Expected inhead-noscript content",
    "char-in-head-noscript":
        "Unexpected non-space character. Expected inhead-noscript content",
    "XXX-undefined-error":
        "Undefined error (this sucks and should be fixed)",
}

namespaces = {
    "html": "http://www.w3.org/1999/xhtml",
    "mathml": "http://www.w3.org/1998/Math/MathML",
    "svg": "http://www.w3.org/2000/svg",
    "xlink": "http://www.w3.org/1999/xlink",
    "xml": "http://www.w3.org/XML/1998/namespace",
    "xmlns": "http://www.w3.org/2000/xmlns/"
}

scopingElements = frozenset([
    (namespaces["html"], "applet"),
    (namespaces["html"], "caption"),
    (namespaces["html"], "html"),
    (namespaces["html"], "marquee"),
    (namespaces["html"], "object"),
    (namespaces["html"], "table"),
    (namespaces["html"], "td"),
    (namespaces["html"], "th"),
    (namespaces["mathml"], "mi"),
    (namespaces["mathml"], "mo"),
    (namespaces["mathml"], "mn"),
    (namespaces["mathml"], "ms"),
    (namespaces["mathml"], "mtext"),
    (namespaces["mathml"], "annotation-xml"),
    (namespaces["svg"], "foreignObject"),
    (namespaces["svg"], "desc"),
    (namespaces["svg"], "title"),
])

formattingElements = frozenset([
    (namespaces["html"], "a"),
    (namespaces["html"], "b"),
    (namespaces["html"], "big"),
    (namespaces["html"], "code"),
    (namespaces["html"], "em"),
    (namespaces["html"], "font"),
    (namespaces["html"], "i"),
    (namespaces["html"], "nobr"),
    (namespaces["html"], "s"),
    (namespaces["html"], "small"),
    (namespaces["html"], "strike"),
    (namespaces["html"], "strong"),
    (namespaces["html"], "tt"),
    (namespaces["html"], "u")
])

specialElements = frozenset([
    (namespaces["html"], "address"),
    (namespaces["html"], "applet"),
    (namespaces["html"], "area"),
    (namespaces["html"], "article"),
    (namespaces["html"], "aside"),
    (namespaces["html"], "base"),
    (namespaces["html"], "basefont"),
    (namespaces["html"], "bgsound"),
    (namespaces["html"], "blockquote"),
    (namespaces["html"], "body"),
    (namespaces["html"], "br"),
    (namespaces["html"], "button"),
    (namespaces["html"], "caption"),
    (namespaces["html"], "center"),
    (namespaces["html"], "col"),
    (namespaces["html"], "colgroup"),
    (namespaces["html"], "command"),
    (namespaces["html"], "dd"),
    (namespaces["html"], "details"),
    (namespaces["html"], "dir"),
    (namespaces["html"], "div"),
    (namespaces["html"], "dl"),
    (namespaces["html"], "dt"),
    (namespaces["html"], "embed"),
    (namespaces["html"], "fieldset"),
    (namespaces["html"], "figure"),
    (namespaces["html"], "footer"),
    (namespaces["html"], "form"),
    (namespaces["html"], "frame"),
    (namespaces["html"], "frameset"),
    (namespaces["html"], "h1"),
    (namespaces["html"], "h2"),
    (namespaces["html"], "h3"),
    (namespaces["html"], "h4"),
    (namespaces["html"], "h5"),
    (namespaces["html"], "h6"),
    (namespaces["html"], "head"),
    (namespaces["html"], "header"),
    (namespaces["html"], "hr"),
    (namespaces["html"], "html"),
    (namespaces["html"], "iframe"),
    # Note that image is commented out in the spec as "this isn't an
    # element that can end up on the stack, so it doesn't matter,"
    (namespaces["html"], "image"),
    (namespaces["html"], "img"),
    (namespaces["html"], "input"),
    (namespaces["html"], "isindex"),
    (namespaces["html"], "li"),
    (namespaces["html"], "link"),
    (namespaces["html"], "listing"),
    (namespaces["html"], "marquee"),
    (namespaces["html"], "menu"),
    (namespaces["html"], "meta"),
    (namespaces["html"], "nav"),
    (namespaces["html"], "noembed"),
    (namespaces["html"], "noframes"),
    (namespaces["html"], "noscript"),
    (namespaces["html"], "object"),
    (namespaces["html"], "ol"),
    (namespaces["html"], "p"),
    (namespaces["html"], "param"),
    (namespaces["html"], "plaintext"),
    (namespaces["html"], "pre"),
    (namespaces["html"], "script"),
    (namespaces["html"], "section"),
    (namespaces["html"], "select"),
    (namespaces["html"], "style"),
    (namespaces["html"], "table"),
    (namespaces["html"], "tbody"),
    (namespaces["html"], "td"),
    (namespaces["html"], "textarea"),
    (namespaces["html"], "tfoot"),
    (namespaces["html"], "th"),
    (namespaces["html"], "thead"),
    (namespaces["html"], "title"),
    (namespaces["html"], "tr"),
    (namespaces["html"], "ul"),
    (namespaces["html"], "wbr"),
    (namespaces["html"], "xmp"),
    (namespaces["svg"], "foreignObject")
])

htmlIntegrationPointElements = frozenset([
    (namespaces["mathml"], "annotation-xml"),
    (namespaces["svg"], "foreignObject"),
    (namespaces["svg"], "desc"),
    (namespaces["svg"], "title")
])

mathmlTextIntegrationPointElements = frozenset([
    (namespaces["mathml"], "mi"),
    (namespaces["mathml"], "mo"),
    (namespaces["mathml"], "mn"),
    (namespaces["mathml"], "ms"),
    (namespaces["mathml"], "mtext")
])

adjustSVGAttributes = {
    "attributename": "attributeName",
    "attributetype": "attributeType",
    "basefrequency": "baseFrequency",
    "baseprofile": "baseProfile",
    "calcmode": "calcMode",
    "clippathunits": "clipPathUnits",
    "contentscripttype": "contentScriptType",
    "contentstyletype": "contentStyleType",
    "diffuseconstant": "diffuseConstant",
    "edgemode": "edgeMode",
    "externalresourcesrequired": "externalResourcesRequired",
    "filterres": "filterRes",
    "filterunits": "filterUnits",
    "glyphref": "glyphRef",
    "gradienttransform": "gradientTransform",
    "gradientunits": "gradientUnits",
    "kernelmatrix": "kernelMatrix",
    "kernelunitlength": "kernelUnitLength",
    "keypoints": "keyPoints",
    "keysplines": "keySplines",
    "keytimes": "keyTimes",
    "lengthadjust": "lengthAdjust",
    "limitingconeangle": "limitingConeAngle",
    "markerheight": "markerHeight",
    "markerunits": "markerUnits",
    "markerwidth": "markerWidth",
    "maskcontentunits": "maskContentUnits",
    "maskunits": "maskUnits",
    "numoctaves": "numOctaves",
    "pathlength": "pathLength",
    "patterncontentunits": "patternContentUnits",
    "patterntransform": "patternTransform",
    "patternunits": "patternUnits",
    "pointsatx": "pointsAtX",
    "pointsaty": "pointsAtY",
    "pointsatz": "pointsAtZ",
    "preservealpha": "preserveAlpha",
    "preserveaspectratio": "preserveAspectRatio",
    "primitiveunits": "primitiveUnits",
    "refx": "refX",
    "refy": "refY",
    "repeatcount": "repeatCount",
    "repeatdur": "repeatDur",
    "requiredextensions": "requiredExtensions",
    "requiredfeatures": "requiredFeatures",
    "specularconstant": "specularConstant",
    "specularexponent": "specularExponent",
    "spreadmethod": "spreadMethod",
    "startoffset": "startOffset",
    "stddeviation": "stdDeviation",
    "stitchtiles": "stitchTiles",
    "surfacescale": "surfaceScale",
    "systemlanguage": "systemLanguage",
    "tablevalues": "tableValues",
    "targetx": "targetX",
    "targety": "targetY",
    "textlength": "textLength",
    "viewbox": "viewBox",
    "viewtarget": "viewTarget",
    "xchannelselector": "xChannelSelector",
    "ychannelselector": "yChannelSelector",
    "zoomandpan": "zoomAndPan"
}

adjustMathMLAttributes = {"definitionurl": "definitionURL"}

adjustForeignAttributes = {
    "xlink:actuate": ("xlink", "actuate", namespaces["xlink"]),
    "xlink:arcrole": ("xlink", "arcrole", namespaces["xlink"]),
    "xlink:href": ("xlink", "href", namespaces["xlink"]),
    "xlink:role": ("xlink", "role", namespaces["xlink"]),
    "xlink:show": ("xlink", "show", namespaces["xlink"]),
    "xlink:title": ("xlink", "title", namespaces["xlink"]),
    "xlink:type": ("xlink", "type", namespaces["xlink"]),
    "xml:base": ("xml", "base", namespaces["xml"]),
    "xml:lang": ("xml", "lang", namespaces["xml"]),
    "xml:space": ("xml", "space", namespaces["xml"]),
    "xmlns": (None, "xmlns", namespaces["xmlns"]),
    "xmlns:xlink": ("xmlns", "xlink", namespaces["xmlns"])
}

unadjustForeignAttributes = {(ns, local): qname for qname, (prefix, local, ns) in
                             adjustForeignAttributes.items()}

spaceCharacters = frozenset([
    "\t",
    "\n",
    "\u000C",
    " ",
    "\r"
])

tableInsertModeElements = frozenset([
    "table",
    "tbody",
    "tfoot",
    "thead",
    "tr"
])

asciiLowercase = frozenset(string.ascii_lowercase)
asciiUppercase = frozenset(string.ascii_uppercase)
asciiLetters = frozenset(string.ascii_letters)
digits = frozenset(string.digits)
hexDigits = frozenset(string.hexdigits)

asciiUpper2Lower = {ord(c): ord(c.lower()) for c in string.ascii_uppercase}

# Heading elements need to be ordered
headingElements = (
    "h1",
    "h2",
    "h3",
    "h4",
    "h5",
    "h6"
)

voidElements = frozenset([
    "base",
    "command",
    "event-source",
    "link",
    "meta",
    "hr",
    "br",
    "img",
    "embed",
    "param",
    "area",
    "col",
    "input",
    "source",
    "track"
])

cdataElements = frozenset(['title', 'textarea'])

rcdataElements = frozenset([
    'style',
    'script',
    'xmp',
    'iframe',
    'noembed',
    'noframes',
    'noscript'
])

booleanAttributes = {
    "": frozenset(["irrelevant", "itemscope"]),
    "style": frozenset(["scoped"]),
    "img": frozenset(["ismap"]),
    "audio": frozenset(["autoplay", "controls"]),
    "video": frozenset(["autoplay", "controls"]),
    "script": frozenset(["defer", "async"]),
    "details": frozenset(["open"]),
    "datagrid": frozenset(["multiple", "disabled"]),
    "command": frozenset(["hidden", "disabled", "checked", "default"]),
    "hr": frozenset(["noshade"]),
    "menu": frozenset(["autosubmit"]),
    "fieldset": frozenset(["disabled", "readonly"]),
    "option": frozenset(["disabled", "readonly", "selected"]),
    "optgroup": frozenset(["disabled", "readonly"]),
    "button": frozenset(["disabled", "autofocus"]),
    "input": frozenset(["disabled", "readonly", "required", "autofocus", "checked", "ismap"]),
    "select": frozenset(["disabled", "readonly", "autofocus", "multiple"]),
    "output": frozenset(["disabled", "readonly"]),
    "iframe": frozenset(["seamless"]),
}

# entitiesWindows1252 has to be _ordered_ and needs to have an index. It
# therefore can't be a frozenset.
entitiesWindows1252 = (
    8364,   # 0x80  0x20AC  EURO SIGN
    65533,  # 0x81          UNDEFINED
    8218,   # 0x82  0x201A  SINGLE LOW-9 QUOTATION MARK
    402,    # 0x83  0x0192  LATIN SMALL LETTER F WITH HOOK
    8222,   # 0x84  0x201E  DOUBLE LOW-9 QUOTATION MARK
    8230,   # 0x85  0x2026  HORIZONTAL ELLIPSIS
    8224,   # 0x86  0x2020  DAGGER
    8225,   # 0x87  0x2021  DOUBLE DAGGER
    710,    # 0x88  0x02C6  MODIFIER LETTER CIRCUMFLEX ACCENT
    8240,   # 0x89  0x2030  PER MILLE SIGN
    352,    # 0x8A  0x0160  LATIN CAPITAL LETTER S WITH CARON
    8249,   # 0x8B  0x2039  SINGLE LEFT-POINTING ANGLE QUOTATION MARK
    338,    # 0x8C  0x0152  LATIN CAPITAL LIGATURE OE
    65533,  # 0x8D          UNDEFINED
    381,    # 0x8E  0x017D  LATIN CAPITAL LETTER Z WITH CARON
    65533,  # 0x8F          UNDEFINED
    65533,  # 0x90          UNDEFINED
    8216,   # 0x91  0x2018  LEFT SINGLE QUOTATION MARK
    8217,   # 0x92  0x2019  RIGHT SINGLE QUOTATION MARK
    8220,   # 0x93  0x201C  LEFT DOUBLE QUOTATION MARK
    8221,   # 0x94  0x201D  RIGHT DOUBLE QUOTATION MARK
    8226,   # 0x95  0x2022  BULLET
    8211,   # 0x96  0x2013  EN DASH
    8212,   # 0x97  0x2014  EM DASH
    732,    # 0x98  0x02DC  SMALL TILDE
    8482,   # 0x99  0x2122  TRADE MARK SIGN
    353,    # 0x9A  0x0161  LATIN SMALL LETTER S WITH CARON
    8250,   # 0x9B  0x203A  SINGLE RIGHT-POINTING ANGLE QUOTATION MARK
    339,    # 0x9C  0x0153  LATIN SMALL LIGATURE OE
    65533,  # 0x9D          UNDEFINED
    382,    # 0x9E  0x017E  LATIN SMALL LETTER Z WITH CARON
    376     # 0x9F  0x0178  LATIN CAPITAL LETTER Y WITH DIAERESIS
)

xmlEntities = frozenset(['lt;', 'gt;', 'amp;', 'apos;', 'quot;'])

entities = {
    "AElig": "\xc6",
    "AElig;": "\xc6",
    "AMP": "&",
    "AMP;": "&",
    "Aacute": "\xc1",
    "Aacute;": "\xc1",
    "Abreve;": "\u0102",
    "Acirc": "\xc2",
    "Acirc;": "\xc2",
    "Acy;": "\u0410",
    "Afr;": "\U0001d504",
    "Agrave": "\xc0",
    "Agrave;": "\xc0",
    "Alpha;": "\u0391",
    "Amacr;": "\u0100",
    "And;": "\u2a53",
    "Aogon;": "\u0104",
    "Aopf;": "\U0001d538",
    "ApplyFunction;": "\u2061",
    "Aring": "\xc5",
    "Aring;": "\xc5",
    "Ascr;": "\U0001d49c",
    "Assign;": "\u2254",
    "Atilde": "\xc3",
    "Atilde;": "\xc3",
    "Auml": "\xc4",
    "Auml;": "\xc4",
    "Backslash;": "\u2216",
    "Barv;": "\u2ae7",
    "Barwed;": "\u2306",
    "Bcy;": "\u0411",
    "Because;": "\u2235",
    "Bernoullis;": "\u212c",
    "Beta;": "\u0392",
    "Bfr;": "\U0001d505",
    "Bopf;": "\U0001d539",
    "Breve;": "\u02d8",
    "Bscr;": "\u212c",
    "Bumpeq;": "\u224e",
    "CHcy;": "\u0427",
    "COPY": "\xa9",
    "COPY;": "\xa9",
    "Cacute;": "\u0106",
    "Cap;": "\u22d2",
    "CapitalDifferentialD;": "\u2145",
    "Cayleys;": "\u212d",
    "Ccaron;": "\u010c",
    "Ccedil": "\xc7",
    "Ccedil;": "\xc7",
    "Ccirc;": "\u0108",
    "Cconint;": "\u2230",
    "Cdot;": "\u010a",
    "Cedilla;": "\xb8",
    "CenterDot;": "\xb7",
    "Cfr;": "\u212d",
    "Chi;": "\u03a7",
    "CircleDot;": "\u2299",
    "CircleMinus;": "\u2296",
    "CirclePlus;": "\u2295",
    "CircleTimes;": "\u2297",
    "ClockwiseContourIntegral;": "\u2232",
    "CloseCurlyDoubleQuote;": "\u201d",
    "CloseCurlyQuote;": "\u2019",
    "Colon;": "\u2237",
    "Colone;": "\u2a74",
    "Congruent;": "\u2261",
    "Conint;": "\u222f",
    "ContourIntegral;": "\u222e",
    "Copf;": "\u2102",
    "Coproduct;": "\u2210",
    "CounterClockwiseContourIntegral;": "\u2233",
    "Cross;": "\u2a2f",
    "Cscr;": "\U0001d49e",
    "Cup;": "\u22d3",
    "CupCap;": "\u224d",
    "DD;": "\u2145",
    "DDotrahd;": "\u2911",
    "DJcy;": "\u0402",
    "DScy;": "\u0405",
    "DZcy;": "\u040f",
    "Dagger;": "\u2021",
    "Darr;": "\u21a1",
    "Dashv;": "\u2ae4",
    "Dcaron;": "\u010e",
    "Dcy;": "\u0414",
    "Del;": "\u2207",
    "Delta;": "\u0394",
    "Dfr;": "\U0001d507",
    "DiacriticalAcute;": "\xb4",
    "DiacriticalDot;": "\u02d9",
    "DiacriticalDoubleAcute;": "\u02dd",
    "DiacriticalGrave;": "`",
    "DiacriticalTilde;": "\u02dc",
    "Diamond;": "\u22c4",
    "DifferentialD;": "\u2146",
    "Dopf;": "\U0001d53b",
    "Dot;": "\xa8",
    "DotDot;": "\u20dc",
    "DotEqual;": "\u2250",
    "DoubleContourIntegral;": "\u222f",
    "DoubleDot;": "\xa8",
    "DoubleDownArrow;": "\u21d3",
    "DoubleLeftArrow;": "\u21d0",
    "DoubleLeftRightArrow;": "\u21d4",
    "DoubleLeftTee;": "\u2ae4",
    "DoubleLongLeftArrow;": "\u27f8",
    "DoubleLongLeftRightArrow;": "\u27fa",
    "DoubleLongRightArrow;": "\u27f9",
    "DoubleRightArrow;": "\u21d2",
    "DoubleRightTee;": "\u22a8",
    "DoubleUpArrow;": "\u21d1",
    "DoubleUpDownArrow;": "\u21d5",
    "DoubleVerticalBar;": "\u2225",
    "DownArrow;": "\u2193",
    "DownArrowBar;": "\u2913",
    "DownArrowUpArrow;": "\u21f5",
    "DownBreve;": "\u0311",
    "DownLeftRightVector;": "\u2950",
    "DownLeftTeeVector;": "\u295e",
    "DownLeftVector;": "\u21bd",
    "DownLeftVectorBar;": "\u2956",
    "DownRightTeeVector;": "\u295f",
    "DownRightVector;": "\u21c1",
    "DownRightVectorBar;": "\u2957",
    "DownTee;": "\u22a4",
    "DownTeeArrow;": "\u21a7",
    "Downarrow;": "\u21d3",
    "Dscr;": "\U0001d49f",
    "Dstrok;": "\u0110",
    "ENG;": "\u014a",
    "ETH": "\xd0",
    "ETH;": "\xd0",
    "Eacute": "\xc9",
    "Eacute;": "\xc9",
    "Ecaron;": "\u011a",
    "Ecirc": "\xca",
    "Ecirc;": "\xca",
    "Ecy;": "\u042d",
    "Edot;": "\u0116",
    "Efr;": "\U0001d508",
    "Egrave": "\xc8",
    "Egrave;": "\xc8",
    "Element;": "\u2208",
    "Emacr;": "\u0112",
    "EmptySmallSquare;": "\u25fb",
    "EmptyVerySmallSquare;": "\u25ab",
    "Eogon;": "\u0118",
    "Eopf;": "\U0001d53c",
    "Epsilon;": "\u0395",
    "Equal;": "\u2a75",
    "EqualTilde;": "\u2242",
    "Equilibrium;": "\u21cc",
    "Escr;": "\u2130",
    "Esim;": "\u2a73",
    "Eta;": "\u0397",
    "Euml": "\xcb",
    "Euml;": "\xcb",
    "Exists;": "\u2203",
    "ExponentialE;": "\u2147",
    "Fcy;": "\u0424",
    "Ffr;": "\U0001d509",
    "FilledSmallSquare;": "\u25fc",
    "FilledVerySmallSquare;": "\u25aa",
    "Fopf;": "\U0001d53d",
    "ForAll;": "\u2200",
    "Fouriertrf;": "\u2131",
    "Fscr;": "\u2131",
    "GJcy;": "\u0403",
    "GT": ">",
    "GT;": ">",
    "Gamma;": "\u0393",
    "Gammad;": "\u03dc",
    "Gbreve;": "\u011e",
    "Gcedil;": "\u0122",
    "Gcirc;": "\u011c",
    "Gcy;": "\u0413",
    "Gdot;": "\u0120",
    "Gfr;": "\U0001d50a",
    "Gg;": "\u22d9",
    "Gopf;": "\U0001d53e",
    "GreaterEqual;": "\u2265",
    "GreaterEqualLess;": "\u22db",
    "GreaterFullEqual;": "\u2267",
    "GreaterGreater;": "\u2aa2",
    "GreaterLess;": "\u2277",
    "GreaterSlantEqual;": "\u2a7e",
    "GreaterTilde;": "\u2273",
    "Gscr;": "\U0001d4a2",
    "Gt;": "\u226b",
    "HARDcy;": "\u042a",
    "Hacek;": "\u02c7",
    "Hat;": "^",
    "Hcirc;": "\u0124",
    "Hfr;": "\u210c",
    "HilbertSpace;": "\u210b",
    "Hopf;": "\u210d",
    "HorizontalLine;": "\u2500",
    "Hscr;": "\u210b",
    "Hstrok;": "\u0126",
    "HumpDownHump;": "\u224e",
    "HumpEqual;": "\u224f",
    "IEcy;": "\u0415",
    "IJlig;": "\u0132",
    "IOcy;": "\u0401",
    "Iacute": "\xcd",
    "Iacute;": "\xcd",
    "Icirc": "\xce",
    "Icirc;": "\xce",
    "Icy;": "\u0418",
    "Idot;": "\u0130",
    "Ifr;": "\u2111",
    "Igrave": "\xcc",
    "Igrave;": "\xcc",
    "Im;": "\u2111",
    "Imacr;": "\u012a",
    "ImaginaryI;": "\u2148",
    "Implies;": "\u21d2",
    "Int;": "\u222c",
    "Integral;": "\u222b",
    "Intersection;": "\u22c2",
    "InvisibleComma;": "\u2063",
    "InvisibleTimes;": "\u2062",
    "Iogon;": "\u012e",
    "Iopf;": "\U0001d540",
    "Iota;": "\u0399",
    "Iscr;": "\u2110",
    "Itilde;": "\u0128",
    "Iukcy;": "\u0406",
    "Iuml": "\xcf",
    "Iuml;": "\xcf",
    "Jcirc;": "\u0134",
    "Jcy;": "\u0419",
    "Jfr;": "\U0001d50d",
    "Jopf;": "\U0001d541",
    "Jscr;": "\U0001d4a5",
    "Jsercy;": "\u0408",
    "Jukcy;": "\u0404",
    "KHcy;": "\u0425",
    "KJcy;": "\u040c",
    "Kappa;": "\u039a",
    "Kcedil;": "\u0136",
    "Kcy;": "\u041a",
    "Kfr;": "\U0001d50e",
    "Kopf;": "\U0001d542",
    "Kscr;": "\U0001d4a6",
    "LJcy;": "\u0409",
    "LT": "<",
    "LT;": "<",
    "Lacute;": "\u0139",
    "Lambda;": "\u039b",
    "Lang;": "\u27ea",
    "Laplacetrf;": "\u2112",
    "Larr;": "\u219e",
    "Lcaron;": "\u013d",
    "Lcedil;": "\u013b",
    "Lcy;": "\u041b",
    "LeftAngleBracket;": "\u27e8",
    "LeftArrow;": "\u2190",
    "LeftArrowBar;": "\u21e4",
    "LeftArrowRightArrow;": "\u21c6",
    "LeftCeiling;": "\u2308",
    "LeftDoubleBracket;": "\u27e6",
    "LeftDownTeeVector;": "\u2961",
    "LeftDownVector;": "\u21c3",
    "LeftDownVectorBar;": "\u2959",
    "LeftFloor;": "\u230a",
    "LeftRightArrow;": "\u2194",
    "LeftRightVector;": "\u294e",
    "LeftTee;": "\u22a3",
    "LeftTeeArrow;": "\u21a4",
    "LeftTeeVector;": "\u295a",
    "LeftTriangle;": "\u22b2",
    "LeftTriangleBar;": "\u29cf",
    "LeftTriangleEqual;": "\u22b4",
    "LeftUpDownVector;": "\u2951",
    "LeftUpTeeVector;": "\u2960",
    "LeftUpVector;": "\u21bf",
    "LeftUpVectorBar;": "\u2958",
    "LeftVector;": "\u21bc",
    "LeftVectorBar;": "\u2952",
    "Leftarrow;": "\u21d0",
    "Leftrightarrow;": "\u21d4",
    "LessEqualGreater;": "\u22da",
    "LessFullEqual;": "\u2266",
    "LessGreater;": "\u2276",
    "LessLess;": "\u2aa1",
    "LessSlantEqual;": "\u2a7d",
    "LessTilde;": "\u2272",
    "Lfr;": "\U0001d50f",
    "Ll;": "\u22d8",
    "Lleftarrow;": "\u21da",
    "Lmidot;": "\u013f",
    "LongLeftArrow;": "\u27f5",
    "LongLeftRightArrow;": "\u27f7",
    "LongRightArrow;": "\u27f6",
    "Longleftarrow;": "\u27f8",
    "Longleftrightarrow;": "\u27fa",
    "Longrightarrow;": "\u27f9",
    "Lopf;": "\U0001d543",
    "LowerLeftArrow;": "\u2199",
    "LowerRightArrow;": "\u2198",
    "Lscr;": "\u2112",
    "Lsh;": "\u21b0",
    "Lstrok;": "\u0141",
    "Lt;": "\u226a",
    "Map;": "\u2905",
    "Mcy;": "\u041c",
    "MediumSpace;": "\u205f",
    "Mellintrf;": "\u2133",
    "Mfr;": "\U0001d510",
    "MinusPlus;": "\u2213",
    "Mopf;": "\U0001d544",
    "Mscr;": "\u2133",
    "Mu;": "\u039c",
    "NJcy;": "\u040a",
    "Nacute;": "\u0143",
    "Ncaron;": "\u0147",
    "Ncedil;": "\u0145",
    "Ncy;": "\u041d",
    "NegativeMediumSpace;": "\u200b",
    "NegativeThickSpace;": "\u200b",
    "NegativeThinSpace;": "\u200b",
    "NegativeVeryThinSpace;": "\u200b",
    "NestedGreaterGreater;": "\u226b",
    "NestedLessLess;": "\u226a",
    "NewLine;": "\n",
    "Nfr;": "\U0001d511",
    "NoBreak;": "\u2060",
    "NonBreakingSpace;": "\xa0",
    "Nopf;": "\u2115",
    "Not;": "\u2aec",
    "NotCongruent;": "\u2262",
    "NotCupCap;": "\u226d",
    "NotDoubleVerticalBar;": "\u2226",
    "NotElement;": "\u2209",
    "NotEqual;": "\u2260",
    "NotEqualTilde;": "\u2242\u0338",
    "NotExists;": "\u2204",
    "NotGreater;": "\u226f",
    "NotGreaterEqual;": "\u2271",
    "NotGreaterFullEqual;": "\u2267\u0338",
    "NotGreaterGreater;": "\u226b\u0338",
    "NotGreaterLess;": "\u2279",
    "NotGreaterSlantEqual;": "\u2a7e\u0338",
    "NotGreaterTilde;": "\u2275",
    "NotHumpDownHump;": "\u224e\u0338",
    "NotHumpEqual;": "\u224f\u0338",
    "NotLeftTriangle;": "\u22ea",
    "NotLeftTriangleBar;": "\u29cf\u0338",
    "NotLeftTriangleEqual;": "\u22ec",
    "NotLess;": "\u226e",
    "NotLessEqual;": "\u2270",
    "NotLessGreater;": "\u2278",
    "NotLessLess;": "\u226a\u0338",
    "NotLessSlantEqual;": "\u2a7d\u0338",
    "NotLessTilde;": "\u2274",
    "NotNestedGreaterGreater;": "\u2aa2\u0338",
    "NotNestedLessLess;": "\u2aa1\u0338",
    "NotPrecedes;": "\u2280",
    "NotPrecedesEqual;": "\u2aaf\u0338",
    "NotPrecedesSlantEqual;": "\u22e0",
    "NotReverseElement;": "\u220c",
    "NotRightTriangle;": "\u22eb",
    "NotRightTriangleBar;": "\u29d0\u0338",
    "NotRightTriangleEqual;": "\u22ed",
    "NotSquareSubset;": "\u228f\u0338",
    "NotSquareSubsetEqual;": "\u22e2",
    "NotSquareSuperset;": "\u2290\u0338",
    "NotSquareSupersetEqual;": "\u22e3",
    "NotSubset;": "\u2282\u20d2",
    "NotSubsetEqual;": "\u2288",
    "NotSucceeds;": "\u2281",
    "NotSucceedsEqual;": "\u2ab0\u0338",
    "NotSucceedsSlantEqual;": "\u22e1",
    "NotSucceedsTilde;": "\u227f\u0338",
    "NotSuperset;": "\u2283\u20d2",
    "NotSupersetEqual;": "\u2289",
    "NotTilde;": "\u2241",
    "NotTildeEqual;": "\u2244",
    "NotTildeFullEqual;": "\u2247",
    "NotTildeTilde;": "\u2249",
    "NotVerticalBar;": "\u2224",
    "Nscr;": "\U0001d4a9",
    "Ntilde": "\xd1",
    "Ntilde;": "\xd1",
    "Nu;": "\u039d",
    "OElig;": "\u0152",
    "Oacute": "\xd3",
    "Oacute;": "\xd3",
    "Ocirc": "\xd4",
    "Ocirc;": "\xd4",
    "Ocy;": "\u041e",
    "Odblac;": "\u0150",
    "Ofr;": "\U0001d512",
    "Ograve": "\xd2",
    "Ograve;": "\xd2",
    "Omacr;": "\u014c",
    "Omega;": "\u03a9",
    "Omicron;": "\u039f",
    "Oopf;": "\U0001d546",
    "OpenCurlyDoubleQuote;": "\u201c",
    "OpenCurlyQuote;": "\u2018",
    "Or;": "\u2a54",
    "Oscr;": "\U0001d4aa",
    "Oslash": "\xd8",
    "Oslash;": "\xd8",
    "Otilde": "\xd5",
    "Otilde;": "\xd5",
    "Otimes;": "\u2a37",
    "Ouml": "\xd6",
    "Ouml;": "\xd6",
    "OverBar;": "\u203e",
    "OverBrace;": "\u23de",
    "OverBracket;": "\u23b4",
    "OverParenthesis;": "\u23dc",
    "PartialD;": "\u2202",
    "Pcy;": "\u041f",
    "Pfr;": "\U0001d513",
    "Phi;": "\u03a6",
    "Pi;": "\u03a0",
    "PlusMinus;": "\xb1",
    "Poincareplane;": "\u210c",
    "Popf;": "\u2119",
    "Pr;": "\u2abb",
    "Precedes;": "\u227a",
    "PrecedesEqual;": "\u2aaf",
    "PrecedesSlantEqual;": "\u227c",
    "PrecedesTilde;": "\u227e",
    "Prime;": "\u2033",
    "Product;": "\u220f",
    "Proportion;": "\u2237",
    "Proportional;": "\u221d",
    "Pscr;": "\U0001d4ab",
    "Psi;": "\u03a8",
    "QUOT": "\"",
    "QUOT;": "\"",
    "Qfr;": "\U0001d514",
    "Qopf;": "\u211a",
    "Qscr;": "\U0001d4ac",
    "RBarr;": "\u2910",
    "REG": "\xae",
    "REG;": "\xae",
    "Racute;": "\u0154",
    "Rang;": "\u27eb",
    "Rarr;": "\u21a0",
    "Rarrtl;": "\u2916",
    "Rcaron;": "\u0158",
    "Rcedil;": "\u0156",
    "Rcy;": "\u0420",
    "Re;": "\u211c",
    "ReverseElement;": "\u220b",
    "ReverseEquilibrium;": "\u21cb",
    "ReverseUpEquilibrium;": "\u296f",
    "Rfr;": "\u211c",
    "Rho;": "\u03a1",
    "RightAngleBracket;": "\u27e9",
    "RightArrow;": "\u2192",
    "RightArrowBar;": "\u21e5",
    "RightArrowLeftArrow;": "\u21c4",
    "RightCeiling;": "\u2309",
    "RightDoubleBracket;": "\u27e7",
    "RightDownTeeVector;": "\u295d",
    "RightDownVector;": "\u21c2",
    "RightDownVectorBar;": "\u2955",
    "RightFloor;": "\u230b",
    "RightTee;": "\u22a2",
    "RightTeeArrow;": "\u21a6",
    "RightTeeVector;": "\u295b",
    "RightTriangle;": "\u22b3",
    "RightTriangleBar;": "\u29d0",
    "RightTriangleEqual;": "\u22b5",
    "RightUpDownVector;": "\u294f",
    "RightUpTeeVector;": "\u295c",
    "RightUpVector;": "\u21be",
    "RightUpVectorBar;": "\u2954",
    "RightVector;": "\u21c0",
    "RightVectorBar;": "\u2953",
    "Rightarrow;": "\u21d2",
    "Ropf;": "\u211d",
    "RoundImplies;": "\u2970",
    "Rrightarrow;": "\u21db",
    "Rscr;": "\u211b",
    "Rsh;": "\u21b1",
    "RuleDelayed;": "\u29f4",
    "SHCHcy;": "\u0429",
    "SHcy;": "\u0428",
    "SOFTcy;": "\u042c",
    "Sacute;": "\u015a",
    "Sc;": "\u2abc",
    "Scaron;": "\u0160",
    "Scedil;": "\u015e",
    "Scirc;": "\u015c",
    "Scy;": "\u0421",
    "Sfr;": "\U0001d516",
    "ShortDownArrow;": "\u2193",
    "ShortLeftArrow;": "\u2190",
    "ShortRightArrow;": "\u2192",
    "ShortUpArrow;": "\u2191",
    "Sigma;": "\u03a3",
    "SmallCircle;": "\u2218",
    "Sopf;": "\U0001d54a",
    "Sqrt;": "\u221a",
    "Square;": "\u25a1",
    "SquareIntersection;": "\u2293",
    "SquareSubset;": "\u228f",
    "SquareSubsetEqual;": "\u2291",
    "SquareSuperset;": "\u2290",
    "SquareSupersetEqual;": "\u2292",
    "SquareUnion;": "\u2294",
    "Sscr;": "\U0001d4ae",
    "Star;": "\u22c6",
    "Sub;": "\u22d0",
    "Subset;": "\u22d0",
    "SubsetEqual;": "\u2286",
    "Succeeds;": "\u227b",
    "SucceedsEqual;": "\u2ab0",
    "SucceedsSlantEqual;": "\u227d",
    "SucceedsTilde;": "\u227f",
    "SuchThat;": "\u220b",
    "Sum;": "\u2211",
    "Sup;": "\u22d1",
    "Superset;": "\u2283",
    "SupersetEqual;": "\u2287",
    "Supset;": "\u22d1",
    "THORN": "\xde",
    "THORN;": "\xde",
    "TRADE;": "\u2122",
    "TSHcy;": "\u040b",
    "TScy;": "\u0426",
    "Tab;": "\t",
    "Tau;": "\u03a4",
    "Tcaron;": "\u0164",
    "Tcedil;": "\u0162",
    "Tcy;": "\u0422",
    "Tfr;": "\U0001d517",
    "Therefore;": "\u2234",
    "Theta;": "\u0398",
    "ThickSpace;": "\u205f\u200a",
    "ThinSpace;": "\u2009",
    "Tilde;": "\u223c",
    "TildeEqual;": "\u2243",
    "TildeFullEqual;": "\u2245",
    "TildeTilde;": "\u2248",
    "Topf;": "\U0001d54b",
    "TripleDot;": "\u20db",
    "Tscr;": "\U0001d4af",
    "Tstrok;": "\u0166",
    "Uacute": "\xda",
    "Uacute;": "\xda",
    "Uarr;": "\u219f",
    "Uarrocir;": "\u2949",
    "Ubrcy;": "\u040e",
    "Ubreve;": "\u016c",
    "Ucirc": "\xdb",
    "Ucirc;": "\xdb",
    "Ucy;": "\u0423",
    "Udblac;": "\u0170",
    "Ufr;": "\U0001d518",
    "Ugrave": "\xd9",
    "Ugrave;": "\xd9",
    "Umacr;": "\u016a",
    "UnderBar;": "_",
    "UnderBrace;": "\u23df",
    "UnderBracket;": "\u23b5",
    "UnderParenthesis;": "\u23dd",
    "Union;": "\u22c3",
    "UnionPlus;": "\u228e",
    "Uogon;": "\u0172",
    "Uopf;": "\U0001d54c",
    "UpArrow;": "\u2191",
    "UpArrowBar;": "\u2912",
    "UpArrowDownArrow;": "\u21c5",
    "UpDownArrow;": "\u2195",
    "UpEquilibrium;": "\u296e",
    "UpTee;": "\u22a5",
    "UpTeeArrow;": "\u21a5",
    "Uparrow;": "\u21d1",
    "Updownarrow;": "\u21d5",
    "UpperLeftArrow;": "\u2196",
    "UpperRightArrow;": "\u2197",
    "Upsi;": "\u03d2",
    "Upsilon;": "\u03a5",
    "Uring;": "\u016e",
    "Uscr;": "\U0001d4b0",
    "Utilde;": "\u0168",
    "Uuml": "\xdc",
    "Uuml;": "\xdc",
    "VDash;": "\u22ab",
    "Vbar;": "\u2aeb",
    "Vcy;": "\u0412",
    "Vdash;": "\u22a9",
    "Vdashl;": "\u2ae6",
    "Vee;": "\u22c1",
    "Verbar;": "\u2016",
    "Vert;": "\u2016",
    "VerticalBar;": "\u2223",
    "VerticalLine;": "|",
    "VerticalSeparator;": "\u2758",
    "VerticalTilde;": "\u2240",
    "VeryThinSpace;": "\u200a",
    "Vfr;": "\U0001d519",
    "Vopf;": "\U0001d54d",
    "Vscr;": "\U0001d4b1",
    "Vvdash;": "\u22aa",
    "Wcirc;": "\u0174",
    "Wedge;": "\u22c0",
    "Wfr;": "\U0001d51a",
    "Wopf;": "\U0001d54e",
    "Wscr;": "\U0001d4b2",
    "Xfr;": "\U0001d51b",
    "Xi;": "\u039e",
    "Xopf;": "\U0001d54f",
    "Xscr;": "\U0001d4b3",
    "YAcy;": "\u042f",
    "YIcy;": "\u0407",
    "YUcy;": "\u042e",
    "Yacute": "\xdd",
    "Yacute;": "\xdd",
    "Ycirc;": "\u0176",
    "Ycy;": "\u042b",
    "Yfr;": "\U0001d51c",
    "Yopf;": "\U0001d550",
    "Yscr;": "\U0001d4b4",
    "Yuml;": "\u0178",
    "ZHcy;": "\u0416",
    "Zacute;": "\u0179",
    "Zcaron;": "\u017d",
    "Zcy;": "\u0417",
    "Zdot;": "\u017b",
    "ZeroWidthSpace;": "\u200b",
    "Zeta;": "\u0396",
    "Zfr;": "\u2128",
    "Zopf;": "\u2124",
    "Zscr;": "\U0001d4b5",
    "aacute": "\xe1",
    "aacute;": "\xe1",
    "abreve;": "\u0103",
    "ac;": "\u223e",
    "acE;": "\u223e\u0333",
    "acd;": "\u223f",
    "acirc": "\xe2",
    "acirc;": "\xe2",
    "acute": "\xb4",
    "acute;": "\xb4",
    "acy;": "\u0430",
    "aelig": "\xe6",
    "aelig;": "\xe6",
    "af;": "\u2061",
    "afr;": "\U0001d51e",
    "agrave": "\xe0",
    "agrave;": "\xe0",
    "alefsym;": "\u2135",
    "aleph;": "\u2135",
    "alpha;": "\u03b1",
    "amacr;": "\u0101",
    "amalg;": "\u2a3f",
    "amp": "&",
    "amp;": "&",
    "and;": "\u2227",
    "andand;": "\u2a55",
    "andd;": "\u2a5c",
    "andslope;": "\u2a58",
    "andv;": "\u2a5a",
    "ang;": "\u2220",
    "ange;": "\u29a4",
    "angle;": "\u2220",
    "angmsd;": "\u2221",
    "angmsdaa;": "\u29a8",
    "angmsdab;": "\u29a9",
    "angmsdac;": "\u29aa",
    "angmsdad;": "\u29ab",
    "angmsdae;": "\u29ac",
    "angmsdaf;": "\u29ad",
    "angmsdag;": "\u29ae",
    "angmsdah;": "\u29af",
    "angrt;": "\u221f",
    "angrtvb;": "\u22be",
    "angrtvbd;": "\u299d",
    "angsph;": "\u2222",
    "angst;": "\xc5",
    "angzarr;": "\u237c",
    "aogon;": "\u0105",
    "aopf;": "\U0001d552",
    "ap;": "\u2248",
    "apE;": "\u2a70",
    "apacir;": "\u2a6f",
    "ape;": "\u224a",
    "apid;": "\u224b",
    "apos;": "'",
    "approx;": "\u2248",
    "approxeq;": "\u224a",
    "aring": "\xe5",
    "aring;": "\xe5",
    "ascr;": "\U0001d4b6",
    "ast;": "*",
    "asymp;": "\u2248",
    "asympeq;": "\u224d",
    "atilde": "\xe3",
    "atilde;": "\xe3",
    "auml": "\xe4",
    "auml;": "\xe4",
    "awconint;": "\u2233",
    "awint;": "\u2a11",
    "bNot;": "\u2aed",
    "backcong;": "\u224c",
    "backepsilon;": "\u03f6",
    "backprime;": "\u2035",
    "backsim;": "\u223d",
    "backsimeq;": "\u22cd",
    "barvee;": "\u22bd",
    "barwed;": "\u2305",
    "barwedge;": "\u2305",
    "bbrk;": "\u23b5",
    "bbrktbrk;": "\u23b6",
    "bcong;": "\u224c",
    "bcy;": "\u0431",
    "bdquo;": "\u201e",
    "becaus;": "\u2235",
    "because;": "\u2235",
    "bemptyv;": "\u29b0",
    "bepsi;": "\u03f6",
    "bernou;": "\u212c",
    "beta;": "\u03b2",
    "beth;": "\u2136",
    "between;": "\u226c",
    "bfr;": "\U0001d51f",
    "bigcap;": "\u22c2",
    "bigcirc;": "\u25ef",
    "bigcup;": "\u22c3",
    "bigodot;": "\u2a00",
    "bigoplus;": "\u2a01",
    "bigotimes;": "\u2a02",
    "bigsqcup;": "\u2a06",
    "bigstar;": "\u2605",
    "bigtriangledown;": "\u25bd",
    "bigtriangleup;": "\u25b3",
    "biguplus;": "\u2a04",
    "bigvee;": "\u22c1",
    "bigwedge;": "\u22c0",
    "bkarow;": "\u290d",
    "blacklozenge;": "\u29eb",
    "blacksquare;": "\u25aa",
    "blacktriangle;": "\u25b4",
    "blacktriangledown;": "\u25be",
    "blacktriangleleft;": "\u25c2",
    "blacktriangleright;": "\u25b8",
    "blank;": "\u2423",
    "blk12;": "\u2592",
    "blk14;": "\u2591",
    "blk34;": "\u2593",
    "block;": "\u2588",
    "bne;": "=\u20e5",
    "bnequiv;": "\u2261\u20e5",
    "bnot;": "\u2310",
    "bopf;": "\U0001d553",
    "bot;": "\u22a5",
    "bottom;": "\u22a5",
    "bowtie;": "\u22c8",
    "boxDL;": "\u2557",
    "boxDR;": "\u2554",
    "boxDl;": "\u2556",
    "boxDr;": "\u2553",
    "boxH;": "\u2550",
    "boxHD;": "\u2566",
    "boxHU;": "\u2569",
    "boxHd;": "\u2564",
    "boxHu;": "\u2567",
    "boxUL;": "\u255d",
    "boxUR;": "\u255a",
    "boxUl;": "\u255c",
    "boxUr;": "\u2559",
    "boxV;": "\u2551",
    "boxVH;": "\u256c",
    "boxVL;": "\u2563",
    "boxVR;": "\u2560",
    "boxVh;": "\u256b",
    "boxVl;": "\u2562",
    "boxVr;": "\u255f",
    "boxbox;": "\u29c9",
    "boxdL;": "\u2555",
    "boxdR;": "\u2552",
    "boxdl;": "\u2510",
    "boxdr;": "\u250c",
    "boxh;": "\u2500",
    "boxhD;": "\u2565",
    "boxhU;": "\u2568",
    "boxhd;": "\u252c",
    "boxhu;": "\u2534",
    "boxminus;": "\u229f",
    "boxplus;": "\u229e",
    "boxtimes;": "\u22a0",
    "boxuL;": "\u255b",
    "boxuR;": "\u2558",
    "boxul;": "\u2518",
    "boxur;": "\u2514",
    "boxv;": "\u2502",
    "boxvH;": "\u256a",
    "boxvL;": "\u2561",
    "boxvR;": "\u255e",
    "boxvh;": "\u253c",
    "boxvl;": "\u2524",
    "boxvr;": "\u251c",
    "bprime;": "\u2035",
    "breve;": "\u02d8",
    "brvbar": "\xa6",
    "brvbar;": "\xa6",
    "bscr;": "\U0001d4b7",
    "bsemi;": "\u204f",
    "bsim;": "\u223d",
    "bsime;": "\u22cd",
    "bsol;": "\\",
    "bsolb;": "\u29c5",
    "bsolhsub;": "\u27c8",
    "bull;": "\u2022",
    "bullet;": "\u2022",
    "bump;": "\u224e",
    "bumpE;": "\u2aae",
    "bumpe;": "\u224f",
    "bumpeq;": "\u224f",
    "cacute;": "\u0107",
    "cap;": "\u2229",
    "capand;": "\u2a44",
    "capbrcup;": "\u2a49",
    "capcap;": "\u2a4b",
    "capcup;": "\u2a47",
    "capdot;": "\u2a40",
    "caps;": "\u2229\ufe00",
    "caret;": "\u2041",
    "caron;": "\u02c7",
    "ccaps;": "\u2a4d",
    "ccaron;": "\u010d",
    "ccedil": "\xe7",
    "ccedil;": "\xe7",
    "ccirc;": "\u0109",
    "ccups;": "\u2a4c",
    "ccupssm;": "\u2a50",
    "cdot;": "\u010b",
    "cedil": "\xb8",
    "cedil;": "\xb8",
    "cemptyv;": "\u29b2",
    "cent": "\xa2",
    "cent;": "\xa2",
    "centerdot;": "\xb7",
    "cfr;": "\U0001d520",
    "chcy;": "\u0447",
    "check;": "\u2713",
    "checkmark;": "\u2713",
    "chi;": "\u03c7",
    "cir;": "\u25cb",
    "cirE;": "\u29c3",
    "circ;": "\u02c6",
    "circeq;": "\u2257",
    "circlearrowleft;": "\u21ba",
    "circlearrowright;": "\u21bb",
    "circledR;": "\xae",
    "circledS;": "\u24c8",
    "circledast;": "\u229b",
    "circledcirc;": "\u229a",
    "circleddash;": "\u229d",
    "cire;": "\u2257",
    "cirfnint;": "\u2a10",
    "cirmid;": "\u2aef",
    "cirscir;": "\u29c2",
    "clubs;": "\u2663",
    "clubsuit;": "\u2663",
    "colon;": ":",
    "colone;": "\u2254",
    "coloneq;": "\u2254",
    "comma;": ",",
    "commat;": "@",
    "comp;": "\u2201",
    "compfn;": "\u2218",
    "complement;": "\u2201",
    "complexes;": "\u2102",
    "cong;": "\u2245",
    "congdot;": "\u2a6d",
    "conint;": "\u222e",
    "copf;": "\U0001d554",
    "coprod;": "\u2210",
    "copy": "\xa9",
    "copy;": "\xa9",
    "copysr;": "\u2117",
    "crarr;": "\u21b5",
    "cross;": "\u2717",
    "cscr;": "\U0001d4b8",
    "csub;": "\u2acf",
    "csube;": "\u2ad1",
    "csup;": "\u2ad0",
    "csupe;": "\u2ad2",
    "ctdot;": "\u22ef",
    "cudarrl;": "\u2938",
    "cudarrr;": "\u2935",
    "cuepr;": "\u22de",
    "cuesc;": "\u22df",
    "cularr;": "\u21b6",
    "cularrp;": "\u293d",
    "cup;": "\u222a",
    "cupbrcap;": "\u2a48",
    "cupcap;": "\u2a46",
    "cupcup;": "\u2a4a",
    "cupdot;": "\u228d",
    "cupor;": "\u2a45",
    "cups;": "\u222a\ufe00",
    "curarr;": "\u21b7",
    "curarrm;": "\u293c",
    "curlyeqprec;": "\u22de",
    "curlyeqsucc;": "\u22df",
    "curlyvee;": "\u22ce",
    "curlywedge;": "\u22cf",
    "curren": "\xa4",
    "curren;": "\xa4",
    "curvearrowleft;": "\u21b6",
    "curvearrowright;": "\u21b7",
    "cuvee;": "\u22ce",
    "cuwed;": "\u22cf",
    "cwconint;": "\u2232",
    "cwint;": "\u2231",
    "cylcty;": "\u232d",
    "dArr;": "\u21d3",
    "dHar;": "\u2965",
    "dagger;": "\u2020",
    "daleth;": "\u2138",
    "darr;": "\u2193",
    "dash;": "\u2010",
    "dashv;": "\u22a3",
    "dbkarow;": "\u290f",
    "dblac;": "\u02dd",
    "dcaron;": "\u010f",
    "dcy;": "\u0434",
    "dd;": "\u2146",
    "ddagger;": "\u2021",
    "ddarr;": "\u21ca",
    "ddotseq;": "\u2a77",
    "deg": "\xb0",
    "deg;": "\xb0",
    "delta;": "\u03b4",
    "demptyv;": "\u29b1",
    "dfisht;": "\u297f",
    "dfr;": "\U0001d521",
    "dharl;": "\u21c3",
    "dharr;": "\u21c2",
    "diam;": "\u22c4",
    "diamond;": "\u22c4",
    "diamondsuit;": "\u2666",
    "diams;": "\u2666",
    "die;": "\xa8",
    "digamma;": "\u03dd",
    "disin;": "\u22f2",
    "div;": "\xf7",
    "divide": "\xf7",
    "divide;": "\xf7",
    "divideontimes;": "\u22c7",
    "divonx;": "\u22c7",
    "djcy;": "\u0452",
    "dlcorn;": "\u231e",
    "dlcrop;": "\u230d",
    "dollar;": "$",
    "dopf;": "\U0001d555",
    "dot;": "\u02d9",
    "doteq;": "\u2250",
    "doteqdot;": "\u2251",
    "dotminus;": "\u2238",
    "dotplus;": "\u2214",
    "dotsquare;": "\u22a1",
    "doublebarwedge;": "\u2306",
    "downarrow;": "\u2193",
    "downdownarrows;": "\u21ca",
    "downharpoonleft;": "\u21c3",
    "downharpoonright;": "\u21c2",
    "drbkarow;": "\u2910",
    "drcorn;": "\u231f",
    "drcrop;": "\u230c",
    "dscr;": "\U0001d4b9",
    "dscy;": "\u0455",
    "dsol;": "\u29f6",
    "dstrok;": "\u0111",
    "dtdot;": "\u22f1",
    "dtri;": "\u25bf",
    "dtrif;": "\u25be",
    "duarr;": "\u21f5",
    "duhar;": "\u296f",
    "dwangle;": "\u29a6",
    "dzcy;": "\u045f",
    "dzigrarr;": "\u27ff",
    "eDDot;": "\u2a77",
    "eDot;": "\u2251",
    "eacute": "\xe9",
    "eacute;": "\xe9",
    "easter;": "\u2a6e",
    "ecaron;": "\u011b",
    "ecir;": "\u2256",
    "ecirc": "\xea",
    "ecirc;": "\xea",
    "ecolon;": "\u2255",
    "ecy;": "\u044d",
    "edot;": "\u0117",
    "ee;": "\u2147",
    "efDot;": "\u2252",
    "efr;": "\U0001d522",
    "eg;": "\u2a9a",
    "egrave": "\xe8",
    "egrave;": "\xe8",
    "egs;": "\u2a96",
    "egsdot;": "\u2a98",
    "el;": "\u2a99",
    "elinters;": "\u23e7",
    "ell;": "\u2113",
    "els;": "\u2a95",
    "elsdot;": "\u2a97",
    "emacr;": "\u0113",
    "empty;": "\u2205",
    "emptyset;": "\u2205",
    "emptyv;": "\u2205",
    "emsp13;": "\u2004",
    "emsp14;": "\u2005",
    "emsp;": "\u2003",
    "eng;": "\u014b",
    "ensp;": "\u2002",
    "eogon;": "\u0119",
    "eopf;": "\U0001d556",
    "epar;": "\u22d5",
    "eparsl;": "\u29e3",
    "eplus;": "\u2a71",
    "epsi;": "\u03b5",
    "epsilon;": "\u03b5",
    "epsiv;": "\u03f5",
    "eqcirc;": "\u2256",
    "eqcolon;": "\u2255",
    "eqsim;": "\u2242",
    "eqslantgtr;": "\u2a96",
    "eqslantless;": "\u2a95",
    "equals;": "=",
    "equest;": "\u225f",
    "equiv;": "\u2261",
    "equivDD;": "\u2a78",
    "eqvparsl;": "\u29e5",
    "erDot;": "\u2253",
    "erarr;": "\u2971",
    "escr;": "\u212f",
    "esdot;": "\u2250",
    "esim;": "\u2242",
    "eta;": "\u03b7",
    "eth": "\xf0",
    "eth;": "\xf0",
    "euml": "\xeb",
    "euml;": "\xeb",
    "euro;": "\u20ac",
    "excl;": "!",
    "exist;": "\u2203",
    "expectation;": "\u2130",
    "exponentiale;": "\u2147",
    "fallingdotseq;": "\u2252",
    "fcy;": "\u0444",
    "female;": "\u2640",
    "ffilig;": "\ufb03",
    "fflig;": "\ufb00",
    "ffllig;": "\ufb04",
    "ffr;": "\U0001d523",
    "filig;": "\ufb01",
    "fjlig;": "fj",
    "flat;": "\u266d",
    "fllig;": "\ufb02",
    "fltns;": "\u25b1",
    "fnof;": "\u0192",
    "fopf;": "\U0001d557",
    "forall;": "\u2200",
    "fork;": "\u22d4",
    "forkv;": "\u2ad9",
    "fpartint;": "\u2a0d",
    "frac12": "\xbd",
    "frac12;": "\xbd",
    "frac13;": "\u2153",
    "frac14": "\xbc",
    "frac14;": "\xbc",
    "frac15;": "\u2155",
    "frac16;": "\u2159",
    "frac18;": "\u215b",
    "frac23;": "\u2154",
    "frac25;": "\u2156",
    "frac34": "\xbe",
    "frac34;": "\xbe",
    "frac35;": "\u2157",
    "frac38;": "\u215c",
    "frac45;": "\u2158",
    "frac56;": "\u215a",
    "frac58;": "\u215d",
    "frac78;": "\u215e",
    "frasl;": "\u2044",
    "frown;": "\u2322",
    "fscr;": "\U0001d4bb",
    "gE;": "\u2267",
    "gEl;": "\u2a8c",
    "gacute;": "\u01f5",
    "gamma;": "\u03b3",
    "gammad;": "\u03dd",
    "gap;": "\u2a86",
    "gbreve;": "\u011f",
    "gcirc;": "\u011d",
    "gcy;": "\u0433",
    "gdot;": "\u0121",
    "ge;": "\u2265",
    "gel;": "\u22db",
    "geq;": "\u2265",
    "geqq;": "\u2267",
    "geqslant;": "\u2a7e",
    "ges;": "\u2a7e",
    "gescc;": "\u2aa9",
    "gesdot;": "\u2a80",
    "gesdoto;": "\u2a82",
    "gesdotol;": "\u2a84",
    "gesl;": "\u22db\ufe00",
    "gesles;": "\u2a94",
    "gfr;": "\U0001d524",
    "gg;": "\u226b",
    "ggg;": "\u22d9",
    "gimel;": "\u2137",
    "gjcy;": "\u0453",
    "gl;": "\u2277",
    "glE;": "\u2a92",
    "gla;": "\u2aa5",
    "glj;": "\u2aa4",
    "gnE;": "\u2269",
    "gnap;": "\u2a8a",
    "gnapprox;": "\u2a8a",
    "gne;": "\u2a88",
    "gneq;": "\u2a88",
    "gneqq;": "\u2269",
    "gnsim;": "\u22e7",
    "gopf;": "\U0001d558",
    "grave;": "`",
    "gscr;": "\u210a",
    "gsim;": "\u2273",
    "gsime;": "\u2a8e",
    "gsiml;": "\u2a90",
    "gt": ">",
    "gt;": ">",
    "gtcc;": "\u2aa7",
    "gtcir;": "\u2a7a",
    "gtdot;": "\u22d7",
    "gtlPar;": "\u2995",
    "gtquest;": "\u2a7c",
    "gtrapprox;": "\u2a86",
    "gtrarr;": "\u2978",
    "gtrdot;": "\u22d7",
    "gtreqless;": "\u22db",
    "gtreqqless;": "\u2a8c",
    "gtrless;": "\u2277",
    "gtrsim;": "\u2273",
    "gvertneqq;": "\u2269\ufe00",
    "gvnE;": "\u2269\ufe00",
    "hArr;": "\u21d4",
    "hairsp;": "\u200a",
    "half;": "\xbd",
    "hamilt;": "\u210b",
    "hardcy;": "\u044a",
    "harr;": "\u2194",
    "harrcir;": "\u2948",
    "harrw;": "\u21ad",
    "hbar;": "\u210f",
    "hcirc;": "\u0125",
    "hearts;": "\u2665",
    "heartsuit;": "\u2665",
    "hellip;": "\u2026",
    "hercon;": "\u22b9",
    "hfr;": "\U0001d525",
    "hksearow;": "\u2925",
    "hkswarow;": "\u2926",
    "hoarr;": "\u21ff",
    "homtht;": "\u223b",
    "hookleftarrow;": "\u21a9",
    "hookrightarrow;": "\u21aa",
    "hopf;": "\U0001d559",
    "horbar;": "\u2015",
    "hscr;": "\U0001d4bd",
    "hslash;": "\u210f",
    "hstrok;": "\u0127",
    "hybull;": "\u2043",
    "hyphen;": "\u2010",
    "iacute": "\xed",
    "iacute;": "\xed",
    "ic;": "\u2063",
    "icirc": "\xee",
    "icirc;": "\xee",
    "icy;": "\u0438",
    "iecy;": "\u0435",
    "iexcl": "\xa1",
    "iexcl;": "\xa1",
    "iff;": "\u21d4",
    "ifr;": "\U0001d526",
    "igrave": "\xec",
    "igrave;": "\xec",
    "ii;": "\u2148",
    "iiiint;": "\u2a0c",
    "iiint;": "\u222d",
    "iinfin;": "\u29dc",
    "iiota;": "\u2129",
    "ijlig;": "\u0133",
    "imacr;": "\u012b",
    "image;": "\u2111",
    "imagline;": "\u2110",
    "imagpart;": "\u2111",
    "imath;": "\u0131",
    "imof;": "\u22b7",
    "imped;": "\u01b5",
    "in;": "\u2208",
    "incare;": "\u2105",
    "infin;": "\u221e",
    "infintie;": "\u29dd",
    "inodot;": "\u0131",
    "int;": "\u222b",
    "intcal;": "\u22ba",
    "integers;": "\u2124",
    "intercal;": "\u22ba",
    "intlarhk;": "\u2a17",
    "intprod;": "\u2a3c",
    "iocy;": "\u0451",
    "iogon;": "\u012f",
    "iopf;": "\U0001d55a",
    "iota;": "\u03b9",
    "iprod;": "\u2a3c",
    "iquest": "\xbf",
    "iquest;": "\xbf",
    "iscr;": "\U0001d4be",
    "isin;": "\u2208",
    "isinE;": "\u22f9",
    "isindot;": "\u22f5",
    "isins;": "\u22f4",
    "isinsv;": "\u22f3",
    "isinv;": "\u2208",
    "it;": "\u2062",
    "itilde;": "\u0129",
    "iukcy;": "\u0456",
    "iuml": "\xef",
    "iuml;": "\xef",
    "jcirc;": "\u0135",
    "jcy;": "\u0439",
    "jfr;": "\U0001d527",
    "jmath;": "\u0237",
    "jopf;": "\U0001d55b",
    "jscr;": "\U0001d4bf",
    "jsercy;": "\u0458",
    "jukcy;": "\u0454",
    "kappa;": "\u03ba",
    "kappav;": "\u03f0",
    "kcedil;": "\u0137",
    "kcy;": "\u043a",
    "kfr;": "\U0001d528",
    "kgreen;": "\u0138",
    "khcy;": "\u0445",
    "kjcy;": "\u045c",
    "kopf;": "\U0001d55c",
    "kscr;": "\U0001d4c0",
    "lAarr;": "\u21da",
    "lArr;": "\u21d0",
    "lAtail;": "\u291b",
    "lBarr;": "\u290e",
    "lE;": "\u2266",
    "lEg;": "\u2a8b",
    "lHar;": "\u2962",
    "lacute;": "\u013a",
    "laemptyv;": "\u29b4",
    "lagran;": "\u2112",
    "lambda;": "\u03bb",
    "lang;": "\u27e8",
    "langd;": "\u2991",
    "langle;": "\u27e8",
    "lap;": "\u2a85",
    "laquo": "\xab",
    "laquo;": "\xab",
    "larr;": "\u2190",
    "larrb;": "\u21e4",
    "larrbfs;": "\u291f",
    "larrfs;": "\u291d",
    "larrhk;": "\u21a9",
    "larrlp;": "\u21ab",
    "larrpl;": "\u2939",
    "larrsim;": "\u2973",
    "larrtl;": "\u21a2",
    "lat;": "\u2aab",
    "latail;": "\u2919",
    "late;": "\u2aad",
    "lates;": "\u2aad\ufe00",
    "lbarr;": "\u290c",
    "lbbrk;": "\u2772",
    "lbrace;": "{",
    "lbrack;": "[",
    "lbrke;": "\u298b",
    "lbrksld;": "\u298f",
    "lbrkslu;": "\u298d",
    "lcaron;": "\u013e",
    "lcedil;": "\u013c",
    "lceil;": "\u2308",
    "lcub;": "{",
    "lcy;": "\u043b",
    "ldca;": "\u2936",
    "ldquo;": "\u201c",
    "ldquor;": "\u201e",
    "ldrdhar;": "\u2967",
    "ldrushar;": "\u294b",
    "ldsh;": "\u21b2",
    "le;": "\u2264",
    "leftarrow;": "\u2190",
    "leftarrowtail;": "\u21a2",
    "leftharpoondown;": "\u21bd",
    "leftharpoonup;": "\u21bc",
    "leftleftarrows;": "\u21c7",
    "leftrightarrow;": "\u2194",
    "leftrightarrows;": "\u21c6",
    "leftrightharpoons;": "\u21cb",
    "leftrightsquigarrow;": "\u21ad",
    "leftthreetimes;": "\u22cb",
    "leg;": "\u22da",
    "leq;": "\u2264",
    "leqq;": "\u2266",
    "leqslant;": "\u2a7d",
    "les;": "\u2a7d",
    "lescc;": "\u2aa8",
    "lesdot;": "\u2a7f",
    "lesdoto;": "\u2a81",
    "lesdotor;": "\u2a83",
    "lesg;": "\u22da\ufe00",
    "lesges;": "\u2a93",
    "lessapprox;": "\u2a85",
    "lessdot;": "\u22d6",
    "lesseqgtr;": "\u22da",
    "lesseqqgtr;": "\u2a8b",
    "lessgtr;": "\u2276",
    "lesssim;": "\u2272",
    "lfisht;": "\u297c",
    "lfloor;": "\u230a",
    "lfr;": "\U0001d529",
    "lg;": "\u2276",
    "lgE;": "\u2a91",
    "lhard;": "\u21bd",
    "lharu;": "\u21bc",
    "lharul;": "\u296a",
    "lhblk;": "\u2584",
    "ljcy;": "\u0459",
    "ll;": "\u226a",
    "llarr;": "\u21c7",
    "llcorner;": "\u231e",
    "llhard;": "\u296b",
    "lltri;": "\u25fa",
    "lmidot;": "\u0140",
    "lmoust;": "\u23b0",
    "lmoustache;": "\u23b0",
    "lnE;": "\u2268",
    "lnap;": "\u2a89",
    "lnapprox;": "\u2a89",
    "lne;": "\u2a87",
    "lneq;": "\u2a87",
    "lneqq;": "\u2268",
    "lnsim;": "\u22e6",
    "loang;": "\u27ec",
    "loarr;": "\u21fd",
    "lobrk;": "\u27e6",
    "longleftarrow;": "\u27f5",
    "longleftrightarrow;": "\u27f7",
    "longmapsto;": "\u27fc",
    "longrightarrow;": "\u27f6",
    "looparrowleft;": "\u21ab",
    "looparrowright;": "\u21ac",
    "lopar;": "\u2985",
    "lopf;": "\U0001d55d",
    "loplus;": "\u2a2d",
    "lotimes;": "\u2a34",
    "lowast;": "\u2217",
    "lowbar;": "_",
    "loz;": "\u25ca",
    "lozenge;": "\u25ca",
    "lozf;": "\u29eb",
    "lpar;": "(",
    "lparlt;": "\u2993",
    "lrarr;": "\u21c6",
    "lrcorner;": "\u231f",
    "lrhar;": "\u21cb",
    "lrhard;": "\u296d",
    "lrm;": "\u200e",
    "lrtri;": "\u22bf",
    "lsaquo;": "\u2039",
    "lscr;": "\U0001d4c1",
    "lsh;": "\u21b0",
    "lsim;": "\u2272",
    "lsime;": "\u2a8d",
    "lsimg;": "\u2a8f",
    "lsqb;": "[",
    "lsquo;": "\u2018",
    "lsquor;": "\u201a",
    "lstrok;": "\u0142",
    "lt": "<",
    "lt;": "<",
    "ltcc;": "\u2aa6",
    "ltcir;": "\u2a79",
    "ltdot;": "\u22d6",
    "lthree;": "\u22cb",
    "ltimes;": "\u22c9",
    "ltlarr;": "\u2976",
    "ltquest;": "\u2a7b",
    "ltrPar;": "\u2996",
    "ltri;": "\u25c3",
    "ltrie;": "\u22b4",
    "ltrif;": "\u25c2",
    "lurdshar;": "\u294a",
    "luruhar;": "\u2966",
    "lvertneqq;": "\u2268\ufe00",
    "lvnE;": "\u2268\ufe00",
    "mDDot;": "\u223a",
    "macr": "\xaf",
    "macr;": "\xaf",
    "male;": "\u2642",
    "malt;": "\u2720",
    "maltese;": "\u2720",
    "map;": "\u21a6",
    "mapsto;": "\u21a6",
    "mapstodown;": "\u21a7",
    "mapstoleft;": "\u21a4",
    "mapstoup;": "\u21a5",
    "marker;": "\u25ae",
    "mcomma;": "\u2a29",
    "mcy;": "\u043c",
    "mdash;": "\u2014",
    "measuredangle;": "\u2221",
    "mfr;": "\U0001d52a",
    "mho;": "\u2127",
    "micro": "\xb5",
    "micro;": "\xb5",
    "mid;": "\u2223",
    "midast;": "*",
    "midcir;": "\u2af0",
    "middot": "\xb7",
    "middot;": "\xb7",
    "minus;": "\u2212",
    "minusb;": "\u229f",
    "minusd;": "\u2238",
    "minusdu;": "\u2a2a",
    "mlcp;": "\u2adb",
    "mldr;": "\u2026",
    "mnplus;": "\u2213",
    "models;": "\u22a7",
    "mopf;": "\U0001d55e",
    "mp;": "\u2213",
    "mscr;": "\U0001d4c2",
    "mstpos;": "\u223e",
    "mu;": "\u03bc",
    "multimap;": "\u22b8",
    "mumap;": "\u22b8",
    "nGg;": "\u22d9\u0338",
    "nGt;": "\u226b\u20d2",
    "nGtv;": "\u226b\u0338",
    "nLeftarrow;": "\u21cd",
    "nLeftrightarrow;": "\u21ce",
    "nLl;": "\u22d8\u0338",
    "nLt;": "\u226a\u20d2",
    "nLtv;": "\u226a\u0338",
    "nRightarrow;": "\u21cf",
    "nVDash;": "\u22af",
    "nVdash;": "\u22ae",
    "nabla;": "\u2207",
    "nacute;": "\u0144",
    "nang;": "\u2220\u20d2",
    "nap;": "\u2249",
    "napE;": "\u2a70\u0338",
    "napid;": "\u224b\u0338",
    "napos;": "\u0149",
    "napprox;": "\u2249",
    "natur;": "\u266e",
    "natural;": "\u266e",
    "naturals;": "\u2115",
    "nbsp": "\xa0",
    "nbsp;": "\xa0",
    "nbump;": "\u224e\u0338",
    "nbumpe;": "\u224f\u0338",
    "ncap;": "\u2a43",
    "ncaron;": "\u0148",
    "ncedil;": "\u0146",
    "ncong;": "\u2247",
    "ncongdot;": "\u2a6d\u0338",
    "ncup;": "\u2a42",
    "ncy;": "\u043d",
    "ndash;": "\u2013",
    "ne;": "\u2260",
    "neArr;": "\u21d7",
    "nearhk;": "\u2924",
    "nearr;": "\u2197",
    "nearrow;": "\u2197",
    "nedot;": "\u2250\u0338",
    "nequiv;": "\u2262",
    "nesear;": "\u2928",
    "nesim;": "\u2242\u0338",
    "nexist;": "\u2204",
    "nexists;": "\u2204",
    "nfr;": "\U0001d52b",
    "ngE;": "\u2267\u0338",
    "nge;": "\u2271",
    "ngeq;": "\u2271",
    "ngeqq;": "\u2267\u0338",
    "ngeqslant;": "\u2a7e\u0338",
    "nges;": "\u2a7e\u0338",
    "ngsim;": "\u2275",
    "ngt;": "\u226f",
    "ngtr;": "\u226f",
    "nhArr;": "\u21ce",
    "nharr;": "\u21ae",
    "nhpar;": "\u2af2",
    "ni;": "\u220b",
    "nis;": "\u22fc",
    "nisd;": "\u22fa",
    "niv;": "\u220b",
    "njcy;": "\u045a",
    "nlArr;": "\u21cd",
    "nlE;": "\u2266\u0338",
    "nlarr;": "\u219a",
    "nldr;": "\u2025",
    "nle;": "\u2270",
    "nleftarrow;": "\u219a",
    "nleftrightarrow;": "\u21ae",
    "nleq;": "\u2270",
    "nleqq;": "\u2266\u0338",
    "nleqslant;": "\u2a7d\u0338",
    "nles;": "\u2a7d\u0338",
    "nless;": "\u226e",
    "nlsim;": "\u2274",
    "nlt;": "\u226e",
    "nltri;": "\u22ea",
    "nltrie;": "\u22ec",
    "nmid;": "\u2224",
    "nopf;": "\U0001d55f",
    "not": "\xac",
    "not;": "\xac",
    "notin;": "\u2209",
    "notinE;": "\u22f9\u0338",
    "notindot;": "\u22f5\u0338",
    "notinva;": "\u2209",
    "notinvb;": "\u22f7",
    "notinvc;": "\u22f6",
    "notni;": "\u220c",
    "notniva;": "\u220c",
    "notnivb;": "\u22fe",
    "notnivc;": "\u22fd",
    "npar;": "\u2226",
    "nparallel;": "\u2226",
    "nparsl;": "\u2afd\u20e5",
    "npart;": "\u2202\u0338",
    "npolint;": "\u2a14",
    "npr;": "\u2280",
    "nprcue;": "\u22e0",
    "npre;": "\u2aaf\u0338",
    "nprec;": "\u2280",
    "npreceq;": "\u2aaf\u0338",
    "nrArr;": "\u21cf",
    "nrarr;": "\u219b",
    "nrarrc;": "\u2933\u0338",
    "nrarrw;": "\u219d\u0338",
    "nrightarrow;": "\u219b",
    "nrtri;": "\u22eb",
    "nrtrie;": "\u22ed",
    "nsc;": "\u2281",
    "nsccue;": "\u22e1",
    "nsce;": "\u2ab0\u0338",
    "nscr;": "\U0001d4c3",
    "nshortmid;": "\u2224",
    "nshortparallel;": "\u2226",
    "nsim;": "\u2241",
    "nsime;": "\u2244",
    "nsimeq;": "\u2244",
    "nsmid;": "\u2224",
    "nspar;": "\u2226",
    "nsqsube;": "\u22e2",
    "nsqsupe;": "\u22e3",
    "nsub;": "\u2284",
    "nsubE;": "\u2ac5\u0338",
    "nsube;": "\u2288",
    "nsubset;": "\u2282\u20d2",
    "nsubseteq;": "\u2288",
    "nsubseteqq;": "\u2ac5\u0338",
    "nsucc;": "\u2281",
    "nsucceq;": "\u2ab0\u0338",
    "nsup;": "\u2285",
    "nsupE;": "\u2ac6\u0338",
    "nsupe;": "\u2289",
    "nsupset;": "\u2283\u20d2",
    "nsupseteq;": "\u2289",
    "nsupseteqq;": "\u2ac6\u0338",
    "ntgl;": "\u2279",
    "ntilde": "\xf1",
    "ntilde;": "\xf1",
    "ntlg;": "\u2278",
    "ntriangleleft;": "\u22ea",
    "ntrianglelefteq;": "\u22ec",
    "ntriangleright;": "\u22eb",
    "ntrianglerighteq;": "\u22ed",
    "nu;": "\u03bd",
    "num;": "#",
    "numero;": "\u2116",
    "numsp;": "\u2007",
    "nvDash;": "\u22ad",
    "nvHarr;": "\u2904",
    "nvap;": "\u224d\u20d2",
    "nvdash;": "\u22ac",
    "nvge;": "\u2265\u20d2",
    "nvgt;": ">\u20d2",
    "nvinfin;": "\u29de",
    "nvlArr;": "\u2902",
    "nvle;": "\u2264\u20d2",
    "nvlt;": "<\u20d2",
    "nvltrie;": "\u22b4\u20d2",
    "nvrArr;": "\u2903",
    "nvrtrie;": "\u22b5\u20d2",
    "nvsim;": "\u223c\u20d2",
    "nwArr;": "\u21d6",
    "nwarhk;": "\u2923",
    "nwarr;": "\u2196",
    "nwarrow;": "\u2196",
    "nwnear;": "\u2927",
    "oS;": "\u24c8",
    "oacute": "\xf3",
    "oacute;": "\xf3",
    "oast;": "\u229b",
    "ocir;": "\u229a",
    "ocirc": "\xf4",
    "ocirc;": "\xf4",
    "ocy;": "\u043e",
    "odash;": "\u229d",
    "odblac;": "\u0151",
    "odiv;": "\u2a38",
    "odot;": "\u2299",
    "odsold;": "\u29bc",
    "oelig;": "\u0153",
    "ofcir;": "\u29bf",
    "ofr;": "\U0001d52c",
    "ogon;": "\u02db",
    "ograve": "\xf2",
    "ograve;": "\xf2",
    "ogt;": "\u29c1",
    "ohbar;": "\u29b5",
    "ohm;": "\u03a9",
    "oint;": "\u222e",
    "olarr;": "\u21ba",
    "olcir;": "\u29be",
    "olcross;": "\u29bb",
    "oline;": "\u203e",
    "olt;": "\u29c0",
    "omacr;": "\u014d",
    "omega;": "\u03c9",
    "omicron;": "\u03bf",
    "omid;": "\u29b6",
    "ominus;": "\u2296",
    "oopf;": "\U0001d560",
    "opar;": "\u29b7",
    "operp;": "\u29b9",
    "oplus;": "\u2295",
    "or;": "\u2228",
    "orarr;": "\u21bb",
    "ord;": "\u2a5d",
    "order;": "\u2134",
    "orderof;": "\u2134",
    "ordf": "\xaa",
    "ordf;": "\xaa",
    "ordm": "\xba",
    "ordm;": "\xba",
    "origof;": "\u22b6",
    "oror;": "\u2a56",
    "orslope;": "\u2a57",
    "orv;": "\u2a5b",
    "oscr;": "\u2134",
    "oslash": "\xf8",
    "oslash;": "\xf8",
    "osol;": "\u2298",
    "otilde": "\xf5",
    "otilde;": "\xf5",
    "otimes;": "\u2297",
    "otimesas;": "\u2a36",
    "ouml": "\xf6",
    "ouml;": "\xf6",
    "ovbar;": "\u233d",
    "par;": "\u2225",
    "para": "\xb6",
    "para;": "\xb6",
    "parallel;": "\u2225",
    "parsim;": "\u2af3",
    "parsl;": "\u2afd",
    "part;": "\u2202",
    "pcy;": "\u043f",
    "percnt;": "%",
    "period;": ".",
    "permil;": "\u2030",
    "perp;": "\u22a5",
    "pertenk;": "\u2031",
    "pfr;": "\U0001d52d",
    "phi;": "\u03c6",
    "phiv;": "\u03d5",
    "phmmat;": "\u2133",
    "phone;": "\u260e",
    "pi;": "\u03c0",
    "pitchfork;": "\u22d4",
    "piv;": "\u03d6",
    "planck;": "\u210f",
    "planckh;": "\u210e",
    "plankv;": "\u210f",
    "plus;": "+",
    "plusacir;": "\u2a23",
    "plusb;": "\u229e",
    "pluscir;": "\u2a22",
    "plusdo;": "\u2214",
    "plusdu;": "\u2a25",
    "pluse;": "\u2a72",
    "plusmn": "\xb1",
    "plusmn;": "\xb1",
    "plussim;": "\u2a26",
    "plustwo;": "\u2a27",
    "pm;": "\xb1",
    "pointint;": "\u2a15",
    "popf;": "\U0001d561",
    "pound": "\xa3",
    "pound;": "\xa3",
    "pr;": "\u227a",
    "prE;": "\u2ab3",
    "prap;": "\u2ab7",
    "prcue;": "\u227c",
    "pre;": "\u2aaf",
    "prec;": "\u227a",
    "precapprox;": "\u2ab7",
    "preccurlyeq;": "\u227c",
    "preceq;": "\u2aaf",
    "precnapprox;": "\u2ab9",
    "precneqq;": "\u2ab5",
    "precnsim;": "\u22e8",
    "precsim;": "\u227e",
    "prime;": "\u2032",
    "primes;": "\u2119",
    "prnE;": "\u2ab5",
    "prnap;": "\u2ab9",
    "prnsim;": "\u22e8",
    "prod;": "\u220f",
    "profalar;": "\u232e",
    "profline;": "\u2312",
    "profsurf;": "\u2313",
    "prop;": "\u221d",
    "propto;": "\u221d",
    "prsim;": "\u227e",
    "prurel;": "\u22b0",
    "pscr;": "\U0001d4c5",
    "psi;": "\u03c8",
    "puncsp;": "\u2008",
    "qfr;": "\U0001d52e",
    "qint;": "\u2a0c",
    "qopf;": "\U0001d562",
    "qprime;": "\u2057",
    "qscr;": "\U0001d4c6",
    "quaternions;": "\u210d",
    "quatint;": "\u2a16",
    "quest;": "?",
    "questeq;": "\u225f",
    "quot": "\"",
    "quot;": "\"",
    "rAarr;": "\u21db",
    "rArr;": "\u21d2",
    "rAtail;": "\u291c",
    "rBarr;": "\u290f",
    "rHar;": "\u2964",
    "race;": "\u223d\u0331",
    "racute;": "\u0155",
    "radic;": "\u221a",
    "raemptyv;": "\u29b3",
    "rang;": "\u27e9",
    "rangd;": "\u2992",
    "range;": "\u29a5",
    "rangle;": "\u27e9",
    "raquo": "\xbb",
    "raquo;": "\xbb",
    "rarr;": "\u2192",
    "rarrap;": "\u2975",
    "rarrb;": "\u21e5",
    "rarrbfs;": "\u2920",
    "rarrc;": "\u2933",
    "rarrfs;": "\u291e",
    "rarrhk;": "\u21aa",
    "rarrlp;": "\u21ac",
    "rarrpl;": "\u2945",
    "rarrsim;": "\u2974",
    "rarrtl;": "\u21a3",
    "rarrw;": "\u219d",
    "ratail;": "\u291a",
    "ratio;": "\u2236",
    "rationals;": "\u211a",
    "rbarr;": "\u290d",
    "rbbrk;": "\u2773",
    "rbrace;": "}",
    "rbrack;": "]",
    "rbrke;": "\u298c",
    "rbrksld;": "\u298e",
    "rbrkslu;": "\u2990",
    "rcaron;": "\u0159",
    "rcedil;": "\u0157",
    "rceil;": "\u2309",
    "rcub;": "}",
    "rcy;": "\u0440",
    "rdca;": "\u2937",
    "rdldhar;": "\u2969",
    "rdquo;": "\u201d",
    "rdquor;": "\u201d",
    "rdsh;": "\u21b3",
    "real;": "\u211c",
    "realine;": "\u211b",
    "realpart;": "\u211c",
    "reals;": "\u211d",
    "rect;": "\u25ad",
    "reg": "\xae",
    "reg;": "\xae",
    "rfisht;": "\u297d",
    "rfloor;": "\u230b",
    "rfr;": "\U0001d52f",
    "rhard;": "\u21c1",
    "rharu;": "\u21c0",
    "rharul;": "\u296c",
    "rho;": "\u03c1",
    "rhov;": "\u03f1",
    "rightarrow;": "\u2192",
    "rightarrowtail;": "\u21a3",
    "rightharpoondown;": "\u21c1",
    "rightharpoonup;": "\u21c0",
    "rightleftarrows;": "\u21c4",
    "rightleftharpoons;": "\u21cc",
    "rightrightarrows;": "\u21c9",
    "rightsquigarrow;": "\u219d",
    "rightthreetimes;": "\u22cc",
    "ring;": "\u02da",
    "risingdotseq;": "\u2253",
    "rlarr;": "\u21c4",
    "rlhar;": "\u21cc",
    "rlm;": "\u200f",
    "rmoust;": "\u23b1",
    "rmoustache;": "\u23b1",
    "rnmid;": "\u2aee",
    "roang;": "\u27ed",
    "roarr;": "\u21fe",
    "robrk;": "\u27e7",
    "ropar;": "\u2986",
    "ropf;": "\U0001d563",
    "roplus;": "\u2a2e",
    "rotimes;": "\u2a35",
    "rpar;": ")",
    "rpargt;": "\u2994",
    "rppolint;": "\u2a12",
    "rrarr;": "\u21c9",
    "rsaquo;": "\u203a",
    "rscr;": "\U0001d4c7",
    "rsh;": "\u21b1",
    "rsqb;": "]",
    "rsquo;": "\u2019",
    "rsquor;": "\u2019",
    "rthree;": "\u22cc",
    "rtimes;": "\u22ca",
    "rtri;": "\u25b9",
    "rtrie;": "\u22b5",
    "rtrif;": "\u25b8",
    "rtriltri;": "\u29ce",
    "ruluhar;": "\u2968",
    "rx;": "\u211e",
    "sacute;": "\u015b",
    "sbquo;": "\u201a",
    "sc;": "\u227b",
    "scE;": "\u2ab4",
    "scap;": "\u2ab8",
    "scaron;": "\u0161",
    "sccue;": "\u227d",
    "sce;": "\u2ab0",
    "scedil;": "\u015f",
    "scirc;": "\u015d",
    "scnE;": "\u2ab6",
    "scnap;": "\u2aba",
    "scnsim;": "\u22e9",
    "scpolint;": "\u2a13",
    "scsim;": "\u227f",
    "scy;": "\u0441",
    "sdot;": "\u22c5",
    "sdotb;": "\u22a1",
    "sdote;": "\u2a66",
    "seArr;": "\u21d8",
    "searhk;": "\u2925",
    "searr;": "\u2198",
    "searrow;": "\u2198",
    "sect": "\xa7",
    "sect;": "\xa7",
    "semi;": ";",
    "seswar;": "\u2929",
    "setminus;": "\u2216",
    "setmn;": "\u2216",
    "sext;": "\u2736",
    "sfr;": "\U0001d530",
    "sfrown;": "\u2322",
    "sharp;": "\u266f",
    "shchcy;": "\u0449",
    "shcy;": "\u0448",
    "shortmid;": "\u2223",
    "shortparallel;": "\u2225",
    "shy": "\xad",
    "shy;": "\xad",
    "sigma;": "\u03c3",
    "sigmaf;": "\u03c2",
    "sigmav;": "\u03c2",
    "sim;": "\u223c",
    "simdot;": "\u2a6a",
    "sime;": "\u2243",
    "simeq;": "\u2243",
    "simg;": "\u2a9e",
    "simgE;": "\u2aa0",
    "siml;": "\u2a9d",
    "simlE;": "\u2a9f",
    "simne;": "\u2246",
    "simplus;": "\u2a24",
    "simrarr;": "\u2972",
    "slarr;": "\u2190",
    "smallsetminus;": "\u2216",
    "smashp;": "\u2a33",
    "smeparsl;": "\u29e4",
    "smid;": "\u2223",
    "smile;": "\u2323",
    "smt;": "\u2aaa",
    "smte;": "\u2aac",
    "smtes;": "\u2aac\ufe00",
    "softcy;": "\u044c",
    "sol;": "/",
    "solb;": "\u29c4",
    "solbar;": "\u233f",
    "sopf;": "\U0001d564",
    "spades;": "\u2660",
    "spadesuit;": "\u2660",
    "spar;": "\u2225",
    "sqcap;": "\u2293",
    "sqcaps;": "\u2293\ufe00",
    "sqcup;": "\u2294",
    "sqcups;": "\u2294\ufe00",
    "sqsub;": "\u228f",
    "sqsube;": "\u2291",
    "sqsubset;": "\u228f",
    "sqsubseteq;": "\u2291",
    "sqsup;": "\u2290",
    "sqsupe;": "\u2292",
    "sqsupset;": "\u2290",
    "sqsupseteq;": "\u2292",
    "squ;": "\u25a1",
    "square;": "\u25a1",
    "squarf;": "\u25aa",
    "squf;": "\u25aa",
    "srarr;": "\u2192",
    "sscr;": "\U0001d4c8",
    "ssetmn;": "\u2216",
    "ssmile;": "\u2323",
    "sstarf;": "\u22c6",
    "star;": "\u2606",
    "starf;": "\u2605",
    "straightepsilon;": "\u03f5",
    "straightphi;": "\u03d5",
    "strns;": "\xaf",
    "sub;": "\u2282",
    "subE;": "\u2ac5",
    "subdot;": "\u2abd",
    "sube;": "\u2286",
    "subedot;": "\u2ac3",
    "submult;": "\u2ac1",
    "subnE;": "\u2acb",
    "subne;": "\u228a",
    "subplus;": "\u2abf",
    "subrarr;": "\u2979",
    "subset;": "\u2282",
    "subseteq;": "\u2286",
    "subseteqq;": "\u2ac5",
    "subsetneq;": "\u228a",
    "subsetneqq;": "\u2acb",
    "subsim;": "\u2ac7",
    "subsub;": "\u2ad5",
    "subsup;": "\u2ad3",
    "succ;": "\u227b",
    "succapprox;": "\u2ab8",
    "succcurlyeq;": "\u227d",
    "succeq;": "\u2ab0",
    "succnapprox;": "\u2aba",
    "succneqq;": "\u2ab6",
    "succnsim;": "\u22e9",
    "succsim;": "\u227f",
    "sum;": "\u2211",
    "sung;": "\u266a",
    "sup1": "\xb9",
    "sup1;": "\xb9",
    "sup2": "\xb2",
    "sup2;": "\xb2",
    "sup3": "\xb3",
    "sup3;": "\xb3",
    "sup;": "\u2283",
    "supE;": "\u2ac6",
    "supdot;": "\u2abe",
    "supdsub;": "\u2ad8",
    "supe;": "\u2287",
    "supedot;": "\u2ac4",
    "suphsol;": "\u27c9",
    "suphsub;": "\u2ad7",
    "suplarr;": "\u297b",
    "supmult;": "\u2ac2",
    "supnE;": "\u2acc",
    "supne;": "\u228b",
    "supplus;": "\u2ac0",
    "supset;": "\u2283",
    "supseteq;": "\u2287",
    "supseteqq;": "\u2ac6",
    "supsetneq;": "\u228b",
    "supsetneqq;": "\u2acc",
    "supsim;": "\u2ac8",
    "supsub;": "\u2ad4",
    "supsup;": "\u2ad6",
    "swArr;": "\u21d9",
    "swarhk;": "\u2926",
    "swarr;": "\u2199",
    "swarrow;": "\u2199",
    "swnwar;": "\u292a",
    "szlig": "\xdf",
    "szlig;": "\xdf",
    "target;": "\u2316",
    "tau;": "\u03c4",
    "tbrk;": "\u23b4",
    "tcaron;": "\u0165",
    "tcedil;": "\u0163",
    "tcy;": "\u0442",
    "tdot;": "\u20db",
    "telrec;": "\u2315",
    "tfr;": "\U0001d531",
    "there4;": "\u2234",
    "therefore;": "\u2234",
    "theta;": "\u03b8",
    "thetasym;": "\u03d1",
    "thetav;": "\u03d1",
    "thickapprox;": "\u2248",
    "thicksim;": "\u223c",
    "thinsp;": "\u2009",
    "thkap;": "\u2248",
    "thksim;": "\u223c",
    "thorn": "\xfe",
    "thorn;": "\xfe",
    "tilde;": "\u02dc",
    "times": "\xd7",
    "times;": "\xd7",
    "timesb;": "\u22a0",
    "timesbar;": "\u2a31",
    "timesd;": "\u2a30",
    "tint;": "\u222d",
    "toea;": "\u2928",
    "top;": "\u22a4",
    "topbot;": "\u2336",
    "topcir;": "\u2af1",
    "topf;": "\U0001d565",
    "topfork;": "\u2ada",
    "tosa;": "\u2929",
    "tprime;": "\u2034",
    "trade;": "\u2122",
    "triangle;": "\u25b5",
    "triangledown;": "\u25bf",
    "triangleleft;": "\u25c3",
    "trianglelefteq;": "\u22b4",
    "triangleq;": "\u225c",
    "triangleright;": "\u25b9",
    "trianglerighteq;": "\u22b5",
    "tridot;": "\u25ec",
    "trie;": "\u225c",
    "triminus;": "\u2a3a",
    "triplus;": "\u2a39",
    "trisb;": "\u29cd",
    "tritime;": "\u2a3b",
    "trpezium;": "\u23e2",
    "tscr;": "\U0001d4c9",
    "tscy;": "\u0446",
    "tshcy;": "\u045b",
    "tstrok;": "\u0167",
    "twixt;": "\u226c",
    "twoheadleftarrow;": "\u219e",
    "twoheadrightarrow;": "\u21a0",
    "uArr;": "\u21d1",
    "uHar;": "\u2963",
    "uacute": "\xfa",
    "uacute;": "\xfa",
    "uarr;": "\u2191",
    "ubrcy;": "\u045e",
    "ubreve;": "\u016d",
    "ucirc": "\xfb",
    "ucirc;": "\xfb",
    "ucy;": "\u0443",
    "udarr;": "\u21c5",
    "udblac;": "\u0171",
    "udhar;": "\u296e",
    "ufisht;": "\u297e",
    "ufr;": "\U0001d532",
    "ugrave": "\xf9",
    "ugrave;": "\xf9",
    "uharl;": "\u21bf",
    "uharr;": "\u21be",
    "uhblk;": "\u2580",
    "ulcorn;": "\u231c",
    "ulcorner;": "\u231c",
    "ulcrop;": "\u230f",
    "ultri;": "\u25f8",
    "umacr;": "\u016b",
    "uml": "\xa8",
    "uml;": "\xa8",
    "uogon;": "\u0173",
    "uopf;": "\U0001d566",
    "uparrow;": "\u2191",
    "updownarrow;": "\u2195",
    "upharpoonleft;": "\u21bf",
    "upharpoonright;": "\u21be",
    "uplus;": "\u228e",
    "upsi;": "\u03c5",
    "upsih;": "\u03d2",
    "upsilon;": "\u03c5",
    "upuparrows;": "\u21c8",
    "urcorn;": "\u231d",
    "urcorner;": "\u231d",
    "urcrop;": "\u230e",
    "uring;": "\u016f",
    "urtri;": "\u25f9",
    "uscr;": "\U0001d4ca",
    "utdot;": "\u22f0",
    "utilde;": "\u0169",
    "utri;": "\u25b5",
    "utrif;": "\u25b4",
    "uuarr;": "\u21c8",
    "uuml": "\xfc",
    "uuml;": "\xfc",
    "uwangle;": "\u29a7",
    "vArr;": "\u21d5",
    "vBar;": "\u2ae8",
    "vBarv;": "\u2ae9",
    "vDash;": "\u22a8",
    "vangrt;": "\u299c",
    "varepsilon;": "\u03f5",
    "varkappa;": "\u03f0",
    "varnothing;": "\u2205",
    "varphi;": "\u03d5",
    "varpi;": "\u03d6",
    "varpropto;": "\u221d",
    "varr;": "\u2195",
    "varrho;": "\u03f1",
    "varsigma;": "\u03c2",
    "varsubsetneq;": "\u228a\ufe00",
    "varsubsetneqq;": "\u2acb\ufe00",
    "varsupsetneq;": "\u228b\ufe00",
    "varsupsetneqq;": "\u2acc\ufe00",
    "vartheta;": "\u03d1",
    "vartriangleleft;": "\u22b2",
    "vartriangleright;": "\u22b3",
    "vcy;": "\u0432",
    "vdash;": "\u22a2",
    "vee;": "\u2228",
    "veebar;": "\u22bb",
    "veeeq;": "\u225a",
    "vellip;": "\u22ee",
    "verbar;": "|",
    "vert;": "|",
    "vfr;": "\U0001d533",
    "vltri;": "\u22b2",
    "vnsub;": "\u2282\u20d2",
    "vnsup;": "\u2283\u20d2",
    "vopf;": "\U0001d567",
    "vprop;": "\u221d",
    "vrtri;": "\u22b3",
    "vscr;": "\U0001d4cb",
    "vsubnE;": "\u2acb\ufe00",
    "vsubne;": "\u228a\ufe00",
    "vsupnE;": "\u2acc\ufe00",
    "vsupne;": "\u228b\ufe00",
    "vzigzag;": "\u299a",
    "wcirc;": "\u0175",
    "wedbar;": "\u2a5f",
    "wedge;": "\u2227",
    "wedgeq;": "\u2259",
    "weierp;": "\u2118",
    "wfr;": "\U0001d534",
    "wopf;": "\U0001d568",
    "wp;": "\u2118",
    "wr;": "\u2240",
    "wreath;": "\u2240",
    "wscr;": "\U0001d4cc",
    "xcap;": "\u22c2",
    "xcirc;": "\u25ef",
    "xcup;": "\u22c3",
    "xdtri;": "\u25bd",
    "xfr;": "\U0001d535",
    "xhArr;": "\u27fa",
    "xharr;": "\u27f7",
    "xi;": "\u03be",
    "xlArr;": "\u27f8",
    "xlarr;": "\u27f5",
    "xmap;": "\u27fc",
    "xnis;": "\u22fb",
    "xodot;": "\u2a00",
    "xopf;": "\U0001d569",
    "xoplus;": "\u2a01",
    "xotime;": "\u2a02",
    "xrArr;": "\u27f9",
    "xrarr;": "\u27f6",
    "xscr;": "\U0001d4cd",
    "xsqcup;": "\u2a06",
    "xuplus;": "\u2a04",
    "xutri;": "\u25b3",
    "xvee;": "\u22c1",
    "xwedge;": "\u22c0",
    "yacute": "\xfd",
    "yacute;": "\xfd",
    "yacy;": "\u044f",
    "ycirc;": "\u0177",
    "ycy;": "\u044b",
    "yen": "\xa5",
    "yen;": "\xa5",
    "yfr;": "\U0001d536",
    "yicy;": "\u0457",
    "yopf;": "\U0001d56a",
    "yscr;": "\U0001d4ce",
    "yucy;": "\u044e",
    "yuml": "\xff",
    "yuml;": "\xff",
    "zacute;": "\u017a",
    "zcaron;": "\u017e",
    "zcy;": "\u0437",
    "zdot;": "\u017c",
    "zeetrf;": "\u2128",
    "zeta;": "\u03b6",
    "zfr;": "\U0001d537",
    "zhcy;": "\u0436",
    "zigrarr;": "\u21dd",
    "zopf;": "\U0001d56b",
    "zscr;": "\U0001d4cf",
    "zwj;": "\u200d",
    "zwnj;": "\u200c",
}

replacementCharacters = {
    0x0: "\uFFFD",
    0x0d: "\u000D",
    0x80: "\u20AC",
    0x81: "\u0081",
    0x82: "\u201A",
    0x83: "\u0192",
    0x84: "\u201E",
    0x85: "\u2026",
    0x86: "\u2020",
    0x87: "\u2021",
    0x88: "\u02C6",
    0x89: "\u2030",
    0x8A: "\u0160",
    0x8B: "\u2039",
    0x8C: "\u0152",
    0x8D: "\u008D",
    0x8E: "\u017D",
    0x8F: "\u008F",
    0x90: "\u0090",
    0x91: "\u2018",
    0x92: "\u2019",
    0x93: "\u201C",
    0x94: "\u201D",
    0x95: "\u2022",
    0x96: "\u2013",
    0x97: "\u2014",
    0x98: "\u02DC",
    0x99: "\u2122",
    0x9A: "\u0161",
    0x9B: "\u203A",
    0x9C: "\u0153",
    0x9D: "\u009D",
    0x9E: "\u017E",
    0x9F: "\u0178",
}

tokenTypes = {
    "Doctype": 0,
    "Characters": 1,
    "SpaceCharacters": 2,
    "StartTag": 3,
    "EndTag": 4,
    "EmptyTag": 5,
    "Comment": 6,
    "ParseError": 7
}

tagTokenTypes = frozenset([tokenTypes["StartTag"], tokenTypes["EndTag"],
                           tokenTypes["EmptyTag"]])


prefixes = {v: k for k, v in namespaces.items()}
prefixes["http://www.w3.org/1998/Math/MathML"] = "math"


class DataLossWarning(UserWarning):
    """Raised when the current tree is unable to represent the input data"""
    pass


class _ReparseException(Exception):
    pass
site-packages/pip/_vendor/html5lib/treeadapters/__pycache__/sax.cpython-39.pyc000064400000002737150564745010023367 0ustar00a

��Re��@sdddlmZmZmZddlmZddlmZmZiZ	e�
�D]\ZZZ
edur<e
e	e<q<dd�ZdS)�)�absolute_import�division�unicode_literals)�AttributesNSImpl�)�adjustForeignAttributes�unadjustForeignAttributesNcCs|��t��D]\}}|�||�q|D]�}|d}|dkrBq*q*|dvr�t|dt�}|�|d|df|d|�|dkr�|�|d|df|d�q*|dkr�|�|d|df|d�q*|d	vr�|�|d�q*|d
kr�q*ds*Jd��q*t��D]\}}|�	|�q�|�
�d
S)z�Call SAX-like content handler based on treewalker walker

    :arg walker: the treewalker to use to walk the tree to convert it

    :arg handler: SAX handler to use

    �type�Doctype)�StartTag�EmptyTag�data�	namespace�namer�EndTag)�
Characters�SpaceCharacters�CommentFzUnknown token typeN)�
startDocument�prefix_mapping�items�startPrefixMappingrr�startElementNS�endElementNS�
characters�endPrefixMapping�endDocument)�walker�handler�prefixr�tokenr	�attrs�r"��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treeadapters/sax.py�to_sax
s>����r$)�
__future__rrrZxml.sax.xmlreaderr�	constantsrrr�valuesrZ	localNamerr$r"r"r"r#�<module>s
site-packages/pip/_vendor/html5lib/treeadapters/__pycache__/genshi.cpython-39.pyc000064400000003060150564745010024037 0ustar00a

��Re��@sLddlmZmZmZddlmZmZddlmZmZm	Z	m
Z
mZdd�ZdS)�)�absolute_import�division�unicode_literals)�QName�Attrs)�START�END�TEXT�COMMENT�DOCTYPEccsRg}|D�]*}|d}|dvr.|�|d�n|rHtd�|�dfVg}|dvr�|drnd|d|d	f}n|d	}td
d�|d��D��}tt|�|fdfV|dkr�d
}|d
kr�|dr�d|d|d	f}n|d	}tt|�dfVq|dk�rt|ddfVq|dkrt	|d	|d|dfdfVqq|�rNtd�|�dfVdS)z�Convert a tree to a genshi tree

    :arg walker: the treewalker to use to walk the tree to convert it

    :returns: generator of genshi nodes

    �type)�
Characters�SpaceCharacters�data�)N���r)�StartTag�EmptyTag�	namespace�{%s}%s�namecSs4g|],\}}t|ddur"d|n|d�|f�qS)rNr�)r)�.0�attr�value�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treeadapters/genshi.py�
<listcomp>s�zto_genshi.<locals>.<listcomp>r�EndTag�Comment�Doctype�publicId�systemIdN)
�appendr	�joinr�itemsrrrr
r)�walker�text�tokenrr�attrsrrr�	to_genshisD

�
��r*N)
�
__future__rrrZgenshi.corerrrrr	r
rr*rrrr�<module>ssite-packages/pip/_vendor/html5lib/treeadapters/__pycache__/__init__.cpython-39.pyc000064400000001720150564745010024322 0ustar00a

��Re��@s\dZddlmZmZmZddlmZdgZzddlmZWne	yLYn0e�
d�dS)	a�Tree adapters let you convert from one tree structure to another

Example:

.. code-block:: python

   from pip._vendor import html5lib
   from pip._vendor.html5lib.treeadapters import genshi

   doc = '<html><body>Hi!</body></html>'
   treebuilder = html5lib.getTreeBuilder('etree')
   parser = html5lib.HTMLParser(tree=treebuilder)
   tree = parser.parse(doc)
   TreeWalker = html5lib.getTreeWalker('etree')

   genshi_tree = genshi.to_genshi(TreeWalker(tree))

�)�absolute_import�division�unicode_literals�)�saxr)�genshirN)�__doc__�
__future__rrr�r�__all__r�ImportError�append�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treeadapters/__init__.py�<module>ssite-packages/pip/_vendor/html5lib/treeadapters/genshi.py000064400000003263150564745010017555 0ustar00from __future__ import absolute_import, division, unicode_literals

from genshi.core import QName, Attrs
from genshi.core import START, END, TEXT, COMMENT, DOCTYPE


def to_genshi(walker):
    """Convert a tree to a genshi tree

    :arg walker: the treewalker to use to walk the tree to convert it

    :returns: generator of genshi nodes

    """
    text = []
    for token in walker:
        type = token["type"]
        if type in ("Characters", "SpaceCharacters"):
            text.append(token["data"])
        elif text:
            yield TEXT, "".join(text), (None, -1, -1)
            text = []

        if type in ("StartTag", "EmptyTag"):
            if token["namespace"]:
                name = "{%s}%s" % (token["namespace"], token["name"])
            else:
                name = token["name"]
            attrs = Attrs([(QName("{%s}%s" % attr if attr[0] is not None else attr[1]), value)
                           for attr, value in token["data"].items()])
            yield (START, (QName(name), attrs), (None, -1, -1))
            if type == "EmptyTag":
                type = "EndTag"

        if type == "EndTag":
            if token["namespace"]:
                name = "{%s}%s" % (token["namespace"], token["name"])
            else:
                name = token["name"]

            yield END, QName(name), (None, -1, -1)

        elif type == "Comment":
            yield COMMENT, token["data"], (None, -1, -1)

        elif type == "Doctype":
            yield DOCTYPE, (token["name"], token["publicId"],
                            token["systemId"]), (None, -1, -1)

        else:
            pass  # FIXME: What to do?

    if text:
        yield TEXT, "".join(text), (None, -1, -1)
site-packages/pip/_vendor/html5lib/treeadapters/sax.py000064400000003360150564745010017071 0ustar00from __future__ import absolute_import, division, unicode_literals

from xml.sax.xmlreader import AttributesNSImpl

from ..constants import adjustForeignAttributes, unadjustForeignAttributes

prefix_mapping = {}
for prefix, localName, namespace in adjustForeignAttributes.values():
    if prefix is not None:
        prefix_mapping[prefix] = namespace


def to_sax(walker, handler):
    """Call SAX-like content handler based on treewalker walker

    :arg walker: the treewalker to use to walk the tree to convert it

    :arg handler: SAX handler to use

    """
    handler.startDocument()
    for prefix, namespace in prefix_mapping.items():
        handler.startPrefixMapping(prefix, namespace)

    for token in walker:
        type = token["type"]
        if type == "Doctype":
            continue
        elif type in ("StartTag", "EmptyTag"):
            attrs = AttributesNSImpl(token["data"],
                                     unadjustForeignAttributes)
            handler.startElementNS((token["namespace"], token["name"]),
                                   token["name"],
                                   attrs)
            if type == "EmptyTag":
                handler.endElementNS((token["namespace"], token["name"]),
                                     token["name"])
        elif type == "EndTag":
            handler.endElementNS((token["namespace"], token["name"]),
                                 token["name"])
        elif type in ("Characters", "SpaceCharacters"):
            handler.characters(token["data"])
        elif type == "Comment":
            pass
        else:
            assert False, "Unknown token type"

    for prefix, namespace in prefix_mapping.items():
        handler.endPrefixMapping(prefix)
    handler.endDocument()
site-packages/pip/_vendor/html5lib/treeadapters/__init__.py000064400000001247150564745010020037 0ustar00"""Tree adapters let you convert from one tree structure to another

Example:

.. code-block:: python

   from pip._vendor import html5lib
   from pip._vendor.html5lib.treeadapters import genshi

   doc = '<html><body>Hi!</body></html>'
   treebuilder = html5lib.getTreeBuilder('etree')
   parser = html5lib.HTMLParser(tree=treebuilder)
   tree = parser.parse(doc)
   TreeWalker = html5lib.getTreeWalker('etree')

   genshi_tree = genshi.to_genshi(TreeWalker(tree))

"""
from __future__ import absolute_import, division, unicode_literals

from . import sax

__all__ = ["sax"]

try:
    from . import genshi  # noqa
except ImportError:
    pass
else:
    __all__.append("genshi")
site-packages/pip/_vendor/html5lib/__init__.py000064400000002210150564745010015343 0ustar00"""
HTML parsing library based on the `WHATWG HTML specification
<https://whatwg.org/html>`_. The parser is designed to be compatible with
existing HTML found in the wild and implements well-defined error recovery that
is largely compatible with modern desktop web browsers.

Example usage::

    from pip._vendor import html5lib
    with open("my_document.html", "rb") as f:
        tree = html5lib.parse(f)

For convenience, this module re-exports the following names:

* :func:`~.html5parser.parse`
* :func:`~.html5parser.parseFragment`
* :class:`~.html5parser.HTMLParser`
* :func:`~.treebuilders.getTreeBuilder`
* :func:`~.treewalkers.getTreeWalker`
* :func:`~.serializer.serialize`
"""

from __future__ import absolute_import, division, unicode_literals

from .html5parser import HTMLParser, parse, parseFragment
from .treebuilders import getTreeBuilder
from .treewalkers import getTreeWalker
from .serializer import serialize

__all__ = ["HTMLParser", "parse", "parseFragment", "getTreeBuilder",
           "getTreeWalker", "serialize"]

# this has to be at the top level, see how setup.py parses this
#: Distribution version number.
__version__ = "1.1"
site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/base.cpython-39.pyc000064400000026133150564745010023510 0ustar00a

��Re�8�@s�ddlmZmZmZddlmZddlmZmZm	Z	dZ
ee�dfeee	ddfhB�dfeee	dd	fe	dd
fhB�dfee	ddfe	ddfg�dfee	ddfe	dd
fg�dfd�ZGdd�de
�ZGdd�de�ZGdd�de
�ZdS)�)�absolute_import�division�unicode_literals)�	text_type�)�scopingElements�tableInsertModeElements�
namespacesNF�html�button�ol�ul�table�optgroup�optionT)Nr�listr�selectc@sbeZdZdZdd�Zdd�Zdd�Zdd	�Zddd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
d
S)�NodezRepresents an item in the treecCs(||_d|_d|_i|_g|_g|_dS)zRCreates a Node

        :arg name: The tag name associated with the node

        N)�name�parent�value�
attributes�
childNodes�_flags)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treebuilders/base.py�__init__sz
Node.__init__cCs:d�dd�|j��D��}|r,d|j|fSd|jSdS)N� cSsg|]\}}d||f�qS)z%s="%s"r)�.0rrrrr�
<listcomp>.s�z Node.__str__.<locals>.<listcomp>z<%s %s>�<%s>)�joinr�itemsr)r�
attributesStrrrr�__str__-s
�zNode.__str__cCs
d|jS)Nr!)r�rrrr�__repr__6sz
Node.__repr__cCst�dS)z[Insert node as a child of the current node

        :arg node: the node to insert

        N��NotImplementedError�r�noderrr�appendChild9szNode.appendChildNcCst�dS)aBInsert data as text in the current node, positioned before the
        start of node insertBefore or to the end of the node's text.

        :arg data: the data to insert

        :arg insertBefore: True if you want to insert the text before the node
            and False if you want to insert it after the node

        Nr()r�data�insertBeforerrr�
insertTextAs
zNode.insertTextcCst�dS)aInsert node as a child of the current node, before refNode in the
        list of child nodes. Raises ValueError if refNode is not a child of
        the current node

        :arg node: the node to insert

        :arg refNode: the child node to insert the node before

        Nr()rr+�refNoderrrr.Ms
zNode.insertBeforecCst�dS)zhRemove node from the children of the current node

        :arg node: the child node to remove

        Nr(r*rrr�removeChildYszNode.removeChildcCs |jD]}|�|�qg|_dS)z�Move all the children of the current node to newParent.
        This is needed so that trees that don't store text as nodes move the
        text in the correct way

        :arg newParent: the node to move all this node's children to

        N)rr,)r�	newParent�childrrr�reparentChildrenas	
zNode.reparentChildrencCst�dS)z�Return a shallow copy of the current node i.e. a node with the same
        name and attributes but with no parent or child nodes
        Nr(r&rrr�	cloneNodenszNode.cloneNodecCst�dS)zFReturn true if the node has children or text, false otherwise
        Nr(r&rrr�
hasContenttszNode.hasContent)N)�__name__�
__module__�__qualname__�__doc__rr%r'r,r/r.r1r4r5r6rrrrrs	

rc@seZdZdd�Zdd�ZdS)�ActiveFormattingElementscCsfd}|tkrV|ddd�D]:}|tkr*qV|�||�r>|d7}|dkr|�|�qVqt�||�dS)Nr�����)�Marker�
nodesEqual�remover�append)rr+�
equalCount�elementrrrrB{s
zActiveFormattingElements.appendcCs$|j|jksdS|j|jks dSdS)NFT)�	nameTupler)r�node1�node2rrrr@�s
z#ActiveFormattingElements.nodesEqualN)r7r8r9rBr@rrrrr;zs
r;c@s�eZdZdZdZdZdZdZdZdd�Z	dd�Z
d+dd�Zd	d
�Zdd�Z
d
d�Zdd�Zdd�Zd,dd�Zdd�Zdd�Zdd�Zeee�Zdd�Zdd�Zd-dd �Zd!d"�Zd.d#d$�Zd%d&�Zd'd(�Zd)d*�ZdS)/�TreeBuilderaBase treebuilder implementation

    * documentClass - the class to use for the bottommost node of a document
    * elementClass - the class to use for HTML Elements
    * commentClass - the class to use for comments
    * doctypeClass - the class to use for doctypes

    NcCs|rd|_nd|_|��dS)zmCreate a TreeBuilder

        :arg namespaceHTMLElements: whether or not to namespace HTML elements

        zhttp://www.w3.org/1999/xhtmlN)�defaultNamespace�reset)r�namespaceHTMLElementsrrrr�szTreeBuilder.__init__cCs.g|_t�|_d|_d|_d|_|��|_dS)NF)�openElementsr;�activeFormattingElements�headPointer�formPointer�insertFromTable�
documentClass�documentr&rrrrJ�szTreeBuilder.resetcCs�t|d�}|s2t|t�r$td|f}t|t�s2J�t|\}}t|j�D]>}|r^||kr^dS|sr|j|krrdS||j|vArHdSqHds�J�dS)NrEr
TF)	�hasattr�
isinstancerr	�tuple�listElementsMap�reversedrLrE)r�target�variant�	exactNode�listElements�invertr+rrr�elementInScope�s

zTreeBuilder.elementInScopecCs�|js
dSt|j�d}|j|}|tks4||jvr8dS|tkrl||jvrl|dkrXd}ql|d8}|j|}q8|d7}|j|}|��}|�d|j|j|jd��}||j|<||jdkrlq�qldS)Nr=rr<�StartTag)�typer�	namespacer-)	rM�lenr?rLr5�
insertElementrr`r)r�i�entry�clonerDrrr�#reconstructActiveFormattingElements�s.

�
z/TreeBuilder.reconstructActiveFormattingElementscCs(|j��}|jr$|tkr$|j��}q
dS�N)rM�popr?)rrdrrr�clearActiveFormattingElementss
z)TreeBuilder.clearActiveFormattingElementscCs:|jddd�D]$}|tkr"q6q|j|kr|SqdS)z�Check if an element exists between the end of the active
        formatting elements and the last marker. If it does, return it, else
        return falseNr<F)rMr?r)rr�itemrrr�!elementInActiveFormattingElements
s

z-TreeBuilder.elementInActiveFormattingElementscCs&|�|�}|j�|�|j�|�dSrg)�
createElementrLrBrRr,)r�tokenrDrrr�
insertRoots
zTreeBuilder.insertRootcCs6|d}|d}|d}|�|||�}|j�|�dS)Nr�publicId�systemId)�doctypeClassrRr,)rrmrrorp�doctyperrr�
insertDoctype s
zTreeBuilder.insertDoctypecCs*|dur|jd}|�|�|d��dS)Nr<r-)rLr,�commentClass)rrmrrrr�
insertComment(s
zTreeBuilder.insertCommentcCs0|d}|�d|j�}|�||�}|d|_|S)z.Create an element but don't insert it anywhererr`r-)�getrI�elementClassr�rrmrr`rDrrrrl-s

zTreeBuilder.createElementcCs|jSrg)�_insertFromTabler&rrr�_getInsertFromTable5szTreeBuilder._getInsertFromTablecCs ||_|r|j|_n|j|_dS)zsSwitch the function used to insert an element from the
        normal one to the misnested table one and back againN)ry�insertElementTablerb�insertElementNormal)rrrrr�_setInsertFromTable8s
zTreeBuilder._setInsertFromTablecCsb|d}t|t�sJd|��|�d|j�}|�||�}|d|_|jd�|�|j�|�|S)NrzElement %s not unicoder`r-r<)	rTrrvrIrwrrLr,rBrxrrrr|Cs
zTreeBuilder.insertElementNormalcCs`|�|�}|jdjtvr$|�|�S|��\}}|durD|�|�n|�||�|j�|�|S)z-Create an element and insert it into the treer<N)	rlrLrrr|�getTableMisnestedNodePositionr,r.rB)rrmrDrr.rrrr{Ms

zTreeBuilder.insertElementTablecCsV|dur|jd}|jr.|jr:|jdjtvr:|�|�n|��\}}|�||�dS)zInsert text data.Nr<)rLrPrrr/r~)rr-rr.rrrr/]s

��zTreeBuilder.insertTextcCstd}d}d}|jddd�D]}|jdkr|}q4q|rb|jrJ|j}|}ql|j|j�|�d}n
|jd}||fS)zsGet the foster parent element, and sibling to insert before
        (or None) when inserting a misnested table nodeNr<rr=r)rLrr�index)r�	lastTable�fosterParentr.�elmrrrr~ls 
�
z)TreeBuilder.getTableMisnestedNodePositioncCs8|jdj}|td�vr4||kr4|j��|�|�dS)Nr<)�dd�dt�lirr�p�rp�rt)rLr�	frozensetrh�generateImpliedEndTags)r�excluderrrrr��s�
z"TreeBuilder.generateImpliedEndTagscCs|jS)zReturn the final tree)rRr&rrr�getDocument�szTreeBuilder.getDocumentcCs|��}|jd�|�|S)zReturn the final fragmentr)�
fragmentClassrLr4)r�fragmentrrr�getFragment�szTreeBuilder.getFragmentcCst�dS)z�Serialize the subtree of node in the format required by unit tests

        :arg node: the node from which to start serializing

        Nr(r*rrr�testSerializer�szTreeBuilder.testSerializer)N)N)N)N)r7r8r9r:rQrwrtrqr�rrJr]rfrirkrnrsrurlrzr}�propertyrPr|r{r/r~r�r�r�r�rrrrrH�s6
.
	




rH)�
__future__rrrZpip._vendor.sixr�	constantsrrr	r?r�rV�objectrrr;rHrrrr�<module>s0

��
��
���csite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree_lxml.cpython-39.pyc000064400000031363150564745010024737 0ustar00a

��Re�9�@sdZddlmZmZmZddlZddlZddlZzddlm	Z	Wne
y^ddlm	Z	Yn0ddlm
Z
ddlmZdd	lmZdd
lmZddlmZddlmZddlmZmZd
Ze�d�Ze�d�jZGdd�de�ZGdd�de�Zdd�Z dd�Z!Gdd�de
j"�Z"dS)a�Module for supporting the lxml.etree library. The idea here is to use as much
of the native library as possible, without using fragile hacks like custom element
names that break between releases. The downside of this is that we cannot represent
all possible trees; specifically the following are known to cause problems:

Text or comments as siblings of the root element
Docypes with no name

When any of these things occur, we emit a DataLossWarning
�)�absolute_import�division�unicode_literalsN)�MutableMapping�)�base�)�DataLossWarning)�	constants)�etree)�	_ihatexml)�PY3�binary_typeTz
{([^}]*)}(.*)Zasdc@seZdZdd�ZdS)�DocumentTypecCs||_||_||_dS�N)�name�publicId�systemId)�selfrrr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treebuilders/etree_lxml.py�__init__)szDocumentType.__init__N)�__name__�
__module__�__qualname__rrrrrr(src@s,eZdZdd�Zdd�Zdd�Zee�ZdS)�DocumentcCsd|_g|_dSr)�_elementTree�_childNodes�rrrrr0szDocument.__init__cCs.|j��}|j����D]}q|�|j�dSr)r�getrootZitersiblingsZaddnext�_element)r�element�lastrrr�appendChild4s
zDocument.appendChildcCs|jSr)rrrrr�_getChildNodes;szDocument._getChildNodesN)rrrrr#r$�property�
childNodesrrrrr/srcs6g�tjdd��d���fdd�	��|d�d���S)NT�ZpreventDoubleDashCommentsrc
s,t|d��s,t|d�r���d�|jjrz|jjsF|jjsFd|jj}nd|jj|jj|jjf}��dd|d|f�|��}|��dur�|��}q�|dur��||d�|�	�}q�nnt
|t�s�t
|t��rt
|t�s�t
jd	dks�J���d
d||f�n$��d�|D]}�||d��q�n�|jtk�r���dd||jf�t|d
��r(|j�r(��d
d||jf��n�t
|tj��s�J�tj�|j�}|du�r�|�d�}|�d�}tj|}��dd||��|�f�n��dd|��|j�f�t|d��r�g}|j��D]d\}	}
t�|	�}|du�rj|��\}}	��|	�}	tj|}d||	f}n
��|	�}|�||
f��q t|�D]&\}	}
��dd|d|	|
f��q�|j�rڈ�d
d|d|jf�|d7}|D]}�||��q�t|d
��r(|j�r(��d
d|d|jf�dS)N�tagrz	#document�
<!DOCTYPE %s>z<!DOCTYPE %s "%s" "%s">z|%s%s� rrz|%s"%s"z#document-fragmentz|%s<!-- %s -->�tailrz
|%s<%s %s>z|%s<%s>�attribz%s %sz
|%s%s="%s") �hasattr�append�docinfo�internalDTD�	public_idZ
system_url�	root_namerZgetpreviousZgetnext�
isinstance�str�bytes�sys�version_infor(�comment_type�textr+rZ_Element�etree_builders�
tag_regexp�match�groupr
�prefixes�fromXmlNamer,�items�groups�sorted)
r!�indent�dtd_strZnext_elementZnsmatch�nsr(�prefix�
attributesr�valueZattr_string�child��
infosetFilter�rv�serializeElementrrrMEs|

��





�

�




 z(testSerializer.<locals>.serializeElement�
)r)r�
InfosetFilter�join�r!rrJr�testSerializerAs
F
rRcs$g���fdd���|�d���S)z4Serialize an element and its child nodes to a stringcst|d�sH|jjr:|jjr$|jj}nd|jj}��|��|���n�|jtkrf��d|j	f�n~|j
s���d|jf�n.d�dd�|j
��D��}��d|j|f�|j	r���|j	�|D]}�|�qĈ�d	|jf�t|d
��r|j
�r��|j
�dS)Nr(r)z	<!--%s-->z<%s>r*cSsg|]\}}d||f�qS)z%s="%s"r)�.0rrHrrr�
<listcomp>�s�z6tostring.<locals>.serializeElement.<locals>.<listcomp>z<%s %s>z</%s>r+)r-r/r0�doctyper2r.rr(r8r9r,rPr@r+)r!rD�attrrI�rLrMrrrM�s,




�
z"tostring.<locals>.serializeElement�)rPrQrrWr�tostring�s rYcszeZdZeZeZdZdZeZ	e
Zddd�Zdd�Z
dd�Zd	d
�Zdd�Zd
d�Zddd�Zd�fdd�	Zdd�Z�ZS)�TreeBuilderNFcs�tjt|d��tjdd��|_||_G�fdd�dt��G���fdd�d�j�}G��fdd	�d	�j	�}||_
||_tj
�||�dS)
N)�fullTreeTr'csPeZdZdd�Z�fdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dS)z(TreeBuilder.__init__.<locals>.AttributescSs
||_dSr)r �rr!rrrr�sz1TreeBuilder.__init__.<locals>.Attributes.__init__cs4t|t�r&d|d��|d�f}n
��|�}|S)N�{%s}%srr)r3�tupleZcoerceAttribute)r�keyr�rKrr�
_coerceKey�s

z3TreeBuilder.__init__.<locals>.Attributes._coerceKeycSs0|jjj|�|�}ts,t|t�r,|�d�}|S)N�ascii)r r,rar
r3r�decode�rr_rHrrr�__getitem__�s
z4TreeBuilder.__init__.<locals>.Attributes.__getitem__cSs||jjj|�|�<dSr�r r,rardrrr�__setitem__�sz4TreeBuilder.__init__.<locals>.Attributes.__setitem__cSs|jjj|�|�=dSrrf)rr_rrr�__delitem__�sz4TreeBuilder.__init__.<locals>.Attributes.__delitem__cSst|jjj�Sr)�iterr r,rrrr�__iter__�sz1TreeBuilder.__init__.<locals>.Attributes.__iter__cSst|jjj�Sr)�lenr r,rrrr�__len__�sz0TreeBuilder.__init__.<locals>.Attributes.__len__cSs|jjj��Sr)r r,�clearrrrrrm�sz.TreeBuilder.__init__.<locals>.Attributes.clearN)rrrrrarergrhrjrlrmrr`rr�
Attributes�srncspeZdZ���fdd�Z�fdd�Z�fdd�Zeee�Zdd�Zd	d
�Z	eee	�Z
d��fdd
�	Zdd�ZdS)z%TreeBuilder.__init__.<locals>.Elementcs*��|�}�jj|||d��|�|_dS)N)�	namespace)�
coerceElement�Elementr�_attributes)rrro�rn�builderrKrrr�s
z.TreeBuilder.__init__.<locals>.Element.__init__cs$��|�|_|�|j|j�|j_dSr)rp�_nameZ_getETreeTagZ
_namespacer r()rrr`rr�_setName�s�z.TreeBuilder.__init__.<locals>.Element._setNamecs��|j�Sr)r?rurr`rr�_getName�sz.TreeBuilder.__init__.<locals>.Element._getNamecSs|jSr)rrrrrr�_getAttributes�sz4TreeBuilder.__init__.<locals>.Element._getAttributescSs|j}|��|�|�dSr)rGrm�update)rrHrGrrr�_setAttributes�sz4TreeBuilder.__init__.<locals>.Element._setAttributesNcs��|�}�j�|||�dSr)ZcoerceCharactersrq�
insertText)r�data�insertBefore�rtrKrrr{s
z0TreeBuilder.__init__.<locals>.Element.insertTextcSs0t|�|j|j�}|jjr,|jj�|jj�|Sr)�typerror r,ryr\rrr�	cloneNodesz/TreeBuilder.__init__.<locals>.Element.cloneNode)N)
rrrrrvrwr%rrxrzrGr{r�rrsrrrq�s

rqcs8eZdZ��fdd�Z�fdd�Zdd�Zeee�ZdS)z%TreeBuilder.__init__.<locals>.Commentcs��|�}�j�||�dSr)�
coerceComment�Commentr�rr|r~rrrs
z.TreeBuilder.__init__.<locals>.Comment.__init__cs��|�}||j_dSr)r�r r9r�r`rr�_setDatas
z.TreeBuilder.__init__.<locals>.Comment._setDatacSs|jjSr)r r9rrrr�_getDatasz.TreeBuilder.__init__.<locals>.Comment._getDataN)rrrrr�r�r%r|rr~rrr�
sr�)r:�getETreeModulerrrOrK�namespaceHTMLElementsrrqr��elementClass�commentClassrrZr)rr�r[rqr�rrsrr�s $zTreeBuilder.__init__cCs$tj�|�|j|_g|_d|_dSr)rrZ�reset�insertCommentInitial�
insertComment�initial_commentsrUrrrrr�szTreeBuilder.resetcCst|�Sr)rRr\rrrrR#szTreeBuilder.testSerializercCstr|jjS|jj��SdSr)r[�documentrrrrrr�getDocument&szTreeBuilder.getDocumentcCsFg}|jdj}|jr"|�|j�|�t|��|jrB|�|j�|S)Nr)�openElementsr r9r.�extend�listr+)r�fragmentr!rrr�getFragment,szTreeBuilder.getFragmentcCsh|d}|d}|d}|s0t�dt�d|_n4|j�|�}||krPt�dt�|�|||�}||_dS)Nrrrz#lxml cannot represent empty doctypez%lxml cannot represent non-xml doctype)�warnings�warnr	rUrKrp�doctypeClass)r�tokenrrrZcoercedNamerUrrr�
insertDoctype6szTreeBuilder.insertDoctypecCs6|dus||jusJ�|jjdus&J�|j�|�dSr)r�rr�r.�rr|�parentrrrr�Fsz TreeBuilder.insertCommentInitialcsB||jkr,|jj��djtkr,t�dt�tt	|��
||�dS)N���z@lxml cannot represent adjacent comments beyond the root elements)r�rrr(r8r�r�r	�superrZr�r���	__class__rr�insertCommentMainKs

�zTreeBuilder.insertCommentMaincCs�d}|jr�|jjsJ�|d|jj7}|jjdus>|jjdur�|d|j�|jjpRd�7}|jjr�|jj}|�d�dkr�|�d�dkr�t�dt	�|�
dd�}|�d�dkr�|d	|7}q�|d
|7}n|d7}|d7}|jj|d
kr�t�dt	�|d7}t�|�}|j
D] }|�|d�}|�|j��q|��|_|��|j_|d
}|�d|j�}|du�rj|}	nd||f}	|	|_|�||�}
||
_|jj�|
�|j�|
�|j|_dS)NrXz<!DOCTYPE %sz
 PUBLIC "%s" �'r�"z6DOCTYPE system cannot contain single and double quotesZU00027z"%s"z'%s'z''�>rzGlxml cannot represent doctype with a different name to the root elementz$<THIS_SHOULD_NEVER_APPEAR_PUBLICLY/>r|ror])rUrrrrKZcoercePubid�findr�r�r	�replacer�
fromstringr�r�Zaddpreviousr �
documentClassr�Zgetroottreer�get�defaultNamespacer(r�rr.r�r�r�)rr�ZdocStr�sysid�root�
comment_token�commentrroZ	etree_tagZroot_elementrrr�
insertRootQsP
��



zTreeBuilder.insertRoot)F)N)N)rrrrr�rr�r�r��
fragmentClassr�implementationrr�rRr�r�r�r�r�r��
__classcell__rrr�rrZ�s
\

rZ)#�__doc__�
__future__rrrr��rer6�collections.abcr�ImportError�collectionsrXrr
r	rr:rZ
lxml.etreeZpip._vendor.sixr
rr[�compiler;r�r(r8�objectrrrRrYrZrrrr�<module>s.
O)site-packages/pip/_vendor/html5lib/treebuilders/__pycache__/etree.cpython-39.pyc000064400000027124150564745010023703 0ustar00a

��Re$2�@s�ddlmZmZmZddlmZddlZddlmZddlm	Z	ddlm
Z
dd	lmZdd
lmZddl
mZe�d�Zddd�Zee�ZdS)�)�absolute_import�division�unicode_literals)�	text_typeN)�copy�)�base�)�	_ihatexml)�	constants)�
namespaces)�moduleFactoryFactoryz
{([^}]*)}(.*)Fc	s�����d�j�G�fdd�dtj��G�fdd�d���G�fdd�d���G�fdd	�d	���G�fd
d�d����fdd
��	��fdd�}G��������	fdd�dtj�}t�S)NZasdcs�eZdZd$�fdd�	Zdd�Zdd�Zdd	�Zeee�Zd
d�Z	dd
�Z
ee
e	�Zdd�Zdd�Z
eee
�Zdd�Zdd�Zeee�Zdd�Zdd�Zdd�Zdd�Zd%dd�Zd d!�Zd"d#�ZdS)&z getETreeBuilder.<locals>.ElementNcs^||_||_��|�||��|_|dur:td|jf|_n|j|jf|_d|_g|_g|_	dS)N�html)
�_name�
_namespace�Element�_getETreeTag�_elementr�	nameTuple�parent�_childNodes�_flags)�self�name�	namespace��ElementTree���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treebuilders/etree.py�__init__s
�z)getETreeBuilder.<locals>.Element.__init__cSs|dur|}nd||f}|S)N�{%s}%sr)rrrZ	etree_tagrrrr%sz-getETreeBuilder.<locals>.Element._getETreeTagcSs||_|�|j|j�|j_dS�N)rrrr�tag)rrrrr�_setName,sz)getETreeBuilder.<locals>.Element._setNamecSs|jSr!)r�rrrr�_getName0sz)getETreeBuilder.<locals>.Element._getNamecSs||_|�|j|j�|j_dSr!)rrrrr")rrrrr�
_setNamespace5sz.getETreeBuilder.<locals>.Element._setNamespacecSs|jSr!)rr$rrr�
_getNamespace9sz.getETreeBuilder.<locals>.Element._getNamespacecSs|jjSr!)r�attribr$rrr�_getAttributes>sz/getETreeBuilder.<locals>.Element._getAttributescSsV|jj}|��|rR|��D]4\}}t|t�rDd|d|df}n|}|||<qdS)Nr r	r)rr(�clear�items�
isinstance�tuple)r�
attributesZ	el_attrib�key�valuerrrr�_setAttributesAs
z/getETreeBuilder.<locals>.Element._setAttributescSs|jSr!)rr$rrr�_getChildNodesPsz/getETreeBuilder.<locals>.Element._getChildNodescSs*|jdd�=g|_|D]}|�|�qdSr!)rrZinsertChild)rr0�elementrrr�_setChildNodesSsz/getETreeBuilder.<locals>.Element._setChildNodescSst|jjpt|j��S)z,Return true if the node has children or text)�boolr�text�lenr$rrr�
hasContent[sz+getETreeBuilder.<locals>.Element.hasContentcSs$|j�|�|j�|j�||_dSr!)r�appendrr�r�noderrr�appendChild_sz,getETreeBuilder.<locals>.Element.appendChildcSs,t|j��|j�}|j�||j�||_dSr!)�listr�index�insertr)rr;�refNoder>rrr�insertBeforedsz-getETreeBuilder.<locals>.Element.insertBeforecSs$|j�|�|j�|j�d|_dSr!)r�removerrr:rrr�removeChildisz,getETreeBuilder.<locals>.Element.removeChildcSs�t|j�s,|jjsd|j_|jj|7_n�|durb|jdjsLd|jd_|jdj|7_nxt|j�}|�|j�}|dkr�|j|djs�d|j|d_|j|dj|7_n |jjs�d|j_|jj|7_dS)N����rr)r7rr6�tailr=r>)r�datarA�childrenr>rrr�
insertTextns"

z+getETreeBuilder.<locals>.Element.insertTextcSs.t|�|j|j�}|jjr*t|jj�|j_|Sr!)�typerrrr(r�rr3rrr�	cloneNode�sz*getETreeBuilder.<locals>.Element.cloneNodecSsl|jr"|jdjj|jj7_n0|jjs2d|j_|jjdurR|jj|jj7_d|j_tj�||�dS)NrErD)�
childNodesrrFr6r�Node�reparentChildren)r�	newParentrrrrO�sz1getETreeBuilder.<locals>.Element.reparentChildren)N)N)�__name__�
__module__�__qualname__rrr#r%�propertyrr&r'rr)r1r.r2r4rMr8r<rArCrIrLrOrrrrrs*





rcs2eZdZ�fdd�Zdd�Zdd�Zeee�ZdS)z getETreeBuilder.<locals>.Commentcs"��|�|_d|_g|_g|_dSr!)�Commentrrrr)rrGrrrr�sz)getETreeBuilder.<locals>.Comment.__init__cSs|jjSr!�rr6r$rrr�_getData�sz)getETreeBuilder.<locals>.Comment._getDatacSs||j_dSr!rV�rr0rrr�_setData�sz)getETreeBuilder.<locals>.Comment._setDataN)rQrRrSrrWrYrTrGrrrrrU�srUcsLeZdZ�fdd�Zdd�Zdd�Zeee�Zdd�Zd	d
�Z	eee	�Z
dS)z%getETreeBuilder.<locals>.DocumentTypecs$��|d�||j_||_||_dS)N�
<!DOCTYPE>)rrr6�publicId�systemId)rrr[r\�rrrr�sz.getETreeBuilder.<locals>.DocumentType.__init__cSs|j�dd�S)Nr[rD�r�getr$rrr�_getPublicId�sz2getETreeBuilder.<locals>.DocumentType._getPublicIdcSs|dur|j�d|�dS)Nr[�r�setrXrrr�_setPublicId�sz2getETreeBuilder.<locals>.DocumentType._setPublicIdcSs|j�dd�S)Nr\rDr^r$rrr�_getSystemId�sz2getETreeBuilder.<locals>.DocumentType._getSystemIdcSs|dur|j�d|�dS)Nr\rarXrrr�_setSystemId�sz2getETreeBuilder.<locals>.DocumentType._setSystemIdN)rQrRrSrr`rcrTr[rdrer\rr]rr�DocumentType�s
rfcseZdZ�fdd�ZdS)z!getETreeBuilder.<locals>.Documentcs��|d�dS)N�
DOCUMENT_ROOT�rr$r]rrr�sz*getETreeBuilder.<locals>.Document.__init__N�rQrRrSrrr]rr�Document�srjcseZdZ�fdd�ZdS)z)getETreeBuilder.<locals>.DocumentFragmentcs��|d�dS)NZDOCUMENT_FRAGMENTrhr$r]rrr�sz2getETreeBuilder.<locals>.DocumentFragment.__init__Nrirr]rr�DocumentFragment�srkcs*g�d���fdd�	��|d�d���S)Nrcs�t|d�s|��}|jdkrz|�d�s0|�d�rd|�d�p<d}|�d�pJd}��d|j||f�n��d|jf��n�|jdkr��d	�|jdur���d
d|d|jf�|jdur�td
��t|d�r�t|j	�r�td���nh|j�k�r��dd||jf��n@t
|jt��s4Jdt|j�|jf��t
�|j�}|du�rR|j}n"|��\}}tj|}d||f}��dd||f�t|d��r*g}|j	��D]T\}}	t
�|�}|du�r�|��\}}tj|}d||f}
n|}
|�|
|	f��q�t|�D]&\}}	��dd|d||	f��q|j�rN��d
d|d|jf�|d7}|D]}�||��qZ|j�r���d
d|d|jf�dS)Nr"rZr[r\rDz<!DOCTYPE %s "%s" "%s">�
<!DOCTYPE %s>rgz	#documentz|%s"%s"� r	�Document node cannot have tailr(�$Document node cannot have attributesz|%s<!-- %s -->zExpected unicode, got %s, %sz%s %sz|%s<%s>z
|%s%s="%s")�hasattr�getrootr"r_r9r6rF�	TypeErrorr7r(r,rrJ�
tag_regexp�match�groupsr�prefixesr+�sorted)r3�indentr[r\Znsmatchr�ns�prefixr.r0Zattr_string�child)�ElementTreeCommentType�rv�serializeElementrrr~�sb


�



�




 zAgetETreeBuilder.<locals>.testSerializer.<locals>.serializeElement�
)r)�join�r3)r|)r}r~r�testSerializer�s7
z'getETreeBuilder.<locals>.testSerializercs2g�t��������fdd���|�d���S)z4Serialize an element and its child nodes to a stringcs�t|�j�r|��}|jdkr||�d�s2|�d�rf|�d�p>d}|�d�pLd}��d|j||f�n��d|jf��n|jdkr�|jdur���|j�|jdur�td��t	|d	�r�t
|j�r�td
��|D]}�|�q�n�|j�kr���d|jf�n�|j�s��d��|j�f�n2d
�
�fdd�|j��D��}��d|j|f�|j�rd��|j�|D]}�|��qh��d|jf�|j�r���|j�dS)NrZr[r\rDz<!DOCTYPE %s PUBLIC "%s" "%s">rlrgrnr(roz	<!--%s-->z<%s>rmcs"g|]\}}d��|�|f�qS)z%s="%s")�fromXmlName)�.0rr0)�filterrr�
<listcomp>)s�
�zOgetETreeBuilder.<locals>.tostring.<locals>.serializeElement.<locals>.<listcomp>z<%s %s>z</%s>)r,rrqr"r_r9r6rFrrrpr7r(r�r�r+)r3r[r\r{�attr)rr|r�r}r~rrr~sD

�



�z;getETreeBuilder.<locals>.tostring.<locals>.serializeElementrD)r
Z
InfosetFilterr�r�)rr|)r�r}r~r�tostrings
-z!getETreeBuilder.<locals>.tostringcsDeZdZ�Z�Z�Z�Z�Z�Z�fdd�Z	�fdd�Z
dd�ZdS)z$getETreeBuilder.<locals>.TreeBuildercs�|�Sr!rrK)r�rrr�Dsz3getETreeBuilder.<locals>.TreeBuilder.testSerializercs<�r|jjS|jdur*|jj�d|j�S|jj�d�SdS)Nz{%s}htmlr)�documentr�defaultNamespace�findr$)�fullTreerr�getDocumentGs
�z0getETreeBuilder.<locals>.TreeBuilder.getDocumentcSstj�|�jSr!)r�TreeBuilder�getFragmentrr$rrrr�Qsz0getETreeBuilder.<locals>.TreeBuilder.getFragmentN)rQrRrS�
documentClass�doctypeClass�elementClass�commentClass�
fragmentClass�implementationr�r�r�r)rUrjrkrfr�ElementTreeImplementationr�r�rrr�<s
r�)rUr"rrNr��locals)r�r�r�r�r)
rUrjrkrfrrr|r�r�r�r�getETreeBuilders>6$r�)F)�
__future__rrrZpip._vendor.sixr�rerrDrr
rr�_utilsr
�compilersr��getETreeModulerrrr�<module>s

Fsite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/__init__.cpython-39.pyc000064400000006453150564745010024340 0ustar00a

��Re�@s6dZddlmZmZmZddlmZiZddd�ZdS)	a)A collection of modules for building different kinds of trees from HTML
documents.

To create a treebuilder for a new type of tree, you need to do
implement several things:

1. A set of classes for various types of elements: Document, Doctype, Comment,
   Element. These must implement the interface of ``base.treebuilders.Node``
   (although comment nodes have a different signature for their constructor,
   see ``treebuilders.etree.Comment``) Textual content may also be implemented
   as another node type, or not, as your tree implementation requires.

2. A treebuilder object (called ``TreeBuilder`` by convention) that inherits
   from ``treebuilders.base.TreeBuilder``. This has 4 required attributes:

   * ``documentClass`` - the class to use for the bottommost node of a document
   * ``elementClass`` - the class to use for HTML Elements
   * ``commentClass`` - the class to use for comments
   * ``doctypeClass`` - the class to use for doctypes

   It also has one required method:

   * ``getDocument`` - Returns the root node of the complete document tree

3. If you wish to run the unit tests, you must also create a ``testSerializer``
   method on your treebuilder which accepts a node and returns a string
   containing Node and its children serialized according to the format used in
   the unittests

�)�absolute_import�division�unicode_literals�)�
default_etreeNcKs�|��}|tvr�|dkrPddlm}|dur<ddlm}|}|j|fi|��jS|dkrpddlm}|jt|<n@|d	kr�dd
lm	}|dur�t
}|j|fi|��jStd|��t�
|�S)aGet a TreeBuilder class for various types of trees with built-in support

    :arg treeType: the name of the tree type required (case-insensitive). Supported
        values are:

        * "dom" - A generic builder for DOM implementations, defaulting to a
          xml.dom.minidom based implementation.
        * "etree" - A generic builder for tree implementations exposing an
          ElementTree-like interface, defaulting to xml.etree.cElementTree if
          available and xml.etree.ElementTree if not.
        * "lxml" - A etree-based builder for lxml.etree, handling limitations
          of lxml's implementation.

    :arg implementation: (Currently applies to the "etree" and "dom" tree
        types). A module implementing the tree type e.g. xml.etree.ElementTree
        or xml.etree.cElementTree.

    :arg kwargs: Any additional options to pass to the TreeBuilder when
        creating it.

    Example:

    >>> from html5lib.treebuilders import getTreeBuilder
    >>> builder = getTreeBuilder('etree')

    �dom�)rNr)�minidom�lxml)�
etree_lxml�etree)rzUnrecognised treebuilder "%s" )�lower�treeBuilderCache�r�xml.domr	�getDomModule�TreeBuilderrrr�getETreeModule�
ValueError�get)�treeType�implementation�kwargsrr	rr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treebuilders/__init__.py�getTreeBuilder's$r)N)	�__doc__�
__future__rrr�_utilsrrrrrrr�<module>ssite-packages/pip/_vendor/html5lib/treebuilders/__pycache__/dom.cpython-39.pyc000064400000022424150564745010023354 0ustar00a

��Re�"�@s�ddlmZmZmZzddlmZWneyBddlmZYn0ddlm	Z	m
Z
ddlZddlm
Z
ddlmZdd	lmZdd
lmZdd�Zee�ZdS)
�)�absolute_import�division�unicode_literals)�MutableMapping)�minidom�NodeN�)�base�)�	constants)�
namespaces)�moduleFactoryFactorycsV��Gdd�dt��G��fdd�dtj��G����fdd�dtj�}dd��t�S)	Nc@sLeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dS)zgetDomBuilder.<locals>.AttrListcSs
||_dS�N��element��selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/html5lib/treebuilders/dom.py�__init__sz(getDomBuilder.<locals>.AttrList.__init__cSst|jj���Sr)�iterr�
attributes�keys�rrrr�__iter__sz(getDomBuilder.<locals>.AttrList.__iter__cSs4t|t�rt�n |jj�|�}||_||jj|<dSr)�
isinstance�tuple�NotImplementedErrorr�
ownerDocumentZcreateAttribute�valuer)r�namer�attrrrr�__setitem__s

z+getDomBuilder.<locals>.AttrList.__setitem__cSst|jj�Sr)�lenrrrrrr�__len__#sz'getDomBuilder.<locals>.AttrList.__len__cSst|jj���Sr)�listrr�itemsrrrrr&&sz%getDomBuilder.<locals>.AttrList.itemscSst|jj���Sr)r%rr�valuesrrrrr')sz&getDomBuilder.<locals>.AttrList.valuescSs"t|t�rt�n|jj|jSdSr)rrrrrr�rr rrr�__getitem__,s
z+getDomBuilder.<locals>.AttrList.__getitem__cSst|t�rt�n
|jj|=dSr)rrrrrr(rrr�__delitem__2s
z+getDomBuilder.<locals>.AttrList.__delitem__N)�__name__�
__module__�__qualname__rrr"r$r&r'r)r*rrrr�AttrListsr.cs�eZdZdd�Zedd��Zdd�Zddd	�Zd
d�Zdd
�Z	dd�Z
�fdd�Zdd�Zeee�Z
�fdd�Zdd�Zdd�Zee�ZdS)z"getDomBuilder.<locals>.NodeBuildercSstj�||j�||_dSr)r	rr�nodeNamerrrrrr9sz+getDomBuilder.<locals>.NodeBuilder.__init__cSst|jd�r|jjpdS)N�namespaceURI)�hasattrrr0rrrr�<lambda>=s�z+getDomBuilder.<locals>.NodeBuilder.<lambda>cSs||_|j�|j�dSr)�parentr�appendChild�r�noderrrr4@sz.getDomBuilder.<locals>.NodeBuilder.appendChildNcSs4|jj�|�}|r$|j�||j�n|j�|�dSr)rr�createTextNode�insertBeforer4)r�datar8�textrrr�
insertTextDsz-getDomBuilder.<locals>.NodeBuilder.insertTextcSs|j�|j|j�||_dSr)rr8r3)rr6�refNoderrrr8Ksz/getDomBuilder.<locals>.NodeBuilder.insertBeforecSs&|jj|jkr|j�|j�d|_dSr)rZ
parentNode�removeChildr3r5rrrr=Osz.getDomBuilder.<locals>.NodeBuilder.removeChildcSs6|j��r,|jj}|j�|�|j�|�qg|_dSr)r�
hasChildNodesZ
firstChildr=r4�
childNodes)r�	newParent�childrrr�reparentChildrenTs

z3getDomBuilder.<locals>.NodeBuilder.reparentChildrencs
�|j�Srrr)r.rr�
getAttributes[sz0getDomBuilder.<locals>.NodeBuilder.getAttributescSsv|rrt|���D]`\}}t|t�rb|ddurD|dd|d}n|d}|j�|d||�q|j�||�qdS)Nr�:rr
)r%r&rrrZsetAttributeNSZsetAttribute)rrr rZ
qualifiedNamerrr�
setAttributes^s
��z0getDomBuilder.<locals>.NodeBuilder.setAttributescs�|j�d��S)NF)r�	cloneNoder��NodeBuilderrrrFmsz,getDomBuilder.<locals>.NodeBuilder.cloneNodecSs
|j��Sr)rr>rrrr�
hasContentpsz-getDomBuilder.<locals>.NodeBuilder.hasContentcSs(|jdurtd|jfS|j|jfSdS)N�html)�	namespacerr rrrr�getNameTupless
z/getDomBuilder.<locals>.NodeBuilder.getNameTuple)N)r+r,r-r�propertyrKr4r;r8r=rBrCrErrFrIrL�	nameTupler)r.rHrrrH8s

rHcs�eZdZ�fdd�Z��fdd�Zd�fdd�	Z�fdd	�Z�fd
d�Zdd
�Z�fdd�Z	dd�Z
dd�Zddd�Z�Z
dZdS)z"getDomBuilder.<locals>.TreeBuildercs����ddd�|_t�|�Sr)�getDOMImplementationZcreateDocument�dom�weakref�proxyr)�Domrr�
documentClass|sz0getDomBuilder.<locals>.TreeBuilder.documentClasscsR|d}|d}|d}���}|�|||�}|j��|���tkrN|j|_dS)Nr �publicId�systemId)rOZcreateDocumentType�documentr4rrPr)r�tokenr rUrVZdomimpl�doctype)rSrHrr�
insertDoctype�sz0getDomBuilder.<locals>.TreeBuilder.insertDoctypeNcs6|dur |jdur |j�|�}n|j�||�}�|�Sr)�defaultNamespacerP�
createElementZcreateElementNS)rr rKr6rGrr�elementClass�sz/getDomBuilder.<locals>.TreeBuilder.elementClasscs�|j�|��Sr)rPZ
createComment)rr9rGrr�commentClass�sz/getDomBuilder.<locals>.TreeBuilder.commentClasscs�|j���Sr)rPZcreateDocumentFragmentrrGrr�
fragmentClass�sz0getDomBuilder.<locals>.TreeBuilder.fragmentClasscSs|j�|j�dSr)rPr4rr5rrrr4�sz.getDomBuilder.<locals>.TreeBuilder.appendChildcs�|�Srrr)�testSerializerrrr`�sz1getDomBuilder.<locals>.TreeBuilder.testSerializercSs|jSr)rPrrrr�getDocument�sz.getDomBuilder.<locals>.TreeBuilder.getDocumentcSstj�|�jSr)r	�TreeBuilder�getFragmentrrrrrrc�sz.getDomBuilder.<locals>.TreeBuilder.getFragmentcSsp|}||krtj�|||�nNt|jd�rXtj|jjvrXt|jj�|j_|jj�	tj�|j�
|j�|��dS)N�_child_node_types)r	rbr;r1rPr�	TEXT_NODErdr%�appendr4r7)rr9r3rrrr;�sz-getDomBuilder.<locals>.TreeBuilder.insertText)N)N)r+r,r-rTrZr]r^r_r4r`rarcr;�implementationr r)rS�DomImplementationrHr`rrrb{s

rbcs0|��g�d��fdd�	��|d�d���S)Nrcs|jtjkr�|jrj|js|jrP|jp&d}|jp0d}��dd||j||f�q~��dd||jf�n��dd|f��nr|jtjkr���d��nX|jtjkr���d��n>|jtj	krވ�dd||j
f��n|jtjk�r��d	d||j
f�n�t|d
��r6|j
du�r6dtj|j
|jf}n|j}��dd||f�|���r�g}tt|j��D]T}|j�|�}|j}|j}|j
}	|	�r�dtj|	|jf}n|j}|�||f��qnt|�D]&\}}��d
d|d||f��q�|d7}|jD]}
�|
|��qdS)N�z|%s<!DOCTYPE %s "%s" "%s">� z|%s<!DOCTYPE %s>z|%s<!DOCTYPE >z	#documentz#document-fragmentz|%s<!-- %s -->z|%s"%s"r0z%s %sz|%s<%s>z
|%s%s="%s"r
)ZnodeTyperZDOCUMENT_TYPE_NODEr rUrVrfZ
DOCUMENT_NODEZDOCUMENT_FRAGMENT_NODEZCOMMENT_NODEZ	nodeValuerer1r0r�prefixesr/Z
hasAttributes�ranger#r�itemrZ	localName�sortedr?)r�indentrUrVr r�ir!r�nsrA��rv�serializeElementrrrt�sT

���
 
z?getDomBuilder.<locals>.testSerializer.<locals>.serializeElement�
)r)�	normalize�joinrrrrrr`�s
.
z%getDomBuilder.<locals>.testSerializer)rr	rrb�locals)rhrbr)r.rSrhrHr`r�
getDomBuilders$C:6ry)�
__future__rrr�collections.abcr�ImportError�collections�xml.domrrrQrir	rr�_utilsr
ry�getDomModulerrrr�<module>s_site-packages/pip/_vendor/html5lib/treebuilders/base.py000064400000034345150564745010017225 0ustar00from __future__ import absolute_import, division, unicode_literals
from pip._vendor.six import text_type

from ..constants import scopingElements, tableInsertModeElements, namespaces

# The scope markers are inserted when entering object elements,
# marquees, table cells, and table captions, and are used to prevent formatting
# from "leaking" into tables, object elements, and marquees.
Marker = None

listElementsMap = {
    None: (frozenset(scopingElements), False),
    "button": (frozenset(scopingElements | {(namespaces["html"], "button")}), False),
    "list": (frozenset(scopingElements | {(namespaces["html"], "ol"),
                                          (namespaces["html"], "ul")}), False),
    "table": (frozenset([(namespaces["html"], "html"),
                         (namespaces["html"], "table")]), False),
    "select": (frozenset([(namespaces["html"], "optgroup"),
                          (namespaces["html"], "option")]), True)
}


class Node(object):
    """Represents an item in the tree"""
    def __init__(self, name):
        """Creates a Node

        :arg name: The tag name associated with the node

        """
        # The tag name associated with the node
        self.name = name
        # The parent of the current node (or None for the document node)
        self.parent = None
        # The value of the current node (applies to text nodes and comments)
        self.value = None
        # A dict holding name -> value pairs for attributes of the node
        self.attributes = {}
        # A list of child nodes of the current node. This must include all
        # elements but not necessarily other node types.
        self.childNodes = []
        # A list of miscellaneous flags that can be set on the node.
        self._flags = []

    def __str__(self):
        attributesStr = " ".join(["%s=\"%s\"" % (name, value)
                                  for name, value in
                                  self.attributes.items()])
        if attributesStr:
            return "<%s %s>" % (self.name, attributesStr)
        else:
            return "<%s>" % (self.name)

    def __repr__(self):
        return "<%s>" % (self.name)

    def appendChild(self, node):
        """Insert node as a child of the current node

        :arg node: the node to insert

        """
        raise NotImplementedError

    def insertText(self, data, insertBefore=None):
        """Insert data as text in the current node, positioned before the
        start of node insertBefore or to the end of the node's text.

        :arg data: the data to insert

        :arg insertBefore: True if you want to insert the text before the node
            and False if you want to insert it after the node

        """
        raise NotImplementedError

    def insertBefore(self, node, refNode):
        """Insert node as a child of the current node, before refNode in the
        list of child nodes. Raises ValueError if refNode is not a child of
        the current node

        :arg node: the node to insert

        :arg refNode: the child node to insert the node before

        """
        raise NotImplementedError

    def removeChild(self, node):
        """Remove node from the children of the current node

        :arg node: the child node to remove

        """
        raise NotImplementedError

    def reparentChildren(self, newParent):
        """Move all the children of the current node to newParent.
        This is needed so that trees that don't store text as nodes move the
        text in the correct way

        :arg newParent: the node to move all this node's children to

        """
        # XXX - should this method be made more general?
        for child in self.childNodes:
            newParent.appendChild(child)
        self.childNodes = []

    def cloneNode(self):
        """Return a shallow copy of the current node i.e. a node with the same
        name and attributes but with no parent or child nodes
        """
        raise NotImplementedError

    def hasContent(self):
        """Return true if the node has children or text, false otherwise
        """
        raise NotImplementedError


class ActiveFormattingElements(list):
    def append(self, node):
        equalCount = 0
        if node != Marker:
            for element in self[::-1]:
                if element == Marker:
                    break
                if self.nodesEqual(element, node):
                    equalCount += 1
                if equalCount == 3:
                    self.remove(element)
                    break
        list.append(self, node)

    def nodesEqual(self, node1, node2):
        if not node1.nameTuple == node2.nameTuple:
            return False

        if not node1.attributes == node2.attributes:
            return False

        return True


class TreeBuilder(object):
    """Base treebuilder implementation

    * documentClass - the class to use for the bottommost node of a document
    * elementClass - the class to use for HTML Elements
    * commentClass - the class to use for comments
    * doctypeClass - the class to use for doctypes

    """
    # pylint:disable=not-callable

    # Document class
    documentClass = None

    # The class to use for creating a node
    elementClass = None

    # The class to use for creating comments
    commentClass = None

    # The class to use for creating doctypes
    doctypeClass = None

    # Fragment class
    fragmentClass = None

    def __init__(self, namespaceHTMLElements):
        """Create a TreeBuilder

        :arg namespaceHTMLElements: whether or not to namespace HTML elements

        """
        if namespaceHTMLElements:
            self.defaultNamespace = "http://www.w3.org/1999/xhtml"
        else:
            self.defaultNamespace = None
        self.reset()

    def reset(self):
        self.openElements = []
        self.activeFormattingElements = ActiveFormattingElements()

        # XXX - rename these to headElement, formElement
        self.headPointer = None
        self.formPointer = None

        self.insertFromTable = False

        self.document = self.documentClass()

    def elementInScope(self, target, variant=None):

        # If we pass a node in we match that. if we pass a string
        # match any node with that name
        exactNode = hasattr(target, "nameTuple")
        if not exactNode:
            if isinstance(target, text_type):
                target = (namespaces["html"], target)
            assert isinstance(target, tuple)

        listElements, invert = listElementsMap[variant]

        for node in reversed(self.openElements):
            if exactNode and node == target:
                return True
            elif not exactNode and node.nameTuple == target:
                return True
            elif (invert ^ (node.nameTuple in listElements)):
                return False

        assert False  # We should never reach this point

    def reconstructActiveFormattingElements(self):
        # Within this algorithm the order of steps described in the
        # specification is not quite the same as the order of steps in the
        # code. It should still do the same though.

        # Step 1: stop the algorithm when there's nothing to do.
        if not self.activeFormattingElements:
            return

        # Step 2 and step 3: we start with the last element. So i is -1.
        i = len(self.activeFormattingElements) - 1
        entry = self.activeFormattingElements[i]
        if entry == Marker or entry in self.openElements:
            return

        # Step 6
        while entry != Marker and entry not in self.openElements:
            if i == 0:
                # This will be reset to 0 below
                i = -1
                break
            i -= 1
            # Step 5: let entry be one earlier in the list.
            entry = self.activeFormattingElements[i]

        while True:
            # Step 7
            i += 1

            # Step 8
            entry = self.activeFormattingElements[i]
            clone = entry.cloneNode()  # Mainly to get a new copy of the attributes

            # Step 9
            element = self.insertElement({"type": "StartTag",
                                          "name": clone.name,
                                          "namespace": clone.namespace,
                                          "data": clone.attributes})

            # Step 10
            self.activeFormattingElements[i] = element

            # Step 11
            if element == self.activeFormattingElements[-1]:
                break

    def clearActiveFormattingElements(self):
        entry = self.activeFormattingElements.pop()
        while self.activeFormattingElements and entry != Marker:
            entry = self.activeFormattingElements.pop()

    def elementInActiveFormattingElements(self, name):
        """Check if an element exists between the end of the active
        formatting elements and the last marker. If it does, return it, else
        return false"""

        for item in self.activeFormattingElements[::-1]:
            # Check for Marker first because if it's a Marker it doesn't have a
            # name attribute.
            if item == Marker:
                break
            elif item.name == name:
                return item
        return False

    def insertRoot(self, token):
        element = self.createElement(token)
        self.openElements.append(element)
        self.document.appendChild(element)

    def insertDoctype(self, token):
        name = token["name"]
        publicId = token["publicId"]
        systemId = token["systemId"]

        doctype = self.doctypeClass(name, publicId, systemId)
        self.document.appendChild(doctype)

    def insertComment(self, token, parent=None):
        if parent is None:
            parent = self.openElements[-1]
        parent.appendChild(self.commentClass(token["data"]))

    def createElement(self, token):
        """Create an element but don't insert it anywhere"""
        name = token["name"]
        namespace = token.get("namespace", self.defaultNamespace)
        element = self.elementClass(name, namespace)
        element.attributes = token["data"]
        return element

    def _getInsertFromTable(self):
        return self._insertFromTable

    def _setInsertFromTable(self, value):
        """Switch the function used to insert an element from the
        normal one to the misnested table one and back again"""
        self._insertFromTable = value
        if value:
            self.insertElement = self.insertElementTable
        else:
            self.insertElement = self.insertElementNormal

    insertFromTable = property(_getInsertFromTable, _setInsertFromTable)

    def insertElementNormal(self, token):
        name = token["name"]
        assert isinstance(name, text_type), "Element %s not unicode" % name
        namespace = token.get("namespace", self.defaultNamespace)
        element = self.elementClass(name, namespace)
        element.attributes = token["data"]
        self.openElements[-1].appendChild(element)
        self.openElements.append(element)
        return element

    def insertElementTable(self, token):
        """Create an element and insert it into the tree"""
        element = self.createElement(token)
        if self.openElements[-1].name not in tableInsertModeElements:
            return self.insertElementNormal(token)
        else:
            # We should be in the InTable mode. This means we want to do
            # special magic element rearranging
            parent, insertBefore = self.getTableMisnestedNodePosition()
            if insertBefore is None:
                parent.appendChild(element)
            else:
                parent.insertBefore(element, insertBefore)
            self.openElements.append(element)
        return element

    def insertText(self, data, parent=None):
        """Insert text data."""
        if parent is None:
            parent = self.openElements[-1]

        if (not self.insertFromTable or (self.insertFromTable and
                                         self.openElements[-1].name
                                         not in tableInsertModeElements)):
            parent.insertText(data)
        else:
            # We should be in the InTable mode. This means we want to do
            # special magic element rearranging
            parent, insertBefore = self.getTableMisnestedNodePosition()
            parent.insertText(data, insertBefore)

    def getTableMisnestedNodePosition(self):
        """Get the foster parent element, and sibling to insert before
        (or None) when inserting a misnested table node"""
        # The foster parent element is the one which comes before the most
        # recently opened table element
        # XXX - this is really inelegant
        lastTable = None
        fosterParent = None
        insertBefore = None
        for elm in self.openElements[::-1]:
            if elm.name == "table":
                lastTable = elm
                break
        if lastTable:
            # XXX - we should really check that this parent is actually a
            # node here
            if lastTable.parent:
                fosterParent = lastTable.parent
                insertBefore = lastTable
            else:
                fosterParent = self.openElements[
                    self.openElements.index(lastTable) - 1]
        else:
            fosterParent = self.openElements[0]
        return fosterParent, insertBefore

    def generateImpliedEndTags(self, exclude=None):
        name = self.openElements[-1].name
        # XXX td, th and tr are not actually needed
        if (name in frozenset(("dd", "dt", "li", "option", "optgroup", "p", "rp", "rt")) and
                name != exclude):
            self.openElements.pop()
            # XXX This is not entirely what the specification says. We should
            # investigate it more closely.
            self.generateImpliedEndTags(exclude)

    def getDocument(self):
        """Return the final tree"""
        return self.document

    def getFragment(self):
        """Return the final fragment"""
        # assert self.innerHTML
        fragment = self.fragmentClass()
        self.openElements[0].reparentChildren(fragment)
        return fragment

    def testSerializer(self, node):
        """Serialize the subtree of node in the format required by unit tests

        :arg node: the node from which to start serializing

        """
        raise NotImplementedError
site-packages/pip/_vendor/html5lib/treebuilders/etree.py000064400000031044150564745010017410 0ustar00from __future__ import absolute_import, division, unicode_literals
# pylint:disable=protected-access

from pip._vendor.six import text_type

import re

from copy import copy

from . import base
from .. import _ihatexml
from .. import constants
from ..constants import namespaces
from .._utils import moduleFactoryFactory

tag_regexp = re.compile("{([^}]*)}(.*)")


def getETreeBuilder(ElementTreeImplementation, fullTree=False):
    ElementTree = ElementTreeImplementation
    ElementTreeCommentType = ElementTree.Comment("asd").tag

    class Element(base.Node):
        def __init__(self, name, namespace=None):
            self._name = name
            self._namespace = namespace
            self._element = ElementTree.Element(self._getETreeTag(name,
                                                                  namespace))
            if namespace is None:
                self.nameTuple = namespaces["html"], self._name
            else:
                self.nameTuple = self._namespace, self._name
            self.parent = None
            self._childNodes = []
            self._flags = []

        def _getETreeTag(self, name, namespace):
            if namespace is None:
                etree_tag = name
            else:
                etree_tag = "{%s}%s" % (namespace, name)
            return etree_tag

        def _setName(self, name):
            self._name = name
            self._element.tag = self._getETreeTag(self._name, self._namespace)

        def _getName(self):
            return self._name

        name = property(_getName, _setName)

        def _setNamespace(self, namespace):
            self._namespace = namespace
            self._element.tag = self._getETreeTag(self._name, self._namespace)

        def _getNamespace(self):
            return self._namespace

        namespace = property(_getNamespace, _setNamespace)

        def _getAttributes(self):
            return self._element.attrib

        def _setAttributes(self, attributes):
            el_attrib = self._element.attrib
            el_attrib.clear()
            if attributes:
                # calling .items _always_ allocates, and the above truthy check is cheaper than the
                # allocation on average
                for key, value in attributes.items():
                    if isinstance(key, tuple):
                        name = "{%s}%s" % (key[2], key[1])
                    else:
                        name = key
                    el_attrib[name] = value

        attributes = property(_getAttributes, _setAttributes)

        def _getChildNodes(self):
            return self._childNodes

        def _setChildNodes(self, value):
            del self._element[:]
            self._childNodes = []
            for element in value:
                self.insertChild(element)

        childNodes = property(_getChildNodes, _setChildNodes)

        def hasContent(self):
            """Return true if the node has children or text"""
            return bool(self._element.text or len(self._element))

        def appendChild(self, node):
            self._childNodes.append(node)
            self._element.append(node._element)
            node.parent = self

        def insertBefore(self, node, refNode):
            index = list(self._element).index(refNode._element)
            self._element.insert(index, node._element)
            node.parent = self

        def removeChild(self, node):
            self._childNodes.remove(node)
            self._element.remove(node._element)
            node.parent = None

        def insertText(self, data, insertBefore=None):
            if not(len(self._element)):
                if not self._element.text:
                    self._element.text = ""
                self._element.text += data
            elif insertBefore is None:
                # Insert the text as the tail of the last child element
                if not self._element[-1].tail:
                    self._element[-1].tail = ""
                self._element[-1].tail += data
            else:
                # Insert the text before the specified node
                children = list(self._element)
                index = children.index(insertBefore._element)
                if index > 0:
                    if not self._element[index - 1].tail:
                        self._element[index - 1].tail = ""
                    self._element[index - 1].tail += data
                else:
                    if not self._element.text:
                        self._element.text = ""
                    self._element.text += data

        def cloneNode(self):
            element = type(self)(self.name, self.namespace)
            if self._element.attrib:
                element._element.attrib = copy(self._element.attrib)
            return element

        def reparentChildren(self, newParent):
            if newParent.childNodes:
                newParent.childNodes[-1]._element.tail += self._element.text
            else:
                if not newParent._element.text:
                    newParent._element.text = ""
                if self._element.text is not None:
                    newParent._element.text += self._element.text
            self._element.text = ""
            base.Node.reparentChildren(self, newParent)

    class Comment(Element):
        def __init__(self, data):
            # Use the superclass constructor to set all properties on the
            # wrapper element
            self._element = ElementTree.Comment(data)
            self.parent = None
            self._childNodes = []
            self._flags = []

        def _getData(self):
            return self._element.text

        def _setData(self, value):
            self._element.text = value

        data = property(_getData, _setData)

    class DocumentType(Element):
        def __init__(self, name, publicId, systemId):
            Element.__init__(self, "<!DOCTYPE>")
            self._element.text = name
            self.publicId = publicId
            self.systemId = systemId

        def _getPublicId(self):
            return self._element.get("publicId", "")

        def _setPublicId(self, value):
            if value is not None:
                self._element.set("publicId", value)

        publicId = property(_getPublicId, _setPublicId)

        def _getSystemId(self):
            return self._element.get("systemId", "")

        def _setSystemId(self, value):
            if value is not None:
                self._element.set("systemId", value)

        systemId = property(_getSystemId, _setSystemId)

    class Document(Element):
        def __init__(self):
            Element.__init__(self, "DOCUMENT_ROOT")

    class DocumentFragment(Element):
        def __init__(self):
            Element.__init__(self, "DOCUMENT_FRAGMENT")

    def testSerializer(element):
        rv = []

        def serializeElement(element, indent=0):
            if not(hasattr(element, "tag")):
                element = element.getroot()
            if element.tag == "<!DOCTYPE>":
                if element.get("publicId") or element.get("systemId"):
                    publicId = element.get("publicId") or ""
                    systemId = element.get("systemId") or ""
                    rv.append("""<!DOCTYPE %s "%s" "%s">""" %
                              (element.text, publicId, systemId))
                else:
                    rv.append("<!DOCTYPE %s>" % (element.text,))
            elif element.tag == "DOCUMENT_ROOT":
                rv.append("#document")
                if element.text is not None:
                    rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text))
                if element.tail is not None:
                    raise TypeError("Document node cannot have tail")
                if hasattr(element, "attrib") and len(element.attrib):
                    raise TypeError("Document node cannot have attributes")
            elif element.tag == ElementTreeCommentType:
                rv.append("|%s<!-- %s -->" % (' ' * indent, element.text))
            else:
                assert isinstance(element.tag, text_type), \
                    "Expected unicode, got %s, %s" % (type(element.tag), element.tag)
                nsmatch = tag_regexp.match(element.tag)

                if nsmatch is None:
                    name = element.tag
                else:
                    ns, name = nsmatch.groups()
                    prefix = constants.prefixes[ns]
                    name = "%s %s" % (prefix, name)
                rv.append("|%s<%s>" % (' ' * indent, name))

                if hasattr(element, "attrib"):
                    attributes = []
                    for name, value in element.attrib.items():
                        nsmatch = tag_regexp.match(name)
                        if nsmatch is not None:
                            ns, name = nsmatch.groups()
                            prefix = constants.prefixes[ns]
                            attr_string = "%s %s" % (prefix, name)
                        else:
                            attr_string = name
                        attributes.append((attr_string, value))

                    for name, value in sorted(attributes):
                        rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value))
                if element.text:
                    rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text))
            indent += 2
            for child in element:
                serializeElement(child, indent)
            if element.tail:
                rv.append("|%s\"%s\"" % (' ' * (indent - 2), element.tail))
        serializeElement(element, 0)

        return "\n".join(rv)

    def tostring(element):  # pylint:disable=unused-variable
        """Serialize an element and its child nodes to a string"""
        rv = []
        filter = _ihatexml.InfosetFilter()

        def serializeElement(element):
            if isinstance(element, ElementTree.ElementTree):
                element = element.getroot()

            if element.tag == "<!DOCTYPE>":
                if element.get("publicId") or element.get("systemId"):
                    publicId = element.get("publicId") or ""
                    systemId = element.get("systemId") or ""
                    rv.append("""<!DOCTYPE %s PUBLIC "%s" "%s">""" %
                              (element.text, publicId, systemId))
                else:
                    rv.append("<!DOCTYPE %s>" % (element.text,))
            elif element.tag == "DOCUMENT_ROOT":
                if element.text is not None:
                    rv.append(element.text)
                if element.tail is not None:
                    raise TypeError("Document node cannot have tail")
                if hasattr(element, "attrib") and len(element.attrib):
                    raise TypeError("Document node cannot have attributes")

                for child in element:
                    serializeElement(child)

            elif element.tag == ElementTreeCommentType:
                rv.append("<!--%s-->" % (element.text,))
            else:
                # This is assumed to be an ordinary element
                if not element.attrib:
                    rv.append("<%s>" % (filter.fromXmlName(element.tag),))
                else:
                    attr = " ".join(["%s=\"%s\"" % (
                        filter.fromXmlName(name), value)
                        for name, value in element.attrib.items()])
                    rv.append("<%s %s>" % (element.tag, attr))
                if element.text:
                    rv.append(element.text)

                for child in element:
                    serializeElement(child)

                rv.append("</%s>" % (element.tag,))

            if element.tail:
                rv.append(element.tail)

        serializeElement(element)

        return "".join(rv)

    class TreeBuilder(base.TreeBuilder):  # pylint:disable=unused-variable
        documentClass = Document
        doctypeClass = DocumentType
        elementClass = Element
        commentClass = Comment
        fragmentClass = DocumentFragment
        implementation = ElementTreeImplementation

        def testSerializer(self, element):
            return testSerializer(element)

        def getDocument(self):
            if fullTree:
                return self.document._element
            else:
                if self.defaultNamespace is not None:
                    return self.document._element.find(
                        "{%s}html" % self.defaultNamespace)
                else:
                    return self.document._element.find("html")

        def getFragment(self):
            return base.TreeBuilder.getFragment(self)._element

    return locals()


getETreeModule = moduleFactoryFactory(getETreeBuilder)
site-packages/pip/_vendor/html5lib/treebuilders/dom.py000064400000021335150564745010017065 0ustar00from __future__ import absolute_import, division, unicode_literals


try:
    from collections.abc import MutableMapping
except ImportError:  # Python 2.7
    from collections import MutableMapping
from xml.dom import minidom, Node
import weakref

from . import base
from .. import constants
from ..constants import namespaces
from .._utils import moduleFactoryFactory


def getDomBuilder(DomImplementation):
    Dom = DomImplementation

    class AttrList(MutableMapping):
        def __init__(self, element):
            self.element = element

        def __iter__(self):
            return iter(self.element.attributes.keys())

        def __setitem__(self, name, value):
            if isinstance(name, tuple):
                raise NotImplementedError
            else:
                attr = self.element.ownerDocument.createAttribute(name)
                attr.value = value
                self.element.attributes[name] = attr

        def __len__(self):
            return len(self.element.attributes)

        def items(self):
            return list(self.element.attributes.items())

        def values(self):
            return list(self.element.attributes.values())

        def __getitem__(self, name):
            if isinstance(name, tuple):
                raise NotImplementedError
            else:
                return self.element.attributes[name].value

        def __delitem__(self, name):
            if isinstance(name, tuple):
                raise NotImplementedError
            else:
                del self.element.attributes[name]

    class NodeBuilder(base.Node):
        def __init__(self, element):
            base.Node.__init__(self, element.nodeName)
            self.element = element

        namespace = property(lambda self: hasattr(self.element, "namespaceURI") and
                             self.element.namespaceURI or None)

        def appendChild(self, node):
            node.parent = self
            self.element.appendChild(node.element)

        def insertText(self, data, insertBefore=None):
            text = self.element.ownerDocument.createTextNode(data)
            if insertBefore:
                self.element.insertBefore(text, insertBefore.element)
            else:
                self.element.appendChild(text)

        def insertBefore(self, node, refNode):
            self.element.insertBefore(node.element, refNode.element)
            node.parent = self

        def removeChild(self, node):
            if node.element.parentNode == self.element:
                self.element.removeChild(node.element)
            node.parent = None

        def reparentChildren(self, newParent):
            while self.element.hasChildNodes():
                child = self.element.firstChild
                self.element.removeChild(child)
                newParent.element.appendChild(child)
            self.childNodes = []

        def getAttributes(self):
            return AttrList(self.element)

        def setAttributes(self, attributes):
            if attributes:
                for name, value in list(attributes.items()):
                    if isinstance(name, tuple):
                        if name[0] is not None:
                            qualifiedName = (name[0] + ":" + name[1])
                        else:
                            qualifiedName = name[1]
                        self.element.setAttributeNS(name[2], qualifiedName,
                                                    value)
                    else:
                        self.element.setAttribute(
                            name, value)
        attributes = property(getAttributes, setAttributes)

        def cloneNode(self):
            return NodeBuilder(self.element.cloneNode(False))

        def hasContent(self):
            return self.element.hasChildNodes()

        def getNameTuple(self):
            if self.namespace is None:
                return namespaces["html"], self.name
            else:
                return self.namespace, self.name

        nameTuple = property(getNameTuple)

    class TreeBuilder(base.TreeBuilder):  # pylint:disable=unused-variable
        def documentClass(self):
            self.dom = Dom.getDOMImplementation().createDocument(None, None, None)
            return weakref.proxy(self)

        def insertDoctype(self, token):
            name = token["name"]
            publicId = token["publicId"]
            systemId = token["systemId"]

            domimpl = Dom.getDOMImplementation()
            doctype = domimpl.createDocumentType(name, publicId, systemId)
            self.document.appendChild(NodeBuilder(doctype))
            if Dom == minidom:
                doctype.ownerDocument = self.dom

        def elementClass(self, name, namespace=None):
            if namespace is None and self.defaultNamespace is None:
                node = self.dom.createElement(name)
            else:
                node = self.dom.createElementNS(namespace, name)

            return NodeBuilder(node)

        def commentClass(self, data):
            return NodeBuilder(self.dom.createComment(data))

        def fragmentClass(self):
            return NodeBuilder(self.dom.createDocumentFragment())

        def appendChild(self, node):
            self.dom.appendChild(node.element)

        def testSerializer(self, element):
            return testSerializer(element)

        def getDocument(self):
            return self.dom

        def getFragment(self):
            return base.TreeBuilder.getFragment(self).element

        def insertText(self, data, parent=None):
            data = data
            if parent != self:
                base.TreeBuilder.insertText(self, data, parent)
            else:
                # HACK: allow text nodes as children of the document node
                if hasattr(self.dom, '_child_node_types'):
                    # pylint:disable=protected-access
                    if Node.TEXT_NODE not in self.dom._child_node_types:
                        self.dom._child_node_types = list(self.dom._child_node_types)
                        self.dom._child_node_types.append(Node.TEXT_NODE)
                self.dom.appendChild(self.dom.createTextNode(data))

        implementation = DomImplementation
        name = None

    def testSerializer(element):
        element.normalize()
        rv = []

        def serializeElement(element, indent=0):
            if element.nodeType == Node.DOCUMENT_TYPE_NODE:
                if element.name:
                    if element.publicId or element.systemId:
                        publicId = element.publicId or ""
                        systemId = element.systemId or ""
                        rv.append("""|%s<!DOCTYPE %s "%s" "%s">""" %
                                  (' ' * indent, element.name, publicId, systemId))
                    else:
                        rv.append("|%s<!DOCTYPE %s>" % (' ' * indent, element.name))
                else:
                    rv.append("|%s<!DOCTYPE >" % (' ' * indent,))
            elif element.nodeType == Node.DOCUMENT_NODE:
                rv.append("#document")
            elif element.nodeType == Node.DOCUMENT_FRAGMENT_NODE:
                rv.append("#document-fragment")
            elif element.nodeType == Node.COMMENT_NODE:
                rv.append("|%s<!-- %s -->" % (' ' * indent, element.nodeValue))
            elif element.nodeType == Node.TEXT_NODE:
                rv.append("|%s\"%s\"" % (' ' * indent, element.nodeValue))
            else:
                if (hasattr(element, "namespaceURI") and
                        element.namespaceURI is not None):
                    name = "%s %s" % (constants.prefixes[element.namespaceURI],
                                      element.nodeName)
                else:
                    name = element.nodeName
                rv.append("|%s<%s>" % (' ' * indent, name))
                if element.hasAttributes():
                    attributes = []
                    for i in range(len(element.attributes)):
                        attr = element.attributes.item(i)
                        name = attr.nodeName
                        value = attr.value
                        ns = attr.namespaceURI
                        if ns:
                            name = "%s %s" % (constants.prefixes[ns], attr.localName)
                        else:
                            name = attr.nodeName
                        attributes.append((name, value))

                    for name, value in sorted(attributes):
                        rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value))
            indent += 2
            for child in element.childNodes:
                serializeElement(child, indent)
        serializeElement(element, 0)

        return "\n".join(rv)

    return locals()


# The actual means to get a module!
getDomModule = moduleFactoryFactory(getDomBuilder)
site-packages/pip/_vendor/html5lib/treebuilders/__init__.py000064400000007010150564745010020037 0ustar00"""A collection of modules for building different kinds of trees from HTML
documents.

To create a treebuilder for a new type of tree, you need to do
implement several things:

1. A set of classes for various types of elements: Document, Doctype, Comment,
   Element. These must implement the interface of ``base.treebuilders.Node``
   (although comment nodes have a different signature for their constructor,
   see ``treebuilders.etree.Comment``) Textual content may also be implemented
   as another node type, or not, as your tree implementation requires.

2. A treebuilder object (called ``TreeBuilder`` by convention) that inherits
   from ``treebuilders.base.TreeBuilder``. This has 4 required attributes:

   * ``documentClass`` - the class to use for the bottommost node of a document
   * ``elementClass`` - the class to use for HTML Elements
   * ``commentClass`` - the class to use for comments
   * ``doctypeClass`` - the class to use for doctypes

   It also has one required method:

   * ``getDocument`` - Returns the root node of the complete document tree

3. If you wish to run the unit tests, you must also create a ``testSerializer``
   method on your treebuilder which accepts a node and returns a string
   containing Node and its children serialized according to the format used in
   the unittests

"""

from __future__ import absolute_import, division, unicode_literals

from .._utils import default_etree

treeBuilderCache = {}


def getTreeBuilder(treeType, implementation=None, **kwargs):
    """Get a TreeBuilder class for various types of trees with built-in support

    :arg treeType: the name of the tree type required (case-insensitive). Supported
        values are:

        * "dom" - A generic builder for DOM implementations, defaulting to a
          xml.dom.minidom based implementation.
        * "etree" - A generic builder for tree implementations exposing an
          ElementTree-like interface, defaulting to xml.etree.cElementTree if
          available and xml.etree.ElementTree if not.
        * "lxml" - A etree-based builder for lxml.etree, handling limitations
          of lxml's implementation.

    :arg implementation: (Currently applies to the "etree" and "dom" tree
        types). A module implementing the tree type e.g. xml.etree.ElementTree
        or xml.etree.cElementTree.

    :arg kwargs: Any additional options to pass to the TreeBuilder when
        creating it.

    Example:

    >>> from html5lib.treebuilders import getTreeBuilder
    >>> builder = getTreeBuilder('etree')

    """

    treeType = treeType.lower()
    if treeType not in treeBuilderCache:
        if treeType == "dom":
            from . import dom
            # Come up with a sane default (pref. from the stdlib)
            if implementation is None:
                from xml.dom import minidom
                implementation = minidom
            # NEVER cache here, caching is done in the dom submodule
            return dom.getDomModule(implementation, **kwargs).TreeBuilder
        elif treeType == "lxml":
            from . import etree_lxml
            treeBuilderCache[treeType] = etree_lxml.TreeBuilder
        elif treeType == "etree":
            from . import etree
            if implementation is None:
                implementation = default_etree
            # NEVER cache here, caching is done in the etree submodule
            return etree.getETreeModule(implementation, **kwargs).TreeBuilder
        else:
            raise ValueError("""Unrecognised treebuilder "%s" """ % treeType)
    return treeBuilderCache.get(treeType)
site-packages/pip/_vendor/html5lib/treebuilders/etree_lxml.py000064400000034656150564745010020460 0ustar00"""Module for supporting the lxml.etree library. The idea here is to use as much
of the native library as possible, without using fragile hacks like custom element
names that break between releases. The downside of this is that we cannot represent
all possible trees; specifically the following are known to cause problems:

Text or comments as siblings of the root element
Docypes with no name

When any of these things occur, we emit a DataLossWarning
"""

from __future__ import absolute_import, division, unicode_literals
# pylint:disable=protected-access

import warnings
import re
import sys

try:
    from collections.abc import MutableMapping
except ImportError:
    from collections import MutableMapping

from . import base
from ..constants import DataLossWarning
from .. import constants
from . import etree as etree_builders
from .. import _ihatexml

import lxml.etree as etree
from pip._vendor.six import PY3, binary_type


fullTree = True
tag_regexp = re.compile("{([^}]*)}(.*)")

comment_type = etree.Comment("asd").tag


class DocumentType(object):
    def __init__(self, name, publicId, systemId):
        self.name = name
        self.publicId = publicId
        self.systemId = systemId


class Document(object):
    def __init__(self):
        self._elementTree = None
        self._childNodes = []

    def appendChild(self, element):
        last = self._elementTree.getroot()
        for last in self._elementTree.getroot().itersiblings():
            pass

        last.addnext(element._element)

    def _getChildNodes(self):
        return self._childNodes

    childNodes = property(_getChildNodes)


def testSerializer(element):
    rv = []
    infosetFilter = _ihatexml.InfosetFilter(preventDoubleDashComments=True)

    def serializeElement(element, indent=0):
        if not hasattr(element, "tag"):
            if hasattr(element, "getroot"):
                # Full tree case
                rv.append("#document")
                if element.docinfo.internalDTD:
                    if not (element.docinfo.public_id or
                            element.docinfo.system_url):
                        dtd_str = "<!DOCTYPE %s>" % element.docinfo.root_name
                    else:
                        dtd_str = """<!DOCTYPE %s "%s" "%s">""" % (
                            element.docinfo.root_name,
                            element.docinfo.public_id,
                            element.docinfo.system_url)
                    rv.append("|%s%s" % (' ' * (indent + 2), dtd_str))
                next_element = element.getroot()
                while next_element.getprevious() is not None:
                    next_element = next_element.getprevious()
                while next_element is not None:
                    serializeElement(next_element, indent + 2)
                    next_element = next_element.getnext()
            elif isinstance(element, str) or isinstance(element, bytes):
                # Text in a fragment
                assert isinstance(element, str) or sys.version_info[0] == 2
                rv.append("|%s\"%s\"" % (' ' * indent, element))
            else:
                # Fragment case
                rv.append("#document-fragment")
                for next_element in element:
                    serializeElement(next_element, indent + 2)
        elif element.tag == comment_type:
            rv.append("|%s<!-- %s -->" % (' ' * indent, element.text))
            if hasattr(element, "tail") and element.tail:
                rv.append("|%s\"%s\"" % (' ' * indent, element.tail))
        else:
            assert isinstance(element, etree._Element)
            nsmatch = etree_builders.tag_regexp.match(element.tag)
            if nsmatch is not None:
                ns = nsmatch.group(1)
                tag = nsmatch.group(2)
                prefix = constants.prefixes[ns]
                rv.append("|%s<%s %s>" % (' ' * indent, prefix,
                                          infosetFilter.fromXmlName(tag)))
            else:
                rv.append("|%s<%s>" % (' ' * indent,
                                       infosetFilter.fromXmlName(element.tag)))

            if hasattr(element, "attrib"):
                attributes = []
                for name, value in element.attrib.items():
                    nsmatch = tag_regexp.match(name)
                    if nsmatch is not None:
                        ns, name = nsmatch.groups()
                        name = infosetFilter.fromXmlName(name)
                        prefix = constants.prefixes[ns]
                        attr_string = "%s %s" % (prefix, name)
                    else:
                        attr_string = infosetFilter.fromXmlName(name)
                    attributes.append((attr_string, value))

                for name, value in sorted(attributes):
                    rv.append('|%s%s="%s"' % (' ' * (indent + 2), name, value))

            if element.text:
                rv.append("|%s\"%s\"" % (' ' * (indent + 2), element.text))
            indent += 2
            for child in element:
                serializeElement(child, indent)
            if hasattr(element, "tail") and element.tail:
                rv.append("|%s\"%s\"" % (' ' * (indent - 2), element.tail))
    serializeElement(element, 0)

    return "\n".join(rv)


def tostring(element):
    """Serialize an element and its child nodes to a string"""
    rv = []

    def serializeElement(element):
        if not hasattr(element, "tag"):
            if element.docinfo.internalDTD:
                if element.docinfo.doctype:
                    dtd_str = element.docinfo.doctype
                else:
                    dtd_str = "<!DOCTYPE %s>" % element.docinfo.root_name
                rv.append(dtd_str)
            serializeElement(element.getroot())

        elif element.tag == comment_type:
            rv.append("<!--%s-->" % (element.text,))

        else:
            # This is assumed to be an ordinary element
            if not element.attrib:
                rv.append("<%s>" % (element.tag,))
            else:
                attr = " ".join(["%s=\"%s\"" % (name, value)
                                 for name, value in element.attrib.items()])
                rv.append("<%s %s>" % (element.tag, attr))
            if element.text:
                rv.append(element.text)

            for child in element:
                serializeElement(child)

            rv.append("</%s>" % (element.tag,))

        if hasattr(element, "tail") and element.tail:
            rv.append(element.tail)

    serializeElement(element)

    return "".join(rv)


class TreeBuilder(base.TreeBuilder):
    documentClass = Document
    doctypeClass = DocumentType
    elementClass = None
    commentClass = None
    fragmentClass = Document
    implementation = etree

    def __init__(self, namespaceHTMLElements, fullTree=False):
        builder = etree_builders.getETreeModule(etree, fullTree=fullTree)
        infosetFilter = self.infosetFilter = _ihatexml.InfosetFilter(preventDoubleDashComments=True)
        self.namespaceHTMLElements = namespaceHTMLElements

        class Attributes(MutableMapping):
            def __init__(self, element):
                self._element = element

            def _coerceKey(self, key):
                if isinstance(key, tuple):
                    name = "{%s}%s" % (key[2], infosetFilter.coerceAttribute(key[1]))
                else:
                    name = infosetFilter.coerceAttribute(key)
                return name

            def __getitem__(self, key):
                value = self._element._element.attrib[self._coerceKey(key)]
                if not PY3 and isinstance(value, binary_type):
                    value = value.decode("ascii")
                return value

            def __setitem__(self, key, value):
                self._element._element.attrib[self._coerceKey(key)] = value

            def __delitem__(self, key):
                del self._element._element.attrib[self._coerceKey(key)]

            def __iter__(self):
                return iter(self._element._element.attrib)

            def __len__(self):
                return len(self._element._element.attrib)

            def clear(self):
                return self._element._element.attrib.clear()

        class Element(builder.Element):
            def __init__(self, name, namespace):
                name = infosetFilter.coerceElement(name)
                builder.Element.__init__(self, name, namespace=namespace)
                self._attributes = Attributes(self)

            def _setName(self, name):
                self._name = infosetFilter.coerceElement(name)
                self._element.tag = self._getETreeTag(
                    self._name, self._namespace)

            def _getName(self):
                return infosetFilter.fromXmlName(self._name)

            name = property(_getName, _setName)

            def _getAttributes(self):
                return self._attributes

            def _setAttributes(self, value):
                attributes = self.attributes
                attributes.clear()
                attributes.update(value)

            attributes = property(_getAttributes, _setAttributes)

            def insertText(self, data, insertBefore=None):
                data = infosetFilter.coerceCharacters(data)
                builder.Element.insertText(self, data, insertBefore)

            def cloneNode(self):
                element = type(self)(self.name, self.namespace)
                if self._element.attrib:
                    element._element.attrib.update(self._element.attrib)
                return element

        class Comment(builder.Comment):
            def __init__(self, data):
                data = infosetFilter.coerceComment(data)
                builder.Comment.__init__(self, data)

            def _setData(self, data):
                data = infosetFilter.coerceComment(data)
                self._element.text = data

            def _getData(self):
                return self._element.text

            data = property(_getData, _setData)

        self.elementClass = Element
        self.commentClass = Comment
        # self.fragmentClass = builder.DocumentFragment
        base.TreeBuilder.__init__(self, namespaceHTMLElements)

    def reset(self):
        base.TreeBuilder.reset(self)
        self.insertComment = self.insertCommentInitial
        self.initial_comments = []
        self.doctype = None

    def testSerializer(self, element):
        return testSerializer(element)

    def getDocument(self):
        if fullTree:
            return self.document._elementTree
        else:
            return self.document._elementTree.getroot()

    def getFragment(self):
        fragment = []
        element = self.openElements[0]._element
        if element.text:
            fragment.append(element.text)
        fragment.extend(list(element))
        if element.tail:
            fragment.append(element.tail)
        return fragment

    def insertDoctype(self, token):
        name = token["name"]
        publicId = token["publicId"]
        systemId = token["systemId"]

        if not name:
            warnings.warn("lxml cannot represent empty doctype", DataLossWarning)
            self.doctype = None
        else:
            coercedName = self.infosetFilter.coerceElement(name)
            if coercedName != name:
                warnings.warn("lxml cannot represent non-xml doctype", DataLossWarning)

            doctype = self.doctypeClass(coercedName, publicId, systemId)
            self.doctype = doctype

    def insertCommentInitial(self, data, parent=None):
        assert parent is None or parent is self.document
        assert self.document._elementTree is None
        self.initial_comments.append(data)

    def insertCommentMain(self, data, parent=None):
        if (parent == self.document and
                self.document._elementTree.getroot()[-1].tag == comment_type):
            warnings.warn("lxml cannot represent adjacent comments beyond the root elements", DataLossWarning)
        super(TreeBuilder, self).insertComment(data, parent)

    def insertRoot(self, token):
        # Because of the way libxml2 works, it doesn't seem to be possible to
        # alter information like the doctype after the tree has been parsed.
        # Therefore we need to use the built-in parser to create our initial
        # tree, after which we can add elements like normal
        docStr = ""
        if self.doctype:
            assert self.doctype.name
            docStr += "<!DOCTYPE %s" % self.doctype.name
            if (self.doctype.publicId is not None or
                    self.doctype.systemId is not None):
                docStr += (' PUBLIC "%s" ' %
                           (self.infosetFilter.coercePubid(self.doctype.publicId or "")))
                if self.doctype.systemId:
                    sysid = self.doctype.systemId
                    if sysid.find("'") >= 0 and sysid.find('"') >= 0:
                        warnings.warn("DOCTYPE system cannot contain single and double quotes", DataLossWarning)
                        sysid = sysid.replace("'", 'U00027')
                    if sysid.find("'") >= 0:
                        docStr += '"%s"' % sysid
                    else:
                        docStr += "'%s'" % sysid
                else:
                    docStr += "''"
            docStr += ">"
            if self.doctype.name != token["name"]:
                warnings.warn("lxml cannot represent doctype with a different name to the root element", DataLossWarning)
        docStr += "<THIS_SHOULD_NEVER_APPEAR_PUBLICLY/>"
        root = etree.fromstring(docStr)

        # Append the initial comments:
        for comment_token in self.initial_comments:
            comment = self.commentClass(comment_token["data"])
            root.addprevious(comment._element)

        # Create the root document and add the ElementTree to it
        self.document = self.documentClass()
        self.document._elementTree = root.getroottree()

        # Give the root element the right name
        name = token["name"]
        namespace = token.get("namespace", self.defaultNamespace)
        if namespace is None:
            etree_tag = name
        else:
            etree_tag = "{%s}%s" % (namespace, name)
        root.tag = etree_tag

        # Add the root element to the internal child/open data structures
        root_element = self.elementClass(name, namespace)
        root_element._element = root
        self.document._childNodes.append(root_element)
        self.openElements.append(root_element)

        # Reset to the default insert comment function
        self.insertComment = self.insertCommentMain
site-packages/pip/_vendor/html5lib/_utils.py000064400000011503150564745010015110 0ustar00from __future__ import absolute_import, division, unicode_literals

from types import ModuleType

try:
    from collections.abc import Mapping
except ImportError:
    from collections import Mapping

from pip._vendor.six import text_type, PY3

if PY3:
    import xml.etree.ElementTree as default_etree
else:
    try:
        import xml.etree.cElementTree as default_etree
    except ImportError:
        import xml.etree.ElementTree as default_etree


__all__ = ["default_etree", "MethodDispatcher", "isSurrogatePair",
           "surrogatePairToCodepoint", "moduleFactoryFactory",
           "supports_lone_surrogates"]


# Platforms not supporting lone surrogates (\uD800-\uDFFF) should be
# caught by the below test. In general this would be any platform
# using UTF-16 as its encoding of unicode strings, such as
# Jython. This is because UTF-16 itself is based on the use of such
# surrogates, and there is no mechanism to further escape such
# escapes.
try:
    _x = eval('"\\uD800"')  # pylint:disable=eval-used
    if not isinstance(_x, text_type):
        # We need this with u"" because of http://bugs.jython.org/issue2039
        _x = eval('u"\\uD800"')  # pylint:disable=eval-used
        assert isinstance(_x, text_type)
except Exception:
    supports_lone_surrogates = False
else:
    supports_lone_surrogates = True


class MethodDispatcher(dict):
    """Dict with 2 special properties:

    On initiation, keys that are lists, sets or tuples are converted to
    multiple keys so accessing any one of the items in the original
    list-like object returns the matching value

    md = MethodDispatcher({("foo", "bar"):"baz"})
    md["foo"] == "baz"

    A default value which can be set through the default attribute.
    """

    def __init__(self, items=()):
        _dictEntries = []
        for name, value in items:
            if isinstance(name, (list, tuple, frozenset, set)):
                for item in name:
                    _dictEntries.append((item, value))
            else:
                _dictEntries.append((name, value))
        dict.__init__(self, _dictEntries)
        assert len(self) == len(_dictEntries)
        self.default = None

    def __getitem__(self, key):
        return dict.get(self, key, self.default)

    def __get__(self, instance, owner=None):
        return BoundMethodDispatcher(instance, self)


class BoundMethodDispatcher(Mapping):
    """Wraps a MethodDispatcher, binding its return values to `instance`"""
    def __init__(self, instance, dispatcher):
        self.instance = instance
        self.dispatcher = dispatcher

    def __getitem__(self, key):
        # see https://docs.python.org/3/reference/datamodel.html#object.__get__
        # on a function, __get__ is used to bind a function to an instance as a bound method
        return self.dispatcher[key].__get__(self.instance)

    def get(self, key, default):
        if key in self.dispatcher:
            return self[key]
        else:
            return default

    def __iter__(self):
        return iter(self.dispatcher)

    def __len__(self):
        return len(self.dispatcher)

    def __contains__(self, key):
        return key in self.dispatcher


# Some utility functions to deal with weirdness around UCS2 vs UCS4
# python builds

def isSurrogatePair(data):
    return (len(data) == 2 and
            ord(data[0]) >= 0xD800 and ord(data[0]) <= 0xDBFF and
            ord(data[1]) >= 0xDC00 and ord(data[1]) <= 0xDFFF)


def surrogatePairToCodepoint(data):
    char_val = (0x10000 + (ord(data[0]) - 0xD800) * 0x400 +
                (ord(data[1]) - 0xDC00))
    return char_val

# Module Factory Factory (no, this isn't Java, I know)
# Here to stop this being duplicated all over the place.


def moduleFactoryFactory(factory):
    moduleCache = {}

    def moduleFactory(baseModule, *args, **kwargs):
        if isinstance(ModuleType.__name__, type("")):
            name = "_%s_factory" % baseModule.__name__
        else:
            name = b"_%s_factory" % baseModule.__name__

        kwargs_tuple = tuple(kwargs.items())

        try:
            return moduleCache[name][args][kwargs_tuple]
        except KeyError:
            mod = ModuleType(name)
            objs = factory(baseModule, *args, **kwargs)
            mod.__dict__.update(objs)
            if "name" not in moduleCache:
                moduleCache[name] = {}
            if "args" not in moduleCache[name]:
                moduleCache[name][args] = {}
            if "kwargs" not in moduleCache[name][args]:
                moduleCache[name][args][kwargs_tuple] = {}
            moduleCache[name][args][kwargs_tuple] = mod
            return mod

    return moduleFactory


def memoize(func):
    cache = {}

    def wrapped(*args, **kwargs):
        key = (tuple(args), tuple(kwargs.items()))
        if key not in cache:
            cache[key] = func(*args, **kwargs)
        return cache[key]

    return wrapped
site-packages/pip/_vendor/html5lib/_ihatexml.py000064400000040530150564745010015565 0ustar00from __future__ import absolute_import, division, unicode_literals

import re
import warnings

from .constants import DataLossWarning

baseChar = """
[#x0041-#x005A] | [#x0061-#x007A] | [#x00C0-#x00D6] | [#x00D8-#x00F6] |
[#x00F8-#x00FF] | [#x0100-#x0131] | [#x0134-#x013E] | [#x0141-#x0148] |
[#x014A-#x017E] | [#x0180-#x01C3] | [#x01CD-#x01F0] | [#x01F4-#x01F5] |
[#x01FA-#x0217] | [#x0250-#x02A8] | [#x02BB-#x02C1] | #x0386 |
[#x0388-#x038A] | #x038C | [#x038E-#x03A1] | [#x03A3-#x03CE] |
[#x03D0-#x03D6] | #x03DA | #x03DC | #x03DE | #x03E0 | [#x03E2-#x03F3] |
[#x0401-#x040C] | [#x040E-#x044F] | [#x0451-#x045C] | [#x045E-#x0481] |
[#x0490-#x04C4] | [#x04C7-#x04C8] | [#x04CB-#x04CC] | [#x04D0-#x04EB] |
[#x04EE-#x04F5] | [#x04F8-#x04F9] | [#x0531-#x0556] | #x0559 |
[#x0561-#x0586] | [#x05D0-#x05EA] | [#x05F0-#x05F2] | [#x0621-#x063A] |
[#x0641-#x064A] | [#x0671-#x06B7] | [#x06BA-#x06BE] | [#x06C0-#x06CE] |
[#x06D0-#x06D3] | #x06D5 | [#x06E5-#x06E6] | [#x0905-#x0939] | #x093D |
[#x0958-#x0961] | [#x0985-#x098C] | [#x098F-#x0990] | [#x0993-#x09A8] |
[#x09AA-#x09B0] | #x09B2 | [#x09B6-#x09B9] | [#x09DC-#x09DD] |
[#x09DF-#x09E1] | [#x09F0-#x09F1] | [#x0A05-#x0A0A] | [#x0A0F-#x0A10] |
[#x0A13-#x0A28] | [#x0A2A-#x0A30] | [#x0A32-#x0A33] | [#x0A35-#x0A36] |
[#x0A38-#x0A39] | [#x0A59-#x0A5C] | #x0A5E | [#x0A72-#x0A74] |
[#x0A85-#x0A8B] | #x0A8D | [#x0A8F-#x0A91] | [#x0A93-#x0AA8] |
[#x0AAA-#x0AB0] | [#x0AB2-#x0AB3] | [#x0AB5-#x0AB9] | #x0ABD | #x0AE0 |
[#x0B05-#x0B0C] | [#x0B0F-#x0B10] | [#x0B13-#x0B28] | [#x0B2A-#x0B30] |
[#x0B32-#x0B33] | [#x0B36-#x0B39] | #x0B3D | [#x0B5C-#x0B5D] |
[#x0B5F-#x0B61] | [#x0B85-#x0B8A] | [#x0B8E-#x0B90] | [#x0B92-#x0B95] |
[#x0B99-#x0B9A] | #x0B9C | [#x0B9E-#x0B9F] | [#x0BA3-#x0BA4] |
[#x0BA8-#x0BAA] | [#x0BAE-#x0BB5] | [#x0BB7-#x0BB9] | [#x0C05-#x0C0C] |
[#x0C0E-#x0C10] | [#x0C12-#x0C28] | [#x0C2A-#x0C33] | [#x0C35-#x0C39] |
[#x0C60-#x0C61] | [#x0C85-#x0C8C] | [#x0C8E-#x0C90] | [#x0C92-#x0CA8] |
[#x0CAA-#x0CB3] | [#x0CB5-#x0CB9] | #x0CDE | [#x0CE0-#x0CE1] |
[#x0D05-#x0D0C] | [#x0D0E-#x0D10] | [#x0D12-#x0D28] | [#x0D2A-#x0D39] |
[#x0D60-#x0D61] | [#x0E01-#x0E2E] | #x0E30 | [#x0E32-#x0E33] |
[#x0E40-#x0E45] | [#x0E81-#x0E82] | #x0E84 | [#x0E87-#x0E88] | #x0E8A |
#x0E8D | [#x0E94-#x0E97] | [#x0E99-#x0E9F] | [#x0EA1-#x0EA3] | #x0EA5 |
#x0EA7 | [#x0EAA-#x0EAB] | [#x0EAD-#x0EAE] | #x0EB0 | [#x0EB2-#x0EB3] |
#x0EBD | [#x0EC0-#x0EC4] | [#x0F40-#x0F47] | [#x0F49-#x0F69] |
[#x10A0-#x10C5] | [#x10D0-#x10F6] | #x1100 | [#x1102-#x1103] |
[#x1105-#x1107] | #x1109 | [#x110B-#x110C] | [#x110E-#x1112] | #x113C |
#x113E | #x1140 | #x114C | #x114E | #x1150 | [#x1154-#x1155] | #x1159 |
[#x115F-#x1161] | #x1163 | #x1165 | #x1167 | #x1169 | [#x116D-#x116E] |
[#x1172-#x1173] | #x1175 | #x119E | #x11A8 | #x11AB | [#x11AE-#x11AF] |
[#x11B7-#x11B8] | #x11BA | [#x11BC-#x11C2] | #x11EB | #x11F0 | #x11F9 |
[#x1E00-#x1E9B] | [#x1EA0-#x1EF9] | [#x1F00-#x1F15] | [#x1F18-#x1F1D] |
[#x1F20-#x1F45] | [#x1F48-#x1F4D] | [#x1F50-#x1F57] | #x1F59 | #x1F5B |
#x1F5D | [#x1F5F-#x1F7D] | [#x1F80-#x1FB4] | [#x1FB6-#x1FBC] | #x1FBE |
[#x1FC2-#x1FC4] | [#x1FC6-#x1FCC] | [#x1FD0-#x1FD3] | [#x1FD6-#x1FDB] |
[#x1FE0-#x1FEC] | [#x1FF2-#x1FF4] | [#x1FF6-#x1FFC] | #x2126 |
[#x212A-#x212B] | #x212E | [#x2180-#x2182] | [#x3041-#x3094] |
[#x30A1-#x30FA] | [#x3105-#x312C] | [#xAC00-#xD7A3]"""

ideographic = """[#x4E00-#x9FA5] | #x3007 | [#x3021-#x3029]"""

combiningCharacter = """
[#x0300-#x0345] | [#x0360-#x0361] | [#x0483-#x0486] | [#x0591-#x05A1] |
[#x05A3-#x05B9] | [#x05BB-#x05BD] | #x05BF | [#x05C1-#x05C2] | #x05C4 |
[#x064B-#x0652] | #x0670 | [#x06D6-#x06DC] | [#x06DD-#x06DF] |
[#x06E0-#x06E4] | [#x06E7-#x06E8] | [#x06EA-#x06ED] | [#x0901-#x0903] |
#x093C | [#x093E-#x094C] | #x094D | [#x0951-#x0954] | [#x0962-#x0963] |
[#x0981-#x0983] | #x09BC | #x09BE | #x09BF | [#x09C0-#x09C4] |
[#x09C7-#x09C8] | [#x09CB-#x09CD] | #x09D7 | [#x09E2-#x09E3] | #x0A02 |
#x0A3C | #x0A3E | #x0A3F | [#x0A40-#x0A42] | [#x0A47-#x0A48] |
[#x0A4B-#x0A4D] | [#x0A70-#x0A71] | [#x0A81-#x0A83] | #x0ABC |
[#x0ABE-#x0AC5] | [#x0AC7-#x0AC9] | [#x0ACB-#x0ACD] | [#x0B01-#x0B03] |
#x0B3C | [#x0B3E-#x0B43] | [#x0B47-#x0B48] | [#x0B4B-#x0B4D] |
[#x0B56-#x0B57] | [#x0B82-#x0B83] | [#x0BBE-#x0BC2] | [#x0BC6-#x0BC8] |
[#x0BCA-#x0BCD] | #x0BD7 | [#x0C01-#x0C03] | [#x0C3E-#x0C44] |
[#x0C46-#x0C48] | [#x0C4A-#x0C4D] | [#x0C55-#x0C56] | [#x0C82-#x0C83] |
[#x0CBE-#x0CC4] | [#x0CC6-#x0CC8] | [#x0CCA-#x0CCD] | [#x0CD5-#x0CD6] |
[#x0D02-#x0D03] | [#x0D3E-#x0D43] | [#x0D46-#x0D48] | [#x0D4A-#x0D4D] |
#x0D57 | #x0E31 | [#x0E34-#x0E3A] | [#x0E47-#x0E4E] | #x0EB1 |
[#x0EB4-#x0EB9] | [#x0EBB-#x0EBC] | [#x0EC8-#x0ECD] | [#x0F18-#x0F19] |
#x0F35 | #x0F37 | #x0F39 | #x0F3E | #x0F3F | [#x0F71-#x0F84] |
[#x0F86-#x0F8B] | [#x0F90-#x0F95] | #x0F97 | [#x0F99-#x0FAD] |
[#x0FB1-#x0FB7] | #x0FB9 | [#x20D0-#x20DC] | #x20E1 | [#x302A-#x302F] |
#x3099 | #x309A"""

digit = """
[#x0030-#x0039] | [#x0660-#x0669] | [#x06F0-#x06F9] | [#x0966-#x096F] |
[#x09E6-#x09EF] | [#x0A66-#x0A6F] | [#x0AE6-#x0AEF] | [#x0B66-#x0B6F] |
[#x0BE7-#x0BEF] | [#x0C66-#x0C6F] | [#x0CE6-#x0CEF] | [#x0D66-#x0D6F] |
[#x0E50-#x0E59] | [#x0ED0-#x0ED9] | [#x0F20-#x0F29]"""

extender = """
#x00B7 | #x02D0 | #x02D1 | #x0387 | #x0640 | #x0E46 | #x0EC6 | #x3005 |
#[#x3031-#x3035] | [#x309D-#x309E] | [#x30FC-#x30FE]"""

letter = " | ".join([baseChar, ideographic])

# Without the
name = " | ".join([letter, digit, ".", "-", "_", combiningCharacter,
                   extender])
nameFirst = " | ".join([letter, "_"])

reChar = re.compile(r"#x([\d|A-F]{4,4})")
reCharRange = re.compile(r"\[#x([\d|A-F]{4,4})-#x([\d|A-F]{4,4})\]")


def charStringToList(chars):
    charRanges = [item.strip() for item in chars.split(" | ")]
    rv = []
    for item in charRanges:
        foundMatch = False
        for regexp in (reChar, reCharRange):
            match = regexp.match(item)
            if match is not None:
                rv.append([hexToInt(item) for item in match.groups()])
                if len(rv[-1]) == 1:
                    rv[-1] = rv[-1] * 2
                foundMatch = True
                break
        if not foundMatch:
            assert len(item) == 1

            rv.append([ord(item)] * 2)
    rv = normaliseCharList(rv)
    return rv


def normaliseCharList(charList):
    charList = sorted(charList)
    for item in charList:
        assert item[1] >= item[0]
    rv = []
    i = 0
    while i < len(charList):
        j = 1
        rv.append(charList[i])
        while i + j < len(charList) and charList[i + j][0] <= rv[-1][1] + 1:
            rv[-1][1] = charList[i + j][1]
            j += 1
        i += j
    return rv


# We don't really support characters above the BMP :(
max_unicode = int("FFFF", 16)


def missingRanges(charList):
    rv = []
    if charList[0] != 0:
        rv.append([0, charList[0][0] - 1])
    for i, item in enumerate(charList[:-1]):
        rv.append([item[1] + 1, charList[i + 1][0] - 1])
    if charList[-1][1] != max_unicode:
        rv.append([charList[-1][1] + 1, max_unicode])
    return rv


def listToRegexpStr(charList):
    rv = []
    for item in charList:
        if item[0] == item[1]:
            rv.append(escapeRegexp(chr(item[0])))
        else:
            rv.append(escapeRegexp(chr(item[0])) + "-" +
                      escapeRegexp(chr(item[1])))
    return "[%s]" % "".join(rv)


def hexToInt(hex_str):
    return int(hex_str, 16)


def escapeRegexp(string):
    specialCharacters = (".", "^", "$", "*", "+", "?", "{", "}",
                         "[", "]", "|", "(", ")", "-")
    for char in specialCharacters:
        string = string.replace(char, "\\" + char)

    return string

# output from the above
nonXmlNameBMPRegexp = re.compile('[\x00-,/:-@\\[-\\^`\\{-\xb6\xb8-\xbf\xd7\xf7\u0132-\u0133\u013f-\u0140\u0149\u017f\u01c4-\u01cc\u01f1-\u01f3\u01f6-\u01f9\u0218-\u024f\u02a9-\u02ba\u02c2-\u02cf\u02d2-\u02ff\u0346-\u035f\u0362-\u0385\u038b\u038d\u03a2\u03cf\u03d7-\u03d9\u03db\u03dd\u03df\u03e1\u03f4-\u0400\u040d\u0450\u045d\u0482\u0487-\u048f\u04c5-\u04c6\u04c9-\u04ca\u04cd-\u04cf\u04ec-\u04ed\u04f6-\u04f7\u04fa-\u0530\u0557-\u0558\u055a-\u0560\u0587-\u0590\u05a2\u05ba\u05be\u05c0\u05c3\u05c5-\u05cf\u05eb-\u05ef\u05f3-\u0620\u063b-\u063f\u0653-\u065f\u066a-\u066f\u06b8-\u06b9\u06bf\u06cf\u06d4\u06e9\u06ee-\u06ef\u06fa-\u0900\u0904\u093a-\u093b\u094e-\u0950\u0955-\u0957\u0964-\u0965\u0970-\u0980\u0984\u098d-\u098e\u0991-\u0992\u09a9\u09b1\u09b3-\u09b5\u09ba-\u09bb\u09bd\u09c5-\u09c6\u09c9-\u09ca\u09ce-\u09d6\u09d8-\u09db\u09de\u09e4-\u09e5\u09f2-\u0a01\u0a03-\u0a04\u0a0b-\u0a0e\u0a11-\u0a12\u0a29\u0a31\u0a34\u0a37\u0a3a-\u0a3b\u0a3d\u0a43-\u0a46\u0a49-\u0a4a\u0a4e-\u0a58\u0a5d\u0a5f-\u0a65\u0a75-\u0a80\u0a84\u0a8c\u0a8e\u0a92\u0aa9\u0ab1\u0ab4\u0aba-\u0abb\u0ac6\u0aca\u0ace-\u0adf\u0ae1-\u0ae5\u0af0-\u0b00\u0b04\u0b0d-\u0b0e\u0b11-\u0b12\u0b29\u0b31\u0b34-\u0b35\u0b3a-\u0b3b\u0b44-\u0b46\u0b49-\u0b4a\u0b4e-\u0b55\u0b58-\u0b5b\u0b5e\u0b62-\u0b65\u0b70-\u0b81\u0b84\u0b8b-\u0b8d\u0b91\u0b96-\u0b98\u0b9b\u0b9d\u0ba0-\u0ba2\u0ba5-\u0ba7\u0bab-\u0bad\u0bb6\u0bba-\u0bbd\u0bc3-\u0bc5\u0bc9\u0bce-\u0bd6\u0bd8-\u0be6\u0bf0-\u0c00\u0c04\u0c0d\u0c11\u0c29\u0c34\u0c3a-\u0c3d\u0c45\u0c49\u0c4e-\u0c54\u0c57-\u0c5f\u0c62-\u0c65\u0c70-\u0c81\u0c84\u0c8d\u0c91\u0ca9\u0cb4\u0cba-\u0cbd\u0cc5\u0cc9\u0cce-\u0cd4\u0cd7-\u0cdd\u0cdf\u0ce2-\u0ce5\u0cf0-\u0d01\u0d04\u0d0d\u0d11\u0d29\u0d3a-\u0d3d\u0d44-\u0d45\u0d49\u0d4e-\u0d56\u0d58-\u0d5f\u0d62-\u0d65\u0d70-\u0e00\u0e2f\u0e3b-\u0e3f\u0e4f\u0e5a-\u0e80\u0e83\u0e85-\u0e86\u0e89\u0e8b-\u0e8c\u0e8e-\u0e93\u0e98\u0ea0\u0ea4\u0ea6\u0ea8-\u0ea9\u0eac\u0eaf\u0eba\u0ebe-\u0ebf\u0ec5\u0ec7\u0ece-\u0ecf\u0eda-\u0f17\u0f1a-\u0f1f\u0f2a-\u0f34\u0f36\u0f38\u0f3a-\u0f3d\u0f48\u0f6a-\u0f70\u0f85\u0f8c-\u0f8f\u0f96\u0f98\u0fae-\u0fb0\u0fb8\u0fba-\u109f\u10c6-\u10cf\u10f7-\u10ff\u1101\u1104\u1108\u110a\u110d\u1113-\u113b\u113d\u113f\u1141-\u114b\u114d\u114f\u1151-\u1153\u1156-\u1158\u115a-\u115e\u1162\u1164\u1166\u1168\u116a-\u116c\u116f-\u1171\u1174\u1176-\u119d\u119f-\u11a7\u11a9-\u11aa\u11ac-\u11ad\u11b0-\u11b6\u11b9\u11bb\u11c3-\u11ea\u11ec-\u11ef\u11f1-\u11f8\u11fa-\u1dff\u1e9c-\u1e9f\u1efa-\u1eff\u1f16-\u1f17\u1f1e-\u1f1f\u1f46-\u1f47\u1f4e-\u1f4f\u1f58\u1f5a\u1f5c\u1f5e\u1f7e-\u1f7f\u1fb5\u1fbd\u1fbf-\u1fc1\u1fc5\u1fcd-\u1fcf\u1fd4-\u1fd5\u1fdc-\u1fdf\u1fed-\u1ff1\u1ff5\u1ffd-\u20cf\u20dd-\u20e0\u20e2-\u2125\u2127-\u2129\u212c-\u212d\u212f-\u217f\u2183-\u3004\u3006\u3008-\u3020\u3030\u3036-\u3040\u3095-\u3098\u309b-\u309c\u309f-\u30a0\u30fb\u30ff-\u3104\u312d-\u4dff\u9fa6-\uabff\ud7a4-\uffff]')  # noqa

nonXmlNameFirstBMPRegexp = re.compile('[\x00-@\\[-\\^`\\{-\xbf\xd7\xf7\u0132-\u0133\u013f-\u0140\u0149\u017f\u01c4-\u01cc\u01f1-\u01f3\u01f6-\u01f9\u0218-\u024f\u02a9-\u02ba\u02c2-\u0385\u0387\u038b\u038d\u03a2\u03cf\u03d7-\u03d9\u03db\u03dd\u03df\u03e1\u03f4-\u0400\u040d\u0450\u045d\u0482-\u048f\u04c5-\u04c6\u04c9-\u04ca\u04cd-\u04cf\u04ec-\u04ed\u04f6-\u04f7\u04fa-\u0530\u0557-\u0558\u055a-\u0560\u0587-\u05cf\u05eb-\u05ef\u05f3-\u0620\u063b-\u0640\u064b-\u0670\u06b8-\u06b9\u06bf\u06cf\u06d4\u06d6-\u06e4\u06e7-\u0904\u093a-\u093c\u093e-\u0957\u0962-\u0984\u098d-\u098e\u0991-\u0992\u09a9\u09b1\u09b3-\u09b5\u09ba-\u09db\u09de\u09e2-\u09ef\u09f2-\u0a04\u0a0b-\u0a0e\u0a11-\u0a12\u0a29\u0a31\u0a34\u0a37\u0a3a-\u0a58\u0a5d\u0a5f-\u0a71\u0a75-\u0a84\u0a8c\u0a8e\u0a92\u0aa9\u0ab1\u0ab4\u0aba-\u0abc\u0abe-\u0adf\u0ae1-\u0b04\u0b0d-\u0b0e\u0b11-\u0b12\u0b29\u0b31\u0b34-\u0b35\u0b3a-\u0b3c\u0b3e-\u0b5b\u0b5e\u0b62-\u0b84\u0b8b-\u0b8d\u0b91\u0b96-\u0b98\u0b9b\u0b9d\u0ba0-\u0ba2\u0ba5-\u0ba7\u0bab-\u0bad\u0bb6\u0bba-\u0c04\u0c0d\u0c11\u0c29\u0c34\u0c3a-\u0c5f\u0c62-\u0c84\u0c8d\u0c91\u0ca9\u0cb4\u0cba-\u0cdd\u0cdf\u0ce2-\u0d04\u0d0d\u0d11\u0d29\u0d3a-\u0d5f\u0d62-\u0e00\u0e2f\u0e31\u0e34-\u0e3f\u0e46-\u0e80\u0e83\u0e85-\u0e86\u0e89\u0e8b-\u0e8c\u0e8e-\u0e93\u0e98\u0ea0\u0ea4\u0ea6\u0ea8-\u0ea9\u0eac\u0eaf\u0eb1\u0eb4-\u0ebc\u0ebe-\u0ebf\u0ec5-\u0f3f\u0f48\u0f6a-\u109f\u10c6-\u10cf\u10f7-\u10ff\u1101\u1104\u1108\u110a\u110d\u1113-\u113b\u113d\u113f\u1141-\u114b\u114d\u114f\u1151-\u1153\u1156-\u1158\u115a-\u115e\u1162\u1164\u1166\u1168\u116a-\u116c\u116f-\u1171\u1174\u1176-\u119d\u119f-\u11a7\u11a9-\u11aa\u11ac-\u11ad\u11b0-\u11b6\u11b9\u11bb\u11c3-\u11ea\u11ec-\u11ef\u11f1-\u11f8\u11fa-\u1dff\u1e9c-\u1e9f\u1efa-\u1eff\u1f16-\u1f17\u1f1e-\u1f1f\u1f46-\u1f47\u1f4e-\u1f4f\u1f58\u1f5a\u1f5c\u1f5e\u1f7e-\u1f7f\u1fb5\u1fbd\u1fbf-\u1fc1\u1fc5\u1fcd-\u1fcf\u1fd4-\u1fd5\u1fdc-\u1fdf\u1fed-\u1ff1\u1ff5\u1ffd-\u2125\u2127-\u2129\u212c-\u212d\u212f-\u217f\u2183-\u3006\u3008-\u3020\u302a-\u3040\u3095-\u30a0\u30fb-\u3104\u312d-\u4dff\u9fa6-\uabff\ud7a4-\uffff]')  # noqa

# Simpler things
nonPubidCharRegexp = re.compile("[^\x20\x0D\x0Aa-zA-Z0-9\\-'()+,./:=?;!*#@$_%]")


class InfosetFilter(object):
    replacementRegexp = re.compile(r"U[\dA-F]{5,5}")

    def __init__(self,
                 dropXmlnsLocalName=False,
                 dropXmlnsAttrNs=False,
                 preventDoubleDashComments=False,
                 preventDashAtCommentEnd=False,
                 replaceFormFeedCharacters=True,
                 preventSingleQuotePubid=False):

        self.dropXmlnsLocalName = dropXmlnsLocalName
        self.dropXmlnsAttrNs = dropXmlnsAttrNs

        self.preventDoubleDashComments = preventDoubleDashComments
        self.preventDashAtCommentEnd = preventDashAtCommentEnd

        self.replaceFormFeedCharacters = replaceFormFeedCharacters

        self.preventSingleQuotePubid = preventSingleQuotePubid

        self.replaceCache = {}

    def coerceAttribute(self, name, namespace=None):
        if self.dropXmlnsLocalName and name.startswith("xmlns:"):
            warnings.warn("Attributes cannot begin with xmlns", DataLossWarning)
            return None
        elif (self.dropXmlnsAttrNs and
              namespace == "http://www.w3.org/2000/xmlns/"):
            warnings.warn("Attributes cannot be in the xml namespace", DataLossWarning)
            return None
        else:
            return self.toXmlName(name)

    def coerceElement(self, name):
        return self.toXmlName(name)

    def coerceComment(self, data):
        if self.preventDoubleDashComments:
            while "--" in data:
                warnings.warn("Comments cannot contain adjacent dashes", DataLossWarning)
                data = data.replace("--", "- -")
            if data.endswith("-"):
                warnings.warn("Comments cannot end in a dash", DataLossWarning)
                data += " "
        return data

    def coerceCharacters(self, data):
        if self.replaceFormFeedCharacters:
            for _ in range(data.count("\x0C")):
                warnings.warn("Text cannot contain U+000C", DataLossWarning)
            data = data.replace("\x0C", " ")
        # Other non-xml characters
        return data

    def coercePubid(self, data):
        dataOutput = data
        for char in nonPubidCharRegexp.findall(data):
            warnings.warn("Coercing non-XML pubid", DataLossWarning)
            replacement = self.getReplacementCharacter(char)
            dataOutput = dataOutput.replace(char, replacement)
        if self.preventSingleQuotePubid and dataOutput.find("'") >= 0:
            warnings.warn("Pubid cannot contain single quote", DataLossWarning)
            dataOutput = dataOutput.replace("'", self.getReplacementCharacter("'"))
        return dataOutput

    def toXmlName(self, name):
        nameFirst = name[0]
        nameRest = name[1:]
        m = nonXmlNameFirstBMPRegexp.match(nameFirst)
        if m:
            warnings.warn("Coercing non-XML name: %s" % name, DataLossWarning)
            nameFirstOutput = self.getReplacementCharacter(nameFirst)
        else:
            nameFirstOutput = nameFirst

        nameRestOutput = nameRest
        replaceChars = set(nonXmlNameBMPRegexp.findall(nameRest))
        for char in replaceChars:
            warnings.warn("Coercing non-XML name: %s" % name, DataLossWarning)
            replacement = self.getReplacementCharacter(char)
            nameRestOutput = nameRestOutput.replace(char, replacement)
        return nameFirstOutput + nameRestOutput

    def getReplacementCharacter(self, char):
        if char in self.replaceCache:
            replacement = self.replaceCache[char]
        else:
            replacement = self.escapeChar(char)
        return replacement

    def fromXmlName(self, name):
        for item in set(self.replacementRegexp.findall(name)):
            name = name.replace(item, self.unescapeChar(item))
        return name

    def escapeChar(self, char):
        replacement = "U%05X" % ord(char)
        self.replaceCache[char] = replacement
        return replacement

    def unescapeChar(self, charcode):
        return chr(int(charcode[1:], 16))
site-packages/pip/_vendor/html5lib/html5parser.py000064400000344702150564745010016071 0ustar00from __future__ import absolute_import, division, unicode_literals
from pip._vendor.six import with_metaclass, viewkeys

import types

from . import _inputstream
from . import _tokenizer

from . import treebuilders
from .treebuilders.base import Marker

from . import _utils
from .constants import (
    spaceCharacters, asciiUpper2Lower,
    specialElements, headingElements, cdataElements, rcdataElements,
    tokenTypes, tagTokenTypes,
    namespaces,
    htmlIntegrationPointElements, mathmlTextIntegrationPointElements,
    adjustForeignAttributes as adjustForeignAttributesMap,
    adjustMathMLAttributes, adjustSVGAttributes,
    E,
    _ReparseException
)


def parse(doc, treebuilder="etree", namespaceHTMLElements=True, **kwargs):
    """Parse an HTML document as a string or file-like object into a tree

    :arg doc: the document to parse as a string or file-like object

    :arg treebuilder: the treebuilder to use when parsing

    :arg namespaceHTMLElements: whether or not to namespace HTML elements

    :returns: parsed tree

    Example:

    >>> from html5lib.html5parser import parse
    >>> parse('<html><body><p>This is a doc</p></body></html>')
    <Element u'{http://www.w3.org/1999/xhtml}html' at 0x7feac4909db0>

    """
    tb = treebuilders.getTreeBuilder(treebuilder)
    p = HTMLParser(tb, namespaceHTMLElements=namespaceHTMLElements)
    return p.parse(doc, **kwargs)


def parseFragment(doc, container="div", treebuilder="etree", namespaceHTMLElements=True, **kwargs):
    """Parse an HTML fragment as a string or file-like object into a tree

    :arg doc: the fragment to parse as a string or file-like object

    :arg container: the container context to parse the fragment in

    :arg treebuilder: the treebuilder to use when parsing

    :arg namespaceHTMLElements: whether or not to namespace HTML elements

    :returns: parsed tree

    Example:

    >>> from html5lib.html5libparser import parseFragment
    >>> parseFragment('<b>this is a fragment</b>')
    <Element u'DOCUMENT_FRAGMENT' at 0x7feac484b090>

    """
    tb = treebuilders.getTreeBuilder(treebuilder)
    p = HTMLParser(tb, namespaceHTMLElements=namespaceHTMLElements)
    return p.parseFragment(doc, container=container, **kwargs)


def method_decorator_metaclass(function):
    class Decorated(type):
        def __new__(meta, classname, bases, classDict):
            for attributeName, attribute in classDict.items():
                if isinstance(attribute, types.FunctionType):
                    attribute = function(attribute)

                classDict[attributeName] = attribute
            return type.__new__(meta, classname, bases, classDict)
    return Decorated


class HTMLParser(object):
    """HTML parser

    Generates a tree structure from a stream of (possibly malformed) HTML.

    """

    def __init__(self, tree=None, strict=False, namespaceHTMLElements=True, debug=False):
        """
        :arg tree: a treebuilder class controlling the type of tree that will be
            returned. Built in treebuilders can be accessed through
            html5lib.treebuilders.getTreeBuilder(treeType)

        :arg strict: raise an exception when a parse error is encountered

        :arg namespaceHTMLElements: whether or not to namespace HTML elements

        :arg debug: whether or not to enable debug mode which logs things

        Example:

        >>> from html5lib.html5parser import HTMLParser
        >>> parser = HTMLParser()                     # generates parser with etree builder
        >>> parser = HTMLParser('lxml', strict=True)  # generates parser with lxml builder which is strict

        """

        # Raise an exception on the first error encountered
        self.strict = strict

        if tree is None:
            tree = treebuilders.getTreeBuilder("etree")
        self.tree = tree(namespaceHTMLElements)
        self.errors = []

        self.phases = {name: cls(self, self.tree) for name, cls in
                       getPhases(debug).items()}

    def _parse(self, stream, innerHTML=False, container="div", scripting=False, **kwargs):

        self.innerHTMLMode = innerHTML
        self.container = container
        self.scripting = scripting
        self.tokenizer = _tokenizer.HTMLTokenizer(stream, parser=self, **kwargs)
        self.reset()

        try:
            self.mainLoop()
        except _ReparseException:
            self.reset()
            self.mainLoop()

    def reset(self):
        self.tree.reset()
        self.firstStartTag = False
        self.errors = []
        self.log = []  # only used with debug mode
        # "quirks" / "limited quirks" / "no quirks"
        self.compatMode = "no quirks"

        if self.innerHTMLMode:
            self.innerHTML = self.container.lower()

            if self.innerHTML in cdataElements:
                self.tokenizer.state = self.tokenizer.rcdataState
            elif self.innerHTML in rcdataElements:
                self.tokenizer.state = self.tokenizer.rawtextState
            elif self.innerHTML == 'plaintext':
                self.tokenizer.state = self.tokenizer.plaintextState
            else:
                # state already is data state
                # self.tokenizer.state = self.tokenizer.dataState
                pass
            self.phase = self.phases["beforeHtml"]
            self.phase.insertHtmlElement()
            self.resetInsertionMode()
        else:
            self.innerHTML = False  # pylint:disable=redefined-variable-type
            self.phase = self.phases["initial"]

        self.lastPhase = None

        self.beforeRCDataPhase = None

        self.framesetOK = True

    @property
    def documentEncoding(self):
        """Name of the character encoding that was used to decode the input stream, or
        :obj:`None` if that is not determined yet

        """
        if not hasattr(self, 'tokenizer'):
            return None
        return self.tokenizer.stream.charEncoding[0].name

    def isHTMLIntegrationPoint(self, element):
        if (element.name == "annotation-xml" and
                element.namespace == namespaces["mathml"]):
            return ("encoding" in element.attributes and
                    element.attributes["encoding"].translate(
                        asciiUpper2Lower) in
                    ("text/html", "application/xhtml+xml"))
        else:
            return (element.namespace, element.name) in htmlIntegrationPointElements

    def isMathMLTextIntegrationPoint(self, element):
        return (element.namespace, element.name) in mathmlTextIntegrationPointElements

    def mainLoop(self):
        CharactersToken = tokenTypes["Characters"]
        SpaceCharactersToken = tokenTypes["SpaceCharacters"]
        StartTagToken = tokenTypes["StartTag"]
        EndTagToken = tokenTypes["EndTag"]
        CommentToken = tokenTypes["Comment"]
        DoctypeToken = tokenTypes["Doctype"]
        ParseErrorToken = tokenTypes["ParseError"]

        for token in self.tokenizer:
            prev_token = None
            new_token = token
            while new_token is not None:
                prev_token = new_token
                currentNode = self.tree.openElements[-1] if self.tree.openElements else None
                currentNodeNamespace = currentNode.namespace if currentNode else None
                currentNodeName = currentNode.name if currentNode else None

                type = new_token["type"]

                if type == ParseErrorToken:
                    self.parseError(new_token["data"], new_token.get("datavars", {}))
                    new_token = None
                else:
                    if (len(self.tree.openElements) == 0 or
                        currentNodeNamespace == self.tree.defaultNamespace or
                        (self.isMathMLTextIntegrationPoint(currentNode) and
                         ((type == StartTagToken and
                           token["name"] not in frozenset(["mglyph", "malignmark"])) or
                          type in (CharactersToken, SpaceCharactersToken))) or
                        (currentNodeNamespace == namespaces["mathml"] and
                         currentNodeName == "annotation-xml" and
                         type == StartTagToken and
                         token["name"] == "svg") or
                        (self.isHTMLIntegrationPoint(currentNode) and
                         type in (StartTagToken, CharactersToken, SpaceCharactersToken))):
                        phase = self.phase
                    else:
                        phase = self.phases["inForeignContent"]

                    if type == CharactersToken:
                        new_token = phase.processCharacters(new_token)
                    elif type == SpaceCharactersToken:
                        new_token = phase.processSpaceCharacters(new_token)
                    elif type == StartTagToken:
                        new_token = phase.processStartTag(new_token)
                    elif type == EndTagToken:
                        new_token = phase.processEndTag(new_token)
                    elif type == CommentToken:
                        new_token = phase.processComment(new_token)
                    elif type == DoctypeToken:
                        new_token = phase.processDoctype(new_token)

            if (type == StartTagToken and prev_token["selfClosing"] and
                    not prev_token["selfClosingAcknowledged"]):
                self.parseError("non-void-element-with-trailing-solidus",
                                {"name": prev_token["name"]})

        # When the loop finishes it's EOF
        reprocess = True
        phases = []
        while reprocess:
            phases.append(self.phase)
            reprocess = self.phase.processEOF()
            if reprocess:
                assert self.phase not in phases

    def parse(self, stream, *args, **kwargs):
        """Parse a HTML document into a well-formed tree

        :arg stream: a file-like object or string containing the HTML to be parsed

            The optional encoding parameter must be a string that indicates
            the encoding.  If specified, that encoding will be used,
            regardless of any BOM or later declaration (such as in a meta
            element).

        :arg scripting: treat noscript elements as if JavaScript was turned on

        :returns: parsed tree

        Example:

        >>> from html5lib.html5parser import HTMLParser
        >>> parser = HTMLParser()
        >>> parser.parse('<html><body><p>This is a doc</p></body></html>')
        <Element u'{http://www.w3.org/1999/xhtml}html' at 0x7feac4909db0>

        """
        self._parse(stream, False, None, *args, **kwargs)
        return self.tree.getDocument()

    def parseFragment(self, stream, *args, **kwargs):
        """Parse a HTML fragment into a well-formed tree fragment

        :arg container: name of the element we're setting the innerHTML
            property if set to None, default to 'div'

        :arg stream: a file-like object or string containing the HTML to be parsed

            The optional encoding parameter must be a string that indicates
            the encoding.  If specified, that encoding will be used,
            regardless of any BOM or later declaration (such as in a meta
            element)

        :arg scripting: treat noscript elements as if JavaScript was turned on

        :returns: parsed tree

        Example:

        >>> from html5lib.html5libparser import HTMLParser
        >>> parser = HTMLParser()
        >>> parser.parseFragment('<b>this is a fragment</b>')
        <Element u'DOCUMENT_FRAGMENT' at 0x7feac484b090>

        """
        self._parse(stream, True, *args, **kwargs)
        return self.tree.getFragment()

    def parseError(self, errorcode="XXX-undefined-error", datavars=None):
        # XXX The idea is to make errorcode mandatory.
        if datavars is None:
            datavars = {}
        self.errors.append((self.tokenizer.stream.position(), errorcode, datavars))
        if self.strict:
            raise ParseError(E[errorcode] % datavars)

    def adjustMathMLAttributes(self, token):
        adjust_attributes(token, adjustMathMLAttributes)

    def adjustSVGAttributes(self, token):
        adjust_attributes(token, adjustSVGAttributes)

    def adjustForeignAttributes(self, token):
        adjust_attributes(token, adjustForeignAttributesMap)

    def reparseTokenNormal(self, token):
        # pylint:disable=unused-argument
        self.parser.phase()

    def resetInsertionMode(self):
        # The name of this method is mostly historical. (It's also used in the
        # specification.)
        last = False
        newModes = {
            "select": "inSelect",
            "td": "inCell",
            "th": "inCell",
            "tr": "inRow",
            "tbody": "inTableBody",
            "thead": "inTableBody",
            "tfoot": "inTableBody",
            "caption": "inCaption",
            "colgroup": "inColumnGroup",
            "table": "inTable",
            "head": "inBody",
            "body": "inBody",
            "frameset": "inFrameset",
            "html": "beforeHead"
        }
        for node in self.tree.openElements[::-1]:
            nodeName = node.name
            new_phase = None
            if node == self.tree.openElements[0]:
                assert self.innerHTML
                last = True
                nodeName = self.innerHTML
            # Check for conditions that should only happen in the innerHTML
            # case
            if nodeName in ("select", "colgroup", "head", "html"):
                assert self.innerHTML

            if not last and node.namespace != self.tree.defaultNamespace:
                continue

            if nodeName in newModes:
                new_phase = self.phases[newModes[nodeName]]
                break
            elif last:
                new_phase = self.phases["inBody"]
                break

        self.phase = new_phase

    def parseRCDataRawtext(self, token, contentType):
        # Generic RCDATA/RAWTEXT Parsing algorithm
        assert contentType in ("RAWTEXT", "RCDATA")

        self.tree.insertElement(token)

        if contentType == "RAWTEXT":
            self.tokenizer.state = self.tokenizer.rawtextState
        else:
            self.tokenizer.state = self.tokenizer.rcdataState

        self.originalPhase = self.phase

        self.phase = self.phases["text"]


@_utils.memoize
def getPhases(debug):
    def log(function):
        """Logger that records which phase processes each token"""
        type_names = {value: key for key, value in tokenTypes.items()}

        def wrapped(self, *args, **kwargs):
            if function.__name__.startswith("process") and len(args) > 0:
                token = args[0]
                info = {"type": type_names[token['type']]}
                if token['type'] in tagTokenTypes:
                    info["name"] = token['name']

                self.parser.log.append((self.parser.tokenizer.state.__name__,
                                        self.parser.phase.__class__.__name__,
                                        self.__class__.__name__,
                                        function.__name__,
                                        info))
                return function(self, *args, **kwargs)
            else:
                return function(self, *args, **kwargs)
        return wrapped

    def getMetaclass(use_metaclass, metaclass_func):
        if use_metaclass:
            return method_decorator_metaclass(metaclass_func)
        else:
            return type

    # pylint:disable=unused-argument
    class Phase(with_metaclass(getMetaclass(debug, log))):
        """Base class for helper object that implements each phase of processing
        """
        __slots__ = ("parser", "tree", "__startTagCache", "__endTagCache")

        def __init__(self, parser, tree):
            self.parser = parser
            self.tree = tree
            self.__startTagCache = {}
            self.__endTagCache = {}

        def processEOF(self):
            raise NotImplementedError

        def processComment(self, token):
            # For most phases the following is correct. Where it's not it will be
            # overridden.
            self.tree.insertComment(token, self.tree.openElements[-1])

        def processDoctype(self, token):
            self.parser.parseError("unexpected-doctype")

        def processCharacters(self, token):
            self.tree.insertText(token["data"])

        def processSpaceCharacters(self, token):
            self.tree.insertText(token["data"])

        def processStartTag(self, token):
            # Note the caching is done here rather than BoundMethodDispatcher as doing it there
            # requires a circular reference to the Phase, and this ends up with a significant
            # (CPython 2.7, 3.8) GC cost when parsing many short inputs
            name = token["name"]
            # In Py2, using `in` is quicker in general than try/except KeyError
            # In Py3, `in` is quicker when there are few cache hits (typically short inputs)
            if name in self.__startTagCache:
                func = self.__startTagCache[name]
            else:
                func = self.__startTagCache[name] = self.startTagHandler[name]
                # bound the cache size in case we get loads of unknown tags
                while len(self.__startTagCache) > len(self.startTagHandler) * 1.1:
                    # this makes the eviction policy random on Py < 3.7 and FIFO >= 3.7
                    self.__startTagCache.pop(next(iter(self.__startTagCache)))
            return func(token)

        def startTagHtml(self, token):
            if not self.parser.firstStartTag and token["name"] == "html":
                self.parser.parseError("non-html-root")
            # XXX Need a check here to see if the first start tag token emitted is
            # this token... If it's not, invoke self.parser.parseError().
            for attr, value in token["data"].items():
                if attr not in self.tree.openElements[0].attributes:
                    self.tree.openElements[0].attributes[attr] = value
            self.parser.firstStartTag = False

        def processEndTag(self, token):
            # Note the caching is done here rather than BoundMethodDispatcher as doing it there
            # requires a circular reference to the Phase, and this ends up with a significant
            # (CPython 2.7, 3.8) GC cost when parsing many short inputs
            name = token["name"]
            # In Py2, using `in` is quicker in general than try/except KeyError
            # In Py3, `in` is quicker when there are few cache hits (typically short inputs)
            if name in self.__endTagCache:
                func = self.__endTagCache[name]
            else:
                func = self.__endTagCache[name] = self.endTagHandler[name]
                # bound the cache size in case we get loads of unknown tags
                while len(self.__endTagCache) > len(self.endTagHandler) * 1.1:
                    # this makes the eviction policy random on Py < 3.7 and FIFO >= 3.7
                    self.__endTagCache.pop(next(iter(self.__endTagCache)))
            return func(token)

    class InitialPhase(Phase):
        __slots__ = tuple()

        def processSpaceCharacters(self, token):
            pass

        def processComment(self, token):
            self.tree.insertComment(token, self.tree.document)

        def processDoctype(self, token):
            name = token["name"]
            publicId = token["publicId"]
            systemId = token["systemId"]
            correct = token["correct"]

            if (name != "html" or publicId is not None or
                    systemId is not None and systemId != "about:legacy-compat"):
                self.parser.parseError("unknown-doctype")

            if publicId is None:
                publicId = ""

            self.tree.insertDoctype(token)

            if publicId != "":
                publicId = publicId.translate(asciiUpper2Lower)

            if (not correct or token["name"] != "html" or
                    publicId.startswith(
                        ("+//silmaril//dtd html pro v0r11 19970101//",
                         "-//advasoft ltd//dtd html 3.0 aswedit + extensions//",
                         "-//as//dtd html 3.0 aswedit + extensions//",
                         "-//ietf//dtd html 2.0 level 1//",
                         "-//ietf//dtd html 2.0 level 2//",
                         "-//ietf//dtd html 2.0 strict level 1//",
                         "-//ietf//dtd html 2.0 strict level 2//",
                         "-//ietf//dtd html 2.0 strict//",
                         "-//ietf//dtd html 2.0//",
                         "-//ietf//dtd html 2.1e//",
                         "-//ietf//dtd html 3.0//",
                         "-//ietf//dtd html 3.2 final//",
                         "-//ietf//dtd html 3.2//",
                         "-//ietf//dtd html 3//",
                         "-//ietf//dtd html level 0//",
                         "-//ietf//dtd html level 1//",
                         "-//ietf//dtd html level 2//",
                         "-//ietf//dtd html level 3//",
                         "-//ietf//dtd html strict level 0//",
                         "-//ietf//dtd html strict level 1//",
                         "-//ietf//dtd html strict level 2//",
                         "-//ietf//dtd html strict level 3//",
                         "-//ietf//dtd html strict//",
                         "-//ietf//dtd html//",
                         "-//metrius//dtd metrius presentational//",
                         "-//microsoft//dtd internet explorer 2.0 html strict//",
                         "-//microsoft//dtd internet explorer 2.0 html//",
                         "-//microsoft//dtd internet explorer 2.0 tables//",
                         "-//microsoft//dtd internet explorer 3.0 html strict//",
                         "-//microsoft//dtd internet explorer 3.0 html//",
                         "-//microsoft//dtd internet explorer 3.0 tables//",
                         "-//netscape comm. corp.//dtd html//",
                         "-//netscape comm. corp.//dtd strict html//",
                         "-//o'reilly and associates//dtd html 2.0//",
                         "-//o'reilly and associates//dtd html extended 1.0//",
                         "-//o'reilly and associates//dtd html extended relaxed 1.0//",
                         "-//softquad software//dtd hotmetal pro 6.0::19990601::extensions to html 4.0//",
                         "-//softquad//dtd hotmetal pro 4.0::19971010::extensions to html 4.0//",
                         "-//spyglass//dtd html 2.0 extended//",
                         "-//sq//dtd html 2.0 hotmetal + extensions//",
                         "-//sun microsystems corp.//dtd hotjava html//",
                         "-//sun microsystems corp.//dtd hotjava strict html//",
                         "-//w3c//dtd html 3 1995-03-24//",
                         "-//w3c//dtd html 3.2 draft//",
                         "-//w3c//dtd html 3.2 final//",
                         "-//w3c//dtd html 3.2//",
                         "-//w3c//dtd html 3.2s draft//",
                         "-//w3c//dtd html 4.0 frameset//",
                         "-//w3c//dtd html 4.0 transitional//",
                         "-//w3c//dtd html experimental 19960712//",
                         "-//w3c//dtd html experimental 970421//",
                         "-//w3c//dtd w3 html//",
                         "-//w3o//dtd w3 html 3.0//",
                         "-//webtechs//dtd mozilla html 2.0//",
                         "-//webtechs//dtd mozilla html//")) or
                    publicId in ("-//w3o//dtd w3 html strict 3.0//en//",
                                 "-/w3c/dtd html 4.0 transitional/en",
                                 "html") or
                    publicId.startswith(
                        ("-//w3c//dtd html 4.01 frameset//",
                         "-//w3c//dtd html 4.01 transitional//")) and
                    systemId is None or
                    systemId and systemId.lower() == "http://www.ibm.com/data/dtd/v11/ibmxhtml1-transitional.dtd"):
                self.parser.compatMode = "quirks"
            elif (publicId.startswith(
                    ("-//w3c//dtd xhtml 1.0 frameset//",
                     "-//w3c//dtd xhtml 1.0 transitional//")) or
                  publicId.startswith(
                      ("-//w3c//dtd html 4.01 frameset//",
                       "-//w3c//dtd html 4.01 transitional//")) and
                  systemId is not None):
                self.parser.compatMode = "limited quirks"

            self.parser.phase = self.parser.phases["beforeHtml"]

        def anythingElse(self):
            self.parser.compatMode = "quirks"
            self.parser.phase = self.parser.phases["beforeHtml"]

        def processCharacters(self, token):
            self.parser.parseError("expected-doctype-but-got-chars")
            self.anythingElse()
            return token

        def processStartTag(self, token):
            self.parser.parseError("expected-doctype-but-got-start-tag",
                                   {"name": token["name"]})
            self.anythingElse()
            return token

        def processEndTag(self, token):
            self.parser.parseError("expected-doctype-but-got-end-tag",
                                   {"name": token["name"]})
            self.anythingElse()
            return token

        def processEOF(self):
            self.parser.parseError("expected-doctype-but-got-eof")
            self.anythingElse()
            return True

    class BeforeHtmlPhase(Phase):
        __slots__ = tuple()

        # helper methods
        def insertHtmlElement(self):
            self.tree.insertRoot(impliedTagToken("html", "StartTag"))
            self.parser.phase = self.parser.phases["beforeHead"]

        # other
        def processEOF(self):
            self.insertHtmlElement()
            return True

        def processComment(self, token):
            self.tree.insertComment(token, self.tree.document)

        def processSpaceCharacters(self, token):
            pass

        def processCharacters(self, token):
            self.insertHtmlElement()
            return token

        def processStartTag(self, token):
            if token["name"] == "html":
                self.parser.firstStartTag = True
            self.insertHtmlElement()
            return token

        def processEndTag(self, token):
            if token["name"] not in ("head", "body", "html", "br"):
                self.parser.parseError("unexpected-end-tag-before-html",
                                       {"name": token["name"]})
            else:
                self.insertHtmlElement()
                return token

    class BeforeHeadPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            self.startTagHead(impliedTagToken("head", "StartTag"))
            return True

        def processSpaceCharacters(self, token):
            pass

        def processCharacters(self, token):
            self.startTagHead(impliedTagToken("head", "StartTag"))
            return token

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagHead(self, token):
            self.tree.insertElement(token)
            self.tree.headPointer = self.tree.openElements[-1]
            self.parser.phase = self.parser.phases["inHead"]

        def startTagOther(self, token):
            self.startTagHead(impliedTagToken("head", "StartTag"))
            return token

        def endTagImplyHead(self, token):
            self.startTagHead(impliedTagToken("head", "StartTag"))
            return token

        def endTagOther(self, token):
            self.parser.parseError("end-tag-after-implied-root",
                                   {"name": token["name"]})

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml),
            ("head", startTagHead)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            (("head", "body", "html", "br"), endTagImplyHead)
        ])
        endTagHandler.default = endTagOther

    class InHeadPhase(Phase):
        __slots__ = tuple()

        # the real thing
        def processEOF(self):
            self.anythingElse()
            return True

        def processCharacters(self, token):
            self.anythingElse()
            return token

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagHead(self, token):
            self.parser.parseError("two-heads-are-not-better-than-one")

        def startTagBaseLinkCommand(self, token):
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True

        def startTagMeta(self, token):
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True

            attributes = token["data"]
            if self.parser.tokenizer.stream.charEncoding[1] == "tentative":
                if "charset" in attributes:
                    self.parser.tokenizer.stream.changeEncoding(attributes["charset"])
                elif ("content" in attributes and
                      "http-equiv" in attributes and
                      attributes["http-equiv"].lower() == "content-type"):
                    # Encoding it as UTF-8 here is a hack, as really we should pass
                    # the abstract Unicode string, and just use the
                    # ContentAttrParser on that, but using UTF-8 allows all chars
                    # to be encoded and as a ASCII-superset works.
                    data = _inputstream.EncodingBytes(attributes["content"].encode("utf-8"))
                    parser = _inputstream.ContentAttrParser(data)
                    codec = parser.parse()
                    self.parser.tokenizer.stream.changeEncoding(codec)

        def startTagTitle(self, token):
            self.parser.parseRCDataRawtext(token, "RCDATA")

        def startTagNoFramesStyle(self, token):
            # Need to decide whether to implement the scripting-disabled case
            self.parser.parseRCDataRawtext(token, "RAWTEXT")

        def startTagNoscript(self, token):
            if self.parser.scripting:
                self.parser.parseRCDataRawtext(token, "RAWTEXT")
            else:
                self.tree.insertElement(token)
                self.parser.phase = self.parser.phases["inHeadNoscript"]

        def startTagScript(self, token):
            self.tree.insertElement(token)
            self.parser.tokenizer.state = self.parser.tokenizer.scriptDataState
            self.parser.originalPhase = self.parser.phase
            self.parser.phase = self.parser.phases["text"]

        def startTagOther(self, token):
            self.anythingElse()
            return token

        def endTagHead(self, token):
            node = self.parser.tree.openElements.pop()
            assert node.name == "head", "Expected head got %s" % node.name
            self.parser.phase = self.parser.phases["afterHead"]

        def endTagHtmlBodyBr(self, token):
            self.anythingElse()
            return token

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def anythingElse(self):
            self.endTagHead(impliedTagToken("head"))

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml),
            ("title", startTagTitle),
            (("noframes", "style"), startTagNoFramesStyle),
            ("noscript", startTagNoscript),
            ("script", startTagScript),
            (("base", "basefont", "bgsound", "command", "link"),
             startTagBaseLinkCommand),
            ("meta", startTagMeta),
            ("head", startTagHead)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("head", endTagHead),
            (("br", "html", "body"), endTagHtmlBodyBr)
        ])
        endTagHandler.default = endTagOther

    class InHeadNoscriptPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            self.parser.parseError("eof-in-head-noscript")
            self.anythingElse()
            return True

        def processComment(self, token):
            return self.parser.phases["inHead"].processComment(token)

        def processCharacters(self, token):
            self.parser.parseError("char-in-head-noscript")
            self.anythingElse()
            return token

        def processSpaceCharacters(self, token):
            return self.parser.phases["inHead"].processSpaceCharacters(token)

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagBaseLinkCommand(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagHeadNoscript(self, token):
            self.parser.parseError("unexpected-start-tag", {"name": token["name"]})

        def startTagOther(self, token):
            self.parser.parseError("unexpected-inhead-noscript-tag", {"name": token["name"]})
            self.anythingElse()
            return token

        def endTagNoscript(self, token):
            node = self.parser.tree.openElements.pop()
            assert node.name == "noscript", "Expected noscript got %s" % node.name
            self.parser.phase = self.parser.phases["inHead"]

        def endTagBr(self, token):
            self.parser.parseError("unexpected-inhead-noscript-tag", {"name": token["name"]})
            self.anythingElse()
            return token

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def anythingElse(self):
            # Caller must raise parse error first!
            self.endTagNoscript(impliedTagToken("noscript"))

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml),
            (("basefont", "bgsound", "link", "meta", "noframes", "style"), startTagBaseLinkCommand),
            (("head", "noscript"), startTagHeadNoscript),
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("noscript", endTagNoscript),
            ("br", endTagBr),
        ])
        endTagHandler.default = endTagOther

    class AfterHeadPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            self.anythingElse()
            return True

        def processCharacters(self, token):
            self.anythingElse()
            return token

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagBody(self, token):
            self.parser.framesetOK = False
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inBody"]

        def startTagFrameset(self, token):
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inFrameset"]

        def startTagFromHead(self, token):
            self.parser.parseError("unexpected-start-tag-out-of-my-head",
                                   {"name": token["name"]})
            self.tree.openElements.append(self.tree.headPointer)
            self.parser.phases["inHead"].processStartTag(token)
            for node in self.tree.openElements[::-1]:
                if node.name == "head":
                    self.tree.openElements.remove(node)
                    break

        def startTagHead(self, token):
            self.parser.parseError("unexpected-start-tag", {"name": token["name"]})

        def startTagOther(self, token):
            self.anythingElse()
            return token

        def endTagHtmlBodyBr(self, token):
            self.anythingElse()
            return token

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def anythingElse(self):
            self.tree.insertElement(impliedTagToken("body", "StartTag"))
            self.parser.phase = self.parser.phases["inBody"]
            self.parser.framesetOK = True

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml),
            ("body", startTagBody),
            ("frameset", startTagFrameset),
            (("base", "basefont", "bgsound", "link", "meta", "noframes", "script",
              "style", "title"),
             startTagFromHead),
            ("head", startTagHead)
        ])
        startTagHandler.default = startTagOther
        endTagHandler = _utils.MethodDispatcher([(("body", "html", "br"),
                                                  endTagHtmlBodyBr)])
        endTagHandler.default = endTagOther

    class InBodyPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#parsing-main-inbody
        # the really-really-really-very crazy mode
        __slots__ = ("processSpaceCharacters",)

        def __init__(self, *args, **kwargs):
            super(InBodyPhase, self).__init__(*args, **kwargs)
            # Set this to the default handler
            self.processSpaceCharacters = self.processSpaceCharactersNonPre

        def isMatchingFormattingElement(self, node1, node2):
            return (node1.name == node2.name and
                    node1.namespace == node2.namespace and
                    node1.attributes == node2.attributes)

        # helper
        def addFormattingElement(self, token):
            self.tree.insertElement(token)
            element = self.tree.openElements[-1]

            matchingElements = []
            for node in self.tree.activeFormattingElements[::-1]:
                if node is Marker:
                    break
                elif self.isMatchingFormattingElement(node, element):
                    matchingElements.append(node)

            assert len(matchingElements) <= 3
            if len(matchingElements) == 3:
                self.tree.activeFormattingElements.remove(matchingElements[-1])
            self.tree.activeFormattingElements.append(element)

        # the real deal
        def processEOF(self):
            allowed_elements = frozenset(("dd", "dt", "li", "p", "tbody", "td",
                                          "tfoot", "th", "thead", "tr", "body",
                                          "html"))
            for node in self.tree.openElements[::-1]:
                if node.name not in allowed_elements:
                    self.parser.parseError("expected-closing-tag-but-got-eof")
                    break
            # Stop parsing

        def processSpaceCharactersDropNewline(self, token):
            # Sometimes (start of <pre>, <listing>, and <textarea> blocks) we
            # want to drop leading newlines
            data = token["data"]
            self.processSpaceCharacters = self.processSpaceCharactersNonPre
            if (data.startswith("\n") and
                self.tree.openElements[-1].name in ("pre", "listing", "textarea") and
                    not self.tree.openElements[-1].hasContent()):
                data = data[1:]
            if data:
                self.tree.reconstructActiveFormattingElements()
                self.tree.insertText(data)

        def processCharacters(self, token):
            if token["data"] == "\u0000":
                # The tokenizer should always emit null on its own
                return
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertText(token["data"])
            # This must be bad for performance
            if (self.parser.framesetOK and
                any([char not in spaceCharacters
                     for char in token["data"]])):
                self.parser.framesetOK = False

        def processSpaceCharactersNonPre(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertText(token["data"])

        def startTagProcessInHead(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagBody(self, token):
            self.parser.parseError("unexpected-start-tag", {"name": "body"})
            if (len(self.tree.openElements) == 1 or
                    self.tree.openElements[1].name != "body"):
                assert self.parser.innerHTML
            else:
                self.parser.framesetOK = False
                for attr, value in token["data"].items():
                    if attr not in self.tree.openElements[1].attributes:
                        self.tree.openElements[1].attributes[attr] = value

        def startTagFrameset(self, token):
            self.parser.parseError("unexpected-start-tag", {"name": "frameset"})
            if (len(self.tree.openElements) == 1 or self.tree.openElements[1].name != "body"):
                assert self.parser.innerHTML
            elif not self.parser.framesetOK:
                pass
            else:
                if self.tree.openElements[1].parent:
                    self.tree.openElements[1].parent.removeChild(self.tree.openElements[1])
                while self.tree.openElements[-1].name != "html":
                    self.tree.openElements.pop()
                self.tree.insertElement(token)
                self.parser.phase = self.parser.phases["inFrameset"]

        def startTagCloseP(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            self.tree.insertElement(token)

        def startTagPreListing(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            self.tree.insertElement(token)
            self.parser.framesetOK = False
            self.processSpaceCharacters = self.processSpaceCharactersDropNewline

        def startTagForm(self, token):
            if self.tree.formPointer:
                self.parser.parseError("unexpected-start-tag", {"name": "form"})
            else:
                if self.tree.elementInScope("p", variant="button"):
                    self.endTagP(impliedTagToken("p"))
                self.tree.insertElement(token)
                self.tree.formPointer = self.tree.openElements[-1]

        def startTagListItem(self, token):
            self.parser.framesetOK = False

            stopNamesMap = {"li": ["li"],
                            "dt": ["dt", "dd"],
                            "dd": ["dt", "dd"]}
            stopNames = stopNamesMap[token["name"]]
            for node in reversed(self.tree.openElements):
                if node.name in stopNames:
                    self.parser.phase.processEndTag(
                        impliedTagToken(node.name, "EndTag"))
                    break
                if (node.nameTuple in specialElements and
                        node.name not in ("address", "div", "p")):
                    break

            if self.tree.elementInScope("p", variant="button"):
                self.parser.phase.processEndTag(
                    impliedTagToken("p", "EndTag"))

            self.tree.insertElement(token)

        def startTagPlaintext(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            self.tree.insertElement(token)
            self.parser.tokenizer.state = self.parser.tokenizer.plaintextState

        def startTagHeading(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            if self.tree.openElements[-1].name in headingElements:
                self.parser.parseError("unexpected-start-tag", {"name": token["name"]})
                self.tree.openElements.pop()
            self.tree.insertElement(token)

        def startTagA(self, token):
            afeAElement = self.tree.elementInActiveFormattingElements("a")
            if afeAElement:
                self.parser.parseError("unexpected-start-tag-implies-end-tag",
                                       {"startName": "a", "endName": "a"})
                self.endTagFormatting(impliedTagToken("a"))
                if afeAElement in self.tree.openElements:
                    self.tree.openElements.remove(afeAElement)
                if afeAElement in self.tree.activeFormattingElements:
                    self.tree.activeFormattingElements.remove(afeAElement)
            self.tree.reconstructActiveFormattingElements()
            self.addFormattingElement(token)

        def startTagFormatting(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.addFormattingElement(token)

        def startTagNobr(self, token):
            self.tree.reconstructActiveFormattingElements()
            if self.tree.elementInScope("nobr"):
                self.parser.parseError("unexpected-start-tag-implies-end-tag",
                                       {"startName": "nobr", "endName": "nobr"})
                self.processEndTag(impliedTagToken("nobr"))
                # XXX Need tests that trigger the following
                self.tree.reconstructActiveFormattingElements()
            self.addFormattingElement(token)

        def startTagButton(self, token):
            if self.tree.elementInScope("button"):
                self.parser.parseError("unexpected-start-tag-implies-end-tag",
                                       {"startName": "button", "endName": "button"})
                self.processEndTag(impliedTagToken("button"))
                return token
            else:
                self.tree.reconstructActiveFormattingElements()
                self.tree.insertElement(token)
                self.parser.framesetOK = False

        def startTagAppletMarqueeObject(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertElement(token)
            self.tree.activeFormattingElements.append(Marker)
            self.parser.framesetOK = False

        def startTagXmp(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            self.tree.reconstructActiveFormattingElements()
            self.parser.framesetOK = False
            self.parser.parseRCDataRawtext(token, "RAWTEXT")

        def startTagTable(self, token):
            if self.parser.compatMode != "quirks":
                if self.tree.elementInScope("p", variant="button"):
                    self.processEndTag(impliedTagToken("p"))
            self.tree.insertElement(token)
            self.parser.framesetOK = False
            self.parser.phase = self.parser.phases["inTable"]

        def startTagVoidFormatting(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True
            self.parser.framesetOK = False

        def startTagInput(self, token):
            framesetOK = self.parser.framesetOK
            self.startTagVoidFormatting(token)
            if ("type" in token["data"] and
                    token["data"]["type"].translate(asciiUpper2Lower) == "hidden"):
                # input type=hidden doesn't change framesetOK
                self.parser.framesetOK = framesetOK

        def startTagParamSource(self, token):
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True

        def startTagHr(self, token):
            if self.tree.elementInScope("p", variant="button"):
                self.endTagP(impliedTagToken("p"))
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True
            self.parser.framesetOK = False

        def startTagImage(self, token):
            # No really...
            self.parser.parseError("unexpected-start-tag-treated-as",
                                   {"originalName": "image", "newName": "img"})
            self.processStartTag(impliedTagToken("img", "StartTag",
                                                 attributes=token["data"],
                                                 selfClosing=token["selfClosing"]))

        def startTagIsIndex(self, token):
            self.parser.parseError("deprecated-tag", {"name": "isindex"})
            if self.tree.formPointer:
                return
            form_attrs = {}
            if "action" in token["data"]:
                form_attrs["action"] = token["data"]["action"]
            self.processStartTag(impliedTagToken("form", "StartTag",
                                                 attributes=form_attrs))
            self.processStartTag(impliedTagToken("hr", "StartTag"))
            self.processStartTag(impliedTagToken("label", "StartTag"))
            # XXX Localization ...
            if "prompt" in token["data"]:
                prompt = token["data"]["prompt"]
            else:
                prompt = "This is a searchable index. Enter search keywords: "
            self.processCharacters(
                {"type": tokenTypes["Characters"], "data": prompt})
            attributes = token["data"].copy()
            if "action" in attributes:
                del attributes["action"]
            if "prompt" in attributes:
                del attributes["prompt"]
            attributes["name"] = "isindex"
            self.processStartTag(impliedTagToken("input", "StartTag",
                                                 attributes=attributes,
                                                 selfClosing=token["selfClosing"]))
            self.processEndTag(impliedTagToken("label"))
            self.processStartTag(impliedTagToken("hr", "StartTag"))
            self.processEndTag(impliedTagToken("form"))

        def startTagTextarea(self, token):
            self.tree.insertElement(token)
            self.parser.tokenizer.state = self.parser.tokenizer.rcdataState
            self.processSpaceCharacters = self.processSpaceCharactersDropNewline
            self.parser.framesetOK = False

        def startTagIFrame(self, token):
            self.parser.framesetOK = False
            self.startTagRawtext(token)

        def startTagNoscript(self, token):
            if self.parser.scripting:
                self.startTagRawtext(token)
            else:
                self.startTagOther(token)

        def startTagRawtext(self, token):
            """iframe, noembed noframes, noscript(if scripting enabled)"""
            self.parser.parseRCDataRawtext(token, "RAWTEXT")

        def startTagOpt(self, token):
            if self.tree.openElements[-1].name == "option":
                self.parser.phase.processEndTag(impliedTagToken("option"))
            self.tree.reconstructActiveFormattingElements()
            self.parser.tree.insertElement(token)

        def startTagSelect(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertElement(token)
            self.parser.framesetOK = False
            if self.parser.phase in (self.parser.phases["inTable"],
                                     self.parser.phases["inCaption"],
                                     self.parser.phases["inColumnGroup"],
                                     self.parser.phases["inTableBody"],
                                     self.parser.phases["inRow"],
                                     self.parser.phases["inCell"]):
                self.parser.phase = self.parser.phases["inSelectInTable"]
            else:
                self.parser.phase = self.parser.phases["inSelect"]

        def startTagRpRt(self, token):
            if self.tree.elementInScope("ruby"):
                self.tree.generateImpliedEndTags()
                if self.tree.openElements[-1].name != "ruby":
                    self.parser.parseError()
            self.tree.insertElement(token)

        def startTagMath(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.parser.adjustMathMLAttributes(token)
            self.parser.adjustForeignAttributes(token)
            token["namespace"] = namespaces["mathml"]
            self.tree.insertElement(token)
            # Need to get the parse error right for the case where the token
            # has a namespace not equal to the xmlns attribute
            if token["selfClosing"]:
                self.tree.openElements.pop()
                token["selfClosingAcknowledged"] = True

        def startTagSvg(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.parser.adjustSVGAttributes(token)
            self.parser.adjustForeignAttributes(token)
            token["namespace"] = namespaces["svg"]
            self.tree.insertElement(token)
            # Need to get the parse error right for the case where the token
            # has a namespace not equal to the xmlns attribute
            if token["selfClosing"]:
                self.tree.openElements.pop()
                token["selfClosingAcknowledged"] = True

        def startTagMisplaced(self, token):
            """ Elements that should be children of other elements that have a
            different insertion mode; here they are ignored
            "caption", "col", "colgroup", "frame", "frameset", "head",
            "option", "optgroup", "tbody", "td", "tfoot", "th", "thead",
            "tr", "noscript"
            """
            self.parser.parseError("unexpected-start-tag-ignored", {"name": token["name"]})

        def startTagOther(self, token):
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertElement(token)

        def endTagP(self, token):
            if not self.tree.elementInScope("p", variant="button"):
                self.startTagCloseP(impliedTagToken("p", "StartTag"))
                self.parser.parseError("unexpected-end-tag", {"name": "p"})
                self.endTagP(impliedTagToken("p", "EndTag"))
            else:
                self.tree.generateImpliedEndTags("p")
                if self.tree.openElements[-1].name != "p":
                    self.parser.parseError("unexpected-end-tag", {"name": "p"})
                node = self.tree.openElements.pop()
                while node.name != "p":
                    node = self.tree.openElements.pop()

        def endTagBody(self, token):
            if not self.tree.elementInScope("body"):
                self.parser.parseError()
                return
            elif self.tree.openElements[-1].name != "body":
                for node in self.tree.openElements[2:]:
                    if node.name not in frozenset(("dd", "dt", "li", "optgroup",
                                                   "option", "p", "rp", "rt",
                                                   "tbody", "td", "tfoot",
                                                   "th", "thead", "tr", "body",
                                                   "html")):
                        # Not sure this is the correct name for the parse error
                        self.parser.parseError(
                            "expected-one-end-tag-but-got-another",
                            {"gotName": "body", "expectedName": node.name})
                        break
            self.parser.phase = self.parser.phases["afterBody"]

        def endTagHtml(self, token):
            # We repeat the test for the body end tag token being ignored here
            if self.tree.elementInScope("body"):
                self.endTagBody(impliedTagToken("body"))
                return token

        def endTagBlock(self, token):
            # Put us back in the right whitespace handling mode
            if token["name"] == "pre":
                self.processSpaceCharacters = self.processSpaceCharactersNonPre
            inScope = self.tree.elementInScope(token["name"])
            if inScope:
                self.tree.generateImpliedEndTags()
            if self.tree.openElements[-1].name != token["name"]:
                self.parser.parseError("end-tag-too-early", {"name": token["name"]})
            if inScope:
                node = self.tree.openElements.pop()
                while node.name != token["name"]:
                    node = self.tree.openElements.pop()

        def endTagForm(self, token):
            node = self.tree.formPointer
            self.tree.formPointer = None
            if node is None or not self.tree.elementInScope(node):
                self.parser.parseError("unexpected-end-tag",
                                       {"name": "form"})
            else:
                self.tree.generateImpliedEndTags()
                if self.tree.openElements[-1] != node:
                    self.parser.parseError("end-tag-too-early-ignored",
                                           {"name": "form"})
                self.tree.openElements.remove(node)

        def endTagListItem(self, token):
            if token["name"] == "li":
                variant = "list"
            else:
                variant = None
            if not self.tree.elementInScope(token["name"], variant=variant):
                self.parser.parseError("unexpected-end-tag", {"name": token["name"]})
            else:
                self.tree.generateImpliedEndTags(exclude=token["name"])
                if self.tree.openElements[-1].name != token["name"]:
                    self.parser.parseError(
                        "end-tag-too-early",
                        {"name": token["name"]})
                node = self.tree.openElements.pop()
                while node.name != token["name"]:
                    node = self.tree.openElements.pop()

        def endTagHeading(self, token):
            for item in headingElements:
                if self.tree.elementInScope(item):
                    self.tree.generateImpliedEndTags()
                    break
            if self.tree.openElements[-1].name != token["name"]:
                self.parser.parseError("end-tag-too-early", {"name": token["name"]})

            for item in headingElements:
                if self.tree.elementInScope(item):
                    item = self.tree.openElements.pop()
                    while item.name not in headingElements:
                        item = self.tree.openElements.pop()
                    break

        def endTagFormatting(self, token):
            """The much-feared adoption agency algorithm"""
            # http://svn.whatwg.org/webapps/complete.html#adoptionAgency revision 7867
            # XXX Better parseError messages appreciated.

            # Step 1
            outerLoopCounter = 0

            # Step 2
            while outerLoopCounter < 8:

                # Step 3
                outerLoopCounter += 1

                # Step 4:

                # Let the formatting element be the last element in
                # the list of active formatting elements that:
                # - is between the end of the list and the last scope
                # marker in the list, if any, or the start of the list
                # otherwise, and
                # - has the same tag name as the token.
                formattingElement = self.tree.elementInActiveFormattingElements(
                    token["name"])
                if (not formattingElement or
                    (formattingElement in self.tree.openElements and
                     not self.tree.elementInScope(formattingElement.name))):
                    # If there is no such node, then abort these steps
                    # and instead act as described in the "any other
                    # end tag" entry below.
                    self.endTagOther(token)
                    return

                # Otherwise, if there is such a node, but that node is
                # not in the stack of open elements, then this is a
                # parse error; remove the element from the list, and
                # abort these steps.
                elif formattingElement not in self.tree.openElements:
                    self.parser.parseError("adoption-agency-1.2", {"name": token["name"]})
                    self.tree.activeFormattingElements.remove(formattingElement)
                    return

                # Otherwise, if there is such a node, and that node is
                # also in the stack of open elements, but the element
                # is not in scope, then this is a parse error; ignore
                # the token, and abort these steps.
                elif not self.tree.elementInScope(formattingElement.name):
                    self.parser.parseError("adoption-agency-4.4", {"name": token["name"]})
                    return

                # Otherwise, there is a formatting element and that
                # element is in the stack and is in scope. If the
                # element is not the current node, this is a parse
                # error. In any case, proceed with the algorithm as
                # written in the following steps.
                else:
                    if formattingElement != self.tree.openElements[-1]:
                        self.parser.parseError("adoption-agency-1.3", {"name": token["name"]})

                # Step 5:

                # Let the furthest block be the topmost node in the
                # stack of open elements that is lower in the stack
                # than the formatting element, and is an element in
                # the special category. There might not be one.
                afeIndex = self.tree.openElements.index(formattingElement)
                furthestBlock = None
                for element in self.tree.openElements[afeIndex:]:
                    if element.nameTuple in specialElements:
                        furthestBlock = element
                        break

                # Step 6:

                # If there is no furthest block, then the UA must
                # first pop all the nodes from the bottom of the stack
                # of open elements, from the current node up to and
                # including the formatting element, then remove the
                # formatting element from the list of active
                # formatting elements, and finally abort these steps.
                if furthestBlock is None:
                    element = self.tree.openElements.pop()
                    while element != formattingElement:
                        element = self.tree.openElements.pop()
                    self.tree.activeFormattingElements.remove(element)
                    return

                # Step 7
                commonAncestor = self.tree.openElements[afeIndex - 1]

                # Step 8:
                # The bookmark is supposed to help us identify where to reinsert
                # nodes in step 15. We have to ensure that we reinsert nodes after
                # the node before the active formatting element. Note the bookmark
                # can move in step 9.7
                bookmark = self.tree.activeFormattingElements.index(formattingElement)

                # Step 9
                lastNode = node = furthestBlock
                innerLoopCounter = 0

                index = self.tree.openElements.index(node)
                while innerLoopCounter < 3:
                    innerLoopCounter += 1
                    # Node is element before node in open elements
                    index -= 1
                    node = self.tree.openElements[index]
                    if node not in self.tree.activeFormattingElements:
                        self.tree.openElements.remove(node)
                        continue
                    # Step 9.6
                    if node == formattingElement:
                        break
                    # Step 9.7
                    if lastNode == furthestBlock:
                        bookmark = self.tree.activeFormattingElements.index(node) + 1
                    # Step 9.8
                    clone = node.cloneNode()
                    # Replace node with clone
                    self.tree.activeFormattingElements[
                        self.tree.activeFormattingElements.index(node)] = clone
                    self.tree.openElements[
                        self.tree.openElements.index(node)] = clone
                    node = clone
                    # Step 9.9
                    # Remove lastNode from its parents, if any
                    if lastNode.parent:
                        lastNode.parent.removeChild(lastNode)
                    node.appendChild(lastNode)
                    # Step 9.10
                    lastNode = node

                # Step 10
                # Foster parent lastNode if commonAncestor is a
                # table, tbody, tfoot, thead, or tr we need to foster
                # parent the lastNode
                if lastNode.parent:
                    lastNode.parent.removeChild(lastNode)

                if commonAncestor.name in frozenset(("table", "tbody", "tfoot", "thead", "tr")):
                    parent, insertBefore = self.tree.getTableMisnestedNodePosition()
                    parent.insertBefore(lastNode, insertBefore)
                else:
                    commonAncestor.appendChild(lastNode)

                # Step 11
                clone = formattingElement.cloneNode()

                # Step 12
                furthestBlock.reparentChildren(clone)

                # Step 13
                furthestBlock.appendChild(clone)

                # Step 14
                self.tree.activeFormattingElements.remove(formattingElement)
                self.tree.activeFormattingElements.insert(bookmark, clone)

                # Step 15
                self.tree.openElements.remove(formattingElement)
                self.tree.openElements.insert(
                    self.tree.openElements.index(furthestBlock) + 1, clone)

        def endTagAppletMarqueeObject(self, token):
            if self.tree.elementInScope(token["name"]):
                self.tree.generateImpliedEndTags()
            if self.tree.openElements[-1].name != token["name"]:
                self.parser.parseError("end-tag-too-early", {"name": token["name"]})

            if self.tree.elementInScope(token["name"]):
                element = self.tree.openElements.pop()
                while element.name != token["name"]:
                    element = self.tree.openElements.pop()
                self.tree.clearActiveFormattingElements()

        def endTagBr(self, token):
            self.parser.parseError("unexpected-end-tag-treated-as",
                                   {"originalName": "br", "newName": "br element"})
            self.tree.reconstructActiveFormattingElements()
            self.tree.insertElement(impliedTagToken("br", "StartTag"))
            self.tree.openElements.pop()

        def endTagOther(self, token):
            for node in self.tree.openElements[::-1]:
                if node.name == token["name"]:
                    self.tree.generateImpliedEndTags(exclude=token["name"])
                    if self.tree.openElements[-1].name != token["name"]:
                        self.parser.parseError("unexpected-end-tag", {"name": token["name"]})
                    while self.tree.openElements.pop() != node:
                        pass
                    break
                else:
                    if node.nameTuple in specialElements:
                        self.parser.parseError("unexpected-end-tag", {"name": token["name"]})
                        break

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            (("base", "basefont", "bgsound", "command", "link", "meta",
              "script", "style", "title"),
             startTagProcessInHead),
            ("body", startTagBody),
            ("frameset", startTagFrameset),
            (("address", "article", "aside", "blockquote", "center", "details",
              "dir", "div", "dl", "fieldset", "figcaption", "figure",
              "footer", "header", "hgroup", "main", "menu", "nav", "ol", "p",
              "section", "summary", "ul"),
             startTagCloseP),
            (headingElements, startTagHeading),
            (("pre", "listing"), startTagPreListing),
            ("form", startTagForm),
            (("li", "dd", "dt"), startTagListItem),
            ("plaintext", startTagPlaintext),
            ("a", startTagA),
            (("b", "big", "code", "em", "font", "i", "s", "small", "strike",
              "strong", "tt", "u"), startTagFormatting),
            ("nobr", startTagNobr),
            ("button", startTagButton),
            (("applet", "marquee", "object"), startTagAppletMarqueeObject),
            ("xmp", startTagXmp),
            ("table", startTagTable),
            (("area", "br", "embed", "img", "keygen", "wbr"),
             startTagVoidFormatting),
            (("param", "source", "track"), startTagParamSource),
            ("input", startTagInput),
            ("hr", startTagHr),
            ("image", startTagImage),
            ("isindex", startTagIsIndex),
            ("textarea", startTagTextarea),
            ("iframe", startTagIFrame),
            ("noscript", startTagNoscript),
            (("noembed", "noframes"), startTagRawtext),
            ("select", startTagSelect),
            (("rp", "rt"), startTagRpRt),
            (("option", "optgroup"), startTagOpt),
            (("math"), startTagMath),
            (("svg"), startTagSvg),
            (("caption", "col", "colgroup", "frame", "head",
              "tbody", "td", "tfoot", "th", "thead",
              "tr"), startTagMisplaced)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("body", endTagBody),
            ("html", endTagHtml),
            (("address", "article", "aside", "blockquote", "button", "center",
              "details", "dialog", "dir", "div", "dl", "fieldset", "figcaption", "figure",
              "footer", "header", "hgroup", "listing", "main", "menu", "nav", "ol", "pre",
              "section", "summary", "ul"), endTagBlock),
            ("form", endTagForm),
            ("p", endTagP),
            (("dd", "dt", "li"), endTagListItem),
            (headingElements, endTagHeading),
            (("a", "b", "big", "code", "em", "font", "i", "nobr", "s", "small",
              "strike", "strong", "tt", "u"), endTagFormatting),
            (("applet", "marquee", "object"), endTagAppletMarqueeObject),
            ("br", endTagBr),
        ])
        endTagHandler.default = endTagOther

    class TextPhase(Phase):
        __slots__ = tuple()

        def processCharacters(self, token):
            self.tree.insertText(token["data"])

        def processEOF(self):
            self.parser.parseError("expected-named-closing-tag-but-got-eof",
                                   {"name": self.tree.openElements[-1].name})
            self.tree.openElements.pop()
            self.parser.phase = self.parser.originalPhase
            return True

        def startTagOther(self, token):
            assert False, "Tried to process start tag %s in RCDATA/RAWTEXT mode" % token['name']

        def endTagScript(self, token):
            node = self.tree.openElements.pop()
            assert node.name == "script"
            self.parser.phase = self.parser.originalPhase
            # The rest of this method is all stuff that only happens if
            # document.write works

        def endTagOther(self, token):
            self.tree.openElements.pop()
            self.parser.phase = self.parser.originalPhase

        startTagHandler = _utils.MethodDispatcher([])
        startTagHandler.default = startTagOther
        endTagHandler = _utils.MethodDispatcher([
            ("script", endTagScript)])
        endTagHandler.default = endTagOther

    class InTablePhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-table
        __slots__ = tuple()

        # helper methods
        def clearStackToTableContext(self):
            # "clear the stack back to a table context"
            while self.tree.openElements[-1].name not in ("table", "html"):
                # self.parser.parseError("unexpected-implied-end-tag-in-table",
                #  {"name":  self.tree.openElements[-1].name})
                self.tree.openElements.pop()
            # When the current node is <html> it's an innerHTML case

        # processing methods
        def processEOF(self):
            if self.tree.openElements[-1].name != "html":
                self.parser.parseError("eof-in-table")
            else:
                assert self.parser.innerHTML
            # Stop parsing

        def processSpaceCharacters(self, token):
            originalPhase = self.parser.phase
            self.parser.phase = self.parser.phases["inTableText"]
            self.parser.phase.originalPhase = originalPhase
            self.parser.phase.processSpaceCharacters(token)

        def processCharacters(self, token):
            originalPhase = self.parser.phase
            self.parser.phase = self.parser.phases["inTableText"]
            self.parser.phase.originalPhase = originalPhase
            self.parser.phase.processCharacters(token)

        def insertText(self, token):
            # If we get here there must be at least one non-whitespace character
            # Do the table magic!
            self.tree.insertFromTable = True
            self.parser.phases["inBody"].processCharacters(token)
            self.tree.insertFromTable = False

        def startTagCaption(self, token):
            self.clearStackToTableContext()
            self.tree.activeFormattingElements.append(Marker)
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inCaption"]

        def startTagColgroup(self, token):
            self.clearStackToTableContext()
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inColumnGroup"]

        def startTagCol(self, token):
            self.startTagColgroup(impliedTagToken("colgroup", "StartTag"))
            return token

        def startTagRowGroup(self, token):
            self.clearStackToTableContext()
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inTableBody"]

        def startTagImplyTbody(self, token):
            self.startTagRowGroup(impliedTagToken("tbody", "StartTag"))
            return token

        def startTagTable(self, token):
            self.parser.parseError("unexpected-start-tag-implies-end-tag",
                                   {"startName": "table", "endName": "table"})
            self.parser.phase.processEndTag(impliedTagToken("table"))
            if not self.parser.innerHTML:
                return token

        def startTagStyleScript(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagInput(self, token):
            if ("type" in token["data"] and
                    token["data"]["type"].translate(asciiUpper2Lower) == "hidden"):
                self.parser.parseError("unexpected-hidden-input-in-table")
                self.tree.insertElement(token)
                # XXX associate with form
                self.tree.openElements.pop()
            else:
                self.startTagOther(token)

        def startTagForm(self, token):
            self.parser.parseError("unexpected-form-in-table")
            if self.tree.formPointer is None:
                self.tree.insertElement(token)
                self.tree.formPointer = self.tree.openElements[-1]
                self.tree.openElements.pop()

        def startTagOther(self, token):
            self.parser.parseError("unexpected-start-tag-implies-table-voodoo", {"name": token["name"]})
            # Do the table magic!
            self.tree.insertFromTable = True
            self.parser.phases["inBody"].processStartTag(token)
            self.tree.insertFromTable = False

        def endTagTable(self, token):
            if self.tree.elementInScope("table", variant="table"):
                self.tree.generateImpliedEndTags()
                if self.tree.openElements[-1].name != "table":
                    self.parser.parseError("end-tag-too-early-named",
                                           {"gotName": "table",
                                            "expectedName": self.tree.openElements[-1].name})
                while self.tree.openElements[-1].name != "table":
                    self.tree.openElements.pop()
                self.tree.openElements.pop()
                self.parser.resetInsertionMode()
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def endTagIgnore(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag-implies-table-voodoo", {"name": token["name"]})
            # Do the table magic!
            self.tree.insertFromTable = True
            self.parser.phases["inBody"].processEndTag(token)
            self.tree.insertFromTable = False

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("caption", startTagCaption),
            ("colgroup", startTagColgroup),
            ("col", startTagCol),
            (("tbody", "tfoot", "thead"), startTagRowGroup),
            (("td", "th", "tr"), startTagImplyTbody),
            ("table", startTagTable),
            (("style", "script"), startTagStyleScript),
            ("input", startTagInput),
            ("form", startTagForm)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("table", endTagTable),
            (("body", "caption", "col", "colgroup", "html", "tbody", "td",
              "tfoot", "th", "thead", "tr"), endTagIgnore)
        ])
        endTagHandler.default = endTagOther

    class InTableTextPhase(Phase):
        __slots__ = ("originalPhase", "characterTokens")

        def __init__(self, *args, **kwargs):
            super(InTableTextPhase, self).__init__(*args, **kwargs)
            self.originalPhase = None
            self.characterTokens = []

        def flushCharacters(self):
            data = "".join([item["data"] for item in self.characterTokens])
            if any([item not in spaceCharacters for item in data]):
                token = {"type": tokenTypes["Characters"], "data": data}
                self.parser.phases["inTable"].insertText(token)
            elif data:
                self.tree.insertText(data)
            self.characterTokens = []

        def processComment(self, token):
            self.flushCharacters()
            self.parser.phase = self.originalPhase
            return token

        def processEOF(self):
            self.flushCharacters()
            self.parser.phase = self.originalPhase
            return True

        def processCharacters(self, token):
            if token["data"] == "\u0000":
                return
            self.characterTokens.append(token)

        def processSpaceCharacters(self, token):
            # pretty sure we should never reach here
            self.characterTokens.append(token)
    #        assert False

        def processStartTag(self, token):
            self.flushCharacters()
            self.parser.phase = self.originalPhase
            return token

        def processEndTag(self, token):
            self.flushCharacters()
            self.parser.phase = self.originalPhase
            return token

    class InCaptionPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-caption
        __slots__ = tuple()

        def ignoreEndTagCaption(self):
            return not self.tree.elementInScope("caption", variant="table")

        def processEOF(self):
            self.parser.phases["inBody"].processEOF()

        def processCharacters(self, token):
            return self.parser.phases["inBody"].processCharacters(token)

        def startTagTableElement(self, token):
            self.parser.parseError()
            # XXX Have to duplicate logic here to find out if the tag is ignored
            ignoreEndTag = self.ignoreEndTagCaption()
            self.parser.phase.processEndTag(impliedTagToken("caption"))
            if not ignoreEndTag:
                return token

        def startTagOther(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def endTagCaption(self, token):
            if not self.ignoreEndTagCaption():
                # AT this code is quite similar to endTagTable in "InTable"
                self.tree.generateImpliedEndTags()
                if self.tree.openElements[-1].name != "caption":
                    self.parser.parseError("expected-one-end-tag-but-got-another",
                                           {"gotName": "caption",
                                            "expectedName": self.tree.openElements[-1].name})
                while self.tree.openElements[-1].name != "caption":
                    self.tree.openElements.pop()
                self.tree.openElements.pop()
                self.tree.clearActiveFormattingElements()
                self.parser.phase = self.parser.phases["inTable"]
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def endTagTable(self, token):
            self.parser.parseError()
            ignoreEndTag = self.ignoreEndTagCaption()
            self.parser.phase.processEndTag(impliedTagToken("caption"))
            if not ignoreEndTag:
                return token

        def endTagIgnore(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def endTagOther(self, token):
            return self.parser.phases["inBody"].processEndTag(token)

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            (("caption", "col", "colgroup", "tbody", "td", "tfoot", "th",
              "thead", "tr"), startTagTableElement)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("caption", endTagCaption),
            ("table", endTagTable),
            (("body", "col", "colgroup", "html", "tbody", "td", "tfoot", "th",
              "thead", "tr"), endTagIgnore)
        ])
        endTagHandler.default = endTagOther

    class InColumnGroupPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-column
        __slots__ = tuple()

        def ignoreEndTagColgroup(self):
            return self.tree.openElements[-1].name == "html"

        def processEOF(self):
            if self.tree.openElements[-1].name == "html":
                assert self.parser.innerHTML
                return
            else:
                ignoreEndTag = self.ignoreEndTagColgroup()
                self.endTagColgroup(impliedTagToken("colgroup"))
                if not ignoreEndTag:
                    return True

        def processCharacters(self, token):
            ignoreEndTag = self.ignoreEndTagColgroup()
            self.endTagColgroup(impliedTagToken("colgroup"))
            if not ignoreEndTag:
                return token

        def startTagCol(self, token):
            self.tree.insertElement(token)
            self.tree.openElements.pop()
            token["selfClosingAcknowledged"] = True

        def startTagOther(self, token):
            ignoreEndTag = self.ignoreEndTagColgroup()
            self.endTagColgroup(impliedTagToken("colgroup"))
            if not ignoreEndTag:
                return token

        def endTagColgroup(self, token):
            if self.ignoreEndTagColgroup():
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()
            else:
                self.tree.openElements.pop()
                self.parser.phase = self.parser.phases["inTable"]

        def endTagCol(self, token):
            self.parser.parseError("no-end-tag", {"name": "col"})

        def endTagOther(self, token):
            ignoreEndTag = self.ignoreEndTagColgroup()
            self.endTagColgroup(impliedTagToken("colgroup"))
            if not ignoreEndTag:
                return token

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("col", startTagCol)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("colgroup", endTagColgroup),
            ("col", endTagCol)
        ])
        endTagHandler.default = endTagOther

    class InTableBodyPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-table0
        __slots__ = tuple()

        # helper methods
        def clearStackToTableBodyContext(self):
            while self.tree.openElements[-1].name not in ("tbody", "tfoot",
                                                          "thead", "html"):
                # self.parser.parseError("unexpected-implied-end-tag-in-table",
                #  {"name": self.tree.openElements[-1].name})
                self.tree.openElements.pop()
            if self.tree.openElements[-1].name == "html":
                assert self.parser.innerHTML

        # the rest
        def processEOF(self):
            self.parser.phases["inTable"].processEOF()

        def processSpaceCharacters(self, token):
            return self.parser.phases["inTable"].processSpaceCharacters(token)

        def processCharacters(self, token):
            return self.parser.phases["inTable"].processCharacters(token)

        def startTagTr(self, token):
            self.clearStackToTableBodyContext()
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inRow"]

        def startTagTableCell(self, token):
            self.parser.parseError("unexpected-cell-in-table-body",
                                   {"name": token["name"]})
            self.startTagTr(impliedTagToken("tr", "StartTag"))
            return token

        def startTagTableOther(self, token):
            # XXX AT Any ideas on how to share this with endTagTable?
            if (self.tree.elementInScope("tbody", variant="table") or
                self.tree.elementInScope("thead", variant="table") or
                    self.tree.elementInScope("tfoot", variant="table")):
                self.clearStackToTableBodyContext()
                self.endTagTableRowGroup(
                    impliedTagToken(self.tree.openElements[-1].name))
                return token
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def startTagOther(self, token):
            return self.parser.phases["inTable"].processStartTag(token)

        def endTagTableRowGroup(self, token):
            if self.tree.elementInScope(token["name"], variant="table"):
                self.clearStackToTableBodyContext()
                self.tree.openElements.pop()
                self.parser.phase = self.parser.phases["inTable"]
            else:
                self.parser.parseError("unexpected-end-tag-in-table-body",
                                       {"name": token["name"]})

        def endTagTable(self, token):
            if (self.tree.elementInScope("tbody", variant="table") or
                self.tree.elementInScope("thead", variant="table") or
                    self.tree.elementInScope("tfoot", variant="table")):
                self.clearStackToTableBodyContext()
                self.endTagTableRowGroup(
                    impliedTagToken(self.tree.openElements[-1].name))
                return token
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def endTagIgnore(self, token):
            self.parser.parseError("unexpected-end-tag-in-table-body",
                                   {"name": token["name"]})

        def endTagOther(self, token):
            return self.parser.phases["inTable"].processEndTag(token)

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("tr", startTagTr),
            (("td", "th"), startTagTableCell),
            (("caption", "col", "colgroup", "tbody", "tfoot", "thead"),
             startTagTableOther)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            (("tbody", "tfoot", "thead"), endTagTableRowGroup),
            ("table", endTagTable),
            (("body", "caption", "col", "colgroup", "html", "td", "th",
              "tr"), endTagIgnore)
        ])
        endTagHandler.default = endTagOther

    class InRowPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-row
        __slots__ = tuple()

        # helper methods (XXX unify this with other table helper methods)
        def clearStackToTableRowContext(self):
            while self.tree.openElements[-1].name not in ("tr", "html"):
                self.parser.parseError("unexpected-implied-end-tag-in-table-row",
                                       {"name": self.tree.openElements[-1].name})
                self.tree.openElements.pop()

        def ignoreEndTagTr(self):
            return not self.tree.elementInScope("tr", variant="table")

        # the rest
        def processEOF(self):
            self.parser.phases["inTable"].processEOF()

        def processSpaceCharacters(self, token):
            return self.parser.phases["inTable"].processSpaceCharacters(token)

        def processCharacters(self, token):
            return self.parser.phases["inTable"].processCharacters(token)

        def startTagTableCell(self, token):
            self.clearStackToTableRowContext()
            self.tree.insertElement(token)
            self.parser.phase = self.parser.phases["inCell"]
            self.tree.activeFormattingElements.append(Marker)

        def startTagTableOther(self, token):
            ignoreEndTag = self.ignoreEndTagTr()
            self.endTagTr(impliedTagToken("tr"))
            # XXX how are we sure it's always ignored in the innerHTML case?
            if not ignoreEndTag:
                return token

        def startTagOther(self, token):
            return self.parser.phases["inTable"].processStartTag(token)

        def endTagTr(self, token):
            if not self.ignoreEndTagTr():
                self.clearStackToTableRowContext()
                self.tree.openElements.pop()
                self.parser.phase = self.parser.phases["inTableBody"]
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def endTagTable(self, token):
            ignoreEndTag = self.ignoreEndTagTr()
            self.endTagTr(impliedTagToken("tr"))
            # Reprocess the current tag if the tr end tag was not ignored
            # XXX how are we sure it's always ignored in the innerHTML case?
            if not ignoreEndTag:
                return token

        def endTagTableRowGroup(self, token):
            if self.tree.elementInScope(token["name"], variant="table"):
                self.endTagTr(impliedTagToken("tr"))
                return token
            else:
                self.parser.parseError()

        def endTagIgnore(self, token):
            self.parser.parseError("unexpected-end-tag-in-table-row",
                                   {"name": token["name"]})

        def endTagOther(self, token):
            return self.parser.phases["inTable"].processEndTag(token)

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            (("td", "th"), startTagTableCell),
            (("caption", "col", "colgroup", "tbody", "tfoot", "thead",
              "tr"), startTagTableOther)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("tr", endTagTr),
            ("table", endTagTable),
            (("tbody", "tfoot", "thead"), endTagTableRowGroup),
            (("body", "caption", "col", "colgroup", "html", "td", "th"),
             endTagIgnore)
        ])
        endTagHandler.default = endTagOther

    class InCellPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-cell
        __slots__ = tuple()

        # helper
        def closeCell(self):
            if self.tree.elementInScope("td", variant="table"):
                self.endTagTableCell(impliedTagToken("td"))
            elif self.tree.elementInScope("th", variant="table"):
                self.endTagTableCell(impliedTagToken("th"))

        # the rest
        def processEOF(self):
            self.parser.phases["inBody"].processEOF()

        def processCharacters(self, token):
            return self.parser.phases["inBody"].processCharacters(token)

        def startTagTableOther(self, token):
            if (self.tree.elementInScope("td", variant="table") or
                    self.tree.elementInScope("th", variant="table")):
                self.closeCell()
                return token
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def startTagOther(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def endTagTableCell(self, token):
            if self.tree.elementInScope(token["name"], variant="table"):
                self.tree.generateImpliedEndTags(token["name"])
                if self.tree.openElements[-1].name != token["name"]:
                    self.parser.parseError("unexpected-cell-end-tag",
                                           {"name": token["name"]})
                    while True:
                        node = self.tree.openElements.pop()
                        if node.name == token["name"]:
                            break
                else:
                    self.tree.openElements.pop()
                self.tree.clearActiveFormattingElements()
                self.parser.phase = self.parser.phases["inRow"]
            else:
                self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def endTagIgnore(self, token):
            self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

        def endTagImply(self, token):
            if self.tree.elementInScope(token["name"], variant="table"):
                self.closeCell()
                return token
            else:
                # sometimes innerHTML case
                self.parser.parseError()

        def endTagOther(self, token):
            return self.parser.phases["inBody"].processEndTag(token)

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            (("caption", "col", "colgroup", "tbody", "td", "tfoot", "th",
              "thead", "tr"), startTagTableOther)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            (("td", "th"), endTagTableCell),
            (("body", "caption", "col", "colgroup", "html"), endTagIgnore),
            (("table", "tbody", "tfoot", "thead", "tr"), endTagImply)
        ])
        endTagHandler.default = endTagOther

    class InSelectPhase(Phase):
        __slots__ = tuple()

        # http://www.whatwg.org/specs/web-apps/current-work/#in-select
        def processEOF(self):
            if self.tree.openElements[-1].name != "html":
                self.parser.parseError("eof-in-select")
            else:
                assert self.parser.innerHTML

        def processCharacters(self, token):
            if token["data"] == "\u0000":
                return
            self.tree.insertText(token["data"])

        def startTagOption(self, token):
            # We need to imply </option> if <option> is the current node.
            if self.tree.openElements[-1].name == "option":
                self.tree.openElements.pop()
            self.tree.insertElement(token)

        def startTagOptgroup(self, token):
            if self.tree.openElements[-1].name == "option":
                self.tree.openElements.pop()
            if self.tree.openElements[-1].name == "optgroup":
                self.tree.openElements.pop()
            self.tree.insertElement(token)

        def startTagSelect(self, token):
            self.parser.parseError("unexpected-select-in-select")
            self.endTagSelect(impliedTagToken("select"))

        def startTagInput(self, token):
            self.parser.parseError("unexpected-input-in-select")
            if self.tree.elementInScope("select", variant="select"):
                self.endTagSelect(impliedTagToken("select"))
                return token
            else:
                assert self.parser.innerHTML

        def startTagScript(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("unexpected-start-tag-in-select",
                                   {"name": token["name"]})

        def endTagOption(self, token):
            if self.tree.openElements[-1].name == "option":
                self.tree.openElements.pop()
            else:
                self.parser.parseError("unexpected-end-tag-in-select",
                                       {"name": "option"})

        def endTagOptgroup(self, token):
            # </optgroup> implicitly closes <option>
            if (self.tree.openElements[-1].name == "option" and
                    self.tree.openElements[-2].name == "optgroup"):
                self.tree.openElements.pop()
            # It also closes </optgroup>
            if self.tree.openElements[-1].name == "optgroup":
                self.tree.openElements.pop()
            # But nothing else
            else:
                self.parser.parseError("unexpected-end-tag-in-select",
                                       {"name": "optgroup"})

        def endTagSelect(self, token):
            if self.tree.elementInScope("select", variant="select"):
                node = self.tree.openElements.pop()
                while node.name != "select":
                    node = self.tree.openElements.pop()
                self.parser.resetInsertionMode()
            else:
                # innerHTML case
                assert self.parser.innerHTML
                self.parser.parseError()

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag-in-select",
                                   {"name": token["name"]})

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("option", startTagOption),
            ("optgroup", startTagOptgroup),
            ("select", startTagSelect),
            (("input", "keygen", "textarea"), startTagInput),
            ("script", startTagScript)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("option", endTagOption),
            ("optgroup", endTagOptgroup),
            ("select", endTagSelect)
        ])
        endTagHandler.default = endTagOther

    class InSelectInTablePhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            self.parser.phases["inSelect"].processEOF()

        def processCharacters(self, token):
            return self.parser.phases["inSelect"].processCharacters(token)

        def startTagTable(self, token):
            self.parser.parseError("unexpected-table-element-start-tag-in-select-in-table", {"name": token["name"]})
            self.endTagOther(impliedTagToken("select"))
            return token

        def startTagOther(self, token):
            return self.parser.phases["inSelect"].processStartTag(token)

        def endTagTable(self, token):
            self.parser.parseError("unexpected-table-element-end-tag-in-select-in-table", {"name": token["name"]})
            if self.tree.elementInScope(token["name"], variant="table"):
                self.endTagOther(impliedTagToken("select"))
                return token

        def endTagOther(self, token):
            return self.parser.phases["inSelect"].processEndTag(token)

        startTagHandler = _utils.MethodDispatcher([
            (("caption", "table", "tbody", "tfoot", "thead", "tr", "td", "th"),
             startTagTable)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            (("caption", "table", "tbody", "tfoot", "thead", "tr", "td", "th"),
             endTagTable)
        ])
        endTagHandler.default = endTagOther

    class InForeignContentPhase(Phase):
        __slots__ = tuple()

        breakoutElements = frozenset(["b", "big", "blockquote", "body", "br",
                                      "center", "code", "dd", "div", "dl", "dt",
                                      "em", "embed", "h1", "h2", "h3",
                                      "h4", "h5", "h6", "head", "hr", "i", "img",
                                      "li", "listing", "menu", "meta", "nobr",
                                      "ol", "p", "pre", "ruby", "s", "small",
                                      "span", "strong", "strike", "sub", "sup",
                                      "table", "tt", "u", "ul", "var"])

        def adjustSVGTagNames(self, token):
            replacements = {"altglyph": "altGlyph",
                            "altglyphdef": "altGlyphDef",
                            "altglyphitem": "altGlyphItem",
                            "animatecolor": "animateColor",
                            "animatemotion": "animateMotion",
                            "animatetransform": "animateTransform",
                            "clippath": "clipPath",
                            "feblend": "feBlend",
                            "fecolormatrix": "feColorMatrix",
                            "fecomponenttransfer": "feComponentTransfer",
                            "fecomposite": "feComposite",
                            "feconvolvematrix": "feConvolveMatrix",
                            "fediffuselighting": "feDiffuseLighting",
                            "fedisplacementmap": "feDisplacementMap",
                            "fedistantlight": "feDistantLight",
                            "feflood": "feFlood",
                            "fefunca": "feFuncA",
                            "fefuncb": "feFuncB",
                            "fefuncg": "feFuncG",
                            "fefuncr": "feFuncR",
                            "fegaussianblur": "feGaussianBlur",
                            "feimage": "feImage",
                            "femerge": "feMerge",
                            "femergenode": "feMergeNode",
                            "femorphology": "feMorphology",
                            "feoffset": "feOffset",
                            "fepointlight": "fePointLight",
                            "fespecularlighting": "feSpecularLighting",
                            "fespotlight": "feSpotLight",
                            "fetile": "feTile",
                            "feturbulence": "feTurbulence",
                            "foreignobject": "foreignObject",
                            "glyphref": "glyphRef",
                            "lineargradient": "linearGradient",
                            "radialgradient": "radialGradient",
                            "textpath": "textPath"}

            if token["name"] in replacements:
                token["name"] = replacements[token["name"]]

        def processCharacters(self, token):
            if token["data"] == "\u0000":
                token["data"] = "\uFFFD"
            elif (self.parser.framesetOK and
                  any(char not in spaceCharacters for char in token["data"])):
                self.parser.framesetOK = False
            Phase.processCharacters(self, token)

        def processStartTag(self, token):
            currentNode = self.tree.openElements[-1]
            if (token["name"] in self.breakoutElements or
                (token["name"] == "font" and
                 set(token["data"].keys()) & {"color", "face", "size"})):
                self.parser.parseError("unexpected-html-element-in-foreign-content",
                                       {"name": token["name"]})
                while (self.tree.openElements[-1].namespace !=
                       self.tree.defaultNamespace and
                       not self.parser.isHTMLIntegrationPoint(self.tree.openElements[-1]) and
                       not self.parser.isMathMLTextIntegrationPoint(self.tree.openElements[-1])):
                    self.tree.openElements.pop()
                return token

            else:
                if currentNode.namespace == namespaces["mathml"]:
                    self.parser.adjustMathMLAttributes(token)
                elif currentNode.namespace == namespaces["svg"]:
                    self.adjustSVGTagNames(token)
                    self.parser.adjustSVGAttributes(token)
                self.parser.adjustForeignAttributes(token)
                token["namespace"] = currentNode.namespace
                self.tree.insertElement(token)
                if token["selfClosing"]:
                    self.tree.openElements.pop()
                    token["selfClosingAcknowledged"] = True

        def processEndTag(self, token):
            nodeIndex = len(self.tree.openElements) - 1
            node = self.tree.openElements[-1]
            if node.name.translate(asciiUpper2Lower) != token["name"]:
                self.parser.parseError("unexpected-end-tag", {"name": token["name"]})

            while True:
                if node.name.translate(asciiUpper2Lower) == token["name"]:
                    # XXX this isn't in the spec but it seems necessary
                    if self.parser.phase == self.parser.phases["inTableText"]:
                        self.parser.phase.flushCharacters()
                        self.parser.phase = self.parser.phase.originalPhase
                    while self.tree.openElements.pop() != node:
                        assert self.tree.openElements
                    new_token = None
                    break
                nodeIndex -= 1

                node = self.tree.openElements[nodeIndex]
                if node.namespace != self.tree.defaultNamespace:
                    continue
                else:
                    new_token = self.parser.phase.processEndTag(token)
                    break
            return new_token

    class AfterBodyPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            # Stop parsing
            pass

        def processComment(self, token):
            # This is needed because data is to be appended to the <html> element
            # here and not to whatever is currently open.
            self.tree.insertComment(token, self.tree.openElements[0])

        def processCharacters(self, token):
            self.parser.parseError("unexpected-char-after-body")
            self.parser.phase = self.parser.phases["inBody"]
            return token

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("unexpected-start-tag-after-body",
                                   {"name": token["name"]})
            self.parser.phase = self.parser.phases["inBody"]
            return token

        def endTagHtml(self, name):
            if self.parser.innerHTML:
                self.parser.parseError("unexpected-end-tag-after-body-innerhtml")
            else:
                self.parser.phase = self.parser.phases["afterAfterBody"]

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag-after-body",
                                   {"name": token["name"]})
            self.parser.phase = self.parser.phases["inBody"]
            return token

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([("html", endTagHtml)])
        endTagHandler.default = endTagOther

    class InFramesetPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#in-frameset
        __slots__ = tuple()

        def processEOF(self):
            if self.tree.openElements[-1].name != "html":
                self.parser.parseError("eof-in-frameset")
            else:
                assert self.parser.innerHTML

        def processCharacters(self, token):
            self.parser.parseError("unexpected-char-in-frameset")

        def startTagFrameset(self, token):
            self.tree.insertElement(token)

        def startTagFrame(self, token):
            self.tree.insertElement(token)
            self.tree.openElements.pop()

        def startTagNoframes(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("unexpected-start-tag-in-frameset",
                                   {"name": token["name"]})

        def endTagFrameset(self, token):
            if self.tree.openElements[-1].name == "html":
                # innerHTML case
                self.parser.parseError("unexpected-frameset-in-frameset-innerhtml")
            else:
                self.tree.openElements.pop()
            if (not self.parser.innerHTML and
                    self.tree.openElements[-1].name != "frameset"):
                # If we're not in innerHTML mode and the current node is not a
                # "frameset" element (anymore) then switch.
                self.parser.phase = self.parser.phases["afterFrameset"]

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag-in-frameset",
                                   {"name": token["name"]})

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("frameset", startTagFrameset),
            ("frame", startTagFrame),
            ("noframes", startTagNoframes)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("frameset", endTagFrameset)
        ])
        endTagHandler.default = endTagOther

    class AfterFramesetPhase(Phase):
        # http://www.whatwg.org/specs/web-apps/current-work/#after3
        __slots__ = tuple()

        def processEOF(self):
            # Stop parsing
            pass

        def processCharacters(self, token):
            self.parser.parseError("unexpected-char-after-frameset")

        def startTagNoframes(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("unexpected-start-tag-after-frameset",
                                   {"name": token["name"]})

        def endTagHtml(self, token):
            self.parser.phase = self.parser.phases["afterAfterFrameset"]

        def endTagOther(self, token):
            self.parser.parseError("unexpected-end-tag-after-frameset",
                                   {"name": token["name"]})

        startTagHandler = _utils.MethodDispatcher([
            ("html", Phase.startTagHtml),
            ("noframes", startTagNoframes)
        ])
        startTagHandler.default = startTagOther

        endTagHandler = _utils.MethodDispatcher([
            ("html", endTagHtml)
        ])
        endTagHandler.default = endTagOther

    class AfterAfterBodyPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            pass

        def processComment(self, token):
            self.tree.insertComment(token, self.tree.document)

        def processSpaceCharacters(self, token):
            return self.parser.phases["inBody"].processSpaceCharacters(token)

        def processCharacters(self, token):
            self.parser.parseError("expected-eof-but-got-char")
            self.parser.phase = self.parser.phases["inBody"]
            return token

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("expected-eof-but-got-start-tag",
                                   {"name": token["name"]})
            self.parser.phase = self.parser.phases["inBody"]
            return token

        def processEndTag(self, token):
            self.parser.parseError("expected-eof-but-got-end-tag",
                                   {"name": token["name"]})
            self.parser.phase = self.parser.phases["inBody"]
            return token

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml)
        ])
        startTagHandler.default = startTagOther

    class AfterAfterFramesetPhase(Phase):
        __slots__ = tuple()

        def processEOF(self):
            pass

        def processComment(self, token):
            self.tree.insertComment(token, self.tree.document)

        def processSpaceCharacters(self, token):
            return self.parser.phases["inBody"].processSpaceCharacters(token)

        def processCharacters(self, token):
            self.parser.parseError("expected-eof-but-got-char")

        def startTagHtml(self, token):
            return self.parser.phases["inBody"].processStartTag(token)

        def startTagNoFrames(self, token):
            return self.parser.phases["inHead"].processStartTag(token)

        def startTagOther(self, token):
            self.parser.parseError("expected-eof-but-got-start-tag",
                                   {"name": token["name"]})

        def processEndTag(self, token):
            self.parser.parseError("expected-eof-but-got-end-tag",
                                   {"name": token["name"]})

        startTagHandler = _utils.MethodDispatcher([
            ("html", startTagHtml),
            ("noframes", startTagNoFrames)
        ])
        startTagHandler.default = startTagOther

    # pylint:enable=unused-argument

    return {
        "initial": InitialPhase,
        "beforeHtml": BeforeHtmlPhase,
        "beforeHead": BeforeHeadPhase,
        "inHead": InHeadPhase,
        "inHeadNoscript": InHeadNoscriptPhase,
        "afterHead": AfterHeadPhase,
        "inBody": InBodyPhase,
        "text": TextPhase,
        "inTable": InTablePhase,
        "inTableText": InTableTextPhase,
        "inCaption": InCaptionPhase,
        "inColumnGroup": InColumnGroupPhase,
        "inTableBody": InTableBodyPhase,
        "inRow": InRowPhase,
        "inCell": InCellPhase,
        "inSelect": InSelectPhase,
        "inSelectInTable": InSelectInTablePhase,
        "inForeignContent": InForeignContentPhase,
        "afterBody": AfterBodyPhase,
        "inFrameset": InFramesetPhase,
        "afterFrameset": AfterFramesetPhase,
        "afterAfterBody": AfterAfterBodyPhase,
        "afterAfterFrameset": AfterAfterFramesetPhase,
        # XXX after after frameset
    }


def adjust_attributes(token, replacements):
    needs_adjustment = viewkeys(token['data']) & viewkeys(replacements)
    if needs_adjustment:
        token['data'] = type(token['data'])((replacements.get(k, k), v)
                                            for k, v in token['data'].items())


def impliedTagToken(name, type="EndTag", attributes=None,
                    selfClosing=False):
    if attributes is None:
        attributes = {}
    return {"type": tokenTypes[type], "name": name, "data": attributes,
            "selfClosing": selfClosing}


class ParseError(Exception):
    """Error in parsed document"""
    pass
site-packages/pip/_vendor/html5lib/_tokenizer.py000064400000226360150564745010015773 0ustar00from __future__ import absolute_import, division, unicode_literals

from pip._vendor.six import unichr as chr

from collections import deque, OrderedDict
from sys import version_info

from .constants import spaceCharacters
from .constants import entities
from .constants import asciiLetters, asciiUpper2Lower
from .constants import digits, hexDigits, EOF
from .constants import tokenTypes, tagTokenTypes
from .constants import replacementCharacters

from ._inputstream import HTMLInputStream

from ._trie import Trie

entitiesTrie = Trie(entities)

if version_info >= (3, 7):
    attributeMap = dict
else:
    attributeMap = OrderedDict


class HTMLTokenizer(object):
    """ This class takes care of tokenizing HTML.

    * self.currentToken
      Holds the token that is currently being processed.

    * self.state
      Holds a reference to the method to be invoked... XXX

    * self.stream
      Points to HTMLInputStream object.
    """

    def __init__(self, stream, parser=None, **kwargs):

        self.stream = HTMLInputStream(stream, **kwargs)
        self.parser = parser

        # Setup the initial tokenizer state
        self.escapeFlag = False
        self.lastFourChars = []
        self.state = self.dataState
        self.escape = False

        # The current token being created
        self.currentToken = None
        super(HTMLTokenizer, self).__init__()

    def __iter__(self):
        """ This is where the magic happens.

        We do our usually processing through the states and when we have a token
        to return we yield the token which pauses processing until the next token
        is requested.
        """
        self.tokenQueue = deque([])
        # Start processing. When EOF is reached self.state will return False
        # instead of True and the loop will terminate.
        while self.state():
            while self.stream.errors:
                yield {"type": tokenTypes["ParseError"], "data": self.stream.errors.pop(0)}
            while self.tokenQueue:
                yield self.tokenQueue.popleft()

    def consumeNumberEntity(self, isHex):
        """This function returns either U+FFFD or the character based on the
        decimal or hexadecimal representation. It also discards ";" if present.
        If not present self.tokenQueue.append({"type": tokenTypes["ParseError"]}) is invoked.
        """

        allowed = digits
        radix = 10
        if isHex:
            allowed = hexDigits
            radix = 16

        charStack = []

        # Consume all the characters that are in range while making sure we
        # don't hit an EOF.
        c = self.stream.char()
        while c in allowed and c is not EOF:
            charStack.append(c)
            c = self.stream.char()

        # Convert the set of characters consumed to an int.
        charAsInt = int("".join(charStack), radix)

        # Certain characters get replaced with others
        if charAsInt in replacementCharacters:
            char = replacementCharacters[charAsInt]
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "illegal-codepoint-for-numeric-entity",
                                    "datavars": {"charAsInt": charAsInt}})
        elif ((0xD800 <= charAsInt <= 0xDFFF) or
              (charAsInt > 0x10FFFF)):
            char = "\uFFFD"
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "illegal-codepoint-for-numeric-entity",
                                    "datavars": {"charAsInt": charAsInt}})
        else:
            # Should speed up this check somehow (e.g. move the set to a constant)
            if ((0x0001 <= charAsInt <= 0x0008) or
                (0x000E <= charAsInt <= 0x001F) or
                (0x007F <= charAsInt <= 0x009F) or
                (0xFDD0 <= charAsInt <= 0xFDEF) or
                charAsInt in frozenset([0x000B, 0xFFFE, 0xFFFF, 0x1FFFE,
                                        0x1FFFF, 0x2FFFE, 0x2FFFF, 0x3FFFE,
                                        0x3FFFF, 0x4FFFE, 0x4FFFF, 0x5FFFE,
                                        0x5FFFF, 0x6FFFE, 0x6FFFF, 0x7FFFE,
                                        0x7FFFF, 0x8FFFE, 0x8FFFF, 0x9FFFE,
                                        0x9FFFF, 0xAFFFE, 0xAFFFF, 0xBFFFE,
                                        0xBFFFF, 0xCFFFE, 0xCFFFF, 0xDFFFE,
                                        0xDFFFF, 0xEFFFE, 0xEFFFF, 0xFFFFE,
                                        0xFFFFF, 0x10FFFE, 0x10FFFF])):
                self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                        "data":
                                        "illegal-codepoint-for-numeric-entity",
                                        "datavars": {"charAsInt": charAsInt}})
            try:
                # Try/except needed as UCS-2 Python builds' unichar only works
                # within the BMP.
                char = chr(charAsInt)
            except ValueError:
                v = charAsInt - 0x10000
                char = chr(0xD800 | (v >> 10)) + chr(0xDC00 | (v & 0x3FF))

        # Discard the ; if present. Otherwise, put it back on the queue and
        # invoke parseError on parser.
        if c != ";":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "numeric-entity-without-semicolon"})
            self.stream.unget(c)

        return char

    def consumeEntity(self, allowedChar=None, fromAttribute=False):
        # Initialise to the default output for when no entity is matched
        output = "&"

        charStack = [self.stream.char()]
        if (charStack[0] in spaceCharacters or charStack[0] in (EOF, "<", "&") or
                (allowedChar is not None and allowedChar == charStack[0])):
            self.stream.unget(charStack[0])

        elif charStack[0] == "#":
            # Read the next character to see if it's hex or decimal
            hex = False
            charStack.append(self.stream.char())
            if charStack[-1] in ("x", "X"):
                hex = True
                charStack.append(self.stream.char())

            # charStack[-1] should be the first digit
            if (hex and charStack[-1] in hexDigits) \
                    or (not hex and charStack[-1] in digits):
                # At least one digit found, so consume the whole number
                self.stream.unget(charStack[-1])
                output = self.consumeNumberEntity(hex)
            else:
                # No digits found
                self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                        "data": "expected-numeric-entity"})
                self.stream.unget(charStack.pop())
                output = "&" + "".join(charStack)

        else:
            # At this point in the process might have named entity. Entities
            # are stored in the global variable "entities".
            #
            # Consume characters and compare to these to a substring of the
            # entity names in the list until the substring no longer matches.
            while (charStack[-1] is not EOF):
                if not entitiesTrie.has_keys_with_prefix("".join(charStack)):
                    break
                charStack.append(self.stream.char())

            # At this point we have a string that starts with some characters
            # that may match an entity
            # Try to find the longest entity the string will match to take care
            # of &noti for instance.
            try:
                entityName = entitiesTrie.longest_prefix("".join(charStack[:-1]))
                entityLength = len(entityName)
            except KeyError:
                entityName = None

            if entityName is not None:
                if entityName[-1] != ";":
                    self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                            "named-entity-without-semicolon"})
                if (entityName[-1] != ";" and fromAttribute and
                    (charStack[entityLength] in asciiLetters or
                     charStack[entityLength] in digits or
                     charStack[entityLength] == "=")):
                    self.stream.unget(charStack.pop())
                    output = "&" + "".join(charStack)
                else:
                    output = entities[entityName]
                    self.stream.unget(charStack.pop())
                    output += "".join(charStack[entityLength:])
            else:
                self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                        "expected-named-entity"})
                self.stream.unget(charStack.pop())
                output = "&" + "".join(charStack)

        if fromAttribute:
            self.currentToken["data"][-1][1] += output
        else:
            if output in spaceCharacters:
                tokenType = "SpaceCharacters"
            else:
                tokenType = "Characters"
            self.tokenQueue.append({"type": tokenTypes[tokenType], "data": output})

    def processEntityInAttribute(self, allowedChar):
        """This method replaces the need for "entityInAttributeValueState".
        """
        self.consumeEntity(allowedChar=allowedChar, fromAttribute=True)

    def emitCurrentToken(self):
        """This method is a generic handler for emitting the tags. It also sets
        the state to "data" because that's what's needed after a token has been
        emitted.
        """
        token = self.currentToken
        # Add token to the queue to be yielded
        if (token["type"] in tagTokenTypes):
            token["name"] = token["name"].translate(asciiUpper2Lower)
            if token["type"] == tokenTypes["StartTag"]:
                raw = token["data"]
                data = attributeMap(raw)
                if len(raw) > len(data):
                    # we had some duplicated attribute, fix so first wins
                    data.update(raw[::-1])
                token["data"] = data

            if token["type"] == tokenTypes["EndTag"]:
                if token["data"]:
                    self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                            "data": "attributes-in-end-tag"})
                if token["selfClosing"]:
                    self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                            "data": "self-closing-flag-on-end-tag"})
        self.tokenQueue.append(token)
        self.state = self.dataState

    # Below are the various tokenizer states worked out.
    def dataState(self):
        data = self.stream.char()
        if data == "&":
            self.state = self.entityDataState
        elif data == "<":
            self.state = self.tagOpenState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\u0000"})
        elif data is EOF:
            # Tokenization ends.
            return False
        elif data in spaceCharacters:
            # Directly after emitting a token you switch back to the "data
            # state". At that point spaceCharacters are important so they are
            # emitted separately.
            self.tokenQueue.append({"type": tokenTypes["SpaceCharacters"], "data":
                                    data + self.stream.charsUntil(spaceCharacters, True)})
            # No need to update lastFourChars here, since the first space will
            # have already been appended to lastFourChars and will have broken
            # any <!-- or --> sequences
        else:
            chars = self.stream.charsUntil(("&", "<", "\u0000"))
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + chars})
        return True

    def entityDataState(self):
        self.consumeEntity()
        self.state = self.dataState
        return True

    def rcdataState(self):
        data = self.stream.char()
        if data == "&":
            self.state = self.characterReferenceInRcdata
        elif data == "<":
            self.state = self.rcdataLessThanSignState
        elif data == EOF:
            # Tokenization ends.
            return False
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        elif data in spaceCharacters:
            # Directly after emitting a token you switch back to the "data
            # state". At that point spaceCharacters are important so they are
            # emitted separately.
            self.tokenQueue.append({"type": tokenTypes["SpaceCharacters"], "data":
                                    data + self.stream.charsUntil(spaceCharacters, True)})
            # No need to update lastFourChars here, since the first space will
            # have already been appended to lastFourChars and will have broken
            # any <!-- or --> sequences
        else:
            chars = self.stream.charsUntil(("&", "<", "\u0000"))
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + chars})
        return True

    def characterReferenceInRcdata(self):
        self.consumeEntity()
        self.state = self.rcdataState
        return True

    def rawtextState(self):
        data = self.stream.char()
        if data == "<":
            self.state = self.rawtextLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        elif data == EOF:
            # Tokenization ends.
            return False
        else:
            chars = self.stream.charsUntil(("<", "\u0000"))
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + chars})
        return True

    def scriptDataState(self):
        data = self.stream.char()
        if data == "<":
            self.state = self.scriptDataLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        elif data == EOF:
            # Tokenization ends.
            return False
        else:
            chars = self.stream.charsUntil(("<", "\u0000"))
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + chars})
        return True

    def plaintextState(self):
        data = self.stream.char()
        if data == EOF:
            # Tokenization ends.
            return False
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + self.stream.charsUntil("\u0000")})
        return True

    def tagOpenState(self):
        data = self.stream.char()
        if data == "!":
            self.state = self.markupDeclarationOpenState
        elif data == "/":
            self.state = self.closeTagOpenState
        elif data in asciiLetters:
            self.currentToken = {"type": tokenTypes["StartTag"],
                                 "name": data, "data": [],
                                 "selfClosing": False,
                                 "selfClosingAcknowledged": False}
            self.state = self.tagNameState
        elif data == ">":
            # XXX In theory it could be something besides a tag name. But
            # do we really care?
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-tag-name-but-got-right-bracket"})
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<>"})
            self.state = self.dataState
        elif data == "?":
            # XXX In theory it could be something besides a tag name. But
            # do we really care?
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-tag-name-but-got-question-mark"})
            self.stream.unget(data)
            self.state = self.bogusCommentState
        else:
            # XXX
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-tag-name"})
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.stream.unget(data)
            self.state = self.dataState
        return True

    def closeTagOpenState(self):
        data = self.stream.char()
        if data in asciiLetters:
            self.currentToken = {"type": tokenTypes["EndTag"], "name": data,
                                 "data": [], "selfClosing": False}
            self.state = self.tagNameState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-closing-tag-but-got-right-bracket"})
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-closing-tag-but-got-eof"})
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "</"})
            self.state = self.dataState
        else:
            # XXX data can be _'_...
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-closing-tag-but-got-char",
                                    "datavars": {"data": data}})
            self.stream.unget(data)
            self.state = self.bogusCommentState
        return True

    def tagNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeAttributeNameState
        elif data == ">":
            self.emitCurrentToken()
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-tag-name"})
            self.state = self.dataState
        elif data == "/":
            self.state = self.selfClosingStartTagState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["name"] += "\uFFFD"
        else:
            self.currentToken["name"] += data
            # (Don't use charsUntil here, because tag names are
            # very short and it's faster to not do anything fancy)
        return True

    def rcdataLessThanSignState(self):
        data = self.stream.char()
        if data == "/":
            self.temporaryBuffer = ""
            self.state = self.rcdataEndTagOpenState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.stream.unget(data)
            self.state = self.rcdataState
        return True

    def rcdataEndTagOpenState(self):
        data = self.stream.char()
        if data in asciiLetters:
            self.temporaryBuffer += data
            self.state = self.rcdataEndTagNameState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "</"})
            self.stream.unget(data)
            self.state = self.rcdataState
        return True

    def rcdataEndTagNameState(self):
        appropriate = self.currentToken and self.currentToken["name"].lower() == self.temporaryBuffer.lower()
        data = self.stream.char()
        if data in spaceCharacters and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.beforeAttributeNameState
        elif data == "/" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.selfClosingStartTagState
        elif data == ">" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.emitCurrentToken()
            self.state = self.dataState
        elif data in asciiLetters:
            self.temporaryBuffer += data
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "</" + self.temporaryBuffer})
            self.stream.unget(data)
            self.state = self.rcdataState
        return True

    def rawtextLessThanSignState(self):
        data = self.stream.char()
        if data == "/":
            self.temporaryBuffer = ""
            self.state = self.rawtextEndTagOpenState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.stream.unget(data)
            self.state = self.rawtextState
        return True

    def rawtextEndTagOpenState(self):
        data = self.stream.char()
        if data in asciiLetters:
            self.temporaryBuffer += data
            self.state = self.rawtextEndTagNameState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "</"})
            self.stream.unget(data)
            self.state = self.rawtextState
        return True

    def rawtextEndTagNameState(self):
        appropriate = self.currentToken and self.currentToken["name"].lower() == self.temporaryBuffer.lower()
        data = self.stream.char()
        if data in spaceCharacters and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.beforeAttributeNameState
        elif data == "/" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.selfClosingStartTagState
        elif data == ">" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.emitCurrentToken()
            self.state = self.dataState
        elif data in asciiLetters:
            self.temporaryBuffer += data
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "</" + self.temporaryBuffer})
            self.stream.unget(data)
            self.state = self.rawtextState
        return True

    def scriptDataLessThanSignState(self):
        data = self.stream.char()
        if data == "/":
            self.temporaryBuffer = ""
            self.state = self.scriptDataEndTagOpenState
        elif data == "!":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<!"})
            self.state = self.scriptDataEscapeStartState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.stream.unget(data)
            self.state = self.scriptDataState
        return True

    def scriptDataEndTagOpenState(self):
        data = self.stream.char()
        if data in asciiLetters:
            self.temporaryBuffer += data
            self.state = self.scriptDataEndTagNameState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "</"})
            self.stream.unget(data)
            self.state = self.scriptDataState
        return True

    def scriptDataEndTagNameState(self):
        appropriate = self.currentToken and self.currentToken["name"].lower() == self.temporaryBuffer.lower()
        data = self.stream.char()
        if data in spaceCharacters and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.beforeAttributeNameState
        elif data == "/" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.selfClosingStartTagState
        elif data == ">" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.emitCurrentToken()
            self.state = self.dataState
        elif data in asciiLetters:
            self.temporaryBuffer += data
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "</" + self.temporaryBuffer})
            self.stream.unget(data)
            self.state = self.scriptDataState
        return True

    def scriptDataEscapeStartState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataEscapeStartDashState
        else:
            self.stream.unget(data)
            self.state = self.scriptDataState
        return True

    def scriptDataEscapeStartDashState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataEscapedDashDashState
        else:
            self.stream.unget(data)
            self.state = self.scriptDataState
        return True

    def scriptDataEscapedState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataEscapedDashState
        elif data == "<":
            self.state = self.scriptDataEscapedLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        elif data == EOF:
            self.state = self.dataState
        else:
            chars = self.stream.charsUntil(("<", "-", "\u0000"))
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data":
                                    data + chars})
        return True

    def scriptDataEscapedDashState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataEscapedDashDashState
        elif data == "<":
            self.state = self.scriptDataEscapedLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
            self.state = self.scriptDataEscapedState
        elif data == EOF:
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataEscapedDashDashState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
        elif data == "<":
            self.state = self.scriptDataEscapedLessThanSignState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": ">"})
            self.state = self.scriptDataState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
            self.state = self.scriptDataEscapedState
        elif data == EOF:
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataEscapedLessThanSignState(self):
        data = self.stream.char()
        if data == "/":
            self.temporaryBuffer = ""
            self.state = self.scriptDataEscapedEndTagOpenState
        elif data in asciiLetters:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<" + data})
            self.temporaryBuffer = data
            self.state = self.scriptDataDoubleEscapeStartState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.stream.unget(data)
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataEscapedEndTagOpenState(self):
        data = self.stream.char()
        if data in asciiLetters:
            self.temporaryBuffer = data
            self.state = self.scriptDataEscapedEndTagNameState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "</"})
            self.stream.unget(data)
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataEscapedEndTagNameState(self):
        appropriate = self.currentToken and self.currentToken["name"].lower() == self.temporaryBuffer.lower()
        data = self.stream.char()
        if data in spaceCharacters and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.beforeAttributeNameState
        elif data == "/" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.state = self.selfClosingStartTagState
        elif data == ">" and appropriate:
            self.currentToken = {"type": tokenTypes["EndTag"],
                                 "name": self.temporaryBuffer,
                                 "data": [], "selfClosing": False}
            self.emitCurrentToken()
            self.state = self.dataState
        elif data in asciiLetters:
            self.temporaryBuffer += data
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "</" + self.temporaryBuffer})
            self.stream.unget(data)
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataDoubleEscapeStartState(self):
        data = self.stream.char()
        if data in (spaceCharacters | frozenset(("/", ">"))):
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            if self.temporaryBuffer.lower() == "script":
                self.state = self.scriptDataDoubleEscapedState
            else:
                self.state = self.scriptDataEscapedState
        elif data in asciiLetters:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.temporaryBuffer += data
        else:
            self.stream.unget(data)
            self.state = self.scriptDataEscapedState
        return True

    def scriptDataDoubleEscapedState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataDoubleEscapedDashState
        elif data == "<":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.state = self.scriptDataDoubleEscapedLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
        elif data == EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-script-in-script"})
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
        return True

    def scriptDataDoubleEscapedDashState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
            self.state = self.scriptDataDoubleEscapedDashDashState
        elif data == "<":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.state = self.scriptDataDoubleEscapedLessThanSignState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
            self.state = self.scriptDataDoubleEscapedState
        elif data == EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-script-in-script"})
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.state = self.scriptDataDoubleEscapedState
        return True

    def scriptDataDoubleEscapedDashDashState(self):
        data = self.stream.char()
        if data == "-":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "-"})
        elif data == "<":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "<"})
            self.state = self.scriptDataDoubleEscapedLessThanSignState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": ">"})
            self.state = self.scriptDataState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": "\uFFFD"})
            self.state = self.scriptDataDoubleEscapedState
        elif data == EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-script-in-script"})
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.state = self.scriptDataDoubleEscapedState
        return True

    def scriptDataDoubleEscapedLessThanSignState(self):
        data = self.stream.char()
        if data == "/":
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": "/"})
            self.temporaryBuffer = ""
            self.state = self.scriptDataDoubleEscapeEndState
        else:
            self.stream.unget(data)
            self.state = self.scriptDataDoubleEscapedState
        return True

    def scriptDataDoubleEscapeEndState(self):
        data = self.stream.char()
        if data in (spaceCharacters | frozenset(("/", ">"))):
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            if self.temporaryBuffer.lower() == "script":
                self.state = self.scriptDataEscapedState
            else:
                self.state = self.scriptDataDoubleEscapedState
        elif data in asciiLetters:
            self.tokenQueue.append({"type": tokenTypes["Characters"], "data": data})
            self.temporaryBuffer += data
        else:
            self.stream.unget(data)
            self.state = self.scriptDataDoubleEscapedState
        return True

    def beforeAttributeNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.stream.charsUntil(spaceCharacters, True)
        elif data in asciiLetters:
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        elif data == ">":
            self.emitCurrentToken()
        elif data == "/":
            self.state = self.selfClosingStartTagState
        elif data in ("'", '"', "=", "<"):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "invalid-character-in-attribute-name"})
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"].append(["\uFFFD", ""])
            self.state = self.attributeNameState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-attribute-name-but-got-eof"})
            self.state = self.dataState
        else:
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        return True

    def attributeNameState(self):
        data = self.stream.char()
        leavingThisState = True
        emitToken = False
        if data == "=":
            self.state = self.beforeAttributeValueState
        elif data in asciiLetters:
            self.currentToken["data"][-1][0] += data +\
                self.stream.charsUntil(asciiLetters, True)
            leavingThisState = False
        elif data == ">":
            # XXX If we emit here the attributes are converted to a dict
            # without being checked and when the code below runs we error
            # because data is a dict not a list
            emitToken = True
        elif data in spaceCharacters:
            self.state = self.afterAttributeNameState
        elif data == "/":
            self.state = self.selfClosingStartTagState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"][-1][0] += "\uFFFD"
            leavingThisState = False
        elif data in ("'", '"', "<"):
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data":
                                    "invalid-character-in-attribute-name"})
            self.currentToken["data"][-1][0] += data
            leavingThisState = False
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "eof-in-attribute-name"})
            self.state = self.dataState
        else:
            self.currentToken["data"][-1][0] += data
            leavingThisState = False

        if leavingThisState:
            # Attributes are not dropped at this stage. That happens when the
            # start tag token is emitted so values can still be safely appended
            # to attributes, but we do want to report the parse error in time.
            self.currentToken["data"][-1][0] = (
                self.currentToken["data"][-1][0].translate(asciiUpper2Lower))
            for name, _ in self.currentToken["data"][:-1]:
                if self.currentToken["data"][-1][0] == name:
                    self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                            "duplicate-attribute"})
                    break
            # XXX Fix for above XXX
            if emitToken:
                self.emitCurrentToken()
        return True

    def afterAttributeNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.stream.charsUntil(spaceCharacters, True)
        elif data == "=":
            self.state = self.beforeAttributeValueState
        elif data == ">":
            self.emitCurrentToken()
        elif data in asciiLetters:
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        elif data == "/":
            self.state = self.selfClosingStartTagState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"].append(["\uFFFD", ""])
            self.state = self.attributeNameState
        elif data in ("'", '"', "<"):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "invalid-character-after-attribute-name"})
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-end-of-tag-but-got-eof"})
            self.state = self.dataState
        else:
            self.currentToken["data"].append([data, ""])
            self.state = self.attributeNameState
        return True

    def beforeAttributeValueState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.stream.charsUntil(spaceCharacters, True)
        elif data == "\"":
            self.state = self.attributeValueDoubleQuotedState
        elif data == "&":
            self.state = self.attributeValueUnQuotedState
            self.stream.unget(data)
        elif data == "'":
            self.state = self.attributeValueSingleQuotedState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-attribute-value-but-got-right-bracket"})
            self.emitCurrentToken()
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"][-1][1] += "\uFFFD"
            self.state = self.attributeValueUnQuotedState
        elif data in ("=", "<", "`"):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "equals-in-unquoted-attribute-value"})
            self.currentToken["data"][-1][1] += data
            self.state = self.attributeValueUnQuotedState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-attribute-value-but-got-eof"})
            self.state = self.dataState
        else:
            self.currentToken["data"][-1][1] += data
            self.state = self.attributeValueUnQuotedState
        return True

    def attributeValueDoubleQuotedState(self):
        data = self.stream.char()
        if data == "\"":
            self.state = self.afterAttributeValueState
        elif data == "&":
            self.processEntityInAttribute('"')
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"][-1][1] += "\uFFFD"
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-attribute-value-double-quote"})
            self.state = self.dataState
        else:
            self.currentToken["data"][-1][1] += data +\
                self.stream.charsUntil(("\"", "&", "\u0000"))
        return True

    def attributeValueSingleQuotedState(self):
        data = self.stream.char()
        if data == "'":
            self.state = self.afterAttributeValueState
        elif data == "&":
            self.processEntityInAttribute("'")
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"][-1][1] += "\uFFFD"
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-attribute-value-single-quote"})
            self.state = self.dataState
        else:
            self.currentToken["data"][-1][1] += data +\
                self.stream.charsUntil(("'", "&", "\u0000"))
        return True

    def attributeValueUnQuotedState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeAttributeNameState
        elif data == "&":
            self.processEntityInAttribute(">")
        elif data == ">":
            self.emitCurrentToken()
        elif data in ('"', "'", "=", "<", "`"):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-character-in-unquoted-attribute-value"})
            self.currentToken["data"][-1][1] += data
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"][-1][1] += "\uFFFD"
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-attribute-value-no-quotes"})
            self.state = self.dataState
        else:
            self.currentToken["data"][-1][1] += data + self.stream.charsUntil(
                frozenset(("&", ">", '"', "'", "=", "<", "`", "\u0000")) | spaceCharacters)
        return True

    def afterAttributeValueState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeAttributeNameState
        elif data == ">":
            self.emitCurrentToken()
        elif data == "/":
            self.state = self.selfClosingStartTagState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-EOF-after-attribute-value"})
            self.stream.unget(data)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-character-after-attribute-value"})
            self.stream.unget(data)
            self.state = self.beforeAttributeNameState
        return True

    def selfClosingStartTagState(self):
        data = self.stream.char()
        if data == ">":
            self.currentToken["selfClosing"] = True
            self.emitCurrentToken()
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data":
                                    "unexpected-EOF-after-solidus-in-tag"})
            self.stream.unget(data)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-character-after-solidus-in-tag"})
            self.stream.unget(data)
            self.state = self.beforeAttributeNameState
        return True

    def bogusCommentState(self):
        # Make a new comment token and give it as value all the characters
        # until the first > or EOF (charsUntil checks for EOF automatically)
        # and emit it.
        data = self.stream.charsUntil(">")
        data = data.replace("\u0000", "\uFFFD")
        self.tokenQueue.append(
            {"type": tokenTypes["Comment"], "data": data})

        # Eat the character directly after the bogus comment which is either a
        # ">" or an EOF.
        self.stream.char()
        self.state = self.dataState
        return True

    def markupDeclarationOpenState(self):
        charStack = [self.stream.char()]
        if charStack[-1] == "-":
            charStack.append(self.stream.char())
            if charStack[-1] == "-":
                self.currentToken = {"type": tokenTypes["Comment"], "data": ""}
                self.state = self.commentStartState
                return True
        elif charStack[-1] in ('d', 'D'):
            matched = True
            for expected in (('o', 'O'), ('c', 'C'), ('t', 'T'),
                             ('y', 'Y'), ('p', 'P'), ('e', 'E')):
                charStack.append(self.stream.char())
                if charStack[-1] not in expected:
                    matched = False
                    break
            if matched:
                self.currentToken = {"type": tokenTypes["Doctype"],
                                     "name": "",
                                     "publicId": None, "systemId": None,
                                     "correct": True}
                self.state = self.doctypeState
                return True
        elif (charStack[-1] == "[" and
              self.parser is not None and
              self.parser.tree.openElements and
              self.parser.tree.openElements[-1].namespace != self.parser.tree.defaultNamespace):
            matched = True
            for expected in ["C", "D", "A", "T", "A", "["]:
                charStack.append(self.stream.char())
                if charStack[-1] != expected:
                    matched = False
                    break
            if matched:
                self.state = self.cdataSectionState
                return True

        self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                "expected-dashes-or-doctype"})

        while charStack:
            self.stream.unget(charStack.pop())
        self.state = self.bogusCommentState
        return True

    def commentStartState(self):
        data = self.stream.char()
        if data == "-":
            self.state = self.commentStartDashState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "incorrect-comment"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-comment"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["data"] += data
            self.state = self.commentState
        return True

    def commentStartDashState(self):
        data = self.stream.char()
        if data == "-":
            self.state = self.commentEndState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "-\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "incorrect-comment"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-comment"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["data"] += "-" + data
            self.state = self.commentState
        return True

    def commentState(self):
        data = self.stream.char()
        if data == "-":
            self.state = self.commentEndDashState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "\uFFFD"
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "eof-in-comment"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["data"] += data + \
                self.stream.charsUntil(("-", "\u0000"))
        return True

    def commentEndDashState(self):
        data = self.stream.char()
        if data == "-":
            self.state = self.commentEndState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "-\uFFFD"
            self.state = self.commentState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-comment-end-dash"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["data"] += "-" + data
            self.state = self.commentState
        return True

    def commentEndState(self):
        data = self.stream.char()
        if data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "--\uFFFD"
            self.state = self.commentState
        elif data == "!":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-bang-after-double-dash-in-comment"})
            self.state = self.commentEndBangState
        elif data == "-":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-dash-after-double-dash-in-comment"})
            self.currentToken["data"] += data
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-comment-double-dash"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            # XXX
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-comment"})
            self.currentToken["data"] += "--" + data
            self.state = self.commentState
        return True

    def commentEndBangState(self):
        data = self.stream.char()
        if data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == "-":
            self.currentToken["data"] += "--!"
            self.state = self.commentEndDashState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["data"] += "--!\uFFFD"
            self.state = self.commentState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-comment-end-bang-state"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["data"] += "--!" + data
            self.state = self.commentState
        return True

    def doctypeState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeDoctypeNameState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-doctype-name-but-got-eof"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "need-space-after-doctype"})
            self.stream.unget(data)
            self.state = self.beforeDoctypeNameState
        return True

    def beforeDoctypeNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-doctype-name-but-got-right-bracket"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["name"] = "\uFFFD"
            self.state = self.doctypeNameState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-doctype-name-but-got-eof"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["name"] = data
            self.state = self.doctypeNameState
        return True

    def doctypeNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower)
            self.state = self.afterDoctypeNameState
        elif data == ">":
            self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower)
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["name"] += "\uFFFD"
            self.state = self.doctypeNameState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype-name"})
            self.currentToken["correct"] = False
            self.currentToken["name"] = self.currentToken["name"].translate(asciiUpper2Lower)
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["name"] += data
        return True

    def afterDoctypeNameState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.currentToken["correct"] = False
            self.stream.unget(data)
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            if data in ("p", "P"):
                matched = True
                for expected in (("u", "U"), ("b", "B"), ("l", "L"),
                                 ("i", "I"), ("c", "C")):
                    data = self.stream.char()
                    if data not in expected:
                        matched = False
                        break
                if matched:
                    self.state = self.afterDoctypePublicKeywordState
                    return True
            elif data in ("s", "S"):
                matched = True
                for expected in (("y", "Y"), ("s", "S"), ("t", "T"),
                                 ("e", "E"), ("m", "M")):
                    data = self.stream.char()
                    if data not in expected:
                        matched = False
                        break
                if matched:
                    self.state = self.afterDoctypeSystemKeywordState
                    return True

            # All the characters read before the current 'data' will be
            # [a-zA-Z], so they're garbage in the bogus doctype and can be
            # discarded; only the latest character might be '>' or EOF
            # and needs to be ungetted
            self.stream.unget(data)
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "expected-space-or-right-bracket-in-doctype", "datavars":
                                    {"data": data}})
            self.currentToken["correct"] = False
            self.state = self.bogusDoctypeState

        return True

    def afterDoctypePublicKeywordState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeDoctypePublicIdentifierState
        elif data in ("'", '"'):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.stream.unget(data)
            self.state = self.beforeDoctypePublicIdentifierState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.stream.unget(data)
            self.state = self.beforeDoctypePublicIdentifierState
        return True

    def beforeDoctypePublicIdentifierState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == "\"":
            self.currentToken["publicId"] = ""
            self.state = self.doctypePublicIdentifierDoubleQuotedState
        elif data == "'":
            self.currentToken["publicId"] = ""
            self.state = self.doctypePublicIdentifierSingleQuotedState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-end-of-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["correct"] = False
            self.state = self.bogusDoctypeState
        return True

    def doctypePublicIdentifierDoubleQuotedState(self):
        data = self.stream.char()
        if data == "\"":
            self.state = self.afterDoctypePublicIdentifierState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["publicId"] += "\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-end-of-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["publicId"] += data
        return True

    def doctypePublicIdentifierSingleQuotedState(self):
        data = self.stream.char()
        if data == "'":
            self.state = self.afterDoctypePublicIdentifierState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["publicId"] += "\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-end-of-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["publicId"] += data
        return True

    def afterDoctypePublicIdentifierState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.betweenDoctypePublicAndSystemIdentifiersState
        elif data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == '"':
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierDoubleQuotedState
        elif data == "'":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierSingleQuotedState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["correct"] = False
            self.state = self.bogusDoctypeState
        return True

    def betweenDoctypePublicAndSystemIdentifiersState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data == '"':
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierDoubleQuotedState
        elif data == "'":
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierSingleQuotedState
        elif data == EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["correct"] = False
            self.state = self.bogusDoctypeState
        return True

    def afterDoctypeSystemKeywordState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            self.state = self.beforeDoctypeSystemIdentifierState
        elif data in ("'", '"'):
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.stream.unget(data)
            self.state = self.beforeDoctypeSystemIdentifierState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.stream.unget(data)
            self.state = self.beforeDoctypeSystemIdentifierState
        return True

    def beforeDoctypeSystemIdentifierState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == "\"":
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierDoubleQuotedState
        elif data == "'":
            self.currentToken["systemId"] = ""
            self.state = self.doctypeSystemIdentifierSingleQuotedState
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.currentToken["correct"] = False
            self.state = self.bogusDoctypeState
        return True

    def doctypeSystemIdentifierDoubleQuotedState(self):
        data = self.stream.char()
        if data == "\"":
            self.state = self.afterDoctypeSystemIdentifierState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["systemId"] += "\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-end-of-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["systemId"] += data
        return True

    def doctypeSystemIdentifierSingleQuotedState(self):
        data = self.stream.char()
        if data == "'":
            self.state = self.afterDoctypeSystemIdentifierState
        elif data == "\u0000":
            self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                    "data": "invalid-codepoint"})
            self.currentToken["systemId"] += "\uFFFD"
        elif data == ">":
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-end-of-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.currentToken["systemId"] += data
        return True

    def afterDoctypeSystemIdentifierState(self):
        data = self.stream.char()
        if data in spaceCharacters:
            pass
        elif data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "eof-in-doctype"})
            self.currentToken["correct"] = False
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            self.tokenQueue.append({"type": tokenTypes["ParseError"], "data":
                                    "unexpected-char-in-doctype"})
            self.state = self.bogusDoctypeState
        return True

    def bogusDoctypeState(self):
        data = self.stream.char()
        if data == ">":
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        elif data is EOF:
            # XXX EMIT
            self.stream.unget(data)
            self.tokenQueue.append(self.currentToken)
            self.state = self.dataState
        else:
            pass
        return True

    def cdataSectionState(self):
        data = []
        while True:
            data.append(self.stream.charsUntil("]"))
            data.append(self.stream.charsUntil(">"))
            char = self.stream.char()
            if char == EOF:
                break
            else:
                assert char == ">"
                if data[-1][-2:] == "]]":
                    data[-1] = data[-1][:-2]
                    break
                else:
                    data.append(char)

        data = "".join(data)  # pylint:disable=redefined-variable-type
        # Deal with null here rather than in the parser
        nullCount = data.count("\u0000")
        if nullCount > 0:
            for _ in range(nullCount):
                self.tokenQueue.append({"type": tokenTypes["ParseError"],
                                        "data": "invalid-codepoint"})
            data = data.replace("\u0000", "\uFFFD")
        if data:
            self.tokenQueue.append({"type": tokenTypes["Characters"],
                                    "data": data})
        self.state = self.dataState
        return True
site-packages/pip/_vendor/pyparsing.py000064400001025762150564745010014122 0ustar00# -*- coding: utf-8 -*-
# module pyparsing.py
#
# Copyright (c) 2003-2019  Paul T. McGuire
#
# Permission is hereby granted, free of charge, to any person obtaining
# a copy of this software and associated documentation files (the
# "Software"), to deal in the Software without restriction, including
# without limitation the rights to use, copy, modify, merge, publish,
# distribute, sublicense, and/or sell copies of the Software, and to
# permit persons to whom the Software is furnished to do so, subject to
# the following conditions:
#
# The above copyright notice and this permission notice shall be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
# EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
# IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY
# CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT,
# TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE
# SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
#

__doc__ = \
"""
pyparsing module - Classes and methods to define and execute parsing grammars
=============================================================================

The pyparsing module is an alternative approach to creating and
executing simple grammars, vs. the traditional lex/yacc approach, or the
use of regular expressions.  With pyparsing, you don't need to learn
a new syntax for defining grammars or matching expressions - the parsing
module provides a library of classes that you use to construct the
grammar directly in Python.

Here is a program to parse "Hello, World!" (or any greeting of the form
``"<salutation>, <addressee>!"``), built up using :class:`Word`,
:class:`Literal`, and :class:`And` elements
(the :class:`'+'<ParserElement.__add__>` operators create :class:`And` expressions,
and the strings are auto-converted to :class:`Literal` expressions)::

    from pip._vendor.pyparsing import Word, alphas

    # define grammar of a greeting
    greet = Word(alphas) + "," + Word(alphas) + "!"

    hello = "Hello, World!"
    print (hello, "->", greet.parseString(hello))

The program outputs the following::

    Hello, World! -> ['Hello', ',', 'World', '!']

The Python representation of the grammar is quite readable, owing to the
self-explanatory class names, and the use of '+', '|' and '^' operators.

The :class:`ParseResults` object returned from
:class:`ParserElement.parseString` can be
accessed as a nested list, a dictionary, or an object with named
attributes.

The pyparsing module handles some of the problems that are typically
vexing when writing text parsers:

  - extra or missing whitespace (the above program will also handle
    "Hello,World!", "Hello  ,  World  !", etc.)
  - quoted strings
  - embedded comments


Getting Started -
-----------------
Visit the classes :class:`ParserElement` and :class:`ParseResults` to
see the base classes that most other pyparsing
classes inherit from. Use the docstrings for examples of how to:

 - construct literal match expressions from :class:`Literal` and
   :class:`CaselessLiteral` classes
 - construct character word-group expressions using the :class:`Word`
   class
 - see how to create repetitive expressions using :class:`ZeroOrMore`
   and :class:`OneOrMore` classes
 - use :class:`'+'<And>`, :class:`'|'<MatchFirst>`, :class:`'^'<Or>`,
   and :class:`'&'<Each>` operators to combine simple expressions into
   more complex ones
 - associate names with your parsed results using
   :class:`ParserElement.setResultsName`
 - access the parsed data, which is returned as a :class:`ParseResults`
   object
 - find some helpful expression short-cuts like :class:`delimitedList`
   and :class:`oneOf`
 - find more useful common expressions in the :class:`pyparsing_common`
   namespace class
"""

__version__ = "2.4.7"
__versionTime__ = "30 Mar 2020 00:43 UTC"
__author__ = "Paul McGuire <ptmcg@users.sourceforge.net>"

import string
from weakref import ref as wkref
import copy
import sys
import warnings
import re
import sre_constants
import collections
import pprint
import traceback
import types
from datetime import datetime
from operator import itemgetter
import itertools
from functools import wraps
from contextlib import contextmanager

try:
    # Python 3
    from itertools import filterfalse
except ImportError:
    from itertools import ifilterfalse as filterfalse

try:
    from _thread import RLock
except ImportError:
    from threading import RLock

try:
    # Python 3
    from collections.abc import Iterable
    from collections.abc import MutableMapping, Mapping
except ImportError:
    # Python 2.7
    from collections import Iterable
    from collections import MutableMapping, Mapping

try:
    from collections import OrderedDict as _OrderedDict
except ImportError:
    try:
        from ordereddict import OrderedDict as _OrderedDict
    except ImportError:
        _OrderedDict = None

try:
    from types import SimpleNamespace
except ImportError:
    class SimpleNamespace: pass

# version compatibility configuration
__compat__ = SimpleNamespace()
__compat__.__doc__ = """
    A cross-version compatibility configuration for pyparsing features that will be
    released in a future version. By setting values in this configuration to True,
    those features can be enabled in prior versions for compatibility development
    and testing.

     - collect_all_And_tokens - flag to enable fix for Issue #63 that fixes erroneous grouping
       of results names when an And expression is nested within an Or or MatchFirst; set to
       True to enable bugfix released in pyparsing 2.3.0, or False to preserve
       pre-2.3.0 handling of named results
"""
__compat__.collect_all_And_tokens = True

__diag__ = SimpleNamespace()
__diag__.__doc__ = """
Diagnostic configuration (all default to False)
     - warn_multiple_tokens_in_named_alternation - flag to enable warnings when a results
       name is defined on a MatchFirst or Or expression with one or more And subexpressions
       (only warns if __compat__.collect_all_And_tokens is False)
     - warn_ungrouped_named_tokens_in_collection - flag to enable warnings when a results
       name is defined on a containing expression with ungrouped subexpressions that also
       have results names
     - warn_name_set_on_empty_Forward - flag to enable warnings whan a Forward is defined
       with a results name, but has no contents defined
     - warn_on_multiple_string_args_to_oneof - flag to enable warnings whan oneOf is
       incorrectly called with multiple str arguments
     - enable_debug_on_named_expressions - flag to auto-enable debug on all subsequent
       calls to ParserElement.setName()
"""
__diag__.warn_multiple_tokens_in_named_alternation = False
__diag__.warn_ungrouped_named_tokens_in_collection = False
__diag__.warn_name_set_on_empty_Forward = False
__diag__.warn_on_multiple_string_args_to_oneof = False
__diag__.enable_debug_on_named_expressions = False
__diag__._all_names = [nm for nm in vars(__diag__) if nm.startswith("enable_") or nm.startswith("warn_")]

def _enable_all_warnings():
    __diag__.warn_multiple_tokens_in_named_alternation = True
    __diag__.warn_ungrouped_named_tokens_in_collection = True
    __diag__.warn_name_set_on_empty_Forward = True
    __diag__.warn_on_multiple_string_args_to_oneof = True
__diag__.enable_all_warnings = _enable_all_warnings


__all__ = ['__version__', '__versionTime__', '__author__', '__compat__', '__diag__',
           'And', 'CaselessKeyword', 'CaselessLiteral', 'CharsNotIn', 'Combine', 'Dict', 'Each', 'Empty',
           'FollowedBy', 'Forward', 'GoToColumn', 'Group', 'Keyword', 'LineEnd', 'LineStart', 'Literal',
           'PrecededBy', 'MatchFirst', 'NoMatch', 'NotAny', 'OneOrMore', 'OnlyOnce', 'Optional', 'Or',
           'ParseBaseException', 'ParseElementEnhance', 'ParseException', 'ParseExpression', 'ParseFatalException',
           'ParseResults', 'ParseSyntaxException', 'ParserElement', 'QuotedString', 'RecursiveGrammarException',
           'Regex', 'SkipTo', 'StringEnd', 'StringStart', 'Suppress', 'Token', 'TokenConverter',
           'White', 'Word', 'WordEnd', 'WordStart', 'ZeroOrMore', 'Char',
           'alphanums', 'alphas', 'alphas8bit', 'anyCloseTag', 'anyOpenTag', 'cStyleComment', 'col',
           'commaSeparatedList', 'commonHTMLEntity', 'countedArray', 'cppStyleComment', 'dblQuotedString',
           'dblSlashComment', 'delimitedList', 'dictOf', 'downcaseTokens', 'empty', 'hexnums',
           'htmlComment', 'javaStyleComment', 'line', 'lineEnd', 'lineStart', 'lineno',
           'makeHTMLTags', 'makeXMLTags', 'matchOnlyAtCol', 'matchPreviousExpr', 'matchPreviousLiteral',
           'nestedExpr', 'nullDebugAction', 'nums', 'oneOf', 'opAssoc', 'operatorPrecedence', 'printables',
           'punc8bit', 'pythonStyleComment', 'quotedString', 'removeQuotes', 'replaceHTMLEntity',
           'replaceWith', 'restOfLine', 'sglQuotedString', 'srange', 'stringEnd',
           'stringStart', 'traceParseAction', 'unicodeString', 'upcaseTokens', 'withAttribute',
           'indentedBlock', 'originalTextFor', 'ungroup', 'infixNotation', 'locatedExpr', 'withClass',
           'CloseMatch', 'tokenMap', 'pyparsing_common', 'pyparsing_unicode', 'unicode_set',
           'conditionAsParseAction', 're',
           ]

system_version = tuple(sys.version_info)[:3]
PY_3 = system_version[0] == 3
if PY_3:
    _MAX_INT = sys.maxsize
    basestring = str
    unichr = chr
    unicode = str
    _ustr = str

    # build list of single arg builtins, that can be used as parse actions
    singleArgBuiltins = [sum, len, sorted, reversed, list, tuple, set, any, all, min, max]

else:
    _MAX_INT = sys.maxint
    range = xrange

    def _ustr(obj):
        """Drop-in replacement for str(obj) that tries to be Unicode
        friendly. It first tries str(obj). If that fails with
        a UnicodeEncodeError, then it tries unicode(obj). It then
        < returns the unicode object | encodes it with the default
        encoding | ... >.
        """
        if isinstance(obj, unicode):
            return obj

        try:
            # If this works, then _ustr(obj) has the same behaviour as str(obj), so
            # it won't break any existing code.
            return str(obj)

        except UnicodeEncodeError:
            # Else encode it
            ret = unicode(obj).encode(sys.getdefaultencoding(), 'xmlcharrefreplace')
            xmlcharref = Regex(r'&#\d+;')
            xmlcharref.setParseAction(lambda t: '\\u' + hex(int(t[0][2:-1]))[2:])
            return xmlcharref.transformString(ret)

    # build list of single arg builtins, tolerant of Python version, that can be used as parse actions
    singleArgBuiltins = []
    import __builtin__

    for fname in "sum len sorted reversed list tuple set any all min max".split():
        try:
            singleArgBuiltins.append(getattr(__builtin__, fname))
        except AttributeError:
            continue

_generatorType = type((y for y in range(1)))

def _xml_escape(data):
    """Escape &, <, >, ", ', etc. in a string of data."""

    # ampersand must be replaced first
    from_symbols = '&><"\''
    to_symbols = ('&' + s + ';' for s in "amp gt lt quot apos".split())
    for from_, to_ in zip(from_symbols, to_symbols):
        data = data.replace(from_, to_)
    return data

alphas = string.ascii_uppercase + string.ascii_lowercase
nums = "0123456789"
hexnums = nums + "ABCDEFabcdef"
alphanums = alphas + nums
_bslash = chr(92)
printables = "".join(c for c in string.printable if c not in string.whitespace)


def conditionAsParseAction(fn, message=None, fatal=False):
    msg = message if message is not None else "failed user-defined condition"
    exc_type = ParseFatalException if fatal else ParseException
    fn = _trim_arity(fn)

    @wraps(fn)
    def pa(s, l, t):
        if not bool(fn(s, l, t)):
            raise exc_type(s, l, msg)

    return pa

class ParseBaseException(Exception):
    """base exception class for all parsing runtime exceptions"""
    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__(self, pstr, loc=0, msg=None, elem=None):
        self.loc = loc
        if msg is None:
            self.msg = pstr
            self.pstr = ""
        else:
            self.msg = msg
            self.pstr = pstr
        self.parserElement = elem
        self.args = (pstr, loc, msg)

    @classmethod
    def _from_exception(cls, pe):
        """
        internal factory method to simplify creating one type of ParseException
        from another - avoids having __init__ signature conflicts among subclasses
        """
        return cls(pe.pstr, pe.loc, pe.msg, pe.parserElement)

    def __getattr__(self, aname):
        """supported attributes by name are:
           - lineno - returns the line number of the exception text
           - col - returns the column number of the exception text
           - line - returns the line containing the exception text
        """
        if aname == "lineno":
            return lineno(self.loc, self.pstr)
        elif aname in ("col", "column"):
            return col(self.loc, self.pstr)
        elif aname == "line":
            return line(self.loc, self.pstr)
        else:
            raise AttributeError(aname)

    def __str__(self):
        if self.pstr:
            if self.loc >= len(self.pstr):
                foundstr = ', found end of text'
            else:
                foundstr = (', found %r' % self.pstr[self.loc:self.loc + 1]).replace(r'\\', '\\')
        else:
            foundstr = ''
        return ("%s%s  (at char %d), (line:%d, col:%d)" %
                   (self.msg, foundstr, self.loc, self.lineno, self.column))
    def __repr__(self):
        return _ustr(self)
    def markInputline(self, markerString=">!<"):
        """Extracts the exception line from the input string, and marks
           the location of the exception with a special symbol.
        """
        line_str = self.line
        line_column = self.column - 1
        if markerString:
            line_str = "".join((line_str[:line_column],
                                markerString, line_str[line_column:]))
        return line_str.strip()
    def __dir__(self):
        return "lineno col line".split() + dir(type(self))

class ParseException(ParseBaseException):
    """
    Exception thrown when parse expressions don't match class;
    supported attributes by name are:
    - lineno - returns the line number of the exception text
    - col - returns the column number of the exception text
    - line - returns the line containing the exception text

    Example::

        try:
            Word(nums).setName("integer").parseString("ABC")
        except ParseException as pe:
            print(pe)
            print("column: {}".format(pe.col))

    prints::

       Expected integer (at char 0), (line:1, col:1)
        column: 1

    """

    @staticmethod
    def explain(exc, depth=16):
        """
        Method to take an exception and translate the Python internal traceback into a list
        of the pyparsing expressions that caused the exception to be raised.

        Parameters:

         - exc - exception raised during parsing (need not be a ParseException, in support
           of Python exceptions that might be raised in a parse action)
         - depth (default=16) - number of levels back in the stack trace to list expression
           and function names; if None, the full stack trace names will be listed; if 0, only
           the failing input line, marker, and exception string will be shown

        Returns a multi-line string listing the ParserElements and/or function names in the
        exception's stack trace.

        Note: the diagnostic output will include string representations of the expressions
        that failed to parse. These representations will be more helpful if you use `setName` to
        give identifiable names to your expressions. Otherwise they will use the default string
        forms, which may be cryptic to read.

        explain() is only supported under Python 3.
        """
        import inspect

        if depth is None:
            depth = sys.getrecursionlimit()
        ret = []
        if isinstance(exc, ParseBaseException):
            ret.append(exc.line)
            ret.append(' ' * (exc.col - 1) + '^')
        ret.append("{0}: {1}".format(type(exc).__name__, exc))

        if depth > 0:
            callers = inspect.getinnerframes(exc.__traceback__, context=depth)
            seen = set()
            for i, ff in enumerate(callers[-depth:]):
                frm = ff[0]

                f_self = frm.f_locals.get('self', None)
                if isinstance(f_self, ParserElement):
                    if frm.f_code.co_name not in ('parseImpl', '_parseNoCache'):
                        continue
                    if f_self in seen:
                        continue
                    seen.add(f_self)

                    self_type = type(f_self)
                    ret.append("{0}.{1} - {2}".format(self_type.__module__,
                                                      self_type.__name__,
                                                      f_self))
                elif f_self is not None:
                    self_type = type(f_self)
                    ret.append("{0}.{1}".format(self_type.__module__,
                                                self_type.__name__))
                else:
                    code = frm.f_code
                    if code.co_name in ('wrapper', '<module>'):
                        continue

                    ret.append("{0}".format(code.co_name))

                depth -= 1
                if not depth:
                    break

        return '\n'.join(ret)


class ParseFatalException(ParseBaseException):
    """user-throwable exception thrown when inconsistent parse content
       is found; stops all parsing immediately"""
    pass

class ParseSyntaxException(ParseFatalException):
    """just like :class:`ParseFatalException`, but thrown internally
    when an :class:`ErrorStop<And._ErrorStop>` ('-' operator) indicates
    that parsing is to stop immediately because an unbacktrackable
    syntax error has been found.
    """
    pass

#~ class ReparseException(ParseBaseException):
    #~ """Experimental class - parse actions can raise this exception to cause
       #~ pyparsing to reparse the input string:
        #~ - with a modified input string, and/or
        #~ - with a modified start location
       #~ Set the values of the ReparseException in the constructor, and raise the
       #~ exception in a parse action to cause pyparsing to use the new string/location.
       #~ Setting the values as None causes no change to be made.
       #~ """
    #~ def __init_( self, newstring, restartLoc ):
        #~ self.newParseText = newstring
        #~ self.reparseLoc = restartLoc

class RecursiveGrammarException(Exception):
    """exception thrown by :class:`ParserElement.validate` if the
    grammar could be improperly recursive
    """
    def __init__(self, parseElementList):
        self.parseElementTrace = parseElementList

    def __str__(self):
        return "RecursiveGrammarException: %s" % self.parseElementTrace

class _ParseResultsWithOffset(object):
    def __init__(self, p1, p2):
        self.tup = (p1, p2)
    def __getitem__(self, i):
        return self.tup[i]
    def __repr__(self):
        return repr(self.tup[0])
    def setOffset(self, i):
        self.tup = (self.tup[0], i)

class ParseResults(object):
    """Structured parse results, to provide multiple means of access to
    the parsed data:

       - as a list (``len(results)``)
       - by list index (``results[0], results[1]``, etc.)
       - by attribute (``results.<resultsName>`` - see :class:`ParserElement.setResultsName`)

    Example::

        integer = Word(nums)
        date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))
        # equivalent form:
        # date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

        # parseString returns a ParseResults object
        result = date_str.parseString("1999/12/31")

        def test(s, fn=repr):
            print("%s -> %s" % (s, fn(eval(s))))
        test("list(result)")
        test("result[0]")
        test("result['month']")
        test("result.day")
        test("'month' in result")
        test("'minutes' in result")
        test("result.dump()", str)

    prints::

        list(result) -> ['1999', '/', '12', '/', '31']
        result[0] -> '1999'
        result['month'] -> '12'
        result.day -> '31'
        'month' in result -> True
        'minutes' in result -> False
        result.dump() -> ['1999', '/', '12', '/', '31']
        - day: 31
        - month: 12
        - year: 1999
    """
    def __new__(cls, toklist=None, name=None, asList=True, modal=True):
        if isinstance(toklist, cls):
            return toklist
        retobj = object.__new__(cls)
        retobj.__doinit = True
        return retobj

    # Performance tuning: we construct a *lot* of these, so keep this
    # constructor as small and fast as possible
    def __init__(self, toklist=None, name=None, asList=True, modal=True, isinstance=isinstance):
        if self.__doinit:
            self.__doinit = False
            self.__name = None
            self.__parent = None
            self.__accumNames = {}
            self.__asList = asList
            self.__modal = modal
            if toklist is None:
                toklist = []
            if isinstance(toklist, list):
                self.__toklist = toklist[:]
            elif isinstance(toklist, _generatorType):
                self.__toklist = list(toklist)
            else:
                self.__toklist = [toklist]
            self.__tokdict = dict()

        if name is not None and name:
            if not modal:
                self.__accumNames[name] = 0
            if isinstance(name, int):
                name = _ustr(name)  # will always return a str, but use _ustr for consistency
            self.__name = name
            if not (isinstance(toklist, (type(None), basestring, list)) and toklist in (None, '', [])):
                if isinstance(toklist, basestring):
                    toklist = [toklist]
                if asList:
                    if isinstance(toklist, ParseResults):
                        self[name] = _ParseResultsWithOffset(ParseResults(toklist.__toklist), 0)
                    else:
                        self[name] = _ParseResultsWithOffset(ParseResults(toklist[0]), 0)
                    self[name].__name = name
                else:
                    try:
                        self[name] = toklist[0]
                    except (KeyError, TypeError, IndexError):
                        self[name] = toklist

    def __getitem__(self, i):
        if isinstance(i, (int, slice)):
            return self.__toklist[i]
        else:
            if i not in self.__accumNames:
                return self.__tokdict[i][-1][0]
            else:
                return ParseResults([v[0] for v in self.__tokdict[i]])

    def __setitem__(self, k, v, isinstance=isinstance):
        if isinstance(v, _ParseResultsWithOffset):
            self.__tokdict[k] = self.__tokdict.get(k, list()) + [v]
            sub = v[0]
        elif isinstance(k, (int, slice)):
            self.__toklist[k] = v
            sub = v
        else:
            self.__tokdict[k] = self.__tokdict.get(k, list()) + [_ParseResultsWithOffset(v, 0)]
            sub = v
        if isinstance(sub, ParseResults):
            sub.__parent = wkref(self)

    def __delitem__(self, i):
        if isinstance(i, (int, slice)):
            mylen = len(self.__toklist)
            del self.__toklist[i]

            # convert int to slice
            if isinstance(i, int):
                if i < 0:
                    i += mylen
                i = slice(i, i + 1)
            # get removed indices
            removed = list(range(*i.indices(mylen)))
            removed.reverse()
            # fixup indices in token dictionary
            for name, occurrences in self.__tokdict.items():
                for j in removed:
                    for k, (value, position) in enumerate(occurrences):
                        occurrences[k] = _ParseResultsWithOffset(value, position - (position > j))
        else:
            del self.__tokdict[i]

    def __contains__(self, k):
        return k in self.__tokdict

    def __len__(self):
        return len(self.__toklist)

    def __bool__(self):
        return (not not self.__toklist)
    __nonzero__ = __bool__

    def __iter__(self):
        return iter(self.__toklist)

    def __reversed__(self):
        return iter(self.__toklist[::-1])

    def _iterkeys(self):
        if hasattr(self.__tokdict, "iterkeys"):
            return self.__tokdict.iterkeys()
        else:
            return iter(self.__tokdict)

    def _itervalues(self):
        return (self[k] for k in self._iterkeys())

    def _iteritems(self):
        return ((k, self[k]) for k in self._iterkeys())

    if PY_3:
        keys = _iterkeys
        """Returns an iterator of all named result keys."""

        values = _itervalues
        """Returns an iterator of all named result values."""

        items = _iteritems
        """Returns an iterator of all named result key-value tuples."""

    else:
        iterkeys = _iterkeys
        """Returns an iterator of all named result keys (Python 2.x only)."""

        itervalues = _itervalues
        """Returns an iterator of all named result values (Python 2.x only)."""

        iteritems = _iteritems
        """Returns an iterator of all named result key-value tuples (Python 2.x only)."""

        def keys(self):
            """Returns all named result keys (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iterkeys())

        def values(self):
            """Returns all named result values (as a list in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.itervalues())

        def items(self):
            """Returns all named result key-values (as a list of tuples in Python 2.x, as an iterator in Python 3.x)."""
            return list(self.iteritems())

    def haskeys(self):
        """Since keys() returns an iterator, this method is helpful in bypassing
           code that looks for the existence of any defined results names."""
        return bool(self.__tokdict)

    def pop(self, *args, **kwargs):
        """
        Removes and returns item at specified index (default= ``last``).
        Supports both ``list`` and ``dict`` semantics for ``pop()``. If
        passed no argument or an integer argument, it will use ``list``
        semantics and pop tokens from the list of parsed tokens. If passed
        a non-integer argument (most likely a string), it will use ``dict``
        semantics and pop the corresponding value from any defined results
        names. A second default return value argument is supported, just as in
        ``dict.pop()``.

        Example::

            def remove_first(tokens):
                tokens.pop(0)
            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']
            print(OneOrMore(Word(nums)).addParseAction(remove_first).parseString("0 123 321")) # -> ['123', '321']

            label = Word(alphas)
            patt = label("LABEL") + OneOrMore(Word(nums))
            print(patt.parseString("AAB 123 321").dump())

            # Use pop() in a parse action to remove named result (note that corresponding value is not
            # removed from list form of results)
            def remove_LABEL(tokens):
                tokens.pop("LABEL")
                return tokens
            patt.addParseAction(remove_LABEL)
            print(patt.parseString("AAB 123 321").dump())

        prints::

            ['AAB', '123', '321']
            - LABEL: AAB

            ['AAB', '123', '321']
        """
        if not args:
            args = [-1]
        for k, v in kwargs.items():
            if k == 'default':
                args = (args[0], v)
            else:
                raise TypeError("pop() got an unexpected keyword argument '%s'" % k)
        if (isinstance(args[0], int)
                or len(args) == 1
                or args[0] in self):
            index = args[0]
            ret = self[index]
            del self[index]
            return ret
        else:
            defaultvalue = args[1]
            return defaultvalue

    def get(self, key, defaultValue=None):
        """
        Returns named result matching the given key, or if there is no
        such name, then returns the given ``defaultValue`` or ``None`` if no
        ``defaultValue`` is specified.

        Similar to ``dict.get()``.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString("1999/12/31")
            print(result.get("year")) # -> '1999'
            print(result.get("hour", "not specified")) # -> 'not specified'
            print(result.get("hour")) # -> None
        """
        if key in self:
            return self[key]
        else:
            return defaultValue

    def insert(self, index, insStr):
        """
        Inserts new element at location index in the list of parsed tokens.

        Similar to ``list.insert()``.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to insert the parse location in the front of the parsed results
            def insert_locn(locn, tokens):
                tokens.insert(0, locn)
            print(OneOrMore(Word(nums)).addParseAction(insert_locn).parseString("0 123 321")) # -> [0, '0', '123', '321']
        """
        self.__toklist.insert(index, insStr)
        # fixup indices in token dictionary
        for name, occurrences in self.__tokdict.items():
            for k, (value, position) in enumerate(occurrences):
                occurrences[k] = _ParseResultsWithOffset(value, position + (position > index))

    def append(self, item):
        """
        Add single element to end of ParseResults list of elements.

        Example::

            print(OneOrMore(Word(nums)).parseString("0 123 321")) # -> ['0', '123', '321']

            # use a parse action to compute the sum of the parsed integers, and add it to the end
            def append_sum(tokens):
                tokens.append(sum(map(int, tokens)))
            print(OneOrMore(Word(nums)).addParseAction(append_sum).parseString("0 123 321")) # -> ['0', '123', '321', 444]
        """
        self.__toklist.append(item)

    def extend(self, itemseq):
        """
        Add sequence of elements to end of ParseResults list of elements.

        Example::

            patt = OneOrMore(Word(alphas))

            # use a parse action to append the reverse of the matched strings, to make a palindrome
            def make_palindrome(tokens):
                tokens.extend(reversed([t[::-1] for t in tokens]))
                return ''.join(tokens)
            print(patt.addParseAction(make_palindrome).parseString("lskdj sdlkjf lksd")) # -> 'lskdjsdlkjflksddsklfjkldsjdksl'
        """
        if isinstance(itemseq, ParseResults):
            self.__iadd__(itemseq)
        else:
            self.__toklist.extend(itemseq)

    def clear(self):
        """
        Clear all elements and results names.
        """
        del self.__toklist[:]
        self.__tokdict.clear()

    def __getattr__(self, name):
        try:
            return self[name]
        except KeyError:
            return ""

    def __add__(self, other):
        ret = self.copy()
        ret += other
        return ret

    def __iadd__(self, other):
        if other.__tokdict:
            offset = len(self.__toklist)
            addoffset = lambda a: offset if a < 0 else a + offset
            otheritems = other.__tokdict.items()
            otherdictitems = [(k, _ParseResultsWithOffset(v[0], addoffset(v[1])))
                              for k, vlist in otheritems for v in vlist]
            for k, v in otherdictitems:
                self[k] = v
                if isinstance(v[0], ParseResults):
                    v[0].__parent = wkref(self)

        self.__toklist += other.__toklist
        self.__accumNames.update(other.__accumNames)
        return self

    def __radd__(self, other):
        if isinstance(other, int) and other == 0:
            # useful for merging many ParseResults using sum() builtin
            return self.copy()
        else:
            # this may raise a TypeError - so be it
            return other + self

    def __repr__(self):
        return "(%s, %s)" % (repr(self.__toklist), repr(self.__tokdict))

    def __str__(self):
        return '[' + ', '.join(_ustr(i) if isinstance(i, ParseResults) else repr(i) for i in self.__toklist) + ']'

    def _asStringList(self, sep=''):
        out = []
        for item in self.__toklist:
            if out and sep:
                out.append(sep)
            if isinstance(item, ParseResults):
                out += item._asStringList()
            else:
                out.append(_ustr(item))
        return out

    def asList(self):
        """
        Returns the parse results as a nested list of matching tokens, all converted to strings.

        Example::

            patt = OneOrMore(Word(alphas))
            result = patt.parseString("sldkj lsdkj sldkj")
            # even though the result prints in string-like form, it is actually a pyparsing ParseResults
            print(type(result), result) # -> <class 'pyparsing.ParseResults'> ['sldkj', 'lsdkj', 'sldkj']

            # Use asList() to create an actual list
            result_list = result.asList()
            print(type(result_list), result_list) # -> <class 'list'> ['sldkj', 'lsdkj', 'sldkj']
        """
        return [res.asList() if isinstance(res, ParseResults) else res for res in self.__toklist]

    def asDict(self):
        """
        Returns the named parse results as a nested dictionary.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString('12/31/1999')
            print(type(result), repr(result)) # -> <class 'pyparsing.ParseResults'> (['12', '/', '31', '/', '1999'], {'day': [('1999', 4)], 'year': [('12', 0)], 'month': [('31', 2)]})

            result_dict = result.asDict()
            print(type(result_dict), repr(result_dict)) # -> <class 'dict'> {'day': '1999', 'year': '12', 'month': '31'}

            # even though a ParseResults supports dict-like access, sometime you just need to have a dict
            import json
            print(json.dumps(result)) # -> Exception: TypeError: ... is not JSON serializable
            print(json.dumps(result.asDict())) # -> {"month": "31", "day": "1999", "year": "12"}
        """
        if PY_3:
            item_fn = self.items
        else:
            item_fn = self.iteritems

        def toItem(obj):
            if isinstance(obj, ParseResults):
                if obj.haskeys():
                    return obj.asDict()
                else:
                    return [toItem(v) for v in obj]
            else:
                return obj

        return dict((k, toItem(v)) for k, v in item_fn())

    def copy(self):
        """
        Returns a new copy of a :class:`ParseResults` object.
        """
        ret = ParseResults(self.__toklist)
        ret.__tokdict = dict(self.__tokdict.items())
        ret.__parent = self.__parent
        ret.__accumNames.update(self.__accumNames)
        ret.__name = self.__name
        return ret

    def asXML(self, doctag=None, namedItemsOnly=False, indent="", formatted=True):
        """
        (Deprecated) Returns the parse results as XML. Tags are created for tokens and lists that have defined results names.
        """
        nl = "\n"
        out = []
        namedItems = dict((v[1], k) for (k, vlist) in self.__tokdict.items()
                          for v in vlist)
        nextLevelIndent = indent + "  "

        # collapse out indents if formatting is not desired
        if not formatted:
            indent = ""
            nextLevelIndent = ""
            nl = ""

        selfTag = None
        if doctag is not None:
            selfTag = doctag
        else:
            if self.__name:
                selfTag = self.__name

        if not selfTag:
            if namedItemsOnly:
                return ""
            else:
                selfTag = "ITEM"

        out += [nl, indent, "<", selfTag, ">"]

        for i, res in enumerate(self.__toklist):
            if isinstance(res, ParseResults):
                if i in namedItems:
                    out += [res.asXML(namedItems[i],
                                      namedItemsOnly and doctag is None,
                                      nextLevelIndent,
                                      formatted)]
                else:
                    out += [res.asXML(None,
                                      namedItemsOnly and doctag is None,
                                      nextLevelIndent,
                                      formatted)]
            else:
                # individual token, see if there is a name for it
                resTag = None
                if i in namedItems:
                    resTag = namedItems[i]
                if not resTag:
                    if namedItemsOnly:
                        continue
                    else:
                        resTag = "ITEM"
                xmlBodyText = _xml_escape(_ustr(res))
                out += [nl, nextLevelIndent, "<", resTag, ">",
                        xmlBodyText,
                                                "</", resTag, ">"]

        out += [nl, indent, "</", selfTag, ">"]
        return "".join(out)

    def __lookup(self, sub):
        for k, vlist in self.__tokdict.items():
            for v, loc in vlist:
                if sub is v:
                    return k
        return None

    def getName(self):
        r"""
        Returns the results name for this token expression. Useful when several
        different expressions might match at a particular location.

        Example::

            integer = Word(nums)
            ssn_expr = Regex(r"\d\d\d-\d\d-\d\d\d\d")
            house_number_expr = Suppress('#') + Word(nums, alphanums)
            user_data = (Group(house_number_expr)("house_number")
                        | Group(ssn_expr)("ssn")
                        | Group(integer)("age"))
            user_info = OneOrMore(user_data)

            result = user_info.parseString("22 111-22-3333 #221B")
            for item in result:
                print(item.getName(), ':', item[0])

        prints::

            age : 22
            ssn : 111-22-3333
            house_number : 221B
        """
        if self.__name:
            return self.__name
        elif self.__parent:
            par = self.__parent()
            if par:
                return par.__lookup(self)
            else:
                return None
        elif (len(self) == 1
              and len(self.__tokdict) == 1
              and next(iter(self.__tokdict.values()))[0][1] in (0, -1)):
            return next(iter(self.__tokdict.keys()))
        else:
            return None

    def dump(self, indent='', full=True, include_list=True, _depth=0):
        """
        Diagnostic method for listing out the contents of
        a :class:`ParseResults`. Accepts an optional ``indent`` argument so
        that this string can be embedded in a nested display of other data.

        Example::

            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            result = date_str.parseString('12/31/1999')
            print(result.dump())

        prints::

            ['12', '/', '31', '/', '1999']
            - day: 1999
            - month: 31
            - year: 12
        """
        out = []
        NL = '\n'
        if include_list:
            out.append(indent + _ustr(self.asList()))
        else:
            out.append('')

        if full:
            if self.haskeys():
                items = sorted((str(k), v) for k, v in self.items())
                for k, v in items:
                    if out:
                        out.append(NL)
                    out.append("%s%s- %s: " % (indent, ('  ' * _depth), k))
                    if isinstance(v, ParseResults):
                        if v:
                            out.append(v.dump(indent=indent, full=full, include_list=include_list, _depth=_depth + 1))
                        else:
                            out.append(_ustr(v))
                    else:
                        out.append(repr(v))
            elif any(isinstance(vv, ParseResults) for vv in self):
                v = self
                for i, vv in enumerate(v):
                    if isinstance(vv, ParseResults):
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,
                                                            ('  ' * (_depth)),
                                                            i,
                                                            indent,
                                                            ('  ' * (_depth + 1)),
                                                            vv.dump(indent=indent,
                                                                    full=full,
                                                                    include_list=include_list,
                                                                    _depth=_depth + 1)))
                    else:
                        out.append("\n%s%s[%d]:\n%s%s%s" % (indent,
                                                            ('  ' * (_depth)),
                                                            i,
                                                            indent,
                                                            ('  ' * (_depth + 1)),
                                                            _ustr(vv)))

        return "".join(out)

    def pprint(self, *args, **kwargs):
        """
        Pretty-printer for parsed results as a list, using the
        `pprint <https://docs.python.org/3/library/pprint.html>`_ module.
        Accepts additional positional or keyword args as defined for
        `pprint.pprint <https://docs.python.org/3/library/pprint.html#pprint.pprint>`_ .

        Example::

            ident = Word(alphas, alphanums)
            num = Word(nums)
            func = Forward()
            term = ident | num | Group('(' + func + ')')
            func <<= ident + Group(Optional(delimitedList(term)))
            result = func.parseString("fna a,b,(fnb c,d,200),100")
            result.pprint(width=40)

        prints::

            ['fna',
             ['a',
              'b',
              ['(', 'fnb', ['c', 'd', '200'], ')'],
              '100']]
        """
        pprint.pprint(self.asList(), *args, **kwargs)

    # add support for pickle protocol
    def __getstate__(self):
        return (self.__toklist,
                (self.__tokdict.copy(),
                 self.__parent is not None and self.__parent() or None,
                 self.__accumNames,
                 self.__name))

    def __setstate__(self, state):
        self.__toklist = state[0]
        self.__tokdict, par, inAccumNames, self.__name = state[1]
        self.__accumNames = {}
        self.__accumNames.update(inAccumNames)
        if par is not None:
            self.__parent = wkref(par)
        else:
            self.__parent = None

    def __getnewargs__(self):
        return self.__toklist, self.__name, self.__asList, self.__modal

    def __dir__(self):
        return dir(type(self)) + list(self.keys())

    @classmethod
    def from_dict(cls, other, name=None):
        """
        Helper classmethod to construct a ParseResults from a dict, preserving the
        name-value relations as results names. If an optional 'name' argument is
        given, a nested ParseResults will be returned
        """
        def is_iterable(obj):
            try:
                iter(obj)
            except Exception:
                return False
            else:
                if PY_3:
                    return not isinstance(obj, (str, bytes))
                else:
                    return not isinstance(obj, basestring)

        ret = cls([])
        for k, v in other.items():
            if isinstance(v, Mapping):
                ret += cls.from_dict(v, name=k)
            else:
                ret += cls([v], name=k, asList=is_iterable(v))
        if name is not None:
            ret = cls([ret], name=name)
        return ret

MutableMapping.register(ParseResults)

def col (loc, strg):
    """Returns current column within a string, counting newlines as line separators.
   The first column is number 1.

   Note: the default parsing behavior is to expand tabs in the input string
   before starting the parsing process.  See
   :class:`ParserElement.parseString` for more
   information on parsing strings containing ``<TAB>`` s, and suggested
   methods to maintain a consistent view of the parsed string, the parse
   location, and line and column positions within the parsed string.
   """
    s = strg
    return 1 if 0 < loc < len(s) and s[loc-1] == '\n' else loc - s.rfind("\n", 0, loc)

def lineno(loc, strg):
    """Returns current line number within a string, counting newlines as line separators.
    The first line is number 1.

    Note - the default parsing behavior is to expand tabs in the input string
    before starting the parsing process.  See :class:`ParserElement.parseString`
    for more information on parsing strings containing ``<TAB>`` s, and
    suggested methods to maintain a consistent view of the parsed string, the
    parse location, and line and column positions within the parsed string.
    """
    return strg.count("\n", 0, loc) + 1

def line(loc, strg):
    """Returns the line of text containing loc within a string, counting newlines as line separators.
       """
    lastCR = strg.rfind("\n", 0, loc)
    nextCR = strg.find("\n", loc)
    if nextCR >= 0:
        return strg[lastCR + 1:nextCR]
    else:
        return strg[lastCR + 1:]

def _defaultStartDebugAction(instring, loc, expr):
    print(("Match " + _ustr(expr) + " at loc " + _ustr(loc) + "(%d,%d)" % (lineno(loc, instring), col(loc, instring))))

def _defaultSuccessDebugAction(instring, startloc, endloc, expr, toks):
    print("Matched " + _ustr(expr) + " -> " + str(toks.asList()))

def _defaultExceptionDebugAction(instring, loc, expr, exc):
    print("Exception raised:" + _ustr(exc))

def nullDebugAction(*args):
    """'Do-nothing' debug action, to suppress debugging output during parsing."""
    pass

# Only works on Python 3.x - nonlocal is toxic to Python 2 installs
#~ 'decorator to trim function calls to match the arity of the target'
#~ def _trim_arity(func, maxargs=3):
    #~ if func in singleArgBuiltins:
        #~ return lambda s,l,t: func(t)
    #~ limit = 0
    #~ foundArity = False
    #~ def wrapper(*args):
        #~ nonlocal limit,foundArity
        #~ while 1:
            #~ try:
                #~ ret = func(*args[limit:])
                #~ foundArity = True
                #~ return ret
            #~ except TypeError:
                #~ if limit == maxargs or foundArity:
                    #~ raise
                #~ limit += 1
                #~ continue
    #~ return wrapper

# this version is Python 2.x-3.x cross-compatible
'decorator to trim function calls to match the arity of the target'
def _trim_arity(func, maxargs=2):
    if func in singleArgBuiltins:
        return lambda s, l, t: func(t)
    limit = [0]
    foundArity = [False]

    # traceback return data structure changed in Py3.5 - normalize back to plain tuples
    if system_version[:2] >= (3, 5):
        def extract_stack(limit=0):
            # special handling for Python 3.5.0 - extra deep call stack by 1
            offset = -3 if system_version == (3, 5, 0) else -2
            frame_summary = traceback.extract_stack(limit=-offset + limit - 1)[offset]
            return [frame_summary[:2]]
        def extract_tb(tb, limit=0):
            frames = traceback.extract_tb(tb, limit=limit)
            frame_summary = frames[-1]
            return [frame_summary[:2]]
    else:
        extract_stack = traceback.extract_stack
        extract_tb = traceback.extract_tb

    # synthesize what would be returned by traceback.extract_stack at the call to
    # user's parse action 'func', so that we don't incur call penalty at parse time

    LINE_DIFF = 6
    # IF ANY CODE CHANGES, EVEN JUST COMMENTS OR BLANK LINES, BETWEEN THE NEXT LINE AND
    # THE CALL TO FUNC INSIDE WRAPPER, LINE_DIFF MUST BE MODIFIED!!!!
    this_line = extract_stack(limit=2)[-1]
    pa_call_line_synth = (this_line[0], this_line[1] + LINE_DIFF)

    def wrapper(*args):
        while 1:
            try:
                ret = func(*args[limit[0]:])
                foundArity[0] = True
                return ret
            except TypeError:
                # re-raise TypeErrors if they did not come from our arity testing
                if foundArity[0]:
                    raise
                else:
                    try:
                        tb = sys.exc_info()[-1]
                        if not extract_tb(tb, limit=2)[-1][:2] == pa_call_line_synth:
                            raise
                    finally:
                        try:
                            del tb
                        except NameError:
                            pass

                if limit[0] <= maxargs:
                    limit[0] += 1
                    continue
                raise

    # copy func name to wrapper for sensible debug output
    func_name = "<parse action>"
    try:
        func_name = getattr(func, '__name__',
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    wrapper.__name__ = func_name

    return wrapper


class ParserElement(object):
    """Abstract base level parser element class."""
    DEFAULT_WHITE_CHARS = " \n\t\r"
    verbose_stacktrace = False

    @staticmethod
    def setDefaultWhitespaceChars(chars):
        r"""
        Overrides the default whitespace chars

        Example::

            # default whitespace chars are space, <TAB> and newline
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def', 'ghi', 'jkl']

            # change to just treat newline as significant
            ParserElement.setDefaultWhitespaceChars(" \t")
            OneOrMore(Word(alphas)).parseString("abc def\nghi jkl")  # -> ['abc', 'def']
        """
        ParserElement.DEFAULT_WHITE_CHARS = chars

    @staticmethod
    def inlineLiteralsUsing(cls):
        """
        Set class to be used for inclusion of string literals into a parser.

        Example::

            # default literal class used is Literal
            integer = Word(nums)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']


            # change to Suppress
            ParserElement.inlineLiteralsUsing(Suppress)
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")

            date_str.parseString("1999/12/31")  # -> ['1999', '12', '31']
        """
        ParserElement._literalStringClass = cls

    @classmethod
    def _trim_traceback(cls, tb):
        while tb.tb_next:
            tb = tb.tb_next
        return tb

    def __init__(self, savelist=False):
        self.parseAction = list()
        self.failAction = None
        # ~ self.name = "<unknown>"  # don't define self.name, let subclasses try/except upcall
        self.strRepr = None
        self.resultsName = None
        self.saveAsList = savelist
        self.skipWhitespace = True
        self.whiteChars = set(ParserElement.DEFAULT_WHITE_CHARS)
        self.copyDefaultWhiteChars = True
        self.mayReturnEmpty = False # used when checking for left-recursion
        self.keepTabs = False
        self.ignoreExprs = list()
        self.debug = False
        self.streamlined = False
        self.mayIndexError = True # used to optimize exception handling for subclasses that don't advance parse index
        self.errmsg = ""
        self.modalResults = True # used to mark results names as modal (report only last) or cumulative (list all)
        self.debugActions = (None, None, None)  # custom debug actions
        self.re = None
        self.callPreparse = True # used to avoid redundant calls to preParse
        self.callDuringTry = False

    def copy(self):
        """
        Make a copy of this :class:`ParserElement`.  Useful for defining
        different parse actions for the same parsing pattern, using copies of
        the original parse element.

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            integerK = integer.copy().addParseAction(lambda toks: toks[0] * 1024) + Suppress("K")
            integerM = integer.copy().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")

            print(OneOrMore(integerK | integerM | integer).parseString("5K 100 640K 256M"))

        prints::

            [5120, 100, 655360, 268435456]

        Equivalent form of ``expr.copy()`` is just ``expr()``::

            integerM = integer().addParseAction(lambda toks: toks[0] * 1024 * 1024) + Suppress("M")
        """
        cpy = copy.copy(self)
        cpy.parseAction = self.parseAction[:]
        cpy.ignoreExprs = self.ignoreExprs[:]
        if self.copyDefaultWhiteChars:
            cpy.whiteChars = ParserElement.DEFAULT_WHITE_CHARS
        return cpy

    def setName(self, name):
        """
        Define name for this expression, makes debugging and exception messages clearer.

        Example::

            Word(nums).parseString("ABC")  # -> Exception: Expected W:(0123...) (at char 0), (line:1, col:1)
            Word(nums).setName("integer").parseString("ABC")  # -> Exception: Expected integer (at char 0), (line:1, col:1)
        """
        self.name = name
        self.errmsg = "Expected " + self.name
        if __diag__.enable_debug_on_named_expressions:
            self.setDebug()
        return self

    def setResultsName(self, name, listAllMatches=False):
        """
        Define name for referencing matching tokens as a nested attribute
        of the returned parse results.
        NOTE: this returns a *copy* of the original :class:`ParserElement` object;
        this is so that the client can define a basic element, such as an
        integer, and reference it in multiple places with different names.

        You can also set results names using the abbreviated syntax,
        ``expr("name")`` in place of ``expr.setResultsName("name")``
        - see :class:`__call__`.

        Example::

            date_str = (integer.setResultsName("year") + '/'
                        + integer.setResultsName("month") + '/'
                        + integer.setResultsName("day"))

            # equivalent form:
            date_str = integer("year") + '/' + integer("month") + '/' + integer("day")
        """
        return self._setResultsName(name, listAllMatches)

    def _setResultsName(self, name, listAllMatches=False):
        newself = self.copy()
        if name.endswith("*"):
            name = name[:-1]
            listAllMatches = True
        newself.resultsName = name
        newself.modalResults = not listAllMatches
        return newself

    def setBreak(self, breakFlag=True):
        """Method to invoke the Python pdb debugger when this element is
           about to be parsed. Set ``breakFlag`` to True to enable, False to
           disable.
        """
        if breakFlag:
            _parseMethod = self._parse
            def breaker(instring, loc, doActions=True, callPreParse=True):
                import pdb
                # this call to pdb.set_trace() is intentional, not a checkin error
                pdb.set_trace()
                return _parseMethod(instring, loc, doActions, callPreParse)
            breaker._originalParseMethod = _parseMethod
            self._parse = breaker
        else:
            if hasattr(self._parse, "_originalParseMethod"):
                self._parse = self._parse._originalParseMethod
        return self

    def setParseAction(self, *fns, **kwargs):
        """
        Define one or more actions to perform when successfully matching parse element definition.
        Parse action fn is a callable method with 0-3 arguments, called as ``fn(s, loc, toks)`` ,
        ``fn(loc, toks)`` , ``fn(toks)`` , or just ``fn()`` , where:

        - s   = the original string being parsed (see note below)
        - loc = the location of the matching substring
        - toks = a list of the matched tokens, packaged as a :class:`ParseResults` object

        If the functions in fns modify the tokens, they can return them as the return
        value from fn, and the modified list of tokens will replace the original.
        Otherwise, fn does not need to return any value.

        If None is passed as the parse action, all previously added parse actions for this
        expression are cleared.

        Optional keyword arguments:
        - callDuringTry = (default= ``False``) indicate if parse action should be run during lookaheads and alternate testing

        Note: the default parsing behavior is to expand tabs in the input string
        before starting the parsing process.  See :class:`parseString for more
        information on parsing strings containing ``<TAB>`` s, and suggested
        methods to maintain a consistent view of the parsed string, the parse
        location, and line and column positions within the parsed string.

        Example::

            integer = Word(nums)
            date_str = integer + '/' + integer + '/' + integer

            date_str.parseString("1999/12/31")  # -> ['1999', '/', '12', '/', '31']

            # use parse action to convert to ints at parse time
            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            date_str = integer + '/' + integer + '/' + integer

            # note that integer fields are now ints, not strings
            date_str.parseString("1999/12/31")  # -> [1999, '/', 12, '/', 31]
        """
        if list(fns) == [None,]:
            self.parseAction = []
        else:
            if not all(callable(fn) for fn in fns):
                raise TypeError("parse actions must be callable")
            self.parseAction = list(map(_trim_arity, list(fns)))
            self.callDuringTry = kwargs.get("callDuringTry", False)
        return self

    def addParseAction(self, *fns, **kwargs):
        """
        Add one or more parse actions to expression's list of parse actions. See :class:`setParseAction`.

        See examples in :class:`copy`.
        """
        self.parseAction += list(map(_trim_arity, list(fns)))
        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def addCondition(self, *fns, **kwargs):
        """Add a boolean predicate function to expression's list of parse actions. See
        :class:`setParseAction` for function call signatures. Unlike ``setParseAction``,
        functions passed to ``addCondition`` need to return boolean success/fail of the condition.

        Optional keyword arguments:
        - message = define a custom message to be used in the raised exception
        - fatal   = if True, will raise ParseFatalException to stop parsing immediately; otherwise will raise ParseException

        Example::

            integer = Word(nums).setParseAction(lambda toks: int(toks[0]))
            year_int = integer.copy()
            year_int.addCondition(lambda toks: toks[0] >= 2000, message="Only support years 2000 and later")
            date_str = year_int + '/' + integer + '/' + integer

            result = date_str.parseString("1999/12/31")  # -> Exception: Only support years 2000 and later (at char 0), (line:1, col:1)
        """
        for fn in fns:
            self.parseAction.append(conditionAsParseAction(fn, message=kwargs.get('message'),
                                                           fatal=kwargs.get('fatal', False)))

        self.callDuringTry = self.callDuringTry or kwargs.get("callDuringTry", False)
        return self

    def setFailAction(self, fn):
        """Define action to perform if parsing fails at this expression.
           Fail acton fn is a callable function that takes the arguments
           ``fn(s, loc, expr, err)`` where:
           - s = string being parsed
           - loc = location where expression match was attempted and failed
           - expr = the parse expression that failed
           - err = the exception thrown
           The function returns no value.  It may throw :class:`ParseFatalException`
           if it is desired to stop parsing immediately."""
        self.failAction = fn
        return self

    def _skipIgnorables(self, instring, loc):
        exprsFound = True
        while exprsFound:
            exprsFound = False
            for e in self.ignoreExprs:
                try:
                    while 1:
                        loc, dummy = e._parse(instring, loc)
                        exprsFound = True
                except ParseException:
                    pass
        return loc

    def preParse(self, instring, loc):
        if self.ignoreExprs:
            loc = self._skipIgnorables(instring, loc)

        if self.skipWhitespace:
            wt = self.whiteChars
            instrlen = len(instring)
            while loc < instrlen and instring[loc] in wt:
                loc += 1

        return loc

    def parseImpl(self, instring, loc, doActions=True):
        return loc, []

    def postParse(self, instring, loc, tokenlist):
        return tokenlist

    # ~ @profile
    def _parseNoCache(self, instring, loc, doActions=True, callPreParse=True):
        TRY, MATCH, FAIL = 0, 1, 2
        debugging = (self.debug)  # and doActions)

        if debugging or self.failAction:
            # ~ print ("Match", self, "at loc", loc, "(%d, %d)" % (lineno(loc, instring), col(loc, instring)))
            if self.debugActions[TRY]:
                self.debugActions[TRY](instring, loc, self)
            try:
                if callPreParse and self.callPreparse:
                    preloc = self.preParse(instring, loc)
                else:
                    preloc = loc
                tokensStart = preloc
                if self.mayIndexError or preloc >= len(instring):
                    try:
                        loc, tokens = self.parseImpl(instring, preloc, doActions)
                    except IndexError:
                        raise ParseException(instring, len(instring), self.errmsg, self)
                else:
                    loc, tokens = self.parseImpl(instring, preloc, doActions)
            except Exception as err:
                # ~ print ("Exception raised:", err)
                if self.debugActions[FAIL]:
                    self.debugActions[FAIL](instring, tokensStart, self, err)
                if self.failAction:
                    self.failAction(instring, tokensStart, self, err)
                raise
        else:
            if callPreParse and self.callPreparse:
                preloc = self.preParse(instring, loc)
            else:
                preloc = loc
            tokensStart = preloc
            if self.mayIndexError or preloc >= len(instring):
                try:
                    loc, tokens = self.parseImpl(instring, preloc, doActions)
                except IndexError:
                    raise ParseException(instring, len(instring), self.errmsg, self)
            else:
                loc, tokens = self.parseImpl(instring, preloc, doActions)

        tokens = self.postParse(instring, loc, tokens)

        retTokens = ParseResults(tokens, self.resultsName, asList=self.saveAsList, modal=self.modalResults)
        if self.parseAction and (doActions or self.callDuringTry):
            if debugging:
                try:
                    for fn in self.parseAction:
                        try:
                            tokens = fn(instring, tokensStart, retTokens)
                        except IndexError as parse_action_exc:
                            exc = ParseException("exception raised in parse action")
                            exc.__cause__ = parse_action_exc
                            raise exc

                        if tokens is not None and tokens is not retTokens:
                            retTokens = ParseResults(tokens,
                                                      self.resultsName,
                                                      asList=self.saveAsList and isinstance(tokens, (ParseResults, list)),
                                                      modal=self.modalResults)
                except Exception as err:
                    # ~ print "Exception raised in user parse action:", err
                    if self.debugActions[FAIL]:
                        self.debugActions[FAIL](instring, tokensStart, self, err)
                    raise
            else:
                for fn in self.parseAction:
                    try:
                        tokens = fn(instring, tokensStart, retTokens)
                    except IndexError as parse_action_exc:
                        exc = ParseException("exception raised in parse action")
                        exc.__cause__ = parse_action_exc
                        raise exc

                    if tokens is not None and tokens is not retTokens:
                        retTokens = ParseResults(tokens,
                                                  self.resultsName,
                                                  asList=self.saveAsList and isinstance(tokens, (ParseResults, list)),
                                                  modal=self.modalResults)
        if debugging:
            # ~ print ("Matched", self, "->", retTokens.asList())
            if self.debugActions[MATCH]:
                self.debugActions[MATCH](instring, tokensStart, loc, self, retTokens)

        return loc, retTokens

    def tryParse(self, instring, loc):
        try:
            return self._parse(instring, loc, doActions=False)[0]
        except ParseFatalException:
            raise ParseException(instring, loc, self.errmsg, self)

    def canParseNext(self, instring, loc):
        try:
            self.tryParse(instring, loc)
        except (ParseException, IndexError):
            return False
        else:
            return True

    class _UnboundedCache(object):
        def __init__(self):
            cache = {}
            self.not_in_cache = not_in_cache = object()

            def get(self, key):
                return cache.get(key, not_in_cache)

            def set(self, key, value):
                cache[key] = value

            def clear(self):
                cache.clear()

            def cache_len(self):
                return len(cache)

            self.get = types.MethodType(get, self)
            self.set = types.MethodType(set, self)
            self.clear = types.MethodType(clear, self)
            self.__len__ = types.MethodType(cache_len, self)

    if _OrderedDict is not None:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = _OrderedDict()

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(cache) > size:
                        try:
                            cache.popitem(False)
                        except KeyError:
                            pass

                def clear(self):
                    cache.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    else:
        class _FifoCache(object):
            def __init__(self, size):
                self.not_in_cache = not_in_cache = object()

                cache = {}
                key_fifo = collections.deque([], size)

                def get(self, key):
                    return cache.get(key, not_in_cache)

                def set(self, key, value):
                    cache[key] = value
                    while len(key_fifo) > size:
                        cache.pop(key_fifo.popleft(), None)
                    key_fifo.append(key)

                def clear(self):
                    cache.clear()
                    key_fifo.clear()

                def cache_len(self):
                    return len(cache)

                self.get = types.MethodType(get, self)
                self.set = types.MethodType(set, self)
                self.clear = types.MethodType(clear, self)
                self.__len__ = types.MethodType(cache_len, self)

    # argument cache for optimizing repeated calls when backtracking through recursive expressions
    packrat_cache = {} # this is set later by enabledPackrat(); this is here so that resetCache() doesn't fail
    packrat_cache_lock = RLock()
    packrat_cache_stats = [0, 0]

    # this method gets repeatedly called during backtracking with the same arguments -
    # we can cache these arguments and save ourselves the trouble of re-parsing the contained expression
    def _parseCache(self, instring, loc, doActions=True, callPreParse=True):
        HIT, MISS = 0, 1
        lookup = (self, instring, loc, callPreParse, doActions)
        with ParserElement.packrat_cache_lock:
            cache = ParserElement.packrat_cache
            value = cache.get(lookup)
            if value is cache.not_in_cache:
                ParserElement.packrat_cache_stats[MISS] += 1
                try:
                    value = self._parseNoCache(instring, loc, doActions, callPreParse)
                except ParseBaseException as pe:
                    # cache a copy of the exception, without the traceback
                    cache.set(lookup, pe.__class__(*pe.args))
                    raise
                else:
                    cache.set(lookup, (value[0], value[1].copy()))
                    return value
            else:
                ParserElement.packrat_cache_stats[HIT] += 1
                if isinstance(value, Exception):
                    raise value
                return value[0], value[1].copy()

    _parse = _parseNoCache

    @staticmethod
    def resetCache():
        ParserElement.packrat_cache.clear()
        ParserElement.packrat_cache_stats[:] = [0] * len(ParserElement.packrat_cache_stats)

    _packratEnabled = False
    @staticmethod
    def enablePackrat(cache_size_limit=128):
        """Enables "packrat" parsing, which adds memoizing to the parsing logic.
           Repeated parse attempts at the same string location (which happens
           often in many complex grammars) can immediately return a cached value,
           instead of re-executing parsing/validating code.  Memoizing is done of
           both valid results and parsing exceptions.

           Parameters:

           - cache_size_limit - (default= ``128``) - if an integer value is provided
             will limit the size of the packrat cache; if None is passed, then
             the cache size will be unbounded; if 0 is passed, the cache will
             be effectively disabled.

           This speedup may break existing programs that use parse actions that
           have side-effects.  For this reason, packrat parsing is disabled when
           you first import pyparsing.  To activate the packrat feature, your
           program must call the class method :class:`ParserElement.enablePackrat`.
           For best results, call ``enablePackrat()`` immediately after
           importing pyparsing.

           Example::

               from pip._vendor import pyparsing
               pyparsing.ParserElement.enablePackrat()
        """
        if not ParserElement._packratEnabled:
            ParserElement._packratEnabled = True
            if cache_size_limit is None:
                ParserElement.packrat_cache = ParserElement._UnboundedCache()
            else:
                ParserElement.packrat_cache = ParserElement._FifoCache(cache_size_limit)
            ParserElement._parse = ParserElement._parseCache

    def parseString(self, instring, parseAll=False):
        """
        Execute the parse expression with the given string.
        This is the main interface to the client code, once the complete
        expression has been built.

        Returns the parsed data as a :class:`ParseResults` object, which may be
        accessed as a list, or as a dict or object with attributes if the given parser
        includes results names.

        If you want the grammar to require that the entire input string be
        successfully parsed, then set ``parseAll`` to True (equivalent to ending
        the grammar with ``StringEnd()``).

        Note: ``parseString`` implicitly calls ``expandtabs()`` on the input string,
        in order to report proper column numbers in parse actions.
        If the input string contains tabs and
        the grammar uses parse actions that use the ``loc`` argument to index into the
        string being parsed, you can ensure you have a consistent view of the input
        string by:

        - calling ``parseWithTabs`` on your grammar before calling ``parseString``
          (see :class:`parseWithTabs`)
        - define your parse action using the full ``(s, loc, toks)`` signature, and
          reference the input string using the parse action's ``s`` argument
        - explictly expand the tabs in your input string before calling
          ``parseString``

        Example::

            Word('a').parseString('aaaaabaaa')  # -> ['aaaaa']
            Word('a').parseString('aaaaabaaa', parseAll=True)  # -> Exception: Expected end of text
        """
        ParserElement.resetCache()
        if not self.streamlined:
            self.streamline()
            # ~ self.saveAsList = True
        for e in self.ignoreExprs:
            e.streamline()
        if not self.keepTabs:
            instring = instring.expandtabs()
        try:
            loc, tokens = self._parse(instring, 0)
            if parseAll:
                loc = self.preParse(instring, loc)
                se = Empty() + StringEnd()
                se._parse(instring, loc)
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clearing out pyparsing internal stack trace
                if getattr(exc, '__traceback__', None) is not None:
                    exc.__traceback__ = self._trim_traceback(exc.__traceback__)
                raise exc
        else:
            return tokens

    def scanString(self, instring, maxMatches=_MAX_INT, overlap=False):
        """
        Scan the input string for expression matches.  Each match will return the
        matching tokens, start location, and end location.  May be called with optional
        ``maxMatches`` argument, to clip scanning after 'n' matches are found.  If
        ``overlap`` is specified, then overlapping matches will be reported.

        Note that the start and end locations are reported relative to the string
        being parsed.  See :class:`parseString` for more information on parsing
        strings with embedded tabs.

        Example::

            source = "sldjf123lsdjjkf345sldkjf879lkjsfd987"
            print(source)
            for tokens, start, end in Word(alphas).scanString(source):
                print(' '*start + '^'*(end-start))
                print(' '*start + tokens[0])

        prints::

            sldjf123lsdjjkf345sldkjf879lkjsfd987
            ^^^^^
            sldjf
                    ^^^^^^^
                    lsdjjkf
                              ^^^^^^
                              sldkjf
                                       ^^^^^^
                                       lkjsfd
        """
        if not self.streamlined:
            self.streamline()
        for e in self.ignoreExprs:
            e.streamline()

        if not self.keepTabs:
            instring = _ustr(instring).expandtabs()
        instrlen = len(instring)
        loc = 0
        preparseFn = self.preParse
        parseFn = self._parse
        ParserElement.resetCache()
        matches = 0
        try:
            while loc <= instrlen and matches < maxMatches:
                try:
                    preloc = preparseFn(instring, loc)
                    nextLoc, tokens = parseFn(instring, preloc, callPreParse=False)
                except ParseException:
                    loc = preloc + 1
                else:
                    if nextLoc > loc:
                        matches += 1
                        yield tokens, preloc, nextLoc
                        if overlap:
                            nextloc = preparseFn(instring, loc)
                            if nextloc > loc:
                                loc = nextLoc
                            else:
                                loc += 1
                        else:
                            loc = nextLoc
                    else:
                        loc = preloc + 1
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clearing out pyparsing internal stack trace
                if getattr(exc, '__traceback__', None) is not None:
                    exc.__traceback__ = self._trim_traceback(exc.__traceback__)
                raise exc

    def transformString(self, instring):
        """
        Extension to :class:`scanString`, to modify matching text with modified tokens that may
        be returned from a parse action.  To use ``transformString``, define a grammar and
        attach a parse action to it that modifies the returned token list.
        Invoking ``transformString()`` on a target string will then scan for matches,
        and replace the matched text patterns according to the logic in the parse
        action.  ``transformString()`` returns the resulting transformed string.

        Example::

            wd = Word(alphas)
            wd.setParseAction(lambda toks: toks[0].title())

            print(wd.transformString("now is the winter of our discontent made glorious summer by this sun of york."))

        prints::

            Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York.
        """
        out = []
        lastE = 0
        # force preservation of <TAB>s, to minimize unwanted transformation of string, and to
        # keep string locs straight between transformString and scanString
        self.keepTabs = True
        try:
            for t, s, e in self.scanString(instring):
                out.append(instring[lastE:s])
                if t:
                    if isinstance(t, ParseResults):
                        out += t.asList()
                    elif isinstance(t, list):
                        out += t
                    else:
                        out.append(t)
                lastE = e
            out.append(instring[lastE:])
            out = [o for o in out if o]
            return "".join(map(_ustr, _flatten(out)))
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clearing out pyparsing internal stack trace
                if getattr(exc, '__traceback__', None) is not None:
                    exc.__traceback__ = self._trim_traceback(exc.__traceback__)
                raise exc

    def searchString(self, instring, maxMatches=_MAX_INT):
        """
        Another extension to :class:`scanString`, simplifying the access to the tokens found
        to match the given parse expression.  May be called with optional
        ``maxMatches`` argument, to clip searching after 'n' matches are found.

        Example::

            # a capitalized word starts with an uppercase letter, followed by zero or more lowercase letters
            cap_word = Word(alphas.upper(), alphas.lower())

            print(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity"))

            # the sum() builtin can be used to merge results into a single ParseResults object
            print(sum(cap_word.searchString("More than Iron, more than Lead, more than Gold I need Electricity")))

        prints::

            [['More'], ['Iron'], ['Lead'], ['Gold'], ['I'], ['Electricity']]
            ['More', 'Iron', 'Lead', 'Gold', 'I', 'Electricity']
        """
        try:
            return ParseResults([t for t, s, e in self.scanString(instring, maxMatches)])
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clearing out pyparsing internal stack trace
                if getattr(exc, '__traceback__', None) is not None:
                    exc.__traceback__ = self._trim_traceback(exc.__traceback__)
                raise exc

    def split(self, instring, maxsplit=_MAX_INT, includeSeparators=False):
        """
        Generator method to split a string using the given expression as a separator.
        May be called with optional ``maxsplit`` argument, to limit the number of splits;
        and the optional ``includeSeparators`` argument (default= ``False``), if the separating
        matching text should be included in the split results.

        Example::

            punc = oneOf(list(".,;:/-!?"))
            print(list(punc.split("This, this?, this sentence, is badly punctuated!")))

        prints::

            ['This', ' this', '', ' this sentence', ' is badly punctuated', '']
        """
        splits = 0
        last = 0
        for t, s, e in self.scanString(instring, maxMatches=maxsplit):
            yield instring[last:s]
            if includeSeparators:
                yield t[0]
            last = e
        yield instring[last:]

    def __add__(self, other):
        """
        Implementation of + operator - returns :class:`And`. Adding strings to a ParserElement
        converts them to :class:`Literal`s by default.

        Example::

            greet = Word(alphas) + "," + Word(alphas) + "!"
            hello = "Hello, World!"
            print (hello, "->", greet.parseString(hello))

        prints::

            Hello, World! -> ['Hello', ',', 'World', '!']

        ``...`` may be used as a parse expression as a short form of :class:`SkipTo`.

            Literal('start') + ... + Literal('end')

        is equivalent to:

            Literal('start') + SkipTo('end')("_skipped*") + Literal('end')

        Note that the skipped text is returned with '_skipped' as a results name,
        and to support having multiple skips in the same parser, the value returned is
        a list of all skipped text.
        """
        if other is Ellipsis:
            return _PendingSkip(self)

        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return And([self, other])

    def __radd__(self, other):
        """
        Implementation of + operator when left operand is not a :class:`ParserElement`
        """
        if other is Ellipsis:
            return SkipTo(self)("_skipped*") + self

        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return other + self

    def __sub__(self, other):
        """
        Implementation of - operator, returns :class:`And` with error stop
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return self + And._ErrorStop() + other

    def __rsub__(self, other):
        """
        Implementation of - operator when left operand is not a :class:`ParserElement`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return other - self

    def __mul__(self, other):
        """
        Implementation of * operator, allows use of ``expr * 3`` in place of
        ``expr + expr + expr``.  Expressions may also me multiplied by a 2-integer
        tuple, similar to ``{min, max}`` multipliers in regular expressions.  Tuples
        may also include ``None`` as in:
         - ``expr*(n, None)`` or ``expr*(n, )`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr*(None, n)`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr*(None, None)`` is equivalent to ``ZeroOrMore(expr)``
         - ``expr*(1, None)`` is equivalent to ``OneOrMore(expr)``

        Note that ``expr*(None, n)`` does not raise an exception if
        more than n exprs exist in the input stream; that is,
        ``expr*(None, n)`` does not enforce a maximum number of expr
        occurrences.  If this behavior is desired, then write
        ``expr*(None, n) + ~expr``
        """
        if other is Ellipsis:
            other = (0, None)
        elif isinstance(other, tuple) and other[:1] == (Ellipsis,):
            other = ((0, ) + other[1:] + (None,))[:2]

        if isinstance(other, int):
            minElements, optElements = other, 0
        elif isinstance(other, tuple):
            other = tuple(o if o is not Ellipsis else None for o in other)
            other = (other + (None, None))[:2]
            if other[0] is None:
                other = (0, other[1])
            if isinstance(other[0], int) and other[1] is None:
                if other[0] == 0:
                    return ZeroOrMore(self)
                if other[0] == 1:
                    return OneOrMore(self)
                else:
                    return self * other[0] + ZeroOrMore(self)
            elif isinstance(other[0], int) and isinstance(other[1], int):
                minElements, optElements = other
                optElements -= minElements
            else:
                raise TypeError("cannot multiply 'ParserElement' and ('%s', '%s') objects", type(other[0]), type(other[1]))
        else:
            raise TypeError("cannot multiply 'ParserElement' and '%s' objects", type(other))

        if minElements < 0:
            raise ValueError("cannot multiply ParserElement by negative value")
        if optElements < 0:
            raise ValueError("second tuple value must be greater or equal to first tuple value")
        if minElements == optElements == 0:
            raise ValueError("cannot multiply ParserElement by 0 or (0, 0)")

        if optElements:
            def makeOptionalList(n):
                if n > 1:
                    return Optional(self + makeOptionalList(n - 1))
                else:
                    return Optional(self)
            if minElements:
                if minElements == 1:
                    ret = self + makeOptionalList(optElements)
                else:
                    ret = And([self] * minElements) + makeOptionalList(optElements)
            else:
                ret = makeOptionalList(optElements)
        else:
            if minElements == 1:
                ret = self
            else:
                ret = And([self] * minElements)
        return ret

    def __rmul__(self, other):
        return self.__mul__(other)

    def __or__(self, other):
        """
        Implementation of | operator - returns :class:`MatchFirst`
        """
        if other is Ellipsis:
            return _PendingSkip(self, must_skip=True)

        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return MatchFirst([self, other])

    def __ror__(self, other):
        """
        Implementation of | operator when left operand is not a :class:`ParserElement`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return other | self

    def __xor__(self, other):
        """
        Implementation of ^ operator - returns :class:`Or`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return Or([self, other])

    def __rxor__(self, other):
        """
        Implementation of ^ operator when left operand is not a :class:`ParserElement`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return other ^ self

    def __and__(self, other):
        """
        Implementation of & operator - returns :class:`Each`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return Each([self, other])

    def __rand__(self, other):
        """
        Implementation of & operator when left operand is not a :class:`ParserElement`
        """
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        if not isinstance(other, ParserElement):
            warnings.warn("Cannot combine element of type %s with ParserElement" % type(other),
                          SyntaxWarning, stacklevel=2)
            return None
        return other & self

    def __invert__(self):
        """
        Implementation of ~ operator - returns :class:`NotAny`
        """
        return NotAny(self)

    def __iter__(self):
        # must implement __iter__ to override legacy use of sequential access to __getitem__ to
        # iterate over a sequence
        raise TypeError('%r object is not iterable' % self.__class__.__name__)

    def __getitem__(self, key):
        """
        use ``[]`` indexing notation as a short form for expression repetition:
         - ``expr[n]`` is equivalent to ``expr*n``
         - ``expr[m, n]`` is equivalent to ``expr*(m, n)``
         - ``expr[n, ...]`` or ``expr[n,]`` is equivalent
              to ``expr*n + ZeroOrMore(expr)``
              (read as "at least n instances of ``expr``")
         - ``expr[..., n]`` is equivalent to ``expr*(0, n)``
              (read as "0 to n instances of ``expr``")
         - ``expr[...]`` and ``expr[0, ...]`` are equivalent to ``ZeroOrMore(expr)``
         - ``expr[1, ...]`` is equivalent to ``OneOrMore(expr)``
         ``None`` may be used in place of ``...``.

        Note that ``expr[..., n]`` and ``expr[m, n]``do not raise an exception
        if more than ``n`` ``expr``s exist in the input stream.  If this behavior is
        desired, then write ``expr[..., n] + ~expr``.
       """

        # convert single arg keys to tuples
        try:
            if isinstance(key, str):
                key = (key,)
            iter(key)
        except TypeError:
            key = (key, key)

        if len(key) > 2:
            warnings.warn("only 1 or 2 index arguments supported ({0}{1})".format(key[:5],
                                                                                '... [{0}]'.format(len(key))
                                                                                if len(key) > 5 else ''))

        # clip to 2 elements
        ret = self * tuple(key[:2])
        return ret

    def __call__(self, name=None):
        """
        Shortcut for :class:`setResultsName`, with ``listAllMatches=False``.

        If ``name`` is given with a trailing ``'*'`` character, then ``listAllMatches`` will be
        passed as ``True``.

        If ``name` is omitted, same as calling :class:`copy`.

        Example::

            # these are equivalent
            userdata = Word(alphas).setResultsName("name") + Word(nums + "-").setResultsName("socsecno")
            userdata = Word(alphas)("name") + Word(nums + "-")("socsecno")
        """
        if name is not None:
            return self._setResultsName(name)
        else:
            return self.copy()

    def suppress(self):
        """
        Suppresses the output of this :class:`ParserElement`; useful to keep punctuation from
        cluttering up returned output.
        """
        return Suppress(self)

    def leaveWhitespace(self):
        """
        Disables the skipping of whitespace before matching the characters in the
        :class:`ParserElement`'s defined pattern.  This is normally only used internally by
        the pyparsing module, but may be needed in some whitespace-sensitive grammars.
        """
        self.skipWhitespace = False
        return self

    def setWhitespaceChars(self, chars):
        """
        Overrides the default whitespace chars
        """
        self.skipWhitespace = True
        self.whiteChars = chars
        self.copyDefaultWhiteChars = False
        return self

    def parseWithTabs(self):
        """
        Overrides default behavior to expand ``<TAB>``s to spaces before parsing the input string.
        Must be called before ``parseString`` when the input grammar contains elements that
        match ``<TAB>`` characters.
        """
        self.keepTabs = True
        return self

    def ignore(self, other):
        """
        Define expression to be ignored (e.g., comments) while doing pattern
        matching; may be called repeatedly, to define multiple comment or other
        ignorable patterns.

        Example::

            patt = OneOrMore(Word(alphas))
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj']

            patt.ignore(cStyleComment)
            patt.parseString('ablaj /* comment */ lskjd') # -> ['ablaj', 'lskjd']
        """
        if isinstance(other, basestring):
            other = Suppress(other)

        if isinstance(other, Suppress):
            if other not in self.ignoreExprs:
                self.ignoreExprs.append(other)
        else:
            self.ignoreExprs.append(Suppress(other.copy()))
        return self

    def setDebugActions(self, startAction, successAction, exceptionAction):
        """
        Enable display of debugging messages while doing pattern matching.
        """
        self.debugActions = (startAction or _defaultStartDebugAction,
                             successAction or _defaultSuccessDebugAction,
                             exceptionAction or _defaultExceptionDebugAction)
        self.debug = True
        return self

    def setDebug(self, flag=True):
        """
        Enable display of debugging messages while doing pattern matching.
        Set ``flag`` to True to enable, False to disable.

        Example::

            wd = Word(alphas).setName("alphaword")
            integer = Word(nums).setName("numword")
            term = wd | integer

            # turn on debugging for wd
            wd.setDebug()

            OneOrMore(term).parseString("abc 123 xyz 890")

        prints::

            Match alphaword at loc 0(1,1)
            Matched alphaword -> ['abc']
            Match alphaword at loc 3(1,4)
            Exception raised:Expected alphaword (at char 4), (line:1, col:5)
            Match alphaword at loc 7(1,8)
            Matched alphaword -> ['xyz']
            Match alphaword at loc 11(1,12)
            Exception raised:Expected alphaword (at char 12), (line:1, col:13)
            Match alphaword at loc 15(1,16)
            Exception raised:Expected alphaword (at char 15), (line:1, col:16)

        The output shown is that produced by the default debug actions - custom debug actions can be
        specified using :class:`setDebugActions`. Prior to attempting
        to match the ``wd`` expression, the debugging message ``"Match <exprname> at loc <n>(<line>,<col>)"``
        is shown. Then if the parse succeeds, a ``"Matched"`` message is shown, or an ``"Exception raised"``
        message is shown. Also note the use of :class:`setName` to assign a human-readable name to the expression,
        which makes debugging and exception messages easier to understand - for instance, the default
        name created for the :class:`Word` expression without calling ``setName`` is ``"W:(ABCD...)"``.
        """
        if flag:
            self.setDebugActions(_defaultStartDebugAction, _defaultSuccessDebugAction, _defaultExceptionDebugAction)
        else:
            self.debug = False
        return self

    def __str__(self):
        return self.name

    def __repr__(self):
        return _ustr(self)

    def streamline(self):
        self.streamlined = True
        self.strRepr = None
        return self

    def checkRecursion(self, parseElementList):
        pass

    def validate(self, validateTrace=None):
        """
        Check defined expressions for valid structure, check for infinite recursive definitions.
        """
        self.checkRecursion([])

    def parseFile(self, file_or_filename, parseAll=False):
        """
        Execute the parse expression on the given file or filename.
        If a filename is specified (instead of a file object),
        the entire file is opened, read, and closed before parsing.
        """
        try:
            file_contents = file_or_filename.read()
        except AttributeError:
            with open(file_or_filename, "r") as f:
                file_contents = f.read()
        try:
            return self.parseString(file_contents, parseAll)
        except ParseBaseException as exc:
            if ParserElement.verbose_stacktrace:
                raise
            else:
                # catch and re-raise exception from here, clearing out pyparsing internal stack trace
                if getattr(exc, '__traceback__', None) is not None:
                    exc.__traceback__ = self._trim_traceback(exc.__traceback__)
                raise exc

    def __eq__(self, other):
        if self is other:
            return True
        elif isinstance(other, basestring):
            return self.matches(other)
        elif isinstance(other, ParserElement):
            return vars(self) == vars(other)
        return False

    def __ne__(self, other):
        return not (self == other)

    def __hash__(self):
        return id(self)

    def __req__(self, other):
        return self == other

    def __rne__(self, other):
        return not (self == other)

    def matches(self, testString, parseAll=True):
        """
        Method for quick testing of a parser against a test string. Good for simple
        inline microtests of sub expressions while building up larger parser.

        Parameters:
         - testString - to test against this expression for a match
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests

        Example::

            expr = Word(nums)
            assert expr.matches("100")
        """
        try:
            self.parseString(_ustr(testString), parseAll=parseAll)
            return True
        except ParseBaseException:
            return False

    def runTests(self, tests, parseAll=True, comment='#',
                 fullDump=True, printResults=True, failureTests=False, postParse=None,
                 file=None):
        """
        Execute the parse expression on a series of test strings, showing each
        test, the parsed results or where the parse failed. Quick and easy way to
        run a parse expression against a list of sample strings.

        Parameters:
         - tests - a list of separate test strings, or a multiline string of test strings
         - parseAll - (default= ``True``) - flag to pass to :class:`parseString` when running tests
         - comment - (default= ``'#'``) - expression for indicating embedded comments in the test
              string; pass None to disable comment filtering
         - fullDump - (default= ``True``) - dump results as list followed by results names in nested outline;
              if False, only dump nested list
         - printResults - (default= ``True``) prints test output to stdout
         - failureTests - (default= ``False``) indicates if these tests are expected to fail parsing
         - postParse - (default= ``None``) optional callback for successful parse results; called as
              `fn(test_string, parse_results)` and returns a string to be added to the test output
         - file - (default=``None``) optional file-like object to which test output will be written;
              if None, will default to ``sys.stdout``

        Returns: a (success, results) tuple, where success indicates that all tests succeeded
        (or failed if ``failureTests`` is True), and the results contain a list of lines of each
        test's output

        Example::

            number_expr = pyparsing_common.number.copy()

            result = number_expr.runTests('''
                # unsigned integer
                100
                # negative integer
                -100
                # float with scientific notation
                6.02e23
                # integer with scientific notation
                1e-12
                ''')
            print("Success" if result[0] else "Failed!")

            result = number_expr.runTests('''
                # stray character
                100Z
                # missing leading digit before '.'
                -.100
                # too many '.'
                3.14.159
                ''', failureTests=True)
            print("Success" if result[0] else "Failed!")

        prints::

            # unsigned integer
            100
            [100]

            # negative integer
            -100
            [-100]

            # float with scientific notation
            6.02e23
            [6.02e+23]

            # integer with scientific notation
            1e-12
            [1e-12]

            Success

            # stray character
            100Z
               ^
            FAIL: Expected end of text (at char 3), (line:1, col:4)

            # missing leading digit before '.'
            -.100
            ^
            FAIL: Expected {real number with scientific notation | real number | signed integer} (at char 0), (line:1, col:1)

            # too many '.'
            3.14.159
                ^
            FAIL: Expected end of text (at char 4), (line:1, col:5)

            Success

        Each test string must be on a single line. If you want to test a string that spans multiple
        lines, create a test like this::

            expr.runTest(r"this is a test\\n of strings that spans \\n 3 lines")

        (Note that this is a raw string literal, you must include the leading 'r'.)
        """
        if isinstance(tests, basestring):
            tests = list(map(str.strip, tests.rstrip().splitlines()))
        if isinstance(comment, basestring):
            comment = Literal(comment)
        if file is None:
            file = sys.stdout
        print_ = file.write

        allResults = []
        comments = []
        success = True
        NL = Literal(r'\n').addParseAction(replaceWith('\n')).ignore(quotedString)
        BOM = u'\ufeff'
        for t in tests:
            if comment is not None and comment.matches(t, False) or comments and not t:
                comments.append(t)
                continue
            if not t:
                continue
            out = ['\n' + '\n'.join(comments) if comments else '', t]
            comments = []
            try:
                # convert newline marks to actual newlines, and strip leading BOM if present
                t = NL.transformString(t.lstrip(BOM))
                result = self.parseString(t, parseAll=parseAll)
            except ParseBaseException as pe:
                fatal = "(FATAL)" if isinstance(pe, ParseFatalException) else ""
                if '\n' in t:
                    out.append(line(pe.loc, t))
                    out.append(' ' * (col(pe.loc, t) - 1) + '^' + fatal)
                else:
                    out.append(' ' * pe.loc + '^' + fatal)
                out.append("FAIL: " + str(pe))
                success = success and failureTests
                result = pe
            except Exception as exc:
                out.append("FAIL-EXCEPTION: " + str(exc))
                success = success and failureTests
                result = exc
            else:
                success = success and not failureTests
                if postParse is not None:
                    try:
                        pp_value = postParse(t, result)
                        if pp_value is not None:
                            if isinstance(pp_value, ParseResults):
                                out.append(pp_value.dump())
                            else:
                                out.append(str(pp_value))
                        else:
                            out.append(result.dump())
                    except Exception as e:
                        out.append(result.dump(full=fullDump))
                        out.append("{0} failed: {1}: {2}".format(postParse.__name__, type(e).__name__, e))
                else:
                    out.append(result.dump(full=fullDump))

            if printResults:
                if fullDump:
                    out.append('')
                print_('\n'.join(out))

            allResults.append((t, result))

        return success, allResults


class _PendingSkip(ParserElement):
    # internal placeholder class to hold a place were '...' is added to a parser element,
    # once another ParserElement is added, this placeholder will be replaced with a SkipTo
    def __init__(self, expr, must_skip=False):
        super(_PendingSkip, self).__init__()
        self.strRepr = str(expr + Empty()).replace('Empty', '...')
        self.name = self.strRepr
        self.anchor = expr
        self.must_skip = must_skip

    def __add__(self, other):
        skipper = SkipTo(other).setName("...")("_skipped*")
        if self.must_skip:
            def must_skip(t):
                if not t._skipped or t._skipped.asList() == ['']:
                    del t[0]
                    t.pop("_skipped", None)
            def show_skip(t):
                if t._skipped.asList()[-1:] == ['']:
                    skipped = t.pop('_skipped')
                    t['_skipped'] = 'missing <' + repr(self.anchor) + '>'
            return (self.anchor + skipper().addParseAction(must_skip)
                    | skipper().addParseAction(show_skip)) + other

        return self.anchor + skipper + other

    def __repr__(self):
        return self.strRepr

    def parseImpl(self, *args):
        raise Exception("use of `...` expression without following SkipTo target expression")


class Token(ParserElement):
    """Abstract :class:`ParserElement` subclass, for defining atomic
    matching patterns.
    """
    def __init__(self):
        super(Token, self).__init__(savelist=False)


class Empty(Token):
    """An empty token, will always match.
    """
    def __init__(self):
        super(Empty, self).__init__()
        self.name = "Empty"
        self.mayReturnEmpty = True
        self.mayIndexError = False


class NoMatch(Token):
    """A token that will never match.
    """
    def __init__(self):
        super(NoMatch, self).__init__()
        self.name = "NoMatch"
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.errmsg = "Unmatchable token"

    def parseImpl(self, instring, loc, doActions=True):
        raise ParseException(instring, loc, self.errmsg, self)


class Literal(Token):
    """Token to exactly match a specified string.

    Example::

        Literal('blah').parseString('blah')  # -> ['blah']
        Literal('blah').parseString('blahfooblah')  # -> ['blah']
        Literal('blah').parseString('bla')  # -> Exception: Expected "blah"

    For case-insensitive matching, use :class:`CaselessLiteral`.

    For keyword matching (force word break before and after the matched string),
    use :class:`Keyword` or :class:`CaselessKeyword`.
    """
    def __init__(self, matchString):
        super(Literal, self).__init__()
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Literal; use Empty() instead",
                            SyntaxWarning, stacklevel=2)
            self.__class__ = Empty
        self.name = '"%s"' % _ustr(self.match)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False

        # Performance tuning: modify __class__ to select
        # a parseImpl optimized for single-character check
        if self.matchLen == 1 and type(self) is Literal:
            self.__class__ = _SingleCharLiteral

    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc] == self.firstMatchChar and instring.startswith(self.match, loc):
            return loc + self.matchLen, self.match
        raise ParseException(instring, loc, self.errmsg, self)

class _SingleCharLiteral(Literal):
    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc] == self.firstMatchChar:
            return loc + 1, self.match
        raise ParseException(instring, loc, self.errmsg, self)

_L = Literal
ParserElement._literalStringClass = Literal

class Keyword(Token):
    """Token to exactly match a specified string as a keyword, that is,
    it must be immediately followed by a non-keyword character.  Compare
    with :class:`Literal`:

     - ``Literal("if")`` will match the leading ``'if'`` in
       ``'ifAndOnlyIf'``.
     - ``Keyword("if")`` will not; it will only match the leading
       ``'if'`` in ``'if x=1'``, or ``'if(y==2)'``

    Accepts two optional constructor arguments in addition to the
    keyword string:

     - ``identChars`` is a string of characters that would be valid
       identifier characters, defaulting to all alphanumerics + "_" and
       "$"
     - ``caseless`` allows case-insensitive matching, default is ``False``.

    Example::

        Keyword("start").parseString("start")  # -> ['start']
        Keyword("start").parseString("starting")  # -> Exception

    For case-insensitive matching, use :class:`CaselessKeyword`.
    """
    DEFAULT_KEYWORD_CHARS = alphanums + "_$"

    def __init__(self, matchString, identChars=None, caseless=False):
        super(Keyword, self).__init__()
        if identChars is None:
            identChars = Keyword.DEFAULT_KEYWORD_CHARS
        self.match = matchString
        self.matchLen = len(matchString)
        try:
            self.firstMatchChar = matchString[0]
        except IndexError:
            warnings.warn("null string passed to Keyword; use Empty() instead",
                          SyntaxWarning, stacklevel=2)
        self.name = '"%s"' % self.match
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = False
        self.mayIndexError = False
        self.caseless = caseless
        if caseless:
            self.caselessmatch = matchString.upper()
            identChars = identChars.upper()
        self.identChars = set(identChars)

    def parseImpl(self, instring, loc, doActions=True):
        if self.caseless:
            if ((instring[loc:loc + self.matchLen].upper() == self.caselessmatch)
                    and (loc >= len(instring) - self.matchLen
                         or instring[loc + self.matchLen].upper() not in self.identChars)
                    and (loc == 0
                         or instring[loc - 1].upper() not in self.identChars)):
                return loc + self.matchLen, self.match

        else:
            if instring[loc] == self.firstMatchChar:
                if ((self.matchLen == 1 or instring.startswith(self.match, loc))
                        and (loc >= len(instring) - self.matchLen
                             or instring[loc + self.matchLen] not in self.identChars)
                        and (loc == 0 or instring[loc - 1] not in self.identChars)):
                    return loc + self.matchLen, self.match

        raise ParseException(instring, loc, self.errmsg, self)

    def copy(self):
        c = super(Keyword, self).copy()
        c.identChars = Keyword.DEFAULT_KEYWORD_CHARS
        return c

    @staticmethod
    def setDefaultKeywordChars(chars):
        """Overrides the default Keyword chars
        """
        Keyword.DEFAULT_KEYWORD_CHARS = chars

class CaselessLiteral(Literal):
    """Token to match a specified string, ignoring case of letters.
    Note: the matched results will always be in the case of the given
    match string, NOT the case of the input text.

    Example::

        OneOrMore(CaselessLiteral("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD', 'CMD']

    (Contrast with example for :class:`CaselessKeyword`.)
    """
    def __init__(self, matchString):
        super(CaselessLiteral, self).__init__(matchString.upper())
        # Preserve the defining literal.
        self.returnString = matchString
        self.name = "'%s'" % self.returnString
        self.errmsg = "Expected " + self.name

    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc:loc + self.matchLen].upper() == self.match:
            return loc + self.matchLen, self.returnString
        raise ParseException(instring, loc, self.errmsg, self)

class CaselessKeyword(Keyword):
    """
    Caseless version of :class:`Keyword`.

    Example::

        OneOrMore(CaselessKeyword("CMD")).parseString("cmd CMD Cmd10") # -> ['CMD', 'CMD']

    (Contrast with example for :class:`CaselessLiteral`.)
    """
    def __init__(self, matchString, identChars=None):
        super(CaselessKeyword, self).__init__(matchString, identChars, caseless=True)

class CloseMatch(Token):
    """A variation on :class:`Literal` which matches "close" matches,
    that is, strings with at most 'n' mismatching characters.
    :class:`CloseMatch` takes parameters:

     - ``match_string`` - string to be matched
     - ``maxMismatches`` - (``default=1``) maximum number of
       mismatches allowed to count as a match

    The results from a successful parse will contain the matched text
    from the input string and the following named results:

     - ``mismatches`` - a list of the positions within the
       match_string where mismatches were found
     - ``original`` - the original match_string used to compare
       against the input string

    If ``mismatches`` is an empty list, then the match was an exact
    match.

    Example::

        patt = CloseMatch("ATCATCGAATGGA")
        patt.parseString("ATCATCGAAXGGA") # -> (['ATCATCGAAXGGA'], {'mismatches': [[9]], 'original': ['ATCATCGAATGGA']})
        patt.parseString("ATCAXCGAAXGGA") # -> Exception: Expected 'ATCATCGAATGGA' (with up to 1 mismatches) (at char 0), (line:1, col:1)

        # exact match
        patt.parseString("ATCATCGAATGGA") # -> (['ATCATCGAATGGA'], {'mismatches': [[]], 'original': ['ATCATCGAATGGA']})

        # close match allowing up to 2 mismatches
        patt = CloseMatch("ATCATCGAATGGA", maxMismatches=2)
        patt.parseString("ATCAXCGAAXGGA") # -> (['ATCAXCGAAXGGA'], {'mismatches': [[4, 9]], 'original': ['ATCATCGAATGGA']})
    """
    def __init__(self, match_string, maxMismatches=1):
        super(CloseMatch, self).__init__()
        self.name = match_string
        self.match_string = match_string
        self.maxMismatches = maxMismatches
        self.errmsg = "Expected %r (with up to %d mismatches)" % (self.match_string, self.maxMismatches)
        self.mayIndexError = False
        self.mayReturnEmpty = False

    def parseImpl(self, instring, loc, doActions=True):
        start = loc
        instrlen = len(instring)
        maxloc = start + len(self.match_string)

        if maxloc <= instrlen:
            match_string = self.match_string
            match_stringloc = 0
            mismatches = []
            maxMismatches = self.maxMismatches

            for match_stringloc, s_m in enumerate(zip(instring[loc:maxloc], match_string)):
                src, mat = s_m
                if src != mat:
                    mismatches.append(match_stringloc)
                    if len(mismatches) > maxMismatches:
                        break
            else:
                loc = match_stringloc + 1
                results = ParseResults([instring[start:loc]])
                results['original'] = match_string
                results['mismatches'] = mismatches
                return loc, results

        raise ParseException(instring, loc, self.errmsg, self)


class Word(Token):
    """Token for matching words composed of allowed character sets.
    Defined with string containing all allowed initial characters, an
    optional string containing allowed body characters (if omitted,
    defaults to the initial character set), and an optional minimum,
    maximum, and/or exact length.  The default value for ``min`` is
    1 (a minimum value < 1 is not valid); the default values for
    ``max`` and ``exact`` are 0, meaning no maximum or exact
    length restriction. An optional ``excludeChars`` parameter can
    list characters that might be found in the input ``bodyChars``
    string; useful to define a word of all printables except for one or
    two characters, for instance.

    :class:`srange` is useful for defining custom character set strings
    for defining ``Word`` expressions, using range notation from
    regular expression character sets.

    A common mistake is to use :class:`Word` to match a specific literal
    string, as in ``Word("Address")``. Remember that :class:`Word`
    uses the string argument to define *sets* of matchable characters.
    This expression would match "Add", "AAA", "dAred", or any other word
    made up of the characters 'A', 'd', 'r', 'e', and 's'. To match an
    exact literal string, use :class:`Literal` or :class:`Keyword`.

    pyparsing includes helper strings for building Words:

     - :class:`alphas`
     - :class:`nums`
     - :class:`alphanums`
     - :class:`hexnums`
     - :class:`alphas8bit` (alphabetic characters in ASCII range 128-255
       - accented, tilded, umlauted, etc.)
     - :class:`punc8bit` (non-alphabetic characters in ASCII range
       128-255 - currency, symbols, superscripts, diacriticals, etc.)
     - :class:`printables` (any non-whitespace character)

    Example::

        # a word composed of digits
        integer = Word(nums) # equivalent to Word("0123456789") or Word(srange("0-9"))

        # a word with a leading capital, and zero or more lowercase
        capital_word = Word(alphas.upper(), alphas.lower())

        # hostnames are alphanumeric, with leading alpha, and '-'
        hostname = Word(alphas, alphanums + '-')

        # roman numeral (not a strict parser, accepts invalid mix of characters)
        roman = Word("IVXLCDM")

        # any string of non-whitespace characters, except for ','
        csv_value = Word(printables, excludeChars=",")
    """
    def __init__(self, initChars, bodyChars=None, min=1, max=0, exact=0, asKeyword=False, excludeChars=None):
        super(Word, self).__init__()
        if excludeChars:
            excludeChars = set(excludeChars)
            initChars = ''.join(c for c in initChars if c not in excludeChars)
            if bodyChars:
                bodyChars = ''.join(c for c in bodyChars if c not in excludeChars)
        self.initCharsOrig = initChars
        self.initChars = set(initChars)
        if bodyChars:
            self.bodyCharsOrig = bodyChars
            self.bodyChars = set(bodyChars)
        else:
            self.bodyCharsOrig = initChars
            self.bodyChars = set(initChars)

        self.maxSpecified = max > 0

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use Optional(Word()) if zero-length word is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.asKeyword = asKeyword

        if ' ' not in self.initCharsOrig + self.bodyCharsOrig and (min == 1 and max == 0 and exact == 0):
            if self.bodyCharsOrig == self.initCharsOrig:
                self.reString = "[%s]+" % _escapeRegexRangeChars(self.initCharsOrig)
            elif len(self.initCharsOrig) == 1:
                self.reString = "%s[%s]*" % (re.escape(self.initCharsOrig),
                                             _escapeRegexRangeChars(self.bodyCharsOrig),)
            else:
                self.reString = "[%s][%s]*" % (_escapeRegexRangeChars(self.initCharsOrig),
                                               _escapeRegexRangeChars(self.bodyCharsOrig),)
            if self.asKeyword:
                self.reString = r"\b" + self.reString + r"\b"

            try:
                self.re = re.compile(self.reString)
            except Exception:
                self.re = None
            else:
                self.re_match = self.re.match
                self.__class__ = _WordRegex

    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc] not in self.initChars:
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        instrlen = len(instring)
        bodychars = self.bodyChars
        maxloc = start + self.maxLen
        maxloc = min(maxloc, instrlen)
        while loc < maxloc and instring[loc] in bodychars:
            loc += 1

        throwException = False
        if loc - start < self.minLen:
            throwException = True
        elif self.maxSpecified and loc < instrlen and instring[loc] in bodychars:
            throwException = True
        elif self.asKeyword:
            if (start > 0 and instring[start - 1] in bodychars
                    or loc < instrlen and instring[loc] in bodychars):
                throwException = True

        if throwException:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__(self):
        try:
            return super(Word, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:

            def charsAsStr(s):
                if len(s) > 4:
                    return s[:4] + "..."
                else:
                    return s

            if self.initCharsOrig != self.bodyCharsOrig:
                self.strRepr = "W:(%s, %s)" % (charsAsStr(self.initCharsOrig), charsAsStr(self.bodyCharsOrig))
            else:
                self.strRepr = "W:(%s)" % charsAsStr(self.initCharsOrig)

        return self.strRepr

class _WordRegex(Word):
    def parseImpl(self, instring, loc, doActions=True):
        result = self.re_match(instring, loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        return loc, result.group()


class Char(_WordRegex):
    """A short-cut class for defining ``Word(characters, exact=1)``,
    when defining a match of any single character in a string of
    characters.
    """
    def __init__(self, charset, asKeyword=False, excludeChars=None):
        super(Char, self).__init__(charset, exact=1, asKeyword=asKeyword, excludeChars=excludeChars)
        self.reString = "[%s]" % _escapeRegexRangeChars(''.join(self.initChars))
        if asKeyword:
            self.reString = r"\b%s\b" % self.reString
        self.re = re.compile(self.reString)
        self.re_match = self.re.match


class Regex(Token):
    r"""Token for matching strings that match a given regular
    expression. Defined with string specifying the regular expression in
    a form recognized by the stdlib Python  `re module <https://docs.python.org/3/library/re.html>`_.
    If the given regex contains named groups (defined using ``(?P<name>...)``),
    these will be preserved as named parse results.

    If instead of the Python stdlib re module you wish to use a different RE module
    (such as the `regex` module), you can replace it by either building your
    Regex object with a compiled RE that was compiled using regex:

    Example::

        realnum = Regex(r"[+-]?\d+\.\d*")
        date = Regex(r'(?P<year>\d{4})-(?P<month>\d\d?)-(?P<day>\d\d?)')
        # ref: https://stackoverflow.com/questions/267399/how-do-you-match-only-valid-roman-numerals-with-a-regular-expression
        roman = Regex(r"M{0,4}(CM|CD|D?{0,3})(XC|XL|L?X{0,3})(IX|IV|V?I{0,3})")

        # use regex module instead of stdlib re module to construct a Regex using
        # a compiled regular expression
        import regex
        parser = pp.Regex(regex.compile(r'[0-9]'))

    """
    def __init__(self, pattern, flags=0, asGroupList=False, asMatch=False):
        """The parameters ``pattern`` and ``flags`` are passed
        to the ``re.compile()`` function as-is. See the Python
        `re module <https://docs.python.org/3/library/re.html>`_ module for an
        explanation of the acceptable patterns and flags.
        """
        super(Regex, self).__init__()

        if isinstance(pattern, basestring):
            if not pattern:
                warnings.warn("null string passed to Regex; use Empty() instead",
                              SyntaxWarning, stacklevel=2)

            self.pattern = pattern
            self.flags = flags

            try:
                self.re = re.compile(self.pattern, self.flags)
                self.reString = self.pattern
            except sre_constants.error:
                warnings.warn("invalid pattern (%s) passed to Regex" % pattern,
                              SyntaxWarning, stacklevel=2)
                raise

        elif hasattr(pattern, 'pattern') and hasattr(pattern, 'match'):
            self.re = pattern
            self.pattern = self.reString = pattern.pattern
            self.flags = flags

        else:
            raise TypeError("Regex may only be constructed with a string or a compiled RE object")

        self.re_match = self.re.match

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = self.re_match("") is not None
        self.asGroupList = asGroupList
        self.asMatch = asMatch
        if self.asGroupList:
            self.parseImpl = self.parseImplAsGroupList
        if self.asMatch:
            self.parseImpl = self.parseImplAsMatch

    def parseImpl(self, instring, loc, doActions=True):
        result = self.re_match(instring, loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = ParseResults(result.group())
        d = result.groupdict()
        if d:
            for k, v in d.items():
                ret[k] = v
        return loc, ret

    def parseImplAsGroupList(self, instring, loc, doActions=True):
        result = self.re_match(instring, loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = result.groups()
        return loc, ret

    def parseImplAsMatch(self, instring, loc, doActions=True):
        result = self.re_match(instring, loc)
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = result
        return loc, ret

    def __str__(self):
        try:
            return super(Regex, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "Re:(%s)" % repr(self.pattern)

        return self.strRepr

    def sub(self, repl):
        r"""
        Return Regex with an attached parse action to transform the parsed
        result as if called using `re.sub(expr, repl, string) <https://docs.python.org/3/library/re.html#re.sub>`_.

        Example::

            make_html = Regex(r"(\w+):(.*?):").sub(r"<\1>\2</\1>")
            print(make_html.transformString("h1:main title:"))
            # prints "<h1>main title</h1>"
        """
        if self.asGroupList:
            warnings.warn("cannot use sub() with Regex(asGroupList=True)",
                          SyntaxWarning, stacklevel=2)
            raise SyntaxError()

        if self.asMatch and callable(repl):
            warnings.warn("cannot use sub() with a callable with Regex(asMatch=True)",
                          SyntaxWarning, stacklevel=2)
            raise SyntaxError()

        if self.asMatch:
            def pa(tokens):
                return tokens[0].expand(repl)
        else:
            def pa(tokens):
                return self.re.sub(repl, tokens[0])
        return self.addParseAction(pa)

class QuotedString(Token):
    r"""
    Token for matching strings that are delimited by quoting characters.

    Defined with the following parameters:

        - quoteChar - string of one or more characters defining the
          quote delimiting string
        - escChar - character to escape quotes, typically backslash
          (default= ``None``)
        - escQuote - special quote sequence to escape an embedded quote
          string (such as SQL's ``""`` to escape an embedded ``"``)
          (default= ``None``)
        - multiline - boolean indicating whether quotes can span
          multiple lines (default= ``False``)
        - unquoteResults - boolean indicating whether the matched text
          should be unquoted (default= ``True``)
        - endQuoteChar - string of one or more characters defining the
          end of the quote delimited string (default= ``None``  => same as
          quoteChar)
        - convertWhitespaceEscapes - convert escaped whitespace
          (``'\t'``, ``'\n'``, etc.) to actual whitespace
          (default= ``True``)

    Example::

        qs = QuotedString('"')
        print(qs.searchString('lsjdf "This is the quote" sldjf'))
        complex_qs = QuotedString('{{', endQuoteChar='}}')
        print(complex_qs.searchString('lsjdf {{This is the "quote"}} sldjf'))
        sql_qs = QuotedString('"', escQuote='""')
        print(sql_qs.searchString('lsjdf "This is the quote with ""embedded"" quotes" sldjf'))

    prints::

        [['This is the quote']]
        [['This is the "quote"']]
        [['This is the quote with "embedded" quotes']]
    """
    def __init__(self, quoteChar, escChar=None, escQuote=None, multiline=False,
                 unquoteResults=True, endQuoteChar=None, convertWhitespaceEscapes=True):
        super(QuotedString, self).__init__()

        # remove white space from quote chars - wont work anyway
        quoteChar = quoteChar.strip()
        if not quoteChar:
            warnings.warn("quoteChar cannot be the empty string", SyntaxWarning, stacklevel=2)
            raise SyntaxError()

        if endQuoteChar is None:
            endQuoteChar = quoteChar
        else:
            endQuoteChar = endQuoteChar.strip()
            if not endQuoteChar:
                warnings.warn("endQuoteChar cannot be the empty string", SyntaxWarning, stacklevel=2)
                raise SyntaxError()

        self.quoteChar = quoteChar
        self.quoteCharLen = len(quoteChar)
        self.firstQuoteChar = quoteChar[0]
        self.endQuoteChar = endQuoteChar
        self.endQuoteCharLen = len(endQuoteChar)
        self.escChar = escChar
        self.escQuote = escQuote
        self.unquoteResults = unquoteResults
        self.convertWhitespaceEscapes = convertWhitespaceEscapes

        if multiline:
            self.flags = re.MULTILINE | re.DOTALL
            self.pattern = r'%s(?:[^%s%s]' % (re.escape(self.quoteChar),
                                              _escapeRegexRangeChars(self.endQuoteChar[0]),
                                              (escChar is not None and _escapeRegexRangeChars(escChar) or ''))
        else:
            self.flags = 0
            self.pattern = r'%s(?:[^%s\n\r%s]' % (re.escape(self.quoteChar),
                                                  _escapeRegexRangeChars(self.endQuoteChar[0]),
                                                  (escChar is not None and _escapeRegexRangeChars(escChar) or ''))
        if len(self.endQuoteChar) > 1:
            self.pattern += (
                '|(?:' + ')|(?:'.join("%s[^%s]" % (re.escape(self.endQuoteChar[:i]),
                                                   _escapeRegexRangeChars(self.endQuoteChar[i]))
                                      for i in range(len(self.endQuoteChar) - 1, 0, -1)) + ')')

        if escQuote:
            self.pattern += (r'|(?:%s)' % re.escape(escQuote))
        if escChar:
            self.pattern += (r'|(?:%s.)' % re.escape(escChar))
            self.escCharReplacePattern = re.escape(self.escChar) + "(.)"
        self.pattern += (r')*%s' % re.escape(self.endQuoteChar))

        try:
            self.re = re.compile(self.pattern, self.flags)
            self.reString = self.pattern
            self.re_match = self.re.match
        except sre_constants.error:
            warnings.warn("invalid pattern (%s) passed to Regex" % self.pattern,
                          SyntaxWarning, stacklevel=2)
            raise

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayIndexError = False
        self.mayReturnEmpty = True

    def parseImpl(self, instring, loc, doActions=True):
        result = instring[loc] == self.firstQuoteChar and self.re_match(instring, loc) or None
        if not result:
            raise ParseException(instring, loc, self.errmsg, self)

        loc = result.end()
        ret = result.group()

        if self.unquoteResults:

            # strip off quotes
            ret = ret[self.quoteCharLen: -self.endQuoteCharLen]

            if isinstance(ret, basestring):
                # replace escaped whitespace
                if '\\' in ret and self.convertWhitespaceEscapes:
                    ws_map = {
                        r'\t': '\t',
                        r'\n': '\n',
                        r'\f': '\f',
                        r'\r': '\r',
                    }
                    for wslit, wschar in ws_map.items():
                        ret = ret.replace(wslit, wschar)

                # replace escaped characters
                if self.escChar:
                    ret = re.sub(self.escCharReplacePattern, r"\g<1>", ret)

                # replace escaped quotes
                if self.escQuote:
                    ret = ret.replace(self.escQuote, self.endQuoteChar)

        return loc, ret

    def __str__(self):
        try:
            return super(QuotedString, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "quoted string, starting with %s ending with %s" % (self.quoteChar, self.endQuoteChar)

        return self.strRepr


class CharsNotIn(Token):
    """Token for matching words composed of characters *not* in a given
    set (will include whitespace in matched characters if not listed in
    the provided exclusion set - see example). Defined with string
    containing all disallowed characters, and an optional minimum,
    maximum, and/or exact length.  The default value for ``min`` is
    1 (a minimum value < 1 is not valid); the default values for
    ``max`` and ``exact`` are 0, meaning no maximum or exact
    length restriction.

    Example::

        # define a comma-separated-value as anything that is not a ','
        csv_value = CharsNotIn(',')
        print(delimitedList(csv_value).parseString("dkls,lsdkjf,s12 34,@!#,213"))

    prints::

        ['dkls', 'lsdkjf', 's12 34', '@!#', '213']
    """
    def __init__(self, notChars, min=1, max=0, exact=0):
        super(CharsNotIn, self).__init__()
        self.skipWhitespace = False
        self.notChars = notChars

        if min < 1:
            raise ValueError("cannot specify a minimum length < 1; use "
                             "Optional(CharsNotIn()) if zero-length char group is permitted")

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

        self.name = _ustr(self)
        self.errmsg = "Expected " + self.name
        self.mayReturnEmpty = (self.minLen == 0)
        self.mayIndexError = False

    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc] in self.notChars:
            raise ParseException(instring, loc, self.errmsg, self)

        start = loc
        loc += 1
        notchars = self.notChars
        maxlen = min(start + self.maxLen, len(instring))
        while loc < maxlen and instring[loc] not in notchars:
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]

    def __str__(self):
        try:
            return super(CharsNotIn, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            if len(self.notChars) > 4:
                self.strRepr = "!W:(%s...)" % self.notChars[:4]
            else:
                self.strRepr = "!W:(%s)" % self.notChars

        return self.strRepr

class White(Token):
    """Special matching class for matching whitespace.  Normally,
    whitespace is ignored by pyparsing grammars.  This class is included
    when some whitespace structures are significant.  Define with
    a string containing the whitespace characters to be matched; default
    is ``" \\t\\r\\n"``.  Also takes optional ``min``,
    ``max``, and ``exact`` arguments, as defined for the
    :class:`Word` class.
    """
    whiteStrs = {
        ' ' : '<SP>',
        '\t': '<TAB>',
        '\n': '<LF>',
        '\r': '<CR>',
        '\f': '<FF>',
        u'\u00A0': '<NBSP>',
        u'\u1680': '<OGHAM_SPACE_MARK>',
        u'\u180E': '<MONGOLIAN_VOWEL_SEPARATOR>',
        u'\u2000': '<EN_QUAD>',
        u'\u2001': '<EM_QUAD>',
        u'\u2002': '<EN_SPACE>',
        u'\u2003': '<EM_SPACE>',
        u'\u2004': '<THREE-PER-EM_SPACE>',
        u'\u2005': '<FOUR-PER-EM_SPACE>',
        u'\u2006': '<SIX-PER-EM_SPACE>',
        u'\u2007': '<FIGURE_SPACE>',
        u'\u2008': '<PUNCTUATION_SPACE>',
        u'\u2009': '<THIN_SPACE>',
        u'\u200A': '<HAIR_SPACE>',
        u'\u200B': '<ZERO_WIDTH_SPACE>',
        u'\u202F': '<NNBSP>',
        u'\u205F': '<MMSP>',
        u'\u3000': '<IDEOGRAPHIC_SPACE>',
        }
    def __init__(self, ws=" \t\r\n", min=1, max=0, exact=0):
        super(White, self).__init__()
        self.matchWhite = ws
        self.setWhitespaceChars("".join(c for c in self.whiteChars if c not in self.matchWhite))
        # ~ self.leaveWhitespace()
        self.name = ("".join(White.whiteStrs[c] for c in self.matchWhite))
        self.mayReturnEmpty = True
        self.errmsg = "Expected " + self.name

        self.minLen = min

        if max > 0:
            self.maxLen = max
        else:
            self.maxLen = _MAX_INT

        if exact > 0:
            self.maxLen = exact
            self.minLen = exact

    def parseImpl(self, instring, loc, doActions=True):
        if instring[loc] not in self.matchWhite:
            raise ParseException(instring, loc, self.errmsg, self)
        start = loc
        loc += 1
        maxloc = start + self.maxLen
        maxloc = min(maxloc, len(instring))
        while loc < maxloc and instring[loc] in self.matchWhite:
            loc += 1

        if loc - start < self.minLen:
            raise ParseException(instring, loc, self.errmsg, self)

        return loc, instring[start:loc]


class _PositionToken(Token):
    def __init__(self):
        super(_PositionToken, self).__init__()
        self.name = self.__class__.__name__
        self.mayReturnEmpty = True
        self.mayIndexError = False

class GoToColumn(_PositionToken):
    """Token to advance to a specific column of input text; useful for
    tabular report scraping.
    """
    def __init__(self, colno):
        super(GoToColumn, self).__init__()
        self.col = colno

    def preParse(self, instring, loc):
        if col(loc, instring) != self.col:
            instrlen = len(instring)
            if self.ignoreExprs:
                loc = self._skipIgnorables(instring, loc)
            while loc < instrlen and instring[loc].isspace() and col(loc, instring) != self.col:
                loc += 1
        return loc

    def parseImpl(self, instring, loc, doActions=True):
        thiscol = col(loc, instring)
        if thiscol > self.col:
            raise ParseException(instring, loc, "Text not in expected column", self)
        newloc = loc + self.col - thiscol
        ret = instring[loc: newloc]
        return newloc, ret


class LineStart(_PositionToken):
    r"""Matches if current position is at the beginning of a line within
    the parse string

    Example::

        test = '''\
        AAA this line
        AAA and this line
          AAA but not this one
        B AAA and definitely not this one
        '''

        for t in (LineStart() + 'AAA' + restOfLine).searchString(test):
            print(t)

    prints::

        ['AAA', ' this line']
        ['AAA', ' and this line']

    """
    def __init__(self):
        super(LineStart, self).__init__()
        self.errmsg = "Expected start of line"

    def parseImpl(self, instring, loc, doActions=True):
        if col(loc, instring) == 1:
            return loc, []
        raise ParseException(instring, loc, self.errmsg, self)

class LineEnd(_PositionToken):
    """Matches if current position is at the end of a line within the
    parse string
    """
    def __init__(self):
        super(LineEnd, self).__init__()
        self.setWhitespaceChars(ParserElement.DEFAULT_WHITE_CHARS.replace("\n", ""))
        self.errmsg = "Expected end of line"

    def parseImpl(self, instring, loc, doActions=True):
        if loc < len(instring):
            if instring[loc] == "\n":
                return loc + 1, "\n"
            else:
                raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc + 1, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class StringStart(_PositionToken):
    """Matches if current position is at the beginning of the parse
    string
    """
    def __init__(self):
        super(StringStart, self).__init__()
        self.errmsg = "Expected start of text"

    def parseImpl(self, instring, loc, doActions=True):
        if loc != 0:
            # see if entire string up to here is just whitespace and ignoreables
            if loc != self.preParse(instring, 0):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class StringEnd(_PositionToken):
    """Matches if current position is at the end of the parse string
    """
    def __init__(self):
        super(StringEnd, self).__init__()
        self.errmsg = "Expected end of text"

    def parseImpl(self, instring, loc, doActions=True):
        if loc < len(instring):
            raise ParseException(instring, loc, self.errmsg, self)
        elif loc == len(instring):
            return loc + 1, []
        elif loc > len(instring):
            return loc, []
        else:
            raise ParseException(instring, loc, self.errmsg, self)

class WordStart(_PositionToken):
    """Matches if the current position is at the beginning of a Word,
    and is not preceded by any character in a given set of
    ``wordChars`` (default= ``printables``). To emulate the
    ``\b`` behavior of regular expressions, use
    ``WordStart(alphanums)``. ``WordStart`` will also match at
    the beginning of the string being parsed, or at the beginning of
    a line.
    """
    def __init__(self, wordChars=printables):
        super(WordStart, self).__init__()
        self.wordChars = set(wordChars)
        self.errmsg = "Not at the start of a word"

    def parseImpl(self, instring, loc, doActions=True):
        if loc != 0:
            if (instring[loc - 1] in self.wordChars
                    or instring[loc] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

class WordEnd(_PositionToken):
    """Matches if the current position is at the end of a Word, and is
    not followed by any character in a given set of ``wordChars``
    (default= ``printables``). To emulate the ``\b`` behavior of
    regular expressions, use ``WordEnd(alphanums)``. ``WordEnd``
    will also match at the end of the string being parsed, or at the end
    of a line.
    """
    def __init__(self, wordChars=printables):
        super(WordEnd, self).__init__()
        self.wordChars = set(wordChars)
        self.skipWhitespace = False
        self.errmsg = "Not at the end of a word"

    def parseImpl(self, instring, loc, doActions=True):
        instrlen = len(instring)
        if instrlen > 0 and loc < instrlen:
            if (instring[loc] in self.wordChars or
                    instring[loc - 1] not in self.wordChars):
                raise ParseException(instring, loc, self.errmsg, self)
        return loc, []


class ParseExpression(ParserElement):
    """Abstract subclass of ParserElement, for combining and
    post-processing parsed tokens.
    """
    def __init__(self, exprs, savelist=False):
        super(ParseExpression, self).__init__(savelist)
        if isinstance(exprs, _generatorType):
            exprs = list(exprs)

        if isinstance(exprs, basestring):
            self.exprs = [self._literalStringClass(exprs)]
        elif isinstance(exprs, ParserElement):
            self.exprs = [exprs]
        elif isinstance(exprs, Iterable):
            exprs = list(exprs)
            # if sequence of strings provided, wrap with Literal
            if any(isinstance(expr, basestring) for expr in exprs):
                exprs = (self._literalStringClass(e) if isinstance(e, basestring) else e for e in exprs)
            self.exprs = list(exprs)
        else:
            try:
                self.exprs = list(exprs)
            except TypeError:
                self.exprs = [exprs]
        self.callPreparse = False

    def append(self, other):
        self.exprs.append(other)
        self.strRepr = None
        return self

    def leaveWhitespace(self):
        """Extends ``leaveWhitespace`` defined in base class, and also invokes ``leaveWhitespace`` on
           all contained expressions."""
        self.skipWhitespace = False
        self.exprs = [e.copy() for e in self.exprs]
        for e in self.exprs:
            e.leaveWhitespace()
        return self

    def ignore(self, other):
        if isinstance(other, Suppress):
            if other not in self.ignoreExprs:
                super(ParseExpression, self).ignore(other)
                for e in self.exprs:
                    e.ignore(self.ignoreExprs[-1])
        else:
            super(ParseExpression, self).ignore(other)
            for e in self.exprs:
                e.ignore(self.ignoreExprs[-1])
        return self

    def __str__(self):
        try:
            return super(ParseExpression, self).__str__()
        except Exception:
            pass

        if self.strRepr is None:
            self.strRepr = "%s:(%s)" % (self.__class__.__name__, _ustr(self.exprs))
        return self.strRepr

    def streamline(self):
        super(ParseExpression, self).streamline()

        for e in self.exprs:
            e.streamline()

        # collapse nested And's of the form And(And(And(a, b), c), d) to And(a, b, c, d)
        # but only if there are no parse actions or resultsNames on the nested And's
        # (likewise for Or's and MatchFirst's)
        if len(self.exprs) == 2:
            other = self.exprs[0]
            if (isinstance(other, self.__class__)
                    and not other.parseAction
                    and other.resultsName is None
                    and not other.debug):
                self.exprs = other.exprs[:] + [self.exprs[1]]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

            other = self.exprs[-1]
            if (isinstance(other, self.__class__)
                    and not other.parseAction
                    and other.resultsName is None
                    and not other.debug):
                self.exprs = self.exprs[:-1] + other.exprs[:]
                self.strRepr = None
                self.mayReturnEmpty |= other.mayReturnEmpty
                self.mayIndexError  |= other.mayIndexError

        self.errmsg = "Expected " + _ustr(self)

        return self

    def validate(self, validateTrace=None):
        tmp = (validateTrace if validateTrace is not None else [])[:] + [self]
        for e in self.exprs:
            e.validate(tmp)
        self.checkRecursion([])

    def copy(self):
        ret = super(ParseExpression, self).copy()
        ret.exprs = [e.copy() for e in self.exprs]
        return ret

    def _setResultsName(self, name, listAllMatches=False):
        if __diag__.warn_ungrouped_named_tokens_in_collection:
            for e in self.exprs:
                if isinstance(e, ParserElement) and e.resultsName:
                    warnings.warn("{0}: setting results name {1!r} on {2} expression "
                                  "collides with {3!r} on contained expression".format("warn_ungrouped_named_tokens_in_collection",
                                                                                       name,
                                                                                       type(self).__name__,
                                                                                       e.resultsName),
                                  stacklevel=3)

        return super(ParseExpression, self)._setResultsName(name, listAllMatches)


class And(ParseExpression):
    """
    Requires all given :class:`ParseExpression` s to be found in the given order.
    Expressions may be separated by whitespace.
    May be constructed using the ``'+'`` operator.
    May also be constructed using the ``'-'`` operator, which will
    suppress backtracking.

    Example::

        integer = Word(nums)
        name_expr = OneOrMore(Word(alphas))

        expr = And([integer("id"), name_expr("name"), integer("age")])
        # more easily written as:
        expr = integer("id") + name_expr("name") + integer("age")
    """

    class _ErrorStop(Empty):
        def __init__(self, *args, **kwargs):
            super(And._ErrorStop, self).__init__(*args, **kwargs)
            self.name = '-'
            self.leaveWhitespace()

    def __init__(self, exprs, savelist=True):
        exprs = list(exprs)
        if exprs and Ellipsis in exprs:
            tmp = []
            for i, expr in enumerate(exprs):
                if expr is Ellipsis:
                    if i < len(exprs) - 1:
                        skipto_arg = (Empty() + exprs[i + 1]).exprs[-1]
                        tmp.append(SkipTo(skipto_arg)("_skipped*"))
                    else:
                        raise Exception("cannot construct And with sequence ending in ...")
                else:
                    tmp.append(expr)
            exprs[:] = tmp
        super(And, self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.setWhitespaceChars(self.exprs[0].whiteChars)
        self.skipWhitespace = self.exprs[0].skipWhitespace
        self.callPreparse = True

    def streamline(self):
        # collapse any _PendingSkip's
        if self.exprs:
            if any(isinstance(e, ParseExpression) and e.exprs and isinstance(e.exprs[-1], _PendingSkip)
                   for e in self.exprs[:-1]):
                for i, e in enumerate(self.exprs[:-1]):
                    if e is None:
                        continue
                    if (isinstance(e, ParseExpression)
                            and e.exprs and isinstance(e.exprs[-1], _PendingSkip)):
                        e.exprs[-1] = e.exprs[-1] + self.exprs[i + 1]
                        self.exprs[i + 1] = None
                self.exprs = [e for e in self.exprs if e is not None]

        super(And, self).streamline()
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        return self

    def parseImpl(self, instring, loc, doActions=True):
        # pass False as last arg to _parse for first element, since we already
        # pre-parsed the string as part of our And pre-parsing
        loc, resultlist = self.exprs[0]._parse(instring, loc, doActions, callPreParse=False)
        errorStop = False
        for e in self.exprs[1:]:
            if isinstance(e, And._ErrorStop):
                errorStop = True
                continue
            if errorStop:
                try:
                    loc, exprtokens = e._parse(instring, loc, doActions)
                except ParseSyntaxException:
                    raise
                except ParseBaseException as pe:
                    pe.__traceback__ = None
                    raise ParseSyntaxException._from_exception(pe)
                except IndexError:
                    raise ParseSyntaxException(instring, len(instring), self.errmsg, self)
            else:
                loc, exprtokens = e._parse(instring, loc, doActions)
            if exprtokens or exprtokens.haskeys():
                resultlist += exprtokens
        return loc, resultlist

    def __iadd__(self, other):
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        return self.append(other)  # And([self, other])

    def checkRecursion(self, parseElementList):
        subRecCheckList = parseElementList[:] + [self]
        for e in self.exprs:
            e.checkRecursion(subRecCheckList)
            if not e.mayReturnEmpty:
                break

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr


class Or(ParseExpression):
    """Requires that at least one :class:`ParseExpression` is found. If
    two expressions match, the expression that matches the longest
    string will be used. May be constructed using the ``'^'``
    operator.

    Example::

        # construct Or using '^' operator

        number = Word(nums) ^ Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789"))

    prints::

        [['123'], ['3.1416'], ['789']]
    """
    def __init__(self, exprs, savelist=False):
        super(Or, self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def streamline(self):
        super(Or, self).streamline()
        if __compat__.collect_all_And_tokens:
            self.saveAsList = any(e.saveAsList for e in self.exprs)
        return self

    def parseImpl(self, instring, loc, doActions=True):
        maxExcLoc = -1
        maxException = None
        matches = []
        for e in self.exprs:
            try:
                loc2 = e.tryParse(instring, loc)
            except ParseException as err:
                err.__traceback__ = None
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring, len(instring), e.errmsg, self)
                    maxExcLoc = len(instring)
            else:
                # save match among all matches, to retry longest to shortest
                matches.append((loc2, e))

        if matches:
            # re-evaluate all matches in descending order of length of match, in case attached actions
            # might change whether or how much they match of the input.
            matches.sort(key=itemgetter(0), reverse=True)

            if not doActions:
                # no further conditions or parse actions to change the selection of
                # alternative, so the first match will be the best match
                best_expr = matches[0][1]
                return best_expr._parse(instring, loc, doActions)

            longest = -1, None
            for loc1, expr1 in matches:
                if loc1 <= longest[0]:
                    # already have a longer match than this one will deliver, we are done
                    return longest

                try:
                    loc2, toks = expr1._parse(instring, loc, doActions)
                except ParseException as err:
                    err.__traceback__ = None
                    if err.loc > maxExcLoc:
                        maxException = err
                        maxExcLoc = err.loc
                else:
                    if loc2 >= loc1:
                        return loc2, toks
                    # didn't match as much as before
                    elif loc2 > longest[0]:
                        longest = loc2, toks

            if longest != (-1, None):
                return longest

        if maxException is not None:
            maxException.msg = self.errmsg
            raise maxException
        else:
            raise ParseException(instring, loc, "no defined alternatives to match", self)


    def __ixor__(self, other):
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        return self.append(other)  # Or([self, other])

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " ^ ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion(self, parseElementList):
        subRecCheckList = parseElementList[:] + [self]
        for e in self.exprs:
            e.checkRecursion(subRecCheckList)

    def _setResultsName(self, name, listAllMatches=False):
        if (not __compat__.collect_all_And_tokens
                and __diag__.warn_multiple_tokens_in_named_alternation):
            if any(isinstance(e, And) for e in self.exprs):
                warnings.warn("{0}: setting results name {1!r} on {2} expression "
                              "may only return a single token for an And alternative, "
                              "in future will return the full list of tokens".format(
                    "warn_multiple_tokens_in_named_alternation", name, type(self).__name__),
                    stacklevel=3)

        return super(Or, self)._setResultsName(name, listAllMatches)


class MatchFirst(ParseExpression):
    """Requires that at least one :class:`ParseExpression` is found. If
    two expressions match, the first one listed is the one that will
    match. May be constructed using the ``'|'`` operator.

    Example::

        # construct MatchFirst using '|' operator

        # watch the order of expressions to match
        number = Word(nums) | Combine(Word(nums) + '.' + Word(nums))
        print(number.searchString("123 3.1416 789")) #  Fail! -> [['123'], ['3'], ['1416'], ['789']]

        # put more selective expression first
        number = Combine(Word(nums) + '.' + Word(nums)) | Word(nums)
        print(number.searchString("123 3.1416 789")) #  Better -> [['123'], ['3.1416'], ['789']]
    """
    def __init__(self, exprs, savelist=False):
        super(MatchFirst, self).__init__(exprs, savelist)
        if self.exprs:
            self.mayReturnEmpty = any(e.mayReturnEmpty for e in self.exprs)
        else:
            self.mayReturnEmpty = True

    def streamline(self):
        super(MatchFirst, self).streamline()
        if __compat__.collect_all_And_tokens:
            self.saveAsList = any(e.saveAsList for e in self.exprs)
        return self

    def parseImpl(self, instring, loc, doActions=True):
        maxExcLoc = -1
        maxException = None
        for e in self.exprs:
            try:
                ret = e._parse(instring, loc, doActions)
                return ret
            except ParseException as err:
                if err.loc > maxExcLoc:
                    maxException = err
                    maxExcLoc = err.loc
            except IndexError:
                if len(instring) > maxExcLoc:
                    maxException = ParseException(instring, len(instring), e.errmsg, self)
                    maxExcLoc = len(instring)

        # only got here if no expression matched, raise exception for match that made it the furthest
        else:
            if maxException is not None:
                maxException.msg = self.errmsg
                raise maxException
            else:
                raise ParseException(instring, loc, "no defined alternatives to match", self)

    def __ior__(self, other):
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        return self.append(other)  # MatchFirst([self, other])

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " | ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion(self, parseElementList):
        subRecCheckList = parseElementList[:] + [self]
        for e in self.exprs:
            e.checkRecursion(subRecCheckList)

    def _setResultsName(self, name, listAllMatches=False):
        if (not __compat__.collect_all_And_tokens
                and __diag__.warn_multiple_tokens_in_named_alternation):
            if any(isinstance(e, And) for e in self.exprs):
                warnings.warn("{0}: setting results name {1!r} on {2} expression "
                              "may only return a single token for an And alternative, "
                              "in future will return the full list of tokens".format(
                    "warn_multiple_tokens_in_named_alternation", name, type(self).__name__),
                    stacklevel=3)

        return super(MatchFirst, self)._setResultsName(name, listAllMatches)


class Each(ParseExpression):
    """Requires all given :class:`ParseExpression` s to be found, but in
    any order. Expressions may be separated by whitespace.

    May be constructed using the ``'&'`` operator.

    Example::

        color = oneOf("RED ORANGE YELLOW GREEN BLUE PURPLE BLACK WHITE BROWN")
        shape_type = oneOf("SQUARE CIRCLE TRIANGLE STAR HEXAGON OCTAGON")
        integer = Word(nums)
        shape_attr = "shape:" + shape_type("shape")
        posn_attr = "posn:" + Group(integer("x") + ',' + integer("y"))("posn")
        color_attr = "color:" + color("color")
        size_attr = "size:" + integer("size")

        # use Each (using operator '&') to accept attributes in any order
        # (shape and posn are required, color and size are optional)
        shape_spec = shape_attr & posn_attr & Optional(color_attr) & Optional(size_attr)

        shape_spec.runTests('''
            shape: SQUARE color: BLACK posn: 100, 120
            shape: CIRCLE size: 50 color: BLUE posn: 50,80
            color:GREEN size:20 shape:TRIANGLE posn:20,40
            '''
            )

    prints::

        shape: SQUARE color: BLACK posn: 100, 120
        ['shape:', 'SQUARE', 'color:', 'BLACK', 'posn:', ['100', ',', '120']]
        - color: BLACK
        - posn: ['100', ',', '120']
          - x: 100
          - y: 120
        - shape: SQUARE


        shape: CIRCLE size: 50 color: BLUE posn: 50,80
        ['shape:', 'CIRCLE', 'size:', '50', 'color:', 'BLUE', 'posn:', ['50', ',', '80']]
        - color: BLUE
        - posn: ['50', ',', '80']
          - x: 50
          - y: 80
        - shape: CIRCLE
        - size: 50


        color: GREEN size: 20 shape: TRIANGLE posn: 20,40
        ['color:', 'GREEN', 'size:', '20', 'shape:', 'TRIANGLE', 'posn:', ['20', ',', '40']]
        - color: GREEN
        - posn: ['20', ',', '40']
          - x: 20
          - y: 40
        - shape: TRIANGLE
        - size: 20
    """
    def __init__(self, exprs, savelist=True):
        super(Each, self).__init__(exprs, savelist)
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        self.skipWhitespace = True
        self.initExprGroups = True
        self.saveAsList = True

    def streamline(self):
        super(Each, self).streamline()
        self.mayReturnEmpty = all(e.mayReturnEmpty for e in self.exprs)
        return self

    def parseImpl(self, instring, loc, doActions=True):
        if self.initExprGroups:
            self.opt1map = dict((id(e.expr), e) for e in self.exprs if isinstance(e, Optional))
            opt1 = [e.expr for e in self.exprs if isinstance(e, Optional)]
            opt2 = [e for e in self.exprs if e.mayReturnEmpty and not isinstance(e, (Optional, Regex))]
            self.optionals = opt1 + opt2
            self.multioptionals = [e.expr for e in self.exprs if isinstance(e, ZeroOrMore)]
            self.multirequired = [e.expr for e in self.exprs if isinstance(e, OneOrMore)]
            self.required = [e for e in self.exprs if not isinstance(e, (Optional, ZeroOrMore, OneOrMore))]
            self.required += self.multirequired
            self.initExprGroups = False
        tmpLoc = loc
        tmpReqd = self.required[:]
        tmpOpt  = self.optionals[:]
        matchOrder = []

        keepMatching = True
        while keepMatching:
            tmpExprs = tmpReqd + tmpOpt + self.multioptionals + self.multirequired
            failed = []
            for e in tmpExprs:
                try:
                    tmpLoc = e.tryParse(instring, tmpLoc)
                except ParseException:
                    failed.append(e)
                else:
                    matchOrder.append(self.opt1map.get(id(e), e))
                    if e in tmpReqd:
                        tmpReqd.remove(e)
                    elif e in tmpOpt:
                        tmpOpt.remove(e)
            if len(failed) == len(tmpExprs):
                keepMatching = False

        if tmpReqd:
            missing = ", ".join(_ustr(e) for e in tmpReqd)
            raise ParseException(instring, loc, "Missing one or more required elements (%s)" % missing)

        # add any unmatched Optionals, in case they have default values defined
        matchOrder += [e for e in self.exprs if isinstance(e, Optional) and e.expr in tmpOpt]

        resultlist = []
        for e in matchOrder:
            loc, results = e._parse(instring, loc, doActions)
            resultlist.append(results)

        finalResults = sum(resultlist, ParseResults([]))
        return loc, finalResults

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + " & ".join(_ustr(e) for e in self.exprs) + "}"

        return self.strRepr

    def checkRecursion(self, parseElementList):
        subRecCheckList = parseElementList[:] + [self]
        for e in self.exprs:
            e.checkRecursion(subRecCheckList)


class ParseElementEnhance(ParserElement):
    """Abstract subclass of :class:`ParserElement`, for combining and
    post-processing parsed tokens.
    """
    def __init__(self, expr, savelist=False):
        super(ParseElementEnhance, self).__init__(savelist)
        if isinstance(expr, basestring):
            if issubclass(self._literalStringClass, Token):
                expr = self._literalStringClass(expr)
            else:
                expr = self._literalStringClass(Literal(expr))
        self.expr = expr
        self.strRepr = None
        if expr is not None:
            self.mayIndexError = expr.mayIndexError
            self.mayReturnEmpty = expr.mayReturnEmpty
            self.setWhitespaceChars(expr.whiteChars)
            self.skipWhitespace = expr.skipWhitespace
            self.saveAsList = expr.saveAsList
            self.callPreparse = expr.callPreparse
            self.ignoreExprs.extend(expr.ignoreExprs)

    def parseImpl(self, instring, loc, doActions=True):
        if self.expr is not None:
            return self.expr._parse(instring, loc, doActions, callPreParse=False)
        else:
            raise ParseException("", loc, self.errmsg, self)

    def leaveWhitespace(self):
        self.skipWhitespace = False
        self.expr = self.expr.copy()
        if self.expr is not None:
            self.expr.leaveWhitespace()
        return self

    def ignore(self, other):
        if isinstance(other, Suppress):
            if other not in self.ignoreExprs:
                super(ParseElementEnhance, self).ignore(other)
                if self.expr is not None:
                    self.expr.ignore(self.ignoreExprs[-1])
        else:
            super(ParseElementEnhance, self).ignore(other)
            if self.expr is not None:
                self.expr.ignore(self.ignoreExprs[-1])
        return self

    def streamline(self):
        super(ParseElementEnhance, self).streamline()
        if self.expr is not None:
            self.expr.streamline()
        return self

    def checkRecursion(self, parseElementList):
        if self in parseElementList:
            raise RecursiveGrammarException(parseElementList + [self])
        subRecCheckList = parseElementList[:] + [self]
        if self.expr is not None:
            self.expr.checkRecursion(subRecCheckList)

    def validate(self, validateTrace=None):
        if validateTrace is None:
            validateTrace = []
        tmp = validateTrace[:] + [self]
        if self.expr is not None:
            self.expr.validate(tmp)
        self.checkRecursion([])

    def __str__(self):
        try:
            return super(ParseElementEnhance, self).__str__()
        except Exception:
            pass

        if self.strRepr is None and self.expr is not None:
            self.strRepr = "%s:(%s)" % (self.__class__.__name__, _ustr(self.expr))
        return self.strRepr


class FollowedBy(ParseElementEnhance):
    """Lookahead matching of the given parse expression.
    ``FollowedBy`` does *not* advance the parsing position within
    the input string, it only verifies that the specified parse
    expression matches at the current position.  ``FollowedBy``
    always returns a null token list. If any results names are defined
    in the lookahead expression, those *will* be returned for access by
    name.

    Example::

        # use FollowedBy to match a label only if it is followed by a ':'
        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))

        OneOrMore(attr_expr).parseString("shape: SQUARE color: BLACK posn: upper left").pprint()

    prints::

        [['shape', 'SQUARE'], ['color', 'BLACK'], ['posn', 'upper left']]
    """
    def __init__(self, expr):
        super(FollowedBy, self).__init__(expr)
        self.mayReturnEmpty = True

    def parseImpl(self, instring, loc, doActions=True):
        # by using self._expr.parse and deleting the contents of the returned ParseResults list
        # we keep any named results that were defined in the FollowedBy expression
        _, ret = self.expr._parse(instring, loc, doActions=doActions)
        del ret[:]

        return loc, ret


class PrecededBy(ParseElementEnhance):
    """Lookbehind matching of the given parse expression.
    ``PrecededBy`` does not advance the parsing position within the
    input string, it only verifies that the specified parse expression
    matches prior to the current position.  ``PrecededBy`` always
    returns a null token list, but if a results name is defined on the
    given expression, it is returned.

    Parameters:

     - expr - expression that must match prior to the current parse
       location
     - retreat - (default= ``None``) - (int) maximum number of characters
       to lookbehind prior to the current parse location

    If the lookbehind expression is a string, Literal, Keyword, or
    a Word or CharsNotIn with a specified exact or maximum length, then
    the retreat parameter is not required. Otherwise, retreat must be
    specified to give a maximum number of characters to look back from
    the current parse position for a lookbehind match.

    Example::

        # VB-style variable names with type prefixes
        int_var = PrecededBy("#") + pyparsing_common.identifier
        str_var = PrecededBy("$") + pyparsing_common.identifier

    """
    def __init__(self, expr, retreat=None):
        super(PrecededBy, self).__init__(expr)
        self.expr = self.expr().leaveWhitespace()
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.exact = False
        if isinstance(expr, str):
            retreat = len(expr)
            self.exact = True
        elif isinstance(expr, (Literal, Keyword)):
            retreat = expr.matchLen
            self.exact = True
        elif isinstance(expr, (Word, CharsNotIn)) and expr.maxLen != _MAX_INT:
            retreat = expr.maxLen
            self.exact = True
        elif isinstance(expr, _PositionToken):
            retreat = 0
            self.exact = True
        self.retreat = retreat
        self.errmsg = "not preceded by " + str(expr)
        self.skipWhitespace = False
        self.parseAction.append(lambda s, l, t: t.__delitem__(slice(None, None)))

    def parseImpl(self, instring, loc=0, doActions=True):
        if self.exact:
            if loc < self.retreat:
                raise ParseException(instring, loc, self.errmsg)
            start = loc - self.retreat
            _, ret = self.expr._parse(instring, start)
        else:
            # retreat specified a maximum lookbehind window, iterate
            test_expr = self.expr + StringEnd()
            instring_slice = instring[max(0, loc - self.retreat):loc]
            last_expr = ParseException(instring, loc, self.errmsg)
            for offset in range(1, min(loc, self.retreat + 1)+1):
                try:
                    # print('trying', offset, instring_slice, repr(instring_slice[loc - offset:]))
                    _, ret = test_expr._parse(instring_slice, len(instring_slice) - offset)
                except ParseBaseException as pbe:
                    last_expr = pbe
                else:
                    break
            else:
                raise last_expr
        return loc, ret


class NotAny(ParseElementEnhance):
    """Lookahead to disallow matching with the given parse expression.
    ``NotAny`` does *not* advance the parsing position within the
    input string, it only verifies that the specified parse expression
    does *not* match at the current position.  Also, ``NotAny`` does
    *not* skip over leading whitespace. ``NotAny`` always returns
    a null token list.  May be constructed using the '~' operator.

    Example::

        AND, OR, NOT = map(CaselessKeyword, "AND OR NOT".split())

        # take care not to mistake keywords for identifiers
        ident = ~(AND | OR | NOT) + Word(alphas)
        boolean_term = Optional(NOT) + ident

        # very crude boolean expression - to support parenthesis groups and
        # operation hierarchy, use infixNotation
        boolean_expr = boolean_term + ZeroOrMore((AND | OR) + boolean_term)

        # integers that are followed by "." are actually floats
        integer = Word(nums) + ~Char(".")
    """
    def __init__(self, expr):
        super(NotAny, self).__init__(expr)
        # ~ self.leaveWhitespace()
        self.skipWhitespace = False  # do NOT use self.leaveWhitespace(), don't want to propagate to exprs
        self.mayReturnEmpty = True
        self.errmsg = "Found unwanted token, " + _ustr(self.expr)

    def parseImpl(self, instring, loc, doActions=True):
        if self.expr.canParseNext(instring, loc):
            raise ParseException(instring, loc, self.errmsg, self)
        return loc, []

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "~{" + _ustr(self.expr) + "}"

        return self.strRepr

class _MultipleMatch(ParseElementEnhance):
    def __init__(self, expr, stopOn=None):
        super(_MultipleMatch, self).__init__(expr)
        self.saveAsList = True
        ender = stopOn
        if isinstance(ender, basestring):
            ender = self._literalStringClass(ender)
        self.stopOn(ender)

    def stopOn(self, ender):
        if isinstance(ender, basestring):
            ender = self._literalStringClass(ender)
        self.not_ender = ~ender if ender is not None else None
        return self

    def parseImpl(self, instring, loc, doActions=True):
        self_expr_parse = self.expr._parse
        self_skip_ignorables = self._skipIgnorables
        check_ender = self.not_ender is not None
        if check_ender:
            try_not_ender = self.not_ender.tryParse

        # must be at least one (but first see if we are the stopOn sentinel;
        # if so, fail)
        if check_ender:
            try_not_ender(instring, loc)
        loc, tokens = self_expr_parse(instring, loc, doActions, callPreParse=False)
        try:
            hasIgnoreExprs = (not not self.ignoreExprs)
            while 1:
                if check_ender:
                    try_not_ender(instring, loc)
                if hasIgnoreExprs:
                    preloc = self_skip_ignorables(instring, loc)
                else:
                    preloc = loc
                loc, tmptokens = self_expr_parse(instring, preloc, doActions)
                if tmptokens or tmptokens.haskeys():
                    tokens += tmptokens
        except (ParseException, IndexError):
            pass

        return loc, tokens

    def _setResultsName(self, name, listAllMatches=False):
        if __diag__.warn_ungrouped_named_tokens_in_collection:
            for e in [self.expr] + getattr(self.expr, 'exprs', []):
                if isinstance(e, ParserElement) and e.resultsName:
                    warnings.warn("{0}: setting results name {1!r} on {2} expression "
                                  "collides with {3!r} on contained expression".format("warn_ungrouped_named_tokens_in_collection",
                                                                                       name,
                                                                                       type(self).__name__,
                                                                                       e.resultsName),
                                  stacklevel=3)

        return super(_MultipleMatch, self)._setResultsName(name, listAllMatches)


class OneOrMore(_MultipleMatch):
    """Repetition of one or more of the given expression.

    Parameters:
     - expr - expression that must match one or more times
     - stopOn - (default= ``None``) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition
          expression)

    Example::

        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: BLACK"
        OneOrMore(attr_expr).parseString(text).pprint()  # Fail! read 'color' as data instead of next label -> [['shape', 'SQUARE color']]

        # use stopOn attribute for OneOrMore to avoid reading label string as part of the data
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        OneOrMore(attr_expr).parseString(text).pprint() # Better -> [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'BLACK']]

        # could also be written as
        (attr_expr * (1,)).parseString(text).pprint()
    """

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "{" + _ustr(self.expr) + "}..."

        return self.strRepr

class ZeroOrMore(_MultipleMatch):
    """Optional repetition of zero or more of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - stopOn - (default= ``None``) - expression for a terminating sentinel
          (only required if the sentinel would ordinarily match the repetition
          expression)

    Example: similar to :class:`OneOrMore`
    """
    def __init__(self, expr, stopOn=None):
        super(ZeroOrMore, self).__init__(expr, stopOn=stopOn)
        self.mayReturnEmpty = True

    def parseImpl(self, instring, loc, doActions=True):
        try:
            return super(ZeroOrMore, self).parseImpl(instring, loc, doActions)
        except (ParseException, IndexError):
            return loc, []

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]..."

        return self.strRepr


class _NullToken(object):
    def __bool__(self):
        return False
    __nonzero__ = __bool__
    def __str__(self):
        return ""

class Optional(ParseElementEnhance):
    """Optional matching of the given expression.

    Parameters:
     - expr - expression that must match zero or more times
     - default (optional) - value to be returned if the optional expression is not found.

    Example::

        # US postal code can be a 5-digit zip, plus optional 4-digit qualifier
        zip = Combine(Word(nums, exact=5) + Optional('-' + Word(nums, exact=4)))
        zip.runTests('''
            # traditional ZIP code
            12345

            # ZIP+4 form
            12101-0001

            # invalid ZIP
            98765-
            ''')

    prints::

        # traditional ZIP code
        12345
        ['12345']

        # ZIP+4 form
        12101-0001
        ['12101-0001']

        # invalid ZIP
        98765-
             ^
        FAIL: Expected end of text (at char 5), (line:1, col:6)
    """
    __optionalNotMatched = _NullToken()

    def __init__(self, expr, default=__optionalNotMatched):
        super(Optional, self).__init__(expr, savelist=False)
        self.saveAsList = self.expr.saveAsList
        self.defaultValue = default
        self.mayReturnEmpty = True

    def parseImpl(self, instring, loc, doActions=True):
        try:
            loc, tokens = self.expr._parse(instring, loc, doActions, callPreParse=False)
        except (ParseException, IndexError):
            if self.defaultValue is not self.__optionalNotMatched:
                if self.expr.resultsName:
                    tokens = ParseResults([self.defaultValue])
                    tokens[self.expr.resultsName] = self.defaultValue
                else:
                    tokens = [self.defaultValue]
            else:
                tokens = []
        return loc, tokens

    def __str__(self):
        if hasattr(self, "name"):
            return self.name

        if self.strRepr is None:
            self.strRepr = "[" + _ustr(self.expr) + "]"

        return self.strRepr

class SkipTo(ParseElementEnhance):
    """Token for skipping over all undefined text until the matched
    expression is found.

    Parameters:
     - expr - target expression marking the end of the data to be skipped
     - include - (default= ``False``) if True, the target expression is also parsed
          (the skipped text and target expression are returned as a 2-element list).
     - ignore - (default= ``None``) used to define grammars (typically quoted strings and
          comments) that might contain false matches to the target expression
     - failOn - (default= ``None``) define expressions that are not allowed to be
          included in the skipped test; if found before the target expression is found,
          the SkipTo is not a match

    Example::

        report = '''
            Outstanding Issues Report - 1 Jan 2000

               # | Severity | Description                               |  Days Open
            -----+----------+-------------------------------------------+-----------
             101 | Critical | Intermittent system crash                 |          6
              94 | Cosmetic | Spelling error on Login ('log|n')         |         14
              79 | Minor    | System slow when running too many reports |         47
            '''
        integer = Word(nums)
        SEP = Suppress('|')
        # use SkipTo to simply match everything up until the next SEP
        # - ignore quoted strings, so that a '|' character inside a quoted string does not match
        # - parse action will call token.strip() for each matched token, i.e., the description body
        string_data = SkipTo(SEP, ignore=quotedString)
        string_data.setParseAction(tokenMap(str.strip))
        ticket_expr = (integer("issue_num") + SEP
                      + string_data("sev") + SEP
                      + string_data("desc") + SEP
                      + integer("days_open"))

        for tkt in ticket_expr.searchString(report):
            print tkt.dump()

    prints::

        ['101', 'Critical', 'Intermittent system crash', '6']
        - days_open: 6
        - desc: Intermittent system crash
        - issue_num: 101
        - sev: Critical
        ['94', 'Cosmetic', "Spelling error on Login ('log|n')", '14']
        - days_open: 14
        - desc: Spelling error on Login ('log|n')
        - issue_num: 94
        - sev: Cosmetic
        ['79', 'Minor', 'System slow when running too many reports', '47']
        - days_open: 47
        - desc: System slow when running too many reports
        - issue_num: 79
        - sev: Minor
    """
    def __init__(self, other, include=False, ignore=None, failOn=None):
        super(SkipTo, self).__init__(other)
        self.ignoreExpr = ignore
        self.mayReturnEmpty = True
        self.mayIndexError = False
        self.includeMatch = include
        self.saveAsList = False
        if isinstance(failOn, basestring):
            self.failOn = self._literalStringClass(failOn)
        else:
            self.failOn = failOn
        self.errmsg = "No match found for " + _ustr(self.expr)

    def parseImpl(self, instring, loc, doActions=True):
        startloc = loc
        instrlen = len(instring)
        expr = self.expr
        expr_parse = self.expr._parse
        self_failOn_canParseNext = self.failOn.canParseNext if self.failOn is not None else None
        self_ignoreExpr_tryParse = self.ignoreExpr.tryParse if self.ignoreExpr is not None else None

        tmploc = loc
        while tmploc <= instrlen:
            if self_failOn_canParseNext is not None:
                # break if failOn expression matches
                if self_failOn_canParseNext(instring, tmploc):
                    break

            if self_ignoreExpr_tryParse is not None:
                # advance past ignore expressions
                while 1:
                    try:
                        tmploc = self_ignoreExpr_tryParse(instring, tmploc)
                    except ParseBaseException:
                        break

            try:
                expr_parse(instring, tmploc, doActions=False, callPreParse=False)
            except (ParseException, IndexError):
                # no match, advance loc in string
                tmploc += 1
            else:
                # matched skipto expr, done
                break

        else:
            # ran off the end of the input string without matching skipto expr, fail
            raise ParseException(instring, loc, self.errmsg, self)

        # build up return values
        loc = tmploc
        skiptext = instring[startloc:loc]
        skipresult = ParseResults(skiptext)

        if self.includeMatch:
            loc, mat = expr_parse(instring, loc, doActions, callPreParse=False)
            skipresult += mat

        return loc, skipresult

class Forward(ParseElementEnhance):
    """Forward declaration of an expression to be defined later -
    used for recursive grammars, such as algebraic infix notation.
    When the expression is known, it is assigned to the ``Forward``
    variable using the '<<' operator.

    Note: take care when assigning to ``Forward`` not to overlook
    precedence of operators.

    Specifically, '|' has a lower precedence than '<<', so that::

        fwdExpr << a | b | c

    will actually be evaluated as::

        (fwdExpr << a) | b | c

    thereby leaving b and c out as parseable alternatives.  It is recommended that you
    explicitly group the values inserted into the ``Forward``::

        fwdExpr << (a | b | c)

    Converting to use the '<<=' operator instead will avoid this problem.

    See :class:`ParseResults.pprint` for an example of a recursive
    parser created using ``Forward``.
    """
    def __init__(self, other=None):
        super(Forward, self).__init__(other, savelist=False)

    def __lshift__(self, other):
        if isinstance(other, basestring):
            other = self._literalStringClass(other)
        self.expr = other
        self.strRepr = None
        self.mayIndexError = self.expr.mayIndexError
        self.mayReturnEmpty = self.expr.mayReturnEmpty
        self.setWhitespaceChars(self.expr.whiteChars)
        self.skipWhitespace = self.expr.skipWhitespace
        self.saveAsList = self.expr.saveAsList
        self.ignoreExprs.extend(self.expr.ignoreExprs)
        return self

    def __ilshift__(self, other):
        return self << other

    def leaveWhitespace(self):
        self.skipWhitespace = False
        return self

    def streamline(self):
        if not self.streamlined:
            self.streamlined = True
            if self.expr is not None:
                self.expr.streamline()
        return self

    def validate(self, validateTrace=None):
        if validateTrace is None:
            validateTrace = []

        if self not in validateTrace:
            tmp = validateTrace[:] + [self]
            if self.expr is not None:
                self.expr.validate(tmp)
        self.checkRecursion([])

    def __str__(self):
        if hasattr(self, "name"):
            return self.name
        if self.strRepr is not None:
            return self.strRepr

        # Avoid infinite recursion by setting a temporary strRepr
        self.strRepr = ": ..."

        # Use the string representation of main expression.
        retString = '...'
        try:
            if self.expr is not None:
                retString = _ustr(self.expr)[:1000]
            else:
                retString = "None"
        finally:
            self.strRepr = self.__class__.__name__ + ": " + retString
        return self.strRepr

    def copy(self):
        if self.expr is not None:
            return super(Forward, self).copy()
        else:
            ret = Forward()
            ret <<= self
            return ret

    def _setResultsName(self, name, listAllMatches=False):
        if __diag__.warn_name_set_on_empty_Forward:
            if self.expr is None:
                warnings.warn("{0}: setting results name {0!r} on {1} expression "
                              "that has no contained expression".format("warn_name_set_on_empty_Forward",
                                                                        name,
                                                                        type(self).__name__),
                              stacklevel=3)

        return super(Forward, self)._setResultsName(name, listAllMatches)

class TokenConverter(ParseElementEnhance):
    """
    Abstract subclass of :class:`ParseExpression`, for converting parsed results.
    """
    def __init__(self, expr, savelist=False):
        super(TokenConverter, self).__init__(expr)  # , savelist)
        self.saveAsList = False

class Combine(TokenConverter):
    """Converter to concatenate all matching tokens to a single string.
    By default, the matching patterns must also be contiguous in the
    input string; this can be disabled by specifying
    ``'adjacent=False'`` in the constructor.

    Example::

        real = Word(nums) + '.' + Word(nums)
        print(real.parseString('3.1416')) # -> ['3', '.', '1416']
        # will also erroneously match the following
        print(real.parseString('3. 1416')) # -> ['3', '.', '1416']

        real = Combine(Word(nums) + '.' + Word(nums))
        print(real.parseString('3.1416')) # -> ['3.1416']
        # no match when there are internal spaces
        print(real.parseString('3. 1416')) # -> Exception: Expected W:(0123...)
    """
    def __init__(self, expr, joinString="", adjacent=True):
        super(Combine, self).__init__(expr)
        # suppress whitespace-stripping in contained parse expressions, but re-enable it on the Combine itself
        if adjacent:
            self.leaveWhitespace()
        self.adjacent = adjacent
        self.skipWhitespace = True
        self.joinString = joinString
        self.callPreparse = True

    def ignore(self, other):
        if self.adjacent:
            ParserElement.ignore(self, other)
        else:
            super(Combine, self).ignore(other)
        return self

    def postParse(self, instring, loc, tokenlist):
        retToks = tokenlist.copy()
        del retToks[:]
        retToks += ParseResults(["".join(tokenlist._asStringList(self.joinString))], modal=self.modalResults)

        if self.resultsName and retToks.haskeys():
            return [retToks]
        else:
            return retToks

class Group(TokenConverter):
    """Converter to return the matched tokens as a list - useful for
    returning tokens of :class:`ZeroOrMore` and :class:`OneOrMore` expressions.

    Example::

        ident = Word(alphas)
        num = Word(nums)
        term = ident | num
        func = ident + Optional(delimitedList(term))
        print(func.parseString("fn a, b, 100"))  # -> ['fn', 'a', 'b', '100']

        func = ident + Group(Optional(delimitedList(term)))
        print(func.parseString("fn a, b, 100"))  # -> ['fn', ['a', 'b', '100']]
    """
    def __init__(self, expr):
        super(Group, self).__init__(expr)
        self.saveAsList = True

    def postParse(self, instring, loc, tokenlist):
        return [tokenlist]

class Dict(TokenConverter):
    """Converter to return a repetitive expression as a list, but also
    as a dictionary. Each element can also be referenced using the first
    token in the expression as its key. Useful for tabular report
    scraping when the first column can be used as a item key.

    Example::

        data_word = Word(alphas)
        label = data_word + FollowedBy(':')
        attr_expr = Group(label + Suppress(':') + OneOrMore(data_word).setParseAction(' '.join))

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))

        # print attributes as plain groups
        print(OneOrMore(attr_expr).parseString(text).dump())

        # instead of OneOrMore(expr), parse using Dict(OneOrMore(Group(expr))) - Dict will auto-assign names
        result = Dict(OneOrMore(Group(attr_expr))).parseString(text)
        print(result.dump())

        # access named fields as dict entries, or output as dict
        print(result['shape'])
        print(result.asDict())

    prints::

        ['shape', 'SQUARE', 'posn', 'upper left', 'color', 'light blue', 'texture', 'burlap']
        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        {'color': 'light blue', 'posn': 'upper left', 'texture': 'burlap', 'shape': 'SQUARE'}

    See more examples at :class:`ParseResults` of accessing fields by results name.
    """
    def __init__(self, expr):
        super(Dict, self).__init__(expr)
        self.saveAsList = True

    def postParse(self, instring, loc, tokenlist):
        for i, tok in enumerate(tokenlist):
            if len(tok) == 0:
                continue
            ikey = tok[0]
            if isinstance(ikey, int):
                ikey = _ustr(tok[0]).strip()
            if len(tok) == 1:
                tokenlist[ikey] = _ParseResultsWithOffset("", i)
            elif len(tok) == 2 and not isinstance(tok[1], ParseResults):
                tokenlist[ikey] = _ParseResultsWithOffset(tok[1], i)
            else:
                dictvalue = tok.copy()  # ParseResults(i)
                del dictvalue[0]
                if len(dictvalue) != 1 or (isinstance(dictvalue, ParseResults) and dictvalue.haskeys()):
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue, i)
                else:
                    tokenlist[ikey] = _ParseResultsWithOffset(dictvalue[0], i)

        if self.resultsName:
            return [tokenlist]
        else:
            return tokenlist


class Suppress(TokenConverter):
    """Converter for ignoring the results of a parsed expression.

    Example::

        source = "a, b, c,d"
        wd = Word(alphas)
        wd_list1 = wd + ZeroOrMore(',' + wd)
        print(wd_list1.parseString(source))

        # often, delimiters that are useful during parsing are just in the
        # way afterward - use Suppress to keep them out of the parsed output
        wd_list2 = wd + ZeroOrMore(Suppress(',') + wd)
        print(wd_list2.parseString(source))

    prints::

        ['a', ',', 'b', ',', 'c', ',', 'd']
        ['a', 'b', 'c', 'd']

    (See also :class:`delimitedList`.)
    """
    def postParse(self, instring, loc, tokenlist):
        return []

    def suppress(self):
        return self


class OnlyOnce(object):
    """Wrapper for parse actions, to ensure they are only called once.
    """
    def __init__(self, methodCall):
        self.callable = _trim_arity(methodCall)
        self.called = False
    def __call__(self, s, l, t):
        if not self.called:
            results = self.callable(s, l, t)
            self.called = True
            return results
        raise ParseException(s, l, "")
    def reset(self):
        self.called = False

def traceParseAction(f):
    """Decorator for debugging parse actions.

    When the parse action is called, this decorator will print
    ``">> entering method-name(line:<current_source_line>, <parse_location>, <matched_tokens>)"``.
    When the parse action completes, the decorator will print
    ``"<<"`` followed by the returned value, or any exception that the parse action raised.

    Example::

        wd = Word(alphas)

        @traceParseAction
        def remove_duplicate_chars(tokens):
            return ''.join(sorted(set(''.join(tokens))))

        wds = OneOrMore(wd).setParseAction(remove_duplicate_chars)
        print(wds.parseString("slkdjs sld sldd sdlf sdljf"))

    prints::

        >>entering remove_duplicate_chars(line: 'slkdjs sld sldd sdlf sdljf', 0, (['slkdjs', 'sld', 'sldd', 'sdlf', 'sdljf'], {}))
        <<leaving remove_duplicate_chars (ret: 'dfjkls')
        ['dfjkls']
    """
    f = _trim_arity(f)
    def z(*paArgs):
        thisFunc = f.__name__
        s, l, t = paArgs[-3:]
        if len(paArgs) > 3:
            thisFunc = paArgs[0].__class__.__name__ + '.' + thisFunc
        sys.stderr.write(">>entering %s(line: '%s', %d, %r)\n" % (thisFunc, line(l, s), l, t))
        try:
            ret = f(*paArgs)
        except Exception as exc:
            sys.stderr.write("<<leaving %s (exception: %s)\n" % (thisFunc, exc))
            raise
        sys.stderr.write("<<leaving %s (ret: %r)\n" % (thisFunc, ret))
        return ret
    try:
        z.__name__ = f.__name__
    except AttributeError:
        pass
    return z

#
# global helpers
#
def delimitedList(expr, delim=",", combine=False):
    """Helper to define a delimited list of expressions - the delimiter
    defaults to ','. By default, the list elements and delimiters can
    have intervening whitespace, and comments, but this can be
    overridden by passing ``combine=True`` in the constructor. If
    ``combine`` is set to ``True``, the matching tokens are
    returned as a single token string, with the delimiters included;
    otherwise, the matching tokens are returned as a list of tokens,
    with the delimiters suppressed.

    Example::

        delimitedList(Word(alphas)).parseString("aa,bb,cc") # -> ['aa', 'bb', 'cc']
        delimitedList(Word(hexnums), delim=':', combine=True).parseString("AA:BB:CC:DD:EE") # -> ['AA:BB:CC:DD:EE']
    """
    dlName = _ustr(expr) + " [" + _ustr(delim) + " " + _ustr(expr) + "]..."
    if combine:
        return Combine(expr + ZeroOrMore(delim + expr)).setName(dlName)
    else:
        return (expr + ZeroOrMore(Suppress(delim) + expr)).setName(dlName)

def countedArray(expr, intExpr=None):
    """Helper to define a counted list of expressions.

    This helper defines a pattern of the form::

        integer expr expr expr...

    where the leading integer tells how many expr expressions follow.
    The matched tokens returns the array of expr tokens as a list - the
    leading count token is suppressed.

    If ``intExpr`` is specified, it should be a pyparsing expression
    that produces an integer value.

    Example::

        countedArray(Word(alphas)).parseString('2 ab cd ef')  # -> ['ab', 'cd']

        # in this parser, the leading integer value is given in binary,
        # '10' indicating that 2 values are in the array
        binaryConstant = Word('01').setParseAction(lambda t: int(t[0], 2))
        countedArray(Word(alphas), intExpr=binaryConstant).parseString('10 ab cd ef')  # -> ['ab', 'cd']
    """
    arrayExpr = Forward()
    def countFieldParseAction(s, l, t):
        n = t[0]
        arrayExpr << (n and Group(And([expr] * n)) or Group(empty))
        return []
    if intExpr is None:
        intExpr = Word(nums).setParseAction(lambda t: int(t[0]))
    else:
        intExpr = intExpr.copy()
    intExpr.setName("arrayLen")
    intExpr.addParseAction(countFieldParseAction, callDuringTry=True)
    return (intExpr + arrayExpr).setName('(len) ' + _ustr(expr) + '...')

def _flatten(L):
    ret = []
    for i in L:
        if isinstance(i, list):
            ret.extend(_flatten(i))
        else:
            ret.append(i)
    return ret

def matchPreviousLiteral(expr):
    """Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks for
    a 'repeat' of a previous expression.  For example::

        first = Word(nums)
        second = matchPreviousLiteral(first)
        matchExpr = first + ":" + second

    will match ``"1:1"``, but not ``"1:2"``.  Because this
    matches a previous literal, will also match the leading
    ``"1:1"`` in ``"1:10"``. If this is not desired, use
    :class:`matchPreviousExpr`. Do *not* use with packrat parsing
    enabled.
    """
    rep = Forward()
    def copyTokenToRepeater(s, l, t):
        if t:
            if len(t) == 1:
                rep << t[0]
            else:
                # flatten t tokens
                tflat = _flatten(t.asList())
                rep << And(Literal(tt) for tt in tflat)
        else:
            rep << Empty()
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def matchPreviousExpr(expr):
    """Helper to define an expression that is indirectly defined from
    the tokens matched in a previous expression, that is, it looks for
    a 'repeat' of a previous expression.  For example::

        first = Word(nums)
        second = matchPreviousExpr(first)
        matchExpr = first + ":" + second

    will match ``"1:1"``, but not ``"1:2"``.  Because this
    matches by expressions, will *not* match the leading ``"1:1"``
    in ``"1:10"``; the expressions are evaluated first, and then
    compared, so ``"1"`` is compared with ``"10"``. Do *not* use
    with packrat parsing enabled.
    """
    rep = Forward()
    e2 = expr.copy()
    rep <<= e2
    def copyTokenToRepeater(s, l, t):
        matchTokens = _flatten(t.asList())
        def mustMatchTheseTokens(s, l, t):
            theseTokens = _flatten(t.asList())
            if theseTokens != matchTokens:
                raise ParseException('', 0, '')
        rep.setParseAction(mustMatchTheseTokens, callDuringTry=True)
    expr.addParseAction(copyTokenToRepeater, callDuringTry=True)
    rep.setName('(prev) ' + _ustr(expr))
    return rep

def _escapeRegexRangeChars(s):
    # ~  escape these chars: ^-[]
    for c in r"\^-[]":
        s = s.replace(c, _bslash + c)
    s = s.replace("\n", r"\n")
    s = s.replace("\t", r"\t")
    return _ustr(s)

def oneOf(strs, caseless=False, useRegex=True, asKeyword=False):
    """Helper to quickly define a set of alternative Literals, and makes
    sure to do longest-first testing when there is a conflict,
    regardless of the input order, but returns
    a :class:`MatchFirst` for best performance.

    Parameters:

     - strs - a string of space-delimited literals, or a collection of
       string literals
     - caseless - (default= ``False``) - treat all literals as
       caseless
     - useRegex - (default= ``True``) - as an optimization, will
       generate a Regex object; otherwise, will generate
       a :class:`MatchFirst` object (if ``caseless=True`` or ``asKeyword=True``, or if
       creating a :class:`Regex` raises an exception)
     - asKeyword - (default=``False``) - enforce Keyword-style matching on the
       generated expressions

    Example::

        comp_oper = oneOf("< = > <= >= !=")
        var = Word(alphas)
        number = Word(nums)
        term = var | number
        comparison_expr = term + comp_oper + term
        print(comparison_expr.searchString("B = 12  AA=23 B<=AA AA>12"))

    prints::

        [['B', '=', '12'], ['AA', '=', '23'], ['B', '<=', 'AA'], ['AA', '>', '12']]
    """
    if isinstance(caseless, basestring):
        warnings.warn("More than one string argument passed to oneOf, pass "
                      "choices as a list or space-delimited string", stacklevel=2)

    if caseless:
        isequal = (lambda a, b: a.upper() == b.upper())
        masks = (lambda a, b: b.upper().startswith(a.upper()))
        parseElementClass = CaselessKeyword if asKeyword else CaselessLiteral
    else:
        isequal = (lambda a, b: a == b)
        masks = (lambda a, b: b.startswith(a))
        parseElementClass = Keyword if asKeyword else Literal

    symbols = []
    if isinstance(strs, basestring):
        symbols = strs.split()
    elif isinstance(strs, Iterable):
        symbols = list(strs)
    else:
        warnings.warn("Invalid argument to oneOf, expected string or iterable",
                      SyntaxWarning, stacklevel=2)
    if not symbols:
        return NoMatch()

    if not asKeyword:
        # if not producing keywords, need to reorder to take care to avoid masking
        # longer choices with shorter ones
        i = 0
        while i < len(symbols) - 1:
            cur = symbols[i]
            for j, other in enumerate(symbols[i + 1:]):
                if isequal(other, cur):
                    del symbols[i + j + 1]
                    break
                elif masks(cur, other):
                    del symbols[i + j + 1]
                    symbols.insert(i, other)
                    break
            else:
                i += 1

    if not (caseless or asKeyword) and useRegex:
        # ~ print (strs, "->", "|".join([_escapeRegexChars(sym) for sym in symbols]))
        try:
            if len(symbols) == len("".join(symbols)):
                return Regex("[%s]" % "".join(_escapeRegexRangeChars(sym) for sym in symbols)).setName(' | '.join(symbols))
            else:
                return Regex("|".join(re.escape(sym) for sym in symbols)).setName(' | '.join(symbols))
        except Exception:
            warnings.warn("Exception creating Regex for oneOf, building MatchFirst",
                    SyntaxWarning, stacklevel=2)

    # last resort, just use MatchFirst
    return MatchFirst(parseElementClass(sym) for sym in symbols).setName(' | '.join(symbols))

def dictOf(key, value):
    """Helper to easily and clearly define a dictionary by specifying
    the respective patterns for the key and value.  Takes care of
    defining the :class:`Dict`, :class:`ZeroOrMore`, and
    :class:`Group` tokens in the proper order.  The key pattern
    can include delimiting markers or punctuation, as long as they are
    suppressed, thereby leaving the significant key text.  The value
    pattern can include named results, so that the :class:`Dict` results
    can include named token fields.

    Example::

        text = "shape: SQUARE posn: upper left color: light blue texture: burlap"
        attr_expr = (label + Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join))
        print(OneOrMore(attr_expr).parseString(text).dump())

        attr_label = label
        attr_value = Suppress(':') + OneOrMore(data_word, stopOn=label).setParseAction(' '.join)

        # similar to Dict, but simpler call format
        result = dictOf(attr_label, attr_value).parseString(text)
        print(result.dump())
        print(result['shape'])
        print(result.shape)  # object attribute access works too
        print(result.asDict())

    prints::

        [['shape', 'SQUARE'], ['posn', 'upper left'], ['color', 'light blue'], ['texture', 'burlap']]
        - color: light blue
        - posn: upper left
        - shape: SQUARE
        - texture: burlap
        SQUARE
        SQUARE
        {'color': 'light blue', 'shape': 'SQUARE', 'posn': 'upper left', 'texture': 'burlap'}
    """
    return Dict(OneOrMore(Group(key + value)))

def originalTextFor(expr, asString=True):
    """Helper to return the original, untokenized text for a given
    expression.  Useful to restore the parsed fields of an HTML start
    tag into the raw tag text itself, or to revert separate tokens with
    intervening whitespace back to the original matching input text. By
    default, returns astring containing the original parsed text.

    If the optional ``asString`` argument is passed as
    ``False``, then the return value is
    a :class:`ParseResults` containing any results names that
    were originally matched, and a single token containing the original
    matched text from the input string.  So if the expression passed to
    :class:`originalTextFor` contains expressions with defined
    results names, you must set ``asString`` to ``False`` if you
    want to preserve those results name values.

    Example::

        src = "this is test <b> bold <i>text</i> </b> normal text "
        for tag in ("b", "i"):
            opener, closer = makeHTMLTags(tag)
            patt = originalTextFor(opener + SkipTo(closer) + closer)
            print(patt.searchString(src)[0])

    prints::

        ['<b> bold <i>text</i> </b>']
        ['<i>text</i>']
    """
    locMarker = Empty().setParseAction(lambda s, loc, t: loc)
    endlocMarker = locMarker.copy()
    endlocMarker.callPreparse = False
    matchExpr = locMarker("_original_start") + expr + endlocMarker("_original_end")
    if asString:
        extractText = lambda s, l, t: s[t._original_start: t._original_end]
    else:
        def extractText(s, l, t):
            t[:] = [s[t.pop('_original_start'):t.pop('_original_end')]]
    matchExpr.setParseAction(extractText)
    matchExpr.ignoreExprs = expr.ignoreExprs
    return matchExpr

def ungroup(expr):
    """Helper to undo pyparsing's default grouping of And expressions,
    even if all but one are non-empty.
    """
    return TokenConverter(expr).addParseAction(lambda t: t[0])

def locatedExpr(expr):
    """Helper to decorate a returned token with its starting and ending
    locations in the input string.

    This helper adds the following results names:

     - locn_start = location where matched expression begins
     - locn_end = location where matched expression ends
     - value = the actual parsed results

    Be careful if the input text contains ``<TAB>`` characters, you
    may want to call :class:`ParserElement.parseWithTabs`

    Example::

        wd = Word(alphas)
        for match in locatedExpr(wd).searchString("ljsdf123lksdjjf123lkkjj1222"):
            print(match)

    prints::

        [[0, 'ljsdf', 5]]
        [[8, 'lksdjjf', 15]]
        [[18, 'lkkjj', 23]]
    """
    locator = Empty().setParseAction(lambda s, l, t: l)
    return Group(locator("locn_start") + expr("value") + locator.copy().leaveWhitespace()("locn_end"))


# convenience constants for positional expressions
empty       = Empty().setName("empty")
lineStart   = LineStart().setName("lineStart")
lineEnd     = LineEnd().setName("lineEnd")
stringStart = StringStart().setName("stringStart")
stringEnd   = StringEnd().setName("stringEnd")

_escapedPunc = Word(_bslash, r"\[]-*.$+^?()~ ", exact=2).setParseAction(lambda s, l, t: t[0][1])
_escapedHexChar = Regex(r"\\0?[xX][0-9a-fA-F]+").setParseAction(lambda s, l, t: unichr(int(t[0].lstrip(r'\0x'), 16)))
_escapedOctChar = Regex(r"\\0[0-7]+").setParseAction(lambda s, l, t: unichr(int(t[0][1:], 8)))
_singleChar = _escapedPunc | _escapedHexChar | _escapedOctChar | CharsNotIn(r'\]', exact=1)
_charRange = Group(_singleChar + Suppress("-") + _singleChar)
_reBracketExpr = Literal("[") + Optional("^").setResultsName("negate") + Group(OneOrMore(_charRange | _singleChar)).setResultsName("body") + "]"

def srange(s):
    r"""Helper to easily define string ranges for use in Word
    construction. Borrows syntax from regexp '[]' string range
    definitions::

        srange("[0-9]")   -> "0123456789"
        srange("[a-z]")   -> "abcdefghijklmnopqrstuvwxyz"
        srange("[a-z$_]") -> "abcdefghijklmnopqrstuvwxyz$_"

    The input string must be enclosed in []'s, and the returned string
    is the expanded character set joined into a single string. The
    values enclosed in the []'s may be:

     - a single character
     - an escaped character with a leading backslash (such as ``\-``
       or ``\]``)
     - an escaped hex character with a leading ``'\x'``
       (``\x21``, which is a ``'!'`` character) (``\0x##``
       is also supported for backwards compatibility)
     - an escaped octal character with a leading ``'\0'``
       (``\041``, which is a ``'!'`` character)
     - a range of any of the above, separated by a dash (``'a-z'``,
       etc.)
     - any combination of the above (``'aeiouy'``,
       ``'a-zA-Z0-9_$'``, etc.)
    """
    _expanded = lambda p: p if not isinstance(p, ParseResults) else ''.join(unichr(c) for c in range(ord(p[0]), ord(p[1]) + 1))
    try:
        return "".join(_expanded(part) for part in _reBracketExpr.parseString(s).body)
    except Exception:
        return ""

def matchOnlyAtCol(n):
    """Helper method for defining parse actions that require matching at
    a specific column in the input text.
    """
    def verifyCol(strg, locn, toks):
        if col(locn, strg) != n:
            raise ParseException(strg, locn, "matched token not at column %d" % n)
    return verifyCol

def replaceWith(replStr):
    """Helper method for common parse actions that simply return
    a literal value.  Especially useful when used with
    :class:`transformString<ParserElement.transformString>` ().

    Example::

        num = Word(nums).setParseAction(lambda toks: int(toks[0]))
        na = oneOf("N/A NA").setParseAction(replaceWith(math.nan))
        term = na | num

        OneOrMore(term).parseString("324 234 N/A 234") # -> [324, 234, nan, 234]
    """
    return lambda s, l, t: [replStr]

def removeQuotes(s, l, t):
    """Helper parse action for removing quotation marks from parsed
    quoted strings.

    Example::

        # by default, quotation marks are included in parsed results
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["'Now is the Winter of our Discontent'"]

        # use removeQuotes to strip quotation marks from parsed results
        quotedString.setParseAction(removeQuotes)
        quotedString.parseString("'Now is the Winter of our Discontent'") # -> ["Now is the Winter of our Discontent"]
    """
    return t[0][1:-1]

def tokenMap(func, *args):
    """Helper to define a parse action by mapping a function to all
    elements of a ParseResults list. If any additional args are passed,
    they are forwarded to the given function as additional arguments
    after the token, as in
    ``hex_integer = Word(hexnums).setParseAction(tokenMap(int, 16))``,
    which will convert the parsed data to an integer using base 16.

    Example (compare the last to example in :class:`ParserElement.transformString`::

        hex_ints = OneOrMore(Word(hexnums)).setParseAction(tokenMap(int, 16))
        hex_ints.runTests('''
            00 11 22 aa FF 0a 0d 1a
            ''')

        upperword = Word(alphas).setParseAction(tokenMap(str.upper))
        OneOrMore(upperword).runTests('''
            my kingdom for a horse
            ''')

        wd = Word(alphas).setParseAction(tokenMap(str.title))
        OneOrMore(wd).setParseAction(' '.join).runTests('''
            now is the winter of our discontent made glorious summer by this sun of york
            ''')

    prints::

        00 11 22 aa FF 0a 0d 1a
        [0, 17, 34, 170, 255, 10, 13, 26]

        my kingdom for a horse
        ['MY', 'KINGDOM', 'FOR', 'A', 'HORSE']

        now is the winter of our discontent made glorious summer by this sun of york
        ['Now Is The Winter Of Our Discontent Made Glorious Summer By This Sun Of York']
    """
    def pa(s, l, t):
        return [func(tokn, *args) for tokn in t]

    try:
        func_name = getattr(func, '__name__',
                            getattr(func, '__class__').__name__)
    except Exception:
        func_name = str(func)
    pa.__name__ = func_name

    return pa

upcaseTokens = tokenMap(lambda t: _ustr(t).upper())
"""(Deprecated) Helper parse action to convert tokens to upper case.
Deprecated in favor of :class:`pyparsing_common.upcaseTokens`"""

downcaseTokens = tokenMap(lambda t: _ustr(t).lower())
"""(Deprecated) Helper parse action to convert tokens to lower case.
Deprecated in favor of :class:`pyparsing_common.downcaseTokens`"""

def _makeTags(tagStr, xml,
              suppress_LT=Suppress("<"),
              suppress_GT=Suppress(">")):
    """Internal helper to construct opening and closing tag expressions, given a tag name"""
    if isinstance(tagStr, basestring):
        resname = tagStr
        tagStr = Keyword(tagStr, caseless=not xml)
    else:
        resname = tagStr.name

    tagAttrName = Word(alphas, alphanums + "_-:")
    if xml:
        tagAttrValue = dblQuotedString.copy().setParseAction(removeQuotes)
        openTag = (suppress_LT
                   + tagStr("tag")
                   + Dict(ZeroOrMore(Group(tagAttrName + Suppress("=") + tagAttrValue)))
                   + Optional("/", default=[False])("empty").setParseAction(lambda s, l, t: t[0] == '/')
                   + suppress_GT)
    else:
        tagAttrValue = quotedString.copy().setParseAction(removeQuotes) | Word(printables, excludeChars=">")
        openTag = (suppress_LT
                   + tagStr("tag")
                   + Dict(ZeroOrMore(Group(tagAttrName.setParseAction(downcaseTokens)
                                           + Optional(Suppress("=") + tagAttrValue))))
                   + Optional("/", default=[False])("empty").setParseAction(lambda s, l, t: t[0] == '/')
                   + suppress_GT)
    closeTag = Combine(_L("</") + tagStr + ">", adjacent=False)

    openTag.setName("<%s>" % resname)
    # add start<tagname> results name in parse action now that ungrouped names are not reported at two levels
    openTag.addParseAction(lambda t: t.__setitem__("start" + "".join(resname.replace(":", " ").title().split()), t.copy()))
    closeTag = closeTag("end" + "".join(resname.replace(":", " ").title().split())).setName("</%s>" % resname)
    openTag.tag = resname
    closeTag.tag = resname
    openTag.tag_body = SkipTo(closeTag())
    return openTag, closeTag

def makeHTMLTags(tagStr):
    """Helper to construct opening and closing tag expressions for HTML,
    given a tag name. Matches tags in either upper or lower case,
    attributes with namespaces and with quoted or unquoted values.

    Example::

        text = '<td>More info at the <a href="https://github.com/pyparsing/pyparsing/wiki">pyparsing</a> wiki page</td>'
        # makeHTMLTags returns pyparsing expressions for the opening and
        # closing tags as a 2-tuple
        a, a_end = makeHTMLTags("A")
        link_expr = a + SkipTo(a_end)("link_text") + a_end

        for link in link_expr.searchString(text):
            # attributes in the <A> tag (like "href" shown here) are
            # also accessible as named results
            print(link.link_text, '->', link.href)

    prints::

        pyparsing -> https://github.com/pyparsing/pyparsing/wiki
    """
    return _makeTags(tagStr, False)

def makeXMLTags(tagStr):
    """Helper to construct opening and closing tag expressions for XML,
    given a tag name. Matches tags only in the given upper/lower case.

    Example: similar to :class:`makeHTMLTags`
    """
    return _makeTags(tagStr, True)

def withAttribute(*args, **attrDict):
    """Helper to create a validating parse action to be used with start
    tags created with :class:`makeXMLTags` or
    :class:`makeHTMLTags`. Use ``withAttribute`` to qualify
    a starting tag with a required attribute value, to avoid false
    matches on common tags such as ``<TD>`` or ``<DIV>``.

    Call ``withAttribute`` with a series of attribute names and
    values. Specify the list of filter attributes names and values as:

     - keyword arguments, as in ``(align="right")``, or
     - as an explicit dict with ``**`` operator, when an attribute
       name is also a Python reserved word, as in ``**{"class":"Customer", "align":"right"}``
     - a list of name-value tuples, as in ``(("ns1:class", "Customer"), ("ns2:align", "right"))``

    For attribute names with a namespace prefix, you must use the second
    form.  Attribute names are matched insensitive to upper/lower case.

    If just testing for ``class`` (with or without a namespace), use
    :class:`withClass`.

    To verify that the attribute exists, but without specifying a value,
    pass ``withAttribute.ANY_VALUE`` as the value.

    Example::

        html = '''
            <div>
            Some text
            <div type="grid">1 4 0 1 0</div>
            <div type="graph">1,3 2,3 1,1</div>
            <div>this has no type</div>
            </div>

        '''
        div,div_end = makeHTMLTags("div")

        # only match div tag having a type attribute with value "grid"
        div_grid = div().setParseAction(withAttribute(type="grid"))
        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)

        # construct a match with any div tag having a type attribute, regardless of the value
        div_any_type = div().setParseAction(withAttribute(type=withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)

    prints::

        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    if args:
        attrs = args[:]
    else:
        attrs = attrDict.items()
    attrs = [(k, v) for k, v in attrs]
    def pa(s, l, tokens):
        for attrName, attrValue in attrs:
            if attrName not in tokens:
                raise ParseException(s, l, "no matching attribute " + attrName)
            if attrValue != withAttribute.ANY_VALUE and tokens[attrName] != attrValue:
                raise ParseException(s, l, "attribute '%s' has value '%s', must be '%s'" %
                                            (attrName, tokens[attrName], attrValue))
    return pa
withAttribute.ANY_VALUE = object()

def withClass(classname, namespace=''):
    """Simplified version of :class:`withAttribute` when
    matching on a div class - made difficult because ``class`` is
    a reserved word in Python.

    Example::

        html = '''
            <div>
            Some text
            <div class="grid">1 4 0 1 0</div>
            <div class="graph">1,3 2,3 1,1</div>
            <div>this &lt;div&gt; has no class</div>
            </div>

        '''
        div,div_end = makeHTMLTags("div")
        div_grid = div().setParseAction(withClass("grid"))

        grid_expr = div_grid + SkipTo(div | div_end)("body")
        for grid_header in grid_expr.searchString(html):
            print(grid_header.body)

        div_any_type = div().setParseAction(withClass(withAttribute.ANY_VALUE))
        div_expr = div_any_type + SkipTo(div | div_end)("body")
        for div_header in div_expr.searchString(html):
            print(div_header.body)

    prints::

        1 4 0 1 0

        1 4 0 1 0
        1,3 2,3 1,1
    """
    classattr = "%s:class" % namespace if namespace else "class"
    return withAttribute(**{classattr: classname})

opAssoc = SimpleNamespace()
opAssoc.LEFT = object()
opAssoc.RIGHT = object()

def infixNotation(baseExpr, opList, lpar=Suppress('('), rpar=Suppress(')')):
    """Helper method for constructing grammars of expressions made up of
    operators working in a precedence hierarchy.  Operators may be unary
    or binary, left- or right-associative.  Parse actions can also be
    attached to operator expressions. The generated parser will also
    recognize the use of parentheses to override operator precedences
    (see example below).

    Note: if you define a deep operator list, you may see performance
    issues when using infixNotation. See
    :class:`ParserElement.enablePackrat` for a mechanism to potentially
    improve your parser performance.

    Parameters:
     - baseExpr - expression representing the most basic element for the
       nested
     - opList - list of tuples, one for each operator precedence level
       in the expression grammar; each tuple is of the form ``(opExpr,
       numTerms, rightLeftAssoc, parseAction)``, where:

       - opExpr is the pyparsing expression for the operator; may also
         be a string, which will be converted to a Literal; if numTerms
         is 3, opExpr is a tuple of two expressions, for the two
         operators separating the 3 terms
       - numTerms is the number of terms for this operator (must be 1,
         2, or 3)
       - rightLeftAssoc is the indicator whether the operator is right
         or left associative, using the pyparsing-defined constants
         ``opAssoc.RIGHT`` and ``opAssoc.LEFT``.
       - parseAction is the parse action to be associated with
         expressions matching this operator expression (the parse action
         tuple member may be omitted); if the parse action is passed
         a tuple or list of functions, this is equivalent to calling
         ``setParseAction(*fn)``
         (:class:`ParserElement.setParseAction`)
     - lpar - expression for matching left-parentheses
       (default= ``Suppress('(')``)
     - rpar - expression for matching right-parentheses
       (default= ``Suppress(')')``)

    Example::

        # simple example of four-function arithmetic with ints and
        # variable names
        integer = pyparsing_common.signed_integer
        varname = pyparsing_common.identifier

        arith_expr = infixNotation(integer | varname,
            [
            ('-', 1, opAssoc.RIGHT),
            (oneOf('* /'), 2, opAssoc.LEFT),
            (oneOf('+ -'), 2, opAssoc.LEFT),
            ])

        arith_expr.runTests('''
            5+3*6
            (5+3)*6
            -2--11
            ''', fullDump=False)

    prints::

        5+3*6
        [[5, '+', [3, '*', 6]]]

        (5+3)*6
        [[[5, '+', 3], '*', 6]]

        -2--11
        [[['-', 2], '-', ['-', 11]]]
    """
    # captive version of FollowedBy that does not do parse actions or capture results names
    class _FB(FollowedBy):
        def parseImpl(self, instring, loc, doActions=True):
            self.expr.tryParse(instring, loc)
            return loc, []

    ret = Forward()
    lastExpr = baseExpr | (lpar + ret + rpar)
    for i, operDef in enumerate(opList):
        opExpr, arity, rightLeftAssoc, pa = (operDef + (None, ))[:4]
        termName = "%s term" % opExpr if arity < 3 else "%s%s term" % opExpr
        if arity == 3:
            if opExpr is None or len(opExpr) != 2:
                raise ValueError(
                    "if numterms=3, opExpr must be a tuple or list of two expressions")
            opExpr1, opExpr2 = opExpr
        thisExpr = Forward().setName(termName)
        if rightLeftAssoc == opAssoc.LEFT:
            if arity == 1:
                matchExpr = _FB(lastExpr + opExpr) + Group(lastExpr + OneOrMore(opExpr))
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = _FB(lastExpr + opExpr + lastExpr) + Group(lastExpr + OneOrMore(opExpr + lastExpr))
                else:
                    matchExpr = _FB(lastExpr + lastExpr) + Group(lastExpr + OneOrMore(lastExpr))
            elif arity == 3:
                matchExpr = (_FB(lastExpr + opExpr1 + lastExpr + opExpr2 + lastExpr)
                             + Group(lastExpr + OneOrMore(opExpr1 + lastExpr + opExpr2 + lastExpr)))
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        elif rightLeftAssoc == opAssoc.RIGHT:
            if arity == 1:
                # try to avoid LR with this extra test
                if not isinstance(opExpr, Optional):
                    opExpr = Optional(opExpr)
                matchExpr = _FB(opExpr.expr + thisExpr) + Group(opExpr + thisExpr)
            elif arity == 2:
                if opExpr is not None:
                    matchExpr = _FB(lastExpr + opExpr + thisExpr) + Group(lastExpr + OneOrMore(opExpr + thisExpr))
                else:
                    matchExpr = _FB(lastExpr + thisExpr) + Group(lastExpr + OneOrMore(thisExpr))
            elif arity == 3:
                matchExpr = (_FB(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr)
                             + Group(lastExpr + opExpr1 + thisExpr + opExpr2 + thisExpr))
            else:
                raise ValueError("operator must be unary (1), binary (2), or ternary (3)")
        else:
            raise ValueError("operator must indicate right or left associativity")
        if pa:
            if isinstance(pa, (tuple, list)):
                matchExpr.setParseAction(*pa)
            else:
                matchExpr.setParseAction(pa)
        thisExpr <<= (matchExpr.setName(termName) | lastExpr)
        lastExpr = thisExpr
    ret <<= lastExpr
    return ret

operatorPrecedence = infixNotation
"""(Deprecated) Former name of :class:`infixNotation`, will be
dropped in a future release."""

dblQuotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*') + '"').setName("string enclosed in double quotes")
sglQuotedString = Combine(Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*") + "'").setName("string enclosed in single quotes")
quotedString = Combine(Regex(r'"(?:[^"\n\r\\]|(?:"")|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*') + '"'
                       | Regex(r"'(?:[^'\n\r\\]|(?:'')|(?:\\(?:[^x]|x[0-9a-fA-F]+)))*") + "'").setName("quotedString using single or double quotes")
unicodeString = Combine(_L('u') + quotedString.copy()).setName("unicode string literal")

def nestedExpr(opener="(", closer=")", content=None, ignoreExpr=quotedString.copy()):
    """Helper method for defining nested lists enclosed in opening and
    closing delimiters ("(" and ")" are the default).

    Parameters:
     - opener - opening character for a nested list
       (default= ``"("``); can also be a pyparsing expression
     - closer - closing character for a nested list
       (default= ``")"``); can also be a pyparsing expression
     - content - expression for items within the nested lists
       (default= ``None``)
     - ignoreExpr - expression for ignoring opening and closing
       delimiters (default= :class:`quotedString`)

    If an expression is not provided for the content argument, the
    nested expression will capture all whitespace-delimited content
    between delimiters as a list of separate values.

    Use the ``ignoreExpr`` argument to define expressions that may
    contain opening or closing characters that should not be treated as
    opening or closing characters for nesting, such as quotedString or
    a comment expression.  Specify multiple expressions using an
    :class:`Or` or :class:`MatchFirst`. The default is
    :class:`quotedString`, but if no expressions are to be ignored, then
    pass ``None`` for this argument.

    Example::

        data_type = oneOf("void int short long char float double")
        decl_data_type = Combine(data_type + Optional(Word('*')))
        ident = Word(alphas+'_', alphanums+'_')
        number = pyparsing_common.number
        arg = Group(decl_data_type + ident)
        LPAR, RPAR = map(Suppress, "()")

        code_body = nestedExpr('{', '}', ignoreExpr=(quotedString | cStyleComment))

        c_function = (decl_data_type("type")
                      + ident("name")
                      + LPAR + Optional(delimitedList(arg), [])("args") + RPAR
                      + code_body("body"))
        c_function.ignore(cStyleComment)

        source_code = '''
            int is_odd(int x) {
                return (x%2);
            }

            int dec_to_hex(char hchar) {
                if (hchar >= '0' && hchar <= '9') {
                    return (ord(hchar)-ord('0'));
                } else {
                    return (10+ord(hchar)-ord('A'));
                }
            }
        '''
        for func in c_function.searchString(source_code):
            print("%(name)s (%(type)s) args: %(args)s" % func)


    prints::

        is_odd (int) args: [['int', 'x']]
        dec_to_hex (int) args: [['char', 'hchar']]
    """
    if opener == closer:
        raise ValueError("opening and closing strings cannot be the same")
    if content is None:
        if isinstance(opener, basestring) and isinstance(closer, basestring):
            if len(opener) == 1 and len(closer) == 1:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr
                                                 + CharsNotIn(opener
                                                              + closer
                                                              + ParserElement.DEFAULT_WHITE_CHARS, exact=1)
                                                 )
                                       ).setParseAction(lambda t: t[0].strip()))
                else:
                    content = (empty.copy() + CharsNotIn(opener
                                                         + closer
                                                         + ParserElement.DEFAULT_WHITE_CHARS
                                                         ).setParseAction(lambda t: t[0].strip()))
            else:
                if ignoreExpr is not None:
                    content = (Combine(OneOrMore(~ignoreExpr
                                                 + ~Literal(opener)
                                                 + ~Literal(closer)
                                                 + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS, exact=1))
                                       ).setParseAction(lambda t: t[0].strip()))
                else:
                    content = (Combine(OneOrMore(~Literal(opener)
                                                 + ~Literal(closer)
                                                 + CharsNotIn(ParserElement.DEFAULT_WHITE_CHARS, exact=1))
                                       ).setParseAction(lambda t: t[0].strip()))
        else:
            raise ValueError("opening and closing arguments must be strings if no content expression is given")
    ret = Forward()
    if ignoreExpr is not None:
        ret <<= Group(Suppress(opener) + ZeroOrMore(ignoreExpr | ret | content) + Suppress(closer))
    else:
        ret <<= Group(Suppress(opener) + ZeroOrMore(ret | content)  + Suppress(closer))
    ret.setName('nested %s%s expression' % (opener, closer))
    return ret

def indentedBlock(blockStatementExpr, indentStack, indent=True):
    """Helper method for defining space-delimited indentation blocks,
    such as those used to define block statements in Python source code.

    Parameters:

     - blockStatementExpr - expression defining syntax of statement that
       is repeated within the indented block
     - indentStack - list created by caller to manage indentation stack
       (multiple statementWithIndentedBlock expressions within a single
       grammar should share a common indentStack)
     - indent - boolean indicating whether block must be indented beyond
       the current level; set to False for block of left-most
       statements (default= ``True``)

    A valid block must contain at least one ``blockStatement``.

    Example::

        data = '''
        def A(z):
          A1
          B = 100
          G = A2
          A2
          A3
        B
        def BB(a,b,c):
          BB1
          def BBA():
            bba1
            bba2
            bba3
        C
        D
        def spam(x,y):
             def eggs(z):
                 pass
        '''


        indentStack = [1]
        stmt = Forward()

        identifier = Word(alphas, alphanums)
        funcDecl = ("def" + identifier + Group("(" + Optional(delimitedList(identifier)) + ")") + ":")
        func_body = indentedBlock(stmt, indentStack)
        funcDef = Group(funcDecl + func_body)

        rvalue = Forward()
        funcCall = Group(identifier + "(" + Optional(delimitedList(rvalue)) + ")")
        rvalue << (funcCall | identifier | Word(nums))
        assignment = Group(identifier + "=" + rvalue)
        stmt << (funcDef | assignment | identifier)

        module_body = OneOrMore(stmt)

        parseTree = module_body.parseString(data)
        parseTree.pprint()

    prints::

        [['def',
          'A',
          ['(', 'z', ')'],
          ':',
          [['A1'], [['B', '=', '100']], [['G', '=', 'A2']], ['A2'], ['A3']]],
         'B',
         ['def',
          'BB',
          ['(', 'a', 'b', 'c', ')'],
          ':',
          [['BB1'], [['def', 'BBA', ['(', ')'], ':', [['bba1'], ['bba2'], ['bba3']]]]]],
         'C',
         'D',
         ['def',
          'spam',
          ['(', 'x', 'y', ')'],
          ':',
          [[['def', 'eggs', ['(', 'z', ')'], ':', [['pass']]]]]]]
    """
    backup_stack = indentStack[:]

    def reset_stack():
        indentStack[:] = backup_stack

    def checkPeerIndent(s, l, t):
        if l >= len(s): return
        curCol = col(l, s)
        if curCol != indentStack[-1]:
            if curCol > indentStack[-1]:
                raise ParseException(s, l, "illegal nesting")
            raise ParseException(s, l, "not a peer entry")

    def checkSubIndent(s, l, t):
        curCol = col(l, s)
        if curCol > indentStack[-1]:
            indentStack.append(curCol)
        else:
            raise ParseException(s, l, "not a subentry")

    def checkUnindent(s, l, t):
        if l >= len(s): return
        curCol = col(l, s)
        if not(indentStack and curCol in indentStack):
            raise ParseException(s, l, "not an unindent")
        if curCol < indentStack[-1]:
            indentStack.pop()

    NL = OneOrMore(LineEnd().setWhitespaceChars("\t ").suppress(), stopOn=StringEnd())
    INDENT = (Empty() + Empty().setParseAction(checkSubIndent)).setName('INDENT')
    PEER   = Empty().setParseAction(checkPeerIndent).setName('')
    UNDENT = Empty().setParseAction(checkUnindent).setName('UNINDENT')
    if indent:
        smExpr = Group(Optional(NL)
                       + INDENT
                       + OneOrMore(PEER + Group(blockStatementExpr) + Optional(NL), stopOn=StringEnd())
                       + UNDENT)
    else:
        smExpr = Group(Optional(NL)
                       + OneOrMore(PEER + Group(blockStatementExpr) + Optional(NL), stopOn=StringEnd())
                       + UNDENT)
    smExpr.setFailAction(lambda a, b, c, d: reset_stack())
    blockStatementExpr.ignore(_bslash + LineEnd())
    return smExpr.setName('indented block')

alphas8bit = srange(r"[\0xc0-\0xd6\0xd8-\0xf6\0xf8-\0xff]")
punc8bit = srange(r"[\0xa1-\0xbf\0xd7\0xf7]")

anyOpenTag, anyCloseTag = makeHTMLTags(Word(alphas, alphanums + "_:").setName('any tag'))
_htmlEntityMap = dict(zip("gt lt amp nbsp quot apos".split(), '><& "\''))
commonHTMLEntity = Regex('&(?P<entity>' + '|'.join(_htmlEntityMap.keys()) +");").setName("common HTML entity")
def replaceHTMLEntity(t):
    """Helper parser action to replace common HTML entities with their special characters"""
    return _htmlEntityMap.get(t.entity)

# it's easy to get these comment structures wrong - they're very common, so may as well make them available
cStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/').setName("C style comment")
"Comment of the form ``/* ... */``"

htmlComment = Regex(r"<!--[\s\S]*?-->").setName("HTML comment")
"Comment of the form ``<!-- ... -->``"

restOfLine = Regex(r".*").leaveWhitespace().setName("rest of line")
dblSlashComment = Regex(r"//(?:\\\n|[^\n])*").setName("// comment")
"Comment of the form ``// ... (to end of line)``"

cppStyleComment = Combine(Regex(r"/\*(?:[^*]|\*(?!/))*") + '*/' | dblSlashComment).setName("C++ style comment")
"Comment of either form :class:`cStyleComment` or :class:`dblSlashComment`"

javaStyleComment = cppStyleComment
"Same as :class:`cppStyleComment`"

pythonStyleComment = Regex(r"#.*").setName("Python style comment")
"Comment of the form ``# ... (to end of line)``"

_commasepitem = Combine(OneOrMore(Word(printables, excludeChars=',')
                                  + Optional(Word(" \t")
                                             + ~Literal(",") + ~LineEnd()))).streamline().setName("commaItem")
commaSeparatedList = delimitedList(Optional(quotedString.copy() | _commasepitem, default="")).setName("commaSeparatedList")
"""(Deprecated) Predefined expression of 1 or more printable words or
quoted strings, separated by commas.

This expression is deprecated in favor of :class:`pyparsing_common.comma_separated_list`.
"""

# some other useful expressions - using lower-case class name since we are really using this as a namespace
class pyparsing_common:
    """Here are some common low-level expressions that may be useful in
    jump-starting parser development:

     - numeric forms (:class:`integers<integer>`, :class:`reals<real>`,
       :class:`scientific notation<sci_real>`)
     - common :class:`programming identifiers<identifier>`
     - network addresses (:class:`MAC<mac_address>`,
       :class:`IPv4<ipv4_address>`, :class:`IPv6<ipv6_address>`)
     - ISO8601 :class:`dates<iso8601_date>` and
       :class:`datetime<iso8601_datetime>`
     - :class:`UUID<uuid>`
     - :class:`comma-separated list<comma_separated_list>`

    Parse actions:

     - :class:`convertToInteger`
     - :class:`convertToFloat`
     - :class:`convertToDate`
     - :class:`convertToDatetime`
     - :class:`stripHTMLTags`
     - :class:`upcaseTokens`
     - :class:`downcaseTokens`

    Example::

        pyparsing_common.number.runTests('''
            # any int or real number, returned as the appropriate type
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.fnumber.runTests('''
            # any int or real number, returned as float
            100
            -100
            +100
            3.14159
            6.02e23
            1e-12
            ''')

        pyparsing_common.hex_integer.runTests('''
            # hex numbers
            100
            FF
            ''')

        pyparsing_common.fraction.runTests('''
            # fractions
            1/2
            -3/4
            ''')

        pyparsing_common.mixed_integer.runTests('''
            # mixed fractions
            1
            1/2
            -3/4
            1-3/4
            ''')

        import uuid
        pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
        pyparsing_common.uuid.runTests('''
            # uuid
            12345678-1234-5678-1234-567812345678
            ''')

    prints::

        # any int or real number, returned as the appropriate type
        100
        [100]

        -100
        [-100]

        +100
        [100]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # any int or real number, returned as float
        100
        [100.0]

        -100
        [-100.0]

        +100
        [100.0]

        3.14159
        [3.14159]

        6.02e23
        [6.02e+23]

        1e-12
        [1e-12]

        # hex numbers
        100
        [256]

        FF
        [255]

        # fractions
        1/2
        [0.5]

        -3/4
        [-0.75]

        # mixed fractions
        1
        [1]

        1/2
        [0.5]

        -3/4
        [-0.75]

        1-3/4
        [1.75]

        # uuid
        12345678-1234-5678-1234-567812345678
        [UUID('12345678-1234-5678-1234-567812345678')]
    """

    convertToInteger = tokenMap(int)
    """
    Parse action for converting parsed integers to Python int
    """

    convertToFloat = tokenMap(float)
    """
    Parse action for converting parsed numbers to Python float
    """

    integer = Word(nums).setName("integer").setParseAction(convertToInteger)
    """expression that parses an unsigned integer, returns an int"""

    hex_integer = Word(hexnums).setName("hex integer").setParseAction(tokenMap(int, 16))
    """expression that parses a hexadecimal integer, returns an int"""

    signed_integer = Regex(r'[+-]?\d+').setName("signed integer").setParseAction(convertToInteger)
    """expression that parses an integer with optional leading sign, returns an int"""

    fraction = (signed_integer().setParseAction(convertToFloat) + '/' + signed_integer().setParseAction(convertToFloat)).setName("fraction")
    """fractional expression of an integer divided by an integer, returns a float"""
    fraction.addParseAction(lambda t: t[0]/t[-1])

    mixed_integer = (fraction | signed_integer + Optional(Optional('-').suppress() + fraction)).setName("fraction or mixed integer-fraction")
    """mixed integer of the form 'integer - fraction', with optional leading integer, returns float"""
    mixed_integer.addParseAction(sum)

    real = Regex(r'[+-]?(?:\d+\.\d*|\.\d+)').setName("real number").setParseAction(convertToFloat)
    """expression that parses a floating point number and returns a float"""

    sci_real = Regex(r'[+-]?(?:\d+(?:[eE][+-]?\d+)|(?:\d+\.\d*|\.\d+)(?:[eE][+-]?\d+)?)').setName("real number with scientific notation").setParseAction(convertToFloat)
    """expression that parses a floating point number with optional
    scientific notation and returns a float"""

    # streamlining this expression makes the docs nicer-looking
    number = (sci_real | real | signed_integer).streamline()
    """any numeric expression, returns the corresponding Python type"""

    fnumber = Regex(r'[+-]?\d+\.?\d*([eE][+-]?\d+)?').setName("fnumber").setParseAction(convertToFloat)
    """any int or real number, returned as float"""

    identifier = Word(alphas + '_', alphanums + '_').setName("identifier")
    """typical code identifier (leading alpha or '_', followed by 0 or more alphas, nums, or '_')"""

    ipv4_address = Regex(r'(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})(\.(25[0-5]|2[0-4][0-9]|1?[0-9]{1,2})){3}').setName("IPv4 address")
    "IPv4 address (``0.0.0.0 - 255.255.255.255``)"

    _ipv6_part = Regex(r'[0-9a-fA-F]{1,4}').setName("hex_integer")
    _full_ipv6_address = (_ipv6_part + (':' + _ipv6_part) * 7).setName("full IPv6 address")
    _short_ipv6_address = (Optional(_ipv6_part + (':' + _ipv6_part) * (0, 6))
                           + "::"
                           + Optional(_ipv6_part + (':' + _ipv6_part) * (0, 6))
                           ).setName("short IPv6 address")
    _short_ipv6_address.addCondition(lambda t: sum(1 for tt in t if pyparsing_common._ipv6_part.matches(tt)) < 8)
    _mixed_ipv6_address = ("::ffff:" + ipv4_address).setName("mixed IPv6 address")
    ipv6_address = Combine((_full_ipv6_address | _mixed_ipv6_address | _short_ipv6_address).setName("IPv6 address")).setName("IPv6 address")
    "IPv6 address (long, short, or mixed form)"

    mac_address = Regex(r'[0-9a-fA-F]{2}([:.-])[0-9a-fA-F]{2}(?:\1[0-9a-fA-F]{2}){4}').setName("MAC address")
    "MAC address xx:xx:xx:xx:xx (may also have '-' or '.' delimiters)"

    @staticmethod
    def convertToDate(fmt="%Y-%m-%d"):
        """
        Helper to create a parse action for converting parsed date string to Python datetime.date

        Params -
         - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%d"``)

        Example::

            date_expr = pyparsing_common.iso8601_date.copy()
            date_expr.setParseAction(pyparsing_common.convertToDate())
            print(date_expr.parseString("1999-12-31"))

        prints::

            [datetime.date(1999, 12, 31)]
        """
        def cvt_fn(s, l, t):
            try:
                return datetime.strptime(t[0], fmt).date()
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    @staticmethod
    def convertToDatetime(fmt="%Y-%m-%dT%H:%M:%S.%f"):
        """Helper to create a parse action for converting parsed
        datetime string to Python datetime.datetime

        Params -
         - fmt - format to be passed to datetime.strptime (default= ``"%Y-%m-%dT%H:%M:%S.%f"``)

        Example::

            dt_expr = pyparsing_common.iso8601_datetime.copy()
            dt_expr.setParseAction(pyparsing_common.convertToDatetime())
            print(dt_expr.parseString("1999-12-31T23:59:59.999"))

        prints::

            [datetime.datetime(1999, 12, 31, 23, 59, 59, 999000)]
        """
        def cvt_fn(s, l, t):
            try:
                return datetime.strptime(t[0], fmt)
            except ValueError as ve:
                raise ParseException(s, l, str(ve))
        return cvt_fn

    iso8601_date = Regex(r'(?P<year>\d{4})(?:-(?P<month>\d\d)(?:-(?P<day>\d\d))?)?').setName("ISO8601 date")
    "ISO8601 date (``yyyy-mm-dd``)"

    iso8601_datetime = Regex(r'(?P<year>\d{4})-(?P<month>\d\d)-(?P<day>\d\d)[T ](?P<hour>\d\d):(?P<minute>\d\d)(:(?P<second>\d\d(\.\d*)?)?)?(?P<tz>Z|[+-]\d\d:?\d\d)?').setName("ISO8601 datetime")
    "ISO8601 datetime (``yyyy-mm-ddThh:mm:ss.s(Z|+-00:00)``) - trailing seconds, milliseconds, and timezone optional; accepts separating ``'T'`` or ``' '``"

    uuid = Regex(r'[0-9a-fA-F]{8}(-[0-9a-fA-F]{4}){3}-[0-9a-fA-F]{12}').setName("UUID")
    "UUID (``xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx``)"

    _html_stripper = anyOpenTag.suppress() | anyCloseTag.suppress()
    @staticmethod
    def stripHTMLTags(s, l, tokens):
        """Parse action to remove HTML tags from web page HTML source

        Example::

            # strip HTML links from normal text
            text = '<td>More info at the <a href="https://github.com/pyparsing/pyparsing/wiki">pyparsing</a> wiki page</td>'
            td, td_end = makeHTMLTags("TD")
            table_text = td + SkipTo(td_end).setParseAction(pyparsing_common.stripHTMLTags)("body") + td_end
            print(table_text.parseString(text).body)

        Prints::

            More info at the pyparsing wiki page
        """
        return pyparsing_common._html_stripper.transformString(tokens[0])

    _commasepitem = Combine(OneOrMore(~Literal(",")
                                      + ~LineEnd()
                                      + Word(printables, excludeChars=',')
                                      + Optional(White(" \t")))).streamline().setName("commaItem")
    comma_separated_list = delimitedList(Optional(quotedString.copy()
                                                  | _commasepitem, default='')
                                         ).setName("comma separated list")
    """Predefined expression of 1 or more printable words or quoted strings, separated by commas."""

    upcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).upper()))
    """Parse action to convert tokens to upper case."""

    downcaseTokens = staticmethod(tokenMap(lambda t: _ustr(t).lower()))
    """Parse action to convert tokens to lower case."""


class _lazyclassproperty(object):
    def __init__(self, fn):
        self.fn = fn
        self.__doc__ = fn.__doc__
        self.__name__ = fn.__name__

    def __get__(self, obj, cls):
        if cls is None:
            cls = type(obj)
        if not hasattr(cls, '_intern') or any(cls._intern is getattr(superclass, '_intern', [])
                                              for superclass in cls.__mro__[1:]):
            cls._intern = {}
        attrname = self.fn.__name__
        if attrname not in cls._intern:
            cls._intern[attrname] = self.fn(cls)
        return cls._intern[attrname]


class unicode_set(object):
    """
    A set of Unicode characters, for language-specific strings for
    ``alphas``, ``nums``, ``alphanums``, and ``printables``.
    A unicode_set is defined by a list of ranges in the Unicode character
    set, in a class attribute ``_ranges``, such as::

        _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),]

    A unicode set can also be defined using multiple inheritance of other unicode sets::

        class CJK(Chinese, Japanese, Korean):
            pass
    """
    _ranges = []

    @classmethod
    def _get_chars_for_ranges(cls):
        ret = []
        for cc in cls.__mro__:
            if cc is unicode_set:
                break
            for rr in cc._ranges:
                ret.extend(range(rr[0], rr[-1] + 1))
        return [unichr(c) for c in sorted(set(ret))]

    @_lazyclassproperty
    def printables(cls):
        "all non-whitespace characters in this range"
        return u''.join(filterfalse(unicode.isspace, cls._get_chars_for_ranges()))

    @_lazyclassproperty
    def alphas(cls):
        "all alphabetic characters in this range"
        return u''.join(filter(unicode.isalpha, cls._get_chars_for_ranges()))

    @_lazyclassproperty
    def nums(cls):
        "all numeric digit characters in this range"
        return u''.join(filter(unicode.isdigit, cls._get_chars_for_ranges()))

    @_lazyclassproperty
    def alphanums(cls):
        "all alphanumeric characters in this range"
        return cls.alphas + cls.nums


class pyparsing_unicode(unicode_set):
    """
    A namespace class for defining common language unicode_sets.
    """
    _ranges = [(32, sys.maxunicode)]

    class Latin1(unicode_set):
        "Unicode set for Latin-1 Unicode Character Range"
        _ranges = [(0x0020, 0x007e), (0x00a0, 0x00ff),]

    class LatinA(unicode_set):
        "Unicode set for Latin-A Unicode Character Range"
        _ranges = [(0x0100, 0x017f),]

    class LatinB(unicode_set):
        "Unicode set for Latin-B Unicode Character Range"
        _ranges = [(0x0180, 0x024f),]

    class Greek(unicode_set):
        "Unicode set for Greek Unicode Character Ranges"
        _ranges = [
            (0x0370, 0x03ff), (0x1f00, 0x1f15), (0x1f18, 0x1f1d), (0x1f20, 0x1f45), (0x1f48, 0x1f4d),
            (0x1f50, 0x1f57), (0x1f59,), (0x1f5b,), (0x1f5d,), (0x1f5f, 0x1f7d), (0x1f80, 0x1fb4), (0x1fb6, 0x1fc4),
            (0x1fc6, 0x1fd3), (0x1fd6, 0x1fdb), (0x1fdd, 0x1fef), (0x1ff2, 0x1ff4), (0x1ff6, 0x1ffe),
        ]

    class Cyrillic(unicode_set):
        "Unicode set for Cyrillic Unicode Character Range"
        _ranges = [(0x0400, 0x04ff)]

    class Chinese(unicode_set):
        "Unicode set for Chinese Unicode Character Range"
        _ranges = [(0x4e00, 0x9fff), (0x3000, 0x303f),]

    class Japanese(unicode_set):
        "Unicode set for Japanese Unicode Character Range, combining Kanji, Hiragana, and Katakana ranges"
        _ranges = []

        class Kanji(unicode_set):
            "Unicode set for Kanji Unicode Character Range"
            _ranges = [(0x4E00, 0x9Fbf), (0x3000, 0x303f),]

        class Hiragana(unicode_set):
            "Unicode set for Hiragana Unicode Character Range"
            _ranges = [(0x3040, 0x309f),]

        class Katakana(unicode_set):
            "Unicode set for Katakana  Unicode Character Range"
            _ranges = [(0x30a0, 0x30ff),]

    class Korean(unicode_set):
        "Unicode set for Korean Unicode Character Range"
        _ranges = [(0xac00, 0xd7af), (0x1100, 0x11ff), (0x3130, 0x318f), (0xa960, 0xa97f), (0xd7b0, 0xd7ff), (0x3000, 0x303f),]

    class CJK(Chinese, Japanese, Korean):
        "Unicode set for combined Chinese, Japanese, and Korean (CJK) Unicode Character Range"
        pass

    class Thai(unicode_set):
        "Unicode set for Thai Unicode Character Range"
        _ranges = [(0x0e01, 0x0e3a), (0x0e3f, 0x0e5b),]

    class Arabic(unicode_set):
        "Unicode set for Arabic Unicode Character Range"
        _ranges = [(0x0600, 0x061b), (0x061e, 0x06ff), (0x0700, 0x077f),]

    class Hebrew(unicode_set):
        "Unicode set for Hebrew Unicode Character Range"
        _ranges = [(0x0590, 0x05ff),]

    class Devanagari(unicode_set):
        "Unicode set for Devanagari Unicode Character Range"
        _ranges = [(0x0900, 0x097f), (0xa8e0, 0xa8ff)]

pyparsing_unicode.Japanese._ranges = (pyparsing_unicode.Japanese.Kanji._ranges
                                      + pyparsing_unicode.Japanese.Hiragana._ranges
                                      + pyparsing_unicode.Japanese.Katakana._ranges)

# define ranges in language character sets
if PY_3:
    setattr(pyparsing_unicode, u"العربية", pyparsing_unicode.Arabic)
    setattr(pyparsing_unicode, u"中文", pyparsing_unicode.Chinese)
    setattr(pyparsing_unicode, u"кириллица", pyparsing_unicode.Cyrillic)
    setattr(pyparsing_unicode, u"Ελληνικά", pyparsing_unicode.Greek)
    setattr(pyparsing_unicode, u"עִברִית", pyparsing_unicode.Hebrew)
    setattr(pyparsing_unicode, u"日本語", pyparsing_unicode.Japanese)
    setattr(pyparsing_unicode.Japanese, u"漢字", pyparsing_unicode.Japanese.Kanji)
    setattr(pyparsing_unicode.Japanese, u"カタカナ", pyparsing_unicode.Japanese.Katakana)
    setattr(pyparsing_unicode.Japanese, u"ひらがな", pyparsing_unicode.Japanese.Hiragana)
    setattr(pyparsing_unicode, u"한국어", pyparsing_unicode.Korean)
    setattr(pyparsing_unicode, u"ไทย", pyparsing_unicode.Thai)
    setattr(pyparsing_unicode, u"देवनागरी", pyparsing_unicode.Devanagari)


class pyparsing_test:
    """
    namespace class for classes useful in writing unit tests
    """

    class reset_pyparsing_context:
        """
        Context manager to be used when writing unit tests that modify pyparsing config values:
         - packrat parsing
         - default whitespace characters.
         - default keyword characters
         - literal string auto-conversion class
         - __diag__ settings

        Example:
            with reset_pyparsing_context():
                # test that literals used to construct a grammar are automatically suppressed
                ParserElement.inlineLiteralsUsing(Suppress)

                term = Word(alphas) | Word(nums)
                group = Group('(' + term[...] + ')')

                # assert that the '()' characters are not included in the parsed tokens
                self.assertParseAndCheckLisst(group, "(abc 123 def)", ['abc', '123', 'def'])

            # after exiting context manager, literals are converted to Literal expressions again
        """

        def __init__(self):
            self._save_context = {}

        def save(self):
            self._save_context["default_whitespace"] = ParserElement.DEFAULT_WHITE_CHARS
            self._save_context["default_keyword_chars"] = Keyword.DEFAULT_KEYWORD_CHARS
            self._save_context[
                "literal_string_class"
            ] = ParserElement._literalStringClass
            self._save_context["packrat_enabled"] = ParserElement._packratEnabled
            self._save_context["packrat_parse"] = ParserElement._parse
            self._save_context["__diag__"] = {
                name: getattr(__diag__, name) for name in __diag__._all_names
            }
            self._save_context["__compat__"] = {
                "collect_all_And_tokens": __compat__.collect_all_And_tokens
            }
            return self

        def restore(self):
            # reset pyparsing global state
            if (
                ParserElement.DEFAULT_WHITE_CHARS
                != self._save_context["default_whitespace"]
            ):
                ParserElement.setDefaultWhitespaceChars(
                    self._save_context["default_whitespace"]
                )
            Keyword.DEFAULT_KEYWORD_CHARS = self._save_context["default_keyword_chars"]
            ParserElement.inlineLiteralsUsing(
                self._save_context["literal_string_class"]
            )
            for name, value in self._save_context["__diag__"].items():
                setattr(__diag__, name, value)
            ParserElement._packratEnabled = self._save_context["packrat_enabled"]
            ParserElement._parse = self._save_context["packrat_parse"]
            __compat__.collect_all_And_tokens = self._save_context["__compat__"]

        def __enter__(self):
            return self.save()

        def __exit__(self, *args):
            return self.restore()

    class TestParseResultsAsserts:
        """
        A mixin class to add parse results assertion methods to normal unittest.TestCase classes.
        """
        def assertParseResultsEquals(
            self, result, expected_list=None, expected_dict=None, msg=None
        ):
            """
            Unit test assertion to compare a ParseResults object with an optional expected_list,
            and compare any defined results names with an optional expected_dict.
            """
            if expected_list is not None:
                self.assertEqual(expected_list, result.asList(), msg=msg)
            if expected_dict is not None:
                self.assertEqual(expected_dict, result.asDict(), msg=msg)

        def assertParseAndCheckList(
            self, expr, test_string, expected_list, msg=None, verbose=True
        ):
            """
            Convenience wrapper assert to test a parser element and input string, and assert that
            the resulting ParseResults.asList() is equal to the expected_list.
            """
            result = expr.parseString(test_string, parseAll=True)
            if verbose:
                print(result.dump())
            self.assertParseResultsEquals(result, expected_list=expected_list, msg=msg)

        def assertParseAndCheckDict(
            self, expr, test_string, expected_dict, msg=None, verbose=True
        ):
            """
            Convenience wrapper assert to test a parser element and input string, and assert that
            the resulting ParseResults.asDict() is equal to the expected_dict.
            """
            result = expr.parseString(test_string, parseAll=True)
            if verbose:
                print(result.dump())
            self.assertParseResultsEquals(result, expected_dict=expected_dict, msg=msg)

        def assertRunTestResults(
            self, run_tests_report, expected_parse_results=None, msg=None
        ):
            """
            Unit test assertion to evaluate output of ParserElement.runTests(). If a list of
            list-dict tuples is given as the expected_parse_results argument, then these are zipped
            with the report tuples returned by runTests and evaluated using assertParseResultsEquals.
            Finally, asserts that the overall runTests() success value is True.

            :param run_tests_report: tuple(bool, [tuple(str, ParseResults or Exception)]) returned from runTests
            :param expected_parse_results (optional): [tuple(str, list, dict, Exception)]
            """
            run_test_success, run_test_results = run_tests_report

            if expected_parse_results is not None:
                merged = [
                    (rpt[0], rpt[1], expected)
                    for rpt, expected in zip(run_test_results, expected_parse_results)
                ]
                for test_string, result, expected in merged:
                    # expected should be a tuple containing a list and/or a dict or an exception,
                    # and optional failure message string
                    # an empty tuple will skip any result validation
                    fail_msg = next(
                        (exp for exp in expected if isinstance(exp, str)), None
                    )
                    expected_exception = next(
                        (
                            exp
                            for exp in expected
                            if isinstance(exp, type) and issubclass(exp, Exception)
                        ),
                        None,
                    )
                    if expected_exception is not None:
                        with self.assertRaises(
                            expected_exception=expected_exception, msg=fail_msg or msg
                        ):
                            if isinstance(result, Exception):
                                raise result
                    else:
                        expected_list = next(
                            (exp for exp in expected if isinstance(exp, list)), None
                        )
                        expected_dict = next(
                            (exp for exp in expected if isinstance(exp, dict)), None
                        )
                        if (expected_list, expected_dict) != (None, None):
                            self.assertParseResultsEquals(
                                result,
                                expected_list=expected_list,
                                expected_dict=expected_dict,
                                msg=fail_msg or msg,
                            )
                        else:
                            # warning here maybe?
                            print("no validation for {!r}".format(test_string))

            # do this last, in case some specific test results can be reported instead
            self.assertTrue(
                run_test_success, msg=msg if msg is not None else "failed runTests"
            )

        @contextmanager
        def assertRaisesParseException(self, exc_type=ParseException, msg=None):
            with self.assertRaises(exc_type, msg=msg):
                yield


if __name__ == "__main__":

    selectToken    = CaselessLiteral("select")
    fromToken      = CaselessLiteral("from")

    ident          = Word(alphas, alphanums + "_$")

    columnName     = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    columnNameList = Group(delimitedList(columnName)).setName("columns")
    columnSpec     = ('*' | columnNameList)

    tableName      = delimitedList(ident, ".", combine=True).setParseAction(upcaseTokens)
    tableNameList  = Group(delimitedList(tableName)).setName("tables")

    simpleSQL      = selectToken("command") + columnSpec("columns") + fromToken + tableNameList("tables")

    # demo runTests method, including embedded comments in test string
    simpleSQL.runTests("""
        # '*' as column list and dotted table name
        select * from SYS.XYZZY

        # caseless match on "SELECT", and casts back to "select"
        SELECT * from XYZZY, ABC

        # list of column names, and mixed case SELECT keyword
        Select AA,BB,CC from Sys.dual

        # multiple tables
        Select A, B, C from Sys.dual, Table2

        # invalid SELECT keyword - should fail
        Xelect A, B, C from Sys.dual

        # incomplete command - should fail
        Select

        # invalid column name - should fail
        Select ^^^ frox Sys.dual

        """)

    pyparsing_common.number.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    # any int or real number, returned as float
    pyparsing_common.fnumber.runTests("""
        100
        -100
        +100
        3.14159
        6.02e23
        1e-12
        """)

    pyparsing_common.hex_integer.runTests("""
        100
        FF
        """)

    import uuid
    pyparsing_common.uuid.setParseAction(tokenMap(uuid.UUID))
    pyparsing_common.uuid.runTests("""
        12345678-1234-5678-1234-567812345678
        """)
site-packages/pip/_vendor/idna/__pycache__/intranges.cpython-39.pyc000064400000003553150564745010021273 0ustar00a

��Re��@s@dZddlZddlmZmZdd�Zdd�Zdd	�Zd
d�ZdS)a	
Given a list of integers, made up of (hopefully) a small number of long runs
of consecutive integers, compute a representation of the form
((start1, end1), (start2, end2) ...). Then answer the question "was x present
in the original list?" in time O(log(# runs)).
�N)�List�TuplecCs�t|�}g}d}tt|��D]b}|dt|�krJ||||ddkrJq||d|d�}|�t|d|dd��|}qt|�S)aRepresent a list of integers as a sequence of ranges:
    ((start_0, end_0), (start_1, end_1), ...), such that the original
    integers are exactly those x such that start_i <= x < end_i for some i.

    Ranges are encoded as single integers (start << 32 | end), not as tuples.
    ����r)�sorted�range�len�append�
_encode_range�tuple)�list_�sorted_list�ranges�
last_write�i�
current_range�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/intranges.py�intranges_from_lists	rcCs|d>|BS)N� r)�start�endrrrr
!sr
cCs|d?|d@fS)Nrl��r)�rrrr�
_decode_range%srcCszt|d�}t�||�}|dkrNt||d�\}}||krF|krNnndS|t|�krvt||�\}}||krvdSdS)z=Determine if `int_` falls into one of the ranges in `ranges`.rrTF)r
�bisect�bisect_leftrr)�int_r�tuple_�pos�left�right�_rrr�intranges_contain*s
r")	�__doc__r�typingrrrr
rr"rrrr�<module>ssite-packages/pip/_vendor/idna/__pycache__/uts46data.cpython-39.pyc000064400000435442150564745010021126 0ustar00a

��Rey�@s�ddlmZmZmZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�Zd)d*�Zd+d,�Zd-d.�Zd/d0�Zd1d2�Zd3d4�Zd5d6�Zd7d8�Zd9d:�Z d;d<�Z!d=d>�Z"d?d@�Z#dAdB�Z$dCdD�Z%dEdF�Z&dGdH�Z'dIdJ�Z(dKdL�Z)dMdN�Z*dOdP�Z+dQdR�Z,dSdT�Z-dUdV�Z.dWdX�Z/dYdZ�Z0d[d\�Z1d]d^�Z2d_d`�Z3dadb�Z4dcdd�Z5dedf�Z6dgdh�Z7didj�Z8dkdl�Z9dmdn�Z:dodp�Z;dqdr�Z<dsdt�Z=dudv�Z>dwdx�Z?dydz�Z@d{d|�ZAd}d~�ZBdd��ZCd�d��ZDd�d��ZEd�d��ZFd�d��ZGd�d��ZHd�d��ZId�d��ZJd�d��ZKd�d��ZLd�d��ZMd�d��ZNd�d��ZOd�d��ZPd�d��ZQd�d��ZRd�d��ZSd�d��ZTeUe�e�e�e�e	�e
�e�e�e
�e�e�e�e�e�e�e�e�e�e�e�e�e�e�e�e�e�e�e �e!�e"�e#�e$�e%�e&�e'�e(�e)�e*�e+�e,�e-�e.�e/�e0�e1�e2�e3�e4�e5�e6�e7�e8�e9�e:�e;�e<�e=�e>�e?�e@�eA�eB�eC�eD�eE�eF�eG�eH�eI�eJ�eK�eL�eM�eN�eO�eP�eQ�eR�eS�eT��ZVd�S)��)�List�Tuple�Unionz13.0.0cCsgd�S)N)d)r�3)�r)�r)�r)�r)�r)�r)�r)�r)�	r)�
r)�r)�r)�
r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)�r)� r)�!r)�"r)�#r)�$r)�%r)�&r)�'r)�(r)�)r)�*r)�+r)�,r)�-�V)�.r3)�/r)�0r3)�1r3)�2r3)�3r3)�4r3)�5r3)�6r3)�7r3)�8r3)�9r3)�:r)�;r)�<r)�=r)�>r)�?r)�@r)�A�M�a)�BrH�b)�CrH�c)�DrH�d)�ErH�e)�FrH�f)�GrH�g)�HrH�h)�IrH�i)�JrH�j)�KrH�k)�LrH�l)�MrH�m)�NrH�n)�OrH�o)�PrH�p)�QrH�q)�RrH�r)�SrH�s)�TrH�t)�UrH�u)�VrH�v)�WrH�w)�XrH�x)�YrH�y)�ZrH�z)�[r)�\r)�]r)�^r)�_r)�`r)�ar3)�br3)�cr3�r�r�r���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/uts46data.py�_seg_0	sr�cCsgd�S)N)d)�dr3)�er3)�fr3)�gr3)�hr3)�ir3)�jr3)�kr3)�lr3)�mr3)�nr3)�or3)�pr3)�qr3)�rr3)�sr3)�tr3)�ur3)�vr3)�wr3)�xr3)�yr3)�zr3)�{r)�|r)�}r)�~r)�r)��X)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r�)�r� )�r3)�r3)�r3)�r3)�r3)�r3)�r3)�ru ̈)�r3)�rHrI)�r3)�r3)��I)�r3)�r� ̄)�r3)�r3)�rH�2)�rHr)�r� ́)�rH�μ)�r3)�r3)�ru ̧)�rH�1)�rHre)�r3)�rHu1⁄4)�rHu1⁄2)�rHu3⁄4)�r3)�rH�à)�rH�á)��rH�â)��rH�ã)��rH�ä)��rH�å)��rH�æ)��rH�çr�r�r�r�r��_seg_1rsr�cCsgd�S)N)d)��rH�è)��rH�é)��rH�ê)��rH�ë)��rH�ì)��rH�í)��rH�î)��rH�ï)��rH�ð)��rH�ñ)��rH�ò)��rH�ó)��rH�ô)��rH�õ)��rH�ö)��r3)��rH�ø)��rH�ù)��rH�ú)��rH�û)��rH�ü)��rH�ý)��rH�þ)���D�ss)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)��r3)�r3)�r3)�r3)�r3)�r3)�r3)�r3)�r3)�r3)�r3)�r3)�rHuā)ir3)irHuă)ir3)irHuą)ir3)irHuć)ir3)irHuĉ)i	r3)i
rHuċ)ir3)irHuč)i
r3)irHuď)ir3)irHuđ)ir3)irHuē)ir3)irHuĕ)ir3)irHuė)ir3)irHuę)ir3)irHuě)ir3)irHuĝ)ir3)irHuğ)ir3)i rHuġ)i!r3)i"rHuģ)i#r3)i$rHuĥ)i%r3)i&rH�ħ)i'r3)i(rHuĩ)i)r3)i*rHuī)i+r3r�r�r�r�r��_seg_2�srOcCsgd�S)N)d)i,rHuĭ)i-r3)i.rHuį)i/r3)i0rHui̇)i1r3)i2rHZij)i4rHuĵ)i5r3)i6rHuķ)i7r3)i9rHuĺ)i:r3)i;rHuļ)i<r3)i=rHuľ)i>r3)i?rHul·)iArHuł)iBr3)iCrHuń)iDr3)iErHuņ)iFr3)iGrHuň)iHr3)iIrHuʼn)iJrH�ŋ)iKr3)iLrHuō)iMr3)iNrHuŏ)iOr3)iPrHuő)iQr3)iRrH�œ)iSr3)iTrHuŕ)iUr3)iVrHuŗ)iWr3)iXrHuř)iYr3)iZrHuś)i[r3)i\rHuŝ)i]r3)i^rHuş)i_r3)i`rHuš)iar3)ibrHuţ)icr3)idrHuť)ier3)ifrHuŧ)igr3)ihrHuũ)iir3)ijrHuū)ikr3)ilrHuŭ)imr3)inrHuů)ior3)iprHuű)iqr3)irrHuų)isr3)itrHuŵ)iur3)ivrHuŷ)iwr3)ixrH�ÿ)iyrHuź)izr3)i{rHuż)i|r3)i}rHuž)i~r3)irHrm)i�r3)i�rHuɓ)i�rHuƃ)i�r3)i�rHuƅ)i�r3)i�rH�ɔ)i�rHuƈ)i�r3)i�rHuɖ)i�rHuɗ)i�rHuƌ)i�r3)i�rH�ǝ)i�rH�ə)i�rH�ɛ)i�rHuƒ)i�r3)i�rHuɠr�r�r�r�r��_seg_3DsrWcCsgd�S)N)d)i�rH�ɣ)i�r3)i�rH�ɩ)i�rH�ɨ)i�rHuƙ)i�r3)i�rH�ɯ)i�rH�ɲ)i�r3)i�rH�ɵ)i�rHuơ)i�r3)i�rHuƣ)i�r3)i�rHuƥ)i�r3)i�rHuʀ)i�rHuƨ)i�r3)i�rH�ʃ)i�r3)i�rHuƭ)i�r3)i�rHuʈ)i�rHuư)i�r3)i�rH�ʊ)i�rH�ʋ)i�rHuƴ)i�r3)i�rHuƶ)i�r3)i�rH�ʒ)i�rHuƹ)i�r3)i�rHuƽ)i�r3)i�rHudž)i�rHZlj)i�rHZnj)i�rHuǎ)i�r3)i�rHuǐ)i�r3)i�rHuǒ)i�r3)i�rHuǔ)i�r3)i�rHuǖ)i�r3)i�rHuǘ)i�r3)i�rHuǚ)i�r3)i�rHuǜ)i�r3)i�rHuǟ)i�r3)i�rHuǡ)i�r3)i�rHuǣ)i�r3)i�rHuǥ)i�r3)i�rHuǧ)i�r3)i�rHuǩ)i�r3)i�rHuǫ)i�r3)i�rHuǭ)i�r3)i�rHuǯ)i�r3)i�rHZdz)i�rHuǵ)i�r3)i�rHuƕ)i�rHuƿ)i�rHuǹ)i�r3)i�rHuǻ)i�r3)i�rHuǽ)i�r3)i�rHuǿ)i�r3)irHuȁ)ir3)irHuȃ)ir3)irHuȅ)ir3)irHuȇ)ir3)irHuȉ)i	r3)i
rHuȋ)ir3)irHuȍr�r�r�r�r��_seg_4�srbcCsgd�S)N)d)i
r3)irHuȏ)ir3)irHuȑ)ir3)irHuȓ)ir3)irHuȕ)ir3)irHuȗ)ir3)irHuș)ir3)irHuț)ir3)irHuȝ)ir3)irHuȟ)ir3)i rHuƞ)i!r3)i"rH�ȣ)i#r3)i$rHuȥ)i%r3)i&rHuȧ)i'r3)i(rHuȩ)i)r3)i*rHuȫ)i+r3)i,rHuȭ)i-r3)i.rHuȯ)i/r3)i0rHuȱ)i1r3)i2rHuȳ)i3r3)i:rHuⱥ)i;rHuȼ)i<r3)i=rHuƚ)i>rHuⱦ)i?r3)iArHuɂ)iBr3)iCrHuƀ)iDrH�ʉ)iErH�ʌ)iFrHuɇ)iGr3)iHrHuɉ)iIr3)iJrHuɋ)iKr3)iLrHuɍ)iMr3)iNrHuɏ)iOr3)i�rHrW)i�rH�ɦ)i�rHr[)i�rHrk)i�rHuɹ)i�rHuɻ)i�rHuʁ)i�rHru)i�rHry)i�r3)i�ru ̆)i�ru ̇)i�ru ̊)i�ru ̨)i�ru ̃)i�ru ̋)i�r3)i�rHrX)i�rHr_)i�rHrm)i�rHrw)i�rHuʕ)i�r3)i@rHù)iArHú)iBr3)iCrHu̓)iDrHǘ)iErH�ι)iFr3)iOr�)iPr3)iprHuͱ)iqr3)irrHuͳ)isr3)itrHuʹ)iur3)ivrHuͷ)iwr3r�r�r�r�r��_seg_5srhcCsgd�S)N)d)ixr�)izru ι)i{r3)i~r�;)irHuϳ)i�r�)i�rr�)i�r� ̈́)i�rH�ά)i�rH�·)i�rH�έ)i�rH�ή)i�rH�ί)i�r�)i�rH�ό)i�r�)i�rH�ύ)i�rH�ώ)i�r3)i�rH�α)i�rH�β)i�rH�γ)i�rH�δ)i�rH�ε)i�rH�ζ)i�rH�η)i�rH�θ)i�rHrg)i�rH�κ)i�rH�λ)i�rHr�)i�rH�ν)i�rH�ξ)i�rH�ο)i�rH�π)i�rH�ρ)i�r�)i�rH�σ)i�rH�τ)i�rH�υ)i�rH�φ)i�rH�χ)i�rH�ψ)i�rH�ω)i�rHuϊ)i�rH�ϋ)i�r3)i�r+r�)i�r3)i�rHuϗ)i�rHrt)i�rHrz)i�rHr�)i�rHrq)i�rHr�)i�rHr�)i�rHr�)i�r3)i�rHuϙ)i�r3)i�rHuϛ)i�r3)i�rH�ϝ)i�r3)i�rHuϟ)i�r3)i�rHuϡ)i�r3)i�rHuϣ)i�r3)i�rHuϥ)i�r3)i�rHuϧ)i�r3)i�rHuϩ)i�r3)i�rHuϫ)i�r3)i�rHuϭ)i�r3)i�rHuϯ)i�r3)i�rHr{)i�rHr�)i�rHr�)i�r3)i�rHrz)i�rHrw)i�r3)i�rHuϸ)i�r3)i�rHr�)i�rHuϻ)i�r3)i�rHuͻ)i�rHuͼ)i�rHuͽ)irHuѐ)irHuё)irHuђr�r�r�r�r��_seg_6sr�cCsgd�S)N)d)irHuѓ)irHuє)irHuѕ)irHuі)irHuї)irHuј)i	rHuљ)i
rHuњ)irHuћ)irHuќ)i
rHuѝ)irHuў)irHuџ)irHuа)irHuб)irH�в)irHuг)irH�д)irHuе)irHuж)irHuз)irHuи)irHuй)irHuк)irHuл)irHuм)irH�н)irH�о)irHuп)i rHuр)i!rH�с)i"rH�т)i#rHuу)i$rHuф)i%rHuх)i&rHuц)i'rHuч)i(rHuш)i)rHuщ)i*rH�ъ)i+rHuы)i,rH�ь)i-rHuэ)i.rHuю)i/rHuя)i0r3)i`rHuѡ)iar3)ibrH�ѣ)icr3)idrHuѥ)ier3)ifrHuѧ)igr3)ihrHuѩ)iir3)ijrHuѫ)ikr3)ilrHuѭ)imr3)inrHuѯ)ior3)iprHuѱ)iqr3)irrHuѳ)isr3)itrHuѵ)iur3)ivrHuѷ)iwr3)ixrHuѹ)iyr3)izrHuѻ)i{r3)i|rHuѽ)i}r3)i~rHuѿ)ir3)i�rHuҁ)i�r3)i�rHuҋ)i�r3)i�rHuҍ)i�r3)i�rHuҏ)i�r3)i�rHuґ)i�r3)i�rHuғ)i�r3)i�rHuҕ)i�r3)i�rHuҗ)i�r3)i�rHuҙ)i�r3)i�rHuқ)i�r3)i�rHuҝ)i�r3r�r�r�r�r��_seg_7�sr�cCsgd�S)N)d)i�rHuҟ)i�r3)i�rHuҡ)i�r3)i�rHuң)i�r3)i�rHuҥ)i�r3)i�rHuҧ)i�r3)i�rHuҩ)i�r3)i�rHuҫ)i�r3)i�rHuҭ)i�r3)i�rHuү)i�r3)i�rHuұ)i�r3)i�rHuҳ)i�r3)i�rHuҵ)i�r3)i�rHuҷ)i�r3)i�rHuҹ)i�r3)i�rHuһ)i�r3)i�rHuҽ)i�r3)i�rHuҿ)i�r3)i�r�)i�rHuӂ)i�r3)i�rHuӄ)i�r3)i�rHuӆ)i�r3)i�rHuӈ)i�r3)i�rHuӊ)i�r3)i�rHuӌ)i�r3)i�rHuӎ)i�r3)i�rHuӑ)i�r3)i�rHuӓ)i�r3)i�rHuӕ)i�r3)i�rHuӗ)i�r3)i�rHuә)i�r3)i�rHuӛ)i�r3)i�rHuӝ)i�r3)i�rHuӟ)i�r3)i�rHuӡ)i�r3)i�rHuӣ)i�r3)i�rHuӥ)i�r3)i�rHuӧ)i�r3)i�rHuө)i�r3)i�rHuӫ)i�r3)i�rHuӭ)i�r3)i�rHuӯ)i�r3)i�rHuӱ)i�r3)i�rHuӳ)i�r3)i�rHuӵ)i�r3)i�rHuӷ)i�r3)i�rHuӹ)i�r3)i�rHuӻ)i�r3)i�rHuӽ)i�r3)i�rHuӿ)i�r3)irHuԁ)ir3)irHuԃr�r�r�r�r��_seg_8Qsr�cCsgd�S)N)d)ir3)irHuԅ)ir3)irHuԇ)ir3)irHuԉ)i	r3)i
rHuԋ)ir3)irHuԍ)i
r3)irHuԏ)ir3)irHuԑ)ir3)irHuԓ)ir3)irHuԕ)ir3)irHuԗ)ir3)irHuԙ)ir3)irHuԛ)ir3)irHuԝ)ir3)irHuԟ)ir3)i rHuԡ)i!r3)i"rHuԣ)i#r3)i$rHuԥ)i%r3)i&rHuԧ)i'r3)i(rHuԩ)i)r3)i*rHuԫ)i+r3)i,rHuԭ)i-r3)i.rHuԯ)i/r3)i0r�)i1rHuա)i2rHuբ)i3rHuգ)i4rHuդ)i5rHuե)i6rHuզ)i7rHuէ)i8rHuը)i9rHuթ)i:rHuժ)i;rHuի)i<rHuլ)i=rHuխ)i>rHuծ)i?rHuկ)i@rHuհ)iArHuձ)iBrHuղ)iCrHuճ)iDrHuմ)iErHuյ)iFrHuն)iGrHuշ)iHrHuո)iIrHuչ)iJrHuպ)iKrHuջ)iLrHuռ)iMrHuս)iNrHuվ)iOrHuտ)iPrHuր)iQrHuց)iRrHuւ)iSrHuփ)iTrHuք)iUrHuօ)iVrHuֆ)iWr�)iYr3)i�rHuեւ)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)ir3r�r�r�r�r��_seg_9�sr�cCsgd�S)N)d)iurHuاٴ)ivrHuوٴ)iwrH�ۇٴ)ixrHuيٴ)iyr3)i�r�)i�r3)ir�)ir3)iKr�)iMr3)i�r�)i�r3)i�r�)i�r3)i.r�)i0r3)i?r�)i@r3)i\r�)i^r3)i_r�)i`r3)ikr�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)iX	rHuक़)iY	rHuख़)iZ	rHuग़)i[	rHuज़)i\	rHuड़)i]	rHuढ़)i^	rHuफ़)i_	rHuय़)i`	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	rHuড়)i�	rHuঢ়)i�	r�)i�	rHuয়)i�	r3)i�	r�)i�	r3)i�	r�)i
r3)i
r�)i
r3)i
r�)i
r3)i
r�)i
r3)i)
r�)i*
r3)i1
r�)i2
r3)i3
rHuਲ਼)i4
r�)i5
r3)i6
rHuਸ਼)i7
r�)i8
r3)i:
r�)i<
r3)i=
r�)i>
r3)iC
r�)iG
r3)iI
r�)iK
r3)iN
r�)iQ
r3)iR
r�)iY
rHuਖ਼)iZ
rHuਗ਼)i[
rHuਜ਼r�r�r�r�r��_seg_10#sr�cCsgd�S)N)d)i\
r3)i]
r�)i^
rHuਫ਼)i_
r�)if
r3)iw
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)ir�)ir3)ir�)ir3)i
r�)ir3)ir�)ir3)i)r�)i*r3)i1r�)i2r3)i4r�)i5r3)i:r�)i<r3)iEr�)iGr3)iIr�)iKr3)iNr�)iUr3)iXr�)i\rHuଡ଼)i]rHuଢ଼)i^r�)i_r3)idr�)ifr3)ixr�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i
r�)ir3)ir�)ir3r�r�r�r�r��_seg_11�sr�cCsgd�S)N)d)i)r�)i*r3)i:r�)i=r3)iEr�)iFr3)iIr�)iJr3)iNr�)iUr3)iWr�)iXr3)i[r�)i`r3)idr�)ifr3)ipr�)iwr3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i
r3)i

r�)i
r3)i
r�)i
r3)iE
r�)iF
r3)iI
r�)iJ
r3)iP
r�)iT
r3)id
r�)if
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)ir3)i3rHuํา)i4r3)i;r�)i?r3)i\r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�rHuໍາ)i�r3r�r�r�r�r��_seg_12�sr�cCsgd�S)N)d)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�rHuຫນ)i�rHuຫມ)i�r3)i�r�)ir3)irHu་)i
r3)iCrHuགྷ)iDr3)iHr�)iIr3)iMrHuཌྷ)iNr3)iRrHuདྷ)iSr3)iWrHuབྷ)iXr3)i\rHuཛྷ)i]r3)iirHuཀྵ)ijr3)imr�)iqr3)isrHuཱི)itr3)iurHuཱུ)ivrHuྲྀ)iwrHu	ྲཱྀ)ixrHuླྀ)iyrHu	ླཱྀ)izr3)i�rHuཱྀ)i�r3)i�rHuྒྷ)i�r3)i�r�)i�r3)i�rHuྜྷ)i�r3)i�rHuྡྷ)i�r3)i�rHuྦྷ)i�r3)i�rHuྫྷ)i�r3)i�rHuྐྵ)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i�r�)i�rHuⴧ)i�r�)i�rHuⴭ)i�r�)i�r3)i�rH�ნ)i�r3)i_r�)iar3)iIr�)iJr3)iNr�)iPr3)iWr�)iXr3)iYr�)iZr3)i^r�)i`r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)ir�)ir3r�r�r�r�r��_seg_13^sr�cCsgd�S)N)d)ir�)ir3)i[r�)i]r3)i}r�)i�r3)i�r�)i�r3)i�r�)i�rHuᏰ)i�rHuᏱ)i�rHuᏲ)i�rHuᏳ)i�rHuᏴ)i�rHuᏵ)i�r�)ir3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i
r�)ir3)ir�)i r3)i7r�)i@r3)iTr�)i`r3)imr�)inr3)iqr�)irr3)itr�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)ir3)ir�)ir�)ir3)ir�)i r3)iyr�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)i r3)i,r�)i0r3)i<r�)i@r3)iAr�)iDr3)inr�)ipr3)iur�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)ir�)ir3)i_r�)i`r3)i}r�)ir3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)iLr�)iPr3)i}r�)i�r3)i�r�)i�r3)i8r�)i;r3)iJr�)iMr3r�r�r�r�r��_seg_14�sr�cCsgd�S)N)d)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�ꙋ)i�r�)i�rHuა)i�rHuბ)i�rHuგ)i�rHuდ)i�rHuე)i�rHuვ)i�rHuზ)i�rHuთ)i�rHuი)i�rHuკ)i�rHuლ)i�rHuმ)i�rHr�)i�rHuო)i�rHuპ)i�rHuჟ)i�rHuრ)i�rHuს)i�rHuტ)i�rHuუ)i�rHuფ)i�rHuქ)i�rHuღ)i�rHuყ)i�rHuშ)i�rHuჩ)i�rHuც)i�rHuძ)i�rHuწ)i�rHuჭ)i�rHuხ)i�rHuჯ)i�rHuჰ)i�rHuჱ)i�rHuჲ)i�rHuჳ)i�rHuჴ)i�rHuჵ)i�rHuჶ)i�rHuჷ)i�rHuჸ)i�rHuჹ)i�rHuჺ)i�r�)i�rHuჽ)i�rHuჾ)i�rHuჿ)i�r3)i�r�)i�r3)i�r�)ir3)i,rHrI)i-rHr�)i.rHrK)i/r3)i0rHrO)i1rHrQ)i2rHrT)i3rHrU)i4rHrW)i5rHrY)i6rHr[)i7rHr])i8rHr_)i9rHra)i:rHrc)i;r3)i<rHre)i=rHrc)i>rHrg)i?rHrk)i@rHro)iArHrq)iBrHru)iCrHrI)iDrH�ɐ)iErH�ɑ)iFrHuᴂ)iGrHrK)iHrHrO)iIrHrQ)iJrHrU)iKrHrV)iLrH�ɜ)iMrHrU)iNr3)iOrHr])iPrHra)iQrHrP)iRrHrer�r�r�r�r��_seg_150sr�cCsgd�S)N)d)iSrHrS)iTrHuᴖ)iUrHuᴗ)iVrHrg)iWrHro)iXrHrq)iYrHuᴝ)iZrHr[)i[rHrs)i\rHuᴥ)i]rHrt)i^rHru)i_rHrv)i`rHr�)iarHr�)ibrHrY)icrHrk)idrHrq)ierHrs)ifrHrt)igrHru)ihrHr�)iirHr�)ijrHr�)ikr3)ixrHr�)iyr3)i�rH�ɒ)i�rHrM)i�rHuɕ)i�rHr)i�rHr�)i�rHrS)i�rHuɟ)i�rH�ɡ)i�rH�ɥ)i�rHrZ)i�rHrY)i�rH�ɪ)i�rHuᵻ)i�rH�ʝ)i�rHuɭ)i�rHuᶅ)i�rHuʟ)i�rH�ɱ)i�rHuɰ)i�rHr\)i�rHuɳ)i�rHuɴ)i�rHr])i�rHuɸ)i�rH�ʂ)i�rHr^)i�rHuƫ)i�rHrd)i�rHr_)i�rHuᴜ)i�rHr`)i�rHre)i�rHr{)i�rHuʐ)i�rHuʑ)i�rHra)i�rHrz)i�r3)i�r�)i�r3)irHuḁ)ir3)irHuḃ)ir3)irHuḅ)ir3)irHuḇ)ir3)irHuḉ)i	r3)i
rHuḋ)ir3)irHuḍ)i
r3)irHuḏ)ir3)irHuḑ)ir3)irHuḓ)ir3)irHuḕ)ir3)irHuḗ)ir3)irHuḙ)ir3)irHuḛ)ir3)irHuḝ)ir3)irHuḟ)ir3)i rHuḡr�r�r�r�r��_seg_16�sr�cCsgd�S)N)d)i!r3)i"rHuḣ)i#r3)i$rHuḥ)i%r3)i&rHuḧ)i'r3)i(rHuḩ)i)r3)i*rHuḫ)i+r3)i,rHuḭ)i-r3)i.rHuḯ)i/r3)i0rHuḱ)i1r3)i2rHuḳ)i3r3)i4rHuḵ)i5r3)i6rHuḷ)i7r3)i8rHuḹ)i9r3)i:rHuḻ)i;r3)i<rHuḽ)i=r3)i>rHuḿ)i?r3)i@rHuṁ)iAr3)iBrHuṃ)iCr3)iDrHuṅ)iEr3)iFrHuṇ)iGr3)iHrHuṉ)iIr3)iJrHuṋ)iKr3)iLrHuṍ)iMr3)iNrHuṏ)iOr3)iPrHuṑ)iQr3)iRrHuṓ)iSr3)iTrHuṕ)iUr3)iVrHuṗ)iWr3)iXrHuṙ)iYr3)iZrHuṛ)i[r3)i\rHuṝ)i]r3)i^rHuṟ)i_r3)i`rH�ṡ)iar3)ibrHuṣ)icr3)idrHuṥ)ier3)ifrHuṧ)igr3)ihrHuṩ)iir3)ijrHuṫ)ikr3)ilrHuṭ)imr3)inrHuṯ)ior3)iprHuṱ)iqr3)irrHuṳ)isr3)itrHuṵ)iur3)ivrHuṷ)iwr3)ixrHuṹ)iyr3)izrHuṻ)i{r3)i|rHuṽ)i}r3)i~rHuṿ)ir3)i�rHuẁ)i�r3)i�rHuẃ)i�r3)i�rHuẅr�r�r�r�r��_seg_17sr�cCsgd�S)N)d)i�r3)i�rHuẇ)i�r3)i�rHuẉ)i�r3)i�rHuẋ)i�r3)i�rHuẍ)i�r3)i�rHuẏ)i�r3)i�rHuẑ)i�r3)i�rHuẓ)i�r3)i�rHuẕ)i�r3)i�rHuaʾ)i�rHr�)i�r3)i�rHr,)i�r3)i�rHuạ)i�r3)i�rHuả)i�r3)i�rHuấ)i�r3)i�rHuầ)i�r3)i�rHuẩ)i�r3)i�rHuẫ)i�r3)i�rHuậ)i�r3)i�rHuắ)i�r3)i�rHuằ)i�r3)i�rHuẳ)i�r3)i�rHuẵ)i�r3)i�rHuặ)i�r3)i�rHuẹ)i�r3)i�rHuẻ)i�r3)i�rHuẽ)i�r3)i�rHuế)i�r3)i�rHuề)i�r3)i�rHuể)i�r3)i�rHuễ)i�r3)i�rHuệ)i�r3)i�rHuỉ)i�r3)i�rHuị)i�r3)i�rHuọ)i�r3)i�rHuỏ)i�r3)i�rHuố)i�r3)i�rHuồ)i�r3)i�rHuổ)i�r3)i�rHuỗ)i�r3)i�rHuộ)i�r3)i�rHuớ)i�r3)i�rHuờ)i�r3)i�rHuở)i�r3)i�rHuỡ)i�r3)i�rHuợ)i�r3)i�rHuụ)i�r3)i�rHuủ)i�r3)i�rHuứ)i�r3)i�rHuừ)i�r3)i�rHuử)i�r3r�r�r�r�r��_seg_18ksr�cCsgd�S)N)d)i�rHuữ)i�r3)i�rHuự)i�r3)i�rHuỳ)i�r3)i�rHuỵ)i�r3)i�rHuỷ)i�r3)i�rHuỹ)i�r3)i�rHuỻ)i�r3)i�rHuỽ)i�r3)i�rHuỿ)i�r3)irHuἀ)i	rHuἁ)i
rHuἂ)irHuἃ)irHuἄ)i
rHuἅ)irHuἆ)irHuἇ)ir3)ir�)irHuἐ)irHuἑ)irHuἒ)irHuἓ)irHuἔ)irHuἕ)ir�)i r3)i(rHuἠ)i)rHuἡ)i*rHuἢ)i+rHuἣ)i,rHuἤ)i-rHuἥ)i.rHuἦ)i/rHuἧ)i0r3)i8rHuἰ)i9rHuἱ)i:rHuἲ)i;rHuἳ)i<rHuἴ)i=rHuἵ)i>rHuἶ)i?rHuἷ)i@r3)iFr�)iHrHuὀ)iIrHuὁ)iJrHuὂ)iKrHuὃ)iLrHuὄ)iMrHuὅ)iNr�)iPr3)iXr�)iYrHuὑ)iZr�)i[rHuὓ)i\r�)i]rHuὕ)i^r�)i_rHuὗ)i`r3)ihrHuὠ)iirHuὡ)ijrHuὢ)ikrHuὣ)ilrHuὤ)imrHuὥ)inrHuὦ)iorHuὧ)ipr3)iqrHrk)irr3)isrHrm)itr3)iurHrn)ivr3)iwrHro)ixr3)iyrHrp)izr3)i{rHrq)i|r3)i}rHrr)i~r�)i�rH�ἀι)i�rH�ἁι)i�rH�ἂι)i�rH�ἃι)i�rH�ἄιr�r�r�r�r��_seg_19�sr�cCsgd�S)N)d)i�rH�ἅι)i�rH�ἆι)i�rH�ἇι)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�ἠι)i�rH�ἡι)i�rH�ἢι)i�rH�ἣι)i�rH�ἤι)i�rH�ἥι)i�rH�ἦι)i�rH�ἧι)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�ὠι)i�rH�ὡι)i�rH�ὢι)i�rH�ὣι)i�rH�ὤι)i�rH�ὥι)i�rH�ὦι)i�rH�ὧι)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�r3)i�rHuὰι)i�rH�αι)i�rHuάι)i�r�)i�r3)i�rHuᾶι)i�rHuᾰ)i�rHuᾱ)i�rHuὰ)i�rHrk)i�rHr�)i�r� ̓)i�rHrg)i�rr�)i�ru ͂)i�ru ̈͂)i�rHuὴι)i�rH�ηι)i�rHuήι)i�r�)i�r3)i�rHuῆι)i�rHuὲ)i�rHrm)i�rHuὴ)i�rHrn)i�rHr�)i�ru ̓̀)i�ru ̓́)i�ru ̓͂)i�r3)i�rHuΐ)i�r�)i�r3)i�rHuῐ)i�rHuῑ)i�rHuὶ)i�rHro)i�r�)i�ru ̔̀)i�ru ̔́)i�ru ̔͂)i�r3)i�rHuΰ)i�r3)i�rHuῠ)i�rHuῡ)i�rHuὺ)i�rHrq)i�rHuῥ)i�ru ̈̀)i�rrj)i�r�`)i�r�)i�rHuὼι)i�rH�ωιr�r�r�r�r��_seg_20=sr�cCsgd�S)N)d)i�rHuώι)i�r�)i�r3)i�rHuῶι)i�rHuὸ)i�rHrp)i�rHuὼ)i�rHrr)i�rHr�)i�rr�)i�ru ̔)i�r�)i rr�)i r�)i r+�)i r�)i r3)i rHu‐)i r3)i ru ̳)i r3)i$ r�)i' r3)i( r�)i/ rr�)i0 r3)i3 rHu′′)i4 rHu	′′′)i5 r3)i6 rHu‵‵)i7 rHu	‵‵‵)i8 r3)i< rz!!)i= r3)i> r� ̅)i? r3)iG rz??)iH rz?!)iI rz!?)iJ r3)iW rHu′′′′)iX r3)i_ rr�)i` r�)ia r�)id r�)ie r�)ip rH�0)iq rHrY)ir r�)it rH�4)iu rH�5)iv rH�6)iw rH�7)ix rH�8)iy rH�9)iz r�+)i{ rH�−)i| r�=)i} r�()i~ r�))i rHrc)i� rHr�)i� rHr�)i� rHr�)i� rHr)i� rHr�)i� rHr�)i� rHr�)i� rHr�)i� rHr�)i� rHr�)i� rr�)i� rHr�)i� rr�)i� rr�)i� rr�)i� r�)i� rHrI)i� rHrQ)i� rHre)i� rHrw)i� rHrU)i� rHrW)i� rHr])i� rHr_)i� rHra)i� rHrc)i� rHrg)i� rHrm)i� rHro)i� r�)i� r3)i� rH�rs)i� r3)i� r�)i� r3)i� r�)i!rza/c)i!rza/sr�r�r�r�r��_seg_21�sr�cCsgd�S)N)d)i!rHrM)i!rHu°c)i!r3)i!rzc/o)i!rzc/u)i!rHrV)i!r3)i	!rHu°f)i
!rHrU)i!rHrW)i!rHrN)i!rHrY)i!rHr_)i!r3)i!rHrc)i!rH�no)i!r3)i!rHrg)i!rHri)i!rHrk)i!r3)i !rH�sm)i!!rH�tel)i"!rH�tm)i#!r3)i$!rHr{)i%!r3)i&!rHr�)i'!r3)i(!rHr{)i)!r3)i*!rHr])i+!rHr�)i,!rHrK)i-!rHrM)i.!r3)i/!rHrQ)i1!rHrS)i2!r�)i3!rHra)i4!rHre)i5!rH�א)i6!rHuב)i7!rHuג)i8!rH�ד)i9!rHrY)i:!r3)i;!rHZfax)i<!rHr�)i=!rHru)i?!rHr�)i@!rHu∑)iA!r3)iE!rHrO)iG!rHrQ)iH!rHrY)iI!rHr[)iJ!r3)iP!rHu1⁄7)iQ!rHu1⁄9)iR!rHu1⁄10)iS!rHu1⁄3)iT!rHu2⁄3)iU!rHu1⁄5)iV!rHu2⁄5)iW!rHu3⁄5)iX!rHu4⁄5)iY!rHu1⁄6)iZ!rHu5⁄6)i[!rHu1⁄8)i\!rHu3⁄8)i]!rHu5⁄8)i^!rHu7⁄8)i_!rHu1⁄)i`!rHrY)ia!rH�ii)ib!rH�iii)ic!rH�iv)id!rHrs)ie!rH�vi)if!rH�vii)ig!rH�viii)ih!rH�ix)ii!rHrw)ij!rH�xi)ik!rH�xii)il!rHr_)im!rHrM)in!rHrO)io!rHra)ip!rHrY)iq!rHr�)ir!rHr�)is!rHr�)it!rHrs)iu!rHr�)iv!rHr�)iw!rHr�)ix!rHr�)iy!rHrwr�r�r�r�r��_seg_22	sr�cCsgd�S)N)d)iz!rHr�)i{!rHr�)i|!rHr_)i}!rHrM)i~!rHrO)i!rHra)i�!r3)i�!r�)i�!r3)i�!rHu0⁄3)i�!r3)i�!r�)i�!r3)i,"rHu∫∫)i-"rHu	∫∫∫)i."r3)i/"rHu∮∮)i0"rHu	∮∮∮)i1"r3)i`"r)ia"r3)in"r)ip"r3)i)#rH�〈)i*#rH�〉)i+#r3)i'$r�)i@$r3)iK$r�)i`$rHr�)ia$rHr�)ib$rHr)ic$rHr�)id$rHr�)ie$rHr�)if$rHr�)ig$rHr�)ih$rHr�)ii$rH�10)ij$rHZ11)ik$rHZ12)il$rHZ13)im$rHZ14)in$rHZ15)io$rHZ16)ip$rHZ17)iq$rHZ18)ir$rHZ19)is$rHZ20)it$rz(1))iu$rz(2))iv$rz(3))iw$rz(4))ix$rz(5))iy$rz(6))iz$rz(7))i{$rz(8))i|$rz(9))i}$rz(10))i~$rz(11))i$rz(12))i�$rz(13))i�$rz(14))i�$rz(15))i�$rz(16))i�$rz(17))i�$rz(18))i�$rz(19))i�$rz(20))i�$r�)i�$r�(a))i�$r�(b))i�$r�(c))i�$r�(d))i�$r�(e))i�$r�(f))i�$r�(g))i�$r�(h))i�$r�(i))i�$r�(j))i�$r�(k))i�$r�(l))i�$r�(m))i�$r�(n))i�$r�(o))i�$r�(p))i�$r�(q))i�$r�(r))i�$r�(s))i�$r�(t))i�$r�(u))i�$r�(v))i�$r�(w))i�$r�(x))i�$r�(y))i�$r�(z))i�$rHrI)i�$rHrK)i�$rHrM)i�$rHrOr�r�r�r�r��_seg_23x	srcCsgd�S)N)d)i�$rHrQ)i�$rHrS)i�$rHrU)i�$rHrW)i�$rHrY)i�$rHr[)i�$rHr])i�$rHr_)i�$rHra)i�$rHrc)i�$rHre)i�$rHrg)i�$rHri)i�$rHrk)i�$rHrm)i�$rHro)i�$rHrq)i�$rHrs)i�$rHru)i�$rHrw)i�$rHry)i�$rHr{)i�$rHrI)i�$rHrK)i�$rHrM)i�$rHrO)i�$rHrQ)i�$rHrS)i�$rHrU)i�$rHrW)i�$rHrY)i�$rHr[)i�$rHr])i�$rHr_)i�$rHra)i�$rHrc)i�$rHre)i�$rHrg)i�$rHri)i�$rHrk)i�$rHrm)i�$rHro)i�$rHrq)i�$rHrs)i�$rHru)i�$rHrw)i�$rHry)i�$rHr{)i�$rHr�)i�$r3)i*rHu∫∫∫∫)i
*r3)it*rz::=)iu*rz==)iv*rz===)iw*r3)i�*rHu⫝̸)i�*r3)it+r�)iv+r3)i�+r�)i�+r3)i,rHuⰰ)i,rHuⰱ)i,rHuⰲ)i,rHuⰳ)i,rHuⰴ)i,rHuⰵ)i,rHuⰶ)i,rHuⰷ)i,rHuⰸ)i	,rHuⰹ)i
,rHuⰺ)i,rHuⰻ)i,rHuⰼ)i
,rHuⰽ)i,rHuⰾ)i,rHuⰿ)i,rHuⱀ)i,rHuⱁ)i,rHuⱂ)i,rHuⱃ)i,rHuⱄ)i,rHuⱅ)i,rHuⱆ)i,rHuⱇ)i,rHuⱈ)i,rHuⱉ)i,rHuⱊ)i,rHuⱋ)i,rHuⱌ)i,rHuⱍ)i,rHuⱎ)i,rHuⱏ)i ,rHuⱐ)i!,rHuⱑ)i",rHuⱒ)i#,rHuⱓ)i$,rHuⱔ)i%,rHuⱕr�r�r�r�r��_seg_24�	srcCsgd�S)N)d)i&,rHuⱖ)i',rHuⱗ)i(,rHuⱘ)i),rHuⱙ)i*,rHuⱚ)i+,rHuⱛ)i,,rHuⱜ)i-,rHuⱝ)i.,rHuⱞ)i/,r�)i0,r3)i_,r�)i`,rHuⱡ)ia,r3)ib,rH�ɫ)ic,rHuᵽ)id,rHuɽ)ie,r3)ig,rHuⱨ)ih,r3)ii,rHuⱪ)ij,r3)ik,rHuⱬ)il,r3)im,rHr�)in,rHr�)io,rHr�)ip,rHr�)iq,r3)ir,rHuⱳ)is,r3)iu,rHuⱶ)iv,r3)i|,rHr[)i},rHrs)i~,rHuȿ)i,rHuɀ)i�,rHuⲁ)i�,r3)i�,rHuⲃ)i�,r3)i�,rHuⲅ)i�,r3)i�,rHuⲇ)i�,r3)i�,rHuⲉ)i�,r3)i�,rHuⲋ)i�,r3)i�,rHuⲍ)i�,r3)i�,rHuⲏ)i�,r3)i�,rHuⲑ)i�,r3)i�,rHuⲓ)i�,r3)i�,rHuⲕ)i�,r3)i�,rHuⲗ)i�,r3)i�,rHuⲙ)i�,r3)i�,rHuⲛ)i�,r3)i�,rHuⲝ)i�,r3)i�,rHuⲟ)i�,r3)i�,rHuⲡ)i�,r3)i�,rHuⲣ)i�,r3)i�,rHuⲥ)i�,r3)i�,rHuⲧ)i�,r3)i�,rHuⲩ)i�,r3)i�,rHuⲫ)i�,r3)i�,rHuⲭ)i�,r3)i�,rHuⲯ)i�,r3)i�,rHuⲱ)i�,r3)i�,rHuⲳ)i�,r3)i�,rHuⲵ)i�,r3)i�,rHuⲷ)i�,r3)i�,rHuⲹ)i�,r3)i�,rHuⲻ)i�,r3)i�,rHuⲽ)i�,r3)i�,rHuⲿr�r�r�r�r��_seg_25J
srcCsgd�S)N)d)i�,r3)i�,rHuⳁ)i�,r3)i�,rHuⳃ)i�,r3)i�,rHuⳅ)i�,r3)i�,rHuⳇ)i�,r3)i�,rHuⳉ)i�,r3)i�,rHuⳋ)i�,r3)i�,rHuⳍ)i�,r3)i�,rHuⳏ)i�,r3)i�,rHuⳑ)i�,r3)i�,rHuⳓ)i�,r3)i�,rHuⳕ)i�,r3)i�,rHuⳗ)i�,r3)i�,rHuⳙ)i�,r3)i�,rHuⳛ)i�,r3)i�,rHuⳝ)i�,r3)i�,rHuⳟ)i�,r3)i�,rHuⳡ)i�,r3)i�,rHuⳣ)i�,r3)i�,rHuⳬ)i�,r3)i�,rHuⳮ)i�,r3)i�,rHuⳳ)i�,r3)i�,r�)i�,r3)i&-r�)i'-r3)i(-r�)i--r3)i.-r�)i0-r3)ih-r�)io-rHuⵡ)ip-r3)iq-r�)i-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)i�-r�)i�-r3)iS.r�)i�.r3)i�.r�)i�.r3)i�.rHu母)i�.r3)i�.rHu龟)i�.r�)i/rH�一)i/rHu丨)i/rHu丶)i/rHu丿)i/rH�乙)i/rHu亅)i/rH�二)i/rHu亠)i/rH�人)i	/rHu儿)i
/rHu入)i/rH�八)i/rHu冂)i
/rHu冖)i/rHu冫)i/rHu几)i/rH�凵)i/rHu刀r�r�r�r�r��_seg_26�
srcCsgd�S)N)d)i/rH�力)i/rHu勹)i/rHu匕)i/rHu匚)i/rHu匸)i/rH�十)i/rHu卜)i/rHu卩)i/rHu厂)i/rHu厶)i/rHu又)i/rHu口)i/rHu囗)i/rH�土)i /rHu士)i!/rHu夂)i"/rHu夊)i#/rHu夕)i$/rHu大)i%/rH�女)i&/rHu子)i'/rHu宀)i(/rHu寸)i)/rHu小)i*/rH�尢)i+/rHu尸)i,/rH�屮)i-/rHu山)i./rHu巛)i//rHu工)i0/rHu己)i1/rHu巾)i2/rHu干)i3/rHu幺)i4/rHu广)i5/rHu廴)i6/rH�廾)i7/rHu弋)i8/rHu弓)i9/rHu彐)i:/rHu彡)i;/rHu彳)i</rHu心)i=/rHu戈)i>/rHu戶)i?/rH�手)i@/rHu支)iA/rHu攴)iB/rH�文)iC/rHu斗)iD/rHu斤)iE/rHu方)iF/rHu无)iG/rH�日)iH/rHu曰)iI/rH�月)iJ/rH�木)iK/rHu欠)iL/rHu止)iM/rH�歹)iN/rHu殳)iO/rHu毋)iP/rHu比)iQ/rHu毛)iR/rHu氏)iS/rHu气)iT/rH�水)iU/rH�火)iV/rHu爪)iW/rHu父)iX/rHu爻)iY/rHu爿)iZ/rHu片)i[/rHu牙)i\/rHu牛)i]/rHu犬)i^/rHu玄)i_/rHu玉)i`/rHu瓜)ia/rHu瓦)ib/rHu甘)ic/rH�生)id/rHu用)ie/rHu田)if/rHu疋)ig/rHu疒)ih/rHu癶)ii/rHu白)ij/rHu皮)ik/rHu皿)il/rHu目)im/rHu矛)in/rHu矢)io/rHu石)ip/rHu示)iq/rHu禸)ir/rHu禾)is/rHu穴)it/rH�立)iu/rHu竹r�r�r�r�r��_seg_27sr'cCsgd�S)N)d)iv/rHu米)iw/rHu糸)ix/rHu缶)iy/rHu网)iz/rHu羊)i{/rH�羽)i|/rH�老)i}/rHu而)i~/rHu耒)i/rHu耳)i�/rHu聿)i�/rHu肉)i�/rHu臣)i�/rHu自)i�/rHu至)i�/rHu臼)i�/rHu舌)i�/rHu舛)i�/rHu舟)i�/rHu艮)i�/rHu色)i�/rHu艸)i�/rHu虍)i�/rHu虫)i�/rHu血)i�/rH�行)i�/rH�衣)i�/rHu襾)i�/rH�見)i�/rHu角)i�/rHu言)i�/rHu谷)i�/rHu豆)i�/rH�豕)i�/rHu豸)i�/rHu貝)i�/rHu赤)i�/rH�走)i�/rHu足)i�/rHu身)i�/rH�車)i�/rHu辛)i�/rH�辰)i�/rHu辵)i�/rHu邑)i�/rHu酉)i�/rHu釆)i�/rH�里)i�/rH�金)i�/rHu長)i�/rHu門)i�/rHu阜)i�/rHu隶)i�/rHu隹)i�/rHu雨)i�/rHu靑)i�/rHu非)i�/rHu面)i�/rHu革)i�/rHu韋)i�/rHu韭)i�/rHu音)i�/rHu頁)i�/rHu風)i�/rHu飛)i�/rHu食)i�/rHu首)i�/rHu香)i�/rHu馬)i�/rHu骨)i�/rHu高)i�/rHu髟)i�/rHu鬥)i�/rHu鬯)i�/rHu鬲)i�/rHu鬼)i�/rHu魚)i�/rHu鳥)i�/rHu鹵)i�/rH�鹿)i�/rHu麥)i�/rH�麻)i�/rHu黃)i�/rHu黍)i�/rHu黑)i�/rH�黹)i�/rHu黽)i�/rHu鼎)i�/rHu鼓)i�/rHu鼠)i�/rH�鼻)i�/rHu齊)i�/rHu齒)i�/rH�龍)i�/rH�龜)i�/rHu龠)i�/r�)i0rr�)i0r3)i0rH�.r�r�r�r�r��_seg_28�sr:cCsgd�S)N)d)i0r3)i60rHu〒)i70r3)i80rHr)i90rHu卄)i:0rHu卅)i;0r3)i@0r�)iA0r3)i�0r�)i�0r3)i�0ru ゙)i�0ru ゚)i�0r3)i�0rHuより)i�0r3)i�0rHuコト)i1r�)i1r3)i01r�)i11rH�ᄀ)i21rH�ᄁ)i31rH�ᆪ)i41rH�ᄂ)i51rH�ᆬ)i61rH�ᆭ)i71rH�ᄃ)i81rH�ᄄ)i91rH�ᄅ)i:1rH�ᆰ)i;1rH�ᆱ)i<1rH�ᆲ)i=1rH�ᆳ)i>1rH�ᆴ)i?1rH�ᆵ)i@1rH�ᄚ)iA1rH�ᄆ)iB1rH�ᄇ)iC1rH�ᄈ)iD1rH�ᄡ)iE1rH�ᄉ)iF1rH�ᄊ)iG1rH�ᄋ)iH1rH�ᄌ)iI1rH�ᄍ)iJ1rH�ᄎ)iK1rH�ᄏ)iL1rH�ᄐ)iM1rH�ᄑ)iN1rH�ᄒ)iO1rH�ᅡ)iP1rH�ᅢ)iQ1rH�ᅣ)iR1rH�ᅤ)iS1rH�ᅥ)iT1rH�ᅦ)iU1rH�ᅧ)iV1rH�ᅨ)iW1rH�ᅩ)iX1rH�ᅪ)iY1rH�ᅫ)iZ1rH�ᅬ)i[1rH�ᅭ)i\1rH�ᅮ)i]1rH�ᅯ)i^1rH�ᅰ)i_1rH�ᅱ)i`1rH�ᅲ)ia1rH�ᅳ)ib1rH�ᅴ)ic1rH�ᅵ)id1r�)ie1rHuᄔ)if1rHuᄕ)ig1rHuᇇ)ih1rHuᇈ)ii1rHuᇌ)ij1rHuᇎ)ik1rHuᇓ)il1rHuᇗ)im1rHuᇙ)in1rHuᄜ)io1rHuᇝ)ip1rHuᇟ)iq1rHuᄝ)ir1rHuᄞ)is1rHuᄠ)it1rHuᄢ)iu1rHuᄣ)iv1rHuᄧ)iw1rHuᄩ)ix1rHuᄫ)iy1rHuᄬ)iz1rHuᄭ)i{1rHuᄮ)i|1rHuᄯ)i}1rHuᄲ)i~1rHuᄶ)i1rHuᅀ)i�1rHuᅇr�r�r�r�r��_seg_29�srncCsgd�S)N)d)i�1rHuᅌ)i�1rHuᇱ)i�1rHuᇲ)i�1rHuᅗ)i�1rHuᅘ)i�1rHuᅙ)i�1rHuᆄ)i�1rHuᆅ)i�1rHuᆈ)i�1rHuᆑ)i�1rHuᆒ)i�1rHuᆔ)i�1rHuᆞ)i�1rHuᆡ)i�1r�)i�1r3)i�1rHr)i�1rHr)i�1rH�三)i�1rH�四)i�1rH�上)i�1rH�中)i�1rH�下)i�1rHu甲)i�1rHr)i�1rHu丙)i�1rHu丁)i�1rH�天)i�1rHu地)i�1rHr)i�1r3)i�1r�)i�1r3)i2ru(ᄀ))i2ru(ᄂ))i2ru(ᄃ))i2ru(ᄅ))i2ru(ᄆ))i2ru(ᄇ))i2ru(ᄉ))i2ru(ᄋ))i2ru(ᄌ))i	2ru(ᄎ))i
2ru(ᄏ))i2ru(ᄐ))i2ru(ᄑ))i
2ru(ᄒ))i2ru(가))i2ru(나))i2ru(다))i2ru(라))i2ru(마))i2ru(바))i2ru(사))i2ru(아))i2ru(자))i2ru(차))i2ru(카))i2ru(타))i2ru(파))i2ru(하))i2ru(주))i2ru(오전))i2ru(오후))i2r�)i 2ru(一))i!2ru(二))i"2ru(三))i#2ru(四))i$2ru(五))i%2ru(六))i&2ru(七))i'2ru(八))i(2ru(九))i)2ru(十))i*2ru(月))i+2ru(火))i,2ru(水))i-2ru(木))i.2ru(金))i/2ru(土))i02ru(日))i12ru(株))i22ru(有))i32ru(社))i42ru(名))i52ru(特))i62ru(財))i72ru(祝))i82ru(労))i92ru(代))i:2ru(呼))i;2ru(学))i<2ru(監))i=2ru(企))i>2ru(資))i?2ru(協))i@2ru(祭))iA2ru(休))iB2ru(自)r�r�r�r�r��_seg_30WsrucCsgd�S)N)d)iC2ru(至))iD2rHu問)iE2rHu幼)iF2rHr)iG2rHu箏)iH2r3)iP2rHZpte)iQ2rHZ21)iR2rHZ22)iS2rHZ23)iT2rHZ24)iU2rHZ25)iV2rHZ26)iW2rH�27)iX2rHZ28)iY2rHZ29)iZ2rHZ30)i[2rHZ31)i\2rHZ32)i]2rHZ33)i^2rHZ34)i_2rHZ35)i`2rHr;)ia2rHr>)ib2rHrA)ic2rHrC)id2rHrK)ie2rHrL)if2rHrO)ig2rHrQ)ih2rHrR)ii2rHrT)ij2rHrU)ik2rHrV)il2rHrW)im2rHrX)in2rHu가)io2rHu나)ip2rHu다)iq2rHu라)ir2rHu마)is2rHu바)it2rHu사)iu2rHu아)iv2rHu자)iw2rHu차)ix2rHu카)iy2rHu타)iz2rHu파)i{2rHu하)i|2rHu참고)i}2rHu주의)i~2rHu우)i2r3)i�2rHr)i�2rHr)i�2rHro)i�2rHrp)i�2rHu五)i�2rH�六)i�2rHu七)i�2rHr)i�2rHu九)i�2rHr)i�2rHr )i�2rHr$)i�2rHr#)i�2rHr!)i�2rHr2)i�2rHr)i�2rHr)i�2rHu株)i�2rH�有)i�2rH�社)i�2rHu名)i�2rHu特)i�2rHu財)i�2rH�祝)i�2rHu労)i�2rHu秘)i�2rHu男)i�2rHr)i�2rHu適)i�2rHu優)i�2rHu印)i�2rHu注)i�2rHu項)i�2rHu休)i�2rHu写)i�2rHu正)i�2rHrq)i�2rHrr)i�2rHrs)i�2rH�左)i�2rH�右)i�2rHu医)i�2rHu宗)i�2rHu学)i�2rHu監)i�2rHu企r�r�r�r�r��_seg_31�sr}cCsgd�S)N)d)i�2rHu資)i�2rHu協)i�2rHu夜)i�2rHZ36)i�2rHZ37)i�2rHZ38)i�2rHZ39)i�2rHZ40)i�2rHZ41)i�2rHZ42)i�2rHZ43)i�2rHZ44)i�2rHZ45)i�2rHZ46)i�2rHZ47)i�2rHZ48)i�2rHZ49)i�2rHZ50)i�2rHu1月)i�2rHu2月)i�2rHu3月)i�2rHu4月)i�2rHu5月)i�2rHu6月)i�2rHu7月)i�2rHu8月)i�2rHu9月)i�2rHu10月)i�2rHu11月)i�2rHu12月)i�2rH�hg)i�2rHZerg)i�2rHZev)i�2rHZltd)i�2rH�ア)i�2rH�イ)i�2rH�ウ)i�2rH�エ)i�2rH�オ)i�2rH�カ)i�2rH�キ)i�2rH�ク)i�2rH�ケ)i�2rH�コ)i�2rH�サ)i�2rH�シ)i�2rH�ス)i�2rH�セ)i�2rH�ソ)i�2rH�タ)i�2rH�チ)i�2rH�ツ)i�2rH�テ)i�2rH�ト)i�2rH�ナ)i�2rH�ニ)i�2rH�ヌ)i�2rH�ネ)i�2rH�ノ)i�2rH�ハ)i�2rH�ヒ)i�2rH�フ)i�2rH�ヘ)i�2rH�ホ)i�2rH�マ)i�2rH�ミ)i�2rH�ム)i�2rH�メ)i�2rH�モ)i�2rH�ヤ)i�2rH�ユ)i�2rH�ヨ)i�2rH�ラ)i�2rH�リ)i�2rH�ル)i�2rH�レ)i�2rH�ロ)i�2rH�ワ)i�2rHuヰ)i�2rHuヱ)i�2rH�ヲ)i�2rHu令和)i3rHuアパート)i3rHuアルファ)i3rHuアンペア)i3rHu	アール)i3rHuイニング)i3rHu	インチ)i3rHu	ウォン)i3rHuエスクード)i3rHuエーカー)i	3rHu	オンス)i
3rHu	オーム)i3rHu	カイリ)i3rHuカラット)i
3rHuカロリー)i3rHu	ガロン)i3rHu	ガンマ)i3rHuギガ)i3rHu	ギニーr�r�r�r�r��_seg_32)
sr�cCsgd�S)N)d)i3rHuキュリー)i3rHuギルダー)i3rHuキロ)i3rHuキログラム)i3rHuキロメートル)i3rHuキロワット)i3rHu	グラム)i3rHuグラムトン)i3rHuクルゼイロ)i3rHuクローネ)i3rHu	ケース)i3rHu	コルナ)i3rHu	コーポ)i3rHuサイクル)i 3rHuサンチーム)i!3rHuシリング)i"3rHu	センチ)i#3rHu	セント)i$3rHu	ダース)i%3rHuデシ)i&3rHuドル)i'3rHuトン)i(3rHuナノ)i)3rHu	ノット)i*3rHu	ハイツ)i+3rHuパーセント)i,3rHu	パーツ)i-3rHuバーレル)i.3rHuピアストル)i/3rHu	ピクル)i03rHuピコ)i13rHuビル)i23rHuファラッド)i33rHuフィート)i43rHuブッシェル)i53rHu	フラン)i63rHuヘクタール)i73rHuペソ)i83rHu	ペニヒ)i93rHu	ヘルツ)i:3rHu	ペンス)i;3rHu	ページ)i<3rHu	ベータ)i=3rHuポイント)i>3rHu	ボルト)i?3rHuホン)i@3rHu	ポンド)iA3rHu	ホール)iB3rHu	ホーン)iC3rHuマイクロ)iD3rHu	マイル)iE3rHu	マッハ)iF3rHu	マルク)iG3rHuマンション)iH3rHuミクロン)iI3rHuミリ)iJ3rHuミリバール)iK3rHuメガ)iL3rHuメガトン)iM3rHuメートル)iN3rHu	ヤード)iO3rHu	ヤール)iP3rHu	ユアン)iQ3rHuリットル)iR3rHuリラ)iS3rHu	ルピー)iT3rHuルーブル)iU3rHuレム)iV3rHuレントゲン)iW3rHu	ワット)iX3rHu0点)iY3rHu1点)iZ3rHu2点)i[3rHu3点)i\3rHu4点)i]3rHu5点)i^3rHu6点)i_3rHu7点)i`3rHu8点)ia3rHu9点)ib3rHu10点)ic3rHu11点)id3rHu12点)ie3rHu13点)if3rHu14点)ig3rHu15点)ih3rHu16点)ii3rHu17点)ij3rHu18点)ik3rHu19点)il3rHu20点)im3rHu21点)in3rHu22点)io3rHu23点)ip3rHu24点)iq3rH�hpa)ir3rH�da)is3rH�au)it3rH�bar)iu3rH�ovr�r�r�r�r��_seg_33�
sr�cCsgd�S)N)d)iv3rHZpc)iw3rH�dm)ix3rHZdm2)iy3rHZdm3)iz3rH�iu)i{3rHu平成)i|3rHu昭和)i}3rHu大正)i~3rHu明治)i3rHu株式会社)i�3rH�pa)i�3rH�na)i�3rHuμa)i�3rH�ma)i�3rH�ka)i�3rH�kb)i�3rH�mb)i�3rH�gb)i�3rH�cal)i�3rHZkcal)i�3rH�pf)i�3rH�nf)i�3rHuμf)i�3rHuμg)i�3rHZmg)i�3rHZkg)i�3rH�hz)i�3rH�khz)i�3rH�mhz)i�3rH�ghz)i�3rH�thz)i�3rHuμl)i�3rH�ml)i�3rH�dl)i�3rH�kl)i�3rHZfm)i�3rH�nm)i�3rHuμm)i�3rH�mm)i�3rH�cm)i�3rH�km)i�3rHZmm2)i�3rHZcm2)i�3rH�m2)i�3rHZkm2)i�3rHZmm3)i�3rHZcm3)i�3rHZm3)i�3rHZkm3)i�3rHum∕s)i�3rHum∕s2)i�3rHr�)i�3rH�kpa)i�3rH�mpa)i�3rH�gpa)i�3rHZrad)i�3rHurad∕s)i�3rHurad∕s2)i�3rHZps)i�3rH�ns)i�3rHuμs)i�3rH�ms)i�3rH�pv)i�3rH�nv)i�3rHuμv)i�3rH�mv)i�3rH�kv)i�3rHr�)i�3rH�pw)i�3rH�nw)i�3rHuμw)i�3rH�mw)i�3rH�kw)i�3rHr�)i�3rHukω)i�3rHumω)i�3r�)i�3rH�bq)i�3rH�cc)i�3rH�cd)i�3rHuc∕kg)i�3r�)i�3rH�db)i�3rH�gy)i�3rHZha)i�3rH�hp)i�3rH�in)i�3rH�kk)i�3rHr�)i�3rHZkt)i�3rHZlm)i�3rH�ln)i�3rH�log)i�3rHZlx)i�3rHr�)i�3rH�mil)i�3rHZmol)i�3rH�ph)i�3r�)i�3rH�ppmr�r�r�r�r��_seg_34�
sr�cCsgd�S)N)d)i�3rH�pr)i�3rH�sr)i�3rH�sv)i�3rH�wb)i�3rHuv∕m)i�3rHua∕m)i�3rHu1日)i�3rHu2日)i�3rHu3日)i�3rHu4日)i�3rHu5日)i�3rHu6日)i�3rHu7日)i�3rHu8日)i�3rHu9日)i�3rHu10日)i�3rHu11日)i�3rHu12日)i�3rHu13日)i�3rHu14日)i�3rHu15日)i�3rHu16日)i�3rHu17日)i�3rHu18日)i�3rHu19日)i�3rHu20日)i�3rHu21日)i�3rHu22日)i�3rHu23日)i�3rHu24日)i�3rHu25日)i�3rHu26日)i�3rHu27日)i�3rHu28日)i�3rHu29日)i�3rHu30日)i�3rHu31日)i�3rHZgal)i4r3)i��r�)i�r3)i��r�)i��r3)iǤr�)iФr3)i,�r�)i@�rHuꙁ)iA�r3)iB�rHuꙃ)iC�r3)iD�rHuꙅ)iE�r3)iF�rHuꙇ)iG�r3)iH�rHuꙉ)iI�r3)iJ�rHr�)iK�r3)iL�rHuꙍ)iM�r3)iN�rHuꙏ)iO�r3)iP�rHuꙑ)iQ�r3)iR�rHuꙓ)iS�r3)iT�rHuꙕ)iU�r3)iV�rHuꙗ)iW�r3)iX�rHuꙙ)iY�r3)iZ�rHuꙛ)i[�r3)i\�rHuꙝ)i]�r3)i^�rHuꙟ)i_�r3)i`�rHuꙡ)ia�r3)ib�rHuꙣ)ic�r3)id�rHuꙥ)ie�r3)if�rHuꙧ)ig�r3)ih�rHuꙩ)ii�r3)ij�rHuꙫ)ik�r3)il�rHuꙭ)im�r3)i��rHuꚁ)i��r3)i��rHuꚃ)i��r3)i��rHuꚅ)i��r3)i��rHuꚇ)i��r3r�r�r�r�r��_seg_35dsr�cCsgd�S)N)d)i��rHuꚉ)i��r3)i��rHuꚋ)i��r3)i��rHuꚍ)i��r3)i��rHuꚏ)i��r3)i��rHuꚑ)i��r3)i��rHuꚓ)i��r3)i��rHuꚕ)i��r3)i��rHuꚗ)i��r3)i��rHuꚙ)i��r3)i��rHuꚛ)i��r3)i��rHr�)i��rHr�)i��r3)i��r�)i�r3)i"�rHuꜣ)i#�r3)i$�rHuꜥ)i%�r3)i&�rH�ꜧ)i'�r3)i(�rHuꜩ)i)�r3)i*�rHuꜫ)i+�r3)i,�rHuꜭ)i-�r3)i.�rHuꜯ)i/�r3)i2�rHuꜳ)i3�r3)i4�rHuꜵ)i5�r3)i6�rHuꜷ)i7�r3)i8�rHuꜹ)i9�r3)i:�rHuꜻ)i;�r3)i<�rHuꜽ)i=�r3)i>�rHuꜿ)i?�r3)i@�rHuꝁ)iA�r3)iB�rHuꝃ)iC�r3)iD�rHuꝅ)iE�r3)iF�rHuꝇ)iG�r3)iH�rHuꝉ)iI�r3)iJ�rHuꝋ)iK�r3)iL�rHuꝍ)iM�r3)iN�rHuꝏ)iO�r3)iP�rHuꝑ)iQ�r3)iR�rHuꝓ)iS�r3)iT�rHuꝕ)iU�r3)iV�rHuꝗ)iW�r3)iX�rHuꝙ)iY�r3)iZ�rHuꝛ)i[�r3)i\�rHuꝝ)i]�r3)i^�rHuꝟ)i_�r3)i`�rHuꝡ)ia�r3)ib�rHuꝣ)ic�r3)id�rHuꝥ)ie�r3)if�rHuꝧ)ig�r3)ih�rHuꝩ)ii�r3)ij�rHuꝫ)ik�r3)il�rHuꝭ)im�r3)in�rH�ꝯr�r�r�r�r��_seg_36�sr�cCsgd�S)N)d)io�r3)ip�rHr�)iq�r3)iy�rHuꝺ)iz�r3)i{�rHuꝼ)i|�r3)i}�rHuᵹ)i~�rHuꝿ)i�r3)i��rHuꞁ)i��r3)i��rHuꞃ)i��r3)i��rHuꞅ)i��r3)i��rHuꞇ)i��r3)i��rHuꞌ)i��r3)i��rHr�)i��r3)i��rHuꞑ)i��r3)i��rHuꞓ)i��r3)i��rHuꞗ)i��r3)i��rHuꞙ)i��r3)i��rHuꞛ)i��r3)i��rHuꞝ)i��r3)i��rHuꞟ)i��r3)i��rHuꞡ)i��r3)i��rHuꞣ)i��r3)i��rHuꞥ)i��r3)i��rHuꞧ)i��r3)i��rHuꞩ)i��r3)i��rHrf)i��rHr�)i��rHr�)i��rHuɬ)i��rHr�)i��r3)i��rHuʞ)i��rHuʇ)i��rHr�)i��rHuꭓ)i��rHuꞵ)i��r3)i��rHuꞷ)i��r3)i��rHuꞹ)i��r3)i��rHuꞻ)i��r3)i��rHuꞽ)i��r3)i��rHuꞿ)i��r3)i��r�)i§rHuꟃ)içr3)iħrHuꞔ)iŧrHr�)iƧrHuᶎ)iǧrHuꟈ)iȧr3)iɧrHuꟊ)iʧr3)i˧r�)i��rHuꟶ)i��r3)i��rHrN)i��rHrQ)i��r3)i-�r�)i0�r3)i:�r�)i@�r3)ix�r�)i��r3)iƨr�)iΨr3)iڨr�)i�r3)iT�r�)i_�r3)i}�r�)i��r3)iΩr�)iϩr3r�r�r�r�r��_seg_376sr�cCsgd�S)N)d)iکr�)iީr3)i��r�)i�r3)i7�r�)i@�r3)iN�r�)iP�r3)iZ�r�)i\�r3)iêr�)i۪r3)i��r�)i�r3)i�r�)i	�r3)i�r�)i�r3)i�r�)i �r3)i'�r�)i(�r3)i/�r�)i0�r3)i\�rHr�)i]�rHuꬷ)i^�rHr
)i_�rHuꭒ)i`�r3)ii�rHuʍ)ij�r3)il�r�)ip�rHuᎠ)iq�rHuᎡ)ir�rHuᎢ)is�rHuᎣ)it�rHuᎤ)iu�rHuᎥ)iv�rHuᎦ)iw�rHuᎧ)ix�rHuᎨ)iy�rHuᎩ)iz�rHuᎪ)i{�rHuᎫ)i|�rHuᎬ)i}�rHuᎭ)i~�rHuᎮ)i�rHuᎯ)i��rHuᎰ)i��rHuᎱ)i��rHuᎲ)i��rHuᎳ)i��rHuᎴ)i��rHuᎵ)i��rHuᎶ)i��rHuᎷ)i��rHuᎸ)i��rHuᎹ)i��rHuᎺ)i��rHuᎻ)i��rHuᎼ)i��rHuᎽ)i��rHuᎾ)i��rHuᎿ)i��rHuᏀ)i��rHuᏁ)i��rHuᏂ)i��rHuᏃ)i��rHuᏄ)i��rHuᏅ)i��rHuᏆ)i��rHuᏇ)i��rHuᏈ)i��rHuᏉ)i��rHuᏊ)i��rHuᏋ)i��rHuᏌ)i��rHuᏍ)i��rHuᏎ)i��rHuᏏ)i��rHuᏐ)i��rHuᏑ)i��rHuᏒ)i��rHuᏓ)i��rHuᏔ)i��rHuᏕ)i��rHuᏖ)i��rHuᏗ)i��rHuᏘ)i��rHuᏙ)i��rHuᏚ)i��rHuᏛ)i��rHuᏜ)i��rHuᏝ)i��rHuᏞ)i��rHuᏟ)i��rHuᏠ)i��rHuᏡ)i��rHuᏢ)i��rHuᏣr�r�r�r�r��_seg_38�sr�cCsgd�S)N)d)i��rHuᏤ)i��rHuᏥ)i��rHuᏦ)i��rHuᏧ)i��rHuᏨ)i��rHuᏩ)i��rHuᏪ)i��rHuᏫ)i��rHuᏬ)i��rHuᏭ)i��rHuᏮ)i��rHuᏯ)i��r3)i�r�)i�r3)i��r�)i�r3)i��r�)i��r3)i��r�)i��r3)i��r�)i�rHu豈)i�rHu更)i�rHr/)i�rHu賈)i�rHu滑)i�rHu串)i�rHu句)i�rHr8)i	�rHu契)i
�rHr2)i�rHu喇)i�rHu奈)i
�rH�懶)i�rHu癩)i�rHu羅)i�rHu蘿)i�rHu螺)i�rHu裸)i�rHu邏)i�rH�樂)i�rHu洛)i�rHu烙)i�rHu珞)i�rHu落)i�rHu酪)i�rHu駱)i�rHu亂)i�rHu卵)i�rHu欄)i�rHu爛)i�rHu蘭)i �rHu鸞)i!�rHu嵐)i"�rHu濫)i#�rHu藍)i$�rHu襤)i%�rHu拉)i&�rHu臘)i'�rHu蠟)i(�rH�廊)i)�rH�朗)i*�rHu浪)i+�rHu狼)i,�rHu郎)i-�rHu來)i.�rHu冷)i/�rHu勞)i0�rHu擄)i1�rHu櫓)i2�rHu爐)i3�rHu盧)i4�rHr))i5�rHu蘆)i6�rH�虜)i7�rHu路)i8�rHu露)i9�rHu魯)i:�rHu鷺)i;�rH�碌)i<�rHu祿)i=�rHu綠)i>�rHu菉)i?�rHu錄)i@�rHr3)iA�rHu論)iB�rHu壟)iC�rHu弄)iD�rHu籠)iE�rHu聾)iF�rHu牢)iG�rHu磊)iH�rHu賂)iI�rHu雷)iJ�rHu壘)iK�rHu屢)iL�rHu樓)iM�rHu淚)iN�rHu漏r�r�r�r�r��_seg_39sr�cCsgd�S)N)d)iO�rHu累)iP�rHu縷)iQ�rHu陋)iR�rHu勒)iS�rHu肋)iT�rHu凜)iU�rHu凌)iV�rHu稜)iW�rHu綾)iX�rHu菱)iY�rHu陵)iZ�rHu讀)i[�rHu拏)i\�rHr�)i]�rH�諾)i^�rHu丹)i_�rH�寧)i`�rHu怒)ia�rH�率)ib�rH�異)ic�rH�北)id�rHu磻)ie�rHu便)if�rHu復)ig�rHu不)ih�rHu泌)ii�rHu數)ij�rHu索)ik�rHu參)il�rHu塞)im�rHu省)in�rHu葉)io�rH�說)ip�rH�殺)iq�rHr0)ir�rHu沈)is�rHu拾)it�rH�若)iu�rHu掠)iv�rHu略)iw�rHu亮)ix�rHu兩)iy�rHu凉)iz�rHu梁)i{�rHu糧)i|�rHu良)i}�rHu諒)i~�rHu量)i�rHu勵)i��rHu呂)i��rHr)i��rHu廬)i��rHu旅)i��rHu濾)i��rHu礪)i��rHu閭)i��rHu驪)i��rHu麗)i��rHu黎)i��rHr)i��rHu曆)i��rHu歷)i��rHu轢)i��rHu年)i��rHu憐)i��rHu戀)i��rHu撚)i��rHu漣)i��rHu煉)i��rHu璉)i��rHu秊)i��rH�練)i��rHu聯)i��rHu輦)i��rHu蓮)i��rHu連)i��rHu鍊)i��rHu列)i��rHu劣)i��rHu咽)i��rHu烈)i��rHu裂)i��rHr�)i��rHu廉)i��rHu念)i��rHu捻)i��rHu殮)i��rHu簾)i��rHu獵)i��rHu令)i��rHu囹)i��rHr�)i��rHu嶺)i��rHu怜)i��rHu玲)i��rHu瑩)i��rHu羚)i��rHu聆)i��rHu鈴)i��rHu零r�r�r�r�r��_seg_40qsrcCsgd�S)N)d)i��rHu靈)i��rHu領)i��rHu例)i��rHu禮)i��rHu醴)i��rHu隸)i��rHu惡)i��rHu了)i��rHu僚)i��rHu寮)i��rHu尿)i��rH�料)i��rHr�)i��rHu燎)i��rHu療)i�rHu蓼)i�rHu遼)i�rHr7)i�rHu暈)i�rHu阮)i�rHu劉)i�rHu杻)i�rHu柳)i�rH�流)i�rHu溜)i�rHu琉)i�rHu留)i�rHu硫)i�rHu紐)i�rHu類)i�rHrw)i�rHu戮)i�rHu陸)i�rHu倫)i�rHu崙)i�rHu淪)i�rHu輪)i�rHu律)i�rHu慄)i�rHu栗)i�rHr�)i�rHu隆)i�rHu利)i�rHu吏)i�rHu履)i�rHu易)i�rHu李)i�rHu梨)i�rHu泥)i�rHu理)i�rHu痢)i�rHu罹)i�rHu裏)i�rHu裡)i�rHr1)i�rHu離)i�rHu匿)i�rHu溺)i�rHu吝)i�rHu燐)i�rHu璘)i�rHu藺)i�rHu隣)i�rHu鱗)i�rHu麟)i�rHu林)i��rHu淋)i��rHu臨)i��rHr&)i��rHu笠)i��rHu粒)i��rHu狀)i��rHu炙)i��rHu識)i��rHu什)i��rHu茶)i��rHu刺)i�rH�切)i�rHu度)i�rHu拓)i�rHu糖)i�rHu宅)i�rHu洞)i�rHu暴)i�rHu輻)i�rHr*)i	�rHu降)i
�rHr,)i�rHu廓)i�rHu兀)i
�rHu嗀)i�r3)i�rH�塚)i�r3)i�rH�晴)i�r3)i�rHu凞)i�rH�猪)i�rH�益)i�rHu礼r�r�r�r�r��_seg_41�sr	cCsgd�S)N)d)i�rHu神)i�rHu祥)i�rH�福)i�rH�靖)i�rHu精)i�rHr()i�r3)i �rHu蘒)i!�r3)i"�rH�諸)i#�r3)i%�rH�逸)i&�rHu都)i'�r3)i*�rHu飯)i+�rHu飼)i,�rHu館)i-�rHu鶴)i.�rHu郞)i/�rHu隷)i0�rH�侮)i1�rH�僧)i2�rH�免)i3�rH�勉)i4�rH�勤)i5�rH�卑)i6�rH�喝)i7�rH�嘆)i8�rHu器)i9�rHu塀)i:�rHu墨)i;�rHu層)i<�rHr)i=�rH�悔)i>�rHu慨)i?�rH�憎)i@�rH�懲)iA�rH�敏)iB�rHu既)iC�rH�暑)iD�rH�梅)iE�rH�海)iF�rHu渚)iG�rH�漢)iH�rH�煮)iI�rHu爫)iJ�rHu琢)iK�rHu碑)iL�rHry)iM�rHu祉)iN�rHu祈)iO�rHu祐)iP�rH�祖)iQ�rHrz)iR�rHu禍)iS�rHu禎)iT�rH�穀)iU�rHu突)iV�rH�節)iW�rHr)iX�rHu縉)iY�rHu繁)iZ�rHu署)i[�rH�者)i\�rHu臭)i]�rHu艹)i_�rH�著)i`�rHu褐)ia�rH�視)ib�rH�謁)ic�rH�謹)id�rHu賓)ie�rH�贈)if�rHu辶)ig�rHr
)ih�rH�難)ii�rH�響)ij�rH�頻)ik�rHu恵)il�rHu𤋮)im�rHu舘)in�r�)ip�rHu並)iq�rH�况)ir�rHu全)is�rHu侀)it�rHu充)iu�rHu冀)iv�rH�勇)iw�rH�勺)ix�rHr)iy�rHu啕)iz�rH�喙)i{�rHu嗢)i|�rHr)i}�rHu墳)i~�rHu奄)i�rHu奔)i��rHu婢)i��rHu嬨r�r�r�r�r��_seg_42Csr/cCsgd�S)N)d)i��rHu廒)i��rHu廙)i��rHu彩)i��rHu徭)i��rHu惘)i��rH�慎)i��rHu愈)i��rHr)i��rHu慠)i��rHr)i��rHu戴)i��rHu揄)i��rHu搜)i��rHu摒)i��rHu敖)i��rHr)i��rHr�)i��rH�望)i��rHu杖)i��rHr")i��rHr�)i��rHr)i��rHu滛)i��rH�滋)i��rHr)i��rH�瀞)i��rHr)i��rHu瞧)i��rH�爵)i��rHu犯)i��rHr)i��rH�瑱)i��rHu甆)i��rHu画)i��rHu瘝)i��rHu瘟)i��rHr)i��rHu盛)i��rH�直)i��rH�睊)i��rHu着)i��rH�磌)i��rHu窱)i��rHr!)i��rHu类)i��rHu絛)i��rHr)i��rHu缾)i��rHr")i��rHu荒)i��rHu華)i��rH�蝹)i��rHu襁)i��rHu覆)i��rHr$)i��rHu調)i��rHr)i��rHu請)i��rHr%)i��rHr�)i��rH�諭)i��rHr&)i��rH�變)i��rHr')i�rH�輸)i�rHu遲)i�rHu醙)i�rHu鉶)i�rHu陼)i�rHr()i�rHr)i�rHu韛)i�rHr))i�rH�頋)i�rHr*)i�rH�鬒)i�rHr8)i�rHu𢡊)i�rHu𢡄)i�rHu𣏕)i�rH�㮝)i�rHu䀘)i�rH�䀹)i�rHu𥉉)i�rHu𥳐)i�rHu𧻓)i�rHu齃)i�rHu龎)i�r�)i�rH�ff)i�rH�fi)i�rH�fl)i�rH�ffi)i�rH�ffl)i�rH�st)i�r�)i�rHuմն)i�rHuմե)i�rHuմի)i�rHuվնr�r�r�r�r��_seg_43�srGcCsgd�S)N)d)i�rHuմխ)i�r�)i�rHuיִ)i�r3)i�rHuײַ)i �rHuע)i!�rHr�)i"�rHr�)i#�rHuה)i$�rHuכ)i%�rHuל)i&�rHuם)i'�rHuר)i(�rHuת)i)�rr�)i*�rHuשׁ)i+�rHuשׂ)i,�rHuשּׁ)i-�rHuשּׂ)i.�rHuאַ)i/�rHuאָ)i0�rHuאּ)i1�rHuבּ)i2�rHuגּ)i3�rHuדּ)i4�rHuהּ)i5�rHuוּ)i6�rHuזּ)i7�r�)i8�rHuטּ)i9�rHuיּ)i:�rHuךּ)i;�rHuכּ)i<�rHuלּ)i=�r�)i>�rHuמּ)i?�r�)i@�rHuנּ)iA�rHuסּ)iB�r�)iC�rHuףּ)iD�rHuפּ)iE�r�)iF�rHuצּ)iG�rHuקּ)iH�rHuרּ)iI�rHuשּ)iJ�rHuתּ)iK�rHuוֹ)iL�rHuבֿ)iM�rHuכֿ)iN�rHuפֿ)iO�rHuאל)iP�rHuٱ)iR�rHuٻ)iV�rHuپ)iZ�rHuڀ)i^�rHuٺ)ib�rHuٿ)if�rHuٹ)ij�rHuڤ)in�rHuڦ)ir�rHuڄ)iv�rHuڃ)iz�rHuچ)i~�rHuڇ)i��rHuڍ)i��rHuڌ)i��rHuڎ)i��rHuڈ)i��rHuژ)i��rHuڑ)i��rHuک)i��rHuگ)i��rHuڳ)i��rHuڱ)i��rH�ں)i��rHuڻ)i��rHuۀ)i��rHuہ)i��rHuھ)i��rHuے)i��rHuۓ)i��r3)i�r�)i�rHuڭ)i�rHuۇ)i�rHuۆ)i�rHuۈ)i�rHr�)i�rHuۋ)i�rHuۅ)i�rHuۉ)i�rHuې)i�rH�ى)i�rHuئا)i�rHuئە)i�rHuئو)i�rHuئۇ)i�rHuئۆr�r�r�r�r��_seg_44srJcCsgd�S)N)d)i�rHuئۈ)i��rHuئې)i��rH�ئى)i��rHuی)i�rH�ئج)i�rH�ئح)i�rH�ئم)i�rHrK)i�rH�ئي)i�rH�بج)i�rH�بح)i�rH�بخ)i�rH�بم)i	�rH�بى)i
�rH�بي)i�rH�تج)i�rH�تح)i
�rH�تخ)i�rH�تم)i�rH�تى)i�rH�تي)i�rHuثج)i�rH�ثم)i�rH�ثى)i�rH�ثي)i�rH�جح)i�rH�جم)i�rH�حج)i�rH�حم)i�rH�خج)i�rHuخح)i�rH�خم)i�rH�سج)i�rH�سح)i�rH�سخ)i�rH�سم)i �rH�صح)i!�rH�صم)i"�rH�ضج)i#�rH�ضح)i$�rH�ضخ)i%�rH�ضم)i&�rH�طح)i'�rH�طم)i(�rH�ظم)i)�rH�عج)i*�rH�عم)i+�rH�غج)i,�rH�غم)i-�rH�فج)i.�rH�فح)i/�rH�فخ)i0�rH�فم)i1�rH�فى)i2�rH�في)i3�rH�قح)i4�rH�قم)i5�rH�قى)i6�rH�قي)i7�rH�كا)i8�rH�كج)i9�rH�كح)i:�rH�كخ)i;�rH�كل)i<�rH�كم)i=�rH�كى)i>�rH�كي)i?�rH�لج)i@�rH�لح)iA�rH�لخ)iB�rH�لم)iC�rH�لى)iD�rH�لي)iE�rH�مج)iF�rH�مح)iG�rH�مخ)iH�rH�مم)iI�rHuمى)iJ�rHuمي)iK�rH�نج)iL�rH�نح)iM�rH�نخ)iN�rH�نم)iO�rH�نى)iP�rH�ني)iQ�rH�هج)iR�rH�هم)iS�rHuهى)iT�rHuهي)iU�rH�يج)iV�rH�يح)iW�rH�يخ)iX�rH�يم)iY�rH�يى)iZ�rH�يي)i[�rHuذٰ)i\�rHuرٰ)i]�rH�ىٰ)i^�ru ٌّ)i_�ru ٍّr�r�r�r�r��_seg_45~sr�cCsgd�S)N)d)i`�ru َّ)ia�ru ُّ)ib�ru ِّ)ic�ru ّٰ)id�rHuئر)ie�rHuئز)if�rHrN)ig�rHuئن)ih�rHrK)ii�rHrO)ij�rHuبر)ik�rHuبز)il�rHrS)im�rHuبن)in�rHrT)io�rHrU)ip�rHuتر)iq�rHuتز)ir�rHrY)is�rHuتن)it�rHrZ)iu�rHr[)iv�rHuثر)iw�rHuثز)ix�rHr\)iy�rHuثن)iz�rHr])i{�rHr^)i|�rHrz)i}�rHr{)i~�rHr~)i�rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHuما)i��rHr�)i��rHuنر)i��rHuنز)i��rHr�)i��rHuنن)i��rHr�)i��rHr�)i��rHr�)i��rHuير)i��rHuيز)i��rHr�)i��rHuين)i��rHr�)i��rHr�)i��rHrL)i��rHrM)i��rHuئخ)i��rHrN)i��rH�ئه)i��rHrP)i��rHrQ)i��rHrR)i��rHrS)i��rH�به)i��rHrV)i��rHrW)i��rHrX)i��rHrY)i��rH�ته)i��rHr\)i��rHr_)i��rHr`)i��rHra)i��rHrb)i��rHrc)i��rHrd)i��rHre)i��rHrf)i��rHrg)i��rHrh)i��rHri)i��rHuصخ)i��rHrj)i��rHrk)i��rHrl)i��rHrm)i��rHrn)i��rHro)i��rHrq)i��rHrr)i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i�rHr|)i�rHr}r�r�r�r�r��_seg_46�sr�cCsgd�S)N)d)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHuله)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�نه)i�rHr�)i�rHr�)i�rHuهٰ)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�يه)i�rHrN)i�rHr�)i�rHrS)i�rHr�)i�rHrY)i�rHr�)i�rHr\)i�rHuثه)i�rHrh)i�rH�سه)i�rH�شم)i�rH�شه)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHuـَّ)i�rHuـُّ)i�rHuـِّ)i��rH�طى)i��rH�طي)i��rH�عى)i��rH�عي)i��rH�غى)i��rH�غي)i��rH�سى)i��rH�سي)i��rH�شى)i��rH�شي)i��rH�حى)i�rH�حي)i�rH�جى)i�rH�جي)i�rH�خى)i�rH�خي)i�rH�صى)i�rH�صي)i�rH�ضى)i�rH�ضي)i	�rH�شج)i
�rH�شح)i�rH�شخ)i�rHr�)i
�rH�شر)i�rH�سر)i�rH�صر)i�rH�ضر)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i �rHr�)i!�rHr�)i"�rHr�)i#�rHr�)i$�rHr�)i%�rHr�)i&�rHr�)i'�rHr�r�r�r�r�r��_seg_47Psr�cCsgd�S)N)d)i(�rHr�)i)�rHr�)i*�rHr�)i+�rHr�)i,�rHr�)i-�rHr�)i.�rHr�)i/�rHr�)i0�rHr�)i1�rHr�)i2�rHr�)i3�rHrp)i4�rHre)i5�rHrf)i6�rHrg)i7�rHr�)i8�rHr�)i9�rHr�)i:�rHrp)i;�rHrq)i<�rHuاً)i>�r3)i@�r�)iP�rHuتجم)iQ�rHuتحج)iS�rHuتحم)iT�rHuتخم)iU�rHuتمج)iV�rHuتمح)iW�rHuتمخ)iX�rHuجمح)iZ�rHuحمي)i[�rHuحمى)i\�rHuسحج)i]�rHuسجح)i^�rHuسجى)i_�rHuسمح)ia�rHuسمج)ib�rHuسمم)id�rHuصحح)if�rH�صمم)ig�rHuشحم)ii�rHuشجي)ij�rHuشمخ)il�rHuشمم)in�rHuضحى)io�rHuضخم)iq�rHuطمح)is�rHuطمم)it�rHuطمي)iu�rH�عجم)iv�rHuعمم)ix�rHuعمى)iy�rHuغمم)iz�rHuغمي)i{�rHuغمى)i|�rHuفخم)i~�rH�قمح)i�rHuقمم)i��rH�لحم)i��rHuلحي)i��rHuلحى)i��rHuلجج)i��rHuلخم)i��rHuلمح)i��rHuمحج)i��rHuمحم)i��rHuمحي)i��rHuمجح)i��rHuمجم)i��rHuمخج)i��rHuمخم)i��r�)i��rHuمجخ)i��rHuهمج)i��rHuهمم)i��rHuنحم)i��rHuنحى)i��rHuنجم)i��rHuنجى)i��rHuنمي)i��rHuنمى)i��rHuيمم)i��rHuبخي)i��rHuتجي)i��rHuتجى)i��rHuتخي)i��rHuتخى)i��rHuتمي)i��rHuتمى)i��rHuجمي)i��rHuجحى)i��rHuجمى)i��rHuسخى)i��rHuصحي)i��rHuشحي)i��rHuضحي)i��rHuلجي)i��rHuلمي)i��rHuيحيr�r�r�r�r��_seg_48�sr�cCsgd�S)N)d)i��rHuيجي)i��rHuيمي)i��rHuممي)i��rHuقمي)i��rHuنحي)i��rHr�)i��rHr�)i��rHuعمي)i��rHuكمي)i��rH�نجح)i��rHuمخي)i��rH�لجم)i��rH�كمم)i��rHr�)i��rHr�)i��rHuجحي)i��rHuحجي)i��rHuمجي)i��rHuفمي)i�rHuبحي)i�rHr�)i�rHr�)i�rHr�)i�rHuسخي)i�rHuنجي)i�r�)i�rHuصلے)i�rHuقلے)i�rHuالله)i�rHuاكبر)i�rHuمحمد)i��rHuصلعم)i��rHuرسول)i��rHuعليه)i��rHuوسلم)i��rHuصلى)i��ru!صلى الله عليه وسلم)i��ruجل جلاله)i��rHuریال)i��r3)i��r�)i�r�)i�r�,)i�rH�、)i�r�)i�r�:)i�rri)i�r�!)i�r�?)i�rHu〖)i�rHu〗)i�r�)i �r3)i0�r�)i1�rH�—)i2�rHu–)i3�r�_)i5�rr�)i6�rr�)i7�r�{)i8�r�})i9�rH�〔)i:�rH�〕)i;�rHu【)i<�rHu】)i=�rHu《)i>�rHu》)i?�rHr�)i@�rHr�)iA�rH�「)iB�rH�」)iC�rHu『)iD�rHu』)iE�r3)iG�r�[)iH�r�])iI�rr�)iM�rr�)iP�rr�)iQ�rHr�)iR�r�)iT�rri)iU�rr�)iV�rr�)iW�rr�)iX�rHr�)iY�rr�)iZ�rr�)i[�rr�)i\�rr�)i]�rHr�)i^�rHr�)i_�r�#)i`�r�&)ia�r�*)ib�rr�)ic�rH�-)id�r�<)ie�r�>)if�rr�r�r�r�r�r��_seg_49"sr�cCsgd�S)N)d)ig�r�)ih�r�\)ii�r�$)ij�r�%)ik�r�@)il�r�)ip�ru ً)iq�rHuـً)ir�ru ٌ)is�r3)it�ru ٍ)iu�r�)iv�ru َ)iw�rHuـَ)ix�ru ُ)iy�rHuـُ)iz�ru ِ)i{�rHuـِ)i|�ru ّ)i}�rHuـّ)i~�ru ْ)i�rHuـْ)i��rHuء)i��rHuآ)i��rHuأ)i��rHuؤ)i��rHuإ)i��rHuئ)i��rH�ا)i��rH�ب)i��rHuة)i��rH�ت)i��rH�ث)i��rH�ج)i��rH�ح)i��rH�خ)i��rH�د)i��rH�ذ)i��rH�ر)i��rH�ز)i��rH�س)i��rH�ش)i��rH�ص)i��rH�ض)i��rH�ط)i�rH�ظ)i�rH�ع)i�rH�غ)i�rH�ف)i�rH�ق)i�rH�ك)i�rH�ل)i�rH�م)i�rH�ن)i�rH�ه)i�rH�و)i�rHrI)i�rH�ي)i��rHuلآ)i��rHuلأ)i��rHuلإ)i��rHuلا)i��r�)i��r�)i�r�)i�rr�)i�r�")i�rr�)i�rr�)i�rr�)i�rr�)i�r�')i�rr�)i	�rr�)i
�rr�)i�rr�)i�rr�)i
�rHr�)i�rHr9)i�r�/)i�rHr�)i�rHr�)i�rHr�)i�rHr)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rr�)i�rri)i�rr�)i�rr�)i�rr�)i�rr�)i �rr�)i!�rHrI)i"�rHrK)i#�rHrMr�r�r�r�r��_seg_50�srcCsgd�S)N)d)i$�rHrO)i%�rHrQ)i&�rHrS)i'�rHrU)i(�rHrW)i)�rHrY)i*�rHr[)i+�rHr])i,�rHr_)i-�rHra)i.�rHrc)i/�rHre)i0�rHrg)i1�rHri)i2�rHrk)i3�rHrm)i4�rHro)i5�rHrq)i6�rHrs)i7�rHru)i8�rHrw)i9�rHry)i:�rHr{)i;�rr�)i<�rr�)i=�rr�)i>�r�^)i?�rr�)i@�rr�)iA�rHrI)iB�rHrK)iC�rHrM)iD�rHrO)iE�rHrQ)iF�rHrS)iG�rHrU)iH�rHrW)iI�rHrY)iJ�rHr[)iK�rHr])iL�rHr_)iM�rHra)iN�rHrc)iO�rHre)iP�rHrg)iQ�rHri)iR�rHrk)iS�rHrm)iT�rHro)iU�rHrq)iV�rHrs)iW�rHru)iX�rHrw)iY�rHry)iZ�rHr{)i[�rr�)i\�r�|)i]�rr�)i^�r�~)i_�rHu⦅)i`�rHu⦆)ia�rHr9)ib�rHr�)ic�rHr�)id�rHr�)ie�rHu・)if�rHr�)ig�rHuァ)ih�rHuィ)ii�rHuゥ)ij�rHuェ)ik�rHuォ)il�rHuャ)im�rHuュ)in�rHuョ)io�rHuッ)ip�rHuー)iq�rHr)ir�rHr�)is�rHr�)it�rHr�)iu�rHr�)iv�rHr�)iw�rHr�)ix�rHr�)iy�rHr�)iz�rHr�)i{�rHr�)i|�rHr�)i}�rHr�)i~�rHr�)i�rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�r�r�r�r�r��_seg_51�srcCsgd�S)N)d)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHuン)i��rHu゙)i��rHu゚)i��r�)i��rHr;)i��rHr<)i��rHr=)i��rHr>)i��rHr?)i��rHr@)i��rHrA)i��rHrB)i��rHrC)i��rHrD)i��rHrE)i��rHrF)i��rHrG)i��rHrH)i��rHrI)i��rHrJ)i��rHrK)i��rHrL)i��rHrM)i��rHrN)i��rHrO)i��rHrP)i��rHrQ)i��rHrR)i��rHrS)i��rHrT)i��rHrU)i��rHrV)i��rHrW)i��rHrX)i��r�)i�rHrY)i�rHrZ)i�rHr[)i�rHr\)i�rHr])i�rHr^)i�r�)i�rHr_)i�rHr`)i�rHra)i�rHrb)i�rHrc)i�rHrd)i�r�)i�rHre)i�rHrf)i�rHrg)i�rHrh)i�rHri)i�rHrj)i�r�)i�rHrk)i�rHrl)i�rHrm)i�r�)i�rH�¢)i�rH�£)i�rH�¬)i�rr�)i�rH�¦)i�rH�¥)i�rHu₩)i�r�)i�rHu│)i�rHu←)i�rHu↑)i�rHu→)i�rHu↓)i�rHu■)i�rHu○)i�r�)ir3)ir�)i
r3r�r�r�r�r��_seg_52]srcCsgd�S)N)d)i'r�)i(r3)i;r�)i<r3)i>r�)i?r3)iNr�)iPr3)i^r�)i�r3)i�r�)ir3)ir�)ir3)i4r�)i7r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i$r�)i-r3)iKr�)iPr3)i{r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)irHu𐐨)irHu𐐩)irHu𐐪)irHu𐐫)irHu𐐬)irHu𐐭)irHu𐐮)irHu𐐯)irHu𐐰)i	rHu𐐱)i
rHu𐐲)irHu𐐳)irHu𐐴)i
rHu𐐵)irHu𐐶)irHu𐐷)irHu𐐸)irHu𐐹)irHu𐐺)irHu𐐻)irHu𐐼)irHu𐐽)irHu𐐾)irHu𐐿)irHu𐑀)irHu𐑁)irHu𐑂)irHu𐑃)irHu𐑄)irHu𐑅)irHu𐑆)irHu𐑇)i rHu𐑈)i!rHu𐑉)i"rHu𐑊)i#rHu𐑋)i$rHu𐑌)i%rHu𐑍)i&rHu𐑎)i'rHu𐑏)i(r3)i�r�)i�r3)i�r�)i�rHu𐓘)i�rHu𐓙)i�rHu𐓚)i�rHu𐓛)i�rHu𐓜)i�rHu𐓝)i�rHu𐓞)i�rHu𐓟)i�rHu𐓠)i�rHu𐓡)i�rHu𐓢)i�rHu𐓣)i�rHu𐓤)i�rHu𐓥)i�rHu𐓦r�r�r�r�r��_seg_53�srcCsgd�S)N)d)i�rHu𐓧)i�rHu𐓨)i�rHu𐓩)i�rHu𐓪)i�rHu𐓫)i�rHu𐓬)i�rHu𐓭)i�rHu𐓮)i�rHu𐓯)i�rHu𐓰)i�rHu𐓱)i�rHu𐓲)i�rHu𐓳)i�rHu𐓴)i�rHu𐓵)i�rHu𐓶)i�rHu𐓷)i�rHu𐓸)i�rHu𐓹)i�rHu𐓺)i�rHu𐓻)i�r�)i�r3)i�r�)ir3)i(r�)i0r3)idr�)ior3)ipr�)ir3)i7r�)i@r3)iVr�)i`r3)ihr�)ir3)ir�)ir3)i	r�)i
r3)i6r�)i7r3)i9r�)i<r3)i=r�)i?r3)iVr�)iWr3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i	r�)i	r3)i:	r�)i?	r3)i@	r�)i�	r3)i�	r�)i�	r3)i�	r�)i�	r3)i
r�)i
r3)i
r�)i
r3)i
r�)i
r3)i
r�)i
r3)i6
r�)i8
r3)i;
r�)i?
r3)iI
r�)iP
r3)iY
r�)i`
r3)i�
r�)i�
r3)i�
r�)i�
r3)i�
r�)ir3)i6r�)i9r3)iVr�)iXr3)isr�)ixr3)i�r�)i�r3)i�r�)i�r3)i�r�r�r�r�r�r��_seg_54/srcCsgd�S)N)d)ir3)iIr�)i�rHu𐳀)i�rHu𐳁)i�rHu𐳂)i�rHu𐳃)i�rHu𐳄)i�rHu𐳅)i�rHu𐳆)i�rHu𐳇)i�rHu𐳈)i�rHu𐳉)i�rHu𐳊)i�rHu𐳋)i�rHu𐳌)i�rHu𐳍)i�rHu𐳎)i�rHu𐳏)i�rHu𐳐)i�rHu𐳑)i�rHu𐳒)i�rHu𐳓)i�rHu𐳔)i�rHu𐳕)i�rHu𐳖)i�rHu𐳗)i�rHu𐳘)i�rHu𐳙)i�rHu𐳚)i�rHu𐳛)i�rHu𐳜)i�rHu𐳝)i�rHu𐳞)i�rHu𐳟)i�rHu𐳠)i�rHu𐳡)i�rHu𐳢)i�rHu𐳣)i�rHu𐳤)i�rHu𐳥)i�rHu𐳦)i�rHu𐳧)i�rHu𐳨)i�rHu𐳩)i�rHu𐳪)i�rHu𐳫)i�rHu𐳬)i�rHu𐳭)i�rHu𐳮)i�rHu𐳯)i�rHu𐳰)i�rHu𐳱)i�rHu𐳲)i�r�)i�r3)i�r�)i�r3)i(
r�)i0
r3)i:
r�)i`r3)ir�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i(r�)i0r3)iZr�)i�r3)i�r�)i�r3)i�r�)ir3)iNr�)iRr3)ipr�)ir3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)i5r�)i6r3)iHr�)iPr3)iwr�)i�r3)i�r�)i�r3)i�r�)ir3)ir�r�r�r�r�r��_seg_55�srcCsgd�S)N)d)ir3)i?r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)ir3)i
r�)ir3)ir�)ir3)i)r�)i*r3)i1r�)i2r3)i4r�)i5r3)i:r�)i;r3)iEr�)iGr3)iIr�)iKr3)iNr�)iPr3)iQr�)iWr3)iXr�)i]r3)idr�)ifr3)imr�)ipr3)iur�)ir3)i\r�)i]r3)ibr�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)iEr�)iPr3)iZr�)i`r3)imr�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)ir3)i,r�)i0r3)i@r�)ir3)i<r�)i�rHu𑣀)i�rHu𑣁)i�rHu𑣂)i�rHu𑣃)i�rHu𑣄)i�rHu𑣅)i�rHu𑣆)i�rHu𑣇)i�rHu𑣈)i�rHu𑣉)i�rHu𑣊)i�rHu𑣋)i�rHu𑣌)i�rHu𑣍)i�rHu𑣎)i�rHu𑣏)i�rHu𑣐)i�rHu𑣑)i�rHu𑣒)i�rHu𑣓)i�rHu𑣔)i�rHu𑣕)i�rHu𑣖)i�rHu𑣗r�r�r�r�r��_seg_56srcCsgd�S)N)d)i�rHu𑣘)i�rHu𑣙)i�rHu𑣚)i�rHu𑣛)i�rHu𑣜)i�rHu𑣝)i�rHu𑣞)i�rHu𑣟)i�r3)i�r�)i�r3)ir�)i	r3)i
r�)ir3)ir�)ir3)ir�)ir3)i6r�)i7r3)i9r�)i;r3)iGr�)iPr3)iZr�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)iHr�)iPr3)i�r�)i�r3)i�r�)ir3)i	r�)i
r3)i7r�)i8r3)iFr�)iPr3)imr�)ipr3)i�r�)i�r3)i�r�)i�r3)i�r�)ir3)ir�)ir3)i
r�)ir3)i7r�)i:r3)i;r�)i<r3)i>r�)i?r3)iHr�)iPr3)iZr�)i`r3)ifr�)igr3)iir�)ijr3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i�#r�)i$r3)io$r�)ip$r3)iu$r�)i�$r3)iD%r�)i0r3)i/4r�)iDr3)iGFr�)ihr3)i9jr�)i@jr3)i_jr�r�r�r�r�r��_seg_57jsrcCsgd�S)N)d)i`jr3)ijjr�)injr3)ipjr�)i�jr3)i�jr�)i�jr3)i�jr�)ikr3)iFkr�)iPkr3)iZkr�)i[kr3)ibkr�)ickr3)ixkr�)i}kr3)i�kr�)i@nrHu𖹠)iAnrHu𖹡)iBnrHu𖹢)iCnrHu𖹣)iDnrHu𖹤)iEnrHu𖹥)iFnrHu𖹦)iGnrHu𖹧)iHnrHu𖹨)iInrHu𖹩)iJnrHu𖹪)iKnrHu𖹫)iLnrHu𖹬)iMnrHu𖹭)iNnrHu𖹮)iOnrHu𖹯)iPnrHu𖹰)iQnrHu𖹱)iRnrHu𖹲)iSnrHu𖹳)iTnrHu𖹴)iUnrHu𖹵)iVnrHu𖹶)iWnrHu𖹷)iXnrHu𖹸)iYnrHu𖹹)iZnrHu𖹺)i[nrHu𖹻)i\nrHu𖹼)i]nrHu𖹽)i^nrHu𖹾)i_nrHu𖹿)i`nr3)i�nr�)ior3)iKor�)iOor3)i�or�)i�or3)i�or�)i�or3)i�or�)i�or3)i�or�)ipr3)i��r�)i�r3)i֌r�)i�r3)i	�r�)i�r3)i�r�)iP�r3)iS�r�)id�r3)ih�r�)ip�r3)i��r�)i�r3)ik�r�)ip�r3)i}�r�)i��r3)i��r�)i��r3)i��r�)i��r3)i��r�)i��r�)i�r3)i��r�)i�r3)i'�r�)i)�r3)i^�rHu𝅗𝅥)i_�rHu𝅘𝅥)i`�rHu𝅘𝅥𝅮)ia�rHu𝅘𝅥𝅯)ib�rHu𝅘𝅥𝅰)ic�rHu𝅘𝅥𝅱)id�rHu𝅘𝅥𝅲)ie�r3r�r�r�r�r��_seg_58�srcCsgd�S)N)d)is�r�)i{�r3)i��rHu𝆹𝅥)i��rHu𝆺𝅥)i��rHu𝆹𝅥𝅮)i��rHu𝆺𝅥𝅮)i��rHu𝆹𝅥𝅯)i��rHu𝆺𝅥𝅯)i��r3)i��r�)i�r3)iF�r�)i��r3)i��r�)i�r3)iW�r�)i`�r3)iy�r�)i�rHrI)i�rHrK)i�rHrM)i�rHrO)i�rHrQ)i�rHrS)i�rHrU)i�rHrW)i�rHrY)i	�rHr[)i
�rHr])i�rHr_)i�rHra)i
�rHrc)i�rHre)i�rHrg)i�rHri)i�rHrk)i�rHrm)i�rHro)i�rHrq)i�rHrs)i�rHru)i�rHrw)i�rHry)i�rHr{)i�rHrI)i�rHrK)i�rHrM)i�rHrO)i�rHrQ)i�rHrS)i �rHrU)i!�rHrW)i"�rHrY)i#�rHr[)i$�rHr])i%�rHr_)i&�rHra)i'�rHrc)i(�rHre)i)�rHrg)i*�rHri)i+�rHrk)i,�rHrm)i-�rHro)i.�rHrq)i/�rHrs)i0�rHru)i1�rHrw)i2�rHry)i3�rHr{)i4�rHrI)i5�rHrK)i6�rHrM)i7�rHrO)i8�rHrQ)i9�rHrS)i:�rHrU)i;�rHrW)i<�rHrY)i=�rHr[)i>�rHr])i?�rHr_)i@�rHra)iA�rHrc)iB�rHre)iC�rHrg)iD�rHri)iE�rHrk)iF�rHrm)iG�rHro)iH�rHrq)iI�rHrs)iJ�rHru)iK�rHrw)iL�rHry)iM�rHr{)iN�rHrI)iO�rHrK)iP�rHrM)iQ�rHrOr�r�r�r�r��_seg_59<srcCsgd�S)N)d)iR�rHrQ)iS�rHrS)iT�rHrU)iU�r�)iV�rHrY)iW�rHr[)iX�rHr])iY�rHr_)iZ�rHra)i[�rHrc)i\�rHre)i]�rHrg)i^�rHri)i_�rHrk)i`�rHrm)ia�rHro)ib�rHrq)ic�rHrs)id�rHru)ie�rHrw)if�rHry)ig�rHr{)ih�rHrI)ii�rHrK)ij�rHrM)ik�rHrO)il�rHrQ)im�rHrS)in�rHrU)io�rHrW)ip�rHrY)iq�rHr[)ir�rHr])is�rHr_)it�rHra)iu�rHrc)iv�rHre)iw�rHrg)ix�rHri)iy�rHrk)iz�rHrm)i{�rHro)i|�rHrq)i}�rHrs)i~�rHru)i�rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��r�)i��rHrM)i��rHrO)i��r�)i��rHrU)i��r�)i��rHr[)i��rHr])i��r�)i��rHrc)i��rHre)i��rHrg)i��rHri)i��r�)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrMr�r�r�r�r��_seg_60�srcCsgd�S)N)d)i��rHrO)i��r�)i��rHrS)i��r�)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��r�)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i�rHru)i�rHrw)i�rHry)i�rHr{)i�rHrI)i�rHrK)i�r�)i�rHrO)i�rHrQ)i	�rHrS)i
�rHrU)i�r�)i
�rHr[)i�rHr])i�rHr_)i�rHra)i�rHrc)i�rHre)i�rHrg)i�rHri)i�r�)i�rHrm)i�rHro)i�rHrq)i�rHrs)i�rHru)i�rHrw)i�rHry)i�r�r�r�r�r�r��_seg_61srcCsgd�S)N)d)i�rHrI)i�rHrK)i �rHrM)i!�rHrO)i"�rHrQ)i#�rHrS)i$�rHrU)i%�rHrW)i&�rHrY)i'�rHr[)i(�rHr])i)�rHr_)i*�rHra)i+�rHrc)i,�rHre)i-�rHrg)i.�rHri)i/�rHrk)i0�rHrm)i1�rHro)i2�rHrq)i3�rHrs)i4�rHru)i5�rHrw)i6�rHry)i7�rHr{)i8�rHrI)i9�rHrK)i:�r�)i;�rHrO)i<�rHrQ)i=�rHrS)i>�rHrU)i?�r�)i@�rHrY)iA�rHr[)iB�rHr])iC�rHr_)iD�rHra)iE�r�)iF�rHre)iG�r�)iJ�rHrm)iK�rHro)iL�rHrq)iM�rHrs)iN�rHru)iO�rHrw)iP�rHry)iQ�r�)iR�rHrI)iS�rHrK)iT�rHrM)iU�rHrO)iV�rHrQ)iW�rHrS)iX�rHrU)iY�rHrW)iZ�rHrY)i[�rHr[)i\�rHr])i]�rHr_)i^�rHra)i_�rHrc)i`�rHre)ia�rHrg)ib�rHri)ic�rHrk)id�rHrm)ie�rHro)if�rHrq)ig�rHrs)ih�rHru)ii�rHrw)ij�rHry)ik�rHr{)il�rHrI)im�rHrK)in�rHrM)io�rHrO)ip�rHrQ)iq�rHrS)ir�rHrU)is�rHrW)it�rHrY)iu�rHr[)iv�rHr])iw�rHr_)ix�rHra)iy�rHrc)iz�rHre)i{�rHrg)i|�rHri)i}�rHrk)i~�rHrm)i�rHro)i��rHrq)i��rHrs)i��rHru)i��rHrwr�r�r�r�r��_seg_62wsrcCsgd�S)N)d)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHror�r�r�r�r��_seg_63�srcCsgd�S)N)d)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i�rHrm)i�rHro)i�rHrq)i�rHrs)i�rHru)i�rHrw)i�rHry)i�rHr{)i�rHrI)i	�rHrK)i
�rHrM)i�rHrO)i�rHrQ)i
�rHrS)i�rHrU)i�rHrW)i�rHrY)i�rHr[)i�rHr])i�rHr_)i�rHra)i�rHrc)i�rHre)i�rHrg)i�rHri)i�rHrk)i�rHrm)i�rHro)i�rHrq)i�rHrs)i�rHru)i�rHrw)i �rHry)i!�rHr{)i"�rHrI)i#�rHrK)i$�rHrM)i%�rHrO)i&�rHrQ)i'�rHrS)i(�rHrU)i)�rHrW)i*�rHrY)i+�rHr[)i,�rHr])i-�rHr_)i.�rHra)i/�rHrc)i0�rHre)i1�rHrg)i2�rHri)i3�rHrk)i4�rHrm)i5�rHro)i6�rHrq)i7�rHrs)i8�rHru)i9�rHrw)i:�rHry)i;�rHr{)i<�rHrI)i=�rHrK)i>�rHrM)i?�rHrO)i@�rHrQ)iA�rHrS)iB�rHrU)iC�rHrW)iD�rHrY)iE�rHr[)iF�rHr])iG�rHr_)iH�rHra)iI�rHrc)iJ�rHre)iK�rHrgr�r�r�r�r��_seg_64IsrcCsgd�S)N)d)iL�rHri)iM�rHrk)iN�rHrm)iO�rHro)iP�rHrq)iQ�rHrs)iR�rHru)iS�rHrw)iT�rHry)iU�rHr{)iV�rHrI)iW�rHrK)iX�rHrM)iY�rHrO)iZ�rHrQ)i[�rHrS)i\�rHrU)i]�rHrW)i^�rHrY)i_�rHr[)i`�rHr])ia�rHr_)ib�rHra)ic�rHrc)id�rHre)ie�rHrg)if�rHri)ig�rHrk)ih�rHrm)ii�rHro)ij�rHrq)ik�rHrs)il�rHru)im�rHrw)in�rHry)io�rHr{)ip�rHrI)iq�rHrK)ir�rHrM)is�rHrO)it�rHrQ)iu�rHrS)iv�rHrU)iw�rHrW)ix�rHrY)iy�rHr[)iz�rHr])i{�rHr_)i|�rHra)i}�rHrc)i~�rHre)i�rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHrI)i��rHrK)i��rHrM)i��rHrO)i��rHrQ)i��rHrS)i��rHrU)i��rHrW)i��rHrY)i��rHr[)i��rHr])i��rHr_)i��rHra)i��rHrc)i��rHre)i��rHrg)i��rHri)i��rHrk)i��rHrm)i��rHro)i��rHrq)i��rHrs)i��rHru)i��rHrw)i��rHry)i��rHr{)i��rHuı)i��rHuȷ)i��r�)i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i��rHrz)i��rHrgr�r�r�r�r��_seg_65�srcCsgd�S)N)d)i��rHr{)i��rHr|)i��rHr�)i��rHr})i��rHr~)i��rHr)i��rHr�)i��rHr�)i��rHrz)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rH�∇)i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i��rHrz)i��rHrg)i��rHr{)i��rHr|)i��rHr�)i��rHr})i��rHr~)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rH�∂)i��rHrw)i��rHrz)i��rHr{)i��rHr�)i��rHr�)i��rHr�)i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i��rHrz)i��rHrg)i��rHr{)i��rHr|)i��rHr�)i��rHr})i��rHr~)i��rHr)i��rHr�)i��rHr�)i��rHrz)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr )i��rHrs)i��rHrt)i��rHru)i��rHrv)i�rHrw)i�rHrx)i�rHry)i�rHrz)i�rHrg)i�rHr{)i�rHr|)i�rHr�)i�rHr})i	�rHr~)i
�rHr)i�rHr�)i�rHr�)i
�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr!)i�rHrwr�r�r�r�r��_seg_66sr"cCsgd�S)N)d)i�rHrz)i�rHr{)i�rHr�)i�rHr�)i�rHr�)i�rHrs)i�rHrt)i�rHru)i�rHrv)i �rHrw)i!�rHrx)i"�rHry)i#�rHrz)i$�rHrg)i%�rHr{)i&�rHr|)i'�rHr�)i(�rHr})i)�rHr~)i*�rHr)i+�rHr�)i,�rHr�)i-�rHrz)i.�rHr�)i/�rHr�)i0�rHr�)i1�rHr�)i2�rHr�)i3�rHr�)i4�rHr�)i5�rHr )i6�rHrs)i7�rHrt)i8�rHru)i9�rHrv)i:�rHrw)i;�rHrx)i<�rHry)i=�rHrz)i>�rHrg)i?�rHr{)i@�rHr|)iA�rHr�)iB�rHr})iC�rHr~)iD�rHr)iE�rHr�)iF�rHr�)iG�rHr�)iI�rHr�)iJ�rHr�)iK�rHr�)iL�rHr�)iM�rHr�)iN�rHr�)iO�rHr!)iP�rHrw)iQ�rHrz)iR�rHr{)iS�rHr�)iT�rHr�)iU�rHr�)iV�rHrs)iW�rHrt)iX�rHru)iY�rHrv)iZ�rHrw)i[�rHrx)i\�rHry)i]�rHrz)i^�rHrg)i_�rHr{)i`�rHr|)ia�rHr�)ib�rHr})ic�rHr~)id�rHr)ie�rHr�)if�rHr�)ig�rHrz)ih�rHr�)ii�rHr�)ij�rHr�)ik�rHr�)il�rHr�)im�rHr�)in�rHr�)io�rHr )ip�rHrs)iq�rHrt)ir�rHru)is�rHrv)it�rHrw)iu�rHrx)iv�rHry)iw�rHrz)ix�rHrg)iy�rHr{)iz�rHr|)i{�rHr�r�r�r�r�r��_seg_67�sr#cCsgd�S)N)d)i|�rHr})i}�rHr~)i~�rHr)i�rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr!)i��rHrw)i��rHrz)i��rHr{)i��rHr�)i��rHr�)i��rHr�)i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i��rHrz)i��rHrg)i��rHr{)i��rHr|)i��rHr�)i��rHr})i��rHr~)i��rHr)i��rHr�)i��rHr�)i��rHrz)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr )i��rHrs)i��rHrt)i��rHru)i��rHrv)i��rHrw)i��rHrx)i��rHry)i��rHrz)i��rHrg)i��rHr{)i��rHr|)i��rHr�)i��rHr})i��rHr~)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr!)i��rHrw)i��rHrz)i��rHr{)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��r�)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�r�r�r�r�r��_seg_68�sr$cCsgd�S)N)d)i��rHr�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i�r3)i��r�)i��r3)i��r�)i��r3)i��r�)i�r3)i�r�)i�r3)i�r�)i�r3)i"�r�)i#�r3)i%�r�)i&�r3)i+�r�)i�r3)i-�r�)i0�r3)i>�r�)i@�r3)iJ�r�)iN�r3)iP�r�)i��r3)i��r�)i��r3)i�r�)i�r3)i��r�)i��r3)i��r�)i�rHu𞤢)i�rHu𞤣)i�rHu𞤤)i�rHu𞤥)i�rHu𞤦)i�rHu𞤧)i�rHu𞤨)i�rHu𞤩)i�rHu𞤪)i	�rHu𞤫)i
�rHu𞤬)i�rHu𞤭)i�rHu𞤮)i
�rHu𞤯)i�rHu𞤰)i�rHu𞤱)i�rHu𞤲)i�rHu𞤳)i�rHu𞤴)i�rHu𞤵)i�rHu𞤶)i�rHu𞤷)i�rHu𞤸)i�rHu𞤹)i�rHu𞤺)i�rHu𞤻)i�rHu𞤼)i�rHu𞤽)i�rHu𞤾)i�rHu𞤿)i�rHu𞥀)i�rHu𞥁)i �rHu𞥂)i!�rHu𞥃)i"�r3)iL�r�)iP�r3)iZ�r�)i^�r3)i`�r�r�r�r�r�r��_seg_69Vsr%cCsgd�S)N)d)iq�r3)i��r�)i�r3)i>�r�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�r�)i�rHr)i�rHr�)i�rHr�)i�rHr�)i	�rHr)i
�rHr�)i�rHr�)i�rHr)i
�rHr)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rHr�)i�rH�ٮ)i�rHrH)i�rH�ڡ)i�rH�ٯ)i �r�)i!�rHr�)i"�rHr�)i#�r�)i$�rHr)i%�r�)i'�rHr�)i(�r�)i)�rHr)i*�rHr�)i+�rHr�)i,�rHr)i-�rHr)i.�rHr�)i/�rHr�)i0�rHr�)i1�rHr�)i2�rHr�)i3�r�)i4�rHr�)i5�rHr�)i6�rHr�)i7�rHr�)i8�r�)i9�rHr�)i:�r�)i;�rHr�)i<�r�)iB�rHr�)iC�r�)iG�rHr�)iH�r�)iI�rHr)iJ�r�)iK�rHr�)iL�r�)iM�rHr)iN�rHr�)iO�rHr�)iP�r�)iQ�rHr�)iR�rHr�)iS�r�)iT�rHr�)iU�r�)iW�rHr�)iX�r�)iY�rHr�)iZ�r�)i[�rHr�)i\�r�)i]�rHrH)i^�r�)i_�rHr()i`�r�)ia�rHr�)ib�rHr�)ic�r�)id�rHr)ie�r�)ig�rHr�)ih�rHr�)ii�rHr)ij�rHr�r�r�r�r�r��_seg_70�sr)cCsgd�S)N)d)ik�r�)il�rHr)im�rHr)in�rHr�)io�rHr�)ip�rHr�)iq�rHr�)ir�rHr�)is�r�)it�rHr�)iu�rHr�)iv�rHr�)iw�rHr�)ix�r�)iy�rHr�)iz�rHr�)i{�rHr�)i|�rHr&)i}�r�)i~�rHr')i�r�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��r�)i��rHr�)i��rHr)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��r�)i��rHr�)i��rHr�)i��rHr�)i��r�)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr)i��r�)i��rHr�)i��rHr)i��rHr)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��r�)i��r3)i��r�)i�r3)i,�r�)i0�r3)i��r�)i��r3)i��r�)i��r3)i��r�)i��r3)i��r�)i��r3)i��r�)i�rz0,)i�rz1,)i�rz2,)i�rz3,)i�rz4,)i�rz5,)i�rz6,)i�rz7,r�r�r�r�r��_seg_71(sr*cCsgd�S)N)d)i	�rz8,)i
�rz9,)i�r3)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr�)i�rr)i �rr)i!�rr)i"�rr)i#�rr)i$�rr)i%�rr)i&�rr)i'�rr)i(�rr	)i)�rr
)i*�rHu〔s〕)i+�rHrM)i,�rHrk)i-�rHr�)i.�rHZwz)i/�r3)i0�rHrI)i1�rHrK)i2�rHrM)i3�rHrO)i4�rHrQ)i5�rHrS)i6�rHrU)i7�rHrW)i8�rHrY)i9�rHr[)i:�rHr])i;�rHr_)i<�rHra)i=�rHrc)i>�rHre)i?�rHrg)i@�rHri)iA�rHrk)iB�rHrm)iC�rHro)iD�rHrq)iE�rHrs)iF�rHru)iG�rHrw)iH�rHry)iI�rHr{)iJ�rHZhv)iK�rHr�)iL�rH�sd)iM�rHr,)iN�rHZppv)iO�rHZwc)iP�r3)ij�rHZmc)ik�rHZmd)il�rH�mr)im�r3)i��rHZdj)i��r3)i��r�)i��r3)i�rHuほか)i�rHuココ)i�rHr�)i�r�)i�rHr)i�rHu字)i�rHu双)i�rHuデ)i�rHr)i�rH�多)i�rHu解)i�rHrt)i�rHu交)i�rHu映)i�rHu無)i�rHr)i�rHu前)i�rHu後)i�rH�再)i�rHu新)i �rHu初)i!�rHu終)i"�rHr%)i#�rHu販r�r�r�r�r��_seg_72�sr/cCsgd�S)N)d)i$�rHu声)i%�rHu吹)i&�rHu演)i'�rHu投)i(�rHu捕)i)�rHr)i*�rHro)i+�rHu遊)i,�rHr{)i-�rHrr)i.�rHr|)i/�rHu指)i0�rHr.)i1�rHu打)i2�rHu禁)i3�rHu空)i4�rHu合)i5�rHu満)i6�rHrx)i7�rHr )i8�rHu申)i9�rH�割)i:�rHu営)i;�rHu配)i<�r�)i@�rHu	〔本〕)iA�rHu	〔三〕)iB�rHu	〔二〕)iC�rHu	〔安〕)iD�rHu	〔点〕)iE�rHu	〔打〕)iF�rHu	〔盗〕)iG�rHu	〔勝〕)iH�rHu	〔敗〕)iI�r�)iP�rHu得)iQ�rHu可)iR�r�)i`�r3)if�r�)i�r3)i�r�)i�r3)i�r�)i�r3)i��r�)i�r3)it�r�)i��r3)i�r�)i�r3)i�r�)i�r3)i�r�)i�r3)iH�r�)iP�r3)iZ�r�)i`�r3)i��r�)i��r3)i��r�)i��r3)i��r�)i�r3)iy�r�)iz�r3)i�r�)i�r3)iT�r�)i`�r3)in�r�)ip�r3)iu�r�)ix�r3)i{�r�)i��r3)i��r�)i��r3)i��r�)i��r3)i��r�)i��r3)i�r�)i�r3)i�r�)i�r3)i��r�)i��r3)i�r�)i�rHr�)i�rHr�)i�rHr�)i�rHr)i�rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�)i��rHr�r�r�r�r�r��_seg_73�sr1cCsgd�S)N)d)i��r�)ir3)iަr�)i�r3)i5�r�)i@�r3)i�r�)i �r3)i��r�)i��r3)i��r�)i�rHu丽)i�rHu丸)i�rHu乁)i�rHu𠄢)i�rHu你)i�rHr)i�rHu侻)i�rHu倂)i�rHu偺)i	�rHu備)i
�rHr)i�rHu像)i�rHu㒞)i
�rHu𠘺)i�rHr)i�rHu兔)i�rHu兤)i�rHu具)i�rHu𠔜)i�rHu㒹)i�rHu內)i�rHr.)i�rHu𠕋)i�rHu冗)i�rHu冤)i�rHu仌)i�rHu冬)i�rHr+)i�rHu𩇟)i�rHr)i�rHu刃)i�rHu㓟)i �rHu刻)i!�rHu剆)i"�rHr0)i#�rHu剷)i$�rHu㔕)i%�rHr,)i&�rHr)i'�rHr)i(�rHr-)i)�rHu包)i*�rHu匆)i+�rHr�)i,�rHu卉)i-�rHr)i.�rHu博)i/�rHu即)i0�rHu卽)i1�rHu卿)i4�rHu𠨬)i5�rHu灰)i6�rHu及)i7�rHu叟)i8�rHu𠭣)i9�rHu叫)i:�rHu叱)i;�rHu吆)i<�rHu咞)i=�rHu吸)i>�rHu呈)i?�rHu周)i@�rHu咢)iA�rHu哶)iB�rHu唐)iC�rHu啓)iD�rHu啣)iE�rHu善)iG�rHr.)iH�rHu喫)iI�rHu喳)iJ�rHu嗂)iK�rHu圖)iL�rHr)iM�rHu圗)iN�rHu噑)iO�rHu噴)iP�rHr)iQ�rHu壮)iR�rHu城)iS�rHu埴)iT�rHu堍)iU�rHu型)iV�rHu堲)iW�rHu報)iX�rHu墬)iY�rHu𡓤)iZ�rHu売)i[�rHu壷r�r�r�r�r��_seg_74csr2cCsgd�S)N)d)i\�rHu夆)i]�rHr-)i^�rHu夢)i_�rHu奢)i`�rHu𡚨)ia�rHu𡛪)ib�rHu姬)ic�rHu娛)id�rHu娧)ie�rHu姘)if�rHu婦)ig�rHu㛮)ih�r�)ii�rHu嬈)ij�rHu嬾)il�rHu𡧈)im�rHu寃)in�rHu寘)io�rHr�)ip�rHu寳)iq�rHu𡬘)ir�rHu寿)is�rHu将)it�r�)iu�rHr)iv�rHu㞁)iw�rHu屠)ix�rHr)iy�rHu峀)iz�rHu岍)i{�rHu𡷤)i|�rHu嵃)i}�rHu𡷦)i~�rHu嵮)i�rHu嵫)i��rHu嵼)i��rHu巡)i��rHu巢)i��rHu㠯)i��rHu巽)i��rHu帨)i��rHu帽)i��rHu幩)i��rHu㡢)i��rHu𢆃)i��rHu㡼)i��rHu庰)i��rHu庳)i��rHu庶)i��rHr�)i��rHu𪎒)i��rHr)i��rHu𢌱)i��rH�舁)i��rHu弢)i��rHu㣇)i��rHu𣊸)i��rHu𦇚)i��rHu形)i��rHu彫)i��rHu㣣)i��rHu徚)i��rHu忍)i��rHu志)i��rHu忹)i��rHu悁)i��rHu㤺)i��rHu㤜)i��rHr)i��rHu𢛔)i��rHu惇)i��rHu慈)i��rH�慌)i��rHr0)i��rHr4)i��rHu慺)i��rHr)i��rHu憲)i��rHu憤)i��rHu憯)i��rHu懞)i��rHr)i��rHr�)i��rHu成)i��rHu戛)i��rHu扝)i��rHu抱)i��rHu拔)i��rHu捐)i��rHu𢬌)i��rHu挽)i��rHu拼)i��rHu捨)i��rHu掃)i��rHu揤)i��rHu𢯱)i��rHu搢)i��rHu揅)i��rHu掩)i�rHu㨮r�r�r�r�r��_seg_75�sr5cCsgd�S)N)d)i�rHu摩)i�rHu摾)i�rHu撝)i�rHu摷)i�rHu㩬)i�rHr)i�rHu敬)i�rHu𣀊)i�rHu旣)i�rHu書)i�rHu晉)i�rHu㬙)i�rHr)i�rHu㬈)i�rHu㫤)i�rHu冒)i�rHu冕)i�rHu最)i�rHu暜)i�rHu肭)i�rHu䏙)i�rHr�)i�rHr1)i�rHu朡)i�rHu杞)i�rHu杓)i�rHu𣏃)i�rHu㭉)i�rHu柺)i�rHu枅)i�rHu桒)i�rHr)i�rHu𣑭)i�rHu梎)i�rHu栟)i�rHu椔)i�rHr?)i�rHu楂)i�rHu榣)i�rHu槪)i�rHu檨)i�rHu𣚣)i�rHu櫛)i�rHu㰘)i�rHu次)i�rHu𣢧)i�rHu歔)i�rHu㱎)i�rHu歲)i�rHu殟)i��rHr�)i��rHu殻)i��rHu𣪍)i��rHu𡴋)i��rHu𣫺)i��rHu汎)i��rHu𣲼)i��rHu沿)i��rHu泍)i��rHu汧)i��rHu洖)i�rHu派)i�rHr)i�rHr)i�rHu浩)i�rHu浸)i�rHu涅)i�rHu𣴞)i�rHu洴)i�rHu港)i	�rHu湮)i
�rHu㴳)i�rHr2)i�rHu滇)i
�rHu𣻑)i�rHu淹)i�rHu潮)i�rHu𣽞)i�rHu𣾎)i�rHu濆)i�rHu瀹)i�rHr3)i�rHu瀛)i�rHu㶖)i�rHu灊)i�rHu災)i�rHu灷)i�rHu炭)i�rHu𠔥)i�rHu煅)i�rHu𤉣)i�rHu熜)i�r�)i �rHu爨)i!�rHr4)i"�rHu牐)i#�rHu𤘈)i$�rHu犀)i%�rHu犕)i&�rHu𤜵r�r�r�r�r��_seg_765sr6cCsgd�S)N)d)i'�rHu𤠔)i(�rHu獺)i)�rHu王)i*�rHu㺬)i+�rHu玥)i,�rHu㺸)i.�rHu瑇)i/�rHu瑜)i0�rHr5)i1�rHu璅)i2�rHu瓊)i3�rHu㼛)i4�rHu甤)i5�rHu𤰶)i6�rHu甾)i7�rHu𤲒)i8�rHr�)i9�rHu𢆟)i:�rHu瘐)i;�rHu𤾡)i<�rHu𤾸)i=�rHu𥁄)i>�rHu㿼)i?�rHu䀈)i@�rHr6)iA�rHu𥃳)iB�rHu𥃲)iC�rHu𥄙)iD�rHu𥄳)iE�rHu眞)iF�rHu真)iH�rHr7)iI�rHr@)iJ�rHu瞋)iK�rHu䁆)iL�rHu䂖)iM�rHu𥐝)iN�rHu硎)iO�rHr�)iP�rHr8)iQ�rHu䃣)iR�rHu𥘦)iS�rHr)iT�rHu𥚚)iU�rHu𥛅)iV�rHr
)iW�rHu秫)iX�rHu䄯)iY�rHr )iZ�rHu穊)i[�rHu穏)i\�rHu𥥼)i]�rHu𥪧)i_�r�)i`�rHu䈂)ia�rHu𥮫)ib�rHu篆)ic�rHu築)id�rHu䈧)ie�rHu𥲀)if�rHu糒)ig�rHu䊠)ih�rHu糨)ii�rHu糣)ij�rHu紀)ik�rHu𥾆)il�rHu絣)im�rHu䌁)in�rHu緇)io�rHu縂)ip�rHu繅)iq�rHu䌴)ir�rHu𦈨)is�rHu𦉇)it�rHu䍙)iu�rHu𦋙)iv�rHu罺)iw�rHu𦌾)ix�rHu羕)iy�rHu翺)iz�rHr")i{�rHu𦓚)i|�rHu𦔣)i}�rHu聠)i~�rHu𦖨)i�rHu聰)i��rHu𣍟)i��rHu䏕)i��rHu育)i��rHu脃)i��rHu䐋)i��rHu脾)i��rHu媵)i��rHu𦞧)i��rHu𦞵)i��rHu𣎓)i��rHu𣎜)i��rHr3)i��rHu舄)i��rHu辞r�r�r�r�r��_seg_77�sr7cCsgd�S)N)d)i��rHu䑫)i��rHu芑)i��rHu芋)i��rHu芝)i��rHu劳)i��rHu花)i��rHu芳)i��rHu芽)i��rHu苦)i��rHu𦬼)i��rHr�)i��rHu茝)i��rHu荣)i��rHu莭)i��rHu茣)i��rHu莽)i��rHu菧)i��rHr#)i��rHu荓)i��rHu菊)i��rHu菌)i��rHu菜)i��rHu𦰶)i��rHu𦵫)i��rHu𦳕)i��rHu䔫)i��rHu蓱)i��rHu蓳)i��rHu蔖)i��rHu𧏊)i��rHu蕤)i��rHu𦼬)i��rHu䕝)i��rHu䕡)i��rHu𦾱)i��rHu𧃒)i��rHu䕫)i��rHu虐)i��rHr�)i��rHu虧)i��rHu虩)i��rHu蚩)i��rHu蚈)i��rHu蜎)i��rHu蛢)i��rHr9)i��rHu蜨)i��rHu蝫)i��rHu螆)i��r�)i��rHu蟡)i��rHu蠁)i�rHu䗹)i�rHu衠)i�rHr+)i�rHu𧙧)i�rHu裗)i�rHu裞)i�rHu䘵)i�rHu裺)i�rHu㒻)i�rHu𧢮)i�rHu𧥦)i�rHu䚾)i�rHu䛇)i�rHu誠)i�rHr:)i�rHr;)i�rHr-)i�rHu𧲨)i�rHu貫)i�rHu賁)i�rHu贛)i�rHu起)i�rHu𧼯)i�rHu𠠄)i�rHu跋)i�rHu趼)i�rHu跰)i�rHu𠣞)i�rHu軔)i�rHr<)i�rHu𨗒)i�rHu𨗭)i�rHu邔)i�rHu郱)i�rHu鄑)i�rHu𨜮)i�rHu鄛)i�rHu鈸)i�rHu鋗)i�rHu鋘)i�rHu鉼)i�rHu鏹)i�rHu鐕)i�rHu𨯺)i�rHu開)i�rHu䦕)i�rHu閷)i�rHu𨵷r�r�r�r�r��_seg_78 sr8cCsgd�S)N)0)i�rHu䧦)i�rHu雃)i�rHu嶲)i��rHu霣)i��rHu𩅅)i��rHu𩈚)i��rHu䩮)i��rHu䩶)i��rHu韠)i��rHu𩐊)i��rHu䪲)i��rHu𩒖)i��rHr=)i�rHu頩)i�rHu𩖶)i�rHu飢)i�rHu䬳)i�rHu餩)i�rHu馧)i�rHu駂)i�rHu駾)i�rHu䯎)i	�rHu𩬰)i
�rHr>)i�rHu鱀)i�rHu鳽)i
�rHu䳎)i�rHu䳭)i�rHu鵧)i�rHu𪃎)i�rHu䳸)i�rHu𪄅)i�rHu𪈎)i�rHu𪊑)i�rHr4)i�rHu䵖)i�rHr5)i�rHu黾)i�rHu鼅)i�rHu鼏)i�rHu鼖)i�rHr6)i�rHu𪘀)i�r�)ir3)iKr�)ir�)i�r�r�r�r�r�r��_seg_79p sr9N)W�typingrrr�__version__r�r�rOrWrbrhr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rrrrr'r:rnrur}r�r�r�r�r�r�r�r�rr	r/rGrJr�r�r�r�r�rrrrrrrrrrrrrrrrr"r#r$r%r)r*r/r1r2r5r6r7r8r9�tuple�	uts46datar�r�r�r��<module>s�iiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii5��������	�
���
������������������� �!�"�#�$�%�&�'�(�)�*�+�,�-�.�/�0�1�2�3�4�5�6�7�8�9�:�;�<�=�>�?�@�A�B�C�D�E�F�G�H�I�J�K�L�M�N�O��site-packages/pip/_vendor/idna/__pycache__/compat.cpython-39.pyc000064400000001323150564745010020555 0ustar00a

��Reh�@s<ddlTddlTddlmZmZdd�Zdd�Zdd	�Zd
S)�)�*�)�Any�UnioncCst|�S�N)�encode��label�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/compat.py�ToASCIIsrcCst|�Sr)�decoderr
r
r�	ToUnicode	srcCstd��dS)Nz,IDNA 2008 does not utilise nameprep protocol)�NotImplementedError)�sr
r
r�nameprep
srN)�core�codec�typingrrrrrr
r
r
r�<module>s
site-packages/pip/_vendor/idna/__pycache__/idnadata.cpython-39.pyc000064400000053241150564745010021045 0ustar00a

��Ren��@s:dZdddddd�Zddddddddddd	d	d	d	dd	dd	dddddd	d	d	d	dddddddddddddd
dddddddd	ddddd	d	d	dd	d	d	ddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	d	ddddddddddddddddddddddddddddddddddddddd	ddd	d	d	d	d	d	d	d	d	dd	dd	ddd	d	d	dd	d	dddddd	dddd	d	d	d	d	ddddd	dddddddddd	dd	dd	ddd	d	dddddddddddd	d	d	dddddddddddddddd	d	ddddd	dd	d	dddd	d	dddddddddddddddddddddddddddddddddddddddd
d	dddddd	d	dd	ddddddddddd	dd	d	d	dddddddd	dd	d	ddddddddddd	d	d	dd	ddd	d	dddddd	dddddddddddddddddd
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
ddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	dd	dd	d	dddd	d	d	d	d	ddddddddddd	dddd	ddd	ddddd	dd	dd	d	d	dddd	ddd	dd	d	dd	d	d	d	d	dddddddddddddddddddddddddddddddddddddd	ddddd	dddddddddddddddddddddd	ddddd	d	d	ddd	d	ddd	ddddd	d	dddddd	dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd
��Zdddd�ZdS)z13.0.0)$lt�
lx�
l~�
l��
l�l�l� l�0l�8l��l�l+�tlbtulk�ul��vl|l`|lF�|lN }�X@}lZd}l\l}l^t}l~|}l�~l��~l�l�Xl�tl��l��l'!�lf+�-l�l��lFRH)l�.:l�.l:l�/<l0@l0@l*0�@l<0�@��MP��8lnzdl�z�i��o�?��&�57�8]��N�`��k�:lz`�K)	l�D��@l��l7{tll={�ll?{�llB{mlE{mlP{m)��0Al�0tBl1@
�S1@E
lrH)	��0�Bl1�C�2�Gl�2@KlX3Llp�}l��}l0@
�h1�E
)�Greek�Han�Hebrew�Hiragana�Katakana�U�D�R�C�T�L(iiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJinioiqirisitiuiviwixiyizi{i|i}i~ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�iiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/iMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXi`iaibicidieifigihiiiji�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�iii
ii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixi�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i i
 i/ if ig ih ii i@�iA�iB�iC�iD�iE�iF�iG�iH�iI�iJ�iK�iL�iM�iN�iO�iP�iQ�iR�iS�iT�iU�iV�iW�iX�iY�iZ�i[�i\�i]�i^�i_�i`�ia�ib�ic�id�ie�if�ig�ih�ii�ij�ik�il�im�in�io�ip�iq�ir�is�i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�
i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i
i
i
i
i
i
i
i
i
i	
i

i
i
i

i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i
i 
i!
i"
i#
i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiQiRiSiTi�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i	�i
�i�i�i
�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i �i!�i"�i#�i$�i%�i&�i'�i(�i)�i*�i+�i,�i-�i.�i/�i0�i1�i2�i3�i4�i5�i6�i7�i8�i9�i:�i;�i<�i=�i>�i?�i@�iA�iB�iC�iK�(�l.�l:�l{�l�|l�lllll
$l,l4l<lDlLlTl\ldllltl |l"�l$�l&�l(�l*�l,�l.�l0�l2�l6�l9�l;�l=�l?�lClElGlI lL,lN4lP<lRDlTLlVTlX\lZdl\ll^tl`|lb�ld�lf�lh�lj�ll�ln�lp�lr�lt�lv�lx�l{�l}�l�l�l�l�l� l�0l�Hl�Tl�dl�xl��l��l��l��l��l��l��l��l��l��l��l�8l�@l�Hl�Pl�Xl�`l�hl�pl�|l��l��l��l��l��l��l��l��l��l��l��l��l�lllll
$l,l4l<lDlLlTl\ldllltl |l"�l$�l&�l(�l*�l,�l.�l0�l2�l:�l=�lA�lC	lH	lJ$	lL,	lN4	l�<	l��
l�l��l��l@lC
lO
lp@
lr�
lt�
lx�
l~�
l�@l��l�\l�dl�ll�tl�|l��l��l��l��l��l��l��l��l��l��l��l`�lb�ld�lf�lh�lj�ll�ln�lp�lr�lt�lv�lx�lz�l|�l~�l��l�l�l�,l�4l�<l�Dl�Ll�Tl�\l�dl�ll�tl�|l��l��l��l��l��l��l��l��l��l��l��l��l��l��l��l��l�l�l�l� l�(l�0l�8l�Dl�Ll�Tl�\l�dl�ll�tl�|l��l��l��l��l��l��l��l��l��l��l��l��l��l��l��l�lllll
$l,l4l<lDlLlTl\ldllltl |l"�l$�l&�l(�l*�l,�l.�l0�lZdl��l� l�Dl��l�l�l�rl��l@l@�l`lu�l��l�Tl�|l��l�lK@l�4l�l��l. l\!lk�!l��"l��"l�L#lX	�#ld	�%lp	�%l�	�%l�	&l�	<&l�	L&l�	�&l�	�&l�	�&l�	�&l�	'l�	,'l�	\'l�	�'l�	�'l�	�'l�	�'l
(l
(l
<(l)
L(l1
�(l3
�(l6
�(l:
�(l=
�(lC
�(lI
)lN
,)lR
D)l]
p)lv
�)l�
*l�
*l�
<*l�
L*l�
�*l�
�*l�
�*l�
�*l�
+l�
,+l�
@+l�
�+l�
�+l�+l,l
,l<,l)L,l1�,l4�,l:�,lE�,lI-lN,-lXT-ld|-lp�-lr�-l�.l�.l�8.l�H.l�d.l�p.l�x.l��.l��.l��.l��.l�/l�(/l�@/l�\/l��/l
0l80l)H0l:�0lE�0lI1lN(1lWT1l[`1ld�1lp�1l�2l�2l�82l�H2l��2l��2l��2l�3l�(3l�T3l�x3l��3l��3l��3l

4l
84lE
H4lI
5lO
(5lX
P5ld
|5lp
�5l�
�5l�
6l�
6l�
h6l�
�6l�
�6l�
7l�
(7l�
<7l�
X7l�
`7l�
�7l�
�7l38l;�8lO9lZ@9l�:l�:l�:l�0:l��:l��:l��:l�;l�;l� ;l�@;l�x;l<l,<l`<l*�<l6�<l8�<l:�<lC�<lH=lM$=lR8=lWL=l\`=lit=lm�=ls�=lu�=l��=l�>l�>l�P>l�d>l�x>l��>l��>l��>l��>l�?lJ@l�@Al�@Cl�ClIHlN(IlW@IlY`Il^hIl��Il�(Jl�@Jl��Jl��Jl�Kl�Kl� Kl`KlHLl[`Ll`tMl�Nl��NlmPl��Yl�Zl��Zl��[l
\l8\l5�\lT]lm�]lq�]lt�]l�^l��^l�\_l�p_l��_l@`ly�`l�bl��bldl,�dl<�dlnelu�el�fl��fl�@glhl_�hl}�il��il�@jl��jl��jl��jlLllZ@mlt�ml�nl8plJql~4ql�@sl�Psl,tl0�tl<�tlO8ulx�ul��ul�wl�wlxlxlxlxl
$xl,xl4xl<xlDxlLxlTxl\xldxllxltxl |xl"�xl$�xl&�xl(�xl*�xl,�xl.�xl0�xl2�xl4�xl6�xl8�xl:�xl<�xl>�xl@�xlBylDylFylHylJ$ylL,ylN4ylP<ylRDylTLylVTylX\ylZdyl\lyl^tyl`|ylb�yld�ylf�ylh�ylj�yll�yln�ylp�ylr�ylt�ylv�ylx�ylz�yl|�yl~�yl��yl�zl�zl�zl�zl�$zl�,zl�4zl�<zl�Dzl�Lzl�Tzl�pzl�|zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl��zl�{l�{l�{l�{l�${l�,{l�4{l�<{l�D{l�L{l�T{l�\{l�d{l�l{l�t{l�|{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l��{l�{l@|l(�|l8�|lF}rlh�}lq�}ls�}lu�}lw�}ly�}l{�}l}�}l��~l��~l�l�@l�Xl��l��l��lO!8l�!l_,�0lb,�1lg,�1li,�1lk,�1lm,�1lr,�1lu,�1l|,�1l�,2l�,2l�,2l�,2l�,$2l�,,2l�,42l�,<2l�,D2l�,L2l�,T2l�,\2l�,d2l�,l2l�,t2l�,|2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,�2l�,3l�,3l�,3l�,3l�,$3l�,,3l�,43l�,<3l�,D3l�,L3l�,T3l�,\3l�,d3l�,l3l�,t3l�,|3l�,�3l�,�3l�,�3l�,�3l�,�3l&-4l(-�4l.-�4lh-�4l�-�5l�-�6l�-�6l�-�6l�-�6l�-7l�- 7l�-@7l�-`7l.�7l0.�8l0@l.0�@l=0�@rl�0dBl�0tBrl�0�Cl01Dl�1�Frrrl�$l�$@l
&l,&@lB&lD&lF&lH&lJ&$lL&,lN&4lP&<lR&DlT&LlV&TlX&\lZ&dl\&ll^&tl`&|lb&�ld&�lf&�lh&�lj&�ll&�lp&�l~&�l�&�l�&l�&l�&l�&l�&$l�&,l�&4l�&<l�&Dl�&Ll�&Tl�&\l�&dl�&ll�&xl�&�l '\l$'�l&'�l('�l*'�l,'�l.'�l2'�l4'�l6'�l8'�l:'�l<'�l>'�l@'�lB'lD'lF'lH'lJ'$lL',lN'4lP'<lR'DlT'LlV'TlX'\lZ'dl\'ll^'tl`'|lb'�ld'�lf'�lh'�lj'�ll'�ln'�lp'�ly'�l{'�l}'�l�'�l�'l�'l�'l�'l�'0l�'8l�'Dl�'Ll�'\l�'dl�'ll�'tl�'|l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'�l�'l�' l�'(l�'�l((�l-(� lt(!l�("l�(@#l�(�#l�(�#l.)�#lT)�$l�)&l�)<'l�)�'l7*(lN*)lZ*@)lw*�)l�*�)l�*l+l�*�+l�*�+l+,l+$,l+D,l'+�,l/+�,l[+�,lj+�-l�+/l�+�/l�+�/l�W0lz8hlzDhlzLhl z|hl"z�hl%z�hl*z�hl{xll0~�xlt~�yll'4l;�l>�lN�l^@l�l��l�
l��
l��l lA�lJ
l{@
l�l��l� l��l��l�`l(ld�l7lVlh�l l	  l6( l9� l=� lV� lw�!l�"l��#l��#l	$l:	�$l�	&l�	�&l
(l
(l
0(l
T(l6
d(l;
�(l@
�(l}
�)l�
*l�
+l�
$+l6,lV-ls�-l�.lI0l�3l(
4l:
�4l�:l��:l��:l<l(�<lQ�<l��>l��?lG@lp�Al��Al�@Cl��Cl5Dl@�DlHElt@Elw�El�Fl�$Gl�8Gl�pGlHl8LHl?�Hl�Jl� Jl�(Jl�<Jl�|Jl��Jl��KlLl
Ll<Ll)LLl1�Ll4�Ll:�LlE�LlIMlN,MlQ@MlX\MldtMlm�Mlu�MlKPlZ@QlbxQl�Rl�Sl�@Sl�Vl��Vl�`WlAXlEYlZ@Yl�Zl�[l\l,t\l:�\l;`l�cl�cl
$dl0dlTdl6`dl9�dlD�dlZ@el��fl��fl�hgl��gl?hlHil�@il�tjl�kl	pl7(plA�plZ@ql��ql�Hrl��rltl
 tl7,tl;�tl>�tlH�tlZ@ulf�uli�ul��ul�@vl�Lvl��vl��{l��~l�#	lD%	l/4@	lGF
l9j l_j)ljj�)l�j@+l�j�+l7k,lDk-lZk@-lxk�-l�k�-l�n�9lKo<l�o<=l�o<>l�o�?l�o�?rl�@l� l	
4l1@
r
rl�2�E
lk<p
l}<�q
l�<r
l�<@r
l�<tr
l7ZhlmZ�hlvZ�il�Zjl�Zljl�Z�jl`l` l"`ll%`�l+`�l-al>a�lJalOa8l�bl�h l�h@#lLi�$lZi@%l�{�orrrrr	r
)l 0)l��lv�
l��lj�l��l�0�C)�PVALID�CONTEXTJ�CONTEXTON)�__version__�scripts�
joining_types�codepoint_classes�r#r#��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/idnadata.py�<module>s`&�[�����������������site-packages/pip/_vendor/idna/__pycache__/package_data.cpython-39.pyc000064400000000372150564745010021661 0ustar00a

��Re�@sdZdS)z3.2N)�__version__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/package_data.py�<module>�site-packages/pip/_vendor/idna/__pycache__/codec.cpython-39.pyc000064400000005545150564745010020361 0ustar00a

��Re}
�@s�ddlmZmZmZmZmZddlZddlZddlm	Z	m
Z
e�d�ZGdd�dej
�Z
Gdd	�d	ej�ZGd
d�dej�ZGdd
�d
e
ej�ZGdd�de
ej�Zdd�ZdS)�)�encode�decode�alabel�ulabel�	IDNAError�N)�Tuple�Optionalu[.。.。]c@s eZdZddd�Zddd�ZdS)	�Codec�strictcCs.|dkrtd�|���|sdSt|�t|�fS)Nr�Unsupported error handling "{}")�r)r�formatr�len��self�data�errors�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/codec.pyr
s
zCodec.encodecCs.|dkrtd�|���|sdSt|�t|�fS)Nrr��r)rrrrrrrrrs
zCodec.decodeN)r)r)�__name__�
__module__�__qualname__rrrrrrr
s

r
c@seZdZdd�ZdS)�IncrementalEncoderc
Cs�|dkrtd�|���|sdSt�|�}d}|rV|dsDd}|d=n|sV|d=|rVd}g}d}|D]*}|�t|��|r�|d7}|t|�7}qbd�|�|}	|t|�7}|	|fS�	Nrrrr����.rr)rr�_unicode_dots_re�split�appendrr�join�
rrr�final�labels�trailing_dot�result�size�labelZ
result_strrrr�_buffer_encodes0
z!IncrementalEncoder._buffer_encodeN)rrrr*rrrrrsrc@seZdZdd�ZdS)�IncrementalDecoderc
Cs�|dkrtd�|���|sdSt�|�}d}|rV|dsDd}|d=n|sV|d=|rVd}g}d}|D]*}|�t|��|r�|d7}|t|�7}qbd�|�|}	|t|�7}|	|fSr)rrrr r!rrr"r#rrr�_buffer_decodeAs0
z!IncrementalDecoder._buffer_decodeN)rrrr,rrrrr+@sr+c@seZdZdS)�StreamWriterN�rrrrrrrr-bsr-c@seZdZdS)�StreamReaderNr.rrrrr/fsr/c	Cs tjdt�jt�jttttd�S)N�idna)�namerr�incrementalencoder�incrementaldecoder�streamwriter�streamreader)	�codecs�	CodecInfor
rrrr+r-r/rrrr�getregentryjs�r8)�corerrrrrr6�re�typingrr	�compilerr
�BufferedIncrementalEncoderr�BufferedIncrementalDecoderr+r-r/r8rrrr�<module>s
""site-packages/pip/_vendor/idna/__pycache__/__init__.cpython-39.pyc000064400000001566150564745010021042 0ustar00a

��ReQ�@stddlmZddlmZmZmZmZmZmZm	Z	m
Z
mZmZm
Z
mZmZmZmZmZmZmZddlmZgd�ZdS)�)�__version__)�
IDNABidiError�	IDNAError�InvalidCodepoint�InvalidCodepointContext�alabel�
check_bidi�check_hyphen_ok�check_initial_combiner�check_label�	check_nfc�decode�encode�ulabel�uts46_remap�valid_contextj�valid_contexto�valid_label_length�valid_string_length)�intranges_contain)rrrrrrr	r
rrr
rrrrrrrrN)�package_datar�corerrrrrrr	r
rrr
rrrrrrr�	intrangesr�__all__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/__init__.py�<module>sPsite-packages/pip/_vendor/idna/__pycache__/core.cpython-39.pyc000064400000021747150564745010020236 0ustar00a

��Re2�@s0ddlmZddlZddlZddlZddlmZmZddlm	Z	dZ
dZe�d�Z
Gd	d
�d
e�ZGdd�de�ZGd
d�de�ZGdd�de�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd7dd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd8d(d)�Zd*d+�Zd,d-�Z d.d/�Z!d9d1d2�Z"d:d3d4�Z#d;d5d6�Z$dS)<�)�idnadata�N)�Union�Optional)�intranges_contain�	sxn--u[.。.。]c@seZdZdZdS)�	IDNAErrorz7 Base exception for all IDNA-encoding related problems N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/idna/core.pyrsrc@seZdZdZdS)�
IDNABidiErrorz= Exception when bidirectional requirements are not satisfied Nr	rrrrrsrc@seZdZdZdS)�InvalidCodepointz> Exception when a disallowed or unallocated codepoint is used Nr	rrrrrsrc@seZdZdZdS)�InvalidCodepointContextzE Exception when the codepoint is not valid in the context it is used Nr	rrrrrsrcCs0t�t|��}|dkr,t�t|��s,td��|S)Nrz Unknown character in unicodedata)�unicodedata�	combining�chr�name�
ValueError)�cp�vrrr�_combining_class s
rcCstt|�tj|�S)N)r�ordr�scripts)r�scriptrrr�
_is_script(srcCs
|�d�S)N�punycode)�encode��srrr�	_punycode,sr#cCs
d�|�S)NzU+{:04X})�formatr!rrr�_unot0sr%cCst|�dkrdSdS)N�?FT��len��labelrrr�valid_label_length5sr+cCst|�|rdndkrdSdS)N��FTr')r*�trailing_dotrrr�valid_string_length<sr/Fc	Cs^d}t|d�D]:\}}t�|�}|dkr<td�t|�|���|dvrd}q|sV|sVdSt�|d�}|dvrrd}n |d	kr�d}ntd
�t|����d}d}t|d�D]�\}}t�|�}|�r|dvr�td�|���|d
vr�d}n|dkr�d}|dv�rJ|�s|}n||k�rJtd��q�|dv�r.td�|���|dv�r>d}q�|dkr�d}q�|�sZtd��dS)NFr�z1Unknown directionality in label {} at position {})�R�AL�ANTr)r1r2�Lz=First codepoint in label {} must be directionality L, R or AL)
r1r2r3�EN�ES�CS�ET�ON�BN�NSMzGInvalid direction for codepoint at position {} in a right-to-left label)r1r2r5r3r;)r3r5z2Can not mix numeral types in a right-to-left label)r4r5r6r7r8r9r:r;zGInvalid direction for codepoint at position {} in a left-to-right label)r4r5z0Label ends with illegal codepoint directionality)�	enumerater�
bidirectionalrr$�repr)	r*�	check_ltr�
bidi_label�idxr�	direction�rtl�valid_ending�number_typerrr�
check_bidiCsP






rFcCs"t�|d�ddkrtd��dS)Nr�Mz0Label begins with an illegal combining characterT)r�categoryrr)rrr�check_initial_combinersrIcCs<|dd�dkrtd��|ddks0|ddkr8td��d	S)
N��z--z4Label has disallowed hyphens in 3rd and 4th positionr�-���z)Label must not start or end with a hyphenT)rr)rrr�check_hyphen_ok�s
rNcCst�d|�|krtd��dS)N�NFCz%Label must be in Normalization Form C)r�	normalizerr)rrr�	check_nfc�srQcCs4t||�}|dkr�|dkr8tt||d��tkr8dSd}t|ddd�D]B}tj�t||��}|td�krrqL|td�td	�fvrLd}q�qL|s�dSd}t|dt|��D]B}tj�t||��}|td�kr�q�|td
�td	�fvr�d}q�q�|S|dk�r,|dk�r(tt||d��tk�r(dSdSdSdS)Ni rrTFrM�Tr4�Dr1i
 )rr�_virama_combining_class�ranger�
joining_types�getr()r*�pos�cp_value�ok�i�joining_typerrr�valid_contextj�s<

r]cCs�t||�}|dkrdd|kr0t|�dkr`nn,t||d�dkr`t||d�dkr`dSdS|dkr�|t|�dkr�t|�dkr�t||dd�SdS|d	ks�|d
kr�|dkr�t||dd�SdS|dk�r|D]6}|d
kr�q�t|d��s
t|d��s
t|d�r�dSq�dSd|k�r.dk�rhnn6|D],}dt|�k�rVdk�r6nndS�q6dSd|k�r�dk�r�nn6|D],}dt|�k�r�dk�r�nndS�q�dSdS)N�rr�lTFiu�Greeki�i��Hebrewi�0u・�Hiragana�Katakana�Hani`iii�i�)rr(r)r*rX�	exceptionrYrrrr�valid_contexto�s@ (
" 
 
rfcCsJt|ttf�r|�d�}t|�dkr,td��t|�t|�t|�t	|�D]�\}}t
|�}t|tj
d�rpqLqLt|tj
d�r�z,t||�s�td�t|�|dt|����Wn0ty�td�t|�|dt|����Yn0qLt|tj
d	��rt||��s<td
�t|�|dt|����qLtd�t|�|dt|����qLt|�dS)Nzutf-8rzEmpty Label�PVALID�CONTEXTJz*Joiner {} not allowed at position {} in {}rz<Unknown codepoint adjacent to joiner {} at position {} in {}�CONTEXTO�-Codepoint {} not allowed at position {} in {}z-Codepoint {} at position {} of {} not allowed)�
isinstance�bytes�	bytearray�decoder(rrQrNrIr<rrr�codepoint_classesr]rr$r%r>rrfrrF)r*rXrrYrrr�check_label�s4

�
�  rpcCs|z(|�d�}t|�t|�s$td��|WSty:Yn0|sHtd��t|�}t|�t|�}t|}t|�sxtd��|S)N�asciizLabel too longzNo Input)	r �ulabelr+r�UnicodeEncodeError�strrpr#�_alabel_prefix�r*�label_bytesrrr�alabels"
rxcCs�t|ttf�s>z|�d�}WqBty:t|�|YS0n|}|��}|�t�r�|t	t�d�}|spt
d��|�d�ddkr�t
d��nt|�|�d�S|�d�}t|�|S)Nrqz5Malformed A-label, no Punycode eligible content foundrMrLz"A-label must not end with a hyphenr)rkrlrmr rsrp�lower�
startswithrur(rrnrvrrrrr%s&



rrTcCs$ddlm}d}t|�D]�\}}t|�}z�||dkr8|nt�||df�d}|d}	d}
t|�dkrn|d}
|	d	ks�|	d
kr�|r�|	dkr�|s�|
dur�||7}n@|
dur�|	dks�|	dkr�|r�|	d
kr�|r�||
7}n|	d
kr�t��Wqt�ytd�t	|�|dt
|����Yq0qt�d|�S)zBRe-map the characters in the string according to UTS46 processing.r)�	uts46datar0��ZN�rJ�VrS�3rG�IrjrO)
r{r<r�bisect�bisect_leftr(�
IndexErrorrr$r%r>rrP)�domain�
std3_rules�transitionalr{�outputrX�char�
code_point�uts46row�status�replacementrrr�uts46_remap@sV������
����

��r�c	Cs�t|ttf�r|�d�}|r(t|||�}d}g}|r@|�d�}n
t�|�}|rX|dgkr`td��|ddkrv|d=d}|D]$}t|�}|r�|�	|�qztd��qz|r�|�	d	�d
�
|�}t||�s�td��|S)NrqF�.r0�Empty domainrMT�Empty label��.zDomain too long)rkrlrmrnr��split�_unicode_dots_rerrx�append�joinr/)	r"�strict�uts46r�r�r.�result�labelsr*rrrr as2





r cCs�t|ttf�r|�d�}|r(t||d�}d}g}|s@t�|�}n
|�d�}|rX|dgkr`td��|dsr|d=d}|D]$}t|�}|r�|�	|�qvtd��qv|r�|�	d�d�
|�S)	NrqFr�r0r�rMTr�)rkrlrmrnr�r�r�rrrr�r�)r"r�r�r�r.r�r�r*rrrrn�s,



rn)F)F)TF)FFFF)FFF)%r0rr�r�re�typingrr�	intrangesrrTru�compiler��UnicodeErrorrrrrrrr#r%r+r/rFrIrNrQr]rfrprxrrr�r rnrrrr�<module>s<

<	,
+ 
!
site-packages/pip/_vendor/idna/uts46data.py000064400000612171150564745010014633 0ustar00# This file is automatically generated by tools/idna-data

from typing import List, Tuple, Union

"""IDNA Mapping Table from UTS46."""


__version__ = '13.0.0'
def _seg_0():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x0, '3'),
    (0x1, '3'),
    (0x2, '3'),
    (0x3, '3'),
    (0x4, '3'),
    (0x5, '3'),
    (0x6, '3'),
    (0x7, '3'),
    (0x8, '3'),
    (0x9, '3'),
    (0xA, '3'),
    (0xB, '3'),
    (0xC, '3'),
    (0xD, '3'),
    (0xE, '3'),
    (0xF, '3'),
    (0x10, '3'),
    (0x11, '3'),
    (0x12, '3'),
    (0x13, '3'),
    (0x14, '3'),
    (0x15, '3'),
    (0x16, '3'),
    (0x17, '3'),
    (0x18, '3'),
    (0x19, '3'),
    (0x1A, '3'),
    (0x1B, '3'),
    (0x1C, '3'),
    (0x1D, '3'),
    (0x1E, '3'),
    (0x1F, '3'),
    (0x20, '3'),
    (0x21, '3'),
    (0x22, '3'),
    (0x23, '3'),
    (0x24, '3'),
    (0x25, '3'),
    (0x26, '3'),
    (0x27, '3'),
    (0x28, '3'),
    (0x29, '3'),
    (0x2A, '3'),
    (0x2B, '3'),
    (0x2C, '3'),
    (0x2D, 'V'),
    (0x2E, 'V'),
    (0x2F, '3'),
    (0x30, 'V'),
    (0x31, 'V'),
    (0x32, 'V'),
    (0x33, 'V'),
    (0x34, 'V'),
    (0x35, 'V'),
    (0x36, 'V'),
    (0x37, 'V'),
    (0x38, 'V'),
    (0x39, 'V'),
    (0x3A, '3'),
    (0x3B, '3'),
    (0x3C, '3'),
    (0x3D, '3'),
    (0x3E, '3'),
    (0x3F, '3'),
    (0x40, '3'),
    (0x41, 'M', 'a'),
    (0x42, 'M', 'b'),
    (0x43, 'M', 'c'),
    (0x44, 'M', 'd'),
    (0x45, 'M', 'e'),
    (0x46, 'M', 'f'),
    (0x47, 'M', 'g'),
    (0x48, 'M', 'h'),
    (0x49, 'M', 'i'),
    (0x4A, 'M', 'j'),
    (0x4B, 'M', 'k'),
    (0x4C, 'M', 'l'),
    (0x4D, 'M', 'm'),
    (0x4E, 'M', 'n'),
    (0x4F, 'M', 'o'),
    (0x50, 'M', 'p'),
    (0x51, 'M', 'q'),
    (0x52, 'M', 'r'),
    (0x53, 'M', 's'),
    (0x54, 'M', 't'),
    (0x55, 'M', 'u'),
    (0x56, 'M', 'v'),
    (0x57, 'M', 'w'),
    (0x58, 'M', 'x'),
    (0x59, 'M', 'y'),
    (0x5A, 'M', 'z'),
    (0x5B, '3'),
    (0x5C, '3'),
    (0x5D, '3'),
    (0x5E, '3'),
    (0x5F, '3'),
    (0x60, '3'),
    (0x61, 'V'),
    (0x62, 'V'),
    (0x63, 'V'),
    ]

def _seg_1():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x64, 'V'),
    (0x65, 'V'),
    (0x66, 'V'),
    (0x67, 'V'),
    (0x68, 'V'),
    (0x69, 'V'),
    (0x6A, 'V'),
    (0x6B, 'V'),
    (0x6C, 'V'),
    (0x6D, 'V'),
    (0x6E, 'V'),
    (0x6F, 'V'),
    (0x70, 'V'),
    (0x71, 'V'),
    (0x72, 'V'),
    (0x73, 'V'),
    (0x74, 'V'),
    (0x75, 'V'),
    (0x76, 'V'),
    (0x77, 'V'),
    (0x78, 'V'),
    (0x79, 'V'),
    (0x7A, 'V'),
    (0x7B, '3'),
    (0x7C, '3'),
    (0x7D, '3'),
    (0x7E, '3'),
    (0x7F, '3'),
    (0x80, 'X'),
    (0x81, 'X'),
    (0x82, 'X'),
    (0x83, 'X'),
    (0x84, 'X'),
    (0x85, 'X'),
    (0x86, 'X'),
    (0x87, 'X'),
    (0x88, 'X'),
    (0x89, 'X'),
    (0x8A, 'X'),
    (0x8B, 'X'),
    (0x8C, 'X'),
    (0x8D, 'X'),
    (0x8E, 'X'),
    (0x8F, 'X'),
    (0x90, 'X'),
    (0x91, 'X'),
    (0x92, 'X'),
    (0x93, 'X'),
    (0x94, 'X'),
    (0x95, 'X'),
    (0x96, 'X'),
    (0x97, 'X'),
    (0x98, 'X'),
    (0x99, 'X'),
    (0x9A, 'X'),
    (0x9B, 'X'),
    (0x9C, 'X'),
    (0x9D, 'X'),
    (0x9E, 'X'),
    (0x9F, 'X'),
    (0xA0, '3', ' '),
    (0xA1, 'V'),
    (0xA2, 'V'),
    (0xA3, 'V'),
    (0xA4, 'V'),
    (0xA5, 'V'),
    (0xA6, 'V'),
    (0xA7, 'V'),
    (0xA8, '3', ' ̈'),
    (0xA9, 'V'),
    (0xAA, 'M', 'a'),
    (0xAB, 'V'),
    (0xAC, 'V'),
    (0xAD, 'I'),
    (0xAE, 'V'),
    (0xAF, '3', ' ̄'),
    (0xB0, 'V'),
    (0xB1, 'V'),
    (0xB2, 'M', '2'),
    (0xB3, 'M', '3'),
    (0xB4, '3', ' ́'),
    (0xB5, 'M', 'μ'),
    (0xB6, 'V'),
    (0xB7, 'V'),
    (0xB8, '3', ' ̧'),
    (0xB9, 'M', '1'),
    (0xBA, 'M', 'o'),
    (0xBB, 'V'),
    (0xBC, 'M', '1⁄4'),
    (0xBD, 'M', '1⁄2'),
    (0xBE, 'M', '3⁄4'),
    (0xBF, 'V'),
    (0xC0, 'M', 'à'),
    (0xC1, 'M', 'á'),
    (0xC2, 'M', 'â'),
    (0xC3, 'M', 'ã'),
    (0xC4, 'M', 'ä'),
    (0xC5, 'M', 'å'),
    (0xC6, 'M', 'æ'),
    (0xC7, 'M', 'ç'),
    ]

def _seg_2():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xC8, 'M', 'è'),
    (0xC9, 'M', 'é'),
    (0xCA, 'M', 'ê'),
    (0xCB, 'M', 'ë'),
    (0xCC, 'M', 'ì'),
    (0xCD, 'M', 'í'),
    (0xCE, 'M', 'î'),
    (0xCF, 'M', 'ï'),
    (0xD0, 'M', 'ð'),
    (0xD1, 'M', 'ñ'),
    (0xD2, 'M', 'ò'),
    (0xD3, 'M', 'ó'),
    (0xD4, 'M', 'ô'),
    (0xD5, 'M', 'õ'),
    (0xD6, 'M', 'ö'),
    (0xD7, 'V'),
    (0xD8, 'M', 'ø'),
    (0xD9, 'M', 'ù'),
    (0xDA, 'M', 'ú'),
    (0xDB, 'M', 'û'),
    (0xDC, 'M', 'ü'),
    (0xDD, 'M', 'ý'),
    (0xDE, 'M', 'þ'),
    (0xDF, 'D', 'ss'),
    (0xE0, 'V'),
    (0xE1, 'V'),
    (0xE2, 'V'),
    (0xE3, 'V'),
    (0xE4, 'V'),
    (0xE5, 'V'),
    (0xE6, 'V'),
    (0xE7, 'V'),
    (0xE8, 'V'),
    (0xE9, 'V'),
    (0xEA, 'V'),
    (0xEB, 'V'),
    (0xEC, 'V'),
    (0xED, 'V'),
    (0xEE, 'V'),
    (0xEF, 'V'),
    (0xF0, 'V'),
    (0xF1, 'V'),
    (0xF2, 'V'),
    (0xF3, 'V'),
    (0xF4, 'V'),
    (0xF5, 'V'),
    (0xF6, 'V'),
    (0xF7, 'V'),
    (0xF8, 'V'),
    (0xF9, 'V'),
    (0xFA, 'V'),
    (0xFB, 'V'),
    (0xFC, 'V'),
    (0xFD, 'V'),
    (0xFE, 'V'),
    (0xFF, 'V'),
    (0x100, 'M', 'ā'),
    (0x101, 'V'),
    (0x102, 'M', 'ă'),
    (0x103, 'V'),
    (0x104, 'M', 'ą'),
    (0x105, 'V'),
    (0x106, 'M', 'ć'),
    (0x107, 'V'),
    (0x108, 'M', 'ĉ'),
    (0x109, 'V'),
    (0x10A, 'M', 'ċ'),
    (0x10B, 'V'),
    (0x10C, 'M', 'č'),
    (0x10D, 'V'),
    (0x10E, 'M', 'ď'),
    (0x10F, 'V'),
    (0x110, 'M', 'đ'),
    (0x111, 'V'),
    (0x112, 'M', 'ē'),
    (0x113, 'V'),
    (0x114, 'M', 'ĕ'),
    (0x115, 'V'),
    (0x116, 'M', 'ė'),
    (0x117, 'V'),
    (0x118, 'M', 'ę'),
    (0x119, 'V'),
    (0x11A, 'M', 'ě'),
    (0x11B, 'V'),
    (0x11C, 'M', 'ĝ'),
    (0x11D, 'V'),
    (0x11E, 'M', 'ğ'),
    (0x11F, 'V'),
    (0x120, 'M', 'ġ'),
    (0x121, 'V'),
    (0x122, 'M', 'ģ'),
    (0x123, 'V'),
    (0x124, 'M', 'ĥ'),
    (0x125, 'V'),
    (0x126, 'M', 'ħ'),
    (0x127, 'V'),
    (0x128, 'M', 'ĩ'),
    (0x129, 'V'),
    (0x12A, 'M', 'ī'),
    (0x12B, 'V'),
    ]

def _seg_3():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x12C, 'M', 'ĭ'),
    (0x12D, 'V'),
    (0x12E, 'M', 'į'),
    (0x12F, 'V'),
    (0x130, 'M', 'i̇'),
    (0x131, 'V'),
    (0x132, 'M', 'ij'),
    (0x134, 'M', 'ĵ'),
    (0x135, 'V'),
    (0x136, 'M', 'ķ'),
    (0x137, 'V'),
    (0x139, 'M', 'ĺ'),
    (0x13A, 'V'),
    (0x13B, 'M', 'ļ'),
    (0x13C, 'V'),
    (0x13D, 'M', 'ľ'),
    (0x13E, 'V'),
    (0x13F, 'M', 'l·'),
    (0x141, 'M', 'ł'),
    (0x142, 'V'),
    (0x143, 'M', 'ń'),
    (0x144, 'V'),
    (0x145, 'M', 'ņ'),
    (0x146, 'V'),
    (0x147, 'M', 'ň'),
    (0x148, 'V'),
    (0x149, 'M', 'ʼn'),
    (0x14A, 'M', 'ŋ'),
    (0x14B, 'V'),
    (0x14C, 'M', 'ō'),
    (0x14D, 'V'),
    (0x14E, 'M', 'ŏ'),
    (0x14F, 'V'),
    (0x150, 'M', 'ő'),
    (0x151, 'V'),
    (0x152, 'M', 'œ'),
    (0x153, 'V'),
    (0x154, 'M', 'ŕ'),
    (0x155, 'V'),
    (0x156, 'M', 'ŗ'),
    (0x157, 'V'),
    (0x158, 'M', 'ř'),
    (0x159, 'V'),
    (0x15A, 'M', 'ś'),
    (0x15B, 'V'),
    (0x15C, 'M', 'ŝ'),
    (0x15D, 'V'),
    (0x15E, 'M', 'ş'),
    (0x15F, 'V'),
    (0x160, 'M', 'š'),
    (0x161, 'V'),
    (0x162, 'M', 'ţ'),
    (0x163, 'V'),
    (0x164, 'M', 'ť'),
    (0x165, 'V'),
    (0x166, 'M', 'ŧ'),
    (0x167, 'V'),
    (0x168, 'M', 'ũ'),
    (0x169, 'V'),
    (0x16A, 'M', 'ū'),
    (0x16B, 'V'),
    (0x16C, 'M', 'ŭ'),
    (0x16D, 'V'),
    (0x16E, 'M', 'ů'),
    (0x16F, 'V'),
    (0x170, 'M', 'ű'),
    (0x171, 'V'),
    (0x172, 'M', 'ų'),
    (0x173, 'V'),
    (0x174, 'M', 'ŵ'),
    (0x175, 'V'),
    (0x176, 'M', 'ŷ'),
    (0x177, 'V'),
    (0x178, 'M', 'ÿ'),
    (0x179, 'M', 'ź'),
    (0x17A, 'V'),
    (0x17B, 'M', 'ż'),
    (0x17C, 'V'),
    (0x17D, 'M', 'ž'),
    (0x17E, 'V'),
    (0x17F, 'M', 's'),
    (0x180, 'V'),
    (0x181, 'M', 'ɓ'),
    (0x182, 'M', 'ƃ'),
    (0x183, 'V'),
    (0x184, 'M', 'ƅ'),
    (0x185, 'V'),
    (0x186, 'M', 'ɔ'),
    (0x187, 'M', 'ƈ'),
    (0x188, 'V'),
    (0x189, 'M', 'ɖ'),
    (0x18A, 'M', 'ɗ'),
    (0x18B, 'M', 'ƌ'),
    (0x18C, 'V'),
    (0x18E, 'M', 'ǝ'),
    (0x18F, 'M', 'ə'),
    (0x190, 'M', 'ɛ'),
    (0x191, 'M', 'ƒ'),
    (0x192, 'V'),
    (0x193, 'M', 'ɠ'),
    ]

def _seg_4():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x194, 'M', 'ɣ'),
    (0x195, 'V'),
    (0x196, 'M', 'ɩ'),
    (0x197, 'M', 'ɨ'),
    (0x198, 'M', 'ƙ'),
    (0x199, 'V'),
    (0x19C, 'M', 'ɯ'),
    (0x19D, 'M', 'ɲ'),
    (0x19E, 'V'),
    (0x19F, 'M', 'ɵ'),
    (0x1A0, 'M', 'ơ'),
    (0x1A1, 'V'),
    (0x1A2, 'M', 'ƣ'),
    (0x1A3, 'V'),
    (0x1A4, 'M', 'ƥ'),
    (0x1A5, 'V'),
    (0x1A6, 'M', 'ʀ'),
    (0x1A7, 'M', 'ƨ'),
    (0x1A8, 'V'),
    (0x1A9, 'M', 'ʃ'),
    (0x1AA, 'V'),
    (0x1AC, 'M', 'ƭ'),
    (0x1AD, 'V'),
    (0x1AE, 'M', 'ʈ'),
    (0x1AF, 'M', 'ư'),
    (0x1B0, 'V'),
    (0x1B1, 'M', 'ʊ'),
    (0x1B2, 'M', 'ʋ'),
    (0x1B3, 'M', 'ƴ'),
    (0x1B4, 'V'),
    (0x1B5, 'M', 'ƶ'),
    (0x1B6, 'V'),
    (0x1B7, 'M', 'ʒ'),
    (0x1B8, 'M', 'ƹ'),
    (0x1B9, 'V'),
    (0x1BC, 'M', 'ƽ'),
    (0x1BD, 'V'),
    (0x1C4, 'M', 'dž'),
    (0x1C7, 'M', 'lj'),
    (0x1CA, 'M', 'nj'),
    (0x1CD, 'M', 'ǎ'),
    (0x1CE, 'V'),
    (0x1CF, 'M', 'ǐ'),
    (0x1D0, 'V'),
    (0x1D1, 'M', 'ǒ'),
    (0x1D2, 'V'),
    (0x1D3, 'M', 'ǔ'),
    (0x1D4, 'V'),
    (0x1D5, 'M', 'ǖ'),
    (0x1D6, 'V'),
    (0x1D7, 'M', 'ǘ'),
    (0x1D8, 'V'),
    (0x1D9, 'M', 'ǚ'),
    (0x1DA, 'V'),
    (0x1DB, 'M', 'ǜ'),
    (0x1DC, 'V'),
    (0x1DE, 'M', 'ǟ'),
    (0x1DF, 'V'),
    (0x1E0, 'M', 'ǡ'),
    (0x1E1, 'V'),
    (0x1E2, 'M', 'ǣ'),
    (0x1E3, 'V'),
    (0x1E4, 'M', 'ǥ'),
    (0x1E5, 'V'),
    (0x1E6, 'M', 'ǧ'),
    (0x1E7, 'V'),
    (0x1E8, 'M', 'ǩ'),
    (0x1E9, 'V'),
    (0x1EA, 'M', 'ǫ'),
    (0x1EB, 'V'),
    (0x1EC, 'M', 'ǭ'),
    (0x1ED, 'V'),
    (0x1EE, 'M', 'ǯ'),
    (0x1EF, 'V'),
    (0x1F1, 'M', 'dz'),
    (0x1F4, 'M', 'ǵ'),
    (0x1F5, 'V'),
    (0x1F6, 'M', 'ƕ'),
    (0x1F7, 'M', 'ƿ'),
    (0x1F8, 'M', 'ǹ'),
    (0x1F9, 'V'),
    (0x1FA, 'M', 'ǻ'),
    (0x1FB, 'V'),
    (0x1FC, 'M', 'ǽ'),
    (0x1FD, 'V'),
    (0x1FE, 'M', 'ǿ'),
    (0x1FF, 'V'),
    (0x200, 'M', 'ȁ'),
    (0x201, 'V'),
    (0x202, 'M', 'ȃ'),
    (0x203, 'V'),
    (0x204, 'M', 'ȅ'),
    (0x205, 'V'),
    (0x206, 'M', 'ȇ'),
    (0x207, 'V'),
    (0x208, 'M', 'ȉ'),
    (0x209, 'V'),
    (0x20A, 'M', 'ȋ'),
    (0x20B, 'V'),
    (0x20C, 'M', 'ȍ'),
    ]

def _seg_5():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x20D, 'V'),
    (0x20E, 'M', 'ȏ'),
    (0x20F, 'V'),
    (0x210, 'M', 'ȑ'),
    (0x211, 'V'),
    (0x212, 'M', 'ȓ'),
    (0x213, 'V'),
    (0x214, 'M', 'ȕ'),
    (0x215, 'V'),
    (0x216, 'M', 'ȗ'),
    (0x217, 'V'),
    (0x218, 'M', 'ș'),
    (0x219, 'V'),
    (0x21A, 'M', 'ț'),
    (0x21B, 'V'),
    (0x21C, 'M', 'ȝ'),
    (0x21D, 'V'),
    (0x21E, 'M', 'ȟ'),
    (0x21F, 'V'),
    (0x220, 'M', 'ƞ'),
    (0x221, 'V'),
    (0x222, 'M', 'ȣ'),
    (0x223, 'V'),
    (0x224, 'M', 'ȥ'),
    (0x225, 'V'),
    (0x226, 'M', 'ȧ'),
    (0x227, 'V'),
    (0x228, 'M', 'ȩ'),
    (0x229, 'V'),
    (0x22A, 'M', 'ȫ'),
    (0x22B, 'V'),
    (0x22C, 'M', 'ȭ'),
    (0x22D, 'V'),
    (0x22E, 'M', 'ȯ'),
    (0x22F, 'V'),
    (0x230, 'M', 'ȱ'),
    (0x231, 'V'),
    (0x232, 'M', 'ȳ'),
    (0x233, 'V'),
    (0x23A, 'M', 'ⱥ'),
    (0x23B, 'M', 'ȼ'),
    (0x23C, 'V'),
    (0x23D, 'M', 'ƚ'),
    (0x23E, 'M', 'ⱦ'),
    (0x23F, 'V'),
    (0x241, 'M', 'ɂ'),
    (0x242, 'V'),
    (0x243, 'M', 'ƀ'),
    (0x244, 'M', 'ʉ'),
    (0x245, 'M', 'ʌ'),
    (0x246, 'M', 'ɇ'),
    (0x247, 'V'),
    (0x248, 'M', 'ɉ'),
    (0x249, 'V'),
    (0x24A, 'M', 'ɋ'),
    (0x24B, 'V'),
    (0x24C, 'M', 'ɍ'),
    (0x24D, 'V'),
    (0x24E, 'M', 'ɏ'),
    (0x24F, 'V'),
    (0x2B0, 'M', 'h'),
    (0x2B1, 'M', 'ɦ'),
    (0x2B2, 'M', 'j'),
    (0x2B3, 'M', 'r'),
    (0x2B4, 'M', 'ɹ'),
    (0x2B5, 'M', 'ɻ'),
    (0x2B6, 'M', 'ʁ'),
    (0x2B7, 'M', 'w'),
    (0x2B8, 'M', 'y'),
    (0x2B9, 'V'),
    (0x2D8, '3', ' ̆'),
    (0x2D9, '3', ' ̇'),
    (0x2DA, '3', ' ̊'),
    (0x2DB, '3', ' ̨'),
    (0x2DC, '3', ' ̃'),
    (0x2DD, '3', ' ̋'),
    (0x2DE, 'V'),
    (0x2E0, 'M', 'ɣ'),
    (0x2E1, 'M', 'l'),
    (0x2E2, 'M', 's'),
    (0x2E3, 'M', 'x'),
    (0x2E4, 'M', 'ʕ'),
    (0x2E5, 'V'),
    (0x340, 'M', '̀'),
    (0x341, 'M', '́'),
    (0x342, 'V'),
    (0x343, 'M', '̓'),
    (0x344, 'M', '̈́'),
    (0x345, 'M', 'ι'),
    (0x346, 'V'),
    (0x34F, 'I'),
    (0x350, 'V'),
    (0x370, 'M', 'ͱ'),
    (0x371, 'V'),
    (0x372, 'M', 'ͳ'),
    (0x373, 'V'),
    (0x374, 'M', 'ʹ'),
    (0x375, 'V'),
    (0x376, 'M', 'ͷ'),
    (0x377, 'V'),
    ]

def _seg_6():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x378, 'X'),
    (0x37A, '3', ' ι'),
    (0x37B, 'V'),
    (0x37E, '3', ';'),
    (0x37F, 'M', 'ϳ'),
    (0x380, 'X'),
    (0x384, '3', ' ́'),
    (0x385, '3', ' ̈́'),
    (0x386, 'M', 'ά'),
    (0x387, 'M', '·'),
    (0x388, 'M', 'έ'),
    (0x389, 'M', 'ή'),
    (0x38A, 'M', 'ί'),
    (0x38B, 'X'),
    (0x38C, 'M', 'ό'),
    (0x38D, 'X'),
    (0x38E, 'M', 'ύ'),
    (0x38F, 'M', 'ώ'),
    (0x390, 'V'),
    (0x391, 'M', 'α'),
    (0x392, 'M', 'β'),
    (0x393, 'M', 'γ'),
    (0x394, 'M', 'δ'),
    (0x395, 'M', 'ε'),
    (0x396, 'M', 'ζ'),
    (0x397, 'M', 'η'),
    (0x398, 'M', 'θ'),
    (0x399, 'M', 'ι'),
    (0x39A, 'M', 'κ'),
    (0x39B, 'M', 'λ'),
    (0x39C, 'M', 'μ'),
    (0x39D, 'M', 'ν'),
    (0x39E, 'M', 'ξ'),
    (0x39F, 'M', 'ο'),
    (0x3A0, 'M', 'π'),
    (0x3A1, 'M', 'ρ'),
    (0x3A2, 'X'),
    (0x3A3, 'M', 'σ'),
    (0x3A4, 'M', 'τ'),
    (0x3A5, 'M', 'υ'),
    (0x3A6, 'M', 'φ'),
    (0x3A7, 'M', 'χ'),
    (0x3A8, 'M', 'ψ'),
    (0x3A9, 'M', 'ω'),
    (0x3AA, 'M', 'ϊ'),
    (0x3AB, 'M', 'ϋ'),
    (0x3AC, 'V'),
    (0x3C2, 'D', 'σ'),
    (0x3C3, 'V'),
    (0x3CF, 'M', 'ϗ'),
    (0x3D0, 'M', 'β'),
    (0x3D1, 'M', 'θ'),
    (0x3D2, 'M', 'υ'),
    (0x3D3, 'M', 'ύ'),
    (0x3D4, 'M', 'ϋ'),
    (0x3D5, 'M', 'φ'),
    (0x3D6, 'M', 'π'),
    (0x3D7, 'V'),
    (0x3D8, 'M', 'ϙ'),
    (0x3D9, 'V'),
    (0x3DA, 'M', 'ϛ'),
    (0x3DB, 'V'),
    (0x3DC, 'M', 'ϝ'),
    (0x3DD, 'V'),
    (0x3DE, 'M', 'ϟ'),
    (0x3DF, 'V'),
    (0x3E0, 'M', 'ϡ'),
    (0x3E1, 'V'),
    (0x3E2, 'M', 'ϣ'),
    (0x3E3, 'V'),
    (0x3E4, 'M', 'ϥ'),
    (0x3E5, 'V'),
    (0x3E6, 'M', 'ϧ'),
    (0x3E7, 'V'),
    (0x3E8, 'M', 'ϩ'),
    (0x3E9, 'V'),
    (0x3EA, 'M', 'ϫ'),
    (0x3EB, 'V'),
    (0x3EC, 'M', 'ϭ'),
    (0x3ED, 'V'),
    (0x3EE, 'M', 'ϯ'),
    (0x3EF, 'V'),
    (0x3F0, 'M', 'κ'),
    (0x3F1, 'M', 'ρ'),
    (0x3F2, 'M', 'σ'),
    (0x3F3, 'V'),
    (0x3F4, 'M', 'θ'),
    (0x3F5, 'M', 'ε'),
    (0x3F6, 'V'),
    (0x3F7, 'M', 'ϸ'),
    (0x3F8, 'V'),
    (0x3F9, 'M', 'σ'),
    (0x3FA, 'M', 'ϻ'),
    (0x3FB, 'V'),
    (0x3FD, 'M', 'ͻ'),
    (0x3FE, 'M', 'ͼ'),
    (0x3FF, 'M', 'ͽ'),
    (0x400, 'M', 'ѐ'),
    (0x401, 'M', 'ё'),
    (0x402, 'M', 'ђ'),
    ]

def _seg_7():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x403, 'M', 'ѓ'),
    (0x404, 'M', 'є'),
    (0x405, 'M', 'ѕ'),
    (0x406, 'M', 'і'),
    (0x407, 'M', 'ї'),
    (0x408, 'M', 'ј'),
    (0x409, 'M', 'љ'),
    (0x40A, 'M', 'њ'),
    (0x40B, 'M', 'ћ'),
    (0x40C, 'M', 'ќ'),
    (0x40D, 'M', 'ѝ'),
    (0x40E, 'M', 'ў'),
    (0x40F, 'M', 'џ'),
    (0x410, 'M', 'а'),
    (0x411, 'M', 'б'),
    (0x412, 'M', 'в'),
    (0x413, 'M', 'г'),
    (0x414, 'M', 'д'),
    (0x415, 'M', 'е'),
    (0x416, 'M', 'ж'),
    (0x417, 'M', 'з'),
    (0x418, 'M', 'и'),
    (0x419, 'M', 'й'),
    (0x41A, 'M', 'к'),
    (0x41B, 'M', 'л'),
    (0x41C, 'M', 'м'),
    (0x41D, 'M', 'н'),
    (0x41E, 'M', 'о'),
    (0x41F, 'M', 'п'),
    (0x420, 'M', 'р'),
    (0x421, 'M', 'с'),
    (0x422, 'M', 'т'),
    (0x423, 'M', 'у'),
    (0x424, 'M', 'ф'),
    (0x425, 'M', 'х'),
    (0x426, 'M', 'ц'),
    (0x427, 'M', 'ч'),
    (0x428, 'M', 'ш'),
    (0x429, 'M', 'щ'),
    (0x42A, 'M', 'ъ'),
    (0x42B, 'M', 'ы'),
    (0x42C, 'M', 'ь'),
    (0x42D, 'M', 'э'),
    (0x42E, 'M', 'ю'),
    (0x42F, 'M', 'я'),
    (0x430, 'V'),
    (0x460, 'M', 'ѡ'),
    (0x461, 'V'),
    (0x462, 'M', 'ѣ'),
    (0x463, 'V'),
    (0x464, 'M', 'ѥ'),
    (0x465, 'V'),
    (0x466, 'M', 'ѧ'),
    (0x467, 'V'),
    (0x468, 'M', 'ѩ'),
    (0x469, 'V'),
    (0x46A, 'M', 'ѫ'),
    (0x46B, 'V'),
    (0x46C, 'M', 'ѭ'),
    (0x46D, 'V'),
    (0x46E, 'M', 'ѯ'),
    (0x46F, 'V'),
    (0x470, 'M', 'ѱ'),
    (0x471, 'V'),
    (0x472, 'M', 'ѳ'),
    (0x473, 'V'),
    (0x474, 'M', 'ѵ'),
    (0x475, 'V'),
    (0x476, 'M', 'ѷ'),
    (0x477, 'V'),
    (0x478, 'M', 'ѹ'),
    (0x479, 'V'),
    (0x47A, 'M', 'ѻ'),
    (0x47B, 'V'),
    (0x47C, 'M', 'ѽ'),
    (0x47D, 'V'),
    (0x47E, 'M', 'ѿ'),
    (0x47F, 'V'),
    (0x480, 'M', 'ҁ'),
    (0x481, 'V'),
    (0x48A, 'M', 'ҋ'),
    (0x48B, 'V'),
    (0x48C, 'M', 'ҍ'),
    (0x48D, 'V'),
    (0x48E, 'M', 'ҏ'),
    (0x48F, 'V'),
    (0x490, 'M', 'ґ'),
    (0x491, 'V'),
    (0x492, 'M', 'ғ'),
    (0x493, 'V'),
    (0x494, 'M', 'ҕ'),
    (0x495, 'V'),
    (0x496, 'M', 'җ'),
    (0x497, 'V'),
    (0x498, 'M', 'ҙ'),
    (0x499, 'V'),
    (0x49A, 'M', 'қ'),
    (0x49B, 'V'),
    (0x49C, 'M', 'ҝ'),
    (0x49D, 'V'),
    ]

def _seg_8():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x49E, 'M', 'ҟ'),
    (0x49F, 'V'),
    (0x4A0, 'M', 'ҡ'),
    (0x4A1, 'V'),
    (0x4A2, 'M', 'ң'),
    (0x4A3, 'V'),
    (0x4A4, 'M', 'ҥ'),
    (0x4A5, 'V'),
    (0x4A6, 'M', 'ҧ'),
    (0x4A7, 'V'),
    (0x4A8, 'M', 'ҩ'),
    (0x4A9, 'V'),
    (0x4AA, 'M', 'ҫ'),
    (0x4AB, 'V'),
    (0x4AC, 'M', 'ҭ'),
    (0x4AD, 'V'),
    (0x4AE, 'M', 'ү'),
    (0x4AF, 'V'),
    (0x4B0, 'M', 'ұ'),
    (0x4B1, 'V'),
    (0x4B2, 'M', 'ҳ'),
    (0x4B3, 'V'),
    (0x4B4, 'M', 'ҵ'),
    (0x4B5, 'V'),
    (0x4B6, 'M', 'ҷ'),
    (0x4B7, 'V'),
    (0x4B8, 'M', 'ҹ'),
    (0x4B9, 'V'),
    (0x4BA, 'M', 'һ'),
    (0x4BB, 'V'),
    (0x4BC, 'M', 'ҽ'),
    (0x4BD, 'V'),
    (0x4BE, 'M', 'ҿ'),
    (0x4BF, 'V'),
    (0x4C0, 'X'),
    (0x4C1, 'M', 'ӂ'),
    (0x4C2, 'V'),
    (0x4C3, 'M', 'ӄ'),
    (0x4C4, 'V'),
    (0x4C5, 'M', 'ӆ'),
    (0x4C6, 'V'),
    (0x4C7, 'M', 'ӈ'),
    (0x4C8, 'V'),
    (0x4C9, 'M', 'ӊ'),
    (0x4CA, 'V'),
    (0x4CB, 'M', 'ӌ'),
    (0x4CC, 'V'),
    (0x4CD, 'M', 'ӎ'),
    (0x4CE, 'V'),
    (0x4D0, 'M', 'ӑ'),
    (0x4D1, 'V'),
    (0x4D2, 'M', 'ӓ'),
    (0x4D3, 'V'),
    (0x4D4, 'M', 'ӕ'),
    (0x4D5, 'V'),
    (0x4D6, 'M', 'ӗ'),
    (0x4D7, 'V'),
    (0x4D8, 'M', 'ә'),
    (0x4D9, 'V'),
    (0x4DA, 'M', 'ӛ'),
    (0x4DB, 'V'),
    (0x4DC, 'M', 'ӝ'),
    (0x4DD, 'V'),
    (0x4DE, 'M', 'ӟ'),
    (0x4DF, 'V'),
    (0x4E0, 'M', 'ӡ'),
    (0x4E1, 'V'),
    (0x4E2, 'M', 'ӣ'),
    (0x4E3, 'V'),
    (0x4E4, 'M', 'ӥ'),
    (0x4E5, 'V'),
    (0x4E6, 'M', 'ӧ'),
    (0x4E7, 'V'),
    (0x4E8, 'M', 'ө'),
    (0x4E9, 'V'),
    (0x4EA, 'M', 'ӫ'),
    (0x4EB, 'V'),
    (0x4EC, 'M', 'ӭ'),
    (0x4ED, 'V'),
    (0x4EE, 'M', 'ӯ'),
    (0x4EF, 'V'),
    (0x4F0, 'M', 'ӱ'),
    (0x4F1, 'V'),
    (0x4F2, 'M', 'ӳ'),
    (0x4F3, 'V'),
    (0x4F4, 'M', 'ӵ'),
    (0x4F5, 'V'),
    (0x4F6, 'M', 'ӷ'),
    (0x4F7, 'V'),
    (0x4F8, 'M', 'ӹ'),
    (0x4F9, 'V'),
    (0x4FA, 'M', 'ӻ'),
    (0x4FB, 'V'),
    (0x4FC, 'M', 'ӽ'),
    (0x4FD, 'V'),
    (0x4FE, 'M', 'ӿ'),
    (0x4FF, 'V'),
    (0x500, 'M', 'ԁ'),
    (0x501, 'V'),
    (0x502, 'M', 'ԃ'),
    ]

def _seg_9():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x503, 'V'),
    (0x504, 'M', 'ԅ'),
    (0x505, 'V'),
    (0x506, 'M', 'ԇ'),
    (0x507, 'V'),
    (0x508, 'M', 'ԉ'),
    (0x509, 'V'),
    (0x50A, 'M', 'ԋ'),
    (0x50B, 'V'),
    (0x50C, 'M', 'ԍ'),
    (0x50D, 'V'),
    (0x50E, 'M', 'ԏ'),
    (0x50F, 'V'),
    (0x510, 'M', 'ԑ'),
    (0x511, 'V'),
    (0x512, 'M', 'ԓ'),
    (0x513, 'V'),
    (0x514, 'M', 'ԕ'),
    (0x515, 'V'),
    (0x516, 'M', 'ԗ'),
    (0x517, 'V'),
    (0x518, 'M', 'ԙ'),
    (0x519, 'V'),
    (0x51A, 'M', 'ԛ'),
    (0x51B, 'V'),
    (0x51C, 'M', 'ԝ'),
    (0x51D, 'V'),
    (0x51E, 'M', 'ԟ'),
    (0x51F, 'V'),
    (0x520, 'M', 'ԡ'),
    (0x521, 'V'),
    (0x522, 'M', 'ԣ'),
    (0x523, 'V'),
    (0x524, 'M', 'ԥ'),
    (0x525, 'V'),
    (0x526, 'M', 'ԧ'),
    (0x527, 'V'),
    (0x528, 'M', 'ԩ'),
    (0x529, 'V'),
    (0x52A, 'M', 'ԫ'),
    (0x52B, 'V'),
    (0x52C, 'M', 'ԭ'),
    (0x52D, 'V'),
    (0x52E, 'M', 'ԯ'),
    (0x52F, 'V'),
    (0x530, 'X'),
    (0x531, 'M', 'ա'),
    (0x532, 'M', 'բ'),
    (0x533, 'M', 'գ'),
    (0x534, 'M', 'դ'),
    (0x535, 'M', 'ե'),
    (0x536, 'M', 'զ'),
    (0x537, 'M', 'է'),
    (0x538, 'M', 'ը'),
    (0x539, 'M', 'թ'),
    (0x53A, 'M', 'ժ'),
    (0x53B, 'M', 'ի'),
    (0x53C, 'M', 'լ'),
    (0x53D, 'M', 'խ'),
    (0x53E, 'M', 'ծ'),
    (0x53F, 'M', 'կ'),
    (0x540, 'M', 'հ'),
    (0x541, 'M', 'ձ'),
    (0x542, 'M', 'ղ'),
    (0x543, 'M', 'ճ'),
    (0x544, 'M', 'մ'),
    (0x545, 'M', 'յ'),
    (0x546, 'M', 'ն'),
    (0x547, 'M', 'շ'),
    (0x548, 'M', 'ո'),
    (0x549, 'M', 'չ'),
    (0x54A, 'M', 'պ'),
    (0x54B, 'M', 'ջ'),
    (0x54C, 'M', 'ռ'),
    (0x54D, 'M', 'ս'),
    (0x54E, 'M', 'վ'),
    (0x54F, 'M', 'տ'),
    (0x550, 'M', 'ր'),
    (0x551, 'M', 'ց'),
    (0x552, 'M', 'ւ'),
    (0x553, 'M', 'փ'),
    (0x554, 'M', 'ք'),
    (0x555, 'M', 'օ'),
    (0x556, 'M', 'ֆ'),
    (0x557, 'X'),
    (0x559, 'V'),
    (0x587, 'M', 'եւ'),
    (0x588, 'V'),
    (0x58B, 'X'),
    (0x58D, 'V'),
    (0x590, 'X'),
    (0x591, 'V'),
    (0x5C8, 'X'),
    (0x5D0, 'V'),
    (0x5EB, 'X'),
    (0x5EF, 'V'),
    (0x5F5, 'X'),
    (0x606, 'V'),
    (0x61C, 'X'),
    (0x61E, 'V'),
    ]

def _seg_10():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x675, 'M', 'اٴ'),
    (0x676, 'M', 'وٴ'),
    (0x677, 'M', 'ۇٴ'),
    (0x678, 'M', 'يٴ'),
    (0x679, 'V'),
    (0x6DD, 'X'),
    (0x6DE, 'V'),
    (0x70E, 'X'),
    (0x710, 'V'),
    (0x74B, 'X'),
    (0x74D, 'V'),
    (0x7B2, 'X'),
    (0x7C0, 'V'),
    (0x7FB, 'X'),
    (0x7FD, 'V'),
    (0x82E, 'X'),
    (0x830, 'V'),
    (0x83F, 'X'),
    (0x840, 'V'),
    (0x85C, 'X'),
    (0x85E, 'V'),
    (0x85F, 'X'),
    (0x860, 'V'),
    (0x86B, 'X'),
    (0x8A0, 'V'),
    (0x8B5, 'X'),
    (0x8B6, 'V'),
    (0x8C8, 'X'),
    (0x8D3, 'V'),
    (0x8E2, 'X'),
    (0x8E3, 'V'),
    (0x958, 'M', 'क़'),
    (0x959, 'M', 'ख़'),
    (0x95A, 'M', 'ग़'),
    (0x95B, 'M', 'ज़'),
    (0x95C, 'M', 'ड़'),
    (0x95D, 'M', 'ढ़'),
    (0x95E, 'M', 'फ़'),
    (0x95F, 'M', 'य़'),
    (0x960, 'V'),
    (0x984, 'X'),
    (0x985, 'V'),
    (0x98D, 'X'),
    (0x98F, 'V'),
    (0x991, 'X'),
    (0x993, 'V'),
    (0x9A9, 'X'),
    (0x9AA, 'V'),
    (0x9B1, 'X'),
    (0x9B2, 'V'),
    (0x9B3, 'X'),
    (0x9B6, 'V'),
    (0x9BA, 'X'),
    (0x9BC, 'V'),
    (0x9C5, 'X'),
    (0x9C7, 'V'),
    (0x9C9, 'X'),
    (0x9CB, 'V'),
    (0x9CF, 'X'),
    (0x9D7, 'V'),
    (0x9D8, 'X'),
    (0x9DC, 'M', 'ড়'),
    (0x9DD, 'M', 'ঢ়'),
    (0x9DE, 'X'),
    (0x9DF, 'M', 'য়'),
    (0x9E0, 'V'),
    (0x9E4, 'X'),
    (0x9E6, 'V'),
    (0x9FF, 'X'),
    (0xA01, 'V'),
    (0xA04, 'X'),
    (0xA05, 'V'),
    (0xA0B, 'X'),
    (0xA0F, 'V'),
    (0xA11, 'X'),
    (0xA13, 'V'),
    (0xA29, 'X'),
    (0xA2A, 'V'),
    (0xA31, 'X'),
    (0xA32, 'V'),
    (0xA33, 'M', 'ਲ਼'),
    (0xA34, 'X'),
    (0xA35, 'V'),
    (0xA36, 'M', 'ਸ਼'),
    (0xA37, 'X'),
    (0xA38, 'V'),
    (0xA3A, 'X'),
    (0xA3C, 'V'),
    (0xA3D, 'X'),
    (0xA3E, 'V'),
    (0xA43, 'X'),
    (0xA47, 'V'),
    (0xA49, 'X'),
    (0xA4B, 'V'),
    (0xA4E, 'X'),
    (0xA51, 'V'),
    (0xA52, 'X'),
    (0xA59, 'M', 'ਖ਼'),
    (0xA5A, 'M', 'ਗ਼'),
    (0xA5B, 'M', 'ਜ਼'),
    ]

def _seg_11():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xA5C, 'V'),
    (0xA5D, 'X'),
    (0xA5E, 'M', 'ਫ਼'),
    (0xA5F, 'X'),
    (0xA66, 'V'),
    (0xA77, 'X'),
    (0xA81, 'V'),
    (0xA84, 'X'),
    (0xA85, 'V'),
    (0xA8E, 'X'),
    (0xA8F, 'V'),
    (0xA92, 'X'),
    (0xA93, 'V'),
    (0xAA9, 'X'),
    (0xAAA, 'V'),
    (0xAB1, 'X'),
    (0xAB2, 'V'),
    (0xAB4, 'X'),
    (0xAB5, 'V'),
    (0xABA, 'X'),
    (0xABC, 'V'),
    (0xAC6, 'X'),
    (0xAC7, 'V'),
    (0xACA, 'X'),
    (0xACB, 'V'),
    (0xACE, 'X'),
    (0xAD0, 'V'),
    (0xAD1, 'X'),
    (0xAE0, 'V'),
    (0xAE4, 'X'),
    (0xAE6, 'V'),
    (0xAF2, 'X'),
    (0xAF9, 'V'),
    (0xB00, 'X'),
    (0xB01, 'V'),
    (0xB04, 'X'),
    (0xB05, 'V'),
    (0xB0D, 'X'),
    (0xB0F, 'V'),
    (0xB11, 'X'),
    (0xB13, 'V'),
    (0xB29, 'X'),
    (0xB2A, 'V'),
    (0xB31, 'X'),
    (0xB32, 'V'),
    (0xB34, 'X'),
    (0xB35, 'V'),
    (0xB3A, 'X'),
    (0xB3C, 'V'),
    (0xB45, 'X'),
    (0xB47, 'V'),
    (0xB49, 'X'),
    (0xB4B, 'V'),
    (0xB4E, 'X'),
    (0xB55, 'V'),
    (0xB58, 'X'),
    (0xB5C, 'M', 'ଡ଼'),
    (0xB5D, 'M', 'ଢ଼'),
    (0xB5E, 'X'),
    (0xB5F, 'V'),
    (0xB64, 'X'),
    (0xB66, 'V'),
    (0xB78, 'X'),
    (0xB82, 'V'),
    (0xB84, 'X'),
    (0xB85, 'V'),
    (0xB8B, 'X'),
    (0xB8E, 'V'),
    (0xB91, 'X'),
    (0xB92, 'V'),
    (0xB96, 'X'),
    (0xB99, 'V'),
    (0xB9B, 'X'),
    (0xB9C, 'V'),
    (0xB9D, 'X'),
    (0xB9E, 'V'),
    (0xBA0, 'X'),
    (0xBA3, 'V'),
    (0xBA5, 'X'),
    (0xBA8, 'V'),
    (0xBAB, 'X'),
    (0xBAE, 'V'),
    (0xBBA, 'X'),
    (0xBBE, 'V'),
    (0xBC3, 'X'),
    (0xBC6, 'V'),
    (0xBC9, 'X'),
    (0xBCA, 'V'),
    (0xBCE, 'X'),
    (0xBD0, 'V'),
    (0xBD1, 'X'),
    (0xBD7, 'V'),
    (0xBD8, 'X'),
    (0xBE6, 'V'),
    (0xBFB, 'X'),
    (0xC00, 'V'),
    (0xC0D, 'X'),
    (0xC0E, 'V'),
    (0xC11, 'X'),
    (0xC12, 'V'),
    ]

def _seg_12():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xC29, 'X'),
    (0xC2A, 'V'),
    (0xC3A, 'X'),
    (0xC3D, 'V'),
    (0xC45, 'X'),
    (0xC46, 'V'),
    (0xC49, 'X'),
    (0xC4A, 'V'),
    (0xC4E, 'X'),
    (0xC55, 'V'),
    (0xC57, 'X'),
    (0xC58, 'V'),
    (0xC5B, 'X'),
    (0xC60, 'V'),
    (0xC64, 'X'),
    (0xC66, 'V'),
    (0xC70, 'X'),
    (0xC77, 'V'),
    (0xC8D, 'X'),
    (0xC8E, 'V'),
    (0xC91, 'X'),
    (0xC92, 'V'),
    (0xCA9, 'X'),
    (0xCAA, 'V'),
    (0xCB4, 'X'),
    (0xCB5, 'V'),
    (0xCBA, 'X'),
    (0xCBC, 'V'),
    (0xCC5, 'X'),
    (0xCC6, 'V'),
    (0xCC9, 'X'),
    (0xCCA, 'V'),
    (0xCCE, 'X'),
    (0xCD5, 'V'),
    (0xCD7, 'X'),
    (0xCDE, 'V'),
    (0xCDF, 'X'),
    (0xCE0, 'V'),
    (0xCE4, 'X'),
    (0xCE6, 'V'),
    (0xCF0, 'X'),
    (0xCF1, 'V'),
    (0xCF3, 'X'),
    (0xD00, 'V'),
    (0xD0D, 'X'),
    (0xD0E, 'V'),
    (0xD11, 'X'),
    (0xD12, 'V'),
    (0xD45, 'X'),
    (0xD46, 'V'),
    (0xD49, 'X'),
    (0xD4A, 'V'),
    (0xD50, 'X'),
    (0xD54, 'V'),
    (0xD64, 'X'),
    (0xD66, 'V'),
    (0xD80, 'X'),
    (0xD81, 'V'),
    (0xD84, 'X'),
    (0xD85, 'V'),
    (0xD97, 'X'),
    (0xD9A, 'V'),
    (0xDB2, 'X'),
    (0xDB3, 'V'),
    (0xDBC, 'X'),
    (0xDBD, 'V'),
    (0xDBE, 'X'),
    (0xDC0, 'V'),
    (0xDC7, 'X'),
    (0xDCA, 'V'),
    (0xDCB, 'X'),
    (0xDCF, 'V'),
    (0xDD5, 'X'),
    (0xDD6, 'V'),
    (0xDD7, 'X'),
    (0xDD8, 'V'),
    (0xDE0, 'X'),
    (0xDE6, 'V'),
    (0xDF0, 'X'),
    (0xDF2, 'V'),
    (0xDF5, 'X'),
    (0xE01, 'V'),
    (0xE33, 'M', 'ํา'),
    (0xE34, 'V'),
    (0xE3B, 'X'),
    (0xE3F, 'V'),
    (0xE5C, 'X'),
    (0xE81, 'V'),
    (0xE83, 'X'),
    (0xE84, 'V'),
    (0xE85, 'X'),
    (0xE86, 'V'),
    (0xE8B, 'X'),
    (0xE8C, 'V'),
    (0xEA4, 'X'),
    (0xEA5, 'V'),
    (0xEA6, 'X'),
    (0xEA7, 'V'),
    (0xEB3, 'M', 'ໍາ'),
    (0xEB4, 'V'),
    ]

def _seg_13():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xEBE, 'X'),
    (0xEC0, 'V'),
    (0xEC5, 'X'),
    (0xEC6, 'V'),
    (0xEC7, 'X'),
    (0xEC8, 'V'),
    (0xECE, 'X'),
    (0xED0, 'V'),
    (0xEDA, 'X'),
    (0xEDC, 'M', 'ຫນ'),
    (0xEDD, 'M', 'ຫມ'),
    (0xEDE, 'V'),
    (0xEE0, 'X'),
    (0xF00, 'V'),
    (0xF0C, 'M', '་'),
    (0xF0D, 'V'),
    (0xF43, 'M', 'གྷ'),
    (0xF44, 'V'),
    (0xF48, 'X'),
    (0xF49, 'V'),
    (0xF4D, 'M', 'ཌྷ'),
    (0xF4E, 'V'),
    (0xF52, 'M', 'དྷ'),
    (0xF53, 'V'),
    (0xF57, 'M', 'བྷ'),
    (0xF58, 'V'),
    (0xF5C, 'M', 'ཛྷ'),
    (0xF5D, 'V'),
    (0xF69, 'M', 'ཀྵ'),
    (0xF6A, 'V'),
    (0xF6D, 'X'),
    (0xF71, 'V'),
    (0xF73, 'M', 'ཱི'),
    (0xF74, 'V'),
    (0xF75, 'M', 'ཱུ'),
    (0xF76, 'M', 'ྲྀ'),
    (0xF77, 'M', 'ྲཱྀ'),
    (0xF78, 'M', 'ླྀ'),
    (0xF79, 'M', 'ླཱྀ'),
    (0xF7A, 'V'),
    (0xF81, 'M', 'ཱྀ'),
    (0xF82, 'V'),
    (0xF93, 'M', 'ྒྷ'),
    (0xF94, 'V'),
    (0xF98, 'X'),
    (0xF99, 'V'),
    (0xF9D, 'M', 'ྜྷ'),
    (0xF9E, 'V'),
    (0xFA2, 'M', 'ྡྷ'),
    (0xFA3, 'V'),
    (0xFA7, 'M', 'ྦྷ'),
    (0xFA8, 'V'),
    (0xFAC, 'M', 'ྫྷ'),
    (0xFAD, 'V'),
    (0xFB9, 'M', 'ྐྵ'),
    (0xFBA, 'V'),
    (0xFBD, 'X'),
    (0xFBE, 'V'),
    (0xFCD, 'X'),
    (0xFCE, 'V'),
    (0xFDB, 'X'),
    (0x1000, 'V'),
    (0x10A0, 'X'),
    (0x10C7, 'M', 'ⴧ'),
    (0x10C8, 'X'),
    (0x10CD, 'M', 'ⴭ'),
    (0x10CE, 'X'),
    (0x10D0, 'V'),
    (0x10FC, 'M', 'ნ'),
    (0x10FD, 'V'),
    (0x115F, 'X'),
    (0x1161, 'V'),
    (0x1249, 'X'),
    (0x124A, 'V'),
    (0x124E, 'X'),
    (0x1250, 'V'),
    (0x1257, 'X'),
    (0x1258, 'V'),
    (0x1259, 'X'),
    (0x125A, 'V'),
    (0x125E, 'X'),
    (0x1260, 'V'),
    (0x1289, 'X'),
    (0x128A, 'V'),
    (0x128E, 'X'),
    (0x1290, 'V'),
    (0x12B1, 'X'),
    (0x12B2, 'V'),
    (0x12B6, 'X'),
    (0x12B8, 'V'),
    (0x12BF, 'X'),
    (0x12C0, 'V'),
    (0x12C1, 'X'),
    (0x12C2, 'V'),
    (0x12C6, 'X'),
    (0x12C8, 'V'),
    (0x12D7, 'X'),
    (0x12D8, 'V'),
    (0x1311, 'X'),
    (0x1312, 'V'),
    ]

def _seg_14():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1316, 'X'),
    (0x1318, 'V'),
    (0x135B, 'X'),
    (0x135D, 'V'),
    (0x137D, 'X'),
    (0x1380, 'V'),
    (0x139A, 'X'),
    (0x13A0, 'V'),
    (0x13F6, 'X'),
    (0x13F8, 'M', 'Ᏸ'),
    (0x13F9, 'M', 'Ᏹ'),
    (0x13FA, 'M', 'Ᏺ'),
    (0x13FB, 'M', 'Ᏻ'),
    (0x13FC, 'M', 'Ᏼ'),
    (0x13FD, 'M', 'Ᏽ'),
    (0x13FE, 'X'),
    (0x1400, 'V'),
    (0x1680, 'X'),
    (0x1681, 'V'),
    (0x169D, 'X'),
    (0x16A0, 'V'),
    (0x16F9, 'X'),
    (0x1700, 'V'),
    (0x170D, 'X'),
    (0x170E, 'V'),
    (0x1715, 'X'),
    (0x1720, 'V'),
    (0x1737, 'X'),
    (0x1740, 'V'),
    (0x1754, 'X'),
    (0x1760, 'V'),
    (0x176D, 'X'),
    (0x176E, 'V'),
    (0x1771, 'X'),
    (0x1772, 'V'),
    (0x1774, 'X'),
    (0x1780, 'V'),
    (0x17B4, 'X'),
    (0x17B6, 'V'),
    (0x17DE, 'X'),
    (0x17E0, 'V'),
    (0x17EA, 'X'),
    (0x17F0, 'V'),
    (0x17FA, 'X'),
    (0x1800, 'V'),
    (0x1806, 'X'),
    (0x1807, 'V'),
    (0x180B, 'I'),
    (0x180E, 'X'),
    (0x1810, 'V'),
    (0x181A, 'X'),
    (0x1820, 'V'),
    (0x1879, 'X'),
    (0x1880, 'V'),
    (0x18AB, 'X'),
    (0x18B0, 'V'),
    (0x18F6, 'X'),
    (0x1900, 'V'),
    (0x191F, 'X'),
    (0x1920, 'V'),
    (0x192C, 'X'),
    (0x1930, 'V'),
    (0x193C, 'X'),
    (0x1940, 'V'),
    (0x1941, 'X'),
    (0x1944, 'V'),
    (0x196E, 'X'),
    (0x1970, 'V'),
    (0x1975, 'X'),
    (0x1980, 'V'),
    (0x19AC, 'X'),
    (0x19B0, 'V'),
    (0x19CA, 'X'),
    (0x19D0, 'V'),
    (0x19DB, 'X'),
    (0x19DE, 'V'),
    (0x1A1C, 'X'),
    (0x1A1E, 'V'),
    (0x1A5F, 'X'),
    (0x1A60, 'V'),
    (0x1A7D, 'X'),
    (0x1A7F, 'V'),
    (0x1A8A, 'X'),
    (0x1A90, 'V'),
    (0x1A9A, 'X'),
    (0x1AA0, 'V'),
    (0x1AAE, 'X'),
    (0x1AB0, 'V'),
    (0x1AC1, 'X'),
    (0x1B00, 'V'),
    (0x1B4C, 'X'),
    (0x1B50, 'V'),
    (0x1B7D, 'X'),
    (0x1B80, 'V'),
    (0x1BF4, 'X'),
    (0x1BFC, 'V'),
    (0x1C38, 'X'),
    (0x1C3B, 'V'),
    (0x1C4A, 'X'),
    (0x1C4D, 'V'),
    ]

def _seg_15():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1C80, 'M', 'в'),
    (0x1C81, 'M', 'д'),
    (0x1C82, 'M', 'о'),
    (0x1C83, 'M', 'с'),
    (0x1C84, 'M', 'т'),
    (0x1C86, 'M', 'ъ'),
    (0x1C87, 'M', 'ѣ'),
    (0x1C88, 'M', 'ꙋ'),
    (0x1C89, 'X'),
    (0x1C90, 'M', 'ა'),
    (0x1C91, 'M', 'ბ'),
    (0x1C92, 'M', 'გ'),
    (0x1C93, 'M', 'დ'),
    (0x1C94, 'M', 'ე'),
    (0x1C95, 'M', 'ვ'),
    (0x1C96, 'M', 'ზ'),
    (0x1C97, 'M', 'თ'),
    (0x1C98, 'M', 'ი'),
    (0x1C99, 'M', 'კ'),
    (0x1C9A, 'M', 'ლ'),
    (0x1C9B, 'M', 'მ'),
    (0x1C9C, 'M', 'ნ'),
    (0x1C9D, 'M', 'ო'),
    (0x1C9E, 'M', 'პ'),
    (0x1C9F, 'M', 'ჟ'),
    (0x1CA0, 'M', 'რ'),
    (0x1CA1, 'M', 'ს'),
    (0x1CA2, 'M', 'ტ'),
    (0x1CA3, 'M', 'უ'),
    (0x1CA4, 'M', 'ფ'),
    (0x1CA5, 'M', 'ქ'),
    (0x1CA6, 'M', 'ღ'),
    (0x1CA7, 'M', 'ყ'),
    (0x1CA8, 'M', 'შ'),
    (0x1CA9, 'M', 'ჩ'),
    (0x1CAA, 'M', 'ც'),
    (0x1CAB, 'M', 'ძ'),
    (0x1CAC, 'M', 'წ'),
    (0x1CAD, 'M', 'ჭ'),
    (0x1CAE, 'M', 'ხ'),
    (0x1CAF, 'M', 'ჯ'),
    (0x1CB0, 'M', 'ჰ'),
    (0x1CB1, 'M', 'ჱ'),
    (0x1CB2, 'M', 'ჲ'),
    (0x1CB3, 'M', 'ჳ'),
    (0x1CB4, 'M', 'ჴ'),
    (0x1CB5, 'M', 'ჵ'),
    (0x1CB6, 'M', 'ჶ'),
    (0x1CB7, 'M', 'ჷ'),
    (0x1CB8, 'M', 'ჸ'),
    (0x1CB9, 'M', 'ჹ'),
    (0x1CBA, 'M', 'ჺ'),
    (0x1CBB, 'X'),
    (0x1CBD, 'M', 'ჽ'),
    (0x1CBE, 'M', 'ჾ'),
    (0x1CBF, 'M', 'ჿ'),
    (0x1CC0, 'V'),
    (0x1CC8, 'X'),
    (0x1CD0, 'V'),
    (0x1CFB, 'X'),
    (0x1D00, 'V'),
    (0x1D2C, 'M', 'a'),
    (0x1D2D, 'M', 'æ'),
    (0x1D2E, 'M', 'b'),
    (0x1D2F, 'V'),
    (0x1D30, 'M', 'd'),
    (0x1D31, 'M', 'e'),
    (0x1D32, 'M', 'ǝ'),
    (0x1D33, 'M', 'g'),
    (0x1D34, 'M', 'h'),
    (0x1D35, 'M', 'i'),
    (0x1D36, 'M', 'j'),
    (0x1D37, 'M', 'k'),
    (0x1D38, 'M', 'l'),
    (0x1D39, 'M', 'm'),
    (0x1D3A, 'M', 'n'),
    (0x1D3B, 'V'),
    (0x1D3C, 'M', 'o'),
    (0x1D3D, 'M', 'ȣ'),
    (0x1D3E, 'M', 'p'),
    (0x1D3F, 'M', 'r'),
    (0x1D40, 'M', 't'),
    (0x1D41, 'M', 'u'),
    (0x1D42, 'M', 'w'),
    (0x1D43, 'M', 'a'),
    (0x1D44, 'M', 'ɐ'),
    (0x1D45, 'M', 'ɑ'),
    (0x1D46, 'M', 'ᴂ'),
    (0x1D47, 'M', 'b'),
    (0x1D48, 'M', 'd'),
    (0x1D49, 'M', 'e'),
    (0x1D4A, 'M', 'ə'),
    (0x1D4B, 'M', 'ɛ'),
    (0x1D4C, 'M', 'ɜ'),
    (0x1D4D, 'M', 'g'),
    (0x1D4E, 'V'),
    (0x1D4F, 'M', 'k'),
    (0x1D50, 'M', 'm'),
    (0x1D51, 'M', 'ŋ'),
    (0x1D52, 'M', 'o'),
    ]

def _seg_16():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D53, 'M', 'ɔ'),
    (0x1D54, 'M', 'ᴖ'),
    (0x1D55, 'M', 'ᴗ'),
    (0x1D56, 'M', 'p'),
    (0x1D57, 'M', 't'),
    (0x1D58, 'M', 'u'),
    (0x1D59, 'M', 'ᴝ'),
    (0x1D5A, 'M', 'ɯ'),
    (0x1D5B, 'M', 'v'),
    (0x1D5C, 'M', 'ᴥ'),
    (0x1D5D, 'M', 'β'),
    (0x1D5E, 'M', 'γ'),
    (0x1D5F, 'M', 'δ'),
    (0x1D60, 'M', 'φ'),
    (0x1D61, 'M', 'χ'),
    (0x1D62, 'M', 'i'),
    (0x1D63, 'M', 'r'),
    (0x1D64, 'M', 'u'),
    (0x1D65, 'M', 'v'),
    (0x1D66, 'M', 'β'),
    (0x1D67, 'M', 'γ'),
    (0x1D68, 'M', 'ρ'),
    (0x1D69, 'M', 'φ'),
    (0x1D6A, 'M', 'χ'),
    (0x1D6B, 'V'),
    (0x1D78, 'M', 'н'),
    (0x1D79, 'V'),
    (0x1D9B, 'M', 'ɒ'),
    (0x1D9C, 'M', 'c'),
    (0x1D9D, 'M', 'ɕ'),
    (0x1D9E, 'M', 'ð'),
    (0x1D9F, 'M', 'ɜ'),
    (0x1DA0, 'M', 'f'),
    (0x1DA1, 'M', 'ɟ'),
    (0x1DA2, 'M', 'ɡ'),
    (0x1DA3, 'M', 'ɥ'),
    (0x1DA4, 'M', 'ɨ'),
    (0x1DA5, 'M', 'ɩ'),
    (0x1DA6, 'M', 'ɪ'),
    (0x1DA7, 'M', 'ᵻ'),
    (0x1DA8, 'M', 'ʝ'),
    (0x1DA9, 'M', 'ɭ'),
    (0x1DAA, 'M', 'ᶅ'),
    (0x1DAB, 'M', 'ʟ'),
    (0x1DAC, 'M', 'ɱ'),
    (0x1DAD, 'M', 'ɰ'),
    (0x1DAE, 'M', 'ɲ'),
    (0x1DAF, 'M', 'ɳ'),
    (0x1DB0, 'M', 'ɴ'),
    (0x1DB1, 'M', 'ɵ'),
    (0x1DB2, 'M', 'ɸ'),
    (0x1DB3, 'M', 'ʂ'),
    (0x1DB4, 'M', 'ʃ'),
    (0x1DB5, 'M', 'ƫ'),
    (0x1DB6, 'M', 'ʉ'),
    (0x1DB7, 'M', 'ʊ'),
    (0x1DB8, 'M', 'ᴜ'),
    (0x1DB9, 'M', 'ʋ'),
    (0x1DBA, 'M', 'ʌ'),
    (0x1DBB, 'M', 'z'),
    (0x1DBC, 'M', 'ʐ'),
    (0x1DBD, 'M', 'ʑ'),
    (0x1DBE, 'M', 'ʒ'),
    (0x1DBF, 'M', 'θ'),
    (0x1DC0, 'V'),
    (0x1DFA, 'X'),
    (0x1DFB, 'V'),
    (0x1E00, 'M', 'ḁ'),
    (0x1E01, 'V'),
    (0x1E02, 'M', 'ḃ'),
    (0x1E03, 'V'),
    (0x1E04, 'M', 'ḅ'),
    (0x1E05, 'V'),
    (0x1E06, 'M', 'ḇ'),
    (0x1E07, 'V'),
    (0x1E08, 'M', 'ḉ'),
    (0x1E09, 'V'),
    (0x1E0A, 'M', 'ḋ'),
    (0x1E0B, 'V'),
    (0x1E0C, 'M', 'ḍ'),
    (0x1E0D, 'V'),
    (0x1E0E, 'M', 'ḏ'),
    (0x1E0F, 'V'),
    (0x1E10, 'M', 'ḑ'),
    (0x1E11, 'V'),
    (0x1E12, 'M', 'ḓ'),
    (0x1E13, 'V'),
    (0x1E14, 'M', 'ḕ'),
    (0x1E15, 'V'),
    (0x1E16, 'M', 'ḗ'),
    (0x1E17, 'V'),
    (0x1E18, 'M', 'ḙ'),
    (0x1E19, 'V'),
    (0x1E1A, 'M', 'ḛ'),
    (0x1E1B, 'V'),
    (0x1E1C, 'M', 'ḝ'),
    (0x1E1D, 'V'),
    (0x1E1E, 'M', 'ḟ'),
    (0x1E1F, 'V'),
    (0x1E20, 'M', 'ḡ'),
    ]

def _seg_17():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1E21, 'V'),
    (0x1E22, 'M', 'ḣ'),
    (0x1E23, 'V'),
    (0x1E24, 'M', 'ḥ'),
    (0x1E25, 'V'),
    (0x1E26, 'M', 'ḧ'),
    (0x1E27, 'V'),
    (0x1E28, 'M', 'ḩ'),
    (0x1E29, 'V'),
    (0x1E2A, 'M', 'ḫ'),
    (0x1E2B, 'V'),
    (0x1E2C, 'M', 'ḭ'),
    (0x1E2D, 'V'),
    (0x1E2E, 'M', 'ḯ'),
    (0x1E2F, 'V'),
    (0x1E30, 'M', 'ḱ'),
    (0x1E31, 'V'),
    (0x1E32, 'M', 'ḳ'),
    (0x1E33, 'V'),
    (0x1E34, 'M', 'ḵ'),
    (0x1E35, 'V'),
    (0x1E36, 'M', 'ḷ'),
    (0x1E37, 'V'),
    (0x1E38, 'M', 'ḹ'),
    (0x1E39, 'V'),
    (0x1E3A, 'M', 'ḻ'),
    (0x1E3B, 'V'),
    (0x1E3C, 'M', 'ḽ'),
    (0x1E3D, 'V'),
    (0x1E3E, 'M', 'ḿ'),
    (0x1E3F, 'V'),
    (0x1E40, 'M', 'ṁ'),
    (0x1E41, 'V'),
    (0x1E42, 'M', 'ṃ'),
    (0x1E43, 'V'),
    (0x1E44, 'M', 'ṅ'),
    (0x1E45, 'V'),
    (0x1E46, 'M', 'ṇ'),
    (0x1E47, 'V'),
    (0x1E48, 'M', 'ṉ'),
    (0x1E49, 'V'),
    (0x1E4A, 'M', 'ṋ'),
    (0x1E4B, 'V'),
    (0x1E4C, 'M', 'ṍ'),
    (0x1E4D, 'V'),
    (0x1E4E, 'M', 'ṏ'),
    (0x1E4F, 'V'),
    (0x1E50, 'M', 'ṑ'),
    (0x1E51, 'V'),
    (0x1E52, 'M', 'ṓ'),
    (0x1E53, 'V'),
    (0x1E54, 'M', 'ṕ'),
    (0x1E55, 'V'),
    (0x1E56, 'M', 'ṗ'),
    (0x1E57, 'V'),
    (0x1E58, 'M', 'ṙ'),
    (0x1E59, 'V'),
    (0x1E5A, 'M', 'ṛ'),
    (0x1E5B, 'V'),
    (0x1E5C, 'M', 'ṝ'),
    (0x1E5D, 'V'),
    (0x1E5E, 'M', 'ṟ'),
    (0x1E5F, 'V'),
    (0x1E60, 'M', 'ṡ'),
    (0x1E61, 'V'),
    (0x1E62, 'M', 'ṣ'),
    (0x1E63, 'V'),
    (0x1E64, 'M', 'ṥ'),
    (0x1E65, 'V'),
    (0x1E66, 'M', 'ṧ'),
    (0x1E67, 'V'),
    (0x1E68, 'M', 'ṩ'),
    (0x1E69, 'V'),
    (0x1E6A, 'M', 'ṫ'),
    (0x1E6B, 'V'),
    (0x1E6C, 'M', 'ṭ'),
    (0x1E6D, 'V'),
    (0x1E6E, 'M', 'ṯ'),
    (0x1E6F, 'V'),
    (0x1E70, 'M', 'ṱ'),
    (0x1E71, 'V'),
    (0x1E72, 'M', 'ṳ'),
    (0x1E73, 'V'),
    (0x1E74, 'M', 'ṵ'),
    (0x1E75, 'V'),
    (0x1E76, 'M', 'ṷ'),
    (0x1E77, 'V'),
    (0x1E78, 'M', 'ṹ'),
    (0x1E79, 'V'),
    (0x1E7A, 'M', 'ṻ'),
    (0x1E7B, 'V'),
    (0x1E7C, 'M', 'ṽ'),
    (0x1E7D, 'V'),
    (0x1E7E, 'M', 'ṿ'),
    (0x1E7F, 'V'),
    (0x1E80, 'M', 'ẁ'),
    (0x1E81, 'V'),
    (0x1E82, 'M', 'ẃ'),
    (0x1E83, 'V'),
    (0x1E84, 'M', 'ẅ'),
    ]

def _seg_18():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1E85, 'V'),
    (0x1E86, 'M', 'ẇ'),
    (0x1E87, 'V'),
    (0x1E88, 'M', 'ẉ'),
    (0x1E89, 'V'),
    (0x1E8A, 'M', 'ẋ'),
    (0x1E8B, 'V'),
    (0x1E8C, 'M', 'ẍ'),
    (0x1E8D, 'V'),
    (0x1E8E, 'M', 'ẏ'),
    (0x1E8F, 'V'),
    (0x1E90, 'M', 'ẑ'),
    (0x1E91, 'V'),
    (0x1E92, 'M', 'ẓ'),
    (0x1E93, 'V'),
    (0x1E94, 'M', 'ẕ'),
    (0x1E95, 'V'),
    (0x1E9A, 'M', 'aʾ'),
    (0x1E9B, 'M', 'ṡ'),
    (0x1E9C, 'V'),
    (0x1E9E, 'M', 'ss'),
    (0x1E9F, 'V'),
    (0x1EA0, 'M', 'ạ'),
    (0x1EA1, 'V'),
    (0x1EA2, 'M', 'ả'),
    (0x1EA3, 'V'),
    (0x1EA4, 'M', 'ấ'),
    (0x1EA5, 'V'),
    (0x1EA6, 'M', 'ầ'),
    (0x1EA7, 'V'),
    (0x1EA8, 'M', 'ẩ'),
    (0x1EA9, 'V'),
    (0x1EAA, 'M', 'ẫ'),
    (0x1EAB, 'V'),
    (0x1EAC, 'M', 'ậ'),
    (0x1EAD, 'V'),
    (0x1EAE, 'M', 'ắ'),
    (0x1EAF, 'V'),
    (0x1EB0, 'M', 'ằ'),
    (0x1EB1, 'V'),
    (0x1EB2, 'M', 'ẳ'),
    (0x1EB3, 'V'),
    (0x1EB4, 'M', 'ẵ'),
    (0x1EB5, 'V'),
    (0x1EB6, 'M', 'ặ'),
    (0x1EB7, 'V'),
    (0x1EB8, 'M', 'ẹ'),
    (0x1EB9, 'V'),
    (0x1EBA, 'M', 'ẻ'),
    (0x1EBB, 'V'),
    (0x1EBC, 'M', 'ẽ'),
    (0x1EBD, 'V'),
    (0x1EBE, 'M', 'ế'),
    (0x1EBF, 'V'),
    (0x1EC0, 'M', 'ề'),
    (0x1EC1, 'V'),
    (0x1EC2, 'M', 'ể'),
    (0x1EC3, 'V'),
    (0x1EC4, 'M', 'ễ'),
    (0x1EC5, 'V'),
    (0x1EC6, 'M', 'ệ'),
    (0x1EC7, 'V'),
    (0x1EC8, 'M', 'ỉ'),
    (0x1EC9, 'V'),
    (0x1ECA, 'M', 'ị'),
    (0x1ECB, 'V'),
    (0x1ECC, 'M', 'ọ'),
    (0x1ECD, 'V'),
    (0x1ECE, 'M', 'ỏ'),
    (0x1ECF, 'V'),
    (0x1ED0, 'M', 'ố'),
    (0x1ED1, 'V'),
    (0x1ED2, 'M', 'ồ'),
    (0x1ED3, 'V'),
    (0x1ED4, 'M', 'ổ'),
    (0x1ED5, 'V'),
    (0x1ED6, 'M', 'ỗ'),
    (0x1ED7, 'V'),
    (0x1ED8, 'M', 'ộ'),
    (0x1ED9, 'V'),
    (0x1EDA, 'M', 'ớ'),
    (0x1EDB, 'V'),
    (0x1EDC, 'M', 'ờ'),
    (0x1EDD, 'V'),
    (0x1EDE, 'M', 'ở'),
    (0x1EDF, 'V'),
    (0x1EE0, 'M', 'ỡ'),
    (0x1EE1, 'V'),
    (0x1EE2, 'M', 'ợ'),
    (0x1EE3, 'V'),
    (0x1EE4, 'M', 'ụ'),
    (0x1EE5, 'V'),
    (0x1EE6, 'M', 'ủ'),
    (0x1EE7, 'V'),
    (0x1EE8, 'M', 'ứ'),
    (0x1EE9, 'V'),
    (0x1EEA, 'M', 'ừ'),
    (0x1EEB, 'V'),
    (0x1EEC, 'M', 'ử'),
    (0x1EED, 'V'),
    ]

def _seg_19():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1EEE, 'M', 'ữ'),
    (0x1EEF, 'V'),
    (0x1EF0, 'M', 'ự'),
    (0x1EF1, 'V'),
    (0x1EF2, 'M', 'ỳ'),
    (0x1EF3, 'V'),
    (0x1EF4, 'M', 'ỵ'),
    (0x1EF5, 'V'),
    (0x1EF6, 'M', 'ỷ'),
    (0x1EF7, 'V'),
    (0x1EF8, 'M', 'ỹ'),
    (0x1EF9, 'V'),
    (0x1EFA, 'M', 'ỻ'),
    (0x1EFB, 'V'),
    (0x1EFC, 'M', 'ỽ'),
    (0x1EFD, 'V'),
    (0x1EFE, 'M', 'ỿ'),
    (0x1EFF, 'V'),
    (0x1F08, 'M', 'ἀ'),
    (0x1F09, 'M', 'ἁ'),
    (0x1F0A, 'M', 'ἂ'),
    (0x1F0B, 'M', 'ἃ'),
    (0x1F0C, 'M', 'ἄ'),
    (0x1F0D, 'M', 'ἅ'),
    (0x1F0E, 'M', 'ἆ'),
    (0x1F0F, 'M', 'ἇ'),
    (0x1F10, 'V'),
    (0x1F16, 'X'),
    (0x1F18, 'M', 'ἐ'),
    (0x1F19, 'M', 'ἑ'),
    (0x1F1A, 'M', 'ἒ'),
    (0x1F1B, 'M', 'ἓ'),
    (0x1F1C, 'M', 'ἔ'),
    (0x1F1D, 'M', 'ἕ'),
    (0x1F1E, 'X'),
    (0x1F20, 'V'),
    (0x1F28, 'M', 'ἠ'),
    (0x1F29, 'M', 'ἡ'),
    (0x1F2A, 'M', 'ἢ'),
    (0x1F2B, 'M', 'ἣ'),
    (0x1F2C, 'M', 'ἤ'),
    (0x1F2D, 'M', 'ἥ'),
    (0x1F2E, 'M', 'ἦ'),
    (0x1F2F, 'M', 'ἧ'),
    (0x1F30, 'V'),
    (0x1F38, 'M', 'ἰ'),
    (0x1F39, 'M', 'ἱ'),
    (0x1F3A, 'M', 'ἲ'),
    (0x1F3B, 'M', 'ἳ'),
    (0x1F3C, 'M', 'ἴ'),
    (0x1F3D, 'M', 'ἵ'),
    (0x1F3E, 'M', 'ἶ'),
    (0x1F3F, 'M', 'ἷ'),
    (0x1F40, 'V'),
    (0x1F46, 'X'),
    (0x1F48, 'M', 'ὀ'),
    (0x1F49, 'M', 'ὁ'),
    (0x1F4A, 'M', 'ὂ'),
    (0x1F4B, 'M', 'ὃ'),
    (0x1F4C, 'M', 'ὄ'),
    (0x1F4D, 'M', 'ὅ'),
    (0x1F4E, 'X'),
    (0x1F50, 'V'),
    (0x1F58, 'X'),
    (0x1F59, 'M', 'ὑ'),
    (0x1F5A, 'X'),
    (0x1F5B, 'M', 'ὓ'),
    (0x1F5C, 'X'),
    (0x1F5D, 'M', 'ὕ'),
    (0x1F5E, 'X'),
    (0x1F5F, 'M', 'ὗ'),
    (0x1F60, 'V'),
    (0x1F68, 'M', 'ὠ'),
    (0x1F69, 'M', 'ὡ'),
    (0x1F6A, 'M', 'ὢ'),
    (0x1F6B, 'M', 'ὣ'),
    (0x1F6C, 'M', 'ὤ'),
    (0x1F6D, 'M', 'ὥ'),
    (0x1F6E, 'M', 'ὦ'),
    (0x1F6F, 'M', 'ὧ'),
    (0x1F70, 'V'),
    (0x1F71, 'M', 'ά'),
    (0x1F72, 'V'),
    (0x1F73, 'M', 'έ'),
    (0x1F74, 'V'),
    (0x1F75, 'M', 'ή'),
    (0x1F76, 'V'),
    (0x1F77, 'M', 'ί'),
    (0x1F78, 'V'),
    (0x1F79, 'M', 'ό'),
    (0x1F7A, 'V'),
    (0x1F7B, 'M', 'ύ'),
    (0x1F7C, 'V'),
    (0x1F7D, 'M', 'ώ'),
    (0x1F7E, 'X'),
    (0x1F80, 'M', 'ἀι'),
    (0x1F81, 'M', 'ἁι'),
    (0x1F82, 'M', 'ἂι'),
    (0x1F83, 'M', 'ἃι'),
    (0x1F84, 'M', 'ἄι'),
    ]

def _seg_20():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1F85, 'M', 'ἅι'),
    (0x1F86, 'M', 'ἆι'),
    (0x1F87, 'M', 'ἇι'),
    (0x1F88, 'M', 'ἀι'),
    (0x1F89, 'M', 'ἁι'),
    (0x1F8A, 'M', 'ἂι'),
    (0x1F8B, 'M', 'ἃι'),
    (0x1F8C, 'M', 'ἄι'),
    (0x1F8D, 'M', 'ἅι'),
    (0x1F8E, 'M', 'ἆι'),
    (0x1F8F, 'M', 'ἇι'),
    (0x1F90, 'M', 'ἠι'),
    (0x1F91, 'M', 'ἡι'),
    (0x1F92, 'M', 'ἢι'),
    (0x1F93, 'M', 'ἣι'),
    (0x1F94, 'M', 'ἤι'),
    (0x1F95, 'M', 'ἥι'),
    (0x1F96, 'M', 'ἦι'),
    (0x1F97, 'M', 'ἧι'),
    (0x1F98, 'M', 'ἠι'),
    (0x1F99, 'M', 'ἡι'),
    (0x1F9A, 'M', 'ἢι'),
    (0x1F9B, 'M', 'ἣι'),
    (0x1F9C, 'M', 'ἤι'),
    (0x1F9D, 'M', 'ἥι'),
    (0x1F9E, 'M', 'ἦι'),
    (0x1F9F, 'M', 'ἧι'),
    (0x1FA0, 'M', 'ὠι'),
    (0x1FA1, 'M', 'ὡι'),
    (0x1FA2, 'M', 'ὢι'),
    (0x1FA3, 'M', 'ὣι'),
    (0x1FA4, 'M', 'ὤι'),
    (0x1FA5, 'M', 'ὥι'),
    (0x1FA6, 'M', 'ὦι'),
    (0x1FA7, 'M', 'ὧι'),
    (0x1FA8, 'M', 'ὠι'),
    (0x1FA9, 'M', 'ὡι'),
    (0x1FAA, 'M', 'ὢι'),
    (0x1FAB, 'M', 'ὣι'),
    (0x1FAC, 'M', 'ὤι'),
    (0x1FAD, 'M', 'ὥι'),
    (0x1FAE, 'M', 'ὦι'),
    (0x1FAF, 'M', 'ὧι'),
    (0x1FB0, 'V'),
    (0x1FB2, 'M', 'ὰι'),
    (0x1FB3, 'M', 'αι'),
    (0x1FB4, 'M', 'άι'),
    (0x1FB5, 'X'),
    (0x1FB6, 'V'),
    (0x1FB7, 'M', 'ᾶι'),
    (0x1FB8, 'M', 'ᾰ'),
    (0x1FB9, 'M', 'ᾱ'),
    (0x1FBA, 'M', 'ὰ'),
    (0x1FBB, 'M', 'ά'),
    (0x1FBC, 'M', 'αι'),
    (0x1FBD, '3', ' ̓'),
    (0x1FBE, 'M', 'ι'),
    (0x1FBF, '3', ' ̓'),
    (0x1FC0, '3', ' ͂'),
    (0x1FC1, '3', ' ̈͂'),
    (0x1FC2, 'M', 'ὴι'),
    (0x1FC3, 'M', 'ηι'),
    (0x1FC4, 'M', 'ήι'),
    (0x1FC5, 'X'),
    (0x1FC6, 'V'),
    (0x1FC7, 'M', 'ῆι'),
    (0x1FC8, 'M', 'ὲ'),
    (0x1FC9, 'M', 'έ'),
    (0x1FCA, 'M', 'ὴ'),
    (0x1FCB, 'M', 'ή'),
    (0x1FCC, 'M', 'ηι'),
    (0x1FCD, '3', ' ̓̀'),
    (0x1FCE, '3', ' ̓́'),
    (0x1FCF, '3', ' ̓͂'),
    (0x1FD0, 'V'),
    (0x1FD3, 'M', 'ΐ'),
    (0x1FD4, 'X'),
    (0x1FD6, 'V'),
    (0x1FD8, 'M', 'ῐ'),
    (0x1FD9, 'M', 'ῑ'),
    (0x1FDA, 'M', 'ὶ'),
    (0x1FDB, 'M', 'ί'),
    (0x1FDC, 'X'),
    (0x1FDD, '3', ' ̔̀'),
    (0x1FDE, '3', ' ̔́'),
    (0x1FDF, '3', ' ̔͂'),
    (0x1FE0, 'V'),
    (0x1FE3, 'M', 'ΰ'),
    (0x1FE4, 'V'),
    (0x1FE8, 'M', 'ῠ'),
    (0x1FE9, 'M', 'ῡ'),
    (0x1FEA, 'M', 'ὺ'),
    (0x1FEB, 'M', 'ύ'),
    (0x1FEC, 'M', 'ῥ'),
    (0x1FED, '3', ' ̈̀'),
    (0x1FEE, '3', ' ̈́'),
    (0x1FEF, '3', '`'),
    (0x1FF0, 'X'),
    (0x1FF2, 'M', 'ὼι'),
    (0x1FF3, 'M', 'ωι'),
    ]

def _seg_21():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1FF4, 'M', 'ώι'),
    (0x1FF5, 'X'),
    (0x1FF6, 'V'),
    (0x1FF7, 'M', 'ῶι'),
    (0x1FF8, 'M', 'ὸ'),
    (0x1FF9, 'M', 'ό'),
    (0x1FFA, 'M', 'ὼ'),
    (0x1FFB, 'M', 'ώ'),
    (0x1FFC, 'M', 'ωι'),
    (0x1FFD, '3', ' ́'),
    (0x1FFE, '3', ' ̔'),
    (0x1FFF, 'X'),
    (0x2000, '3', ' '),
    (0x200B, 'I'),
    (0x200C, 'D', ''),
    (0x200E, 'X'),
    (0x2010, 'V'),
    (0x2011, 'M', '‐'),
    (0x2012, 'V'),
    (0x2017, '3', ' ̳'),
    (0x2018, 'V'),
    (0x2024, 'X'),
    (0x2027, 'V'),
    (0x2028, 'X'),
    (0x202F, '3', ' '),
    (0x2030, 'V'),
    (0x2033, 'M', '′′'),
    (0x2034, 'M', '′′′'),
    (0x2035, 'V'),
    (0x2036, 'M', '‵‵'),
    (0x2037, 'M', '‵‵‵'),
    (0x2038, 'V'),
    (0x203C, '3', '!!'),
    (0x203D, 'V'),
    (0x203E, '3', ' ̅'),
    (0x203F, 'V'),
    (0x2047, '3', '??'),
    (0x2048, '3', '?!'),
    (0x2049, '3', '!?'),
    (0x204A, 'V'),
    (0x2057, 'M', '′′′′'),
    (0x2058, 'V'),
    (0x205F, '3', ' '),
    (0x2060, 'I'),
    (0x2061, 'X'),
    (0x2064, 'I'),
    (0x2065, 'X'),
    (0x2070, 'M', '0'),
    (0x2071, 'M', 'i'),
    (0x2072, 'X'),
    (0x2074, 'M', '4'),
    (0x2075, 'M', '5'),
    (0x2076, 'M', '6'),
    (0x2077, 'M', '7'),
    (0x2078, 'M', '8'),
    (0x2079, 'M', '9'),
    (0x207A, '3', '+'),
    (0x207B, 'M', '−'),
    (0x207C, '3', '='),
    (0x207D, '3', '('),
    (0x207E, '3', ')'),
    (0x207F, 'M', 'n'),
    (0x2080, 'M', '0'),
    (0x2081, 'M', '1'),
    (0x2082, 'M', '2'),
    (0x2083, 'M', '3'),
    (0x2084, 'M', '4'),
    (0x2085, 'M', '5'),
    (0x2086, 'M', '6'),
    (0x2087, 'M', '7'),
    (0x2088, 'M', '8'),
    (0x2089, 'M', '9'),
    (0x208A, '3', '+'),
    (0x208B, 'M', '−'),
    (0x208C, '3', '='),
    (0x208D, '3', '('),
    (0x208E, '3', ')'),
    (0x208F, 'X'),
    (0x2090, 'M', 'a'),
    (0x2091, 'M', 'e'),
    (0x2092, 'M', 'o'),
    (0x2093, 'M', 'x'),
    (0x2094, 'M', 'ə'),
    (0x2095, 'M', 'h'),
    (0x2096, 'M', 'k'),
    (0x2097, 'M', 'l'),
    (0x2098, 'M', 'm'),
    (0x2099, 'M', 'n'),
    (0x209A, 'M', 'p'),
    (0x209B, 'M', 's'),
    (0x209C, 'M', 't'),
    (0x209D, 'X'),
    (0x20A0, 'V'),
    (0x20A8, 'M', 'rs'),
    (0x20A9, 'V'),
    (0x20C0, 'X'),
    (0x20D0, 'V'),
    (0x20F1, 'X'),
    (0x2100, '3', 'a/c'),
    (0x2101, '3', 'a/s'),
    ]

def _seg_22():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2102, 'M', 'c'),
    (0x2103, 'M', '°c'),
    (0x2104, 'V'),
    (0x2105, '3', 'c/o'),
    (0x2106, '3', 'c/u'),
    (0x2107, 'M', 'ɛ'),
    (0x2108, 'V'),
    (0x2109, 'M', '°f'),
    (0x210A, 'M', 'g'),
    (0x210B, 'M', 'h'),
    (0x210F, 'M', 'ħ'),
    (0x2110, 'M', 'i'),
    (0x2112, 'M', 'l'),
    (0x2114, 'V'),
    (0x2115, 'M', 'n'),
    (0x2116, 'M', 'no'),
    (0x2117, 'V'),
    (0x2119, 'M', 'p'),
    (0x211A, 'M', 'q'),
    (0x211B, 'M', 'r'),
    (0x211E, 'V'),
    (0x2120, 'M', 'sm'),
    (0x2121, 'M', 'tel'),
    (0x2122, 'M', 'tm'),
    (0x2123, 'V'),
    (0x2124, 'M', 'z'),
    (0x2125, 'V'),
    (0x2126, 'M', 'ω'),
    (0x2127, 'V'),
    (0x2128, 'M', 'z'),
    (0x2129, 'V'),
    (0x212A, 'M', 'k'),
    (0x212B, 'M', 'å'),
    (0x212C, 'M', 'b'),
    (0x212D, 'M', 'c'),
    (0x212E, 'V'),
    (0x212F, 'M', 'e'),
    (0x2131, 'M', 'f'),
    (0x2132, 'X'),
    (0x2133, 'M', 'm'),
    (0x2134, 'M', 'o'),
    (0x2135, 'M', 'א'),
    (0x2136, 'M', 'ב'),
    (0x2137, 'M', 'ג'),
    (0x2138, 'M', 'ד'),
    (0x2139, 'M', 'i'),
    (0x213A, 'V'),
    (0x213B, 'M', 'fax'),
    (0x213C, 'M', 'π'),
    (0x213D, 'M', 'γ'),
    (0x213F, 'M', 'π'),
    (0x2140, 'M', '∑'),
    (0x2141, 'V'),
    (0x2145, 'M', 'd'),
    (0x2147, 'M', 'e'),
    (0x2148, 'M', 'i'),
    (0x2149, 'M', 'j'),
    (0x214A, 'V'),
    (0x2150, 'M', '1⁄7'),
    (0x2151, 'M', '1⁄9'),
    (0x2152, 'M', '1⁄10'),
    (0x2153, 'M', '1⁄3'),
    (0x2154, 'M', '2⁄3'),
    (0x2155, 'M', '1⁄5'),
    (0x2156, 'M', '2⁄5'),
    (0x2157, 'M', '3⁄5'),
    (0x2158, 'M', '4⁄5'),
    (0x2159, 'M', '1⁄6'),
    (0x215A, 'M', '5⁄6'),
    (0x215B, 'M', '1⁄8'),
    (0x215C, 'M', '3⁄8'),
    (0x215D, 'M', '5⁄8'),
    (0x215E, 'M', '7⁄8'),
    (0x215F, 'M', '1⁄'),
    (0x2160, 'M', 'i'),
    (0x2161, 'M', 'ii'),
    (0x2162, 'M', 'iii'),
    (0x2163, 'M', 'iv'),
    (0x2164, 'M', 'v'),
    (0x2165, 'M', 'vi'),
    (0x2166, 'M', 'vii'),
    (0x2167, 'M', 'viii'),
    (0x2168, 'M', 'ix'),
    (0x2169, 'M', 'x'),
    (0x216A, 'M', 'xi'),
    (0x216B, 'M', 'xii'),
    (0x216C, 'M', 'l'),
    (0x216D, 'M', 'c'),
    (0x216E, 'M', 'd'),
    (0x216F, 'M', 'm'),
    (0x2170, 'M', 'i'),
    (0x2171, 'M', 'ii'),
    (0x2172, 'M', 'iii'),
    (0x2173, 'M', 'iv'),
    (0x2174, 'M', 'v'),
    (0x2175, 'M', 'vi'),
    (0x2176, 'M', 'vii'),
    (0x2177, 'M', 'viii'),
    (0x2178, 'M', 'ix'),
    (0x2179, 'M', 'x'),
    ]

def _seg_23():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x217A, 'M', 'xi'),
    (0x217B, 'M', 'xii'),
    (0x217C, 'M', 'l'),
    (0x217D, 'M', 'c'),
    (0x217E, 'M', 'd'),
    (0x217F, 'M', 'm'),
    (0x2180, 'V'),
    (0x2183, 'X'),
    (0x2184, 'V'),
    (0x2189, 'M', '0⁄3'),
    (0x218A, 'V'),
    (0x218C, 'X'),
    (0x2190, 'V'),
    (0x222C, 'M', '∫∫'),
    (0x222D, 'M', '∫∫∫'),
    (0x222E, 'V'),
    (0x222F, 'M', '∮∮'),
    (0x2230, 'M', '∮∮∮'),
    (0x2231, 'V'),
    (0x2260, '3'),
    (0x2261, 'V'),
    (0x226E, '3'),
    (0x2270, 'V'),
    (0x2329, 'M', '〈'),
    (0x232A, 'M', '〉'),
    (0x232B, 'V'),
    (0x2427, 'X'),
    (0x2440, 'V'),
    (0x244B, 'X'),
    (0x2460, 'M', '1'),
    (0x2461, 'M', '2'),
    (0x2462, 'M', '3'),
    (0x2463, 'M', '4'),
    (0x2464, 'M', '5'),
    (0x2465, 'M', '6'),
    (0x2466, 'M', '7'),
    (0x2467, 'M', '8'),
    (0x2468, 'M', '9'),
    (0x2469, 'M', '10'),
    (0x246A, 'M', '11'),
    (0x246B, 'M', '12'),
    (0x246C, 'M', '13'),
    (0x246D, 'M', '14'),
    (0x246E, 'M', '15'),
    (0x246F, 'M', '16'),
    (0x2470, 'M', '17'),
    (0x2471, 'M', '18'),
    (0x2472, 'M', '19'),
    (0x2473, 'M', '20'),
    (0x2474, '3', '(1)'),
    (0x2475, '3', '(2)'),
    (0x2476, '3', '(3)'),
    (0x2477, '3', '(4)'),
    (0x2478, '3', '(5)'),
    (0x2479, '3', '(6)'),
    (0x247A, '3', '(7)'),
    (0x247B, '3', '(8)'),
    (0x247C, '3', '(9)'),
    (0x247D, '3', '(10)'),
    (0x247E, '3', '(11)'),
    (0x247F, '3', '(12)'),
    (0x2480, '3', '(13)'),
    (0x2481, '3', '(14)'),
    (0x2482, '3', '(15)'),
    (0x2483, '3', '(16)'),
    (0x2484, '3', '(17)'),
    (0x2485, '3', '(18)'),
    (0x2486, '3', '(19)'),
    (0x2487, '3', '(20)'),
    (0x2488, 'X'),
    (0x249C, '3', '(a)'),
    (0x249D, '3', '(b)'),
    (0x249E, '3', '(c)'),
    (0x249F, '3', '(d)'),
    (0x24A0, '3', '(e)'),
    (0x24A1, '3', '(f)'),
    (0x24A2, '3', '(g)'),
    (0x24A3, '3', '(h)'),
    (0x24A4, '3', '(i)'),
    (0x24A5, '3', '(j)'),
    (0x24A6, '3', '(k)'),
    (0x24A7, '3', '(l)'),
    (0x24A8, '3', '(m)'),
    (0x24A9, '3', '(n)'),
    (0x24AA, '3', '(o)'),
    (0x24AB, '3', '(p)'),
    (0x24AC, '3', '(q)'),
    (0x24AD, '3', '(r)'),
    (0x24AE, '3', '(s)'),
    (0x24AF, '3', '(t)'),
    (0x24B0, '3', '(u)'),
    (0x24B1, '3', '(v)'),
    (0x24B2, '3', '(w)'),
    (0x24B3, '3', '(x)'),
    (0x24B4, '3', '(y)'),
    (0x24B5, '3', '(z)'),
    (0x24B6, 'M', 'a'),
    (0x24B7, 'M', 'b'),
    (0x24B8, 'M', 'c'),
    (0x24B9, 'M', 'd'),
    ]

def _seg_24():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x24BA, 'M', 'e'),
    (0x24BB, 'M', 'f'),
    (0x24BC, 'M', 'g'),
    (0x24BD, 'M', 'h'),
    (0x24BE, 'M', 'i'),
    (0x24BF, 'M', 'j'),
    (0x24C0, 'M', 'k'),
    (0x24C1, 'M', 'l'),
    (0x24C2, 'M', 'm'),
    (0x24C3, 'M', 'n'),
    (0x24C4, 'M', 'o'),
    (0x24C5, 'M', 'p'),
    (0x24C6, 'M', 'q'),
    (0x24C7, 'M', 'r'),
    (0x24C8, 'M', 's'),
    (0x24C9, 'M', 't'),
    (0x24CA, 'M', 'u'),
    (0x24CB, 'M', 'v'),
    (0x24CC, 'M', 'w'),
    (0x24CD, 'M', 'x'),
    (0x24CE, 'M', 'y'),
    (0x24CF, 'M', 'z'),
    (0x24D0, 'M', 'a'),
    (0x24D1, 'M', 'b'),
    (0x24D2, 'M', 'c'),
    (0x24D3, 'M', 'd'),
    (0x24D4, 'M', 'e'),
    (0x24D5, 'M', 'f'),
    (0x24D6, 'M', 'g'),
    (0x24D7, 'M', 'h'),
    (0x24D8, 'M', 'i'),
    (0x24D9, 'M', 'j'),
    (0x24DA, 'M', 'k'),
    (0x24DB, 'M', 'l'),
    (0x24DC, 'M', 'm'),
    (0x24DD, 'M', 'n'),
    (0x24DE, 'M', 'o'),
    (0x24DF, 'M', 'p'),
    (0x24E0, 'M', 'q'),
    (0x24E1, 'M', 'r'),
    (0x24E2, 'M', 's'),
    (0x24E3, 'M', 't'),
    (0x24E4, 'M', 'u'),
    (0x24E5, 'M', 'v'),
    (0x24E6, 'M', 'w'),
    (0x24E7, 'M', 'x'),
    (0x24E8, 'M', 'y'),
    (0x24E9, 'M', 'z'),
    (0x24EA, 'M', '0'),
    (0x24EB, 'V'),
    (0x2A0C, 'M', '∫∫∫∫'),
    (0x2A0D, 'V'),
    (0x2A74, '3', '::='),
    (0x2A75, '3', '=='),
    (0x2A76, '3', '==='),
    (0x2A77, 'V'),
    (0x2ADC, 'M', '⫝̸'),
    (0x2ADD, 'V'),
    (0x2B74, 'X'),
    (0x2B76, 'V'),
    (0x2B96, 'X'),
    (0x2B97, 'V'),
    (0x2C00, 'M', 'ⰰ'),
    (0x2C01, 'M', 'ⰱ'),
    (0x2C02, 'M', 'ⰲ'),
    (0x2C03, 'M', 'ⰳ'),
    (0x2C04, 'M', 'ⰴ'),
    (0x2C05, 'M', 'ⰵ'),
    (0x2C06, 'M', 'ⰶ'),
    (0x2C07, 'M', 'ⰷ'),
    (0x2C08, 'M', 'ⰸ'),
    (0x2C09, 'M', 'ⰹ'),
    (0x2C0A, 'M', 'ⰺ'),
    (0x2C0B, 'M', 'ⰻ'),
    (0x2C0C, 'M', 'ⰼ'),
    (0x2C0D, 'M', 'ⰽ'),
    (0x2C0E, 'M', 'ⰾ'),
    (0x2C0F, 'M', 'ⰿ'),
    (0x2C10, 'M', 'ⱀ'),
    (0x2C11, 'M', 'ⱁ'),
    (0x2C12, 'M', 'ⱂ'),
    (0x2C13, 'M', 'ⱃ'),
    (0x2C14, 'M', 'ⱄ'),
    (0x2C15, 'M', 'ⱅ'),
    (0x2C16, 'M', 'ⱆ'),
    (0x2C17, 'M', 'ⱇ'),
    (0x2C18, 'M', 'ⱈ'),
    (0x2C19, 'M', 'ⱉ'),
    (0x2C1A, 'M', 'ⱊ'),
    (0x2C1B, 'M', 'ⱋ'),
    (0x2C1C, 'M', 'ⱌ'),
    (0x2C1D, 'M', 'ⱍ'),
    (0x2C1E, 'M', 'ⱎ'),
    (0x2C1F, 'M', 'ⱏ'),
    (0x2C20, 'M', 'ⱐ'),
    (0x2C21, 'M', 'ⱑ'),
    (0x2C22, 'M', 'ⱒ'),
    (0x2C23, 'M', 'ⱓ'),
    (0x2C24, 'M', 'ⱔ'),
    (0x2C25, 'M', 'ⱕ'),
    ]

def _seg_25():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2C26, 'M', 'ⱖ'),
    (0x2C27, 'M', 'ⱗ'),
    (0x2C28, 'M', 'ⱘ'),
    (0x2C29, 'M', 'ⱙ'),
    (0x2C2A, 'M', 'ⱚ'),
    (0x2C2B, 'M', 'ⱛ'),
    (0x2C2C, 'M', 'ⱜ'),
    (0x2C2D, 'M', 'ⱝ'),
    (0x2C2E, 'M', 'ⱞ'),
    (0x2C2F, 'X'),
    (0x2C30, 'V'),
    (0x2C5F, 'X'),
    (0x2C60, 'M', 'ⱡ'),
    (0x2C61, 'V'),
    (0x2C62, 'M', 'ɫ'),
    (0x2C63, 'M', 'ᵽ'),
    (0x2C64, 'M', 'ɽ'),
    (0x2C65, 'V'),
    (0x2C67, 'M', 'ⱨ'),
    (0x2C68, 'V'),
    (0x2C69, 'M', 'ⱪ'),
    (0x2C6A, 'V'),
    (0x2C6B, 'M', 'ⱬ'),
    (0x2C6C, 'V'),
    (0x2C6D, 'M', 'ɑ'),
    (0x2C6E, 'M', 'ɱ'),
    (0x2C6F, 'M', 'ɐ'),
    (0x2C70, 'M', 'ɒ'),
    (0x2C71, 'V'),
    (0x2C72, 'M', 'ⱳ'),
    (0x2C73, 'V'),
    (0x2C75, 'M', 'ⱶ'),
    (0x2C76, 'V'),
    (0x2C7C, 'M', 'j'),
    (0x2C7D, 'M', 'v'),
    (0x2C7E, 'M', 'ȿ'),
    (0x2C7F, 'M', 'ɀ'),
    (0x2C80, 'M', 'ⲁ'),
    (0x2C81, 'V'),
    (0x2C82, 'M', 'ⲃ'),
    (0x2C83, 'V'),
    (0x2C84, 'M', 'ⲅ'),
    (0x2C85, 'V'),
    (0x2C86, 'M', 'ⲇ'),
    (0x2C87, 'V'),
    (0x2C88, 'M', 'ⲉ'),
    (0x2C89, 'V'),
    (0x2C8A, 'M', 'ⲋ'),
    (0x2C8B, 'V'),
    (0x2C8C, 'M', 'ⲍ'),
    (0x2C8D, 'V'),
    (0x2C8E, 'M', 'ⲏ'),
    (0x2C8F, 'V'),
    (0x2C90, 'M', 'ⲑ'),
    (0x2C91, 'V'),
    (0x2C92, 'M', 'ⲓ'),
    (0x2C93, 'V'),
    (0x2C94, 'M', 'ⲕ'),
    (0x2C95, 'V'),
    (0x2C96, 'M', 'ⲗ'),
    (0x2C97, 'V'),
    (0x2C98, 'M', 'ⲙ'),
    (0x2C99, 'V'),
    (0x2C9A, 'M', 'ⲛ'),
    (0x2C9B, 'V'),
    (0x2C9C, 'M', 'ⲝ'),
    (0x2C9D, 'V'),
    (0x2C9E, 'M', 'ⲟ'),
    (0x2C9F, 'V'),
    (0x2CA0, 'M', 'ⲡ'),
    (0x2CA1, 'V'),
    (0x2CA2, 'M', 'ⲣ'),
    (0x2CA3, 'V'),
    (0x2CA4, 'M', 'ⲥ'),
    (0x2CA5, 'V'),
    (0x2CA6, 'M', 'ⲧ'),
    (0x2CA7, 'V'),
    (0x2CA8, 'M', 'ⲩ'),
    (0x2CA9, 'V'),
    (0x2CAA, 'M', 'ⲫ'),
    (0x2CAB, 'V'),
    (0x2CAC, 'M', 'ⲭ'),
    (0x2CAD, 'V'),
    (0x2CAE, 'M', 'ⲯ'),
    (0x2CAF, 'V'),
    (0x2CB0, 'M', 'ⲱ'),
    (0x2CB1, 'V'),
    (0x2CB2, 'M', 'ⲳ'),
    (0x2CB3, 'V'),
    (0x2CB4, 'M', 'ⲵ'),
    (0x2CB5, 'V'),
    (0x2CB6, 'M', 'ⲷ'),
    (0x2CB7, 'V'),
    (0x2CB8, 'M', 'ⲹ'),
    (0x2CB9, 'V'),
    (0x2CBA, 'M', 'ⲻ'),
    (0x2CBB, 'V'),
    (0x2CBC, 'M', 'ⲽ'),
    (0x2CBD, 'V'),
    (0x2CBE, 'M', 'ⲿ'),
    ]

def _seg_26():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2CBF, 'V'),
    (0x2CC0, 'M', 'ⳁ'),
    (0x2CC1, 'V'),
    (0x2CC2, 'M', 'ⳃ'),
    (0x2CC3, 'V'),
    (0x2CC4, 'M', 'ⳅ'),
    (0x2CC5, 'V'),
    (0x2CC6, 'M', 'ⳇ'),
    (0x2CC7, 'V'),
    (0x2CC8, 'M', 'ⳉ'),
    (0x2CC9, 'V'),
    (0x2CCA, 'M', 'ⳋ'),
    (0x2CCB, 'V'),
    (0x2CCC, 'M', 'ⳍ'),
    (0x2CCD, 'V'),
    (0x2CCE, 'M', 'ⳏ'),
    (0x2CCF, 'V'),
    (0x2CD0, 'M', 'ⳑ'),
    (0x2CD1, 'V'),
    (0x2CD2, 'M', 'ⳓ'),
    (0x2CD3, 'V'),
    (0x2CD4, 'M', 'ⳕ'),
    (0x2CD5, 'V'),
    (0x2CD6, 'M', 'ⳗ'),
    (0x2CD7, 'V'),
    (0x2CD8, 'M', 'ⳙ'),
    (0x2CD9, 'V'),
    (0x2CDA, 'M', 'ⳛ'),
    (0x2CDB, 'V'),
    (0x2CDC, 'M', 'ⳝ'),
    (0x2CDD, 'V'),
    (0x2CDE, 'M', 'ⳟ'),
    (0x2CDF, 'V'),
    (0x2CE0, 'M', 'ⳡ'),
    (0x2CE1, 'V'),
    (0x2CE2, 'M', 'ⳣ'),
    (0x2CE3, 'V'),
    (0x2CEB, 'M', 'ⳬ'),
    (0x2CEC, 'V'),
    (0x2CED, 'M', 'ⳮ'),
    (0x2CEE, 'V'),
    (0x2CF2, 'M', 'ⳳ'),
    (0x2CF3, 'V'),
    (0x2CF4, 'X'),
    (0x2CF9, 'V'),
    (0x2D26, 'X'),
    (0x2D27, 'V'),
    (0x2D28, 'X'),
    (0x2D2D, 'V'),
    (0x2D2E, 'X'),
    (0x2D30, 'V'),
    (0x2D68, 'X'),
    (0x2D6F, 'M', 'ⵡ'),
    (0x2D70, 'V'),
    (0x2D71, 'X'),
    (0x2D7F, 'V'),
    (0x2D97, 'X'),
    (0x2DA0, 'V'),
    (0x2DA7, 'X'),
    (0x2DA8, 'V'),
    (0x2DAF, 'X'),
    (0x2DB0, 'V'),
    (0x2DB7, 'X'),
    (0x2DB8, 'V'),
    (0x2DBF, 'X'),
    (0x2DC0, 'V'),
    (0x2DC7, 'X'),
    (0x2DC8, 'V'),
    (0x2DCF, 'X'),
    (0x2DD0, 'V'),
    (0x2DD7, 'X'),
    (0x2DD8, 'V'),
    (0x2DDF, 'X'),
    (0x2DE0, 'V'),
    (0x2E53, 'X'),
    (0x2E80, 'V'),
    (0x2E9A, 'X'),
    (0x2E9B, 'V'),
    (0x2E9F, 'M', '母'),
    (0x2EA0, 'V'),
    (0x2EF3, 'M', '龟'),
    (0x2EF4, 'X'),
    (0x2F00, 'M', '一'),
    (0x2F01, 'M', '丨'),
    (0x2F02, 'M', '丶'),
    (0x2F03, 'M', '丿'),
    (0x2F04, 'M', '乙'),
    (0x2F05, 'M', '亅'),
    (0x2F06, 'M', '二'),
    (0x2F07, 'M', '亠'),
    (0x2F08, 'M', '人'),
    (0x2F09, 'M', '儿'),
    (0x2F0A, 'M', '入'),
    (0x2F0B, 'M', '八'),
    (0x2F0C, 'M', '冂'),
    (0x2F0D, 'M', '冖'),
    (0x2F0E, 'M', '冫'),
    (0x2F0F, 'M', '几'),
    (0x2F10, 'M', '凵'),
    (0x2F11, 'M', '刀'),
    ]

def _seg_27():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F12, 'M', '力'),
    (0x2F13, 'M', '勹'),
    (0x2F14, 'M', '匕'),
    (0x2F15, 'M', '匚'),
    (0x2F16, 'M', '匸'),
    (0x2F17, 'M', '十'),
    (0x2F18, 'M', '卜'),
    (0x2F19, 'M', '卩'),
    (0x2F1A, 'M', '厂'),
    (0x2F1B, 'M', '厶'),
    (0x2F1C, 'M', '又'),
    (0x2F1D, 'M', '口'),
    (0x2F1E, 'M', '囗'),
    (0x2F1F, 'M', '土'),
    (0x2F20, 'M', '士'),
    (0x2F21, 'M', '夂'),
    (0x2F22, 'M', '夊'),
    (0x2F23, 'M', '夕'),
    (0x2F24, 'M', '大'),
    (0x2F25, 'M', '女'),
    (0x2F26, 'M', '子'),
    (0x2F27, 'M', '宀'),
    (0x2F28, 'M', '寸'),
    (0x2F29, 'M', '小'),
    (0x2F2A, 'M', '尢'),
    (0x2F2B, 'M', '尸'),
    (0x2F2C, 'M', '屮'),
    (0x2F2D, 'M', '山'),
    (0x2F2E, 'M', '巛'),
    (0x2F2F, 'M', '工'),
    (0x2F30, 'M', '己'),
    (0x2F31, 'M', '巾'),
    (0x2F32, 'M', '干'),
    (0x2F33, 'M', '幺'),
    (0x2F34, 'M', '广'),
    (0x2F35, 'M', '廴'),
    (0x2F36, 'M', '廾'),
    (0x2F37, 'M', '弋'),
    (0x2F38, 'M', '弓'),
    (0x2F39, 'M', '彐'),
    (0x2F3A, 'M', '彡'),
    (0x2F3B, 'M', '彳'),
    (0x2F3C, 'M', '心'),
    (0x2F3D, 'M', '戈'),
    (0x2F3E, 'M', '戶'),
    (0x2F3F, 'M', '手'),
    (0x2F40, 'M', '支'),
    (0x2F41, 'M', '攴'),
    (0x2F42, 'M', '文'),
    (0x2F43, 'M', '斗'),
    (0x2F44, 'M', '斤'),
    (0x2F45, 'M', '方'),
    (0x2F46, 'M', '无'),
    (0x2F47, 'M', '日'),
    (0x2F48, 'M', '曰'),
    (0x2F49, 'M', '月'),
    (0x2F4A, 'M', '木'),
    (0x2F4B, 'M', '欠'),
    (0x2F4C, 'M', '止'),
    (0x2F4D, 'M', '歹'),
    (0x2F4E, 'M', '殳'),
    (0x2F4F, 'M', '毋'),
    (0x2F50, 'M', '比'),
    (0x2F51, 'M', '毛'),
    (0x2F52, 'M', '氏'),
    (0x2F53, 'M', '气'),
    (0x2F54, 'M', '水'),
    (0x2F55, 'M', '火'),
    (0x2F56, 'M', '爪'),
    (0x2F57, 'M', '父'),
    (0x2F58, 'M', '爻'),
    (0x2F59, 'M', '爿'),
    (0x2F5A, 'M', '片'),
    (0x2F5B, 'M', '牙'),
    (0x2F5C, 'M', '牛'),
    (0x2F5D, 'M', '犬'),
    (0x2F5E, 'M', '玄'),
    (0x2F5F, 'M', '玉'),
    (0x2F60, 'M', '瓜'),
    (0x2F61, 'M', '瓦'),
    (0x2F62, 'M', '甘'),
    (0x2F63, 'M', '生'),
    (0x2F64, 'M', '用'),
    (0x2F65, 'M', '田'),
    (0x2F66, 'M', '疋'),
    (0x2F67, 'M', '疒'),
    (0x2F68, 'M', '癶'),
    (0x2F69, 'M', '白'),
    (0x2F6A, 'M', '皮'),
    (0x2F6B, 'M', '皿'),
    (0x2F6C, 'M', '目'),
    (0x2F6D, 'M', '矛'),
    (0x2F6E, 'M', '矢'),
    (0x2F6F, 'M', '石'),
    (0x2F70, 'M', '示'),
    (0x2F71, 'M', '禸'),
    (0x2F72, 'M', '禾'),
    (0x2F73, 'M', '穴'),
    (0x2F74, 'M', '立'),
    (0x2F75, 'M', '竹'),
    ]

def _seg_28():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F76, 'M', '米'),
    (0x2F77, 'M', '糸'),
    (0x2F78, 'M', '缶'),
    (0x2F79, 'M', '网'),
    (0x2F7A, 'M', '羊'),
    (0x2F7B, 'M', '羽'),
    (0x2F7C, 'M', '老'),
    (0x2F7D, 'M', '而'),
    (0x2F7E, 'M', '耒'),
    (0x2F7F, 'M', '耳'),
    (0x2F80, 'M', '聿'),
    (0x2F81, 'M', '肉'),
    (0x2F82, 'M', '臣'),
    (0x2F83, 'M', '自'),
    (0x2F84, 'M', '至'),
    (0x2F85, 'M', '臼'),
    (0x2F86, 'M', '舌'),
    (0x2F87, 'M', '舛'),
    (0x2F88, 'M', '舟'),
    (0x2F89, 'M', '艮'),
    (0x2F8A, 'M', '色'),
    (0x2F8B, 'M', '艸'),
    (0x2F8C, 'M', '虍'),
    (0x2F8D, 'M', '虫'),
    (0x2F8E, 'M', '血'),
    (0x2F8F, 'M', '行'),
    (0x2F90, 'M', '衣'),
    (0x2F91, 'M', '襾'),
    (0x2F92, 'M', '見'),
    (0x2F93, 'M', '角'),
    (0x2F94, 'M', '言'),
    (0x2F95, 'M', '谷'),
    (0x2F96, 'M', '豆'),
    (0x2F97, 'M', '豕'),
    (0x2F98, 'M', '豸'),
    (0x2F99, 'M', '貝'),
    (0x2F9A, 'M', '赤'),
    (0x2F9B, 'M', '走'),
    (0x2F9C, 'M', '足'),
    (0x2F9D, 'M', '身'),
    (0x2F9E, 'M', '車'),
    (0x2F9F, 'M', '辛'),
    (0x2FA0, 'M', '辰'),
    (0x2FA1, 'M', '辵'),
    (0x2FA2, 'M', '邑'),
    (0x2FA3, 'M', '酉'),
    (0x2FA4, 'M', '釆'),
    (0x2FA5, 'M', '里'),
    (0x2FA6, 'M', '金'),
    (0x2FA7, 'M', '長'),
    (0x2FA8, 'M', '門'),
    (0x2FA9, 'M', '阜'),
    (0x2FAA, 'M', '隶'),
    (0x2FAB, 'M', '隹'),
    (0x2FAC, 'M', '雨'),
    (0x2FAD, 'M', '靑'),
    (0x2FAE, 'M', '非'),
    (0x2FAF, 'M', '面'),
    (0x2FB0, 'M', '革'),
    (0x2FB1, 'M', '韋'),
    (0x2FB2, 'M', '韭'),
    (0x2FB3, 'M', '音'),
    (0x2FB4, 'M', '頁'),
    (0x2FB5, 'M', '風'),
    (0x2FB6, 'M', '飛'),
    (0x2FB7, 'M', '食'),
    (0x2FB8, 'M', '首'),
    (0x2FB9, 'M', '香'),
    (0x2FBA, 'M', '馬'),
    (0x2FBB, 'M', '骨'),
    (0x2FBC, 'M', '高'),
    (0x2FBD, 'M', '髟'),
    (0x2FBE, 'M', '鬥'),
    (0x2FBF, 'M', '鬯'),
    (0x2FC0, 'M', '鬲'),
    (0x2FC1, 'M', '鬼'),
    (0x2FC2, 'M', '魚'),
    (0x2FC3, 'M', '鳥'),
    (0x2FC4, 'M', '鹵'),
    (0x2FC5, 'M', '鹿'),
    (0x2FC6, 'M', '麥'),
    (0x2FC7, 'M', '麻'),
    (0x2FC8, 'M', '黃'),
    (0x2FC9, 'M', '黍'),
    (0x2FCA, 'M', '黑'),
    (0x2FCB, 'M', '黹'),
    (0x2FCC, 'M', '黽'),
    (0x2FCD, 'M', '鼎'),
    (0x2FCE, 'M', '鼓'),
    (0x2FCF, 'M', '鼠'),
    (0x2FD0, 'M', '鼻'),
    (0x2FD1, 'M', '齊'),
    (0x2FD2, 'M', '齒'),
    (0x2FD3, 'M', '龍'),
    (0x2FD4, 'M', '龜'),
    (0x2FD5, 'M', '龠'),
    (0x2FD6, 'X'),
    (0x3000, '3', ' '),
    (0x3001, 'V'),
    (0x3002, 'M', '.'),
    ]

def _seg_29():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x3003, 'V'),
    (0x3036, 'M', '〒'),
    (0x3037, 'V'),
    (0x3038, 'M', '十'),
    (0x3039, 'M', '卄'),
    (0x303A, 'M', '卅'),
    (0x303B, 'V'),
    (0x3040, 'X'),
    (0x3041, 'V'),
    (0x3097, 'X'),
    (0x3099, 'V'),
    (0x309B, '3', ' ゙'),
    (0x309C, '3', ' ゚'),
    (0x309D, 'V'),
    (0x309F, 'M', 'より'),
    (0x30A0, 'V'),
    (0x30FF, 'M', 'コト'),
    (0x3100, 'X'),
    (0x3105, 'V'),
    (0x3130, 'X'),
    (0x3131, 'M', 'ᄀ'),
    (0x3132, 'M', 'ᄁ'),
    (0x3133, 'M', 'ᆪ'),
    (0x3134, 'M', 'ᄂ'),
    (0x3135, 'M', 'ᆬ'),
    (0x3136, 'M', 'ᆭ'),
    (0x3137, 'M', 'ᄃ'),
    (0x3138, 'M', 'ᄄ'),
    (0x3139, 'M', 'ᄅ'),
    (0x313A, 'M', 'ᆰ'),
    (0x313B, 'M', 'ᆱ'),
    (0x313C, 'M', 'ᆲ'),
    (0x313D, 'M', 'ᆳ'),
    (0x313E, 'M', 'ᆴ'),
    (0x313F, 'M', 'ᆵ'),
    (0x3140, 'M', 'ᄚ'),
    (0x3141, 'M', 'ᄆ'),
    (0x3142, 'M', 'ᄇ'),
    (0x3143, 'M', 'ᄈ'),
    (0x3144, 'M', 'ᄡ'),
    (0x3145, 'M', 'ᄉ'),
    (0x3146, 'M', 'ᄊ'),
    (0x3147, 'M', 'ᄋ'),
    (0x3148, 'M', 'ᄌ'),
    (0x3149, 'M', 'ᄍ'),
    (0x314A, 'M', 'ᄎ'),
    (0x314B, 'M', 'ᄏ'),
    (0x314C, 'M', 'ᄐ'),
    (0x314D, 'M', 'ᄑ'),
    (0x314E, 'M', 'ᄒ'),
    (0x314F, 'M', 'ᅡ'),
    (0x3150, 'M', 'ᅢ'),
    (0x3151, 'M', 'ᅣ'),
    (0x3152, 'M', 'ᅤ'),
    (0x3153, 'M', 'ᅥ'),
    (0x3154, 'M', 'ᅦ'),
    (0x3155, 'M', 'ᅧ'),
    (0x3156, 'M', 'ᅨ'),
    (0x3157, 'M', 'ᅩ'),
    (0x3158, 'M', 'ᅪ'),
    (0x3159, 'M', 'ᅫ'),
    (0x315A, 'M', 'ᅬ'),
    (0x315B, 'M', 'ᅭ'),
    (0x315C, 'M', 'ᅮ'),
    (0x315D, 'M', 'ᅯ'),
    (0x315E, 'M', 'ᅰ'),
    (0x315F, 'M', 'ᅱ'),
    (0x3160, 'M', 'ᅲ'),
    (0x3161, 'M', 'ᅳ'),
    (0x3162, 'M', 'ᅴ'),
    (0x3163, 'M', 'ᅵ'),
    (0x3164, 'X'),
    (0x3165, 'M', 'ᄔ'),
    (0x3166, 'M', 'ᄕ'),
    (0x3167, 'M', 'ᇇ'),
    (0x3168, 'M', 'ᇈ'),
    (0x3169, 'M', 'ᇌ'),
    (0x316A, 'M', 'ᇎ'),
    (0x316B, 'M', 'ᇓ'),
    (0x316C, 'M', 'ᇗ'),
    (0x316D, 'M', 'ᇙ'),
    (0x316E, 'M', 'ᄜ'),
    (0x316F, 'M', 'ᇝ'),
    (0x3170, 'M', 'ᇟ'),
    (0x3171, 'M', 'ᄝ'),
    (0x3172, 'M', 'ᄞ'),
    (0x3173, 'M', 'ᄠ'),
    (0x3174, 'M', 'ᄢ'),
    (0x3175, 'M', 'ᄣ'),
    (0x3176, 'M', 'ᄧ'),
    (0x3177, 'M', 'ᄩ'),
    (0x3178, 'M', 'ᄫ'),
    (0x3179, 'M', 'ᄬ'),
    (0x317A, 'M', 'ᄭ'),
    (0x317B, 'M', 'ᄮ'),
    (0x317C, 'M', 'ᄯ'),
    (0x317D, 'M', 'ᄲ'),
    (0x317E, 'M', 'ᄶ'),
    (0x317F, 'M', 'ᅀ'),
    (0x3180, 'M', 'ᅇ'),
    ]

def _seg_30():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x3181, 'M', 'ᅌ'),
    (0x3182, 'M', 'ᇱ'),
    (0x3183, 'M', 'ᇲ'),
    (0x3184, 'M', 'ᅗ'),
    (0x3185, 'M', 'ᅘ'),
    (0x3186, 'M', 'ᅙ'),
    (0x3187, 'M', 'ᆄ'),
    (0x3188, 'M', 'ᆅ'),
    (0x3189, 'M', 'ᆈ'),
    (0x318A, 'M', 'ᆑ'),
    (0x318B, 'M', 'ᆒ'),
    (0x318C, 'M', 'ᆔ'),
    (0x318D, 'M', 'ᆞ'),
    (0x318E, 'M', 'ᆡ'),
    (0x318F, 'X'),
    (0x3190, 'V'),
    (0x3192, 'M', '一'),
    (0x3193, 'M', '二'),
    (0x3194, 'M', '三'),
    (0x3195, 'M', '四'),
    (0x3196, 'M', '上'),
    (0x3197, 'M', '中'),
    (0x3198, 'M', '下'),
    (0x3199, 'M', '甲'),
    (0x319A, 'M', '乙'),
    (0x319B, 'M', '丙'),
    (0x319C, 'M', '丁'),
    (0x319D, 'M', '天'),
    (0x319E, 'M', '地'),
    (0x319F, 'M', '人'),
    (0x31A0, 'V'),
    (0x31E4, 'X'),
    (0x31F0, 'V'),
    (0x3200, '3', '(ᄀ)'),
    (0x3201, '3', '(ᄂ)'),
    (0x3202, '3', '(ᄃ)'),
    (0x3203, '3', '(ᄅ)'),
    (0x3204, '3', '(ᄆ)'),
    (0x3205, '3', '(ᄇ)'),
    (0x3206, '3', '(ᄉ)'),
    (0x3207, '3', '(ᄋ)'),
    (0x3208, '3', '(ᄌ)'),
    (0x3209, '3', '(ᄎ)'),
    (0x320A, '3', '(ᄏ)'),
    (0x320B, '3', '(ᄐ)'),
    (0x320C, '3', '(ᄑ)'),
    (0x320D, '3', '(ᄒ)'),
    (0x320E, '3', '(가)'),
    (0x320F, '3', '(나)'),
    (0x3210, '3', '(다)'),
    (0x3211, '3', '(라)'),
    (0x3212, '3', '(마)'),
    (0x3213, '3', '(바)'),
    (0x3214, '3', '(사)'),
    (0x3215, '3', '(아)'),
    (0x3216, '3', '(자)'),
    (0x3217, '3', '(차)'),
    (0x3218, '3', '(카)'),
    (0x3219, '3', '(타)'),
    (0x321A, '3', '(파)'),
    (0x321B, '3', '(하)'),
    (0x321C, '3', '(주)'),
    (0x321D, '3', '(오전)'),
    (0x321E, '3', '(오후)'),
    (0x321F, 'X'),
    (0x3220, '3', '(一)'),
    (0x3221, '3', '(二)'),
    (0x3222, '3', '(三)'),
    (0x3223, '3', '(四)'),
    (0x3224, '3', '(五)'),
    (0x3225, '3', '(六)'),
    (0x3226, '3', '(七)'),
    (0x3227, '3', '(八)'),
    (0x3228, '3', '(九)'),
    (0x3229, '3', '(十)'),
    (0x322A, '3', '(月)'),
    (0x322B, '3', '(火)'),
    (0x322C, '3', '(水)'),
    (0x322D, '3', '(木)'),
    (0x322E, '3', '(金)'),
    (0x322F, '3', '(土)'),
    (0x3230, '3', '(日)'),
    (0x3231, '3', '(株)'),
    (0x3232, '3', '(有)'),
    (0x3233, '3', '(社)'),
    (0x3234, '3', '(名)'),
    (0x3235, '3', '(特)'),
    (0x3236, '3', '(財)'),
    (0x3237, '3', '(祝)'),
    (0x3238, '3', '(労)'),
    (0x3239, '3', '(代)'),
    (0x323A, '3', '(呼)'),
    (0x323B, '3', '(学)'),
    (0x323C, '3', '(監)'),
    (0x323D, '3', '(企)'),
    (0x323E, '3', '(資)'),
    (0x323F, '3', '(協)'),
    (0x3240, '3', '(祭)'),
    (0x3241, '3', '(休)'),
    (0x3242, '3', '(自)'),
    ]

def _seg_31():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x3243, '3', '(至)'),
    (0x3244, 'M', '問'),
    (0x3245, 'M', '幼'),
    (0x3246, 'M', '文'),
    (0x3247, 'M', '箏'),
    (0x3248, 'V'),
    (0x3250, 'M', 'pte'),
    (0x3251, 'M', '21'),
    (0x3252, 'M', '22'),
    (0x3253, 'M', '23'),
    (0x3254, 'M', '24'),
    (0x3255, 'M', '25'),
    (0x3256, 'M', '26'),
    (0x3257, 'M', '27'),
    (0x3258, 'M', '28'),
    (0x3259, 'M', '29'),
    (0x325A, 'M', '30'),
    (0x325B, 'M', '31'),
    (0x325C, 'M', '32'),
    (0x325D, 'M', '33'),
    (0x325E, 'M', '34'),
    (0x325F, 'M', '35'),
    (0x3260, 'M', 'ᄀ'),
    (0x3261, 'M', 'ᄂ'),
    (0x3262, 'M', 'ᄃ'),
    (0x3263, 'M', 'ᄅ'),
    (0x3264, 'M', 'ᄆ'),
    (0x3265, 'M', 'ᄇ'),
    (0x3266, 'M', 'ᄉ'),
    (0x3267, 'M', 'ᄋ'),
    (0x3268, 'M', 'ᄌ'),
    (0x3269, 'M', 'ᄎ'),
    (0x326A, 'M', 'ᄏ'),
    (0x326B, 'M', 'ᄐ'),
    (0x326C, 'M', 'ᄑ'),
    (0x326D, 'M', 'ᄒ'),
    (0x326E, 'M', '가'),
    (0x326F, 'M', '나'),
    (0x3270, 'M', '다'),
    (0x3271, 'M', '라'),
    (0x3272, 'M', '마'),
    (0x3273, 'M', '바'),
    (0x3274, 'M', '사'),
    (0x3275, 'M', '아'),
    (0x3276, 'M', '자'),
    (0x3277, 'M', '차'),
    (0x3278, 'M', '카'),
    (0x3279, 'M', '타'),
    (0x327A, 'M', '파'),
    (0x327B, 'M', '하'),
    (0x327C, 'M', '참고'),
    (0x327D, 'M', '주의'),
    (0x327E, 'M', '우'),
    (0x327F, 'V'),
    (0x3280, 'M', '一'),
    (0x3281, 'M', '二'),
    (0x3282, 'M', '三'),
    (0x3283, 'M', '四'),
    (0x3284, 'M', '五'),
    (0x3285, 'M', '六'),
    (0x3286, 'M', '七'),
    (0x3287, 'M', '八'),
    (0x3288, 'M', '九'),
    (0x3289, 'M', '十'),
    (0x328A, 'M', '月'),
    (0x328B, 'M', '火'),
    (0x328C, 'M', '水'),
    (0x328D, 'M', '木'),
    (0x328E, 'M', '金'),
    (0x328F, 'M', '土'),
    (0x3290, 'M', '日'),
    (0x3291, 'M', '株'),
    (0x3292, 'M', '有'),
    (0x3293, 'M', '社'),
    (0x3294, 'M', '名'),
    (0x3295, 'M', '特'),
    (0x3296, 'M', '財'),
    (0x3297, 'M', '祝'),
    (0x3298, 'M', '労'),
    (0x3299, 'M', '秘'),
    (0x329A, 'M', '男'),
    (0x329B, 'M', '女'),
    (0x329C, 'M', '適'),
    (0x329D, 'M', '優'),
    (0x329E, 'M', '印'),
    (0x329F, 'M', '注'),
    (0x32A0, 'M', '項'),
    (0x32A1, 'M', '休'),
    (0x32A2, 'M', '写'),
    (0x32A3, 'M', '正'),
    (0x32A4, 'M', '上'),
    (0x32A5, 'M', '中'),
    (0x32A6, 'M', '下'),
    (0x32A7, 'M', '左'),
    (0x32A8, 'M', '右'),
    (0x32A9, 'M', '医'),
    (0x32AA, 'M', '宗'),
    (0x32AB, 'M', '学'),
    (0x32AC, 'M', '監'),
    (0x32AD, 'M', '企'),
    ]

def _seg_32():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x32AE, 'M', '資'),
    (0x32AF, 'M', '協'),
    (0x32B0, 'M', '夜'),
    (0x32B1, 'M', '36'),
    (0x32B2, 'M', '37'),
    (0x32B3, 'M', '38'),
    (0x32B4, 'M', '39'),
    (0x32B5, 'M', '40'),
    (0x32B6, 'M', '41'),
    (0x32B7, 'M', '42'),
    (0x32B8, 'M', '43'),
    (0x32B9, 'M', '44'),
    (0x32BA, 'M', '45'),
    (0x32BB, 'M', '46'),
    (0x32BC, 'M', '47'),
    (0x32BD, 'M', '48'),
    (0x32BE, 'M', '49'),
    (0x32BF, 'M', '50'),
    (0x32C0, 'M', '1月'),
    (0x32C1, 'M', '2月'),
    (0x32C2, 'M', '3月'),
    (0x32C3, 'M', '4月'),
    (0x32C4, 'M', '5月'),
    (0x32C5, 'M', '6月'),
    (0x32C6, 'M', '7月'),
    (0x32C7, 'M', '8月'),
    (0x32C8, 'M', '9月'),
    (0x32C9, 'M', '10月'),
    (0x32CA, 'M', '11月'),
    (0x32CB, 'M', '12月'),
    (0x32CC, 'M', 'hg'),
    (0x32CD, 'M', 'erg'),
    (0x32CE, 'M', 'ev'),
    (0x32CF, 'M', 'ltd'),
    (0x32D0, 'M', 'ア'),
    (0x32D1, 'M', 'イ'),
    (0x32D2, 'M', 'ウ'),
    (0x32D3, 'M', 'エ'),
    (0x32D4, 'M', 'オ'),
    (0x32D5, 'M', 'カ'),
    (0x32D6, 'M', 'キ'),
    (0x32D7, 'M', 'ク'),
    (0x32D8, 'M', 'ケ'),
    (0x32D9, 'M', 'コ'),
    (0x32DA, 'M', 'サ'),
    (0x32DB, 'M', 'シ'),
    (0x32DC, 'M', 'ス'),
    (0x32DD, 'M', 'セ'),
    (0x32DE, 'M', 'ソ'),
    (0x32DF, 'M', 'タ'),
    (0x32E0, 'M', 'チ'),
    (0x32E1, 'M', 'ツ'),
    (0x32E2, 'M', 'テ'),
    (0x32E3, 'M', 'ト'),
    (0x32E4, 'M', 'ナ'),
    (0x32E5, 'M', 'ニ'),
    (0x32E6, 'M', 'ヌ'),
    (0x32E7, 'M', 'ネ'),
    (0x32E8, 'M', 'ノ'),
    (0x32E9, 'M', 'ハ'),
    (0x32EA, 'M', 'ヒ'),
    (0x32EB, 'M', 'フ'),
    (0x32EC, 'M', 'ヘ'),
    (0x32ED, 'M', 'ホ'),
    (0x32EE, 'M', 'マ'),
    (0x32EF, 'M', 'ミ'),
    (0x32F0, 'M', 'ム'),
    (0x32F1, 'M', 'メ'),
    (0x32F2, 'M', 'モ'),
    (0x32F3, 'M', 'ヤ'),
    (0x32F4, 'M', 'ユ'),
    (0x32F5, 'M', 'ヨ'),
    (0x32F6, 'M', 'ラ'),
    (0x32F7, 'M', 'リ'),
    (0x32F8, 'M', 'ル'),
    (0x32F9, 'M', 'レ'),
    (0x32FA, 'M', 'ロ'),
    (0x32FB, 'M', 'ワ'),
    (0x32FC, 'M', 'ヰ'),
    (0x32FD, 'M', 'ヱ'),
    (0x32FE, 'M', 'ヲ'),
    (0x32FF, 'M', '令和'),
    (0x3300, 'M', 'アパート'),
    (0x3301, 'M', 'アルファ'),
    (0x3302, 'M', 'アンペア'),
    (0x3303, 'M', 'アール'),
    (0x3304, 'M', 'イニング'),
    (0x3305, 'M', 'インチ'),
    (0x3306, 'M', 'ウォン'),
    (0x3307, 'M', 'エスクード'),
    (0x3308, 'M', 'エーカー'),
    (0x3309, 'M', 'オンス'),
    (0x330A, 'M', 'オーム'),
    (0x330B, 'M', 'カイリ'),
    (0x330C, 'M', 'カラット'),
    (0x330D, 'M', 'カロリー'),
    (0x330E, 'M', 'ガロン'),
    (0x330F, 'M', 'ガンマ'),
    (0x3310, 'M', 'ギガ'),
    (0x3311, 'M', 'ギニー'),
    ]

def _seg_33():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x3312, 'M', 'キュリー'),
    (0x3313, 'M', 'ギルダー'),
    (0x3314, 'M', 'キロ'),
    (0x3315, 'M', 'キログラム'),
    (0x3316, 'M', 'キロメートル'),
    (0x3317, 'M', 'キロワット'),
    (0x3318, 'M', 'グラム'),
    (0x3319, 'M', 'グラムトン'),
    (0x331A, 'M', 'クルゼイロ'),
    (0x331B, 'M', 'クローネ'),
    (0x331C, 'M', 'ケース'),
    (0x331D, 'M', 'コルナ'),
    (0x331E, 'M', 'コーポ'),
    (0x331F, 'M', 'サイクル'),
    (0x3320, 'M', 'サンチーム'),
    (0x3321, 'M', 'シリング'),
    (0x3322, 'M', 'センチ'),
    (0x3323, 'M', 'セント'),
    (0x3324, 'M', 'ダース'),
    (0x3325, 'M', 'デシ'),
    (0x3326, 'M', 'ドル'),
    (0x3327, 'M', 'トン'),
    (0x3328, 'M', 'ナノ'),
    (0x3329, 'M', 'ノット'),
    (0x332A, 'M', 'ハイツ'),
    (0x332B, 'M', 'パーセント'),
    (0x332C, 'M', 'パーツ'),
    (0x332D, 'M', 'バーレル'),
    (0x332E, 'M', 'ピアストル'),
    (0x332F, 'M', 'ピクル'),
    (0x3330, 'M', 'ピコ'),
    (0x3331, 'M', 'ビル'),
    (0x3332, 'M', 'ファラッド'),
    (0x3333, 'M', 'フィート'),
    (0x3334, 'M', 'ブッシェル'),
    (0x3335, 'M', 'フラン'),
    (0x3336, 'M', 'ヘクタール'),
    (0x3337, 'M', 'ペソ'),
    (0x3338, 'M', 'ペニヒ'),
    (0x3339, 'M', 'ヘルツ'),
    (0x333A, 'M', 'ペンス'),
    (0x333B, 'M', 'ページ'),
    (0x333C, 'M', 'ベータ'),
    (0x333D, 'M', 'ポイント'),
    (0x333E, 'M', 'ボルト'),
    (0x333F, 'M', 'ホン'),
    (0x3340, 'M', 'ポンド'),
    (0x3341, 'M', 'ホール'),
    (0x3342, 'M', 'ホーン'),
    (0x3343, 'M', 'マイクロ'),
    (0x3344, 'M', 'マイル'),
    (0x3345, 'M', 'マッハ'),
    (0x3346, 'M', 'マルク'),
    (0x3347, 'M', 'マンション'),
    (0x3348, 'M', 'ミクロン'),
    (0x3349, 'M', 'ミリ'),
    (0x334A, 'M', 'ミリバール'),
    (0x334B, 'M', 'メガ'),
    (0x334C, 'M', 'メガトン'),
    (0x334D, 'M', 'メートル'),
    (0x334E, 'M', 'ヤード'),
    (0x334F, 'M', 'ヤール'),
    (0x3350, 'M', 'ユアン'),
    (0x3351, 'M', 'リットル'),
    (0x3352, 'M', 'リラ'),
    (0x3353, 'M', 'ルピー'),
    (0x3354, 'M', 'ルーブル'),
    (0x3355, 'M', 'レム'),
    (0x3356, 'M', 'レントゲン'),
    (0x3357, 'M', 'ワット'),
    (0x3358, 'M', '0点'),
    (0x3359, 'M', '1点'),
    (0x335A, 'M', '2点'),
    (0x335B, 'M', '3点'),
    (0x335C, 'M', '4点'),
    (0x335D, 'M', '5点'),
    (0x335E, 'M', '6点'),
    (0x335F, 'M', '7点'),
    (0x3360, 'M', '8点'),
    (0x3361, 'M', '9点'),
    (0x3362, 'M', '10点'),
    (0x3363, 'M', '11点'),
    (0x3364, 'M', '12点'),
    (0x3365, 'M', '13点'),
    (0x3366, 'M', '14点'),
    (0x3367, 'M', '15点'),
    (0x3368, 'M', '16点'),
    (0x3369, 'M', '17点'),
    (0x336A, 'M', '18点'),
    (0x336B, 'M', '19点'),
    (0x336C, 'M', '20点'),
    (0x336D, 'M', '21点'),
    (0x336E, 'M', '22点'),
    (0x336F, 'M', '23点'),
    (0x3370, 'M', '24点'),
    (0x3371, 'M', 'hpa'),
    (0x3372, 'M', 'da'),
    (0x3373, 'M', 'au'),
    (0x3374, 'M', 'bar'),
    (0x3375, 'M', 'ov'),
    ]

def _seg_34():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x3376, 'M', 'pc'),
    (0x3377, 'M', 'dm'),
    (0x3378, 'M', 'dm2'),
    (0x3379, 'M', 'dm3'),
    (0x337A, 'M', 'iu'),
    (0x337B, 'M', '平成'),
    (0x337C, 'M', '昭和'),
    (0x337D, 'M', '大正'),
    (0x337E, 'M', '明治'),
    (0x337F, 'M', '株式会社'),
    (0x3380, 'M', 'pa'),
    (0x3381, 'M', 'na'),
    (0x3382, 'M', 'μa'),
    (0x3383, 'M', 'ma'),
    (0x3384, 'M', 'ka'),
    (0x3385, 'M', 'kb'),
    (0x3386, 'M', 'mb'),
    (0x3387, 'M', 'gb'),
    (0x3388, 'M', 'cal'),
    (0x3389, 'M', 'kcal'),
    (0x338A, 'M', 'pf'),
    (0x338B, 'M', 'nf'),
    (0x338C, 'M', 'μf'),
    (0x338D, 'M', 'μg'),
    (0x338E, 'M', 'mg'),
    (0x338F, 'M', 'kg'),
    (0x3390, 'M', 'hz'),
    (0x3391, 'M', 'khz'),
    (0x3392, 'M', 'mhz'),
    (0x3393, 'M', 'ghz'),
    (0x3394, 'M', 'thz'),
    (0x3395, 'M', 'μl'),
    (0x3396, 'M', 'ml'),
    (0x3397, 'M', 'dl'),
    (0x3398, 'M', 'kl'),
    (0x3399, 'M', 'fm'),
    (0x339A, 'M', 'nm'),
    (0x339B, 'M', 'μm'),
    (0x339C, 'M', 'mm'),
    (0x339D, 'M', 'cm'),
    (0x339E, 'M', 'km'),
    (0x339F, 'M', 'mm2'),
    (0x33A0, 'M', 'cm2'),
    (0x33A1, 'M', 'm2'),
    (0x33A2, 'M', 'km2'),
    (0x33A3, 'M', 'mm3'),
    (0x33A4, 'M', 'cm3'),
    (0x33A5, 'M', 'm3'),
    (0x33A6, 'M', 'km3'),
    (0x33A7, 'M', 'm∕s'),
    (0x33A8, 'M', 'm∕s2'),
    (0x33A9, 'M', 'pa'),
    (0x33AA, 'M', 'kpa'),
    (0x33AB, 'M', 'mpa'),
    (0x33AC, 'M', 'gpa'),
    (0x33AD, 'M', 'rad'),
    (0x33AE, 'M', 'rad∕s'),
    (0x33AF, 'M', 'rad∕s2'),
    (0x33B0, 'M', 'ps'),
    (0x33B1, 'M', 'ns'),
    (0x33B2, 'M', 'μs'),
    (0x33B3, 'M', 'ms'),
    (0x33B4, 'M', 'pv'),
    (0x33B5, 'M', 'nv'),
    (0x33B6, 'M', 'μv'),
    (0x33B7, 'M', 'mv'),
    (0x33B8, 'M', 'kv'),
    (0x33B9, 'M', 'mv'),
    (0x33BA, 'M', 'pw'),
    (0x33BB, 'M', 'nw'),
    (0x33BC, 'M', 'μw'),
    (0x33BD, 'M', 'mw'),
    (0x33BE, 'M', 'kw'),
    (0x33BF, 'M', 'mw'),
    (0x33C0, 'M', 'kω'),
    (0x33C1, 'M', 'mω'),
    (0x33C2, 'X'),
    (0x33C3, 'M', 'bq'),
    (0x33C4, 'M', 'cc'),
    (0x33C5, 'M', 'cd'),
    (0x33C6, 'M', 'c∕kg'),
    (0x33C7, 'X'),
    (0x33C8, 'M', 'db'),
    (0x33C9, 'M', 'gy'),
    (0x33CA, 'M', 'ha'),
    (0x33CB, 'M', 'hp'),
    (0x33CC, 'M', 'in'),
    (0x33CD, 'M', 'kk'),
    (0x33CE, 'M', 'km'),
    (0x33CF, 'M', 'kt'),
    (0x33D0, 'M', 'lm'),
    (0x33D1, 'M', 'ln'),
    (0x33D2, 'M', 'log'),
    (0x33D3, 'M', 'lx'),
    (0x33D4, 'M', 'mb'),
    (0x33D5, 'M', 'mil'),
    (0x33D6, 'M', 'mol'),
    (0x33D7, 'M', 'ph'),
    (0x33D8, 'X'),
    (0x33D9, 'M', 'ppm'),
    ]

def _seg_35():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x33DA, 'M', 'pr'),
    (0x33DB, 'M', 'sr'),
    (0x33DC, 'M', 'sv'),
    (0x33DD, 'M', 'wb'),
    (0x33DE, 'M', 'v∕m'),
    (0x33DF, 'M', 'a∕m'),
    (0x33E0, 'M', '1日'),
    (0x33E1, 'M', '2日'),
    (0x33E2, 'M', '3日'),
    (0x33E3, 'M', '4日'),
    (0x33E4, 'M', '5日'),
    (0x33E5, 'M', '6日'),
    (0x33E6, 'M', '7日'),
    (0x33E7, 'M', '8日'),
    (0x33E8, 'M', '9日'),
    (0x33E9, 'M', '10日'),
    (0x33EA, 'M', '11日'),
    (0x33EB, 'M', '12日'),
    (0x33EC, 'M', '13日'),
    (0x33ED, 'M', '14日'),
    (0x33EE, 'M', '15日'),
    (0x33EF, 'M', '16日'),
    (0x33F0, 'M', '17日'),
    (0x33F1, 'M', '18日'),
    (0x33F2, 'M', '19日'),
    (0x33F3, 'M', '20日'),
    (0x33F4, 'M', '21日'),
    (0x33F5, 'M', '22日'),
    (0x33F6, 'M', '23日'),
    (0x33F7, 'M', '24日'),
    (0x33F8, 'M', '25日'),
    (0x33F9, 'M', '26日'),
    (0x33FA, 'M', '27日'),
    (0x33FB, 'M', '28日'),
    (0x33FC, 'M', '29日'),
    (0x33FD, 'M', '30日'),
    (0x33FE, 'M', '31日'),
    (0x33FF, 'M', 'gal'),
    (0x3400, 'V'),
    (0x9FFD, 'X'),
    (0xA000, 'V'),
    (0xA48D, 'X'),
    (0xA490, 'V'),
    (0xA4C7, 'X'),
    (0xA4D0, 'V'),
    (0xA62C, 'X'),
    (0xA640, 'M', 'ꙁ'),
    (0xA641, 'V'),
    (0xA642, 'M', 'ꙃ'),
    (0xA643, 'V'),
    (0xA644, 'M', 'ꙅ'),
    (0xA645, 'V'),
    (0xA646, 'M', 'ꙇ'),
    (0xA647, 'V'),
    (0xA648, 'M', 'ꙉ'),
    (0xA649, 'V'),
    (0xA64A, 'M', 'ꙋ'),
    (0xA64B, 'V'),
    (0xA64C, 'M', 'ꙍ'),
    (0xA64D, 'V'),
    (0xA64E, 'M', 'ꙏ'),
    (0xA64F, 'V'),
    (0xA650, 'M', 'ꙑ'),
    (0xA651, 'V'),
    (0xA652, 'M', 'ꙓ'),
    (0xA653, 'V'),
    (0xA654, 'M', 'ꙕ'),
    (0xA655, 'V'),
    (0xA656, 'M', 'ꙗ'),
    (0xA657, 'V'),
    (0xA658, 'M', 'ꙙ'),
    (0xA659, 'V'),
    (0xA65A, 'M', 'ꙛ'),
    (0xA65B, 'V'),
    (0xA65C, 'M', 'ꙝ'),
    (0xA65D, 'V'),
    (0xA65E, 'M', 'ꙟ'),
    (0xA65F, 'V'),
    (0xA660, 'M', 'ꙡ'),
    (0xA661, 'V'),
    (0xA662, 'M', 'ꙣ'),
    (0xA663, 'V'),
    (0xA664, 'M', 'ꙥ'),
    (0xA665, 'V'),
    (0xA666, 'M', 'ꙧ'),
    (0xA667, 'V'),
    (0xA668, 'M', 'ꙩ'),
    (0xA669, 'V'),
    (0xA66A, 'M', 'ꙫ'),
    (0xA66B, 'V'),
    (0xA66C, 'M', 'ꙭ'),
    (0xA66D, 'V'),
    (0xA680, 'M', 'ꚁ'),
    (0xA681, 'V'),
    (0xA682, 'M', 'ꚃ'),
    (0xA683, 'V'),
    (0xA684, 'M', 'ꚅ'),
    (0xA685, 'V'),
    (0xA686, 'M', 'ꚇ'),
    (0xA687, 'V'),
    ]

def _seg_36():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xA688, 'M', 'ꚉ'),
    (0xA689, 'V'),
    (0xA68A, 'M', 'ꚋ'),
    (0xA68B, 'V'),
    (0xA68C, 'M', 'ꚍ'),
    (0xA68D, 'V'),
    (0xA68E, 'M', 'ꚏ'),
    (0xA68F, 'V'),
    (0xA690, 'M', 'ꚑ'),
    (0xA691, 'V'),
    (0xA692, 'M', 'ꚓ'),
    (0xA693, 'V'),
    (0xA694, 'M', 'ꚕ'),
    (0xA695, 'V'),
    (0xA696, 'M', 'ꚗ'),
    (0xA697, 'V'),
    (0xA698, 'M', 'ꚙ'),
    (0xA699, 'V'),
    (0xA69A, 'M', 'ꚛ'),
    (0xA69B, 'V'),
    (0xA69C, 'M', 'ъ'),
    (0xA69D, 'M', 'ь'),
    (0xA69E, 'V'),
    (0xA6F8, 'X'),
    (0xA700, 'V'),
    (0xA722, 'M', 'ꜣ'),
    (0xA723, 'V'),
    (0xA724, 'M', 'ꜥ'),
    (0xA725, 'V'),
    (0xA726, 'M', 'ꜧ'),
    (0xA727, 'V'),
    (0xA728, 'M', 'ꜩ'),
    (0xA729, 'V'),
    (0xA72A, 'M', 'ꜫ'),
    (0xA72B, 'V'),
    (0xA72C, 'M', 'ꜭ'),
    (0xA72D, 'V'),
    (0xA72E, 'M', 'ꜯ'),
    (0xA72F, 'V'),
    (0xA732, 'M', 'ꜳ'),
    (0xA733, 'V'),
    (0xA734, 'M', 'ꜵ'),
    (0xA735, 'V'),
    (0xA736, 'M', 'ꜷ'),
    (0xA737, 'V'),
    (0xA738, 'M', 'ꜹ'),
    (0xA739, 'V'),
    (0xA73A, 'M', 'ꜻ'),
    (0xA73B, 'V'),
    (0xA73C, 'M', 'ꜽ'),
    (0xA73D, 'V'),
    (0xA73E, 'M', 'ꜿ'),
    (0xA73F, 'V'),
    (0xA740, 'M', 'ꝁ'),
    (0xA741, 'V'),
    (0xA742, 'M', 'ꝃ'),
    (0xA743, 'V'),
    (0xA744, 'M', 'ꝅ'),
    (0xA745, 'V'),
    (0xA746, 'M', 'ꝇ'),
    (0xA747, 'V'),
    (0xA748, 'M', 'ꝉ'),
    (0xA749, 'V'),
    (0xA74A, 'M', 'ꝋ'),
    (0xA74B, 'V'),
    (0xA74C, 'M', 'ꝍ'),
    (0xA74D, 'V'),
    (0xA74E, 'M', 'ꝏ'),
    (0xA74F, 'V'),
    (0xA750, 'M', 'ꝑ'),
    (0xA751, 'V'),
    (0xA752, 'M', 'ꝓ'),
    (0xA753, 'V'),
    (0xA754, 'M', 'ꝕ'),
    (0xA755, 'V'),
    (0xA756, 'M', 'ꝗ'),
    (0xA757, 'V'),
    (0xA758, 'M', 'ꝙ'),
    (0xA759, 'V'),
    (0xA75A, 'M', 'ꝛ'),
    (0xA75B, 'V'),
    (0xA75C, 'M', 'ꝝ'),
    (0xA75D, 'V'),
    (0xA75E, 'M', 'ꝟ'),
    (0xA75F, 'V'),
    (0xA760, 'M', 'ꝡ'),
    (0xA761, 'V'),
    (0xA762, 'M', 'ꝣ'),
    (0xA763, 'V'),
    (0xA764, 'M', 'ꝥ'),
    (0xA765, 'V'),
    (0xA766, 'M', 'ꝧ'),
    (0xA767, 'V'),
    (0xA768, 'M', 'ꝩ'),
    (0xA769, 'V'),
    (0xA76A, 'M', 'ꝫ'),
    (0xA76B, 'V'),
    (0xA76C, 'M', 'ꝭ'),
    (0xA76D, 'V'),
    (0xA76E, 'M', 'ꝯ'),
    ]

def _seg_37():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xA76F, 'V'),
    (0xA770, 'M', 'ꝯ'),
    (0xA771, 'V'),
    (0xA779, 'M', 'ꝺ'),
    (0xA77A, 'V'),
    (0xA77B, 'M', 'ꝼ'),
    (0xA77C, 'V'),
    (0xA77D, 'M', 'ᵹ'),
    (0xA77E, 'M', 'ꝿ'),
    (0xA77F, 'V'),
    (0xA780, 'M', 'ꞁ'),
    (0xA781, 'V'),
    (0xA782, 'M', 'ꞃ'),
    (0xA783, 'V'),
    (0xA784, 'M', 'ꞅ'),
    (0xA785, 'V'),
    (0xA786, 'M', 'ꞇ'),
    (0xA787, 'V'),
    (0xA78B, 'M', 'ꞌ'),
    (0xA78C, 'V'),
    (0xA78D, 'M', 'ɥ'),
    (0xA78E, 'V'),
    (0xA790, 'M', 'ꞑ'),
    (0xA791, 'V'),
    (0xA792, 'M', 'ꞓ'),
    (0xA793, 'V'),
    (0xA796, 'M', 'ꞗ'),
    (0xA797, 'V'),
    (0xA798, 'M', 'ꞙ'),
    (0xA799, 'V'),
    (0xA79A, 'M', 'ꞛ'),
    (0xA79B, 'V'),
    (0xA79C, 'M', 'ꞝ'),
    (0xA79D, 'V'),
    (0xA79E, 'M', 'ꞟ'),
    (0xA79F, 'V'),
    (0xA7A0, 'M', 'ꞡ'),
    (0xA7A1, 'V'),
    (0xA7A2, 'M', 'ꞣ'),
    (0xA7A3, 'V'),
    (0xA7A4, 'M', 'ꞥ'),
    (0xA7A5, 'V'),
    (0xA7A6, 'M', 'ꞧ'),
    (0xA7A7, 'V'),
    (0xA7A8, 'M', 'ꞩ'),
    (0xA7A9, 'V'),
    (0xA7AA, 'M', 'ɦ'),
    (0xA7AB, 'M', 'ɜ'),
    (0xA7AC, 'M', 'ɡ'),
    (0xA7AD, 'M', 'ɬ'),
    (0xA7AE, 'M', 'ɪ'),
    (0xA7AF, 'V'),
    (0xA7B0, 'M', 'ʞ'),
    (0xA7B1, 'M', 'ʇ'),
    (0xA7B2, 'M', 'ʝ'),
    (0xA7B3, 'M', 'ꭓ'),
    (0xA7B4, 'M', 'ꞵ'),
    (0xA7B5, 'V'),
    (0xA7B6, 'M', 'ꞷ'),
    (0xA7B7, 'V'),
    (0xA7B8, 'M', 'ꞹ'),
    (0xA7B9, 'V'),
    (0xA7BA, 'M', 'ꞻ'),
    (0xA7BB, 'V'),
    (0xA7BC, 'M', 'ꞽ'),
    (0xA7BD, 'V'),
    (0xA7BE, 'M', 'ꞿ'),
    (0xA7BF, 'V'),
    (0xA7C0, 'X'),
    (0xA7C2, 'M', 'ꟃ'),
    (0xA7C3, 'V'),
    (0xA7C4, 'M', 'ꞔ'),
    (0xA7C5, 'M', 'ʂ'),
    (0xA7C6, 'M', 'ᶎ'),
    (0xA7C7, 'M', 'ꟈ'),
    (0xA7C8, 'V'),
    (0xA7C9, 'M', 'ꟊ'),
    (0xA7CA, 'V'),
    (0xA7CB, 'X'),
    (0xA7F5, 'M', 'ꟶ'),
    (0xA7F6, 'V'),
    (0xA7F8, 'M', 'ħ'),
    (0xA7F9, 'M', 'œ'),
    (0xA7FA, 'V'),
    (0xA82D, 'X'),
    (0xA830, 'V'),
    (0xA83A, 'X'),
    (0xA840, 'V'),
    (0xA878, 'X'),
    (0xA880, 'V'),
    (0xA8C6, 'X'),
    (0xA8CE, 'V'),
    (0xA8DA, 'X'),
    (0xA8E0, 'V'),
    (0xA954, 'X'),
    (0xA95F, 'V'),
    (0xA97D, 'X'),
    (0xA980, 'V'),
    (0xA9CE, 'X'),
    (0xA9CF, 'V'),
    ]

def _seg_38():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xA9DA, 'X'),
    (0xA9DE, 'V'),
    (0xA9FF, 'X'),
    (0xAA00, 'V'),
    (0xAA37, 'X'),
    (0xAA40, 'V'),
    (0xAA4E, 'X'),
    (0xAA50, 'V'),
    (0xAA5A, 'X'),
    (0xAA5C, 'V'),
    (0xAAC3, 'X'),
    (0xAADB, 'V'),
    (0xAAF7, 'X'),
    (0xAB01, 'V'),
    (0xAB07, 'X'),
    (0xAB09, 'V'),
    (0xAB0F, 'X'),
    (0xAB11, 'V'),
    (0xAB17, 'X'),
    (0xAB20, 'V'),
    (0xAB27, 'X'),
    (0xAB28, 'V'),
    (0xAB2F, 'X'),
    (0xAB30, 'V'),
    (0xAB5C, 'M', 'ꜧ'),
    (0xAB5D, 'M', 'ꬷ'),
    (0xAB5E, 'M', 'ɫ'),
    (0xAB5F, 'M', 'ꭒ'),
    (0xAB60, 'V'),
    (0xAB69, 'M', 'ʍ'),
    (0xAB6A, 'V'),
    (0xAB6C, 'X'),
    (0xAB70, 'M', 'Ꭰ'),
    (0xAB71, 'M', 'Ꭱ'),
    (0xAB72, 'M', 'Ꭲ'),
    (0xAB73, 'M', 'Ꭳ'),
    (0xAB74, 'M', 'Ꭴ'),
    (0xAB75, 'M', 'Ꭵ'),
    (0xAB76, 'M', 'Ꭶ'),
    (0xAB77, 'M', 'Ꭷ'),
    (0xAB78, 'M', 'Ꭸ'),
    (0xAB79, 'M', 'Ꭹ'),
    (0xAB7A, 'M', 'Ꭺ'),
    (0xAB7B, 'M', 'Ꭻ'),
    (0xAB7C, 'M', 'Ꭼ'),
    (0xAB7D, 'M', 'Ꭽ'),
    (0xAB7E, 'M', 'Ꭾ'),
    (0xAB7F, 'M', 'Ꭿ'),
    (0xAB80, 'M', 'Ꮀ'),
    (0xAB81, 'M', 'Ꮁ'),
    (0xAB82, 'M', 'Ꮂ'),
    (0xAB83, 'M', 'Ꮃ'),
    (0xAB84, 'M', 'Ꮄ'),
    (0xAB85, 'M', 'Ꮅ'),
    (0xAB86, 'M', 'Ꮆ'),
    (0xAB87, 'M', 'Ꮇ'),
    (0xAB88, 'M', 'Ꮈ'),
    (0xAB89, 'M', 'Ꮉ'),
    (0xAB8A, 'M', 'Ꮊ'),
    (0xAB8B, 'M', 'Ꮋ'),
    (0xAB8C, 'M', 'Ꮌ'),
    (0xAB8D, 'M', 'Ꮍ'),
    (0xAB8E, 'M', 'Ꮎ'),
    (0xAB8F, 'M', 'Ꮏ'),
    (0xAB90, 'M', 'Ꮐ'),
    (0xAB91, 'M', 'Ꮑ'),
    (0xAB92, 'M', 'Ꮒ'),
    (0xAB93, 'M', 'Ꮓ'),
    (0xAB94, 'M', 'Ꮔ'),
    (0xAB95, 'M', 'Ꮕ'),
    (0xAB96, 'M', 'Ꮖ'),
    (0xAB97, 'M', 'Ꮗ'),
    (0xAB98, 'M', 'Ꮘ'),
    (0xAB99, 'M', 'Ꮙ'),
    (0xAB9A, 'M', 'Ꮚ'),
    (0xAB9B, 'M', 'Ꮛ'),
    (0xAB9C, 'M', 'Ꮜ'),
    (0xAB9D, 'M', 'Ꮝ'),
    (0xAB9E, 'M', 'Ꮞ'),
    (0xAB9F, 'M', 'Ꮟ'),
    (0xABA0, 'M', 'Ꮠ'),
    (0xABA1, 'M', 'Ꮡ'),
    (0xABA2, 'M', 'Ꮢ'),
    (0xABA3, 'M', 'Ꮣ'),
    (0xABA4, 'M', 'Ꮤ'),
    (0xABA5, 'M', 'Ꮥ'),
    (0xABA6, 'M', 'Ꮦ'),
    (0xABA7, 'M', 'Ꮧ'),
    (0xABA8, 'M', 'Ꮨ'),
    (0xABA9, 'M', 'Ꮩ'),
    (0xABAA, 'M', 'Ꮪ'),
    (0xABAB, 'M', 'Ꮫ'),
    (0xABAC, 'M', 'Ꮬ'),
    (0xABAD, 'M', 'Ꮭ'),
    (0xABAE, 'M', 'Ꮮ'),
    (0xABAF, 'M', 'Ꮯ'),
    (0xABB0, 'M', 'Ꮰ'),
    (0xABB1, 'M', 'Ꮱ'),
    (0xABB2, 'M', 'Ꮲ'),
    (0xABB3, 'M', 'Ꮳ'),
    ]

def _seg_39():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xABB4, 'M', 'Ꮴ'),
    (0xABB5, 'M', 'Ꮵ'),
    (0xABB6, 'M', 'Ꮶ'),
    (0xABB7, 'M', 'Ꮷ'),
    (0xABB8, 'M', 'Ꮸ'),
    (0xABB9, 'M', 'Ꮹ'),
    (0xABBA, 'M', 'Ꮺ'),
    (0xABBB, 'M', 'Ꮻ'),
    (0xABBC, 'M', 'Ꮼ'),
    (0xABBD, 'M', 'Ꮽ'),
    (0xABBE, 'M', 'Ꮾ'),
    (0xABBF, 'M', 'Ꮿ'),
    (0xABC0, 'V'),
    (0xABEE, 'X'),
    (0xABF0, 'V'),
    (0xABFA, 'X'),
    (0xAC00, 'V'),
    (0xD7A4, 'X'),
    (0xD7B0, 'V'),
    (0xD7C7, 'X'),
    (0xD7CB, 'V'),
    (0xD7FC, 'X'),
    (0xF900, 'M', '豈'),
    (0xF901, 'M', '更'),
    (0xF902, 'M', '車'),
    (0xF903, 'M', '賈'),
    (0xF904, 'M', '滑'),
    (0xF905, 'M', '串'),
    (0xF906, 'M', '句'),
    (0xF907, 'M', '龜'),
    (0xF909, 'M', '契'),
    (0xF90A, 'M', '金'),
    (0xF90B, 'M', '喇'),
    (0xF90C, 'M', '奈'),
    (0xF90D, 'M', '懶'),
    (0xF90E, 'M', '癩'),
    (0xF90F, 'M', '羅'),
    (0xF910, 'M', '蘿'),
    (0xF911, 'M', '螺'),
    (0xF912, 'M', '裸'),
    (0xF913, 'M', '邏'),
    (0xF914, 'M', '樂'),
    (0xF915, 'M', '洛'),
    (0xF916, 'M', '烙'),
    (0xF917, 'M', '珞'),
    (0xF918, 'M', '落'),
    (0xF919, 'M', '酪'),
    (0xF91A, 'M', '駱'),
    (0xF91B, 'M', '亂'),
    (0xF91C, 'M', '卵'),
    (0xF91D, 'M', '欄'),
    (0xF91E, 'M', '爛'),
    (0xF91F, 'M', '蘭'),
    (0xF920, 'M', '鸞'),
    (0xF921, 'M', '嵐'),
    (0xF922, 'M', '濫'),
    (0xF923, 'M', '藍'),
    (0xF924, 'M', '襤'),
    (0xF925, 'M', '拉'),
    (0xF926, 'M', '臘'),
    (0xF927, 'M', '蠟'),
    (0xF928, 'M', '廊'),
    (0xF929, 'M', '朗'),
    (0xF92A, 'M', '浪'),
    (0xF92B, 'M', '狼'),
    (0xF92C, 'M', '郎'),
    (0xF92D, 'M', '來'),
    (0xF92E, 'M', '冷'),
    (0xF92F, 'M', '勞'),
    (0xF930, 'M', '擄'),
    (0xF931, 'M', '櫓'),
    (0xF932, 'M', '爐'),
    (0xF933, 'M', '盧'),
    (0xF934, 'M', '老'),
    (0xF935, 'M', '蘆'),
    (0xF936, 'M', '虜'),
    (0xF937, 'M', '路'),
    (0xF938, 'M', '露'),
    (0xF939, 'M', '魯'),
    (0xF93A, 'M', '鷺'),
    (0xF93B, 'M', '碌'),
    (0xF93C, 'M', '祿'),
    (0xF93D, 'M', '綠'),
    (0xF93E, 'M', '菉'),
    (0xF93F, 'M', '錄'),
    (0xF940, 'M', '鹿'),
    (0xF941, 'M', '論'),
    (0xF942, 'M', '壟'),
    (0xF943, 'M', '弄'),
    (0xF944, 'M', '籠'),
    (0xF945, 'M', '聾'),
    (0xF946, 'M', '牢'),
    (0xF947, 'M', '磊'),
    (0xF948, 'M', '賂'),
    (0xF949, 'M', '雷'),
    (0xF94A, 'M', '壘'),
    (0xF94B, 'M', '屢'),
    (0xF94C, 'M', '樓'),
    (0xF94D, 'M', '淚'),
    (0xF94E, 'M', '漏'),
    ]

def _seg_40():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xF94F, 'M', '累'),
    (0xF950, 'M', '縷'),
    (0xF951, 'M', '陋'),
    (0xF952, 'M', '勒'),
    (0xF953, 'M', '肋'),
    (0xF954, 'M', '凜'),
    (0xF955, 'M', '凌'),
    (0xF956, 'M', '稜'),
    (0xF957, 'M', '綾'),
    (0xF958, 'M', '菱'),
    (0xF959, 'M', '陵'),
    (0xF95A, 'M', '讀'),
    (0xF95B, 'M', '拏'),
    (0xF95C, 'M', '樂'),
    (0xF95D, 'M', '諾'),
    (0xF95E, 'M', '丹'),
    (0xF95F, 'M', '寧'),
    (0xF960, 'M', '怒'),
    (0xF961, 'M', '率'),
    (0xF962, 'M', '異'),
    (0xF963, 'M', '北'),
    (0xF964, 'M', '磻'),
    (0xF965, 'M', '便'),
    (0xF966, 'M', '復'),
    (0xF967, 'M', '不'),
    (0xF968, 'M', '泌'),
    (0xF969, 'M', '數'),
    (0xF96A, 'M', '索'),
    (0xF96B, 'M', '參'),
    (0xF96C, 'M', '塞'),
    (0xF96D, 'M', '省'),
    (0xF96E, 'M', '葉'),
    (0xF96F, 'M', '說'),
    (0xF970, 'M', '殺'),
    (0xF971, 'M', '辰'),
    (0xF972, 'M', '沈'),
    (0xF973, 'M', '拾'),
    (0xF974, 'M', '若'),
    (0xF975, 'M', '掠'),
    (0xF976, 'M', '略'),
    (0xF977, 'M', '亮'),
    (0xF978, 'M', '兩'),
    (0xF979, 'M', '凉'),
    (0xF97A, 'M', '梁'),
    (0xF97B, 'M', '糧'),
    (0xF97C, 'M', '良'),
    (0xF97D, 'M', '諒'),
    (0xF97E, 'M', '量'),
    (0xF97F, 'M', '勵'),
    (0xF980, 'M', '呂'),
    (0xF981, 'M', '女'),
    (0xF982, 'M', '廬'),
    (0xF983, 'M', '旅'),
    (0xF984, 'M', '濾'),
    (0xF985, 'M', '礪'),
    (0xF986, 'M', '閭'),
    (0xF987, 'M', '驪'),
    (0xF988, 'M', '麗'),
    (0xF989, 'M', '黎'),
    (0xF98A, 'M', '力'),
    (0xF98B, 'M', '曆'),
    (0xF98C, 'M', '歷'),
    (0xF98D, 'M', '轢'),
    (0xF98E, 'M', '年'),
    (0xF98F, 'M', '憐'),
    (0xF990, 'M', '戀'),
    (0xF991, 'M', '撚'),
    (0xF992, 'M', '漣'),
    (0xF993, 'M', '煉'),
    (0xF994, 'M', '璉'),
    (0xF995, 'M', '秊'),
    (0xF996, 'M', '練'),
    (0xF997, 'M', '聯'),
    (0xF998, 'M', '輦'),
    (0xF999, 'M', '蓮'),
    (0xF99A, 'M', '連'),
    (0xF99B, 'M', '鍊'),
    (0xF99C, 'M', '列'),
    (0xF99D, 'M', '劣'),
    (0xF99E, 'M', '咽'),
    (0xF99F, 'M', '烈'),
    (0xF9A0, 'M', '裂'),
    (0xF9A1, 'M', '說'),
    (0xF9A2, 'M', '廉'),
    (0xF9A3, 'M', '念'),
    (0xF9A4, 'M', '捻'),
    (0xF9A5, 'M', '殮'),
    (0xF9A6, 'M', '簾'),
    (0xF9A7, 'M', '獵'),
    (0xF9A8, 'M', '令'),
    (0xF9A9, 'M', '囹'),
    (0xF9AA, 'M', '寧'),
    (0xF9AB, 'M', '嶺'),
    (0xF9AC, 'M', '怜'),
    (0xF9AD, 'M', '玲'),
    (0xF9AE, 'M', '瑩'),
    (0xF9AF, 'M', '羚'),
    (0xF9B0, 'M', '聆'),
    (0xF9B1, 'M', '鈴'),
    (0xF9B2, 'M', '零'),
    ]

def _seg_41():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xF9B3, 'M', '靈'),
    (0xF9B4, 'M', '領'),
    (0xF9B5, 'M', '例'),
    (0xF9B6, 'M', '禮'),
    (0xF9B7, 'M', '醴'),
    (0xF9B8, 'M', '隸'),
    (0xF9B9, 'M', '惡'),
    (0xF9BA, 'M', '了'),
    (0xF9BB, 'M', '僚'),
    (0xF9BC, 'M', '寮'),
    (0xF9BD, 'M', '尿'),
    (0xF9BE, 'M', '料'),
    (0xF9BF, 'M', '樂'),
    (0xF9C0, 'M', '燎'),
    (0xF9C1, 'M', '療'),
    (0xF9C2, 'M', '蓼'),
    (0xF9C3, 'M', '遼'),
    (0xF9C4, 'M', '龍'),
    (0xF9C5, 'M', '暈'),
    (0xF9C6, 'M', '阮'),
    (0xF9C7, 'M', '劉'),
    (0xF9C8, 'M', '杻'),
    (0xF9C9, 'M', '柳'),
    (0xF9CA, 'M', '流'),
    (0xF9CB, 'M', '溜'),
    (0xF9CC, 'M', '琉'),
    (0xF9CD, 'M', '留'),
    (0xF9CE, 'M', '硫'),
    (0xF9CF, 'M', '紐'),
    (0xF9D0, 'M', '類'),
    (0xF9D1, 'M', '六'),
    (0xF9D2, 'M', '戮'),
    (0xF9D3, 'M', '陸'),
    (0xF9D4, 'M', '倫'),
    (0xF9D5, 'M', '崙'),
    (0xF9D6, 'M', '淪'),
    (0xF9D7, 'M', '輪'),
    (0xF9D8, 'M', '律'),
    (0xF9D9, 'M', '慄'),
    (0xF9DA, 'M', '栗'),
    (0xF9DB, 'M', '率'),
    (0xF9DC, 'M', '隆'),
    (0xF9DD, 'M', '利'),
    (0xF9DE, 'M', '吏'),
    (0xF9DF, 'M', '履'),
    (0xF9E0, 'M', '易'),
    (0xF9E1, 'M', '李'),
    (0xF9E2, 'M', '梨'),
    (0xF9E3, 'M', '泥'),
    (0xF9E4, 'M', '理'),
    (0xF9E5, 'M', '痢'),
    (0xF9E6, 'M', '罹'),
    (0xF9E7, 'M', '裏'),
    (0xF9E8, 'M', '裡'),
    (0xF9E9, 'M', '里'),
    (0xF9EA, 'M', '離'),
    (0xF9EB, 'M', '匿'),
    (0xF9EC, 'M', '溺'),
    (0xF9ED, 'M', '吝'),
    (0xF9EE, 'M', '燐'),
    (0xF9EF, 'M', '璘'),
    (0xF9F0, 'M', '藺'),
    (0xF9F1, 'M', '隣'),
    (0xF9F2, 'M', '鱗'),
    (0xF9F3, 'M', '麟'),
    (0xF9F4, 'M', '林'),
    (0xF9F5, 'M', '淋'),
    (0xF9F6, 'M', '臨'),
    (0xF9F7, 'M', '立'),
    (0xF9F8, 'M', '笠'),
    (0xF9F9, 'M', '粒'),
    (0xF9FA, 'M', '狀'),
    (0xF9FB, 'M', '炙'),
    (0xF9FC, 'M', '識'),
    (0xF9FD, 'M', '什'),
    (0xF9FE, 'M', '茶'),
    (0xF9FF, 'M', '刺'),
    (0xFA00, 'M', '切'),
    (0xFA01, 'M', '度'),
    (0xFA02, 'M', '拓'),
    (0xFA03, 'M', '糖'),
    (0xFA04, 'M', '宅'),
    (0xFA05, 'M', '洞'),
    (0xFA06, 'M', '暴'),
    (0xFA07, 'M', '輻'),
    (0xFA08, 'M', '行'),
    (0xFA09, 'M', '降'),
    (0xFA0A, 'M', '見'),
    (0xFA0B, 'M', '廓'),
    (0xFA0C, 'M', '兀'),
    (0xFA0D, 'M', '嗀'),
    (0xFA0E, 'V'),
    (0xFA10, 'M', '塚'),
    (0xFA11, 'V'),
    (0xFA12, 'M', '晴'),
    (0xFA13, 'V'),
    (0xFA15, 'M', '凞'),
    (0xFA16, 'M', '猪'),
    (0xFA17, 'M', '益'),
    (0xFA18, 'M', '礼'),
    ]

def _seg_42():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFA19, 'M', '神'),
    (0xFA1A, 'M', '祥'),
    (0xFA1B, 'M', '福'),
    (0xFA1C, 'M', '靖'),
    (0xFA1D, 'M', '精'),
    (0xFA1E, 'M', '羽'),
    (0xFA1F, 'V'),
    (0xFA20, 'M', '蘒'),
    (0xFA21, 'V'),
    (0xFA22, 'M', '諸'),
    (0xFA23, 'V'),
    (0xFA25, 'M', '逸'),
    (0xFA26, 'M', '都'),
    (0xFA27, 'V'),
    (0xFA2A, 'M', '飯'),
    (0xFA2B, 'M', '飼'),
    (0xFA2C, 'M', '館'),
    (0xFA2D, 'M', '鶴'),
    (0xFA2E, 'M', '郞'),
    (0xFA2F, 'M', '隷'),
    (0xFA30, 'M', '侮'),
    (0xFA31, 'M', '僧'),
    (0xFA32, 'M', '免'),
    (0xFA33, 'M', '勉'),
    (0xFA34, 'M', '勤'),
    (0xFA35, 'M', '卑'),
    (0xFA36, 'M', '喝'),
    (0xFA37, 'M', '嘆'),
    (0xFA38, 'M', '器'),
    (0xFA39, 'M', '塀'),
    (0xFA3A, 'M', '墨'),
    (0xFA3B, 'M', '層'),
    (0xFA3C, 'M', '屮'),
    (0xFA3D, 'M', '悔'),
    (0xFA3E, 'M', '慨'),
    (0xFA3F, 'M', '憎'),
    (0xFA40, 'M', '懲'),
    (0xFA41, 'M', '敏'),
    (0xFA42, 'M', '既'),
    (0xFA43, 'M', '暑'),
    (0xFA44, 'M', '梅'),
    (0xFA45, 'M', '海'),
    (0xFA46, 'M', '渚'),
    (0xFA47, 'M', '漢'),
    (0xFA48, 'M', '煮'),
    (0xFA49, 'M', '爫'),
    (0xFA4A, 'M', '琢'),
    (0xFA4B, 'M', '碑'),
    (0xFA4C, 'M', '社'),
    (0xFA4D, 'M', '祉'),
    (0xFA4E, 'M', '祈'),
    (0xFA4F, 'M', '祐'),
    (0xFA50, 'M', '祖'),
    (0xFA51, 'M', '祝'),
    (0xFA52, 'M', '禍'),
    (0xFA53, 'M', '禎'),
    (0xFA54, 'M', '穀'),
    (0xFA55, 'M', '突'),
    (0xFA56, 'M', '節'),
    (0xFA57, 'M', '練'),
    (0xFA58, 'M', '縉'),
    (0xFA59, 'M', '繁'),
    (0xFA5A, 'M', '署'),
    (0xFA5B, 'M', '者'),
    (0xFA5C, 'M', '臭'),
    (0xFA5D, 'M', '艹'),
    (0xFA5F, 'M', '著'),
    (0xFA60, 'M', '褐'),
    (0xFA61, 'M', '視'),
    (0xFA62, 'M', '謁'),
    (0xFA63, 'M', '謹'),
    (0xFA64, 'M', '賓'),
    (0xFA65, 'M', '贈'),
    (0xFA66, 'M', '辶'),
    (0xFA67, 'M', '逸'),
    (0xFA68, 'M', '難'),
    (0xFA69, 'M', '響'),
    (0xFA6A, 'M', '頻'),
    (0xFA6B, 'M', '恵'),
    (0xFA6C, 'M', '𤋮'),
    (0xFA6D, 'M', '舘'),
    (0xFA6E, 'X'),
    (0xFA70, 'M', '並'),
    (0xFA71, 'M', '况'),
    (0xFA72, 'M', '全'),
    (0xFA73, 'M', '侀'),
    (0xFA74, 'M', '充'),
    (0xFA75, 'M', '冀'),
    (0xFA76, 'M', '勇'),
    (0xFA77, 'M', '勺'),
    (0xFA78, 'M', '喝'),
    (0xFA79, 'M', '啕'),
    (0xFA7A, 'M', '喙'),
    (0xFA7B, 'M', '嗢'),
    (0xFA7C, 'M', '塚'),
    (0xFA7D, 'M', '墳'),
    (0xFA7E, 'M', '奄'),
    (0xFA7F, 'M', '奔'),
    (0xFA80, 'M', '婢'),
    (0xFA81, 'M', '嬨'),
    ]

def _seg_43():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFA82, 'M', '廒'),
    (0xFA83, 'M', '廙'),
    (0xFA84, 'M', '彩'),
    (0xFA85, 'M', '徭'),
    (0xFA86, 'M', '惘'),
    (0xFA87, 'M', '慎'),
    (0xFA88, 'M', '愈'),
    (0xFA89, 'M', '憎'),
    (0xFA8A, 'M', '慠'),
    (0xFA8B, 'M', '懲'),
    (0xFA8C, 'M', '戴'),
    (0xFA8D, 'M', '揄'),
    (0xFA8E, 'M', '搜'),
    (0xFA8F, 'M', '摒'),
    (0xFA90, 'M', '敖'),
    (0xFA91, 'M', '晴'),
    (0xFA92, 'M', '朗'),
    (0xFA93, 'M', '望'),
    (0xFA94, 'M', '杖'),
    (0xFA95, 'M', '歹'),
    (0xFA96, 'M', '殺'),
    (0xFA97, 'M', '流'),
    (0xFA98, 'M', '滛'),
    (0xFA99, 'M', '滋'),
    (0xFA9A, 'M', '漢'),
    (0xFA9B, 'M', '瀞'),
    (0xFA9C, 'M', '煮'),
    (0xFA9D, 'M', '瞧'),
    (0xFA9E, 'M', '爵'),
    (0xFA9F, 'M', '犯'),
    (0xFAA0, 'M', '猪'),
    (0xFAA1, 'M', '瑱'),
    (0xFAA2, 'M', '甆'),
    (0xFAA3, 'M', '画'),
    (0xFAA4, 'M', '瘝'),
    (0xFAA5, 'M', '瘟'),
    (0xFAA6, 'M', '益'),
    (0xFAA7, 'M', '盛'),
    (0xFAA8, 'M', '直'),
    (0xFAA9, 'M', '睊'),
    (0xFAAA, 'M', '着'),
    (0xFAAB, 'M', '磌'),
    (0xFAAC, 'M', '窱'),
    (0xFAAD, 'M', '節'),
    (0xFAAE, 'M', '类'),
    (0xFAAF, 'M', '絛'),
    (0xFAB0, 'M', '練'),
    (0xFAB1, 'M', '缾'),
    (0xFAB2, 'M', '者'),
    (0xFAB3, 'M', '荒'),
    (0xFAB4, 'M', '華'),
    (0xFAB5, 'M', '蝹'),
    (0xFAB6, 'M', '襁'),
    (0xFAB7, 'M', '覆'),
    (0xFAB8, 'M', '視'),
    (0xFAB9, 'M', '調'),
    (0xFABA, 'M', '諸'),
    (0xFABB, 'M', '請'),
    (0xFABC, 'M', '謁'),
    (0xFABD, 'M', '諾'),
    (0xFABE, 'M', '諭'),
    (0xFABF, 'M', '謹'),
    (0xFAC0, 'M', '變'),
    (0xFAC1, 'M', '贈'),
    (0xFAC2, 'M', '輸'),
    (0xFAC3, 'M', '遲'),
    (0xFAC4, 'M', '醙'),
    (0xFAC5, 'M', '鉶'),
    (0xFAC6, 'M', '陼'),
    (0xFAC7, 'M', '難'),
    (0xFAC8, 'M', '靖'),
    (0xFAC9, 'M', '韛'),
    (0xFACA, 'M', '響'),
    (0xFACB, 'M', '頋'),
    (0xFACC, 'M', '頻'),
    (0xFACD, 'M', '鬒'),
    (0xFACE, 'M', '龜'),
    (0xFACF, 'M', '𢡊'),
    (0xFAD0, 'M', '𢡄'),
    (0xFAD1, 'M', '𣏕'),
    (0xFAD2, 'M', '㮝'),
    (0xFAD3, 'M', '䀘'),
    (0xFAD4, 'M', '䀹'),
    (0xFAD5, 'M', '𥉉'),
    (0xFAD6, 'M', '𥳐'),
    (0xFAD7, 'M', '𧻓'),
    (0xFAD8, 'M', '齃'),
    (0xFAD9, 'M', '龎'),
    (0xFADA, 'X'),
    (0xFB00, 'M', 'ff'),
    (0xFB01, 'M', 'fi'),
    (0xFB02, 'M', 'fl'),
    (0xFB03, 'M', 'ffi'),
    (0xFB04, 'M', 'ffl'),
    (0xFB05, 'M', 'st'),
    (0xFB07, 'X'),
    (0xFB13, 'M', 'մն'),
    (0xFB14, 'M', 'մե'),
    (0xFB15, 'M', 'մի'),
    (0xFB16, 'M', 'վն'),
    ]

def _seg_44():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFB17, 'M', 'մխ'),
    (0xFB18, 'X'),
    (0xFB1D, 'M', 'יִ'),
    (0xFB1E, 'V'),
    (0xFB1F, 'M', 'ײַ'),
    (0xFB20, 'M', 'ע'),
    (0xFB21, 'M', 'א'),
    (0xFB22, 'M', 'ד'),
    (0xFB23, 'M', 'ה'),
    (0xFB24, 'M', 'כ'),
    (0xFB25, 'M', 'ל'),
    (0xFB26, 'M', 'ם'),
    (0xFB27, 'M', 'ר'),
    (0xFB28, 'M', 'ת'),
    (0xFB29, '3', '+'),
    (0xFB2A, 'M', 'שׁ'),
    (0xFB2B, 'M', 'שׂ'),
    (0xFB2C, 'M', 'שּׁ'),
    (0xFB2D, 'M', 'שּׂ'),
    (0xFB2E, 'M', 'אַ'),
    (0xFB2F, 'M', 'אָ'),
    (0xFB30, 'M', 'אּ'),
    (0xFB31, 'M', 'בּ'),
    (0xFB32, 'M', 'גּ'),
    (0xFB33, 'M', 'דּ'),
    (0xFB34, 'M', 'הּ'),
    (0xFB35, 'M', 'וּ'),
    (0xFB36, 'M', 'זּ'),
    (0xFB37, 'X'),
    (0xFB38, 'M', 'טּ'),
    (0xFB39, 'M', 'יּ'),
    (0xFB3A, 'M', 'ךּ'),
    (0xFB3B, 'M', 'כּ'),
    (0xFB3C, 'M', 'לּ'),
    (0xFB3D, 'X'),
    (0xFB3E, 'M', 'מּ'),
    (0xFB3F, 'X'),
    (0xFB40, 'M', 'נּ'),
    (0xFB41, 'M', 'סּ'),
    (0xFB42, 'X'),
    (0xFB43, 'M', 'ףּ'),
    (0xFB44, 'M', 'פּ'),
    (0xFB45, 'X'),
    (0xFB46, 'M', 'צּ'),
    (0xFB47, 'M', 'קּ'),
    (0xFB48, 'M', 'רּ'),
    (0xFB49, 'M', 'שּ'),
    (0xFB4A, 'M', 'תּ'),
    (0xFB4B, 'M', 'וֹ'),
    (0xFB4C, 'M', 'בֿ'),
    (0xFB4D, 'M', 'כֿ'),
    (0xFB4E, 'M', 'פֿ'),
    (0xFB4F, 'M', 'אל'),
    (0xFB50, 'M', 'ٱ'),
    (0xFB52, 'M', 'ٻ'),
    (0xFB56, 'M', 'پ'),
    (0xFB5A, 'M', 'ڀ'),
    (0xFB5E, 'M', 'ٺ'),
    (0xFB62, 'M', 'ٿ'),
    (0xFB66, 'M', 'ٹ'),
    (0xFB6A, 'M', 'ڤ'),
    (0xFB6E, 'M', 'ڦ'),
    (0xFB72, 'M', 'ڄ'),
    (0xFB76, 'M', 'ڃ'),
    (0xFB7A, 'M', 'چ'),
    (0xFB7E, 'M', 'ڇ'),
    (0xFB82, 'M', 'ڍ'),
    (0xFB84, 'M', 'ڌ'),
    (0xFB86, 'M', 'ڎ'),
    (0xFB88, 'M', 'ڈ'),
    (0xFB8A, 'M', 'ژ'),
    (0xFB8C, 'M', 'ڑ'),
    (0xFB8E, 'M', 'ک'),
    (0xFB92, 'M', 'گ'),
    (0xFB96, 'M', 'ڳ'),
    (0xFB9A, 'M', 'ڱ'),
    (0xFB9E, 'M', 'ں'),
    (0xFBA0, 'M', 'ڻ'),
    (0xFBA4, 'M', 'ۀ'),
    (0xFBA6, 'M', 'ہ'),
    (0xFBAA, 'M', 'ھ'),
    (0xFBAE, 'M', 'ے'),
    (0xFBB0, 'M', 'ۓ'),
    (0xFBB2, 'V'),
    (0xFBC2, 'X'),
    (0xFBD3, 'M', 'ڭ'),
    (0xFBD7, 'M', 'ۇ'),
    (0xFBD9, 'M', 'ۆ'),
    (0xFBDB, 'M', 'ۈ'),
    (0xFBDD, 'M', 'ۇٴ'),
    (0xFBDE, 'M', 'ۋ'),
    (0xFBE0, 'M', 'ۅ'),
    (0xFBE2, 'M', 'ۉ'),
    (0xFBE4, 'M', 'ې'),
    (0xFBE8, 'M', 'ى'),
    (0xFBEA, 'M', 'ئا'),
    (0xFBEC, 'M', 'ئە'),
    (0xFBEE, 'M', 'ئو'),
    (0xFBF0, 'M', 'ئۇ'),
    (0xFBF2, 'M', 'ئۆ'),
    ]

def _seg_45():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFBF4, 'M', 'ئۈ'),
    (0xFBF6, 'M', 'ئې'),
    (0xFBF9, 'M', 'ئى'),
    (0xFBFC, 'M', 'ی'),
    (0xFC00, 'M', 'ئج'),
    (0xFC01, 'M', 'ئح'),
    (0xFC02, 'M', 'ئم'),
    (0xFC03, 'M', 'ئى'),
    (0xFC04, 'M', 'ئي'),
    (0xFC05, 'M', 'بج'),
    (0xFC06, 'M', 'بح'),
    (0xFC07, 'M', 'بخ'),
    (0xFC08, 'M', 'بم'),
    (0xFC09, 'M', 'بى'),
    (0xFC0A, 'M', 'بي'),
    (0xFC0B, 'M', 'تج'),
    (0xFC0C, 'M', 'تح'),
    (0xFC0D, 'M', 'تخ'),
    (0xFC0E, 'M', 'تم'),
    (0xFC0F, 'M', 'تى'),
    (0xFC10, 'M', 'تي'),
    (0xFC11, 'M', 'ثج'),
    (0xFC12, 'M', 'ثم'),
    (0xFC13, 'M', 'ثى'),
    (0xFC14, 'M', 'ثي'),
    (0xFC15, 'M', 'جح'),
    (0xFC16, 'M', 'جم'),
    (0xFC17, 'M', 'حج'),
    (0xFC18, 'M', 'حم'),
    (0xFC19, 'M', 'خج'),
    (0xFC1A, 'M', 'خح'),
    (0xFC1B, 'M', 'خم'),
    (0xFC1C, 'M', 'سج'),
    (0xFC1D, 'M', 'سح'),
    (0xFC1E, 'M', 'سخ'),
    (0xFC1F, 'M', 'سم'),
    (0xFC20, 'M', 'صح'),
    (0xFC21, 'M', 'صم'),
    (0xFC22, 'M', 'ضج'),
    (0xFC23, 'M', 'ضح'),
    (0xFC24, 'M', 'ضخ'),
    (0xFC25, 'M', 'ضم'),
    (0xFC26, 'M', 'طح'),
    (0xFC27, 'M', 'طم'),
    (0xFC28, 'M', 'ظم'),
    (0xFC29, 'M', 'عج'),
    (0xFC2A, 'M', 'عم'),
    (0xFC2B, 'M', 'غج'),
    (0xFC2C, 'M', 'غم'),
    (0xFC2D, 'M', 'فج'),
    (0xFC2E, 'M', 'فح'),
    (0xFC2F, 'M', 'فخ'),
    (0xFC30, 'M', 'فم'),
    (0xFC31, 'M', 'فى'),
    (0xFC32, 'M', 'في'),
    (0xFC33, 'M', 'قح'),
    (0xFC34, 'M', 'قم'),
    (0xFC35, 'M', 'قى'),
    (0xFC36, 'M', 'قي'),
    (0xFC37, 'M', 'كا'),
    (0xFC38, 'M', 'كج'),
    (0xFC39, 'M', 'كح'),
    (0xFC3A, 'M', 'كخ'),
    (0xFC3B, 'M', 'كل'),
    (0xFC3C, 'M', 'كم'),
    (0xFC3D, 'M', 'كى'),
    (0xFC3E, 'M', 'كي'),
    (0xFC3F, 'M', 'لج'),
    (0xFC40, 'M', 'لح'),
    (0xFC41, 'M', 'لخ'),
    (0xFC42, 'M', 'لم'),
    (0xFC43, 'M', 'لى'),
    (0xFC44, 'M', 'لي'),
    (0xFC45, 'M', 'مج'),
    (0xFC46, 'M', 'مح'),
    (0xFC47, 'M', 'مخ'),
    (0xFC48, 'M', 'مم'),
    (0xFC49, 'M', 'مى'),
    (0xFC4A, 'M', 'مي'),
    (0xFC4B, 'M', 'نج'),
    (0xFC4C, 'M', 'نح'),
    (0xFC4D, 'M', 'نخ'),
    (0xFC4E, 'M', 'نم'),
    (0xFC4F, 'M', 'نى'),
    (0xFC50, 'M', 'ني'),
    (0xFC51, 'M', 'هج'),
    (0xFC52, 'M', 'هم'),
    (0xFC53, 'M', 'هى'),
    (0xFC54, 'M', 'هي'),
    (0xFC55, 'M', 'يج'),
    (0xFC56, 'M', 'يح'),
    (0xFC57, 'M', 'يخ'),
    (0xFC58, 'M', 'يم'),
    (0xFC59, 'M', 'يى'),
    (0xFC5A, 'M', 'يي'),
    (0xFC5B, 'M', 'ذٰ'),
    (0xFC5C, 'M', 'رٰ'),
    (0xFC5D, 'M', 'ىٰ'),
    (0xFC5E, '3', ' ٌّ'),
    (0xFC5F, '3', ' ٍّ'),
    ]

def _seg_46():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFC60, '3', ' َّ'),
    (0xFC61, '3', ' ُّ'),
    (0xFC62, '3', ' ِّ'),
    (0xFC63, '3', ' ّٰ'),
    (0xFC64, 'M', 'ئر'),
    (0xFC65, 'M', 'ئز'),
    (0xFC66, 'M', 'ئم'),
    (0xFC67, 'M', 'ئن'),
    (0xFC68, 'M', 'ئى'),
    (0xFC69, 'M', 'ئي'),
    (0xFC6A, 'M', 'بر'),
    (0xFC6B, 'M', 'بز'),
    (0xFC6C, 'M', 'بم'),
    (0xFC6D, 'M', 'بن'),
    (0xFC6E, 'M', 'بى'),
    (0xFC6F, 'M', 'بي'),
    (0xFC70, 'M', 'تر'),
    (0xFC71, 'M', 'تز'),
    (0xFC72, 'M', 'تم'),
    (0xFC73, 'M', 'تن'),
    (0xFC74, 'M', 'تى'),
    (0xFC75, 'M', 'تي'),
    (0xFC76, 'M', 'ثر'),
    (0xFC77, 'M', 'ثز'),
    (0xFC78, 'M', 'ثم'),
    (0xFC79, 'M', 'ثن'),
    (0xFC7A, 'M', 'ثى'),
    (0xFC7B, 'M', 'ثي'),
    (0xFC7C, 'M', 'فى'),
    (0xFC7D, 'M', 'في'),
    (0xFC7E, 'M', 'قى'),
    (0xFC7F, 'M', 'قي'),
    (0xFC80, 'M', 'كا'),
    (0xFC81, 'M', 'كل'),
    (0xFC82, 'M', 'كم'),
    (0xFC83, 'M', 'كى'),
    (0xFC84, 'M', 'كي'),
    (0xFC85, 'M', 'لم'),
    (0xFC86, 'M', 'لى'),
    (0xFC87, 'M', 'لي'),
    (0xFC88, 'M', 'ما'),
    (0xFC89, 'M', 'مم'),
    (0xFC8A, 'M', 'نر'),
    (0xFC8B, 'M', 'نز'),
    (0xFC8C, 'M', 'نم'),
    (0xFC8D, 'M', 'نن'),
    (0xFC8E, 'M', 'نى'),
    (0xFC8F, 'M', 'ني'),
    (0xFC90, 'M', 'ىٰ'),
    (0xFC91, 'M', 'ير'),
    (0xFC92, 'M', 'يز'),
    (0xFC93, 'M', 'يم'),
    (0xFC94, 'M', 'ين'),
    (0xFC95, 'M', 'يى'),
    (0xFC96, 'M', 'يي'),
    (0xFC97, 'M', 'ئج'),
    (0xFC98, 'M', 'ئح'),
    (0xFC99, 'M', 'ئخ'),
    (0xFC9A, 'M', 'ئم'),
    (0xFC9B, 'M', 'ئه'),
    (0xFC9C, 'M', 'بج'),
    (0xFC9D, 'M', 'بح'),
    (0xFC9E, 'M', 'بخ'),
    (0xFC9F, 'M', 'بم'),
    (0xFCA0, 'M', 'به'),
    (0xFCA1, 'M', 'تج'),
    (0xFCA2, 'M', 'تح'),
    (0xFCA3, 'M', 'تخ'),
    (0xFCA4, 'M', 'تم'),
    (0xFCA5, 'M', 'ته'),
    (0xFCA6, 'M', 'ثم'),
    (0xFCA7, 'M', 'جح'),
    (0xFCA8, 'M', 'جم'),
    (0xFCA9, 'M', 'حج'),
    (0xFCAA, 'M', 'حم'),
    (0xFCAB, 'M', 'خج'),
    (0xFCAC, 'M', 'خم'),
    (0xFCAD, 'M', 'سج'),
    (0xFCAE, 'M', 'سح'),
    (0xFCAF, 'M', 'سخ'),
    (0xFCB0, 'M', 'سم'),
    (0xFCB1, 'M', 'صح'),
    (0xFCB2, 'M', 'صخ'),
    (0xFCB3, 'M', 'صم'),
    (0xFCB4, 'M', 'ضج'),
    (0xFCB5, 'M', 'ضح'),
    (0xFCB6, 'M', 'ضخ'),
    (0xFCB7, 'M', 'ضم'),
    (0xFCB8, 'M', 'طح'),
    (0xFCB9, 'M', 'ظم'),
    (0xFCBA, 'M', 'عج'),
    (0xFCBB, 'M', 'عم'),
    (0xFCBC, 'M', 'غج'),
    (0xFCBD, 'M', 'غم'),
    (0xFCBE, 'M', 'فج'),
    (0xFCBF, 'M', 'فح'),
    (0xFCC0, 'M', 'فخ'),
    (0xFCC1, 'M', 'فم'),
    (0xFCC2, 'M', 'قح'),
    (0xFCC3, 'M', 'قم'),
    ]

def _seg_47():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFCC4, 'M', 'كج'),
    (0xFCC5, 'M', 'كح'),
    (0xFCC6, 'M', 'كخ'),
    (0xFCC7, 'M', 'كل'),
    (0xFCC8, 'M', 'كم'),
    (0xFCC9, 'M', 'لج'),
    (0xFCCA, 'M', 'لح'),
    (0xFCCB, 'M', 'لخ'),
    (0xFCCC, 'M', 'لم'),
    (0xFCCD, 'M', 'له'),
    (0xFCCE, 'M', 'مج'),
    (0xFCCF, 'M', 'مح'),
    (0xFCD0, 'M', 'مخ'),
    (0xFCD1, 'M', 'مم'),
    (0xFCD2, 'M', 'نج'),
    (0xFCD3, 'M', 'نح'),
    (0xFCD4, 'M', 'نخ'),
    (0xFCD5, 'M', 'نم'),
    (0xFCD6, 'M', 'نه'),
    (0xFCD7, 'M', 'هج'),
    (0xFCD8, 'M', 'هم'),
    (0xFCD9, 'M', 'هٰ'),
    (0xFCDA, 'M', 'يج'),
    (0xFCDB, 'M', 'يح'),
    (0xFCDC, 'M', 'يخ'),
    (0xFCDD, 'M', 'يم'),
    (0xFCDE, 'M', 'يه'),
    (0xFCDF, 'M', 'ئم'),
    (0xFCE0, 'M', 'ئه'),
    (0xFCE1, 'M', 'بم'),
    (0xFCE2, 'M', 'به'),
    (0xFCE3, 'M', 'تم'),
    (0xFCE4, 'M', 'ته'),
    (0xFCE5, 'M', 'ثم'),
    (0xFCE6, 'M', 'ثه'),
    (0xFCE7, 'M', 'سم'),
    (0xFCE8, 'M', 'سه'),
    (0xFCE9, 'M', 'شم'),
    (0xFCEA, 'M', 'شه'),
    (0xFCEB, 'M', 'كل'),
    (0xFCEC, 'M', 'كم'),
    (0xFCED, 'M', 'لم'),
    (0xFCEE, 'M', 'نم'),
    (0xFCEF, 'M', 'نه'),
    (0xFCF0, 'M', 'يم'),
    (0xFCF1, 'M', 'يه'),
    (0xFCF2, 'M', 'ـَّ'),
    (0xFCF3, 'M', 'ـُّ'),
    (0xFCF4, 'M', 'ـِّ'),
    (0xFCF5, 'M', 'طى'),
    (0xFCF6, 'M', 'طي'),
    (0xFCF7, 'M', 'عى'),
    (0xFCF8, 'M', 'عي'),
    (0xFCF9, 'M', 'غى'),
    (0xFCFA, 'M', 'غي'),
    (0xFCFB, 'M', 'سى'),
    (0xFCFC, 'M', 'سي'),
    (0xFCFD, 'M', 'شى'),
    (0xFCFE, 'M', 'شي'),
    (0xFCFF, 'M', 'حى'),
    (0xFD00, 'M', 'حي'),
    (0xFD01, 'M', 'جى'),
    (0xFD02, 'M', 'جي'),
    (0xFD03, 'M', 'خى'),
    (0xFD04, 'M', 'خي'),
    (0xFD05, 'M', 'صى'),
    (0xFD06, 'M', 'صي'),
    (0xFD07, 'M', 'ضى'),
    (0xFD08, 'M', 'ضي'),
    (0xFD09, 'M', 'شج'),
    (0xFD0A, 'M', 'شح'),
    (0xFD0B, 'M', 'شخ'),
    (0xFD0C, 'M', 'شم'),
    (0xFD0D, 'M', 'شر'),
    (0xFD0E, 'M', 'سر'),
    (0xFD0F, 'M', 'صر'),
    (0xFD10, 'M', 'ضر'),
    (0xFD11, 'M', 'طى'),
    (0xFD12, 'M', 'طي'),
    (0xFD13, 'M', 'عى'),
    (0xFD14, 'M', 'عي'),
    (0xFD15, 'M', 'غى'),
    (0xFD16, 'M', 'غي'),
    (0xFD17, 'M', 'سى'),
    (0xFD18, 'M', 'سي'),
    (0xFD19, 'M', 'شى'),
    (0xFD1A, 'M', 'شي'),
    (0xFD1B, 'M', 'حى'),
    (0xFD1C, 'M', 'حي'),
    (0xFD1D, 'M', 'جى'),
    (0xFD1E, 'M', 'جي'),
    (0xFD1F, 'M', 'خى'),
    (0xFD20, 'M', 'خي'),
    (0xFD21, 'M', 'صى'),
    (0xFD22, 'M', 'صي'),
    (0xFD23, 'M', 'ضى'),
    (0xFD24, 'M', 'ضي'),
    (0xFD25, 'M', 'شج'),
    (0xFD26, 'M', 'شح'),
    (0xFD27, 'M', 'شخ'),
    ]

def _seg_48():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFD28, 'M', 'شم'),
    (0xFD29, 'M', 'شر'),
    (0xFD2A, 'M', 'سر'),
    (0xFD2B, 'M', 'صر'),
    (0xFD2C, 'M', 'ضر'),
    (0xFD2D, 'M', 'شج'),
    (0xFD2E, 'M', 'شح'),
    (0xFD2F, 'M', 'شخ'),
    (0xFD30, 'M', 'شم'),
    (0xFD31, 'M', 'سه'),
    (0xFD32, 'M', 'شه'),
    (0xFD33, 'M', 'طم'),
    (0xFD34, 'M', 'سج'),
    (0xFD35, 'M', 'سح'),
    (0xFD36, 'M', 'سخ'),
    (0xFD37, 'M', 'شج'),
    (0xFD38, 'M', 'شح'),
    (0xFD39, 'M', 'شخ'),
    (0xFD3A, 'M', 'طم'),
    (0xFD3B, 'M', 'ظم'),
    (0xFD3C, 'M', 'اً'),
    (0xFD3E, 'V'),
    (0xFD40, 'X'),
    (0xFD50, 'M', 'تجم'),
    (0xFD51, 'M', 'تحج'),
    (0xFD53, 'M', 'تحم'),
    (0xFD54, 'M', 'تخم'),
    (0xFD55, 'M', 'تمج'),
    (0xFD56, 'M', 'تمح'),
    (0xFD57, 'M', 'تمخ'),
    (0xFD58, 'M', 'جمح'),
    (0xFD5A, 'M', 'حمي'),
    (0xFD5B, 'M', 'حمى'),
    (0xFD5C, 'M', 'سحج'),
    (0xFD5D, 'M', 'سجح'),
    (0xFD5E, 'M', 'سجى'),
    (0xFD5F, 'M', 'سمح'),
    (0xFD61, 'M', 'سمج'),
    (0xFD62, 'M', 'سمم'),
    (0xFD64, 'M', 'صحح'),
    (0xFD66, 'M', 'صمم'),
    (0xFD67, 'M', 'شحم'),
    (0xFD69, 'M', 'شجي'),
    (0xFD6A, 'M', 'شمخ'),
    (0xFD6C, 'M', 'شمم'),
    (0xFD6E, 'M', 'ضحى'),
    (0xFD6F, 'M', 'ضخم'),
    (0xFD71, 'M', 'طمح'),
    (0xFD73, 'M', 'طمم'),
    (0xFD74, 'M', 'طمي'),
    (0xFD75, 'M', 'عجم'),
    (0xFD76, 'M', 'عمم'),
    (0xFD78, 'M', 'عمى'),
    (0xFD79, 'M', 'غمم'),
    (0xFD7A, 'M', 'غمي'),
    (0xFD7B, 'M', 'غمى'),
    (0xFD7C, 'M', 'فخم'),
    (0xFD7E, 'M', 'قمح'),
    (0xFD7F, 'M', 'قمم'),
    (0xFD80, 'M', 'لحم'),
    (0xFD81, 'M', 'لحي'),
    (0xFD82, 'M', 'لحى'),
    (0xFD83, 'M', 'لجج'),
    (0xFD85, 'M', 'لخم'),
    (0xFD87, 'M', 'لمح'),
    (0xFD89, 'M', 'محج'),
    (0xFD8A, 'M', 'محم'),
    (0xFD8B, 'M', 'محي'),
    (0xFD8C, 'M', 'مجح'),
    (0xFD8D, 'M', 'مجم'),
    (0xFD8E, 'M', 'مخج'),
    (0xFD8F, 'M', 'مخم'),
    (0xFD90, 'X'),
    (0xFD92, 'M', 'مجخ'),
    (0xFD93, 'M', 'همج'),
    (0xFD94, 'M', 'همم'),
    (0xFD95, 'M', 'نحم'),
    (0xFD96, 'M', 'نحى'),
    (0xFD97, 'M', 'نجم'),
    (0xFD99, 'M', 'نجى'),
    (0xFD9A, 'M', 'نمي'),
    (0xFD9B, 'M', 'نمى'),
    (0xFD9C, 'M', 'يمم'),
    (0xFD9E, 'M', 'بخي'),
    (0xFD9F, 'M', 'تجي'),
    (0xFDA0, 'M', 'تجى'),
    (0xFDA1, 'M', 'تخي'),
    (0xFDA2, 'M', 'تخى'),
    (0xFDA3, 'M', 'تمي'),
    (0xFDA4, 'M', 'تمى'),
    (0xFDA5, 'M', 'جمي'),
    (0xFDA6, 'M', 'جحى'),
    (0xFDA7, 'M', 'جمى'),
    (0xFDA8, 'M', 'سخى'),
    (0xFDA9, 'M', 'صحي'),
    (0xFDAA, 'M', 'شحي'),
    (0xFDAB, 'M', 'ضحي'),
    (0xFDAC, 'M', 'لجي'),
    (0xFDAD, 'M', 'لمي'),
    (0xFDAE, 'M', 'يحي'),
    ]

def _seg_49():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFDAF, 'M', 'يجي'),
    (0xFDB0, 'M', 'يمي'),
    (0xFDB1, 'M', 'ممي'),
    (0xFDB2, 'M', 'قمي'),
    (0xFDB3, 'M', 'نحي'),
    (0xFDB4, 'M', 'قمح'),
    (0xFDB5, 'M', 'لحم'),
    (0xFDB6, 'M', 'عمي'),
    (0xFDB7, 'M', 'كمي'),
    (0xFDB8, 'M', 'نجح'),
    (0xFDB9, 'M', 'مخي'),
    (0xFDBA, 'M', 'لجم'),
    (0xFDBB, 'M', 'كمم'),
    (0xFDBC, 'M', 'لجم'),
    (0xFDBD, 'M', 'نجح'),
    (0xFDBE, 'M', 'جحي'),
    (0xFDBF, 'M', 'حجي'),
    (0xFDC0, 'M', 'مجي'),
    (0xFDC1, 'M', 'فمي'),
    (0xFDC2, 'M', 'بحي'),
    (0xFDC3, 'M', 'كمم'),
    (0xFDC4, 'M', 'عجم'),
    (0xFDC5, 'M', 'صمم'),
    (0xFDC6, 'M', 'سخي'),
    (0xFDC7, 'M', 'نجي'),
    (0xFDC8, 'X'),
    (0xFDF0, 'M', 'صلے'),
    (0xFDF1, 'M', 'قلے'),
    (0xFDF2, 'M', 'الله'),
    (0xFDF3, 'M', 'اكبر'),
    (0xFDF4, 'M', 'محمد'),
    (0xFDF5, 'M', 'صلعم'),
    (0xFDF6, 'M', 'رسول'),
    (0xFDF7, 'M', 'عليه'),
    (0xFDF8, 'M', 'وسلم'),
    (0xFDF9, 'M', 'صلى'),
    (0xFDFA, '3', 'صلى الله عليه وسلم'),
    (0xFDFB, '3', 'جل جلاله'),
    (0xFDFC, 'M', 'ریال'),
    (0xFDFD, 'V'),
    (0xFDFE, 'X'),
    (0xFE00, 'I'),
    (0xFE10, '3', ','),
    (0xFE11, 'M', '、'),
    (0xFE12, 'X'),
    (0xFE13, '3', ':'),
    (0xFE14, '3', ';'),
    (0xFE15, '3', '!'),
    (0xFE16, '3', '?'),
    (0xFE17, 'M', '〖'),
    (0xFE18, 'M', '〗'),
    (0xFE19, 'X'),
    (0xFE20, 'V'),
    (0xFE30, 'X'),
    (0xFE31, 'M', '—'),
    (0xFE32, 'M', '–'),
    (0xFE33, '3', '_'),
    (0xFE35, '3', '('),
    (0xFE36, '3', ')'),
    (0xFE37, '3', '{'),
    (0xFE38, '3', '}'),
    (0xFE39, 'M', '〔'),
    (0xFE3A, 'M', '〕'),
    (0xFE3B, 'M', '【'),
    (0xFE3C, 'M', '】'),
    (0xFE3D, 'M', '《'),
    (0xFE3E, 'M', '》'),
    (0xFE3F, 'M', '〈'),
    (0xFE40, 'M', '〉'),
    (0xFE41, 'M', '「'),
    (0xFE42, 'M', '」'),
    (0xFE43, 'M', '『'),
    (0xFE44, 'M', '』'),
    (0xFE45, 'V'),
    (0xFE47, '3', '['),
    (0xFE48, '3', ']'),
    (0xFE49, '3', ' ̅'),
    (0xFE4D, '3', '_'),
    (0xFE50, '3', ','),
    (0xFE51, 'M', '、'),
    (0xFE52, 'X'),
    (0xFE54, '3', ';'),
    (0xFE55, '3', ':'),
    (0xFE56, '3', '?'),
    (0xFE57, '3', '!'),
    (0xFE58, 'M', '—'),
    (0xFE59, '3', '('),
    (0xFE5A, '3', ')'),
    (0xFE5B, '3', '{'),
    (0xFE5C, '3', '}'),
    (0xFE5D, 'M', '〔'),
    (0xFE5E, 'M', '〕'),
    (0xFE5F, '3', '#'),
    (0xFE60, '3', '&'),
    (0xFE61, '3', '*'),
    (0xFE62, '3', '+'),
    (0xFE63, 'M', '-'),
    (0xFE64, '3', '<'),
    (0xFE65, '3', '>'),
    (0xFE66, '3', '='),
    ]

def _seg_50():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFE67, 'X'),
    (0xFE68, '3', '\\'),
    (0xFE69, '3', '$'),
    (0xFE6A, '3', '%'),
    (0xFE6B, '3', '@'),
    (0xFE6C, 'X'),
    (0xFE70, '3', ' ً'),
    (0xFE71, 'M', 'ـً'),
    (0xFE72, '3', ' ٌ'),
    (0xFE73, 'V'),
    (0xFE74, '3', ' ٍ'),
    (0xFE75, 'X'),
    (0xFE76, '3', ' َ'),
    (0xFE77, 'M', 'ـَ'),
    (0xFE78, '3', ' ُ'),
    (0xFE79, 'M', 'ـُ'),
    (0xFE7A, '3', ' ِ'),
    (0xFE7B, 'M', 'ـِ'),
    (0xFE7C, '3', ' ّ'),
    (0xFE7D, 'M', 'ـّ'),
    (0xFE7E, '3', ' ْ'),
    (0xFE7F, 'M', 'ـْ'),
    (0xFE80, 'M', 'ء'),
    (0xFE81, 'M', 'آ'),
    (0xFE83, 'M', 'أ'),
    (0xFE85, 'M', 'ؤ'),
    (0xFE87, 'M', 'إ'),
    (0xFE89, 'M', 'ئ'),
    (0xFE8D, 'M', 'ا'),
    (0xFE8F, 'M', 'ب'),
    (0xFE93, 'M', 'ة'),
    (0xFE95, 'M', 'ت'),
    (0xFE99, 'M', 'ث'),
    (0xFE9D, 'M', 'ج'),
    (0xFEA1, 'M', 'ح'),
    (0xFEA5, 'M', 'خ'),
    (0xFEA9, 'M', 'د'),
    (0xFEAB, 'M', 'ذ'),
    (0xFEAD, 'M', 'ر'),
    (0xFEAF, 'M', 'ز'),
    (0xFEB1, 'M', 'س'),
    (0xFEB5, 'M', 'ش'),
    (0xFEB9, 'M', 'ص'),
    (0xFEBD, 'M', 'ض'),
    (0xFEC1, 'M', 'ط'),
    (0xFEC5, 'M', 'ظ'),
    (0xFEC9, 'M', 'ع'),
    (0xFECD, 'M', 'غ'),
    (0xFED1, 'M', 'ف'),
    (0xFED5, 'M', 'ق'),
    (0xFED9, 'M', 'ك'),
    (0xFEDD, 'M', 'ل'),
    (0xFEE1, 'M', 'م'),
    (0xFEE5, 'M', 'ن'),
    (0xFEE9, 'M', 'ه'),
    (0xFEED, 'M', 'و'),
    (0xFEEF, 'M', 'ى'),
    (0xFEF1, 'M', 'ي'),
    (0xFEF5, 'M', 'لآ'),
    (0xFEF7, 'M', 'لأ'),
    (0xFEF9, 'M', 'لإ'),
    (0xFEFB, 'M', 'لا'),
    (0xFEFD, 'X'),
    (0xFEFF, 'I'),
    (0xFF00, 'X'),
    (0xFF01, '3', '!'),
    (0xFF02, '3', '"'),
    (0xFF03, '3', '#'),
    (0xFF04, '3', '$'),
    (0xFF05, '3', '%'),
    (0xFF06, '3', '&'),
    (0xFF07, '3', '\''),
    (0xFF08, '3', '('),
    (0xFF09, '3', ')'),
    (0xFF0A, '3', '*'),
    (0xFF0B, '3', '+'),
    (0xFF0C, '3', ','),
    (0xFF0D, 'M', '-'),
    (0xFF0E, 'M', '.'),
    (0xFF0F, '3', '/'),
    (0xFF10, 'M', '0'),
    (0xFF11, 'M', '1'),
    (0xFF12, 'M', '2'),
    (0xFF13, 'M', '3'),
    (0xFF14, 'M', '4'),
    (0xFF15, 'M', '5'),
    (0xFF16, 'M', '6'),
    (0xFF17, 'M', '7'),
    (0xFF18, 'M', '8'),
    (0xFF19, 'M', '9'),
    (0xFF1A, '3', ':'),
    (0xFF1B, '3', ';'),
    (0xFF1C, '3', '<'),
    (0xFF1D, '3', '='),
    (0xFF1E, '3', '>'),
    (0xFF1F, '3', '?'),
    (0xFF20, '3', '@'),
    (0xFF21, 'M', 'a'),
    (0xFF22, 'M', 'b'),
    (0xFF23, 'M', 'c'),
    ]

def _seg_51():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFF24, 'M', 'd'),
    (0xFF25, 'M', 'e'),
    (0xFF26, 'M', 'f'),
    (0xFF27, 'M', 'g'),
    (0xFF28, 'M', 'h'),
    (0xFF29, 'M', 'i'),
    (0xFF2A, 'M', 'j'),
    (0xFF2B, 'M', 'k'),
    (0xFF2C, 'M', 'l'),
    (0xFF2D, 'M', 'm'),
    (0xFF2E, 'M', 'n'),
    (0xFF2F, 'M', 'o'),
    (0xFF30, 'M', 'p'),
    (0xFF31, 'M', 'q'),
    (0xFF32, 'M', 'r'),
    (0xFF33, 'M', 's'),
    (0xFF34, 'M', 't'),
    (0xFF35, 'M', 'u'),
    (0xFF36, 'M', 'v'),
    (0xFF37, 'M', 'w'),
    (0xFF38, 'M', 'x'),
    (0xFF39, 'M', 'y'),
    (0xFF3A, 'M', 'z'),
    (0xFF3B, '3', '['),
    (0xFF3C, '3', '\\'),
    (0xFF3D, '3', ']'),
    (0xFF3E, '3', '^'),
    (0xFF3F, '3', '_'),
    (0xFF40, '3', '`'),
    (0xFF41, 'M', 'a'),
    (0xFF42, 'M', 'b'),
    (0xFF43, 'M', 'c'),
    (0xFF44, 'M', 'd'),
    (0xFF45, 'M', 'e'),
    (0xFF46, 'M', 'f'),
    (0xFF47, 'M', 'g'),
    (0xFF48, 'M', 'h'),
    (0xFF49, 'M', 'i'),
    (0xFF4A, 'M', 'j'),
    (0xFF4B, 'M', 'k'),
    (0xFF4C, 'M', 'l'),
    (0xFF4D, 'M', 'm'),
    (0xFF4E, 'M', 'n'),
    (0xFF4F, 'M', 'o'),
    (0xFF50, 'M', 'p'),
    (0xFF51, 'M', 'q'),
    (0xFF52, 'M', 'r'),
    (0xFF53, 'M', 's'),
    (0xFF54, 'M', 't'),
    (0xFF55, 'M', 'u'),
    (0xFF56, 'M', 'v'),
    (0xFF57, 'M', 'w'),
    (0xFF58, 'M', 'x'),
    (0xFF59, 'M', 'y'),
    (0xFF5A, 'M', 'z'),
    (0xFF5B, '3', '{'),
    (0xFF5C, '3', '|'),
    (0xFF5D, '3', '}'),
    (0xFF5E, '3', '~'),
    (0xFF5F, 'M', '⦅'),
    (0xFF60, 'M', '⦆'),
    (0xFF61, 'M', '.'),
    (0xFF62, 'M', '「'),
    (0xFF63, 'M', '」'),
    (0xFF64, 'M', '、'),
    (0xFF65, 'M', '・'),
    (0xFF66, 'M', 'ヲ'),
    (0xFF67, 'M', 'ァ'),
    (0xFF68, 'M', 'ィ'),
    (0xFF69, 'M', 'ゥ'),
    (0xFF6A, 'M', 'ェ'),
    (0xFF6B, 'M', 'ォ'),
    (0xFF6C, 'M', 'ャ'),
    (0xFF6D, 'M', 'ュ'),
    (0xFF6E, 'M', 'ョ'),
    (0xFF6F, 'M', 'ッ'),
    (0xFF70, 'M', 'ー'),
    (0xFF71, 'M', 'ア'),
    (0xFF72, 'M', 'イ'),
    (0xFF73, 'M', 'ウ'),
    (0xFF74, 'M', 'エ'),
    (0xFF75, 'M', 'オ'),
    (0xFF76, 'M', 'カ'),
    (0xFF77, 'M', 'キ'),
    (0xFF78, 'M', 'ク'),
    (0xFF79, 'M', 'ケ'),
    (0xFF7A, 'M', 'コ'),
    (0xFF7B, 'M', 'サ'),
    (0xFF7C, 'M', 'シ'),
    (0xFF7D, 'M', 'ス'),
    (0xFF7E, 'M', 'セ'),
    (0xFF7F, 'M', 'ソ'),
    (0xFF80, 'M', 'タ'),
    (0xFF81, 'M', 'チ'),
    (0xFF82, 'M', 'ツ'),
    (0xFF83, 'M', 'テ'),
    (0xFF84, 'M', 'ト'),
    (0xFF85, 'M', 'ナ'),
    (0xFF86, 'M', 'ニ'),
    (0xFF87, 'M', 'ヌ'),
    ]

def _seg_52():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0xFF88, 'M', 'ネ'),
    (0xFF89, 'M', 'ノ'),
    (0xFF8A, 'M', 'ハ'),
    (0xFF8B, 'M', 'ヒ'),
    (0xFF8C, 'M', 'フ'),
    (0xFF8D, 'M', 'ヘ'),
    (0xFF8E, 'M', 'ホ'),
    (0xFF8F, 'M', 'マ'),
    (0xFF90, 'M', 'ミ'),
    (0xFF91, 'M', 'ム'),
    (0xFF92, 'M', 'メ'),
    (0xFF93, 'M', 'モ'),
    (0xFF94, 'M', 'ヤ'),
    (0xFF95, 'M', 'ユ'),
    (0xFF96, 'M', 'ヨ'),
    (0xFF97, 'M', 'ラ'),
    (0xFF98, 'M', 'リ'),
    (0xFF99, 'M', 'ル'),
    (0xFF9A, 'M', 'レ'),
    (0xFF9B, 'M', 'ロ'),
    (0xFF9C, 'M', 'ワ'),
    (0xFF9D, 'M', 'ン'),
    (0xFF9E, 'M', '゙'),
    (0xFF9F, 'M', '゚'),
    (0xFFA0, 'X'),
    (0xFFA1, 'M', 'ᄀ'),
    (0xFFA2, 'M', 'ᄁ'),
    (0xFFA3, 'M', 'ᆪ'),
    (0xFFA4, 'M', 'ᄂ'),
    (0xFFA5, 'M', 'ᆬ'),
    (0xFFA6, 'M', 'ᆭ'),
    (0xFFA7, 'M', 'ᄃ'),
    (0xFFA8, 'M', 'ᄄ'),
    (0xFFA9, 'M', 'ᄅ'),
    (0xFFAA, 'M', 'ᆰ'),
    (0xFFAB, 'M', 'ᆱ'),
    (0xFFAC, 'M', 'ᆲ'),
    (0xFFAD, 'M', 'ᆳ'),
    (0xFFAE, 'M', 'ᆴ'),
    (0xFFAF, 'M', 'ᆵ'),
    (0xFFB0, 'M', 'ᄚ'),
    (0xFFB1, 'M', 'ᄆ'),
    (0xFFB2, 'M', 'ᄇ'),
    (0xFFB3, 'M', 'ᄈ'),
    (0xFFB4, 'M', 'ᄡ'),
    (0xFFB5, 'M', 'ᄉ'),
    (0xFFB6, 'M', 'ᄊ'),
    (0xFFB7, 'M', 'ᄋ'),
    (0xFFB8, 'M', 'ᄌ'),
    (0xFFB9, 'M', 'ᄍ'),
    (0xFFBA, 'M', 'ᄎ'),
    (0xFFBB, 'M', 'ᄏ'),
    (0xFFBC, 'M', 'ᄐ'),
    (0xFFBD, 'M', 'ᄑ'),
    (0xFFBE, 'M', 'ᄒ'),
    (0xFFBF, 'X'),
    (0xFFC2, 'M', 'ᅡ'),
    (0xFFC3, 'M', 'ᅢ'),
    (0xFFC4, 'M', 'ᅣ'),
    (0xFFC5, 'M', 'ᅤ'),
    (0xFFC6, 'M', 'ᅥ'),
    (0xFFC7, 'M', 'ᅦ'),
    (0xFFC8, 'X'),
    (0xFFCA, 'M', 'ᅧ'),
    (0xFFCB, 'M', 'ᅨ'),
    (0xFFCC, 'M', 'ᅩ'),
    (0xFFCD, 'M', 'ᅪ'),
    (0xFFCE, 'M', 'ᅫ'),
    (0xFFCF, 'M', 'ᅬ'),
    (0xFFD0, 'X'),
    (0xFFD2, 'M', 'ᅭ'),
    (0xFFD3, 'M', 'ᅮ'),
    (0xFFD4, 'M', 'ᅯ'),
    (0xFFD5, 'M', 'ᅰ'),
    (0xFFD6, 'M', 'ᅱ'),
    (0xFFD7, 'M', 'ᅲ'),
    (0xFFD8, 'X'),
    (0xFFDA, 'M', 'ᅳ'),
    (0xFFDB, 'M', 'ᅴ'),
    (0xFFDC, 'M', 'ᅵ'),
    (0xFFDD, 'X'),
    (0xFFE0, 'M', '¢'),
    (0xFFE1, 'M', '£'),
    (0xFFE2, 'M', '¬'),
    (0xFFE3, '3', ' ̄'),
    (0xFFE4, 'M', '¦'),
    (0xFFE5, 'M', '¥'),
    (0xFFE6, 'M', '₩'),
    (0xFFE7, 'X'),
    (0xFFE8, 'M', '│'),
    (0xFFE9, 'M', '←'),
    (0xFFEA, 'M', '↑'),
    (0xFFEB, 'M', '→'),
    (0xFFEC, 'M', '↓'),
    (0xFFED, 'M', '■'),
    (0xFFEE, 'M', '○'),
    (0xFFEF, 'X'),
    (0x10000, 'V'),
    (0x1000C, 'X'),
    (0x1000D, 'V'),
    ]

def _seg_53():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x10027, 'X'),
    (0x10028, 'V'),
    (0x1003B, 'X'),
    (0x1003C, 'V'),
    (0x1003E, 'X'),
    (0x1003F, 'V'),
    (0x1004E, 'X'),
    (0x10050, 'V'),
    (0x1005E, 'X'),
    (0x10080, 'V'),
    (0x100FB, 'X'),
    (0x10100, 'V'),
    (0x10103, 'X'),
    (0x10107, 'V'),
    (0x10134, 'X'),
    (0x10137, 'V'),
    (0x1018F, 'X'),
    (0x10190, 'V'),
    (0x1019D, 'X'),
    (0x101A0, 'V'),
    (0x101A1, 'X'),
    (0x101D0, 'V'),
    (0x101FE, 'X'),
    (0x10280, 'V'),
    (0x1029D, 'X'),
    (0x102A0, 'V'),
    (0x102D1, 'X'),
    (0x102E0, 'V'),
    (0x102FC, 'X'),
    (0x10300, 'V'),
    (0x10324, 'X'),
    (0x1032D, 'V'),
    (0x1034B, 'X'),
    (0x10350, 'V'),
    (0x1037B, 'X'),
    (0x10380, 'V'),
    (0x1039E, 'X'),
    (0x1039F, 'V'),
    (0x103C4, 'X'),
    (0x103C8, 'V'),
    (0x103D6, 'X'),
    (0x10400, 'M', '𐐨'),
    (0x10401, 'M', '𐐩'),
    (0x10402, 'M', '𐐪'),
    (0x10403, 'M', '𐐫'),
    (0x10404, 'M', '𐐬'),
    (0x10405, 'M', '𐐭'),
    (0x10406, 'M', '𐐮'),
    (0x10407, 'M', '𐐯'),
    (0x10408, 'M', '𐐰'),
    (0x10409, 'M', '𐐱'),
    (0x1040A, 'M', '𐐲'),
    (0x1040B, 'M', '𐐳'),
    (0x1040C, 'M', '𐐴'),
    (0x1040D, 'M', '𐐵'),
    (0x1040E, 'M', '𐐶'),
    (0x1040F, 'M', '𐐷'),
    (0x10410, 'M', '𐐸'),
    (0x10411, 'M', '𐐹'),
    (0x10412, 'M', '𐐺'),
    (0x10413, 'M', '𐐻'),
    (0x10414, 'M', '𐐼'),
    (0x10415, 'M', '𐐽'),
    (0x10416, 'M', '𐐾'),
    (0x10417, 'M', '𐐿'),
    (0x10418, 'M', '𐑀'),
    (0x10419, 'M', '𐑁'),
    (0x1041A, 'M', '𐑂'),
    (0x1041B, 'M', '𐑃'),
    (0x1041C, 'M', '𐑄'),
    (0x1041D, 'M', '𐑅'),
    (0x1041E, 'M', '𐑆'),
    (0x1041F, 'M', '𐑇'),
    (0x10420, 'M', '𐑈'),
    (0x10421, 'M', '𐑉'),
    (0x10422, 'M', '𐑊'),
    (0x10423, 'M', '𐑋'),
    (0x10424, 'M', '𐑌'),
    (0x10425, 'M', '𐑍'),
    (0x10426, 'M', '𐑎'),
    (0x10427, 'M', '𐑏'),
    (0x10428, 'V'),
    (0x1049E, 'X'),
    (0x104A0, 'V'),
    (0x104AA, 'X'),
    (0x104B0, 'M', '𐓘'),
    (0x104B1, 'M', '𐓙'),
    (0x104B2, 'M', '𐓚'),
    (0x104B3, 'M', '𐓛'),
    (0x104B4, 'M', '𐓜'),
    (0x104B5, 'M', '𐓝'),
    (0x104B6, 'M', '𐓞'),
    (0x104B7, 'M', '𐓟'),
    (0x104B8, 'M', '𐓠'),
    (0x104B9, 'M', '𐓡'),
    (0x104BA, 'M', '𐓢'),
    (0x104BB, 'M', '𐓣'),
    (0x104BC, 'M', '𐓤'),
    (0x104BD, 'M', '𐓥'),
    (0x104BE, 'M', '𐓦'),
    ]

def _seg_54():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x104BF, 'M', '𐓧'),
    (0x104C0, 'M', '𐓨'),
    (0x104C1, 'M', '𐓩'),
    (0x104C2, 'M', '𐓪'),
    (0x104C3, 'M', '𐓫'),
    (0x104C4, 'M', '𐓬'),
    (0x104C5, 'M', '𐓭'),
    (0x104C6, 'M', '𐓮'),
    (0x104C7, 'M', '𐓯'),
    (0x104C8, 'M', '𐓰'),
    (0x104C9, 'M', '𐓱'),
    (0x104CA, 'M', '𐓲'),
    (0x104CB, 'M', '𐓳'),
    (0x104CC, 'M', '𐓴'),
    (0x104CD, 'M', '𐓵'),
    (0x104CE, 'M', '𐓶'),
    (0x104CF, 'M', '𐓷'),
    (0x104D0, 'M', '𐓸'),
    (0x104D1, 'M', '𐓹'),
    (0x104D2, 'M', '𐓺'),
    (0x104D3, 'M', '𐓻'),
    (0x104D4, 'X'),
    (0x104D8, 'V'),
    (0x104FC, 'X'),
    (0x10500, 'V'),
    (0x10528, 'X'),
    (0x10530, 'V'),
    (0x10564, 'X'),
    (0x1056F, 'V'),
    (0x10570, 'X'),
    (0x10600, 'V'),
    (0x10737, 'X'),
    (0x10740, 'V'),
    (0x10756, 'X'),
    (0x10760, 'V'),
    (0x10768, 'X'),
    (0x10800, 'V'),
    (0x10806, 'X'),
    (0x10808, 'V'),
    (0x10809, 'X'),
    (0x1080A, 'V'),
    (0x10836, 'X'),
    (0x10837, 'V'),
    (0x10839, 'X'),
    (0x1083C, 'V'),
    (0x1083D, 'X'),
    (0x1083F, 'V'),
    (0x10856, 'X'),
    (0x10857, 'V'),
    (0x1089F, 'X'),
    (0x108A7, 'V'),
    (0x108B0, 'X'),
    (0x108E0, 'V'),
    (0x108F3, 'X'),
    (0x108F4, 'V'),
    (0x108F6, 'X'),
    (0x108FB, 'V'),
    (0x1091C, 'X'),
    (0x1091F, 'V'),
    (0x1093A, 'X'),
    (0x1093F, 'V'),
    (0x10940, 'X'),
    (0x10980, 'V'),
    (0x109B8, 'X'),
    (0x109BC, 'V'),
    (0x109D0, 'X'),
    (0x109D2, 'V'),
    (0x10A04, 'X'),
    (0x10A05, 'V'),
    (0x10A07, 'X'),
    (0x10A0C, 'V'),
    (0x10A14, 'X'),
    (0x10A15, 'V'),
    (0x10A18, 'X'),
    (0x10A19, 'V'),
    (0x10A36, 'X'),
    (0x10A38, 'V'),
    (0x10A3B, 'X'),
    (0x10A3F, 'V'),
    (0x10A49, 'X'),
    (0x10A50, 'V'),
    (0x10A59, 'X'),
    (0x10A60, 'V'),
    (0x10AA0, 'X'),
    (0x10AC0, 'V'),
    (0x10AE7, 'X'),
    (0x10AEB, 'V'),
    (0x10AF7, 'X'),
    (0x10B00, 'V'),
    (0x10B36, 'X'),
    (0x10B39, 'V'),
    (0x10B56, 'X'),
    (0x10B58, 'V'),
    (0x10B73, 'X'),
    (0x10B78, 'V'),
    (0x10B92, 'X'),
    (0x10B99, 'V'),
    (0x10B9D, 'X'),
    (0x10BA9, 'V'),
    (0x10BB0, 'X'),
    ]

def _seg_55():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x10C00, 'V'),
    (0x10C49, 'X'),
    (0x10C80, 'M', '𐳀'),
    (0x10C81, 'M', '𐳁'),
    (0x10C82, 'M', '𐳂'),
    (0x10C83, 'M', '𐳃'),
    (0x10C84, 'M', '𐳄'),
    (0x10C85, 'M', '𐳅'),
    (0x10C86, 'M', '𐳆'),
    (0x10C87, 'M', '𐳇'),
    (0x10C88, 'M', '𐳈'),
    (0x10C89, 'M', '𐳉'),
    (0x10C8A, 'M', '𐳊'),
    (0x10C8B, 'M', '𐳋'),
    (0x10C8C, 'M', '𐳌'),
    (0x10C8D, 'M', '𐳍'),
    (0x10C8E, 'M', '𐳎'),
    (0x10C8F, 'M', '𐳏'),
    (0x10C90, 'M', '𐳐'),
    (0x10C91, 'M', '𐳑'),
    (0x10C92, 'M', '𐳒'),
    (0x10C93, 'M', '𐳓'),
    (0x10C94, 'M', '𐳔'),
    (0x10C95, 'M', '𐳕'),
    (0x10C96, 'M', '𐳖'),
    (0x10C97, 'M', '𐳗'),
    (0x10C98, 'M', '𐳘'),
    (0x10C99, 'M', '𐳙'),
    (0x10C9A, 'M', '𐳚'),
    (0x10C9B, 'M', '𐳛'),
    (0x10C9C, 'M', '𐳜'),
    (0x10C9D, 'M', '𐳝'),
    (0x10C9E, 'M', '𐳞'),
    (0x10C9F, 'M', '𐳟'),
    (0x10CA0, 'M', '𐳠'),
    (0x10CA1, 'M', '𐳡'),
    (0x10CA2, 'M', '𐳢'),
    (0x10CA3, 'M', '𐳣'),
    (0x10CA4, 'M', '𐳤'),
    (0x10CA5, 'M', '𐳥'),
    (0x10CA6, 'M', '𐳦'),
    (0x10CA7, 'M', '𐳧'),
    (0x10CA8, 'M', '𐳨'),
    (0x10CA9, 'M', '𐳩'),
    (0x10CAA, 'M', '𐳪'),
    (0x10CAB, 'M', '𐳫'),
    (0x10CAC, 'M', '𐳬'),
    (0x10CAD, 'M', '𐳭'),
    (0x10CAE, 'M', '𐳮'),
    (0x10CAF, 'M', '𐳯'),
    (0x10CB0, 'M', '𐳰'),
    (0x10CB1, 'M', '𐳱'),
    (0x10CB2, 'M', '𐳲'),
    (0x10CB3, 'X'),
    (0x10CC0, 'V'),
    (0x10CF3, 'X'),
    (0x10CFA, 'V'),
    (0x10D28, 'X'),
    (0x10D30, 'V'),
    (0x10D3A, 'X'),
    (0x10E60, 'V'),
    (0x10E7F, 'X'),
    (0x10E80, 'V'),
    (0x10EAA, 'X'),
    (0x10EAB, 'V'),
    (0x10EAE, 'X'),
    (0x10EB0, 'V'),
    (0x10EB2, 'X'),
    (0x10F00, 'V'),
    (0x10F28, 'X'),
    (0x10F30, 'V'),
    (0x10F5A, 'X'),
    (0x10FB0, 'V'),
    (0x10FCC, 'X'),
    (0x10FE0, 'V'),
    (0x10FF7, 'X'),
    (0x11000, 'V'),
    (0x1104E, 'X'),
    (0x11052, 'V'),
    (0x11070, 'X'),
    (0x1107F, 'V'),
    (0x110BD, 'X'),
    (0x110BE, 'V'),
    (0x110C2, 'X'),
    (0x110D0, 'V'),
    (0x110E9, 'X'),
    (0x110F0, 'V'),
    (0x110FA, 'X'),
    (0x11100, 'V'),
    (0x11135, 'X'),
    (0x11136, 'V'),
    (0x11148, 'X'),
    (0x11150, 'V'),
    (0x11177, 'X'),
    (0x11180, 'V'),
    (0x111E0, 'X'),
    (0x111E1, 'V'),
    (0x111F5, 'X'),
    (0x11200, 'V'),
    (0x11212, 'X'),
    ]

def _seg_56():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x11213, 'V'),
    (0x1123F, 'X'),
    (0x11280, 'V'),
    (0x11287, 'X'),
    (0x11288, 'V'),
    (0x11289, 'X'),
    (0x1128A, 'V'),
    (0x1128E, 'X'),
    (0x1128F, 'V'),
    (0x1129E, 'X'),
    (0x1129F, 'V'),
    (0x112AA, 'X'),
    (0x112B0, 'V'),
    (0x112EB, 'X'),
    (0x112F0, 'V'),
    (0x112FA, 'X'),
    (0x11300, 'V'),
    (0x11304, 'X'),
    (0x11305, 'V'),
    (0x1130D, 'X'),
    (0x1130F, 'V'),
    (0x11311, 'X'),
    (0x11313, 'V'),
    (0x11329, 'X'),
    (0x1132A, 'V'),
    (0x11331, 'X'),
    (0x11332, 'V'),
    (0x11334, 'X'),
    (0x11335, 'V'),
    (0x1133A, 'X'),
    (0x1133B, 'V'),
    (0x11345, 'X'),
    (0x11347, 'V'),
    (0x11349, 'X'),
    (0x1134B, 'V'),
    (0x1134E, 'X'),
    (0x11350, 'V'),
    (0x11351, 'X'),
    (0x11357, 'V'),
    (0x11358, 'X'),
    (0x1135D, 'V'),
    (0x11364, 'X'),
    (0x11366, 'V'),
    (0x1136D, 'X'),
    (0x11370, 'V'),
    (0x11375, 'X'),
    (0x11400, 'V'),
    (0x1145C, 'X'),
    (0x1145D, 'V'),
    (0x11462, 'X'),
    (0x11480, 'V'),
    (0x114C8, 'X'),
    (0x114D0, 'V'),
    (0x114DA, 'X'),
    (0x11580, 'V'),
    (0x115B6, 'X'),
    (0x115B8, 'V'),
    (0x115DE, 'X'),
    (0x11600, 'V'),
    (0x11645, 'X'),
    (0x11650, 'V'),
    (0x1165A, 'X'),
    (0x11660, 'V'),
    (0x1166D, 'X'),
    (0x11680, 'V'),
    (0x116B9, 'X'),
    (0x116C0, 'V'),
    (0x116CA, 'X'),
    (0x11700, 'V'),
    (0x1171B, 'X'),
    (0x1171D, 'V'),
    (0x1172C, 'X'),
    (0x11730, 'V'),
    (0x11740, 'X'),
    (0x11800, 'V'),
    (0x1183C, 'X'),
    (0x118A0, 'M', '𑣀'),
    (0x118A1, 'M', '𑣁'),
    (0x118A2, 'M', '𑣂'),
    (0x118A3, 'M', '𑣃'),
    (0x118A4, 'M', '𑣄'),
    (0x118A5, 'M', '𑣅'),
    (0x118A6, 'M', '𑣆'),
    (0x118A7, 'M', '𑣇'),
    (0x118A8, 'M', '𑣈'),
    (0x118A9, 'M', '𑣉'),
    (0x118AA, 'M', '𑣊'),
    (0x118AB, 'M', '𑣋'),
    (0x118AC, 'M', '𑣌'),
    (0x118AD, 'M', '𑣍'),
    (0x118AE, 'M', '𑣎'),
    (0x118AF, 'M', '𑣏'),
    (0x118B0, 'M', '𑣐'),
    (0x118B1, 'M', '𑣑'),
    (0x118B2, 'M', '𑣒'),
    (0x118B3, 'M', '𑣓'),
    (0x118B4, 'M', '𑣔'),
    (0x118B5, 'M', '𑣕'),
    (0x118B6, 'M', '𑣖'),
    (0x118B7, 'M', '𑣗'),
    ]

def _seg_57():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x118B8, 'M', '𑣘'),
    (0x118B9, 'M', '𑣙'),
    (0x118BA, 'M', '𑣚'),
    (0x118BB, 'M', '𑣛'),
    (0x118BC, 'M', '𑣜'),
    (0x118BD, 'M', '𑣝'),
    (0x118BE, 'M', '𑣞'),
    (0x118BF, 'M', '𑣟'),
    (0x118C0, 'V'),
    (0x118F3, 'X'),
    (0x118FF, 'V'),
    (0x11907, 'X'),
    (0x11909, 'V'),
    (0x1190A, 'X'),
    (0x1190C, 'V'),
    (0x11914, 'X'),
    (0x11915, 'V'),
    (0x11917, 'X'),
    (0x11918, 'V'),
    (0x11936, 'X'),
    (0x11937, 'V'),
    (0x11939, 'X'),
    (0x1193B, 'V'),
    (0x11947, 'X'),
    (0x11950, 'V'),
    (0x1195A, 'X'),
    (0x119A0, 'V'),
    (0x119A8, 'X'),
    (0x119AA, 'V'),
    (0x119D8, 'X'),
    (0x119DA, 'V'),
    (0x119E5, 'X'),
    (0x11A00, 'V'),
    (0x11A48, 'X'),
    (0x11A50, 'V'),
    (0x11AA3, 'X'),
    (0x11AC0, 'V'),
    (0x11AF9, 'X'),
    (0x11C00, 'V'),
    (0x11C09, 'X'),
    (0x11C0A, 'V'),
    (0x11C37, 'X'),
    (0x11C38, 'V'),
    (0x11C46, 'X'),
    (0x11C50, 'V'),
    (0x11C6D, 'X'),
    (0x11C70, 'V'),
    (0x11C90, 'X'),
    (0x11C92, 'V'),
    (0x11CA8, 'X'),
    (0x11CA9, 'V'),
    (0x11CB7, 'X'),
    (0x11D00, 'V'),
    (0x11D07, 'X'),
    (0x11D08, 'V'),
    (0x11D0A, 'X'),
    (0x11D0B, 'V'),
    (0x11D37, 'X'),
    (0x11D3A, 'V'),
    (0x11D3B, 'X'),
    (0x11D3C, 'V'),
    (0x11D3E, 'X'),
    (0x11D3F, 'V'),
    (0x11D48, 'X'),
    (0x11D50, 'V'),
    (0x11D5A, 'X'),
    (0x11D60, 'V'),
    (0x11D66, 'X'),
    (0x11D67, 'V'),
    (0x11D69, 'X'),
    (0x11D6A, 'V'),
    (0x11D8F, 'X'),
    (0x11D90, 'V'),
    (0x11D92, 'X'),
    (0x11D93, 'V'),
    (0x11D99, 'X'),
    (0x11DA0, 'V'),
    (0x11DAA, 'X'),
    (0x11EE0, 'V'),
    (0x11EF9, 'X'),
    (0x11FB0, 'V'),
    (0x11FB1, 'X'),
    (0x11FC0, 'V'),
    (0x11FF2, 'X'),
    (0x11FFF, 'V'),
    (0x1239A, 'X'),
    (0x12400, 'V'),
    (0x1246F, 'X'),
    (0x12470, 'V'),
    (0x12475, 'X'),
    (0x12480, 'V'),
    (0x12544, 'X'),
    (0x13000, 'V'),
    (0x1342F, 'X'),
    (0x14400, 'V'),
    (0x14647, 'X'),
    (0x16800, 'V'),
    (0x16A39, 'X'),
    (0x16A40, 'V'),
    (0x16A5F, 'X'),
    ]

def _seg_58():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x16A60, 'V'),
    (0x16A6A, 'X'),
    (0x16A6E, 'V'),
    (0x16A70, 'X'),
    (0x16AD0, 'V'),
    (0x16AEE, 'X'),
    (0x16AF0, 'V'),
    (0x16AF6, 'X'),
    (0x16B00, 'V'),
    (0x16B46, 'X'),
    (0x16B50, 'V'),
    (0x16B5A, 'X'),
    (0x16B5B, 'V'),
    (0x16B62, 'X'),
    (0x16B63, 'V'),
    (0x16B78, 'X'),
    (0x16B7D, 'V'),
    (0x16B90, 'X'),
    (0x16E40, 'M', '𖹠'),
    (0x16E41, 'M', '𖹡'),
    (0x16E42, 'M', '𖹢'),
    (0x16E43, 'M', '𖹣'),
    (0x16E44, 'M', '𖹤'),
    (0x16E45, 'M', '𖹥'),
    (0x16E46, 'M', '𖹦'),
    (0x16E47, 'M', '𖹧'),
    (0x16E48, 'M', '𖹨'),
    (0x16E49, 'M', '𖹩'),
    (0x16E4A, 'M', '𖹪'),
    (0x16E4B, 'M', '𖹫'),
    (0x16E4C, 'M', '𖹬'),
    (0x16E4D, 'M', '𖹭'),
    (0x16E4E, 'M', '𖹮'),
    (0x16E4F, 'M', '𖹯'),
    (0x16E50, 'M', '𖹰'),
    (0x16E51, 'M', '𖹱'),
    (0x16E52, 'M', '𖹲'),
    (0x16E53, 'M', '𖹳'),
    (0x16E54, 'M', '𖹴'),
    (0x16E55, 'M', '𖹵'),
    (0x16E56, 'M', '𖹶'),
    (0x16E57, 'M', '𖹷'),
    (0x16E58, 'M', '𖹸'),
    (0x16E59, 'M', '𖹹'),
    (0x16E5A, 'M', '𖹺'),
    (0x16E5B, 'M', '𖹻'),
    (0x16E5C, 'M', '𖹼'),
    (0x16E5D, 'M', '𖹽'),
    (0x16E5E, 'M', '𖹾'),
    (0x16E5F, 'M', '𖹿'),
    (0x16E60, 'V'),
    (0x16E9B, 'X'),
    (0x16F00, 'V'),
    (0x16F4B, 'X'),
    (0x16F4F, 'V'),
    (0x16F88, 'X'),
    (0x16F8F, 'V'),
    (0x16FA0, 'X'),
    (0x16FE0, 'V'),
    (0x16FE5, 'X'),
    (0x16FF0, 'V'),
    (0x16FF2, 'X'),
    (0x17000, 'V'),
    (0x187F8, 'X'),
    (0x18800, 'V'),
    (0x18CD6, 'X'),
    (0x18D00, 'V'),
    (0x18D09, 'X'),
    (0x1B000, 'V'),
    (0x1B11F, 'X'),
    (0x1B150, 'V'),
    (0x1B153, 'X'),
    (0x1B164, 'V'),
    (0x1B168, 'X'),
    (0x1B170, 'V'),
    (0x1B2FC, 'X'),
    (0x1BC00, 'V'),
    (0x1BC6B, 'X'),
    (0x1BC70, 'V'),
    (0x1BC7D, 'X'),
    (0x1BC80, 'V'),
    (0x1BC89, 'X'),
    (0x1BC90, 'V'),
    (0x1BC9A, 'X'),
    (0x1BC9C, 'V'),
    (0x1BCA0, 'I'),
    (0x1BCA4, 'X'),
    (0x1D000, 'V'),
    (0x1D0F6, 'X'),
    (0x1D100, 'V'),
    (0x1D127, 'X'),
    (0x1D129, 'V'),
    (0x1D15E, 'M', '𝅗𝅥'),
    (0x1D15F, 'M', '𝅘𝅥'),
    (0x1D160, 'M', '𝅘𝅥𝅮'),
    (0x1D161, 'M', '𝅘𝅥𝅯'),
    (0x1D162, 'M', '𝅘𝅥𝅰'),
    (0x1D163, 'M', '𝅘𝅥𝅱'),
    (0x1D164, 'M', '𝅘𝅥𝅲'),
    (0x1D165, 'V'),
    ]

def _seg_59():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D173, 'X'),
    (0x1D17B, 'V'),
    (0x1D1BB, 'M', '𝆹𝅥'),
    (0x1D1BC, 'M', '𝆺𝅥'),
    (0x1D1BD, 'M', '𝆹𝅥𝅮'),
    (0x1D1BE, 'M', '𝆺𝅥𝅮'),
    (0x1D1BF, 'M', '𝆹𝅥𝅯'),
    (0x1D1C0, 'M', '𝆺𝅥𝅯'),
    (0x1D1C1, 'V'),
    (0x1D1E9, 'X'),
    (0x1D200, 'V'),
    (0x1D246, 'X'),
    (0x1D2E0, 'V'),
    (0x1D2F4, 'X'),
    (0x1D300, 'V'),
    (0x1D357, 'X'),
    (0x1D360, 'V'),
    (0x1D379, 'X'),
    (0x1D400, 'M', 'a'),
    (0x1D401, 'M', 'b'),
    (0x1D402, 'M', 'c'),
    (0x1D403, 'M', 'd'),
    (0x1D404, 'M', 'e'),
    (0x1D405, 'M', 'f'),
    (0x1D406, 'M', 'g'),
    (0x1D407, 'M', 'h'),
    (0x1D408, 'M', 'i'),
    (0x1D409, 'M', 'j'),
    (0x1D40A, 'M', 'k'),
    (0x1D40B, 'M', 'l'),
    (0x1D40C, 'M', 'm'),
    (0x1D40D, 'M', 'n'),
    (0x1D40E, 'M', 'o'),
    (0x1D40F, 'M', 'p'),
    (0x1D410, 'M', 'q'),
    (0x1D411, 'M', 'r'),
    (0x1D412, 'M', 's'),
    (0x1D413, 'M', 't'),
    (0x1D414, 'M', 'u'),
    (0x1D415, 'M', 'v'),
    (0x1D416, 'M', 'w'),
    (0x1D417, 'M', 'x'),
    (0x1D418, 'M', 'y'),
    (0x1D419, 'M', 'z'),
    (0x1D41A, 'M', 'a'),
    (0x1D41B, 'M', 'b'),
    (0x1D41C, 'M', 'c'),
    (0x1D41D, 'M', 'd'),
    (0x1D41E, 'M', 'e'),
    (0x1D41F, 'M', 'f'),
    (0x1D420, 'M', 'g'),
    (0x1D421, 'M', 'h'),
    (0x1D422, 'M', 'i'),
    (0x1D423, 'M', 'j'),
    (0x1D424, 'M', 'k'),
    (0x1D425, 'M', 'l'),
    (0x1D426, 'M', 'm'),
    (0x1D427, 'M', 'n'),
    (0x1D428, 'M', 'o'),
    (0x1D429, 'M', 'p'),
    (0x1D42A, 'M', 'q'),
    (0x1D42B, 'M', 'r'),
    (0x1D42C, 'M', 's'),
    (0x1D42D, 'M', 't'),
    (0x1D42E, 'M', 'u'),
    (0x1D42F, 'M', 'v'),
    (0x1D430, 'M', 'w'),
    (0x1D431, 'M', 'x'),
    (0x1D432, 'M', 'y'),
    (0x1D433, 'M', 'z'),
    (0x1D434, 'M', 'a'),
    (0x1D435, 'M', 'b'),
    (0x1D436, 'M', 'c'),
    (0x1D437, 'M', 'd'),
    (0x1D438, 'M', 'e'),
    (0x1D439, 'M', 'f'),
    (0x1D43A, 'M', 'g'),
    (0x1D43B, 'M', 'h'),
    (0x1D43C, 'M', 'i'),
    (0x1D43D, 'M', 'j'),
    (0x1D43E, 'M', 'k'),
    (0x1D43F, 'M', 'l'),
    (0x1D440, 'M', 'm'),
    (0x1D441, 'M', 'n'),
    (0x1D442, 'M', 'o'),
    (0x1D443, 'M', 'p'),
    (0x1D444, 'M', 'q'),
    (0x1D445, 'M', 'r'),
    (0x1D446, 'M', 's'),
    (0x1D447, 'M', 't'),
    (0x1D448, 'M', 'u'),
    (0x1D449, 'M', 'v'),
    (0x1D44A, 'M', 'w'),
    (0x1D44B, 'M', 'x'),
    (0x1D44C, 'M', 'y'),
    (0x1D44D, 'M', 'z'),
    (0x1D44E, 'M', 'a'),
    (0x1D44F, 'M', 'b'),
    (0x1D450, 'M', 'c'),
    (0x1D451, 'M', 'd'),
    ]

def _seg_60():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D452, 'M', 'e'),
    (0x1D453, 'M', 'f'),
    (0x1D454, 'M', 'g'),
    (0x1D455, 'X'),
    (0x1D456, 'M', 'i'),
    (0x1D457, 'M', 'j'),
    (0x1D458, 'M', 'k'),
    (0x1D459, 'M', 'l'),
    (0x1D45A, 'M', 'm'),
    (0x1D45B, 'M', 'n'),
    (0x1D45C, 'M', 'o'),
    (0x1D45D, 'M', 'p'),
    (0x1D45E, 'M', 'q'),
    (0x1D45F, 'M', 'r'),
    (0x1D460, 'M', 's'),
    (0x1D461, 'M', 't'),
    (0x1D462, 'M', 'u'),
    (0x1D463, 'M', 'v'),
    (0x1D464, 'M', 'w'),
    (0x1D465, 'M', 'x'),
    (0x1D466, 'M', 'y'),
    (0x1D467, 'M', 'z'),
    (0x1D468, 'M', 'a'),
    (0x1D469, 'M', 'b'),
    (0x1D46A, 'M', 'c'),
    (0x1D46B, 'M', 'd'),
    (0x1D46C, 'M', 'e'),
    (0x1D46D, 'M', 'f'),
    (0x1D46E, 'M', 'g'),
    (0x1D46F, 'M', 'h'),
    (0x1D470, 'M', 'i'),
    (0x1D471, 'M', 'j'),
    (0x1D472, 'M', 'k'),
    (0x1D473, 'M', 'l'),
    (0x1D474, 'M', 'm'),
    (0x1D475, 'M', 'n'),
    (0x1D476, 'M', 'o'),
    (0x1D477, 'M', 'p'),
    (0x1D478, 'M', 'q'),
    (0x1D479, 'M', 'r'),
    (0x1D47A, 'M', 's'),
    (0x1D47B, 'M', 't'),
    (0x1D47C, 'M', 'u'),
    (0x1D47D, 'M', 'v'),
    (0x1D47E, 'M', 'w'),
    (0x1D47F, 'M', 'x'),
    (0x1D480, 'M', 'y'),
    (0x1D481, 'M', 'z'),
    (0x1D482, 'M', 'a'),
    (0x1D483, 'M', 'b'),
    (0x1D484, 'M', 'c'),
    (0x1D485, 'M', 'd'),
    (0x1D486, 'M', 'e'),
    (0x1D487, 'M', 'f'),
    (0x1D488, 'M', 'g'),
    (0x1D489, 'M', 'h'),
    (0x1D48A, 'M', 'i'),
    (0x1D48B, 'M', 'j'),
    (0x1D48C, 'M', 'k'),
    (0x1D48D, 'M', 'l'),
    (0x1D48E, 'M', 'm'),
    (0x1D48F, 'M', 'n'),
    (0x1D490, 'M', 'o'),
    (0x1D491, 'M', 'p'),
    (0x1D492, 'M', 'q'),
    (0x1D493, 'M', 'r'),
    (0x1D494, 'M', 's'),
    (0x1D495, 'M', 't'),
    (0x1D496, 'M', 'u'),
    (0x1D497, 'M', 'v'),
    (0x1D498, 'M', 'w'),
    (0x1D499, 'M', 'x'),
    (0x1D49A, 'M', 'y'),
    (0x1D49B, 'M', 'z'),
    (0x1D49C, 'M', 'a'),
    (0x1D49D, 'X'),
    (0x1D49E, 'M', 'c'),
    (0x1D49F, 'M', 'd'),
    (0x1D4A0, 'X'),
    (0x1D4A2, 'M', 'g'),
    (0x1D4A3, 'X'),
    (0x1D4A5, 'M', 'j'),
    (0x1D4A6, 'M', 'k'),
    (0x1D4A7, 'X'),
    (0x1D4A9, 'M', 'n'),
    (0x1D4AA, 'M', 'o'),
    (0x1D4AB, 'M', 'p'),
    (0x1D4AC, 'M', 'q'),
    (0x1D4AD, 'X'),
    (0x1D4AE, 'M', 's'),
    (0x1D4AF, 'M', 't'),
    (0x1D4B0, 'M', 'u'),
    (0x1D4B1, 'M', 'v'),
    (0x1D4B2, 'M', 'w'),
    (0x1D4B3, 'M', 'x'),
    (0x1D4B4, 'M', 'y'),
    (0x1D4B5, 'M', 'z'),
    (0x1D4B6, 'M', 'a'),
    (0x1D4B7, 'M', 'b'),
    (0x1D4B8, 'M', 'c'),
    ]

def _seg_61():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D4B9, 'M', 'd'),
    (0x1D4BA, 'X'),
    (0x1D4BB, 'M', 'f'),
    (0x1D4BC, 'X'),
    (0x1D4BD, 'M', 'h'),
    (0x1D4BE, 'M', 'i'),
    (0x1D4BF, 'M', 'j'),
    (0x1D4C0, 'M', 'k'),
    (0x1D4C1, 'M', 'l'),
    (0x1D4C2, 'M', 'm'),
    (0x1D4C3, 'M', 'n'),
    (0x1D4C4, 'X'),
    (0x1D4C5, 'M', 'p'),
    (0x1D4C6, 'M', 'q'),
    (0x1D4C7, 'M', 'r'),
    (0x1D4C8, 'M', 's'),
    (0x1D4C9, 'M', 't'),
    (0x1D4CA, 'M', 'u'),
    (0x1D4CB, 'M', 'v'),
    (0x1D4CC, 'M', 'w'),
    (0x1D4CD, 'M', 'x'),
    (0x1D4CE, 'M', 'y'),
    (0x1D4CF, 'M', 'z'),
    (0x1D4D0, 'M', 'a'),
    (0x1D4D1, 'M', 'b'),
    (0x1D4D2, 'M', 'c'),
    (0x1D4D3, 'M', 'd'),
    (0x1D4D4, 'M', 'e'),
    (0x1D4D5, 'M', 'f'),
    (0x1D4D6, 'M', 'g'),
    (0x1D4D7, 'M', 'h'),
    (0x1D4D8, 'M', 'i'),
    (0x1D4D9, 'M', 'j'),
    (0x1D4DA, 'M', 'k'),
    (0x1D4DB, 'M', 'l'),
    (0x1D4DC, 'M', 'm'),
    (0x1D4DD, 'M', 'n'),
    (0x1D4DE, 'M', 'o'),
    (0x1D4DF, 'M', 'p'),
    (0x1D4E0, 'M', 'q'),
    (0x1D4E1, 'M', 'r'),
    (0x1D4E2, 'M', 's'),
    (0x1D4E3, 'M', 't'),
    (0x1D4E4, 'M', 'u'),
    (0x1D4E5, 'M', 'v'),
    (0x1D4E6, 'M', 'w'),
    (0x1D4E7, 'M', 'x'),
    (0x1D4E8, 'M', 'y'),
    (0x1D4E9, 'M', 'z'),
    (0x1D4EA, 'M', 'a'),
    (0x1D4EB, 'M', 'b'),
    (0x1D4EC, 'M', 'c'),
    (0x1D4ED, 'M', 'd'),
    (0x1D4EE, 'M', 'e'),
    (0x1D4EF, 'M', 'f'),
    (0x1D4F0, 'M', 'g'),
    (0x1D4F1, 'M', 'h'),
    (0x1D4F2, 'M', 'i'),
    (0x1D4F3, 'M', 'j'),
    (0x1D4F4, 'M', 'k'),
    (0x1D4F5, 'M', 'l'),
    (0x1D4F6, 'M', 'm'),
    (0x1D4F7, 'M', 'n'),
    (0x1D4F8, 'M', 'o'),
    (0x1D4F9, 'M', 'p'),
    (0x1D4FA, 'M', 'q'),
    (0x1D4FB, 'M', 'r'),
    (0x1D4FC, 'M', 's'),
    (0x1D4FD, 'M', 't'),
    (0x1D4FE, 'M', 'u'),
    (0x1D4FF, 'M', 'v'),
    (0x1D500, 'M', 'w'),
    (0x1D501, 'M', 'x'),
    (0x1D502, 'M', 'y'),
    (0x1D503, 'M', 'z'),
    (0x1D504, 'M', 'a'),
    (0x1D505, 'M', 'b'),
    (0x1D506, 'X'),
    (0x1D507, 'M', 'd'),
    (0x1D508, 'M', 'e'),
    (0x1D509, 'M', 'f'),
    (0x1D50A, 'M', 'g'),
    (0x1D50B, 'X'),
    (0x1D50D, 'M', 'j'),
    (0x1D50E, 'M', 'k'),
    (0x1D50F, 'M', 'l'),
    (0x1D510, 'M', 'm'),
    (0x1D511, 'M', 'n'),
    (0x1D512, 'M', 'o'),
    (0x1D513, 'M', 'p'),
    (0x1D514, 'M', 'q'),
    (0x1D515, 'X'),
    (0x1D516, 'M', 's'),
    (0x1D517, 'M', 't'),
    (0x1D518, 'M', 'u'),
    (0x1D519, 'M', 'v'),
    (0x1D51A, 'M', 'w'),
    (0x1D51B, 'M', 'x'),
    (0x1D51C, 'M', 'y'),
    (0x1D51D, 'X'),
    ]

def _seg_62():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D51E, 'M', 'a'),
    (0x1D51F, 'M', 'b'),
    (0x1D520, 'M', 'c'),
    (0x1D521, 'M', 'd'),
    (0x1D522, 'M', 'e'),
    (0x1D523, 'M', 'f'),
    (0x1D524, 'M', 'g'),
    (0x1D525, 'M', 'h'),
    (0x1D526, 'M', 'i'),
    (0x1D527, 'M', 'j'),
    (0x1D528, 'M', 'k'),
    (0x1D529, 'M', 'l'),
    (0x1D52A, 'M', 'm'),
    (0x1D52B, 'M', 'n'),
    (0x1D52C, 'M', 'o'),
    (0x1D52D, 'M', 'p'),
    (0x1D52E, 'M', 'q'),
    (0x1D52F, 'M', 'r'),
    (0x1D530, 'M', 's'),
    (0x1D531, 'M', 't'),
    (0x1D532, 'M', 'u'),
    (0x1D533, 'M', 'v'),
    (0x1D534, 'M', 'w'),
    (0x1D535, 'M', 'x'),
    (0x1D536, 'M', 'y'),
    (0x1D537, 'M', 'z'),
    (0x1D538, 'M', 'a'),
    (0x1D539, 'M', 'b'),
    (0x1D53A, 'X'),
    (0x1D53B, 'M', 'd'),
    (0x1D53C, 'M', 'e'),
    (0x1D53D, 'M', 'f'),
    (0x1D53E, 'M', 'g'),
    (0x1D53F, 'X'),
    (0x1D540, 'M', 'i'),
    (0x1D541, 'M', 'j'),
    (0x1D542, 'M', 'k'),
    (0x1D543, 'M', 'l'),
    (0x1D544, 'M', 'm'),
    (0x1D545, 'X'),
    (0x1D546, 'M', 'o'),
    (0x1D547, 'X'),
    (0x1D54A, 'M', 's'),
    (0x1D54B, 'M', 't'),
    (0x1D54C, 'M', 'u'),
    (0x1D54D, 'M', 'v'),
    (0x1D54E, 'M', 'w'),
    (0x1D54F, 'M', 'x'),
    (0x1D550, 'M', 'y'),
    (0x1D551, 'X'),
    (0x1D552, 'M', 'a'),
    (0x1D553, 'M', 'b'),
    (0x1D554, 'M', 'c'),
    (0x1D555, 'M', 'd'),
    (0x1D556, 'M', 'e'),
    (0x1D557, 'M', 'f'),
    (0x1D558, 'M', 'g'),
    (0x1D559, 'M', 'h'),
    (0x1D55A, 'M', 'i'),
    (0x1D55B, 'M', 'j'),
    (0x1D55C, 'M', 'k'),
    (0x1D55D, 'M', 'l'),
    (0x1D55E, 'M', 'm'),
    (0x1D55F, 'M', 'n'),
    (0x1D560, 'M', 'o'),
    (0x1D561, 'M', 'p'),
    (0x1D562, 'M', 'q'),
    (0x1D563, 'M', 'r'),
    (0x1D564, 'M', 's'),
    (0x1D565, 'M', 't'),
    (0x1D566, 'M', 'u'),
    (0x1D567, 'M', 'v'),
    (0x1D568, 'M', 'w'),
    (0x1D569, 'M', 'x'),
    (0x1D56A, 'M', 'y'),
    (0x1D56B, 'M', 'z'),
    (0x1D56C, 'M', 'a'),
    (0x1D56D, 'M', 'b'),
    (0x1D56E, 'M', 'c'),
    (0x1D56F, 'M', 'd'),
    (0x1D570, 'M', 'e'),
    (0x1D571, 'M', 'f'),
    (0x1D572, 'M', 'g'),
    (0x1D573, 'M', 'h'),
    (0x1D574, 'M', 'i'),
    (0x1D575, 'M', 'j'),
    (0x1D576, 'M', 'k'),
    (0x1D577, 'M', 'l'),
    (0x1D578, 'M', 'm'),
    (0x1D579, 'M', 'n'),
    (0x1D57A, 'M', 'o'),
    (0x1D57B, 'M', 'p'),
    (0x1D57C, 'M', 'q'),
    (0x1D57D, 'M', 'r'),
    (0x1D57E, 'M', 's'),
    (0x1D57F, 'M', 't'),
    (0x1D580, 'M', 'u'),
    (0x1D581, 'M', 'v'),
    (0x1D582, 'M', 'w'),
    (0x1D583, 'M', 'x'),
    ]

def _seg_63():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D584, 'M', 'y'),
    (0x1D585, 'M', 'z'),
    (0x1D586, 'M', 'a'),
    (0x1D587, 'M', 'b'),
    (0x1D588, 'M', 'c'),
    (0x1D589, 'M', 'd'),
    (0x1D58A, 'M', 'e'),
    (0x1D58B, 'M', 'f'),
    (0x1D58C, 'M', 'g'),
    (0x1D58D, 'M', 'h'),
    (0x1D58E, 'M', 'i'),
    (0x1D58F, 'M', 'j'),
    (0x1D590, 'M', 'k'),
    (0x1D591, 'M', 'l'),
    (0x1D592, 'M', 'm'),
    (0x1D593, 'M', 'n'),
    (0x1D594, 'M', 'o'),
    (0x1D595, 'M', 'p'),
    (0x1D596, 'M', 'q'),
    (0x1D597, 'M', 'r'),
    (0x1D598, 'M', 's'),
    (0x1D599, 'M', 't'),
    (0x1D59A, 'M', 'u'),
    (0x1D59B, 'M', 'v'),
    (0x1D59C, 'M', 'w'),
    (0x1D59D, 'M', 'x'),
    (0x1D59E, 'M', 'y'),
    (0x1D59F, 'M', 'z'),
    (0x1D5A0, 'M', 'a'),
    (0x1D5A1, 'M', 'b'),
    (0x1D5A2, 'M', 'c'),
    (0x1D5A3, 'M', 'd'),
    (0x1D5A4, 'M', 'e'),
    (0x1D5A5, 'M', 'f'),
    (0x1D5A6, 'M', 'g'),
    (0x1D5A7, 'M', 'h'),
    (0x1D5A8, 'M', 'i'),
    (0x1D5A9, 'M', 'j'),
    (0x1D5AA, 'M', 'k'),
    (0x1D5AB, 'M', 'l'),
    (0x1D5AC, 'M', 'm'),
    (0x1D5AD, 'M', 'n'),
    (0x1D5AE, 'M', 'o'),
    (0x1D5AF, 'M', 'p'),
    (0x1D5B0, 'M', 'q'),
    (0x1D5B1, 'M', 'r'),
    (0x1D5B2, 'M', 's'),
    (0x1D5B3, 'M', 't'),
    (0x1D5B4, 'M', 'u'),
    (0x1D5B5, 'M', 'v'),
    (0x1D5B6, 'M', 'w'),
    (0x1D5B7, 'M', 'x'),
    (0x1D5B8, 'M', 'y'),
    (0x1D5B9, 'M', 'z'),
    (0x1D5BA, 'M', 'a'),
    (0x1D5BB, 'M', 'b'),
    (0x1D5BC, 'M', 'c'),
    (0x1D5BD, 'M', 'd'),
    (0x1D5BE, 'M', 'e'),
    (0x1D5BF, 'M', 'f'),
    (0x1D5C0, 'M', 'g'),
    (0x1D5C1, 'M', 'h'),
    (0x1D5C2, 'M', 'i'),
    (0x1D5C3, 'M', 'j'),
    (0x1D5C4, 'M', 'k'),
    (0x1D5C5, 'M', 'l'),
    (0x1D5C6, 'M', 'm'),
    (0x1D5C7, 'M', 'n'),
    (0x1D5C8, 'M', 'o'),
    (0x1D5C9, 'M', 'p'),
    (0x1D5CA, 'M', 'q'),
    (0x1D5CB, 'M', 'r'),
    (0x1D5CC, 'M', 's'),
    (0x1D5CD, 'M', 't'),
    (0x1D5CE, 'M', 'u'),
    (0x1D5CF, 'M', 'v'),
    (0x1D5D0, 'M', 'w'),
    (0x1D5D1, 'M', 'x'),
    (0x1D5D2, 'M', 'y'),
    (0x1D5D3, 'M', 'z'),
    (0x1D5D4, 'M', 'a'),
    (0x1D5D5, 'M', 'b'),
    (0x1D5D6, 'M', 'c'),
    (0x1D5D7, 'M', 'd'),
    (0x1D5D8, 'M', 'e'),
    (0x1D5D9, 'M', 'f'),
    (0x1D5DA, 'M', 'g'),
    (0x1D5DB, 'M', 'h'),
    (0x1D5DC, 'M', 'i'),
    (0x1D5DD, 'M', 'j'),
    (0x1D5DE, 'M', 'k'),
    (0x1D5DF, 'M', 'l'),
    (0x1D5E0, 'M', 'm'),
    (0x1D5E1, 'M', 'n'),
    (0x1D5E2, 'M', 'o'),
    (0x1D5E3, 'M', 'p'),
    (0x1D5E4, 'M', 'q'),
    (0x1D5E5, 'M', 'r'),
    (0x1D5E6, 'M', 's'),
    (0x1D5E7, 'M', 't'),
    ]

def _seg_64():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D5E8, 'M', 'u'),
    (0x1D5E9, 'M', 'v'),
    (0x1D5EA, 'M', 'w'),
    (0x1D5EB, 'M', 'x'),
    (0x1D5EC, 'M', 'y'),
    (0x1D5ED, 'M', 'z'),
    (0x1D5EE, 'M', 'a'),
    (0x1D5EF, 'M', 'b'),
    (0x1D5F0, 'M', 'c'),
    (0x1D5F1, 'M', 'd'),
    (0x1D5F2, 'M', 'e'),
    (0x1D5F3, 'M', 'f'),
    (0x1D5F4, 'M', 'g'),
    (0x1D5F5, 'M', 'h'),
    (0x1D5F6, 'M', 'i'),
    (0x1D5F7, 'M', 'j'),
    (0x1D5F8, 'M', 'k'),
    (0x1D5F9, 'M', 'l'),
    (0x1D5FA, 'M', 'm'),
    (0x1D5FB, 'M', 'n'),
    (0x1D5FC, 'M', 'o'),
    (0x1D5FD, 'M', 'p'),
    (0x1D5FE, 'M', 'q'),
    (0x1D5FF, 'M', 'r'),
    (0x1D600, 'M', 's'),
    (0x1D601, 'M', 't'),
    (0x1D602, 'M', 'u'),
    (0x1D603, 'M', 'v'),
    (0x1D604, 'M', 'w'),
    (0x1D605, 'M', 'x'),
    (0x1D606, 'M', 'y'),
    (0x1D607, 'M', 'z'),
    (0x1D608, 'M', 'a'),
    (0x1D609, 'M', 'b'),
    (0x1D60A, 'M', 'c'),
    (0x1D60B, 'M', 'd'),
    (0x1D60C, 'M', 'e'),
    (0x1D60D, 'M', 'f'),
    (0x1D60E, 'M', 'g'),
    (0x1D60F, 'M', 'h'),
    (0x1D610, 'M', 'i'),
    (0x1D611, 'M', 'j'),
    (0x1D612, 'M', 'k'),
    (0x1D613, 'M', 'l'),
    (0x1D614, 'M', 'm'),
    (0x1D615, 'M', 'n'),
    (0x1D616, 'M', 'o'),
    (0x1D617, 'M', 'p'),
    (0x1D618, 'M', 'q'),
    (0x1D619, 'M', 'r'),
    (0x1D61A, 'M', 's'),
    (0x1D61B, 'M', 't'),
    (0x1D61C, 'M', 'u'),
    (0x1D61D, 'M', 'v'),
    (0x1D61E, 'M', 'w'),
    (0x1D61F, 'M', 'x'),
    (0x1D620, 'M', 'y'),
    (0x1D621, 'M', 'z'),
    (0x1D622, 'M', 'a'),
    (0x1D623, 'M', 'b'),
    (0x1D624, 'M', 'c'),
    (0x1D625, 'M', 'd'),
    (0x1D626, 'M', 'e'),
    (0x1D627, 'M', 'f'),
    (0x1D628, 'M', 'g'),
    (0x1D629, 'M', 'h'),
    (0x1D62A, 'M', 'i'),
    (0x1D62B, 'M', 'j'),
    (0x1D62C, 'M', 'k'),
    (0x1D62D, 'M', 'l'),
    (0x1D62E, 'M', 'm'),
    (0x1D62F, 'M', 'n'),
    (0x1D630, 'M', 'o'),
    (0x1D631, 'M', 'p'),
    (0x1D632, 'M', 'q'),
    (0x1D633, 'M', 'r'),
    (0x1D634, 'M', 's'),
    (0x1D635, 'M', 't'),
    (0x1D636, 'M', 'u'),
    (0x1D637, 'M', 'v'),
    (0x1D638, 'M', 'w'),
    (0x1D639, 'M', 'x'),
    (0x1D63A, 'M', 'y'),
    (0x1D63B, 'M', 'z'),
    (0x1D63C, 'M', 'a'),
    (0x1D63D, 'M', 'b'),
    (0x1D63E, 'M', 'c'),
    (0x1D63F, 'M', 'd'),
    (0x1D640, 'M', 'e'),
    (0x1D641, 'M', 'f'),
    (0x1D642, 'M', 'g'),
    (0x1D643, 'M', 'h'),
    (0x1D644, 'M', 'i'),
    (0x1D645, 'M', 'j'),
    (0x1D646, 'M', 'k'),
    (0x1D647, 'M', 'l'),
    (0x1D648, 'M', 'm'),
    (0x1D649, 'M', 'n'),
    (0x1D64A, 'M', 'o'),
    (0x1D64B, 'M', 'p'),
    ]

def _seg_65():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D64C, 'M', 'q'),
    (0x1D64D, 'M', 'r'),
    (0x1D64E, 'M', 's'),
    (0x1D64F, 'M', 't'),
    (0x1D650, 'M', 'u'),
    (0x1D651, 'M', 'v'),
    (0x1D652, 'M', 'w'),
    (0x1D653, 'M', 'x'),
    (0x1D654, 'M', 'y'),
    (0x1D655, 'M', 'z'),
    (0x1D656, 'M', 'a'),
    (0x1D657, 'M', 'b'),
    (0x1D658, 'M', 'c'),
    (0x1D659, 'M', 'd'),
    (0x1D65A, 'M', 'e'),
    (0x1D65B, 'M', 'f'),
    (0x1D65C, 'M', 'g'),
    (0x1D65D, 'M', 'h'),
    (0x1D65E, 'M', 'i'),
    (0x1D65F, 'M', 'j'),
    (0x1D660, 'M', 'k'),
    (0x1D661, 'M', 'l'),
    (0x1D662, 'M', 'm'),
    (0x1D663, 'M', 'n'),
    (0x1D664, 'M', 'o'),
    (0x1D665, 'M', 'p'),
    (0x1D666, 'M', 'q'),
    (0x1D667, 'M', 'r'),
    (0x1D668, 'M', 's'),
    (0x1D669, 'M', 't'),
    (0x1D66A, 'M', 'u'),
    (0x1D66B, 'M', 'v'),
    (0x1D66C, 'M', 'w'),
    (0x1D66D, 'M', 'x'),
    (0x1D66E, 'M', 'y'),
    (0x1D66F, 'M', 'z'),
    (0x1D670, 'M', 'a'),
    (0x1D671, 'M', 'b'),
    (0x1D672, 'M', 'c'),
    (0x1D673, 'M', 'd'),
    (0x1D674, 'M', 'e'),
    (0x1D675, 'M', 'f'),
    (0x1D676, 'M', 'g'),
    (0x1D677, 'M', 'h'),
    (0x1D678, 'M', 'i'),
    (0x1D679, 'M', 'j'),
    (0x1D67A, 'M', 'k'),
    (0x1D67B, 'M', 'l'),
    (0x1D67C, 'M', 'm'),
    (0x1D67D, 'M', 'n'),
    (0x1D67E, 'M', 'o'),
    (0x1D67F, 'M', 'p'),
    (0x1D680, 'M', 'q'),
    (0x1D681, 'M', 'r'),
    (0x1D682, 'M', 's'),
    (0x1D683, 'M', 't'),
    (0x1D684, 'M', 'u'),
    (0x1D685, 'M', 'v'),
    (0x1D686, 'M', 'w'),
    (0x1D687, 'M', 'x'),
    (0x1D688, 'M', 'y'),
    (0x1D689, 'M', 'z'),
    (0x1D68A, 'M', 'a'),
    (0x1D68B, 'M', 'b'),
    (0x1D68C, 'M', 'c'),
    (0x1D68D, 'M', 'd'),
    (0x1D68E, 'M', 'e'),
    (0x1D68F, 'M', 'f'),
    (0x1D690, 'M', 'g'),
    (0x1D691, 'M', 'h'),
    (0x1D692, 'M', 'i'),
    (0x1D693, 'M', 'j'),
    (0x1D694, 'M', 'k'),
    (0x1D695, 'M', 'l'),
    (0x1D696, 'M', 'm'),
    (0x1D697, 'M', 'n'),
    (0x1D698, 'M', 'o'),
    (0x1D699, 'M', 'p'),
    (0x1D69A, 'M', 'q'),
    (0x1D69B, 'M', 'r'),
    (0x1D69C, 'M', 's'),
    (0x1D69D, 'M', 't'),
    (0x1D69E, 'M', 'u'),
    (0x1D69F, 'M', 'v'),
    (0x1D6A0, 'M', 'w'),
    (0x1D6A1, 'M', 'x'),
    (0x1D6A2, 'M', 'y'),
    (0x1D6A3, 'M', 'z'),
    (0x1D6A4, 'M', 'ı'),
    (0x1D6A5, 'M', 'ȷ'),
    (0x1D6A6, 'X'),
    (0x1D6A8, 'M', 'α'),
    (0x1D6A9, 'M', 'β'),
    (0x1D6AA, 'M', 'γ'),
    (0x1D6AB, 'M', 'δ'),
    (0x1D6AC, 'M', 'ε'),
    (0x1D6AD, 'M', 'ζ'),
    (0x1D6AE, 'M', 'η'),
    (0x1D6AF, 'M', 'θ'),
    (0x1D6B0, 'M', 'ι'),
    ]

def _seg_66():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D6B1, 'M', 'κ'),
    (0x1D6B2, 'M', 'λ'),
    (0x1D6B3, 'M', 'μ'),
    (0x1D6B4, 'M', 'ν'),
    (0x1D6B5, 'M', 'ξ'),
    (0x1D6B6, 'M', 'ο'),
    (0x1D6B7, 'M', 'π'),
    (0x1D6B8, 'M', 'ρ'),
    (0x1D6B9, 'M', 'θ'),
    (0x1D6BA, 'M', 'σ'),
    (0x1D6BB, 'M', 'τ'),
    (0x1D6BC, 'M', 'υ'),
    (0x1D6BD, 'M', 'φ'),
    (0x1D6BE, 'M', 'χ'),
    (0x1D6BF, 'M', 'ψ'),
    (0x1D6C0, 'M', 'ω'),
    (0x1D6C1, 'M', '∇'),
    (0x1D6C2, 'M', 'α'),
    (0x1D6C3, 'M', 'β'),
    (0x1D6C4, 'M', 'γ'),
    (0x1D6C5, 'M', 'δ'),
    (0x1D6C6, 'M', 'ε'),
    (0x1D6C7, 'M', 'ζ'),
    (0x1D6C8, 'M', 'η'),
    (0x1D6C9, 'M', 'θ'),
    (0x1D6CA, 'M', 'ι'),
    (0x1D6CB, 'M', 'κ'),
    (0x1D6CC, 'M', 'λ'),
    (0x1D6CD, 'M', 'μ'),
    (0x1D6CE, 'M', 'ν'),
    (0x1D6CF, 'M', 'ξ'),
    (0x1D6D0, 'M', 'ο'),
    (0x1D6D1, 'M', 'π'),
    (0x1D6D2, 'M', 'ρ'),
    (0x1D6D3, 'M', 'σ'),
    (0x1D6D5, 'M', 'τ'),
    (0x1D6D6, 'M', 'υ'),
    (0x1D6D7, 'M', 'φ'),
    (0x1D6D8, 'M', 'χ'),
    (0x1D6D9, 'M', 'ψ'),
    (0x1D6DA, 'M', 'ω'),
    (0x1D6DB, 'M', '∂'),
    (0x1D6DC, 'M', 'ε'),
    (0x1D6DD, 'M', 'θ'),
    (0x1D6DE, 'M', 'κ'),
    (0x1D6DF, 'M', 'φ'),
    (0x1D6E0, 'M', 'ρ'),
    (0x1D6E1, 'M', 'π'),
    (0x1D6E2, 'M', 'α'),
    (0x1D6E3, 'M', 'β'),
    (0x1D6E4, 'M', 'γ'),
    (0x1D6E5, 'M', 'δ'),
    (0x1D6E6, 'M', 'ε'),
    (0x1D6E7, 'M', 'ζ'),
    (0x1D6E8, 'M', 'η'),
    (0x1D6E9, 'M', 'θ'),
    (0x1D6EA, 'M', 'ι'),
    (0x1D6EB, 'M', 'κ'),
    (0x1D6EC, 'M', 'λ'),
    (0x1D6ED, 'M', 'μ'),
    (0x1D6EE, 'M', 'ν'),
    (0x1D6EF, 'M', 'ξ'),
    (0x1D6F0, 'M', 'ο'),
    (0x1D6F1, 'M', 'π'),
    (0x1D6F2, 'M', 'ρ'),
    (0x1D6F3, 'M', 'θ'),
    (0x1D6F4, 'M', 'σ'),
    (0x1D6F5, 'M', 'τ'),
    (0x1D6F6, 'M', 'υ'),
    (0x1D6F7, 'M', 'φ'),
    (0x1D6F8, 'M', 'χ'),
    (0x1D6F9, 'M', 'ψ'),
    (0x1D6FA, 'M', 'ω'),
    (0x1D6FB, 'M', '∇'),
    (0x1D6FC, 'M', 'α'),
    (0x1D6FD, 'M', 'β'),
    (0x1D6FE, 'M', 'γ'),
    (0x1D6FF, 'M', 'δ'),
    (0x1D700, 'M', 'ε'),
    (0x1D701, 'M', 'ζ'),
    (0x1D702, 'M', 'η'),
    (0x1D703, 'M', 'θ'),
    (0x1D704, 'M', 'ι'),
    (0x1D705, 'M', 'κ'),
    (0x1D706, 'M', 'λ'),
    (0x1D707, 'M', 'μ'),
    (0x1D708, 'M', 'ν'),
    (0x1D709, 'M', 'ξ'),
    (0x1D70A, 'M', 'ο'),
    (0x1D70B, 'M', 'π'),
    (0x1D70C, 'M', 'ρ'),
    (0x1D70D, 'M', 'σ'),
    (0x1D70F, 'M', 'τ'),
    (0x1D710, 'M', 'υ'),
    (0x1D711, 'M', 'φ'),
    (0x1D712, 'M', 'χ'),
    (0x1D713, 'M', 'ψ'),
    (0x1D714, 'M', 'ω'),
    (0x1D715, 'M', '∂'),
    (0x1D716, 'M', 'ε'),
    ]

def _seg_67():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D717, 'M', 'θ'),
    (0x1D718, 'M', 'κ'),
    (0x1D719, 'M', 'φ'),
    (0x1D71A, 'M', 'ρ'),
    (0x1D71B, 'M', 'π'),
    (0x1D71C, 'M', 'α'),
    (0x1D71D, 'M', 'β'),
    (0x1D71E, 'M', 'γ'),
    (0x1D71F, 'M', 'δ'),
    (0x1D720, 'M', 'ε'),
    (0x1D721, 'M', 'ζ'),
    (0x1D722, 'M', 'η'),
    (0x1D723, 'M', 'θ'),
    (0x1D724, 'M', 'ι'),
    (0x1D725, 'M', 'κ'),
    (0x1D726, 'M', 'λ'),
    (0x1D727, 'M', 'μ'),
    (0x1D728, 'M', 'ν'),
    (0x1D729, 'M', 'ξ'),
    (0x1D72A, 'M', 'ο'),
    (0x1D72B, 'M', 'π'),
    (0x1D72C, 'M', 'ρ'),
    (0x1D72D, 'M', 'θ'),
    (0x1D72E, 'M', 'σ'),
    (0x1D72F, 'M', 'τ'),
    (0x1D730, 'M', 'υ'),
    (0x1D731, 'M', 'φ'),
    (0x1D732, 'M', 'χ'),
    (0x1D733, 'M', 'ψ'),
    (0x1D734, 'M', 'ω'),
    (0x1D735, 'M', '∇'),
    (0x1D736, 'M', 'α'),
    (0x1D737, 'M', 'β'),
    (0x1D738, 'M', 'γ'),
    (0x1D739, 'M', 'δ'),
    (0x1D73A, 'M', 'ε'),
    (0x1D73B, 'M', 'ζ'),
    (0x1D73C, 'M', 'η'),
    (0x1D73D, 'M', 'θ'),
    (0x1D73E, 'M', 'ι'),
    (0x1D73F, 'M', 'κ'),
    (0x1D740, 'M', 'λ'),
    (0x1D741, 'M', 'μ'),
    (0x1D742, 'M', 'ν'),
    (0x1D743, 'M', 'ξ'),
    (0x1D744, 'M', 'ο'),
    (0x1D745, 'M', 'π'),
    (0x1D746, 'M', 'ρ'),
    (0x1D747, 'M', 'σ'),
    (0x1D749, 'M', 'τ'),
    (0x1D74A, 'M', 'υ'),
    (0x1D74B, 'M', 'φ'),
    (0x1D74C, 'M', 'χ'),
    (0x1D74D, 'M', 'ψ'),
    (0x1D74E, 'M', 'ω'),
    (0x1D74F, 'M', '∂'),
    (0x1D750, 'M', 'ε'),
    (0x1D751, 'M', 'θ'),
    (0x1D752, 'M', 'κ'),
    (0x1D753, 'M', 'φ'),
    (0x1D754, 'M', 'ρ'),
    (0x1D755, 'M', 'π'),
    (0x1D756, 'M', 'α'),
    (0x1D757, 'M', 'β'),
    (0x1D758, 'M', 'γ'),
    (0x1D759, 'M', 'δ'),
    (0x1D75A, 'M', 'ε'),
    (0x1D75B, 'M', 'ζ'),
    (0x1D75C, 'M', 'η'),
    (0x1D75D, 'M', 'θ'),
    (0x1D75E, 'M', 'ι'),
    (0x1D75F, 'M', 'κ'),
    (0x1D760, 'M', 'λ'),
    (0x1D761, 'M', 'μ'),
    (0x1D762, 'M', 'ν'),
    (0x1D763, 'M', 'ξ'),
    (0x1D764, 'M', 'ο'),
    (0x1D765, 'M', 'π'),
    (0x1D766, 'M', 'ρ'),
    (0x1D767, 'M', 'θ'),
    (0x1D768, 'M', 'σ'),
    (0x1D769, 'M', 'τ'),
    (0x1D76A, 'M', 'υ'),
    (0x1D76B, 'M', 'φ'),
    (0x1D76C, 'M', 'χ'),
    (0x1D76D, 'M', 'ψ'),
    (0x1D76E, 'M', 'ω'),
    (0x1D76F, 'M', '∇'),
    (0x1D770, 'M', 'α'),
    (0x1D771, 'M', 'β'),
    (0x1D772, 'M', 'γ'),
    (0x1D773, 'M', 'δ'),
    (0x1D774, 'M', 'ε'),
    (0x1D775, 'M', 'ζ'),
    (0x1D776, 'M', 'η'),
    (0x1D777, 'M', 'θ'),
    (0x1D778, 'M', 'ι'),
    (0x1D779, 'M', 'κ'),
    (0x1D77A, 'M', 'λ'),
    (0x1D77B, 'M', 'μ'),
    ]

def _seg_68():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D77C, 'M', 'ν'),
    (0x1D77D, 'M', 'ξ'),
    (0x1D77E, 'M', 'ο'),
    (0x1D77F, 'M', 'π'),
    (0x1D780, 'M', 'ρ'),
    (0x1D781, 'M', 'σ'),
    (0x1D783, 'M', 'τ'),
    (0x1D784, 'M', 'υ'),
    (0x1D785, 'M', 'φ'),
    (0x1D786, 'M', 'χ'),
    (0x1D787, 'M', 'ψ'),
    (0x1D788, 'M', 'ω'),
    (0x1D789, 'M', '∂'),
    (0x1D78A, 'M', 'ε'),
    (0x1D78B, 'M', 'θ'),
    (0x1D78C, 'M', 'κ'),
    (0x1D78D, 'M', 'φ'),
    (0x1D78E, 'M', 'ρ'),
    (0x1D78F, 'M', 'π'),
    (0x1D790, 'M', 'α'),
    (0x1D791, 'M', 'β'),
    (0x1D792, 'M', 'γ'),
    (0x1D793, 'M', 'δ'),
    (0x1D794, 'M', 'ε'),
    (0x1D795, 'M', 'ζ'),
    (0x1D796, 'M', 'η'),
    (0x1D797, 'M', 'θ'),
    (0x1D798, 'M', 'ι'),
    (0x1D799, 'M', 'κ'),
    (0x1D79A, 'M', 'λ'),
    (0x1D79B, 'M', 'μ'),
    (0x1D79C, 'M', 'ν'),
    (0x1D79D, 'M', 'ξ'),
    (0x1D79E, 'M', 'ο'),
    (0x1D79F, 'M', 'π'),
    (0x1D7A0, 'M', 'ρ'),
    (0x1D7A1, 'M', 'θ'),
    (0x1D7A2, 'M', 'σ'),
    (0x1D7A3, 'M', 'τ'),
    (0x1D7A4, 'M', 'υ'),
    (0x1D7A5, 'M', 'φ'),
    (0x1D7A6, 'M', 'χ'),
    (0x1D7A7, 'M', 'ψ'),
    (0x1D7A8, 'M', 'ω'),
    (0x1D7A9, 'M', '∇'),
    (0x1D7AA, 'M', 'α'),
    (0x1D7AB, 'M', 'β'),
    (0x1D7AC, 'M', 'γ'),
    (0x1D7AD, 'M', 'δ'),
    (0x1D7AE, 'M', 'ε'),
    (0x1D7AF, 'M', 'ζ'),
    (0x1D7B0, 'M', 'η'),
    (0x1D7B1, 'M', 'θ'),
    (0x1D7B2, 'M', 'ι'),
    (0x1D7B3, 'M', 'κ'),
    (0x1D7B4, 'M', 'λ'),
    (0x1D7B5, 'M', 'μ'),
    (0x1D7B6, 'M', 'ν'),
    (0x1D7B7, 'M', 'ξ'),
    (0x1D7B8, 'M', 'ο'),
    (0x1D7B9, 'M', 'π'),
    (0x1D7BA, 'M', 'ρ'),
    (0x1D7BB, 'M', 'σ'),
    (0x1D7BD, 'M', 'τ'),
    (0x1D7BE, 'M', 'υ'),
    (0x1D7BF, 'M', 'φ'),
    (0x1D7C0, 'M', 'χ'),
    (0x1D7C1, 'M', 'ψ'),
    (0x1D7C2, 'M', 'ω'),
    (0x1D7C3, 'M', '∂'),
    (0x1D7C4, 'M', 'ε'),
    (0x1D7C5, 'M', 'θ'),
    (0x1D7C6, 'M', 'κ'),
    (0x1D7C7, 'M', 'φ'),
    (0x1D7C8, 'M', 'ρ'),
    (0x1D7C9, 'M', 'π'),
    (0x1D7CA, 'M', 'ϝ'),
    (0x1D7CC, 'X'),
    (0x1D7CE, 'M', '0'),
    (0x1D7CF, 'M', '1'),
    (0x1D7D0, 'M', '2'),
    (0x1D7D1, 'M', '3'),
    (0x1D7D2, 'M', '4'),
    (0x1D7D3, 'M', '5'),
    (0x1D7D4, 'M', '6'),
    (0x1D7D5, 'M', '7'),
    (0x1D7D6, 'M', '8'),
    (0x1D7D7, 'M', '9'),
    (0x1D7D8, 'M', '0'),
    (0x1D7D9, 'M', '1'),
    (0x1D7DA, 'M', '2'),
    (0x1D7DB, 'M', '3'),
    (0x1D7DC, 'M', '4'),
    (0x1D7DD, 'M', '5'),
    (0x1D7DE, 'M', '6'),
    (0x1D7DF, 'M', '7'),
    (0x1D7E0, 'M', '8'),
    (0x1D7E1, 'M', '9'),
    (0x1D7E2, 'M', '0'),
    (0x1D7E3, 'M', '1'),
    ]

def _seg_69():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1D7E4, 'M', '2'),
    (0x1D7E5, 'M', '3'),
    (0x1D7E6, 'M', '4'),
    (0x1D7E7, 'M', '5'),
    (0x1D7E8, 'M', '6'),
    (0x1D7E9, 'M', '7'),
    (0x1D7EA, 'M', '8'),
    (0x1D7EB, 'M', '9'),
    (0x1D7EC, 'M', '0'),
    (0x1D7ED, 'M', '1'),
    (0x1D7EE, 'M', '2'),
    (0x1D7EF, 'M', '3'),
    (0x1D7F0, 'M', '4'),
    (0x1D7F1, 'M', '5'),
    (0x1D7F2, 'M', '6'),
    (0x1D7F3, 'M', '7'),
    (0x1D7F4, 'M', '8'),
    (0x1D7F5, 'M', '9'),
    (0x1D7F6, 'M', '0'),
    (0x1D7F7, 'M', '1'),
    (0x1D7F8, 'M', '2'),
    (0x1D7F9, 'M', '3'),
    (0x1D7FA, 'M', '4'),
    (0x1D7FB, 'M', '5'),
    (0x1D7FC, 'M', '6'),
    (0x1D7FD, 'M', '7'),
    (0x1D7FE, 'M', '8'),
    (0x1D7FF, 'M', '9'),
    (0x1D800, 'V'),
    (0x1DA8C, 'X'),
    (0x1DA9B, 'V'),
    (0x1DAA0, 'X'),
    (0x1DAA1, 'V'),
    (0x1DAB0, 'X'),
    (0x1E000, 'V'),
    (0x1E007, 'X'),
    (0x1E008, 'V'),
    (0x1E019, 'X'),
    (0x1E01B, 'V'),
    (0x1E022, 'X'),
    (0x1E023, 'V'),
    (0x1E025, 'X'),
    (0x1E026, 'V'),
    (0x1E02B, 'X'),
    (0x1E100, 'V'),
    (0x1E12D, 'X'),
    (0x1E130, 'V'),
    (0x1E13E, 'X'),
    (0x1E140, 'V'),
    (0x1E14A, 'X'),
    (0x1E14E, 'V'),
    (0x1E150, 'X'),
    (0x1E2C0, 'V'),
    (0x1E2FA, 'X'),
    (0x1E2FF, 'V'),
    (0x1E300, 'X'),
    (0x1E800, 'V'),
    (0x1E8C5, 'X'),
    (0x1E8C7, 'V'),
    (0x1E8D7, 'X'),
    (0x1E900, 'M', '𞤢'),
    (0x1E901, 'M', '𞤣'),
    (0x1E902, 'M', '𞤤'),
    (0x1E903, 'M', '𞤥'),
    (0x1E904, 'M', '𞤦'),
    (0x1E905, 'M', '𞤧'),
    (0x1E906, 'M', '𞤨'),
    (0x1E907, 'M', '𞤩'),
    (0x1E908, 'M', '𞤪'),
    (0x1E909, 'M', '𞤫'),
    (0x1E90A, 'M', '𞤬'),
    (0x1E90B, 'M', '𞤭'),
    (0x1E90C, 'M', '𞤮'),
    (0x1E90D, 'M', '𞤯'),
    (0x1E90E, 'M', '𞤰'),
    (0x1E90F, 'M', '𞤱'),
    (0x1E910, 'M', '𞤲'),
    (0x1E911, 'M', '𞤳'),
    (0x1E912, 'M', '𞤴'),
    (0x1E913, 'M', '𞤵'),
    (0x1E914, 'M', '𞤶'),
    (0x1E915, 'M', '𞤷'),
    (0x1E916, 'M', '𞤸'),
    (0x1E917, 'M', '𞤹'),
    (0x1E918, 'M', '𞤺'),
    (0x1E919, 'M', '𞤻'),
    (0x1E91A, 'M', '𞤼'),
    (0x1E91B, 'M', '𞤽'),
    (0x1E91C, 'M', '𞤾'),
    (0x1E91D, 'M', '𞤿'),
    (0x1E91E, 'M', '𞥀'),
    (0x1E91F, 'M', '𞥁'),
    (0x1E920, 'M', '𞥂'),
    (0x1E921, 'M', '𞥃'),
    (0x1E922, 'V'),
    (0x1E94C, 'X'),
    (0x1E950, 'V'),
    (0x1E95A, 'X'),
    (0x1E95E, 'V'),
    (0x1E960, 'X'),
    ]

def _seg_70():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1EC71, 'V'),
    (0x1ECB5, 'X'),
    (0x1ED01, 'V'),
    (0x1ED3E, 'X'),
    (0x1EE00, 'M', 'ا'),
    (0x1EE01, 'M', 'ب'),
    (0x1EE02, 'M', 'ج'),
    (0x1EE03, 'M', 'د'),
    (0x1EE04, 'X'),
    (0x1EE05, 'M', 'و'),
    (0x1EE06, 'M', 'ز'),
    (0x1EE07, 'M', 'ح'),
    (0x1EE08, 'M', 'ط'),
    (0x1EE09, 'M', 'ي'),
    (0x1EE0A, 'M', 'ك'),
    (0x1EE0B, 'M', 'ل'),
    (0x1EE0C, 'M', 'م'),
    (0x1EE0D, 'M', 'ن'),
    (0x1EE0E, 'M', 'س'),
    (0x1EE0F, 'M', 'ع'),
    (0x1EE10, 'M', 'ف'),
    (0x1EE11, 'M', 'ص'),
    (0x1EE12, 'M', 'ق'),
    (0x1EE13, 'M', 'ر'),
    (0x1EE14, 'M', 'ش'),
    (0x1EE15, 'M', 'ت'),
    (0x1EE16, 'M', 'ث'),
    (0x1EE17, 'M', 'خ'),
    (0x1EE18, 'M', 'ذ'),
    (0x1EE19, 'M', 'ض'),
    (0x1EE1A, 'M', 'ظ'),
    (0x1EE1B, 'M', 'غ'),
    (0x1EE1C, 'M', 'ٮ'),
    (0x1EE1D, 'M', 'ں'),
    (0x1EE1E, 'M', 'ڡ'),
    (0x1EE1F, 'M', 'ٯ'),
    (0x1EE20, 'X'),
    (0x1EE21, 'M', 'ب'),
    (0x1EE22, 'M', 'ج'),
    (0x1EE23, 'X'),
    (0x1EE24, 'M', 'ه'),
    (0x1EE25, 'X'),
    (0x1EE27, 'M', 'ح'),
    (0x1EE28, 'X'),
    (0x1EE29, 'M', 'ي'),
    (0x1EE2A, 'M', 'ك'),
    (0x1EE2B, 'M', 'ل'),
    (0x1EE2C, 'M', 'م'),
    (0x1EE2D, 'M', 'ن'),
    (0x1EE2E, 'M', 'س'),
    (0x1EE2F, 'M', 'ع'),
    (0x1EE30, 'M', 'ف'),
    (0x1EE31, 'M', 'ص'),
    (0x1EE32, 'M', 'ق'),
    (0x1EE33, 'X'),
    (0x1EE34, 'M', 'ش'),
    (0x1EE35, 'M', 'ت'),
    (0x1EE36, 'M', 'ث'),
    (0x1EE37, 'M', 'خ'),
    (0x1EE38, 'X'),
    (0x1EE39, 'M', 'ض'),
    (0x1EE3A, 'X'),
    (0x1EE3B, 'M', 'غ'),
    (0x1EE3C, 'X'),
    (0x1EE42, 'M', 'ج'),
    (0x1EE43, 'X'),
    (0x1EE47, 'M', 'ح'),
    (0x1EE48, 'X'),
    (0x1EE49, 'M', 'ي'),
    (0x1EE4A, 'X'),
    (0x1EE4B, 'M', 'ل'),
    (0x1EE4C, 'X'),
    (0x1EE4D, 'M', 'ن'),
    (0x1EE4E, 'M', 'س'),
    (0x1EE4F, 'M', 'ع'),
    (0x1EE50, 'X'),
    (0x1EE51, 'M', 'ص'),
    (0x1EE52, 'M', 'ق'),
    (0x1EE53, 'X'),
    (0x1EE54, 'M', 'ش'),
    (0x1EE55, 'X'),
    (0x1EE57, 'M', 'خ'),
    (0x1EE58, 'X'),
    (0x1EE59, 'M', 'ض'),
    (0x1EE5A, 'X'),
    (0x1EE5B, 'M', 'غ'),
    (0x1EE5C, 'X'),
    (0x1EE5D, 'M', 'ں'),
    (0x1EE5E, 'X'),
    (0x1EE5F, 'M', 'ٯ'),
    (0x1EE60, 'X'),
    (0x1EE61, 'M', 'ب'),
    (0x1EE62, 'M', 'ج'),
    (0x1EE63, 'X'),
    (0x1EE64, 'M', 'ه'),
    (0x1EE65, 'X'),
    (0x1EE67, 'M', 'ح'),
    (0x1EE68, 'M', 'ط'),
    (0x1EE69, 'M', 'ي'),
    (0x1EE6A, 'M', 'ك'),
    ]

def _seg_71():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1EE6B, 'X'),
    (0x1EE6C, 'M', 'م'),
    (0x1EE6D, 'M', 'ن'),
    (0x1EE6E, 'M', 'س'),
    (0x1EE6F, 'M', 'ع'),
    (0x1EE70, 'M', 'ف'),
    (0x1EE71, 'M', 'ص'),
    (0x1EE72, 'M', 'ق'),
    (0x1EE73, 'X'),
    (0x1EE74, 'M', 'ش'),
    (0x1EE75, 'M', 'ت'),
    (0x1EE76, 'M', 'ث'),
    (0x1EE77, 'M', 'خ'),
    (0x1EE78, 'X'),
    (0x1EE79, 'M', 'ض'),
    (0x1EE7A, 'M', 'ظ'),
    (0x1EE7B, 'M', 'غ'),
    (0x1EE7C, 'M', 'ٮ'),
    (0x1EE7D, 'X'),
    (0x1EE7E, 'M', 'ڡ'),
    (0x1EE7F, 'X'),
    (0x1EE80, 'M', 'ا'),
    (0x1EE81, 'M', 'ب'),
    (0x1EE82, 'M', 'ج'),
    (0x1EE83, 'M', 'د'),
    (0x1EE84, 'M', 'ه'),
    (0x1EE85, 'M', 'و'),
    (0x1EE86, 'M', 'ز'),
    (0x1EE87, 'M', 'ح'),
    (0x1EE88, 'M', 'ط'),
    (0x1EE89, 'M', 'ي'),
    (0x1EE8A, 'X'),
    (0x1EE8B, 'M', 'ل'),
    (0x1EE8C, 'M', 'م'),
    (0x1EE8D, 'M', 'ن'),
    (0x1EE8E, 'M', 'س'),
    (0x1EE8F, 'M', 'ع'),
    (0x1EE90, 'M', 'ف'),
    (0x1EE91, 'M', 'ص'),
    (0x1EE92, 'M', 'ق'),
    (0x1EE93, 'M', 'ر'),
    (0x1EE94, 'M', 'ش'),
    (0x1EE95, 'M', 'ت'),
    (0x1EE96, 'M', 'ث'),
    (0x1EE97, 'M', 'خ'),
    (0x1EE98, 'M', 'ذ'),
    (0x1EE99, 'M', 'ض'),
    (0x1EE9A, 'M', 'ظ'),
    (0x1EE9B, 'M', 'غ'),
    (0x1EE9C, 'X'),
    (0x1EEA1, 'M', 'ب'),
    (0x1EEA2, 'M', 'ج'),
    (0x1EEA3, 'M', 'د'),
    (0x1EEA4, 'X'),
    (0x1EEA5, 'M', 'و'),
    (0x1EEA6, 'M', 'ز'),
    (0x1EEA7, 'M', 'ح'),
    (0x1EEA8, 'M', 'ط'),
    (0x1EEA9, 'M', 'ي'),
    (0x1EEAA, 'X'),
    (0x1EEAB, 'M', 'ل'),
    (0x1EEAC, 'M', 'م'),
    (0x1EEAD, 'M', 'ن'),
    (0x1EEAE, 'M', 'س'),
    (0x1EEAF, 'M', 'ع'),
    (0x1EEB0, 'M', 'ف'),
    (0x1EEB1, 'M', 'ص'),
    (0x1EEB2, 'M', 'ق'),
    (0x1EEB3, 'M', 'ر'),
    (0x1EEB4, 'M', 'ش'),
    (0x1EEB5, 'M', 'ت'),
    (0x1EEB6, 'M', 'ث'),
    (0x1EEB7, 'M', 'خ'),
    (0x1EEB8, 'M', 'ذ'),
    (0x1EEB9, 'M', 'ض'),
    (0x1EEBA, 'M', 'ظ'),
    (0x1EEBB, 'M', 'غ'),
    (0x1EEBC, 'X'),
    (0x1EEF0, 'V'),
    (0x1EEF2, 'X'),
    (0x1F000, 'V'),
    (0x1F02C, 'X'),
    (0x1F030, 'V'),
    (0x1F094, 'X'),
    (0x1F0A0, 'V'),
    (0x1F0AF, 'X'),
    (0x1F0B1, 'V'),
    (0x1F0C0, 'X'),
    (0x1F0C1, 'V'),
    (0x1F0D0, 'X'),
    (0x1F0D1, 'V'),
    (0x1F0F6, 'X'),
    (0x1F101, '3', '0,'),
    (0x1F102, '3', '1,'),
    (0x1F103, '3', '2,'),
    (0x1F104, '3', '3,'),
    (0x1F105, '3', '4,'),
    (0x1F106, '3', '5,'),
    (0x1F107, '3', '6,'),
    (0x1F108, '3', '7,'),
    ]

def _seg_72():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1F109, '3', '8,'),
    (0x1F10A, '3', '9,'),
    (0x1F10B, 'V'),
    (0x1F110, '3', '(a)'),
    (0x1F111, '3', '(b)'),
    (0x1F112, '3', '(c)'),
    (0x1F113, '3', '(d)'),
    (0x1F114, '3', '(e)'),
    (0x1F115, '3', '(f)'),
    (0x1F116, '3', '(g)'),
    (0x1F117, '3', '(h)'),
    (0x1F118, '3', '(i)'),
    (0x1F119, '3', '(j)'),
    (0x1F11A, '3', '(k)'),
    (0x1F11B, '3', '(l)'),
    (0x1F11C, '3', '(m)'),
    (0x1F11D, '3', '(n)'),
    (0x1F11E, '3', '(o)'),
    (0x1F11F, '3', '(p)'),
    (0x1F120, '3', '(q)'),
    (0x1F121, '3', '(r)'),
    (0x1F122, '3', '(s)'),
    (0x1F123, '3', '(t)'),
    (0x1F124, '3', '(u)'),
    (0x1F125, '3', '(v)'),
    (0x1F126, '3', '(w)'),
    (0x1F127, '3', '(x)'),
    (0x1F128, '3', '(y)'),
    (0x1F129, '3', '(z)'),
    (0x1F12A, 'M', '〔s〕'),
    (0x1F12B, 'M', 'c'),
    (0x1F12C, 'M', 'r'),
    (0x1F12D, 'M', 'cd'),
    (0x1F12E, 'M', 'wz'),
    (0x1F12F, 'V'),
    (0x1F130, 'M', 'a'),
    (0x1F131, 'M', 'b'),
    (0x1F132, 'M', 'c'),
    (0x1F133, 'M', 'd'),
    (0x1F134, 'M', 'e'),
    (0x1F135, 'M', 'f'),
    (0x1F136, 'M', 'g'),
    (0x1F137, 'M', 'h'),
    (0x1F138, 'M', 'i'),
    (0x1F139, 'M', 'j'),
    (0x1F13A, 'M', 'k'),
    (0x1F13B, 'M', 'l'),
    (0x1F13C, 'M', 'm'),
    (0x1F13D, 'M', 'n'),
    (0x1F13E, 'M', 'o'),
    (0x1F13F, 'M', 'p'),
    (0x1F140, 'M', 'q'),
    (0x1F141, 'M', 'r'),
    (0x1F142, 'M', 's'),
    (0x1F143, 'M', 't'),
    (0x1F144, 'M', 'u'),
    (0x1F145, 'M', 'v'),
    (0x1F146, 'M', 'w'),
    (0x1F147, 'M', 'x'),
    (0x1F148, 'M', 'y'),
    (0x1F149, 'M', 'z'),
    (0x1F14A, 'M', 'hv'),
    (0x1F14B, 'M', 'mv'),
    (0x1F14C, 'M', 'sd'),
    (0x1F14D, 'M', 'ss'),
    (0x1F14E, 'M', 'ppv'),
    (0x1F14F, 'M', 'wc'),
    (0x1F150, 'V'),
    (0x1F16A, 'M', 'mc'),
    (0x1F16B, 'M', 'md'),
    (0x1F16C, 'M', 'mr'),
    (0x1F16D, 'V'),
    (0x1F190, 'M', 'dj'),
    (0x1F191, 'V'),
    (0x1F1AE, 'X'),
    (0x1F1E6, 'V'),
    (0x1F200, 'M', 'ほか'),
    (0x1F201, 'M', 'ココ'),
    (0x1F202, 'M', 'サ'),
    (0x1F203, 'X'),
    (0x1F210, 'M', '手'),
    (0x1F211, 'M', '字'),
    (0x1F212, 'M', '双'),
    (0x1F213, 'M', 'デ'),
    (0x1F214, 'M', '二'),
    (0x1F215, 'M', '多'),
    (0x1F216, 'M', '解'),
    (0x1F217, 'M', '天'),
    (0x1F218, 'M', '交'),
    (0x1F219, 'M', '映'),
    (0x1F21A, 'M', '無'),
    (0x1F21B, 'M', '料'),
    (0x1F21C, 'M', '前'),
    (0x1F21D, 'M', '後'),
    (0x1F21E, 'M', '再'),
    (0x1F21F, 'M', '新'),
    (0x1F220, 'M', '初'),
    (0x1F221, 'M', '終'),
    (0x1F222, 'M', '生'),
    (0x1F223, 'M', '販'),
    ]

def _seg_73():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1F224, 'M', '声'),
    (0x1F225, 'M', '吹'),
    (0x1F226, 'M', '演'),
    (0x1F227, 'M', '投'),
    (0x1F228, 'M', '捕'),
    (0x1F229, 'M', '一'),
    (0x1F22A, 'M', '三'),
    (0x1F22B, 'M', '遊'),
    (0x1F22C, 'M', '左'),
    (0x1F22D, 'M', '中'),
    (0x1F22E, 'M', '右'),
    (0x1F22F, 'M', '指'),
    (0x1F230, 'M', '走'),
    (0x1F231, 'M', '打'),
    (0x1F232, 'M', '禁'),
    (0x1F233, 'M', '空'),
    (0x1F234, 'M', '合'),
    (0x1F235, 'M', '満'),
    (0x1F236, 'M', '有'),
    (0x1F237, 'M', '月'),
    (0x1F238, 'M', '申'),
    (0x1F239, 'M', '割'),
    (0x1F23A, 'M', '営'),
    (0x1F23B, 'M', '配'),
    (0x1F23C, 'X'),
    (0x1F240, 'M', '〔本〕'),
    (0x1F241, 'M', '〔三〕'),
    (0x1F242, 'M', '〔二〕'),
    (0x1F243, 'M', '〔安〕'),
    (0x1F244, 'M', '〔点〕'),
    (0x1F245, 'M', '〔打〕'),
    (0x1F246, 'M', '〔盗〕'),
    (0x1F247, 'M', '〔勝〕'),
    (0x1F248, 'M', '〔敗〕'),
    (0x1F249, 'X'),
    (0x1F250, 'M', '得'),
    (0x1F251, 'M', '可'),
    (0x1F252, 'X'),
    (0x1F260, 'V'),
    (0x1F266, 'X'),
    (0x1F300, 'V'),
    (0x1F6D8, 'X'),
    (0x1F6E0, 'V'),
    (0x1F6ED, 'X'),
    (0x1F6F0, 'V'),
    (0x1F6FD, 'X'),
    (0x1F700, 'V'),
    (0x1F774, 'X'),
    (0x1F780, 'V'),
    (0x1F7D9, 'X'),
    (0x1F7E0, 'V'),
    (0x1F7EC, 'X'),
    (0x1F800, 'V'),
    (0x1F80C, 'X'),
    (0x1F810, 'V'),
    (0x1F848, 'X'),
    (0x1F850, 'V'),
    (0x1F85A, 'X'),
    (0x1F860, 'V'),
    (0x1F888, 'X'),
    (0x1F890, 'V'),
    (0x1F8AE, 'X'),
    (0x1F8B0, 'V'),
    (0x1F8B2, 'X'),
    (0x1F900, 'V'),
    (0x1F979, 'X'),
    (0x1F97A, 'V'),
    (0x1F9CC, 'X'),
    (0x1F9CD, 'V'),
    (0x1FA54, 'X'),
    (0x1FA60, 'V'),
    (0x1FA6E, 'X'),
    (0x1FA70, 'V'),
    (0x1FA75, 'X'),
    (0x1FA78, 'V'),
    (0x1FA7B, 'X'),
    (0x1FA80, 'V'),
    (0x1FA87, 'X'),
    (0x1FA90, 'V'),
    (0x1FAA9, 'X'),
    (0x1FAB0, 'V'),
    (0x1FAB7, 'X'),
    (0x1FAC0, 'V'),
    (0x1FAC3, 'X'),
    (0x1FAD0, 'V'),
    (0x1FAD7, 'X'),
    (0x1FB00, 'V'),
    (0x1FB93, 'X'),
    (0x1FB94, 'V'),
    (0x1FBCB, 'X'),
    (0x1FBF0, 'M', '0'),
    (0x1FBF1, 'M', '1'),
    (0x1FBF2, 'M', '2'),
    (0x1FBF3, 'M', '3'),
    (0x1FBF4, 'M', '4'),
    (0x1FBF5, 'M', '5'),
    (0x1FBF6, 'M', '6'),
    (0x1FBF7, 'M', '7'),
    (0x1FBF8, 'M', '8'),
    (0x1FBF9, 'M', '9'),
    ]

def _seg_74():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x1FBFA, 'X'),
    (0x20000, 'V'),
    (0x2A6DE, 'X'),
    (0x2A700, 'V'),
    (0x2B735, 'X'),
    (0x2B740, 'V'),
    (0x2B81E, 'X'),
    (0x2B820, 'V'),
    (0x2CEA2, 'X'),
    (0x2CEB0, 'V'),
    (0x2EBE1, 'X'),
    (0x2F800, 'M', '丽'),
    (0x2F801, 'M', '丸'),
    (0x2F802, 'M', '乁'),
    (0x2F803, 'M', '𠄢'),
    (0x2F804, 'M', '你'),
    (0x2F805, 'M', '侮'),
    (0x2F806, 'M', '侻'),
    (0x2F807, 'M', '倂'),
    (0x2F808, 'M', '偺'),
    (0x2F809, 'M', '備'),
    (0x2F80A, 'M', '僧'),
    (0x2F80B, 'M', '像'),
    (0x2F80C, 'M', '㒞'),
    (0x2F80D, 'M', '𠘺'),
    (0x2F80E, 'M', '免'),
    (0x2F80F, 'M', '兔'),
    (0x2F810, 'M', '兤'),
    (0x2F811, 'M', '具'),
    (0x2F812, 'M', '𠔜'),
    (0x2F813, 'M', '㒹'),
    (0x2F814, 'M', '內'),
    (0x2F815, 'M', '再'),
    (0x2F816, 'M', '𠕋'),
    (0x2F817, 'M', '冗'),
    (0x2F818, 'M', '冤'),
    (0x2F819, 'M', '仌'),
    (0x2F81A, 'M', '冬'),
    (0x2F81B, 'M', '况'),
    (0x2F81C, 'M', '𩇟'),
    (0x2F81D, 'M', '凵'),
    (0x2F81E, 'M', '刃'),
    (0x2F81F, 'M', '㓟'),
    (0x2F820, 'M', '刻'),
    (0x2F821, 'M', '剆'),
    (0x2F822, 'M', '割'),
    (0x2F823, 'M', '剷'),
    (0x2F824, 'M', '㔕'),
    (0x2F825, 'M', '勇'),
    (0x2F826, 'M', '勉'),
    (0x2F827, 'M', '勤'),
    (0x2F828, 'M', '勺'),
    (0x2F829, 'M', '包'),
    (0x2F82A, 'M', '匆'),
    (0x2F82B, 'M', '北'),
    (0x2F82C, 'M', '卉'),
    (0x2F82D, 'M', '卑'),
    (0x2F82E, 'M', '博'),
    (0x2F82F, 'M', '即'),
    (0x2F830, 'M', '卽'),
    (0x2F831, 'M', '卿'),
    (0x2F834, 'M', '𠨬'),
    (0x2F835, 'M', '灰'),
    (0x2F836, 'M', '及'),
    (0x2F837, 'M', '叟'),
    (0x2F838, 'M', '𠭣'),
    (0x2F839, 'M', '叫'),
    (0x2F83A, 'M', '叱'),
    (0x2F83B, 'M', '吆'),
    (0x2F83C, 'M', '咞'),
    (0x2F83D, 'M', '吸'),
    (0x2F83E, 'M', '呈'),
    (0x2F83F, 'M', '周'),
    (0x2F840, 'M', '咢'),
    (0x2F841, 'M', '哶'),
    (0x2F842, 'M', '唐'),
    (0x2F843, 'M', '啓'),
    (0x2F844, 'M', '啣'),
    (0x2F845, 'M', '善'),
    (0x2F847, 'M', '喙'),
    (0x2F848, 'M', '喫'),
    (0x2F849, 'M', '喳'),
    (0x2F84A, 'M', '嗂'),
    (0x2F84B, 'M', '圖'),
    (0x2F84C, 'M', '嘆'),
    (0x2F84D, 'M', '圗'),
    (0x2F84E, 'M', '噑'),
    (0x2F84F, 'M', '噴'),
    (0x2F850, 'M', '切'),
    (0x2F851, 'M', '壮'),
    (0x2F852, 'M', '城'),
    (0x2F853, 'M', '埴'),
    (0x2F854, 'M', '堍'),
    (0x2F855, 'M', '型'),
    (0x2F856, 'M', '堲'),
    (0x2F857, 'M', '報'),
    (0x2F858, 'M', '墬'),
    (0x2F859, 'M', '𡓤'),
    (0x2F85A, 'M', '売'),
    (0x2F85B, 'M', '壷'),
    ]

def _seg_75():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F85C, 'M', '夆'),
    (0x2F85D, 'M', '多'),
    (0x2F85E, 'M', '夢'),
    (0x2F85F, 'M', '奢'),
    (0x2F860, 'M', '𡚨'),
    (0x2F861, 'M', '𡛪'),
    (0x2F862, 'M', '姬'),
    (0x2F863, 'M', '娛'),
    (0x2F864, 'M', '娧'),
    (0x2F865, 'M', '姘'),
    (0x2F866, 'M', '婦'),
    (0x2F867, 'M', '㛮'),
    (0x2F868, 'X'),
    (0x2F869, 'M', '嬈'),
    (0x2F86A, 'M', '嬾'),
    (0x2F86C, 'M', '𡧈'),
    (0x2F86D, 'M', '寃'),
    (0x2F86E, 'M', '寘'),
    (0x2F86F, 'M', '寧'),
    (0x2F870, 'M', '寳'),
    (0x2F871, 'M', '𡬘'),
    (0x2F872, 'M', '寿'),
    (0x2F873, 'M', '将'),
    (0x2F874, 'X'),
    (0x2F875, 'M', '尢'),
    (0x2F876, 'M', '㞁'),
    (0x2F877, 'M', '屠'),
    (0x2F878, 'M', '屮'),
    (0x2F879, 'M', '峀'),
    (0x2F87A, 'M', '岍'),
    (0x2F87B, 'M', '𡷤'),
    (0x2F87C, 'M', '嵃'),
    (0x2F87D, 'M', '𡷦'),
    (0x2F87E, 'M', '嵮'),
    (0x2F87F, 'M', '嵫'),
    (0x2F880, 'M', '嵼'),
    (0x2F881, 'M', '巡'),
    (0x2F882, 'M', '巢'),
    (0x2F883, 'M', '㠯'),
    (0x2F884, 'M', '巽'),
    (0x2F885, 'M', '帨'),
    (0x2F886, 'M', '帽'),
    (0x2F887, 'M', '幩'),
    (0x2F888, 'M', '㡢'),
    (0x2F889, 'M', '𢆃'),
    (0x2F88A, 'M', '㡼'),
    (0x2F88B, 'M', '庰'),
    (0x2F88C, 'M', '庳'),
    (0x2F88D, 'M', '庶'),
    (0x2F88E, 'M', '廊'),
    (0x2F88F, 'M', '𪎒'),
    (0x2F890, 'M', '廾'),
    (0x2F891, 'M', '𢌱'),
    (0x2F893, 'M', '舁'),
    (0x2F894, 'M', '弢'),
    (0x2F896, 'M', '㣇'),
    (0x2F897, 'M', '𣊸'),
    (0x2F898, 'M', '𦇚'),
    (0x2F899, 'M', '形'),
    (0x2F89A, 'M', '彫'),
    (0x2F89B, 'M', '㣣'),
    (0x2F89C, 'M', '徚'),
    (0x2F89D, 'M', '忍'),
    (0x2F89E, 'M', '志'),
    (0x2F89F, 'M', '忹'),
    (0x2F8A0, 'M', '悁'),
    (0x2F8A1, 'M', '㤺'),
    (0x2F8A2, 'M', '㤜'),
    (0x2F8A3, 'M', '悔'),
    (0x2F8A4, 'M', '𢛔'),
    (0x2F8A5, 'M', '惇'),
    (0x2F8A6, 'M', '慈'),
    (0x2F8A7, 'M', '慌'),
    (0x2F8A8, 'M', '慎'),
    (0x2F8A9, 'M', '慌'),
    (0x2F8AA, 'M', '慺'),
    (0x2F8AB, 'M', '憎'),
    (0x2F8AC, 'M', '憲'),
    (0x2F8AD, 'M', '憤'),
    (0x2F8AE, 'M', '憯'),
    (0x2F8AF, 'M', '懞'),
    (0x2F8B0, 'M', '懲'),
    (0x2F8B1, 'M', '懶'),
    (0x2F8B2, 'M', '成'),
    (0x2F8B3, 'M', '戛'),
    (0x2F8B4, 'M', '扝'),
    (0x2F8B5, 'M', '抱'),
    (0x2F8B6, 'M', '拔'),
    (0x2F8B7, 'M', '捐'),
    (0x2F8B8, 'M', '𢬌'),
    (0x2F8B9, 'M', '挽'),
    (0x2F8BA, 'M', '拼'),
    (0x2F8BB, 'M', '捨'),
    (0x2F8BC, 'M', '掃'),
    (0x2F8BD, 'M', '揤'),
    (0x2F8BE, 'M', '𢯱'),
    (0x2F8BF, 'M', '搢'),
    (0x2F8C0, 'M', '揅'),
    (0x2F8C1, 'M', '掩'),
    (0x2F8C2, 'M', '㨮'),
    ]

def _seg_76():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F8C3, 'M', '摩'),
    (0x2F8C4, 'M', '摾'),
    (0x2F8C5, 'M', '撝'),
    (0x2F8C6, 'M', '摷'),
    (0x2F8C7, 'M', '㩬'),
    (0x2F8C8, 'M', '敏'),
    (0x2F8C9, 'M', '敬'),
    (0x2F8CA, 'M', '𣀊'),
    (0x2F8CB, 'M', '旣'),
    (0x2F8CC, 'M', '書'),
    (0x2F8CD, 'M', '晉'),
    (0x2F8CE, 'M', '㬙'),
    (0x2F8CF, 'M', '暑'),
    (0x2F8D0, 'M', '㬈'),
    (0x2F8D1, 'M', '㫤'),
    (0x2F8D2, 'M', '冒'),
    (0x2F8D3, 'M', '冕'),
    (0x2F8D4, 'M', '最'),
    (0x2F8D5, 'M', '暜'),
    (0x2F8D6, 'M', '肭'),
    (0x2F8D7, 'M', '䏙'),
    (0x2F8D8, 'M', '朗'),
    (0x2F8D9, 'M', '望'),
    (0x2F8DA, 'M', '朡'),
    (0x2F8DB, 'M', '杞'),
    (0x2F8DC, 'M', '杓'),
    (0x2F8DD, 'M', '𣏃'),
    (0x2F8DE, 'M', '㭉'),
    (0x2F8DF, 'M', '柺'),
    (0x2F8E0, 'M', '枅'),
    (0x2F8E1, 'M', '桒'),
    (0x2F8E2, 'M', '梅'),
    (0x2F8E3, 'M', '𣑭'),
    (0x2F8E4, 'M', '梎'),
    (0x2F8E5, 'M', '栟'),
    (0x2F8E6, 'M', '椔'),
    (0x2F8E7, 'M', '㮝'),
    (0x2F8E8, 'M', '楂'),
    (0x2F8E9, 'M', '榣'),
    (0x2F8EA, 'M', '槪'),
    (0x2F8EB, 'M', '檨'),
    (0x2F8EC, 'M', '𣚣'),
    (0x2F8ED, 'M', '櫛'),
    (0x2F8EE, 'M', '㰘'),
    (0x2F8EF, 'M', '次'),
    (0x2F8F0, 'M', '𣢧'),
    (0x2F8F1, 'M', '歔'),
    (0x2F8F2, 'M', '㱎'),
    (0x2F8F3, 'M', '歲'),
    (0x2F8F4, 'M', '殟'),
    (0x2F8F5, 'M', '殺'),
    (0x2F8F6, 'M', '殻'),
    (0x2F8F7, 'M', '𣪍'),
    (0x2F8F8, 'M', '𡴋'),
    (0x2F8F9, 'M', '𣫺'),
    (0x2F8FA, 'M', '汎'),
    (0x2F8FB, 'M', '𣲼'),
    (0x2F8FC, 'M', '沿'),
    (0x2F8FD, 'M', '泍'),
    (0x2F8FE, 'M', '汧'),
    (0x2F8FF, 'M', '洖'),
    (0x2F900, 'M', '派'),
    (0x2F901, 'M', '海'),
    (0x2F902, 'M', '流'),
    (0x2F903, 'M', '浩'),
    (0x2F904, 'M', '浸'),
    (0x2F905, 'M', '涅'),
    (0x2F906, 'M', '𣴞'),
    (0x2F907, 'M', '洴'),
    (0x2F908, 'M', '港'),
    (0x2F909, 'M', '湮'),
    (0x2F90A, 'M', '㴳'),
    (0x2F90B, 'M', '滋'),
    (0x2F90C, 'M', '滇'),
    (0x2F90D, 'M', '𣻑'),
    (0x2F90E, 'M', '淹'),
    (0x2F90F, 'M', '潮'),
    (0x2F910, 'M', '𣽞'),
    (0x2F911, 'M', '𣾎'),
    (0x2F912, 'M', '濆'),
    (0x2F913, 'M', '瀹'),
    (0x2F914, 'M', '瀞'),
    (0x2F915, 'M', '瀛'),
    (0x2F916, 'M', '㶖'),
    (0x2F917, 'M', '灊'),
    (0x2F918, 'M', '災'),
    (0x2F919, 'M', '灷'),
    (0x2F91A, 'M', '炭'),
    (0x2F91B, 'M', '𠔥'),
    (0x2F91C, 'M', '煅'),
    (0x2F91D, 'M', '𤉣'),
    (0x2F91E, 'M', '熜'),
    (0x2F91F, 'X'),
    (0x2F920, 'M', '爨'),
    (0x2F921, 'M', '爵'),
    (0x2F922, 'M', '牐'),
    (0x2F923, 'M', '𤘈'),
    (0x2F924, 'M', '犀'),
    (0x2F925, 'M', '犕'),
    (0x2F926, 'M', '𤜵'),
    ]

def _seg_77():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F927, 'M', '𤠔'),
    (0x2F928, 'M', '獺'),
    (0x2F929, 'M', '王'),
    (0x2F92A, 'M', '㺬'),
    (0x2F92B, 'M', '玥'),
    (0x2F92C, 'M', '㺸'),
    (0x2F92E, 'M', '瑇'),
    (0x2F92F, 'M', '瑜'),
    (0x2F930, 'M', '瑱'),
    (0x2F931, 'M', '璅'),
    (0x2F932, 'M', '瓊'),
    (0x2F933, 'M', '㼛'),
    (0x2F934, 'M', '甤'),
    (0x2F935, 'M', '𤰶'),
    (0x2F936, 'M', '甾'),
    (0x2F937, 'M', '𤲒'),
    (0x2F938, 'M', '異'),
    (0x2F939, 'M', '𢆟'),
    (0x2F93A, 'M', '瘐'),
    (0x2F93B, 'M', '𤾡'),
    (0x2F93C, 'M', '𤾸'),
    (0x2F93D, 'M', '𥁄'),
    (0x2F93E, 'M', '㿼'),
    (0x2F93F, 'M', '䀈'),
    (0x2F940, 'M', '直'),
    (0x2F941, 'M', '𥃳'),
    (0x2F942, 'M', '𥃲'),
    (0x2F943, 'M', '𥄙'),
    (0x2F944, 'M', '𥄳'),
    (0x2F945, 'M', '眞'),
    (0x2F946, 'M', '真'),
    (0x2F948, 'M', '睊'),
    (0x2F949, 'M', '䀹'),
    (0x2F94A, 'M', '瞋'),
    (0x2F94B, 'M', '䁆'),
    (0x2F94C, 'M', '䂖'),
    (0x2F94D, 'M', '𥐝'),
    (0x2F94E, 'M', '硎'),
    (0x2F94F, 'M', '碌'),
    (0x2F950, 'M', '磌'),
    (0x2F951, 'M', '䃣'),
    (0x2F952, 'M', '𥘦'),
    (0x2F953, 'M', '祖'),
    (0x2F954, 'M', '𥚚'),
    (0x2F955, 'M', '𥛅'),
    (0x2F956, 'M', '福'),
    (0x2F957, 'M', '秫'),
    (0x2F958, 'M', '䄯'),
    (0x2F959, 'M', '穀'),
    (0x2F95A, 'M', '穊'),
    (0x2F95B, 'M', '穏'),
    (0x2F95C, 'M', '𥥼'),
    (0x2F95D, 'M', '𥪧'),
    (0x2F95F, 'X'),
    (0x2F960, 'M', '䈂'),
    (0x2F961, 'M', '𥮫'),
    (0x2F962, 'M', '篆'),
    (0x2F963, 'M', '築'),
    (0x2F964, 'M', '䈧'),
    (0x2F965, 'M', '𥲀'),
    (0x2F966, 'M', '糒'),
    (0x2F967, 'M', '䊠'),
    (0x2F968, 'M', '糨'),
    (0x2F969, 'M', '糣'),
    (0x2F96A, 'M', '紀'),
    (0x2F96B, 'M', '𥾆'),
    (0x2F96C, 'M', '絣'),
    (0x2F96D, 'M', '䌁'),
    (0x2F96E, 'M', '緇'),
    (0x2F96F, 'M', '縂'),
    (0x2F970, 'M', '繅'),
    (0x2F971, 'M', '䌴'),
    (0x2F972, 'M', '𦈨'),
    (0x2F973, 'M', '𦉇'),
    (0x2F974, 'M', '䍙'),
    (0x2F975, 'M', '𦋙'),
    (0x2F976, 'M', '罺'),
    (0x2F977, 'M', '𦌾'),
    (0x2F978, 'M', '羕'),
    (0x2F979, 'M', '翺'),
    (0x2F97A, 'M', '者'),
    (0x2F97B, 'M', '𦓚'),
    (0x2F97C, 'M', '𦔣'),
    (0x2F97D, 'M', '聠'),
    (0x2F97E, 'M', '𦖨'),
    (0x2F97F, 'M', '聰'),
    (0x2F980, 'M', '𣍟'),
    (0x2F981, 'M', '䏕'),
    (0x2F982, 'M', '育'),
    (0x2F983, 'M', '脃'),
    (0x2F984, 'M', '䐋'),
    (0x2F985, 'M', '脾'),
    (0x2F986, 'M', '媵'),
    (0x2F987, 'M', '𦞧'),
    (0x2F988, 'M', '𦞵'),
    (0x2F989, 'M', '𣎓'),
    (0x2F98A, 'M', '𣎜'),
    (0x2F98B, 'M', '舁'),
    (0x2F98C, 'M', '舄'),
    (0x2F98D, 'M', '辞'),
    ]

def _seg_78():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F98E, 'M', '䑫'),
    (0x2F98F, 'M', '芑'),
    (0x2F990, 'M', '芋'),
    (0x2F991, 'M', '芝'),
    (0x2F992, 'M', '劳'),
    (0x2F993, 'M', '花'),
    (0x2F994, 'M', '芳'),
    (0x2F995, 'M', '芽'),
    (0x2F996, 'M', '苦'),
    (0x2F997, 'M', '𦬼'),
    (0x2F998, 'M', '若'),
    (0x2F999, 'M', '茝'),
    (0x2F99A, 'M', '荣'),
    (0x2F99B, 'M', '莭'),
    (0x2F99C, 'M', '茣'),
    (0x2F99D, 'M', '莽'),
    (0x2F99E, 'M', '菧'),
    (0x2F99F, 'M', '著'),
    (0x2F9A0, 'M', '荓'),
    (0x2F9A1, 'M', '菊'),
    (0x2F9A2, 'M', '菌'),
    (0x2F9A3, 'M', '菜'),
    (0x2F9A4, 'M', '𦰶'),
    (0x2F9A5, 'M', '𦵫'),
    (0x2F9A6, 'M', '𦳕'),
    (0x2F9A7, 'M', '䔫'),
    (0x2F9A8, 'M', '蓱'),
    (0x2F9A9, 'M', '蓳'),
    (0x2F9AA, 'M', '蔖'),
    (0x2F9AB, 'M', '𧏊'),
    (0x2F9AC, 'M', '蕤'),
    (0x2F9AD, 'M', '𦼬'),
    (0x2F9AE, 'M', '䕝'),
    (0x2F9AF, 'M', '䕡'),
    (0x2F9B0, 'M', '𦾱'),
    (0x2F9B1, 'M', '𧃒'),
    (0x2F9B2, 'M', '䕫'),
    (0x2F9B3, 'M', '虐'),
    (0x2F9B4, 'M', '虜'),
    (0x2F9B5, 'M', '虧'),
    (0x2F9B6, 'M', '虩'),
    (0x2F9B7, 'M', '蚩'),
    (0x2F9B8, 'M', '蚈'),
    (0x2F9B9, 'M', '蜎'),
    (0x2F9BA, 'M', '蛢'),
    (0x2F9BB, 'M', '蝹'),
    (0x2F9BC, 'M', '蜨'),
    (0x2F9BD, 'M', '蝫'),
    (0x2F9BE, 'M', '螆'),
    (0x2F9BF, 'X'),
    (0x2F9C0, 'M', '蟡'),
    (0x2F9C1, 'M', '蠁'),
    (0x2F9C2, 'M', '䗹'),
    (0x2F9C3, 'M', '衠'),
    (0x2F9C4, 'M', '衣'),
    (0x2F9C5, 'M', '𧙧'),
    (0x2F9C6, 'M', '裗'),
    (0x2F9C7, 'M', '裞'),
    (0x2F9C8, 'M', '䘵'),
    (0x2F9C9, 'M', '裺'),
    (0x2F9CA, 'M', '㒻'),
    (0x2F9CB, 'M', '𧢮'),
    (0x2F9CC, 'M', '𧥦'),
    (0x2F9CD, 'M', '䚾'),
    (0x2F9CE, 'M', '䛇'),
    (0x2F9CF, 'M', '誠'),
    (0x2F9D0, 'M', '諭'),
    (0x2F9D1, 'M', '變'),
    (0x2F9D2, 'M', '豕'),
    (0x2F9D3, 'M', '𧲨'),
    (0x2F9D4, 'M', '貫'),
    (0x2F9D5, 'M', '賁'),
    (0x2F9D6, 'M', '贛'),
    (0x2F9D7, 'M', '起'),
    (0x2F9D8, 'M', '𧼯'),
    (0x2F9D9, 'M', '𠠄'),
    (0x2F9DA, 'M', '跋'),
    (0x2F9DB, 'M', '趼'),
    (0x2F9DC, 'M', '跰'),
    (0x2F9DD, 'M', '𠣞'),
    (0x2F9DE, 'M', '軔'),
    (0x2F9DF, 'M', '輸'),
    (0x2F9E0, 'M', '𨗒'),
    (0x2F9E1, 'M', '𨗭'),
    (0x2F9E2, 'M', '邔'),
    (0x2F9E3, 'M', '郱'),
    (0x2F9E4, 'M', '鄑'),
    (0x2F9E5, 'M', '𨜮'),
    (0x2F9E6, 'M', '鄛'),
    (0x2F9E7, 'M', '鈸'),
    (0x2F9E8, 'M', '鋗'),
    (0x2F9E9, 'M', '鋘'),
    (0x2F9EA, 'M', '鉼'),
    (0x2F9EB, 'M', '鏹'),
    (0x2F9EC, 'M', '鐕'),
    (0x2F9ED, 'M', '𨯺'),
    (0x2F9EE, 'M', '開'),
    (0x2F9EF, 'M', '䦕'),
    (0x2F9F0, 'M', '閷'),
    (0x2F9F1, 'M', '𨵷'),
    ]

def _seg_79():
    # type: () -> List[Union[Tuple[int, str], Tuple[int, str, str]]]
    return [
    (0x2F9F2, 'M', '䧦'),
    (0x2F9F3, 'M', '雃'),
    (0x2F9F4, 'M', '嶲'),
    (0x2F9F5, 'M', '霣'),
    (0x2F9F6, 'M', '𩅅'),
    (0x2F9F7, 'M', '𩈚'),
    (0x2F9F8, 'M', '䩮'),
    (0x2F9F9, 'M', '䩶'),
    (0x2F9FA, 'M', '韠'),
    (0x2F9FB, 'M', '𩐊'),
    (0x2F9FC, 'M', '䪲'),
    (0x2F9FD, 'M', '𩒖'),
    (0x2F9FE, 'M', '頋'),
    (0x2FA00, 'M', '頩'),
    (0x2FA01, 'M', '𩖶'),
    (0x2FA02, 'M', '飢'),
    (0x2FA03, 'M', '䬳'),
    (0x2FA04, 'M', '餩'),
    (0x2FA05, 'M', '馧'),
    (0x2FA06, 'M', '駂'),
    (0x2FA07, 'M', '駾'),
    (0x2FA08, 'M', '䯎'),
    (0x2FA09, 'M', '𩬰'),
    (0x2FA0A, 'M', '鬒'),
    (0x2FA0B, 'M', '鱀'),
    (0x2FA0C, 'M', '鳽'),
    (0x2FA0D, 'M', '䳎'),
    (0x2FA0E, 'M', '䳭'),
    (0x2FA0F, 'M', '鵧'),
    (0x2FA10, 'M', '𪃎'),
    (0x2FA11, 'M', '䳸'),
    (0x2FA12, 'M', '𪄅'),
    (0x2FA13, 'M', '𪈎'),
    (0x2FA14, 'M', '𪊑'),
    (0x2FA15, 'M', '麻'),
    (0x2FA16, 'M', '䵖'),
    (0x2FA17, 'M', '黹'),
    (0x2FA18, 'M', '黾'),
    (0x2FA19, 'M', '鼅'),
    (0x2FA1A, 'M', '鼏'),
    (0x2FA1B, 'M', '鼖'),
    (0x2FA1C, 'M', '鼻'),
    (0x2FA1D, 'M', '𪘀'),
    (0x2FA1E, 'X'),
    (0x30000, 'V'),
    (0x3134B, 'X'),
    (0xE0100, 'I'),
    (0xE01F0, 'X'),
    ]

uts46data = tuple(
    _seg_0()
    + _seg_1()
    + _seg_2()
    + _seg_3()
    + _seg_4()
    + _seg_5()
    + _seg_6()
    + _seg_7()
    + _seg_8()
    + _seg_9()
    + _seg_10()
    + _seg_11()
    + _seg_12()
    + _seg_13()
    + _seg_14()
    + _seg_15()
    + _seg_16()
    + _seg_17()
    + _seg_18()
    + _seg_19()
    + _seg_20()
    + _seg_21()
    + _seg_22()
    + _seg_23()
    + _seg_24()
    + _seg_25()
    + _seg_26()
    + _seg_27()
    + _seg_28()
    + _seg_29()
    + _seg_30()
    + _seg_31()
    + _seg_32()
    + _seg_33()
    + _seg_34()
    + _seg_35()
    + _seg_36()
    + _seg_37()
    + _seg_38()
    + _seg_39()
    + _seg_40()
    + _seg_41()
    + _seg_42()
    + _seg_43()
    + _seg_44()
    + _seg_45()
    + _seg_46()
    + _seg_47()
    + _seg_48()
    + _seg_49()
    + _seg_50()
    + _seg_51()
    + _seg_52()
    + _seg_53()
    + _seg_54()
    + _seg_55()
    + _seg_56()
    + _seg_57()
    + _seg_58()
    + _seg_59()
    + _seg_60()
    + _seg_61()
    + _seg_62()
    + _seg_63()
    + _seg_64()
    + _seg_65()
    + _seg_66()
    + _seg_67()
    + _seg_68()
    + _seg_69()
    + _seg_70()
    + _seg_71()
    + _seg_72()
    + _seg_73()
    + _seg_74()
    + _seg_75()
    + _seg_76()
    + _seg_77()
    + _seg_78()
    + _seg_79()
)  # type: Tuple[Union[Tuple[int, str], Tuple[int, str, str]], ...]
site-packages/pip/_vendor/idna/compat.py000064400000000550150564745010014267 0ustar00from .core import *
from .codec import *
from typing import Any, Union

def ToASCII(label):
    # type: (str) -> bytes
    return encode(label)

def ToUnicode(label):
    # type: (Union[bytes, bytearray]) -> str
    return decode(label)

def nameprep(s):
    # type: (Any) -> None
    raise NotImplementedError('IDNA 2008 does not utilise nameprep protocol')

site-packages/pip/_vendor/idna/codec.py000064400000006575150564745010014076 0ustar00from .core import encode, decode, alabel, ulabel, IDNAError
import codecs
import re
from typing import Tuple, Optional

_unicode_dots_re = re.compile('[\u002e\u3002\uff0e\uff61]')

class Codec(codecs.Codec):

    def encode(self, data, errors='strict'):
        # type: (str, str) -> Tuple[bytes, int]
        if errors != 'strict':
            raise IDNAError('Unsupported error handling \"{}\"'.format(errors))

        if not data:
            return b"", 0

        return encode(data), len(data)

    def decode(self, data, errors='strict'):
        # type: (bytes, str) -> Tuple[str, int]
        if errors != 'strict':
            raise IDNAError('Unsupported error handling \"{}\"'.format(errors))

        if not data:
            return '', 0

        return decode(data), len(data)

class IncrementalEncoder(codecs.BufferedIncrementalEncoder):
    def _buffer_encode(self, data, errors, final):  # type: ignore
        # type: (str, str, bool) -> Tuple[str, int]
        if errors != 'strict':
            raise IDNAError('Unsupported error handling \"{}\"'.format(errors))

        if not data:
            return "", 0

        labels = _unicode_dots_re.split(data)
        trailing_dot = ''
        if labels:
            if not labels[-1]:
                trailing_dot = '.'
                del labels[-1]
            elif not final:
                # Keep potentially unfinished label until the next call
                del labels[-1]
                if labels:
                    trailing_dot = '.'

        result = []
        size = 0
        for label in labels:
            result.append(alabel(label))
            if size:
                size += 1
            size += len(label)

        # Join with U+002E
        result_str = '.'.join(result) + trailing_dot  # type: ignore
        size += len(trailing_dot)
        return result_str, size

class IncrementalDecoder(codecs.BufferedIncrementalDecoder):
    def _buffer_decode(self, data, errors, final):  # type: ignore
        # type: (str, str, bool) -> Tuple[str, int]
        if errors != 'strict':
            raise IDNAError('Unsupported error handling \"{}\"'.format(errors))

        if not data:
            return ('', 0)

        labels = _unicode_dots_re.split(data)
        trailing_dot = ''
        if labels:
            if not labels[-1]:
                trailing_dot = '.'
                del labels[-1]
            elif not final:
                # Keep potentially unfinished label until the next call
                del labels[-1]
                if labels:
                    trailing_dot = '.'

        result = []
        size = 0
        for label in labels:
            result.append(ulabel(label))
            if size:
                size += 1
            size += len(label)

        result_str = '.'.join(result) + trailing_dot
        size += len(trailing_dot)
        return (result_str, size)


class StreamWriter(Codec, codecs.StreamWriter):
    pass


class StreamReader(Codec, codecs.StreamReader):
    pass


def getregentry():
    # type: () -> codecs.CodecInfo
    # Compatibility as a search_function for codecs.register()
    return codecs.CodecInfo(
        name='idna',
        encode=Codec().encode,  # type: ignore
        decode=Codec().decode,  # type: ignore
        incrementalencoder=IncrementalEncoder,
        incrementaldecoder=IncrementalDecoder,
        streamwriter=StreamWriter,
        streamreader=StreamReader,
    )
site-packages/pip/_vendor/idna/package_data.py000064400000000025150564745010015365 0ustar00__version__ = '3.2'

site-packages/pip/_vendor/idna/__init__.py000064400000001521150564745010014542 0ustar00from .package_data import __version__
from .core import (
    IDNABidiError,
    IDNAError,
    InvalidCodepoint,
    InvalidCodepointContext,
    alabel,
    check_bidi,
    check_hyphen_ok,
    check_initial_combiner,
    check_label,
    check_nfc,
    decode,
    encode,
    ulabel,
    uts46_remap,
    valid_contextj,
    valid_contexto,
    valid_label_length,
    valid_string_length,
)
from .intranges import intranges_contain

__all__ = [
    "IDNABidiError",
    "IDNAError",
    "InvalidCodepoint",
    "InvalidCodepointContext",
    "alabel",
    "check_bidi",
    "check_hyphen_ok",
    "check_initial_combiner",
    "check_label",
    "check_nfc",
    "decode",
    "encode",
    "intranges_contain",
    "ulabel",
    "uts46_remap",
    "valid_contextj",
    "valid_contexto",
    "valid_label_length",
    "valid_string_length",
]
site-packages/pip/_vendor/idna/idnadata.py000064400000122556150564745010014564 0ustar00# This file is automatically generated by tools/idna-data

__version__ = '13.0.0'
scripts = {
    'Greek': (
        0x37000000374,
        0x37500000378,
        0x37a0000037e,
        0x37f00000380,
        0x38400000385,
        0x38600000387,
        0x3880000038b,
        0x38c0000038d,
        0x38e000003a2,
        0x3a3000003e2,
        0x3f000000400,
        0x1d2600001d2b,
        0x1d5d00001d62,
        0x1d6600001d6b,
        0x1dbf00001dc0,
        0x1f0000001f16,
        0x1f1800001f1e,
        0x1f2000001f46,
        0x1f4800001f4e,
        0x1f5000001f58,
        0x1f5900001f5a,
        0x1f5b00001f5c,
        0x1f5d00001f5e,
        0x1f5f00001f7e,
        0x1f8000001fb5,
        0x1fb600001fc5,
        0x1fc600001fd4,
        0x1fd600001fdc,
        0x1fdd00001ff0,
        0x1ff200001ff5,
        0x1ff600001fff,
        0x212600002127,
        0xab650000ab66,
        0x101400001018f,
        0x101a0000101a1,
        0x1d2000001d246,
    ),
    'Han': (
        0x2e8000002e9a,
        0x2e9b00002ef4,
        0x2f0000002fd6,
        0x300500003006,
        0x300700003008,
        0x30210000302a,
        0x30380000303c,
        0x340000004dc0,
        0x4e0000009ffd,
        0xf9000000fa6e,
        0xfa700000fada,
        0x16ff000016ff2,
        0x200000002a6de,
        0x2a7000002b735,
        0x2b7400002b81e,
        0x2b8200002cea2,
        0x2ceb00002ebe1,
        0x2f8000002fa1e,
        0x300000003134b,
    ),
    'Hebrew': (
        0x591000005c8,
        0x5d0000005eb,
        0x5ef000005f5,
        0xfb1d0000fb37,
        0xfb380000fb3d,
        0xfb3e0000fb3f,
        0xfb400000fb42,
        0xfb430000fb45,
        0xfb460000fb50,
    ),
    'Hiragana': (
        0x304100003097,
        0x309d000030a0,
        0x1b0010001b11f,
        0x1b1500001b153,
        0x1f2000001f201,
    ),
    'Katakana': (
        0x30a1000030fb,
        0x30fd00003100,
        0x31f000003200,
        0x32d0000032ff,
        0x330000003358,
        0xff660000ff70,
        0xff710000ff9e,
        0x1b0000001b001,
        0x1b1640001b168,
    ),
}
joining_types = {
    0x600: 85,
    0x601: 85,
    0x602: 85,
    0x603: 85,
    0x604: 85,
    0x605: 85,
    0x608: 85,
    0x60b: 85,
    0x620: 68,
    0x621: 85,
    0x622: 82,
    0x623: 82,
    0x624: 82,
    0x625: 82,
    0x626: 68,
    0x627: 82,
    0x628: 68,
    0x629: 82,
    0x62a: 68,
    0x62b: 68,
    0x62c: 68,
    0x62d: 68,
    0x62e: 68,
    0x62f: 82,
    0x630: 82,
    0x631: 82,
    0x632: 82,
    0x633: 68,
    0x634: 68,
    0x635: 68,
    0x636: 68,
    0x637: 68,
    0x638: 68,
    0x639: 68,
    0x63a: 68,
    0x63b: 68,
    0x63c: 68,
    0x63d: 68,
    0x63e: 68,
    0x63f: 68,
    0x640: 67,
    0x641: 68,
    0x642: 68,
    0x643: 68,
    0x644: 68,
    0x645: 68,
    0x646: 68,
    0x647: 68,
    0x648: 82,
    0x649: 68,
    0x64a: 68,
    0x66e: 68,
    0x66f: 68,
    0x671: 82,
    0x672: 82,
    0x673: 82,
    0x674: 85,
    0x675: 82,
    0x676: 82,
    0x677: 82,
    0x678: 68,
    0x679: 68,
    0x67a: 68,
    0x67b: 68,
    0x67c: 68,
    0x67d: 68,
    0x67e: 68,
    0x67f: 68,
    0x680: 68,
    0x681: 68,
    0x682: 68,
    0x683: 68,
    0x684: 68,
    0x685: 68,
    0x686: 68,
    0x687: 68,
    0x688: 82,
    0x689: 82,
    0x68a: 82,
    0x68b: 82,
    0x68c: 82,
    0x68d: 82,
    0x68e: 82,
    0x68f: 82,
    0x690: 82,
    0x691: 82,
    0x692: 82,
    0x693: 82,
    0x694: 82,
    0x695: 82,
    0x696: 82,
    0x697: 82,
    0x698: 82,
    0x699: 82,
    0x69a: 68,
    0x69b: 68,
    0x69c: 68,
    0x69d: 68,
    0x69e: 68,
    0x69f: 68,
    0x6a0: 68,
    0x6a1: 68,
    0x6a2: 68,
    0x6a3: 68,
    0x6a4: 68,
    0x6a5: 68,
    0x6a6: 68,
    0x6a7: 68,
    0x6a8: 68,
    0x6a9: 68,
    0x6aa: 68,
    0x6ab: 68,
    0x6ac: 68,
    0x6ad: 68,
    0x6ae: 68,
    0x6af: 68,
    0x6b0: 68,
    0x6b1: 68,
    0x6b2: 68,
    0x6b3: 68,
    0x6b4: 68,
    0x6b5: 68,
    0x6b6: 68,
    0x6b7: 68,
    0x6b8: 68,
    0x6b9: 68,
    0x6ba: 68,
    0x6bb: 68,
    0x6bc: 68,
    0x6bd: 68,
    0x6be: 68,
    0x6bf: 68,
    0x6c0: 82,
    0x6c1: 68,
    0x6c2: 68,
    0x6c3: 82,
    0x6c4: 82,
    0x6c5: 82,
    0x6c6: 82,
    0x6c7: 82,
    0x6c8: 82,
    0x6c9: 82,
    0x6ca: 82,
    0x6cb: 82,
    0x6cc: 68,
    0x6cd: 82,
    0x6ce: 68,
    0x6cf: 82,
    0x6d0: 68,
    0x6d1: 68,
    0x6d2: 82,
    0x6d3: 82,
    0x6d5: 82,
    0x6dd: 85,
    0x6ee: 82,
    0x6ef: 82,
    0x6fa: 68,
    0x6fb: 68,
    0x6fc: 68,
    0x6ff: 68,
    0x70f: 84,
    0x710: 82,
    0x712: 68,
    0x713: 68,
    0x714: 68,
    0x715: 82,
    0x716: 82,
    0x717: 82,
    0x718: 82,
    0x719: 82,
    0x71a: 68,
    0x71b: 68,
    0x71c: 68,
    0x71d: 68,
    0x71e: 82,
    0x71f: 68,
    0x720: 68,
    0x721: 68,
    0x722: 68,
    0x723: 68,
    0x724: 68,
    0x725: 68,
    0x726: 68,
    0x727: 68,
    0x728: 82,
    0x729: 68,
    0x72a: 82,
    0x72b: 68,
    0x72c: 82,
    0x72d: 68,
    0x72e: 68,
    0x72f: 82,
    0x74d: 82,
    0x74e: 68,
    0x74f: 68,
    0x750: 68,
    0x751: 68,
    0x752: 68,
    0x753: 68,
    0x754: 68,
    0x755: 68,
    0x756: 68,
    0x757: 68,
    0x758: 68,
    0x759: 82,
    0x75a: 82,
    0x75b: 82,
    0x75c: 68,
    0x75d: 68,
    0x75e: 68,
    0x75f: 68,
    0x760: 68,
    0x761: 68,
    0x762: 68,
    0x763: 68,
    0x764: 68,
    0x765: 68,
    0x766: 68,
    0x767: 68,
    0x768: 68,
    0x769: 68,
    0x76a: 68,
    0x76b: 82,
    0x76c: 82,
    0x76d: 68,
    0x76e: 68,
    0x76f: 68,
    0x770: 68,
    0x771: 82,
    0x772: 68,
    0x773: 82,
    0x774: 82,
    0x775: 68,
    0x776: 68,
    0x777: 68,
    0x778: 82,
    0x779: 82,
    0x77a: 68,
    0x77b: 68,
    0x77c: 68,
    0x77d: 68,
    0x77e: 68,
    0x77f: 68,
    0x7ca: 68,
    0x7cb: 68,
    0x7cc: 68,
    0x7cd: 68,
    0x7ce: 68,
    0x7cf: 68,
    0x7d0: 68,
    0x7d1: 68,
    0x7d2: 68,
    0x7d3: 68,
    0x7d4: 68,
    0x7d5: 68,
    0x7d6: 68,
    0x7d7: 68,
    0x7d8: 68,
    0x7d9: 68,
    0x7da: 68,
    0x7db: 68,
    0x7dc: 68,
    0x7dd: 68,
    0x7de: 68,
    0x7df: 68,
    0x7e0: 68,
    0x7e1: 68,
    0x7e2: 68,
    0x7e3: 68,
    0x7e4: 68,
    0x7e5: 68,
    0x7e6: 68,
    0x7e7: 68,
    0x7e8: 68,
    0x7e9: 68,
    0x7ea: 68,
    0x7fa: 67,
    0x840: 82,
    0x841: 68,
    0x842: 68,
    0x843: 68,
    0x844: 68,
    0x845: 68,
    0x846: 82,
    0x847: 82,
    0x848: 68,
    0x849: 82,
    0x84a: 68,
    0x84b: 68,
    0x84c: 68,
    0x84d: 68,
    0x84e: 68,
    0x84f: 68,
    0x850: 68,
    0x851: 68,
    0x852: 68,
    0x853: 68,
    0x854: 82,
    0x855: 68,
    0x856: 82,
    0x857: 82,
    0x858: 82,
    0x860: 68,
    0x861: 85,
    0x862: 68,
    0x863: 68,
    0x864: 68,
    0x865: 68,
    0x866: 85,
    0x867: 82,
    0x868: 68,
    0x869: 82,
    0x86a: 82,
    0x8a0: 68,
    0x8a1: 68,
    0x8a2: 68,
    0x8a3: 68,
    0x8a4: 68,
    0x8a5: 68,
    0x8a6: 68,
    0x8a7: 68,
    0x8a8: 68,
    0x8a9: 68,
    0x8aa: 82,
    0x8ab: 82,
    0x8ac: 82,
    0x8ad: 85,
    0x8ae: 82,
    0x8af: 68,
    0x8b0: 68,
    0x8b1: 82,
    0x8b2: 82,
    0x8b3: 68,
    0x8b4: 68,
    0x8b6: 68,
    0x8b7: 68,
    0x8b8: 68,
    0x8b9: 82,
    0x8ba: 68,
    0x8bb: 68,
    0x8bc: 68,
    0x8bd: 68,
    0x8be: 68,
    0x8bf: 68,
    0x8c0: 68,
    0x8c1: 68,
    0x8c2: 68,
    0x8c3: 68,
    0x8c4: 68,
    0x8c5: 68,
    0x8c6: 68,
    0x8c7: 68,
    0x8e2: 85,
    0x1806: 85,
    0x1807: 68,
    0x180a: 67,
    0x180e: 85,
    0x1820: 68,
    0x1821: 68,
    0x1822: 68,
    0x1823: 68,
    0x1824: 68,
    0x1825: 68,
    0x1826: 68,
    0x1827: 68,
    0x1828: 68,
    0x1829: 68,
    0x182a: 68,
    0x182b: 68,
    0x182c: 68,
    0x182d: 68,
    0x182e: 68,
    0x182f: 68,
    0x1830: 68,
    0x1831: 68,
    0x1832: 68,
    0x1833: 68,
    0x1834: 68,
    0x1835: 68,
    0x1836: 68,
    0x1837: 68,
    0x1838: 68,
    0x1839: 68,
    0x183a: 68,
    0x183b: 68,
    0x183c: 68,
    0x183d: 68,
    0x183e: 68,
    0x183f: 68,
    0x1840: 68,
    0x1841: 68,
    0x1842: 68,
    0x1843: 68,
    0x1844: 68,
    0x1845: 68,
    0x1846: 68,
    0x1847: 68,
    0x1848: 68,
    0x1849: 68,
    0x184a: 68,
    0x184b: 68,
    0x184c: 68,
    0x184d: 68,
    0x184e: 68,
    0x184f: 68,
    0x1850: 68,
    0x1851: 68,
    0x1852: 68,
    0x1853: 68,
    0x1854: 68,
    0x1855: 68,
    0x1856: 68,
    0x1857: 68,
    0x1858: 68,
    0x1859: 68,
    0x185a: 68,
    0x185b: 68,
    0x185c: 68,
    0x185d: 68,
    0x185e: 68,
    0x185f: 68,
    0x1860: 68,
    0x1861: 68,
    0x1862: 68,
    0x1863: 68,
    0x1864: 68,
    0x1865: 68,
    0x1866: 68,
    0x1867: 68,
    0x1868: 68,
    0x1869: 68,
    0x186a: 68,
    0x186b: 68,
    0x186c: 68,
    0x186d: 68,
    0x186e: 68,
    0x186f: 68,
    0x1870: 68,
    0x1871: 68,
    0x1872: 68,
    0x1873: 68,
    0x1874: 68,
    0x1875: 68,
    0x1876: 68,
    0x1877: 68,
    0x1878: 68,
    0x1880: 85,
    0x1881: 85,
    0x1882: 85,
    0x1883: 85,
    0x1884: 85,
    0x1885: 84,
    0x1886: 84,
    0x1887: 68,
    0x1888: 68,
    0x1889: 68,
    0x188a: 68,
    0x188b: 68,
    0x188c: 68,
    0x188d: 68,
    0x188e: 68,
    0x188f: 68,
    0x1890: 68,
    0x1891: 68,
    0x1892: 68,
    0x1893: 68,
    0x1894: 68,
    0x1895: 68,
    0x1896: 68,
    0x1897: 68,
    0x1898: 68,
    0x1899: 68,
    0x189a: 68,
    0x189b: 68,
    0x189c: 68,
    0x189d: 68,
    0x189e: 68,
    0x189f: 68,
    0x18a0: 68,
    0x18a1: 68,
    0x18a2: 68,
    0x18a3: 68,
    0x18a4: 68,
    0x18a5: 68,
    0x18a6: 68,
    0x18a7: 68,
    0x18a8: 68,
    0x18aa: 68,
    0x200c: 85,
    0x200d: 67,
    0x202f: 85,
    0x2066: 85,
    0x2067: 85,
    0x2068: 85,
    0x2069: 85,
    0xa840: 68,
    0xa841: 68,
    0xa842: 68,
    0xa843: 68,
    0xa844: 68,
    0xa845: 68,
    0xa846: 68,
    0xa847: 68,
    0xa848: 68,
    0xa849: 68,
    0xa84a: 68,
    0xa84b: 68,
    0xa84c: 68,
    0xa84d: 68,
    0xa84e: 68,
    0xa84f: 68,
    0xa850: 68,
    0xa851: 68,
    0xa852: 68,
    0xa853: 68,
    0xa854: 68,
    0xa855: 68,
    0xa856: 68,
    0xa857: 68,
    0xa858: 68,
    0xa859: 68,
    0xa85a: 68,
    0xa85b: 68,
    0xa85c: 68,
    0xa85d: 68,
    0xa85e: 68,
    0xa85f: 68,
    0xa860: 68,
    0xa861: 68,
    0xa862: 68,
    0xa863: 68,
    0xa864: 68,
    0xa865: 68,
    0xa866: 68,
    0xa867: 68,
    0xa868: 68,
    0xa869: 68,
    0xa86a: 68,
    0xa86b: 68,
    0xa86c: 68,
    0xa86d: 68,
    0xa86e: 68,
    0xa86f: 68,
    0xa870: 68,
    0xa871: 68,
    0xa872: 76,
    0xa873: 85,
    0x10ac0: 68,
    0x10ac1: 68,
    0x10ac2: 68,
    0x10ac3: 68,
    0x10ac4: 68,
    0x10ac5: 82,
    0x10ac6: 85,
    0x10ac7: 82,
    0x10ac8: 85,
    0x10ac9: 82,
    0x10aca: 82,
    0x10acb: 85,
    0x10acc: 85,
    0x10acd: 76,
    0x10ace: 82,
    0x10acf: 82,
    0x10ad0: 82,
    0x10ad1: 82,
    0x10ad2: 82,
    0x10ad3: 68,
    0x10ad4: 68,
    0x10ad5: 68,
    0x10ad6: 68,
    0x10ad7: 76,
    0x10ad8: 68,
    0x10ad9: 68,
    0x10ada: 68,
    0x10adb: 68,
    0x10adc: 68,
    0x10add: 82,
    0x10ade: 68,
    0x10adf: 68,
    0x10ae0: 68,
    0x10ae1: 82,
    0x10ae2: 85,
    0x10ae3: 85,
    0x10ae4: 82,
    0x10aeb: 68,
    0x10aec: 68,
    0x10aed: 68,
    0x10aee: 68,
    0x10aef: 82,
    0x10b80: 68,
    0x10b81: 82,
    0x10b82: 68,
    0x10b83: 82,
    0x10b84: 82,
    0x10b85: 82,
    0x10b86: 68,
    0x10b87: 68,
    0x10b88: 68,
    0x10b89: 82,
    0x10b8a: 68,
    0x10b8b: 68,
    0x10b8c: 82,
    0x10b8d: 68,
    0x10b8e: 82,
    0x10b8f: 82,
    0x10b90: 68,
    0x10b91: 82,
    0x10ba9: 82,
    0x10baa: 82,
    0x10bab: 82,
    0x10bac: 82,
    0x10bad: 68,
    0x10bae: 68,
    0x10baf: 85,
    0x10d00: 76,
    0x10d01: 68,
    0x10d02: 68,
    0x10d03: 68,
    0x10d04: 68,
    0x10d05: 68,
    0x10d06: 68,
    0x10d07: 68,
    0x10d08: 68,
    0x10d09: 68,
    0x10d0a: 68,
    0x10d0b: 68,
    0x10d0c: 68,
    0x10d0d: 68,
    0x10d0e: 68,
    0x10d0f: 68,
    0x10d10: 68,
    0x10d11: 68,
    0x10d12: 68,
    0x10d13: 68,
    0x10d14: 68,
    0x10d15: 68,
    0x10d16: 68,
    0x10d17: 68,
    0x10d18: 68,
    0x10d19: 68,
    0x10d1a: 68,
    0x10d1b: 68,
    0x10d1c: 68,
    0x10d1d: 68,
    0x10d1e: 68,
    0x10d1f: 68,
    0x10d20: 68,
    0x10d21: 68,
    0x10d22: 82,
    0x10d23: 68,
    0x10f30: 68,
    0x10f31: 68,
    0x10f32: 68,
    0x10f33: 82,
    0x10f34: 68,
    0x10f35: 68,
    0x10f36: 68,
    0x10f37: 68,
    0x10f38: 68,
    0x10f39: 68,
    0x10f3a: 68,
    0x10f3b: 68,
    0x10f3c: 68,
    0x10f3d: 68,
    0x10f3e: 68,
    0x10f3f: 68,
    0x10f40: 68,
    0x10f41: 68,
    0x10f42: 68,
    0x10f43: 68,
    0x10f44: 68,
    0x10f45: 85,
    0x10f51: 68,
    0x10f52: 68,
    0x10f53: 68,
    0x10f54: 82,
    0x10fb0: 68,
    0x10fb1: 85,
    0x10fb2: 68,
    0x10fb3: 68,
    0x10fb4: 82,
    0x10fb5: 82,
    0x10fb6: 82,
    0x10fb7: 85,
    0x10fb8: 68,
    0x10fb9: 82,
    0x10fba: 82,
    0x10fbb: 68,
    0x10fbc: 68,
    0x10fbd: 82,
    0x10fbe: 68,
    0x10fbf: 68,
    0x10fc0: 85,
    0x10fc1: 68,
    0x10fc2: 82,
    0x10fc3: 82,
    0x10fc4: 68,
    0x10fc5: 85,
    0x10fc6: 85,
    0x10fc7: 85,
    0x10fc8: 85,
    0x10fc9: 82,
    0x10fca: 68,
    0x10fcb: 76,
    0x110bd: 85,
    0x110cd: 85,
    0x1e900: 68,
    0x1e901: 68,
    0x1e902: 68,
    0x1e903: 68,
    0x1e904: 68,
    0x1e905: 68,
    0x1e906: 68,
    0x1e907: 68,
    0x1e908: 68,
    0x1e909: 68,
    0x1e90a: 68,
    0x1e90b: 68,
    0x1e90c: 68,
    0x1e90d: 68,
    0x1e90e: 68,
    0x1e90f: 68,
    0x1e910: 68,
    0x1e911: 68,
    0x1e912: 68,
    0x1e913: 68,
    0x1e914: 68,
    0x1e915: 68,
    0x1e916: 68,
    0x1e917: 68,
    0x1e918: 68,
    0x1e919: 68,
    0x1e91a: 68,
    0x1e91b: 68,
    0x1e91c: 68,
    0x1e91d: 68,
    0x1e91e: 68,
    0x1e91f: 68,
    0x1e920: 68,
    0x1e921: 68,
    0x1e922: 68,
    0x1e923: 68,
    0x1e924: 68,
    0x1e925: 68,
    0x1e926: 68,
    0x1e927: 68,
    0x1e928: 68,
    0x1e929: 68,
    0x1e92a: 68,
    0x1e92b: 68,
    0x1e92c: 68,
    0x1e92d: 68,
    0x1e92e: 68,
    0x1e92f: 68,
    0x1e930: 68,
    0x1e931: 68,
    0x1e932: 68,
    0x1e933: 68,
    0x1e934: 68,
    0x1e935: 68,
    0x1e936: 68,
    0x1e937: 68,
    0x1e938: 68,
    0x1e939: 68,
    0x1e93a: 68,
    0x1e93b: 68,
    0x1e93c: 68,
    0x1e93d: 68,
    0x1e93e: 68,
    0x1e93f: 68,
    0x1e940: 68,
    0x1e941: 68,
    0x1e942: 68,
    0x1e943: 68,
    0x1e94b: 84,
}
codepoint_classes = {
    'PVALID': (
        0x2d0000002e,
        0x300000003a,
        0x610000007b,
        0xdf000000f7,
        0xf800000100,
        0x10100000102,
        0x10300000104,
        0x10500000106,
        0x10700000108,
        0x1090000010a,
        0x10b0000010c,
        0x10d0000010e,
        0x10f00000110,
        0x11100000112,
        0x11300000114,
        0x11500000116,
        0x11700000118,
        0x1190000011a,
        0x11b0000011c,
        0x11d0000011e,
        0x11f00000120,
        0x12100000122,
        0x12300000124,
        0x12500000126,
        0x12700000128,
        0x1290000012a,
        0x12b0000012c,
        0x12d0000012e,
        0x12f00000130,
        0x13100000132,
        0x13500000136,
        0x13700000139,
        0x13a0000013b,
        0x13c0000013d,
        0x13e0000013f,
        0x14200000143,
        0x14400000145,
        0x14600000147,
        0x14800000149,
        0x14b0000014c,
        0x14d0000014e,
        0x14f00000150,
        0x15100000152,
        0x15300000154,
        0x15500000156,
        0x15700000158,
        0x1590000015a,
        0x15b0000015c,
        0x15d0000015e,
        0x15f00000160,
        0x16100000162,
        0x16300000164,
        0x16500000166,
        0x16700000168,
        0x1690000016a,
        0x16b0000016c,
        0x16d0000016e,
        0x16f00000170,
        0x17100000172,
        0x17300000174,
        0x17500000176,
        0x17700000178,
        0x17a0000017b,
        0x17c0000017d,
        0x17e0000017f,
        0x18000000181,
        0x18300000184,
        0x18500000186,
        0x18800000189,
        0x18c0000018e,
        0x19200000193,
        0x19500000196,
        0x1990000019c,
        0x19e0000019f,
        0x1a1000001a2,
        0x1a3000001a4,
        0x1a5000001a6,
        0x1a8000001a9,
        0x1aa000001ac,
        0x1ad000001ae,
        0x1b0000001b1,
        0x1b4000001b5,
        0x1b6000001b7,
        0x1b9000001bc,
        0x1bd000001c4,
        0x1ce000001cf,
        0x1d0000001d1,
        0x1d2000001d3,
        0x1d4000001d5,
        0x1d6000001d7,
        0x1d8000001d9,
        0x1da000001db,
        0x1dc000001de,
        0x1df000001e0,
        0x1e1000001e2,
        0x1e3000001e4,
        0x1e5000001e6,
        0x1e7000001e8,
        0x1e9000001ea,
        0x1eb000001ec,
        0x1ed000001ee,
        0x1ef000001f1,
        0x1f5000001f6,
        0x1f9000001fa,
        0x1fb000001fc,
        0x1fd000001fe,
        0x1ff00000200,
        0x20100000202,
        0x20300000204,
        0x20500000206,
        0x20700000208,
        0x2090000020a,
        0x20b0000020c,
        0x20d0000020e,
        0x20f00000210,
        0x21100000212,
        0x21300000214,
        0x21500000216,
        0x21700000218,
        0x2190000021a,
        0x21b0000021c,
        0x21d0000021e,
        0x21f00000220,
        0x22100000222,
        0x22300000224,
        0x22500000226,
        0x22700000228,
        0x2290000022a,
        0x22b0000022c,
        0x22d0000022e,
        0x22f00000230,
        0x23100000232,
        0x2330000023a,
        0x23c0000023d,
        0x23f00000241,
        0x24200000243,
        0x24700000248,
        0x2490000024a,
        0x24b0000024c,
        0x24d0000024e,
        0x24f000002b0,
        0x2b9000002c2,
        0x2c6000002d2,
        0x2ec000002ed,
        0x2ee000002ef,
        0x30000000340,
        0x34200000343,
        0x3460000034f,
        0x35000000370,
        0x37100000372,
        0x37300000374,
        0x37700000378,
        0x37b0000037e,
        0x39000000391,
        0x3ac000003cf,
        0x3d7000003d8,
        0x3d9000003da,
        0x3db000003dc,
        0x3dd000003de,
        0x3df000003e0,
        0x3e1000003e2,
        0x3e3000003e4,
        0x3e5000003e6,
        0x3e7000003e8,
        0x3e9000003ea,
        0x3eb000003ec,
        0x3ed000003ee,
        0x3ef000003f0,
        0x3f3000003f4,
        0x3f8000003f9,
        0x3fb000003fd,
        0x43000000460,
        0x46100000462,
        0x46300000464,
        0x46500000466,
        0x46700000468,
        0x4690000046a,
        0x46b0000046c,
        0x46d0000046e,
        0x46f00000470,
        0x47100000472,
        0x47300000474,
        0x47500000476,
        0x47700000478,
        0x4790000047a,
        0x47b0000047c,
        0x47d0000047e,
        0x47f00000480,
        0x48100000482,
        0x48300000488,
        0x48b0000048c,
        0x48d0000048e,
        0x48f00000490,
        0x49100000492,
        0x49300000494,
        0x49500000496,
        0x49700000498,
        0x4990000049a,
        0x49b0000049c,
        0x49d0000049e,
        0x49f000004a0,
        0x4a1000004a2,
        0x4a3000004a4,
        0x4a5000004a6,
        0x4a7000004a8,
        0x4a9000004aa,
        0x4ab000004ac,
        0x4ad000004ae,
        0x4af000004b0,
        0x4b1000004b2,
        0x4b3000004b4,
        0x4b5000004b6,
        0x4b7000004b8,
        0x4b9000004ba,
        0x4bb000004bc,
        0x4bd000004be,
        0x4bf000004c0,
        0x4c2000004c3,
        0x4c4000004c5,
        0x4c6000004c7,
        0x4c8000004c9,
        0x4ca000004cb,
        0x4cc000004cd,
        0x4ce000004d0,
        0x4d1000004d2,
        0x4d3000004d4,
        0x4d5000004d6,
        0x4d7000004d8,
        0x4d9000004da,
        0x4db000004dc,
        0x4dd000004de,
        0x4df000004e0,
        0x4e1000004e2,
        0x4e3000004e4,
        0x4e5000004e6,
        0x4e7000004e8,
        0x4e9000004ea,
        0x4eb000004ec,
        0x4ed000004ee,
        0x4ef000004f0,
        0x4f1000004f2,
        0x4f3000004f4,
        0x4f5000004f6,
        0x4f7000004f8,
        0x4f9000004fa,
        0x4fb000004fc,
        0x4fd000004fe,
        0x4ff00000500,
        0x50100000502,
        0x50300000504,
        0x50500000506,
        0x50700000508,
        0x5090000050a,
        0x50b0000050c,
        0x50d0000050e,
        0x50f00000510,
        0x51100000512,
        0x51300000514,
        0x51500000516,
        0x51700000518,
        0x5190000051a,
        0x51b0000051c,
        0x51d0000051e,
        0x51f00000520,
        0x52100000522,
        0x52300000524,
        0x52500000526,
        0x52700000528,
        0x5290000052a,
        0x52b0000052c,
        0x52d0000052e,
        0x52f00000530,
        0x5590000055a,
        0x56000000587,
        0x58800000589,
        0x591000005be,
        0x5bf000005c0,
        0x5c1000005c3,
        0x5c4000005c6,
        0x5c7000005c8,
        0x5d0000005eb,
        0x5ef000005f3,
        0x6100000061b,
        0x62000000640,
        0x64100000660,
        0x66e00000675,
        0x679000006d4,
        0x6d5000006dd,
        0x6df000006e9,
        0x6ea000006f0,
        0x6fa00000700,
        0x7100000074b,
        0x74d000007b2,
        0x7c0000007f6,
        0x7fd000007fe,
        0x8000000082e,
        0x8400000085c,
        0x8600000086b,
        0x8a0000008b5,
        0x8b6000008c8,
        0x8d3000008e2,
        0x8e300000958,
        0x96000000964,
        0x96600000970,
        0x97100000984,
        0x9850000098d,
        0x98f00000991,
        0x993000009a9,
        0x9aa000009b1,
        0x9b2000009b3,
        0x9b6000009ba,
        0x9bc000009c5,
        0x9c7000009c9,
        0x9cb000009cf,
        0x9d7000009d8,
        0x9e0000009e4,
        0x9e6000009f2,
        0x9fc000009fd,
        0x9fe000009ff,
        0xa0100000a04,
        0xa0500000a0b,
        0xa0f00000a11,
        0xa1300000a29,
        0xa2a00000a31,
        0xa3200000a33,
        0xa3500000a36,
        0xa3800000a3a,
        0xa3c00000a3d,
        0xa3e00000a43,
        0xa4700000a49,
        0xa4b00000a4e,
        0xa5100000a52,
        0xa5c00000a5d,
        0xa6600000a76,
        0xa8100000a84,
        0xa8500000a8e,
        0xa8f00000a92,
        0xa9300000aa9,
        0xaaa00000ab1,
        0xab200000ab4,
        0xab500000aba,
        0xabc00000ac6,
        0xac700000aca,
        0xacb00000ace,
        0xad000000ad1,
        0xae000000ae4,
        0xae600000af0,
        0xaf900000b00,
        0xb0100000b04,
        0xb0500000b0d,
        0xb0f00000b11,
        0xb1300000b29,
        0xb2a00000b31,
        0xb3200000b34,
        0xb3500000b3a,
        0xb3c00000b45,
        0xb4700000b49,
        0xb4b00000b4e,
        0xb5500000b58,
        0xb5f00000b64,
        0xb6600000b70,
        0xb7100000b72,
        0xb8200000b84,
        0xb8500000b8b,
        0xb8e00000b91,
        0xb9200000b96,
        0xb9900000b9b,
        0xb9c00000b9d,
        0xb9e00000ba0,
        0xba300000ba5,
        0xba800000bab,
        0xbae00000bba,
        0xbbe00000bc3,
        0xbc600000bc9,
        0xbca00000bce,
        0xbd000000bd1,
        0xbd700000bd8,
        0xbe600000bf0,
        0xc0000000c0d,
        0xc0e00000c11,
        0xc1200000c29,
        0xc2a00000c3a,
        0xc3d00000c45,
        0xc4600000c49,
        0xc4a00000c4e,
        0xc5500000c57,
        0xc5800000c5b,
        0xc6000000c64,
        0xc6600000c70,
        0xc8000000c84,
        0xc8500000c8d,
        0xc8e00000c91,
        0xc9200000ca9,
        0xcaa00000cb4,
        0xcb500000cba,
        0xcbc00000cc5,
        0xcc600000cc9,
        0xcca00000cce,
        0xcd500000cd7,
        0xcde00000cdf,
        0xce000000ce4,
        0xce600000cf0,
        0xcf100000cf3,
        0xd0000000d0d,
        0xd0e00000d11,
        0xd1200000d45,
        0xd4600000d49,
        0xd4a00000d4f,
        0xd5400000d58,
        0xd5f00000d64,
        0xd6600000d70,
        0xd7a00000d80,
        0xd8100000d84,
        0xd8500000d97,
        0xd9a00000db2,
        0xdb300000dbc,
        0xdbd00000dbe,
        0xdc000000dc7,
        0xdca00000dcb,
        0xdcf00000dd5,
        0xdd600000dd7,
        0xdd800000de0,
        0xde600000df0,
        0xdf200000df4,
        0xe0100000e33,
        0xe3400000e3b,
        0xe4000000e4f,
        0xe5000000e5a,
        0xe8100000e83,
        0xe8400000e85,
        0xe8600000e8b,
        0xe8c00000ea4,
        0xea500000ea6,
        0xea700000eb3,
        0xeb400000ebe,
        0xec000000ec5,
        0xec600000ec7,
        0xec800000ece,
        0xed000000eda,
        0xede00000ee0,
        0xf0000000f01,
        0xf0b00000f0c,
        0xf1800000f1a,
        0xf2000000f2a,
        0xf3500000f36,
        0xf3700000f38,
        0xf3900000f3a,
        0xf3e00000f43,
        0xf4400000f48,
        0xf4900000f4d,
        0xf4e00000f52,
        0xf5300000f57,
        0xf5800000f5c,
        0xf5d00000f69,
        0xf6a00000f6d,
        0xf7100000f73,
        0xf7400000f75,
        0xf7a00000f81,
        0xf8200000f85,
        0xf8600000f93,
        0xf9400000f98,
        0xf9900000f9d,
        0xf9e00000fa2,
        0xfa300000fa7,
        0xfa800000fac,
        0xfad00000fb9,
        0xfba00000fbd,
        0xfc600000fc7,
        0x10000000104a,
        0x10500000109e,
        0x10d0000010fb,
        0x10fd00001100,
        0x120000001249,
        0x124a0000124e,
        0x125000001257,
        0x125800001259,
        0x125a0000125e,
        0x126000001289,
        0x128a0000128e,
        0x1290000012b1,
        0x12b2000012b6,
        0x12b8000012bf,
        0x12c0000012c1,
        0x12c2000012c6,
        0x12c8000012d7,
        0x12d800001311,
        0x131200001316,
        0x13180000135b,
        0x135d00001360,
        0x138000001390,
        0x13a0000013f6,
        0x14010000166d,
        0x166f00001680,
        0x16810000169b,
        0x16a0000016eb,
        0x16f1000016f9,
        0x17000000170d,
        0x170e00001715,
        0x172000001735,
        0x174000001754,
        0x17600000176d,
        0x176e00001771,
        0x177200001774,
        0x1780000017b4,
        0x17b6000017d4,
        0x17d7000017d8,
        0x17dc000017de,
        0x17e0000017ea,
        0x18100000181a,
        0x182000001879,
        0x1880000018ab,
        0x18b0000018f6,
        0x19000000191f,
        0x19200000192c,
        0x19300000193c,
        0x19460000196e,
        0x197000001975,
        0x1980000019ac,
        0x19b0000019ca,
        0x19d0000019da,
        0x1a0000001a1c,
        0x1a2000001a5f,
        0x1a6000001a7d,
        0x1a7f00001a8a,
        0x1a9000001a9a,
        0x1aa700001aa8,
        0x1ab000001abe,
        0x1abf00001ac1,
        0x1b0000001b4c,
        0x1b5000001b5a,
        0x1b6b00001b74,
        0x1b8000001bf4,
        0x1c0000001c38,
        0x1c4000001c4a,
        0x1c4d00001c7e,
        0x1cd000001cd3,
        0x1cd400001cfb,
        0x1d0000001d2c,
        0x1d2f00001d30,
        0x1d3b00001d3c,
        0x1d4e00001d4f,
        0x1d6b00001d78,
        0x1d7900001d9b,
        0x1dc000001dfa,
        0x1dfb00001e00,
        0x1e0100001e02,
        0x1e0300001e04,
        0x1e0500001e06,
        0x1e0700001e08,
        0x1e0900001e0a,
        0x1e0b00001e0c,
        0x1e0d00001e0e,
        0x1e0f00001e10,
        0x1e1100001e12,
        0x1e1300001e14,
        0x1e1500001e16,
        0x1e1700001e18,
        0x1e1900001e1a,
        0x1e1b00001e1c,
        0x1e1d00001e1e,
        0x1e1f00001e20,
        0x1e2100001e22,
        0x1e2300001e24,
        0x1e2500001e26,
        0x1e2700001e28,
        0x1e2900001e2a,
        0x1e2b00001e2c,
        0x1e2d00001e2e,
        0x1e2f00001e30,
        0x1e3100001e32,
        0x1e3300001e34,
        0x1e3500001e36,
        0x1e3700001e38,
        0x1e3900001e3a,
        0x1e3b00001e3c,
        0x1e3d00001e3e,
        0x1e3f00001e40,
        0x1e4100001e42,
        0x1e4300001e44,
        0x1e4500001e46,
        0x1e4700001e48,
        0x1e4900001e4a,
        0x1e4b00001e4c,
        0x1e4d00001e4e,
        0x1e4f00001e50,
        0x1e5100001e52,
        0x1e5300001e54,
        0x1e5500001e56,
        0x1e5700001e58,
        0x1e5900001e5a,
        0x1e5b00001e5c,
        0x1e5d00001e5e,
        0x1e5f00001e60,
        0x1e6100001e62,
        0x1e6300001e64,
        0x1e6500001e66,
        0x1e6700001e68,
        0x1e6900001e6a,
        0x1e6b00001e6c,
        0x1e6d00001e6e,
        0x1e6f00001e70,
        0x1e7100001e72,
        0x1e7300001e74,
        0x1e7500001e76,
        0x1e7700001e78,
        0x1e7900001e7a,
        0x1e7b00001e7c,
        0x1e7d00001e7e,
        0x1e7f00001e80,
        0x1e8100001e82,
        0x1e8300001e84,
        0x1e8500001e86,
        0x1e8700001e88,
        0x1e8900001e8a,
        0x1e8b00001e8c,
        0x1e8d00001e8e,
        0x1e8f00001e90,
        0x1e9100001e92,
        0x1e9300001e94,
        0x1e9500001e9a,
        0x1e9c00001e9e,
        0x1e9f00001ea0,
        0x1ea100001ea2,
        0x1ea300001ea4,
        0x1ea500001ea6,
        0x1ea700001ea8,
        0x1ea900001eaa,
        0x1eab00001eac,
        0x1ead00001eae,
        0x1eaf00001eb0,
        0x1eb100001eb2,
        0x1eb300001eb4,
        0x1eb500001eb6,
        0x1eb700001eb8,
        0x1eb900001eba,
        0x1ebb00001ebc,
        0x1ebd00001ebe,
        0x1ebf00001ec0,
        0x1ec100001ec2,
        0x1ec300001ec4,
        0x1ec500001ec6,
        0x1ec700001ec8,
        0x1ec900001eca,
        0x1ecb00001ecc,
        0x1ecd00001ece,
        0x1ecf00001ed0,
        0x1ed100001ed2,
        0x1ed300001ed4,
        0x1ed500001ed6,
        0x1ed700001ed8,
        0x1ed900001eda,
        0x1edb00001edc,
        0x1edd00001ede,
        0x1edf00001ee0,
        0x1ee100001ee2,
        0x1ee300001ee4,
        0x1ee500001ee6,
        0x1ee700001ee8,
        0x1ee900001eea,
        0x1eeb00001eec,
        0x1eed00001eee,
        0x1eef00001ef0,
        0x1ef100001ef2,
        0x1ef300001ef4,
        0x1ef500001ef6,
        0x1ef700001ef8,
        0x1ef900001efa,
        0x1efb00001efc,
        0x1efd00001efe,
        0x1eff00001f08,
        0x1f1000001f16,
        0x1f2000001f28,
        0x1f3000001f38,
        0x1f4000001f46,
        0x1f5000001f58,
        0x1f6000001f68,
        0x1f7000001f71,
        0x1f7200001f73,
        0x1f7400001f75,
        0x1f7600001f77,
        0x1f7800001f79,
        0x1f7a00001f7b,
        0x1f7c00001f7d,
        0x1fb000001fb2,
        0x1fb600001fb7,
        0x1fc600001fc7,
        0x1fd000001fd3,
        0x1fd600001fd8,
        0x1fe000001fe3,
        0x1fe400001fe8,
        0x1ff600001ff7,
        0x214e0000214f,
        0x218400002185,
        0x2c3000002c5f,
        0x2c6100002c62,
        0x2c6500002c67,
        0x2c6800002c69,
        0x2c6a00002c6b,
        0x2c6c00002c6d,
        0x2c7100002c72,
        0x2c7300002c75,
        0x2c7600002c7c,
        0x2c8100002c82,
        0x2c8300002c84,
        0x2c8500002c86,
        0x2c8700002c88,
        0x2c8900002c8a,
        0x2c8b00002c8c,
        0x2c8d00002c8e,
        0x2c8f00002c90,
        0x2c9100002c92,
        0x2c9300002c94,
        0x2c9500002c96,
        0x2c9700002c98,
        0x2c9900002c9a,
        0x2c9b00002c9c,
        0x2c9d00002c9e,
        0x2c9f00002ca0,
        0x2ca100002ca2,
        0x2ca300002ca4,
        0x2ca500002ca6,
        0x2ca700002ca8,
        0x2ca900002caa,
        0x2cab00002cac,
        0x2cad00002cae,
        0x2caf00002cb0,
        0x2cb100002cb2,
        0x2cb300002cb4,
        0x2cb500002cb6,
        0x2cb700002cb8,
        0x2cb900002cba,
        0x2cbb00002cbc,
        0x2cbd00002cbe,
        0x2cbf00002cc0,
        0x2cc100002cc2,
        0x2cc300002cc4,
        0x2cc500002cc6,
        0x2cc700002cc8,
        0x2cc900002cca,
        0x2ccb00002ccc,
        0x2ccd00002cce,
        0x2ccf00002cd0,
        0x2cd100002cd2,
        0x2cd300002cd4,
        0x2cd500002cd6,
        0x2cd700002cd8,
        0x2cd900002cda,
        0x2cdb00002cdc,
        0x2cdd00002cde,
        0x2cdf00002ce0,
        0x2ce100002ce2,
        0x2ce300002ce5,
        0x2cec00002ced,
        0x2cee00002cf2,
        0x2cf300002cf4,
        0x2d0000002d26,
        0x2d2700002d28,
        0x2d2d00002d2e,
        0x2d3000002d68,
        0x2d7f00002d97,
        0x2da000002da7,
        0x2da800002daf,
        0x2db000002db7,
        0x2db800002dbf,
        0x2dc000002dc7,
        0x2dc800002dcf,
        0x2dd000002dd7,
        0x2dd800002ddf,
        0x2de000002e00,
        0x2e2f00002e30,
        0x300500003008,
        0x302a0000302e,
        0x303c0000303d,
        0x304100003097,
        0x30990000309b,
        0x309d0000309f,
        0x30a1000030fb,
        0x30fc000030ff,
        0x310500003130,
        0x31a0000031c0,
        0x31f000003200,
        0x340000004dc0,
        0x4e0000009ffd,
        0xa0000000a48d,
        0xa4d00000a4fe,
        0xa5000000a60d,
        0xa6100000a62c,
        0xa6410000a642,
        0xa6430000a644,
        0xa6450000a646,
        0xa6470000a648,
        0xa6490000a64a,
        0xa64b0000a64c,
        0xa64d0000a64e,
        0xa64f0000a650,
        0xa6510000a652,
        0xa6530000a654,
        0xa6550000a656,
        0xa6570000a658,
        0xa6590000a65a,
        0xa65b0000a65c,
        0xa65d0000a65e,
        0xa65f0000a660,
        0xa6610000a662,
        0xa6630000a664,
        0xa6650000a666,
        0xa6670000a668,
        0xa6690000a66a,
        0xa66b0000a66c,
        0xa66d0000a670,
        0xa6740000a67e,
        0xa67f0000a680,
        0xa6810000a682,
        0xa6830000a684,
        0xa6850000a686,
        0xa6870000a688,
        0xa6890000a68a,
        0xa68b0000a68c,
        0xa68d0000a68e,
        0xa68f0000a690,
        0xa6910000a692,
        0xa6930000a694,
        0xa6950000a696,
        0xa6970000a698,
        0xa6990000a69a,
        0xa69b0000a69c,
        0xa69e0000a6e6,
        0xa6f00000a6f2,
        0xa7170000a720,
        0xa7230000a724,
        0xa7250000a726,
        0xa7270000a728,
        0xa7290000a72a,
        0xa72b0000a72c,
        0xa72d0000a72e,
        0xa72f0000a732,
        0xa7330000a734,
        0xa7350000a736,
        0xa7370000a738,
        0xa7390000a73a,
        0xa73b0000a73c,
        0xa73d0000a73e,
        0xa73f0000a740,
        0xa7410000a742,
        0xa7430000a744,
        0xa7450000a746,
        0xa7470000a748,
        0xa7490000a74a,
        0xa74b0000a74c,
        0xa74d0000a74e,
        0xa74f0000a750,
        0xa7510000a752,
        0xa7530000a754,
        0xa7550000a756,
        0xa7570000a758,
        0xa7590000a75a,
        0xa75b0000a75c,
        0xa75d0000a75e,
        0xa75f0000a760,
        0xa7610000a762,
        0xa7630000a764,
        0xa7650000a766,
        0xa7670000a768,
        0xa7690000a76a,
        0xa76b0000a76c,
        0xa76d0000a76e,
        0xa76f0000a770,
        0xa7710000a779,
        0xa77a0000a77b,
        0xa77c0000a77d,
        0xa77f0000a780,
        0xa7810000a782,
        0xa7830000a784,
        0xa7850000a786,
        0xa7870000a789,
        0xa78c0000a78d,
        0xa78e0000a790,
        0xa7910000a792,
        0xa7930000a796,
        0xa7970000a798,
        0xa7990000a79a,
        0xa79b0000a79c,
        0xa79d0000a79e,
        0xa79f0000a7a0,
        0xa7a10000a7a2,
        0xa7a30000a7a4,
        0xa7a50000a7a6,
        0xa7a70000a7a8,
        0xa7a90000a7aa,
        0xa7af0000a7b0,
        0xa7b50000a7b6,
        0xa7b70000a7b8,
        0xa7b90000a7ba,
        0xa7bb0000a7bc,
        0xa7bd0000a7be,
        0xa7bf0000a7c0,
        0xa7c30000a7c4,
        0xa7c80000a7c9,
        0xa7ca0000a7cb,
        0xa7f60000a7f8,
        0xa7fa0000a828,
        0xa82c0000a82d,
        0xa8400000a874,
        0xa8800000a8c6,
        0xa8d00000a8da,
        0xa8e00000a8f8,
        0xa8fb0000a8fc,
        0xa8fd0000a92e,
        0xa9300000a954,
        0xa9800000a9c1,
        0xa9cf0000a9da,
        0xa9e00000a9ff,
        0xaa000000aa37,
        0xaa400000aa4e,
        0xaa500000aa5a,
        0xaa600000aa77,
        0xaa7a0000aac3,
        0xaadb0000aade,
        0xaae00000aaf0,
        0xaaf20000aaf7,
        0xab010000ab07,
        0xab090000ab0f,
        0xab110000ab17,
        0xab200000ab27,
        0xab280000ab2f,
        0xab300000ab5b,
        0xab600000ab6a,
        0xabc00000abeb,
        0xabec0000abee,
        0xabf00000abfa,
        0xac000000d7a4,
        0xfa0e0000fa10,
        0xfa110000fa12,
        0xfa130000fa15,
        0xfa1f0000fa20,
        0xfa210000fa22,
        0xfa230000fa25,
        0xfa270000fa2a,
        0xfb1e0000fb1f,
        0xfe200000fe30,
        0xfe730000fe74,
        0x100000001000c,
        0x1000d00010027,
        0x100280001003b,
        0x1003c0001003e,
        0x1003f0001004e,
        0x100500001005e,
        0x10080000100fb,
        0x101fd000101fe,
        0x102800001029d,
        0x102a0000102d1,
        0x102e0000102e1,
        0x1030000010320,
        0x1032d00010341,
        0x103420001034a,
        0x103500001037b,
        0x103800001039e,
        0x103a0000103c4,
        0x103c8000103d0,
        0x104280001049e,
        0x104a0000104aa,
        0x104d8000104fc,
        0x1050000010528,
        0x1053000010564,
        0x1060000010737,
        0x1074000010756,
        0x1076000010768,
        0x1080000010806,
        0x1080800010809,
        0x1080a00010836,
        0x1083700010839,
        0x1083c0001083d,
        0x1083f00010856,
        0x1086000010877,
        0x108800001089f,
        0x108e0000108f3,
        0x108f4000108f6,
        0x1090000010916,
        0x109200001093a,
        0x10980000109b8,
        0x109be000109c0,
        0x10a0000010a04,
        0x10a0500010a07,
        0x10a0c00010a14,
        0x10a1500010a18,
        0x10a1900010a36,
        0x10a3800010a3b,
        0x10a3f00010a40,
        0x10a6000010a7d,
        0x10a8000010a9d,
        0x10ac000010ac8,
        0x10ac900010ae7,
        0x10b0000010b36,
        0x10b4000010b56,
        0x10b6000010b73,
        0x10b8000010b92,
        0x10c0000010c49,
        0x10cc000010cf3,
        0x10d0000010d28,
        0x10d3000010d3a,
        0x10e8000010eaa,
        0x10eab00010ead,
        0x10eb000010eb2,
        0x10f0000010f1d,
        0x10f2700010f28,
        0x10f3000010f51,
        0x10fb000010fc5,
        0x10fe000010ff7,
        0x1100000011047,
        0x1106600011070,
        0x1107f000110bb,
        0x110d0000110e9,
        0x110f0000110fa,
        0x1110000011135,
        0x1113600011140,
        0x1114400011148,
        0x1115000011174,
        0x1117600011177,
        0x11180000111c5,
        0x111c9000111cd,
        0x111ce000111db,
        0x111dc000111dd,
        0x1120000011212,
        0x1121300011238,
        0x1123e0001123f,
        0x1128000011287,
        0x1128800011289,
        0x1128a0001128e,
        0x1128f0001129e,
        0x1129f000112a9,
        0x112b0000112eb,
        0x112f0000112fa,
        0x1130000011304,
        0x113050001130d,
        0x1130f00011311,
        0x1131300011329,
        0x1132a00011331,
        0x1133200011334,
        0x113350001133a,
        0x1133b00011345,
        0x1134700011349,
        0x1134b0001134e,
        0x1135000011351,
        0x1135700011358,
        0x1135d00011364,
        0x113660001136d,
        0x1137000011375,
        0x114000001144b,
        0x114500001145a,
        0x1145e00011462,
        0x11480000114c6,
        0x114c7000114c8,
        0x114d0000114da,
        0x11580000115b6,
        0x115b8000115c1,
        0x115d8000115de,
        0x1160000011641,
        0x1164400011645,
        0x116500001165a,
        0x11680000116b9,
        0x116c0000116ca,
        0x117000001171b,
        0x1171d0001172c,
        0x117300001173a,
        0x118000001183b,
        0x118c0000118ea,
        0x118ff00011907,
        0x119090001190a,
        0x1190c00011914,
        0x1191500011917,
        0x1191800011936,
        0x1193700011939,
        0x1193b00011944,
        0x119500001195a,
        0x119a0000119a8,
        0x119aa000119d8,
        0x119da000119e2,
        0x119e3000119e5,
        0x11a0000011a3f,
        0x11a4700011a48,
        0x11a5000011a9a,
        0x11a9d00011a9e,
        0x11ac000011af9,
        0x11c0000011c09,
        0x11c0a00011c37,
        0x11c3800011c41,
        0x11c5000011c5a,
        0x11c7200011c90,
        0x11c9200011ca8,
        0x11ca900011cb7,
        0x11d0000011d07,
        0x11d0800011d0a,
        0x11d0b00011d37,
        0x11d3a00011d3b,
        0x11d3c00011d3e,
        0x11d3f00011d48,
        0x11d5000011d5a,
        0x11d6000011d66,
        0x11d6700011d69,
        0x11d6a00011d8f,
        0x11d9000011d92,
        0x11d9300011d99,
        0x11da000011daa,
        0x11ee000011ef7,
        0x11fb000011fb1,
        0x120000001239a,
        0x1248000012544,
        0x130000001342f,
        0x1440000014647,
        0x1680000016a39,
        0x16a4000016a5f,
        0x16a6000016a6a,
        0x16ad000016aee,
        0x16af000016af5,
        0x16b0000016b37,
        0x16b4000016b44,
        0x16b5000016b5a,
        0x16b6300016b78,
        0x16b7d00016b90,
        0x16e6000016e80,
        0x16f0000016f4b,
        0x16f4f00016f88,
        0x16f8f00016fa0,
        0x16fe000016fe2,
        0x16fe300016fe5,
        0x16ff000016ff2,
        0x17000000187f8,
        0x1880000018cd6,
        0x18d0000018d09,
        0x1b0000001b11f,
        0x1b1500001b153,
        0x1b1640001b168,
        0x1b1700001b2fc,
        0x1bc000001bc6b,
        0x1bc700001bc7d,
        0x1bc800001bc89,
        0x1bc900001bc9a,
        0x1bc9d0001bc9f,
        0x1da000001da37,
        0x1da3b0001da6d,
        0x1da750001da76,
        0x1da840001da85,
        0x1da9b0001daa0,
        0x1daa10001dab0,
        0x1e0000001e007,
        0x1e0080001e019,
        0x1e01b0001e022,
        0x1e0230001e025,
        0x1e0260001e02b,
        0x1e1000001e12d,
        0x1e1300001e13e,
        0x1e1400001e14a,
        0x1e14e0001e14f,
        0x1e2c00001e2fa,
        0x1e8000001e8c5,
        0x1e8d00001e8d7,
        0x1e9220001e94c,
        0x1e9500001e95a,
        0x1fbf00001fbfa,
        0x200000002a6de,
        0x2a7000002b735,
        0x2b7400002b81e,
        0x2b8200002cea2,
        0x2ceb00002ebe1,
        0x300000003134b,
    ),
    'CONTEXTJ': (
        0x200c0000200e,
    ),
    'CONTEXTO': (
        0xb7000000b8,
        0x37500000376,
        0x5f3000005f5,
        0x6600000066a,
        0x6f0000006fa,
        0x30fb000030fc,
    ),
}
site-packages/pip/_vendor/idna/intranges.py000064400000003615150564745010015003 0ustar00"""
Given a list of integers, made up of (hopefully) a small number of long runs
of consecutive integers, compute a representation of the form
((start1, end1), (start2, end2) ...). Then answer the question "was x present
in the original list?" in time O(log(# runs)).
"""

import bisect
from typing import List, Tuple

def intranges_from_list(list_):
    # type: (List[int]) -> Tuple[int, ...]
    """Represent a list of integers as a sequence of ranges:
    ((start_0, end_0), (start_1, end_1), ...), such that the original
    integers are exactly those x such that start_i <= x < end_i for some i.

    Ranges are encoded as single integers (start << 32 | end), not as tuples.
    """

    sorted_list = sorted(list_)
    ranges = []
    last_write = -1
    for i in range(len(sorted_list)):
        if i+1 < len(sorted_list):
            if sorted_list[i] == sorted_list[i+1]-1:
                continue
        current_range = sorted_list[last_write+1:i+1]
        ranges.append(_encode_range(current_range[0], current_range[-1] + 1))
        last_write = i

    return tuple(ranges)

def _encode_range(start, end):
    # type: (int, int) -> int
    return (start << 32) | end

def _decode_range(r):
    # type: (int) -> Tuple[int, int]
    return (r >> 32), (r & ((1 << 32) - 1))


def intranges_contain(int_, ranges):
    # type: (int, Tuple[int, ...]) -> bool
    """Determine if `int_` falls into one of the ranges in `ranges`."""
    tuple_ = _encode_range(int_, 0)
    pos = bisect.bisect_left(ranges, tuple_)
    # we could be immediately ahead of a tuple (start, end)
    # with start < int_ <= end
    if pos > 0:
        left, right = _decode_range(ranges[pos-1])
        if left <= int_ < right:
            return True
    # or we could be immediately behind a tuple (int_, end)
    if pos < len(ranges):
        left, _ = _decode_range(ranges[pos])
        if left == int_:
            return True
    return False
site-packages/pip/_vendor/idna/core.py000064400000031032150564745010013733 0ustar00from . import idnadata
import bisect
import unicodedata
import re
from typing import Union, Optional
from .intranges import intranges_contain

_virama_combining_class = 9
_alabel_prefix = b'xn--'
_unicode_dots_re = re.compile('[\u002e\u3002\uff0e\uff61]')

class IDNAError(UnicodeError):
    """ Base exception for all IDNA-encoding related problems """
    pass


class IDNABidiError(IDNAError):
    """ Exception when bidirectional requirements are not satisfied """
    pass


class InvalidCodepoint(IDNAError):
    """ Exception when a disallowed or unallocated codepoint is used """
    pass


class InvalidCodepointContext(IDNAError):
    """ Exception when the codepoint is not valid in the context it is used """
    pass


def _combining_class(cp):
    # type: (int) -> int
    v = unicodedata.combining(chr(cp))
    if v == 0:
        if not unicodedata.name(chr(cp)):
            raise ValueError('Unknown character in unicodedata')
    return v

def _is_script(cp, script):
    # type: (str, str) -> bool
    return intranges_contain(ord(cp), idnadata.scripts[script])

def _punycode(s):
    # type: (str) -> bytes
    return s.encode('punycode')

def _unot(s):
    # type: (int) -> str
    return 'U+{:04X}'.format(s)


def valid_label_length(label):
    # type: (Union[bytes, str]) -> bool
    if len(label) > 63:
        return False
    return True


def valid_string_length(label, trailing_dot):
    # type: (Union[bytes, str], bool) -> bool
    if len(label) > (254 if trailing_dot else 253):
        return False
    return True


def check_bidi(label, check_ltr=False):
    # type: (str, bool) -> bool
    # Bidi rules should only be applied if string contains RTL characters
    bidi_label = False
    for (idx, cp) in enumerate(label, 1):
        direction = unicodedata.bidirectional(cp)
        if direction == '':
            # String likely comes from a newer version of Unicode
            raise IDNABidiError('Unknown directionality in label {} at position {}'.format(repr(label), idx))
        if direction in ['R', 'AL', 'AN']:
            bidi_label = True
    if not bidi_label and not check_ltr:
        return True

    # Bidi rule 1
    direction = unicodedata.bidirectional(label[0])
    if direction in ['R', 'AL']:
        rtl = True
    elif direction == 'L':
        rtl = False
    else:
        raise IDNABidiError('First codepoint in label {} must be directionality L, R or AL'.format(repr(label)))

    valid_ending = False
    number_type = None  # type: Optional[str]
    for (idx, cp) in enumerate(label, 1):
        direction = unicodedata.bidirectional(cp)

        if rtl:
            # Bidi rule 2
            if not direction in ['R', 'AL', 'AN', 'EN', 'ES', 'CS', 'ET', 'ON', 'BN', 'NSM']:
                raise IDNABidiError('Invalid direction for codepoint at position {} in a right-to-left label'.format(idx))
            # Bidi rule 3
            if direction in ['R', 'AL', 'EN', 'AN']:
                valid_ending = True
            elif direction != 'NSM':
                valid_ending = False
            # Bidi rule 4
            if direction in ['AN', 'EN']:
                if not number_type:
                    number_type = direction
                else:
                    if number_type != direction:
                        raise IDNABidiError('Can not mix numeral types in a right-to-left label')
        else:
            # Bidi rule 5
            if not direction in ['L', 'EN', 'ES', 'CS', 'ET', 'ON', 'BN', 'NSM']:
                raise IDNABidiError('Invalid direction for codepoint at position {} in a left-to-right label'.format(idx))
            # Bidi rule 6
            if direction in ['L', 'EN']:
                valid_ending = True
            elif direction != 'NSM':
                valid_ending = False

    if not valid_ending:
        raise IDNABidiError('Label ends with illegal codepoint directionality')

    return True


def check_initial_combiner(label):
    # type: (str) -> bool
    if unicodedata.category(label[0])[0] == 'M':
        raise IDNAError('Label begins with an illegal combining character')
    return True


def check_hyphen_ok(label):
    # type: (str) -> bool
    if label[2:4] == '--':
        raise IDNAError('Label has disallowed hyphens in 3rd and 4th position')
    if label[0] == '-' or label[-1] == '-':
        raise IDNAError('Label must not start or end with a hyphen')
    return True


def check_nfc(label):
    # type: (str) -> None
    if unicodedata.normalize('NFC', label) != label:
        raise IDNAError('Label must be in Normalization Form C')


def valid_contextj(label, pos):
    # type: (str, int) -> bool
    cp_value = ord(label[pos])

    if cp_value == 0x200c:

        if pos > 0:
            if _combining_class(ord(label[pos - 1])) == _virama_combining_class:
                return True

        ok = False
        for i in range(pos-1, -1, -1):
            joining_type = idnadata.joining_types.get(ord(label[i]))
            if joining_type == ord('T'):
                continue
            if joining_type in [ord('L'), ord('D')]:
                ok = True
                break

        if not ok:
            return False

        ok = False
        for i in range(pos+1, len(label)):
            joining_type = idnadata.joining_types.get(ord(label[i]))
            if joining_type == ord('T'):
                continue
            if joining_type in [ord('R'), ord('D')]:
                ok = True
                break
        return ok

    if cp_value == 0x200d:

        if pos > 0:
            if _combining_class(ord(label[pos - 1])) == _virama_combining_class:
                return True
        return False

    else:

        return False


def valid_contexto(label, pos, exception=False):
    # type: (str, int, bool) -> bool
    cp_value = ord(label[pos])

    if cp_value == 0x00b7:
        if 0 < pos < len(label)-1:
            if ord(label[pos - 1]) == 0x006c and ord(label[pos + 1]) == 0x006c:
                return True
        return False

    elif cp_value == 0x0375:
        if pos < len(label)-1 and len(label) > 1:
            return _is_script(label[pos + 1], 'Greek')
        return False

    elif cp_value == 0x05f3 or cp_value == 0x05f4:
        if pos > 0:
            return _is_script(label[pos - 1], 'Hebrew')
        return False

    elif cp_value == 0x30fb:
        for cp in label:
            if cp == '\u30fb':
                continue
            if _is_script(cp, 'Hiragana') or _is_script(cp, 'Katakana') or _is_script(cp, 'Han'):
                return True
        return False

    elif 0x660 <= cp_value <= 0x669:
        for cp in label:
            if 0x6f0 <= ord(cp) <= 0x06f9:
                return False
        return True

    elif 0x6f0 <= cp_value <= 0x6f9:
        for cp in label:
            if 0x660 <= ord(cp) <= 0x0669:
                return False
        return True

    return False


def check_label(label):
    # type: (Union[str, bytes, bytearray]) -> None
    if isinstance(label, (bytes, bytearray)):
        label = label.decode('utf-8')
    if len(label) == 0:
        raise IDNAError('Empty Label')

    check_nfc(label)
    check_hyphen_ok(label)
    check_initial_combiner(label)

    for (pos, cp) in enumerate(label):
        cp_value = ord(cp)
        if intranges_contain(cp_value, idnadata.codepoint_classes['PVALID']):
            continue
        elif intranges_contain(cp_value, idnadata.codepoint_classes['CONTEXTJ']):
            try:
                if not valid_contextj(label, pos):
                    raise InvalidCodepointContext('Joiner {} not allowed at position {} in {}'.format(
                        _unot(cp_value), pos+1, repr(label)))
            except ValueError:
                raise IDNAError('Unknown codepoint adjacent to joiner {} at position {} in {}'.format(
                    _unot(cp_value), pos+1, repr(label)))
        elif intranges_contain(cp_value, idnadata.codepoint_classes['CONTEXTO']):
            if not valid_contexto(label, pos):
                raise InvalidCodepointContext('Codepoint {} not allowed at position {} in {}'.format(_unot(cp_value), pos+1, repr(label)))
        else:
            raise InvalidCodepoint('Codepoint {} at position {} of {} not allowed'.format(_unot(cp_value), pos+1, repr(label)))

    check_bidi(label)


def alabel(label):
    # type: (str) -> bytes
    try:
        label_bytes = label.encode('ascii')
        ulabel(label_bytes)
        if not valid_label_length(label_bytes):
            raise IDNAError('Label too long')
        return label_bytes
    except UnicodeEncodeError:
        pass

    if not label:
        raise IDNAError('No Input')

    label = str(label)
    check_label(label)
    label_bytes = _punycode(label)
    label_bytes = _alabel_prefix + label_bytes

    if not valid_label_length(label_bytes):
        raise IDNAError('Label too long')

    return label_bytes


def ulabel(label):
    # type: (Union[str, bytes, bytearray]) -> str
    if not isinstance(label, (bytes, bytearray)):
        try:
            label_bytes = label.encode('ascii')
        except UnicodeEncodeError:
            check_label(label)
            return label
    else:
        label_bytes = label

    label_bytes = label_bytes.lower()
    if label_bytes.startswith(_alabel_prefix):
        label_bytes = label_bytes[len(_alabel_prefix):]
        if not label_bytes:
            raise IDNAError('Malformed A-label, no Punycode eligible content found')
        if label_bytes.decode('ascii')[-1] == '-':
            raise IDNAError('A-label must not end with a hyphen')
    else:
        check_label(label_bytes)
        return label_bytes.decode('ascii')

    label = label_bytes.decode('punycode')
    check_label(label)
    return label


def uts46_remap(domain, std3_rules=True, transitional=False):
    # type: (str, bool, bool) -> str
    """Re-map the characters in the string according to UTS46 processing."""
    from .uts46data import uts46data
    output = ''

    for pos, char in enumerate(domain):
        code_point = ord(char)
        try:
            uts46row = uts46data[code_point if code_point < 256 else
                bisect.bisect_left(uts46data, (code_point, 'Z')) - 1]
            status = uts46row[1]
            replacement = None  # type: Optional[str]
            if len(uts46row) == 3:
                replacement = uts46row[2]  # type: ignore
            if (status == 'V' or
                    (status == 'D' and not transitional) or
                    (status == '3' and not std3_rules and replacement is None)):
                output += char
            elif replacement is not None and (status == 'M' or
                    (status == '3' and not std3_rules) or
                    (status == 'D' and transitional)):
                output += replacement
            elif status != 'I':
                raise IndexError()
        except IndexError:
            raise InvalidCodepoint(
                'Codepoint {} not allowed at position {} in {}'.format(
                _unot(code_point), pos + 1, repr(domain)))

    return unicodedata.normalize('NFC', output)


def encode(s, strict=False, uts46=False, std3_rules=False, transitional=False):
    # type: (Union[str, bytes, bytearray], bool, bool, bool, bool) -> bytes
    if isinstance(s, (bytes, bytearray)):
        s = s.decode('ascii')
    if uts46:
        s = uts46_remap(s, std3_rules, transitional)
    trailing_dot = False
    result = []
    if strict:
        labels = s.split('.')
    else:
        labels = _unicode_dots_re.split(s)
    if not labels or labels == ['']:
        raise IDNAError('Empty domain')
    if labels[-1] == '':
        del labels[-1]
        trailing_dot = True
    for label in labels:
        s = alabel(label)
        if s:
            result.append(s)
        else:
            raise IDNAError('Empty label')
    if trailing_dot:
        result.append(b'')
    s = b'.'.join(result)
    if not valid_string_length(s, trailing_dot):
        raise IDNAError('Domain too long')
    return s


def decode(s, strict=False, uts46=False, std3_rules=False):
    # type: (Union[str, bytes, bytearray], bool, bool, bool) -> str
    if isinstance(s, (bytes, bytearray)):
        s = s.decode('ascii')
    if uts46:
        s = uts46_remap(s, std3_rules, False)
    trailing_dot = False
    result = []
    if not strict:
        labels = _unicode_dots_re.split(s)
    else:
        labels = s.split('.')
    if not labels or labels == ['']:
        raise IDNAError('Empty domain')
    if not labels[-1]:
        del labels[-1]
        trailing_dot = True
    for label in labels:
        s = ulabel(label)
        if s:
            result.append(s)
        else:
            raise IDNAError('Empty label')
    if trailing_dot:
        result.append('')
    return '.'.join(result)
site-packages/pip/_vendor/platformdirs/__pycache__/__main__.cpython-39.pyc000064400000002271150564745010022610 0ustar00a

��Ret�@s4ddlmZmZdZdd�dd�Zedkr0e�dS)�)�PlatformDirs�__version__)	�
user_data_dir�user_config_dir�user_cache_dir�user_state_dir�user_log_dir�user_documents_dir�user_runtime_dir�
site_data_dir�site_config_dirN)�returncCs�d}d}tdt�d��td�t||dd�}tD]}t|�dt||����q2td	�t||�}tD]}t|�dt||����qftd
�t|�}tD]}t|�dt||����q�td�t|dd
�}tD]}t|�dt||����q�dS)NZMyAppZ	MyCompanyz-- platformdirs z --z%-- app dirs (with optional 'version')z1.0)�versionz: z)
-- app dirs (without optional 'version')z+
-- app dirs (without optional 'appauthor')z(
-- app dirs (with disabled 'appauthor')F)�	appauthor)�printrr�PROPS�getattr)Zapp_nameZ
app_author�dirs�prop�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/__main__.py�mains&
r�__main__)Zpip._vendor.platformdirsrrrr�__name__rrrr�<module>s
site-packages/pip/_vendor/platformdirs/__pycache__/windows.cpython-39.pyc000064400000014457150564745010022573 0ustar00a

��Re�@s�ddlZddlZddlmZddlmZmZddlmZGdd�de�Z	e
e
d�d	d
�Ze
e
d�dd�Ze
e
d�d
d�Z
ee
ge
fd�dd�Zedd�e��ZdgZdS)�N)�	lru_cache)�Callable�Optional�)�PlatformDirsABCc@s�eZdZdZeed�dd��Zdd�eeeed�dd	�Zeed�d
d��Z	eed�dd
��Z
eed�dd��Zeed�dd��Zeed�dd��Z
eed�dd��Zeed�dd��Zeed�dd��ZdS)�Windowsa�`MSDN on where to store app data files
    <http://support.microsoft.com/default.aspx?scid=kb;en-us;310294#XSLTH3194121123120121120120>`_.
    Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>`,
    `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`,
    `version <platformdirs.api.PlatformDirsABC.version>`,
    `roaming <platformdirs.api.PlatformDirsABC.roaming>`,
    `opinion <platformdirs.api.PlatformDirsABC.opinion>`.��returncCs(|jr
dnd}tj�t|��}|�|�S)z�
        :return: data directory tied to the user, e.g.
         ``%USERPROFILE%\AppData\Local\$appauthor\$appname`` (not roaming) or
         ``%USERPROFILE%\AppData\Roaming\$appauthor\$appname`` (roaming)
        �
CSIDL_APPDATA�CSIDL_LOCAL_APPDATA)�roaming�os�path�normpath�get_win_folder�
_append_parts)�self�constr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/windows.py�
user_data_dirszWindows.user_data_dirN��
opinion_value)rrr	cCstg}|jr`|jdur*|jp|j}|�|�|�|j�|durN|jrN|�|�|jr`|�|j�tjj|g|�R�S)NF)�appname�	appauthor�append�opinion�versionr
r�join)rrr�params�authorrrrrs


zWindows._append_partscCstj�td��}|�|�S)zT:return: data directory shared by users, e.g. ``C:\ProgramData\$appauthor\$appname``�CSIDL_COMMON_APPDATA�r
rrrr�rrrrr�
site_data_dir+szWindows.site_data_dircCs|jS)zC:return: config directory tied to the user, same as `user_data_dir`�r�rrrr�user_config_dir1szWindows.user_config_dircCs|jS)zF:return: config directory shared by the users, same as `site_data_dir`)r$r&rrr�site_config_dir6szWindows.site_config_dircCstj�td��}|j|dd�S)z�
        :return: cache directory tied to the user (if opinionated with ``Cache`` folder within ``$appname``) e.g.
         ``%USERPROFILE%\AppData\Local\$appauthor\$appname\Cache\$version``
        r�Cacherr"r#rrr�user_cache_dir;szWindows.user_cache_dircCs|jS)zB:return: state directory tied to the user, same as `user_data_dir`r%r&rrr�user_state_dirDszWindows.user_state_dircCs|j}|jrtj�|d�}|S)zy
        :return: log directory tied to the user, same as `user_data_dir` if not opinionated else ``Logs`` in it
        ZLogs)rrr
rrr#rrr�user_log_dirIszWindows.user_log_dircCstj�td��S)z`
        :return: documents directory tied to the user e.g. ``%USERPROFILE%\Documents``
        �CSIDL_PERSONAL)r
rrrr&rrr�user_documents_dirSszWindows.user_documents_dircCs$tj�tj�td�d��}|�|�S)z�
        :return: runtime directory tied to the user, e.g.
         ``%USERPROFILE%\AppData\Local\Temp\$appauthor\$appname``
        rZTemp)r
rrrrrr#rrr�user_runtime_dirZszWindows.user_runtime_dir)�__name__�
__module__�__qualname__�__doc__�property�strrrrr$r'r(r*r+r,r.r/rrrrr	s(	

	r)�
csidl_namer	cCsr|dkr$tj�tj�tjd�d�Sdddd��|�}|durLtd	|����tj�|�}|durntd
|����|S)z&Get folder from environment variables.r-�USERPROFILEZ	Documents�APPDATAZALLUSERSPROFILE�LOCALAPPDATA)r
r!rN�Unknown CSIDL name: zUnset environment variable: )r
rrr�environ�get�
ValueError)r6Zenv_var_name�resultrrr�get_win_folder_from_env_varsds��r?cCsXddddd��|�}|dur*td|����ddl}|�|jd	�}|�||�\}}t|�S)
z�Get folder from the registry.

    This is a fallback technique at best. I'm not sure if using the
    registry for this guarantees us the correct answer for all CSIDL_*
    names.
    ZAppDatazCommon AppDataz
Local AppDataZPersonal�r
r!rr-Nr:rz@Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders)r<r=�winreg�OpenKey�HKEY_CURRENT_USER�QueryValueExr5)r6Zshell_folder_namerA�key�	directory�_rrr�get_win_folder_from_registryvs��rHcCs�ddddd��|�}|dur*td|����t�d�}ttd	�}|j�d|dd
|�tdd�|D��r�t�d�}|j�	|j
|d�r�|}|j
S)
zGet folder with ctypes.��#��r@Nr:i�windllrcss|]}t|�dkVqdS)�N)�ord)�.0�crrr�	<genexpr>��z,get_win_folder_via_ctypes.<locals>.<genexpr>)r<r=�ctypes�create_unicode_buffer�getattrZshell32ZSHGetFolderPathW�anyZkernel32ZGetShortPathNameW�value)r6Zcsidl_const�bufrMZbuf2rrr�get_win_folder_via_ctypes�s"��


rZrcCs:ttd�rtSzddl}Wnty0tYS0tSdS)NrMr)�hasattrrTrZrA�ImportErrorr?rH)rArrr�_pick_get_win_folder�s

r])�maxsize)rTr
�	functoolsr�typingrr�apirrr5r?rHrZr]r�__all__rrrr�<module>s[�site-packages/pip/_vendor/platformdirs/__pycache__/unix.cpython-39.pyc000064400000015646150564745010022065 0ustar00a

��Re��@s�ddlZddlZddlmZddlmZddlmZddlm	Z	ej
�d�rZddlmZne
d	�d
d�ZGdd
�d
e	�Zeeed�dd�Zd
gZdS)�N)�ConfigParser)�Path)�Optional�)�PlatformDirsABC�linux)�getuid��returncCstd��dS)Nzshould only be used on Linux)�RuntimeError�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/unix.pyr
src@s�eZdZdZeed�dd��Zeed�dd��Zeed�dd	�Zeed�d
d��Z	eed�dd
��Z
eed�dd��Zeed�dd��Zeed�dd��Z
eed�dd��Zeed�dd��Zeed�dd��Zeed�dd��Zeed�dd�ZdS) �UnixaD
    On Unix/Linux, we follow the
    `XDG Basedir Spec <https://specifications.freedesktop.org/basedir-spec/basedir-spec-latest.html>`_. The spec allows
    overriding directories with environment variables. The examples show are the default values, alongside the name of
    the environment variable that overrides them. Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>`,
    `version <platformdirs.api.PlatformDirsABC.version>`,
    `multipath <platformdirs.api.PlatformDirsABC.multipath>`,
    `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    r	cCs,tj�dd�}|��s"tj�d�}|�|�S)z�
        :return: data directory tied to the user, e.g. ``~/.local/share/$appname/$version`` or
         ``$XDG_DATA_HOME/$appname/$version``
        �
XDG_DATA_HOME�z~/.local/share��os�environ�get�strip�path�
expanduser�_append_app_name_and_version��selfrrrr
�
user_data_dirszUnix.user_data_dircCs.tj�dd�}|��s$dtj�d�}|�|�S)aY
        :return: data directories shared by users (if `multipath <platformdirs.api.PlatformDirsABC.multipath>` is
         enabled and ``XDG_DATA_DIR`` is set and a multi path the response is also a multi path separated by the OS
         path separator), e.g. ``/usr/local/share/$appname/$version`` or ``/usr/share/$appname/$version``
        �
XDG_DATA_DIRSrz/usr/local/sharez
/usr/share)rrrr�pathsep�_with_multi_pathrrrr
�
site_data_dir(szUnix.site_data_dir)rr
cs<|�tj�}�js|dd�}�fdd�|D�}tj�|�S)Nrrcsg|]}��tj�|���qSr)rrrr)�.0�p�rrr
�
<listcomp>9�z)Unix._with_multi_path.<locals>.<listcomp>)�splitrr�	multipath�join)rr�	path_listrr"r
r5s
zUnix._with_multi_pathcCs,tj�dd�}|��s"tj�d�}|�|�S)z�
        :return: config directory tied to the user, e.g. ``~/.config/$appname/$version`` or
         ``$XDG_CONFIG_HOME/$appname/$version``
        �XDG_CONFIG_HOMErz	~/.configrrrrr
�user_config_dir<szUnix.user_config_dircCs$tj�dd�}|��sd}|�|�S)a/
        :return: config directories shared by users (if `multipath <platformdirs.api.PlatformDirsABC.multipath>`
         is enabled and ``XDG_DATA_DIR`` is set and a multi path the response is also a multi path separated by the OS
         path separator), e.g. ``/etc/xdg/$appname/$version``
        �XDG_CONFIG_DIRSrz/etc/xdg)rrrrrrrrr
�site_config_dirGszUnix.site_config_dircCs,tj�dd�}|��s"tj�d�}|�|�S)z�
        :return: cache directory tied to the user, e.g. ``~/.cache/$appname/$version`` or
         ``~/$XDG_CACHE_HOME/$appname/$version``
        �XDG_CACHE_HOMErz~/.cacherrrrr
�user_cache_dirTszUnix.user_cache_dircCs,tj�dd�}|��s"tj�d�}|�|�S)z�
        :return: state directory tied to the user, e.g. ``~/.local/state/$appname/$version`` or
         ``$XDG_STATE_HOME/$appname/$version``
        �XDG_STATE_HOMErz~/.local/staterrrrr
�user_state_dir_szUnix.user_state_dircCs|j}|jrtj�|d�}|S)zx
        :return: log directory tied to the user, same as `user_data_dir` if not opinionated else ``log`` in it
        �log)r.�opinionrrr'rrrr
�user_log_dirjszUnix.user_log_dircCs6td�}|dur2tj�dd���}|s2tj�d�}|S)zU
        :return: documents directory tied to the user, e.g. ``~/Documents``
        �XDG_DOCUMENTS_DIRNrz~/Documents)�_get_user_dirs_folderrrrrrr)r�
documents_dirrrr
�user_documents_dirtszUnix.user_documents_dircCs,tj�dd�}|��s"dt���}|�|�S)z�
        :return: runtime directory tied to the user, e.g. ``/run/user/$(id -u)/$appname/$version`` or
         ``$XDG_RUNTIME_DIR/$appname/$version``
        �XDG_RUNTIME_DIRrz
/run/user/)rrrrrrrrrr
�user_runtime_dir�szUnix.user_runtime_dircCs|�|j�S)zd:return: data path shared by users. Only return first item, even if ``multipath`` is set to ``True``)� _first_item_as_path_if_multipathrr"rrr
�site_data_path�szUnix.site_data_pathcCs|�|j�S)zj:return: config path shared by the users. Only return first item, even if ``multipath`` is set to ``True``)r:r,r"rrr
�site_config_path�szUnix.site_config_path)�	directoryr
cCs|jr|�tj�d}t|�S)Nr)r&r%rrr)rr=rrr
r:�sz%Unix._first_item_as_path_if_multipathN)�__name__�
__module__�__qualname__�__doc__�property�strrrrr*r,r.r0r3r7r9rr;r<r:rrrr
rs2



	
r)�keyr
cCs�tj�t�jd�}tj�|�r�t�}t|��$}|�d|�	����Wd�n1sV0Y||dvrpdS|d|�
d�}|�dtj�d��}|SdS)zjReturn directory from user-dirs.dirs config file. See https://freedesktop.org/wiki/Software/xdg-user-dirs/zuser-dirs.dirsz[top]
N�top�"z$HOME�~)
rrr'rr*�existsr�open�read_string�readr�replacer)rD�user_dirs_config_path�parser�streamrrrr
r5�s
2r5)r�sys�configparserr�pathlibr�typingr�apir�platform�
startswithr�intrrCr5�__all__rrrr
�<module>s
�site-packages/pip/_vendor/platformdirs/__pycache__/android.cpython-39.pyc000064400000010414150564745010022506 0ustar00a

��Re��@szddlZddlZddlZddlmZddlmZGdd�de�Zedd�ed�d	d
��Z	edd�ed�dd��Z
dgZdS)
�N)�	lru_cache�)�PlatformDirsABCc@s�eZdZdZeed�dd��Zeed�dd��Zeed�dd��Zeed�d	d
��Z	eed�dd��Z
eed�d
d��Zeed�dd��Zeed�dd��Z
eed�dd��ZdS)�Androidz�
    Follows the guidance `from here <https://android.stackexchange.com/a/216132>`_. Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>` and
    `version <platformdirs.api.PlatformDirsABC.version>`.
    ��returncCs|�t�d�S)zd:return: data directory tied to the user, e.g. ``/data/user/<userid>/<packagename>/files/<AppName>``�files��_append_app_name_and_version�_android_folder��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/android.py�
user_data_dirszAndroid.user_data_dircCs|jS)z@:return: data directory shared by users, same as `user_data_dir`�rrrrr�
site_data_dirszAndroid.site_data_dircCs|�t�d�S)z
        :return: config directory tied to the user, e.g. ``/data/user/<userid>/<packagename>/shared_prefs/<AppName>``
        Zshared_prefsr	rrrr�user_config_dirszAndroid.user_config_dircCs|jS)zH:return: config directory shared by the users, same as `user_config_dir`)rrrrr�site_config_dir!szAndroid.site_config_dircCs|�t�d�S)zj:return: cache directory tied to the user, e.g. e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>``�cacher	rrrr�user_cache_dir&szAndroid.user_cache_dircCs|jS)zB:return: state directory tied to the user, same as `user_data_dir`rrrrr�user_state_dir+szAndroid.user_state_dircCs|j}|jrtj�|d�}|S)z�
        :return: log directory tied to the user, same as `user_cache_dir` if not opinionated else ``log`` in it,
          e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>/log``
        �log�r�opinion�os�path�join�r
rrrr�user_log_dir0szAndroid.user_log_dircCst�S)zf
        :return: documents directory tied to the user e.g. ``/storage/emulated/0/Documents``
        )�_android_documents_folderrrrr�user_documents_dir;szAndroid.user_documents_dircCs|j}|jrtj�|d�}|S)z�
        :return: runtime directory tied to the user, same as `user_cache_dir` if not opinionated else ``tmp`` in it,
          e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>/tmp``
        �tmprrrrr�user_runtime_dirBszAndroid.user_runtime_dirN)�__name__�
__module__�__qualname__�__doc__�property�strrrrrrrrr!r#rrrrr	s&
r)�maxsizercCszz(ddlm}|d�}|������}WnLtytt�d�}tj	D] }|�
|�rF|�d�d}qpqFtd��Yn0|S)z':return: base folder for the Android OSr��	autoclass�android.content.Contextz /data/(data|user/\d+)/(.+)/filesz/filesz&Cannot find path to android app folder)
�jniusr,ZgetFilesDirZ
getParentFile�getAbsolutePath�	Exception�re�compile�sysr�match�split�OSError)r,�Context�result�patternrrrrrNs


rcCsLz0ddlm}|d�}|d�}|�|j���}WntyFd}Yn0|S)z,:return: documents folder for the Android OSrr+r-zandroid.os.Environmentz/storage/emulated/0/Documents)r.r,ZgetExternalFilesDirZDIRECTORY_DOCUMENTSr/r0)r,r7�Environment�
documents_dirrrrr cs
r )rr1r3�	functoolsr�apirrr)rr �__all__rrrr�<module>sE�site-packages/pip/_vendor/platformdirs/__pycache__/version.cpython-39.pyc000064400000000516150564745010022555 0ustar00a

��ReP�@sdZdZdZdS)z Version information z2.4.0)���N)�__doc__�__version__�__version_info__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/version.py�<module>ssite-packages/pip/_vendor/platformdirs/__pycache__/api.cpython-39.pyc000064400000012335150564745010021643 0ustar00a

��Re:�@sfddlZddlZddlmZmZddlmZddlmZm	Z	ej
dkrRddlmZGdd�de�ZdS)	�N)�ABC�abstractmethod)�Path)�Optional�Union)��)�Literalc@s�eZdZdZd1eeeeddfeeeeed�dd�Zeed	�d
d�Z	e
eed�d
d���Ze
eed�dd���Z
e
eed�dd���Ze
eed�dd���Ze
eed�dd���Ze
eed�dd���Ze
eed�dd���Ze
eed�dd���Ze
eed�dd���Ze
ed�dd ��Ze
ed�d!d"��Ze
ed�d#d$��Ze
ed�d%d&��Ze
ed�d'd(��Ze
ed�d)d*��Ze
ed�d+d,��Ze
ed�d-d.��Ze
ed�d/d0��ZdS)2�PlatformDirsABCz7
    Abstract base class for platform directories.
    NFTzLiteral[False]��appname�	appauthor�version�roaming�	multipath�opinioncCs(||_||_||_||_||_||_dS)a%
        Create a new platform directory.

        :param appname: See `appname`.
        :param appauthor: See `appauthor`.
        :param version: See `version`.
        :param roaming: See `roaming`.
        :param multipath: See `multipath`.
        :param opinion: See `opinion`.
        Nr)�selfrr
rrrr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/api.py�__init__szPlatformDirsABC.__init__)�base�returncGsLt|dd��}|jr4|�|j�|jr4|�|j�tjj|dg|�R�S)N�r)�listr�appendr�os�path�join)rr�paramsrrr�_append_app_name_and_version;sz,PlatformDirsABC._append_app_name_and_version)rcCsdS)z(:return: data directory tied to the userNr�rrrr�
user_data_dirCszPlatformDirsABC.user_data_dircCsdS)z':return: data directory shared by usersNrr rrr�
site_data_dirHszPlatformDirsABC.site_data_dircCsdS)z*:return: config directory tied to the userNrr rrr�user_config_dirMszPlatformDirsABC.user_config_dircCsdS)z-:return: config directory shared by the usersNrr rrr�site_config_dirRszPlatformDirsABC.site_config_dircCsdS)z):return: cache directory tied to the userNrr rrr�user_cache_dirWszPlatformDirsABC.user_cache_dircCsdS)z):return: state directory tied to the userNrr rrr�user_state_dir\szPlatformDirsABC.user_state_dircCsdS)z':return: log directory tied to the userNrr rrr�user_log_diraszPlatformDirsABC.user_log_dircCsdS)z-:return: documents directory tied to the userNrr rrr�user_documents_dirfsz"PlatformDirsABC.user_documents_dircCsdS)z+:return: runtime directory tied to the userNrr rrr�user_runtime_dirksz PlatformDirsABC.user_runtime_dircCs
t|j�S)z#:return: data path tied to the user)rr!r rrr�user_data_pathpszPlatformDirsABC.user_data_pathcCs
t|j�S)z":return: data path shared by users)rr"r rrr�site_data_pathuszPlatformDirsABC.site_data_pathcCs
t|j�S)z%:return: config path tied to the user)rr#r rrr�user_config_pathzsz PlatformDirsABC.user_config_pathcCs
t|j�S)z(:return: config path shared by the users)rr$r rrr�site_config_pathsz PlatformDirsABC.site_config_pathcCs
t|j�S)z$:return: cache path tied to the user)rr%r rrr�user_cache_path�szPlatformDirsABC.user_cache_pathcCs
t|j�S)z$:return: state path tied to the user)rr&r rrr�user_state_path�szPlatformDirsABC.user_state_pathcCs
t|j�S)z":return: log path tied to the user)rr'r rrr�
user_log_path�szPlatformDirsABC.user_log_pathcCs
t|j�S)z(:return: documents path tied to the user)rr(r rrr�user_documents_path�sz#PlatformDirsABC.user_documents_pathcCs
t|j�S)z&:return: runtime path tied to the user)rr)r rrr�user_runtime_path�sz!PlatformDirsABC.user_runtime_path)NNNFFT)�__name__�
__module__�__qualname__�__doc__r�strr�boolrr�propertyrr!r"r#r$r%r&r'r(r)rr*r+r,r-r.r/r0r1r2rrrrr
sz��+r
)
r�sys�abcrr�pathlibr�typingrr�version_infor	r
rrrr�<module>s
site-packages/pip/_vendor/platformdirs/__pycache__/__init__.cpython-39.pyc000064400000025551150564745010022635 0ustar00a

��Re;2�@sPdZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
erPddlmZddl
mZddlmZmZe	ed	�d
d�Ze�ZeZd7eee
edd
feeeed�dd�Zd8eee
edd
feeeed�dd�Zd9eee
edd
feeeed�dd�Zd:eee
edd
feeeed�dd�Zd;eee
edd
feeeed�dd�Zd<eee
edd
feeeed�dd�Zd=eee
edd
feeeed�dd�Zed	�d d!�Zd>eee
edd
feeeed�d"d#�Zd?eee
edd
feeeed�d$d%�Z d@eee
edd
feeeed�d&d'�Z!dAeee
edd
feeeed�d(d)�Z"dBeee
edd
feeeed�d*d+�Z#dCeee
edd
feeeed�d,d-�Z$dDeee
edd
feeeed�d.d/�Z%dEeee
edd
feeeed�d0d1�Z&ed	�d2d3�Z'dFeee
edd
feeeed�d4d5�Z(gd6�Z)dS)Gz�
Utilities for determining application-specific dirs. See <https://github.com/platformdirs/platformdirs> for details and
usage.
�N)�Path)�
TYPE_CHECKING�Optional�Type�Union)�Literal�)�PlatformDirsABC)�__version__�__version_info__)�returncCsjt�d�dkr&t�d�dkr&d\}}n0tjdkr:d\}}ntjdkrNd	\}}nd
\}}tt�|�|�}|S)N�ANDROID_DATAz/data�ANDROID_ROOTz/system)z pip._vendor.platformdirs.android�Android�win32)z pip._vendor.platformdirs.windows�Windows�darwin)zpip._vendor.platformdirs.macos�MacOS)zpip._vendor.platformdirs.unix�Unix)�os�getenv�sys�platform�getattr�	importlib�
import_module)�module�name�result�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/__init__.py�_set_platform_dir_classs




r!FzLiteral[False])�appname�	appauthor�version�roamingrcCst||||d�jS)aq
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: data directory tied to the user
    �r"r#r$r%)�PlatformDirs�
user_data_dirr&rrr r(#s
r()r"r#r$�	multipathrcCst||||d�jS)at
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: data directory shared by users
    �r"r#r$r))r'�
site_data_dirr*rrr r+3s
r+cCst||||d�jS)as
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: config directory tied to the user
    r&)r'�user_config_dirr&rrr r,Cs
r,cCst||||d�jS)az
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: config directory shared by the users
    r*)r'�site_config_dirr*rrr r-Ss
r-T)r"r#r$�opinionrcCst||||d�jS)ar
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: cache directory tied to the user
    �r"r#r$r.)r'�user_cache_dirr/rrr r0cs
r0cCst||||d�jS)ar
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: state directory tied to the user
    r&)r'�user_state_dirr&rrr r1ss
r1cCst||||d�jS)ap
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: log directory tied to the user
    r/)r'�user_log_dirr/rrr r2�s
r2cCst�jS)z8
    :returns: documents directory tied to the user
    )r'�user_documents_dirrrrr r3�sr3cCst||||d�jS)at
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: runtime directory tied to the user
    r/)r'�user_runtime_dirr/rrr r4�s
r4cCst||||d�jS)al
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: data path tied to the user
    r&)r'�user_data_pathr&rrr r5�s
r5cCst||||d�jS)aq
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `multipath <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: data path shared by users
    r*)r'�site_data_pathr*rrr r6�s
r6cCst||||d�jS)an
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: config path tied to the user
    r&)r'�user_config_pathr&rrr r7�s
r7cCst||||d�jS)au
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: config path shared by the users
    r*)r'�site_config_pathr*rrr r8�s
r8cCst||||d�jS)am
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: cache path tied to the user
    r/)r'�user_cache_pathr/rrr r9�s
r9cCst||||d�jS)am
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: state path tied to the user
    r&)r'�user_state_pathr&rrr r:�s
r:cCst||||d�jS)ak
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: log path tied to the user
    r/)r'�
user_log_pathr/rrr r;
s
r;cCst�jS)z3
    :returns: documents path tied to the user
    )r'�user_documents_pathrrrr r<sr<cCst||||d�jS)ao
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: runtime path tied to the user
    r/)r'�user_runtime_pathr/rrr r=!s
r=)r
rr'�AppDirsr	r(r,r0r1r2r3r4r+r-r5r7r9r:r;r<r=r6r8)NNNF)NNNF)NNNF)NNNF)NNNT)NNNF)NNNT)NNNT)NNNF)NNNF)NNNF)NNNF)NNNT)NNNF)NNNT)NNNT)*�__doc__rrr�pathlibr�typingrrrrZtyping_extensionsr�apir	r$r
rr!r'r>�str�boolr(r+r,r-r0r1r2r3r4r5r6r7r8r9r:r;r<r=�__all__rrrr �<module>s~
��������������������������������site-packages/pip/_vendor/platformdirs/__pycache__/macos.cpython-39.pyc000064400000006343150564745010022176 0ustar00a

��Re;
�@s.ddlZddlmZGdd�de�ZdgZdS)�N�)�PlatformDirsABCc@s�eZdZdZeed�dd��Zeed�dd��Zeed�dd��Zeed�d	d
��Z	eed�dd��Z
eed�d
d��Zeed�dd��Zeed�dd��Z
eed�dd��ZdS)�MacOSa�
    Platform directories for the macOS operating system. Follows the guidance from `Apple documentation
    <https://developer.apple.com/library/archive/documentation/FileManagement/Conceptual/FileSystemProgrammingGuide/MacOSXDirectories/MacOSXDirectories.html>`_.
    Makes use of the `appname <platformdirs.api.PlatformDirsABC.appname>` and
    `version <platformdirs.api.PlatformDirsABC.version>`.
    )�returncCs|�tj�d��S)zb:return: data directory tied to the user, e.g. ``~/Library/Application Support/$appname/$version``z~/Library/Application Support/��_append_app_name_and_version�os�path�
expanduser��self�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/platformdirs/macos.py�
user_data_dirszMacOS.user_data_dircCs
|�d�S)z`:return: data directory shared by users, e.g. ``/Library/Application Support/$appname/$version``z/Library/Application Support�rrr
r
r�
site_data_dirszMacOS.site_data_dircCs|�tj�d��S)z\:return: config directory tied to the user, e.g. ``~/Library/Preferences/$appname/$version``z~/Library/Preferences/rrr
r
r�user_config_dirszMacOS.user_config_dircCs
|�d�S)zU:return: config directory shared by the users, e.g. ``/Library/Preferences/$appname``z/Library/Preferencesrrr
r
r�site_config_dirszMacOS.site_config_dircCs|�tj�d��S)zV:return: cache directory tied to the user, e.g. ``~/Library/Caches/$appname/$version``z~/Library/Cachesrrr
r
r�user_cache_dir"szMacOS.user_cache_dircCs|jS)zB:return: state directory tied to the user, same as `user_data_dir`)rrr
r
r�user_state_dir'szMacOS.user_state_dircCs|�tj�d��S)zR:return: log directory tied to the user, e.g. ``~/Library/Logs/$appname/$version``z~/Library/Logsrrr
r
r�user_log_dir,szMacOS.user_log_dircCstj�d�S)zC:return: documents directory tied to the user, e.g. ``~/Documents``z~/Documents)rr	r
rr
r
r�user_documents_dir1szMacOS.user_documents_dircCs|�tj�d��S)zg:return: runtime directory tied to the user, e.g. ``~/Library/Caches/TemporaryItems/$appname/$version``z~/Library/Caches/TemporaryItemsrrr
r
r�user_runtime_dir6szMacOS.user_runtime_dirN)�__name__�
__module__�__qualname__�__doc__�property�strrrrrrrrrrr
r
r
rrs&r)r�apirr�__all__r
r
r
r�<module>s7�site-packages/pip/_vendor/platformdirs/unix.py000064400000015371150564745010015571 0ustar00import os
import sys
from configparser import ConfigParser
from pathlib import Path
from typing import Optional

from .api import PlatformDirsABC

if sys.platform.startswith("linux"):  # pragma: no branch # no op check, only to please the type checker
    from os import getuid
else:

    def getuid() -> int:
        raise RuntimeError("should only be used on Linux")


class Unix(PlatformDirsABC):
    """
    On Unix/Linux, we follow the
    `XDG Basedir Spec <https://specifications.freedesktop.org/basedir-spec/basedir-spec-latest.html>`_. The spec allows
    overriding directories with environment variables. The examples show are the default values, alongside the name of
    the environment variable that overrides them. Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>`,
    `version <platformdirs.api.PlatformDirsABC.version>`,
    `multipath <platformdirs.api.PlatformDirsABC.multipath>`,
    `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    """

    @property
    def user_data_dir(self) -> str:
        """
        :return: data directory tied to the user, e.g. ``~/.local/share/$appname/$version`` or
         ``$XDG_DATA_HOME/$appname/$version``
        """
        path = os.environ.get("XDG_DATA_HOME", "")
        if not path.strip():
            path = os.path.expanduser("~/.local/share")
        return self._append_app_name_and_version(path)

    @property
    def site_data_dir(self) -> str:
        """
        :return: data directories shared by users (if `multipath <platformdirs.api.PlatformDirsABC.multipath>` is
         enabled and ``XDG_DATA_DIR`` is set and a multi path the response is also a multi path separated by the OS
         path separator), e.g. ``/usr/local/share/$appname/$version`` or ``/usr/share/$appname/$version``
        """
        # XDG default for $XDG_DATA_DIRS; only first, if multipath is False
        path = os.environ.get("XDG_DATA_DIRS", "")
        if not path.strip():
            path = f"/usr/local/share{os.pathsep}/usr/share"
        return self._with_multi_path(path)

    def _with_multi_path(self, path: str) -> str:
        path_list = path.split(os.pathsep)
        if not self.multipath:
            path_list = path_list[0:1]
        path_list = [self._append_app_name_and_version(os.path.expanduser(p)) for p in path_list]
        return os.pathsep.join(path_list)

    @property
    def user_config_dir(self) -> str:
        """
        :return: config directory tied to the user, e.g. ``~/.config/$appname/$version`` or
         ``$XDG_CONFIG_HOME/$appname/$version``
        """
        path = os.environ.get("XDG_CONFIG_HOME", "")
        if not path.strip():
            path = os.path.expanduser("~/.config")
        return self._append_app_name_and_version(path)

    @property
    def site_config_dir(self) -> str:
        """
        :return: config directories shared by users (if `multipath <platformdirs.api.PlatformDirsABC.multipath>`
         is enabled and ``XDG_DATA_DIR`` is set and a multi path the response is also a multi path separated by the OS
         path separator), e.g. ``/etc/xdg/$appname/$version``
        """
        # XDG default for $XDG_CONFIG_DIRS only first, if multipath is False
        path = os.environ.get("XDG_CONFIG_DIRS", "")
        if not path.strip():
            path = "/etc/xdg"
        return self._with_multi_path(path)

    @property
    def user_cache_dir(self) -> str:
        """
        :return: cache directory tied to the user, e.g. ``~/.cache/$appname/$version`` or
         ``~/$XDG_CACHE_HOME/$appname/$version``
        """
        path = os.environ.get("XDG_CACHE_HOME", "")
        if not path.strip():
            path = os.path.expanduser("~/.cache")
        return self._append_app_name_and_version(path)

    @property
    def user_state_dir(self) -> str:
        """
        :return: state directory tied to the user, e.g. ``~/.local/state/$appname/$version`` or
         ``$XDG_STATE_HOME/$appname/$version``
        """
        path = os.environ.get("XDG_STATE_HOME", "")
        if not path.strip():
            path = os.path.expanduser("~/.local/state")
        return self._append_app_name_and_version(path)

    @property
    def user_log_dir(self) -> str:
        """
        :return: log directory tied to the user, same as `user_data_dir` if not opinionated else ``log`` in it
        """
        path = self.user_cache_dir
        if self.opinion:
            path = os.path.join(path, "log")
        return path

    @property
    def user_documents_dir(self) -> str:
        """
        :return: documents directory tied to the user, e.g. ``~/Documents``
        """
        documents_dir = _get_user_dirs_folder("XDG_DOCUMENTS_DIR")
        if documents_dir is None:
            documents_dir = os.environ.get("XDG_DOCUMENTS_DIR", "").strip()
            if not documents_dir:
                documents_dir = os.path.expanduser("~/Documents")

        return documents_dir

    @property
    def user_runtime_dir(self) -> str:
        """
        :return: runtime directory tied to the user, e.g. ``/run/user/$(id -u)/$appname/$version`` or
         ``$XDG_RUNTIME_DIR/$appname/$version``
        """
        path = os.environ.get("XDG_RUNTIME_DIR", "")
        if not path.strip():
            path = f"/run/user/{getuid()}"
        return self._append_app_name_and_version(path)

    @property
    def site_data_path(self) -> Path:
        """:return: data path shared by users. Only return first item, even if ``multipath`` is set to ``True``"""
        return self._first_item_as_path_if_multipath(self.site_data_dir)

    @property
    def site_config_path(self) -> Path:
        """:return: config path shared by the users. Only return first item, even if ``multipath`` is set to ``True``"""
        return self._first_item_as_path_if_multipath(self.site_config_dir)

    def _first_item_as_path_if_multipath(self, directory: str) -> Path:
        if self.multipath:
            # If multipath is True, the first path is returned.
            directory = directory.split(os.pathsep)[0]
        return Path(directory)


def _get_user_dirs_folder(key: str) -> Optional[str]:
    """Return directory from user-dirs.dirs config file. See https://freedesktop.org/wiki/Software/xdg-user-dirs/"""
    user_dirs_config_path = os.path.join(Unix().user_config_dir, "user-dirs.dirs")
    if os.path.exists(user_dirs_config_path):
        parser = ConfigParser()

        with open(user_dirs_config_path) as stream:
            # Add fake section header, so ConfigParser doesn't complain
            parser.read_string(f"[top]\n{stream.read()}")

        if key not in parser["top"]:
            return None

        path = parser["top"][key].strip('"')
        # Handle relative home paths
        path = path.replace("$HOME", os.path.expanduser("~"))
        return path

    return None


__all__ = [
    "Unix",
]
site-packages/pip/_vendor/platformdirs/macos.py000064400000005073150564745010015706 0ustar00import os

from .api import PlatformDirsABC


class MacOS(PlatformDirsABC):
    """
    Platform directories for the macOS operating system. Follows the guidance from `Apple documentation
    <https://developer.apple.com/library/archive/documentation/FileManagement/Conceptual/FileSystemProgrammingGuide/MacOSXDirectories/MacOSXDirectories.html>`_.
    Makes use of the `appname <platformdirs.api.PlatformDirsABC.appname>` and
    `version <platformdirs.api.PlatformDirsABC.version>`.
    """

    @property
    def user_data_dir(self) -> str:
        """:return: data directory tied to the user, e.g. ``~/Library/Application Support/$appname/$version``"""
        return self._append_app_name_and_version(os.path.expanduser("~/Library/Application Support/"))

    @property
    def site_data_dir(self) -> str:
        """:return: data directory shared by users, e.g. ``/Library/Application Support/$appname/$version``"""
        return self._append_app_name_and_version("/Library/Application Support")

    @property
    def user_config_dir(self) -> str:
        """:return: config directory tied to the user, e.g. ``~/Library/Preferences/$appname/$version``"""
        return self._append_app_name_and_version(os.path.expanduser("~/Library/Preferences/"))

    @property
    def site_config_dir(self) -> str:
        """:return: config directory shared by the users, e.g. ``/Library/Preferences/$appname``"""
        return self._append_app_name_and_version("/Library/Preferences")

    @property
    def user_cache_dir(self) -> str:
        """:return: cache directory tied to the user, e.g. ``~/Library/Caches/$appname/$version``"""
        return self._append_app_name_and_version(os.path.expanduser("~/Library/Caches"))

    @property
    def user_state_dir(self) -> str:
        """:return: state directory tied to the user, same as `user_data_dir`"""
        return self.user_data_dir

    @property
    def user_log_dir(self) -> str:
        """:return: log directory tied to the user, e.g. ``~/Library/Logs/$appname/$version``"""
        return self._append_app_name_and_version(os.path.expanduser("~/Library/Logs"))

    @property
    def user_documents_dir(self) -> str:
        """:return: documents directory tied to the user, e.g. ``~/Documents``"""
        return os.path.expanduser("~/Documents")

    @property
    def user_runtime_dir(self) -> str:
        """:return: runtime directory tied to the user, e.g. ``~/Library/Caches/TemporaryItems/$appname/$version``"""
        return self._append_app_name_and_version(os.path.expanduser("~/Library/Caches/TemporaryItems"))


__all__ = [
    "MacOS",
]
site-packages/pip/_vendor/platformdirs/api.py000064400000011472150564745010015355 0ustar00import os
import sys
from abc import ABC, abstractmethod
from pathlib import Path
from typing import Optional, Union

if sys.version_info >= (3, 8):  # pragma: no branch
    from typing import Literal  # pragma: no cover


class PlatformDirsABC(ABC):
    """
    Abstract base class for platform directories.
    """

    def __init__(
        self,
        appname: Optional[str] = None,
        appauthor: Union[str, None, "Literal[False]"] = None,
        version: Optional[str] = None,
        roaming: bool = False,
        multipath: bool = False,
        opinion: bool = True,
    ):
        """
        Create a new platform directory.

        :param appname: See `appname`.
        :param appauthor: See `appauthor`.
        :param version: See `version`.
        :param roaming: See `roaming`.
        :param multipath: See `multipath`.
        :param opinion: See `opinion`.
        """
        self.appname = appname  #: The name of application.
        self.appauthor = appauthor
        """
        The name of the app author or distributing body for this application. Typically, it is the owning company name.
        Defaults to `appname`. You may pass ``False`` to disable it.
        """
        self.version = version
        """
        An optional version path element to append to the path. You might want to use this if you want multiple versions
        of your app to be able to run independently. If used, this would typically be ``<major>.<minor>``.
        """
        self.roaming = roaming
        """
        Whether to use the roaming appdata directory on Windows. That means that for users on a Windows network setup
        for roaming profiles, this user data will be synced on login (see
        `here <http://technet.microsoft.com/en-us/library/cc766489(WS.10).aspx>`_).
        """
        self.multipath = multipath
        """
        An optional parameter only applicable to Unix/Linux which indicates that the entire list of data dirs should be
        returned. By default, the first item would only be returned.
        """
        self.opinion = opinion  #: A flag to indicating to use opinionated values.

    def _append_app_name_and_version(self, *base: str) -> str:
        params = list(base[1:])
        if self.appname:
            params.append(self.appname)
            if self.version:
                params.append(self.version)
        return os.path.join(base[0], *params)

    @property
    @abstractmethod
    def user_data_dir(self) -> str:
        """:return: data directory tied to the user"""

    @property
    @abstractmethod
    def site_data_dir(self) -> str:
        """:return: data directory shared by users"""

    @property
    @abstractmethod
    def user_config_dir(self) -> str:
        """:return: config directory tied to the user"""

    @property
    @abstractmethod
    def site_config_dir(self) -> str:
        """:return: config directory shared by the users"""

    @property
    @abstractmethod
    def user_cache_dir(self) -> str:
        """:return: cache directory tied to the user"""

    @property
    @abstractmethod
    def user_state_dir(self) -> str:
        """:return: state directory tied to the user"""

    @property
    @abstractmethod
    def user_log_dir(self) -> str:
        """:return: log directory tied to the user"""

    @property
    @abstractmethod
    def user_documents_dir(self) -> str:
        """:return: documents directory tied to the user"""

    @property
    @abstractmethod
    def user_runtime_dir(self) -> str:
        """:return: runtime directory tied to the user"""

    @property
    def user_data_path(self) -> Path:
        """:return: data path tied to the user"""
        return Path(self.user_data_dir)

    @property
    def site_data_path(self) -> Path:
        """:return: data path shared by users"""
        return Path(self.site_data_dir)

    @property
    def user_config_path(self) -> Path:
        """:return: config path tied to the user"""
        return Path(self.user_config_dir)

    @property
    def site_config_path(self) -> Path:
        """:return: config path shared by the users"""
        return Path(self.site_config_dir)

    @property
    def user_cache_path(self) -> Path:
        """:return: cache path tied to the user"""
        return Path(self.user_cache_dir)

    @property
    def user_state_path(self) -> Path:
        """:return: state path tied to the user"""
        return Path(self.user_state_dir)

    @property
    def user_log_path(self) -> Path:
        """:return: log path tied to the user"""
        return Path(self.user_log_dir)

    @property
    def user_documents_path(self) -> Path:
        """:return: documents path tied to the user"""
        return Path(self.user_documents_dir)

    @property
    def user_runtime_path(self) -> Path:
        """:return: runtime path tied to the user"""
        return Path(self.user_runtime_dir)
site-packages/pip/_vendor/platformdirs/version.py000064400000000120150564745010016255 0ustar00""" Version information """

__version__ = "2.4.0"
__version_info__ = (2, 4, 0)
site-packages/pip/_vendor/platformdirs/android.py000064400000007632150564745010016227 0ustar00import os
import re
import sys
from functools import lru_cache

from .api import PlatformDirsABC


class Android(PlatformDirsABC):
    """
    Follows the guidance `from here <https://android.stackexchange.com/a/216132>`_. Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>` and
    `version <platformdirs.api.PlatformDirsABC.version>`.
    """

    @property
    def user_data_dir(self) -> str:
        """:return: data directory tied to the user, e.g. ``/data/user/<userid>/<packagename>/files/<AppName>``"""
        return self._append_app_name_and_version(_android_folder(), "files")

    @property
    def site_data_dir(self) -> str:
        """:return: data directory shared by users, same as `user_data_dir`"""
        return self.user_data_dir

    @property
    def user_config_dir(self) -> str:
        """
        :return: config directory tied to the user, e.g. ``/data/user/<userid>/<packagename>/shared_prefs/<AppName>``
        """
        return self._append_app_name_and_version(_android_folder(), "shared_prefs")

    @property
    def site_config_dir(self) -> str:
        """:return: config directory shared by the users, same as `user_config_dir`"""
        return self.user_config_dir

    @property
    def user_cache_dir(self) -> str:
        """:return: cache directory tied to the user, e.g. e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>``"""
        return self._append_app_name_and_version(_android_folder(), "cache")

    @property
    def user_state_dir(self) -> str:
        """:return: state directory tied to the user, same as `user_data_dir`"""
        return self.user_data_dir

    @property
    def user_log_dir(self) -> str:
        """
        :return: log directory tied to the user, same as `user_cache_dir` if not opinionated else ``log`` in it,
          e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>/log``
        """
        path = self.user_cache_dir
        if self.opinion:
            path = os.path.join(path, "log")
        return path

    @property
    def user_documents_dir(self) -> str:
        """
        :return: documents directory tied to the user e.g. ``/storage/emulated/0/Documents``
        """
        return _android_documents_folder()

    @property
    def user_runtime_dir(self) -> str:
        """
        :return: runtime directory tied to the user, same as `user_cache_dir` if not opinionated else ``tmp`` in it,
          e.g. ``/data/user/<userid>/<packagename>/cache/<AppName>/tmp``
        """
        path = self.user_cache_dir
        if self.opinion:
            path = os.path.join(path, "tmp")
        return path


@lru_cache(maxsize=1)
def _android_folder() -> str:
    """:return: base folder for the Android OS"""
    try:
        # First try to get path to android app via pyjnius
        from jnius import autoclass  # noqa: SC200

        Context = autoclass("android.content.Context")  # noqa: SC200
        result: str = Context.getFilesDir().getParentFile().getAbsolutePath()
    except Exception:
        # if fails find an android folder looking path on the sys.path
        pattern = re.compile(r"/data/(data|user/\d+)/(.+)/files")
        for path in sys.path:
            if pattern.match(path):
                result = path.split("/files")[0]
                break
        else:
            raise OSError("Cannot find path to android app folder")
    return result


@lru_cache(maxsize=1)
def _android_documents_folder() -> str:
    """:return: documents folder for the Android OS"""
    # Get directories with pyjnius
    try:
        from jnius import autoclass  # noqa: SC200

        Context = autoclass("android.content.Context")  # noqa: SC200
        Environment = autoclass("android.os.Environment")
        documents_dir: str = Context.getExternalFilesDir(Environment.DIRECTORY_DOCUMENTS).getAbsolutePath()
    except Exception:
        documents_dir = "/storage/emulated/0/Documents"

    return documents_dir


__all__ = [
    "Android",
]
site-packages/pip/_vendor/platformdirs/__init__.py000064400000031073150564745010016342 0ustar00"""
Utilities for determining application-specific dirs. See <https://github.com/platformdirs/platformdirs> for details and
usage.
"""
import importlib
import os
import sys
from pathlib import Path
from typing import TYPE_CHECKING, Optional, Type, Union

if TYPE_CHECKING:
    from typing_extensions import Literal  # pragma: no cover

from .api import PlatformDirsABC
from .version import __version__, __version_info__


def _set_platform_dir_class() -> Type[PlatformDirsABC]:
    if os.getenv("ANDROID_DATA") == "/data" and os.getenv("ANDROID_ROOT") == "/system":
        module, name = "pip._vendor.platformdirs.android", "Android"
    elif sys.platform == "win32":
        module, name = "pip._vendor.platformdirs.windows", "Windows"
    elif sys.platform == "darwin":
        module, name = "pip._vendor.platformdirs.macos", "MacOS"
    else:
        module, name = "pip._vendor.platformdirs.unix", "Unix"
    result: Type[PlatformDirsABC] = getattr(importlib.import_module(module), name)
    return result


PlatformDirs = _set_platform_dir_class()  #: Currently active platform
AppDirs = PlatformDirs  #: Backwards compatibility with appdirs


def user_data_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: data directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_data_dir


def site_data_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    multipath: bool = False,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: data directory shared by users
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, multipath=multipath).site_data_dir


def user_config_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: config directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_config_dir


def site_config_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    multipath: bool = False,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: config directory shared by the users
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, multipath=multipath).site_config_dir


def user_cache_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: cache directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_cache_dir


def user_state_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: state directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_state_dir


def user_log_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: log directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_log_dir


def user_documents_dir() -> str:
    """
    :returns: documents directory tied to the user
    """
    return PlatformDirs().user_documents_dir


def user_runtime_dir(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> str:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: runtime directory tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_runtime_dir


def user_data_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: data path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_data_path


def site_data_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    multipath: bool = False,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `multipath <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: data path shared by users
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, multipath=multipath).site_data_path


def user_config_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: config path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_config_path


def site_config_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    multipath: bool = False,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param multipath: See `roaming <platformdirs.api.PlatformDirsABC.multipath>`.
    :returns: config path shared by the users
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, multipath=multipath).site_config_path


def user_cache_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: cache path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_cache_path


def user_state_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    roaming: bool = False,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param roaming: See `roaming <platformdirs.api.PlatformDirsABC.version>`.
    :returns: state path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, roaming=roaming).user_state_path


def user_log_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `roaming <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: log path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_log_path


def user_documents_path() -> Path:
    """
    :returns: documents path tied to the user
    """
    return PlatformDirs().user_documents_path


def user_runtime_path(
    appname: Optional[str] = None,
    appauthor: Union[str, None, "Literal[False]"] = None,
    version: Optional[str] = None,
    opinion: bool = True,
) -> Path:
    """
    :param appname: See `appname <platformdirs.api.PlatformDirsABC.appname>`.
    :param appauthor: See `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`.
    :param version: See `version <platformdirs.api.PlatformDirsABC.version>`.
    :param opinion: See `opinion <platformdirs.api.PlatformDirsABC.opinion>`.
    :returns: runtime path tied to the user
    """
    return PlatformDirs(appname=appname, appauthor=appauthor, version=version, opinion=opinion).user_runtime_path


__all__ = [
    "__version__",
    "__version_info__",
    "PlatformDirs",
    "AppDirs",
    "PlatformDirsABC",
    "user_data_dir",
    "user_config_dir",
    "user_cache_dir",
    "user_state_dir",
    "user_log_dir",
    "user_documents_dir",
    "user_runtime_dir",
    "site_data_dir",
    "site_config_dir",
    "user_data_path",
    "user_config_path",
    "user_cache_path",
    "user_state_path",
    "user_log_path",
    "user_documents_path",
    "user_runtime_path",
    "site_data_path",
    "site_config_path",
]
site-packages/pip/_vendor/platformdirs/windows.py000064400000014420150564745010016272 0ustar00import ctypes
import os
from functools import lru_cache
from typing import Callable, Optional

from .api import PlatformDirsABC


class Windows(PlatformDirsABC):
    """`MSDN on where to store app data files
    <http://support.microsoft.com/default.aspx?scid=kb;en-us;310294#XSLTH3194121123120121120120>`_.
    Makes use of the
    `appname <platformdirs.api.PlatformDirsABC.appname>`,
    `appauthor <platformdirs.api.PlatformDirsABC.appauthor>`,
    `version <platformdirs.api.PlatformDirsABC.version>`,
    `roaming <platformdirs.api.PlatformDirsABC.roaming>`,
    `opinion <platformdirs.api.PlatformDirsABC.opinion>`."""

    @property
    def user_data_dir(self) -> str:
        """
        :return: data directory tied to the user, e.g.
         ``%USERPROFILE%\\AppData\\Local\\$appauthor\\$appname`` (not roaming) or
         ``%USERPROFILE%\\AppData\\Roaming\\$appauthor\\$appname`` (roaming)
        """
        const = "CSIDL_APPDATA" if self.roaming else "CSIDL_LOCAL_APPDATA"
        path = os.path.normpath(get_win_folder(const))
        return self._append_parts(path)

    def _append_parts(self, path: str, *, opinion_value: Optional[str] = None) -> str:
        params = []
        if self.appname:
            if self.appauthor is not False:
                author = self.appauthor or self.appname
                params.append(author)
            params.append(self.appname)
            if opinion_value is not None and self.opinion:
                params.append(opinion_value)
            if self.version:
                params.append(self.version)
        return os.path.join(path, *params)

    @property
    def site_data_dir(self) -> str:
        """:return: data directory shared by users, e.g. ``C:\\ProgramData\\$appauthor\\$appname``"""
        path = os.path.normpath(get_win_folder("CSIDL_COMMON_APPDATA"))
        return self._append_parts(path)

    @property
    def user_config_dir(self) -> str:
        """:return: config directory tied to the user, same as `user_data_dir`"""
        return self.user_data_dir

    @property
    def site_config_dir(self) -> str:
        """:return: config directory shared by the users, same as `site_data_dir`"""
        return self.site_data_dir

    @property
    def user_cache_dir(self) -> str:
        """
        :return: cache directory tied to the user (if opinionated with ``Cache`` folder within ``$appname``) e.g.
         ``%USERPROFILE%\\AppData\\Local\\$appauthor\\$appname\\Cache\\$version``
        """
        path = os.path.normpath(get_win_folder("CSIDL_LOCAL_APPDATA"))
        return self._append_parts(path, opinion_value="Cache")

    @property
    def user_state_dir(self) -> str:
        """:return: state directory tied to the user, same as `user_data_dir`"""
        return self.user_data_dir

    @property
    def user_log_dir(self) -> str:
        """
        :return: log directory tied to the user, same as `user_data_dir` if not opinionated else ``Logs`` in it
        """
        path = self.user_data_dir
        if self.opinion:
            path = os.path.join(path, "Logs")
        return path

    @property
    def user_documents_dir(self) -> str:
        """
        :return: documents directory tied to the user e.g. ``%USERPROFILE%\\Documents``
        """
        return os.path.normpath(get_win_folder("CSIDL_PERSONAL"))

    @property
    def user_runtime_dir(self) -> str:
        """
        :return: runtime directory tied to the user, e.g.
         ``%USERPROFILE%\\AppData\\Local\\Temp\\$appauthor\\$appname``
        """
        path = os.path.normpath(os.path.join(get_win_folder("CSIDL_LOCAL_APPDATA"), "Temp"))
        return self._append_parts(path)


def get_win_folder_from_env_vars(csidl_name: str) -> str:
    """Get folder from environment variables."""
    if csidl_name == "CSIDL_PERSONAL":  # does not have an environment name
        return os.path.join(os.path.normpath(os.environ["USERPROFILE"]), "Documents")

    env_var_name = {
        "CSIDL_APPDATA": "APPDATA",
        "CSIDL_COMMON_APPDATA": "ALLUSERSPROFILE",
        "CSIDL_LOCAL_APPDATA": "LOCALAPPDATA",
    }.get(csidl_name)
    if env_var_name is None:
        raise ValueError(f"Unknown CSIDL name: {csidl_name}")
    result = os.environ.get(env_var_name)
    if result is None:
        raise ValueError(f"Unset environment variable: {env_var_name}")
    return result


def get_win_folder_from_registry(csidl_name: str) -> str:
    """Get folder from the registry.

    This is a fallback technique at best. I'm not sure if using the
    registry for this guarantees us the correct answer for all CSIDL_*
    names.
    """
    shell_folder_name = {
        "CSIDL_APPDATA": "AppData",
        "CSIDL_COMMON_APPDATA": "Common AppData",
        "CSIDL_LOCAL_APPDATA": "Local AppData",
        "CSIDL_PERSONAL": "Personal",
    }.get(csidl_name)
    if shell_folder_name is None:
        raise ValueError(f"Unknown CSIDL name: {csidl_name}")

    import winreg

    key = winreg.OpenKey(winreg.HKEY_CURRENT_USER, r"Software\Microsoft\Windows\CurrentVersion\Explorer\Shell Folders")
    directory, _ = winreg.QueryValueEx(key, shell_folder_name)
    return str(directory)


def get_win_folder_via_ctypes(csidl_name: str) -> str:
    """Get folder with ctypes."""
    csidl_const = {
        "CSIDL_APPDATA": 26,
        "CSIDL_COMMON_APPDATA": 35,
        "CSIDL_LOCAL_APPDATA": 28,
        "CSIDL_PERSONAL": 5,
    }.get(csidl_name)
    if csidl_const is None:
        raise ValueError(f"Unknown CSIDL name: {csidl_name}")

    buf = ctypes.create_unicode_buffer(1024)
    windll = getattr(ctypes, "windll")  # noqa: B009 # using getattr to avoid false positive with mypy type checker
    windll.shell32.SHGetFolderPathW(None, csidl_const, None, 0, buf)

    # Downgrade to short path name if it has highbit chars.
    if any(ord(c) > 255 for c in buf):
        buf2 = ctypes.create_unicode_buffer(1024)
        if windll.kernel32.GetShortPathNameW(buf.value, buf2, 1024):
            buf = buf2

    return buf.value


def _pick_get_win_folder() -> Callable[[str], str]:
    if hasattr(ctypes, "windll"):
        return get_win_folder_via_ctypes
    try:
        import winreg  # noqa: F401
    except ImportError:
        return get_win_folder_from_env_vars
    else:
        return get_win_folder_from_registry


get_win_folder = lru_cache(maxsize=None)(_pick_get_win_folder())

__all__ = [
    "Windows",
]
site-packages/pip/_vendor/platformdirs/__main__.py000064400000002164150564745010016322 0ustar00from pip._vendor.platformdirs import PlatformDirs, __version__

PROPS = (
    "user_data_dir",
    "user_config_dir",
    "user_cache_dir",
    "user_state_dir",
    "user_log_dir",
    "user_documents_dir",
    "user_runtime_dir",
    "site_data_dir",
    "site_config_dir",
)


def main() -> None:
    app_name = "MyApp"
    app_author = "MyCompany"

    print(f"-- platformdirs {__version__} --")

    print("-- app dirs (with optional 'version')")
    dirs = PlatformDirs(app_name, app_author, version="1.0")
    for prop in PROPS:
        print(f"{prop}: {getattr(dirs, prop)}")

    print("\n-- app dirs (without optional 'version')")
    dirs = PlatformDirs(app_name, app_author)
    for prop in PROPS:
        print(f"{prop}: {getattr(dirs, prop)}")

    print("\n-- app dirs (without optional 'appauthor')")
    dirs = PlatformDirs(app_name)
    for prop in PROPS:
        print(f"{prop}: {getattr(dirs, prop)}")

    print("\n-- app dirs (with disabled 'appauthor')")
    dirs = PlatformDirs(app_name, appauthor=False)
    for prop in PROPS:
        print(f"{prop}: {getattr(dirs, prop)}")


if __name__ == "__main__":
    main()
site-packages/pip/_vendor/distlib/__pycache__/scripts.cpython-39.pyc000064400000026021150564745010021502 0ustar00a

��Re8E�@s�ddlmZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
ddlmZddl
mZmZmZmZmZmZe�e�Zd��Ze�d�Zd	Zd
d�ZeZGdd
�d
e�ZdS)�)�BytesION�)�	sysconfig�detect_encoding�ZipFile)�finder)�FileOperator�get_export_entry�convert_path�get_executable�get_platform�in_venva�
<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
 <assemblyIdentity version="1.0.0.0"
 processorArchitecture="X86"
 name="%s"
 type="win32"/>

 <!-- Identify the application security requirements. -->
 <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
 <security>
 <requestedPrivileges>
 <requestedExecutionLevel level="asInvoker" uiAccess="false"/>
 </requestedPrivileges>
 </security>
 </trustInfo>
</assembly>s^#!.*pythonw?[0-9.]*([ 	].*)?$z�# -*- coding: utf-8 -*-
import re
import sys
from %(module)s import %(import_name)s
if __name__ == '__main__':
    sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
    sys.exit(%(func)s())
cCsXd|vrT|�d�rB|�dd�\}}d|vrT|�d�sTd||f}n|�d�sTd|}|S)N� z
/usr/bin/env r�"z%s "%s"z"%s")�
startswith�split)�
executable�env�_executable�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/scripts.py�enquote_executable3s

rc@s�eZdZdZeZdZd*dd�Zdd�Ze	j
�d	�rBd
d�Zdd
�Z
dd�Zd+dd�Zdd�ZeZdd�Zdd�ZdZdd�Zd,dd�Zdd�Zed d!��Zejd"d!��Zejd#ks�ejd	kr�ejd#kr�d$d%�Zd-d&d'�Z d.d(d)�Z!dS)/�ScriptMakerz_
    A class to copy or create scripts from source scripts or callable
    specifications.
    NTFcCs�||_||_||_d|_d|_tjdkp:tjdko:tjdk|_t	d�|_
|pRt|�|_tjdkprtjdkortjdk|_
tj|_dS)NF�posix�java)��X.Y�nt)�
source_dir�
target_dir�
add_launchers�force�clobber�os�name�_name�set_mode�set�variantsr�_fileop�_is_nt�sys�version_info)�selfrrr �dry_run�fileoprrr�__init__Ns�

�zScriptMaker.__init__cCs@|�dd�r<|jr<tj�|�\}}|�dd�}tj�||�}|S)N�guiF�python�pythonw)�getr*r#�pathr�replace�join)r-r�options�dn�fnrrr�_get_alternate_executable_s
z%ScriptMaker._get_alternate_executablerc	Csfz<t|�� }|�d�dkWd�WS1s00YWn$ttfy`t�d|�YdS0dS)zl
            Determine if the specified executable is a script
            (contains a #! line)
            �z#!NzFailed to open %sF)�open�read�OSError�IOError�logger�warning)r-r�fprrr�	_is_shellgs
2zScriptMaker._is_shellcCsD|�|�r*ddl}|jj�d�dkr<|Sn|���d�r<|Sd|S)Nrzos.name�Linuxz
jython.exez/usr/bin/env %s)rDr�lang�System�getProperty�lower�endswith)r-rrrrr�_fix_jython_executabless
z"ScriptMaker._fix_jython_executablecCs�tjdkrd}n8t|�t|�d}tjdkr4d}nd}d|voF||k}|r^d||d	}n d
}|d||d7}|d
7}|S)a
        Build a shebang line. In the simple case (on Windows, or a shebang line
        which is not too long or contains spaces) use a simple formulation for
        the shebang. Otherwise, use /bin/sh as the executable, with a contrived
        shebang which allows the script to run either under Python or sh, using
        suitable quoting. Thanks to Harald Nordgren for his input.

        See also: http://www.in-ulm.de/~mascheck/various/shebang/#length
                  https://hg.mozilla.org/mozilla-central/file/tip/mach
        rT��darwini�� s#!�
s
#!/bin/sh
s	'''exec' s "$0" "$@"
s' ''')r#r$�lenr+�platform)r-r�post_interp�simple_shebang�shebang_length�max_shebang_length�resultrrr�_build_shebangs

�zScriptMaker._build_shebang�cCszd}|jr|j}d}n�t��s&t�}nxt�rLtj�t�d�dt�	d��}nRtj�t�	d�dt�	d�t�	d�f�}tj�
|�s�tj�t�	d�dt�	d��}|r�|�||�}tj
�d	�r�|�|�}|r�t|�}|�d
�}tj
dkr�d|vr�d
|vr�|d7}|�||�}z|�d
�Wn t�y6td|��Yn0|d
k�rvz|�|�Wn$t�yttd||f��Yn0|S)NTF�scriptszpython%s�EXE�BINDIRz
python%s%s�VERSIONr�utf-8�cliz	-X:Framesz
-X:FullFramess
 -X:Framesz,The shebang (%r) is not decodable from utf-8z?The shebang (%r) is not decodable from the script encoding (%r))rr�is_python_buildrr
r#r5r7�get_path�get_config_var�isfiler;r+rRrrKr�encoderX�decode�UnicodeDecodeError�
ValueError)r-�encodingrSr8�enquoter�shebangrrr�_get_shebang�s`�
���


��

��
zScriptMaker._get_shebangcCs$|jt|j|j�d�d|jd�S)N�.r)�module�import_name�func)�script_template�dict�prefix�suffixr)r-�entryrrr�_get_script_text�s
�zScriptMaker._get_script_textcCstj�|�}|j|S�N)r#r5�basename�manifest)r-�exename�baserrr�get_manifest�szScriptMaker.get_manifestcCs|jo
|j}tj�d�}|�|�s*||7}|s8||}nn|dkrL|�d�}n
|�d�}t�}	t|	d��}
|
�	d|�Wd�n1s�0Y|	�
�}|||}|D�]R}tj�|j
|�}
|�r�tj�|
�\}}|�d�r�|}
d|
}
z|j�|
|�Wn�t�y�t�d�d	|
}tj�|��r8t�|�t�|
|�|j�|
|�t�d
�zt�|�Wnt�y~Yn0Yn0nl|j�r�|
�d|��s�d|
|f}
tj�|
��r�|j�s�t�d
|
�q�|j�|
|�|j�r�|j�|
g�|�|
�q�dS)Nr^�py�t�wz__main__.pyz.pyz%s.exez:Failed to write executable - trying to use .deleteme logicz%s.deletemez0Able to replace executable using .deleteme logicrlz%s.%szSkipping existing file %s)r r*r#�lineseprdrJ�
_get_launcherrr�writestr�getvaluer5r7r�splitextrr)�write_binary_file�	ExceptionrArB�exists�remove�rename�debugr"r&�set_executable_mode�append)r-�namesrj�script_bytes�	filenames�ext�use_launcherr�launcher�stream�zf�zip_datar$�outname�n�e�dfnamerrr�
_write_script�sX


*




zScriptMaker._write_script�-cCsnt�}d|jvr|�|�d|jvr<|�d||jdf�d|jvrj|�d||j|jd|jdf�|S)Nr�Xz%s%srrz	%s%s%s.%sr)r'r(�addr,�variant_separator)r-r$rWrrr�get_script_filenames$s



�z ScriptMaker.get_script_filenamesc
Cs�d}|r0|�dg�}|r0dd�|�}|�d�}|jd||d�}|�|��d�}|�|j�}|rr|�dd�rrd	}	nd
}	|�|||||	�dS)NrY�interpreter_argsz %srr^�r8r1F�pywr|)r4r7rdrkrur�r$r�)
r-rtr�r8rS�argsrj�script�scriptnamesr�rrr�_make_script/s
zScriptMaker._make_scriptcCs�d}tj�|jt|��}tj�|jtj�|��}|jsT|j�	||�sTt
�d|�dSzt|d�}Wnt
y�|jsx�d}YnF0|��}|s�t
�d|�dSt�|�dd��}|r�d}|�d�p�d	}|�s|r�|��|j�||�|jr�|j�|g�|�|�n�t
�d
||j�|jj�s|t|j�\}	}
|�d�|�|	|�}d|v�rTd
}nd}tj�|�}
|�|
g||��||�|�r�|��dS)NFznot copying %s (up-to-date)�rbz%s is an empty file (skipping)s
rPTrrYzcopying and adjusting %s -> %srspythonwr�r|) r#r5r7rr
rrwr!r)�newerrAr�r=r@r.�readlinerB�
FIRST_LINE_RE�matchr6�group�close�	copy_filer&r�r��infor�seekrkr�r>)r-r�r��adjustr��f�
first_liner�rSrh�linesrjr�r�rrr�_copy_script?sR
�


zScriptMaker._copy_scriptcCs|jjSrv�r)r.)r-rrrr.rszScriptMaker.dry_runcCs||j_dSrvr�)r-�valuerrrr.vsrcCstt�d�dkrd}nd}t�dkr&dnd}d|||f}t�d	d
�d}t|��|�}|snd||f}t|��|jS)
N�P�Z64Z32z	win-arm64z-armrz
%s%s%s.exerlrrz(Unable to find resource %s in package %s)	�struct�calcsizer�__name__�rsplitr�findrg�bytes)r-�kind�bitsZplatform_suffixr$Zdistlib_package�resource�msgrrrr�~s�zScriptMaker._get_launchercCs6g}t|�}|dur"|�||�n|j|||d�|S)a�
        Make a script.

        :param specification: The specification, which is either a valid export
                              entry specification (to make a script from a
                              callable) or a filename (to make a script by
                              copying from a source location).
        :param options: A dictionary of options controlling script generation.
        :return: A list of all absolute pathnames written to.
        Nr�)r	r�r�)r-�
specificationr8r�rtrrr�make�szScriptMaker.makecCs$g}|D]}|�|�||��q|S)z�
        Take a list of specifications and make scripts from them,
        :param specifications: A list of specifications.
        :return: A list of all absolute pathnames written to,
        )�extendr�)r-�specificationsr8r�r�rrr�
make_multiple�szScriptMaker.make_multiple)TFN)rYN)N)N)N)"r��
__module__�__qualname__�__doc__�SCRIPT_TEMPLATErprr0r;r+rRrrDrKrXrkru�_DEFAULT_MANIFESTrxr{r�r�r�r�r��propertyr.�setterr#r$r%r�r�r�rrrrrEs8�

E4
3


r) �ior�loggingr#�rer�r+�compatrrr�	resourcesr�utilrr	r
rrr
�	getLoggerr�rA�stripr��compiler�r�r�_enquote_executable�objectrrrrr�<module>s 


site-packages/pip/_vendor/distlib/__pycache__/database.cpython-39.pyc000064400000123027150564745010021563 0ustar00a

��Res��@s�dZddlmZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
mZddlmZddlmZmZddlmZmZmZmZdd	lmZmZmZmZmZmZm Z gd
�Z!e�"e#�Z$dZ%dZ&d
eddde%dfZ'dZ(Gdd�de)�Z*Gdd�de)�Z+Gdd�de)�Z,Gdd�de,�Z-Gdd�de-�Z.Gdd�de-�Z/e.Z0e/Z1Gdd �d e)�Z2d*d"d#�Z3d$d%�Z4d&d'�Z5d(d)�Z6dS)+zPEP 376 implementation.�)�unicode_literalsN�)�DistlibException�	resources)�StringIO)�
get_scheme�UnsupportedVersionError)�Metadata�METADATA_FILENAME�WHEEL_METADATA_FILENAME�LEGACY_METADATA_FILENAME)�parse_requirement�cached_property�parse_name_and_version�read_exports�
write_exports�	CSVReader�	CSVWriter)�Distribution�BaseInstalledDistribution�InstalledDistribution�EggInfoDistribution�DistributionPathzpydist-exports.jsonzpydist-commands.json�	INSTALLER�RECORD�	REQUESTED�	RESOURCES�SHAREDz
.dist-infoc@s(eZdZdZdd�Zdd�Zdd�ZdS)	�_CachezL
    A simple cache mapping names and .dist-info paths to distributions
    cCsi|_i|_d|_dS)zZ
        Initialise an instance. There is normally one for each DistributionPath.
        FN)�name�path�	generated��self�r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/database.py�__init__1sz_Cache.__init__cCs|j��|j��d|_dS)zC
        Clear the cache, setting it to its initial state.
        FN)r�clearr r!r"r$r$r%r'9s

z_Cache.clearcCs2|j|jvr.||j|j<|j�|jg��|�dS)z`
        Add a distribution to the cache.
        :param dist: The distribution to add.
        N)r r�
setdefault�key�append�r#�distr$r$r%�addAsz
_Cache.addN)�__name__�
__module__�__qualname__�__doc__r&r'r-r$r$r$r%r-src@s�eZdZdZddd�Zdd�Zdd	�Zeee�Zd
d�Z	dd
�Z
dd�Zedd��Z
dd�Zdd�Zddd�Zdd�Zddd�ZdS)rzU
    Represents a set of distributions installed on a path (typically sys.path).
    NFcCsD|durtj}||_d|_||_t�|_t�|_d|_td�|_	dS)a�
        Create an instance from a path, optionally including legacy (distutils/
        setuptools/distribute) distributions.
        :param path: The path to use, as a list of directories. If not specified,
                     sys.path is used.
        :param include_egg: If True, this instance will look for and return legacy
                            distributions as well as those based on PEP 376.
        NT�default)
�sysr �
_include_dist�_include_eggr�_cache�
_cache_egg�_cache_enabledr�_scheme)r#r Zinclude_eggr$r$r%r&Os	zDistributionPath.__init__cCs|jS�N�r8r"r$r$r%�_get_cache_enabledcsz#DistributionPath._get_cache_enabledcCs
||_dSr:r;)r#�valuer$r$r%�_set_cache_enabledfsz#DistributionPath._set_cache_enabledcCs|j��|j��dS)z,
        Clears the internal cache.
        N)r6r'r7r"r$r$r%�clear_cacheks
zDistributionPath.clear_cachec

csZt�}|jD�]F}t�|�}|dur&q|�d�}|r|js<qt|j�}|D�]}|�|�}|rJ|j|vrjqJ|j�r|�t	��rt
ttg}|D]"}t
�||�}	|�|	�}
|
r�q�q�qJt�|
����}t|dd�}Wd�n1s�0Yt�d|j�|�|j�t|j||d�VqJ|jrJ|�d�rJt�d|j�|�|j�t|j|�VqJqdS)zD
        Yield .dist-info and/or .egg(-info) distributions.
        N��legacy��fileobj�schemezFound %s)�metadata�env)�	.egg-info�.egg)�setr r�finder_for_path�find�is_container�sortedr4�endswith�DISTINFO_EXTr
rr�	posixpath�join�
contextlib�closing�	as_streamr	�logger�debugr-�new_dist_classr5�old_dist_class)
r#�seenr �finder�rZrset�entryZpossible_filenamesZmetadata_filenameZ
metadata_pathZpydist�streamrEr$r$r%�_yield_distributionsssF





�
*�
z%DistributionPath._yield_distributionscCsp|jj}|jo|jj}|s"|rl|��D](}t|t�rF|j�|�q*|j�|�q*|r`d|j_|rld|j_dS)zk
        Scan the path for distributions and populate the cache with
        those that are found.
        TN)r6r!r5r7r^�
isinstancerr-)r#Zgen_distZgen_eggr,r$r$r%�_generate_cache�s

z DistributionPath._generate_cachecCs|�dd�}d�||g�tS)ao
        The *name* and *version* parameters are converted into their
        filename-escaped form, i.e. any ``'-'`` characters are replaced
        with ``'_'`` other than the one in ``'dist-info'`` and the one
        separating the name from the version number.

        :parameter name: is converted to a standard distribution name by replacing
                         any runs of non- alphanumeric characters with a single
                         ``'-'``.
        :type name: string
        :parameter version: is converted to a standard version string. Spaces
                            become dots, and all other non-alphanumeric characters
                            (except dots) become dashes, with runs of multiple
                            dashes condensed to a single dash.
        :type version: string
        :returns: directory name
        :rtype: string�-�_)�replacerQrO)�clsr�versionr$r$r%�distinfo_dirname�sz!DistributionPath.distinfo_dirnameccs^|js|��D]
}|Vqn>|��|jj��D]
}|Vq0|jrZ|jj��D]
}|VqNdS)a5
        Provides an iterator that looks for distributions and returns
        :class:`InstalledDistribution` or
        :class:`EggInfoDistribution` instances for each one of them.

        :rtype: iterator of :class:`InstalledDistribution` and
                :class:`EggInfoDistribution` instances
        N)r8r^r`r6r �valuesr5r7r+r$r$r%�get_distributions�s	
z"DistributionPath.get_distributionscCs�d}|��}|js4|��D]}|j|kr|}q|qnH|��||jjvrZ|jj|d}n"|jr|||jjvr||jj|d}|S)a=
        Looks for a named distribution on the path.

        This function only returns the first result found, as no more than one
        value is expected. If nothing is found, ``None`` is returned.

        :rtype: :class:`InstalledDistribution`, :class:`EggInfoDistribution`
                or ``None``
        Nr)	�lowerr8r^r)r`r6rr5r7)r#r�resultr,r$r$r%�get_distribution�s

z!DistributionPath.get_distributionc	cs�d}|durHz|j�d||f�}Wn"tyFtd||f��Yn0|��D]p}t|d�slt�d|�qP|j}|D]H}t	|�\}}|dur�||kr�|VqPqv||krv|�
|�rv|VqPqvqPdS)a
        Iterates over all distributions to find which distributions provide *name*.
        If a *version* is provided, it will be used to filter the results.

        This function only returns the first result found, since no more than
        one values are expected. If the directory is not found, returns ``None``.

        :parameter version: a version specifier that indicates the version
                            required, conforming to the format in ``PEP-345``

        :type name: string
        :type version: string
        N�%s (%s)zinvalid name or version: %r, %r�provideszNo "provides": %s)r9�matcher�
ValueErrorrrh�hasattrrUrVrmr�match)	r#rrernr,�provided�p�p_name�p_verr$r$r%�provides_distribution�s*�
z&DistributionPath.provides_distributioncCs(|�|�}|durtd|��|�|�S)z5
        Return the path to a resource file.
        Nzno distribution named %r found)rk�LookupError�get_resource_path)r#r�
relative_pathr,r$r$r%�
get_file_path!s
zDistributionPath.get_file_pathccsX|��D]J}|j}||vr||}|dur>||vrR||Vq|��D]
}|VqFqdS)z�
        Return all of the exported entries in a particular category.

        :param category: The category to search for entries.
        :param name: If specified, only entries with that name are returned.
        N)rh�exportsrg)r#�categoryrr,r[�d�vr$r$r%�get_exported_entries*sz%DistributionPath.get_exported_entries)NF)N)N)r.r/r0r1r&r<r>�propertyZ
cache_enabledr?r^r`�classmethodrfrhrkrvrzrr$r$r$r%rKs

,

)	rc@s�eZdZdZdZdZdd�Zedd��ZeZ	edd��Z
ed	d
��Zdd�Zed
d��Z
edd��Zedd��Zedd��Zedd��Zdd�Zdd�Zdd�Zdd�ZdS) rz�
    A base class for distributions, whether installed or from indexes.
    Either way, it must have some metadata, so that's all that's needed
    for construction.
    FcCsL||_|j|_|j��|_|j|_d|_d|_d|_d|_t	�|_
i|_dS)z�
        Initialise an instance.
        :param metadata: The instance of :class:`Metadata` describing this
        distribution.
        N)rErrir)re�locator�digest�extras�contextrIZ
download_urls�digests)r#rEr$r$r%r&OszDistribution.__init__cCs|jjS)zH
        The source archive download URL for this distribution.
        )rE�
source_urlr"r$r$r%r�`szDistribution.source_urlcCsd|j|jfS)zX
        A utility property which displays the name and version in parentheses.
        rl�rrer"r$r$r%�name_and_versioniszDistribution.name_and_versioncCs.|jj}d|j|jf}||vr*|�|�|S)z�
        A set of distribution names and versions provided by this distribution.
        :return: A set of "name (version)" strings.
        rl)rErmrrer*)r#�plist�sr$r$r%rmps

zDistribution.providescCs8|j}t�d|���t||�}t|j||j|jd��S)Nz%Getting requirements from metadata %r)r�rF)	rErUrVZtodict�getattrrI�get_requirementsr�r�)r#Zreq_attr�mdZreqtsr$r$r%�_get_requirements|s
�zDistribution._get_requirementscCs
|�d�S)N�run_requires�r�r"r$r$r%r��szDistribution.run_requirescCs
|�d�S)N�
meta_requiresr�r"r$r$r%r��szDistribution.meta_requirescCs
|�d�S)N�build_requiresr�r"r$r$r%r��szDistribution.build_requirescCs
|�d�S)N�
test_requiresr�r"r$r$r%r��szDistribution.test_requirescCs
|�d�S)N�dev_requiresr�r"r$r$r%r��szDistribution.dev_requiresc
	Cs�t|�}t|jj�}z|�|j�}Wn4tyXt�d|�|�	�d}|�|�}Yn0|j
}d}|jD]B}t|�\}}	||kr�qjz|�
|	�}Wq�Wqjty�Yqj0qj|S)z�
        Say if this instance matches (fulfills) a requirement.
        :param req: The requirement to match.
        :rtype req: str
        :return: True if it matches, else False.
        �+could not read version %r - using name onlyrF)r
rrErDrn�requirementrrU�warning�splitr)rmrrq)
r#�reqr[rDrnrrjrsrtrur$r$r%�matches_requirement�s,	�


z Distribution.matches_requirementcCs(|jrd|j}nd}d|j|j|fS)zC
        Return a textual representation of this instance,
        z [%s]r@z<Distribution %s (%s)%s>)r�rre)r#�suffixr$r$r%�__repr__�szDistribution.__repr__cCs>t|�t|�urd}n$|j|jko8|j|jko8|j|jk}|S)a<
        See if this distribution is the same as another.
        :param other: The distribution to compare with. To be equal to one
                      another. distributions must have the same type, name,
                      version and source_url.
        :return: True if it is the same, else False.
        F)�typerrer�)r#�otherrjr$r$r%�__eq__�s
�
�zDistribution.__eq__cCst|j�t|j�t|j�S)zH
        Compute hash in a way which matches the equality test.
        )�hashrrer�r"r$r$r%�__hash__�szDistribution.__hash__N)r.r/r0r1Zbuild_time_dependency�	requestedr&r�r��download_urlr�rmr�r�r�r�r�r�r�r�r�r�r$r$r$r%r=s4







"
rcs0eZdZdZdZd�fdd�	Zddd�Z�ZS)	rz]
    This is the base class for installed distributions (whether PEP 376 or
    legacy).
    Ncs tt|��|�||_||_dS)a
        Initialise an instance.
        :param metadata: An instance of :class:`Metadata` which describes the
                         distribution. This will normally have been initialised
                         from a metadata file in the ``path``.
        :param path:     The path of the ``.dist-info`` or ``.egg-info``
                         directory for the distribution.
        :param env:      This is normally the :class:`DistributionPath`
                         instance where this distribution was found.
        N)�superrr&r �	dist_path)r#rEr rF��	__class__r$r%r&�sz"BaseInstalledDistribution.__init__cCsd|dur|j}|dur"tj}d}ntt|�}d|j}||���}t�|��d��d�}d||fS)a�
        Get the hash of some data, using a particular hash algorithm, if
        specified.

        :param data: The data to be hashed.
        :type data: bytes
        :param hasher: The name of a hash implementation, supported by hashlib,
                       or ``None``. Examples of valid values are ``'sha1'``,
                       ``'sha224'``, ``'sha384'``, '``sha256'``, ``'md5'`` and
                       ``'sha512'``. If no hasher is specified, the ``hasher``
                       attribute of the :class:`InstalledDistribution` instance
                       is used. If the hasher is determined to be ``None``, MD5
                       is used as the hashing algorithm.
        :returns: The hash of the data. If a hasher was explicitly specified,
                  the returned hash will be prefixed with the specified hasher
                  followed by '='.
        :rtype: str
        Nr@z%s=�=�ascii�%s%s)	�hasher�hashlib�md5r�r��base64�urlsafe_b64encode�rstrip�decode)r#�datar��prefixr�r$r$r%�get_hash�s

z"BaseInstalledDistribution.get_hash)N)N)r.r/r0r1r�r&r��
__classcell__r$r$r�r%r�srcs�eZdZdZdZd'�fdd�	Zdd�Zdd	�Zd
d�Ze	dd
��Z
dd�Zdd�Zdd�Z
dd�Zd(dd�Zdd�Ze	dd��Zd)dd�Zdd �Zd!d"�Zd#d$�Zd%d&�ZejZ�ZS)*ra
    Created with the *path* of the ``.dist-info`` directory provided to the
    constructor. It reads the metadata contained in ``pydist.json`` when it is
    instantiated., or uses a passed in Metadata instance (useful for when
    dry-run mode is being used).
    �sha256Nc
s~g|_t�|�|_}|dur*td|��|rP|jrP||jjvrP|jj|j}n�|dur�|�	t
�}|durt|�	t�}|dur�|�	t�}|dur�tdt
|f��t
�|����}t|dd�}Wd�n1s�0Ytt|��|||�|�r|j�r|j�|�|�	d�}|du|_tj�|d�}tj�|��rzt|d��}|���d�}	Wd�n1�sf0Y|	��|_dS)	Nzfinder unavailable for %szno %s found in %srArBr�
top_level.txt�rb�utf-8)�modulesrrJrZror8r6r rErKr
rrrRrSrTr	r�rr&r-r��osrQ�exists�open�readr��
splitlines)
r#r rErFrZr[r]rs�fr�r�r$r%r&s8


�*

.zInstalledDistribution.__init__cCsd|j|j|jfS)Nz#<InstalledDistribution %r %s at %r>�rrer r"r$r$r%r�=s�zInstalledDistribution.__repr__cCsd|j|jfS�Nz%s %sr�r"r$r$r%�__str__AszInstalledDistribution.__str__c
	Cs�g}|�d�}t�|����z}t|d��P}|D]:}dd�tt|�d�D�}||\}}}	|�|||	f�q.Wd�n1s~0YWd�n1s�0Y|S)a"
        Get the list of installed files for the distribution
        :return: A list of tuples of path, hash and size. Note that hash and
                 size might be ``None`` for some entries. The path is exactly
                 as stored in the file (which is as in PEP 376).
        r�r]cSsg|]}d�qSr:r$)�.0�ir$r$r%�
<listcomp>S�z6InstalledDistribution._get_records.<locals>.<listcomp>�N)�get_distinfo_resourcerRrSrTr�range�lenr*)
r#�resultsr[r]Z
record_reader�row�missingr �checksum�sizer$r$r%�_get_recordsDs
Nz"InstalledDistribution._get_recordscCsi}|�t�}|r|��}|S)a
        Return the information exported by this distribution.
        :return: A dictionary of exports, mapping an export category to a dict
                 of :class:`ExportEntry` instances describing the individual
                 export entries, and keyed by name.
        )r��EXPORTS_FILENAMEr)r#rjr[r$r$r%r{[s

zInstalledDistribution.exportscCsLi}|�t�}|rHt�|����}t|�}Wd�n1s>0Y|S)z�
        Read exports data from a file in .ini format.

        :return: A dictionary of exports, mapping an export category to a list
                 of :class:`ExportEntry` instances describing the individual
                 export entries.
        N)r�r�rRrSrTr)r#rjr[r]r$r$r%ris
&z"InstalledDistribution.read_exportscCsB|�t�}t|d��}t||�Wd�n1s40YdS)a
        Write a dictionary of exports to a file in .ini format.
        :param exports: A dictionary of exports, mapping an export category to
                        a list of :class:`ExportEntry` instances describing the
                        individual export entries.
        �wN)�get_distinfo_filer�r�r)r#r{�rfr�r$r$r%rxs
z#InstalledDistribution.write_exportsc	Cs�|�d�}t�|����t}t|d��J}|D]4\}}||kr*|Wd�Wd�Sq*Wd�n1st0YWd�n1s�0Ytd|��dS)aW
        NOTE: This API may change in the future.

        Return the absolute path to a resource file with the given relative
        path.

        :param relative_path: The path, relative to .dist-info, of the resource
                              of interest.
        :return: The absolute path where the resource is to be found.
        rr�Nz3no resource file with relative path %r is installed)r�rRrSrTr�KeyError)r#ryr[r]Zresources_reader�relative�destinationr$r$r%rx�s
b�z'InstalledDistribution.get_resource_pathccs|��D]
}|VqdS)z�
        Iterates over the ``RECORD`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: iterator of (path, hash, size)
        N)r�)r#rjr$r$r%�list_installed_files�sz*InstalledDistribution.list_installed_filesFc
CsRtj�|d�}tj�|j�}|�|�}tj�|d�}|�d�}t�d|�|rRdSt|���}|D]�}tj�	|�sz|�
d�r�d}	}
nHdtj�|�}
t|d��}|�
|���}	Wd�n1s�0Y|�|�s�|r�|�|�r�tj�||�}|�||	|
f�q`|�|��rtj�||�}|�|ddf�Wd�n1�sD0Y|S)z�
        Writes the ``RECORD`` file, using the ``paths`` iterable passed in. Any
        existing ``RECORD`` file is silently overwritten.

        prefix is used to determine when to write absolute paths.
        r@r�creating %sN�z.pycz.pyoz%dr�)r�r rQ�dirname�
startswithr�rU�infor�isdirrN�getsizer�r�r��relpath�writerow)r#�pathsr��dry_run�baseZbase_under_prefix�record_path�writerr �
hash_valuer��fpr$r$r%�write_installed_files�s0



,�0z+InstalledDistribution.write_installed_filesc	Cs0g}tj�|j�}|�d�}|��D�]\}}}tj�|�sJtj�||�}||krTq$tj�|�st|�|dddf�q$tj�	|�r$t
tj�|��}|r�||kr�|�|d||f�q$|r$d|vr�|�dd�d}nd	}t
|d
��<}	|�|	��|�}
|
|k�r
|�|d||
f�Wd	�q$1�s 0Yq$|S)�
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        rr�TFr��=rrNr�r�)r�r r�r�r��isabsrQr�r*�isfile�strr�r�r�r�r�)r#�
mismatchesr�r�r r�r�Zactual_sizer�r�Zactual_hashr$r$r%�check_installed_files�s.	

4z+InstalledDistribution.check_installed_filescCs�i}tj�|jd�}tj�|�r�tj|ddd��}|����}Wd�n1sR0Y|D]8}|�dd�\}}|dkr�|�	|g��
|�q`|||<q`|S)	a�
        A dictionary of shared locations whose keys are in the set 'prefix',
        'purelib', 'platlib', 'scripts', 'headers', 'data' and 'namespace'.
        The corresponding value is the absolute path of that category for
        this distribution, and takes into account any paths selected by the
        user at installation time (e.g. via command-line arguments). In the
        case of the 'namespace' key, this would be a list of absolute paths
        for the roots of namespace packages in this distribution.

        The first time this property is accessed, the relevant information is
        read from the SHARED file in the .dist-info directory.
        rr[r���encodingNr�r�	namespace)r�r rQr��codecsr�r�r�r�r(r*)r#rj�shared_pathr��lines�liner)r=r$r$r%�shared_locations�s*
z&InstalledDistribution.shared_locationsc	Cs�tj�|jd�}t�d|�|r$dSg}dD].}||}tj�||�r,|�d||f�q,|�dd�D]}|�d|�qhtj	|d	d
d�� }|�
d�|��Wd�n1s�0Y|S)
aa
        Write shared location information to the SHARED file in .dist-info.
        :param paths: A dictionary as described in the documentation for
        :meth:`shared_locations`.
        :param dry_run: If True, the action is logged but no file is actually
                        written.
        :return: The path of the file written to.
        rr�N)r��lib�headers�scriptsr�z%s=%sr�r$znamespace=%sr�r�r��
)r�r rQrUr�r�r*�getr�r��write)	r#r�r�r�r�r)r �nsr�r$r$r%�write_shared_locationss	.z,InstalledDistribution.write_shared_locationscCsF|tvrtd||jf��t�|j�}|dur<td|j��|�|�S)N�+invalid path for a dist-info file: %r at %rzUnable to get a finder for %s)�
DIST_FILESrr rrJrK)r#r rZr$r$r%r�s�z+InstalledDistribution.get_distinfo_resourcecCs~|�tj�dkrT|�tj�dd�\}}||j�tj�dkrTtd||j|jf��|tvrntd||jf��tj�	|j|�S)a�
        Returns a path located under the ``.dist-info`` directory. Returns a
        string representing the path.

        :parameter path: a ``'/'``-separated path relative to the
                         ``.dist-info`` directory or an absolute path;
                         If *path* is an absolute path and doesn't start
                         with the ``.dist-info`` directory path,
                         a :class:`DistlibException` is raised
        :type path: str
        :rtype: str
        r���N���z;dist-info file %r does not belong to the %r %s distributionr)
rKr��sepr�r rrrerrQ)r#r rfr$r$r%r�$s���z'InstalledDistribution.get_distinfo_fileccsRtj�|j�}|��D]6\}}}tj�|�s:tj�||�}|�|j�r|VqdS)z�
        Iterates over the ``RECORD`` entries and returns paths for each line if
        the path is pointing to a file located in the ``.dist-info`` directory
        or one of its subdirectories.

        :returns: iterator of paths
        N)r�r r�r�r�rQr�)r#r�r r�r�r$r$r%�list_distinfo_filesAsz)InstalledDistribution.list_distinfo_filescCst|t�o|j|jkSr:)r_rr �r#r�r$r$r%r�Qs

�zInstalledDistribution.__eq__)NN)F)F)r.r/r0r1r�r&r�r�r�rr{rrrxr�r�r�r�rr�r�r	r��objectr�r�r$r$r�r%rs,"



##

	rcsjeZdZdZdZiZd�fdd�	Zdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zddd�Zdd�Z
ejZ�ZS)raCreated with the *path* of the ``.egg-info`` directory or file provided
    to the constructor. It reads the metadata contained in the file itself, or
    if the given path happens to be a directory, the metadata is read from the
    file ``PKG-INFO`` under that directory.TNcs�dd�}||_||_|rJ|jrJ||jjvrJ|jj|j}|||j|j�n0|�|�}|||j|j�|rz|jrz|j�|�t	t
|��|||�dS)NcSs||_|��|_||_dSr:)rrir)re)r��nr~r$r$r%�set_name_and_versioncs
z:EggInfoDistribution.__init__.<locals>.set_name_and_version)r r�r8r7rErre�
_get_metadatar-r�rr&)r#r rFr
rEr�r$r%r&bs

zEggInfoDistribution.__init__cs�d}dd���fdd�}d}}|�d�r�tj�|�r�tj�|d�}tj�|d�}t|dd	�}tj�|d
�}	tj�|d�}||	�}nnt�|�}
t|
�	d��
d
��}t|dd�}z,|
�	d�}|
�	d��
d�}�|�
d��}Wnty�d}Yn0nf|�d��rNtj�|��r@tj�|d
�}	||	�}tj�|d�}tj�|d�}t|dd	�}ntd|��|�rj|�
|�|du�r�|du�r�tj�|��r�t|d��}
|
���
d�}Wd�n1�s�0Y|�s�g}n|��}||_|S)NcSs�g}|��}|D]�}|��}|�d�r6t�d|�q�t|�}|sPt�d|�q|jr`t�d�|jst|�|j	�qd�
dd�|jD��}|�d|j	|f�q|S)	z�Create a list of dependencies from a requires.txt file.

            *data*: the contents of a setuptools-produced requires.txt file.
            �[z.Unexpected line: quitting requirement scan: %rz#Not recognised as a requirement: %rz4extra requirements in requires.txt are not supportedz, css|]}d|VqdS)r�Nr$)r��cr$r$r%�	<genexpr>�r�zQEggInfoDistribution._get_metadata.<locals>.parse_requires_data.<locals>.<genexpr>rl)r��stripr�rUr�r
r��constraintsr*rrQ)r��reqsr�r�r[Zconsr$r$r%�parse_requires_datazs(
�
z>EggInfoDistribution._get_metadata.<locals>.parse_requires_datacsZg}z>t�|dd��}�|���}Wd�n1s60YWntyTYn0|S)z�Create a list of dependencies from a requires.txt file.

            *req_path*: the path to a setuptools-produced requires.txt file.
            r[r�N)r�r�r��IOError)�req_pathrr��rr$r%�parse_requires_path�s.z>EggInfoDistribution._get_metadata.<locals>.parse_requires_pathrHzEGG-INFOzPKG-INFOrA)r rDzrequires.txtr�zEGG-INFO/PKG-INFO�utf8rBzEGG-INFO/requires.txtzEGG-INFO/top_level.txtr�rGz,path must end with .egg-info or .egg, got %rr�)rNr�r r�rQr	�	zipimport�zipimporterr�get_datar�rrZadd_requirementsr�r�r�r�r�)r#r �requiresrZtl_pathZtl_datars�	meta_pathrErZzipfrCr�r�r$rr%rwsX


�
�

.z!EggInfoDistribution._get_metadatacCsd|j|j|jfS)Nz!<EggInfoDistribution %r %s at %r>r�r"r$r$r%r��s�zEggInfoDistribution.__repr__cCsd|j|jfSr�r�r"r$r$r%r��szEggInfoDistribution.__str__cCs`g}tj�|jd�}tj�|�r\|��D]2\}}}||kr<q(tj�|�s(|�|dddf�q(|S)r��installed-files.txtr�TF)r�r rQr�r�r*)r#r�r�r rbr$r$r%r��s	z)EggInfoDistribution.check_installed_filescCs�dd�}dd�}tj�|jd�}g}tj�|�r�tj|ddd���}|D]l}|��}tj�tj�|j|��}tj�|�s�t�	d	|�|�
d
�r�qFtj�|�sF|�|||�||�f�qFWd�n1s�0Y|�|ddf�|S)z�
        Iterates over the ``installed-files.txt`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: a list of (path, hash, size)
        cSs8t|d�}z|��}W|��n
|��0t�|���S)Nr�)r�r��closer�r��	hexdigest)r r��contentr$r$r%�_md5�s


z6EggInfoDistribution.list_installed_files.<locals>._md5cSst�|�jSr:)r��stat�st_size)r r$r$r%�_size�sz7EggInfoDistribution.list_installed_files.<locals>._sizer r[r�r�zNon-existent file: %sr�N)
r�r rQr�r�r�r�normpathrUr�rNr�r*)r#r$r'r�rjr�r�rsr$r$r%r��s"
8z(EggInfoDistribution.list_installed_filesFccs�tj�|jd�}tj�|�r�d}tj|ddd��j}|D]T}|��}|dkrPd}q6|s6tj�tj�|j|��}|�|j�r6|r�|Vq6|Vq6Wd�n1s�0YdS)	a
        Iterates over the ``installed-files.txt`` entries and returns paths for
        each line if the path is pointing to a file located in the
        ``.egg-info`` directory or one of its subdirectories.

        :parameter absolute: If *absolute* is ``True``, each returned path is
                          transformed into a local absolute path. Otherwise the
                          raw value from ``installed-files.txt`` is returned.
        :type absolute: boolean
        :returns: iterator of paths
        r Tr[r�r�z./FN)	r�r rQr�r�r�rr(r�)r#�absoluter��skipr�r�rsr$r$r%r	sz'EggInfoDistribution.list_distinfo_filescCst|t�o|j|jkSr:)r_rr r
r$r$r%r�.s

�zEggInfoDistribution.__eq__)N)F)r.r/r0r1r�r�r&rr�r�r�r�r	r�rr�r�r$r$r�r%rYsZ&
rc@s^eZdZdZdd�Zdd�Zddd�Zd	d
�Zdd�Zddd�Z	ddd�Z
dd�Zdd�ZdS)�DependencyGrapha�
    Represents a dependency graph between distributions.

    The dependency relationships are stored in an ``adjacency_list`` that maps
    distributions to a list of ``(other, label)`` tuples where  ``other``
    is a distribution and the edge is labeled with ``label`` (i.e. the version
    specifier, if such was provided). Also, for more efficient traversal, for
    every distribution ``x``, a list of predecessors is kept in
    ``reverse_list[x]``. An edge from distribution ``a`` to
    distribution ``b`` means that ``a`` depends on ``b``. If any missing
    dependencies are found, they are stored in ``missing``, which is a
    dictionary that maps distributions to a list of requirements that were not
    provided by any other distributions.
    cCsi|_i|_i|_dSr:)�adjacency_list�reverse_listr�r"r$r$r%r&IszDependencyGraph.__init__cCsg|j|<g|j|<dS)z�Add the *distribution* to the graph.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        N)r,r-)r#�distributionr$r$r%�add_distributionNs
z DependencyGraph.add_distributionNcCs6|j|�||f�||j|vr2|j|�|�dS)a�Add an edge from distribution *x* to distribution *y* with the given
        *label*.

        :type x: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type y: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type label: ``str`` or ``None``
        N)r,r*r-)r#�x�y�labelr$r$r%�add_edgeXs
zDependencyGraph.add_edgecCs&t�d||�|j�|g��|�dS)a
        Add a missing *requirement* for the given *distribution*.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        :type requirement: ``str``
        z
%s missing %rN)rUrVr�r(r*)r#r.r�r$r$r%�add_missinggszDependencyGraph.add_missingcCsd|j|jfSr�r�r+r$r$r%�
_repr_distrszDependencyGraph._repr_distrcCs�|�|�g}|j|D]h\}}|�|�}|dur<d||f}|�d|t|��|�||d�}|�d�}|�|dd��qd�|�S)zPrints only a subgraphNz%s [%s]z    rr�)r5r,r*r��	repr_noder��extendrQ)r#r,�level�outputr�r2Z	suboutput�subsr$r$r%r6us

zDependencyGraph.repr_nodeTcCs�g}|�d�|j��D]h\}}t|�dkr:|s:|�|�|D]@\}}|durh|�d|j|j|f�q>|�d|j|jf�q>q|s�t|�dkr�|�d�|�d�|�d�|D]}|�d	|j�|�d
�q�|�d�|�d�dS)a9Writes a DOT output for the graph to the provided file *f*.

        If *skip_disconnected* is set to ``True``, then all distributions
        that are not dependent on any other distribution are skipped.

        :type f: has to support ``file``-like operations
        :type skip_disconnected: ``bool``
        zdigraph dependencies {
rNz"%s" -> "%s" [label="%s"]
z
"%s" -> "%s"
zsubgraph disconnected {
zlabel = "Disconnected"
zbgcolor = red
z"%s"r�z}
)rr,�itemsr�r*r)r#r�Zskip_disconnected�disconnectedr,�adjsr�r2r$r$r%�to_dot�s(	

�



zDependencyGraph.to_dotcs�g}i}|j��D]\}}|dd�||<qg�t|���dd�D]\}}|sD��|�||=qD�shq�|��D]\}}�fdd�|D�||<qpt�ddd��D��|���q,|t|���fS)aa
        Perform a topological sort of the graph.
        :return: A tuple, the first element of which is a topologically sorted
                 list of distributions, and the second element of which is a
                 list of distributions that cannot be sorted because they have
                 circular dependencies and so form a cycle.
        Ncs g|]\}}|�vr||f�qSr$r$)r�r}r[��	to_remover$r%r��r�z4DependencyGraph.topological_sort.<locals>.<listcomp>zMoving to result: %scSsg|]}d|j|jf�qS)rlr�)r�r}r$r$r%r��r�)r,r;�listr*rUrVr7�keys)r#rj�alist�kr~r$r?r%�topological_sort�s$
�z DependencyGraph.topological_sortcCs2g}|j��D]\}}|�|�|��qd�|�S)zRepresentation of the graphr�)r,r;r*r6rQ)r#r9r,r=r$r$r%r��szDependencyGraph.__repr__)N)r)T)
r.r/r0r1r&r/r3r4r5r6r>rEr�r$r$r$r%r+9s



 r+r2cCsRt|�}t�}i}|D]L}|�|�|jD]6}t|�\}}t�d|||�|�|g��||f�q*q|D]�}|j	|j
B|jB|jB}|D]�}	z|�
|	�}
Wn4ty�t�d|	�|	��d}|�
|�}
Yn0|
j}d}||v�r:||D]L\}}z|
�|�}
Wnt�yd}
Yn0|
r�|�|||	�d}�q:q�|s�|�||	�q�qh|S)a6Makes a dependency graph from the given distributions.

    :parameter dists: a list of distributions
    :type dists: list of :class:`distutils2.database.InstalledDistribution` and
                 :class:`distutils2.database.EggInfoDistribution` instances
    :rtype: a :class:`DependencyGraph` instance
    zAdd to provided: %s, %s, %sr�rFT)rr+r/rmrrUrVr(r*r�r�r�r�rnrr�r�r)rqr3r4)�distsrD�graphrrr,rsrrerr�rn�matched�providerrqr$r$r%�
make_graph�sN


���

rJcCsv||vrtd|j��t|�}|g}|j|}|rh|��}|�|�|j|D]}||vrN|�|�qNq.|�d�|S)z�Recursively generate a list of distributions from *dists* that are
    dependent on *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    �1given distribution %r is not a member of the listr)rrrJr-�popr*)rFr,rG�dep�todor}�succr$r$r%�get_dependent_dists�s�


rPcCsn||vrtd|j��t|�}g}|j|}|rj|��d}|�|�|j|D]}||vrP|�|�qPq,|S)z�Recursively generate a list of distributions from *dists* that are
    required by *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    rKr)rrrJr,rLr*)rFr,rGr�rNr}�predr$r$r%�get_required_distss�

rRcKs8|�dd�}tfi|��}||_||_|p,d|_t|�S)zO
    A convenience method for making a dist given just a name and version.
    �summaryzPlaceholder for summary)rLr	rrerSr)rre�kwargsrSr�r$r$r%�	make_dist2s
rU)r2)7r1�
__future__rr�r�rRr��loggingr�rPr3rr@rr�compatrrerrrEr	r
rr�utilr
rrrrrr�__all__�	getLoggerr.rUr�ZCOMMANDS_FILENAMErrOrrrrrrrrWrXr+rJrPrRrUr$r$r$r%�<module>sT$
�s7J]
6site-packages/pip/_vendor/distlib/__pycache__/util.cpython-39.pyc000064400000146644150564745010021006 0ustar00a

��Re��@s�ddlZddlmZddlZddlZddlmZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZzddlZWney�dZYn0ddlZddlZddlZddlZddlZzddlZWney�ddlZYn0ddlZddlmZddlmZmZmZmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0e
�1e2�Z3e
�4d�Z5e
�4d�Z6e
�4d	�Z7e
�4d
�Z8e
�4d�Z9e
�4d�Z:e
�4d
�Z;e
�4d�Z<dd�Z=dd�Z>dd�Z?dd�Z@dd�ZAd�dd�ZBdd�ZCdd�ZDdd �ZEejFd!d"��ZGejFd#d$��ZHejFd�d&d'��ZIGd(d)�d)eJ�ZKd*d+�ZLGd,d-�d-eJ�ZMd.d/�ZNGd0d1�d1eJ�ZOe
�4d2e
jP�ZQd3d4�ZRd�d5d6�ZSd7d8�ZTd9d:�ZUd;d<�ZVd=d>�ZWd?d@�ZXe
�4dAe
jY�ZZe
�4dB�Z[d�dCdD�Z\e
�4dE�Z]dFdG�Z^dHdI�Z_dJdK�Z`dLZadMdN�ZbdOdP�ZcGdQdR�dReJ�ZdGdSdT�dTeJ�ZeGdUdV�dVeJ�ZfdWZgd�dYdZ�Zhd[d\�Zid]ZjGd^d_�d_eJ�Zke
�4d`�Zle
�4da�Zme
�4db�Zndcdd�Zdedf�Zoe�r�ddglmpZqmrZrmsZsGdhdi�die$jt�ZtGdjdk�dkeq�ZpGdldm�dmepe'�Zuejvddn�Zwewdok�rGdpdq�dqe$jx�Zxe�rGdrds�dse$jy�ZyGdtdu�due%jz�Zze�r>Gdvdw�dwe%j{�Z{Gdxdy�dye%j|�Z|dzd{�Z}Gd|d}�d}eJ�Z~Gd~d�de~�ZGd�d��d�e~�Z�Gd�d��d�e(�Z�Gd�d��d�eJ�Z�d�d��Z�Gd�d��d�eJ�Z�d�d��Z�d�d��Z�d�d��Z�d�d�d�d��Z�d�d��Z�dS)��N)�deque)�iglob�)�DistlibException)�string_types�	text_type�shutil�	raw_input�StringIO�cache_from_source�urlopen�urljoin�httplib�	xmlrpclib�	splittype�HTTPHandler�BaseConfigurator�valid_ident�	Container�configparser�URLError�ZipFile�fsdecode�unquote�urlparsez^([\w\.-]+)\s*z^([\w\.*+-]+)\s*z^(<=?|>=?|={2,3}|[~!]=)\s*z*^((<=?)|(>=?)|={2,3}|[~!]=|in|not\s+in)\s*z^or\b\s*z	^and\b\s*z(\S+)\s*z(([\s\w\.{}()*+#:;,/?!~`@$%^&=|<>\[\]-]+)cs6dd����fdd���fdd���fdd���|�S)	ae
    Parse a marker string and return a dictionary containing a marker expression.

    The dictionary will contain keys "op", "lhs" and "rhs" for non-terminals in
    the expression grammar, or strings. A string contained in quotes is to be
    interpreted as a literal string, and a string not contained in quotes is a
    variable (such as os_name).
    cSs.t�|�}|r,|��d}||��d�}n�|s:td��n�|d}|dvrVtd|��d�|d�}|dd�}|g}|r�|d|kr��qqt|d|kr�|�|�|dd�}qtt�|�}|s�td|��|�|��d�||��d�}qtd�|�}td|��|�|�d�|�}|dd��	�}||fS)	Nrzunexpected end of inputz'"zinvalid expression: %s�rzerror in string literal: %szunterminated string: %s)
�
IDENTIFIER�match�groups�end�SyntaxError�replace�append�STRING_CHUNK�join�lstrip)�	remaining�m�result�q�oq�parts�s�r-��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/util.py�
marker_varAs:






z parse_marker.<locals>.marker_varcs�|rR|ddkrR�|dd����\}}|ddkr@td|��|dd���}nZ�|�\}}|r�t�|�}|srq�|��d}||��d�}�|�\}}|||d�}q^|}||fS)Nr�(r�)�unterminated parenthesis: %s��op�lhs�rhs)r%r �	MARKER_OPrrr)r&r(r5r'r4r6)�markerr/r-r.�marker_expres 
z!parse_marker.<locals>.marker_exprcsR�|�\}}|rJt�|�}|s qJ||��d�}�|�\}}d||d�}q||fS)N�andr3)�ANDrr�r&r5r'r6)r9r-r.�
marker_andxs
z parse_marker.<locals>.marker_andcsR�|�\}}|rJt�|�}|s qJ||��d�}�|�\}}d||d�}q||fS)N�orr3)�ORrrr<)r=r-r.r8�s
zparse_marker.<locals>.markerr-)�
marker_stringr-)r8r=r9r/r.�parse_marker8s
	$rAcCs0|��}|r|�d�rdSt�|�}|s4td|��|��d}||��d�}d}}}}|�r:|ddk�r:|�dd�}|dkr�td|��|d|�}	||dd���}g}|	�r0t�|	�}|s�td	|	��|�	|��d�|	|��d�}	|	�s�q0|	dd
k�rtd|	��|	dd���}	q�|�s:d}|�r�|ddk�r�|dd���}t
�|�}|�sztd
|��|��d}t|�}
|
j�r�|
j
�s�td|��||��d���}n�dd�}|ddk�r�||�\}}n�|�dd�}|dk�rtd|��|d|�}	||dd���}t�|	��r@||	�\}}nXt�|	�}|�s\td|	��|��d}
|	|��d���}	|	�r�td|	��d|
fg}|�r�|ddk�r�td|��|dd���}t|�\}}|�r�|ddk�r�td|��|�s|}nd|d�dd�|D��f}t||||||d�S)z�
    Parse a requirement passed in as a string. Return a Container
    whose attributes contain the various parts of the requirement.
    �#Nzname expected: %sr�[�]rzunterminated extra: %szmalformed extra: %s�,zcomma expected in extras: %s�@zinvalid URI: %szInvalid URL: %scSs�t�|�}d}|r�g}|��d}||��d�}t�|�}|sLtd|��|��d}|�||f�||��d�}|r�|ddkr�q�|dd���}|s�q�t�|�}|std|��q|s�d}||fS)z|
                Return a list of operator, version tuples if any are
                specified, else None.
                Nrzinvalid version: %srEr�invalid constraint: %s)�
COMPARE_OPrrr�VERSION_IDENTIFIERr r"r%)�
ver_remainingr'�versionsr4�vr-r-r.�get_versions�s.


z'parse_requirement.<locals>.get_versionsr0r1r2rGz~=�;zinvalid requirement: %szunexpected trailing data: %s�%s %s�, cSsg|]}d|�qS)rOr-)�.0�conr-r-r.�
<listcomp>	�z%parse_requirement.<locals>.<listcomp>)�name�extras�constraintsr8�url�requirement)�strip�
startswithrrr rr�findr%r"�	NON_SPACEr�scheme�netlocrHrIrAr$r)�reqr&r'�distnamerV�	mark_exprrK�uri�ir,�trM�_rL�rsr-r-r.�parse_requirement�s�


 


�rhcCs�dd�}i}|D]�\}}}tj�||�}t|�D]p}tj�||�}	t|	�D]T}
|||
�}|durn|�|d�qJ|||
�}|�tjjd��d�}
|
d|||<qJq0q|S)z%Find destinations for resources filescSsD|�tjjd�}|�tjjd�}|�|�s.J�|t|�d��d�S�N�/)r!�os�path�sepr[�lenr%)�rootrlr-r-r.�get_rel_pathsz)get_resources_dests.<locals>.get_rel_pathNrj)rkrlr$r�popr!rm�rstrip)�resources_root�rulesrp�destinations�base�suffix�dest�prefix�abs_base�abs_glob�abs_path�
resource_file�rel_path�rel_destr-r-r.�get_resources_destss

r�cCs(ttd�rd}ntjttdtj�k}|S)N�real_prefixT�base_prefix)�hasattr�sysry�getattr�r(r-r-r.�in_venv(s
r�cCstj}t|t�st|�}|S�N)r��
executable�
isinstancerrr�r-r-r.�get_executable2s
r�cCsN|}t|�}|}|s|r|}|r|d��}||vr6qJ|rd|||f}q|S)Nrz	%c: %s
%s)r	�lower)�prompt�
allowed_chars�error_prompt�default�pr,�cr-r-r.�proceedDsr�cCs8t|t�r|��}i}|D]}||vr||||<q|Sr�)r�r�split)�d�keysr(�keyr-r-r.�extract_by_keyTs
r�cCsvtjddkrt�d�|�}|��}t|�}zlt�|�}|ddd}|��D]B\}}|��D]0\}}d||f}t	|�}	|	dus�J�|	||<q`qP|WSt
y�|�dd�Yn0dd	�}
t�
�}z|
||�Wn:tj�y|��t�|�}t|�}|
||�Yn0i}|��D]R}i||<}|�|�D]4\}
}d|
|f}t	|�}	|	du�sbJ�|	||
<�q8�q|S)
Nr��utf-8�
extensionszpython.exports�exportsz%s = %scSs$t|d�r|�|�n
|�|�dS)N�	read_file)r�r��readfp)�cp�streamr-r-r.�read_streamqs
z!read_exports.<locals>.read_stream)r��version_info�codecs�	getreader�readr
�json�load�items�get_export_entry�	Exception�seekr�ConfigParser�MissingSectionHeaderError�close�textwrap�dedent�sections)r��data�jdatar(�group�entries�krLr,�entryr�r�r�rU�valuer-r-r.�read_exports]sD

r�cCs�tjddkrt�d�|�}t��}|��D]l\}}|�|�|��D]P}|j	dur\|j
}nd|j
|j	f}|jr�d|d�|j�f}|�
||j|�qFq,|�|�dS)Nrr�r�z%s:%sz%s [%s]rP)r�r�r��	getwriterrr�r��add_section�valuesrwry�flagsr$�setrU�write)r�r�r�r�rLr�r,r-r-r.�
write_exports�s

r�c	cs.t��}z|VWt�|�nt�|�0dSr�)�tempfile�mkdtempr�rmtree)�tdr-r-r.�tempdir�sr�c	cs8t��}zt�|�dVWt�|�nt�|�0dSr�)rk�getcwd�chdir)r��cwdr-r-r.r��s

r��c	cs8t��}zt�|�dVWt�|�nt�|�0dSr�)�socket�getdefaulttimeout�setdefaulttimeout)�seconds�ctor-r-r.�socket_timeout�s

r�c@seZdZdd�Zddd�ZdS)�cached_propertycCs
||_dSr�)�func)�selfr�r-r-r.�__init__�szcached_property.__init__NcCs,|dur|S|�|�}t�||jj|�|Sr�)r��object�__setattr__�__name__)r��obj�clsr�r-r-r.�__get__�s

zcached_property.__get__)N)r��
__module__�__qualname__r�r�r-r-r-r.r��sr�cCs~tjdkr|S|s|S|ddkr.td|��|ddkrFtd|��|�d�}tj|vrh|�tj�qP|srtjStjj|�S)a�Return 'pathname' as a name that will work on the native filesystem.

    The path is split on '/' and put back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    rjrzpath '%s' cannot be absolute���zpath '%s' cannot end with '/')rkrm�
ValueErrorr��curdir�removerlr$)�pathname�pathsr-r-r.�convert_path�s



r�c@s�eZdZd$dd�Zdd�Zdd�Zdd	�Zd%dd�Zd&dd�Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
d'dd�Zdd�Zdd�Zd d!�Zd"d#�Zd
S)(�FileOperatorFcCs||_t�|_|��dSr�)�dry_runr��ensured�_init_record)r�r�r-r-r.r��szFileOperator.__init__cCsd|_t�|_t�|_dS�NF)�recordr��
files_written�dirs_created�r�r-r-r.r��szFileOperator._init_recordcCs|jr|j�|�dSr�)r�r��add)r�rlr-r-r.�record_as_written�szFileOperator.record_as_writtencCsHtj�|�s tdtj�|���tj�|�s0dSt�|�jt�|�jkS)a�Tell if the target is newer than the source.

        Returns true if 'source' exists and is more recently modified than
        'target', or if 'source' exists and 'target' doesn't.

        Returns false if both exist and 'target' is the same age or younger
        than 'source'. Raise PackagingFileError if 'source' does not exist.

        Note that this test is not very accurate: files created in the same
        second will have the same "age".
        zfile '%r' does not existT)rkrl�existsr�abspath�stat�st_mtime)r��source�targetr-r-r.�newer�s
�zFileOperator.newerTcCs�|�tj�|��t�d||�|js�d}|rdtj�|�rDd|}n tj�|�rdtj�	|�sdd|}|rtt
|d��t�||�|�
|�dS)z8Copy a file respecting dry-run and force flags.
        zCopying %s to %sNz%s is a symlinkz%s is a non-regular filez which would be overwritten)�
ensure_dirrkrl�dirname�logger�infor��islinkr��isfiler�r�copyfiler�)r��infile�outfile�check�msgr-r-r.�	copy_files
zFileOperator.copy_fileNcCs�tj�|�rJ�|�tj�|��t�d||�|js~|durJt|d�}nt	j|d|d�}zt
�||�W|��n
|��0|�
|�dS)NzCopying stream %s to %s�wb�w��encoding)rkrl�isdirr�r�r�r�r��openr�r�copyfileobjr�r�)r��instreamr�r�	outstreamr-r-r.�copy_streamszFileOperator.copy_streamcCsp|�tj�|��|jsbtj�|�r.t�|�t|d��}|�|�Wd�n1sX0Y|�	|�dS)Nr)
r�rkrlr�r�r�r�rr�r�)r�rlr��fr-r-r.�write_binary_file's
(zFileOperator.write_binary_filecCs|�||�|��dSr�)r�encode)r�rlr�rr-r-r.�write_text_file0szFileOperator.write_text_filecCsntjdkstjdkrjtjdkrj|D]F}|jr:t�d|�q"t�|�j|B|@}t�d||�t�||�q"dS)N�posix�javazchanging mode of %szchanging mode of %s to %o)	rkrU�_namer�r�r�r��st_mode�chmod)r��bits�mask�filesr�moder-r-r.�set_mode3szFileOperator.set_modecCs|�dd|�S)Nimi�)r)r,rr-r-r.�<lambda>?rTzFileOperator.<lambda>cCs|tj�|�}||jvrxtj�|�sx|j�|�tj�|�\}}|�|�t�	d|�|j
sft�|�|jrx|j
�|�dS)NzCreating %s)rkrlr�r�r�r�r�r�r�r�r��mkdirr�r�)r�rlr�rr-r-r.r�As

zFileOperator.ensure_dirc	Cs�t||�}t�d||�|js�|s0|�||�rX|s:d}n|�|�sHJ�|t|�d�}i}|rvttd�rvtj	j
|d<tj|||dfi|��|�|�|S)NzByte-compiling %s to %s�PycInvalidationMode�invalidation_modeT)
rr�r�r�r�r[rnr��
py_compiler�CHECKED_HASH�compiler�)	r�rl�optimize�forcery�hashed_invalidation�dpath�diagpath�compile_kwargsr-r-r.�byte_compileMs
zFileOperator.byte_compilecCs�tj�|�r�tj�|�r^tj�|�s^t�d|�|js@t�	|�|j
r�||jvr�|j�|�nPtj�|�rpd}nd}t�d||�|js�t�|�|j
r�||j
vr�|j
�|�dS)NzRemoving directory tree at %s�link�filezRemoving %s %s)rkrlr�rr�r��debugr�rr�r�r�r�r�)r�rlr,r-r-r.�ensure_removed^s"



zFileOperator.ensure_removedcCsDd}|s@tj�|�r$t�|tj�}q@tj�|�}||kr:q@|}q|Sr�)rkrlr��access�W_OKr�)r�rlr(�parentr-r-r.�is_writablesszFileOperator.is_writablecCs"|js
J�|j|jf}|��|S)zV
        Commit recorded changes, turn off recording, return
        changes.
        )r�r�r�r�)r�r(r-r-r.�commits
zFileOperator.commitcCs�|js�t|j�D]}tj�|�rt�|�qt|jdd�}|D]F}t�	|�}|rz|dgks^J�tj�
||d�}t�|�t�|�q>|��dS)NT)�reverse�__pycache__r)
r��listr�rkrlr�r��sortedr��listdirr$�rmdirr�)r�r�dirsr��flist�sdr-r-r.�rollback�s

zFileOperator.rollback)F)T)N)FFNF)r�r�r�r�r�r�r�rr
rrr�set_executable_moder�r&r*r.r/r9r-r-r-r.r��s 


	

r�cCs^|tjvrtj|}nt|�}|dur,|}n.|�d�}t||�d��}|D]}t||�}qJ|S)N�.r)r��modules�
__import__r�r�rq)�module_name�dotted_path�modr(r+r�r-r-r.�resolve�s

rAc@s6eZdZdd�Zedd��Zdd�Zdd�Zej	Z	d	S)
�ExportEntrycCs||_||_||_||_dSr��rUryrwr�)r�rUryrwr�r-r-r.r��szExportEntry.__init__cCst|j|j�Sr�)rAryrwr�r-r-r.r��szExportEntry.valuecCsd|j|j|j|jfS)Nz<ExportEntry %s = %s:%s %s>rCr�r-r-r.�__repr__�s
�zExportEntry.__repr__cCsDt|t�sd}n0|j|jko>|j|jko>|j|jko>|j|jk}|Sr�)r�rBrUryrwr�)r��otherr(r-r-r.�__eq__�s

�
�
�zExportEntry.__eq__N)
r�r�r�r�r�r�rDrFr��__hash__r-r-r-r.rB�s

rBz�(?P<name>(\w|[-.+])+)
                      \s*=\s*(?P<callable>(\w+)([:\.]\w+)*)
                      \s*(\[\s*(?P<flags>[\w-]+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])?
                      c
Cs�t�|�}|s0d}d|vs"d|vr�td|��n�|��}|d}|d}|�d�}|dkrf|d}}n"|dkrztd|��|�d�\}}|d	}	|	dur�d|vs�d|vr�td|��g}	nd
d�|	�d�D�}	t||||	�}|S)
NrCrDzInvalid specification '%s'rU�callable�:rrr�cSsg|]}|���qSr-�rZ)rQrr-r-r.rS�rTz$get_export_entry.<locals>.<listcomp>rE)�ENTRY_RE�searchr�	groupdict�countr�rB)
�
specificationr'r(r�rUrl�colonsryrwr�r-r-r.r��s8
�
��r�cCs�|durd}tjdkr.dtjvr.tj�d�}ntj�d�}tj�|�rft�|tj�}|s�t	�
d|�n:zt�|�d}Wn&ty�t	j
d	|dd
�d}Yn0|s�t
��}t	�
d|�tj�||�S)
a�
    Return the default base location for distlib caches. If the directory does
    not exist, it is created. Use the suffix provided for the base directory,
    and default to '.distlib' if it isn't provided.

    On Windows, if LOCALAPPDATA is defined in the environment, then it is
    assumed to be a directory, and will be the parent directory of the result.
    On POSIX, and on Windows if LOCALAPPDATA is not defined, the user's home
    directory - using os.expanduser('~') - will be the parent directory of
    the result.

    The result is just the directory '.distlib' in the parent directory as
    determined above, or with the name specified with ``suffix``.
    Nz.distlib�nt�LOCALAPPDATAz
$localappdata�~z(Directory exists but is not writable: %sTzUnable to create %s)�exc_infoFz#Default location unusable, using %s)rkrU�environrl�
expandvars�
expanduserrr+r,r��warning�makedirs�OSErrorr�r�r$)rwr(�usabler-r-r.�get_cache_base�s&

r\cCsBtj�tj�|��\}}|r(|�dd�}|�tjd�}||dS)a
    Convert an absolute path to a directory name for use in a cache.

    The algorithm used is:

    #. On Windows, any ``':'`` in the drive is replaced with ``'---'``.
    #. Any occurrence of ``os.sep`` is replaced with ``'--'``.
    #. ``'.cache'`` is appended.
    rIz---z--z.cache)rkrl�
splitdriver�r!rm)rlr�r�r-r-r.�path_to_cache_dirs

r^cCs|�d�s|dS|Sri)�endswith)r,r-r-r.�ensure_slash$s
r`cCs`d}}d|vr>|�dd�\}}d|vr.|}n|�dd�\}}|rJt|�}|rVt|�}|||fS)NrFrrI)�rsplitr�r)r_�username�passwordryr-r-r.�parse_credentials*srdcCst�d�}t�|�|S)N�)rk�umaskr�r-r-r.�get_process_umask9s

rgcCs<d}d}t|�D]\}}t|t�sd}q,q|dus8J�|S)NTF)�	enumerater�r)�seqr(rdr,r-r-r.�is_string_sequence>s
rjz3([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-([a-z0-9_.+-]+)z
-py(\d\.?\d?)cCs�d}d}t|��dd�}t�|�}|r@|�d�}|d|���}|r�t|�t|�dkr�t�t�	|�d|�}|r�|�
�}|d|�||dd�|f}|dur�t�|�}|r�|�d�|�d�|f}|S)zw
    Extract name, version, python version from a filename (no extension)

    Return name, version, pyver or None
    N� �-rz\br�)rr!�PYTHON_VERSIONrLr��startrn�rer�escaper�PROJECT_NAME_AND_VERSION)�filename�project_namer(�pyverr'�nr-r-r.�split_filenameMs"


rvz-(?P<name>[\w .-]+)\s*\(\s*(?P<ver>[^\s)]+)\)$cCs:t�|�}|std|��|��}|d����|dfS)z�
    A utility method used to get name and version from a string.

    From e.g. a Provides-Dist value.

    :param p: A value in a form 'foo (1.0)'
    :return: The name and version as a tuple.
    z$Ill-formed name/version string: '%s'rU�ver)�NAME_VERSION_RErrrMrZr�)r�r'r�r-r-r.�parse_name_and_versionis
	
rycCs�t�}t|pg�}t|pg�}d|vr8|�d�||O}|D]x}|dkrT|�|�q<|�d�r�|dd�}||vr�t�d|�||vr�|�|�q<||vr�t�d|�|�|�q<|S)N�*rlrzundeclared extra: %s)r�r�r�r[r�rX)�	requested�	availabler(�r�unwantedr-r-r.�
get_extrasxs&

rc
Cs�i}zNt|�}|��}|�d�}|�d�s8t�d|�nt�d�|�}t�	|�}Wn2t
y�}zt�d||�WYd}~n
d}~00|S)NzContent-Typezapplication/jsonz(Unexpected response for JSON request: %sr�z&Failed to get external data for %s: %s)rr��getr[r�r)r�r�r�r�r��	exception)rXr(�resp�headers�ct�reader�er-r-r.�_get_external_data�s

$r�z'https://www.red-dove.com/pypi/projects/cCs*d|d��|f}tt|�}t|�}|S)Nz%s/%s/project.jsonr��upperr
�_external_data_base_urlr�)rUrXr(r-r-r.�get_project_data�s
r�cCs(d|d��||f}tt|�}t|�S)Nz%s/%s/package-%s.jsonrr�)rU�versionrXr-r-r.�get_package_data�s
r�c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�Cachez�
    A class implementing a cache for resources that need to live in the file system
    e.g. shared libraries. This class was moved from resources to here because it
    could be used by other modules, e.g. the wheel module.
    cCsPtj�|�st�|�t�|�jd@dkr6t�d|�tj�tj�	|��|_
dS)zu
        Initialise an instance.

        :param base: The base directory where the cache should be located.
        �?rzDirectory '%s' is not privateN)rkrlrrYr�rr�rXr��normpathrv)r�rvr-r-r.r��s

zCache.__init__cCst|�S)zN
        Converts a resource prefix to a directory name in the cache.
        )r^)r�ryr-r-r.�
prefix_to_dir�szCache.prefix_to_dirc	Cs�g}t�|j�D]p}tj�|j|�}z>tj�|�s>tj�|�rJt�|�ntj�|�r`t	�
|�Wqty~|�|�Yq0q|S)z"
        Clear the cache.
        )
rkr4rvrlr$r�r�r�rrr�r�r")r��not_removed�fnr-r-r.�clear�szCache.clearN)r�r�r��__doc__r�r�r�r-r-r-r.r��sr�c@s:eZdZdZdd�Zddd�Zdd�Zd	d
�Zdd�Zd
S)�
EventMixinz1
    A very simple publish/subscribe system.
    cCs
i|_dSr�)�_subscribersr�r-r-r.r��szEventMixin.__init__TcCsD|j}||vrt|g�||<n"||}|r6|�|�n
|�|�dS)a`
        Add a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be added (and called when the
                           event is published).
        :param append: Whether to append or prepend the subscriber to an
                       existing subscriber list for the event.
        N)r�rr"�
appendleft)r��event�
subscriberr"�subs�sqr-r-r.r��s
zEventMixin.addcCs,|j}||vrtd|��||�|�dS)z�
        Remove a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be removed.
        zNo subscribers: %rN)r�r�r�)r�r�r�r�r-r-r.r��szEventMixin.removecCst|j�|d��S)z�
        Return an iterator for the subscribers for an event.
        :param event: The event to return subscribers for.
        r-)�iterr�r�)r�r�r-r-r.�get_subscribersszEventMixin.get_subscribersc	Ospg}|�|�D]J}z||g|�Ri|��}Wn tyLt�d�d}Yn0|�|�qt�d||||�|S)a^
        Publish a event and return a list of values returned by its
        subscribers.

        :param event: The event to publish.
        :param args: The positional arguments to pass to the event's
                     subscribers.
        :param kwargs: The keyword arguments to pass to the event's
                       subscribers.
        z"Exception during event publicationNz/publish %s: args = %s, kwargs = %s, result = %s)r�r�r�r�r"r))r�r��args�kwargsr(r�r�r-r-r.�publish
s

�zEventMixin.publishN)T)	r�r�r�r�r�r�r�r�r�r-r-r-r.r��s
r�c@s^eZdZdd�Zdd�Zddd�Zdd	�Zd
d�Zdd
�Zdd�Z	e
dd��Ze
dd��ZdS)�	SequencercCsi|_i|_t�|_dSr�)�_preds�_succsr��_nodesr�r-r-r.r�(szSequencer.__init__cCs|j�|�dSr�)r�r��r��noder-r-r.�add_node-szSequencer.add_nodeFcCs�||jvr|j�|�|r�t|j�|d��D]}|�||�q,t|j�|d��D]}|�||�qPt|j���D]\}}|sp|j|=qpt|j���D]\}}|s�|j|=q�dS)Nr-)r�r�r�r�r�r�r2r�)r�r��edgesr�r,r�rLr-r-r.�remove_node0s

zSequencer.remove_nodecCs<||ksJ�|j�|t���|�|j�|t���|�dSr�)r��
setdefaultr�r�r�)r��pred�succr-r-r.r�@sz
Sequencer.addcCs�||ksJ�z|j|}|j|}WntyBtd|��Yn0z|�|�|�|�Wn"ty~td||f��Yn0dS)Nz%r not a successor of anythingz%r not a successor of %r)r�r��KeyErrorr�r�)r�r�r��preds�succsr-r-r.r�Es

zSequencer.removecCs||jvp||jvp||jvSr�)r�r�r�)r��stepr-r-r.�is_stepRs�zSequencer.is_stepcCs�|�|�std|��g}g}t�}|�|�|r�|�d�}||vrb||kr�|�|�|�|�q.|�|�|�|�|j�|d�}|�	|�q.t
|�S)NzUnknown: %rrr-)r�r�r�r"rqr�r�r�r��extend�reversed)r��finalr(�todo�seenr�r�r-r-r.�	get_stepsVs"





zSequencer.get_stepscsRdg�g�i�i�g�|j��������fdd���D]}|�vr8�|�q8�S)Nrcs��d�|<�d�|<�dd7<��|�z�|}WntyTg}Yn0|D]J}|�vr��|�t�|�|��|<qZ|�vrZt�|�|��|<qZ�|�|kr�g}���}|�|�||kr�q�q�t|�}��|�dS)Nrr)r"r��minrq�tuple)r��
successors�	successor�connected_component�	component��graph�index�
index_counter�lowlinksr(�stack�
strongconnectr-r.r�zs*


z3Sequencer.strong_connections.<locals>.strongconnect)r�r�r-r�r.�strong_connectionsos"
zSequencer.strong_connectionscCsfdg}|jD]*}|j|}|D]}|�d||f�qq|jD]}|�d|�q>|�d�d�|�S)Nzdigraph G {z  %s -> %s;z  %s;�}�
)r�r"r�r$)r�r(r�r�r�r�r-r-r.�dot�s



z
Sequencer.dotN)F)
r�r�r�r�r�r�r�r�r�r��propertyr�r�r-r-r-r.r�'s


2r�)�.tar.gz�.tar.bz2�.tar�.zip�.tgz�.tbz�.whlTc
sZ��fdd�}tj����t���d}|dur�|�d�r>d}nH|�d�rRd}d}n4|�d�rfd	}d
}n |�d�rzd}d
}ntd|��z�|dkr�t|d
�}|r�|��}|D]}||�q�n*t�	||�}|r�|�
�}|D]}||�q�|dk�r*tjddk�r*|�
�D]"}	t|	jt��s|	j�d�|	_�q|���W|�rV|��n|�rT|��0dS)NcsRt|t�s|�d�}tj�tj��|��}|���rB|�tjkrNt	d|��dS)Nr�zpath outside destination: %r)
r�r�decoderkrlr�r$r[rmr�)rlr���dest_dir�plenr-r.�
check_path�s


zunarchive.<locals>.check_path)r�r��zip)r�r��tgzzr:gz)r�r��tbzzr:bz2r��tarr}zUnknown format for %rrr�r�)rkrlr�rnr_r�r�namelist�tarfiler�getnamesr�r��
getmembersr�rUrr��
extractallr�)
�archive_filenamer��formatr�r��archiver�namesrU�tarinfor-r�r.�	unarchive�sL






�r�cCs�t��}t|�}t|d��d}t�|�D]H\}}}|D]8}tj�||�}||d�}	tj�|	|�}
|�||
�q4q&Wd�n1s�0Y|S)z*zip a directory tree into a BytesIO objectrN)	�io�BytesIOrnrrk�walkrlr$r�)�	directoryr(�dlen�zfror6rrU�full�relrxr-r-r.�zip_dir�s.r�)r�K�M�G�T�Pc@sreZdZdZddd�Zdd�Zdd	�Zd
d�Zdd
�Ze	dd��Z
e	dd��Zdd�Ze	dd��Z
e	dd��ZdS)�Progress�UNKNOWNr�dcCs<|dus||ksJ�||_|_||_d|_d|_d|_dS)NrF)r��cur�max�started�elapsed�done)r��minval�maxvalr-r-r.r�szProgress.__init__cCsV|j|ksJ�|jdus&||jks&J�||_t��}|jdurF||_n||j|_dSr�)r�r�r��timer�r�)r��curval�nowr-r-r.�update
s
zProgress.updatecCs |dksJ�|�|j|�dS�Nr)rr�)r��incrr-r-r.�	incrementszProgress.incrementcCs|�|j�|Sr�)rr�r�r-r-r.rnszProgress.startcCs |jdur|�|j�d|_dS)NT)r�rr�r�r-r-r.�stops
z
Progress.stopcCs|jdur|jS|jSr�)r��unknownr�r-r-r.�maximum!szProgress.maximumcCsD|jrd}n4|jdurd}n$d|j|j|j|j}d|}|S)Nz100 %z ?? %gY@z%3d %%)r�r�r�r�)r�r(rLr-r-r.�
percentage%s
zProgress.percentagecCs:|dkr|jdus|j|jkr$d}nt�dt�|��}|S)Nrz??:??:??z%H:%M:%S)r�r�r�r��strftime�gmtime)r��durationr(r-r-r.�format_duration0szProgress.format_durationcCs�|jrd}|j}n^d}|jdur&d}nJ|jdks<|j|jkrBd}n.t|j|j�}||j|j}|d|j}d||�|�fS)N�DonezETA r�rrz%s: %s)r�r�r�r�r��floatr)r�ryrer-r-r.�ETA9s
zProgress.ETAcCsL|jdkrd}n|j|j|j}tD]}|dkr6q@|d}q&d||fS)Nrgi�g@�@z%d %sB/s)r�r�r��UNITS)r�r(�unitr-r-r.�speedLs

zProgress.speedN)rr�)r�r�r�rr�rrrnrr�rrrrrr-r-r-r.r��s




	
r�z\{([^}]*)\}z[^/\\,{]\*\*|\*\*[^/\\,}]z^[^{]*\}|\{[^}]*$cCs<t�|�rd}t||��t�|�r4d}t||��t|�S)zAExtended globbing function that supports ** and {opt1,opt2,opt3}.z7invalid glob %r: recursive glob "**" must be used alonez2invalid glob %r: mismatching set marker '{' or '}')�_CHECK_RECURSIVE_GLOBrLr��_CHECK_MISMATCH_SET�_iglob)�	path_globr�r-r-r.ras

rccst�|d�}t|�dkrht|�dks,J|��|\}}}|�d�D]$}td�|||f��D]
}|VqXq@n�d|vr�t|�D]
}|Vqxn~|�dd�\}}|dkr�d}|dkr�d}n|�d�}|�d	�}t�|�D]4\}}}	tj	�
|�}ttj	�||��D]
}
|
Vq�q�dS)
Nrr�rErz**r;rzrj�\)�	RICH_GLOBr�rnrr$�	std_iglobr%rkr�rlr�)r�rich_path_globryr�rw�itemrl�radical�dirrr�r-r-r.rls*



r)�HTTPSHandler�match_hostname�CertificateErrorc@seZdZdZdZdd�ZdS)�HTTPSConnectionNTcCs^t�|j|jf|j�}t|dd�r0||_|��tt	d�sp|j
rHt	j}nt	j}t	j
||j|j|t	j|j
d�|_n�t	�t	j�}tt	d�r�|jt	jO_|jr�|�|j|j�i}|j
r�t	j|_|j|j
d�tt	dd�r�|j|d<|j
|fi|��|_|j
�rZ|j�rZz$t|j��|j�t�d	|j�Wn.t�yX|j�tj�|j���Yn0dS)
N�_tunnel_hostF�
SSLContext)�	cert_reqs�ssl_version�ca_certs�OP_NO_SSLv2)�cafile�HAS_SNI�server_hostnamezHost verified: %s) r��create_connection�host�port�timeoutr��sock�_tunnelr��sslr&�
CERT_REQUIRED�	CERT_NONE�wrap_socket�key_file�	cert_file�PROTOCOL_SSLv23r#�optionsr'�load_cert_chain�verify_mode�load_verify_locations�check_domainr�getpeercertr�r)r �shutdown�	SHUT_RDWRr�)r�r/r$�contextr�r-r-r.�connect�sB
�



zHTTPSConnection.connect)r�r�r�r&r<rAr-r-r-r.r!�sr!c@s&eZdZd	dd�Zdd�Zdd�ZdS)
rTcCst�|�||_||_dSr�)�BaseHTTPSHandlerr�r&r<)r�r&r<r-r-r.r��s
zHTTPSHandler.__init__cOs(t|i|��}|jr$|j|_|j|_|S)a
            This is called to create a connection instance. Normally you'd
            pass a connection class to do_open, but it doesn't actually check for
            a class, and just expects a callable. As long as we behave just as a
            constructor would have, we should be OK. If it ever changes so that
            we *must* pass a class, we'll create an UnsafeHTTPSConnection class
            which just sets check_domain to False in the class definition, and
            choose which one to pass to do_open.
            )r!r&r<)r�r�r�r(r-r-r.�_conn_maker�s

zHTTPSHandler._conn_makerc
CsZz|�|j|�WStyT}z,dt|j�vr>td|j��n�WYd}~n
d}~00dS)Nzcertificate verify failedz*Unable to verify server certificate for %s)�do_openrCr�str�reasonr r,)r�r`r�r-r-r.�
https_open�s�zHTTPSHandler.https_openN)T)r�r�r�r�rCrGr-r-r-r.r�s
rc@seZdZdd�ZdS)�HTTPSOnlyHandlercCstd|��dS)NzAUnexpected HTTP request on what should be a secure connection: %s)r)r�r`r-r-r.�	http_open�s�zHTTPSOnlyHandler.http_openN)r�r�r�rIr-r-r-r.rH�srH��rJ�c@seZdZddd�ZdS)�HTTPrNcKs*|dkrd}|�|j||fi|���dSr��_setupZ_connection_class�r�r,r-r�r-r-r.r��sz
HTTP.__init__)rN�r�r�r�r�r-r-r-r.rM�srMc@seZdZddd�ZdS)�HTTPSrNcKs*|dkrd}|�|j||fi|���dSrrNrPr-r-r.r��szHTTPS.__init__)rNrQr-r-r-r.rR�srRc@seZdZddd�Zdd�ZdS)�	TransportrcCs||_tj�||�dSr�)r.rrSr��r�r.�use_datetimer-r-r.r�szTransport.__init__cCs`|�|�\}}}tdkr(t||jd�}n4|jr<||jdkrR||_|t�|�f|_|jd}|S)NrK)r.rr)�
get_host_info�	_ver_inforMr.�_connection�_extra_headersr�HTTPConnection)r�r,�h�eh�x509r(r-r-r.�make_connections
zTransport.make_connectionN)r�r�r�r�r�r^r-r-r-r.rS�s
rSc@seZdZddd�Zdd�ZdS)�
SafeTransportrcCs||_tj�||�dSr�)r.rr`r�rTr-r-r.r�szSafeTransport.__init__cCs�|�|�\}}}|si}|j|d<tdkr>t|dfi|��}n>|jrR||jdkrr||_|tj|dfi|��f|_|jd}|S)Nr.rKrr)rVr.rWrRrXrYrr!)r�r,r[r\r�r(r-r-r.r^s
�

zSafeTransport.make_connectionN)rr_r-r-r-r.r`s
r`c@seZdZdd�ZdS)�ServerProxycKsx|�dd�|_}|dur^t|�d}|�dd�}|dkr@t}nt}|||d�|d<}||_tjj	||fi|��dS)Nr.rrU�https)rU�	transport)
rqr.rr�r`rSrcrrar�)r�rcr�r.r^rU�tclsrer-r-r.r�&szServerProxy.__init__NrQr-r-r-r.ra%sracKs:tjddkr|d7}nd|d<d|d<t||fi|��S)Nrr��br�newliner�r)r�r�r)r�rr�r-r-r.�	_csv_open;s

rgc@s4eZdZed�ed�ed�d�Zdd�Zdd�Zd	S)
�CSVBaserE�"r�)�	delimiter�	quotechar�lineterminatorcCs|Sr�r-r�r-r-r.�	__enter__MszCSVBase.__enter__cGs|j��dSr�)r�r�)r�rTr-r-r.�__exit__PszCSVBase.__exit__N)r�r�r�rE�defaultsrmrnr-r-r-r.rhFs�rhc@s(eZdZdd�Zdd�Zdd�ZeZdS)�	CSVReadercKs`d|vr4|d}tjddkr,t�d�|�}||_nt|dd�|_tj|jfi|j��|_dS)Nr�rr�r�rlr})	r�r�r�r�r�rg�csvr�ro)r�r�r�r-r-r.r�UszCSVReader.__init__cCs|Sr�r-r�r-r-r.�__iter__`szCSVReader.__iter__cCsFt|j�}tjddkrBt|�D] \}}t|t�s |�d�||<q |S�Nrr�r�)�nextr�r�r�rhr�rr�)r�r(rdrr-r-r.rtcs

zCSVReader.nextN)r�r�r�r�rrrt�__next__r-r-r-r.rpTsrpc@seZdZdd�Zdd�ZdS)�	CSVWritercKs(t|d�|_tj|jfi|j��|_dS)Nr)rgr�rq�writerro)r�r�r�r-r-r.r�nszCSVWriter.__init__cCsNtjddkr>g}|D]"}t|t�r.|�d�}|�|�q|}|j�|�dSrs)r�r�r�rr
r"rw�writerow)r��rowr}rr-r-r.rxrs

zCSVWriter.writerowN)r�r�r�r�rxr-r-r-r.rvmsrvcsHeZdZeej�Zded<d�fdd�	Zdd�Zdd	�Zd
d�Z	�Z
S)
�Configurator�inc_convert�incNcs"tt|��|�|pt��|_dSr�)�superrzr�rkr�rv)r��configrv��	__class__r-r.r��szConfigurator.__init__c
s���fdd����d�}t|�s*��|�}��dd�}��dd�}|r\t�fdd�|D��}��fd	d��D�}t|�}||i|��}|r�|��D]\}}	t||�|	��q�|S)
Ncsvt|ttf�r*t|��fdd�|D��}nHt|t�rhd|vrH��|�}qri}|D]}�||�||<qPn
��|�}|S)Ncsg|]}�|��qSr-r-)rQrd��convertr-r.rS�rTzBConfigurator.configure_custom.<locals>.convert.<locals>.<listcomp>�())r�r2r��type�dict�configure_customr�)�or(r�)r�r�r-r.r��s

z.Configurator.configure_custom.<locals>.convertr�r;z[]r-csg|]}�|��qSr-r-)rQr�r�r-r.rS�rTz1Configurator.configure_custom.<locals>.<listcomp>cs$g|]}t|�r|��|�f�qSr-)r)rQr�)r~r�r-r.rS�rT)rqrHrAr�r�r��setattr)
r�r~r��propsr�r�r�r(rurLr-)r~r�r�r.r��s

zConfigurator.configure_customcCs4|j|}t|t�r0d|vr0|�|�|j|<}|S)Nr�)r~r�r�r�)r�r�r(r-r-r.�__getitem__�s
zConfigurator.__getitem__cCsZtj�|�stj�|j|�}tj|ddd��}t�|�}Wd�n1sL0Y|S)z*Default converter for the inc:// protocol.r}r�rN)	rkrl�isabsr$rvr�rr�r�)r�r�rr(r-r-r.r{�s
(zConfigurator.inc_convert)N)r�r�r�r�r�value_convertersr�r�r�r{�
__classcell__r-r-rr.rz�s
rzc@s*eZdZdZd
dd�Zdd�Zdd	�ZdS)�SubprocessMixinzC
    Mixin for running subprocesses and capturing their output
    FNcCs||_||_dSr�)�verbose�progress)r�r�r�r-r-r.r��szSubprocessMixin.__init__cCsj|j}|j}|��}|sq^|dur.|||�q|s@tj�d�ntj�|�d��tj��q|��dS)z�
        Read lines from a subprocess' output stream and either pass to a progress
        callable (if specified) or write progress information to sys.stderr.
        Nr;r�)	r�r��readliner��stderrr�r��flushr�)r�r�r@r�r�r,r-r-r.r��szSubprocessMixin.readercKs�tj|ftjtjd�|��}tj|j|jdfd�}|��tj|j|jdfd�}|��|�	�|�
�|�
�|jdur�|�dd�n|jr�t
j�d�|S)N)�stdoutr�r�)r�r�r�zdone.�mainzdone.
)�
subprocess�Popen�PIPE�	threading�Threadr�r�rnr��waitr$r�r�r�r�)r��cmdr�r��t1�t2r-r-r.�run_command�s"��
zSubprocessMixin.run_command)FN)r�r�r�r�r�r�r�r-r-r-r.r��s
r�cCst�dd|���S)z,Normalize a python package name a la PEP 503z[-_.]+rl)ro�subr�)rUr-r-r.�normalize_name�sr�c@s.eZdZdZdZd
dd�Zdd�Zdd	�ZdS)�
PyPIRCFilezhttps://upload.pypi.org/legacy/�pypiNcCs.|durtj�tj�d�d�}||_||_dS)NrSz.pypirc)rkrlr$rWrrrX)r�r�rXr-r-r.r��szPyPIRCFile.__init__c
Cs�i}tj�|j��r||jp|j}t��}|�|j�|�	�}d|v�r*|�
dd�}dd�|�d�D�}|gkr~d|vr|dg}n�|D]�}d|i}|�
|d�|d<d	|jfd
|jfdfD].\}}	|�
||�r�|�
||�||<q�|	||<q�|dk�r
||jdfv�r
|j|d	<q�|d|kr�|d	|kr�i}q�nRd|v�r|d}|�
|d	��rT|�
|d	�}n|j}|�
|d�|�
|d
�|||jd�}|S)N�	distutilsz
index-serverscSs g|]}|��dkr|���qS)rrJ)rQ�serverr-r-r.rSs�z#PyPIRCFile.read.<locals>.<listcomp>r�r�r�rb�
repository�realm)rcNzserver-loginrc)rbrcr�r�r�)rkrlr�rrrX�DEFAULT_REPOSITORYr�RawConfigParserr�r�r�r��
DEFAULT_REALM�
has_option)
r�r(r�r~r��
index_servers�_serversr�r�r�r-r-r.r��sX
��

�
�


�zPyPIRCFile.readcCs�t��}|j}|�|�|�d�s,|�d�|�dd|�|�dd|�t|d��}|�|�Wd�n1sr0YdS)Nr�rbrcr)	rr�rrr��has_sectionr�r�rr�)r�rbrcr~r�rr-r-r.r9s


zPyPIRCFile.update)NN)r�r�r�r�r�r�r�rr-r-r-r.r��s

:r�cCst|jd���S)zG
    Read the PyPI access configuration as supported by distutils.
    )rX)r�rXr��r�r-r-r.�_load_pypircEsr�cCst��|j|j�dSr�)r�rrbrcr�r-r-r.�
_store_pypircKsr�cCs�tjdkrFdtj��vrdSdtj��vr.dSdtj��vr@dStjSdtjvrZtjdStjd	ksnttd
�sttjSt��\}}}}}|���	dd�}|�	d
d��	dd�}|dd�dkr�d||fS|dd�dk�r(|ddk�r�d}dt
|d�d|dd�f}ddd�}|d|tj7}n�|dd�dk�rLdd lm
}|�S|dd!�d"k�r�d"}t�d#tj�}|�|�}|�r�|��}n>|dd!�d$k�r�ddl}	ddl}
|	�|
j��|||�\}}}d%|||fS)&a�Return a string that identifies the current platform.  This is used mainly to
    distinguish platform-specific build directories and platform-specific built
    distributions.  Typically includes the OS name and version and the
    architecture (as supplied by 'os.uname()'), although the exact information
    included depends on the OS; eg. on Linux, the kernel version isn't
    particularly important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.

    rQ�amd64�	win-amd64z(arm)�	win-arm32z(arm64)z	win-arm64�_PYTHON_HOST_PLATFORMr�unamerjrrkrfrlN��linuxz%s-%s�sunosr�5�solarisz%d.%sr�rJ�32bit�64bit)i���l����z.%s�aix)�aix_platformrL�cygwinz[\d.]+�darwinz%s-%s-%s)rkrUr�r�r��platformrUr�r�r!�int�maxsize�_aix_supportr�ror�ASCIIrr��_osx_support�distutils.sysconfig�get_platform_osx�	sysconfig�get_config_vars)�osnamer,�releaser��machine�bitnessr��rel_rer'r�r�r-r-r.�get_host_platformSsN


 


�
r��win32r�r�)�x86�x64�armcCs2tjdkrt�Stj�d�}|tvr*t�St|S)NrQ�VSCMD_ARG_TGT_ARCH)rkrUr�rUr��_TARGET_TO_PLAT)�cross_compilation_targetr-r-r.�get_platform�s
r�)NN)r�)N)N)NT)�r��collectionsr�
contextlibrq�globrrr�r��loggingrkrror�r1�ImportErrorr�r�r�r�r�r�Zdummy_threadingr�rr�compatrrrr	r
rrr
rrrrrrrrrrrrr�	getLoggerr�r�rrrIrHr7r?r;r]r#rArhr�r�r�r�r�r�r��contextmanagerr�r�r�r�r�r�r�rArB�VERBOSErKr�r\r^r`rdrgrj�Irqrmrvrxryrr�r�r�r�r�r�r��ARCHIVE_EXTENSIONSr�r�rr�rrrrrrBrr r!rHr�rWrMrRrSr`rargrhrprvrzr�r�r�r�r�r�r�r�r-r-r-r.�<module>s�
\








Y}

	/

		8�
)
�


,H
6]


+)	
7.QR�site-packages/pip/_vendor/distlib/__pycache__/resources.cpython-39.pyc000064400000025451150564745010022033 0ustar00a

��ReD*�	@s�ddlmZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
mZddlm
Z
mZmZe�e�ZdaGdd�de�ZGdd	�d	e�ZGd
d�de�ZGdd
�d
e�ZGdd�de�ZGdd�de�Zed�ee	jeiZzNzddlZWne �yddl!ZYn0eeej"<eeej#<eeej$<[Wne e%f�yRYn0dd�Z&iZ'dd�Z(e�)e*d��Z+dd�Z,dS)�)�unicode_literalsN�)�DistlibException)�cached_property�get_cache_base�Cachecs.eZdZd�fdd�	Zdd�Zdd�Z�ZS)	�
ResourceCacheNcs0|durtj�t�td��}tt|��|�dS)Nzresource-cache)�os�path�joinr�str�superr�__init__)�self�base��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/resources.pyrszResourceCache.__init__cCsdS)z�
        Is the cache stale for the given resource?

        :param resource: The :class:`Resource` being cached.
        :param path: The path of the resource in the cache.
        :return: True if the cache is stale.
        Tr�r�resourcer
rrr�is_stale"s	zResourceCache.is_stalecCs�|j�|�\}}|dur|}n�tj�|j|�|�|�}tj�|�}tj�|�sXt�	|�tj�
|�sjd}n|�||�}|r�t|d��}|�
|j�Wd�n1s�0Y|S)z�
        Get a resource into the cache,

        :param resource: A :class:`Resource` instance.
        :return: The pathname of the resource in the cache.
        NT�wb)�finder�get_cache_infor	r
rr�
prefix_to_dir�dirname�isdir�makedirs�existsr�open�write�bytes)rr�prefixr
�resultr�stale�frrr�get-s
*zResourceCache.get)N)�__name__�
__module__�__qualname__rrr'�
__classcell__rrrrrsrc@seZdZdd�ZdS)�ResourceBasecCs||_||_dS�N)r�name)rrr.rrrrHszResourceBase.__init__N)r(r)r*rrrrrr,Gsr,c@s@eZdZdZdZdd�Zedd��Zedd��Zed	d
��Z	dS)�Resourcez�
    A class representing an in-package resource, such as a data file. This is
    not normally instantiated by user code, but rather by a
    :class:`ResourceFinder` which manages the resource.
    FcCs|j�|�S)z�
        Get the resource as a stream.

        This is not a property to make it obvious that it returns a new stream
        each time.
        )r�
get_stream�rrrr�	as_streamUszResource.as_streamcCstdurt�at�|�Sr-)�cacherr'r1rrr�	file_path^szResource.file_pathcCs|j�|�Sr-)r�	get_bytesr1rrrr"eszResource.bytescCs|j�|�Sr-)r�get_sizer1rrr�sizeisz
Resource.sizeN)
r(r)r*�__doc__�is_containerr2rr4r"r7rrrrr/Ms	

r/c@seZdZdZedd��ZdS)�ResourceContainerTcCs|j�|�Sr-)r�
get_resourcesr1rrr�	resourcesqszResourceContainer.resourcesN)r(r)r*r9rr<rrrrr:nsr:c@s�eZdZdZej�d�rdZndZdd�Zdd�Z	d	d
�Z
dd�Zd
d�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zeejj�Zdd�ZdS)�ResourceFinderz4
    Resource finder for file system resources.
    �java)�.pyc�.pyoz.class)r?r@cCs.||_t|dd�|_tj�t|dd��|_dS)N�
__loader__�__file__�)�module�getattr�loaderr	r
rr)rrDrrrr�szResourceFinder.__init__cCstj�|�Sr-)r	r
�realpath�rr
rrr�_adjust_path�szResourceFinder._adjust_pathcCsBt|t�rd}nd}|�|�}|�d|j�tjj|�}|�|�S)N�/�/r)	�
isinstancer"�split�insertrr	r
rrI)r�
resource_name�sep�partsr$rrr�
_make_path�s

zResourceFinder._make_pathcCstj�|�Sr-)r	r
rrHrrr�_find�szResourceFinder._findcCs
d|jfSr-)r
�rrrrrr�szResourceFinder.get_cache_infocCsD|�|�}|�|�sd}n&|�|�r0t||�}n
t||�}||_|Sr-)rRrS�
_is_directoryr:r/r
)rrOr
r$rrr�find�s



zResourceFinder.findcCst|jd�S�N�rb)r r
rTrrrr0�szResourceFinder.get_streamcCs8t|jd��}|��Wd�S1s*0YdSrW)r r
�read)rrr&rrrr5�szResourceFinder.get_bytescCstj�|j�Sr-)r	r
�getsizerTrrrr6�szResourceFinder.get_sizecs*�fdd��t�fdd�t�|j�D��S)Ncs|dko|��j�S)N�__pycache__)�endswith�skipped_extensions)r&r1rr�allowed�s
�z-ResourceFinder.get_resources.<locals>.allowedcsg|]}�|�r|�qSrr)�.0r&)r^rr�
<listcomp>��z0ResourceFinder.get_resources.<locals>.<listcomp>)�setr	�listdirr
rTr)r^rrr;�szResourceFinder.get_resourcescCs|�|j�Sr-)rUr
rTrrrr9�szResourceFinder.is_containerccs�|�|�}|dur�|g}|r�|�d�}|V|jr|j}|jD]>}|sL|}nd�||g�}|�|�}|jrv|�|�q>|Vq>qdS)NrrK)rV�popr9r.r<r�append)rrOr�todo�rnamer.�new_name�childrrr�iterator�s 



zResourceFinder.iteratorN)r(r)r*r8�sys�platform�
startswithr]rrIrRrSrrVr0r5r6r;r9�staticmethodr	r
rrUrjrrrrr=vs"r=cs`eZdZdZ�fdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Z�Z
S)�ZipResourceFinderz6
    Resource finder for resources in .zip files.
    csZtt|��|�|jj}dt|�|_t|jd�r>|jj|_nt	j
||_t|j�|_dS)Nr�_files)
r
rorrF�archive�len�
prefix_len�hasattrrp�	zipimport�_zip_directory_cache�sorted�index)rrDrqrrrr�szZipResourceFinder.__init__cCs|Sr-rrHrrrrI�szZipResourceFinder._adjust_pathcCs�||jd�}||jvrd}nV|r:|dtjkr:|tj}t�|j|�}z|j|�|�}Wntyrd}Yn0|s�t�	d||j
j�nt�	d||j
j�|S)NT���Fz_find failed: %r %rz_find worked: %r %r)rsrpr	rP�bisectrxrm�
IndexError�logger�debugrFr#)rr
r$�irrrrS�s


zZipResourceFinder._findcCs&|jj}|jdt|�d�}||fS)Nr)rFrqr
rr)rrr#r
rrrr�sz ZipResourceFinder.get_cache_infocCs|j�|j�Sr-)rF�get_datar
rTrrrr5�szZipResourceFinder.get_bytescCst�|�|��Sr-)�io�BytesIOr5rTrrrr0�szZipResourceFinder.get_streamcCs|j|jd�}|j|dS)N�)r
rsrprrrrr6�szZipResourceFinder.get_sizecCs�|j|jd�}|r,|dtjkr,|tj7}t|�}t�}t�|j|�}|t|j�kr�|j|�|�shq�|j||d�}|�	|�
tjd�d�|d7}qH|S)Nryrr)r
rsr	rPrrrbrzrxrm�addrM)rrr
�plenr$r~�srrrr;s

zZipResourceFinder.get_resourcescCsh||jd�}|r*|dtjkr*|tj7}t�|j|�}z|j|�|�}Wntybd}Yn0|S)NryF)rsr	rPrzrxrmr{)rr
r~r$rrrrUs

zZipResourceFinder._is_directory)r(r)r*r8rrIrSrr5r0r6r;rUr+rrrrro�srocCs|tt|�<dSr-)�_finder_registry�type)rF�finder_makerrrr�register_finder2sr�cCs�|tvrt|}nv|tjvr$t|�tj|}t|dd�}|durJtd��t|dd�}t�t|��}|durxtd|��||�}|t|<|S)z�
    Return a resource finder for a package.
    :param package: The name of the package.
    :return: A :class:`ResourceFinder` instance for the package.
    �__path__Nz8You cannot get a finder for a module, only for a packagerAzUnable to locate finder for %r)	�
_finder_cacherk�modules�
__import__rErr�r'r�)�packager$rDr
rFr�rrrr9s


r�	__dummy__cCsRd}t�|�tj�|�}t�t|��}|rNt}tj	�
|d�|_||_||�}|S)z�
    Return a resource finder for a path, which should represent a container.

    :param path: The path.
    :return: A :class:`ResourceFinder` instance for the path.
    NrC)
�pkgutil�get_importerrk�path_importer_cacher'r�r��
_dummy_moduler	r
rrBrA)r
r$rFrrDrrr�finder_for_pathUs
r�)-�
__future__rrzr��loggingr	r�rk�typesrurCr�utilrrr�	getLoggerr(r|r3r�objectr,r/r:r=ror��zipimporterr��_frozen_importlib_externalZ_fi�ImportError�_frozen_importlib�SourceFileLoader�
FileFinder�SourcelessFileLoader�AttributeErrorr�r�r�
ModuleTyperr�r�rrrr�<module>sJ
,!ZO�


site-packages/pip/_vendor/distlib/__pycache__/compat.cpython-39.pyc000064400000076164150564745010021313 0ustar00a

��Re��@sddlmZddlZddlZddlZzddlZWneyFdZYn0ejddk�rpddlmZe	fZ
eZddl
mZddlZddlZddlmZddlmZmZmZmZmZdd	lmZmZmZm Z m!Z!m"Z"m#Z#d
d�Zddl$Z$ddl$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-e�r$dd
l$m.Z.ddl/Z/ddl0Z0ddl1Z2ddl3m3Z3ddl4Z4e5Z5ddl6m7Z8ddl6m9Z:�nddl;mZe<fZ
e<Zddl;m=ZddlZddlZddlZddl>mZmZmZmZmZmZmZm#Z#ddl?m&Z&mZm%Z%m Z m!Z!m)Z)m*Z*m+Z+m,Z,m-Z-e�rdd
l?m.Z.ddl@m(Z(m'Z'm"Z"ddlAmBZ/ddl?mCZ$ddlDmBZ0ddl2Z2ddlEm3Z3ddlFmGZ4eHZ5ddl6m:Z:e8Z8zddlmIZImJZJWn6e�y�Gdd�deK�ZJdbdd�ZLdd�ZIYn0zddl
mMZNWn$e�yGdd�deO�ZNYn0zdd lmPZPWn*e�y<ejQejRBdfd!d"�ZPYn0dd#lSmTZUeVeUd$��r\eUZTn,dd%lSmWZXGd&d'�d'eX�ZWGd(d)�d)eU�ZTzdd*lYmZZZWne�y�d+d,�ZZYn0zddl[Z[Wn e�y�dd-lm[Z[Yn0ze\Z\Wn(e]�ydd.l^m_Z_d/d0�Z\Yn0zej`Z`ejaZaWnFeb�yle�c��pBd1Zdedd2k�rTd3Zend4Zed5d6�Z`d7d8�ZaYn0zdd9lfmgZgWnFe�y�dd:lhmiZimjZjddlZe�kd;�Zld<d=�Zmd>d?�ZgYn0zdd@lnmoZoWn e�y�dd@lpmoZoYn0ejddA�dBk�re3�jqZqnddClnmqZqzddDlrmsZsWn`e�y�ddElrmtZtzddFlumvZwWne�y|dcdHdI�ZwYn0GdJdK�dKet�ZsYn0zddLlxmyZyWnDe�y�zddLlzmyZyWne�y�dddMdN�ZyYn0Yn0zddOlrm{Z{Wn�e�y~zddPl|m}Z~Wn e�y:ddPlm}Z~Yn0zddQl�m�Z�m�Z�m�Z�Wne�yhYn0GdRdS�dSe��Z{Yn0zddTl�m�Z�m�Z�Wnte�ye�kdUej��Z�dVdW�Z�GdXdY�dYe��Z�dedZd[�Z�Gd\d]�d]e��Z�Gd^d_�d_e��Z�Gd`da�daeO�Z�Yn0dS)f�)�absolute_importN�)�StringIO)�FileType�)�shutil)�urlparse�
urlunparse�urljoin�urlsplit�
urlunsplit)�urlretrieve�quote�unquote�url2pathname�pathname2url�ContentTooShortError�	splittypecCst|t�r|�d�}t|�S)N�utf-8)�
isinstance�unicode�encode�_quote)�s�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/compat.pyrs

r)	�Request�urlopen�URLError�	HTTPError�HTTPBasicAuthHandler�HTTPPasswordMgr�HTTPHandler�HTTPRedirectHandler�build_opener)�HTTPSHandler)�
HTMLParser)�ifilter)�ifilterfalse)�
TextIOWrapper)rr	r
rrrrr)
rr
rrrr r!r"r#r$)rrr)�filterfalse)�match_hostname�CertificateErrorc@seZdZdS)r,N)�__name__�
__module__�__qualname__rrrrr,`sr,c
Cs�g}|sdS|�d�}|d|dd�}}|�d�}||krNtdt|���|sb|��|��kS|dkrv|�d�n>|�d	�s�|�d	�r�|�t�|��n|�t�|��	d
d��|D]}|�t�|��q�t�
dd
�|�dtj�}	|	�
|�S)zpMatching according to RFC 6125, section 6.4.3

        http://tools.ietf.org/html/rfc6125#section-6.4.3
        F�.rrN�*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)�split�countr,�repr�lower�append�
startswith�re�escape�replace�compile�join�
IGNORECASE�match)
�dn�hostnameZ
max_wildcardsZpats�partsZleftmost�	remainder�	wildcards�frag�patrrr�_dnsname_matchds*


�rFcCs�|std��g}|�dd�}|D]*\}}|dkr t||�r@dS|�|�q |s�|�dd�D]6}|D],\}}|dkrdt||�r�dS|�|�qdq\t|�dkr�td	|d
�tt|��f��n*t|�dkr�td||df��ntd
��dS)a=Verify that *cert* (in decoded format as returned by
        SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
        rules are followed, but IP addresses are not accepted for *hostname*.

        CertificateError is raised on failure. On success, the function
        returns nothing.
        ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIRED�subjectAltNamer�DNSN�subject�
commonNamerz&hostname %r doesn't match either of %s�, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)	�
ValueError�getrFr6�lenr,r<�mapr4)�certr@�dnsnames�san�key�value�subrrrr+�s2

�
�r+)�SimpleNamespacec@seZdZdZdd�ZdS)�	ContainerzR
        A generic container for when multiple values need to be returned
        cKs|j�|�dS�N��__dict__�update)�self�kwargsrrr�__init__�szContainer.__init__N)r-r.r/�__doc__r^rrrrrW�srW)�whichcs"dd�}tj���r&|�|�r"�SdS|dur>tj�dtj�}|sFdS|�tj�}tj	dkr�tj
|vrt|�dtj
�tj�dd��tj�}t�fd	d
�|D��r��g}q‡fdd�|D�}n�g}t
�}|D]P}tj�|�}||vr�|�|�|D](}	tj�||	�}
||
|�r�|
Sq�q�dS)
aKGiven a command, mode, and a PATH string, return the path which
        conforms to the given mode on the PATH, or None if there is no such
        file.

        `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
        of os.environ.get("PATH"), or can be overridden with a custom search
        path.

        cSs&tj�|�o$t�||�o$tj�|�SrX)�os�path�exists�access�isdir)�fn�moderrr�
_access_check�s�zwhich.<locals>._access_checkN�PATH�win32r�PATHEXT�c3s |]}����|���VqdSrX)r5�endswith��.0�ext��cmdrr�	<genexpr>��zwhich.<locals>.<genexpr>csg|]}�|�qSrrrnrqrr�
<listcomp>�rtzwhich.<locals>.<listcomp>)rarb�dirname�environrM�defpathr2�pathsep�sys�platform�curdir�insert�any�set�normcase�addr<)rrrgrbrh�pathext�files�seen�dir�normdir�thefile�namerrqrr`�s8





r`)�ZipFile�	__enter__)�
ZipExtFilec@s$eZdZdd�Zdd�Zdd�ZdS)r�cCs|j�|j�dSrXrY)r\�baserrrr^szZipExtFile.__init__cCs|SrXr�r\rrrr�szZipExtFile.__enter__cGs|��dSrX��close�r\�exc_inforrr�__exit__szZipExtFile.__exit__N)r-r.r/r^r�r�rrrrr�sr�c@s$eZdZdd�Zdd�Zdd�ZdS)r�cCs|SrXrr�rrrr�$szZipFile.__enter__cGs|��dSrXr�r�rrrr�'szZipFile.__exit__cOs tj|g|�Ri|��}t|�SrX)�BaseZipFile�openr�)r\�argsr]r�rrrr�+szZipFile.openN)r-r.r/r�r�r�rrrrr�#sr�)�python_implementationcCs0dtjvrdStjdkrdStj�d�r,dSdS)z6Return a string identifying the Python implementation.�PyPy�java�Jython�
IronPython�CPython)rz�versionrar�r7rrrrr�2s

r�)�	sysconfig)�CallablecCs
t|t�SrX)rr�)�objrrr�callableFsr�r�mbcs�strict�surrogateescapecCs:t|t�r|St|t�r$|�tt�Stdt|�j��dS�Nzexpect bytes or str, not %s)	r�bytes�	text_typer�_fsencoding�	_fserrors�	TypeError�typer-��filenamerrr�fsencodeZs

�r�cCs:t|t�r|St|t�r$|�tt�Stdt|�j��dSr�)	rr�r��decoder�r�r�r�r-r�rrr�fsdecodecs

�r�)�detect_encoding)�BOM_UTF8�lookupzcoding[:=]\s*([-\w.]+)cCsH|dd����dd�}|dks*|�d�r.dS|dvs@|�d�rDd	S|S)
z(Imitates get_normal_name in tokenizer.c.N��_�-rzutf-8-)zlatin-1�
iso-8859-1ziso-latin-1)zlatin-1-ziso-8859-1-ziso-latin-1-r�)r5r:r7)�orig_enc�encrrr�_get_normal_namets�r�cs�z�jj�Wnty"d�Yn0d�d}d}�fdd�}��fdd�}|�}|�t�rnd�|d	d�}d
}|sz|gfS||�}|r�||gfS|�}|s�||gfS||�}|r�|||gfS|||gfS)a?
        The detect_encoding() function is used to detect the encoding that should
        be used to decode a Python source file.  It requires one argument, readline,
        in the same way as the tokenize() generator.

        It will call readline a maximum of twice, and return the encoding used
        (as a string) and a list of any lines (left as bytes) it has read in.

        It detects the encoding from the presence of a utf-8 bom or an encoding
        cookie as specified in pep-0263.  If both a bom and a cookie are present,
        but disagree, a SyntaxError will be raised.  If the encoding cookie is an
        invalid charset, raise a SyntaxError.  Note that if a utf-8 bom is found,
        'utf-8-sig' is returned.

        If no encoding is specified, then the default of 'utf-8' will be returned.
        NFrcs"z��WStyYdS0dS)Nrt)�
StopIterationr)�readlinerr�read_or_stop�sz%detect_encoding.<locals>.read_or_stopcs�z|�d�}Wn2ty@d}�dur4d�|��}t|��Yn0t�|�}|sTdSt|d�}zt|�}Wn8ty��dur�d|}nd��|�}t|��Yn0�r�|j	dkrԈdur�d}n
d���}t|��|d	7}|S)
Nrz'invalid or missing encoding declarationz{} for {!r}rzunknown encoding: zunknown encoding for {!r}: {}zencoding problem: utf-8z encoding problem for {!r}: utf-8z-sig)
r��UnicodeDecodeError�format�SyntaxError�	cookie_re�findallr�r��LookupErrorr�)�line�line_string�msg�matches�encoding�codec)�	bom_foundr�rr�find_cookie�s8

�

z$detect_encoding.<locals>.find_cookieTrz	utf-8-sig)�__self__r��AttributeErrorr7r�)r�r��defaultr�r��first�secondr)r�r�r�rr�s4
&


r�)r9�)r�)�unescape)�ChainMap)�MutableMapping)�recursive_repr�...cs�fdd�}|S)zm
            Decorator to make a repr function return fillvalue for a recursive
            call
            csLt�����fdd�}t�d�|_t�d�|_t�d�|_t�di�|_|S)Nc	sLt|�t�f}|�vr�S��|�z�|�}W��|�n��|�0|SrX)�id�	get_identr��discard)r\rS�result)�	fillvalue�repr_running�
user_functionrr�wrapper�s

z=_recursive_repr.<locals>.decorating_function.<locals>.wrapperr.r_r-�__annotations__)r�getattrr.r_r-r�)r�r��r�)r�r�r�decorating_function�sz,_recursive_repr.<locals>.decorating_functionr)r�r�rr�r�_recursive_repr�sr�c@s�eZdZdZdd�Zdd�Zdd�Zd'd	d
�Zdd�Zd
d�Z	dd�Z
dd�Ze�dd��Z
edd��Zdd�ZeZdd�Zedd��Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�ZdS)(r�a� A ChainMap groups multiple dicts (or other mappings) together
        to create a single, updateable view.

        The underlying mappings are stored in a list.  That list is public and can
        accessed or updated using the *maps* attribute.  There is no other state.

        Lookups search the underlying mappings successively until a key is found.
        In contrast, writes, updates, and deletions only operate on the first
        mapping.

        cGst|�pig|_dS)z�Initialize a ChainMap by setting *maps* to the given mappings.
            If no mappings are provided, a single empty dictionary is used.

            N)�list�maps)r\r�rrrr^szChainMap.__init__cCst|��dSrX)�KeyError�r\rSrrr�__missing__szChainMap.__missing__c	Cs8|jD]&}z||WSty*Yq0q|�|�SrX)r�r�r�)r\rS�mappingrrr�__getitem__s
zChainMap.__getitem__NcCs||vr||S|SrXr�r\rSr�rrrrM'szChainMap.getcCstt�j|j��SrX)rNr�unionr�r�rrr�__len__*szChainMap.__len__cCstt�j|j��SrX)�iterrr�r�r�rrr�__iter__-szChainMap.__iter__cst�fdd�|jD��S)Nc3s|]}�|vVqdSrXr)ro�m�rSrrrs1rtz(ChainMap.__contains__.<locals>.<genexpr>�r~r�r�rr�r�__contains__0szChainMap.__contains__cCs
t|j�SrXr�r�rrr�__bool__3szChainMap.__bool__cCsd�|d�tt|j���S)Nz{0.__class__.__name__}({1})rK)r�r<rOr4r�r�rrr�__repr__6s�zChainMap.__repr__cGs|tj|g|�R��S)z?Create a ChainMap with a single dict created from the iterable.)�dict�fromkeys)�cls�iterabler�rrrr�;szChainMap.fromkeyscCs&|j|jd��g|jdd��R�S)zHNew ChainMap or subclass with a new copy of maps[0] and refs to maps[1:]rrN)�	__class__r��copyr�rrrr�@sz
ChainMap.copycCs|jig|j�R�S)z;New ChainMap with a new dict followed by all previous maps.�r�r�r�rrr�	new_childFszChainMap.new_childcCs|j|jdd��S)zNew ChainMap from maps[1:].rNr�r�rrr�parentsJszChainMap.parentscCs||jd|<dS)Nr)r�)r\rSrTrrr�__setitem__OszChainMap.__setitem__cCs6z|jd|=Wn ty0td�|���Yn0dS)Nr�(Key not found in the first mapping: {!r})r�r�r�r�rrr�__delitem__RszChainMap.__delitem__cCs0z|jd��WSty*td��Yn0dS)zPRemove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.rz#No keys found in the first mapping.N)r��popitemr�r�rrrrXszChainMap.popitemcGs@z|jdj|g|�R�WSty:td�|���Yn0dS)zWRemove *key* from maps[0] and return its value. Raise KeyError if *key* not in maps[0].rrN)r��popr�r�)r\rSr�rrrr_szChainMap.popcCs|jd��dS)z'Clear maps[0], leaving maps[1:] intact.rN)r��clearr�rrrrfszChainMap.clear)N)r-r.r/r_r^r�r�rMr�r�r�r�r�r��classmethodr�r��__copy__r��propertyr�r�rrrrrrrrr�s.



r�)�cache_from_sourcecCs0|�d�sJ�|durd}|r$d}nd}||S)Nz.pyT�c�o)rm)rb�debug_override�suffixrrrrpsr)�OrderedDict)r�)�KeysView�
ValuesView�	ItemsViewc@s�eZdZdZdd�Zejfdd�Zejfdd�Zdd	�Zd
d�Z	dd
�Z
d6dd�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�ZeZe�Zefdd �Zd7d"d#�Zd8d$d%�Zd&d'�Zd(d)�Zed9d*d+��Zd,d-�Zd.d/�Zd0d1�Zd2d3�Z d4d5�Z!d!S):r
z)Dictionary that remembers insertion ordercOspt|�dkrtdt|���z
|jWn4tyZg|_}||dg|dd�<i|_Yn0|j|i|��dS)z�Initialize an ordered dictionary.  Signature is the same as for
            regular dictionaries, but keyword arguments are not recommended
            because their insertion order is arbitrary.

            rz$expected at most 1 arguments, got %dN)rNr��_OrderedDict__rootr��_OrderedDict__map�_OrderedDict__update)r\r��kwds�rootrrrr^�s

zOrderedDict.__init__cCsF||vr6|j}|d}|||g|d<|d<|j|<||||�dS)z!od.__setitem__(i, y) <==> od[i]=yrrN)rr)r\rSrTZdict_setitemr�lastrrrr��s
 zOrderedDict.__setitem__cCs0|||�|j�|�\}}}||d<||d<dS)z od.__delitem__(y) <==> del od[y]rrN)rr)r\rSZdict_delitem�	link_prev�	link_nextrrrr�s
zOrderedDict.__delitem__ccs.|j}|d}||ur*|dV|d}qdS)zod.__iter__() <==> iter(od)rr�N�r�r\r�currrrrr��s

zOrderedDict.__iter__ccs.|j}|d}||ur*|dV|d}qdS)z#od.__reversed__() <==> reversed(od)rr�Nrrrrr�__reversed__�s

zOrderedDict.__reversed__cCsbz@|j��D]}|dd�=q|j}||dg|dd�<|j��WntyRYn0t�|�dS)z.od.clear() -> None.  Remove all items from od.N)r�
itervaluesrrr�r�)r\�noderrrrr�szOrderedDict.clearTcCs||std��|j}|r8|d}|d}||d<||d<n |d}|d}||d<||d<|d}|j|=t�||�}||fS)z�od.popitem() -> (k, v), return and remove a (key, value) pair.
            Pairs are returned in LIFO order if last is true or FIFO order if false.

            zdictionary is emptyrrr�)r�rrr�r)r\rr�linkrrrSrTrrrr�s 
zOrderedDict.popitemcCst|�S)zod.keys() -> list of keys in od)r�r�rrr�keys�szOrderedDict.keyscs�fdd��D�S)z#od.values() -> list of values in odcsg|]}�|�qSrr�rorSr�rrru�rtz&OrderedDict.values.<locals>.<listcomp>rr�rr�r�values�szOrderedDict.valuescs�fdd��D�S)z.od.items() -> list of (key, value) pairs in odcsg|]}|�|f�qSrrr!r�rrru�rtz%OrderedDict.items.<locals>.<listcomp>rr�rr�r�items�szOrderedDict.itemscCst|�S)z0od.iterkeys() -> an iterator over the keys in od)r�r�rrr�iterkeys�szOrderedDict.iterkeysccs|D]}||VqdS)z2od.itervalues -> an iterator over the values in odNr�r\�krrrr�szOrderedDict.itervaluesccs|D]}|||fVqdS)z=od.iteritems -> an iterator over the (key, value) items in odNrr%rrr�	iteritemsszOrderedDict.iteritemscOs�t|�dkr tdt|�f��n|s,td��|d}d}t|�dkrL|d}t|t�rn|D]}||||<qZn<t|d�r�|��D]}||||<q�n|D]\}}|||<q�|��D]\}}|||<q�dS)	a�od.update(E, **F) -> None.  Update od from dict/iterable E and F.

            If E is a dict instance, does:           for k in E: od[k] = E[k]
            If E has a .keys() method, does:         for k in E.keys(): od[k] = E[k]
            Or if E is an iterable of items, does:   for k, v in E: od[k] = v
            In either case, this is followed by:     for k, v in F.items(): od[k] = v

            r�z8update() takes at most 2 positional arguments (%d given)z,update() takes at least 1 argument (0 given)rrrr N)rNr�rr��hasattrr r#)r�rr\�otherrSrTrrrr[	s(	�


zOrderedDict.updatecCs0||vr||}||=|S||jur,t|��|S)z�od.pop(k[,d]) -> v, remove specified key and return the corresponding value.
            If key is not found, d is returned if given, otherwise KeyError is raised.

            )�_OrderedDict__markerr�)r\rSr�r�rrrr,s
zOrderedDict.popNcCs||vr||S|||<|S)zDod.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in odrr�rrr�
setdefault9szOrderedDict.setdefaultcCsn|si}t|�t�f}||vr"dSd||<z6|sFd|jjfW||=Sd|jj|��fW||=S||=0dS)zod.__repr__() <==> repr(od)r�rz%s()z%s(%r)N)r��
_get_identr�r-r#)r\�
_repr_runningZcall_keyrrrr�@s��zOrderedDict.__repr__csX�fdd��D�}t����}tt��D]}|�|d�q(|rL�j|f|fS�j|ffS)z%Return state information for picklingcsg|]}|�|g�qSrr�ror&r�rrruPrtz*OrderedDict.__reduce__.<locals>.<listcomp>N)�varsr�r
rr�)r\r#�	inst_dictr&rr�r�
__reduce__NszOrderedDict.__reduce__cCs
|�|�S)z!od.copy() -> a shallow copy of od)r�r�rrrr�XszOrderedDict.copycCs|�}|D]}|||<q
|S)z�OD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
            and values equal to v (which defaults to None).

            r)r�r�rT�drSrrrr�\s
zOrderedDict.fromkeyscCs6t|t�r*t|�t|�ko(|��|��kSt�||�S)z�od.__eq__(y) <==> od==y.  Comparison to another OD is order-sensitive
            while comparison to a regular mapping is order-insensitive.

            )rr
rNr#r��__eq__�r\r)rrrr3gs
 zOrderedDict.__eq__cCs
||kSrXrr4rrr�__ne__pszOrderedDict.__ne__cCst|�S)z@od.viewkeys() -> a set-like object providing a view on od's keys)rr�rrr�viewkeysuszOrderedDict.viewkeyscCst|�S)z<od.viewvalues() -> an object providing a view on od's values)rr�rrr�
viewvaluesyszOrderedDict.viewvaluescCst|�S)zBod.viewitems() -> a set-like object providing a view on od's items)rr�rrr�	viewitems}szOrderedDict.viewitems)T)N)N)N)"r-r.r/r_r^r�r�rr�rrrr r"r#r$rr'r[r�objectr*rr+r�r1r�rr�r3r5r6r7r8rrrrr
�s:
	




	r
)�BaseConfigurator�valid_identz^[a-z_][a-z0-9_]*$cCst�|�}|std|��dS)Nz!Not a valid Python identifier: %rT)�
IDENTIFIERr>rL)rr�rrrr;�s
r;c@s"eZdZdZdd�Zddd�ZdS)�ConvertingDictz A converting dictionary wrapper.cCsJt�||�}|j�|�}||urF|||<t|�tttfvrF||_||_	|SrX)
r�r��configurator�convertr�r=�ConvertingList�ConvertingTuple�parentrS�r\rSrTr�rrrr��s
�zConvertingDict.__getitem__NcCsLt�|||�}|j�|�}||urH|||<t|�tttfvrH||_||_	|SrX)
r�rMr>r?r�r=r@rArBrS�r\rSr�rTr�rrrrM�s
�zConvertingDict.get)N)r-r.r/r_r�rMrrrrr=�sr=cCsDt�|||�}|j�|�}||ur@t|�tttfvr@||_||_	|SrX)
r�rr>r?r�r=r@rArBrSrDrrrr�s
�rc@s"eZdZdZdd�Zddd�ZdS)	r@zA converting list wrapper.cCsJt�||�}|j�|�}||urF|||<t|�tttfvrF||_||_	|SrX)
r�r�r>r?r�r=r@rArBrSrCrrrr��s
�zConvertingList.__getitem__���cCs<t�||�}|j�|�}||ur8t|�tttfvr8||_|SrX)	r�rr>r?r�r=r@rArB)r\�idxrTr�rrrr�s
�zConvertingList.popN)rE)r-r.r/r_r�rrrrrr@�sr@c@seZdZdZdd�ZdS)rAzA converting tuple wrapper.cCsBt�||�}|j�|�}||ur>t|�tttfvr>||_||_	|SrX)
�tupler�r>r?r�r=r@rArBrSrCrrrr��s
�zConvertingTuple.__getitem__N)r-r.r/r_r�rrrrrA�srAc@s�eZdZdZe�d�Ze�d�Ze�d�Ze�d�Z	e�d�Z
ddd	�Zee
�Zd
d�Zdd
�Zdd�Zdd�Zdd�Zdd�Zdd�ZdS)r:zQ
        The configurator base class which defines some useful defaults.
        z%^(?P<prefix>[a-z]+)://(?P<suffix>.*)$z^\s*(\w+)\s*z^\.\s*(\w+)\s*z^\[\s*(\w+)\s*\]\s*z^\d+$�ext_convert�cfg_convert)rp�cfgcCst|�|_||j_dSrX)r=�configr>)r\rKrrrr^�s
zBaseConfigurator.__init__c		Cs�|�d�}|�d�}z\|�|�}|D]F}|d|7}zt||�}Wq$tyh|�|�t||�}Yq$0q$|WSty�t��dd�\}}td||f�}|||_	|_
|�Yn0dS)zl
            Resolve strings to objects using standard import and attribute
            syntax.
            r0rrNzCannot resolve %r: %s)r2r�importerr�r��ImportErrorrzr�rL�	__cause__�
__traceback__)	r\rr��used�foundrD�e�tb�vrrr�resolve�s"



zBaseConfigurator.resolvecCs
|�|�S)z*Default converter for the ext:// protocol.)rU�r\rTrrrrHszBaseConfigurator.ext_convertcCs�|}|j�|�}|dur&td|��n�||��d�}|j|��d}|r�|j�|�}|rn||��d}nb|j�|�}|r�|��d}|j�|�s�||}n0zt	|�}||}Wnt
y�||}Yn0|r�||��d�}qHtd||f��qH|S)z*Default converter for the cfg:// protocol.NzUnable to convert %rrzUnable to convert %r at %r)�WORD_PATTERNr>rL�endrK�groups�DOT_PATTERN�
INDEX_PATTERN�
DIGIT_PATTERN�intr�)r\rT�restr�r2rF�nrrrrIs4
�zBaseConfigurator.cfg_convertcCs�t|t�s$t|t�r$t|�}||_n�t|t�sHt|t�rHt|�}||_nzt|t�slt|t�rlt|�}||_nVt|t�r�|j	�
|�}|r�|��}|d}|j�
|d�}|r�|d}t||�}||�}|S)z�
            Convert values to an appropriate type. dicts, lists and tuples are
            replaced by their converting alternatives. Strings are checked to
            see if they have a conversion format and are converted if they do.
            �prefixNr)rr=r�r>r@r�rArG�string_types�CONVERT_PATTERNr>�	groupdict�value_convertersrMr�)r\rTr�r2r`�	converterrrrrr?4s,
�

zBaseConfigurator.convertcsr��d�}t|�s|�|�}��dd�}t�fdd��D��}|fi|��}|rn|��D]\}}t|||�qX|S)z1Configure an object with a user-supplied factory.z()r0Ncs g|]}t|�r|�|f�qSr)r;r.�rKrrruWrtz5BaseConfigurator.configure_custom.<locals>.<listcomp>)rr�rUr�r#�setattr)r\rKr	�propsr]r�r�rTrrfr�configure_customPs

z!BaseConfigurator.configure_customcCst|t�rt|�}|S)z0Utility function which converts lists to tuples.)rr�rGrVrrr�as_tuple^s
zBaseConfigurator.as_tupleN)r-r.r/r_r8r;rbrWrZr[r\rd�staticmethod�
__import__rLr^rUrHrIr?rirjrrrrr:�s"




�"r:)r)r�)N)N)��
__future__rrar8rz�sslrM�version_infor�
basestringrarr��typesr�	file_type�__builtin__�builtins�ConfigParser�configparserZ	_backportrrr	r
rr�urllibr
rrrrrrr�urllib2rrrrr r!r"r#r$r%�httplib�	xmlrpclib�Queue�queuer&�htmlentitydefs�	raw_input�	itertoolsr'�filterr(r*�io�strr)�urllib.parse�urllib.request�urllib.error�http.client�client�request�
xmlrpc.client�html.parser�
html.entities�entities�inputr+r,rLrFrVrWr9r`�F_OK�X_OK�zipfiler�r�r(r�ZBaseZipExtFiler{r�r�r��	NameError�collections.abcr�r�r�r��getfilesystemencodingr�r��tokenizer��codecsr�r�r;r�r��htmlr9�cgir��collectionsr�r��reprlibr�r��importlib.utilr�impr
�threadr�r,�dummy_thread�_abcollrrrr��logging.configr:r;�Ir<r=rr�r@rGrArrrr�<module>s(
$,(0
2+A


		
[
b
w

site-packages/pip/_vendor/distlib/__pycache__/version.cpython-39.pyc000064400000047652150564745010021515 0ustar00a

��Re�[�@sZdZddlZddlZddlmZddlmZgd�Ze�e	�Z
Gdd�de�ZGd	d
�d
e
�ZGdd�de
�Ze�d
�Zdd�ZeZGdd�de�Zdd�ZGdd�de�Ze�d�dfe�d�dfe�d�dfe�d�dfe�d�dfe�d�dfe�d �d!fe�d"�d#fe�d$�d%fe�d&�d'ff
Ze�d(�dfe�d)�dfe�d*�d!fe�d �d!fe�d+�dffZe�d,�Zd-d.�Zd/d0�Ze�d1ej�Zd2d2d3d2d4ddd5�Zd6d7�ZGd8d9�d9e�Z Gd:d;�d;e�Z!e�d<ej�Z"d=d>�Z#d?d@�Z$GdAdB�dBe�Z%GdCdD�dDe�Z&GdEdF�dFe
�Z'e'eee�e'ee!dGdH��e'e$e&e�dI�Z(e(dJe(dK<dLdM�Z)dS)Nz~
Implementation of a flexible versioning scheme providing support for PEP-440,
setuptools-compatible and semantic versioning.
�N�)�string_types��parse_requirement)�NormalizedVersion�NormalizedMatcher�
LegacyVersion�
LegacyMatcher�SemanticVersion�SemanticMatcher�UnsupportedVersionError�
get_schemec@seZdZdZdS)rzThis is an unsupported version.N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/version.pyrsrc@sxeZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zedd��ZdS)�VersioncCs@|��|_}|�|�|_}t|t�s,J�t|�dks<J�dS)Nr)�strip�_string�parse�_parts�
isinstance�tuple�len)�self�s�partsrrr�__init__szVersion.__init__cCstd��dS)Nzplease implement in a subclass��NotImplementedError�rrrrrr%sz
Version.parsecCs$t|�t|�kr td||f��dS)Nzcannot compare %r and %r)�type�	TypeError�r�otherrrr�_check_compatible(szVersion._check_compatiblecCs|�|�|j|jkS�N�r'rr%rrr�__eq__,s
zVersion.__eq__cCs|�|�Sr(�r*r%rrr�__ne__0szVersion.__ne__cCs|�|�|j|jkSr(r)r%rrr�__lt__3s
zVersion.__lt__cCs|�|�p|�|�Sr(�r-r*r%rrr�__gt__7szVersion.__gt__cCs|�|�p|�|�Sr(r.r%rrr�__le__:szVersion.__le__cCs|�|�p|�|�Sr()r/r*r%rrr�__ge__=szVersion.__ge__cCs
t|j�Sr()�hashr�rrrr�__hash__AszVersion.__hash__cCsd|jj|jfS)Nz%s('%s')��	__class__rrr3rrr�__repr__DszVersion.__repr__cCs|jSr(�rr3rrr�__str__GszVersion.__str__cCstd��dS)NzPlease implement in subclasses.r r3rrr�
is_prereleaseJszVersion.is_prereleaseN)rrrrrr'r*r,r-r/r0r1r4r7r9�propertyr:rrrrrsrc	@s�eZdZdZdd�dd�dd�dd�dd�dd�d	d�d
d�d�Zdd
�Zdd�Zdd�Zedd��Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�ZdS) �MatcherNcCs||kSr(r��v�c�prrr�<lambda>T�zMatcher.<lambda>cCs||kSr(rr=rrrrAUrBcCs||kp||kSr(rr=rrrrAVrBcCs||kp||kSr(rr=rrrrAWrBcCs||kSr(rr=rrrrAXrBcCs||kSr(rr=rrrrAYrBcCs||kp||kSr(rr=rrrrA[rBcCs||kSr(rr=rrrrA\rB)�<�>�<=�>=�==�===�~=�!=cCst|�Sr(rr"rrrraszMatcher.parse_requirementcCs�|jdurtd��|��|_}|�|�}|s:td|��|j|_|j��|_g}|jr�|jD]d\}}|�	d�r�|dvr�td|��|dd�d}}|�|�n|�|�d}}|�
|||f�q^t|�|_dS)	NzPlease specify a version classz
Not valid: %rz.*)rGrJz#'.*' not allowed for %r constraints���TF)
�
version_class�
ValueErrorrrr�name�lower�key�constraints�endswith�appendrr)rr�rZclist�opZvn�prefixrrrrds*


�zMatcher.__init__cCsxt|t�r|�|�}|jD]X\}}}|j�|�}t|t�rDt||�}|s`d||jjf}t	|��||||�sdSqdS)z�
        Check if the provided version matches the constraints.

        :param version: The version to match against this instance.
        :type version: String or :class:`Version` instance.
        z%r not implemented for %sFT)
rrrLr�
_operators�get�getattrr6rr!)r�version�operator�
constraintrV�f�msgrrr�match�s




�z
Matcher.matchcCs6d}t|j�dkr2|jdddvr2|jdd}|S)Nrr)rGrH)rr)r�resultrrr�
exact_version�s zMatcher.exact_versioncCs0t|�t|�ks|j|jkr,td||f��dS)Nzcannot compare %s and %s)r#rNr$r%rrrr'�szMatcher._check_compatiblecCs"|�|�|j|jko |j|jkSr()r'rPrr%rrrr*�s
zMatcher.__eq__cCs|�|�Sr(r+r%rrrr,�szMatcher.__ne__cCst|j�t|j�Sr()r2rPrr3rrrr4�szMatcher.__hash__cCsd|jj|jfS)Nz%s(%r)r5r3rrrr7�szMatcher.__repr__cCs|jSr(r8r3rrrr9�szMatcher.__str__)rrrrLrWrrr_r;rar'r*r,r4r7r9rrrrr<Os*�
r<zk^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?(\.(post)(\d+))?(\.(dev)(\d+))?(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$cCs�|��}t�|�}|s"td|��|��}tdd�|d�d�D��}t|�dkrl|ddkrl|dd�}qF|dszd}nt|ddd��}|dd	�}|d
d�}|dd
�}|d}|dkr�d}n|dt|d�f}|dkr�d}n|dt|d�f}|dk�rd}n|dt|d�f}|du�r2d}nHg}	|�d�D]0}
|
�	��r\dt|
�f}
nd|
f}
|	�
|
��q@t|	�}|�s�|�s�|�r�d}nd}|�s�d}|�s�d}||||||fS)NzNot a valid version: %scss|]}t|�VqdSr(��int��.0r>rrr�	<genexpr>�rBz_pep_440_key.<locals>.<genexpr>r�.���r����	�
��
)NNr)�arh)�z)�_)�final)r�PEP440_VERSION_REr_r�groupsr�splitrrc�isdigitrS)r�mru�nums�epoch�pre�post�dev�localr�partrrr�_pep_440_key�sT



r�c@s0eZdZdZdd�Zegd��Zedd��ZdS)raIA rational version.

    Good:
        1.2         # equivalent to "1.2.0"
        1.2.0
        1.2a1
        1.2.3a2
        1.2.3b1
        1.2.3c1
        1.2.3.4
        TODO: fill this out

    Bad:
        1           # minimum two numbers
        1.2a        # release level must have a release serial
        1.2.3b
    cCs<t|�}t�|�}|��}tdd�|d�d�D��|_|S)Ncss|]}t|�VqdSr(rbrdrrrrfrBz*NormalizedVersion.parse.<locals>.<genexpr>rrg)�_normalized_keyrtr_rurrv�_release_clause)rrr`rxrurrrr
s

zNormalizedVersion.parse)rp�br?�rcr}cst�fdd��jD��S)Nc3s |]}|r|d�jvVqdS)rN)�PREREL_TAGS)re�tr3rrrfrBz2NormalizedVersion.is_prerelease.<locals>.<genexpr>)�anyrr3rr3rr:szNormalizedVersion.is_prereleaseN)	rrrrr�setr�r;r:rrrrr�s
rcCs>t|�}t|�}||krdS|�|�s*dSt|�}||dkS)NTFrg)�str�
startswithr)�x�y�nrrr�
_match_prefixs
r�c	@sneZdZeZddddddddd	�Zd
d�Zdd
�Zdd�Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd�ZdS)r�_match_compatible�	_match_lt�	_match_gt�	_match_le�	_match_ge�	_match_eq�_match_arbitrary�	_match_ne)rIrCrDrErFrGrHrJcCsV|rd|vo|jd}n|jdo,|jd}|rN|j�dd�d}|�|�}||fS)N�+rhrr)rrrvrL)rrZr\rVZstrip_localrrrr�
_adjust_local6s
zNormalizedMatcher._adjust_localcCsD|�|||�\}}||krdS|j}d�dd�|D��}t||�S)NFrgcSsg|]}t|��qSr�r��re�irrr�
<listcomp>IrBz/NormalizedMatcher._match_lt.<locals>.<listcomp>�r�r��joinr��rrZr\rVZrelease_clause�pfxrrrr�DszNormalizedMatcher._match_ltcCsD|�|||�\}}||krdS|j}d�dd�|D��}t||�S)NFrgcSsg|]}t|��qSrr�r�rrrr�QrBz/NormalizedMatcher._match_gt.<locals>.<listcomp>r�r�rrrr�LszNormalizedMatcher._match_gtcCs|�|||�\}}||kSr(�r��rrZr\rVrrrr�TszNormalizedMatcher._match_lecCs|�|||�\}}||kSr(r�r�rrrr�XszNormalizedMatcher._match_gecCs.|�|||�\}}|s ||k}n
t||�}|Sr(�r�r��rrZr\rVr`rrrr�\s


zNormalizedMatcher._match_eqcCst|�t|�kSr(r�r�rrrr�dsz"NormalizedMatcher._match_arbitrarycCs0|�|||�\}}|s ||k}nt||�}|Sr(r�r�rrrr�gs

zNormalizedMatcher._match_necCsf|�|||�\}}||krdS||kr*dS|j}t|�dkrH|dd�}d�dd�|D��}t||�S)NTFrrhrgcSsg|]}t|��qSrr�r�rrrr�zrBz7NormalizedMatcher._match_compatible.<locals>.<listcomp>)r�r�rr�r�r�rrrr�osz#NormalizedMatcher._match_compatibleN)rrrrrLrWr�r�r�r�r�r�r�r�r�rrrrr's&�rz[.+-]$�z^[.](\d)z0.\1z^[.-]z
^\((.*)\)$�\1z^v(ersion)?\s*(\d+)z\2z^r(ev)?\s*(\d+)z[.]{2,}rgz\b(alfa|apha)\b�alphaz\b(pre-alpha|prealpha)\bz	pre.alphaz	\(beta\)$�betaz
^[:~._+-]+z
[,*")([\]]z[~:+_ -]z\.$z
(\d+(\.\d+)*)cCsL|����}tD]\}}|�||�}q|s.d}t�|�}|sFd}|}n�|��d�d�}dd�|D�}t|�dkr~|�	d�qft|�dkr�||�
�d�}n8d�dd�|dd�D��||�
�d�}|dd�}d�d	d�|D��}|��}|�rtD]\}}|�||�}q�|�s|}nd
|v�r*dnd}|||}t
|��sHd}|S)
z�
    Try to suggest a semantic form for a version for which
    _suggest_normalized_version couldn't come up with anything.
    z0.0.0rrgcSsg|]}t|��qSrrbr�rrrr��rBz-_suggest_semantic_version.<locals>.<listcomp>�NcSsg|]}t|��qSrr�r�rrrr��rBcSsg|]}t|��qSrr�r�rrrr��rBr}�-r�)rrO�
_REPLACEMENTS�sub�_NUMERIC_PREFIXr_rurvrrS�endr��_SUFFIX_REPLACEMENTS�	is_semver)rr`�pat�replrxrV�suffix�seprrr�_suggest_semantic_version�s:
,
r�cCsdzt|�|WSty Yn0|��}dD]\}}|�||�}q.t�dd|�}t�dd|�}t�dd|�}t�dd	|�}t�d
d|�}|�d�r�|d
d�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd|�}t�dd |�}zt|�Wnt�y^d}Yn0|S)!a�Suggest a normalized version close to the given version string.

    If you have a version string that isn't rational (i.e. NormalizedVersion
    doesn't like it) then you might be able to get an equivalent (or close)
    rational version from this function.

    This does a number of simple normalizations to the given string, based
    on observation of versions currently in use on PyPI. Given a dump of
    those version during PyCon 2009, 4287 of them:
    - 2312 (53.93%) match NormalizedVersion without change
      with the automatic suggestion
    - 3474 (81.04%) match when using this suggestion method

    @param s {str} An irrational version string.
    @returns A rational version string, or None, if couldn't determine one.
    ))z-alpharp)z-betar�)r�rp)r�r�)r�r?)z-finalr�)z-prer?)z-releaser�)z.releaser�)z-stabler�)r�rg)rrrg)� r�)z.finalr�)rsr�zpre$Zpre0zdev$Zdev0z([abc]|rc)[\-\.](\d+)$z\1\2z[\-\.](dev)[\-\.]?r?(\d+)$z.\1\2z[.~]?([abc])\.?r�r>rNz\b0+(\d+)(?!\d)z(\d+[abc])$z\g<1>0z\.?(dev-r|dev\.r)\.?(\d+)$z.dev\2z-(a|b|c)(\d+)$z[\.\-](dev|devel)$z.dev0z(?![\.\-])dev$z(final|stable)$r�z\.?(r|-|-r)\.?(\d+)$z.post\2z\.?(dev|git|bzr)\.?(\d+)$z\.?(pre|preview|-c)(\d+)$zc\g<2>zp(\d+)$z.post\1)r�rrO�replace�rer�r�)r�rs�origr�rrr�_suggest_normalized_version�s>
	
r�z([a-z]+|\d+|[\.-])r?zfinal-�@)r{�previewr�r�r}r�rgcCsrdd�}g}||�D]T}|�d�r^|dkrD|rD|ddkrD|��q*|r^|ddkr^|��qD|�|�qt|�S)NcSstg}t�|���D]R}t�||�}|rd|dd�krBdkrRnn|�d�}nd|}|�|�q|�d�|S)N�0r�9��*�*final)�
_VERSION_PARTrvrO�_VERSION_REPLACErX�zfillrS)rr`r@rrr�	get_partsCs 
z_legacy_key.<locals>.get_partsr�r�rhz*final-�00000000)r��poprSr)rr�r`r@rrr�_legacy_keyBs



r�c@s eZdZdd�Zedd��ZdS)rcCst|�Sr()r�r"rrrr]szLegacyVersion.parsecCs8d}|jD](}t|t�r
|�d�r
|dkr
d}q4q
|S)NFr�r�T)rrrr�)rr`r�rrrr:`s
�zLegacyVersion.is_prereleaseN�rrrrr;r:rrrrr\src@s4eZdZeZeej�Zded<e�	d�Z
dd�ZdS)r	r�rIz^(\d+(\.\d+)*)cCs`||krdS|j�t|��}|s2t�d||�dS|��d}d|vrV|�dd�d}t||�S)NFzACannot compute compatible match for version %s  and constraint %sTrrgr)�
numeric_rer_r��logger�warningru�rsplitr�)rrZr\rVrxrrrrr�ss�zLegacyMatcher._match_compatibleN)rrrrrL�dictr<rWr��compiler�r�rrrrr	ks


r	zN^(\d+)\.(\d+)\.(\d+)(-[a-z0-9]+(\.[a-z0-9-]+)*)?(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$cCs
t�|�Sr()�
_SEMVER_REr_)rrrrr��sr�c	Csndd�}t|�}|st|��|��}dd�|dd�D�\}}}||dd�||dd�}}|||f||fS)	NcSs8|dur|f}n$|dd��d�}tdd�|D��}|S)NrrgcSs"g|]}|��r|�d�n|�qS)r�)rwr�)rer@rrrr��rBz5_semantic_key.<locals>.make_tuple.<locals>.<listcomp>)rvr)rZabsentr`rrrr�
make_tuple�s
z!_semantic_key.<locals>.make_tuplecSsg|]}t|��qSrrbr�rrrr��rBz!_semantic_key.<locals>.<listcomp>r��|�r�)r�rru)	rr�rxru�major�minor�patchr{�buildrrr�
_semantic_key�s
r�c@s eZdZdd�Zedd��ZdS)r
cCst|�Sr()r�r"rrrr�szSemanticVersion.parsecCs|jdddkS)Nrrr�)rr3rrrr:�szSemanticVersion.is_prereleaseNr�rrrrr
�sr
c@seZdZeZdS)rN)rrrr
rLrrrrr�src@s6eZdZddd�Zdd�Zdd�Zdd	�Zd
d�ZdS)
�
VersionSchemeNcCs||_||_||_dSr()rP�matcher�	suggester)rrPr�r�rrrr�szVersionScheme.__init__cCs0z|j�|�d}Wnty*d}Yn0|S�NTF)r�rLr�rrr`rrr�is_valid_version�s
zVersionScheme.is_valid_versioncCs.z|�|�d}Wnty(d}Yn0|Sr�)r�rr�rrr�is_valid_matcher�s

zVersionScheme.is_valid_matchercCs$|�d�r|dd�}|�d|�S)z:
        Used for processing some metadata fields
        �,Nrhzdummy_name (%s))rRr�r"rrr�is_valid_constraint_list�s
z&VersionScheme.is_valid_constraint_listcCs|jdurd}n
|�|�}|Sr()r�r�rrr�suggest�s

zVersionScheme.suggest)N)rrrrr�r�r�r�rrrrr��s

	r�cCs|Sr(rr"rrrrA�rBrA)�
normalized�legacyZsemanticr��defaultcCs|tvrtd|��t|S)Nzunknown scheme name: %r)�_SCHEMESrM)rNrrrr
�sr
)*r�loggingr��compatr�utilr�__all__�	getLoggerrr�rMr�objectrr<r�rtr�r�rr�rr�r�r�r�r��Ir�r�r�rr	r�r�r�r
rr�r�r
rrrr�<module>s�
1d
=$W���
.r��	'���site-packages/pip/_vendor/distlib/__pycache__/locators.cpython-39.pyc000064400000112624150564745010021646 0ustar00a

��Re���@s0ddlZddlmZddlZddlZddlZddlZddlZzddlZWne	ybddl
ZYn0ddlZddlm
Z
ddlmZmZmZmZmZmZmZmZmZmZmZmZmZmZddlmZm Z m!Z!ddl"m#Z#m$Z$ddl%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,m-Z-dd	l.m/Z/m0Z0dd
l1m2Z2m3Z3e�4e5�Z6e�7d�Z8e�7dej9�Z:e�7d
�Z;dZ<d,dd�Z=Gdd�de�Z>Gdd�de?�Z@Gdd�de@�ZAGdd�de@�ZBGdd�de?�ZCGdd�de@�ZDGdd�de@�ZEGdd �d e@�ZFGd!d"�d"e@�ZGGd#d$�d$e@�ZHeHeF�eDd%d&d'�d(d)�ZIeIjJZJGd*d+�d+e?�ZKdS)-�N)�BytesIO�)�DistlibException)�urljoin�urlparse�
urlunparse�url2pathname�pathname2url�queue�quote�unescape�build_opener�HTTPRedirectHandler�	text_type�Request�	HTTPError�URLError)�Distribution�DistributionPath�	make_dist)�Metadata�MetadataInvalidError)�cached_property�ensure_slash�split_filename�get_project_data�parse_requirement�parse_name_and_version�ServerProxy�normalize_name)�
get_scheme�UnsupportedVersionError)�Wheel�
is_compatiblez^(\w+)=([a-f0-9]+)z;\s*charset\s*=\s*(.*)\s*$ztext/html|application/x(ht)?mlzhttps://pypi.org/pypicCs>|durt}t|dd�}z|��W|d��S|d��0dS)z�
    Return all distribution names known by an index.
    :param url: The URL of the index.
    :return: A list of all known distribution names.
    N�@��timeout�close)�
DEFAULT_INDEXr�
list_packages)�url�client�r,��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/locators.py�get_all_distribution_names)s
�r.c@s$eZdZdZdd�ZeZZZdS)�RedirectHandlerzE
    A class to work around a bug in some Python 3.2.x releases.
    c	Cs�d}dD]}||vr||}q"q|dur.dSt|�}|jdkrnt|��|�}t|d�rf|�||�n|||<t�||||||�S)N)�location�uri��replace_header)r�schemer�get_full_url�hasattrr3�BaseRedirectHandler�http_error_302)	�self�req�fp�code�msg�headers�newurl�key�urlpartsr,r,r-r8@s 

�zRedirectHandler.http_error_302N)�__name__�
__module__�__qualname__�__doc__r8�http_error_301�http_error_303�http_error_307r,r,r,r-r/7sr/c@s�eZdZdZdZdZdZdZedZd)dd	�Z	d
d�Z
dd
�Zdd�Zdd�Z
dd�Zee
e�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd d!�Zd"d#�Zd$d%�Zd*d'd(�ZdS)+�LocatorzG
    A base class for locators - things that locate distributions.
    )z.tar.gzz.tar.bz2z.tarz.zipz.tgzz.tbz)z.eggz.exe�.whl)z.pdfN)rJ�defaultcCs,i|_||_tt��|_d|_t��|_dS)a^
        Initialise an instance.
        :param scheme: Because locators look for most recent versions, they
                       need to know the version scheme to use. This specifies
                       the current PEP-recommended scheme - use ``'legacy'``
                       if you need to support existing distributions on PyPI.
        N)	�_cacher4r
r/�opener�matcherr
�Queue�errors)r9r4r,r,r-�__init__fs
zLocator.__init__cCsTg}|j��sPz|j�d�}|�|�Wn|jjyBYqYn0|j��q|S)z8
        Return any errors which have occurred.
        F)rP�empty�get�append�Empty�	task_done)r9�result�er,r,r-�
get_errorsys

zLocator.get_errorscCs|��dS)z>
        Clear any errors which may have been logged.
        N)rY�r9r,r,r-�clear_errors�szLocator.clear_errorscCs|j��dS�N)rL�clearrZr,r,r-�clear_cache�szLocator.clear_cachecCs|jSr\��_schemerZr,r,r-�_get_scheme�szLocator._get_schemecCs
||_dSr\r_)r9�valuer,r,r-�_set_scheme�szLocator._set_schemecCstd��dS)a=
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This should be implemented in subclasses.

        If called from a locate() request, self.matcher will be set to a
        matcher for the requirement to satisfy, otherwise it will be None.
        � Please implement in the subclassN��NotImplementedError)r9�namer,r,r-�_get_project�s
zLocator._get_projectcCstd��dS)�J
        Return all the distribution names known to this locator.
        rdNrerZr,r,r-�get_distribution_names�szLocator.get_distribution_namescCsL|jdur|�|�}n2||jvr,|j|}n|��|�|�}||j|<|S)z�
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This calls _get_project to do all the work, and just implements a caching layer on top.
        N)rLrhr[)r9rgrWr,r,r-�get_project�s



zLocator.get_projectcCs^t|�}t�|j�}d}|�d�}|�|j�}|rBtt|�|j�}|j	dkd|j
v||||fS)zu
        Give an url a score which can be used to choose preferred URLs
        for a given project release.
        TrJ�httpszpypi.org)r�	posixpath�basename�path�endswith�downloadable_extensionsr#r"�
wheel_tagsr4�netloc)r9r*�trn�
compatible�is_wheelZis_downloadabler,r,r-�	score_url�s
�zLocator.score_urlcCsR|}|rN|�|�}|�|�}||kr(|}||kr@t�d||�nt�d||�|S)a{
        Choose one of two URLs where both are candidates for distribution
        archives for the same version of a distribution (for example,
        .tar.gz vs. zip).

        The current implementation favours https:// URLs over http://, archives
        from PyPI over those from other locations, wheel compatibility (if a
        wheel) and then the archive name.
        zNot replacing %r with %rzReplacing %r with %r)rw�logger�debug)r9�url1�url2rW�s1�s2r,r,r-�
prefer_url�s


zLocator.prefer_urlcCs
t||�S)zZ
        Attempt to split a filename in project name, version and Python version.
        )r)r9�filename�project_namer,r,r-r�szLocator.split_filenamecCs dd�}d}t|�\}}}}}	}
|
���d�r<t�d||
�t�|
�}|rX|��\}}
nd\}}
|}|r�|ddkr�|dd�}|�d	��r@z~t	|�}t
||j�s�t�d
|�nX|dur�d}n||j|�}|�r|j|j
|jt|||||	df�d
�dd�|jD��d�}Wn2t�y<}zt�d|�WYd}~n
d}~00n�|�|j��s\t�d|�n�t�|�}}|jD]�}|�|��rp|dt|��}|�||�}|�s�t�d|�nH|\}}}|�r�|||��r�|||t|||||	df�d�}|�r�||d<�q�qp|�r|�r|
|d|<|S)a
        See if a URL is a candidate for a download URL for a project (the URL
        has typically been scraped from an HTML page).

        If it is, a dictionary is returned with keys "name", "version",
        "filename" and "url"; otherwise, None is returned.
        cSst|�t|�kSr\)r)Zname1Zname2r,r,r-�same_project�sz:Locator.convert_url_to_download_info.<locals>.same_projectNzegg=z %s: version hint in fragment: %r)NN����/rJzWheel not compatible: %sTr2z, cSs"g|]}d�t|dd����qS)�.�N)�join�list)�.0�vr,r,r-�
<listcomp>�z8Locator.convert_url_to_download_info.<locals>.<listcomp>)rg�versionrr*�python-versionzinvalid path for wheel: %szNot downloadable: %sz No match for project/version: %s)rgr�rr*r��	%s_digest)r�lower�
startswithrxry�HASHER_HASH�match�groupsrpr"r#rrrgr�rrr��pyver�	Exception�warningrqrmrn�lenr)r9r*r�r�rWr4rsro�params�query�frag�m�algo�digestZorigpath�wheel�includerXr�extrtrgr�r�r,r,r-�convert_url_to_download_info�sv�
���
	$

��
z$Locator.convert_url_to_download_infocCshd}d|vr6|d}dD]}||vr|||f}q6q|sddD]$}d|}||vr>|||f}qdq>|S)z�
        Get a digest from a dictionary by looking at a "digests" dictionary
        or keys of the form 'algo_digest'.

        Returns a 2-tuple (algo, digest) if found, else None. Currently
        looks only for SHA256, then MD5.
        N�digests)�sha256�md5r�r,)r9�inforWr�r�r@r,r,r-�_get_digest1szLocator._get_digestc	Cs�|�d�}|�d�}||vr,||}|j}nt|||jd�}|j}|�|�|_}|d}||d|<|j|dkr�|�|j|�|_|d�|t	���
|�||_|||<dS)z�
        Update a result dictionary (the final result from _get_project) with a
        dictionary for a specific version, which typically holds information
        gleaned from a filename or URL for an archive for the distribution.
        rgr��r4r*r��urlsN)�pop�metadatarr4r�r��
source_urlr~�
setdefault�set�add�locator)	r9rWr�rgr��dist�mdr�r*r,r,r-�_update_version_dataHs

zLocator._update_version_dataFc	Cs�d}t|�}|dur td|��t|j�}|�|j�|_}t�d|t|�j	�|�
|j�}t|�dk�rg}|j
}	|D]X}
|
dvr�qxz(|�|
�s�n|s�|	|
�js�|�|
�Wqxty�t�d||
�Yqx0qxt|�dkr�t||jd�}|�rt�d	|�|d
}||}|�rv|j�r$|j|_|�di��|t��|_i}|�di�}
|jD]}||
v�rR|
|||<�qR||_d|_|S)
a
        Find the most recent distribution which matches the given
        requirement.

        :param requirement: A requirement of the form 'foo (1.0)' or perhaps
                            'foo (>= 1.0, < 2.0, != 1.3)'
        :param prereleases: If ``True``, allow pre-release versions
                            to be located. Otherwise, pre-release versions
                            are not returned.
        :return: A :class:`Distribution` instance, or ``None`` if no such
                 distribution could be located.
        NzNot a valid requirement: %rzmatcher: %s (%s)r��r�r�zerror matching %s with %rr)r@zsorted list: %sr�r�r�)rrr r4rN�requirementrxry�typerBrkrgr�Z
version_classr��
is_prereleaserTr�r��sortedr@�extrasrSr��
download_urlsr�)r9r��prereleasesrW�rr4rN�versions�slistZvcls�kr��d�sdr*r,r,r-�locate_sP




zLocator.locate)rK)F)rBrCrDrE�source_extensions�binary_extensions�excluded_extensionsrrrqrQrYr[r^rarc�propertyr4rhrjrkrwr~rr�r�r�r�r,r,r,r-rIVs.

JrIcs0eZdZdZ�fdd�Zdd�Zdd�Z�ZS)�PyPIRPCLocatorz�
    This locator uses XML-RPC to locate distributions. It therefore
    cannot be used with simple mirrors (that only mirror file content).
    cs.tt|�jfi|��||_t|dd�|_dS)z�
        Initialise an instance.

        :param url: The URL to use for XML-RPC.
        :param kwargs: Passed to the superclass constructor.
        r$r%N)�superr�rQ�base_urlrr+�r9r*�kwargs��	__class__r,r-rQ�szPyPIRPCLocator.__init__cCst|j���S�ri)r�r+r)rZr,r,r-rj�sz%PyPIRPCLocator.get_distribution_namescCsiid�}|j�|d�}|D]�}|j�||�}|j�||�}t|jd�}|d|_|d|_|�d�|_	|�dg�|_
|�d�|_t|�}|r|d	}	|	d
|_
|�|	�|_||_|||<|D]:}	|	d
}
|�|	�}|d�|t���|
�||d|
<q�q|S)
Nr�Tr�rgr��license�keywords�summaryrr*r�r�)r+Zpackage_releasesZrelease_urlsZrelease_datarr4rgr�rSr�r�r�rr�r�r�r�r�r�r�)r9rgrWr�r�r��datar�r�r�r*r�r,r,r-rh�s0




zPyPIRPCLocator._get_project�rBrCrDrErQrjrh�
__classcell__r,r,r�r-r��sr�cs0eZdZdZ�fdd�Zdd�Zdd�Z�ZS)�PyPIJSONLocatorzw
    This locator uses PyPI's JSON interface. It's very limited in functionality
    and probably not worth using.
    cs$tt|�jfi|��t|�|_dSr\)r�r�rQrr�r�r�r,r-rQ�szPyPIJSONLocator.__init__cCstd��dS�rizNot available from this locatorNrerZr,r,r-rj�sz&PyPIJSONLocator.get_distribution_namesc
Csiid�}t|jdt|��}�z�|j�|�}|����}t�|�}t	|j
d�}|d}|d|_|d|_|�
d�|_|�
dg�|_|�
d	�|_t|�}||_|d
}	|||j<|d
D]T}
|
d}|j�|�|�|
�|j|<|d
�|jt���|�|�|
�|d|<q�|d
��D]�\}}||jk�r4�qt	|j
d�}
|j|
_||
_t|
�}||_|||<|D]T}
|
d}|j�|�|�|
�|j|<|d
�|t���|�|�|
�|d|<�qh�qWnBt�y}z(|j�t|��t�d|�WYd}~n
d}~00|S)Nr�z%s/jsonr�r�rgr�r�r�r�r�r*r�ZreleaseszJSON fetch failed: %s) rr�rrM�open�read�decode�json�loadsrr4rgr�rSr�r�r�rr�r�r�r�r�r�r��itemsr�rP�putrrx�	exception)r9rgrWr*�respr�r�r�r�r�r�r��infosZomd�odistrXr,r,r-rh�sT




	"zPyPIJSONLocator._get_projectr�r,r,r�r-r��sr�c@s`eZdZdZe�dejejBejB�Z	e�dejejB�Z
dd�Ze�dej�Ze
dd��Zd	S)
�Pagez4
    This class represents a scraped HTML page.
    z�
(rel\s*=\s*(?:"(?P<rel1>[^"]*)"|'(?P<rel2>[^']*)'|(?P<rel3>[^>\s
]*))\s+)?
href\s*=\s*(?:"(?P<url1>[^"]*)"|'(?P<url2>[^']*)'|(?P<url3>[^>\s
]*))
(\s+rel\s*=\s*(?:"(?P<rel4>[^"]*)"|'(?P<rel5>[^']*)'|(?P<rel6>[^>\s
]*)))?
z!<base\s+href\s*=\s*['"]?([^'">]+)cCs4||_||_|_|j�|j�}|r0|�d�|_dS)zk
        Initialise an instance with the Unicode page contents and the URL they
        came from.
        rN)r�r�r*�_base�search�group)r9r�r*r�r,r,r-rQ s
z
Page.__init__z[^a-z0-9$&+,/:;=?@.#%_\\|-]cCs�dd�}t�}|j�|j�D]�}|�d�}|dpX|dpX|dpX|dpX|dpX|d	}|d
pp|dpp|d}t|j|�}t|�}|j�	d
d�|�}|�
||f�qt|dd�dd�}|S)z�
        Return the URLs of all the links on a page together with information
        about their "rel" attribute, for determining which ones to treat as
        downloads and which ones to queue for further scraping.
        cSs,t|�\}}}}}}t||t|�|||f�S)zTidy up an URL.)rrr)r*r4rsror�r�r�r,r,r-�clean4s�zPage.links.<locals>.cleanr2Zrel1Zrel2Zrel3Zrel4Zrel5Zrel6rzr{Zurl3cSsdt|�d��S)Nz%%%2xr)�ordr�)r�r,r,r-�<lambda>Br�zPage.links.<locals>.<lambda>cSs|dS)Nrr,)rtr,r,r-r�Fr�T)r@�reverse)r��_href�finditerr��	groupdictrr�r�	_clean_re�subr�r�)r9r�rWr�r��relr*r,r,r-�links-s$
���z
Page.linksN)rBrCrDrE�re�compile�I�S�Xr�r�rQr�rr�r,r,r,r-r�s�r�cs�eZdZdZejdd�dd�d�Zd�fdd	�	Zd
d�Zdd
�Z	dd�Z
e�dej
�Zdd�Zdd�Zdd�Zdd�Zdd�Ze�d�Zdd�Z�ZS)�SimpleScrapingLocatorz�
    A locator which scrapes HTML pages to locate downloads for a distribution.
    This runs multiple threads to do the I/O; performance is at least as good
    as pip's PackageFinder, which works in an analogous fashion.
    cCstjt|�d���S)N)�fileobj)�gzip�GzipFilerr���br,r,r-r�Tr�zSimpleScrapingLocator.<lambda>cCs|Sr\r,r�r,r,r-r�Ur�)�deflater��noneN�
csptt|�jfi|��t|�|_||_i|_t�|_t	�
�|_t�|_d|_
||_t��|_t��|_d|_dS)a�
        Initialise an instance.
        :param url: The root URL to use for scraping.
        :param timeout: The timeout, in seconds, to be applied to requests.
                        This defaults to ``None`` (no timeout specified).
        :param num_workers: The number of worker threads you want to do I/O,
                            This defaults to 10.
        :param kwargs: Passed to the superclass.
        FN)r�r�rQrr�r&�_page_cacher��_seenr
rO�	_to_fetch�
_bad_hosts�skip_externals�num_workers�	threading�RLock�_lock�_gplock�platform_check)r9r*r&rr�r�r,r-rQXs




zSimpleScrapingLocator.__init__cCsFg|_t|j�D]0}tj|jd�}|�d�|��|j�|�qdS)z�
        Threads are created only when get_project is called, and terminate
        before it returns. They are there primarily to parallelise I/O (i.e.
        fetching web pages).
        )�targetTN)	�_threads�rangerr�Thread�_fetch�	setDaemon�startrT)r9�irtr,r,r-�_prepare_threadsss
z&SimpleScrapingLocator._prepare_threadscCs6|jD]}|j�d�q|jD]}|��qg|_dS)zu
        Tell all the threads to terminate (by sending a sentinel value) and
        wait for them to do so.
        N)r
rr�r�)r9rtr,r,r-�
_wait_threads�s



z#SimpleScrapingLocator._wait_threadsc	Cs�iid�}|j��||_||_t|jdt|��}|j��|j��|�	�z.t
�d|�|j�
|�|j��W|��n
|��0|`Wd�n1s�0Y|S)Nr�z%s/zQueueing %s)rrWr�rr�rr�r]r�rrxryrr�r�r)r9rgrWr*r,r,r-rh�s


"z"SimpleScrapingLocator._get_projectz<\b(linux_(i\d86|x86_64|arm\w+)|win(32|_amd64)|macosx_?\d+)\bcCs|j�|�S)zD
        Does an URL refer to a platform-specific download?
        )�platform_dependentr�)r9r*r,r,r-�_is_platform_dependent�sz,SimpleScrapingLocator._is_platform_dependentcCsn|jr|�|�rd}n|�||j�}t�d||�|rj|j�|�|j|�Wd�n1s`0Y|S)a%
        See if an URL is a suitable download for a project.

        If it is, register information in the result dictionary (for
        _get_project) about the specific version it's for.

        Note that the return value isn't actually used other than as a boolean
        value.
        Nzprocess_download: %s -> %s)	rrr�r�rxryrr�rW)r9r*r�r,r,r-�_process_download�s
,z'SimpleScrapingLocator._process_downloadc
Cs�t|�\}}}}}}|�|j|j|j�r2d}n||jrJ|�|j�sJd}nd|�|j�s\d}nR|dvrjd}nD|dvrxd}n6|�|�r�d}n&|�	dd�d}	|	�
�dkr�d}nd}t�d	||||�|S)
z�
        Determine whether a link URL from a referring page and with a
        particular "rel" attribute should be queued for scraping.
        F)Zhomepage�download)�httprl�ftp�:rr�	localhostTz#should_queue: %s (%s) from %s -> %s)
rrpr�r�r�rr�r�r�splitr�rxry)
r9�linkZreferrerr�r4rsro�_rW�hostr,r,r-�
_should_queue�s.�

�z#SimpleScrapingLocator._should_queuec
Cs|j��}z�z�|r�|�|�}|dur4WW|j��q|jD]h\}}||jvr:zB|j�|�|�|�s�|�|||�r�t	�
d||�|j�|�Wq:ty�Yq:0q:Wn4t
y�}z|j�t|��WYd}~n
d}~00W|j��n|j��0|sq�qdS)z�
        Get a URL to fetch from the work queue, get the HTML page, examine its
        links for download candidates and candidates for further scraping.

        This is a handy method to run in a thread.
        NzQueueing %s from %s)rrS�get_pagerVr�r�r�rrrxryr�rr�rPr)r9r*�pagerr�rXr,r,r-r
�s0


�

�(zSimpleScrapingLocator._fetchcCs|t|�\}}}}}}|dkr:tj�t|��r:tt|�d�}||jvr`|j|}t�	d||��n|�
dd�d}d}||jvr�t�	d||��n�t|d	d
id�}�z�z�t�	d|�|j
j||jd
�}	t�	d|�|	��}
|
�dd�}t�|��r�|	��}|	��}
|
�d�}|�r"|j|}||
�}
d}t�|�}|�r@|�d�}z|
�|�}
Wnt�yl|
�d�}
Yn0t|
|�}||j|<Wn�t�y�}z&|jdk�r�t�d||�WYd}~n�d}~0t �y*}zNt�d||�|j!�|j�"|�Wd�n1�s0YWYd}~n<d}~0t#�y\}zt�d||�WYd}~n
d}~00W||j|<n||j|<0|S)a
        Get the HTML for an URL, possibly from an in-memory cache.

        XXX TODO Note: this cache is never actually cleared. It's assumed that
        the data won't get stale over the lifetime of a locator instance (not
        necessarily true for the default_locator).
        �filez
index.htmlzReturning %s from cache: %srrrNzSkipping %s due to bad host %szAccept-encoding�identity)r>zFetching %sr%z
Fetched %szContent-Typer2zContent-Encodingzutf-8zlatin-1i�zFetch failed: %s: %s)$r�osro�isdirrrrr�rxryrrrrMr�r&r�rS�HTML_CONTENT_TYPEr��geturlr��decoders�CHARSETr�r�r��UnicodeErrorr�rr<r�rrr�r�)r9r*r4rsrorrWrr:r�r>�content_typeZ	final_urlr��encoding�decoderr�rXr,r,r-r �sZ	







"@&zSimpleScrapingLocator.get_pagez<a href=[^>]*>([^<]+)<cCsLt�}|�|j�}|s$td|j��|j�|j�D]}|�|�d��q2|S)rizUnable to get %sr)	r�r r�r�_distname_rer�r�r�r�)r9rWr!r�r,r,r-rj7sz,SimpleScrapingLocator.get_distribution_names)Nr�)rBrCrDrE�zlib�
decompressr(rQrrrhr�r�r�rrrrr
r r.rjr�r,r,r�r-r�Js&�

�!;
r�cs8eZdZdZ�fdd�Zdd�Zdd�Zdd	�Z�ZS)
�DirectoryLocatorz?
    This class locates distributions in a directory tree.
    csR|�dd�|_tt|�jfi|��tj�|�}tj�|�sHt	d|��||_
dS)a�
        Initialise an instance.
        :param path: The root of the directory tree to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * recursive - if True (the default), subdirectories are
                         recursed into. If False, only the top-level directory
                         is searched,
        �	recursiveTzNot a directory: %rN)r�r2r�r1rQr$ro�abspathr%r�base_dir)r9ror�r�r,r-rQHs
zDirectoryLocator.__init__cCs|�|j�S)z�
        Should a filename be considered as a candidate for a distribution
        archive? As well as the filename, the directory which contains it
        is provided, though not used by the current implementation.
        )rprq)r9r�parentr,r,r-�should_includeYszDirectoryLocator.should_includec		Cs�iid�}t�|j�D]t\}}}|D]Z}|�||�r$tj�||�}tddttj�|��dddf�}|�	||�}|r$|�
||�q$|jsq�q|S)Nr�r"r2)r$�walkr4r6ror�rr	r3r�r�r2)	r9rgrW�root�dirs�files�fnr*r�r,r,r-rhas
�zDirectoryLocator._get_projectc	Cs�t�}t�|j�D]v\}}}|D]\}|�||�r tj�||�}tddttj�	|��dddf�}|�
|d�}|r |�|d�q |jsq�q|S)rir"r2Nrg)
r�r$r7r4r6ror�rr	r3r�r�r2)r9rWr8r9r:r;r*r�r,r,r-rjqs�z'DirectoryLocator.get_distribution_names)	rBrCrDrErQr6rhrjr�r,r,r�r-r1Cs
r1c@s eZdZdZdd�Zdd�ZdS)�JSONLocatora
    This locator uses special extended metadata (not available on PyPI) and is
    the basis of performant dependency resolution in distlib. Other locators
    require archive downloads before dependencies can be determined! As you
    might imagine, that can be slow.
    cCstd��dSr�rerZr,r,r-rj�sz"JSONLocator.get_distribution_namescCs�iid�}t|�}|r�|�dg�D]�}|ddks"|ddkr@q"t|d|d|�d	d
�|jd�}|j}|d|_d
|vr�|d
r�d|d
f|_|�di�|_|�di�|_|||j	<|d�
|j	t���|d�q"|S)Nr�r:�ptype�sdistZ	pyversion�sourcergr�r�zPlaceholder for summary)r�r4r*r�r��requirements�exportsr�)
rrSrr4r�r�r��dependenciesrAr�r�r�r�)r9rgrWr�r�r�r�r,r,r-rh�s*
��

 zJSONLocator._get_projectN)rBrCrDrErjrhr,r,r,r-r<�sr<cs(eZdZdZ�fdd�Zdd�Z�ZS)�DistPathLocatorz�
    This locator finds installed distributions in a path. It can be useful for
    adding to an :class:`AggregatingLocator`.
    cs.tt|�jfi|��t|t�s$J�||_dS)zs
        Initialise an instance.

        :param distpath: A :class:`DistributionPath` instance to search.
        N)r�rCrQ�
isinstancer�distpath)r9rEr�r�r,r-rQ�szDistPathLocator.__init__cCsP|j�|�}|dur iid�}n,|j|d|jt|jg�id|jtdg�ii}|S)Nr�r�r�)rE�get_distributionr�r�r�)r9rgr�rWr,r,r-rh�s�zDistPathLocator._get_project)rBrCrDrErQrhr�r,r,r�r-rC�s
rCcsReZdZdZ�fdd�Z�fdd�Zdd�Zeej	j
e�Z	dd	�Zd
d�Z�Z
S)�AggregatingLocatorzI
    This class allows you to chain and/or merge a list of locators.
    cs.|�dd�|_||_tt|�jfi|��dS)a�
        Initialise an instance.

        :param locators: The list of locators to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * merge - if False (the default), the first successful
                         search from any of the locators is returned. If True,
                         the results from all locators are merged (this can be
                         slow).
        �mergeFN)r�rH�locatorsr�rGrQ)r9rIr�r�r,r-rQ�szAggregatingLocator.__init__cs&tt|���|jD]}|��qdSr\)r�rGr^rI�r9r�r�r,r-r^�s
zAggregatingLocator.clear_cachecCs||_|jD]
}||_qdSr\)r`rIr4)r9rbr�r,r,r-rc�s
zAggregatingLocator._set_schemecCs�i}|jD]�}|�|�}|r
|jr�|�di�}|�di�}|�|�|�d�}|r�|r�|��D]*\}}	||vr�|||	O<q^|	||<q^|�d�}
|r�|
r�|
�|�q
|jdur�d}n"d}|D]}|j�|�r�d}q�q�|r
|}q�q
|S)Nr�r�TF)rIrkrHrS�updater�rNr�)r9rgrWr�r�r:r�Zdfr�r��dd�foundr,r,r-rh�s8






zAggregatingLocator._get_projectc	Cs:t�}|jD](}z||��O}Wqty2Yq0q|Sr�)r�rIrjrf)r9rWr�r,r,r-rjs
z)AggregatingLocator.get_distribution_names)rBrCrDrErQr^rcr�rIr4�fgetrhrjr�r,r,r�r-rG�s,rGzhttps://pypi.org/simple/r$r%�legacyr�c@sLeZdZdZddd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	ddd�Z
dS)�DependencyFinderz0
    Locate dependencies for distributions.
    NcCs|pt|_t|jj�|_dS)zf
        Initialise an instance, using the specified locator
        to locate distributions.
        N)�default_locatorr�r r4rJr,r,r-rQ/s
zDependencyFinder.__init__cCsrt�d|�|j}||j|<||j||jf<|jD]:}t|�\}}t�d|||�|j�	|t
���||f�q2dS)z�
        Add a distribution to the finder. This will update internal information
        about who provides what.
        :param dist: The distribution to add.
        zadding distribution %szAdd to provided: %s, %s, %sN)rxryr@�
dists_by_name�distsr��providesr�providedr�r�r�)r9r�rg�pr�r,r,r-�add_distribution7s

z!DependencyFinder.add_distributioncCsxt�d|�|j}|j|=|j||jf=|jD]D}t|�\}}t�d|||�|j|}|�	||f�|s.|j|=q.dS)z�
        Remove a distribution from the finder. This will update internal
        information about who provides what.
        :param dist: The distribution to remove.
        zremoving distribution %sz Remove from provided: %s, %s, %sN)
rxryr@rRrSr�rTrrU�remove)r9r�rgrVr��sr,r,r-�remove_distributionFs

z$DependencyFinder.remove_distributioncCs@z|j�|�}Wn*ty:|��d}|j�|�}Yn0|S)z�
        Get a version matcher for a requirement.
        :param reqt: The requirement
        :type reqt: str
        :return: A version matcher (an instance of
                 :class:`distlib.version.Matcher`).
        r)r4rNr!r)r9�reqtrNrgr,r,r-�get_matcherXszDependencyFinder.get_matcherc		Csr|�|�}|j}t�}|j}||vrn||D]@\}}z|�|�}WntyXd}Yn0|r,|�|�qnq,|S)z�
        Find the distributions which can fulfill a requirement.

        :param reqt: The requirement.
         :type reqt: str
        :return: A set of distribution which can fulfill the requirement.
        F)r\r@r�rUr�r!r�)	r9r[rNrgrWrUr��providerr�r,r,r-�find_providershs


zDependencyFinder.find_providersc	Cs�|j|}t�}|D]$}|�|�}|�|j�s|�|�q|rZ|�d||t|�f�d}n@|�|�|j|=|D]}|j�|t���|�qp|�	|�d}|S)a�
        Attempt to replace one provider with another. This is typically used
        when resolving dependencies from multiple sources, e.g. A requires
        (B >= 1.0) while C requires (B >= 1.1).

        For successful replacement, ``provider`` must meet all the requirements
        which ``other`` fulfills.

        :param provider: The provider we are trying to replace with.
        :param other: The provider we're trying to replace.
        :param problems: If False is returned, this will contain what
                         problems prevented replacement. This is currently
                         a tuple of the literal string 'cantreplace',
                         ``provider``, ``other``  and the set of requirements
                         that ``provider`` couldn't fulfill.
        :return: True if we can replace ``other`` with ``provider``, else
                 False.
        ZcantreplaceFT)
�reqtsr�r\r�r�r��	frozensetrZr�rW)	r9r]�other�problemsZrlist�	unmatchedrYrNrWr,r,r-�try_to_replace�s$


�

zDependencyFinder.try_to_replaceFcCs�i|_i|_i|_i|_t|p g�}d|vrF|�d�|tgd��O}t|t�rf|}}t�	d|�n4|j
j||d�}}|dur�td|��t�	d|�d|_
t�}t|g�}t|g�}|�r�|��}|j}	|	|jvr�|�|�n"|j|	}
|
|k�r|�||
|�|j|jB}|j}t�}
|�r^||v�r^d	D]*}d
|}||v�r2|
t|d|�O}
�q2||B|
B}|D�].}|�|�}|�sDt�	d|�|j
j||d�}|du�r�|�s�|j
j|dd�}|du�r�t�	d
|�|�d|f�n^|j|j}}||f|jv�r|�|�|�|�||v�rD||v�rD|�|�t�	d|j�|D]R}|j}	|	|jv�rv|j�|t���|�n"|j|	}
|
|k�rH|�||
|��qH�qnq�t|j���}|D]&}||v|_|j�r�t�	d|j��q�t�	d|�||fS)a�
        Find a distribution and all distributions it depends on.

        :param requirement: The requirement specifying the distribution to
                            find, or a Distribution instance.
        :param meta_extras: A list of meta extras such as :test:, :build: and
                            so on.
        :param prereleases: If ``True``, allow pre-release versions to be
                            returned - otherwise, don't return prereleases
                            unless they're all that's available.

        Return a set of :class:`Distribution` instances and a set of
        problems.

        The distributions returned should be such that they have the
        :attr:`required` attribute set to ``True`` if they were
        from the ``requirement`` passed to ``find()``, and they have the
        :attr:`build_time_dependency` attribute set to ``True`` unless they
        are post-installation dependencies of the ``requirement``.

        The problems should be a tuple consisting of the string
        ``'unsatisfied'`` and the requirement which couldn't be satisfied
        by any distribution known to the locator.
        z:*:)z:test:z:build:z:dev:zpassed %s as requirement)r�NzUnable to locate %rz
located %sT)�test�build�devz:%s:z%s_requireszNo providers found for %rzCannot satisfy %rZunsatisfiedzAdding %s to install_distsz#%s is a build-time dependency only.zfind done for %s)rUrSrRr_r�rXrDrrxryr�r�r�	requestedr�r@rWrdZrun_requiresZ
meta_requiresZbuild_requires�getattrr^r�r�Zname_and_versionr��valuesZbuild_time_dependency)r9r�Zmeta_extrasr�r�r�rb�todoZ
install_distsrgraZireqtsZsreqtsZereqtsr@rXZ	all_reqtsr��	providersr]�nr�rVrSr,r,r-�find�s�

�












�


�zDependencyFinder.find)N)NF)rBrCrDrErQrWrZr\r^rdrnr,r,r,r-rP*s
(rP)N)Lr��iorr��loggingr$rmr�r�ImportErrorZdummy_threadingr/r2r�compatrrrrr	r
rrr
rr7rrrrZdatabaserrrr�rr�utilrrrrrrrrr�r r!r�r"r#�	getLoggerrBrxr�r�r�r)r&r(r.r/�objectrIr�r�r�r�r1r<rCrGrQr�rPr,r,r,r-�<module>s\@(



G0E:zA&[��site-packages/pip/_vendor/distlib/__pycache__/manifest.cpython-39.pyc000064400000023767150564745010021637 0ustar00a

��Re�9�@s�dZddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZdgZe�
e�Ze�dej�Ze�d	ejejB�Zejdd
�ZGdd�de�ZdS)zu
Class representing the list of files in a distribution.

Equivalent to distutils.filelist, but fixes some problems.
�N�)�DistlibException)�fsdecode��convert_path�Manifestz\\w*
z#.*?(?=
)|
(?=$)�c@szeZdZdZddd�Zdd�Zdd�Zd	d
�Zddd
�Zdd�Z	dd�Z
dd�Zddd�Zd dd�Z
d!dd�Zdd�ZdS)"rz~A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.
    NcCs>tj�tj�|pt����|_|jtj|_d|_t	�|_
dS)zd
        Initialise an instance.

        :param base: The base directory to explore under.
        N)�os�path�abspath�normpath�getcwd�base�sep�prefix�allfiles�set�files)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/manifest.py�__init__*szManifest.__init__cCs�ddlm}m}m}g|_}|j}|g}|j}|j}|r�|�}t�	|�}	|	D]R}
tj
�||
�}t�|�}|j}
||
�r�|�t
|��qN||
�rN||
�sN||�qNq6dS)zmFind all files under the base and set ``allfiles`` to the absolute
        pathnames of files found.
        r)�S_ISREG�S_ISDIR�S_ISLNKN)�statrrrrr�pop�appendr	�listdirr
�join�st_moder)rrrrr�root�stackr�push�names�name�fullnamer�moderrr�findall9s"


zManifest.findallcCs4|�|j�stj�|j|�}|j�tj�|��dS)zz
        Add a file to the manifest.

        :param item: The pathname to add. This can be relative to the base.
        N)	�
startswithrr	r
rrr�addr)r�itemrrrr*TszManifest.addcCs|D]}|�|�qdS)z�
        Add a list of files to the manifest.

        :param items: The pathnames to add. These can be relative to the base.
        N)r*)r�itemsr+rrr�add_many^szManifest.add_manyFcsb��fdd��t�j�}|rFt�}|D]}�|tj�|��q&||O}dd�tdd�|D��D�S)z8
        Return sorted files in directory order
        csJ|�|�t�d|�|�jkrFtj�|�\}}|dvs<J��||�dS)Nzadd_dir added %s)��/)r*�logger�debugrr	r
�split)�dirs�d�parent�_��add_dirrrrr8ls

z Manifest.sorted.<locals>.add_dircSsg|]}tjj|��qSr)r	r
r)�.0Z
path_tuplerrr�
<listcomp>z�z#Manifest.sorted.<locals>.<listcomp>css|]}tj�|�VqdS)N)r	r
r2)r9r
rrr�	<genexpr>{r;z"Manifest.sorted.<locals>.<genexpr>)rrr	r
�dirname�sorted)rZwantdirs�resultr3�frr7rr>gs
�zManifest.sortedcCst�|_g|_dS)zClear all collected files.N)rrr)rrrr�clear}szManifest.clearcCsj|�|�\}}}}|dkrB|D]}|j|dd�st�d|�q�n$|dkrf|D]}|j|dd�}qN�n|dkr�|D]}|j|dd�srt�d|�qrn�|d	kr�|D]}|j|dd�}q�n�|d
kr�|D] }|j||d�s�t�d||�q�n�|d
k�r
|D]}|j||d�}q�n\|dk�r2|jd|d��sft�d|�n4|dk�rZ|jd|d��sft�d|�ntd|��dS)av
        Process a directive which either adds some files from ``allfiles`` to
        ``files``, or removes some files from ``files``.

        :param directive: The directive to process. This should be in a format
                     compatible with distutils ``MANIFEST.in`` files:

                     http://docs.python.org/distutils/sourcedist.html#commands
        �includeT)�anchorzno files found matching %r�exclude�global-includeFz3no files found matching %r anywhere in distribution�global-exclude�recursive-include)rz-no files found matching %r under directory %r�recursive-exclude�graftNz no directories found matching %r�prunez4no previously-included directories found matching %rzinvalid action %r)�_parse_directive�_include_patternr0�warning�_exclude_patternr)r�	directive�action�patterns�thedirZ
dirpattern�pattern�foundrrr�process_directive�sN��

�
��zManifest.process_directivecCs|��}t|�dkr,|ddvr,|�dd�|d}d}}}|dvrxt|�dkr`td|��d	d
�|dd�D�}n~|dvr�t|�dkr�td
|��t|d�}dd
�|dd�D�}n:|dvr�t|�dkr�td|��t|d�}ntd|��||||fS)z�
        Validate a directive.
        :param directive: The directive to validate.
        :return: A tuple of action, patterns, thedir, dir_patterns
        rr)rBrDrErFrGrHrIrJrBN)rBrDrErFrz$%r expects <pattern1> <pattern2> ...cSsg|]}t|��qSrr�r9�wordrrrr:�r;z-Manifest._parse_directive.<locals>.<listcomp>)rGrH�z*%r expects <dir> <pattern1> <pattern2> ...cSsg|]}t|��qSrrrVrrrr:�r;)rIrJz!%r expects a single <dir_pattern>zunknown action %r)r2�len�insertrr)rrO�wordsrPrQrRZdir_patternrrrrK�s4���zManifest._parse_directiveTcCsPd}|�||||�}|jdur&|��|jD]}|�|�r,|j�|�d}q,|S)a�Select strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.

        Patterns are not quite the same as implemented by the 'fnmatch'
        module: '*' and '?'  match non-special characters, where "special"
        is platform-dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found.
        FNT)�_translate_patternrr(�searchrr*)rrSrCr�is_regexrT�
pattern_rer%rrrrLs


zManifest._include_patterncCsBd}|�||||�}t|j�D]}|�|�r|j�|�d}q|S)atRemove strings (presumably filenames) from 'files' that match
        'pattern'.

        Other parameters are the same as for 'include_pattern()', above.
        The list 'self.files' is modified in place. Return True if files are
        found.

        This API is public to allow e.g. exclusion of SCM subdirs, e.g. when
        packaging source distributions
        FT)r\�listrr]�remove)rrSrCrr^rTr_r@rrrrN)s
zManifest._exclude_patternc
Cs�|rt|t�rt�|�S|Stdkr:|�d��d�\}}}|rj|�|�}tdkrn|�|�rd|�|�snJ�nd}t�	t
j�|j
d��}	|du�rftdkr�|�d�}
|�|�dt|
��}n>|�|�}|�|�r�|�|�s�J�|t|�t|�t|��}t
j}t
jdk�rd}tdk�r4d|	|�|d|f�}n0|t|�t|�t|��}d	||	||||f}n8|�r�tdk�r�d|	|}nd
||	|t|�d�f}t�|�S)aTranslate a shell-like wildcard pattern to a compiled regular
        expression.

        Return the compiled regex.  If 'is_regex' true,
        then 'pattern' is directly compiled to a regex (if it's a string)
        or just returned as-is (assumes it's a regex object).
        )rXrr6r.N�\z\\�^z.*z%s%s%s%s.*%s%sz%s%s%s)�
isinstance�str�re�compile�_PYTHON_VERSION�_glob_to_re�	partitionr)�endswith�escaper	r
rrrYr)
rrSrCrr^�startr6�endr_rZ
empty_patternZ	prefix_rerrrrr\=sF	






�

�
zManifest._translate_patterncCs8t�|�}tj}tjdkrd}d|}t�d||�}|S)z�Translate a shell-like glob pattern to a regular expression.

        Return a string containing the regex.  Differs from
        'fnmatch.translate()' in that '*' does not match "special characters"
        (which are platform-specific).
        rbz\\\\z\1[^%s]z((?<!\\)(\\\\)*)\.)�fnmatch�	translater	rrf�sub)rrSr_r�escapedrrrrits

zManifest._glob_to_re)N)F)TNF)TNF)TNF)�__name__�
__module__�__qualname__�__doc__rr(r*r-r>rArUrKrLrNr\rirrrrr%s&

	
O/�
)�
�
7)rvro�loggingr	rf�sysr.r�compatr�utilr�__all__�	getLoggerrsr0rg�MZ_COLLAPSE_PATTERN�SZ_COMMENTED_LINE�version_inforh�objectrrrrr�<module>s
site-packages/pip/_vendor/distlib/__pycache__/markers.cpython-39.pyc000064400000011625150564745010021463 0ustar00a

��Re}�@s�dZddlZddlZddlZddlZddlmZddlmZm	Z	ddl
mZdgZ
e�d�Zd	d
�Zdd�ZGd
d�de�Zdd�Ze�Z[e�Zddd�ZdS)zG
Parser for the environment markers micro-language defined in PEP 508.
�N�)�string_types)�in_venv�parse_marker)�NormalizedVersion�	interpretz<((\d+(\.\d+)*\w*)|\'(\d+(\.\d+)*\w*)\'|\"(\d+(\.\d+)*\w*)\")cCst|t�r|sdS|ddvS)NFr�'")�
isinstancer)�o�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/markers.py�_is_literalsr
cCs2g}t�|�D]}|�t|��d��qt|�S)Nr)�_VERSION_PATTERN�finditer�append�NV�groups�set)�s�result�mrrr�
_get_versions!src
@sfeZdZdZdd�dd�dd�dd�dd�dd�d	d�d
d�dd�dd�d
d�dd�d�Zdd�ZdS)�	Evaluatorz;
    This class is used to evaluate marker expessions.
    cCs||kS�Nr��x�yrrr�<lambda>-�zEvaluator.<lambda>cCs||kSrrrrrrr.rcCs||kp||kSrrrrrrr/rcCs||kSrrrrrrr0rcCs||kSrrrrrrr1rcCs||kp||kSrrrrrrr2rcCs||kSrrrrrrr3rcCs||kp||kSrrrrrrr4rcCs|o|Srrrrrrr5rcCs|p|Srrrrrrr6rcCs||vSrrrrrrr7rcCs||vSrrrrrrr8r)�==�===�~=�!=�<�<=�>�>=�and�or�in�not inc	Cs"t|t�rB|ddvr$|dd�}n||vr8td|��||}n�t|t�sPJ�|d}||jvrntd|��|d}|d	}t|d�r�t|d	�r�td
|||f��|�||�}|�||�}|dks�|dkr�|dvr�t|�}t|�}n$|dk�r|d
v�rt|�}t	|�}|j|||�}|S)z�
        Evaluate a marker expression returned by the :func:`parse_requirement`
        function in the specified context.
        rrr���zunknown variable: %s�opzop not implemented: %s�lhs�rhszinvalid comparison: %s %s %s�python_version)r#r$r%r&r rr"r!)r)r*)
r	r�SyntaxError�dict�
operations�NotImplementedErrorr
�evaluaterr)	�self�expr�contextrr,ZelhsZerhsr-r.rrrr4;s4


�
zEvaluator.evaluateN)�__name__�
__module__�__qualname__�__doc__r2r4rrrrr's�rc
Cs�dd�}ttd�r(|tjj�}tjj}nd}d}||tjt��t��t�	�t�
�t��tt��t�
�t�
�dd�tjd�}|S)NcSs<d|j|j|jf}|j}|dkr8||dt|j�7}|S)Nz%s.%s.%s�finalr)�major�minor�micro�releaselevel�str�serial)�info�version�kindrrr�format_full_version^s
z,default_context.<locals>.format_full_version�implementation�0��)�implementation_name�implementation_version�os_name�platform_machine�platform_python_implementation�platform_release�platform_system�platform_versionZplatform_in_venv�python_full_versionr/�sys_platform)�hasattr�sysrGrD�name�os�platform�machine�python_implementation�release�systemrArr/)rFrLrKrrrr�default_context]s(

�r^c
Cs�zt|�\}}Wn4tyD}ztd||f��WYd}~n
d}~00|rf|ddkrftd||f��tt�}|r||�|�t�||�S)z�
    Interpret a marker and return a result depending on environment.

    :param marker: The marker to interpret.
    :type marker: str
    :param execution_context: The context used for name lookup.
    :type execution_context: mapping
    z)Unable to interpret marker syntax: %s: %sNr�#z*unexpected trailing data in marker: %s: %s)r�	Exceptionr0r1�DEFAULT_CONTEXT�update�	evaluatorr4)�markerZexecution_contextr6�rest�er7rrrr�s	&
)N)r;rX�rerVrY�compatr�utilrrrDrr�__all__�compilerr
r�objectrr^rarcrrrrr�<module>s"
6site-packages/pip/_vendor/distlib/__pycache__/wheel.cpython-39.pyc000064400000065107150564745010021127 0ustar00a

��Re���@s�ddlmZddlZddlZddlZddlmZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlmZmZddlmZmZmZmZmZddlmZddlmZmZm Z m!Z!dd	l"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*m+Z+m,Z,dd
l-m.Z.m/Z/e
�0e1�Z2da3e4ed��r8dZ5n*ej6�7d
��rLdZ5nej6dk�r^dZ5ndZ5e�8d�Z9e9�s�dej:dd�Z9de9Z;e5e9Z<e,��=dd��=dd�Z>e�8d�Z?e?�r�e?�7d��r�e?�=dd��@d�dZ?ndd�ZAeA�Z?[Ae
�Bde
jCe
jDB�ZEe
�Bde
jCe
jDB�ZFe
�Bd�ZGe
�Bd �ZHd!ZId"ZJejKd#k�rHd$d%�ZLnd&d%�ZLGd'd(�d(eM�ZNeN�ZOGd)d*�d*eM�ZPd+d,�ZQd-d.�ZReR�ZS[Rd1d/d0�ZTdS)2�)�unicode_literalsN)�message_from_file�)�__version__�DistlibException)�	sysconfig�ZipFile�fsdecode�	text_type�filter)�InstalledDistribution)�Metadata�METADATA_FILENAME�WHEEL_METADATA_FILENAME�LEGACY_METADATA_FILENAME)
�FileOperator�convert_path�	CSVReader�	CSVWriter�Cache�cached_property�get_cache_base�read_exports�tempdir�get_platform)�NormalizedVersion�UnsupportedVersionError�pypy_version_info�pp�java�jy�cli�ip�cp�py_version_nodotz%s%s��py�-�_�.�SOABIzcpython-cCsRdtg}t�d�r|�d�t�d�r0|�d�t�d�dkrH|�d�d	�|�S)
Nr#�Py_DEBUG�d�
WITH_PYMALLOC�m�Py_UNICODE_SIZE��u�)�
VER_SUFFIXr�get_config_var�append�join)�parts�r8��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/wheel.py�_derive_abi<s




r:zz
(?P<nm>[^-]+)
-(?P<vn>\d+[^-]*)
(-(?P<bn>\d+[^-]*))?
-(?P<py>\w+\d+(\.\w+\d+)*)
-(?P<bi>\w+)
-(?P<ar>\w+(\.\w+)*)
\.whl$
z7
(?P<nm>[^-]+)
-(?P<vn>\d+[^-]*)
(-(?P<bn>\d+[^-]*))?$
s
\s*#![^\r\n]*s^(\s*#!("[^"]+"|\S+))\s+(.*)$s#!pythons	#!pythonw�/cCs|S�Nr8��or8r8r9�<lambda>^�r?cCs|�tjd�S)Nr;)�replace�os�sepr=r8r8r9r?`r@c@s6eZdZdd�Zdd�Zdd�Zddd	�Zd
d�ZdS)
�MountercCsi|_i|_dSr<)�
impure_wheels�libs��selfr8r8r9�__init__dszMounter.__init__cCs||j|<|j�|�dSr<)rErF�update)rH�pathname�
extensionsr8r8r9�addhs
zMounter.addcCs0|j�|�}|D]\}}||jvr|j|=qdSr<)rE�poprF)rHrKrL�k�vr8r8r9�removels
zMounter.removeNcCs||jvr|}nd}|Sr<)rF)rH�fullname�path�resultr8r8r9�find_modulers
zMounter.find_modulecCsj|tjvrtj|}nP||jvr,td|��t�||j|�}||_|�dd�}t|�dkrf|d|_	|S)Nzunable to find extension for %sr)rr)
�sys�modulesrF�ImportError�impZload_dynamic�
__loader__�rsplit�len�__package__)rHrRrTr7r8r8r9�load_moduleys


zMounter.load_module)N)�__name__�
__module__�__qualname__rIrMrQrUr^r8r8r8r9rDcs

rDc@s�eZdZdZdZdZd4dd�Zedd	��Zed
d��Z	edd
��Z
edd��Zdd�Z
edd��Zdd�Zd5dd�Zdd�Zdd�Zdd�Zd6dd�Zd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd7d,d-�Zd.d/�Zd0d1�Zd8d2d3�ZdS)9�Wheelz@
    Class to build and install from Wheel files (PEP 427).
    )rr�sha256NFcCs8||_||_d|_tg|_dg|_dg|_t��|_	|durRd|_
d|_|j|_
n�t�|�}|r�|�d�}|d|_
|d�d	d
�|_|d|_|j|_
n�tj�|�\}}t�|�}|s�td|��|r�tj�|�|_	||_
|�d�}|d|_
|d|_|d|_|d
�d�|_|d�d�|_|d�d�|_dS)zB
        Initialise an instance using a (valid) filename.
        r2�none�anyN�dummyz0.1�nmZvnr(r'ZbnzInvalid name or filename: %rr&r)Zbi�ar)�signZ
should_verify�buildver�PYVER�pyver�abi�archrB�getcwd�dirname�name�version�filename�	_filename�NAME_VERSION_RE�match�	groupdictrArS�split�FILENAME_REr�abspath)rHrsri�verifyr.�inforpr8r8r9rI�sD







�



zWheel.__init__cCs^|jrd|j}nd}d�|j�}d�|j�}d�|j�}|j�dd�}d|j|||||fS)zJ
        Build and return a filename from the various components.
        r'r2r)r(z%s-%s%s-%s-%s-%s.whl)rjr6rlrmrnrrrArq)rHrjrlrmrnrrr8r8r9rs�s
�zWheel.filenamecCstj�|j|j�}tj�|�Sr<)rBrSr6rprs�isfile)rHrSr8r8r9�exists�szWheel.existsccs4|jD](}|jD]}|jD]}|||fVqqqdSr<)rlrmrn)rHrlrmrnr8r8r9�tags�s


z
Wheel.tagsc
Cs4tj�|j|j�}d|j|jf}d|}t�d�}t	|d���}|�
|�}|d�dd�}tdd	�|D��}t
tg}	d}
|	D]z}zbt�||�}|�|��8}
||
�}t|d
�}
|
r�Wd�Wq�Wd�n1s�0YWq|ty�Yq|0q||
�stdd�|	���Wd�n1�s&0Y|
S)
N�%s-%s�%s.dist-info�utf-8�r�
Wheel-Versionr)rcSsg|]}t|��qSr8��int��.0�ir8r8r9�
<listcomp>�r@z"Wheel.metadata.<locals>.<listcomp>)�fileobjz8Invalid wheel, because metadata is missing: looked in %sz, )rBrSr6rprsrqrr�codecs�	getreaderr�get_wheel_metadatarx�tuplerr�	posixpath�openr
�KeyError�
ValueError)rHrK�name_ver�info_dir�wrapper�zf�wheel_metadata�wv�file_version�fnsrT�fn�metadata_filename�bf�wfr8r8r9�metadata�s2


4�&zWheel.metadatacCsld|j|jf}d|}t�|d�}|�|��&}t�d�|�}t|�}Wd�n1sZ0Yt|�S)Nr�r��WHEELr�)	rqrrr�r6r�r�r�r�dict)rHr�r�r�r�r�r��messager8r8r9r��s&zWheel.get_wheel_metadatacCsJtj�|j|j�}t|d��}|�|�}Wd�n1s<0Y|S)Nr�)rBrSr6rprsrr�)rHrKr�rTr8r8r9r|s(z
Wheel.infocCs�t�|�}|r||��}|d|�||d�}}d|��vrBt}nt}t�|�}|rfd|��d}nd}||}||}nT|�d�}|�d�}	|dks�||	kr�d}
n|||d�d	kr�d	}
nd}
t|
|}|S)
Nspythonw� ���r@�
�
rr%s
)	�
SHEBANG_RErv�end�lower�SHEBANG_PYTHONW�SHEBANG_PYTHON�SHEBANG_DETAIL_RE�groups�find)rH�datar.r��shebangZdata_after_shebangZshebang_python�argsZcr�lfZtermr8r8r9�process_shebangs,




zWheel.process_shebangcCsf|dur|j}ztt|�}Wnty:td|��Yn0||���}t�|��d��	d�}||fS)NzUnsupported hash algorithm: %r�=�ascii)
�	hash_kind�getattr�hashlib�AttributeErrorr�digest�base64�urlsafe_b64encode�rstrip�decode)rHr�r��hasherrTr8r8r9�get_hash%szWheel.get_hashcCsjt|�}ttj�||��}|�|ddf�t|��$}|D]}|�|�q8Wd�n1s\0YdS)Nr2)�list�to_posixrBrS�relpathr5r�writerow)rH�records�record_path�base�p�writer�rowr8r8r9�write_record0s
zWheel.write_recordc	Cs�g}|\}}tt|j�}|D]d\}}	t|	d��}
|
��}Wd�n1sL0Yd|�|�}tj�|	�}
|�	|||
f�qtj�
|d�}	|�||	|�ttj�
|d��}|�	||	f�dS)N�rbz%s=%s�RECORD)
r�r�r�r��readr�rBrS�getsizer5r6r�r�)rHr|�libdir�
archive_pathsr��distinfor�r��apr��fr�r��sizer8r8r9�
write_records8s&zWheel.write_recordscCsZt|dtj��8}|D]"\}}t�d||�|�||�qWd�n1sL0YdS)N�wzWrote %s to %s in wheel)r�zipfile�ZIP_DEFLATED�logger�debug�write)rHrKr�r�r�r�r8r8r9�	build_zipHszWheel.build_zipc"s�|duri}tt�fdd�d��d}|dkrFd}tg}tg}tg}nd}tg}d	g}d
g}|�d|�|_|�d|�|_|�d
|�|_	�|}	d|j
|jf}
d|
}d|
}g}
dD�]
}|�vr�q��|}tj
�|�r�t�|�D]�\}}}|D]�}ttj
�||��}tj
�||�}ttj
�|||��}|
�||f�|dkr�|�d�s�t|d��}|��}Wd�n1�st0Y|�|�}t|d��}|�|�Wd�q�1�s�0Yq�q�q�|	}d}t�|�D]�\}}}||k�r:t|�D]8\}}t|�}|�d��r�tj
�||�}||=�q,�q�|�s:Jd��|D]H}t|��d��rV�q>tj
�||�}ttj
�||��}|
�||f��q>�q�t�|�}|D]B}|dv�r�ttj
�||��}ttj
�||��}|
�||f��q�d|�p�|jdtd|g}|jD] \}}}|�d|||f��qtj
�|d�}t|d�� }|�d �|��Wd�n1�sd0Yttj
�|d��}|
�||f�d!d"�} t|
| d#�}
|� ||f|	|
�tj
�|j!|j"�}!|�#|!|
�|!S)$z�
        Build a wheel from files in specified paths, and use any specified tags
        when determining the name of the wheel.
        Ncs|�vSr<r8r=��pathsr8r9r?Vr@zWheel.build.<locals>.<lambda>)�purelib�platlibrr��false�truerdrerlrmrnr��%s.datar�)r��headers�scriptsr��.exer��wb�
.dist-infoz(.dist-info directory expected, not found)z.pycz.pyo)r��	INSTALLERZSHAREDr�zWheel-Version: %d.%dzGenerator: distlib %szRoot-Is-Purelib: %sz
Tag: %s-%s-%sr�r��
cSs*|d}|�d�}d|vr"|d7}||fS)Nrr;r�i')�count)�tr��nr8r8r9�sorter�s

zWheel.build.<locals>.sorter)�key)$r�r�IMPVER�ABI�ARCHrk�getrlrmrnrqrrrBrS�isdir�walkr	r6r�r�r5�endswithr�r�r�r��	enumerate�listdir�
wheel_versionrr�sortedr�rprsr�)"rHr�rr�Zlibkey�is_pureZ
default_pyverZdefault_abiZdefault_archr�r��data_dirr�r�r�rS�root�dirs�filesr�r��rpr�r�r�r�r��dnr�rlrmrnr�rKr8r�r9�buildNs�
(
0



�0zWheel.buildcCs
|�d�S)zl
        Determine whether an archive entry should be skipped when verifying
        or installing.
        )r;z/RECORD.jws)r�)rH�arcnamer8r8r9�
skip_entry�szWheel.skip_entrycC
Ksx|j}|�d�}|�dd�}|�dd�}tj�|j|j�}d|j|jf}	d|	}
d|	}t	�|t
�}t	�|d�}
t	�|d	�}t�d
�}t
|d����}|�|
�� }||�}t|�}Wd�n1s�0Y|d
�dd�}tdd�|D��}||jk�r|�r||j|�|ddk�r(|d}n|d}i}|�|��X}t|d��,}|D]}|d}|||<�qPWd�n1�s~0YWd�n1�s�0Yt	�|
d�}t	�|d�}t	�|
dd�}t|d�}d|_tj} g}!t��}"|"|_d|_�zF�z|��D�],}#|#j}$t|$t��r,|$}%n
|$�d
�}%|� |%��rF�q||%}|d�rxt!|#j"�|dk�rxt#d|%��|d�r�|d�dd�\}&}'|�|$��}|�$�}(Wd�n1�s�0Y|�%|(|&�\})}*|*|'k�r�t#d|$��|�r|%�&||f��rt'�(d |%��q|%�&|��o,|%�)d!�}+|%�&|��rd|%�d"d�\})},}-tj�||,t*|-��}.n$|%|
|fv�rv�qtj�|t*|%��}.|+�s�|�|$��}|�+||.�Wd�n1�s�0Ytjd#k�r�t�,|.|#j-d$?d%@�|!�.|.�|�s\|d�r\t|.d&��>}|�$�}(|�%|(|&�\})}/|/|*k�r<t#d'|.��Wd�n1�sR0Y| �r:|.�)d(��r:z|j/|.|d)�}0|!�.|0�Wn"t0�y�t'j1d*dd+�Yn0n�tj�2t*|$��}1tj�|"|1�}2|�|$��}|�+||2�Wd�n1�s�0Ytj�|.�\}3}1|3|_|�3|1�}4|�4|4�|!�5|4��q|�rVt'�(d,�d}5�n�d}6|j6d
}|d-k�rPt	�|d.�}7z�|�|7��}t7|�}8Wd�n1�s�0Yi}6d/D]r}9d0|9}:|:|8v�r�i|6d1|9<};|8|:�8�D]<}<d2|<j9|<j:f}=|<j;�r|=d3d4�|<j;�7}=|=|;|<j<�q�q�Wnt0�yLt'�1d5�Yn0nxzX|�|��8}||�}t<�=|��d6�}6|6�r�|6�d7�}6Wd�n1�s�0YWnt0�y�t'�1d8�Yn0|6�r�|6�d9i�}>|6�d:i�}?|>�s�|?�r�|�dd�}@tj�>|@��st?d;��|@|_|>�@�D]*\}:}<d<|:|<f}A|�3|A�}4|�4|4��q"|?�r�d=di}B|?�@�D],\}:}<d<|:|<f}A|�3|A|B�}4|�4|4��qdtj�||�}tA|�}5tB|�}|d=|d=||d><|5�C||�}|�r�|!�.|�|5�D|!|d?|�|5WWtE�F|"�Wd�St0�y8t'�Gd@�|�H��Yn0WtE�F|"�ntE�F|"�0Wd�n1�sj0YdS)Aa~
        Install a wheel to the specified paths. If kwarg ``warner`` is
        specified, it should be a callable, which will be called with two
        tuples indicating the wheel version of this software and the wheel
        version in the file, if there is a discrepancy in the versions.
        This can be used to issue any warnings to raise any exceptions.
        If kwarg ``lib_only`` is True, only the purelib/platlib files are
        installed, and the headers, scripts, data and dist-info metadata are
        not written. If kwarg ``bytecode_hashed_invalidation`` is True, written
        bytecode will try to use file-hash based invalidation (PEP-552) on
        supported interpreter versions (CPython 2.7+).

        The return value is a :class:`InstalledDistribution` instance unless
        ``options.lib_only`` is True, in which case the return value is ``None``.
        �warner�lib_onlyFZbytecode_hashed_invalidationr�r�r�r�r�r�r�Nr�r)rcSsg|]}t|��qSr8r�r�r8r8r9r��r@z!Wheel.install.<locals>.<listcomp>zRoot-Is-Purelibr�r�r���streamrr2r�)�dry_runTr%�size mismatch for %s�=�digest mismatch for %szlib_only: skipping %sr�r;�posix�i�r�zdigest mismatch on write for %sz.py)�hashed_invalidationzByte-compilation failed)�exc_infozlib_only: returning Nonez1.0zentry_points.txt)�console�guiz
%s_scriptszwrap_%sz%s:%sz [%s]�,zAUnable to read legacy script metadata, so cannot generate scriptsrLzpython.commandsz8Unable to read JSON metadata, so cannot generate scriptsZwrap_consoleZwrap_guizValid script path not specifiedz%s = %sr�lib�prefixzinstallation failed.)Irr�rBrSr6rprsrqrrr�rr�r�rr�rrxr�r�rr�recordrV�dont_write_bytecode�tempfile�mkdtemp�
source_dir�
target_dir�infolist�
isinstancer
r�r�str�	file_sizerr�r��
startswithr�r�r�r�copy_stream�chmod�
external_attrr5�byte_compile�	Exception�warning�basename�make�set_executable_mode�extendr|r�valuesr�suffix�flags�json�loadr�r��itemsrr�Zwrite_shared_locationsZwrite_installed_files�shutil�rmtree�	exception�rollback)CrHr��maker�kwargsrrrZbc_hashed_invalidationrKr�r�r��
metadata_name�wheel_metadata_name�record_namer�r��bwfr�r�r�r�r�r�r��readerr�r�Zdata_pfxZinfo_pfxZ
script_pfx�fileopZbcZoutfiles�workdir�zinfor�	u_arcname�kind�valuer�r(r�Z	is_script�wherer�outfileZ	newdigestZpycr�Zworknamer�	filenames�dist�commands�epZepdatar�rOr,rP�s�console_scripts�gui_scriptsZ
script_dir�script�optionsr8r8r9�install�s\

&
L

�
(
�
�,

�&��,




(
.

�
�

z
Wheel.installcCs8tdur4tj�t�td�dtjdd��}t|�atS)Nzdylib-cachez%s.%sr%)	�cacherBrSr6rr rV�version_infor)rHr�r8r8r9�_get_dylib_cache�s�zWheel._get_dylib_cachec	Cs�tj�|j|j�}d|j|jf}d|}t�|d�}t�	d�}g}t
|d���.}�z|�|���}||�}	t�
|	�}
|��}|�|�}tj�|j|�}
tj�|
�s�t�|
�|
��D]�\}}tj�|
t|��}tj�|�s�d}n6t�|�j}tj�|�}|�|�}tj|j�}||k}|�r(|�||
�|�||f�q�Wd�n1�sN0YWnt�ynYn0Wd�n1�s�0Y|S)Nr�r�Z
EXTENSIONSr�r�T)rBrSr6rprsrqrrr�r�r�rr�r0r1rR�
prefix_to_dirr�r��makedirsr2rr~�stat�st_mtime�datetime�
fromtimestamp�getinfo�	date_time�extractr5r�)rHrKr�r�rr�rTr�r�r�rLrPrZ
cache_baserqr��destr[Z	file_timer|Z
wheel_timer8r8r9�_get_extensions�s>




4&zWheel._get_extensionscCst|�S)zM
        Determine if a wheel is compatible with the running system.
        )�
is_compatiblerGr8r8r9r^�szWheel.is_compatiblecCsdS)zP
        Determine if a wheel is asserted as mountable by its metadata.
        Tr8rGr8r8r9�is_mountable�szWheel.is_mountablecCs�tj�tj�|j|j��}|��s2d|}t|��|��sJd|}t|��|t	jvrbt
�d|�nN|rtt	j�|�nt	j�
d|�|��}|r�tt	jvr�t	j�t�t�||�dS)Nz)Wheel %s not compatible with this Python.z$Wheel %s is marked as not mountable.z%s already in pathr)rBrSrzr6rprsr^rr_rVr�r�r5�insertr]�_hook�	meta_pathrM)rHr5rK�msgrLr8r8r9�mount�s"

zWheel.mountcCsrtj�tj�|j|j��}|tjvr2t�d|�n<tj�	|�|t
jvrRt
�	|�t
jsnt
tjvrntj�	t
�dS)Nz%s not in path)
rBrSrzr6rprsrVr�r�rQrarErb)rHrKr8r8r9�unmounts



z
Wheel.unmountc
Csrtj�|j|j�}d|j|jf}d|}d|}t�|t�}t�|d�}t�|d�}t	�
d�}t|d����}	|	�|�� }
||
�}t
|�}Wd�n1s�0Y|d�d	d
�}
tdd�|
D��}i}|	�|��V}t|d
��*}|D]}|d}|||<q�Wd�n1�s0YWd�n1�s60Y|	��D�]}|j}t|t��rf|}n
|�d�}|�d�}d|v�r�td|��|�|��r��qH||}|d�r�t|j�|dk�r�td|��|d
�rH|d
�dd
�\}}|	�|��}|��}Wd�n1�s0Y|�||�\}}||k�rHtd|���qHWd�n1�sd0YdS)Nr�r�r�r�r�r�r�r�r)rcSsg|]}t|��qSr8r�r�r8r8r9r�(r@z Wheel.verify.<locals>.<listcomp>r	rr;�..�invalid entry in wheel: %rr%rr
r)rBrSr6rprsrqrrr�rr�r�rr�rrxr�rrrr
r�rrr r!r�r�)rHrKr�r�r�r9r:r;r�r�r<r�r�r�r�r�r�r=r�r�r@rrArBrCr�r(r�r8r8r9r{s\
&J


��
(
�zWheel.verifyc	Ksdd�}dd�}tj�|j|j�}d|j|jf}d|}t�|d�}	t����}
t	|d���}i}|�
�D]h}
|
j}t|t�r�|}n
|�
d	�}||	kr�qhd
|vr�td|��|�|
|
�tj�|
t|��}|||<qhWd�n1s�0Y|||�\}}||fi|��}|�r�|||�\}}|�r<||k�r<|||�|du�rftjd
d|
d�\}}t�|�n*tj�|��s�td|��tj�||j�}t|���}tj�|
|�}||f}|�||
|�|�||�|du�r�t�||�Wd�n1�s�0Y|S)a�
        Update the contents of a wheel in a generic way. The modifier should
        be a callable which expects a dictionary argument: its keys are
        archive-entry paths, and its values are absolute filesystem paths
        where the contents the corresponding archive entries can be found. The
        modifier is free to change the contents of the files pointed to, add
        new entries and remove entries, before returning. This method will
        extract the entire contents of the wheel to a temporary location, call
        the modifier, and then use the passed (and possibly updated)
        dictionary to write a new wheel. If ``dest_dir`` is specified, the new
        wheel is written there -- otherwise, the original wheel is overwritten.

        The modifier should return True if it updated the wheel, else False.
        This method returns the same value the modifier returns.
        cSsHd}}d|tf}||vr$d|}||vr@||}t|d�j}||fS)Nz%s/%sz%s/PKG-INFO�rS)rr
rr)�path_mapr�rrrSr�r8r8r9�get_version`sz!Wheel.update.<locals>.get_versioncSs�d}z|t|�}|�d�}|dkr*d|}nTdd�||dd��d�D�}|dd7<d	|d|�d�d
d�|D��f}Wnty�t�d|�Yn0|r�t|d
�}||_|�	t
�}|j||d�t�d||�dS)Nr'rz%s+1cSsg|]}t|��qSr8r�)r�rJr8r8r9r�rr@z8Wheel.update.<locals>.update_version.<locals>.<listcomp>rr)r�z%s+%scss|]}t|�VqdSr<)r r�r8r8r9�	<genexpr>ur@z7Wheel.update.<locals>.update_version.<locals>.<genexpr>z0Cannot update non-compliant (PEP-440) version %rrh)rS�legacyzVersion updated from %r to %r)rr�rxr6rr�r�r
rrr�rr�)rrrS�updatedrPr�r7Zmdrlr8r8r9�update_versionjs.

 �
�


�z$Wheel.update.<locals>.update_versionr�r�r�r�r�rfrgNz.whlz
wheel-update-)r.r�dirzNot a directory: %r)rBrSr6rprsrqrrr�rrrrr
r�rr[rr�mkstemp�closer�r�r2r�r�r3�copyfile)rH�modifier�dest_dirr8rjrnrKr�r�r;r?r�rir@rrArSZoriginal_versionr(�modified�current_version�fd�newpathr�r�r|r8r8r9rJOs\



�(

�

,zWheel.update)NFF)N)NN)F)N)r_r`ra�__doc__r�r�rI�propertyrsr~rrr�r�r|r�r�r�r�r�rrrOrRr]r^r_rdrer{rJr8r8r8r9rb�s@
)



	


tn	"
8rbcCsZddl}|��}g}|ddkrV|d�d�D]}|�|��rFt|�nd�q.t|�}|S)Nr�glibcrr))�platform�libc_verrxr5�isdigitr�r�)r|�verrTrJr8r8r9�_get_glibc_version�sr�c
Cshtg}td}ttjdddd�D]}|�d�|t|�g��q$g}t��D]*\}}}|�	d�rN|�|�
dd�d�qN|��tdkr�|�
dt�|�d�g}tg}tjd	k�r�t�d
t�}|�r�|��\}	}}}
t|�}|
g}|
dvr�|�d�|
d
v�r|�d�|
dv�r |�d�|
dv�r4|�d�|
dv�rH|�d�|dk�r�|D]*}d|	|||f}
|
tk�rV|�|
��qV|d8}�qH|D�]}|D�]}
|�d�t|df�||
f�|dk�r�tj�	d��r�|
�dd�}
t�}t|�dk�r�|dk�r$|�d�t|df�|d|
f�|dk�rP|�d�t|df�|d|
f�|dk�r||�d�t|df�|d|
f�|�d�t|df�|d|d|d|
ff��q��q�t|�D]L\}}|�d�t|f�ddf�|dk�r�|�d�t|df�ddf��q�t|�D]L\}}|�d�d |f�ddf�|dk�r|�d�d |df�ddf��qt|�S)!zG
    Return (pyver, abi, arch) tuples compatible with this Python.
    rrr�r2z.abir)r%rd�darwinz(\w+)_(\d+)_(\d+)_(\w+)$)�i386�ppc�fat)r�r��x86_64Zfat3)�ppc64r��fat64)r�r��intel)r�r�r�r�r��	universalz%s_%s_%s_%s�linuxZlinux_)r%�z
manylinux1_%s)r%�zmanylinux2010_%s)r%�zmanylinux2014_%szmanylinux_%s_%s_%srer&)r3�rangerVrQr5r6r rYZget_suffixesr"rx�sortr�r`r�r|�rervr�r��
IMP_PREFIXrAr�r\r��set)�versions�major�minor�abisr.r(rT�archesr.rqrn�matchesrvrJrmr7r�rrr8r8r9�compatible_tags�s�















�
�
���
"
"r�cCs\t|t�st|�}d}|dur"t}|D]0\}}}||jvr&||jvr&||jvr&d}qXq&|S)NFT)rrb�COMPATIBLE_TAGSrlrmrn)�wheelrrTrrmrnr8r8r9r^s
r^)N)U�
__future__rr�r�rW�emailrr�rYr0�loggingrBr�r�r3rVrr�r2rr�compatrrr	r
rZdatabaserr�r
rrr�utilrrrrrrrrrrrrrr�	getLoggerr_r�rP�hasattrr�r|r"r4r3rQrkr�rAr�r�rxr:�compile�
IGNORECASE�VERBOSEryrur�r�r�r�rCr��objectrDrarbr�r�r�r^r8r8r8r9�<module>s�0


	
�

�


#1
Psite-packages/pip/_vendor/distlib/__pycache__/index.cpython-39.pyc000064400000041660150564745010021130 0ustar00a

��ReQ�@s�ddlZddlZddlZddlZddlZddlZzddlmZWney^ddl	mZYn0ddl
mZddlm
Z
mZmZmZmZmZddlmZmZe�e�ZdZdZGd	d
�d
e�ZdS)�N)�Thread�)�DistlibException)�HTTPBasicAuthHandler�Request�HTTPPasswordMgr�urlparse�build_opener�string_types)�zip_dir�ServerProxyzhttps://pypi.org/pypi�pypic@s�eZdZdZdZd*dd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zd+dd�Zd,dd�Z
d-dd�Zd.dd�Zdd�Zd/dd�Zd0d d!�Zd1d"d#�Zd$d%�Zd&d'�Zd2d(d)�ZdS)3�PackageIndexzc
    This class represents a package index compatible with PyPI, the Python
    Package Index.
    s.----------ThIs_Is_tHe_distlib_index_bouNdaRY_$Nc
Cs�|pt|_|��t|j�\}}}}}}|s<|s<|s<|dvrJtd|j��d|_d|_d|_d|_t	t
jd��Z}dD]D}	z,tj
|	dg||d�}
|
dkr�|	|_Wq�Wqtty�Yqt0qtWd�n1s�0YdS)	z�
        Initialise an instance.

        :param url: The URL of the index. If not specified, the URL for PyPI is
                    used.
        )�http�httpszinvalid repository: %sN�w)�gpgZgpg2z	--version��stdout�stderrr)�
DEFAULT_INDEX�url�read_configurationrr�password_handler�ssl_verifierr�gpg_home�open�os�devnull�
subprocess�
check_call�OSError)�selfr�scheme�netloc�path�params�query�fragZsink�s�rc�r+��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/index.py�__init__$s(
�
zPackageIndex.__init__cCsddlm}|�S)zs
        Get the distutils command for interacting with PyPI configurations.
        :return: the command.
        r)�_get_pypirc_command)�utilr.)r"�cmdr+r+r,r.Asz PackageIndex._get_pypirc_commandcCsNddlm}||�}|�d�|_|�d�|_|�dd�|_|�d|j�|_dS)	z�
        Read the PyPI access configuration as supported by distutils. This populates
        ``username``, ``password``, ``realm`` and ``url`` attributes from the
        configuration.
        r)�_load_pypirc�username�password�realmr
�
repositoryN)r/r1�getr2r3r4r)r"r1�cfgr+r+r,rIszPackageIndex.read_configurationcCs |��ddlm}||�dS)z�
        Save the PyPI access configuration. You must have set ``username`` and
        ``password`` attributes before calling this method.
        r)�
_store_pypircN)�check_credentialsr/r8)r"r8r+r+r,�save_configurationVszPackageIndex.save_configurationcCs\|jdus|jdurtd��t�}t|j�\}}}}}}|�|j||j|j�t|�|_	dS)zp
        Check that ``username`` and ``password`` have been set, and raise an
        exception if not.
        Nz!username and password must be set)
r2r3rrrr�add_passwordr4rr)r"Zpm�_r$r+r+r,r9_szPackageIndex.check_credentialscCs\|��|��|��}d|d<|�|��g�}|�|�}d|d<|�|��g�}|�|�S)aq
        Register a distribution on PyPI, using the provided metadata.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the distribution to be
                         registered.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        �verify�:action�submit)r9�validate�todict�encode_request�items�send_request)r"�metadata�d�request�responser+r+r,�registerks

zPackageIndex.registercCsF|��}|sq:|�d���}|�|�t�d||f�q|��dS)ar
        Thread runner for reading lines of from a subprocess into a buffer.

        :param name: The logical name of the stream (used for logging only).
        :param stream: The stream to read from. This will typically a pipe
                       connected to the output stream of a subprocess.
        :param outbuf: The list to append the read lines to.
        �utf-8z%s: %sN)�readline�decode�rstrip�append�logger�debug�close)r"�name�streamZoutbufr)r+r+r,�_readers

zPackageIndex._readerc	Cs�|jdddg}|dur|j}|r.|�d|g�|durD|�gd��t��}tj�|tj�|�d�}|�dd	d
|d||g�t	�
dd
�|��||fS)a�
        Return a suitable command for signing a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The signing command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        �--status-fd�2�--no-ttyN�	--homedir)z--batchz--passphrase-fd�0z.ascz
--detach-signz--armorz--local-userz--output�invoking: %s� )rr�extend�tempfile�mkdtemprr%�join�basenamerOrP)r"�filename�signer�
sign_password�keystorer0�tdZsfr+r+r,�get_sign_command�s
�zPackageIndex.get_sign_commandc	Cs�tjtjd�}|dur tj|d<g}g}tj|fi|��}t|jd|j|fd�}|��t|jd|j|fd�}|��|dur�|j�	|�|j�
�|��|��|��|j
||fS)a�
        Run a command in a child process , passing it any input data specified.

        :param cmd: The command to run.
        :param input_data: If specified, this must be a byte string containing
                           data to be sent to the child process.
        :return: A tuple consisting of the subprocess' exit code, a list of
                 lines read from the subprocess' ``stdout``, and a list of
                 lines read from the subprocess' ``stderr``.
        rN�stdinr)�target�argsr)r�PIPE�PopenrrTr�startrrg�writerQ�waitr_�
returncode)	r"r0Z
input_data�kwargsrr�p�t1�t2r+r+r,�run_command�s&�

zPackageIndex.run_commandc
CsD|�||||�\}}|�||�d��\}}}	|dkr@td|��|S)aR
        Sign a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The absolute pathname of the file where the signature is
                 stored.
        rJrz&sign command failed with error code %s)rfrt�encoder)
r"rarbrcrdr0�sig_filer*rrr+r+r,�	sign_file�s
��
�zPackageIndex.sign_file�sdist�sourcec	CsR|��tj�|�s td|��|��|��}d}	|rZ|jsJt�	d�n|�
||||�}	t|d��}
|
��}Wd�n1s�0Yt
�|���}t
�|���}
|�dd||||
d��dtj�|�|fg}|	�r8t|	d��}
|
��}Wd�n1�s0Y|�d	tj�|	�|f�t�tj�|	��|�|��|�}|�|�S)
a�
        Upload a release file to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the file to be uploaded.
        :param filename: The pathname of the file to be uploaded.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param filetype: The type of the file being uploaded. This is the
                        distutils command which produced that file, e.g.
                        ``sdist`` or ``bdist_wheel``.
        :param pyversion: The version of Python which the release relates
                          to. For code compatible with any Python, this would
                          be ``source``, otherwise it would be e.g. ``3.2``.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        z
not found: %sNz)no signing program available - not signed�rbZfile_upload�1)r>Zprotocol_version�filetype�	pyversion�
md5_digest�
sha256_digest�contentZ
gpg_signature)r9rr%�existsrr@rArrO�warningrwr�read�hashlib�md5�	hexdigest�sha256�updater`rN�shutil�rmtree�dirnamerBrCrD)r"rErarbrcr|r}rdrFrv�fZ	file_datar~r�filesZsig_datarGr+r+r,�upload_file�sD
�&�(�zPackageIndex.upload_filec
Cs�|��tj�|�s td|��tj�|d�}tj�|�sFtd|��|��|j|j	}}t
|���}dd|fd|fg}d||fg}|�||�}	|�
|	�S)a2
        Upload documentation to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the documentation to be
                         uploaded.
        :param doc_dir: The pathname of the directory which contains the
                        documentation. This should be the directory that
                        contains the ``index.html`` for the documentation.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        znot a directory: %rz
index.htmlz
not found: %r)r>Z
doc_uploadrR�versionr�)r9rr%�isdirrr_r�r@rRr�r�getvaluerBrD)
r"rEZdoc_dir�fnrRr��zip_data�fieldsr�rGr+r+r,�upload_documentation!s
�z!PackageIndex.upload_documentationcCsT|jdddg}|dur|j}|r.|�d|g�|�d||g�t�dd�|��|S)	a|
        Return a suitable command for verifying a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The verifying command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        rUrVrWNrXz--verifyrZr[)rrr\rOrPr_)r"�signature_filename�
data_filenamerdr0r+r+r,�get_verify_command=szPackageIndex.get_verify_commandcCsH|jstd��|�|||�}|�|�\}}}|dvr@td|��|dkS)a6
        Verify a signature for a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: True if the signature was verified, else False.
        z0verification unavailable because gpg unavailable)rrz(verify command failed with error code %sr)rrr�rt)r"r�r�rdr0r*rrr+r+r,�verify_signatureUs��zPackageIndex.verify_signaturec	Cs�|durd}t�d�n6t|ttf�r0|\}}nd}tt|��}t�d|�t|d���}|�t	|��}z�|�
�}	d}
d}d}d}
d	|	vr�t|	d
�}|r�||
|
|�|�|
�}|s�q�|t
|�7}|�|�|r�|�|�|
d7}
|r�||
|
|�q�W|��n
|��0Wd�n1�s 0Y|dk�rN||k�rNtd||f��|�r�|��}||k�rztd
||||f��t�d|�dS)a
        This is a convenience method for downloading a file from an URL.
        Normally, this will be a file from the index, though currently
        no check is made for this (i.e. a file can be downloaded from
        anywhere).

        The method is just like the :func:`urlretrieve` function in the
        standard library, except that it allows digest computation to be
        done during download and checking that the downloaded data
        matched any expected value.

        :param url: The URL of the file to be downloaded (assumed to be
                    available via an HTTP GET request).
        :param destfile: The pathname where the downloaded file is to be
                         saved.
        :param digest: If specified, this must be a (hasher, value)
                       tuple, where hasher is the algorithm used (e.g.
                       ``'md5'``) and ``value`` is the expected value.
        :param reporthook: The same as for :func:`urlretrieve` in the
                           standard library.
        NzNo digest specifiedr�zDigest specified: %s�wbi ���rzcontent-lengthzContent-Lengthrz1retrieval incomplete: got only %d out of %d bytesz.%s digest mismatch for %s: expected %s, got %szDigest verified: %s)rOrP�
isinstance�list�tuple�getattrr�rrDr�info�intr��lenrmr�rQrr�)r"r�destfile�digest�
reporthookZdigesterZhasherZdfpZsfp�headers�	blocksize�sizer��blocknum�block�actualr+r+r,�
download_filens\



4��
��zPackageIndex.download_filecCs:g}|jr|�|j�|jr(|�|j�t|�}|�|�S)z�
        Send a standard library :class:`Request` to PyPI and return its
        response.

        :param req: The request to send.
        :return: The HTTP response from PyPI (a standard library HTTPResponse).
        )rrNrr	r)r"�req�handlers�openerr+r+r,rD�szPackageIndex.send_requestc
Cs�g}|j}|D]L\}}t|ttf�s*|g}|D]*}|�d|d|�d�d|�d�f�q.q|D].\}}	}
|�d|d||	f�d�d|
f�q`|�d|ddf�d�|�}d|}|tt|��d�}
t	|j
||
�S)	a&
        Encode fields and files for posting to an HTTP server.

        :param fields: The fields to send as a list of (fieldname, value)
                       tuples.
        :param files: The files to send as a list of (fieldname, filename,
                      file_bytes) tuple.
        s--z)Content-Disposition: form-data; name="%s"rJ�z8Content-Disposition: form-data; name="%s"; filename="%s"s
smultipart/form-data; boundary=)zContent-typezContent-length)�boundaryr�r�r�r\rur_�strr�rr)r"r�r��partsr��k�values�v�keyra�value�body�ctr�r+r+r,rB�sD���
���

�zPackageIndex.encode_requestcCsNt|t�rd|i}t|jdd�}z|�||p.d�W|d��S|d��0dS)NrRg@)�timeout�andrQ)r�r
rr�search)r"Zterms�operatorZ	rpc_proxyr+r+r,r��s

�zPackageIndex.search)N)N)N)N)NNrxryN)N)N)NN)N)�__name__�
__module__�__qualname__�__doc__r�r-r.rr:r9rIrTrfrtrwr�r�r�r�r�rDrBr�r+r+r+r,rs2

	�

#
�
9�
�

M+r)r��loggingrr�rr]�	threadingr�ImportErrorZdummy_threading�r�compatrrrrr	r
r/rr�	getLoggerr�rOr�
DEFAULT_REALM�objectrr+r+r+r,�<module>s  
site-packages/pip/_vendor/distlib/__pycache__/__init__.cpython-39.pyc000064400000002107150564745010021551 0ustar00a

��ReE�@slddlZdZGdd�de�ZzddlmZWn$eyPGdd�dej�ZYn0e�e�Z	e	�
e��dS)�Nz0.3.3c@seZdZdS)�DistlibExceptionN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/__init__.pyrsr)�NullHandlerc@s$eZdZdd�Zdd�Zdd�ZdS)rcCsdS�Nr��self�recordrrr�handle�zNullHandler.handlecCsdSr	rr
rrr�emitrzNullHandler.emitcCs
d|_dSr	)�lock)rrrr�
createLockrzNullHandler.createLockN)rrrr
rrrrrrrsr)�logging�__version__�	Exceptionrr�ImportError�Handler�	getLoggerr�logger�
addHandlerrrrr�<module>s
site-packages/pip/_vendor/distlib/__pycache__/metadata.cpython-39.pyc000064400000064002150564745010021574 0ustar00a

��Re�@s�dZddlmZddlZddlmZddlZddlZddlZddl	m
Z
mZddlm
Z
mZmZddlmZdd	lmZmZdd
lmZmZe�e�ZGdd�de
�ZGd
d�de
�ZGdd�de
�ZGdd�de
�Zgd�ZdZ dZ!e�"d�Z#e�"d�Z$dZ%dZ&dZ'dZ(dZ)dZ*dZ+e*dZ,d Z-e.�Z/e/�0e%�e/�0e&�e/�0e(�e/�0e*�e/�0e,�e�"d!�Z1d"d#�Z2d$d%�Z3d&d'�e/D�Z4d(d'�e4�5�D�Z6d)Z7d*Z8d+Z9d,Z:d-Z;d.Z<d/Z=e>�Z?e�"d0�Z@d;d2d3�ZAGd4d5�d5e>�ZBd6ZCd7ZDd8ZEGd9d:�d:e>�ZFdS)<z�Implementation of the Metadata for Python packages PEPs.

Supports all metadata formats (1.0, 1.1, 1.2, 1.3/2.1 and withdrawn 2.0).
�)�unicode_literalsN)�message_from_file�)�DistlibException�__version__)�StringIO�string_types�	text_type)�	interpret)�extract_by_key�
get_extras)�
get_scheme�PEP440_VERSION_REc@seZdZdZdS)�MetadataMissingErrorzA required metadata is missingN��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/metadata.pyrsrc@seZdZdZdS)�MetadataConflictErrorz>Attempt to read or write metadata fields that are conflictual.Nrrrrrr src@seZdZdZdS)� MetadataUnrecognizedVersionErrorz Unknown metadata version number.Nrrrrrr$src@seZdZdZdS)�MetadataInvalidErrorzA metadata value is invalidNrrrrrr(sr)�Metadata�PKG_INFO_ENCODING�PKG_INFO_PREFERRED_VERSION�utf-8�1.1z

       \|�	
        )�Metadata-Version�Name�Version�Platform�Summary�Description�Keywords�	Home-page�Author�Author-email�License)r r!r"r#�Supported-Platformr$r%r&r'r(r)r*�
Classifier�Download-URL�	Obsoletes�Provides�Requires)r.r/r0r,r-)r r!r"r#r+r$r%r&r'r(r)�
Maintainer�Maintainer-emailr*r,r-�Obsoletes-Dist�Project-URL�
Provides-Dist�
Requires-Dist�Requires-Python�Requires-External)r5r6r7r3r8r1r2r4)r r!r"r#r+r$r%r&r'r(r)r1r2r*r,r-r3r4r5r6r7r8�Private-Version�Obsoleted-By�Setup-Requires-Dist�	Extension�Provides-Extra)r9r=r:r;r<)�Description-Content-Typer0r/r.)r>z"extra\s*==\s*("([^"]+)"|'([^']+)')cCsZ|dkrtS|dkrtS|dkr$tS|dvrBttdd�tD��S|dkrNtSt|��dS)N�1.0r�1.2)�1.3�2.1css|]}|tvr|VqdS�N)�_345_FIELDS)�.0�frrr�	<genexpr>z�z%_version2fieldlist.<locals>.<genexpr>�2.0)�_241_FIELDS�_314_FIELDSrD�tuple�_566_FIELDS�_426_FIELDSr)�versionrrr�_version2fieldlistqsrPc
Cs:dd�}g}|��D]"\}}|gddfvr,q|�|�qgd�}|D]�}|tvrnd|vrn|�d�t�d|�|tvr�d|vr�|�d�t�d	|�|tvr�d
|vr�|�d
�t�d|�|tvr�d|vr�|�d�t�d
|�|tv�rd|v�r|dk�r|�d�t�d|�|t	vrDd|vrD|�d�t�d|�qDt
|�dk�rR|dSt
|�dk�rtt�d|�td��d|v�o�||t�}d
|v�o�||t
�}d|v�o�||t�}d|v�o�||t�}	t|�t|�t|�t|	�dk�r�td��|�s|�s|�s|	�st|v�rtS|�r"dS|�r,d
S|�r6dSdS)z5Detect the best version depending on the fields used.cSs|D]}||vrdSqdS)NTFr)�keys�markers�markerrrr�_has_marker�sz"_best_version.<locals>._has_marker�UNKNOWNN)r?rr@rArIrBr?zRemoved 1.0 due to %srzRemoved 1.1 due to %sr@zRemoved 1.2 due to %srAzRemoved 1.3 due to %srBr%zRemoved 2.1 due to %srIzRemoved 2.0 due to %srrz)Out of options - unknown metadata set: %szUnknown metadata setz,You used incompatible 1.1/1.2/2.0/2.1 fields)�items�appendrJ�remove�logger�debugrKrDrMrN�lenr�_314_MARKERS�_345_MARKERS�_566_MARKERS�_426_MARKERS�intr)
�fieldsrTrQ�key�valueZpossible_versionsZis_1_1Zis_1_2Zis_2_1Zis_2_0rrr�
_best_version�s`






&
rdcCsi|]}|���dd�|�qS)�-�_)�lower�replace)rE�namerrr�
<dictcomp>�srjcCsi|]\}}||�qSrr)rE�attr�fieldrrrrj�rH)r6r3r5)r7)r")r#r,r.r0r/r3r5r6r8r4r+r;r=r<)r4)r&)r(r1r$r%z[^A-Za-z0-9.]+FcCs0|r$t�d|�}t�d|�dd��}d||fS)zhReturn the distribution name with version.

    If for_filename is true, return a filename-escaped form.re� �.z%s-%s)�	_FILESAFE�subrh)rirOZfor_filenamerrr�_get_name_and_version�srqc@s
eZdZdZd?dd�Zdd�Zdd	�Zd
d�Zdd
�Zdd�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zd@dd�Zdd�Zdd �Zd!d"�Zd#d$�ZdAd%d&�ZdBd'd(�ZdCd)d*�Zd+d,�Zefd-d.�ZdDd/d0�ZdEd1d2�Zd3d4�Zd5d6�Zd7d8�Zd9d:�Zd;d<�Z d=d>�Z!dS)F�LegacyMetadataaoThe legacy metadata of a release.

    Supports versions 1.0, 1.1, 1.2, 2.0 and 1.3/2.1 (auto-detected). You can
    instantiate the class with one of these arguments (or none):
    - *path*, the path to a metadata file
    - *fileobj* give a file-like object with metadata as content
    - *mapping* is a dict-like object
    - *scheme* is a version scheme name
    N�defaultcCsz|||g�d�dkrtd��i|_g|_d|_||_|durH|�|�n.|dur\|�|�n|durv|�|�|�	�dS)N��'path, fileobj and mapping are exclusive)
�count�	TypeError�_fieldsZrequires_filesZ
_dependencies�scheme�read�	read_file�update�set_metadata_version)�self�path�fileobj�mappingryrrr�__init__�s
zLegacyMetadata.__init__cCst|j�|jd<dS�Nr )rdrx�r~rrrr}
sz#LegacyMetadata.set_metadata_versioncCs|�d||f�dS)Nz%s: %s
)�write)r~r�rircrrr�_write_field
szLegacyMetadata._write_fieldcCs
|�|�SrC)�get�r~rirrr�__getitem__szLegacyMetadata.__getitem__cCs|�||�SrC)�set)r~rircrrr�__setitem__szLegacyMetadata.__setitem__cCs6|�|�}z|j|=Wnty0t|��Yn0dSrC)�
_convert_namerx�KeyError)r~ri�
field_namerrr�__delitem__s

zLegacyMetadata.__delitem__cCs||jvp|�|�|jvSrC)rxr�r�rrr�__contains__s
�zLegacyMetadata.__contains__cCs(|tvr|S|�dd���}t�||�S)Nrerf)�_ALL_FIELDSrhrg�_ATTR2FIELDr�r�rrrr�!szLegacyMetadata._convert_namecCs|tvs|tvrgSdS)NrU)�_LISTFIELDS�_ELEMENTSFIELDr�rrr�_default_value'szLegacyMetadata._default_valuecCs&|jdvrt�d|�St�d|�SdS)N�r?r�
)�metadata_version�_LINE_PREFIX_PRE_1_2rp�_LINE_PREFIX_1_2�r~rcrrr�_remove_line_prefix,s
z"LegacyMetadata._remove_line_prefixcCs|tvr||St|��dSrC)r��AttributeErrorr�rrr�__getattr__2szLegacyMetadata.__getattr__FcCst|d|d|�S)zhReturn the distribution name with version.

        If filesafe is true, return a filename-escaped form.r!r")rq)r~Zfilesaferrr�get_fullname=szLegacyMetadata.get_fullnamecCs|�|�}|tvS)z+return True if name is a valid metadata key)r�r�r�rrr�is_fieldCs
zLegacyMetadata.is_fieldcCs|�|�}|tvSrC)r�r�r�rrr�is_multi_fieldHs
zLegacyMetadata.is_multi_fieldcCs6tj|ddd�}z|�|�W|��n
|��0dS)z*Read the metadata values from a file path.�rr��encodingN)�codecs�openr{�close)r~�filepath�fprrrrzLszLegacyMetadata.readcCs�t|�}|d|jd<tD]p}||vr(q|tvrf|�|�}|tvrX|durXdd�|D�}|�||�q||}|dur|dkr|�||�q|��}|r�|n|d|d<dS)z,Read the metadata values from a file object.zmetadata-versionr NcSsg|]}t|�d���qS��,)rL�split�rErcrrr�
<listcomp>arHz,LegacyMetadata.read_file.<locals>.<listcomp>rUr%)rrxr�r��get_all�_LISTTUPLEFIELDSr��get_payload)r~Zfileob�msgrl�valuesrc�bodyrrrr{Ts
zLegacyMetadata.read_filecCs8tj|ddd�}z|�||�W|��n
|��0dS)z&Write the metadata fields to filepath.�wrr�N)r�r��
write_filer�)r~r��skip_unknownr�rrrr�pszLegacyMetadata.writecCs�|��t|d�D]�}|�|�}|r8|dgdgfvr8q|tvrV|�||d�|��q|tvr�|dkr�|jdvr~|�dd�}n|�dd�}|g}|t	vr�d	d
�|D�}|D]}|�|||�q�qdS)z0Write the PKG-INFO format data to a file object.r rUr�r%r�r�rz	
       |cSsg|]}d�|��qSr���joinr�rrrr��rHz-LegacyMetadata.write_file.<locals>.<listcomp>N)
r}rPr�r�r�r�r�r�rhr�)r~�
fileobjectr�rlr�rcrrrr�xs$

zLegacyMetadata.write_filecsv�fdd�}|sn@t|d�r:|��D]}||||�q$n|D]\}}|||�q>|rr|��D]\}}|||�q^dS)a�Set metadata values from the given iterable `other` and kwargs.

        Behavior is like `dict.update`: If `other` has a ``keys`` method,
        they are looped over and ``self[key]`` is assigned ``other[key]``.
        Else, ``other`` is an iterable of ``(key, value)`` iterables.

        Keys that don't match a metadata field or that have an empty value are
        dropped.
        cs"|tvr|r����|�|�dSrC)r�r�r�)rbrcr�rr�_set�sz#LegacyMetadata.update.<locals>._setrQN)�hasattrrQrV)r~�other�kwargsr��k�vrr�rr|�s

zLegacyMetadata.updatecCsh|�|�}|tvs|dkrNt|ttf�sNt|t�rHdd�|�d�D�}qzg}n,|tvrzt|ttf�szt|t�rv|g}ng}t�	t
j��r<|d}t|j
�}|tvr�|dur�|D](}|�|�d�d�s�t�d	|||�q�nb|tv�r|du�r|�|��s<t�d
|||�n0|tv�r<|du�r<|�|��s<t�d
|||�|tv�rZ|dk�rZ|�|�}||j|<dS)z"Control then set a metadata field.r#cSsg|]}|���qSr)�strip)rEr�rrrr��rHz&LegacyMetadata.set.<locals>.<listcomp>r�r!N�;rz$'%s': '%s' is not valid (field '%s')z.'%s': '%s' is not a valid version (field '%s')r%)r�r��
isinstance�listrLrr�r�rY�isEnabledFor�logging�WARNINGr
ry�_PREDICATE_FIELDS�is_valid_matcher�warning�_VERSIONS_FIELDS�is_valid_constraint_list�_VERSION_FIELDS�is_valid_version�_UNICODEFIELDSr�rx)r~rirc�project_nameryr�rrrr��sJ
�
�

���


zLegacyMetadata.setcCs�|�|�}||jvr*|tur&|�|�}|S|tvr@|j|}|S|tvr�|j|}|dur^gSg}|D].}|tvr~|�|�qf|�|d|df�qf|S|tvr�|j|}t	|t
�r�|�d�S|j|S)zGet a metadata field.Nrrr�)r�rx�_MISSINGr�r�r�r�rWr�r�rr�)r~rirsrc�res�valrrrr��s.







zLegacyMetadata.getcs�|��gg}}dD]}||vr|�|�q|rP|gkrPdd�|�}t|��dD]}||vrT|�|�qT|ddkr�||fSt|j���fdd�}t|ft�jft	�j
ffD]@\}}|D]2}	|�|	d	�}
|
d	ur�||
�s�|�d
|	|
f�q�q�||fS)zkCheck if the metadata is compliant. If strict is True then raise if
        no Name or Version are provided)r!r"zmissing required metadata: %s�, )r'r(r r@cs(|D]}��|�d�d�sdSqdS)Nr�rFT)r�r�)rcr��ryrr�are_valid_constraintssz3LegacyMetadata.check.<locals>.are_valid_constraintsNzWrong value for '%s': %s)r}rWr�rr
ryr�r�r�r�r�r�)r~�strict�missing�warningsrkr�r�ra�
controllerrlrcrr�r�check�s8

���zLegacyMetadata.checkcCsh|��t|d�}i}|D]F}|r.||jvrt|}|dkrL||||<qdd�||D�||<q|S)aReturn fields as a dict.

        Field names will be converted to use the underscore-lowercase style
        instead of hyphen-mixed case (i.e. home_page instead of Home-page).
        This is as per https://www.python.org/dev/peps/pep-0566/#id17.
        r �project_urlcSsg|]}d�|��qSr�r�)rE�urrrr�5rHz)LegacyMetadata.todict.<locals>.<listcomp>)r}rPrx�_FIELD2ATTR)r~Zskip_missingra�datar�rbrrr�todict"szLegacyMetadata.todictcCs8|ddkr$dD]}||vr||=q|d|7<dS)Nr r)r.r0r/r6r)r~�requirementsrlrrr�add_requirements9s
zLegacyMetadata.add_requirementscCstt|d��Sr�)r�rPr�rrrrQDszLegacyMetadata.keysccs|��D]
}|VqdSrC�rQ)r~rbrrr�__iter__GszLegacyMetadata.__iter__cs�fdd����D�S)Ncsg|]}�|�qSrr�rErbr�rrr�LrHz)LegacyMetadata.values.<locals>.<listcomp>r�r�rr�rr�KszLegacyMetadata.valuescs�fdd����D�S)Ncsg|]}|�|f�qSrrr�r�rrr�OrHz(LegacyMetadata.items.<locals>.<listcomp>r�r�rr�rrVNszLegacyMetadata.itemscCsd|jj|j|jfS)Nz
<%s %s %s>)�	__class__rrirOr�rrr�__repr__Qs�zLegacyMetadata.__repr__)NNNrs)F)F)F)N)F)F)"rrrrr�r}r�r�r�r�r�r�r�r�r�r�r�r�rzr{r�r�r|r�r�r�r�r�r�rQr�r�rVr�rrrrrr�s@�




,
,
rrzpydist.jsonz
metadata.json�METADATAc@s�eZdZdZe�d�Ze�dej�Ze	Z
e�d�ZdZde
Zdddd	�Zd
ZdZedfedfe
dfedfd�Zd
ZdHdd�Zed�ZdefZdefZdefdefeeedefeeeedefddd�
Z[[dd�ZdIdd�Zdd�Zed d!��Z ed"d#��Z!e!j"d$d#��Z!dJd%d&�Z#ed'd(��Z$ed)d*��Z%e%j"d+d*��Z%d,d-�Z&d.d/�Z'd0d1�Z(d2d3�Z)d4d5d6d7d8d9d:d;d<dd=�
Z*d>d?�Z+dKdBdC�Z,dDdE�Z-dFdG�Z.dS)Lrz�
    The metadata of a release. This implementation uses 2.0 (JSON)
    metadata where possible. If not possible, it wraps a LegacyMetadata
    instance which handles the key-value metadata format.
    z
^\d+(\.\d+)*$z!^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$z	.{1,2047}rIzdistlib (%s)r)�legacy)rirO�summaryzqname version license summary description author author_email keywords platform home_page classifiers download_urlzwextras run_requires test_requires build_requires dev_requires provides meta_requires obsoleted_by supports_environments)r�rirOr�)�_legacy�_dataryNrscCs@|||g�d�dkrtd��d|_d|_||_|durxz|�||�||_Wn(tytt||d�|_|��Yn0n�d}|r�t	|d��}|�
�}Wd�q�1s�0Yn|r�|�
�}|dur�|j|jd�|_nbt
|t�s�|�d�}zt�|�|_|�|j|�Wn.t�y:tt|�|d�|_|��Yn0dS)Nrtru)r�ry�rb�r��	generatorr)r�ry)rvrwr�r�ry�_validate_mappingrrr�validater�rz�METADATA_VERSION�	GENERATORr�r	�decode�json�loads�
ValueErrorr)r~rr�r�ryr�rFrrrr��s@
(�


�zMetadata.__init__)rirO�license�keywordsr�r6r;r=r,)r-N)r N)
�run_requires�build_requires�dev_requiresZ
test_requires�
meta_requires�extras�modules�
namespaces�exports�commands�classifiers�
source_urlr�c
CsXt�|d�}t�|d�}||v�r||\}}|jr^|durP|durHdn|�}n|j�|�}n�|durjdn|�}|dvr�|j�||�}n�t�}|}|j�d�}	|	�r
|dkr�|	�d|�}nP|dkr�|	�d�}	|	r�|	�||�}n,|	�d	�}	|	s�|j�d	�}	|	�r
|	�||�}||u�rT|}n:||v�r2t�||�}n"|j�rH|j�|�}n|j�|�}|S)
N�common_keys�mapped_keys�rrr�rr�
extensionsr�python.commandsr�python.details�python.exports)�object�__getattribute__r�r�r�)
r~rb�common�mapped�lk�maker�resultrc�sentinel�drrrr
�sD




zMetadata.__getattribute__cCsH||jvrD|j|\}}|p |j|vrD|�|�}|sDtd||f��dS)Nz.'%s' is an invalid value for the '%s' property)�SYNTAX_VALIDATORSry�matchr)r~rbrcry�pattern�
exclusions�mrrr�_validate_value�s

��zMetadata._validate_valuecCs*|�||�t�|d�}t�|d�}||vr�||\}}|jrV|durJt�||j|<nf|dvrj||j|<nR|j�di�}|dkr�||d<n2|dkr�|�di�}|||<n|�d	i�}|||<nh||vr�t�|||�nP|d
k�rt|t	��r|�
�}|�r|��}ng}|j�r||j|<n
||j|<dS)Nrrrrrr	rr
rr�)rrr
r��NotImplementedErrorr��
setdefault�__setattr__r�rr�r�)r~rbrcrrrrfrrrrr�s<




zMetadata.__setattr__cCst|j|jd�S�NT)rqrirOr�rrr�name_and_version%szMetadata.name_and_versioncCsF|jr|jd}n|j�dg�}d|j|jf}||vrB|�|�|S)Nr5�providesz%s (%s))r�r�rrirOrW)r~r�srrrr )s
zMetadata.providescCs |jr||jd<n
||jd<dS)Nr5r )r�r�r�rrrr 4sc
Cs�|jr|}n�g}t|pg|j�}|D]d}d|vr>d|vr>d}n8d|vrLd}n|�d�|v}|rv|�d�}|rvt||�}|r$|�|d�q$dD]F}d|}	|	|vr�|�|	�|j�d|g�}|�|j|||d��q�|S)	a�
        Base method to get dependencies, given a set of extras
        to satisfy and an optional environment context.
        :param reqts: A list of sometimes-wanted dependencies,
                      perhaps dependent on extras and environment.
        :param extras: A list of optional components being requested.
        :param env: An optional environment for marker evaluation.
        �extra�environmentT�requires)�build�dev�testz:%s:z%s_requires)r��env)	r�rr�r�r
�extendrXr��get_requirements)
r~�reqtsr�r(rr�includerSrb�errrr*;s2	


�
zMetadata.get_requirementscCs|jr|��S|jSrC)r��_from_legacyr�r�rrr�
dictionaryeszMetadata.dictionarycCs|jrt�nt|j|j�SdSrC)r�rrr��DEPENDENCY_KEYSr�rrr�dependencieskszMetadata.dependenciescCs|jrt�n|j�|�dSrC)r�rr�r|r�rrrr1rsc	Cs�|�d�|jkrt��g}|j��D]"\}}||vr$||vr$|�|�q$|rbdd�|�}t|��|��D]\}}|�|||�qjdS)Nr�zMissing metadata items: %sr�)	r�r�r�MANDATORY_KEYSrVrWr�rr)	r~r�ryr�rbrr�r�r�rrrr�yszMetadata._validate_mappingcCsB|jr.|j�d�\}}|s|r>t�d||�n|�|j|j�dS)NTz#Metadata: missing: %s, warnings: %s)r�r�rYr�r�r�ry)r~r�r�rrrr��s�zMetadata.validatecCs(|jr|j�d�St|j|j�}|SdSr)r�r�rr��
INDEX_KEYS)r~rrrrr��szMetadata.todictc
Cs�|jr|jrJ�|j|jd�}|j�d�}dD]*}||vr.|dkrHd}n|}||||<q.|�dg�}|dgkrtg}||d<d	}|D]*\}}||vr�||r�d
||ig||<q�|j|d<i}i}	|S)Nr�T)rirOr�r��description�
classifierr5rr&�r�))Z
requires_distr�)Zsetup_requires_distr�r$r )r�r�r�r�r�r�r )
r~rZlmdr��nk�kwrQ�ok�author�
maintainerrrrr.�s.�

zMetadata._from_legacyr!r"r*r$r%r'r(r)r-)
rirO)rr
r�r�r4)r�python.projectZproject_urlsZHome)rr<�contactsrri)rr<r=r�emailr)rr
rcCs�dd�}|jr|jrJ�t�}|j}|j��D]t\}}t|t�sV||vr�||||<q.|}d}|D]2}z||}Wqbttfy�d}Yq�Yqb0qb|r.|||<q.||j	|j
�}	||j|j�}
|j
r�t|j
�|d<t|	�|d<t|
�|d<|S)NcSs�t�}|D]|}|�d�}|�d�}|d}|D]V}|sF|sF|�|�q.d}|rVd|}|rp|rld||f}n|}|�d�||f��q.q
|S)Nr"r#r$r6z
extra == "%s"z(%s) and %sr�)r�r��addr�)�entriesr+r-r"r(Zrlistr�rSrrr�process_entries�s"

z,Metadata._to_legacy.<locals>.process_entriesTFr=r6r;)r�r�rr�LEGACY_MAPPINGrVr�rLr��
IndexErrorr�r�r�r�r��sorted)r~rArZnmdr7r9r�foundr�Zr1Zr2rrr�
_to_legacy�s2

zMetadata._to_legacyFTcCs�||g�d�dkrtd��|��|r`|jr4|j}n|��}|rP|j||d�q�|j||d�nr|jrp|��}n|j}|r�t	j
||dddd�nBt�|dd��$}t	j
||dddd�Wd�n1s�0YdS)	Nrz)Exactly one of path and fileobj is needed)r�Trt)�ensure_ascii�indent�	sort_keysr�r)
rvr�r�r�rFr�r�r.r�r��dumpr�r�)r~rr�r�r�Z	legacy_mdrrFrrrr��s*
��zMetadata.writecCs�|jr|j�|�nr|j�dg�}d}|D]}d|vr*d|vr*|}qHq*|durfd|i}|�d|�n t|d�t|�B}t|�|d<dS)Nr�r#r"r$r)r�r�r�r�insertr�rD)r~r�r��always�entryZrsetrrrr�
szMetadata.add_requirementscCs*|jpd}|jpd}d|jj|j||fS)Nz	(no name)z
no versionz<%s %s %s (%s)>)rirOr�rr�)r~rirOrrrr�s


�zMetadata.__repr__)NNNrs)N)NN)NNFT)/rrrr�re�compileZMETADATA_VERSION_MATCHER�IZNAME_MATCHERrZVERSION_MATCHERZSUMMARY_MATCHERr�rr�r2r3r0r�	__slots__r�r�rr�Z	none_list�dictZ	none_dictrr
rr�propertyrr �setterr*r/r1r�r�r�r.rBrFr�r�r�rrrrr[s�

���
-�+

'




*


	�
2
r)F)Gr�
__future__rr�r>rr�r�rNr6rr�compatrrr	rRr
�utilrrrOr
r�	getLoggerrrYrrrr�__all__rrrOr�r�rJrKr\rDr]rNr_rMr^r�r�r|ZEXTRA_RErPrdr�rVr�r�r�r�r�r�r�r�rr�rorqrrZMETADATA_FILENAMEZWHEEL_METADATA_FILENAMEZLEGACY_METADATA_FILENAMErrrrr�<module>sx









I�


jsite-packages/pip/_vendor/distlib/metadata.py000064400000114305150564745010015307 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Implementation of the Metadata for Python packages PEPs.

Supports all metadata formats (1.0, 1.1, 1.2, 1.3/2.1 and withdrawn 2.0).
"""
from __future__ import unicode_literals

import codecs
from email import message_from_file
import json
import logging
import re


from . import DistlibException, __version__
from .compat import StringIO, string_types, text_type
from .markers import interpret
from .util import extract_by_key, get_extras
from .version import get_scheme, PEP440_VERSION_RE

logger = logging.getLogger(__name__)


class MetadataMissingError(DistlibException):
    """A required metadata is missing"""


class MetadataConflictError(DistlibException):
    """Attempt to read or write metadata fields that are conflictual."""


class MetadataUnrecognizedVersionError(DistlibException):
    """Unknown metadata version number."""


class MetadataInvalidError(DistlibException):
    """A metadata value is invalid"""

# public API of this module
__all__ = ['Metadata', 'PKG_INFO_ENCODING', 'PKG_INFO_PREFERRED_VERSION']

# Encoding used for the PKG-INFO files
PKG_INFO_ENCODING = 'utf-8'

# preferred version. Hopefully will be changed
# to 1.2 once PEP 345 is supported everywhere
PKG_INFO_PREFERRED_VERSION = '1.1'

_LINE_PREFIX_1_2 = re.compile('\n       \\|')
_LINE_PREFIX_PRE_1_2 = re.compile('\n        ')
_241_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'License')

_314_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'License', 'Classifier', 'Download-URL', 'Obsoletes',
               'Provides', 'Requires')

_314_MARKERS = ('Obsoletes', 'Provides', 'Requires', 'Classifier',
                'Download-URL')

_345_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'Maintainer', 'Maintainer-email', 'License',
               'Classifier', 'Download-URL', 'Obsoletes-Dist',
               'Project-URL', 'Provides-Dist', 'Requires-Dist',
               'Requires-Python', 'Requires-External')

_345_MARKERS = ('Provides-Dist', 'Requires-Dist', 'Requires-Python',
                'Obsoletes-Dist', 'Requires-External', 'Maintainer',
                'Maintainer-email', 'Project-URL')

_426_FIELDS = ('Metadata-Version', 'Name', 'Version', 'Platform',
               'Supported-Platform', 'Summary', 'Description',
               'Keywords', 'Home-page', 'Author', 'Author-email',
               'Maintainer', 'Maintainer-email', 'License',
               'Classifier', 'Download-URL', 'Obsoletes-Dist',
               'Project-URL', 'Provides-Dist', 'Requires-Dist',
               'Requires-Python', 'Requires-External', 'Private-Version',
               'Obsoleted-By', 'Setup-Requires-Dist', 'Extension',
               'Provides-Extra')

_426_MARKERS = ('Private-Version', 'Provides-Extra', 'Obsoleted-By',
                'Setup-Requires-Dist', 'Extension')

# See issue #106: Sometimes 'Requires' and 'Provides' occur wrongly in
# the metadata. Include them in the tuple literal below to allow them
# (for now).
# Ditto for Obsoletes - see issue #140.
_566_FIELDS = _426_FIELDS + ('Description-Content-Type',
                             'Requires', 'Provides', 'Obsoletes')

_566_MARKERS = ('Description-Content-Type',)

_ALL_FIELDS = set()
_ALL_FIELDS.update(_241_FIELDS)
_ALL_FIELDS.update(_314_FIELDS)
_ALL_FIELDS.update(_345_FIELDS)
_ALL_FIELDS.update(_426_FIELDS)
_ALL_FIELDS.update(_566_FIELDS)

EXTRA_RE = re.compile(r'''extra\s*==\s*("([^"]+)"|'([^']+)')''')


def _version2fieldlist(version):
    if version == '1.0':
        return _241_FIELDS
    elif version == '1.1':
        return _314_FIELDS
    elif version == '1.2':
        return _345_FIELDS
    elif version in ('1.3', '2.1'):
        # avoid adding field names if already there
        return _345_FIELDS + tuple(f for f in _566_FIELDS if f not in _345_FIELDS)
    elif version == '2.0':
        return _426_FIELDS
    raise MetadataUnrecognizedVersionError(version)


def _best_version(fields):
    """Detect the best version depending on the fields used."""
    def _has_marker(keys, markers):
        for marker in markers:
            if marker in keys:
                return True
        return False

    keys = []
    for key, value in fields.items():
        if value in ([], 'UNKNOWN', None):
            continue
        keys.append(key)

    possible_versions = ['1.0', '1.1', '1.2', '1.3', '2.0', '2.1']

    # first let's try to see if a field is not part of one of the version
    for key in keys:
        if key not in _241_FIELDS and '1.0' in possible_versions:
            possible_versions.remove('1.0')
            logger.debug('Removed 1.0 due to %s', key)
        if key not in _314_FIELDS and '1.1' in possible_versions:
            possible_versions.remove('1.1')
            logger.debug('Removed 1.1 due to %s', key)
        if key not in _345_FIELDS and '1.2' in possible_versions:
            possible_versions.remove('1.2')
            logger.debug('Removed 1.2 due to %s', key)
        if key not in _566_FIELDS and '1.3' in possible_versions:
            possible_versions.remove('1.3')
            logger.debug('Removed 1.3 due to %s', key)
        if key not in _566_FIELDS and '2.1' in possible_versions:
            if key != 'Description':  # In 2.1, description allowed after headers
                possible_versions.remove('2.1')
                logger.debug('Removed 2.1 due to %s', key)
        if key not in _426_FIELDS and '2.0' in possible_versions:
            possible_versions.remove('2.0')
            logger.debug('Removed 2.0 due to %s', key)

    # possible_version contains qualified versions
    if len(possible_versions) == 1:
        return possible_versions[0]   # found !
    elif len(possible_versions) == 0:
        logger.debug('Out of options - unknown metadata set: %s', fields)
        raise MetadataConflictError('Unknown metadata set')

    # let's see if one unique marker is found
    is_1_1 = '1.1' in possible_versions and _has_marker(keys, _314_MARKERS)
    is_1_2 = '1.2' in possible_versions and _has_marker(keys, _345_MARKERS)
    is_2_1 = '2.1' in possible_versions and _has_marker(keys, _566_MARKERS)
    is_2_0 = '2.0' in possible_versions and _has_marker(keys, _426_MARKERS)
    if int(is_1_1) + int(is_1_2) + int(is_2_1) + int(is_2_0) > 1:
        raise MetadataConflictError('You used incompatible 1.1/1.2/2.0/2.1 fields')

    # we have the choice, 1.0, or 1.2, or 2.0
    #   - 1.0 has a broken Summary field but works with all tools
    #   - 1.1 is to avoid
    #   - 1.2 fixes Summary but has little adoption
    #   - 2.0 adds more features and is very new
    if not is_1_1 and not is_1_2 and not is_2_1 and not is_2_0:
        # we couldn't find any specific marker
        if PKG_INFO_PREFERRED_VERSION in possible_versions:
            return PKG_INFO_PREFERRED_VERSION
    if is_1_1:
        return '1.1'
    if is_1_2:
        return '1.2'
    if is_2_1:
        return '2.1'

    return '2.0'

# This follows the rules about transforming keys as described in
# https://www.python.org/dev/peps/pep-0566/#id17
_ATTR2FIELD = {
    name.lower().replace("-", "_"): name for name in _ALL_FIELDS
}
_FIELD2ATTR = {field: attr for attr, field in _ATTR2FIELD.items()}

_PREDICATE_FIELDS = ('Requires-Dist', 'Obsoletes-Dist', 'Provides-Dist')
_VERSIONS_FIELDS = ('Requires-Python',)
_VERSION_FIELDS = ('Version',)
_LISTFIELDS = ('Platform', 'Classifier', 'Obsoletes',
               'Requires', 'Provides', 'Obsoletes-Dist',
               'Provides-Dist', 'Requires-Dist', 'Requires-External',
               'Project-URL', 'Supported-Platform', 'Setup-Requires-Dist',
               'Provides-Extra', 'Extension')
_LISTTUPLEFIELDS = ('Project-URL',)

_ELEMENTSFIELD = ('Keywords',)

_UNICODEFIELDS = ('Author', 'Maintainer', 'Summary', 'Description')

_MISSING = object()

_FILESAFE = re.compile('[^A-Za-z0-9.]+')


def _get_name_and_version(name, version, for_filename=False):
    """Return the distribution name with version.

    If for_filename is true, return a filename-escaped form."""
    if for_filename:
        # For both name and version any runs of non-alphanumeric or '.'
        # characters are replaced with a single '-'.  Additionally any
        # spaces in the version string become '.'
        name = _FILESAFE.sub('-', name)
        version = _FILESAFE.sub('-', version.replace(' ', '.'))
    return '%s-%s' % (name, version)


class LegacyMetadata(object):
    """The legacy metadata of a release.

    Supports versions 1.0, 1.1, 1.2, 2.0 and 1.3/2.1 (auto-detected). You can
    instantiate the class with one of these arguments (or none):
    - *path*, the path to a metadata file
    - *fileobj* give a file-like object with metadata as content
    - *mapping* is a dict-like object
    - *scheme* is a version scheme name
    """
    # TODO document the mapping API and UNKNOWN default key

    def __init__(self, path=None, fileobj=None, mapping=None,
                 scheme='default'):
        if [path, fileobj, mapping].count(None) < 2:
            raise TypeError('path, fileobj and mapping are exclusive')
        self._fields = {}
        self.requires_files = []
        self._dependencies = None
        self.scheme = scheme
        if path is not None:
            self.read(path)
        elif fileobj is not None:
            self.read_file(fileobj)
        elif mapping is not None:
            self.update(mapping)
            self.set_metadata_version()

    def set_metadata_version(self):
        self._fields['Metadata-Version'] = _best_version(self._fields)

    def _write_field(self, fileobj, name, value):
        fileobj.write('%s: %s\n' % (name, value))

    def __getitem__(self, name):
        return self.get(name)

    def __setitem__(self, name, value):
        return self.set(name, value)

    def __delitem__(self, name):
        field_name = self._convert_name(name)
        try:
            del self._fields[field_name]
        except KeyError:
            raise KeyError(name)

    def __contains__(self, name):
        return (name in self._fields or
                self._convert_name(name) in self._fields)

    def _convert_name(self, name):
        if name in _ALL_FIELDS:
            return name
        name = name.replace('-', '_').lower()
        return _ATTR2FIELD.get(name, name)

    def _default_value(self, name):
        if name in _LISTFIELDS or name in _ELEMENTSFIELD:
            return []
        return 'UNKNOWN'

    def _remove_line_prefix(self, value):
        if self.metadata_version in ('1.0', '1.1'):
            return _LINE_PREFIX_PRE_1_2.sub('\n', value)
        else:
            return _LINE_PREFIX_1_2.sub('\n', value)

    def __getattr__(self, name):
        if name in _ATTR2FIELD:
            return self[name]
        raise AttributeError(name)

    #
    # Public API
    #

#    dependencies = property(_get_dependencies, _set_dependencies)

    def get_fullname(self, filesafe=False):
        """Return the distribution name with version.

        If filesafe is true, return a filename-escaped form."""
        return _get_name_and_version(self['Name'], self['Version'], filesafe)

    def is_field(self, name):
        """return True if name is a valid metadata key"""
        name = self._convert_name(name)
        return name in _ALL_FIELDS

    def is_multi_field(self, name):
        name = self._convert_name(name)
        return name in _LISTFIELDS

    def read(self, filepath):
        """Read the metadata values from a file path."""
        fp = codecs.open(filepath, 'r', encoding='utf-8')
        try:
            self.read_file(fp)
        finally:
            fp.close()

    def read_file(self, fileob):
        """Read the metadata values from a file object."""
        msg = message_from_file(fileob)
        self._fields['Metadata-Version'] = msg['metadata-version']

        # When reading, get all the fields we can
        for field in _ALL_FIELDS:
            if field not in msg:
                continue
            if field in _LISTFIELDS:
                # we can have multiple lines
                values = msg.get_all(field)
                if field in _LISTTUPLEFIELDS and values is not None:
                    values = [tuple(value.split(',')) for value in values]
                self.set(field, values)
            else:
                # single line
                value = msg[field]
                if value is not None and value != 'UNKNOWN':
                    self.set(field, value)

        # PEP 566 specifies that the body be used for the description, if
        # available
        body = msg.get_payload()
        self["Description"] = body if body else self["Description"]
        # logger.debug('Attempting to set metadata for %s', self)
        # self.set_metadata_version()

    def write(self, filepath, skip_unknown=False):
        """Write the metadata fields to filepath."""
        fp = codecs.open(filepath, 'w', encoding='utf-8')
        try:
            self.write_file(fp, skip_unknown)
        finally:
            fp.close()

    def write_file(self, fileobject, skip_unknown=False):
        """Write the PKG-INFO format data to a file object."""
        self.set_metadata_version()

        for field in _version2fieldlist(self['Metadata-Version']):
            values = self.get(field)
            if skip_unknown and values in ('UNKNOWN', [], ['UNKNOWN']):
                continue
            if field in _ELEMENTSFIELD:
                self._write_field(fileobject, field, ','.join(values))
                continue
            if field not in _LISTFIELDS:
                if field == 'Description':
                    if self.metadata_version in ('1.0', '1.1'):
                        values = values.replace('\n', '\n        ')
                    else:
                        values = values.replace('\n', '\n       |')
                values = [values]

            if field in _LISTTUPLEFIELDS:
                values = [','.join(value) for value in values]

            for value in values:
                self._write_field(fileobject, field, value)

    def update(self, other=None, **kwargs):
        """Set metadata values from the given iterable `other` and kwargs.

        Behavior is like `dict.update`: If `other` has a ``keys`` method,
        they are looped over and ``self[key]`` is assigned ``other[key]``.
        Else, ``other`` is an iterable of ``(key, value)`` iterables.

        Keys that don't match a metadata field or that have an empty value are
        dropped.
        """
        def _set(key, value):
            if key in _ATTR2FIELD and value:
                self.set(self._convert_name(key), value)

        if not other:
            # other is None or empty container
            pass
        elif hasattr(other, 'keys'):
            for k in other.keys():
                _set(k, other[k])
        else:
            for k, v in other:
                _set(k, v)

        if kwargs:
            for k, v in kwargs.items():
                _set(k, v)

    def set(self, name, value):
        """Control then set a metadata field."""
        name = self._convert_name(name)

        if ((name in _ELEMENTSFIELD or name == 'Platform') and
            not isinstance(value, (list, tuple))):
            if isinstance(value, string_types):
                value = [v.strip() for v in value.split(',')]
            else:
                value = []
        elif (name in _LISTFIELDS and
              not isinstance(value, (list, tuple))):
            if isinstance(value, string_types):
                value = [value]
            else:
                value = []

        if logger.isEnabledFor(logging.WARNING):
            project_name = self['Name']

            scheme = get_scheme(self.scheme)
            if name in _PREDICATE_FIELDS and value is not None:
                for v in value:
                    # check that the values are valid
                    if not scheme.is_valid_matcher(v.split(';')[0]):
                        logger.warning(
                            "'%s': '%s' is not valid (field '%s')",
                            project_name, v, name)
            # FIXME this rejects UNKNOWN, is that right?
            elif name in _VERSIONS_FIELDS and value is not None:
                if not scheme.is_valid_constraint_list(value):
                    logger.warning("'%s': '%s' is not a valid version (field '%s')",
                                   project_name, value, name)
            elif name in _VERSION_FIELDS and value is not None:
                if not scheme.is_valid_version(value):
                    logger.warning("'%s': '%s' is not a valid version (field '%s')",
                                   project_name, value, name)

        if name in _UNICODEFIELDS:
            if name == 'Description':
                value = self._remove_line_prefix(value)

        self._fields[name] = value

    def get(self, name, default=_MISSING):
        """Get a metadata field."""
        name = self._convert_name(name)
        if name not in self._fields:
            if default is _MISSING:
                default = self._default_value(name)
            return default
        if name in _UNICODEFIELDS:
            value = self._fields[name]
            return value
        elif name in _LISTFIELDS:
            value = self._fields[name]
            if value is None:
                return []
            res = []
            for val in value:
                if name not in _LISTTUPLEFIELDS:
                    res.append(val)
                else:
                    # That's for Project-URL
                    res.append((val[0], val[1]))
            return res

        elif name in _ELEMENTSFIELD:
            value = self._fields[name]
            if isinstance(value, string_types):
                return value.split(',')
        return self._fields[name]

    def check(self, strict=False):
        """Check if the metadata is compliant. If strict is True then raise if
        no Name or Version are provided"""
        self.set_metadata_version()

        # XXX should check the versions (if the file was loaded)
        missing, warnings = [], []

        for attr in ('Name', 'Version'):  # required by PEP 345
            if attr not in self:
                missing.append(attr)

        if strict and missing != []:
            msg = 'missing required metadata: %s' % ', '.join(missing)
            raise MetadataMissingError(msg)

        for attr in ('Home-page', 'Author'):
            if attr not in self:
                missing.append(attr)

        # checking metadata 1.2 (XXX needs to check 1.1, 1.0)
        if self['Metadata-Version'] != '1.2':
            return missing, warnings

        scheme = get_scheme(self.scheme)

        def are_valid_constraints(value):
            for v in value:
                if not scheme.is_valid_matcher(v.split(';')[0]):
                    return False
            return True

        for fields, controller in ((_PREDICATE_FIELDS, are_valid_constraints),
                                   (_VERSIONS_FIELDS,
                                    scheme.is_valid_constraint_list),
                                   (_VERSION_FIELDS,
                                    scheme.is_valid_version)):
            for field in fields:
                value = self.get(field, None)
                if value is not None and not controller(value):
                    warnings.append("Wrong value for '%s': %s" % (field, value))

        return missing, warnings

    def todict(self, skip_missing=False):
        """Return fields as a dict.

        Field names will be converted to use the underscore-lowercase style
        instead of hyphen-mixed case (i.e. home_page instead of Home-page).
        This is as per https://www.python.org/dev/peps/pep-0566/#id17.
        """
        self.set_metadata_version()

        fields = _version2fieldlist(self['Metadata-Version'])

        data = {}

        for field_name in fields:
            if not skip_missing or field_name in self._fields:
                key = _FIELD2ATTR[field_name]
                if key != 'project_url':
                    data[key] = self[field_name]
                else:
                    data[key] = [','.join(u) for u in self[field_name]]

        return data

    def add_requirements(self, requirements):
        if self['Metadata-Version'] == '1.1':
            # we can't have 1.1 metadata *and* Setuptools requires
            for field in ('Obsoletes', 'Requires', 'Provides'):
                if field in self:
                    del self[field]
        self['Requires-Dist'] += requirements

    # Mapping API
    # TODO could add iter* variants

    def keys(self):
        return list(_version2fieldlist(self['Metadata-Version']))

    def __iter__(self):
        for key in self.keys():
            yield key

    def values(self):
        return [self[key] for key in self.keys()]

    def items(self):
        return [(key, self[key]) for key in self.keys()]

    def __repr__(self):
        return '<%s %s %s>' % (self.__class__.__name__, self.name,
                               self.version)


METADATA_FILENAME = 'pydist.json'
WHEEL_METADATA_FILENAME = 'metadata.json'
LEGACY_METADATA_FILENAME = 'METADATA'


class Metadata(object):
    """
    The metadata of a release. This implementation uses 2.0 (JSON)
    metadata where possible. If not possible, it wraps a LegacyMetadata
    instance which handles the key-value metadata format.
    """

    METADATA_VERSION_MATCHER = re.compile(r'^\d+(\.\d+)*$')

    NAME_MATCHER = re.compile('^[0-9A-Z]([0-9A-Z_.-]*[0-9A-Z])?$', re.I)

    VERSION_MATCHER = PEP440_VERSION_RE

    SUMMARY_MATCHER = re.compile('.{1,2047}')

    METADATA_VERSION = '2.0'

    GENERATOR = 'distlib (%s)' % __version__

    MANDATORY_KEYS = {
        'name': (),
        'version': (),
        'summary': ('legacy',),
    }

    INDEX_KEYS = ('name version license summary description author '
                  'author_email keywords platform home_page classifiers '
                  'download_url')

    DEPENDENCY_KEYS = ('extras run_requires test_requires build_requires '
                       'dev_requires provides meta_requires obsoleted_by '
                       'supports_environments')

    SYNTAX_VALIDATORS = {
        'metadata_version': (METADATA_VERSION_MATCHER, ()),
        'name': (NAME_MATCHER, ('legacy',)),
        'version': (VERSION_MATCHER, ('legacy',)),
        'summary': (SUMMARY_MATCHER, ('legacy',)),
    }

    __slots__ = ('_legacy', '_data', 'scheme')

    def __init__(self, path=None, fileobj=None, mapping=None,
                 scheme='default'):
        if [path, fileobj, mapping].count(None) < 2:
            raise TypeError('path, fileobj and mapping are exclusive')
        self._legacy = None
        self._data = None
        self.scheme = scheme
        #import pdb; pdb.set_trace()
        if mapping is not None:
            try:
                self._validate_mapping(mapping, scheme)
                self._data = mapping
            except MetadataUnrecognizedVersionError:
                self._legacy = LegacyMetadata(mapping=mapping, scheme=scheme)
                self.validate()
        else:
            data = None
            if path:
                with open(path, 'rb') as f:
                    data = f.read()
            elif fileobj:
                data = fileobj.read()
            if data is None:
                # Initialised with no args - to be added
                self._data = {
                    'metadata_version': self.METADATA_VERSION,
                    'generator': self.GENERATOR,
                }
            else:
                if not isinstance(data, text_type):
                    data = data.decode('utf-8')
                try:
                    self._data = json.loads(data)
                    self._validate_mapping(self._data, scheme)
                except ValueError:
                    # Note: MetadataUnrecognizedVersionError does not
                    # inherit from ValueError (it's a DistlibException,
                    # which should not inherit from ValueError).
                    # The ValueError comes from the json.load - if that
                    # succeeds and we get a validation error, we want
                    # that to propagate
                    self._legacy = LegacyMetadata(fileobj=StringIO(data),
                                                  scheme=scheme)
                    self.validate()

    common_keys = set(('name', 'version', 'license', 'keywords', 'summary'))

    none_list = (None, list)
    none_dict = (None, dict)

    mapped_keys = {
        'run_requires': ('Requires-Dist', list),
        'build_requires': ('Setup-Requires-Dist', list),
        'dev_requires': none_list,
        'test_requires': none_list,
        'meta_requires': none_list,
        'extras': ('Provides-Extra', list),
        'modules': none_list,
        'namespaces': none_list,
        'exports': none_dict,
        'commands': none_dict,
        'classifiers': ('Classifier', list),
        'source_url': ('Download-URL', None),
        'metadata_version': ('Metadata-Version', None),
    }

    del none_list, none_dict

    def __getattribute__(self, key):
        common = object.__getattribute__(self, 'common_keys')
        mapped = object.__getattribute__(self, 'mapped_keys')
        if key in mapped:
            lk, maker = mapped[key]
            if self._legacy:
                if lk is None:
                    result = None if maker is None else maker()
                else:
                    result = self._legacy.get(lk)
            else:
                value = None if maker is None else maker()
                if key not in ('commands', 'exports', 'modules', 'namespaces',
                               'classifiers'):
                    result = self._data.get(key, value)
                else:
                    # special cases for PEP 459
                    sentinel = object()
                    result = sentinel
                    d = self._data.get('extensions')
                    if d:
                        if key == 'commands':
                            result = d.get('python.commands', value)
                        elif key == 'classifiers':
                            d = d.get('python.details')
                            if d:
                                result = d.get(key, value)
                        else:
                            d = d.get('python.exports')
                            if not d:
                                d = self._data.get('python.exports')
                            if d:
                                result = d.get(key, value)
                    if result is sentinel:
                        result = value
        elif key not in common:
            result = object.__getattribute__(self, key)
        elif self._legacy:
            result = self._legacy.get(key)
        else:
            result = self._data.get(key)
        return result

    def _validate_value(self, key, value, scheme=None):
        if key in self.SYNTAX_VALIDATORS:
            pattern, exclusions = self.SYNTAX_VALIDATORS[key]
            if (scheme or self.scheme) not in exclusions:
                m = pattern.match(value)
                if not m:
                    raise MetadataInvalidError("'%s' is an invalid value for "
                                               "the '%s' property" % (value,
                                                                    key))

    def __setattr__(self, key, value):
        self._validate_value(key, value)
        common = object.__getattribute__(self, 'common_keys')
        mapped = object.__getattribute__(self, 'mapped_keys')
        if key in mapped:
            lk, _ = mapped[key]
            if self._legacy:
                if lk is None:
                    raise NotImplementedError
                self._legacy[lk] = value
            elif key not in ('commands', 'exports', 'modules', 'namespaces',
                             'classifiers'):
                self._data[key] = value
            else:
                # special cases for PEP 459
                d = self._data.setdefault('extensions', {})
                if key == 'commands':
                    d['python.commands'] = value
                elif key == 'classifiers':
                    d = d.setdefault('python.details', {})
                    d[key] = value
                else:
                    d = d.setdefault('python.exports', {})
                    d[key] = value
        elif key not in common:
            object.__setattr__(self, key, value)
        else:
            if key == 'keywords':
                if isinstance(value, string_types):
                    value = value.strip()
                    if value:
                        value = value.split()
                    else:
                        value = []
            if self._legacy:
                self._legacy[key] = value
            else:
                self._data[key] = value

    @property
    def name_and_version(self):
        return _get_name_and_version(self.name, self.version, True)

    @property
    def provides(self):
        if self._legacy:
            result = self._legacy['Provides-Dist']
        else:
            result = self._data.setdefault('provides', [])
        s = '%s (%s)' % (self.name, self.version)
        if s not in result:
            result.append(s)
        return result

    @provides.setter
    def provides(self, value):
        if self._legacy:
            self._legacy['Provides-Dist'] = value
        else:
            self._data['provides'] = value

    def get_requirements(self, reqts, extras=None, env=None):
        """
        Base method to get dependencies, given a set of extras
        to satisfy and an optional environment context.
        :param reqts: A list of sometimes-wanted dependencies,
                      perhaps dependent on extras and environment.
        :param extras: A list of optional components being requested.
        :param env: An optional environment for marker evaluation.
        """
        if self._legacy:
            result = reqts
        else:
            result = []
            extras = get_extras(extras or [], self.extras)
            for d in reqts:
                if 'extra' not in d and 'environment' not in d:
                    # unconditional
                    include = True
                else:
                    if 'extra' not in d:
                        # Not extra-dependent - only environment-dependent
                        include = True
                    else:
                        include = d.get('extra') in extras
                    if include:
                        # Not excluded because of extras, check environment
                        marker = d.get('environment')
                        if marker:
                            include = interpret(marker, env)
                if include:
                    result.extend(d['requires'])
            for key in ('build', 'dev', 'test'):
                e = ':%s:' % key
                if e in extras:
                    extras.remove(e)
                    # A recursive call, but it should terminate since 'test'
                    # has been removed from the extras
                    reqts = self._data.get('%s_requires' % key, [])
                    result.extend(self.get_requirements(reqts, extras=extras,
                                                        env=env))
        return result

    @property
    def dictionary(self):
        if self._legacy:
            return self._from_legacy()
        return self._data

    @property
    def dependencies(self):
        if self._legacy:
            raise NotImplementedError
        else:
            return extract_by_key(self._data, self.DEPENDENCY_KEYS)

    @dependencies.setter
    def dependencies(self, value):
        if self._legacy:
            raise NotImplementedError
        else:
            self._data.update(value)

    def _validate_mapping(self, mapping, scheme):
        if mapping.get('metadata_version') != self.METADATA_VERSION:
            raise MetadataUnrecognizedVersionError()
        missing = []
        for key, exclusions in self.MANDATORY_KEYS.items():
            if key not in mapping:
                if scheme not in exclusions:
                    missing.append(key)
        if missing:
            msg = 'Missing metadata items: %s' % ', '.join(missing)
            raise MetadataMissingError(msg)
        for k, v in mapping.items():
            self._validate_value(k, v, scheme)

    def validate(self):
        if self._legacy:
            missing, warnings = self._legacy.check(True)
            if missing or warnings:
                logger.warning('Metadata: missing: %s, warnings: %s',
                               missing, warnings)
        else:
            self._validate_mapping(self._data, self.scheme)

    def todict(self):
        if self._legacy:
            return self._legacy.todict(True)
        else:
            result = extract_by_key(self._data, self.INDEX_KEYS)
            return result

    def _from_legacy(self):
        assert self._legacy and not self._data
        result = {
            'metadata_version': self.METADATA_VERSION,
            'generator': self.GENERATOR,
        }
        lmd = self._legacy.todict(True)     # skip missing ones
        for k in ('name', 'version', 'license', 'summary', 'description',
                  'classifier'):
            if k in lmd:
                if k == 'classifier':
                    nk = 'classifiers'
                else:
                    nk = k
                result[nk] = lmd[k]
        kw = lmd.get('Keywords', [])
        if kw == ['']:
            kw = []
        result['keywords'] = kw
        keys = (('requires_dist', 'run_requires'),
                ('setup_requires_dist', 'build_requires'))
        for ok, nk in keys:
            if ok in lmd and lmd[ok]:
                result[nk] = [{'requires': lmd[ok]}]
        result['provides'] = self.provides
        author = {}
        maintainer = {}
        return result

    LEGACY_MAPPING = {
        'name': 'Name',
        'version': 'Version',
        ('extensions', 'python.details', 'license'): 'License',
        'summary': 'Summary',
        'description': 'Description',
        ('extensions', 'python.project', 'project_urls', 'Home'): 'Home-page',
        ('extensions', 'python.project', 'contacts', 0, 'name'): 'Author',
        ('extensions', 'python.project', 'contacts', 0, 'email'): 'Author-email',
        'source_url': 'Download-URL',
        ('extensions', 'python.details', 'classifiers'): 'Classifier',
    }

    def _to_legacy(self):
        def process_entries(entries):
            reqts = set()
            for e in entries:
                extra = e.get('extra')
                env = e.get('environment')
                rlist = e['requires']
                for r in rlist:
                    if not env and not extra:
                        reqts.add(r)
                    else:
                        marker = ''
                        if extra:
                            marker = 'extra == "%s"' % extra
                        if env:
                            if marker:
                                marker = '(%s) and %s' % (env, marker)
                            else:
                                marker = env
                        reqts.add(';'.join((r, marker)))
            return reqts

        assert self._data and not self._legacy
        result = LegacyMetadata()
        nmd = self._data
        # import pdb; pdb.set_trace()
        for nk, ok in self.LEGACY_MAPPING.items():
            if not isinstance(nk, tuple):
                if nk in nmd:
                    result[ok] = nmd[nk]
            else:
                d = nmd
                found = True
                for k in nk:
                    try:
                        d = d[k]
                    except (KeyError, IndexError):
                        found = False
                        break
                if found:
                    result[ok] = d
        r1 = process_entries(self.run_requires + self.meta_requires)
        r2 = process_entries(self.build_requires + self.dev_requires)
        if self.extras:
            result['Provides-Extra'] = sorted(self.extras)
        result['Requires-Dist'] = sorted(r1)
        result['Setup-Requires-Dist'] = sorted(r2)
        # TODO: any other fields wanted
        return result

    def write(self, path=None, fileobj=None, legacy=False, skip_unknown=True):
        if [path, fileobj].count(None) != 1:
            raise ValueError('Exactly one of path and fileobj is needed')
        self.validate()
        if legacy:
            if self._legacy:
                legacy_md = self._legacy
            else:
                legacy_md = self._to_legacy()
            if path:
                legacy_md.write(path, skip_unknown=skip_unknown)
            else:
                legacy_md.write_file(fileobj, skip_unknown=skip_unknown)
        else:
            if self._legacy:
                d = self._from_legacy()
            else:
                d = self._data
            if fileobj:
                json.dump(d, fileobj, ensure_ascii=True, indent=2,
                          sort_keys=True)
            else:
                with codecs.open(path, 'w', 'utf-8') as f:
                    json.dump(d, f, ensure_ascii=True, indent=2,
                              sort_keys=True)

    def add_requirements(self, requirements):
        if self._legacy:
            self._legacy.add_requirements(requirements)
        else:
            run_requires = self._data.setdefault('run_requires', [])
            always = None
            for entry in run_requires:
                if 'environment' not in entry and 'extra' not in entry:
                    always = entry
                    break
            if always is None:
                always = { 'requires': requirements }
                run_requires.insert(0, always)
            else:
                rset = set(always['requires']) | set(requirements)
                always['requires'] = sorted(rset)

    def __repr__(self):
        name = self.name or '(no name)'
        version = self.version or 'no version'
        return '<%s %s %s (%s)>' % (self.__class__.__name__,
                                    self.metadata_version, name, version)
site-packages/pip/_vendor/distlib/manifest.py000064400000034733150564745010015343 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2013 Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""
Class representing the list of files in a distribution.

Equivalent to distutils.filelist, but fixes some problems.
"""
import fnmatch
import logging
import os
import re
import sys

from . import DistlibException
from .compat import fsdecode
from .util import convert_path


__all__ = ['Manifest']

logger = logging.getLogger(__name__)

# a \ followed by some spaces + EOL
_COLLAPSE_PATTERN = re.compile('\\\\w*\n', re.M)
_COMMENTED_LINE = re.compile('#.*?(?=\n)|\n(?=$)', re.M | re.S)

#
# Due to the different results returned by fnmatch.translate, we need
# to do slightly different processing for Python 2.7 and 3.2 ... this needed
# to be brought in for Python 3.6 onwards.
#
_PYTHON_VERSION = sys.version_info[:2]

class Manifest(object):
    """A list of files built by on exploring the filesystem and filtered by
    applying various patterns to what we find there.
    """

    def __init__(self, base=None):
        """
        Initialise an instance.

        :param base: The base directory to explore under.
        """
        self.base = os.path.abspath(os.path.normpath(base or os.getcwd()))
        self.prefix = self.base + os.sep
        self.allfiles = None
        self.files = set()

    #
    # Public API
    #

    def findall(self):
        """Find all files under the base and set ``allfiles`` to the absolute
        pathnames of files found.
        """
        from stat import S_ISREG, S_ISDIR, S_ISLNK

        self.allfiles = allfiles = []
        root = self.base
        stack = [root]
        pop = stack.pop
        push = stack.append

        while stack:
            root = pop()
            names = os.listdir(root)

            for name in names:
                fullname = os.path.join(root, name)

                # Avoid excess stat calls -- just one will do, thank you!
                stat = os.stat(fullname)
                mode = stat.st_mode
                if S_ISREG(mode):
                    allfiles.append(fsdecode(fullname))
                elif S_ISDIR(mode) and not S_ISLNK(mode):
                    push(fullname)

    def add(self, item):
        """
        Add a file to the manifest.

        :param item: The pathname to add. This can be relative to the base.
        """
        if not item.startswith(self.prefix):
            item = os.path.join(self.base, item)
        self.files.add(os.path.normpath(item))

    def add_many(self, items):
        """
        Add a list of files to the manifest.

        :param items: The pathnames to add. These can be relative to the base.
        """
        for item in items:
            self.add(item)

    def sorted(self, wantdirs=False):
        """
        Return sorted files in directory order
        """

        def add_dir(dirs, d):
            dirs.add(d)
            logger.debug('add_dir added %s', d)
            if d != self.base:
                parent, _ = os.path.split(d)
                assert parent not in ('', '/')
                add_dir(dirs, parent)

        result = set(self.files)    # make a copy!
        if wantdirs:
            dirs = set()
            for f in result:
                add_dir(dirs, os.path.dirname(f))
            result |= dirs
        return [os.path.join(*path_tuple) for path_tuple in
                sorted(os.path.split(path) for path in result)]

    def clear(self):
        """Clear all collected files."""
        self.files = set()
        self.allfiles = []

    def process_directive(self, directive):
        """
        Process a directive which either adds some files from ``allfiles`` to
        ``files``, or removes some files from ``files``.

        :param directive: The directive to process. This should be in a format
                     compatible with distutils ``MANIFEST.in`` files:

                     http://docs.python.org/distutils/sourcedist.html#commands
        """
        # Parse the line: split it up, make sure the right number of words
        # is there, and return the relevant words.  'action' is always
        # defined: it's the first word of the line.  Which of the other
        # three are defined depends on the action; it'll be either
        # patterns, (dir and patterns), or (dirpattern).
        action, patterns, thedir, dirpattern = self._parse_directive(directive)

        # OK, now we know that the action is valid and we have the
        # right number of words on the line for that action -- so we
        # can proceed with minimal error-checking.
        if action == 'include':
            for pattern in patterns:
                if not self._include_pattern(pattern, anchor=True):
                    logger.warning('no files found matching %r', pattern)

        elif action == 'exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, anchor=True)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'found matching %r', pattern)

        elif action == 'global-include':
            for pattern in patterns:
                if not self._include_pattern(pattern, anchor=False):
                    logger.warning('no files found matching %r '
                                   'anywhere in distribution', pattern)

        elif action == 'global-exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, anchor=False)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'matching %r found anywhere in '
                #                   'distribution', pattern)

        elif action == 'recursive-include':
            for pattern in patterns:
                if not self._include_pattern(pattern, prefix=thedir):
                    logger.warning('no files found matching %r '
                                   'under directory %r', pattern, thedir)

        elif action == 'recursive-exclude':
            for pattern in patterns:
                found = self._exclude_pattern(pattern, prefix=thedir)
                #if not found:
                #    logger.warning('no previously-included files '
                #                   'matching %r found under directory %r',
                #                   pattern, thedir)

        elif action == 'graft':
            if not self._include_pattern(None, prefix=dirpattern):
                logger.warning('no directories found matching %r',
                               dirpattern)

        elif action == 'prune':
            if not self._exclude_pattern(None, prefix=dirpattern):
                logger.warning('no previously-included directories found '
                               'matching %r', dirpattern)
        else:   # pragma: no cover
            # This should never happen, as it should be caught in
            # _parse_template_line
            raise DistlibException(
                'invalid action %r' % action)

    #
    # Private API
    #

    def _parse_directive(self, directive):
        """
        Validate a directive.
        :param directive: The directive to validate.
        :return: A tuple of action, patterns, thedir, dir_patterns
        """
        words = directive.split()
        if len(words) == 1 and words[0] not in ('include', 'exclude',
                                                'global-include',
                                                'global-exclude',
                                                'recursive-include',
                                                'recursive-exclude',
                                                'graft', 'prune'):
            # no action given, let's use the default 'include'
            words.insert(0, 'include')

        action = words[0]
        patterns = thedir = dir_pattern = None

        if action in ('include', 'exclude',
                      'global-include', 'global-exclude'):
            if len(words) < 2:
                raise DistlibException(
                    '%r expects <pattern1> <pattern2> ...' % action)

            patterns = [convert_path(word) for word in words[1:]]

        elif action in ('recursive-include', 'recursive-exclude'):
            if len(words) < 3:
                raise DistlibException(
                    '%r expects <dir> <pattern1> <pattern2> ...' % action)

            thedir = convert_path(words[1])
            patterns = [convert_path(word) for word in words[2:]]

        elif action in ('graft', 'prune'):
            if len(words) != 2:
                raise DistlibException(
                    '%r expects a single <dir_pattern>' % action)

            dir_pattern = convert_path(words[1])

        else:
            raise DistlibException('unknown action %r' % action)

        return action, patterns, thedir, dir_pattern

    def _include_pattern(self, pattern, anchor=True, prefix=None,
                         is_regex=False):
        """Select strings (presumably filenames) from 'self.files' that
        match 'pattern', a Unix-style wildcard (glob) pattern.

        Patterns are not quite the same as implemented by the 'fnmatch'
        module: '*' and '?'  match non-special characters, where "special"
        is platform-dependent: slash on Unix; colon, slash, and backslash on
        DOS/Windows; and colon on Mac OS.

        If 'anchor' is true (the default), then the pattern match is more
        stringent: "*.py" will match "foo.py" but not "foo/bar.py".  If
        'anchor' is false, both of these will match.

        If 'prefix' is supplied, then only filenames starting with 'prefix'
        (itself a pattern) and ending with 'pattern', with anything in between
        them, will match.  'anchor' is ignored in this case.

        If 'is_regex' is true, 'anchor' and 'prefix' are ignored, and
        'pattern' is assumed to be either a string containing a regex or a
        regex object -- no translation is done, the regex is just compiled
        and used as-is.

        Selected strings will be added to self.files.

        Return True if files are found.
        """
        # XXX docstring lying about what the special chars are?
        found = False
        pattern_re = self._translate_pattern(pattern, anchor, prefix, is_regex)

        # delayed loading of allfiles list
        if self.allfiles is None:
            self.findall()

        for name in self.allfiles:
            if pattern_re.search(name):
                self.files.add(name)
                found = True
        return found

    def _exclude_pattern(self, pattern, anchor=True, prefix=None,
                         is_regex=False):
        """Remove strings (presumably filenames) from 'files' that match
        'pattern'.

        Other parameters are the same as for 'include_pattern()', above.
        The list 'self.files' is modified in place. Return True if files are
        found.

        This API is public to allow e.g. exclusion of SCM subdirs, e.g. when
        packaging source distributions
        """
        found = False
        pattern_re = self._translate_pattern(pattern, anchor, prefix, is_regex)
        for f in list(self.files):
            if pattern_re.search(f):
                self.files.remove(f)
                found = True
        return found

    def _translate_pattern(self, pattern, anchor=True, prefix=None,
                           is_regex=False):
        """Translate a shell-like wildcard pattern to a compiled regular
        expression.

        Return the compiled regex.  If 'is_regex' true,
        then 'pattern' is directly compiled to a regex (if it's a string)
        or just returned as-is (assumes it's a regex object).
        """
        if is_regex:
            if isinstance(pattern, str):
                return re.compile(pattern)
            else:
                return pattern

        if _PYTHON_VERSION > (3, 2):
            # ditch start and end characters
            start, _, end = self._glob_to_re('_').partition('_')

        if pattern:
            pattern_re = self._glob_to_re(pattern)
            if _PYTHON_VERSION > (3, 2):
                assert pattern_re.startswith(start) and pattern_re.endswith(end)
        else:
            pattern_re = ''

        base = re.escape(os.path.join(self.base, ''))
        if prefix is not None:
            # ditch end of pattern character
            if _PYTHON_VERSION <= (3, 2):
                empty_pattern = self._glob_to_re('')
                prefix_re = self._glob_to_re(prefix)[:-len(empty_pattern)]
            else:
                prefix_re = self._glob_to_re(prefix)
                assert prefix_re.startswith(start) and prefix_re.endswith(end)
                prefix_re = prefix_re[len(start): len(prefix_re) - len(end)]
            sep = os.sep
            if os.sep == '\\':
                sep = r'\\'
            if _PYTHON_VERSION <= (3, 2):
                pattern_re = '^' + base + sep.join((prefix_re,
                                                    '.*' + pattern_re))
            else:
                pattern_re = pattern_re[len(start): len(pattern_re) - len(end)]
                pattern_re = r'%s%s%s%s.*%s%s' % (start, base, prefix_re, sep,
                                                  pattern_re, end)
        else:  # no prefix -- respect anchor flag
            if anchor:
                if _PYTHON_VERSION <= (3, 2):
                    pattern_re = '^' + base + pattern_re
                else:
                    pattern_re = r'%s%s%s' % (start, base, pattern_re[len(start):])

        return re.compile(pattern_re)

    def _glob_to_re(self, pattern):
        """Translate a shell-like glob pattern to a regular expression.

        Return a string containing the regex.  Differs from
        'fnmatch.translate()' in that '*' does not match "special characters"
        (which are platform-specific).
        """
        pattern_re = fnmatch.translate(pattern)

        # '?' and '*' in the glob pattern become '.' and '.*' in the RE, which
        # IMHO is wrong -- '?' and '*' aren't supposed to match slash in Unix,
        # and by extension they shouldn't match such "special characters" under
        # any OS.  So change all non-escaped dots in the RE to match any
        # character except the special characters (currently: just os.sep).
        sep = os.sep
        if os.sep == '\\':
            # we're using a regex to manipulate a regex, so we need
            # to escape the backslash twice
            sep = r'\\\\'
        escaped = r'\1[^%s]' % sep
        pattern_re = re.sub(r'((?<!\\)(\\\\)*)\.', escaped, pattern_re)
        return pattern_re
site-packages/pip/_vendor/distlib/markers.py000064400000011575150564745010015200 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""
Parser for the environment markers micro-language defined in PEP 508.
"""

# Note: In PEP 345, the micro-language was Python compatible, so the ast
# module could be used to parse it. However, PEP 508 introduced operators such
# as ~= and === which aren't in Python, necessitating a different approach.

import os
import re
import sys
import platform

from .compat import string_types
from .util import in_venv, parse_marker
from .version import NormalizedVersion as NV

__all__ = ['interpret']

_VERSION_PATTERN = re.compile(r'((\d+(\.\d+)*\w*)|\'(\d+(\.\d+)*\w*)\'|\"(\d+(\.\d+)*\w*)\")')

def _is_literal(o):
    if not isinstance(o, string_types) or not o:
        return False
    return o[0] in '\'"'

def _get_versions(s):
    result = []
    for m in _VERSION_PATTERN.finditer(s):
        result.append(NV(m.groups()[0]))
    return set(result)

class Evaluator(object):
    """
    This class is used to evaluate marker expessions.
    """

    operations = {
        '==': lambda x, y: x == y,
        '===': lambda x, y: x == y,
        '~=': lambda x, y: x == y or x > y,
        '!=': lambda x, y: x != y,
        '<':  lambda x, y: x < y,
        '<=':  lambda x, y: x == y or x < y,
        '>':  lambda x, y: x > y,
        '>=':  lambda x, y: x == y or x > y,
        'and': lambda x, y: x and y,
        'or': lambda x, y: x or y,
        'in': lambda x, y: x in y,
        'not in': lambda x, y: x not in y,
    }

    def evaluate(self, expr, context):
        """
        Evaluate a marker expression returned by the :func:`parse_requirement`
        function in the specified context.
        """
        if isinstance(expr, string_types):
            if expr[0] in '\'"':
                result = expr[1:-1]
            else:
                if expr not in context:
                    raise SyntaxError('unknown variable: %s' % expr)
                result = context[expr]
        else:
            assert isinstance(expr, dict)
            op = expr['op']
            if op not in self.operations:
                raise NotImplementedError('op not implemented: %s' % op)
            elhs = expr['lhs']
            erhs = expr['rhs']
            if _is_literal(expr['lhs']) and _is_literal(expr['rhs']):
                raise SyntaxError('invalid comparison: %s %s %s' % (elhs, op, erhs))

            lhs = self.evaluate(elhs, context)
            rhs = self.evaluate(erhs, context)
            if ((elhs == 'python_version' or erhs == 'python_version') and
                op in ('<', '<=', '>', '>=', '===', '==', '!=', '~=')):
                lhs = NV(lhs)
                rhs = NV(rhs)
            elif elhs == 'python_version' and op in ('in', 'not in'):
                lhs = NV(lhs)
                rhs = _get_versions(rhs)
            result = self.operations[op](lhs, rhs)
        return result

def default_context():
    def format_full_version(info):
        version = '%s.%s.%s' % (info.major, info.minor, info.micro)
        kind = info.releaselevel
        if kind != 'final':
            version += kind[0] + str(info.serial)
        return version

    if hasattr(sys, 'implementation'):
        implementation_version = format_full_version(sys.implementation.version)
        implementation_name = sys.implementation.name
    else:
        implementation_version = '0'
        implementation_name = ''

    result = {
        'implementation_name': implementation_name,
        'implementation_version': implementation_version,
        'os_name': os.name,
        'platform_machine': platform.machine(),
        'platform_python_implementation': platform.python_implementation(),
        'platform_release': platform.release(),
        'platform_system': platform.system(),
        'platform_version': platform.version(),
        'platform_in_venv': str(in_venv()),
        'python_full_version': platform.python_version(),
        'python_version': platform.python_version()[:3],
        'sys_platform': sys.platform,
    }
    return result

DEFAULT_CONTEXT = default_context()
del default_context

evaluator = Evaluator()

def interpret(marker, execution_context=None):
    """
    Interpret a marker and return a result depending on environment.

    :param marker: The marker to interpret.
    :type marker: str
    :param execution_context: The context used for name lookup.
    :type execution_context: mapping
    """
    try:
        expr, rest = parse_marker(marker)
    except Exception as e:
        raise SyntaxError('Unable to interpret marker syntax: %s: %s' % (marker, e))
    if rest and rest[0] != '#':
        raise SyntaxError('unexpected trailing data in marker: %s: %s' % (marker, rest))
    context = dict(DEFAULT_CONTEXT)
    if execution_context:
        context.update(execution_context)
    return evaluator.evaluate(expr, context)
site-packages/pip/_vendor/distlib/index.py000064400000050403150564745010014634 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import hashlib
import logging
import os
import shutil
import subprocess
import tempfile
try:
    from threading import Thread
except ImportError:
    from dummy_threading import Thread

from . import DistlibException
from .compat import (HTTPBasicAuthHandler, Request, HTTPPasswordMgr,
                     urlparse, build_opener, string_types)
from .util import zip_dir, ServerProxy

logger = logging.getLogger(__name__)

DEFAULT_INDEX = 'https://pypi.org/pypi'
DEFAULT_REALM = 'pypi'

class PackageIndex(object):
    """
    This class represents a package index compatible with PyPI, the Python
    Package Index.
    """

    boundary = b'----------ThIs_Is_tHe_distlib_index_bouNdaRY_$'

    def __init__(self, url=None):
        """
        Initialise an instance.

        :param url: The URL of the index. If not specified, the URL for PyPI is
                    used.
        """
        self.url = url or DEFAULT_INDEX
        self.read_configuration()
        scheme, netloc, path, params, query, frag = urlparse(self.url)
        if params or query or frag or scheme not in ('http', 'https'):
            raise DistlibException('invalid repository: %s' % self.url)
        self.password_handler = None
        self.ssl_verifier = None
        self.gpg = None
        self.gpg_home = None
        with open(os.devnull, 'w') as sink:
            # Use gpg by default rather than gpg2, as gpg2 insists on
            # prompting for passwords
            for s in ('gpg', 'gpg2'):
                try:
                    rc = subprocess.check_call([s, '--version'], stdout=sink,
                                               stderr=sink)
                    if rc == 0:
                        self.gpg = s
                        break
                except OSError:
                    pass

    def _get_pypirc_command(self):
        """
        Get the distutils command for interacting with PyPI configurations.
        :return: the command.
        """
        from .util import _get_pypirc_command as cmd
        return cmd()

    def read_configuration(self):
        """
        Read the PyPI access configuration as supported by distutils. This populates
        ``username``, ``password``, ``realm`` and ``url`` attributes from the
        configuration.
        """
        from .util import _load_pypirc
        cfg = _load_pypirc(self)
        self.username = cfg.get('username')
        self.password = cfg.get('password')
        self.realm = cfg.get('realm', 'pypi')
        self.url = cfg.get('repository', self.url)

    def save_configuration(self):
        """
        Save the PyPI access configuration. You must have set ``username`` and
        ``password`` attributes before calling this method.
        """
        self.check_credentials()
        from .util import _store_pypirc
        _store_pypirc(self)

    def check_credentials(self):
        """
        Check that ``username`` and ``password`` have been set, and raise an
        exception if not.
        """
        if self.username is None or self.password is None:
            raise DistlibException('username and password must be set')
        pm = HTTPPasswordMgr()
        _, netloc, _, _, _, _ = urlparse(self.url)
        pm.add_password(self.realm, netloc, self.username, self.password)
        self.password_handler = HTTPBasicAuthHandler(pm)

    def register(self, metadata):
        """
        Register a distribution on PyPI, using the provided metadata.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the distribution to be
                         registered.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        metadata.validate()
        d = metadata.todict()
        d[':action'] = 'verify'
        request = self.encode_request(d.items(), [])
        response = self.send_request(request)
        d[':action'] = 'submit'
        request = self.encode_request(d.items(), [])
        return self.send_request(request)

    def _reader(self, name, stream, outbuf):
        """
        Thread runner for reading lines of from a subprocess into a buffer.

        :param name: The logical name of the stream (used for logging only).
        :param stream: The stream to read from. This will typically a pipe
                       connected to the output stream of a subprocess.
        :param outbuf: The list to append the read lines to.
        """
        while True:
            s = stream.readline()
            if not s:
                break
            s = s.decode('utf-8').rstrip()
            outbuf.append(s)
            logger.debug('%s: %s' % (name, s))
        stream.close()

    def get_sign_command(self, filename, signer, sign_password,
                         keystore=None):
        """
        Return a suitable command for signing a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The signing command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        """
        cmd = [self.gpg, '--status-fd', '2', '--no-tty']
        if keystore is None:
            keystore = self.gpg_home
        if keystore:
            cmd.extend(['--homedir', keystore])
        if sign_password is not None:
            cmd.extend(['--batch', '--passphrase-fd', '0'])
        td = tempfile.mkdtemp()
        sf = os.path.join(td, os.path.basename(filename) + '.asc')
        cmd.extend(['--detach-sign', '--armor', '--local-user',
                    signer, '--output', sf, filename])
        logger.debug('invoking: %s', ' '.join(cmd))
        return cmd, sf

    def run_command(self, cmd, input_data=None):
        """
        Run a command in a child process , passing it any input data specified.

        :param cmd: The command to run.
        :param input_data: If specified, this must be a byte string containing
                           data to be sent to the child process.
        :return: A tuple consisting of the subprocess' exit code, a list of
                 lines read from the subprocess' ``stdout``, and a list of
                 lines read from the subprocess' ``stderr``.
        """
        kwargs = {
            'stdout': subprocess.PIPE,
            'stderr': subprocess.PIPE,
        }
        if input_data is not None:
            kwargs['stdin'] = subprocess.PIPE
        stdout = []
        stderr = []
        p = subprocess.Popen(cmd, **kwargs)
        # We don't use communicate() here because we may need to
        # get clever with interacting with the command
        t1 = Thread(target=self._reader, args=('stdout', p.stdout, stdout))
        t1.start()
        t2 = Thread(target=self._reader, args=('stderr', p.stderr, stderr))
        t2.start()
        if input_data is not None:
            p.stdin.write(input_data)
            p.stdin.close()

        p.wait()
        t1.join()
        t2.join()
        return p.returncode, stdout, stderr

    def sign_file(self, filename, signer, sign_password, keystore=None):
        """
        Sign a file.

        :param filename: The pathname to the file to be signed.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The absolute pathname of the file where the signature is
                 stored.
        """
        cmd, sig_file = self.get_sign_command(filename, signer, sign_password,
                                              keystore)
        rc, stdout, stderr = self.run_command(cmd,
                                              sign_password.encode('utf-8'))
        if rc != 0:
            raise DistlibException('sign command failed with error '
                                   'code %s' % rc)
        return sig_file

    def upload_file(self, metadata, filename, signer=None, sign_password=None,
                    filetype='sdist', pyversion='source', keystore=None):
        """
        Upload a release file to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the file to be uploaded.
        :param filename: The pathname of the file to be uploaded.
        :param signer: The identifier of the signer of the file.
        :param sign_password: The passphrase for the signer's
                              private key used for signing.
        :param filetype: The type of the file being uploaded. This is the
                        distutils command which produced that file, e.g.
                        ``sdist`` or ``bdist_wheel``.
        :param pyversion: The version of Python which the release relates
                          to. For code compatible with any Python, this would
                          be ``source``, otherwise it would be e.g. ``3.2``.
        :param keystore: The path to a directory which contains the keys
                         used in signing. If not specified, the instance's
                         ``gpg_home`` attribute is used instead.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        if not os.path.exists(filename):
            raise DistlibException('not found: %s' % filename)
        metadata.validate()
        d = metadata.todict()
        sig_file = None
        if signer:
            if not self.gpg:
                logger.warning('no signing program available - not signed')
            else:
                sig_file = self.sign_file(filename, signer, sign_password,
                                          keystore)
        with open(filename, 'rb') as f:
            file_data = f.read()
        md5_digest = hashlib.md5(file_data).hexdigest()
        sha256_digest = hashlib.sha256(file_data).hexdigest()
        d.update({
            ':action': 'file_upload',
            'protocol_version': '1',
            'filetype': filetype,
            'pyversion': pyversion,
            'md5_digest': md5_digest,
            'sha256_digest': sha256_digest,
        })
        files = [('content', os.path.basename(filename), file_data)]
        if sig_file:
            with open(sig_file, 'rb') as f:
                sig_data = f.read()
            files.append(('gpg_signature', os.path.basename(sig_file),
                         sig_data))
            shutil.rmtree(os.path.dirname(sig_file))
        request = self.encode_request(d.items(), files)
        return self.send_request(request)

    def upload_documentation(self, metadata, doc_dir):
        """
        Upload documentation to the index.

        :param metadata: A :class:`Metadata` instance defining at least a name
                         and version number for the documentation to be
                         uploaded.
        :param doc_dir: The pathname of the directory which contains the
                        documentation. This should be the directory that
                        contains the ``index.html`` for the documentation.
        :return: The HTTP response received from PyPI upon submission of the
                request.
        """
        self.check_credentials()
        if not os.path.isdir(doc_dir):
            raise DistlibException('not a directory: %r' % doc_dir)
        fn = os.path.join(doc_dir, 'index.html')
        if not os.path.exists(fn):
            raise DistlibException('not found: %r' % fn)
        metadata.validate()
        name, version = metadata.name, metadata.version
        zip_data = zip_dir(doc_dir).getvalue()
        fields = [(':action', 'doc_upload'),
                  ('name', name), ('version', version)]
        files = [('content', name, zip_data)]
        request = self.encode_request(fields, files)
        return self.send_request(request)

    def get_verify_command(self, signature_filename, data_filename,
                           keystore=None):
        """
        Return a suitable command for verifying a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: The verifying command as a list suitable to be
                 passed to :class:`subprocess.Popen`.
        """
        cmd = [self.gpg, '--status-fd', '2', '--no-tty']
        if keystore is None:
            keystore = self.gpg_home
        if keystore:
            cmd.extend(['--homedir', keystore])
        cmd.extend(['--verify', signature_filename, data_filename])
        logger.debug('invoking: %s', ' '.join(cmd))
        return cmd

    def verify_signature(self, signature_filename, data_filename,
                         keystore=None):
        """
        Verify a signature for a file.

        :param signature_filename: The pathname to the file containing the
                                   signature.
        :param data_filename: The pathname to the file containing the
                              signed data.
        :param keystore: The path to a directory which contains the keys
                         used in verification. If not specified, the
                         instance's ``gpg_home`` attribute is used instead.
        :return: True if the signature was verified, else False.
        """
        if not self.gpg:
            raise DistlibException('verification unavailable because gpg '
                                   'unavailable')
        cmd = self.get_verify_command(signature_filename, data_filename,
                                      keystore)
        rc, stdout, stderr = self.run_command(cmd)
        if rc not in (0, 1):
            raise DistlibException('verify command failed with error '
                             'code %s' % rc)
        return rc == 0

    def download_file(self, url, destfile, digest=None, reporthook=None):
        """
        This is a convenience method for downloading a file from an URL.
        Normally, this will be a file from the index, though currently
        no check is made for this (i.e. a file can be downloaded from
        anywhere).

        The method is just like the :func:`urlretrieve` function in the
        standard library, except that it allows digest computation to be
        done during download and checking that the downloaded data
        matched any expected value.

        :param url: The URL of the file to be downloaded (assumed to be
                    available via an HTTP GET request).
        :param destfile: The pathname where the downloaded file is to be
                         saved.
        :param digest: If specified, this must be a (hasher, value)
                       tuple, where hasher is the algorithm used (e.g.
                       ``'md5'``) and ``value`` is the expected value.
        :param reporthook: The same as for :func:`urlretrieve` in the
                           standard library.
        """
        if digest is None:
            digester = None
            logger.debug('No digest specified')
        else:
            if isinstance(digest, (list, tuple)):
                hasher, digest = digest
            else:
                hasher = 'md5'
            digester = getattr(hashlib, hasher)()
            logger.debug('Digest specified: %s' % digest)
        # The following code is equivalent to urlretrieve.
        # We need to do it this way so that we can compute the
        # digest of the file as we go.
        with open(destfile, 'wb') as dfp:
            # addinfourl is not a context manager on 2.x
            # so we have to use try/finally
            sfp = self.send_request(Request(url))
            try:
                headers = sfp.info()
                blocksize = 8192
                size = -1
                read = 0
                blocknum = 0
                if "content-length" in headers:
                    size = int(headers["Content-Length"])
                if reporthook:
                    reporthook(blocknum, blocksize, size)
                while True:
                    block = sfp.read(blocksize)
                    if not block:
                        break
                    read += len(block)
                    dfp.write(block)
                    if digester:
                        digester.update(block)
                    blocknum += 1
                    if reporthook:
                        reporthook(blocknum, blocksize, size)
            finally:
                sfp.close()

        # check that we got the whole file, if we can
        if size >= 0 and read < size:
            raise DistlibException(
                'retrieval incomplete: got only %d out of %d bytes'
                % (read, size))
        # if we have a digest, it must match.
        if digester:
            actual = digester.hexdigest()
            if digest != actual:
                raise DistlibException('%s digest mismatch for %s: expected '
                                       '%s, got %s' % (hasher, destfile,
                                                       digest, actual))
            logger.debug('Digest verified: %s', digest)

    def send_request(self, req):
        """
        Send a standard library :class:`Request` to PyPI and return its
        response.

        :param req: The request to send.
        :return: The HTTP response from PyPI (a standard library HTTPResponse).
        """
        handlers = []
        if self.password_handler:
            handlers.append(self.password_handler)
        if self.ssl_verifier:
            handlers.append(self.ssl_verifier)
        opener = build_opener(*handlers)
        return opener.open(req)

    def encode_request(self, fields, files):
        """
        Encode fields and files for posting to an HTTP server.

        :param fields: The fields to send as a list of (fieldname, value)
                       tuples.
        :param files: The files to send as a list of (fieldname, filename,
                      file_bytes) tuple.
        """
        # Adapted from packaging, which in turn was adapted from
        # http://code.activestate.com/recipes/146306

        parts = []
        boundary = self.boundary
        for k, values in fields:
            if not isinstance(values, (list, tuple)):
                values = [values]

            for v in values:
                parts.extend((
                    b'--' + boundary,
                    ('Content-Disposition: form-data; name="%s"' %
                     k).encode('utf-8'),
                    b'',
                    v.encode('utf-8')))
        for key, filename, value in files:
            parts.extend((
                b'--' + boundary,
                ('Content-Disposition: form-data; name="%s"; filename="%s"' %
                 (key, filename)).encode('utf-8'),
                b'',
                value))

        parts.extend((b'--' + boundary + b'--', b''))

        body = b'\r\n'.join(parts)
        ct = b'multipart/form-data; boundary=' + boundary
        headers = {
            'Content-type': ct,
            'Content-length': str(len(body))
        }
        return Request(self.url, body, headers)

    def search(self, terms, operator=None):
        if isinstance(terms, string_types):
            terms = {'name': terms}
        rpc_proxy = ServerProxy(self.url, timeout=3.0)
        try:
            return rpc_proxy.search(terms, operator or 'and')
        finally:
            rpc_proxy('close')()
site-packages/pip/_vendor/distlib/wheel.py000064400000123677150564745010014647 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2020 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import unicode_literals

import base64
import codecs
import datetime
from email import message_from_file
import hashlib
import imp
import json
import logging
import os
import posixpath
import re
import shutil
import sys
import tempfile
import zipfile

from . import __version__, DistlibException
from .compat import sysconfig, ZipFile, fsdecode, text_type, filter
from .database import InstalledDistribution
from .metadata import (Metadata, METADATA_FILENAME, WHEEL_METADATA_FILENAME,
                       LEGACY_METADATA_FILENAME)
from .util import (FileOperator, convert_path, CSVReader, CSVWriter, Cache,
                   cached_property, get_cache_base, read_exports, tempdir,
                   get_platform)
from .version import NormalizedVersion, UnsupportedVersionError

logger = logging.getLogger(__name__)

cache = None    # created when needed

if hasattr(sys, 'pypy_version_info'):  # pragma: no cover
    IMP_PREFIX = 'pp'
elif sys.platform.startswith('java'):  # pragma: no cover
    IMP_PREFIX = 'jy'
elif sys.platform == 'cli':  # pragma: no cover
    IMP_PREFIX = 'ip'
else:
    IMP_PREFIX = 'cp'

VER_SUFFIX = sysconfig.get_config_var('py_version_nodot')
if not VER_SUFFIX:   # pragma: no cover
    VER_SUFFIX = '%s%s' % sys.version_info[:2]
PYVER = 'py' + VER_SUFFIX
IMPVER = IMP_PREFIX + VER_SUFFIX

ARCH = get_platform().replace('-', '_').replace('.', '_')

ABI = sysconfig.get_config_var('SOABI')
if ABI and ABI.startswith('cpython-'):
    ABI = ABI.replace('cpython-', 'cp').split('-')[0]
else:
    def _derive_abi():
        parts = ['cp', VER_SUFFIX]
        if sysconfig.get_config_var('Py_DEBUG'):
            parts.append('d')
        if sysconfig.get_config_var('WITH_PYMALLOC'):
            parts.append('m')
        if sysconfig.get_config_var('Py_UNICODE_SIZE') == 4:
            parts.append('u')
        return ''.join(parts)
    ABI = _derive_abi()
    del _derive_abi

FILENAME_RE = re.compile(r'''
(?P<nm>[^-]+)
-(?P<vn>\d+[^-]*)
(-(?P<bn>\d+[^-]*))?
-(?P<py>\w+\d+(\.\w+\d+)*)
-(?P<bi>\w+)
-(?P<ar>\w+(\.\w+)*)
\.whl$
''', re.IGNORECASE | re.VERBOSE)

NAME_VERSION_RE = re.compile(r'''
(?P<nm>[^-]+)
-(?P<vn>\d+[^-]*)
(-(?P<bn>\d+[^-]*))?$
''', re.IGNORECASE | re.VERBOSE)

SHEBANG_RE = re.compile(br'\s*#![^\r\n]*')
SHEBANG_DETAIL_RE = re.compile(br'^(\s*#!("[^"]+"|\S+))\s+(.*)$')
SHEBANG_PYTHON = b'#!python'
SHEBANG_PYTHONW = b'#!pythonw'

if os.sep == '/':
    to_posix = lambda o: o
else:
    to_posix = lambda o: o.replace(os.sep, '/')


class Mounter(object):
    def __init__(self):
        self.impure_wheels = {}
        self.libs = {}

    def add(self, pathname, extensions):
        self.impure_wheels[pathname] = extensions
        self.libs.update(extensions)

    def remove(self, pathname):
        extensions = self.impure_wheels.pop(pathname)
        for k, v in extensions:
            if k in self.libs:
                del self.libs[k]

    def find_module(self, fullname, path=None):
        if fullname in self.libs:
            result = self
        else:
            result = None
        return result

    def load_module(self, fullname):
        if fullname in sys.modules:
            result = sys.modules[fullname]
        else:
            if fullname not in self.libs:
                raise ImportError('unable to find extension for %s' % fullname)
            result = imp.load_dynamic(fullname, self.libs[fullname])
            result.__loader__ = self
            parts = fullname.rsplit('.', 1)
            if len(parts) > 1:
                result.__package__ = parts[0]
        return result

_hook = Mounter()


class Wheel(object):
    """
    Class to build and install from Wheel files (PEP 427).
    """

    wheel_version = (1, 1)
    hash_kind = 'sha256'

    def __init__(self, filename=None, sign=False, verify=False):
        """
        Initialise an instance using a (valid) filename.
        """
        self.sign = sign
        self.should_verify = verify
        self.buildver = ''
        self.pyver = [PYVER]
        self.abi = ['none']
        self.arch = ['any']
        self.dirname = os.getcwd()
        if filename is None:
            self.name = 'dummy'
            self.version = '0.1'
            self._filename = self.filename
        else:
            m = NAME_VERSION_RE.match(filename)
            if m:
                info = m.groupdict('')
                self.name = info['nm']
                # Reinstate the local version separator
                self.version = info['vn'].replace('_', '-')
                self.buildver = info['bn']
                self._filename = self.filename
            else:
                dirname, filename = os.path.split(filename)
                m = FILENAME_RE.match(filename)
                if not m:
                    raise DistlibException('Invalid name or '
                                           'filename: %r' % filename)
                if dirname:
                    self.dirname = os.path.abspath(dirname)
                self._filename = filename
                info = m.groupdict('')
                self.name = info['nm']
                self.version = info['vn']
                self.buildver = info['bn']
                self.pyver = info['py'].split('.')
                self.abi = info['bi'].split('.')
                self.arch = info['ar'].split('.')

    @property
    def filename(self):
        """
        Build and return a filename from the various components.
        """
        if self.buildver:
            buildver = '-' + self.buildver
        else:
            buildver = ''
        pyver = '.'.join(self.pyver)
        abi = '.'.join(self.abi)
        arch = '.'.join(self.arch)
        # replace - with _ as a local version separator
        version = self.version.replace('-', '_')
        return '%s-%s%s-%s-%s-%s.whl' % (self.name, version, buildver,
                                         pyver, abi, arch)

    @property
    def exists(self):
        path = os.path.join(self.dirname, self.filename)
        return os.path.isfile(path)

    @property
    def tags(self):
        for pyver in self.pyver:
            for abi in self.abi:
                for arch in self.arch:
                    yield pyver, abi, arch

    @cached_property
    def metadata(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        wrapper = codecs.getreader('utf-8')
        with ZipFile(pathname, 'r') as zf:
            wheel_metadata = self.get_wheel_metadata(zf)
            wv = wheel_metadata['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            # if file_version < (1, 1):
                # fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME,
                       # LEGACY_METADATA_FILENAME]
            # else:
                # fns = [WHEEL_METADATA_FILENAME, METADATA_FILENAME]
            fns = [WHEEL_METADATA_FILENAME, LEGACY_METADATA_FILENAME]
            result = None
            for fn in fns:
                try:
                    metadata_filename = posixpath.join(info_dir, fn)
                    with zf.open(metadata_filename) as bf:
                        wf = wrapper(bf)
                        result = Metadata(fileobj=wf)
                        if result:
                            break
                except KeyError:
                    pass
            if not result:
                raise ValueError('Invalid wheel, because metadata is '
                                 'missing: looked in %s' % ', '.join(fns))
        return result

    def get_wheel_metadata(self, zf):
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        metadata_filename = posixpath.join(info_dir, 'WHEEL')
        with zf.open(metadata_filename) as bf:
            wf = codecs.getreader('utf-8')(bf)
            message = message_from_file(wf)
        return dict(message)

    @cached_property
    def info(self):
        pathname = os.path.join(self.dirname, self.filename)
        with ZipFile(pathname, 'r') as zf:
            result = self.get_wheel_metadata(zf)
        return result

    def process_shebang(self, data):
        m = SHEBANG_RE.match(data)
        if m:
            end = m.end()
            shebang, data_after_shebang = data[:end], data[end:]
            # Preserve any arguments after the interpreter
            if b'pythonw' in shebang.lower():
                shebang_python = SHEBANG_PYTHONW
            else:
                shebang_python = SHEBANG_PYTHON
            m = SHEBANG_DETAIL_RE.match(shebang)
            if m:
                args = b' ' + m.groups()[-1]
            else:
                args = b''
            shebang = shebang_python + args
            data = shebang + data_after_shebang
        else:
            cr = data.find(b'\r')
            lf = data.find(b'\n')
            if cr < 0 or cr > lf:
                term = b'\n'
            else:
                if data[cr:cr + 2] == b'\r\n':
                    term = b'\r\n'
                else:
                    term = b'\r'
            data = SHEBANG_PYTHON + term + data
        return data

    def get_hash(self, data, hash_kind=None):
        if hash_kind is None:
            hash_kind = self.hash_kind
        try:
            hasher = getattr(hashlib, hash_kind)
        except AttributeError:
            raise DistlibException('Unsupported hash algorithm: %r' % hash_kind)
        result = hasher(data).digest()
        result = base64.urlsafe_b64encode(result).rstrip(b'=').decode('ascii')
        return hash_kind, result

    def write_record(self, records, record_path, base):
        records = list(records) # make a copy, as mutated
        p = to_posix(os.path.relpath(record_path, base))
        records.append((p, '', ''))
        with CSVWriter(record_path) as writer:
            for row in records:
                writer.writerow(row)

    def write_records(self, info, libdir, archive_paths):
        records = []
        distinfo, info_dir = info
        hasher = getattr(hashlib, self.hash_kind)
        for ap, p in archive_paths:
            with open(p, 'rb') as f:
                data = f.read()
            digest = '%s=%s' % self.get_hash(data)
            size = os.path.getsize(p)
            records.append((ap, digest, size))

        p = os.path.join(distinfo, 'RECORD')
        self.write_record(records, p, libdir)
        ap = to_posix(os.path.join(info_dir, 'RECORD'))
        archive_paths.append((ap, p))

    def build_zip(self, pathname, archive_paths):
        with ZipFile(pathname, 'w', zipfile.ZIP_DEFLATED) as zf:
            for ap, p in archive_paths:
                logger.debug('Wrote %s to %s in wheel', p, ap)
                zf.write(p, ap)

    def build(self, paths, tags=None, wheel_version=None):
        """
        Build a wheel from files in specified paths, and use any specified tags
        when determining the name of the wheel.
        """
        if tags is None:
            tags = {}

        libkey = list(filter(lambda o: o in paths, ('purelib', 'platlib')))[0]
        if libkey == 'platlib':
            is_pure = 'false'
            default_pyver = [IMPVER]
            default_abi = [ABI]
            default_arch = [ARCH]
        else:
            is_pure = 'true'
            default_pyver = [PYVER]
            default_abi = ['none']
            default_arch = ['any']

        self.pyver = tags.get('pyver', default_pyver)
        self.abi = tags.get('abi', default_abi)
        self.arch = tags.get('arch', default_arch)

        libdir = paths[libkey]

        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        archive_paths = []

        # First, stuff which is not in site-packages
        for key in ('data', 'headers', 'scripts'):
            if key not in paths:
                continue
            path = paths[key]
            if os.path.isdir(path):
                for root, dirs, files in os.walk(path):
                    for fn in files:
                        p = fsdecode(os.path.join(root, fn))
                        rp = os.path.relpath(p, path)
                        ap = to_posix(os.path.join(data_dir, key, rp))
                        archive_paths.append((ap, p))
                        if key == 'scripts' and not p.endswith('.exe'):
                            with open(p, 'rb') as f:
                                data = f.read()
                            data = self.process_shebang(data)
                            with open(p, 'wb') as f:
                                f.write(data)

        # Now, stuff which is in site-packages, other than the
        # distinfo stuff.
        path = libdir
        distinfo = None
        for root, dirs, files in os.walk(path):
            if root == path:
                # At the top level only, save distinfo for later
                # and skip it for now
                for i, dn in enumerate(dirs):
                    dn = fsdecode(dn)
                    if dn.endswith('.dist-info'):
                        distinfo = os.path.join(root, dn)
                        del dirs[i]
                        break
                assert distinfo, '.dist-info directory expected, not found'

            for fn in files:
                # comment out next suite to leave .pyc files in
                if fsdecode(fn).endswith(('.pyc', '.pyo')):
                    continue
                p = os.path.join(root, fn)
                rp = to_posix(os.path.relpath(p, path))
                archive_paths.append((rp, p))

        # Now distinfo. Assumed to be flat, i.e. os.listdir is enough.
        files = os.listdir(distinfo)
        for fn in files:
            if fn not in ('RECORD', 'INSTALLER', 'SHARED', 'WHEEL'):
                p = fsdecode(os.path.join(distinfo, fn))
                ap = to_posix(os.path.join(info_dir, fn))
                archive_paths.append((ap, p))

        wheel_metadata = [
            'Wheel-Version: %d.%d' % (wheel_version or self.wheel_version),
            'Generator: distlib %s' % __version__,
            'Root-Is-Purelib: %s' % is_pure,
        ]
        for pyver, abi, arch in self.tags:
            wheel_metadata.append('Tag: %s-%s-%s' % (pyver, abi, arch))
        p = os.path.join(distinfo, 'WHEEL')
        with open(p, 'w') as f:
            f.write('\n'.join(wheel_metadata))
        ap = to_posix(os.path.join(info_dir, 'WHEEL'))
        archive_paths.append((ap, p))

        # sort the entries by archive path. Not needed by any spec, but it
        # keeps the archive listing and RECORD tidier than they would otherwise
        # be. Use the number of path segments to keep directory entries together,
        # and keep the dist-info stuff at the end.
        def sorter(t):
            ap = t[0]
            n = ap.count('/')
            if '.dist-info' in ap:
                n += 10000
            return (n, ap)
        archive_paths = sorted(archive_paths, key=sorter)

        # Now, at last, RECORD.
        # Paths in here are archive paths - nothing else makes sense.
        self.write_records((distinfo, info_dir), libdir, archive_paths)
        # Now, ready to build the zip file
        pathname = os.path.join(self.dirname, self.filename)
        self.build_zip(pathname, archive_paths)
        return pathname

    def skip_entry(self, arcname):
        """
        Determine whether an archive entry should be skipped when verifying
        or installing.
        """
        # The signature file won't be in RECORD,
        # and we  don't currently don't do anything with it
        # We also skip directories, as they won't be in RECORD
        # either. See:
        #
        # https://github.com/pypa/wheel/issues/294
        # https://github.com/pypa/wheel/issues/287
        # https://github.com/pypa/wheel/pull/289
        #
        return arcname.endswith(('/', '/RECORD.jws'))

    def install(self, paths, maker, **kwargs):
        """
        Install a wheel to the specified paths. If kwarg ``warner`` is
        specified, it should be a callable, which will be called with two
        tuples indicating the wheel version of this software and the wheel
        version in the file, if there is a discrepancy in the versions.
        This can be used to issue any warnings to raise any exceptions.
        If kwarg ``lib_only`` is True, only the purelib/platlib files are
        installed, and the headers, scripts, data and dist-info metadata are
        not written. If kwarg ``bytecode_hashed_invalidation`` is True, written
        bytecode will try to use file-hash based invalidation (PEP-552) on
        supported interpreter versions (CPython 2.7+).

        The return value is a :class:`InstalledDistribution` instance unless
        ``options.lib_only`` is True, in which case the return value is ``None``.
        """

        dry_run = maker.dry_run
        warner = kwargs.get('warner')
        lib_only = kwargs.get('lib_only', False)
        bc_hashed_invalidation = kwargs.get('bytecode_hashed_invalidation', False)

        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        metadata_name = posixpath.join(info_dir, LEGACY_METADATA_FILENAME)
        wheel_metadata_name = posixpath.join(info_dir, 'WHEEL')
        record_name = posixpath.join(info_dir, 'RECORD')

        wrapper = codecs.getreader('utf-8')

        with ZipFile(pathname, 'r') as zf:
            with zf.open(wheel_metadata_name) as bwf:
                wf = wrapper(bwf)
                message = message_from_file(wf)
            wv = message['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            if (file_version != self.wheel_version) and warner:
                warner(self.wheel_version, file_version)

            if message['Root-Is-Purelib'] == 'true':
                libdir = paths['purelib']
            else:
                libdir = paths['platlib']

            records = {}
            with zf.open(record_name) as bf:
                with CSVReader(stream=bf) as reader:
                    for row in reader:
                        p = row[0]
                        records[p] = row

            data_pfx = posixpath.join(data_dir, '')
            info_pfx = posixpath.join(info_dir, '')
            script_pfx = posixpath.join(data_dir, 'scripts', '')

            # make a new instance rather than a copy of maker's,
            # as we mutate it
            fileop = FileOperator(dry_run=dry_run)
            fileop.record = True    # so we can rollback if needed

            bc = not sys.dont_write_bytecode    # Double negatives. Lovely!

            outfiles = []   # for RECORD writing

            # for script copying/shebang processing
            workdir = tempfile.mkdtemp()
            # set target dir later
            # we default add_launchers to False, as the
            # Python Launcher should be used instead
            maker.source_dir = workdir
            maker.target_dir = None
            try:
                for zinfo in zf.infolist():
                    arcname = zinfo.filename
                    if isinstance(arcname, text_type):
                        u_arcname = arcname
                    else:
                        u_arcname = arcname.decode('utf-8')
                    if self.skip_entry(u_arcname):
                        continue
                    row = records[u_arcname]
                    if row[2] and str(zinfo.file_size) != row[2]:
                        raise DistlibException('size mismatch for '
                                               '%s' % u_arcname)
                    if row[1]:
                        kind, value = row[1].split('=', 1)
                        with zf.open(arcname) as bf:
                            data = bf.read()
                        _, digest = self.get_hash(data, kind)
                        if digest != value:
                            raise DistlibException('digest mismatch for '
                                                   '%s' % arcname)

                    if lib_only and u_arcname.startswith((info_pfx, data_pfx)):
                        logger.debug('lib_only: skipping %s', u_arcname)
                        continue
                    is_script = (u_arcname.startswith(script_pfx)
                                 and not u_arcname.endswith('.exe'))

                    if u_arcname.startswith(data_pfx):
                        _, where, rp = u_arcname.split('/', 2)
                        outfile = os.path.join(paths[where], convert_path(rp))
                    else:
                        # meant for site-packages.
                        if u_arcname in (wheel_metadata_name, record_name):
                            continue
                        outfile = os.path.join(libdir, convert_path(u_arcname))
                    if not is_script:
                        with zf.open(arcname) as bf:
                            fileop.copy_stream(bf, outfile)
                        # Issue #147: permission bits aren't preserved. Using
                        # zf.extract(zinfo, libdir) should have worked, but didn't,
                        # see https://www.thetopsites.net/article/53834422.shtml
                        # So ... manually preserve permission bits as given in zinfo
                        if os.name == 'posix':
                            # just set the normal permission bits
                            os.chmod(outfile, (zinfo.external_attr >> 16) & 0x1FF)
                        outfiles.append(outfile)
                        # Double check the digest of the written file
                        if not dry_run and row[1]:
                            with open(outfile, 'rb') as bf:
                                data = bf.read()
                                _, newdigest = self.get_hash(data, kind)
                                if newdigest != digest:
                                    raise DistlibException('digest mismatch '
                                                           'on write for '
                                                           '%s' % outfile)
                        if bc and outfile.endswith('.py'):
                            try:
                                pyc = fileop.byte_compile(outfile,
                                                          hashed_invalidation=bc_hashed_invalidation)
                                outfiles.append(pyc)
                            except Exception:
                                # Don't give up if byte-compilation fails,
                                # but log it and perhaps warn the user
                                logger.warning('Byte-compilation failed',
                                               exc_info=True)
                    else:
                        fn = os.path.basename(convert_path(arcname))
                        workname = os.path.join(workdir, fn)
                        with zf.open(arcname) as bf:
                            fileop.copy_stream(bf, workname)

                        dn, fn = os.path.split(outfile)
                        maker.target_dir = dn
                        filenames = maker.make(fn)
                        fileop.set_executable_mode(filenames)
                        outfiles.extend(filenames)

                if lib_only:
                    logger.debug('lib_only: returning None')
                    dist = None
                else:
                    # Generate scripts

                    # Try to get pydist.json so we can see if there are
                    # any commands to generate. If this fails (e.g. because
                    # of a legacy wheel), log a warning but don't give up.
                    commands = None
                    file_version = self.info['Wheel-Version']
                    if file_version == '1.0':
                        # Use legacy info
                        ep = posixpath.join(info_dir, 'entry_points.txt')
                        try:
                            with zf.open(ep) as bwf:
                                epdata = read_exports(bwf)
                            commands = {}
                            for key in ('console', 'gui'):
                                k = '%s_scripts' % key
                                if k in epdata:
                                    commands['wrap_%s' % key] = d = {}
                                    for v in epdata[k].values():
                                        s = '%s:%s' % (v.prefix, v.suffix)
                                        if v.flags:
                                            s += ' [%s]' % ','.join(v.flags)
                                        d[v.name] = s
                        except Exception:
                            logger.warning('Unable to read legacy script '
                                           'metadata, so cannot generate '
                                           'scripts')
                    else:
                        try:
                            with zf.open(metadata_name) as bwf:
                                wf = wrapper(bwf)
                                commands = json.load(wf).get('extensions')
                                if commands:
                                    commands = commands.get('python.commands')
                        except Exception:
                            logger.warning('Unable to read JSON metadata, so '
                                           'cannot generate scripts')
                    if commands:
                        console_scripts = commands.get('wrap_console', {})
                        gui_scripts = commands.get('wrap_gui', {})
                        if console_scripts or gui_scripts:
                            script_dir = paths.get('scripts', '')
                            if not os.path.isdir(script_dir):
                                raise ValueError('Valid script path not '
                                                 'specified')
                            maker.target_dir = script_dir
                            for k, v in console_scripts.items():
                                script = '%s = %s' % (k, v)
                                filenames = maker.make(script)
                                fileop.set_executable_mode(filenames)

                            if gui_scripts:
                                options = {'gui': True }
                                for k, v in gui_scripts.items():
                                    script = '%s = %s' % (k, v)
                                    filenames = maker.make(script, options)
                                    fileop.set_executable_mode(filenames)

                    p = os.path.join(libdir, info_dir)
                    dist = InstalledDistribution(p)

                    # Write SHARED
                    paths = dict(paths)     # don't change passed in dict
                    del paths['purelib']
                    del paths['platlib']
                    paths['lib'] = libdir
                    p = dist.write_shared_locations(paths, dry_run)
                    if p:
                        outfiles.append(p)

                    # Write RECORD
                    dist.write_installed_files(outfiles, paths['prefix'],
                                               dry_run)
                return dist
            except Exception:  # pragma: no cover
                logger.exception('installation failed.')
                fileop.rollback()
                raise
            finally:
                shutil.rmtree(workdir)

    def _get_dylib_cache(self):
        global cache
        if cache is None:
            # Use native string to avoid issues on 2.x: see Python #20140.
            base = os.path.join(get_cache_base(), str('dylib-cache'),
                                '%s.%s' % sys.version_info[:2])
            cache = Cache(base)
        return cache

    def _get_extensions(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        arcname = posixpath.join(info_dir, 'EXTENSIONS')
        wrapper = codecs.getreader('utf-8')
        result = []
        with ZipFile(pathname, 'r') as zf:
            try:
                with zf.open(arcname) as bf:
                    wf = wrapper(bf)
                    extensions = json.load(wf)
                    cache = self._get_dylib_cache()
                    prefix = cache.prefix_to_dir(pathname)
                    cache_base = os.path.join(cache.base, prefix)
                    if not os.path.isdir(cache_base):
                        os.makedirs(cache_base)
                    for name, relpath in extensions.items():
                        dest = os.path.join(cache_base, convert_path(relpath))
                        if not os.path.exists(dest):
                            extract = True
                        else:
                            file_time = os.stat(dest).st_mtime
                            file_time = datetime.datetime.fromtimestamp(file_time)
                            info = zf.getinfo(relpath)
                            wheel_time = datetime.datetime(*info.date_time)
                            extract = wheel_time > file_time
                        if extract:
                            zf.extract(relpath, cache_base)
                        result.append((name, dest))
            except KeyError:
                pass
        return result

    def is_compatible(self):
        """
        Determine if a wheel is compatible with the running system.
        """
        return is_compatible(self)

    def is_mountable(self):
        """
        Determine if a wheel is asserted as mountable by its metadata.
        """
        return True # for now - metadata details TBD

    def mount(self, append=False):
        pathname = os.path.abspath(os.path.join(self.dirname, self.filename))
        if not self.is_compatible():
            msg = 'Wheel %s not compatible with this Python.' % pathname
            raise DistlibException(msg)
        if not self.is_mountable():
            msg = 'Wheel %s is marked as not mountable.' % pathname
            raise DistlibException(msg)
        if pathname in sys.path:
            logger.debug('%s already in path', pathname)
        else:
            if append:
                sys.path.append(pathname)
            else:
                sys.path.insert(0, pathname)
            extensions = self._get_extensions()
            if extensions:
                if _hook not in sys.meta_path:
                    sys.meta_path.append(_hook)
                _hook.add(pathname, extensions)

    def unmount(self):
        pathname = os.path.abspath(os.path.join(self.dirname, self.filename))
        if pathname not in sys.path:
            logger.debug('%s not in path', pathname)
        else:
            sys.path.remove(pathname)
            if pathname in _hook.impure_wheels:
                _hook.remove(pathname)
            if not _hook.impure_wheels:
                if _hook in sys.meta_path:
                    sys.meta_path.remove(_hook)

    def verify(self):
        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        data_dir = '%s.data' % name_ver
        info_dir = '%s.dist-info' % name_ver

        metadata_name = posixpath.join(info_dir, LEGACY_METADATA_FILENAME)
        wheel_metadata_name = posixpath.join(info_dir, 'WHEEL')
        record_name = posixpath.join(info_dir, 'RECORD')

        wrapper = codecs.getreader('utf-8')

        with ZipFile(pathname, 'r') as zf:
            with zf.open(wheel_metadata_name) as bwf:
                wf = wrapper(bwf)
                message = message_from_file(wf)
            wv = message['Wheel-Version'].split('.', 1)
            file_version = tuple([int(i) for i in wv])
            # TODO version verification

            records = {}
            with zf.open(record_name) as bf:
                with CSVReader(stream=bf) as reader:
                    for row in reader:
                        p = row[0]
                        records[p] = row

            for zinfo in zf.infolist():
                arcname = zinfo.filename
                if isinstance(arcname, text_type):
                    u_arcname = arcname
                else:
                    u_arcname = arcname.decode('utf-8')
                # See issue #115: some wheels have .. in their entries, but
                # in the filename ... e.g. __main__..py ! So the check is
                # updated to look for .. in the directory portions
                p = u_arcname.split('/')
                if '..' in p:
                    raise DistlibException('invalid entry in '
                                           'wheel: %r' % u_arcname)

                if self.skip_entry(u_arcname):
                    continue
                row = records[u_arcname]
                if row[2] and str(zinfo.file_size) != row[2]:
                    raise DistlibException('size mismatch for '
                                           '%s' % u_arcname)
                if row[1]:
                    kind, value = row[1].split('=', 1)
                    with zf.open(arcname) as bf:
                        data = bf.read()
                    _, digest = self.get_hash(data, kind)
                    if digest != value:
                        raise DistlibException('digest mismatch for '
                                               '%s' % arcname)

    def update(self, modifier, dest_dir=None, **kwargs):
        """
        Update the contents of a wheel in a generic way. The modifier should
        be a callable which expects a dictionary argument: its keys are
        archive-entry paths, and its values are absolute filesystem paths
        where the contents the corresponding archive entries can be found. The
        modifier is free to change the contents of the files pointed to, add
        new entries and remove entries, before returning. This method will
        extract the entire contents of the wheel to a temporary location, call
        the modifier, and then use the passed (and possibly updated)
        dictionary to write a new wheel. If ``dest_dir`` is specified, the new
        wheel is written there -- otherwise, the original wheel is overwritten.

        The modifier should return True if it updated the wheel, else False.
        This method returns the same value the modifier returns.
        """

        def get_version(path_map, info_dir):
            version = path = None
            key = '%s/%s' % (info_dir, LEGACY_METADATA_FILENAME)
            if key not in path_map:
                key = '%s/PKG-INFO' % info_dir
            if key in path_map:
                path = path_map[key]
                version = Metadata(path=path).version
            return version, path

        def update_version(version, path):
            updated = None
            try:
                v = NormalizedVersion(version)
                i = version.find('-')
                if i < 0:
                    updated = '%s+1' % version
                else:
                    parts = [int(s) for s in version[i + 1:].split('.')]
                    parts[-1] += 1
                    updated = '%s+%s' % (version[:i],
                                         '.'.join(str(i) for i in parts))
            except UnsupportedVersionError:
                logger.debug('Cannot update non-compliant (PEP-440) '
                             'version %r', version)
            if updated:
                md = Metadata(path=path)
                md.version = updated
                legacy = path.endswith(LEGACY_METADATA_FILENAME)
                md.write(path=path, legacy=legacy)
                logger.debug('Version updated from %r to %r', version,
                             updated)

        pathname = os.path.join(self.dirname, self.filename)
        name_ver = '%s-%s' % (self.name, self.version)
        info_dir = '%s.dist-info' % name_ver
        record_name = posixpath.join(info_dir, 'RECORD')
        with tempdir() as workdir:
            with ZipFile(pathname, 'r') as zf:
                path_map = {}
                for zinfo in zf.infolist():
                    arcname = zinfo.filename
                    if isinstance(arcname, text_type):
                        u_arcname = arcname
                    else:
                        u_arcname = arcname.decode('utf-8')
                    if u_arcname == record_name:
                        continue
                    if '..' in u_arcname:
                        raise DistlibException('invalid entry in '
                                               'wheel: %r' % u_arcname)
                    zf.extract(zinfo, workdir)
                    path = os.path.join(workdir, convert_path(u_arcname))
                    path_map[u_arcname] = path

            # Remember the version.
            original_version, _ = get_version(path_map, info_dir)
            # Files extracted. Call the modifier.
            modified = modifier(path_map, **kwargs)
            if modified:
                # Something changed - need to build a new wheel.
                current_version, path = get_version(path_map, info_dir)
                if current_version and (current_version == original_version):
                    # Add or update local version to signify changes.
                    update_version(current_version, path)
                # Decide where the new wheel goes.
                if dest_dir is None:
                    fd, newpath = tempfile.mkstemp(suffix='.whl',
                                                   prefix='wheel-update-',
                                                   dir=workdir)
                    os.close(fd)
                else:
                    if not os.path.isdir(dest_dir):
                        raise DistlibException('Not a directory: %r' % dest_dir)
                    newpath = os.path.join(dest_dir, self.filename)
                archive_paths = list(path_map.items())
                distinfo = os.path.join(workdir, info_dir)
                info = distinfo, info_dir
                self.write_records(info, workdir, archive_paths)
                self.build_zip(newpath, archive_paths)
                if dest_dir is None:
                    shutil.copyfile(newpath, pathname)
        return modified

def _get_glibc_version():
    import platform
    ver = platform.libc_ver()
    result = []
    if ver[0] == 'glibc':
        for s in ver[1].split('.'):
            result.append(int(s) if s.isdigit() else 0)
        result = tuple(result)
    return result

def compatible_tags():
    """
    Return (pyver, abi, arch) tuples compatible with this Python.
    """
    versions = [VER_SUFFIX]
    major = VER_SUFFIX[0]
    for minor in range(sys.version_info[1] - 1, - 1, -1):
        versions.append(''.join([major, str(minor)]))

    abis = []
    for suffix, _, _ in imp.get_suffixes():
        if suffix.startswith('.abi'):
            abis.append(suffix.split('.', 2)[1])
    abis.sort()
    if ABI != 'none':
        abis.insert(0, ABI)
    abis.append('none')
    result = []

    arches = [ARCH]
    if sys.platform == 'darwin':
        m = re.match(r'(\w+)_(\d+)_(\d+)_(\w+)$', ARCH)
        if m:
            name, major, minor, arch = m.groups()
            minor = int(minor)
            matches = [arch]
            if arch in ('i386', 'ppc'):
                matches.append('fat')
            if arch in ('i386', 'ppc', 'x86_64'):
                matches.append('fat3')
            if arch in ('ppc64', 'x86_64'):
                matches.append('fat64')
            if arch in ('i386', 'x86_64'):
                matches.append('intel')
            if arch in ('i386', 'x86_64', 'intel', 'ppc', 'ppc64'):
                matches.append('universal')
            while minor >= 0:
                for match in matches:
                    s = '%s_%s_%s_%s' % (name, major, minor, match)
                    if s != ARCH:   # already there
                        arches.append(s)
                minor -= 1

    # Most specific - our Python version, ABI and arch
    for abi in abis:
        for arch in arches:
            result.append((''.join((IMP_PREFIX, versions[0])), abi, arch))
            # manylinux
            if abi != 'none' and sys.platform.startswith('linux'):
                arch = arch.replace('linux_', '')
                parts = _get_glibc_version()
                if len(parts) == 2:
                    if parts >= (2, 5):
                        result.append((''.join((IMP_PREFIX, versions[0])), abi,
                                       'manylinux1_%s' % arch))
                    if parts >= (2, 12):
                        result.append((''.join((IMP_PREFIX, versions[0])), abi,
                                       'manylinux2010_%s' % arch))
                    if parts >= (2, 17):
                        result.append((''.join((IMP_PREFIX, versions[0])), abi,
                                       'manylinux2014_%s' % arch))
                    result.append((''.join((IMP_PREFIX, versions[0])), abi,
                                   'manylinux_%s_%s_%s' % (parts[0], parts[1],
                                                           arch)))

    # where no ABI / arch dependency, but IMP_PREFIX dependency
    for i, version in enumerate(versions):
        result.append((''.join((IMP_PREFIX, version)), 'none', 'any'))
        if i == 0:
            result.append((''.join((IMP_PREFIX, version[0])), 'none', 'any'))

    # no IMP_PREFIX, ABI or arch dependency
    for i, version in enumerate(versions):
        result.append((''.join(('py', version)), 'none', 'any'))
        if i == 0:
            result.append((''.join(('py', version[0])), 'none', 'any'))

    return set(result)


COMPATIBLE_TAGS = compatible_tags()

del compatible_tags


def is_compatible(wheel, tags=None):
    if not isinstance(wheel, Wheel):
        wheel = Wheel(wheel)    # assume it's a filename
    result = False
    if tags is None:
        tags = COMPATIBLE_TAGS
    for ver, abi, arch in tags:
        if ver in wheel.pyver and abi in wheel.abi and arch in wheel.arch:
            result = True
            break
    return result
site-packages/pip/_vendor/distlib/scripts.py000064400000042470150564745010015221 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2015 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from io import BytesIO
import logging
import os
import re
import struct
import sys

from .compat import sysconfig, detect_encoding, ZipFile
from .resources import finder
from .util import (FileOperator, get_export_entry, convert_path,
                   get_executable, get_platform, in_venv)

logger = logging.getLogger(__name__)

_DEFAULT_MANIFEST = '''
<?xml version="1.0" encoding="UTF-8" standalone="yes"?>
<assembly xmlns="urn:schemas-microsoft-com:asm.v1" manifestVersion="1.0">
 <assemblyIdentity version="1.0.0.0"
 processorArchitecture="X86"
 name="%s"
 type="win32"/>

 <!-- Identify the application security requirements. -->
 <trustInfo xmlns="urn:schemas-microsoft-com:asm.v3">
 <security>
 <requestedPrivileges>
 <requestedExecutionLevel level="asInvoker" uiAccess="false"/>
 </requestedPrivileges>
 </security>
 </trustInfo>
</assembly>'''.strip()

# check if Python is called on the first line with this expression
FIRST_LINE_RE = re.compile(b'^#!.*pythonw?[0-9.]*([ \t].*)?$')
SCRIPT_TEMPLATE = r'''# -*- coding: utf-8 -*-
import re
import sys
from %(module)s import %(import_name)s
if __name__ == '__main__':
    sys.argv[0] = re.sub(r'(-script\.pyw|\.exe)?$', '', sys.argv[0])
    sys.exit(%(func)s())
'''


def enquote_executable(executable):
    if ' ' in executable:
        # make sure we quote only the executable in case of env
        # for example /usr/bin/env "/dir with spaces/bin/jython"
        # instead of "/usr/bin/env /dir with spaces/bin/jython"
        # otherwise whole
        if executable.startswith('/usr/bin/env '):
            env, _executable = executable.split(' ', 1)
            if ' ' in _executable and not _executable.startswith('"'):
                executable = '%s "%s"' % (env, _executable)
        else:
            if not executable.startswith('"'):
                executable = '"%s"' % executable
    return executable

# Keep the old name around (for now), as there is at least one project using it!
_enquote_executable = enquote_executable

class ScriptMaker(object):
    """
    A class to copy or create scripts from source scripts or callable
    specifications.
    """
    script_template = SCRIPT_TEMPLATE

    executable = None  # for shebangs

    def __init__(self, source_dir, target_dir, add_launchers=True,
                 dry_run=False, fileop=None):
        self.source_dir = source_dir
        self.target_dir = target_dir
        self.add_launchers = add_launchers
        self.force = False
        self.clobber = False
        # It only makes sense to set mode bits on POSIX.
        self.set_mode = (os.name == 'posix') or (os.name == 'java' and
                                                 os._name == 'posix')
        self.variants = set(('', 'X.Y'))
        self._fileop = fileop or FileOperator(dry_run)

        self._is_nt = os.name == 'nt' or (
            os.name == 'java' and os._name == 'nt')
        self.version_info = sys.version_info

    def _get_alternate_executable(self, executable, options):
        if options.get('gui', False) and self._is_nt:  # pragma: no cover
            dn, fn = os.path.split(executable)
            fn = fn.replace('python', 'pythonw')
            executable = os.path.join(dn, fn)
        return executable

    if sys.platform.startswith('java'):  # pragma: no cover
        def _is_shell(self, executable):
            """
            Determine if the specified executable is a script
            (contains a #! line)
            """
            try:
                with open(executable) as fp:
                    return fp.read(2) == '#!'
            except (OSError, IOError):
                logger.warning('Failed to open %s', executable)
                return False

        def _fix_jython_executable(self, executable):
            if self._is_shell(executable):
                # Workaround for Jython is not needed on Linux systems.
                import java

                if java.lang.System.getProperty('os.name') == 'Linux':
                    return executable
            elif executable.lower().endswith('jython.exe'):
                # Use wrapper exe for Jython on Windows
                return executable
            return '/usr/bin/env %s' % executable

    def _build_shebang(self, executable, post_interp):
        """
        Build a shebang line. In the simple case (on Windows, or a shebang line
        which is not too long or contains spaces) use a simple formulation for
        the shebang. Otherwise, use /bin/sh as the executable, with a contrived
        shebang which allows the script to run either under Python or sh, using
        suitable quoting. Thanks to Harald Nordgren for his input.

        See also: http://www.in-ulm.de/~mascheck/various/shebang/#length
                  https://hg.mozilla.org/mozilla-central/file/tip/mach
        """
        if os.name != 'posix':
            simple_shebang = True
        else:
            # Add 3 for '#!' prefix and newline suffix.
            shebang_length = len(executable) + len(post_interp) + 3
            if sys.platform == 'darwin':
                max_shebang_length = 512
            else:
                max_shebang_length = 127
            simple_shebang = ((b' ' not in executable) and
                              (shebang_length <= max_shebang_length))

        if simple_shebang:
            result = b'#!' + executable + post_interp + b'\n'
        else:
            result = b'#!/bin/sh\n'
            result += b"'''exec' " + executable + post_interp + b' "$0" "$@"\n'
            result += b"' '''"
        return result

    def _get_shebang(self, encoding, post_interp=b'', options=None):
        enquote = True
        if self.executable:
            executable = self.executable
            enquote = False     # assume this will be taken care of
        elif not sysconfig.is_python_build():
            executable = get_executable()
        elif in_venv():  # pragma: no cover
            executable = os.path.join(sysconfig.get_path('scripts'),
                            'python%s' % sysconfig.get_config_var('EXE'))
        else:  # pragma: no cover
            executable = os.path.join(
                sysconfig.get_config_var('BINDIR'),
               'python%s%s' % (sysconfig.get_config_var('VERSION'),
                               sysconfig.get_config_var('EXE')))
            if not os.path.isfile(executable):
                # for Python builds from source on Windows, no Python executables with
                # a version suffix are created, so we use python.exe
                executable = os.path.join(sysconfig.get_config_var('BINDIR'),
                                'python%s' % (sysconfig.get_config_var('EXE')))
        if options:
            executable = self._get_alternate_executable(executable, options)

        if sys.platform.startswith('java'):  # pragma: no cover
            executable = self._fix_jython_executable(executable)

        # Normalise case for Windows - COMMENTED OUT
        # executable = os.path.normcase(executable)
        # N.B. The normalising operation above has been commented out: See
        # issue #124. Although paths in Windows are generally case-insensitive,
        # they aren't always. For example, a path containing a ẞ (which is a
        # LATIN CAPITAL LETTER SHARP S - U+1E9E) is normcased to ß (which is a
        # LATIN SMALL LETTER SHARP S' - U+00DF). The two are not considered by
        # Windows as equivalent in path names.

        # If the user didn't specify an executable, it may be necessary to
        # cater for executable paths with spaces (not uncommon on Windows)
        if enquote:
            executable = enquote_executable(executable)
        # Issue #51: don't use fsencode, since we later try to
        # check that the shebang is decodable using utf-8.
        executable = executable.encode('utf-8')
        # in case of IronPython, play safe and enable frames support
        if (sys.platform == 'cli' and '-X:Frames' not in post_interp
            and '-X:FullFrames' not in post_interp):  # pragma: no cover
            post_interp += b' -X:Frames'
        shebang = self._build_shebang(executable, post_interp)
        # Python parser starts to read a script using UTF-8 until
        # it gets a #coding:xxx cookie. The shebang has to be the
        # first line of a file, the #coding:xxx cookie cannot be
        # written before. So the shebang has to be decodable from
        # UTF-8.
        try:
            shebang.decode('utf-8')
        except UnicodeDecodeError:  # pragma: no cover
            raise ValueError(
                'The shebang (%r) is not decodable from utf-8' % shebang)
        # If the script is encoded to a custom encoding (use a
        # #coding:xxx cookie), the shebang has to be decodable from
        # the script encoding too.
        if encoding != 'utf-8':
            try:
                shebang.decode(encoding)
            except UnicodeDecodeError:  # pragma: no cover
                raise ValueError(
                    'The shebang (%r) is not decodable '
                    'from the script encoding (%r)' % (shebang, encoding))
        return shebang

    def _get_script_text(self, entry):
        return self.script_template % dict(module=entry.prefix,
                                           import_name=entry.suffix.split('.')[0],
                                           func=entry.suffix)

    manifest = _DEFAULT_MANIFEST

    def get_manifest(self, exename):
        base = os.path.basename(exename)
        return self.manifest % base

    def _write_script(self, names, shebang, script_bytes, filenames, ext):
        use_launcher = self.add_launchers and self._is_nt
        linesep = os.linesep.encode('utf-8')
        if not shebang.endswith(linesep):
            shebang += linesep
        if not use_launcher:
            script_bytes = shebang + script_bytes
        else:  # pragma: no cover
            if ext == 'py':
                launcher = self._get_launcher('t')
            else:
                launcher = self._get_launcher('w')
            stream = BytesIO()
            with ZipFile(stream, 'w') as zf:
                zf.writestr('__main__.py', script_bytes)
            zip_data = stream.getvalue()
            script_bytes = launcher + shebang + zip_data
        for name in names:
            outname = os.path.join(self.target_dir, name)
            if use_launcher:  # pragma: no cover
                n, e = os.path.splitext(outname)
                if e.startswith('.py'):
                    outname = n
                outname = '%s.exe' % outname
                try:
                    self._fileop.write_binary_file(outname, script_bytes)
                except Exception:
                    # Failed writing an executable - it might be in use.
                    logger.warning('Failed to write executable - trying to '
                                   'use .deleteme logic')
                    dfname = '%s.deleteme' % outname
                    if os.path.exists(dfname):
                        os.remove(dfname)       # Not allowed to fail here
                    os.rename(outname, dfname)  # nor here
                    self._fileop.write_binary_file(outname, script_bytes)
                    logger.debug('Able to replace executable using '
                                 '.deleteme logic')
                    try:
                        os.remove(dfname)
                    except Exception:
                        pass    # still in use - ignore error
            else:
                if self._is_nt and not outname.endswith('.' + ext):  # pragma: no cover
                    outname = '%s.%s' % (outname, ext)
                if os.path.exists(outname) and not self.clobber:
                    logger.warning('Skipping existing file %s', outname)
                    continue
                self._fileop.write_binary_file(outname, script_bytes)
                if self.set_mode:
                    self._fileop.set_executable_mode([outname])
            filenames.append(outname)

    variant_separator = '-'

    def get_script_filenames(self, name):
        result = set()
        if '' in self.variants:
            result.add(name)
        if 'X' in self.variants:
            result.add('%s%s' % (name, self.version_info[0]))
        if 'X.Y' in self.variants:
            result.add('%s%s%s.%s' % (name, self.variant_separator,
                                      self.version_info[0], self.version_info[1]))
        return result

    def _make_script(self, entry, filenames, options=None):
        post_interp = b''
        if options:
            args = options.get('interpreter_args', [])
            if args:
                args = ' %s' % ' '.join(args)
                post_interp = args.encode('utf-8')
        shebang = self._get_shebang('utf-8', post_interp, options=options)
        script = self._get_script_text(entry).encode('utf-8')
        scriptnames = self.get_script_filenames(entry.name)
        if options and options.get('gui', False):
            ext = 'pyw'
        else:
            ext = 'py'
        self._write_script(scriptnames, shebang, script, filenames, ext)

    def _copy_script(self, script, filenames):
        adjust = False
        script = os.path.join(self.source_dir, convert_path(script))
        outname = os.path.join(self.target_dir, os.path.basename(script))
        if not self.force and not self._fileop.newer(script, outname):
            logger.debug('not copying %s (up-to-date)', script)
            return

        # Always open the file, but ignore failures in dry-run mode --
        # that way, we'll get accurate feedback if we can read the
        # script.
        try:
            f = open(script, 'rb')
        except IOError:  # pragma: no cover
            if not self.dry_run:
                raise
            f = None
        else:
            first_line = f.readline()
            if not first_line:  # pragma: no cover
                logger.warning('%s is an empty file (skipping)', script)
                return

            match = FIRST_LINE_RE.match(first_line.replace(b'\r\n', b'\n'))
            if match:
                adjust = True
                post_interp = match.group(1) or b''

        if not adjust:
            if f:
                f.close()
            self._fileop.copy_file(script, outname)
            if self.set_mode:
                self._fileop.set_executable_mode([outname])
            filenames.append(outname)
        else:
            logger.info('copying and adjusting %s -> %s', script,
                        self.target_dir)
            if not self._fileop.dry_run:
                encoding, lines = detect_encoding(f.readline)
                f.seek(0)
                shebang = self._get_shebang(encoding, post_interp)
                if b'pythonw' in first_line:  # pragma: no cover
                    ext = 'pyw'
                else:
                    ext = 'py'
                n = os.path.basename(outname)
                self._write_script([n], shebang, f.read(), filenames, ext)
            if f:
                f.close()

    @property
    def dry_run(self):
        return self._fileop.dry_run

    @dry_run.setter
    def dry_run(self, value):
        self._fileop.dry_run = value

    if os.name == 'nt' or (os.name == 'java' and os._name == 'nt'):  # pragma: no cover
        # Executable launcher support.
        # Launchers are from https://bitbucket.org/vinay.sajip/simple_launcher/

        def _get_launcher(self, kind):
            if struct.calcsize('P') == 8:   # 64-bit
                bits = '64'
            else:
                bits = '32'
            platform_suffix = '-arm' if get_platform() == 'win-arm64' else ''
            name = '%s%s%s.exe' % (kind, bits, platform_suffix)
            # Issue 31: don't hardcode an absolute package name, but
            # determine it relative to the current package
            distlib_package = __name__.rsplit('.', 1)[0]
            resource = finder(distlib_package).find(name)
            if not resource:
                msg = ('Unable to find resource %s in package %s' % (name,
                       distlib_package))
                raise ValueError(msg)
            return resource.bytes

    # Public API follows

    def make(self, specification, options=None):
        """
        Make a script.

        :param specification: The specification, which is either a valid export
                              entry specification (to make a script from a
                              callable) or a filename (to make a script by
                              copying from a source location).
        :param options: A dictionary of options controlling script generation.
        :return: A list of all absolute pathnames written to.
        """
        filenames = []
        entry = get_export_entry(specification)
        if entry is None:
            self._copy_script(specification, filenames)
        else:
            self._make_script(entry, filenames, options=options)
        return filenames

    def make_multiple(self, specifications, options=None):
        """
        Take a list of specifications and make scripts from them,
        :param specifications: A list of specifications.
        :return: A list of all absolute pathnames written to,
        """
        filenames = []
        for specification in specifications:
            filenames.extend(self.make(specification, options))
        return filenames
site-packages/pip/_vendor/distlib/compat.py000064400000121027150564745010015011 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import absolute_import

import os
import re
import sys

try:
    import ssl
except ImportError:  # pragma: no cover
    ssl = None

if sys.version_info[0] < 3:  # pragma: no cover
    from StringIO import StringIO
    string_types = basestring,
    text_type = unicode
    from types import FileType as file_type
    import __builtin__ as builtins
    import ConfigParser as configparser
    from ._backport import shutil
    from urlparse import urlparse, urlunparse, urljoin, urlsplit, urlunsplit
    from urllib import (urlretrieve, quote as _quote, unquote, url2pathname,
                        pathname2url, ContentTooShortError, splittype)

    def quote(s):
        if isinstance(s, unicode):
            s = s.encode('utf-8')
        return _quote(s)

    import urllib2
    from urllib2 import (Request, urlopen, URLError, HTTPError,
                         HTTPBasicAuthHandler, HTTPPasswordMgr,
                         HTTPHandler, HTTPRedirectHandler,
                         build_opener)
    if ssl:
        from urllib2 import HTTPSHandler
    import httplib
    import xmlrpclib
    import Queue as queue
    from HTMLParser import HTMLParser
    import htmlentitydefs
    raw_input = raw_input
    from itertools import ifilter as filter
    from itertools import ifilterfalse as filterfalse

    # Leaving this around for now, in case it needs resurrecting in some way
    # _userprog = None
    # def splituser(host):
        # """splituser('user[:passwd]@host[:port]') --> 'user[:passwd]', 'host[:port]'."""
        # global _userprog
        # if _userprog is None:
            # import re
            # _userprog = re.compile('^(.*)@(.*)$')

        # match = _userprog.match(host)
        # if match: return match.group(1, 2)
        # return None, host

else:  # pragma: no cover
    from io import StringIO
    string_types = str,
    text_type = str
    from io import TextIOWrapper as file_type
    import builtins
    import configparser
    import shutil
    from urllib.parse import (urlparse, urlunparse, urljoin, quote,
                              unquote, urlsplit, urlunsplit, splittype)
    from urllib.request import (urlopen, urlretrieve, Request, url2pathname,
                                pathname2url,
                                HTTPBasicAuthHandler, HTTPPasswordMgr,
                                HTTPHandler, HTTPRedirectHandler,
                                build_opener)
    if ssl:
        from urllib.request import HTTPSHandler
    from urllib.error import HTTPError, URLError, ContentTooShortError
    import http.client as httplib
    import urllib.request as urllib2
    import xmlrpc.client as xmlrpclib
    import queue
    from html.parser import HTMLParser
    import html.entities as htmlentitydefs
    raw_input = input
    from itertools import filterfalse
    filter = filter


try:
    from ssl import match_hostname, CertificateError
except ImportError: # pragma: no cover
    class CertificateError(ValueError):
        pass


    def _dnsname_match(dn, hostname, max_wildcards=1):
        """Matching according to RFC 6125, section 6.4.3

        http://tools.ietf.org/html/rfc6125#section-6.4.3
        """
        pats = []
        if not dn:
            return False

        parts = dn.split('.')
        leftmost, remainder = parts[0], parts[1:]

        wildcards = leftmost.count('*')
        if wildcards > max_wildcards:
            # Issue #17980: avoid denials of service by refusing more
            # than one wildcard per fragment.  A survey of established
            # policy among SSL implementations showed it to be a
            # reasonable choice.
            raise CertificateError(
                "too many wildcards in certificate DNS name: " + repr(dn))

        # speed up common case w/o wildcards
        if not wildcards:
            return dn.lower() == hostname.lower()

        # RFC 6125, section 6.4.3, subitem 1.
        # The client SHOULD NOT attempt to match a presented identifier in which
        # the wildcard character comprises a label other than the left-most label.
        if leftmost == '*':
            # When '*' is a fragment by itself, it matches a non-empty dotless
            # fragment.
            pats.append('[^.]+')
        elif leftmost.startswith('xn--') or hostname.startswith('xn--'):
            # RFC 6125, section 6.4.3, subitem 3.
            # The client SHOULD NOT attempt to match a presented identifier
            # where the wildcard character is embedded within an A-label or
            # U-label of an internationalized domain name.
            pats.append(re.escape(leftmost))
        else:
            # Otherwise, '*' matches any dotless string, e.g. www*
            pats.append(re.escape(leftmost).replace(r'\*', '[^.]*'))

        # add the remaining fragments, ignore any wildcards
        for frag in remainder:
            pats.append(re.escape(frag))

        pat = re.compile(r'\A' + r'\.'.join(pats) + r'\Z', re.IGNORECASE)
        return pat.match(hostname)


    def match_hostname(cert, hostname):
        """Verify that *cert* (in decoded format as returned by
        SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
        rules are followed, but IP addresses are not accepted for *hostname*.

        CertificateError is raised on failure. On success, the function
        returns nothing.
        """
        if not cert:
            raise ValueError("empty or no certificate, match_hostname needs a "
                             "SSL socket or SSL context with either "
                             "CERT_OPTIONAL or CERT_REQUIRED")
        dnsnames = []
        san = cert.get('subjectAltName', ())
        for key, value in san:
            if key == 'DNS':
                if _dnsname_match(value, hostname):
                    return
                dnsnames.append(value)
        if not dnsnames:
            # The subject is only checked when there is no dNSName entry
            # in subjectAltName
            for sub in cert.get('subject', ()):
                for key, value in sub:
                    # XXX according to RFC 2818, the most specific Common Name
                    # must be used.
                    if key == 'commonName':
                        if _dnsname_match(value, hostname):
                            return
                        dnsnames.append(value)
        if len(dnsnames) > 1:
            raise CertificateError("hostname %r "
                "doesn't match either of %s"
                % (hostname, ', '.join(map(repr, dnsnames))))
        elif len(dnsnames) == 1:
            raise CertificateError("hostname %r "
                "doesn't match %r"
                % (hostname, dnsnames[0]))
        else:
            raise CertificateError("no appropriate commonName or "
                "subjectAltName fields were found")


try:
    from types import SimpleNamespace as Container
except ImportError:  # pragma: no cover
    class Container(object):
        """
        A generic container for when multiple values need to be returned
        """
        def __init__(self, **kwargs):
            self.__dict__.update(kwargs)


try:
    from shutil import which
except ImportError:  # pragma: no cover
    # Implementation from Python 3.3
    def which(cmd, mode=os.F_OK | os.X_OK, path=None):
        """Given a command, mode, and a PATH string, return the path which
        conforms to the given mode on the PATH, or None if there is no such
        file.

        `mode` defaults to os.F_OK | os.X_OK. `path` defaults to the result
        of os.environ.get("PATH"), or can be overridden with a custom search
        path.

        """
        # Check that a given file can be accessed with the correct mode.
        # Additionally check that `file` is not a directory, as on Windows
        # directories pass the os.access check.
        def _access_check(fn, mode):
            return (os.path.exists(fn) and os.access(fn, mode)
                    and not os.path.isdir(fn))

        # If we're given a path with a directory part, look it up directly rather
        # than referring to PATH directories. This includes checking relative to the
        # current directory, e.g. ./script
        if os.path.dirname(cmd):
            if _access_check(cmd, mode):
                return cmd
            return None

        if path is None:
            path = os.environ.get("PATH", os.defpath)
        if not path:
            return None
        path = path.split(os.pathsep)

        if sys.platform == "win32":
            # The current directory takes precedence on Windows.
            if not os.curdir in path:
                path.insert(0, os.curdir)

            # PATHEXT is necessary to check on Windows.
            pathext = os.environ.get("PATHEXT", "").split(os.pathsep)
            # See if the given file matches any of the expected path extensions.
            # This will allow us to short circuit when given "python.exe".
            # If it does match, only test that one, otherwise we have to try
            # others.
            if any(cmd.lower().endswith(ext.lower()) for ext in pathext):
                files = [cmd]
            else:
                files = [cmd + ext for ext in pathext]
        else:
            # On other platforms you don't have things like PATHEXT to tell you
            # what file suffixes are executable, so just pass on cmd as-is.
            files = [cmd]

        seen = set()
        for dir in path:
            normdir = os.path.normcase(dir)
            if not normdir in seen:
                seen.add(normdir)
                for thefile in files:
                    name = os.path.join(dir, thefile)
                    if _access_check(name, mode):
                        return name
        return None


# ZipFile is a context manager in 2.7, but not in 2.6

from zipfile import ZipFile as BaseZipFile

if hasattr(BaseZipFile, '__enter__'):  # pragma: no cover
    ZipFile = BaseZipFile
else:  # pragma: no cover
    from zipfile import ZipExtFile as BaseZipExtFile

    class ZipExtFile(BaseZipExtFile):
        def __init__(self, base):
            self.__dict__.update(base.__dict__)

        def __enter__(self):
            return self

        def __exit__(self, *exc_info):
            self.close()
            # return None, so if an exception occurred, it will propagate

    class ZipFile(BaseZipFile):
        def __enter__(self):
            return self

        def __exit__(self, *exc_info):
            self.close()
            # return None, so if an exception occurred, it will propagate

        def open(self, *args, **kwargs):
            base = BaseZipFile.open(self, *args, **kwargs)
            return ZipExtFile(base)

try:
    from platform import python_implementation
except ImportError: # pragma: no cover
    def python_implementation():
        """Return a string identifying the Python implementation."""
        if 'PyPy' in sys.version:
            return 'PyPy'
        if os.name == 'java':
            return 'Jython'
        if sys.version.startswith('IronPython'):
            return 'IronPython'
        return 'CPython'

try:
    import sysconfig
except ImportError: # pragma: no cover
    from ._backport import sysconfig

try:
    callable = callable
except NameError:   # pragma: no cover
    from collections.abc import Callable

    def callable(obj):
        return isinstance(obj, Callable)


try:
    fsencode = os.fsencode
    fsdecode = os.fsdecode
except AttributeError:  # pragma: no cover
    # Issue #99: on some systems (e.g. containerised),
    # sys.getfilesystemencoding() returns None, and we need a real value,
    # so fall back to utf-8. From the CPython 2.7 docs relating to Unix and
    # sys.getfilesystemencoding(): the return value is "the user’s preference
    # according to the result of nl_langinfo(CODESET), or None if the
    # nl_langinfo(CODESET) failed."
    _fsencoding = sys.getfilesystemencoding() or 'utf-8'
    if _fsencoding == 'mbcs':
        _fserrors = 'strict'
    else:
        _fserrors = 'surrogateescape'

    def fsencode(filename):
        if isinstance(filename, bytes):
            return filename
        elif isinstance(filename, text_type):
            return filename.encode(_fsencoding, _fserrors)
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)

    def fsdecode(filename):
        if isinstance(filename, text_type):
            return filename
        elif isinstance(filename, bytes):
            return filename.decode(_fsencoding, _fserrors)
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)

try:
    from tokenize import detect_encoding
except ImportError: # pragma: no cover
    from codecs import BOM_UTF8, lookup
    import re

    cookie_re = re.compile(r"coding[:=]\s*([-\w.]+)")

    def _get_normal_name(orig_enc):
        """Imitates get_normal_name in tokenizer.c."""
        # Only care about the first 12 characters.
        enc = orig_enc[:12].lower().replace("_", "-")
        if enc == "utf-8" or enc.startswith("utf-8-"):
            return "utf-8"
        if enc in ("latin-1", "iso-8859-1", "iso-latin-1") or \
           enc.startswith(("latin-1-", "iso-8859-1-", "iso-latin-1-")):
            return "iso-8859-1"
        return orig_enc

    def detect_encoding(readline):
        """
        The detect_encoding() function is used to detect the encoding that should
        be used to decode a Python source file.  It requires one argument, readline,
        in the same way as the tokenize() generator.

        It will call readline a maximum of twice, and return the encoding used
        (as a string) and a list of any lines (left as bytes) it has read in.

        It detects the encoding from the presence of a utf-8 bom or an encoding
        cookie as specified in pep-0263.  If both a bom and a cookie are present,
        but disagree, a SyntaxError will be raised.  If the encoding cookie is an
        invalid charset, raise a SyntaxError.  Note that if a utf-8 bom is found,
        'utf-8-sig' is returned.

        If no encoding is specified, then the default of 'utf-8' will be returned.
        """
        try:
            filename = readline.__self__.name
        except AttributeError:
            filename = None
        bom_found = False
        encoding = None
        default = 'utf-8'
        def read_or_stop():
            try:
                return readline()
            except StopIteration:
                return b''

        def find_cookie(line):
            try:
                # Decode as UTF-8. Either the line is an encoding declaration,
                # in which case it should be pure ASCII, or it must be UTF-8
                # per default encoding.
                line_string = line.decode('utf-8')
            except UnicodeDecodeError:
                msg = "invalid or missing encoding declaration"
                if filename is not None:
                    msg = '{} for {!r}'.format(msg, filename)
                raise SyntaxError(msg)

            matches = cookie_re.findall(line_string)
            if not matches:
                return None
            encoding = _get_normal_name(matches[0])
            try:
                codec = lookup(encoding)
            except LookupError:
                # This behaviour mimics the Python interpreter
                if filename is None:
                    msg = "unknown encoding: " + encoding
                else:
                    msg = "unknown encoding for {!r}: {}".format(filename,
                            encoding)
                raise SyntaxError(msg)

            if bom_found:
                if codec.name != 'utf-8':
                    # This behaviour mimics the Python interpreter
                    if filename is None:
                        msg = 'encoding problem: utf-8'
                    else:
                        msg = 'encoding problem for {!r}: utf-8'.format(filename)
                    raise SyntaxError(msg)
                encoding += '-sig'
            return encoding

        first = read_or_stop()
        if first.startswith(BOM_UTF8):
            bom_found = True
            first = first[3:]
            default = 'utf-8-sig'
        if not first:
            return default, []

        encoding = find_cookie(first)
        if encoding:
            return encoding, [first]

        second = read_or_stop()
        if not second:
            return default, [first]

        encoding = find_cookie(second)
        if encoding:
            return encoding, [first, second]

        return default, [first, second]

# For converting & <-> &amp; etc.
try:
    from html import escape
except ImportError:
    from cgi import escape
if sys.version_info[:2] < (3, 4):
    unescape = HTMLParser().unescape
else:
    from html import unescape

try:
    from collections import ChainMap
except ImportError: # pragma: no cover
    from collections import MutableMapping

    try:
        from reprlib import recursive_repr as _recursive_repr
    except ImportError:
        def _recursive_repr(fillvalue='...'):
            '''
            Decorator to make a repr function return fillvalue for a recursive
            call
            '''

            def decorating_function(user_function):
                repr_running = set()

                def wrapper(self):
                    key = id(self), get_ident()
                    if key in repr_running:
                        return fillvalue
                    repr_running.add(key)
                    try:
                        result = user_function(self)
                    finally:
                        repr_running.discard(key)
                    return result

                # Can't use functools.wraps() here because of bootstrap issues
                wrapper.__module__ = getattr(user_function, '__module__')
                wrapper.__doc__ = getattr(user_function, '__doc__')
                wrapper.__name__ = getattr(user_function, '__name__')
                wrapper.__annotations__ = getattr(user_function, '__annotations__', {})
                return wrapper

            return decorating_function

    class ChainMap(MutableMapping):
        ''' A ChainMap groups multiple dicts (or other mappings) together
        to create a single, updateable view.

        The underlying mappings are stored in a list.  That list is public and can
        accessed or updated using the *maps* attribute.  There is no other state.

        Lookups search the underlying mappings successively until a key is found.
        In contrast, writes, updates, and deletions only operate on the first
        mapping.

        '''

        def __init__(self, *maps):
            '''Initialize a ChainMap by setting *maps* to the given mappings.
            If no mappings are provided, a single empty dictionary is used.

            '''
            self.maps = list(maps) or [{}]          # always at least one map

        def __missing__(self, key):
            raise KeyError(key)

        def __getitem__(self, key):
            for mapping in self.maps:
                try:
                    return mapping[key]             # can't use 'key in mapping' with defaultdict
                except KeyError:
                    pass
            return self.__missing__(key)            # support subclasses that define __missing__

        def get(self, key, default=None):
            return self[key] if key in self else default

        def __len__(self):
            return len(set().union(*self.maps))     # reuses stored hash values if possible

        def __iter__(self):
            return iter(set().union(*self.maps))

        def __contains__(self, key):
            return any(key in m for m in self.maps)

        def __bool__(self):
            return any(self.maps)

        @_recursive_repr()
        def __repr__(self):
            return '{0.__class__.__name__}({1})'.format(
                self, ', '.join(map(repr, self.maps)))

        @classmethod
        def fromkeys(cls, iterable, *args):
            'Create a ChainMap with a single dict created from the iterable.'
            return cls(dict.fromkeys(iterable, *args))

        def copy(self):
            'New ChainMap or subclass with a new copy of maps[0] and refs to maps[1:]'
            return self.__class__(self.maps[0].copy(), *self.maps[1:])

        __copy__ = copy

        def new_child(self):                        # like Django's Context.push()
            'New ChainMap with a new dict followed by all previous maps.'
            return self.__class__({}, *self.maps)

        @property
        def parents(self):                          # like Django's Context.pop()
            'New ChainMap from maps[1:].'
            return self.__class__(*self.maps[1:])

        def __setitem__(self, key, value):
            self.maps[0][key] = value

        def __delitem__(self, key):
            try:
                del self.maps[0][key]
            except KeyError:
                raise KeyError('Key not found in the first mapping: {!r}'.format(key))

        def popitem(self):
            'Remove and return an item pair from maps[0]. Raise KeyError is maps[0] is empty.'
            try:
                return self.maps[0].popitem()
            except KeyError:
                raise KeyError('No keys found in the first mapping.')

        def pop(self, key, *args):
            'Remove *key* from maps[0] and return its value. Raise KeyError if *key* not in maps[0].'
            try:
                return self.maps[0].pop(key, *args)
            except KeyError:
                raise KeyError('Key not found in the first mapping: {!r}'.format(key))

        def clear(self):
            'Clear maps[0], leaving maps[1:] intact.'
            self.maps[0].clear()

try:
    from importlib.util import cache_from_source  # Python >= 3.4
except ImportError:  # pragma: no cover
    try:
        from imp import cache_from_source
    except ImportError:  # pragma: no cover
        def cache_from_source(path, debug_override=None):
            assert path.endswith('.py')
            if debug_override is None:
                debug_override = __debug__
            if debug_override:
                suffix = 'c'
            else:
                suffix = 'o'
            return path + suffix

try:
    from collections import OrderedDict
except ImportError: # pragma: no cover
## {{{ http://code.activestate.com/recipes/576693/ (r9)
# Backport of OrderedDict() class that runs on Python 2.4, 2.5, 2.6, 2.7 and pypy.
# Passes Python2.7's test suite and incorporates all the latest updates.
    try:
        from thread import get_ident as _get_ident
    except ImportError:
        from dummy_thread import get_ident as _get_ident

    try:
        from _abcoll import KeysView, ValuesView, ItemsView
    except ImportError:
        pass


    class OrderedDict(dict):
        'Dictionary that remembers insertion order'
        # An inherited dict maps keys to values.
        # The inherited dict provides __getitem__, __len__, __contains__, and get.
        # The remaining methods are order-aware.
        # Big-O running times for all methods are the same as for regular dictionaries.

        # The internal self.__map dictionary maps keys to links in a doubly linked list.
        # The circular doubly linked list starts and ends with a sentinel element.
        # The sentinel element never gets deleted (this simplifies the algorithm).
        # Each link is stored as a list of length three:  [PREV, NEXT, KEY].

        def __init__(self, *args, **kwds):
            '''Initialize an ordered dictionary.  Signature is the same as for
            regular dictionaries, but keyword arguments are not recommended
            because their insertion order is arbitrary.

            '''
            if len(args) > 1:
                raise TypeError('expected at most 1 arguments, got %d' % len(args))
            try:
                self.__root
            except AttributeError:
                self.__root = root = []                     # sentinel node
                root[:] = [root, root, None]
                self.__map = {}
            self.__update(*args, **kwds)

        def __setitem__(self, key, value, dict_setitem=dict.__setitem__):
            'od.__setitem__(i, y) <==> od[i]=y'
            # Setting a new item creates a new link which goes at the end of the linked
            # list, and the inherited dictionary is updated with the new key/value pair.
            if key not in self:
                root = self.__root
                last = root[0]
                last[1] = root[0] = self.__map[key] = [last, root, key]
            dict_setitem(self, key, value)

        def __delitem__(self, key, dict_delitem=dict.__delitem__):
            'od.__delitem__(y) <==> del od[y]'
            # Deleting an existing item uses self.__map to find the link which is
            # then removed by updating the links in the predecessor and successor nodes.
            dict_delitem(self, key)
            link_prev, link_next, key = self.__map.pop(key)
            link_prev[1] = link_next
            link_next[0] = link_prev

        def __iter__(self):
            'od.__iter__() <==> iter(od)'
            root = self.__root
            curr = root[1]
            while curr is not root:
                yield curr[2]
                curr = curr[1]

        def __reversed__(self):
            'od.__reversed__() <==> reversed(od)'
            root = self.__root
            curr = root[0]
            while curr is not root:
                yield curr[2]
                curr = curr[0]

        def clear(self):
            'od.clear() -> None.  Remove all items from od.'
            try:
                for node in self.__map.itervalues():
                    del node[:]
                root = self.__root
                root[:] = [root, root, None]
                self.__map.clear()
            except AttributeError:
                pass
            dict.clear(self)

        def popitem(self, last=True):
            '''od.popitem() -> (k, v), return and remove a (key, value) pair.
            Pairs are returned in LIFO order if last is true or FIFO order if false.

            '''
            if not self:
                raise KeyError('dictionary is empty')
            root = self.__root
            if last:
                link = root[0]
                link_prev = link[0]
                link_prev[1] = root
                root[0] = link_prev
            else:
                link = root[1]
                link_next = link[1]
                root[1] = link_next
                link_next[0] = root
            key = link[2]
            del self.__map[key]
            value = dict.pop(self, key)
            return key, value

        # -- the following methods do not depend on the internal structure --

        def keys(self):
            'od.keys() -> list of keys in od'
            return list(self)

        def values(self):
            'od.values() -> list of values in od'
            return [self[key] for key in self]

        def items(self):
            'od.items() -> list of (key, value) pairs in od'
            return [(key, self[key]) for key in self]

        def iterkeys(self):
            'od.iterkeys() -> an iterator over the keys in od'
            return iter(self)

        def itervalues(self):
            'od.itervalues -> an iterator over the values in od'
            for k in self:
                yield self[k]

        def iteritems(self):
            'od.iteritems -> an iterator over the (key, value) items in od'
            for k in self:
                yield (k, self[k])

        def update(*args, **kwds):
            '''od.update(E, **F) -> None.  Update od from dict/iterable E and F.

            If E is a dict instance, does:           for k in E: od[k] = E[k]
            If E has a .keys() method, does:         for k in E.keys(): od[k] = E[k]
            Or if E is an iterable of items, does:   for k, v in E: od[k] = v
            In either case, this is followed by:     for k, v in F.items(): od[k] = v

            '''
            if len(args) > 2:
                raise TypeError('update() takes at most 2 positional '
                                'arguments (%d given)' % (len(args),))
            elif not args:
                raise TypeError('update() takes at least 1 argument (0 given)')
            self = args[0]
            # Make progressively weaker assumptions about "other"
            other = ()
            if len(args) == 2:
                other = args[1]
            if isinstance(other, dict):
                for key in other:
                    self[key] = other[key]
            elif hasattr(other, 'keys'):
                for key in other.keys():
                    self[key] = other[key]
            else:
                for key, value in other:
                    self[key] = value
            for key, value in kwds.items():
                self[key] = value

        __update = update  # let subclasses override update without breaking __init__

        __marker = object()

        def pop(self, key, default=__marker):
            '''od.pop(k[,d]) -> v, remove specified key and return the corresponding value.
            If key is not found, d is returned if given, otherwise KeyError is raised.

            '''
            if key in self:
                result = self[key]
                del self[key]
                return result
            if default is self.__marker:
                raise KeyError(key)
            return default

        def setdefault(self, key, default=None):
            'od.setdefault(k[,d]) -> od.get(k,d), also set od[k]=d if k not in od'
            if key in self:
                return self[key]
            self[key] = default
            return default

        def __repr__(self, _repr_running=None):
            'od.__repr__() <==> repr(od)'
            if not _repr_running: _repr_running = {}
            call_key = id(self), _get_ident()
            if call_key in _repr_running:
                return '...'
            _repr_running[call_key] = 1
            try:
                if not self:
                    return '%s()' % (self.__class__.__name__,)
                return '%s(%r)' % (self.__class__.__name__, self.items())
            finally:
                del _repr_running[call_key]

        def __reduce__(self):
            'Return state information for pickling'
            items = [[k, self[k]] for k in self]
            inst_dict = vars(self).copy()
            for k in vars(OrderedDict()):
                inst_dict.pop(k, None)
            if inst_dict:
                return (self.__class__, (items,), inst_dict)
            return self.__class__, (items,)

        def copy(self):
            'od.copy() -> a shallow copy of od'
            return self.__class__(self)

        @classmethod
        def fromkeys(cls, iterable, value=None):
            '''OD.fromkeys(S[, v]) -> New ordered dictionary with keys from S
            and values equal to v (which defaults to None).

            '''
            d = cls()
            for key in iterable:
                d[key] = value
            return d

        def __eq__(self, other):
            '''od.__eq__(y) <==> od==y.  Comparison to another OD is order-sensitive
            while comparison to a regular mapping is order-insensitive.

            '''
            if isinstance(other, OrderedDict):
                return len(self)==len(other) and self.items() == other.items()
            return dict.__eq__(self, other)

        def __ne__(self, other):
            return not self == other

        # -- the following methods are only used in Python 2.7 --

        def viewkeys(self):
            "od.viewkeys() -> a set-like object providing a view on od's keys"
            return KeysView(self)

        def viewvalues(self):
            "od.viewvalues() -> an object providing a view on od's values"
            return ValuesView(self)

        def viewitems(self):
            "od.viewitems() -> a set-like object providing a view on od's items"
            return ItemsView(self)

try:
    from logging.config import BaseConfigurator, valid_ident
except ImportError: # pragma: no cover
    IDENTIFIER = re.compile('^[a-z_][a-z0-9_]*$', re.I)


    def valid_ident(s):
        m = IDENTIFIER.match(s)
        if not m:
            raise ValueError('Not a valid Python identifier: %r' % s)
        return True


    # The ConvertingXXX classes are wrappers around standard Python containers,
    # and they serve to convert any suitable values in the container. The
    # conversion converts base dicts, lists and tuples to their wrapped
    # equivalents, whereas strings which match a conversion format are converted
    # appropriately.
    #
    # Each wrapper should have a configurator attribute holding the actual
    # configurator to use for conversion.

    class ConvertingDict(dict):
        """A converting dictionary wrapper."""

        def __getitem__(self, key):
            value = dict.__getitem__(self, key)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

        def get(self, key, default=None):
            value = dict.get(self, key, default)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

    def pop(self, key, default=None):
        value = dict.pop(self, key, default)
        result = self.configurator.convert(value)
        if value is not result:
            if type(result) in (ConvertingDict, ConvertingList,
                                ConvertingTuple):
                result.parent = self
                result.key = key
        return result

    class ConvertingList(list):
        """A converting list wrapper."""
        def __getitem__(self, key):
            value = list.__getitem__(self, key)
            result = self.configurator.convert(value)
            #If the converted value is different, save for next time
            if value is not result:
                self[key] = result
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

        def pop(self, idx=-1):
            value = list.pop(self, idx)
            result = self.configurator.convert(value)
            if value is not result:
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
            return result

    class ConvertingTuple(tuple):
        """A converting tuple wrapper."""
        def __getitem__(self, key):
            value = tuple.__getitem__(self, key)
            result = self.configurator.convert(value)
            if value is not result:
                if type(result) in (ConvertingDict, ConvertingList,
                                    ConvertingTuple):
                    result.parent = self
                    result.key = key
            return result

    class BaseConfigurator(object):
        """
        The configurator base class which defines some useful defaults.
        """

        CONVERT_PATTERN = re.compile(r'^(?P<prefix>[a-z]+)://(?P<suffix>.*)$')

        WORD_PATTERN = re.compile(r'^\s*(\w+)\s*')
        DOT_PATTERN = re.compile(r'^\.\s*(\w+)\s*')
        INDEX_PATTERN = re.compile(r'^\[\s*(\w+)\s*\]\s*')
        DIGIT_PATTERN = re.compile(r'^\d+$')

        value_converters = {
            'ext' : 'ext_convert',
            'cfg' : 'cfg_convert',
        }

        # We might want to use a different one, e.g. importlib
        importer = staticmethod(__import__)

        def __init__(self, config):
            self.config = ConvertingDict(config)
            self.config.configurator = self

        def resolve(self, s):
            """
            Resolve strings to objects using standard import and attribute
            syntax.
            """
            name = s.split('.')
            used = name.pop(0)
            try:
                found = self.importer(used)
                for frag in name:
                    used += '.' + frag
                    try:
                        found = getattr(found, frag)
                    except AttributeError:
                        self.importer(used)
                        found = getattr(found, frag)
                return found
            except ImportError:
                e, tb = sys.exc_info()[1:]
                v = ValueError('Cannot resolve %r: %s' % (s, e))
                v.__cause__, v.__traceback__ = e, tb
                raise v

        def ext_convert(self, value):
            """Default converter for the ext:// protocol."""
            return self.resolve(value)

        def cfg_convert(self, value):
            """Default converter for the cfg:// protocol."""
            rest = value
            m = self.WORD_PATTERN.match(rest)
            if m is None:
                raise ValueError("Unable to convert %r" % value)
            else:
                rest = rest[m.end():]
                d = self.config[m.groups()[0]]
                #print d, rest
                while rest:
                    m = self.DOT_PATTERN.match(rest)
                    if m:
                        d = d[m.groups()[0]]
                    else:
                        m = self.INDEX_PATTERN.match(rest)
                        if m:
                            idx = m.groups()[0]
                            if not self.DIGIT_PATTERN.match(idx):
                                d = d[idx]
                            else:
                                try:
                                    n = int(idx) # try as number first (most likely)
                                    d = d[n]
                                except TypeError:
                                    d = d[idx]
                    if m:
                        rest = rest[m.end():]
                    else:
                        raise ValueError('Unable to convert '
                                         '%r at %r' % (value, rest))
            #rest should be empty
            return d

        def convert(self, value):
            """
            Convert values to an appropriate type. dicts, lists and tuples are
            replaced by their converting alternatives. Strings are checked to
            see if they have a conversion format and are converted if they do.
            """
            if not isinstance(value, ConvertingDict) and isinstance(value, dict):
                value = ConvertingDict(value)
                value.configurator = self
            elif not isinstance(value, ConvertingList) and isinstance(value, list):
                value = ConvertingList(value)
                value.configurator = self
            elif not isinstance(value, ConvertingTuple) and\
                     isinstance(value, tuple):
                value = ConvertingTuple(value)
                value.configurator = self
            elif isinstance(value, string_types):
                m = self.CONVERT_PATTERN.match(value)
                if m:
                    d = m.groupdict()
                    prefix = d['prefix']
                    converter = self.value_converters.get(prefix, None)
                    if converter:
                        suffix = d['suffix']
                        converter = getattr(self, converter)
                        value = converter(suffix)
            return value

        def configure_custom(self, config):
            """Configure an object with a user-supplied factory."""
            c = config.pop('()')
            if not callable(c):
                c = self.resolve(c)
            props = config.pop('.', None)
            # Check for valid identifiers
            kwargs = dict([(k, config[k]) for k in config if valid_ident(k)])
            result = c(**kwargs)
            if props:
                for name, value in props.items():
                    setattr(result, name, value)
            return result

        def as_tuple(self, value):
            """Utility function which converts lists to tuples."""
            if isinstance(value, list):
                value = tuple(value)
            return value
site-packages/pip/_vendor/distlib/database.py000064400000143563150564745010015303 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2017 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""PEP 376 implementation."""

from __future__ import unicode_literals

import base64
import codecs
import contextlib
import hashlib
import logging
import os
import posixpath
import sys
import zipimport

from . import DistlibException, resources
from .compat import StringIO
from .version import get_scheme, UnsupportedVersionError
from .metadata import (Metadata, METADATA_FILENAME, WHEEL_METADATA_FILENAME,
                       LEGACY_METADATA_FILENAME)
from .util import (parse_requirement, cached_property, parse_name_and_version,
                   read_exports, write_exports, CSVReader, CSVWriter)


__all__ = ['Distribution', 'BaseInstalledDistribution',
           'InstalledDistribution', 'EggInfoDistribution',
           'DistributionPath']


logger = logging.getLogger(__name__)

EXPORTS_FILENAME = 'pydist-exports.json'
COMMANDS_FILENAME = 'pydist-commands.json'

DIST_FILES = ('INSTALLER', METADATA_FILENAME, 'RECORD', 'REQUESTED',
              'RESOURCES', EXPORTS_FILENAME, 'SHARED')

DISTINFO_EXT = '.dist-info'


class _Cache(object):
    """
    A simple cache mapping names and .dist-info paths to distributions
    """
    def __init__(self):
        """
        Initialise an instance. There is normally one for each DistributionPath.
        """
        self.name = {}
        self.path = {}
        self.generated = False

    def clear(self):
        """
        Clear the cache, setting it to its initial state.
        """
        self.name.clear()
        self.path.clear()
        self.generated = False

    def add(self, dist):
        """
        Add a distribution to the cache.
        :param dist: The distribution to add.
        """
        if dist.path not in self.path:
            self.path[dist.path] = dist
            self.name.setdefault(dist.key, []).append(dist)


class DistributionPath(object):
    """
    Represents a set of distributions installed on a path (typically sys.path).
    """
    def __init__(self, path=None, include_egg=False):
        """
        Create an instance from a path, optionally including legacy (distutils/
        setuptools/distribute) distributions.
        :param path: The path to use, as a list of directories. If not specified,
                     sys.path is used.
        :param include_egg: If True, this instance will look for and return legacy
                            distributions as well as those based on PEP 376.
        """
        if path is None:
            path = sys.path
        self.path = path
        self._include_dist = True
        self._include_egg = include_egg

        self._cache = _Cache()
        self._cache_egg = _Cache()
        self._cache_enabled = True
        self._scheme = get_scheme('default')

    def _get_cache_enabled(self):
        return self._cache_enabled

    def _set_cache_enabled(self, value):
        self._cache_enabled = value

    cache_enabled = property(_get_cache_enabled, _set_cache_enabled)

    def clear_cache(self):
        """
        Clears the internal cache.
        """
        self._cache.clear()
        self._cache_egg.clear()


    def _yield_distributions(self):
        """
        Yield .dist-info and/or .egg(-info) distributions.
        """
        # We need to check if we've seen some resources already, because on
        # some Linux systems (e.g. some Debian/Ubuntu variants) there are
        # symlinks which alias other files in the environment.
        seen = set()
        for path in self.path:
            finder = resources.finder_for_path(path)
            if finder is None:
                continue
            r = finder.find('')
            if not r or not r.is_container:
                continue
            rset = sorted(r.resources)
            for entry in rset:
                r = finder.find(entry)
                if not r or r.path in seen:
                    continue
                if self._include_dist and entry.endswith(DISTINFO_EXT):
                    possible_filenames = [METADATA_FILENAME,
                                          WHEEL_METADATA_FILENAME,
                                          LEGACY_METADATA_FILENAME]
                    for metadata_filename in possible_filenames:
                        metadata_path = posixpath.join(entry, metadata_filename)
                        pydist = finder.find(metadata_path)
                        if pydist:
                            break
                    else:
                        continue

                    with contextlib.closing(pydist.as_stream()) as stream:
                        metadata = Metadata(fileobj=stream, scheme='legacy')
                    logger.debug('Found %s', r.path)
                    seen.add(r.path)
                    yield new_dist_class(r.path, metadata=metadata,
                                         env=self)
                elif self._include_egg and entry.endswith(('.egg-info',
                                                          '.egg')):
                    logger.debug('Found %s', r.path)
                    seen.add(r.path)
                    yield old_dist_class(r.path, self)

    def _generate_cache(self):
        """
        Scan the path for distributions and populate the cache with
        those that are found.
        """
        gen_dist = not self._cache.generated
        gen_egg = self._include_egg and not self._cache_egg.generated
        if gen_dist or gen_egg:
            for dist in self._yield_distributions():
                if isinstance(dist, InstalledDistribution):
                    self._cache.add(dist)
                else:
                    self._cache_egg.add(dist)

            if gen_dist:
                self._cache.generated = True
            if gen_egg:
                self._cache_egg.generated = True

    @classmethod
    def distinfo_dirname(cls, name, version):
        """
        The *name* and *version* parameters are converted into their
        filename-escaped form, i.e. any ``'-'`` characters are replaced
        with ``'_'`` other than the one in ``'dist-info'`` and the one
        separating the name from the version number.

        :parameter name: is converted to a standard distribution name by replacing
                         any runs of non- alphanumeric characters with a single
                         ``'-'``.
        :type name: string
        :parameter version: is converted to a standard version string. Spaces
                            become dots, and all other non-alphanumeric characters
                            (except dots) become dashes, with runs of multiple
                            dashes condensed to a single dash.
        :type version: string
        :returns: directory name
        :rtype: string"""
        name = name.replace('-', '_')
        return '-'.join([name, version]) + DISTINFO_EXT

    def get_distributions(self):
        """
        Provides an iterator that looks for distributions and returns
        :class:`InstalledDistribution` or
        :class:`EggInfoDistribution` instances for each one of them.

        :rtype: iterator of :class:`InstalledDistribution` and
                :class:`EggInfoDistribution` instances
        """
        if not self._cache_enabled:
            for dist in self._yield_distributions():
                yield dist
        else:
            self._generate_cache()

            for dist in self._cache.path.values():
                yield dist

            if self._include_egg:
                for dist in self._cache_egg.path.values():
                    yield dist

    def get_distribution(self, name):
        """
        Looks for a named distribution on the path.

        This function only returns the first result found, as no more than one
        value is expected. If nothing is found, ``None`` is returned.

        :rtype: :class:`InstalledDistribution`, :class:`EggInfoDistribution`
                or ``None``
        """
        result = None
        name = name.lower()
        if not self._cache_enabled:
            for dist in self._yield_distributions():
                if dist.key == name:
                    result = dist
                    break
        else:
            self._generate_cache()

            if name in self._cache.name:
                result = self._cache.name[name][0]
            elif self._include_egg and name in self._cache_egg.name:
                result = self._cache_egg.name[name][0]
        return result

    def provides_distribution(self, name, version=None):
        """
        Iterates over all distributions to find which distributions provide *name*.
        If a *version* is provided, it will be used to filter the results.

        This function only returns the first result found, since no more than
        one values are expected. If the directory is not found, returns ``None``.

        :parameter version: a version specifier that indicates the version
                            required, conforming to the format in ``PEP-345``

        :type name: string
        :type version: string
        """
        matcher = None
        if version is not None:
            try:
                matcher = self._scheme.matcher('%s (%s)' % (name, version))
            except ValueError:
                raise DistlibException('invalid name or version: %r, %r' %
                                      (name, version))

        for dist in self.get_distributions():
            # We hit a problem on Travis where enum34 was installed and doesn't
            # have a provides attribute ...
            if not hasattr(dist, 'provides'):
                logger.debug('No "provides": %s', dist)
            else:
                provided = dist.provides

                for p in provided:
                    p_name, p_ver = parse_name_and_version(p)
                    if matcher is None:
                        if p_name == name:
                            yield dist
                            break
                    else:
                        if p_name == name and matcher.match(p_ver):
                            yield dist
                            break

    def get_file_path(self, name, relative_path):
        """
        Return the path to a resource file.
        """
        dist = self.get_distribution(name)
        if dist is None:
            raise LookupError('no distribution named %r found' % name)
        return dist.get_resource_path(relative_path)

    def get_exported_entries(self, category, name=None):
        """
        Return all of the exported entries in a particular category.

        :param category: The category to search for entries.
        :param name: If specified, only entries with that name are returned.
        """
        for dist in self.get_distributions():
            r = dist.exports
            if category in r:
                d = r[category]
                if name is not None:
                    if name in d:
                        yield d[name]
                else:
                    for v in d.values():
                        yield v


class Distribution(object):
    """
    A base class for distributions, whether installed or from indexes.
    Either way, it must have some metadata, so that's all that's needed
    for construction.
    """

    build_time_dependency = False
    """
    Set to True if it's known to be only a build-time dependency (i.e.
    not needed after installation).
    """

    requested = False
    """A boolean that indicates whether the ``REQUESTED`` metadata file is
    present (in other words, whether the package was installed by user
    request or it was installed as a dependency)."""

    def __init__(self, metadata):
        """
        Initialise an instance.
        :param metadata: The instance of :class:`Metadata` describing this
        distribution.
        """
        self.metadata = metadata
        self.name = metadata.name
        self.key = self.name.lower()    # for case-insensitive comparisons
        self.version = metadata.version
        self.locator = None
        self.digest = None
        self.extras = None      # additional features requested
        self.context = None     # environment marker overrides
        self.download_urls = set()
        self.digests = {}

    @property
    def source_url(self):
        """
        The source archive download URL for this distribution.
        """
        return self.metadata.source_url

    download_url = source_url   # Backward compatibility

    @property
    def name_and_version(self):
        """
        A utility property which displays the name and version in parentheses.
        """
        return '%s (%s)' % (self.name, self.version)

    @property
    def provides(self):
        """
        A set of distribution names and versions provided by this distribution.
        :return: A set of "name (version)" strings.
        """
        plist = self.metadata.provides
        s = '%s (%s)' % (self.name, self.version)
        if s not in plist:
            plist.append(s)
        return plist

    def _get_requirements(self, req_attr):
        md = self.metadata
        logger.debug('Getting requirements from metadata %r', md.todict())
        reqts = getattr(md, req_attr)
        return set(md.get_requirements(reqts, extras=self.extras,
                                       env=self.context))

    @property
    def run_requires(self):
        return self._get_requirements('run_requires')

    @property
    def meta_requires(self):
        return self._get_requirements('meta_requires')

    @property
    def build_requires(self):
        return self._get_requirements('build_requires')

    @property
    def test_requires(self):
        return self._get_requirements('test_requires')

    @property
    def dev_requires(self):
        return self._get_requirements('dev_requires')

    def matches_requirement(self, req):
        """
        Say if this instance matches (fulfills) a requirement.
        :param req: The requirement to match.
        :rtype req: str
        :return: True if it matches, else False.
        """
        # Requirement may contain extras - parse to lose those
        # from what's passed to the matcher
        r = parse_requirement(req)
        scheme = get_scheme(self.metadata.scheme)
        try:
            matcher = scheme.matcher(r.requirement)
        except UnsupportedVersionError:
            # XXX compat-mode if cannot read the version
            logger.warning('could not read version %r - using name only',
                           req)
            name = req.split()[0]
            matcher = scheme.matcher(name)

        name = matcher.key   # case-insensitive

        result = False
        for p in self.provides:
            p_name, p_ver = parse_name_and_version(p)
            if p_name != name:
                continue
            try:
                result = matcher.match(p_ver)
                break
            except UnsupportedVersionError:
                pass
        return result

    def __repr__(self):
        """
        Return a textual representation of this instance,
        """
        if self.source_url:
            suffix = ' [%s]' % self.source_url
        else:
            suffix = ''
        return '<Distribution %s (%s)%s>' % (self.name, self.version, suffix)

    def __eq__(self, other):
        """
        See if this distribution is the same as another.
        :param other: The distribution to compare with. To be equal to one
                      another. distributions must have the same type, name,
                      version and source_url.
        :return: True if it is the same, else False.
        """
        if type(other) is not type(self):
            result = False
        else:
            result = (self.name == other.name and
                      self.version == other.version and
                      self.source_url == other.source_url)
        return result

    def __hash__(self):
        """
        Compute hash in a way which matches the equality test.
        """
        return hash(self.name) + hash(self.version) + hash(self.source_url)


class BaseInstalledDistribution(Distribution):
    """
    This is the base class for installed distributions (whether PEP 376 or
    legacy).
    """

    hasher = None

    def __init__(self, metadata, path, env=None):
        """
        Initialise an instance.
        :param metadata: An instance of :class:`Metadata` which describes the
                         distribution. This will normally have been initialised
                         from a metadata file in the ``path``.
        :param path:     The path of the ``.dist-info`` or ``.egg-info``
                         directory for the distribution.
        :param env:      This is normally the :class:`DistributionPath`
                         instance where this distribution was found.
        """
        super(BaseInstalledDistribution, self).__init__(metadata)
        self.path = path
        self.dist_path = env

    def get_hash(self, data, hasher=None):
        """
        Get the hash of some data, using a particular hash algorithm, if
        specified.

        :param data: The data to be hashed.
        :type data: bytes
        :param hasher: The name of a hash implementation, supported by hashlib,
                       or ``None``. Examples of valid values are ``'sha1'``,
                       ``'sha224'``, ``'sha384'``, '``sha256'``, ``'md5'`` and
                       ``'sha512'``. If no hasher is specified, the ``hasher``
                       attribute of the :class:`InstalledDistribution` instance
                       is used. If the hasher is determined to be ``None``, MD5
                       is used as the hashing algorithm.
        :returns: The hash of the data. If a hasher was explicitly specified,
                  the returned hash will be prefixed with the specified hasher
                  followed by '='.
        :rtype: str
        """
        if hasher is None:
            hasher = self.hasher
        if hasher is None:
            hasher = hashlib.md5
            prefix = ''
        else:
            hasher = getattr(hashlib, hasher)
            prefix = '%s=' % self.hasher
        digest = hasher(data).digest()
        digest = base64.urlsafe_b64encode(digest).rstrip(b'=').decode('ascii')
        return '%s%s' % (prefix, digest)


class InstalledDistribution(BaseInstalledDistribution):
    """
    Created with the *path* of the ``.dist-info`` directory provided to the
    constructor. It reads the metadata contained in ``pydist.json`` when it is
    instantiated., or uses a passed in Metadata instance (useful for when
    dry-run mode is being used).
    """

    hasher = 'sha256'

    def __init__(self, path, metadata=None, env=None):
        self.modules = []
        self.finder = finder = resources.finder_for_path(path)
        if finder is None:
            raise ValueError('finder unavailable for %s' % path)
        if env and env._cache_enabled and path in env._cache.path:
            metadata = env._cache.path[path].metadata
        elif metadata is None:
            r = finder.find(METADATA_FILENAME)
            # Temporary - for Wheel 0.23 support
            if r is None:
                r = finder.find(WHEEL_METADATA_FILENAME)
            # Temporary - for legacy support
            if r is None:
                r = finder.find(LEGACY_METADATA_FILENAME)
            if r is None:
                raise ValueError('no %s found in %s' % (METADATA_FILENAME,
                                                        path))
            with contextlib.closing(r.as_stream()) as stream:
                metadata = Metadata(fileobj=stream, scheme='legacy')

        super(InstalledDistribution, self).__init__(metadata, path, env)

        if env and env._cache_enabled:
            env._cache.add(self)

        r = finder.find('REQUESTED')
        self.requested = r is not None
        p  = os.path.join(path, 'top_level.txt')
        if os.path.exists(p):
            with open(p, 'rb') as f:
                data = f.read().decode('utf-8')
            self.modules = data.splitlines()

    def __repr__(self):
        return '<InstalledDistribution %r %s at %r>' % (
            self.name, self.version, self.path)

    def __str__(self):
        return "%s %s" % (self.name, self.version)

    def _get_records(self):
        """
        Get the list of installed files for the distribution
        :return: A list of tuples of path, hash and size. Note that hash and
                 size might be ``None`` for some entries. The path is exactly
                 as stored in the file (which is as in PEP 376).
        """
        results = []
        r = self.get_distinfo_resource('RECORD')
        with contextlib.closing(r.as_stream()) as stream:
            with CSVReader(stream=stream) as record_reader:
                # Base location is parent dir of .dist-info dir
                #base_location = os.path.dirname(self.path)
                #base_location = os.path.abspath(base_location)
                for row in record_reader:
                    missing = [None for i in range(len(row), 3)]
                    path, checksum, size = row + missing
                    #if not os.path.isabs(path):
                    #    path = path.replace('/', os.sep)
                    #    path = os.path.join(base_location, path)
                    results.append((path, checksum, size))
        return results

    @cached_property
    def exports(self):
        """
        Return the information exported by this distribution.
        :return: A dictionary of exports, mapping an export category to a dict
                 of :class:`ExportEntry` instances describing the individual
                 export entries, and keyed by name.
        """
        result = {}
        r = self.get_distinfo_resource(EXPORTS_FILENAME)
        if r:
            result = self.read_exports()
        return result

    def read_exports(self):
        """
        Read exports data from a file in .ini format.

        :return: A dictionary of exports, mapping an export category to a list
                 of :class:`ExportEntry` instances describing the individual
                 export entries.
        """
        result = {}
        r = self.get_distinfo_resource(EXPORTS_FILENAME)
        if r:
            with contextlib.closing(r.as_stream()) as stream:
                result = read_exports(stream)
        return result

    def write_exports(self, exports):
        """
        Write a dictionary of exports to a file in .ini format.
        :param exports: A dictionary of exports, mapping an export category to
                        a list of :class:`ExportEntry` instances describing the
                        individual export entries.
        """
        rf = self.get_distinfo_file(EXPORTS_FILENAME)
        with open(rf, 'w') as f:
            write_exports(exports, f)

    def get_resource_path(self, relative_path):
        """
        NOTE: This API may change in the future.

        Return the absolute path to a resource file with the given relative
        path.

        :param relative_path: The path, relative to .dist-info, of the resource
                              of interest.
        :return: The absolute path where the resource is to be found.
        """
        r = self.get_distinfo_resource('RESOURCES')
        with contextlib.closing(r.as_stream()) as stream:
            with CSVReader(stream=stream) as resources_reader:
                for relative, destination in resources_reader:
                    if relative == relative_path:
                        return destination
        raise KeyError('no resource file with relative path %r '
                       'is installed' % relative_path)

    def list_installed_files(self):
        """
        Iterates over the ``RECORD`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: iterator of (path, hash, size)
        """
        for result in self._get_records():
            yield result

    def write_installed_files(self, paths, prefix, dry_run=False):
        """
        Writes the ``RECORD`` file, using the ``paths`` iterable passed in. Any
        existing ``RECORD`` file is silently overwritten.

        prefix is used to determine when to write absolute paths.
        """
        prefix = os.path.join(prefix, '')
        base = os.path.dirname(self.path)
        base_under_prefix = base.startswith(prefix)
        base = os.path.join(base, '')
        record_path = self.get_distinfo_file('RECORD')
        logger.info('creating %s', record_path)
        if dry_run:
            return None
        with CSVWriter(record_path) as writer:
            for path in paths:
                if os.path.isdir(path) or path.endswith(('.pyc', '.pyo')):
                    # do not put size and hash, as in PEP-376
                    hash_value = size = ''
                else:
                    size = '%d' % os.path.getsize(path)
                    with open(path, 'rb') as fp:
                        hash_value = self.get_hash(fp.read())
                if path.startswith(base) or (base_under_prefix and
                                             path.startswith(prefix)):
                    path = os.path.relpath(path, base)
                writer.writerow((path, hash_value, size))

            # add the RECORD file itself
            if record_path.startswith(base):
                record_path = os.path.relpath(record_path, base)
            writer.writerow((record_path, '', ''))
        return record_path

    def check_installed_files(self):
        """
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        """
        mismatches = []
        base = os.path.dirname(self.path)
        record_path = self.get_distinfo_file('RECORD')
        for path, hash_value, size in self.list_installed_files():
            if not os.path.isabs(path):
                path = os.path.join(base, path)
            if path == record_path:
                continue
            if not os.path.exists(path):
                mismatches.append((path, 'exists', True, False))
            elif os.path.isfile(path):
                actual_size = str(os.path.getsize(path))
                if size and actual_size != size:
                    mismatches.append((path, 'size', size, actual_size))
                elif hash_value:
                    if '=' in hash_value:
                        hasher = hash_value.split('=', 1)[0]
                    else:
                        hasher = None

                    with open(path, 'rb') as f:
                        actual_hash = self.get_hash(f.read(), hasher)
                        if actual_hash != hash_value:
                            mismatches.append((path, 'hash', hash_value, actual_hash))
        return mismatches

    @cached_property
    def shared_locations(self):
        """
        A dictionary of shared locations whose keys are in the set 'prefix',
        'purelib', 'platlib', 'scripts', 'headers', 'data' and 'namespace'.
        The corresponding value is the absolute path of that category for
        this distribution, and takes into account any paths selected by the
        user at installation time (e.g. via command-line arguments). In the
        case of the 'namespace' key, this would be a list of absolute paths
        for the roots of namespace packages in this distribution.

        The first time this property is accessed, the relevant information is
        read from the SHARED file in the .dist-info directory.
        """
        result = {}
        shared_path = os.path.join(self.path, 'SHARED')
        if os.path.isfile(shared_path):
            with codecs.open(shared_path, 'r', encoding='utf-8') as f:
                lines = f.read().splitlines()
            for line in lines:
                key, value = line.split('=', 1)
                if key == 'namespace':
                    result.setdefault(key, []).append(value)
                else:
                    result[key] = value
        return result

    def write_shared_locations(self, paths, dry_run=False):
        """
        Write shared location information to the SHARED file in .dist-info.
        :param paths: A dictionary as described in the documentation for
        :meth:`shared_locations`.
        :param dry_run: If True, the action is logged but no file is actually
                        written.
        :return: The path of the file written to.
        """
        shared_path = os.path.join(self.path, 'SHARED')
        logger.info('creating %s', shared_path)
        if dry_run:
            return None
        lines = []
        for key in ('prefix', 'lib', 'headers', 'scripts', 'data'):
            path = paths[key]
            if os.path.isdir(paths[key]):
                lines.append('%s=%s' % (key,  path))
        for ns in paths.get('namespace', ()):
            lines.append('namespace=%s' % ns)

        with codecs.open(shared_path, 'w', encoding='utf-8') as f:
            f.write('\n'.join(lines))
        return shared_path

    def get_distinfo_resource(self, path):
        if path not in DIST_FILES:
            raise DistlibException('invalid path for a dist-info file: '
                                   '%r at %r' % (path, self.path))
        finder = resources.finder_for_path(self.path)
        if finder is None:
            raise DistlibException('Unable to get a finder for %s' % self.path)
        return finder.find(path)

    def get_distinfo_file(self, path):
        """
        Returns a path located under the ``.dist-info`` directory. Returns a
        string representing the path.

        :parameter path: a ``'/'``-separated path relative to the
                         ``.dist-info`` directory or an absolute path;
                         If *path* is an absolute path and doesn't start
                         with the ``.dist-info`` directory path,
                         a :class:`DistlibException` is raised
        :type path: str
        :rtype: str
        """
        # Check if it is an absolute path  # XXX use relpath, add tests
        if path.find(os.sep) >= 0:
            # it's an absolute path?
            distinfo_dirname, path = path.split(os.sep)[-2:]
            if distinfo_dirname != self.path.split(os.sep)[-1]:
                raise DistlibException(
                    'dist-info file %r does not belong to the %r %s '
                    'distribution' % (path, self.name, self.version))

        # The file must be relative
        if path not in DIST_FILES:
            raise DistlibException('invalid path for a dist-info file: '
                                   '%r at %r' % (path, self.path))

        return os.path.join(self.path, path)

    def list_distinfo_files(self):
        """
        Iterates over the ``RECORD`` entries and returns paths for each line if
        the path is pointing to a file located in the ``.dist-info`` directory
        or one of its subdirectories.

        :returns: iterator of paths
        """
        base = os.path.dirname(self.path)
        for path, checksum, size in self._get_records():
            # XXX add separator or use real relpath algo
            if not os.path.isabs(path):
                path = os.path.join(base, path)
            if path.startswith(self.path):
                yield path

    def __eq__(self, other):
        return (isinstance(other, InstalledDistribution) and
                self.path == other.path)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    __hash__ = object.__hash__


class EggInfoDistribution(BaseInstalledDistribution):
    """Created with the *path* of the ``.egg-info`` directory or file provided
    to the constructor. It reads the metadata contained in the file itself, or
    if the given path happens to be a directory, the metadata is read from the
    file ``PKG-INFO`` under that directory."""

    requested = True    # as we have no way of knowing, assume it was
    shared_locations = {}

    def __init__(self, path, env=None):
        def set_name_and_version(s, n, v):
            s.name = n
            s.key = n.lower()   # for case-insensitive comparisons
            s.version = v

        self.path = path
        self.dist_path = env
        if env and env._cache_enabled and path in env._cache_egg.path:
            metadata = env._cache_egg.path[path].metadata
            set_name_and_version(self, metadata.name, metadata.version)
        else:
            metadata = self._get_metadata(path)

            # Need to be set before caching
            set_name_and_version(self, metadata.name, metadata.version)

            if env and env._cache_enabled:
                env._cache_egg.add(self)
        super(EggInfoDistribution, self).__init__(metadata, path, env)

    def _get_metadata(self, path):
        requires = None

        def parse_requires_data(data):
            """Create a list of dependencies from a requires.txt file.

            *data*: the contents of a setuptools-produced requires.txt file.
            """
            reqs = []
            lines = data.splitlines()
            for line in lines:
                line = line.strip()
                if line.startswith('['):
                    logger.warning('Unexpected line: quitting requirement scan: %r',
                                   line)
                    break
                r = parse_requirement(line)
                if not r:
                    logger.warning('Not recognised as a requirement: %r', line)
                    continue
                if r.extras:
                    logger.warning('extra requirements in requires.txt are '
                                   'not supported')
                if not r.constraints:
                    reqs.append(r.name)
                else:
                    cons = ', '.join('%s%s' % c for c in r.constraints)
                    reqs.append('%s (%s)' % (r.name, cons))
            return reqs

        def parse_requires_path(req_path):
            """Create a list of dependencies from a requires.txt file.

            *req_path*: the path to a setuptools-produced requires.txt file.
            """

            reqs = []
            try:
                with codecs.open(req_path, 'r', 'utf-8') as fp:
                    reqs = parse_requires_data(fp.read())
            except IOError:
                pass
            return reqs

        tl_path = tl_data = None
        if path.endswith('.egg'):
            if os.path.isdir(path):
                p = os.path.join(path, 'EGG-INFO')
                meta_path = os.path.join(p, 'PKG-INFO')
                metadata = Metadata(path=meta_path, scheme='legacy')
                req_path = os.path.join(p, 'requires.txt')
                tl_path = os.path.join(p, 'top_level.txt')
                requires = parse_requires_path(req_path)
            else:
                # FIXME handle the case where zipfile is not available
                zipf = zipimport.zipimporter(path)
                fileobj = StringIO(
                    zipf.get_data('EGG-INFO/PKG-INFO').decode('utf8'))
                metadata = Metadata(fileobj=fileobj, scheme='legacy')
                try:
                    data = zipf.get_data('EGG-INFO/requires.txt')
                    tl_data = zipf.get_data('EGG-INFO/top_level.txt').decode('utf-8')
                    requires = parse_requires_data(data.decode('utf-8'))
                except IOError:
                    requires = None
        elif path.endswith('.egg-info'):
            if os.path.isdir(path):
                req_path = os.path.join(path, 'requires.txt')
                requires = parse_requires_path(req_path)
                path = os.path.join(path, 'PKG-INFO')
                tl_path = os.path.join(path, 'top_level.txt')
            metadata = Metadata(path=path, scheme='legacy')
        else:
            raise DistlibException('path must end with .egg-info or .egg, '
                                   'got %r' % path)

        if requires:
            metadata.add_requirements(requires)
        # look for top-level modules in top_level.txt, if present
        if tl_data is None:
            if tl_path is not None and os.path.exists(tl_path):
                with open(tl_path, 'rb') as f:
                    tl_data = f.read().decode('utf-8')
        if not tl_data:
            tl_data = []
        else:
            tl_data = tl_data.splitlines()
        self.modules = tl_data
        return metadata

    def __repr__(self):
        return '<EggInfoDistribution %r %s at %r>' % (
            self.name, self.version, self.path)

    def __str__(self):
        return "%s %s" % (self.name, self.version)

    def check_installed_files(self):
        """
        Checks that the hashes and sizes of the files in ``RECORD`` are
        matched by the files themselves. Returns a (possibly empty) list of
        mismatches. Each entry in the mismatch list will be a tuple consisting
        of the path, 'exists', 'size' or 'hash' according to what didn't match
        (existence is checked first, then size, then hash), the expected
        value and the actual value.
        """
        mismatches = []
        record_path = os.path.join(self.path, 'installed-files.txt')
        if os.path.exists(record_path):
            for path, _, _ in self.list_installed_files():
                if path == record_path:
                    continue
                if not os.path.exists(path):
                    mismatches.append((path, 'exists', True, False))
        return mismatches

    def list_installed_files(self):
        """
        Iterates over the ``installed-files.txt`` entries and returns a tuple
        ``(path, hash, size)`` for each line.

        :returns: a list of (path, hash, size)
        """

        def _md5(path):
            f = open(path, 'rb')
            try:
                content = f.read()
            finally:
                f.close()
            return hashlib.md5(content).hexdigest()

        def _size(path):
            return os.stat(path).st_size

        record_path = os.path.join(self.path, 'installed-files.txt')
        result = []
        if os.path.exists(record_path):
            with codecs.open(record_path, 'r', encoding='utf-8') as f:
                for line in f:
                    line = line.strip()
                    p = os.path.normpath(os.path.join(self.path, line))
                    # "./" is present as a marker between installed files
                    # and installation metadata files
                    if not os.path.exists(p):
                        logger.warning('Non-existent file: %s', p)
                        if p.endswith(('.pyc', '.pyo')):
                            continue
                        #otherwise fall through and fail
                    if not os.path.isdir(p):
                        result.append((p, _md5(p), _size(p)))
            result.append((record_path, None, None))
        return result

    def list_distinfo_files(self, absolute=False):
        """
        Iterates over the ``installed-files.txt`` entries and returns paths for
        each line if the path is pointing to a file located in the
        ``.egg-info`` directory or one of its subdirectories.

        :parameter absolute: If *absolute* is ``True``, each returned path is
                          transformed into a local absolute path. Otherwise the
                          raw value from ``installed-files.txt`` is returned.
        :type absolute: boolean
        :returns: iterator of paths
        """
        record_path = os.path.join(self.path, 'installed-files.txt')
        if os.path.exists(record_path):
            skip = True
            with codecs.open(record_path, 'r', encoding='utf-8') as f:
                for line in f:
                    line = line.strip()
                    if line == './':
                        skip = False
                        continue
                    if not skip:
                        p = os.path.normpath(os.path.join(self.path, line))
                        if p.startswith(self.path):
                            if absolute:
                                yield p
                            else:
                                yield line

    def __eq__(self, other):
        return (isinstance(other, EggInfoDistribution) and
                self.path == other.path)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    __hash__ = object.__hash__

new_dist_class = InstalledDistribution
old_dist_class = EggInfoDistribution


class DependencyGraph(object):
    """
    Represents a dependency graph between distributions.

    The dependency relationships are stored in an ``adjacency_list`` that maps
    distributions to a list of ``(other, label)`` tuples where  ``other``
    is a distribution and the edge is labeled with ``label`` (i.e. the version
    specifier, if such was provided). Also, for more efficient traversal, for
    every distribution ``x``, a list of predecessors is kept in
    ``reverse_list[x]``. An edge from distribution ``a`` to
    distribution ``b`` means that ``a`` depends on ``b``. If any missing
    dependencies are found, they are stored in ``missing``, which is a
    dictionary that maps distributions to a list of requirements that were not
    provided by any other distributions.
    """

    def __init__(self):
        self.adjacency_list = {}
        self.reverse_list = {}
        self.missing = {}

    def add_distribution(self, distribution):
        """Add the *distribution* to the graph.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        """
        self.adjacency_list[distribution] = []
        self.reverse_list[distribution] = []
        #self.missing[distribution] = []

    def add_edge(self, x, y, label=None):
        """Add an edge from distribution *x* to distribution *y* with the given
        *label*.

        :type x: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type y: :class:`distutils2.database.InstalledDistribution` or
                 :class:`distutils2.database.EggInfoDistribution`
        :type label: ``str`` or ``None``
        """
        self.adjacency_list[x].append((y, label))
        # multiple edges are allowed, so be careful
        if x not in self.reverse_list[y]:
            self.reverse_list[y].append(x)

    def add_missing(self, distribution, requirement):
        """
        Add a missing *requirement* for the given *distribution*.

        :type distribution: :class:`distutils2.database.InstalledDistribution`
                            or :class:`distutils2.database.EggInfoDistribution`
        :type requirement: ``str``
        """
        logger.debug('%s missing %r', distribution, requirement)
        self.missing.setdefault(distribution, []).append(requirement)

    def _repr_dist(self, dist):
        return '%s %s' % (dist.name, dist.version)

    def repr_node(self, dist, level=1):
        """Prints only a subgraph"""
        output = [self._repr_dist(dist)]
        for other, label in self.adjacency_list[dist]:
            dist = self._repr_dist(other)
            if label is not None:
                dist = '%s [%s]' % (dist, label)
            output.append('    ' * level + str(dist))
            suboutput = self.repr_node(other, level + 1)
            subs = suboutput.split('\n')
            output.extend(subs[1:])
        return '\n'.join(output)

    def to_dot(self, f, skip_disconnected=True):
        """Writes a DOT output for the graph to the provided file *f*.

        If *skip_disconnected* is set to ``True``, then all distributions
        that are not dependent on any other distribution are skipped.

        :type f: has to support ``file``-like operations
        :type skip_disconnected: ``bool``
        """
        disconnected = []

        f.write("digraph dependencies {\n")
        for dist, adjs in self.adjacency_list.items():
            if len(adjs) == 0 and not skip_disconnected:
                disconnected.append(dist)
            for other, label in adjs:
                if not label is None:
                    f.write('"%s" -> "%s" [label="%s"]\n' %
                            (dist.name, other.name, label))
                else:
                    f.write('"%s" -> "%s"\n' % (dist.name, other.name))
        if not skip_disconnected and len(disconnected) > 0:
            f.write('subgraph disconnected {\n')
            f.write('label = "Disconnected"\n')
            f.write('bgcolor = red\n')

            for dist in disconnected:
                f.write('"%s"' % dist.name)
                f.write('\n')
            f.write('}\n')
        f.write('}\n')

    def topological_sort(self):
        """
        Perform a topological sort of the graph.
        :return: A tuple, the first element of which is a topologically sorted
                 list of distributions, and the second element of which is a
                 list of distributions that cannot be sorted because they have
                 circular dependencies and so form a cycle.
        """
        result = []
        # Make a shallow copy of the adjacency list
        alist = {}
        for k, v in self.adjacency_list.items():
            alist[k] = v[:]
        while True:
            # See what we can remove in this run
            to_remove = []
            for k, v in list(alist.items())[:]:
                if not v:
                    to_remove.append(k)
                    del alist[k]
            if not to_remove:
                # What's left in alist (if anything) is a cycle.
                break
            # Remove from the adjacency list of others
            for k, v in alist.items():
                alist[k] = [(d, r) for d, r in v if d not in to_remove]
            logger.debug('Moving to result: %s',
                         ['%s (%s)' % (d.name, d.version) for d in to_remove])
            result.extend(to_remove)
        return result, list(alist.keys())

    def __repr__(self):
        """Representation of the graph"""
        output = []
        for dist, adjs in self.adjacency_list.items():
            output.append(self.repr_node(dist))
        return '\n'.join(output)


def make_graph(dists, scheme='default'):
    """Makes a dependency graph from the given distributions.

    :parameter dists: a list of distributions
    :type dists: list of :class:`distutils2.database.InstalledDistribution` and
                 :class:`distutils2.database.EggInfoDistribution` instances
    :rtype: a :class:`DependencyGraph` instance
    """
    scheme = get_scheme(scheme)
    graph = DependencyGraph()
    provided = {}  # maps names to lists of (version, dist) tuples

    # first, build the graph and find out what's provided
    for dist in dists:
        graph.add_distribution(dist)

        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Add to provided: %s, %s, %s', name, version, dist)
            provided.setdefault(name, []).append((version, dist))

    # now make the edges
    for dist in dists:
        requires = (dist.run_requires | dist.meta_requires |
                    dist.build_requires | dist.dev_requires)
        for req in requires:
            try:
                matcher = scheme.matcher(req)
            except UnsupportedVersionError:
                # XXX compat-mode if cannot read the version
                logger.warning('could not read version %r - using name only',
                               req)
                name = req.split()[0]
                matcher = scheme.matcher(name)

            name = matcher.key   # case-insensitive

            matched = False
            if name in provided:
                for version, provider in provided[name]:
                    try:
                        match = matcher.match(version)
                    except UnsupportedVersionError:
                        match = False

                    if match:
                        graph.add_edge(dist, provider, req)
                        matched = True
                        break
            if not matched:
                graph.add_missing(dist, req)
    return graph


def get_dependent_dists(dists, dist):
    """Recursively generate a list of distributions from *dists* that are
    dependent on *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    """
    if dist not in dists:
        raise DistlibException('given distribution %r is not a member '
                               'of the list' % dist.name)
    graph = make_graph(dists)

    dep = [dist]  # dependent distributions
    todo = graph.reverse_list[dist]  # list of nodes we should inspect

    while todo:
        d = todo.pop()
        dep.append(d)
        for succ in graph.reverse_list[d]:
            if succ not in dep:
                todo.append(succ)

    dep.pop(0)  # remove dist from dep, was there to prevent infinite loops
    return dep


def get_required_dists(dists, dist):
    """Recursively generate a list of distributions from *dists* that are
    required by *dist*.

    :param dists: a list of distributions
    :param dist: a distribution, member of *dists* for which we are interested
    """
    if dist not in dists:
        raise DistlibException('given distribution %r is not a member '
                               'of the list' % dist.name)
    graph = make_graph(dists)

    req = []  # required distributions
    todo = graph.adjacency_list[dist]  # list of nodes we should inspect

    while todo:
        d = todo.pop()[0]
        req.append(d)
        for pred in graph.adjacency_list[d]:
            if pred not in req:
                todo.append(pred)

    return req


def make_dist(name, version, **kwargs):
    """
    A convenience method for making a dist given just a name and version.
    """
    summary = kwargs.pop('summary', 'Placeholder for summary')
    md = Metadata(**kwargs)
    md.name = name
    md.version = version
    md.summary = summary or 'Placeholder for summary'
    return Distribution(md)
site-packages/pip/_vendor/distlib/version.py000064400000055731150564745010015223 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2017 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""
Implementation of a flexible versioning scheme providing support for PEP-440,
setuptools-compatible and semantic versioning.
"""

import logging
import re

from .compat import string_types
from .util import parse_requirement

__all__ = ['NormalizedVersion', 'NormalizedMatcher',
           'LegacyVersion', 'LegacyMatcher',
           'SemanticVersion', 'SemanticMatcher',
           'UnsupportedVersionError', 'get_scheme']

logger = logging.getLogger(__name__)


class UnsupportedVersionError(ValueError):
    """This is an unsupported version."""
    pass


class Version(object):
    def __init__(self, s):
        self._string = s = s.strip()
        self._parts = parts = self.parse(s)
        assert isinstance(parts, tuple)
        assert len(parts) > 0

    def parse(self, s):
        raise NotImplementedError('please implement in a subclass')

    def _check_compatible(self, other):
        if type(self) != type(other):
            raise TypeError('cannot compare %r and %r' % (self, other))

    def __eq__(self, other):
        self._check_compatible(other)
        return self._parts == other._parts

    def __ne__(self, other):
        return not self.__eq__(other)

    def __lt__(self, other):
        self._check_compatible(other)
        return self._parts < other._parts

    def __gt__(self, other):
        return not (self.__lt__(other) or self.__eq__(other))

    def __le__(self, other):
        return self.__lt__(other) or self.__eq__(other)

    def __ge__(self, other):
        return self.__gt__(other) or self.__eq__(other)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    def __hash__(self):
        return hash(self._parts)

    def __repr__(self):
        return "%s('%s')" % (self.__class__.__name__, self._string)

    def __str__(self):
        return self._string

    @property
    def is_prerelease(self):
        raise NotImplementedError('Please implement in subclasses.')


class Matcher(object):
    version_class = None

    # value is either a callable or the name of a method
    _operators = {
        '<': lambda v, c, p: v < c,
        '>': lambda v, c, p: v > c,
        '<=': lambda v, c, p: v == c or v < c,
        '>=': lambda v, c, p: v == c or v > c,
        '==': lambda v, c, p: v == c,
        '===': lambda v, c, p: v == c,
        # by default, compatible => >=.
        '~=': lambda v, c, p: v == c or v > c,
        '!=': lambda v, c, p: v != c,
    }

    # this is a method only to support alternative implementations
    # via overriding
    def parse_requirement(self, s):
        return parse_requirement(s)

    def __init__(self, s):
        if self.version_class is None:
            raise ValueError('Please specify a version class')
        self._string = s = s.strip()
        r = self.parse_requirement(s)
        if not r:
            raise ValueError('Not valid: %r' % s)
        self.name = r.name
        self.key = self.name.lower()    # for case-insensitive comparisons
        clist = []
        if r.constraints:
            # import pdb; pdb.set_trace()
            for op, s in r.constraints:
                if s.endswith('.*'):
                    if op not in ('==', '!='):
                        raise ValueError('\'.*\' not allowed for '
                                         '%r constraints' % op)
                    # Could be a partial version (e.g. for '2.*') which
                    # won't parse as a version, so keep it as a string
                    vn, prefix = s[:-2], True
                    # Just to check that vn is a valid version
                    self.version_class(vn)
                else:
                    # Should parse as a version, so we can create an
                    # instance for the comparison
                    vn, prefix = self.version_class(s), False
                clist.append((op, vn, prefix))
        self._parts = tuple(clist)

    def match(self, version):
        """
        Check if the provided version matches the constraints.

        :param version: The version to match against this instance.
        :type version: String or :class:`Version` instance.
        """
        if isinstance(version, string_types):
            version = self.version_class(version)
        for operator, constraint, prefix in self._parts:
            f = self._operators.get(operator)
            if isinstance(f, string_types):
                f = getattr(self, f)
            if not f:
                msg = ('%r not implemented '
                       'for %s' % (operator, self.__class__.__name__))
                raise NotImplementedError(msg)
            if not f(version, constraint, prefix):
                return False
        return True

    @property
    def exact_version(self):
        result = None
        if len(self._parts) == 1 and self._parts[0][0] in ('==', '==='):
            result = self._parts[0][1]
        return result

    def _check_compatible(self, other):
        if type(self) != type(other) or self.name != other.name:
            raise TypeError('cannot compare %s and %s' % (self, other))

    def __eq__(self, other):
        self._check_compatible(other)
        return self.key == other.key and self._parts == other._parts

    def __ne__(self, other):
        return not self.__eq__(other)

    # See http://docs.python.org/reference/datamodel#object.__hash__
    def __hash__(self):
        return hash(self.key) + hash(self._parts)

    def __repr__(self):
        return "%s(%r)" % (self.__class__.__name__, self._string)

    def __str__(self):
        return self._string


PEP440_VERSION_RE = re.compile(r'^v?(\d+!)?(\d+(\.\d+)*)((a|b|c|rc)(\d+))?'
                               r'(\.(post)(\d+))?(\.(dev)(\d+))?'
                               r'(\+([a-zA-Z\d]+(\.[a-zA-Z\d]+)?))?$')


def _pep_440_key(s):
    s = s.strip()
    m = PEP440_VERSION_RE.match(s)
    if not m:
        raise UnsupportedVersionError('Not a valid version: %s' % s)
    groups = m.groups()
    nums = tuple(int(v) for v in groups[1].split('.'))
    while len(nums) > 1 and nums[-1] == 0:
        nums = nums[:-1]

    if not groups[0]:
        epoch = 0
    else:
        epoch = int(groups[0][:-1])
    pre = groups[4:6]
    post = groups[7:9]
    dev = groups[10:12]
    local = groups[13]
    if pre == (None, None):
        pre = ()
    else:
        pre = pre[0], int(pre[1])
    if post == (None, None):
        post = ()
    else:
        post = post[0], int(post[1])
    if dev == (None, None):
        dev = ()
    else:
        dev = dev[0], int(dev[1])
    if local is None:
        local = ()
    else:
        parts = []
        for part in local.split('.'):
            # to ensure that numeric compares as > lexicographic, avoid
            # comparing them directly, but encode a tuple which ensures
            # correct sorting
            if part.isdigit():
                part = (1, int(part))
            else:
                part = (0, part)
            parts.append(part)
        local = tuple(parts)
    if not pre:
        # either before pre-release, or final release and after
        if not post and dev:
            # before pre-release
            pre = ('a', -1)     # to sort before a0
        else:
            pre = ('z',)        # to sort after all pre-releases
    # now look at the state of post and dev.
    if not post:
        post = ('_',)   # sort before 'a'
    if not dev:
        dev = ('final',)

    #print('%s -> %s' % (s, m.groups()))
    return epoch, nums, pre, post, dev, local


_normalized_key = _pep_440_key


class NormalizedVersion(Version):
    """A rational version.

    Good:
        1.2         # equivalent to "1.2.0"
        1.2.0
        1.2a1
        1.2.3a2
        1.2.3b1
        1.2.3c1
        1.2.3.4
        TODO: fill this out

    Bad:
        1           # minimum two numbers
        1.2a        # release level must have a release serial
        1.2.3b
    """
    def parse(self, s):
        result = _normalized_key(s)
        # _normalized_key loses trailing zeroes in the release
        # clause, since that's needed to ensure that X.Y == X.Y.0 == X.Y.0.0
        # However, PEP 440 prefix matching needs it: for example,
        # (~= 1.4.5.0) matches differently to (~= 1.4.5.0.0).
        m = PEP440_VERSION_RE.match(s)      # must succeed
        groups = m.groups()
        self._release_clause = tuple(int(v) for v in groups[1].split('.'))
        return result

    PREREL_TAGS = set(['a', 'b', 'c', 'rc', 'dev'])

    @property
    def is_prerelease(self):
        return any(t[0] in self.PREREL_TAGS for t in self._parts if t)


def _match_prefix(x, y):
    x = str(x)
    y = str(y)
    if x == y:
        return True
    if not x.startswith(y):
        return False
    n = len(y)
    return x[n] == '.'


class NormalizedMatcher(Matcher):
    version_class = NormalizedVersion

    # value is either a callable or the name of a method
    _operators = {
        '~=': '_match_compatible',
        '<': '_match_lt',
        '>': '_match_gt',
        '<=': '_match_le',
        '>=': '_match_ge',
        '==': '_match_eq',
        '===': '_match_arbitrary',
        '!=': '_match_ne',
    }

    def _adjust_local(self, version, constraint, prefix):
        if prefix:
            strip_local = '+' not in constraint and version._parts[-1]
        else:
            # both constraint and version are
            # NormalizedVersion instances.
            # If constraint does not have a local component,
            # ensure the version doesn't, either.
            strip_local = not constraint._parts[-1] and version._parts[-1]
        if strip_local:
            s = version._string.split('+', 1)[0]
            version = self.version_class(s)
        return version, constraint

    def _match_lt(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version >= constraint:
            return False
        release_clause = constraint._release_clause
        pfx = '.'.join([str(i) for i in release_clause])
        return not _match_prefix(version, pfx)

    def _match_gt(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version <= constraint:
            return False
        release_clause = constraint._release_clause
        pfx = '.'.join([str(i) for i in release_clause])
        return not _match_prefix(version, pfx)

    def _match_le(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        return version <= constraint

    def _match_ge(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        return version >= constraint

    def _match_eq(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if not prefix:
            result = (version == constraint)
        else:
            result = _match_prefix(version, constraint)
        return result

    def _match_arbitrary(self, version, constraint, prefix):
        return str(version) == str(constraint)

    def _match_ne(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if not prefix:
            result = (version != constraint)
        else:
            result = not _match_prefix(version, constraint)
        return result

    def _match_compatible(self, version, constraint, prefix):
        version, constraint = self._adjust_local(version, constraint, prefix)
        if version == constraint:
            return True
        if version < constraint:
            return False
#        if not prefix:
#            return True
        release_clause = constraint._release_clause
        if len(release_clause) > 1:
            release_clause = release_clause[:-1]
        pfx = '.'.join([str(i) for i in release_clause])
        return _match_prefix(version, pfx)

_REPLACEMENTS = (
    (re.compile('[.+-]$'), ''),                     # remove trailing puncts
    (re.compile(r'^[.](\d)'), r'0.\1'),             # .N -> 0.N at start
    (re.compile('^[.-]'), ''),                      # remove leading puncts
    (re.compile(r'^\((.*)\)$'), r'\1'),             # remove parentheses
    (re.compile(r'^v(ersion)?\s*(\d+)'), r'\2'),    # remove leading v(ersion)
    (re.compile(r'^r(ev)?\s*(\d+)'), r'\2'),        # remove leading v(ersion)
    (re.compile('[.]{2,}'), '.'),                   # multiple runs of '.'
    (re.compile(r'\b(alfa|apha)\b'), 'alpha'),      # misspelt alpha
    (re.compile(r'\b(pre-alpha|prealpha)\b'),
                'pre.alpha'),                       # standardise
    (re.compile(r'\(beta\)$'), 'beta'),             # remove parentheses
)

_SUFFIX_REPLACEMENTS = (
    (re.compile('^[:~._+-]+'), ''),                   # remove leading puncts
    (re.compile('[,*")([\\]]'), ''),                  # remove unwanted chars
    (re.compile('[~:+_ -]'), '.'),                    # replace illegal chars
    (re.compile('[.]{2,}'), '.'),                   # multiple runs of '.'
    (re.compile(r'\.$'), ''),                       # trailing '.'
)

_NUMERIC_PREFIX = re.compile(r'(\d+(\.\d+)*)')


def _suggest_semantic_version(s):
    """
    Try to suggest a semantic form for a version for which
    _suggest_normalized_version couldn't come up with anything.
    """
    result = s.strip().lower()
    for pat, repl in _REPLACEMENTS:
        result = pat.sub(repl, result)
    if not result:
        result = '0.0.0'

    # Now look for numeric prefix, and separate it out from
    # the rest.
    #import pdb; pdb.set_trace()
    m = _NUMERIC_PREFIX.match(result)
    if not m:
        prefix = '0.0.0'
        suffix = result
    else:
        prefix = m.groups()[0].split('.')
        prefix = [int(i) for i in prefix]
        while len(prefix) < 3:
            prefix.append(0)
        if len(prefix) == 3:
            suffix = result[m.end():]
        else:
            suffix = '.'.join([str(i) for i in prefix[3:]]) + result[m.end():]
            prefix = prefix[:3]
        prefix = '.'.join([str(i) for i in prefix])
        suffix = suffix.strip()
    if suffix:
        #import pdb; pdb.set_trace()
        # massage the suffix.
        for pat, repl in _SUFFIX_REPLACEMENTS:
            suffix = pat.sub(repl, suffix)

    if not suffix:
        result = prefix
    else:
        sep = '-' if 'dev' in suffix else '+'
        result = prefix + sep + suffix
    if not is_semver(result):
        result = None
    return result


def _suggest_normalized_version(s):
    """Suggest a normalized version close to the given version string.

    If you have a version string that isn't rational (i.e. NormalizedVersion
    doesn't like it) then you might be able to get an equivalent (or close)
    rational version from this function.

    This does a number of simple normalizations to the given string, based
    on observation of versions currently in use on PyPI. Given a dump of
    those version during PyCon 2009, 4287 of them:
    - 2312 (53.93%) match NormalizedVersion without change
      with the automatic suggestion
    - 3474 (81.04%) match when using this suggestion method

    @param s {str} An irrational version string.
    @returns A rational version string, or None, if couldn't determine one.
    """
    try:
        _normalized_key(s)
        return s   # already rational
    except UnsupportedVersionError:
        pass

    rs = s.lower()

    # part of this could use maketrans
    for orig, repl in (('-alpha', 'a'), ('-beta', 'b'), ('alpha', 'a'),
                       ('beta', 'b'), ('rc', 'c'), ('-final', ''),
                       ('-pre', 'c'),
                       ('-release', ''), ('.release', ''), ('-stable', ''),
                       ('+', '.'), ('_', '.'), (' ', ''), ('.final', ''),
                       ('final', '')):
        rs = rs.replace(orig, repl)

    # if something ends with dev or pre, we add a 0
    rs = re.sub(r"pre$", r"pre0", rs)
    rs = re.sub(r"dev$", r"dev0", rs)

    # if we have something like "b-2" or "a.2" at the end of the
    # version, that is probably beta, alpha, etc
    # let's remove the dash or dot
    rs = re.sub(r"([abc]|rc)[\-\.](\d+)$", r"\1\2", rs)

    # 1.0-dev-r371 -> 1.0.dev371
    # 0.1-dev-r79 -> 0.1.dev79
    rs = re.sub(r"[\-\.](dev)[\-\.]?r?(\d+)$", r".\1\2", rs)

    # Clean: 2.0.a.3, 2.0.b1, 0.9.0~c1
    rs = re.sub(r"[.~]?([abc])\.?", r"\1", rs)

    # Clean: v0.3, v1.0
    if rs.startswith('v'):
        rs = rs[1:]

    # Clean leading '0's on numbers.
    #TODO: unintended side-effect on, e.g., "2003.05.09"
    # PyPI stats: 77 (~2%) better
    rs = re.sub(r"\b0+(\d+)(?!\d)", r"\1", rs)

    # Clean a/b/c with no version. E.g. "1.0a" -> "1.0a0". Setuptools infers
    # zero.
    # PyPI stats: 245 (7.56%) better
    rs = re.sub(r"(\d+[abc])$", r"\g<1>0", rs)

    # the 'dev-rNNN' tag is a dev tag
    rs = re.sub(r"\.?(dev-r|dev\.r)\.?(\d+)$", r".dev\2", rs)

    # clean the - when used as a pre delimiter
    rs = re.sub(r"-(a|b|c)(\d+)$", r"\1\2", rs)

    # a terminal "dev" or "devel" can be changed into ".dev0"
    rs = re.sub(r"[\.\-](dev|devel)$", r".dev0", rs)

    # a terminal "dev" can be changed into ".dev0"
    rs = re.sub(r"(?![\.\-])dev$", r".dev0", rs)

    # a terminal "final" or "stable" can be removed
    rs = re.sub(r"(final|stable)$", "", rs)

    # The 'r' and the '-' tags are post release tags
    #   0.4a1.r10       ->  0.4a1.post10
    #   0.9.33-17222    ->  0.9.33.post17222
    #   0.9.33-r17222   ->  0.9.33.post17222
    rs = re.sub(r"\.?(r|-|-r)\.?(\d+)$", r".post\2", rs)

    # Clean 'r' instead of 'dev' usage:
    #   0.9.33+r17222   ->  0.9.33.dev17222
    #   1.0dev123       ->  1.0.dev123
    #   1.0.git123      ->  1.0.dev123
    #   1.0.bzr123      ->  1.0.dev123
    #   0.1a0dev.123    ->  0.1a0.dev123
    # PyPI stats:  ~150 (~4%) better
    rs = re.sub(r"\.?(dev|git|bzr)\.?(\d+)$", r".dev\2", rs)

    # Clean '.pre' (normalized from '-pre' above) instead of 'c' usage:
    #   0.2.pre1        ->  0.2c1
    #   0.2-c1         ->  0.2c1
    #   1.0preview123   ->  1.0c123
    # PyPI stats: ~21 (0.62%) better
    rs = re.sub(r"\.?(pre|preview|-c)(\d+)$", r"c\g<2>", rs)

    # Tcl/Tk uses "px" for their post release markers
    rs = re.sub(r"p(\d+)$", r".post\1", rs)

    try:
        _normalized_key(rs)
    except UnsupportedVersionError:
        rs = None
    return rs

#
#   Legacy version processing (distribute-compatible)
#

_VERSION_PART = re.compile(r'([a-z]+|\d+|[\.-])', re.I)
_VERSION_REPLACE = {
    'pre': 'c',
    'preview': 'c',
    '-': 'final-',
    'rc': 'c',
    'dev': '@',
    '': None,
    '.': None,
}


def _legacy_key(s):
    def get_parts(s):
        result = []
        for p in _VERSION_PART.split(s.lower()):
            p = _VERSION_REPLACE.get(p, p)
            if p:
                if '0' <= p[:1] <= '9':
                    p = p.zfill(8)
                else:
                    p = '*' + p
                result.append(p)
        result.append('*final')
        return result

    result = []
    for p in get_parts(s):
        if p.startswith('*'):
            if p < '*final':
                while result and result[-1] == '*final-':
                    result.pop()
            while result and result[-1] == '00000000':
                result.pop()
        result.append(p)
    return tuple(result)


class LegacyVersion(Version):
    def parse(self, s):
        return _legacy_key(s)

    @property
    def is_prerelease(self):
        result = False
        for x in self._parts:
            if (isinstance(x, string_types) and x.startswith('*') and
                x < '*final'):
                result = True
                break
        return result


class LegacyMatcher(Matcher):
    version_class = LegacyVersion

    _operators = dict(Matcher._operators)
    _operators['~='] = '_match_compatible'

    numeric_re = re.compile(r'^(\d+(\.\d+)*)')

    def _match_compatible(self, version, constraint, prefix):
        if version < constraint:
            return False
        m = self.numeric_re.match(str(constraint))
        if not m:
            logger.warning('Cannot compute compatible match for version %s '
                           ' and constraint %s', version, constraint)
            return True
        s = m.groups()[0]
        if '.' in s:
            s = s.rsplit('.', 1)[0]
        return _match_prefix(version, s)

#
#   Semantic versioning
#

_SEMVER_RE = re.compile(r'^(\d+)\.(\d+)\.(\d+)'
                        r'(-[a-z0-9]+(\.[a-z0-9-]+)*)?'
                        r'(\+[a-z0-9]+(\.[a-z0-9-]+)*)?$', re.I)


def is_semver(s):
    return _SEMVER_RE.match(s)


def _semantic_key(s):
    def make_tuple(s, absent):
        if s is None:
            result = (absent,)
        else:
            parts = s[1:].split('.')
            # We can't compare ints and strings on Python 3, so fudge it
            # by zero-filling numeric values so simulate a numeric comparison
            result = tuple([p.zfill(8) if p.isdigit() else p for p in parts])
        return result

    m = is_semver(s)
    if not m:
        raise UnsupportedVersionError(s)
    groups = m.groups()
    major, minor, patch = [int(i) for i in groups[:3]]
    # choose the '|' and '*' so that versions sort correctly
    pre, build = make_tuple(groups[3], '|'), make_tuple(groups[5], '*')
    return (major, minor, patch), pre, build


class SemanticVersion(Version):
    def parse(self, s):
        return _semantic_key(s)

    @property
    def is_prerelease(self):
        return self._parts[1][0] != '|'


class SemanticMatcher(Matcher):
    version_class = SemanticVersion


class VersionScheme(object):
    def __init__(self, key, matcher, suggester=None):
        self.key = key
        self.matcher = matcher
        self.suggester = suggester

    def is_valid_version(self, s):
        try:
            self.matcher.version_class(s)
            result = True
        except UnsupportedVersionError:
            result = False
        return result

    def is_valid_matcher(self, s):
        try:
            self.matcher(s)
            result = True
        except UnsupportedVersionError:
            result = False
        return result

    def is_valid_constraint_list(self, s):
        """
        Used for processing some metadata fields
        """
        # See issue #140. Be tolerant of a single trailing comma.
        if s.endswith(','):
            s = s[:-1]
        return self.is_valid_matcher('dummy_name (%s)' % s)

    def suggest(self, s):
        if self.suggester is None:
            result = None
        else:
            result = self.suggester(s)
        return result

_SCHEMES = {
    'normalized': VersionScheme(_normalized_key, NormalizedMatcher,
                                _suggest_normalized_version),
    'legacy': VersionScheme(_legacy_key, LegacyMatcher, lambda self, s: s),
    'semantic': VersionScheme(_semantic_key, SemanticMatcher,
                              _suggest_semantic_version),
}

_SCHEMES['default'] = _SCHEMES['normalized']


def get_scheme(name):
    if name not in _SCHEMES:
        raise ValueError('unknown scheme name: %r' % name)
    return _SCHEMES[name]
site-packages/pip/_vendor/distlib/__init__.py000064400000001105150564745010015257 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2019 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import logging

__version__ = '0.3.3'

class DistlibException(Exception):
    pass

try:
    from logging import NullHandler
except ImportError: # pragma: no cover
    class NullHandler(logging.Handler):
        def handle(self, record): pass
        def emit(self, record): pass
        def createLock(self): self.lock = None

logger = logging.getLogger(__name__)
logger.addHandler(NullHandler())
site-packages/pip/_vendor/distlib/util.py000064400000204266150564745010014512 0ustar00#
# Copyright (C) 2012-2021 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
import codecs
from collections import deque
import contextlib
import csv
from glob import iglob as std_iglob
import io
import json
import logging
import os
import py_compile
import re
import socket
try:
    import ssl
except ImportError:  # pragma: no cover
    ssl = None
import subprocess
import sys
import tarfile
import tempfile
import textwrap

try:
    import threading
except ImportError:  # pragma: no cover
    import dummy_threading as threading
import time

from . import DistlibException
from .compat import (string_types, text_type, shutil, raw_input, StringIO,
                     cache_from_source, urlopen, urljoin, httplib, xmlrpclib,
                     splittype, HTTPHandler, BaseConfigurator, valid_ident,
                     Container, configparser, URLError, ZipFile, fsdecode,
                     unquote, urlparse)

logger = logging.getLogger(__name__)

#
# Requirement parsing code as per PEP 508
#

IDENTIFIER = re.compile(r'^([\w\.-]+)\s*')
VERSION_IDENTIFIER = re.compile(r'^([\w\.*+-]+)\s*')
COMPARE_OP = re.compile(r'^(<=?|>=?|={2,3}|[~!]=)\s*')
MARKER_OP = re.compile(r'^((<=?)|(>=?)|={2,3}|[~!]=|in|not\s+in)\s*')
OR = re.compile(r'^or\b\s*')
AND = re.compile(r'^and\b\s*')
NON_SPACE = re.compile(r'(\S+)\s*')
STRING_CHUNK = re.compile(r'([\s\w\.{}()*+#:;,/?!~`@$%^&=|<>\[\]-]+)')


def parse_marker(marker_string):
    """
    Parse a marker string and return a dictionary containing a marker expression.

    The dictionary will contain keys "op", "lhs" and "rhs" for non-terminals in
    the expression grammar, or strings. A string contained in quotes is to be
    interpreted as a literal string, and a string not contained in quotes is a
    variable (such as os_name).
    """
    def marker_var(remaining):
        # either identifier, or literal string
        m = IDENTIFIER.match(remaining)
        if m:
            result = m.groups()[0]
            remaining = remaining[m.end():]
        elif not remaining:
            raise SyntaxError('unexpected end of input')
        else:
            q = remaining[0]
            if q not in '\'"':
                raise SyntaxError('invalid expression: %s' % remaining)
            oq = '\'"'.replace(q, '')
            remaining = remaining[1:]
            parts = [q]
            while remaining:
                # either a string chunk, or oq, or q to terminate
                if remaining[0] == q:
                    break
                elif remaining[0] == oq:
                    parts.append(oq)
                    remaining = remaining[1:]
                else:
                    m = STRING_CHUNK.match(remaining)
                    if not m:
                        raise SyntaxError('error in string literal: %s' % remaining)
                    parts.append(m.groups()[0])
                    remaining = remaining[m.end():]
            else:
                s = ''.join(parts)
                raise SyntaxError('unterminated string: %s' % s)
            parts.append(q)
            result = ''.join(parts)
            remaining = remaining[1:].lstrip() # skip past closing quote
        return result, remaining

    def marker_expr(remaining):
        if remaining and remaining[0] == '(':
            result, remaining = marker(remaining[1:].lstrip())
            if remaining[0] != ')':
                raise SyntaxError('unterminated parenthesis: %s' % remaining)
            remaining = remaining[1:].lstrip()
        else:
            lhs, remaining = marker_var(remaining)
            while remaining:
                m = MARKER_OP.match(remaining)
                if not m:
                    break
                op = m.groups()[0]
                remaining = remaining[m.end():]
                rhs, remaining = marker_var(remaining)
                lhs = {'op': op, 'lhs': lhs, 'rhs': rhs}
            result = lhs
        return result, remaining

    def marker_and(remaining):
        lhs, remaining = marker_expr(remaining)
        while remaining:
            m = AND.match(remaining)
            if not m:
                break
            remaining = remaining[m.end():]
            rhs, remaining = marker_expr(remaining)
            lhs = {'op': 'and', 'lhs': lhs, 'rhs': rhs}
        return lhs, remaining

    def marker(remaining):
        lhs, remaining = marker_and(remaining)
        while remaining:
            m = OR.match(remaining)
            if not m:
                break
            remaining = remaining[m.end():]
            rhs, remaining = marker_and(remaining)
            lhs = {'op': 'or', 'lhs': lhs, 'rhs': rhs}
        return lhs, remaining

    return marker(marker_string)


def parse_requirement(req):
    """
    Parse a requirement passed in as a string. Return a Container
    whose attributes contain the various parts of the requirement.
    """
    remaining = req.strip()
    if not remaining or remaining.startswith('#'):
        return None
    m = IDENTIFIER.match(remaining)
    if not m:
        raise SyntaxError('name expected: %s' % remaining)
    distname = m.groups()[0]
    remaining = remaining[m.end():]
    extras = mark_expr = versions = uri = None
    if remaining and remaining[0] == '[':
        i = remaining.find(']', 1)
        if i < 0:
            raise SyntaxError('unterminated extra: %s' % remaining)
        s = remaining[1:i]
        remaining = remaining[i + 1:].lstrip()
        extras = []
        while s:
            m = IDENTIFIER.match(s)
            if not m:
                raise SyntaxError('malformed extra: %s' % s)
            extras.append(m.groups()[0])
            s = s[m.end():]
            if not s:
                break
            if s[0] != ',':
                raise SyntaxError('comma expected in extras: %s' % s)
            s = s[1:].lstrip()
        if not extras:
            extras = None
    if remaining:
        if remaining[0] == '@':
            # it's a URI
            remaining = remaining[1:].lstrip()
            m = NON_SPACE.match(remaining)
            if not m:
                raise SyntaxError('invalid URI: %s' % remaining)
            uri = m.groups()[0]
            t = urlparse(uri)
            # there are issues with Python and URL parsing, so this test
            # is a bit crude. See bpo-20271, bpo-23505. Python doesn't
            # always parse invalid URLs correctly - it should raise
            # exceptions for malformed URLs
            if not (t.scheme and t.netloc):
                raise SyntaxError('Invalid URL: %s' % uri)
            remaining = remaining[m.end():].lstrip()
        else:

            def get_versions(ver_remaining):
                """
                Return a list of operator, version tuples if any are
                specified, else None.
                """
                m = COMPARE_OP.match(ver_remaining)
                versions = None
                if m:
                    versions = []
                    while True:
                        op = m.groups()[0]
                        ver_remaining = ver_remaining[m.end():]
                        m = VERSION_IDENTIFIER.match(ver_remaining)
                        if not m:
                            raise SyntaxError('invalid version: %s' % ver_remaining)
                        v = m.groups()[0]
                        versions.append((op, v))
                        ver_remaining = ver_remaining[m.end():]
                        if not ver_remaining or ver_remaining[0] != ',':
                            break
                        ver_remaining = ver_remaining[1:].lstrip()
                        # Some packages have a trailing comma which would break things
                        # See issue #148
                        if not ver_remaining:
                            break
                        m = COMPARE_OP.match(ver_remaining)
                        if not m:
                            raise SyntaxError('invalid constraint: %s' % ver_remaining)
                    if not versions:
                        versions = None
                return versions, ver_remaining

            if remaining[0] != '(':
                versions, remaining = get_versions(remaining)
            else:
                i = remaining.find(')', 1)
                if i < 0:
                    raise SyntaxError('unterminated parenthesis: %s' % remaining)
                s = remaining[1:i]
                remaining = remaining[i + 1:].lstrip()
                # As a special diversion from PEP 508, allow a version number
                # a.b.c in parentheses as a synonym for ~= a.b.c (because this
                # is allowed in earlier PEPs)
                if COMPARE_OP.match(s):
                    versions, _ = get_versions(s)
                else:
                    m = VERSION_IDENTIFIER.match(s)
                    if not m:
                        raise SyntaxError('invalid constraint: %s' % s)
                    v = m.groups()[0]
                    s = s[m.end():].lstrip()
                    if s:
                        raise SyntaxError('invalid constraint: %s' % s)
                    versions = [('~=', v)]

    if remaining:
        if remaining[0] != ';':
            raise SyntaxError('invalid requirement: %s' % remaining)
        remaining = remaining[1:].lstrip()

        mark_expr, remaining = parse_marker(remaining)

    if remaining and remaining[0] != '#':
        raise SyntaxError('unexpected trailing data: %s' % remaining)

    if not versions:
        rs = distname
    else:
        rs = '%s %s' % (distname, ', '.join(['%s %s' % con for con in versions]))
    return Container(name=distname, extras=extras, constraints=versions,
                     marker=mark_expr, url=uri, requirement=rs)


def get_resources_dests(resources_root, rules):
    """Find destinations for resources files"""

    def get_rel_path(root, path):
        # normalizes and returns a lstripped-/-separated path
        root = root.replace(os.path.sep, '/')
        path = path.replace(os.path.sep, '/')
        assert path.startswith(root)
        return path[len(root):].lstrip('/')

    destinations = {}
    for base, suffix, dest in rules:
        prefix = os.path.join(resources_root, base)
        for abs_base in iglob(prefix):
            abs_glob = os.path.join(abs_base, suffix)
            for abs_path in iglob(abs_glob):
                resource_file = get_rel_path(resources_root, abs_path)
                if dest is None:  # remove the entry if it was here
                    destinations.pop(resource_file, None)
                else:
                    rel_path = get_rel_path(abs_base, abs_path)
                    rel_dest = dest.replace(os.path.sep, '/').rstrip('/')
                    destinations[resource_file] = rel_dest + '/' + rel_path
    return destinations


def in_venv():
    if hasattr(sys, 'real_prefix'):
        # virtualenv venvs
        result = True
    else:
        # PEP 405 venvs
        result = sys.prefix != getattr(sys, 'base_prefix', sys.prefix)
    return result


def get_executable():
# The __PYVENV_LAUNCHER__ dance is apparently no longer needed, as
# changes to the stub launcher mean that sys.executable always points
# to the stub on OS X
#    if sys.platform == 'darwin' and ('__PYVENV_LAUNCHER__'
#                                     in os.environ):
#        result =  os.environ['__PYVENV_LAUNCHER__']
#    else:
#        result = sys.executable
#    return result
    # Avoid normcasing: see issue #143
    # result = os.path.normcase(sys.executable)
    result = sys.executable
    if not isinstance(result, text_type):
        result = fsdecode(result)
    return result


def proceed(prompt, allowed_chars, error_prompt=None, default=None):
    p = prompt
    while True:
        s = raw_input(p)
        p = prompt
        if not s and default:
            s = default
        if s:
            c = s[0].lower()
            if c in allowed_chars:
                break
            if error_prompt:
                p = '%c: %s\n%s' % (c, error_prompt, prompt)
    return c


def extract_by_key(d, keys):
    if isinstance(keys, string_types):
        keys = keys.split()
    result = {}
    for key in keys:
        if key in d:
            result[key] = d[key]
    return result

def read_exports(stream):
    if sys.version_info[0] >= 3:
        # needs to be a text stream
        stream = codecs.getreader('utf-8')(stream)
    # Try to load as JSON, falling back on legacy format
    data = stream.read()
    stream = StringIO(data)
    try:
        jdata = json.load(stream)
        result = jdata['extensions']['python.exports']['exports']
        for group, entries in result.items():
            for k, v in entries.items():
                s = '%s = %s' % (k, v)
                entry = get_export_entry(s)
                assert entry is not None
                entries[k] = entry
        return result
    except Exception:
        stream.seek(0, 0)

    def read_stream(cp, stream):
        if hasattr(cp, 'read_file'):
            cp.read_file(stream)
        else:
            cp.readfp(stream)

    cp = configparser.ConfigParser()
    try:
        read_stream(cp, stream)
    except configparser.MissingSectionHeaderError:
        stream.close()
        data = textwrap.dedent(data)
        stream = StringIO(data)
        read_stream(cp, stream)

    result = {}
    for key in cp.sections():
        result[key] = entries = {}
        for name, value in cp.items(key):
            s = '%s = %s' % (name, value)
            entry = get_export_entry(s)
            assert entry is not None
            #entry.dist = self
            entries[name] = entry
    return result


def write_exports(exports, stream):
    if sys.version_info[0] >= 3:
        # needs to be a text stream
        stream = codecs.getwriter('utf-8')(stream)
    cp = configparser.ConfigParser()
    for k, v in exports.items():
        # TODO check k, v for valid values
        cp.add_section(k)
        for entry in v.values():
            if entry.suffix is None:
                s = entry.prefix
            else:
                s = '%s:%s' % (entry.prefix, entry.suffix)
            if entry.flags:
                s = '%s [%s]' % (s, ', '.join(entry.flags))
            cp.set(k, entry.name, s)
    cp.write(stream)


@contextlib.contextmanager
def tempdir():
    td = tempfile.mkdtemp()
    try:
        yield td
    finally:
        shutil.rmtree(td)

@contextlib.contextmanager
def chdir(d):
    cwd = os.getcwd()
    try:
        os.chdir(d)
        yield
    finally:
        os.chdir(cwd)


@contextlib.contextmanager
def socket_timeout(seconds=15):
    cto = socket.getdefaulttimeout()
    try:
        socket.setdefaulttimeout(seconds)
        yield
    finally:
        socket.setdefaulttimeout(cto)


class cached_property(object):
    def __init__(self, func):
        self.func = func
        #for attr in ('__name__', '__module__', '__doc__'):
        #    setattr(self, attr, getattr(func, attr, None))

    def __get__(self, obj, cls=None):
        if obj is None:
            return self
        value = self.func(obj)
        object.__setattr__(obj, self.func.__name__, value)
        #obj.__dict__[self.func.__name__] = value = self.func(obj)
        return value

def convert_path(pathname):
    """Return 'pathname' as a name that will work on the native filesystem.

    The path is split on '/' and put back together again using the current
    directory separator.  Needed because filenames in the setup script are
    always supplied in Unix style, and have to be converted to the local
    convention before we can actually use them in the filesystem.  Raises
    ValueError on non-Unix-ish systems if 'pathname' either starts or
    ends with a slash.
    """
    if os.sep == '/':
        return pathname
    if not pathname:
        return pathname
    if pathname[0] == '/':
        raise ValueError("path '%s' cannot be absolute" % pathname)
    if pathname[-1] == '/':
        raise ValueError("path '%s' cannot end with '/'" % pathname)

    paths = pathname.split('/')
    while os.curdir in paths:
        paths.remove(os.curdir)
    if not paths:
        return os.curdir
    return os.path.join(*paths)


class FileOperator(object):
    def __init__(self, dry_run=False):
        self.dry_run = dry_run
        self.ensured = set()
        self._init_record()

    def _init_record(self):
        self.record = False
        self.files_written = set()
        self.dirs_created = set()

    def record_as_written(self, path):
        if self.record:
            self.files_written.add(path)

    def newer(self, source, target):
        """Tell if the target is newer than the source.

        Returns true if 'source' exists and is more recently modified than
        'target', or if 'source' exists and 'target' doesn't.

        Returns false if both exist and 'target' is the same age or younger
        than 'source'. Raise PackagingFileError if 'source' does not exist.

        Note that this test is not very accurate: files created in the same
        second will have the same "age".
        """
        if not os.path.exists(source):
            raise DistlibException("file '%r' does not exist" %
                                   os.path.abspath(source))
        if not os.path.exists(target):
            return True

        return os.stat(source).st_mtime > os.stat(target).st_mtime

    def copy_file(self, infile, outfile, check=True):
        """Copy a file respecting dry-run and force flags.
        """
        self.ensure_dir(os.path.dirname(outfile))
        logger.info('Copying %s to %s', infile, outfile)
        if not self.dry_run:
            msg = None
            if check:
                if os.path.islink(outfile):
                    msg = '%s is a symlink' % outfile
                elif os.path.exists(outfile) and not os.path.isfile(outfile):
                    msg = '%s is a non-regular file' % outfile
            if msg:
                raise ValueError(msg + ' which would be overwritten')
            shutil.copyfile(infile, outfile)
        self.record_as_written(outfile)

    def copy_stream(self, instream, outfile, encoding=None):
        assert not os.path.isdir(outfile)
        self.ensure_dir(os.path.dirname(outfile))
        logger.info('Copying stream %s to %s', instream, outfile)
        if not self.dry_run:
            if encoding is None:
                outstream = open(outfile, 'wb')
            else:
                outstream = codecs.open(outfile, 'w', encoding=encoding)
            try:
                shutil.copyfileobj(instream, outstream)
            finally:
                outstream.close()
        self.record_as_written(outfile)

    def write_binary_file(self, path, data):
        self.ensure_dir(os.path.dirname(path))
        if not self.dry_run:
            if os.path.exists(path):
                os.remove(path)
            with open(path, 'wb') as f:
                f.write(data)
        self.record_as_written(path)

    def write_text_file(self, path, data, encoding):
        self.write_binary_file(path, data.encode(encoding))

    def set_mode(self, bits, mask, files):
        if os.name == 'posix' or (os.name == 'java' and os._name == 'posix'):
            # Set the executable bits (owner, group, and world) on
            # all the files specified.
            for f in files:
                if self.dry_run:
                    logger.info("changing mode of %s", f)
                else:
                    mode = (os.stat(f).st_mode | bits) & mask
                    logger.info("changing mode of %s to %o", f, mode)
                    os.chmod(f, mode)

    set_executable_mode = lambda s, f: s.set_mode(0o555, 0o7777, f)

    def ensure_dir(self, path):
        path = os.path.abspath(path)
        if path not in self.ensured and not os.path.exists(path):
            self.ensured.add(path)
            d, f = os.path.split(path)
            self.ensure_dir(d)
            logger.info('Creating %s' % path)
            if not self.dry_run:
                os.mkdir(path)
            if self.record:
                self.dirs_created.add(path)

    def byte_compile(self, path, optimize=False, force=False, prefix=None, hashed_invalidation=False):
        dpath = cache_from_source(path, not optimize)
        logger.info('Byte-compiling %s to %s', path, dpath)
        if not self.dry_run:
            if force or self.newer(path, dpath):
                if not prefix:
                    diagpath = None
                else:
                    assert path.startswith(prefix)
                    diagpath = path[len(prefix):]
            compile_kwargs = {}
            if hashed_invalidation and hasattr(py_compile, 'PycInvalidationMode'):
                compile_kwargs['invalidation_mode'] = py_compile.PycInvalidationMode.CHECKED_HASH
            py_compile.compile(path, dpath, diagpath, True, **compile_kwargs)     # raise error
        self.record_as_written(dpath)
        return dpath

    def ensure_removed(self, path):
        if os.path.exists(path):
            if os.path.isdir(path) and not os.path.islink(path):
                logger.debug('Removing directory tree at %s', path)
                if not self.dry_run:
                    shutil.rmtree(path)
                if self.record:
                    if path in self.dirs_created:
                        self.dirs_created.remove(path)
            else:
                if os.path.islink(path):
                    s = 'link'
                else:
                    s = 'file'
                logger.debug('Removing %s %s', s, path)
                if not self.dry_run:
                    os.remove(path)
                if self.record:
                    if path in self.files_written:
                        self.files_written.remove(path)

    def is_writable(self, path):
        result = False
        while not result:
            if os.path.exists(path):
                result = os.access(path, os.W_OK)
                break
            parent = os.path.dirname(path)
            if parent == path:
                break
            path = parent
        return result

    def commit(self):
        """
        Commit recorded changes, turn off recording, return
        changes.
        """
        assert self.record
        result = self.files_written, self.dirs_created
        self._init_record()
        return result

    def rollback(self):
        if not self.dry_run:
            for f in list(self.files_written):
                if os.path.exists(f):
                    os.remove(f)
            # dirs should all be empty now, except perhaps for
            # __pycache__ subdirs
            # reverse so that subdirs appear before their parents
            dirs = sorted(self.dirs_created, reverse=True)
            for d in dirs:
                flist = os.listdir(d)
                if flist:
                    assert flist == ['__pycache__']
                    sd = os.path.join(d, flist[0])
                    os.rmdir(sd)
                os.rmdir(d)     # should fail if non-empty
        self._init_record()

def resolve(module_name, dotted_path):
    if module_name in sys.modules:
        mod = sys.modules[module_name]
    else:
        mod = __import__(module_name)
    if dotted_path is None:
        result = mod
    else:
        parts = dotted_path.split('.')
        result = getattr(mod, parts.pop(0))
        for p in parts:
            result = getattr(result, p)
    return result


class ExportEntry(object):
    def __init__(self, name, prefix, suffix, flags):
        self.name = name
        self.prefix = prefix
        self.suffix = suffix
        self.flags = flags

    @cached_property
    def value(self):
        return resolve(self.prefix, self.suffix)

    def __repr__(self):  # pragma: no cover
        return '<ExportEntry %s = %s:%s %s>' % (self.name, self.prefix,
                                                self.suffix, self.flags)

    def __eq__(self, other):
        if not isinstance(other, ExportEntry):
            result = False
        else:
            result = (self.name == other.name and
                      self.prefix == other.prefix and
                      self.suffix == other.suffix and
                      self.flags == other.flags)
        return result

    __hash__ = object.__hash__


ENTRY_RE = re.compile(r'''(?P<name>(\w|[-.+])+)
                      \s*=\s*(?P<callable>(\w+)([:\.]\w+)*)
                      \s*(\[\s*(?P<flags>[\w-]+(=\w+)?(,\s*\w+(=\w+)?)*)\s*\])?
                      ''', re.VERBOSE)

def get_export_entry(specification):
    m = ENTRY_RE.search(specification)
    if not m:
        result = None
        if '[' in specification or ']' in specification:
            raise DistlibException("Invalid specification "
                                   "'%s'" % specification)
    else:
        d = m.groupdict()
        name = d['name']
        path = d['callable']
        colons = path.count(':')
        if colons == 0:
            prefix, suffix = path, None
        else:
            if colons != 1:
                raise DistlibException("Invalid specification "
                                       "'%s'" % specification)
            prefix, suffix = path.split(':')
        flags = d['flags']
        if flags is None:
            if '[' in specification or ']' in specification:
                raise DistlibException("Invalid specification "
                                       "'%s'" % specification)
            flags = []
        else:
            flags = [f.strip() for f in flags.split(',')]
        result = ExportEntry(name, prefix, suffix, flags)
    return result


def get_cache_base(suffix=None):
    """
    Return the default base location for distlib caches. If the directory does
    not exist, it is created. Use the suffix provided for the base directory,
    and default to '.distlib' if it isn't provided.

    On Windows, if LOCALAPPDATA is defined in the environment, then it is
    assumed to be a directory, and will be the parent directory of the result.
    On POSIX, and on Windows if LOCALAPPDATA is not defined, the user's home
    directory - using os.expanduser('~') - will be the parent directory of
    the result.

    The result is just the directory '.distlib' in the parent directory as
    determined above, or with the name specified with ``suffix``.
    """
    if suffix is None:
        suffix = '.distlib'
    if os.name == 'nt' and 'LOCALAPPDATA' in os.environ:
        result = os.path.expandvars('$localappdata')
    else:
        # Assume posix, or old Windows
        result = os.path.expanduser('~')
    # we use 'isdir' instead of 'exists', because we want to
    # fail if there's a file with that name
    if os.path.isdir(result):
        usable = os.access(result, os.W_OK)
        if not usable:
            logger.warning('Directory exists but is not writable: %s', result)
    else:
        try:
            os.makedirs(result)
            usable = True
        except OSError:
            logger.warning('Unable to create %s', result, exc_info=True)
            usable = False
    if not usable:
        result = tempfile.mkdtemp()
        logger.warning('Default location unusable, using %s', result)
    return os.path.join(result, suffix)


def path_to_cache_dir(path):
    """
    Convert an absolute path to a directory name for use in a cache.

    The algorithm used is:

    #. On Windows, any ``':'`` in the drive is replaced with ``'---'``.
    #. Any occurrence of ``os.sep`` is replaced with ``'--'``.
    #. ``'.cache'`` is appended.
    """
    d, p = os.path.splitdrive(os.path.abspath(path))
    if d:
        d = d.replace(':', '---')
    p = p.replace(os.sep, '--')
    return d + p + '.cache'


def ensure_slash(s):
    if not s.endswith('/'):
        return s + '/'
    return s


def parse_credentials(netloc):
    username = password = None
    if '@' in netloc:
        prefix, netloc = netloc.rsplit('@', 1)
        if ':' not in prefix:
            username = prefix
        else:
            username, password = prefix.split(':', 1)
    if username:
        username = unquote(username)
    if password:
        password = unquote(password)
    return username, password, netloc


def get_process_umask():
    result = os.umask(0o22)
    os.umask(result)
    return result

def is_string_sequence(seq):
    result = True
    i = None
    for i, s in enumerate(seq):
        if not isinstance(s, string_types):
            result = False
            break
    assert i is not None
    return result

PROJECT_NAME_AND_VERSION = re.compile('([a-z0-9_]+([.-][a-z_][a-z0-9_]*)*)-'
                                      '([a-z0-9_.+-]+)', re.I)
PYTHON_VERSION = re.compile(r'-py(\d\.?\d?)')


def split_filename(filename, project_name=None):
    """
    Extract name, version, python version from a filename (no extension)

    Return name, version, pyver or None
    """
    result = None
    pyver = None
    filename = unquote(filename).replace(' ', '-')
    m = PYTHON_VERSION.search(filename)
    if m:
        pyver = m.group(1)
        filename = filename[:m.start()]
    if project_name and len(filename) > len(project_name) + 1:
        m = re.match(re.escape(project_name) + r'\b', filename)
        if m:
            n = m.end()
            result = filename[:n], filename[n + 1:], pyver
    if result is None:
        m = PROJECT_NAME_AND_VERSION.match(filename)
        if m:
            result = m.group(1), m.group(3), pyver
    return result

# Allow spaces in name because of legacy dists like "Twisted Core"
NAME_VERSION_RE = re.compile(r'(?P<name>[\w .-]+)\s*'
                             r'\(\s*(?P<ver>[^\s)]+)\)$')

def parse_name_and_version(p):
    """
    A utility method used to get name and version from a string.

    From e.g. a Provides-Dist value.

    :param p: A value in a form 'foo (1.0)'
    :return: The name and version as a tuple.
    """
    m = NAME_VERSION_RE.match(p)
    if not m:
        raise DistlibException('Ill-formed name/version string: \'%s\'' % p)
    d = m.groupdict()
    return d['name'].strip().lower(), d['ver']

def get_extras(requested, available):
    result = set()
    requested = set(requested or [])
    available = set(available or [])
    if '*' in requested:
        requested.remove('*')
        result |= available
    for r in requested:
        if r == '-':
            result.add(r)
        elif r.startswith('-'):
            unwanted = r[1:]
            if unwanted not in available:
                logger.warning('undeclared extra: %s' % unwanted)
            if unwanted in result:
                result.remove(unwanted)
        else:
            if r not in available:
                logger.warning('undeclared extra: %s' % r)
            result.add(r)
    return result
#
# Extended metadata functionality
#

def _get_external_data(url):
    result = {}
    try:
        # urlopen might fail if it runs into redirections,
        # because of Python issue #13696. Fixed in locators
        # using a custom redirect handler.
        resp = urlopen(url)
        headers = resp.info()
        ct = headers.get('Content-Type')
        if not ct.startswith('application/json'):
            logger.debug('Unexpected response for JSON request: %s', ct)
        else:
            reader = codecs.getreader('utf-8')(resp)
            #data = reader.read().decode('utf-8')
            #result = json.loads(data)
            result = json.load(reader)
    except Exception as e:
        logger.exception('Failed to get external data for %s: %s', url, e)
    return result

_external_data_base_url = 'https://www.red-dove.com/pypi/projects/'

def get_project_data(name):
    url = '%s/%s/project.json' % (name[0].upper(), name)
    url = urljoin(_external_data_base_url, url)
    result = _get_external_data(url)
    return result

def get_package_data(name, version):
    url = '%s/%s/package-%s.json' % (name[0].upper(), name, version)
    url = urljoin(_external_data_base_url, url)
    return _get_external_data(url)


class Cache(object):
    """
    A class implementing a cache for resources that need to live in the file system
    e.g. shared libraries. This class was moved from resources to here because it
    could be used by other modules, e.g. the wheel module.
    """

    def __init__(self, base):
        """
        Initialise an instance.

        :param base: The base directory where the cache should be located.
        """
        # we use 'isdir' instead of 'exists', because we want to
        # fail if there's a file with that name
        if not os.path.isdir(base):  # pragma: no cover
            os.makedirs(base)
        if (os.stat(base).st_mode & 0o77) != 0:
            logger.warning('Directory \'%s\' is not private', base)
        self.base = os.path.abspath(os.path.normpath(base))

    def prefix_to_dir(self, prefix):
        """
        Converts a resource prefix to a directory name in the cache.
        """
        return path_to_cache_dir(prefix)

    def clear(self):
        """
        Clear the cache.
        """
        not_removed = []
        for fn in os.listdir(self.base):
            fn = os.path.join(self.base, fn)
            try:
                if os.path.islink(fn) or os.path.isfile(fn):
                    os.remove(fn)
                elif os.path.isdir(fn):
                    shutil.rmtree(fn)
            except Exception:
                not_removed.append(fn)
        return not_removed


class EventMixin(object):
    """
    A very simple publish/subscribe system.
    """
    def __init__(self):
        self._subscribers = {}

    def add(self, event, subscriber, append=True):
        """
        Add a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be added (and called when the
                           event is published).
        :param append: Whether to append or prepend the subscriber to an
                       existing subscriber list for the event.
        """
        subs = self._subscribers
        if event not in subs:
            subs[event] = deque([subscriber])
        else:
            sq = subs[event]
            if append:
                sq.append(subscriber)
            else:
                sq.appendleft(subscriber)

    def remove(self, event, subscriber):
        """
        Remove a subscriber for an event.

        :param event: The name of an event.
        :param subscriber: The subscriber to be removed.
        """
        subs = self._subscribers
        if event not in subs:
            raise ValueError('No subscribers: %r' % event)
        subs[event].remove(subscriber)

    def get_subscribers(self, event):
        """
        Return an iterator for the subscribers for an event.
        :param event: The event to return subscribers for.
        """
        return iter(self._subscribers.get(event, ()))

    def publish(self, event, *args, **kwargs):
        """
        Publish a event and return a list of values returned by its
        subscribers.

        :param event: The event to publish.
        :param args: The positional arguments to pass to the event's
                     subscribers.
        :param kwargs: The keyword arguments to pass to the event's
                       subscribers.
        """
        result = []
        for subscriber in self.get_subscribers(event):
            try:
                value = subscriber(event, *args, **kwargs)
            except Exception:
                logger.exception('Exception during event publication')
                value = None
            result.append(value)
        logger.debug('publish %s: args = %s, kwargs = %s, result = %s',
                     event, args, kwargs, result)
        return result

#
# Simple sequencing
#
class Sequencer(object):
    def __init__(self):
        self._preds = {}
        self._succs = {}
        self._nodes = set()     # nodes with no preds/succs

    def add_node(self, node):
        self._nodes.add(node)

    def remove_node(self, node, edges=False):
        if node in self._nodes:
            self._nodes.remove(node)
        if edges:
            for p in set(self._preds.get(node, ())):
                self.remove(p, node)
            for s in set(self._succs.get(node, ())):
                self.remove(node, s)
            # Remove empties
            for k, v in list(self._preds.items()):
                if not v:
                    del self._preds[k]
            for k, v in list(self._succs.items()):
                if not v:
                    del self._succs[k]

    def add(self, pred, succ):
        assert pred != succ
        self._preds.setdefault(succ, set()).add(pred)
        self._succs.setdefault(pred, set()).add(succ)

    def remove(self, pred, succ):
        assert pred != succ
        try:
            preds = self._preds[succ]
            succs = self._succs[pred]
        except KeyError:  # pragma: no cover
            raise ValueError('%r not a successor of anything' % succ)
        try:
            preds.remove(pred)
            succs.remove(succ)
        except KeyError:  # pragma: no cover
            raise ValueError('%r not a successor of %r' % (succ, pred))

    def is_step(self, step):
        return (step in self._preds or step in self._succs or
                step in self._nodes)

    def get_steps(self, final):
        if not self.is_step(final):
            raise ValueError('Unknown: %r' % final)
        result = []
        todo = []
        seen = set()
        todo.append(final)
        while todo:
            step = todo.pop(0)
            if step in seen:
                # if a step was already seen,
                # move it to the end (so it will appear earlier
                # when reversed on return) ... but not for the
                # final step, as that would be confusing for
                # users
                if step != final:
                    result.remove(step)
                    result.append(step)
            else:
                seen.add(step)
                result.append(step)
                preds = self._preds.get(step, ())
                todo.extend(preds)
        return reversed(result)

    @property
    def strong_connections(self):
        #http://en.wikipedia.org/wiki/Tarjan%27s_strongly_connected_components_algorithm
        index_counter = [0]
        stack = []
        lowlinks = {}
        index = {}
        result = []

        graph = self._succs

        def strongconnect(node):
            # set the depth index for this node to the smallest unused index
            index[node] = index_counter[0]
            lowlinks[node] = index_counter[0]
            index_counter[0] += 1
            stack.append(node)

            # Consider successors
            try:
                successors = graph[node]
            except Exception:
                successors = []
            for successor in successors:
                if successor not in lowlinks:
                    # Successor has not yet been visited
                    strongconnect(successor)
                    lowlinks[node] = min(lowlinks[node],lowlinks[successor])
                elif successor in stack:
                    # the successor is in the stack and hence in the current
                    # strongly connected component (SCC)
                    lowlinks[node] = min(lowlinks[node],index[successor])

            # If `node` is a root node, pop the stack and generate an SCC
            if lowlinks[node] == index[node]:
                connected_component = []

                while True:
                    successor = stack.pop()
                    connected_component.append(successor)
                    if successor == node: break
                component = tuple(connected_component)
                # storing the result
                result.append(component)

        for node in graph:
            if node not in lowlinks:
                strongconnect(node)

        return result

    @property
    def dot(self):
        result = ['digraph G {']
        for succ in self._preds:
            preds = self._preds[succ]
            for pred in preds:
                result.append('  %s -> %s;' % (pred, succ))
        for node in self._nodes:
            result.append('  %s;' % node)
        result.append('}')
        return '\n'.join(result)

#
# Unarchiving functionality for zip, tar, tgz, tbz, whl
#

ARCHIVE_EXTENSIONS = ('.tar.gz', '.tar.bz2', '.tar', '.zip',
                      '.tgz', '.tbz', '.whl')

def unarchive(archive_filename, dest_dir, format=None, check=True):

    def check_path(path):
        if not isinstance(path, text_type):
            path = path.decode('utf-8')
        p = os.path.abspath(os.path.join(dest_dir, path))
        if not p.startswith(dest_dir) or p[plen] != os.sep:
            raise ValueError('path outside destination: %r' % p)

    dest_dir = os.path.abspath(dest_dir)
    plen = len(dest_dir)
    archive = None
    if format is None:
        if archive_filename.endswith(('.zip', '.whl')):
            format = 'zip'
        elif archive_filename.endswith(('.tar.gz', '.tgz')):
            format = 'tgz'
            mode = 'r:gz'
        elif archive_filename.endswith(('.tar.bz2', '.tbz')):
            format = 'tbz'
            mode = 'r:bz2'
        elif archive_filename.endswith('.tar'):
            format = 'tar'
            mode = 'r'
        else:  # pragma: no cover
            raise ValueError('Unknown format for %r' % archive_filename)
    try:
        if format == 'zip':
            archive = ZipFile(archive_filename, 'r')
            if check:
                names = archive.namelist()
                for name in names:
                    check_path(name)
        else:
            archive = tarfile.open(archive_filename, mode)
            if check:
                names = archive.getnames()
                for name in names:
                    check_path(name)
        if format != 'zip' and sys.version_info[0] < 3:
            # See Python issue 17153. If the dest path contains Unicode,
            # tarfile extraction fails on Python 2.x if a member path name
            # contains non-ASCII characters - it leads to an implicit
            # bytes -> unicode conversion using ASCII to decode.
            for tarinfo in archive.getmembers():
                if not isinstance(tarinfo.name, text_type):
                    tarinfo.name = tarinfo.name.decode('utf-8')
        archive.extractall(dest_dir)

    finally:
        if archive:
            archive.close()


def zip_dir(directory):
    """zip a directory tree into a BytesIO object"""
    result = io.BytesIO()
    dlen = len(directory)
    with ZipFile(result, "w") as zf:
        for root, dirs, files in os.walk(directory):
            for name in files:
                full = os.path.join(root, name)
                rel = root[dlen:]
                dest = os.path.join(rel, name)
                zf.write(full, dest)
    return result

#
# Simple progress bar
#

UNITS = ('', 'K', 'M', 'G','T','P')


class Progress(object):
    unknown = 'UNKNOWN'

    def __init__(self, minval=0, maxval=100):
        assert maxval is None or maxval >= minval
        self.min = self.cur = minval
        self.max = maxval
        self.started = None
        self.elapsed = 0
        self.done = False

    def update(self, curval):
        assert self.min <= curval
        assert self.max is None or curval <= self.max
        self.cur = curval
        now = time.time()
        if self.started is None:
            self.started = now
        else:
            self.elapsed = now - self.started

    def increment(self, incr):
        assert incr >= 0
        self.update(self.cur + incr)

    def start(self):
        self.update(self.min)
        return self

    def stop(self):
        if self.max is not None:
            self.update(self.max)
        self.done = True

    @property
    def maximum(self):
        return self.unknown if self.max is None else self.max

    @property
    def percentage(self):
        if self.done:
            result = '100 %'
        elif self.max is None:
            result = ' ?? %'
        else:
            v = 100.0 * (self.cur - self.min) / (self.max - self.min)
            result = '%3d %%' % v
        return result

    def format_duration(self, duration):
        if (duration <= 0) and self.max is None or self.cur == self.min:
            result = '??:??:??'
        #elif duration < 1:
        #    result = '--:--:--'
        else:
            result = time.strftime('%H:%M:%S', time.gmtime(duration))
        return result

    @property
    def ETA(self):
        if self.done:
            prefix = 'Done'
            t = self.elapsed
            #import pdb; pdb.set_trace()
        else:
            prefix = 'ETA '
            if self.max is None:
                t = -1
            elif self.elapsed == 0 or (self.cur == self.min):
                t = 0
            else:
                #import pdb; pdb.set_trace()
                t = float(self.max - self.min)
                t /= self.cur - self.min
                t = (t - 1) * self.elapsed
        return '%s: %s' % (prefix, self.format_duration(t))

    @property
    def speed(self):
        if self.elapsed == 0:
            result = 0.0
        else:
            result = (self.cur - self.min) / self.elapsed
        for unit in UNITS:
            if result < 1000:
                break
            result /= 1000.0
        return '%d %sB/s' % (result, unit)

#
# Glob functionality
#

RICH_GLOB = re.compile(r'\{([^}]*)\}')
_CHECK_RECURSIVE_GLOB = re.compile(r'[^/\\,{]\*\*|\*\*[^/\\,}]')
_CHECK_MISMATCH_SET = re.compile(r'^[^{]*\}|\{[^}]*$')


def iglob(path_glob):
    """Extended globbing function that supports ** and {opt1,opt2,opt3}."""
    if _CHECK_RECURSIVE_GLOB.search(path_glob):
        msg = """invalid glob %r: recursive glob "**" must be used alone"""
        raise ValueError(msg % path_glob)
    if _CHECK_MISMATCH_SET.search(path_glob):
        msg = """invalid glob %r: mismatching set marker '{' or '}'"""
        raise ValueError(msg % path_glob)
    return _iglob(path_glob)


def _iglob(path_glob):
    rich_path_glob = RICH_GLOB.split(path_glob, 1)
    if len(rich_path_glob) > 1:
        assert len(rich_path_glob) == 3, rich_path_glob
        prefix, set, suffix = rich_path_glob
        for item in set.split(','):
            for path in _iglob(''.join((prefix, item, suffix))):
                yield path
    else:
        if '**' not in path_glob:
            for item in std_iglob(path_glob):
                yield item
        else:
            prefix, radical = path_glob.split('**', 1)
            if prefix == '':
                prefix = '.'
            if radical == '':
                radical = '*'
            else:
                # we support both
                radical = radical.lstrip('/')
                radical = radical.lstrip('\\')
            for path, dir, files in os.walk(prefix):
                path = os.path.normpath(path)
                for fn in _iglob(os.path.join(path, radical)):
                    yield fn

if ssl:
    from .compat import (HTTPSHandler as BaseHTTPSHandler, match_hostname,
                         CertificateError)


#
# HTTPSConnection which verifies certificates/matches domains
#

    class HTTPSConnection(httplib.HTTPSConnection):
        ca_certs = None # set this to the path to the certs file (.pem)
        check_domain = True # only used if ca_certs is not None

        # noinspection PyPropertyAccess
        def connect(self):
            sock = socket.create_connection((self.host, self.port), self.timeout)
            if getattr(self, '_tunnel_host', False):
                self.sock = sock
                self._tunnel()

            if not hasattr(ssl, 'SSLContext'):
                # For 2.x
                if self.ca_certs:
                    cert_reqs = ssl.CERT_REQUIRED
                else:
                    cert_reqs = ssl.CERT_NONE
                self.sock = ssl.wrap_socket(sock, self.key_file, self.cert_file,
                                            cert_reqs=cert_reqs,
                                            ssl_version=ssl.PROTOCOL_SSLv23,
                                            ca_certs=self.ca_certs)
            else:  # pragma: no cover
                context = ssl.SSLContext(ssl.PROTOCOL_SSLv23)
                if hasattr(ssl, 'OP_NO_SSLv2'):
                    context.options |= ssl.OP_NO_SSLv2
                if self.cert_file:
                    context.load_cert_chain(self.cert_file, self.key_file)
                kwargs = {}
                if self.ca_certs:
                    context.verify_mode = ssl.CERT_REQUIRED
                    context.load_verify_locations(cafile=self.ca_certs)
                    if getattr(ssl, 'HAS_SNI', False):
                        kwargs['server_hostname'] = self.host
                self.sock = context.wrap_socket(sock, **kwargs)
            if self.ca_certs and self.check_domain:
                try:
                    match_hostname(self.sock.getpeercert(), self.host)
                    logger.debug('Host verified: %s', self.host)
                except CertificateError:  # pragma: no cover
                    self.sock.shutdown(socket.SHUT_RDWR)
                    self.sock.close()
                    raise

    class HTTPSHandler(BaseHTTPSHandler):
        def __init__(self, ca_certs, check_domain=True):
            BaseHTTPSHandler.__init__(self)
            self.ca_certs = ca_certs
            self.check_domain = check_domain

        def _conn_maker(self, *args, **kwargs):
            """
            This is called to create a connection instance. Normally you'd
            pass a connection class to do_open, but it doesn't actually check for
            a class, and just expects a callable. As long as we behave just as a
            constructor would have, we should be OK. If it ever changes so that
            we *must* pass a class, we'll create an UnsafeHTTPSConnection class
            which just sets check_domain to False in the class definition, and
            choose which one to pass to do_open.
            """
            result = HTTPSConnection(*args, **kwargs)
            if self.ca_certs:
                result.ca_certs = self.ca_certs
                result.check_domain = self.check_domain
            return result

        def https_open(self, req):
            try:
                return self.do_open(self._conn_maker, req)
            except URLError as e:
                if 'certificate verify failed' in str(e.reason):
                    raise CertificateError('Unable to verify server certificate '
                                           'for %s' % req.host)
                else:
                    raise

    #
    # To prevent against mixing HTTP traffic with HTTPS (examples: A Man-In-The-
    # Middle proxy using HTTP listens on port 443, or an index mistakenly serves
    # HTML containing a http://xyz link when it should be https://xyz),
    # you can use the following handler class, which does not allow HTTP traffic.
    #
    # It works by inheriting from HTTPHandler - so build_opener won't add a
    # handler for HTTP itself.
    #
    class HTTPSOnlyHandler(HTTPSHandler, HTTPHandler):
        def http_open(self, req):
            raise URLError('Unexpected HTTP request on what should be a secure '
                           'connection: %s' % req)

#
# XML-RPC with timeouts
#

_ver_info = sys.version_info[:2]

if _ver_info == (2, 6):
    class HTTP(httplib.HTTP):
        def __init__(self, host='', port=None, **kwargs):
            if port == 0:   # 0 means use port 0, not the default port
                port = None
            self._setup(self._connection_class(host, port, **kwargs))


    if ssl:
        class HTTPS(httplib.HTTPS):
            def __init__(self, host='', port=None, **kwargs):
                if port == 0:   # 0 means use port 0, not the default port
                    port = None
                self._setup(self._connection_class(host, port, **kwargs))


class Transport(xmlrpclib.Transport):
    def __init__(self, timeout, use_datetime=0):
        self.timeout = timeout
        xmlrpclib.Transport.__init__(self, use_datetime)

    def make_connection(self, host):
        h, eh, x509 = self.get_host_info(host)
        if _ver_info == (2, 6):
            result = HTTP(h, timeout=self.timeout)
        else:
            if not self._connection or host != self._connection[0]:
                self._extra_headers = eh
                self._connection = host, httplib.HTTPConnection(h)
            result = self._connection[1]
        return result

if ssl:
    class SafeTransport(xmlrpclib.SafeTransport):
        def __init__(self, timeout, use_datetime=0):
            self.timeout = timeout
            xmlrpclib.SafeTransport.__init__(self, use_datetime)

        def make_connection(self, host):
            h, eh, kwargs = self.get_host_info(host)
            if not kwargs:
                kwargs = {}
            kwargs['timeout'] = self.timeout
            if _ver_info == (2, 6):
                result = HTTPS(host, None, **kwargs)
            else:
                if not self._connection or host != self._connection[0]:
                    self._extra_headers = eh
                    self._connection = host, httplib.HTTPSConnection(h, None,
                                                                     **kwargs)
                result = self._connection[1]
            return result


class ServerProxy(xmlrpclib.ServerProxy):
    def __init__(self, uri, **kwargs):
        self.timeout = timeout = kwargs.pop('timeout', None)
        # The above classes only come into play if a timeout
        # is specified
        if timeout is not None:
            # scheme = splittype(uri)  # deprecated as of Python 3.8
            scheme = urlparse(uri)[0]
            use_datetime = kwargs.get('use_datetime', 0)
            if scheme == 'https':
                tcls = SafeTransport
            else:
                tcls = Transport
            kwargs['transport'] = t = tcls(timeout, use_datetime=use_datetime)
            self.transport = t
        xmlrpclib.ServerProxy.__init__(self, uri, **kwargs)

#
# CSV functionality. This is provided because on 2.x, the csv module can't
# handle Unicode. However, we need to deal with Unicode in e.g. RECORD files.
#

def _csv_open(fn, mode, **kwargs):
    if sys.version_info[0] < 3:
        mode += 'b'
    else:
        kwargs['newline'] = ''
        # Python 3 determines encoding from locale. Force 'utf-8'
        # file encoding to match other forced utf-8 encoding
        kwargs['encoding'] = 'utf-8'
    return open(fn, mode, **kwargs)


class CSVBase(object):
    defaults = {
        'delimiter': str(','),      # The strs are used because we need native
        'quotechar': str('"'),      # str in the csv API (2.x won't take
        'lineterminator': str('\n') # Unicode)
    }

    def __enter__(self):
        return self

    def __exit__(self, *exc_info):
        self.stream.close()


class CSVReader(CSVBase):
    def __init__(self, **kwargs):
        if 'stream' in kwargs:
            stream = kwargs['stream']
            if sys.version_info[0] >= 3:
                # needs to be a text stream
                stream = codecs.getreader('utf-8')(stream)
            self.stream = stream
        else:
            self.stream = _csv_open(kwargs['path'], 'r')
        self.reader = csv.reader(self.stream, **self.defaults)

    def __iter__(self):
        return self

    def next(self):
        result = next(self.reader)
        if sys.version_info[0] < 3:
            for i, item in enumerate(result):
                if not isinstance(item, text_type):
                    result[i] = item.decode('utf-8')
        return result

    __next__ = next

class CSVWriter(CSVBase):
    def __init__(self, fn, **kwargs):
        self.stream = _csv_open(fn, 'w')
        self.writer = csv.writer(self.stream, **self.defaults)

    def writerow(self, row):
        if sys.version_info[0] < 3:
            r = []
            for item in row:
                if isinstance(item, text_type):
                    item = item.encode('utf-8')
                r.append(item)
            row = r
        self.writer.writerow(row)

#
#   Configurator functionality
#

class Configurator(BaseConfigurator):

    value_converters = dict(BaseConfigurator.value_converters)
    value_converters['inc'] = 'inc_convert'

    def __init__(self, config, base=None):
        super(Configurator, self).__init__(config)
        self.base = base or os.getcwd()

    def configure_custom(self, config):
        def convert(o):
            if isinstance(o, (list, tuple)):
                result = type(o)([convert(i) for i in o])
            elif isinstance(o, dict):
                if '()' in o:
                    result = self.configure_custom(o)
                else:
                    result = {}
                    for k in o:
                        result[k] = convert(o[k])
            else:
                result = self.convert(o)
            return result

        c = config.pop('()')
        if not callable(c):
            c = self.resolve(c)
        props = config.pop('.', None)
        # Check for valid identifiers
        args = config.pop('[]', ())
        if args:
            args = tuple([convert(o) for o in args])
        items = [(k, convert(config[k])) for k in config if valid_ident(k)]
        kwargs = dict(items)
        result = c(*args, **kwargs)
        if props:
            for n, v in props.items():
                setattr(result, n, convert(v))
        return result

    def __getitem__(self, key):
        result = self.config[key]
        if isinstance(result, dict) and '()' in result:
            self.config[key] = result = self.configure_custom(result)
        return result

    def inc_convert(self, value):
        """Default converter for the inc:// protocol."""
        if not os.path.isabs(value):
            value = os.path.join(self.base, value)
        with codecs.open(value, 'r', encoding='utf-8') as f:
            result = json.load(f)
        return result


class SubprocessMixin(object):
    """
    Mixin for running subprocesses and capturing their output
    """
    def __init__(self, verbose=False, progress=None):
        self.verbose = verbose
        self.progress = progress

    def reader(self, stream, context):
        """
        Read lines from a subprocess' output stream and either pass to a progress
        callable (if specified) or write progress information to sys.stderr.
        """
        progress = self.progress
        verbose = self.verbose
        while True:
            s = stream.readline()
            if not s:
                break
            if progress is not None:
                progress(s, context)
            else:
                if not verbose:
                    sys.stderr.write('.')
                else:
                    sys.stderr.write(s.decode('utf-8'))
                sys.stderr.flush()
        stream.close()

    def run_command(self, cmd, **kwargs):
        p = subprocess.Popen(cmd, stdout=subprocess.PIPE,
                             stderr=subprocess.PIPE, **kwargs)
        t1 = threading.Thread(target=self.reader, args=(p.stdout, 'stdout'))
        t1.start()
        t2 = threading.Thread(target=self.reader, args=(p.stderr, 'stderr'))
        t2.start()
        p.wait()
        t1.join()
        t2.join()
        if self.progress is not None:
            self.progress('done.', 'main')
        elif self.verbose:
            sys.stderr.write('done.\n')
        return p


def normalize_name(name):
    """Normalize a python package name a la PEP 503"""
    # https://www.python.org/dev/peps/pep-0503/#normalized-names
    return re.sub('[-_.]+', '-', name).lower()

# def _get_pypirc_command():
    # """
    # Get the distutils command for interacting with PyPI configurations.
    # :return: the command.
    # """
    # from distutils.core import Distribution
    # from distutils.config import PyPIRCCommand
    # d = Distribution()
    # return PyPIRCCommand(d)

class PyPIRCFile(object):

    DEFAULT_REPOSITORY = 'https://upload.pypi.org/legacy/'
    DEFAULT_REALM = 'pypi'

    def __init__(self, fn=None, url=None):
        if fn is None:
            fn = os.path.join(os.path.expanduser('~'), '.pypirc')
        self.filename = fn
        self.url = url

    def read(self):
        result = {}

        if os.path.exists(self.filename):
            repository = self.url or self.DEFAULT_REPOSITORY

            config = configparser.RawConfigParser()
            config.read(self.filename)
            sections = config.sections()
            if 'distutils' in sections:
                # let's get the list of servers
                index_servers = config.get('distutils', 'index-servers')
                _servers = [server.strip() for server in
                            index_servers.split('\n')
                            if server.strip() != '']
                if _servers == []:
                    # nothing set, let's try to get the default pypi
                    if 'pypi' in sections:
                        _servers = ['pypi']
                else:
                    for server in _servers:
                        result = {'server': server}
                        result['username'] = config.get(server, 'username')

                        # optional params
                        for key, default in (('repository', self.DEFAULT_REPOSITORY),
                                             ('realm', self.DEFAULT_REALM),
                                             ('password', None)):
                            if config.has_option(server, key):
                                result[key] = config.get(server, key)
                            else:
                                result[key] = default

                        # work around people having "repository" for the "pypi"
                        # section of their config set to the HTTP (rather than
                        # HTTPS) URL
                        if (server == 'pypi' and
                            repository in (self.DEFAULT_REPOSITORY, 'pypi')):
                            result['repository'] = self.DEFAULT_REPOSITORY
                        elif (result['server'] != repository and
                              result['repository'] != repository):
                            result = {}
            elif 'server-login' in sections:
                # old format
                server = 'server-login'
                if config.has_option(server, 'repository'):
                    repository = config.get(server, 'repository')
                else:
                    repository = self.DEFAULT_REPOSITORY
                result = {
                    'username': config.get(server, 'username'),
                    'password': config.get(server, 'password'),
                    'repository': repository,
                    'server': server,
                    'realm': self.DEFAULT_REALM
                }
        return result

    def update(self, username, password):
        # import pdb; pdb.set_trace()
        config = configparser.RawConfigParser()
        fn = self.filename
        config.read(fn)
        if not config.has_section('pypi'):
            config.add_section('pypi')
        config.set('pypi', 'username', username)
        config.set('pypi', 'password', password)
        with open(fn, 'w') as f:
            config.write(f)

def _load_pypirc(index):
    """
    Read the PyPI access configuration as supported by distutils.
    """
    return PyPIRCFile(url=index.url).read()

def _store_pypirc(index):
    PyPIRCFile().update(index.username, index.password)

#
# get_platform()/get_host_platform() copied from Python 3.10.a0 source, with some minor
# tweaks
#

def get_host_platform():
    """Return a string that identifies the current platform.  This is used mainly to
    distinguish platform-specific build directories and platform-specific built
    distributions.  Typically includes the OS name and version and the
    architecture (as supplied by 'os.uname()'), although the exact information
    included depends on the OS; eg. on Linux, the kernel version isn't
    particularly important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.

    """
    if os.name == 'nt':
        if 'amd64' in sys.version.lower():
            return 'win-amd64'
        if '(arm)' in sys.version.lower():
            return 'win-arm32'
        if '(arm64)' in sys.version.lower():
            return 'win-arm64'
        return sys.platform

    # Set for cross builds explicitly
    if "_PYTHON_HOST_PLATFORM" in os.environ:
        return os.environ["_PYTHON_HOST_PLATFORM"]

    if os.name != 'posix' or not hasattr(os, 'uname'):
        # XXX what about the architecture? NT is Intel or Alpha,
        # Mac OS is M68k or PPC, etc.
        return sys.platform

    # Try to distinguish various flavours of Unix

    (osname, host, release, version, machine) = os.uname()

    # Convert the OS name to lowercase, remove '/' characters, and translate
    # spaces (for "Power Macintosh")
    osname = osname.lower().replace('/', '')
    machine = machine.replace(' ', '_').replace('/', '-')

    if osname[:5] == 'linux':
        # At least on Linux/Intel, 'machine' is the processor --
        # i386, etc.
        # XXX what about Alpha, SPARC, etc?
        return  "%s-%s" % (osname, machine)

    elif osname[:5] == 'sunos':
        if release[0] >= '5':           # SunOS 5 == Solaris 2
            osname = 'solaris'
            release = '%d.%s' % (int(release[0]) - 3, release[2:])
            # We can't use 'platform.architecture()[0]' because a
            # bootstrap problem. We use a dict to get an error
            # if some suspicious happens.
            bitness = {2147483647:'32bit', 9223372036854775807:'64bit'}
            machine += '.%s' % bitness[sys.maxsize]
        # fall through to standard osname-release-machine representation
    elif osname[:3] == 'aix':
        from _aix_support import aix_platform
        return aix_platform()
    elif osname[:6] == 'cygwin':
        osname = 'cygwin'
        rel_re = re.compile (r'[\d.]+', re.ASCII)
        m = rel_re.match(release)
        if m:
            release = m.group()
    elif osname[:6] == 'darwin':
        import _osx_support, distutils.sysconfig
        osname, release, machine = _osx_support.get_platform_osx(
                                        distutils.sysconfig.get_config_vars(),
                                        osname, release, machine)

    return '%s-%s-%s' % (osname, release, machine)


_TARGET_TO_PLAT = {
    'x86' : 'win32',
    'x64' : 'win-amd64',
    'arm' : 'win-arm32',
}


def get_platform():
    if os.name != 'nt':
        return get_host_platform()
    cross_compilation_target = os.environ.get('VSCMD_ARG_TGT_ARCH')
    if cross_compilation_target not in _TARGET_TO_PLAT:
        return get_host_platform()
    return _TARGET_TO_PLAT[cross_compilation_target]
site-packages/pip/_vendor/distlib/_backport/__pycache__/shutil.cpython-39.pyc000064400000052324150564745010023274 0ustar00a

��Rekd�@sdZddlZddlZddlZddlmZddlZzddlmZWne	y^ddl
mZYn0ddlZddlm
Z
zddlZdZWne	y�dZYn0zdd	lmZWne	y�dZYn0zdd
lmZWne	y�dZYn0gd�ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZzeWne�yddZYn0dfdd�Zdd�Zdd�Z dd�Z!dd �Z"d!d"�Z#d#d$�Z$d%d&�Z%dde$dfd'd(�Z&dgd)d*�Z'd+d,�Z(d-d.�Z)d/d0�Z*d1d2�Z+d3d4�Z,dhd6d7�Z-did8d9�Z.djd:d;�Z/e-d<gd=fe-d>gd?fe-d@gdAfe/gdBfdC�Z0e�rLe-d>gd?fe0dD<dEdF�Z1dkdHdI�Z2dJdK�Z3dldLdM�Z4dNdO�Z5dPdQ�Z6dmdRdS�Z7dTdU�Z8dVdW�Z9dXdY�Z:dZd[�Z;d\d]ge;gd=fd^ge;gdAfd_ge:gdBfd`�Z<e�r�dage;gd?fe<dD<dbdc�Z=dnddde�Z>dS)oz�Utility functions for copying and archiving files and directory trees.

XXX The functions here don't copy the resource fork or other metadata on Mac.

�N)�abspath)�Callable�)�tarfileTF)�getpwnam)�getgrnam)�copyfileobj�copyfile�copymode�copystat�copy�copy2�copytree�move�rmtree�Error�SpecialFileError�	ExecError�make_archive�get_archive_formats�register_archive_format�unregister_archive_format�get_unpack_formats�register_unpack_format�unregister_unpack_format�unpack_archive�ignore_patternsc@seZdZdS)rN)�__name__�
__module__�__qualname__�r r ��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/_backport/shutil.pyr/src@seZdZdZdS)rz|Raised when trying to do a kind of operation (e.g. copying) which is
    not supported on a special file (e.g. a named pipe)N�rrr�__doc__r r r r!r2src@seZdZdZdS)rz+Raised when a command could not be executedNr"r r r r!r6src@seZdZdZdS)�	ReadErrorz%Raised when an archive cannot be readNr"r r r r!r$9sr$c@seZdZdZdS)�
RegistryErrorzVRaised when a registry operation with the archiving
    and unpacking registries failsNr"r r r r!r%<sr%�@cCs |�|�}|sq|�|�qdS)z=copy data from file-like object fsrc to file-like object fdstN)�read�write)�fsrc�fdst�length�bufr r r!rFs
rcCsZttjd�r2ztj�||�WSty0YdS0tj�tj�|��tj�tj�|��kS)N�samefileF)�hasattr�os�pathr-�OSError�normcaser��src�dstr r r!�	_samefileNs�r6c	Cs�t||�rtd||f��||fD]>}zt�|�}WntyFYq"0t�|j�r"td|��q"t|d��D}t|d��}t	||�Wd�n1s�0YWd�n1s�0YdS)zCopy data from src to dstz`%s` and `%s` are the same filez`%s` is a named pipe�rb�wbN)
r6rr/�statr1�S_ISFIFO�st_moder�openr)r4r5�fn�str)r*r r r!r	Zs
r	cCs0ttd�r,t�|�}t�|j�}t�||�dS)zCopy mode bits from src to dst�chmodN)r.r/r9�S_IMODEr;r?)r4r5r>�moder r r!r
ns

r
c
Cs�t�|�}t�|j�}ttd�r4t�||j|jf�ttd�rJt�||�ttd�r�t|d�r�zt�	||j
�Wn<ty�}z$ttd�r�|jtj
kr��WYd}~n
d}~00dS)zCCopy all stat info (mode bits, atime, mtime, flags) from src to dst�utimer?�chflags�st_flags�
EOPNOTSUPPN)r/r9r@r;r.rB�st_atime�st_mtimer?rCrDr1�errnorE)r4r5r>rA�whyr r r!rus




�rcCs:tj�|�r"tj�|tj�|��}t||�t||�dS)zVCopy data and mode bits ("cp src dst").

    The destination may be a directory.

    N)r/r0�isdir�join�basenamer	r
r3r r r!r�s
rcCs:tj�|�r"tj�|tj�|��}t||�t||�dS)z]Copy data and all stat info ("cp -p src dst").

    The destination may be a directory.

    N)r/r0rJrKrLr	rr3r r r!r
�s
r
cs�fdd�}|S)z�Function that can be used as copytree() ignore parameter.

    Patterns is a sequence of glob-style patterns
    that are used to exclude filescs(g}�D]}|�t�||��qt|�S�N)�extend�fnmatch�filter�set)r0�names�
ignored_names�pattern��patternsr r!�_ignore_patterns�sz)ignore_patterns.<locals>._ignore_patternsr )rVrWr rUr!r�srcCs�t�|�}|dur|||�}nt�}t�|�g}|D�]}	|	|vrFq6tj�||	�}
tj�||	�}zttj�|
�r�t�|
�}|r�t�||�q�tj�	|�s�|r�Wq6||
|�n(tj�
|
�r�t|
||||�n
||
|�Wq6t�y}
z|�
|
jd�WYd}
~
q6d}
~
0t�yD}z |�|
|t|�f�WYd}~q6d}~00q6zt||�WnRt�y�}z8tdu�r�t|t��r�n|�
||t|�f�WYd}~n
d}~00|�r�t|��dS)a�Recursively copy a directory tree.

    The destination directory must not already exist.
    If exception(s) occur, an Error is raised with a list of reasons.

    If the optional symlinks flag is true, symbolic links in the
    source tree result in symbolic links in the destination tree; if
    it is false, the contents of the files pointed to by symbolic
    links are copied. If the file pointed by the symlink doesn't
    exist, an exception will be added in the list of errors raised in
    an Error exception at the end of the copy process.

    You can set the optional ignore_dangling_symlinks flag to true if you
    want to silence this exception. Notice that this has no effect on
    platforms that don't support os.symlink.

    The optional ignore argument is a callable. If given, it
    is called with the `src` parameter, which is the directory
    being visited by copytree(), and `names` which is the list of
    `src` contents, as returned by os.listdir():

        callable(src, names) -> ignored_names

    Since copytree() is called recursively, the callable will be
    called once for each directory that is copied. It returns a
    list of names relative to the `src` directory that should
    not be copied.

    The optional copy_function argument is a callable that will be used
    to copy each file. It will be called with the source path and the
    destination path as arguments. By default, copy2() is used, but any
    function that supports the same signature (like copy()) can be used.

    Nr)r/�listdirrQ�makedirsr0rK�islink�readlink�symlink�existsrJrrrN�args�EnvironmentError�append�strrr1�WindowsError�
isinstance)r4r5�symlinks�ignore�
copy_function�ignore_dangling_symlinksrRrS�errors�name�srcname�dstname�linkto�errrIr r r!r�sD$



$,*rc	Csh|rdd�}n|durdd�}ztj�|�r4td��Wn(ty^|tjj|t���YdS0g}zt�|�}Wn&tjy�|tj|t���Yn0|D]�}tj�||�}zt�	|�j
}Wntjy�d}Yn0t�|�r�t
|||�q�zt�|�Wq�tj�y(|tj|t���Yq�0q�zt�|�Wn(tj�yb|tj|t���Yn0dS)a�Recursively delete a directory tree.

    If ignore_errors is set, errors are ignored; otherwise, if onerror
    is set, it is called to handle the error with arguments (func,
    path, exc_info) where func is os.listdir, os.remove, or os.rmdir;
    path is the argument to that function that caused it to fail; and
    exc_info is a tuple returned by sys.exc_info().  If ignore_errors
    is false and onerror is None, an exception is raised.

    cWsdSrMr �r^r r r!�onerrorszrmtree.<locals>.onerrorNcWs�dSrMr rnr r r!rosz%Cannot call rmtree on a symbolic linkr)r/r0rZr1�sys�exc_inforX�errorrK�lstatr;r9�S_ISDIRr�remove�rmdir)r0�
ignore_errorsrorRri�fullnamerAr r r!r�s>


rcCstj�|�tjj��SrM)r/r0rL�rstrip�sep)r0r r r!�	_basename*sr{cCs�|}tj�|�rTt||�r*t�||�dStj�|t|��}tj�|�rTtd|��zt�||�Wndt	y�tj�|�r�t
||�r�td||f��t||dd�t|�nt
||�t�|�Yn0dS)a�Recursively move a file or directory to another location. This is
    similar to the Unix "mv" command.

    If the destination is a directory or a symlink to a directory, the source
    is moved inside the directory. The destination path must not already
    exist.

    If the destination already exists but is not a directory, it may be
    overwritten depending on os.rename() semantics.

    If the destination is on our current filesystem, then rename() is used.
    Otherwise, src is copied to the destination and then removed.
    A lot more could be done here...  A look at a mv.c shows a lot of
    the issues this implementation glosses over.

    Nz$Destination path '%s' already existsz.Cannot move a directory '%s' into itself '%s'.T)rd)r/r0rJr6�renamerKr{r]rr1�
_destinsrcrrr
�unlink)r4r5�real_dstr r r!r/s$



rcCsNt|�}t|�}|�tjj�s*|tjj7}|�tjj�sD|tjj7}|�|�SrM)r�endswithr/r0rz�
startswithr3r r r!r}Wsr}cCsLtdus|durdSzt|�}Wnty6d}Yn0|durH|dSdS)z"Returns a gid, given a group name.N�)r�KeyError�ri�resultr r r!�_get_gid`s
r�cCsLtdus|durdSzt|�}Wnty6d}Yn0|durH|dSdS)z"Returns an uid, given a user name.Nr�)rr�r�r r r!�_get_uidls
r��gzipcs
ddd�}ddi}	tr&d|d<d|	d<|d	urD||	vrDtd
�|���|d|	�|d�}
tj�|
�}tj�|�s�|d	ur�|�d|�|s�t�	|�|d	ur�|�d
�t
���t�������fdd�}|�st�
|
d||�}
z|
j||d�W|
��n
|
��0|
S)a�Create a (possibly compressed) tar file from all the files under
    'base_dir'.

    'compress' must be "gzip" (the default), "bzip2", or None.

    'owner' and 'group' can be used to define an owner and a group for the
    archive that is being built. If not provided, the current owner and group
    will be used.

    The output tar file will be named 'base_name' +  ".tar", possibly plus
    the appropriate compression extension (".gz", or ".bz2").

    Returns the output filename.
    �gz�)r�Nr�z.gz�bz2�bzip2�.bz2NzCbad value for 'compress', or compression format not supported : {0}�.tar�creating %szCreating tar archivecs,�dur�|_�|_�dur(�|_�|_|SrM)�gid�gname�uid�uname)�tarinfo�r��group�ownerr�r r!�_set_uid_gid�sz#_make_tarball.<locals>._set_uid_gidzw|%s)rP)�_BZ2_SUPPORTED�
ValueError�format�getr/r0�dirnamer]�inforYr�r�rr<�add�close)�	base_name�base_dir�compress�verbose�dry_runr�r��logger�tar_compression�compress_ext�archive_name�archive_dirr��tarr r�r!�
_make_tarballxs6
�

	r�cCsb|r
d}nd}ddlm}ddlm}z|d|||g|d�Wn|y\td�|�Yn0dS)	Nz-rz-rqr)�DistutilsExecError)�spawn�zip)r�zkunable to create zip file '%s': could neither import the 'zipfile' module nor find a standalone zip utility)�distutils.errorsr��distutils.spawnr�r)r��zip_filenamer�r��
zipoptionsr�r�r r r!�_call_external_zip�s�r�cCs|d}tj�|�}tj�|�sB|dur4|�d|�|sBt�|�zddl}Wntydd}Yn0|dur~t||||�n�|dur�|�d||�|�s|j	|d|j
d�}t�|�D]V\}	}
}|D]F}tj�tj�
|	|��}
tj�|
�r�|�|
|
�|dur�|�d|
�q�q�|��|S)	amCreate a zip file from all the files under 'base_dir'.

    The output zip file will be named 'base_name' + ".zip".  Uses either the
    "zipfile" Python module (if available) or the InfoZIP "zip" utility
    (if installed and found on the default search path).  If neither tool is
    available, raises ExecError.  Returns the name of the output zip
    file.
    �.zipNr�rz#creating '%s' and adding '%s' to it�w)�compressionzadding '%s')r/r0r�r]r�rY�zipfile�ImportErrorr��ZipFile�ZIP_DEFLATED�walk�normpathrK�isfiler(r�)r�r�r�r�r�r�r�r�r��dirpath�dirnames�	filenamesrir0r r r!�
_make_zipfile�s<	

��r�)r�r�zgzip'ed tar-file)r�r�zbzip2'ed tar-file)r�Nzuncompressed tar filezZIP file)�gztar�bztarr�r�r�cCsdd�t��D�}|��|S)z�Returns a list of supported formats for archiving and unarchiving.

    Each element of the returned sequence is a tuple (name, description)
    cSsg|]\}}||df�qS)r�r )�.0ri�registryr r r!�
<listcomp>�z'get_archive_formats.<locals>.<listcomp>)�_ARCHIVE_FORMATS�items�sort��formatsr r r!rs
�rr�cCsv|durg}t|t�s"td|��t|ttf�s8td��|D]&}t|ttf�rZt|�dkr<td��q<|||ft|<dS)auRegisters an archive format.

    name is the name of the format. function is the callable that will be
    used to create archives. If provided, extra_args is a sequence of
    (name, value) tuples that will be passed as arguments to the callable.
    description can be provided to describe the format, and will be returned
    by the get_archive_formats() function.
    NzThe %s object is not callablez!extra_args needs to be a sequencer�z+extra_args elements are : (arg_name, value))rcr�	TypeError�tuple�list�lenr�)ri�function�
extra_args�description�elementr r r!rs	

rcCs
t|=dSrM)r��rir r r!r src	
Cs*t��}	|dur>|dur$|�d|�tj�|�}|s>t�|�|durLtj}||d�}
zt|}Wnty�t	d|��Yn0|d}|dD]\}
}||
|
<q�|dkr�||
d<||
d	<z<|||fi|
��}W|dur�|dur�|�d
|	�t�|	�n,|du�r$|du�r|�d
|	�t�|	�0|S)a�Create an archive file (eg. zip or tar).

    'base_name' is the name of the file to create, minus any format-specific
    extension; 'format' is the archive format: one of "zip", "tar", "bztar"
    or "gztar".

    'root_dir' is a directory that will be the root directory of the
    archive; ie. we typically chdir into 'root_dir' before creating the
    archive.  'base_dir' is the directory where we start archiving from;
    ie. 'base_dir' will be the common prefix of all files and
    directories in the archive.  'root_dir' and 'base_dir' both default
    to the current directory.  Returns the name of the archive file.

    'owner' and 'group' are used when creating a tar archive. By default,
    uses the current owner and group.
    Nzchanging into '%s')r�r�zunknown archive format '%s'rrr�r�r�zchanging back to '%s')
r/�getcwd�debugr0r�chdir�curdirr�r�r�)r�r��root_dirr�r�r�r�r�r��save_cwd�kwargs�format_info�func�arg�val�filenamer r r!r#s>


�

rcCsdd�t��D�}|��|S)z�Returns a list of supported formats for unpacking.

    Each element of the returned sequence is a tuple
    (name, extensions, description)
    cSs"g|]\}}||d|df�qS)r�r )r�rir�r r r!r�`r�z&get_unpack_formats.<locals>.<listcomp>)�_UNPACK_FORMATSr�r�r�r r r!rZs
�rc	Csli}t��D]\}}|dD]}|||<qq|D]$}||vr0d}t||||f��q0t|t�shtd��dS)z+Checks what gets registered as an unpacker.rz!%s is already registered for "%s"z*The registered function must be a callableN)r�r�r%rcrr�)	�
extensionsr�r��existing_extensionsrir��ext�	extension�msgr r r!�_check_unpack_optionses�

r�cCs,|durg}t|||�||||ft|<dS)aMRegisters an unpack format.

    `name` is the name of the format. `extensions` is a list of extensions
    corresponding to the format.

    `function` is the callable that will be
    used to unpack archives. The callable will receive archives to unpack.
    If it's unable to handle an archive, it needs to raise a ReadError
    exception.

    If provided, `extra_args` is a sequence of
    (name, value) tuples that will be passed as arguments to the callable.
    description can be provided to describe the format, and will be returned
    by the get_unpack_formats() function.
    N)r�r�)rir�r�r�r�r r r!rwsrcCs
t|=dS)z*Removes the pack format from the registry.N)r�r�r r r!r�srcCs&tj�|�}tj�|�s"t�|�dS)z1Ensure that the parent directory of `path` existsN)r/r0r�rJrY)r0r�r r r!�_ensure_directory�sr�c		Cs�zddl}Wnty&td��Yn0|�|�s>td|��|�|�}z�|��D]�}|j}|�d�sRd|vrpqRtj	j
|g|�d��R�}|s�qRt|�|�
d�sR|�|j�}t|d�}z|�|�W|��~qR|��~0qRW|��n
|��0dS)z+Unpack zip `filename` to `extract_dir`
    rNz/zlib not supported, cannot unpack this archive.z%s is not a zip file�/z..r8)r�r�r$�
is_zipfiler��infolistr�r�r/r0rK�splitr�r�r'r<r(r�)	r��extract_dirr�r�r�ri�target�data�fr r r!�_unpack_zipfile�s4



�r�cCsVzt�|�}Wn tjy.td|��Yn0z|�|�W|��n
|��0dS)z:Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir`
    z/%s is not a compressed or uncompressed tar fileN)rr<�TarErrorr$�
extractallr�)r�r��tarobjr r r!�_unpack_tarfile�s�
r�z.tar.gzz.tgzr�r�)r�r�r�r�cCs:t��D],\}}|dD]}|�|�r|SqqdS)Nr)r�r�r�)r�rir�r�r r r!�_find_unpack_format�s

r�cCs�|durt��}|durjzt|}Wn tyDtd�|���Yn0|d}|||fit|d���nLt|�}|dur�td�|���t|d}tt|d�}|||fi|��dS)a�Unpack an archive.

    `filename` is the name of the archive.

    `extract_dir` is the name of the target directory, where the archive
    is unpacked. If not provided, the current working directory is used.

    `format` is the archive format: one of "zip", "tar", or "gztar". Or any
    other registered format. If not provided, unpack_archive will use the
    filename extension and see if an unpacker was registered for that
    extension.

    In case none is found, a ValueError is raised.
    NzUnknown unpack format '{0}'rr�zUnknown archive format '{0}')	r/r�r�r�r�r��dictr�r$)r�r�r�r�r�r�r r r!r�sr)r&)FN)r�rrNNN)FF)rrN)Nr�)NNrrNNN)Nr�)NN)?r#r/rpr9�os.pathrrO�collections.abcrr��collectionsrHr�rr�r��pwdr�grpr�__all__r_rrrr$�	Exceptionr%rb�	NameErrorrr6r	r
rrr
rrrr{rr}r�r�r�r�r�r�rrrrrr�rrr�r�r�r�r�rr r r r!�<module>s�




�
R
1(	�
>

0


��

�
7�
%��site-packages/pip/_vendor/distlib/_backport/__pycache__/tarfile.cpython-39.pyc000064400000172462150564745010023420 0ustar00a

��Re�i�
@sFddlmZdZdZdZdZdZdZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZzddlZddlZWney�dZZYn0eefZzeef7ZWney�Yn0gd	�Zejdd
kr�ddlZnddlZejZdZdZed
Z dZ!dZ"dZ#dZ$dZ%dZ&dZ'dZ(dZ)dZ*dZ+dZ,dZ-dZ.dZ/dZ0dZ1dZ2dZ3dZ4dZ5d Z6d!Z7e6Z8e&e'e(e)e,e-e.e*e+e/e0e1fZ9e&e'e.e1fZ:e/e0e1fZ;d"Z<e=d#�Z>e?e?e?e@e@e@d$�ZAd%ZBd&ZCd'ZDd(ZEd)ZFd*ZGd+ZHd,ZIdZJd-ZKd.ZLd/ZMd0ZNd1ZOd2ZPd3ZQd!ZRd ZSe	jTd4v�rd5ZUne�V�ZUd6d7�ZWd8d9�ZXd:d;�ZYd2e8fd<d=�ZZd>d?�Z[d{d@dA�Z\eBdBfeCdCfeDdDfeEdEfeFdFfeGdGffeKdHffeLdIffeMeHBdJfeHdKfeMdLffeNdHffeOdIffePeIBdJfeIdKfePdLffeQdHffeRdIffeSeJBdMfeJdNfeSdLfff
Z]dOdP�Z^GdQdR�dRe_�Z`GdSdT�dTe`�ZaGdUdV�dVe`�ZbGdWdX�dXe`�ZcGdYdZ�dZe`�ZdGd[d\�d\e`�ZeGd]d^�d^ee�ZfGd_d`�d`ee�ZgGdadb�dbee�ZhGdcdd�ddee�ZiGdedf�dfee�ZjGdgdh�dhek�ZlGdidj�djek�ZmGdkdl�dlek�ZnGdmdn�dnek�ZoGdodp�dpek�ZpGdqdr�drek�ZqGdsdt�dtek�ZrGdudv�dvek�ZsGdwdx�dxek�Ztdydz�ZueZvesjZdS)|�)�print_functionz
$Revision$z0.9.0u"Lars Gustäbel (lars@gustaebel.de)z5$Date: 2011-02-25 17:42:01 +0200 (Fri, 25 Feb 2011) $z?$Id: tarfile.py 88586 2011-02-25 15:42:01Z marc-andre.lemburg $u4Gustavo Niemeyer, Niels Gustäbel, Richard Townsend.N)�TarFile�TarInfo�
is_tarfile�TarError����sustar  sustar00�d��0�1�2�3�4�5�6�7�L�K�S�x�g�X��)�path�linkpath�size�mtime�uid�gid�uname�gname)rrr#r$)�atime�ctimer r!r"ri�i�i`�@i ii����@� ���)�ntZcezutf-8cCs(|�||�}|d|�|t|�tS)z8Convert a string to a null-terminated bytes object.
    N)�encode�len�NUL)�s�length�encoding�errors�r8��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/_backport/tarfile.py�stn�sr:cCs*|�d�}|dkr|d|�}|�||�S)z8Convert a null-terminated bytes object to a string.
    r���N)�find�decode)r4r6r7�pr8r8r9�nts�s
r?cCs�|dtd�krHztt|dd�p"dd�}Wq~tyDtd��Yq~0n6d}tt|�d�D] }|dK}|t||d�7}q\|S)	z/Convert a number field to a python number.
    rr*�ascii�strict�0r.zinvalid headerr)�chr�intr?�
ValueError�InvalidHeaderError�ranger2�ord)r4�n�ir8r8r9�nti�srKcCs�d|krd|dkr<nnd|d|f�d�t}n||tksT|d|dkr\td��|dkr|t�dt�d	|��d}t�}t|d�D]}|�	d|d
@�|dL}q�|�	dd�|S)z/Convert a python number to a number field.
    rr.rz%0*or@r)zoverflow in number field�L�l�r*)
r1r3�
GNU_FORMATrE�struct�unpack�pack�	bytearrayrG�insert)rI�digits�formatr4rJr8r8r9�itn�s	 
rWc	Cshdtt�d|dd��t�d|dd���}dtt�d|dd��t�d	|dd���}||fS)
a�Calculate the checksum for a member's header by summing up all
       characters except for the chksum field which is treated as if
       it was filled with spaces. According to the GNU tar sources,
       some tars (Sun and NeXT) calculate chksum with signed char,
       which will be different if there are chars in the buffer with
       the high bit set. So we calculate two checksums, unsigned and
       signed.
    r)Z148BN�Z356B�r	Z148bZ356b)�sumrPrQ)�buf�unsigned_chksum�
signed_chksumr8r8r9�calc_chksums�s	00r^cCs�|dkrdS|dur4|�d�}|s$q0|�|�qdSd}t||�\}}t|�D],}|�|�}t|�|krptd��|�|�qN|dkr�|�|�}t|�|kr�td��|�|�dS)zjCopy length bytes from fileobj src to fileobj dst.
       If length is None, copy the entire content.
    rNr'zend of file reached)�read�write�divmodrGr2�IOError)�src�dstr5r[�BUFSIZE�blocks�	remainder�br8r8r9�copyfileobjs*



rirM�-rh�d�cr>�r�wr4�S�x�t�TcCsJg}tD]6}|D]"\}}||@|kr|�|�qq|�d�qd�|�S)zcConvert a file's mode to a string of the form
       -rwxrwxrwx.
       Used by TarFile.list()
    rj�)�filemode_table�append�join)�mode�perm�table�bit�charr8r8r9�filemode8s
r|c@seZdZdZdS)rzBase exception.N��__name__�
__module__�__qualname__�__doc__r8r8r8r9rGsrc@seZdZdZdS)�ExtractErrorz%General exception for extract errors.Nr}r8r8r8r9r�Jsr�c@seZdZdZdS)�	ReadErrorz&Exception for unreadable tar archives.Nr}r8r8r8r9r�Msr�c@seZdZdZdS)�CompressionErrorz.Exception for unavailable compression methods.Nr}r8r8r8r9r�Psr�c@seZdZdZdS)�StreamErrorz=Exception for unsupported operations on stream-like TarFiles.Nr}r8r8r8r9r�Ssr�c@seZdZdZdS)�HeaderErrorz!Base exception for header errors.Nr}r8r8r8r9r�Vsr�c@seZdZdZdS)�EmptyHeaderErrorzException for empty headers.Nr}r8r8r8r9r�Ysr�c@seZdZdZdS)�TruncatedHeaderErrorz Exception for truncated headers.Nr}r8r8r8r9r�\sr�c@seZdZdZdS)�EOFHeaderErrorz"Exception for end of file headers.Nr}r8r8r8r9r�_sr�c@seZdZdZdS)rFzException for invalid headers.Nr}r8r8r8r9rFbsrFc@seZdZdZdS)�SubsequentHeaderErrorz3Exception for missing and invalid extended headers.Nr}r8r8r8r9r�esr�c@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)�
_LowLevelFilez�Low-level file object. Supports reading and writing.
       It is used instead of a regular file object for streaming
       access.
    cCsFtjtjtjBtjBd�|}ttd�r2|tjO}t�||d�|_dS)N)rmrn�O_BINARYi�)	�os�O_RDONLY�O_WRONLY�O_CREAT�O_TRUNC�hasattrr��open�fd)�self�namerwr8r8r9�__init__rs��

z_LowLevelFile.__init__cCst�|j�dS�N)r��closer��r�r8r8r9r�{sz_LowLevelFile.closecCst�|j|�Sr�)r�r_r��r�rr8r8r9r_~sz_LowLevelFile.readcCst�|j|�dSr�)r�r`r��r�r4r8r8r9r`�sz_LowLevelFile.writeN)r~rr�r�r�r�r_r`r8r8r8r9r�ls
	r�c@steZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zddd�Zddd�Z
dd�Zdd�ZdS)�_Streama�Class that serves as an adapter between TarFile and
       a stream-like object.  The stream-like object only
       needs to have a read() or write() method and is accessed
       blockwise.  Use of gzip or bzip2 compression is possible.
       A stream-like object could be for example: sys.stdin,
       sys.stdout, a socket, a tape device etc.

       _Stream is intended to be used only internally.
    cCsNd|_|durt||�}d|_|dkr6t|�}|��}|p<d|_||_||_||_||_d|_	d|_
d|_z�|dkr�zddl}Wnt
y�td	��Yn0||_|�d�|_|d
kr�|��n|��|dk�r zddl}Wnt
y�td��Yn0|d
k�rd|_|��|_n
|��|_Wn&|j�s<|j��d|_�Yn0dS)
z$Construct a _Stream object.
        TNF�*rs�r�gzzzlib module is not availablerm�bz2�bz2 module is not available)�_extfileobjr��_StreamProxy�getcomptyper�rw�comptype�fileobj�bufsizer[�pos�closed�zlib�ImportErrorr��crc32�crc�
_init_read_gz�_init_write_gzr��dbuf�BZ2Decompressor�cmp�
BZ2Compressorr�)r�r�rwr�r�r�r�r�r8r8r9r��sP





z_Stream.__init__cCst|d�r|js|��dS)Nr�)r�r�r�r�r8r8r9�__del__�sz_Stream.__del__cCs�|j�d|jj|jj|jjd�|_t�dtt	�	���}|�
d|d�|j�d�rf|jdd�|_|�
|j�
d	d
�t�dS)z6Initialize for writing with gzip compression.
        �	r�<Ls�s�z.gzN���z
iso-8859-1�replace)r��compressobj�DEFLATED�	MAX_WBITS�
DEF_MEM_LEVELr�rPrRrD�time�_Stream__writer��endswithr1r3)r��	timestampr8r8r9r��s�z_Stream._init_write_gzcCsR|jdkr|j�||j�|_|jt|�7_|jdkrD|j�|�}|�|�dS)z&Write string s to the stream.
        r��tarN)	r�r�r�r�r�r2r��compressr�r�r8r8r9r`�s

z
_Stream.writecCsN|j|7_t|j�|jkrJ|j�|jd|j��|j|jd�|_qdS)z]Write string s to the stream if a whole new block
           is ready to be written.
        N)r[r2r�r�r`r�r8r8r9�__write�sz_Stream.__writecCs�|jr
dS|jdkr2|jdkr2|j|j��7_|jdkr�|jr�|j�|j�d|_|jdkr�|j�t�	d|j
d@��|j�t�	d|jd@��|js�|j�
�d|_dS)	z[Close the _Stream object. No operation should be
           done on it afterwards.
        Nrnr�r�r�r�l��T)r�rwr�r[r��flushr�r`rPrRr�r�r�r�r�r8r8r9r��s

z
_Stream.closecCs�|j�|jj�|_d|_|�d�dkr0td��|�d�dkrFtd��t|�d��}|�d�|d	@r�t|�d��d
t|�d��}|�	|�|d@r�|�d�}|r�|t
kr�q�q�|d@r�|�d�}|r�|t
kr�q�q�|d@r�|�d�d
S)z:Initialize for reading a gzip compressed fileobj.
        r�rs��not a gzip filer�zunsupported compression method�r/r)r.r-N)r��
decompressobjr�r�r��
_Stream__readr�r�rHr_r3)r��flag�xlenr4r8r8r9r�s*
 


z_Stream._init_read_gzcCs|jS)z3Return the stream's file pointer position.
        �r�r�r8r8r9�tell#sz_Stream.tellrcCsX||jdkrJt||j|j�\}}t|�D]}|�|j�q,|�|�ntd��|jS)zXSet the stream's file pointer to pos. Negative seeking
           is forbidden.
        rz seeking backwards is not allowed)r�rar�rGr_r�)r�r�rfrgrJr8r8r9�seek(sz_Stream.seekNcCsV|dur6g}|�|j�}|sq*|�|�qd�|�}n
|�|�}|jt|�7_|S)z�Return the next size number of bytes from the stream.
           If size is not defined, return all bytes of the stream
           up to EOF.
        Nrs)�_readr�rurvr�r2)r�rrqr[r8r8r9r_5s
z_Stream.readcCs�|jdkr|�|�St|j�}||kr�|�|j�}|s8q�z|j�|�}Wntybtd��Yn0|j|7_|t|�7}q|jd|�}|j|d�|_|S)z+Return size bytes from the stream.
        r�zinvalid compressed dataN)	r�r�r2r�r�r��
decompressrbr��r�rrlr[r8r8r9r�Gs 


z
_Stream._readcCsdt|j�}||krB|j�|j�}|s&qB|j|7_|t|�7}q
|jd|�}|j|d�|_|S)zsReturn size bytes from stream. If internal buffer is empty,
           read another block from the stream.
        N)r2r[r�r_r�r�r8r8r9�__read\s
z_Stream.__read)r)N)r~rr�r�r�r�r�r`r�r�r�r�r�r_r�r�r8r8r8r9r��s
4
	

r�c@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)r�zsSmall proxy class that enables transparent compression
       detection for the Stream interface (mode 'r|*').
    cCs||_|j�t�|_dSr�)r�r_�	BLOCKSIZEr[)r�r�r8r8r9r�qsz_StreamProxy.__init__cCs|jj|_|jSr�)r�r_r[r�r8r8r9r_us
z_StreamProxy.readcCs$|j�d�rdS|j�d�r dSdS)Ns�r�sBZh91r�r�)r[�
startswithr�r8r8r9r�ys
z_StreamProxy.getcomptypecCs|j��dSr�)r�r�r�r8r8r9r��sz_StreamProxy.closeN)r~rr�r�r�r_r�r�r8r8r8r9r�ls
r�c@sLeZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�ZdS)�	_BZ2ProxyaSmall proxy class that enables external file object
       support for "r:bz2" and "w:bz2" modes. This is actually
       a workaround for a limitation in bz2 module's BZ2File
       class which (unlike gzip.GzipFile) has no support for
       a file object argument.
    r'cCs(||_||_t|jdd�|_|��dS)Nr�)r�rw�getattrr��init)r�r�rwr8r8r9r��sz_BZ2Proxy.__init__cCsDddl}d|_|jdkr6|��|_|j�d�d|_n
|��|_dS)Nrrmr�)	r�r�rwr��bz2objr�r�r[r�)r�r�r8r8r9r��s

z_BZ2Proxy.initcCs�t|j�}||krN|j�|j�}|s&qN|j�|�}|j|7_|t|�7}q
|jd|�}|j|d�|_|jt|�7_|Sr�)r2r[r�r_�	blocksizer�r�r�)r�rrp�raw�datar[r8r8r9r_�s
z_BZ2Proxy.readcCs&||jkr|��|�||j�dSr�)r�r�r_)r�r�r8r8r9r��s
z_BZ2Proxy.seekcCs|jSr�r�r�r8r8r9r��sz_BZ2Proxy.tellcCs.|jt|�7_|j�|�}|j�|�dSr�)r�r2r�r�r�r`)r�r�r�r8r8r9r`�sz_BZ2Proxy.writecCs$|jdkr |j��}|j�|�dS)Nrn)rwr�r�r�r`)r�r�r8r8r9r��s

z_BZ2Proxy.closeN)r~rr�r�r�r�r�r_r�r�r`r�r8r8r8r9r��s
r�c@s<eZdZdZd
dd�Zdd�Zdd�Zd	d
�Zddd�ZdS)�_FileInFilezA thin wrapper around an existing file object that
       provides a part of its data as an individual file
       object.
    NcCs�||_||_||_d|_|dur*d|fg}d|_g|_d}|j}|D]L\}}||krh|j�d||df�|j�d||||f�||7}||}qD||jkr�|j�d||jdf�dS)NrFT)r��offsetr�position�	map_index�mapru)r�r�r�r�	blockinfo�lastpos�realposr8r8r9r��s$


z_FileInFile.__init__cCst|jd�sdS|j��S)N�seekableT)r�r�r�r�r8r8r9r��sz_FileInFile.seekablecCs|jS)�*Return the current file position.
        �r�r�r8r8r9r��sz_FileInFile.tellcCs
||_dS)�(Seek to a position in the file.
        Nr�)r�r�r8r8r9r��sz_FileInFile.seekcCs�|dur|j|j}nt||j|j�}d}|dkr�|j|j\}}}}||jkr^|krfq�nq�q4|jd7_|jt|j�kr4d|_q4t|||j�}|r�|j�||j|�||j�|�7}n|t	|7}||8}|j|7_q,|S)z!Read data from the file.
        Nr�rr)
rr��minr�r�r2r�r�r_r3)r�rr[r��start�stopr�r5r8r8r9r_�s&z_FileInFile.read)N)N)	r~rr�r�r�r�r�r�r_r8r8r8r9r��s
r�c@szeZdZdZdZdd�Zdd�Zdd�Zd	d
�Zddd
�Z	e	Z
ddd�Zdd�Zdd�Z
ejfdd�Zdd�Zdd�ZdS)�ExFileObjectzaFile-like object for reading an archive member.
       Is returned by TarFile.extractfile().
    r(cCsDt|j|j|j|j�|_|j|_d|_d|_|j|_d|_d|_	dS)NrmFrr�)
r�r��offset_datar�sparser�rwr�r��buffer)r��tarfile�tarinfor8r8r9r�s�zExFileObject.__init__cCsdS�NTr8r�r8r8r9�readable!szExFileObject.readablecCsdS)NFr8r�r8r8r9�writable$szExFileObject.writablecCs
|j��Sr�)r�r�r�r8r8r9r�'szExFileObject.seekableNcCs�|jrtd��d}|jrL|dur.|j}d|_n|jd|�}|j|d�|_|durd||j��7}n||j�|t|��7}|jt|�7_|S)z~Read at most size bytes from the file. If size is not
           present or None, read all data until EOF is reached.
        �I/O operation on closed filer�N)r�rEr�r�r_r2r�)r�rr[r8r8r9r_*szExFileObject.readr;cCs�|jrtd��|j�d�d}|dkrt|j�|j�}|j|7_|rNd|vr&|j�d�d}|dkrtt|j�}qtq&|dkr�t||�}|jd|�}|j|d�|_|j	t|�7_	|S)z�Read one entire line from the file. If size is present
           and non-negative, return a string with at most that
           size, which may be an incomplete line.
        r��
rrr;N)
r�rEr�r<r�r_r�r2r�r�)r�rr�r[r8r8r9�readlineEs"

zExFileObject.readlinecCs"g}|��}|sq|�|�q|S)z0Return a list with all remaining lines.
        )r�ru)r��result�liner8r8r9�	readlinesbs
zExFileObject.readlinescCs|jrtd��|jS)r�r�)r�rEr�r�r8r8r9r�lszExFileObject.tellcCs�|jrtd��|tjkr.tt|d�|j�|_nj|tjkrj|dkrTt|j|d�|_q�t|j||j�|_n.|tj	kr�tt|j||j�d�|_ntd��d|_
|j�|j�dS)r�r�rzInvalid argumentr�N)
r�rEr��SEEK_SETr��maxrr��SEEK_CUR�SEEK_ENDr�r�r�)r�r��whencer8r8r9r�ts


zExFileObject.seekcCs
d|_dS)zClose the file object.
        TN)r�r�r8r8r9r��szExFileObject.closeccs|��}|sq|VqdS)z/Get an iterator over the file's lines.
        N)r�)r�r�r8r8r9�__iter__�szExFileObject.__iter__)N)r;)r~rr�r�r�r�r�r�r�r_�read1r�r�r�r�r�r�r�rr8r8r8r9r�s



r�c@s�eZdZdZdZdVdd�Zdd�Zdd	�Zeee�Z	d
d�Z
dd
�Zee
e�Zdd�Z
dd�Zeedfdd�Zdd�Zdd�Zdd�Zedd��Zdd�Zedd ��Zed!d"��Zed#d$��Zed%d&��Zed'd(��Zed)d*��Zd+d,�Zd-d.�Z d/d0�Z!d1d2�Z"d3d4�Z#d5d6�Z$d7d8�Z%d9d:�Z&d;d<�Z'd=d>�Z(d?d@�Z)dAdB�Z*dCdD�Z+dEdF�Z,dGdH�Z-dIdJ�Z.dKdL�Z/dMdN�Z0dOdP�Z1dQdR�Z2dSdT�Z3dUS)WraInformational class which holds the details about an
       archive member given by a tar header block.
       TarInfo objects are returned by TarFile.getmember(),
       TarFile.getmembers() and TarFile.gettarinfo() and are
       usually created internally.
    )r�rwr!r"rr �chksum�type�linknamer#r$�devmajor�devminorr�r��pax_headersr�r��_sparse_structs�_link_targetrscCsj||_d|_d|_d|_d|_d|_d|_t|_d|_	d|_
d|_d|_d|_
d|_d|_d|_i|_dS)zXConstruct a TarInfo object. name is the optional name
           of the member.
        i�rrsN)r�rwr!r"rr r�REGTYPErrr#r$rrr�r�r�r�r�r�r8r8r9r��s"zTarInfo.__init__cCs|jSr��r�r�r8r8r9�_getpath�szTarInfo._getpathcCs
||_dSr�rrr8r8r9�_setpath�szTarInfo._setpathcCs|jSr��rr�r8r8r9�_getlinkpath�szTarInfo._getlinkpathcCs
||_dSr�r)r�rr8r8r9�_setlinkpath�szTarInfo._setlinkpathcCsd|jj|jt|�fS)Nz<%s %r at %#x>)�	__class__r~r��idr�r8r8r9�__repr__�szTarInfo.__repr__cCsl|j|jd@|j|j|j|j|j|j|j|j	|j
|j|jd�
}|dt
krh|d�d�sh|dd7<|S)z9Return the TarInfo's attributes as a dictionary.
        �)
r�rwr!r"rr rrrr#r$rrrr��/)r�rwr!r"rr rrrr#r$rr�DIRTYPEr�)r��infor8r8r9�get_info�s"�zTarInfo.get_info�surrogateescapecCsT|��}|tkr|�|||�S|tkr4|�|||�S|tkrH|�||�Std��dS)z<Return a tar header as a string of 512 byte blocks.
        zinvalid formatN)r�USTAR_FORMAT�create_ustar_headerrO�create_gnu_header�
PAX_FORMAT�create_pax_headerrE)r�rVr6r7rr8r8r9�tobuf�sz
TarInfo.tobufcCsZt|d<t|d�tkr td��t|d�tkrJ|�|d�\|d<|d<|�|t||�S)z3Return the object as a ustar header block.
        �magicrzlinkname is too longr��prefix)�POSIX_MAGICr2�LENGTH_LINKrE�LENGTH_NAME�_posix_split_name�_create_headerr)r�rr6r7r8r8r9r�szTarInfo.create_ustar_headercCspt|d<d}t|d�tkr4||�|dt||�7}t|d�tkr\||�|dt||�7}||�|t||�S)z:Return the object as a GNU header block sequence.
        r!r�rr�)	�	GNU_MAGICr2r$�_create_gnu_long_header�GNUTYPE_LONGLINKr%�GNUTYPE_LONGNAMEr'rO)r�rr6r7r[r8r8r9rszTarInfo.create_gnu_headerc
	Cs(t|d<|j��}ddtfddtfddfD]h\}}}||vr>q*z||�dd	�Wn"tyt||||<Yq*Yn0t||�|kr*||||<q*d
D]`\}}||vr�d||<q�||}d|kr�d|d
kr�nn
t|t	�r�t
|�||<d||<q�|�r|�|t|�}	nd}	|	|�
|tdd�S)z�Return the object as a ustar header block. If it cannot be
           represented this way, prepend a pax extended header sequence
           with supplement information.
        r!r�rrr)r#r#r,)r$r$r,r@rA))r!r.)r"r.)r�)r r,rr.rr�r�)r#r�copyr%r$r1�UnicodeEncodeErrorr2�
isinstance�float�str�_create_pax_generic_header�XHDTYPEr'r)
r�rr6rr��hnamer5rU�valr[r8r8r9rs4
�
*
zTarInfo.create_pax_headercCs|�|td�S)zAReturn the object as a pax global header block sequence.
        �utf8)r2�XGLTYPE)�clsrr8r8r9�create_pax_global_headerDsz TarInfo.create_pax_global_headercCsj|dtd�}|r.|ddkr.|dd�}q|t|�d�}|dd�}|rZt|�tkrbtd��||fS)zUSplit a name longer than 100 chars into a prefix
           and a name part.
        Nrr;rzname is too long)�
LENGTH_PREFIXr2r%rE)r�r�r"r8r8r9r&JszTarInfo._posix_split_namecCsVt|�dd�d||�t|�dd�d@d|�t|�dd�d|�t|�d	d�d|�t|�d
d�d|�t|�dd�d|�d
|�dt�t|�dd�d||�|�dt�t|�dd�d||�t|�dd�d||�t|�dd�d|�t|�dd�d|�t|�dd�d||�g}t�dtd�|��}t	|td��d}|dd�d|�
d�|dd�}|S)z�Return a header block. info is a dictionary with file
           information, format must be one of the *_FORMAT constants.
        r�rsrrwrrr.r!r"rr,r s        rrr!r#r,r$rrr"rz%dsr�Ni����z%06or@i����)r:�getrWr
r#rPrRr�rvr^r1)rrVr6r7�partsr[rr8r8r9r'Ys(

�&zTarInfo._create_headercCs.tt|�t�\}}|dkr*|t|t7}|S)zdReturn the string payload filled with zero bytes
           up to the next 512 byte border.
        r)rar2r�r3)�payloadrfrgr8r8r9�_create_payloaduszTarInfo._create_payloadcCsR|�||�t}i}d|d<||d<t|�|d<t|d<|�|t||�|�|�S)zTReturn a GNUTYPE_LONGNAME or GNUTYPE_LONGLINK sequence
           for name.
        z
././@LongLinkr�rrr!)r1r3r2r(r'rr>)r8r�rr6r7rr8r8r9r)s�zTarInfo._create_gnu_long_headerc	Cs0d}|��D]6\}}z|�dd�Wqty@d}YqDYq0qd}|rT|d7}|��D]�\}}|�d�}|r�|�|d�}n
|�d�}t|�t|�d}d	}	}
|tt|
��}	|	|
kr�q�|	}
q�|tt|
�d
�d|d|d
7}q\i}d|d<||d<t|�|d<t|d<|�|td
d�|�	|�S)z�Return a POSIX.1-2008 extended or global header sequence
           that contains a list of keyword, value pairs. The values
           must be strings.
        Fr6rATr�s21 hdrcharset=BINARY
rrrr@� �=r�z././@PaxHeaderr�rrr!r�)
�itemsr1r.r2r1�bytesr#r'rr>)r8rrr6�binary�keyword�value�recordsrMrIr>rr8r8r9r2�s<

(�z"TarInfo._create_pax_generic_headerc	Cstt|�dkrtd��t|�tkr(td��|�t�tkr>td��t|dd��}|t|�vrbt	d��|�}t
|dd�||�|_t|dd	��|_t|d	d
��|_
t|d
d��|_t|dd��|_t|dd��|_||_|dd
�|_t
|d
d�||�|_t
|dd�||�|_t
|dd�||�|_t|dd��|_t|dd��|_t
|dd�||�}|jtk�r�|j�d��r�t|_|jtk�r6d}g}td�D]j}	z0t|||d��}
t||d|d��}Wnt�y�Y�qYn0|�|
|f�|d7}�q�t|d�}t|dd��}
|||
f|_ |�!��rN|j�"d�|_|�rp|jt#v�rp|d|j|_|S)zAConstruct a TarInfo object from a 512 byte bytes object.
        rzempty headerztruncated headerzend of file headerrXrYzbad checksumr�l�t�|��ii	i)iIiQiYi�ri�r/r,�i�i�i�)$r2r�r�r��countr3r�rKr^rFr?r�rwr!r"rr rrrr#r$rr�AREGTYPEr�r�GNUTYPE_SPARSErGrEru�boolr�isdir�rstrip�	GNU_TYPES)r8r[r6r7r�objr"r��structsrJr��numbytes�
isextended�origsizer8r8r9�frombuf�sZ
zTarInfo.frombufcCs8|j�t�}|�||j|j�}|j��t|_|�|�S)zOReturn the next TarInfo object from TarFile object
           tarfile.
        )	r�r_r�rYr6r7r�r��_proc_member)r8r�r[rTr8r8r9�fromtarfileszTarInfo.fromtarfilecCsT|jttfvr|�|�S|jtkr,|�|�S|jtttfvrF|�	|�S|�
|�SdS)zYChoose the right processing method depending on
           the type and call it.
        N)rr+r*�
_proc_gnulongrO�_proc_sparser3r7�SOLARIS_XHDTYPE�	_proc_pax�
_proc_builtin�r�r�r8r8r9rZs



zTarInfo._proc_membercCsR|j��|_|j}|��s$|jtvr4||�|j�7}||_|�	|j
|j|j�|S)zfProcess a builtin type or an unknown type which
           will be treated as a regular file.
        )
r�r�r��isregr�SUPPORTED_TYPES�_blockrr��_apply_pax_inforr6r7)r�r�r�r8r8r9r`$szTarInfo._proc_builtincCs�|j�|�|j��}z|�|�}Wnty<td��Yn0|j|_|jt	krdt
||j|j�|_
n|jtkr�t
||j|j�|_|S)zSProcess the blocks that hold a GNU longname
           or longlink member.
        � missing or bad subsequent header)r�r_rdrr[r�r�r�rr+r?r6r7r�r*r)r�r�r[�nextr8r8r9r\5s

zTarInfo._proc_gnulongc
	Cs�|j\}}}|`|r�|j�t�}d}td�D]l}z0t|||d��}t||d|d��}	WntyxYq�Yn0|r�|	r�|�||	f�|d7}q,t|d�}q||_	|j�
�|_|j|�|j
�|_||_
|S)z8Process a GNU sparse header plus extra headers.
        r�r,rLi�)rr�r_r�rGrKrErurPr�r�r�rdrr�)
r�r�rUrWrXr[r�rJr�rVr8r8r9r]Ks(
zTarInfo._proc_sparsecCs|j�|�|j��}|jtkr&|j}n
|j��}t�	d|�}|durX|�
d��d�|d<|�d�}|dkrr|j
}nd}t�d�}d}|�||�}|s��q&|��\}	}
t|	�}	||�d	�d|�d�|	d�}|�|
dd|j�}
|
tv�r|�|||j
|j�}n|�|dd|j�}|||
<||	7}q�z|�|�}Wnt�yPtd
��Yn0d|v�rj|�||�nHd|v�r�|�|||�n.|�d
�dk�r�|�d�dk�r�|�|||�|jttfv�r|�||j
|j�|j |_ d|v�r|j!}
|�"��s|jt#v�r|
|�|j�7}
|
|_ |S)zVProcess an extended or global header as described in
           POSIX.1-2008.
        s\d+ hdrcharset=([^\n]+)\nNrr6�
hdrcharset�BINARYs(\d+) ([^=]+)=rrrf�GNU.sparse.map�GNU.sparse.sizezGNU.sparse.major�1zGNU.sparse.minorrBr)$r�r_rdrrr7rr-�re�search�groupr=r;r6�compile�match�groupsrD�endr��_decode_pax_fieldr7�PAX_NAME_FIELDSr[r�r��_proc_gnusparse_01�_proc_gnusparse_00�_proc_gnusparse_10r3r^rer�r�rbrc)r�r�r[rrrrir6�regexr�r5rDrErgr�r8r8r9r_gsd



$	
�
�
�


 
zTarInfo._proc_paxcCshg}t�d|�D]}|�t|�d���qg}t�d|�D]}|�t|�d���q:tt||��|_dS)z?Process a GNU tar extended sparse header, version 0.0.
        s\d+ GNU.sparse.offset=(\d+)\nrs\d+ GNU.sparse.numbytes=(\d+)\nN)rn�finditerrurDrp�list�zipr�)r�rgrr[�offsetsrrrVr8r8r9rx�szTarInfo._proc_gnusparse_00cCs@dd�|d�d�D�}tt|ddd�|ddd���|_dS)z?Process a GNU tar extended sparse header, version 0.1.
        cSsg|]}t|��qSr8)rD)�.0rpr8r8r9�
<listcomp>�r�z.TarInfo._proc_gnusparse_01.<locals>.<listcomp>rk�,Nrr)�splitr|r}r�)r�rgrr�r8r8r9rw�szTarInfo._proc_gnusparse_01cCs�d}g}|j�t�}|�dd�\}}t|�}t|�|dkrtd|vrT||j�t�7}|�dd�\}}|�t|��q,|j��|_t	t
|ddd�|ddd���|_dS)z?Process a GNU tar extended sparse header, version 1.0.
        Nr�rr)r�r_r�r�rDr2rur�r�r|r}r�)r�rgrr��fieldsr�r[�numberr8r8r9ry�szTarInfo._proc_gnusparse_10c	Cs�|��D]�\}}|dkr&t|d|�q|dkr@t|dt|��q|dkrZt|dt|��q|tvr|tvr�zt||�}Wnty�d}Yn0|dkr�|�d�}t|||�q|��|_dS)	zoReplace fields with supplemental information from a previous
           pax extended or global header.
        zGNU.sparse.namerrlrzGNU.sparse.realsizerrN)	rA�setattrrD�
PAX_FIELDS�PAX_NUMBER_FIELDSrErRr-r)r�rr6r7rDrEr8r8r9re�s"

zTarInfo._apply_pax_infocCs2z|�|d�WSty,|�||�YS0dS)z1Decode a single field from a pax record.
        rAN)r=�UnicodeDecodeError)r�rEr6�fallback_encoding�fallback_errorsr8r8r9ruszTarInfo._decode_pax_fieldcCs"t|t�\}}|r|d7}|tS)z_Round up a byte count by BLOCKSIZE and return it,
           e.g. _block(834) => 1024.
        r)rar�)r�rMrfrgr8r8r9rd
szTarInfo._blockcCs
|jtvSr�)r�
REGULAR_TYPESr�r8r8r9rbsz
TarInfo.isregcCs|��Sr�)rbr�r8r8r9�isfileszTarInfo.isfilecCs
|jtkSr�)rrr�r8r8r9rQsz
TarInfo.isdircCs
|jtkSr�)r�SYMTYPEr�r8r8r9�issymsz
TarInfo.issymcCs
|jtkSr�)r�LNKTYPEr�r8r8r9�islnksz
TarInfo.islnkcCs
|jtkSr�)r�CHRTYPEr�r8r8r9�ischr sz
TarInfo.ischrcCs
|jtkSr�)r�BLKTYPEr�r8r8r9�isblk"sz
TarInfo.isblkcCs
|jtkSr�)r�FIFOTYPEr�r8r8r9�isfifo$szTarInfo.isfifocCs
|jduSr�)r�r�r8r8r9�issparse&szTarInfo.issparsecCs|jtttfvSr�)rr�r�r�r�r8r8r9�isdev(sz
TarInfo.isdevN)rs)4r~rr�r��	__slots__r�r
r�propertyrrrrrr�DEFAULT_FORMAT�ENCODINGr rrr�classmethodr9r&�staticmethodr'r>r)r2rYr[rZr`r\r]r_rxrwryrerurdrbr�rQr�r�r�r�r�r�r�r8r8r8r9r�sf



1


	

2
>

f	rc@s�eZdZdZdZdZdZdZeZ	e
ZdZe
ZeZdVdd	�Zedddefd
d��ZedWdd
��ZedXdd��ZedYdd��Zdddd�Zdd�Zdd�Zdd�Zdd�ZdZdd �Zd[d"d#�Zd\d$d%�Zd]d&d'�Z d^d)d*�Z!d_d,d-�Z"d.d/�Z#d`d0d1�Z$d2d3�Z%d4d5�Z&d6d7�Z'd8d9�Z(d:d;�Z)d<d=�Z*d>d?�Z+d@dA�Z,dBdC�Z-dDdE�Z.dadFdG�Z/dHdI�Z0dbdJdK�Z1dLdM�Z2dNdO�Z3dPdQ�Z4dRdS�Z5dTdU�Z6dS)crz=The TarFile Class provides an interface to tar archives.
    rFrNrmrc

Cs�t|�dks|dvrtd��||_dddd�||_|sn|jdkrZtj�|�sZd	|_d|_t||j�}d
|_n0|dur�t	|d�r�|j
}t	|d
�r�|j|_d|_|r�tj�|�nd|_
||_|dur�||_
|dur�||_|dur�||_|dur�||_|du�r||_|	|_|
du�r&|j
tk�r&|
|_ni|_|du�r<||_|du�rL||_d
|_g|_d
|_|j��|_i|_z�|jdk�r�d|_|��|_|jdk�r"|j�|j�z|j� |�}|j�!|�WnXt"�y�|j�|j�Y�q"Yn2t#�y}
zt$t%|
���WYd}
~
n
d}
~
00�q�|jdv�rld|_|j�rl|j�&|j�'��}|j�(|�|jt|�7_Wn&|j�s�|j�)�d|_�Yn0dS)a�Open an (uncompressed) tar archive `name'. `mode' is either 'r' to
           read from an existing archive, 'a' to append data to an existing
           file or 'w' to create a new file overwriting an existing one. `mode'
           defaults to 'r'.
           If `fileobj' is given, it is used for reading or writing data. If it
           can be determined, `mode' is overridden by `fileobj's mode.
           `fileobj' is not closed, when TarFile is closed.
        rr��mode must be 'r', 'a' or 'w'�rbzr+b�wb)rm�arnr�rnFNr�rwTrm�aw)*r2rErw�_moder�r�exists�	bltn_openr�r�r��abspathr�rVr��dereference�ignore_zerosr6r7rr�debug�
errorlevelr��members�_loadedr�r��inodes�firstmemberrgr�r[rur�r�r�r1r9r-r`r�)r�r�rwr�rVr�r�r�r6r7rr�r��er[r8r8r9r�Fs�





&
zTarFile.__init__c
Ks�|s|std��|dvr�|jD]�}t||j|�}|durB|��}z||d|fi|��WSttfy�}	z*|dur�|�|�WYd}	~	qWYd}	~	qd}	~	00qtd���nd|v�r|�dd�\}
}|
p�d}
|p�d}||jvr�t||j|�}ntd	|��|||
|fi|��Sd
|v�r�|�d
d�\}
}|
�p:d}
|�pDd}|
dv�rXtd��t||
|||�}z|||
|fi|��}Wn|�	��Yn0d
|_
|S|dv�r�|j|||fi|��Std��dS)a|Open a tar archive for reading, writing or appending. Return
           an appropriate TarFile class.

           mode:
           'r' or 'r:*' open for reading with transparent compression
           'r:'         open for reading exclusively uncompressed
           'r:gz'       open for reading with gzip compression
           'r:bz2'      open for reading with bzip2 compression
           'a' or 'a:'  open for appending, creating the file if necessary
           'w' or 'w:'  open for writing without compression
           'w:gz'       open for writing with gzip compression
           'w:bz2'      open for writing with bzip2 compression

           'r|*'        open a stream of tar blocks with transparent compression
           'r|'         open an uncompressed stream of tar blocks for reading
           'r|gz'       open a gzip compressed stream of tar blocks
           'r|bz2'      open a bzip2 compressed stream of tar blocks
           'w|'         open an uncompressed stream for writing
           'w|gz'       open a gzip compressed stream for writing
           'w|bz2'      open a bzip2 compressed stream for writing
        znothing to open)rmzr:*Nrmz%file could not be opened successfully�:rr�zunknown compression type %r�|�rw�mode must be 'r' or 'w'Fr�zundiscernible mode)rE�	OPEN_METHr�r�r�r�r�r�r�r�r��taropen)
r8r�rwr�r��kwargsr��func�	saved_posr�r|�streamrqr8r8r9r��sN

$






zTarFile.opencKs0t|�dks|dvrtd��||||fi|��S)zCOpen uncompressed tar archive name for reading or writing.
        rr�r�)r2rE)r8r�rwr�r�r8r8r9r��szTarFile.taropenr�c		Ks�t|�dks|dvrtd��zddl}|jWnttfyLtd��Yn0|du}z.|�||d||�}|j|||fi|��}WnXty�|s�|dur�|�	�|dur��t
d��Yn"|s�|dur�|�	��Yn0||_|S)	zkOpen gzip compressed tar archive name for reading or writing.
           Appending is not allowed.
        rr�r�rNzgzip module is not availablerhr�)r2rE�gzip�GzipFiler��AttributeErrorr�r�rbr�r�r�)	r8r�rwr��
compresslevelr�r�Z
extfileobjrqr8r8r9�gzopens.
zTarFile.gzopenc	Ks�t|�dks|dvrtd��zddl}WntyBtd��Yn0|durXt||�}n|j|||d�}z|j|||fi|��}Wn&tt	fy�|�
�td��Yn0d	|_|S)
zlOpen bzip2 compressed tar archive name for reading or writing.
           Appending is not allowed.
        rr�zmode must be 'r' or 'w'.rNr�)r�znot a bzip2 fileF)
r2rEr�r�r�r��BZ2Filer�rb�EOFErrorr�r�r�)r8r�rwr�r�r�r�rqr8r8r9�bz2open$s zTarFile.bz2openr�r�r�)r�r�r�cCs�|jr
dS|jdvrf|j�ttd�|jtd7_t|jt�\}}|dkrf|j�tt|�|j	sv|j�
�d|_dS)zlClose the TarFile. In write-mode, two finishing zero blocks are
           appended to the archive.
        Nr�rrT)r�rwr�r`r3r�r�ra�
RECORDSIZEr�r�)r�rfrgr8r8r9r�Hs

z
TarFile.closecCs"|�|�}|durtd|��|S)aReturn a TarInfo object for member `name'. If `name' can not be
           found in the archive, KeyError is raised. If a member occurs more
           than once in the archive, its last occurrence is assumed to be the
           most up-to-date version.
        Nzfilename %r not found)�
_getmember�KeyError)r�r�r�r8r8r9�	getmember\s
zTarFile.getmembercCs|��|js|��|jS)z�Return the members of the archive as a list of TarInfo objects. The
           list has the same order as the members in the archive.
        )�_checkr��_loadr�r�r8r8r9�
getmembersgszTarFile.getmemberscCsdd�|��D�S)z�Return the members of the archive as a list of their names. It has
           the same order as the list returned by getmembers().
        cSsg|]
}|j�qSr8r)rr�r8r8r9r�ur�z$TarFile.getnames.<locals>.<listcomp>)r�r�r8r8r9�getnamesqszTarFile.getnamescCsh|�d�|dur|j}|dur$|}tj�|�\}}|�tjd�}|�d�}|��}||_	|dur�t
td�r~|js~t�|�}q�t�
|�}nt�|���}d}|j}t
�|��r|j|jf}	|js�|jdkr�|	|jvr�||j|	kr�t}
|j|	}nt}
|	d�rt||j|	<nht
�|��rt}
nVt
�|��r0t}
nDt
�|��rLt}
t�|�}n(t
�|��r^t }
nt
�!|��rpt"}
ndS||_||_#|j$|_%|j&|_'|
tk�r�|j(|_)nd|_)|j*|_+|
|_,||_-t.�r�zt.�/|j%�d|_0Wnt1�y�Yn0t2�r"zt2�3|j'�d|_4Wnt1�y Yn0|
t t"fv�rdt
td��rdt
td	��rdt�5|j6�|_7t�8|j6�|_9|S)
aOCreate a TarInfo object for either the file `name' or the file
           object `fileobj' (using os.fstat on its file descriptor). You can
           modify some of the TarInfo's attributes before you add it using
           addfile(). If given, `arcname' specifies an alternative name for the
           file in the archive.
        r�Nr�lstatrsrr�major�minor):r�r�r�r�
splitdriver��sep�lstripr�r�r�r�r��stat�fstat�fileno�st_mode�S_ISREG�st_ino�st_dev�st_nlinkr�r�r
�S_ISDIRr�S_ISFIFOr��S_ISLNKr��readlink�S_ISCHRr��S_ISBLKr�rw�st_uidr!�st_gidr"�st_sizer�st_mtimer rr�pwd�getpwuidr#r��grp�getgrgidr$r��st_rdevrr�r)r�r��arcnamer��drvr��statresr�stmd�inoderr8r8r9�
gettarinfows�

��


zTarFile.gettarinfoTcCs|��|D]�}|r�tt|j�dd�td|jp4|j|jp>|jfdd�|��sZ|�	�rxtdd|j
|jfdd�ntd|jdd�tdt
�|j�dd	�dd�t|j|��r�d
nddd�|r�|��r�td|jdd�|��r�td
|jdd�t�qdS)z�Print a table of contents to sys.stdout. If `verbose' is False, only
           the names of the members are printed. If it is True, an `ls -l'-like
           output is produced.
        � )rtz%s/%sz%10sz%d,%dz%10dz%d-%02d-%02d %02d:%02d:%02dNr�rrsz->zlink to)r��printr|rwr#r!r$r"r�r�rrrr��	localtimer r�rQr�rr�)r��verboser�r8r8r9r|�s8
��
����zTarFile.listc	Csn|�d�|dur|}|durPddl}|�dtd�||�rP|�dd|�dS|jdur�tj�|�|jkr�|�dd|�dS|�d|�|�	||�}|dur�|�dd	|�dS|dur�||�}|dur�|�dd|�dS|�
��r
t|d
�}|�||�|�
�n`|���r`|�|�|�rjt�|�D].}|jtj�||�tj�||�|||d��q.n
|�|�dS)a~Add the file `name' to the archive. `name' may be any type of file
           (directory, fifo, symbolic link, etc.). If given, `arcname'
           specifies an alternative name for the file in the archive.
           Directories are added recursively by default. This can be avoided by
           setting `recursive' to False. `exclude' is a function that should
           return True for each filename to be excluded. `filter' is a function
           that expects a TarInfo object argument and returns the changed
           TarInfo object, if it returns None the TarInfo object will be
           excluded from the archive.
        r�Nrzuse the filter argument insteadrztarfile: Excluded %rztarfile: Skipped %rrztarfile: Unsupported type %rr�)�filter)r��warnings�warn�DeprecationWarning�_dbgr�r�rr�r�rbr��addfiler�rQ�listdir�addrv)	r�r�r��	recursive�excluder�r�r��fr8r8r9r��sH
�




�zTarFile.addcCs�|�d�t�|�}|�|j|j|j�}|j�|�|jt	|�7_|dur�t
||j|j�t|jt
�\}}|dkr�|j�tt
|�|d7}|j|t
7_|j�|�dS)a]Add the TarInfo object `tarinfo' to the archive. If `fileobj' is
           given, tarinfo.size bytes are read from it and added to the archive.
           You can create TarInfo objects using gettarinfo().
           On Windows platforms, `fileobj' should always be opened with mode
           'rb' to avoid irritation about the file size.
        r�Nrr)r�r-r rVr6r7r�r`r�r2rirrar�r3r�ru)r�r�r�r[rfrgr8r8r9r�4s

zTarFile.addfile�.cCs�g}|dur|}|D]<}|��r:|�|�t�|�}d|_|j|||��d�q|jdd�d�|��|D]�}tj�	||j
�}z(|�||�|�||�|�
||�Wqnty�}z*|jdkrȂn|�dd|�WYd}~qnd}~00qndS)	aMExtract all members from the archive to the current working
           directory and set owner, modification time and permissions on
           directories afterwards. `path' specifies a different directory
           to extract to. `members' is optional and must be a subset of the
           list returned by getmembers().
        N���	set_attrscSs|jSr�r)r�r8r8r9�<lambda>dr�z$TarFile.extractall.<locals>.<lambda>)�keyr�tarfile: %s)rQrur-rw�extract�sort�reverser�rrvr��chown�utime�chmodr�r�r�)r�rr��directoriesr��dirpathr�r8r8r9�
extractallNs*


zTarFile.extractallrsc
Cs|�d�t|t�r |�|�}n|}|��r>tj�||j�|_	z |j
|tj�||j�|d�Wn�ty�}zP|j
dkr|�n6|jdur�|�dd|j�n|�dd|j|jf�WYd}~nLd}~0t�y}z*|j
dkr�n|�dd|�WYd}~n
d}~00dS)axExtract a member from the archive to the current working directory,
           using its full name. Its file information is extracted as accurately
           as possible. `member' may be a filename or a TarInfo object. You can
           specify a different directory using `path'. File attributes (owner,
           mtime, mode) are set unless `set_attrs' is False.
        rmr�rNrrztarfile: %s %r)r�r/r1r�r�r�rrvrr	�_extract_memberr��EnvironmentErrorr��filenamer��strerrorr�)r��memberrrr�r�r8r8r9rts(

�


,
zTarFile.extractcCs�|�d�t|t�r |�|�}n|}|��r8|�||�S|jtvrN|�||�S|��s^|�	�r�t|j
t�rttd��q�|�
|�|��SndSdS)a�Extract a member from the archive as a file object. `member' may be
           a filename or a TarInfo object. If `member' is a regular file, a
           file-like object is returned. If `member' is a link, a file-like
           object is constructed from the link's target. If `member' is none of
           the above, None is returned.
           The file-like object is read-only and provides the following
           methods: read(), readline(), readlines(), seek() and tell()
        rmz'cannot extract (sym)link as file objectN)r�r/r1r�rb�
fileobjectrrcr�r�r�r�r��extractfile�_find_link_target)r�rr�r8r8r9r�s	



zTarFile.extractfilecCsR|�d�}|�dtj�}tj�|�}|r>tj�|�s>t�|�|��sN|�	�rh|�
dd|j|jf�n|�
d|j�|�
�r�|�||�n�|��r�|�||�nx|��r�|�||�nb|��s�|��r�|�||�nD|��s�|�	�r�|�||�n&|jtv�r|�||�n|�||�|�rN|�||�|�	��sN|�||�|�||�dS)z\Extract the TarInfo object tarinfo to a physical
           file called targetpath.
        rrz%s -> %sN)rRr�r�r�r�dirnamer��makedirsr�r�r�r�rrb�makefilerQ�makedirr��makefifor�r��makedev�makelinkrrc�makeunknownrr	r)r�r��
targetpathr�	upperdirsr8r8r9r
�s4


zTarFile._extract_memberc
CsHzt�|d�Wn2tyB}z|jtjkr.�WYd}~n
d}~00dS)z,Make a directory called targetpath.
        r�N)r��mkdirr�errno�EEXIST�r�r�rr�r8r8r9r�s
zTarFile.makedircCs||j}|�|j�t|d�}|jdurN|jD]\}}|�|�t|||�q,nt|||j�|�|j�|��|��dS)z'Make a file called targetpath.
        r�N)	r�r�r�r�r�rir�truncater�)r�r�r�source�targetr�rr8r8r9r�s


zTarFile.makefilecCs"|�||�|�dd|j�dS)zYMake a file from a TarInfo object with an unknown type
           at targetpath.
        rz9tarfile: Unknown file type %r, extracted as regular file.N)rr�r�r�r�rr8r8r9r	s�zTarFile.makeunknowncCs"ttd�rt�|�ntd��dS)z'Make a fifo called targetpath.
        �mkfifozfifo not supported by systemN)r�r�r'r�r&r8r8r9r	s
zTarFile.makefifocCs^ttd�rttd�std��|j}|��r6|tjO}n
|tjO}t�||t�	|j
|j��dS)z<Make a character or block device called targetpath.
        �mknodrz'special devices not supported by systemN)r�r�r�rwr�r��S_IFBLK�S_IFCHRr(rrr)r�r�rrwr8r8r9r	s
�zTarFile.makedevcCs�zL|��rt�|j|�n0tj�|j�r8t�|j|�n|�|�	|�|�Wn<t
y�|��r~tj�tj�|j
�|j�}n|j}Yn40z|�|�	|�|�Wnty�td��Yn0dS)z�Make a (symbolic) link called targetpath. If it cannot be created
          (platform limitation), we try to make a copy of the referenced file
          instead of a link.
        z%unable to resolve link inside archiveN)r�r��symlinkrrr�r	�linkr
r�symlink_exceptionrvrr�r�r�)r�r�rrr8r8r9r'	s(���zTarFile.makelinkc
Cs�tr�ttd�r�t��dkr�zt�|j�d}WntyF|j}Yn0zt�	|j
�d}Wntyt|j}Yn0z>|��r�ttd�r�t�
|||�ntjdkr�t�|||�Wn,ty�}ztd��WYd}~n
d}~00dS)z6Set owner of targetpath according to tarinfo.
        �geteuidrr�lchownZos2emxzcould not change ownerN)r�r�r�r.r��getgrnamr$r�r"�getpwnamr#r!r�r/�sys�platformrrr�)r�r�r�g�ur�r8r8r9rD	s 
z
TarFile.chownc
CsNttd�rJzt�||j�Wn,tyH}ztd��WYd}~n
d}~00dS)zASet file permissions of targetpath according to tarinfo.
        r	zcould not change modeN)r�r�r	rwrr�r"r8r8r9r	Z	s

z
TarFile.chmodc
CsXttd�sdSzt�||j|jf�Wn,tyR}ztd��WYd}~n
d}~00dS)zBSet modification time of targetpath according to tarinfo.
        rNz"could not change modification time)r�r�rr rr�r"r8r8r9rc	s
z
TarFile.utimec
Cs�|�d�|jdur$|j}d|_|S|j�|j�d}z|j�|�}W�q�ty�}zD|jr�|�	dd|j|f�|jt
7_WYd}~q6WYd}~�q�d}~0t�y}z\|jr�|�	dd|j|f�|jt
7_WYd}~q6n|jdk�rtt
|���WYd}~n�d}~0t�y>|jdk�r:td��Ynnt�yz}z$|jdk�rftt
|���WYd}~n:d}~0t�y�}ztt
|���WYd}~n
d}~00�q�q6|du�r�|j�|�nd|_|S)z�Return the next member of the archive as a TarInfo object, when
           TarFile is opened for reading. Return None if there is no more
           available.
        �raNrz0x%X: %srz
empty fileT)r�r�r�r�r�r�r[r�r�r�r�rFr�r1r�r�r�r�rur�)r��mr�r�r8r8r9rgn	sD

"  "
zTarFile.nextcCsn|��}|dur"|d|�|��}|r2tj�|�}t|�D].}|rRtj�|j�}n|j}||kr:|Sq:dS)z}Find an archive member by name from bottom to top.
           If tarinfo is given, it is used as the starting point.
        N)r��indexr�r�normpath�reversedr�)r�r�r��	normalizer�r�member_namer8r8r9r��	szTarFile._getmembercCs|��}|durqqd|_dS)zWRead through the entire archive file and look for readable
           members.
        NT)rgr��r�r�r8r8r9r��	sz
TarFile._loadcCs:|jrtd|jj��|dur6|j|vr6td|j��dS)znCheck if TarFile is still open, and if the operation's mode
           corresponds to TarFile's mode.
        z%s is closedNzbad operation for mode %r)r�rbrr~rw)r�rwr8r8r9r��	szTarFile._checkcCsX|��r&tj�|j�d|j}d}n
|j}|}|j||dd�}|durTtd|��|S)zZFind the target member of a symlink or hardlink member in the
           archive.
        rNT)r�r;zlinkname %r not found)r�r�rrr�rr�r�)r�r�r�limitrr8r8r9r�	szTarFile._find_link_targetcCs|jrt|j�St|�SdS)z$Provide an iterator object.
        N)r��iterr��TarIterr�r8r8r9r�	s
zTarFile.__iter__cCs||jkrt|tjd�dS)z.Write debugging output to sys.stderr.
        )�fileN)r�r�r2�stderr)r��level�msgr8r8r9r��	s
zTarFile._dbgcCs|��|Sr�)r�r�r8r8r9�	__enter__�	szTarFile.__enter__cCs,|dur|��n|js"|j��d|_dSr�)r�r�r�r�)r�rrE�	tracebackr8r8r9�__exit__�	s


zTarFile.__exit__)NrmNNNNNNrNNN)rmN)rmNr�)rmNr�)NNN)T)NTNN)N)r�N)rsT)T)NF)N)7r~rr�r�r�r�r�r�r�rVr�r6r7rr�r�rr�r�r�r�r�r�r�r�r�r�r�r�r�r|r�r�rrrr
rrrrrrrr	rrgr�r�r�rrr�rErGr8r8r8r9r,sp�
kK�	

b

>

&
#&
0	1


	rc@s,eZdZdZdd�Zdd�Zdd�ZeZdS)	r@zMIterator Class.

       for tarinfo in TarFile(...):
           suite...
    cCs||_d|_dS)z$Construct a TarIter object.
        rN)r�r8rar8r8r9r�
szTarIter.__init__cCs|S)z Return iterator object.
        r8r�r8r8r9r
szTarIter.__iter__cCs`|jjs$|j��}|sNd|j_t�n*z|jj|j}WntyLt�Yn0|jd7_|S)z�Return the next item using TarFile's next() method.
           When all members have been read, set TarFile as _loaded.
        Tr)r�r�rg�
StopIterationr�r8�
IndexErrorr=r8r8r9�__next__

s

zTarIter.__next__N)r~rr�r�r�rrJrgr8r8r8r9r@�	s
r@cCs0zt|�}|��WdSty*YdS0dS)zfReturn True if name points to a tar archive that we
       are able to handle, else return False.
    TFN)r�r�r)r�rqr8r8r9r#
sr)N)w�
__future__r�__version__�version�
__author__�__date__Z	__cvsid__�__credits__r2r�r�r r�rPr-rnr�r�r�r��NotImplementedErrorr-ZWindowsError�	NameError�__all__�version_info�__builtin__�builtinsr��_openr3r�r�r(r#r%r$r:r
rNr�r�r�r�rr��CONTTYPEr+r*rOr3r7r^rrOrr�rcr�rSr��setrvr0rDr��S_IFLNK�S_IFREGr)�S_IFDIRr*�S_IFIFOZTSUIDZTSGIDZTSVTXZTUREADZTUWRITEZTUEXECZTGREADZTGWRITEZTGEXECZTOREADZTOWRITEZTOEXECr�r��getfilesystemencodingr:r?rKrWr^rirtr|�	Exceptionrr�r�r�r�r�r�r�r�rFr��objectr�r�r�r�r�r�rrr@rr�r8r8r8r9�<module>sN
����
�
�
�
��i?KT*site-packages/pip/_vendor/distlib/_backport/__pycache__/misc.cpython-39.pyc000064400000002166150564745010022716 0ustar00a

��Re��@s�dZddlZddlZgd�ZzddlmZWneyHd
dd�ZYn0zeZWn&eyxddl	m
Z
d	d
�ZYn0z
ejZWney�dd�ZYn0dS)z/Backports for individual classes and functions.�N)�cache_from_source�callable�fsencode)rTcCs|rdp
d}||S)N�c�o�)Zpy_file�debug�extrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/_backport/misc.pyrsr)�CallablecCs
t|t�S)N)�
isinstancer)�objrrr
rsrcCs<t|t�r|St|t�r&|�t���Stdt|�j��dS)Nzexpect bytes or str, not %s)	r�bytes�str�encode�sys�getfilesystemencoding�	TypeError�type�__name__)�filenamerrr
r"s

�r)T)
�__doc__�osr�__all__�impr�ImportErrorr�	NameError�collectionsrr�AttributeErrorrrrr
�<module>s 
site-packages/pip/_vendor/distlib/_backport/__pycache__/__init__.cpython-39.pyc000064400000001012150564745010023507 0ustar00a

��Re�@sdZdS)aModules copied from Python 3 standard libraries, for internal use only.

Individual classes and functions are found in d2._backport.misc.  Intended
usage is to always import things missing from 3.1 from that module: the
built-in/stdlib objects will be used if found.
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/_backport/__init__.py�<module>�site-packages/pip/_vendor/distlib/_backport/__pycache__/sysconfig.cpython-39.pyc000064400000037206150564745010023772 0ustar00a

��Re�h�@s�dZddlZddlZddlZddlZddlmZmZzddlZWne	yZddl
ZYn0gd�Zdd�Zej
r�ej�eej
��Znee���Zejdkr�ded	d���vr�eej�ee��Zejdkr�d
edd���vr�eej�eee��Zejdk�r*ded
d���v�r*eej�eee��Zdd�Ze�Zdadd�Ze��Ze�d�Zdd�Zdejdd�Zdejdd�Z dejdd�Z!ej�"ej#�Z$ej�"ej%�Z&da'dZ(dd�Z)dd�Z*dd �Z+d!d"�Z,d#d$�Z-d%d&�Z.dId'd(�Z/d)d*�Z0d+d,�Z1d-d.�Z2dJd/d0�Z3d1d2�Z4d3d4�Z5d5d6�Z6e-�dd7fd8d9�Z7e-�dd7fd:d;�Z8d<d=�Z9d>d?�Z:d@dA�Z;dBdC�Z<dDdE�Z=dFdG�Z>e?dHk�r�e>�dS)Kz-Access to Python's configuration information.�N)�pardir�realpath)�get_config_h_filename�get_config_var�get_config_vars�get_makefile_filename�get_path�get_path_names�	get_paths�get_platform�get_python_version�get_scheme_names�parse_config_hcCs&z
t|�WSty |YS0dS�N)r�OSError)�path�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/distlib/_backport/sysconfig.py�_safe_realpath"s
r�ntZpcbuildi����z\pc\vi����z\pcbuild\amd64i�cCs,dD]"}tj�tj�td|��rdSqdS)N)z
Setup.distzSetup.local�ModulesTF)�osr�isfile�join�
_PROJECT_BASE)�fnrrr�is_python_build:srFcCs�ts�ddlm}t�dd�d}||�}|�d�}|s>Jd��|���}t�|�Wd�n1sf0Yt	r�dD] }t�
|d	d
�t�
|dd�qxd
adS)N�)�finder�.�rz
sysconfig.cfgzsysconfig.cfg exists)�posix_prefix�
posix_home�includez{srcdir}/Include�platincludez{projectbase}/.T)�	_cfg_read�	resourcesr�__name__�rsplit�find�	as_stream�_SCHEMES�readfp�
_PYTHON_BUILD�set)rZbackport_package�_finderZ_cfgfile�s�schemerrr�_ensure_cfg_readDs

(r2z\{([^{]*?)\}c
s�t�|�d�r|�d�}nt�}|��}|D]8}|dkr<q.|D]$\}}|�||�rVq@|�|||�q@q.|�d�|��D]H}t|�|����fdd�}|�|�D]\}}|�||t	�
||��q�qzdS)N�globalscs$|�d�}|�vr�|S|�d�S�Nr r��group�Zmatchobj�name��	variablesrr�	_replaceros
z"_expand_globals.<locals>._replacer)r2�has_section�items�tuple�sections�
has_optionr.�remove_section�dict�	_VAR_REPL�sub)�configr3r?�section�option�valuer;rr9r�_expand_globalsYs$

rIz%s.%s.%s�z%s.%srz%s%scs�fdd�}t�||�S)z�In the string `path`, replace tokens like {some.thing} with the
    corresponding value from the map `local_vars`.

    If there is no corresponding value, leave the token unchanged.
    cs8|�d�}|�vr�|S|tjvr.tj|S|�d�Sr4)r6r�environr7��
local_varsrrr;�s


z_subst_vars.<locals>._replacer�rCrD)rrMr;rrLr�_subst_vars�srOcCs0|��}|��D]\}}||vr"q|||<qdSr)�keysr=)�target_dict�
other_dict�target_keys�keyrHrrr�_extend_dict�s
rUcCs`i}|duri}t|t��t�|�D]4\}}tjdvrDtj�|�}tj�t	||��||<q&|S)N)�posixr)
rUrr+r=rr8r�
expanduser�normpathrO)r1�vars�resrTrHrrr�_expand_vars�s
r[cs�fdd�}t�||�S)Ncs$|�d�}|�vr�|S|�d�Sr4r5r7�rYrrr;�s
zformat_value.<locals>._replacerrN)rHrYr;rr\r�format_value�sr]cCstjdkrdStjS)NrVr!)rr8rrrr�_get_default_scheme�s
r^cCs�tj�dd�}dd�}tjdkrBtj�d�p.d}|r8|S||d�Stjdkr|td	�}|r||r`|S|dd
|dtjdd��S|r�|S|dd
�SdS)N�PYTHONUSERBASEcWstj�tjj|��Sr)rrrWr)�argsrrr�joinuser�sz_getuserbase.<locals>.joinuserr�APPDATA�~�Python�darwin�PYTHONFRAMEWORK�Libraryz%d.%drz.local)rrK�getr8�sys�platformr�version_info)�env_basera�baseZ	frameworkrrr�_getuserbase�s$



�rnc	Cst�d�}t�d�}t�d�}|dur*i}i}i}tj|ddd��}|��}Wd�n1s`0Y|D]�}	|	�d�sn|	��d	kr�qn|�|	�}
|
rn|
�d
d�\}}|��}|�	dd	�}
d
|
vr�|||<qnzt
|�}Wn"ty�|�	dd
�||<Yqn0|||<qnt|�
��}d}t|�dk�r�t|�D�]�}||}|�|��pP|�|�}
|
du�r�|
�d
�}d}||v�r�t||�}n�||v�r�d}nx|tjv�r�tj|}n`||v�r�|�d��r�|dd�|v�r�d	}n$d||v�r�d}nt|d|�}nd	||<}|�r�||
��d�}|d|
���||}d
|v�rL|||<n|zt
|�}Wn t�yx|��||<Yn
0|||<|�|�|�d��r�|dd�|v�r�|dd�}||v�r�|||<n|||<|�|��q.�q|��D]"\}}t|t��r�|��||<�q�|�|�|S)z�Parse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    z"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)z\$\(([A-Za-z][A-Za-z0-9_]*)\)z\${([A-Za-z][A-Za-z0-9_]*)}Nzutf-8�surrogateescape)�encoding�errors�#�r rz$$�$)�CFLAGS�LDFLAGS�CPPFLAGSrTF�PY_rJ)�re�compile�codecs�open�	readlines�
startswith�strip�matchr6�replace�int�
ValueError�listrP�lenr>�search�strrrK�end�start�remover=�
isinstance�update)�filenamerY�_variable_rx�_findvar1_rx�_findvar2_rx�done�notdone�f�lines�line�m�n�v�tmpvr:�renamed_variablesr8rH�found�item�after�krrr�_parse_makefile�s�	


&







�


�


r�cCsDtrtj�td�Sttd�r,dttjf}nd}tj�t	d�|d�S)z Return the path of the Makefile.�Makefile�abiflagszconfig-%s%srE�stdlib)
r-rrrr�hasattrri�_PY_VERSION_SHORTr�r)�config_dir_namerrrrKs
rc
Cst�}zt||�WnLty`}z4d|}t|d�rD|d|j}t|��WYd}~n
d}~00t�}z6t|��}t||�Wd�n1s�0YWnLty�}z4d|}t|d�r�|d|j}t|��WYd}~n
d}~00tr�|d|d<dS)z7Initialize the module as appropriate for POSIX systems.z.invalid Python installation: unable to open %s�strerrorz (%s)N�	BLDSHARED�LDSHARED)	rr��IOErrorr�r�rr|rr-)rY�makefile�e�msg�config_hr�rrr�_init_posixVs&

,
r�cCsVtd�|d<td�|d<td�|d<d|d<d	|d
<t|d<tj�ttj��|d<d
S)z+Initialize the module as appropriate for NTr��LIBDEST�
platstdlib�
BINLIBDESTr#�	INCLUDEPYz.pyd�SOz.exe�EXE�VERSION�BINDIRN)r�_PY_VERSION_SHORT_NO_DOTrr�dirnamerri�
executabler\rrr�_init_non_posixrsr�cCs�|duri}t�d�}t�d�}|��}|s.q�|�|�}|rv|�dd�\}}zt|�}WntyjYn0|||<q |�|�}|r d||�d�<q |S)z�Parse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    Nz"#define ([A-Z][A-Za-z0-9_]+) (.*)
z&/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/
r rr)ryrz�readliner�r6r�r�)�fprY�	define_rx�undef_rxr�r�r�r�rrrr�s&




rcCs:tr$tjdkrtj�td�}q,t}ntd�}tj�|d�S)zReturn the path of pyconfig.h.r�PCr$z
pyconfig.h)r-rr8rrrr)�inc_dirrrrr�s
rcCsttt����S)z,Return a tuple containing the schemes names.)r>�sortedr+r?rrrrr
�sr
cCs
t�d�S)z*Return a tuple containing the paths names.r!)r+�optionsrrrrr	�sr	TcCs&t�|rt||�Stt�|��SdS)z�Return a mapping containing an install scheme.

    ``scheme`` is the install scheme name. If not provided, it will
    return the default scheme for the current platform.
    N)r2r[rBr+r=)r1rY�expandrrrr
�s
r
cCst|||�|S)z[Return a path corresponding to the scheme.

    ``scheme`` is the install scheme name.
    )r
)r8r1rYr�rrrr�srcGs�tdu�rviattd<ttd<ttd<ttd<tdtdtd<ttd	<ttd
<ttd<ztjtd<Wnty�d
td<Yn0t	j
dvr�tt�t	j
dkr�tt�tj
dkr�t�td<dtvr�ttd<nttd�td<t�rXt	j
dk�rXt}zt	��}Wnt�yd}Yn0t	j�td��sX||k�rXt	j�|td�}t	j�|�td<tjdk�rvt	��d}t|�d�d�}|dk�r�dD]2}t|}t�dd|�}t�dd|�}|t|<�q�n�dt	jv�rt	jd}dD]0}t|}t�dd|�}|d|}|t|<�q�t�dd
�}	t�d|	�}
|
du�rv|
� d�}t	j�!|��svdD]$}t|}t�dd|�}|t|<�qP|�r�g}|D]}
|�"t�|
���q�|StSdS)ayWith no arguments, return a dictionary of all configuration
    variables relevant for the current platform.

    On Unix, this means every variable defined in Python's installed Makefile;
    On Windows and Mac OS it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    N�prefix�exec_prefix�
py_version�py_version_shortrr�py_version_nodotrm�platbase�projectbaser�rs)rZos2rVz2.6�userbase�srcdirrer�)rv�
BASECFLAGSru�	PY_CFLAGSr�z
-arch\s+\w+\s� z-isysroot [^ 	]*Z	ARCHFLAGSruz-isysroot\s+(\S+)r z-isysroot\s+\S+(\s|$))#�_CONFIG_VARS�_PREFIX�_EXEC_PREFIX�_PY_VERSIONr�rrir��AttributeErrorrr8r�r��versionrnrr-�getcwdrr�isabsrrXrj�unamer��splitryrDrKrhr�r6�exists�append)r`rm�cwdr�Zkernel_version�
major_versionrT�flags�archrur�Zsdk�valsr8rrrr�s�






�



rcCst��|�S)z�Return the value of a single variable using the dictionary returned by
    'get_config_vars()'.

    Equivalent to get_config_vars().get(name)
    )rrh)r8rrrrPsrcCsdtjdkrnd}tj�|�}|dkr(tjStj�d|�}tj|t|�|���}|dkr\dS|dkrhdStjStjd	ks�ttd
�s�tjSt�	�\}}}}}|���
dd�}|�
d
d�}|�
dd�}|dd�dkr�d||fS|dd�dk�r&|ddk�rVd}dt|d�d|dd�f}�n0|dd�dk�rDd||fS|dd�dk�rdd|||fS|dd�d k�r�d }t�
d!�}	|	�|�}
|
�rV|
��}�n�|dd�d"k�rVt�}|�d#�}|}
ztd$�}Wnt�y�YnR0zt�d%|���}
W|��n
|��0|
du�r8d&�|
�d'��d&�dd��}
|�sB|
}|�rV|}d(}|
d&d)k�rd*t��d+d���v�rd,}t��d+�}t�d-|�}ttt|���}t|�d'k�r�|d}n^|d.k�r�d,}nN|d/k�r�d0}n>|d1k�r�d2}n.|d3k�r�d4}n|d5k�r
d6}ntd7|f��n<|d8k�r6tjd9k�rVd:}n |d;v�rVtjd9k�rRd<}nd=}d>|||fS)?a�Return a string that identifies the current platform.

    This is used mainly to distinguish platform-specific build directories and
    platform-specific built distributions.  Typically includes the OS name
    and version and the architecture (as supplied by 'os.uname()'),
    although the exact information included depends on the OS; eg. for IRIX
    the architecture isn't particularly important (IRIX only runs on SGI
    hardware), but for Linux the kernel version isn't particularly
    important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u
       irix-5.3
       irix64-6.2

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win-ia64 (64bit Windows on Itanium)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.
    rz bit (����)�amd64z	win-amd64Zitaniumzwin-ia64rVr��/rsr��_�-N��linuxz%s-%s�sunosr�5�solarisz%d.%srJr�Zirix�aixz%s-%s.%s��cygwinz[\d.]+re�MACOSX_DEPLOYMENT_TARGETz0/System/Library/CoreServices/SystemVersion.plistz=<key>ProductUserVisibleVersion</key>\s*<string>(.*?)</string>rr �macosxz10.4.z-archru�fatz
-arch\s+(\S+))�i386�ppc)r��x86_64�intel)r�r�r�Zfat3)�ppc64r��fat64)r�r�r�r��	universalz%Don't know machine value for archs=%rr�lr�)�PowerPC�Power_Macintoshr�r�z%s-%s-%s) rr8rir�r)rjr��lowerr�r�r�r�ryrzr�r6rrhr|r�r��read�closerr�r�findallr>r�r.r��maxsize)r��i�jZlook�osname�host�releaser��machine�rel_rer�ZcfgvarsZmacverZ
macreleaser��cflagsZarchsrrrrYs�
$


�
�





�

rcCstSr)r�rrrrr�srcCsFtt|����D]0\}\}}|dkr0td|�td||f�qdS)Nrz%s: z
	%s = "%s")�	enumerater�r=�print)�title�data�indexrTrHrrr�_print_dict�srcCsRtdt��tdt��tdt��t�tdt��t�tdt��dS)z*Display all information sysconfig detains.zPlatform: "%s"zPython version: "%s"z!Current installation scheme: "%s"�Paths�	VariablesN)rrrr^rr
rrrrr�_mainsr�__main__)N)N)@�__doc__r{rryri�os.pathrr�configparser�ImportError�ConfigParser�__all__rr�rr�rr�r8r�rrr-r%r2�RawConfigParserr+rzrCrIrkr�r�r�rXr�r�r�r�r��
_USER_BASErOrUr[r]r^rnr�rr�r�rrr
r	r
rrrrrrrr'rrrr�<module>sx"
!
	
v

	#
site-packages/pip/_vendor/distlib/_backport/shutil.py000064400000062153150564745010017006 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Utility functions for copying and archiving files and directory trees.

XXX The functions here don't copy the resource fork or other metadata on Mac.

"""

import os
import sys
import stat
from os.path import abspath
import fnmatch
try:
    from collections.abc import Callable
except ImportError:
    from collections import Callable
import errno
from . import tarfile

try:
    import bz2
    _BZ2_SUPPORTED = True
except ImportError:
    _BZ2_SUPPORTED = False

try:
    from pwd import getpwnam
except ImportError:
    getpwnam = None

try:
    from grp import getgrnam
except ImportError:
    getgrnam = None

__all__ = ["copyfileobj", "copyfile", "copymode", "copystat", "copy", "copy2",
           "copytree", "move", "rmtree", "Error", "SpecialFileError",
           "ExecError", "make_archive", "get_archive_formats",
           "register_archive_format", "unregister_archive_format",
           "get_unpack_formats", "register_unpack_format",
           "unregister_unpack_format", "unpack_archive", "ignore_patterns"]

class Error(EnvironmentError):
    pass

class SpecialFileError(EnvironmentError):
    """Raised when trying to do a kind of operation (e.g. copying) which is
    not supported on a special file (e.g. a named pipe)"""

class ExecError(EnvironmentError):
    """Raised when a command could not be executed"""

class ReadError(EnvironmentError):
    """Raised when an archive cannot be read"""

class RegistryError(Exception):
    """Raised when a registry operation with the archiving
    and unpacking registries fails"""


try:
    WindowsError
except NameError:
    WindowsError = None

def copyfileobj(fsrc, fdst, length=16*1024):
    """copy data from file-like object fsrc to file-like object fdst"""
    while 1:
        buf = fsrc.read(length)
        if not buf:
            break
        fdst.write(buf)

def _samefile(src, dst):
    # Macintosh, Unix.
    if hasattr(os.path, 'samefile'):
        try:
            return os.path.samefile(src, dst)
        except OSError:
            return False

    # All other platforms: check for same pathname.
    return (os.path.normcase(os.path.abspath(src)) ==
            os.path.normcase(os.path.abspath(dst)))

def copyfile(src, dst):
    """Copy data from src to dst"""
    if _samefile(src, dst):
        raise Error("`%s` and `%s` are the same file" % (src, dst))

    for fn in [src, dst]:
        try:
            st = os.stat(fn)
        except OSError:
            # File most likely does not exist
            pass
        else:
            # XXX What about other special files? (sockets, devices...)
            if stat.S_ISFIFO(st.st_mode):
                raise SpecialFileError("`%s` is a named pipe" % fn)

    with open(src, 'rb') as fsrc:
        with open(dst, 'wb') as fdst:
            copyfileobj(fsrc, fdst)

def copymode(src, dst):
    """Copy mode bits from src to dst"""
    if hasattr(os, 'chmod'):
        st = os.stat(src)
        mode = stat.S_IMODE(st.st_mode)
        os.chmod(dst, mode)

def copystat(src, dst):
    """Copy all stat info (mode bits, atime, mtime, flags) from src to dst"""
    st = os.stat(src)
    mode = stat.S_IMODE(st.st_mode)
    if hasattr(os, 'utime'):
        os.utime(dst, (st.st_atime, st.st_mtime))
    if hasattr(os, 'chmod'):
        os.chmod(dst, mode)
    if hasattr(os, 'chflags') and hasattr(st, 'st_flags'):
        try:
            os.chflags(dst, st.st_flags)
        except OSError as why:
            if (not hasattr(errno, 'EOPNOTSUPP') or
                why.errno != errno.EOPNOTSUPP):
                raise

def copy(src, dst):
    """Copy data and mode bits ("cp src dst").

    The destination may be a directory.

    """
    if os.path.isdir(dst):
        dst = os.path.join(dst, os.path.basename(src))
    copyfile(src, dst)
    copymode(src, dst)

def copy2(src, dst):
    """Copy data and all stat info ("cp -p src dst").

    The destination may be a directory.

    """
    if os.path.isdir(dst):
        dst = os.path.join(dst, os.path.basename(src))
    copyfile(src, dst)
    copystat(src, dst)

def ignore_patterns(*patterns):
    """Function that can be used as copytree() ignore parameter.

    Patterns is a sequence of glob-style patterns
    that are used to exclude files"""
    def _ignore_patterns(path, names):
        ignored_names = []
        for pattern in patterns:
            ignored_names.extend(fnmatch.filter(names, pattern))
        return set(ignored_names)
    return _ignore_patterns

def copytree(src, dst, symlinks=False, ignore=None, copy_function=copy2,
             ignore_dangling_symlinks=False):
    """Recursively copy a directory tree.

    The destination directory must not already exist.
    If exception(s) occur, an Error is raised with a list of reasons.

    If the optional symlinks flag is true, symbolic links in the
    source tree result in symbolic links in the destination tree; if
    it is false, the contents of the files pointed to by symbolic
    links are copied. If the file pointed by the symlink doesn't
    exist, an exception will be added in the list of errors raised in
    an Error exception at the end of the copy process.

    You can set the optional ignore_dangling_symlinks flag to true if you
    want to silence this exception. Notice that this has no effect on
    platforms that don't support os.symlink.

    The optional ignore argument is a callable. If given, it
    is called with the `src` parameter, which is the directory
    being visited by copytree(), and `names` which is the list of
    `src` contents, as returned by os.listdir():

        callable(src, names) -> ignored_names

    Since copytree() is called recursively, the callable will be
    called once for each directory that is copied. It returns a
    list of names relative to the `src` directory that should
    not be copied.

    The optional copy_function argument is a callable that will be used
    to copy each file. It will be called with the source path and the
    destination path as arguments. By default, copy2() is used, but any
    function that supports the same signature (like copy()) can be used.

    """
    names = os.listdir(src)
    if ignore is not None:
        ignored_names = ignore(src, names)
    else:
        ignored_names = set()

    os.makedirs(dst)
    errors = []
    for name in names:
        if name in ignored_names:
            continue
        srcname = os.path.join(src, name)
        dstname = os.path.join(dst, name)
        try:
            if os.path.islink(srcname):
                linkto = os.readlink(srcname)
                if symlinks:
                    os.symlink(linkto, dstname)
                else:
                    # ignore dangling symlink if the flag is on
                    if not os.path.exists(linkto) and ignore_dangling_symlinks:
                        continue
                    # otherwise let the copy occurs. copy2 will raise an error
                    copy_function(srcname, dstname)
            elif os.path.isdir(srcname):
                copytree(srcname, dstname, symlinks, ignore, copy_function)
            else:
                # Will raise a SpecialFileError for unsupported file types
                copy_function(srcname, dstname)
        # catch the Error from the recursive copytree so that we can
        # continue with other files
        except Error as err:
            errors.extend(err.args[0])
        except EnvironmentError as why:
            errors.append((srcname, dstname, str(why)))
    try:
        copystat(src, dst)
    except OSError as why:
        if WindowsError is not None and isinstance(why, WindowsError):
            # Copying file access times may fail on Windows
            pass
        else:
            errors.extend((src, dst, str(why)))
    if errors:
        raise Error(errors)

def rmtree(path, ignore_errors=False, onerror=None):
    """Recursively delete a directory tree.

    If ignore_errors is set, errors are ignored; otherwise, if onerror
    is set, it is called to handle the error with arguments (func,
    path, exc_info) where func is os.listdir, os.remove, or os.rmdir;
    path is the argument to that function that caused it to fail; and
    exc_info is a tuple returned by sys.exc_info().  If ignore_errors
    is false and onerror is None, an exception is raised.

    """
    if ignore_errors:
        def onerror(*args):
            pass
    elif onerror is None:
        def onerror(*args):
            raise
    try:
        if os.path.islink(path):
            # symlinks to directories are forbidden, see bug #1669
            raise OSError("Cannot call rmtree on a symbolic link")
    except OSError:
        onerror(os.path.islink, path, sys.exc_info())
        # can't continue even if onerror hook returns
        return
    names = []
    try:
        names = os.listdir(path)
    except os.error:
        onerror(os.listdir, path, sys.exc_info())
    for name in names:
        fullname = os.path.join(path, name)
        try:
            mode = os.lstat(fullname).st_mode
        except os.error:
            mode = 0
        if stat.S_ISDIR(mode):
            rmtree(fullname, ignore_errors, onerror)
        else:
            try:
                os.remove(fullname)
            except os.error:
                onerror(os.remove, fullname, sys.exc_info())
    try:
        os.rmdir(path)
    except os.error:
        onerror(os.rmdir, path, sys.exc_info())


def _basename(path):
    # A basename() variant which first strips the trailing slash, if present.
    # Thus we always get the last component of the path, even for directories.
    return os.path.basename(path.rstrip(os.path.sep))

def move(src, dst):
    """Recursively move a file or directory to another location. This is
    similar to the Unix "mv" command.

    If the destination is a directory or a symlink to a directory, the source
    is moved inside the directory. The destination path must not already
    exist.

    If the destination already exists but is not a directory, it may be
    overwritten depending on os.rename() semantics.

    If the destination is on our current filesystem, then rename() is used.
    Otherwise, src is copied to the destination and then removed.
    A lot more could be done here...  A look at a mv.c shows a lot of
    the issues this implementation glosses over.

    """
    real_dst = dst
    if os.path.isdir(dst):
        if _samefile(src, dst):
            # We might be on a case insensitive filesystem,
            # perform the rename anyway.
            os.rename(src, dst)
            return

        real_dst = os.path.join(dst, _basename(src))
        if os.path.exists(real_dst):
            raise Error("Destination path '%s' already exists" % real_dst)
    try:
        os.rename(src, real_dst)
    except OSError:
        if os.path.isdir(src):
            if _destinsrc(src, dst):
                raise Error("Cannot move a directory '%s' into itself '%s'." % (src, dst))
            copytree(src, real_dst, symlinks=True)
            rmtree(src)
        else:
            copy2(src, real_dst)
            os.unlink(src)

def _destinsrc(src, dst):
    src = abspath(src)
    dst = abspath(dst)
    if not src.endswith(os.path.sep):
        src += os.path.sep
    if not dst.endswith(os.path.sep):
        dst += os.path.sep
    return dst.startswith(src)

def _get_gid(name):
    """Returns a gid, given a group name."""
    if getgrnam is None or name is None:
        return None
    try:
        result = getgrnam(name)
    except KeyError:
        result = None
    if result is not None:
        return result[2]
    return None

def _get_uid(name):
    """Returns an uid, given a user name."""
    if getpwnam is None or name is None:
        return None
    try:
        result = getpwnam(name)
    except KeyError:
        result = None
    if result is not None:
        return result[2]
    return None

def _make_tarball(base_name, base_dir, compress="gzip", verbose=0, dry_run=0,
                  owner=None, group=None, logger=None):
    """Create a (possibly compressed) tar file from all the files under
    'base_dir'.

    'compress' must be "gzip" (the default), "bzip2", or None.

    'owner' and 'group' can be used to define an owner and a group for the
    archive that is being built. If not provided, the current owner and group
    will be used.

    The output tar file will be named 'base_name' +  ".tar", possibly plus
    the appropriate compression extension (".gz", or ".bz2").

    Returns the output filename.
    """
    tar_compression = {'gzip': 'gz', None: ''}
    compress_ext = {'gzip': '.gz'}

    if _BZ2_SUPPORTED:
        tar_compression['bzip2'] = 'bz2'
        compress_ext['bzip2'] = '.bz2'

    # flags for compression program, each element of list will be an argument
    if compress is not None and compress not in compress_ext:
        raise ValueError("bad value for 'compress', or compression format not "
                         "supported : {0}".format(compress))

    archive_name = base_name + '.tar' + compress_ext.get(compress, '')
    archive_dir = os.path.dirname(archive_name)

    if not os.path.exists(archive_dir):
        if logger is not None:
            logger.info("creating %s", archive_dir)
        if not dry_run:
            os.makedirs(archive_dir)

    # creating the tarball
    if logger is not None:
        logger.info('Creating tar archive')

    uid = _get_uid(owner)
    gid = _get_gid(group)

    def _set_uid_gid(tarinfo):
        if gid is not None:
            tarinfo.gid = gid
            tarinfo.gname = group
        if uid is not None:
            tarinfo.uid = uid
            tarinfo.uname = owner
        return tarinfo

    if not dry_run:
        tar = tarfile.open(archive_name, 'w|%s' % tar_compression[compress])
        try:
            tar.add(base_dir, filter=_set_uid_gid)
        finally:
            tar.close()

    return archive_name

def _call_external_zip(base_dir, zip_filename, verbose=False, dry_run=False):
    # XXX see if we want to keep an external call here
    if verbose:
        zipoptions = "-r"
    else:
        zipoptions = "-rq"
    from distutils.errors import DistutilsExecError
    from distutils.spawn import spawn
    try:
        spawn(["zip", zipoptions, zip_filename, base_dir], dry_run=dry_run)
    except DistutilsExecError:
        # XXX really should distinguish between "couldn't find
        # external 'zip' command" and "zip failed".
        raise ExecError("unable to create zip file '%s': "
            "could neither import the 'zipfile' module nor "
            "find a standalone zip utility") % zip_filename

def _make_zipfile(base_name, base_dir, verbose=0, dry_run=0, logger=None):
    """Create a zip file from all the files under 'base_dir'.

    The output zip file will be named 'base_name' + ".zip".  Uses either the
    "zipfile" Python module (if available) or the InfoZIP "zip" utility
    (if installed and found on the default search path).  If neither tool is
    available, raises ExecError.  Returns the name of the output zip
    file.
    """
    zip_filename = base_name + ".zip"
    archive_dir = os.path.dirname(base_name)

    if not os.path.exists(archive_dir):
        if logger is not None:
            logger.info("creating %s", archive_dir)
        if not dry_run:
            os.makedirs(archive_dir)

    # If zipfile module is not available, try spawning an external 'zip'
    # command.
    try:
        import zipfile
    except ImportError:
        zipfile = None

    if zipfile is None:
        _call_external_zip(base_dir, zip_filename, verbose, dry_run)
    else:
        if logger is not None:
            logger.info("creating '%s' and adding '%s' to it",
                        zip_filename, base_dir)

        if not dry_run:
            zip = zipfile.ZipFile(zip_filename, "w",
                                  compression=zipfile.ZIP_DEFLATED)

            for dirpath, dirnames, filenames in os.walk(base_dir):
                for name in filenames:
                    path = os.path.normpath(os.path.join(dirpath, name))
                    if os.path.isfile(path):
                        zip.write(path, path)
                        if logger is not None:
                            logger.info("adding '%s'", path)
            zip.close()

    return zip_filename

_ARCHIVE_FORMATS = {
    'gztar': (_make_tarball, [('compress', 'gzip')], "gzip'ed tar-file"),
    'bztar': (_make_tarball, [('compress', 'bzip2')], "bzip2'ed tar-file"),
    'tar':   (_make_tarball, [('compress', None)], "uncompressed tar file"),
    'zip':   (_make_zipfile, [], "ZIP file"),
    }

if _BZ2_SUPPORTED:
    _ARCHIVE_FORMATS['bztar'] = (_make_tarball, [('compress', 'bzip2')],
                                "bzip2'ed tar-file")

def get_archive_formats():
    """Returns a list of supported formats for archiving and unarchiving.

    Each element of the returned sequence is a tuple (name, description)
    """
    formats = [(name, registry[2]) for name, registry in
               _ARCHIVE_FORMATS.items()]
    formats.sort()
    return formats

def register_archive_format(name, function, extra_args=None, description=''):
    """Registers an archive format.

    name is the name of the format. function is the callable that will be
    used to create archives. If provided, extra_args is a sequence of
    (name, value) tuples that will be passed as arguments to the callable.
    description can be provided to describe the format, and will be returned
    by the get_archive_formats() function.
    """
    if extra_args is None:
        extra_args = []
    if not isinstance(function, Callable):
        raise TypeError('The %s object is not callable' % function)
    if not isinstance(extra_args, (tuple, list)):
        raise TypeError('extra_args needs to be a sequence')
    for element in extra_args:
        if not isinstance(element, (tuple, list)) or len(element) !=2:
            raise TypeError('extra_args elements are : (arg_name, value)')

    _ARCHIVE_FORMATS[name] = (function, extra_args, description)

def unregister_archive_format(name):
    del _ARCHIVE_FORMATS[name]

def make_archive(base_name, format, root_dir=None, base_dir=None, verbose=0,
                 dry_run=0, owner=None, group=None, logger=None):
    """Create an archive file (eg. zip or tar).

    'base_name' is the name of the file to create, minus any format-specific
    extension; 'format' is the archive format: one of "zip", "tar", "bztar"
    or "gztar".

    'root_dir' is a directory that will be the root directory of the
    archive; ie. we typically chdir into 'root_dir' before creating the
    archive.  'base_dir' is the directory where we start archiving from;
    ie. 'base_dir' will be the common prefix of all files and
    directories in the archive.  'root_dir' and 'base_dir' both default
    to the current directory.  Returns the name of the archive file.

    'owner' and 'group' are used when creating a tar archive. By default,
    uses the current owner and group.
    """
    save_cwd = os.getcwd()
    if root_dir is not None:
        if logger is not None:
            logger.debug("changing into '%s'", root_dir)
        base_name = os.path.abspath(base_name)
        if not dry_run:
            os.chdir(root_dir)

    if base_dir is None:
        base_dir = os.curdir

    kwargs = {'dry_run': dry_run, 'logger': logger}

    try:
        format_info = _ARCHIVE_FORMATS[format]
    except KeyError:
        raise ValueError("unknown archive format '%s'" % format)

    func = format_info[0]
    for arg, val in format_info[1]:
        kwargs[arg] = val

    if format != 'zip':
        kwargs['owner'] = owner
        kwargs['group'] = group

    try:
        filename = func(base_name, base_dir, **kwargs)
    finally:
        if root_dir is not None:
            if logger is not None:
                logger.debug("changing back to '%s'", save_cwd)
            os.chdir(save_cwd)

    return filename


def get_unpack_formats():
    """Returns a list of supported formats for unpacking.

    Each element of the returned sequence is a tuple
    (name, extensions, description)
    """
    formats = [(name, info[0], info[3]) for name, info in
               _UNPACK_FORMATS.items()]
    formats.sort()
    return formats

def _check_unpack_options(extensions, function, extra_args):
    """Checks what gets registered as an unpacker."""
    # first make sure no other unpacker is registered for this extension
    existing_extensions = {}
    for name, info in _UNPACK_FORMATS.items():
        for ext in info[0]:
            existing_extensions[ext] = name

    for extension in extensions:
        if extension in existing_extensions:
            msg = '%s is already registered for "%s"'
            raise RegistryError(msg % (extension,
                                       existing_extensions[extension]))

    if not isinstance(function, Callable):
        raise TypeError('The registered function must be a callable')


def register_unpack_format(name, extensions, function, extra_args=None,
                           description=''):
    """Registers an unpack format.

    `name` is the name of the format. `extensions` is a list of extensions
    corresponding to the format.

    `function` is the callable that will be
    used to unpack archives. The callable will receive archives to unpack.
    If it's unable to handle an archive, it needs to raise a ReadError
    exception.

    If provided, `extra_args` is a sequence of
    (name, value) tuples that will be passed as arguments to the callable.
    description can be provided to describe the format, and will be returned
    by the get_unpack_formats() function.
    """
    if extra_args is None:
        extra_args = []
    _check_unpack_options(extensions, function, extra_args)
    _UNPACK_FORMATS[name] = extensions, function, extra_args, description

def unregister_unpack_format(name):
    """Removes the pack format from the registry."""
    del _UNPACK_FORMATS[name]

def _ensure_directory(path):
    """Ensure that the parent directory of `path` exists"""
    dirname = os.path.dirname(path)
    if not os.path.isdir(dirname):
        os.makedirs(dirname)

def _unpack_zipfile(filename, extract_dir):
    """Unpack zip `filename` to `extract_dir`
    """
    try:
        import zipfile
    except ImportError:
        raise ReadError('zlib not supported, cannot unpack this archive.')

    if not zipfile.is_zipfile(filename):
        raise ReadError("%s is not a zip file" % filename)

    zip = zipfile.ZipFile(filename)
    try:
        for info in zip.infolist():
            name = info.filename

            # don't extract absolute paths or ones with .. in them
            if name.startswith('/') or '..' in name:
                continue

            target = os.path.join(extract_dir, *name.split('/'))
            if not target:
                continue

            _ensure_directory(target)
            if not name.endswith('/'):
                # file
                data = zip.read(info.filename)
                f = open(target, 'wb')
                try:
                    f.write(data)
                finally:
                    f.close()
                    del data
    finally:
        zip.close()

def _unpack_tarfile(filename, extract_dir):
    """Unpack tar/tar.gz/tar.bz2 `filename` to `extract_dir`
    """
    try:
        tarobj = tarfile.open(filename)
    except tarfile.TarError:
        raise ReadError(
            "%s is not a compressed or uncompressed tar file" % filename)
    try:
        tarobj.extractall(extract_dir)
    finally:
        tarobj.close()

_UNPACK_FORMATS = {
    'gztar': (['.tar.gz', '.tgz'], _unpack_tarfile, [], "gzip'ed tar-file"),
    'tar':   (['.tar'], _unpack_tarfile, [], "uncompressed tar file"),
    'zip':   (['.zip'], _unpack_zipfile, [], "ZIP file")
    }

if _BZ2_SUPPORTED:
    _UNPACK_FORMATS['bztar'] = (['.bz2'], _unpack_tarfile, [],
                                "bzip2'ed tar-file")

def _find_unpack_format(filename):
    for name, info in _UNPACK_FORMATS.items():
        for extension in info[0]:
            if filename.endswith(extension):
                return name
    return None

def unpack_archive(filename, extract_dir=None, format=None):
    """Unpack an archive.

    `filename` is the name of the archive.

    `extract_dir` is the name of the target directory, where the archive
    is unpacked. If not provided, the current working directory is used.

    `format` is the archive format: one of "zip", "tar", or "gztar". Or any
    other registered format. If not provided, unpack_archive will use the
    filename extension and see if an unpacker was registered for that
    extension.

    In case none is found, a ValueError is raised.
    """
    if extract_dir is None:
        extract_dir = os.getcwd()

    if format is not None:
        try:
            format_info = _UNPACK_FORMATS[format]
        except KeyError:
            raise ValueError("Unknown unpack format '{0}'".format(format))

        func = format_info[1]
        func(filename, extract_dir, **dict(format_info[2]))
    else:
        # we need to look at the registered unpackers supported extensions
        format = _find_unpack_format(filename)
        if format is None:
            raise ReadError("Unknown archive format '{0}'".format(filename))

        func = _UNPACK_FORMATS[format][1]
        kwargs = dict(_UNPACK_FORMATS[format][2])
        func(filename, extract_dir, **kwargs)
site-packages/pip/_vendor/distlib/_backport/misc.py000064400000001713150564745010016424 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Backports for individual classes and functions."""

import os
import sys

__all__ = ['cache_from_source', 'callable', 'fsencode']


try:
    from imp import cache_from_source
except ImportError:
    def cache_from_source(py_file, debug=__debug__):
        ext = debug and 'c' or 'o'
        return py_file + ext


try:
    callable = callable
except NameError:
    from collections import Callable

    def callable(obj):
        return isinstance(obj, Callable)


try:
    fsencode = os.fsencode
except AttributeError:
    def fsencode(filename):
        if isinstance(filename, bytes):
            return filename
        elif isinstance(filename, str):
            return filename.encode(sys.getfilesystemencoding())
        else:
            raise TypeError("expect bytes or str, not %s" %
                            type(filename).__name__)
site-packages/pip/_vendor/distlib/_backport/__init__.py000064400000000422150564745010017224 0ustar00"""Modules copied from Python 3 standard libraries, for internal use only.

Individual classes and functions are found in d2._backport.misc.  Intended
usage is to always import things missing from 3.1 from that module: the
built-in/stdlib objects will be used if found.
"""
site-packages/pip/_vendor/distlib/_backport/sysconfig.py000064400000064346150564745010017510 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012 The Python Software Foundation.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
"""Access to Python's configuration information."""

import codecs
import os
import re
import sys
from os.path import pardir, realpath
try:
    import configparser
except ImportError:
    import ConfigParser as configparser


__all__ = [
    'get_config_h_filename',
    'get_config_var',
    'get_config_vars',
    'get_makefile_filename',
    'get_path',
    'get_path_names',
    'get_paths',
    'get_platform',
    'get_python_version',
    'get_scheme_names',
    'parse_config_h',
]


def _safe_realpath(path):
    try:
        return realpath(path)
    except OSError:
        return path


if sys.executable:
    _PROJECT_BASE = os.path.dirname(_safe_realpath(sys.executable))
else:
    # sys.executable can be empty if argv[0] has been changed and Python is
    # unable to retrieve the real program name
    _PROJECT_BASE = _safe_realpath(os.getcwd())

if os.name == "nt" and "pcbuild" in _PROJECT_BASE[-8:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir))
# PC/VS7.1
if os.name == "nt" and "\\pc\\v" in _PROJECT_BASE[-10:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir))
# PC/AMD64
if os.name == "nt" and "\\pcbuild\\amd64" in _PROJECT_BASE[-14:].lower():
    _PROJECT_BASE = _safe_realpath(os.path.join(_PROJECT_BASE, pardir, pardir))


def is_python_build():
    for fn in ("Setup.dist", "Setup.local"):
        if os.path.isfile(os.path.join(_PROJECT_BASE, "Modules", fn)):
            return True
    return False

_PYTHON_BUILD = is_python_build()

_cfg_read = False

def _ensure_cfg_read():
    global _cfg_read
    if not _cfg_read:
        from ..resources import finder
        backport_package = __name__.rsplit('.', 1)[0]
        _finder = finder(backport_package)
        _cfgfile = _finder.find('sysconfig.cfg')
        assert _cfgfile, 'sysconfig.cfg exists'
        with _cfgfile.as_stream() as s:
            _SCHEMES.readfp(s)
        if _PYTHON_BUILD:
            for scheme in ('posix_prefix', 'posix_home'):
                _SCHEMES.set(scheme, 'include', '{srcdir}/Include')
                _SCHEMES.set(scheme, 'platinclude', '{projectbase}/.')

        _cfg_read = True


_SCHEMES = configparser.RawConfigParser()
_VAR_REPL = re.compile(r'\{([^{]*?)\}')

def _expand_globals(config):
    _ensure_cfg_read()
    if config.has_section('globals'):
        globals = config.items('globals')
    else:
        globals = tuple()

    sections = config.sections()
    for section in sections:
        if section == 'globals':
            continue
        for option, value in globals:
            if config.has_option(section, option):
                continue
            config.set(section, option, value)
    config.remove_section('globals')

    # now expanding local variables defined in the cfg file
    #
    for section in config.sections():
        variables = dict(config.items(section))

        def _replacer(matchobj):
            name = matchobj.group(1)
            if name in variables:
                return variables[name]
            return matchobj.group(0)

        for option, value in config.items(section):
            config.set(section, option, _VAR_REPL.sub(_replacer, value))

#_expand_globals(_SCHEMES)

_PY_VERSION = '%s.%s.%s' % sys.version_info[:3]
_PY_VERSION_SHORT = '%s.%s' % sys.version_info[:2]
_PY_VERSION_SHORT_NO_DOT = '%s%s' % sys.version_info[:2]
_PREFIX = os.path.normpath(sys.prefix)
_EXEC_PREFIX = os.path.normpath(sys.exec_prefix)
_CONFIG_VARS = None
_USER_BASE = None


def _subst_vars(path, local_vars):
    """In the string `path`, replace tokens like {some.thing} with the
    corresponding value from the map `local_vars`.

    If there is no corresponding value, leave the token unchanged.
    """
    def _replacer(matchobj):
        name = matchobj.group(1)
        if name in local_vars:
            return local_vars[name]
        elif name in os.environ:
            return os.environ[name]
        return matchobj.group(0)
    return _VAR_REPL.sub(_replacer, path)


def _extend_dict(target_dict, other_dict):
    target_keys = target_dict.keys()
    for key, value in other_dict.items():
        if key in target_keys:
            continue
        target_dict[key] = value


def _expand_vars(scheme, vars):
    res = {}
    if vars is None:
        vars = {}
    _extend_dict(vars, get_config_vars())

    for key, value in _SCHEMES.items(scheme):
        if os.name in ('posix', 'nt'):
            value = os.path.expanduser(value)
        res[key] = os.path.normpath(_subst_vars(value, vars))
    return res


def format_value(value, vars):
    def _replacer(matchobj):
        name = matchobj.group(1)
        if name in vars:
            return vars[name]
        return matchobj.group(0)
    return _VAR_REPL.sub(_replacer, value)


def _get_default_scheme():
    if os.name == 'posix':
        # the default scheme for posix is posix_prefix
        return 'posix_prefix'
    return os.name


def _getuserbase():
    env_base = os.environ.get("PYTHONUSERBASE", None)

    def joinuser(*args):
        return os.path.expanduser(os.path.join(*args))

    # what about 'os2emx', 'riscos' ?
    if os.name == "nt":
        base = os.environ.get("APPDATA") or "~"
        if env_base:
            return env_base
        else:
            return joinuser(base, "Python")

    if sys.platform == "darwin":
        framework = get_config_var("PYTHONFRAMEWORK")
        if framework:
            if env_base:
                return env_base
            else:
                return joinuser("~", "Library", framework, "%d.%d" %
                                sys.version_info[:2])

    if env_base:
        return env_base
    else:
        return joinuser("~", ".local")


def _parse_makefile(filename, vars=None):
    """Parse a Makefile-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    # Regexes needed for parsing Makefile (and similar syntaxes,
    # like old-style Setup files).
    _variable_rx = re.compile(r"([a-zA-Z][a-zA-Z0-9_]+)\s*=\s*(.*)")
    _findvar1_rx = re.compile(r"\$\(([A-Za-z][A-Za-z0-9_]*)\)")
    _findvar2_rx = re.compile(r"\${([A-Za-z][A-Za-z0-9_]*)}")

    if vars is None:
        vars = {}
    done = {}
    notdone = {}

    with codecs.open(filename, encoding='utf-8', errors="surrogateescape") as f:
        lines = f.readlines()

    for line in lines:
        if line.startswith('#') or line.strip() == '':
            continue
        m = _variable_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            v = v.strip()
            # `$$' is a literal `$' in make
            tmpv = v.replace('$$', '')

            if "$" in tmpv:
                notdone[n] = v
            else:
                try:
                    v = int(v)
                except ValueError:
                    # insert literal `$'
                    done[n] = v.replace('$$', '$')
                else:
                    done[n] = v

    # do variable interpolation here
    variables = list(notdone.keys())

    # Variables with a 'PY_' prefix in the makefile. These need to
    # be made available without that prefix through sysconfig.
    # Special care is needed to ensure that variable expansion works, even
    # if the expansion uses the name without a prefix.
    renamed_variables = ('CFLAGS', 'LDFLAGS', 'CPPFLAGS')

    while len(variables) > 0:
        for name in tuple(variables):
            value = notdone[name]
            m = _findvar1_rx.search(value) or _findvar2_rx.search(value)
            if m is not None:
                n = m.group(1)
                found = True
                if n in done:
                    item = str(done[n])
                elif n in notdone:
                    # get it on a subsequent round
                    found = False
                elif n in os.environ:
                    # do it like make: fall back to environment
                    item = os.environ[n]

                elif n in renamed_variables:
                    if (name.startswith('PY_') and
                        name[3:] in renamed_variables):
                        item = ""

                    elif 'PY_' + n in notdone:
                        found = False

                    else:
                        item = str(done['PY_' + n])

                else:
                    done[n] = item = ""

                if found:
                    after = value[m.end():]
                    value = value[:m.start()] + item + after
                    if "$" in after:
                        notdone[name] = value
                    else:
                        try:
                            value = int(value)
                        except ValueError:
                            done[name] = value.strip()
                        else:
                            done[name] = value
                        variables.remove(name)

                        if (name.startswith('PY_') and
                            name[3:] in renamed_variables):

                            name = name[3:]
                            if name not in done:
                                done[name] = value

            else:
                # bogus variable reference (e.g. "prefix=$/opt/python");
                # just drop it since we can't deal
                done[name] = value
                variables.remove(name)

    # strip spurious spaces
    for k, v in done.items():
        if isinstance(v, str):
            done[k] = v.strip()

    # save the results in the global dictionary
    vars.update(done)
    return vars


def get_makefile_filename():
    """Return the path of the Makefile."""
    if _PYTHON_BUILD:
        return os.path.join(_PROJECT_BASE, "Makefile")
    if hasattr(sys, 'abiflags'):
        config_dir_name = 'config-%s%s' % (_PY_VERSION_SHORT, sys.abiflags)
    else:
        config_dir_name = 'config'
    return os.path.join(get_path('stdlib'), config_dir_name, 'Makefile')


def _init_posix(vars):
    """Initialize the module as appropriate for POSIX systems."""
    # load the installed Makefile:
    makefile = get_makefile_filename()
    try:
        _parse_makefile(makefile, vars)
    except IOError as e:
        msg = "invalid Python installation: unable to open %s" % makefile
        if hasattr(e, "strerror"):
            msg = msg + " (%s)" % e.strerror
        raise IOError(msg)
    # load the installed pyconfig.h:
    config_h = get_config_h_filename()
    try:
        with open(config_h) as f:
            parse_config_h(f, vars)
    except IOError as e:
        msg = "invalid Python installation: unable to open %s" % config_h
        if hasattr(e, "strerror"):
            msg = msg + " (%s)" % e.strerror
        raise IOError(msg)
    # On AIX, there are wrong paths to the linker scripts in the Makefile
    # -- these paths are relative to the Python source, but when installed
    # the scripts are in another directory.
    if _PYTHON_BUILD:
        vars['LDSHARED'] = vars['BLDSHARED']


def _init_non_posix(vars):
    """Initialize the module as appropriate for NT"""
    # set basic install directories
    vars['LIBDEST'] = get_path('stdlib')
    vars['BINLIBDEST'] = get_path('platstdlib')
    vars['INCLUDEPY'] = get_path('include')
    vars['SO'] = '.pyd'
    vars['EXE'] = '.exe'
    vars['VERSION'] = _PY_VERSION_SHORT_NO_DOT
    vars['BINDIR'] = os.path.dirname(_safe_realpath(sys.executable))

#
# public APIs
#


def parse_config_h(fp, vars=None):
    """Parse a config.h-style file.

    A dictionary containing name/value pairs is returned.  If an
    optional dictionary is passed in as the second argument, it is
    used instead of a new dictionary.
    """
    if vars is None:
        vars = {}
    define_rx = re.compile("#define ([A-Z][A-Za-z0-9_]+) (.*)\n")
    undef_rx = re.compile("/[*] #undef ([A-Z][A-Za-z0-9_]+) [*]/\n")

    while True:
        line = fp.readline()
        if not line:
            break
        m = define_rx.match(line)
        if m:
            n, v = m.group(1, 2)
            try:
                v = int(v)
            except ValueError:
                pass
            vars[n] = v
        else:
            m = undef_rx.match(line)
            if m:
                vars[m.group(1)] = 0
    return vars


def get_config_h_filename():
    """Return the path of pyconfig.h."""
    if _PYTHON_BUILD:
        if os.name == "nt":
            inc_dir = os.path.join(_PROJECT_BASE, "PC")
        else:
            inc_dir = _PROJECT_BASE
    else:
        inc_dir = get_path('platinclude')
    return os.path.join(inc_dir, 'pyconfig.h')


def get_scheme_names():
    """Return a tuple containing the schemes names."""
    return tuple(sorted(_SCHEMES.sections()))


def get_path_names():
    """Return a tuple containing the paths names."""
    # xxx see if we want a static list
    return _SCHEMES.options('posix_prefix')


def get_paths(scheme=_get_default_scheme(), vars=None, expand=True):
    """Return a mapping containing an install scheme.

    ``scheme`` is the install scheme name. If not provided, it will
    return the default scheme for the current platform.
    """
    _ensure_cfg_read()
    if expand:
        return _expand_vars(scheme, vars)
    else:
        return dict(_SCHEMES.items(scheme))


def get_path(name, scheme=_get_default_scheme(), vars=None, expand=True):
    """Return a path corresponding to the scheme.

    ``scheme`` is the install scheme name.
    """
    return get_paths(scheme, vars, expand)[name]


def get_config_vars(*args):
    """With no arguments, return a dictionary of all configuration
    variables relevant for the current platform.

    On Unix, this means every variable defined in Python's installed Makefile;
    On Windows and Mac OS it's a much smaller set.

    With arguments, return a list of values that result from looking up
    each argument in the configuration variable dictionary.
    """
    global _CONFIG_VARS
    if _CONFIG_VARS is None:
        _CONFIG_VARS = {}
        # Normalized versions of prefix and exec_prefix are handy to have;
        # in fact, these are the standard versions used most places in the
        # distutils2 module.
        _CONFIG_VARS['prefix'] = _PREFIX
        _CONFIG_VARS['exec_prefix'] = _EXEC_PREFIX
        _CONFIG_VARS['py_version'] = _PY_VERSION
        _CONFIG_VARS['py_version_short'] = _PY_VERSION_SHORT
        _CONFIG_VARS['py_version_nodot'] = _PY_VERSION[0] + _PY_VERSION[2]
        _CONFIG_VARS['base'] = _PREFIX
        _CONFIG_VARS['platbase'] = _EXEC_PREFIX
        _CONFIG_VARS['projectbase'] = _PROJECT_BASE
        try:
            _CONFIG_VARS['abiflags'] = sys.abiflags
        except AttributeError:
            # sys.abiflags may not be defined on all platforms.
            _CONFIG_VARS['abiflags'] = ''

        if os.name in ('nt', 'os2'):
            _init_non_posix(_CONFIG_VARS)
        if os.name == 'posix':
            _init_posix(_CONFIG_VARS)
        # Setting 'userbase' is done below the call to the
        # init function to enable using 'get_config_var' in
        # the init-function.
        if sys.version >= '2.6':
            _CONFIG_VARS['userbase'] = _getuserbase()

        if 'srcdir' not in _CONFIG_VARS:
            _CONFIG_VARS['srcdir'] = _PROJECT_BASE
        else:
            _CONFIG_VARS['srcdir'] = _safe_realpath(_CONFIG_VARS['srcdir'])

        # Convert srcdir into an absolute path if it appears necessary.
        # Normally it is relative to the build directory.  However, during
        # testing, for example, we might be running a non-installed python
        # from a different directory.
        if _PYTHON_BUILD and os.name == "posix":
            base = _PROJECT_BASE
            try:
                cwd = os.getcwd()
            except OSError:
                cwd = None
            if (not os.path.isabs(_CONFIG_VARS['srcdir']) and
                base != cwd):
                # srcdir is relative and we are not in the same directory
                # as the executable. Assume executable is in the build
                # directory and make srcdir absolute.
                srcdir = os.path.join(base, _CONFIG_VARS['srcdir'])
                _CONFIG_VARS['srcdir'] = os.path.normpath(srcdir)

        if sys.platform == 'darwin':
            kernel_version = os.uname()[2]  # Kernel version (8.4.3)
            major_version = int(kernel_version.split('.')[0])

            if major_version < 8:
                # On Mac OS X before 10.4, check if -arch and -isysroot
                # are in CFLAGS or LDFLAGS and remove them if they are.
                # This is needed when building extensions on a 10.3 system
                # using a universal build of python.
                for key in ('LDFLAGS', 'BASECFLAGS',
                        # a number of derived variables. These need to be
                        # patched up as well.
                        'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):
                    flags = _CONFIG_VARS[key]
                    flags = re.sub(r'-arch\s+\w+\s', ' ', flags)
                    flags = re.sub('-isysroot [^ \t]*', ' ', flags)
                    _CONFIG_VARS[key] = flags
            else:
                # Allow the user to override the architecture flags using
                # an environment variable.
                # NOTE: This name was introduced by Apple in OSX 10.5 and
                # is used by several scripting languages distributed with
                # that OS release.
                if 'ARCHFLAGS' in os.environ:
                    arch = os.environ['ARCHFLAGS']
                    for key in ('LDFLAGS', 'BASECFLAGS',
                        # a number of derived variables. These need to be
                        # patched up as well.
                        'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):

                        flags = _CONFIG_VARS[key]
                        flags = re.sub(r'-arch\s+\w+\s', ' ', flags)
                        flags = flags + ' ' + arch
                        _CONFIG_VARS[key] = flags

                # If we're on OSX 10.5 or later and the user tries to
                # compiles an extension using an SDK that is not present
                # on the current machine it is better to not use an SDK
                # than to fail.
                #
                # The major usecase for this is users using a Python.org
                # binary installer  on OSX 10.6: that installer uses
                # the 10.4u SDK, but that SDK is not installed by default
                # when you install Xcode.
                #
                CFLAGS = _CONFIG_VARS.get('CFLAGS', '')
                m = re.search(r'-isysroot\s+(\S+)', CFLAGS)
                if m is not None:
                    sdk = m.group(1)
                    if not os.path.exists(sdk):
                        for key in ('LDFLAGS', 'BASECFLAGS',
                             # a number of derived variables. These need to be
                             # patched up as well.
                            'CFLAGS', 'PY_CFLAGS', 'BLDSHARED'):

                            flags = _CONFIG_VARS[key]
                            flags = re.sub(r'-isysroot\s+\S+(\s|$)', ' ', flags)
                            _CONFIG_VARS[key] = flags

    if args:
        vals = []
        for name in args:
            vals.append(_CONFIG_VARS.get(name))
        return vals
    else:
        return _CONFIG_VARS


def get_config_var(name):
    """Return the value of a single variable using the dictionary returned by
    'get_config_vars()'.

    Equivalent to get_config_vars().get(name)
    """
    return get_config_vars().get(name)


def get_platform():
    """Return a string that identifies the current platform.

    This is used mainly to distinguish platform-specific build directories and
    platform-specific built distributions.  Typically includes the OS name
    and version and the architecture (as supplied by 'os.uname()'),
    although the exact information included depends on the OS; eg. for IRIX
    the architecture isn't particularly important (IRIX only runs on SGI
    hardware), but for Linux the kernel version isn't particularly
    important.

    Examples of returned values:
       linux-i586
       linux-alpha (?)
       solaris-2.6-sun4u
       irix-5.3
       irix64-6.2

    Windows will return one of:
       win-amd64 (64bit Windows on AMD64 (aka x86_64, Intel64, EM64T, etc)
       win-ia64 (64bit Windows on Itanium)
       win32 (all others - specifically, sys.platform is returned)

    For other non-POSIX platforms, currently just returns 'sys.platform'.
    """
    if os.name == 'nt':
        # sniff sys.version for architecture.
        prefix = " bit ("
        i = sys.version.find(prefix)
        if i == -1:
            return sys.platform
        j = sys.version.find(")", i)
        look = sys.version[i+len(prefix):j].lower()
        if look == 'amd64':
            return 'win-amd64'
        if look == 'itanium':
            return 'win-ia64'
        return sys.platform

    if os.name != "posix" or not hasattr(os, 'uname'):
        # XXX what about the architecture? NT is Intel or Alpha,
        # Mac OS is M68k or PPC, etc.
        return sys.platform

    # Try to distinguish various flavours of Unix
    osname, host, release, version, machine = os.uname()

    # Convert the OS name to lowercase, remove '/' characters
    # (to accommodate BSD/OS), and translate spaces (for "Power Macintosh")
    osname = osname.lower().replace('/', '')
    machine = machine.replace(' ', '_')
    machine = machine.replace('/', '-')

    if osname[:5] == "linux":
        # At least on Linux/Intel, 'machine' is the processor --
        # i386, etc.
        # XXX what about Alpha, SPARC, etc?
        return  "%s-%s" % (osname, machine)
    elif osname[:5] == "sunos":
        if release[0] >= "5":           # SunOS 5 == Solaris 2
            osname = "solaris"
            release = "%d.%s" % (int(release[0]) - 3, release[2:])
        # fall through to standard osname-release-machine representation
    elif osname[:4] == "irix":              # could be "irix64"!
        return "%s-%s" % (osname, release)
    elif osname[:3] == "aix":
        return "%s-%s.%s" % (osname, version, release)
    elif osname[:6] == "cygwin":
        osname = "cygwin"
        rel_re = re.compile(r'[\d.]+')
        m = rel_re.match(release)
        if m:
            release = m.group()
    elif osname[:6] == "darwin":
        #
        # For our purposes, we'll assume that the system version from
        # distutils' perspective is what MACOSX_DEPLOYMENT_TARGET is set
        # to. This makes the compatibility story a bit more sane because the
        # machine is going to compile and link as if it were
        # MACOSX_DEPLOYMENT_TARGET.
        cfgvars = get_config_vars()
        macver = cfgvars.get('MACOSX_DEPLOYMENT_TARGET')

        if True:
            # Always calculate the release of the running machine,
            # needed to determine if we can build fat binaries or not.

            macrelease = macver
            # Get the system version. Reading this plist is a documented
            # way to get the system version (see the documentation for
            # the Gestalt Manager)
            try:
                f = open('/System/Library/CoreServices/SystemVersion.plist')
            except IOError:
                # We're on a plain darwin box, fall back to the default
                # behaviour.
                pass
            else:
                try:
                    m = re.search(r'<key>ProductUserVisibleVersion</key>\s*'
                                  r'<string>(.*?)</string>', f.read())
                finally:
                    f.close()
                if m is not None:
                    macrelease = '.'.join(m.group(1).split('.')[:2])
                # else: fall back to the default behaviour

        if not macver:
            macver = macrelease

        if macver:
            release = macver
            osname = "macosx"

            if ((macrelease + '.') >= '10.4.' and
                '-arch' in get_config_vars().get('CFLAGS', '').strip()):
                # The universal build will build fat binaries, but not on
                # systems before 10.4
                #
                # Try to detect 4-way universal builds, those have machine-type
                # 'universal' instead of 'fat'.

                machine = 'fat'
                cflags = get_config_vars().get('CFLAGS')

                archs = re.findall(r'-arch\s+(\S+)', cflags)
                archs = tuple(sorted(set(archs)))

                if len(archs) == 1:
                    machine = archs[0]
                elif archs == ('i386', 'ppc'):
                    machine = 'fat'
                elif archs == ('i386', 'x86_64'):
                    machine = 'intel'
                elif archs == ('i386', 'ppc', 'x86_64'):
                    machine = 'fat3'
                elif archs == ('ppc64', 'x86_64'):
                    machine = 'fat64'
                elif archs == ('i386', 'ppc', 'ppc64', 'x86_64'):
                    machine = 'universal'
                else:
                    raise ValueError(
                       "Don't know machine value for archs=%r" % (archs,))

            elif machine == 'i386':
                # On OSX the machine type returned by uname is always the
                # 32-bit variant, even if the executable architecture is
                # the 64-bit variant
                if sys.maxsize >= 2**32:
                    machine = 'x86_64'

            elif machine in ('PowerPC', 'Power_Macintosh'):
                # Pick a sane name for the PPC architecture.
                # See 'i386' case
                if sys.maxsize >= 2**32:
                    machine = 'ppc64'
                else:
                    machine = 'ppc'

    return "%s-%s-%s" % (osname, release, machine)


def get_python_version():
    return _PY_VERSION_SHORT


def _print_dict(title, data):
    for index, (key, value) in enumerate(sorted(data.items())):
        if index == 0:
            print('%s: ' % (title))
        print('\t%s = "%s"' % (key, value))


def _main():
    """Display all information sysconfig detains."""
    print('Platform: "%s"' % get_platform())
    print('Python version: "%s"' % get_python_version())
    print('Current installation scheme: "%s"' % _get_default_scheme())
    print()
    _print_dict('Paths', get_paths())
    print()
    _print_dict('Variables', get_config_vars())


if __name__ == '__main__':
    _main()
site-packages/pip/_vendor/distlib/_backport/sysconfig.cfg000064400000005071150564745010017605 0ustar00[posix_prefix]
# Configuration directories.  Some of these come straight out of the
# configure script.  They are for implementing the other variables, not to
# be used directly in [resource_locations].
confdir = /etc
datadir = /usr/share
libdir = /usr/lib
statedir = /var
# User resource directory
local = ~/.local/{distribution.name}

stdlib = {base}/lib/python{py_version_short}
platstdlib = {platbase}/lib/python{py_version_short}
purelib = {base}/lib/python{py_version_short}/site-packages
platlib = {platbase}/lib/python{py_version_short}/site-packages
include = {base}/include/python{py_version_short}{abiflags}
platinclude = {platbase}/include/python{py_version_short}{abiflags}
data = {base}

[posix_home]
stdlib = {base}/lib/python
platstdlib = {base}/lib/python
purelib = {base}/lib/python
platlib = {base}/lib/python
include = {base}/include/python
platinclude = {base}/include/python
scripts = {base}/bin
data = {base}

[nt]
stdlib = {base}/Lib
platstdlib = {base}/Lib
purelib = {base}/Lib/site-packages
platlib = {base}/Lib/site-packages
include = {base}/Include
platinclude = {base}/Include
scripts = {base}/Scripts
data = {base}

[os2]
stdlib = {base}/Lib
platstdlib = {base}/Lib
purelib = {base}/Lib/site-packages
platlib = {base}/Lib/site-packages
include = {base}/Include
platinclude = {base}/Include
scripts = {base}/Scripts
data = {base}

[os2_home]
stdlib = {userbase}/lib/python{py_version_short}
platstdlib = {userbase}/lib/python{py_version_short}
purelib = {userbase}/lib/python{py_version_short}/site-packages
platlib = {userbase}/lib/python{py_version_short}/site-packages
include = {userbase}/include/python{py_version_short}
scripts = {userbase}/bin
data = {userbase}

[nt_user]
stdlib = {userbase}/Python{py_version_nodot}
platstdlib = {userbase}/Python{py_version_nodot}
purelib = {userbase}/Python{py_version_nodot}/site-packages
platlib = {userbase}/Python{py_version_nodot}/site-packages
include = {userbase}/Python{py_version_nodot}/Include
scripts = {userbase}/Scripts
data = {userbase}

[posix_user]
stdlib = {userbase}/lib/python{py_version_short}
platstdlib = {userbase}/lib/python{py_version_short}
purelib = {userbase}/lib/python{py_version_short}/site-packages
platlib = {userbase}/lib/python{py_version_short}/site-packages
include = {userbase}/include/python{py_version_short}
scripts = {userbase}/bin
data = {userbase}

[osx_framework_user]
stdlib = {userbase}/lib/python
platstdlib = {userbase}/lib/python
purelib = {userbase}/lib/python/site-packages
platlib = {userbase}/lib/python/site-packages
include = {userbase}/include
scripts = {userbase}/bin
data = {userbase}
site-packages/pip/_vendor/distlib/_backport/tarfile.py000064400000264724150564745010017134 0ustar00#-------------------------------------------------------------------
# tarfile.py
#-------------------------------------------------------------------
# Copyright (C) 2002 Lars Gustaebel <lars@gustaebel.de>
# All rights reserved.
#
# Permission  is  hereby granted,  free  of charge,  to  any person
# obtaining a  copy of  this software  and associated documentation
# files  (the  "Software"),  to   deal  in  the  Software   without
# restriction,  including  without limitation  the  rights to  use,
# copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies  of  the  Software,  and to  permit  persons  to  whom the
# Software  is  furnished  to  do  so,  subject  to  the  following
# conditions:
#
# The above copyright  notice and this  permission notice shall  be
# included in all copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS  IS", WITHOUT WARRANTY OF ANY  KIND,
# EXPRESS OR IMPLIED, INCLUDING  BUT NOT LIMITED TO  THE WARRANTIES
# OF  MERCHANTABILITY,  FITNESS   FOR  A  PARTICULAR   PURPOSE  AND
# NONINFRINGEMENT.  IN  NO  EVENT SHALL  THE  AUTHORS  OR COPYRIGHT
# HOLDERS  BE LIABLE  FOR ANY  CLAIM, DAMAGES  OR OTHER  LIABILITY,
# WHETHER  IN AN  ACTION OF  CONTRACT, TORT  OR OTHERWISE,  ARISING
# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
# OTHER DEALINGS IN THE SOFTWARE.
#
from __future__ import print_function

"""Read from and write to tar format archives.
"""

__version__ = "$Revision$"

version     = "0.9.0"
__author__  = "Lars Gust\u00e4bel (lars@gustaebel.de)"
__date__    = "$Date: 2011-02-25 17:42:01 +0200 (Fri, 25 Feb 2011) $"
__cvsid__   = "$Id: tarfile.py 88586 2011-02-25 15:42:01Z marc-andre.lemburg $"
__credits__ = "Gustavo Niemeyer, Niels Gust\u00e4bel, Richard Townsend."

#---------
# Imports
#---------
import sys
import os
import stat
import errno
import time
import struct
import copy
import re

try:
    import grp, pwd
except ImportError:
    grp = pwd = None

# os.symlink on Windows prior to 6.0 raises NotImplementedError
symlink_exception = (AttributeError, NotImplementedError)
try:
    # WindowsError (1314) will be raised if the caller does not hold the
    # SeCreateSymbolicLinkPrivilege privilege
    symlink_exception += (WindowsError,)
except NameError:
    pass

# from tarfile import *
__all__ = ["TarFile", "TarInfo", "is_tarfile", "TarError"]

if sys.version_info[0] < 3:
    import __builtin__ as builtins
else:
    import builtins

_open = builtins.open   # Since 'open' is TarFile.open

#---------------------------------------------------------
# tar constants
#---------------------------------------------------------
NUL = b"\0"                     # the null character
BLOCKSIZE = 512                 # length of processing blocks
RECORDSIZE = BLOCKSIZE * 20     # length of records
GNU_MAGIC = b"ustar  \0"        # magic gnu tar string
POSIX_MAGIC = b"ustar\x0000"    # magic posix tar string

LENGTH_NAME = 100               # maximum length of a filename
LENGTH_LINK = 100               # maximum length of a linkname
LENGTH_PREFIX = 155             # maximum length of the prefix field

REGTYPE = b"0"                  # regular file
AREGTYPE = b"\0"                # regular file
LNKTYPE = b"1"                  # link (inside tarfile)
SYMTYPE = b"2"                  # symbolic link
CHRTYPE = b"3"                  # character special device
BLKTYPE = b"4"                  # block special device
DIRTYPE = b"5"                  # directory
FIFOTYPE = b"6"                 # fifo special device
CONTTYPE = b"7"                 # contiguous file

GNUTYPE_LONGNAME = b"L"         # GNU tar longname
GNUTYPE_LONGLINK = b"K"         # GNU tar longlink
GNUTYPE_SPARSE = b"S"           # GNU tar sparse file

XHDTYPE = b"x"                  # POSIX.1-2001 extended header
XGLTYPE = b"g"                  # POSIX.1-2001 global header
SOLARIS_XHDTYPE = b"X"          # Solaris extended header

USTAR_FORMAT = 0                # POSIX.1-1988 (ustar) format
GNU_FORMAT = 1                  # GNU tar format
PAX_FORMAT = 2                  # POSIX.1-2001 (pax) format
DEFAULT_FORMAT = GNU_FORMAT

#---------------------------------------------------------
# tarfile constants
#---------------------------------------------------------
# File types that tarfile supports:
SUPPORTED_TYPES = (REGTYPE, AREGTYPE, LNKTYPE,
                   SYMTYPE, DIRTYPE, FIFOTYPE,
                   CONTTYPE, CHRTYPE, BLKTYPE,
                   GNUTYPE_LONGNAME, GNUTYPE_LONGLINK,
                   GNUTYPE_SPARSE)

# File types that will be treated as a regular file.
REGULAR_TYPES = (REGTYPE, AREGTYPE,
                 CONTTYPE, GNUTYPE_SPARSE)

# File types that are part of the GNU tar format.
GNU_TYPES = (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK,
             GNUTYPE_SPARSE)

# Fields from a pax header that override a TarInfo attribute.
PAX_FIELDS = ("path", "linkpath", "size", "mtime",
              "uid", "gid", "uname", "gname")

# Fields from a pax header that are affected by hdrcharset.
PAX_NAME_FIELDS = set(("path", "linkpath", "uname", "gname"))

# Fields in a pax header that are numbers, all other fields
# are treated as strings.
PAX_NUMBER_FIELDS = {
    "atime": float,
    "ctime": float,
    "mtime": float,
    "uid": int,
    "gid": int,
    "size": int
}

#---------------------------------------------------------
# Bits used in the mode field, values in octal.
#---------------------------------------------------------
S_IFLNK = 0o120000        # symbolic link
S_IFREG = 0o100000        # regular file
S_IFBLK = 0o060000        # block device
S_IFDIR = 0o040000        # directory
S_IFCHR = 0o020000        # character device
S_IFIFO = 0o010000        # fifo

TSUID   = 0o4000          # set UID on execution
TSGID   = 0o2000          # set GID on execution
TSVTX   = 0o1000          # reserved

TUREAD  = 0o400           # read by owner
TUWRITE = 0o200           # write by owner
TUEXEC  = 0o100           # execute/search by owner
TGREAD  = 0o040           # read by group
TGWRITE = 0o020           # write by group
TGEXEC  = 0o010           # execute/search by group
TOREAD  = 0o004           # read by other
TOWRITE = 0o002           # write by other
TOEXEC  = 0o001           # execute/search by other

#---------------------------------------------------------
# initialization
#---------------------------------------------------------
if os.name in ("nt", "ce"):
    ENCODING = "utf-8"
else:
    ENCODING = sys.getfilesystemencoding()

#---------------------------------------------------------
# Some useful functions
#---------------------------------------------------------

def stn(s, length, encoding, errors):
    """Convert a string to a null-terminated bytes object.
    """
    s = s.encode(encoding, errors)
    return s[:length] + (length - len(s)) * NUL

def nts(s, encoding, errors):
    """Convert a null-terminated bytes object to a string.
    """
    p = s.find(b"\0")
    if p != -1:
        s = s[:p]
    return s.decode(encoding, errors)

def nti(s):
    """Convert a number field to a python number.
    """
    # There are two possible encodings for a number field, see
    # itn() below.
    if s[0] != chr(0o200):
        try:
            n = int(nts(s, "ascii", "strict") or "0", 8)
        except ValueError:
            raise InvalidHeaderError("invalid header")
    else:
        n = 0
        for i in range(len(s) - 1):
            n <<= 8
            n += ord(s[i + 1])
    return n

def itn(n, digits=8, format=DEFAULT_FORMAT):
    """Convert a python number to a number field.
    """
    # POSIX 1003.1-1988 requires numbers to be encoded as a string of
    # octal digits followed by a null-byte, this allows values up to
    # (8**(digits-1))-1. GNU tar allows storing numbers greater than
    # that if necessary. A leading 0o200 byte indicates this particular
    # encoding, the following digits-1 bytes are a big-endian
    # representation. This allows values up to (256**(digits-1))-1.
    if 0 <= n < 8 ** (digits - 1):
        s = ("%0*o" % (digits - 1, n)).encode("ascii") + NUL
    else:
        if format != GNU_FORMAT or n >= 256 ** (digits - 1):
            raise ValueError("overflow in number field")

        if n < 0:
            # XXX We mimic GNU tar's behaviour with negative numbers,
            # this could raise OverflowError.
            n = struct.unpack("L", struct.pack("l", n))[0]

        s = bytearray()
        for i in range(digits - 1):
            s.insert(0, n & 0o377)
            n >>= 8
        s.insert(0, 0o200)
    return s

def calc_chksums(buf):
    """Calculate the checksum for a member's header by summing up all
       characters except for the chksum field which is treated as if
       it was filled with spaces. According to the GNU tar sources,
       some tars (Sun and NeXT) calculate chksum with signed char,
       which will be different if there are chars in the buffer with
       the high bit set. So we calculate two checksums, unsigned and
       signed.
    """
    unsigned_chksum = 256 + sum(struct.unpack("148B", buf[:148]) + struct.unpack("356B", buf[156:512]))
    signed_chksum = 256 + sum(struct.unpack("148b", buf[:148]) + struct.unpack("356b", buf[156:512]))
    return unsigned_chksum, signed_chksum

def copyfileobj(src, dst, length=None):
    """Copy length bytes from fileobj src to fileobj dst.
       If length is None, copy the entire content.
    """
    if length == 0:
        return
    if length is None:
        while True:
            buf = src.read(16*1024)
            if not buf:
                break
            dst.write(buf)
        return

    BUFSIZE = 16 * 1024
    blocks, remainder = divmod(length, BUFSIZE)
    for b in range(blocks):
        buf = src.read(BUFSIZE)
        if len(buf) < BUFSIZE:
            raise IOError("end of file reached")
        dst.write(buf)

    if remainder != 0:
        buf = src.read(remainder)
        if len(buf) < remainder:
            raise IOError("end of file reached")
        dst.write(buf)
    return

filemode_table = (
    ((S_IFLNK,      "l"),
     (S_IFREG,      "-"),
     (S_IFBLK,      "b"),
     (S_IFDIR,      "d"),
     (S_IFCHR,      "c"),
     (S_IFIFO,      "p")),

    ((TUREAD,       "r"),),
    ((TUWRITE,      "w"),),
    ((TUEXEC|TSUID, "s"),
     (TSUID,        "S"),
     (TUEXEC,       "x")),

    ((TGREAD,       "r"),),
    ((TGWRITE,      "w"),),
    ((TGEXEC|TSGID, "s"),
     (TSGID,        "S"),
     (TGEXEC,       "x")),

    ((TOREAD,       "r"),),
    ((TOWRITE,      "w"),),
    ((TOEXEC|TSVTX, "t"),
     (TSVTX,        "T"),
     (TOEXEC,       "x"))
)

def filemode(mode):
    """Convert a file's mode to a string of the form
       -rwxrwxrwx.
       Used by TarFile.list()
    """
    perm = []
    for table in filemode_table:
        for bit, char in table:
            if mode & bit == bit:
                perm.append(char)
                break
        else:
            perm.append("-")
    return "".join(perm)

class TarError(Exception):
    """Base exception."""
    pass
class ExtractError(TarError):
    """General exception for extract errors."""
    pass
class ReadError(TarError):
    """Exception for unreadable tar archives."""
    pass
class CompressionError(TarError):
    """Exception for unavailable compression methods."""
    pass
class StreamError(TarError):
    """Exception for unsupported operations on stream-like TarFiles."""
    pass
class HeaderError(TarError):
    """Base exception for header errors."""
    pass
class EmptyHeaderError(HeaderError):
    """Exception for empty headers."""
    pass
class TruncatedHeaderError(HeaderError):
    """Exception for truncated headers."""
    pass
class EOFHeaderError(HeaderError):
    """Exception for end of file headers."""
    pass
class InvalidHeaderError(HeaderError):
    """Exception for invalid headers."""
    pass
class SubsequentHeaderError(HeaderError):
    """Exception for missing and invalid extended headers."""
    pass

#---------------------------
# internal stream interface
#---------------------------
class _LowLevelFile(object):
    """Low-level file object. Supports reading and writing.
       It is used instead of a regular file object for streaming
       access.
    """

    def __init__(self, name, mode):
        mode = {
            "r": os.O_RDONLY,
            "w": os.O_WRONLY | os.O_CREAT | os.O_TRUNC,
        }[mode]
        if hasattr(os, "O_BINARY"):
            mode |= os.O_BINARY
        self.fd = os.open(name, mode, 0o666)

    def close(self):
        os.close(self.fd)

    def read(self, size):
        return os.read(self.fd, size)

    def write(self, s):
        os.write(self.fd, s)

class _Stream(object):
    """Class that serves as an adapter between TarFile and
       a stream-like object.  The stream-like object only
       needs to have a read() or write() method and is accessed
       blockwise.  Use of gzip or bzip2 compression is possible.
       A stream-like object could be for example: sys.stdin,
       sys.stdout, a socket, a tape device etc.

       _Stream is intended to be used only internally.
    """

    def __init__(self, name, mode, comptype, fileobj, bufsize):
        """Construct a _Stream object.
        """
        self._extfileobj = True
        if fileobj is None:
            fileobj = _LowLevelFile(name, mode)
            self._extfileobj = False

        if comptype == '*':
            # Enable transparent compression detection for the
            # stream interface
            fileobj = _StreamProxy(fileobj)
            comptype = fileobj.getcomptype()

        self.name     = name or ""
        self.mode     = mode
        self.comptype = comptype
        self.fileobj  = fileobj
        self.bufsize  = bufsize
        self.buf      = b""
        self.pos      = 0
        self.closed   = False

        try:
            if comptype == "gz":
                try:
                    import zlib
                except ImportError:
                    raise CompressionError("zlib module is not available")
                self.zlib = zlib
                self.crc = zlib.crc32(b"")
                if mode == "r":
                    self._init_read_gz()
                else:
                    self._init_write_gz()

            if comptype == "bz2":
                try:
                    import bz2
                except ImportError:
                    raise CompressionError("bz2 module is not available")
                if mode == "r":
                    self.dbuf = b""
                    self.cmp = bz2.BZ2Decompressor()
                else:
                    self.cmp = bz2.BZ2Compressor()
        except:
            if not self._extfileobj:
                self.fileobj.close()
            self.closed = True
            raise

    def __del__(self):
        if hasattr(self, "closed") and not self.closed:
            self.close()

    def _init_write_gz(self):
        """Initialize for writing with gzip compression.
        """
        self.cmp = self.zlib.compressobj(9, self.zlib.DEFLATED,
                                            -self.zlib.MAX_WBITS,
                                            self.zlib.DEF_MEM_LEVEL,
                                            0)
        timestamp = struct.pack("<L", int(time.time()))
        self.__write(b"\037\213\010\010" + timestamp + b"\002\377")
        if self.name.endswith(".gz"):
            self.name = self.name[:-3]
        # RFC1952 says we must use ISO-8859-1 for the FNAME field.
        self.__write(self.name.encode("iso-8859-1", "replace") + NUL)

    def write(self, s):
        """Write string s to the stream.
        """
        if self.comptype == "gz":
            self.crc = self.zlib.crc32(s, self.crc)
        self.pos += len(s)
        if self.comptype != "tar":
            s = self.cmp.compress(s)
        self.__write(s)

    def __write(self, s):
        """Write string s to the stream if a whole new block
           is ready to be written.
        """
        self.buf += s
        while len(self.buf) > self.bufsize:
            self.fileobj.write(self.buf[:self.bufsize])
            self.buf = self.buf[self.bufsize:]

    def close(self):
        """Close the _Stream object. No operation should be
           done on it afterwards.
        """
        if self.closed:
            return

        if self.mode == "w" and self.comptype != "tar":
            self.buf += self.cmp.flush()

        if self.mode == "w" and self.buf:
            self.fileobj.write(self.buf)
            self.buf = b""
            if self.comptype == "gz":
                # The native zlib crc is an unsigned 32-bit integer, but
                # the Python wrapper implicitly casts that to a signed C
                # long.  So, on a 32-bit box self.crc may "look negative",
                # while the same crc on a 64-bit box may "look positive".
                # To avoid irksome warnings from the `struct` module, force
                # it to look positive on all boxes.
                self.fileobj.write(struct.pack("<L", self.crc & 0xffffffff))
                self.fileobj.write(struct.pack("<L", self.pos & 0xffffFFFF))

        if not self._extfileobj:
            self.fileobj.close()

        self.closed = True

    def _init_read_gz(self):
        """Initialize for reading a gzip compressed fileobj.
        """
        self.cmp = self.zlib.decompressobj(-self.zlib.MAX_WBITS)
        self.dbuf = b""

        # taken from gzip.GzipFile with some alterations
        if self.__read(2) != b"\037\213":
            raise ReadError("not a gzip file")
        if self.__read(1) != b"\010":
            raise CompressionError("unsupported compression method")

        flag = ord(self.__read(1))
        self.__read(6)

        if flag & 4:
            xlen = ord(self.__read(1)) + 256 * ord(self.__read(1))
            self.read(xlen)
        if flag & 8:
            while True:
                s = self.__read(1)
                if not s or s == NUL:
                    break
        if flag & 16:
            while True:
                s = self.__read(1)
                if not s or s == NUL:
                    break
        if flag & 2:
            self.__read(2)

    def tell(self):
        """Return the stream's file pointer position.
        """
        return self.pos

    def seek(self, pos=0):
        """Set the stream's file pointer to pos. Negative seeking
           is forbidden.
        """
        if pos - self.pos >= 0:
            blocks, remainder = divmod(pos - self.pos, self.bufsize)
            for i in range(blocks):
                self.read(self.bufsize)
            self.read(remainder)
        else:
            raise StreamError("seeking backwards is not allowed")
        return self.pos

    def read(self, size=None):
        """Return the next size number of bytes from the stream.
           If size is not defined, return all bytes of the stream
           up to EOF.
        """
        if size is None:
            t = []
            while True:
                buf = self._read(self.bufsize)
                if not buf:
                    break
                t.append(buf)
            buf = "".join(t)
        else:
            buf = self._read(size)
        self.pos += len(buf)
        return buf

    def _read(self, size):
        """Return size bytes from the stream.
        """
        if self.comptype == "tar":
            return self.__read(size)

        c = len(self.dbuf)
        while c < size:
            buf = self.__read(self.bufsize)
            if not buf:
                break
            try:
                buf = self.cmp.decompress(buf)
            except IOError:
                raise ReadError("invalid compressed data")
            self.dbuf += buf
            c += len(buf)
        buf = self.dbuf[:size]
        self.dbuf = self.dbuf[size:]
        return buf

    def __read(self, size):
        """Return size bytes from stream. If internal buffer is empty,
           read another block from the stream.
        """
        c = len(self.buf)
        while c < size:
            buf = self.fileobj.read(self.bufsize)
            if not buf:
                break
            self.buf += buf
            c += len(buf)
        buf = self.buf[:size]
        self.buf = self.buf[size:]
        return buf
# class _Stream

class _StreamProxy(object):
    """Small proxy class that enables transparent compression
       detection for the Stream interface (mode 'r|*').
    """

    def __init__(self, fileobj):
        self.fileobj = fileobj
        self.buf = self.fileobj.read(BLOCKSIZE)

    def read(self, size):
        self.read = self.fileobj.read
        return self.buf

    def getcomptype(self):
        if self.buf.startswith(b"\037\213\010"):
            return "gz"
        if self.buf.startswith(b"BZh91"):
            return "bz2"
        return "tar"

    def close(self):
        self.fileobj.close()
# class StreamProxy

class _BZ2Proxy(object):
    """Small proxy class that enables external file object
       support for "r:bz2" and "w:bz2" modes. This is actually
       a workaround for a limitation in bz2 module's BZ2File
       class which (unlike gzip.GzipFile) has no support for
       a file object argument.
    """

    blocksize = 16 * 1024

    def __init__(self, fileobj, mode):
        self.fileobj = fileobj
        self.mode = mode
        self.name = getattr(self.fileobj, "name", None)
        self.init()

    def init(self):
        import bz2
        self.pos = 0
        if self.mode == "r":
            self.bz2obj = bz2.BZ2Decompressor()
            self.fileobj.seek(0)
            self.buf = b""
        else:
            self.bz2obj = bz2.BZ2Compressor()

    def read(self, size):
        x = len(self.buf)
        while x < size:
            raw = self.fileobj.read(self.blocksize)
            if not raw:
                break
            data = self.bz2obj.decompress(raw)
            self.buf += data
            x += len(data)

        buf = self.buf[:size]
        self.buf = self.buf[size:]
        self.pos += len(buf)
        return buf

    def seek(self, pos):
        if pos < self.pos:
            self.init()
        self.read(pos - self.pos)

    def tell(self):
        return self.pos

    def write(self, data):
        self.pos += len(data)
        raw = self.bz2obj.compress(data)
        self.fileobj.write(raw)

    def close(self):
        if self.mode == "w":
            raw = self.bz2obj.flush()
            self.fileobj.write(raw)
# class _BZ2Proxy

#------------------------
# Extraction file object
#------------------------
class _FileInFile(object):
    """A thin wrapper around an existing file object that
       provides a part of its data as an individual file
       object.
    """

    def __init__(self, fileobj, offset, size, blockinfo=None):
        self.fileobj = fileobj
        self.offset = offset
        self.size = size
        self.position = 0

        if blockinfo is None:
            blockinfo = [(0, size)]

        # Construct a map with data and zero blocks.
        self.map_index = 0
        self.map = []
        lastpos = 0
        realpos = self.offset
        for offset, size in blockinfo:
            if offset > lastpos:
                self.map.append((False, lastpos, offset, None))
            self.map.append((True, offset, offset + size, realpos))
            realpos += size
            lastpos = offset + size
        if lastpos < self.size:
            self.map.append((False, lastpos, self.size, None))

    def seekable(self):
        if not hasattr(self.fileobj, "seekable"):
            # XXX gzip.GzipFile and bz2.BZ2File
            return True
        return self.fileobj.seekable()

    def tell(self):
        """Return the current file position.
        """
        return self.position

    def seek(self, position):
        """Seek to a position in the file.
        """
        self.position = position

    def read(self, size=None):
        """Read data from the file.
        """
        if size is None:
            size = self.size - self.position
        else:
            size = min(size, self.size - self.position)

        buf = b""
        while size > 0:
            while True:
                data, start, stop, offset = self.map[self.map_index]
                if start <= self.position < stop:
                    break
                else:
                    self.map_index += 1
                    if self.map_index == len(self.map):
                        self.map_index = 0
            length = min(size, stop - self.position)
            if data:
                self.fileobj.seek(offset + (self.position - start))
                buf += self.fileobj.read(length)
            else:
                buf += NUL * length
            size -= length
            self.position += length
        return buf
#class _FileInFile


class ExFileObject(object):
    """File-like object for reading an archive member.
       Is returned by TarFile.extractfile().
    """
    blocksize = 1024

    def __init__(self, tarfile, tarinfo):
        self.fileobj = _FileInFile(tarfile.fileobj,
                                   tarinfo.offset_data,
                                   tarinfo.size,
                                   tarinfo.sparse)
        self.name = tarinfo.name
        self.mode = "r"
        self.closed = False
        self.size = tarinfo.size

        self.position = 0
        self.buffer = b""

    def readable(self):
        return True

    def writable(self):
        return False

    def seekable(self):
        return self.fileobj.seekable()

    def read(self, size=None):
        """Read at most size bytes from the file. If size is not
           present or None, read all data until EOF is reached.
        """
        if self.closed:
            raise ValueError("I/O operation on closed file")

        buf = b""
        if self.buffer:
            if size is None:
                buf = self.buffer
                self.buffer = b""
            else:
                buf = self.buffer[:size]
                self.buffer = self.buffer[size:]

        if size is None:
            buf += self.fileobj.read()
        else:
            buf += self.fileobj.read(size - len(buf))

        self.position += len(buf)
        return buf

    # XXX TextIOWrapper uses the read1() method.
    read1 = read

    def readline(self, size=-1):
        """Read one entire line from the file. If size is present
           and non-negative, return a string with at most that
           size, which may be an incomplete line.
        """
        if self.closed:
            raise ValueError("I/O operation on closed file")

        pos = self.buffer.find(b"\n") + 1
        if pos == 0:
            # no newline found.
            while True:
                buf = self.fileobj.read(self.blocksize)
                self.buffer += buf
                if not buf or b"\n" in buf:
                    pos = self.buffer.find(b"\n") + 1
                    if pos == 0:
                        # no newline found.
                        pos = len(self.buffer)
                    break

        if size != -1:
            pos = min(size, pos)

        buf = self.buffer[:pos]
        self.buffer = self.buffer[pos:]
        self.position += len(buf)
        return buf

    def readlines(self):
        """Return a list with all remaining lines.
        """
        result = []
        while True:
            line = self.readline()
            if not line: break
            result.append(line)
        return result

    def tell(self):
        """Return the current file position.
        """
        if self.closed:
            raise ValueError("I/O operation on closed file")

        return self.position

    def seek(self, pos, whence=os.SEEK_SET):
        """Seek to a position in the file.
        """
        if self.closed:
            raise ValueError("I/O operation on closed file")

        if whence == os.SEEK_SET:
            self.position = min(max(pos, 0), self.size)
        elif whence == os.SEEK_CUR:
            if pos < 0:
                self.position = max(self.position + pos, 0)
            else:
                self.position = min(self.position + pos, self.size)
        elif whence == os.SEEK_END:
            self.position = max(min(self.size + pos, self.size), 0)
        else:
            raise ValueError("Invalid argument")

        self.buffer = b""
        self.fileobj.seek(self.position)

    def close(self):
        """Close the file object.
        """
        self.closed = True

    def __iter__(self):
        """Get an iterator over the file's lines.
        """
        while True:
            line = self.readline()
            if not line:
                break
            yield line
#class ExFileObject

#------------------
# Exported Classes
#------------------
class TarInfo(object):
    """Informational class which holds the details about an
       archive member given by a tar header block.
       TarInfo objects are returned by TarFile.getmember(),
       TarFile.getmembers() and TarFile.gettarinfo() and are
       usually created internally.
    """

    __slots__ = ("name", "mode", "uid", "gid", "size", "mtime",
                 "chksum", "type", "linkname", "uname", "gname",
                 "devmajor", "devminor",
                 "offset", "offset_data", "pax_headers", "sparse",
                 "tarfile", "_sparse_structs", "_link_target")

    def __init__(self, name=""):
        """Construct a TarInfo object. name is the optional name
           of the member.
        """
        self.name = name        # member name
        self.mode = 0o644       # file permissions
        self.uid = 0            # user id
        self.gid = 0            # group id
        self.size = 0           # file size
        self.mtime = 0          # modification time
        self.chksum = 0         # header checksum
        self.type = REGTYPE     # member type
        self.linkname = ""      # link name
        self.uname = ""         # user name
        self.gname = ""         # group name
        self.devmajor = 0       # device major number
        self.devminor = 0       # device minor number

        self.offset = 0         # the tar header starts here
        self.offset_data = 0    # the file's data starts here

        self.sparse = None      # sparse member information
        self.pax_headers = {}   # pax header information

    # In pax headers the "name" and "linkname" field are called
    # "path" and "linkpath".
    def _getpath(self):
        return self.name
    def _setpath(self, name):
        self.name = name
    path = property(_getpath, _setpath)

    def _getlinkpath(self):
        return self.linkname
    def _setlinkpath(self, linkname):
        self.linkname = linkname
    linkpath = property(_getlinkpath, _setlinkpath)

    def __repr__(self):
        return "<%s %r at %#x>" % (self.__class__.__name__,self.name,id(self))

    def get_info(self):
        """Return the TarInfo's attributes as a dictionary.
        """
        info = {
            "name":     self.name,
            "mode":     self.mode & 0o7777,
            "uid":      self.uid,
            "gid":      self.gid,
            "size":     self.size,
            "mtime":    self.mtime,
            "chksum":   self.chksum,
            "type":     self.type,
            "linkname": self.linkname,
            "uname":    self.uname,
            "gname":    self.gname,
            "devmajor": self.devmajor,
            "devminor": self.devminor
        }

        if info["type"] == DIRTYPE and not info["name"].endswith("/"):
            info["name"] += "/"

        return info

    def tobuf(self, format=DEFAULT_FORMAT, encoding=ENCODING, errors="surrogateescape"):
        """Return a tar header as a string of 512 byte blocks.
        """
        info = self.get_info()

        if format == USTAR_FORMAT:
            return self.create_ustar_header(info, encoding, errors)
        elif format == GNU_FORMAT:
            return self.create_gnu_header(info, encoding, errors)
        elif format == PAX_FORMAT:
            return self.create_pax_header(info, encoding)
        else:
            raise ValueError("invalid format")

    def create_ustar_header(self, info, encoding, errors):
        """Return the object as a ustar header block.
        """
        info["magic"] = POSIX_MAGIC

        if len(info["linkname"]) > LENGTH_LINK:
            raise ValueError("linkname is too long")

        if len(info["name"]) > LENGTH_NAME:
            info["prefix"], info["name"] = self._posix_split_name(info["name"])

        return self._create_header(info, USTAR_FORMAT, encoding, errors)

    def create_gnu_header(self, info, encoding, errors):
        """Return the object as a GNU header block sequence.
        """
        info["magic"] = GNU_MAGIC

        buf = b""
        if len(info["linkname"]) > LENGTH_LINK:
            buf += self._create_gnu_long_header(info["linkname"], GNUTYPE_LONGLINK, encoding, errors)

        if len(info["name"]) > LENGTH_NAME:
            buf += self._create_gnu_long_header(info["name"], GNUTYPE_LONGNAME, encoding, errors)

        return buf + self._create_header(info, GNU_FORMAT, encoding, errors)

    def create_pax_header(self, info, encoding):
        """Return the object as a ustar header block. If it cannot be
           represented this way, prepend a pax extended header sequence
           with supplement information.
        """
        info["magic"] = POSIX_MAGIC
        pax_headers = self.pax_headers.copy()

        # Test string fields for values that exceed the field length or cannot
        # be represented in ASCII encoding.
        for name, hname, length in (
                ("name", "path", LENGTH_NAME), ("linkname", "linkpath", LENGTH_LINK),
                ("uname", "uname", 32), ("gname", "gname", 32)):

            if hname in pax_headers:
                # The pax header has priority.
                continue

            # Try to encode the string as ASCII.
            try:
                info[name].encode("ascii", "strict")
            except UnicodeEncodeError:
                pax_headers[hname] = info[name]
                continue

            if len(info[name]) > length:
                pax_headers[hname] = info[name]

        # Test number fields for values that exceed the field limit or values
        # that like to be stored as float.
        for name, digits in (("uid", 8), ("gid", 8), ("size", 12), ("mtime", 12)):
            if name in pax_headers:
                # The pax header has priority. Avoid overflow.
                info[name] = 0
                continue

            val = info[name]
            if not 0 <= val < 8 ** (digits - 1) or isinstance(val, float):
                pax_headers[name] = str(val)
                info[name] = 0

        # Create a pax extended header if necessary.
        if pax_headers:
            buf = self._create_pax_generic_header(pax_headers, XHDTYPE, encoding)
        else:
            buf = b""

        return buf + self._create_header(info, USTAR_FORMAT, "ascii", "replace")

    @classmethod
    def create_pax_global_header(cls, pax_headers):
        """Return the object as a pax global header block sequence.
        """
        return cls._create_pax_generic_header(pax_headers, XGLTYPE, "utf8")

    def _posix_split_name(self, name):
        """Split a name longer than 100 chars into a prefix
           and a name part.
        """
        prefix = name[:LENGTH_PREFIX + 1]
        while prefix and prefix[-1] != "/":
            prefix = prefix[:-1]

        name = name[len(prefix):]
        prefix = prefix[:-1]

        if not prefix or len(name) > LENGTH_NAME:
            raise ValueError("name is too long")
        return prefix, name

    @staticmethod
    def _create_header(info, format, encoding, errors):
        """Return a header block. info is a dictionary with file
           information, format must be one of the *_FORMAT constants.
        """
        parts = [
            stn(info.get("name", ""), 100, encoding, errors),
            itn(info.get("mode", 0) & 0o7777, 8, format),
            itn(info.get("uid", 0), 8, format),
            itn(info.get("gid", 0), 8, format),
            itn(info.get("size", 0), 12, format),
            itn(info.get("mtime", 0), 12, format),
            b"        ", # checksum field
            info.get("type", REGTYPE),
            stn(info.get("linkname", ""), 100, encoding, errors),
            info.get("magic", POSIX_MAGIC),
            stn(info.get("uname", ""), 32, encoding, errors),
            stn(info.get("gname", ""), 32, encoding, errors),
            itn(info.get("devmajor", 0), 8, format),
            itn(info.get("devminor", 0), 8, format),
            stn(info.get("prefix", ""), 155, encoding, errors)
        ]

        buf = struct.pack("%ds" % BLOCKSIZE, b"".join(parts))
        chksum = calc_chksums(buf[-BLOCKSIZE:])[0]
        buf = buf[:-364] + ("%06o\0" % chksum).encode("ascii") + buf[-357:]
        return buf

    @staticmethod
    def _create_payload(payload):
        """Return the string payload filled with zero bytes
           up to the next 512 byte border.
        """
        blocks, remainder = divmod(len(payload), BLOCKSIZE)
        if remainder > 0:
            payload += (BLOCKSIZE - remainder) * NUL
        return payload

    @classmethod
    def _create_gnu_long_header(cls, name, type, encoding, errors):
        """Return a GNUTYPE_LONGNAME or GNUTYPE_LONGLINK sequence
           for name.
        """
        name = name.encode(encoding, errors) + NUL

        info = {}
        info["name"] = "././@LongLink"
        info["type"] = type
        info["size"] = len(name)
        info["magic"] = GNU_MAGIC

        # create extended header + name blocks.
        return cls._create_header(info, USTAR_FORMAT, encoding, errors) + \
                cls._create_payload(name)

    @classmethod
    def _create_pax_generic_header(cls, pax_headers, type, encoding):
        """Return a POSIX.1-2008 extended or global header sequence
           that contains a list of keyword, value pairs. The values
           must be strings.
        """
        # Check if one of the fields contains surrogate characters and thereby
        # forces hdrcharset=BINARY, see _proc_pax() for more information.
        binary = False
        for keyword, value in pax_headers.items():
            try:
                value.encode("utf8", "strict")
            except UnicodeEncodeError:
                binary = True
                break

        records = b""
        if binary:
            # Put the hdrcharset field at the beginning of the header.
            records += b"21 hdrcharset=BINARY\n"

        for keyword, value in pax_headers.items():
            keyword = keyword.encode("utf8")
            if binary:
                # Try to restore the original byte representation of `value'.
                # Needless to say, that the encoding must match the string.
                value = value.encode(encoding, "surrogateescape")
            else:
                value = value.encode("utf8")

            l = len(keyword) + len(value) + 3   # ' ' + '=' + '\n'
            n = p = 0
            while True:
                n = l + len(str(p))
                if n == p:
                    break
                p = n
            records += bytes(str(p), "ascii") + b" " + keyword + b"=" + value + b"\n"

        # We use a hardcoded "././@PaxHeader" name like star does
        # instead of the one that POSIX recommends.
        info = {}
        info["name"] = "././@PaxHeader"
        info["type"] = type
        info["size"] = len(records)
        info["magic"] = POSIX_MAGIC

        # Create pax header + record blocks.
        return cls._create_header(info, USTAR_FORMAT, "ascii", "replace") + \
                cls._create_payload(records)

    @classmethod
    def frombuf(cls, buf, encoding, errors):
        """Construct a TarInfo object from a 512 byte bytes object.
        """
        if len(buf) == 0:
            raise EmptyHeaderError("empty header")
        if len(buf) != BLOCKSIZE:
            raise TruncatedHeaderError("truncated header")
        if buf.count(NUL) == BLOCKSIZE:
            raise EOFHeaderError("end of file header")

        chksum = nti(buf[148:156])
        if chksum not in calc_chksums(buf):
            raise InvalidHeaderError("bad checksum")

        obj = cls()
        obj.name = nts(buf[0:100], encoding, errors)
        obj.mode = nti(buf[100:108])
        obj.uid = nti(buf[108:116])
        obj.gid = nti(buf[116:124])
        obj.size = nti(buf[124:136])
        obj.mtime = nti(buf[136:148])
        obj.chksum = chksum
        obj.type = buf[156:157]
        obj.linkname = nts(buf[157:257], encoding, errors)
        obj.uname = nts(buf[265:297], encoding, errors)
        obj.gname = nts(buf[297:329], encoding, errors)
        obj.devmajor = nti(buf[329:337])
        obj.devminor = nti(buf[337:345])
        prefix = nts(buf[345:500], encoding, errors)

        # Old V7 tar format represents a directory as a regular
        # file with a trailing slash.
        if obj.type == AREGTYPE and obj.name.endswith("/"):
            obj.type = DIRTYPE

        # The old GNU sparse format occupies some of the unused
        # space in the buffer for up to 4 sparse structures.
        # Save the them for later processing in _proc_sparse().
        if obj.type == GNUTYPE_SPARSE:
            pos = 386
            structs = []
            for i in range(4):
                try:
                    offset = nti(buf[pos:pos + 12])
                    numbytes = nti(buf[pos + 12:pos + 24])
                except ValueError:
                    break
                structs.append((offset, numbytes))
                pos += 24
            isextended = bool(buf[482])
            origsize = nti(buf[483:495])
            obj._sparse_structs = (structs, isextended, origsize)

        # Remove redundant slashes from directories.
        if obj.isdir():
            obj.name = obj.name.rstrip("/")

        # Reconstruct a ustar longname.
        if prefix and obj.type not in GNU_TYPES:
            obj.name = prefix + "/" + obj.name
        return obj

    @classmethod
    def fromtarfile(cls, tarfile):
        """Return the next TarInfo object from TarFile object
           tarfile.
        """
        buf = tarfile.fileobj.read(BLOCKSIZE)
        obj = cls.frombuf(buf, tarfile.encoding, tarfile.errors)
        obj.offset = tarfile.fileobj.tell() - BLOCKSIZE
        return obj._proc_member(tarfile)

    #--------------------------------------------------------------------------
    # The following are methods that are called depending on the type of a
    # member. The entry point is _proc_member() which can be overridden in a
    # subclass to add custom _proc_*() methods. A _proc_*() method MUST
    # implement the following
    # operations:
    # 1. Set self.offset_data to the position where the data blocks begin,
    #    if there is data that follows.
    # 2. Set tarfile.offset to the position where the next member's header will
    #    begin.
    # 3. Return self or another valid TarInfo object.
    def _proc_member(self, tarfile):
        """Choose the right processing method depending on
           the type and call it.
        """
        if self.type in (GNUTYPE_LONGNAME, GNUTYPE_LONGLINK):
            return self._proc_gnulong(tarfile)
        elif self.type == GNUTYPE_SPARSE:
            return self._proc_sparse(tarfile)
        elif self.type in (XHDTYPE, XGLTYPE, SOLARIS_XHDTYPE):
            return self._proc_pax(tarfile)
        else:
            return self._proc_builtin(tarfile)

    def _proc_builtin(self, tarfile):
        """Process a builtin type or an unknown type which
           will be treated as a regular file.
        """
        self.offset_data = tarfile.fileobj.tell()
        offset = self.offset_data
        if self.isreg() or self.type not in SUPPORTED_TYPES:
            # Skip the following data blocks.
            offset += self._block(self.size)
        tarfile.offset = offset

        # Patch the TarInfo object with saved global
        # header information.
        self._apply_pax_info(tarfile.pax_headers, tarfile.encoding, tarfile.errors)

        return self

    def _proc_gnulong(self, tarfile):
        """Process the blocks that hold a GNU longname
           or longlink member.
        """
        buf = tarfile.fileobj.read(self._block(self.size))

        # Fetch the next header and process it.
        try:
            next = self.fromtarfile(tarfile)
        except HeaderError:
            raise SubsequentHeaderError("missing or bad subsequent header")

        # Patch the TarInfo object from the next header with
        # the longname information.
        next.offset = self.offset
        if self.type == GNUTYPE_LONGNAME:
            next.name = nts(buf, tarfile.encoding, tarfile.errors)
        elif self.type == GNUTYPE_LONGLINK:
            next.linkname = nts(buf, tarfile.encoding, tarfile.errors)

        return next

    def _proc_sparse(self, tarfile):
        """Process a GNU sparse header plus extra headers.
        """
        # We already collected some sparse structures in frombuf().
        structs, isextended, origsize = self._sparse_structs
        del self._sparse_structs

        # Collect sparse structures from extended header blocks.
        while isextended:
            buf = tarfile.fileobj.read(BLOCKSIZE)
            pos = 0
            for i in range(21):
                try:
                    offset = nti(buf[pos:pos + 12])
                    numbytes = nti(buf[pos + 12:pos + 24])
                except ValueError:
                    break
                if offset and numbytes:
                    structs.append((offset, numbytes))
                pos += 24
            isextended = bool(buf[504])
        self.sparse = structs

        self.offset_data = tarfile.fileobj.tell()
        tarfile.offset = self.offset_data + self._block(self.size)
        self.size = origsize
        return self

    def _proc_pax(self, tarfile):
        """Process an extended or global header as described in
           POSIX.1-2008.
        """
        # Read the header information.
        buf = tarfile.fileobj.read(self._block(self.size))

        # A pax header stores supplemental information for either
        # the following file (extended) or all following files
        # (global).
        if self.type == XGLTYPE:
            pax_headers = tarfile.pax_headers
        else:
            pax_headers = tarfile.pax_headers.copy()

        # Check if the pax header contains a hdrcharset field. This tells us
        # the encoding of the path, linkpath, uname and gname fields. Normally,
        # these fields are UTF-8 encoded but since POSIX.1-2008 tar
        # implementations are allowed to store them as raw binary strings if
        # the translation to UTF-8 fails.
        match = re.search(br"\d+ hdrcharset=([^\n]+)\n", buf)
        if match is not None:
            pax_headers["hdrcharset"] = match.group(1).decode("utf8")

        # For the time being, we don't care about anything other than "BINARY".
        # The only other value that is currently allowed by the standard is
        # "ISO-IR 10646 2000 UTF-8" in other words UTF-8.
        hdrcharset = pax_headers.get("hdrcharset")
        if hdrcharset == "BINARY":
            encoding = tarfile.encoding
        else:
            encoding = "utf8"

        # Parse pax header information. A record looks like that:
        # "%d %s=%s\n" % (length, keyword, value). length is the size
        # of the complete record including the length field itself and
        # the newline. keyword and value are both UTF-8 encoded strings.
        regex = re.compile(br"(\d+) ([^=]+)=")
        pos = 0
        while True:
            match = regex.match(buf, pos)
            if not match:
                break

            length, keyword = match.groups()
            length = int(length)
            value = buf[match.end(2) + 1:match.start(1) + length - 1]

            # Normally, we could just use "utf8" as the encoding and "strict"
            # as the error handler, but we better not take the risk. For
            # example, GNU tar <= 1.23 is known to store filenames it cannot
            # translate to UTF-8 as raw strings (unfortunately without a
            # hdrcharset=BINARY header).
            # We first try the strict standard encoding, and if that fails we
            # fall back on the user's encoding and error handler.
            keyword = self._decode_pax_field(keyword, "utf8", "utf8",
                    tarfile.errors)
            if keyword in PAX_NAME_FIELDS:
                value = self._decode_pax_field(value, encoding, tarfile.encoding,
                        tarfile.errors)
            else:
                value = self._decode_pax_field(value, "utf8", "utf8",
                        tarfile.errors)

            pax_headers[keyword] = value
            pos += length

        # Fetch the next header.
        try:
            next = self.fromtarfile(tarfile)
        except HeaderError:
            raise SubsequentHeaderError("missing or bad subsequent header")

        # Process GNU sparse information.
        if "GNU.sparse.map" in pax_headers:
            # GNU extended sparse format version 0.1.
            self._proc_gnusparse_01(next, pax_headers)

        elif "GNU.sparse.size" in pax_headers:
            # GNU extended sparse format version 0.0.
            self._proc_gnusparse_00(next, pax_headers, buf)

        elif pax_headers.get("GNU.sparse.major") == "1" and pax_headers.get("GNU.sparse.minor") == "0":
            # GNU extended sparse format version 1.0.
            self._proc_gnusparse_10(next, pax_headers, tarfile)

        if self.type in (XHDTYPE, SOLARIS_XHDTYPE):
            # Patch the TarInfo object with the extended header info.
            next._apply_pax_info(pax_headers, tarfile.encoding, tarfile.errors)
            next.offset = self.offset

            if "size" in pax_headers:
                # If the extended header replaces the size field,
                # we need to recalculate the offset where the next
                # header starts.
                offset = next.offset_data
                if next.isreg() or next.type not in SUPPORTED_TYPES:
                    offset += next._block(next.size)
                tarfile.offset = offset

        return next

    def _proc_gnusparse_00(self, next, pax_headers, buf):
        """Process a GNU tar extended sparse header, version 0.0.
        """
        offsets = []
        for match in re.finditer(br"\d+ GNU.sparse.offset=(\d+)\n", buf):
            offsets.append(int(match.group(1)))
        numbytes = []
        for match in re.finditer(br"\d+ GNU.sparse.numbytes=(\d+)\n", buf):
            numbytes.append(int(match.group(1)))
        next.sparse = list(zip(offsets, numbytes))

    def _proc_gnusparse_01(self, next, pax_headers):
        """Process a GNU tar extended sparse header, version 0.1.
        """
        sparse = [int(x) for x in pax_headers["GNU.sparse.map"].split(",")]
        next.sparse = list(zip(sparse[::2], sparse[1::2]))

    def _proc_gnusparse_10(self, next, pax_headers, tarfile):
        """Process a GNU tar extended sparse header, version 1.0.
        """
        fields = None
        sparse = []
        buf = tarfile.fileobj.read(BLOCKSIZE)
        fields, buf = buf.split(b"\n", 1)
        fields = int(fields)
        while len(sparse) < fields * 2:
            if b"\n" not in buf:
                buf += tarfile.fileobj.read(BLOCKSIZE)
            number, buf = buf.split(b"\n", 1)
            sparse.append(int(number))
        next.offset_data = tarfile.fileobj.tell()
        next.sparse = list(zip(sparse[::2], sparse[1::2]))

    def _apply_pax_info(self, pax_headers, encoding, errors):
        """Replace fields with supplemental information from a previous
           pax extended or global header.
        """
        for keyword, value in pax_headers.items():
            if keyword == "GNU.sparse.name":
                setattr(self, "path", value)
            elif keyword == "GNU.sparse.size":
                setattr(self, "size", int(value))
            elif keyword == "GNU.sparse.realsize":
                setattr(self, "size", int(value))
            elif keyword in PAX_FIELDS:
                if keyword in PAX_NUMBER_FIELDS:
                    try:
                        value = PAX_NUMBER_FIELDS[keyword](value)
                    except ValueError:
                        value = 0
                if keyword == "path":
                    value = value.rstrip("/")
                setattr(self, keyword, value)

        self.pax_headers = pax_headers.copy()

    def _decode_pax_field(self, value, encoding, fallback_encoding, fallback_errors):
        """Decode a single field from a pax record.
        """
        try:
            return value.decode(encoding, "strict")
        except UnicodeDecodeError:
            return value.decode(fallback_encoding, fallback_errors)

    def _block(self, count):
        """Round up a byte count by BLOCKSIZE and return it,
           e.g. _block(834) => 1024.
        """
        blocks, remainder = divmod(count, BLOCKSIZE)
        if remainder:
            blocks += 1
        return blocks * BLOCKSIZE

    def isreg(self):
        return self.type in REGULAR_TYPES
    def isfile(self):
        return self.isreg()
    def isdir(self):
        return self.type == DIRTYPE
    def issym(self):
        return self.type == SYMTYPE
    def islnk(self):
        return self.type == LNKTYPE
    def ischr(self):
        return self.type == CHRTYPE
    def isblk(self):
        return self.type == BLKTYPE
    def isfifo(self):
        return self.type == FIFOTYPE
    def issparse(self):
        return self.sparse is not None
    def isdev(self):
        return self.type in (CHRTYPE, BLKTYPE, FIFOTYPE)
# class TarInfo

class TarFile(object):
    """The TarFile Class provides an interface to tar archives.
    """

    debug = 0                   # May be set from 0 (no msgs) to 3 (all msgs)

    dereference = False         # If true, add content of linked file to the
                                # tar file, else the link.

    ignore_zeros = False        # If true, skips empty or invalid blocks and
                                # continues processing.

    errorlevel = 1              # If 0, fatal errors only appear in debug
                                # messages (if debug >= 0). If > 0, errors
                                # are passed to the caller as exceptions.

    format = DEFAULT_FORMAT     # The format to use when creating an archive.

    encoding = ENCODING         # Encoding for 8-bit character strings.

    errors = None               # Error handler for unicode conversion.

    tarinfo = TarInfo           # The default TarInfo class to use.

    fileobject = ExFileObject   # The default ExFileObject class to use.

    def __init__(self, name=None, mode="r", fileobj=None, format=None,
            tarinfo=None, dereference=None, ignore_zeros=None, encoding=None,
            errors="surrogateescape", pax_headers=None, debug=None, errorlevel=None):
        """Open an (uncompressed) tar archive `name'. `mode' is either 'r' to
           read from an existing archive, 'a' to append data to an existing
           file or 'w' to create a new file overwriting an existing one. `mode'
           defaults to 'r'.
           If `fileobj' is given, it is used for reading or writing data. If it
           can be determined, `mode' is overridden by `fileobj's mode.
           `fileobj' is not closed, when TarFile is closed.
        """
        if len(mode) > 1 or mode not in "raw":
            raise ValueError("mode must be 'r', 'a' or 'w'")
        self.mode = mode
        self._mode = {"r": "rb", "a": "r+b", "w": "wb"}[mode]

        if not fileobj:
            if self.mode == "a" and not os.path.exists(name):
                # Create nonexistent files in append mode.
                self.mode = "w"
                self._mode = "wb"
            fileobj = bltn_open(name, self._mode)
            self._extfileobj = False
        else:
            if name is None and hasattr(fileobj, "name"):
                name = fileobj.name
            if hasattr(fileobj, "mode"):
                self._mode = fileobj.mode
            self._extfileobj = True
        self.name = os.path.abspath(name) if name else None
        self.fileobj = fileobj

        # Init attributes.
        if format is not None:
            self.format = format
        if tarinfo is not None:
            self.tarinfo = tarinfo
        if dereference is not None:
            self.dereference = dereference
        if ignore_zeros is not None:
            self.ignore_zeros = ignore_zeros
        if encoding is not None:
            self.encoding = encoding
        self.errors = errors

        if pax_headers is not None and self.format == PAX_FORMAT:
            self.pax_headers = pax_headers
        else:
            self.pax_headers = {}

        if debug is not None:
            self.debug = debug
        if errorlevel is not None:
            self.errorlevel = errorlevel

        # Init datastructures.
        self.closed = False
        self.members = []       # list of members as TarInfo objects
        self._loaded = False    # flag if all members have been read
        self.offset = self.fileobj.tell()
                                # current position in the archive file
        self.inodes = {}        # dictionary caching the inodes of
                                # archive members already added

        try:
            if self.mode == "r":
                self.firstmember = None
                self.firstmember = self.next()

            if self.mode == "a":
                # Move to the end of the archive,
                # before the first empty block.
                while True:
                    self.fileobj.seek(self.offset)
                    try:
                        tarinfo = self.tarinfo.fromtarfile(self)
                        self.members.append(tarinfo)
                    except EOFHeaderError:
                        self.fileobj.seek(self.offset)
                        break
                    except HeaderError as e:
                        raise ReadError(str(e))

            if self.mode in "aw":
                self._loaded = True

                if self.pax_headers:
                    buf = self.tarinfo.create_pax_global_header(self.pax_headers.copy())
                    self.fileobj.write(buf)
                    self.offset += len(buf)
        except:
            if not self._extfileobj:
                self.fileobj.close()
            self.closed = True
            raise

    #--------------------------------------------------------------------------
    # Below are the classmethods which act as alternate constructors to the
    # TarFile class. The open() method is the only one that is needed for
    # public use; it is the "super"-constructor and is able to select an
    # adequate "sub"-constructor for a particular compression using the mapping
    # from OPEN_METH.
    #
    # This concept allows one to subclass TarFile without losing the comfort of
    # the super-constructor. A sub-constructor is registered and made available
    # by adding it to the mapping in OPEN_METH.

    @classmethod
    def open(cls, name=None, mode="r", fileobj=None, bufsize=RECORDSIZE, **kwargs):
        """Open a tar archive for reading, writing or appending. Return
           an appropriate TarFile class.

           mode:
           'r' or 'r:*' open for reading with transparent compression
           'r:'         open for reading exclusively uncompressed
           'r:gz'       open for reading with gzip compression
           'r:bz2'      open for reading with bzip2 compression
           'a' or 'a:'  open for appending, creating the file if necessary
           'w' or 'w:'  open for writing without compression
           'w:gz'       open for writing with gzip compression
           'w:bz2'      open for writing with bzip2 compression

           'r|*'        open a stream of tar blocks with transparent compression
           'r|'         open an uncompressed stream of tar blocks for reading
           'r|gz'       open a gzip compressed stream of tar blocks
           'r|bz2'      open a bzip2 compressed stream of tar blocks
           'w|'         open an uncompressed stream for writing
           'w|gz'       open a gzip compressed stream for writing
           'w|bz2'      open a bzip2 compressed stream for writing
        """

        if not name and not fileobj:
            raise ValueError("nothing to open")

        if mode in ("r", "r:*"):
            # Find out which *open() is appropriate for opening the file.
            for comptype in cls.OPEN_METH:
                func = getattr(cls, cls.OPEN_METH[comptype])
                if fileobj is not None:
                    saved_pos = fileobj.tell()
                try:
                    return func(name, "r", fileobj, **kwargs)
                except (ReadError, CompressionError) as e:
                    if fileobj is not None:
                        fileobj.seek(saved_pos)
                    continue
            raise ReadError("file could not be opened successfully")

        elif ":" in mode:
            filemode, comptype = mode.split(":", 1)
            filemode = filemode or "r"
            comptype = comptype or "tar"

            # Select the *open() function according to
            # given compression.
            if comptype in cls.OPEN_METH:
                func = getattr(cls, cls.OPEN_METH[comptype])
            else:
                raise CompressionError("unknown compression type %r" % comptype)
            return func(name, filemode, fileobj, **kwargs)

        elif "|" in mode:
            filemode, comptype = mode.split("|", 1)
            filemode = filemode or "r"
            comptype = comptype or "tar"

            if filemode not in "rw":
                raise ValueError("mode must be 'r' or 'w'")

            stream = _Stream(name, filemode, comptype, fileobj, bufsize)
            try:
                t = cls(name, filemode, stream, **kwargs)
            except:
                stream.close()
                raise
            t._extfileobj = False
            return t

        elif mode in "aw":
            return cls.taropen(name, mode, fileobj, **kwargs)

        raise ValueError("undiscernible mode")

    @classmethod
    def taropen(cls, name, mode="r", fileobj=None, **kwargs):
        """Open uncompressed tar archive name for reading or writing.
        """
        if len(mode) > 1 or mode not in "raw":
            raise ValueError("mode must be 'r', 'a' or 'w'")
        return cls(name, mode, fileobj, **kwargs)

    @classmethod
    def gzopen(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs):
        """Open gzip compressed tar archive name for reading or writing.
           Appending is not allowed.
        """
        if len(mode) > 1 or mode not in "rw":
            raise ValueError("mode must be 'r' or 'w'")

        try:
            import gzip
            gzip.GzipFile
        except (ImportError, AttributeError):
            raise CompressionError("gzip module is not available")

        extfileobj = fileobj is not None
        try:
            fileobj = gzip.GzipFile(name, mode + "b", compresslevel, fileobj)
            t = cls.taropen(name, mode, fileobj, **kwargs)
        except IOError:
            if not extfileobj and fileobj is not None:
                fileobj.close()
            if fileobj is None:
                raise
            raise ReadError("not a gzip file")
        except:
            if not extfileobj and fileobj is not None:
                fileobj.close()
            raise
        t._extfileobj = extfileobj
        return t

    @classmethod
    def bz2open(cls, name, mode="r", fileobj=None, compresslevel=9, **kwargs):
        """Open bzip2 compressed tar archive name for reading or writing.
           Appending is not allowed.
        """
        if len(mode) > 1 or mode not in "rw":
            raise ValueError("mode must be 'r' or 'w'.")

        try:
            import bz2
        except ImportError:
            raise CompressionError("bz2 module is not available")

        if fileobj is not None:
            fileobj = _BZ2Proxy(fileobj, mode)
        else:
            fileobj = bz2.BZ2File(name, mode, compresslevel=compresslevel)

        try:
            t = cls.taropen(name, mode, fileobj, **kwargs)
        except (IOError, EOFError):
            fileobj.close()
            raise ReadError("not a bzip2 file")
        t._extfileobj = False
        return t

    # All *open() methods are registered here.
    OPEN_METH = {
        "tar": "taropen",   # uncompressed tar
        "gz":  "gzopen",    # gzip compressed tar
        "bz2": "bz2open"    # bzip2 compressed tar
    }

    #--------------------------------------------------------------------------
    # The public methods which TarFile provides:

    def close(self):
        """Close the TarFile. In write-mode, two finishing zero blocks are
           appended to the archive.
        """
        if self.closed:
            return

        if self.mode in "aw":
            self.fileobj.write(NUL * (BLOCKSIZE * 2))
            self.offset += (BLOCKSIZE * 2)
            # fill up the end with zero-blocks
            # (like option -b20 for tar does)
            blocks, remainder = divmod(self.offset, RECORDSIZE)
            if remainder > 0:
                self.fileobj.write(NUL * (RECORDSIZE - remainder))

        if not self._extfileobj:
            self.fileobj.close()
        self.closed = True

    def getmember(self, name):
        """Return a TarInfo object for member `name'. If `name' can not be
           found in the archive, KeyError is raised. If a member occurs more
           than once in the archive, its last occurrence is assumed to be the
           most up-to-date version.
        """
        tarinfo = self._getmember(name)
        if tarinfo is None:
            raise KeyError("filename %r not found" % name)
        return tarinfo

    def getmembers(self):
        """Return the members of the archive as a list of TarInfo objects. The
           list has the same order as the members in the archive.
        """
        self._check()
        if not self._loaded:    # if we want to obtain a list of
            self._load()        # all members, we first have to
                                # scan the whole archive.
        return self.members

    def getnames(self):
        """Return the members of the archive as a list of their names. It has
           the same order as the list returned by getmembers().
        """
        return [tarinfo.name for tarinfo in self.getmembers()]

    def gettarinfo(self, name=None, arcname=None, fileobj=None):
        """Create a TarInfo object for either the file `name' or the file
           object `fileobj' (using os.fstat on its file descriptor). You can
           modify some of the TarInfo's attributes before you add it using
           addfile(). If given, `arcname' specifies an alternative name for the
           file in the archive.
        """
        self._check("aw")

        # When fileobj is given, replace name by
        # fileobj's real name.
        if fileobj is not None:
            name = fileobj.name

        # Building the name of the member in the archive.
        # Backward slashes are converted to forward slashes,
        # Absolute paths are turned to relative paths.
        if arcname is None:
            arcname = name
        drv, arcname = os.path.splitdrive(arcname)
        arcname = arcname.replace(os.sep, "/")
        arcname = arcname.lstrip("/")

        # Now, fill the TarInfo object with
        # information specific for the file.
        tarinfo = self.tarinfo()
        tarinfo.tarfile = self

        # Use os.stat or os.lstat, depending on platform
        # and if symlinks shall be resolved.
        if fileobj is None:
            if hasattr(os, "lstat") and not self.dereference:
                statres = os.lstat(name)
            else:
                statres = os.stat(name)
        else:
            statres = os.fstat(fileobj.fileno())
        linkname = ""

        stmd = statres.st_mode
        if stat.S_ISREG(stmd):
            inode = (statres.st_ino, statres.st_dev)
            if not self.dereference and statres.st_nlink > 1 and \
                    inode in self.inodes and arcname != self.inodes[inode]:
                # Is it a hardlink to an already
                # archived file?
                type = LNKTYPE
                linkname = self.inodes[inode]
            else:
                # The inode is added only if its valid.
                # For win32 it is always 0.
                type = REGTYPE
                if inode[0]:
                    self.inodes[inode] = arcname
        elif stat.S_ISDIR(stmd):
            type = DIRTYPE
        elif stat.S_ISFIFO(stmd):
            type = FIFOTYPE
        elif stat.S_ISLNK(stmd):
            type = SYMTYPE
            linkname = os.readlink(name)
        elif stat.S_ISCHR(stmd):
            type = CHRTYPE
        elif stat.S_ISBLK(stmd):
            type = BLKTYPE
        else:
            return None

        # Fill the TarInfo object with all
        # information we can get.
        tarinfo.name = arcname
        tarinfo.mode = stmd
        tarinfo.uid = statres.st_uid
        tarinfo.gid = statres.st_gid
        if type == REGTYPE:
            tarinfo.size = statres.st_size
        else:
            tarinfo.size = 0
        tarinfo.mtime = statres.st_mtime
        tarinfo.type = type
        tarinfo.linkname = linkname
        if pwd:
            try:
                tarinfo.uname = pwd.getpwuid(tarinfo.uid)[0]
            except KeyError:
                pass
        if grp:
            try:
                tarinfo.gname = grp.getgrgid(tarinfo.gid)[0]
            except KeyError:
                pass

        if type in (CHRTYPE, BLKTYPE):
            if hasattr(os, "major") and hasattr(os, "minor"):
                tarinfo.devmajor = os.major(statres.st_rdev)
                tarinfo.devminor = os.minor(statres.st_rdev)
        return tarinfo

    def list(self, verbose=True):
        """Print a table of contents to sys.stdout. If `verbose' is False, only
           the names of the members are printed. If it is True, an `ls -l'-like
           output is produced.
        """
        self._check()

        for tarinfo in self:
            if verbose:
                print(filemode(tarinfo.mode), end=' ')
                print("%s/%s" % (tarinfo.uname or tarinfo.uid,
                                 tarinfo.gname or tarinfo.gid), end=' ')
                if tarinfo.ischr() or tarinfo.isblk():
                    print("%10s" % ("%d,%d" \
                                    % (tarinfo.devmajor, tarinfo.devminor)), end=' ')
                else:
                    print("%10d" % tarinfo.size, end=' ')
                print("%d-%02d-%02d %02d:%02d:%02d" \
                      % time.localtime(tarinfo.mtime)[:6], end=' ')

            print(tarinfo.name + ("/" if tarinfo.isdir() else ""), end=' ')

            if verbose:
                if tarinfo.issym():
                    print("->", tarinfo.linkname, end=' ')
                if tarinfo.islnk():
                    print("link to", tarinfo.linkname, end=' ')
            print()

    def add(self, name, arcname=None, recursive=True, exclude=None, filter=None):
        """Add the file `name' to the archive. `name' may be any type of file
           (directory, fifo, symbolic link, etc.). If given, `arcname'
           specifies an alternative name for the file in the archive.
           Directories are added recursively by default. This can be avoided by
           setting `recursive' to False. `exclude' is a function that should
           return True for each filename to be excluded. `filter' is a function
           that expects a TarInfo object argument and returns the changed
           TarInfo object, if it returns None the TarInfo object will be
           excluded from the archive.
        """
        self._check("aw")

        if arcname is None:
            arcname = name

        # Exclude pathnames.
        if exclude is not None:
            import warnings
            warnings.warn("use the filter argument instead",
                    DeprecationWarning, 2)
            if exclude(name):
                self._dbg(2, "tarfile: Excluded %r" % name)
                return

        # Skip if somebody tries to archive the archive...
        if self.name is not None and os.path.abspath(name) == self.name:
            self._dbg(2, "tarfile: Skipped %r" % name)
            return

        self._dbg(1, name)

        # Create a TarInfo object from the file.
        tarinfo = self.gettarinfo(name, arcname)

        if tarinfo is None:
            self._dbg(1, "tarfile: Unsupported type %r" % name)
            return

        # Change or exclude the TarInfo object.
        if filter is not None:
            tarinfo = filter(tarinfo)
            if tarinfo is None:
                self._dbg(2, "tarfile: Excluded %r" % name)
                return

        # Append the tar header and data to the archive.
        if tarinfo.isreg():
            f = bltn_open(name, "rb")
            self.addfile(tarinfo, f)
            f.close()

        elif tarinfo.isdir():
            self.addfile(tarinfo)
            if recursive:
                for f in os.listdir(name):
                    self.add(os.path.join(name, f), os.path.join(arcname, f),
                            recursive, exclude, filter=filter)

        else:
            self.addfile(tarinfo)

    def addfile(self, tarinfo, fileobj=None):
        """Add the TarInfo object `tarinfo' to the archive. If `fileobj' is
           given, tarinfo.size bytes are read from it and added to the archive.
           You can create TarInfo objects using gettarinfo().
           On Windows platforms, `fileobj' should always be opened with mode
           'rb' to avoid irritation about the file size.
        """
        self._check("aw")

        tarinfo = copy.copy(tarinfo)

        buf = tarinfo.tobuf(self.format, self.encoding, self.errors)
        self.fileobj.write(buf)
        self.offset += len(buf)

        # If there's data to follow, append it.
        if fileobj is not None:
            copyfileobj(fileobj, self.fileobj, tarinfo.size)
            blocks, remainder = divmod(tarinfo.size, BLOCKSIZE)
            if remainder > 0:
                self.fileobj.write(NUL * (BLOCKSIZE - remainder))
                blocks += 1
            self.offset += blocks * BLOCKSIZE

        self.members.append(tarinfo)

    def extractall(self, path=".", members=None):
        """Extract all members from the archive to the current working
           directory and set owner, modification time and permissions on
           directories afterwards. `path' specifies a different directory
           to extract to. `members' is optional and must be a subset of the
           list returned by getmembers().
        """
        directories = []

        if members is None:
            members = self

        for tarinfo in members:
            if tarinfo.isdir():
                # Extract directories with a safe mode.
                directories.append(tarinfo)
                tarinfo = copy.copy(tarinfo)
                tarinfo.mode = 0o700
            # Do not set_attrs directories, as we will do that further down
            self.extract(tarinfo, path, set_attrs=not tarinfo.isdir())

        # Reverse sort directories.
        directories.sort(key=lambda a: a.name)
        directories.reverse()

        # Set correct owner, mtime and filemode on directories.
        for tarinfo in directories:
            dirpath = os.path.join(path, tarinfo.name)
            try:
                self.chown(tarinfo, dirpath)
                self.utime(tarinfo, dirpath)
                self.chmod(tarinfo, dirpath)
            except ExtractError as e:
                if self.errorlevel > 1:
                    raise
                else:
                    self._dbg(1, "tarfile: %s" % e)

    def extract(self, member, path="", set_attrs=True):
        """Extract a member from the archive to the current working directory,
           using its full name. Its file information is extracted as accurately
           as possible. `member' may be a filename or a TarInfo object. You can
           specify a different directory using `path'. File attributes (owner,
           mtime, mode) are set unless `set_attrs' is False.
        """
        self._check("r")

        if isinstance(member, str):
            tarinfo = self.getmember(member)
        else:
            tarinfo = member

        # Prepare the link target for makelink().
        if tarinfo.islnk():
            tarinfo._link_target = os.path.join(path, tarinfo.linkname)

        try:
            self._extract_member(tarinfo, os.path.join(path, tarinfo.name),
                                 set_attrs=set_attrs)
        except EnvironmentError as e:
            if self.errorlevel > 0:
                raise
            else:
                if e.filename is None:
                    self._dbg(1, "tarfile: %s" % e.strerror)
                else:
                    self._dbg(1, "tarfile: %s %r" % (e.strerror, e.filename))
        except ExtractError as e:
            if self.errorlevel > 1:
                raise
            else:
                self._dbg(1, "tarfile: %s" % e)

    def extractfile(self, member):
        """Extract a member from the archive as a file object. `member' may be
           a filename or a TarInfo object. If `member' is a regular file, a
           file-like object is returned. If `member' is a link, a file-like
           object is constructed from the link's target. If `member' is none of
           the above, None is returned.
           The file-like object is read-only and provides the following
           methods: read(), readline(), readlines(), seek() and tell()
        """
        self._check("r")

        if isinstance(member, str):
            tarinfo = self.getmember(member)
        else:
            tarinfo = member

        if tarinfo.isreg():
            return self.fileobject(self, tarinfo)

        elif tarinfo.type not in SUPPORTED_TYPES:
            # If a member's type is unknown, it is treated as a
            # regular file.
            return self.fileobject(self, tarinfo)

        elif tarinfo.islnk() or tarinfo.issym():
            if isinstance(self.fileobj, _Stream):
                # A small but ugly workaround for the case that someone tries
                # to extract a (sym)link as a file-object from a non-seekable
                # stream of tar blocks.
                raise StreamError("cannot extract (sym)link as file object")
            else:
                # A (sym)link's file object is its target's file object.
                return self.extractfile(self._find_link_target(tarinfo))
        else:
            # If there's no data associated with the member (directory, chrdev,
            # blkdev, etc.), return None instead of a file object.
            return None

    def _extract_member(self, tarinfo, targetpath, set_attrs=True):
        """Extract the TarInfo object tarinfo to a physical
           file called targetpath.
        """
        # Fetch the TarInfo object for the given name
        # and build the destination pathname, replacing
        # forward slashes to platform specific separators.
        targetpath = targetpath.rstrip("/")
        targetpath = targetpath.replace("/", os.sep)

        # Create all upper directories.
        upperdirs = os.path.dirname(targetpath)
        if upperdirs and not os.path.exists(upperdirs):
            # Create directories that are not part of the archive with
            # default permissions.
            os.makedirs(upperdirs)

        if tarinfo.islnk() or tarinfo.issym():
            self._dbg(1, "%s -> %s" % (tarinfo.name, tarinfo.linkname))
        else:
            self._dbg(1, tarinfo.name)

        if tarinfo.isreg():
            self.makefile(tarinfo, targetpath)
        elif tarinfo.isdir():
            self.makedir(tarinfo, targetpath)
        elif tarinfo.isfifo():
            self.makefifo(tarinfo, targetpath)
        elif tarinfo.ischr() or tarinfo.isblk():
            self.makedev(tarinfo, targetpath)
        elif tarinfo.islnk() or tarinfo.issym():
            self.makelink(tarinfo, targetpath)
        elif tarinfo.type not in SUPPORTED_TYPES:
            self.makeunknown(tarinfo, targetpath)
        else:
            self.makefile(tarinfo, targetpath)

        if set_attrs:
            self.chown(tarinfo, targetpath)
            if not tarinfo.issym():
                self.chmod(tarinfo, targetpath)
                self.utime(tarinfo, targetpath)

    #--------------------------------------------------------------------------
    # Below are the different file methods. They are called via
    # _extract_member() when extract() is called. They can be replaced in a
    # subclass to implement other functionality.

    def makedir(self, tarinfo, targetpath):
        """Make a directory called targetpath.
        """
        try:
            # Use a safe mode for the directory, the real mode is set
            # later in _extract_member().
            os.mkdir(targetpath, 0o700)
        except EnvironmentError as e:
            if e.errno != errno.EEXIST:
                raise

    def makefile(self, tarinfo, targetpath):
        """Make a file called targetpath.
        """
        source = self.fileobj
        source.seek(tarinfo.offset_data)
        target = bltn_open(targetpath, "wb")
        if tarinfo.sparse is not None:
            for offset, size in tarinfo.sparse:
                target.seek(offset)
                copyfileobj(source, target, size)
        else:
            copyfileobj(source, target, tarinfo.size)
        target.seek(tarinfo.size)
        target.truncate()
        target.close()

    def makeunknown(self, tarinfo, targetpath):
        """Make a file from a TarInfo object with an unknown type
           at targetpath.
        """
        self.makefile(tarinfo, targetpath)
        self._dbg(1, "tarfile: Unknown file type %r, " \
                     "extracted as regular file." % tarinfo.type)

    def makefifo(self, tarinfo, targetpath):
        """Make a fifo called targetpath.
        """
        if hasattr(os, "mkfifo"):
            os.mkfifo(targetpath)
        else:
            raise ExtractError("fifo not supported by system")

    def makedev(self, tarinfo, targetpath):
        """Make a character or block device called targetpath.
        """
        if not hasattr(os, "mknod") or not hasattr(os, "makedev"):
            raise ExtractError("special devices not supported by system")

        mode = tarinfo.mode
        if tarinfo.isblk():
            mode |= stat.S_IFBLK
        else:
            mode |= stat.S_IFCHR

        os.mknod(targetpath, mode,
                 os.makedev(tarinfo.devmajor, tarinfo.devminor))

    def makelink(self, tarinfo, targetpath):
        """Make a (symbolic) link called targetpath. If it cannot be created
          (platform limitation), we try to make a copy of the referenced file
          instead of a link.
        """
        try:
            # For systems that support symbolic and hard links.
            if tarinfo.issym():
                os.symlink(tarinfo.linkname, targetpath)
            else:
                # See extract().
                if os.path.exists(tarinfo._link_target):
                    os.link(tarinfo._link_target, targetpath)
                else:
                    self._extract_member(self._find_link_target(tarinfo),
                                         targetpath)
        except symlink_exception:
            if tarinfo.issym():
                linkpath = os.path.join(os.path.dirname(tarinfo.name),
                                        tarinfo.linkname)
            else:
                linkpath = tarinfo.linkname
        else:
            try:
                self._extract_member(self._find_link_target(tarinfo),
                                     targetpath)
            except KeyError:
                raise ExtractError("unable to resolve link inside archive")

    def chown(self, tarinfo, targetpath):
        """Set owner of targetpath according to tarinfo.
        """
        if pwd and hasattr(os, "geteuid") and os.geteuid() == 0:
            # We have to be root to do so.
            try:
                g = grp.getgrnam(tarinfo.gname)[2]
            except KeyError:
                g = tarinfo.gid
            try:
                u = pwd.getpwnam(tarinfo.uname)[2]
            except KeyError:
                u = tarinfo.uid
            try:
                if tarinfo.issym() and hasattr(os, "lchown"):
                    os.lchown(targetpath, u, g)
                else:
                    if sys.platform != "os2emx":
                        os.chown(targetpath, u, g)
            except EnvironmentError as e:
                raise ExtractError("could not change owner")

    def chmod(self, tarinfo, targetpath):
        """Set file permissions of targetpath according to tarinfo.
        """
        if hasattr(os, 'chmod'):
            try:
                os.chmod(targetpath, tarinfo.mode)
            except EnvironmentError as e:
                raise ExtractError("could not change mode")

    def utime(self, tarinfo, targetpath):
        """Set modification time of targetpath according to tarinfo.
        """
        if not hasattr(os, 'utime'):
            return
        try:
            os.utime(targetpath, (tarinfo.mtime, tarinfo.mtime))
        except EnvironmentError as e:
            raise ExtractError("could not change modification time")

    #--------------------------------------------------------------------------
    def next(self):
        """Return the next member of the archive as a TarInfo object, when
           TarFile is opened for reading. Return None if there is no more
           available.
        """
        self._check("ra")
        if self.firstmember is not None:
            m = self.firstmember
            self.firstmember = None
            return m

        # Read the next block.
        self.fileobj.seek(self.offset)
        tarinfo = None
        while True:
            try:
                tarinfo = self.tarinfo.fromtarfile(self)
            except EOFHeaderError as e:
                if self.ignore_zeros:
                    self._dbg(2, "0x%X: %s" % (self.offset, e))
                    self.offset += BLOCKSIZE
                    continue
            except InvalidHeaderError as e:
                if self.ignore_zeros:
                    self._dbg(2, "0x%X: %s" % (self.offset, e))
                    self.offset += BLOCKSIZE
                    continue
                elif self.offset == 0:
                    raise ReadError(str(e))
            except EmptyHeaderError:
                if self.offset == 0:
                    raise ReadError("empty file")
            except TruncatedHeaderError as e:
                if self.offset == 0:
                    raise ReadError(str(e))
            except SubsequentHeaderError as e:
                raise ReadError(str(e))
            break

        if tarinfo is not None:
            self.members.append(tarinfo)
        else:
            self._loaded = True

        return tarinfo

    #--------------------------------------------------------------------------
    # Little helper methods:

    def _getmember(self, name, tarinfo=None, normalize=False):
        """Find an archive member by name from bottom to top.
           If tarinfo is given, it is used as the starting point.
        """
        # Ensure that all members have been loaded.
        members = self.getmembers()

        # Limit the member search list up to tarinfo.
        if tarinfo is not None:
            members = members[:members.index(tarinfo)]

        if normalize:
            name = os.path.normpath(name)

        for member in reversed(members):
            if normalize:
                member_name = os.path.normpath(member.name)
            else:
                member_name = member.name

            if name == member_name:
                return member

    def _load(self):
        """Read through the entire archive file and look for readable
           members.
        """
        while True:
            tarinfo = self.next()
            if tarinfo is None:
                break
        self._loaded = True

    def _check(self, mode=None):
        """Check if TarFile is still open, and if the operation's mode
           corresponds to TarFile's mode.
        """
        if self.closed:
            raise IOError("%s is closed" % self.__class__.__name__)
        if mode is not None and self.mode not in mode:
            raise IOError("bad operation for mode %r" % self.mode)

    def _find_link_target(self, tarinfo):
        """Find the target member of a symlink or hardlink member in the
           archive.
        """
        if tarinfo.issym():
            # Always search the entire archive.
            linkname = os.path.dirname(tarinfo.name) + "/" + tarinfo.linkname
            limit = None
        else:
            # Search the archive before the link, because a hard link is
            # just a reference to an already archived file.
            linkname = tarinfo.linkname
            limit = tarinfo

        member = self._getmember(linkname, tarinfo=limit, normalize=True)
        if member is None:
            raise KeyError("linkname %r not found" % linkname)
        return member

    def __iter__(self):
        """Provide an iterator object.
        """
        if self._loaded:
            return iter(self.members)
        else:
            return TarIter(self)

    def _dbg(self, level, msg):
        """Write debugging output to sys.stderr.
        """
        if level <= self.debug:
            print(msg, file=sys.stderr)

    def __enter__(self):
        self._check()
        return self

    def __exit__(self, type, value, traceback):
        if type is None:
            self.close()
        else:
            # An exception occurred. We must not call close() because
            # it would try to write end-of-archive blocks and padding.
            if not self._extfileobj:
                self.fileobj.close()
            self.closed = True
# class TarFile

class TarIter(object):
    """Iterator Class.

       for tarinfo in TarFile(...):
           suite...
    """

    def __init__(self, tarfile):
        """Construct a TarIter object.
        """
        self.tarfile = tarfile
        self.index = 0
    def __iter__(self):
        """Return iterator object.
        """
        return self

    def __next__(self):
        """Return the next item using TarFile's next() method.
           When all members have been read, set TarFile as _loaded.
        """
        # Fix for SF #1100429: Under rare circumstances it can
        # happen that getmembers() is called during iteration,
        # which will cause TarIter to stop prematurely.
        if not self.tarfile._loaded:
            tarinfo = self.tarfile.next()
            if not tarinfo:
                self.tarfile._loaded = True
                raise StopIteration
        else:
            try:
                tarinfo = self.tarfile.members[self.index]
            except IndexError:
                raise StopIteration
        self.index += 1
        return tarinfo

    next = __next__ # for Python 2.x

#--------------------
# exported functions
#--------------------
def is_tarfile(name):
    """Return True if name points to a tar archive that we
       are able to handle, else return False.
    """
    try:
        t = open(name)
        t.close()
        return True
    except TarError:
        return False

bltn_open = open
open = TarFile.open
site-packages/pip/_vendor/distlib/locators.py000064400000145375150564745010015370 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2012-2015 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#

import gzip
from io import BytesIO
import json
import logging
import os
import posixpath
import re
try:
    import threading
except ImportError:  # pragma: no cover
    import dummy_threading as threading
import zlib

from . import DistlibException
from .compat import (urljoin, urlparse, urlunparse, url2pathname, pathname2url,
                     queue, quote, unescape, build_opener,
                     HTTPRedirectHandler as BaseRedirectHandler, text_type,
                     Request, HTTPError, URLError)
from .database import Distribution, DistributionPath, make_dist
from .metadata import Metadata, MetadataInvalidError
from .util import (cached_property, ensure_slash, split_filename, get_project_data,
                   parse_requirement, parse_name_and_version, ServerProxy,
                   normalize_name)
from .version import get_scheme, UnsupportedVersionError
from .wheel import Wheel, is_compatible

logger = logging.getLogger(__name__)

HASHER_HASH = re.compile(r'^(\w+)=([a-f0-9]+)')
CHARSET = re.compile(r';\s*charset\s*=\s*(.*)\s*$', re.I)
HTML_CONTENT_TYPE = re.compile('text/html|application/x(ht)?ml')
DEFAULT_INDEX = 'https://pypi.org/pypi'

def get_all_distribution_names(url=None):
    """
    Return all distribution names known by an index.
    :param url: The URL of the index.
    :return: A list of all known distribution names.
    """
    if url is None:
        url = DEFAULT_INDEX
    client = ServerProxy(url, timeout=3.0)
    try:
        return client.list_packages()
    finally:
        client('close')()

class RedirectHandler(BaseRedirectHandler):
    """
    A class to work around a bug in some Python 3.2.x releases.
    """
    # There's a bug in the base version for some 3.2.x
    # (e.g. 3.2.2 on Ubuntu Oneiric). If a Location header
    # returns e.g. /abc, it bails because it says the scheme ''
    # is bogus, when actually it should use the request's
    # URL for the scheme. See Python issue #13696.
    def http_error_302(self, req, fp, code, msg, headers):
        # Some servers (incorrectly) return multiple Location headers
        # (so probably same goes for URI).  Use first header.
        newurl = None
        for key in ('location', 'uri'):
            if key in headers:
                newurl = headers[key]
                break
        if newurl is None:  # pragma: no cover
            return
        urlparts = urlparse(newurl)
        if urlparts.scheme == '':
            newurl = urljoin(req.get_full_url(), newurl)
            if hasattr(headers, 'replace_header'):
                headers.replace_header(key, newurl)
            else:
                headers[key] = newurl
        return BaseRedirectHandler.http_error_302(self, req, fp, code, msg,
                                                  headers)

    http_error_301 = http_error_303 = http_error_307 = http_error_302

class Locator(object):
    """
    A base class for locators - things that locate distributions.
    """
    source_extensions = ('.tar.gz', '.tar.bz2', '.tar', '.zip', '.tgz', '.tbz')
    binary_extensions = ('.egg', '.exe', '.whl')
    excluded_extensions = ('.pdf',)

    # A list of tags indicating which wheels you want to match. The default
    # value of None matches against the tags compatible with the running
    # Python. If you want to match other values, set wheel_tags on a locator
    # instance to a list of tuples (pyver, abi, arch) which you want to match.
    wheel_tags = None

    downloadable_extensions = source_extensions + ('.whl',)

    def __init__(self, scheme='default'):
        """
        Initialise an instance.
        :param scheme: Because locators look for most recent versions, they
                       need to know the version scheme to use. This specifies
                       the current PEP-recommended scheme - use ``'legacy'``
                       if you need to support existing distributions on PyPI.
        """
        self._cache = {}
        self.scheme = scheme
        # Because of bugs in some of the handlers on some of the platforms,
        # we use our own opener rather than just using urlopen.
        self.opener = build_opener(RedirectHandler())
        # If get_project() is called from locate(), the matcher instance
        # is set from the requirement passed to locate(). See issue #18 for
        # why this can be useful to know.
        self.matcher = None
        self.errors = queue.Queue()

    def get_errors(self):
        """
        Return any errors which have occurred.
        """
        result = []
        while not self.errors.empty():  # pragma: no cover
            try:
                e = self.errors.get(False)
                result.append(e)
            except self.errors.Empty:
                continue
            self.errors.task_done()
        return result

    def clear_errors(self):
        """
        Clear any errors which may have been logged.
        """
        # Just get the errors and throw them away
        self.get_errors()

    def clear_cache(self):
        self._cache.clear()

    def _get_scheme(self):
        return self._scheme

    def _set_scheme(self, value):
        self._scheme = value

    scheme = property(_get_scheme, _set_scheme)

    def _get_project(self, name):
        """
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This should be implemented in subclasses.

        If called from a locate() request, self.matcher will be set to a
        matcher for the requirement to satisfy, otherwise it will be None.
        """
        raise NotImplementedError('Please implement in the subclass')

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Please implement in the subclass')

    def get_project(self, name):
        """
        For a given project, get a dictionary mapping available versions to Distribution
        instances.

        This calls _get_project to do all the work, and just implements a caching layer on top.
        """
        if self._cache is None:  # pragma: no cover
            result = self._get_project(name)
        elif name in self._cache:
            result = self._cache[name]
        else:
            self.clear_errors()
            result = self._get_project(name)
            self._cache[name] = result
        return result

    def score_url(self, url):
        """
        Give an url a score which can be used to choose preferred URLs
        for a given project release.
        """
        t = urlparse(url)
        basename = posixpath.basename(t.path)
        compatible = True
        is_wheel = basename.endswith('.whl')
        is_downloadable = basename.endswith(self.downloadable_extensions)
        if is_wheel:
            compatible = is_compatible(Wheel(basename), self.wheel_tags)
        return (t.scheme == 'https', 'pypi.org' in t.netloc,
                is_downloadable, is_wheel, compatible, basename)

    def prefer_url(self, url1, url2):
        """
        Choose one of two URLs where both are candidates for distribution
        archives for the same version of a distribution (for example,
        .tar.gz vs. zip).

        The current implementation favours https:// URLs over http://, archives
        from PyPI over those from other locations, wheel compatibility (if a
        wheel) and then the archive name.
        """
        result = url2
        if url1:
            s1 = self.score_url(url1)
            s2 = self.score_url(url2)
            if s1 > s2:
                result = url1
            if result != url2:
                logger.debug('Not replacing %r with %r', url1, url2)
            else:
                logger.debug('Replacing %r with %r', url1, url2)
        return result

    def split_filename(self, filename, project_name):
        """
        Attempt to split a filename in project name, version and Python version.
        """
        return split_filename(filename, project_name)

    def convert_url_to_download_info(self, url, project_name):
        """
        See if a URL is a candidate for a download URL for a project (the URL
        has typically been scraped from an HTML page).

        If it is, a dictionary is returned with keys "name", "version",
        "filename" and "url"; otherwise, None is returned.
        """
        def same_project(name1, name2):
            return normalize_name(name1) == normalize_name(name2)

        result = None
        scheme, netloc, path, params, query, frag = urlparse(url)
        if frag.lower().startswith('egg='):  # pragma: no cover
            logger.debug('%s: version hint in fragment: %r',
                         project_name, frag)
        m = HASHER_HASH.match(frag)
        if m:
            algo, digest = m.groups()
        else:
            algo, digest = None, None
        origpath = path
        if path and path[-1] == '/':  # pragma: no cover
            path = path[:-1]
        if path.endswith('.whl'):
            try:
                wheel = Wheel(path)
                if not is_compatible(wheel, self.wheel_tags):
                    logger.debug('Wheel not compatible: %s', path)
                else:
                    if project_name is None:
                        include = True
                    else:
                        include = same_project(wheel.name, project_name)
                    if include:
                        result = {
                            'name': wheel.name,
                            'version': wheel.version,
                            'filename': wheel.filename,
                            'url': urlunparse((scheme, netloc, origpath,
                                               params, query, '')),
                            'python-version': ', '.join(
                                ['.'.join(list(v[2:])) for v in wheel.pyver]),
                        }
            except Exception as e:  # pragma: no cover
                logger.warning('invalid path for wheel: %s', path)
        elif not path.endswith(self.downloadable_extensions):  # pragma: no cover
            logger.debug('Not downloadable: %s', path)
        else:  # downloadable extension
            path = filename = posixpath.basename(path)
            for ext in self.downloadable_extensions:
                if path.endswith(ext):
                    path = path[:-len(ext)]
                    t = self.split_filename(path, project_name)
                    if not t:  # pragma: no cover
                        logger.debug('No match for project/version: %s', path)
                    else:
                        name, version, pyver = t
                        if not project_name or same_project(project_name, name):
                            result = {
                                'name': name,
                                'version': version,
                                'filename': filename,
                                'url': urlunparse((scheme, netloc, origpath,
                                                   params, query, '')),
                                #'packagetype': 'sdist',
                            }
                            if pyver:  # pragma: no cover
                                result['python-version'] = pyver
                    break
        if result and algo:
            result['%s_digest' % algo] = digest
        return result

    def _get_digest(self, info):
        """
        Get a digest from a dictionary by looking at a "digests" dictionary
        or keys of the form 'algo_digest'.

        Returns a 2-tuple (algo, digest) if found, else None. Currently
        looks only for SHA256, then MD5.
        """
        result = None
        if 'digests' in info:
            digests = info['digests']
            for algo in ('sha256', 'md5'):
                if algo in digests:
                    result = (algo, digests[algo])
                    break
        if not result:
            for algo in ('sha256', 'md5'):
                key = '%s_digest' % algo
                if key in info:
                    result = (algo, info[key])
                    break
        return result

    def _update_version_data(self, result, info):
        """
        Update a result dictionary (the final result from _get_project) with a
        dictionary for a specific version, which typically holds information
        gleaned from a filename or URL for an archive for the distribution.
        """
        name = info.pop('name')
        version = info.pop('version')
        if version in result:
            dist = result[version]
            md = dist.metadata
        else:
            dist = make_dist(name, version, scheme=self.scheme)
            md = dist.metadata
        dist.digest = digest = self._get_digest(info)
        url = info['url']
        result['digests'][url] = digest
        if md.source_url != info['url']:
            md.source_url = self.prefer_url(md.source_url, url)
            result['urls'].setdefault(version, set()).add(url)
        dist.locator = self
        result[version] = dist

    def locate(self, requirement, prereleases=False):
        """
        Find the most recent distribution which matches the given
        requirement.

        :param requirement: A requirement of the form 'foo (1.0)' or perhaps
                            'foo (>= 1.0, < 2.0, != 1.3)'
        :param prereleases: If ``True``, allow pre-release versions
                            to be located. Otherwise, pre-release versions
                            are not returned.
        :return: A :class:`Distribution` instance, or ``None`` if no such
                 distribution could be located.
        """
        result = None
        r = parse_requirement(requirement)
        if r is None:  # pragma: no cover
            raise DistlibException('Not a valid requirement: %r' % requirement)
        scheme = get_scheme(self.scheme)
        self.matcher = matcher = scheme.matcher(r.requirement)
        logger.debug('matcher: %s (%s)', matcher, type(matcher).__name__)
        versions = self.get_project(r.name)
        if len(versions) > 2:   # urls and digests keys are present
            # sometimes, versions are invalid
            slist = []
            vcls = matcher.version_class
            for k in versions:
                if k in ('urls', 'digests'):
                    continue
                try:
                    if not matcher.match(k):
                        pass  # logger.debug('%s did not match %r', matcher, k)
                    else:
                        if prereleases or not vcls(k).is_prerelease:
                            slist.append(k)
                        # else:
                            # logger.debug('skipping pre-release '
                                         # 'version %s of %s', k, matcher.name)
                except Exception:  # pragma: no cover
                    logger.warning('error matching %s with %r', matcher, k)
                    pass # slist.append(k)
            if len(slist) > 1:
                slist = sorted(slist, key=scheme.key)
            if slist:
                logger.debug('sorted list: %s', slist)
                version = slist[-1]
                result = versions[version]
        if result:
            if r.extras:
                result.extras = r.extras
            result.download_urls = versions.get('urls', {}).get(version, set())
            d = {}
            sd = versions.get('digests', {})
            for url in result.download_urls:
                if url in sd:  # pragma: no cover
                    d[url] = sd[url]
            result.digests = d
        self.matcher = None
        return result


class PyPIRPCLocator(Locator):
    """
    This locator uses XML-RPC to locate distributions. It therefore
    cannot be used with simple mirrors (that only mirror file content).
    """
    def __init__(self, url, **kwargs):
        """
        Initialise an instance.

        :param url: The URL to use for XML-RPC.
        :param kwargs: Passed to the superclass constructor.
        """
        super(PyPIRPCLocator, self).__init__(**kwargs)
        self.base_url = url
        self.client = ServerProxy(url, timeout=3.0)

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        return set(self.client.list_packages())

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        versions = self.client.package_releases(name, True)
        for v in versions:
            urls = self.client.release_urls(name, v)
            data = self.client.release_data(name, v)
            metadata = Metadata(scheme=self.scheme)
            metadata.name = data['name']
            metadata.version = data['version']
            metadata.license = data.get('license')
            metadata.keywords = data.get('keywords', [])
            metadata.summary = data.get('summary')
            dist = Distribution(metadata)
            if urls:
                info = urls[0]
                metadata.source_url = info['url']
                dist.digest = self._get_digest(info)
                dist.locator = self
                result[v] = dist
                for info in urls:
                    url = info['url']
                    digest = self._get_digest(info)
                    result['urls'].setdefault(v, set()).add(url)
                    result['digests'][url] = digest
        return result

class PyPIJSONLocator(Locator):
    """
    This locator uses PyPI's JSON interface. It's very limited in functionality
    and probably not worth using.
    """
    def __init__(self, url, **kwargs):
        super(PyPIJSONLocator, self).__init__(**kwargs)
        self.base_url = ensure_slash(url)

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Not available from this locator')

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        url = urljoin(self.base_url, '%s/json' % quote(name))
        try:
            resp = self.opener.open(url)
            data = resp.read().decode() # for now
            d = json.loads(data)
            md = Metadata(scheme=self.scheme)
            data = d['info']
            md.name = data['name']
            md.version = data['version']
            md.license = data.get('license')
            md.keywords = data.get('keywords', [])
            md.summary = data.get('summary')
            dist = Distribution(md)
            dist.locator = self
            urls = d['urls']
            result[md.version] = dist
            for info in d['urls']:
                url = info['url']
                dist.download_urls.add(url)
                dist.digests[url] = self._get_digest(info)
                result['urls'].setdefault(md.version, set()).add(url)
                result['digests'][url] = self._get_digest(info)
            # Now get other releases
            for version, infos in d['releases'].items():
                if version == md.version:
                    continue    # already done
                omd = Metadata(scheme=self.scheme)
                omd.name = md.name
                omd.version = version
                odist = Distribution(omd)
                odist.locator = self
                result[version] = odist
                for info in infos:
                    url = info['url']
                    odist.download_urls.add(url)
                    odist.digests[url] = self._get_digest(info)
                    result['urls'].setdefault(version, set()).add(url)
                    result['digests'][url] = self._get_digest(info)
#            for info in urls:
#                md.source_url = info['url']
#                dist.digest = self._get_digest(info)
#                dist.locator = self
#                for info in urls:
#                    url = info['url']
#                    result['urls'].setdefault(md.version, set()).add(url)
#                    result['digests'][url] = self._get_digest(info)
        except Exception as e:
            self.errors.put(text_type(e))
            logger.exception('JSON fetch failed: %s', e)
        return result


class Page(object):
    """
    This class represents a scraped HTML page.
    """
    # The following slightly hairy-looking regex just looks for the contents of
    # an anchor link, which has an attribute "href" either immediately preceded
    # or immediately followed by a "rel" attribute. The attribute values can be
    # declared with double quotes, single quotes or no quotes - which leads to
    # the length of the expression.
    _href = re.compile("""
(rel\\s*=\\s*(?:"(?P<rel1>[^"]*)"|'(?P<rel2>[^']*)'|(?P<rel3>[^>\\s\n]*))\\s+)?
href\\s*=\\s*(?:"(?P<url1>[^"]*)"|'(?P<url2>[^']*)'|(?P<url3>[^>\\s\n]*))
(\\s+rel\\s*=\\s*(?:"(?P<rel4>[^"]*)"|'(?P<rel5>[^']*)'|(?P<rel6>[^>\\s\n]*)))?
""", re.I | re.S | re.X)
    _base = re.compile(r"""<base\s+href\s*=\s*['"]?([^'">]+)""", re.I | re.S)

    def __init__(self, data, url):
        """
        Initialise an instance with the Unicode page contents and the URL they
        came from.
        """
        self.data = data
        self.base_url = self.url = url
        m = self._base.search(self.data)
        if m:
            self.base_url = m.group(1)

    _clean_re = re.compile(r'[^a-z0-9$&+,/:;=?@.#%_\\|-]', re.I)

    @cached_property
    def links(self):
        """
        Return the URLs of all the links on a page together with information
        about their "rel" attribute, for determining which ones to treat as
        downloads and which ones to queue for further scraping.
        """
        def clean(url):
            "Tidy up an URL."
            scheme, netloc, path, params, query, frag = urlparse(url)
            return urlunparse((scheme, netloc, quote(path),
                               params, query, frag))

        result = set()
        for match in self._href.finditer(self.data):
            d = match.groupdict('')
            rel = (d['rel1'] or d['rel2'] or d['rel3'] or
                   d['rel4'] or d['rel5'] or d['rel6'])
            url = d['url1'] or d['url2'] or d['url3']
            url = urljoin(self.base_url, url)
            url = unescape(url)
            url = self._clean_re.sub(lambda m: '%%%2x' % ord(m.group(0)), url)
            result.add((url, rel))
        # We sort the result, hoping to bring the most recent versions
        # to the front
        result = sorted(result, key=lambda t: t[0], reverse=True)
        return result


class SimpleScrapingLocator(Locator):
    """
    A locator which scrapes HTML pages to locate downloads for a distribution.
    This runs multiple threads to do the I/O; performance is at least as good
    as pip's PackageFinder, which works in an analogous fashion.
    """

    # These are used to deal with various Content-Encoding schemes.
    decoders = {
        'deflate': zlib.decompress,
        'gzip': lambda b: gzip.GzipFile(fileobj=BytesIO(b)).read(),
        'none': lambda b: b,
    }

    def __init__(self, url, timeout=None, num_workers=10, **kwargs):
        """
        Initialise an instance.
        :param url: The root URL to use for scraping.
        :param timeout: The timeout, in seconds, to be applied to requests.
                        This defaults to ``None`` (no timeout specified).
        :param num_workers: The number of worker threads you want to do I/O,
                            This defaults to 10.
        :param kwargs: Passed to the superclass.
        """
        super(SimpleScrapingLocator, self).__init__(**kwargs)
        self.base_url = ensure_slash(url)
        self.timeout = timeout
        self._page_cache = {}
        self._seen = set()
        self._to_fetch = queue.Queue()
        self._bad_hosts = set()
        self.skip_externals = False
        self.num_workers = num_workers
        self._lock = threading.RLock()
        # See issue #45: we need to be resilient when the locator is used
        # in a thread, e.g. with concurrent.futures. We can't use self._lock
        # as it is for coordinating our internal threads - the ones created
        # in _prepare_threads.
        self._gplock = threading.RLock()
        self.platform_check = False  # See issue #112

    def _prepare_threads(self):
        """
        Threads are created only when get_project is called, and terminate
        before it returns. They are there primarily to parallelise I/O (i.e.
        fetching web pages).
        """
        self._threads = []
        for i in range(self.num_workers):
            t = threading.Thread(target=self._fetch)
            t.setDaemon(True)
            t.start()
            self._threads.append(t)

    def _wait_threads(self):
        """
        Tell all the threads to terminate (by sending a sentinel value) and
        wait for them to do so.
        """
        # Note that you need two loops, since you can't say which
        # thread will get each sentinel
        for t in self._threads:
            self._to_fetch.put(None)    # sentinel
        for t in self._threads:
            t.join()
        self._threads = []

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        with self._gplock:
            self.result = result
            self.project_name = name
            url = urljoin(self.base_url, '%s/' % quote(name))
            self._seen.clear()
            self._page_cache.clear()
            self._prepare_threads()
            try:
                logger.debug('Queueing %s', url)
                self._to_fetch.put(url)
                self._to_fetch.join()
            finally:
                self._wait_threads()
            del self.result
        return result

    platform_dependent = re.compile(r'\b(linux_(i\d86|x86_64|arm\w+)|'
                                    r'win(32|_amd64)|macosx_?\d+)\b', re.I)

    def _is_platform_dependent(self, url):
        """
        Does an URL refer to a platform-specific download?
        """
        return self.platform_dependent.search(url)

    def _process_download(self, url):
        """
        See if an URL is a suitable download for a project.

        If it is, register information in the result dictionary (for
        _get_project) about the specific version it's for.

        Note that the return value isn't actually used other than as a boolean
        value.
        """
        if self.platform_check and self._is_platform_dependent(url):
            info = None
        else:
            info = self.convert_url_to_download_info(url, self.project_name)
        logger.debug('process_download: %s -> %s', url, info)
        if info:
            with self._lock:    # needed because self.result is shared
                self._update_version_data(self.result, info)
        return info

    def _should_queue(self, link, referrer, rel):
        """
        Determine whether a link URL from a referring page and with a
        particular "rel" attribute should be queued for scraping.
        """
        scheme, netloc, path, _, _, _ = urlparse(link)
        if path.endswith(self.source_extensions + self.binary_extensions +
                         self.excluded_extensions):
            result = False
        elif self.skip_externals and not link.startswith(self.base_url):
            result = False
        elif not referrer.startswith(self.base_url):
            result = False
        elif rel not in ('homepage', 'download'):
            result = False
        elif scheme not in ('http', 'https', 'ftp'):
            result = False
        elif self._is_platform_dependent(link):
            result = False
        else:
            host = netloc.split(':', 1)[0]
            if host.lower() == 'localhost':
                result = False
            else:
                result = True
        logger.debug('should_queue: %s (%s) from %s -> %s', link, rel,
                     referrer, result)
        return result

    def _fetch(self):
        """
        Get a URL to fetch from the work queue, get the HTML page, examine its
        links for download candidates and candidates for further scraping.

        This is a handy method to run in a thread.
        """
        while True:
            url = self._to_fetch.get()
            try:
                if url:
                    page = self.get_page(url)
                    if page is None:    # e.g. after an error
                        continue
                    for link, rel in page.links:
                        if link not in self._seen:
                            try:
                                self._seen.add(link)
                                if (not self._process_download(link) and
                                    self._should_queue(link, url, rel)):
                                    logger.debug('Queueing %s from %s', link, url)
                                    self._to_fetch.put(link)
                            except MetadataInvalidError:  # e.g. invalid versions
                                pass
            except Exception as e:  # pragma: no cover
                self.errors.put(text_type(e))
            finally:
                # always do this, to avoid hangs :-)
                self._to_fetch.task_done()
            if not url:
                #logger.debug('Sentinel seen, quitting.')
                break

    def get_page(self, url):
        """
        Get the HTML for an URL, possibly from an in-memory cache.

        XXX TODO Note: this cache is never actually cleared. It's assumed that
        the data won't get stale over the lifetime of a locator instance (not
        necessarily true for the default_locator).
        """
        # http://peak.telecommunity.com/DevCenter/EasyInstall#package-index-api
        scheme, netloc, path, _, _, _ = urlparse(url)
        if scheme == 'file' and os.path.isdir(url2pathname(path)):
            url = urljoin(ensure_slash(url), 'index.html')

        if url in self._page_cache:
            result = self._page_cache[url]
            logger.debug('Returning %s from cache: %s', url, result)
        else:
            host = netloc.split(':', 1)[0]
            result = None
            if host in self._bad_hosts:
                logger.debug('Skipping %s due to bad host %s', url, host)
            else:
                req = Request(url, headers={'Accept-encoding': 'identity'})
                try:
                    logger.debug('Fetching %s', url)
                    resp = self.opener.open(req, timeout=self.timeout)
                    logger.debug('Fetched %s', url)
                    headers = resp.info()
                    content_type = headers.get('Content-Type', '')
                    if HTML_CONTENT_TYPE.match(content_type):
                        final_url = resp.geturl()
                        data = resp.read()
                        encoding = headers.get('Content-Encoding')
                        if encoding:
                            decoder = self.decoders[encoding]   # fail if not found
                            data = decoder(data)
                        encoding = 'utf-8'
                        m = CHARSET.search(content_type)
                        if m:
                            encoding = m.group(1)
                        try:
                            data = data.decode(encoding)
                        except UnicodeError:  # pragma: no cover
                            data = data.decode('latin-1')    # fallback
                        result = Page(data, final_url)
                        self._page_cache[final_url] = result
                except HTTPError as e:
                    if e.code != 404:
                        logger.exception('Fetch failed: %s: %s', url, e)
                except URLError as e:  # pragma: no cover
                    logger.exception('Fetch failed: %s: %s', url, e)
                    with self._lock:
                        self._bad_hosts.add(host)
                except Exception as e:  # pragma: no cover
                    logger.exception('Fetch failed: %s: %s', url, e)
                finally:
                    self._page_cache[url] = result   # even if None (failure)
        return result

    _distname_re = re.compile('<a href=[^>]*>([^<]+)<')

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        page = self.get_page(self.base_url)
        if not page:
            raise DistlibException('Unable to get %s' % self.base_url)
        for match in self._distname_re.finditer(page.data):
            result.add(match.group(1))
        return result

class DirectoryLocator(Locator):
    """
    This class locates distributions in a directory tree.
    """

    def __init__(self, path, **kwargs):
        """
        Initialise an instance.
        :param path: The root of the directory tree to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * recursive - if True (the default), subdirectories are
                         recursed into. If False, only the top-level directory
                         is searched,
        """
        self.recursive = kwargs.pop('recursive', True)
        super(DirectoryLocator, self).__init__(**kwargs)
        path = os.path.abspath(path)
        if not os.path.isdir(path):  # pragma: no cover
            raise DistlibException('Not a directory: %r' % path)
        self.base_dir = path

    def should_include(self, filename, parent):
        """
        Should a filename be considered as a candidate for a distribution
        archive? As well as the filename, the directory which contains it
        is provided, though not used by the current implementation.
        """
        return filename.endswith(self.downloadable_extensions)

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        for root, dirs, files in os.walk(self.base_dir):
            for fn in files:
                if self.should_include(fn, root):
                    fn = os.path.join(root, fn)
                    url = urlunparse(('file', '',
                                      pathname2url(os.path.abspath(fn)),
                                      '', '', ''))
                    info = self.convert_url_to_download_info(url, name)
                    if info:
                        self._update_version_data(result, info)
            if not self.recursive:
                break
        return result

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        for root, dirs, files in os.walk(self.base_dir):
            for fn in files:
                if self.should_include(fn, root):
                    fn = os.path.join(root, fn)
                    url = urlunparse(('file', '',
                                      pathname2url(os.path.abspath(fn)),
                                      '', '', ''))
                    info = self.convert_url_to_download_info(url, None)
                    if info:
                        result.add(info['name'])
            if not self.recursive:
                break
        return result

class JSONLocator(Locator):
    """
    This locator uses special extended metadata (not available on PyPI) and is
    the basis of performant dependency resolution in distlib. Other locators
    require archive downloads before dependencies can be determined! As you
    might imagine, that can be slow.
    """
    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        raise NotImplementedError('Not available from this locator')

    def _get_project(self, name):
        result = {'urls': {}, 'digests': {}}
        data = get_project_data(name)
        if data:
            for info in data.get('files', []):
                if info['ptype'] != 'sdist' or info['pyversion'] != 'source':
                    continue
                # We don't store summary in project metadata as it makes
                # the data bigger for no benefit during dependency
                # resolution
                dist = make_dist(data['name'], info['version'],
                                 summary=data.get('summary',
                                                  'Placeholder for summary'),
                                 scheme=self.scheme)
                md = dist.metadata
                md.source_url = info['url']
                # TODO SHA256 digest
                if 'digest' in info and info['digest']:
                    dist.digest = ('md5', info['digest'])
                md.dependencies = info.get('requirements', {})
                dist.exports = info.get('exports', {})
                result[dist.version] = dist
                result['urls'].setdefault(dist.version, set()).add(info['url'])
        return result

class DistPathLocator(Locator):
    """
    This locator finds installed distributions in a path. It can be useful for
    adding to an :class:`AggregatingLocator`.
    """
    def __init__(self, distpath, **kwargs):
        """
        Initialise an instance.

        :param distpath: A :class:`DistributionPath` instance to search.
        """
        super(DistPathLocator, self).__init__(**kwargs)
        assert isinstance(distpath, DistributionPath)
        self.distpath = distpath

    def _get_project(self, name):
        dist = self.distpath.get_distribution(name)
        if dist is None:
            result = {'urls': {}, 'digests': {}}
        else:
            result = {
                dist.version: dist,
                'urls': {dist.version: set([dist.source_url])},
                'digests': {dist.version: set([None])}
            }
        return result


class AggregatingLocator(Locator):
    """
    This class allows you to chain and/or merge a list of locators.
    """
    def __init__(self, *locators, **kwargs):
        """
        Initialise an instance.

        :param locators: The list of locators to search.
        :param kwargs: Passed to the superclass constructor,
                       except for:
                       * merge - if False (the default), the first successful
                         search from any of the locators is returned. If True,
                         the results from all locators are merged (this can be
                         slow).
        """
        self.merge = kwargs.pop('merge', False)
        self.locators = locators
        super(AggregatingLocator, self).__init__(**kwargs)

    def clear_cache(self):
        super(AggregatingLocator, self).clear_cache()
        for locator in self.locators:
            locator.clear_cache()

    def _set_scheme(self, value):
        self._scheme = value
        for locator in self.locators:
            locator.scheme = value

    scheme = property(Locator.scheme.fget, _set_scheme)

    def _get_project(self, name):
        result = {}
        for locator in self.locators:
            d = locator.get_project(name)
            if d:
                if self.merge:
                    files = result.get('urls', {})
                    digests = result.get('digests', {})
                    # next line could overwrite result['urls'], result['digests']
                    result.update(d)
                    df = result.get('urls')
                    if files and df:
                        for k, v in files.items():
                            if k in df:
                                df[k] |= v
                            else:
                                df[k] = v
                    dd = result.get('digests')
                    if digests and dd:
                        dd.update(digests)
                else:
                    # See issue #18. If any dists are found and we're looking
                    # for specific constraints, we only return something if
                    # a match is found. For example, if a DirectoryLocator
                    # returns just foo (1.0) while we're looking for
                    # foo (>= 2.0), we'll pretend there was nothing there so
                    # that subsequent locators can be queried. Otherwise we
                    # would just return foo (1.0) which would then lead to a
                    # failure to find foo (>= 2.0), because other locators
                    # weren't searched. Note that this only matters when
                    # merge=False.
                    if self.matcher is None:
                        found = True
                    else:
                        found = False
                        for k in d:
                            if self.matcher.match(k):
                                found = True
                                break
                    if found:
                        result = d
                        break
        return result

    def get_distribution_names(self):
        """
        Return all the distribution names known to this locator.
        """
        result = set()
        for locator in self.locators:
            try:
                result |= locator.get_distribution_names()
            except NotImplementedError:
                pass
        return result


# We use a legacy scheme simply because most of the dists on PyPI use legacy
# versions which don't conform to PEP 426 / PEP 440.
default_locator = AggregatingLocator(
                    JSONLocator(),
                    SimpleScrapingLocator('https://pypi.org/simple/',
                                          timeout=3.0),
                    scheme='legacy')

locate = default_locator.locate


class DependencyFinder(object):
    """
    Locate dependencies for distributions.
    """

    def __init__(self, locator=None):
        """
        Initialise an instance, using the specified locator
        to locate distributions.
        """
        self.locator = locator or default_locator
        self.scheme = get_scheme(self.locator.scheme)

    def add_distribution(self, dist):
        """
        Add a distribution to the finder. This will update internal information
        about who provides what.
        :param dist: The distribution to add.
        """
        logger.debug('adding distribution %s', dist)
        name = dist.key
        self.dists_by_name[name] = dist
        self.dists[(name, dist.version)] = dist
        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Add to provided: %s, %s, %s', name, version, dist)
            self.provided.setdefault(name, set()).add((version, dist))

    def remove_distribution(self, dist):
        """
        Remove a distribution from the finder. This will update internal
        information about who provides what.
        :param dist: The distribution to remove.
        """
        logger.debug('removing distribution %s', dist)
        name = dist.key
        del self.dists_by_name[name]
        del self.dists[(name, dist.version)]
        for p in dist.provides:
            name, version = parse_name_and_version(p)
            logger.debug('Remove from provided: %s, %s, %s', name, version, dist)
            s = self.provided[name]
            s.remove((version, dist))
            if not s:
                del self.provided[name]

    def get_matcher(self, reqt):
        """
        Get a version matcher for a requirement.
        :param reqt: The requirement
        :type reqt: str
        :return: A version matcher (an instance of
                 :class:`distlib.version.Matcher`).
        """
        try:
            matcher = self.scheme.matcher(reqt)
        except UnsupportedVersionError:  # pragma: no cover
            # XXX compat-mode if cannot read the version
            name = reqt.split()[0]
            matcher = self.scheme.matcher(name)
        return matcher

    def find_providers(self, reqt):
        """
        Find the distributions which can fulfill a requirement.

        :param reqt: The requirement.
         :type reqt: str
        :return: A set of distribution which can fulfill the requirement.
        """
        matcher = self.get_matcher(reqt)
        name = matcher.key   # case-insensitive
        result = set()
        provided = self.provided
        if name in provided:
            for version, provider in provided[name]:
                try:
                    match = matcher.match(version)
                except UnsupportedVersionError:
                    match = False

                if match:
                    result.add(provider)
                    break
        return result

    def try_to_replace(self, provider, other, problems):
        """
        Attempt to replace one provider with another. This is typically used
        when resolving dependencies from multiple sources, e.g. A requires
        (B >= 1.0) while C requires (B >= 1.1).

        For successful replacement, ``provider`` must meet all the requirements
        which ``other`` fulfills.

        :param provider: The provider we are trying to replace with.
        :param other: The provider we're trying to replace.
        :param problems: If False is returned, this will contain what
                         problems prevented replacement. This is currently
                         a tuple of the literal string 'cantreplace',
                         ``provider``, ``other``  and the set of requirements
                         that ``provider`` couldn't fulfill.
        :return: True if we can replace ``other`` with ``provider``, else
                 False.
        """
        rlist = self.reqts[other]
        unmatched = set()
        for s in rlist:
            matcher = self.get_matcher(s)
            if not matcher.match(provider.version):
                unmatched.add(s)
        if unmatched:
            # can't replace other with provider
            problems.add(('cantreplace', provider, other,
                          frozenset(unmatched)))
            result = False
        else:
            # can replace other with provider
            self.remove_distribution(other)
            del self.reqts[other]
            for s in rlist:
                self.reqts.setdefault(provider, set()).add(s)
            self.add_distribution(provider)
            result = True
        return result

    def find(self, requirement, meta_extras=None, prereleases=False):
        """
        Find a distribution and all distributions it depends on.

        :param requirement: The requirement specifying the distribution to
                            find, or a Distribution instance.
        :param meta_extras: A list of meta extras such as :test:, :build: and
                            so on.
        :param prereleases: If ``True``, allow pre-release versions to be
                            returned - otherwise, don't return prereleases
                            unless they're all that's available.

        Return a set of :class:`Distribution` instances and a set of
        problems.

        The distributions returned should be such that they have the
        :attr:`required` attribute set to ``True`` if they were
        from the ``requirement`` passed to ``find()``, and they have the
        :attr:`build_time_dependency` attribute set to ``True`` unless they
        are post-installation dependencies of the ``requirement``.

        The problems should be a tuple consisting of the string
        ``'unsatisfied'`` and the requirement which couldn't be satisfied
        by any distribution known to the locator.
        """

        self.provided = {}
        self.dists = {}
        self.dists_by_name = {}
        self.reqts = {}

        meta_extras = set(meta_extras or [])
        if ':*:' in meta_extras:
            meta_extras.remove(':*:')
            # :meta: and :run: are implicitly included
            meta_extras |= set([':test:', ':build:', ':dev:'])

        if isinstance(requirement, Distribution):
            dist = odist = requirement
            logger.debug('passed %s as requirement', odist)
        else:
            dist = odist = self.locator.locate(requirement,
                                               prereleases=prereleases)
            if dist is None:
                raise DistlibException('Unable to locate %r' % requirement)
            logger.debug('located %s', odist)
        dist.requested = True
        problems = set()
        todo = set([dist])
        install_dists = set([odist])
        while todo:
            dist = todo.pop()
            name = dist.key     # case-insensitive
            if name not in self.dists_by_name:
                self.add_distribution(dist)
            else:
                #import pdb; pdb.set_trace()
                other = self.dists_by_name[name]
                if other != dist:
                    self.try_to_replace(dist, other, problems)

            ireqts = dist.run_requires | dist.meta_requires
            sreqts = dist.build_requires
            ereqts = set()
            if meta_extras and dist in install_dists:
                for key in ('test', 'build', 'dev'):
                    e = ':%s:' % key
                    if e in meta_extras:
                        ereqts |= getattr(dist, '%s_requires' % key)
            all_reqts = ireqts | sreqts | ereqts
            for r in all_reqts:
                providers = self.find_providers(r)
                if not providers:
                    logger.debug('No providers found for %r', r)
                    provider = self.locator.locate(r, prereleases=prereleases)
                    # If no provider is found and we didn't consider
                    # prereleases, consider them now.
                    if provider is None and not prereleases:
                        provider = self.locator.locate(r, prereleases=True)
                    if provider is None:
                        logger.debug('Cannot satisfy %r', r)
                        problems.add(('unsatisfied', r))
                    else:
                        n, v = provider.key, provider.version
                        if (n, v) not in self.dists:
                            todo.add(provider)
                        providers.add(provider)
                        if r in ireqts and dist in install_dists:
                            install_dists.add(provider)
                            logger.debug('Adding %s to install_dists',
                                         provider.name_and_version)
                for p in providers:
                    name = p.key
                    if name not in self.dists_by_name:
                        self.reqts.setdefault(p, set()).add(r)
                    else:
                        other = self.dists_by_name[name]
                        if other != p:
                            # see if other can be replaced by p
                            self.try_to_replace(p, other, problems)

        dists = set(self.dists.values())
        for dist in dists:
            dist.build_time_dependency = dist not in install_dists
            if dist.build_time_dependency:
                logger.debug('%s is a build-time dependency only.',
                             dist.name_and_version)
        logger.debug('find done for %s', odist)
        return dists, problems
site-packages/pip/_vendor/distlib/resources.py000064400000025104150564745010015537 0ustar00# -*- coding: utf-8 -*-
#
# Copyright (C) 2013-2017 Vinay Sajip.
# Licensed to the Python Software Foundation under a contributor agreement.
# See LICENSE.txt and CONTRIBUTORS.txt.
#
from __future__ import unicode_literals

import bisect
import io
import logging
import os
import pkgutil
import sys
import types
import zipimport

from . import DistlibException
from .util import cached_property, get_cache_base, Cache

logger = logging.getLogger(__name__)


cache = None    # created when needed


class ResourceCache(Cache):
    def __init__(self, base=None):
        if base is None:
            # Use native string to avoid issues on 2.x: see Python #20140.
            base = os.path.join(get_cache_base(), str('resource-cache'))
        super(ResourceCache, self).__init__(base)

    def is_stale(self, resource, path):
        """
        Is the cache stale for the given resource?

        :param resource: The :class:`Resource` being cached.
        :param path: The path of the resource in the cache.
        :return: True if the cache is stale.
        """
        # Cache invalidation is a hard problem :-)
        return True

    def get(self, resource):
        """
        Get a resource into the cache,

        :param resource: A :class:`Resource` instance.
        :return: The pathname of the resource in the cache.
        """
        prefix, path = resource.finder.get_cache_info(resource)
        if prefix is None:
            result = path
        else:
            result = os.path.join(self.base, self.prefix_to_dir(prefix), path)
            dirname = os.path.dirname(result)
            if not os.path.isdir(dirname):
                os.makedirs(dirname)
            if not os.path.exists(result):
                stale = True
            else:
                stale = self.is_stale(resource, path)
            if stale:
                # write the bytes of the resource to the cache location
                with open(result, 'wb') as f:
                    f.write(resource.bytes)
        return result


class ResourceBase(object):
    def __init__(self, finder, name):
        self.finder = finder
        self.name = name


class Resource(ResourceBase):
    """
    A class representing an in-package resource, such as a data file. This is
    not normally instantiated by user code, but rather by a
    :class:`ResourceFinder` which manages the resource.
    """
    is_container = False        # Backwards compatibility

    def as_stream(self):
        """
        Get the resource as a stream.

        This is not a property to make it obvious that it returns a new stream
        each time.
        """
        return self.finder.get_stream(self)

    @cached_property
    def file_path(self):
        global cache
        if cache is None:
            cache = ResourceCache()
        return cache.get(self)

    @cached_property
    def bytes(self):
        return self.finder.get_bytes(self)

    @cached_property
    def size(self):
        return self.finder.get_size(self)


class ResourceContainer(ResourceBase):
    is_container = True     # Backwards compatibility

    @cached_property
    def resources(self):
        return self.finder.get_resources(self)


class ResourceFinder(object):
    """
    Resource finder for file system resources.
    """

    if sys.platform.startswith('java'):
        skipped_extensions = ('.pyc', '.pyo', '.class')
    else:
        skipped_extensions = ('.pyc', '.pyo')

    def __init__(self, module):
        self.module = module
        self.loader = getattr(module, '__loader__', None)
        self.base = os.path.dirname(getattr(module, '__file__', ''))

    def _adjust_path(self, path):
        return os.path.realpath(path)

    def _make_path(self, resource_name):
        # Issue #50: need to preserve type of path on Python 2.x
        # like os.path._get_sep
        if isinstance(resource_name, bytes):    # should only happen on 2.x
            sep = b'/'
        else:
            sep = '/'
        parts = resource_name.split(sep)
        parts.insert(0, self.base)
        result = os.path.join(*parts)
        return self._adjust_path(result)

    def _find(self, path):
        return os.path.exists(path)

    def get_cache_info(self, resource):
        return None, resource.path

    def find(self, resource_name):
        path = self._make_path(resource_name)
        if not self._find(path):
            result = None
        else:
            if self._is_directory(path):
                result = ResourceContainer(self, resource_name)
            else:
                result = Resource(self, resource_name)
            result.path = path
        return result

    def get_stream(self, resource):
        return open(resource.path, 'rb')

    def get_bytes(self, resource):
        with open(resource.path, 'rb') as f:
            return f.read()

    def get_size(self, resource):
        return os.path.getsize(resource.path)

    def get_resources(self, resource):
        def allowed(f):
            return (f != '__pycache__' and not
                    f.endswith(self.skipped_extensions))
        return set([f for f in os.listdir(resource.path) if allowed(f)])

    def is_container(self, resource):
        return self._is_directory(resource.path)

    _is_directory = staticmethod(os.path.isdir)

    def iterator(self, resource_name):
        resource = self.find(resource_name)
        if resource is not None:
            todo = [resource]
            while todo:
                resource = todo.pop(0)
                yield resource
                if resource.is_container:
                    rname = resource.name
                    for name in resource.resources:
                        if not rname:
                            new_name = name
                        else:
                            new_name = '/'.join([rname, name])
                        child = self.find(new_name)
                        if child.is_container:
                            todo.append(child)
                        else:
                            yield child


class ZipResourceFinder(ResourceFinder):
    """
    Resource finder for resources in .zip files.
    """
    def __init__(self, module):
        super(ZipResourceFinder, self).__init__(module)
        archive = self.loader.archive
        self.prefix_len = 1 + len(archive)
        # PyPy doesn't have a _files attr on zipimporter, and you can't set one
        if hasattr(self.loader, '_files'):
            self._files = self.loader._files
        else:
            self._files = zipimport._zip_directory_cache[archive]
        self.index = sorted(self._files)

    def _adjust_path(self, path):
        return path

    def _find(self, path):
        path = path[self.prefix_len:]
        if path in self._files:
            result = True
        else:
            if path and path[-1] != os.sep:
                path = path + os.sep
            i = bisect.bisect(self.index, path)
            try:
                result = self.index[i].startswith(path)
            except IndexError:
                result = False
        if not result:
            logger.debug('_find failed: %r %r', path, self.loader.prefix)
        else:
            logger.debug('_find worked: %r %r', path, self.loader.prefix)
        return result

    def get_cache_info(self, resource):
        prefix = self.loader.archive
        path = resource.path[1 + len(prefix):]
        return prefix, path

    def get_bytes(self, resource):
        return self.loader.get_data(resource.path)

    def get_stream(self, resource):
        return io.BytesIO(self.get_bytes(resource))

    def get_size(self, resource):
        path = resource.path[self.prefix_len:]
        return self._files[path][3]

    def get_resources(self, resource):
        path = resource.path[self.prefix_len:]
        if path and path[-1] != os.sep:
            path += os.sep
        plen = len(path)
        result = set()
        i = bisect.bisect(self.index, path)
        while i < len(self.index):
            if not self.index[i].startswith(path):
                break
            s = self.index[i][plen:]
            result.add(s.split(os.sep, 1)[0])   # only immediate children
            i += 1
        return result

    def _is_directory(self, path):
        path = path[self.prefix_len:]
        if path and path[-1] != os.sep:
            path += os.sep
        i = bisect.bisect(self.index, path)
        try:
            result = self.index[i].startswith(path)
        except IndexError:
            result = False
        return result


_finder_registry = {
    type(None): ResourceFinder,
    zipimport.zipimporter: ZipResourceFinder
}

try:
    # In Python 3.6, _frozen_importlib -> _frozen_importlib_external
    try:
        import _frozen_importlib_external as _fi
    except ImportError:
        import _frozen_importlib as _fi
    _finder_registry[_fi.SourceFileLoader] = ResourceFinder
    _finder_registry[_fi.FileFinder] = ResourceFinder
    # See issue #146
    _finder_registry[_fi.SourcelessFileLoader] = ResourceFinder
    del _fi
except (ImportError, AttributeError):
    pass


def register_finder(loader, finder_maker):
    _finder_registry[type(loader)] = finder_maker


_finder_cache = {}


def finder(package):
    """
    Return a resource finder for a package.
    :param package: The name of the package.
    :return: A :class:`ResourceFinder` instance for the package.
    """
    if package in _finder_cache:
        result = _finder_cache[package]
    else:
        if package not in sys.modules:
            __import__(package)
        module = sys.modules[package]
        path = getattr(module, '__path__', None)
        if path is None:
            raise DistlibException('You cannot get a finder for a module, '
                                   'only for a package')
        loader = getattr(module, '__loader__', None)
        finder_maker = _finder_registry.get(type(loader))
        if finder_maker is None:
            raise DistlibException('Unable to locate finder for %r' % package)
        result = finder_maker(module)
        _finder_cache[package] = result
    return result


_dummy_module = types.ModuleType(str('__dummy__'))


def finder_for_path(path):
    """
    Return a resource finder for a path, which should represent a container.

    :param path: The path.
    :return: A :class:`ResourceFinder` instance for the path.
    """
    result = None
    # calls any path hooks, gets importer into cache
    pkgutil.get_importer(path)
    loader = sys.path_importer_cache.get(path)
    finder = _finder_registry.get(type(loader))
    if finder:
        module = _dummy_module
        module.__file__ = os.path.join(path, '')
        module.__loader__ = loader
        result = finder(module)
    return result
site-packages/pip/_vendor/resolvelib/providers.py000064400000013360150564745010016257 0ustar00class AbstractProvider(object):
    """Delegate class to provide requirement interface for the resolver."""

    def identify(self, requirement_or_candidate):
        """Given a requirement, return an identifier for it.

        This is used to identify a requirement, e.g. whether two requirements
        should have their specifier parts merged.
        """
        raise NotImplementedError

    def get_preference(
        self,
        identifier,
        resolutions,
        candidates,
        information,
        backtrack_causes,
    ):
        """Produce a sort key for given requirement based on preference.

        The preference is defined as "I think this requirement should be
        resolved first". The lower the return value is, the more preferred
        this group of arguments is.

        :param identifier: An identifier as returned by ``identify()``. This
            identifies the dependency matches of which should be returned.
        :param resolutions: Mapping of candidates currently pinned by the
            resolver. Each key is an identifier, and the value a candidate.
            The candidate may conflict with requirements from ``information``.
        :param candidates: Mapping of each dependency's possible candidates.
            Each value is an iterator of candidates.
        :param information: Mapping of requirement information of each package.
            Each value is an iterator of *requirement information*.
        :param backtrack_causes: Sequence of requirement information that were
            the requirements that caused the resolver to most recently backtrack.

        A *requirement information* instance is a named tuple with two members:

        * ``requirement`` specifies a requirement contributing to the current
          list of candidates.
        * ``parent`` specifies the candidate that provides (dependend on) the
          requirement, or ``None`` to indicate a root requirement.

        The preference could depend on a various of issues, including (not
        necessarily in this order):

        * Is this package pinned in the current resolution result?
        * How relaxed is the requirement? Stricter ones should probably be
          worked on first? (I don't know, actually.)
        * How many possibilities are there to satisfy this requirement? Those
          with few left should likely be worked on first, I guess?
        * Are there any known conflicts for this requirement? We should
          probably work on those with the most known conflicts.

        A sortable value should be returned (this will be used as the ``key``
        parameter of the built-in sorting function). The smaller the value is,
        the more preferred this requirement is (i.e. the sorting function
        is called with ``reverse=False``).
        """
        raise NotImplementedError

    def find_matches(self, identifier, requirements, incompatibilities):
        """Find all possible candidates that satisfy given constraints.

        :param identifier: An identifier as returned by ``identify()``. This
            identifies the dependency matches of which should be returned.
        :param requirements: A mapping of requirements that all returned
            candidates must satisfy. Each key is an identifier, and the value
            an iterator of requirements for that dependency.
        :param incompatibilities: A mapping of known incompatibilities of
            each dependency. Each key is an identifier, and the value an
            iterator of incompatibilities known to the resolver. All
            incompatibilities *must* be excluded from the return value.

        This should try to get candidates based on the requirements' types.
        For VCS, local, and archive requirements, the one-and-only match is
        returned, and for a "named" requirement, the index(es) should be
        consulted to find concrete candidates for this requirement.

        The return value should produce candidates ordered by preference; the
        most preferred candidate should come first. The return type may be one
        of the following:

        * A callable that returns an iterator that yields candidates.
        * An collection of candidates.
        * An iterable of candidates. This will be consumed immediately into a
          list of candidates.
        """
        raise NotImplementedError

    def is_satisfied_by(self, requirement, candidate):
        """Whether the given requirement can be satisfied by a candidate.

        The candidate is guarenteed to have been generated from the
        requirement.

        A boolean should be returned to indicate whether ``candidate`` is a
        viable solution to the requirement.
        """
        raise NotImplementedError

    def get_dependencies(self, candidate):
        """Get dependencies of a candidate.

        This should return a collection of requirements that `candidate`
        specifies as its dependencies.
        """
        raise NotImplementedError


class AbstractResolver(object):
    """The thing that performs the actual resolution work."""

    base_exception = Exception

    def __init__(self, provider, reporter):
        self.provider = provider
        self.reporter = reporter

    def resolve(self, requirements, **kwargs):
        """Take a collection of constraints, spit out the resolution result.

        This returns a representation of the final resolution state, with one
        guarenteed attribute ``mapping`` that contains resolved candidates as
        values. The keys are their respective identifiers.

        :param requirements: A collection of constraints.
        :param kwargs: Additional keyword arguments that subclasses may accept.

        :raises: ``self.base_exception`` or its subclass.
        """
        raise NotImplementedError
site-packages/pip/_vendor/resolvelib/__pycache__/structs.cpython-39.pyc000064400000016226150564745010022244 0ustar00a

��Re��@sbddlZddlmZGdd�de�ZGdd�dej�ZGdd	�d	e�ZGd
d�de�Zdd
�Z	dS)�N�)�collections_abcc@speZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zdd�ZdS)�
DirectedGraphz&A graph structure with directed edges.cCst�|_i|_i|_dS�N)�set�	_vertices�	_forwards�
_backwards��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/structs.py�__init__	szDirectedGraph.__init__cCs
t|j�Sr)�iterrr
rrr
�__iter__szDirectedGraph.__iter__cCs
t|j�Sr)�lenrr
rrr
�__len__szDirectedGraph.__len__cCs
||jvSr)r�r�keyrrr
�__contains__szDirectedGraph.__contains__cCsBt�}t|j�|_dd�|j��D�|_dd�|j��D�|_|S)z$Return a shallow copy of this graph.cSsi|]\}}|t|��qSr�r��.0�k�vrrr
�
<dictcomp>�z&DirectedGraph.copy.<locals>.<dictcomp>cSsi|]\}}|t|��qSrrrrrr
rr)rrrr�itemsr	)r�otherrrr
�copys
zDirectedGraph.copycCs:||jvrtd��|j�|�t�|j|<t�|j|<dS)zAdd a new vertex to the graph.z
vertex existsN)r�
ValueError�addrrr	rrrr
r!s

zDirectedGraph.addcCsT|j�|�|j�|�D]}|j|�|�q|j�|�D]}|j|�|�q:dS)zCRemove a vertex from the graph, disconnecting all edges from/to it.N)r�remover�popr	)rr�f�trrr
r"'s
zDirectedGraph.removecCs||j|vo||j|vSr)r	r�rr$r%rrr
�	connected/szDirectedGraph.connectedcCs6||jvrt|��|j|�|�|j|�|�dS)zgConnect two existing vertices.

        Nothing happens if the vertices are already connected.
        N)r�KeyErrorrr!r	r&rrr
�connect2s
zDirectedGraph.connectccs,|j��D]\}}|D]}||fVqq
dSr)rr)rr$�childrenr%rrr
�
iter_edges<szDirectedGraph.iter_edgescCst|j|�Sr)rrrrrr
�
iter_childrenAszDirectedGraph.iter_childrencCst|j|�Sr)rr	rrrr
�iter_parentsDszDirectedGraph.iter_parentsN)�__name__�
__module__�__qualname__�__doc__rrrrrr!r"r'r)r+r,r-rrrr
rs
rc@sJeZdZddd�Zdd�Zdd�ZeZdd	�Zd
d�Zdd
�Z	dd�Z
dS)�IteratorMappingNcCs||_||_|pi|_dSr)�_mapping�	_accessor�_appends)r�mapping�accessor�appendsrrr
rIszIteratorMapping.__init__cCsd�|j|j|j�S)Nz!IteratorMapping({!r}, {!r}, {!r}))�formatr3r4r5r
rrr
�__repr__Ns
�zIteratorMapping.__repr__cCst|jp|j�Sr)�boolr3r5r
rrr
�__bool__UszIteratorMapping.__bool__cCs||jvp||jvSr)r3r5rrrr
rZszIteratorMapping.__contains__cCsLz|j|}Wn ty.t|j|�YS0t�|�|�|j�|d��S)Nr)r3r(rr5�	itertools�chainr4�get)rrrrrr
�__getitem__]s
zIteratorMapping.__getitem__cs"�fdd��jD�}t��j|�S)Nc3s|]}|�jvr|VqdSr�r3�rrr
rr
�	<genexpr>erz+IteratorMapping.__iter__.<locals>.<genexpr>)r5r=r>r3�r�morerr
r
rdszIteratorMapping.__iter__cs&t�fdd��jD��}t�j�|S)Nc3s|]}|�jvrdVqdS)rNrArBr
rr
rCirz*IteratorMapping.__len__.<locals>.<genexpr>)�sumr5rr3rDrr
r
rhszIteratorMapping.__len__)N)r.r/r0rr:r<�__nonzero__rr@rrrrrr
r2Hs
r2c@s4eZdZdZdd�Zdd�Zdd�ZeZdd	�Zd
S)�_FactoryIterableViewa:Wrap an iterator factory returned by `find_matches()`.

    Calling `iter()` on this class would invoke the underlying iterator
    factory, making it a "collection with ordering" that can be iterated
    through multiple times, but lacks random access methods presented in
    built-in Python sequence types.
    cCs
||_dSr��_factory)r�factoryrrr
rvsz_FactoryIterableView.__init__cCsd�t|�jt|����S�Nz{}({}))r9�typer.�listrJr
rrr
r:ysz_FactoryIterableView.__repr__cCs*zt|���Wnty$YdS0dS)NFT)�nextrJ�
StopIterationr
rrr
r<|s
z_FactoryIterableView.__bool__cCs|��SrrIr
rrr
r�sz_FactoryIterableView.__iter__N�	r.r/r0r1rr:r<rGrrrrr
rHmsrHc@s4eZdZdZdd�Zdd�Zdd�ZeZdd	�Zd
S)�_SequenceIterableViewz�Wrap an iterable returned by find_matches().

    This is essentially just a proxy to the underlying sequence that provides
    the same interface as `_FactoryIterableView`.
    cCs
||_dSr)�	_sequence)r�sequencerrr
r�sz_SequenceIterableView.__init__cCsd�t|�j|j�SrL)r9rMr.rSr
rrr
r:�sz_SequenceIterableView.__repr__cCs
t|j�Sr)r;rSr
rrr
r<�sz_SequenceIterableView.__bool__cCs
t|j�Sr)rrSr
rrr
r�sz_SequenceIterableView.__iter__NrQrrrr
rR�srRcCs,t|�rt|�St|tj�s$t|�}t|�S)zCBuild an iterable view from the value returned by `find_matches()`.)�callablerH�
isinstancer�SequencerNrR)�matchesrrr
�build_iter_view�s
rY)
r=�compatr�objectr�Mappingr2rHrRrYrrrr
�<module>sB%site-packages/pip/_vendor/resolvelib/__pycache__/reporters.cpython-39.pyc000064400000004445150564745010022562 0ustar00a

��ReT�@sGdd�de�ZdS)c@sHeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dS)�BaseReporterz?Delegate class to provider progress reporting for the resolver.cCsdS)z-Called before the resolution actually starts.N�)�selfrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/reporters.py�startingszBaseReporter.startingcCsdS)zYCalled before each round of resolution starts.

        The index is zero-based.
        Nr)r�indexrrr�starting_roundszBaseReporter.starting_roundcCsdS)z�Called before each round of resolution ends.

        This is NOT called if the resolution ends at this round. Use `ending`
        if you want to report finalization. The index is zero-based.
        Nr)rr�staterrr�ending_round
szBaseReporter.ending_roundcCsdS)z/Called before the resolution ends successfully.Nr)rrrrr�endingszBaseReporter.endingcCsdS)a�Called when adding a new requirement into the resolve criteria.

        :param requirement: The additional requirement to be applied to filter
            the available candidaites.
        :param parent: The candidate that requires ``requirement`` as a
            dependency, or None if ``requirement`` is one of the root
            requirements passed in from ``Resolver.resolve()``.
        Nr)r�requirement�parentrrr�adding_requirementszBaseReporter.adding_requirementcCsdS)z6Called when rejecting a candidate during backtracking.Nr�r�	candidaterrr�backtracking!szBaseReporter.backtrackingcCsdS)z9Called when adding a candidate to the potential solution.Nrrrrr�pinning$szBaseReporter.pinningN)�__name__�
__module__�__qualname__�__doc__rrr	r
r
rrrrrrrs
rN)�objectrrrrr�<module>�site-packages/pip/_vendor/resolvelib/__pycache__/providers.cpython-39.pyc000064400000015130150564745010022543 0ustar00a

��Re��@s$Gdd�de�ZGdd�de�ZdS)c@s8eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�ZdS)
�AbstractProviderzADelegate class to provide requirement interface for the resolver.cCst�dS)z�Given a requirement, return an identifier for it.

        This is used to identify a requirement, e.g. whether two requirements
        should have their specifier parts merged.
        N��NotImplementedError)�self�requirement_or_candidate�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/providers.py�identifyszAbstractProvider.identifycCst�dS)a�Produce a sort key for given requirement based on preference.

        The preference is defined as "I think this requirement should be
        resolved first". The lower the return value is, the more preferred
        this group of arguments is.

        :param identifier: An identifier as returned by ``identify()``. This
            identifies the dependency matches of which should be returned.
        :param resolutions: Mapping of candidates currently pinned by the
            resolver. Each key is an identifier, and the value a candidate.
            The candidate may conflict with requirements from ``information``.
        :param candidates: Mapping of each dependency's possible candidates.
            Each value is an iterator of candidates.
        :param information: Mapping of requirement information of each package.
            Each value is an iterator of *requirement information*.
        :param backtrack_causes: Sequence of requirement information that were
            the requirements that caused the resolver to most recently backtrack.

        A *requirement information* instance is a named tuple with two members:

        * ``requirement`` specifies a requirement contributing to the current
          list of candidates.
        * ``parent`` specifies the candidate that provides (dependend on) the
          requirement, or ``None`` to indicate a root requirement.

        The preference could depend on a various of issues, including (not
        necessarily in this order):

        * Is this package pinned in the current resolution result?
        * How relaxed is the requirement? Stricter ones should probably be
          worked on first? (I don't know, actually.)
        * How many possibilities are there to satisfy this requirement? Those
          with few left should likely be worked on first, I guess?
        * Are there any known conflicts for this requirement? We should
          probably work on those with the most known conflicts.

        A sortable value should be returned (this will be used as the ``key``
        parameter of the built-in sorting function). The smaller the value is,
        the more preferred this requirement is (i.e. the sorting function
        is called with ``reverse=False``).
        Nr)r�
identifier�resolutions�
candidates�information�backtrack_causesrrr�get_preferences1zAbstractProvider.get_preferencecCst�dS)a�Find all possible candidates that satisfy given constraints.

        :param identifier: An identifier as returned by ``identify()``. This
            identifies the dependency matches of which should be returned.
        :param requirements: A mapping of requirements that all returned
            candidates must satisfy. Each key is an identifier, and the value
            an iterator of requirements for that dependency.
        :param incompatibilities: A mapping of known incompatibilities of
            each dependency. Each key is an identifier, and the value an
            iterator of incompatibilities known to the resolver. All
            incompatibilities *must* be excluded from the return value.

        This should try to get candidates based on the requirements' types.
        For VCS, local, and archive requirements, the one-and-only match is
        returned, and for a "named" requirement, the index(es) should be
        consulted to find concrete candidates for this requirement.

        The return value should produce candidates ordered by preference; the
        most preferred candidate should come first. The return type may be one
        of the following:

        * A callable that returns an iterator that yields candidates.
        * An collection of candidates.
        * An iterable of candidates. This will be consumed immediately into a
          list of candidates.
        Nr)rr	�requirements�incompatibilitiesrrr�find_matches?szAbstractProvider.find_matchescCst�dS)aWhether the given requirement can be satisfied by a candidate.

        The candidate is guarenteed to have been generated from the
        requirement.

        A boolean should be returned to indicate whether ``candidate`` is a
        viable solution to the requirement.
        Nr)r�requirement�	candidaterrr�is_satisfied_by\s	z AbstractProvider.is_satisfied_bycCst�dS)z�Get dependencies of a candidate.

        This should return a collection of requirements that `candidate`
        specifies as its dependencies.
        Nr)rrrrr�get_dependenciesgsz!AbstractProvider.get_dependenciesN)	�__name__�
__module__�__qualname__�__doc__rrrrrrrrrrs3rc@s$eZdZdZeZdd�Zdd�ZdS)�AbstractResolverz3The thing that performs the actual resolution work.cCs||_||_dS)N)�provider�reporter)rrrrrr�__init__uszAbstractResolver.__init__cKst�dS)a�Take a collection of constraints, spit out the resolution result.

        This returns a representation of the final resolution state, with one
        guarenteed attribute ``mapping`` that contains resolved candidates as
        values. The keys are their respective identifiers.

        :param requirements: A collection of constraints.
        :param kwargs: Additional keyword arguments that subclasses may accept.

        :raises: ``self.base_exception`` or its subclass.
        Nr)rr�kwargsrrr�resolveyszAbstractResolver.resolveN)rrrr�	Exception�base_exceptionrrrrrrrpsrN)�objectrrrrrr�<module>sosite-packages/pip/_vendor/resolvelib/__pycache__/resolvers.cpython-39.pyc000064400000035702150564745010022561 0ustar00a

��Re�D�@s�ddlZddlZddlmZddlmZmZmZe�dddg�Z	Gdd	�d	e
�ZGd
d�de�ZGdd
�d
e�Z
Gdd�de�ZGdd�de�ZGdd�de�ZGdd�de�Ze�dd�ZGdd�de�Zdd�Ze�dd�Zdd�ZGd d!�d!e�ZdS)"�N�)�AbstractResolver)�
DirectedGraph�IteratorMapping�build_iter_view�RequirementInformation�requirement�parentc@seZdZdZdS)�ResolverExceptionz�A base class for all exceptions raised by this module.

    Exceptions derived by this class should all be handled in this module. Any
    bubbling pass the resolver should be treated as a bug.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/resolvers.pyr

sr
cs$eZdZ�fdd�Zdd�Z�ZS)�RequirementsConflictedcstt|��|�||_dS�N)�superr�__init__�	criterion)�selfr��	__class__rrrszRequirementsConflicted.__init__cCs d�d�dd�|j��D���S)NzRequirements conflict: {}�, css|]}t|�VqdSr��repr��.0�rrrr�	<genexpr>�z1RequirementsConflicted.__str__.<locals>.<genexpr>)�format�joinr�iter_requirement�rrrr�__str__s�zRequirementsConflicted.__str__�rrr
rr%�
__classcell__rrrrrsrcs$eZdZ�fdd�Zdd�Z�ZS)�InconsistentCandidatecs"tt|��||�||_||_dSr)rr(r�	candidater)rr)rrrrr!szInconsistentCandidate.__init__cCs$d�|jd�dd�|j��D���S)Nz+Provided candidate {!r} does not satisfy {}rcss|]}t|�VqdSrrrrrrr)r z0InconsistentCandidate.__str__.<locals>.<genexpr>)r!r)r"rr#r$rrrr%&s�zInconsistentCandidate.__str__r&rrrrr( sr(c@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)�	CriterionaQRepresentation of possible resolution results of a package.

    This holds three attributes:

    * `information` is a collection of `RequirementInformation` pairs.
      Each pair is a requirement contributing to this criterion, and the
      candidate that provides the requirement.
    * `incompatibilities` is a collection of all known not-to-work candidates
      to exclude from consideration.
    * `candidates` is a collection containing all possible candidates deducted
      from the union of contributing requirements and known incompatibilities.
      It should never be empty, except when the criterion is an attribute of a
      raised `RequirementsConflicted` (in which case it is always empty).

    .. note::
        This class is intended to be externally immutable. **Do not** mutate
        any of its attribute containers.
    cCs||_||_||_dSr��
candidates�information�incompatibilities)rr,r-r.rrrrAszCriterion.__init__cCs d�dd�|jD��}d�|�S)Nrcss|]\}}d�||�VqdS)z({!r}, via={!r})N)r!)r�reqr	rrrrGs�z%Criterion.__repr__.<locals>.<genexpr>z
Criterion({}))r"r-r!)r�requirementsrrr�__repr__Fs
�zCriterion.__repr__cCsdd�|jD�S)Ncss|]}|jVqdSr)r�r�irrrrNr z-Criterion.iter_requirement.<locals>.<genexpr>�r-r$rrrr#MszCriterion.iter_requirementcCsdd�|jD�S)Ncss|]}|jVqdSr�r	r2rrrrQr z(Criterion.iter_parent.<locals>.<genexpr>r4r$rrr�iter_parentPszCriterion.iter_parentN)rrr
rrr1r#r6rrrrr*-s
r*c@seZdZdS)�ResolutionErrorN)rrr
rrrrr7Tsr7cseZdZ�fdd�Z�ZS)�ResolutionImpossiblecstt|��|�||_dSr)rr8r�causes)rr9rrrrYszResolutionImpossible.__init__�rrr
rr'rrrrr8Xsr8cseZdZ�fdd�Z�ZS)�ResolutionTooDeepcstt|��|�||_dSr)rr;r�round_count)rr<rrrr`szResolutionTooDeep.__init__r:rrrrr;_sr;�Statez!mapping criteria backtrack_causesc@sdeZdZdZdd�Zedd��Zdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dd�ZdS)�
Resolutionz�Stateful resolution object.

    This is designed as a one-off object that holds information to kick start
    the resolution process, and holds the results afterwards.
    cCs||_||_g|_dSr)�_p�_r�_states)r�provider�reporterrrrrpszResolution.__init__cCs,z|jdWSty&td��Yn0dS)N����state)rA�
IndexError�AttributeErrorr$rrrrEuszResolution.statecCs>|jd}t|j��|j��|jdd�d�}|j�|�dS)z�Push a new state into history.

        This new state will be used to hold resolution results of the next
        coming round.
        rDN��mapping�criteria�backtrack_causes)rAr=rI�copyrJrK�append)r�baserErrr�_push_new_state|s
�zResolution._push_new_statec	Cs�|jj||d�|jj|d�}|�|�}|r8t|j�}ng}|jj|t|t	�
d�||gi�t|t	�d�||i�d�}|r�t|j�}|�
t||��nt||�g}tt|�||d�}|js�t|��|||<dS)N)rr	)�requirement_or_candidater#r.��
identifierr0r.r+)r@�adding_requirementr?�identify�get�listr.�find_matchesr�operator�methodcaller�
attrgetterr-rMrr*rr,r)	rrJrr	rRrr.�matchesr-rrr�_add_to_criteria�s>
���
�zResolution._add_to_criteriac	Cs>|jj||jjt|jjt�d��t|jjt�d��|jjd�S)Nr,r-)rR�resolutionsr,r-rK)	r?�get_preferencerErIrrJrXrZrK)r�namerrr�_get_preference�s���zResolution._get_preferencecsBz�jj|�Wnty$YdS0t��fdd�|��D��S)NFc3s|]}�jj|�d�VqdS�)rr)N�r?�is_satisfied_byr��current_pinrrrr�s�z8Resolution._is_current_pin_satisfying.<locals>.<genexpr>)rErI�KeyError�allr#)rr_rrrdr�_is_current_pin_satisfying�s�z%Resolution._is_current_pin_satisfyingcCs4|jj��}|jj|d�D]}|j|||d�q|S)N�r)r5)rErJrLr?�get_dependenciesr\)rr)rJrrrr�_get_updated_criteria�sz Resolution._get_updated_criteriacs��jj|}g}|jD]��z����}Wn<tyd}z$|�|j�WYd}~qWYd}~n
d}~00t��fdd�|��D��}|s�t	�|���j
j�d��jj�|��jj
�|d���jj
|<gS|S)Nc3s|]}�jj|�d�VqdSrarbr�r)rrrr�s�z7Resolution._attempt_to_pin_criterion.<locals>.<genexpr>ri)rErJr,rkrrMrrgr#r(r@�pinning�updaterI�pop)rr_rr9rJ�e�	satisfiedrrlr�_attempt_to_pin_criterion�s&
"�
z$Resolution._attempt_to_pin_criterioncs�t�j�dkr��jd=�j��}|j��\}}dd�|j��D����||gf��jj	|d���fdd�}��
�|�}|rdSqd	S)
a�Perform backtracking.

        When we enter here, the stack is like this::

            [ state Z ]
            [ state Y ]
            [ state X ]
            .... earlier states are irrelevant.

        1. No pins worked for Z, so it does not have a pin.
        2. We want to reset state Y to unpinned, and pin another candidate.
        3. State X holds what state Y was before the pin, but does not
           have the incompatibility information gathered in state Y.

        Each iteration of the loop will:

        1.  Discard Z.
        2.  Discard Y but remember its incompatibility information gathered
            previously, and the failure we're dealing with right now.
        3.  Push a new state Y' based on X, and apply the incompatibility
            information from Y to Y'.
        4a. If this causes Y' to conflict, we need to backtrack again. Make Y'
            the new Z and go back to step 2.
        4b. If the incompatibilities apply cleanly, end backtracking.
        �rDcSsg|]\}}|t|j�f�qSr)rVr.�r�k�vrrr�
<listcomp>s�z)Resolution._backtrack.<locals>.<listcomp>ric	s��D]�\}}|sqz�jj|}Wnty8YqYn0�jj|t�jjt�d��t�jjt�d�||i�d�}t	|�}|s�dS|�
|j�t|t
|j�|d��jj|<qdS)Nr#r.rQFr+T)rErJrfr?rWrrXrYrZr�extendr.r*rVr-)rur.rr[r,��incompatibilities_from_brokenrrr�_patch_criteria!s:
����z.Resolution._backtrack.<locals>._patch_criteriaTF)�lenrArorI�popitemrJ�itemsrMr@�backtrackingrO)r�broken_stater_r)r{�successrryr�
_backtrack�s
�zResolution._backtrackc
sX�jrtd���j��tt��igd�g�_|D]N}z�j�jj	|dd�Wq2t
y~}zt|jj
��WYd}~q2d}~00q2���t|�D]�}�jj|d��fdd��jj	��D�}|s܈jj�jd��jSt|�jd�}��|�}|�r0���}	d	d�|D��jjdd�<|	�s8t�jj��n����jj|�jd
�q�t|��dS)Nzalready resolvedrHr5)�indexcs g|]\}}��||�s|�qSr)rh)r�keyrr$rrrwis�z&Resolution.resolve.<locals>.<listcomp>)rE)r�cSsg|]}|jD]}|�qqSrr4)r�cr3rrrrw|s)r�rE)rA�RuntimeErrorr@�startingr=�collections�OrderedDictr\rErJrr8rr-rO�range�starting_roundr~�ending�minr`rrr�rK�ending_roundr;)
rr0�
max_roundsrrp�round_index�unsatisfied_namesr_�failure_causesr�rr$r�resolveMsF
��$

�

�zResolution.resolveN)rrr
rr�propertyrErOr\r`rhrkrrr�r�rrrrr>is
'
$Yr>c	Cs�||vrdS||vrdS||��D]b}z|t|�}WntyNYq$Yn0||vrh|�|�dSt||||�r$|�|�dSq$dS)NTF)r6�idrf�add�_has_route_to_root)rJr��all_keys�	connected�p�pkeyrrrr��s 


r��Resultzmapping graph criteriac
s�|j}dd�|��D�}d|td�<t�}|�d�dh�|j��D]�\}}t|j||��s^qD||vrp|�|�|��D]J}z|t|�}Wnty�YqxYn0||vr�|�|�|�	||�qxqDt
�fdd�|��D�||jd�S)NcSsi|]\}}t|�|�qSr)r�rtrrr�
<dictcomp>�r z!_build_result.<locals>.<dictcomp>csi|]\}}|�vr||�qSrrrt�r�rrr��r )rI�graphrJ)rIr~r�rr�rJr�r6rf�connectr�)rErIr�r�r�rr�r�rr�r�
_build_result�s0



�r�c@seZdZdZeZddd�ZdS)�Resolverz3The thing that performs the actual resolution work.�dcCs$t|j|j�}|j||d�}t|�S)aTake a collection of constraints, spit out the resolution result.

        The return value is a representation to the final resolution result. It
        is a tuple subclass with three public members:

        * `mapping`: A dict of resolved candidates. Each key is an identifier
            of a requirement (as returned by the provider's `identify` method),
            and the value is the resolved candidate.
        * `graph`: A `DirectedGraph` instance representing the dependency tree.
            The vertices are keys of `mapping`, and each edge represents *why*
            a particular package is included. A special vertex `None` is
            included to represent parents of user-supplied requirements.
        * `criteria`: A dict of "criteria" that hold detailed information on
            how edges in the graph are derived. Each key is an identifier of a
            requirement, and the value is a `Criterion` instance.

        The following exceptions may be raised if a resolution cannot be found:

        * `ResolutionImpossible`: A resolution cannot be found for the given
            combination of requirements. The `causes` attribute of the
            exception is a list of (requirement, parent), giving the
            requirements that could not be satisfied.
        * `ResolutionTooDeep`: The dependency tree is too deeply nested and
            the resolver gave up. This is usually caused by a circular
            dependency, but you can try to resolve this by increasing the
            `max_rounds` argument.
        )r�)r>rBrCr�r�)rr0r��
resolutionrErrrr��szResolver.resolveN)r�)rrr
rr
�base_exceptionr�rrrrr��sr�)r�rX�	providersr�structsrrr�
namedtupler�	Exceptionr
rr(�objectr*r7r8r;r=r>r�r�r�r�rrrr�<module>s*�
'%site-packages/pip/_vendor/resolvelib/__pycache__/__init__.cpython-39.pyc000064400000001205150564745010022263 0ustar00a

��Re�@sLgd�ZdZddlmZmZddlmZddlmZm	Z	m
Z
mZmZm
Z
dS))
�__version__�AbstractProvider�AbstractResolver�BaseReporter�InconsistentCandidate�Resolver�RequirementsConflicted�ResolutionError�ResolutionImpossible�ResolutionTooDeepz0.8.0�)rr)r)rrrrr	r
N)�__all__r�	providersrr�	reportersr�	resolversrrrrr	r
�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/__init__.py�<module>s
site-packages/pip/_vendor/resolvelib/structs.py000064400000011272150564745010015751 0ustar00import itertools

from .compat import collections_abc


class DirectedGraph(object):
    """A graph structure with directed edges."""

    def __init__(self):
        self._vertices = set()
        self._forwards = {}  # <key> -> Set[<key>]
        self._backwards = {}  # <key> -> Set[<key>]

    def __iter__(self):
        return iter(self._vertices)

    def __len__(self):
        return len(self._vertices)

    def __contains__(self, key):
        return key in self._vertices

    def copy(self):
        """Return a shallow copy of this graph."""
        other = DirectedGraph()
        other._vertices = set(self._vertices)
        other._forwards = {k: set(v) for k, v in self._forwards.items()}
        other._backwards = {k: set(v) for k, v in self._backwards.items()}
        return other

    def add(self, key):
        """Add a new vertex to the graph."""
        if key in self._vertices:
            raise ValueError("vertex exists")
        self._vertices.add(key)
        self._forwards[key] = set()
        self._backwards[key] = set()

    def remove(self, key):
        """Remove a vertex from the graph, disconnecting all edges from/to it."""
        self._vertices.remove(key)
        for f in self._forwards.pop(key):
            self._backwards[f].remove(key)
        for t in self._backwards.pop(key):
            self._forwards[t].remove(key)

    def connected(self, f, t):
        return f in self._backwards[t] and t in self._forwards[f]

    def connect(self, f, t):
        """Connect two existing vertices.

        Nothing happens if the vertices are already connected.
        """
        if t not in self._vertices:
            raise KeyError(t)
        self._forwards[f].add(t)
        self._backwards[t].add(f)

    def iter_edges(self):
        for f, children in self._forwards.items():
            for t in children:
                yield f, t

    def iter_children(self, key):
        return iter(self._forwards[key])

    def iter_parents(self, key):
        return iter(self._backwards[key])


class IteratorMapping(collections_abc.Mapping):
    def __init__(self, mapping, accessor, appends=None):
        self._mapping = mapping
        self._accessor = accessor
        self._appends = appends or {}

    def __repr__(self):
        return "IteratorMapping({!r}, {!r}, {!r})".format(
            self._mapping,
            self._accessor,
            self._appends,
        )

    def __bool__(self):
        return bool(self._mapping or self._appends)

    __nonzero__ = __bool__  # XXX: Python 2.

    def __contains__(self, key):
        return key in self._mapping or key in self._appends

    def __getitem__(self, k):
        try:
            v = self._mapping[k]
        except KeyError:
            return iter(self._appends[k])
        return itertools.chain(self._accessor(v), self._appends.get(k, ()))

    def __iter__(self):
        more = (k for k in self._appends if k not in self._mapping)
        return itertools.chain(self._mapping, more)

    def __len__(self):
        more = sum(1 for k in self._appends if k not in self._mapping)
        return len(self._mapping) + more


class _FactoryIterableView(object):
    """Wrap an iterator factory returned by `find_matches()`.

    Calling `iter()` on this class would invoke the underlying iterator
    factory, making it a "collection with ordering" that can be iterated
    through multiple times, but lacks random access methods presented in
    built-in Python sequence types.
    """

    def __init__(self, factory):
        self._factory = factory

    def __repr__(self):
        return "{}({})".format(type(self).__name__, list(self._factory()))

    def __bool__(self):
        try:
            next(self._factory())
        except StopIteration:
            return False
        return True

    __nonzero__ = __bool__  # XXX: Python 2.

    def __iter__(self):
        return self._factory()


class _SequenceIterableView(object):
    """Wrap an iterable returned by find_matches().

    This is essentially just a proxy to the underlying sequence that provides
    the same interface as `_FactoryIterableView`.
    """

    def __init__(self, sequence):
        self._sequence = sequence

    def __repr__(self):
        return "{}({})".format(type(self).__name__, self._sequence)

    def __bool__(self):
        return bool(self._sequence)

    __nonzero__ = __bool__  # XXX: Python 2.

    def __iter__(self):
        return iter(self._sequence)


def build_iter_view(matches):
    """Build an iterable view from the value returned by `find_matches()`."""
    if callable(matches):
        return _FactoryIterableView(matches)
    if not isinstance(matches, collections_abc.Sequence):
        matches = list(matches)
    return _SequenceIterableView(matches)
site-packages/pip/_vendor/resolvelib/__init__.py000064400000001031150564745010015771 0ustar00__all__ = [
    "__version__",
    "AbstractProvider",
    "AbstractResolver",
    "BaseReporter",
    "InconsistentCandidate",
    "Resolver",
    "RequirementsConflicted",
    "ResolutionError",
    "ResolutionImpossible",
    "ResolutionTooDeep",
]

__version__ = "0.8.0"


from .providers import AbstractProvider, AbstractResolver
from .reporters import BaseReporter
from .resolvers import (
    InconsistentCandidate,
    RequirementsConflicted,
    Resolver,
    ResolutionError,
    ResolutionImpossible,
    ResolutionTooDeep,
)
site-packages/pip/_vendor/resolvelib/resolvers.py000064400000042204150564745010016265 0ustar00import collections
import operator

from .providers import AbstractResolver
from .structs import DirectedGraph, IteratorMapping, build_iter_view


RequirementInformation = collections.namedtuple(
    "RequirementInformation", ["requirement", "parent"]
)


class ResolverException(Exception):
    """A base class for all exceptions raised by this module.

    Exceptions derived by this class should all be handled in this module. Any
    bubbling pass the resolver should be treated as a bug.
    """


class RequirementsConflicted(ResolverException):
    def __init__(self, criterion):
        super(RequirementsConflicted, self).__init__(criterion)
        self.criterion = criterion

    def __str__(self):
        return "Requirements conflict: {}".format(
            ", ".join(repr(r) for r in self.criterion.iter_requirement()),
        )


class InconsistentCandidate(ResolverException):
    def __init__(self, candidate, criterion):
        super(InconsistentCandidate, self).__init__(candidate, criterion)
        self.candidate = candidate
        self.criterion = criterion

    def __str__(self):
        return "Provided candidate {!r} does not satisfy {}".format(
            self.candidate,
            ", ".join(repr(r) for r in self.criterion.iter_requirement()),
        )


class Criterion(object):
    """Representation of possible resolution results of a package.

    This holds three attributes:

    * `information` is a collection of `RequirementInformation` pairs.
      Each pair is a requirement contributing to this criterion, and the
      candidate that provides the requirement.
    * `incompatibilities` is a collection of all known not-to-work candidates
      to exclude from consideration.
    * `candidates` is a collection containing all possible candidates deducted
      from the union of contributing requirements and known incompatibilities.
      It should never be empty, except when the criterion is an attribute of a
      raised `RequirementsConflicted` (in which case it is always empty).

    .. note::
        This class is intended to be externally immutable. **Do not** mutate
        any of its attribute containers.
    """

    def __init__(self, candidates, information, incompatibilities):
        self.candidates = candidates
        self.information = information
        self.incompatibilities = incompatibilities

    def __repr__(self):
        requirements = ", ".join(
            "({!r}, via={!r})".format(req, parent)
            for req, parent in self.information
        )
        return "Criterion({})".format(requirements)

    def iter_requirement(self):
        return (i.requirement for i in self.information)

    def iter_parent(self):
        return (i.parent for i in self.information)


class ResolutionError(ResolverException):
    pass


class ResolutionImpossible(ResolutionError):
    def __init__(self, causes):
        super(ResolutionImpossible, self).__init__(causes)
        # causes is a list of RequirementInformation objects
        self.causes = causes


class ResolutionTooDeep(ResolutionError):
    def __init__(self, round_count):
        super(ResolutionTooDeep, self).__init__(round_count)
        self.round_count = round_count


# Resolution state in a round.
State = collections.namedtuple("State", "mapping criteria backtrack_causes")


class Resolution(object):
    """Stateful resolution object.

    This is designed as a one-off object that holds information to kick start
    the resolution process, and holds the results afterwards.
    """

    def __init__(self, provider, reporter):
        self._p = provider
        self._r = reporter
        self._states = []

    @property
    def state(self):
        try:
            return self._states[-1]
        except IndexError:
            raise AttributeError("state")

    def _push_new_state(self):
        """Push a new state into history.

        This new state will be used to hold resolution results of the next
        coming round.
        """
        base = self._states[-1]
        state = State(
            mapping=base.mapping.copy(),
            criteria=base.criteria.copy(),
            backtrack_causes=base.backtrack_causes[:],
        )
        self._states.append(state)

    def _add_to_criteria(self, criteria, requirement, parent):
        self._r.adding_requirement(requirement=requirement, parent=parent)

        identifier = self._p.identify(requirement_or_candidate=requirement)
        criterion = criteria.get(identifier)
        if criterion:
            incompatibilities = list(criterion.incompatibilities)
        else:
            incompatibilities = []

        matches = self._p.find_matches(
            identifier=identifier,
            requirements=IteratorMapping(
                criteria,
                operator.methodcaller("iter_requirement"),
                {identifier: [requirement]},
            ),
            incompatibilities=IteratorMapping(
                criteria,
                operator.attrgetter("incompatibilities"),
                {identifier: incompatibilities},
            ),
        )

        if criterion:
            information = list(criterion.information)
            information.append(RequirementInformation(requirement, parent))
        else:
            information = [RequirementInformation(requirement, parent)]

        criterion = Criterion(
            candidates=build_iter_view(matches),
            information=information,
            incompatibilities=incompatibilities,
        )
        if not criterion.candidates:
            raise RequirementsConflicted(criterion)
        criteria[identifier] = criterion

    def _get_preference(self, name):
        return self._p.get_preference(
            identifier=name,
            resolutions=self.state.mapping,
            candidates=IteratorMapping(
                self.state.criteria,
                operator.attrgetter("candidates"),
            ),
            information=IteratorMapping(
                self.state.criteria,
                operator.attrgetter("information"),
            ),
            backtrack_causes=self.state.backtrack_causes,
        )

    def _is_current_pin_satisfying(self, name, criterion):
        try:
            current_pin = self.state.mapping[name]
        except KeyError:
            return False
        return all(
            self._p.is_satisfied_by(requirement=r, candidate=current_pin)
            for r in criterion.iter_requirement()
        )

    def _get_updated_criteria(self, candidate):
        criteria = self.state.criteria.copy()
        for requirement in self._p.get_dependencies(candidate=candidate):
            self._add_to_criteria(criteria, requirement, parent=candidate)
        return criteria

    def _attempt_to_pin_criterion(self, name):
        criterion = self.state.criteria[name]

        causes = []
        for candidate in criterion.candidates:
            try:
                criteria = self._get_updated_criteria(candidate)
            except RequirementsConflicted as e:
                causes.append(e.criterion)
                continue

            # Check the newly-pinned candidate actually works. This should
            # always pass under normal circumstances, but in the case of a
            # faulty provider, we will raise an error to notify the implementer
            # to fix find_matches() and/or is_satisfied_by().
            satisfied = all(
                self._p.is_satisfied_by(requirement=r, candidate=candidate)
                for r in criterion.iter_requirement()
            )
            if not satisfied:
                raise InconsistentCandidate(candidate, criterion)

            self._r.pinning(candidate=candidate)
            self.state.criteria.update(criteria)

            # Put newly-pinned candidate at the end. This is essential because
            # backtracking looks at this mapping to get the last pin.
            self.state.mapping.pop(name, None)
            self.state.mapping[name] = candidate

            return []

        # All candidates tried, nothing works. This criterion is a dead
        # end, signal for backtracking.
        return causes

    def _backtrack(self):
        """Perform backtracking.

        When we enter here, the stack is like this::

            [ state Z ]
            [ state Y ]
            [ state X ]
            .... earlier states are irrelevant.

        1. No pins worked for Z, so it does not have a pin.
        2. We want to reset state Y to unpinned, and pin another candidate.
        3. State X holds what state Y was before the pin, but does not
           have the incompatibility information gathered in state Y.

        Each iteration of the loop will:

        1.  Discard Z.
        2.  Discard Y but remember its incompatibility information gathered
            previously, and the failure we're dealing with right now.
        3.  Push a new state Y' based on X, and apply the incompatibility
            information from Y to Y'.
        4a. If this causes Y' to conflict, we need to backtrack again. Make Y'
            the new Z and go back to step 2.
        4b. If the incompatibilities apply cleanly, end backtracking.
        """
        while len(self._states) >= 3:
            # Remove the state that triggered backtracking.
            del self._states[-1]

            # Retrieve the last candidate pin and known incompatibilities.
            broken_state = self._states.pop()
            name, candidate = broken_state.mapping.popitem()
            incompatibilities_from_broken = [
                (k, list(v.incompatibilities))
                for k, v in broken_state.criteria.items()
            ]

            # Also mark the newly known incompatibility.
            incompatibilities_from_broken.append((name, [candidate]))

            self._r.backtracking(candidate=candidate)

            # Create a new state from the last known-to-work one, and apply
            # the previously gathered incompatibility information.
            def _patch_criteria():
                for k, incompatibilities in incompatibilities_from_broken:
                    if not incompatibilities:
                        continue
                    try:
                        criterion = self.state.criteria[k]
                    except KeyError:
                        continue
                    matches = self._p.find_matches(
                        identifier=k,
                        requirements=IteratorMapping(
                            self.state.criteria,
                            operator.methodcaller("iter_requirement"),
                        ),
                        incompatibilities=IteratorMapping(
                            self.state.criteria,
                            operator.attrgetter("incompatibilities"),
                            {k: incompatibilities},
                        ),
                    )
                    candidates = build_iter_view(matches)
                    if not candidates:
                        return False
                    incompatibilities.extend(criterion.incompatibilities)
                    self.state.criteria[k] = Criterion(
                        candidates=candidates,
                        information=list(criterion.information),
                        incompatibilities=incompatibilities,
                    )
                return True

            self._push_new_state()
            success = _patch_criteria()

            # It works! Let's work on this new state.
            if success:
                return True

            # State does not work after applying known incompatibilities.
            # Try the still previous state.

        # No way to backtrack anymore.
        return False

    def resolve(self, requirements, max_rounds):
        if self._states:
            raise RuntimeError("already resolved")

        self._r.starting()

        # Initialize the root state.
        self._states = [
            State(
                mapping=collections.OrderedDict(),
                criteria={},
                backtrack_causes=[],
            )
        ]
        for r in requirements:
            try:
                self._add_to_criteria(self.state.criteria, r, parent=None)
            except RequirementsConflicted as e:
                raise ResolutionImpossible(e.criterion.information)

        # The root state is saved as a sentinel so the first ever pin can have
        # something to backtrack to if it fails. The root state is basically
        # pinning the virtual "root" package in the graph.
        self._push_new_state()

        for round_index in range(max_rounds):
            self._r.starting_round(index=round_index)

            unsatisfied_names = [
                key
                for key, criterion in self.state.criteria.items()
                if not self._is_current_pin_satisfying(key, criterion)
            ]

            # All criteria are accounted for. Nothing more to pin, we are done!
            if not unsatisfied_names:
                self._r.ending(state=self.state)
                return self.state

            # Choose the most preferred unpinned criterion to try.
            name = min(unsatisfied_names, key=self._get_preference)
            failure_causes = self._attempt_to_pin_criterion(name)

            if failure_causes:
                # Backtrack if pinning fails. The backtrack process puts us in
                # an unpinned state, so we can work on it in the next round.
                success = self._backtrack()
                self.state.backtrack_causes[:] = [
                    i for c in failure_causes for i in c.information
                ]

                # Dead ends everywhere. Give up.
                if not success:
                    raise ResolutionImpossible(self.state.backtrack_causes)
            else:
                # Pinning was successful. Push a new state to do another pin.
                self._push_new_state()

            self._r.ending_round(index=round_index, state=self.state)

        raise ResolutionTooDeep(max_rounds)


def _has_route_to_root(criteria, key, all_keys, connected):
    if key in connected:
        return True
    if key not in criteria:
        return False
    for p in criteria[key].iter_parent():
        try:
            pkey = all_keys[id(p)]
        except KeyError:
            continue
        if pkey in connected:
            connected.add(key)
            return True
        if _has_route_to_root(criteria, pkey, all_keys, connected):
            connected.add(key)
            return True
    return False


Result = collections.namedtuple("Result", "mapping graph criteria")


def _build_result(state):
    mapping = state.mapping
    all_keys = {id(v): k for k, v in mapping.items()}
    all_keys[id(None)] = None

    graph = DirectedGraph()
    graph.add(None)  # Sentinel as root dependencies' parent.

    connected = {None}
    for key, criterion in state.criteria.items():
        if not _has_route_to_root(state.criteria, key, all_keys, connected):
            continue
        if key not in graph:
            graph.add(key)
        for p in criterion.iter_parent():
            try:
                pkey = all_keys[id(p)]
            except KeyError:
                continue
            if pkey not in graph:
                graph.add(pkey)
            graph.connect(pkey, key)

    return Result(
        mapping={k: v for k, v in mapping.items() if k in connected},
        graph=graph,
        criteria=state.criteria,
    )


class Resolver(AbstractResolver):
    """The thing that performs the actual resolution work."""

    base_exception = ResolverException

    def resolve(self, requirements, max_rounds=100):
        """Take a collection of constraints, spit out the resolution result.

        The return value is a representation to the final resolution result. It
        is a tuple subclass with three public members:

        * `mapping`: A dict of resolved candidates. Each key is an identifier
            of a requirement (as returned by the provider's `identify` method),
            and the value is the resolved candidate.
        * `graph`: A `DirectedGraph` instance representing the dependency tree.
            The vertices are keys of `mapping`, and each edge represents *why*
            a particular package is included. A special vertex `None` is
            included to represent parents of user-supplied requirements.
        * `criteria`: A dict of "criteria" that hold detailed information on
            how edges in the graph are derived. Each key is an identifier of a
            requirement, and the value is a `Criterion` instance.

        The following exceptions may be raised if a resolution cannot be found:

        * `ResolutionImpossible`: A resolution cannot be found for the given
            combination of requirements. The `causes` attribute of the
            exception is a list of (requirement, parent), giving the
            requirements that could not be satisfied.
        * `ResolutionTooDeep`: The dependency tree is too deeply nested and
            the resolver gave up. This is usually caused by a circular
            dependency, but you can try to resolve this by increasing the
            `max_rounds` argument.
        """
        resolution = Resolution(self.provider, self.reporter)
        state = resolution.resolve(requirements, max_rounds=max_rounds)
        return _build_result(state)
site-packages/pip/_vendor/resolvelib/compat/__pycache__/collections_abc.cpython-39.pyc000064400000000636150564745010025141 0ustar00a

��Re��@sDddgZzddlmZmZWn"ey>ddlmZmZYn0dS)�Mapping�Sequence�)rrN)�__all__�collections.abcrr�ImportError�collections�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/compat/collections_abc.py�<module>ssite-packages/pip/_vendor/resolvelib/compat/__pycache__/__init__.cpython-39.pyc000064400000000360150564745010023547 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/resolvelib/compat/__init__.py�<module>�site-packages/pip/_vendor/resolvelib/compat/collections_abc.py000064400000000234150564745010020644 0ustar00__all__ = ["Mapping", "Sequence"]

try:
    from collections.abc import Mapping, Sequence
except ImportError:
    from collections import Mapping, Sequence
site-packages/pip/_vendor/resolvelib/compat/__init__.py000064400000000000150564745010017247 0ustar00site-packages/pip/_vendor/resolvelib/reporters.py000064400000002524150564745010016267 0ustar00class BaseReporter(object):
    """Delegate class to provider progress reporting for the resolver."""

    def starting(self):
        """Called before the resolution actually starts."""

    def starting_round(self, index):
        """Called before each round of resolution starts.

        The index is zero-based.
        """

    def ending_round(self, index, state):
        """Called before each round of resolution ends.

        This is NOT called if the resolution ends at this round. Use `ending`
        if you want to report finalization. The index is zero-based.
        """

    def ending(self, state):
        """Called before the resolution ends successfully."""

    def adding_requirement(self, requirement, parent):
        """Called when adding a new requirement into the resolve criteria.

        :param requirement: The additional requirement to be applied to filter
            the available candidaites.
        :param parent: The candidate that requires ``requirement`` as a
            dependency, or None if ``requirement`` is one of the root
            requirements passed in from ``Resolver.resolve()``.
        """

    def backtracking(self, candidate):
        """Called when rejecting a candidate during backtracking."""

    def pinning(self, candidate):
        """Called when adding a candidate to the potential solution."""
site-packages/pip/_vendor/colorama/__pycache__/initialise.cpython-39.pyc000064400000003325150564745010022312 0ustar00a

��Re{�@stddlZddlZddlZddlmZdadadadada	dd�Z
ddd	�Zd
d�Zej
dd
��Zdd�Zdd�ZdS)�N�)�AnsiToWin32FcCstdurtt���dS�N)r�orig_stdout�	reset_all�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/initialise.pyrsrTcCs�|st|||g�rtd��tjatjatjdur6dantt||||�t_atjdur\da	ntt||||�t_a	t
s�t�t
�da
dS)Nz,wrap=False conflicts with any other arg=TrueT)�any�
ValueError�sys�stdoutr�stderr�orig_stderr�wrapped_stdout�wrap_stream�wrapped_stderr�atexit_done�atexit�registerr)�	autoreset�convert�strip�wraprrr�inits
�
�
rcCs tdurtt_tdurtt_dSr)rrrrr
rrrr�deinit3srcos,t|i|��zdVWt�nt�0dSr)rr)�args�kwargsrrr�
colorama_text:srcCs tdurtt_tdurtt_dSr)rrrrr
rrrr�reinitCsrcCs&|r"t||||d�}|��r"|j}|S)N)rrr)r�should_wrap�stream)r rrrr�wrapperrrrrJs�r)FNNT)r�
contextlibr�ansitowin32rrrrrrrrr�contextmanagerrrrrrrr�<module>s

site-packages/pip/_vendor/colorama/__pycache__/ansi.cpython-39.pyc000064400000006303150564745010021111 0ustar00a

��Re�	�@s�dZdZdZdZdd�Zdd�Zdd	d
�Zddd�ZGd
d�de�Z	Gdd�de�Z
Gdd�de	�ZGdd�de	�ZGdd�de	�Z
e�Ze�Ze
�Ze
�ZdS)z�
This module generates ANSI character codes to printing colors to terminals.
See: http://en.wikipedia.org/wiki/ANSI_escape_code
z]�cCstt|�dS)N�m��CSI�str)�code�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/ansi.py�
code_to_charssr	cCstd|tS)Nz2;)�OSC�BEL)�titlerrr�	set_titlesr
�cCstt|�dS)N�Jr��moderrr�clear_screensrcCstt|�dS)N�Krrrrr�
clear_linesrc@seZdZdd�ZdS)�	AnsiCodescCs6t|�D](}|�d�st||�}t||t|��qdS)N�_)�dir�
startswith�getattr�setattrr	)�self�name�valuerrr�__init__s

zAnsiCodes.__init__N)�__name__�
__module__�__qualname__rrrrrrsrc@s>eZdZd
dd�Zddd�Zddd�Zddd	�Zdd
d�ZdS)�
AnsiCursor�cCstt|�dS)N�Ar�r�nrrr�UP%sz
AnsiCursor.UPcCstt|�dS)N�Brr%rrr�DOWN'szAnsiCursor.DOWNcCstt|�dS)N�Crr%rrr�FORWARD)szAnsiCursor.FORWARDcCstt|�dS)N�Drr%rrr�BACK+szAnsiCursor.BACKcCstt|�dt|�dS)N�;�Hr)r�x�yrrr�POS-szAnsiCursor.POSN)r#)r#)r#)r#)r#r#)rr r!r'r)r+r-r2rrrrr"$s




r"c@sPeZdZdZdZdZdZdZdZdZ	dZ
d	Zd
ZdZ
dZd
ZdZdZdZdZdS)�AnsiFore��� �!�"�#�$�%�'�Z�[�\�]�^�_�`�aN�rr r!�BLACK�RED�GREEN�YELLOW�BLUE�MAGENTA�CYAN�WHITE�RESET�
LIGHTBLACK_EX�LIGHTRED_EX�
LIGHTGREEN_EX�LIGHTYELLOW_EX�LIGHTBLUE_EX�LIGHTMAGENTA_EX�LIGHTCYAN_EX�
LIGHTWHITE_EXrrrrr31s"r3c@sPeZdZdZdZdZdZdZdZdZ	dZ
d	Zd
ZdZ
dZd
ZdZdZdZdZdS)�AnsiBack�(�)�*�+�,�-�.�/�1�d�e�f�g�h�i�j�kNrErrrrrWGs"rWc@seZdZdZdZdZdZdS)�	AnsiStyler#r��N)rr r!�BRIGHT�DIM�NORMAL�	RESET_ALLrrrrri]sriN)r)r)�__doc__rr
rr	r
rr�objectrr"r3rWri�Fore�Back�Style�Cursorrrrr�<module>s 


site-packages/pip/_vendor/colorama/__pycache__/ansitowin32.cpython-39.pyc000064400000017065150564745010022346 0ustar00a

��Re)�@s�ddlZddlZddlZddlmZmZmZmZmZddl	m
Z
mZmZddl
mZmZdZ	edurje
�Z	Gdd�de�ZGdd	�d	e�ZdS)
�N�)�AnsiFore�AnsiBack�	AnsiStyle�Style�BEL)�WinTerm�WinColor�WinStyle)�windll�winapi_testc@sLeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	e
dd��ZdS)�
StreamWrapperz�
    Wraps a stream (such as stdout), acting as a transparent proxy for all
    attribute access apart from method 'write()', which is delegated to our
    Converter instance.
    cCs||_||_dS�N)�_StreamWrapper__wrapped�_StreamWrapper__convertor)�self�wrapped�	converter�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/ansitowin32.py�__init__szStreamWrapper.__init__cCst|j|�Sr)�getattrr)r�namerrr�__getattr__szStreamWrapper.__getattr__cOs|jj|i|��Sr)r�	__enter__�r�args�kwargsrrrrszStreamWrapper.__enter__cOs|jj|i|��Sr)r�__exit__rrrrr%szStreamWrapper.__exit__cCs|j�|�dSr)r�write�r�textrrrr(szStreamWrapper.writecCsZ|j}dtjvr0|dur0|tjus,|tjur0dSz
|j}WntyNYdS0|�SdS)N�PYCHARM_HOSTEDTF)r�os�environ�sys�
__stdout__�
__stderr__�isatty�AttributeError)r�stream�
stream_isattyrrrr(+s

zStreamWrapper.isattycCs(|j}z|jWSty"YdS0dS�NT)r�closedr))rr*rrrr-7s
zStreamWrapper.closedN)�__name__�
__module__�__qualname__�__doc__rrrrrr(�propertyr-rrrrr
sr
c@s~eZdZdZe�d�Ze�d�Zddd�Zdd	�Z	d
d�Z
dd
�Zdd�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�ZdS)�AnsiToWin32z�
    Implements a 'write()' method which, on Windows, will strip ANSI character
    sequences from the text, and if outputting to a tty, will convert them into
    win32 function calls.
    z?\[((?:\d|;)*)([a-zA-Z])?z?\]([^]*)()?NFcCs�||_||_t||�|_tjdk}|o*t�}|durN|pL|jjoL|j��}||_	|durt|or|jjor|j��}||_
|��|_|jt
ju|_dS)N�nt)r�	autoresetr
r*r#rrr-r(�strip�convert�get_win32_calls�win32_callsr%�stderr�	on_stderr)rrr7r6r5�
on_windows�conversion_supportedrrrrIs


zAnsiToWin32.__init__cCs|jp|jp|jS)aj
        True if this class is actually needed. If false, then the output
        stream will not be affected, nor will win32 calls be issued, so
        wrapping stdout is not actually required. This will generally be
        False on non-Windows platforms, unless optional functionality like
        autoreset has been requested using kwargs to init()
        )r7r6r5�rrrr�should_wrapjszAnsiToWin32.should_wrapcNCs@|j�r:t�r:tjtjftjtjtjftjtjtj	ftj	tjtj	ft
jtjt
jft
jtjt
jft
jtjt
jft
jtjt
jft
jtjt
jft
jtjt
jft
jtjt
jft
jtjt
jft
jtjft
jtjt
jdft
jtjt
jdft
jtjt
jdft
jtjt
jdft
jtjt
jdft
jtjt
jdft
jtjt
jdft
jtjt
jdftjtj t
jftjtj t
jftjtj t
jftjtj t
jftjtj t
jftjtj t
jftjtj t
jftjtj t
jftjtj ftjtj t
jdftjtj t
jdftjtj t
jdftjtj t
jdftjtj t
jdftjtj t
jdftjtj t
jdftjtj t
jdfi&St!�Sr,)"r7�wintermr�	RESET_ALL�	reset_all�BRIGHT�styler
�DIM�NORMALr�BLACK�forer	�RED�GREEN�YELLOW�BLUE�MAGENTA�CYAN�WHITE�GREY�RESET�
LIGHTBLACK_EX�LIGHTRED_EX�
LIGHTGREEN_EX�LIGHTYELLOW_EX�LIGHTBLUE_EX�LIGHTMAGENTA_EX�LIGHTCYAN_EX�
LIGHTWHITE_EXr�back�dictr>rrrr8tsR


�(zAnsiToWin32.get_win32_callscCs@|js|jr|�|�n|j�|�|j��|jr<|��dSr)r6r7�write_and_convertrr�flushr5rBr rrrr�s
zAnsiToWin32.writecCs4|jr|�dd�n|js0|jjs0|j�tj�dS)N�m�r)	r7�
call_win32r6r*r-rrrrAr>rrrrB�szAnsiToWin32.reset_allcCsbd}|�|�}|j�|�D]0}|��\}}|�|||�|j|���|}q|�||t|��dS)z�
        Write the given text to our wrapped stream, stripping any ANSI
        sequences from the text, and optionally converting them into win32
        calls.
        rN)�convert_osc�ANSI_CSI_RE�finditer�span�write_plain_text�convert_ansi�groups�len)rr!�cursor�match�start�endrrrr\�s
zAnsiToWin32.write_and_convertcCs*||kr&|j�|||��|j��dSr)rrr])rr!rkrlrrrre�szAnsiToWin32.write_plain_textcCs"|jr|�||�}|�||�dSr)r7�extract_paramsr`)r�paramstring�command�paramsrrrrf�szAnsiToWin32.convert_ansicCsz|dvr8tdd�|�d�D��}t|�dkrv|d}q n>tdd�|�d�D��}t|�dkrv|d	vrjd
}n|dvrvd}|S)N�Hfcss&|]}t|�dkrt|�ndVqdS)rrN�rh�int��.0�prrr�	<genexpr>��z-AnsiToWin32.extract_params.<locals>.<genexpr>�;�)rcss"|]}t|�dkrt|�VqdS)rNrrrtrrrrw�rxr�JKmr_�ABCD)�tuple�splitrh)rrornrprrrrm�szAnsiToWin32.extract_paramscCs�|dkrV|D]F}||jvr|j|}|d}|dd�}t|jd�}||i|��qn�|dvrttj|d|jd�n�|dvr�tj|d|jd�nf|dvr�tj||jd�nL|dvr�|d}d|fd|f|df|dfd	�|\}	}
tj|	|
|jd�dS)
Nr^rr)r;�J�Krqr|)�A�B�C�D)r9r[r;r@�erase_screen�
erase_line�set_cursor_position�
cursor_adjust)rrorp�param�	func_args�funcrr�n�x�yrrrr`�s$

*zAnsiToWin32.call_win32cCs�|j�|�D]n}|��\}}|d|�||d�}|��\}}|tkr|�d�dkr|�d�}|ddvrt�|d�q|S)Nryrr�02)	�ANSI_OSC_RErcrdrgr�countr~r@�	set_title)rr!rjrkrlrnrorprrrra�s
zAnsiToWin32.convert_osc)NNF)r.r/r0r1�re�compilerbr�rr?r8rrBr\rerfrmr`rarrrrr3@s


!
,
r3)r�r%r#�ansirrrrrr@rr	r
�win32rr�objectr
r3rrrr�<module>s0site-packages/pip/_vendor/colorama/__pycache__/winterm.cpython-39.pyc000064400000011137150564745010021645 0ustar00a

��Re&�@s@ddlmZGdd�de�ZGdd�de�ZGdd�de�ZdS)	�)�win32c@s,eZdZdZdZdZdZdZdZdZ	dZ
d	S)
�WinColor�r������N)�__name__�
__module__�__qualname__�BLACK�BLUE�GREEN�CYAN�RED�MAGENTA�YELLOW�GREY�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/winterm.pyrsrc@seZdZdZdZdZdS)�WinStyler��N)rrr
�NORMAL�BRIGHT�BRIGHT_BACKGROUNDrrrrrsrc@s�eZdZdd�Zdd�Zdd�Zd dd	�Zd!dd�Zd"d
d�Zd#dd�Z	d$dd�Z
dd�Zd%dd�Zd&dd�Z
d'dd�Zd(dd�Zdd�ZdS))�WinTermcCs>t�tj�j|_|�|j�|j|_|j|_	|j
|_d|_dS)Nr)
r�GetConsoleScreenBufferInfo�STDOUT�wAttributes�_default�	set_attrs�_fore�
_default_fore�_back�
_default_back�_style�_default_style�_light��selfrrr�__init__szWinTerm.__init__cCs|j|jd|j|jBS)N�)r$r&r(r*r+rrr�	get_attrs$szWinTerm.get_attrscCs.|d@|_|d?d@|_|tjtjB@|_dS)Nr
r)r$r&rrrr()r,�valuerrrr#'s
zWinTerm.set_attrsNcCs$|�|j�|j|jd�d|_dS)N)�attrsr)r#r"�set_consoler*)r,�	on_stderrrrr�	reset_all,szWinTerm.reset_allFcCsL|dur|j}||_|r*|jtjO_n|jtjM_|j|d�dS�N)r3)r%r$r*rrr2)r,�fore�lightr3rrrr61szWinTerm.forecCsL|dur|j}||_|r*|jtjO_n|jtjM_|j|d�dSr5)r'r&r*rrr2)r,�backr7r3rrrr8<szWinTerm.backcCs$|dur|j}||_|j|d�dSr5)r)r(r2)r,�styler3rrrr9Gsz
WinTerm.stylecCs0|dur|��}tj}|r tj}t�||�dS�N)r/rr �STDERR�SetConsoleTextAttribute)r,r1r3�handlerrrr2MszWinTerm.set_consolecCs,t�|�j}|jd7_|jd7_|S)Nr)rr�dwCursorPosition�X�Y)r,r=�positionrrr�get_positionUszWinTerm.get_positioncCs,|durdStj}|rtj}t�||�dSr:)rr r;�SetConsoleCursorPosition)r,rAr3r=rrr�set_cursor_position]szWinTerm.set_cursor_positioncCsBtj}|rtj}|�|�}|j||j|f}tj||dd�dS)NF)�adjust)rr r;rBr@r?rC)r,�x�yr3r=rA�adjusted_positionrrr�
cursor_adjustgs
zWinTerm.cursor_adjustrc	Cs�tj}|rtj}t�|�}|jj|jj}|jj|jj|jj}|dkrZ|j}||}n8|dkrtt�dd�}|}n|dkr�t�dd�}|}ndSt�	|d||�t�
||��||�|dkr�t�|d�dS)Nrrr� )rr)
rr r;r�dwSizer?r@r>�COORD�FillConsoleOutputCharacter�FillConsoleOutputAttributer/rC)	r,�moder3r=�csbi�cells_in_screen�cells_before_cursor�
from_coord�cells_to_eraserrr�erase_screenos(

zWinTerm.erase_screencCs�tj}|rtj}t�|�}|dkr:|j}|jj|jj}nH|dkr\t�d|jj�}|jj}n&|dkr~t�d|jj�}|jj}ndSt�	|d||�t�
||��||�dS)NrrrrJ)rr r;rr>rKr?rLr@rMrNr/)r,rOr3r=rPrSrTrrr�
erase_line�s 


zWinTerm.erase_linecCst�|�dSr:)r�SetConsoleTitle)r,�titlerrr�	set_title�szWinTerm.set_title)N)NFF)NFF)NF)NF)NF)F)rF)rF)rrr
r-r/r#r4r6r8r9r2rBrDrIrUrVrYrrrrrs








 
rN)�r�objectrrrrrrr�<module>ssite-packages/pip/_vendor/colorama/__pycache__/__init__.cpython-39.pyc000064400000000742150564745010021717 0ustar00a

��Re��@sDddlmZmZmZmZddlmZmZmZm	Z	ddl
mZdZdS)�)�init�deinit�reinit�
colorama_text)�Fore�Back�Style�Cursor)�AnsiToWin32z0.4.4N)
�
initialiserrrr�ansirrrr	�ansitowin32r
�__version__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/__init__.py�<module>ssite-packages/pip/_vendor/colorama/__pycache__/win32.cpython-39.pyc000064400000007615150564745010021130 0ustar00a

��Re�	@s�dZdZz.ddlZddlmZeej�ZddlmZWn,eefybdZdd�Z	dd�Z
Y�nZ0dd	lmZmZm
Z
mZejZGd
d�de�ZejjZejge_eje_ejjZejee�ge_eje_ejj	Zejejge_eje_ejjZejege_eje_ejj Z!eje
ejeeej�ge!_eje!_ejj"Z#ejejejeeej�ge#_eje#_ejj$Z%ej&ge%_eje%_dd
�Z'dd�Z
efdd�Zdd�Z	ddd�Zdd�Z(dd�Z"dd�Z)dS)i����i��N)�
LibraryLoader)�wintypescGsdS�N���_rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/colorama/win32.py�<lambda>�r	cGsdSrrrrrrr	r
)�byref�	Structure�c_char�POINTERc@s>eZdZdZdefdefdejfdejfdefgZdd�Z	d	S)
�CONSOLE_SCREEN_BUFFER_INFOzstruct in wincon.h.�dwSize�dwCursorPosition�wAttributes�srWindow�dwMaximumWindowSizecCsHd|jj|jj|jj|jj|j|jj|jj|jj|jj	|j
j|j
jfS)Nz"(%d,%d,%d,%d,%d,%d,%d,%d,%d,%d,%d))r�Y�Xrrr�Top�LeftZBottomZRightr)�selfrrr�__str__s�z"CONSOLE_SCREEN_BUFFER_INFO.__str__N)
�__name__�
__module__�__qualname__�__doc__�COORDr�WORDZ
SMALL_RECT�_fields_rrrrrrs�rcCst�}t|t|��}t|�Sr)r�_GetConsoleScreenBufferInfor�bool)�handle�csbi�successrrr�_winapi_test\s
�r'cCstdd�tt�tt�fD��S)Ncss|]}t|�VqdSr)r')�.0�hrrr�	<genexpr>cr
zwinapi_test.<locals>.<genexpr>)�any�
_GetStdHandle�STDOUT�STDERRrrrr�winapi_testbs�r/cCs t|�}t�}t|t|��}|Sr)r,rr"r)�	stream_idr$r%r&rrr�GetConsoleScreenBufferInfofs�r1cCst|�}t||�Sr)r,�_SetConsoleTextAttribute)r0�attrsr$rrr�SetConsoleTextAttributemsr4TcCsvt|�}|jdks|jdkr dSt|jd|jd�}|rdtt�j}|j|j7_|j|j7_t|�}t	||�S)Nr�)
rrrr1r-rrrr,�_SetConsoleCursorPosition)r0�position�adjust�adjusted_position�srr$rrr�SetConsoleCursorPositionqs
r;cCsBt|�}t|���}t�|�}t�d�}t||||t|��}|jS)Nr)r,r
�encoder�DWORD�_FillConsoleOutputCharacterAr�value)r0�char�length�startr$�num_writtenr&rrr�FillConsoleOutputCharacter�s

�rDcCs:t|�}t�|�}t�|�}t�d�}t||||t|��S)za FillConsoleOutputAttribute( hConsole, csbi.wAttributes, dwConSize, coordScreen, &cCharsWritten )r)r,rr r=�_FillConsoleOutputAttributer)r0�attrrArBr$�	attributerCrrr�FillConsoleOutputAttribute�s


�rHcCst|�Sr)�_SetConsoleTitleW)�titlerrr�SetConsoleTitle�srK)T)*r-r.�ctypesrZWinDLL�windllr�AttributeError�ImportErrorr4r/rrr
rZ_COORDrrZkernel32ZGetStdHandler,r=�argtypesZHANDLE�restyper1r"ZBOOLr2r r;r6ZFillConsoleOutputCharacterAr>rHrEZSetConsoleTitleWrIZLPCWSTRr'rDrKrrrr�<module>sx
�������


site-packages/pip/_vendor/colorama/winterm.py000064400000014446150564745010015364 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.
from . import win32


# from wincon.h
class WinColor(object):
    BLACK   = 0
    BLUE    = 1
    GREEN   = 2
    CYAN    = 3
    RED     = 4
    MAGENTA = 5
    YELLOW  = 6
    GREY    = 7

# from wincon.h
class WinStyle(object):
    NORMAL              = 0x00 # dim text, dim background
    BRIGHT              = 0x08 # bright text, dim background
    BRIGHT_BACKGROUND   = 0x80 # dim text, bright background

class WinTerm(object):

    def __init__(self):
        self._default = win32.GetConsoleScreenBufferInfo(win32.STDOUT).wAttributes
        self.set_attrs(self._default)
        self._default_fore = self._fore
        self._default_back = self._back
        self._default_style = self._style
        # In order to emulate LIGHT_EX in windows, we borrow the BRIGHT style.
        # So that LIGHT_EX colors and BRIGHT style do not clobber each other,
        # we track them separately, since LIGHT_EX is overwritten by Fore/Back
        # and BRIGHT is overwritten by Style codes.
        self._light = 0

    def get_attrs(self):
        return self._fore + self._back * 16 + (self._style | self._light)

    def set_attrs(self, value):
        self._fore = value & 7
        self._back = (value >> 4) & 7
        self._style = value & (WinStyle.BRIGHT | WinStyle.BRIGHT_BACKGROUND)

    def reset_all(self, on_stderr=None):
        self.set_attrs(self._default)
        self.set_console(attrs=self._default)
        self._light = 0

    def fore(self, fore=None, light=False, on_stderr=False):
        if fore is None:
            fore = self._default_fore
        self._fore = fore
        # Emulate LIGHT_EX with BRIGHT Style
        if light:
            self._light |= WinStyle.BRIGHT
        else:
            self._light &= ~WinStyle.BRIGHT
        self.set_console(on_stderr=on_stderr)

    def back(self, back=None, light=False, on_stderr=False):
        if back is None:
            back = self._default_back
        self._back = back
        # Emulate LIGHT_EX with BRIGHT_BACKGROUND Style
        if light:
            self._light |= WinStyle.BRIGHT_BACKGROUND
        else:
            self._light &= ~WinStyle.BRIGHT_BACKGROUND
        self.set_console(on_stderr=on_stderr)

    def style(self, style=None, on_stderr=False):
        if style is None:
            style = self._default_style
        self._style = style
        self.set_console(on_stderr=on_stderr)

    def set_console(self, attrs=None, on_stderr=False):
        if attrs is None:
            attrs = self.get_attrs()
        handle = win32.STDOUT
        if on_stderr:
            handle = win32.STDERR
        win32.SetConsoleTextAttribute(handle, attrs)

    def get_position(self, handle):
        position = win32.GetConsoleScreenBufferInfo(handle).dwCursorPosition
        # Because Windows coordinates are 0-based,
        # and win32.SetConsoleCursorPosition expects 1-based.
        position.X += 1
        position.Y += 1
        return position

    def set_cursor_position(self, position=None, on_stderr=False):
        if position is None:
            # I'm not currently tracking the position, so there is no default.
            # position = self.get_position()
            return
        handle = win32.STDOUT
        if on_stderr:
            handle = win32.STDERR
        win32.SetConsoleCursorPosition(handle, position)

    def cursor_adjust(self, x, y, on_stderr=False):
        handle = win32.STDOUT
        if on_stderr:
            handle = win32.STDERR
        position = self.get_position(handle)
        adjusted_position = (position.Y + y, position.X + x)
        win32.SetConsoleCursorPosition(handle, adjusted_position, adjust=False)

    def erase_screen(self, mode=0, on_stderr=False):
        # 0 should clear from the cursor to the end of the screen.
        # 1 should clear from the cursor to the beginning of the screen.
        # 2 should clear the entire screen, and move cursor to (1,1)
        handle = win32.STDOUT
        if on_stderr:
            handle = win32.STDERR
        csbi = win32.GetConsoleScreenBufferInfo(handle)
        # get the number of character cells in the current buffer
        cells_in_screen = csbi.dwSize.X * csbi.dwSize.Y
        # get number of character cells before current cursor position
        cells_before_cursor = csbi.dwSize.X * csbi.dwCursorPosition.Y + csbi.dwCursorPosition.X
        if mode == 0:
            from_coord = csbi.dwCursorPosition
            cells_to_erase = cells_in_screen - cells_before_cursor
        elif mode == 1:
            from_coord = win32.COORD(0, 0)
            cells_to_erase = cells_before_cursor
        elif mode == 2:
            from_coord = win32.COORD(0, 0)
            cells_to_erase = cells_in_screen
        else:
            # invalid mode
            return
        # fill the entire screen with blanks
        win32.FillConsoleOutputCharacter(handle, ' ', cells_to_erase, from_coord)
        # now set the buffer's attributes accordingly
        win32.FillConsoleOutputAttribute(handle, self.get_attrs(), cells_to_erase, from_coord)
        if mode == 2:
            # put the cursor where needed
            win32.SetConsoleCursorPosition(handle, (1, 1))

    def erase_line(self, mode=0, on_stderr=False):
        # 0 should clear from the cursor to the end of the line.
        # 1 should clear from the cursor to the beginning of the line.
        # 2 should clear the entire line.
        handle = win32.STDOUT
        if on_stderr:
            handle = win32.STDERR
        csbi = win32.GetConsoleScreenBufferInfo(handle)
        if mode == 0:
            from_coord = csbi.dwCursorPosition
            cells_to_erase = csbi.dwSize.X - csbi.dwCursorPosition.X
        elif mode == 1:
            from_coord = win32.COORD(0, csbi.dwCursorPosition.Y)
            cells_to_erase = csbi.dwCursorPosition.X
        elif mode == 2:
            from_coord = win32.COORD(0, csbi.dwCursorPosition.Y)
            cells_to_erase = csbi.dwSize.X
        else:
            # invalid mode
            return
        # fill the entire screen with blanks
        win32.FillConsoleOutputCharacter(handle, ' ', cells_to_erase, from_coord)
        # now set the buffer's attributes accordingly
        win32.FillConsoleOutputAttribute(handle, self.get_attrs(), cells_to_erase, from_coord)

    def set_title(self, title):
        win32.SetConsoleTitle(title)
site-packages/pip/_vendor/colorama/ansi.py000064400000004732150564745010014626 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.
'''
This module generates ANSI character codes to printing colors to terminals.
See: http://en.wikipedia.org/wiki/ANSI_escape_code
'''

CSI = '\033['
OSC = '\033]'
BEL = '\a'


def code_to_chars(code):
    return CSI + str(code) + 'm'

def set_title(title):
    return OSC + '2;' + title + BEL

def clear_screen(mode=2):
    return CSI + str(mode) + 'J'

def clear_line(mode=2):
    return CSI + str(mode) + 'K'


class AnsiCodes(object):
    def __init__(self):
        # the subclasses declare class attributes which are numbers.
        # Upon instantiation we define instance attributes, which are the same
        # as the class attributes but wrapped with the ANSI escape sequence
        for name in dir(self):
            if not name.startswith('_'):
                value = getattr(self, name)
                setattr(self, name, code_to_chars(value))


class AnsiCursor(object):
    def UP(self, n=1):
        return CSI + str(n) + 'A'
    def DOWN(self, n=1):
        return CSI + str(n) + 'B'
    def FORWARD(self, n=1):
        return CSI + str(n) + 'C'
    def BACK(self, n=1):
        return CSI + str(n) + 'D'
    def POS(self, x=1, y=1):
        return CSI + str(y) + ';' + str(x) + 'H'


class AnsiFore(AnsiCodes):
    BLACK           = 30
    RED             = 31
    GREEN           = 32
    YELLOW          = 33
    BLUE            = 34
    MAGENTA         = 35
    CYAN            = 36
    WHITE           = 37
    RESET           = 39

    # These are fairly well supported, but not part of the standard.
    LIGHTBLACK_EX   = 90
    LIGHTRED_EX     = 91
    LIGHTGREEN_EX   = 92
    LIGHTYELLOW_EX  = 93
    LIGHTBLUE_EX    = 94
    LIGHTMAGENTA_EX = 95
    LIGHTCYAN_EX    = 96
    LIGHTWHITE_EX   = 97


class AnsiBack(AnsiCodes):
    BLACK           = 40
    RED             = 41
    GREEN           = 42
    YELLOW          = 43
    BLUE            = 44
    MAGENTA         = 45
    CYAN            = 46
    WHITE           = 47
    RESET           = 49

    # These are fairly well supported, but not part of the standard.
    LIGHTBLACK_EX   = 100
    LIGHTRED_EX     = 101
    LIGHTGREEN_EX   = 102
    LIGHTYELLOW_EX  = 103
    LIGHTBLUE_EX    = 104
    LIGHTMAGENTA_EX = 105
    LIGHTCYAN_EX    = 106
    LIGHTWHITE_EX   = 107


class AnsiStyle(AnsiCodes):
    BRIGHT    = 1
    DIM       = 2
    NORMAL    = 22
    RESET_ALL = 0

Fore   = AnsiFore()
Back   = AnsiBack()
Style  = AnsiStyle()
Cursor = AnsiCursor()
site-packages/pip/_vendor/colorama/ansitowin32.py000064400000024425150564745010016055 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.
import re
import sys
import os

from .ansi import AnsiFore, AnsiBack, AnsiStyle, Style, BEL
from .winterm import WinTerm, WinColor, WinStyle
from .win32 import windll, winapi_test


winterm = None
if windll is not None:
    winterm = WinTerm()


class StreamWrapper(object):
    '''
    Wraps a stream (such as stdout), acting as a transparent proxy for all
    attribute access apart from method 'write()', which is delegated to our
    Converter instance.
    '''
    def __init__(self, wrapped, converter):
        # double-underscore everything to prevent clashes with names of
        # attributes on the wrapped stream object.
        self.__wrapped = wrapped
        self.__convertor = converter

    def __getattr__(self, name):
        return getattr(self.__wrapped, name)

    def __enter__(self, *args, **kwargs):
        # special method lookup bypasses __getattr__/__getattribute__, see
        # https://stackoverflow.com/questions/12632894/why-doesnt-getattr-work-with-exit
        # thus, contextlib magic methods are not proxied via __getattr__
        return self.__wrapped.__enter__(*args, **kwargs)

    def __exit__(self, *args, **kwargs):
        return self.__wrapped.__exit__(*args, **kwargs)

    def write(self, text):
        self.__convertor.write(text)

    def isatty(self):
        stream = self.__wrapped
        if 'PYCHARM_HOSTED' in os.environ:
            if stream is not None and (stream is sys.__stdout__ or stream is sys.__stderr__):
                return True
        try:
            stream_isatty = stream.isatty
        except AttributeError:
            return False
        else:
            return stream_isatty()

    @property
    def closed(self):
        stream = self.__wrapped
        try:
            return stream.closed
        except AttributeError:
            return True


class AnsiToWin32(object):
    '''
    Implements a 'write()' method which, on Windows, will strip ANSI character
    sequences from the text, and if outputting to a tty, will convert them into
    win32 function calls.
    '''
    ANSI_CSI_RE = re.compile('\001?\033\\[((?:\\d|;)*)([a-zA-Z])\002?')   # Control Sequence Introducer
    ANSI_OSC_RE = re.compile('\001?\033\\]([^\a]*)(\a)\002?')             # Operating System Command

    def __init__(self, wrapped, convert=None, strip=None, autoreset=False):
        # The wrapped stream (normally sys.stdout or sys.stderr)
        self.wrapped = wrapped

        # should we reset colors to defaults after every .write()
        self.autoreset = autoreset

        # create the proxy wrapping our output stream
        self.stream = StreamWrapper(wrapped, self)

        on_windows = os.name == 'nt'
        # We test if the WinAPI works, because even if we are on Windows
        # we may be using a terminal that doesn't support the WinAPI
        # (e.g. Cygwin Terminal). In this case it's up to the terminal
        # to support the ANSI codes.
        conversion_supported = on_windows and winapi_test()

        # should we strip ANSI sequences from our output?
        if strip is None:
            strip = conversion_supported or (not self.stream.closed and not self.stream.isatty())
        self.strip = strip

        # should we should convert ANSI sequences into win32 calls?
        if convert is None:
            convert = conversion_supported and not self.stream.closed and self.stream.isatty()
        self.convert = convert

        # dict of ansi codes to win32 functions and parameters
        self.win32_calls = self.get_win32_calls()

        # are we wrapping stderr?
        self.on_stderr = self.wrapped is sys.stderr

    def should_wrap(self):
        '''
        True if this class is actually needed. If false, then the output
        stream will not be affected, nor will win32 calls be issued, so
        wrapping stdout is not actually required. This will generally be
        False on non-Windows platforms, unless optional functionality like
        autoreset has been requested using kwargs to init()
        '''
        return self.convert or self.strip or self.autoreset

    def get_win32_calls(self):
        if self.convert and winterm:
            return {
                AnsiStyle.RESET_ALL: (winterm.reset_all, ),
                AnsiStyle.BRIGHT: (winterm.style, WinStyle.BRIGHT),
                AnsiStyle.DIM: (winterm.style, WinStyle.NORMAL),
                AnsiStyle.NORMAL: (winterm.style, WinStyle.NORMAL),
                AnsiFore.BLACK: (winterm.fore, WinColor.BLACK),
                AnsiFore.RED: (winterm.fore, WinColor.RED),
                AnsiFore.GREEN: (winterm.fore, WinColor.GREEN),
                AnsiFore.YELLOW: (winterm.fore, WinColor.YELLOW),
                AnsiFore.BLUE: (winterm.fore, WinColor.BLUE),
                AnsiFore.MAGENTA: (winterm.fore, WinColor.MAGENTA),
                AnsiFore.CYAN: (winterm.fore, WinColor.CYAN),
                AnsiFore.WHITE: (winterm.fore, WinColor.GREY),
                AnsiFore.RESET: (winterm.fore, ),
                AnsiFore.LIGHTBLACK_EX: (winterm.fore, WinColor.BLACK, True),
                AnsiFore.LIGHTRED_EX: (winterm.fore, WinColor.RED, True),
                AnsiFore.LIGHTGREEN_EX: (winterm.fore, WinColor.GREEN, True),
                AnsiFore.LIGHTYELLOW_EX: (winterm.fore, WinColor.YELLOW, True),
                AnsiFore.LIGHTBLUE_EX: (winterm.fore, WinColor.BLUE, True),
                AnsiFore.LIGHTMAGENTA_EX: (winterm.fore, WinColor.MAGENTA, True),
                AnsiFore.LIGHTCYAN_EX: (winterm.fore, WinColor.CYAN, True),
                AnsiFore.LIGHTWHITE_EX: (winterm.fore, WinColor.GREY, True),
                AnsiBack.BLACK: (winterm.back, WinColor.BLACK),
                AnsiBack.RED: (winterm.back, WinColor.RED),
                AnsiBack.GREEN: (winterm.back, WinColor.GREEN),
                AnsiBack.YELLOW: (winterm.back, WinColor.YELLOW),
                AnsiBack.BLUE: (winterm.back, WinColor.BLUE),
                AnsiBack.MAGENTA: (winterm.back, WinColor.MAGENTA),
                AnsiBack.CYAN: (winterm.back, WinColor.CYAN),
                AnsiBack.WHITE: (winterm.back, WinColor.GREY),
                AnsiBack.RESET: (winterm.back, ),
                AnsiBack.LIGHTBLACK_EX: (winterm.back, WinColor.BLACK, True),
                AnsiBack.LIGHTRED_EX: (winterm.back, WinColor.RED, True),
                AnsiBack.LIGHTGREEN_EX: (winterm.back, WinColor.GREEN, True),
                AnsiBack.LIGHTYELLOW_EX: (winterm.back, WinColor.YELLOW, True),
                AnsiBack.LIGHTBLUE_EX: (winterm.back, WinColor.BLUE, True),
                AnsiBack.LIGHTMAGENTA_EX: (winterm.back, WinColor.MAGENTA, True),
                AnsiBack.LIGHTCYAN_EX: (winterm.back, WinColor.CYAN, True),
                AnsiBack.LIGHTWHITE_EX: (winterm.back, WinColor.GREY, True),
            }
        return dict()

    def write(self, text):
        if self.strip or self.convert:
            self.write_and_convert(text)
        else:
            self.wrapped.write(text)
            self.wrapped.flush()
        if self.autoreset:
            self.reset_all()


    def reset_all(self):
        if self.convert:
            self.call_win32('m', (0,))
        elif not self.strip and not self.stream.closed:
            self.wrapped.write(Style.RESET_ALL)


    def write_and_convert(self, text):
        '''
        Write the given text to our wrapped stream, stripping any ANSI
        sequences from the text, and optionally converting them into win32
        calls.
        '''
        cursor = 0
        text = self.convert_osc(text)
        for match in self.ANSI_CSI_RE.finditer(text):
            start, end = match.span()
            self.write_plain_text(text, cursor, start)
            self.convert_ansi(*match.groups())
            cursor = end
        self.write_plain_text(text, cursor, len(text))


    def write_plain_text(self, text, start, end):
        if start < end:
            self.wrapped.write(text[start:end])
            self.wrapped.flush()


    def convert_ansi(self, paramstring, command):
        if self.convert:
            params = self.extract_params(command, paramstring)
            self.call_win32(command, params)


    def extract_params(self, command, paramstring):
        if command in 'Hf':
            params = tuple(int(p) if len(p) != 0 else 1 for p in paramstring.split(';'))
            while len(params) < 2:
                # defaults:
                params = params + (1,)
        else:
            params = tuple(int(p) for p in paramstring.split(';') if len(p) != 0)
            if len(params) == 0:
                # defaults:
                if command in 'JKm':
                    params = (0,)
                elif command in 'ABCD':
                    params = (1,)

        return params


    def call_win32(self, command, params):
        if command == 'm':
            for param in params:
                if param in self.win32_calls:
                    func_args = self.win32_calls[param]
                    func = func_args[0]
                    args = func_args[1:]
                    kwargs = dict(on_stderr=self.on_stderr)
                    func(*args, **kwargs)
        elif command in 'J':
            winterm.erase_screen(params[0], on_stderr=self.on_stderr)
        elif command in 'K':
            winterm.erase_line(params[0], on_stderr=self.on_stderr)
        elif command in 'Hf':     # cursor position - absolute
            winterm.set_cursor_position(params, on_stderr=self.on_stderr)
        elif command in 'ABCD':   # cursor position - relative
            n = params[0]
            # A - up, B - down, C - forward, D - back
            x, y = {'A': (0, -n), 'B': (0, n), 'C': (n, 0), 'D': (-n, 0)}[command]
            winterm.cursor_adjust(x, y, on_stderr=self.on_stderr)


    def convert_osc(self, text):
        for match in self.ANSI_OSC_RE.finditer(text):
            start, end = match.span()
            text = text[:start] + text[end:]
            paramstring, command = match.groups()
            if command == BEL:
                if paramstring.count(";") == 1:
                    params = paramstring.split(";")
                    # 0 - change title and icon (we will only change title)
                    # 1 - change icon (we don't support this)
                    # 2 - change title
                    if params[0] in '02':
                        winterm.set_title(params[1])
        return text
site-packages/pip/_vendor/colorama/__init__.py000064400000000357150564745010015432 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.
from .initialise import init, deinit, reinit, colorama_text
from .ansi import Fore, Back, Style, Cursor
from .ansitowin32 import AnsiToWin32

__version__ = '0.4.4'
site-packages/pip/_vendor/colorama/win32.py000064400000012434150564745010014634 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.

# from winbase.h
STDOUT = -11
STDERR = -12

try:
    import ctypes
    from ctypes import LibraryLoader
    windll = LibraryLoader(ctypes.WinDLL)
    from ctypes import wintypes
except (AttributeError, ImportError):
    windll = None
    SetConsoleTextAttribute = lambda *_: None
    winapi_test = lambda *_: None
else:
    from ctypes import byref, Structure, c_char, POINTER

    COORD = wintypes._COORD

    class CONSOLE_SCREEN_BUFFER_INFO(Structure):
        """struct in wincon.h."""
        _fields_ = [
            ("dwSize", COORD),
            ("dwCursorPosition", COORD),
            ("wAttributes", wintypes.WORD),
            ("srWindow", wintypes.SMALL_RECT),
            ("dwMaximumWindowSize", COORD),
        ]
        def __str__(self):
            return '(%d,%d,%d,%d,%d,%d,%d,%d,%d,%d,%d)' % (
                self.dwSize.Y, self.dwSize.X
                , self.dwCursorPosition.Y, self.dwCursorPosition.X
                , self.wAttributes
                , self.srWindow.Top, self.srWindow.Left, self.srWindow.Bottom, self.srWindow.Right
                , self.dwMaximumWindowSize.Y, self.dwMaximumWindowSize.X
            )

    _GetStdHandle = windll.kernel32.GetStdHandle
    _GetStdHandle.argtypes = [
        wintypes.DWORD,
    ]
    _GetStdHandle.restype = wintypes.HANDLE

    _GetConsoleScreenBufferInfo = windll.kernel32.GetConsoleScreenBufferInfo
    _GetConsoleScreenBufferInfo.argtypes = [
        wintypes.HANDLE,
        POINTER(CONSOLE_SCREEN_BUFFER_INFO),
    ]
    _GetConsoleScreenBufferInfo.restype = wintypes.BOOL

    _SetConsoleTextAttribute = windll.kernel32.SetConsoleTextAttribute
    _SetConsoleTextAttribute.argtypes = [
        wintypes.HANDLE,
        wintypes.WORD,
    ]
    _SetConsoleTextAttribute.restype = wintypes.BOOL

    _SetConsoleCursorPosition = windll.kernel32.SetConsoleCursorPosition
    _SetConsoleCursorPosition.argtypes = [
        wintypes.HANDLE,
        COORD,
    ]
    _SetConsoleCursorPosition.restype = wintypes.BOOL

    _FillConsoleOutputCharacterA = windll.kernel32.FillConsoleOutputCharacterA
    _FillConsoleOutputCharacterA.argtypes = [
        wintypes.HANDLE,
        c_char,
        wintypes.DWORD,
        COORD,
        POINTER(wintypes.DWORD),
    ]
    _FillConsoleOutputCharacterA.restype = wintypes.BOOL

    _FillConsoleOutputAttribute = windll.kernel32.FillConsoleOutputAttribute
    _FillConsoleOutputAttribute.argtypes = [
        wintypes.HANDLE,
        wintypes.WORD,
        wintypes.DWORD,
        COORD,
        POINTER(wintypes.DWORD),
    ]
    _FillConsoleOutputAttribute.restype = wintypes.BOOL

    _SetConsoleTitleW = windll.kernel32.SetConsoleTitleW
    _SetConsoleTitleW.argtypes = [
        wintypes.LPCWSTR
    ]
    _SetConsoleTitleW.restype = wintypes.BOOL

    def _winapi_test(handle):
        csbi = CONSOLE_SCREEN_BUFFER_INFO()
        success = _GetConsoleScreenBufferInfo(
            handle, byref(csbi))
        return bool(success)

    def winapi_test():
        return any(_winapi_test(h) for h in
                   (_GetStdHandle(STDOUT), _GetStdHandle(STDERR)))

    def GetConsoleScreenBufferInfo(stream_id=STDOUT):
        handle = _GetStdHandle(stream_id)
        csbi = CONSOLE_SCREEN_BUFFER_INFO()
        success = _GetConsoleScreenBufferInfo(
            handle, byref(csbi))
        return csbi

    def SetConsoleTextAttribute(stream_id, attrs):
        handle = _GetStdHandle(stream_id)
        return _SetConsoleTextAttribute(handle, attrs)

    def SetConsoleCursorPosition(stream_id, position, adjust=True):
        position = COORD(*position)
        # If the position is out of range, do nothing.
        if position.Y <= 0 or position.X <= 0:
            return
        # Adjust for Windows' SetConsoleCursorPosition:
        #    1. being 0-based, while ANSI is 1-based.
        #    2. expecting (x,y), while ANSI uses (y,x).
        adjusted_position = COORD(position.Y - 1, position.X - 1)
        if adjust:
            # Adjust for viewport's scroll position
            sr = GetConsoleScreenBufferInfo(STDOUT).srWindow
            adjusted_position.Y += sr.Top
            adjusted_position.X += sr.Left
        # Resume normal processing
        handle = _GetStdHandle(stream_id)
        return _SetConsoleCursorPosition(handle, adjusted_position)

    def FillConsoleOutputCharacter(stream_id, char, length, start):
        handle = _GetStdHandle(stream_id)
        char = c_char(char.encode())
        length = wintypes.DWORD(length)
        num_written = wintypes.DWORD(0)
        # Note that this is hard-coded for ANSI (vs wide) bytes.
        success = _FillConsoleOutputCharacterA(
            handle, char, length, start, byref(num_written))
        return num_written.value

    def FillConsoleOutputAttribute(stream_id, attr, length, start):
        ''' FillConsoleOutputAttribute( hConsole, csbi.wAttributes, dwConSize, coordScreen, &cCharsWritten )'''
        handle = _GetStdHandle(stream_id)
        attribute = wintypes.WORD(attr)
        length = wintypes.DWORD(length)
        num_written = wintypes.DWORD(0)
        # Note that this is hard-coded for ANSI (vs wide) bytes.
        return _FillConsoleOutputAttribute(
            handle, attribute, length, start, byref(num_written))

    def SetConsoleTitle(title):
        return _SetConsoleTitleW(title)
site-packages/pip/_vendor/colorama/initialise.py000064400000003573150564745010016030 0ustar00# Copyright Jonathan Hartley 2013. BSD 3-Clause license, see LICENSE file.
import atexit
import contextlib
import sys

from .ansitowin32 import AnsiToWin32


orig_stdout = None
orig_stderr = None

wrapped_stdout = None
wrapped_stderr = None

atexit_done = False


def reset_all():
    if AnsiToWin32 is not None:    # Issue #74: objects might become None at exit
        AnsiToWin32(orig_stdout).reset_all()


def init(autoreset=False, convert=None, strip=None, wrap=True):

    if not wrap and any([autoreset, convert, strip]):
        raise ValueError('wrap=False conflicts with any other arg=True')

    global wrapped_stdout, wrapped_stderr
    global orig_stdout, orig_stderr

    orig_stdout = sys.stdout
    orig_stderr = sys.stderr

    if sys.stdout is None:
        wrapped_stdout = None
    else:
        sys.stdout = wrapped_stdout = \
            wrap_stream(orig_stdout, convert, strip, autoreset, wrap)
    if sys.stderr is None:
        wrapped_stderr = None
    else:
        sys.stderr = wrapped_stderr = \
            wrap_stream(orig_stderr, convert, strip, autoreset, wrap)

    global atexit_done
    if not atexit_done:
        atexit.register(reset_all)
        atexit_done = True


def deinit():
    if orig_stdout is not None:
        sys.stdout = orig_stdout
    if orig_stderr is not None:
        sys.stderr = orig_stderr


@contextlib.contextmanager
def colorama_text(*args, **kwargs):
    init(*args, **kwargs)
    try:
        yield
    finally:
        deinit()


def reinit():
    if wrapped_stdout is not None:
        sys.stdout = wrapped_stdout
    if wrapped_stderr is not None:
        sys.stderr = wrapped_stderr


def wrap_stream(stream, convert, strip, autoreset, wrap):
    if wrap:
        wrapper = AnsiToWin32(stream,
            convert=convert, strip=strip, autoreset=autoreset)
        if wrapper.should_wrap():
            stream = wrapper.stream
    return stream
site-packages/pip/_vendor/msgpack/__pycache__/ext.cpython-39.pyc000064400000014257150564745010020616 0ustar00a

��Re��@s�ddlmZddlZddlZddlZejddkZerDeefZ	dZ
n4eZ	zejjZ
Wn"e
yve�e�d��Z
Yn0Gdd�dedd��ZGdd�de�ZdS)	�)�
namedtupleN�cs eZdZdZ�fdd�Z�ZS)�ExtTypez'ExtType represents ext type in msgpack.csVt|t�std��t|t�s$td��d|kr8dksBntd��tt|��|||�S)Nzcode must be intzdata must be bytesr�zcode must be 0~127)�
isinstance�int�	TypeError�bytes�
ValueError�superr�__new__)�cls�code�data��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/msgpack/ext.pyrs

zExtType.__new__)�__name__�
__module__�__qualname__�__doc__r�
__classcell__rrrrrsrz	code datac@s�eZdZdZddgZd dd�Zdd�Zd	d
�Zdd�Zd
d�Z	e
dd��Zdd�Ze
dd��Z
dd�Ze
dd��Zdd�Zdd�Ze
dd��ZdS)!�	TimestampaRTimestamp represents the Timestamp extension type in msgpack.

    When built with Cython, msgpack uses C methods to pack and unpack `Timestamp`. When using pure-Python
    msgpack, :func:`to_bytes` and :func:`from_bytes` are used to pack and unpack `Timestamp`.

    This class is immutable: Do not override seconds and nanoseconds.
    �seconds�nanosecondsrcCsRt|t�std��t|t�s$td��d|kr8dksBntd��||_||_dS)a�Initialize a Timestamp object.

        :param int seconds:
            Number of seconds since the UNIX epoch (00:00:00 UTC Jan 1 1970, minus leap seconds).
            May be negative.

        :param int nanoseconds:
            Number of nanoseconds to add to `seconds` to get fractional time.
            Maximum is 999_999_999.  Default is 0.

        Note: Negative times (before the UNIX epoch) are represented as negative seconds + positive ns.
        zseconds must be an intergerznanoseconds must be an integerr�ʚ;z?nanoseconds must be a non-negative integer less than 999999999.N)r�	int_typesrr
rr)�selfrrrrr�__init__-s

�zTimestamp.__init__cCsd�|j|j�S)z#String representation of Timestamp.z'Timestamp(seconds={0}, nanoseconds={1}))�formatrr�rrrr�__repr__Es�zTimestamp.__repr__cCs*t|�|jur&|j|jko$|j|jkSdS)z0Check for equality with another Timestamp objectF)�typerrr�r�otherrrr�__eq__Ks�zTimestamp.__eq__cCs|�|�S)z(not-equals method (see :func:`__eq__()`))r&r$rrr�__ne__SszTimestamp.__ne__cCst|j|jf�S)N)�hashrrr!rrr�__hash__WszTimestamp.__hash__cCs�t|�dkr"t�d|�d}d}nTt|�dkrPt�d|�d}|d@}|d?}n&t|�dkrnt�d	|�\}}ntd
��t||�S)aUnpack bytes into a `Timestamp` object.

        Used for pure-Python msgpack unpacking.

        :param b: Payload from msgpack ext message with code -1
        :type b: bytes

        :returns: Timestamp object unpacked from msgpack ext payload
        :rtype: Timestamp
        ��!Lr��!Ql���"��!IqzFTimestamp type can only be created from 32, 64, or 96-bit byte objects)�len�struct�unpackr
r)�brr�data64rrr�
from_bytesZs
�zTimestamp.from_bytescCs\|jd?dkrF|jd>|jB}|d@dkr8t�d|�}qXt�d|�}nt�d|j|j�}|S)z�Pack this Timestamp object into bytes.

        Used for pure-Python msgpack packing.

        :returns data: Payload for EXT message with code -1 (timestamp type)
        :rtype: bytes
        r.rl��r+r-r0)rrr2�pack)rr5rrrr�to_bytesuszTimestamp.to_bytescCs&t|d�}t|dd�}t||�S)z�Create a Timestamp from posix timestamp in seconds.

        :param unix_float: Posix timestamp in seconds.
        :type unix_float: int or float.
        �r)rr)�unix_secrrrrr�	from_unix�szTimestamp.from_unixcCs|j|jdS)znGet the timestamp as a floating-point value.

        :returns: posix timestamp
        :rtype: float
        ge��A�rrr!rrr�to_unix�szTimestamp.to_unixcCstt|d��S)z�Create a Timestamp from posix timestamp in nanoseconds.

        :param int unix_ns: Posix timestamp in nanoseconds.
        :rtype: Timestamp
        r)r�divmod)�unix_nsrrr�from_unix_nano�szTimestamp.from_unix_nanocCs|jd|jS)z~Get the timestamp as a unixtime in nanoseconds.

        :returns: posix timestamp in nanoseconds
        :rtype: int
        rr<r!rrr�to_unix_nano�szTimestamp.to_unix_nanocCstj�dt�tj|��d�S)zlGet the timestamp as a UTC datetime.

        Python 2 is not supported.

        :rtype: datetime.
        r)r)�datetime�
fromtimestamp�_utc�	timedeltar=r!rrr�to_datetime�s�zTimestamp.to_datetimecCst�|���S)zuCreate a Timestamp from datetime with tzinfo.

        Python 2 is not supported.

        :rtype: Timestamp
        )rr;�	timestamp)�dtrrr�
from_datetime�szTimestamp.from_datetimeN)r)rrrr�	__slots__rr"r&r'r)�staticmethodr6r8r;r=r@rArFrIrrrrr"s&




r)�collectionsrrB�sysr2�version_info�PY2r�longrrD�timezone�utc�AttributeErrorrEr�objectrrrrr�<module>s
site-packages/pip/_vendor/msgpack/__pycache__/fallback.cpython-39.pyc000064400000064215150564745010021554 0ustar00a

��Re���@s�dZddlmZddlZddlZejddkZerDeefZ	dd�Z
neZ	eZe
Zdd�Z
ejdkrpeZd	d
�Zndd
�Zeed�r�dd
lmZzddlmZWney�ddlmZYn0dZGdd�de�ZndZddlmZdd�ZddlmZm Z m!Z!m"Z"m#Z#ddl$m%Z%m&Z&dZ'dZ(dZ)dZ*dZ+dZ,dZ-dZ.dZ/dZ0dZ1e2e3fdd�Z4d d!�Z5d"d#�Z6ejd$k�rzd+d%d&�Z7nej8Z7Gd'd(�d(e�Z9Gd)d*�d*e�Z:dS),z.Fallback pure Python implementation of msgpack�)�datetimeN�cCs|��S�N)�	iteritems��d�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/msgpack/fallback.py�dict_iteritemssr
cCs|��Sr)�itemsrrrr	r
s)��cCs.t|j�dko,t|jdt�o,|jd�d�S)N�rz maximum recursion depth exceeded)�len�args�
isinstance�str�
startswith��errr	�_is_recursionerrors
��rcCsdS)NTrrrrr	r's�pypy_version_info)�newlist_hint)�BytesBuilder)�
StringBuilderTc@s&eZdZd	dd�Zdd�Zdd�ZdS)
�StringIO�cCs,|r tt|��|_|j�|�nt�|_dSr)rr�builder�append��self�srrr	�__init__7szStringIO.__init__cCs6t|t�r|��}nt|t�r&t|�}|j�|�dSr)r�
memoryview�tobytes�	bytearray�bytesrrrrrr	�write>s



zStringIO.writecCs
|j��Sr)r�build�r rrr	�getvalueEszStringIO.getvalueN)r)�__name__�
__module__�__qualname__r"r'r*rrrr	r6s
rF)�BytesIOcCsgSrr)�sizerrr	�<lambda>Mrr0r)�
BufferFull�	OutOfData�	ExtraData�FormatError�
StackError)�ExtType�	Timestampr�r
i�cCs(||�|ur||�|vS||�|uSdSrr)�obj�t�type�tuplerrr	�_check_type_strictdsr=cCst|�}|jdkrtd��|S)Nrz$cannot unpack from multi-byte object)r#�itemsize�
ValueError)r9�viewrrr	�_get_data_from_bufferks
rAc
Ks�tddt|�i|��}|�|�z|��}WnJtyFtd��Yn2tyv}zt|�r`t��WYd}~n
d}~00|�	�r�t
||����|S)aw
    Unpack an object from `packed`.

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``ValueError`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.

    See :class:`Unpacker` for options.
    N�max_buffer_sizezUnpack failed: incomplete input)N)�Unpackerr�feed�_unpackr2r?�RecursionErrorrr5�_got_extradatar3�_get_extradata)�packed�kwargs�unpacker�retrrrr	�unpackbrs
rM)r��cCst�|t|�|�S)z0Explicit type cast for legacy struct.unpack_from)�struct�unpack_fromr&)�f�b�orrr	�_unpack_from�srUc@s�eZdZdZdddddddddddedddddfdd	�Zd
d�Zdd
�Zdd�Zdd�Z	dd�Z
d*dd�Zd+dd�Ze
fdd�Ze
fdd�Zdd�Zdd�ZeZd d!�Zd"d#�Zd$d%�Zd&d'�Zd(d)�ZdS),rCa�Streaming unpacker.

    Arguments:

    :param file_like:
        File-like object having `.read(n)` method.
        If specified, unpacker reads serialized data from it and :meth:`feed()` is not usable.

    :param int read_size:
        Used as `file_like.read(read_size)`. (default: `min(16*1024, max_buffer_size)`)

    :param bool use_list:
        If true, unpack msgpack array to Python list.
        Otherwise, unpack to Python tuple. (default: True)

    :param bool raw:
        If true, unpack msgpack raw to Python bytes.
        Otherwise, unpack to Python str by decoding with UTF-8 encoding (default).

    :param int timestamp:
        Control how timestamp type is unpacked:

            0 - Timestamp
            1 - float  (Seconds from the EPOCH)
            2 - int  (Nanoseconds from the EPOCH)
            3 - datetime.datetime  (UTC).  Python 2 is not supported.

    :param bool strict_map_key:
        If true (default), only str or bytes are accepted for map (dict) keys.

    :param callable object_hook:
        When specified, it should be callable.
        Unpacker calls it with a dict argument after unpacking msgpack map.
        (See also simplejson)

    :param callable object_pairs_hook:
        When specified, it should be callable.
        Unpacker calls it with a list of key-value pairs after unpacking msgpack map.
        (See also simplejson)

    :param str unicode_errors:
        The error handler for decoding unicode. (default: 'strict')
        This option should be used only when you have msgpack data which
        contains invalid UTF-8 string.

    :param int max_buffer_size:
        Limits size of data waiting unpacked.  0 means 2**32-1.
        The default value is 100*1024*1024 (100MiB).
        Raises `BufferFull` exception when it is insufficient.
        You should set this parameter when unpacking data from untrusted source.

    :param int max_str_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max length of str. (default: max_buffer_size)

    :param int max_bin_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max length of bin. (default: max_buffer_size)

    :param int max_array_len:
        Limits max length of array.
        (default: max_buffer_size)

    :param int max_map_len:
        Limits max length of map.
        (default: max_buffer_size//2)

    :param int max_ext_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max size of ext type.  (default: max_buffer_size)

    Example of streaming deserialize from file-like object::

        unpacker = Unpacker(file_like)
        for o in unpacker:
            process(o)

    Example of streaming deserialize from socket::

        unpacker = Unpacker(max_buffer_size)
        while True:
            buf = sock.recv(1024**2)
            if not buf:
                break
            unpacker.feed(buf)
            for o in unpacker:
                process(o)

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``OutOfData`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.
    NrTFi@���cCs�|
durd}
|durd|_nt|j�s.td��||_d|_t�|_d|_d|_|sVd}|
dkrb|}
|dkrn|}|dkrz|}|dkr�|d}|dkr�|}||_	||j	kr�t
d	��|p�t|j	d
�|_t
|�|_t
|�|_|
|_||_d|kr�dks�nt
d��||_|	|_||_||_||_|
|_||_||_||_||_d|_|	du�r\t|	��s\td
��|du�rxt|��sxtd��|du�r�t|��s�td��|du�r�|du�r�td��t|��s�td��dS)N�strictTz!`file_like.read` must be callableFri���rVrz.read_size must be smaller than max_buffer_sizei@rztimestamp must be 0..3z`list_hook` is not callablez`object_hook` is not callablez#`object_pairs_hook` is not callablez8object_pairs_hook and object_hook are mutually exclusivez`ext_hook` is not callable)�_feeding�callable�read�	TypeError�	file_liker%�_buffer�_buff_i�_buf_checkpoint�_max_buffer_sizer?�min�
_read_size�bool�_raw�_strict_map_key�_unicode_errors�	_use_list�
_timestamp�
_list_hook�_object_hook�_object_pairs_hook�	_ext_hook�_max_str_len�_max_bin_len�_max_array_len�_max_map_len�_max_ext_len�_stream_offset)r r\�	read_size�use_list�raw�	timestamp�strict_map_key�object_hook�object_pairs_hook�	list_hook�unicode_errorsrB�ext_hook�max_str_len�max_bin_len�
max_array_len�max_map_len�max_ext_lenrrr	r"�sp
	


�
zUnpacker.__init__cCsr|js
J�t|�}t|j�|jt|�|jkr4t�|jdkrb|jd|j�=|j|j8_d|_|j�|�dS�Nr)	rXrArr]r^r`r1r_�extend)r �
next_bytesr@rrr	rDUs

z
Unpacker.feedcCs"|j|j|j7_|j|_dS)z+ Gets rid of the used parts of the buffer. N)rrr^r_r)rrr	�_consumedszUnpacker._consumecCs|jt|j�kSr)r^rr]r)rrr	rGiszUnpacker._got_extradatacCs|j|jd�Sr)r]r^r)rrr	rHlszUnpacker._get_extradatacCs|j|dd�}|��|S)NF��raise_outofdata)�_readr��r �nrLrrr	�
read_bytesoszUnpacker.read_bytescCs8|j||d�|j}|j|||�}|t|�|_|S)Nr�)�_reserver^r]r)r r�r��irLrrr	r�ts
zUnpacker._readcCs�t|j�|j|}|dkr dS|jr2|j|_t�|jdkr`|jd|j�=|j|j8_d|_|}|dkr�t|j|�}|j�	|�}|s�q�t
|t�s�J�|j|7_|t|�8}qft|j�||jkr�|r�d|_t�dSr�)rr]r^rXr_r2�maxrbr\rZrr&)r r�r��remain_bytes�
to_read_bytes�	read_datarrr	r�|s,
zUnpacker._reservecCs&
t}d}d}|�d�|j|j}|jd7_|d@dkrD|}�	n�|d@dkr`d|dA}�	n�|d@dkr�|d@}t}||jkr�td	||j��|�|�}�	n~|d
@dkr�|d@}t}||j	kr�td
||j	���	nJ|d
@dk�r
|d@}t
}||jk�
rtd||j���	n|dk�rd}�	n|dk�r.d}�n�|dk�r@d}�n�|dk�r�t}|�d�|j|j}|jd7_||j
k�r�td||j
f��|�|�}�n~|dk�rt}|�d�td|j|j�d}|jd7_||j
k�r�td||j
f��|�|�}�n|dk�rjt}|�d�td|j|j�d}|jd7_||j
k�r\td||j
f��|�|�}�n�|dk�r�t}|�d�td|j|j�\}}|jd7_||jk�r�td||jf��|�|�}�nL|dk�r6t}|�d �td!|j|j�\}}|jd 7_||jk�r(td||jf��|�|�}�n�|d"k�r�t}|�d#�td$|j|j�\}}|jd#7_||jk�r�td||jf��|�|�}�n�|d%k�r�|�d�td&|j|j�d}|jd7_�nF|d'k�r|�d(�td)|j|j�d}|jd(7_�n|d*k�rB|�d�|j|j}|jd7_�n�|d+k�r||�d�td|j|j�d}|jd7_�n�|d,k�r�|�d�td|j|j�d}|jd7_�nf|d-k�r�|�d(�td.|j|j�d}|jd(7_�n,|d/k�r*|�d�td0|j|j�d}|jd7_�n�|d1k�rd|�d�td2|j|j�d}|jd7_�n�|d3k�r�|�d�td4|j|j�d}|jd7_�n~|d5k�r�|�d(�td6|j|j�d}|jd(7_�nD|d7k�r4t}|jdk�rtdd|jf��|�d�td8|j|j�\}}|jd7_�n�|d9k�r�t}|jdk�r`tdd|jf��|�d �td:|j|j�\}}|jd 7_�n�|d;k�r�t}|jdk�r�tdd|jf��|�d#�td<|j|j�\}}|jd#7_�n0|d=k�rHt}|jd(k�rtdd(|jf��|�d>�td?|j|j�\}}|jd>7_�n�|d@k�r�t}|jdAk�rttddA|jf��|�dB�tdC|j|j�\}}|jdB7_�nx|dDk�r�t}|�d�|j|j}|jd7_||jk�r�td	||j��|�|�}�n|dEk�r^t}|�d�td|j|j�\}|jd7_||jk�rPtd	||j��|�|�}�n�|dFk�r�t}|�d�td|j|j�\}|jd7_||jk�r�td	||j��|�|�}�n^|dGk�	rt}|�d�td|j|j�\}|jd7_||j	k�
rtd
||j	���n|dHk�	rht}|�d�td|j|j�\}|jd7_||j	k�
rtd
||j	��n�|dIk�	r�|�d�td|j|j�\}|jd7_||jk�	r�td||j��t
}n`|dJk�
r|�d�td|j|j�\}|jd7_||jk�
r
td||j��t
}ntdK|��|||fS)LNrr���rV���z%s exceeds max_str_len(%s)����z%s exceeds max_array_len(%s)z%s exceeds max_map_len(%s)���F��T��z%s exceeds max_bin_len(%s)��r�>H��r8�>I���Bbz%s exceeds max_ext_len(%s)��rz>Hb��r
z>Ib��z>f���z>d��������z>Q��rS��z>h��z>i��z>q���b1s���b2s���b4s���	�b8s�����b16s��������������zUnknown header: 0x%x)�TYPE_IMMEDIATEr�r]r^�TYPE_RAWrmr?r��
TYPE_ARRAYro�TYPE_MAPrp�TYPE_BINrnrU�TYPE_EXTrqr4)r �execute�typr�r9rS�Lrrr	�_read_header�s�





























































zUnpacker._read_headerc
s���|�\}}}|tkr,|tkr(td��|S|tkrH|tkrDtd��|S|tkr�|tkrtt|�D]}��t�q`dSt	|�}t|�D]}|�
��t��q��jdur���|�}�j
r�|St|�S|tk�r�|tkr�t|�D]}��t���t�q�dS�jdu�r"���fdd�t|�D��}n�i}t|�D]j}��t�}�j�rjt|�ttfv�rjtdtt|����t�s�t|�tu�r�t�|�}��t�||<�q.�jdu�r���|�}|S|tk�r�dS|tk�r�j�r�t|�}n|�d�j�}|S|tk�rt|�S|tk�rv|dk�rft� t|��}	�j!dk�r8|	�"�S�j!d	k�rL|	�#�S�j!d
k�r`|	�$�S|	Sn��%|t|��S|t&k�s�J�|S)NzExpected arrayzExpected mapc3s"|]}��t���t�fVqdSr)rE�EX_CONSTRUCT)�.0�_r)rr	�	<genexpr>�s�z#Unpacker._unpack.<locals>.<genexpr>z%s is not allowed for map key�utf_8rVrrr)'r��EX_READ_ARRAY_HEADERr�r?�EX_READ_MAP_HEADERr��EX_SKIP�xrangerErrr�rirgr<rkrer;�unicoder&r�PY2�sys�internrjr�rd�decoderfr�r�r7�
from_bytesrh�to_unix�to_unix_nano�to_datetimerlr�)
r r�r�r�r9r�rLr��key�tsrr)r	rEps�



�

�







zUnpacker._unpackcCs|Srrr)rrr	�__iter__�szUnpacker.__iter__cCsPz|�t�}|��|WSty6|��t�YntyJt�Yn0dSr)rEr�r�r2�
StopIterationrFr5�r rLrrr	�__next__�s
zUnpacker.__next__cCs|�t�|��dSr)rEr�r�r)rrr	�skip�s
z
Unpacker.skipcCs2z|�t�}Wnty$t�Yn0|��|Sr)rEr�rFr5r�r�rrr	�unpack�s
zUnpacker.unpackcCs|�t�}|��|Sr)rEr�r�r�rrr	�read_array_header�s
zUnpacker.read_array_headercCs|�t�}|��|Sr)rEr�r�r�rrr	�read_map_header�s
zUnpacker.read_map_headercCs|jSr)rrr)rrr	�tell�sz
Unpacker.tell)T)T)r+r,r-�__doc__r6r"rDr�rGrHr�r�r�r�r�rEr�r��nextr�r�r�r�r�rrrr	rC�sJa�
]

 UMrCc@s�eZdZdZd#dd�Zeeefdd�Zd	d
�Z	dd�Z
d
d�Zdd�Zdd�Z
dd�Zdd�Zefdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�ZdS)$�Packera8	
    MessagePack Packer

    Usage::

        packer = Packer()
        astream.write(packer.pack(a))
        astream.write(packer.pack(b))

    Packer's constructor has some keyword arguments:

    :param callable default:
        Convert user type to builtin type that Packer supports.
        See also simplejson's document.

    :param bool use_single_float:
        Use single precision float type for float. (default: False)

    :param bool autoreset:
        Reset buffer after each pack and return its content as `bytes`. (default: True).
        If set this to false, use `bytes()` to get content and `.reset()` to clear buffer.

    :param bool use_bin_type:
        Use bin type introduced in msgpack spec 2.0 for bytes.
        It also enables str8 type for unicode. (default: True)

    :param bool strict_types:
        If set to true, types will be checked to be exact. Derived classes
        from serializable types will not be serialized and will be
        treated as unsupported type and forwarded to default.
        Additionally tuples will not be serialized as lists.
        This is useful when trying to implement accurate serialization
        for python types.

    :param bool datetime:
        If set to true, datetime with tzinfo is packed into Timestamp type.
        Note that the tzinfo is stripped in the timestamp.
        You can get UTC datetime with `timestamp=3` option of the Unpacker.
        (Python 2 is not supported).

    :param str unicode_errors:
        The error handler for encoding unicode. (default: 'strict')
        DO NOT USE THIS!!  This option is kept for very specific usage.

    Example of streaming deserialize from file-like object::

        unpacker = Unpacker(file_like)
        for o in unpacker:
            process(o)

    Example of streaming deserialize from socket::

        unpacker = Unpacker()
        while True:
            buf = sock.recv(1024**2)
            if not buf:
                break
            unpacker.feed(buf)
            for o in unpacker:
                process(o)

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``OutOfData`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.
    NFTcCsf||_||_||_||_t�|_tr0|r0td��t|�|_	|p@d|_
|dur\t|�s\td��||_
dS)Nz%datetime is not supported in Python 2rWzdefault must be callable)�
_strict_types�
_use_float�
_autoreset�
_use_bin_typerr]r�r?rc�	_datetimerfrYr[�_default)r �default�use_single_float�	autoreset�use_bin_type�strict_typesrr{rrr	r",s


zPacker.__init__cCs�d}|jr|}t}nttf}|dkr,td��|dur@|j�d�S||t�rf|rZ|j�d�S|j�d�S||t��r�d|kr�dkr�nn|j�t�	d|��Sd	|kr�dkr�nn|j�t�	d
|��Sd|kr�dkr�nn|j�t�	dd
|��Sd|k�rdk�r*nn|j�t�	dd|��Sd|k�rBdk�r\nn|j�t�	dd|��Sd|k�rtdk�r�nn|j�t�	dd|��Sd|k�r�dk�r�nn|j�t�	dd|��Sd|k�r�dk�r�nn|j�t�	dd|��Sd|k�r
dk�r$nn|j�t�	dd|��Sd |k�r<dk�rVnn|j�t�	d!d"|��S|�sx|j
du�rx|�
|�}d#}qtd$��||tt
f��r�t|�}|d%k�r�td&t|�j��|�|�|j�|�S||t��r|�d'|j�}t|�}|d%k�r�td(��|�|�|j�|�S||t��rVt|�|j}|d%k�r@td)��|�|�|j�|�S||t��r�|j�r�|j�t�	d*d+|��S|j�t�	d,d-|��S||ttf��r�||t��r�d.}|��}	n|j}|j}	t|t ��s�J�t|	t��s�J�t|	�}
|
d/k�r|j�d0�n�|
d1k�r$|j�d2�n�|
d3k�r<|j�d4�n�|
d5k�rT|j�d6�nr|
d7k�rl|j�d8�nZ|
dk�r�|j�t�	d9d:|
��n8|
dk�r�|j�t�	dd;|
��n|j�t�	dd<|
��|j�t�	d
|��|j�|	�dS|||��r0t|�}|�!|�t"|�D]}|�#|||d/��qdS||t$��rV|�%t|�t&|�|d/�S|j'�r�||t(��r�|j)du�r�t�*|�}d/}q|�s�|j
du�r�|�
|�}d/}qt+d=|f��qdS)>NFrzrecursion limit exceeded������r��Bi��rSr��BBr�i����z>Bbr���>BHr�i���z>Bhr�����>BIr�i�z>Bir�l����z>BQr�l����z>Bqr�TzInteger value out of range�z%s is too largezutf-8zString is too largezMemoryview is too largez>Bfr�z>Bdr�rVr��r��r8��r���r���>BBr�r�r�zCannot serialize %r),r��listr<r?r]r'rc�	int_typesrP�packr��
OverflowErrorr&r%rr;r+�_pack_bin_headerr��encoderf�_pack_raw_headerr#r>�floatr�r6r7�to_bytes�code�datar�int�_pack_array_headerr��_pack�dict�_pack_map_pairsr
r��	_DateTime�tzinfo�
from_datetimer[)r r9�
nest_limit�check�check_type_strict�default_used�
list_typesr�rrr�r�rrr	rDs�
















� 

zPacker._packcCsFz|�|�Wnt�|_�Yn0|jrB|j��}t�|_|SdSr)rrr]r�r*)r r9rLrrr	r�s
zPacker.packcCs0|�t|�|�|jr,|j��}t�|_|SdSr)rrr�r]r*r)r �pairsrLrrr	�pack_map_pairs�s

zPacker.pack_map_pairscCs6|dkrt�|�|�|jr2|j��}t�|_|SdS�Nr)r?rr�r]r*rr�rrr	�pack_array_header�s

zPacker.pack_array_headercCs6|dkrt�|�|�|jr2|j��}t�|_|SdSr$)r?�_pack_map_headerr�r]r*rr�rrr	�pack_map_header�s

zPacker.pack_map_headercCsJt|t�std��d|kr&dks0ntd��t|t�sBtd��t|�}|dkrZtd��|dkrp|j�d	�n�|d
kr�|j�d�n�|dkr�|j�d
�n�|dkr�|j�d�nt|dkr�|j�d�n^|dkr�|j�dt�	d|��n<|dk�r|j�dt�	d|��n|j�dt�	d|��|j�t�	d|��|j�|�dS)Nztypecode must have int type.r�ztypecode should be 0-127zdata must have bytes typerzToo large datarrrrr8rr�rr�rr���r�r���r���r�)
rrr[r?r&rr]r'rPr)r �typecoderr�rrr	�
pack_ext_type�s4


zPacker.pack_ext_typecCsh|dkr |j�t�dd|��S|dkr>|j�t�dd|��S|dkr\|j�t�dd	|��Std
��dS)Nr�r�r�r�rr�rrr�zArray is too large�r]r'rPrr?�r r�rrr	r�szPacker._pack_array_headercCsh|dkr |j�t�dd|��S|dkr>|j�t�dd|��S|dkr\|j�t�dd	|��Std
��dS)Nr�r�r�r�rr�rrr�zDict is too larger.r/rrr	r&szPacker._pack_map_headercCs<|�|�|D](\}}|�||d�|�||d�qdS)Nr)r&r)r r�r"r�k�vrrr	rs
zPacker._pack_map_pairscCs�|dkr"|j�t�dd|��nn|jrH|dkrH|j�t�dd|��nH|dkrh|j�t�dd	|��n(|d
kr�|j�t�dd|��ntd
��dS)Nr�r�r�r�r	r�r�rr�rrr�zRaw is too large)r]r'rPrr�r?r/rrr	rszPacker._pack_raw_headercCsv|js|�|�S|dkr.|j�t�dd|��S|dkrL|j�t�dd|��S|dkrj|j�t�dd	|��Std
��dS)Nr�r	r�r�rr�rrr�zBin is too large)r�rr]r'rPrr?r/rrr	r#s
zPacker._pack_bin_headercCs
|j��S)z/Return internal buffer contents as bytes object)r]r*r)rrr	r&/szPacker.bytescCst�|_dS)zYReset internal buffer.

        This method is useful only when autoreset=False.
        N)rr]r)rrr	�reset3szPacker.resetcCs"tstrt|���S|j��SdS)zReturn view of internal buffer.N)�USING_STRINGBUILDERr�r#r&r]�	getbufferr)rrr	r4:szPacker.getbuffer)NFTTFFN)r+r,r-r�r"�DEFAULT_RECURSE_LIMITrr=rrr#r%r'r-rr&rrrr&r2r4rrrr	r��s4F�
�
z				r�)r);r�rrr�rP�version_infor�r�longrr
rr��ranger��RuntimeErrorrFr�hasattrZ__pypy__rZ__pypy__.buildersrr�ImportErrorr3�objectr�ior.�
exceptionsr1r2r3r4r5�extr6r7r�r�r�r�r�r�r�r�r�r�r5r;r<r=rArMrUrQrCr�rrrr	�<module>sf




Ssite-packages/pip/_vendor/msgpack/__pycache__/exceptions.cpython-39.pyc000064400000003543150564745010022173 0ustar00a

��Re9�@sxGdd�de�ZGdd�de�ZGdd�de�ZGdd�dee�ZGdd	�d	ee�ZeZGd
d�de�ZeZ	eZ
eZdS)
c@seZdZdZdS)�UnpackExceptionz�Base class for some exceptions raised while unpacking.

    NOTE: unpack may raise exception other than subclass of
    UnpackException.  If you want to catch all error, catch
    Exception instead.
    N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/msgpack/exceptions.pyrsrc@seZdZdS)�
BufferFullN�rrrrrrrr	
sr	c@seZdZdS)�	OutOfDataNr
rrrrrsrc@seZdZdZdS)�FormatErrorzInvalid msgpack formatNrrrrrrsrc@seZdZdZdS)�
StackErrorz
Too nestedNrrrrrr
sr
c@s eZdZdZdd�Zdd�ZdS)�	ExtraDataz�ExtraData is raised when there is trailing data.

    This exception is raised while only one-shot (not streaming)
    unpack.
    cCs||_||_dS)N)�unpacked�extra)�selfrrrrr�__init__%szExtraData.__init__cCsdS)Nzunpack(b) received extra data.r)rrrr�__str__)szExtraData.__str__N)rrrrrrrrrrrsrN)
�	Exceptionrr	r�
ValueErrorrr
�UnpackValueErrorr�
PackException�PackValueError�
OverflowError�PackOverflowErrorrrrr�<module>s	site-packages/pip/_vendor/msgpack/__pycache__/__init__.cpython-39.pyc000064400000002655150564745010021554 0ustar00a

��Re^�@s�ddlmZddlTddlmZmZddlZddlZej�	d�sNej
ddkrdddlmZm
Z
mZn@zddlmZm
Z
mZWn&ey�ddlmZm
Z
mZYn0d	d
�Zdd�Zd
d�ZeZe
ZeZeZdS)�)�version)�*)�ExtType�	Timestamp�NZMSGPACK_PUREPYTHON�)�Packer�unpackb�UnpackercKs"tfi|��}|�|�|��dS)zX
    Pack object `o` and write it to `stream`

    See :class:`Packer` for options.
    N)r�write�pack)�o�stream�kwargs�packer�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/msgpack/__init__.pyrsrcKstfi|���|�S)zW
    Pack object `o` and return packed bytes

    See :class:`Packer` for options.
    )rr)r
rrrr�packbsrcKs|��}t|fi|��S)z�
    Unpack an object from `stream`.

    Raises `ExtraData` when `stream` contains extra bytes.
    See :class:`Unpacker` for options.
    )�readr	)rr�datarrr�unpack&sr)�_versionr�
exceptions�extrr�os�sys�environ�get�version_info�fallbackrr	r
Z	_cmsgpack�ImportErrorrrr�load�loads�dump�dumpsrrrr�<module>s"
	site-packages/pip/_vendor/msgpack/__pycache__/_version.cpython-39.pyc000064400000000401150564745010021624 0ustar00a

��Re�@sdZdS))���N)�version�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/msgpack/_version.py�<module>�site-packages/pip/_vendor/msgpack/ext.py000064400000013710150564745010014320 0ustar00# coding: utf-8
from collections import namedtuple
import datetime
import sys
import struct


PY2 = sys.version_info[0] == 2

if PY2:
    int_types = (int, long)
    _utc = None
else:
    int_types = int
    try:
        _utc = datetime.timezone.utc
    except AttributeError:
        _utc = datetime.timezone(datetime.timedelta(0))


class ExtType(namedtuple("ExtType", "code data")):
    """ExtType represents ext type in msgpack."""

    def __new__(cls, code, data):
        if not isinstance(code, int):
            raise TypeError("code must be int")
        if not isinstance(data, bytes):
            raise TypeError("data must be bytes")
        if not 0 <= code <= 127:
            raise ValueError("code must be 0~127")
        return super(ExtType, cls).__new__(cls, code, data)


class Timestamp(object):
    """Timestamp represents the Timestamp extension type in msgpack.

    When built with Cython, msgpack uses C methods to pack and unpack `Timestamp`. When using pure-Python
    msgpack, :func:`to_bytes` and :func:`from_bytes` are used to pack and unpack `Timestamp`.

    This class is immutable: Do not override seconds and nanoseconds.
    """

    __slots__ = ["seconds", "nanoseconds"]

    def __init__(self, seconds, nanoseconds=0):
        """Initialize a Timestamp object.

        :param int seconds:
            Number of seconds since the UNIX epoch (00:00:00 UTC Jan 1 1970, minus leap seconds).
            May be negative.

        :param int nanoseconds:
            Number of nanoseconds to add to `seconds` to get fractional time.
            Maximum is 999_999_999.  Default is 0.

        Note: Negative times (before the UNIX epoch) are represented as negative seconds + positive ns.
        """
        if not isinstance(seconds, int_types):
            raise TypeError("seconds must be an interger")
        if not isinstance(nanoseconds, int_types):
            raise TypeError("nanoseconds must be an integer")
        if not (0 <= nanoseconds < 10 ** 9):
            raise ValueError(
                "nanoseconds must be a non-negative integer less than 999999999."
            )
        self.seconds = seconds
        self.nanoseconds = nanoseconds

    def __repr__(self):
        """String representation of Timestamp."""
        return "Timestamp(seconds={0}, nanoseconds={1})".format(
            self.seconds, self.nanoseconds
        )

    def __eq__(self, other):
        """Check for equality with another Timestamp object"""
        if type(other) is self.__class__:
            return (
                self.seconds == other.seconds and self.nanoseconds == other.nanoseconds
            )
        return False

    def __ne__(self, other):
        """not-equals method (see :func:`__eq__()`)"""
        return not self.__eq__(other)

    def __hash__(self):
        return hash((self.seconds, self.nanoseconds))

    @staticmethod
    def from_bytes(b):
        """Unpack bytes into a `Timestamp` object.

        Used for pure-Python msgpack unpacking.

        :param b: Payload from msgpack ext message with code -1
        :type b: bytes

        :returns: Timestamp object unpacked from msgpack ext payload
        :rtype: Timestamp
        """
        if len(b) == 4:
            seconds = struct.unpack("!L", b)[0]
            nanoseconds = 0
        elif len(b) == 8:
            data64 = struct.unpack("!Q", b)[0]
            seconds = data64 & 0x00000003FFFFFFFF
            nanoseconds = data64 >> 34
        elif len(b) == 12:
            nanoseconds, seconds = struct.unpack("!Iq", b)
        else:
            raise ValueError(
                "Timestamp type can only be created from 32, 64, or 96-bit byte objects"
            )
        return Timestamp(seconds, nanoseconds)

    def to_bytes(self):
        """Pack this Timestamp object into bytes.

        Used for pure-Python msgpack packing.

        :returns data: Payload for EXT message with code -1 (timestamp type)
        :rtype: bytes
        """
        if (self.seconds >> 34) == 0:  # seconds is non-negative and fits in 34 bits
            data64 = self.nanoseconds << 34 | self.seconds
            if data64 & 0xFFFFFFFF00000000 == 0:
                # nanoseconds is zero and seconds < 2**32, so timestamp 32
                data = struct.pack("!L", data64)
            else:
                # timestamp 64
                data = struct.pack("!Q", data64)
        else:
            # timestamp 96
            data = struct.pack("!Iq", self.nanoseconds, self.seconds)
        return data

    @staticmethod
    def from_unix(unix_sec):
        """Create a Timestamp from posix timestamp in seconds.

        :param unix_float: Posix timestamp in seconds.
        :type unix_float: int or float.
        """
        seconds = int(unix_sec // 1)
        nanoseconds = int((unix_sec % 1) * 10 ** 9)
        return Timestamp(seconds, nanoseconds)

    def to_unix(self):
        """Get the timestamp as a floating-point value.

        :returns: posix timestamp
        :rtype: float
        """
        return self.seconds + self.nanoseconds / 1e9

    @staticmethod
    def from_unix_nano(unix_ns):
        """Create a Timestamp from posix timestamp in nanoseconds.

        :param int unix_ns: Posix timestamp in nanoseconds.
        :rtype: Timestamp
        """
        return Timestamp(*divmod(unix_ns, 10 ** 9))

    def to_unix_nano(self):
        """Get the timestamp as a unixtime in nanoseconds.

        :returns: posix timestamp in nanoseconds
        :rtype: int
        """
        return self.seconds * 10 ** 9 + self.nanoseconds

    def to_datetime(self):
        """Get the timestamp as a UTC datetime.

        Python 2 is not supported.

        :rtype: datetime.
        """
        return datetime.datetime.fromtimestamp(0, _utc) + datetime.timedelta(
            seconds=self.to_unix()
        )

    @staticmethod
    def from_datetime(dt):
        """Create a Timestamp from datetime with tzinfo.

        Python 2 is not supported.

        :rtype: Timestamp
        """
        return Timestamp.from_unix(dt.timestamp())
site-packages/pip/_vendor/msgpack/exceptions.py000064400000002071150564745010015677 0ustar00class UnpackException(Exception):
    """Base class for some exceptions raised while unpacking.

    NOTE: unpack may raise exception other than subclass of
    UnpackException.  If you want to catch all error, catch
    Exception instead.
    """


class BufferFull(UnpackException):
    pass


class OutOfData(UnpackException):
    pass


class FormatError(ValueError, UnpackException):
    """Invalid msgpack format"""


class StackError(ValueError, UnpackException):
    """Too nested"""


# Deprecated.  Use ValueError instead
UnpackValueError = ValueError


class ExtraData(UnpackValueError):
    """ExtraData is raised when there is trailing data.

    This exception is raised while only one-shot (not streaming)
    unpack.
    """

    def __init__(self, unpacked, extra):
        self.unpacked = unpacked
        self.extra = extra

    def __str__(self):
        return "unpack(b) received extra data."


# Deprecated.  Use Exception instead to catch all exception during packing.
PackException = Exception
PackValueError = ValueError
PackOverflowError = OverflowError
site-packages/pip/_vendor/msgpack/fallback.py000064400000112212150564745010015254 0ustar00"""Fallback pure Python implementation of msgpack"""

from datetime import datetime as _DateTime
import sys
import struct


PY2 = sys.version_info[0] == 2
if PY2:
    int_types = (int, long)

    def dict_iteritems(d):
        return d.iteritems()


else:
    int_types = int
    unicode = str
    xrange = range

    def dict_iteritems(d):
        return d.items()


if sys.version_info < (3, 5):
    # Ugly hack...
    RecursionError = RuntimeError

    def _is_recursionerror(e):
        return (
            len(e.args) == 1
            and isinstance(e.args[0], str)
            and e.args[0].startswith("maximum recursion depth exceeded")
        )


else:

    def _is_recursionerror(e):
        return True


if hasattr(sys, "pypy_version_info"):
    # StringIO is slow on PyPy, StringIO is faster.  However: PyPy's own
    # StringBuilder is fastest.
    from __pypy__ import newlist_hint

    try:
        from __pypy__.builders import BytesBuilder as StringBuilder
    except ImportError:
        from __pypy__.builders import StringBuilder
    USING_STRINGBUILDER = True

    class StringIO(object):
        def __init__(self, s=b""):
            if s:
                self.builder = StringBuilder(len(s))
                self.builder.append(s)
            else:
                self.builder = StringBuilder()

        def write(self, s):
            if isinstance(s, memoryview):
                s = s.tobytes()
            elif isinstance(s, bytearray):
                s = bytes(s)
            self.builder.append(s)

        def getvalue(self):
            return self.builder.build()


else:
    USING_STRINGBUILDER = False
    from io import BytesIO as StringIO

    newlist_hint = lambda size: []


from .exceptions import BufferFull, OutOfData, ExtraData, FormatError, StackError

from .ext import ExtType, Timestamp


EX_SKIP = 0
EX_CONSTRUCT = 1
EX_READ_ARRAY_HEADER = 2
EX_READ_MAP_HEADER = 3

TYPE_IMMEDIATE = 0
TYPE_ARRAY = 1
TYPE_MAP = 2
TYPE_RAW = 3
TYPE_BIN = 4
TYPE_EXT = 5

DEFAULT_RECURSE_LIMIT = 511


def _check_type_strict(obj, t, type=type, tuple=tuple):
    if type(t) is tuple:
        return type(obj) in t
    else:
        return type(obj) is t


def _get_data_from_buffer(obj):
    view = memoryview(obj)
    if view.itemsize != 1:
        raise ValueError("cannot unpack from multi-byte object")
    return view


def unpackb(packed, **kwargs):
    """
    Unpack an object from `packed`.

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``ValueError`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.

    See :class:`Unpacker` for options.
    """
    unpacker = Unpacker(None, max_buffer_size=len(packed), **kwargs)
    unpacker.feed(packed)
    try:
        ret = unpacker._unpack()
    except OutOfData:
        raise ValueError("Unpack failed: incomplete input")
    except RecursionError as e:
        if _is_recursionerror(e):
            raise StackError
        raise
    if unpacker._got_extradata():
        raise ExtraData(ret, unpacker._get_extradata())
    return ret


if sys.version_info < (2, 7, 6):

    def _unpack_from(f, b, o=0):
        """Explicit type cast for legacy struct.unpack_from"""
        return struct.unpack_from(f, bytes(b), o)


else:
    _unpack_from = struct.unpack_from


class Unpacker(object):
    """Streaming unpacker.

    Arguments:

    :param file_like:
        File-like object having `.read(n)` method.
        If specified, unpacker reads serialized data from it and :meth:`feed()` is not usable.

    :param int read_size:
        Used as `file_like.read(read_size)`. (default: `min(16*1024, max_buffer_size)`)

    :param bool use_list:
        If true, unpack msgpack array to Python list.
        Otherwise, unpack to Python tuple. (default: True)

    :param bool raw:
        If true, unpack msgpack raw to Python bytes.
        Otherwise, unpack to Python str by decoding with UTF-8 encoding (default).

    :param int timestamp:
        Control how timestamp type is unpacked:

            0 - Timestamp
            1 - float  (Seconds from the EPOCH)
            2 - int  (Nanoseconds from the EPOCH)
            3 - datetime.datetime  (UTC).  Python 2 is not supported.

    :param bool strict_map_key:
        If true (default), only str or bytes are accepted for map (dict) keys.

    :param callable object_hook:
        When specified, it should be callable.
        Unpacker calls it with a dict argument after unpacking msgpack map.
        (See also simplejson)

    :param callable object_pairs_hook:
        When specified, it should be callable.
        Unpacker calls it with a list of key-value pairs after unpacking msgpack map.
        (See also simplejson)

    :param str unicode_errors:
        The error handler for decoding unicode. (default: 'strict')
        This option should be used only when you have msgpack data which
        contains invalid UTF-8 string.

    :param int max_buffer_size:
        Limits size of data waiting unpacked.  0 means 2**32-1.
        The default value is 100*1024*1024 (100MiB).
        Raises `BufferFull` exception when it is insufficient.
        You should set this parameter when unpacking data from untrusted source.

    :param int max_str_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max length of str. (default: max_buffer_size)

    :param int max_bin_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max length of bin. (default: max_buffer_size)

    :param int max_array_len:
        Limits max length of array.
        (default: max_buffer_size)

    :param int max_map_len:
        Limits max length of map.
        (default: max_buffer_size//2)

    :param int max_ext_len:
        Deprecated, use *max_buffer_size* instead.
        Limits max size of ext type.  (default: max_buffer_size)

    Example of streaming deserialize from file-like object::

        unpacker = Unpacker(file_like)
        for o in unpacker:
            process(o)

    Example of streaming deserialize from socket::

        unpacker = Unpacker(max_buffer_size)
        while True:
            buf = sock.recv(1024**2)
            if not buf:
                break
            unpacker.feed(buf)
            for o in unpacker:
                process(o)

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``OutOfData`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.
    """

    def __init__(
        self,
        file_like=None,
        read_size=0,
        use_list=True,
        raw=False,
        timestamp=0,
        strict_map_key=True,
        object_hook=None,
        object_pairs_hook=None,
        list_hook=None,
        unicode_errors=None,
        max_buffer_size=100 * 1024 * 1024,
        ext_hook=ExtType,
        max_str_len=-1,
        max_bin_len=-1,
        max_array_len=-1,
        max_map_len=-1,
        max_ext_len=-1,
    ):
        if unicode_errors is None:
            unicode_errors = "strict"

        if file_like is None:
            self._feeding = True
        else:
            if not callable(file_like.read):
                raise TypeError("`file_like.read` must be callable")
            self.file_like = file_like
            self._feeding = False

        #: array of bytes fed.
        self._buffer = bytearray()
        #: Which position we currently reads
        self._buff_i = 0

        # When Unpacker is used as an iterable, between the calls to next(),
        # the buffer is not "consumed" completely, for efficiency sake.
        # Instead, it is done sloppily.  To make sure we raise BufferFull at
        # the correct moments, we have to keep track of how sloppy we were.
        # Furthermore, when the buffer is incomplete (that is: in the case
        # we raise an OutOfData) we need to rollback the buffer to the correct
        # state, which _buf_checkpoint records.
        self._buf_checkpoint = 0

        if not max_buffer_size:
            max_buffer_size = 2 ** 31 - 1
        if max_str_len == -1:
            max_str_len = max_buffer_size
        if max_bin_len == -1:
            max_bin_len = max_buffer_size
        if max_array_len == -1:
            max_array_len = max_buffer_size
        if max_map_len == -1:
            max_map_len = max_buffer_size // 2
        if max_ext_len == -1:
            max_ext_len = max_buffer_size

        self._max_buffer_size = max_buffer_size
        if read_size > self._max_buffer_size:
            raise ValueError("read_size must be smaller than max_buffer_size")
        self._read_size = read_size or min(self._max_buffer_size, 16 * 1024)
        self._raw = bool(raw)
        self._strict_map_key = bool(strict_map_key)
        self._unicode_errors = unicode_errors
        self._use_list = use_list
        if not (0 <= timestamp <= 3):
            raise ValueError("timestamp must be 0..3")
        self._timestamp = timestamp
        self._list_hook = list_hook
        self._object_hook = object_hook
        self._object_pairs_hook = object_pairs_hook
        self._ext_hook = ext_hook
        self._max_str_len = max_str_len
        self._max_bin_len = max_bin_len
        self._max_array_len = max_array_len
        self._max_map_len = max_map_len
        self._max_ext_len = max_ext_len
        self._stream_offset = 0

        if list_hook is not None and not callable(list_hook):
            raise TypeError("`list_hook` is not callable")
        if object_hook is not None and not callable(object_hook):
            raise TypeError("`object_hook` is not callable")
        if object_pairs_hook is not None and not callable(object_pairs_hook):
            raise TypeError("`object_pairs_hook` is not callable")
        if object_hook is not None and object_pairs_hook is not None:
            raise TypeError(
                "object_pairs_hook and object_hook are mutually " "exclusive"
            )
        if not callable(ext_hook):
            raise TypeError("`ext_hook` is not callable")

    def feed(self, next_bytes):
        assert self._feeding
        view = _get_data_from_buffer(next_bytes)
        if len(self._buffer) - self._buff_i + len(view) > self._max_buffer_size:
            raise BufferFull

        # Strip buffer before checkpoint before reading file.
        if self._buf_checkpoint > 0:
            del self._buffer[: self._buf_checkpoint]
            self._buff_i -= self._buf_checkpoint
            self._buf_checkpoint = 0

        # Use extend here: INPLACE_ADD += doesn't reliably typecast memoryview in jython
        self._buffer.extend(view)

    def _consume(self):
        """ Gets rid of the used parts of the buffer. """
        self._stream_offset += self._buff_i - self._buf_checkpoint
        self._buf_checkpoint = self._buff_i

    def _got_extradata(self):
        return self._buff_i < len(self._buffer)

    def _get_extradata(self):
        return self._buffer[self._buff_i :]

    def read_bytes(self, n):
        ret = self._read(n, raise_outofdata=False)
        self._consume()
        return ret

    def _read(self, n, raise_outofdata=True):
        # (int) -> bytearray
        self._reserve(n, raise_outofdata=raise_outofdata)
        i = self._buff_i
        ret = self._buffer[i : i + n]
        self._buff_i = i + len(ret)
        return ret

    def _reserve(self, n, raise_outofdata=True):
        remain_bytes = len(self._buffer) - self._buff_i - n

        # Fast path: buffer has n bytes already
        if remain_bytes >= 0:
            return

        if self._feeding:
            self._buff_i = self._buf_checkpoint
            raise OutOfData

        # Strip buffer before checkpoint before reading file.
        if self._buf_checkpoint > 0:
            del self._buffer[: self._buf_checkpoint]
            self._buff_i -= self._buf_checkpoint
            self._buf_checkpoint = 0

        # Read from file
        remain_bytes = -remain_bytes
        while remain_bytes > 0:
            to_read_bytes = max(self._read_size, remain_bytes)
            read_data = self.file_like.read(to_read_bytes)
            if not read_data:
                break
            assert isinstance(read_data, bytes)
            self._buffer += read_data
            remain_bytes -= len(read_data)

        if len(self._buffer) < n + self._buff_i and raise_outofdata:
            self._buff_i = 0  # rollback
            raise OutOfData

    def _read_header(self, execute=EX_CONSTRUCT):
        typ = TYPE_IMMEDIATE
        n = 0
        obj = None
        self._reserve(1)
        b = self._buffer[self._buff_i]
        self._buff_i += 1
        if b & 0b10000000 == 0:
            obj = b
        elif b & 0b11100000 == 0b11100000:
            obj = -1 - (b ^ 0xFF)
        elif b & 0b11100000 == 0b10100000:
            n = b & 0b00011111
            typ = TYPE_RAW
            if n > self._max_str_len:
                raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len)
            obj = self._read(n)
        elif b & 0b11110000 == 0b10010000:
            n = b & 0b00001111
            typ = TYPE_ARRAY
            if n > self._max_array_len:
                raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len)
        elif b & 0b11110000 == 0b10000000:
            n = b & 0b00001111
            typ = TYPE_MAP
            if n > self._max_map_len:
                raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len)
        elif b == 0xC0:
            obj = None
        elif b == 0xC2:
            obj = False
        elif b == 0xC3:
            obj = True
        elif b == 0xC4:
            typ = TYPE_BIN
            self._reserve(1)
            n = self._buffer[self._buff_i]
            self._buff_i += 1
            if n > self._max_bin_len:
                raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len))
            obj = self._read(n)
        elif b == 0xC5:
            typ = TYPE_BIN
            self._reserve(2)
            n = _unpack_from(">H", self._buffer, self._buff_i)[0]
            self._buff_i += 2
            if n > self._max_bin_len:
                raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len))
            obj = self._read(n)
        elif b == 0xC6:
            typ = TYPE_BIN
            self._reserve(4)
            n = _unpack_from(">I", self._buffer, self._buff_i)[0]
            self._buff_i += 4
            if n > self._max_bin_len:
                raise ValueError("%s exceeds max_bin_len(%s)" % (n, self._max_bin_len))
            obj = self._read(n)
        elif b == 0xC7:  # ext 8
            typ = TYPE_EXT
            self._reserve(2)
            L, n = _unpack_from("Bb", self._buffer, self._buff_i)
            self._buff_i += 2
            if L > self._max_ext_len:
                raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len))
            obj = self._read(L)
        elif b == 0xC8:  # ext 16
            typ = TYPE_EXT
            self._reserve(3)
            L, n = _unpack_from(">Hb", self._buffer, self._buff_i)
            self._buff_i += 3
            if L > self._max_ext_len:
                raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len))
            obj = self._read(L)
        elif b == 0xC9:  # ext 32
            typ = TYPE_EXT
            self._reserve(5)
            L, n = _unpack_from(">Ib", self._buffer, self._buff_i)
            self._buff_i += 5
            if L > self._max_ext_len:
                raise ValueError("%s exceeds max_ext_len(%s)" % (L, self._max_ext_len))
            obj = self._read(L)
        elif b == 0xCA:
            self._reserve(4)
            obj = _unpack_from(">f", self._buffer, self._buff_i)[0]
            self._buff_i += 4
        elif b == 0xCB:
            self._reserve(8)
            obj = _unpack_from(">d", self._buffer, self._buff_i)[0]
            self._buff_i += 8
        elif b == 0xCC:
            self._reserve(1)
            obj = self._buffer[self._buff_i]
            self._buff_i += 1
        elif b == 0xCD:
            self._reserve(2)
            obj = _unpack_from(">H", self._buffer, self._buff_i)[0]
            self._buff_i += 2
        elif b == 0xCE:
            self._reserve(4)
            obj = _unpack_from(">I", self._buffer, self._buff_i)[0]
            self._buff_i += 4
        elif b == 0xCF:
            self._reserve(8)
            obj = _unpack_from(">Q", self._buffer, self._buff_i)[0]
            self._buff_i += 8
        elif b == 0xD0:
            self._reserve(1)
            obj = _unpack_from("b", self._buffer, self._buff_i)[0]
            self._buff_i += 1
        elif b == 0xD1:
            self._reserve(2)
            obj = _unpack_from(">h", self._buffer, self._buff_i)[0]
            self._buff_i += 2
        elif b == 0xD2:
            self._reserve(4)
            obj = _unpack_from(">i", self._buffer, self._buff_i)[0]
            self._buff_i += 4
        elif b == 0xD3:
            self._reserve(8)
            obj = _unpack_from(">q", self._buffer, self._buff_i)[0]
            self._buff_i += 8
        elif b == 0xD4:  # fixext 1
            typ = TYPE_EXT
            if self._max_ext_len < 1:
                raise ValueError("%s exceeds max_ext_len(%s)" % (1, self._max_ext_len))
            self._reserve(2)
            n, obj = _unpack_from("b1s", self._buffer, self._buff_i)
            self._buff_i += 2
        elif b == 0xD5:  # fixext 2
            typ = TYPE_EXT
            if self._max_ext_len < 2:
                raise ValueError("%s exceeds max_ext_len(%s)" % (2, self._max_ext_len))
            self._reserve(3)
            n, obj = _unpack_from("b2s", self._buffer, self._buff_i)
            self._buff_i += 3
        elif b == 0xD6:  # fixext 4
            typ = TYPE_EXT
            if self._max_ext_len < 4:
                raise ValueError("%s exceeds max_ext_len(%s)" % (4, self._max_ext_len))
            self._reserve(5)
            n, obj = _unpack_from("b4s", self._buffer, self._buff_i)
            self._buff_i += 5
        elif b == 0xD7:  # fixext 8
            typ = TYPE_EXT
            if self._max_ext_len < 8:
                raise ValueError("%s exceeds max_ext_len(%s)" % (8, self._max_ext_len))
            self._reserve(9)
            n, obj = _unpack_from("b8s", self._buffer, self._buff_i)
            self._buff_i += 9
        elif b == 0xD8:  # fixext 16
            typ = TYPE_EXT
            if self._max_ext_len < 16:
                raise ValueError("%s exceeds max_ext_len(%s)" % (16, self._max_ext_len))
            self._reserve(17)
            n, obj = _unpack_from("b16s", self._buffer, self._buff_i)
            self._buff_i += 17
        elif b == 0xD9:
            typ = TYPE_RAW
            self._reserve(1)
            n = self._buffer[self._buff_i]
            self._buff_i += 1
            if n > self._max_str_len:
                raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len)
            obj = self._read(n)
        elif b == 0xDA:
            typ = TYPE_RAW
            self._reserve(2)
            (n,) = _unpack_from(">H", self._buffer, self._buff_i)
            self._buff_i += 2
            if n > self._max_str_len:
                raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len)
            obj = self._read(n)
        elif b == 0xDB:
            typ = TYPE_RAW
            self._reserve(4)
            (n,) = _unpack_from(">I", self._buffer, self._buff_i)
            self._buff_i += 4
            if n > self._max_str_len:
                raise ValueError("%s exceeds max_str_len(%s)", n, self._max_str_len)
            obj = self._read(n)
        elif b == 0xDC:
            typ = TYPE_ARRAY
            self._reserve(2)
            (n,) = _unpack_from(">H", self._buffer, self._buff_i)
            self._buff_i += 2
            if n > self._max_array_len:
                raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len)
        elif b == 0xDD:
            typ = TYPE_ARRAY
            self._reserve(4)
            (n,) = _unpack_from(">I", self._buffer, self._buff_i)
            self._buff_i += 4
            if n > self._max_array_len:
                raise ValueError("%s exceeds max_array_len(%s)", n, self._max_array_len)
        elif b == 0xDE:
            self._reserve(2)
            (n,) = _unpack_from(">H", self._buffer, self._buff_i)
            self._buff_i += 2
            if n > self._max_map_len:
                raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len)
            typ = TYPE_MAP
        elif b == 0xDF:
            self._reserve(4)
            (n,) = _unpack_from(">I", self._buffer, self._buff_i)
            self._buff_i += 4
            if n > self._max_map_len:
                raise ValueError("%s exceeds max_map_len(%s)", n, self._max_map_len)
            typ = TYPE_MAP
        else:
            raise FormatError("Unknown header: 0x%x" % b)
        return typ, n, obj

    def _unpack(self, execute=EX_CONSTRUCT):
        typ, n, obj = self._read_header(execute)

        if execute == EX_READ_ARRAY_HEADER:
            if typ != TYPE_ARRAY:
                raise ValueError("Expected array")
            return n
        if execute == EX_READ_MAP_HEADER:
            if typ != TYPE_MAP:
                raise ValueError("Expected map")
            return n
        # TODO should we eliminate the recursion?
        if typ == TYPE_ARRAY:
            if execute == EX_SKIP:
                for i in xrange(n):
                    # TODO check whether we need to call `list_hook`
                    self._unpack(EX_SKIP)
                return
            ret = newlist_hint(n)
            for i in xrange(n):
                ret.append(self._unpack(EX_CONSTRUCT))
            if self._list_hook is not None:
                ret = self._list_hook(ret)
            # TODO is the interaction between `list_hook` and `use_list` ok?
            return ret if self._use_list else tuple(ret)
        if typ == TYPE_MAP:
            if execute == EX_SKIP:
                for i in xrange(n):
                    # TODO check whether we need to call hooks
                    self._unpack(EX_SKIP)
                    self._unpack(EX_SKIP)
                return
            if self._object_pairs_hook is not None:
                ret = self._object_pairs_hook(
                    (self._unpack(EX_CONSTRUCT), self._unpack(EX_CONSTRUCT))
                    for _ in xrange(n)
                )
            else:
                ret = {}
                for _ in xrange(n):
                    key = self._unpack(EX_CONSTRUCT)
                    if self._strict_map_key and type(key) not in (unicode, bytes):
                        raise ValueError(
                            "%s is not allowed for map key" % str(type(key))
                        )
                    if not PY2 and type(key) is str:
                        key = sys.intern(key)
                    ret[key] = self._unpack(EX_CONSTRUCT)
                if self._object_hook is not None:
                    ret = self._object_hook(ret)
            return ret
        if execute == EX_SKIP:
            return
        if typ == TYPE_RAW:
            if self._raw:
                obj = bytes(obj)
            else:
                obj = obj.decode("utf_8", self._unicode_errors)
            return obj
        if typ == TYPE_BIN:
            return bytes(obj)
        if typ == TYPE_EXT:
            if n == -1:  # timestamp
                ts = Timestamp.from_bytes(bytes(obj))
                if self._timestamp == 1:
                    return ts.to_unix()
                elif self._timestamp == 2:
                    return ts.to_unix_nano()
                elif self._timestamp == 3:
                    return ts.to_datetime()
                else:
                    return ts
            else:
                return self._ext_hook(n, bytes(obj))
        assert typ == TYPE_IMMEDIATE
        return obj

    def __iter__(self):
        return self

    def __next__(self):
        try:
            ret = self._unpack(EX_CONSTRUCT)
            self._consume()
            return ret
        except OutOfData:
            self._consume()
            raise StopIteration
        except RecursionError:
            raise StackError

    next = __next__

    def skip(self):
        self._unpack(EX_SKIP)
        self._consume()

    def unpack(self):
        try:
            ret = self._unpack(EX_CONSTRUCT)
        except RecursionError:
            raise StackError
        self._consume()
        return ret

    def read_array_header(self):
        ret = self._unpack(EX_READ_ARRAY_HEADER)
        self._consume()
        return ret

    def read_map_header(self):
        ret = self._unpack(EX_READ_MAP_HEADER)
        self._consume()
        return ret

    def tell(self):
        return self._stream_offset


class Packer(object):
    """
    MessagePack Packer

    Usage::

        packer = Packer()
        astream.write(packer.pack(a))
        astream.write(packer.pack(b))

    Packer's constructor has some keyword arguments:

    :param callable default:
        Convert user type to builtin type that Packer supports.
        See also simplejson's document.

    :param bool use_single_float:
        Use single precision float type for float. (default: False)

    :param bool autoreset:
        Reset buffer after each pack and return its content as `bytes`. (default: True).
        If set this to false, use `bytes()` to get content and `.reset()` to clear buffer.

    :param bool use_bin_type:
        Use bin type introduced in msgpack spec 2.0 for bytes.
        It also enables str8 type for unicode. (default: True)

    :param bool strict_types:
        If set to true, types will be checked to be exact. Derived classes
        from serializable types will not be serialized and will be
        treated as unsupported type and forwarded to default.
        Additionally tuples will not be serialized as lists.
        This is useful when trying to implement accurate serialization
        for python types.

    :param bool datetime:
        If set to true, datetime with tzinfo is packed into Timestamp type.
        Note that the tzinfo is stripped in the timestamp.
        You can get UTC datetime with `timestamp=3` option of the Unpacker.
        (Python 2 is not supported).

    :param str unicode_errors:
        The error handler for encoding unicode. (default: 'strict')
        DO NOT USE THIS!!  This option is kept for very specific usage.

    Example of streaming deserialize from file-like object::

        unpacker = Unpacker(file_like)
        for o in unpacker:
            process(o)

    Example of streaming deserialize from socket::

        unpacker = Unpacker()
        while True:
            buf = sock.recv(1024**2)
            if not buf:
                break
            unpacker.feed(buf)
            for o in unpacker:
                process(o)

    Raises ``ExtraData`` when *packed* contains extra bytes.
    Raises ``OutOfData`` when *packed* is incomplete.
    Raises ``FormatError`` when *packed* is not valid msgpack.
    Raises ``StackError`` when *packed* contains too nested.
    Other exceptions can be raised during unpacking.
    """

    def __init__(
        self,
        default=None,
        use_single_float=False,
        autoreset=True,
        use_bin_type=True,
        strict_types=False,
        datetime=False,
        unicode_errors=None,
    ):
        self._strict_types = strict_types
        self._use_float = use_single_float
        self._autoreset = autoreset
        self._use_bin_type = use_bin_type
        self._buffer = StringIO()
        if PY2 and datetime:
            raise ValueError("datetime is not supported in Python 2")
        self._datetime = bool(datetime)
        self._unicode_errors = unicode_errors or "strict"
        if default is not None:
            if not callable(default):
                raise TypeError("default must be callable")
        self._default = default

    def _pack(
        self,
        obj,
        nest_limit=DEFAULT_RECURSE_LIMIT,
        check=isinstance,
        check_type_strict=_check_type_strict,
    ):
        default_used = False
        if self._strict_types:
            check = check_type_strict
            list_types = list
        else:
            list_types = (list, tuple)
        while True:
            if nest_limit < 0:
                raise ValueError("recursion limit exceeded")
            if obj is None:
                return self._buffer.write(b"\xc0")
            if check(obj, bool):
                if obj:
                    return self._buffer.write(b"\xc3")
                return self._buffer.write(b"\xc2")
            if check(obj, int_types):
                if 0 <= obj < 0x80:
                    return self._buffer.write(struct.pack("B", obj))
                if -0x20 <= obj < 0:
                    return self._buffer.write(struct.pack("b", obj))
                if 0x80 <= obj <= 0xFF:
                    return self._buffer.write(struct.pack("BB", 0xCC, obj))
                if -0x80 <= obj < 0:
                    return self._buffer.write(struct.pack(">Bb", 0xD0, obj))
                if 0xFF < obj <= 0xFFFF:
                    return self._buffer.write(struct.pack(">BH", 0xCD, obj))
                if -0x8000 <= obj < -0x80:
                    return self._buffer.write(struct.pack(">Bh", 0xD1, obj))
                if 0xFFFF < obj <= 0xFFFFFFFF:
                    return self._buffer.write(struct.pack(">BI", 0xCE, obj))
                if -0x80000000 <= obj < -0x8000:
                    return self._buffer.write(struct.pack(">Bi", 0xD2, obj))
                if 0xFFFFFFFF < obj <= 0xFFFFFFFFFFFFFFFF:
                    return self._buffer.write(struct.pack(">BQ", 0xCF, obj))
                if -0x8000000000000000 <= obj < -0x80000000:
                    return self._buffer.write(struct.pack(">Bq", 0xD3, obj))
                if not default_used and self._default is not None:
                    obj = self._default(obj)
                    default_used = True
                    continue
                raise OverflowError("Integer value out of range")
            if check(obj, (bytes, bytearray)):
                n = len(obj)
                if n >= 2 ** 32:
                    raise ValueError("%s is too large" % type(obj).__name__)
                self._pack_bin_header(n)
                return self._buffer.write(obj)
            if check(obj, unicode):
                obj = obj.encode("utf-8", self._unicode_errors)
                n = len(obj)
                if n >= 2 ** 32:
                    raise ValueError("String is too large")
                self._pack_raw_header(n)
                return self._buffer.write(obj)
            if check(obj, memoryview):
                n = len(obj) * obj.itemsize
                if n >= 2 ** 32:
                    raise ValueError("Memoryview is too large")
                self._pack_bin_header(n)
                return self._buffer.write(obj)
            if check(obj, float):
                if self._use_float:
                    return self._buffer.write(struct.pack(">Bf", 0xCA, obj))
                return self._buffer.write(struct.pack(">Bd", 0xCB, obj))
            if check(obj, (ExtType, Timestamp)):
                if check(obj, Timestamp):
                    code = -1
                    data = obj.to_bytes()
                else:
                    code = obj.code
                    data = obj.data
                assert isinstance(code, int)
                assert isinstance(data, bytes)
                L = len(data)
                if L == 1:
                    self._buffer.write(b"\xd4")
                elif L == 2:
                    self._buffer.write(b"\xd5")
                elif L == 4:
                    self._buffer.write(b"\xd6")
                elif L == 8:
                    self._buffer.write(b"\xd7")
                elif L == 16:
                    self._buffer.write(b"\xd8")
                elif L <= 0xFF:
                    self._buffer.write(struct.pack(">BB", 0xC7, L))
                elif L <= 0xFFFF:
                    self._buffer.write(struct.pack(">BH", 0xC8, L))
                else:
                    self._buffer.write(struct.pack(">BI", 0xC9, L))
                self._buffer.write(struct.pack("b", code))
                self._buffer.write(data)
                return
            if check(obj, list_types):
                n = len(obj)
                self._pack_array_header(n)
                for i in xrange(n):
                    self._pack(obj[i], nest_limit - 1)
                return
            if check(obj, dict):
                return self._pack_map_pairs(
                    len(obj), dict_iteritems(obj), nest_limit - 1
                )

            if self._datetime and check(obj, _DateTime) and obj.tzinfo is not None:
                obj = Timestamp.from_datetime(obj)
                default_used = 1
                continue

            if not default_used and self._default is not None:
                obj = self._default(obj)
                default_used = 1
                continue
            raise TypeError("Cannot serialize %r" % (obj,))

    def pack(self, obj):
        try:
            self._pack(obj)
        except:
            self._buffer = StringIO()  # force reset
            raise
        if self._autoreset:
            ret = self._buffer.getvalue()
            self._buffer = StringIO()
            return ret

    def pack_map_pairs(self, pairs):
        self._pack_map_pairs(len(pairs), pairs)
        if self._autoreset:
            ret = self._buffer.getvalue()
            self._buffer = StringIO()
            return ret

    def pack_array_header(self, n):
        if n >= 2 ** 32:
            raise ValueError
        self._pack_array_header(n)
        if self._autoreset:
            ret = self._buffer.getvalue()
            self._buffer = StringIO()
            return ret

    def pack_map_header(self, n):
        if n >= 2 ** 32:
            raise ValueError
        self._pack_map_header(n)
        if self._autoreset:
            ret = self._buffer.getvalue()
            self._buffer = StringIO()
            return ret

    def pack_ext_type(self, typecode, data):
        if not isinstance(typecode, int):
            raise TypeError("typecode must have int type.")
        if not 0 <= typecode <= 127:
            raise ValueError("typecode should be 0-127")
        if not isinstance(data, bytes):
            raise TypeError("data must have bytes type")
        L = len(data)
        if L > 0xFFFFFFFF:
            raise ValueError("Too large data")
        if L == 1:
            self._buffer.write(b"\xd4")
        elif L == 2:
            self._buffer.write(b"\xd5")
        elif L == 4:
            self._buffer.write(b"\xd6")
        elif L == 8:
            self._buffer.write(b"\xd7")
        elif L == 16:
            self._buffer.write(b"\xd8")
        elif L <= 0xFF:
            self._buffer.write(b"\xc7" + struct.pack("B", L))
        elif L <= 0xFFFF:
            self._buffer.write(b"\xc8" + struct.pack(">H", L))
        else:
            self._buffer.write(b"\xc9" + struct.pack(">I", L))
        self._buffer.write(struct.pack("B", typecode))
        self._buffer.write(data)

    def _pack_array_header(self, n):
        if n <= 0x0F:
            return self._buffer.write(struct.pack("B", 0x90 + n))
        if n <= 0xFFFF:
            return self._buffer.write(struct.pack(">BH", 0xDC, n))
        if n <= 0xFFFFFFFF:
            return self._buffer.write(struct.pack(">BI", 0xDD, n))
        raise ValueError("Array is too large")

    def _pack_map_header(self, n):
        if n <= 0x0F:
            return self._buffer.write(struct.pack("B", 0x80 + n))
        if n <= 0xFFFF:
            return self._buffer.write(struct.pack(">BH", 0xDE, n))
        if n <= 0xFFFFFFFF:
            return self._buffer.write(struct.pack(">BI", 0xDF, n))
        raise ValueError("Dict is too large")

    def _pack_map_pairs(self, n, pairs, nest_limit=DEFAULT_RECURSE_LIMIT):
        self._pack_map_header(n)
        for (k, v) in pairs:
            self._pack(k, nest_limit - 1)
            self._pack(v, nest_limit - 1)

    def _pack_raw_header(self, n):
        if n <= 0x1F:
            self._buffer.write(struct.pack("B", 0xA0 + n))
        elif self._use_bin_type and n <= 0xFF:
            self._buffer.write(struct.pack(">BB", 0xD9, n))
        elif n <= 0xFFFF:
            self._buffer.write(struct.pack(">BH", 0xDA, n))
        elif n <= 0xFFFFFFFF:
            self._buffer.write(struct.pack(">BI", 0xDB, n))
        else:
            raise ValueError("Raw is too large")

    def _pack_bin_header(self, n):
        if not self._use_bin_type:
            return self._pack_raw_header(n)
        elif n <= 0xFF:
            return self._buffer.write(struct.pack(">BB", 0xC4, n))
        elif n <= 0xFFFF:
            return self._buffer.write(struct.pack(">BH", 0xC5, n))
        elif n <= 0xFFFFFFFF:
            return self._buffer.write(struct.pack(">BI", 0xC6, n))
        else:
            raise ValueError("Bin is too large")

    def bytes(self):
        """Return internal buffer contents as bytes object"""
        return self._buffer.getvalue()

    def reset(self):
        """Reset internal buffer.

        This method is useful only when autoreset=False.
        """
        self._buffer = StringIO()

    def getbuffer(self):
        """Return view of internal buffer."""
        if USING_STRINGBUILDER or PY2:
            return memoryview(self.bytes())
        else:
            return self._buffer.getbuffer()
site-packages/pip/_vendor/msgpack/__init__.py000064400000002136150564745010015257 0ustar00# coding: utf-8
from ._version import version
from .exceptions import *
from .ext import ExtType, Timestamp

import os
import sys


if os.environ.get("MSGPACK_PUREPYTHON") or sys.version_info[0] == 2:
    from .fallback import Packer, unpackb, Unpacker
else:
    try:
        from ._cmsgpack import Packer, unpackb, Unpacker
    except ImportError:
        from .fallback import Packer, unpackb, Unpacker


def pack(o, stream, **kwargs):
    """
    Pack object `o` and write it to `stream`

    See :class:`Packer` for options.
    """
    packer = Packer(**kwargs)
    stream.write(packer.pack(o))


def packb(o, **kwargs):
    """
    Pack object `o` and return packed bytes

    See :class:`Packer` for options.
    """
    return Packer(**kwargs).pack(o)


def unpack(stream, **kwargs):
    """
    Unpack an object from `stream`.

    Raises `ExtraData` when `stream` contains extra bytes.
    See :class:`Unpacker` for options.
    """
    data = stream.read()
    return unpackb(data, **kwargs)


# alias for compatibility to simplejson/marshal/pickle.
load = unpack
loads = unpackb

dump = pack
dumps = packb
site-packages/pip/_vendor/msgpack/_version.py000064400000000024150564745010015336 0ustar00version = (1, 0, 2)
site-packages/pip/_vendor/webencodings/__pycache__/labels.cpython-39.pyc000064400000007444150564745010022302 0ustar00a

��Re#��@s�dZddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	d	d
d
d
ddddddddddd
d
d
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd d!d!d!d!d!d"d"d"d#d#d$d$d$d$d$d$d$d%d%d%d%d%d%d%d%d%d%d&d&d'd(d(d)d*��Zd+S),z�

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

�utf-8�ibm866�
iso-8859-2�
iso-8859-3�
iso-8859-4�
iso-8859-5�
iso-8859-6�
iso-8859-7�
iso-8859-8�iso-8859-8-i�iso-8859-10�iso-8859-13�iso-8859-14�iso-8859-15�iso-8859-16�koi8-r�koi8-u�	macintosh�windows-874�windows-1250�windows-1251�windows-1252�windows-1253�windows-1254�windows-1255�windows-1256�windows-1257�windows-1258�x-mac-cyrillic�gbk�gb18030�
hz-gb-2312�big5�euc-jp�iso-2022-jp�	shift_jis�euc-kr�iso-2022-kr�utf-16be�utf-16le�x-user-defined)�zunicode-1-1-utf-8r�utf8�866�cp866�csibm866r�csisolatin2rz
iso-ir-101z	iso8859-2�iso88592z
iso_8859-2ziso_8859-2:1987�l2�latin2�csisolatin3rz
iso-ir-109z	iso8859-3�iso88593z
iso_8859-3ziso_8859-3:1988�l3�latin3�csisolatin4rz
iso-ir-110z	iso8859-4�iso88594z
iso_8859-4ziso_8859-4:1988�l4�latin4�csisolatincyrillic�cyrillicrz
iso-ir-144z	iso8859-5�iso88595z
iso_8859-5ziso_8859-5:1988�arabiczasmo-708�csiso88596e�csiso88596i�csisolatinarabiczecma-114rziso-8859-6-eziso-8859-6-iz
iso-ir-127z	iso8859-6�iso88596z
iso_8859-6ziso_8859-6:1987�csisolatingreekzecma-118�elot_928�greek�greek8rz
iso-ir-126z	iso8859-7�iso88597z
iso_8859-7ziso_8859-7:1987�sun_eu_greek�csiso88598e�csisolatinhebrew�hebrewr	ziso-8859-8-ez
iso-ir-138z	iso8859-8�iso88598z
iso_8859-8ziso_8859-8:1988�visual�csiso88598ir
�logical�csisolatin6rz
iso-ir-157z
iso8859-10�	iso885910�l6�latin6rz
iso8859-13�	iso885913r
z
iso8859-14�	iso885914�csisolatin9rz
iso8859-15�	iso885915ziso_8859-15�l9r�cskoi8r�koi�koi8r�koi8_rr�csmacintosh�macrzx-mac-romanzdos-874ziso-8859-11z
iso8859-11�	iso885911ztis-620r�cp1250rzx-cp1250�cp1251rzx-cp1251zansi_x3.4-1968�ascii�cp1252�cp819�csisolatin1�ibm819z
iso-8859-1z
iso-ir-100z	iso8859-1�iso88591z
iso_8859-1ziso_8859-1:1987�l1�latin1zus-asciirzx-cp1252�cp1253rzx-cp1253�cp1254�csisolatin5z
iso-8859-9z
iso-ir-148z	iso8859-9�iso88599z
iso_8859-9ziso_8859-9:1989�l5�latin5rzx-cp1254�cp1255rzx-cp1255�cp1256rzx-cp1256�cp1257rzx-cp1257�cp1258rzx-cp1258rzx-mac-ukrainian�chinese�csgb2312�csiso58gb231280�gb2312�gb_2312z
gb_2312-80rz	iso-ir-58zx-gbkrr r!z
big5-hkscszcn-big5�csbig5zx-x-big5�cseucpkdfmtjapaneser"zx-euc-jp�csiso2022jpr#�
csshiftjis�ms_kanjiz	shift-jisr$�sjiszwindows-31jzx-sjis�cseuckr�
csksc56011987r%z
iso-ir-149�koreanzks_c_5601-1987zks_c_5601-1989�ksc5601�ksc_5601zwindows-949�csiso2022krr&r'zutf-16r(r)N)�__doc__�LABELS�r�r���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/webencodings/labels.py�<module>s���site-packages/pip/_vendor/webencodings/__pycache__/mklabels.cpython-39.pyc000064400000003632150564745010022625 0ustar00a

��Re�@sddZddlZzddlmZWney:ddlmZYn0dd�Zdd�Zedkr`e	ed	��dS)
z�

    webencodings.mklabels
    ~~~~~~~~~~~~~~~~~~~~~

    Regenarate the webencodings.labels module.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

�N)�urlopencCs||��ksJ�|S�N)�lower)�string�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/webencodings/mklabels.py�assert_lowersrcsfdg}dd�t�t|����d��D�}tdd�|D���|��fdd�|D��|�d�d	�|�S)
Na"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
cSsLg|]D}|dD]6}|dD](}tt|���d�t|d��d�f�qqqS)�	encodings�labels�u�name)�reprr�lstrip)�.0�category�encoding�labelrrr�
<listcomp>,s��zgenerate.<locals>.<listcomp>�asciicss|]\}}t|�VqdSr��len�rrrrrr�	<genexpr>2�zgenerate.<locals>.<genexpr>c3s,|]$\}}d|d�t|�|fVqdS)z    %s:%s %s,
� Nrr��max_lenrrr3s��}�)	�json�loadsr�read�decode�max�extend�append�join)�url�partsr
rrr�generates��
r)�__main__z.http://encoding.spec.whatwg.org/encodings.json)
�__doc__r�urllibr�ImportError�urllib.requestrr)�__name__�printrrrr�<module>s!site-packages/pip/_vendor/webencodings/__pycache__/x_user_defined.cpython-39.pyc000064400000005222150564745010024013 0ustar00a

��Re��	@s�dZddlmZddlZGdd�dej�ZGdd�dej�ZGdd	�d	ej�ZGd
d�deej�ZGdd
�d
eej�Zej	de�j
e�jeeeed�ZdZ
e�e
�ZdS)z�

    webencodings.x_user_defined
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~

    An implementation of the x-user-defined encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

�)�unicode_literalsNc@s eZdZddd�Zddd�ZdS)	�Codec�strictcCst�||t�S�N)�codecs�charmap_encode�encoding_table��self�input�errors�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/webencodings/x_user_defined.py�encodeszCodec.encodecCst�||t�Sr)r�charmap_decode�decoding_tabler	r
r
r�decodeszCodec.decodeN)r)r)�__name__�
__module__�__qualname__rrr
r
r
rrs
rc@seZdZddd�ZdS)�IncrementalEncoderFcCst�||jt�dS�Nr)rrrr�r
r�finalr
r
rrszIncrementalEncoder.encodeN)F)rrrrr
r
r
rrsrc@seZdZddd�ZdS)�IncrementalDecoderFcCst�||jt�dSr)rrrrrr
r
rr$szIncrementalDecoder.decodeN)F)rrrrr
r
r
rr#src@seZdZdS)�StreamWriterN�rrrr
r
r
rr(src@seZdZdS)�StreamReaderNrr
r
r
rr,srzx-user-defined)�namerr�incrementalencoder�incrementaldecoder�streamreader�streamwriteru	

 !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuvwxyz{|}~)�__doc__�
__future__rrrrrrr�	CodecInforr�
codec_infor�
charmap_buildrr
r
r
r�<module>s*	��site-packages/pip/_vendor/webencodings/__pycache__/__init__.cpython-39.pyc000064400000023042150564745010022567 0ustar00a

��ReS)�@s�dZddlmZddlZddlmZdZddd	d
d�ZiZdd
�Z	dd�Z
dd�ZGdd�de�Z
e
d�Ze
d�Ze
d�Zd+dd�Zdd�Zedfdd�Zd,dd �Zd!d"�Zedfd#d$�Zd%d&�ZGd'd(�d(e�ZGd)d*�d*e�ZdS)-a

    webencodings
    ~~~~~~~~~~~~

    This is a Python implementation of the `WHATWG Encoding standard
    <http://encoding.spec.whatwg.org/>`. See README for details.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

�)�unicode_literalsN�)�LABELSz0.5.1z
iso-8859-8zmac-cyrillicz	mac-roman�cp874)ziso-8859-8-izx-mac-cyrillic�	macintoshzwindows-874cCs|�d����d�S)a9Transform (only) ASCII letters to lower case: A-Z is mapped to a-z.

    :param string: An Unicode string.
    :returns: A new Unicode string.

    This is used for `ASCII case-insensitive
    <http://encoding.spec.whatwg.org/#ascii-case-insensitive>`_
    matching of encoding labels.
    The same matching is also used, among other things,
    for `CSS keywords <http://dev.w3.org/csswg/css-values/#keywords>`_.

    This is different from the :meth:`~py:str.lower` method of Unicode strings
    which also affect non-ASCII characters,
    sometimes mapping them into the ASCII range:

        >>> keyword = u'Bac\N{KELVIN SIGN}ground'
        >>> assert keyword.lower() == u'background'
        >>> assert ascii_lower(keyword) != keyword.lower()
        >>> assert ascii_lower(keyword) == u'bac\N{KELVIN SIGN}ground'

    �utf8)�encode�lower�decode)�string�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/webencodings/__init__.py�ascii_lower#srcCsxt|�d��}t�|�}|dur$dSt�|�}|durt|dkrLddlm}nt�||�}t�	|�}t
||�}|t|<|S)u<
    Look for an encoding by its label.
    This is the spec’s `get an encoding
    <http://encoding.spec.whatwg.org/#concept-encoding-get>`_ algorithm.
    Supported labels are listed there.

    :param label: A string.
    :returns:
        An :class:`Encoding` object, or :obj:`None` for an unknown label.

    z	

 Nzx-user-definedr)�
codec_info)r�stripr�get�CACHE�x_user_definedr�PYTHON_NAMES�codecs�lookup�Encoding)�label�name�encodingr�python_namerrr
r=s




rcCs.t|d�r|St|�}|dur*td|��|S)z�
    Accept either an encoding object or label.

    :param encoding: An :class:`Encoding` object or a label string.
    :returns: An :class:`Encoding` object.
    :raises: :exc:`~exceptions.LookupError` for an unknown label.

    rNzUnknown encoding label: %r)�hasattrr�LookupError)�encoding_or_labelrrrr
�
_get_encoding[s	
rc@s eZdZdZdd�Zdd�ZdS)raOReresents a character encoding such as UTF-8,
    that can be used for decoding or encoding.

    .. attribute:: name

        Canonical name of the encoding

    .. attribute:: codec_info

        The actual implementation of the encoding,
        a stdlib :class:`~codecs.CodecInfo` object.
        See :func:`codecs.register`.

    cCs||_||_dS�N)rr)�selfrrrrr
�__init__|szEncoding.__init__cCs
d|jS)Nz
<Encoding %s>)r)r!rrr
�__repr__�szEncoding.__repr__N)�__name__�
__module__�__qualname__�__doc__r"r#rrrr
rmsrzutf-8zutf-16lezutf-16be�replacecCs2t|�}t|�\}}|p|}|j�||�d|fS)a�
    Decode a single string.

    :param input: A byte string
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return:
        A ``(output, encoding)`` tuple of an Unicode string
        and an :obj:`Encoding`.

    r)r�_detect_bomrr
)�input�fallback_encoding�errors�bom_encodingrrrr
r
�sr
cCsV|�d�rt|dd�fS|�d�r4t|dd�fS|�d�rNt|dd�fSd|fS)zBReturn (bom_encoding, input), with any BOM removed from the input.s���Ns��s�)�
startswith�_UTF16LE�_UTF16BE�UTF8)r*rrr
r)�s


r)�strictcCst|�j�||�dS)a;
    Encode a single string.

    :param input: An Unicode string.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return: A byte string.

    r)rrr)r*rr,rrr
r�srcCs$t||�}t||�}t|�}||fS)a�
    "Pull"-based decoder.

    :param input:
        An iterable of byte strings.

        The input is first consumed just enough to determine the encoding
        based on the precense of a BOM,
        then consumed on demand when the return value is.
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns:
        An ``(output, encoding)`` tuple.
        :obj:`output` is an iterable of Unicode strings,
        :obj:`encoding` is the :obj:`Encoding` that is being used.

    )�IncrementalDecoder�_iter_decode_generator�next)r*r+r,�decoder�	generatorrrrr
�iter_decode�s

r:ccs�|j}t|�}|D]0}||�}|r|jdus0J�|jV|Vqtq|ddd�}|jdus^J�|jV|rp|VdS|D]}||�}|rx|Vqx|ddd�}|r�|VdS)zqReturn a generator that first yields the :obj:`Encoding`,
    then yields output chukns as Unicode strings.

    N�T��final)r
�iterr)r*r8r
�chunck�outputrrr
r6�s,r6cCst||�j}t||�S)uY
    “Pull”-based encoder.

    :param input: An iterable of Unicode strings.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns: An iterable of byte strings.

    )�IncrementalEncoderr�_iter_encode_generator)r*rr,rrrr
�iter_encode�srCccs6|D]}||�}|r|Vq|ddd�}|r2|VdS)N�Tr<r)r*rr?r@rrr
rBsrBc@s$eZdZdZd	dd�Zd
dd�ZdS)r5uO
    “Push”-based decoder.

    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    r(cCs&t|�|_||_d|_d|_d|_dS)Nr;)r�_fallback_encoding�_errors�_buffer�_decoderr)r!r+r,rrr
r"s

zIncrementalDecoder.__init__FcCs||j}|dur|||�S|j|}t|�\}}|durVt|�dkrP|sP||_dS|j}|j�|j�j}||_||_	|||�S)z�Decode one chunk of the input.

        :param input: A byte string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: An Unicode string.

        Nr/rD)
rHrGr)�lenrEr�incrementaldecoderrFr
r)r!r*r=r8rrrr
r
's


zIncrementalDecoder.decodeN)r()F)r$r%r&r'r"r
rrrr
r5s

r5c@seZdZdZedfdd�ZdS)rAu�
    “Push”-based encoder.

    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    .. method:: encode(input, final=False)

        :param input: An Unicode string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: A byte string.

    r4cCst|�}|j�|�j|_dSr )rr�incrementalencoderr)r!rr,rrr
r"TszIncrementalEncoder.__init__N)r$r%r&r'r3r"rrrr
rACsrA)r()r()r'�
__future__rr�labelsr�VERSIONrrrrr�objectrr3r1r2r
r)rr:r6rCrBr5rArrrr
�<module>s4
�

 
3site-packages/pip/_vendor/webencodings/__pycache__/tests.cpython-39.pyc000064400000011766150564745010022204 0ustar00a

��Re��@s�dZddlmZddlmZmZmZmZmZm	Z	m
Z
mZmZdd�Z
dd�Zd	d
�Zdd�Zd
d�Zdd�Zdd�Zdd�Zdd�ZdS)z�

    webencodings.tests
    ~~~~~~~~~~~~~~~~~~

    A basic test suite for Encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

�)�unicode_literals�)	�lookup�LABELS�decode�encode�iter_decode�iter_encode�IncrementalDecoder�IncrementalEncoder�UTF8cOs8z||i|��Wn|y&YdS0td|��dS)NzDid not raise %s.)�AssertionError)�	exception�function�args�kwargs�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/webencodings/tests.py�
assert_raisess
rcCstd�jdksJ�td�jdks$J�td�jdks6J�td�jdksHJ�td�jdksZJ�td�jdkslJ�td�jdks~J�td�dus�J�td�dus�J�td	�jd
ks�J�td�jd
ks�J�td�jd
ks�J�td
�jd
ks�J�td�dus�J�td�du�sJ�dS)Nzutf-8zUtf-8zUTF-8�utf8zutf8 z 
utf8	�u8uutf-8 zUS-ASCIIzwindows-1252z
iso-8859-1�latin1ZLATIN1zlatin-1uLATİN1)r�namerrrr�test_labelssrcCstD]�}td|�dt|�fks"J�td|�dks4J�dD]D}tdg||�\}}t|�gks`J�ttdg||��gks8J�q8t|�}|�d�dks�J�|jddd�dks�J�t|�}|�d�dks�J�|jddd�dksJ�qt	t�
��D]}t|�j|ks�J�q�dS)N��)rr�T)�final)rrrrr�listr	r
r�set�valuesr)�label�repeat�output�_�decoder�encoderrrrr�test_all_labels0sr'cCsTtttdd�tttdd�tttgd�tttgd�tttd�tttd�dS)N�é�invalid�é)r�LookupErrorrrrr	r
rrrrr�test_invalid_labelCsr,cCs�tdd�dtd�fksJ�tdtd��dtd�fks8J�tdd�dtd�fksRJ�tdt�dtd�fkslJ�tdd�dtd�fks�J�td	d�dtd�fks�J�td
d�dtd�fks�J�tdd�dtd
�fks�J�tdd�dtd�fks�J�tdd�dtd
�fk�s
J�tdd�dtd�fk�s&J�tdd�dtd
�fk�sBJ�tdd�dtd
�fk�s^J�tdd�dtd�fk�szJ�tdd�dtd
�fk�s�J�tdd�dtd
�fk�s�J�dS)N��ru€�érr*�ascii�ér(����utf-16be�����utf-16les���us�����zUTF-16BE��zUTF-16LEzUTF-16)rrrrrrr�test_decodeLs r7cCsptdd�dksJ�tdd�dks$J�tdd�dks6J�tdd�dksHJ�tdd�dksZJ�tdd	�d
kslJ�dS)Nr*r��rr.�utf-16r6r4r2r5)rrrrr�test_encodebsr:cCsxdd�}|gd�dksJ�|dgd�dks.J�|dgd�dksBJ�|dgd�d	ksVJ�|d
dgd�d	kslJ�|dd
gd�d	ks�J�|dgd�dks�J�|dgd�dks�J�|gd�d�dks�J�|gd�d�dks�J�|gd�d�dks�J�|dgd�dk�sJ�|dgd�dk�sJ�|dgd�dk�s.J�|dgd�dk�sDJ�|gd�d�dk�s\J�|gd�d�dk�stJ�dS)NcSst||�\}}d�|�S)Nr)r�join)�input�fallback_encodingr#�	_encodingrrr�iter_decode_to_stringlsz/test_iter_decode.<locals>.iter_decode_to_stringrrrr8r*shelloZhelloshe�lloshell�or.r0r()�����)rB�arCua�)r��rrs���rDrBs�uï»r1r3)r��rrs���)rsh�r@�x-user-defineduhllor)r?rrr�test_iter_decodeksL��������
��rJcCs�d�tgd��dksJ�d�tdgd��dks2J�d�tdgd��dksLJ�d�tgd�d��dkshJ�d�tgd�d��dks�J�d�tgd�d	��dks�J�d�tgd�d
��dks�J�d�tgd�d
��dks�J�dS)Nrrrr*r8)rr*rrr9r6r4r2r5)ruhrZllorIsh�llo)r;r	rrrr�test_iter_encode�s��rKcCs@d}d}d}d}t|d�|td�fks*J�t|d�|ks<J�dS)Ns2,O�#�ɻtϨ�u2,O#tsaaZaarI)rrr)�encoded�decodedrrr�test_x_user_defined�srNN)�__doc__�
__future__rrrrrrrr	r
rrrrr'r,r7r:rJrKrNrrrr�<module>s,			site-packages/pip/_vendor/webencodings/labels.py000064400000021423150564745010016004 0ustar00"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
    'unicode-1-1-utf-8':   'utf-8',
    'utf-8':               'utf-8',
    'utf8':                'utf-8',
    '866':                 'ibm866',
    'cp866':               'ibm866',
    'csibm866':            'ibm866',
    'ibm866':              'ibm866',
    'csisolatin2':         'iso-8859-2',
    'iso-8859-2':          'iso-8859-2',
    'iso-ir-101':          'iso-8859-2',
    'iso8859-2':           'iso-8859-2',
    'iso88592':            'iso-8859-2',
    'iso_8859-2':          'iso-8859-2',
    'iso_8859-2:1987':     'iso-8859-2',
    'l2':                  'iso-8859-2',
    'latin2':              'iso-8859-2',
    'csisolatin3':         'iso-8859-3',
    'iso-8859-3':          'iso-8859-3',
    'iso-ir-109':          'iso-8859-3',
    'iso8859-3':           'iso-8859-3',
    'iso88593':            'iso-8859-3',
    'iso_8859-3':          'iso-8859-3',
    'iso_8859-3:1988':     'iso-8859-3',
    'l3':                  'iso-8859-3',
    'latin3':              'iso-8859-3',
    'csisolatin4':         'iso-8859-4',
    'iso-8859-4':          'iso-8859-4',
    'iso-ir-110':          'iso-8859-4',
    'iso8859-4':           'iso-8859-4',
    'iso88594':            'iso-8859-4',
    'iso_8859-4':          'iso-8859-4',
    'iso_8859-4:1988':     'iso-8859-4',
    'l4':                  'iso-8859-4',
    'latin4':              'iso-8859-4',
    'csisolatincyrillic':  'iso-8859-5',
    'cyrillic':            'iso-8859-5',
    'iso-8859-5':          'iso-8859-5',
    'iso-ir-144':          'iso-8859-5',
    'iso8859-5':           'iso-8859-5',
    'iso88595':            'iso-8859-5',
    'iso_8859-5':          'iso-8859-5',
    'iso_8859-5:1988':     'iso-8859-5',
    'arabic':              'iso-8859-6',
    'asmo-708':            'iso-8859-6',
    'csiso88596e':         'iso-8859-6',
    'csiso88596i':         'iso-8859-6',
    'csisolatinarabic':    'iso-8859-6',
    'ecma-114':            'iso-8859-6',
    'iso-8859-6':          'iso-8859-6',
    'iso-8859-6-e':        'iso-8859-6',
    'iso-8859-6-i':        'iso-8859-6',
    'iso-ir-127':          'iso-8859-6',
    'iso8859-6':           'iso-8859-6',
    'iso88596':            'iso-8859-6',
    'iso_8859-6':          'iso-8859-6',
    'iso_8859-6:1987':     'iso-8859-6',
    'csisolatingreek':     'iso-8859-7',
    'ecma-118':            'iso-8859-7',
    'elot_928':            'iso-8859-7',
    'greek':               'iso-8859-7',
    'greek8':              'iso-8859-7',
    'iso-8859-7':          'iso-8859-7',
    'iso-ir-126':          'iso-8859-7',
    'iso8859-7':           'iso-8859-7',
    'iso88597':            'iso-8859-7',
    'iso_8859-7':          'iso-8859-7',
    'iso_8859-7:1987':     'iso-8859-7',
    'sun_eu_greek':        'iso-8859-7',
    'csiso88598e':         'iso-8859-8',
    'csisolatinhebrew':    'iso-8859-8',
    'hebrew':              'iso-8859-8',
    'iso-8859-8':          'iso-8859-8',
    'iso-8859-8-e':        'iso-8859-8',
    'iso-ir-138':          'iso-8859-8',
    'iso8859-8':           'iso-8859-8',
    'iso88598':            'iso-8859-8',
    'iso_8859-8':          'iso-8859-8',
    'iso_8859-8:1988':     'iso-8859-8',
    'visual':              'iso-8859-8',
    'csiso88598i':         'iso-8859-8-i',
    'iso-8859-8-i':        'iso-8859-8-i',
    'logical':             'iso-8859-8-i',
    'csisolatin6':         'iso-8859-10',
    'iso-8859-10':         'iso-8859-10',
    'iso-ir-157':          'iso-8859-10',
    'iso8859-10':          'iso-8859-10',
    'iso885910':           'iso-8859-10',
    'l6':                  'iso-8859-10',
    'latin6':              'iso-8859-10',
    'iso-8859-13':         'iso-8859-13',
    'iso8859-13':          'iso-8859-13',
    'iso885913':           'iso-8859-13',
    'iso-8859-14':         'iso-8859-14',
    'iso8859-14':          'iso-8859-14',
    'iso885914':           'iso-8859-14',
    'csisolatin9':         'iso-8859-15',
    'iso-8859-15':         'iso-8859-15',
    'iso8859-15':          'iso-8859-15',
    'iso885915':           'iso-8859-15',
    'iso_8859-15':         'iso-8859-15',
    'l9':                  'iso-8859-15',
    'iso-8859-16':         'iso-8859-16',
    'cskoi8r':             'koi8-r',
    'koi':                 'koi8-r',
    'koi8':                'koi8-r',
    'koi8-r':              'koi8-r',
    'koi8_r':              'koi8-r',
    'koi8-u':              'koi8-u',
    'csmacintosh':         'macintosh',
    'mac':                 'macintosh',
    'macintosh':           'macintosh',
    'x-mac-roman':         'macintosh',
    'dos-874':             'windows-874',
    'iso-8859-11':         'windows-874',
    'iso8859-11':          'windows-874',
    'iso885911':           'windows-874',
    'tis-620':             'windows-874',
    'windows-874':         'windows-874',
    'cp1250':              'windows-1250',
    'windows-1250':        'windows-1250',
    'x-cp1250':            'windows-1250',
    'cp1251':              'windows-1251',
    'windows-1251':        'windows-1251',
    'x-cp1251':            'windows-1251',
    'ansi_x3.4-1968':      'windows-1252',
    'ascii':               'windows-1252',
    'cp1252':              'windows-1252',
    'cp819':               'windows-1252',
    'csisolatin1':         'windows-1252',
    'ibm819':              'windows-1252',
    'iso-8859-1':          'windows-1252',
    'iso-ir-100':          'windows-1252',
    'iso8859-1':           'windows-1252',
    'iso88591':            'windows-1252',
    'iso_8859-1':          'windows-1252',
    'iso_8859-1:1987':     'windows-1252',
    'l1':                  'windows-1252',
    'latin1':              'windows-1252',
    'us-ascii':            'windows-1252',
    'windows-1252':        'windows-1252',
    'x-cp1252':            'windows-1252',
    'cp1253':              'windows-1253',
    'windows-1253':        'windows-1253',
    'x-cp1253':            'windows-1253',
    'cp1254':              'windows-1254',
    'csisolatin5':         'windows-1254',
    'iso-8859-9':          'windows-1254',
    'iso-ir-148':          'windows-1254',
    'iso8859-9':           'windows-1254',
    'iso88599':            'windows-1254',
    'iso_8859-9':          'windows-1254',
    'iso_8859-9:1989':     'windows-1254',
    'l5':                  'windows-1254',
    'latin5':              'windows-1254',
    'windows-1254':        'windows-1254',
    'x-cp1254':            'windows-1254',
    'cp1255':              'windows-1255',
    'windows-1255':        'windows-1255',
    'x-cp1255':            'windows-1255',
    'cp1256':              'windows-1256',
    'windows-1256':        'windows-1256',
    'x-cp1256':            'windows-1256',
    'cp1257':              'windows-1257',
    'windows-1257':        'windows-1257',
    'x-cp1257':            'windows-1257',
    'cp1258':              'windows-1258',
    'windows-1258':        'windows-1258',
    'x-cp1258':            'windows-1258',
    'x-mac-cyrillic':      'x-mac-cyrillic',
    'x-mac-ukrainian':     'x-mac-cyrillic',
    'chinese':             'gbk',
    'csgb2312':            'gbk',
    'csiso58gb231280':     'gbk',
    'gb2312':              'gbk',
    'gb_2312':             'gbk',
    'gb_2312-80':          'gbk',
    'gbk':                 'gbk',
    'iso-ir-58':           'gbk',
    'x-gbk':               'gbk',
    'gb18030':             'gb18030',
    'hz-gb-2312':          'hz-gb-2312',
    'big5':                'big5',
    'big5-hkscs':          'big5',
    'cn-big5':             'big5',
    'csbig5':              'big5',
    'x-x-big5':            'big5',
    'cseucpkdfmtjapanese': 'euc-jp',
    'euc-jp':              'euc-jp',
    'x-euc-jp':            'euc-jp',
    'csiso2022jp':         'iso-2022-jp',
    'iso-2022-jp':         'iso-2022-jp',
    'csshiftjis':          'shift_jis',
    'ms_kanji':            'shift_jis',
    'shift-jis':           'shift_jis',
    'shift_jis':           'shift_jis',
    'sjis':                'shift_jis',
    'windows-31j':         'shift_jis',
    'x-sjis':              'shift_jis',
    'cseuckr':             'euc-kr',
    'csksc56011987':       'euc-kr',
    'euc-kr':              'euc-kr',
    'iso-ir-149':          'euc-kr',
    'korean':              'euc-kr',
    'ks_c_5601-1987':      'euc-kr',
    'ks_c_5601-1989':      'euc-kr',
    'ksc5601':             'euc-kr',
    'ksc_5601':            'euc-kr',
    'windows-949':         'euc-kr',
    'csiso2022kr':         'iso-2022-kr',
    'iso-2022-kr':         'iso-2022-kr',
    'utf-16be':            'utf-16be',
    'utf-16':              'utf-16le',
    'utf-16le':            'utf-16le',
    'x-user-defined':      'x-user-defined',
}
site-packages/pip/_vendor/webencodings/__init__.py000064400000024523150564745010016305 0ustar00# coding: utf-8
"""

    webencodings
    ~~~~~~~~~~~~

    This is a Python implementation of the `WHATWG Encoding standard
    <http://encoding.spec.whatwg.org/>`. See README for details.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

import codecs

from .labels import LABELS


VERSION = '0.5.1'


# Some names in Encoding are not valid Python aliases. Remap these.
PYTHON_NAMES = {
    'iso-8859-8-i': 'iso-8859-8',
    'x-mac-cyrillic': 'mac-cyrillic',
    'macintosh': 'mac-roman',
    'windows-874': 'cp874'}

CACHE = {}


def ascii_lower(string):
    r"""Transform (only) ASCII letters to lower case: A-Z is mapped to a-z.

    :param string: An Unicode string.
    :returns: A new Unicode string.

    This is used for `ASCII case-insensitive
    <http://encoding.spec.whatwg.org/#ascii-case-insensitive>`_
    matching of encoding labels.
    The same matching is also used, among other things,
    for `CSS keywords <http://dev.w3.org/csswg/css-values/#keywords>`_.

    This is different from the :meth:`~py:str.lower` method of Unicode strings
    which also affect non-ASCII characters,
    sometimes mapping them into the ASCII range:

        >>> keyword = u'Bac\N{KELVIN SIGN}ground'
        >>> assert keyword.lower() == u'background'
        >>> assert ascii_lower(keyword) != keyword.lower()
        >>> assert ascii_lower(keyword) == u'bac\N{KELVIN SIGN}ground'

    """
    # This turns out to be faster than unicode.translate()
    return string.encode('utf8').lower().decode('utf8')


def lookup(label):
    """
    Look for an encoding by its label.
    This is the spec’s `get an encoding
    <http://encoding.spec.whatwg.org/#concept-encoding-get>`_ algorithm.
    Supported labels are listed there.

    :param label: A string.
    :returns:
        An :class:`Encoding` object, or :obj:`None` for an unknown label.

    """
    # Only strip ASCII whitespace: U+0009, U+000A, U+000C, U+000D, and U+0020.
    label = ascii_lower(label.strip('\t\n\f\r '))
    name = LABELS.get(label)
    if name is None:
        return None
    encoding = CACHE.get(name)
    if encoding is None:
        if name == 'x-user-defined':
            from .x_user_defined import codec_info
        else:
            python_name = PYTHON_NAMES.get(name, name)
            # Any python_name value that gets to here should be valid.
            codec_info = codecs.lookup(python_name)
        encoding = Encoding(name, codec_info)
        CACHE[name] = encoding
    return encoding


def _get_encoding(encoding_or_label):
    """
    Accept either an encoding object or label.

    :param encoding: An :class:`Encoding` object or a label string.
    :returns: An :class:`Encoding` object.
    :raises: :exc:`~exceptions.LookupError` for an unknown label.

    """
    if hasattr(encoding_or_label, 'codec_info'):
        return encoding_or_label

    encoding = lookup(encoding_or_label)
    if encoding is None:
        raise LookupError('Unknown encoding label: %r' % encoding_or_label)
    return encoding


class Encoding(object):
    """Reresents a character encoding such as UTF-8,
    that can be used for decoding or encoding.

    .. attribute:: name

        Canonical name of the encoding

    .. attribute:: codec_info

        The actual implementation of the encoding,
        a stdlib :class:`~codecs.CodecInfo` object.
        See :func:`codecs.register`.

    """
    def __init__(self, name, codec_info):
        self.name = name
        self.codec_info = codec_info

    def __repr__(self):
        return '<Encoding %s>' % self.name


#: The UTF-8 encoding. Should be used for new content and formats.
UTF8 = lookup('utf-8')

_UTF16LE = lookup('utf-16le')
_UTF16BE = lookup('utf-16be')


def decode(input, fallback_encoding, errors='replace'):
    """
    Decode a single string.

    :param input: A byte string
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return:
        A ``(output, encoding)`` tuple of an Unicode string
        and an :obj:`Encoding`.

    """
    # Fail early if `encoding` is an invalid label.
    fallback_encoding = _get_encoding(fallback_encoding)
    bom_encoding, input = _detect_bom(input)
    encoding = bom_encoding or fallback_encoding
    return encoding.codec_info.decode(input, errors)[0], encoding


def _detect_bom(input):
    """Return (bom_encoding, input), with any BOM removed from the input."""
    if input.startswith(b'\xFF\xFE'):
        return _UTF16LE, input[2:]
    if input.startswith(b'\xFE\xFF'):
        return _UTF16BE, input[2:]
    if input.startswith(b'\xEF\xBB\xBF'):
        return UTF8, input[3:]
    return None, input


def encode(input, encoding=UTF8, errors='strict'):
    """
    Encode a single string.

    :param input: An Unicode string.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :return: A byte string.

    """
    return _get_encoding(encoding).codec_info.encode(input, errors)[0]


def iter_decode(input, fallback_encoding, errors='replace'):
    """
    "Pull"-based decoder.

    :param input:
        An iterable of byte strings.

        The input is first consumed just enough to determine the encoding
        based on the precense of a BOM,
        then consumed on demand when the return value is.
    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns:
        An ``(output, encoding)`` tuple.
        :obj:`output` is an iterable of Unicode strings,
        :obj:`encoding` is the :obj:`Encoding` that is being used.

    """

    decoder = IncrementalDecoder(fallback_encoding, errors)
    generator = _iter_decode_generator(input, decoder)
    encoding = next(generator)
    return generator, encoding


def _iter_decode_generator(input, decoder):
    """Return a generator that first yields the :obj:`Encoding`,
    then yields output chukns as Unicode strings.

    """
    decode = decoder.decode
    input = iter(input)
    for chunck in input:
        output = decode(chunck)
        if output:
            assert decoder.encoding is not None
            yield decoder.encoding
            yield output
            break
    else:
        # Input exhausted without determining the encoding
        output = decode(b'', final=True)
        assert decoder.encoding is not None
        yield decoder.encoding
        if output:
            yield output
        return

    for chunck in input:
        output = decode(chunck)
        if output:
            yield output
    output = decode(b'', final=True)
    if output:
        yield output


def iter_encode(input, encoding=UTF8, errors='strict'):
    """
    “Pull”-based encoder.

    :param input: An iterable of Unicode strings.
    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.
    :returns: An iterable of byte strings.

    """
    # Fail early if `encoding` is an invalid label.
    encode = IncrementalEncoder(encoding, errors).encode
    return _iter_encode_generator(input, encode)


def _iter_encode_generator(input, encode):
    for chunck in input:
        output = encode(chunck)
        if output:
            yield output
    output = encode('', final=True)
    if output:
        yield output


class IncrementalDecoder(object):
    """
    “Push”-based decoder.

    :param fallback_encoding:
        An :class:`Encoding` object or a label string.
        The encoding to use if :obj:`input` does note have a BOM.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    """
    def __init__(self, fallback_encoding, errors='replace'):
        # Fail early if `encoding` is an invalid label.
        self._fallback_encoding = _get_encoding(fallback_encoding)
        self._errors = errors
        self._buffer = b''
        self._decoder = None
        #: The actual :class:`Encoding` that is being used,
        #: or :obj:`None` if that is not determined yet.
        #: (Ie. if there is not enough input yet to determine
        #: if there is a BOM.)
        self.encoding = None  # Not known yet.

    def decode(self, input, final=False):
        """Decode one chunk of the input.

        :param input: A byte string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: An Unicode string.

        """
        decoder = self._decoder
        if decoder is not None:
            return decoder(input, final)

        input = self._buffer + input
        encoding, input = _detect_bom(input)
        if encoding is None:
            if len(input) < 3 and not final:  # Not enough data yet.
                self._buffer = input
                return ''
            else:  # No BOM
                encoding = self._fallback_encoding
        decoder = encoding.codec_info.incrementaldecoder(self._errors).decode
        self._decoder = decoder
        self.encoding = encoding
        return decoder(input, final)


class IncrementalEncoder(object):
    """
    “Push”-based encoder.

    :param encoding: An :class:`Encoding` object or a label string.
    :param errors: Type of error handling. See :func:`codecs.register`.
    :raises: :exc:`~exceptions.LookupError` for an unknown encoding label.

    .. method:: encode(input, final=False)

        :param input: An Unicode string.
        :param final:
            Indicate that no more input is available.
            Must be :obj:`True` if this is the last call.
        :returns: A byte string.

    """
    def __init__(self, encoding=UTF8, errors='strict'):
        encoding = _get_encoding(encoding)
        self.encode = encoding.codec_info.incrementalencoder(errors).encode
site-packages/pip/_vendor/webencodings/mklabels.py000064400000002431150564745010016332 0ustar00"""

    webencodings.mklabels
    ~~~~~~~~~~~~~~~~~~~~~

    Regenarate the webencodings.labels module.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

import json
try:
    from urllib import urlopen
except ImportError:
    from urllib.request import urlopen


def assert_lower(string):
    assert string == string.lower()
    return string


def generate(url):
    parts = ['''\
"""

    webencodings.labels
    ~~~~~~~~~~~~~~~~~~~

    Map encoding labels to their name.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

# XXX Do not edit!
# This file is automatically generated by mklabels.py

LABELS = {
''']
    labels = [
        (repr(assert_lower(label)).lstrip('u'),
         repr(encoding['name']).lstrip('u'))
        for category in json.loads(urlopen(url).read().decode('ascii'))
        for encoding in category['encodings']
        for label in encoding['labels']]
    max_len = max(len(label) for label, name in labels)
    parts.extend(
        '    %s:%s %s,\n' % (label, ' ' * (max_len - len(label)), name)
        for label, name in labels)
    parts.append('}')
    return ''.join(parts)


if __name__ == '__main__':
    print(generate('http://encoding.spec.whatwg.org/encodings.json'))
site-packages/pip/_vendor/webencodings/x_user_defined.py000064400000010323150564745010017522 0ustar00# coding: utf-8
"""

    webencodings.x_user_defined
    ~~~~~~~~~~~~~~~~~~~~~~~~~~~

    An implementation of the x-user-defined encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

import codecs


### Codec APIs

class Codec(codecs.Codec):

    def encode(self, input, errors='strict'):
        return codecs.charmap_encode(input, errors, encoding_table)

    def decode(self, input, errors='strict'):
        return codecs.charmap_decode(input, errors, decoding_table)


class IncrementalEncoder(codecs.IncrementalEncoder):
    def encode(self, input, final=False):
        return codecs.charmap_encode(input, self.errors, encoding_table)[0]


class IncrementalDecoder(codecs.IncrementalDecoder):
    def decode(self, input, final=False):
        return codecs.charmap_decode(input, self.errors, decoding_table)[0]


class StreamWriter(Codec, codecs.StreamWriter):
    pass


class StreamReader(Codec, codecs.StreamReader):
    pass


### encodings module API

codec_info = codecs.CodecInfo(
    name='x-user-defined',
    encode=Codec().encode,
    decode=Codec().decode,
    incrementalencoder=IncrementalEncoder,
    incrementaldecoder=IncrementalDecoder,
    streamreader=StreamReader,
    streamwriter=StreamWriter,
)


### Decoding Table

# Python 3:
# for c in range(256): print('    %r' % chr(c if c < 128 else c + 0xF700))
decoding_table = (
    '\x00'
    '\x01'
    '\x02'
    '\x03'
    '\x04'
    '\x05'
    '\x06'
    '\x07'
    '\x08'
    '\t'
    '\n'
    '\x0b'
    '\x0c'
    '\r'
    '\x0e'
    '\x0f'
    '\x10'
    '\x11'
    '\x12'
    '\x13'
    '\x14'
    '\x15'
    '\x16'
    '\x17'
    '\x18'
    '\x19'
    '\x1a'
    '\x1b'
    '\x1c'
    '\x1d'
    '\x1e'
    '\x1f'
    ' '
    '!'
    '"'
    '#'
    '$'
    '%'
    '&'
    "'"
    '('
    ')'
    '*'
    '+'
    ','
    '-'
    '.'
    '/'
    '0'
    '1'
    '2'
    '3'
    '4'
    '5'
    '6'
    '7'
    '8'
    '9'
    ':'
    ';'
    '<'
    '='
    '>'
    '?'
    '@'
    'A'
    'B'
    'C'
    'D'
    'E'
    'F'
    'G'
    'H'
    'I'
    'J'
    'K'
    'L'
    'M'
    'N'
    'O'
    'P'
    'Q'
    'R'
    'S'
    'T'
    'U'
    'V'
    'W'
    'X'
    'Y'
    'Z'
    '['
    '\\'
    ']'
    '^'
    '_'
    '`'
    'a'
    'b'
    'c'
    'd'
    'e'
    'f'
    'g'
    'h'
    'i'
    'j'
    'k'
    'l'
    'm'
    'n'
    'o'
    'p'
    'q'
    'r'
    's'
    't'
    'u'
    'v'
    'w'
    'x'
    'y'
    'z'
    '{'
    '|'
    '}'
    '~'
    '\x7f'
    '\uf780'
    '\uf781'
    '\uf782'
    '\uf783'
    '\uf784'
    '\uf785'
    '\uf786'
    '\uf787'
    '\uf788'
    '\uf789'
    '\uf78a'
    '\uf78b'
    '\uf78c'
    '\uf78d'
    '\uf78e'
    '\uf78f'
    '\uf790'
    '\uf791'
    '\uf792'
    '\uf793'
    '\uf794'
    '\uf795'
    '\uf796'
    '\uf797'
    '\uf798'
    '\uf799'
    '\uf79a'
    '\uf79b'
    '\uf79c'
    '\uf79d'
    '\uf79e'
    '\uf79f'
    '\uf7a0'
    '\uf7a1'
    '\uf7a2'
    '\uf7a3'
    '\uf7a4'
    '\uf7a5'
    '\uf7a6'
    '\uf7a7'
    '\uf7a8'
    '\uf7a9'
    '\uf7aa'
    '\uf7ab'
    '\uf7ac'
    '\uf7ad'
    '\uf7ae'
    '\uf7af'
    '\uf7b0'
    '\uf7b1'
    '\uf7b2'
    '\uf7b3'
    '\uf7b4'
    '\uf7b5'
    '\uf7b6'
    '\uf7b7'
    '\uf7b8'
    '\uf7b9'
    '\uf7ba'
    '\uf7bb'
    '\uf7bc'
    '\uf7bd'
    '\uf7be'
    '\uf7bf'
    '\uf7c0'
    '\uf7c1'
    '\uf7c2'
    '\uf7c3'
    '\uf7c4'
    '\uf7c5'
    '\uf7c6'
    '\uf7c7'
    '\uf7c8'
    '\uf7c9'
    '\uf7ca'
    '\uf7cb'
    '\uf7cc'
    '\uf7cd'
    '\uf7ce'
    '\uf7cf'
    '\uf7d0'
    '\uf7d1'
    '\uf7d2'
    '\uf7d3'
    '\uf7d4'
    '\uf7d5'
    '\uf7d6'
    '\uf7d7'
    '\uf7d8'
    '\uf7d9'
    '\uf7da'
    '\uf7db'
    '\uf7dc'
    '\uf7dd'
    '\uf7de'
    '\uf7df'
    '\uf7e0'
    '\uf7e1'
    '\uf7e2'
    '\uf7e3'
    '\uf7e4'
    '\uf7e5'
    '\uf7e6'
    '\uf7e7'
    '\uf7e8'
    '\uf7e9'
    '\uf7ea'
    '\uf7eb'
    '\uf7ec'
    '\uf7ed'
    '\uf7ee'
    '\uf7ef'
    '\uf7f0'
    '\uf7f1'
    '\uf7f2'
    '\uf7f3'
    '\uf7f4'
    '\uf7f5'
    '\uf7f6'
    '\uf7f7'
    '\uf7f8'
    '\uf7f9'
    '\uf7fa'
    '\uf7fb'
    '\uf7fc'
    '\uf7fd'
    '\uf7fe'
    '\uf7ff'
)

### Encoding table
encoding_table = codecs.charmap_build(decoding_table)
site-packages/pip/_vendor/webencodings/tests.py000064400000014643150564745010015712 0ustar00# coding: utf-8
"""

    webencodings.tests
    ~~~~~~~~~~~~~~~~~~

    A basic test suite for Encoding.

    :copyright: Copyright 2012 by Simon Sapin
    :license: BSD, see LICENSE for details.

"""

from __future__ import unicode_literals

from . import (lookup, LABELS, decode, encode, iter_decode, iter_encode,
               IncrementalDecoder, IncrementalEncoder, UTF8)


def assert_raises(exception, function, *args, **kwargs):
    try:
        function(*args, **kwargs)
    except exception:
        return
    else:  # pragma: no cover
        raise AssertionError('Did not raise %s.' % exception)


def test_labels():
    assert lookup('utf-8').name == 'utf-8'
    assert lookup('Utf-8').name == 'utf-8'
    assert lookup('UTF-8').name == 'utf-8'
    assert lookup('utf8').name == 'utf-8'
    assert lookup('utf8').name == 'utf-8'
    assert lookup('utf8 ').name == 'utf-8'
    assert lookup(' \r\nutf8\t').name == 'utf-8'
    assert lookup('u8') is None  # Python label.
    assert lookup('utf-8 ') is None  # Non-ASCII white space.

    assert lookup('US-ASCII').name == 'windows-1252'
    assert lookup('iso-8859-1').name == 'windows-1252'
    assert lookup('latin1').name == 'windows-1252'
    assert lookup('LATIN1').name == 'windows-1252'
    assert lookup('latin-1') is None
    assert lookup('LATİN1') is None  # ASCII-only case insensitivity.


def test_all_labels():
    for label in LABELS:
        assert decode(b'', label) == ('', lookup(label))
        assert encode('', label) == b''
        for repeat in [0, 1, 12]:
            output, _ = iter_decode([b''] * repeat, label)
            assert list(output) == []
            assert list(iter_encode([''] * repeat, label)) == []
        decoder = IncrementalDecoder(label)
        assert decoder.decode(b'') == ''
        assert decoder.decode(b'', final=True) == ''
        encoder = IncrementalEncoder(label)
        assert encoder.encode('') == b''
        assert encoder.encode('', final=True) == b''
    # All encoding names are valid labels too:
    for name in set(LABELS.values()):
        assert lookup(name).name == name


def test_invalid_label():
    assert_raises(LookupError, decode, b'\xEF\xBB\xBF\xc3\xa9', 'invalid')
    assert_raises(LookupError, encode, 'é', 'invalid')
    assert_raises(LookupError, iter_decode, [], 'invalid')
    assert_raises(LookupError, iter_encode, [], 'invalid')
    assert_raises(LookupError, IncrementalDecoder, 'invalid')
    assert_raises(LookupError, IncrementalEncoder, 'invalid')


def test_decode():
    assert decode(b'\x80', 'latin1') == ('€', lookup('latin1'))
    assert decode(b'\x80', lookup('latin1')) == ('€', lookup('latin1'))
    assert decode(b'\xc3\xa9', 'utf8') == ('é', lookup('utf8'))
    assert decode(b'\xc3\xa9', UTF8) == ('é', lookup('utf8'))
    assert decode(b'\xc3\xa9', 'ascii') == ('é', lookup('ascii'))
    assert decode(b'\xEF\xBB\xBF\xc3\xa9', 'ascii') == ('é', lookup('utf8'))  # UTF-8 with BOM

    assert decode(b'\xFE\xFF\x00\xe9', 'ascii') == ('é', lookup('utf-16be'))  # UTF-16-BE with BOM
    assert decode(b'\xFF\xFE\xe9\x00', 'ascii') == ('é', lookup('utf-16le'))  # UTF-16-LE with BOM
    assert decode(b'\xFE\xFF\xe9\x00', 'ascii') == ('\ue900', lookup('utf-16be'))
    assert decode(b'\xFF\xFE\x00\xe9', 'ascii') == ('\ue900', lookup('utf-16le'))

    assert decode(b'\x00\xe9', 'UTF-16BE') == ('é', lookup('utf-16be'))
    assert decode(b'\xe9\x00', 'UTF-16LE') == ('é', lookup('utf-16le'))
    assert decode(b'\xe9\x00', 'UTF-16') == ('é', lookup('utf-16le'))

    assert decode(b'\xe9\x00', 'UTF-16BE') == ('\ue900', lookup('utf-16be'))
    assert decode(b'\x00\xe9', 'UTF-16LE') == ('\ue900', lookup('utf-16le'))
    assert decode(b'\x00\xe9', 'UTF-16') == ('\ue900', lookup('utf-16le'))


def test_encode():
    assert encode('é', 'latin1') == b'\xe9'
    assert encode('é', 'utf8') == b'\xc3\xa9'
    assert encode('é', 'utf8') == b'\xc3\xa9'
    assert encode('é', 'utf-16') == b'\xe9\x00'
    assert encode('é', 'utf-16le') == b'\xe9\x00'
    assert encode('é', 'utf-16be') == b'\x00\xe9'


def test_iter_decode():
    def iter_decode_to_string(input, fallback_encoding):
        output, _encoding = iter_decode(input, fallback_encoding)
        return ''.join(output)
    assert iter_decode_to_string([], 'latin1') == ''
    assert iter_decode_to_string([b''], 'latin1') == ''
    assert iter_decode_to_string([b'\xe9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'hello'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'he', b'llo'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'hell', b'o'], 'latin1') == 'hello'
    assert iter_decode_to_string([b'\xc3\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xEF\xBB\xBF\xc3\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'\xEF\xBB\xBF', b'\xc3', b'\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'\xEF\xBB\xBF', b'a', b'\xc3'], 'latin1') == 'a\uFFFD'
    assert iter_decode_to_string([
        b'', b'\xEF', b'', b'', b'\xBB\xBF\xc3', b'\xa9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xEF\xBB\xBF'], 'latin1') == ''
    assert iter_decode_to_string([b'\xEF\xBB'], 'latin1') == 'ï»'
    assert iter_decode_to_string([b'\xFE\xFF\x00\xe9'], 'latin1') == 'é'
    assert iter_decode_to_string([b'\xFF\xFE\xe9\x00'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'', b'\xFF', b'', b'', b'\xFE\xe9', b'\x00'], 'latin1') == 'é'
    assert iter_decode_to_string([
        b'', b'h\xe9', b'llo'], 'x-user-defined') == 'h\uF7E9llo'


def test_iter_encode():
    assert b''.join(iter_encode([], 'latin1')) == b''
    assert b''.join(iter_encode([''], 'latin1')) == b''
    assert b''.join(iter_encode(['é'], 'latin1')) == b'\xe9'
    assert b''.join(iter_encode(['', 'é', '', ''], 'latin1')) == b'\xe9'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16')) == b'\xe9\x00'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16le')) == b'\xe9\x00'
    assert b''.join(iter_encode(['', 'é', '', ''], 'utf-16be')) == b'\x00\xe9'
    assert b''.join(iter_encode([
        '', 'h\uF7E9', '', 'llo'], 'x-user-defined')) == b'h\xe9llo'


def test_x_user_defined():
    encoded = b'2,\x0c\x0b\x1aO\xd9#\xcb\x0f\xc9\xbbt\xcf\xa8\xca'
    decoded = '2,\x0c\x0b\x1aO\uf7d9#\uf7cb\x0f\uf7c9\uf7bbt\uf7cf\uf7a8\uf7ca'
    encoded = b'aa'
    decoded = 'aa'
    assert decode(encoded, 'x-user-defined') == (decoded, lookup('x-user-defined'))
    assert encode(decoded, 'x-user-defined') == encoded
site-packages/pip/_vendor/certifi/__pycache__/__main__.cpython-39.pyc000064400000000753150564745010021532 0ustar00a

��Re��@sXddlZddlmZmZe��Zejdddd�e��ZejrJe	e��n
e	e��dS)�N)�contents�wherez-cz
--contents�
store_true)�action)
�argparse�pip._vendor.certifirr�ArgumentParser�parser�add_argument�
parse_args�args�print�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/certifi/__main__.py�<module>ssite-packages/pip/_vendor/certifi/__pycache__/__init__.cpython-39.pyc000064400000000472150564745010021547 0ustar00a

��Re>�@sddlmZmZdZdS)�)�contents�wherez
2021.05.30N)�corerr�__version__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/certifi/__init__.py�<module>ssite-packages/pip/_vendor/certifi/__pycache__/core.cpython-39.pyc000064400000003031150564745010020732 0ustar00a

��Re�@s�dZddlZGdd�de�ZzFe�ej�d�Zer@dd�Ze��ddl	m
ZmZda
dad	d�ZWn4eytYn$ey�ddd�Zd
d�ZYn0dd�ZdS)ze
certifi.py
~~~~~~~~~~

This module returns the installation location of cacert.pem or its contents.
�Nc@seZdZdS)�_PipPatchedCertificateN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/certifi/core.pyrsr�_PIP_STANDALONE_CERTcCstS)N)rrrrr�wheresr	)�path�	read_textcCs"tdurtdd�att���atS)Nzpip._vendor.certifi�
cacert.pem)�_CACERT_PATH�get_path�_CACERT_CTX�str�	__enter__rrrrr	 s
�asciicCs<tt�d|d��}|��Wd�S1s.0YdS)N�r��encoding)�openr	�read)�_module�_pathr�datarrrr@srcCsdS)Nz /etc/pki/tls/certs/ca-bundle.crtrrrrrr	FscCstdddd�S)N�certifirrr)rrrrr�contentsJsr)r)�__doc__�os�	Exceptionr�ImportError�environ�getrr	�importlib.resourcesr
rrrr
rrrrr�<module>s$
site-packages/pip/_vendor/certifi/__init__.py000064400000000076150564745010015260 0ustar00from .core import contents, where

__version__ = "2021.05.30"
site-packages/pip/_vendor/certifi/__main__.py000064400000000377150564745010015245 0ustar00import argparse

from pip._vendor.certifi import contents, where

parser = argparse.ArgumentParser()
parser.add_argument("-c", "--contents", action="store_true")
args = parser.parse_args()

if args.contents:
    print(contents())
else:
    print(where())
site-packages/pip/_vendor/certifi/core.py000064400000005436150564745010014456 0ustar00# -*- coding: utf-8 -*-

"""
certifi.py
~~~~~~~~~~

This module returns the installation location of cacert.pem or its contents.
"""
import os


class _PipPatchedCertificate(Exception):
    pass


try:
    raise ImportError  # force fallback
    # Return a certificate file on disk for a standalone pip zipapp running in
    # an isolated build environment to use. Passing --cert to the standalone
    # pip does not work since requests calls where() unconditionally on import.
    _PIP_STANDALONE_CERT = os.environ.get("_PIP_STANDALONE_CERT")
    if _PIP_STANDALONE_CERT:
        def where():
            return _PIP_STANDALONE_CERT
        raise _PipPatchedCertificate()

    from importlib.resources import path as get_path, read_text

    _CACERT_CTX = None
    _CACERT_PATH = None

    def where():
        # This is slightly terrible, but we want to delay extracting the file
        # in cases where we're inside of a zipimport situation until someone
        # actually calls where(), but we don't want to re-extract the file
        # on every call of where(), so we'll do it once then store it in a
        # global variable.
        global _CACERT_CTX
        global _CACERT_PATH
        if _CACERT_PATH is None:
            # This is slightly janky, the importlib.resources API wants you to
            # manage the cleanup of this file, so it doesn't actually return a
            # path, it returns a context manager that will give you the path
            # when you enter it and will do any cleanup when you leave it. In
            # the common case of not needing a temporary file, it will just
            # return the file system location and the __exit__() is a no-op.
            #
            # We also have to hold onto the actual context manager, because
            # it will do the cleanup whenever it gets garbage collected, so
            # we will also store that at the global level as well.
            _CACERT_CTX = get_path("pip._vendor.certifi", "cacert.pem")
            _CACERT_PATH = str(_CACERT_CTX.__enter__())

        return _CACERT_PATH

except _PipPatchedCertificate:
    pass

except ImportError:
    # This fallback will work for Python versions prior to 3.7 that lack the
    # importlib.resources module but relies on the existing `where` function
    # so won't address issues with environments like PyOxidizer that don't set
    # __file__ on modules.
    def read_text(_module, _path, encoding="ascii"):
        with open(where(), "r", encoding=encoding) as data:
            return data.read()

    # If we don't have importlib.resources, then we will just do the old logic
    # of assuming we're on the filesystem and munge the path directly.
    def where():
        return '/etc/pki/tls/certs/ca-bundle.crt'


def contents():
    return read_text("certifi", "cacert.pem", encoding="ascii")
site-packages/pip/_vendor/six.py000064400000103365150564745010012704 0ustar00# Copyright (c) 2010-2020 Benjamin Peterson
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.

"""Utilities for writing code that runs on Python 2 and 3"""

from __future__ import absolute_import

import functools
import itertools
import operator
import sys
import types

__author__ = "Benjamin Peterson <benjamin@python.org>"
__version__ = "1.16.0"


# Useful for very coarse version differentiation.
PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3
PY34 = sys.version_info[0:2] >= (3, 4)

if PY3:
    string_types = str,
    integer_types = int,
    class_types = type,
    text_type = str
    binary_type = bytes

    MAXSIZE = sys.maxsize
else:
    string_types = basestring,
    integer_types = (int, long)
    class_types = (type, types.ClassType)
    text_type = unicode
    binary_type = str

    if sys.platform.startswith("java"):
        # Jython always uses 32 bits.
        MAXSIZE = int((1 << 31) - 1)
    else:
        # It's possible to have sizeof(long) != sizeof(Py_ssize_t).
        class X(object):

            def __len__(self):
                return 1 << 31
        try:
            len(X())
        except OverflowError:
            # 32-bit
            MAXSIZE = int((1 << 31) - 1)
        else:
            # 64-bit
            MAXSIZE = int((1 << 63) - 1)
        del X

if PY34:
    from importlib.util import spec_from_loader
else:
    spec_from_loader = None


def _add_doc(func, doc):
    """Add documentation to a function."""
    func.__doc__ = doc


def _import_module(name):
    """Import module, returning the module after the last dot."""
    __import__(name)
    return sys.modules[name]


class _LazyDescr(object):

    def __init__(self, name):
        self.name = name

    def __get__(self, obj, tp):
        result = self._resolve()
        setattr(obj, self.name, result)  # Invokes __set__.
        try:
            # This is a bit ugly, but it avoids running this again by
            # removing this descriptor.
            delattr(obj.__class__, self.name)
        except AttributeError:
            pass
        return result


class MovedModule(_LazyDescr):

    def __init__(self, name, old, new=None):
        super(MovedModule, self).__init__(name)
        if PY3:
            if new is None:
                new = name
            self.mod = new
        else:
            self.mod = old

    def _resolve(self):
        return _import_module(self.mod)

    def __getattr__(self, attr):
        _module = self._resolve()
        value = getattr(_module, attr)
        setattr(self, attr, value)
        return value


class _LazyModule(types.ModuleType):

    def __init__(self, name):
        super(_LazyModule, self).__init__(name)
        self.__doc__ = self.__class__.__doc__

    def __dir__(self):
        attrs = ["__doc__", "__name__"]
        attrs += [attr.name for attr in self._moved_attributes]
        return attrs

    # Subclasses should override this
    _moved_attributes = []


class MovedAttribute(_LazyDescr):

    def __init__(self, name, old_mod, new_mod, old_attr=None, new_attr=None):
        super(MovedAttribute, self).__init__(name)
        if PY3:
            if new_mod is None:
                new_mod = name
            self.mod = new_mod
            if new_attr is None:
                if old_attr is None:
                    new_attr = name
                else:
                    new_attr = old_attr
            self.attr = new_attr
        else:
            self.mod = old_mod
            if old_attr is None:
                old_attr = name
            self.attr = old_attr

    def _resolve(self):
        module = _import_module(self.mod)
        return getattr(module, self.attr)


class _SixMetaPathImporter(object):

    """
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    """

    def __init__(self, six_module_name):
        self.name = six_module_name
        self.known_modules = {}

    def _add_module(self, mod, *fullnames):
        for fullname in fullnames:
            self.known_modules[self.name + "." + fullname] = mod

    def _get_module(self, fullname):
        return self.known_modules[self.name + "." + fullname]

    def find_module(self, fullname, path=None):
        if fullname in self.known_modules:
            return self
        return None

    def find_spec(self, fullname, path, target=None):
        if fullname in self.known_modules:
            return spec_from_loader(fullname, self)
        return None

    def __get_module(self, fullname):
        try:
            return self.known_modules[fullname]
        except KeyError:
            raise ImportError("This loader does not know module " + fullname)

    def load_module(self, fullname):
        try:
            # in case of a reload
            return sys.modules[fullname]
        except KeyError:
            pass
        mod = self.__get_module(fullname)
        if isinstance(mod, MovedModule):
            mod = mod._resolve()
        else:
            mod.__loader__ = self
        sys.modules[fullname] = mod
        return mod

    def is_package(self, fullname):
        """
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        """
        return hasattr(self.__get_module(fullname), "__path__")

    def get_code(self, fullname):
        """Return None

        Required, if is_package is implemented"""
        self.__get_module(fullname)  # eventually raises ImportError
        return None
    get_source = get_code  # same as get_code

    def create_module(self, spec):
        return self.load_module(spec.name)

    def exec_module(self, module):
        pass

_importer = _SixMetaPathImporter(__name__)


class _MovedItems(_LazyModule):

    """Lazy loading of moved objects"""
    __path__ = []  # mark as package


_moved_attributes = [
    MovedAttribute("cStringIO", "cStringIO", "io", "StringIO"),
    MovedAttribute("filter", "itertools", "builtins", "ifilter", "filter"),
    MovedAttribute("filterfalse", "itertools", "itertools", "ifilterfalse", "filterfalse"),
    MovedAttribute("input", "__builtin__", "builtins", "raw_input", "input"),
    MovedAttribute("intern", "__builtin__", "sys"),
    MovedAttribute("map", "itertools", "builtins", "imap", "map"),
    MovedAttribute("getcwd", "os", "os", "getcwdu", "getcwd"),
    MovedAttribute("getcwdb", "os", "os", "getcwd", "getcwdb"),
    MovedAttribute("getoutput", "commands", "subprocess"),
    MovedAttribute("range", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute("reload_module", "__builtin__", "importlib" if PY34 else "imp", "reload"),
    MovedAttribute("reduce", "__builtin__", "functools"),
    MovedAttribute("shlex_quote", "pipes", "shlex", "quote"),
    MovedAttribute("StringIO", "StringIO", "io"),
    MovedAttribute("UserDict", "UserDict", "collections"),
    MovedAttribute("UserList", "UserList", "collections"),
    MovedAttribute("UserString", "UserString", "collections"),
    MovedAttribute("xrange", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute("zip", "itertools", "builtins", "izip", "zip"),
    MovedAttribute("zip_longest", "itertools", "itertools", "izip_longest", "zip_longest"),
    MovedModule("builtins", "__builtin__"),
    MovedModule("configparser", "ConfigParser"),
    MovedModule("collections_abc", "collections", "collections.abc" if sys.version_info >= (3, 3) else "collections"),
    MovedModule("copyreg", "copy_reg"),
    MovedModule("dbm_gnu", "gdbm", "dbm.gnu"),
    MovedModule("dbm_ndbm", "dbm", "dbm.ndbm"),
    MovedModule("_dummy_thread", "dummy_thread", "_dummy_thread" if sys.version_info < (3, 9) else "_thread"),
    MovedModule("http_cookiejar", "cookielib", "http.cookiejar"),
    MovedModule("http_cookies", "Cookie", "http.cookies"),
    MovedModule("html_entities", "htmlentitydefs", "html.entities"),
    MovedModule("html_parser", "HTMLParser", "html.parser"),
    MovedModule("http_client", "httplib", "http.client"),
    MovedModule("email_mime_base", "email.MIMEBase", "email.mime.base"),
    MovedModule("email_mime_image", "email.MIMEImage", "email.mime.image"),
    MovedModule("email_mime_multipart", "email.MIMEMultipart", "email.mime.multipart"),
    MovedModule("email_mime_nonmultipart", "email.MIMENonMultipart", "email.mime.nonmultipart"),
    MovedModule("email_mime_text", "email.MIMEText", "email.mime.text"),
    MovedModule("BaseHTTPServer", "BaseHTTPServer", "http.server"),
    MovedModule("CGIHTTPServer", "CGIHTTPServer", "http.server"),
    MovedModule("SimpleHTTPServer", "SimpleHTTPServer", "http.server"),
    MovedModule("cPickle", "cPickle", "pickle"),
    MovedModule("queue", "Queue"),
    MovedModule("reprlib", "repr"),
    MovedModule("socketserver", "SocketServer"),
    MovedModule("_thread", "thread", "_thread"),
    MovedModule("tkinter", "Tkinter"),
    MovedModule("tkinter_dialog", "Dialog", "tkinter.dialog"),
    MovedModule("tkinter_filedialog", "FileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_scrolledtext", "ScrolledText", "tkinter.scrolledtext"),
    MovedModule("tkinter_simpledialog", "SimpleDialog", "tkinter.simpledialog"),
    MovedModule("tkinter_tix", "Tix", "tkinter.tix"),
    MovedModule("tkinter_ttk", "ttk", "tkinter.ttk"),
    MovedModule("tkinter_constants", "Tkconstants", "tkinter.constants"),
    MovedModule("tkinter_dnd", "Tkdnd", "tkinter.dnd"),
    MovedModule("tkinter_colorchooser", "tkColorChooser",
                "tkinter.colorchooser"),
    MovedModule("tkinter_commondialog", "tkCommonDialog",
                "tkinter.commondialog"),
    MovedModule("tkinter_tkfiledialog", "tkFileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_font", "tkFont", "tkinter.font"),
    MovedModule("tkinter_messagebox", "tkMessageBox", "tkinter.messagebox"),
    MovedModule("tkinter_tksimpledialog", "tkSimpleDialog",
                "tkinter.simpledialog"),
    MovedModule("urllib_parse", __name__ + ".moves.urllib_parse", "urllib.parse"),
    MovedModule("urllib_error", __name__ + ".moves.urllib_error", "urllib.error"),
    MovedModule("urllib", __name__ + ".moves.urllib", __name__ + ".moves.urllib"),
    MovedModule("urllib_robotparser", "robotparser", "urllib.robotparser"),
    MovedModule("xmlrpc_client", "xmlrpclib", "xmlrpc.client"),
    MovedModule("xmlrpc_server", "SimpleXMLRPCServer", "xmlrpc.server"),
]
# Add windows specific modules.
if sys.platform == "win32":
    _moved_attributes += [
        MovedModule("winreg", "_winreg"),
    ]

for attr in _moved_attributes:
    setattr(_MovedItems, attr.name, attr)
    if isinstance(attr, MovedModule):
        _importer._add_module(attr, "moves." + attr.name)
del attr

_MovedItems._moved_attributes = _moved_attributes

moves = _MovedItems(__name__ + ".moves")
_importer._add_module(moves, "moves")


class Module_six_moves_urllib_parse(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_parse"""


_urllib_parse_moved_attributes = [
    MovedAttribute("ParseResult", "urlparse", "urllib.parse"),
    MovedAttribute("SplitResult", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qs", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qsl", "urlparse", "urllib.parse"),
    MovedAttribute("urldefrag", "urlparse", "urllib.parse"),
    MovedAttribute("urljoin", "urlparse", "urllib.parse"),
    MovedAttribute("urlparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlsplit", "urlparse", "urllib.parse"),
    MovedAttribute("urlunparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlunsplit", "urlparse", "urllib.parse"),
    MovedAttribute("quote", "urllib", "urllib.parse"),
    MovedAttribute("quote_plus", "urllib", "urllib.parse"),
    MovedAttribute("unquote", "urllib", "urllib.parse"),
    MovedAttribute("unquote_plus", "urllib", "urllib.parse"),
    MovedAttribute("unquote_to_bytes", "urllib", "urllib.parse", "unquote", "unquote_to_bytes"),
    MovedAttribute("urlencode", "urllib", "urllib.parse"),
    MovedAttribute("splitquery", "urllib", "urllib.parse"),
    MovedAttribute("splittag", "urllib", "urllib.parse"),
    MovedAttribute("splituser", "urllib", "urllib.parse"),
    MovedAttribute("splitvalue", "urllib", "urllib.parse"),
    MovedAttribute("uses_fragment", "urlparse", "urllib.parse"),
    MovedAttribute("uses_netloc", "urlparse", "urllib.parse"),
    MovedAttribute("uses_params", "urlparse", "urllib.parse"),
    MovedAttribute("uses_query", "urlparse", "urllib.parse"),
    MovedAttribute("uses_relative", "urlparse", "urllib.parse"),
]
for attr in _urllib_parse_moved_attributes:
    setattr(Module_six_moves_urllib_parse, attr.name, attr)
del attr

Module_six_moves_urllib_parse._moved_attributes = _urllib_parse_moved_attributes

_importer._add_module(Module_six_moves_urllib_parse(__name__ + ".moves.urllib_parse"),
                      "moves.urllib_parse", "moves.urllib.parse")


class Module_six_moves_urllib_error(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_error"""


_urllib_error_moved_attributes = [
    MovedAttribute("URLError", "urllib2", "urllib.error"),
    MovedAttribute("HTTPError", "urllib2", "urllib.error"),
    MovedAttribute("ContentTooShortError", "urllib", "urllib.error"),
]
for attr in _urllib_error_moved_attributes:
    setattr(Module_six_moves_urllib_error, attr.name, attr)
del attr

Module_six_moves_urllib_error._moved_attributes = _urllib_error_moved_attributes

_importer._add_module(Module_six_moves_urllib_error(__name__ + ".moves.urllib.error"),
                      "moves.urllib_error", "moves.urllib.error")


class Module_six_moves_urllib_request(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_request"""


_urllib_request_moved_attributes = [
    MovedAttribute("urlopen", "urllib2", "urllib.request"),
    MovedAttribute("install_opener", "urllib2", "urllib.request"),
    MovedAttribute("build_opener", "urllib2", "urllib.request"),
    MovedAttribute("pathname2url", "urllib", "urllib.request"),
    MovedAttribute("url2pathname", "urllib", "urllib.request"),
    MovedAttribute("getproxies", "urllib", "urllib.request"),
    MovedAttribute("Request", "urllib2", "urllib.request"),
    MovedAttribute("OpenerDirector", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDefaultErrorHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPRedirectHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPCookieProcessor", "urllib2", "urllib.request"),
    MovedAttribute("ProxyHandler", "urllib2", "urllib.request"),
    MovedAttribute("BaseHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgr", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgrWithDefaultRealm", "urllib2", "urllib.request"),
    MovedAttribute("AbstractBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("AbstractDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPSHandler", "urllib2", "urllib.request"),
    MovedAttribute("FileHandler", "urllib2", "urllib.request"),
    MovedAttribute("FTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("CacheFTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("UnknownHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPErrorProcessor", "urllib2", "urllib.request"),
    MovedAttribute("urlretrieve", "urllib", "urllib.request"),
    MovedAttribute("urlcleanup", "urllib", "urllib.request"),
    MovedAttribute("URLopener", "urllib", "urllib.request"),
    MovedAttribute("FancyURLopener", "urllib", "urllib.request"),
    MovedAttribute("proxy_bypass", "urllib", "urllib.request"),
    MovedAttribute("parse_http_list", "urllib2", "urllib.request"),
    MovedAttribute("parse_keqv_list", "urllib2", "urllib.request"),
]
for attr in _urllib_request_moved_attributes:
    setattr(Module_six_moves_urllib_request, attr.name, attr)
del attr

Module_six_moves_urllib_request._moved_attributes = _urllib_request_moved_attributes

_importer._add_module(Module_six_moves_urllib_request(__name__ + ".moves.urllib.request"),
                      "moves.urllib_request", "moves.urllib.request")


class Module_six_moves_urllib_response(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_response"""


_urllib_response_moved_attributes = [
    MovedAttribute("addbase", "urllib", "urllib.response"),
    MovedAttribute("addclosehook", "urllib", "urllib.response"),
    MovedAttribute("addinfo", "urllib", "urllib.response"),
    MovedAttribute("addinfourl", "urllib", "urllib.response"),
]
for attr in _urllib_response_moved_attributes:
    setattr(Module_six_moves_urllib_response, attr.name, attr)
del attr

Module_six_moves_urllib_response._moved_attributes = _urllib_response_moved_attributes

_importer._add_module(Module_six_moves_urllib_response(__name__ + ".moves.urllib.response"),
                      "moves.urllib_response", "moves.urllib.response")


class Module_six_moves_urllib_robotparser(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_robotparser"""


_urllib_robotparser_moved_attributes = [
    MovedAttribute("RobotFileParser", "robotparser", "urllib.robotparser"),
]
for attr in _urllib_robotparser_moved_attributes:
    setattr(Module_six_moves_urllib_robotparser, attr.name, attr)
del attr

Module_six_moves_urllib_robotparser._moved_attributes = _urllib_robotparser_moved_attributes

_importer._add_module(Module_six_moves_urllib_robotparser(__name__ + ".moves.urllib.robotparser"),
                      "moves.urllib_robotparser", "moves.urllib.robotparser")


class Module_six_moves_urllib(types.ModuleType):

    """Create a six.moves.urllib namespace that resembles the Python 3 namespace"""
    __path__ = []  # mark as package
    parse = _importer._get_module("moves.urllib_parse")
    error = _importer._get_module("moves.urllib_error")
    request = _importer._get_module("moves.urllib_request")
    response = _importer._get_module("moves.urllib_response")
    robotparser = _importer._get_module("moves.urllib_robotparser")

    def __dir__(self):
        return ['parse', 'error', 'request', 'response', 'robotparser']

_importer._add_module(Module_six_moves_urllib(__name__ + ".moves.urllib"),
                      "moves.urllib")


def add_move(move):
    """Add an item to six.moves."""
    setattr(_MovedItems, move.name, move)


def remove_move(name):
    """Remove item from six.moves."""
    try:
        delattr(_MovedItems, name)
    except AttributeError:
        try:
            del moves.__dict__[name]
        except KeyError:
            raise AttributeError("no such move, %r" % (name,))


if PY3:
    _meth_func = "__func__"
    _meth_self = "__self__"

    _func_closure = "__closure__"
    _func_code = "__code__"
    _func_defaults = "__defaults__"
    _func_globals = "__globals__"
else:
    _meth_func = "im_func"
    _meth_self = "im_self"

    _func_closure = "func_closure"
    _func_code = "func_code"
    _func_defaults = "func_defaults"
    _func_globals = "func_globals"


try:
    advance_iterator = next
except NameError:
    def advance_iterator(it):
        return it.next()
next = advance_iterator


try:
    callable = callable
except NameError:
    def callable(obj):
        return any("__call__" in klass.__dict__ for klass in type(obj).__mro__)


if PY3:
    def get_unbound_function(unbound):
        return unbound

    create_bound_method = types.MethodType

    def create_unbound_method(func, cls):
        return func

    Iterator = object
else:
    def get_unbound_function(unbound):
        return unbound.im_func

    def create_bound_method(func, obj):
        return types.MethodType(func, obj, obj.__class__)

    def create_unbound_method(func, cls):
        return types.MethodType(func, None, cls)

    class Iterator(object):

        def next(self):
            return type(self).__next__(self)

    callable = callable
_add_doc(get_unbound_function,
         """Get the function out of a possibly unbound function""")


get_method_function = operator.attrgetter(_meth_func)
get_method_self = operator.attrgetter(_meth_self)
get_function_closure = operator.attrgetter(_func_closure)
get_function_code = operator.attrgetter(_func_code)
get_function_defaults = operator.attrgetter(_func_defaults)
get_function_globals = operator.attrgetter(_func_globals)


if PY3:
    def iterkeys(d, **kw):
        return iter(d.keys(**kw))

    def itervalues(d, **kw):
        return iter(d.values(**kw))

    def iteritems(d, **kw):
        return iter(d.items(**kw))

    def iterlists(d, **kw):
        return iter(d.lists(**kw))

    viewkeys = operator.methodcaller("keys")

    viewvalues = operator.methodcaller("values")

    viewitems = operator.methodcaller("items")
else:
    def iterkeys(d, **kw):
        return d.iterkeys(**kw)

    def itervalues(d, **kw):
        return d.itervalues(**kw)

    def iteritems(d, **kw):
        return d.iteritems(**kw)

    def iterlists(d, **kw):
        return d.iterlists(**kw)

    viewkeys = operator.methodcaller("viewkeys")

    viewvalues = operator.methodcaller("viewvalues")

    viewitems = operator.methodcaller("viewitems")

_add_doc(iterkeys, "Return an iterator over the keys of a dictionary.")
_add_doc(itervalues, "Return an iterator over the values of a dictionary.")
_add_doc(iteritems,
         "Return an iterator over the (key, value) pairs of a dictionary.")
_add_doc(iterlists,
         "Return an iterator over the (key, [values]) pairs of a dictionary.")


if PY3:
    def b(s):
        return s.encode("latin-1")

    def u(s):
        return s
    unichr = chr
    import struct
    int2byte = struct.Struct(">B").pack
    del struct
    byte2int = operator.itemgetter(0)
    indexbytes = operator.getitem
    iterbytes = iter
    import io
    StringIO = io.StringIO
    BytesIO = io.BytesIO
    del io
    _assertCountEqual = "assertCountEqual"
    if sys.version_info[1] <= 1:
        _assertRaisesRegex = "assertRaisesRegexp"
        _assertRegex = "assertRegexpMatches"
        _assertNotRegex = "assertNotRegexpMatches"
    else:
        _assertRaisesRegex = "assertRaisesRegex"
        _assertRegex = "assertRegex"
        _assertNotRegex = "assertNotRegex"
else:
    def b(s):
        return s
    # Workaround for standalone backslash

    def u(s):
        return unicode(s.replace(r'\\', r'\\\\'), "unicode_escape")
    unichr = unichr
    int2byte = chr

    def byte2int(bs):
        return ord(bs[0])

    def indexbytes(buf, i):
        return ord(buf[i])
    iterbytes = functools.partial(itertools.imap, ord)
    import StringIO
    StringIO = BytesIO = StringIO.StringIO
    _assertCountEqual = "assertItemsEqual"
    _assertRaisesRegex = "assertRaisesRegexp"
    _assertRegex = "assertRegexpMatches"
    _assertNotRegex = "assertNotRegexpMatches"
_add_doc(b, """Byte literal""")
_add_doc(u, """Text literal""")


def assertCountEqual(self, *args, **kwargs):
    return getattr(self, _assertCountEqual)(*args, **kwargs)


def assertRaisesRegex(self, *args, **kwargs):
    return getattr(self, _assertRaisesRegex)(*args, **kwargs)


def assertRegex(self, *args, **kwargs):
    return getattr(self, _assertRegex)(*args, **kwargs)


def assertNotRegex(self, *args, **kwargs):
    return getattr(self, _assertNotRegex)(*args, **kwargs)


if PY3:
    exec_ = getattr(moves.builtins, "exec")

    def reraise(tp, value, tb=None):
        try:
            if value is None:
                value = tp()
            if value.__traceback__ is not tb:
                raise value.with_traceback(tb)
            raise value
        finally:
            value = None
            tb = None

else:
    def exec_(_code_, _globs_=None, _locs_=None):
        """Execute code in a namespace."""
        if _globs_ is None:
            frame = sys._getframe(1)
            _globs_ = frame.f_globals
            if _locs_ is None:
                _locs_ = frame.f_locals
            del frame
        elif _locs_ is None:
            _locs_ = _globs_
        exec("""exec _code_ in _globs_, _locs_""")

    exec_("""def reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
""")


if sys.version_info[:2] > (3,):
    exec_("""def raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
""")
else:
    def raise_from(value, from_value):
        raise value


print_ = getattr(moves.builtins, "print", None)
if print_ is None:
    def print_(*args, **kwargs):
        """The new-style print function for Python 2.4 and 2.5."""
        fp = kwargs.pop("file", sys.stdout)
        if fp is None:
            return

        def write(data):
            if not isinstance(data, basestring):
                data = str(data)
            # If the file has an encoding, encode unicode with it.
            if (isinstance(fp, file) and
                    isinstance(data, unicode) and
                    fp.encoding is not None):
                errors = getattr(fp, "errors", None)
                if errors is None:
                    errors = "strict"
                data = data.encode(fp.encoding, errors)
            fp.write(data)
        want_unicode = False
        sep = kwargs.pop("sep", None)
        if sep is not None:
            if isinstance(sep, unicode):
                want_unicode = True
            elif not isinstance(sep, str):
                raise TypeError("sep must be None or a string")
        end = kwargs.pop("end", None)
        if end is not None:
            if isinstance(end, unicode):
                want_unicode = True
            elif not isinstance(end, str):
                raise TypeError("end must be None or a string")
        if kwargs:
            raise TypeError("invalid keyword arguments to print()")
        if not want_unicode:
            for arg in args:
                if isinstance(arg, unicode):
                    want_unicode = True
                    break
        if want_unicode:
            newline = unicode("\n")
            space = unicode(" ")
        else:
            newline = "\n"
            space = " "
        if sep is None:
            sep = space
        if end is None:
            end = newline
        for i, arg in enumerate(args):
            if i:
                write(sep)
            write(arg)
        write(end)
if sys.version_info[:2] < (3, 3):
    _print = print_

    def print_(*args, **kwargs):
        fp = kwargs.get("file", sys.stdout)
        flush = kwargs.pop("flush", False)
        _print(*args, **kwargs)
        if flush and fp is not None:
            fp.flush()

_add_doc(reraise, """Reraise an exception.""")

if sys.version_info[0:2] < (3, 4):
    # This does exactly the same what the :func:`py3:functools.update_wrapper`
    # function does on Python versions after 3.2. It sets the ``__wrapped__``
    # attribute on ``wrapper`` object and it doesn't raise an error if any of
    # the attributes mentioned in ``assigned`` and ``updated`` are missing on
    # ``wrapped`` object.
    def _update_wrapper(wrapper, wrapped,
                        assigned=functools.WRAPPER_ASSIGNMENTS,
                        updated=functools.WRAPPER_UPDATES):
        for attr in assigned:
            try:
                value = getattr(wrapped, attr)
            except AttributeError:
                continue
            else:
                setattr(wrapper, attr, value)
        for attr in updated:
            getattr(wrapper, attr).update(getattr(wrapped, attr, {}))
        wrapper.__wrapped__ = wrapped
        return wrapper
    _update_wrapper.__doc__ = functools.update_wrapper.__doc__

    def wraps(wrapped, assigned=functools.WRAPPER_ASSIGNMENTS,
              updated=functools.WRAPPER_UPDATES):
        return functools.partial(_update_wrapper, wrapped=wrapped,
                                 assigned=assigned, updated=updated)
    wraps.__doc__ = functools.wraps.__doc__

else:
    wraps = functools.wraps


def with_metaclass(meta, *bases):
    """Create a base class with a metaclass."""
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(type):

        def __new__(cls, name, this_bases, d):
            if sys.version_info[:2] >= (3, 7):
                # This version introduced PEP 560 that requires a bit
                # of extra care (we mimic what is done by __build_class__).
                resolved_bases = types.resolve_bases(bases)
                if resolved_bases is not bases:
                    d['__orig_bases__'] = bases
            else:
                resolved_bases = bases
            return meta(name, resolved_bases, d)

        @classmethod
        def __prepare__(cls, name, this_bases):
            return meta.__prepare__(name, bases)
    return type.__new__(metaclass, 'temporary_class', (), {})


def add_metaclass(metaclass):
    """Class decorator for creating a class with a metaclass."""
    def wrapper(cls):
        orig_vars = cls.__dict__.copy()
        slots = orig_vars.get('__slots__')
        if slots is not None:
            if isinstance(slots, str):
                slots = [slots]
            for slots_var in slots:
                orig_vars.pop(slots_var)
        orig_vars.pop('__dict__', None)
        orig_vars.pop('__weakref__', None)
        if hasattr(cls, '__qualname__'):
            orig_vars['__qualname__'] = cls.__qualname__
        return metaclass(cls.__name__, cls.__bases__, orig_vars)
    return wrapper


def ensure_binary(s, encoding='utf-8', errors='strict'):
    """Coerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    """
    if isinstance(s, binary_type):
        return s
    if isinstance(s, text_type):
        return s.encode(encoding, errors)
    raise TypeError("not expecting type '%s'" % type(s))


def ensure_str(s, encoding='utf-8', errors='strict'):
    """Coerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    # Optimization: Fast return for the common case.
    if type(s) is str:
        return s
    if PY2 and isinstance(s, text_type):
        return s.encode(encoding, errors)
    elif PY3 and isinstance(s, binary_type):
        return s.decode(encoding, errors)
    elif not isinstance(s, (text_type, binary_type)):
        raise TypeError("not expecting type '%s'" % type(s))
    return s


def ensure_text(s, encoding='utf-8', errors='strict'):
    """Coerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    if isinstance(s, binary_type):
        return s.decode(encoding, errors)
    elif isinstance(s, text_type):
        return s
    else:
        raise TypeError("not expecting type '%s'" % type(s))


def python_2_unicode_compatible(klass):
    """
    A class decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    """
    if PY2:
        if '__str__' not in klass.__dict__:
            raise ValueError("@python_2_unicode_compatible cannot be applied "
                             "to %s because it doesn't define __str__()." %
                             klass.__name__)
        klass.__unicode__ = klass.__str__
        klass.__str__ = lambda self: self.__unicode__().encode('utf-8')
    return klass


# Complete the moves implementation.
# This code is at the end of this module to speed up module loading.
# Turn this module into a package.
__path__ = []  # required for PEP 302 and PEP 451
__package__ = __name__  # see PEP 366 @ReservedAssignment
if globals().get("__spec__") is not None:
    __spec__.submodule_search_locations = []  # PEP 451 @UndefinedVariable
# Remove other six meta path importers, since they cause problems. This can
# happen if six is removed from sys.modules and then reloaded. (Setuptools does
# this for some reason.)
if sys.meta_path:
    for i, importer in enumerate(sys.meta_path):
        # Here's some real nastiness: Another "instance" of the six module might
        # be floating around. Therefore, we can't use isinstance() to check for
        # the six meta path importer, since the other six instance will have
        # inserted an importer with different class.
        if (type(importer).__name__ == "_SixMetaPathImporter" and
                importer.name == __name__):
            del sys.meta_path[i]
            break
    del i, importer
# Finally, add the importer to the meta path import hook.
sys.meta_path.append(_importer)
site-packages/pip/_vendor/requests/packages.py000064400000001267150564745010015530 0ustar00import sys

# This code exists for backwards compatibility reasons.
# I don't like it either. Just look the other way. :)

for package in ('urllib3', 'idna', 'chardet'):
    vendored_package = "pip._vendor." + package
    locals()[package] = __import__(vendored_package)
    # This traversal is apparently necessary such that the identities are
    # preserved (requests.packages.urllib3.* is urllib3.*)
    for mod in list(sys.modules):
        if mod == vendored_package or mod.startswith(vendored_package + '.'):
            unprefixed_mod = mod[len("pip._vendor."):]
            sys.modules['pip._vendor.requests.packages.' + unprefixed_mod] = sys.modules[mod]

# Kinda cool, though, right?
site-packages/pip/_vendor/requests/__pycache__/cookies.cpython-39.pyc000064400000044655150564745010021705 0ustar00a

��Re�G�@s�dZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
mZzddlZWne
yjddlZYn0Gdd�de�ZGdd	�d	e�Zd
d�Zdd
�Zddd�ZGdd�de�ZGdd�deje�Zdd�Zdd�Zdd�Zd dd�Zdd�ZdS)!z�
requests.cookies
~~~~~~~~~~~~~~~~

Compatibility code to be able to use `cookielib.CookieJar` with requests.

requests.utils imports from here, so be careful with imports.
�N�)�to_native_string)�	cookielib�urlparse�
urlunparse�Morsel�MutableMappingc@s�eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
ddd�Zdd�Zdd�Z
dd�Zedd��Zedd��Zedd��ZdS) �MockRequesta�Wraps a `requests.Request` to mimic a `urllib2.Request`.

    The code in `cookielib.CookieJar` expects this interface in order to correctly
    manage cookie policies, i.e., determine whether a cookie can be set, given the
    domains of the request and the cookie.

    The original request object is read-only. The client is responsible for collecting
    the new headers via `get_new_headers()` and interpreting them appropriately. You
    probably want `get_cookie_header`, defined below.
    cCs ||_i|_t|jj�j|_dS�N)�_r�_new_headersr�url�scheme�type)�self�request�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/cookies.py�__init__%szMockRequest.__init__cCs|jSr
)r�rrrr�get_type*szMockRequest.get_typecCst|jj�jSr
)rrr
�netlocrrrr�get_host-szMockRequest.get_hostcCs|��Sr
�rrrrr�get_origin_req_host0szMockRequest.get_origin_req_hostcCsT|jj�d�s|jjSt|jjddd�}t|jj�}t|j||j|j	|j
|jg�S)N�Hostzutf-8)�encoding)r�headers�getr
rrrr�path�params�query�fragment)r�host�parsedrrr�get_full_url3s�zMockRequest.get_full_urlcCsdS�NTrrrrr�is_unverifiableAszMockRequest.is_unverifiablecCs||jjvp||jvSr
)rrr�r�namerrr�
has_headerDszMockRequest.has_headerNcCs|jj�||j�||��Sr
)rrrr)rr)�defaultrrr�
get_headerGszMockRequest.get_headercCstd��dS)zMcookielib has no legitimate use for this method; add it back if you find one.z=Cookie headers should be added with add_unredirected_header()N)�NotImplementedError)r�key�valrrr�
add_headerJszMockRequest.add_headercCs||j|<dSr
�r�rr)�valuerrr�add_unredirected_headerNsz#MockRequest.add_unredirected_headercCs|jSr
r1rrrr�get_new_headersQszMockRequest.get_new_headerscCs|��Sr
)r'rrrr�unverifiableTszMockRequest.unverifiablecCs|��Sr
)rrrrr�origin_req_hostXszMockRequest.origin_req_hostcCs|��Sr
rrrrrr#\szMockRequest.host)N)�__name__�
__module__�__qualname__�__doc__rrrrr%r'r*r,r0r4r5�propertyr6r7r#rrrrr	s$


r	c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�MockResponsez�Wraps a `httplib.HTTPMessage` to mimic a `urllib.addinfourl`.

    ...what? Basically, expose the parsed HTTP headers from the server response
    the way `cookielib` expects to see them.
    cCs
||_dS)z�Make a MockResponse for `cookielib` to read.

        :param headers: a httplib.HTTPMessage or analogous carrying the headers
        N��_headers)rrrrrrhszMockResponse.__init__cCs|jSr
r>rrrr�infooszMockResponse.infocCs|j�|�dSr
)r?�
getheadersr(rrrrArszMockResponse.getheadersN)r8r9r:r;rr@rArrrrr=asr=cCs8t|d�r|jsdSt|�}t|jj�}|�||�dS)z�Extract the cookies from the response into a CookieJar.

    :param jar: cookielib.CookieJar (not necessarily a RequestsCookieJar)
    :param request: our own requests.Request object
    :param response: urllib3.HTTPResponse object
    �_original_responseN)�hasattrrBr	r=�msg�extract_cookies)�jarr�response�req�resrrr�extract_cookies_to_jarvs
�rJcCs t|�}|�|�|���d�S)zj
    Produce an appropriate Cookie header string to be sent with `request`, or None.

    :rtype: str
    �Cookie)r	�add_cookie_headerr5r)rFr�rrrr�get_cookie_header�s
rNcCszg}|D]N}|j|krq|dur,||jkr,q|dur@||jkr@q|�|j|j|jf�q|D]\}}}|�|||�q\dS)zkUnsets a cookie by name, by default over all domains and paths.

    Wraps CookieJar.clear(), is O(n).
    N)r)�domainr�append�clear)�	cookiejarr)rOr�
clearables�cookierrr�remove_cookie_by_name�s
rUc@seZdZdZdS)�CookieConflictErrorz�There are two cookies that meet the criteria specified in the cookie jar.
    Use .get and .set and include domain and path args in order to be more specific.
    N)r8r9r:r;rrrrrV�srVcs�eZdZdZd3dd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
dd�Zd4dd�Z�fdd�Zdd�Zdd �Zd!d"�Z�fd#d$�Z�fd%d&�Zd5d'd(�Zd6d)d*�Zd+d,�Zd-d.�Zd/d0�Zd1d2�Z�ZS)7�RequestsCookieJara�Compatibility class; is a cookielib.CookieJar, but exposes a dict
    interface.

    This is the CookieJar we create by default for requests and sessions that
    don't specify one, since some clients may expect response.cookies and
    session.cookies to support dict operations.

    Requests does not use the dict interface internally; it's just for
    compatibility with external client code. All requests code should work
    out of the box with externally provided instances of ``CookieJar``, e.g.
    ``LWPCookieJar`` and ``FileCookieJar``.

    Unlike a regular CookieJar, this class is pickleable.

    .. warning:: dictionary operations that are normally O(1) may be O(n).
    NcCs,z|�|||�WSty&|YS0dS)z�Dict-like get() that also supports optional domain and path args in
        order to resolve naming collisions from using one cookie jar over
        multiple domains.

        .. warning:: operation is O(n), not O(1).
        N)�_find_no_duplicates�KeyError)rr)r+rOrrrrr�szRequestsCookieJar.getcKs\|dur(t|||�d�|�d�d�dSt|t�r<t|�}nt||fi|��}|�|�|S)z�Dict-like set() that also supports optional domain and path args in
        order to resolve naming collisions from using one cookie jar over
        multiple domains.
        NrOr)rOr)rUr�
isinstancer�morsel_to_cookie�
create_cookie�
set_cookie)rr)r3�kwargs�crrr�set�s


zRequestsCookieJar.setccst|�D]}|jVqdS)z�Dict-like iterkeys() that returns an iterator of names of cookies
        from the jar.

        .. seealso:: itervalues() and iteritems().
        N)�iterr)�rrTrrr�iterkeys�szRequestsCookieJar.iterkeyscCst|���S)z�Dict-like keys() that returns a list of names of cookies from the
        jar.

        .. seealso:: values() and items().
        )�listrcrrrr�keys�szRequestsCookieJar.keysccst|�D]}|jVqdS)z�Dict-like itervalues() that returns an iterator of values of cookies
        from the jar.

        .. seealso:: iterkeys() and iteritems().
        N)rar3rbrrr�
itervalues�szRequestsCookieJar.itervaluescCst|���S)z�Dict-like values() that returns a list of values of cookies from the
        jar.

        .. seealso:: keys() and items().
        )rdrfrrrr�values�szRequestsCookieJar.valuesccs t|�D]}|j|jfVqdS)z�Dict-like iteritems() that returns an iterator of name-value tuples
        from the jar.

        .. seealso:: iterkeys() and itervalues().
        N)rar)r3rbrrr�	iteritems�szRequestsCookieJar.iteritemscCst|���S)z�Dict-like items() that returns a list of name-value tuples from the
        jar. Allows client-code to call ``dict(RequestsCookieJar)`` and get a
        vanilla python dict of key value pairs.

        .. seealso:: keys() and values().
        )rdrhrrrr�itemsszRequestsCookieJar.itemscCs,g}t|�D]}|j|vr|�|j�q|S)z2Utility method to list all the domains in the jar.�rarOrP�r�domainsrTrrr�list_domainss

zRequestsCookieJar.list_domainscCs,g}t|�D]}|j|vr|�|j�q|S)z0Utility method to list all the paths in the jar.)rarrP)r�pathsrTrrr�
list_pathss

zRequestsCookieJar.list_pathscCs<g}t|�D]*}|jdur*|j|vr*dS|�|j�qdS)zvReturns True if there are multiple domains in the jar.
        Returns False otherwise.

        :rtype: bool
        NTFrjrkrrr�multiple_domainssz"RequestsCookieJar.multiple_domainscCsFi}t|�D]4}|dus"|j|kr|dus4|j|kr|j||j<q|S)z�Takes as an argument an optional domain and path and returns a plain
        old Python dict of name-value pairs of cookies that meet the
        requirements.

        :rtype: dict
        N)rarOrr3r))rrOr�
dictionaryrTrrr�get_dict+s����zRequestsCookieJar.get_dictcs,ztt|��|�WSty&YdS0dSr&)�superrW�__contains__rVr(��	__class__rrrt;szRequestsCookieJar.__contains__cCs
|�|�S)z�Dict-like __getitem__() for compatibility with client code. Throws
        exception if there are more than one cookie with name. In that case,
        use the more explicit get() method instead.

        .. warning:: operation is O(n), not O(1).
        )rXr(rrr�__getitem__AszRequestsCookieJar.__getitem__cCs|�||�dS)z�Dict-like __setitem__ for compatibility with client code. Throws
        exception if there is already a cookie of that name in the jar. In that
        case, use the more explicit set() method instead.
        N)r`r2rrr�__setitem__JszRequestsCookieJar.__setitem__cCst||�dS)zlDeletes a cookie given a name. Wraps ``cookielib.CookieJar``'s
        ``remove_cookie_by_name()``.
        N)rUr(rrr�__delitem__QszRequestsCookieJar.__delitem__csRt|jd�r4|j�d�r4|j�d�r4|j�dd�|_tt|�j|g|�Ri|��S)N�
startswith�"z\"�)rCr3rz�endswith�replacersrWr])rrT�argsr^rurrr]Ws$zRequestsCookieJar.set_cookiecs<t|tj�r(|D]}|�t�|��qntt|��|�dS)zAUpdates this jar with cookies from another CookieJar or dict-likeN)rZr�	CookieJarr]�copyrsrW�update)r�otherrTrurrr�\szRequestsCookieJar.updatecCs\t|�D]<}|j|kr|dus(|j|kr|dus:|j|kr|jSqtd|||f��dS)a�Requests uses this method internally to get cookie values.

        If there are conflicting cookies, _find arbitrarily chooses one.
        See _find_no_duplicates if you want an exception thrown if there are
        conflicting cookies.

        :param name: a string containing name of cookie
        :param domain: (optional) string containing domain of cookie
        :param path: (optional) string containing path of cookie
        :return: cookie.value
        N�name=%r, domain=%r, path=%r)rar)rOrr3rY)rr)rOrrTrrr�_findds
zRequestsCookieJar._findcCsxd}t|�D]L}|j|kr|dus,|j|kr|dus>|j|kr|durRtd|��|j}q|rb|Std|||f��dS)a�Both ``__get_item__`` and ``get`` call this function: it's never
        used elsewhere in Requests.

        :param name: a string containing name of cookie
        :param domain: (optional) string containing domain of cookie
        :param path: (optional) string containing path of cookie
        :raises KeyError: if cookie is not found
        :raises CookieConflictError: if there are multiple cookies
            that match name and optionally domain and path
        :return: cookie.value
        Nz(There are multiple cookies with name, %rr�)rar)rOrrVr3rY)rr)rOr�toReturnrTrrrrXxs
z%RequestsCookieJar._find_no_duplicatescCs|j��}|�d�|S)�4Unlike a normal CookieJar, this class is pickleable.�
_cookies_lock)�__dict__r��pop�r�staterrr�__getstate__�s

zRequestsCookieJar.__getstate__cCs$|j�|�d|jvr t��|_dS)r�r�N)r�r��	threading�RLockr�r�rrr�__setstate__�s
zRequestsCookieJar.__setstate__cCs"t�}|�|���|�|�|S)z(Return a copy of this RequestsCookieJar.)rW�
set_policy�
get_policyr�)r�new_cjrrrr��s
zRequestsCookieJar.copycCs|jS)z&Return the CookiePolicy instance used.)�_policyrrrrr��szRequestsCookieJar.get_policy)NNN)NN)NN)NN)r8r9r:r;rr`rcrerfrgrhrirmrorprrrtrwrxryr]r�r�rXr�r�r�r��
__classcell__rrrurrW�s2
				
	

rWcCsN|durdSt|d�r|��St�|�}|��|D]}|�t�|��q4|S)Nr�)rCr�rQr])rF�new_jarrTrrr�_copy_cookie_jar�s

r�cKs�d||ddddddddddidd�
}t|�t|�}|rLd	}t|t|���|�|�t|d
�|d<t|d�|d
<|d�d�|d<t|d�|d<tjfi|��S)z�Make a cookie from underspecified parameters.

    By default, the pair of `name` and `value` will be set for the domain ''
    and sent on every request (this is sometimes called a "supercookie").
    rNr|�/FT�HttpOnly)
�versionr)r3�portrOr�secure�expires�discard�comment�comment_url�rest�rfc2109z4create_cookie() got unexpected keyword arguments: %sr��port_specifiedrO�domain_specified�.�domain_initial_dotr�path_specified)r`�	TypeErrorrdr��boolrzrrK)r)r3r^�result�badargs�errrrrr\�s0�
r\cCs�d}|drNztt��t|d��}WqptyJtd|d��Yqp0n"|drpd}t�t�|d|��}t|dt|d�d|d||j	|d	dd
|didt|d�|j
|d
p�dd�
S)zBConvert a Morsel object into a Cookie containing the one k/v pair.Nzmax-agezmax-age: %s must be integerr�z%a, %d-%b-%Y %H:%M:%S GMTr�FrOrr��httponlyr�r�r)
r�r�r�rOr�r)rr�r�r�r�r3r�)�int�time�
ValueErrorr��calendar�timegm�strptimer\r�r.r3)�morselr��
time_templaterrrr[�s4�



�r[TcCsR|durt�}|durNdd�|D�}|D]$}|s8||vr(|�t|||��q(|S)aCReturns a CookieJar from a key/value dictionary.

    :param cookie_dict: Dict of key/values to insert into CookieJar.
    :param cookiejar: (optional) A cookiejar to add the cookies to.
    :param overwrite: (optional) If False, will not replace cookies
        already in the jar with new ones.
    :rtype: CookieJar
    NcSsg|]
}|j�qSr)r))�.0rTrrr�
<listcomp>	�z'cookiejar_from_dict.<locals>.<listcomp>)rWr]r\)�cookie_dictrR�	overwrite�names_from_jarr)rrr�cookiejar_from_dict�s	r�cCstt|tj�std��t|t�r.t||dd�}nBt|tj�rpz|�|�Wn&tyn|D]}|�|�qZYn0|S)z�Add cookies to cookiejar and returns a merged CookieJar.

    :param cookiejar: CookieJar object to add the cookies to.
    :param cookies: Dictionary or CookieJar object to be added.
    :rtype: CookieJar
    z!You can only merge into CookieJarF)rRr�)	rZrr�r��dictr�r��AttributeErrorr])rR�cookies�
cookie_in_jarrrr�
merge_cookiess
�r�)NN)NT)r;r�r�r��_internal_utilsr�compatrrrrrr��ImportErrorZdummy_threading�objectr	r=rJrNrU�RuntimeErrorrVr�rWr�r\r[r�r�rrrr�<module>s.	H
$
site-packages/pip/_vendor/requests/__pycache__/utils.cpython-39.pyc000064400000055503150564745010021403 0ustar00a

��Re�z�@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
ddlmZddlmZddlmZddlmZdd	lmZdd
lmZmZmZmZmZmZmZmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%ddl&m'Z'ddl(m)Z)dd
l*m+Z+m,Z,m-Z-m.Z.dZ/e�0�Z1ddd�Z2d�3e�4dedd�d��Z5ej6dk�rbdd�Z7dd�Zdd�Z8dd�Z9dqd!d"�Z:d#d$�Z;d%d&�Z<ej=d'd(��Z>d)d*�Z?d+d,�Z@d-d.�ZAd/d0�ZBdrd1d2�ZCd3d4�ZDd5d6�ZEd7d8�ZFd9d:�ZGd;d<�ZHd=d>�ZId?d@�ZJdAdB�ZKeLdC�ZMdDdE�ZNdFdG�ZOdHdI�ZPdJdK�ZQdLdM�ZRdNdO�ZSej=dPdQ��ZTdRdS�ZUdsdTdU�ZVdVdW�ZWdtdYdZ�ZXd[d\�ZYd]d^�ZZd_�[d`�Z\e\daZ]e\dbZ^dcdd�Z_dedf�Z`dgdh�Zae�bdi�Zce�bdj�Zddkdl�Zedmdn�Zfdodp�ZgdS)uz�
requests.utils
~~~~~~~~~~~~~~

This module provides utility functions that are used within Requests
that are also useful for external consumption.
�N)�OrderedDict)�make_headers���__version__)�certs)�to_native_string)�parse_http_list)�quote�urlparse�bytes�str�unquote�
getproxies�proxy_bypass�
urlunparse�
basestring�
integer_types�is_py3�proxy_bypass_environment�getproxies_environment�Mapping��cookiejar_from_dict)�CaseInsensitiveDict)�
InvalidURL�
InvalidHeader�FileModeWarning�UnrewindableBodyError)z.netrc�_netrc�Pi�)�http�httpsz, z,\s*T)�accept_encodingzaccept-encoding�win32cCs�ztrddl}nddl}Wnty.YdS0z6|�|jd�}t|�|d�d�}|�|d�d}WntyzYdS0|r�|s�dS|�	d�}|D]T}|dkr�d|vr�d	S|�
dd
�}|�
dd�}|�
d
d�}t�||tj
�r�d	Sq�dS)NrFz;Software\Microsoft\Windows\CurrentVersion\Internet SettingsZProxyEnableZ
ProxyOverride�;z<local>�.Tz\.�*z.*�?)r�winreg�_winreg�ImportError�OpenKey�HKEY_CURRENT_USER�int�QueryValueEx�OSError�split�replace�re�match�I)�hostr)ZinternetSettingsZproxyEnableZ
proxyOverride�test�r8��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/utils.py�proxy_bypass_registry6sF
�����
r:cCst�rt|�St|�SdS)z�Return True, if the host should be bypassed.

        Checks proxy settings gathered from the environment, if specified,
        or the registry.
        N)rrr:)r6r8r8r9r]srcCst|d�r|��}|S)z/Returns an internal sequence dictionary update.�items)�hasattrr;)�dr8r8r9�dict_to_sequenceis
r>c	Cs.d}d}t|d�rt|�}n`t|d�r.|j}nNt|d�r|z|��}WntjyXYn$0t�|�j}d|jvr|t	�
dt�t|d��rz|��}Wn"t
tfy�|dur�|}Yn\0t|d��r|du�rz&|�dd	�|��}|�|p�d�Wnt
tf�yd}Yn0|du�r d}td||�S)
Nr�__len__�len�fileno�ba%Requests has determined the content-length for this request using the binary size of the file: however, the file has been opened in text mode (i.e. without the 'b' flag in the mode). This may lead to an incorrect content-length. In Requests 3.0, support will be removed for files in text mode.�tell�seek�)r<r@rA�io�UnsupportedOperation�os�fstat�st_size�mode�warnings�warnrrCr0�IOErrorrD�max)�o�total_length�current_positionrAr8r8r9�	super_lenrsB




�



rSFc	CsFtj�d�}|dur|f}ndd�tD�}�z�ddlm}m}d}|D]B}ztj�|�}WntyrYWdS0tj�	|�rF|}q�qF|dur�WdSt
|�}	d}
t|t�r�|
�
d�}
|	j�|
�d}z<||��|�}|�r|dr�dnd	}
||
|d
fWWSWn |tf�y$|�r �Yn0Wnttf�y@Yn0dS)z;Returns the Requests tuple auth for a given url from netrc.�NETRCNcss|]}d�|�VqdS)z~/{}N)�format)�.0�fr8r8r9�	<genexpr>��z!get_netrc_auth.<locals>.<genexpr>r)�netrc�NetrcParseError�:�asciirrE)rH�environ�get�NETRC_FILESrZr[�path�
expanduser�KeyError�existsr�
isinstancer
�decode�netlocr1�authenticatorsrNr+�AttributeError)�url�raise_errors�
netrc_file�netrc_locationsrZr[�
netrc_pathrW�loc�ri�splitstrr6r�login_ir8r8r9�get_netrc_auth�s@

rscCsBt|dd�}|r>t|t�r>|ddkr>|ddkr>tj�|�SdS)z0Tries to guess the filename of the given object.�nameNr�<����>)�getattrrerrHra�basename)�objrtr8r8r9�guess_filename�s

�r{cCs�tj�|�r|Stj�|�\}}|rPtj�|�sPtj�|�\}}d�||g�}q t�|�s^|St�|�}||��vrx|St	�
�}tj�||�d�d�}tj�|�s�t|�� }|�|�
|��Wd�n1s�0Y|S)z�Replace nonexistent paths that look like they refer to a member of a zip
    archive with the location of an extracted copy of the target, or else
    just return the provided path unchanged.
    �/rvN)rHrardr1�join�zipfile�
is_zipfile�ZipFile�namelist�tempfile�
gettempdir�atomic_open�write�read)ra�archive�member�prefix�zip_file�tmp�extracted_path�file_handlerr8r8r9�extract_zipped_paths�s"


.r�ccs�tjddkrtjntj}tjtj�|�d�\}}z@t�	|d��}|VWd�n1s\0Y|||�Wnt
y�t�|��Yn0dS)z-Write a file to the disk in an atomic fashionrrE)�dir�wbN)�sys�version_inforH�renamer2r��mkstempra�dirname�fdopen�
BaseException�remove)�filename�replacer�tmp_descriptor�tmp_name�tmp_handlerr8r8r9r�s$
r�cCs.|durdSt|ttttf�r&td��t|�S)aTake an object and test to see if it can be represented as a
    dictionary. Unless it can not be represented as such, return an
    OrderedDict, e.g.,

    ::

        >>> from_key_val_list([('key', 'val')])
        OrderedDict([('key', 'val')])
        >>> from_key_val_list('string')
        Traceback (most recent call last):
        ...
        ValueError: cannot encode objects that are not 2-tuples
        >>> from_key_val_list({'key': 'val'})
        OrderedDict([('key', 'val')])

    :rtype: OrderedDict
    N�+cannot encode objects that are not 2-tuples)rer
r�boolr.�
ValueErrorr��valuer8r8r9�from_key_val_lists
r�cCs@|durdSt|ttttf�r&td��t|t�r8|��}t|�S)a�Take an object and test to see if it can be represented as a
    dictionary. If it can be, return a list of tuples, e.g.,

    ::

        >>> to_key_val_list([('key', 'val')])
        [('key', 'val')]
        >>> to_key_val_list({'key': 'val'})
        [('key', 'val')]
        >>> to_key_val_list('string')
        Traceback (most recent call last):
        ...
        ValueError: cannot encode objects that are not 2-tuples

    :rtype: list
    Nr�)	rer
rr�r.r�rr;�listr�r8r8r9�to_key_val_list:s
r�cCsXg}t|�D]F}|dd�|dd�kr4dkrHnnt|dd��}|�|�q|S)aParse lists as described by RFC 2068 Section 2.

    In particular, parse comma-separated lists where the elements of
    the list may include quoted-strings.  A quoted-string could
    contain a comma.  A non-quoted string could have quotes in the
    middle.  Quotes are removed automatically after parsing.

    It basically works like :func:`parse_set_header` just that items
    may appear multiple times and case sensitivity is preserved.

    The return value is a standard :class:`list`:

    >>> parse_list_header('token, "quoted value"')
    ['token', 'quoted value']

    To create a header from the :class:`list` again, use the
    :func:`dump_header` function.

    :param value: a string with a list header.
    :return: :class:`list`
    :rtype: list
    Nrrv�")�_parse_list_header�unquote_header_value�append)r��result�itemr8r8r9�parse_list_headerXs(r�cCsxi}t|�D]f}d|vr"d||<q|�dd�\}}|dd�|dd�krVdkrjnnt|dd��}|||<q|S)a^Parse lists of key, value pairs as described by RFC 2068 Section 2 and
    convert them into a python dict:

    >>> d = parse_dict_header('foo="is a fish", bar="as well"')
    >>> type(d) is dict
    True
    >>> sorted(d.items())
    [('bar', 'as well'), ('foo', 'is a fish')]

    If there is no value for a key it will be `None`:

    >>> parse_dict_header('key_without_value')
    {'key_without_value': None}

    To create a header from the :class:`dict` again, use the
    :func:`dump_header` function.

    :param value: a string with a dict header.
    :return: :class:`dict`
    :rtype: dict
    �=Nrrvr�)r�r1r�)r�r�r�rtr8r8r9�parse_dict_headerxs(
r�cCs\|rX|d|dkr dkrXnn4|dd�}|rD|dd�dkrX|�dd��d	d�S|S)
z�Unquotes a header value.  (Reversal of :func:`quote_header_value`).
    This does not use the real unquoting but what browsers are actually
    using for quoting.

    :param value: the header value to unquote.
    :rtype: str
    rrvr�rNrEz\\�\z\")r2)r��is_filenamer8r8r9r��s
$r�cCsi}|D]}|j||j<q|S)z�Returns a key/value dictionary from a CookieJar.

    :param cj: CookieJar object to extract cookies from.
    :rtype: dict
    )r�rt)�cj�cookie_dict�cookier8r8r9�dict_from_cookiejar�sr�cCs
t||�S)z�Returns a CookieJar from a key/value dictionary.

    :param cj: CookieJar to insert cookies into.
    :param cookie_dict: Dict of key/values to insert into CookieJar.
    :rtype: CookieJar
    r)r�r�r8r8r9�add_dict_to_cookiejar�sr�cCsTt�dt�tjdtjd�}tjdtjd�}t�d�}|�|�|�|�|�|�S)zlReturns encodings from given content string.

    :param content: bytestring to extract encodings from.
    z�In requests 3.0, get_encodings_from_content will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)z!<meta.*?charset=["\']*(.+?)["\'>])�flagsz+<meta.*?content=["\']*;?charset=(.+?)["\'>]z$^<\?xml.*?encoding=["\']*(.+?)["\'>])rLrM�DeprecationWarningr3�compiler5�findall)�content�
charset_re�	pragma_re�xml_rer8r8r9�get_encodings_from_content�s�
��r�c
Cs�|�d�}|d��|dd�}}i}d}|D]`}|��}|r0|d}}|�d�}	|	dkr�|d|	��|�}||	dd��|�}|||��<q0||fS)	z�Returns content type and parameters from given header

    :param header: string
    :return: tuple containing content type and dictionary of
         parameters
    r%rrNz"' Tr�rv)r1�strip�find�lower)
�header�tokens�content_type�params�params_dict�items_to_strip�param�keyr��index_of_equalsr8r8r9�_parse_content_type_header�s


r�cCsP|�d�}|sdSt|�\}}d|vr4|d�d�Sd|vr@dSd|vrLdSdS)	z}Returns encodings from given HTTP Header Dict.

    :param headers: dictionary to extract encoding from.
    :rtype: str
    zcontent-typeN�charsetz'"�textz
ISO-8859-1zapplication/json�utf-8)r_r�r�)�headersr�r�r8r8r9�get_encoding_from_headers�s
r�ccsj|jdur|D]
}|VqdSt�|j�dd�}|D]}|�|�}|r4|Vq4|jddd�}|rf|VdS)zStream decodes a iterator.Nr2��errorsrYT)�final)�encoding�codecs�getincrementaldecoderrf)�iterator�rr��decoder�chunk�rvr8r8r9�stream_decode_response_unicodes

r�ccsHd}|dus|dkrt|�}|t|�krD||||�V||7}qdS)z Iterate over slices of a string.rN)r@)�string�slice_length�posr8r8r9�iter_slices&sr�cCszt�dt�g}t|j�}|rJzt|j|�WStyH|�|�Yn0zt|j|dd�WSt	yt|jYS0dS)z�Returns the requested content back in unicode.

    :param r: Response object to get unicode content from.

    Tried:

    1. charset from content-type
    2. fall back and replace all unicode characters

    :rtype: str
    z�In requests 3.0, get_unicode_from_response will be removed. For more information, please see the discussion on issue #2266. (This warning should only appear once.)r2r�N)
rLrMr�r�r�r
r��UnicodeErrorr��	TypeError)r��tried_encodingsr�r8r8r9�get_unicode_from_response0s�
r�zBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789-._~c	Cs�|�d�}tdt|��D]�}||dd�}t|�dkr�|��r�ztt|d��}Wntyptd|��Yn0|tvr�|||dd�||<q�d||||<qd||||<qd�	|�S)	z�Un-escape any percent-escape sequences in a URI that are unreserved
    characters. This leaves all reserved, illegal and non-ASCII bytes encoded.

    :rtype: str
    �%rrrE�z%Invalid percent-escape sequence: '%s'N�)
r1�ranger@�isalnum�chrr.r�r�UNRESERVED_SETr})�uri�parts�i�h�cr8r8r9�unquote_unreservedYs
r�cCs>d}d}ztt|�|d�WSty8t||d�YS0dS)z�Re-quote the given URI.

    This function passes the given URI through an unquote/quote cycle to
    ensure that it is fully and consistently quoted.

    :rtype: str
    z!#$%&'()*+,/:;=?@[]~z!#$&'()*+,/:;=?@[]~)�safeN)r
r�r)r��safe_with_percent�safe_without_percentr8r8r9�requote_uriqsr�cCslt�dt�|��d}|�d�\}}t�dt�tt|����d}t�dt�|��d|@}||@||@kS)z�This function allows you to check if an IP belongs to a network subnet

    Example: returns True if ip = 192.168.1.1 and net = 192.168.1.0/24
             returns False if ip = 192.168.1.1 and net = 192.168.100.0/24

    :rtype: bool
    z=Lrr|)�struct�unpack�socket�	inet_atonr1�dotted_netmaskr.)�ip�net�ipaddr�netaddr�bits�netmask�networkr8r8r9�address_in_network�s
rcCs&ddd|>dA}t�t�d|��S)z�Converts mask from /xx format to xxx.xxx.xxx.xxx

    Example: if mask is 24 function returns 255.255.255.0

    :rtype: str
    l��r� z>I)r��	inet_ntoar��pack)�maskr�r8r8r9r��sr�cCs*zt�|�Wntjy$YdS0dS)z
    :rtype: bool
    FT)r�r��error)�	string_ipr8r8r9�is_ipv4_address�s
r	cCs�|�d�dkr�zt|�d�d�}Wnty8YdS0|dksJ|dkrNdSzt�|�d�d�Wq�tjy|YdS0ndSdS)zV
    Very simple check of the cidr format in no_proxy variable.

    :rtype: bool
    r|rFrrT)�countr.r1r�r�r�r)�string_networkrr8r8r9�
is_valid_cidr�s
rc	cst|du}|r"tj�|�}|tj|<z*dVW|rp|durBtj|=qp|tj|<n"|rn|durdtj|=n
|tj|<0dS)z�Set the environment variable 'env_name' to 'value'

    Save previous value, yield, and then restore the previous value stored in
    the environment variable 'env_name'.

    If 'value' is None, do nothingN)rHr^r_)�env_namer��
value_changed�	old_valuer8r8r9�set_environ�s

�
rc	
Cs>dd�}|}|dur|d�}t|�}|jdur2dS|r�dd�|�dd	��d
�D�}t|j�r�|D]0}t|�r�t|j|�r�dSq`|j|kr`dSq`nB|j}|jr�|d�|j�7}|D] }|j�	|�s�|�	|�r�dSq�t
d|��>zt|j�}Wntt
jf�yd}Yn0Wd�n1�s&0Y|�r:dSdS)
zL
    Returns whether we should bypass proxies or not.

    :rtype: bool
    cSstj�|�ptj�|���S�N)rHr^r_�upper)�kr8r8r9�<lambda>�rYz'should_bypass_proxies.<locals>.<lambda>N�no_proxyTcss|]}|r|VqdSrr8)rVr6r8r8r9rX�sz(should_bypass_proxies.<locals>.<genexpr>� r��,z:{}F)r�hostnamer2r1r	rr�portrU�endswithrrr�r��gaierror)	rjr�	get_proxy�no_proxy_arg�parsed�proxy_ip�host_with_portr6�bypassr8r8r9�should_bypass_proxies�s@
�


*r"cCst||d�riSt�SdS)zA
    Return a dict of environment proxies.

    :rtype: dict
    )rN)r"r)rjrr8r8r9�get_environ_proxiessr#cCst|pi}t|�}|jdur.|�|j|�d��S|jd|j|jd|jdg}d}|D]}||vrV||}qpqV|S)z�Select a proxy for the url, if applicable.

    :param url: The url being for the request
    :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs
    N�allz://zall://)rrr_�scheme)rj�proxies�urlparts�
proxy_keys�proxy�	proxy_keyr8r8r9�select_proxy#s
�r+�python-requestscCsd|tfS)zO
    Return a string representing the default user agent.

    :rtype: str
    z%s/%sr)rtr8r8r9�default_user_agent=sr-cCstt�tddd��S)z9
    :rtype: requests.structures.CaseInsensitiveDict
    z*/*z
keep-alive)z
User-AgentzAccept-Encoding�Accept�
Connection)rr-�DEFAULT_ACCEPT_ENCODINGr8r8r8r9�default_headersFs�r1c	
Cs�g}d}|�|�}|s|St�d|�D]�}z|�dd�\}}WntyZ|d}}Yn0d|�d�i}|�d�D]D}z|�d�\}}Wnty�Yq�Yn0|�|�||�|�<qt|�|�q&|S)	z�Return a list of parsed link headers proxies.

    i.e. Link: <http:/.../front.jpeg>; rel=front; type="image/jpeg",<http://.../back.jpeg>; rel=back;type="image/jpeg"

    :rtype: list
    z '"z, *<r%rr�rjz<> '"r�)r�r3r1r�r�)	r��links�
replace_chars�valrjr��linkr�r�r8r8r9�parse_header_linksRs&
r6�r]rE�cCs�|dd�}|tjtjfvr dS|dd�tjkr6dS|dd�tjtjfvrRdS|�t�}|dkrhd	S|dkr�|ddd�tkr�d
S|ddd�tkr�dS|dkr�|dd�t	kr�d
S|dd�t	kr�dSdS)z
    :rtype: str
    N�zutf-32r8z	utf-8-sigrEzutf-16rr�z	utf-16-berz	utf-16-lez	utf-32-bez	utf-32-le)
r��BOM_UTF32_LE�BOM_UTF32_BE�BOM_UTF8�BOM_UTF16_LE�BOM_UTF16_BEr
�_null�_null2�_null3)�data�sample�	nullcountr8r8r9�guess_json_utf}s*
rEcCs8t||�\}}}}}}|s$||}}t||||||f�S)z�Given a URL that may or may not have a scheme, prepend the given scheme.
    Does not replace a present scheme with the one provided as an argument.

    :rtype: str
    )rr)rj�
new_schemer%rgrar��query�fragmentr8r8r9�prepend_scheme_if_needed�s
rIc	Cs@t|�}zt|j�t|j�f}Wnttfy:d}Yn0|S)z{Given a url with authentication components, extract them into a tuple of
    username,password.

    :rtype: (str,str)
    )r�r�)rr�username�passwordrir�)rjr�authr8r8r9�get_auth_from_url�s
rMs^\S[^\r\n]*$|^$z^\S[^\r\n]*$|^$c	Csd|\}}t|t�rt}nt}z|�|�s4td|��Wn(ty^td||t|�f��Yn0dS)z�Verifies that header value is a string which doesn't contain
    leading whitespace or return characters. This prevents unintended
    header injection.

    :param header: tuple, in the format (name, value).
    z7Invalid return character or leading space in header: %sz>Value for header {%s: %s} must be of type str or bytes, not %sN)rer�_CLEAN_HEADER_REGEX_BYTE�_CLEAN_HEADER_REGEX_STRr4rr��type)r�rtr��patr8r8r9�check_header_validity�s

�rRcCsFt|�\}}}}}}|s"||}}|�dd�d}t|||||df�S)zW
    Given a url remove the fragment and the authentication part.

    :rtype: str
    �@rrvr�)r�rsplitr)rjr%rgrar�rGrHr8r8r9�
urldefragauth�s

rUc	Cs^t|jdd�}|durRt|jt�rRz||j�WqZttfyNtd��YqZ0ntd��dS)zfMove file pointer back to its recorded starting position
    so it can be read again on redirect.
    rDNz;An error occurred when rewinding request body for redirect.z+Unable to rewind request body for redirect.)rx�bodyre�_body_positionrrNr0r)�prepared_request�	body_seekr8r8r9�rewind_body�srZ)F)F)N)r,)h�__doc__r��
contextlibrFrHr3r�r�r�r�rLr~�collectionsr�pip._vendor.urllib3.utilrrr�r�_internal_utilsr�compatr	r�r
rrr
rrrrrrrrrr�cookiesr�
structuresr�
exceptionsrrrrr`�where�DEFAULT_CA_BUNDLE_PATH�
DEFAULT_PORTSr}r1r0�platformr:r>rSrsr{r��contextmanagerr�r�r�r�r�r�r�r�r�r�r�r�r�r��	frozensetr�r�r�rr�r	rrr"r#r+r-r1r6�encoder?r@rArErIrMr�rNrOrRrUrZr8r8r8r9�<module>s�@
�'	=
9!

 #

%�
=

	&
 

site-packages/pip/_vendor/requests/__pycache__/help.cpython-39.pyc000064400000005564150564745010021175 0ustar00a

��Re��@s�dZddlmZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZdZ
zddlmZWney~dZYn0zdd	lmZWney�dZdZdZYn0ddlZddlZd
d�Zdd
�Zdd�Zedkr�e�dS)z'Module containing bug report helper(s).�)�print_functionN)�idna)�urllib3�)�__version__)�chardet)�	pyopensslcCs�t��}|dkrt��}nj|dkr\dtjjtjjtjjf}tjjdkr�d�	|tjjg�}n(|dkrnt��}n|dkr�t��}nd}||d	�S)
a�Return a dict with the Python implementation and version.

    Provide both the name and the version of the Python implementation
    currently running. For example, on CPython 2.7.5 it will return
    {'name': 'CPython', 'version': '2.7.5'}.

    This function works best on CPython and PyPy: in particular, it probably
    doesn't work for Jython or IronPython. Future investigation should be done
    to work out the correct shape of the code for those platforms.
    �CPython�PyPyz%s.%s.%s�final��Jython�
IronPython�Unknown)�name�version)
�platform�python_implementation�python_version�sys�pypy_version_info�major�minor�micro�releaselevel�join)�implementation�implementation_version�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/help.py�_implementation s$
��

r c

Cs�zt��t��d�}Wnty2ddd�}Yn0t�}dtji}ddi}ddi}trbdtji}trpdtji}ddd�}t	r�t	jdt	j
jd�}dtt
dd�i}dttdd�i}tj}d|dur�d|ndi}	|||	tdutdu||||||dtid	�S)
z&Generate information for a bug report.)�system�releaserrNr)r�openssl_versionz%xr)rr�
system_sslZusing_pyopensslZusing_charset_normalizerZ	pyOpenSSLrr�charset_normalizer�cryptographyr�requests)rr!r"�IOErrorr rrr%r�OpenSSLZSSL�OPENSSL_VERSION_NUMBER�getattrr&r�sslr�requests_version)
Z
platform_infoZimplementation_infoZurllib3_infoZcharset_normalizer_infoZchardet_infoZpyopenssl_infoZcryptography_infoZ	idna_infor$Zsystem_ssl_inforrr�infoAsX�
�


�
������r.cCsttjt�ddd��dS)z)Pretty-print the bug information as JSON.T�)�	sort_keys�indentN)�print�json�dumpsr.rrrr�main~sr5�__main__)�__doc__�
__future__rr3rrr,�pip._vendorrrrrr-r%r�ImportError�pip._vendor.urllib3.contribrr)r&r r.r5�__name__rrrr�<module>s4

!=site-packages/pip/_vendor/requests/__pycache__/exceptions.cpython-39.pyc000064400000012504150564745010022416 0ustar00a

��Re��@s�dZddlmZGdd�de�ZGdd�de�ZGdd�de�ZGd	d
�d
e�ZGdd�de�ZGd
d�de�Z	Gdd�de�Z
Gdd�dee
�ZGdd�de
�ZGdd�de�Z
Gdd�de�ZGdd�dee�ZGdd�dee�ZGdd�dee�ZGdd �d ee�ZGd!d"�d"e�ZGd#d$�d$e�ZGd%d&�d&ee�ZGd'd(�d(ee�ZGd)d*�d*e�ZGd+d,�d,e�ZGd-d.�d.e�ZGd/d0�d0ee�ZGd1d2�d2e�Zd3S)4z`
requests.exceptions
~~~~~~~~~~~~~~~~~~~

This module contains the set of Requests' exceptions.
�)�	HTTPErrorcs eZdZdZ�fdd�Z�ZS)�RequestExceptionzTThere was an ambiguous exception that occurred while handling your
    request.
    cs\|�dd�}||_|�dd�|_|durB|jsBt|d�rB|jj|_tt|�j|i|��dS)zBInitialize RequestException with `request` and `response` objects.�responseN�request)�poprr�hasattr�superr�__init__)�self�args�kwargsr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/exceptions.pyr	s�
zRequestException.__init__)�__name__�
__module__�__qualname__�__doc__r	�
__classcell__rrr
rrsrc@seZdZdZdS)�InvalidJSONErrorzA JSON error occurred.N�rrrrrrrrrsrc@seZdZdZdS)rzAn HTTP error occurred.Nrrrrrr src@seZdZdZdS)�ConnectionErrorzA Connection error occurred.Nrrrrrr$src@seZdZdZdS)�
ProxyErrorzA proxy error occurred.Nrrrrrr(src@seZdZdZdS)�SSLErrorzAn SSL error occurred.Nrrrrrr,src@seZdZdZdS)�Timeoutz�The request timed out.

    Catching this error will catch both
    :exc:`~requests.exceptions.ConnectTimeout` and
    :exc:`~requests.exceptions.ReadTimeout` errors.
    Nrrrrrr0src@seZdZdZdS)�ConnectTimeoutz�The request timed out while trying to connect to the remote server.

    Requests that produced this error are safe to retry.
    Nrrrrrr9src@seZdZdZdS)�ReadTimeoutz@The server did not send any data in the allotted amount of time.Nrrrrrr@src@seZdZdZdS)�URLRequiredz*A valid URL is required to make a request.NrrrrrrDsrc@seZdZdZdS)�TooManyRedirectszToo many redirects.NrrrrrrHsrc@seZdZdZdS)�
MissingSchemaz/The URL schema (e.g. http or https) is missing.Nrrrrrr Lsr c@seZdZdZdS)�
InvalidSchemaz"See defaults.py for valid schemas.Nrrrrrr!Psr!c@seZdZdZdS)�
InvalidURLz%The URL provided was somehow invalid.Nrrrrrr"Tsr"c@seZdZdZdS)�
InvalidHeaderz.The header value provided was somehow invalid.Nrrrrrr#Xsr#c@seZdZdZdS)�InvalidProxyURLz"The proxy URL provided is invalid.Nrrrrrr$\sr$c@seZdZdZdS)�ChunkedEncodingErrorz?The server declared chunked encoding but sent an invalid chunk.Nrrrrrr%`sr%c@seZdZdZdS)�ContentDecodingErrorz"Failed to decode response content.Nrrrrrr&dsr&c@seZdZdZdS)�StreamConsumedErrorz3The content for this response was already consumed.Nrrrrrr'hsr'c@seZdZdZdS)�
RetryErrorzCustom retries logic failedNrrrrrr(lsr(c@seZdZdZdS)�UnrewindableBodyErrorz;Requests encountered an error when trying to rewind a body.Nrrrrrr)psr)c@seZdZdZdS)�RequestsWarningzBase warning for Requests.Nrrrrrr*vsr*c@seZdZdZdS)�FileModeWarningzJA file was opened in text mode, but Requests determined its binary length.Nrrrrrr+zsr+c@seZdZdZdS)�RequestsDependencyWarningz@An imported dependency doesn't match the expected version range.Nrrrrrr,~sr,N) rZpip._vendor.urllib3.exceptionsr�
BaseHTTPError�IOErrorrrrrrrrrrr�
ValueErrorr r!r"r#r$r%r&�	TypeErrorr'r(r)�Warningr*�DeprecationWarningr+r,rrrr�<module>s2	site-packages/pip/_vendor/requests/__pycache__/compat.cpython-39.pyc000064400000003163150564745010021521 0ustar00a

��Re��@s�dZddlmZddlZejZeddkZeddkZddlZer�ddl	m
Z
mZmZm
Z
mZmZmZmZmZddlmZmZmZmZmZddlmZddlZdd	lmZdd
lmZddlmZm Z m!Z!m"Z"e#Z$e#Z%e&Z#e'Z'e(e)e*fZ+e(e)fZ,n�e�r�ddl-mZmZmZmZmZm
Z
mZmZm
Z
mZdd
l.mZmZmZmZmZddl/m0Zdd	l1mZdd
l2mZddlm"Z"ddl3mZm Z m!Z!e#Z$e#Z#e%Z%e#e%fZ'e(e*fZ+e(fZ,dS)zq
requests.compat
~~~~~~~~~~~~~~~

This module handles import compatibility issues between Python 2 and
Python 3.
�)�chardetN��)	�quote�unquote�
quote_plus�unquote_plus�	urlencode�
getproxies�proxy_bypass�proxy_bypass_environment�getproxies_environment)�urlparse�
urlunparse�urljoin�urlsplit�	urldefrag)�parse_http_list)�Morsel)�StringIO)�Callable�Mapping�MutableMapping�OrderedDict)
rrrrr	rrrrr)rr
rrr
)�	cookiejar)r)rrr)4�__doc__�pip._vendorr�sys�version_info�_ver�is_py2�is_py3�json�urllibrrrrr	r
rrr
rrrrr�urllib2r�	cookielib�Cookierr�collectionsrrrr�str�builtin_str�bytes�unicode�
basestring�int�long�float�
numeric_types�
integer_types�urllib.parse�urllib.request�httpr�http.cookies�io�collections.abc�r8r8��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/compat.py�<module>sD	,

0site-packages/pip/_vendor/requests/__pycache__/_internal_utils.cpython-39.pyc000064400000002475150564745010023436 0ustar00a

��ReH�@s.dZddlmZmZmZd	dd�Zdd�ZdS)
z�
requests._internal_utils
~~~~~~~~~~~~~~

Provides utility functions that are consumed internally by Requests
which depend on extremely few external helpers (such as compat)
�)�is_py2�builtin_str�str�asciicCs.t|t�r|}ntr |�|�}n
|�|�}|S)z�Given a string object, regardless of type, returns a representation of
    that string in the native string type, encoding and decoding where
    necessary. This assumes ASCII unless told otherwise.
    )�
isinstancerr�encode�decode)�string�encoding�out�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/_internal_utils.py�to_native_strings

rcCs8t|t�sJ�z|�d�WdSty2YdS0dS)z�Determine if unicode string only contains ASCII characters.

    :param str u_string: unicode string to check. Must be unicode
        and not Python 2 `str`.
    :rtype: bool
    rTFN)rrr�UnicodeEncodeError)�u_stringrrr
�unicode_is_asciis
rN)r)�__doc__�compatrrrrrrrrr
�<module>s
site-packages/pip/_vendor/requests/__pycache__/sessions.cpython-39.pyc000064400000046670150564745010022116 0ustar00a

��Re�u�@s�dZddlZddlZddlZddlmZddlmZddlm	Z	ddl
mZmZm
Z
mZmZddlmZmZmZmZdd	lmZmZmZdd
lmZmZddlmZddlmZm Z m!Z!dd
l"m#Z#m$Z$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddlm+Z+m,Z,m-Z-m.Z.m/Z/m0Z0ddl1m2Z2ddlm3Z3ej4dk�rTz
ej5Z6Wne7�yPej8Z6Yn0nejZ6efdd�Z9efdd�Z:Gdd�de;�Z<Gdd�de<�Z=dd�Z>dS)z�
requests.sessions
~~~~~~~~~~~~~~~~~

This module provides a Session object to manage and persist settings across
requests (cookies, auth, proxies).
�N)�	timedelta)�OrderedDict�)�_basic_auth_str)�	cookielib�is_py3�urljoin�urlparse�Mapping)�cookiejar_from_dict�extract_cookies_to_jar�RequestsCookieJar�
merge_cookies)�Request�PreparedRequest�DEFAULT_REDIRECT_LIMIT)�
default_hooks�
dispatch_hook)�to_native_string)�to_key_val_list�default_headers�
DEFAULT_PORTS)�TooManyRedirects�
InvalidSchema�ChunkedEncodingError�ContentDecodingError)�CaseInsensitiveDict)�HTTPAdapter)�requote_uri�get_environ_proxies�get_netrc_auth�should_bypass_proxies�get_auth_from_url�rewind_body)�codes)�REDIRECT_STATI�win32cCsp|dur|S|dur|St|t�r,t|t�s0|S|t|��}|�t|��dd�|��D�}|D]
}||=q`|S)z�Determines appropriate setting for a given request, taking into account
    the explicit setting on that request, and the setting in the session. If a
    setting is a dictionary, they will be merged together using `dict_class`
    NcSsg|]\}}|dur|�qS�N�)�.0�k�vr(r(��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/sessions.py�
<listcomp>J�z!merge_setting.<locals>.<listcomp>)�
isinstancer
r�update�items)�request_setting�session_setting�
dict_class�merged_setting�	none_keys�keyr(r(r,�
merge_setting2s��r8cCs@|dus|�d�gkr|S|dus0|�d�gkr4|St|||�S)z�Properly merges both requests and session hooks.

    This is necessary because when request_hooks == {'response': []}, the
    merge breaks Session hooks entirely.
    N�response)�getr8)�
request_hooks�
session_hooksr4r(r(r,�merge_hooksQs
r=c@s>eZdZdd�Zdd�Zddd	�Zd
d�Zdd
�Zdd�ZdS)�SessionRedirectMixincCs,|jr(|jd}tr|�d�}t|d�SdS)z7Receives a Response. Returns a redirect URI or ``None``�location�latin1�utf8N)�is_redirect�headersr�encoder)�self�respr?r(r(r,�get_redirect_targetbs


z(SessionRedirectMixin.get_redirect_targetcCs�t|�}t|�}|j|jkr dS|jdkrL|jdvrL|jdkrL|jdvrLdS|j|jk}|j|jk}t�|jd�df}|s�|j|vr�|j|vr�dS|p�|S)zFDecide whether Authorization header should be removed when redirectingT�http)�PN�https)i�NFN)r	�hostname�scheme�portrr:)rE�old_url�new_url�
old_parsed�
new_parsed�changed_port�changed_scheme�default_portr(r(r,�should_strip_authws$���z&SessionRedirectMixin.should_strip_authFNTc	
ksFg}
|�|�}t|j�j}|�rB|��}
|
�|�|
dd�|_z
|jWn&tt	t
fyp|jjdd�Yn0t
|j�|jkr�td�|j�|d��|��|�d�r�t|j�}d�t|j�|g�}t|�}|jd	kr�|r�|j|d
�}n|jr�|j}|��}|j�st|jt|��}nt|�}t|�|
_|�|
|�|jtjtjfv�rnd}|D]}|
j �!|d��qRd|
_"|
j }|�!dd�t#|
j$||j�t%|
j$|j&�|
�'|
j$�|�(|
|�}|�)|
|�|
j*du�o�d
|v�p�d|v}|�r�t+|
�|
}|�r|Vq|j,|f|||||dd�|	��}t#|j&|
|j�|�|�}|VqdS)zBReceives a Response. Returns a generator of Responses or Requests.rNF)�decode_contentzExceeded {} redirects.)r9z//�:�)�fragment)�Content-LengthzContent-Type�Transfer-Encoding�CookierZr[)�stream�timeout�verify�cert�proxies�allow_redirects)-rGr	�urlrY�copy�append�history�contentrr�RuntimeError�raw�read�len�
max_redirectsr�format�close�
startswith�joinrrL�_replace�geturl�netlocrr�rebuild_method�status_coder$�temporary_redirect�permanent_redirectrC�pop�bodyr�_cookiesr�cookies�prepare_cookies�rebuild_proxies�rebuild_auth�_body_positionr#�send)rErF�reqr]r^r_r`ra�yield_requests�adapter_kwargs�histrc�previous_fragment�prepared_request�parsed_rurl�parsed�purged_headers�headerrC�
rewindabler(r(r,�resolve_redirects�s~





����
z&SessionRedirectMixin.resolve_redirectscCsR|j}|j}d|vr*|�|jj|�r*|d=|jr8t|�nd}|durN|�|�dS)z�When being redirected we may want to strip authentication from the
        request to avoid leaking credentials. This method intelligently removes
        and reapplies authentication where possible to avoid credential loss.
        �
AuthorizationN)rCrcrU�request�	trust_envr �prepare_auth)rEr�r9rCrc�new_authr(r(r,r~�sz!SessionRedirectMixin.rebuild_authc
Cs�|dur|ni}|j}|j}t|�j}|��}|�d�}t||d�}|jr||s|t||d�}	|	�||	�d��}
|
r||�	||
�d|vr�|d=zt
||�\}}Wnty�d\}}Yn0|r�|r�t||�|d<|S)a�This method re-evaluates the proxy configuration by considering the
        environment variables. If we are redirected to a URL covered by
        NO_PROXY, we strip the proxy configuration. Otherwise, we set missing
        proxy keys for this URL (in case they were stripped by a previous
        redirect).

        This method also replaces the Proxy-Authorization header where
        necessary.

        :rtype: dict
        N�no_proxy�r��allzProxy-Authorization)NN)
rCrcr	rLrdr:r!r�r�
setdefaultr"�KeyErrorr)
rEr�rarCrcrL�new_proxiesr��bypass_proxy�environ_proxies�proxy�username�passwordr(r(r,r}s*


z$SessionRedirectMixin.rebuild_proxiescCsX|j}|jtjkr|dkrd}|jtjkr6|dkr6d}|jtjkrN|dkrNd}||_dS)z�When being redirected we may want to change the method of the request
        based on certain specs or browser behavior.
        �HEAD�GET�POSTN)�methodrur$�	see_other�found�moved)rEr�r9r�r(r(r,rt:sz#SessionRedirectMixin.rebuild_method)FNTNNF)	�__name__�
__module__�__qualname__rGrUr�r~r}rtr(r(r(r,r>`s�
n)r>c@s�eZdZdZgd�Zdd�Zdd�Zdd�Zd	d
�Zd+d
d�Z	dd�Z
dd�Zdd�Zd,dd�Z
d-dd�Zd.dd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zd'd(�Zd)d*�ZdS)/�Sessiona�A Requests session.

    Provides cookie persistence, connection-pooling, and configuration.

    Basic Usage::

      >>> import requests
      >>> s = requests.Session()
      >>> s.get('https://httpbin.org/get')
      <Response [200]>

    Or as a context manager::

      >>> with requests.Session() as s:
      ...     s.get('https://httpbin.org/get')
      <Response [200]>
    )rCr{�authra�hooks�paramsr_r`�adaptersr]r�rlcCsrt�|_d|_i|_t�|_i|_d|_d|_d|_	t
|_d|_t
i�|_t�|_|�dt��|�dt��dS)NFTzhttps://zhttp://)rrCr�rarr�r�r]r_r`rrlr�rr{rr��mountr�rEr(r(r,�__init__js

zSession.__init__cCs|Sr'r(r�r(r(r,�	__enter__�szSession.__enter__cGs|��dSr')rn)rE�argsr(r(r,�__exit__�szSession.__exit__c
Cs�|jpi}t|tj�st|�}ttt�|j�|�}|j}|jrR|sR|jsRt	|j
�}t�}|j|j
��|j
|j|j|jt|j|jtd�t|j|j�t||j�|t|j|j�d�
|S)a�Constructs a :class:`PreparedRequest <PreparedRequest>` for
        transmission and returns it. The :class:`PreparedRequest` has settings
        merged from the :class:`Request <Request>` instance and those of the
        :class:`Session`.

        :param request: :class:`Request` instance to prepare with this
            session's settings.
        :rtype: requests.PreparedRequest
        )r4)
r�rc�files�data�jsonrCr�r�r{r�)r{r/r�	CookieJarrrr
r�r�r rcr�preparer��upperr�r�r�r8rCrr�r=r�)rEr�r{�merged_cookiesr��pr(r(r,�prepare_request�s.

�

�zSession.prepare_requestNTcCsxt|��||||pi||pi|||d�
}|�|�}|p8i}|�|j||
||�}|	|
d�}|�|�|j|fi|��}|S)ax	Constructs a :class:`Request <Request>`, prepares it and sends it.
        Returns :class:`Response <Response>` object.

        :param method: method for the new :class:`Request` object.
        :param url: URL for the new :class:`Request` object.
        :param params: (optional) Dictionary or bytes to be sent in the query
            string for the :class:`Request`.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param json: (optional) json to send in the body of the
            :class:`Request`.
        :param headers: (optional) Dictionary of HTTP Headers to send with the
            :class:`Request`.
        :param cookies: (optional) Dict or CookieJar object to send with the
            :class:`Request`.
        :param files: (optional) Dictionary of ``'filename': file-like-objects``
            for multipart encoding upload.
        :param auth: (optional) Auth tuple or callable to enable
            Basic/Digest/Custom HTTP Auth.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple
        :param allow_redirects: (optional) Set to True by default.
        :type allow_redirects: bool
        :param proxies: (optional) Dictionary mapping protocol or protocol and
            hostname to the URL of the proxy.
        :param stream: (optional) whether to immediately download the response
            content. Defaults to ``False``.
        :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use. Defaults to ``True``. When set to
            ``False``, requests will accept any TLS certificate presented by
            the server, and will ignore hostname mismatches and/or expired
            certificates, which will make your application vulnerable to
            man-in-the-middle (MitM) attacks. Setting verify to ``False`` 
            may be useful during local development or testing.
        :param cert: (optional) if String, path to ssl client cert file (.pem).
            If Tuple, ('cert', 'key') pair.
        :rtype: requests.Response
        )
r�rcrCr�r�r�r�r�r{r�)r^rb)rr�r��merge_environment_settingsrcr0r�)rEr�rcr�r�rCr{r�r�r^rbrar�r]r_r`r�r��prep�settings�send_kwargsrFr(r(r,r��s..�
��
zSession.requestcKs |�dd�|jd|fi|��S)z�Sends a GET request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        rbTr��r�r��rErc�kwargsr(r(r,r:"szSession.getcKs |�dd�|jd|fi|��S)z�Sends a OPTIONS request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        rbT�OPTIONSr�r�r(r(r,�options-szSession.optionscKs |�dd�|jd|fi|��S)z�Sends a HEAD request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        rbFr�r�r�r(r(r,�head8szSession.headcKs|jd|f||d�|��S)a�Sends a POST request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param json: (optional) json to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        r�)r�r��r�)rErcr�r�r�r(r(r,�postCszSession.postcKs|jd|fd|i|��S)auSends a PUT request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        �PUTr�r��rErcr�r�r(r(r,�putPs
zSession.putcKs|jd|fd|i|��S)awSends a PATCH request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        �PATCHr�r�r�r(r(r,�patch\s
z
Session.patchcKs|jd|fi|��S)z�Sends a DELETE request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        �DELETEr�r�r(r(r,�deletehszSession.deletec
Ks�|�d|j�|�d|j�|�d|j�|�d|�||j��t|t�rRtd��|�	dd�}|�
d�}|j}|j|j
d�}t�}|j|fi|��}t�|}	t|	d	�|_td
||fi|��}|jr�|jD]}
t|j|
j|
j�q�t|j||j�|�r|j||fi|��}dd�|D�}ng}|�r@|�d
|�|�	�}||_|�s~z"t|j||fddi|���|_Wnt�y|Yn0|�s�|j|S)zISend a given PreparedRequest.

        :rtype: requests.Response
        r]r_r`raz#You can only send PreparedRequests.rbT)rc)�secondsr9cSsg|]}|�qSr(r()r)rFr(r(r,r-�r.z Session.send.<locals>.<listcomp>rr�)r�r]r_r`r}rar/r�
ValueErrorrxr:r��get_adapterrc�preferred_clockr�r�elapsedrrfrr{r�rir��insert�next�_next�
StopIterationrg)
rEr�r�rbr]r��adapter�start�rr�rF�genrfr(r(r,r�rsF



"zSession.sendc
Cs�|jrn|dur|�d�nd}t||d�}|��D]\}}	|�||	�q0|dusV|durntj�d�pltj�d�}t||j�}t||j	�}t||j
�}t||j�}||||d�S)z^
        Check the environment and merge it with some settings.

        :rtype: dict
        Nr�r�T�REQUESTS_CA_BUNDLE�CURL_CA_BUNDLE)r_rar]r`)r�r:rr1r��os�environr8rar]r_r`)
rErcrar]r_r`r��env_proxiesr*r+r(r(r,r��s 
��z"Session.merge_environment_settingscCs@|j��D]"\}}|���|���r
|Sq
td�|���dS)z~
        Returns the appropriate connection adapter for the given URL.

        :rtype: requests.adapters.BaseAdapter
        z*No connection adapters were found for {!r}N)r�r1�lowerrorrm)rErc�prefixr�r(r(r,r��s
zSession.get_adaptercCs|j��D]}|��q
dS)z+Closes all adapters and as such the sessionN)r��valuesrn)rEr+r(r(r,rn�sz
Session.closecs>||j�<�fdd�|jD�}|D]}|j�|�|j|<q"dS)zwRegisters a connection adapter to a prefix.

        Adapters are sorted in descending order by prefix length.
        cs g|]}t|�t��kr|�qSr()rk)r)r*�r�r(r,r-�r.z!Session.mount.<locals>.<listcomp>N)r�rx)rEr�r��keys_to_mover7r(r�r,r��s
z
Session.mountcs�fdd��jD�}|S)Ncsi|]}|t�|d��qSr')�getattr)r)�attrr�r(r,�
<dictcomp>�r.z(Session.__getstate__.<locals>.<dictcomp>)�	__attrs__)rE�stater(r�r,�__getstate__�szSession.__getstate__cCs"|��D]\}}t|||�qdSr')r1�setattr)rEr�r��valuer(r(r,�__setstate__�szSession.__setstate__)NNNNNNNTNNNNNN)NN)N)N)r�r�r��__doc__r�r�r�r�r�r�r:r�r�r�r�r�r�r�r�r�rnr�r�r�r(r(r(r,r�Qs0>)�
L



Kr�cCst�S)aZ
    Returns a :class:`Session` for context-management.

    .. deprecated:: 1.0.0

        This method has been deprecated since version 1.0.0 and is only kept for
        backwards compatibility. New code should use :class:`~requests.sessions.Session`
        to create a session. This may be removed at a future date.

    :rtype: Session
    )r�r(r(r(r,�sessionsr�)?r�r��sys�time�datetimer�collectionsrr�r�compatrrrr	r
r{rrr
r�modelsrrrr�rr�_internal_utilsr�utilsrrr�
exceptionsrrrr�
structuresrr�rrrr r!r"r#�status_codesr$r%�platform�perf_counterr��AttributeErrorZclockr8r=�objectr>r�r�r(r(r(r,�<module>sB 
r3site-packages/pip/_vendor/requests/__pycache__/status_codes.cpython-39.pyc000064400000010266150564745010022740 0ustar00a

��Re\�E@s�daddlmZddddddd	d
ddd
ddddddddddddddddddd d!d"d#d$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=d>d?d@dAdBdCdDdEdFdG�DZedHdI�ZdJdK�Ze�dLS)Ma
The ``codes`` object defines a mapping from common names for HTTP statuses
to their numerical codes, accessible either as attributes or as dictionary
items.

Example::

    >>> import requests
    >>> requests.codes['temporary_redirect']
    307
    >>> requests.codes.teapot
    418
    >>> requests.codes['\o/']
    200

Some codes have multiple names, and both upper- and lower-case versions of
the names are allowed. For example, ``codes.ok``, ``codes.OK``, and
``codes.okay`` all correspond to the HTTP status code 200.
�)�
LookupDict)�continue)�switching_protocols)�
processing)�
checkpoint)�uri_too_long�request_uri_too_long)�ok�okay�all_ok�all_okay�all_goodz\o/u✓)�created)�accepted)�non_authoritative_info�non_authoritative_information)�
no_content)�
reset_content�reset)�partial_content�partial)�multi_status�multiple_status�multi_stati�multiple_stati)�already_reported)�im_used)�multiple_choices)�moved_permanently�movedz\o-)�found)�	see_other�other)�not_modified)�	use_proxy)�switch_proxy)�temporary_redirect�temporary_moved�	temporary)�permanent_redirect�resume_incomplete�resume)�bad_request�bad)�unauthorized)�payment_required�payment)�	forbidden)�	not_foundz-o-)�method_not_allowed�not_allowed)�not_acceptable)�proxy_authentication_required�
proxy_auth�proxy_authentication)�request_timeout�timeout)�conflict)�gone)�length_required)�precondition_failed�precondition)�request_entity_too_large)�request_uri_too_large)�unsupported_media_type�unsupported_media�
media_type)�requested_range_not_satisfiable�requested_range�range_not_satisfiable)�expectation_failed)�im_a_teapot�teapot�
i_am_a_teapot)�misdirected_request)�unprocessable_entity�
unprocessable)�locked)�failed_dependency�
dependency)�unordered_collection�	unordered)�upgrade_required�upgrade)�precondition_requiredr?)�too_many_requests�too_many)�header_fields_too_large�fields_too_large)�no_response�none)�
retry_with�retry)�$blocked_by_windows_parental_controls�parental_controls)�unavailable_for_legal_reasons�
legal_reasons)�client_closed_request)�internal_server_error�server_errorz/o\u✗)�not_implemented)�bad_gateway)�service_unavailable�unavailable)�gateway_timeout)�http_version_not_supported�http_version)�variant_also_negotiates)�insufficient_storage)�bandwidth_limit_exceeded�	bandwidth)�not_extended)�network_authentication_required�network_auth�network_authentication)D�d�e�f�g�z��������������������i,i-i.i/i0i1i2i3i4i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i��status_codes)�namecs~t��D]8\}}|D]*}tt||�|�d�stt|��|�qqdd��tdurvtdd��fdd�tt�D��ndadS)N)�\�/cSs$d�dd�t|D��}d||fS)Nz, css|]}d|VqdS)z``%s``N�)�.0�nr�r���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/status_codes.py�	<genexpr>s�z%_init.<locals>.doc.<locals>.<genexpr>z* %d: %s)�join�_codes)�code�namesr�r�r��docrsz_init.<locals>.doc�
c3s|]}�|�VqdS)Nr�)r�r��r�r�r�r�xr�z_init.<locals>.<genexpr>)	r��items�setattr�codes�
startswith�upper�__doc__r��sorted)r��titles�titler�r�r��_initks
���r�N)r��
structuresrr�r�r�r�r�r�r��<module>s��P
site-packages/pip/_vendor/requests/__pycache__/__version__.cpython-39.pyc000064400000001122150564745010022510 0ustar00a

��Re��@s,dZdZdZdZdZdZdZdZdZd	Z	d
S)�requestszPython HTTP for Humans.zhttps://requests.readthedocs.ioz2.26.0i&z
Kenneth Reitzzme@kennethreitz.orgz
Apache 2.0zCopyright 2020 Kenneth Reitzu✨ 🍰 ✨N)
�	__title__�__description__�__url__�__version__�	__build__�
__author__�__author_email__�__license__�
__copyright__�__cake__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/__version__.py�<module>ssite-packages/pip/_vendor/requests/__pycache__/certs.cpython-39.pyc000064400000001240150564745010021350 0ustar00a

��Re��@s&dZddlmZedkr"ee��dS)uF
requests.certs
~~~~~~~~~~~~~~

This module returns the preferred default CA certificate bundle. There is
only one — the one from the certifi package.

If you are packaging Requests, e.g., for a Linux distribution or a managed
environment, you can change the definition of where() to return a separately
packaged CA bundle.
�)�where�__main__N)�__doc__�pip._vendor.certifir�__name__�print�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/certs.py�<module>ssite-packages/pip/_vendor/requests/__pycache__/hooks.cpython-39.pyc000064400000002005150564745010021353 0ustar00a

��Re��@sdZdgZdd�Zdd�ZdS)z�
requests.hooks
~~~~~~~~~~~~~~

This module provides the capabilities for the Requests hooks system.

Available hooks:

``response``:
    The response generated from a Request.
�responsecCsdd�tD�S)NcSsi|]
}|g�qS�r)�.0�eventrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/hooks.py�
<dictcomp>�z!default_hooks.<locals>.<dictcomp>)�HOOKSrrrr�
default_hookssr	cKsP|pi}|�|�}|rLt|d�r&|g}|D] }||fi|��}|dur*|}q*|S)z6Dispatches a hook dictionary on a given piece of data.�__call__N)�get�hasattr)�key�hooks�	hook_data�kwargs�hook�
_hook_datarrr�
dispatch_hooks

rN)�__doc__rr	rrrrr�<module>ssite-packages/pip/_vendor/requests/__pycache__/packages.cpython-39.pyc000064400000001035150564745010022010 0ustar00a

��Re��@svddlZdD]dZdeZee�e�e<eej�D]>ZeeksJe�ed�r0ee	d�d�Z
ejeejde
<q0qdS)�N)�urllib3�idna�chardetzpip._vendor.�.zpip._vendor.requests.packages.)�sys�package�vendored_package�
__import__�locals�list�modules�mod�
startswith�len�unprefixed_mod�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/packages.py�<module>ssite-packages/pip/_vendor/requests/__pycache__/auth.cpython-39.pyc000064400000020262150564745010021176 0ustar00a

��Re�'�@s�dZddlZddlZddlZddlZddlZddlZddlmZddl	m
Z
mZmZddl
mZddlmZddlmZd	Zd
Zdd�ZGd
d�de�ZGdd�de�ZGdd�de�ZGdd�de�ZdS)z]
requests.auth
~~~~~~~~~~~~~

This module contains the authentication handlers for Requests.
�N)�	b64encode�)�urlparse�str�
basestring)�extract_cookies_to_jar)�to_native_string)�parse_dict_headerz!application/x-www-form-urlencodedzmultipart/form-datacCs�t|t�s&tjd�|�td�t|�}t|t�sPtjd�t|��td�t|�}t|t�rd|�d�}t|t�rx|�d�}dt	t
d�||f�����}|S)zReturns a Basic Auth string.z�Non-string usernames will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.)�categoryz�Non-string passwords will no longer be supported in Requests 3.0.0. Please convert the object you've passed in ({!r}) to a string or bytes object in the near future to avoid problems.�latin1zBasic �:)
�
isinstancer�warnings�warn�format�DeprecationWarningr�type�encoderr�join�strip)�username�password�authstr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/auth.py�_basic_auth_strs0

��
��



�rc@seZdZdZdd�ZdS)�AuthBasez4Base class that all auth implementations derive fromcCstd��dS)NzAuth hooks must be callable.)�NotImplementedError��self�rrrr�__call__KszAuthBase.__call__N��__name__�
__module__�__qualname__�__doc__r!rrrrrHsrc@s0eZdZdZdd�Zdd�Zdd�Zdd	�Zd
S)�
HTTPBasicAuthz?Attaches HTTP Basic Authentication to the given Request object.cCs||_||_dS�N)rr�rrrrrr�__init__RszHTTPBasicAuth.__init__cCs(t|jt|dd�k|jt|dd�kg�S�Nrr��allr�getattrr�r�otherrrr�__eq__Vs�zHTTPBasicAuth.__eq__cCs
||kSr(rr/rrr�__ne__\szHTTPBasicAuth.__ne__cCst|j|j�|jd<|S)N�
Authorization�rrr�headersrrrrr!_szHTTPBasicAuth.__call__N)r#r$r%r&r*r1r2r!rrrrr'Os
r'c@seZdZdZdd�ZdS)�
HTTPProxyAuthz=Attaches HTTP Proxy Authentication to a given Request object.cCst|j|j�|jd<|S)NzProxy-Authorizationr4rrrrr!gszHTTPProxyAuth.__call__Nr"rrrrr6dsr6c@sPeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dd�Z
dd�ZdS)�HTTPDigestAuthz@Attaches HTTP Digest Authentication to the given Request object.cCs||_||_t��|_dSr()rr�	threading�local�
_thread_localr)rrrr*oszHTTPDigestAuth.__init__cCs@t|jd�s<d|j_d|j_d|j_i|j_d|j_d|j_dS)N�initT�r)�hasattrr:r;�
last_nonce�nonce_count�chal�pos�
num_401_calls)rrrr�init_per_thread_stateusz$HTTPDigestAuth.init_per_thread_statecs�|jjd}|jjd}|jj�d�}|jj�d�}|jj�d�}d�|durTd}n|��}|dksl|dkrzd	d
�}	|	�n@|dkr�dd
�}
|
�n*|dkr�dd�}|�n|dkr�dd�}|��fdd�}
�dur�dSd}t|�}|jp�d}|jr�|d|j7}d|j||jf}d||f}�|�}�|�}||jj	k�rJ|jj
d7_
nd|j_
d|jj
}t|jj
��d�}||�d�7}|t
���d�7}|t�d�7}t�|���dd�}|dk�rЈd|||f�}|�s�|
|d||f�}n<|dk�sd|�d �v�r"d!|||d|f}|
||�}ndS||j_	d"|j||||f}|�rT|d#|7}|�rf|d$|7}|�rx|d%|7}|�r�|d&||f7}d'|S)(z
        :rtype: str
        �realm�nonce�qop�	algorithm�opaqueN�MD5zMD5-SESScSs"t|t�r|�d�}t�|���S�N�utf-8)r
rr�hashlib�md5�	hexdigest��xrrr�md5_utf8�s

z4HTTPDigestAuth.build_digest_header.<locals>.md5_utf8�SHAcSs"t|t�r|�d�}t�|���SrJ)r
rrrL�sha1rNrOrrr�sha_utf8�s

z4HTTPDigestAuth.build_digest_header.<locals>.sha_utf8zSHA-256cSs"t|t�r|�d�}t�|���SrJ)r
rrrL�sha256rNrOrrr�sha256_utf8�s

z7HTTPDigestAuth.build_digest_header.<locals>.sha256_utf8zSHA-512cSs"t|t�r|�d�}t�|���SrJ)r
rrrL�sha512rNrOrrr�sha512_utf8�s

z7HTTPDigestAuth.build_digest_header.<locals>.sha512_utf8cs�d||f�S)N�%s:%sr)�s�d��	hash_utf8rr�<lambda>��z4HTTPDigestAuth.build_digest_header.<locals>.<lambda>�/�?z%s:%s:%srYrz%08xrK���auth�,z%s:%s:%s:%s:%sz>username="%s", realm="%s", nonce="%s", uri="%s", response="%s"z
, opaque="%s"z, algorithm="%s"z
, digest="%s"z , qop="auth", nc=%s, cnonce="%s"z	Digest %s)r:r@�get�upperr�path�queryrrr>r?rr�time�ctime�os�urandomrLrSrN�split)r�method�urlrDrErFrGrH�
_algorithmrQrTrVrX�KD�entdig�p_parsedrh�A1�A2�HA1�HA2�ncvaluerZ�cnonce�respdig�noncebit�baserr\r�build_digest_headers�


��z"HTTPDigestAuth.build_digest_headercKs|jrd|j_dS)z)Reset num_401_calls counter on redirects.rN)�is_redirectr:rB)rr �kwargsrrr�handle_redirect�szHTTPDigestAuth.handle_redirectcKs&d|jkrdks$nd|j_|S|jjdurB|jj�|jj�|j�dd�}d|�	�v�r|jjdk�r|jjd7_t
jd	t
jd
�}t
|jd|dd��|j_|j|��|j��}t|j|j|j�|�|j�|�|j|j�|jd<|jj|fi|��}|j�|�||_|Sd|j_|S)
zo
        Takes the given response and tries digest-auth, if needed.

        :rtype: requests.Response
        i�i�rNzwww-authenticater<�digest�zdigest )�flags)�countr3)�status_coder:rBrA�request�body�seekr5rf�lower�re�compile�
IGNORECASEr	�subr@�content�close�copyr�_cookies�raw�prepare_cookiesr~rorp�
connection�send�history�append)rr r��s_auth�pat�prep�_rrrr�
handle_401�s0	
�
zHTTPDigestAuth.handle_401cCs||��|jjr&|�|j|j�|jd<z|j��|j_	Wnt
yRd|j_	Yn0|�d|j�|�d|j
�d|j_|S)Nr3�responser)rCr:r>r~rorpr5r��tellrA�AttributeError�
register_hookr�r�rBrrrrr!szHTTPDigestAuth.__call__cCs(t|jt|dd�k|jt|dd�kg�Sr+r,r/rrrr1*s�zHTTPDigestAuth.__eq__cCs
||kSr(rr/rrrr20szHTTPDigestAuth.__ne__N)r#r$r%r&r*rCr~r�r�r!r1r2rrrrr7ls
f,r7)r&rlr�rjrLr8r�base64r�compatrrr�cookiesr�_internal_utilsr�utilsr	�CONTENT_TYPE_FORM_URLENCODED�CONTENT_TYPE_MULTI_PARTr�objectrr'r6r7rrrr�<module>s$,site-packages/pip/_vendor/requests/__pycache__/models.cpython-39.pyc000064400000057675150564745010021542 0ustar00a

��Rel��@s�dZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZm
Z
mZmZddlmZdd	lmZdd
lmZddlmZddlmZmZmZdd
lmZmZmZm Z m!Z!m"Z"m#Z#m$Z$ddl%m&Z&m'Z'ddl(m)Z)m*Z*m+Z+m,Z,m-Z-m.Z.m/Z/m0Z0m1Z1m2Z2ddl3m4Z4m5Z5m6Z6m7Z7m8Z8m9Z9m:Z:m;Z;m<Z<m=Z=m>Z>m?Z?ddl3m@ZAddlBmCZCeCjDeCjEeCjFeCjGeCjHfZIdZJdZKdZLGdd�deM�ZNGdd�deM�ZOGdd�deO�ZPGdd�deNeO�ZQGdd�deM�ZRdS) z`
requests.models
~~~~~~~~~~~~~~~

This module contains the primary objects that power Requests.
�N)�RequestField)�encode_multipart_formdata)�	parse_url)�DecodeError�ReadTimeoutError�
ProtocolError�LocationParseError)�UnsupportedOperation�)�
default_hooks)�CaseInsensitiveDict)�
HTTPBasicAuth)�cookiejar_from_dict�get_cookie_header�_copy_cookie_jar)�	HTTPError�
MissingSchema�
InvalidURL�ChunkedEncodingError�ContentDecodingError�ConnectionError�StreamConsumedError�InvalidJSONError)�to_native_string�unicode_is_ascii)
�guess_filename�get_auth_from_url�requote_uri�stream_decode_response_unicode�to_key_val_list�parse_header_links�iter_slices�guess_json_utf�	super_len�check_header_validity)�Callable�Mapping�	cookielib�
urlunparse�urlsplit�	urlencode�str�bytes�is_py2�chardet�builtin_str�
basestring)�json)�codes�i(ic@s0eZdZedd��Zedd��Zedd��ZdS)�RequestEncodingMixincCsNg}t|j�}|j}|sd}|�|�|j}|rD|�d�|�|�d�|�S)zBuild the path URL to use.�/�?�)r)�url�path�append�query�join)�selfr8�pr9r;�r?��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/models.py�path_url=s



zRequestEncodingMixin.path_urlcCs�t|ttf�r|St|d�r |St|d�r�g}t|�D]j\}}t|t�sRt|d�sX|g}|D]B}|dur\|�t|t�r�|�d�n|t|t�r�|�d�n|f�q\q6t|dd�S|SdS)z�Encode parameters in a piece of data.

        Will successfully encode parameters when passed as a dict or a list of
        2-tuples. Order is retained if data is a list of 2-tuples but arbitrary
        if parameters are supplied as a dict.
        �read�__iter__N�utf-8T)�doseq)	�
isinstancer+r,�hasattrrr0r:�encoder*)�data�result�k�vs�vr?r?r@�_encode_paramsRs$	

��z#RequestEncodingMixin._encode_paramscCs�|std��nt|t�r td��g}t|p,i�}t|p8i�}|D]|\}}t|t�s\t|d�sb|g}|D]T}|durft|t�s�t|�}|�t|t�r�|�d�n|t|t�r�|�	d�n|f�qfq@|D]�\}}d}d}	t|t
tf��r t|�dkr�|\}
}n&t|�dk�r|\}
}}n|\}
}}}	nt
|��p,|}
|}t|tttf��rJ|}n(t|d��r`|��}n|du�rnq�n|}t|||
|	d	�}
|
j|d
�|�|
�q�t|�\}}||fS)a�Build the body for a multipart/form-data request.

        Will successfully encode files when passed as a dict or a list of
        tuples. Order is retained if data is a list of tuples but arbitrary
        if parameters are supplied as a dict.
        The tuples may be 2-tuples (filename, fileobj), 3-tuples (filename, fileobj, contentype)
        or 4-tuples (filename, fileobj, contentype, custom_headers).
        zFiles must be provided.zData must not be a string.rCNrD��rB)�namerI�filename�headers)�content_type)�
ValueErrorrFr0rrGr,r+r:�decoderH�tuple�list�lenr�	bytearrayrBr�make_multipartr)�filesrI�
new_fields�fields�field�valrMrK�ft�fh�fn�fp�fdata�rf�bodyrTr?r?r@�
_encode_filesmsT



��


z"RequestEncodingMixin._encode_filesN)�__name__�
__module__�__qualname__�propertyrA�staticmethodrNrhr?r?r?r@r4<s

r4c@seZdZdd�Zdd�ZdS)�RequestHooksMixincCsZ||jvrtd|��t|t�r2|j|�|�n$t|d�rV|j|�dd�|D��dS)zProperly register a hook.z1Unsupported event specified, with event name "%s"rCcss|]}t|t�r|VqdS�N)rFr%)�.0�hr?r?r@�	<genexpr>��z2RequestHooksMixin.register_hook.<locals>.<genexpr>N)�hooksrUrFr%r:rG�extend�r=�event�hookr?r?r@�
register_hook�s


zRequestHooksMixin.register_hookcCs0z|j|�|�WdSty*YdS0dS)ziDeregister a previously registered hook.
        Returns True if the hook existed, False if not.
        TFN)rt�removerUrvr?r?r@�deregister_hook�s
z!RequestHooksMixin.deregister_hookN)rirjrkryr{r?r?r?r@rn�srnc
@s*eZdZdZd	dd�Zdd�Zdd�ZdS)
�Requesta{A user-created :class:`Request <Request>` object.

    Used to prepare a :class:`PreparedRequest <PreparedRequest>`, which is sent to the server.

    :param method: HTTP method to use.
    :param url: URL to send.
    :param headers: dictionary of headers to send.
    :param files: dictionary of {filename: fileobject} files to multipart upload.
    :param data: the body to attach to the request. If a dictionary or
        list of tuples ``[(key, value)]`` is provided, form-encoding will
        take place.
    :param json: json for the body to attach to the request (if files or data is not specified).
    :param params: URL parameters to append to the URL. If a dictionary or
        list of tuples ``[(key, value)]`` is provided, form-encoding will
        take place.
    :param auth: Auth handler or (user, pass) tuple.
    :param cookies: dictionary or CookieJar of cookies to attach to this request.
    :param hooks: dictionary of callback hooks, for internal usage.

    Usage::

      >>> import requests
      >>> req = requests.Request('GET', 'https://httpbin.org/get')
      >>> req.prepare()
      <PreparedRequest [GET]>
    Nc
Cs�|durgn|}|durgn|}|dur,in|}|dur<in|}|	durLin|	}	t�|_t|	���D]\}}|j||d�qd||_||_||_||_||_	|
|_
||_||_||_
dS)N)rwrx)rrtrX�itemsry�methodr8rSr\rIr1�params�auth�cookies)
r=r~r8rSr\rIrr�r�rtr1rKrMr?r?r@�__init__�s"zRequest.__init__cCs
d|jS)Nz<Request [%s]>�r~�r=r?r?r@�__repr__�szRequest.__repr__cCs<t�}|j|j|j|j|j|j|j|j|j	|j
|jd�
|S)zXConstructs a :class:`PreparedRequest <PreparedRequest>` for transmission and returns it.)
r~r8rSr\rIr1rr�r�rt)�PreparedRequest�preparer~r8rSr\rIr1rr�r�rt�r=r>r?r?r@r��s�zRequest.prepare)
NNNNNNNNNN)rirjrk�__doc__r�r�r�r?r?r?r@r|�s�
r|c
@s�eZdZdZdd�Zddd�Zdd�Zd	d
�Zdd�Ze	d
d��Z
dd�Zdd�Zddd�Z
dd�Zd dd�Zdd�Zdd�ZdS)!r�a)The fully mutable :class:`PreparedRequest <PreparedRequest>` object,
    containing the exact bytes that will be sent to the server.

    Instances are generated from a :class:`Request <Request>` object, and
    should not be instantiated manually; doing so may produce undesirable
    effects.

    Usage::

      >>> import requests
      >>> req = requests.Request('GET', 'https://httpbin.org/get')
      >>> r = req.prepare()
      >>> r
      <PreparedRequest [GET]>

      >>> s = requests.Session()
      >>> s.send(r)
      <Response [200]>
    cCs0d|_d|_d|_d|_d|_t�|_d|_dSro)r~r8rS�_cookiesrgrrt�_body_positionr�r?r?r@r�%szPreparedRequest.__init__NcCsR|�|�|�||�|�|�|�|�|�|||
�|�||�|�|	�dS)z6Prepares the entire request with the given parameters.N)�prepare_method�prepare_url�prepare_headers�prepare_cookies�prepare_body�prepare_auth�
prepare_hooks)r=r~r8rSr\rIrr�r�rtr1r?r?r@r�6s


zPreparedRequest.preparecCs
d|jS)Nz<PreparedRequest [%s]>r�r�r?r?r@r�HszPreparedRequest.__repr__cCsXt�}|j|_|j|_|jdur*|j��nd|_t|j�|_|j|_|j|_|j	|_	|Sro)
r�r~r8rS�copyrr�rgrtr�r�r?r?r@r�KszPreparedRequest.copycCs$||_|jdur t|j���|_dS)zPrepares the given HTTP method.N)r~r�upper)r=r~r?r?r@r�Vs
zPreparedRequest.prepare_methodcCsBddlm}z|j|dd��d�}Wn|jy<t�Yn0|S)Nr)�idnaT)�uts46rD)�pip._vendorr�rHrV�	IDNAError�UnicodeError)�hostr�r?r?r@�_get_idna_encoded_host\s
z&PreparedRequest._get_idna_encoded_hostc
Cs,t|t�r|�d�}ntr"t|�nt|�}|��}d|vrR|���d�sR||_	dSzt
|�\}}}}}}}	Wn.ty�}
zt|
j
��WYd}
~
n
d}
~
00|s�d}|�t|d��}t|��|s�td|��t|��sz|�|�}Wnty�td��Yn0n|�d��rtd��|�pd	}|�r.|d
7}||7}|�rL|dt|�7}|�sVd}t�r�t|t��rr|�d�}t|t��r�|�d�}t|t��r�|�d�}t|t��r�|�d�}t|	t��r�|	�d�}	t|ttf��r�t|�}|�|�}
|
�r
|�rd
||
f}n|
}tt|||d||	g��}||_	dS)zPrepares the given HTTP URL.�utf8�:�httpNzDInvalid URL {0!r}: No schema supplied. Perhaps you meant http://{0}?z Invalid URL %r: No host suppliedzURL has an invalid label.�*r7�@r5rDz%s&%s)rFr,rVr-�unicoder+�lstrip�lower�
startswithr8rrr�args�formatrrrr�r�rHrNrr()r=r8r�schemer�r��portr9r;�fragment�e�error�netloc�
enc_paramsr?r?r@r�fsh
 







zPreparedRequest.prepare_urlcCs<t�|_|r8|��D]"}t|�|\}}||jt|�<qdS)z Prepares the given HTTP headers.N)rrSr}r$r)r=rS�headerrQ�valuer?r?r@r��szPreparedRequest.prepare_headersc	
Cs�d}d}|sp|durpd}ztj|dd�}Wn0tyZ}zt||d��WYd}~n
d}~00t|t�sp|�d�}tt|d�t|t	t
ttf�g�}|�r8zt
|�}Wntttfy�d}Yn0|}t|dd�du�rz|��|_Wn ttf�yt�|_Yn0|�rtd	��|�r,t|�|jd
<n
d|jd<np|�rP|�||�\}}n2|�r�|�|�}t|t	��sxt|d
��r~d}nd}|�|�|�r�d|jv�r�||jd<||_dS)z"Prepares the given HTTP body data.Nzapplication/jsonF)�	allow_nan)�requestrDrC�tellz1Streamed bodies and files are mutually exclusive.�Content-Length�chunkedzTransfer-EncodingrBz!application/x-www-form-urlencodedzcontent-typezContent-Type)�complexjson�dumpsrUrrFr,rH�allrGr0rXrWr&r#�	TypeError�AttributeErrorr	�getattrr�r��IOError�OSError�object�NotImplementedErrorr/rSrhrN�prepare_content_lengthrg)	r=rIr\r1rgrT�ve�	is_stream�lengthr?r?r@r��sR"

�



zPreparedRequest.prepare_bodycCsL|dur$t|�}|rHt|�|jd<n$|jdvrH|j�d�durHd|jd<dS)z>Prepare Content-Length header based on request method and bodyNr�)�GET�HEAD�0)r#r/rSr~�get)r=rgr�r?r?r@r�sz&PreparedRequest.prepare_content_lengthr7cCsj|dur"t|j�}t|�r|nd}|rft|t�rDt|�dkrDt|�}||�}|j�|j�|�	|j
�dS)z"Prepares the given HTTP auth data.NrO)rr8�anyrFrWrYr
�__dict__�updater�rg)r=r�r8�url_auth�rr?r?r@r�s
zPreparedRequest.prepare_authcCs@t|tj�r||_n
t|�|_t|j|�}|dur<||jd<dS)aPrepares the given HTTP cookie data.

        This function eventually generates a ``Cookie`` header from the
        given cookies using cookielib. Due to cookielib's design, the header
        will not be regenerated if it already exists, meaning this function
        can only be called once for the life of the
        :class:`PreparedRequest <PreparedRequest>` object. Any subsequent calls
        to ``prepare_cookies`` will have no actual effect, unless the "Cookie"
        header is removed beforehand.
        N�Cookie)rFr'�	CookieJarr�rrrS)r=r��
cookie_headerr?r?r@r�4s
zPreparedRequest.prepare_cookiescCs&|pg}|D]}|�|||�qdS)zPrepares the given hooks.N)ry)r=rtrwr?r?r@r�HszPreparedRequest.prepare_hooks)
NNNNNNNNNN)N)r7)rirjrkr�r�r�r�r�r�rmr�r�r�r�r�r�r�r�r?r?r?r@r�s"�

	V
J
r�c@s�eZdZdZgd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zdd�Z
edd��Zedd��Zedd��Zedd��Zedd��Zd2d!d"�Zed d#fd$d%�Zed&d'��Zed(d)��Zd*d+�Zed,d-��Zd.d/�Zd0d1�Zd#S)3�ResponsezhThe :class:`Response <Response>` object, which contains a
    server's response to an HTTP request.
    )
�_content�status_coderSr8�history�encoding�reasonr��elapsedr�cCs^d|_d|_d|_d|_t�|_d|_d|_d|_g|_	d|_
ti�|_t
�d�|_d|_dS)NFr)r��_content_consumed�_nextr�rrS�rawr8r�r�r�rr��datetime�	timedeltar�r�r�r?r?r@r�\s
zResponse.__init__cCs|Sror?r�r?r?r@�	__enter__�szResponse.__enter__cGs|��dSro)�close)r=r�r?r?r@�__exit__�szResponse.__exit__cs �js�j�fdd��jD�S)Ncsi|]}|t�|d��qSro)r�)rp�attrr�r?r@�
<dictcomp>�rsz)Response.__getstate__.<locals>.<dictcomp>)r��content�	__attrs__r�r?r�r@�__getstate__�szResponse.__getstate__cCs:|��D]\}}t|||�qt|dd�t|dd�dS)Nr�Tr�)r}�setattr)r=�staterQr�r?r?r@�__setstate__�szResponse.__setstate__cCs
d|jS)Nz<Response [%s]>)r�r�r?r?r@r��szResponse.__repr__cCs|jS�akReturns True if :attr:`status_code` is less than 400.

        This attribute checks if the status code of the response is between
        400 and 600 to see if there was a client error or a server error. If
        the status code, is between 200 and 400, this will return True. This
        is **not** a check to see if the response code is ``200 OK``.
        ��okr�r?r?r@�__bool__�szResponse.__bool__cCs|jSr�r�r�r?r?r@�__nonzero__�szResponse.__nonzero__cCs
|�d�S)z,Allows you to use a response as an iterator.�)�iter_contentr�r?r?r@rC�szResponse.__iter__cCs&z|��Wnty YdS0dS)axReturns True if :attr:`status_code` is less than 400, False if not.

        This attribute checks if the status code of the response is between
        400 and 600 to see if there was a client error or a server error. If
        the status code is between 200 and 400, this will return True. This
        is **not** a check to see if the response code is ``200 OK``.
        FT)�raise_for_statusrr�r?r?r@r��s
	zResponse.okcCsd|jvo|jtvS)z�True if this Response is a well-formed HTTP redirect that could have
        been processed automatically (by :meth:`Session.resolve_redirects`).
        �location)rSr��REDIRECT_STATIr�r?r?r@�is_redirect�szResponse.is_redirectcCsd|jvo|jtjtjfvS)z@True if this Response one of the permanent versions of redirect.r�)rSr�r2�moved_permanently�permanent_redirectr�r?r?r@�is_permanent_redirect�szResponse.is_permanent_redirectcCs|jS)zTReturns a PreparedRequest for the next request in a redirect chain, if there is one.)r�r�r?r?r@�next�sz
Response.nextcCst�|j�dS)zOThe apparent encoding, provided by the charset_normalizer or chardet libraries.r�)r.�detectr�r�r?r?r@�apparent_encoding�szResponse.apparent_encodingr
Fcs|��fdd�}�jr(t�jt�r(t��n"�durJt�t�sJtdt����t�j��}|�}�jrf|n|}|rxt	|��}|S)a�Iterates over the response data.  When stream=True is set on the
        request, this avoids reading the content at once into memory for
        large responses.  The chunk size is the number of bytes it should
        read into memory.  This is not necessarily the length of each item
        returned as decoding can take place.

        chunk_size must be of type int or None. A value of None will
        function differently depending on the value of `stream`.
        stream=True will read data as it arrives in whatever size the
        chunks are received. If stream=False, data is returned as
        a single chunk.

        If decode_unicode is True, content will be decoded using the best
        available encoding based on the response.
        c
3s�t�jd�r�z �jj�dd�D]
}|VqWq�tyX}zt|��WYd}~q�d}~0ty�}zt|��WYd}~q�d}~0ty�}zt|��WYd}~q�d}~00n�j�	��}|s�q�|Vq�d�_
dS)N�streamT)�decode_content)rGr�r�rrrrrrrBr�)�chunkr���
chunk_sizer=r?r@�generate�s z'Response.iter_content.<locals>.generateNz.chunk_size must be an int, it is instead a %s.)
r�rFr��boolr�intr��typer!r)r=r��decode_unicoder��
reused_chunks�
stream_chunks�chunksr?r�r@r��s
zResponse.iter_contentNccs�d}|j||d�D]n}|dur&||}|r6|�|�}n|��}|rl|drl|rl|dd|dkrl|��}nd}|D]
}|Vqtq|dur�|VdS)z�Iterates over the response data, one line at a time.  When
        stream=True is set on the request, this avoids reading the
        content at once into memory for large responses.

        .. note:: This method is not reentrant safe.
        N)r�r���)r��split�
splitlines�pop)r=r�r�	delimiter�pendingr��lines�liner?r?r@�
iter_liness$

zResponse.iter_linescCsV|jdurJ|jrtd��|jdks,|jdur4d|_nd�|�t��pFd|_d|_|jS)z"Content of the response, in bytes.Fz2The content for this response was already consumedrNrsT)r�r��RuntimeErrorr�r�r<r��CONTENT_CHUNK_SIZEr�r?r?r@r�7s
�zResponse.contentc	Csfd}|j}|jstd�S|jdur(|j}zt|j|dd�}Wn$ttfy`t|jdd�}Yn0|S)a�Content of the response, in unicode.

        If Response.encoding is None, encoding will be guessed using
        ``charset_normalizer`` or ``chardet``.

        The encoding of the response content is determined based solely on HTTP
        headers, following RFC 2616 to the letter. If you can take advantage of
        non-HTTP knowledge to make a better guess at the encoding, you should
        set ``r.encoding`` appropriately before accessing this property.
        Nr7�replace)�errors)r�r�r+r��LookupErrorr�)r=r�r�r?r?r@�textKs
z
Response.textcKsp|js\|jr\t|j�dkr\t|j�}|dur\ztj|j�|�fi|��WStyZYn0tj|jfi|��S)a.Returns the json-encoded content of a response, if any.

        :param \*\*kwargs: Optional arguments that ``json.loads`` takes.
        :raises simplejson.JSONDecodeError: If the response body does not
            contain valid json and simplejson is installed.
        :raises json.JSONDecodeError: If the response body does not contain
            valid json and simplejson is not installed on Python 3.
        :raises ValueError: If the response body does not contain valid
            json and simplejson is not installed on Python 2.        
        rPN)	r�r�rYr"r��loadsrV�UnicodeDecodeErrorr)r=�kwargsr�r?r?r@r1qs

��z
Response.jsoncCsF|j�d�}i}|rBt|�}|D] }|�d�p6|�d�}|||<q |S)z8Returns the parsed header links of the response, if any.�link�relr8)rSr�r )r=r��l�linksr�keyr?r?r@r�s
zResponse.linkscCs�d}t|jt�rBz|j�d�}WqHty>|j�d�}YqH0n|j}d|jkr^dkrvnnd|j||jf}n,d|jkr�dkr�nnd|j||jf}|r�t||d	��d
S)z+Raises :class:`HTTPError`, if one occurred.r7rDz
iso-8859-1i�i�z%s Client Error: %s for url: %siXz%s Server Error: %s for url: %s)�responseN)rFr�r,rVrr�r8r)r=�http_error_msgr�r?r?r@r��szResponse.raise_for_statuscCs0|js|j��t|jdd�}|dur,|�dS)z�Releases the connection back to the pool. Once this method has been
        called the underlying ``raw`` object must not be accessed again.

        *Note: Should not normally need to be called explicitly.*
        �release_connN)r�r�r�r�)r=rr?r?r@r��s

zResponse.close)r
F)rirjrkr�r�r�r�r�r�r�r�r�r�rCrlr�r�r�r�r�r��ITER_CHUNK_SIZErr�rr1rr�r�r?r?r?r@r�Rs@/







7

%
r�)Sr�r��sys�encodings.idna�	encodingsZpip._vendor.urllib3.fieldsrZpip._vendor.urllib3.filepostr�pip._vendor.urllib3.utilrZpip._vendor.urllib3.exceptionsrrrr�ior	rtr�
structuresrr�r
r�rrr�
exceptionsrrrrrrrr�_internal_utilsrr�utilsrrrrrr r!r"r#r$�compatr%r&r'r(r)r*r+r,r-r.r/r0r1r��status_codesr2�moved�found�other�temporary_redirectr�r��DEFAULT_REDIRECT_LIMITrrr�r4rnr|r�r�r?r?r?r@�<module>sD(08�rJDsite-packages/pip/_vendor/requests/__pycache__/adapters.cpython-39.pyc000064400000041164150564745010022044 0ustar00a

��Re,T�@s�dZddlZddlZddlmZmZddlmZddl	m
Z
ddl	mZddl
mZddlmZdd	lmZdd
lmZddlmZddlmZdd
lmZddlmZddlmZddlmZddlmZddlmZddlmZddl m!Z!m"Z"ddl#m$Z$m%Z%m&Z&m'Z'm(Z(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1m2Z2mZmZm3Z3m4Z4m5Z5m6Z6ddl7m8Z8zddl9m:Z:Wne;�y�dd�Z:Yn0dZ<dZ=dZ>dZ?Gd d!�d!e@�ZAGd"d#�d#eA�ZBdS)$z�
requests.adapters
~~~~~~~~~~~~~~~~~

This module contains the transport adapters that Requests uses to define
and maintain connections.
�N)�PoolManager�proxy_from_url)�HTTPResponse)�	parse_url)�Timeout)�Retry)�ClosedPoolError)�ConnectTimeoutError)�	HTTPError)�
MaxRetryError)�NewConnectionError)�
ProxyError)�
ProtocolError)�ReadTimeoutError)�SSLError)�
ResponseError)�LocationValueError�)�Response)�urlparse�
basestring)�DEFAULT_CA_BUNDLE_PATH�extract_zipped_paths�get_encoding_from_headers�prepend_scheme_if_needed�get_auth_from_url�
urldefragauth�select_proxy)�CaseInsensitiveDict)�extract_cookies_to_jar)	�ConnectionError�ConnectTimeout�ReadTimeoutrr
�
RetryError�
InvalidSchema�InvalidProxyURL�
InvalidURL)�_basic_auth_str)�SOCKSProxyManagercOstd��dS)Nz'Missing dependencies for SOCKS support.)r$)�args�kwargs�r+��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/adapters.pyr(.sr(F�
cs2eZdZdZ�fdd�Zddd�Zd	d
�Z�ZS)�BaseAdapterzThe Base Transport Adaptercstt|���dS�N)�superr.�__init__��self��	__class__r+r,r1:szBaseAdapter.__init__FNTcCst�dS)aCSends PreparedRequest object. Returns Response object.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param stream: (optional) Whether to stream the request content.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple
        :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use
        :param cert: (optional) Any user-provided SSL certificate to be trusted.
        :param proxies: (optional) The proxies dictionary to apply to the request.
        N��NotImplementedError)r3�request�stream�timeout�verify�cert�proxiesr+r+r,�send=szBaseAdapter.sendcCst�dS)z!Cleans up adapter specific items.Nr6r2r+r+r,�closeOszBaseAdapter.close)FNTNN)�__name__�
__module__�__qualname__�__doc__r1r>r?�
__classcell__r+r+r4r,r.7s�
r.cs�eZdZdZgd�Zeeeef�fdd�	Zdd�Z	dd�Z
efd	d
�Zdd�Zd
d�Z
dd�Zd dd�Zdd�Zdd�Zdd�Zdd�Zd!dd�Z�ZS)"�HTTPAdaptera�The built-in HTTP Adapter for urllib3.

    Provides a general-case interface for Requests sessions to contact HTTP and
    HTTPS urls by implementing the Transport Adapter interface. This class will
    usually be created by the :class:`Session <Session>` class under the
    covers.

    :param pool_connections: The number of urllib3 connection pools to cache.
    :param pool_maxsize: The maximum number of connections to save in the pool.
    :param max_retries: The maximum number of retries each connection
        should attempt. Note, this applies only to failed DNS lookups, socket
        connections and connection timeouts, never to requests where data has
        made it to the server. By default, Requests does not retry failed
        connections. If you need granular control over the conditions under
        which we retry a request, import urllib3's ``Retry`` class and pass
        that instead.
    :param pool_block: Whether the connection pool should block for connections.

    Usage::

      >>> import requests
      >>> s = requests.Session()
      >>> a = requests.adapters.HTTPAdapter(max_retries=3)
      >>> s.mount('http://', a)
    )�max_retries�config�_pool_connections�
_pool_maxsize�_pool_blockcsd|tkrtddd�|_nt�|�|_i|_i|_tt|���||_	||_
||_|j|||d�dS)NrF)�read��block)
�DEFAULT_RETRIESrrF�from_intrG�
proxy_managerr0rEr1rHrIrJ�init_poolmanager)r3�pool_connections�pool_maxsizerF�
pool_blockr4r+r,r1qszHTTPAdapter.__init__cs�fdd��jD�S)Ncsi|]}|t�|d��qSr/)�getattr)�.0�attrr2r+r,�
<dictcomp>��z,HTTPAdapter.__getstate__.<locals>.<dictcomp>)�	__attrs__r2r+r2r,�__getstate__�szHTTPAdapter.__getstate__cCsDi|_i|_|��D]\}}t|||�q|j|j|j|jd�dS)NrL)rPrG�items�setattrrQrHrIrJ)r3�staterW�valuer+r+r,�__setstate__�s�zHTTPAdapter.__setstate__cKs0||_||_||_tf|||dd�|��|_dS)aInitializes a urllib3 PoolManager.

        This method should not be called from user code, and is only
        exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param connections: The number of urllib3 connection pools to cache.
        :param maxsize: The maximum number of connections to save in the pool.
        :param block: Block when no free connections are available.
        :param pool_kwargs: Extra keyword arguments used to initialize the Pool Manager.
        T)�	num_pools�maxsizerM�strictN)rHrIrJr�poolmanager)r3�connectionsrbrM�pool_kwargsr+r+r,rQ�s
��zHTTPAdapter.init_poolmanagercKs�||jvr|j|}n||���d�r^t|�\}}t|f|||j|j|jd�|��}|j|<n4|�|�}t	|f||j|j|jd�|��}|j|<|S)a�Return urllib3 ProxyManager for the given proxy.

        This method should not be called from user code, and is only
        exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param proxy: The proxy to return a urllib3 ProxyManager for.
        :param proxy_kwargs: Extra keyword arguments used to configure the Proxy Manager.
        :returns: ProxyManager
        :rtype: urllib3.ProxyManager
        �socks)�username�passwordrarbrM)�
proxy_headersrarbrM)
rP�lower�
startswithrr(rHrIrJrjr)r3�proxy�proxy_kwargs�managerrhrirjr+r+r,�proxy_manager_for�s6
���

���zHTTPAdapter.proxy_manager_forcCs�|���d�rn|rnd}|dur"|}|s.tt�}|r>tj�|�sLtd�|���d|_	tj�
|�sf||_q�||_nd|_	d|_d|_|r�t
|t�s�|d|_|d|_n||_d|_|jr�tj�|j�s�td	�|j���|jr�tj�|j�s�td
�|j���dS)aAVerify a SSL certificate. This method should not be called from user
        code, and is only exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param conn: The urllib3 connection object associated with the cert.
        :param url: The requested URL.
        :param verify: Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use
        :param cert: The SSL certificate to verify.
        �httpsNTzECould not find a suitable TLS CA certificate bundle, invalid path: {}�
CERT_REQUIRED�	CERT_NONErrz9Could not find the TLS certificate file, invalid path: {}z1Could not find the TLS key file, invalid path: {})rkrlrr�os�path�exists�IOError�format�	cert_reqs�isdir�ca_certs�ca_cert_dir�
isinstancer�	cert_file�key_file)r3�conn�urlr;r<�cert_locr+r+r,�cert_verify�s>�

��zHTTPAdapter.cert_verifycCs�t�}t|dd�|_tt|di��|_t|j�|_||_|jj|_t	|j
t�r^|j
�d�|_
n|j
|_
t
|j||�||_||_|S)a�Builds a :class:`Response <requests.Response>` object from a urllib3
        response. This should not be called from user code, and is only exposed
        for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`

        :param req: The :class:`PreparedRequest <PreparedRequest>` used to generate the response.
        :param resp: The urllib3 response object.
        :rtype: requests.Response
        �statusN�headers�utf-8)rrU�status_coderr�r�encoding�raw�reasonr}r��bytes�decoder�cookiesr8�
connection)r3�req�resp�responser+r+r,�build_response�s

zHTTPAdapter.build_responseNcCsdt||�}|rDt|d�}t|�}|js.td��|�|�}|�|�}nt|�}|��}|j	�|�}|S)a�Returns a urllib3 connection for the given URL. This should not be
        called from user code, and is only exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param url: The URL to connect to.
        :param proxies: (optional) A Requests-style dictionary of proxies used on this request.
        :rtype: urllib3.ConnectionPool
        �httpzFPlease check proxy URL. It is malformed and could be missing the host.)
rrr�hostr%rp�connection_from_urlr�geturlrd)r3r�r=rm�	proxy_urlrPr��parsedr+r+r,�get_connection$s	


zHTTPAdapter.get_connectioncCs&|j��|j��D]}|��qdS)z�Disposes of any internal state.

        Currently, this closes the PoolManager and any active ProxyManager,
        which closes any pooled connections.
        N)rd�clearrP�values)r3rmr+r+r,r??s
zHTTPAdapter.closec	Cs`t|j|�}t|j�j}|o"|dk}d}|rDt|�j��}|�d�}|j}|r\|s\t|j�}|S)a?Obtain the url to use when making the final request.

        If the message is being sent through a HTTP proxy, the full URL has to
        be used. Otherwise, we should only use the path portion of the URL.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs.
        :rtype: str
        rqFrg)rr�r�schemerkrl�path_urlr)	r3r8r=rmr��is_proxied_http_request�using_socks_proxy�proxy_schemer�r+r+r,�request_urlIs

zHTTPAdapter.request_urlcKsdS)a"Add any headers needed by the connection. As of v2.0 this does
        nothing by default, but is left for overriding by users that subclass
        the :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param request: The :class:`PreparedRequest <PreparedRequest>` to add headers to.
        :param kwargs: The keyword arguments from the call to send().
        Nr+)r3r8r*r+r+r,�add_headersfszHTTPAdapter.add_headerscCs&i}t|�\}}|r"t||�|d<|S)aReturns a dictionary of the headers to add to any request sent
        through a proxy. This works with urllib3 magic to ensure that they are
        correctly sent to the proxy, rather than in a tunnelled request if
        CONNECT is being used.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param proxy: The url of the proxy being used for this request.
        :rtype: dict
        zProxy-Authorization)rr')r3rmr�rhrir+r+r,rjts
�zHTTPAdapter.proxy_headersFTcCs�z|�|j|�}Wn0tyB}zt||d��WYd}~n
d}~00|�||j||�|�||�}	|j||||||d�|jdup�d|jv}
t	|t
�r�z|\}}t||d�}Wn6ty�}zd�
|�}
t|
��WYd}~n
d}~00nt	|t�r�nt||d�}�zD|
�s6|j|j|	|j|jdddd|j|d�
}�nt|d	��rH|j}|jtd
�}z�|j|j|	dd�|j��D]\}}|�||��qr|��|jD]D}|�tt|��d
d��d��|�d�|�|�|�d��q�|�d�z|jdd�}Wnt�y|��}Yn0tj |||ddd�}Wn|�!��Yn0W�n�t"t#j$f�y�}
zt%|
|d��WYd}
~
�njd}
~
0t&�y(}z�t	|j't(��r�t	|j't)��s�t*||d��t	|j't+��r�t,||d��t	|j't-��r�t.||d��t	|j't/��rt0||d��t%||d��WYd}~n�d}~0t1�yX}zt%||d��WYd}~n�d}~0t-�y�}zt.|��WYd}~nhd}~0t/t2f�y�}zBt	|t/��r�t0||d��nt	|t3��r�t4||d��n�WYd}~n
d}~00|�5||�S)aSends PreparedRequest object. Returns Response object.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param stream: (optional) Whether to stream the request content.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple or urllib3 Timeout object
        :param verify: (optional) Either a boolean, in which case it controls whether
            we verify the server's TLS certificate, or a string, in which case it
            must be a path to a CA bundle to use
        :param cert: (optional) Any user-provided SSL certificate to be trusted.
        :param proxies: (optional) The proxies dictionary to apply to the request.
        :rtype: requests.Response
        )r8N)r9r:r;r<r=zContent-Length)�connectrKzrInvalid timeout {}. Pass a (connect, read) timeout tuple, or a single float to set both timeouts to the same valueF)
�methodr��bodyr��redirect�assert_same_host�preload_content�decode_content�retriesr:�
proxy_pool)r:T)�skip_accept_encoding�r�s
s0

)�	buffering)�poolr�r�r�)6r�r�rr&r�r�r�r�r�r}�tuple�TimeoutSauce�
ValueErrorrx�urlopenr�rF�hasattrr��	_get_conn�DEFAULT_POOL_TIMEOUT�
putrequestr\�	putheader�
endheadersr>�hex�len�encode�getresponse�	TypeErrorr�from_httplibr?r�socket�errorr rr�r	rr!rr#�_ProxyErrorr
�	_SSLErrorrr�
_HTTPErrorrr"r�)r3r8r9r:r;r<r=r��er��chunkedr�rK�errr��low_conn�headerr_�i�rr+r+r,r>�s�"
� 
�
�
 


�
"  zHTTPAdapter.send)N)FNTNN)r@rArBrCrZ�DEFAULT_POOLSIZErN�DEFAULT_POOLBLOCKr1r[r`rQrpr�r�r�r?r�r�rjr>rDr+r+r4r,rETs$�%4%

rE)CrC�os.pathrtr�Zpip._vendor.urllib3.poolmanagerrrZpip._vendor.urllib3.responser�pip._vendor.urllib3.utilrrr�Zpip._vendor.urllib3.util.retryrZpip._vendor.urllib3.exceptionsrr	r
r�rrr
r�rrrr�rr�modelsr�compatrr�utilsrrrrrrr�
structuresrr�r�
exceptionsr r!r"r#r$r%r&�authr'Z!pip._vendor.urllib3.contrib.socksr(�ImportErrorr�r�rNr��objectr.rEr+r+r+r,�<module>sF$,site-packages/pip/_vendor/requests/__pycache__/api.cpython-39.pyc000064400000015146150564745010021013 0ustar00a

��Re�@s\dZddlmZdd�Zddd�Zdd	�Zd
d�Zddd
�Zddd�Zddd�Z	dd�Z
dS)z�
requests.api
~~~~~~~~~~~~

This module implements the Requests API.

:copyright: (c) 2012 by Kenneth Reitz.
:license: Apache2, see LICENSE for more details.
�)�sessionscKsBt���&}|jf||d�|��Wd�S1s40YdS)a�	Constructs and sends a :class:`Request <Request>`.

    :param method: method for the new :class:`Request` object: ``GET``, ``OPTIONS``, ``HEAD``, ``POST``, ``PUT``, ``PATCH``, or ``DELETE``.
    :param url: URL for the new :class:`Request` object.
    :param params: (optional) Dictionary, list of tuples or bytes to send
        in the query string for the :class:`Request`.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) A JSON serializable Python object to send in the body of the :class:`Request`.
    :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`.
    :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`.
    :param files: (optional) Dictionary of ``'name': file-like-objects`` (or ``{'name': file-tuple}``) for multipart encoding upload.
        ``file-tuple`` can be a 2-tuple ``('filename', fileobj)``, 3-tuple ``('filename', fileobj, 'content_type')``
        or a 4-tuple ``('filename', fileobj, 'content_type', custom_headers)``, where ``'content-type'`` is a string
        defining the content type of the given file and ``custom_headers`` a dict-like object containing additional headers
        to add for the file.
    :param auth: (optional) Auth tuple to enable Basic/Digest/Custom HTTP Auth.
    :param timeout: (optional) How many seconds to wait for the server to send data
        before giving up, as a float, or a :ref:`(connect timeout, read
        timeout) <timeouts>` tuple.
    :type timeout: float or tuple
    :param allow_redirects: (optional) Boolean. Enable/disable GET/OPTIONS/POST/PUT/PATCH/DELETE/HEAD redirection. Defaults to ``True``.
    :type allow_redirects: bool
    :param proxies: (optional) Dictionary mapping protocol to the URL of the proxy.
    :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use. Defaults to ``True``.
    :param stream: (optional) if ``False``, the response content will be immediately downloaded.
    :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response

    Usage::

      >>> import requests
      >>> req = requests.request('GET', 'https://httpbin.org/get')
      >>> req
      <Response [200]>
    )�method�urlN)r�Session�request)rr�kwargs�session�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/api.pyrs,
rNcKstd|fd|i|��S)adSends a GET request.

    :param url: URL for the new :class:`Request` object.
    :param params: (optional) Dictionary, list of tuples or bytes to send
        in the query string for the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �get�params�r)rrrr	r	r
r@srcKstd|fi|��S)z�Sends an OPTIONS request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �optionsr
�rrr	r	r
rNs	rcKs|�dd�td|fi|��S)akSends a HEAD request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes. If
        `allow_redirects` is not provided, it will be set to `False` (as
        opposed to the default :meth:`request` behavior).
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �allow_redirectsF�head)�
setdefaultrrr	r	r
rZsrcKstd|f||d�|��S)a�Sends a POST request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �post)�data�jsonr
)rrrrr	r	r
risrcKstd|fd|i|��S)a�Sends a PUT request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �putrr
�rrrr	r	r
rxsrcKstd|fd|i|��S)a�Sends a PATCH request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �patchrr
rr	r	r
r�srcKstd|fi|��S)z�Sends a DELETE request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    �deleter
rr	r	r
r�s	r)N)NN)N)N)�__doc__�rrrrrrrrrr	r	r	r
�<module>s
0



site-packages/pip/_vendor/requests/__pycache__/__init__.cpython-39.pyc000064400000007711150564745010022000 0ustar00a

��Re��
@s\dZddlmZddlZddlmZdZzddlmZ	Wne
yNdZ	Yn0dd�Zd	d
�Zzeeje	e�Wn.e
efy�e�d�eje	e�e�Yn0zvddlmZes�e
d
��zddlZWne
y�dZYn0eedd��sddlmZe��ddlmZee�Wne
�y,Yn0ddlmZe�de�ddlmZmZmZmZddlm Z m!Z!m"Z"m#Z#ddlm$Z$m%Z%ddl&m'Z'ddl&m(Z(ddl)m*Z*m+Z+m,Z,ddl-m.Z.m/Z/m0Z0m1Z1m2Z2m3Z3m4Z4m5Z5ddl6m7Z7m8Z8ddl9m:Z:ddlm;Z;m<Z<m=Z=m>Z>m?Z?m@Z@mAZAmBZBmCZCddlDZDddlDmEZEeD�FeG��HeE��ejdeAdd �dS)!a�
Requests HTTP Library
~~~~~~~~~~~~~~~~~~~~~

Requests is an HTTP library, written in Python, for human beings.
Basic GET usage:

   >>> import requests
   >>> r = requests.get('https://www.python.org')
   >>> r.status_code
   200
   >>> b'Python is a programming language' in r.content
   True

... or POST:

   >>> payload = dict(key1='value1', key2='value2')
   >>> r = requests.post('https://httpbin.org/post', data=payload)
   >>> print(r.text)
   {
     ...
     "form": {
       "key1": "value1",
       "key2": "value2"
     },
     ...
   }

The other HTTP methods are supported - see `requests.api`. Full documentation
is at <https://requests.readthedocs.io>.

:copyright: (c) 2017 by Kenneth Reitz.
:license: Apache 2.0, see LICENSE for more details.
�)�urllib3N�)�RequestsDependencyWarning)�__version__cCs>|�d�}|dgksJ�t|�dkr.|�d�|\}}}t|�t|�t|�}}}|dks`J�|dkslJ�|dksxJ�|r�|�d�dd�\}}}t|�t|�t|�}}}d	|||fkr�d
ks�nJ�nh|�r2|�d�dd�\}}}t|�t|�t|�}}}d|||fk�r*dk�s:nJ�ntd
��dS)N�.�dev��0r���)rrr)�rr)rrr)rrrz7You need either charset_normalizer or chardet installed)�split�len�append�int�	Exception)�urllib3_version�chardet_version�charset_normalizer_version�major�minor�patch�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/__init__.py�check_compatibility6s$


"&rcCsTzttt|�d���}Wnty,YdS0|gd�krPd�|�}t�|t�dS)Nr)rr�z4Old version of cryptography ({}) may cause slowdown.)	�list�maprr�
ValueError�format�warnings�warnr)�cryptography_version�warningrrr�_check_cryptographyTs
r%zWurllib3 ({}) or chardet ({})/charset_normalizer ({}) doesn't match a supported version!)�WINDOWSz3pip internals: don't import cryptography on Windows�HAS_SNIF)�	pyopenssl)�DependencyWarning�ignore)�	__title__�__description__�__url__r)�	__build__�
__author__�__author_email__�__license__)�
__copyright__�__cake__)�utils)�packages)�Request�Response�PreparedRequest)�request�get�head�postr�put�delete�options)�session�Session)�codes)	�RequestException�Timeout�URLRequired�TooManyRedirects�	HTTPError�ConnectionError�FileModeWarning�ConnectTimeout�ReadTimeout)�NullHandler�defaultT)r)I�__doc__�pip._vendorrr!�
exceptionsrr�pip._vendor.chardetrr�ImportErrorrr%�AssertionErrorrr"r �pip._internal.utils.compatr&�ssl�getattr�pip._vendor.urllib3.contribr(�inject_into_urllib3Zcryptographyr#Zpip._vendor.urllib3.exceptionsr)�simplefilterr+r,r-r.r/r0r1r2r3�r4r5�modelsr6r7r8�apir9r:r;r<rr=r>r?�sessionsr@rA�status_codesrBrCrDrErFrGrHrIrJrK�loggingrL�	getLogger�__name__�
addHandlerrrrr�<module>sb#
��

(,site-packages/pip/_vendor/requests/__pycache__/structures.cpython-39.pyc000064400000010623150564745010022460 0ustar00a

��Re��@sDdZddlmZddlmZmZGdd�de�ZGdd�de�Zd	S)
zO
requests.structures
~~~~~~~~~~~~~~~~~~~

Data structures that power Requests.
�)�OrderedDict�)�Mapping�MutableMappingc@sbeZdZdZddd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
dS)�CaseInsensitiveDicta�A case-insensitive ``dict``-like object.

    Implements all methods and operations of
    ``MutableMapping`` as well as dict's ``copy``. Also
    provides ``lower_items``.

    All keys are expected to be strings. The structure remembers the
    case of the last key to be set, and ``iter(instance)``,
    ``keys()``, ``items()``, ``iterkeys()``, and ``iteritems()``
    will contain case-sensitive keys. However, querying and contains
    testing is case insensitive::

        cid = CaseInsensitiveDict()
        cid['Accept'] = 'application/json'
        cid['aCCEPT'] == 'application/json'  # True
        list(cid) == ['Accept']  # True

    For example, ``headers['content-encoding']`` will return the
    value of a ``'Content-Encoding'`` response header, regardless
    of how the header name was originally stored.

    If the constructor, ``.update``, or equality comparison
    operations are given keys that have equal ``.lower()``s, the
    behavior is undefined.
    NcKs*t�|_|duri}|j|fi|��dS�N)r�_store�update)�self�data�kwargs�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/requests/structures.py�__init__*szCaseInsensitiveDict.__init__cCs||f|j|��<dSr�r�lower)r
�key�valuer
r
r�__setitem__0szCaseInsensitiveDict.__setitem__cCs|j|��dS)Nrr�r
rr
r
r�__getitem__5szCaseInsensitiveDict.__getitem__cCs|j|��=dSrrrr
r
r�__delitem__8szCaseInsensitiveDict.__delitem__cCsdd�|j��D�S)Ncss|]\}}|VqdSrr
)�.0�casedkey�mappedvaluer
r
r�	<genexpr><�z/CaseInsensitiveDict.__iter__.<locals>.<genexpr>)r�values�r
r
r
r�__iter__;szCaseInsensitiveDict.__iter__cCs
t|j�Sr)�lenrrr
r
r�__len__>szCaseInsensitiveDict.__len__cCsdd�|j��D�S)z.Like iteritems(), but with all lowercase keys.css|]\}}||dfVqdS)rNr
)r�lowerkey�keyvalr
r
rrCs�z2CaseInsensitiveDict.lower_items.<locals>.<genexpr>)r�itemsrr
r
r�lower_itemsAs�zCaseInsensitiveDict.lower_itemscCs0t|t�rt|�}ntSt|���t|���kSr)�
isinstancerr�NotImplemented�dictr%)r
�otherr
r
r�__eq__Is

zCaseInsensitiveDict.__eq__cCst|j���Sr)rrrrr
r
r�copyRszCaseInsensitiveDict.copycCstt|����Sr)�strr(r$rr
r
r�__repr__UszCaseInsensitiveDict.__repr__)N)�__name__�
__module__�__qualname__�__doc__rrrrrr!r%r*r+r-r
r
r
rrs
	rcs<eZdZdZd�fdd�	Zdd�Zdd�Zdd	d
�Z�ZS)
�
LookupDictzDictionary lookup object.Ncs||_tt|���dSr)�name�superr2r)r
r3��	__class__r
rr\szLookupDict.__init__cCs
d|jS)Nz
<lookup '%s'>)r3rr
r
rr-`szLookupDict.__repr__cCs|j�|d�Sr��__dict__�getrr
r
rrcszLookupDict.__getitem__cCs|j�||�Srr7)r
r�defaultr
r
rr9hszLookupDict.get)N)N)	r.r/r0r1rr-rr9�
__classcell__r
r
r5rr2Ys
r2N)	r1�collectionsr�compatrrrr(r2r
r
r
r�<module>sJsite-packages/pip/_vendor/requests/certs.py000064400000000721150564745010015064 0ustar00#!/usr/bin/env python
# -*- coding: utf-8 -*-

"""
requests.certs
~~~~~~~~~~~~~~

This module returns the preferred default CA certificate bundle. There is
only one — the one from the certifi package.

If you are packaging Requests, e.g., for a Linux distribution or a managed
environment, you can change the definition of where() to return a separately
packaged CA bundle.
"""
from pip._vendor.certifi import where

if __name__ == '__main__':
    print(where())
site-packages/pip/_vendor/requests/structures.py000064400000005675150564745010016204 0ustar00# -*- coding: utf-8 -*-

"""
requests.structures
~~~~~~~~~~~~~~~~~~~

Data structures that power Requests.
"""

from collections import OrderedDict

from .compat import Mapping, MutableMapping


class CaseInsensitiveDict(MutableMapping):
    """A case-insensitive ``dict``-like object.

    Implements all methods and operations of
    ``MutableMapping`` as well as dict's ``copy``. Also
    provides ``lower_items``.

    All keys are expected to be strings. The structure remembers the
    case of the last key to be set, and ``iter(instance)``,
    ``keys()``, ``items()``, ``iterkeys()``, and ``iteritems()``
    will contain case-sensitive keys. However, querying and contains
    testing is case insensitive::

        cid = CaseInsensitiveDict()
        cid['Accept'] = 'application/json'
        cid['aCCEPT'] == 'application/json'  # True
        list(cid) == ['Accept']  # True

    For example, ``headers['content-encoding']`` will return the
    value of a ``'Content-Encoding'`` response header, regardless
    of how the header name was originally stored.

    If the constructor, ``.update``, or equality comparison
    operations are given keys that have equal ``.lower()``s, the
    behavior is undefined.
    """

    def __init__(self, data=None, **kwargs):
        self._store = OrderedDict()
        if data is None:
            data = {}
        self.update(data, **kwargs)

    def __setitem__(self, key, value):
        # Use the lowercased key for lookups, but store the actual
        # key alongside the value.
        self._store[key.lower()] = (key, value)

    def __getitem__(self, key):
        return self._store[key.lower()][1]

    def __delitem__(self, key):
        del self._store[key.lower()]

    def __iter__(self):
        return (casedkey for casedkey, mappedvalue in self._store.values())

    def __len__(self):
        return len(self._store)

    def lower_items(self):
        """Like iteritems(), but with all lowercase keys."""
        return (
            (lowerkey, keyval[1])
            for (lowerkey, keyval)
            in self._store.items()
        )

    def __eq__(self, other):
        if isinstance(other, Mapping):
            other = CaseInsensitiveDict(other)
        else:
            return NotImplemented
        # Compare insensitively
        return dict(self.lower_items()) == dict(other.lower_items())

    # Copy is required
    def copy(self):
        return CaseInsensitiveDict(self._store.values())

    def __repr__(self):
        return str(dict(self.items()))


class LookupDict(dict):
    """Dictionary lookup object."""

    def __init__(self, name=None):
        self.name = name
        super(LookupDict, self).__init__()

    def __repr__(self):
        return '<lookup \'%s\'>' % (self.name)

    def __getitem__(self, key):
        # We allow fall-through here, so values default to None

        return self.__dict__.get(key, None)

    def get(self, key, default=None):
        return self.__dict__.get(key, default)
site-packages/pip/_vendor/requests/cookies.py000064400000043776150564745010015421 0ustar00# -*- coding: utf-8 -*-

"""
requests.cookies
~~~~~~~~~~~~~~~~

Compatibility code to be able to use `cookielib.CookieJar` with requests.

requests.utils imports from here, so be careful with imports.
"""

import copy
import time
import calendar

from ._internal_utils import to_native_string
from .compat import cookielib, urlparse, urlunparse, Morsel, MutableMapping

try:
    import threading
except ImportError:
    import dummy_threading as threading


class MockRequest(object):
    """Wraps a `requests.Request` to mimic a `urllib2.Request`.

    The code in `cookielib.CookieJar` expects this interface in order to correctly
    manage cookie policies, i.e., determine whether a cookie can be set, given the
    domains of the request and the cookie.

    The original request object is read-only. The client is responsible for collecting
    the new headers via `get_new_headers()` and interpreting them appropriately. You
    probably want `get_cookie_header`, defined below.
    """

    def __init__(self, request):
        self._r = request
        self._new_headers = {}
        self.type = urlparse(self._r.url).scheme

    def get_type(self):
        return self.type

    def get_host(self):
        return urlparse(self._r.url).netloc

    def get_origin_req_host(self):
        return self.get_host()

    def get_full_url(self):
        # Only return the response's URL if the user hadn't set the Host
        # header
        if not self._r.headers.get('Host'):
            return self._r.url
        # If they did set it, retrieve it and reconstruct the expected domain
        host = to_native_string(self._r.headers['Host'], encoding='utf-8')
        parsed = urlparse(self._r.url)
        # Reconstruct the URL as we expect it
        return urlunparse([
            parsed.scheme, host, parsed.path, parsed.params, parsed.query,
            parsed.fragment
        ])

    def is_unverifiable(self):
        return True

    def has_header(self, name):
        return name in self._r.headers or name in self._new_headers

    def get_header(self, name, default=None):
        return self._r.headers.get(name, self._new_headers.get(name, default))

    def add_header(self, key, val):
        """cookielib has no legitimate use for this method; add it back if you find one."""
        raise NotImplementedError("Cookie headers should be added with add_unredirected_header()")

    def add_unredirected_header(self, name, value):
        self._new_headers[name] = value

    def get_new_headers(self):
        return self._new_headers

    @property
    def unverifiable(self):
        return self.is_unverifiable()

    @property
    def origin_req_host(self):
        return self.get_origin_req_host()

    @property
    def host(self):
        return self.get_host()


class MockResponse(object):
    """Wraps a `httplib.HTTPMessage` to mimic a `urllib.addinfourl`.

    ...what? Basically, expose the parsed HTTP headers from the server response
    the way `cookielib` expects to see them.
    """

    def __init__(self, headers):
        """Make a MockResponse for `cookielib` to read.

        :param headers: a httplib.HTTPMessage or analogous carrying the headers
        """
        self._headers = headers

    def info(self):
        return self._headers

    def getheaders(self, name):
        self._headers.getheaders(name)


def extract_cookies_to_jar(jar, request, response):
    """Extract the cookies from the response into a CookieJar.

    :param jar: cookielib.CookieJar (not necessarily a RequestsCookieJar)
    :param request: our own requests.Request object
    :param response: urllib3.HTTPResponse object
    """
    if not (hasattr(response, '_original_response') and
            response._original_response):
        return
    # the _original_response field is the wrapped httplib.HTTPResponse object,
    req = MockRequest(request)
    # pull out the HTTPMessage with the headers and put it in the mock:
    res = MockResponse(response._original_response.msg)
    jar.extract_cookies(res, req)


def get_cookie_header(jar, request):
    """
    Produce an appropriate Cookie header string to be sent with `request`, or None.

    :rtype: str
    """
    r = MockRequest(request)
    jar.add_cookie_header(r)
    return r.get_new_headers().get('Cookie')


def remove_cookie_by_name(cookiejar, name, domain=None, path=None):
    """Unsets a cookie by name, by default over all domains and paths.

    Wraps CookieJar.clear(), is O(n).
    """
    clearables = []
    for cookie in cookiejar:
        if cookie.name != name:
            continue
        if domain is not None and domain != cookie.domain:
            continue
        if path is not None and path != cookie.path:
            continue
        clearables.append((cookie.domain, cookie.path, cookie.name))

    for domain, path, name in clearables:
        cookiejar.clear(domain, path, name)


class CookieConflictError(RuntimeError):
    """There are two cookies that meet the criteria specified in the cookie jar.
    Use .get and .set and include domain and path args in order to be more specific.
    """


class RequestsCookieJar(cookielib.CookieJar, MutableMapping):
    """Compatibility class; is a cookielib.CookieJar, but exposes a dict
    interface.

    This is the CookieJar we create by default for requests and sessions that
    don't specify one, since some clients may expect response.cookies and
    session.cookies to support dict operations.

    Requests does not use the dict interface internally; it's just for
    compatibility with external client code. All requests code should work
    out of the box with externally provided instances of ``CookieJar``, e.g.
    ``LWPCookieJar`` and ``FileCookieJar``.

    Unlike a regular CookieJar, this class is pickleable.

    .. warning:: dictionary operations that are normally O(1) may be O(n).
    """

    def get(self, name, default=None, domain=None, path=None):
        """Dict-like get() that also supports optional domain and path args in
        order to resolve naming collisions from using one cookie jar over
        multiple domains.

        .. warning:: operation is O(n), not O(1).
        """
        try:
            return self._find_no_duplicates(name, domain, path)
        except KeyError:
            return default

    def set(self, name, value, **kwargs):
        """Dict-like set() that also supports optional domain and path args in
        order to resolve naming collisions from using one cookie jar over
        multiple domains.
        """
        # support client code that unsets cookies by assignment of a None value:
        if value is None:
            remove_cookie_by_name(self, name, domain=kwargs.get('domain'), path=kwargs.get('path'))
            return

        if isinstance(value, Morsel):
            c = morsel_to_cookie(value)
        else:
            c = create_cookie(name, value, **kwargs)
        self.set_cookie(c)
        return c

    def iterkeys(self):
        """Dict-like iterkeys() that returns an iterator of names of cookies
        from the jar.

        .. seealso:: itervalues() and iteritems().
        """
        for cookie in iter(self):
            yield cookie.name

    def keys(self):
        """Dict-like keys() that returns a list of names of cookies from the
        jar.

        .. seealso:: values() and items().
        """
        return list(self.iterkeys())

    def itervalues(self):
        """Dict-like itervalues() that returns an iterator of values of cookies
        from the jar.

        .. seealso:: iterkeys() and iteritems().
        """
        for cookie in iter(self):
            yield cookie.value

    def values(self):
        """Dict-like values() that returns a list of values of cookies from the
        jar.

        .. seealso:: keys() and items().
        """
        return list(self.itervalues())

    def iteritems(self):
        """Dict-like iteritems() that returns an iterator of name-value tuples
        from the jar.

        .. seealso:: iterkeys() and itervalues().
        """
        for cookie in iter(self):
            yield cookie.name, cookie.value

    def items(self):
        """Dict-like items() that returns a list of name-value tuples from the
        jar. Allows client-code to call ``dict(RequestsCookieJar)`` and get a
        vanilla python dict of key value pairs.

        .. seealso:: keys() and values().
        """
        return list(self.iteritems())

    def list_domains(self):
        """Utility method to list all the domains in the jar."""
        domains = []
        for cookie in iter(self):
            if cookie.domain not in domains:
                domains.append(cookie.domain)
        return domains

    def list_paths(self):
        """Utility method to list all the paths in the jar."""
        paths = []
        for cookie in iter(self):
            if cookie.path not in paths:
                paths.append(cookie.path)
        return paths

    def multiple_domains(self):
        """Returns True if there are multiple domains in the jar.
        Returns False otherwise.

        :rtype: bool
        """
        domains = []
        for cookie in iter(self):
            if cookie.domain is not None and cookie.domain in domains:
                return True
            domains.append(cookie.domain)
        return False  # there is only one domain in jar

    def get_dict(self, domain=None, path=None):
        """Takes as an argument an optional domain and path and returns a plain
        old Python dict of name-value pairs of cookies that meet the
        requirements.

        :rtype: dict
        """
        dictionary = {}
        for cookie in iter(self):
            if (
                (domain is None or cookie.domain == domain) and
                (path is None or cookie.path == path)
            ):
                dictionary[cookie.name] = cookie.value
        return dictionary

    def __contains__(self, name):
        try:
            return super(RequestsCookieJar, self).__contains__(name)
        except CookieConflictError:
            return True

    def __getitem__(self, name):
        """Dict-like __getitem__() for compatibility with client code. Throws
        exception if there are more than one cookie with name. In that case,
        use the more explicit get() method instead.

        .. warning:: operation is O(n), not O(1).
        """
        return self._find_no_duplicates(name)

    def __setitem__(self, name, value):
        """Dict-like __setitem__ for compatibility with client code. Throws
        exception if there is already a cookie of that name in the jar. In that
        case, use the more explicit set() method instead.
        """
        self.set(name, value)

    def __delitem__(self, name):
        """Deletes a cookie given a name. Wraps ``cookielib.CookieJar``'s
        ``remove_cookie_by_name()``.
        """
        remove_cookie_by_name(self, name)

    def set_cookie(self, cookie, *args, **kwargs):
        if hasattr(cookie.value, 'startswith') and cookie.value.startswith('"') and cookie.value.endswith('"'):
            cookie.value = cookie.value.replace('\\"', '')
        return super(RequestsCookieJar, self).set_cookie(cookie, *args, **kwargs)

    def update(self, other):
        """Updates this jar with cookies from another CookieJar or dict-like"""
        if isinstance(other, cookielib.CookieJar):
            for cookie in other:
                self.set_cookie(copy.copy(cookie))
        else:
            super(RequestsCookieJar, self).update(other)

    def _find(self, name, domain=None, path=None):
        """Requests uses this method internally to get cookie values.

        If there are conflicting cookies, _find arbitrarily chooses one.
        See _find_no_duplicates if you want an exception thrown if there are
        conflicting cookies.

        :param name: a string containing name of cookie
        :param domain: (optional) string containing domain of cookie
        :param path: (optional) string containing path of cookie
        :return: cookie.value
        """
        for cookie in iter(self):
            if cookie.name == name:
                if domain is None or cookie.domain == domain:
                    if path is None or cookie.path == path:
                        return cookie.value

        raise KeyError('name=%r, domain=%r, path=%r' % (name, domain, path))

    def _find_no_duplicates(self, name, domain=None, path=None):
        """Both ``__get_item__`` and ``get`` call this function: it's never
        used elsewhere in Requests.

        :param name: a string containing name of cookie
        :param domain: (optional) string containing domain of cookie
        :param path: (optional) string containing path of cookie
        :raises KeyError: if cookie is not found
        :raises CookieConflictError: if there are multiple cookies
            that match name and optionally domain and path
        :return: cookie.value
        """
        toReturn = None
        for cookie in iter(self):
            if cookie.name == name:
                if domain is None or cookie.domain == domain:
                    if path is None or cookie.path == path:
                        if toReturn is not None:  # if there are multiple cookies that meet passed in criteria
                            raise CookieConflictError('There are multiple cookies with name, %r' % (name))
                        toReturn = cookie.value  # we will eventually return this as long as no cookie conflict

        if toReturn:
            return toReturn
        raise KeyError('name=%r, domain=%r, path=%r' % (name, domain, path))

    def __getstate__(self):
        """Unlike a normal CookieJar, this class is pickleable."""
        state = self.__dict__.copy()
        # remove the unpickleable RLock object
        state.pop('_cookies_lock')
        return state

    def __setstate__(self, state):
        """Unlike a normal CookieJar, this class is pickleable."""
        self.__dict__.update(state)
        if '_cookies_lock' not in self.__dict__:
            self._cookies_lock = threading.RLock()

    def copy(self):
        """Return a copy of this RequestsCookieJar."""
        new_cj = RequestsCookieJar()
        new_cj.set_policy(self.get_policy())
        new_cj.update(self)
        return new_cj

    def get_policy(self):
        """Return the CookiePolicy instance used."""
        return self._policy


def _copy_cookie_jar(jar):
    if jar is None:
        return None

    if hasattr(jar, 'copy'):
        # We're dealing with an instance of RequestsCookieJar
        return jar.copy()
    # We're dealing with a generic CookieJar instance
    new_jar = copy.copy(jar)
    new_jar.clear()
    for cookie in jar:
        new_jar.set_cookie(copy.copy(cookie))
    return new_jar


def create_cookie(name, value, **kwargs):
    """Make a cookie from underspecified parameters.

    By default, the pair of `name` and `value` will be set for the domain ''
    and sent on every request (this is sometimes called a "supercookie").
    """
    result = {
        'version': 0,
        'name': name,
        'value': value,
        'port': None,
        'domain': '',
        'path': '/',
        'secure': False,
        'expires': None,
        'discard': True,
        'comment': None,
        'comment_url': None,
        'rest': {'HttpOnly': None},
        'rfc2109': False,
    }

    badargs = set(kwargs) - set(result)
    if badargs:
        err = 'create_cookie() got unexpected keyword arguments: %s'
        raise TypeError(err % list(badargs))

    result.update(kwargs)
    result['port_specified'] = bool(result['port'])
    result['domain_specified'] = bool(result['domain'])
    result['domain_initial_dot'] = result['domain'].startswith('.')
    result['path_specified'] = bool(result['path'])

    return cookielib.Cookie(**result)


def morsel_to_cookie(morsel):
    """Convert a Morsel object into a Cookie containing the one k/v pair."""

    expires = None
    if morsel['max-age']:
        try:
            expires = int(time.time() + int(morsel['max-age']))
        except ValueError:
            raise TypeError('max-age: %s must be integer' % morsel['max-age'])
    elif morsel['expires']:
        time_template = '%a, %d-%b-%Y %H:%M:%S GMT'
        expires = calendar.timegm(
            time.strptime(morsel['expires'], time_template)
        )
    return create_cookie(
        comment=morsel['comment'],
        comment_url=bool(morsel['comment']),
        discard=False,
        domain=morsel['domain'],
        expires=expires,
        name=morsel.key,
        path=morsel['path'],
        port=None,
        rest={'HttpOnly': morsel['httponly']},
        rfc2109=False,
        secure=bool(morsel['secure']),
        value=morsel.value,
        version=morsel['version'] or 0,
    )


def cookiejar_from_dict(cookie_dict, cookiejar=None, overwrite=True):
    """Returns a CookieJar from a key/value dictionary.

    :param cookie_dict: Dict of key/values to insert into CookieJar.
    :param cookiejar: (optional) A cookiejar to add the cookies to.
    :param overwrite: (optional) If False, will not replace cookies
        already in the jar with new ones.
    :rtype: CookieJar
    """
    if cookiejar is None:
        cookiejar = RequestsCookieJar()

    if cookie_dict is not None:
        names_from_jar = [cookie.name for cookie in cookiejar]
        for name in cookie_dict:
            if overwrite or (name not in names_from_jar):
                cookiejar.set_cookie(create_cookie(name, cookie_dict[name]))

    return cookiejar


def merge_cookies(cookiejar, cookies):
    """Add cookies to cookiejar and returns a merged CookieJar.

    :param cookiejar: CookieJar object to add the cookies to.
    :param cookies: Dictionary or CookieJar object to be added.
    :rtype: CookieJar
    """
    if not isinstance(cookiejar, cookielib.CookieJar):
        raise ValueError('You can only merge into CookieJar')

    if isinstance(cookies, dict):
        cookiejar = cookiejar_from_dict(
            cookies, cookiejar=cookiejar, overwrite=False)
    elif isinstance(cookies, cookielib.CookieJar):
        try:
            cookiejar.update(cookies)
        except AttributeError:
            for cookie_in_jar in cookies:
                cookiejar.set_cookie(cookie_in_jar)

    return cookiejar
site-packages/pip/_vendor/requests/exceptions.py000064400000006262150564745010016133 0ustar00# -*- coding: utf-8 -*-

"""
requests.exceptions
~~~~~~~~~~~~~~~~~~~

This module contains the set of Requests' exceptions.
"""
from pip._vendor.urllib3.exceptions import HTTPError as BaseHTTPError


class RequestException(IOError):
    """There was an ambiguous exception that occurred while handling your
    request.
    """

    def __init__(self, *args, **kwargs):
        """Initialize RequestException with `request` and `response` objects."""
        response = kwargs.pop('response', None)
        self.response = response
        self.request = kwargs.pop('request', None)
        if (response is not None and not self.request and
                hasattr(response, 'request')):
            self.request = self.response.request
        super(RequestException, self).__init__(*args, **kwargs)


class InvalidJSONError(RequestException):
    """A JSON error occurred."""


class HTTPError(RequestException):
    """An HTTP error occurred."""


class ConnectionError(RequestException):
    """A Connection error occurred."""


class ProxyError(ConnectionError):
    """A proxy error occurred."""


class SSLError(ConnectionError):
    """An SSL error occurred."""


class Timeout(RequestException):
    """The request timed out.

    Catching this error will catch both
    :exc:`~requests.exceptions.ConnectTimeout` and
    :exc:`~requests.exceptions.ReadTimeout` errors.
    """


class ConnectTimeout(ConnectionError, Timeout):
    """The request timed out while trying to connect to the remote server.

    Requests that produced this error are safe to retry.
    """


class ReadTimeout(Timeout):
    """The server did not send any data in the allotted amount of time."""


class URLRequired(RequestException):
    """A valid URL is required to make a request."""


class TooManyRedirects(RequestException):
    """Too many redirects."""


class MissingSchema(RequestException, ValueError):
    """The URL schema (e.g. http or https) is missing."""


class InvalidSchema(RequestException, ValueError):
    """See defaults.py for valid schemas."""


class InvalidURL(RequestException, ValueError):
    """The URL provided was somehow invalid."""


class InvalidHeader(RequestException, ValueError):
    """The header value provided was somehow invalid."""


class InvalidProxyURL(InvalidURL):
    """The proxy URL provided is invalid."""


class ChunkedEncodingError(RequestException):
    """The server declared chunked encoding but sent an invalid chunk."""


class ContentDecodingError(RequestException, BaseHTTPError):
    """Failed to decode response content."""


class StreamConsumedError(RequestException, TypeError):
    """The content for this response was already consumed."""


class RetryError(RequestException):
    """Custom retries logic failed"""


class UnrewindableBodyError(RequestException):
    """Requests encountered an error when trying to rewind a body."""

# Warnings


class RequestsWarning(Warning):
    """Base warning for Requests."""


class FileModeWarning(RequestsWarning, DeprecationWarning):
    """A file was opened in text mode, but Requests determined its binary length."""


class RequestsDependencyWarning(RequestsWarning):
    """An imported dependency doesn't match the expected version range."""
site-packages/pip/_vendor/requests/help.py000064400000007604150564745010014703 0ustar00"""Module containing bug report helper(s)."""
from __future__ import print_function

import json
import platform
import sys
import ssl

from pip._vendor import idna
from pip._vendor import urllib3

from . import __version__ as requests_version

charset_normalizer = None

try:
    from pip._vendor import chardet
except ImportError:
    chardet = None

try:
    from pip._vendor.urllib3.contrib import pyopenssl
except ImportError:
    pyopenssl = None
    OpenSSL = None
    cryptography = None
else:
    import OpenSSL
    import cryptography


def _implementation():
    """Return a dict with the Python implementation and version.

    Provide both the name and the version of the Python implementation
    currently running. For example, on CPython 2.7.5 it will return
    {'name': 'CPython', 'version': '2.7.5'}.

    This function works best on CPython and PyPy: in particular, it probably
    doesn't work for Jython or IronPython. Future investigation should be done
    to work out the correct shape of the code for those platforms.
    """
    implementation = platform.python_implementation()

    if implementation == 'CPython':
        implementation_version = platform.python_version()
    elif implementation == 'PyPy':
        implementation_version = '%s.%s.%s' % (sys.pypy_version_info.major,
                                               sys.pypy_version_info.minor,
                                               sys.pypy_version_info.micro)
        if sys.pypy_version_info.releaselevel != 'final':
            implementation_version = ''.join([
                implementation_version, sys.pypy_version_info.releaselevel
            ])
    elif implementation == 'Jython':
        implementation_version = platform.python_version()  # Complete Guess
    elif implementation == 'IronPython':
        implementation_version = platform.python_version()  # Complete Guess
    else:
        implementation_version = 'Unknown'

    return {'name': implementation, 'version': implementation_version}


def info():
    """Generate information for a bug report."""
    try:
        platform_info = {
            'system': platform.system(),
            'release': platform.release(),
        }
    except IOError:
        platform_info = {
            'system': 'Unknown',
            'release': 'Unknown',
        }

    implementation_info = _implementation()
    urllib3_info = {'version': urllib3.__version__}
    charset_normalizer_info = {'version': None}
    chardet_info = {'version': None}
    if charset_normalizer:
        charset_normalizer_info = {'version': charset_normalizer.__version__}
    if chardet:
        chardet_info = {'version': chardet.__version__}

    pyopenssl_info = {
        'version': None,
        'openssl_version': '',
    }
    if OpenSSL:
        pyopenssl_info = {
            'version': OpenSSL.__version__,
            'openssl_version': '%x' % OpenSSL.SSL.OPENSSL_VERSION_NUMBER,
        }
    cryptography_info = {
        'version': getattr(cryptography, '__version__', ''),
    }
    idna_info = {
        'version': getattr(idna, '__version__', ''),
    }

    system_ssl = ssl.OPENSSL_VERSION_NUMBER
    system_ssl_info = {
        'version': '%x' % system_ssl if system_ssl is not None else ''
    }

    return {
        'platform': platform_info,
        'implementation': implementation_info,
        'system_ssl': system_ssl_info,
        'using_pyopenssl': pyopenssl is not None,
        'using_charset_normalizer': chardet is None,
        'pyOpenSSL': pyopenssl_info,
        'urllib3': urllib3_info,
        'chardet': chardet_info,
        'charset_normalizer': charset_normalizer_info,
        'cryptography': cryptography_info,
        'idna': idna_info,
        'requests': {
            'version': requests_version,
        },
    }


def main():
    """Pretty-print the bug information as JSON."""
    print(json.dumps(info(), sort_keys=True, indent=2))


if __name__ == '__main__':
    main()
site-packages/pip/_vendor/requests/adapters.py000064400000052054150564745010015555 0ustar00# -*- coding: utf-8 -*-

"""
requests.adapters
~~~~~~~~~~~~~~~~~

This module contains the transport adapters that Requests uses to define
and maintain connections.
"""

import os.path
import socket

from pip._vendor.urllib3.poolmanager import PoolManager, proxy_from_url
from pip._vendor.urllib3.response import HTTPResponse
from pip._vendor.urllib3.util import parse_url
from pip._vendor.urllib3.util import Timeout as TimeoutSauce
from pip._vendor.urllib3.util.retry import Retry
from pip._vendor.urllib3.exceptions import ClosedPoolError
from pip._vendor.urllib3.exceptions import ConnectTimeoutError
from pip._vendor.urllib3.exceptions import HTTPError as _HTTPError
from pip._vendor.urllib3.exceptions import MaxRetryError
from pip._vendor.urllib3.exceptions import NewConnectionError
from pip._vendor.urllib3.exceptions import ProxyError as _ProxyError
from pip._vendor.urllib3.exceptions import ProtocolError
from pip._vendor.urllib3.exceptions import ReadTimeoutError
from pip._vendor.urllib3.exceptions import SSLError as _SSLError
from pip._vendor.urllib3.exceptions import ResponseError
from pip._vendor.urllib3.exceptions import LocationValueError

from .models import Response
from .compat import urlparse, basestring
from .utils import (DEFAULT_CA_BUNDLE_PATH, extract_zipped_paths,
                    get_encoding_from_headers, prepend_scheme_if_needed,
                    get_auth_from_url, urldefragauth, select_proxy)
from .structures import CaseInsensitiveDict
from .cookies import extract_cookies_to_jar
from .exceptions import (ConnectionError, ConnectTimeout, ReadTimeout, SSLError,
                         ProxyError, RetryError, InvalidSchema, InvalidProxyURL,
                         InvalidURL)
from .auth import _basic_auth_str

try:
    from pip._vendor.urllib3.contrib.socks import SOCKSProxyManager
except ImportError:
    def SOCKSProxyManager(*args, **kwargs):
        raise InvalidSchema("Missing dependencies for SOCKS support.")

DEFAULT_POOLBLOCK = False
DEFAULT_POOLSIZE = 10
DEFAULT_RETRIES = 0
DEFAULT_POOL_TIMEOUT = None


class BaseAdapter(object):
    """The Base Transport Adapter"""

    def __init__(self):
        super(BaseAdapter, self).__init__()

    def send(self, request, stream=False, timeout=None, verify=True,
             cert=None, proxies=None):
        """Sends PreparedRequest object. Returns Response object.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param stream: (optional) Whether to stream the request content.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple
        :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use
        :param cert: (optional) Any user-provided SSL certificate to be trusted.
        :param proxies: (optional) The proxies dictionary to apply to the request.
        """
        raise NotImplementedError

    def close(self):
        """Cleans up adapter specific items."""
        raise NotImplementedError


class HTTPAdapter(BaseAdapter):
    """The built-in HTTP Adapter for urllib3.

    Provides a general-case interface for Requests sessions to contact HTTP and
    HTTPS urls by implementing the Transport Adapter interface. This class will
    usually be created by the :class:`Session <Session>` class under the
    covers.

    :param pool_connections: The number of urllib3 connection pools to cache.
    :param pool_maxsize: The maximum number of connections to save in the pool.
    :param max_retries: The maximum number of retries each connection
        should attempt. Note, this applies only to failed DNS lookups, socket
        connections and connection timeouts, never to requests where data has
        made it to the server. By default, Requests does not retry failed
        connections. If you need granular control over the conditions under
        which we retry a request, import urllib3's ``Retry`` class and pass
        that instead.
    :param pool_block: Whether the connection pool should block for connections.

    Usage::

      >>> import requests
      >>> s = requests.Session()
      >>> a = requests.adapters.HTTPAdapter(max_retries=3)
      >>> s.mount('http://', a)
    """
    __attrs__ = ['max_retries', 'config', '_pool_connections', '_pool_maxsize',
                 '_pool_block']

    def __init__(self, pool_connections=DEFAULT_POOLSIZE,
                 pool_maxsize=DEFAULT_POOLSIZE, max_retries=DEFAULT_RETRIES,
                 pool_block=DEFAULT_POOLBLOCK):
        if max_retries == DEFAULT_RETRIES:
            self.max_retries = Retry(0, read=False)
        else:
            self.max_retries = Retry.from_int(max_retries)
        self.config = {}
        self.proxy_manager = {}

        super(HTTPAdapter, self).__init__()

        self._pool_connections = pool_connections
        self._pool_maxsize = pool_maxsize
        self._pool_block = pool_block

        self.init_poolmanager(pool_connections, pool_maxsize, block=pool_block)

    def __getstate__(self):
        return {attr: getattr(self, attr, None) for attr in self.__attrs__}

    def __setstate__(self, state):
        # Can't handle by adding 'proxy_manager' to self.__attrs__ because
        # self.poolmanager uses a lambda function, which isn't pickleable.
        self.proxy_manager = {}
        self.config = {}

        for attr, value in state.items():
            setattr(self, attr, value)

        self.init_poolmanager(self._pool_connections, self._pool_maxsize,
                              block=self._pool_block)

    def init_poolmanager(self, connections, maxsize, block=DEFAULT_POOLBLOCK, **pool_kwargs):
        """Initializes a urllib3 PoolManager.

        This method should not be called from user code, and is only
        exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param connections: The number of urllib3 connection pools to cache.
        :param maxsize: The maximum number of connections to save in the pool.
        :param block: Block when no free connections are available.
        :param pool_kwargs: Extra keyword arguments used to initialize the Pool Manager.
        """
        # save these values for pickling
        self._pool_connections = connections
        self._pool_maxsize = maxsize
        self._pool_block = block

        self.poolmanager = PoolManager(num_pools=connections, maxsize=maxsize,
                                       block=block, strict=True, **pool_kwargs)

    def proxy_manager_for(self, proxy, **proxy_kwargs):
        """Return urllib3 ProxyManager for the given proxy.

        This method should not be called from user code, and is only
        exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param proxy: The proxy to return a urllib3 ProxyManager for.
        :param proxy_kwargs: Extra keyword arguments used to configure the Proxy Manager.
        :returns: ProxyManager
        :rtype: urllib3.ProxyManager
        """
        if proxy in self.proxy_manager:
            manager = self.proxy_manager[proxy]
        elif proxy.lower().startswith('socks'):
            username, password = get_auth_from_url(proxy)
            manager = self.proxy_manager[proxy] = SOCKSProxyManager(
                proxy,
                username=username,
                password=password,
                num_pools=self._pool_connections,
                maxsize=self._pool_maxsize,
                block=self._pool_block,
                **proxy_kwargs
            )
        else:
            proxy_headers = self.proxy_headers(proxy)
            manager = self.proxy_manager[proxy] = proxy_from_url(
                proxy,
                proxy_headers=proxy_headers,
                num_pools=self._pool_connections,
                maxsize=self._pool_maxsize,
                block=self._pool_block,
                **proxy_kwargs)

        return manager

    def cert_verify(self, conn, url, verify, cert):
        """Verify a SSL certificate. This method should not be called from user
        code, and is only exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param conn: The urllib3 connection object associated with the cert.
        :param url: The requested URL.
        :param verify: Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use
        :param cert: The SSL certificate to verify.
        """
        if url.lower().startswith('https') and verify:

            cert_loc = None

            # Allow self-specified cert location.
            if verify is not True:
                cert_loc = verify

            if not cert_loc:
                cert_loc = extract_zipped_paths(DEFAULT_CA_BUNDLE_PATH)

            if not cert_loc or not os.path.exists(cert_loc):
                raise IOError("Could not find a suitable TLS CA certificate bundle, "
                              "invalid path: {}".format(cert_loc))

            conn.cert_reqs = 'CERT_REQUIRED'

            if not os.path.isdir(cert_loc):
                conn.ca_certs = cert_loc
            else:
                conn.ca_cert_dir = cert_loc
        else:
            conn.cert_reqs = 'CERT_NONE'
            conn.ca_certs = None
            conn.ca_cert_dir = None

        if cert:
            if not isinstance(cert, basestring):
                conn.cert_file = cert[0]
                conn.key_file = cert[1]
            else:
                conn.cert_file = cert
                conn.key_file = None
            if conn.cert_file and not os.path.exists(conn.cert_file):
                raise IOError("Could not find the TLS certificate file, "
                              "invalid path: {}".format(conn.cert_file))
            if conn.key_file and not os.path.exists(conn.key_file):
                raise IOError("Could not find the TLS key file, "
                              "invalid path: {}".format(conn.key_file))

    def build_response(self, req, resp):
        """Builds a :class:`Response <requests.Response>` object from a urllib3
        response. This should not be called from user code, and is only exposed
        for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`

        :param req: The :class:`PreparedRequest <PreparedRequest>` used to generate the response.
        :param resp: The urllib3 response object.
        :rtype: requests.Response
        """
        response = Response()

        # Fallback to None if there's no status_code, for whatever reason.
        response.status_code = getattr(resp, 'status', None)

        # Make headers case-insensitive.
        response.headers = CaseInsensitiveDict(getattr(resp, 'headers', {}))

        # Set encoding.
        response.encoding = get_encoding_from_headers(response.headers)
        response.raw = resp
        response.reason = response.raw.reason

        if isinstance(req.url, bytes):
            response.url = req.url.decode('utf-8')
        else:
            response.url = req.url

        # Add new cookies from the server.
        extract_cookies_to_jar(response.cookies, req, resp)

        # Give the Response some context.
        response.request = req
        response.connection = self

        return response

    def get_connection(self, url, proxies=None):
        """Returns a urllib3 connection for the given URL. This should not be
        called from user code, and is only exposed for use when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param url: The URL to connect to.
        :param proxies: (optional) A Requests-style dictionary of proxies used on this request.
        :rtype: urllib3.ConnectionPool
        """
        proxy = select_proxy(url, proxies)

        if proxy:
            proxy = prepend_scheme_if_needed(proxy, 'http')
            proxy_url = parse_url(proxy)
            if not proxy_url.host:
                raise InvalidProxyURL("Please check proxy URL. It is malformed"
                                      " and could be missing the host.")
            proxy_manager = self.proxy_manager_for(proxy)
            conn = proxy_manager.connection_from_url(url)
        else:
            # Only scheme should be lower case
            parsed = urlparse(url)
            url = parsed.geturl()
            conn = self.poolmanager.connection_from_url(url)

        return conn

    def close(self):
        """Disposes of any internal state.

        Currently, this closes the PoolManager and any active ProxyManager,
        which closes any pooled connections.
        """
        self.poolmanager.clear()
        for proxy in self.proxy_manager.values():
            proxy.clear()

    def request_url(self, request, proxies):
        """Obtain the url to use when making the final request.

        If the message is being sent through a HTTP proxy, the full URL has to
        be used. Otherwise, we should only use the path portion of the URL.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs.
        :rtype: str
        """
        proxy = select_proxy(request.url, proxies)
        scheme = urlparse(request.url).scheme

        is_proxied_http_request = (proxy and scheme != 'https')
        using_socks_proxy = False
        if proxy:
            proxy_scheme = urlparse(proxy).scheme.lower()
            using_socks_proxy = proxy_scheme.startswith('socks')

        url = request.path_url
        if is_proxied_http_request and not using_socks_proxy:
            url = urldefragauth(request.url)

        return url

    def add_headers(self, request, **kwargs):
        """Add any headers needed by the connection. As of v2.0 this does
        nothing by default, but is left for overriding by users that subclass
        the :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param request: The :class:`PreparedRequest <PreparedRequest>` to add headers to.
        :param kwargs: The keyword arguments from the call to send().
        """
        pass

    def proxy_headers(self, proxy):
        """Returns a dictionary of the headers to add to any request sent
        through a proxy. This works with urllib3 magic to ensure that they are
        correctly sent to the proxy, rather than in a tunnelled request if
        CONNECT is being used.

        This should not be called from user code, and is only exposed for use
        when subclassing the
        :class:`HTTPAdapter <requests.adapters.HTTPAdapter>`.

        :param proxy: The url of the proxy being used for this request.
        :rtype: dict
        """
        headers = {}
        username, password = get_auth_from_url(proxy)

        if username:
            headers['Proxy-Authorization'] = _basic_auth_str(username,
                                                             password)

        return headers

    def send(self, request, stream=False, timeout=None, verify=True, cert=None, proxies=None):
        """Sends PreparedRequest object. Returns Response object.

        :param request: The :class:`PreparedRequest <PreparedRequest>` being sent.
        :param stream: (optional) Whether to stream the request content.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple or urllib3 Timeout object
        :param verify: (optional) Either a boolean, in which case it controls whether
            we verify the server's TLS certificate, or a string, in which case it
            must be a path to a CA bundle to use
        :param cert: (optional) Any user-provided SSL certificate to be trusted.
        :param proxies: (optional) The proxies dictionary to apply to the request.
        :rtype: requests.Response
        """

        try:
            conn = self.get_connection(request.url, proxies)
        except LocationValueError as e:
            raise InvalidURL(e, request=request)

        self.cert_verify(conn, request.url, verify, cert)
        url = self.request_url(request, proxies)
        self.add_headers(request, stream=stream, timeout=timeout, verify=verify, cert=cert, proxies=proxies)

        chunked = not (request.body is None or 'Content-Length' in request.headers)

        if isinstance(timeout, tuple):
            try:
                connect, read = timeout
                timeout = TimeoutSauce(connect=connect, read=read)
            except ValueError as e:
                # this may raise a string formatting error.
                err = ("Invalid timeout {}. Pass a (connect, read) "
                       "timeout tuple, or a single float to set "
                       "both timeouts to the same value".format(timeout))
                raise ValueError(err)
        elif isinstance(timeout, TimeoutSauce):
            pass
        else:
            timeout = TimeoutSauce(connect=timeout, read=timeout)

        try:
            if not chunked:
                resp = conn.urlopen(
                    method=request.method,
                    url=url,
                    body=request.body,
                    headers=request.headers,
                    redirect=False,
                    assert_same_host=False,
                    preload_content=False,
                    decode_content=False,
                    retries=self.max_retries,
                    timeout=timeout
                )

            # Send the request.
            else:
                if hasattr(conn, 'proxy_pool'):
                    conn = conn.proxy_pool

                low_conn = conn._get_conn(timeout=DEFAULT_POOL_TIMEOUT)

                try:
                    low_conn.putrequest(request.method,
                                        url,
                                        skip_accept_encoding=True)

                    for header, value in request.headers.items():
                        low_conn.putheader(header, value)

                    low_conn.endheaders()

                    for i in request.body:
                        low_conn.send(hex(len(i))[2:].encode('utf-8'))
                        low_conn.send(b'\r\n')
                        low_conn.send(i)
                        low_conn.send(b'\r\n')
                    low_conn.send(b'0\r\n\r\n')

                    # Receive the response from the server
                    try:
                        # For Python 2.7, use buffering of HTTP responses
                        r = low_conn.getresponse(buffering=True)
                    except TypeError:
                        # For compatibility with Python 3.3+
                        r = low_conn.getresponse()

                    resp = HTTPResponse.from_httplib(
                        r,
                        pool=conn,
                        connection=low_conn,
                        preload_content=False,
                        decode_content=False
                    )
                except:
                    # If we hit any problems here, clean up the connection.
                    # Then, reraise so that we can handle the actual exception.
                    low_conn.close()
                    raise

        except (ProtocolError, socket.error) as err:
            raise ConnectionError(err, request=request)

        except MaxRetryError as e:
            if isinstance(e.reason, ConnectTimeoutError):
                # TODO: Remove this in 3.0.0: see #2811
                if not isinstance(e.reason, NewConnectionError):
                    raise ConnectTimeout(e, request=request)

            if isinstance(e.reason, ResponseError):
                raise RetryError(e, request=request)

            if isinstance(e.reason, _ProxyError):
                raise ProxyError(e, request=request)

            if isinstance(e.reason, _SSLError):
                # This branch is for urllib3 v1.22 and later.
                raise SSLError(e, request=request)

            raise ConnectionError(e, request=request)

        except ClosedPoolError as e:
            raise ConnectionError(e, request=request)

        except _ProxyError as e:
            raise ProxyError(e)

        except (_SSLError, _HTTPError) as e:
            if isinstance(e, _SSLError):
                # This branch is for urllib3 versions earlier than v1.22
                raise SSLError(e, request=request)
            elif isinstance(e, ReadTimeoutError):
                raise ReadTimeout(e, request=request)
            else:
                raise

        return self.build_response(request, resp)
site-packages/pip/_vendor/requests/status_codes.py000064400000010134150564745010016443 0ustar00# -*- coding: utf-8 -*-

r"""
The ``codes`` object defines a mapping from common names for HTTP statuses
to their numerical codes, accessible either as attributes or as dictionary
items.

Example::

    >>> import requests
    >>> requests.codes['temporary_redirect']
    307
    >>> requests.codes.teapot
    418
    >>> requests.codes['\o/']
    200

Some codes have multiple names, and both upper- and lower-case versions of
the names are allowed. For example, ``codes.ok``, ``codes.OK``, and
``codes.okay`` all correspond to the HTTP status code 200.
"""

from .structures import LookupDict

_codes = {

    # Informational.
    100: ('continue',),
    101: ('switching_protocols',),
    102: ('processing',),
    103: ('checkpoint',),
    122: ('uri_too_long', 'request_uri_too_long'),
    200: ('ok', 'okay', 'all_ok', 'all_okay', 'all_good', '\\o/', '✓'),
    201: ('created',),
    202: ('accepted',),
    203: ('non_authoritative_info', 'non_authoritative_information'),
    204: ('no_content',),
    205: ('reset_content', 'reset'),
    206: ('partial_content', 'partial'),
    207: ('multi_status', 'multiple_status', 'multi_stati', 'multiple_stati'),
    208: ('already_reported',),
    226: ('im_used',),

    # Redirection.
    300: ('multiple_choices',),
    301: ('moved_permanently', 'moved', '\\o-'),
    302: ('found',),
    303: ('see_other', 'other'),
    304: ('not_modified',),
    305: ('use_proxy',),
    306: ('switch_proxy',),
    307: ('temporary_redirect', 'temporary_moved', 'temporary'),
    308: ('permanent_redirect',
          'resume_incomplete', 'resume',),  # These 2 to be removed in 3.0

    # Client Error.
    400: ('bad_request', 'bad'),
    401: ('unauthorized',),
    402: ('payment_required', 'payment'),
    403: ('forbidden',),
    404: ('not_found', '-o-'),
    405: ('method_not_allowed', 'not_allowed'),
    406: ('not_acceptable',),
    407: ('proxy_authentication_required', 'proxy_auth', 'proxy_authentication'),
    408: ('request_timeout', 'timeout'),
    409: ('conflict',),
    410: ('gone',),
    411: ('length_required',),
    412: ('precondition_failed', 'precondition'),
    413: ('request_entity_too_large',),
    414: ('request_uri_too_large',),
    415: ('unsupported_media_type', 'unsupported_media', 'media_type'),
    416: ('requested_range_not_satisfiable', 'requested_range', 'range_not_satisfiable'),
    417: ('expectation_failed',),
    418: ('im_a_teapot', 'teapot', 'i_am_a_teapot'),
    421: ('misdirected_request',),
    422: ('unprocessable_entity', 'unprocessable'),
    423: ('locked',),
    424: ('failed_dependency', 'dependency'),
    425: ('unordered_collection', 'unordered'),
    426: ('upgrade_required', 'upgrade'),
    428: ('precondition_required', 'precondition'),
    429: ('too_many_requests', 'too_many'),
    431: ('header_fields_too_large', 'fields_too_large'),
    444: ('no_response', 'none'),
    449: ('retry_with', 'retry'),
    450: ('blocked_by_windows_parental_controls', 'parental_controls'),
    451: ('unavailable_for_legal_reasons', 'legal_reasons'),
    499: ('client_closed_request',),

    # Server Error.
    500: ('internal_server_error', 'server_error', '/o\\', '✗'),
    501: ('not_implemented',),
    502: ('bad_gateway',),
    503: ('service_unavailable', 'unavailable'),
    504: ('gateway_timeout',),
    505: ('http_version_not_supported', 'http_version'),
    506: ('variant_also_negotiates',),
    507: ('insufficient_storage',),
    509: ('bandwidth_limit_exceeded', 'bandwidth'),
    510: ('not_extended',),
    511: ('network_authentication_required', 'network_auth', 'network_authentication'),
}

codes = LookupDict(name='status_codes')

def _init():
    for code, titles in _codes.items():
        for title in titles:
            setattr(codes, title, code)
            if not title.startswith(('\\', '/')):
                setattr(codes, title.upper(), code)

    def doc(code):
        names = ', '.join('``%s``' % n for n in _codes[code])
        return '* %d: %s' % (code, names)

    global __doc__
    __doc__ = (__doc__ + '\n' +
               '\n'.join(doc(code) for code in sorted(_codes))
               if __doc__ is not None else None)

_init()
site-packages/pip/_vendor/requests/utils.py000064400000075242150564745010015116 0ustar00# -*- coding: utf-8 -*-

"""
requests.utils
~~~~~~~~~~~~~~

This module provides utility functions that are used within Requests
that are also useful for external consumption.
"""

import codecs
import contextlib
import io
import os
import re
import socket
import struct
import sys
import tempfile
import warnings
import zipfile
from collections import OrderedDict
from pip._vendor.urllib3.util import make_headers

from .__version__ import __version__
from . import certs
# to_native_string is unused here, but imported here for backwards compatibility
from ._internal_utils import to_native_string
from .compat import parse_http_list as _parse_list_header
from .compat import (
    quote, urlparse, bytes, str, unquote, getproxies,
    proxy_bypass, urlunparse, basestring, integer_types, is_py3,
    proxy_bypass_environment, getproxies_environment, Mapping)
from .cookies import cookiejar_from_dict
from .structures import CaseInsensitiveDict
from .exceptions import (
    InvalidURL, InvalidHeader, FileModeWarning, UnrewindableBodyError)

NETRC_FILES = ('.netrc', '_netrc')

DEFAULT_CA_BUNDLE_PATH = certs.where()

DEFAULT_PORTS = {'http': 80, 'https': 443}

# Ensure that ', ' is used to preserve previous delimiter behavior.
DEFAULT_ACCEPT_ENCODING = ", ".join(
    re.split(r",\s*", make_headers(accept_encoding=True)["accept-encoding"])
)


if sys.platform == 'win32':
    # provide a proxy_bypass version on Windows without DNS lookups

    def proxy_bypass_registry(host):
        try:
            if is_py3:
                import winreg
            else:
                import _winreg as winreg
        except ImportError:
            return False

        try:
            internetSettings = winreg.OpenKey(winreg.HKEY_CURRENT_USER,
                r'Software\Microsoft\Windows\CurrentVersion\Internet Settings')
            # ProxyEnable could be REG_SZ or REG_DWORD, normalizing it
            proxyEnable = int(winreg.QueryValueEx(internetSettings,
                                              'ProxyEnable')[0])
            # ProxyOverride is almost always a string
            proxyOverride = winreg.QueryValueEx(internetSettings,
                                                'ProxyOverride')[0]
        except OSError:
            return False
        if not proxyEnable or not proxyOverride:
            return False

        # make a check value list from the registry entry: replace the
        # '<local>' string by the localhost entry and the corresponding
        # canonical entry.
        proxyOverride = proxyOverride.split(';')
        # now check if we match one of the registry values.
        for test in proxyOverride:
            if test == '<local>':
                if '.' not in host:
                    return True
            test = test.replace(".", r"\.")     # mask dots
            test = test.replace("*", r".*")     # change glob sequence
            test = test.replace("?", r".")      # change glob char
            if re.match(test, host, re.I):
                return True
        return False

    def proxy_bypass(host):  # noqa
        """Return True, if the host should be bypassed.

        Checks proxy settings gathered from the environment, if specified,
        or the registry.
        """
        if getproxies_environment():
            return proxy_bypass_environment(host)
        else:
            return proxy_bypass_registry(host)


def dict_to_sequence(d):
    """Returns an internal sequence dictionary update."""

    if hasattr(d, 'items'):
        d = d.items()

    return d


def super_len(o):
    total_length = None
    current_position = 0

    if hasattr(o, '__len__'):
        total_length = len(o)

    elif hasattr(o, 'len'):
        total_length = o.len

    elif hasattr(o, 'fileno'):
        try:
            fileno = o.fileno()
        except io.UnsupportedOperation:
            pass
        else:
            total_length = os.fstat(fileno).st_size

            # Having used fstat to determine the file length, we need to
            # confirm that this file was opened up in binary mode.
            if 'b' not in o.mode:
                warnings.warn((
                    "Requests has determined the content-length for this "
                    "request using the binary size of the file: however, the "
                    "file has been opened in text mode (i.e. without the 'b' "
                    "flag in the mode). This may lead to an incorrect "
                    "content-length. In Requests 3.0, support will be removed "
                    "for files in text mode."),
                    FileModeWarning
                )

    if hasattr(o, 'tell'):
        try:
            current_position = o.tell()
        except (OSError, IOError):
            # This can happen in some weird situations, such as when the file
            # is actually a special file descriptor like stdin. In this
            # instance, we don't know what the length is, so set it to zero and
            # let requests chunk it instead.
            if total_length is not None:
                current_position = total_length
        else:
            if hasattr(o, 'seek') and total_length is None:
                # StringIO and BytesIO have seek but no useable fileno
                try:
                    # seek to end of file
                    o.seek(0, 2)
                    total_length = o.tell()

                    # seek back to current position to support
                    # partially read file-like objects
                    o.seek(current_position or 0)
                except (OSError, IOError):
                    total_length = 0

    if total_length is None:
        total_length = 0

    return max(0, total_length - current_position)


def get_netrc_auth(url, raise_errors=False):
    """Returns the Requests tuple auth for a given url from netrc."""

    netrc_file = os.environ.get('NETRC')
    if netrc_file is not None:
        netrc_locations = (netrc_file,)
    else:
        netrc_locations = ('~/{}'.format(f) for f in NETRC_FILES)

    try:
        from netrc import netrc, NetrcParseError

        netrc_path = None

        for f in netrc_locations:
            try:
                loc = os.path.expanduser(f)
            except KeyError:
                # os.path.expanduser can fail when $HOME is undefined and
                # getpwuid fails. See https://bugs.python.org/issue20164 &
                # https://github.com/psf/requests/issues/1846
                return

            if os.path.exists(loc):
                netrc_path = loc
                break

        # Abort early if there isn't one.
        if netrc_path is None:
            return

        ri = urlparse(url)

        # Strip port numbers from netloc. This weird `if...encode`` dance is
        # used for Python 3.2, which doesn't support unicode literals.
        splitstr = b':'
        if isinstance(url, str):
            splitstr = splitstr.decode('ascii')
        host = ri.netloc.split(splitstr)[0]

        try:
            _netrc = netrc(netrc_path).authenticators(host)
            if _netrc:
                # Return with login / password
                login_i = (0 if _netrc[0] else 1)
                return (_netrc[login_i], _netrc[2])
        except (NetrcParseError, IOError):
            # If there was a parsing error or a permissions issue reading the file,
            # we'll just skip netrc auth unless explicitly asked to raise errors.
            if raise_errors:
                raise

    # App Engine hackiness.
    except (ImportError, AttributeError):
        pass


def guess_filename(obj):
    """Tries to guess the filename of the given object."""
    name = getattr(obj, 'name', None)
    if (name and isinstance(name, basestring) and name[0] != '<' and
            name[-1] != '>'):
        return os.path.basename(name)


def extract_zipped_paths(path):
    """Replace nonexistent paths that look like they refer to a member of a zip
    archive with the location of an extracted copy of the target, or else
    just return the provided path unchanged.
    """
    if os.path.exists(path):
        # this is already a valid path, no need to do anything further
        return path

    # find the first valid part of the provided path and treat that as a zip archive
    # assume the rest of the path is the name of a member in the archive
    archive, member = os.path.split(path)
    while archive and not os.path.exists(archive):
        archive, prefix = os.path.split(archive)
        member = '/'.join([prefix, member])

    if not zipfile.is_zipfile(archive):
        return path

    zip_file = zipfile.ZipFile(archive)
    if member not in zip_file.namelist():
        return path

    # we have a valid zip archive and a valid member of that archive
    tmp = tempfile.gettempdir()
    extracted_path = os.path.join(tmp, member.split('/')[-1])
    if not os.path.exists(extracted_path):
        # use read + write to avoid the creating nested folders, we only want the file, avoids mkdir racing condition
        with atomic_open(extracted_path) as file_handler:
            file_handler.write(zip_file.read(member))
    return extracted_path


@contextlib.contextmanager
def atomic_open(filename):
    """Write a file to the disk in an atomic fashion"""
    replacer = os.rename if sys.version_info[0] == 2 else os.replace
    tmp_descriptor, tmp_name = tempfile.mkstemp(dir=os.path.dirname(filename))
    try:
        with os.fdopen(tmp_descriptor, 'wb') as tmp_handler:
            yield tmp_handler
        replacer(tmp_name, filename)
    except BaseException:
        os.remove(tmp_name)
        raise


def from_key_val_list(value):
    """Take an object and test to see if it can be represented as a
    dictionary. Unless it can not be represented as such, return an
    OrderedDict, e.g.,

    ::

        >>> from_key_val_list([('key', 'val')])
        OrderedDict([('key', 'val')])
        >>> from_key_val_list('string')
        Traceback (most recent call last):
        ...
        ValueError: cannot encode objects that are not 2-tuples
        >>> from_key_val_list({'key': 'val'})
        OrderedDict([('key', 'val')])

    :rtype: OrderedDict
    """
    if value is None:
        return None

    if isinstance(value, (str, bytes, bool, int)):
        raise ValueError('cannot encode objects that are not 2-tuples')

    return OrderedDict(value)


def to_key_val_list(value):
    """Take an object and test to see if it can be represented as a
    dictionary. If it can be, return a list of tuples, e.g.,

    ::

        >>> to_key_val_list([('key', 'val')])
        [('key', 'val')]
        >>> to_key_val_list({'key': 'val'})
        [('key', 'val')]
        >>> to_key_val_list('string')
        Traceback (most recent call last):
        ...
        ValueError: cannot encode objects that are not 2-tuples

    :rtype: list
    """
    if value is None:
        return None

    if isinstance(value, (str, bytes, bool, int)):
        raise ValueError('cannot encode objects that are not 2-tuples')

    if isinstance(value, Mapping):
        value = value.items()

    return list(value)


# From mitsuhiko/werkzeug (used with permission).
def parse_list_header(value):
    """Parse lists as described by RFC 2068 Section 2.

    In particular, parse comma-separated lists where the elements of
    the list may include quoted-strings.  A quoted-string could
    contain a comma.  A non-quoted string could have quotes in the
    middle.  Quotes are removed automatically after parsing.

    It basically works like :func:`parse_set_header` just that items
    may appear multiple times and case sensitivity is preserved.

    The return value is a standard :class:`list`:

    >>> parse_list_header('token, "quoted value"')
    ['token', 'quoted value']

    To create a header from the :class:`list` again, use the
    :func:`dump_header` function.

    :param value: a string with a list header.
    :return: :class:`list`
    :rtype: list
    """
    result = []
    for item in _parse_list_header(value):
        if item[:1] == item[-1:] == '"':
            item = unquote_header_value(item[1:-1])
        result.append(item)
    return result


# From mitsuhiko/werkzeug (used with permission).
def parse_dict_header(value):
    """Parse lists of key, value pairs as described by RFC 2068 Section 2 and
    convert them into a python dict:

    >>> d = parse_dict_header('foo="is a fish", bar="as well"')
    >>> type(d) is dict
    True
    >>> sorted(d.items())
    [('bar', 'as well'), ('foo', 'is a fish')]

    If there is no value for a key it will be `None`:

    >>> parse_dict_header('key_without_value')
    {'key_without_value': None}

    To create a header from the :class:`dict` again, use the
    :func:`dump_header` function.

    :param value: a string with a dict header.
    :return: :class:`dict`
    :rtype: dict
    """
    result = {}
    for item in _parse_list_header(value):
        if '=' not in item:
            result[item] = None
            continue
        name, value = item.split('=', 1)
        if value[:1] == value[-1:] == '"':
            value = unquote_header_value(value[1:-1])
        result[name] = value
    return result


# From mitsuhiko/werkzeug (used with permission).
def unquote_header_value(value, is_filename=False):
    r"""Unquotes a header value.  (Reversal of :func:`quote_header_value`).
    This does not use the real unquoting but what browsers are actually
    using for quoting.

    :param value: the header value to unquote.
    :rtype: str
    """
    if value and value[0] == value[-1] == '"':
        # this is not the real unquoting, but fixing this so that the
        # RFC is met will result in bugs with internet explorer and
        # probably some other browsers as well.  IE for example is
        # uploading files with "C:\foo\bar.txt" as filename
        value = value[1:-1]

        # if this is a filename and the starting characters look like
        # a UNC path, then just return the value without quotes.  Using the
        # replace sequence below on a UNC path has the effect of turning
        # the leading double slash into a single slash and then
        # _fix_ie_filename() doesn't work correctly.  See #458.
        if not is_filename or value[:2] != '\\\\':
            return value.replace('\\\\', '\\').replace('\\"', '"')
    return value


def dict_from_cookiejar(cj):
    """Returns a key/value dictionary from a CookieJar.

    :param cj: CookieJar object to extract cookies from.
    :rtype: dict
    """

    cookie_dict = {}

    for cookie in cj:
        cookie_dict[cookie.name] = cookie.value

    return cookie_dict


def add_dict_to_cookiejar(cj, cookie_dict):
    """Returns a CookieJar from a key/value dictionary.

    :param cj: CookieJar to insert cookies into.
    :param cookie_dict: Dict of key/values to insert into CookieJar.
    :rtype: CookieJar
    """

    return cookiejar_from_dict(cookie_dict, cj)


def get_encodings_from_content(content):
    """Returns encodings from given content string.

    :param content: bytestring to extract encodings from.
    """
    warnings.warn((
        'In requests 3.0, get_encodings_from_content will be removed. For '
        'more information, please see the discussion on issue #2266. (This'
        ' warning should only appear once.)'),
        DeprecationWarning)

    charset_re = re.compile(r'<meta.*?charset=["\']*(.+?)["\'>]', flags=re.I)
    pragma_re = re.compile(r'<meta.*?content=["\']*;?charset=(.+?)["\'>]', flags=re.I)
    xml_re = re.compile(r'^<\?xml.*?encoding=["\']*(.+?)["\'>]')

    return (charset_re.findall(content) +
            pragma_re.findall(content) +
            xml_re.findall(content))


def _parse_content_type_header(header):
    """Returns content type and parameters from given header

    :param header: string
    :return: tuple containing content type and dictionary of
         parameters
    """

    tokens = header.split(';')
    content_type, params = tokens[0].strip(), tokens[1:]
    params_dict = {}
    items_to_strip = "\"' "

    for param in params:
        param = param.strip()
        if param:
            key, value = param, True
            index_of_equals = param.find("=")
            if index_of_equals != -1:
                key = param[:index_of_equals].strip(items_to_strip)
                value = param[index_of_equals + 1:].strip(items_to_strip)
            params_dict[key.lower()] = value
    return content_type, params_dict


def get_encoding_from_headers(headers):
    """Returns encodings from given HTTP Header Dict.

    :param headers: dictionary to extract encoding from.
    :rtype: str
    """

    content_type = headers.get('content-type')

    if not content_type:
        return None

    content_type, params = _parse_content_type_header(content_type)

    if 'charset' in params:
        return params['charset'].strip("'\"")

    if 'text' in content_type:
        return 'ISO-8859-1'

    if 'application/json' in content_type:
        # Assume UTF-8 based on RFC 4627: https://www.ietf.org/rfc/rfc4627.txt since the charset was unset
        return 'utf-8'


def stream_decode_response_unicode(iterator, r):
    """Stream decodes a iterator."""

    if r.encoding is None:
        for item in iterator:
            yield item
        return

    decoder = codecs.getincrementaldecoder(r.encoding)(errors='replace')
    for chunk in iterator:
        rv = decoder.decode(chunk)
        if rv:
            yield rv
    rv = decoder.decode(b'', final=True)
    if rv:
        yield rv


def iter_slices(string, slice_length):
    """Iterate over slices of a string."""
    pos = 0
    if slice_length is None or slice_length <= 0:
        slice_length = len(string)
    while pos < len(string):
        yield string[pos:pos + slice_length]
        pos += slice_length


def get_unicode_from_response(r):
    """Returns the requested content back in unicode.

    :param r: Response object to get unicode content from.

    Tried:

    1. charset from content-type
    2. fall back and replace all unicode characters

    :rtype: str
    """
    warnings.warn((
        'In requests 3.0, get_unicode_from_response will be removed. For '
        'more information, please see the discussion on issue #2266. (This'
        ' warning should only appear once.)'),
        DeprecationWarning)

    tried_encodings = []

    # Try charset from content-type
    encoding = get_encoding_from_headers(r.headers)

    if encoding:
        try:
            return str(r.content, encoding)
        except UnicodeError:
            tried_encodings.append(encoding)

    # Fall back:
    try:
        return str(r.content, encoding, errors='replace')
    except TypeError:
        return r.content


# The unreserved URI characters (RFC 3986)
UNRESERVED_SET = frozenset(
    "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz" + "0123456789-._~")


def unquote_unreserved(uri):
    """Un-escape any percent-escape sequences in a URI that are unreserved
    characters. This leaves all reserved, illegal and non-ASCII bytes encoded.

    :rtype: str
    """
    parts = uri.split('%')
    for i in range(1, len(parts)):
        h = parts[i][0:2]
        if len(h) == 2 and h.isalnum():
            try:
                c = chr(int(h, 16))
            except ValueError:
                raise InvalidURL("Invalid percent-escape sequence: '%s'" % h)

            if c in UNRESERVED_SET:
                parts[i] = c + parts[i][2:]
            else:
                parts[i] = '%' + parts[i]
        else:
            parts[i] = '%' + parts[i]
    return ''.join(parts)


def requote_uri(uri):
    """Re-quote the given URI.

    This function passes the given URI through an unquote/quote cycle to
    ensure that it is fully and consistently quoted.

    :rtype: str
    """
    safe_with_percent = "!#$%&'()*+,/:;=?@[]~"
    safe_without_percent = "!#$&'()*+,/:;=?@[]~"
    try:
        # Unquote only the unreserved characters
        # Then quote only illegal characters (do not quote reserved,
        # unreserved, or '%')
        return quote(unquote_unreserved(uri), safe=safe_with_percent)
    except InvalidURL:
        # We couldn't unquote the given URI, so let's try quoting it, but
        # there may be unquoted '%'s in the URI. We need to make sure they're
        # properly quoted so they do not cause issues elsewhere.
        return quote(uri, safe=safe_without_percent)


def address_in_network(ip, net):
    """This function allows you to check if an IP belongs to a network subnet

    Example: returns True if ip = 192.168.1.1 and net = 192.168.1.0/24
             returns False if ip = 192.168.1.1 and net = 192.168.100.0/24

    :rtype: bool
    """
    ipaddr = struct.unpack('=L', socket.inet_aton(ip))[0]
    netaddr, bits = net.split('/')
    netmask = struct.unpack('=L', socket.inet_aton(dotted_netmask(int(bits))))[0]
    network = struct.unpack('=L', socket.inet_aton(netaddr))[0] & netmask
    return (ipaddr & netmask) == (network & netmask)


def dotted_netmask(mask):
    """Converts mask from /xx format to xxx.xxx.xxx.xxx

    Example: if mask is 24 function returns 255.255.255.0

    :rtype: str
    """
    bits = 0xffffffff ^ (1 << 32 - mask) - 1
    return socket.inet_ntoa(struct.pack('>I', bits))


def is_ipv4_address(string_ip):
    """
    :rtype: bool
    """
    try:
        socket.inet_aton(string_ip)
    except socket.error:
        return False
    return True


def is_valid_cidr(string_network):
    """
    Very simple check of the cidr format in no_proxy variable.

    :rtype: bool
    """
    if string_network.count('/') == 1:
        try:
            mask = int(string_network.split('/')[1])
        except ValueError:
            return False

        if mask < 1 or mask > 32:
            return False

        try:
            socket.inet_aton(string_network.split('/')[0])
        except socket.error:
            return False
    else:
        return False
    return True


@contextlib.contextmanager
def set_environ(env_name, value):
    """Set the environment variable 'env_name' to 'value'

    Save previous value, yield, and then restore the previous value stored in
    the environment variable 'env_name'.

    If 'value' is None, do nothing"""
    value_changed = value is not None
    if value_changed:
        old_value = os.environ.get(env_name)
        os.environ[env_name] = value
    try:
        yield
    finally:
        if value_changed:
            if old_value is None:
                del os.environ[env_name]
            else:
                os.environ[env_name] = old_value


def should_bypass_proxies(url, no_proxy):
    """
    Returns whether we should bypass proxies or not.

    :rtype: bool
    """
    # Prioritize lowercase environment variables over uppercase
    # to keep a consistent behaviour with other http projects (curl, wget).
    get_proxy = lambda k: os.environ.get(k) or os.environ.get(k.upper())

    # First check whether no_proxy is defined. If it is, check that the URL
    # we're getting isn't in the no_proxy list.
    no_proxy_arg = no_proxy
    if no_proxy is None:
        no_proxy = get_proxy('no_proxy')
    parsed = urlparse(url)

    if parsed.hostname is None:
        # URLs don't always have hostnames, e.g. file:/// urls.
        return True

    if no_proxy:
        # We need to check whether we match here. We need to see if we match
        # the end of the hostname, both with and without the port.
        no_proxy = (
            host for host in no_proxy.replace(' ', '').split(',') if host
        )

        if is_ipv4_address(parsed.hostname):
            for proxy_ip in no_proxy:
                if is_valid_cidr(proxy_ip):
                    if address_in_network(parsed.hostname, proxy_ip):
                        return True
                elif parsed.hostname == proxy_ip:
                    # If no_proxy ip was defined in plain IP notation instead of cidr notation &
                    # matches the IP of the index
                    return True
        else:
            host_with_port = parsed.hostname
            if parsed.port:
                host_with_port += ':{}'.format(parsed.port)

            for host in no_proxy:
                if parsed.hostname.endswith(host) or host_with_port.endswith(host):
                    # The URL does match something in no_proxy, so we don't want
                    # to apply the proxies on this URL.
                    return True

    with set_environ('no_proxy', no_proxy_arg):
        # parsed.hostname can be `None` in cases such as a file URI.
        try:
            bypass = proxy_bypass(parsed.hostname)
        except (TypeError, socket.gaierror):
            bypass = False

    if bypass:
        return True

    return False


def get_environ_proxies(url, no_proxy=None):
    """
    Return a dict of environment proxies.

    :rtype: dict
    """
    if should_bypass_proxies(url, no_proxy=no_proxy):
        return {}
    else:
        return getproxies()


def select_proxy(url, proxies):
    """Select a proxy for the url, if applicable.

    :param url: The url being for the request
    :param proxies: A dictionary of schemes or schemes and hosts to proxy URLs
    """
    proxies = proxies or {}
    urlparts = urlparse(url)
    if urlparts.hostname is None:
        return proxies.get(urlparts.scheme, proxies.get('all'))

    proxy_keys = [
        urlparts.scheme + '://' + urlparts.hostname,
        urlparts.scheme,
        'all://' + urlparts.hostname,
        'all',
    ]
    proxy = None
    for proxy_key in proxy_keys:
        if proxy_key in proxies:
            proxy = proxies[proxy_key]
            break

    return proxy


def default_user_agent(name="python-requests"):
    """
    Return a string representing the default user agent.

    :rtype: str
    """
    return '%s/%s' % (name, __version__)


def default_headers():
    """
    :rtype: requests.structures.CaseInsensitiveDict
    """
    return CaseInsensitiveDict({
        'User-Agent': default_user_agent(),
        'Accept-Encoding': DEFAULT_ACCEPT_ENCODING,
        'Accept': '*/*',
        'Connection': 'keep-alive',
    })


def parse_header_links(value):
    """Return a list of parsed link headers proxies.

    i.e. Link: <http:/.../front.jpeg>; rel=front; type="image/jpeg",<http://.../back.jpeg>; rel=back;type="image/jpeg"

    :rtype: list
    """

    links = []

    replace_chars = ' \'"'

    value = value.strip(replace_chars)
    if not value:
        return links

    for val in re.split(', *<', value):
        try:
            url, params = val.split(';', 1)
        except ValueError:
            url, params = val, ''

        link = {'url': url.strip('<> \'"')}

        for param in params.split(';'):
            try:
                key, value = param.split('=')
            except ValueError:
                break

            link[key.strip(replace_chars)] = value.strip(replace_chars)

        links.append(link)

    return links


# Null bytes; no need to recreate these on each call to guess_json_utf
_null = '\x00'.encode('ascii')  # encoding to ASCII for Python 3
_null2 = _null * 2
_null3 = _null * 3


def guess_json_utf(data):
    """
    :rtype: str
    """
    # JSON always starts with two ASCII characters, so detection is as
    # easy as counting the nulls and from their location and count
    # determine the encoding. Also detect a BOM, if present.
    sample = data[:4]
    if sample in (codecs.BOM_UTF32_LE, codecs.BOM_UTF32_BE):
        return 'utf-32'     # BOM included
    if sample[:3] == codecs.BOM_UTF8:
        return 'utf-8-sig'  # BOM included, MS style (discouraged)
    if sample[:2] in (codecs.BOM_UTF16_LE, codecs.BOM_UTF16_BE):
        return 'utf-16'     # BOM included
    nullcount = sample.count(_null)
    if nullcount == 0:
        return 'utf-8'
    if nullcount == 2:
        if sample[::2] == _null2:   # 1st and 3rd are null
            return 'utf-16-be'
        if sample[1::2] == _null2:  # 2nd and 4th are null
            return 'utf-16-le'
        # Did not detect 2 valid UTF-16 ascii-range characters
    if nullcount == 3:
        if sample[:3] == _null3:
            return 'utf-32-be'
        if sample[1:] == _null3:
            return 'utf-32-le'
        # Did not detect a valid UTF-32 ascii-range character
    return None


def prepend_scheme_if_needed(url, new_scheme):
    """Given a URL that may or may not have a scheme, prepend the given scheme.
    Does not replace a present scheme with the one provided as an argument.

    :rtype: str
    """
    scheme, netloc, path, params, query, fragment = urlparse(url, new_scheme)

    # urlparse is a finicky beast, and sometimes decides that there isn't a
    # netloc present. Assume that it's being over-cautious, and switch netloc
    # and path if urlparse decided there was no netloc.
    if not netloc:
        netloc, path = path, netloc

    return urlunparse((scheme, netloc, path, params, query, fragment))


def get_auth_from_url(url):
    """Given a url with authentication components, extract them into a tuple of
    username,password.

    :rtype: (str,str)
    """
    parsed = urlparse(url)

    try:
        auth = (unquote(parsed.username), unquote(parsed.password))
    except (AttributeError, TypeError):
        auth = ('', '')

    return auth


# Moved outside of function to avoid recompile every call
_CLEAN_HEADER_REGEX_BYTE = re.compile(b'^\\S[^\\r\\n]*$|^$')
_CLEAN_HEADER_REGEX_STR = re.compile(r'^\S[^\r\n]*$|^$')


def check_header_validity(header):
    """Verifies that header value is a string which doesn't contain
    leading whitespace or return characters. This prevents unintended
    header injection.

    :param header: tuple, in the format (name, value).
    """
    name, value = header

    if isinstance(value, bytes):
        pat = _CLEAN_HEADER_REGEX_BYTE
    else:
        pat = _CLEAN_HEADER_REGEX_STR
    try:
        if not pat.match(value):
            raise InvalidHeader("Invalid return character or leading space in header: %s" % name)
    except TypeError:
        raise InvalidHeader("Value for header {%s: %s} must be of type str or "
                            "bytes, not %s" % (name, value, type(value)))


def urldefragauth(url):
    """
    Given a url remove the fragment and the authentication part.

    :rtype: str
    """
    scheme, netloc, path, params, query, fragment = urlparse(url)

    # see func:`prepend_scheme_if_needed`
    if not netloc:
        netloc, path = path, netloc

    netloc = netloc.rsplit('@', 1)[-1]

    return urlunparse((scheme, netloc, path, params, query, ''))


def rewind_body(prepared_request):
    """Move file pointer back to its recorded starting position
    so it can be read again on redirect.
    """
    body_seek = getattr(prepared_request.body, 'seek', None)
    if body_seek is not None and isinstance(prepared_request._body_position, integer_types):
        try:
            body_seek(prepared_request._body_position)
        except (IOError, OSError):
            raise UnrewindableBodyError("An error occurred when rewinding request "
                                        "body for redirect.")
    else:
        raise UnrewindableBodyError("Unable to rewind request body for redirect.")
site-packages/pip/_vendor/requests/compat.py000064400000003775150564745010015243 0ustar00# -*- coding: utf-8 -*-

"""
requests.compat
~~~~~~~~~~~~~~~

This module handles import compatibility issues between Python 2 and
Python 3.
"""

from pip._vendor import chardet

import sys

# -------
# Pythons
# -------

# Syntax sugar.
_ver = sys.version_info

#: Python 2.x?
is_py2 = (_ver[0] == 2)

#: Python 3.x?
is_py3 = (_ver[0] == 3)

# Note: We've patched out simplejson support in pip because it prevents
#       upgrading simplejson on Windows.
# try:
#     import simplejson as json
# except (ImportError, SyntaxError):
#     # simplejson does not support Python 3.2, it throws a SyntaxError
#     # because of u'...' Unicode literals.
import json

# ---------
# Specifics
# ---------

if is_py2:
    from urllib import (
        quote, unquote, quote_plus, unquote_plus, urlencode, getproxies,
        proxy_bypass, proxy_bypass_environment, getproxies_environment)
    from urlparse import urlparse, urlunparse, urljoin, urlsplit, urldefrag
    from urllib2 import parse_http_list
    import cookielib
    from Cookie import Morsel
    from StringIO import StringIO
    # Keep OrderedDict for backwards compatibility.
    from collections import Callable, Mapping, MutableMapping, OrderedDict


    builtin_str = str
    bytes = str
    str = unicode
    basestring = basestring
    numeric_types = (int, long, float)
    integer_types = (int, long)

elif is_py3:
    from urllib.parse import urlparse, urlunparse, urljoin, urlsplit, urlencode, quote, unquote, quote_plus, unquote_plus, urldefrag
    from urllib.request import parse_http_list, getproxies, proxy_bypass, proxy_bypass_environment, getproxies_environment
    from http import cookiejar as cookielib
    from http.cookies import Morsel
    from io import StringIO
    # Keep OrderedDict for backwards compatibility.
    from collections import OrderedDict
    from collections.abc import Callable, Mapping, MutableMapping

    builtin_str = str
    str = str
    bytes = bytes
    basestring = (str, bytes)
    numeric_types = (int, float)
    integer_types = (int,)
site-packages/pip/_vendor/requests/sessions.py000064400000072730150564745010015623 0ustar00# -*- coding: utf-8 -*-

"""
requests.sessions
~~~~~~~~~~~~~~~~~

This module provides a Session object to manage and persist settings across
requests (cookies, auth, proxies).
"""
import os
import sys
import time
from datetime import timedelta
from collections import OrderedDict

from .auth import _basic_auth_str
from .compat import cookielib, is_py3, urljoin, urlparse, Mapping
from .cookies import (
    cookiejar_from_dict, extract_cookies_to_jar, RequestsCookieJar, merge_cookies)
from .models import Request, PreparedRequest, DEFAULT_REDIRECT_LIMIT
from .hooks import default_hooks, dispatch_hook
from ._internal_utils import to_native_string
from .utils import to_key_val_list, default_headers, DEFAULT_PORTS
from .exceptions import (
    TooManyRedirects, InvalidSchema, ChunkedEncodingError, ContentDecodingError)

from .structures import CaseInsensitiveDict
from .adapters import HTTPAdapter

from .utils import (
    requote_uri, get_environ_proxies, get_netrc_auth, should_bypass_proxies,
    get_auth_from_url, rewind_body
)

from .status_codes import codes

# formerly defined here, reexposed here for backward compatibility
from .models import REDIRECT_STATI

# Preferred clock, based on which one is more accurate on a given system.
if sys.platform == 'win32':
    try:  # Python 3.4+
        preferred_clock = time.perf_counter
    except AttributeError:  # Earlier than Python 3.
        preferred_clock = time.clock
else:
    preferred_clock = time.time


def merge_setting(request_setting, session_setting, dict_class=OrderedDict):
    """Determines appropriate setting for a given request, taking into account
    the explicit setting on that request, and the setting in the session. If a
    setting is a dictionary, they will be merged together using `dict_class`
    """

    if session_setting is None:
        return request_setting

    if request_setting is None:
        return session_setting

    # Bypass if not a dictionary (e.g. verify)
    if not (
            isinstance(session_setting, Mapping) and
            isinstance(request_setting, Mapping)
    ):
        return request_setting

    merged_setting = dict_class(to_key_val_list(session_setting))
    merged_setting.update(to_key_val_list(request_setting))

    # Remove keys that are set to None. Extract keys first to avoid altering
    # the dictionary during iteration.
    none_keys = [k for (k, v) in merged_setting.items() if v is None]
    for key in none_keys:
        del merged_setting[key]

    return merged_setting


def merge_hooks(request_hooks, session_hooks, dict_class=OrderedDict):
    """Properly merges both requests and session hooks.

    This is necessary because when request_hooks == {'response': []}, the
    merge breaks Session hooks entirely.
    """
    if session_hooks is None or session_hooks.get('response') == []:
        return request_hooks

    if request_hooks is None or request_hooks.get('response') == []:
        return session_hooks

    return merge_setting(request_hooks, session_hooks, dict_class)


class SessionRedirectMixin(object):

    def get_redirect_target(self, resp):
        """Receives a Response. Returns a redirect URI or ``None``"""
        # Due to the nature of how requests processes redirects this method will
        # be called at least once upon the original response and at least twice
        # on each subsequent redirect response (if any).
        # If a custom mixin is used to handle this logic, it may be advantageous
        # to cache the redirect location onto the response object as a private
        # attribute.
        if resp.is_redirect:
            location = resp.headers['location']
            # Currently the underlying http module on py3 decode headers
            # in latin1, but empirical evidence suggests that latin1 is very
            # rarely used with non-ASCII characters in HTTP headers.
            # It is more likely to get UTF8 header rather than latin1.
            # This causes incorrect handling of UTF8 encoded location headers.
            # To solve this, we re-encode the location in latin1.
            if is_py3:
                location = location.encode('latin1')
            return to_native_string(location, 'utf8')
        return None

    def should_strip_auth(self, old_url, new_url):
        """Decide whether Authorization header should be removed when redirecting"""
        old_parsed = urlparse(old_url)
        new_parsed = urlparse(new_url)
        if old_parsed.hostname != new_parsed.hostname:
            return True
        # Special case: allow http -> https redirect when using the standard
        # ports. This isn't specified by RFC 7235, but is kept to avoid
        # breaking backwards compatibility with older versions of requests
        # that allowed any redirects on the same host.
        if (old_parsed.scheme == 'http' and old_parsed.port in (80, None)
                and new_parsed.scheme == 'https' and new_parsed.port in (443, None)):
            return False

        # Handle default port usage corresponding to scheme.
        changed_port = old_parsed.port != new_parsed.port
        changed_scheme = old_parsed.scheme != new_parsed.scheme
        default_port = (DEFAULT_PORTS.get(old_parsed.scheme, None), None)
        if (not changed_scheme and old_parsed.port in default_port
                and new_parsed.port in default_port):
            return False

        # Standard case: root URI must match
        return changed_port or changed_scheme

    def resolve_redirects(self, resp, req, stream=False, timeout=None,
                          verify=True, cert=None, proxies=None, yield_requests=False, **adapter_kwargs):
        """Receives a Response. Returns a generator of Responses or Requests."""

        hist = []  # keep track of history

        url = self.get_redirect_target(resp)
        previous_fragment = urlparse(req.url).fragment
        while url:
            prepared_request = req.copy()

            # Update history and keep track of redirects.
            # resp.history must ignore the original request in this loop
            hist.append(resp)
            resp.history = hist[1:]

            try:
                resp.content  # Consume socket so it can be released
            except (ChunkedEncodingError, ContentDecodingError, RuntimeError):
                resp.raw.read(decode_content=False)

            if len(resp.history) >= self.max_redirects:
                raise TooManyRedirects('Exceeded {} redirects.'.format(self.max_redirects), response=resp)

            # Release the connection back into the pool.
            resp.close()

            # Handle redirection without scheme (see: RFC 1808 Section 4)
            if url.startswith('//'):
                parsed_rurl = urlparse(resp.url)
                url = ':'.join([to_native_string(parsed_rurl.scheme), url])

            # Normalize url case and attach previous fragment if needed (RFC 7231 7.1.2)
            parsed = urlparse(url)
            if parsed.fragment == '' and previous_fragment:
                parsed = parsed._replace(fragment=previous_fragment)
            elif parsed.fragment:
                previous_fragment = parsed.fragment
            url = parsed.geturl()

            # Facilitate relative 'location' headers, as allowed by RFC 7231.
            # (e.g. '/path/to/resource' instead of 'http://domain.tld/path/to/resource')
            # Compliant with RFC3986, we percent encode the url.
            if not parsed.netloc:
                url = urljoin(resp.url, requote_uri(url))
            else:
                url = requote_uri(url)

            prepared_request.url = to_native_string(url)

            self.rebuild_method(prepared_request, resp)

            # https://github.com/psf/requests/issues/1084
            if resp.status_code not in (codes.temporary_redirect, codes.permanent_redirect):
                # https://github.com/psf/requests/issues/3490
                purged_headers = ('Content-Length', 'Content-Type', 'Transfer-Encoding')
                for header in purged_headers:
                    prepared_request.headers.pop(header, None)
                prepared_request.body = None

            headers = prepared_request.headers
            headers.pop('Cookie', None)

            # Extract any cookies sent on the response to the cookiejar
            # in the new request. Because we've mutated our copied prepared
            # request, use the old one that we haven't yet touched.
            extract_cookies_to_jar(prepared_request._cookies, req, resp.raw)
            merge_cookies(prepared_request._cookies, self.cookies)
            prepared_request.prepare_cookies(prepared_request._cookies)

            # Rebuild auth and proxy information.
            proxies = self.rebuild_proxies(prepared_request, proxies)
            self.rebuild_auth(prepared_request, resp)

            # A failed tell() sets `_body_position` to `object()`. This non-None
            # value ensures `rewindable` will be True, allowing us to raise an
            # UnrewindableBodyError, instead of hanging the connection.
            rewindable = (
                prepared_request._body_position is not None and
                ('Content-Length' in headers or 'Transfer-Encoding' in headers)
            )

            # Attempt to rewind consumed file-like object.
            if rewindable:
                rewind_body(prepared_request)

            # Override the original request.
            req = prepared_request

            if yield_requests:
                yield req
            else:

                resp = self.send(
                    req,
                    stream=stream,
                    timeout=timeout,
                    verify=verify,
                    cert=cert,
                    proxies=proxies,
                    allow_redirects=False,
                    **adapter_kwargs
                )

                extract_cookies_to_jar(self.cookies, prepared_request, resp.raw)

                # extract redirect url, if any, for the next loop
                url = self.get_redirect_target(resp)
                yield resp

    def rebuild_auth(self, prepared_request, response):
        """When being redirected we may want to strip authentication from the
        request to avoid leaking credentials. This method intelligently removes
        and reapplies authentication where possible to avoid credential loss.
        """
        headers = prepared_request.headers
        url = prepared_request.url

        if 'Authorization' in headers and self.should_strip_auth(response.request.url, url):
            # If we get redirected to a new host, we should strip out any
            # authentication headers.
            del headers['Authorization']

        # .netrc might have more auth for us on our new host.
        new_auth = get_netrc_auth(url) if self.trust_env else None
        if new_auth is not None:
            prepared_request.prepare_auth(new_auth)


    def rebuild_proxies(self, prepared_request, proxies):
        """This method re-evaluates the proxy configuration by considering the
        environment variables. If we are redirected to a URL covered by
        NO_PROXY, we strip the proxy configuration. Otherwise, we set missing
        proxy keys for this URL (in case they were stripped by a previous
        redirect).

        This method also replaces the Proxy-Authorization header where
        necessary.

        :rtype: dict
        """
        proxies = proxies if proxies is not None else {}
        headers = prepared_request.headers
        url = prepared_request.url
        scheme = urlparse(url).scheme
        new_proxies = proxies.copy()
        no_proxy = proxies.get('no_proxy')

        bypass_proxy = should_bypass_proxies(url, no_proxy=no_proxy)
        if self.trust_env and not bypass_proxy:
            environ_proxies = get_environ_proxies(url, no_proxy=no_proxy)

            proxy = environ_proxies.get(scheme, environ_proxies.get('all'))

            if proxy:
                new_proxies.setdefault(scheme, proxy)

        if 'Proxy-Authorization' in headers:
            del headers['Proxy-Authorization']

        try:
            username, password = get_auth_from_url(new_proxies[scheme])
        except KeyError:
            username, password = None, None

        if username and password:
            headers['Proxy-Authorization'] = _basic_auth_str(username, password)

        return new_proxies

    def rebuild_method(self, prepared_request, response):
        """When being redirected we may want to change the method of the request
        based on certain specs or browser behavior.
        """
        method = prepared_request.method

        # https://tools.ietf.org/html/rfc7231#section-6.4.4
        if response.status_code == codes.see_other and method != 'HEAD':
            method = 'GET'

        # Do what the browsers do, despite standards...
        # First, turn 302s into GETs.
        if response.status_code == codes.found and method != 'HEAD':
            method = 'GET'

        # Second, if a POST is responded to with a 301, turn it into a GET.
        # This bizarre behaviour is explained in Issue 1704.
        if response.status_code == codes.moved and method == 'POST':
            method = 'GET'

        prepared_request.method = method


class Session(SessionRedirectMixin):
    """A Requests session.

    Provides cookie persistence, connection-pooling, and configuration.

    Basic Usage::

      >>> import requests
      >>> s = requests.Session()
      >>> s.get('https://httpbin.org/get')
      <Response [200]>

    Or as a context manager::

      >>> with requests.Session() as s:
      ...     s.get('https://httpbin.org/get')
      <Response [200]>
    """

    __attrs__ = [
        'headers', 'cookies', 'auth', 'proxies', 'hooks', 'params', 'verify',
        'cert', 'adapters', 'stream', 'trust_env',
        'max_redirects',
    ]

    def __init__(self):

        #: A case-insensitive dictionary of headers to be sent on each
        #: :class:`Request <Request>` sent from this
        #: :class:`Session <Session>`.
        self.headers = default_headers()

        #: Default Authentication tuple or object to attach to
        #: :class:`Request <Request>`.
        self.auth = None

        #: Dictionary mapping protocol or protocol and host to the URL of the proxy
        #: (e.g. {'http': 'foo.bar:3128', 'http://host.name': 'foo.bar:4012'}) to
        #: be used on each :class:`Request <Request>`.
        self.proxies = {}

        #: Event-handling hooks.
        self.hooks = default_hooks()

        #: Dictionary of querystring data to attach to each
        #: :class:`Request <Request>`. The dictionary values may be lists for
        #: representing multivalued query parameters.
        self.params = {}

        #: Stream response content default.
        self.stream = False

        #: SSL Verification default.
        #: Defaults to `True`, requiring requests to verify the TLS certificate at the
        #: remote end.
        #: If verify is set to `False`, requests will accept any TLS certificate
        #: presented by the server, and will ignore hostname mismatches and/or
        #: expired certificates, which will make your application vulnerable to
        #: man-in-the-middle (MitM) attacks.
        #: Only set this to `False` for testing.
        self.verify = True

        #: SSL client certificate default, if String, path to ssl client
        #: cert file (.pem). If Tuple, ('cert', 'key') pair.
        self.cert = None

        #: Maximum number of redirects allowed. If the request exceeds this
        #: limit, a :class:`TooManyRedirects` exception is raised.
        #: This defaults to requests.models.DEFAULT_REDIRECT_LIMIT, which is
        #: 30.
        self.max_redirects = DEFAULT_REDIRECT_LIMIT

        #: Trust environment settings for proxy configuration, default
        #: authentication and similar.
        self.trust_env = True

        #: A CookieJar containing all currently outstanding cookies set on this
        #: session. By default it is a
        #: :class:`RequestsCookieJar <requests.cookies.RequestsCookieJar>`, but
        #: may be any other ``cookielib.CookieJar`` compatible object.
        self.cookies = cookiejar_from_dict({})

        # Default connection adapters.
        self.adapters = OrderedDict()
        self.mount('https://', HTTPAdapter())
        self.mount('http://', HTTPAdapter())

    def __enter__(self):
        return self

    def __exit__(self, *args):
        self.close()

    def prepare_request(self, request):
        """Constructs a :class:`PreparedRequest <PreparedRequest>` for
        transmission and returns it. The :class:`PreparedRequest` has settings
        merged from the :class:`Request <Request>` instance and those of the
        :class:`Session`.

        :param request: :class:`Request` instance to prepare with this
            session's settings.
        :rtype: requests.PreparedRequest
        """
        cookies = request.cookies or {}

        # Bootstrap CookieJar.
        if not isinstance(cookies, cookielib.CookieJar):
            cookies = cookiejar_from_dict(cookies)

        # Merge with session cookies
        merged_cookies = merge_cookies(
            merge_cookies(RequestsCookieJar(), self.cookies), cookies)

        # Set environment's basic authentication if not explicitly set.
        auth = request.auth
        if self.trust_env and not auth and not self.auth:
            auth = get_netrc_auth(request.url)

        p = PreparedRequest()
        p.prepare(
            method=request.method.upper(),
            url=request.url,
            files=request.files,
            data=request.data,
            json=request.json,
            headers=merge_setting(request.headers, self.headers, dict_class=CaseInsensitiveDict),
            params=merge_setting(request.params, self.params),
            auth=merge_setting(auth, self.auth),
            cookies=merged_cookies,
            hooks=merge_hooks(request.hooks, self.hooks),
        )
        return p

    def request(self, method, url,
            params=None, data=None, headers=None, cookies=None, files=None,
            auth=None, timeout=None, allow_redirects=True, proxies=None,
            hooks=None, stream=None, verify=None, cert=None, json=None):
        """Constructs a :class:`Request <Request>`, prepares it and sends it.
        Returns :class:`Response <Response>` object.

        :param method: method for the new :class:`Request` object.
        :param url: URL for the new :class:`Request` object.
        :param params: (optional) Dictionary or bytes to be sent in the query
            string for the :class:`Request`.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param json: (optional) json to send in the body of the
            :class:`Request`.
        :param headers: (optional) Dictionary of HTTP Headers to send with the
            :class:`Request`.
        :param cookies: (optional) Dict or CookieJar object to send with the
            :class:`Request`.
        :param files: (optional) Dictionary of ``'filename': file-like-objects``
            for multipart encoding upload.
        :param auth: (optional) Auth tuple or callable to enable
            Basic/Digest/Custom HTTP Auth.
        :param timeout: (optional) How long to wait for the server to send
            data before giving up, as a float, or a :ref:`(connect timeout,
            read timeout) <timeouts>` tuple.
        :type timeout: float or tuple
        :param allow_redirects: (optional) Set to True by default.
        :type allow_redirects: bool
        :param proxies: (optional) Dictionary mapping protocol or protocol and
            hostname to the URL of the proxy.
        :param stream: (optional) whether to immediately download the response
            content. Defaults to ``False``.
        :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use. Defaults to ``True``. When set to
            ``False``, requests will accept any TLS certificate presented by
            the server, and will ignore hostname mismatches and/or expired
            certificates, which will make your application vulnerable to
            man-in-the-middle (MitM) attacks. Setting verify to ``False`` 
            may be useful during local development or testing.
        :param cert: (optional) if String, path to ssl client cert file (.pem).
            If Tuple, ('cert', 'key') pair.
        :rtype: requests.Response
        """
        # Create the Request.
        req = Request(
            method=method.upper(),
            url=url,
            headers=headers,
            files=files,
            data=data or {},
            json=json,
            params=params or {},
            auth=auth,
            cookies=cookies,
            hooks=hooks,
        )
        prep = self.prepare_request(req)

        proxies = proxies or {}

        settings = self.merge_environment_settings(
            prep.url, proxies, stream, verify, cert
        )

        # Send the request.
        send_kwargs = {
            'timeout': timeout,
            'allow_redirects': allow_redirects,
        }
        send_kwargs.update(settings)
        resp = self.send(prep, **send_kwargs)

        return resp

    def get(self, url, **kwargs):
        r"""Sends a GET request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        kwargs.setdefault('allow_redirects', True)
        return self.request('GET', url, **kwargs)

    def options(self, url, **kwargs):
        r"""Sends a OPTIONS request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        kwargs.setdefault('allow_redirects', True)
        return self.request('OPTIONS', url, **kwargs)

    def head(self, url, **kwargs):
        r"""Sends a HEAD request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        kwargs.setdefault('allow_redirects', False)
        return self.request('HEAD', url, **kwargs)

    def post(self, url, data=None, json=None, **kwargs):
        r"""Sends a POST request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param json: (optional) json to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        return self.request('POST', url, data=data, json=json, **kwargs)

    def put(self, url, data=None, **kwargs):
        r"""Sends a PUT request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        return self.request('PUT', url, data=data, **kwargs)

    def patch(self, url, data=None, **kwargs):
        r"""Sends a PATCH request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param data: (optional) Dictionary, list of tuples, bytes, or file-like
            object to send in the body of the :class:`Request`.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        return self.request('PATCH', url, data=data, **kwargs)

    def delete(self, url, **kwargs):
        r"""Sends a DELETE request. Returns :class:`Response` object.

        :param url: URL for the new :class:`Request` object.
        :param \*\*kwargs: Optional arguments that ``request`` takes.
        :rtype: requests.Response
        """

        return self.request('DELETE', url, **kwargs)

    def send(self, request, **kwargs):
        """Send a given PreparedRequest.

        :rtype: requests.Response
        """
        # Set defaults that the hooks can utilize to ensure they always have
        # the correct parameters to reproduce the previous request.
        kwargs.setdefault('stream', self.stream)
        kwargs.setdefault('verify', self.verify)
        kwargs.setdefault('cert', self.cert)
        kwargs.setdefault('proxies', self.rebuild_proxies(request, self.proxies))

        # It's possible that users might accidentally send a Request object.
        # Guard against that specific failure case.
        if isinstance(request, Request):
            raise ValueError('You can only send PreparedRequests.')

        # Set up variables needed for resolve_redirects and dispatching of hooks
        allow_redirects = kwargs.pop('allow_redirects', True)
        stream = kwargs.get('stream')
        hooks = request.hooks

        # Get the appropriate adapter to use
        adapter = self.get_adapter(url=request.url)

        # Start time (approximately) of the request
        start = preferred_clock()

        # Send the request
        r = adapter.send(request, **kwargs)

        # Total elapsed time of the request (approximately)
        elapsed = preferred_clock() - start
        r.elapsed = timedelta(seconds=elapsed)

        # Response manipulation hooks
        r = dispatch_hook('response', hooks, r, **kwargs)

        # Persist cookies
        if r.history:

            # If the hooks create history then we want those cookies too
            for resp in r.history:
                extract_cookies_to_jar(self.cookies, resp.request, resp.raw)

        extract_cookies_to_jar(self.cookies, request, r.raw)

        # Resolve redirects if allowed.
        if allow_redirects:
            # Redirect resolving generator.
            gen = self.resolve_redirects(r, request, **kwargs)
            history = [resp for resp in gen]
        else:
            history = []

        # Shuffle things around if there's history.
        if history:
            # Insert the first (original) request at the start
            history.insert(0, r)
            # Get the last request made
            r = history.pop()
            r.history = history

        # If redirects aren't being followed, store the response on the Request for Response.next().
        if not allow_redirects:
            try:
                r._next = next(self.resolve_redirects(r, request, yield_requests=True, **kwargs))
            except StopIteration:
                pass

        if not stream:
            r.content

        return r

    def merge_environment_settings(self, url, proxies, stream, verify, cert):
        """
        Check the environment and merge it with some settings.

        :rtype: dict
        """
        # Gather clues from the surrounding environment.
        if self.trust_env:
            # Set environment's proxies.
            no_proxy = proxies.get('no_proxy') if proxies is not None else None
            env_proxies = get_environ_proxies(url, no_proxy=no_proxy)
            for (k, v) in env_proxies.items():
                proxies.setdefault(k, v)

            # Look for requests environment configuration and be compatible
            # with cURL.
            if verify is True or verify is None:
                verify = (os.environ.get('REQUESTS_CA_BUNDLE') or
                          os.environ.get('CURL_CA_BUNDLE'))

        # Merge all the kwargs.
        proxies = merge_setting(proxies, self.proxies)
        stream = merge_setting(stream, self.stream)
        verify = merge_setting(verify, self.verify)
        cert = merge_setting(cert, self.cert)

        return {'verify': verify, 'proxies': proxies, 'stream': stream,
                'cert': cert}

    def get_adapter(self, url):
        """
        Returns the appropriate connection adapter for the given URL.

        :rtype: requests.adapters.BaseAdapter
        """
        for (prefix, adapter) in self.adapters.items():

            if url.lower().startswith(prefix.lower()):
                return adapter

        # Nothing matches :-/
        raise InvalidSchema("No connection adapters were found for {!r}".format(url))

    def close(self):
        """Closes all adapters and as such the session"""
        for v in self.adapters.values():
            v.close()

    def mount(self, prefix, adapter):
        """Registers a connection adapter to a prefix.

        Adapters are sorted in descending order by prefix length.
        """
        self.adapters[prefix] = adapter
        keys_to_move = [k for k in self.adapters if len(k) < len(prefix)]

        for key in keys_to_move:
            self.adapters[key] = self.adapters.pop(key)

    def __getstate__(self):
        state = {attr: getattr(self, attr, None) for attr in self.__attrs__}
        return state

    def __setstate__(self, state):
        for attr, value in state.items():
            setattr(self, attr, value)


def session():
    """
    Returns a :class:`Session` for context-management.

    .. deprecated:: 1.0.0

        This method has been deprecated since version 1.0.0 and is only kept for
        backwards compatibility. New code should use :class:`~requests.sessions.Session`
        to create a session. This may be removed at a future date.

    :rtype: Session
    """
    return Session()
site-packages/pip/_vendor/requests/hooks.py000064400000001365150564745010015074 0ustar00# -*- coding: utf-8 -*-

"""
requests.hooks
~~~~~~~~~~~~~~

This module provides the capabilities for the Requests hooks system.

Available hooks:

``response``:
    The response generated from a Request.
"""
HOOKS = ['response']


def default_hooks():
    return {event: [] for event in HOOKS}

# TODO: response is the only one


def dispatch_hook(key, hooks, hook_data, **kwargs):
    """Dispatches a hook dictionary on a given piece of data."""
    hooks = hooks or {}
    hooks = hooks.get(key)
    if hooks:
        if hasattr(hooks, '__call__'):
            hooks = [hooks]
        for hook in hooks:
            _hook_data = hook(hook_data, **kwargs)
            if _hook_data is not None:
                hook_data = _hook_data
    return hook_data
site-packages/pip/_vendor/requests/api.py000064400000014402150564745010014516 0ustar00# -*- coding: utf-8 -*-

"""
requests.api
~~~~~~~~~~~~

This module implements the Requests API.

:copyright: (c) 2012 by Kenneth Reitz.
:license: Apache2, see LICENSE for more details.
"""

from . import sessions


def request(method, url, **kwargs):
    """Constructs and sends a :class:`Request <Request>`.

    :param method: method for the new :class:`Request` object: ``GET``, ``OPTIONS``, ``HEAD``, ``POST``, ``PUT``, ``PATCH``, or ``DELETE``.
    :param url: URL for the new :class:`Request` object.
    :param params: (optional) Dictionary, list of tuples or bytes to send
        in the query string for the :class:`Request`.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) A JSON serializable Python object to send in the body of the :class:`Request`.
    :param headers: (optional) Dictionary of HTTP Headers to send with the :class:`Request`.
    :param cookies: (optional) Dict or CookieJar object to send with the :class:`Request`.
    :param files: (optional) Dictionary of ``'name': file-like-objects`` (or ``{'name': file-tuple}``) for multipart encoding upload.
        ``file-tuple`` can be a 2-tuple ``('filename', fileobj)``, 3-tuple ``('filename', fileobj, 'content_type')``
        or a 4-tuple ``('filename', fileobj, 'content_type', custom_headers)``, where ``'content-type'`` is a string
        defining the content type of the given file and ``custom_headers`` a dict-like object containing additional headers
        to add for the file.
    :param auth: (optional) Auth tuple to enable Basic/Digest/Custom HTTP Auth.
    :param timeout: (optional) How many seconds to wait for the server to send data
        before giving up, as a float, or a :ref:`(connect timeout, read
        timeout) <timeouts>` tuple.
    :type timeout: float or tuple
    :param allow_redirects: (optional) Boolean. Enable/disable GET/OPTIONS/POST/PUT/PATCH/DELETE/HEAD redirection. Defaults to ``True``.
    :type allow_redirects: bool
    :param proxies: (optional) Dictionary mapping protocol to the URL of the proxy.
    :param verify: (optional) Either a boolean, in which case it controls whether we verify
            the server's TLS certificate, or a string, in which case it must be a path
            to a CA bundle to use. Defaults to ``True``.
    :param stream: (optional) if ``False``, the response content will be immediately downloaded.
    :param cert: (optional) if String, path to ssl client cert file (.pem). If Tuple, ('cert', 'key') pair.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response

    Usage::

      >>> import requests
      >>> req = requests.request('GET', 'https://httpbin.org/get')
      >>> req
      <Response [200]>
    """

    # By using the 'with' statement we are sure the session is closed, thus we
    # avoid leaving sockets open which can trigger a ResourceWarning in some
    # cases, and look like a memory leak in others.
    with sessions.Session() as session:
        return session.request(method=method, url=url, **kwargs)


def get(url, params=None, **kwargs):
    r"""Sends a GET request.

    :param url: URL for the new :class:`Request` object.
    :param params: (optional) Dictionary, list of tuples or bytes to send
        in the query string for the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('get', url, params=params, **kwargs)


def options(url, **kwargs):
    r"""Sends an OPTIONS request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('options', url, **kwargs)


def head(url, **kwargs):
    r"""Sends a HEAD request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes. If
        `allow_redirects` is not provided, it will be set to `False` (as
        opposed to the default :meth:`request` behavior).
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    kwargs.setdefault('allow_redirects', False)
    return request('head', url, **kwargs)


def post(url, data=None, json=None, **kwargs):
    r"""Sends a POST request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('post', url, data=data, json=json, **kwargs)


def put(url, data=None, **kwargs):
    r"""Sends a PUT request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('put', url, data=data, **kwargs)


def patch(url, data=None, **kwargs):
    r"""Sends a PATCH request.

    :param url: URL for the new :class:`Request` object.
    :param data: (optional) Dictionary, list of tuples, bytes, or file-like
        object to send in the body of the :class:`Request`.
    :param json: (optional) json data to send in the body of the :class:`Request`.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('patch', url, data=data, **kwargs)


def delete(url, **kwargs):
    r"""Sends a DELETE request.

    :param url: URL for the new :class:`Request` object.
    :param \*\*kwargs: Optional arguments that ``request`` takes.
    :return: :class:`Response <Response>` object
    :rtype: requests.Response
    """

    return request('delete', url, **kwargs)
site-packages/pip/_vendor/requests/__init__.py000064400000011771150564745010015512 0ustar00# -*- coding: utf-8 -*-

#   __
#  /__)  _  _     _   _ _/   _
# / (   (- (/ (/ (- _)  /  _)
#          /

"""
Requests HTTP Library
~~~~~~~~~~~~~~~~~~~~~

Requests is an HTTP library, written in Python, for human beings.
Basic GET usage:

   >>> import requests
   >>> r = requests.get('https://www.python.org')
   >>> r.status_code
   200
   >>> b'Python is a programming language' in r.content
   True

... or POST:

   >>> payload = dict(key1='value1', key2='value2')
   >>> r = requests.post('https://httpbin.org/post', data=payload)
   >>> print(r.text)
   {
     ...
     "form": {
       "key1": "value1",
       "key2": "value2"
     },
     ...
   }

The other HTTP methods are supported - see `requests.api`. Full documentation
is at <https://requests.readthedocs.io>.

:copyright: (c) 2017 by Kenneth Reitz.
:license: Apache 2.0, see LICENSE for more details.
"""

from pip._vendor import urllib3
import warnings
from .exceptions import RequestsDependencyWarning

charset_normalizer_version = None

try:
    from pip._vendor.chardet import __version__ as chardet_version
except ImportError:
    chardet_version = None

def check_compatibility(urllib3_version, chardet_version, charset_normalizer_version):
    urllib3_version = urllib3_version.split('.')
    assert urllib3_version != ['dev']  # Verify urllib3 isn't installed from git.

    # Sometimes, urllib3 only reports its version as 16.1.
    if len(urllib3_version) == 2:
        urllib3_version.append('0')

    # Check urllib3 for compatibility.
    major, minor, patch = urllib3_version  # noqa: F811
    major, minor, patch = int(major), int(minor), int(patch)
    # urllib3 >= 1.21.1, <= 1.26
    assert major == 1
    assert minor >= 21
    assert minor <= 26

    # Check charset_normalizer for compatibility.
    if chardet_version:
        major, minor, patch = chardet_version.split('.')[:3]
        major, minor, patch = int(major), int(minor), int(patch)
        # chardet_version >= 3.0.2, < 5.0.0
        assert (3, 0, 2) <= (major, minor, patch) < (5, 0, 0)
    elif charset_normalizer_version:
        major, minor, patch = charset_normalizer_version.split('.')[:3]
        major, minor, patch = int(major), int(minor), int(patch)
        # charset_normalizer >= 2.0.0 < 3.0.0
        assert (2, 0, 0) <= (major, minor, patch) < (3, 0, 0)
    else:
        raise Exception("You need either charset_normalizer or chardet installed")

def _check_cryptography(cryptography_version):
    # cryptography < 1.3.4
    try:
        cryptography_version = list(map(int, cryptography_version.split('.')))
    except ValueError:
        return

    if cryptography_version < [1, 3, 4]:
        warning = 'Old version of cryptography ({}) may cause slowdown.'.format(cryptography_version)
        warnings.warn(warning, RequestsDependencyWarning)

# Check imported dependencies for compatibility.
try:
    check_compatibility(urllib3.__version__, chardet_version, charset_normalizer_version)
except (AssertionError, ValueError):
    warnings.warn("urllib3 ({}) or chardet ({})/charset_normalizer ({}) doesn't match a supported "
                  "version!".format(urllib3.__version__, chardet_version, charset_normalizer_version),
                  RequestsDependencyWarning)

# Attempt to enable urllib3's fallback for SNI support
# if the standard library doesn't support SNI or the
# 'ssl' library isn't available.
try:
    # Note: This logic prevents upgrading cryptography on Windows, if imported
    #       as part of pip.
    from pip._internal.utils.compat import WINDOWS
    if not WINDOWS:
        raise ImportError("pip internals: don't import cryptography on Windows")
    try:
        import ssl
    except ImportError:
        ssl = None

    if not getattr(ssl, "HAS_SNI", False):
        from pip._vendor.urllib3.contrib import pyopenssl
        pyopenssl.inject_into_urllib3()

        # Check cryptography version
        from cryptography import __version__ as cryptography_version
        _check_cryptography(cryptography_version)
except ImportError:
    pass

# urllib3's DependencyWarnings should be silenced.
from pip._vendor.urllib3.exceptions import DependencyWarning
warnings.simplefilter('ignore', DependencyWarning)

from .__version__ import __title__, __description__, __url__, __version__
from .__version__ import __build__, __author__, __author_email__, __license__
from .__version__ import __copyright__, __cake__

from . import utils
from . import packages
from .models import Request, Response, PreparedRequest
from .api import request, get, head, post, patch, put, delete, options
from .sessions import session, Session
from .status_codes import codes
from .exceptions import (
    RequestException, Timeout, URLRequired,
    TooManyRedirects, HTTPError, ConnectionError,
    FileModeWarning, ConnectTimeout, ReadTimeout
)

# Set default logging handler to avoid "No handler found" warnings.
import logging
from logging import NullHandler

logging.getLogger(__name__).addHandler(NullHandler())

# FileModeWarnings go off per the default.
warnings.simplefilter('default', FileModeWarning, append=True)
site-packages/pip/_vendor/requests/auth.py000064400000023737150564745010014721 0ustar00# -*- coding: utf-8 -*-

"""
requests.auth
~~~~~~~~~~~~~

This module contains the authentication handlers for Requests.
"""

import os
import re
import time
import hashlib
import threading
import warnings

from base64 import b64encode

from .compat import urlparse, str, basestring
from .cookies import extract_cookies_to_jar
from ._internal_utils import to_native_string
from .utils import parse_dict_header

CONTENT_TYPE_FORM_URLENCODED = 'application/x-www-form-urlencoded'
CONTENT_TYPE_MULTI_PART = 'multipart/form-data'


def _basic_auth_str(username, password):
    """Returns a Basic Auth string."""

    # "I want us to put a big-ol' comment on top of it that
    # says that this behaviour is dumb but we need to preserve
    # it because people are relying on it."
    #    - Lukasa
    #
    # These are here solely to maintain backwards compatibility
    # for things like ints. This will be removed in 3.0.0.
    if not isinstance(username, basestring):
        warnings.warn(
            "Non-string usernames will no longer be supported in Requests "
            "3.0.0. Please convert the object you've passed in ({!r}) to "
            "a string or bytes object in the near future to avoid "
            "problems.".format(username),
            category=DeprecationWarning,
        )
        username = str(username)

    if not isinstance(password, basestring):
        warnings.warn(
            "Non-string passwords will no longer be supported in Requests "
            "3.0.0. Please convert the object you've passed in ({!r}) to "
            "a string or bytes object in the near future to avoid "
            "problems.".format(type(password)),
            category=DeprecationWarning,
        )
        password = str(password)
    # -- End Removal --

    if isinstance(username, str):
        username = username.encode('latin1')

    if isinstance(password, str):
        password = password.encode('latin1')

    authstr = 'Basic ' + to_native_string(
        b64encode(b':'.join((username, password))).strip()
    )

    return authstr


class AuthBase(object):
    """Base class that all auth implementations derive from"""

    def __call__(self, r):
        raise NotImplementedError('Auth hooks must be callable.')


class HTTPBasicAuth(AuthBase):
    """Attaches HTTP Basic Authentication to the given Request object."""

    def __init__(self, username, password):
        self.username = username
        self.password = password

    def __eq__(self, other):
        return all([
            self.username == getattr(other, 'username', None),
            self.password == getattr(other, 'password', None)
        ])

    def __ne__(self, other):
        return not self == other

    def __call__(self, r):
        r.headers['Authorization'] = _basic_auth_str(self.username, self.password)
        return r


class HTTPProxyAuth(HTTPBasicAuth):
    """Attaches HTTP Proxy Authentication to a given Request object."""

    def __call__(self, r):
        r.headers['Proxy-Authorization'] = _basic_auth_str(self.username, self.password)
        return r


class HTTPDigestAuth(AuthBase):
    """Attaches HTTP Digest Authentication to the given Request object."""

    def __init__(self, username, password):
        self.username = username
        self.password = password
        # Keep state in per-thread local storage
        self._thread_local = threading.local()

    def init_per_thread_state(self):
        # Ensure state is initialized just once per-thread
        if not hasattr(self._thread_local, 'init'):
            self._thread_local.init = True
            self._thread_local.last_nonce = ''
            self._thread_local.nonce_count = 0
            self._thread_local.chal = {}
            self._thread_local.pos = None
            self._thread_local.num_401_calls = None

    def build_digest_header(self, method, url):
        """
        :rtype: str
        """

        realm = self._thread_local.chal['realm']
        nonce = self._thread_local.chal['nonce']
        qop = self._thread_local.chal.get('qop')
        algorithm = self._thread_local.chal.get('algorithm')
        opaque = self._thread_local.chal.get('opaque')
        hash_utf8 = None

        if algorithm is None:
            _algorithm = 'MD5'
        else:
            _algorithm = algorithm.upper()
        # lambdas assume digest modules are imported at the top level
        if _algorithm == 'MD5' or _algorithm == 'MD5-SESS':
            def md5_utf8(x):
                if isinstance(x, str):
                    x = x.encode('utf-8')
                return hashlib.md5(x).hexdigest()
            hash_utf8 = md5_utf8
        elif _algorithm == 'SHA':
            def sha_utf8(x):
                if isinstance(x, str):
                    x = x.encode('utf-8')
                return hashlib.sha1(x).hexdigest()
            hash_utf8 = sha_utf8
        elif _algorithm == 'SHA-256':
            def sha256_utf8(x):
                if isinstance(x, str):
                    x = x.encode('utf-8')
                return hashlib.sha256(x).hexdigest()
            hash_utf8 = sha256_utf8
        elif _algorithm == 'SHA-512':
            def sha512_utf8(x):
                if isinstance(x, str):
                    x = x.encode('utf-8')
                return hashlib.sha512(x).hexdigest()
            hash_utf8 = sha512_utf8

        KD = lambda s, d: hash_utf8("%s:%s" % (s, d))

        if hash_utf8 is None:
            return None

        # XXX not implemented yet
        entdig = None
        p_parsed = urlparse(url)
        #: path is request-uri defined in RFC 2616 which should not be empty
        path = p_parsed.path or "/"
        if p_parsed.query:
            path += '?' + p_parsed.query

        A1 = '%s:%s:%s' % (self.username, realm, self.password)
        A2 = '%s:%s' % (method, path)

        HA1 = hash_utf8(A1)
        HA2 = hash_utf8(A2)

        if nonce == self._thread_local.last_nonce:
            self._thread_local.nonce_count += 1
        else:
            self._thread_local.nonce_count = 1
        ncvalue = '%08x' % self._thread_local.nonce_count
        s = str(self._thread_local.nonce_count).encode('utf-8')
        s += nonce.encode('utf-8')
        s += time.ctime().encode('utf-8')
        s += os.urandom(8)

        cnonce = (hashlib.sha1(s).hexdigest()[:16])
        if _algorithm == 'MD5-SESS':
            HA1 = hash_utf8('%s:%s:%s' % (HA1, nonce, cnonce))

        if not qop:
            respdig = KD(HA1, "%s:%s" % (nonce, HA2))
        elif qop == 'auth' or 'auth' in qop.split(','):
            noncebit = "%s:%s:%s:%s:%s" % (
                nonce, ncvalue, cnonce, 'auth', HA2
            )
            respdig = KD(HA1, noncebit)
        else:
            # XXX handle auth-int.
            return None

        self._thread_local.last_nonce = nonce

        # XXX should the partial digests be encoded too?
        base = 'username="%s", realm="%s", nonce="%s", uri="%s", ' \
               'response="%s"' % (self.username, realm, nonce, path, respdig)
        if opaque:
            base += ', opaque="%s"' % opaque
        if algorithm:
            base += ', algorithm="%s"' % algorithm
        if entdig:
            base += ', digest="%s"' % entdig
        if qop:
            base += ', qop="auth", nc=%s, cnonce="%s"' % (ncvalue, cnonce)

        return 'Digest %s' % (base)

    def handle_redirect(self, r, **kwargs):
        """Reset num_401_calls counter on redirects."""
        if r.is_redirect:
            self._thread_local.num_401_calls = 1

    def handle_401(self, r, **kwargs):
        """
        Takes the given response and tries digest-auth, if needed.

        :rtype: requests.Response
        """

        # If response is not 4xx, do not auth
        # See https://github.com/psf/requests/issues/3772
        if not 400 <= r.status_code < 500:
            self._thread_local.num_401_calls = 1
            return r

        if self._thread_local.pos is not None:
            # Rewind the file position indicator of the body to where
            # it was to resend the request.
            r.request.body.seek(self._thread_local.pos)
        s_auth = r.headers.get('www-authenticate', '')

        if 'digest' in s_auth.lower() and self._thread_local.num_401_calls < 2:

            self._thread_local.num_401_calls += 1
            pat = re.compile(r'digest ', flags=re.IGNORECASE)
            self._thread_local.chal = parse_dict_header(pat.sub('', s_auth, count=1))

            # Consume content and release the original connection
            # to allow our new request to reuse the same one.
            r.content
            r.close()
            prep = r.request.copy()
            extract_cookies_to_jar(prep._cookies, r.request, r.raw)
            prep.prepare_cookies(prep._cookies)

            prep.headers['Authorization'] = self.build_digest_header(
                prep.method, prep.url)
            _r = r.connection.send(prep, **kwargs)
            _r.history.append(r)
            _r.request = prep

            return _r

        self._thread_local.num_401_calls = 1
        return r

    def __call__(self, r):
        # Initialize per-thread state, if needed
        self.init_per_thread_state()
        # If we have a saved nonce, skip the 401
        if self._thread_local.last_nonce:
            r.headers['Authorization'] = self.build_digest_header(r.method, r.url)
        try:
            self._thread_local.pos = r.body.tell()
        except AttributeError:
            # In the case of HTTPDigestAuth being reused and the body of
            # the previous request was a file-like object, pos has the
            # file position of the previous body. Ensure it's set to
            # None.
            self._thread_local.pos = None
        r.register_hook('response', self.handle_401)
        r.register_hook('response', self.handle_redirect)
        self._thread_local.num_401_calls = 1

        return r

    def __eq__(self, other):
        return all([
            self.username == getattr(other, 'username', None),
            self.password == getattr(other, 'password', None)
        ])

    def __ne__(self, other):
        return not self == other
site-packages/pip/_vendor/requests/models.py000064400000104154150564745010015234 0ustar00# -*- coding: utf-8 -*-

"""
requests.models
~~~~~~~~~~~~~~~

This module contains the primary objects that power Requests.
"""

import datetime
import sys

# Import encoding now, to avoid implicit import later.
# Implicit import within threads may cause LookupError when standard library is in a ZIP,
# such as in Embedded Python. See https://github.com/psf/requests/issues/3578.
import encodings.idna

from pip._vendor.urllib3.fields import RequestField
from pip._vendor.urllib3.filepost import encode_multipart_formdata
from pip._vendor.urllib3.util import parse_url
from pip._vendor.urllib3.exceptions import (
    DecodeError, ReadTimeoutError, ProtocolError, LocationParseError)

from io import UnsupportedOperation
from .hooks import default_hooks
from .structures import CaseInsensitiveDict

from .auth import HTTPBasicAuth
from .cookies import cookiejar_from_dict, get_cookie_header, _copy_cookie_jar
from .exceptions import (
    HTTPError, MissingSchema, InvalidURL, ChunkedEncodingError,
    ContentDecodingError, ConnectionError, StreamConsumedError, InvalidJSONError)
from ._internal_utils import to_native_string, unicode_is_ascii
from .utils import (
    guess_filename, get_auth_from_url, requote_uri,
    stream_decode_response_unicode, to_key_val_list, parse_header_links,
    iter_slices, guess_json_utf, super_len, check_header_validity)
from .compat import (
    Callable, Mapping,
    cookielib, urlunparse, urlsplit, urlencode, str, bytes,
    is_py2, chardet, builtin_str, basestring)
from .compat import json as complexjson
from .status_codes import codes

#: The set of HTTP status codes that indicate an automatically
#: processable redirect.
REDIRECT_STATI = (
    codes.moved,               # 301
    codes.found,               # 302
    codes.other,               # 303
    codes.temporary_redirect,  # 307
    codes.permanent_redirect,  # 308
)

DEFAULT_REDIRECT_LIMIT = 30
CONTENT_CHUNK_SIZE = 10 * 1024
ITER_CHUNK_SIZE = 512


class RequestEncodingMixin(object):
    @property
    def path_url(self):
        """Build the path URL to use."""

        url = []

        p = urlsplit(self.url)

        path = p.path
        if not path:
            path = '/'

        url.append(path)

        query = p.query
        if query:
            url.append('?')
            url.append(query)

        return ''.join(url)

    @staticmethod
    def _encode_params(data):
        """Encode parameters in a piece of data.

        Will successfully encode parameters when passed as a dict or a list of
        2-tuples. Order is retained if data is a list of 2-tuples but arbitrary
        if parameters are supplied as a dict.
        """

        if isinstance(data, (str, bytes)):
            return data
        elif hasattr(data, 'read'):
            return data
        elif hasattr(data, '__iter__'):
            result = []
            for k, vs in to_key_val_list(data):
                if isinstance(vs, basestring) or not hasattr(vs, '__iter__'):
                    vs = [vs]
                for v in vs:
                    if v is not None:
                        result.append(
                            (k.encode('utf-8') if isinstance(k, str) else k,
                             v.encode('utf-8') if isinstance(v, str) else v))
            return urlencode(result, doseq=True)
        else:
            return data

    @staticmethod
    def _encode_files(files, data):
        """Build the body for a multipart/form-data request.

        Will successfully encode files when passed as a dict or a list of
        tuples. Order is retained if data is a list of tuples but arbitrary
        if parameters are supplied as a dict.
        The tuples may be 2-tuples (filename, fileobj), 3-tuples (filename, fileobj, contentype)
        or 4-tuples (filename, fileobj, contentype, custom_headers).
        """
        if (not files):
            raise ValueError("Files must be provided.")
        elif isinstance(data, basestring):
            raise ValueError("Data must not be a string.")

        new_fields = []
        fields = to_key_val_list(data or {})
        files = to_key_val_list(files or {})

        for field, val in fields:
            if isinstance(val, basestring) or not hasattr(val, '__iter__'):
                val = [val]
            for v in val:
                if v is not None:
                    # Don't call str() on bytestrings: in Py3 it all goes wrong.
                    if not isinstance(v, bytes):
                        v = str(v)

                    new_fields.append(
                        (field.decode('utf-8') if isinstance(field, bytes) else field,
                         v.encode('utf-8') if isinstance(v, str) else v))

        for (k, v) in files:
            # support for explicit filename
            ft = None
            fh = None
            if isinstance(v, (tuple, list)):
                if len(v) == 2:
                    fn, fp = v
                elif len(v) == 3:
                    fn, fp, ft = v
                else:
                    fn, fp, ft, fh = v
            else:
                fn = guess_filename(v) or k
                fp = v

            if isinstance(fp, (str, bytes, bytearray)):
                fdata = fp
            elif hasattr(fp, 'read'):
                fdata = fp.read()
            elif fp is None:
                continue
            else:
                fdata = fp

            rf = RequestField(name=k, data=fdata, filename=fn, headers=fh)
            rf.make_multipart(content_type=ft)
            new_fields.append(rf)

        body, content_type = encode_multipart_formdata(new_fields)

        return body, content_type


class RequestHooksMixin(object):
    def register_hook(self, event, hook):
        """Properly register a hook."""

        if event not in self.hooks:
            raise ValueError('Unsupported event specified, with event name "%s"' % (event))

        if isinstance(hook, Callable):
            self.hooks[event].append(hook)
        elif hasattr(hook, '__iter__'):
            self.hooks[event].extend(h for h in hook if isinstance(h, Callable))

    def deregister_hook(self, event, hook):
        """Deregister a previously registered hook.
        Returns True if the hook existed, False if not.
        """

        try:
            self.hooks[event].remove(hook)
            return True
        except ValueError:
            return False


class Request(RequestHooksMixin):
    """A user-created :class:`Request <Request>` object.

    Used to prepare a :class:`PreparedRequest <PreparedRequest>`, which is sent to the server.

    :param method: HTTP method to use.
    :param url: URL to send.
    :param headers: dictionary of headers to send.
    :param files: dictionary of {filename: fileobject} files to multipart upload.
    :param data: the body to attach to the request. If a dictionary or
        list of tuples ``[(key, value)]`` is provided, form-encoding will
        take place.
    :param json: json for the body to attach to the request (if files or data is not specified).
    :param params: URL parameters to append to the URL. If a dictionary or
        list of tuples ``[(key, value)]`` is provided, form-encoding will
        take place.
    :param auth: Auth handler or (user, pass) tuple.
    :param cookies: dictionary or CookieJar of cookies to attach to this request.
    :param hooks: dictionary of callback hooks, for internal usage.

    Usage::

      >>> import requests
      >>> req = requests.Request('GET', 'https://httpbin.org/get')
      >>> req.prepare()
      <PreparedRequest [GET]>
    """

    def __init__(self,
            method=None, url=None, headers=None, files=None, data=None,
            params=None, auth=None, cookies=None, hooks=None, json=None):

        # Default empty dicts for dict params.
        data = [] if data is None else data
        files = [] if files is None else files
        headers = {} if headers is None else headers
        params = {} if params is None else params
        hooks = {} if hooks is None else hooks

        self.hooks = default_hooks()
        for (k, v) in list(hooks.items()):
            self.register_hook(event=k, hook=v)

        self.method = method
        self.url = url
        self.headers = headers
        self.files = files
        self.data = data
        self.json = json
        self.params = params
        self.auth = auth
        self.cookies = cookies

    def __repr__(self):
        return '<Request [%s]>' % (self.method)

    def prepare(self):
        """Constructs a :class:`PreparedRequest <PreparedRequest>` for transmission and returns it."""
        p = PreparedRequest()
        p.prepare(
            method=self.method,
            url=self.url,
            headers=self.headers,
            files=self.files,
            data=self.data,
            json=self.json,
            params=self.params,
            auth=self.auth,
            cookies=self.cookies,
            hooks=self.hooks,
        )
        return p


class PreparedRequest(RequestEncodingMixin, RequestHooksMixin):
    """The fully mutable :class:`PreparedRequest <PreparedRequest>` object,
    containing the exact bytes that will be sent to the server.

    Instances are generated from a :class:`Request <Request>` object, and
    should not be instantiated manually; doing so may produce undesirable
    effects.

    Usage::

      >>> import requests
      >>> req = requests.Request('GET', 'https://httpbin.org/get')
      >>> r = req.prepare()
      >>> r
      <PreparedRequest [GET]>

      >>> s = requests.Session()
      >>> s.send(r)
      <Response [200]>
    """

    def __init__(self):
        #: HTTP verb to send to the server.
        self.method = None
        #: HTTP URL to send the request to.
        self.url = None
        #: dictionary of HTTP headers.
        self.headers = None
        # The `CookieJar` used to create the Cookie header will be stored here
        # after prepare_cookies is called
        self._cookies = None
        #: request body to send to the server.
        self.body = None
        #: dictionary of callback hooks, for internal usage.
        self.hooks = default_hooks()
        #: integer denoting starting position of a readable file-like body.
        self._body_position = None

    def prepare(self,
            method=None, url=None, headers=None, files=None, data=None,
            params=None, auth=None, cookies=None, hooks=None, json=None):
        """Prepares the entire request with the given parameters."""

        self.prepare_method(method)
        self.prepare_url(url, params)
        self.prepare_headers(headers)
        self.prepare_cookies(cookies)
        self.prepare_body(data, files, json)
        self.prepare_auth(auth, url)

        # Note that prepare_auth must be last to enable authentication schemes
        # such as OAuth to work on a fully prepared request.

        # This MUST go after prepare_auth. Authenticators could add a hook
        self.prepare_hooks(hooks)

    def __repr__(self):
        return '<PreparedRequest [%s]>' % (self.method)

    def copy(self):
        p = PreparedRequest()
        p.method = self.method
        p.url = self.url
        p.headers = self.headers.copy() if self.headers is not None else None
        p._cookies = _copy_cookie_jar(self._cookies)
        p.body = self.body
        p.hooks = self.hooks
        p._body_position = self._body_position
        return p

    def prepare_method(self, method):
        """Prepares the given HTTP method."""
        self.method = method
        if self.method is not None:
            self.method = to_native_string(self.method.upper())

    @staticmethod
    def _get_idna_encoded_host(host):
        from pip._vendor import idna

        try:
            host = idna.encode(host, uts46=True).decode('utf-8')
        except idna.IDNAError:
            raise UnicodeError
        return host

    def prepare_url(self, url, params):
        """Prepares the given HTTP URL."""
        #: Accept objects that have string representations.
        #: We're unable to blindly call unicode/str functions
        #: as this will include the bytestring indicator (b'')
        #: on python 3.x.
        #: https://github.com/psf/requests/pull/2238
        if isinstance(url, bytes):
            url = url.decode('utf8')
        else:
            url = unicode(url) if is_py2 else str(url)

        # Remove leading whitespaces from url
        url = url.lstrip()

        # Don't do any URL preparation for non-HTTP schemes like `mailto`,
        # `data` etc to work around exceptions from `url_parse`, which
        # handles RFC 3986 only.
        if ':' in url and not url.lower().startswith('http'):
            self.url = url
            return

        # Support for unicode domain names and paths.
        try:
            scheme, auth, host, port, path, query, fragment = parse_url(url)
        except LocationParseError as e:
            raise InvalidURL(*e.args)

        if not scheme:
            error = ("Invalid URL {0!r}: No schema supplied. Perhaps you meant http://{0}?")
            error = error.format(to_native_string(url, 'utf8'))

            raise MissingSchema(error)

        if not host:
            raise InvalidURL("Invalid URL %r: No host supplied" % url)

        # In general, we want to try IDNA encoding the hostname if the string contains
        # non-ASCII characters. This allows users to automatically get the correct IDNA
        # behaviour. For strings containing only ASCII characters, we need to also verify
        # it doesn't start with a wildcard (*), before allowing the unencoded hostname.
        if not unicode_is_ascii(host):
            try:
                host = self._get_idna_encoded_host(host)
            except UnicodeError:
                raise InvalidURL('URL has an invalid label.')
        elif host.startswith(u'*'):
            raise InvalidURL('URL has an invalid label.')

        # Carefully reconstruct the network location
        netloc = auth or ''
        if netloc:
            netloc += '@'
        netloc += host
        if port:
            netloc += ':' + str(port)

        # Bare domains aren't valid URLs.
        if not path:
            path = '/'

        if is_py2:
            if isinstance(scheme, str):
                scheme = scheme.encode('utf-8')
            if isinstance(netloc, str):
                netloc = netloc.encode('utf-8')
            if isinstance(path, str):
                path = path.encode('utf-8')
            if isinstance(query, str):
                query = query.encode('utf-8')
            if isinstance(fragment, str):
                fragment = fragment.encode('utf-8')

        if isinstance(params, (str, bytes)):
            params = to_native_string(params)

        enc_params = self._encode_params(params)
        if enc_params:
            if query:
                query = '%s&%s' % (query, enc_params)
            else:
                query = enc_params

        url = requote_uri(urlunparse([scheme, netloc, path, None, query, fragment]))
        self.url = url

    def prepare_headers(self, headers):
        """Prepares the given HTTP headers."""

        self.headers = CaseInsensitiveDict()
        if headers:
            for header in headers.items():
                # Raise exception on invalid header value.
                check_header_validity(header)
                name, value = header
                self.headers[to_native_string(name)] = value

    def prepare_body(self, data, files, json=None):
        """Prepares the given HTTP body data."""

        # Check if file, fo, generator, iterator.
        # If not, run through normal process.

        # Nottin' on you.
        body = None
        content_type = None

        if not data and json is not None:
            # urllib3 requires a bytes-like body. Python 2's json.dumps
            # provides this natively, but Python 3 gives a Unicode string.
            content_type = 'application/json'

            try:
              body = complexjson.dumps(json, allow_nan=False)
            except ValueError as ve:
              raise InvalidJSONError(ve, request=self)

            if not isinstance(body, bytes):
                body = body.encode('utf-8')

        is_stream = all([
            hasattr(data, '__iter__'),
            not isinstance(data, (basestring, list, tuple, Mapping))
        ])

        if is_stream:
            try:
                length = super_len(data)
            except (TypeError, AttributeError, UnsupportedOperation):
                length = None

            body = data

            if getattr(body, 'tell', None) is not None:
                # Record the current file position before reading.
                # This will allow us to rewind a file in the event
                # of a redirect.
                try:
                    self._body_position = body.tell()
                except (IOError, OSError):
                    # This differentiates from None, allowing us to catch
                    # a failed `tell()` later when trying to rewind the body
                    self._body_position = object()

            if files:
                raise NotImplementedError('Streamed bodies and files are mutually exclusive.')

            if length:
                self.headers['Content-Length'] = builtin_str(length)
            else:
                self.headers['Transfer-Encoding'] = 'chunked'
        else:
            # Multi-part file uploads.
            if files:
                (body, content_type) = self._encode_files(files, data)
            else:
                if data:
                    body = self._encode_params(data)
                    if isinstance(data, basestring) or hasattr(data, 'read'):
                        content_type = None
                    else:
                        content_type = 'application/x-www-form-urlencoded'

            self.prepare_content_length(body)

            # Add content-type if it wasn't explicitly provided.
            if content_type and ('content-type' not in self.headers):
                self.headers['Content-Type'] = content_type

        self.body = body

    def prepare_content_length(self, body):
        """Prepare Content-Length header based on request method and body"""
        if body is not None:
            length = super_len(body)
            if length:
                # If length exists, set it. Otherwise, we fallback
                # to Transfer-Encoding: chunked.
                self.headers['Content-Length'] = builtin_str(length)
        elif self.method not in ('GET', 'HEAD') and self.headers.get('Content-Length') is None:
            # Set Content-Length to 0 for methods that can have a body
            # but don't provide one. (i.e. not GET or HEAD)
            self.headers['Content-Length'] = '0'

    def prepare_auth(self, auth, url=''):
        """Prepares the given HTTP auth data."""

        # If no Auth is explicitly provided, extract it from the URL first.
        if auth is None:
            url_auth = get_auth_from_url(self.url)
            auth = url_auth if any(url_auth) else None

        if auth:
            if isinstance(auth, tuple) and len(auth) == 2:
                # special-case basic HTTP auth
                auth = HTTPBasicAuth(*auth)

            # Allow auth to make its changes.
            r = auth(self)

            # Update self to reflect the auth changes.
            self.__dict__.update(r.__dict__)

            # Recompute Content-Length
            self.prepare_content_length(self.body)

    def prepare_cookies(self, cookies):
        """Prepares the given HTTP cookie data.

        This function eventually generates a ``Cookie`` header from the
        given cookies using cookielib. Due to cookielib's design, the header
        will not be regenerated if it already exists, meaning this function
        can only be called once for the life of the
        :class:`PreparedRequest <PreparedRequest>` object. Any subsequent calls
        to ``prepare_cookies`` will have no actual effect, unless the "Cookie"
        header is removed beforehand.
        """
        if isinstance(cookies, cookielib.CookieJar):
            self._cookies = cookies
        else:
            self._cookies = cookiejar_from_dict(cookies)

        cookie_header = get_cookie_header(self._cookies, self)
        if cookie_header is not None:
            self.headers['Cookie'] = cookie_header

    def prepare_hooks(self, hooks):
        """Prepares the given hooks."""
        # hooks can be passed as None to the prepare method and to this
        # method. To prevent iterating over None, simply use an empty list
        # if hooks is False-y
        hooks = hooks or []
        for event in hooks:
            self.register_hook(event, hooks[event])


class Response(object):
    """The :class:`Response <Response>` object, which contains a
    server's response to an HTTP request.
    """

    __attrs__ = [
        '_content', 'status_code', 'headers', 'url', 'history',
        'encoding', 'reason', 'cookies', 'elapsed', 'request'
    ]

    def __init__(self):
        self._content = False
        self._content_consumed = False
        self._next = None

        #: Integer Code of responded HTTP Status, e.g. 404 or 200.
        self.status_code = None

        #: Case-insensitive Dictionary of Response Headers.
        #: For example, ``headers['content-encoding']`` will return the
        #: value of a ``'Content-Encoding'`` response header.
        self.headers = CaseInsensitiveDict()

        #: File-like object representation of response (for advanced usage).
        #: Use of ``raw`` requires that ``stream=True`` be set on the request.
        #: This requirement does not apply for use internally to Requests.
        self.raw = None

        #: Final URL location of Response.
        self.url = None

        #: Encoding to decode with when accessing r.text.
        self.encoding = None

        #: A list of :class:`Response <Response>` objects from
        #: the history of the Request. Any redirect responses will end
        #: up here. The list is sorted from the oldest to the most recent request.
        self.history = []

        #: Textual reason of responded HTTP Status, e.g. "Not Found" or "OK".
        self.reason = None

        #: A CookieJar of Cookies the server sent back.
        self.cookies = cookiejar_from_dict({})

        #: The amount of time elapsed between sending the request
        #: and the arrival of the response (as a timedelta).
        #: This property specifically measures the time taken between sending
        #: the first byte of the request and finishing parsing the headers. It
        #: is therefore unaffected by consuming the response content or the
        #: value of the ``stream`` keyword argument.
        self.elapsed = datetime.timedelta(0)

        #: The :class:`PreparedRequest <PreparedRequest>` object to which this
        #: is a response.
        self.request = None

    def __enter__(self):
        return self

    def __exit__(self, *args):
        self.close()

    def __getstate__(self):
        # Consume everything; accessing the content attribute makes
        # sure the content has been fully read.
        if not self._content_consumed:
            self.content

        return {attr: getattr(self, attr, None) for attr in self.__attrs__}

    def __setstate__(self, state):
        for name, value in state.items():
            setattr(self, name, value)

        # pickled objects do not have .raw
        setattr(self, '_content_consumed', True)
        setattr(self, 'raw', None)

    def __repr__(self):
        return '<Response [%s]>' % (self.status_code)

    def __bool__(self):
        """Returns True if :attr:`status_code` is less than 400.

        This attribute checks if the status code of the response is between
        400 and 600 to see if there was a client error or a server error. If
        the status code, is between 200 and 400, this will return True. This
        is **not** a check to see if the response code is ``200 OK``.
        """
        return self.ok

    def __nonzero__(self):
        """Returns True if :attr:`status_code` is less than 400.

        This attribute checks if the status code of the response is between
        400 and 600 to see if there was a client error or a server error. If
        the status code, is between 200 and 400, this will return True. This
        is **not** a check to see if the response code is ``200 OK``.
        """
        return self.ok

    def __iter__(self):
        """Allows you to use a response as an iterator."""
        return self.iter_content(128)

    @property
    def ok(self):
        """Returns True if :attr:`status_code` is less than 400, False if not.

        This attribute checks if the status code of the response is between
        400 and 600 to see if there was a client error or a server error. If
        the status code is between 200 and 400, this will return True. This
        is **not** a check to see if the response code is ``200 OK``.
        """
        try:
            self.raise_for_status()
        except HTTPError:
            return False
        return True

    @property
    def is_redirect(self):
        """True if this Response is a well-formed HTTP redirect that could have
        been processed automatically (by :meth:`Session.resolve_redirects`).
        """
        return ('location' in self.headers and self.status_code in REDIRECT_STATI)

    @property
    def is_permanent_redirect(self):
        """True if this Response one of the permanent versions of redirect."""
        return ('location' in self.headers and self.status_code in (codes.moved_permanently, codes.permanent_redirect))

    @property
    def next(self):
        """Returns a PreparedRequest for the next request in a redirect chain, if there is one."""
        return self._next

    @property
    def apparent_encoding(self):
        """The apparent encoding, provided by the charset_normalizer or chardet libraries."""
        return chardet.detect(self.content)['encoding']

    def iter_content(self, chunk_size=1, decode_unicode=False):
        """Iterates over the response data.  When stream=True is set on the
        request, this avoids reading the content at once into memory for
        large responses.  The chunk size is the number of bytes it should
        read into memory.  This is not necessarily the length of each item
        returned as decoding can take place.

        chunk_size must be of type int or None. A value of None will
        function differently depending on the value of `stream`.
        stream=True will read data as it arrives in whatever size the
        chunks are received. If stream=False, data is returned as
        a single chunk.

        If decode_unicode is True, content will be decoded using the best
        available encoding based on the response.
        """

        def generate():
            # Special case for urllib3.
            if hasattr(self.raw, 'stream'):
                try:
                    for chunk in self.raw.stream(chunk_size, decode_content=True):
                        yield chunk
                except ProtocolError as e:
                    raise ChunkedEncodingError(e)
                except DecodeError as e:
                    raise ContentDecodingError(e)
                except ReadTimeoutError as e:
                    raise ConnectionError(e)
            else:
                # Standard file-like object.
                while True:
                    chunk = self.raw.read(chunk_size)
                    if not chunk:
                        break
                    yield chunk

            self._content_consumed = True

        if self._content_consumed and isinstance(self._content, bool):
            raise StreamConsumedError()
        elif chunk_size is not None and not isinstance(chunk_size, int):
            raise TypeError("chunk_size must be an int, it is instead a %s." % type(chunk_size))
        # simulate reading small chunks of the content
        reused_chunks = iter_slices(self._content, chunk_size)

        stream_chunks = generate()

        chunks = reused_chunks if self._content_consumed else stream_chunks

        if decode_unicode:
            chunks = stream_decode_response_unicode(chunks, self)

        return chunks

    def iter_lines(self, chunk_size=ITER_CHUNK_SIZE, decode_unicode=False, delimiter=None):
        """Iterates over the response data, one line at a time.  When
        stream=True is set on the request, this avoids reading the
        content at once into memory for large responses.

        .. note:: This method is not reentrant safe.
        """

        pending = None

        for chunk in self.iter_content(chunk_size=chunk_size, decode_unicode=decode_unicode):

            if pending is not None:
                chunk = pending + chunk

            if delimiter:
                lines = chunk.split(delimiter)
            else:
                lines = chunk.splitlines()

            if lines and lines[-1] and chunk and lines[-1][-1] == chunk[-1]:
                pending = lines.pop()
            else:
                pending = None

            for line in lines:
                yield line

        if pending is not None:
            yield pending

    @property
    def content(self):
        """Content of the response, in bytes."""

        if self._content is False:
            # Read the contents.
            if self._content_consumed:
                raise RuntimeError(
                    'The content for this response was already consumed')

            if self.status_code == 0 or self.raw is None:
                self._content = None
            else:
                self._content = b''.join(self.iter_content(CONTENT_CHUNK_SIZE)) or b''

        self._content_consumed = True
        # don't need to release the connection; that's been handled by urllib3
        # since we exhausted the data.
        return self._content

    @property
    def text(self):
        """Content of the response, in unicode.

        If Response.encoding is None, encoding will be guessed using
        ``charset_normalizer`` or ``chardet``.

        The encoding of the response content is determined based solely on HTTP
        headers, following RFC 2616 to the letter. If you can take advantage of
        non-HTTP knowledge to make a better guess at the encoding, you should
        set ``r.encoding`` appropriately before accessing this property.
        """

        # Try charset from content-type
        content = None
        encoding = self.encoding

        if not self.content:
            return str('')

        # Fallback to auto-detected encoding.
        if self.encoding is None:
            encoding = self.apparent_encoding

        # Decode unicode from given encoding.
        try:
            content = str(self.content, encoding, errors='replace')
        except (LookupError, TypeError):
            # A LookupError is raised if the encoding was not found which could
            # indicate a misspelling or similar mistake.
            #
            # A TypeError can be raised if encoding is None
            #
            # So we try blindly encoding.
            content = str(self.content, errors='replace')

        return content

    def json(self, **kwargs):
        r"""Returns the json-encoded content of a response, if any.

        :param \*\*kwargs: Optional arguments that ``json.loads`` takes.
        :raises simplejson.JSONDecodeError: If the response body does not
            contain valid json and simplejson is installed.
        :raises json.JSONDecodeError: If the response body does not contain
            valid json and simplejson is not installed on Python 3.
        :raises ValueError: If the response body does not contain valid
            json and simplejson is not installed on Python 2.        
        """

        if not self.encoding and self.content and len(self.content) > 3:
            # No encoding set. JSON RFC 4627 section 3 states we should expect
            # UTF-8, -16 or -32. Detect which one to use; If the detection or
            # decoding fails, fall back to `self.text` (using charset_normalizer to make
            # a best guess).
            encoding = guess_json_utf(self.content)
            if encoding is not None:
                try:
                    return complexjson.loads(
                        self.content.decode(encoding), **kwargs
                    )
                except UnicodeDecodeError:
                    # Wrong UTF codec detected; usually because it's not UTF-8
                    # but some other 8-bit codec.  This is an RFC violation,
                    # and the server didn't bother to tell us what codec *was*
                    # used.
                    pass
        return complexjson.loads(self.text, **kwargs)

    @property
    def links(self):
        """Returns the parsed header links of the response, if any."""

        header = self.headers.get('link')

        # l = MultiDict()
        l = {}

        if header:
            links = parse_header_links(header)

            for link in links:
                key = link.get('rel') or link.get('url')
                l[key] = link

        return l

    def raise_for_status(self):
        """Raises :class:`HTTPError`, if one occurred."""

        http_error_msg = ''
        if isinstance(self.reason, bytes):
            # We attempt to decode utf-8 first because some servers
            # choose to localize their reason strings. If the string
            # isn't utf-8, we fall back to iso-8859-1 for all other
            # encodings. (See PR #3538)
            try:
                reason = self.reason.decode('utf-8')
            except UnicodeDecodeError:
                reason = self.reason.decode('iso-8859-1')
        else:
            reason = self.reason

        if 400 <= self.status_code < 500:
            http_error_msg = u'%s Client Error: %s for url: %s' % (self.status_code, reason, self.url)

        elif 500 <= self.status_code < 600:
            http_error_msg = u'%s Server Error: %s for url: %s' % (self.status_code, reason, self.url)

        if http_error_msg:
            raise HTTPError(http_error_msg, response=self)

    def close(self):
        """Releases the connection back to the pool. Once this method has been
        called the underlying ``raw`` object must not be accessed again.

        *Note: Should not normally need to be called explicitly.*
        """
        if not self._content_consumed:
            self.raw.close()

        release_conn = getattr(self.raw, 'release_conn', None)
        if release_conn is not None:
            release_conn()
site-packages/pip/_vendor/requests/_internal_utils.py000064400000002110150564745010017131 0ustar00# -*- coding: utf-8 -*-

"""
requests._internal_utils
~~~~~~~~~~~~~~

Provides utility functions that are consumed internally by Requests
which depend on extremely few external helpers (such as compat)
"""

from .compat import is_py2, builtin_str, str


def to_native_string(string, encoding='ascii'):
    """Given a string object, regardless of type, returns a representation of
    that string in the native string type, encoding and decoding where
    necessary. This assumes ASCII unless told otherwise.
    """
    if isinstance(string, builtin_str):
        out = string
    else:
        if is_py2:
            out = string.encode(encoding)
        else:
            out = string.decode(encoding)

    return out


def unicode_is_ascii(u_string):
    """Determine if unicode string only contains ASCII characters.

    :param str u_string: unicode string to check. Must be unicode
        and not Python 2 `str`.
    :rtype: bool
    """
    assert isinstance(u_string, str)
    try:
        u_string.encode('ascii')
        return True
    except UnicodeEncodeError:
        return False
site-packages/pip/_vendor/requests/__version__.py000064400000000671150564745010016231 0ustar00# .-. .-. .-. . . .-. .-. .-. .-.
# |(  |-  |.| | | |-  `-.  |  `-.
# ' ' `-' `-`.`-' `-' `-'  '  `-'

__title__ = 'requests'
__description__ = 'Python HTTP for Humans.'
__url__ = 'https://requests.readthedocs.io'
__version__ = '2.26.0'
__build__ = 0x022600
__author__ = 'Kenneth Reitz'
__author_email__ = 'me@kennethreitz.org'
__license__ = 'Apache 2.0'
__copyright__ = 'Copyright 2020 Kenneth Reitz'
__cake__ = u'\u2728 \U0001f370 \u2728'
site-packages/pip/_vendor/pep517/meta.py000064400000004637150564745010014052 0ustar00"""Build metadata for a project using PEP 517 hooks.
"""
import argparse
import logging
import os
import shutil
import functools

try:
    import importlib.metadata as imp_meta
except ImportError:
    import importlib_metadata as imp_meta

try:
    from zipfile import Path
except ImportError:
    from zipp import Path

from .envbuild import BuildEnvironment
from .wrappers import Pep517HookCaller, quiet_subprocess_runner
from .dirtools import tempdir, mkdir_p, dir_to_zipfile
from .build import validate_system, load_system, compat_system

log = logging.getLogger(__name__)


def _prep_meta(hooks, env, dest):
    reqs = hooks.get_requires_for_build_wheel({})
    log.info('Got build requires: %s', reqs)

    env.pip_install(reqs)
    log.info('Installed dynamic build dependencies')

    with tempdir() as td:
        log.info('Trying to build metadata in %s', td)
        filename = hooks.prepare_metadata_for_build_wheel(td, {})
        source = os.path.join(td, filename)
        shutil.move(source, os.path.join(dest, os.path.basename(filename)))


def build(source_dir='.', dest=None, system=None):
    system = system or load_system(source_dir)
    dest = os.path.join(source_dir, dest or 'dist')
    mkdir_p(dest)
    validate_system(system)
    hooks = Pep517HookCaller(
        source_dir, system['build-backend'], system.get('backend-path')
    )

    with hooks.subprocess_runner(quiet_subprocess_runner):
        with BuildEnvironment() as env:
            env.pip_install(system['requires'])
            _prep_meta(hooks, env, dest)


def build_as_zip(builder=build):
    with tempdir() as out_dir:
        builder(dest=out_dir)
        return dir_to_zipfile(out_dir)


def load(root):
    """
    Given a source directory (root) of a package,
    return an importlib.metadata.Distribution object
    with metadata build from that package.
    """
    root = os.path.expanduser(root)
    system = compat_system(root)
    builder = functools.partial(build, source_dir=root, system=system)
    path = Path(build_as_zip(builder))
    return imp_meta.PathDistribution(path)


parser = argparse.ArgumentParser()
parser.add_argument(
    'source_dir',
    help="A directory containing pyproject.toml",
)
parser.add_argument(
    '--out-dir', '-o',
    help="Destination in which to save the builds relative to source dir",
)


def main():
    args = parser.parse_args()
    build(args.source_dir, args.out_dir)


if __name__ == '__main__':
    main()
site-packages/pip/_vendor/pep517/__pycache__/colorlog.cpython-39.pyc000064400000005641150564745010021227 0ustar00a

��Re�@sddZddlZddlZzddlZWney6dZYn0dd�ZGdd�dej�Zejfdd�Z	dS)	zONicer log formatting with colours.

Code copied from Tornado, Apache licensed.
�NcCsTd}trPttjd�rPtj��rPzt��t�d�dkr:d}WntyNYn0|S)NF�isatty�colorsrT)�curses�hasattr�sys�stderrrZ	setuptermZtigetnum�	Exception)�color�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/colorlog.py�_stderr_supports_colorsrc@s>eZdZdZejdejdejdejdiZ	ddd�Z
d	d
�ZdS)�LogFormatterz&Log formatter with colour support
    ���TNcCs�tjj||d�i|_|r�t�r�t�d�p6t�d�p6d}dtjkrNdkr\nn
t	|d�}|j
��D] \}}t	t�||�d�|j|<qft	t�d�d�|_
t��}|��d	|_t��nd|_
d
|_dS)aO
        :arg bool color: Enables color support.
        :arg string fmt: Log message format.
        It will be applied to the attributes dict of log records. The
        text between ``%(color)s`` and ``%(end_color)s`` will be colored
        depending on the level if color support is on.
        :arg dict colors: color mappings from logging level to terminal color
        code
        :arg string datefmt: Datetime format.
        Used for formatting ``(asctime)`` placeholder in ``prefix_fmt``.
        .. versionchanged:: 3.2
        Added ``fmt`` and ``datefmt`` arguments.
        )�datefmtZsetafZsetf�)rr)rrr�asciiZsgr0r�FN)�logging�	Formatter�__init__�_colorsrrZtigetstrr�version_info�str�DEFAULT_COLORS�itemsZtparm�_normalZinitscrZgetmaxyx�	termwidthZendwin)�selfr	rZfg_color�levelno�codeZscrr
r
rr2s(

��
�
zLogFormatter.__init__cCs�t|j�}dj|jd|jd�}|t|�|jkrJd|j|t|�}nd}|j|jvrn|j|j}|j}nd}}|j||||S)Nz{initial}-{name}r)�initial�name� z  r)	�len�message�format�	levelnamer#rr rr)r�record�mlenZ
right_text�spaceZstart_colorZ	end_colorr
r
r�
formatMessage]s
�zLogFormatter.formatMessage)TN)�__name__�
__module__�__qualname__�__doc__r�INFO�WARNING�ERROR�CRITICALrrr,r
r
r
rr
(s�
+r
cCs0t��}|�t��tj�|�tj�|�dS)N)r�
StreamHandler�setFormatterr
�root�
addHandler�setLevel)�level�handlerr
r
r�enable_colourful_outputosr<)
r0rrr�ImportErrorrrr
r1r<r
r
r
r�<module>s
Gsite-packages/pip/_vendor/pep517/__pycache__/dirtools.cpython-39.pyc000064400000002552150564745010021244 0ustar00a

��Rei�@sZddlZddlZddlZddlZddlZddlZddlZejdd��Zdd�Z	dd�Z
dS)�Nc	cs.t��}z|VWt�|�nt�|�0dS)z2Create a temporary directory in a context manager.N)�tempfile�mkdtemp�shutil�rmtree)�td�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/dirtools.py�tempdir
sr	c
OsJztj|i|��WStyD}z|jtjkr0�WYd}~n
d}~00dS)zWLike `mkdir`, but does not raise an exception if the
    directory already exists.
    N)�os�mkdir�OSError�errno�EEXIST)�args�kwargs�excrrr�mkdir_ps
rcCs�t��}t�|d�}t�|�D]r\}}}|D]0}tj�||�}tj�||�}|�	|dd�q,|D],}tj�||�}tj�||�}|�
||�qbq|S)z0Construct an in-memory zip file for a directory.�w�/�)�io�BytesIO�zipfile�ZipFiler
�walk�path�join�relpath�writestr�write)�root�buffer�zip_file�dirs�filesrZfs_path�rel_pathrrr�dir_to_zipfilesr&)r
r�
contextlibrrr
r�contextmanagerr	rr&rrrr�<module>s
	site-packages/pip/_vendor/pep517/__pycache__/compat.cpython-39.pyc000064400000003040150564745010020661 0ustar00a

��Re��@s�dZddlZddlZddlZejddkr<dd�Zdd�Zndd�Zd	d�ZzeZWneyje	ZYn0ejd
kr�ddl
mZdd
�Z
ddl
mZnddlmZ
ddlmZdS)zPython 2/3 compatibility�N�cKsFt|ddd��$}tj||fi|��Wd�n1s80YdS)N�w�utf-8��encoding��open�json�dump��obj�path�kwargs�f�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/compat.py�
write_jsonsrcCs<t|ddd��}t�|�Wd�S1s.0YdS)N�rrr�rr	�load�r
rrrr�	read_jsonsrcKsFt|d��(}tj||fddi|��Wd�n1s80YdS)N�wbrrrrrrrrscCs8t|d��}t�|�Wd�S1s*0YdS)N�rbrrrrrrs)r�)rcCs2tj|ddd�}zt|�W|��S|��0dS)N�utf8�)r�newline)�io�
TextIOWrapper�
_toml_load�detach)rrrrr�	toml_load)s�r")�TomlDecodeError)�TOMLDecodeError)�__doc__rr	�sys�version_inforr�FileNotFoundError�	NameError�IOErrorZtomlrr r"r#r$�pip._vendor.tomlirrrr�<module>s$


site-packages/pip/_vendor/pep517/__pycache__/build.cpython-39.pyc000064400000007020150564745010020477 0ustar00a

��Re�
�@sdZddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZmZddl
mZmZe�e�Zdd	�Zd
d�Zdd
�Zdd�Zd"dd�Ze��Zejddd�ejddddd�ejddddd�ejdddd�dd �Zed!k�ree���dS)#z%Build a project using PEP 517 hooks.
�N�)�BuildEnvironment)�Pep517HookCaller)�tempdir�mkdir_p)�FileNotFoundError�	toml_loadcCs4ddh}|t|�ks0dj|t|�d�}t|��dS)z7
    Ensure build system has the requisite fields.
    �requires�
build-backendz"Missing required fields: {missing})�missingN)�set�format�
ValueError)�system�required�message�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/build.py�validate_systems
�rcCsJtj�|d�}t�|d��}t|�}Wd�n1s80Y|dS)zC
    Load the build system from a source dir (pyproject.toml).
    zpyproject.toml�rbNzbuild-system)�os�path�join�io�openr)�
source_dir�	pyproject�fZpyproject_datarrr�load_systems&rc	CsHzt|�}Wnttfy&i}Yn0|�dd�|�dddg�|S)z�
    Given a source dir, attempt to get a build system backend
    and requirements from pyproject.toml. Fallback to
    setuptools but only if the file was not found or a build
    system was not indicated.
    r
z setuptools.build_meta:__legacy__r	�
setuptools�wheel)rr�KeyError�
setdefault)rrrrr�
compat_system's
�r#c
Cs�djfit���}t||�}|i�}t�d|�|�|�t�d�t��p}t�d||�djfit���}t||�}	|	|i�}
tj�	||
�}t
�|tj�	|tj�|
���Wd�n1s�0YdS)Nzget_requires_for_build_{dist}zGot build requires: %sz$Installed dynamic build dependencieszTrying to build %s in %szbuild_{dist})
r
�locals�getattr�log�info�pip_installrrrr�shutil�move�basename)�hooks�env�dist�destZget_requires_name�get_requires�reqs�tdZ
build_name�build�filename�sourcerrr�	_do_build:s




r6cCs�|p
t|�}tj�||pd�}t|�t|�t||d|�d��}t��,}|�	|d�t
||||�Wd�n1s|0YdS)Nr.r
zbackend-pathr	)rrrrrrr�getrr(r6)rr.r/rr,r-rrrr3Ls�r3rz%A directory containing pyproject.toml)�helpz--binaryz-b�
store_trueF)�action�defaultz--sourcez-sz	--out-dirz-oz>Destination in which to save the builds relative to source dircCsZt�d�ttd|js|js dnd|js.|js2dndf��}|D]}t|j||j�q@dS)NzQpep517.build is deprecated. Consider switching to https://pypi.org/project/build/�sdistr )	r&�warning�list�filterr5�binaryr3rZout_dir)�args�distsr.rrr�mainps
�rC�__main__)NN)�__doc__�argparser�loggingrr)Zenvbuildr�wrappersrZdirtoolsrr�compatrr�	getLogger�__name__r&rrr#r6r3�ArgumentParser�parser�add_argumentrC�
parse_argsrrrr�<module>sJ


����
site-packages/pip/_vendor/pep517/__pycache__/check.cpython-39.pyc000064400000012033150564745010020455 0ustar00a

��Re��@s�dZddlZddlZddlZddlZddlmZmZddl	Z	ddl
mZddlZddl
Z
ddlmZddlZddlmZddlmZmZdd	lmZdd
lmZe�e�Zdd�Zd
d�Zdd�Zddd�Z ddddd�Z!dd�Z"edkr�e �dS)zHCheck a project and backend by attempting to build using PEP 517 hooks.
�N)�isfile�join)�CalledProcessError)�mkdtemp�)�enable_colourful_output)�TOMLDecodeError�	toml_load)�BuildEnvironment)�Pep517HookCallerc
CsHt���*}z|�|�t�d�Wn*tyLt�d�YWd�dS0z|�i�}t�d|�Wn.ty�tjddd�YWd�dS0z|�|�t�d�Wn*ty�t�d	�YWd�dS0t�}t�d
|��z$z|�	|i�}t�d|�Wn<t�yJtjddd�YWt
�|�Wd�dS0|�d
��s�t�d|�Wt
�|�Wd�dSt
||�}t|��r�t�d|�n(t�d|�Wt
�|�Wd�dSt�|��r�t�d�n&t�d�Wt
�|�Wd�dSWt
�|�nt
�|�0Wd�dS1�s:0YdS)N�#Installed static build dependencies�+Failed to install static build dependenciesF�Got build requires: %s�'Failure in get_requires_for_build_sdistT��exc_info�$Installed dynamic build dependencies�,Failed to install dynamic build dependencieszTrying to build sdist in %szbuild_sdist returned %rzFailure in build_sdistz.tar.gzz*Filename %s doesn't have .tar.gz extension�Output file %s exists�Output file %s does not existzOutput file is a tar filezOutput file is not a tar file)r
�pip_install�log�infor�error�get_requires_for_build_sdist�	Exceptionr�build_sdist�shutil�rmtree�endswith�pjoinr�tarfile�
is_tarfile��hooksZbuild_sys_requires�env�reqs�td�filename�path�r*��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/check.py�check_build_sdistsh






��
�

	
�

�r,c
CsHt���*}z|�|�t�d�Wn*tyLt�d�YWd�dS0z|�i�}t�d|�Wn.ty�tjddd�YWd�dS0z|�|�t�d�Wn*ty�t�d	�YWd�dS0t�}t�d
|��z$z|�	|i�}t�d|�Wn<t�yJtjddd�YWt
�|�Wd�dS0|�d
��s�t�d|�Wt
�|�Wd�dSt
||�}t|��r�t�d|�n(t�d|�Wt
�|�Wd�dSt�|��r�t�d�n&t�d�Wt
�|�Wd�dSWt
�|�nt
�|�0Wd�dS1�s:0YdS)Nrr
FrrTrrrzTrying to build wheel in %szbuild_wheel returned %rzFailure in build_wheelz.whlz'Filename %s doesn't have .whl extensionrrzOutput file is a zip filezOutput file is not a zip file)r
rrrrr�get_requires_for_build_wheelrr�build_wheelrrrr r�zipfile�
is_zipfiler#r*r*r+�check_build_wheelPsd






�
�

	
�

�r1c	Cs�t|d�}t|�rt�d�nt�d�dSzdt�|d��}t|�}Wd�n1sX0Y|d}|d}|d}|�d	�}t�d
�Wn&t	t
fy�tjddd
�YdS0t|||�}t||�}	t
||�}
|	s�t�d�|
s�t�d�|	S)Nzpyproject.tomlzFound pyproject.tomlzMissing pyproject.tomlF�rbzbuild-system�requiresz
build-backendzbackend-pathzLoaded pyproject.tomlzInvalid pyproject.tomlTrz%Sdist checks failed; scroll up to seezWheel checks failed)r rrrr�io�openr	�getr�KeyErrorrr,r1�warning)�
source_dir�	pyproject�fZpyproject_dataZbuildsysr3�backend�backend_pathr$Zsdist_okZwheel_okr*r*r+�check�s0

&




r>cCsjt�d�t��}|jddd�|�|�}t�t|j�}|rNt	t
dd��nt	t
dd��t�d	�dS)
NzQpep517.check is deprecated. Consider switching to https://pypi.org/project/build/r9z%A directory containing pyproject.toml)�helpz
Checks passed�greenz
Checks failed�redr)
rr8�argparse�ArgumentParser�add_argument�
parse_argsrr>r9�print�ansi�sys�exit)�argvZap�args�okr*r*r+�main�s
�

rMzzzz)�reset�boldrAr@cCs8tjdkr,tj��r,t|t|�tdSt|�SdS)N�ntrN)�os�namerH�stdout�isatty�
ansi_codes�str)�s�attrr*r*r+rG�srG�__main__)N)#�__doc__rBr4�loggingrQ�os.pathrrr r�
subprocessrrHr!�tempfilerr/Zcolorlogr�compatrr	Zenvbuildr
�wrappersr�	getLogger�__name__rr,r1r>rMrUrGr*r*r*r+�<module>s8
98"
�site-packages/pip/_vendor/pep517/__pycache__/envbuild.cpython-39.pyc000064400000010702150564745010021211 0ustar00a

��Re��@s�dZddlZddlZddlZddlZddlmZddlZddlm	Z	ddl
mZddlm
Z
ddlmZmZe�e�Zd	d
�ZGdd�de�Zdd
d�Zddd�ZdS)zIBuild wheels/sdists by installing build deps to a temporary environment.
�N)�
check_call)�	get_paths)�mkdtemp�)�	toml_load)�Pep517HookCaller�
LoggerWrappercCs^t�tj�|d�d��}t|�}Wd�n1s40Y|d}|d|d|�d�fS)Nzpyproject.toml�rbzbuild-system�requiresz
build-backendzbackend-path)�io�open�os�path�joinr�get)�
source_dir�fZpyproject_dataZbuildsys�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/envbuild.py�_load_pyprojects�&�rc@s6eZdZdZdZddd�Zdd�Zdd	�Zd
d�ZdS)
�BuildEnvironmentz�Context manager to install build deps in a simple temporary environment

    Based on code I wrote for pip, which is MIT licensed.
    NTcCs
||_dS)N)�_cleanup)�self�cleanuprrr�__init__=szBuildEnvironment.__init__cCs�tdd�|_t�d|j�tj�dd�|_tj�dd�|_tj	dkrHdnd}t
||j|jd�d	�}|d
}|jr�|tj|jtjd<n|tjtjtjd<|d|dkr�|d}n|dtj|d}|jr�|tj|jtjd<n
|tjd<|S)
Nzpep517-build-env-)�prefixzTemporary build environment: %s�PATH�
PYTHONPATH�nt�posix_prefix)�base�platbase)�vars�scripts�purelib�platlib)
rr�log�infor
�environr�	save_path�save_pythonpath�namer�pathsep�defpath)rZinstall_schemeZinstall_dirsr#�lib_dirsrrr�	__enter__@s0�

��
zBuildEnvironment.__enter__cCsT|sdSt�d|�tjddddd|jgt|�}t|tttj	�tttj
�d�dS)	zAInstall dependencies into this env by calling pip in a subprocessNzCalling pip to install %sz-m�pip�installz--ignore-installedz--prefix)�stdout�stderr)r&r'�sys�
executabler�listrr�logging�INFO�ERROR)r�reqs�cmdrrr�pip_install`s��

�zBuildEnvironment.pip_installcCs~|jo|jduotj�|j�}|r.t�|j�|jdurHtj�dd�n|jtjd<|j	durntj�dd�n|j	tjd<dS)Nrr)
rrr
�isdir�shutil�rmtreer)r(�popr*)r�exc_type�exc_val�exc_tbZ
needs_cleanuprrr�__exit__ns��

zBuildEnvironment.__exit__)T)	�__name__�
__module__�__qualname__�__doc__rrr/r<rDrrrrr!s
 rc	Csz|duri}t|�\}}}t|||�}t��:}|�|�|�|�}|�|�|�||�Wd�S1sl0YdS)amBuild a wheel from a source directory using PEP 517 hooks.

    :param str source_dir: Source directory containing pyproject.toml
    :param str wheel_dir: Target directory to create wheel in
    :param dict config_settings: Options to pass to build backend

    This is a blocking function which will run pip in a subprocess to install
    build requirements.
    N)rrrr<�get_requires_for_build_wheel�build_wheel)	r�	wheel_dir�config_settingsr
�backend�backend_path�hooks�envr:rrrrJ�s



rJc	Csz|duri}t|�\}}}t|||�}t��:}|�|�|�|�}|�|�|�||�Wd�S1sl0YdS)amBuild an sdist from a source directory using PEP 517 hooks.

    :param str source_dir: Source directory containing pyproject.toml
    :param str sdist_dir: Target directory to place sdist in
    :param dict config_settings: Options to pass to build backend

    This is a blocking function which will run pip in a subprocess to install
    build requirements.
    N)rrrr<�get_requires_for_build_sdist�build_sdist)	rZ	sdist_dirrLr
rMrNrOrPr:rrrrR�s



rR)N)N)rHrr
r7r>�
subprocessrr4�	sysconfigr�tempfiler�compatr�wrappersrr�	getLoggerrEr&r�objectrrJrRrrrr�<module>s
a
site-packages/pip/_vendor/pep517/__pycache__/__init__.cpython-39.pyc000064400000000540150564745010021137 0ustar00a

��Re��@sdZdZddlTdS)z6Wrappers to build Python packages using PEP 517 hooks
z0.12.0�)�*N)�__doc__�__version__�wrappers�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/__init__.py�<module>ssite-packages/pip/_vendor/pep517/__pycache__/wrappers.cpython-39.pyc000064400000030364150564745010021252 0ustar00a

��Reu4�@s
ddlZddlmZddlZddlmZmZddlZddl	m
Z
mZmZddl
Z
ddlmZddlmZddlmZgd	�Zed
d��ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�Zddd�Zddd�Zdd�ZGdd�de�ZGdd�dej �Z!dS) �N)�contextmanager)�abspath�join)�
check_call�check_output�STDOUT)�mkdtemp�)�compat)�_in_proc_script_path)�BackendUnavailable�BackendInvalid�HookMissing�UnsupportedOperation�default_subprocess_runner�quiet_subprocess_runner�Pep517HookCallerc	cs,t�}z|VWt�|�nt�|�0dS�N)r�shutil�rmtree)�td�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/wrappers.py�tempdirsrc@seZdZdZdd�ZdS)rzEWill be raised if the backend cannot be imported in the hook process.cCs
||_dSr��	traceback��selfrrrr�__init__#szBackendUnavailable.__init__N��__name__�
__module__�__qualname__�__doc__rrrrrr!src@seZdZdZdd�ZdS)r
z)Will be raised if the backend is invalid.cCs||_||_||_dSr��backend_name�backend_path�message)rr%r&r'rrrr)szBackendInvalid.__init__Nrrrrrr
'sr
cs eZdZdZ�fdd�Z�ZS)rz Will be raised on missing hooks.cstt|��|�||_dSr)�superrr�	hook_name)rr)��	__class__rrr1szHookMissing.__init__)r r!r"r#r�
__classcell__rrr*rr/src@seZdZdZdd�ZdS)rzDMay be raised by build_sdist if the backend indicates that it can't.cCs
||_dSrrrrrrr8szUnsupportedOperation.__init__Nrrrrrr6srcCs*tj��}|r|�|�t|||d�dS)z5The default method of calling the wrapper subprocess.)�cwd�envN)�os�environ�copy�updater��cmdr-�
extra_environr.rrrr<s

rcCs,tj��}|r|�|�t|||td�dS)zDA method of calling the wrapper subprocess while suppressing output.)r-r.�stderrN)r/r0r1r2rrr3rrrrEs

rcCsntj�|�rtd��tj�|�}tj�tj�||��}tj�|�}tj�|�}tj�||g�|krjtd��|S)z�Normalise and check a backend path.

    Ensure that the requested backend path is specified as a relative path,
    and resolves to a location under the given source tree.

    Return an absolute version of the requested path.
    zpaths must be relativez paths must be inside source tree)	r/�path�isabs�
ValueErrorr�normpathr�normcase�commonprefix)�source_tree�	requested�
abs_source�
abs_requested�norm_source�norm_requestedrrr�norm_and_checkNsrCc@s�eZdZdZddd�Zedd��Zdd�Zdd	d
�Zddd
�Z	ddd�Z
d dd�Zd!dd�Zd"dd�Z
d#dd�Zd$dd�Zdd�ZdS)%ra�A wrapper around a source directory to be built with a PEP 517 backend.

    :param source_dir: The path to the source directory, containing
        pyproject.toml.
    :param build_backend: The build backend spec, as per PEP 517, from
        pyproject.toml.
    :param backend_path: The backend path, as per PEP 517, from pyproject.toml.
    :param runner: A callable that invokes the wrapper subprocess.
    :param python_executable: The Python executable used to invoke the backend

    The 'runner', if provided, must expect the following:

    - cmd: a list of strings representing the command and arguments to
      execute, as would be passed to e.g. 'subprocess.check_call'.
    - cwd: a string representing the working directory that must be
      used for the subprocess. Corresponds to the provided source_dir.
    - extra_environ: a dict mapping environment variable names to values
      which must be set for the subprocess execution.
    NcsR|durt}t|��_|�_|r2�fdd�|D�}|�_|�_|sHtj}|�_dS)Ncsg|]}t�j|��qSr)rC�
source_dir)�.0�p�rrr�
<listcomp>�sz-Pep517HookCaller.__init__.<locals>.<listcomp>)	rrrD�
build_backendr&�_subprocess_runner�sys�
executable�python_executable)rrDrIr&�runnerrMrrGrrzs

�zPep517HookCaller.__init__ccs*|j}||_zdVW||_n||_0dS)z\A context manager for temporarily overriding the default subprocess
        runner.
        N)rJ)rrN�prevrrr�subprocess_runner�s
z"Pep517HookCaller.subprocess_runnercCs|�di�S)z>Return the list of optional features supported by the backend.�_supported_features��
_call_hookrGrrrrQ�sz$Pep517HookCaller._supported_featurescCs|�dd|i�S)aIIdentify packages required for building a wheel

        Returns a list of dependency specifications, e.g.::

            ["wheel >= 0.25", "setuptools"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        �get_requires_for_build_wheel�config_settingsrR�rrUrrrrT�s�z-Pep517HookCaller.get_requires_for_build_wheelTcCs|�dt|�||d��S)auPrepare a ``*.dist-info`` folder with metadata for this project.

        Returns the name of the newly created folder.

        If the build backend defines a hook with this name, it will be called
        in a subprocess. If not, the backend will be asked to build a wheel,
        and the dist-info extracted from that (unless _allow_fallback is
        False).
        � prepare_metadata_for_build_wheel��metadata_directoryrU�_allow_fallback�rSr�rrYrUrZrrrrW�s
�z1Pep517HookCaller.prepare_metadata_for_build_wheelcCs(|durt|�}|�dt|�||d��S)avBuild a wheel from this project.

        Returns the name of the newly created file.

        In general, this will call the 'build_wheel' hook in the backend.
        However, if that was previously called by
        'prepare_metadata_for_build_wheel', and the same metadata_directory is
        used, the previously built wheel will be copied to wheel_directory.
        N�build_wheel��wheel_directoryrUrY�rrS�rr_rUrYrrrr]�s�zPep517HookCaller.build_wheelcCs|�dd|i�S)aSIdentify packages required for building an editable wheel

        Returns a list of dependency specifications, e.g.::

            ["wheel >= 0.25", "setuptools"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        �get_requires_for_build_editablerUrRrVrrrrb�s�z0Pep517HookCaller.get_requires_for_build_editablecCs|�dt|�||d��S)aPrepare a ``*.dist-info`` folder with metadata for this project.

        Returns the name of the newly created folder.

        If the build backend defines a hook with this name, it will be called
        in a subprocess. If not, the backend will be asked to build an editable
        wheel, and the dist-info extracted from that (unless _allow_fallback is
        False).
        �#prepare_metadata_for_build_editablerXr[r\rrrrc�s
�z4Pep517HookCaller.prepare_metadata_for_build_editablecCs(|durt|�}|�dt|�||d��S)a�Build an editable wheel from this project.

        Returns the name of the newly created file.

        In general, this will call the 'build_editable' hook in the backend.
        However, if that was previously called by
        'prepare_metadata_for_build_editable', and the same metadata_directory
        is used, the previously built wheel will be copied to wheel_directory.
        N�build_editabler^r`rarrrrd�s�zPep517HookCaller.build_editablecCs|�dd|i�S)a>Identify packages required for building a wheel

        Returns a list of dependency specifications, e.g.::

            ["setuptools >= 26"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        �get_requires_for_build_sdistrUrRrVrrrres�z-Pep517HookCaller.get_requires_for_build_sdistcCs|�dt|�|d��S)z�Build an sdist from this project.

        Returns the name of the newly created file.

        This calls the 'build_sdist' backend hook in a subprocess.
        �build_sdist)�sdist_directoryrUr[)rrgrUrrrrfs�zPep517HookCaller.build_sdistc	Cs�tjddkr|j�d�}n|j}d|i}|jrbtj�|j�}tjddkrZ|�t���}||d<t	���}d|i}t
j|t|d�dd�t
��8}|j}	|j|	tt|��||g|j|d	�Wd�n1s�0Yt
�t|d
��}
|
�d��rt|
�dd
���|
�d��r t|
�dd
���|
�d��rFt|j|j|
�dd
�d��|
�d��rft|
�d��pb|��|
dWd�S1�s�0YdS)Nr��ASCII�PEP517_BUILD_BACKEND�PEP517_BACKEND_PATH�kwargsz
input.json)�indent)r-r5zoutput.json�unsupportedr��
no_backend�backend_invalid�
backend_errorr$�hook_missing�missing_hook_name�
return_val)rK�version_inforI�encoder&r/�pathsepr�getfilesystemencodingrr
�
write_json�pjoinrrMrJr�strrD�	read_json�getrrr
r)rr)rlrIr5r&r�
hook_input�script�python�datarrrrS&sF
��$
�zPep517HookCaller._call_hook)NNN)N)NT)NN)N)NT)NN)N)N)r r!r"r#rrrPrQrTrWr]rbrcrdrerfrSrrrrrfs2�


�
�

�
�


rc@s<eZdZdZdd�Zdd�Zedd��Zdd	�Zd
d�Z	dS)
�
LoggerWrapperzd
    Read messages from a pipe and redirect them
    to a logger (see python's logging module).
    cCsHtj�|�d|_||_||_t��\|_|_	t�
|j�|_|��dS)NT)
�	threading�Threadr�daemon�logger�levelr/�pipe�fd_read�fd_write�fdopen�reader�start)rr�r�rrrr^szLoggerWrapper.__init__cCs|jSr)r�rGrrr�filenokszLoggerWrapper.filenocCs|�tj�r|dd�S|S)N���)�endswithr/�linesep)�msgrrr�remove_newlinenszLoggerWrapper.remove_newlinecCs |jD]}|�|�|��qdSr)r��_writer�)r�linerrr�runrs
zLoggerWrapper.runcCs|j�|j|�dSr)r��logr�)rr'rrrr�vszLoggerWrapper._writeN)
r r!r"r#rr��staticmethodr�r�r�rrrrr�Xs

r�)NN)NN)"r��
contextlibrr/�os.pathrrr{r�
subprocessrrrrK�tempfilerror
�
in_processr�__all__r�	Exceptionrr
rrrrrC�objectrr�r�rrrr�<module>s,

	
	ssite-packages/pip/_vendor/pep517/__pycache__/meta.cpython-39.pyc000064400000005617150564745010020340 0ustar00a

��Re�	�@s>dZddlZddlZddlZddlZddlZzddlmZWne	yVddl
ZYn0zddlmZWne	y�ddl
mZYn0ddlmZddlmZmZddlmZmZmZddlmZmZmZe�e�Zd	d
�Zddd
�Zefdd�Zdd�Z e�!�Z"e"j#ddd�e"j#dddd�dd�Z$edk�r:e$�dS)z2Build metadata for a project using PEP 517 hooks.
�N)�Path�)�BuildEnvironment)�Pep517HookCaller�quiet_subprocess_runner)�tempdir�mkdir_p�dir_to_zipfile)�validate_system�load_system�
compat_systemc
Cs�|�i�}t�d|�|�|�t�d�t��T}t�d|�|�|i�}tj�||�}t	�
|tj�|tj�|���Wd�n1s�0YdS)NzGot build requires: %sz$Installed dynamic build dependencieszTrying to build metadata in %s)�get_requires_for_build_wheel�log�info�pip_installr� prepare_metadata_for_build_wheel�os�path�join�shutil�move�basename)�hooks�env�dest�reqs�td�filename�source�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/meta.py�
_prep_metas


r!�.c	Cs�|p
t|�}tj�||pd�}t|�t|�t||d|�d��}|�t	��Pt
��*}|�|d�t|||�Wd�n1s�0YWd�n1s�0YdS)N�distz
build-backendzbackend-path�requires)
rrrrrr
r�get�subprocess_runnerrrrr!)�
source_dirr�systemrrrrr �build)s�r)cCs<t��"}||d�t|�Wd�S1s.0YdS)N)r)rr	)�builder�out_dirrrr �build_as_zip8s
r,cCs:tj�|�}t|�}tjt||d�}tt|��}t	�
|�S)z�
    Given a source directory (root) of a package,
    return an importlib.metadata.Distribution object
    with metadata build from that package.
    )r'r()rr�
expanduserr�	functools�partialr)rr,�imp_metaZPathDistribution)�rootr(r*rrrr �load>s
r2r'z%A directory containing pyproject.toml)�helpz	--out-dirz-oz>Destination in which to save the builds relative to source dircCst��}t|j|j�dS)N)�parser�
parse_argsr)r'r+)�argsrrr �mainVsr7�__main__)r"NN)%�__doc__�argparse�loggingrrr.�importlib.metadata�metadatar0�ImportErrorZimportlib_metadata�zipfilerZzippZenvbuildr�wrappersrrZdirtoolsrrr	r)r
rr�	getLogger�__name__rr!r,r2�ArgumentParserr4�add_argumentr7rrrr �<module>sD


��
site-packages/pip/_vendor/pep517/check.py000064400000013704150564745010014174 0ustar00"""Check a project and backend by attempting to build using PEP 517 hooks.
"""
import argparse
import io
import logging
import os
from os.path import isfile, join as pjoin
import shutil
from subprocess import CalledProcessError
import sys
import tarfile
from tempfile import mkdtemp
import zipfile

from .colorlog import enable_colourful_output
from .compat import TOMLDecodeError, toml_load
from .envbuild import BuildEnvironment
from .wrappers import Pep517HookCaller

log = logging.getLogger(__name__)


def check_build_sdist(hooks, build_sys_requires):
    with BuildEnvironment() as env:
        try:
            env.pip_install(build_sys_requires)
            log.info('Installed static build dependencies')
        except CalledProcessError:
            log.error('Failed to install static build dependencies')
            return False

        try:
            reqs = hooks.get_requires_for_build_sdist({})
            log.info('Got build requires: %s', reqs)
        except Exception:
            log.error('Failure in get_requires_for_build_sdist', exc_info=True)
            return False

        try:
            env.pip_install(reqs)
            log.info('Installed dynamic build dependencies')
        except CalledProcessError:
            log.error('Failed to install dynamic build dependencies')
            return False

        td = mkdtemp()
        log.info('Trying to build sdist in %s', td)
        try:
            try:
                filename = hooks.build_sdist(td, {})
                log.info('build_sdist returned %r', filename)
            except Exception:
                log.info('Failure in build_sdist', exc_info=True)
                return False

            if not filename.endswith('.tar.gz'):
                log.error(
                    "Filename %s doesn't have .tar.gz extension", filename)
                return False

            path = pjoin(td, filename)
            if isfile(path):
                log.info("Output file %s exists", path)
            else:
                log.error("Output file %s does not exist", path)
                return False

            if tarfile.is_tarfile(path):
                log.info("Output file is a tar file")
            else:
                log.error("Output file is not a tar file")
                return False

        finally:
            shutil.rmtree(td)

        return True


def check_build_wheel(hooks, build_sys_requires):
    with BuildEnvironment() as env:
        try:
            env.pip_install(build_sys_requires)
            log.info('Installed static build dependencies')
        except CalledProcessError:
            log.error('Failed to install static build dependencies')
            return False

        try:
            reqs = hooks.get_requires_for_build_wheel({})
            log.info('Got build requires: %s', reqs)
        except Exception:
            log.error('Failure in get_requires_for_build_sdist', exc_info=True)
            return False

        try:
            env.pip_install(reqs)
            log.info('Installed dynamic build dependencies')
        except CalledProcessError:
            log.error('Failed to install dynamic build dependencies')
            return False

        td = mkdtemp()
        log.info('Trying to build wheel in %s', td)
        try:
            try:
                filename = hooks.build_wheel(td, {})
                log.info('build_wheel returned %r', filename)
            except Exception:
                log.info('Failure in build_wheel', exc_info=True)
                return False

            if not filename.endswith('.whl'):
                log.error("Filename %s doesn't have .whl extension", filename)
                return False

            path = pjoin(td, filename)
            if isfile(path):
                log.info("Output file %s exists", path)
            else:
                log.error("Output file %s does not exist", path)
                return False

            if zipfile.is_zipfile(path):
                log.info("Output file is a zip file")
            else:
                log.error("Output file is not a zip file")
                return False

        finally:
            shutil.rmtree(td)

        return True


def check(source_dir):
    pyproject = pjoin(source_dir, 'pyproject.toml')
    if isfile(pyproject):
        log.info('Found pyproject.toml')
    else:
        log.error('Missing pyproject.toml')
        return False

    try:
        with io.open(pyproject, 'rb') as f:
            pyproject_data = toml_load(f)
        # Ensure the mandatory data can be loaded
        buildsys = pyproject_data['build-system']
        requires = buildsys['requires']
        backend = buildsys['build-backend']
        backend_path = buildsys.get('backend-path')
        log.info('Loaded pyproject.toml')
    except (TOMLDecodeError, KeyError):
        log.error("Invalid pyproject.toml", exc_info=True)
        return False

    hooks = Pep517HookCaller(source_dir, backend, backend_path)

    sdist_ok = check_build_sdist(hooks, requires)
    wheel_ok = check_build_wheel(hooks, requires)

    if not sdist_ok:
        log.warning('Sdist checks failed; scroll up to see')
    if not wheel_ok:
        log.warning('Wheel checks failed')

    return sdist_ok


def main(argv=None):
    log.warning('pep517.check is deprecated. '
                'Consider switching to https://pypi.org/project/build/')

    ap = argparse.ArgumentParser()
    ap.add_argument(
        'source_dir',
        help="A directory containing pyproject.toml")
    args = ap.parse_args(argv)

    enable_colourful_output()

    ok = check(args.source_dir)

    if ok:
        print(ansi('Checks passed', 'green'))
    else:
        print(ansi('Checks failed', 'red'))
        sys.exit(1)


ansi_codes = {
    'reset': '\x1b[0m',
    'bold': '\x1b[1m',
    'red': '\x1b[31m',
    'green': '\x1b[32m',
}


def ansi(s, attr):
    if os.name != 'nt' and sys.stdout.isatty():
        return ansi_codes[attr] + str(s) + ansi_codes['reset']
    else:
        return str(s)


if __name__ == '__main__':
    main()
site-packages/pip/_vendor/pep517/compat.py000064400000002345150564745010014401 0ustar00"""Python 2/3 compatibility"""
import io
import json
import sys


# Handle reading and writing JSON in UTF-8, on Python 3 and 2.

if sys.version_info[0] >= 3:
    # Python 3
    def write_json(obj, path, **kwargs):
        with open(path, 'w', encoding='utf-8') as f:
            json.dump(obj, f, **kwargs)

    def read_json(path):
        with open(path, 'r', encoding='utf-8') as f:
            return json.load(f)

else:
    # Python 2
    def write_json(obj, path, **kwargs):
        with open(path, 'wb') as f:
            json.dump(obj, f, encoding='utf-8', **kwargs)

    def read_json(path):
        with open(path, 'rb') as f:
            return json.load(f)


# FileNotFoundError

try:
    FileNotFoundError = FileNotFoundError
except NameError:
    FileNotFoundError = IOError


if sys.version_info < (3, 6):
    from toml import load as _toml_load  # noqa: F401

    def toml_load(f):
        w = io.TextIOWrapper(f, encoding="utf8", newline="")
        try:
            return _toml_load(w)
        finally:
            w.detach()

    from toml import TomlDecodeError as TOMLDecodeError  # noqa: F401
else:
    from pip._vendor.tomli import load as toml_load  # noqa: F401
    from pip._vendor.tomli import TOMLDecodeError  # noqa: F401
site-packages/pip/_vendor/pep517/dirtools.py000064400000002151150564745010014750 0ustar00import os
import io
import contextlib
import tempfile
import shutil
import errno
import zipfile


@contextlib.contextmanager
def tempdir():
    """Create a temporary directory in a context manager."""
    td = tempfile.mkdtemp()
    try:
        yield td
    finally:
        shutil.rmtree(td)


def mkdir_p(*args, **kwargs):
    """Like `mkdir`, but does not raise an exception if the
    directory already exists.
    """
    try:
        return os.mkdir(*args, **kwargs)
    except OSError as exc:
        if exc.errno != errno.EEXIST:
            raise


def dir_to_zipfile(root):
    """Construct an in-memory zip file for a directory."""
    buffer = io.BytesIO()
    zip_file = zipfile.ZipFile(buffer, 'w')
    for root, dirs, files in os.walk(root):
        for path in dirs:
            fs_path = os.path.join(root, path)
            rel_path = os.path.relpath(fs_path, root)
            zip_file.writestr(rel_path + '/', '')
        for path in files:
            fs_path = os.path.join(root, path)
            rel_path = os.path.relpath(fs_path, root)
            zip_file.write(fs_path, rel_path)
    return zip_file
site-packages/pip/_vendor/pep517/wrappers.py000064400000032165150564745010014764 0ustar00import threading
from contextlib import contextmanager
import os
from os.path import abspath, join as pjoin
import shutil
from subprocess import check_call, check_output, STDOUT
import sys
from tempfile import mkdtemp

from . import compat
from .in_process import _in_proc_script_path

__all__ = [
    'BackendUnavailable',
    'BackendInvalid',
    'HookMissing',
    'UnsupportedOperation',
    'default_subprocess_runner',
    'quiet_subprocess_runner',
    'Pep517HookCaller',
]


@contextmanager
def tempdir():
    td = mkdtemp()
    try:
        yield td
    finally:
        shutil.rmtree(td)


class BackendUnavailable(Exception):
    """Will be raised if the backend cannot be imported in the hook process."""
    def __init__(self, traceback):
        self.traceback = traceback


class BackendInvalid(Exception):
    """Will be raised if the backend is invalid."""
    def __init__(self, backend_name, backend_path, message):
        self.backend_name = backend_name
        self.backend_path = backend_path
        self.message = message


class HookMissing(Exception):
    """Will be raised on missing hooks."""
    def __init__(self, hook_name):
        super(HookMissing, self).__init__(hook_name)
        self.hook_name = hook_name


class UnsupportedOperation(Exception):
    """May be raised by build_sdist if the backend indicates that it can't."""
    def __init__(self, traceback):
        self.traceback = traceback


def default_subprocess_runner(cmd, cwd=None, extra_environ=None):
    """The default method of calling the wrapper subprocess."""
    env = os.environ.copy()
    if extra_environ:
        env.update(extra_environ)

    check_call(cmd, cwd=cwd, env=env)


def quiet_subprocess_runner(cmd, cwd=None, extra_environ=None):
    """A method of calling the wrapper subprocess while suppressing output."""
    env = os.environ.copy()
    if extra_environ:
        env.update(extra_environ)

    check_output(cmd, cwd=cwd, env=env, stderr=STDOUT)


def norm_and_check(source_tree, requested):
    """Normalise and check a backend path.

    Ensure that the requested backend path is specified as a relative path,
    and resolves to a location under the given source tree.

    Return an absolute version of the requested path.
    """
    if os.path.isabs(requested):
        raise ValueError("paths must be relative")

    abs_source = os.path.abspath(source_tree)
    abs_requested = os.path.normpath(os.path.join(abs_source, requested))
    # We have to use commonprefix for Python 2.7 compatibility. So we
    # normalise case to avoid problems because commonprefix is a character
    # based comparison :-(
    norm_source = os.path.normcase(abs_source)
    norm_requested = os.path.normcase(abs_requested)
    if os.path.commonprefix([norm_source, norm_requested]) != norm_source:
        raise ValueError("paths must be inside source tree")

    return abs_requested


class Pep517HookCaller(object):
    """A wrapper around a source directory to be built with a PEP 517 backend.

    :param source_dir: The path to the source directory, containing
        pyproject.toml.
    :param build_backend: The build backend spec, as per PEP 517, from
        pyproject.toml.
    :param backend_path: The backend path, as per PEP 517, from pyproject.toml.
    :param runner: A callable that invokes the wrapper subprocess.
    :param python_executable: The Python executable used to invoke the backend

    The 'runner', if provided, must expect the following:

    - cmd: a list of strings representing the command and arguments to
      execute, as would be passed to e.g. 'subprocess.check_call'.
    - cwd: a string representing the working directory that must be
      used for the subprocess. Corresponds to the provided source_dir.
    - extra_environ: a dict mapping environment variable names to values
      which must be set for the subprocess execution.
    """
    def __init__(
            self,
            source_dir,
            build_backend,
            backend_path=None,
            runner=None,
            python_executable=None,
    ):
        if runner is None:
            runner = default_subprocess_runner

        self.source_dir = abspath(source_dir)
        self.build_backend = build_backend
        if backend_path:
            backend_path = [
                norm_and_check(self.source_dir, p) for p in backend_path
            ]
        self.backend_path = backend_path
        self._subprocess_runner = runner
        if not python_executable:
            python_executable = sys.executable
        self.python_executable = python_executable

    @contextmanager
    def subprocess_runner(self, runner):
        """A context manager for temporarily overriding the default subprocess
        runner.
        """
        prev = self._subprocess_runner
        self._subprocess_runner = runner
        try:
            yield
        finally:
            self._subprocess_runner = prev

    def _supported_features(self):
        """Return the list of optional features supported by the backend."""
        return self._call_hook('_supported_features', {})

    def get_requires_for_build_wheel(self, config_settings=None):
        """Identify packages required for building a wheel

        Returns a list of dependency specifications, e.g.::

            ["wheel >= 0.25", "setuptools"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        """
        return self._call_hook('get_requires_for_build_wheel', {
            'config_settings': config_settings
        })

    def prepare_metadata_for_build_wheel(
            self, metadata_directory, config_settings=None,
            _allow_fallback=True):
        """Prepare a ``*.dist-info`` folder with metadata for this project.

        Returns the name of the newly created folder.

        If the build backend defines a hook with this name, it will be called
        in a subprocess. If not, the backend will be asked to build a wheel,
        and the dist-info extracted from that (unless _allow_fallback is
        False).
        """
        return self._call_hook('prepare_metadata_for_build_wheel', {
            'metadata_directory': abspath(metadata_directory),
            'config_settings': config_settings,
            '_allow_fallback': _allow_fallback,
        })

    def build_wheel(
            self, wheel_directory, config_settings=None,
            metadata_directory=None):
        """Build a wheel from this project.

        Returns the name of the newly created file.

        In general, this will call the 'build_wheel' hook in the backend.
        However, if that was previously called by
        'prepare_metadata_for_build_wheel', and the same metadata_directory is
        used, the previously built wheel will be copied to wheel_directory.
        """
        if metadata_directory is not None:
            metadata_directory = abspath(metadata_directory)
        return self._call_hook('build_wheel', {
            'wheel_directory': abspath(wheel_directory),
            'config_settings': config_settings,
            'metadata_directory': metadata_directory,
        })

    def get_requires_for_build_editable(self, config_settings=None):
        """Identify packages required for building an editable wheel

        Returns a list of dependency specifications, e.g.::

            ["wheel >= 0.25", "setuptools"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        """
        return self._call_hook('get_requires_for_build_editable', {
            'config_settings': config_settings
        })

    def prepare_metadata_for_build_editable(
            self, metadata_directory, config_settings=None,
            _allow_fallback=True):
        """Prepare a ``*.dist-info`` folder with metadata for this project.

        Returns the name of the newly created folder.

        If the build backend defines a hook with this name, it will be called
        in a subprocess. If not, the backend will be asked to build an editable
        wheel, and the dist-info extracted from that (unless _allow_fallback is
        False).
        """
        return self._call_hook('prepare_metadata_for_build_editable', {
            'metadata_directory': abspath(metadata_directory),
            'config_settings': config_settings,
            '_allow_fallback': _allow_fallback,
        })

    def build_editable(
            self, wheel_directory, config_settings=None,
            metadata_directory=None):
        """Build an editable wheel from this project.

        Returns the name of the newly created file.

        In general, this will call the 'build_editable' hook in the backend.
        However, if that was previously called by
        'prepare_metadata_for_build_editable', and the same metadata_directory
        is used, the previously built wheel will be copied to wheel_directory.
        """
        if metadata_directory is not None:
            metadata_directory = abspath(metadata_directory)
        return self._call_hook('build_editable', {
            'wheel_directory': abspath(wheel_directory),
            'config_settings': config_settings,
            'metadata_directory': metadata_directory,
        })

    def get_requires_for_build_sdist(self, config_settings=None):
        """Identify packages required for building a wheel

        Returns a list of dependency specifications, e.g.::

            ["setuptools >= 26"]

        This does not include requirements specified in pyproject.toml.
        It returns the result of calling the equivalently named hook in a
        subprocess.
        """
        return self._call_hook('get_requires_for_build_sdist', {
            'config_settings': config_settings
        })

    def build_sdist(self, sdist_directory, config_settings=None):
        """Build an sdist from this project.

        Returns the name of the newly created file.

        This calls the 'build_sdist' backend hook in a subprocess.
        """
        return self._call_hook('build_sdist', {
            'sdist_directory': abspath(sdist_directory),
            'config_settings': config_settings,
        })

    def _call_hook(self, hook_name, kwargs):
        # On Python 2, pytoml returns Unicode values (which is correct) but the
        # environment passed to check_call needs to contain string values. We
        # convert here by encoding using ASCII (the backend can only contain
        # letters, digits and _, . and : characters, and will be used as a
        # Python identifier, so non-ASCII content is wrong on Python 2 in
        # any case).
        # For backend_path, we use sys.getfilesystemencoding.
        if sys.version_info[0] == 2:
            build_backend = self.build_backend.encode('ASCII')
        else:
            build_backend = self.build_backend
        extra_environ = {'PEP517_BUILD_BACKEND': build_backend}

        if self.backend_path:
            backend_path = os.pathsep.join(self.backend_path)
            if sys.version_info[0] == 2:
                backend_path = backend_path.encode(sys.getfilesystemencoding())
            extra_environ['PEP517_BACKEND_PATH'] = backend_path

        with tempdir() as td:
            hook_input = {'kwargs': kwargs}
            compat.write_json(hook_input, pjoin(td, 'input.json'),
                              indent=2)

            # Run the hook in a subprocess
            with _in_proc_script_path() as script:
                python = self.python_executable
                self._subprocess_runner(
                    [python, abspath(str(script)), hook_name, td],
                    cwd=self.source_dir,
                    extra_environ=extra_environ
                )

            data = compat.read_json(pjoin(td, 'output.json'))
            if data.get('unsupported'):
                raise UnsupportedOperation(data.get('traceback', ''))
            if data.get('no_backend'):
                raise BackendUnavailable(data.get('traceback', ''))
            if data.get('backend_invalid'):
                raise BackendInvalid(
                    backend_name=self.build_backend,
                    backend_path=self.backend_path,
                    message=data.get('backend_error', '')
                )
            if data.get('hook_missing'):
                raise HookMissing(data.get('missing_hook_name') or hook_name)
            return data['return_val']


class LoggerWrapper(threading.Thread):
    """
    Read messages from a pipe and redirect them
    to a logger (see python's logging module).
    """

    def __init__(self, logger, level):
        threading.Thread.__init__(self)
        self.daemon = True

        self.logger = logger
        self.level = level

        # create the pipe and reader
        self.fd_read, self.fd_write = os.pipe()
        self.reader = os.fdopen(self.fd_read)

        self.start()

    def fileno(self):
        return self.fd_write

    @staticmethod
    def remove_newline(msg):
        return msg[:-1] if msg.endswith(os.linesep) else msg

    def run(self):
        for line in self.reader:
            self._write(self.remove_newline(line))

    def _write(self, message):
        self.logger.log(self.level, message)
site-packages/pip/_vendor/pep517/build.py000064400000006601150564745010014214 0ustar00"""Build a project using PEP 517 hooks.
"""
import argparse
import io
import logging
import os
import shutil

from .envbuild import BuildEnvironment
from .wrappers import Pep517HookCaller
from .dirtools import tempdir, mkdir_p
from .compat import FileNotFoundError, toml_load

log = logging.getLogger(__name__)


def validate_system(system):
    """
    Ensure build system has the requisite fields.
    """
    required = {'requires', 'build-backend'}
    if not (required <= set(system)):
        message = "Missing required fields: {missing}".format(
            missing=required-set(system),
        )
        raise ValueError(message)


def load_system(source_dir):
    """
    Load the build system from a source dir (pyproject.toml).
    """
    pyproject = os.path.join(source_dir, 'pyproject.toml')
    with io.open(pyproject, 'rb') as f:
        pyproject_data = toml_load(f)
    return pyproject_data['build-system']


def compat_system(source_dir):
    """
    Given a source dir, attempt to get a build system backend
    and requirements from pyproject.toml. Fallback to
    setuptools but only if the file was not found or a build
    system was not indicated.
    """
    try:
        system = load_system(source_dir)
    except (FileNotFoundError, KeyError):
        system = {}
    system.setdefault(
        'build-backend',
        'setuptools.build_meta:__legacy__',
    )
    system.setdefault('requires', ['setuptools', 'wheel'])
    return system


def _do_build(hooks, env, dist, dest):
    get_requires_name = 'get_requires_for_build_{dist}'.format(**locals())
    get_requires = getattr(hooks, get_requires_name)
    reqs = get_requires({})
    log.info('Got build requires: %s', reqs)

    env.pip_install(reqs)
    log.info('Installed dynamic build dependencies')

    with tempdir() as td:
        log.info('Trying to build %s in %s', dist, td)
        build_name = 'build_{dist}'.format(**locals())
        build = getattr(hooks, build_name)
        filename = build(td, {})
        source = os.path.join(td, filename)
        shutil.move(source, os.path.join(dest, os.path.basename(filename)))


def build(source_dir, dist, dest=None, system=None):
    system = system or load_system(source_dir)
    dest = os.path.join(source_dir, dest or 'dist')
    mkdir_p(dest)

    validate_system(system)
    hooks = Pep517HookCaller(
        source_dir, system['build-backend'], system.get('backend-path')
    )

    with BuildEnvironment() as env:
        env.pip_install(system['requires'])
        _do_build(hooks, env, dist, dest)


parser = argparse.ArgumentParser()
parser.add_argument(
    'source_dir',
    help="A directory containing pyproject.toml",
)
parser.add_argument(
    '--binary', '-b',
    action='store_true',
    default=False,
)
parser.add_argument(
    '--source', '-s',
    action='store_true',
    default=False,
)
parser.add_argument(
    '--out-dir', '-o',
    help="Destination in which to save the builds relative to source dir",
)


def main(args):
    log.warning('pep517.build is deprecated. '
                'Consider switching to https://pypi.org/project/build/')

    # determine which dists to build
    dists = list(filter(None, (
        'sdist' if args.source or not args.binary else None,
        'wheel' if args.binary or not args.source else None,
    )))

    for dist in dists:
        build(args.source_dir, dist, args.out_dir)


if __name__ == '__main__':
    main(parser.parse_args())
site-packages/pip/_vendor/pep517/__init__.py000064400000000202150564745010014643 0ustar00"""Wrappers to build Python packages using PEP 517 hooks
"""

__version__ = '0.12.0'

from .wrappers import *  # noqa: F401, F403
site-packages/pip/_vendor/pep517/envbuild.py000064400000013724150564745010014731 0ustar00"""Build wheels/sdists by installing build deps to a temporary environment.
"""

import io
import os
import logging
import shutil
from subprocess import check_call
import sys
from sysconfig import get_paths
from tempfile import mkdtemp

from .compat import toml_load
from .wrappers import Pep517HookCaller, LoggerWrapper

log = logging.getLogger(__name__)


def _load_pyproject(source_dir):
    with io.open(
            os.path.join(source_dir, 'pyproject.toml'),
            'rb',
            ) as f:
        pyproject_data = toml_load(f)
    buildsys = pyproject_data['build-system']
    return (
        buildsys['requires'],
        buildsys['build-backend'],
        buildsys.get('backend-path'),
    )


class BuildEnvironment(object):
    """Context manager to install build deps in a simple temporary environment

    Based on code I wrote for pip, which is MIT licensed.
    """
    # Copyright (c) 2008-2016 The pip developers (see AUTHORS.txt file)
    #
    # Permission is hereby granted, free of charge, to any person obtaining
    # a copy of this software and associated documentation files (the
    # "Software"), to deal in the Software without restriction, including
    # without limitation the rights to use, copy, modify, merge, publish,
    # distribute, sublicense, and/or sell copies of the Software, and to
    # permit persons to whom the Software is furnished to do so, subject to
    # the following conditions:
    #
    # The above copyright notice and this permission notice shall be
    # included in all copies or substantial portions of the Software.
    #
    # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
    # EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
    # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
    # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE
    # LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
    # OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
    # WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.

    path = None

    def __init__(self, cleanup=True):
        self._cleanup = cleanup

    def __enter__(self):
        self.path = mkdtemp(prefix='pep517-build-env-')
        log.info('Temporary build environment: %s', self.path)

        self.save_path = os.environ.get('PATH', None)
        self.save_pythonpath = os.environ.get('PYTHONPATH', None)

        install_scheme = 'nt' if (os.name == 'nt') else 'posix_prefix'
        install_dirs = get_paths(install_scheme, vars={
            'base': self.path,
            'platbase': self.path,
        })

        scripts = install_dirs['scripts']
        if self.save_path:
            os.environ['PATH'] = scripts + os.pathsep + self.save_path
        else:
            os.environ['PATH'] = scripts + os.pathsep + os.defpath

        if install_dirs['purelib'] == install_dirs['platlib']:
            lib_dirs = install_dirs['purelib']
        else:
            lib_dirs = install_dirs['purelib'] + os.pathsep + \
                install_dirs['platlib']
        if self.save_pythonpath:
            os.environ['PYTHONPATH'] = lib_dirs + os.pathsep + \
                self.save_pythonpath
        else:
            os.environ['PYTHONPATH'] = lib_dirs

        return self

    def pip_install(self, reqs):
        """Install dependencies into this env by calling pip in a subprocess"""
        if not reqs:
            return
        log.info('Calling pip to install %s', reqs)
        cmd = [
            sys.executable, '-m', 'pip', 'install', '--ignore-installed',
            '--prefix', self.path] + list(reqs)
        check_call(
            cmd,
            stdout=LoggerWrapper(log, logging.INFO),
            stderr=LoggerWrapper(log, logging.ERROR),
        )

    def __exit__(self, exc_type, exc_val, exc_tb):
        needs_cleanup = (
            self._cleanup and
            self.path is not None and
            os.path.isdir(self.path)
        )
        if needs_cleanup:
            shutil.rmtree(self.path)

        if self.save_path is None:
            os.environ.pop('PATH', None)
        else:
            os.environ['PATH'] = self.save_path

        if self.save_pythonpath is None:
            os.environ.pop('PYTHONPATH', None)
        else:
            os.environ['PYTHONPATH'] = self.save_pythonpath


def build_wheel(source_dir, wheel_dir, config_settings=None):
    """Build a wheel from a source directory using PEP 517 hooks.

    :param str source_dir: Source directory containing pyproject.toml
    :param str wheel_dir: Target directory to create wheel in
    :param dict config_settings: Options to pass to build backend

    This is a blocking function which will run pip in a subprocess to install
    build requirements.
    """
    if config_settings is None:
        config_settings = {}
    requires, backend, backend_path = _load_pyproject(source_dir)
    hooks = Pep517HookCaller(source_dir, backend, backend_path)

    with BuildEnvironment() as env:
        env.pip_install(requires)
        reqs = hooks.get_requires_for_build_wheel(config_settings)
        env.pip_install(reqs)
        return hooks.build_wheel(wheel_dir, config_settings)


def build_sdist(source_dir, sdist_dir, config_settings=None):
    """Build an sdist from a source directory using PEP 517 hooks.

    :param str source_dir: Source directory containing pyproject.toml
    :param str sdist_dir: Target directory to place sdist in
    :param dict config_settings: Options to pass to build backend

    This is a blocking function which will run pip in a subprocess to install
    build requirements.
    """
    if config_settings is None:
        config_settings = {}
    requires, backend, backend_path = _load_pyproject(source_dir)
    hooks = Pep517HookCaller(source_dir, backend, backend_path)

    with BuildEnvironment() as env:
        env.pip_install(requires)
        reqs = hooks.get_requires_for_build_sdist(config_settings)
        env.pip_install(reqs)
        return hooks.build_sdist(sdist_dir, config_settings)
site-packages/pip/_vendor/pep517/in_process/__pycache__/_in_process.cpython-39.pyc000064400000024104150564745010024051 0ustar00a

��Re�+�@s|dZddlmZddlmZddlZddlZddlZddlmZddl	Z	ddl
Z
ddlZddlZej
ddkr�dd�Zd	d
�Zndd�Zdd
�ZGd
d�de�ZGdd�de�ZGdd�de�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!Zd"d#�Zd$d%�Zd&d'�Zd8d(d)�Zd9d*d+�Z d,d-�Z!Gd.d/�d/e�Z"Gd0d1�d1e�Z#d2d3�Z$hd4�Z%d5d6�Z&e'd7k�rxe&�dS):agThis is invoked in a subprocess to call the build backend hooks.

It expects:
- Command line args: hook_name, control_dir
- Environment variables:
      PEP517_BUILD_BACKEND=entry.point:spec
      PEP517_BACKEND_PATH=paths (separated with os.pathsep)
- control_dir/input.json:
  - {"kwargs": {...}}

Results:
- control_dir/output.json
  - {"return_val": ...}
�)�glob)�
import_moduleN)�join�cKsFt|ddd��$}tj||fi|��Wd�n1s80YdS)N�w�utf-8��encoding��open�json�dump��obj�path�kwargs�f�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/in_process/_in_process.py�
write_json!srcCs<t|ddd��}t�|�Wd�S1s.0YdS)N�rrr�rr�load�rrrrr�	read_json%srcKsFt|d��(}tj||fddi|��Wd�n1s80YdS)N�wbr	rr
rrrrr+scCs8t|d��}t�|�Wd�S1s*0YdS)N�rbrrrrrr/sc@seZdZdZdd�ZdS)�BackendUnavailablez&Raised if we cannot import the backendcCs
||_dS�N��	traceback��selfr rrr�__init__6szBackendUnavailable.__init__N��__name__�
__module__�__qualname__�__doc__r#rrrrr4src@seZdZdZdd�ZdS)�BackendInvalidz Raised if the backend is invalidcCs
||_dSr)�message)r"r*rrrr#<szBackendInvalid.__init__Nr$rrrrr):sr)cs"eZdZdZd�fdd�	Z�ZS)�HookMissingzARaised if a hook is missing and we are not executing the fallbackNcstt|��|�||_dSr)�superr+r#�	hook_name)r"r-��	__class__rrr#BszHookMissing.__init__)N)r%r&r'r(r#�
__classcell__rrr.rr+@sr+cCs<tj�tj�|��}tj�tj�|��}tj�||g�|kS)z5Test if a file is located within the given directory.)�osr�normcase�abspath�commonprefix)�filename�	directoryrrr�contained_inGsr7cs�tj�d�}|r*|�tj�}|tjdd�<tjd}|�d�\}}}zt|��Wnt	ynt
t����Yn0|r�t
�fdd�|D��s�td��|r�|�d	�D]}t�|��q��S)
zFind and load the build backend�PEP517_BACKEND_PATHNr�PEP517_BUILD_BACKEND�:c3s|]}t�j|�VqdSr)r7�__file__)�.0r�rrr�	<genexpr>^s�z!_build_backend.<locals>.<genexpr>z(Backend was not loaded from backend-path�.)r1�environ�get�split�pathsep�sysr�	partitionr�ImportErrorrr �
format_exc�anyr)�getattr)�backend_pathZextra_pathitems�epZmod_path�_Zobj_pathZ	path_partrr=r�_build_backendNs&
�rMcCs"t�}g}t|d�r|�d�|S)z�Return the list of options features supported by the backend.

    Returns a list of strings.
    The only possible value is 'build_editable'.
    �build_editable)rM�hasattr�append)�backend�featuresrrr�_supported_featuresjs


rScCs4t�}z
|j}Wnty&gYS0||�SdS�zfInvoke the optional get_requires_for_build_wheel hook

    Returns [] if the hook is not defined.
    N)rM�get_requires_for_build_wheel�AttributeError��config_settingsrQ�hookrrrrUws

rUcCs4t�}z
|j}Wnty&gYS0||�SdS)ziInvoke the optional get_requires_for_build_editable hook

    Returns [] if the hook is not defined.
    N)rM�get_requires_for_build_editablerVrWrrrrZ�s

rZcCsTt�}z
|j}Wn4tyD|s(t��|�||�}t|||�YS0|||�SdS)z�Invoke optional prepare_metadata_for_build_wheel

    Implements a fallback by building a wheel if the hook isn't defined,
    unless _allow_fallback is False in which case HookMissing is raised.
    N)rM� prepare_metadata_for_build_wheelrVr+�build_wheel�_get_wheel_metadata_from_wheel)�metadata_directoryrX�_allow_fallbackrQrY�whl_basenamerrrr[�s
�
r[cCs~t�}z
|j}Wn^tyn|s(t��z
|j}WntyNtdd��Yn0|||�}t|||�YSYn0|||�SdS)z�Invoke optional prepare_metadata_for_build_editable

    Implements a fallback by building an editable wheel if the hook isn't
    defined, unless _allow_fallback is False in which case HookMissing is
    raised.
    rN)r-N)rM�#prepare_metadata_for_build_editablerVr+rNr])r^rXr_rQrYZ
build_hookr`rrrra�s 


�raZPEP517_ALREADY_BUILT_WHEELcCs@g}|��D]}t�d|�}|r|�|�q|r4|Std��dS)z6Identify the .dist-info folder inside a wheel ZipFile.z[^/\\]+-[^/\\]+\.dist-info/z#No .dist-info folder found in wheelN)�namelist�re�matchrP�	Exception)Zwhl_zip�resr�mrrr�_dist_info_files�srhcCs�ddlm}ttj�|t�d��Wd�n1s60Ytj�||�}||��&}t|�}|j||d�Wd�n1s�0Y|d�	d�dS)z�Extract the metadata from a wheel.

    Fallback for when the build backend does not
    define the 'get_wheel_metadata' hook.
    r)�ZipFilerN)r�members�/)
�zipfilerirr1rr�WHEEL_BUILT_MARKERrh�
extractallrB)r`r^rXriZwhl_fileZzipfZ	dist_inforrrr]�s
,r]cCsl|sdStj�|�}tj�t|t��s*dSttj�|d��}|sLtd�dSt	|�dkrdtd�dS|dS)zHCheck for a wheel already built during the get_wheel_metadata hook.
    Nz*.whlz+Found wheel built marker, but no .whl files�zHFound multiple .whl files; unspecified behaviour. Will call build_wheel.r)
r1r�dirname�isfile�pjoinrmrr�print�len)r^Zmetadata_parentZ	whl_filesrrr�_find_already_built_wheel�srucCs4t|�}|r$t�||�tj�|�St��|||�S)z�Invoke the mandatory build_wheel hook.

    If a wheel was already built in the
    prepare_metadata_for_build_wheel fallback, this
    will copy it rather than rebuilding the wheel.
    )ru�shutil�copy2r1r�basenamerMr\)�wheel_directoryrXr^�prebuilt_whlrrrr\�s
�r\cCs^t�}z
|j}Wnty(t��Yn20t|�}|rNt�||�tj�	|�S||||�SdS)z�Invoke the optional build_editable hook.

    If a wheel was already built in the
    prepare_metadata_for_build_editable fallback, this
    will copy it rather than rebuilding the wheel.
    N)
rMrNrVr+rurvrwr1rrx)ryrXr^rQrYrzrrrrN	s
rNcCs4t�}z
|j}Wnty&gYS0||�SdSrT)rM�get_requires_for_build_sdistrVrWrrrr{s

r{c@seZdZdZdS)�_DummyExceptionz(Nothing should ever raise this exceptionN)r%r&r'r(rrrrr|,sr|c@seZdZdZdd�ZdS)�GotUnsupportedOperationz9For internal use when backend raises UnsupportedOperationcCs
||_dSrrr!rrrr#2sz GotUnsupportedOperation.__init__Nr$rrrrr}0sr}cCs@t�}z|�||�WSt|dt�y:tt����Yn0dS)z&Invoke the mandatory build_sdist hook.�UnsupportedOperationN)rM�build_sdistrIr|r}r rG)�sdist_directoryrXrQrrrr6s
r>	rZrNrarrUr{rSr\r[c
Cstttj�dkrt�d�tjd}tjd}|tvrBt�d|�t�|}tt|d��}ddd�}z|fi|d	��|d
<Wn�ty�}zd|d<|j	|d
<WYd}~n�d}~0t
y�}zd|d<|j|d<WYd}~n|d}~0t�y}zd|d<|j	|d
<WYd}~nFd}~0t
�yZ}z$d|d<|j�p@||d<WYd}~n
d}~00t|t|d�dd�dS)Nrz"Needs args: hook_name, control_dirro�zUnknown hook: %sz
input.jsonF)�unsupported�
return_valrr�T�
no_backendr �backend_invalid�
backend_errorr��hook_missing�missing_hook_namezoutput.json)�indent)rtrD�argv�exit�
HOOK_NAMES�globalsrrrrr r)r*r}r+r-r)r-Zcontrol_dirrY�
hook_inputZjson_out�errr�mainLs0




&r��__main__)N)N)(r(r�	importlibrrr1�os.pathrrrrcrvrDr �version_inforrrerr)r+r7rMrSrUrZr[rarmrhr]rur\rNr{r|r}rr�r�r%rrrr�<module>sN



	

site-packages/pip/_vendor/pep517/in_process/__pycache__/__init__.cpython-39.pyc000064400000001672150564745010023312 0ustar00a

��Re3�@s`dZddlmZmZmZddlmZzddlm	Z	dd�Z
WneyZedd��Z
Yn0dS)z�This is a subpackage because the directory is on sys.path for _in_process.py

The subpackage should stay as empty as possible to avoid shadowing modules that
the backend might import.
�)�dirname�abspath�join)�contextmanagerNcCst�td�S�Nz_in_process.py)�	resources�path�__package__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pep517/in_process/__init__.py�_in_proc_script_pathsrccstttt��d�VdSr)�pjoinrr�__file__r
r
r
rrs)�__doc__�os.pathrrrr
�
contextlibr�importlib.resourcesrr�ImportErrorr
r
r
r�<module>ssite-packages/pip/_vendor/pep517/in_process/_in_process.py000064400000025701150564745010017566 0ustar00"""This is invoked in a subprocess to call the build backend hooks.

It expects:
- Command line args: hook_name, control_dir
- Environment variables:
      PEP517_BUILD_BACKEND=entry.point:spec
      PEP517_BACKEND_PATH=paths (separated with os.pathsep)
- control_dir/input.json:
  - {"kwargs": {...}}

Results:
- control_dir/output.json
  - {"return_val": ...}
"""
from glob import glob
from importlib import import_module
import json
import os
import os.path
from os.path import join as pjoin
import re
import shutil
import sys
import traceback

# This file is run as a script, and `import compat` is not zip-safe, so we
# include write_json() and read_json() from compat.py.
#
# Handle reading and writing JSON in UTF-8, on Python 3 and 2.

if sys.version_info[0] >= 3:
    # Python 3
    def write_json(obj, path, **kwargs):
        with open(path, 'w', encoding='utf-8') as f:
            json.dump(obj, f, **kwargs)

    def read_json(path):
        with open(path, 'r', encoding='utf-8') as f:
            return json.load(f)

else:
    # Python 2
    def write_json(obj, path, **kwargs):
        with open(path, 'wb') as f:
            json.dump(obj, f, encoding='utf-8', **kwargs)

    def read_json(path):
        with open(path, 'rb') as f:
            return json.load(f)


class BackendUnavailable(Exception):
    """Raised if we cannot import the backend"""
    def __init__(self, traceback):
        self.traceback = traceback


class BackendInvalid(Exception):
    """Raised if the backend is invalid"""
    def __init__(self, message):
        self.message = message


class HookMissing(Exception):
    """Raised if a hook is missing and we are not executing the fallback"""
    def __init__(self, hook_name=None):
        super(HookMissing, self).__init__(hook_name)
        self.hook_name = hook_name


def contained_in(filename, directory):
    """Test if a file is located within the given directory."""
    filename = os.path.normcase(os.path.abspath(filename))
    directory = os.path.normcase(os.path.abspath(directory))
    return os.path.commonprefix([filename, directory]) == directory


def _build_backend():
    """Find and load the build backend"""
    # Add in-tree backend directories to the front of sys.path.
    backend_path = os.environ.get('PEP517_BACKEND_PATH')
    if backend_path:
        extra_pathitems = backend_path.split(os.pathsep)
        sys.path[:0] = extra_pathitems

    ep = os.environ['PEP517_BUILD_BACKEND']
    mod_path, _, obj_path = ep.partition(':')
    try:
        obj = import_module(mod_path)
    except ImportError:
        raise BackendUnavailable(traceback.format_exc())

    if backend_path:
        if not any(
            contained_in(obj.__file__, path)
            for path in extra_pathitems
        ):
            raise BackendInvalid("Backend was not loaded from backend-path")

    if obj_path:
        for path_part in obj_path.split('.'):
            obj = getattr(obj, path_part)
    return obj


def _supported_features():
    """Return the list of options features supported by the backend.

    Returns a list of strings.
    The only possible value is 'build_editable'.
    """
    backend = _build_backend()
    features = []
    if hasattr(backend, "build_editable"):
        features.append("build_editable")
    return features


def get_requires_for_build_wheel(config_settings):
    """Invoke the optional get_requires_for_build_wheel hook

    Returns [] if the hook is not defined.
    """
    backend = _build_backend()
    try:
        hook = backend.get_requires_for_build_wheel
    except AttributeError:
        return []
    else:
        return hook(config_settings)


def get_requires_for_build_editable(config_settings):
    """Invoke the optional get_requires_for_build_editable hook

    Returns [] if the hook is not defined.
    """
    backend = _build_backend()
    try:
        hook = backend.get_requires_for_build_editable
    except AttributeError:
        return []
    else:
        return hook(config_settings)


def prepare_metadata_for_build_wheel(
        metadata_directory, config_settings, _allow_fallback):
    """Invoke optional prepare_metadata_for_build_wheel

    Implements a fallback by building a wheel if the hook isn't defined,
    unless _allow_fallback is False in which case HookMissing is raised.
    """
    backend = _build_backend()
    try:
        hook = backend.prepare_metadata_for_build_wheel
    except AttributeError:
        if not _allow_fallback:
            raise HookMissing()
        whl_basename = backend.build_wheel(metadata_directory, config_settings)
        return _get_wheel_metadata_from_wheel(whl_basename, metadata_directory,
                                              config_settings)
    else:
        return hook(metadata_directory, config_settings)


def prepare_metadata_for_build_editable(
        metadata_directory, config_settings, _allow_fallback):
    """Invoke optional prepare_metadata_for_build_editable

    Implements a fallback by building an editable wheel if the hook isn't
    defined, unless _allow_fallback is False in which case HookMissing is
    raised.
    """
    backend = _build_backend()
    try:
        hook = backend.prepare_metadata_for_build_editable
    except AttributeError:
        if not _allow_fallback:
            raise HookMissing()
        try:
            build_hook = backend.build_editable
        except AttributeError:
            raise HookMissing(hook_name='build_editable')
        else:
            whl_basename = build_hook(metadata_directory, config_settings)
            return _get_wheel_metadata_from_wheel(whl_basename,
                                                  metadata_directory,
                                                  config_settings)
    else:
        return hook(metadata_directory, config_settings)


WHEEL_BUILT_MARKER = 'PEP517_ALREADY_BUILT_WHEEL'


def _dist_info_files(whl_zip):
    """Identify the .dist-info folder inside a wheel ZipFile."""
    res = []
    for path in whl_zip.namelist():
        m = re.match(r'[^/\\]+-[^/\\]+\.dist-info/', path)
        if m:
            res.append(path)
    if res:
        return res
    raise Exception("No .dist-info folder found in wheel")


def _get_wheel_metadata_from_wheel(
        whl_basename, metadata_directory, config_settings):
    """Extract the metadata from a wheel.

    Fallback for when the build backend does not
    define the 'get_wheel_metadata' hook.
    """
    from zipfile import ZipFile
    with open(os.path.join(metadata_directory, WHEEL_BUILT_MARKER), 'wb'):
        pass  # Touch marker file

    whl_file = os.path.join(metadata_directory, whl_basename)
    with ZipFile(whl_file) as zipf:
        dist_info = _dist_info_files(zipf)
        zipf.extractall(path=metadata_directory, members=dist_info)
    return dist_info[0].split('/')[0]


def _find_already_built_wheel(metadata_directory):
    """Check for a wheel already built during the get_wheel_metadata hook.
    """
    if not metadata_directory:
        return None
    metadata_parent = os.path.dirname(metadata_directory)
    if not os.path.isfile(pjoin(metadata_parent, WHEEL_BUILT_MARKER)):
        return None

    whl_files = glob(os.path.join(metadata_parent, '*.whl'))
    if not whl_files:
        print('Found wheel built marker, but no .whl files')
        return None
    if len(whl_files) > 1:
        print('Found multiple .whl files; unspecified behaviour. '
              'Will call build_wheel.')
        return None

    # Exactly one .whl file
    return whl_files[0]


def build_wheel(wheel_directory, config_settings, metadata_directory=None):
    """Invoke the mandatory build_wheel hook.

    If a wheel was already built in the
    prepare_metadata_for_build_wheel fallback, this
    will copy it rather than rebuilding the wheel.
    """
    prebuilt_whl = _find_already_built_wheel(metadata_directory)
    if prebuilt_whl:
        shutil.copy2(prebuilt_whl, wheel_directory)
        return os.path.basename(prebuilt_whl)

    return _build_backend().build_wheel(wheel_directory, config_settings,
                                        metadata_directory)


def build_editable(wheel_directory, config_settings, metadata_directory=None):
    """Invoke the optional build_editable hook.

    If a wheel was already built in the
    prepare_metadata_for_build_editable fallback, this
    will copy it rather than rebuilding the wheel.
    """
    backend = _build_backend()
    try:
        hook = backend.build_editable
    except AttributeError:
        raise HookMissing()
    else:
        prebuilt_whl = _find_already_built_wheel(metadata_directory)
        if prebuilt_whl:
            shutil.copy2(prebuilt_whl, wheel_directory)
            return os.path.basename(prebuilt_whl)

        return hook(wheel_directory, config_settings, metadata_directory)


def get_requires_for_build_sdist(config_settings):
    """Invoke the optional get_requires_for_build_wheel hook

    Returns [] if the hook is not defined.
    """
    backend = _build_backend()
    try:
        hook = backend.get_requires_for_build_sdist
    except AttributeError:
        return []
    else:
        return hook(config_settings)


class _DummyException(Exception):
    """Nothing should ever raise this exception"""


class GotUnsupportedOperation(Exception):
    """For internal use when backend raises UnsupportedOperation"""
    def __init__(self, traceback):
        self.traceback = traceback


def build_sdist(sdist_directory, config_settings):
    """Invoke the mandatory build_sdist hook."""
    backend = _build_backend()
    try:
        return backend.build_sdist(sdist_directory, config_settings)
    except getattr(backend, 'UnsupportedOperation', _DummyException):
        raise GotUnsupportedOperation(traceback.format_exc())


HOOK_NAMES = {
    'get_requires_for_build_wheel',
    'prepare_metadata_for_build_wheel',
    'build_wheel',
    'get_requires_for_build_editable',
    'prepare_metadata_for_build_editable',
    'build_editable',
    'get_requires_for_build_sdist',
    'build_sdist',
    '_supported_features',
}


def main():
    if len(sys.argv) < 3:
        sys.exit("Needs args: hook_name, control_dir")
    hook_name = sys.argv[1]
    control_dir = sys.argv[2]
    if hook_name not in HOOK_NAMES:
        sys.exit("Unknown hook: %s" % hook_name)
    hook = globals()[hook_name]

    hook_input = read_json(pjoin(control_dir, 'input.json'))

    json_out = {'unsupported': False, 'return_val': None}
    try:
        json_out['return_val'] = hook(**hook_input['kwargs'])
    except BackendUnavailable as e:
        json_out['no_backend'] = True
        json_out['traceback'] = e.traceback
    except BackendInvalid as e:
        json_out['backend_invalid'] = True
        json_out['backend_error'] = e.message
    except GotUnsupportedOperation as e:
        json_out['unsupported'] = True
        json_out['traceback'] = e.traceback
    except HookMissing as e:
        json_out['hook_missing'] = True
        json_out['missing_hook_name'] = e.hook_name or hook_name

    write_json(json_out, pjoin(control_dir, 'output.json'), indent=2)


if __name__ == '__main__':
    main()
site-packages/pip/_vendor/pep517/in_process/__init__.py000064400000001063150564745010017015 0ustar00"""This is a subpackage because the directory is on sys.path for _in_process.py

The subpackage should stay as empty as possible to avoid shadowing modules that
the backend might import.
"""
from os.path import dirname, abspath, join as pjoin
from contextlib import contextmanager

try:
    import importlib.resources as resources

    def _in_proc_script_path():
        return resources.path(__package__, '_in_process.py')
except ImportError:
    @contextmanager
    def _in_proc_script_path():
        yield pjoin(dirname(abspath(__file__)), '_in_process.py')
site-packages/pip/_vendor/pep517/colorlog.py000064400000010002150564745010014723 0ustar00"""Nicer log formatting with colours.

Code copied from Tornado, Apache licensed.
"""
# Copyright 2012 Facebook
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.

import logging
import sys

try:
    import curses
except ImportError:
    curses = None


def _stderr_supports_color():
    color = False
    if curses and hasattr(sys.stderr, 'isatty') and sys.stderr.isatty():
        try:
            curses.setupterm()
            if curses.tigetnum("colors") > 0:
                color = True
        except Exception:
            pass
    return color


class LogFormatter(logging.Formatter):
    """Log formatter with colour support
    """
    DEFAULT_COLORS = {
        logging.INFO: 2,  # Green
        logging.WARNING: 3,  # Yellow
        logging.ERROR: 1,  # Red
        logging.CRITICAL: 1,
    }

    def __init__(self, color=True, datefmt=None):
        r"""
        :arg bool color: Enables color support.
        :arg string fmt: Log message format.
        It will be applied to the attributes dict of log records. The
        text between ``%(color)s`` and ``%(end_color)s`` will be colored
        depending on the level if color support is on.
        :arg dict colors: color mappings from logging level to terminal color
        code
        :arg string datefmt: Datetime format.
        Used for formatting ``(asctime)`` placeholder in ``prefix_fmt``.
        .. versionchanged:: 3.2
        Added ``fmt`` and ``datefmt`` arguments.
        """
        logging.Formatter.__init__(self, datefmt=datefmt)
        self._colors = {}
        if color and _stderr_supports_color():
            # The curses module has some str/bytes confusion in
            # python3. Until version 3.2.3, most methods return
            # bytes, but only accept strings. In addition, we want to
            # output these strings with the logging module, which
            # works with unicode strings. The explicit calls to
            # unicode() below are harmless in python2 but will do the
            # right conversion in python 3.
            fg_color = (curses.tigetstr("setaf") or
                        curses.tigetstr("setf") or "")
            if (3, 0) < sys.version_info < (3, 2, 3):
                fg_color = str(fg_color, "ascii")

            for levelno, code in self.DEFAULT_COLORS.items():
                self._colors[levelno] = str(
                    curses.tparm(fg_color, code), "ascii")
            self._normal = str(curses.tigetstr("sgr0"), "ascii")

            scr = curses.initscr()
            self.termwidth = scr.getmaxyx()[1]
            curses.endwin()
        else:
            self._normal = ''
            # Default width is usually 80, but too wide is
            # worse than too narrow
            self.termwidth = 70

    def formatMessage(self, record):
        mlen = len(record.message)
        right_text = '{initial}-{name}'.format(initial=record.levelname[0],
                                               name=record.name)
        if mlen + len(right_text) < self.termwidth:
            space = ' ' * (self.termwidth - (mlen + len(right_text)))
        else:
            space = '  '

        if record.levelno in self._colors:
            start_color = self._colors[record.levelno]
            end_color = self._normal
        else:
            start_color = end_color = ''

        return record.message + space + start_color + right_text + end_color


def enable_colourful_output(level=logging.INFO):
    handler = logging.StreamHandler()
    handler.setFormatter(LogFormatter())
    logging.root.addHandler(handler)
    logging.root.setLevel(level)
site-packages/pip/_vendor/vendor.txt000064400000000660150564745010013557 0ustar00CacheControl==0.12.6  # Make sure to update the license in pyproject.toml for this.
colorama==0.4.4
distlib==0.3.3
distro==1.6.0
html5lib==1.1
msgpack==1.0.2
packaging==21.0
pep517==0.12.0
platformdirs==2.4.0
progress==1.6
pyparsing==2.4.7
requests==2.26.0
    certifi==2021.05.30
    chardet==4.0.0
    idna==3.2
    urllib3==1.26.7
resolvelib==0.8.0
setuptools==44.0.0
six==1.16.0
tenacity==8.0.1
tomli==1.0.3
webencodings==0.5.1
site-packages/pip/_vendor/__init__.py000064400000011144150564745010013631 0ustar00"""
pip._vendor is for vendoring dependencies of pip to prevent needing pip to
depend on something external.

Files inside of pip._vendor should be considered immutable and should only be
updated to versions from upstream.
"""
from __future__ import absolute_import

import glob
import os.path
import sys

# Downstream redistributors which have debundled our dependencies should also
# patch this value to be true. This will trigger the additional patching
# to cause things like "six" to be available as pip.
DEBUNDLED = False

# By default, look in this directory for a bunch of .whl files which we will
# add to the beginning of sys.path before attempting to import anything. This
# is done to support downstream re-distributors like Debian and Fedora who
# wish to create their own Wheels for our dependencies to aid in debundling.
WHEEL_DIR = os.path.abspath(os.path.dirname(__file__))


# Define a small helper function to alias our vendored modules to the real ones
# if the vendored ones do not exist. This idea of this was taken from
# https://github.com/kennethreitz/requests/pull/2567.
def vendored(modulename):
    vendored_name = "{0}.{1}".format(__name__, modulename)

    try:
        __import__(modulename, globals(), locals(), level=0)
    except ImportError:
        # We can just silently allow import failures to pass here. If we
        # got to this point it means that ``import pip._vendor.whatever``
        # failed and so did ``import whatever``. Since we're importing this
        # upfront in an attempt to alias imports, not erroring here will
        # just mean we get a regular import error whenever pip *actually*
        # tries to import one of these modules to use it, which actually
        # gives us a better error message than we would have otherwise
        # gotten.
        pass
    else:
        sys.modules[vendored_name] = sys.modules[modulename]
        base, head = vendored_name.rsplit(".", 1)
        setattr(sys.modules[base], head, sys.modules[modulename])


# If we're operating in a debundled setup, then we want to go ahead and trigger
# the aliasing of our vendored libraries as well as looking for wheels to add
# to our sys.path. This will cause all of this code to be a no-op typically
# however downstream redistributors can enable it in a consistent way across
# all platforms.
if DEBUNDLED:
    # Actually look inside of WHEEL_DIR to find .whl files and add them to the
    # front of our sys.path.
    sys.path[:] = glob.glob(os.path.join(WHEEL_DIR, "*.whl")) + sys.path

    # Actually alias all of our vendored dependencies.
    vendored("cachecontrol")
    vendored("certifi")
    vendored("colorama")
    vendored("distlib")
    vendored("distro")
    vendored("html5lib")
    vendored("six")
    vendored("six.moves")
    vendored("six.moves.urllib")
    vendored("six.moves.urllib.parse")
    vendored("packaging")
    vendored("packaging.version")
    vendored("packaging.specifiers")
    vendored("pep517")
    vendored("pkg_resources")
    vendored("platformdirs")
    vendored("progress")
    vendored("requests")
    vendored("requests.exceptions")
    vendored("requests.packages")
    vendored("requests.packages.urllib3")
    vendored("requests.packages.urllib3._collections")
    vendored("requests.packages.urllib3.connection")
    vendored("requests.packages.urllib3.connectionpool")
    vendored("requests.packages.urllib3.contrib")
    vendored("requests.packages.urllib3.contrib.ntlmpool")
    vendored("requests.packages.urllib3.contrib.pyopenssl")
    vendored("requests.packages.urllib3.exceptions")
    vendored("requests.packages.urllib3.fields")
    vendored("requests.packages.urllib3.filepost")
    vendored("requests.packages.urllib3.packages")
    vendored("requests.packages.urllib3.packages.ordered_dict")
    vendored("requests.packages.urllib3.packages.six")
    vendored("requests.packages.urllib3.packages.ssl_match_hostname")
    vendored("requests.packages.urllib3.packages.ssl_match_hostname."
             "_implementation")
    vendored("requests.packages.urllib3.poolmanager")
    vendored("requests.packages.urllib3.request")
    vendored("requests.packages.urllib3.response")
    vendored("requests.packages.urllib3.util")
    vendored("requests.packages.urllib3.util.connection")
    vendored("requests.packages.urllib3.util.request")
    vendored("requests.packages.urllib3.util.response")
    vendored("requests.packages.urllib3.util.retry")
    vendored("requests.packages.urllib3.util.ssl_")
    vendored("requests.packages.urllib3.util.timeout")
    vendored("requests.packages.urllib3.util.url")
    vendored("resolvelib")
    vendored("tenacity")
    vendored("tomli")
    vendored("urllib3")
site-packages/pip/_vendor/chardet/sbcsgroupprober.py000064400000010325150564745010016725 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetgroupprober import CharSetGroupProber
from .hebrewprober import HebrewProber
from .langbulgarianmodel import (ISO_8859_5_BULGARIAN_MODEL,
                                 WINDOWS_1251_BULGARIAN_MODEL)
from .langgreekmodel import ISO_8859_7_GREEK_MODEL, WINDOWS_1253_GREEK_MODEL
from .langhebrewmodel import WINDOWS_1255_HEBREW_MODEL
# from .langhungarianmodel import (ISO_8859_2_HUNGARIAN_MODEL,
#                                  WINDOWS_1250_HUNGARIAN_MODEL)
from .langrussianmodel import (IBM855_RUSSIAN_MODEL, IBM866_RUSSIAN_MODEL,
                               ISO_8859_5_RUSSIAN_MODEL, KOI8_R_RUSSIAN_MODEL,
                               MACCYRILLIC_RUSSIAN_MODEL,
                               WINDOWS_1251_RUSSIAN_MODEL)
from .langthaimodel import TIS_620_THAI_MODEL
from .langturkishmodel import ISO_8859_9_TURKISH_MODEL
from .sbcharsetprober import SingleByteCharSetProber


class SBCSGroupProber(CharSetGroupProber):
    def __init__(self):
        super(SBCSGroupProber, self).__init__()
        hebrew_prober = HebrewProber()
        logical_hebrew_prober = SingleByteCharSetProber(WINDOWS_1255_HEBREW_MODEL,
                                                        False, hebrew_prober)
        # TODO: See if using ISO-8859-8 Hebrew model works better here, since
        #       it's actually the visual one
        visual_hebrew_prober = SingleByteCharSetProber(WINDOWS_1255_HEBREW_MODEL,
                                                       True, hebrew_prober)
        hebrew_prober.set_model_probers(logical_hebrew_prober,
                                        visual_hebrew_prober)
        # TODO: ORDER MATTERS HERE. I changed the order vs what was in master
        #       and several tests failed that did not before. Some thought
        #       should be put into the ordering, and we should consider making
        #       order not matter here, because that is very counter-intuitive.
        self.probers = [
            SingleByteCharSetProber(WINDOWS_1251_RUSSIAN_MODEL),
            SingleByteCharSetProber(KOI8_R_RUSSIAN_MODEL),
            SingleByteCharSetProber(ISO_8859_5_RUSSIAN_MODEL),
            SingleByteCharSetProber(MACCYRILLIC_RUSSIAN_MODEL),
            SingleByteCharSetProber(IBM866_RUSSIAN_MODEL),
            SingleByteCharSetProber(IBM855_RUSSIAN_MODEL),
            SingleByteCharSetProber(ISO_8859_7_GREEK_MODEL),
            SingleByteCharSetProber(WINDOWS_1253_GREEK_MODEL),
            SingleByteCharSetProber(ISO_8859_5_BULGARIAN_MODEL),
            SingleByteCharSetProber(WINDOWS_1251_BULGARIAN_MODEL),
            # TODO: Restore Hungarian encodings (iso-8859-2 and windows-1250)
            #       after we retrain model.
            # SingleByteCharSetProber(ISO_8859_2_HUNGARIAN_MODEL),
            # SingleByteCharSetProber(WINDOWS_1250_HUNGARIAN_MODEL),
            SingleByteCharSetProber(TIS_620_THAI_MODEL),
            SingleByteCharSetProber(ISO_8859_9_TURKISH_MODEL),
            hebrew_prober,
            logical_hebrew_prober,
            visual_hebrew_prober,
        ]
        self.reset()
site-packages/pip/_vendor/chardet/langrussianmodel.py000064400000400126150564745010017055 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

RUSSIAN_LANG_MODEL = {
    37: {  # 'А'
        37: 0,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 1,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 2,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 1,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 1,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 0,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 0,  # 'и'
        23: 1,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 0,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 2,  # 'ф'
        26: 2,  # 'х'
        28: 0,  # 'ц'
        22: 1,  # 'ч'
        25: 2,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    44: {  # 'Б'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 1,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 2,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    33: {  # 'В'
        37: 2,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 1,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 2,  # 'а'
        21: 1,  # 'б'
        10: 1,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 2,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 1,  # 'ъ'
        18: 3,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 0,  # 'ю'
        16: 1,  # 'я'
    },
    46: {  # 'Г'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 2,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 1,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    41: {  # 'Д'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 2,  # 'Е'
        56: 1,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 2,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 3,  # 'ж'
        20: 1,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    48: {  # 'Е'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 1,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 2,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 2,  # 'Р'
        32: 2,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 0,  # 'а'
        21: 0,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 2,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 0,  # 'и'
        23: 2,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 1,  # 'н'
        1: 0,  # 'о'
        15: 1,  # 'п'
        9: 1,  # 'р'
        7: 3,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 2,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    56: {  # 'Ж'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 1,  # 'б'
        10: 0,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 2,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 1,  # 'м'
        5: 0,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 2,  # 'ю'
        16: 0,  # 'я'
    },
    51: {  # 'З'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 0,  # 'г'
        13: 2,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 1,  # 'л'
        12: 1,  # 'м'
        5: 2,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 1,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 1,  # 'я'
    },
    42: {  # 'И'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 2,  # 'Е'
        56: 1,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 2,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 1,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 1,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 2,  # 'з'
        4: 1,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 1,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 1,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    60: {  # 'Й'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 1,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    36: {  # 'К'
        37: 2,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 2,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 1,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 0,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    49: {  # 'Л'
        37: 2,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 1,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 1,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 0,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 0,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 1,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 1,  # 'л'
        12: 0,  # 'м'
        5: 1,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 2,  # 'ю'
        16: 1,  # 'я'
    },
    38: {  # 'М'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 1,  # 'Ф'
        55: 1,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 0,  # 'Ь'
        47: 1,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 1,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 1,  # 'л'
        12: 1,  # 'м'
        5: 2,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 1,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    31: {  # 'Н'
        37: 2,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 1,  # 'З'
        42: 2,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 1,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 1,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 3,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    34: {  # 'О'
        37: 0,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 2,  # 'Д'
        48: 1,  # 'Е'
        56: 1,  # 'Ж'
        51: 1,  # 'З'
        42: 1,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 2,  # 'Л'
        38: 1,  # 'М'
        31: 2,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 2,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 1,  # 'Ф'
        55: 1,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 1,  # 'а'
        21: 2,  # 'б'
        10: 1,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 0,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 0,  # 'и'
        23: 1,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 0,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 1,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 2,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    35: {  # 'П'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 2,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 1,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 0,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 3,  # 'р'
        7: 1,  # 'с'
        6: 1,  # 'т'
        14: 2,  # 'у'
        39: 1,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 0,  # 'ю'
        16: 2,  # 'я'
    },
    45: {  # 'Р'
        37: 2,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 2,  # 'Е'
        56: 1,  # 'Ж'
        51: 0,  # 'З'
        42: 2,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 2,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 1,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 2,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 2,  # 'я'
    },
    32: {  # 'С'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 2,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 1,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 2,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 2,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 1,  # 'с'
        6: 3,  # 'т'
        14: 2,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 1,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 1,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    40: {  # 'Т'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 2,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 1,  # 'Ь'
        47: 1,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 1,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 1,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    52: {  # 'У'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 1,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 1,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 0,  # 'Я'
        3: 1,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 1,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 2,  # 'и'
        23: 1,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 1,  # 'н'
        1: 2,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 0,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    53: {  # 'Ф'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 1,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    55: {  # 'Х'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 2,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 0,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 1,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 1,  # 'ь'
        30: 1,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    58: {  # 'Ц'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 1,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 1,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 0,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 1,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    50: {  # 'Ч'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 1,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 1,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 1,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 1,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 3,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 1,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    57: {  # 'Ш'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 1,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 1,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 1,  # 'н'
        1: 2,  # 'о'
        15: 2,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    63: {  # 'Щ'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 1,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 1,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 1,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 1,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 1,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 1,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    62: {  # 'Ы'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 1,  # 'Ц'
        50: 0,  # 'Ч'
        57: 1,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 0,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 0,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    61: {  # 'Ь'
        37: 0,  # 'А'
        44: 1,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 1,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 1,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 1,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 1,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 0,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 0,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 0,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    47: {  # 'Э'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 1,  # 'Й'
        36: 1,  # 'К'
        49: 1,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 1,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 1,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 0,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 2,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 0,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 1,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    59: {  # 'Ю'
        37: 1,  # 'А'
        44: 1,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 1,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 0,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 1,  # 'б'
        10: 0,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 0,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 2,  # 'н'
        1: 0,  # 'о'
        15: 1,  # 'п'
        9: 1,  # 'р'
        7: 1,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    43: {  # 'Я'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 1,  # 'В'
        46: 1,  # 'Г'
        41: 0,  # 'Д'
        48: 1,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 1,  # 'С'
        40: 1,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 1,  # 'Х'
        58: 0,  # 'Ц'
        50: 1,  # 'Ч'
        57: 0,  # 'Ш'
        63: 1,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 1,  # 'Ю'
        43: 1,  # 'Я'
        3: 0,  # 'а'
        21: 1,  # 'б'
        10: 1,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 0,  # 'е'
        24: 0,  # 'ж'
        20: 1,  # 'з'
        4: 0,  # 'и'
        23: 1,  # 'й'
        11: 1,  # 'к'
        8: 1,  # 'л'
        12: 1,  # 'м'
        5: 2,  # 'н'
        1: 0,  # 'о'
        15: 1,  # 'п'
        9: 1,  # 'р'
        7: 1,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    3: {  # 'а'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 1,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 3,  # 'и'
        23: 3,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 3,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 2,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    21: {  # 'б'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 1,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 1,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 0,  # 'ф'
        26: 2,  # 'х'
        28: 1,  # 'ц'
        22: 1,  # 'ч'
        25: 2,  # 'ш'
        29: 3,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 2,  # 'ю'
        16: 3,  # 'я'
    },
    10: {  # 'в'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 3,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 3,  # 'ш'
        29: 2,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 3,  # 'я'
    },
    19: {  # 'г'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 3,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    13: {  # 'д'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 3,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 3,  # 'ц'
        22: 2,  # 'ч'
        25: 2,  # 'ш'
        29: 1,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 1,  # 'э'
        27: 2,  # 'ю'
        16: 3,  # 'я'
    },
    2: {  # 'е'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 2,  # 'и'
        23: 3,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 2,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 3,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 2,  # 'ю'
        16: 3,  # 'я'
    },
    24: {  # 'ж'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 1,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 1,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 0,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    20: {  # 'з'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 3,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 1,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 3,  # 'я'
    },
    4: {  # 'и'
        37: 1,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 3,  # 'и'
        23: 3,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 2,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 3,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 2,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    23: {  # 'й'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 1,  # 'а'
        21: 1,  # 'б'
        10: 1,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 2,  # 'з'
        4: 1,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 2,  # 'ф'
        26: 1,  # 'х'
        28: 2,  # 'ц'
        22: 3,  # 'ч'
        25: 2,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 2,  # 'я'
    },
    11: {  # 'к'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 3,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 2,  # 'ц'
        22: 1,  # 'ч'
        25: 2,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 1,  # 'ы'
        17: 1,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    8: {  # 'л'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 3,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 1,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 2,  # 'х'
        28: 1,  # 'ц'
        22: 3,  # 'ч'
        25: 2,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 1,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    12: {  # 'м'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 1,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 2,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 2,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 3,  # 'я'
    },
    5: {  # 'н'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 2,  # 'х'
        28: 3,  # 'ц'
        22: 3,  # 'ч'
        25: 2,  # 'ш'
        29: 2,  # 'щ'
        54: 1,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 1,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    1: {  # 'о'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 3,  # 'и'
        23: 3,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 2,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 2,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    15: {  # 'п'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 3,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 3,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 0,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 1,  # 'ш'
        29: 1,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 2,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 3,  # 'я'
    },
    9: {  # 'р'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 2,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 3,  # 'ш'
        29: 2,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 2,  # 'э'
        27: 2,  # 'ю'
        16: 3,  # 'я'
    },
    7: {  # 'с'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 1,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 3,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 3,  # 'ч'
        25: 2,  # 'ш'
        29: 1,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 2,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    6: {  # 'т'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 2,  # 'б'
        10: 3,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 2,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 2,  # 'ш'
        29: 2,  # 'щ'
        54: 2,  # 'ъ'
        18: 3,  # 'ы'
        17: 3,  # 'ь'
        30: 2,  # 'э'
        27: 2,  # 'ю'
        16: 3,  # 'я'
    },
    14: {  # 'у'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 3,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 2,  # 'и'
        23: 2,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 2,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 2,  # 'э'
        27: 3,  # 'ю'
        16: 2,  # 'я'
    },
    39: {  # 'ф'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 0,  # 'в'
        19: 1,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 1,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 2,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 1,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 2,  # 'ы'
        17: 1,  # 'ь'
        30: 2,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    26: {  # 'х'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 3,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 1,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 1,  # 'п'
        9: 3,  # 'р'
        7: 2,  # 'с'
        6: 2,  # 'т'
        14: 2,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 1,  # 'ч'
        25: 2,  # 'ш'
        29: 0,  # 'щ'
        54: 1,  # 'ъ'
        18: 0,  # 'ы'
        17: 1,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    28: {  # 'ц'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 1,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 1,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 2,  # 'к'
        8: 1,  # 'л'
        12: 1,  # 'м'
        5: 1,  # 'н'
        1: 3,  # 'о'
        15: 0,  # 'п'
        9: 1,  # 'р'
        7: 0,  # 'с'
        6: 1,  # 'т'
        14: 3,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 1,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 3,  # 'ы'
        17: 1,  # 'ь'
        30: 0,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    22: {  # 'ч'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 2,  # 'л'
        12: 1,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 1,  # 'с'
        6: 3,  # 'т'
        14: 3,  # 'у'
        39: 1,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 1,  # 'ч'
        25: 2,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 3,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    25: {  # 'ш'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 1,  # 'б'
        10: 2,  # 'в'
        19: 1,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 2,  # 'м'
        5: 3,  # 'н'
        1: 3,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 1,  # 'с'
        6: 2,  # 'т'
        14: 3,  # 'у'
        39: 2,  # 'ф'
        26: 1,  # 'х'
        28: 1,  # 'ц'
        22: 1,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 3,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 0,  # 'я'
    },
    29: {  # 'щ'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 3,  # 'а'
        21: 0,  # 'б'
        10: 1,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 3,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 3,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 1,  # 'м'
        5: 2,  # 'н'
        1: 1,  # 'о'
        15: 0,  # 'п'
        9: 2,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 2,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 2,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 0,  # 'я'
    },
    54: {  # 'ъ'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 0,  # 'б'
        10: 0,  # 'в'
        19: 0,  # 'г'
        13: 0,  # 'д'
        2: 2,  # 'е'
        24: 0,  # 'ж'
        20: 0,  # 'з'
        4: 0,  # 'и'
        23: 0,  # 'й'
        11: 0,  # 'к'
        8: 0,  # 'л'
        12: 0,  # 'м'
        5: 0,  # 'н'
        1: 0,  # 'о'
        15: 0,  # 'п'
        9: 0,  # 'р'
        7: 0,  # 'с'
        6: 0,  # 'т'
        14: 0,  # 'у'
        39: 0,  # 'ф'
        26: 0,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 0,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 1,  # 'ю'
        16: 2,  # 'я'
    },
    18: {  # 'ы'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 3,  # 'б'
        10: 3,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 2,  # 'и'
        23: 3,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 1,  # 'о'
        15: 3,  # 'п'
        9: 3,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 0,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 3,  # 'ч'
        25: 3,  # 'ш'
        29: 2,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 0,  # 'ю'
        16: 2,  # 'я'
    },
    17: {  # 'ь'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 2,  # 'б'
        10: 2,  # 'в'
        19: 2,  # 'г'
        13: 2,  # 'д'
        2: 3,  # 'е'
        24: 1,  # 'ж'
        20: 3,  # 'з'
        4: 2,  # 'и'
        23: 0,  # 'й'
        11: 3,  # 'к'
        8: 0,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 2,  # 'о'
        15: 2,  # 'п'
        9: 1,  # 'р'
        7: 3,  # 'с'
        6: 2,  # 'т'
        14: 0,  # 'у'
        39: 2,  # 'ф'
        26: 1,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 3,  # 'ш'
        29: 2,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 3,  # 'ю'
        16: 3,  # 'я'
    },
    30: {  # 'э'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 1,  # 'М'
        31: 1,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 1,  # 'Р'
        32: 1,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 1,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 1,  # 'б'
        10: 1,  # 'в'
        19: 1,  # 'г'
        13: 2,  # 'д'
        2: 1,  # 'е'
        24: 0,  # 'ж'
        20: 1,  # 'з'
        4: 0,  # 'и'
        23: 2,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 0,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 2,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 2,  # 'ф'
        26: 1,  # 'х'
        28: 0,  # 'ц'
        22: 0,  # 'ч'
        25: 1,  # 'ш'
        29: 0,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 1,  # 'ю'
        16: 1,  # 'я'
    },
    27: {  # 'ю'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 2,  # 'а'
        21: 3,  # 'б'
        10: 1,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 1,  # 'е'
        24: 2,  # 'ж'
        20: 2,  # 'з'
        4: 1,  # 'и'
        23: 1,  # 'й'
        11: 2,  # 'к'
        8: 2,  # 'л'
        12: 2,  # 'м'
        5: 2,  # 'н'
        1: 1,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 0,  # 'у'
        39: 1,  # 'ф'
        26: 2,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 2,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 1,  # 'э'
        27: 2,  # 'ю'
        16: 1,  # 'я'
    },
    16: {  # 'я'
        37: 0,  # 'А'
        44: 0,  # 'Б'
        33: 0,  # 'В'
        46: 0,  # 'Г'
        41: 0,  # 'Д'
        48: 0,  # 'Е'
        56: 0,  # 'Ж'
        51: 0,  # 'З'
        42: 0,  # 'И'
        60: 0,  # 'Й'
        36: 0,  # 'К'
        49: 0,  # 'Л'
        38: 0,  # 'М'
        31: 0,  # 'Н'
        34: 0,  # 'О'
        35: 0,  # 'П'
        45: 0,  # 'Р'
        32: 0,  # 'С'
        40: 0,  # 'Т'
        52: 0,  # 'У'
        53: 0,  # 'Ф'
        55: 0,  # 'Х'
        58: 0,  # 'Ц'
        50: 0,  # 'Ч'
        57: 0,  # 'Ш'
        63: 0,  # 'Щ'
        62: 0,  # 'Ы'
        61: 0,  # 'Ь'
        47: 0,  # 'Э'
        59: 0,  # 'Ю'
        43: 0,  # 'Я'
        3: 0,  # 'а'
        21: 2,  # 'б'
        10: 3,  # 'в'
        19: 2,  # 'г'
        13: 3,  # 'д'
        2: 3,  # 'е'
        24: 3,  # 'ж'
        20: 3,  # 'з'
        4: 2,  # 'и'
        23: 2,  # 'й'
        11: 3,  # 'к'
        8: 3,  # 'л'
        12: 3,  # 'м'
        5: 3,  # 'н'
        1: 0,  # 'о'
        15: 2,  # 'п'
        9: 2,  # 'р'
        7: 3,  # 'с'
        6: 3,  # 'т'
        14: 1,  # 'у'
        39: 1,  # 'ф'
        26: 3,  # 'х'
        28: 2,  # 'ц'
        22: 2,  # 'ч'
        25: 2,  # 'ш'
        29: 3,  # 'щ'
        54: 0,  # 'ъ'
        18: 0,  # 'ы'
        17: 0,  # 'ь'
        30: 0,  # 'э'
        27: 2,  # 'ю'
        16: 2,  # 'я'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
IBM866_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 37,  # 'А'
     129: 44,  # 'Б'
     130: 33,  # 'В'
     131: 46,  # 'Г'
     132: 41,  # 'Д'
     133: 48,  # 'Е'
     134: 56,  # 'Ж'
     135: 51,  # 'З'
     136: 42,  # 'И'
     137: 60,  # 'Й'
     138: 36,  # 'К'
     139: 49,  # 'Л'
     140: 38,  # 'М'
     141: 31,  # 'Н'
     142: 34,  # 'О'
     143: 35,  # 'П'
     144: 45,  # 'Р'
     145: 32,  # 'С'
     146: 40,  # 'Т'
     147: 52,  # 'У'
     148: 53,  # 'Ф'
     149: 55,  # 'Х'
     150: 58,  # 'Ц'
     151: 50,  # 'Ч'
     152: 57,  # 'Ш'
     153: 63,  # 'Щ'
     154: 70,  # 'Ъ'
     155: 62,  # 'Ы'
     156: 61,  # 'Ь'
     157: 47,  # 'Э'
     158: 59,  # 'Ю'
     159: 43,  # 'Я'
     160: 3,  # 'а'
     161: 21,  # 'б'
     162: 10,  # 'в'
     163: 19,  # 'г'
     164: 13,  # 'д'
     165: 2,  # 'е'
     166: 24,  # 'ж'
     167: 20,  # 'з'
     168: 4,  # 'и'
     169: 23,  # 'й'
     170: 11,  # 'к'
     171: 8,  # 'л'
     172: 12,  # 'м'
     173: 5,  # 'н'
     174: 1,  # 'о'
     175: 15,  # 'п'
     176: 191,  # '░'
     177: 192,  # '▒'
     178: 193,  # '▓'
     179: 194,  # '│'
     180: 195,  # '┤'
     181: 196,  # '╡'
     182: 197,  # '╢'
     183: 198,  # '╖'
     184: 199,  # '╕'
     185: 200,  # '╣'
     186: 201,  # '║'
     187: 202,  # '╗'
     188: 203,  # '╝'
     189: 204,  # '╜'
     190: 205,  # '╛'
     191: 206,  # '┐'
     192: 207,  # '└'
     193: 208,  # '┴'
     194: 209,  # '┬'
     195: 210,  # '├'
     196: 211,  # '─'
     197: 212,  # '┼'
     198: 213,  # '╞'
     199: 214,  # '╟'
     200: 215,  # '╚'
     201: 216,  # '╔'
     202: 217,  # '╩'
     203: 218,  # '╦'
     204: 219,  # '╠'
     205: 220,  # '═'
     206: 221,  # '╬'
     207: 222,  # '╧'
     208: 223,  # '╨'
     209: 224,  # '╤'
     210: 225,  # '╥'
     211: 226,  # '╙'
     212: 227,  # '╘'
     213: 228,  # '╒'
     214: 229,  # '╓'
     215: 230,  # '╫'
     216: 231,  # '╪'
     217: 232,  # '┘'
     218: 233,  # '┌'
     219: 234,  # '█'
     220: 235,  # '▄'
     221: 236,  # '▌'
     222: 237,  # '▐'
     223: 238,  # '▀'
     224: 9,  # 'р'
     225: 7,  # 'с'
     226: 6,  # 'т'
     227: 14,  # 'у'
     228: 39,  # 'ф'
     229: 26,  # 'х'
     230: 28,  # 'ц'
     231: 22,  # 'ч'
     232: 25,  # 'ш'
     233: 29,  # 'щ'
     234: 54,  # 'ъ'
     235: 18,  # 'ы'
     236: 17,  # 'ь'
     237: 30,  # 'э'
     238: 27,  # 'ю'
     239: 16,  # 'я'
     240: 239,  # 'Ё'
     241: 68,  # 'ё'
     242: 240,  # 'Є'
     243: 241,  # 'є'
     244: 242,  # 'Ї'
     245: 243,  # 'ї'
     246: 244,  # 'Ў'
     247: 245,  # 'ў'
     248: 246,  # '°'
     249: 247,  # '∙'
     250: 248,  # '·'
     251: 249,  # '√'
     252: 250,  # '№'
     253: 251,  # '¤'
     254: 252,  # '■'
     255: 255,  # '\xa0'
}

IBM866_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='IBM866',
                                              language='Russian',
                                              char_to_order_map=IBM866_RUSSIAN_CHAR_TO_ORDER,
                                              language_model=RUSSIAN_LANG_MODEL,
                                              typical_positive_ratio=0.976601,
                                              keep_ascii_letters=False,
                                              alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

WINDOWS_1251_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 191,  # 'Ђ'
     129: 192,  # 'Ѓ'
     130: 193,  # '‚'
     131: 194,  # 'ѓ'
     132: 195,  # '„'
     133: 196,  # '…'
     134: 197,  # '†'
     135: 198,  # '‡'
     136: 199,  # '€'
     137: 200,  # '‰'
     138: 201,  # 'Љ'
     139: 202,  # '‹'
     140: 203,  # 'Њ'
     141: 204,  # 'Ќ'
     142: 205,  # 'Ћ'
     143: 206,  # 'Џ'
     144: 207,  # 'ђ'
     145: 208,  # '‘'
     146: 209,  # '’'
     147: 210,  # '“'
     148: 211,  # '”'
     149: 212,  # '•'
     150: 213,  # '–'
     151: 214,  # '—'
     152: 215,  # None
     153: 216,  # '™'
     154: 217,  # 'љ'
     155: 218,  # '›'
     156: 219,  # 'њ'
     157: 220,  # 'ќ'
     158: 221,  # 'ћ'
     159: 222,  # 'џ'
     160: 223,  # '\xa0'
     161: 224,  # 'Ў'
     162: 225,  # 'ў'
     163: 226,  # 'Ј'
     164: 227,  # '¤'
     165: 228,  # 'Ґ'
     166: 229,  # '¦'
     167: 230,  # '§'
     168: 231,  # 'Ё'
     169: 232,  # '©'
     170: 233,  # 'Є'
     171: 234,  # '«'
     172: 235,  # '¬'
     173: 236,  # '\xad'
     174: 237,  # '®'
     175: 238,  # 'Ї'
     176: 239,  # '°'
     177: 240,  # '±'
     178: 241,  # 'І'
     179: 242,  # 'і'
     180: 243,  # 'ґ'
     181: 244,  # 'µ'
     182: 245,  # '¶'
     183: 246,  # '·'
     184: 68,  # 'ё'
     185: 247,  # '№'
     186: 248,  # 'є'
     187: 249,  # '»'
     188: 250,  # 'ј'
     189: 251,  # 'Ѕ'
     190: 252,  # 'ѕ'
     191: 253,  # 'ї'
     192: 37,  # 'А'
     193: 44,  # 'Б'
     194: 33,  # 'В'
     195: 46,  # 'Г'
     196: 41,  # 'Д'
     197: 48,  # 'Е'
     198: 56,  # 'Ж'
     199: 51,  # 'З'
     200: 42,  # 'И'
     201: 60,  # 'Й'
     202: 36,  # 'К'
     203: 49,  # 'Л'
     204: 38,  # 'М'
     205: 31,  # 'Н'
     206: 34,  # 'О'
     207: 35,  # 'П'
     208: 45,  # 'Р'
     209: 32,  # 'С'
     210: 40,  # 'Т'
     211: 52,  # 'У'
     212: 53,  # 'Ф'
     213: 55,  # 'Х'
     214: 58,  # 'Ц'
     215: 50,  # 'Ч'
     216: 57,  # 'Ш'
     217: 63,  # 'Щ'
     218: 70,  # 'Ъ'
     219: 62,  # 'Ы'
     220: 61,  # 'Ь'
     221: 47,  # 'Э'
     222: 59,  # 'Ю'
     223: 43,  # 'Я'
     224: 3,  # 'а'
     225: 21,  # 'б'
     226: 10,  # 'в'
     227: 19,  # 'г'
     228: 13,  # 'д'
     229: 2,  # 'е'
     230: 24,  # 'ж'
     231: 20,  # 'з'
     232: 4,  # 'и'
     233: 23,  # 'й'
     234: 11,  # 'к'
     235: 8,  # 'л'
     236: 12,  # 'м'
     237: 5,  # 'н'
     238: 1,  # 'о'
     239: 15,  # 'п'
     240: 9,  # 'р'
     241: 7,  # 'с'
     242: 6,  # 'т'
     243: 14,  # 'у'
     244: 39,  # 'ф'
     245: 26,  # 'х'
     246: 28,  # 'ц'
     247: 22,  # 'ч'
     248: 25,  # 'ш'
     249: 29,  # 'щ'
     250: 54,  # 'ъ'
     251: 18,  # 'ы'
     252: 17,  # 'ь'
     253: 30,  # 'э'
     254: 27,  # 'ю'
     255: 16,  # 'я'
}

WINDOWS_1251_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='windows-1251',
                                                    language='Russian',
                                                    char_to_order_map=WINDOWS_1251_RUSSIAN_CHAR_TO_ORDER,
                                                    language_model=RUSSIAN_LANG_MODEL,
                                                    typical_positive_ratio=0.976601,
                                                    keep_ascii_letters=False,
                                                    alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

IBM855_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 191,  # 'ђ'
     129: 192,  # 'Ђ'
     130: 193,  # 'ѓ'
     131: 194,  # 'Ѓ'
     132: 68,  # 'ё'
     133: 195,  # 'Ё'
     134: 196,  # 'є'
     135: 197,  # 'Є'
     136: 198,  # 'ѕ'
     137: 199,  # 'Ѕ'
     138: 200,  # 'і'
     139: 201,  # 'І'
     140: 202,  # 'ї'
     141: 203,  # 'Ї'
     142: 204,  # 'ј'
     143: 205,  # 'Ј'
     144: 206,  # 'љ'
     145: 207,  # 'Љ'
     146: 208,  # 'њ'
     147: 209,  # 'Њ'
     148: 210,  # 'ћ'
     149: 211,  # 'Ћ'
     150: 212,  # 'ќ'
     151: 213,  # 'Ќ'
     152: 214,  # 'ў'
     153: 215,  # 'Ў'
     154: 216,  # 'џ'
     155: 217,  # 'Џ'
     156: 27,  # 'ю'
     157: 59,  # 'Ю'
     158: 54,  # 'ъ'
     159: 70,  # 'Ъ'
     160: 3,  # 'а'
     161: 37,  # 'А'
     162: 21,  # 'б'
     163: 44,  # 'Б'
     164: 28,  # 'ц'
     165: 58,  # 'Ц'
     166: 13,  # 'д'
     167: 41,  # 'Д'
     168: 2,  # 'е'
     169: 48,  # 'Е'
     170: 39,  # 'ф'
     171: 53,  # 'Ф'
     172: 19,  # 'г'
     173: 46,  # 'Г'
     174: 218,  # '«'
     175: 219,  # '»'
     176: 220,  # '░'
     177: 221,  # '▒'
     178: 222,  # '▓'
     179: 223,  # '│'
     180: 224,  # '┤'
     181: 26,  # 'х'
     182: 55,  # 'Х'
     183: 4,  # 'и'
     184: 42,  # 'И'
     185: 225,  # '╣'
     186: 226,  # '║'
     187: 227,  # '╗'
     188: 228,  # '╝'
     189: 23,  # 'й'
     190: 60,  # 'Й'
     191: 229,  # '┐'
     192: 230,  # '└'
     193: 231,  # '┴'
     194: 232,  # '┬'
     195: 233,  # '├'
     196: 234,  # '─'
     197: 235,  # '┼'
     198: 11,  # 'к'
     199: 36,  # 'К'
     200: 236,  # '╚'
     201: 237,  # '╔'
     202: 238,  # '╩'
     203: 239,  # '╦'
     204: 240,  # '╠'
     205: 241,  # '═'
     206: 242,  # '╬'
     207: 243,  # '¤'
     208: 8,  # 'л'
     209: 49,  # 'Л'
     210: 12,  # 'м'
     211: 38,  # 'М'
     212: 5,  # 'н'
     213: 31,  # 'Н'
     214: 1,  # 'о'
     215: 34,  # 'О'
     216: 15,  # 'п'
     217: 244,  # '┘'
     218: 245,  # '┌'
     219: 246,  # '█'
     220: 247,  # '▄'
     221: 35,  # 'П'
     222: 16,  # 'я'
     223: 248,  # '▀'
     224: 43,  # 'Я'
     225: 9,  # 'р'
     226: 45,  # 'Р'
     227: 7,  # 'с'
     228: 32,  # 'С'
     229: 6,  # 'т'
     230: 40,  # 'Т'
     231: 14,  # 'у'
     232: 52,  # 'У'
     233: 24,  # 'ж'
     234: 56,  # 'Ж'
     235: 10,  # 'в'
     236: 33,  # 'В'
     237: 17,  # 'ь'
     238: 61,  # 'Ь'
     239: 249,  # '№'
     240: 250,  # '\xad'
     241: 18,  # 'ы'
     242: 62,  # 'Ы'
     243: 20,  # 'з'
     244: 51,  # 'З'
     245: 25,  # 'ш'
     246: 57,  # 'Ш'
     247: 30,  # 'э'
     248: 47,  # 'Э'
     249: 29,  # 'щ'
     250: 63,  # 'Щ'
     251: 22,  # 'ч'
     252: 50,  # 'Ч'
     253: 251,  # '§'
     254: 252,  # '■'
     255: 255,  # '\xa0'
}

IBM855_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='IBM855',
                                              language='Russian',
                                              char_to_order_map=IBM855_RUSSIAN_CHAR_TO_ORDER,
                                              language_model=RUSSIAN_LANG_MODEL,
                                              typical_positive_ratio=0.976601,
                                              keep_ascii_letters=False,
                                              alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

KOI8_R_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 191,  # '─'
     129: 192,  # '│'
     130: 193,  # '┌'
     131: 194,  # '┐'
     132: 195,  # '└'
     133: 196,  # '┘'
     134: 197,  # '├'
     135: 198,  # '┤'
     136: 199,  # '┬'
     137: 200,  # '┴'
     138: 201,  # '┼'
     139: 202,  # '▀'
     140: 203,  # '▄'
     141: 204,  # '█'
     142: 205,  # '▌'
     143: 206,  # '▐'
     144: 207,  # '░'
     145: 208,  # '▒'
     146: 209,  # '▓'
     147: 210,  # '⌠'
     148: 211,  # '■'
     149: 212,  # '∙'
     150: 213,  # '√'
     151: 214,  # '≈'
     152: 215,  # '≤'
     153: 216,  # '≥'
     154: 217,  # '\xa0'
     155: 218,  # '⌡'
     156: 219,  # '°'
     157: 220,  # '²'
     158: 221,  # '·'
     159: 222,  # '÷'
     160: 223,  # '═'
     161: 224,  # '║'
     162: 225,  # '╒'
     163: 68,  # 'ё'
     164: 226,  # '╓'
     165: 227,  # '╔'
     166: 228,  # '╕'
     167: 229,  # '╖'
     168: 230,  # '╗'
     169: 231,  # '╘'
     170: 232,  # '╙'
     171: 233,  # '╚'
     172: 234,  # '╛'
     173: 235,  # '╜'
     174: 236,  # '╝'
     175: 237,  # '╞'
     176: 238,  # '╟'
     177: 239,  # '╠'
     178: 240,  # '╡'
     179: 241,  # 'Ё'
     180: 242,  # '╢'
     181: 243,  # '╣'
     182: 244,  # '╤'
     183: 245,  # '╥'
     184: 246,  # '╦'
     185: 247,  # '╧'
     186: 248,  # '╨'
     187: 249,  # '╩'
     188: 250,  # '╪'
     189: 251,  # '╫'
     190: 252,  # '╬'
     191: 253,  # '©'
     192: 27,  # 'ю'
     193: 3,  # 'а'
     194: 21,  # 'б'
     195: 28,  # 'ц'
     196: 13,  # 'д'
     197: 2,  # 'е'
     198: 39,  # 'ф'
     199: 19,  # 'г'
     200: 26,  # 'х'
     201: 4,  # 'и'
     202: 23,  # 'й'
     203: 11,  # 'к'
     204: 8,  # 'л'
     205: 12,  # 'м'
     206: 5,  # 'н'
     207: 1,  # 'о'
     208: 15,  # 'п'
     209: 16,  # 'я'
     210: 9,  # 'р'
     211: 7,  # 'с'
     212: 6,  # 'т'
     213: 14,  # 'у'
     214: 24,  # 'ж'
     215: 10,  # 'в'
     216: 17,  # 'ь'
     217: 18,  # 'ы'
     218: 20,  # 'з'
     219: 25,  # 'ш'
     220: 30,  # 'э'
     221: 29,  # 'щ'
     222: 22,  # 'ч'
     223: 54,  # 'ъ'
     224: 59,  # 'Ю'
     225: 37,  # 'А'
     226: 44,  # 'Б'
     227: 58,  # 'Ц'
     228: 41,  # 'Д'
     229: 48,  # 'Е'
     230: 53,  # 'Ф'
     231: 46,  # 'Г'
     232: 55,  # 'Х'
     233: 42,  # 'И'
     234: 60,  # 'Й'
     235: 36,  # 'К'
     236: 49,  # 'Л'
     237: 38,  # 'М'
     238: 31,  # 'Н'
     239: 34,  # 'О'
     240: 35,  # 'П'
     241: 43,  # 'Я'
     242: 45,  # 'Р'
     243: 32,  # 'С'
     244: 40,  # 'Т'
     245: 52,  # 'У'
     246: 56,  # 'Ж'
     247: 33,  # 'В'
     248: 61,  # 'Ь'
     249: 62,  # 'Ы'
     250: 51,  # 'З'
     251: 57,  # 'Ш'
     252: 47,  # 'Э'
     253: 63,  # 'Щ'
     254: 50,  # 'Ч'
     255: 70,  # 'Ъ'
}

KOI8_R_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='KOI8-R',
                                              language='Russian',
                                              char_to_order_map=KOI8_R_RUSSIAN_CHAR_TO_ORDER,
                                              language_model=RUSSIAN_LANG_MODEL,
                                              typical_positive_ratio=0.976601,
                                              keep_ascii_letters=False,
                                              alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

MACCYRILLIC_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 37,  # 'А'
     129: 44,  # 'Б'
     130: 33,  # 'В'
     131: 46,  # 'Г'
     132: 41,  # 'Д'
     133: 48,  # 'Е'
     134: 56,  # 'Ж'
     135: 51,  # 'З'
     136: 42,  # 'И'
     137: 60,  # 'Й'
     138: 36,  # 'К'
     139: 49,  # 'Л'
     140: 38,  # 'М'
     141: 31,  # 'Н'
     142: 34,  # 'О'
     143: 35,  # 'П'
     144: 45,  # 'Р'
     145: 32,  # 'С'
     146: 40,  # 'Т'
     147: 52,  # 'У'
     148: 53,  # 'Ф'
     149: 55,  # 'Х'
     150: 58,  # 'Ц'
     151: 50,  # 'Ч'
     152: 57,  # 'Ш'
     153: 63,  # 'Щ'
     154: 70,  # 'Ъ'
     155: 62,  # 'Ы'
     156: 61,  # 'Ь'
     157: 47,  # 'Э'
     158: 59,  # 'Ю'
     159: 43,  # 'Я'
     160: 191,  # '†'
     161: 192,  # '°'
     162: 193,  # 'Ґ'
     163: 194,  # '£'
     164: 195,  # '§'
     165: 196,  # '•'
     166: 197,  # '¶'
     167: 198,  # 'І'
     168: 199,  # '®'
     169: 200,  # '©'
     170: 201,  # '™'
     171: 202,  # 'Ђ'
     172: 203,  # 'ђ'
     173: 204,  # '≠'
     174: 205,  # 'Ѓ'
     175: 206,  # 'ѓ'
     176: 207,  # '∞'
     177: 208,  # '±'
     178: 209,  # '≤'
     179: 210,  # '≥'
     180: 211,  # 'і'
     181: 212,  # 'µ'
     182: 213,  # 'ґ'
     183: 214,  # 'Ј'
     184: 215,  # 'Є'
     185: 216,  # 'є'
     186: 217,  # 'Ї'
     187: 218,  # 'ї'
     188: 219,  # 'Љ'
     189: 220,  # 'љ'
     190: 221,  # 'Њ'
     191: 222,  # 'њ'
     192: 223,  # 'ј'
     193: 224,  # 'Ѕ'
     194: 225,  # '¬'
     195: 226,  # '√'
     196: 227,  # 'ƒ'
     197: 228,  # '≈'
     198: 229,  # '∆'
     199: 230,  # '«'
     200: 231,  # '»'
     201: 232,  # '…'
     202: 233,  # '\xa0'
     203: 234,  # 'Ћ'
     204: 235,  # 'ћ'
     205: 236,  # 'Ќ'
     206: 237,  # 'ќ'
     207: 238,  # 'ѕ'
     208: 239,  # '–'
     209: 240,  # '—'
     210: 241,  # '“'
     211: 242,  # '”'
     212: 243,  # '‘'
     213: 244,  # '’'
     214: 245,  # '÷'
     215: 246,  # '„'
     216: 247,  # 'Ў'
     217: 248,  # 'ў'
     218: 249,  # 'Џ'
     219: 250,  # 'џ'
     220: 251,  # '№'
     221: 252,  # 'Ё'
     222: 68,  # 'ё'
     223: 16,  # 'я'
     224: 3,  # 'а'
     225: 21,  # 'б'
     226: 10,  # 'в'
     227: 19,  # 'г'
     228: 13,  # 'д'
     229: 2,  # 'е'
     230: 24,  # 'ж'
     231: 20,  # 'з'
     232: 4,  # 'и'
     233: 23,  # 'й'
     234: 11,  # 'к'
     235: 8,  # 'л'
     236: 12,  # 'м'
     237: 5,  # 'н'
     238: 1,  # 'о'
     239: 15,  # 'п'
     240: 9,  # 'р'
     241: 7,  # 'с'
     242: 6,  # 'т'
     243: 14,  # 'у'
     244: 39,  # 'ф'
     245: 26,  # 'х'
     246: 28,  # 'ц'
     247: 22,  # 'ч'
     248: 25,  # 'ш'
     249: 29,  # 'щ'
     250: 54,  # 'ъ'
     251: 18,  # 'ы'
     252: 17,  # 'ь'
     253: 30,  # 'э'
     254: 27,  # 'ю'
     255: 255,  # '€'
}

MACCYRILLIC_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='MacCyrillic',
                                                   language='Russian',
                                                   char_to_order_map=MACCYRILLIC_RUSSIAN_CHAR_TO_ORDER,
                                                   language_model=RUSSIAN_LANG_MODEL,
                                                   typical_positive_ratio=0.976601,
                                                   keep_ascii_letters=False,
                                                   alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

ISO_8859_5_RUSSIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 142,  # 'A'
     66: 143,  # 'B'
     67: 144,  # 'C'
     68: 145,  # 'D'
     69: 146,  # 'E'
     70: 147,  # 'F'
     71: 148,  # 'G'
     72: 149,  # 'H'
     73: 150,  # 'I'
     74: 151,  # 'J'
     75: 152,  # 'K'
     76: 74,  # 'L'
     77: 153,  # 'M'
     78: 75,  # 'N'
     79: 154,  # 'O'
     80: 155,  # 'P'
     81: 156,  # 'Q'
     82: 157,  # 'R'
     83: 158,  # 'S'
     84: 159,  # 'T'
     85: 160,  # 'U'
     86: 161,  # 'V'
     87: 162,  # 'W'
     88: 163,  # 'X'
     89: 164,  # 'Y'
     90: 165,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 71,  # 'a'
     98: 172,  # 'b'
     99: 66,  # 'c'
     100: 173,  # 'd'
     101: 65,  # 'e'
     102: 174,  # 'f'
     103: 76,  # 'g'
     104: 175,  # 'h'
     105: 64,  # 'i'
     106: 176,  # 'j'
     107: 177,  # 'k'
     108: 77,  # 'l'
     109: 72,  # 'm'
     110: 178,  # 'n'
     111: 69,  # 'o'
     112: 67,  # 'p'
     113: 179,  # 'q'
     114: 78,  # 'r'
     115: 73,  # 's'
     116: 180,  # 't'
     117: 181,  # 'u'
     118: 79,  # 'v'
     119: 182,  # 'w'
     120: 183,  # 'x'
     121: 184,  # 'y'
     122: 185,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 191,  # '\x80'
     129: 192,  # '\x81'
     130: 193,  # '\x82'
     131: 194,  # '\x83'
     132: 195,  # '\x84'
     133: 196,  # '\x85'
     134: 197,  # '\x86'
     135: 198,  # '\x87'
     136: 199,  # '\x88'
     137: 200,  # '\x89'
     138: 201,  # '\x8a'
     139: 202,  # '\x8b'
     140: 203,  # '\x8c'
     141: 204,  # '\x8d'
     142: 205,  # '\x8e'
     143: 206,  # '\x8f'
     144: 207,  # '\x90'
     145: 208,  # '\x91'
     146: 209,  # '\x92'
     147: 210,  # '\x93'
     148: 211,  # '\x94'
     149: 212,  # '\x95'
     150: 213,  # '\x96'
     151: 214,  # '\x97'
     152: 215,  # '\x98'
     153: 216,  # '\x99'
     154: 217,  # '\x9a'
     155: 218,  # '\x9b'
     156: 219,  # '\x9c'
     157: 220,  # '\x9d'
     158: 221,  # '\x9e'
     159: 222,  # '\x9f'
     160: 223,  # '\xa0'
     161: 224,  # 'Ё'
     162: 225,  # 'Ђ'
     163: 226,  # 'Ѓ'
     164: 227,  # 'Є'
     165: 228,  # 'Ѕ'
     166: 229,  # 'І'
     167: 230,  # 'Ї'
     168: 231,  # 'Ј'
     169: 232,  # 'Љ'
     170: 233,  # 'Њ'
     171: 234,  # 'Ћ'
     172: 235,  # 'Ќ'
     173: 236,  # '\xad'
     174: 237,  # 'Ў'
     175: 238,  # 'Џ'
     176: 37,  # 'А'
     177: 44,  # 'Б'
     178: 33,  # 'В'
     179: 46,  # 'Г'
     180: 41,  # 'Д'
     181: 48,  # 'Е'
     182: 56,  # 'Ж'
     183: 51,  # 'З'
     184: 42,  # 'И'
     185: 60,  # 'Й'
     186: 36,  # 'К'
     187: 49,  # 'Л'
     188: 38,  # 'М'
     189: 31,  # 'Н'
     190: 34,  # 'О'
     191: 35,  # 'П'
     192: 45,  # 'Р'
     193: 32,  # 'С'
     194: 40,  # 'Т'
     195: 52,  # 'У'
     196: 53,  # 'Ф'
     197: 55,  # 'Х'
     198: 58,  # 'Ц'
     199: 50,  # 'Ч'
     200: 57,  # 'Ш'
     201: 63,  # 'Щ'
     202: 70,  # 'Ъ'
     203: 62,  # 'Ы'
     204: 61,  # 'Ь'
     205: 47,  # 'Э'
     206: 59,  # 'Ю'
     207: 43,  # 'Я'
     208: 3,  # 'а'
     209: 21,  # 'б'
     210: 10,  # 'в'
     211: 19,  # 'г'
     212: 13,  # 'д'
     213: 2,  # 'е'
     214: 24,  # 'ж'
     215: 20,  # 'з'
     216: 4,  # 'и'
     217: 23,  # 'й'
     218: 11,  # 'к'
     219: 8,  # 'л'
     220: 12,  # 'м'
     221: 5,  # 'н'
     222: 1,  # 'о'
     223: 15,  # 'п'
     224: 9,  # 'р'
     225: 7,  # 'с'
     226: 6,  # 'т'
     227: 14,  # 'у'
     228: 39,  # 'ф'
     229: 26,  # 'х'
     230: 28,  # 'ц'
     231: 22,  # 'ч'
     232: 25,  # 'ш'
     233: 29,  # 'щ'
     234: 54,  # 'ъ'
     235: 18,  # 'ы'
     236: 17,  # 'ь'
     237: 30,  # 'э'
     238: 27,  # 'ю'
     239: 16,  # 'я'
     240: 239,  # '№'
     241: 68,  # 'ё'
     242: 240,  # 'ђ'
     243: 241,  # 'ѓ'
     244: 242,  # 'є'
     245: 243,  # 'ѕ'
     246: 244,  # 'і'
     247: 245,  # 'ї'
     248: 246,  # 'ј'
     249: 247,  # 'љ'
     250: 248,  # 'њ'
     251: 249,  # 'ћ'
     252: 250,  # 'ќ'
     253: 251,  # '§'
     254: 252,  # 'ў'
     255: 255,  # 'џ'
}

ISO_8859_5_RUSSIAN_MODEL = SingleByteCharSetModel(charset_name='ISO-8859-5',
                                                  language='Russian',
                                                  char_to_order_map=ISO_8859_5_RUSSIAN_CHAR_TO_ORDER,
                                                  language_model=RUSSIAN_LANG_MODEL,
                                                  typical_positive_ratio=0.976601,
                                                  keep_ascii_letters=False,
                                                  alphabet='ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё')

site-packages/pip/_vendor/chardet/big5prober.py000064400000003335150564745010015547 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import Big5DistributionAnalysis
from .mbcssm import BIG5_SM_MODEL


class Big5Prober(MultiByteCharSetProber):
    def __init__(self):
        super(Big5Prober, self).__init__()
        self.coding_sm = CodingStateMachine(BIG5_SM_MODEL)
        self.distribution_analyzer = Big5DistributionAnalysis()
        self.reset()

    @property
    def charset_name(self):
        return "Big5"

    @property
    def language(self):
        return "Chinese"
site-packages/pip/_vendor/chardet/mbcsgroupprober.py000064400000003734150564745010016725 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#   Proofpoint, Inc.
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetgroupprober import CharSetGroupProber
from .utf8prober import UTF8Prober
from .sjisprober import SJISProber
from .eucjpprober import EUCJPProber
from .gb2312prober import GB2312Prober
from .euckrprober import EUCKRProber
from .cp949prober import CP949Prober
from .big5prober import Big5Prober
from .euctwprober import EUCTWProber


class MBCSGroupProber(CharSetGroupProber):
    def __init__(self, lang_filter=None):
        super(MBCSGroupProber, self).__init__(lang_filter=lang_filter)
        self.probers = [
            UTF8Prober(),
            SJISProber(),
            EUCJPProber(),
            GB2312Prober(),
            EUCKRProber(),
            CP949Prober(),
            Big5Prober(),
            EUCTWProber()
        ]
        self.reset()
site-packages/pip/_vendor/chardet/jisfreq.py000064400000062261150564745010015155 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

# Sampling from about 20M text materials include literature and computer technology
#
# Japanese frequency table, applied to both S-JIS and EUC-JP
# They are sorted in order.

# 128  --> 0.77094
# 256  --> 0.85710
# 512  --> 0.92635
# 1024 --> 0.97130
# 2048 --> 0.99431
#
# Ideal Distribution Ratio = 0.92635 / (1-0.92635) = 12.58
# Random Distribution Ration = 512 / (2965+62+83+86-512) = 0.191
#
# Typical Distribution Ratio, 25% of IDR

JIS_TYPICAL_DISTRIBUTION_RATIO = 3.0

# Char to FreqOrder table ,
JIS_TABLE_SIZE = 4368

JIS_CHAR_TO_FREQ_ORDER = (
  40,   1,   6, 182, 152, 180, 295,2127, 285, 381,3295,4304,3068,4606,3165,3510, #   16
3511,1822,2785,4607,1193,2226,5070,4608, 171,2996,1247,  18, 179,5071, 856,1661, #   32
1262,5072, 619, 127,3431,3512,3230,1899,1700, 232, 228,1294,1298, 284, 283,2041, #   48
2042,1061,1062,  48,  49,  44,  45, 433, 434,1040,1041, 996, 787,2997,1255,4305, #   64
2108,4609,1684,1648,5073,5074,5075,5076,5077,5078,3687,5079,4610,5080,3927,3928, #   80
5081,3296,3432, 290,2285,1471,2187,5082,2580,2825,1303,2140,1739,1445,2691,3375, #   96
1691,3297,4306,4307,4611, 452,3376,1182,2713,3688,3069,4308,5083,5084,5085,5086, #  112
5087,5088,5089,5090,5091,5092,5093,5094,5095,5096,5097,5098,5099,5100,5101,5102, #  128
5103,5104,5105,5106,5107,5108,5109,5110,5111,5112,4097,5113,5114,5115,5116,5117, #  144
5118,5119,5120,5121,5122,5123,5124,5125,5126,5127,5128,5129,5130,5131,5132,5133, #  160
5134,5135,5136,5137,5138,5139,5140,5141,5142,5143,5144,5145,5146,5147,5148,5149, #  176
5150,5151,5152,4612,5153,5154,5155,5156,5157,5158,5159,5160,5161,5162,5163,5164, #  192
5165,5166,5167,5168,5169,5170,5171,5172,5173,5174,5175,1472, 598, 618, 820,1205, #  208
1309,1412,1858,1307,1692,5176,5177,5178,5179,5180,5181,5182,1142,1452,1234,1172, #  224
1875,2043,2149,1793,1382,2973, 925,2404,1067,1241, 960,1377,2935,1491, 919,1217, #  240
1865,2030,1406,1499,2749,4098,5183,5184,5185,5186,5187,5188,2561,4099,3117,1804, #  256
2049,3689,4309,3513,1663,5189,3166,3118,3298,1587,1561,3433,5190,3119,1625,2998, #  272
3299,4613,1766,3690,2786,4614,5191,5192,5193,5194,2161,  26,3377,   2,3929,  20, #  288
3691,  47,4100,  50,  17,  16,  35, 268,  27, 243,  42, 155,  24, 154,  29, 184, #  304
   4,  91,  14,  92,  53, 396,  33, 289,   9,  37,  64, 620,  21,  39, 321,   5, #  320
  12,  11,  52,  13,   3, 208, 138,   0,   7,  60, 526, 141, 151,1069, 181, 275, #  336
1591,  83, 132,1475, 126, 331, 829,  15,  69, 160,  59,  22, 157,  55,1079, 312, #  352
 109,  38,  23,  25,  10,  19,  79,5195,  61, 382,1124,   8,  30,5196,5197,5198, #  368
5199,5200,5201,5202,5203,5204,5205,5206,  89,  62,  74,  34,2416, 112, 139, 196, #  384
 271, 149,  84, 607, 131, 765,  46,  88, 153, 683,  76, 874, 101, 258,  57,  80, #  400
  32, 364, 121,1508, 169,1547,  68, 235, 145,2999,  41, 360,3027,  70,  63,  31, #  416
  43, 259, 262,1383,  99, 533, 194,  66,  93, 846, 217, 192,  56, 106,  58, 565, #  432
 280, 272, 311, 256, 146,  82, 308,  71, 100, 128, 214, 655, 110, 261, 104,1140, #  448
  54,  51,  36,  87,  67,3070, 185,2618,2936,2020,  28,1066,2390,2059,5207,5208, #  464
5209,5210,5211,5212,5213,5214,5215,5216,4615,5217,5218,5219,5220,5221,5222,5223, #  480
5224,5225,5226,5227,5228,5229,5230,5231,5232,5233,5234,5235,5236,3514,5237,5238, #  496
5239,5240,5241,5242,5243,5244,2297,2031,4616,4310,3692,5245,3071,5246,3598,5247, #  512
4617,3231,3515,5248,4101,4311,4618,3808,4312,4102,5249,4103,4104,3599,5250,5251, #  528
5252,5253,5254,5255,5256,5257,5258,5259,5260,5261,5262,5263,5264,5265,5266,5267, #  544
5268,5269,5270,5271,5272,5273,5274,5275,5276,5277,5278,5279,5280,5281,5282,5283, #  560
5284,5285,5286,5287,5288,5289,5290,5291,5292,5293,5294,5295,5296,5297,5298,5299, #  576
5300,5301,5302,5303,5304,5305,5306,5307,5308,5309,5310,5311,5312,5313,5314,5315, #  592
5316,5317,5318,5319,5320,5321,5322,5323,5324,5325,5326,5327,5328,5329,5330,5331, #  608
5332,5333,5334,5335,5336,5337,5338,5339,5340,5341,5342,5343,5344,5345,5346,5347, #  624
5348,5349,5350,5351,5352,5353,5354,5355,5356,5357,5358,5359,5360,5361,5362,5363, #  640
5364,5365,5366,5367,5368,5369,5370,5371,5372,5373,5374,5375,5376,5377,5378,5379, #  656
5380,5381, 363, 642,2787,2878,2788,2789,2316,3232,2317,3434,2011, 165,1942,3930, #  672
3931,3932,3933,5382,4619,5383,4620,5384,5385,5386,5387,5388,5389,5390,5391,5392, #  688
5393,5394,5395,5396,5397,5398,5399,5400,5401,5402,5403,5404,5405,5406,5407,5408, #  704
5409,5410,5411,5412,5413,5414,5415,5416,5417,5418,5419,5420,5421,5422,5423,5424, #  720
5425,5426,5427,5428,5429,5430,5431,5432,5433,5434,5435,5436,5437,5438,5439,5440, #  736
5441,5442,5443,5444,5445,5446,5447,5448,5449,5450,5451,5452,5453,5454,5455,5456, #  752
5457,5458,5459,5460,5461,5462,5463,5464,5465,5466,5467,5468,5469,5470,5471,5472, #  768
5473,5474,5475,5476,5477,5478,5479,5480,5481,5482,5483,5484,5485,5486,5487,5488, #  784
5489,5490,5491,5492,5493,5494,5495,5496,5497,5498,5499,5500,5501,5502,5503,5504, #  800
5505,5506,5507,5508,5509,5510,5511,5512,5513,5514,5515,5516,5517,5518,5519,5520, #  816
5521,5522,5523,5524,5525,5526,5527,5528,5529,5530,5531,5532,5533,5534,5535,5536, #  832
5537,5538,5539,5540,5541,5542,5543,5544,5545,5546,5547,5548,5549,5550,5551,5552, #  848
5553,5554,5555,5556,5557,5558,5559,5560,5561,5562,5563,5564,5565,5566,5567,5568, #  864
5569,5570,5571,5572,5573,5574,5575,5576,5577,5578,5579,5580,5581,5582,5583,5584, #  880
5585,5586,5587,5588,5589,5590,5591,5592,5593,5594,5595,5596,5597,5598,5599,5600, #  896
5601,5602,5603,5604,5605,5606,5607,5608,5609,5610,5611,5612,5613,5614,5615,5616, #  912
5617,5618,5619,5620,5621,5622,5623,5624,5625,5626,5627,5628,5629,5630,5631,5632, #  928
5633,5634,5635,5636,5637,5638,5639,5640,5641,5642,5643,5644,5645,5646,5647,5648, #  944
5649,5650,5651,5652,5653,5654,5655,5656,5657,5658,5659,5660,5661,5662,5663,5664, #  960
5665,5666,5667,5668,5669,5670,5671,5672,5673,5674,5675,5676,5677,5678,5679,5680, #  976
5681,5682,5683,5684,5685,5686,5687,5688,5689,5690,5691,5692,5693,5694,5695,5696, #  992
5697,5698,5699,5700,5701,5702,5703,5704,5705,5706,5707,5708,5709,5710,5711,5712, # 1008
5713,5714,5715,5716,5717,5718,5719,5720,5721,5722,5723,5724,5725,5726,5727,5728, # 1024
5729,5730,5731,5732,5733,5734,5735,5736,5737,5738,5739,5740,5741,5742,5743,5744, # 1040
5745,5746,5747,5748,5749,5750,5751,5752,5753,5754,5755,5756,5757,5758,5759,5760, # 1056
5761,5762,5763,5764,5765,5766,5767,5768,5769,5770,5771,5772,5773,5774,5775,5776, # 1072
5777,5778,5779,5780,5781,5782,5783,5784,5785,5786,5787,5788,5789,5790,5791,5792, # 1088
5793,5794,5795,5796,5797,5798,5799,5800,5801,5802,5803,5804,5805,5806,5807,5808, # 1104
5809,5810,5811,5812,5813,5814,5815,5816,5817,5818,5819,5820,5821,5822,5823,5824, # 1120
5825,5826,5827,5828,5829,5830,5831,5832,5833,5834,5835,5836,5837,5838,5839,5840, # 1136
5841,5842,5843,5844,5845,5846,5847,5848,5849,5850,5851,5852,5853,5854,5855,5856, # 1152
5857,5858,5859,5860,5861,5862,5863,5864,5865,5866,5867,5868,5869,5870,5871,5872, # 1168
5873,5874,5875,5876,5877,5878,5879,5880,5881,5882,5883,5884,5885,5886,5887,5888, # 1184
5889,5890,5891,5892,5893,5894,5895,5896,5897,5898,5899,5900,5901,5902,5903,5904, # 1200
5905,5906,5907,5908,5909,5910,5911,5912,5913,5914,5915,5916,5917,5918,5919,5920, # 1216
5921,5922,5923,5924,5925,5926,5927,5928,5929,5930,5931,5932,5933,5934,5935,5936, # 1232
5937,5938,5939,5940,5941,5942,5943,5944,5945,5946,5947,5948,5949,5950,5951,5952, # 1248
5953,5954,5955,5956,5957,5958,5959,5960,5961,5962,5963,5964,5965,5966,5967,5968, # 1264
5969,5970,5971,5972,5973,5974,5975,5976,5977,5978,5979,5980,5981,5982,5983,5984, # 1280
5985,5986,5987,5988,5989,5990,5991,5992,5993,5994,5995,5996,5997,5998,5999,6000, # 1296
6001,6002,6003,6004,6005,6006,6007,6008,6009,6010,6011,6012,6013,6014,6015,6016, # 1312
6017,6018,6019,6020,6021,6022,6023,6024,6025,6026,6027,6028,6029,6030,6031,6032, # 1328
6033,6034,6035,6036,6037,6038,6039,6040,6041,6042,6043,6044,6045,6046,6047,6048, # 1344
6049,6050,6051,6052,6053,6054,6055,6056,6057,6058,6059,6060,6061,6062,6063,6064, # 1360
6065,6066,6067,6068,6069,6070,6071,6072,6073,6074,6075,6076,6077,6078,6079,6080, # 1376
6081,6082,6083,6084,6085,6086,6087,6088,6089,6090,6091,6092,6093,6094,6095,6096, # 1392
6097,6098,6099,6100,6101,6102,6103,6104,6105,6106,6107,6108,6109,6110,6111,6112, # 1408
6113,6114,2044,2060,4621, 997,1235, 473,1186,4622, 920,3378,6115,6116, 379,1108, # 1424
4313,2657,2735,3934,6117,3809, 636,3233, 573,1026,3693,3435,2974,3300,2298,4105, # 1440
 854,2937,2463, 393,2581,2417, 539, 752,1280,2750,2480, 140,1161, 440, 708,1569, # 1456
 665,2497,1746,1291,1523,3000, 164,1603, 847,1331, 537,1997, 486, 508,1693,2418, # 1472
1970,2227, 878,1220, 299,1030, 969, 652,2751, 624,1137,3301,2619,  65,3302,2045, # 1488
1761,1859,3120,1930,3694,3516, 663,1767, 852, 835,3695, 269, 767,2826,2339,1305, # 1504
 896,1150, 770,1616,6118, 506,1502,2075,1012,2519, 775,2520,2975,2340,2938,4314, # 1520
3028,2086,1224,1943,2286,6119,3072,4315,2240,1273,1987,3935,1557, 175, 597, 985, # 1536
3517,2419,2521,1416,3029, 585, 938,1931,1007,1052,1932,1685,6120,3379,4316,4623, # 1552
 804, 599,3121,1333,2128,2539,1159,1554,2032,3810, 687,2033,2904, 952, 675,1467, # 1568
3436,6121,2241,1096,1786,2440,1543,1924, 980,1813,2228, 781,2692,1879, 728,1918, # 1584
3696,4624, 548,1950,4625,1809,1088,1356,3303,2522,1944, 502, 972, 373, 513,2827, # 1600
 586,2377,2391,1003,1976,1631,6122,2464,1084, 648,1776,4626,2141, 324, 962,2012, # 1616
2177,2076,1384, 742,2178,1448,1173,1810, 222, 102, 301, 445, 125,2420, 662,2498, # 1632
 277, 200,1476,1165,1068, 224,2562,1378,1446, 450,1880, 659, 791, 582,4627,2939, # 1648
3936,1516,1274, 555,2099,3697,1020,1389,1526,3380,1762,1723,1787,2229, 412,2114, # 1664
1900,2392,3518, 512,2597, 427,1925,2341,3122,1653,1686,2465,2499, 697, 330, 273, # 1680
 380,2162, 951, 832, 780, 991,1301,3073, 965,2270,3519, 668,2523,2636,1286, 535, # 1696
1407, 518, 671, 957,2658,2378, 267, 611,2197,3030,6123, 248,2299, 967,1799,2356, # 1712
 850,1418,3437,1876,1256,1480,2828,1718,6124,6125,1755,1664,2405,6126,4628,2879, # 1728
2829, 499,2179, 676,4629, 557,2329,2214,2090, 325,3234, 464, 811,3001, 992,2342, # 1744
2481,1232,1469, 303,2242, 466,1070,2163, 603,1777,2091,4630,2752,4631,2714, 322, # 1760
2659,1964,1768, 481,2188,1463,2330,2857,3600,2092,3031,2421,4632,2318,2070,1849, # 1776
2598,4633,1302,2254,1668,1701,2422,3811,2905,3032,3123,2046,4106,1763,1694,4634, # 1792
1604, 943,1724,1454, 917, 868,2215,1169,2940, 552,1145,1800,1228,1823,1955, 316, # 1808
1080,2510, 361,1807,2830,4107,2660,3381,1346,1423,1134,4108,6127, 541,1263,1229, # 1824
1148,2540, 545, 465,1833,2880,3438,1901,3074,2482, 816,3937, 713,1788,2500, 122, # 1840
1575, 195,1451,2501,1111,6128, 859, 374,1225,2243,2483,4317, 390,1033,3439,3075, # 1856
2524,1687, 266, 793,1440,2599, 946, 779, 802, 507, 897,1081, 528,2189,1292, 711, # 1872
1866,1725,1167,1640, 753, 398,2661,1053, 246, 348,4318, 137,1024,3440,1600,2077, # 1888
2129, 825,4319, 698, 238, 521, 187,2300,1157,2423,1641,1605,1464,1610,1097,2541, # 1904
1260,1436, 759,2255,1814,2150, 705,3235, 409,2563,3304, 561,3033,2005,2564, 726, # 1920
1956,2343,3698,4109, 949,3812,3813,3520,1669, 653,1379,2525, 881,2198, 632,2256, # 1936
1027, 778,1074, 733,1957, 514,1481,2466, 554,2180, 702,3938,1606,1017,1398,6129, # 1952
1380,3521, 921, 993,1313, 594, 449,1489,1617,1166, 768,1426,1360, 495,1794,3601, # 1968
1177,3602,1170,4320,2344, 476, 425,3167,4635,3168,1424, 401,2662,1171,3382,1998, # 1984
1089,4110, 477,3169, 474,6130,1909, 596,2831,1842, 494, 693,1051,1028,1207,3076, # 2000
 606,2115, 727,2790,1473,1115, 743,3522, 630, 805,1532,4321,2021, 366,1057, 838, # 2016
 684,1114,2142,4322,2050,1492,1892,1808,2271,3814,2424,1971,1447,1373,3305,1090, # 2032
1536,3939,3523,3306,1455,2199, 336, 369,2331,1035, 584,2393, 902, 718,2600,6131, # 2048
2753, 463,2151,1149,1611,2467, 715,1308,3124,1268, 343,1413,3236,1517,1347,2663, # 2064
2093,3940,2022,1131,1553,2100,2941,1427,3441,2942,1323,2484,6132,1980, 872,2368, # 2080
2441,2943, 320,2369,2116,1082, 679,1933,3941,2791,3815, 625,1143,2023, 422,2200, # 2096
3816,6133, 730,1695, 356,2257,1626,2301,2858,2637,1627,1778, 937, 883,2906,2693, # 2112
3002,1769,1086, 400,1063,1325,3307,2792,4111,3077, 456,2345,1046, 747,6134,1524, # 2128
 884,1094,3383,1474,2164,1059, 974,1688,2181,2258,1047, 345,1665,1187, 358, 875, # 2144
3170, 305, 660,3524,2190,1334,1135,3171,1540,1649,2542,1527, 927, 968,2793, 885, # 2160
1972,1850, 482, 500,2638,1218,1109,1085,2543,1654,2034, 876,  78,2287,1482,1277, # 2176
 861,1675,1083,1779, 724,2754, 454, 397,1132,1612,2332, 893, 672,1237, 257,2259, # 2192
2370, 135,3384, 337,2244, 547, 352, 340, 709,2485,1400, 788,1138,2511, 540, 772, # 2208
1682,2260,2272,2544,2013,1843,1902,4636,1999,1562,2288,4637,2201,1403,1533, 407, # 2224
 576,3308,1254,2071, 978,3385, 170, 136,1201,3125,2664,3172,2394, 213, 912, 873, # 2240
3603,1713,2202, 699,3604,3699, 813,3442, 493, 531,1054, 468,2907,1483, 304, 281, # 2256
4112,1726,1252,2094, 339,2319,2130,2639, 756,1563,2944, 748, 571,2976,1588,2425, # 2272
2715,1851,1460,2426,1528,1392,1973,3237, 288,3309, 685,3386, 296, 892,2716,2216, # 2288
1570,2245, 722,1747,2217, 905,3238,1103,6135,1893,1441,1965, 251,1805,2371,3700, # 2304
2601,1919,1078,  75,2182,1509,1592,1270,2640,4638,2152,6136,3310,3817, 524, 706, # 2320
1075, 292,3818,1756,2602, 317,  98,3173,3605,3525,1844,2218,3819,2502, 814, 567, # 2336
 385,2908,1534,6137, 534,1642,3239, 797,6138,1670,1529, 953,4323, 188,1071, 538, # 2352
 178, 729,3240,2109,1226,1374,2000,2357,2977, 731,2468,1116,2014,2051,6139,1261, # 2368
1593, 803,2859,2736,3443, 556, 682, 823,1541,6140,1369,2289,1706,2794, 845, 462, # 2384
2603,2665,1361, 387, 162,2358,1740, 739,1770,1720,1304,1401,3241,1049, 627,1571, # 2400
2427,3526,1877,3942,1852,1500, 431,1910,1503, 677, 297,2795, 286,1433,1038,1198, # 2416
2290,1133,1596,4113,4639,2469,1510,1484,3943,6141,2442, 108, 712,4640,2372, 866, # 2432
3701,2755,3242,1348, 834,1945,1408,3527,2395,3243,1811, 824, 994,1179,2110,1548, # 2448
1453, 790,3003, 690,4324,4325,2832,2909,3820,1860,3821, 225,1748, 310, 346,1780, # 2464
2470, 821,1993,2717,2796, 828, 877,3528,2860,2471,1702,2165,2910,2486,1789, 453, # 2480
 359,2291,1676,  73,1164,1461,1127,3311, 421, 604, 314,1037, 589, 116,2487, 737, # 2496
 837,1180, 111, 244, 735,6142,2261,1861,1362, 986, 523, 418, 581,2666,3822, 103, # 2512
 855, 503,1414,1867,2488,1091, 657,1597, 979, 605,1316,4641,1021,2443,2078,2001, # 2528
1209,  96, 587,2166,1032, 260,1072,2153, 173,  94, 226,3244, 819,2006,4642,4114, # 2544
2203, 231,1744, 782,  97,2667, 786,3387, 887, 391, 442,2219,4326,1425,6143,2694, # 2560
 633,1544,1202, 483,2015, 592,2052,1958,2472,1655, 419, 129,4327,3444,3312,1714, # 2576
1257,3078,4328,1518,1098, 865,1310,1019,1885,1512,1734, 469,2444, 148, 773, 436, # 2592
1815,1868,1128,1055,4329,1245,2756,3445,2154,1934,1039,4643, 579,1238, 932,2320, # 2608
 353, 205, 801, 115,2428, 944,2321,1881, 399,2565,1211, 678, 766,3944, 335,2101, # 2624
1459,1781,1402,3945,2737,2131,1010, 844, 981,1326,1013, 550,1816,1545,2620,1335, # 2640
1008, 371,2881, 936,1419,1613,3529,1456,1395,2273,1834,2604,1317,2738,2503, 416, # 2656
1643,4330, 806,1126, 229, 591,3946,1314,1981,1576,1837,1666, 347,1790, 977,3313, # 2672
 764,2861,1853, 688,2429,1920,1462,  77, 595, 415,2002,3034, 798,1192,4115,6144, # 2688
2978,4331,3035,2695,2582,2072,2566, 430,2430,1727, 842,1396,3947,3702, 613, 377, # 2704
 278, 236,1417,3388,3314,3174, 757,1869, 107,3530,6145,1194, 623,2262, 207,1253, # 2720
2167,3446,3948, 492,1117,1935, 536,1838,2757,1246,4332, 696,2095,2406,1393,1572, # 2736
3175,1782, 583, 190, 253,1390,2230, 830,3126,3389, 934,3245,1703,1749,2979,1870, # 2752
2545,1656,2204, 869,2346,4116,3176,1817, 496,1764,4644, 942,1504, 404,1903,1122, # 2768
1580,3606,2945,1022, 515, 372,1735, 955,2431,3036,6146,2797,1110,2302,2798, 617, # 2784
6147, 441, 762,1771,3447,3607,3608,1904, 840,3037,  86, 939,1385, 572,1370,2445, # 2800
1336, 114,3703, 898, 294, 203,3315, 703,1583,2274, 429, 961,4333,1854,1951,3390, # 2816
2373,3704,4334,1318,1381, 966,1911,2322,1006,1155, 309, 989, 458,2718,1795,1372, # 2832
1203, 252,1689,1363,3177, 517,1936, 168,1490, 562, 193,3823,1042,4117,1835, 551, # 2848
 470,4645, 395, 489,3448,1871,1465,2583,2641, 417,1493, 279,1295, 511,1236,1119, # 2864
  72,1231,1982,1812,3004, 871,1564, 984,3449,1667,2696,2096,4646,2347,2833,1673, # 2880
3609, 695,3246,2668, 807,1183,4647, 890, 388,2333,1801,1457,2911,1765,1477,1031, # 2896
3316,3317,1278,3391,2799,2292,2526, 163,3450,4335,2669,1404,1802,6148,2323,2407, # 2912
1584,1728,1494,1824,1269, 298, 909,3318,1034,1632, 375, 776,1683,2061, 291, 210, # 2928
1123, 809,1249,1002,2642,3038, 206,1011,2132, 144, 975, 882,1565, 342, 667, 754, # 2944
1442,2143,1299,2303,2062, 447, 626,2205,1221,2739,2912,1144,1214,2206,2584, 760, # 2960
1715, 614, 950,1281,2670,2621, 810, 577,1287,2546,4648, 242,2168, 250,2643, 691, # 2976
 123,2644, 647, 313,1029, 689,1357,2946,1650, 216, 771,1339,1306, 808,2063, 549, # 2992
 913,1371,2913,2914,6149,1466,1092,1174,1196,1311,2605,2396,1783,1796,3079, 406, # 3008
2671,2117,3949,4649, 487,1825,2220,6150,2915, 448,2348,1073,6151,2397,1707, 130, # 3024
 900,1598, 329, 176,1959,2527,1620,6152,2275,4336,3319,1983,2191,3705,3610,2155, # 3040
3706,1912,1513,1614,6153,1988, 646, 392,2304,1589,3320,3039,1826,1239,1352,1340, # 3056
2916, 505,2567,1709,1437,2408,2547, 906,6154,2672, 384,1458,1594,1100,1329, 710, # 3072
 423,3531,2064,2231,2622,1989,2673,1087,1882, 333, 841,3005,1296,2882,2379, 580, # 3088
1937,1827,1293,2585, 601, 574, 249,1772,4118,2079,1120, 645, 901,1176,1690, 795, # 3104
2207, 478,1434, 516,1190,1530, 761,2080, 930,1264, 355, 435,1552, 644,1791, 987, # 3120
 220,1364,1163,1121,1538, 306,2169,1327,1222, 546,2645, 218, 241, 610,1704,3321, # 3136
1984,1839,1966,2528, 451,6155,2586,3707,2568, 907,3178, 254,2947, 186,1845,4650, # 3152
 745, 432,1757, 428,1633, 888,2246,2221,2489,3611,2118,1258,1265, 956,3127,1784, # 3168
4337,2490, 319, 510, 119, 457,3612, 274,2035,2007,4651,1409,3128, 970,2758, 590, # 3184
2800, 661,2247,4652,2008,3950,1420,1549,3080,3322,3951,1651,1375,2111, 485,2491, # 3200
1429,1156,6156,2548,2183,1495, 831,1840,2529,2446, 501,1657, 307,1894,3247,1341, # 3216
 666, 899,2156,1539,2549,1559, 886, 349,2208,3081,2305,1736,3824,2170,2759,1014, # 3232
1913,1386, 542,1397,2948, 490, 368, 716, 362, 159, 282,2569,1129,1658,1288,1750, # 3248
2674, 276, 649,2016, 751,1496, 658,1818,1284,1862,2209,2087,2512,3451, 622,2834, # 3264
 376, 117,1060,2053,1208,1721,1101,1443, 247,1250,3179,1792,3952,2760,2398,3953, # 3280
6157,2144,3708, 446,2432,1151,2570,3452,2447,2761,2835,1210,2448,3082, 424,2222, # 3296
1251,2449,2119,2836, 504,1581,4338, 602, 817, 857,3825,2349,2306, 357,3826,1470, # 3312
1883,2883, 255, 958, 929,2917,3248, 302,4653,1050,1271,1751,2307,1952,1430,2697, # 3328
2719,2359, 354,3180, 777, 158,2036,4339,1659,4340,4654,2308,2949,2248,1146,2232, # 3344
3532,2720,1696,2623,3827,6158,3129,1550,2698,1485,1297,1428, 637, 931,2721,2145, # 3360
 914,2550,2587,  81,2450, 612, 827,2646,1242,4655,1118,2884, 472,1855,3181,3533, # 3376
3534, 569,1353,2699,1244,1758,2588,4119,2009,2762,2171,3709,1312,1531,6159,1152, # 3392
1938, 134,1830, 471,3710,2276,1112,1535,3323,3453,3535, 982,1337,2950, 488, 826, # 3408
 674,1058,1628,4120,2017, 522,2399, 211, 568,1367,3454, 350, 293,1872,1139,3249, # 3424
1399,1946,3006,1300,2360,3324, 588, 736,6160,2606, 744, 669,3536,3828,6161,1358, # 3440
 199, 723, 848, 933, 851,1939,1505,1514,1338,1618,1831,4656,1634,3613, 443,2740, # 3456
3829, 717,1947, 491,1914,6162,2551,1542,4121,1025,6163,1099,1223, 198,3040,2722, # 3472
 370, 410,1905,2589, 998,1248,3182,2380, 519,1449,4122,1710, 947, 928,1153,4341, # 3488
2277, 344,2624,1511, 615, 105, 161,1212,1076,1960,3130,2054,1926,1175,1906,2473, # 3504
 414,1873,2801,6164,2309, 315,1319,3325, 318,2018,2146,2157, 963, 631, 223,4342, # 3520
4343,2675, 479,3711,1197,2625,3712,2676,2361,6165,4344,4123,6166,2451,3183,1886, # 3536
2184,1674,1330,1711,1635,1506, 799, 219,3250,3083,3954,1677,3713,3326,2081,3614, # 3552
1652,2073,4657,1147,3041,1752, 643,1961, 147,1974,3955,6167,1716,2037, 918,3007, # 3568
1994, 120,1537, 118, 609,3184,4345, 740,3455,1219, 332,1615,3830,6168,1621,2980, # 3584
1582, 783, 212, 553,2350,3714,1349,2433,2082,4124, 889,6169,2310,1275,1410, 973, # 3600
 166,1320,3456,1797,1215,3185,2885,1846,2590,2763,4658, 629, 822,3008, 763, 940, # 3616
1990,2862, 439,2409,1566,1240,1622, 926,1282,1907,2764, 654,2210,1607, 327,1130, # 3632
3956,1678,1623,6170,2434,2192, 686, 608,3831,3715, 903,3957,3042,6171,2741,1522, # 3648
1915,1105,1555,2552,1359, 323,3251,4346,3457, 738,1354,2553,2311,2334,1828,2003, # 3664
3832,1753,2351,1227,6172,1887,4125,1478,6173,2410,1874,1712,1847, 520,1204,2607, # 3680
 264,4659, 836,2677,2102, 600,4660,3833,2278,3084,6174,4347,3615,1342, 640, 532, # 3696
 543,2608,1888,2400,2591,1009,4348,1497, 341,1737,3616,2723,1394, 529,3252,1321, # 3712
 983,4661,1515,2120, 971,2592, 924, 287,1662,3186,4349,2700,4350,1519, 908,1948, # 3728
2452, 156, 796,1629,1486,2223,2055, 694,4126,1259,1036,3392,1213,2249,2742,1889, # 3744
1230,3958,1015, 910, 408, 559,3617,4662, 746, 725, 935,4663,3959,3009,1289, 563, # 3760
 867,4664,3960,1567,2981,2038,2626, 988,2263,2381,4351, 143,2374, 704,1895,6175, # 3776
1188,3716,2088, 673,3085,2362,4352, 484,1608,1921,2765,2918, 215, 904,3618,3537, # 3792
 894, 509, 976,3043,2701,3961,4353,2837,2982, 498,6176,6177,1102,3538,1332,3393, # 3808
1487,1636,1637, 233, 245,3962, 383, 650, 995,3044, 460,1520,1206,2352, 749,3327, # 3824
 530, 700, 389,1438,1560,1773,3963,2264, 719,2951,2724,3834, 870,1832,1644,1000, # 3840
 839,2474,3717, 197,1630,3394, 365,2886,3964,1285,2133, 734, 922, 818,1106, 732, # 3856
 480,2083,1774,3458, 923,2279,1350, 221,3086,  85,2233,2234,3835,1585,3010,2147, # 3872
1387,1705,2382,1619,2475, 133, 239,2802,1991,1016,2084,2383, 411,2838,1113, 651, # 3888
1985,1160,3328, 990,1863,3087,1048,1276,2647, 265,2627,1599,3253,2056, 150, 638, # 3904
2019, 656, 853, 326,1479, 680,1439,4354,1001,1759, 413,3459,3395,2492,1431, 459, # 3920
4355,1125,3329,2265,1953,1450,2065,2863, 849, 351,2678,3131,3254,3255,1104,1577, # 3936
 227,1351,1645,2453,2193,1421,2887, 812,2121, 634,  95,2435, 201,2312,4665,1646, # 3952
1671,2743,1601,2554,2702,2648,2280,1315,1366,2089,3132,1573,3718,3965,1729,1189, # 3968
 328,2679,1077,1940,1136, 558,1283, 964,1195, 621,2074,1199,1743,3460,3619,1896, # 3984
1916,1890,3836,2952,1154,2112,1064, 862, 378,3011,2066,2113,2803,1568,2839,6178, # 4000
3088,2919,1941,1660,2004,1992,2194, 142, 707,1590,1708,1624,1922,1023,1836,1233, # 4016
1004,2313, 789, 741,3620,6179,1609,2411,1200,4127,3719,3720,4666,2057,3721, 593, # 4032
2840, 367,2920,1878,6180,3461,1521, 628,1168, 692,2211,2649, 300, 720,2067,2571, # 4048
2953,3396, 959,2504,3966,3539,3462,1977, 701,6181, 954,1043, 800, 681, 183,3722, # 4064
1803,1730,3540,4128,2103, 815,2314, 174, 467, 230,2454,1093,2134, 755,3541,3397, # 4080
1141,1162,6182,1738,2039, 270,3256,2513,1005,1647,2185,3837, 858,1679,1897,1719, # 4096
2954,2324,1806, 402, 670, 167,4129,1498,2158,2104, 750,6183, 915, 189,1680,1551, # 4112
 455,4356,1501,2455, 405,1095,2955, 338,1586,1266,1819, 570, 641,1324, 237,1556, # 4128
2650,1388,3723,6184,1368,2384,1343,1978,3089,2436, 879,3724, 792,1191, 758,3012, # 4144
1411,2135,1322,4357, 240,4667,1848,3725,1574,6185, 420,3045,1546,1391, 714,4358, # 4160
1967, 941,1864, 863, 664, 426, 560,1731,2680,1785,2864,1949,2363, 403,3330,1415, # 4176
1279,2136,1697,2335, 204, 721,2097,3838,  90,6186,2085,2505, 191,3967, 124,2148, # 4192
1376,1798,1178,1107,1898,1405, 860,4359,1243,1272,2375,2983,1558,2456,1638, 113, # 4208
3621, 578,1923,2609, 880, 386,4130, 784,2186,2266,1422,2956,2172,1722, 497, 263, # 4224
2514,1267,2412,2610, 177,2703,3542, 774,1927,1344, 616,1432,1595,1018, 172,4360, # 4240
2325, 911,4361, 438,1468,3622, 794,3968,2024,2173,1681,1829,2957, 945, 895,3090, # 4256
 575,2212,2476, 475,2401,2681, 785,2744,1745,2293,2555,1975,3133,2865, 394,4668, # 4272
3839, 635,4131, 639, 202,1507,2195,2766,1345,1435,2572,3726,1908,1184,1181,2457, # 4288
3727,3134,4362, 843,2611, 437, 916,4669, 234, 769,1884,3046,3047,3623, 833,6187, # 4304
1639,2250,2402,1355,1185,2010,2047, 999, 525,1732,1290,1488,2612, 948,1578,3728, # 4320
2413,2477,1216,2725,2159, 334,3840,1328,3624,2921,1525,4132, 564,1056, 891,4363, # 4336
1444,1698,2385,2251,3729,1365,2281,2235,1717,6188, 864,3841,2515, 444, 527,2767, # 4352
2922,3625, 544, 461,6189, 566, 209,2437,3398,2098,1065,2068,3331,3626,3257,2137, # 4368  #last 512
)


site-packages/pip/_vendor/chardet/euckrprober.py000064400000003324150564745010016030 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import EUCKRDistributionAnalysis
from .mbcssm import EUCKR_SM_MODEL


class EUCKRProber(MultiByteCharSetProber):
    def __init__(self):
        super(EUCKRProber, self).__init__()
        self.coding_sm = CodingStateMachine(EUCKR_SM_MODEL)
        self.distribution_analyzer = EUCKRDistributionAnalysis()
        self.reset()

    @property
    def charset_name(self):
        return "EUC-KR"

    @property
    def language(self):
        return "Korean"
site-packages/pip/_vendor/chardet/__pycache__/euckrfreq.cpython-39.pyc000064400000027505150564745010021772 0ustar00a

��Re�4�@sdZdZdZdS)g@�0	(0	�
��xiti�i�i�iHia��i�i�i�i+i��Wiui��hii]i�i�i�i�i�i�i�i�i�i�iviwi�i�i��imiFi!ipi�ii��i�i�i�i�ixi��/i�i�i�i�i�i�i�i9ii�i��ti�i-iyi��Ki��i�i�i�i��iOi�ini�i�i�i�i�ii0�i��<i4�{i�i�i�i�i�iii�iri�i�i�i�i�i�i�i�i�i�i�iXiXi�i�i�i��Yi�i�i&i�iPi�i�i�i�i�i�i�i^i���i�i�i�i�i�i9i�i�i�i�i�i�iQi�i"�it�i��i�ii]i{i7i��i�i{i�i;i�iui�izi/i�i|��i�i�i�i7i�i�i.i�i�i�i�ii�ii�iiiiii{iiii#i|i}i�i~i�iti8ii_i	i
iii
i!iiii_ii�i�i�iii*i�iuii`i"i�ii|iii�i�iiiaiiiiiii?iiRi!i i/i�i�i!i"i=i�i�i#i�i$i%i&i'i(i)i,i�i'ibi$i*i�i+i�i-i�i,i�ii�i�i�i&iUi�i�i#ii�i-i.�'ii�ifi/isi�i0i�i�iii�i i�i9ie�[i1i�i�iiZi�i�i:i�i�i�i2i3i�iGii�i�yi4i�i�i�i5i6i7i,�wi�i�isi8i���i9i:ii�i~i;ii<i;i}i=i>i?ioi)i�i�i@i�iAiBi�i�i2iYiCiDi�i<iEiFiGiHiIi%iJiKiLiMiNiOi�`i>iPi�i=iQi�iRi�iSiTi;i�i�i�i�iiUi�iViWiXi4iY�iiZi[i�i\i]�ii^i�i�i_�"iPi�i�i`i;i�i~iHia�ivi��zi?ii�i�ibi�i�i<icid�dieific�0i�idiigiyi�i�ihii�si0iji=iki�ili�i�i�i<ibi�i��Uii�ii�iIiminioiipiqirisitiu�i�i�i�i6iviwi�i*i��]ixiyi�i�iziZi��-�:�ibi
i{i|�i&i'i�i�i5i�i�ii>i}i~iwiigi�i�i�i6i�i%i�i��(i��vi��i�iwi�i��i��Ei�i�i�i�i�i�i�i�i�if��iVi�i7i�i�i�i�iBi��Ni�i[i�i'i�i�i�i�i�i���iSi�i�i�i�i�iei�iixi�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i?i�i�ii�iqi�ifi�i(i)i�i�i~i�i\i�i�i�i�i�i)ii�i�ii�ii�i�i�i�i��$i��i��i�i�ili�i�i�i�~i�i�iC�i�ii�i@i�i�i�i�i��2i���iKiiziVii�����iQirifiii��i ihi+i3ii1�ii�i�i���gi(i�i�i�izi�i�i�i�i�i�i�i�i�i�i�i���i�i�i�i�iA�i<ijii�iMigi�i2i�i�i�i�i�i�i�i�i�iVihii�iJi�i�i0��i�b��i�i�ii�i�i��iY�i�i��i���ai!�*�i���i�i�iKiDi8iiRiiB�i@i�i�i�ii�i�i�iyi�i�i�i�iiXi:i�i�i#i�i�i�iii��Gi�i�iki�i�i=i�i��i��!i�i�i��J��i��=�i�i�i}i�iijii�i�i�i�i�i�i�i�i�i�i�iEi�i�i�i��ijiO�4i��i�i�i�i�i�i�i�i	ivi]i�iCi�i�i�i�i�i�i�ioi�i��i�i
i�i��iliciAi�i�i�i�i�i�i�i�i�Ti�i�ik�����i�i�i�i3i*i��qi���i>i�ii�i�i�ii+iiiii�;i��piixii��i�iiii�i�i	i�i
ii�ii
i�i���iliii�i�ii�)��i��imi8i�iiDii�iiiiii�ii�ii�7iL�Bi�iiDii�ii�i�itiiii i!i
i"i#i$i%iRi&i'i(i)i*�i+i,�i,i-i.�mi�i
i^i/ici�iEi�i�i�iaimi0iEi1i2i3ii�i�i4iTi�i�i�i5i�ii�i�i6i7iini�i�ioi8i9iFi�i�i�i:iGi;i<i=i>i?i@iAiBiCiDiEi$iFiGi�i�i%i�i�ipiHiIiJiKiLi�i�ii�i�i�i�iMiNiOiPiQiR�i�iSi/iTi�iUi�iiiVi��IiH��ii�i�i�iWi�iX�iqiYiZi[i\i�iri�isi]i^i_i`�iviLia�i��.i�ibiFi>i�i��jic�Zi�i�iBi6i�i`idieii��|iifi�i�i5igi�ihi�ii�H�iji�i�iki�ili�i1iminioipiqiriCisitiuiviwi�ixiyizi{i�i�i|i}i~ii�i�i��iii�iOi�i�i�i�i�i�ii�i�i�i�i�iEi�i�iqi�iiIi�i\i�i-i�i�iSi�i�i��
i�iei�i�i��li�iM���Qi�Piii^ii
i-iFi�i�i�ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i
i.i�iti�i�i�i�iJi�i�i�i�i�i�i�igi�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�ii�i���i�i�iui�i�i�i9i	i�i�i�i	i�i$i�i�i�i�i�5i%�ik�L�i�ii�i�i�i�ii�i�iAi�i�i�i�i�i�i�i�i��R�iui�i�i�ii�i)i�i�i:i�i�i��i���i�i"i$ivi�i���ciZ�i�i�i�i��i*iWiiKi�iLi��+�i���i�i�i�ii�i	iB��i�i?i��i�ii�i�i�i��iMi[i5i�i�i�ini���i�i�i��Ci�i'i�i�i�i�i�i�i�i�i
i�i�i�i	i�ii�i�i�i�i��FiTi/i�i�i�i���8i��ui�i�i�iKi�i(iM�i�ii�i�i�i�i��i�ii�i�iTi��?i��i�i�ii�i�ei i(i�i�i�i�i�i��%i0��i��Oi�i	i8i	i	i	i	i	i��ii	i	i	�i��#���i		i����Vi
�iiPi� �MiWi�i
	ii	i�i|i	i
	�Si��i�ii�i�i�i	i	i�i�i	i	i	i	i	i	i�i	i	��9�Dii���i��i	�3i�i�i	i	i	i{i	ii�i	i	i	i�i 	i��i�ii�i!	iki�i��i�iN��iYi�i"	i#	i$	i%	i&	i'	i(	i]i)	i*	i+	i,	i-	i.	i�i/	i�i��>i���ri1	i�i�i��i\i�i2	i�i3	i�i�i4	i5	i6	iwi7	i�i8	i�i3ii�i�i9	i:	iL����i;	��i<	i�ii�ii�i�ii�i�i�i=	i�iJi>	i?	i@	i�iA	��iB	iC	iD	i�iE	i�i�iF	iG	iH	iI	i�i�i�i�i�i�iJ	iK	ii�iL	iM	iN	iiIiO	iP	i�iQ	iR	iS	i�iT	i�iU	iV	iW	iX	i�iY	iZ	i[	i\	i]	i�i�i^	i_	i`	ia	i�iib	ic	id	ie	ii�if	ig	iUi�ixi ih	i�i�ii	ij	�ii�i4i&��i�i�i!i���i�iSiyii�i�i"i
i�i#ik	i�il	im	in	�i�i�ipiQi�i.io	iU�iOi�ip	i�iq	ir	is	ixit	i�iu	iv	iw	iii$ix	i�iy	i�i`iz	i{	i|	i}	��i~	i	i�	i�	i�	i�	i�	i�	��iCi��i�i�i%iyi�iioi�	i�i�	i�	i�	i�	i�	i�	i�i�	i�	i�	i�	i�i�i�	i�i�i�	i�	i�@i�i�i�	iGi�	i�i�ii2��i�i�i�i�	i�	i�	i�	iNi�	ii�i�i�i�i�	i�	i�	i�	i�	�ri�	i�izin�i�i�iPi�	i#i�	i&iQi�	i�imi�	ici�i�i�	i'i�	iwi�	i�	��i(ili@i�i�i�i)i�i*i�	i�	i�	i�	i�i�	i�	i�	i�	i�i�i�	i�i�i�	i�	i�	iHi�i�	i�	ii�	ii�i�	i�	i�	i�i�	i�	iri�	ii��Ai�	i���i�i�i�	�}i,i�i�i:i�ii��i
iiIi�	iNi�i1i�	iW����i�i�i�i�i�i�	i+i�	i�	�1i�ibi�i�	iqi�	i,i�	i��iei�i��_iidi�	i�	i�	i�	i�	i�	�i�i2i�i�i�i�	�i�	i�	i�	i�	ii�	i�i�	i�	i�i-i�	i�i�i�	i�	i�	i�	i�	i�	i�i�	i�	i�	i�	i�	i�	i�i�	i�i�	i�	���i7i�	i�i�	ii�	i�	i�	i�	i�	i�	i�	��i�i�i�i�i�i����,iGi�ii�i�	i�i^i�	i�	i.i�	�igi�i�	i�i�	i�	i�	i�i�i�	ii�ii_i�i�	i�	i�	i��ihihii{i�	i|i�i��i3i�	i�	i�	i�i�i�	i
i
�\ii�i�ii
i
i
i
i
i
i
�i	
i

i
i
ii�i�i
iji
iii�i4ii
i
i
i
i
i
i��i
i
i
i
i��6id��i/i
i
ii�i��oi
i�iRi��i��	�&i
�k�niz���i�i�i
i�i
i����Xi�idiiSi�i}i
i}ii�i~i�i�ii�i 
i!
iNi"
i�i��i�i�i�i#
i��isi$
i%
i&
�fiDii1i'
i�ii(
i@i)
�^���i�i*
i�i+
i,
i-
i.
i/
iJi+i0
i1
i2
i�iTi�i3
i4
i5
i6
��i7
iAi.i�i�i�i8
i9
ii"i:
i;
i�i<
i=
i>
ii�i0i?
i�i@
iiA
iB
iiC
iD
iE
iF
iG
i�i�i_i[iH
iI
i�ii`iaiJ
i�ii�i�iK
iL
iM
iN
�iO
i�iiiiiiP
iQ
iR
N)� EUCKR_TYPICAL_DISTRIBUTION_RATIO�EUCKR_TABLE_SIZE�EUCKR_CHAR_TO_FREQ_ORDER�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/euckrfreq.py�<module>)ssite-packages/pip/_vendor/chardet/__pycache__/charsetgroupprober.cpython-39.pyc000064400000004373150564745010023721 0ustar00a

��Re��@s,ddlmZddlmZGdd�de�ZdS)�)�ProbingState)�
CharSetProbercsReZdZd�fdd�	Z�fdd�Zedd��Zedd	��Zd
d�Zdd
�Z	�Z
S)�CharSetGroupProberNcs(tt|�j|d�d|_g|_d|_dS)N)�lang_filter�)�superr�__init__�_active_num�probers�_best_guess_prober)�selfr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/charsetgroupprober.pyr!szCharSetGroupProber.__init__csJtt|���d|_|jD]$}|r|��d|_|jd7_qd|_dS)NrTr)rr�resetr	r
�activer)r�proberr
rrr's
zCharSetGroupProber.resetcCs |js|��|jsdS|jjS�N)r�get_confidence�charset_name�rrrrr1s
zCharSetGroupProber.charset_namecCs |js|��|jsdS|jjSr)rr�languagerrrrr9s
zCharSetGroupProber.languagecCs�|jD]~}|sq|jsq|�|�}|s(q|tjkrJ||_tj|_|jS|tjkrd|_|j	d8_	|j	dkrtj|_|jSq|jS)NFrr)
r
r�feedr�FOUND_ITr�_state�state�NOT_MEr	)r�byte_strrrrrrrAs&





zCharSetGroupProber.feedcCs�|j}|tjkrdS|tjkr"dSd}d|_|jD]R}|s<q2|jsT|j�d|j	�q2|�
�}|j�d|j	|j|�||kr2|}||_q2|js�dS|S)Ng�G�z��?g{�G�z�?gz
%s not activez%s %s confidence = %s)rrrrrr
r�logger�debugrrr)rr�	best_confr�confrrrrVs*


z!CharSetGroupProber.get_confidence)N)�__name__�
__module__�__qualname__rr�propertyrrrr�
__classcell__rrr
rr s


rN)�enumsr�
charsetproberrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/mbcssm.cpython-39.pyc000064400000036614150564745010021270 0ustar00a

��Re�c��@sl	ddlmZdZejejejdejejejejejejejejejejejejejejejejejejejejfZdZedeedd�ZdZ	ejejdejejejd	dejd
ejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejfFZ
dZe	de
ed
d�ZdZ
dd	ddejejejejejejejejejejejejejejejejejejejejejejejejejejdejdejejejejejejejf(ZdZe
d
eedd�ZdZejejdejejejejejejejejejejejejejfZdZed	eedd�ZdZejejejdddd	ejejejejejejejejejejejejejejejejejejejejejejejejejdejejejejejejejejejejejejejejejf0ZdZedeedd�ZdZejejejejejejdejejejejejejejejejejejejejejejejejd	ejejejejejejejejejdejejejejejejejejejejejejejf0ZdZedeedd�ZdZejejejdejejejejejejejejejejejejejejejejejejejejfZdZed
eedd�Z dZ!dddejd	dejejejejejejejejejejejejd
d
d
d
ejejd
d
d
d
d
ejd
d
d
d
d
d
dddejddd
d
ejd
d
d
d
d
d
d
ejejejejf8Z"d Z#e!d
e"e#d!d�Z$dZ%d
d
dd
d	dejejejejejejejejejejejejdddejejejdddejdejd
d
dd
dddddejdddejejejdddddejdejejejf8Z&d"Z'e%d
e&e'd#d�Z(d$Z)ejejejejejejd%dd&d'ddd
dd	dejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejddddejejejejejejejejejejejejejdddejejejejejejejejejejejejddddejejejejejejejejejejejejejejddejejejejejejejejejejejejd&d&d&d&ejejejejejejejejejejejejejejejd&ejejejejejejejejejejejejd%d%d%d%ejejejejejejejejejejejejejejejd%ejejejejejejejejejejejejd%d%d%ejejejejejejejejejejejejejejejejejejejejejejejejejejejf�Z*d(Z+e)d)e*e+d*d�Z,d+S),�)�MachineState(rrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrr��Big5)�class_table�class_factor�state_table�char_len_table�name(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrr�	rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)
rrrrrrrrrr�
�CP949(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrzEUC-JP(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrzEUC-KR(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrrrzx-euc-tw(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrr�GB2312(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrr�	Shift_JIS(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)rrrrrrzUTF-16BE)rrrrrrzUTF-16LE(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrr��
rr��rrrrr)rrrrrrrrrrrrrrrr�zUTF-8N)-�enumsr�BIG5_CLS�ERROR�START�ITS_ME�BIG5_ST�BIG5_CHAR_LEN_TABLE�
BIG5_SM_MODEL�	CP949_CLS�CP949_ST�CP949_CHAR_LEN_TABLE�CP949_SM_MODEL�	EUCJP_CLS�EUCJP_ST�EUCJP_CHAR_LEN_TABLE�EUCJP_SM_MODEL�	EUCKR_CLS�EUCKR_ST�EUCKR_CHAR_LEN_TABLE�EUCKR_SM_MODEL�	EUCTW_CLS�EUCTW_ST�EUCTW_CHAR_LEN_TABLE�EUCTW_SM_MODEL�
GB2312_CLS�	GB2312_ST�GB2312_CHAR_LEN_TABLE�GB2312_SM_MODEL�SJIS_CLS�SJIS_ST�SJIS_CHAR_LEN_TABLE�
SJIS_SM_MODEL�
UCS2BE_CLS�	UCS2BE_ST�UCS2BE_CHAR_LEN_TABLE�UCS2BE_SM_MODEL�
UCS2LE_CLS�	UCS2LE_ST�UCS2LE_CHAR_LEN_TABLE�UCS2LE_SM_MODEL�UTF8_CLS�UTF8_ST�UTF8_CHAR_LEN_TABLE�
UTF8_SM_MODEL�rHrH��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/mbcssm.py�<module>sD$  �� ((((((��$  ��$ ��$    �	�$   ��&  ��$ �
�$ �
�$               ��site-packages/pip/_vendor/chardet/__pycache__/utf8prober.cpython-39.pyc000064400000003756150564745010022105 0ustar00a

��Re�
�@sHddlmZddlmZmZddlmZddlmZGdd�de�Z	dS)�)�
CharSetProber)�ProbingState�MachineState)�CodingStateMachine)�
UTF8_SM_MODELcsTeZdZdZ�fdd�Z�fdd�Zedd��Zedd	��Zd
d�Z	dd
�Z
�ZS)�
UTF8Proberg�?cs*tt|���tt�|_d|_|��dS)N)�superr�__init__rr�	coding_sm�
_num_mb_chars�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/utf8prober.pyr	&s
zUTF8Prober.__init__cs"tt|���|j��d|_dS)N�)rrrr
rr
rrrr,s
zUTF8Prober.resetcCsdS)Nzutf-8rr
rrr�charset_name1szUTF8Prober.charset_namecCsdS)N�rr
rrr�language5szUTF8Prober.languagecCs�|D]f}|j�|�}|tjkr,tj|_qlq|tjkrDtj|_qlq|tj	kr|j�
�dkr|jd7_q|jtj
kr�|��|jkr�tj|_|jS)N�r)r
�
next_stater�ERRORr�NOT_ME�_state�ITS_ME�FOUND_IT�START�get_current_charlenr�state�	DETECTING�get_confidence�SHORTCUT_THRESHOLD)r�byte_str�c�coding_staterrr�feed9s


zUTF8Prober.feedcCs.d}|jdkr&||j|j9}d|S|SdS)Ng�G�z��?�g�?)r�
ONE_CHAR_PROB)r�unlikerrrr"Ls

zUTF8Prober.get_confidence)�__name__�
__module__�__qualname__r)r	r�propertyrrr'r"�
__classcell__rrrrr#s

rN)
�
charsetproberr�enumsrr�codingstatemachiner�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/jisfreq.cpython-39.pyc000064400000053235150564745010021445 0ustar00a

��Re�d�@sdZdZdZdS)g@i(�(�����i'iOii}i�i�i�i�i]i�
i�
ii�
i�i�i�i�i�i�i���i�iXi}i�i�ik�ig
i�
i�iki�����iiiii�i�i%i&�0�1�,�-i�i�iii�ii�i�i�i<ii�ipi�i�i�i�i�i�igi�ii�iWiXi�i�ih
i"i�i�i�i�i
i	ii\i�i�i�
i/
i�i�i�i�ii�i0
i�i�
ihi�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�ii�i�i�i�i�i�i�iiiiiiiiii	i
iii
iiiiiiiiiiiiiiiiiii ii!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i�iViji4i�ii�iBii�i8i9i:i;i<i=i>ivi�i�i�iSi�ieiifi�i�id	i+i�i�iaiwi�i�i�iIi�i~i�i�
ii?i@iAiBiCiDi
ii-iiiii�i�
iiEi^i.i�i3iii
iFi/iYi�i�ii�iji�
iiGiHiIiJiq�i1
�iY�ik�/i�2���#i����*�������[��\�5i��!i!�	�%�@il��'iA����4�
�������<i��i-�ii7�S�i��~iKi=��E��;���7i7i8�m�&���
��OiK�=i~id��iLiMiNiOiPiQiRiSiTiUiV�Y�>�J�"ip	�p���i��Ti_�i��.�X�i��Lij�ei�9�P� il�yi��i�D���i��)ihi��F�?��+iiig�ci���B�]iN����8�j�:i5iii7���Ri4�G�d���i��ni�hit�6�3�$�W�Ci��i:
ixi��i*iV	iiWiXiYiZi[i\i]i^i_i`iiaibicidieifigihiiijikiliminioipiqirisiti�
iuiviwixiyizi{i|i�i�ii�ili}i�i~iii	i�i�
i�ii�i
i�i�ii�iiii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�iiiiiiiki�i�
i>i�
i�
i	i�i
	ij
i��i�iZi[i\i]iiiiii	i
iii
iiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�iiiiiiiiii	i
iii
iiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�iiiiiiiiii	i
iii
iiiiiiiiiiiiiiiiiii i!i"i#i$i%i&i'i(i)i*i+i,i-i.i/i0i1i2i3i4i5i6i7i8i9i:i;i<i=i>i?i@iAiBiCiDiEiFiGiHiIiJiKiLiMiNiOiPiQiRiSiTiUiViWiXiYiZi[i\i]i^i_i`iaibicidieifigihiiijikiliminioipiqirisitiuiviwixiyizi{i|i}i~ii�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�i�ii
i�i�i�i�ii�i2
i�i�i{iTi�ia
i�
i^i�i�i|i�i=iimik
i�i�i�i	iViyi�	i�i
iq	ii�ii�
i�	�i�i�i�i!i�i�	i�ii�i��iCiOi3ii�i�i�i�ir	i�i�ini�i+ii�i�i�
ipiqi�i;
�Ai�i�i�iCi0i�ini�
i�i�iTiCioi
i�i
i#	ii�i~iiPi�i�i�ii�i�	ii�	i�i$	izi�i�i&i�i�i�i�ii�i�i�i�i_i�iUi�i�
is	i�	i�i�iIi�i�i�ii�i�i�i3
i�ii$iWi1i5iPi�	i�ii�i�i�i�iXi�i�i�il
i�i�iHi�i�	ii�i�ii�i
i�
iWi�i~ipii$i�iii@iLi�i�	i�i�i�iuiiiJiI	iW	i�i�i_i�i�	i<i�i�ii]iDi�i�i�iihi�i�i�i�i���fi-i��}it	i�i�	i��i�i�i,��i
ibi�i�iXi�iiFii{i`i�i�i+i3iqi�imi�i4
i�i�i�i�i�iBiliX	i�
ii%
i�i�i%	i2iui�i�	i�	i�iJii|iri�i@ii�iii�i�i�
i�i�	iL
iiiii�i�ib
iJ	iici�i�i��i�i�ii4	iRi�im
iTi�i�ii�i�i�i�i�ie	i�ii?i
i�i�i�ii-i	i�i*iEi�i�i+i�i�i&	i�	i�i�i/i�i�i.isi[i�i+ii�
ii�
iBic
i�i�i�i�i�i	i)ii,i�iu	ii	ii9i&
iii�i�i�iv	i�iYi�i3i�i
i�i�iiDi�i�i�i�idi�i�i|i(iyii�ii�i<i8i�	iiiiiid
i5
iBi�inii�ii�i�i|i�	i!i�i)i@in
imii�	i0iai�i�i�	�zi'��i�i�	iWi�i[ivi�i�i�	i�i�i	io
ii�	i�i
ii�i'
i�ii"i�i�i9ii�ii�iJi�i�ihi�i�ie
i�i\i��iip
i@iiQi9i�i���i	�i�i�iw	iiiEi�iJiIi�	i�i�i�i�iifi�i�i�i
i�i1i�i�i
i�i�i'	iri
i�i�i�i�
i�i�ici�	iqi�ixi�ii
i2i�i�ii�i�	i*i�i�ibiFi�ivi�idi�
i�i�i!iRi�i�iQi�ii�iPi�iii�ii�i�i(	i�i�i_ii`i�i�if
i�i6
i�iAii�iai�i�iuiTii2i�i�iii�ii^iCi�i�
i�i[i�i�
ivi%i�i�i�ini!iFi�iZi^i�ii�idii�i�ix	i�i�i]i�iBiici�
i�i�i�iPiqi	iiHiY	i�i�i(
i�i�
i�igi}iKi�	i�ii4i�iWi�i�i�iCig
i-idi�ikii4i}i�iq
i~i+i�	i�i�ihi@	i�	ii@iA	iDi:i�i�iei�
i�iqiwi�i�i�i�i�i�i�idi�iZi�i*iM
i[i�i�isiZi�
i�i�i>i�i'i-i�i�
iii�i)	ii�i�i�itiFi7
i�iti#i�i�i�i�iiYi�i�ifikibi1i�i�
i�i6ioiciiqi�	i�i�i�i�
iui�i:i�i�iN
i�iUi=i�	ivi�il�Ni�i�i�i]i�i;i�i�i�
i�i�iliLi	i}i�i�ii�iB	�i8
iQi�i#i`iTi�i�	ixiiri�	iii�i�i�i�	i�i3inii�ii�ii�i{i�i�i@i�i�ii�i9
��i�i5ih
idiZ	��i�iiii�i�i�iisi-ir
i�iii�i[i�i0iii�i�i.iSi	iRiO
i�ii�i�i;i�i4iy	i�
i;i�iz	i�ipi�i�i i�i�i:
i(i|i�
i�i"i�i�i�i�i�i�iOi�iei�i��i
iC	iti)
ii6�Ki�i�i8i�iP
iihi�i�i�ii�i3i$i�i�i*
i=�bieii�
i4i�i�i�	i.i7i�i\i�i�iiji�ii�i�i�i�i��i/i�i�i�i=i�i^i�i5	i�i�i�	i\i�ii�i�i9i#i+i�
is
i,i�i7ii�iYi�i�i�
iMi�i+
ii
iQi��i6	i�i�i�i�iiyi�iisi#i{	i�
iUifi<i�i�ivi�i�i)i�
ii�ii�i�imi<iii�	i�i�igi�i�	�li�i iD	ibiui�
i�iDiBi�i�i�
i[	i�ii8i�i�i>ii�ii�i�i�i�ii]i�iDi���i�i6iZi�i�	i5i�i�
i�
i<imi�
i,i�	i�iui^i�	i�i�igi�i��Ii�i�igi�i�i\i:i
iM�ti�	i�iEi��o��i�i�i�iEiRi�ii�iEij
i��giWi�i�iKi�	iCi�i=i�i]i$i!i�i�	ii�i��`iKiviii0ii��^��i�i3i�i"ii���i�i�aik
ii;
iwi�i�i�i�i�i�i�
iyii�i�i�iPii�i�	iwi��i�it
i�i�i�ii�i�iJiaii�i]i�i�i�i�	�ii�iiLihii�i�i�
iu
iji�ii#iCi�i�i	ia��i!�si|	i�i	iYi�i
i�i�i�ihiOi5i�i�iziii�
iSi�iLi�i.i�i&ii	i<
i7i�isiAi�i�iMi�
i�isi�i*i,
i%i�
i�	i�iki�i&if��iOiji"i�i(i-i�i[i�i�i�i�i-i=i�i}	i�i��MiSi�i�i�ii�iii�i�i�i�
i
ii
i�i~	i�iJitikivieiyi��i�i<
i�ifi�iM�ki�
ii�ioi���i�iwiv
ili�i]i�ii.i�
i�i�i�i/if	iqi$igi�iG��ini�i>i6i=
i�i�i�i�i�iNi�	ixi�iei*	iihii�i�i$i�i�i�ioibi,ii�i�iiti�i�i	i�ii�
iVi�i�
iiii�i�i�iw
iiipiHi��Vi�iii<iZi�	i8�riwi�i&��i�i�i/i�i�i�i�i>i�i>
iE	ixi�i&iei�iwi	i�i�i5i�i�i�
ii\i��i�iSiiii��i�i2�i�iii+i'i�i%i�i�ix
iOi�i
iQ
i�i�iii�i�i_�Hi�i�ii�igii�iy
i�i�
i0i&i+	ii�ii�i�il
i'i�i'izi�i	i	i�i_i�i�ii�i�i�i?
i�
i�i�	�iz
i�im
i|i
ii	ig	i0i�i�i i�i*i�i�i
i`iwii�i
i#��ici)i�i�iR
i���i�iT�i�iriiVi�i�i�i_ii�ii�iri�i�i�
i`ixi�i�i
i�i�ifi�iin
i=
i*iAii�	i(��ix�iS
i��{iT
i�i9ii�iMi�ir��ii;ii(ii%i�i[iaibii�iDi�i�ii-
i\	i�iii�io
iEimi)i�i!i�iici�i,	i1ii]	i��i�i>iI�i�i�	iTii�i�i�i�i�iyiikizixi�iNi	i�i�i�i	i5i�i�i"i�iHi<idi�i
i�i�ih	i�	i�i
ip
i�i�i:iLi1i�i�i�
ii�i>
i�iq
i?iZiMiIi�iiBiK	iDi�i#i
i
iYi>�i�iii`i�i�i�i�ii�i�i�ii�i�i�i i�i�ici�ii�i�i���iTi�iaii2iyi/i�i"iU
����ibi�i�i�i/i�i�	i�ii
i{i
i�ij�i��i5i*i�i�i�i�iaixi�i�i�	iiFi�i�i�i7i�i�i�	i?i��wi�iii�i�i+i�i8i�i�
iNi�
i�i�i,i�ini�i
ii�ioisi_i?i�i�	i�i�ii�	i�i�i?i0i�	i�	i�iyi3ifi�i=i�i�ilii�	iivi]i�i	i	i�i�izi�
i�iyijiiui�i�ipi�ij�ii	
iiizii�ir
ii�i�i�i�i�iiiFi�i'i�	i{
iniix�ui$ii�i�iMi��i�ikiipi�
i^	iqi
i`i|i�i�	ii

i|
i�	i�
ii�i�	i
i�i�i�i�	iGii�i-i�iZi1iYi�i-	i	iei�i�i[iC�i�i�iei�i.i-ii�i�i	i�i�i�
i�
i7	ibili	�i�i�i{i�i.i	i�i�izi�i�
i�
i�i?
i�ii9ii�
i�ii�i}i�i�
iai�i�	i
�Qi�	idi;iV
i�i/i^iDi�i?imi�
i�
i9iIi�
i�i�i
ii�i�
i{i}i i�ii�i��i&i�i~i�iXi�i�i}
i�
i�i9i�i�i:i�i"i\ii�i
i_	��i8iWi~
i^i%iPisi�iwi�i�ii8	i�iLi�ii.
i�i�i�
i�iiN��i�iPi�iSi�i�i�i:iRi'i0ibii�i�
i�i�i�i�izii�	iiiiiKi���i�i�
iri�iqi
i�i�iniL	ii�ii�i�i�i�i�i�iXi@
i�ig�i�i�i4i�i:ii�i�iri�	i�iQi�
ii	i;i'i�i>i�ibimi�iw��i�i�is
i�ii�iA
i�it
i9	ii�iii�	ioi^i�i�i2i�ici�i��i�iiri�i�i�i!iitii1i{i�i�i�i��i�isii�i�i�i�i��xi�viaipi�i�i
i�iLiOi�iiUi�i.i��i)i.	i�iEi�	i"iiyii	i�i�i��i(i�
ii�iqiEi6i
i�
i2iui6i�i�i�i�i.i�ii	ii�iVi�iisi�
i�i�iGiGijiti�iWii�	i�i�i`i�i�i�iui�ii�
i�i{iQii�	iOiCi�i�i�
i�iJi�	i	i	i$i�i�i�i/	i�ii_ii�iij	iRi�i7ii�i/
ii3iDiu
i6iXi4i�i�iii�ii>i�iii0
i`i`	i
i�i�i�iUi�i i�
irii�i)i�i5i�iHi�i 
i�ii~iri�i�
i�i�i�i�i�	�ii]i�i�ii�ii�ii@
i�i�i�
iai�ivi�i�i�i/i!i6i�i�i�i7iwi�i	i3ici8ixii�i�iB
i�i�iM	i��iF	i�igii�i�i(i�i
i:	ii�iHi�i�
if��i�i"i�
i~i�i�i�i�
iyiii�i�i i!iNi�
i4iA
i�idie���izii�i�i�i�i�i�i0	i�i�ii�i�i�ii�i{i�i�i�i�
i�ifi(ili�iGi�	i���i^iB
imiFi|iiUi�i�i2iRi�i�i#i�i�
i�i�iF��i�Ui�i�i�i1i�iciki�iN	iSi�	���i�
i�i�i$iO	i�iiYi�i�i�i
i�iGiii�iW
i	iC
i?i�i�i~i�i�iUiFi�i�i�ii�i�i�i�
iC
i�	i�i�iiei
i�i�i�ii/iQi_iv
i;i�i�iPi)��iGimi�	i�i�iGi,iIiz�_i�	��i	i9ini�i�
iAi�	i�
iX
i�i#iVi)i<i%i�i}i�i�iHiw
i5i�ipi.ii�i�imii�i�i�
i#ihi|ibi�i�i�i@i(i^izi�iiAi�
i ii"iigi�i|i�i�i��i�i6i�iXi�i�i,i�i�i		ii�i$i#iIik	i�ii�i�i:i	i�iQiioihiVi$i�
i�iti�i�i�iY
i,i�ii
i�iD
i�i�	i~i�
i�
i�i�i%i�ii i��i�ii�i�
i i7i/i
	�i���i�	iEiVi�i�
iE
iui�i&i�i�ii�i�	i�ioi�i�iZi�iii�i�i	ii�i��i!i�ini8i�i'i��i�ii�ii�i�	i�iGi�iRi2i�ii:i�i,��iiZ
ili�i(iXiP	i?i�ii�	ioi�ii�i�i�i�iWi*i��i;i8i�i&i)i�i�i
ioi�ii�i�iHi_i�i�i0i�ix
i�i0i�i;	i�i
i�i�iXi�i	��i�i1i��Zi*i%i�	�i�|idi`ii�iSiji}i\ii�i�iG	i�ii�	if�qi%iBi�i1
ipi�i"ii�i�i�i�i|i�i�ii�	i�il	i2
�i�
i�
ii�i@ihi�i;i��ii	i�i	i�i�i&ii�i�i}i�i%i�i�iii?i�i�	i�ia	iy
ii�
i�i�i�	i�i=i1i�i<i�i{i#i��i�i�i�
iAi�i
i�iti�i�i�	i�i>i
iKi3
i�i�i=��ii\i�i�i'iAi+igi�ib	iKi�i�i�i�i
i�i
i�i4
i�i*i�im	i�	i�i�
ioiNii0i(iii�i$i4i i{ii�i�iQ	i�i�iUi�i�i�i,i`ii�	i�ii�
iji)i i�i-i6��i�	iF
i2i)ii
i*i�iYN)�JIS_TYPICAL_DISTRIBUTION_RATIO�JIS_TABLE_SIZE�JIS_CHAR_TO_FREQ_ORDER�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/jisfreq.py�<module>,ssite-packages/pip/_vendor/chardet/__pycache__/charsetprober.cpython-39.pyc000064400000006701150564745010022641 0ustar00a

��Re��@s0ddlZddlZddlmZGdd�de�ZdS)�N�)�ProbingStatec@sneZdZdZddd�Zdd�Zedd��Zd	d
�Zedd��Z	d
d�Z
edd��Zedd��Z
edd��ZdS)�
CharSetProbergffffff�?NcCsd|_||_t�t�|_dS�N)�_state�lang_filter�logging�	getLogger�__name__�logger)�selfr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/charsetprober.py�__init__'szCharSetProber.__init__cCstj|_dSr)r�	DETECTINGr�rr
r
r�reset,szCharSetProber.resetcCsdSrr
rr
r
r�charset_name/szCharSetProber.charset_namecCsdSrr
)r�bufr
r
r�feed3szCharSetProber.feedcCs|jSr)rrr
r
r�state6szCharSetProber.statecCsdS)Ngr
rr
r
r�get_confidence:szCharSetProber.get_confidencecCst�dd|�}|S)Ns([-])+� )�re�sub)rr
r
r�filter_high_byte_only=sz#CharSetProber.filter_high_byte_onlycCs\t�}t�d|�}|D]@}|�|dd��|dd�}|��sL|dkrLd}|�|�q|S)u9
        We define three types of bytes:
        alphabet: english alphabets [a-zA-Z]
        international: international characters [€-ÿ]
        marker: everything else [^a-zA-Z€-ÿ]

        The input buffer can be thought to contain a series of words delimited
        by markers. This function works to filter all words that contain at
        least one international character. All contiguous sequences of markers
        are replaced by a single space ascii character.

        This filter applies to all scripts which do not use English characters.
        s%[a-zA-Z]*[�-�]+[a-zA-Z]*[^a-zA-Z�-�]?N�����r)�	bytearrayr�findall�extend�isalpha)r�filtered�words�word�	last_charr
r
r�filter_international_wordsBs�z(CharSetProber.filter_international_wordscCs�t�}d}d}tt|��D]n}|||d�}|dkr<d}n|dkrHd}|dkr|��s||kr�|s�|�|||��|�d�|d}q|s�|�||d	��|S)
a�
        Returns a copy of ``buf`` that retains only the sequences of English
        alphabet and high byte characters that are not between <> characters.
        Also retains English alphabet and high byte characters immediately
        before occurrences of >.

        This filter can be applied to all scripts which contain both English
        characters and extended ASCII characters, but is currently only used by
        ``Latin1Prober``.
        Frr�>�<TrrN)r�range�lenr!r )rr"�in_tag�prev�curr�buf_charr
r
r�filter_with_english_lettersgs"

z)CharSetProber.filter_with_english_letters)N)r
�
__module__�__qualname__�SHORTCUT_THRESHOLDrr�propertyrrrr�staticmethodrr&r/r
r
r
rr#s




$r)rr�enumsr�objectrr
r
r
r�<module>ssite-packages/pip/_vendor/chardet/__pycache__/mbcharsetprober.cpython-39.pyc000064400000004374150564745010023164 0ustar00a

��ReU
�@s0ddlmZddlmZmZGdd�de�ZdS)�)�
CharSetProber)�ProbingState�MachineStatecsVeZdZdZd�fdd�	Z�fdd�Zedd��Zed	d
��Zdd�Z	d
d�Z
�ZS)�MultiByteCharSetProberz 
    MultiByteCharSetProber
    Ncs,tt|�j|d�d|_d|_ddg|_dS)N)�lang_filter�)�superr�__init__�distribution_analyzer�	coding_sm�
_last_char)�selfr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/mbcharsetprober.pyr	'szMultiByteCharSetProber.__init__cs<tt|���|jr|j��|jr.|j��ddg|_dS)Nr)rr�resetrr
r�r
rrrr-s

zMultiByteCharSetProber.resetcCst�dS�N��NotImplementedErrorrrrr�charset_name5sz#MultiByteCharSetProber.charset_namecCst�dSrrrrrr�language9szMultiByteCharSetProber.languagecCstt|��D]�}|j�||�}|tjkrN|j�d|j|j	|�t
j|_q�q|tj
krft
j|_q�q|tjkr|j��}|dkr�|d|jd<|j�|j|�q|j�||d|d�|�q|d|jd<|jt
jkr�|j��r�|��|jkr�t
j|_|jS)Nz!%s %s prober hit error at byte %srr���)�range�lenr�
next_stater�ERROR�logger�debugrrr�NOT_ME�_state�ITS_ME�FOUND_IT�START�get_current_charlenrr
�feed�state�	DETECTING�got_enough_data�get_confidence�SHORTCUT_THRESHOLD)r
�byte_str�i�coding_state�char_lenrrrr&=s4

�


�
�zMultiByteCharSetProber.feedcCs
|j��Sr)r
r*rrrrr*Zsz%MultiByteCharSetProber.get_confidence)N)�__name__�
__module__�__qualname__�__doc__r	r�propertyrrr&r*�
__classcell__rrrrr"s

rN)�
charsetproberr�enumsrrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/langhungarianmodel.cpython-39.pyc000064400000052441150564745010023637 0ustar00a

��ReL��@sP$ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#dddddddd$d%d&dd'd(d)d*d+d,d-d.d/d0d1d2d3d4dd5d6d7d8d9d:dddddd;d<d=d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddd�dddddd d!d"d#dddddddd$d%d&dd'd(d)d*d+d,d-d.d/d0d1d2d3d4dd5d6d7d8d9d:dddddd�d�d;d<d=d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdPdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dad�dcdddedfdgdhdidjd�dldmdndodpdqdrdsdtdudvdwdxdydzd{d|dkd~dOd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dQd�d�d�d�d�d�d�d�d�dd�dd�d�d}d�d�d�dbd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel���)?��(�6�-� �2�1�&�'�5�$�)�"�#�/�.�+�!�%�9�0�7�4r���r����	����
����
�r���>���3�,�=�:�;�<�?��������*�8����rrrr	r
rrr
rrrrrrrr�Hrrrrr�@�Drrrrrr r!r"r#r$r%r&r'r(r)r*�Cr+r,r-r.�Ar/r0r1���������������������N��E���������������������L�����������������Q��������������������������������r2�S���P��������r3������r4����������r5���Br6������r7�Fr8�������Tr9�K���G�R���I��r:�U�O�Vr;�M�W���r<�Jr@r=���r>rAr?�(rrrrr(r,r&r%r)r#r+r1r!r'r9r:r0rrr.r"r-r$r*r=r<rr rr?r;r>r
rrrrrr
rrrr@rr3r	rrrrrr2rrrrrArr5r6r7r4r/r8rGrJr�rIrHrar�r�rFr�r�r�rrr�r_r�r�r{r�r�r�r�r�r��X�Y�Z�[�\�]�^�_�`�a�b�c�d�e�f�g�h�i�j�k�l�m�n�o�p�q�r�s�t�u�v�w�x�y�z�{�|�}�~����������������������������������rKrLrMrNrOrPrQrRrSrTrUrVrWrXrYrZr[r\r]r^r`rbrcrdrerfrgrhrirjrkrlrmrnrorprqrsrtrurvrwrxryrzr|r}r~rr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rErDrCrBzwindows-1250Z	Hungariang��(��P�?TuPABCDEFGHIJKLMNOPRSTUVZabcdefghijklmnoprstuvzÁÉÍÓÖÚÜáéíóöúüŐőŰű)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetrrz
ISO-8859-2N)�#pip._vendor.chardet.sbcharsetproberrZHUNGARIAN_LANG_MODELZ$WINDOWS_1250_HUNGARIAN_CHAR_TO_ORDERZWINDOWS_1250_HUNGARIAN_MODELZ"ISO_8859_2_HUNGARIAN_CHAR_TO_ORDERZISO_8859_2_HUNGARIAN_MODEL�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langhungarianmodel.py�<module>s4$
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���	���site-packages/pip/_vendor/chardet/__pycache__/mbcsgroupprober.cpython-39.pyc000064400000002221150564745010023202 0ustar00a

��Re��@s�ddlmZddlmZddlmZddlmZddlm	Z	ddl
mZddlm
Z
ddlmZdd	lmZGd
d�de�ZdS)
�)�CharSetGroupProber)�
UTF8Prober)�
SJISProber)�EUCJPProber)�GB2312Prober)�EUCKRProber)�CP949Prober)�
Big5Prober)�EUCTWProbercseZdZd�fdd�	Z�ZS)�MBCSGroupProberNcsDtt|�j|d�t�t�t�t�t�t�t	�t
�g|_|��dS)N)�lang_filter)
�superr�__init__rrrrrrr	r
�probers�reset)�selfr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/mbcsgroupprober.pyr*s�
zMBCSGroupProber.__init__)N)�__name__�
__module__�__qualname__r�
__classcell__rrrrr)srN)�charsetgroupproberr�
utf8proberr�
sjisproberr�eucjpproberr�gb2312proberr�euckrproberr�cp949proberr�
big5proberr	�euctwproberr
rrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/euckrprober.cpython-39.pyc000064400000002234150564745010022316 0ustar00a

��Re��@sDddlmZddlmZddlmZddlmZGdd�de�ZdS)�)�MultiByteCharSetProber)�CodingStateMachine)�EUCKRDistributionAnalysis)�EUCKR_SM_MODELcs4eZdZ�fdd�Zedd��Zedd��Z�ZS)�EUCKRProbercs,tt|���tt�|_t�|_|��dS)N)	�superr�__init__rr�	coding_smr�distribution_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/euckrprober.pyr#s
zEUCKRProber.__init__cCsdS)NzEUC-KRrrrrr�charset_name)szEUCKRProber.charset_namecCsdS)N�Koreanrrrrr�language-szEUCKRProber.language)�__name__�
__module__�__qualname__r�propertyrr�
__classcell__rrrrr"s

rN)	�mbcharsetproberr�codingstatemachiner�chardistributionr�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/big5prober.cpython-39.pyc000064400000002224150564745010022032 0ustar00a

��Re��@sDddlmZddlmZddlmZddlmZGdd�de�ZdS)�)�MultiByteCharSetProber)�CodingStateMachine)�Big5DistributionAnalysis)�
BIG5_SM_MODELcs4eZdZ�fdd�Zedd��Zedd��Z�ZS)�
Big5Probercs,tt|���tt�|_t�|_|��dS)N)	�superr�__init__rr�	coding_smr�distribution_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/big5prober.pyr#s
zBig5Prober.__init__cCsdS)N�Big5rrrrr�charset_name)szBig5Prober.charset_namecCsdS)N�Chineserrrrr�language-szBig5Prober.language)�__name__�
__module__�__qualname__r�propertyrr�
__classcell__rrrrr"s

rN)	�mbcharsetproberr�codingstatemachiner�chardistributionr�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/sbcharsetprober.cpython-39.pyc000064400000006121150564745010023162 0ustar00a

��Re��@sNddlmZddlmZddlmZmZmZedgd��ZGdd�de�Z	d	S)
�)�
namedtuple�)�
CharSetProber)�CharacterCategory�ProbingState�SequenceLikelihood�SingleByteCharSetModel)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetcsbeZdZdZdZdZdZd�fdd�	Z�fd	d
�Ze	dd��Z
e	d
d��Zdd�Zdd�Z
�ZS)�SingleByteCharSetProber�@igffffff�?g�������?FNcsJtt|���||_||_||_d|_d|_d|_d|_	d|_
|��dS�N)�superr�__init__�_model�	_reversed�_name_prober�_last_order�
_seq_counters�_total_seqs�_total_char�
_freq_char�reset)�self�model�reversed�name_prober��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/sbcharsetprober.pyr3sz SingleByteCharSetProber.__init__cs:tt|���d|_dgt��|_d|_d|_d|_	dS)N�r)
rrrrr�get_num_categoriesrrrr�rr"r$r%rAszSingleByteCharSetProber.resetcCs|jr|jjS|jjSdSr)rr	rr(r$r$r%r	Ksz$SingleByteCharSetProber.charset_namecCs|jr|jjS|jjSdSr)rr
rr(r$r$r%r
Rsz SingleByteCharSetProber.languagec	CsJ|jjs|�|�}|s|jS|jj}|jj}|D]�}|�|tj�}|tj	krZ|j
d7_
||jkr�|jd7_|j
|jkr�|jd7_|js�||j
|}n|||j
}|j|d7<||_
q0|jj}|jtjk�rD|j|jk�rD|��}||jk�r|j�d||�tj|_n(||jk�rD|j�d|||j�tj|_|jS)Nrz$%s confidence = %s, we have a winnerz9%s confidence = %s, below negative shortcut threshhold %s)rr�filter_international_words�staterr�getr�	UNDEFINED�CONTROLr�SAMPLE_SIZErrrrrr	r�	DETECTING�SB_ENOUGH_REL_THRESHOLD�get_confidence�POSITIVE_SHORTCUT_THRESHOLD�logger�debug�FOUND_IT�_state�NEGATIVE_SHORTCUT_THRESHOLD�NOT_ME)	r�byte_strrr�char�order�lm_catr	�
confidencer$r$r%�feedYsH


�
�zSingleByteCharSetProber.feedcCsLd}|jdkrHd|jtj|j|jj}||j|j}|dkrHd}|S)Ng{�G�z�?rg�?g�G�z��?)rrr�POSITIVErr
rr)r�rr$r$r%r1�s
��z&SingleByteCharSetProber.get_confidence)FN)�__name__�
__module__�__qualname__r.r0r2r7rr�propertyr	r
r>r1�
__classcell__r$r$r"r%r-s


0rN)
�collectionsr�
charsetproberr�enumsrrrrrr$r$r$r%�<module>s�
site-packages/pip/_vendor/chardet/__pycache__/sjisprober.cpython-39.pyc000064400000004731150564745010022161 0ustar00a

��Re��@s`ddlmZddlmZddlmZddlmZddlm	Z	ddl
mZmZGdd�de�Z
d	S)
�)�MultiByteCharSetProber)�CodingStateMachine)�SJISDistributionAnalysis)�SJISContextAnalysis)�
SJIS_SM_MODEL)�ProbingState�MachineStatecsPeZdZ�fdd�Z�fdd�Zedd��Zedd��Zd	d
�Zdd�Z	�Z
S)
�
SJISProbercs4tt|���tt�|_t�|_t�|_	|�
�dS�N)�superr	�__init__rr�	coding_smr�distribution_analyzerr�context_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/sjisprober.pyr%s

zSJISProber.__init__cstt|���|j��dSr
)rr	rrrrrrr,szSJISProber.resetcCs|jjSr
)r�charset_namerrrrr0szSJISProber.charset_namecCsdS)N�Japaneserrrrr�language4szSJISProber.languagecCsNtt|��D]�}|j�||�}|tjkrP|j�d|j|j	|�t
j|_�qq|tj
krjt
j|_�qq|tjkr|j��}|dkr�|d|jd<|j�|jd|d�|�|j�|j|�q|j�||d||d|�|�|j�||d|d�|�q|d|jd<|jt
jk�rH|j���rH|��|jk�rHt
j|_|jS)Nz!%s %s prober hit error at byte %s�r�����)�range�lenr
�
next_stater�ERROR�logger�debugrrr�NOT_ME�_state�ITS_ME�FOUND_IT�START�get_current_charlen�
_last_charr�feedr�state�	DETECTING�got_enough_data�get_confidence�SHORTCUT_THRESHOLD)r�byte_str�i�coding_state�char_lenrrrr+8sD

�


�����zSJISProber.feedcCs|j��}|j��}t||�Sr
)rr/r�max)r�context_conf�distrib_confrrrr/Ys

zSJISProber.get_confidence)�__name__�
__module__�__qualname__rr�propertyrrr+r/�
__classcell__rrrrr	$s

!r	N)�mbcharsetproberr�codingstatemachiner�chardistributionr�jpcntxr�mbcssmr�enumsrrr	rrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/chardistribution.cpython-39.pyc000064400000014162150564745010023353 0ustar00a

��Re�$�@s�ddlmZmZmZddlmZmZmZddlm	Z	m
Z
mZddlm
Z
mZmZddlmZmZmZGdd�de�ZGdd	�d	e�ZGd
d�de�ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�ZdS)�)�EUCTW_CHAR_TO_FREQ_ORDER�EUCTW_TABLE_SIZE� EUCTW_TYPICAL_DISTRIBUTION_RATIO)�EUCKR_CHAR_TO_FREQ_ORDER�EUCKR_TABLE_SIZE� EUCKR_TYPICAL_DISTRIBUTION_RATIO)�GB2312_CHAR_TO_FREQ_ORDER�GB2312_TABLE_SIZE�!GB2312_TYPICAL_DISTRIBUTION_RATIO)�BIG5_CHAR_TO_FREQ_ORDER�BIG5_TABLE_SIZE�BIG5_TYPICAL_DISTRIBUTION_RATIO)�JIS_CHAR_TO_FREQ_ORDER�JIS_TABLE_SIZE�JIS_TYPICAL_DISTRIBUTION_RATIOc@sLeZdZdZdZdZdZdd�Zdd�Zd	d
�Z	dd�Z
d
d�Zdd�ZdS)�CharDistributionAnalysisig�G�z��?g{�G�z�?�cCs0d|_d|_d|_d|_d|_d|_|��dS�N)�_char_to_freq_order�_table_size�typical_distribution_ratio�_done�_total_chars�_freq_chars�reset��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/chardistribution.py�__init__.sz!CharDistributionAnalysis.__init__cCsd|_d|_d|_dS)zreset analyser, clear any stateF�N)rrrrrrrr=szCharDistributionAnalysis.resetcCsX|dkr|�|�}nd}|dkrT|jd7_||jkrTd|j|krT|jd7_dS)z"feed a character with known length����r riN)�	get_orderrrrr)r�char�char_len�orderrrr�feedFs
zCharDistributionAnalysis.feedcCsT|jdks|j|jkr|jS|j|jkrN|j|j|j|j}||jkrN|S|jS)z(return confidence based on existing datar )rr�MINIMUM_DATA_THRESHOLD�SURE_NOr�SURE_YES)r�rrrr�get_confidenceTs�
z'CharDistributionAnalysis.get_confidencecCs|j|jkSr)r�ENOUGH_DATA_THRESHOLDrrrr�got_enough_datadsz(CharDistributionAnalysis.got_enough_datacCsdS)Nr"r)r�byte_strrrrr#isz"CharDistributionAnalysis.get_orderN)
�__name__�
__module__�__qualname__r-r*r)r(rrr'r,r.r#rrrrr(s	rcs$eZdZ�fdd�Zdd�Z�ZS)�EUCTWDistributionAnalysiscs$tt|���t|_t|_t|_dSr)	�superr3rrrrrrrr��	__class__rrrrsz"EUCTWDistributionAnalysis.__init__cCs0|d}|dkr(d|d|ddSdSdS)Nr ���^r�r"r�rr/�
first_charrrrr#xsz#EUCTWDistributionAnalysis.get_order�r0r1r2rr#�
__classcell__rrr5rr3qsr3cs$eZdZ�fdd�Zdd�Z�ZS)�EUCKRDistributionAnalysiscs$tt|���t|_t|_t|_dSr)	r4r>rrrrrrrrr5rrr�sz"EUCKRDistributionAnalysis.__init__cCs0|d}|dkr(d|d|ddSdSdS)Nr �r8rr9r"rr:rrrr#�sz#EUCKRDistributionAnalysis.get_orderr<rrr5rr>�sr>cs$eZdZ�fdd�Zdd�Z�ZS)�GB2312DistributionAnalysiscs$tt|���t|_t|_t|_dSr)	r4r@rrrr	rr
rrr5rrr�sz#GB2312DistributionAnalysis.__init__cCs>|d|d}}|dkr6|dkr6d|d|dSdSdS)Nr rr?r9r8r"r�rr/r;�second_charrrrr#�sz$GB2312DistributionAnalysis.get_orderr<rrr5rr@�sr@cs$eZdZ�fdd�Zdd�Z�ZS)�Big5DistributionAnalysiscs$tt|���t|_t|_t|_dSr)	r4rCrrrrrr
rrr5rrr�sz!Big5DistributionAnalysis.__init__cCsX|d|d}}|dkrP|dkr:d|d|ddSd|d|dSndSdS)	Nr r�r9��?�@r"rrArrrr#�sz"Big5DistributionAnalysis.get_orderr<rrr5rrC�srCcs$eZdZ�fdd�Zdd�Z�ZS)�SJISDistributionAnalysiscs$tt|���t|_t|_t|_dSr)	r4rHrrrrrrrrr5rrr�sz!SJISDistributionAnalysis.__init__cCsr|d|d}}|dkr0|dkr0d|d}n&|dkrR|dkrRd|dd}nd	S||d
}|dkrnd	}|S)Nr r��������r"rG�r)rr/r;rBr&rrrr#�sz"SJISDistributionAnalysis.get_orderr<rrr5rrH�srHcs$eZdZ�fdd�Zdd�Z�ZS)�EUCJPDistributionAnalysiscs$tt|���t|_t|_t|_dSr)	r4rPrrrrrrrrr5rrr�sz"EUCJPDistributionAnalysis.__init__cCs0|d}|dkr(d|d|ddSdSdS)Nr �r8r9rr"r)rr/r$rrrr#�sz#EUCJPDistributionAnalysis.get_orderr<rrr5rrP�srPN)�	euctwfreqrrr�	euckrfreqrrr�
gb2312freqrr	r
�big5freqrrr
�jisfreqrrr�objectrr3r>r@rCrHrPrrrr�<module>sIsite-packages/pip/_vendor/chardet/__pycache__/compat.cpython-39.pyc000064400000000665150564745010021264 0ustar00a

��Re��@sLddlZejdkr.dZdZeefZeZej	Z	ndZdZe
efZeZejZ	dS)�N)�rTF)�sys�version_info�PY2�PY3�str�unicode�string_types�	text_type�dict�	iteritems�bytes�items�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/compat.py�<module>s
site-packages/pip/_vendor/chardet/__pycache__/escprober.cpython-39.pyc000064400000005157150564745010021766 0ustar00a

��Ren�@sXddlmZddlmZddlmZmZmZddlm	Z	m
Z
mZmZGdd�de�Z
dS)�)�
CharSetProber)�CodingStateMachine)�LanguageFilter�ProbingState�MachineState)�HZ_SM_MODEL�ISO2022CN_SM_MODEL�ISO2022JP_SM_MODEL�ISO2022KR_SM_MODELcsVeZdZdZd�fdd�	Z�fdd�Zedd��Zed	d
��Zdd�Z	d
d�Z
�ZS)�EscCharSetProberz�
    This CharSetProber uses a "code scheme" approach for detecting encodings,
    whereby easily recognizable escape or shift sequences are relied on to
    identify these encodings.
    Ncs�tt|�j|d�g|_|jtj@rD|j�tt	��|j�tt
��|jtj@r`|j�tt��|jtj
@r||j�tt��d|_d|_d|_d|_|��dS)N)�lang_filter)�superr�__init__�	coding_smrr�CHINESE_SIMPLIFIED�appendrrr�JAPANESEr	�KOREANr
�active_sm_count�_detected_charset�_detected_language�_state�reset)�selfr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/escprober.pyr*szEscCharSetProber.__init__csJtt|���|jD]}|sqd|_|��qt|j�|_d|_d|_dS)NT)	r
rrr�active�lenrrr)rrrrrr:s

zEscCharSetProber.resetcCs|jS�N�r�rrrr�charset_nameEszEscCharSetProber.charset_namecCs|jSr )rr"rrr�languageIszEscCharSetProber.languagecCs|jr
dSdSdS)Ng�G�z��?gr!r"rrr�get_confidenceMszEscCharSetProber.get_confidencecCs�|D]�}|jD]�}|r|jsq|�|�}|tjkrhd|_|jd8_|jdkr�tj|_|j	Sq|tj
krtj|_|��|_
|j|_|j	Sqq|j	S)NFr�)rr�
next_stater�ERRORrr�NOT_MEr�state�ITS_ME�FOUND_IT�get_coding_state_machinerr$r)r�byte_str�cr�coding_staterrr�feedSs"






zEscCharSetProber.feed)N)�__name__�
__module__�__qualname__�__doc__rr�propertyr#r$r%r1�
__classcell__rrrrr#s

rN)�
charsetproberr�codingstatemachiner�enumsrrr�escsmrrr	r
rrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/langrussianmodel.cpython-39.pyc000064400000063434150564745010023353 0ustar00a

��ReV�@s�,ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=dddddd>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]dd^d_d`daddbdcdddedfdgdhdiddjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]dd^d_d`daddbdcdddedfdgdhdiddjd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddkdldmdnd�dodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d\d�dXdd>d^d?d�dTdadBddCd�dRd`dAd�d�d�d�d�d�d�d�dSdddFd�d�d�d�dedGd�d�d�d�d�d�d�dfdHd�d�d�d�d�d�d�d�dgdIdhdJdidKddLdjd�d�d�d�dMd�d�d]d�dNd�dOd�dPd�dQdbdDd_d@d�dZd�d�d�dYdcdEd�dVd�d[d�dWd�dUd�d	dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd�dd^d�dadd�d`d�dddedfdgdhdiddjd�d�d�d�d�dbd_d�d�dcd�d�d�d�d�d\d>d?dTdBdCdRdAdSdFdGdHdIdJdKdLdMd]dNdOdPdQdDd@dZdYdEdVd[dWdUdXd���Z	ed�d�e	ed�d�d�d��Z
ddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=dddddd>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]dkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	d�d�dd^d_d`daddbdcdddedfdgdhdiddjd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]dd^d_d`daddbdcdddedfdgdhdiddjd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd���Z
ed�d�e
ed�d�d�d��Zd�S)��)�SingleByteCharSetModel���)?�%�,�!�.�)�0�8�3�*�<�$�1�&��"�#�-� �(�4�5�7�:�2�9�?�>�=�/�;�+r��
��
r��������r��	����'������6���������������������J��K�������������G��B��A��L��@���M�H��E�C��N�I���O����rrrr	r
rrr
rrrrrrrrrrrrrrrrrr�Fr r!r"r#r$r%r&r'r(r)r*r+r,r-r.r/r0r1���������������������������������������������������������������������������������������������r2r3r4r5r6r7r8r9r:r;r<r=r>r?r@rA���D�����������������(rrrrr+r0r4r3r.r2r&r-r/r(r5r1rAr>r=r'r*r%r9r,r)r:r7r@r8r;r?rrrrrrrrr6rr
rr$rrr	r"rrrr
rrr<rrrrr#rr!r rrhrdrbror�rnrzr`rlrrrQrSrfrkrqru�P�Q�R�S�T�U�V�W�X�Y�Z�[�\�]�^�_�`�a�b�c�d�e�f�g�h�i�j�k�l�m�n�o�p�q�r�s�t�u�v�w�x�y�z�{�|�}�~���������������rFrGrHrIrJrKrLrMrNrOrPrRrTrUrVrWrXrYrZr[r\r]r^r_������rarcrergrirjrmrprsrtrvrwrxry�����r{r|r}r~rr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rErDrCrB�IBM866�Russiangl���P@�?Fu�ЁАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯабвгдежзийклмнопрстуфхцчшщъыьэюяё)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetzwindows-1251�IBM855zKOI8-R�MacCyrillicz
ISO-8859-5N)�#pip._vendor.chardet.sbcharsetproberr�RUSSIAN_LANG_MODEL�IBM866_RUSSIAN_CHAR_TO_ORDER�IBM866_RUSSIAN_MODEL�"WINDOWS_1251_RUSSIAN_CHAR_TO_ORDER�WINDOWS_1251_RUSSIAN_MODEL�IBM855_RUSSIAN_CHAR_TO_ORDER�IBM855_RUSSIAN_MODEL�KOI8_R_RUSSIAN_CHAR_TO_ORDER�KOI8_R_RUSSIAN_MODEL�!MACCYRILLIC_RUSSIAN_CHAR_TO_ORDER�MACCYRILLIC_RUSSIAN_MODEL� ISO_8859_5_RUSSIAN_CHAR_TO_ORDER�ISO_8859_5_RUSSIAN_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langrussianmodel.py�<module>s�,
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���	���	���	���	���	���site-packages/pip/_vendor/chardet/__pycache__/gb2312freq.cpython-39.pyc000064400000045311150564745010021554 0ustar00a

��Re�P�@sdZdZdZdS)g�������?i�(�i�i�i�i<	iTi�i	iQi�i�
i�
i�i�i<iwi�i9	��i�	i�i�i�iZi�i�iqi�i�i�
iW
iyii�ieioi�iv
i�i�i�i�i�iiiLiBiiQ
�i�i�iQi�i�iEiifi�ii�ii�i�i�idi(iyii�i,i	i�i
iEie�i�
iWiRi�i�i|	iR
i�i�
i�i
iX
i�
ijiEi�ii+iPi�i�ii�i�;i�imi�ili�i�i�i��iieiHi�i�i�i�i^iki6i�i"�Fi�i�
i�
i�
i^	i�i�
i�i���i?i`iui$i1i�
i�i�i iUi�i1i�i�igili�ii		i�iGi�i2i�iq	i�i iPi�iL	i�i�i�iiy	i�i�
i�iMi�iki�i4	i�ii�i�i�i
i�	i@	i`i�i5��i�iN
i�i7i3i�iniiuiGi&ibihi0i�iMiNii�	i
iY	i�
i�i�	�
i��i�
�&i�
iQi
i7i^ihi�i�iJi�iPi�i�i�iri�i�i�ii�i�iT	i-iGi,iuii�	i�i$ii�i<iBii$i>ii�ixiLiWii�ini
i�iTi�iZi�i�
i�	i{ii�i�	i��ii�
i?i�
�0ibi;�Di�i��i�iCiSi�iUi>	i|igi�i�i�	i�
iK��i

i�i�i�i%
i5
i�i�ii�ili�ihi
ii�i?�i� iUi���i�i�	i�i*i<
i�i�
i�i
�i
i�i�i�
i�i�
i!	iX
i�i�i�	i�iQigi�i�i�i�	i�iri�
i�i�iqi�i�
i�iz
i�i�	i3�i�i�i�i
i�
i]iciziii7i�i_ii�iCi�i&iaiHi|
ili
i�i=i�i�iIiPi�
i�i~
i8i`i�iw	i�i�
iTi�i�
i�i�i�i�i�
i�ii�i���i[iq�ii�	igi��|iCi	i�i�iNiIi�i�i5i�
i(i�i/iHiK
i
i��ii�i9i�iIi�i�i|i��i�i�
i��iWii�i�i�i^�~iii�i�	i�i=iXi�i3i�
iIi�ii�
�i�i�i�i3ir�i���ioii�i�i{iiiiC���!iHi�	��i�i��f�i=
ii1�[i�i|iDii%i�iti�i�i�
i�i i�i.i��i�yiiTi�i�i9
i�
ii#�i�iJi��ni�ii�i
i�i>iBi"i�i�iai�i�iMiK��iiWi�i�	i�	i!
ir	iFi+	i�
i
ii�ix
�HiJi���i�i�i2�Ki�iR
i$ii9i�
it
i�i�
i�i�i��i�ii�i�ii�i�i%i~i�
i`	�i�i�i�i'iviii�i�i|i�'ii�in
i*i�i�i�iFi�i%i��)i~i�i
i�i�iJi&i�i
i!iDi�iwii[iii�	i�i	i'	i�	i��kii�i	i�
i�
��i��i�ii�i�ih
�Bi_i�i?iigihi�i'i�i�i�i�i�ii�i�i�ii�i��5i]i�ii�i�ii���i�i�i1	i�i)ia	ii�i�i�i;	ii�i_i�i��Qi/iu
i�
i�
iji[i�i}i�i
i�i�i�	ii�i�iQi^
i�iGi�izi�i�i�iViOi�ii�i�i�i�imiui�	i6i�i�i-i$i�i�iRi�i}	i
ipiri]
i�
iDiii�i�i�iiiii�i�ii1ini�i�iitiYiG	i�i�i/i|�ti0i�
i�i�ii�i\i�
i�i�ivi�i�i�ici�ii�i�
i~iAi"i�i"	i!i8�Ci8	ii�	iji�
i1i�ili>i�
ii]
i�i�
i~�iLi(i�i�iKid��viDi�i�i�i%i�	i�i�i�	i,�i2i�i�
i�i�ij�-i.i�imiKii���ioi	i;i	i�i�idi@iai	iM	i1
iLi@i2i�iwi�itiQi
i/i�i i�i~i�i�i�i�i�
i/i�i�
i_	i#i��Oi�iOiG�i�iki[iOioi���i-
i
�ii�	i!i�i�iIi�i�i�i{iri�i.ihi�
iiT
ii�i'iL
�ai�i�
i@�i�i�i�i�i�
iTi/
i�ii�i�ii�iMi�ii
ili�ii�
i{i8i#i�
i�i�i9i�iSiRi�	�wi�iZi�i0
iEigigi�i��
i��Li�i�ipi�i�iSi
i�i�i�i�i�	i@i�iI	iLi;
iOiqi�i�i)	ii:iFi�i5iJidiA
i�i
i)
iEiL�Yi�i�i�i�
im	i�i�i�i�i�i�ici�iRi�i!i�i�i�i�
��i(i�i�i]�Ii�i�
i�	i�iei�ii�i��i�i�i`i�ii�i�i
i�i[i�iC
i(iPi,i�i}iO	i'iLi\
i�
i�ii�i�iiZi�i�isi'i�	iUi�i�i�i�iN
�i�i�i�
i�i�iMi�i�i�i�i�i���i�iii�igi'i�i�i	i�i
��Pi0is	iii�ip	i�	inii�i�ii\i�
i�i�i�
i!ih���imi(i&i
iCiiBi)i5	i�i��ii�i�i
i�i��>i�i���ii�	��i{iNi�iiiEi3��Ji#i�ici�i�
ii$
��i5i�ixii]i�i�ik��i�i�i�i�iP
i�i
iiiri�iLi�izidi�i�i#i�
ii�i[iSiihiiGi�i�ii�i�
iU�+i�	iy�iwi"i�iVi2
i&izi�	i�	i�iJ
iP	i�	i�i���i�i%	iZi$ioiTiKi1i�
i�i`iVi�ii�id�i�i�i�i+i�i�iAi�i�i�i�i�i$i�i�	i_i3i\i�i�����i�ibi�if�i�i�i�i�
iAinii��x�iFi�
iOi[i�i��(i�ii
i#i7i2	ixiif	i�	i�
i!i6i{
��i�iYi��i�i+i�i�i�i�i i�i�i�i�	i	iKiei�i�i�
iTi$i�i�iihii	iRi�i�i"i�i��Zi 
i_i�i@i�iaii�ii3
i�
iki)iji�iiiei�i
iAi^i�i
iW	i�i�ihi:
iz
i)
i�i|i�i�ii�ii8ij��i�
ii�i3i�	i
izi�ipiyi�i�i2ioiD	i=i��.i�i9iaifi�i*i_i�i	i*i�ii�
i�
i�i�i-ii�
i}i�i�
i�i�i*	i�
it�iti6
�qi�i�i�i�i�i�i�iSi 	i�i�i!i)iiyi�
��iiij	i�i�	i�iViIi�i�ig
iiiLi�i�i�ipiDi�i�i�i�i%i�i�iMi�i�i�i$i�i�i�iE	i"iU	i�ie
i	i|i�	iQii�i�i�ii�i�
i1
i�i�i�i��$i i]i4i&i�i�izi�i�i`i�iwi�
iiii�
i)i%
i&i�i�i�i*i�i�i8i2i>
iSi
i�i�i1i#i�i�i
i�i#i	i�
i�i�ii�iF
iii,	i>i+i
ii�iZ�3iq�i�
i$i�i�
iVi�	i;i�i�iNi��miKiWi�i��i5i�
i�	�Ri4i�iIiiDi�i�i�i�
i�i�
i�i�i�ii%iKi|ii�i�
i��i��i9i�in	i�i�i�
i-i3i2
ipi�i?
i�i~i�i�i�i�i�iCi�i[i�i�
i�i�i�i3	i>i�ili
i`ix	i�i�i�i�i#
i�i�i�i0��i�i�i�	iL
i�i�iQi�iRi�i?i(imi�i`
i�i3ii�ii&
i�
iiisiDi�i0	ib	ifi@i�	i�i�i�
ii=i�i
i<��i�i�i�
i�i6ii�isi�i�i�iSi�i�i�
i8iZ
i�i�i�i�i�i�i�ii�
i�i�
i�ii�iNi�	i�i
iLi�
iJ	i�i�i�	ibiigii�i[
i�i�ii�ii~i�i�	i�	i�i�ifi�i�iii�
ii`i4i=	i�i�i�ili?	i�ici�ii�iiii�i�i�i�i�i�i�i�iwi�i�i�i%ii�id	i^i+i�i�i�iZ�iFiI
i�it�i�	�/i�	ii�iyi*
isi&i+ii�i)ii�i�	i�i�ii�
i�ie	i�i�iQi]	ii	i_i7	ii0�ciui�i�iLiki
i�i�i�i��#ii�ii�i��i�i?i��ii�
ini
i�i�i�i	i�i�i�	iHi�iYi�iu	i�i�
i�	i�iJi�i_i_i&i�i�ii�i�i�iZi%�i�	isi�iSiiiiVi�i�i�i�i�i1i'��i�iFi\i
i|i�	iWi�i�i|�iS�Wi�	i*i	iRiliPi#i�ii-iq
i�i��	i�i�i4i�i�i�i�i�iai�i�ii�i@
i�iXiii%i�i�i8
i3i,i�i}i�i�i��dixi�	i�i#	i�i6i�
i�
i.ii�iAi�
i�ir
i�
i�iRidi�ii+
i�i�i�i�
i{i{iHi�	i
i�i�i�i�i�ici	i
i�i�i�i,i�iI
i�icii�i�i�i�i=i�iJi.i,i>iii�iLiei]i�i�
iCi�i�i�i*i;i
i�i i�ipi�
i�i�iRi�i9i�
i�
ii�
i�i�i?i:i�	i�i�iQi'isi�iwi[i�
i�i-i�
i)i�i��Mi�i�i�i�i]�}i6i@i
iiUi�i�i�i/i�i�is
i�	i�i�ik	i?i�i,
i=
i1iMi�i�i�i/iS
id��i�io
i6	ikiui�i�i�iW�{iixiWi�i�ii�i�ixi^i��i[	i+�7i:i�iUi�i�i!i�i�i�
i.iHiTi�iYi�i�i	ii@i"iCisi	i�	iiG
�bi�i7i~i.
iUiiqi�	i�i�i�i{i�i��iYi/i�
i�i�iAi�
i�i*i:i8i'i�i�i�i�iiAi_i0i�i�i�iiii�iKi�
i�i�iti*i�i�i�
i�i�i�i�i�	i�
��i�iDii�ii>i�i�ivii

i�
i'i��Vi�i�i�
i�i�ikiii\i�i�i�i!ii��*iU
i�i�i�i�iizi�i\if
i�i�i�	i
i���i�i�i�i�i�i�i�iiB	iii�i�i�i{�i=i�
��i&i�
i�iVi�	i��"i!i�
i�	i�i�i�
i(	i=i�i�
i�
i�i9ii�i4ia��i�i-iQ	�i�
i�i�
imi�i�i�i�iCi�i�i7i}iqiDi4i��uii�	i�i�ivii�iF	i�iuia
i�iHioi�
i
	iliKiA	i�i�i�	izi4i�	i!
iH
i�i�	��i5iiji�i`i�i�
iBi:iOiai�i#i��i�id�i�i(i�
i&
i�i^imig
iE
iio	i�i(iGi$i@i
iIii�iAii�`�Xii7i�i/i�i$i�i�i�ihi�i�i�i�	iUi"iT�i�i�i�im�i�i!ii�i�ii�i i�iMi�	i��i�i�	�2iqi�i��9i�i�i��i�ii�i/i�i���i�pi@iX	i�ij�ii
i^
ibi�
i�iniiv�i�i�ioio�iivi�i�i0i�i9i�i$	i�ifiSizi
iiDi{i�iX�ii�iVi3iYi
i�i�
iAi�i��i�i�iqiUi�
i:	ii2iji�iii{	i�i�	iFi�i�
i�	ii	i��i�iciiiR	i;ik�hi�ii�
i�iYiOixi��ziZi<iv	ii�
i
i�
i�iIi�i�i4
iMi�i�i�i�ibi�ii�ii�i�i�iCi�i�
i�i�iOi�i8i�i�i�	i.i]iiVic
i�	i�i�i
iH	i�iTi�i�i�	i�i>
i
i�i�i3i�i��,i�i�ifi�i�iiliTiWiHi.i'i>ipi�i�i�i"i�i�	iJi�i�
i�
ivi2i�i�i	ixi	i2i�igi i4i�i�i�
i�
i�i�i{
iz	i�i�	i�
i�i�i�	i�i�ii�i�i?ii��i�	iJi�i�i�i�
i�	i8��iEij
�\i�i8i8
ii i�ii`i�i>ibiSii��i�i�ii�i0i�i�i�iEi�i�idi�i�i�i1�i�i�in�@i�i�i�iTii�	��i�iQ��iCi�i�iKi;iki�i�i�i�i�i_i\	ieiiwi�i�ii�iMi�i.	iriaii�i�
i�i�i�
i�i�i�igi�i�i'i�
i�iii�	�Tii�i4ioi�i�	iXiyi�ixii�	i�
i
iU
�^i:i�i�i�i�ii"i#i�i�iii�	i�i�i�iNiUi�i�ip�i"
i-i�i�	�i�i:i�i
ii�	i�i�i�i�	i��iWi
iG
iLi�i�
iGiw�oi�i�iiMir�i�ii�iQ
i�iIiN�%i
iXiii�i7
i�
iai�i�i�i�	i'
iiihi�i}iNiyi?iti�i�iDi�i�i��i��li�i�i�i	i��ieiJisi2i*iyi�i�i�i�i�ii�iqi�i�i5i}isi$i�i�i�i5
iBi�i��i#idisi}i�i~i�iib
i�iO
i/	�]i�i�i�	i]i2��iifi7i�i�iOi

iRii#
i	i[i��Ei�iif
iS
i�	idiFi�i��iC	iV
i>i�i	i�iBi�ii	i1i6im��ii	i��i�i�ip
i�ii�
i�i<izi�
i|
iX�i�i�isiei]ii��i
i�
iiOi�i+i�	iii�i�i�i:i�	iAi�ii�i�i�i(
i�
imi�4iP�i	ini�i�
ioioi�i�imiiy�ii8i	��i<i�i�i<ic	iCi�i�i&iwi
i�	i$i 
i9i&i�i�i)i`i�i�ii�i�iGibiEi�i:iQioiFi�
i�i�i�	i�
iFi|i;i�
i�ii3i4i%i��:i�	ici�ipi6
i\iCi�i0i9
i�i�i�i�i�i~	i�ii"ihi�i�i�i�	i)ii�	i�
i�i�i*
i�ii
i
i%i�i	i�i�ii�
i.i�i�i9i	i�i�i�
i�i�i	i�i"i!i�i,
i�ig	iJigiF
i�ii�i�
i�i;ii	iZiEi^iA
i�iili5i�i�i�i4i'
i�iiki	i�i�	i2i�i�
i�i	iXi7�i�i�
i{i�iBixi=�ei(i�i�i�i'i�i��i�i�ii�iiii�iD
iiikii�i�i�i-i,i�
iei7
i��i�i��i�i�ini}
ipi7iji�i�	i�i9i�i�
��i�i���i0iPi���i�i�ii�iO
i�
i�i�	��iUi
ii�i<
i�i�i�
iJi�itiNi�i)ii.i��i*iiA��ir
i]i)i�ii�i�iGi�i_
ipik
i�i�i
	i%i
i�i�
i�i>ibi_i^�?ii:iYiDi%�i�i�
i�ii�i��iYi�i�
i�i�ii�i�il
�Ai�i�	i�
i�	iDiBi�	i
i�i	i@
i(iv
i�i�i��gi�
i�	i}ivi�i�i�i�
i�ii�i�ii�i�ii�ii�i�i i-i~i�iyi%i�	i�
�_i�i�ii�i�i�jiMi�i�iGi+i�	ii���i�	i�i�ii3
i;i
�Gi)i�i
i�i�ii�i�ii�i
i
i�i,iIi�	i�i�iiXiAi\i��i�iPivi�i�	iVii=i�
i�
i6i<i�iXii�
i	ii8i-	iY
i�i�i�i�i�i��ici�i�i�i9i�	i-
iui�i	i�i�
iRi�
iii6i�i�	i�ih	i}ii?
iw
ii�i�	iai(i�i�1i�i�
i�i>i�i�i�i�i�i]i�
iXiriPi0i,ii+iji�i`i�i�i�ii.iii
i�i&	iWiB
i�ii|i7�Ui�i�	�iWi�iYi�i�
i�iw
i�
ii�ii�ii�i�iui�ibi�iq
i�i�ibi�i�i���=i$
i�iNiJi�i�iUi�ii�i�	ii��iiKi,i�i�i4iri{i�iiHi�ipi5ii�i=i�	i�iEi9i�i;i�<ii5i�
iiii�i\ii��ipi�ii�i��ri��Ni�ii^i�i�i�i�iKii�i*ii{i+
i�	i�i*i8ii�ii	ii�i�i�ij
ii@il	�6i�i2i8i�i�	i�ii�i'i�i�i�iB
i�iii\i�i�iSiV	ii~i�i�i�i�i�i�
i�iVi�
izi�i^i	i�i�i�i_
i}
iGiq�iIi�i�i��8itihiS	i�i}iRi�	i9��Sin
iPi�iBiii��si�iXi}ifi�i�ii�i�i�i�i���i�iniIi^�iTi�ixi�iri�i�i_�ibi�iOiPiQiRiSN)�!GB2312_TYPICAL_DISTRIBUTION_RATIO�GB2312_TABLE_SIZE�GB2312_CHAR_TO_FREQ_ORDER�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/gb2312freq.py�<module>*ssite-packages/pip/_vendor/chardet/__pycache__/version.cpython-39.pyc000064400000000735150564745010021464 0ustar00a

��Re��@sdZdZe�d�ZdS)z�
This module exists only to simplify retrieving the version number of chardet
from within setup.py and from chardet subpackages.

:author: Dan Blanchard (dan.blanchard@gmail.com)
z4.0.0�.N)�__doc__�__version__�split�VERSION�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/version.py�<module>ssite-packages/pip/_vendor/chardet/__pycache__/langturkishmodel.cpython-39.pyc000064400000050200150564745010023343 0ustar00a

��Re�v�@s2"ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd	d
ddd
ddddddddddddddddddd dddddddd!d"d#dd$d%d&dd'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7dddddd8d9d:d;d<d=d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel���)?��%�/�'��4�$�-�5�<��1��.�*�0�,�#��3�&�>�+�8r���r���r��
��
������	�� �9�:���?�6�2�7�;�!�=�"���)��(��rrrr	r
rrr
rrrrrrrr�Errrrrr�Arrrrr r!r"r#r$r%r&r'r(r)r*�@r+r,r-r.r/r0r1r2r3�����������������������e���������j��������d���������^�P�]��i��r4�������~�}�|�h�I�c�O�U�{r5�z�b�\�y�x�[�g�w�D�v�u�a�t�sr6�Z�r�q�p�or7r>r@�V�Y�Fr8�N�G�R�Xr9�M�B�T�S�n�Kr:�`r=�C�m�J�W�fr;�_�Q�l�L�Hr<r?rA�k(rrrrr(r&r?r+r,r-r%r2r r'r.r)rr<r!rArrr3rr$r#r*r"rr
r=rr/r9r;rrrrr	r@r>rrrr
rrrrr6rrrr5r7rr0r1r8rr:rr4rErDr�r�r�rCr�r�r�r�r�r�r�r�r�r�rwr�r�r�r�r�r�r�r�r�r�r�r�rxrvr�r�r�r�r�rmr\r�r�r�rzrer�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rr~r}r|r{ryrurtrsrrrqrprornrlrkrjrirhrgrfrdrcrbrar`r_r^r]r[rZrYrXrWrVrUrTrSrRrQrPrOrNrMrLrKrJrIrHrGrF�����������������������������������������������������������������������������������������������������������������������������rBz
ISO-8859-9�Turkishg�X4���?TuRABCDEFGHIJKLMNOPRSTUVYZabcdefghijklmnoprstuvyzÂÇÎÖÛÜâçîöûüĞğİıŞş)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetN)�#pip._vendor.chardet.sbcharsetproberr�TURKISH_LANG_MODEL� ISO_8859_9_TURKISH_CHAR_TO_ORDER�ISO_8859_9_TURKISH_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langturkishmodel.py�<module>s"
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���site-packages/pip/_vendor/chardet/__pycache__/langbulgarianmodel.cpython-39.pyc000064400000052530150564745010023626 0ustar00a

��Re˜�@sP$ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=dddddd-d4d9d:d;d<d=d>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVddWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�dd�d�d�d�dd�d�dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dd	dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddCdDdEdFdGdHdIdJd�dKdLdMdNdOdPdQdRd�d�d�d�d�d�d\dSdTdUdVdWdXdYdZd�d[d]d^d_d�d�d�d`dadbdddedcd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d	d�d�dd�d�d�d�dd�d�dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�dd�d�d���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel�)?�?�-�� �#�+�%�,�7�/�(�;�!�.�&�$�)��'��"�3�0�1�5�2�6�9�=�<�8r��	���������
����
������������4�*��:�>r*r'�����M�Z�c�d�H�m�k�e�O��Q�f�L�^�R�n��l�[�J�w�T�`�o��s�A�E�F�Br�D�p�g�\���h�_�V�W�G�t���U�]�a�q�����������������������������������������������������������������������i������������r����rrrr	r
rrr
rrrrrrrrrrrrrrrrrrr ���C��r!r"r#r$r%r&r(r)r+r,r-r.r/r0r1r2r3r4r5r6r7r8r9r:r;r<�Kr=��r>r?rA������r@��b������(rrr*r'r0r4r/r2r3r$r-r&r,r1r.r)r?r<r#r5r%r9r8r(r;r7r+r:rr6rrrrrrrr
rrrrr>r	rrrr
rrrrr=rrrr"rr@rr!r rAr�@r`rcr�rdrarbrmrJ�IrYr�rRrF�NrN�PrPrT�Sr[rprkrl�X�YrGrXrgrqrSrjr\rrr�rHrIrMrQrfrir��jrLrWrKrUr]rers�rr_rn�u�vrZ�x�y�z�{�|�}�~����������������������������������������������������������rOrVr^������rhrortrurvrwrxryrzr{r|r}r~rr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rErDrCrBz
ISO-8859-5�	Bulgariang! _B�?FuxАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЬЮЯабвгдежзийклмнопрстуфхцчшщъьюя)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetr�r�r�r�r�r�r�r�r�r�r�r�r�r�zwindows-1251N)�#pip._vendor.chardet.sbcharsetproberr�BULGARIAN_LANG_MODEL�"ISO_8859_5_BULGARIAN_CHAR_TO_ORDER�ISO_8859_5_BULGARIAN_MODEL�$WINDOWS_1251_BULGARIAN_CHAR_TO_ORDER�WINDOWS_1251_BULGARIAN_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langbulgarianmodel.py�<module>s4$
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���	���site-packages/pip/_vendor/chardet/__pycache__/enums.cpython-39.pyc000064400000005176150564745010021132 0ustar00a

��Re}�@shdZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGd	d
�d
e�ZGdd�de�Zd
S)zr
All of the Enums that are used throughout the chardet package.

:author: Dan Blanchard (dan.blanchard@gmail.com)
c@seZdZdZdZdZdZdS)�
InputStatezS
    This enum represents the different states a universal detector can be in.
    ���N)�__name__�
__module__�__qualname__�__doc__�
PURE_ASCII�	ESC_ASCII�	HIGH_BYTE�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/enums.pyrsrc@s<eZdZdZdZdZdZdZdZdZ	eeBZ
e
eBeBZdS)	�LanguageFilterzj
    This enum represents the different language filters we can apply to a
    ``UniversalDetector``.
    rr����N)rrrr�CHINESE_SIMPLIFIED�CHINESE_TRADITIONAL�JAPANESE�KOREAN�NON_CJK�ALL�CHINESE�CJKrrrr
rsrc@seZdZdZdZdZdZdS)�ProbingStatezG
    This enum represents the different states a prober can be in.
    rrrN)rrrr�	DETECTING�FOUND_IT�NOT_MErrrr
r src@seZdZdZdZdZdZdS)�MachineStatezN
    This enum represents the different states a state machine can be in.
    rrrN)rrrr�START�ERROR�ITS_MErrrr
r)src@s,eZdZdZdZdZdZdZedd��Z	dS)	�SequenceLikelihoodzX
    This enum represents the likelihood of a character following the previous one.
    rrr�cCsdS)z::returns: The number of likelihood categories in the enum.rr)�clsrrr
�get_num_categories;sz%SequenceLikelihood.get_num_categoriesN)
rrrr�NEGATIVE�UNLIKELY�LIKELY�POSITIVE�classmethodr&rrrr
r#2sr#c@s$eZdZdZdZdZdZdZdZdS)�CharacterCategoryz�
    This enum represents the different categories language models for
    ``SingleByteCharsetProber`` put characters into.

    Anything less than CONTROL is considered a letter.
    �����N)	rrrr�	UNDEFINED�
LINE_BREAK�SYMBOL�DIGIT�CONTROLrrrr
r,Asr,N)r�objectrrrrr#r,rrrr
�<module>s			site-packages/pip/_vendor/chardet/__pycache__/eucjpprober.cpython-39.pyc000064400000004665150564745010022325 0ustar00a

��Re��@s`ddlmZmZddlmZddlmZddlmZddl	m
Z
ddlmZGdd�de�Z
d	S)
�)�ProbingState�MachineState)�MultiByteCharSetProber)�CodingStateMachine)�EUCJPDistributionAnalysis)�EUCJPContextAnalysis)�EUCJP_SM_MODELcsPeZdZ�fdd�Z�fdd�Zedd��Zedd��Zd	d
�Zdd�Z	�Z
S)
�EUCJPProbercs4tt|���tt�|_t�|_t�|_	|�
�dS�N)�superr	�__init__rr�	coding_smr�distribution_analyzerr�context_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/eucjpprober.pyr%s

zEUCJPProber.__init__cstt|���|j��dSr
)rr	rrrrrrr,szEUCJPProber.resetcCsdS)NzEUC-JPrrrrr�charset_name0szEUCJPProber.charset_namecCsdS)N�Japaneserrrrr�language4szEUCJPProber.languagecCs6tt|��D]�}|j�||�}|tjkrN|j�d|j|j	|�t
j|_q�q|tj
krft
j|_q�q|tjkr|j��}|dkr�|d|jd<|j�|j|�|j�|j|�q|j�||d|d�|�|j�||d|d�|�q|d|jd<|jt
jk�r0|j���r0|��|jk�r0t
j|_|jS)Nz!%s %s prober hit error at byte %s�r���)�range�lenr
�
next_stater�ERROR�logger�debugrrr�NOT_ME�_state�ITS_ME�FOUND_IT�START�get_current_charlen�
_last_charr�feedr�state�	DETECTING�got_enough_data�get_confidence�SHORTCUT_THRESHOLD)r�byte_str�i�coding_state�char_lenrrrr)8s<

�


���zEUCJPProber.feedcCs|j��}|j��}t||�Sr
)rr-r�max)r�context_conf�distrib_confrrrr-Ys

zEUCJPProber.get_confidence)�__name__�
__module__�__qualname__rr�propertyrrr)r-�
__classcell__rrrrr	$s

!r	N)�enumsrr�mbcharsetproberr�codingstatemachiner�chardistributionr�jpcntxr�mbcssmrr	rrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/big5freq.cpython-39.pyc000064400000065121150564745010021503 0ustar00a

��Rez�@sdZdZdZdS)g�?�(�i	i��i����	�R�i����iai�i�ni��!i�i�L�,iB�i�i�i�i�i��i{
i]
i�i�i
iji�i�i.iNiii�����:i�i�i��?i�i�i=iN�K���ikiq	i�i�i�
i��iib��
i�i��i�
ioi2i�ii�i�ici8i�i�ii��i�i��i{�|��i��"i�
i��@i\i�i�	i�i��i�i�i�iFi�i|
iQ�Hi�i��Piv��i�i�ii�iDi^
i�i�iFi}
i��Ei�iO�i0i��si�i4�<�2i��i&iMi�i�i~
�i�iG�ii[i�i�i?iiiaiiK	i*i�igi�	�Zi�
i:i��i�iKi�	�i�	�ii�i�i�i`ilii�i��i
	i �qi�i~i���i
i�	ii�i�i�ici��u�i�i*ii�i	i�
i��~imi�i��iei��Gi^i�in�i�UiCi�iCi�i�i�i�ii�i �jioi/ii�iPi7	i[i���i�
i?i�i�i�io�Si(i�i�ip�]i�i6ijii@iii�iii8��+i3i[�i�i\i�i�i�i��i]	iAi���i�i1��i�iHi�
i�
idi���i+i�i2i�i�i�i�
iOiL	i�iifi1��i�i�i�i�i�ii3i9���li,��i	ii�ii�i�i�i�ieiz
iQiMi&i��Xi�i�i�i���i
��i�ikipi��M�i�i��%�i�i�i�i�i��'i�	i�i\i�i�7��Ji�i!i�i�ii���iNi�iBiPi_
iq��i�
i�i�i�
i�
i�� i�ii iIi�i8	i�i�
ii{
�i�i
i�i���mi�i�ifi�i�ixi�i�i�i�i�
i�i
i�igi���i
ii�i�i�i�i��&i�i'i�ii�i�i��.�i�ii�iM	i���i�i�i$�#i�iDihiAi�	�i�iri�i�iUiGii�i�i
i�
iQiSiii�i��di�0�Fi�i�iCi�i�ii�i i�iJii�iU�i��iN	i�
i�i�i<i�i�
i:iA
i�
i�i	i 	i9i�i}ii�	i�i�iWiPi�
ii)iDii�i�i4i�
i�	i)i�i�ir	i�isi�i�itii9	i�i�i�is	i*ii�i�i�i�i�i]i�i�i�i�i�	iDiji(
i
i�i�i�
i�iui5iYiji	ili�iuii�i�
i�i�i`
i
i=i�i�i�i�i�i�i�i5i�i!i�i�iTiEixi�i�ie�i�	i�iO	iPi�i|
i6iAi�
i/iki�i�
iili!i	i�iBi�i`i�i
i��iAi�iiii�i�
ivi�	i�i�iwi�i���iMi�
i�i�i�ii�ia
i�i�i�
i}
i�ixi,�}i�i�iB
i�i�i�i�i�i�i�i�iFikimi�	i�i�
i�i	i�i�
i�ibi�i)
i^iP	i
i,i�i�i�i7i5i~
iyi|i�izi�ii�i�
iC
i{iii�i�i�i�i�i�i�i1���bii�	i�
i�i�it	i�i�
i�i_i��i�i�i�ii�
i�
ib
i��i�i5�D�i�ili�i|i[i*
i��i�
i�i%�iGi^	i�ili�i�i�
i�i`i�i�i�ipi��in�cigimii�i'i2i��i�i{i�i�i�i�i�i�iD
i�ifi	i�i|i:	��i�ii�i�
i�i�
i
i�i�i�ini�i�i�i��;i�iIiYi}i�i�
iXi"i�i
i�i�i�i�i�i�i-i�ili�i�i�i�i�ii�i�iEi�i~i�i>i]i�i,i�i�i�iviLiB�ii&i�i�ii�i�i�iGi�i�iiBi�i�
i+
i�ii�i�i�i�i�i�	i iaii�i�i
iai�i}iE
i�iDi�i=i�i0ii�i�
i6i�ii�ivi!�i�i�i��i�i�i��9iHii�i�i�i�
iFi�i�i�i�i�i"i�iHi�i�i�i
ioii�iRi*i�i.i�i�i�i�iiI�3ii,
i�i*iSi�i�iXi�i�i/iibipi�i�	i�i�i�i�i�iRiii�i7i�i2i�i�	i�i�i�i�i�i�i_	i�ibii�
i�i�i�i�	i�i�i�ii�i�i`	�Oi�i�	i�i`i7i�
i�
i�i�i�i�i�iia	i�i�i�i�i�i�igi�i�i�i�i
i_i�i~��i�i�i�i�i�
iai�i�
ib	i�i�i�i�i��i�i�	i;��i�	i-i
i�iQ	i���i�iqii�i�i��`i�i#i#i�ii�iiri�
isii�i�iditi�
i�ic
i�iikiwi�i�i�i�i�i�i�ii�i��gi�iQi�iUi
i�i�i�i���CiSi�ic	i�i�i5i�i�iBi�i_i�iciN��i�iYi�i�i�i�
�iLidi�
i�
i	iKi�i8iaiGi�i�i
i�i�iisi6i�	iti;	�Ni�i�ii	i4i�i:i�i\iqi�i�i�i.i�iiiui�
i�i�i�i�i�
�i
iri�i�i�
i�
i�ici�i�i�i@i�i�i�ii�i�ii�i/i9i�ii�ii�i�i�i�i�
i<	id
ii�i�i8iviiCi�ii
i�imi�i�
i�i�i�i�i�iZi����i�iwi�	�{i�i�i�i��$i�iiMidi0iri�i�i�igiiVi�i�ii�i:iZi�i�i{i$iei0i�ii$i�i�
i�
i�i5i�i[i��vi�iici�iViR	i$i;�i�iifiii�idi�i�iWie
ii�i�iii�i�
iiJi�ii�iiui�i	i�i
i�ii�
iKi�i2i�iLi i.i�i�i2i�ii�i�i��i�i%ii�i�ii
ii�ii�iii��
i�
ixi�iMi�i}ii�iTii�i�id	i�
ii�i�i�i`iii�i�i�	i3iii;i�iiiiyiYiLi�i�i�	i�	iii�iU�\i�i�ii�i�iViNi�ii�
i�i�i 
i�i"ixi%ii;i=	i i�i�i2iEi�i!i�i�i��^i�i"iwi�
i#i�i�i�i�i�i$i<iQi�i�iri%i�i]i�i&i�i<i'ii3i6iBi�i�i(i�i�i�i�i�ii�igi�
i�
i)i�
i	i*i+i!	i�i��i:i�i7i�i^i�ii�ii(i�
i,i+iiixi-i�i/ihi=i.i�i��i�iVii�i�hi�i�i/ii��i�i<i�i�i;i0i�i�i�i
i�	i1izi�ii�i5i6i�
i_�)i2i3i�i4iyi"	�i�i�
i�i^i5i�i�i?i%i+if
i�i6i_iWi�i�i�
i7i>ii�
ig
i#	��iu	i8i�i�i9ih
iF
i�iVi{i�iOi�iI�wi~i�
i^ii�i:i�i�i�i�i�i�i�i�	i;iei1i�i�i�i�i�i�i�i�i�i�iiiPi<i=i�i>i�
i��(i`iRi?iYi�iIi�	��i�i�i7i�i�
i?i�i�i�
i@i�io�iqiYinii�ilii�i�i�i�i�
ii
iEiPiVi�i!
i�i�i�i�i�ini�iAi�i|i�
ii
iBi�i�i�i�ii-
i�	iOi�
i�iCi�iDiWiTiT�i�i�i�i�
i}iEi�i&i�
i�i�i�i�ii�i�
i�i�i~iG
i�
i�ii$	iWi�ii�ioi�i
iFiiGi�ii"
i���i�i#
i�i�iDi-iQi�iHiei�	iiiIi�i�iiJizi
i�iKiiiLi�i$
i�
i�i�i�i�iH
�i�i�ii�i�iMi:i�	ii�iNi�iI
i�i�i�iHi�i�ii�iOi�i�i�	iPiQi�i�ikiRiRi���i�i�
i
i�i�i�i�i�i�ii�iSiTi�i�i�i%
i�i�i�i�i�i&
iSiUii�iS	i�iCi�i�i�iVi~i�i!iXi�i�i�i�i�i,i�ii�i�iOi�i|i�ii�iPi�ie	i�i�i�i�ii�i�i�i�i�i�i�i�iTi�i�i�iWi�i�iiXi�iYi�iRi�
iviii�
ii�i=ii�iZi�iUi�i�i[i�i�i�iVi%	i\i�i�i]i�i�i^i8iT	i.
i'
i_i�ivi�i�i�i�iwitii`iaibi>	i�i�i"i{�i�i�	i�ici�idi9i�iii�i[i9iiidihi
iei�i�i�iJ
i!i�i�ifi���ii�i�i�iTigi�ii�i%i��WiMihii�iii�i�i�i�i��i�i�i�i(
ij
i�iyi7imi�ijiki�i?	iWi"i�i��i�Bi�ili|i��i)
izi���i�
i�i�i�i�ii�i�im�iniuiv	iwi�i	i*
ii�ioipi��iqi7i.iri�i�i
iii�
i�ifi�i�i�isi&	iyii�i i�i�i�ii�i�
ii>i�i�i�i�i�Yi�i�iti�iui/
i+
i�ivi�i�i�i@iwii�	i�	�i�i3i�i�i�i,
ixi�i6i9i�
i	ii��6i��ii!ii�iyi�iQi:ici
i�ii-
�oi�i�iFiXi�
i�i�ifi�iizi�
i�i�i�i�i�i{i/�i�i'	i�	ibii�i`i�i�iiui	ivi�i�ii�i�
i
i
i�i�i�i�i0iE�iSi�i8i[i|i}iGi�ii�i!i~ii�iFiU	i�i�
iii�ieisi�i�i�iw	ii�i�i"iiai	iZi&i�i�i�i�i�	i
i�i�i�iAi�i�i�
iiiii�i�i�i�i�i�ii�i
ioi�ii�ii�
i
i�i	i�iAiiJi0i�ix	i�
i�i�i�iYi�i�i�i�i�i�	i�ii�iZi8ioii�iQi�i�i�i�i�
i���a�Qi�i�ipi;i�iPi�ibi�i�i�i�ii�i�i
i�i�i�i�	i�i�i�i�i�i�i�i�i�i�i�i]ii�ii�i�
�i�ii�i�i'i�ii.
i���i
i�i�ii�i�iLi�i�i�i�ii�	i�i/
i�ii�i�iBi0
i[ii8i�i�i�i�i�	i�ii/i3i�i�i�ii�
ii�ii0i�ini�i�i�ii�i���ii�	i�i�i�i�ii�iii�i4i�i�ii<i�	ii
i�i�iYi�	i�i��iCi�i�ii�
i�i�i�i�i}i1ii�
i�i�i�iK�i���i0
iFi�i�i�i�
ii#i�i\i�i�i��fi�>i�
i�ii�
i�i�i�i�i�ii�
i�i�i�i�iji�i�ii�ii-ipi9i@	i6iyii�i�ii
i1
i�i�ii�i�ii��Vi�i�i�i�i4i�i�i�iKi1
i�i�
iDi�i�i�i�ii�i#i�i�i�i�i2
ii�ii�	iZiHipi�i�igi�i�ii�
iq�ri
i���i�i�i�i�iiK
iV	i�
i\i,i
i�i�i�ii�i]i�i�iu�i�
iii�i�i�i�ii�iZi�i�i�
iii�iiji�i�i0ihihi�i�i�i�i�iEii�i	
i�ii�	i�i�i�i�i�i�i�ii�i�i�
iii�i2
�i�i�i�i�iiMi�iwi�i�i�iy	i?i�i�i�i

itimi(	i�	i�i+i�i�i�i�i�i��iiii�
i�i�	i�i�i�i
i�iL
i�i�i)	i�
ii�
i�iif	i�i�	i�iii�i�i�iRi"iA	i3
i�i�i-i1i�i$i�
i(i�i]i�i#i�i$i�
ii�i�i i*	i�i�i�i�i
i�i�i�
i=ik
i�i!i�i�
i�i�i�i�i�
i�i���_i�i�i�i�i^i2ii�
i�i�i�i9i�i�i�i�i�i�i4
i�i�i�i
�ibii�i�ii�i�ii��5i�i�i�i�i"i	i�
ii�i�i��i4iIii�i#i�i��i�iii�iiji�i�ii$iki�ii�
i�i\i�
igi�iEi5
i�i�i�i�i�i�i�i�i�
i�i�i6
i�i�ii	i�i%ii�i�i�i�i[i�i�i�i�i�ii�i&�ei�i�i�i�i�ioi�i�ili	i}i�i�i�izi�i+	i�i�i�ii�i�ini-i�i�
i'iiii
i�i�i�i(i�i�i�	i�i�ii�i�i�ii�
i�i�i�i�iBiiiqi�i�i�i<i�i*iji)i"i*i�i>i)i7
i�i+i�i�i�ii�i�i��zi�i,i�i�i�iiM
i)iri�i�i-i	i�i�i�i�i�i��Ti��pi�i�i#i#i�i�ii�i�i�i�i�
i�ii�i	i�	i�i�i�iSi�i.i8
i�i�i�i�iz	i9
i=i�i�i�imi�iSiii�i�i�ixi�i�i/i�i�iqi�ii3i�i5i�i:
i�i�
i�i�i�i�i�i�i�i
i;
i�i0i1i�iii�i�ini�i_i�i]ii�ini�i�i�i�i
i%i�i�i�i�i�i`i�i�i�i�iziTisi2i3
i�i
i{	�i5���ii�
ig	i�i�iki�i�i�ii�
i�i�i3i4i�i�i�ici i�i�i
i�
i�iiai:i1i�i�i�i!�Aihi ioi�
i6i"i�i��il
�-i�i
iai5ii�	i�i
i�i�i#i!i�i$i�i�
i�i4
ifi�i	i%i&i�
iN
ri
i�iWi�iii�ii�
i'i�i$i�ii<
i�iNiiiW	iki�iO
i�i�iii�i�iX	i,	ii6i&iti�iP
ii�i7i�
i�i�
i-	ijii	i�
i�iC�[ipi�i�i'i8i
ibi�im
ii�
i�i{i(i�i�	i�i�	i9i�
i�	ii:i
i�
i^i�i�i�iB	i;ii�iii�i�
i���i
i�i	i�i)i<i=i�
ii�
i�i>iiiii�
�=i�i�ii�ii�i�i�i�i�in
i�iHiDi�
i?i�iUiWi�
i=
i�i:i�i�	iNi�iKiEi@i4ici�
i�i>
i�iQ
i
iAii�ihi�ii�
i�
i(i]i�i�i�iBi*i�i�ili�iCiDi�iEih	i�ii�i��i�iXiiii�i+i�i�i�i+iJi�i�iU��ii
i�imi iEiiiiFiiGi
�ki�	ini�i�
idii�ii�i�i�iFi!i i�iri@i�iHi"iCiGi��i�i
i�i^i�
i�i��8iti)i;i�i?
i!i�ioiIi�i"i#i#i$i�iJi�	i%i�
i&iriKi'i5
i(i)i�io
i@
i}iA
i�i,i�i*i4�1i+i�i�i�i�i�i"i�i,ii�
iC	i�iuiriD	i-i�
i-i�iLivi{i�i|	iMi.i.iNiY	i6
iii�i/i0i/i�i�i1i$i�i2i%iGiRiB
i�i_i�i�iC
i�i�iD
i�i3i�i�i�i�i�i�ii.	i�i�	i4i@iwi�i5i�i6i7i�ii	i�i&i0i�i8i	i&i�	i*�i+iGiOiu�i#i�
i1iPiRi�iQi�i�i�i)i�i�i=i9i�ili�i�i�i�iE
iSi�iki�i\i�	i�i�i�
iyi:i�i;i<iF
i%i=i�iRiii>idi'i,i?iei$i�i2i3i-i�i�i4ii(iSi@iAiBi�i.iTi�iOi
i�i�i�i�i�i�
i&ii�i�i�
i<i�
i�ii�iCiiDi�i�iUi�ip
iG
iei�
i�i�i�i�
iEi;iViFi'iGi)i
i�i}	iHi*i�iWi�i�iIi�iTi�i�
i�
i�i�i�i�ii	i�iJi.igi~�iUi�ij	i	iiiXi�i�i8i�i*i�	i�i�iKi~	i�i�i�iLiiHi�
i�ii%��i+ifi�i�i�i�iti�iMiNi�i�i�ii	i�i 
i�i�ik	iYiiYi�i�i��*i/i�iOiR
iZiH
i�i#iIi�
i�	i�i�i5iS
i]iE	i[i\i
isiI
i+iPiQi,i�	i�
iRi�iSiTi�i,i�iUi]iVi�i�iLi-i!i�iT
i\�4iWi�iiXiYiZi�i[i>iwi�i�i�i�i+i^iiJ
i�i_i$i�i\i�i�iili.i]i`i�i^i_iq
i%i�i`iaiii�i�
i6igiU
ibicidiai�
iqixi�
iei
ibiyi�ifihigihi�i�i'iiii�i�i	ijiri�i
i0i	i=i�ibi|�i�i�i�i�i�i	iihiki�
ilii�i�
iai�icimi�ii�ini�
i�iui�i�idiei�isiK
i/ioi7i�i&i�	iJi8ipiqiiijifi�i�	igi	i�i�i(i�i`irisi�i�iii<iFikifi�i��i�i]i�i�iIiti>i�iui�ivi�i�
i�i�i�
i0i�i-i�iwiL
i�i�	i�i�i�i�iGiai�i
iti�i�i�
iSiKixiiyizi�i�iV
��i�iOi�i1�i�
i{i�
i�i�iF	iyi�i9iZ	��i|i}i~iFi�i�i�i�
i�	i:i�iiii2i�i�iipi�i;iG	i�i�i	ihi�i�i�ii7
i�i�i�iiqii�iM
i�iii�iii�i�i�i�
i�i�i�
i(i8
i�iN
i	ii3i�	iiO
i<ii�i�iTiiW
i�	i�i�i�i�i(i�i�i�i1i�i�i�i�i�i�i�iuiZii�
i�	ici�ijiP
ipi�i�i
ii�i�i�iviiiji�i�
i�i�i�i�iQ
iKii�i�	i'i�i�i�ii�i9
ii�i
i�	izi�i�	i
i�i�iki�i�i�i�i�i�i2i	i�ilii�i>i=iCi'i�i�i>i
iiZi�i�i�i�imini�i�ibii�i�	iiqi?i�i
i�ii�i�	i"i�i�
i?i
i}i�i�ioi@i>iiAi�i1i�i)i�i�i�i?i�
i�
i�ii�ixi	i�i	ii�	i3i�iihi�iili�i�i�i)i�i�
ipi�i'i�i�i�i4i�i�i
i�
i�i�i�iAi_i	iBi�iLi�i
iPiqi�ii�iriii�iiX
iicisi!
i�i�iiHi�i	iwi�i�i�iCiDitii�i�i�i�i�ixi
ii�	i/	i�iii5i
i4iuiir
i�ii_i�i��/i�i<i�i~i7iyimi�ii^i5inibi�
iti�iil	i�i�i�i6i�i�	i7i�i�i�i�i�iivi�i�i�i�	i8i:
i�i�
i�i0	i�i1	i�iwi�Ii�i�i�i�i�i(i�iiBi�i�i4isi�
i�i�i�i�i
i�
i�i�i?iCiixi	iEiHii�iFiyiziGi�ii�i�i{i�i�ii|i�i;
i}i9imiH	i~i�izi�i�is
i�	i�iMi�i�ii�i�iim	i�i�iIi�i�iioii�ii�ii�i�i_ii�i�i�i2	i3	iHii*i�ii�
ii�i�i<
i�iii:i�i�
i�i�i�iAi�i�i�iR
i�i�i�iii�
i�i"
i�i�	i�i�i�i�i�i�
ii�
i�i�i�
i�i�iQii�
iY
iii�i�ii�ii{iKi�i�i[iiJi�i�iNi�iS
i�i�i�i�iI	ii�iT
i@ii�
i�i�i�
i�i@iiRi�i�i�	it
i�ii�iJiIi�i�
i�i�i��i�
iU
i�imii�i�i{i�	i�i�	i
i;iV
i�ii�	ii�i<i�ii�
idi�i�i�iXi�i�i9i�i�i6i�i�iUi�i�i�i i�iLi!i=
i�i)i�i�
i	ii�i�i�	i�i�i�
ii=i$i(ivi�ii&i�ici�iii�i	i
i�i>i�i�i�i{i"i�i�i�i�i#i�i�i�i�iW
i�i�	i�i�i�i�	ii$i�
i%ii�i�i�i�i&i^i�i�i[	i�	i�i'ii�i�ii�i�i�i(i i�ii2i\	iJi
ipipin	iiu
iqi)i�i�	iOi>
i�i�
i*iX
i�i�i�	i+i�i�i,i�
i�i�i�
i�i�i�i�iKi�i
i		��iIi-i�i�i!i�i�i�i�i�i.iDi�i�iJisi#
i�i�i�
i�i7i�i�i�i/iwi�i|iii�i�i�i"i4	i|i,iii�i�i�	i�i�ii�iiii�i?iY
iPi0ii1iZ
iZ
i�
i�i�
i�	i�i�
i#i@i2ii|i�iziii�i�iAi'i�i?
i�i	i�i8i�i��yi�iLiVi
i3ii�	i�i$i�ii�iMi�i%i�ii�	i�
ii
i.ii4iNi�iiii�
i�	i�i�ii�i�
i5i9i%ii�i�iKiQii��xi6i7iiBi�i
	iOiv
i[
i�i�i&ii�
ii�
i�iiXi�i*i�i\
iPi�	i�
iNi+iw
iii�iiWiCi�ii�	iQiDii?i]
i�iii^
io	iRi	
i�iEii
iiFii'i�iXiiGiiSi�iiTiRii8ixi�	ikiSi�iTi�ii i�i�i�iDi�i9i!i�i�ii�i(iiLi"i#i$i%i_
i�i&iJi:iUi�i}ii)i�i

i;i<i
iEi+i'i�
i�iHi=iVi*iHiyi,i+iWi3i>i�i�i?i(iIi,i-i`
iqi�ia
ii�ii
i�i)i*iXi�	i+i�iO��iYi,i-i�i.i�i/i�i�i0i.i�i�i�i1i�
iJiri2i3i$
i	i4i�i5i6i7i�
i�ib
i�	i�i�i}i�	ifi�iKi�i�
i�
i�	i�	i8iZi9i�i:i�i�
ic
i�i:i�ii;isiJ	i@i�i�	iti;i�
i/i<id
ii�	i�i�iUi�i0i�ie
i=i�if
iuiMi�i�i�i�i�
i>i	i[iiAiLi�iii?i�	i
i�
i	i@iAiviBiCi\i�iji�i[
i�
ig
iDiEi�i�iFiiBi�iGi�
i�iHi�i(i�iMiCih
i�iIiNiJiiKiLi�i�iMi�i1i2iN�i�idiOidiViDi`i�i#i5	iPi]iQiEini
i�	i�i�i^iRi_iFiSip	ii
i�iTiUi�i

i�iVi�i�i�	i�iGi�i�i�ieiHiOi�	i�iPi`iWiwij
iXi ii3i�
iai�iIiYiJixi7iii�i-i�
i4ik
iibiZi-i�ii.ici�i[i�	iri;i�iKi5iLiXi�i i	i�i�i\i]i�i�i6i%
i^i_i�i�i~i`il
iyiaizi~i�i�ibi?i�ici7idi
i�i�iei�izi�i�iMifigi/ihi|i�i@idi�iiiiNi�i�i�i8ifi�i�ijiki�iOiPi�i�ilimi@ini9i�	ioi�i�	i�i�i<isii\
iQi.iRiNi@i�
i�ipiiSi{iqi�iAi�i�i�ii�iTi
i�i�i�i:i�i�iri�
isi@
iWi;itiui�iviwixiyi|i�i�i�i�izi�i�i3iQi�i�imi�iAi{i|i�i�	i�i}iOiRi�i�igiii~iZi�
ii�i�i�i�i�i\im
i�i}i�i<iXi�i�	i�i�	i�
i�i�i�i
i�iYi�i�i�i0i�i�i�i�i=i�	i�i�i�iLi�i�i�i�ihi�i�i�iii>i>i�i/i?iUi�i�i�i�i�iei)ix
�ti�ifiSi�i�
iy
i�
in
i
ii�iTii@i�i�i�i�i���i�i�i�i�i�	i�i�iziUigi�iVi�io
i�i�i6	ip
iti�i�i�i�i���i�
iAi�iJiVihiZi�	i�iWi�i�i�i�N)�BIG5_TYPICAL_DISTRIBUTION_RATIO�BIG5_TABLE_SIZE�BIG5_CHAR_TO_FREQ_ORDER�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/big5freq.py�<module>+ssite-packages/pip/_vendor/chardet/__pycache__/cp949prober.cpython-39.pyc000064400000002233150564745010022054 0ustar00a

��Re?�@sDddlmZddlmZddlmZddlmZGdd�de�ZdS)�)�EUCKRDistributionAnalysis)�CodingStateMachine)�MultiByteCharSetProber)�CP949_SM_MODELcs4eZdZ�fdd�Zedd��Zedd��Z�ZS)�CP949Probercs,tt|���tt�|_t�|_|��dS)N)	�superr�__init__rr�	coding_smr�distribution_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/cp949prober.pyr#s
zCP949Prober.__init__cCsdS)N�CP949rrrrr�charset_name+szCP949Prober.charset_namecCsdS)N�Koreanrrrrr�language/szCP949Prober.language)�__name__�
__module__�__qualname__r�propertyrr�
__classcell__rrrrr"s

rN)	�chardistributionr�codingstatemachiner�mbcharsetproberr�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/euctwfreq.cpython-39.pyc000064400000065125150564745010022010 0ustar00a

��Re�{�@sdZdZdZdS)g�?i(�ii��i����	�R�i����ii�i�
�ni��!i�i�L�,iA�isi�i�i�i�iL
iS
i�ii�i�i�i�i.iNiii�����:i�i�i��?i�i�i=iN�K���i�il	i�i�i�
i���i�
i ��
ii��i�
ioi$i�ii�i�ici8i�i�ii��i�i��iz�|��it�"i�
ie�@i\i�i�	i�i��i�i�i�iFi�iM
iQ�Hi�i��Piv��ifi�ii�iDiT
i�i�iFiN
i��Ei�iO�i/i��si�i3�<�2i��i&iLi�i�iO
�i�iG�iiMi�i�i?iii`iiF	i*i�igi�	�Zi�
i:i��i�iKi�	�i�	�ii�i�i�i`i�ii�ig�i	i �qi�i~i���iP
i�	ii�i�i�i!i��u�i�i*ii�i	i�
i��~i�i�i��iei��Gi^i�i��i��UiCi�iBi�i�i�i�ii�i �jioi/ii�iOi2	i[i���i�
i&i�i�i�i��Si(i�i��ip�]i�i6iii�
i'iii�iii8��+i%i[�i�i\i�i�i�i�iX	i(i���i�i0��i iHi�
i�
i"i!��i+i�i1i"i�i�i�
iOiG	i�iifi1��i�i�i�i�i�ii2i9���li,��i	ii�ii�i�i}ihi#iq
iQiMi&i��Xi�i�i#i���i
��i�iji�i��M�i�i��%�i�i�i$i�i��'i�	i�iNiii�7��Ji�i!i�i�ii���iMi�i)iPiU
i���i�
i�i%i�
i�
i�� i�i�
i iIi�i3	i�i�
iir
�i�i�i�i���mi�i�i$i�i�ixi�i�i�i�i�
i�i�i�i%i&��i
ii�i�i�i�i��&i�i'i�ii'i�i��.�i�ii�iH	i���i�i�i$�#i�iDi&iAi�	�i�i�i�i�iUiGii�i�i
iQ
iPiSi'i�i��di��0�Fi�i�i*i�i�ii�ii�iJii�iU�i��iI	iR
i�i�i<i�iS
i:i7
i�
i�i	i	i9i�i}ii�	i�i�iViPiT
ii)iCii�i�i&i�
i�	i)i�i�im	i�i�i�i�i�ii4	i�i�i�in	i*ii�i�ii�i�iOi�i�i�i�i�	i+i(i
i�i�i�iU
i(i�i5iYiji�
ili�iuii)i�
i*i+iV
i
i=i�i�i�i�i�i�i�i4i�i!i�i�iTi,ixi�i�ie�i�	i�iJ	iPi�is
i5iAiV
i/iki�i�
iili!i	i�iAi�i`i�i�
i��iAi�i�
iii�i�
i�i�	i�i�i�i�i���iMi�
i�i�i�ii�iW
i�i�i�
it
i�i�i+�}i�iji8
i�i�i�i�i�i�i�i�i-i)imi�	i�iW
i�i		i�i�
i�iai�i
iPiK	i�i,i�i�i�i7i'iu
i�i{iki�i�ii�i�
i9
i�iii�i�i�i�i�i�i�i1���bii�	i�
i�i�io	i�iX
i,iQi��i�i�i�ii�
i�
iX
i��i�i5�D�i�ili�i�i[i 
i��iY
i�i%�i.iY	i�i*i�i�i�
i�iRi�i�i�ipi��in�cigi+ii�i'i2i��i�i{i�ili�imi�i�i:
i�ifi
	i�i|i5	��i�ii�iZ
i�i�
i�i�i�i�i,i�i�i�i��;i�iIiYi�i�i[
iXi"i�i	i�i�i�i~i�i�i,i�iki�i�i-i�i�ii�i�iDi�i�i�i>i]i�i,i�i�i�iviLiB�ii&i�i�i�
i�i�i�iGi�i�iiBi�i�
i!
i�i�iui�i�i�i�i�	i iaii�i�iv
iSi�i}i;
i�iDi�i=i�i0ii�i\
i(i�ii�ivi �i�i�i��i.i�i��9iHii�i�i�i]
iEi�i�i�i�i�i!i�i/i�i�i�ii-ii�iQi*i�i.i/i�i�i�ii0�3ii"
i�i*iRi�i�iWi�i�i/iibi.i�i�	i�i�i�i�i�iRiii�i)i�i2i�i�	i�i�i�i�i�i�iZ	i�iTiiw
i�i�i�i�	i�i�i�ii�i�i[	�Oi�i�	i�i`i7ix
i^
i�i�i�i�i�i�i\	i�i�i�i�i�i�igi�i�ini�i
i_i�i~��i�i�i0i�i�
iai�iy
i]	i�i�i�i�i��i�i�	i;��i�	i-i�i�iL	i���i�i/ii�i�i��`i�i#i"i�ii�ii0i�
isii�i�idisi�
i�iY
i�iikiwioi�i�i�i�i�i1ii2i��gipiQi�iUi�ivi�i�i���CiSi�i^	i�i�i5i�i�iBi�i_i�ibiN��i�iXi�i�i�i�
�iLici�
i�
i	iKiwi*iaiGi3i�i
i�i�ii1i6i�	i2i6	�Ni�i�ii	i3i�i:i�i\iqi�i�ii-i�iii3i�
i�i�i�ixi�
�i�iri�i4i�
i_
i�iUi�i�i�i@i�i5i�ii�i�ii�i/i+i6ii�ii�i�i�i�i�
i7	iZ
ii�i�i8i4iiCi�ii�i�ili�i`
i�i�i�i�i�iYi����i�i5i�	�{i�ii�i��$i7iiMiVi0iri�i�i�igiiVi�i8i�i9i,iZi�i�i{i#iWi0iii$i�i�
i�
i�i4ii[i��vi�iici�iViM	i$i-�i�iiXiii�idi�i�iWi[
ii�i�iiiqi�
ii1i�ii�iiti�ii�i	iii�
i2i�i2i:i3i i.i�i�i2i�i
i�i�i��i�i%ii�i�iii
i�ii�iii��
i�
i6i�i4i�i|ii�iSiii�i_	i�
ii�i�i�i`iii�i�i�	i3iii;iriiii7iYiLi;i�i�	i�	iii�iT�\i�i�ii<i�iVi5i�ii�
i�i�i�ii"ixi%ii;i8	i i�ii2iEi=i!i�ii��^iiwi�
i"i�i�iisi�i#i<iQii�iri$ii]ii%i�i.i&ii3i5iBi�iyi'i�i�i�ii�ii>iYia
ib
i(i�
i	i)i*i	i�i�i:i�i6i�i^iii�ii(i�
i+i+iiixi,i	i.ihi/i-ii�i
iVii�i�hi�izi.ii��i�i<i�i�i;i/i?i�ii
i�	i0i8i�ii�i5i6i�
i_�)i1i2i	i3iyi	�i�i�
ii^i4i�i�i?i$i+i\
i�i5i_iWi�iiz
i6i0ii�
i]
i	��ip	i7i�i�i8i^
i<
i�iVi9i�i6i�iI�wi}ic
i^ii
i9i@i�i
iAii�i�i�	i:iei1iiiBi�i�i�i�ii�i�ii�i7i;i<i�i=i{
i��(i`iRi1iYiiIi�	��ii�i6i�id
i>i�i�ie
i?i�io�iqiYinii�iliiCiDi{i�i�
i_
iEiPiUi�i�i�i�ii�iEimi�i@i�i:i�
ii
iAii�i�i�ii#
i�	iNi�
i�iBi�iCiWiTiT�i�i
i�if
i;iDii&i�
i|i�i�i�ii�i�
i�i�i<i=
i�
i�ii	iVi�i�
iioi�i�	iEi�
iFiii�i���i�i�iiiDi-i8i�iGidi�	ii�
iHiti�iiIizi
iiJi�
i�
iKii�i|
i�ii�i�i>
�i�i�i�iiiLi:i�	i=i�iMii?
i�i�iiHi�i�ii�iNii�i�	iOiPiiFikiQi9i���i�i�
i�	iuiivii�ii�iiRiSi�i�i}i
i>i�i�i�ii
i:iTi	i�iN	i�iCi�ii�iUi~i�i iWi�iGi�i�i�i,iiii�iOi�i|ii�i�iOi?iVi`	i�i i�i~iiwi�ii�i@i!i�i�i;i�i�iiWi�i"i�
iXi�iYiiRig
iuii�
ih
i�
ixi=i
iiZi�i<i�ii[i�i#ii=i 	i\i$ii]iHiIi^i7iO	i$
i
i_i%iviJi�i i!iwitii`iaibi9	i&i�i"i{�i�i�	i'ici�idi8i�ii�i�i[i9i�i�idiZi�	iei�i�i�i@
i!i(i"ifi���ii�i#i$iTigiKii�i%i��WiMihii�iii�i�i�i%i���i�i�i�i
i`
i�iyi7imiLijiki�i:	i>i!i�i��i�Bi)ili|i�i
izi���ii
i�iMi�i�i�
i�i�im�iniuiq	ivi�i�
i
i*iNioipi��iqi7i.iri�iyi�
iii�
i�ifi�i�i�isi!	iyi�
iOi	i�i�i�i+iPij
i�
i>i�i�i�i�i�Yi�iAiti�iui%
i
i�ivi�i�izi2iwi,i�	i�	�i�i3iQi�i�i
ixi�i6i9ik
i�ii&�6i��i-i
i�iRiyi�iQi9ici�
i'i.i
�oi(iBiFi?il
i�i)iei�i�
izi�
iSiTi�i�i�i{i/�i�i"	i�	ibi/i�i`i�i�i0iui�ivi�i�i1i�i�
i�	i�iUi�i*i�i/iE�iSi�i8iZi|i}iFi+ii�i!i~i2i,iFiP	i�im
ii
i�ieisi�iCi�ir	ii�i�ii�iai3iZi%i-i�i�i�i�	i4i�i�i�iAi�i�i�
i�
iii
i�i�i�iDi.i�ii�i�	ini�ii�i�i�
i�i�i	i�i3i5iJi0i�is	i�
i�i�i�i@i�i�iVi�i/i�	i�i6iWiAi7ioi�
i�iPi0i�i1i�in
i���a�Qi2i�ipi:i�iPi�ibi�i�i�i�ii{i3i7iEi�i�i�	i�i�i|iFiGiXi�i�i�i�i�i]i�
i4i�i�i}
�i�i�
i5i�i&i�i8i	
i���i�	i�i�i�i�i�iLi�i�i�i�i�
i�	iYi

i�ii�i�i4i
iBi9i8i�i�i�i�i�	iHi�
i/i3iZi[i�i�
io
i�i�i:i0i�ini\i�i]i�i^i���i�i�	i�i�i�iIi�i�i�i;i�i4i�i<�i=i;i�	ii
i�i�iXi�	i�i��i5i_i�i�
i�
i�i�i�i�i}i0i>ip
iJi�i�iK�i���i&
iFi�i�i�i�
i�
i#i�i[i�i�i��fi�>iq
i�i?ir
i�i6i�i�i�ii�
i�i�i�i�iji�i�ii�i@i-ipi9i;	i6iyiAi�i�ii�	i'
i�i�ii�i�ii��Vi�i�i`i�i4i�i�i�iKi
iai�
i6i�i�i�i�ii�i"i�i�iKi�i

ii�i�i�	iYiGioi�ibifici7iis
ip�ri�	i8��i�i�i�i�iiA
iQ	i�
iCi,i
i�i�i�ii�i\iLi�iu�i~
i�
i�i�i9i�i�iBi�iZi�i�i
iCi�i�iijidi�i0ihigi�iei�iMi}i7i�
i�i
i�iDi�	i�i�i�i�ifi:i�iiNi;i�
iEiFi�i(
�i�i<i�i�iGiMi=iwigi�i~it	i?i�i�i�i
itimi#	i�	i�i+i�i�ihi�i�i��ii[i�
iOi�	i�i�i�i
i�iB
i�i�i$	it
ii�
i�iia	i�i�	i�iHii�i�i�iQi"i<	i
i�i�i-i1i�i#iu
i'i�iDi�ii�i
i�
iIi�i>iJi%	i�i�iPii
iii�i�
i=ia
i�iKi�iv
i�i�i�i�i�
i�i���_i�iji�i?iEi1i�
i�
i�iki�i8iQi�i�ili�i�i
i�imi@i
�ibi�i�i�iiRi�i�i��5i�i�i�i�iLi	iw
iini�i��i4iHii�iMi�i��i�ii�i�ii\i�ioi�iNi]i�ii�
i�i\ix
igi�iEi
i�i�i�i�i�i�ipi�i�
i�iqi
i�i�i�i�i�iOi�i�i�iri�iZi�isiti�i�ii�iP�ei�i�i�i�i�ioi�i�i^ii}i�i�i�iziAi&	i�i�i�i�
i�i�ini-i�i�
iQiiii
i�i�i�iRi�iui�	i�i�iiBi�i�i�
iy
i�i�iviwiBihiqi�i�i�i<i�i*iiiSi"iTiCi>i)i
ixiUi�i�i�i~i�i�i��zi�iVi�i�i�i�iC
i)iqi�i�iWi
	i�i�i�iDi�iE�TiF�pi�i�i#i#iyi�i�
i�i�iGi�iz
i�ii�i�i�	i�i�i�iSi�iXi
i�i�i�i�iu	i
i=iHi�i�i_i�iRiii�i�i�iwi�i�iYizi{iqi�ii3i�i5i�i
i�i{
i�i�i|i}iSi~i�i
i
i�iZi[i�iii�i�i`i�iFi�i]ii�ini�i�iIi�i
ii�i�i�i�i�iGiTiJi�iKiziSiri\i)
ii
iv	�i5���ii�
ib	i�i�iji�i�i�i�i�
i�i�i]i^i�i�i�ici�i�i�ii�
i�i�
iai:i1i�i�i�i��Aihiiai�
i6i�i�i��ib
�-i�i
iHi_ii�	i�i
i�i�i�ii�i�i�i|
i�i*
ifi�i	i�ii�
iD
ii
iLiWi�iiiUii�
ii�i$i�i
i
i�iNiiiR	iki�iE
i�i�iiiMi�iS	i'	ii`i�isi�iF
ii�iai}
i�i�
i(	ijii	i~
iViB�[ibiWi�i�ibi
iIi�ic
ii
i�i{i�i�i�	i�i�	ici�
i�	iidi
i�
i]i�i�i�i=	ieii�iii�i�
i���i
iNi	i�i�ifigi�
i�
i�
i�ihiiiii�
�=i�iXii�ii�i�i�i�i�id
i�iHiCi�
iii�iTiWi�
i
i�i9i�i�	iNi�iKiEiji4iJi�
iYi
i�iG
i	ikii�ihiOii�
i�
ii]iPiZi�ili�iQi[iki�iminiRioic	i�ii�i��iSiXiiii\i+i�i�i�i�iIiTiUiU��ii
i�ili�
iDiiiipiiqi
�ki�	imi�i�
iKii�ii�i�iVi8i�
i i�iri@iWiri�
iCiGi��i�i
i�i^i�
i�i��8itii:i�i
i!i�inisi�i"i�
i#i$i]iti�	i%i�
i&iriui'i+
i(i)i�ie
i
i}i
i�i�i�i*i4�1i+i�i�i�i�i�ii�i,ii�
i>	iXitiri?	i�i�
i-i�iviui{i�iw	iwi�i.ixiT	i,
iii�i/i0i�i�i�i1i�
i�i2i�
iGiRi
i�i^i�iYi
iZi�i
i[i3i^i�i�i�i�i�i
i)	i\i�	i4i@ivi�i5i�i6i7i�id	i]i�
i�i^i8i�i&i�	i�ii9iyiu�ii�
i�iziRi�i{i�i�i�i(i_i�i<i9i�ili�i�i�i_i 
iSi`iki�i[i�	i�i�i�
ixi:i�i;i<i!
i$i=i�i|ii�i>idi�
ii?ieii�i�i�ii�i�i�ii�
i}i@iAiBi�i.i~iaiOi
i�i�i`ibi�i�
i%ii�i�i�
i;i�
i�ii�iCiiDi�i�ii�if
i"
iLi�
i�i�i�i�
iEi;i�iFi&iGi�
i
i�ix	iHi)i�i�ici�iIidiTi�i�
i�
iei�i�i�ii	i�iJiigi~�iUi�ie	i	iii�ifiai8igi�
i�	ibihiKiy	i�i�i�iLii:i�
i�ii��i*iMiii�i�iciti�iMiNi�i�i�i�i	i�i
i�i�if	i�i�iYidiei��*ii�iOiH
i�i#
i�i#i;i�
i�	i�i�i�iI
i\i@	i�i�i�isi$
i�
iPiQi�
i�	i�
iRijiSiTi�i+i�iUi�iVi�i�iLi�
i!i�iJ
i\�4iWi�iiXiYiZi�i[i>iwikifi�i�i+i�ii%
i�i�i$i�i\i�i�iili�
i]i�ili^i_ig
i%i�i`iaigiii�i�
i�iNiK
ibicidi�i�
iciwi�
iei�ibiyi�ifihigihihi�i'iiiimi�iz	ijidiii�ii	i<i�i�i|�i�i�i�i�i�i�iiOiki�
ilii�i�
iai�i�imi�i�i�ini�
i�iui�ini�i�i�iei&
i�
ioi�i�ii�	i<i�ipiqiPiQi�iji�	i�i	i�i�i(i�i_irisi�i�i�ii<iEiRifi�i��i�i]i�i�iIiti=i�iui�ivi�i�
i�ioi�
i�
iki,iliwi'
i�i�	imi�i�i�iFi`i�i�ifi�i�i�
iSiJixi�iyizini�iL
��ipiOi�i�i�
i{i�
i�i�iA	iyi�i�iU	��i|i}i~iFi�iqi�i�
i�	i�i�ii�iii�i�iipi�i�iB	i�ioi	i�i�i�ipii-
i�iri�i�iqiqi�i(
i�iri�i�i�isi�i�i�i�
i�i�i�
i(i.
isi)
i�itii�	ii*
i�iui�i�iTiviM
i{	i�i�iti�i'i�i�i�i�
i�i�i�i�i�i�i�igiZiwi�
i�	ici�iji+
ipi�i�i�ixi�i�i�ihi�i�i�i�
i�i�i�i�i,
i=ii�i�	iiui�i�ii�i/
iyi�i
i�	iyivi�	i�i�iwi�i�i�i�i�i�i�i�
izi�i�ii�i>i�iCi'i�i�i�i{iiZi�i�i�i�i�i�i�i�iaii�i�	i�iqi?i�i
i�i|i�i�	i"i�i�
i�ii}i�i�i�i�i=iiAi�i1i�i)i�i�i�i?i�
i�
i�ii�ixi	i�i	ii�	i�
i�i�ihi�iSi�i�i�i(i�i�
i�i�i'ixiyi�i�
i�i�i
i�
i�i�i�i�i_i	i�i�i>i�i}iPi�i�i�i�i�ii~i�i
iN
iici�i
i�i�i�iHi�i	iii�i�i�i�i�i�ii�i�i�i�i�iji�i�i�	i*	i�i�i�i�
i�ii�iih
i�i�i_i�i��/i�i<i�i~i7ikiTi�ii]iiUibi�
iti�i�ig	i�izi�i�
i�i|	ii�i�i�i�i�i�i�i{i�i�i�	ii0
i�i�
i�i+	i�i,	i�i�i��Ii�i�i�i�i�ii�iiBi�i�i4isi�
i�i�i�i�i�i�
i�i�i?iCi�i�i	i�iGi�i�i�i�i�i�i�i�i�i�i�i�i�ii�i|i1
i�iimiC	i�i�ili�i�ii
i}	i�i?i�i�ii�i�i�ih	i�i�iIi�i�i�iVii�i�i�i�i�i�i^i�i�i�i�i-	i.	i�ii)i�i�i�
i�i�i�i2
i�iiii�i�
i�i�i�iAi�i�i�i-
i�i�i�i�i�i�
i�i
i�i�	i�i�i�i�i�i�
i�i�
i�i�i�
i�i�iQi�i�
iO
i�i�i�i�ii�i�iziKi�i��i[i�iJi�i�i@i�i.
i�i�i�i�iD	ii}i/
i@i�i�
i�i�i�
i�i@iiRi�i�i�	ij
i~i�i�iJi�i�i�
i�i�i���i�
i0
i�imi�i�iimi~	i�i	i�ii1
i�i�i�	i�i�ii�i�i�
idi�i�i�iXi�i�i9i�i�ii�i�iUi�i�i�i�i�iLi�i3
i�ii�i�
i	i�i�i�i�	i��i�i�
i�ii$i(ivi�i�i&i�ibi�i�i�i�i�i�i�ii�i�i�i{i�i�i�i�i�i�i�i�i�i�i2
i�i�	i�i�i�i�	ii�i�
i�i�i�i�i�i�i�i^i�i�iV	i�	i�i�i�i�i�i�i�i�i�i�i�i�ii2iW	i�i
iWioii	iik
iXi�i�i�	iAi4
i�i�
i�i3
i�i�i�	i�i�i�i�i�
i�i�i�
i�i�i�i�i�i�i�i	��iHi�i�i�i�i�iiiii�iDii�iIisi
i�i�i�
i�i i�i�i�i�iwi�i{iii�ii�i�i/	ini,iiii�i	i�	i�i�i
i�iiii
ii4
iBi�ii�iP
i5
i�
i�i�
i�	i�i�
i�i	i�ii|i�iziii�i�i
i'i�i5
i�ii�i!i�i��yi�i�iVii�i�i�	i�i�i�ii�i�i�i�i�ii�	i�
iii-ii�i�i�ii�
ii�
i�	i�i�iii�
i�i"i%ii�i�iJiCii��xi�i�iii�i	iil
i6
i�i�iii�
ii�
i�iiXi�ii�i7
ii�	i�
iNiim
iii�i iWii�i!i�	ii
i�
i>i8
i�i"i#i9
ij	iDi�i�ii$i	
i%ii�
i�i�iXi&ii'iEi�i(iFii)i�ixi�	ikii�ii�i*i+i�i�i�iDi�i�i,i�i�ii�i�iiKi-i.i/i:
i�i0iJi�ii1i|i�i�i�i�i�i�i�iEi*i2i�
i�ii�ii�iHixii�ii3i�i�i�i�i3ii�i�i;
ipi�i<
ii�ii�i�i4i5i	i�	i6i�iO��i
i7i8i�i9i�i:i�i�i;i�i�i�i�i<i�
iiYi=i>i
i	i?i�i@iAiBi�
i�i�i=
i�	i�i�ioi�	ifi�ii�i�
i�
i�	i�	iCiiDi�i#i�i�
i>
i�iEi�i�iFiZiE	i�i�i�	i[i$i�
i�iGi?
ii�	i�i�iGi�i�i�i@
iHi�iA
i\iLi�i�i�i�i�
iIi	iii�ii�iiiJi�	i�i�
i	iKiLi]iMiNi
i�iji�iQ
i�
iB
iOiPi�i�iQi�
i�i�iRi�
iSiTi�i(i�ii�iC
i�iUiiVi�iWiXi�i�iYi�i�i�iZ�i�idi[iciHi�i_i�i#i0	i\ii]i�ini

i�	i�i�ii^ii�i_ik	iD
i�i`iai�i�ibi�i�i�	ici�i�i�i�idi�ii�	i�iiidi^iE
ieiii�i�
ii�i�ifi�i_i7iii�i i�
i�iF
iiigi-i�ii!ii�ihi�	iqi;i�i�i�i�iXi�i i	i�i�iiiji�i�i�i
ikili�i�i~imiG
i`iniyipi�i�ioi>i�ipi�iqi
i�i�iri�iai�i�i�isiti"iui|i�i?ii�iivi�i�i�i�i�iei�i�iwixiyi�i�i�i�izi{i@i|i�i�	i}i�i�	i�i�i%isiiR
i�i.i�iMi?i�
i�i~ii�ibii�i@i�i�i�ii�i�i�i�i�i�i�i�i�i�i�
i�i6
iIi�i�i�i�i�i�i�i�ici�i�i�i�i�i�i�i2ii�i�imi�i@i�i�i�i�	i�i�iNii�i�ifiii�iZi�
i�i�i�i�i�i�i\iH
i�idi�i�iJi�i�	i�i�	i�
i�i�i�i�i�iKi�i�i�i#i�i�i�i�i�i�	i�i�i�iKi�i�i�i�igi�i�i�ihi�i>i�i.i�i�i�i�i�i�i�ii)in
�ti�iii�i�
io
i�
iI
i�i�
i�iiqi�i�iri�i�i���i�i�ip
i�i�	i�i�iziii�iUi�iJ
i�i�i1	iK
iti�i�i�i�i���i�
i�i�iJi�iiLi�	i�i�i�i�i�i�N)� EUCTW_TYPICAL_DISTRIBUTION_RATIO�EUCTW_TABLE_SIZE�EUCTW_CHAR_TO_FREQ_ORDER�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/euctwfreq.py�<module>,ssite-packages/pip/_vendor/chardet/__pycache__/hebrewprober.cpython-39.pyc000064400000005761150564745010022471 0ustar00a

��Re6�@s,ddlmZddlmZGdd�de�ZdS)�)�
CharSetProber)�ProbingStatecs�eZdZdZdZdZdZdZdZdZ	dZ
d	Zd
ZdZ
dZd
ZdZ�fdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zedd��Zedd��Zedd ��Z�ZS)!�HebrewProber�������������������g{�G�z�?z
ISO-8859-8zwindows-1255cs>tt|���d|_d|_d|_d|_d|_d|_|�	�dS�N)
�superr�__init__�_final_char_logical_score�_final_char_visual_score�_prev�_before_prev�_logical_prober�_visual_prober�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/hebrewprober.pyr�szHebrewProber.__init__cCsd|_d|_d|_d|_dS)N�� )rrrrrrrrr�szHebrewProber.resetcCs||_||_dSr)rr)r�
logicalProber�visualProberrrr�set_model_probers�szHebrewProber.set_model_proberscCs||j|j|j|j|jfvSr)�	FINAL_KAF�	FINAL_MEM�	FINAL_NUN�FINAL_PE�FINAL_TSADI�r�crrr�is_final�s�zHebrewProber.is_finalcCs||j|j|j|jfvSr)�
NORMAL_KAF�
NORMAL_MEM�
NORMAL_NUN�	NORMAL_PEr*rrr�is_non_final�s
�zHebrewProber.is_non_finalcCs�|jtjkrtjS|�|�}|D]�}|dkrn|jdkr�|�|j�rR|jd7_q�|�|j�r�|j	d7_	n,|jdkr�|�|j�r�|dkr�|j	d7_	|j|_||_q tj
S)Nr!r)�stater�NOT_ME�filter_high_byte_onlyrr,rrr1r�	DETECTING)r�byte_str�currrr�feed�s&



��zHebrewProber.feedcCsx|j|j}||jkr|jS||jkr.|jS|j��|j��}||jkrR|jS||jkrd|jS|dkrr|jS|jS)Ng)	rr�MIN_FINAL_CHAR_DISTANCE�LOGICAL_HEBREW_NAME�VISUAL_HEBREW_NAMEr�get_confidencer�MIN_MODEL_DISTANCE)r�finalsub�modelsubrrr�charset_name�s
�
zHebrewProber.charset_namecCsdS)N�Hebrewrrrrr�languageszHebrewProber.languagecCs(|jjtjkr"|jjtjkr"tjStjSr)rr2rr3rr5rrrrr2s
�zHebrewProber.state)�__name__�
__module__�__qualname__r%r-r&r.r'r/r(r0r)�NORMAL_TSADIr9r=r;r:rrr$r,r1r8�propertyr@rBr2�
__classcell__rrrrr�s4

;

rN)�
charsetproberr�enumsrrrrrr�<module>scsite-packages/pip/_vendor/chardet/__pycache__/jpcntx.cpython-39.pyc000064400000111416150564745010021304 0ustar00a

��Re�L�@s8dZGdd�de�ZGdd�de�ZGdd�de�ZdS))S)S�rr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Sr�rrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)Srrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrc@sPeZdZdZdZdZdZdZdd�Zdd	�Z	d
d�Z
dd
�Zdd�Zdd�Z
dS)�JapaneseContextAnalysis�����di�rcCs*d|_d|_d|_d|_d|_|��dS�N)�
_total_rel�_rel_sample�_need_to_skip_char_num�_last_char_order�_done�reset��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/jpcntx.py�__init__{sz JapaneseContextAnalysis.__init__cCs*d|_dg|j|_d|_d|_d|_dS)Nrr
F)r
�NUM_OF_CATEGORYrrrrrrrrr�s
zJapaneseContextAnalysis.resetcCs�|jr
dS|j}||kr�|�|||d��\}}||7}||krT|||_d|_q|dkr�|jdkr�|jd7_|j|jkr�d|_q�|jt|j|d7<||_qdS)Nrr
rT)rr�	get_orderrr
�MAX_REL_THRESHOLDr�jp2CharContext)r�byte_str�	num_bytes�i�order�char_lenrrr�feed�s 	
zJapaneseContextAnalysis.feedcCs|j|jkSr)r
�ENOUGH_REL_THRESHOLDrrrr�got_enough_data�sz'JapaneseContextAnalysis.got_enough_datacCs,|j|jkr"|j|jd|jS|jSdS)Nr)r
�MINIMUM_DATA_THRESHOLDr�	DONT_KNOWrrrr�get_confidence�sz&JapaneseContextAnalysis.get_confidencecCsdS)N�r
rr)rrrrrr�sz!JapaneseContextAnalysis.get_orderN)�__name__�
__module__�__qualname__rr%r"rr$rrr!r#r&rrrrrrtsrcs0eZdZ�fdd�Zedd��Zdd�Z�ZS)�SJISContextAnalysiscstt|���d|_dS)N�	SHIFT_JIS)�superr+r�
_charset_namer��	__class__rrr�szSJISContextAnalysis.__init__cCs|jSr)r.rrrr�charset_name�sz SJISContextAnalysis.charset_namecCs�|sdS|d}d|kr$dks>nd|kr:dkrjnn,d}|dksbd	|kr^dkrnnqnd
|_nd}t|�dkr�|d}|dkr�d|kr�d
kr�nn|d|fSd|fS)Nr'r�����r���CP932r����r
)r.�len�rr�
first_charr �second_charrrrr�s.  zSJISContextAnalysis.get_order)r(r)r*r�propertyr1r�
__classcell__rrr/rr+�s
r+c@seZdZdd�ZdS)�EUCJPContextAnalysiscCs�|sdS|d}|dks0d|kr,dkr6nnd}n|dkrDd}nd	}t|�d	kr�|d	}|d
kr�d|krxdkr�nn|d|fSd|fS)
Nr'r���r�rr���r
)r;r<rrrr�s  zEUCJPContextAnalysis.get_orderN)r(r)r*rrrrrrA�srAN)r�objectrr+rArrrr�<module>sVCsite-packages/pip/_vendor/chardet/__pycache__/universaldetector.cpython-39.pyc000064400000013355150564745010023543 0ustar00a

��Re�0�@s�dZddlZddlZddlZddlmZddlmZmZm	Z	ddl
mZddlm
Z
ddlmZdd	lmZGd
d�de�ZdS)a
Module containing the UniversalDetector detector class, which is the primary
class a user of ``chardet`` should use.

:author: Mark Pilgrim (initial port to Python)
:author: Shy Shalom (original C code)
:author: Dan Blanchard (major refactoring for 3.0)
:author: Ian Cordasco
�N�)�CharSetGroupProber)�
InputState�LanguageFilter�ProbingState)�EscCharSetProber)�Latin1Prober)�MBCSGroupProber)�SBCSGroupProberc	@sneZdZdZdZe�d�Ze�d�Ze�d�Z	dddd	d
ddd
d�Z
ejfdd�Z
dd�Zdd�Zdd�ZdS)�UniversalDetectoraq
    The ``UniversalDetector`` class underlies the ``chardet.detect`` function
    and coordinates all of the different charset probers.

    To get a ``dict`` containing an encoding and its confidence, you can simply
    run:

    .. code::

            u = UniversalDetector()
            u.feed(some_bytes)
            u.close()
            detected = u.result

    g�������?s[�-�]s(|~{)s[�-�]zWindows-1252zWindows-1250zWindows-1251zWindows-1256zWindows-1253zWindows-1255zWindows-1254zWindows-1257)z
iso-8859-1z
iso-8859-2z
iso-8859-5z
iso-8859-6z
iso-8859-7z
iso-8859-8z
iso-8859-9ziso-8859-13cCsNd|_g|_d|_d|_d|_d|_d|_||_t�	t
�|_d|_|�
�dS)N)�_esc_charset_prober�_charset_probers�result�done�	_got_data�_input_state�
_last_char�lang_filter�logging�	getLogger�__name__�logger�_has_win_bytes�reset)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/universaldetector.py�__init__QszUniversalDetector.__init__cCsVdddd�|_d|_d|_d|_tj|_d|_|jr>|j�	�|j
D]}|�	�qDdS)z�
        Reset the UniversalDetector and all of its probers back to their
        initial states.  This is called by ``__init__``, so you only need to
        call this directly in between analyses of different documents.
        N���encoding�
confidence�languageF�)rrrrr�
PURE_ASCIIrrrrr
)r�proberrrrr^s

zUniversalDetector.resetcCs>|jr
dSt|�sdSt|t�s(t|�}|js�|�tj�rJdddd�|_nv|�tj	tj
f�rldddd�|_nT|�d�r�dddd�|_n:|�d	�r�d
ddd�|_n |�tjtjf�r�dddd�|_d|_|jd
dur�d|_dS|j
tjk�r.|j�|��rtj|_
n*|j
tjk�r.|j�|j|��r.tj|_
|dd�|_|j
tjk�r�|j�s^t|j�|_|j�|�tjk�r:|jj|j��|jjd�|_d|_n�|j
tjk�r:|j�s�t |j�g|_|jt!j"@�r�|j�#t$��|j�#t%��|jD]:}|�|�tjk�r�|j|��|jd�|_d|_�q&�q�|j&�|��r:d|_'dS)a�
        Takes a chunk of a document and feeds it through all of the relevant
        charset probers.

        After calling ``feed``, you can check the value of the ``done``
        attribute to see if you need to continue feeding the
        ``UniversalDetector`` more data, or if it has made a prediction
        (in the ``result`` attribute).

        .. note::
           You should always call ``close`` when you're done feeding in your
           document if ``done`` is not already ``True``.
        Nz	UTF-8-SIG��?�rzUTF-32s��zX-ISO-10646-UCS-4-3412s��zX-ISO-10646-UCS-4-2143zUTF-16Tr ���)(r�len�
isinstance�	bytearrayr�
startswith�codecs�BOM_UTF8r�BOM_UTF32_LE�BOM_UTF32_BE�BOM_LE�BOM_BErrr$�HIGH_BYTE_DETECTOR�search�	HIGH_BYTE�ESC_DETECTORr�	ESC_ASCIIrrr�feedr�FOUND_IT�charset_name�get_confidencer"r
r	r�NON_CJK�appendr
r�WIN_BYTE_DETECTORr)r�byte_strr%rrrr8os�
�
��

�

�
�
��
�
zUniversalDetector.feedc		Cst|jr|jSd|_|js&|j�d�n�|jtjkrBdddd�|_n�|jtjkr�d}d}d}|j	D]"}|sjq`|�
�}||kr`|}|}q`|r�||jkr�|j}|j�
�}|�
�}|�d	�r�|jr�|j�||�}|||jd�|_|j��tjk�rn|jd
du�rn|j�d�|j	D]`}|�s�qt|t��rP|jD] }|j�d|j|j|�
���q,n|j�d|j|j|�
���q|jS)
z�
        Stop analyzing the current document and come up with a final
        prediction.

        :returns:  The ``result`` attribute, a ``dict`` with the keys
                   `encoding`, `confidence`, and `language`.
        Tzno data received!�asciir&r'rNrziso-8859r z no probers hit minimum thresholdz%s %s confidence = %s)rrrr�debugrrr$r5r
r;�MINIMUM_THRESHOLDr:�lowerr,r�ISO_WIN_MAP�getr"�getEffectiveLevelr�DEBUGr*r�probers)	r�prober_confidence�max_prober_confidence�
max_proberr%r:�lower_charset_namer!�group_proberrrr�close�sj	�



��

�
�zUniversalDetector.closeN)r�
__module__�__qualname__�__doc__rB�re�compiler3r6r>rDr�ALLrrr8rNrrrrr3s$


�	
mr)rQr-rrR�charsetgroupproberr�enumsrrr�	escproberr�latin1proberr�mbcsgroupproberr	�sbcsgroupproberr
�objectrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/codingstatemachine.cpython-39.pyc000064400000005604150564745010023630 0ustar00a

��Re�@s(ddlZddlmZGdd�de�ZdS)�N�)�MachineStatec@sDeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Ze	dd
��Z
dS)�CodingStateMachinea�
    A state machine to verify a byte sequence for a particular encoding. For
    each byte the detector receives, it will feed that byte to every active
    state machine available, one byte at a time. The state machine changes its
    state based on its previous state and the byte it receives. There are 3
    states in a state machine that are of interest to an auto-detector:

    START state: This is the state to start with, or a legal byte sequence
                 (i.e. a valid code point) for character has been identified.

    ME state:  This indicates that the state machine identified a byte sequence
               that is specific to the charset it is designed for and that
               there is no other possible encoding which can contain this byte
               sequence. This will to lead to an immediate positive answer for
               the detector.

    ERROR state: This indicates the state machine identified an illegal byte
                 sequence for that encoding. This will lead to an immediate
                 negative answer for this encoding. Detector will exclude this
                 encoding from consideration from here on.
    cCs0||_d|_d|_d|_t�t�|_|��dS)Nr)	�_model�_curr_byte_pos�_curr_char_len�_curr_state�logging�	getLogger�__name__�logger�reset)�self�sm�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/codingstatemachine.py�__init__7szCodingStateMachine.__init__cCstj|_dS�N)r�STARTr�rrrrr
?szCodingStateMachine.resetcCsh|jd|}|jtjkr0d|_|jd||_|j|jd|}|jd||_|jd7_|jS)N�class_tabler�char_len_table�class_factor�state_tabler)rrrrrr)r�c�
byte_class�
curr_staterrr�
next_stateBs�zCodingStateMachine.next_statecCs|jSr)rrrrr�get_current_charlenPsz&CodingStateMachine.get_current_charlencCs
|jdS)N�name�rrrrr�get_coding_state_machineSsz+CodingStateMachine.get_coding_state_machinecCs
|jdS)N�languager rrrrr"VszCodingStateMachine.languageN)r�
__module__�__qualname__�__doc__rr
rrr!�propertyr"rrrrr!sr)r	�enumsr�objectrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/latin1prober.cpython-39.pyc000064400000005661150564745010022404 0ustar00a

��Re��@s^ddlmZddlmZdZdZdZdZdZdZ	dZ
dZd	Zd
Z
eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee
ee
ee
eeeeeeeeeeeeeeeee
eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee	e	e	e	e	e	e
e
e	e	e	e	e	e	e	e	e
e
e	e	e	e	e	ee	e	e	e	e	e
e
e
eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee�fZdZGdd
�d
e�ZdS)�)�
CharSetProber)�ProbingState��������)@rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrcsLeZdZ�fdd�Zdd�Zedd��Zedd��Zd	d
�Zdd�Z	�Z
S)
�Latin1Probercs&tt|���d|_d|_|��dS)N)�superr�__init__�_last_char_class�
_freq_counter�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/latin1prober.pyraszLatin1Prober.__init__cCs t|_dgt|_t�|�dS)Nr)�OTHr�FREQ_CAT_NUMrrrrrrrrgszLatin1Prober.resetcCsdS)Nz
ISO-8859-1rrrrr�charset_namelszLatin1Prober.charset_namecCsdS)N�rrrrr�languagepszLatin1Prober.languagecCs`|�|�}|D]J}t|}t|jt|}|dkr@tj|_qZ|j|d7<||_q|j	S)Nrr)
�filter_with_english_letters�Latin1_CharToClass�Latin1ClassModelr�	CLASS_NUMr�NOT_ME�_stater�state)r�byte_str�c�
char_class�freqrrr�feedts

�zLatin1Prober.feedcCs\|jtjkrdSt|j�}|dkr(d}n|jd|jdd|}|dkrPd}|d}|S)Ng{�G�z�?grrg4@g\��(\�?)r#rr!�sumr)r�total�
confidencerrr�get_confidence�s
�zLatin1Prober.get_confidence)�__name__�
__module__�__qualname__rr�propertyrrr(r,�
__classcell__rrrrr`s

rN)�
charsetproberr�enumsrr�UDFr�ASC�ASS�ACV�ACO�ASV�ASOr rrrrrrr�<module>s\�'
site-packages/pip/_vendor/chardet/__pycache__/gb2312prober.cpython-39.pyc000064400000002244150564745010022106 0ustar00a

��Re��@sDddlmZddlmZddlmZddlmZGdd�de�ZdS)�)�MultiByteCharSetProber)�CodingStateMachine)�GB2312DistributionAnalysis)�GB2312_SM_MODELcs4eZdZ�fdd�Zedd��Zedd��Z�ZS)�GB2312Probercs,tt|���tt�|_t�|_|��dS)N)	�superr�__init__rr�	coding_smr�distribution_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/gb2312prober.pyr"s
zGB2312Prober.__init__cCsdS)N�GB2312rrrrr�charset_name(szGB2312Prober.charset_namecCsdS)N�Chineserrrrr�language,szGB2312Prober.language)�__name__�
__module__�__qualname__r�propertyrr�
__classcell__rrrrr!s

rN)	�mbcharsetproberr�codingstatemachiner�chardistributionr�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/__init__.cpython-39.pyc000064400000003622150564745010021534 0ustar00a

��Re��@sDddlmZddlmZddlmZmZgd�Zdd�Zdd�Z	d	S)
�)�UniversalDetector)�
InputState)�__version__�VERSION)r�detect�
detect_allrrcCsHt|t�s0t|t�s(td�t|����nt|�}t�}|�|�|��S)z�
    Detect the encoding of the given byte string.

    :param byte_str:     The byte sequence to examine.
    :type byte_str:      ``bytes`` or ``bytearray``
    �3Expected object of type bytes or bytearray, got: {})	�
isinstance�	bytearray�bytes�	TypeError�format�typer�feed�close)�byte_str�detector�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/__init__.pyrs

�
rcCs�t|t�s0t|t�s(td�t|����nt|�}t�}|�|�|��|j	t
jkr�g}|jD]X}|�
�|jkr^|j}|j��}|�d�r�|jr�|j�||�}|�||�
�|jd��q^t|�dkr�t|dd�d�S|jgS)z�
    Detect all the possible encodings of the given byte string.

    :param byte_str:     The byte sequence to examine.
    :type byte_str:      ``bytes`` or ``bytearray``
    rziso-8859)�encoding�
confidence�language�cSs
|dS)Nrr)�resultrrr�<lambda>Q�zdetect_all.<locals>.<lambda>)�key)r	r
rrr
rrrr�_input_stater�	HIGH_BYTE�_charset_probers�get_confidence�MINIMUM_THRESHOLD�charset_name�lower�
startswith�_has_win_bytes�ISO_WIN_MAP�get�appendr�len�sortedr)rr�results�proberr"�lower_charset_namerrrr-s8

�



��
rN)
�universaldetectorr�enumsr�versionrr�__all__rrrrrr�<module>s
site-packages/pip/_vendor/chardet/__pycache__/langhebrewmodel.cpython-39.pyc000064400000050160150564745010023133 0ustar00a

��Re�@s2"ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=dddddd>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvdwdxdydzd{d|d}d~dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�ddd�d�d�dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	d�d�dd���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel��)?�2�<�=�*�5�8�6�1�3�+�,�?�"�7�0�'�9��;�)�!�%�$���#�>��&�-�	����r���r���������
�������
�� �4�/�.�:�(r'�����E�[�O�P�\�Y�a�Z�D�o�p�R�I�_�U�N�y�V�G�C�f�k�T�r�g�sr�Jrrr�L�F�@r	�i�]r
�Arr�B�nr
rrr�Q�M�b�K�l�|��������rAr@���������������������Sr=r>r?�Hr<�^���q���m��������r�t���v�d�����u�w�h�}�����W�c���j�z�{��r�����e�����x��rrr��rrr�Xrrrrrr��rr ���~����r!r"�������������������r#r$r%r&r(r)r*r+r,r-r.r/r0r1r2r3r4r5r6r7r8r9r:r;���`(rrrr'r-r;r/r9r$r#r:r.r1r3r)r,r&r8r5r2r%r7r*r0r(r+r4r6r rrrr<rrrrrr!rrArrrrr"r?r>rrrr
r=r	rrr
rr@rrrrrrcrfrgrYrNrFrbrXr~rRr`rlrarjrUrHrIrirQr}r\rTrWr�r�rKrMrGrJrerrSr�rLrkr�r�r�rZr^r�rdr�r[rmr�rhrOrPr�r]r_r�r�r�r�r�rVr�r�rnr�r�r�r����������������������������������������������������������������������������������rorprqrrrsrtrurvrwrxryrzr{r|r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rErDrCrBzwindows-1255�HebrewgC��|�?Fu<אבגדהוזחטיךכלםמןנסעףפץצקרשתװױײ)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetN)�#pip._vendor.chardet.sbcharsetproberr�HEBREW_LANG_MODEL�!WINDOWS_1255_HEBREW_CHAR_TO_ORDER�WINDOWS_1255_HEBREW_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langhebrewmodel.py�<module>s"
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���site-packages/pip/_vendor/chardet/__pycache__/escsm.cpython-39.pyc000064400000015720150564745010021111 0ustar00a

��Re)�H@s�ddlmZdZejejdejejejejejejejejejejejejejejejejejejejdejdejdejdddejdejdddejdejdejejejejejejejf0ZdZedeedd	d
�ZdZ	ejdejejejejejejejejejejejejejejejejejejejejejejejejejejejejdejejejejejejejejejddejejejejejejejejejejejejejejejejejejejejejejf@Z
dZe	d
e
edd	d
�ZdZ
ejdejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejejdejejejdejejejejejdejejejejejejejejejejejejejejejejejejejejejejejejejejejejfHZdZe
deeddd
�ZdZejdejejejejejejejejejejejejejejejejejejejdejejejejejejdejejejejejejejejejejejf(ZdZedeeddd
�ZdS)�)�MachineState(r�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�r��rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rr�)rrrrrrz
HZ-GB-2312�Chinese)�class_table�class_factor�state_table�char_len_table�name�language(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)	rrrrrrrrr�	zISO-2022-CN(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr�rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrr)
rrrrrrrrrr�
zISO-2022-JP�Japanese(rrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrzISO-2022-KR�KoreanN)�enumsr�HZ_CLS�START�ERROR�ITS_ME�HZ_ST�HZ_CHAR_LEN_TABLE�HZ_SM_MODEL�
ISO2022CN_CLS�ISO2022CN_ST�ISO2022CN_CHAR_LEN_TABLE�ISO2022CN_SM_MODEL�
ISO2022JP_CLS�ISO2022JP_ST�ISO2022JP_CHAR_LEN_TABLE�ISO2022JP_SM_MODEL�
ISO2022KR_CLS�ISO2022KR_ST�ISO2022KR_CHAR_LEN_TABLE�ISO2022KR_SM_MODEL�r*r*��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/escsm.py�<module>s�$ �	�$     ��$      ��$  ��site-packages/pip/_vendor/chardet/__pycache__/sbcsgroupprober.cpython-39.pyc000064400000003312150564745010023212 0ustar00a

��Re��@s�ddlmZddlmZddlmZmZddlmZm	Z	ddl
mZddlm
Z
mZmZmZmZmZddlmZddlmZdd	lmZGd
d�de�ZdS)
�)�CharSetGroupProber)�HebrewProber)�ISO_8859_5_BULGARIAN_MODEL�WINDOWS_1251_BULGARIAN_MODEL)�ISO_8859_7_GREEK_MODEL�WINDOWS_1253_GREEK_MODEL)�WINDOWS_1255_HEBREW_MODEL)�IBM855_RUSSIAN_MODEL�IBM866_RUSSIAN_MODEL�ISO_8859_5_RUSSIAN_MODEL�KOI8_R_RUSSIAN_MODEL�MACCYRILLIC_RUSSIAN_MODEL�WINDOWS_1251_RUSSIAN_MODEL)�TIS_620_THAI_MODEL)�ISO_8859_9_TURKISH_MODEL)�SingleByteCharSetProbercseZdZ�fdd�Z�ZS)�SBCSGroupProbercs�tt|���t�}ttd|�}ttd|�}|�||�tt�tt�tt	�tt
�tt�tt�tt
�tt�tt�tt�tt�tt�|||g|_|��dS)NFT)�superr�__init__rrr�set_model_probersrrrr
r
r	rrrrrr�probers�reset)�self�
hebrew_prober�logical_hebrew_prober�visual_hebrew_prober��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/sbcsgroupprober.pyr/s8����zSBCSGroupProber.__init__)�__name__�
__module__�__qualname__r�
__classcell__rrrrr.srN)�charsetgroupproberr�hebrewproberr�langbulgarianmodelrr�langgreekmodelrr�langhebrewmodelr�langrussianmodelr	r
rrr
r�
langthaimodelr�langturkishmodelr�sbcharsetproberrrrrrr�<module>s site-packages/pip/_vendor/chardet/__pycache__/langgreekmodel.cpython-39.pyc000064400000050054150564745010022756 0ustar00a

��Re݄�@sX"ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�=d�=Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddddddddddddddddddddddddddddddddddd>d?ddddddddddd@dddddddAdddBdCdDdEddFddGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbd-dcdddedfdgdhddidjdkddldmdndodpdqdrdsdtdudvdwdxdyddzd{d|d}d~dd�d�d�d�dd���Zed�d�eed�d�d�d��Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=ddddddddddddddddddddddddddddddddddddddd>d�ddddddddddd@dddddddAd�d?dBdCdDdEddFddGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbd-dcdddedfdgdhddidjdkddldmdndodpdqdrdsdtdudvdwdxdyddzd{d|d}d~dd�d�d�d�dd���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel��)=�<�7�:�$�=�.�6��3�+�)�"�(�4�/�,�5�&�1�;�'�#�0�%�!�-�8�2�9����r����� �
�����
����	���r����*����r)�����R�d�h�^�b�e�t�f�o��u�\�X�q�U�O�v�i�S�C�r�w�_�c�m��H�F�P�Qr�`�]�Y�D�x�a�M�V�Er�N�s�A�Br�L�j�g�W�k�p��r	�J�rr
�G�Ir�l�{�nrr
rrr�[rrrrrrrrrrr�rrrrr �Tr!�yr"r#r$r%�|r&r'r(r*r+r,r-r.r/r0r1r2r3r4r5r6r7r8r9r:r;r<�@�Kr=r>r?(rrrr)r3r-r1r7r5r4r0r.r8r+r6r%r/r"r#r=r'r(r$r:r<r,r>r?r9r&r2rr*rrrrrrrrrr;rrrr
rrrr r
rrrrrr!rrrr	�>�?r�rmrnrWrerjr_rxr^ryrvr�rorhrkrSr`rarDrVrrRrirrrPrd�Zr}rOrcrGrZrbrgrHr[rErIrKrqrFrUrprsrzr\r|rLrtrQrXrlrJrNrTrYrfr��zr{r��}�~������������������������������������������������������������rMr]�����������������������������������������������������������������������������������ru������������������������rw��r~�rCrBrAr@zwindows-1253�Greekg���s�?Fu~ΆΈΉΊΌΎΏΑΒΓΔΕΖΗΘΙΚΛΜΝΞΟΠΡΣΤΥΦΧΨΩάέήίαβγδεζηθικλμνξοπρςστυφχψωόύώ)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetr�r�z
ISO-8859-7N)�#pip._vendor.chardet.sbcharsetproberr�GREEK_LANG_MODEL� WINDOWS_1253_GREEK_CHAR_TO_ORDER�WINDOWS_1253_GREEK_MODEL�ISO_8859_7_GREEK_CHAR_TO_ORDER�ISO_8859_7_GREEK_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langgreekmodel.py�<module>s8"
�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�@�������������������������������,���	���site-packages/pip/_vendor/chardet/__pycache__/euctwprober.cpython-39.pyc000064400000002234150564745010022334 0ustar00a

��Re��@sDddlmZddlmZddlmZddlmZGdd�de�ZdS)�)�MultiByteCharSetProber)�CodingStateMachine)�EUCTWDistributionAnalysis)�EUCTW_SM_MODELcs4eZdZ�fdd�Zedd��Zedd��Z�ZS)�EUCTWProbercs,tt|���tt�|_t�|_|��dS)N)	�superr�__init__rr�	coding_smr�distribution_analyzer�reset��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/euctwprober.pyr"s
zEUCTWProber.__init__cCsdS)NzEUC-TWrrrrr�charset_name(szEUCTWProber.charset_namecCsdS)N�Taiwanrrrrr�language,szEUCTWProber.language)�__name__�
__module__�__qualname__r�propertyrr�
__classcell__rrrrr!s

rN)	�mbcharsetproberr�codingstatemachiner�chardistributionr�mbcssmrrrrrr�<module>ssite-packages/pip/_vendor/chardet/__pycache__/langthaimodel.cpython-39.pyc000064400000050440150564745010022605 0ustar00a

��Rez��@s2"ddlmZdddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddd�?d�?Zddddddddddddddddddddddddddddddddddddddddddddddddd	d	d	d	d	d	d	d	d	d	dddddddd
ddd
ddddddddddddddddddd d!d"d#ddddddd$d%d&d'd(d)d*d+d,d-d.d/d0d1d2d3d4d5d6d7d8d9d:d;d<d=dddddd>d?d@dAdBdCdDdEdFdGdHdIdJdKdLdMdNdOdPdQdRdSdTdUdVdWdXdYdZd[d\d]d^d_d`dadbdcdddedfdgdhdidjdkdldmdndodpdqdrdsdtdudvddwdxdydzd{d|d}d~ddd�d�d�d�d�d�d�d�d�d�d�d�d�d�dd�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d�d	dd���Zed�d�eed�d�d�d��Zd�S)��)�SingleByteCharSetModel���)?������4�"�3�/�:�9�1�5�7�+���,��0r���'�>��6�-�	�r�=���*�.����?��
r�$��
�(�� �#���)��!�2�%���&�8�;�<������j�k�d����e�^���l�m�n�o����Y�_�p�q������@�H�I�r�J�s�t�f�Q���u�Z�g�N�R�`���[�O�T�h�i�a�b�\�������������X���������������v���������c�U�S������������������������rr��r���Kr	r
rrr
�wrrrrrrrrrrrrrrrrrrr r!r"r#r$��r%r&r'r(r)�Lr*�Br+r,r-r.r/r0r1r2r3r4�V����������r5r6r7r8r9�r:r;r<r=�C�Mr>�]���Dr?r@�A�ErA�F�P�G�W����(rrrrr*rr<r=r	r!r-r5r%r0rr$r"rr(rrr)r,r/rrr
r2r6r8rrr3r9rr4r.r;r>rr1r7r&rrr r'rrrr:r
rrrrr?rrr@rAr#rr+r`r�r�r�r�r�r�r�rarbrdr�r�r�rmrrr�rhrnr�rsr�r�r�rrXrkrqrxr�rNrYrorvrwr�rIrMrgrlrtrurGrHrQrRrSrTrZr[rcrerfrjr�r��x�y�z�{�|�}�~�������������������������������������������������������rFrJrKrLrOrPrUrVrWr\r]r^r_������������rirpry����������rzr{r|r}r~r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rErDrCrBzTIS-620�Thaig��@��?Fuกขฃคฅฆงจฉชซฌญฎฏฐฑฒณดตถทธนบปผฝพฟภมยรฤลฦวศษสหฬอฮฯะัาำิีึืฺุู฿เแโใไๅๆ็่้๊๋์ํ๎๏๐๑๒๓๔๕๖๗๘๙๚๛)�charset_name�language�char_to_order_map�language_model�typical_positive_ratio�keep_ascii_letters�alphabetN)�#pip._vendor.chardet.sbcharsetproberr�THAI_LANG_MODEL�TIS_620_THAI_CHAR_TO_ORDER�TIS_620_THAI_MODEL�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/langthaimodel.py�<module>s"
�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B�B���������������������������������*���site-packages/pip/_vendor/chardet/euctwfreq.py000064400000075605150564745010015525 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

# EUCTW frequency table
# Converted from big5 work
# by Taiwan's Mandarin Promotion Council
# <http:#www.edu.tw:81/mandr/>

# 128  --> 0.42261
# 256  --> 0.57851
# 512  --> 0.74851
# 1024 --> 0.89384
# 2048 --> 0.97583
#
# Idea Distribution Ratio = 0.74851/(1-0.74851) =2.98
# Random Distribution Ration = 512/(5401-512)=0.105
#
# Typical Distribution Ratio about 25% of Ideal one, still much higher than RDR

EUCTW_TYPICAL_DISTRIBUTION_RATIO = 0.75

# Char to FreqOrder table ,
EUCTW_TABLE_SIZE = 5376

EUCTW_CHAR_TO_FREQ_ORDER = (
   1,1800,1506, 255,1431, 198,   9,  82,   6,7310, 177, 202,3615,1256,2808, 110,  # 2742
3735,  33,3241, 261,  76,  44,2113,  16,2931,2184,1176, 659,3868,  26,3404,2643,  # 2758
1198,3869,3313,4060, 410,2211, 302, 590, 361,1963,   8, 204,  58,4296,7311,1931,  # 2774
  63,7312,7313, 317,1614,  75, 222, 159,4061,2412,1480,7314,3500,3068, 224,2809,  # 2790
3616,   3,  10,3870,1471,  29,2774,1135,2852,1939, 873, 130,3242,1123, 312,7315,  # 2806
4297,2051, 507, 252, 682,7316, 142,1914, 124, 206,2932,  34,3501,3173,  64, 604,  # 2822
7317,2494,1976,1977, 155,1990, 645, 641,1606,7318,3405, 337,  72, 406,7319,  80,  # 2838
 630, 238,3174,1509, 263, 939,1092,2644, 756,1440,1094,3406, 449,  69,2969, 591,  # 2854
 179,2095, 471, 115,2034,1843,  60,  50,2970, 134, 806,1868, 734,2035,3407, 180,  # 2870
 995,1607, 156, 537,2893, 688,7320, 319,1305, 779,2144, 514,2374, 298,4298, 359,  # 2886
2495,  90,2707,1338, 663,  11, 906,1099,2545,  20,2436, 182, 532,1716,7321, 732,  # 2902
1376,4062,1311,1420,3175,  25,2312,1056, 113, 399, 382,1949, 242,3408,2467, 529,  # 2918
3243, 475,1447,3617,7322, 117,  21, 656, 810,1297,2295,2329,3502,7323, 126,4063,  # 2934
 706, 456, 150, 613,4299,  71,1118,2036,4064, 145,3069,  85, 835, 486,2114,1246,  # 2950
1426, 428, 727,1285,1015, 800, 106, 623, 303,1281,7324,2127,2354, 347,3736, 221,  # 2966
3503,3110,7325,1955,1153,4065,  83, 296,1199,3070, 192, 624,  93,7326, 822,1897,  # 2982
2810,3111, 795,2064, 991,1554,1542,1592,  27,  43,2853, 859, 139,1456, 860,4300,  # 2998
 437, 712,3871, 164,2392,3112, 695, 211,3017,2096, 195,3872,1608,3504,3505,3618,  # 3014
3873, 234, 811,2971,2097,3874,2229,1441,3506,1615,2375, 668,2076,1638, 305, 228,  # 3030
1664,4301, 467, 415,7327, 262,2098,1593, 239, 108, 300, 200,1033, 512,1247,2077,  # 3046
7328,7329,2173,3176,3619,2673, 593, 845,1062,3244,  88,1723,2037,3875,1950, 212,  # 3062
 266, 152, 149, 468,1898,4066,4302,  77, 187,7330,3018,  37,   5,2972,7331,3876,  # 3078
7332,7333,  39,2517,4303,2894,3177,2078,  55, 148,  74,4304, 545, 483,1474,1029,  # 3094
1665, 217,1869,1531,3113,1104,2645,4067,  24, 172,3507, 900,3877,3508,3509,4305,  # 3110
  32,1408,2811,1312, 329, 487,2355,2247,2708, 784,2674,   4,3019,3314,1427,1788,  # 3126
 188, 109, 499,7334,3620,1717,1789, 888,1217,3020,4306,7335,3510,7336,3315,1520,  # 3142
3621,3878, 196,1034, 775,7337,7338, 929,1815, 249, 439,  38,7339,1063,7340, 794,  # 3158
3879,1435,2296,  46, 178,3245,2065,7341,2376,7342, 214,1709,4307, 804,  35, 707,  # 3174
 324,3622,1601,2546, 140, 459,4068,7343,7344,1365, 839, 272, 978,2257,2572,3409,  # 3190
2128,1363,3623,1423, 697, 100,3071,  48,  70,1231, 495,3114,2193,7345,1294,7346,  # 3206
2079, 462, 586,1042,3246, 853, 256, 988, 185,2377,3410,1698, 434,1084,7347,3411,  # 3222
 314,2615,2775,4308,2330,2331, 569,2280, 637,1816,2518, 757,1162,1878,1616,3412,  # 3238
 287,1577,2115, 768,4309,1671,2854,3511,2519,1321,3737, 909,2413,7348,4069, 933,  # 3254
3738,7349,2052,2356,1222,4310, 765,2414,1322, 786,4311,7350,1919,1462,1677,2895,  # 3270
1699,7351,4312,1424,2437,3115,3624,2590,3316,1774,1940,3413,3880,4070, 309,1369,  # 3286
1130,2812, 364,2230,1653,1299,3881,3512,3882,3883,2646, 525,1085,3021, 902,2000,  # 3302
1475, 964,4313, 421,1844,1415,1057,2281, 940,1364,3116, 376,4314,4315,1381,   7,  # 3318
2520, 983,2378, 336,1710,2675,1845, 321,3414, 559,1131,3022,2742,1808,1132,1313,  # 3334
 265,1481,1857,7352, 352,1203,2813,3247, 167,1089, 420,2814, 776, 792,1724,3513,  # 3350
4071,2438,3248,7353,4072,7354, 446, 229, 333,2743, 901,3739,1200,1557,4316,2647,  # 3366
1920, 395,2744,2676,3740,4073,1835, 125, 916,3178,2616,4317,7355,7356,3741,7357,  # 3382
7358,7359,4318,3117,3625,1133,2547,1757,3415,1510,2313,1409,3514,7360,2145, 438,  # 3398
2591,2896,2379,3317,1068, 958,3023, 461, 311,2855,2677,4074,1915,3179,4075,1978,  # 3414
 383, 750,2745,2617,4076, 274, 539, 385,1278,1442,7361,1154,1964, 384, 561, 210,  # 3430
  98,1295,2548,3515,7362,1711,2415,1482,3416,3884,2897,1257, 129,7363,3742, 642,  # 3446
 523,2776,2777,2648,7364, 141,2231,1333,  68, 176, 441, 876, 907,4077, 603,2592,  # 3462
 710, 171,3417, 404, 549,  18,3118,2393,1410,3626,1666,7365,3516,4319,2898,4320,  # 3478
7366,2973, 368,7367, 146, 366,  99, 871,3627,1543, 748, 807,1586,1185,  22,2258,  # 3494
 379,3743,3180,7368,3181, 505,1941,2618,1991,1382,2314,7369, 380,2357, 218, 702,  # 3510
1817,1248,3418,3024,3517,3318,3249,7370,2974,3628, 930,3250,3744,7371,  59,7372,  # 3526
 585, 601,4078, 497,3419,1112,1314,4321,1801,7373,1223,1472,2174,7374, 749,1836,  # 3542
 690,1899,3745,1772,3885,1476, 429,1043,1790,2232,2116, 917,4079, 447,1086,1629,  # 3558
7375, 556,7376,7377,2020,1654, 844,1090, 105, 550, 966,1758,2815,1008,1782, 686,  # 3574
1095,7378,2282, 793,1602,7379,3518,2593,4322,4080,2933,2297,4323,3746, 980,2496,  # 3590
 544, 353, 527,4324, 908,2678,2899,7380, 381,2619,1942,1348,7381,1341,1252, 560,  # 3606
3072,7382,3420,2856,7383,2053, 973, 886,2080, 143,4325,7384,7385, 157,3886, 496,  # 3622
4081,  57, 840, 540,2038,4326,4327,3421,2117,1445, 970,2259,1748,1965,2081,4082,  # 3638
3119,1234,1775,3251,2816,3629, 773,1206,2129,1066,2039,1326,3887,1738,1725,4083,  # 3654
 279,3120,  51,1544,2594, 423,1578,2130,2066, 173,4328,1879,7386,7387,1583, 264,  # 3670
 610,3630,4329,2439, 280, 154,7388,7389,7390,1739, 338,1282,3073, 693,2857,1411,  # 3686
1074,3747,2440,7391,4330,7392,7393,1240, 952,2394,7394,2900,1538,2679, 685,1483,  # 3702
4084,2468,1436, 953,4085,2054,4331, 671,2395,  79,4086,2441,3252, 608, 567,2680,  # 3718
3422,4087,4088,1691, 393,1261,1791,2396,7395,4332,7396,7397,7398,7399,1383,1672,  # 3734
3748,3182,1464, 522,1119, 661,1150, 216, 675,4333,3888,1432,3519, 609,4334,2681,  # 3750
2397,7400,7401,7402,4089,3025,   0,7403,2469, 315, 231,2442, 301,3319,4335,2380,  # 3766
7404, 233,4090,3631,1818,4336,4337,7405,  96,1776,1315,2082,7406, 257,7407,1809,  # 3782
3632,2709,1139,1819,4091,2021,1124,2163,2778,1777,2649,7408,3074, 363,1655,3183,  # 3798
7409,2975,7410,7411,7412,3889,1567,3890, 718, 103,3184, 849,1443, 341,3320,2934,  # 3814
1484,7413,1712, 127,  67, 339,4092,2398, 679,1412, 821,7414,7415, 834, 738, 351,  # 3830
2976,2146, 846, 235,1497,1880, 418,1992,3749,2710, 186,1100,2147,2746,3520,1545,  # 3846
1355,2935,2858,1377, 583,3891,4093,2573,2977,7416,1298,3633,1078,2549,3634,2358,  # 3862
  78,3750,3751, 267,1289,2099,2001,1594,4094, 348, 369,1274,2194,2175,1837,4338,  # 3878
1820,2817,3635,2747,2283,2002,4339,2936,2748, 144,3321, 882,4340,3892,2749,3423,  # 3894
4341,2901,7417,4095,1726, 320,7418,3893,3026, 788,2978,7419,2818,1773,1327,2859,  # 3910
3894,2819,7420,1306,4342,2003,1700,3752,3521,2359,2650, 787,2022, 506, 824,3636,  # 3926
 534, 323,4343,1044,3322,2023,1900, 946,3424,7421,1778,1500,1678,7422,1881,4344,  # 3942
 165, 243,4345,3637,2521, 123, 683,4096, 764,4346,  36,3895,1792, 589,2902, 816,  # 3958
 626,1667,3027,2233,1639,1555,1622,3753,3896,7423,3897,2860,1370,1228,1932, 891,  # 3974
2083,2903, 304,4097,7424, 292,2979,2711,3522, 691,2100,4098,1115,4347, 118, 662,  # 3990
7425, 611,1156, 854,2381,1316,2861,   2, 386, 515,2904,7426,7427,3253, 868,2234,  # 4006
1486, 855,2651, 785,2212,3028,7428,1040,3185,3523,7429,3121, 448,7430,1525,7431,  # 4022
2164,4348,7432,3754,7433,4099,2820,3524,3122, 503, 818,3898,3123,1568, 814, 676,  # 4038
1444, 306,1749,7434,3755,1416,1030, 197,1428, 805,2821,1501,4349,7435,7436,7437,  # 4054
1993,7438,4350,7439,7440,2195,  13,2779,3638,2980,3124,1229,1916,7441,3756,2131,  # 4070
7442,4100,4351,2399,3525,7443,2213,1511,1727,1120,7444,7445, 646,3757,2443, 307,  # 4086
7446,7447,1595,3186,7448,7449,7450,3639,1113,1356,3899,1465,2522,2523,7451, 519,  # 4102
7452, 128,2132,  92,2284,1979,7453,3900,1512, 342,3125,2196,7454,2780,2214,1980,  # 4118
3323,7455, 290,1656,1317, 789, 827,2360,7456,3758,4352, 562, 581,3901,7457, 401,  # 4134
4353,2248,  94,4354,1399,2781,7458,1463,2024,4355,3187,1943,7459, 828,1105,4101,  # 4150
1262,1394,7460,4102, 605,4356,7461,1783,2862,7462,2822, 819,2101, 578,2197,2937,  # 4166
7463,1502, 436,3254,4103,3255,2823,3902,2905,3425,3426,7464,2712,2315,7465,7466,  # 4182
2332,2067,  23,4357, 193, 826,3759,2102, 699,1630,4104,3075, 390,1793,1064,3526,  # 4198
7467,1579,3076,3077,1400,7468,4105,1838,1640,2863,7469,4358,4359, 137,4106, 598,  # 4214
3078,1966, 780, 104, 974,2938,7470, 278, 899, 253, 402, 572, 504, 493,1339,7471,  # 4230
3903,1275,4360,2574,2550,7472,3640,3029,3079,2249, 565,1334,2713, 863,  41,7473,  # 4246
7474,4361,7475,1657,2333,  19, 463,2750,4107, 606,7476,2981,3256,1087,2084,1323,  # 4262
2652,2982,7477,1631,1623,1750,4108,2682,7478,2864, 791,2714,2653,2334, 232,2416,  # 4278
7479,2983,1498,7480,2654,2620, 755,1366,3641,3257,3126,2025,1609, 119,1917,3427,  # 4294
 862,1026,4109,7481,3904,3760,4362,3905,4363,2260,1951,2470,7482,1125, 817,4110,  # 4310
4111,3906,1513,1766,2040,1487,4112,3030,3258,2824,3761,3127,7483,7484,1507,7485,  # 4326
2683, 733,  40,1632,1106,2865, 345,4113, 841,2524, 230,4364,2984,1846,3259,3428,  # 4342
7486,1263, 986,3429,7487, 735, 879, 254,1137, 857, 622,1300,1180,1388,1562,3907,  # 4358
3908,2939, 967,2751,2655,1349, 592,2133,1692,3324,2985,1994,4114,1679,3909,1901,  # 4374
2185,7488, 739,3642,2715,1296,1290,7489,4115,2198,2199,1921,1563,2595,2551,1870,  # 4390
2752,2986,7490, 435,7491, 343,1108, 596,  17,1751,4365,2235,3430,3643,7492,4366,  # 4406
 294,3527,2940,1693, 477, 979, 281,2041,3528, 643,2042,3644,2621,2782,2261,1031,  # 4422
2335,2134,2298,3529,4367, 367,1249,2552,7493,3530,7494,4368,1283,3325,2004, 240,  # 4438
1762,3326,4369,4370, 836,1069,3128, 474,7495,2148,2525, 268,3531,7496,3188,1521,  # 4454
1284,7497,1658,1546,4116,7498,3532,3533,7499,4117,3327,2684,1685,4118, 961,1673,  # 4470
2622, 190,2005,2200,3762,4371,4372,7500, 570,2497,3645,1490,7501,4373,2623,3260,  # 4486
1956,4374, 584,1514, 396,1045,1944,7502,4375,1967,2444,7503,7504,4376,3910, 619,  # 4502
7505,3129,3261, 215,2006,2783,2553,3189,4377,3190,4378, 763,4119,3763,4379,7506,  # 4518
7507,1957,1767,2941,3328,3646,1174, 452,1477,4380,3329,3130,7508,2825,1253,2382,  # 4534
2186,1091,2285,4120, 492,7509, 638,1169,1824,2135,1752,3911, 648, 926,1021,1324,  # 4550
4381, 520,4382, 997, 847,1007, 892,4383,3764,2262,1871,3647,7510,2400,1784,4384,  # 4566
1952,2942,3080,3191,1728,4121,2043,3648,4385,2007,1701,3131,1551,  30,2263,4122,  # 4582
7511,2026,4386,3534,7512, 501,7513,4123, 594,3431,2165,1821,3535,3432,3536,3192,  # 4598
 829,2826,4124,7514,1680,3132,1225,4125,7515,3262,4387,4126,3133,2336,7516,4388,  # 4614
4127,7517,3912,3913,7518,1847,2383,2596,3330,7519,4389, 374,3914, 652,4128,4129,  # 4630
 375,1140, 798,7520,7521,7522,2361,4390,2264, 546,1659, 138,3031,2445,4391,7523,  # 4646
2250, 612,1848, 910, 796,3765,1740,1371, 825,3766,3767,7524,2906,2554,7525, 692,  # 4662
 444,3032,2624, 801,4392,4130,7526,1491, 244,1053,3033,4131,4132, 340,7527,3915,  # 4678
1041,2987, 293,1168,  87,1357,7528,1539, 959,7529,2236, 721, 694,4133,3768, 219,  # 4694
1478, 644,1417,3331,2656,1413,1401,1335,1389,3916,7530,7531,2988,2362,3134,1825,  # 4710
 730,1515, 184,2827,  66,4393,7532,1660,2943, 246,3332, 378,1457, 226,3433, 975,  # 4726
3917,2944,1264,3537, 674, 696,7533, 163,7534,1141,2417,2166, 713,3538,3333,4394,  # 4742
3918,7535,7536,1186,  15,7537,1079,1070,7538,1522,3193,3539, 276,1050,2716, 758,  # 4758
1126, 653,2945,3263,7539,2337, 889,3540,3919,3081,2989, 903,1250,4395,3920,3434,  # 4774
3541,1342,1681,1718, 766,3264, 286,  89,2946,3649,7540,1713,7541,2597,3334,2990,  # 4790
7542,2947,2215,3194,2866,7543,4396,2498,2526, 181, 387,1075,3921, 731,2187,3335,  # 4806
7544,3265, 310, 313,3435,2299, 770,4134,  54,3034, 189,4397,3082,3769,3922,7545,  # 4822
1230,1617,1849, 355,3542,4135,4398,3336, 111,4136,3650,1350,3135,3436,3035,4137,  # 4838
2149,3266,3543,7546,2784,3923,3924,2991, 722,2008,7547,1071, 247,1207,2338,2471,  # 4854
1378,4399,2009, 864,1437,1214,4400, 373,3770,1142,2216, 667,4401, 442,2753,2555,  # 4870
3771,3925,1968,4138,3267,1839, 837, 170,1107, 934,1336,1882,7548,7549,2118,4139,  # 4886
2828, 743,1569,7550,4402,4140, 582,2384,1418,3437,7551,1802,7552, 357,1395,1729,  # 4902
3651,3268,2418,1564,2237,7553,3083,3772,1633,4403,1114,2085,4141,1532,7554, 482,  # 4918
2446,4404,7555,7556,1492, 833,1466,7557,2717,3544,1641,2829,7558,1526,1272,3652,  # 4934
4142,1686,1794, 416,2556,1902,1953,1803,7559,3773,2785,3774,1159,2316,7560,2867,  # 4950
4405,1610,1584,3036,2419,2754, 443,3269,1163,3136,7561,7562,3926,7563,4143,2499,  # 4966
3037,4406,3927,3137,2103,1647,3545,2010,1872,4144,7564,4145, 431,3438,7565, 250,  # 4982
  97,  81,4146,7566,1648,1850,1558, 160, 848,7567, 866, 740,1694,7568,2201,2830,  # 4998
3195,4147,4407,3653,1687, 950,2472, 426, 469,3196,3654,3655,3928,7569,7570,1188,  # 5014
 424,1995, 861,3546,4148,3775,2202,2685, 168,1235,3547,4149,7571,2086,1674,4408,  # 5030
3337,3270, 220,2557,1009,7572,3776, 670,2992, 332,1208, 717,7573,7574,3548,2447,  # 5046
3929,3338,7575, 513,7576,1209,2868,3339,3138,4409,1080,7577,7578,7579,7580,2527,  # 5062
3656,3549, 815,1587,3930,3931,7581,3550,3439,3777,1254,4410,1328,3038,1390,3932,  # 5078
1741,3933,3778,3934,7582, 236,3779,2448,3271,7583,7584,3657,3780,1273,3781,4411,  # 5094
7585, 308,7586,4412, 245,4413,1851,2473,1307,2575, 430, 715,2136,2449,7587, 270,  # 5110
 199,2869,3935,7588,3551,2718,1753, 761,1754, 725,1661,1840,4414,3440,3658,7589,  # 5126
7590, 587,  14,3272, 227,2598, 326, 480,2265, 943,2755,3552, 291, 650,1883,7591,  # 5142
1702,1226, 102,1547,  62,3441, 904,4415,3442,1164,4150,7592,7593,1224,1548,2756,  # 5158
 391, 498,1493,7594,1386,1419,7595,2055,1177,4416, 813, 880,1081,2363, 566,1145,  # 5174
4417,2286,1001,1035,2558,2599,2238, 394,1286,7596,7597,2068,7598,  86,1494,1730,  # 5190
3936, 491,1588, 745, 897,2948, 843,3340,3937,2757,2870,3273,1768, 998,2217,2069,  # 5206
 397,1826,1195,1969,3659,2993,3341, 284,7599,3782,2500,2137,2119,1903,7600,3938,  # 5222
2150,3939,4151,1036,3443,1904, 114,2559,4152, 209,1527,7601,7602,2949,2831,2625,  # 5238
2385,2719,3139, 812,2560,7603,3274,7604,1559, 737,1884,3660,1210, 885,  28,2686,  # 5254
3553,3783,7605,4153,1004,1779,4418,7606, 346,1981,2218,2687,4419,3784,1742, 797,  # 5270
1642,3940,1933,1072,1384,2151, 896,3941,3275,3661,3197,2871,3554,7607,2561,1958,  # 5286
4420,2450,1785,7608,7609,7610,3942,4154,1005,1308,3662,4155,2720,4421,4422,1528,  # 5302
2600, 161,1178,4156,1982, 987,4423,1101,4157, 631,3943,1157,3198,2420,1343,1241,  # 5318
1016,2239,2562, 372, 877,2339,2501,1160, 555,1934, 911,3944,7611, 466,1170, 169,  # 5334
1051,2907,2688,3663,2474,2994,1182,2011,2563,1251,2626,7612, 992,2340,3444,1540,  # 5350
2721,1201,2070,2401,1996,2475,7613,4424, 528,1922,2188,1503,1873,1570,2364,3342,  # 5366
3276,7614, 557,1073,7615,1827,3445,2087,2266,3140,3039,3084, 767,3085,2786,4425,  # 5382
1006,4158,4426,2341,1267,2176,3664,3199, 778,3945,3200,2722,1597,2657,7616,4427,  # 5398
7617,3446,7618,7619,7620,3277,2689,1433,3278, 131,  95,1504,3946, 723,4159,3141,  # 5414
1841,3555,2758,2189,3947,2027,2104,3665,7621,2995,3948,1218,7622,3343,3201,3949,  # 5430
4160,2576, 248,1634,3785, 912,7623,2832,3666,3040,3786, 654,  53,7624,2996,7625,  # 5446
1688,4428, 777,3447,1032,3950,1425,7626, 191, 820,2120,2833, 971,4429, 931,3202,  # 5462
 135, 664, 783,3787,1997, 772,2908,1935,3951,3788,4430,2909,3203, 282,2723, 640,  # 5478
1372,3448,1127, 922, 325,3344,7627,7628, 711,2044,7629,7630,3952,2219,2787,1936,  # 5494
3953,3345,2220,2251,3789,2300,7631,4431,3790,1258,3279,3954,3204,2138,2950,3955,  # 5510
3956,7632,2221, 258,3205,4432, 101,1227,7633,3280,1755,7634,1391,3281,7635,2910,  # 5526
2056, 893,7636,7637,7638,1402,4161,2342,7639,7640,3206,3556,7641,7642, 878,1325,  # 5542
1780,2788,4433, 259,1385,2577, 744,1183,2267,4434,7643,3957,2502,7644, 684,1024,  # 5558
4162,7645, 472,3557,3449,1165,3282,3958,3959, 322,2152, 881, 455,1695,1152,1340,  # 5574
 660, 554,2153,4435,1058,4436,4163, 830,1065,3346,3960,4437,1923,7646,1703,1918,  # 5590
7647, 932,2268, 122,7648,4438, 947, 677,7649,3791,2627, 297,1905,1924,2269,4439,  # 5606
2317,3283,7650,7651,4164,7652,4165,  84,4166, 112, 989,7653, 547,1059,3961, 701,  # 5622
3558,1019,7654,4167,7655,3450, 942, 639, 457,2301,2451, 993,2951, 407, 851, 494,  # 5638
4440,3347, 927,7656,1237,7657,2421,3348, 573,4168, 680, 921,2911,1279,1874, 285,  # 5654
 790,1448,1983, 719,2167,7658,7659,4441,3962,3963,1649,7660,1541, 563,7661,1077,  # 5670
7662,3349,3041,3451, 511,2997,3964,3965,3667,3966,1268,2564,3350,3207,4442,4443,  # 5686
7663, 535,1048,1276,1189,2912,2028,3142,1438,1373,2834,2952,1134,2012,7664,4169,  # 5702
1238,2578,3086,1259,7665, 700,7666,2953,3143,3668,4170,7667,4171,1146,1875,1906,  # 5718
4444,2601,3967, 781,2422, 132,1589, 203, 147, 273,2789,2402, 898,1786,2154,3968,  # 5734
3969,7668,3792,2790,7669,7670,4445,4446,7671,3208,7672,1635,3793, 965,7673,1804,  # 5750
2690,1516,3559,1121,1082,1329,3284,3970,1449,3794,  65,1128,2835,2913,2759,1590,  # 5766
3795,7674,7675,  12,2658,  45, 976,2579,3144,4447, 517,2528,1013,1037,3209,7676,  # 5782
3796,2836,7677,3797,7678,3452,7679,2602, 614,1998,2318,3798,3087,2724,2628,7680,  # 5798
2580,4172, 599,1269,7681,1810,3669,7682,2691,3088, 759,1060, 489,1805,3351,3285,  # 5814
1358,7683,7684,2386,1387,1215,2629,2252, 490,7685,7686,4173,1759,2387,2343,7687,  # 5830
4448,3799,1907,3971,2630,1806,3210,4449,3453,3286,2760,2344, 874,7688,7689,3454,  # 5846
3670,1858,  91,2914,3671,3042,3800,4450,7690,3145,3972,2659,7691,3455,1202,1403,  # 5862
3801,2954,2529,1517,2503,4451,3456,2504,7692,4452,7693,2692,1885,1495,1731,3973,  # 5878
2365,4453,7694,2029,7695,7696,3974,2693,1216, 237,2581,4174,2319,3975,3802,4454,  # 5894
4455,2694,3560,3457, 445,4456,7697,7698,7699,7700,2761,  61,3976,3672,1822,3977,  # 5910
7701, 687,2045, 935, 925, 405,2660, 703,1096,1859,2725,4457,3978,1876,1367,2695,  # 5926
3352, 918,2105,1781,2476, 334,3287,1611,1093,4458, 564,3146,3458,3673,3353, 945,  # 5942
2631,2057,4459,7702,1925, 872,4175,7703,3459,2696,3089, 349,4176,3674,3979,4460,  # 5958
3803,4177,3675,2155,3980,4461,4462,4178,4463,2403,2046, 782,3981, 400, 251,4179,  # 5974
1624,7704,7705, 277,3676, 299,1265, 476,1191,3804,2121,4180,4181,1109, 205,7706,  # 5990
2582,1000,2156,3561,1860,7707,7708,7709,4464,7710,4465,2565, 107,2477,2157,3982,  # 6006
3460,3147,7711,1533, 541,1301, 158, 753,4182,2872,3562,7712,1696, 370,1088,4183,  # 6022
4466,3563, 579, 327, 440, 162,2240, 269,1937,1374,3461, 968,3043,  56,1396,3090,  # 6038
2106,3288,3354,7713,1926,2158,4467,2998,7714,3564,7715,7716,3677,4468,2478,7717,  # 6054
2791,7718,1650,4469,7719,2603,7720,7721,3983,2661,3355,1149,3356,3984,3805,3985,  # 6070
7722,1076,  49,7723, 951,3211,3289,3290, 450,2837, 920,7724,1811,2792,2366,4184,  # 6086
1908,1138,2367,3806,3462,7725,3212,4470,1909,1147,1518,2423,4471,3807,7726,4472,  # 6102
2388,2604, 260,1795,3213,7727,7728,3808,3291, 708,7729,3565,1704,7730,3566,1351,  # 6118
1618,3357,2999,1886, 944,4185,3358,4186,3044,3359,4187,7731,3678, 422, 413,1714,  # 6134
3292, 500,2058,2345,4188,2479,7732,1344,1910, 954,7733,1668,7734,7735,3986,2404,  # 6150
4189,3567,3809,4190,7736,2302,1318,2505,3091, 133,3092,2873,4473, 629,  31,2838,  # 6166
2697,3810,4474, 850, 949,4475,3987,2955,1732,2088,4191,1496,1852,7737,3988, 620,  # 6182
3214, 981,1242,3679,3360,1619,3680,1643,3293,2139,2452,1970,1719,3463,2168,7738,  # 6198
3215,7739,7740,3361,1828,7741,1277,4476,1565,2047,7742,1636,3568,3093,7743, 869,  # 6214
2839, 655,3811,3812,3094,3989,3000,3813,1310,3569,4477,7744,7745,7746,1733, 558,  # 6230
4478,3681, 335,1549,3045,1756,4192,3682,1945,3464,1829,1291,1192, 470,2726,2107,  # 6246
2793, 913,1054,3990,7747,1027,7748,3046,3991,4479, 982,2662,3362,3148,3465,3216,  # 6262
3217,1946,2794,7749, 571,4480,7750,1830,7751,3570,2583,1523,2424,7752,2089, 984,  # 6278
4481,3683,1959,7753,3684, 852, 923,2795,3466,3685, 969,1519, 999,2048,2320,1705,  # 6294
7754,3095, 615,1662, 151, 597,3992,2405,2321,1049, 275,4482,3686,4193, 568,3687,  # 6310
3571,2480,4194,3688,7755,2425,2270, 409,3218,7756,1566,2874,3467,1002, 769,2840,  # 6326
 194,2090,3149,3689,2222,3294,4195, 628,1505,7757,7758,1763,2177,3001,3993, 521,  # 6342
1161,2584,1787,2203,2406,4483,3994,1625,4196,4197, 412,  42,3096, 464,7759,2632,  # 6358
4484,3363,1760,1571,2875,3468,2530,1219,2204,3814,2633,2140,2368,4485,4486,3295,  # 6374
1651,3364,3572,7760,7761,3573,2481,3469,7762,3690,7763,7764,2271,2091, 460,7765,  # 6390
4487,7766,3002, 962, 588,3574, 289,3219,2634,1116,  52,7767,3047,1796,7768,7769,  # 6406
7770,1467,7771,1598,1143,3691,4198,1984,1734,1067,4488,1280,3365, 465,4489,1572,  # 6422
 510,7772,1927,2241,1812,1644,3575,7773,4490,3692,7774,7775,2663,1573,1534,7776,  # 6438
7777,4199, 536,1807,1761,3470,3815,3150,2635,7778,7779,7780,4491,3471,2915,1911,  # 6454
2796,7781,3296,1122, 377,3220,7782, 360,7783,7784,4200,1529, 551,7785,2059,3693,  # 6470
1769,2426,7786,2916,4201,3297,3097,2322,2108,2030,4492,1404, 136,1468,1479, 672,  # 6486
1171,3221,2303, 271,3151,7787,2762,7788,2049, 678,2727, 865,1947,4493,7789,2013,  # 6502
3995,2956,7790,2728,2223,1397,3048,3694,4494,4495,1735,2917,3366,3576,7791,3816,  # 6518
 509,2841,2453,2876,3817,7792,7793,3152,3153,4496,4202,2531,4497,2304,1166,1010,  # 6534
 552, 681,1887,7794,7795,2957,2958,3996,1287,1596,1861,3154, 358, 453, 736, 175,  # 6550
 478,1117, 905,1167,1097,7796,1853,1530,7797,1706,7798,2178,3472,2287,3695,3473,  # 6566
3577,4203,2092,4204,7799,3367,1193,2482,4205,1458,2190,2205,1862,1888,1421,3298,  # 6582
2918,3049,2179,3474, 595,2122,7800,3997,7801,7802,4206,1707,2636, 223,3696,1359,  # 6598
 751,3098, 183,3475,7803,2797,3003, 419,2369, 633, 704,3818,2389, 241,7804,7805,  # 6614
7806, 838,3004,3697,2272,2763,2454,3819,1938,2050,3998,1309,3099,2242,1181,7807,  # 6630
1136,2206,3820,2370,1446,4207,2305,4498,7808,7809,4208,1055,2605, 484,3698,7810,  # 6646
3999, 625,4209,2273,3368,1499,4210,4000,7811,4001,4211,3222,2274,2275,3476,7812,  # 6662
7813,2764, 808,2606,3699,3369,4002,4212,3100,2532, 526,3370,3821,4213, 955,7814,  # 6678
1620,4214,2637,2427,7815,1429,3700,1669,1831, 994, 928,7816,3578,1260,7817,7818,  # 6694
7819,1948,2288, 741,2919,1626,4215,2729,2455, 867,1184, 362,3371,1392,7820,7821,  # 6710
4003,4216,1770,1736,3223,2920,4499,4500,1928,2698,1459,1158,7822,3050,3372,2877,  # 6726
1292,1929,2506,2842,3701,1985,1187,2071,2014,2607,4217,7823,2566,2507,2169,3702,  # 6742
2483,3299,7824,3703,4501,7825,7826, 666,1003,3005,1022,3579,4218,7827,4502,1813,  # 6758
2253, 574,3822,1603, 295,1535, 705,3823,4219, 283, 858, 417,7828,7829,3224,4503,  # 6774
4504,3051,1220,1889,1046,2276,2456,4004,1393,1599, 689,2567, 388,4220,7830,2484,  # 6790
 802,7831,2798,3824,2060,1405,2254,7832,4505,3825,2109,1052,1345,3225,1585,7833,  # 6806
 809,7834,7835,7836, 575,2730,3477, 956,1552,1469,1144,2323,7837,2324,1560,2457,  # 6822
3580,3226,4005, 616,2207,3155,2180,2289,7838,1832,7839,3478,4506,7840,1319,3704,  # 6838
3705,1211,3581,1023,3227,1293,2799,7841,7842,7843,3826, 607,2306,3827, 762,2878,  # 6854
1439,4221,1360,7844,1485,3052,7845,4507,1038,4222,1450,2061,2638,4223,1379,4508,  # 6870
2585,7846,7847,4224,1352,1414,2325,2921,1172,7848,7849,3828,3829,7850,1797,1451,  # 6886
7851,7852,7853,7854,2922,4006,4007,2485,2346, 411,4008,4009,3582,3300,3101,4509,  # 6902
1561,2664,1452,4010,1375,7855,7856,  47,2959, 316,7857,1406,1591,2923,3156,7858,  # 6918
1025,2141,3102,3157, 354,2731, 884,2224,4225,2407, 508,3706, 726,3583, 996,2428,  # 6934
3584, 729,7859, 392,2191,1453,4011,4510,3707,7860,7861,2458,3585,2608,1675,2800,  # 6950
 919,2347,2960,2348,1270,4511,4012,  73,7862,7863, 647,7864,3228,2843,2255,1550,  # 6966
1346,3006,7865,1332, 883,3479,7866,7867,7868,7869,3301,2765,7870,1212, 831,1347,  # 6982
4226,4512,2326,3830,1863,3053, 720,3831,4513,4514,3832,7871,4227,7872,7873,4515,  # 6998
7874,7875,1798,4516,3708,2609,4517,3586,1645,2371,7876,7877,2924, 669,2208,2665,  # 7014
2429,7878,2879,7879,7880,1028,3229,7881,4228,2408,7882,2256,1353,7883,7884,4518,  # 7030
3158, 518,7885,4013,7886,4229,1960,7887,2142,4230,7888,7889,3007,2349,2350,3833,  # 7046
 516,1833,1454,4014,2699,4231,4519,2225,2610,1971,1129,3587,7890,2766,7891,2961,  # 7062
1422, 577,1470,3008,1524,3373,7892,7893, 432,4232,3054,3480,7894,2586,1455,2508,  # 7078
2226,1972,1175,7895,1020,2732,4015,3481,4520,7896,2733,7897,1743,1361,3055,3482,  # 7094
2639,4016,4233,4521,2290, 895, 924,4234,2170, 331,2243,3056, 166,1627,3057,1098,  # 7110
7898,1232,2880,2227,3374,4522, 657, 403,1196,2372, 542,3709,3375,1600,4235,3483,  # 7126
7899,4523,2767,3230, 576, 530,1362,7900,4524,2533,2666,3710,4017,7901, 842,3834,  # 7142
7902,2801,2031,1014,4018, 213,2700,3376, 665, 621,4236,7903,3711,2925,2430,7904,  # 7158
2431,3302,3588,3377,7905,4237,2534,4238,4525,3589,1682,4239,3484,1380,7906, 724,  # 7174
2277, 600,1670,7907,1337,1233,4526,3103,2244,7908,1621,4527,7909, 651,4240,7910,  # 7190
1612,4241,2611,7911,2844,7912,2734,2307,3058,7913, 716,2459,3059, 174,1255,2701,  # 7206
4019,3590, 548,1320,1398, 728,4020,1574,7914,1890,1197,3060,4021,7915,3061,3062,  # 7222
3712,3591,3713, 747,7916, 635,4242,4528,7917,7918,7919,4243,7920,7921,4529,7922,  # 7238
3378,4530,2432, 451,7923,3714,2535,2072,4244,2735,4245,4022,7924,1764,4531,7925,  # 7254
4246, 350,7926,2278,2390,2486,7927,4247,4023,2245,1434,4024, 488,4532, 458,4248,  # 7270
4025,3715, 771,1330,2391,3835,2568,3159,2159,2409,1553,2667,3160,4249,7928,2487,  # 7286
2881,2612,1720,2702,4250,3379,4533,7929,2536,4251,7930,3231,4252,2768,7931,2015,  # 7302
2736,7932,1155,1017,3716,3836,7933,3303,2308, 201,1864,4253,1430,7934,4026,7935,  # 7318
7936,7937,7938,7939,4254,1604,7940, 414,1865, 371,2587,4534,4535,3485,2016,3104,  # 7334
4536,1708, 960,4255, 887, 389,2171,1536,1663,1721,7941,2228,4027,2351,2926,1580,  # 7350
7942,7943,7944,1744,7945,2537,4537,4538,7946,4539,7947,2073,7948,7949,3592,3380,  # 7366
2882,4256,7950,4257,2640,3381,2802, 673,2703,2460, 709,3486,4028,3593,4258,7951,  # 7382
1148, 502, 634,7952,7953,1204,4540,3594,1575,4541,2613,3717,7954,3718,3105, 948,  # 7398
3232, 121,1745,3837,1110,7955,4259,3063,2509,3009,4029,3719,1151,1771,3838,1488,  # 7414
4030,1986,7956,2433,3487,7957,7958,2093,7959,4260,3839,1213,1407,2803, 531,2737,  # 7430
2538,3233,1011,1537,7960,2769,4261,3106,1061,7961,3720,3721,1866,2883,7962,2017,  # 7446
 120,4262,4263,2062,3595,3234,2309,3840,2668,3382,1954,4542,7963,7964,3488,1047,  # 7462
2704,1266,7965,1368,4543,2845, 649,3383,3841,2539,2738,1102,2846,2669,7966,7967,  # 7478
1999,7968,1111,3596,2962,7969,2488,3842,3597,2804,1854,3384,3722,7970,7971,3385,  # 7494
2410,2884,3304,3235,3598,7972,2569,7973,3599,2805,4031,1460, 856,7974,3600,7975,  # 7510
2885,2963,7976,2886,3843,7977,4264, 632,2510, 875,3844,1697,3845,2291,7978,7979,  # 7526
4544,3010,1239, 580,4545,4265,7980, 914, 936,2074,1190,4032,1039,2123,7981,7982,  # 7542
7983,3386,1473,7984,1354,4266,3846,7985,2172,3064,4033, 915,3305,4267,4268,3306,  # 7558
1605,1834,7986,2739, 398,3601,4269,3847,4034, 328,1912,2847,4035,3848,1331,4270,  # 7574
3011, 937,4271,7987,3602,4036,4037,3387,2160,4546,3388, 524, 742, 538,3065,1012,  # 7590
7988,7989,3849,2461,7990, 658,1103, 225,3850,7991,7992,4547,7993,4548,7994,3236,  # 7606
1243,7995,4038, 963,2246,4549,7996,2705,3603,3161,7997,7998,2588,2327,7999,4550,  # 7622
8000,8001,8002,3489,3307, 957,3389,2540,2032,1930,2927,2462, 870,2018,3604,1746,  # 7638
2770,2771,2434,2463,8003,3851,8004,3723,3107,3724,3490,3390,3725,8005,1179,3066,  # 7654
8006,3162,2373,4272,3726,2541,3163,3108,2740,4039,8007,3391,1556,2542,2292, 977,  # 7670
2887,2033,4040,1205,3392,8008,1765,3393,3164,2124,1271,1689, 714,4551,3491,8009,  # 7686
2328,3852, 533,4273,3605,2181, 617,8010,2464,3308,3492,2310,8011,8012,3165,8013,  # 7702
8014,3853,1987, 618, 427,2641,3493,3394,8015,8016,1244,1690,8017,2806,4274,4552,  # 7718
8018,3494,8019,8020,2279,1576, 473,3606,4275,3395, 972,8021,3607,8022,3067,8023,  # 7734
8024,4553,4554,8025,3727,4041,4042,8026, 153,4555, 356,8027,1891,2888,4276,2143,  # 7750
 408, 803,2352,8028,3854,8029,4277,1646,2570,2511,4556,4557,3855,8030,3856,4278,  # 7766
8031,2411,3396, 752,8032,8033,1961,2964,8034, 746,3012,2465,8035,4279,3728, 698,  # 7782
4558,1892,4280,3608,2543,4559,3609,3857,8036,3166,3397,8037,1823,1302,4043,2706,  # 7798
3858,1973,4281,8038,4282,3167, 823,1303,1288,1236,2848,3495,4044,3398, 774,3859,  # 7814
8039,1581,4560,1304,2849,3860,4561,8040,2435,2161,1083,3237,4283,4045,4284, 344,  # 7830
1173, 288,2311, 454,1683,8041,8042,1461,4562,4046,2589,8043,8044,4563, 985, 894,  # 7846
8045,3399,3168,8046,1913,2928,3729,1988,8047,2110,1974,8048,4047,8049,2571,1194,  # 7862
 425,8050,4564,3169,1245,3730,4285,8051,8052,2850,8053, 636,4565,1855,3861, 760,  # 7878
1799,8054,4286,2209,1508,4566,4048,1893,1684,2293,8055,8056,8057,4287,4288,2210,  # 7894
 479,8058,8059, 832,8060,4049,2489,8061,2965,2490,3731, 990,3109, 627,1814,2642,  # 7910
4289,1582,4290,2125,2111,3496,4567,8062, 799,4291,3170,8063,4568,2112,1737,3013,  # 7926
1018, 543, 754,4292,3309,1676,4569,4570,4050,8064,1489,8065,3497,8066,2614,2889,  # 7942
4051,8067,8068,2966,8069,8070,8071,8072,3171,4571,4572,2182,1722,8073,3238,3239,  # 7958
1842,3610,1715, 481, 365,1975,1856,8074,8075,1962,2491,4573,8076,2126,3611,3240,  # 7974
 433,1894,2063,2075,8077, 602,2741,8078,8079,8080,8081,8082,3014,1628,3400,8083,  # 7990
3172,4574,4052,2890,4575,2512,8084,2544,2772,8085,8086,8087,3310,4576,2891,8088,  # 8006
4577,8089,2851,4578,4579,1221,2967,4053,2513,8090,8091,8092,1867,1989,8093,8094,  # 8022
8095,1895,8096,8097,4580,1896,4054, 318,8098,2094,4055,4293,8099,8100, 485,8101,  # 8038
 938,3862, 553,2670, 116,8102,3863,3612,8103,3498,2671,2773,3401,3311,2807,8104,  # 8054
3613,2929,4056,1747,2930,2968,8105,8106, 207,8107,8108,2672,4581,2514,8109,3015,  # 8070
 890,3614,3864,8110,1877,3732,3402,8111,2183,2353,3403,1652,8112,8113,8114, 941,  # 8086
2294, 208,3499,4057,2019, 330,4294,3865,2892,2492,3733,4295,8115,8116,8117,8118,  # 8102
)

site-packages/pip/_vendor/chardet/charsetprober.py000064400000011766150564745010016361 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

import logging
import re

from .enums import ProbingState


class CharSetProber(object):

    SHORTCUT_THRESHOLD = 0.95

    def __init__(self, lang_filter=None):
        self._state = None
        self.lang_filter = lang_filter
        self.logger = logging.getLogger(__name__)

    def reset(self):
        self._state = ProbingState.DETECTING

    @property
    def charset_name(self):
        return None

    def feed(self, buf):
        pass

    @property
    def state(self):
        return self._state

    def get_confidence(self):
        return 0.0

    @staticmethod
    def filter_high_byte_only(buf):
        buf = re.sub(b'([\x00-\x7F])+', b' ', buf)
        return buf

    @staticmethod
    def filter_international_words(buf):
        """
        We define three types of bytes:
        alphabet: english alphabets [a-zA-Z]
        international: international characters [\x80-\xFF]
        marker: everything else [^a-zA-Z\x80-\xFF]

        The input buffer can be thought to contain a series of words delimited
        by markers. This function works to filter all words that contain at
        least one international character. All contiguous sequences of markers
        are replaced by a single space ascii character.

        This filter applies to all scripts which do not use English characters.
        """
        filtered = bytearray()

        # This regex expression filters out only words that have at-least one
        # international character. The word may include one marker character at
        # the end.
        words = re.findall(b'[a-zA-Z]*[\x80-\xFF]+[a-zA-Z]*[^a-zA-Z\x80-\xFF]?',
                           buf)

        for word in words:
            filtered.extend(word[:-1])

            # If the last character in the word is a marker, replace it with a
            # space as markers shouldn't affect our analysis (they are used
            # similarly across all languages and may thus have similar
            # frequencies).
            last_char = word[-1:]
            if not last_char.isalpha() and last_char < b'\x80':
                last_char = b' '
            filtered.extend(last_char)

        return filtered

    @staticmethod
    def filter_with_english_letters(buf):
        """
        Returns a copy of ``buf`` that retains only the sequences of English
        alphabet and high byte characters that are not between <> characters.
        Also retains English alphabet and high byte characters immediately
        before occurrences of >.

        This filter can be applied to all scripts which contain both English
        characters and extended ASCII characters, but is currently only used by
        ``Latin1Prober``.
        """
        filtered = bytearray()
        in_tag = False
        prev = 0

        for curr in range(len(buf)):
            # Slice here to get bytes instead of an int with Python 3
            buf_char = buf[curr:curr + 1]
            # Check if we're coming out of or entering an HTML tag
            if buf_char == b'>':
                in_tag = False
            elif buf_char == b'<':
                in_tag = True

            # If current character is not extended-ASCII and not alphabetic...
            if buf_char < b'\x80' and not buf_char.isalpha():
                # ...and we're not in a tag
                if curr > prev and not in_tag:
                    # Keep everything after last non-extended-ASCII,
                    # non-alphabetic character
                    filtered.extend(buf[prev:curr])
                    # Output a space to delimit stretch we kept
                    filtered.extend(b' ')
                prev = curr + 1

        # If we're not in a tag...
        if not in_tag:
            # Keep everything after last non-extended-ASCII, non-alphabetic
            # character
            filtered.extend(buf[prev:])

        return filtered
site-packages/pip/_vendor/chardet/langhebrewmodel.py000064400000300702150564745010016644 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

HEBREW_LANG_MODEL = {
    50: {  # 'a'
        50: 0,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 2,  # 'l'
        54: 2,  # 'n'
        49: 0,  # 'o'
        51: 2,  # 'r'
        43: 1,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 1,  # 'ק'
        7: 0,  # 'ר'
        10: 1,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    60: {  # 'c'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 0,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 0,  # 'n'
        49: 1,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    61: {  # 'd'
        50: 1,  # 'a'
        60: 0,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 2,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 0,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 1,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    42: {  # 'e'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 2,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 2,  # 'l'
        54: 2,  # 'n'
        49: 1,  # 'o'
        51: 2,  # 'r'
        43: 2,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 1,  # '–'
        52: 2,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    53: {  # 'i'
        50: 1,  # 'a'
        60: 2,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 0,  # 'i'
        56: 1,  # 'l'
        54: 2,  # 'n'
        49: 2,  # 'o'
        51: 1,  # 'r'
        43: 2,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    56: {  # 'l'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 2,  # 'e'
        53: 2,  # 'i'
        56: 2,  # 'l'
        54: 1,  # 'n'
        49: 1,  # 'o'
        51: 0,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    54: {  # 'n'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 1,  # 'o'
        51: 0,  # 'r'
        43: 1,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 2,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    49: {  # 'o'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 2,  # 'n'
        49: 1,  # 'o'
        51: 2,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    51: {  # 'r'
        50: 2,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 2,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 2,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 2,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    43: {  # 's'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 0,  # 'd'
        42: 2,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 1,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 2,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    44: {  # 't'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 0,  # 'd'
        42: 2,  # 'e'
        53: 2,  # 'i'
        56: 1,  # 'l'
        54: 0,  # 'n'
        49: 1,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 1,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 2,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    63: {  # 'u'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 0,  # 'o'
        51: 1,  # 'r'
        43: 2,  # 's'
        44: 1,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    34: {  # '\xa0'
        50: 1,  # 'a'
        60: 0,  # 'c'
        61: 1,  # 'd'
        42: 0,  # 'e'
        53: 1,  # 'i'
        56: 0,  # 'l'
        54: 1,  # 'n'
        49: 1,  # 'o'
        51: 0,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 0,  # 'u'
        34: 2,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 1,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 2,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 2,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 1,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    55: {  # '´'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 1,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 2,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 1,  # 'ן'
        12: 1,  # 'נ'
        19: 1,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    48: {  # '¼'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    39: {  # '½'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    57: {  # '¾'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    30: {  # 'ְ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 2,  # 'ג'
        16: 2,  # 'ד'
        3: 2,  # 'ה'
        2: 2,  # 'ו'
        24: 2,  # 'ז'
        14: 2,  # 'ח'
        22: 2,  # 'ט'
        1: 2,  # 'י'
        25: 2,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 1,  # 'ם'
        6: 2,  # 'מ'
        23: 0,  # 'ן'
        12: 2,  # 'נ'
        19: 2,  # 'ס'
        13: 2,  # 'ע'
        26: 0,  # 'ף'
        18: 2,  # 'פ'
        27: 0,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    59: {  # 'ֱ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 1,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 2,  # 'ל'
        11: 0,  # 'ם'
        6: 2,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    41: {  # 'ֲ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 2,  # 'ב'
        20: 1,  # 'ג'
        16: 2,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 1,  # 'י'
        25: 1,  # 'ך'
        15: 1,  # 'כ'
        4: 2,  # 'ל'
        11: 0,  # 'ם'
        6: 2,  # 'מ'
        23: 0,  # 'ן'
        12: 2,  # 'נ'
        19: 1,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 2,  # 'צ'
        17: 1,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    33: {  # 'ִ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 1,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 1,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 1,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 2,  # 'ב'
        20: 2,  # 'ג'
        16: 2,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 2,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 2,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 2,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    37: {  # 'ֵ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 1,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 1,  # 'ג'
        16: 2,  # 'ד'
        3: 2,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 2,  # 'ח'
        22: 1,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 1,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 1,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 1,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 1,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    36: {  # 'ֶ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 1,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 1,  # 'ג'
        16: 2,  # 'ד'
        3: 2,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 2,  # 'ח'
        22: 1,  # 'ט'
        1: 2,  # 'י'
        25: 2,  # 'ך'
        15: 1,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 2,  # 'ס'
        13: 1,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 2,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    31: {  # 'ַ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 2,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 2,  # 'ג'
        16: 2,  # 'ד'
        3: 2,  # 'ה'
        2: 1,  # 'ו'
        24: 2,  # 'ז'
        14: 2,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 2,  # 'ס'
        13: 2,  # 'ע'
        26: 2,  # 'ף'
        18: 2,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    29: {  # 'ָ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 2,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 1,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 2,  # 'ג'
        16: 2,  # 'ד'
        3: 3,  # 'ה'
        2: 2,  # 'ו'
        24: 2,  # 'ז'
        14: 2,  # 'ח'
        22: 1,  # 'ט'
        1: 2,  # 'י'
        25: 2,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 1,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 2,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    35: {  # 'ֹ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 1,  # 'ג'
        16: 2,  # 'ד'
        3: 2,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 1,  # 'י'
        25: 1,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 2,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 2,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    62: {  # 'ֻ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 1,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 2,  # 'ל'
        11: 1,  # 'ם'
        6: 1,  # 'מ'
        23: 1,  # 'ן'
        12: 1,  # 'נ'
        19: 1,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    28: {  # 'ּ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 3,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 1,  # 'ֲ'
        33: 3,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 3,  # 'ַ'
        29: 3,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 2,  # 'ׁ'
        45: 1,  # 'ׂ'
        9: 2,  # 'א'
        8: 2,  # 'ב'
        20: 1,  # 'ג'
        16: 2,  # 'ד'
        3: 1,  # 'ה'
        2: 2,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 2,  # 'י'
        25: 2,  # 'ך'
        15: 2,  # 'כ'
        4: 2,  # 'ל'
        11: 1,  # 'ם'
        6: 2,  # 'מ'
        23: 1,  # 'ן'
        12: 2,  # 'נ'
        19: 1,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 1,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 2,  # 'ר'
        10: 2,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    38: {  # 'ׁ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 2,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    45: {  # 'ׂ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 1,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 2,  # 'ו'
        24: 0,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 1,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 0,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 0,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    9: {  # 'א'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 2,  # 'ֱ'
        41: 2,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 2,  # 'ע'
        26: 3,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    8: {  # 'ב'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 1,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 3,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 1,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    20: {  # 'ג'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 2,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 1,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 3,  # 'ב'
        20: 2,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 2,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 1,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 2,  # 'פ'
        27: 1,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    16: {  # 'ד'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 1,  # 'ז'
        14: 2,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 2,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 0,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    3: {  # 'ה'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 1,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 1,  # 'ֱ'
        41: 2,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 3,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 0,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    2: {  # 'ו'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 1,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 3,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 3,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 3,  # 'ף'
        18: 3,  # 'פ'
        27: 3,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    24: {  # 'ז'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 1,  # 'ֲ'
        33: 1,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 2,  # 'ב'
        20: 2,  # 'ג'
        16: 2,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 2,  # 'ח'
        22: 1,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 2,  # 'נ'
        19: 1,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 1,  # 'ש'
        5: 2,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    14: {  # 'ח'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 1,  # 'ֱ'
        41: 2,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 3,  # 'ב'
        20: 2,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 2,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 2,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 1,  # 'ע'
        26: 2,  # 'ף'
        18: 2,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    22: {  # 'ט'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 1,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 1,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 1,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 3,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 2,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 3,  # 'ר'
        10: 2,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    1: {  # 'י'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 3,  # 'ף'
        18: 3,  # 'פ'
        27: 3,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    25: {  # 'ך'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 2,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 1,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 1,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    15: {  # 'כ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 3,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 2,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 2,  # 'ע'
        26: 3,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 2,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    4: {  # 'ל'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 3,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    11: {  # 'ם'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 1,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 0,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 1,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    6: {  # 'מ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 0,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    23: {  # 'ן'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 1,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 1,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 1,  # 'ס'
        13: 1,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 1,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    12: {  # 'נ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    19: {  # 'ס'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 1,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 2,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 1,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 3,  # 'ף'
        18: 3,  # 'פ'
        27: 0,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 1,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    13: {  # 'ע'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 1,  # 'ֱ'
        41: 2,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 1,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 2,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 2,  # 'ע'
        26: 1,  # 'ף'
        18: 2,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    26: {  # 'ף'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 1,  # 'ס'
        13: 0,  # 'ע'
        26: 1,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    18: {  # 'פ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 2,  # 'ב'
        20: 3,  # 'ג'
        16: 2,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 2,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 2,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    27: {  # 'ץ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 1,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 0,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    21: {  # 'צ'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 2,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 1,  # 'ז'
        14: 3,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 1,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 1,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 0,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    17: {  # 'ק'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 1,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 2,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 1,  # 'ך'
        15: 1,  # 'כ'
        4: 3,  # 'ל'
        11: 2,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 2,  # 'ץ'
        21: 3,  # 'צ'
        17: 2,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    7: {  # 'ר'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 2,  # '´'
        48: 1,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 1,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 2,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 3,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 3,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 3,  # 'ץ'
        21: 3,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    10: {  # 'ש'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 1,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 1,  # 'ִ'
        37: 1,  # 'ֵ'
        36: 1,  # 'ֶ'
        31: 1,  # 'ַ'
        29: 1,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 3,  # 'ׁ'
        45: 2,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 3,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 3,  # 'ח'
        22: 3,  # 'ט'
        1: 3,  # 'י'
        25: 3,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 2,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 1,  # '…'
    },
    5: {  # 'ת'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 1,  # '\xa0'
        55: 0,  # '´'
        48: 1,  # '¼'
        39: 1,  # '½'
        57: 0,  # '¾'
        30: 2,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 2,  # 'ִ'
        37: 2,  # 'ֵ'
        36: 2,  # 'ֶ'
        31: 2,  # 'ַ'
        29: 2,  # 'ָ'
        35: 1,  # 'ֹ'
        62: 1,  # 'ֻ'
        28: 2,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 3,  # 'א'
        8: 3,  # 'ב'
        20: 3,  # 'ג'
        16: 2,  # 'ד'
        3: 3,  # 'ה'
        2: 3,  # 'ו'
        24: 2,  # 'ז'
        14: 3,  # 'ח'
        22: 2,  # 'ט'
        1: 3,  # 'י'
        25: 2,  # 'ך'
        15: 3,  # 'כ'
        4: 3,  # 'ל'
        11: 3,  # 'ם'
        6: 3,  # 'מ'
        23: 3,  # 'ן'
        12: 3,  # 'נ'
        19: 2,  # 'ס'
        13: 3,  # 'ע'
        26: 2,  # 'ף'
        18: 3,  # 'פ'
        27: 1,  # 'ץ'
        21: 2,  # 'צ'
        17: 3,  # 'ק'
        7: 3,  # 'ר'
        10: 3,  # 'ש'
        5: 3,  # 'ת'
        32: 1,  # '–'
        52: 1,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
    32: {  # '–'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 1,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 1,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 1,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 1,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    52: {  # '’'
        50: 1,  # 'a'
        60: 0,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 1,  # 'r'
        43: 2,  # 's'
        44: 2,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 1,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    47: {  # '“'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 1,  # 'l'
        54: 1,  # 'n'
        49: 1,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 1,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 2,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 1,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 1,  # 'ח'
        22: 1,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 1,  # 'ס'
        13: 1,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 1,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    46: {  # '”'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 1,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 1,  # 'ב'
        20: 1,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 1,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 0,  # '†'
        40: 0,  # '…'
    },
    58: {  # '†'
        50: 0,  # 'a'
        60: 0,  # 'c'
        61: 0,  # 'd'
        42: 0,  # 'e'
        53: 0,  # 'i'
        56: 0,  # 'l'
        54: 0,  # 'n'
        49: 0,  # 'o'
        51: 0,  # 'r'
        43: 0,  # 's'
        44: 0,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 0,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 0,  # 'ה'
        2: 0,  # 'ו'
        24: 0,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 0,  # 'י'
        25: 0,  # 'ך'
        15: 0,  # 'כ'
        4: 0,  # 'ל'
        11: 0,  # 'ם'
        6: 0,  # 'מ'
        23: 0,  # 'ן'
        12: 0,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 0,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 0,  # 'ר'
        10: 0,  # 'ש'
        5: 0,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 0,  # '”'
        58: 2,  # '†'
        40: 0,  # '…'
    },
    40: {  # '…'
        50: 1,  # 'a'
        60: 1,  # 'c'
        61: 1,  # 'd'
        42: 1,  # 'e'
        53: 1,  # 'i'
        56: 0,  # 'l'
        54: 1,  # 'n'
        49: 0,  # 'o'
        51: 1,  # 'r'
        43: 1,  # 's'
        44: 1,  # 't'
        63: 0,  # 'u'
        34: 0,  # '\xa0'
        55: 0,  # '´'
        48: 0,  # '¼'
        39: 0,  # '½'
        57: 0,  # '¾'
        30: 0,  # 'ְ'
        59: 0,  # 'ֱ'
        41: 0,  # 'ֲ'
        33: 0,  # 'ִ'
        37: 0,  # 'ֵ'
        36: 0,  # 'ֶ'
        31: 0,  # 'ַ'
        29: 0,  # 'ָ'
        35: 0,  # 'ֹ'
        62: 0,  # 'ֻ'
        28: 0,  # 'ּ'
        38: 0,  # 'ׁ'
        45: 0,  # 'ׂ'
        9: 1,  # 'א'
        8: 0,  # 'ב'
        20: 0,  # 'ג'
        16: 0,  # 'ד'
        3: 1,  # 'ה'
        2: 1,  # 'ו'
        24: 1,  # 'ז'
        14: 0,  # 'ח'
        22: 0,  # 'ט'
        1: 1,  # 'י'
        25: 0,  # 'ך'
        15: 1,  # 'כ'
        4: 1,  # 'ל'
        11: 0,  # 'ם'
        6: 1,  # 'מ'
        23: 0,  # 'ן'
        12: 1,  # 'נ'
        19: 0,  # 'ס'
        13: 0,  # 'ע'
        26: 0,  # 'ף'
        18: 1,  # 'פ'
        27: 0,  # 'ץ'
        21: 0,  # 'צ'
        17: 0,  # 'ק'
        7: 1,  # 'ר'
        10: 1,  # 'ש'
        5: 1,  # 'ת'
        32: 0,  # '–'
        52: 0,  # '’'
        47: 0,  # '“'
        46: 1,  # '”'
        58: 0,  # '†'
        40: 2,  # '…'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
WINDOWS_1255_HEBREW_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 69,  # 'A'
     66: 91,  # 'B'
     67: 79,  # 'C'
     68: 80,  # 'D'
     69: 92,  # 'E'
     70: 89,  # 'F'
     71: 97,  # 'G'
     72: 90,  # 'H'
     73: 68,  # 'I'
     74: 111,  # 'J'
     75: 112,  # 'K'
     76: 82,  # 'L'
     77: 73,  # 'M'
     78: 95,  # 'N'
     79: 85,  # 'O'
     80: 78,  # 'P'
     81: 121,  # 'Q'
     82: 86,  # 'R'
     83: 71,  # 'S'
     84: 67,  # 'T'
     85: 102,  # 'U'
     86: 107,  # 'V'
     87: 84,  # 'W'
     88: 114,  # 'X'
     89: 103,  # 'Y'
     90: 115,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 50,  # 'a'
     98: 74,  # 'b'
     99: 60,  # 'c'
     100: 61,  # 'd'
     101: 42,  # 'e'
     102: 76,  # 'f'
     103: 70,  # 'g'
     104: 64,  # 'h'
     105: 53,  # 'i'
     106: 105,  # 'j'
     107: 93,  # 'k'
     108: 56,  # 'l'
     109: 65,  # 'm'
     110: 54,  # 'n'
     111: 49,  # 'o'
     112: 66,  # 'p'
     113: 110,  # 'q'
     114: 51,  # 'r'
     115: 43,  # 's'
     116: 44,  # 't'
     117: 63,  # 'u'
     118: 81,  # 'v'
     119: 77,  # 'w'
     120: 98,  # 'x'
     121: 75,  # 'y'
     122: 108,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 124,  # '€'
     129: 202,  # None
     130: 203,  # '‚'
     131: 204,  # 'ƒ'
     132: 205,  # '„'
     133: 40,  # '…'
     134: 58,  # '†'
     135: 206,  # '‡'
     136: 207,  # 'ˆ'
     137: 208,  # '‰'
     138: 209,  # None
     139: 210,  # '‹'
     140: 211,  # None
     141: 212,  # None
     142: 213,  # None
     143: 214,  # None
     144: 215,  # None
     145: 83,  # '‘'
     146: 52,  # '’'
     147: 47,  # '“'
     148: 46,  # '”'
     149: 72,  # '•'
     150: 32,  # '–'
     151: 94,  # '—'
     152: 216,  # '˜'
     153: 113,  # '™'
     154: 217,  # None
     155: 109,  # '›'
     156: 218,  # None
     157: 219,  # None
     158: 220,  # None
     159: 221,  # None
     160: 34,  # '\xa0'
     161: 116,  # '¡'
     162: 222,  # '¢'
     163: 118,  # '£'
     164: 100,  # '₪'
     165: 223,  # '¥'
     166: 224,  # '¦'
     167: 117,  # '§'
     168: 119,  # '¨'
     169: 104,  # '©'
     170: 125,  # '×'
     171: 225,  # '«'
     172: 226,  # '¬'
     173: 87,  # '\xad'
     174: 99,  # '®'
     175: 227,  # '¯'
     176: 106,  # '°'
     177: 122,  # '±'
     178: 123,  # '²'
     179: 228,  # '³'
     180: 55,  # '´'
     181: 229,  # 'µ'
     182: 230,  # '¶'
     183: 101,  # '·'
     184: 231,  # '¸'
     185: 232,  # '¹'
     186: 120,  # '÷'
     187: 233,  # '»'
     188: 48,  # '¼'
     189: 39,  # '½'
     190: 57,  # '¾'
     191: 234,  # '¿'
     192: 30,  # 'ְ'
     193: 59,  # 'ֱ'
     194: 41,  # 'ֲ'
     195: 88,  # 'ֳ'
     196: 33,  # 'ִ'
     197: 37,  # 'ֵ'
     198: 36,  # 'ֶ'
     199: 31,  # 'ַ'
     200: 29,  # 'ָ'
     201: 35,  # 'ֹ'
     202: 235,  # None
     203: 62,  # 'ֻ'
     204: 28,  # 'ּ'
     205: 236,  # 'ֽ'
     206: 126,  # '־'
     207: 237,  # 'ֿ'
     208: 238,  # '׀'
     209: 38,  # 'ׁ'
     210: 45,  # 'ׂ'
     211: 239,  # '׃'
     212: 240,  # 'װ'
     213: 241,  # 'ױ'
     214: 242,  # 'ײ'
     215: 243,  # '׳'
     216: 127,  # '״'
     217: 244,  # None
     218: 245,  # None
     219: 246,  # None
     220: 247,  # None
     221: 248,  # None
     222: 249,  # None
     223: 250,  # None
     224: 9,  # 'א'
     225: 8,  # 'ב'
     226: 20,  # 'ג'
     227: 16,  # 'ד'
     228: 3,  # 'ה'
     229: 2,  # 'ו'
     230: 24,  # 'ז'
     231: 14,  # 'ח'
     232: 22,  # 'ט'
     233: 1,  # 'י'
     234: 25,  # 'ך'
     235: 15,  # 'כ'
     236: 4,  # 'ל'
     237: 11,  # 'ם'
     238: 6,  # 'מ'
     239: 23,  # 'ן'
     240: 12,  # 'נ'
     241: 19,  # 'ס'
     242: 13,  # 'ע'
     243: 26,  # 'ף'
     244: 18,  # 'פ'
     245: 27,  # 'ץ'
     246: 21,  # 'צ'
     247: 17,  # 'ק'
     248: 7,  # 'ר'
     249: 10,  # 'ש'
     250: 5,  # 'ת'
     251: 251,  # None
     252: 252,  # None
     253: 128,  # '\u200e'
     254: 96,  # '\u200f'
     255: 253,  # None
}

WINDOWS_1255_HEBREW_MODEL = SingleByteCharSetModel(charset_name='windows-1255',
                                                   language='Hebrew',
                                                   char_to_order_map=WINDOWS_1255_HEBREW_CHAR_TO_ORDER,
                                                   language_model=HEBREW_LANG_MODEL,
                                                   typical_positive_ratio=0.984004,
                                                   keep_ascii_letters=False,
                                                   alphabet='אבגדהוזחטיךכלםמןנסעףפץצקרשתװױײ')

site-packages/pip/_vendor/chardet/langhungarianmodel.py000064400000310114150564745010017342 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

HUNGARIAN_LANG_MODEL = {
    28: {  # 'A'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 2,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 2,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 2,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 2,  # 'N'
        47: 1,  # 'O'
        46: 2,  # 'P'
        43: 2,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 2,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 2,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 2,  # 'n'
        8: 0,  # 'o'
        23: 2,  # 'p'
        10: 2,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 1,  # 'u'
        19: 1,  # 'v'
        62: 1,  # 'x'
        16: 0,  # 'y'
        11: 3,  # 'z'
        51: 1,  # 'Á'
        44: 0,  # 'É'
        61: 1,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    40: {  # 'B'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 0,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 1,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 3,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    54: {  # 'C'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 0,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 0,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 1,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 3,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 1,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    45: {  # 'D'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 0,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 0,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 1,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 1,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    32: {  # 'E'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 2,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 2,  # 'K'
        41: 2,  # 'L'
        34: 2,  # 'M'
        35: 2,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 1,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 3,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 2,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 1,  # 't'
        21: 2,  # 'u'
        19: 1,  # 'v'
        62: 1,  # 'x'
        16: 0,  # 'y'
        11: 3,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 0,  # 'Ú'
        63: 1,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    50: {  # 'F'
        28: 1,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 0,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 0,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 0,  # 'V'
        55: 1,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 1,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 1,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 0,  # 'Ú'
        63: 1,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    49: {  # 'G'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 2,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 2,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    38: {  # 'H'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 0,  # 'D'
        32: 1,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 1,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 1,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 1,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 0,  # 'V'
        55: 1,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 1,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 0,  # 'n'
        8: 3,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 2,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 2,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    39: {  # 'I'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 2,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 2,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 2,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 0,  # 'e'
        27: 1,  # 'f'
        12: 2,  # 'g'
        20: 1,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    53: {  # 'J'
        28: 2,  # 'A'
        40: 0,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 1,  # 'o'
        23: 0,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 2,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 0,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    36: {  # 'K'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 0,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 1,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 3,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 2,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    41: {  # 'L'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 1,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    34: {  # 'M'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 0,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 3,  # 'a'
        18: 0,  # 'b'
        26: 1,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 3,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 3,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 1,  # 'ű'
    },
    35: {  # 'N'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 2,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 2,  # 'Y'
        52: 1,  # 'Z'
        2: 3,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 2,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 1,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    47: {  # 'O'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 2,  # 'K'
        41: 2,  # 'L'
        34: 2,  # 'M'
        35: 2,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 2,  # 'k'
        6: 2,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 1,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 1,  # 's'
        3: 2,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 1,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 1,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    46: {  # 'P'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 0,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 1,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 1,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 0,  # 'Ú'
        63: 1,  # 'Ü'
        14: 3,  # 'á'
        15: 2,  # 'é'
        30: 0,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 0,  # 'ű'
    },
    43: {  # 'R'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 2,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 2,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    33: {  # 'S'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 3,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 1,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 1,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 1,  # 't'
        21: 1,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    37: {  # 'T'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 1,  # 'S'
        37: 2,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 2,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 1,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 0,  # 't'
        21: 2,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 1,  # 'z'
        51: 2,  # 'Á'
        44: 2,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    57: {  # 'U'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 1,  # 'e'
        27: 0,  # 'f'
        12: 2,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 1,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    48: {  # 'V'
        28: 2,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 0,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 2,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 2,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 2,  # 'o'
        23: 0,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 2,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 0,  # 'Ú'
        63: 1,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 0,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    55: {  # 'Y'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 1,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 2,  # 'Z'
        2: 1,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 1,  # 'd'
        1: 1,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        8: 1,  # 'o'
        23: 1,  # 'p'
        10: 0,  # 'r'
        5: 0,  # 's'
        3: 0,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 1,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    52: {  # 'Z'
        28: 2,  # 'A'
        40: 1,  # 'B'
        54: 0,  # 'C'
        45: 1,  # 'D'
        32: 2,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 2,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 2,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 2,  # 'S'
        37: 1,  # 'T'
        57: 1,  # 'U'
        48: 1,  # 'V'
        55: 1,  # 'Y'
        52: 1,  # 'Z'
        2: 1,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 1,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 1,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 2,  # 's'
        3: 0,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 2,  # 'Á'
        44: 1,  # 'É'
        61: 1,  # 'Í'
        58: 1,  # 'Ó'
        59: 1,  # 'Ö'
        60: 1,  # 'Ú'
        63: 1,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    2: {  # 'a'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 2,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 2,  # 'o'
        23: 3,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 1,  # 'x'
        16: 2,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    18: {  # 'b'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 2,  # 'k'
        6: 2,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 2,  # 's'
        3: 1,  # 't'
        21: 3,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 3,  # 'ó'
        24: 2,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    26: {  # 'c'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 1,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 1,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 1,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 1,  # 'j'
        7: 2,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 2,  # 't'
        21: 2,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 2,  # 'á'
        15: 2,  # 'é'
        30: 2,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    17: {  # 'd'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 2,  # 'k'
        6: 1,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 2,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    1: {  # 'e'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 2,  # 'e'
        27: 3,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 2,  # 'o'
        23: 3,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 2,  # 'u'
        19: 3,  # 'v'
        62: 2,  # 'x'
        16: 2,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    27: {  # 'f'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 3,  # 'o'
        23: 0,  # 'p'
        10: 3,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 2,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 0,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 3,  # 'ö'
        31: 1,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    12: {  # 'g'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 2,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 2,  # 'k'
        6: 3,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 3,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 3,  # 'ó'
        24: 2,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    20: {  # 'h'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 0,  # 'd'
        1: 3,  # 'e'
        27: 0,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 3,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 2,  # 's'
        3: 1,  # 't'
        21: 3,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 2,  # 'y'
        11: 0,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 2,  # 'ó'
        24: 2,  # 'ö'
        31: 2,  # 'ú'
        29: 1,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    9: {  # 'i'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 3,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 2,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 2,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 1,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 3,  # 'ó'
        24: 1,  # 'ö'
        31: 2,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 1,  # 'ű'
    },
    22: {  # 'j'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 1,  # 'i'
        22: 2,  # 'j'
        7: 2,  # 'k'
        6: 2,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 1,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 3,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    7: {  # 'k'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 1,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 2,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 2,  # 'ó'
        24: 3,  # 'ö'
        31: 1,  # 'ú'
        29: 3,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    6: {  # 'l'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 1,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 3,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 2,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 3,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 3,  # 'ő'
        56: 1,  # 'ű'
    },
    13: {  # 'm'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 1,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        8: 3,  # 'o'
        23: 3,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 3,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 2,  # 'ó'
        24: 2,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 2,  # 'ű'
    },
    4: {  # 'n'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 2,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 2,  # 'v'
        62: 1,  # 'x'
        16: 3,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 2,  # 'ó'
        24: 3,  # 'ö'
        31: 2,  # 'ú'
        29: 3,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    8: {  # 'o'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 1,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 2,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 2,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 1,  # 'o'
        23: 3,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 2,  # 'u'
        19: 3,  # 'v'
        62: 1,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    23: {  # 'p'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 1,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 2,  # 'k'
        6: 3,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 3,  # 'o'
        23: 3,  # 'p'
        10: 3,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 3,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 2,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    10: {  # 'r'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 1,  # 'x'
        16: 2,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 3,  # 'ú'
        29: 3,  # 'ü'
        42: 2,  # 'ő'
        56: 2,  # 'ű'
    },
    5: {  # 's'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 2,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 2,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 1,  # 'j'
        7: 3,  # 'k'
        6: 2,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 3,  # 'ú'
        29: 3,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    3: {  # 't'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 1,  # 'g'
        20: 3,  # 'h'
        9: 3,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 3,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 3,  # 'ú'
        29: 3,  # 'ü'
        42: 3,  # 'ő'
        56: 2,  # 'ű'
    },
    21: {  # 'u'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 2,  # 'b'
        26: 2,  # 'c'
        17: 3,  # 'd'
        1: 2,  # 'e'
        27: 1,  # 'f'
        12: 3,  # 'g'
        20: 2,  # 'h'
        9: 2,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 1,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 1,  # 'u'
        19: 3,  # 'v'
        62: 1,  # 'x'
        16: 1,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 2,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 0,  # 'ö'
        31: 1,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    19: {  # 'v'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 3,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 1,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 2,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 2,  # 'ó'
        24: 2,  # 'ö'
        31: 1,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 1,  # 'ű'
    },
    62: {  # 'x'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 0,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 1,  # 'i'
        22: 0,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 1,  # 'o'
        23: 1,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 1,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 1,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    16: {  # 'y'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 3,  # 'e'
        27: 2,  # 'f'
        12: 2,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 2,  # 'j'
        7: 2,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 2,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 2,  # 'í'
        25: 2,  # 'ó'
        24: 3,  # 'ö'
        31: 2,  # 'ú'
        29: 2,  # 'ü'
        42: 1,  # 'ő'
        56: 2,  # 'ű'
    },
    11: {  # 'z'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 3,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 3,  # 'd'
        1: 3,  # 'e'
        27: 1,  # 'f'
        12: 2,  # 'g'
        20: 2,  # 'h'
        9: 3,  # 'i'
        22: 1,  # 'j'
        7: 3,  # 'k'
        6: 2,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 3,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 3,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 3,  # 'á'
        15: 3,  # 'é'
        30: 3,  # 'í'
        25: 3,  # 'ó'
        24: 3,  # 'ö'
        31: 2,  # 'ú'
        29: 3,  # 'ü'
        42: 2,  # 'ő'
        56: 1,  # 'ű'
    },
    51: {  # 'Á'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 1,  # 'F'
        49: 2,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 2,  # 'N'
        47: 0,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 2,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 1,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    44: {  # 'É'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 1,  # 'E'
        50: 0,  # 'F'
        49: 2,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 2,  # 'N'
        47: 0,  # 'O'
        46: 1,  # 'P'
        43: 2,  # 'R'
        33: 2,  # 'S'
        37: 2,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 2,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 3,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 0,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    61: {  # 'Í'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 1,  # 'J'
        36: 0,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 2,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 1,  # 'm'
        4: 0,  # 'n'
        8: 0,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 0,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    58: {  # 'Ó'
        28: 1,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 1,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 2,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 2,  # 'h'
        9: 0,  # 'i'
        22: 0,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 0,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 1,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    59: {  # 'Ö'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 0,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 1,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 0,  # 'b'
        26: 1,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 0,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 0,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 0,  # 'p'
        10: 2,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    60: {  # 'Ú'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 1,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 1,  # 'F'
        49: 1,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 0,  # 'b'
        26: 0,  # 'c'
        17: 0,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 2,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 2,  # 'j'
        7: 0,  # 'k'
        6: 0,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 0,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 0,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    63: {  # 'Ü'
        28: 0,  # 'A'
        40: 1,  # 'B'
        54: 0,  # 'C'
        45: 1,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 1,  # 'G'
        38: 1,  # 'H'
        39: 0,  # 'I'
        53: 1,  # 'J'
        36: 1,  # 'K'
        41: 1,  # 'L'
        34: 1,  # 'M'
        35: 1,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 1,  # 'R'
        33: 1,  # 'S'
        37: 1,  # 'T'
        57: 0,  # 'U'
        48: 1,  # 'V'
        55: 0,  # 'Y'
        52: 1,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 0,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 0,  # 'f'
        12: 1,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 0,  # 'j'
        7: 0,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        8: 0,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 1,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    14: {  # 'á'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 3,  # 'b'
        26: 3,  # 'c'
        17: 3,  # 'd'
        1: 1,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 2,  # 'h'
        9: 2,  # 'i'
        22: 3,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 1,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 2,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 1,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 2,  # 'é'
        30: 1,  # 'í'
        25: 0,  # 'ó'
        24: 1,  # 'ö'
        31: 0,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    15: {  # 'é'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 3,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 3,  # 'g'
        20: 3,  # 'h'
        9: 2,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 1,  # 'o'
        23: 3,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 0,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    30: {  # 'í'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 0,  # 'a'
        18: 1,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 0,  # 'e'
        27: 1,  # 'f'
        12: 3,  # 'g'
        20: 0,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 2,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 2,  # 's'
        3: 3,  # 't'
        21: 0,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    25: {  # 'ó'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 2,  # 'a'
        18: 3,  # 'b'
        26: 2,  # 'c'
        17: 3,  # 'd'
        1: 1,  # 'e'
        27: 2,  # 'f'
        12: 2,  # 'g'
        20: 2,  # 'h'
        9: 2,  # 'i'
        22: 2,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        8: 1,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 1,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 0,  # 'ó'
        24: 1,  # 'ö'
        31: 1,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    24: {  # 'ö'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 0,  # 'a'
        18: 3,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 0,  # 'e'
        27: 1,  # 'f'
        12: 2,  # 'g'
        20: 1,  # 'h'
        9: 0,  # 'i'
        22: 1,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        8: 0,  # 'o'
        23: 2,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 3,  # 't'
        21: 0,  # 'u'
        19: 3,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 3,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    31: {  # 'ú'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 1,  # 'b'
        26: 2,  # 'c'
        17: 1,  # 'd'
        1: 1,  # 'e'
        27: 2,  # 'f'
        12: 3,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 3,  # 'j'
        7: 1,  # 'k'
        6: 3,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 3,  # 'r'
        5: 3,  # 's'
        3: 2,  # 't'
        21: 1,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 1,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    29: {  # 'ü'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 1,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 3,  # 'g'
        20: 2,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 3,  # 'k'
        6: 3,  # 'l'
        13: 1,  # 'm'
        4: 3,  # 'n'
        8: 0,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 0,  # 'u'
        19: 2,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 1,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    42: {  # 'ő'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 2,  # 'b'
        26: 1,  # 'c'
        17: 2,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 2,  # 'k'
        6: 3,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        8: 1,  # 'o'
        23: 1,  # 'p'
        10: 2,  # 'r'
        5: 2,  # 's'
        3: 2,  # 't'
        21: 1,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 1,  # 'é'
        30: 1,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 1,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
    56: {  # 'ű'
        28: 0,  # 'A'
        40: 0,  # 'B'
        54: 0,  # 'C'
        45: 0,  # 'D'
        32: 0,  # 'E'
        50: 0,  # 'F'
        49: 0,  # 'G'
        38: 0,  # 'H'
        39: 0,  # 'I'
        53: 0,  # 'J'
        36: 0,  # 'K'
        41: 0,  # 'L'
        34: 0,  # 'M'
        35: 0,  # 'N'
        47: 0,  # 'O'
        46: 0,  # 'P'
        43: 0,  # 'R'
        33: 0,  # 'S'
        37: 0,  # 'T'
        57: 0,  # 'U'
        48: 0,  # 'V'
        55: 0,  # 'Y'
        52: 0,  # 'Z'
        2: 1,  # 'a'
        18: 1,  # 'b'
        26: 0,  # 'c'
        17: 1,  # 'd'
        1: 1,  # 'e'
        27: 1,  # 'f'
        12: 1,  # 'g'
        20: 1,  # 'h'
        9: 1,  # 'i'
        22: 1,  # 'j'
        7: 1,  # 'k'
        6: 1,  # 'l'
        13: 0,  # 'm'
        4: 2,  # 'n'
        8: 0,  # 'o'
        23: 0,  # 'p'
        10: 1,  # 'r'
        5: 1,  # 's'
        3: 1,  # 't'
        21: 0,  # 'u'
        19: 1,  # 'v'
        62: 0,  # 'x'
        16: 0,  # 'y'
        11: 2,  # 'z'
        51: 0,  # 'Á'
        44: 0,  # 'É'
        61: 0,  # 'Í'
        58: 0,  # 'Ó'
        59: 0,  # 'Ö'
        60: 0,  # 'Ú'
        63: 0,  # 'Ü'
        14: 0,  # 'á'
        15: 0,  # 'é'
        30: 0,  # 'í'
        25: 0,  # 'ó'
        24: 0,  # 'ö'
        31: 0,  # 'ú'
        29: 0,  # 'ü'
        42: 0,  # 'ő'
        56: 0,  # 'ű'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
WINDOWS_1250_HUNGARIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 28,  # 'A'
     66: 40,  # 'B'
     67: 54,  # 'C'
     68: 45,  # 'D'
     69: 32,  # 'E'
     70: 50,  # 'F'
     71: 49,  # 'G'
     72: 38,  # 'H'
     73: 39,  # 'I'
     74: 53,  # 'J'
     75: 36,  # 'K'
     76: 41,  # 'L'
     77: 34,  # 'M'
     78: 35,  # 'N'
     79: 47,  # 'O'
     80: 46,  # 'P'
     81: 72,  # 'Q'
     82: 43,  # 'R'
     83: 33,  # 'S'
     84: 37,  # 'T'
     85: 57,  # 'U'
     86: 48,  # 'V'
     87: 64,  # 'W'
     88: 68,  # 'X'
     89: 55,  # 'Y'
     90: 52,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 2,  # 'a'
     98: 18,  # 'b'
     99: 26,  # 'c'
     100: 17,  # 'd'
     101: 1,  # 'e'
     102: 27,  # 'f'
     103: 12,  # 'g'
     104: 20,  # 'h'
     105: 9,  # 'i'
     106: 22,  # 'j'
     107: 7,  # 'k'
     108: 6,  # 'l'
     109: 13,  # 'm'
     110: 4,  # 'n'
     111: 8,  # 'o'
     112: 23,  # 'p'
     113: 67,  # 'q'
     114: 10,  # 'r'
     115: 5,  # 's'
     116: 3,  # 't'
     117: 21,  # 'u'
     118: 19,  # 'v'
     119: 65,  # 'w'
     120: 62,  # 'x'
     121: 16,  # 'y'
     122: 11,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 161,  # '€'
     129: 162,  # None
     130: 163,  # '‚'
     131: 164,  # None
     132: 165,  # '„'
     133: 166,  # '…'
     134: 167,  # '†'
     135: 168,  # '‡'
     136: 169,  # None
     137: 170,  # '‰'
     138: 171,  # 'Š'
     139: 172,  # '‹'
     140: 173,  # 'Ś'
     141: 174,  # 'Ť'
     142: 175,  # 'Ž'
     143: 176,  # 'Ź'
     144: 177,  # None
     145: 178,  # '‘'
     146: 179,  # '’'
     147: 180,  # '“'
     148: 78,  # '”'
     149: 181,  # '•'
     150: 69,  # '–'
     151: 182,  # '—'
     152: 183,  # None
     153: 184,  # '™'
     154: 185,  # 'š'
     155: 186,  # '›'
     156: 187,  # 'ś'
     157: 188,  # 'ť'
     158: 189,  # 'ž'
     159: 190,  # 'ź'
     160: 191,  # '\xa0'
     161: 192,  # 'ˇ'
     162: 193,  # '˘'
     163: 194,  # 'Ł'
     164: 195,  # '¤'
     165: 196,  # 'Ą'
     166: 197,  # '¦'
     167: 76,  # '§'
     168: 198,  # '¨'
     169: 199,  # '©'
     170: 200,  # 'Ş'
     171: 201,  # '«'
     172: 202,  # '¬'
     173: 203,  # '\xad'
     174: 204,  # '®'
     175: 205,  # 'Ż'
     176: 81,  # '°'
     177: 206,  # '±'
     178: 207,  # '˛'
     179: 208,  # 'ł'
     180: 209,  # '´'
     181: 210,  # 'µ'
     182: 211,  # '¶'
     183: 212,  # '·'
     184: 213,  # '¸'
     185: 214,  # 'ą'
     186: 215,  # 'ş'
     187: 216,  # '»'
     188: 217,  # 'Ľ'
     189: 218,  # '˝'
     190: 219,  # 'ľ'
     191: 220,  # 'ż'
     192: 221,  # 'Ŕ'
     193: 51,  # 'Á'
     194: 83,  # 'Â'
     195: 222,  # 'Ă'
     196: 80,  # 'Ä'
     197: 223,  # 'Ĺ'
     198: 224,  # 'Ć'
     199: 225,  # 'Ç'
     200: 226,  # 'Č'
     201: 44,  # 'É'
     202: 227,  # 'Ę'
     203: 228,  # 'Ë'
     204: 229,  # 'Ě'
     205: 61,  # 'Í'
     206: 230,  # 'Î'
     207: 231,  # 'Ď'
     208: 232,  # 'Đ'
     209: 233,  # 'Ń'
     210: 234,  # 'Ň'
     211: 58,  # 'Ó'
     212: 235,  # 'Ô'
     213: 66,  # 'Ő'
     214: 59,  # 'Ö'
     215: 236,  # '×'
     216: 237,  # 'Ř'
     217: 238,  # 'Ů'
     218: 60,  # 'Ú'
     219: 70,  # 'Ű'
     220: 63,  # 'Ü'
     221: 239,  # 'Ý'
     222: 240,  # 'Ţ'
     223: 241,  # 'ß'
     224: 84,  # 'ŕ'
     225: 14,  # 'á'
     226: 75,  # 'â'
     227: 242,  # 'ă'
     228: 71,  # 'ä'
     229: 82,  # 'ĺ'
     230: 243,  # 'ć'
     231: 73,  # 'ç'
     232: 244,  # 'č'
     233: 15,  # 'é'
     234: 85,  # 'ę'
     235: 79,  # 'ë'
     236: 86,  # 'ě'
     237: 30,  # 'í'
     238: 77,  # 'î'
     239: 87,  # 'ď'
     240: 245,  # 'đ'
     241: 246,  # 'ń'
     242: 247,  # 'ň'
     243: 25,  # 'ó'
     244: 74,  # 'ô'
     245: 42,  # 'ő'
     246: 24,  # 'ö'
     247: 248,  # '÷'
     248: 249,  # 'ř'
     249: 250,  # 'ů'
     250: 31,  # 'ú'
     251: 56,  # 'ű'
     252: 29,  # 'ü'
     253: 251,  # 'ý'
     254: 252,  # 'ţ'
     255: 253,  # '˙'
}

WINDOWS_1250_HUNGARIAN_MODEL = SingleByteCharSetModel(charset_name='windows-1250',
                                                      language='Hungarian',
                                                      char_to_order_map=WINDOWS_1250_HUNGARIAN_CHAR_TO_ORDER,
                                                      language_model=HUNGARIAN_LANG_MODEL,
                                                      typical_positive_ratio=0.947368,
                                                      keep_ascii_letters=True,
                                                      alphabet='ABCDEFGHIJKLMNOPRSTUVZabcdefghijklmnoprstuvzÁÉÍÓÖÚÜáéíóöúüŐőŰű')

ISO_8859_2_HUNGARIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 28,  # 'A'
     66: 40,  # 'B'
     67: 54,  # 'C'
     68: 45,  # 'D'
     69: 32,  # 'E'
     70: 50,  # 'F'
     71: 49,  # 'G'
     72: 38,  # 'H'
     73: 39,  # 'I'
     74: 53,  # 'J'
     75: 36,  # 'K'
     76: 41,  # 'L'
     77: 34,  # 'M'
     78: 35,  # 'N'
     79: 47,  # 'O'
     80: 46,  # 'P'
     81: 71,  # 'Q'
     82: 43,  # 'R'
     83: 33,  # 'S'
     84: 37,  # 'T'
     85: 57,  # 'U'
     86: 48,  # 'V'
     87: 64,  # 'W'
     88: 68,  # 'X'
     89: 55,  # 'Y'
     90: 52,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 2,  # 'a'
     98: 18,  # 'b'
     99: 26,  # 'c'
     100: 17,  # 'd'
     101: 1,  # 'e'
     102: 27,  # 'f'
     103: 12,  # 'g'
     104: 20,  # 'h'
     105: 9,  # 'i'
     106: 22,  # 'j'
     107: 7,  # 'k'
     108: 6,  # 'l'
     109: 13,  # 'm'
     110: 4,  # 'n'
     111: 8,  # 'o'
     112: 23,  # 'p'
     113: 67,  # 'q'
     114: 10,  # 'r'
     115: 5,  # 's'
     116: 3,  # 't'
     117: 21,  # 'u'
     118: 19,  # 'v'
     119: 65,  # 'w'
     120: 62,  # 'x'
     121: 16,  # 'y'
     122: 11,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 159,  # '\x80'
     129: 160,  # '\x81'
     130: 161,  # '\x82'
     131: 162,  # '\x83'
     132: 163,  # '\x84'
     133: 164,  # '\x85'
     134: 165,  # '\x86'
     135: 166,  # '\x87'
     136: 167,  # '\x88'
     137: 168,  # '\x89'
     138: 169,  # '\x8a'
     139: 170,  # '\x8b'
     140: 171,  # '\x8c'
     141: 172,  # '\x8d'
     142: 173,  # '\x8e'
     143: 174,  # '\x8f'
     144: 175,  # '\x90'
     145: 176,  # '\x91'
     146: 177,  # '\x92'
     147: 178,  # '\x93'
     148: 179,  # '\x94'
     149: 180,  # '\x95'
     150: 181,  # '\x96'
     151: 182,  # '\x97'
     152: 183,  # '\x98'
     153: 184,  # '\x99'
     154: 185,  # '\x9a'
     155: 186,  # '\x9b'
     156: 187,  # '\x9c'
     157: 188,  # '\x9d'
     158: 189,  # '\x9e'
     159: 190,  # '\x9f'
     160: 191,  # '\xa0'
     161: 192,  # 'Ą'
     162: 193,  # '˘'
     163: 194,  # 'Ł'
     164: 195,  # '¤'
     165: 196,  # 'Ľ'
     166: 197,  # 'Ś'
     167: 75,  # '§'
     168: 198,  # '¨'
     169: 199,  # 'Š'
     170: 200,  # 'Ş'
     171: 201,  # 'Ť'
     172: 202,  # 'Ź'
     173: 203,  # '\xad'
     174: 204,  # 'Ž'
     175: 205,  # 'Ż'
     176: 79,  # '°'
     177: 206,  # 'ą'
     178: 207,  # '˛'
     179: 208,  # 'ł'
     180: 209,  # '´'
     181: 210,  # 'ľ'
     182: 211,  # 'ś'
     183: 212,  # 'ˇ'
     184: 213,  # '¸'
     185: 214,  # 'š'
     186: 215,  # 'ş'
     187: 216,  # 'ť'
     188: 217,  # 'ź'
     189: 218,  # '˝'
     190: 219,  # 'ž'
     191: 220,  # 'ż'
     192: 221,  # 'Ŕ'
     193: 51,  # 'Á'
     194: 81,  # 'Â'
     195: 222,  # 'Ă'
     196: 78,  # 'Ä'
     197: 223,  # 'Ĺ'
     198: 224,  # 'Ć'
     199: 225,  # 'Ç'
     200: 226,  # 'Č'
     201: 44,  # 'É'
     202: 227,  # 'Ę'
     203: 228,  # 'Ë'
     204: 229,  # 'Ě'
     205: 61,  # 'Í'
     206: 230,  # 'Î'
     207: 231,  # 'Ď'
     208: 232,  # 'Đ'
     209: 233,  # 'Ń'
     210: 234,  # 'Ň'
     211: 58,  # 'Ó'
     212: 235,  # 'Ô'
     213: 66,  # 'Ő'
     214: 59,  # 'Ö'
     215: 236,  # '×'
     216: 237,  # 'Ř'
     217: 238,  # 'Ů'
     218: 60,  # 'Ú'
     219: 69,  # 'Ű'
     220: 63,  # 'Ü'
     221: 239,  # 'Ý'
     222: 240,  # 'Ţ'
     223: 241,  # 'ß'
     224: 82,  # 'ŕ'
     225: 14,  # 'á'
     226: 74,  # 'â'
     227: 242,  # 'ă'
     228: 70,  # 'ä'
     229: 80,  # 'ĺ'
     230: 243,  # 'ć'
     231: 72,  # 'ç'
     232: 244,  # 'č'
     233: 15,  # 'é'
     234: 83,  # 'ę'
     235: 77,  # 'ë'
     236: 84,  # 'ě'
     237: 30,  # 'í'
     238: 76,  # 'î'
     239: 85,  # 'ď'
     240: 245,  # 'đ'
     241: 246,  # 'ń'
     242: 247,  # 'ň'
     243: 25,  # 'ó'
     244: 73,  # 'ô'
     245: 42,  # 'ő'
     246: 24,  # 'ö'
     247: 248,  # '÷'
     248: 249,  # 'ř'
     249: 250,  # 'ů'
     250: 31,  # 'ú'
     251: 56,  # 'ű'
     252: 29,  # 'ü'
     253: 251,  # 'ý'
     254: 252,  # 'ţ'
     255: 253,  # '˙'
}

ISO_8859_2_HUNGARIAN_MODEL = SingleByteCharSetModel(charset_name='ISO-8859-2',
                                                    language='Hungarian',
                                                    char_to_order_map=ISO_8859_2_HUNGARIAN_CHAR_TO_ORDER,
                                                    language_model=HUNGARIAN_LANG_MODEL,
                                                    typical_positive_ratio=0.947368,
                                                    keep_ascii_letters=True,
                                                    alphabet='ABCDEFGHIJKLMNOPRSTUVZabcdefghijklmnoprstuvzÁÉÍÓÖÚÜáéíóöúüŐőŰű')

site-packages/pip/_vendor/chardet/cli/__pycache__/chardetect.cpython-39.pyc000064400000005250150564745010022651 0ustar00a

��Re�
�@srdZddlmZmZmZddlZddlZddlmZddl	m
Z
ddlmZd
dd	�Z
dd
d�Zedkrne�dS)a
Script which takes one or more file paths and reports on their detected
encodings

Example::

    % chardetect somefile someotherfile
    somefile: windows-1252 with confidence 0.5
    someotherfile: ascii with confidence 1.0

If no paths are provided, it takes its input from stdin.

�)�absolute_import�print_function�unicode_literalsN)�__version__)�PY2)�UniversalDetector�stdincCszt�}|D] }t|�}|�|�|jr
q,q
|��|j}trN|�t�	�d�}|drld�
||d|d�Sd�
|�SdS)z�
    Return a string describing the probable encoding of a file or
    list of strings.

    :param lines: The lines to get the encoding of.
    :type lines: Iterable of bytes
    :param name: Name of file or collection of lines
    :type name: str
    �ignore�encodingz{}: {} with confidence {}�
confidencez
{}: no resultN)r�	bytearray�feed�done�close�resultr�decode�sys�getfilesystemencoding�format)�lines�name�u�liner�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/cli/chardetect.py�description_ofs

�rcCs�tjdd�}|jddt�d�dtr(tjntjjgd�|jdd	d
�t	�d�|�
|�}|jD]*}|��rvt
dtjd
�t
t||j��q\dS)z�
    Handles command line arguments and gets things started.

    :param argv: List of arguments, as if specified on the command-line.
                 If None, ``sys.argv[1:]`` is used instead.
    :type argv: list of str
    zVTakes one or more file paths and reports their detected                      encodings)�description�inputz^File whose encoding we would like to determine.                               (default: stdin)�rb�*)�help�type�nargs�defaultz	--version�versionz%(prog)s {})�actionr$z�You are running chardetect interactively. Press CTRL-D twice at the start of a blank line to signal the end of your input. If you want help, run chardetect --help
)�fileN)�argparse�ArgumentParser�add_argument�FileTyperrr�bufferrr�
parse_argsr�isatty�print�stderrrr)�argv�parser�args�frrr�main5s$	�
��

�r4�__main__)r)N)�__doc__�
__future__rrrr'r�pip._vendor.chardetrZpip._vendor.chardet.compatr�%pip._vendor.chardet.universaldetectorrrr4�__name__rrrr�<module>s

site-packages/pip/_vendor/chardet/cli/__pycache__/__init__.cpython-39.pyc000064400000000352150564745010022300 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/cli/__init__.py�<module>�site-packages/pip/_vendor/chardet/cli/__init__.py000064400000000001150564745010016000 0ustar00
site-packages/pip/_vendor/chardet/cli/chardetect.py000064400000005273150564745010016367 0ustar00"""
Script which takes one or more file paths and reports on their detected
encodings

Example::

    % chardetect somefile someotherfile
    somefile: windows-1252 with confidence 0.5
    someotherfile: ascii with confidence 1.0

If no paths are provided, it takes its input from stdin.

"""

from __future__ import absolute_import, print_function, unicode_literals

import argparse
import sys

from pip._vendor.chardet import __version__
from pip._vendor.chardet.compat import PY2
from pip._vendor.chardet.universaldetector import UniversalDetector


def description_of(lines, name='stdin'):
    """
    Return a string describing the probable encoding of a file or
    list of strings.

    :param lines: The lines to get the encoding of.
    :type lines: Iterable of bytes
    :param name: Name of file or collection of lines
    :type name: str
    """
    u = UniversalDetector()
    for line in lines:
        line = bytearray(line)
        u.feed(line)
        # shortcut out of the loop to save reading further - particularly useful if we read a BOM.
        if u.done:
            break
    u.close()
    result = u.result
    if PY2:
        name = name.decode(sys.getfilesystemencoding(), 'ignore')
    if result['encoding']:
        return '{}: {} with confidence {}'.format(name, result['encoding'],
                                                     result['confidence'])
    else:
        return '{}: no result'.format(name)


def main(argv=None):
    """
    Handles command line arguments and gets things started.

    :param argv: List of arguments, as if specified on the command-line.
                 If None, ``sys.argv[1:]`` is used instead.
    :type argv: list of str
    """
    # Get command line arguments
    parser = argparse.ArgumentParser(
        description="Takes one or more file paths and reports their detected \
                     encodings")
    parser.add_argument('input',
                        help='File whose encoding we would like to determine. \
                              (default: stdin)',
                        type=argparse.FileType('rb'), nargs='*',
                        default=[sys.stdin if PY2 else sys.stdin.buffer])
    parser.add_argument('--version', action='version',
                        version='%(prog)s {}'.format(__version__))
    args = parser.parse_args(argv)

    for f in args.input:
        if f.isatty():
            print("You are running chardetect interactively. Press " +
                  "CTRL-D twice at the start of a blank line to signal the " +
                  "end of your input. If you want help, run chardetect " +
                  "--help\n", file=sys.stderr)
        print(description_of(f, f.name))


if __name__ == '__main__':
    main()
site-packages/pip/_vendor/chardet/langthaimodel.py000064400000311572150564745010016324 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

THAI_LANG_MODEL = {
    5: {  # 'ก'
        5: 2,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 2,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 3,  # 'ฎ'
        57: 2,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 2,  # 'ณ'
        20: 2,  # 'ด'
        19: 3,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 1,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 1,  # 'ย'
        2: 3,  # 'ร'
        61: 2,  # 'ฤ'
        15: 3,  # 'ล'
        12: 3,  # 'ว'
        42: 2,  # 'ศ'
        46: 3,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 1,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 3,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 0,  # 'ึ'
        27: 2,  # 'ื'
        32: 2,  # 'ุ'
        35: 1,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 3,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    30: {  # 'ข'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 1,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 2,  # 'ณ'
        20: 0,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 2,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 1,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 2,  # 'ี'
        40: 3,  # 'ึ'
        27: 1,  # 'ื'
        32: 1,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 2,  # '่'
        7: 3,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    24: {  # 'ค'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 2,  # 'ค'
        8: 2,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 2,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 0,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 2,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 3,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 2,  # 'า'
        36: 3,  # 'ำ'
        23: 3,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 1,  # 'เ'
        28: 0,  # 'แ'
        41: 3,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    8: {  # 'ง'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 3,  # 'ค'
        8: 2,  # 'ง'
        26: 2,  # 'จ'
        52: 1,  # 'ฉ'
        34: 2,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 2,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 1,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 2,  # 'ศ'
        46: 1,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 1,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 1,  # 'ื'
        32: 1,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 3,  # 'ๆ'
        37: 0,  # '็'
        6: 2,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    26: {  # 'จ'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 0,  # 'ค'
        8: 2,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 1,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 1,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 1,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 3,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 3,  # 'ำ'
        23: 2,  # 'ิ'
        13: 1,  # 'ี'
        40: 3,  # 'ึ'
        27: 1,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 2,  # '่'
        7: 2,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    52: {  # 'ฉ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 3,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 3,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 1,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 1,  # 'ั'
        1: 1,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 1,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    34: {  # 'ช'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 1,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 1,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 1,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 2,  # 'ั'
        1: 3,  # 'า'
        36: 1,  # 'ำ'
        23: 3,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 1,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    51: {  # 'ซ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 1,  # 'ั'
        1: 1,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 2,  # 'ี'
        40: 3,  # 'ึ'
        27: 2,  # 'ื'
        32: 1,  # 'ุ'
        35: 1,  # 'ู'
        11: 1,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 1,  # '่'
        7: 2,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    47: {  # 'ญ'
        5: 1,  # 'ก'
        30: 1,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 3,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 2,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 2,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 0,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 0,  # '็'
        6: 2,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    58: {  # 'ฎ'
        5: 2,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 1,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    57: {  # 'ฏ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    49: {  # 'ฐ'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 2,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    53: {  # 'ฑ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    55: {  # 'ฒ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    43: {  # 'ณ'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 3,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 3,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 1,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 3,  # 'ะ'
        10: 0,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    20: {  # 'ด'
        5: 2,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 3,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 3,  # 'ั'
        1: 2,  # 'า'
        36: 2,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 1,  # 'ึ'
        27: 2,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 2,  # 'ๆ'
        37: 2,  # '็'
        6: 1,  # '่'
        7: 3,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    19: {  # 'ต'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 1,  # 'ต'
        44: 2,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 2,  # 'ภ'
        9: 1,  # 'ม'
        16: 1,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 0,  # 'ห'
        4: 3,  # 'อ'
        63: 1,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 3,  # 'ิ'
        13: 2,  # 'ี'
        40: 1,  # 'ึ'
        27: 1,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 2,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    44: {  # 'ถ'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 2,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 2,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 1,  # 'ี'
        40: 3,  # 'ึ'
        27: 2,  # 'ื'
        32: 2,  # 'ุ'
        35: 3,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 2,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    14: {  # 'ท'
        5: 1,  # 'ก'
        30: 1,  # 'ข'
        24: 3,  # 'ค'
        8: 1,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 3,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 3,  # 'ย'
        2: 3,  # 'ร'
        61: 1,  # 'ฤ'
        15: 1,  # 'ล'
        12: 2,  # 'ว'
        42: 3,  # 'ศ'
        46: 1,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 3,  # 'ำ'
        23: 2,  # 'ิ'
        13: 3,  # 'ี'
        40: 2,  # 'ึ'
        27: 1,  # 'ื'
        32: 3,  # 'ุ'
        35: 1,  # 'ู'
        11: 0,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    48: {  # 'ธ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 1,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 2,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 2,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    3: {  # 'น'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 3,  # 'ค'
        8: 1,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 1,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 2,  # 'ถ'
        14: 3,  # 'ท'
        48: 3,  # 'ธ'
        3: 2,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 1,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 1,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 3,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 3,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 3,  # 'โ'
        29: 3,  # 'ใ'
        33: 3,  # 'ไ'
        50: 2,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    17: {  # 'บ'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 1,  # 'ง'
        26: 1,  # 'จ'
        52: 1,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 2,  # 'อ'
        63: 1,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 2,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 2,  # 'ื'
        32: 3,  # 'ุ'
        35: 2,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 2,  # '่'
        7: 2,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    25: {  # 'ป'
        5: 2,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 1,  # 'ฎ'
        57: 3,  # 'ฏ'
        49: 1,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 1,  # 'ต'
        44: 1,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 0,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 1,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 3,  # 'ั'
        1: 1,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 3,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 1,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 2,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 2,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 3,  # '็'
        6: 1,  # '่'
        7: 2,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    39: {  # 'ผ'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 1,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 2,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 1,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 1,  # 'ื'
        32: 0,  # 'ุ'
        35: 3,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 1,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    62: {  # 'ฝ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 1,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 1,  # 'ี'
        40: 2,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 2,  # '่'
        7: 1,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    31: {  # 'พ'
        5: 1,  # 'ก'
        30: 1,  # 'ข'
        24: 1,  # 'ค'
        8: 1,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 1,  # 'ณ'
        20: 1,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 0,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 2,  # 'ย'
        2: 3,  # 'ร'
        61: 2,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 1,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 2,  # 'ี'
        40: 1,  # 'ึ'
        27: 3,  # 'ื'
        32: 1,  # 'ุ'
        35: 2,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 1,  # '็'
        6: 0,  # '่'
        7: 1,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    54: {  # 'ฟ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 2,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 2,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 1,  # 'ื'
        32: 1,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 2,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    45: {  # 'ภ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 3,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 2,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    9: {  # 'ม'
        5: 2,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 2,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 1,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 3,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 1,  # 'ย'
        2: 2,  # 'ร'
        61: 2,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 1,  # 'ศ'
        46: 1,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 0,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 3,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 2,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 2,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    16: {  # 'ย'
        5: 3,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 2,  # 'ช'
        51: 0,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 2,  # 'ม'
        16: 0,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 3,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 3,  # 'ี'
        40: 1,  # 'ึ'
        27: 2,  # 'ื'
        32: 2,  # 'ุ'
        35: 3,  # 'ู'
        11: 2,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 2,  # 'ๆ'
        37: 1,  # '็'
        6: 3,  # '่'
        7: 2,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    2: {  # 'ร'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 2,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 3,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 3,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 3,  # 'ถ'
        14: 3,  # 'ท'
        48: 1,  # 'ธ'
        3: 2,  # 'น'
        17: 2,  # 'บ'
        25: 3,  # 'ป'
        39: 2,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 2,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 2,  # 'ศ'
        46: 2,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 1,  # 'ฯ'
        22: 3,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 2,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 3,  # 'ู'
        11: 3,  # 'เ'
        28: 3,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 3,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 3,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    61: {  # 'ฤ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 2,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    15: {  # 'ล'
        5: 2,  # 'ก'
        30: 3,  # 'ข'
        24: 1,  # 'ค'
        8: 3,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 1,  # 'ม'
        16: 3,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 3,  # 'อ'
        63: 2,  # 'ฯ'
        22: 3,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 2,  # 'ึ'
        27: 3,  # 'ื'
        32: 2,  # 'ุ'
        35: 3,  # 'ู'
        11: 2,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 2,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    12: {  # 'ว'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 1,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 1,  # 'ณ'
        20: 2,  # 'ด'
        19: 1,  # 'ต'
        44: 1,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 3,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 3,  # 'ิ'
        13: 2,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 2,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    42: {  # 'ศ'
        5: 1,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 1,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 2,  # 'ว'
        42: 1,  # 'ศ'
        46: 2,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 2,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 2,  # 'ิ'
        13: 0,  # 'ี'
        40: 3,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 2,  # 'ู'
        11: 0,  # 'เ'
        28: 1,  # 'แ'
        41: 0,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    46: {  # 'ษ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 2,  # 'ฎ'
        57: 1,  # 'ฏ'
        49: 2,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 3,  # 'ณ'
        20: 0,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 1,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 2,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    18: {  # 'ส'
        5: 2,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 2,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 3,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 2,  # 'ภ'
        9: 3,  # 'ม'
        16: 1,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 3,  # 'ำ'
        23: 3,  # 'ิ'
        13: 3,  # 'ี'
        40: 2,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 3,  # 'ู'
        11: 2,  # 'เ'
        28: 0,  # 'แ'
        41: 1,  # 'โ'
        29: 0,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 1,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    21: {  # 'ห'
        5: 3,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 1,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 3,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 0,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 0,  # 'ำ'
        23: 1,  # 'ิ'
        13: 1,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 1,  # 'ุ'
        35: 1,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 3,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    4: {  # 'อ'
        5: 3,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 3,  # 'ม'
        16: 3,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 2,  # 'ะ'
        10: 3,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 2,  # 'ิ'
        13: 3,  # 'ี'
        40: 0,  # 'ึ'
        27: 3,  # 'ื'
        32: 3,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 1,  # '็'
        6: 2,  # '่'
        7: 2,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    63: {  # 'ฯ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    22: {  # 'ะ'
        5: 3,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 1,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 3,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 1,  # 'ธ'
        3: 2,  # 'น'
        17: 3,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 2,  # 'อ'
        63: 1,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    10: {  # 'ั'
        5: 3,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 3,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 3,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 2,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 3,  # 'ฒ'
        43: 3,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 3,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 2,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    1: {  # 'า'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 3,  # 'ค'
        8: 3,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 3,  # 'ช'
        51: 1,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 3,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 2,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 3,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 3,  # 'ม'
        16: 3,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 3,  # 'ว'
        42: 2,  # 'ศ'
        46: 3,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 2,  # 'อ'
        63: 1,  # 'ฯ'
        22: 3,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    36: {  # 'ำ'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 3,  # 'ค'
        8: 2,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 1,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 1,  # 'ต'
        44: 1,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 3,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    23: {  # 'ิ'
        5: 3,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 3,  # 'ช'
        51: 0,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 3,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 3,  # 'พ'
        54: 1,  # 'ฟ'
        45: 2,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 3,  # 'ศ'
        46: 2,  # 'ษ'
        18: 2,  # 'ส'
        21: 3,  # 'ห'
        4: 1,  # 'อ'
        63: 1,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 2,  # '้'
        38: 2,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    13: {  # 'ี'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 1,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 3,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 2,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    40: {  # 'ึ'
        5: 3,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 3,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 1,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    27: {  # 'ื'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 3,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    32: {  # 'ุ'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 3,  # 'ค'
        8: 3,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 1,  # 'ฒ'
        43: 3,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 2,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 3,  # 'ม'
        16: 1,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 1,  # 'ว'
        42: 1,  # 'ศ'
        46: 2,  # 'ษ'
        18: 1,  # 'ส'
        21: 1,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 0,  # 'แ'
        41: 1,  # 'โ'
        29: 0,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 2,  # '้'
        38: 1,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    35: {  # 'ู'
        5: 3,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 2,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 2,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 1,  # 'ณ'
        20: 2,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 2,  # 'น'
        17: 0,  # 'บ'
        25: 3,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 1,  # 'แ'
        41: 1,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 3,  # '่'
        7: 3,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    11: {  # 'เ'
        5: 3,  # 'ก'
        30: 3,  # 'ข'
        24: 3,  # 'ค'
        8: 2,  # 'ง'
        26: 3,  # 'จ'
        52: 3,  # 'ฉ'
        34: 3,  # 'ช'
        51: 2,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 1,  # 'ณ'
        20: 3,  # 'ด'
        19: 3,  # 'ต'
        44: 1,  # 'ถ'
        14: 3,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 3,  # 'ป'
        39: 2,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 3,  # 'พ'
        54: 1,  # 'ฟ'
        45: 3,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 3,  # 'ว'
        42: 2,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    28: {  # 'แ'
        5: 3,  # 'ก'
        30: 2,  # 'ข'
        24: 2,  # 'ค'
        8: 1,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 3,  # 'ต'
        44: 2,  # 'ถ'
        14: 3,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 2,  # 'ป'
        39: 3,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 2,  # 'พ'
        54: 2,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    41: {  # 'โ'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 1,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 1,  # 'บ'
        25: 3,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 1,  # 'ภ'
        9: 1,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 3,  # 'ล'
        12: 0,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 0,  # 'ห'
        4: 2,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    29: {  # 'ใ'
        5: 2,  # 'ก'
        30: 0,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 3,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 1,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 3,  # 'ส'
        21: 3,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    33: {  # 'ไ'
        5: 1,  # 'ก'
        30: 2,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 3,  # 'ด'
        19: 1,  # 'ต'
        44: 0,  # 'ถ'
        14: 3,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 1,  # 'บ'
        25: 3,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 2,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 0,  # 'ย'
        2: 3,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 3,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 2,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    50: {  # 'ๆ'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    37: {  # '็'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 2,  # 'ง'
        26: 3,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 1,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 2,  # 'ต'
        44: 0,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 3,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 1,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 2,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 0,  # 'ห'
        4: 1,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 1,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    6: {  # '่'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 1,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 1,  # 'ธ'
        3: 3,  # 'น'
        17: 1,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 1,  # 'ฝ'
        31: 1,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 3,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 2,  # 'ล'
        12: 3,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 1,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 1,  # 'ะ'
        10: 0,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 3,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 1,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    7: {  # '้'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 2,  # 'ค'
        8: 3,  # 'ง'
        26: 2,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 1,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 1,  # 'ด'
        19: 2,  # 'ต'
        44: 1,  # 'ถ'
        14: 2,  # 'ท'
        48: 0,  # 'ธ'
        3: 3,  # 'น'
        17: 2,  # 'บ'
        25: 2,  # 'ป'
        39: 2,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 3,  # 'ม'
        16: 2,  # 'ย'
        2: 2,  # 'ร'
        61: 0,  # 'ฤ'
        15: 1,  # 'ล'
        12: 3,  # 'ว'
        42: 1,  # 'ศ'
        46: 0,  # 'ษ'
        18: 2,  # 'ส'
        21: 2,  # 'ห'
        4: 3,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 3,  # 'า'
        36: 2,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 2,  # 'ใ'
        33: 2,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    38: {  # '์'
        5: 2,  # 'ก'
        30: 1,  # 'ข'
        24: 1,  # 'ค'
        8: 0,  # 'ง'
        26: 1,  # 'จ'
        52: 0,  # 'ฉ'
        34: 1,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 2,  # 'ด'
        19: 1,  # 'ต'
        44: 1,  # 'ถ'
        14: 1,  # 'ท'
        48: 0,  # 'ธ'
        3: 1,  # 'น'
        17: 1,  # 'บ'
        25: 1,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 1,  # 'พ'
        54: 1,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 2,  # 'ม'
        16: 0,  # 'ย'
        2: 1,  # 'ร'
        61: 1,  # 'ฤ'
        15: 1,  # 'ล'
        12: 1,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 1,  # 'ส'
        21: 1,  # 'ห'
        4: 2,  # 'อ'
        63: 1,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 2,  # 'เ'
        28: 2,  # 'แ'
        41: 1,  # 'โ'
        29: 1,  # 'ใ'
        33: 1,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 0,  # '๑'
        59: 0,  # '๒'
        60: 0,  # '๕'
    },
    56: {  # '๑'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 2,  # '๑'
        59: 1,  # '๒'
        60: 1,  # '๕'
    },
    59: {  # '๒'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 1,  # '๑'
        59: 1,  # '๒'
        60: 3,  # '๕'
    },
    60: {  # '๕'
        5: 0,  # 'ก'
        30: 0,  # 'ข'
        24: 0,  # 'ค'
        8: 0,  # 'ง'
        26: 0,  # 'จ'
        52: 0,  # 'ฉ'
        34: 0,  # 'ช'
        51: 0,  # 'ซ'
        47: 0,  # 'ญ'
        58: 0,  # 'ฎ'
        57: 0,  # 'ฏ'
        49: 0,  # 'ฐ'
        53: 0,  # 'ฑ'
        55: 0,  # 'ฒ'
        43: 0,  # 'ณ'
        20: 0,  # 'ด'
        19: 0,  # 'ต'
        44: 0,  # 'ถ'
        14: 0,  # 'ท'
        48: 0,  # 'ธ'
        3: 0,  # 'น'
        17: 0,  # 'บ'
        25: 0,  # 'ป'
        39: 0,  # 'ผ'
        62: 0,  # 'ฝ'
        31: 0,  # 'พ'
        54: 0,  # 'ฟ'
        45: 0,  # 'ภ'
        9: 0,  # 'ม'
        16: 0,  # 'ย'
        2: 0,  # 'ร'
        61: 0,  # 'ฤ'
        15: 0,  # 'ล'
        12: 0,  # 'ว'
        42: 0,  # 'ศ'
        46: 0,  # 'ษ'
        18: 0,  # 'ส'
        21: 0,  # 'ห'
        4: 0,  # 'อ'
        63: 0,  # 'ฯ'
        22: 0,  # 'ะ'
        10: 0,  # 'ั'
        1: 0,  # 'า'
        36: 0,  # 'ำ'
        23: 0,  # 'ิ'
        13: 0,  # 'ี'
        40: 0,  # 'ึ'
        27: 0,  # 'ื'
        32: 0,  # 'ุ'
        35: 0,  # 'ู'
        11: 0,  # 'เ'
        28: 0,  # 'แ'
        41: 0,  # 'โ'
        29: 0,  # 'ใ'
        33: 0,  # 'ไ'
        50: 0,  # 'ๆ'
        37: 0,  # '็'
        6: 0,  # '่'
        7: 0,  # '้'
        38: 0,  # '์'
        56: 2,  # '๑'
        59: 1,  # '๒'
        60: 0,  # '๕'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
TIS_620_THAI_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 182,  # 'A'
     66: 106,  # 'B'
     67: 107,  # 'C'
     68: 100,  # 'D'
     69: 183,  # 'E'
     70: 184,  # 'F'
     71: 185,  # 'G'
     72: 101,  # 'H'
     73: 94,  # 'I'
     74: 186,  # 'J'
     75: 187,  # 'K'
     76: 108,  # 'L'
     77: 109,  # 'M'
     78: 110,  # 'N'
     79: 111,  # 'O'
     80: 188,  # 'P'
     81: 189,  # 'Q'
     82: 190,  # 'R'
     83: 89,  # 'S'
     84: 95,  # 'T'
     85: 112,  # 'U'
     86: 113,  # 'V'
     87: 191,  # 'W'
     88: 192,  # 'X'
     89: 193,  # 'Y'
     90: 194,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 64,  # 'a'
     98: 72,  # 'b'
     99: 73,  # 'c'
     100: 114,  # 'd'
     101: 74,  # 'e'
     102: 115,  # 'f'
     103: 116,  # 'g'
     104: 102,  # 'h'
     105: 81,  # 'i'
     106: 201,  # 'j'
     107: 117,  # 'k'
     108: 90,  # 'l'
     109: 103,  # 'm'
     110: 78,  # 'n'
     111: 82,  # 'o'
     112: 96,  # 'p'
     113: 202,  # 'q'
     114: 91,  # 'r'
     115: 79,  # 's'
     116: 84,  # 't'
     117: 104,  # 'u'
     118: 105,  # 'v'
     119: 97,  # 'w'
     120: 98,  # 'x'
     121: 92,  # 'y'
     122: 203,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 209,  # '\x80'
     129: 210,  # '\x81'
     130: 211,  # '\x82'
     131: 212,  # '\x83'
     132: 213,  # '\x84'
     133: 88,  # '\x85'
     134: 214,  # '\x86'
     135: 215,  # '\x87'
     136: 216,  # '\x88'
     137: 217,  # '\x89'
     138: 218,  # '\x8a'
     139: 219,  # '\x8b'
     140: 220,  # '\x8c'
     141: 118,  # '\x8d'
     142: 221,  # '\x8e'
     143: 222,  # '\x8f'
     144: 223,  # '\x90'
     145: 224,  # '\x91'
     146: 99,  # '\x92'
     147: 85,  # '\x93'
     148: 83,  # '\x94'
     149: 225,  # '\x95'
     150: 226,  # '\x96'
     151: 227,  # '\x97'
     152: 228,  # '\x98'
     153: 229,  # '\x99'
     154: 230,  # '\x9a'
     155: 231,  # '\x9b'
     156: 232,  # '\x9c'
     157: 233,  # '\x9d'
     158: 234,  # '\x9e'
     159: 235,  # '\x9f'
     160: 236,  # None
     161: 5,  # 'ก'
     162: 30,  # 'ข'
     163: 237,  # 'ฃ'
     164: 24,  # 'ค'
     165: 238,  # 'ฅ'
     166: 75,  # 'ฆ'
     167: 8,  # 'ง'
     168: 26,  # 'จ'
     169: 52,  # 'ฉ'
     170: 34,  # 'ช'
     171: 51,  # 'ซ'
     172: 119,  # 'ฌ'
     173: 47,  # 'ญ'
     174: 58,  # 'ฎ'
     175: 57,  # 'ฏ'
     176: 49,  # 'ฐ'
     177: 53,  # 'ฑ'
     178: 55,  # 'ฒ'
     179: 43,  # 'ณ'
     180: 20,  # 'ด'
     181: 19,  # 'ต'
     182: 44,  # 'ถ'
     183: 14,  # 'ท'
     184: 48,  # 'ธ'
     185: 3,  # 'น'
     186: 17,  # 'บ'
     187: 25,  # 'ป'
     188: 39,  # 'ผ'
     189: 62,  # 'ฝ'
     190: 31,  # 'พ'
     191: 54,  # 'ฟ'
     192: 45,  # 'ภ'
     193: 9,  # 'ม'
     194: 16,  # 'ย'
     195: 2,  # 'ร'
     196: 61,  # 'ฤ'
     197: 15,  # 'ล'
     198: 239,  # 'ฦ'
     199: 12,  # 'ว'
     200: 42,  # 'ศ'
     201: 46,  # 'ษ'
     202: 18,  # 'ส'
     203: 21,  # 'ห'
     204: 76,  # 'ฬ'
     205: 4,  # 'อ'
     206: 66,  # 'ฮ'
     207: 63,  # 'ฯ'
     208: 22,  # 'ะ'
     209: 10,  # 'ั'
     210: 1,  # 'า'
     211: 36,  # 'ำ'
     212: 23,  # 'ิ'
     213: 13,  # 'ี'
     214: 40,  # 'ึ'
     215: 27,  # 'ื'
     216: 32,  # 'ุ'
     217: 35,  # 'ู'
     218: 86,  # 'ฺ'
     219: 240,  # None
     220: 241,  # None
     221: 242,  # None
     222: 243,  # None
     223: 244,  # '฿'
     224: 11,  # 'เ'
     225: 28,  # 'แ'
     226: 41,  # 'โ'
     227: 29,  # 'ใ'
     228: 33,  # 'ไ'
     229: 245,  # 'ๅ'
     230: 50,  # 'ๆ'
     231: 37,  # '็'
     232: 6,  # '่'
     233: 7,  # '้'
     234: 67,  # '๊'
     235: 77,  # '๋'
     236: 38,  # '์'
     237: 93,  # 'ํ'
     238: 246,  # '๎'
     239: 247,  # '๏'
     240: 68,  # '๐'
     241: 56,  # '๑'
     242: 59,  # '๒'
     243: 65,  # '๓'
     244: 69,  # '๔'
     245: 60,  # '๕'
     246: 70,  # '๖'
     247: 80,  # '๗'
     248: 71,  # '๘'
     249: 87,  # '๙'
     250: 248,  # '๚'
     251: 249,  # '๛'
     252: 250,  # None
     253: 251,  # None
     254: 252,  # None
     255: 253,  # None
}

TIS_620_THAI_MODEL = SingleByteCharSetModel(charset_name='TIS-620',
                                            language='Thai',
                                            char_to_order_map=TIS_620_THAI_CHAR_TO_ORDER,
                                            language_model=THAI_LANG_MODEL,
                                            typical_positive_ratio=0.926386,
                                            keep_ascii_letters=False,
                                            alphabet='กขฃคฅฆงจฉชซฌญฎฏฐฑฒณดตถทธนบปผฝพฟภมยรฤลฦวศษสหฬอฮฯะัาำิีึืฺุู฿เแโใไๅๆ็่้๊๋์ํ๎๏๐๑๒๓๔๕๖๗๘๙๚๛')

site-packages/pip/_vendor/chardet/big5freq.py000064400000075026150564745010015221 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

# Big5 frequency table
# by Taiwan's Mandarin Promotion Council
# <http://www.edu.tw:81/mandr/>
#
# 128  --> 0.42261
# 256  --> 0.57851
# 512  --> 0.74851
# 1024 --> 0.89384
# 2048 --> 0.97583
#
# Ideal Distribution Ratio = 0.74851/(1-0.74851) =2.98
# Random Distribution Ration = 512/(5401-512)=0.105
#
# Typical Distribution Ratio about 25% of Ideal one, still much higher than RDR

BIG5_TYPICAL_DISTRIBUTION_RATIO = 0.75

#Char to FreqOrder table
BIG5_TABLE_SIZE = 5376

BIG5_CHAR_TO_FREQ_ORDER = (
   1,1801,1506, 255,1431, 198,   9,  82,   6,5008, 177, 202,3681,1256,2821, 110, #   16
3814,  33,3274, 261,  76,  44,2114,  16,2946,2187,1176, 659,3971,  26,3451,2653, #   32
1198,3972,3350,4202, 410,2215, 302, 590, 361,1964,   8, 204,  58,4510,5009,1932, #   48
  63,5010,5011, 317,1614,  75, 222, 159,4203,2417,1480,5012,3555,3091, 224,2822, #   64
3682,   3,  10,3973,1471,  29,2787,1135,2866,1940, 873, 130,3275,1123, 312,5013, #   80
4511,2052, 507, 252, 682,5014, 142,1915, 124, 206,2947,  34,3556,3204,  64, 604, #   96
5015,2501,1977,1978, 155,1991, 645, 641,1606,5016,3452, 337,  72, 406,5017,  80, #  112
 630, 238,3205,1509, 263, 939,1092,2654, 756,1440,1094,3453, 449,  69,2987, 591, #  128
 179,2096, 471, 115,2035,1844,  60,  50,2988, 134, 806,1869, 734,2036,3454, 180, #  144
 995,1607, 156, 537,2907, 688,5018, 319,1305, 779,2145, 514,2379, 298,4512, 359, #  160
2502,  90,2716,1338, 663,  11, 906,1099,2553,  20,2441, 182, 532,1716,5019, 732, #  176
1376,4204,1311,1420,3206,  25,2317,1056, 113, 399, 382,1950, 242,3455,2474, 529, #  192
3276, 475,1447,3683,5020, 117,  21, 656, 810,1297,2300,2334,3557,5021, 126,4205, #  208
 706, 456, 150, 613,4513,  71,1118,2037,4206, 145,3092,  85, 835, 486,2115,1246, #  224
1426, 428, 727,1285,1015, 800, 106, 623, 303,1281,5022,2128,2359, 347,3815, 221, #  240
3558,3135,5023,1956,1153,4207,  83, 296,1199,3093, 192, 624,  93,5024, 822,1898, #  256
2823,3136, 795,2065, 991,1554,1542,1592,  27,  43,2867, 859, 139,1456, 860,4514, #  272
 437, 712,3974, 164,2397,3137, 695, 211,3037,2097, 195,3975,1608,3559,3560,3684, #  288
3976, 234, 811,2989,2098,3977,2233,1441,3561,1615,2380, 668,2077,1638, 305, 228, #  304
1664,4515, 467, 415,5025, 262,2099,1593, 239, 108, 300, 200,1033, 512,1247,2078, #  320
5026,5027,2176,3207,3685,2682, 593, 845,1062,3277,  88,1723,2038,3978,1951, 212, #  336
 266, 152, 149, 468,1899,4208,4516,  77, 187,5028,3038,  37,   5,2990,5029,3979, #  352
5030,5031,  39,2524,4517,2908,3208,2079,  55, 148,  74,4518, 545, 483,1474,1029, #  368
1665, 217,1870,1531,3138,1104,2655,4209,  24, 172,3562, 900,3980,3563,3564,4519, #  384
  32,1408,2824,1312, 329, 487,2360,2251,2717, 784,2683,   4,3039,3351,1427,1789, #  400
 188, 109, 499,5032,3686,1717,1790, 888,1217,3040,4520,5033,3565,5034,3352,1520, #  416
3687,3981, 196,1034, 775,5035,5036, 929,1816, 249, 439,  38,5037,1063,5038, 794, #  432
3982,1435,2301,  46, 178,3278,2066,5039,2381,5040, 214,1709,4521, 804,  35, 707, #  448
 324,3688,1601,2554, 140, 459,4210,5041,5042,1365, 839, 272, 978,2262,2580,3456, #  464
2129,1363,3689,1423, 697, 100,3094,  48,  70,1231, 495,3139,2196,5043,1294,5044, #  480
2080, 462, 586,1042,3279, 853, 256, 988, 185,2382,3457,1698, 434,1084,5045,3458, #  496
 314,2625,2788,4522,2335,2336, 569,2285, 637,1817,2525, 757,1162,1879,1616,3459, #  512
 287,1577,2116, 768,4523,1671,2868,3566,2526,1321,3816, 909,2418,5046,4211, 933, #  528
3817,4212,2053,2361,1222,4524, 765,2419,1322, 786,4525,5047,1920,1462,1677,2909, #  544
1699,5048,4526,1424,2442,3140,3690,2600,3353,1775,1941,3460,3983,4213, 309,1369, #  560
1130,2825, 364,2234,1653,1299,3984,3567,3985,3986,2656, 525,1085,3041, 902,2001, #  576
1475, 964,4527, 421,1845,1415,1057,2286, 940,1364,3141, 376,4528,4529,1381,   7, #  592
2527, 983,2383, 336,1710,2684,1846, 321,3461, 559,1131,3042,2752,1809,1132,1313, #  608
 265,1481,1858,5049, 352,1203,2826,3280, 167,1089, 420,2827, 776, 792,1724,3568, #  624
4214,2443,3281,5050,4215,5051, 446, 229, 333,2753, 901,3818,1200,1557,4530,2657, #  640
1921, 395,2754,2685,3819,4216,1836, 125, 916,3209,2626,4531,5052,5053,3820,5054, #  656
5055,5056,4532,3142,3691,1133,2555,1757,3462,1510,2318,1409,3569,5057,2146, 438, #  672
2601,2910,2384,3354,1068, 958,3043, 461, 311,2869,2686,4217,1916,3210,4218,1979, #  688
 383, 750,2755,2627,4219, 274, 539, 385,1278,1442,5058,1154,1965, 384, 561, 210, #  704
  98,1295,2556,3570,5059,1711,2420,1482,3463,3987,2911,1257, 129,5060,3821, 642, #  720
 523,2789,2790,2658,5061, 141,2235,1333,  68, 176, 441, 876, 907,4220, 603,2602, #  736
 710, 171,3464, 404, 549,  18,3143,2398,1410,3692,1666,5062,3571,4533,2912,4534, #  752
5063,2991, 368,5064, 146, 366,  99, 871,3693,1543, 748, 807,1586,1185,  22,2263, #  768
 379,3822,3211,5065,3212, 505,1942,2628,1992,1382,2319,5066, 380,2362, 218, 702, #  784
1818,1248,3465,3044,3572,3355,3282,5067,2992,3694, 930,3283,3823,5068,  59,5069, #  800
 585, 601,4221, 497,3466,1112,1314,4535,1802,5070,1223,1472,2177,5071, 749,1837, #  816
 690,1900,3824,1773,3988,1476, 429,1043,1791,2236,2117, 917,4222, 447,1086,1629, #  832
5072, 556,5073,5074,2021,1654, 844,1090, 105, 550, 966,1758,2828,1008,1783, 686, #  848
1095,5075,2287, 793,1602,5076,3573,2603,4536,4223,2948,2302,4537,3825, 980,2503, #  864
 544, 353, 527,4538, 908,2687,2913,5077, 381,2629,1943,1348,5078,1341,1252, 560, #  880
3095,5079,3467,2870,5080,2054, 973, 886,2081, 143,4539,5081,5082, 157,3989, 496, #  896
4224,  57, 840, 540,2039,4540,4541,3468,2118,1445, 970,2264,1748,1966,2082,4225, #  912
3144,1234,1776,3284,2829,3695, 773,1206,2130,1066,2040,1326,3990,1738,1725,4226, #  928
 279,3145,  51,1544,2604, 423,1578,2131,2067, 173,4542,1880,5083,5084,1583, 264, #  944
 610,3696,4543,2444, 280, 154,5085,5086,5087,1739, 338,1282,3096, 693,2871,1411, #  960
1074,3826,2445,5088,4544,5089,5090,1240, 952,2399,5091,2914,1538,2688, 685,1483, #  976
4227,2475,1436, 953,4228,2055,4545, 671,2400,  79,4229,2446,3285, 608, 567,2689, #  992
3469,4230,4231,1691, 393,1261,1792,2401,5092,4546,5093,5094,5095,5096,1383,1672, # 1008
3827,3213,1464, 522,1119, 661,1150, 216, 675,4547,3991,1432,3574, 609,4548,2690, # 1024
2402,5097,5098,5099,4232,3045,   0,5100,2476, 315, 231,2447, 301,3356,4549,2385, # 1040
5101, 233,4233,3697,1819,4550,4551,5102,  96,1777,1315,2083,5103, 257,5104,1810, # 1056
3698,2718,1139,1820,4234,2022,1124,2164,2791,1778,2659,5105,3097, 363,1655,3214, # 1072
5106,2993,5107,5108,5109,3992,1567,3993, 718, 103,3215, 849,1443, 341,3357,2949, # 1088
1484,5110,1712, 127,  67, 339,4235,2403, 679,1412, 821,5111,5112, 834, 738, 351, # 1104
2994,2147, 846, 235,1497,1881, 418,1993,3828,2719, 186,1100,2148,2756,3575,1545, # 1120
1355,2950,2872,1377, 583,3994,4236,2581,2995,5113,1298,3699,1078,2557,3700,2363, # 1136
  78,3829,3830, 267,1289,2100,2002,1594,4237, 348, 369,1274,2197,2178,1838,4552, # 1152
1821,2830,3701,2757,2288,2003,4553,2951,2758, 144,3358, 882,4554,3995,2759,3470, # 1168
4555,2915,5114,4238,1726, 320,5115,3996,3046, 788,2996,5116,2831,1774,1327,2873, # 1184
3997,2832,5117,1306,4556,2004,1700,3831,3576,2364,2660, 787,2023, 506, 824,3702, # 1200
 534, 323,4557,1044,3359,2024,1901, 946,3471,5118,1779,1500,1678,5119,1882,4558, # 1216
 165, 243,4559,3703,2528, 123, 683,4239, 764,4560,  36,3998,1793, 589,2916, 816, # 1232
 626,1667,3047,2237,1639,1555,1622,3832,3999,5120,4000,2874,1370,1228,1933, 891, # 1248
2084,2917, 304,4240,5121, 292,2997,2720,3577, 691,2101,4241,1115,4561, 118, 662, # 1264
5122, 611,1156, 854,2386,1316,2875,   2, 386, 515,2918,5123,5124,3286, 868,2238, # 1280
1486, 855,2661, 785,2216,3048,5125,1040,3216,3578,5126,3146, 448,5127,1525,5128, # 1296
2165,4562,5129,3833,5130,4242,2833,3579,3147, 503, 818,4001,3148,1568, 814, 676, # 1312
1444, 306,1749,5131,3834,1416,1030, 197,1428, 805,2834,1501,4563,5132,5133,5134, # 1328
1994,5135,4564,5136,5137,2198,  13,2792,3704,2998,3149,1229,1917,5138,3835,2132, # 1344
5139,4243,4565,2404,3580,5140,2217,1511,1727,1120,5141,5142, 646,3836,2448, 307, # 1360
5143,5144,1595,3217,5145,5146,5147,3705,1113,1356,4002,1465,2529,2530,5148, 519, # 1376
5149, 128,2133,  92,2289,1980,5150,4003,1512, 342,3150,2199,5151,2793,2218,1981, # 1392
3360,4244, 290,1656,1317, 789, 827,2365,5152,3837,4566, 562, 581,4004,5153, 401, # 1408
4567,2252,  94,4568,5154,1399,2794,5155,1463,2025,4569,3218,1944,5156, 828,1105, # 1424
4245,1262,1394,5157,4246, 605,4570,5158,1784,2876,5159,2835, 819,2102, 578,2200, # 1440
2952,5160,1502, 436,3287,4247,3288,2836,4005,2919,3472,3473,5161,2721,2320,5162, # 1456
5163,2337,2068,  23,4571, 193, 826,3838,2103, 699,1630,4248,3098, 390,1794,1064, # 1472
3581,5164,1579,3099,3100,1400,5165,4249,1839,1640,2877,5166,4572,4573, 137,4250, # 1488
 598,3101,1967, 780, 104, 974,2953,5167, 278, 899, 253, 402, 572, 504, 493,1339, # 1504
5168,4006,1275,4574,2582,2558,5169,3706,3049,3102,2253, 565,1334,2722, 863,  41, # 1520
5170,5171,4575,5172,1657,2338,  19, 463,2760,4251, 606,5173,2999,3289,1087,2085, # 1536
1323,2662,3000,5174,1631,1623,1750,4252,2691,5175,2878, 791,2723,2663,2339, 232, # 1552
2421,5176,3001,1498,5177,2664,2630, 755,1366,3707,3290,3151,2026,1609, 119,1918, # 1568
3474, 862,1026,4253,5178,4007,3839,4576,4008,4577,2265,1952,2477,5179,1125, 817, # 1584
4254,4255,4009,1513,1766,2041,1487,4256,3050,3291,2837,3840,3152,5180,5181,1507, # 1600
5182,2692, 733,  40,1632,1106,2879, 345,4257, 841,2531, 230,4578,3002,1847,3292, # 1616
3475,5183,1263, 986,3476,5184, 735, 879, 254,1137, 857, 622,1300,1180,1388,1562, # 1632
4010,4011,2954, 967,2761,2665,1349, 592,2134,1692,3361,3003,1995,4258,1679,4012, # 1648
1902,2188,5185, 739,3708,2724,1296,1290,5186,4259,2201,2202,1922,1563,2605,2559, # 1664
1871,2762,3004,5187, 435,5188, 343,1108, 596,  17,1751,4579,2239,3477,3709,5189, # 1680
4580, 294,3582,2955,1693, 477, 979, 281,2042,3583, 643,2043,3710,2631,2795,2266, # 1696
1031,2340,2135,2303,3584,4581, 367,1249,2560,5190,3585,5191,4582,1283,3362,2005, # 1712
 240,1762,3363,4583,4584, 836,1069,3153, 474,5192,2149,2532, 268,3586,5193,3219, # 1728
1521,1284,5194,1658,1546,4260,5195,3587,3588,5196,4261,3364,2693,1685,4262, 961, # 1744
1673,2632, 190,2006,2203,3841,4585,4586,5197, 570,2504,3711,1490,5198,4587,2633, # 1760
3293,1957,4588, 584,1514, 396,1045,1945,5199,4589,1968,2449,5200,5201,4590,4013, # 1776
 619,5202,3154,3294, 215,2007,2796,2561,3220,4591,3221,4592, 763,4263,3842,4593, # 1792
5203,5204,1958,1767,2956,3365,3712,1174, 452,1477,4594,3366,3155,5205,2838,1253, # 1808
2387,2189,1091,2290,4264, 492,5206, 638,1169,1825,2136,1752,4014, 648, 926,1021, # 1824
1324,4595, 520,4596, 997, 847,1007, 892,4597,3843,2267,1872,3713,2405,1785,4598, # 1840
1953,2957,3103,3222,1728,4265,2044,3714,4599,2008,1701,3156,1551,  30,2268,4266, # 1856
5207,2027,4600,3589,5208, 501,5209,4267, 594,3478,2166,1822,3590,3479,3591,3223, # 1872
 829,2839,4268,5210,1680,3157,1225,4269,5211,3295,4601,4270,3158,2341,5212,4602, # 1888
4271,5213,4015,4016,5214,1848,2388,2606,3367,5215,4603, 374,4017, 652,4272,4273, # 1904
 375,1140, 798,5216,5217,5218,2366,4604,2269, 546,1659, 138,3051,2450,4605,5219, # 1920
2254, 612,1849, 910, 796,3844,1740,1371, 825,3845,3846,5220,2920,2562,5221, 692, # 1936
 444,3052,2634, 801,4606,4274,5222,1491, 244,1053,3053,4275,4276, 340,5223,4018, # 1952
1041,3005, 293,1168,  87,1357,5224,1539, 959,5225,2240, 721, 694,4277,3847, 219, # 1968
1478, 644,1417,3368,2666,1413,1401,1335,1389,4019,5226,5227,3006,2367,3159,1826, # 1984
 730,1515, 184,2840,  66,4607,5228,1660,2958, 246,3369, 378,1457, 226,3480, 975, # 2000
4020,2959,1264,3592, 674, 696,5229, 163,5230,1141,2422,2167, 713,3593,3370,4608, # 2016
4021,5231,5232,1186,  15,5233,1079,1070,5234,1522,3224,3594, 276,1050,2725, 758, # 2032
1126, 653,2960,3296,5235,2342, 889,3595,4022,3104,3007, 903,1250,4609,4023,3481, # 2048
3596,1342,1681,1718, 766,3297, 286,  89,2961,3715,5236,1713,5237,2607,3371,3008, # 2064
5238,2962,2219,3225,2880,5239,4610,2505,2533, 181, 387,1075,4024, 731,2190,3372, # 2080
5240,3298, 310, 313,3482,2304, 770,4278,  54,3054, 189,4611,3105,3848,4025,5241, # 2096
1230,1617,1850, 355,3597,4279,4612,3373, 111,4280,3716,1350,3160,3483,3055,4281, # 2112
2150,3299,3598,5242,2797,4026,4027,3009, 722,2009,5243,1071, 247,1207,2343,2478, # 2128
1378,4613,2010, 864,1437,1214,4614, 373,3849,1142,2220, 667,4615, 442,2763,2563, # 2144
3850,4028,1969,4282,3300,1840, 837, 170,1107, 934,1336,1883,5244,5245,2119,4283, # 2160
2841, 743,1569,5246,4616,4284, 582,2389,1418,3484,5247,1803,5248, 357,1395,1729, # 2176
3717,3301,2423,1564,2241,5249,3106,3851,1633,4617,1114,2086,4285,1532,5250, 482, # 2192
2451,4618,5251,5252,1492, 833,1466,5253,2726,3599,1641,2842,5254,1526,1272,3718, # 2208
4286,1686,1795, 416,2564,1903,1954,1804,5255,3852,2798,3853,1159,2321,5256,2881, # 2224
4619,1610,1584,3056,2424,2764, 443,3302,1163,3161,5257,5258,4029,5259,4287,2506, # 2240
3057,4620,4030,3162,2104,1647,3600,2011,1873,4288,5260,4289, 431,3485,5261, 250, # 2256
  97,  81,4290,5262,1648,1851,1558, 160, 848,5263, 866, 740,1694,5264,2204,2843, # 2272
3226,4291,4621,3719,1687, 950,2479, 426, 469,3227,3720,3721,4031,5265,5266,1188, # 2288
 424,1996, 861,3601,4292,3854,2205,2694, 168,1235,3602,4293,5267,2087,1674,4622, # 2304
3374,3303, 220,2565,1009,5268,3855, 670,3010, 332,1208, 717,5269,5270,3603,2452, # 2320
4032,3375,5271, 513,5272,1209,2882,3376,3163,4623,1080,5273,5274,5275,5276,2534, # 2336
3722,3604, 815,1587,4033,4034,5277,3605,3486,3856,1254,4624,1328,3058,1390,4035, # 2352
1741,4036,3857,4037,5278, 236,3858,2453,3304,5279,5280,3723,3859,1273,3860,4625, # 2368
5281, 308,5282,4626, 245,4627,1852,2480,1307,2583, 430, 715,2137,2454,5283, 270, # 2384
 199,2883,4038,5284,3606,2727,1753, 761,1754, 725,1661,1841,4628,3487,3724,5285, # 2400
5286, 587,  14,3305, 227,2608, 326, 480,2270, 943,2765,3607, 291, 650,1884,5287, # 2416
1702,1226, 102,1547,  62,3488, 904,4629,3489,1164,4294,5288,5289,1224,1548,2766, # 2432
 391, 498,1493,5290,1386,1419,5291,2056,1177,4630, 813, 880,1081,2368, 566,1145, # 2448
4631,2291,1001,1035,2566,2609,2242, 394,1286,5292,5293,2069,5294,  86,1494,1730, # 2464
4039, 491,1588, 745, 897,2963, 843,3377,4040,2767,2884,3306,1768, 998,2221,2070, # 2480
 397,1827,1195,1970,3725,3011,3378, 284,5295,3861,2507,2138,2120,1904,5296,4041, # 2496
2151,4042,4295,1036,3490,1905, 114,2567,4296, 209,1527,5297,5298,2964,2844,2635, # 2512
2390,2728,3164, 812,2568,5299,3307,5300,1559, 737,1885,3726,1210, 885,  28,2695, # 2528
3608,3862,5301,4297,1004,1780,4632,5302, 346,1982,2222,2696,4633,3863,1742, 797, # 2544
1642,4043,1934,1072,1384,2152, 896,4044,3308,3727,3228,2885,3609,5303,2569,1959, # 2560
4634,2455,1786,5304,5305,5306,4045,4298,1005,1308,3728,4299,2729,4635,4636,1528, # 2576
2610, 161,1178,4300,1983, 987,4637,1101,4301, 631,4046,1157,3229,2425,1343,1241, # 2592
1016,2243,2570, 372, 877,2344,2508,1160, 555,1935, 911,4047,5307, 466,1170, 169, # 2608
1051,2921,2697,3729,2481,3012,1182,2012,2571,1251,2636,5308, 992,2345,3491,1540, # 2624
2730,1201,2071,2406,1997,2482,5309,4638, 528,1923,2191,1503,1874,1570,2369,3379, # 2640
3309,5310, 557,1073,5311,1828,3492,2088,2271,3165,3059,3107, 767,3108,2799,4639, # 2656
1006,4302,4640,2346,1267,2179,3730,3230, 778,4048,3231,2731,1597,2667,5312,4641, # 2672
5313,3493,5314,5315,5316,3310,2698,1433,3311, 131,  95,1504,4049, 723,4303,3166, # 2688
1842,3610,2768,2192,4050,2028,2105,3731,5317,3013,4051,1218,5318,3380,3232,4052, # 2704
4304,2584, 248,1634,3864, 912,5319,2845,3732,3060,3865, 654,  53,5320,3014,5321, # 2720
1688,4642, 777,3494,1032,4053,1425,5322, 191, 820,2121,2846, 971,4643, 931,3233, # 2736
 135, 664, 783,3866,1998, 772,2922,1936,4054,3867,4644,2923,3234, 282,2732, 640, # 2752
1372,3495,1127, 922, 325,3381,5323,5324, 711,2045,5325,5326,4055,2223,2800,1937, # 2768
4056,3382,2224,2255,3868,2305,5327,4645,3869,1258,3312,4057,3235,2139,2965,4058, # 2784
4059,5328,2225, 258,3236,4646, 101,1227,5329,3313,1755,5330,1391,3314,5331,2924, # 2800
2057, 893,5332,5333,5334,1402,4305,2347,5335,5336,3237,3611,5337,5338, 878,1325, # 2816
1781,2801,4647, 259,1385,2585, 744,1183,2272,4648,5339,4060,2509,5340, 684,1024, # 2832
4306,5341, 472,3612,3496,1165,3315,4061,4062, 322,2153, 881, 455,1695,1152,1340, # 2848
 660, 554,2154,4649,1058,4650,4307, 830,1065,3383,4063,4651,1924,5342,1703,1919, # 2864
5343, 932,2273, 122,5344,4652, 947, 677,5345,3870,2637, 297,1906,1925,2274,4653, # 2880
2322,3316,5346,5347,4308,5348,4309,  84,4310, 112, 989,5349, 547,1059,4064, 701, # 2896
3613,1019,5350,4311,5351,3497, 942, 639, 457,2306,2456, 993,2966, 407, 851, 494, # 2912
4654,3384, 927,5352,1237,5353,2426,3385, 573,4312, 680, 921,2925,1279,1875, 285, # 2928
 790,1448,1984, 719,2168,5354,5355,4655,4065,4066,1649,5356,1541, 563,5357,1077, # 2944
5358,3386,3061,3498, 511,3015,4067,4068,3733,4069,1268,2572,3387,3238,4656,4657, # 2960
5359, 535,1048,1276,1189,2926,2029,3167,1438,1373,2847,2967,1134,2013,5360,4313, # 2976
1238,2586,3109,1259,5361, 700,5362,2968,3168,3734,4314,5363,4315,1146,1876,1907, # 2992
4658,2611,4070, 781,2427, 132,1589, 203, 147, 273,2802,2407, 898,1787,2155,4071, # 3008
4072,5364,3871,2803,5365,5366,4659,4660,5367,3239,5368,1635,3872, 965,5369,1805, # 3024
2699,1516,3614,1121,1082,1329,3317,4073,1449,3873,  65,1128,2848,2927,2769,1590, # 3040
3874,5370,5371,  12,2668,  45, 976,2587,3169,4661, 517,2535,1013,1037,3240,5372, # 3056
3875,2849,5373,3876,5374,3499,5375,2612, 614,1999,2323,3877,3110,2733,2638,5376, # 3072
2588,4316, 599,1269,5377,1811,3735,5378,2700,3111, 759,1060, 489,1806,3388,3318, # 3088
1358,5379,5380,2391,1387,1215,2639,2256, 490,5381,5382,4317,1759,2392,2348,5383, # 3104
4662,3878,1908,4074,2640,1807,3241,4663,3500,3319,2770,2349, 874,5384,5385,3501, # 3120
3736,1859,  91,2928,3737,3062,3879,4664,5386,3170,4075,2669,5387,3502,1202,1403, # 3136
3880,2969,2536,1517,2510,4665,3503,2511,5388,4666,5389,2701,1886,1495,1731,4076, # 3152
2370,4667,5390,2030,5391,5392,4077,2702,1216, 237,2589,4318,2324,4078,3881,4668, # 3168
4669,2703,3615,3504, 445,4670,5393,5394,5395,5396,2771,  61,4079,3738,1823,4080, # 3184
5397, 687,2046, 935, 925, 405,2670, 703,1096,1860,2734,4671,4081,1877,1367,2704, # 3200
3389, 918,2106,1782,2483, 334,3320,1611,1093,4672, 564,3171,3505,3739,3390, 945, # 3216
2641,2058,4673,5398,1926, 872,4319,5399,3506,2705,3112, 349,4320,3740,4082,4674, # 3232
3882,4321,3741,2156,4083,4675,4676,4322,4677,2408,2047, 782,4084, 400, 251,4323, # 3248
1624,5400,5401, 277,3742, 299,1265, 476,1191,3883,2122,4324,4325,1109, 205,5402, # 3264
2590,1000,2157,3616,1861,5403,5404,5405,4678,5406,4679,2573, 107,2484,2158,4085, # 3280
3507,3172,5407,1533, 541,1301, 158, 753,4326,2886,3617,5408,1696, 370,1088,4327, # 3296
4680,3618, 579, 327, 440, 162,2244, 269,1938,1374,3508, 968,3063,  56,1396,3113, # 3312
2107,3321,3391,5409,1927,2159,4681,3016,5410,3619,5411,5412,3743,4682,2485,5413, # 3328
2804,5414,1650,4683,5415,2613,5416,5417,4086,2671,3392,1149,3393,4087,3884,4088, # 3344
5418,1076,  49,5419, 951,3242,3322,3323, 450,2850, 920,5420,1812,2805,2371,4328, # 3360
1909,1138,2372,3885,3509,5421,3243,4684,1910,1147,1518,2428,4685,3886,5422,4686, # 3376
2393,2614, 260,1796,3244,5423,5424,3887,3324, 708,5425,3620,1704,5426,3621,1351, # 3392
1618,3394,3017,1887, 944,4329,3395,4330,3064,3396,4331,5427,3744, 422, 413,1714, # 3408
3325, 500,2059,2350,4332,2486,5428,1344,1911, 954,5429,1668,5430,5431,4089,2409, # 3424
4333,3622,3888,4334,5432,2307,1318,2512,3114, 133,3115,2887,4687, 629,  31,2851, # 3440
2706,3889,4688, 850, 949,4689,4090,2970,1732,2089,4335,1496,1853,5433,4091, 620, # 3456
3245, 981,1242,3745,3397,1619,3746,1643,3326,2140,2457,1971,1719,3510,2169,5434, # 3472
3246,5435,5436,3398,1829,5437,1277,4690,1565,2048,5438,1636,3623,3116,5439, 869, # 3488
2852, 655,3890,3891,3117,4092,3018,3892,1310,3624,4691,5440,5441,5442,1733, 558, # 3504
4692,3747, 335,1549,3065,1756,4336,3748,1946,3511,1830,1291,1192, 470,2735,2108, # 3520
2806, 913,1054,4093,5443,1027,5444,3066,4094,4693, 982,2672,3399,3173,3512,3247, # 3536
3248,1947,2807,5445, 571,4694,5446,1831,5447,3625,2591,1523,2429,5448,2090, 984, # 3552
4695,3749,1960,5449,3750, 852, 923,2808,3513,3751, 969,1519, 999,2049,2325,1705, # 3568
5450,3118, 615,1662, 151, 597,4095,2410,2326,1049, 275,4696,3752,4337, 568,3753, # 3584
3626,2487,4338,3754,5451,2430,2275, 409,3249,5452,1566,2888,3514,1002, 769,2853, # 3600
 194,2091,3174,3755,2226,3327,4339, 628,1505,5453,5454,1763,2180,3019,4096, 521, # 3616
1161,2592,1788,2206,2411,4697,4097,1625,4340,4341, 412,  42,3119, 464,5455,2642, # 3632
4698,3400,1760,1571,2889,3515,2537,1219,2207,3893,2643,2141,2373,4699,4700,3328, # 3648
1651,3401,3627,5456,5457,3628,2488,3516,5458,3756,5459,5460,2276,2092, 460,5461, # 3664
4701,5462,3020, 962, 588,3629, 289,3250,2644,1116,  52,5463,3067,1797,5464,5465, # 3680
5466,1467,5467,1598,1143,3757,4342,1985,1734,1067,4702,1280,3402, 465,4703,1572, # 3696
 510,5468,1928,2245,1813,1644,3630,5469,4704,3758,5470,5471,2673,1573,1534,5472, # 3712
5473, 536,1808,1761,3517,3894,3175,2645,5474,5475,5476,4705,3518,2929,1912,2809, # 3728
5477,3329,1122, 377,3251,5478, 360,5479,5480,4343,1529, 551,5481,2060,3759,1769, # 3744
2431,5482,2930,4344,3330,3120,2327,2109,2031,4706,1404, 136,1468,1479, 672,1171, # 3760
3252,2308, 271,3176,5483,2772,5484,2050, 678,2736, 865,1948,4707,5485,2014,4098, # 3776
2971,5486,2737,2227,1397,3068,3760,4708,4709,1735,2931,3403,3631,5487,3895, 509, # 3792
2854,2458,2890,3896,5488,5489,3177,3178,4710,4345,2538,4711,2309,1166,1010, 552, # 3808
 681,1888,5490,5491,2972,2973,4099,1287,1596,1862,3179, 358, 453, 736, 175, 478, # 3824
1117, 905,1167,1097,5492,1854,1530,5493,1706,5494,2181,3519,2292,3761,3520,3632, # 3840
4346,2093,4347,5495,3404,1193,2489,4348,1458,2193,2208,1863,1889,1421,3331,2932, # 3856
3069,2182,3521, 595,2123,5496,4100,5497,5498,4349,1707,2646, 223,3762,1359, 751, # 3872
3121, 183,3522,5499,2810,3021, 419,2374, 633, 704,3897,2394, 241,5500,5501,5502, # 3888
 838,3022,3763,2277,2773,2459,3898,1939,2051,4101,1309,3122,2246,1181,5503,1136, # 3904
2209,3899,2375,1446,4350,2310,4712,5504,5505,4351,1055,2615, 484,3764,5506,4102, # 3920
 625,4352,2278,3405,1499,4353,4103,5507,4104,4354,3253,2279,2280,3523,5508,5509, # 3936
2774, 808,2616,3765,3406,4105,4355,3123,2539, 526,3407,3900,4356, 955,5510,1620, # 3952
4357,2647,2432,5511,1429,3766,1669,1832, 994, 928,5512,3633,1260,5513,5514,5515, # 3968
1949,2293, 741,2933,1626,4358,2738,2460, 867,1184, 362,3408,1392,5516,5517,4106, # 3984
4359,1770,1736,3254,2934,4713,4714,1929,2707,1459,1158,5518,3070,3409,2891,1292, # 4000
1930,2513,2855,3767,1986,1187,2072,2015,2617,4360,5519,2574,2514,2170,3768,2490, # 4016
3332,5520,3769,4715,5521,5522, 666,1003,3023,1022,3634,4361,5523,4716,1814,2257, # 4032
 574,3901,1603, 295,1535, 705,3902,4362, 283, 858, 417,5524,5525,3255,4717,4718, # 4048
3071,1220,1890,1046,2281,2461,4107,1393,1599, 689,2575, 388,4363,5526,2491, 802, # 4064
5527,2811,3903,2061,1405,2258,5528,4719,3904,2110,1052,1345,3256,1585,5529, 809, # 4080
5530,5531,5532, 575,2739,3524, 956,1552,1469,1144,2328,5533,2329,1560,2462,3635, # 4096
3257,4108, 616,2210,4364,3180,2183,2294,5534,1833,5535,3525,4720,5536,1319,3770, # 4112
3771,1211,3636,1023,3258,1293,2812,5537,5538,5539,3905, 607,2311,3906, 762,2892, # 4128
1439,4365,1360,4721,1485,3072,5540,4722,1038,4366,1450,2062,2648,4367,1379,4723, # 4144
2593,5541,5542,4368,1352,1414,2330,2935,1172,5543,5544,3907,3908,4724,1798,1451, # 4160
5545,5546,5547,5548,2936,4109,4110,2492,2351, 411,4111,4112,3637,3333,3124,4725, # 4176
1561,2674,1452,4113,1375,5549,5550,  47,2974, 316,5551,1406,1591,2937,3181,5552, # 4192
1025,2142,3125,3182, 354,2740, 884,2228,4369,2412, 508,3772, 726,3638, 996,2433, # 4208
3639, 729,5553, 392,2194,1453,4114,4726,3773,5554,5555,2463,3640,2618,1675,2813, # 4224
 919,2352,2975,2353,1270,4727,4115,  73,5556,5557, 647,5558,3259,2856,2259,1550, # 4240
1346,3024,5559,1332, 883,3526,5560,5561,5562,5563,3334,2775,5564,1212, 831,1347, # 4256
4370,4728,2331,3909,1864,3073, 720,3910,4729,4730,3911,5565,4371,5566,5567,4731, # 4272
5568,5569,1799,4732,3774,2619,4733,3641,1645,2376,4734,5570,2938, 669,2211,2675, # 4288
2434,5571,2893,5572,5573,1028,3260,5574,4372,2413,5575,2260,1353,5576,5577,4735, # 4304
3183, 518,5578,4116,5579,4373,1961,5580,2143,4374,5581,5582,3025,2354,2355,3912, # 4320
 516,1834,1454,4117,2708,4375,4736,2229,2620,1972,1129,3642,5583,2776,5584,2976, # 4336
1422, 577,1470,3026,1524,3410,5585,5586, 432,4376,3074,3527,5587,2594,1455,2515, # 4352
2230,1973,1175,5588,1020,2741,4118,3528,4737,5589,2742,5590,1743,1361,3075,3529, # 4368
2649,4119,4377,4738,2295, 895, 924,4378,2171, 331,2247,3076, 166,1627,3077,1098, # 4384
5591,1232,2894,2231,3411,4739, 657, 403,1196,2377, 542,3775,3412,1600,4379,3530, # 4400
5592,4740,2777,3261, 576, 530,1362,4741,4742,2540,2676,3776,4120,5593, 842,3913, # 4416
5594,2814,2032,1014,4121, 213,2709,3413, 665, 621,4380,5595,3777,2939,2435,5596, # 4432
2436,3335,3643,3414,4743,4381,2541,4382,4744,3644,1682,4383,3531,1380,5597, 724, # 4448
2282, 600,1670,5598,1337,1233,4745,3126,2248,5599,1621,4746,5600, 651,4384,5601, # 4464
1612,4385,2621,5602,2857,5603,2743,2312,3078,5604, 716,2464,3079, 174,1255,2710, # 4480
4122,3645, 548,1320,1398, 728,4123,1574,5605,1891,1197,3080,4124,5606,3081,3082, # 4496
3778,3646,3779, 747,5607, 635,4386,4747,5608,5609,5610,4387,5611,5612,4748,5613, # 4512
3415,4749,2437, 451,5614,3780,2542,2073,4388,2744,4389,4125,5615,1764,4750,5616, # 4528
4390, 350,4751,2283,2395,2493,5617,4391,4126,2249,1434,4127, 488,4752, 458,4392, # 4544
4128,3781, 771,1330,2396,3914,2576,3184,2160,2414,1553,2677,3185,4393,5618,2494, # 4560
2895,2622,1720,2711,4394,3416,4753,5619,2543,4395,5620,3262,4396,2778,5621,2016, # 4576
2745,5622,1155,1017,3782,3915,5623,3336,2313, 201,1865,4397,1430,5624,4129,5625, # 4592
5626,5627,5628,5629,4398,1604,5630, 414,1866, 371,2595,4754,4755,3532,2017,3127, # 4608
4756,1708, 960,4399, 887, 389,2172,1536,1663,1721,5631,2232,4130,2356,2940,1580, # 4624
5632,5633,1744,4757,2544,4758,4759,5634,4760,5635,2074,5636,4761,3647,3417,2896, # 4640
4400,5637,4401,2650,3418,2815, 673,2712,2465, 709,3533,4131,3648,4402,5638,1148, # 4656
 502, 634,5639,5640,1204,4762,3649,1575,4763,2623,3783,5641,3784,3128, 948,3263, # 4672
 121,1745,3916,1110,5642,4403,3083,2516,3027,4132,3785,1151,1771,3917,1488,4133, # 4688
1987,5643,2438,3534,5644,5645,2094,5646,4404,3918,1213,1407,2816, 531,2746,2545, # 4704
3264,1011,1537,4764,2779,4405,3129,1061,5647,3786,3787,1867,2897,5648,2018, 120, # 4720
4406,4407,2063,3650,3265,2314,3919,2678,3419,1955,4765,4134,5649,3535,1047,2713, # 4736
1266,5650,1368,4766,2858, 649,3420,3920,2546,2747,1102,2859,2679,5651,5652,2000, # 4752
5653,1111,3651,2977,5654,2495,3921,3652,2817,1855,3421,3788,5655,5656,3422,2415, # 4768
2898,3337,3266,3653,5657,2577,5658,3654,2818,4135,1460, 856,5659,3655,5660,2899, # 4784
2978,5661,2900,3922,5662,4408, 632,2517, 875,3923,1697,3924,2296,5663,5664,4767, # 4800
3028,1239, 580,4768,4409,5665, 914, 936,2075,1190,4136,1039,2124,5666,5667,5668, # 4816
5669,3423,1473,5670,1354,4410,3925,4769,2173,3084,4137, 915,3338,4411,4412,3339, # 4832
1605,1835,5671,2748, 398,3656,4413,3926,4138, 328,1913,2860,4139,3927,1331,4414, # 4848
3029, 937,4415,5672,3657,4140,4141,3424,2161,4770,3425, 524, 742, 538,3085,1012, # 4864
5673,5674,3928,2466,5675, 658,1103, 225,3929,5676,5677,4771,5678,4772,5679,3267, # 4880
1243,5680,4142, 963,2250,4773,5681,2714,3658,3186,5682,5683,2596,2332,5684,4774, # 4896
5685,5686,5687,3536, 957,3426,2547,2033,1931,2941,2467, 870,2019,3659,1746,2780, # 4912
2781,2439,2468,5688,3930,5689,3789,3130,3790,3537,3427,3791,5690,1179,3086,5691, # 4928
3187,2378,4416,3792,2548,3188,3131,2749,4143,5692,3428,1556,2549,2297, 977,2901, # 4944
2034,4144,1205,3429,5693,1765,3430,3189,2125,1271, 714,1689,4775,3538,5694,2333, # 4960
3931, 533,4417,3660,2184, 617,5695,2469,3340,3539,2315,5696,5697,3190,5698,5699, # 4976
3932,1988, 618, 427,2651,3540,3431,5700,5701,1244,1690,5702,2819,4418,4776,5703, # 4992
3541,4777,5704,2284,1576, 473,3661,4419,3432, 972,5705,3662,5706,3087,5707,5708, # 5008
4778,4779,5709,3793,4145,4146,5710, 153,4780, 356,5711,1892,2902,4420,2144, 408, # 5024
 803,2357,5712,3933,5713,4421,1646,2578,2518,4781,4782,3934,5714,3935,4422,5715, # 5040
2416,3433, 752,5716,5717,1962,3341,2979,5718, 746,3030,2470,4783,4423,3794, 698, # 5056
4784,1893,4424,3663,2550,4785,3664,3936,5719,3191,3434,5720,1824,1302,4147,2715, # 5072
3937,1974,4425,5721,4426,3192, 823,1303,1288,1236,2861,3542,4148,3435, 774,3938, # 5088
5722,1581,4786,1304,2862,3939,4787,5723,2440,2162,1083,3268,4427,4149,4428, 344, # 5104
1173, 288,2316, 454,1683,5724,5725,1461,4788,4150,2597,5726,5727,4789, 985, 894, # 5120
5728,3436,3193,5729,1914,2942,3795,1989,5730,2111,1975,5731,4151,5732,2579,1194, # 5136
 425,5733,4790,3194,1245,3796,4429,5734,5735,2863,5736, 636,4791,1856,3940, 760, # 5152
1800,5737,4430,2212,1508,4792,4152,1894,1684,2298,5738,5739,4793,4431,4432,2213, # 5168
 479,5740,5741, 832,5742,4153,2496,5743,2980,2497,3797, 990,3132, 627,1815,2652, # 5184
4433,1582,4434,2126,2112,3543,4794,5744, 799,4435,3195,5745,4795,2113,1737,3031, # 5200
1018, 543, 754,4436,3342,1676,4796,4797,4154,4798,1489,5746,3544,5747,2624,2903, # 5216
4155,5748,5749,2981,5750,5751,5752,5753,3196,4799,4800,2185,1722,5754,3269,3270, # 5232
1843,3665,1715, 481, 365,1976,1857,5755,5756,1963,2498,4801,5757,2127,3666,3271, # 5248
 433,1895,2064,2076,5758, 602,2750,5759,5760,5761,5762,5763,3032,1628,3437,5764, # 5264
3197,4802,4156,2904,4803,2519,5765,2551,2782,5766,5767,5768,3343,4804,2905,5769, # 5280
4805,5770,2864,4806,4807,1221,2982,4157,2520,5771,5772,5773,1868,1990,5774,5775, # 5296
5776,1896,5777,5778,4808,1897,4158, 318,5779,2095,4159,4437,5780,5781, 485,5782, # 5312
 938,3941, 553,2680, 116,5783,3942,3667,5784,3545,2681,2783,3438,3344,2820,5785, # 5328
3668,2943,4160,1747,2944,2983,5786,5787, 207,5788,4809,5789,4810,2521,5790,3033, # 5344
 890,3669,3943,5791,1878,3798,3439,5792,2186,2358,3440,1652,5793,5794,5795, 941, # 5360
2299, 208,3546,4161,2020, 330,4438,3944,2906,2499,3799,4439,4811,5796,5797,5798, # 5376
)

site-packages/pip/_vendor/chardet/latin1prober.py000064400000012372150564745010016112 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetprober import CharSetProber
from .enums import ProbingState

FREQ_CAT_NUM = 4

UDF = 0  # undefined
OTH = 1  # other
ASC = 2  # ascii capital letter
ASS = 3  # ascii small letter
ACV = 4  # accent capital vowel
ACO = 5  # accent capital other
ASV = 6  # accent small vowel
ASO = 7  # accent small other
CLASS_NUM = 8  # total classes

Latin1_CharToClass = (
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 00 - 07
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 08 - 0F
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 10 - 17
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 18 - 1F
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 20 - 27
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 28 - 2F
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 30 - 37
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 38 - 3F
    OTH, ASC, ASC, ASC, ASC, ASC, ASC, ASC,   # 40 - 47
    ASC, ASC, ASC, ASC, ASC, ASC, ASC, ASC,   # 48 - 4F
    ASC, ASC, ASC, ASC, ASC, ASC, ASC, ASC,   # 50 - 57
    ASC, ASC, ASC, OTH, OTH, OTH, OTH, OTH,   # 58 - 5F
    OTH, ASS, ASS, ASS, ASS, ASS, ASS, ASS,   # 60 - 67
    ASS, ASS, ASS, ASS, ASS, ASS, ASS, ASS,   # 68 - 6F
    ASS, ASS, ASS, ASS, ASS, ASS, ASS, ASS,   # 70 - 77
    ASS, ASS, ASS, OTH, OTH, OTH, OTH, OTH,   # 78 - 7F
    OTH, UDF, OTH, ASO, OTH, OTH, OTH, OTH,   # 80 - 87
    OTH, OTH, ACO, OTH, ACO, UDF, ACO, UDF,   # 88 - 8F
    UDF, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # 90 - 97
    OTH, OTH, ASO, OTH, ASO, UDF, ASO, ACO,   # 98 - 9F
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # A0 - A7
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # A8 - AF
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # B0 - B7
    OTH, OTH, OTH, OTH, OTH, OTH, OTH, OTH,   # B8 - BF
    ACV, ACV, ACV, ACV, ACV, ACV, ACO, ACO,   # C0 - C7
    ACV, ACV, ACV, ACV, ACV, ACV, ACV, ACV,   # C8 - CF
    ACO, ACO, ACV, ACV, ACV, ACV, ACV, OTH,   # D0 - D7
    ACV, ACV, ACV, ACV, ACV, ACO, ACO, ACO,   # D8 - DF
    ASV, ASV, ASV, ASV, ASV, ASV, ASO, ASO,   # E0 - E7
    ASV, ASV, ASV, ASV, ASV, ASV, ASV, ASV,   # E8 - EF
    ASO, ASO, ASV, ASV, ASV, ASV, ASV, OTH,   # F0 - F7
    ASV, ASV, ASV, ASV, ASV, ASO, ASO, ASO,   # F8 - FF
)

# 0 : illegal
# 1 : very unlikely
# 2 : normal
# 3 : very likely
Latin1ClassModel = (
# UDF OTH ASC ASS ACV ACO ASV ASO
    0,  0,  0,  0,  0,  0,  0,  0,  # UDF
    0,  3,  3,  3,  3,  3,  3,  3,  # OTH
    0,  3,  3,  3,  3,  3,  3,  3,  # ASC
    0,  3,  3,  3,  1,  1,  3,  3,  # ASS
    0,  3,  3,  3,  1,  2,  1,  2,  # ACV
    0,  3,  3,  3,  3,  3,  3,  3,  # ACO
    0,  3,  1,  3,  1,  1,  1,  3,  # ASV
    0,  3,  1,  3,  1,  1,  3,  3,  # ASO
)


class Latin1Prober(CharSetProber):
    def __init__(self):
        super(Latin1Prober, self).__init__()
        self._last_char_class = None
        self._freq_counter = None
        self.reset()

    def reset(self):
        self._last_char_class = OTH
        self._freq_counter = [0] * FREQ_CAT_NUM
        CharSetProber.reset(self)

    @property
    def charset_name(self):
        return "ISO-8859-1"

    @property
    def language(self):
        return ""

    def feed(self, byte_str):
        byte_str = self.filter_with_english_letters(byte_str)
        for c in byte_str:
            char_class = Latin1_CharToClass[c]
            freq = Latin1ClassModel[(self._last_char_class * CLASS_NUM)
                                    + char_class]
            if freq == 0:
                self._state = ProbingState.NOT_ME
                break
            self._freq_counter[freq] += 1
            self._last_char_class = char_class

        return self.state

    def get_confidence(self):
        if self.state == ProbingState.NOT_ME:
            return 0.01

        total = sum(self._freq_counter)
        if total < 0.01:
            confidence = 0.0
        else:
            confidence = ((self._freq_counter[3] - self._freq_counter[1] * 20.0)
                          / total)
        if confidence < 0.0:
            confidence = 0.0
        # lower the confidence of latin1 so that other more accurate
        # detector can take priority.
        confidence = confidence * 0.73
        return confidence
site-packages/pip/_vendor/chardet/eucjpprober.py000064400000007245150564745010016033 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .enums import ProbingState, MachineState
from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import EUCJPDistributionAnalysis
from .jpcntx import EUCJPContextAnalysis
from .mbcssm import EUCJP_SM_MODEL


class EUCJPProber(MultiByteCharSetProber):
    def __init__(self):
        super(EUCJPProber, self).__init__()
        self.coding_sm = CodingStateMachine(EUCJP_SM_MODEL)
        self.distribution_analyzer = EUCJPDistributionAnalysis()
        self.context_analyzer = EUCJPContextAnalysis()
        self.reset()

    def reset(self):
        super(EUCJPProber, self).reset()
        self.context_analyzer.reset()

    @property
    def charset_name(self):
        return "EUC-JP"

    @property
    def language(self):
        return "Japanese"

    def feed(self, byte_str):
        for i in range(len(byte_str)):
            # PY3K: byte_str is a byte array, so byte_str[i] is an int, not a byte
            coding_state = self.coding_sm.next_state(byte_str[i])
            if coding_state == MachineState.ERROR:
                self.logger.debug('%s %s prober hit error at byte %s',
                                  self.charset_name, self.language, i)
                self._state = ProbingState.NOT_ME
                break
            elif coding_state == MachineState.ITS_ME:
                self._state = ProbingState.FOUND_IT
                break
            elif coding_state == MachineState.START:
                char_len = self.coding_sm.get_current_charlen()
                if i == 0:
                    self._last_char[1] = byte_str[0]
                    self.context_analyzer.feed(self._last_char, char_len)
                    self.distribution_analyzer.feed(self._last_char, char_len)
                else:
                    self.context_analyzer.feed(byte_str[i - 1:i + 1],
                                                char_len)
                    self.distribution_analyzer.feed(byte_str[i - 1:i + 1],
                                                     char_len)

        self._last_char[0] = byte_str[-1]

        if self.state == ProbingState.DETECTING:
            if (self.context_analyzer.got_enough_data() and
               (self.get_confidence() > self.SHORTCUT_THRESHOLD)):
                self._state = ProbingState.FOUND_IT

        return self.state

    def get_confidence(self):
        context_conf = self.context_analyzer.get_confidence()
        distrib_conf = self.distribution_analyzer.get_confidence()
        return max(context_conf, distrib_conf)
site-packages/pip/_vendor/chardet/compat.py000064400000002260150564745010014766 0ustar00######################## BEGIN LICENSE BLOCK ########################
# Contributor(s):
#   Dan Blanchard
#   Ian Cordasco
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

import sys


if sys.version_info < (3, 0):
    PY2 = True
    PY3 = False
    string_types = (str, unicode)
    text_type = unicode
    iteritems = dict.iteritems
else:
    PY2 = False
    PY3 = True
    string_types = (bytes, str)
    text_type = str
    iteritems = dict.items
site-packages/pip/_vendor/chardet/langbulgarianmodel.py000064400000316313150564745010017341 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

BULGARIAN_LANG_MODEL = {
    63: {  # 'e'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 1,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 0,  # 'и'
        26: 1,  # 'й'
        12: 1,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 1,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 0,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    45: {  # '\xad'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 0,  # 'Л'
        38: 1,  # 'М'
        36: 0,  # 'Н'
        41: 1,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 1,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 0,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 0,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    31: {  # 'А'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 2,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 2,  # 'Е'
        55: 1,  # 'Ж'
        47: 2,  # 'З'
        40: 1,  # 'И'
        59: 1,  # 'Й'
        33: 1,  # 'К'
        46: 2,  # 'Л'
        38: 1,  # 'М'
        36: 2,  # 'Н'
        41: 1,  # 'О'
        30: 2,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 2,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 2,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 1,  # 'а'
        18: 2,  # 'б'
        9: 2,  # 'в'
        20: 2,  # 'г'
        11: 2,  # 'д'
        3: 1,  # 'е'
        23: 1,  # 'ж'
        15: 2,  # 'з'
        2: 0,  # 'и'
        26: 2,  # 'й'
        12: 2,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 0,  # 'о'
        13: 2,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 1,  # 'у'
        29: 2,  # 'ф'
        25: 1,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    32: {  # 'Б'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 2,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 1,  # 'Е'
        55: 1,  # 'Ж'
        47: 2,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 2,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 2,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 0,  # 'Ш'
        57: 1,  # 'Щ'
        61: 2,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 2,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 1,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    35: {  # 'В'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 2,  # 'Ф'
        49: 0,  # 'Х'
        53: 1,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 2,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 2,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 2,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    43: {  # 'Г'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 0,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 1,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 1,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    37: {  # 'Д'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 2,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 2,  # 'Е'
        55: 2,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 2,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 2,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 2,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    44: {  # 'Е'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 2,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 1,  # 'Й'
        33: 2,  # 'К'
        46: 2,  # 'Л'
        38: 1,  # 'М'
        36: 2,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 2,  # 'Ф'
        49: 1,  # 'Х'
        53: 2,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 1,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 0,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 2,  # 'д'
        3: 0,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 0,  # 'и'
        26: 1,  # 'й'
        12: 2,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 2,  # 'н'
        4: 0,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 1,  # 'т'
        19: 1,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    55: {  # 'Ж'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    47: {  # 'З'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 2,  # 'Н'
        41: 1,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 2,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 1,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 1,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    40: {  # 'И'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 1,  # 'Ж'
        47: 2,  # 'З'
        40: 1,  # 'И'
        59: 1,  # 'Й'
        33: 2,  # 'К'
        46: 2,  # 'Л'
        38: 2,  # 'М'
        36: 2,  # 'Н'
        41: 1,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 0,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 1,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 2,  # 'Я'
        1: 1,  # 'а'
        18: 1,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 1,  # 'д'
        3: 1,  # 'е'
        23: 0,  # 'ж'
        15: 3,  # 'з'
        2: 0,  # 'и'
        26: 1,  # 'й'
        12: 1,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 2,  # 'н'
        4: 0,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 0,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    59: {  # 'Й'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 1,  # 'С'
        34: 1,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 1,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 1,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    33: {  # 'К'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 2,  # 'Н'
        41: 2,  # 'О'
        30: 2,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 1,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 2,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 3,  # 'р'
        8: 1,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    46: {  # 'Л'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 2,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 0,  # 'Р'
        28: 1,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    38: {  # 'М'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 2,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 1,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 2,  # 'л'
        14: 0,  # 'м'
        6: 2,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    36: {  # 'Н'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 2,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 2,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 1,  # 'Й'
        33: 2,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 1,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 1,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 2,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    41: {  # 'О'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 2,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 1,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 1,  # 'Й'
        33: 2,  # 'К'
        46: 2,  # 'Л'
        38: 2,  # 'М'
        36: 2,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 0,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 1,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 1,  # 'а'
        18: 2,  # 'б'
        9: 2,  # 'в'
        20: 2,  # 'г'
        11: 1,  # 'д'
        3: 1,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 0,  # 'и'
        26: 1,  # 'й'
        12: 2,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 0,  # 'о'
        13: 2,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 3,  # 'т'
        19: 1,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 1,  # 'ц'
        21: 2,  # 'ч'
        27: 0,  # 'ш'
        24: 2,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    30: {  # 'П'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 2,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 2,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 3,  # 'л'
        14: 0,  # 'м'
        6: 1,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 3,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 2,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    39: {  # 'Р'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 2,  # 'Г'
        37: 2,  # 'Д'
        44: 2,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 0,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 2,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 1,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 1,  # 'с'
        5: 0,  # 'т'
        19: 3,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    28: {  # 'С'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 3,  # 'А'
        32: 2,  # 'Б'
        35: 2,  # 'В'
        43: 1,  # 'Г'
        37: 2,  # 'Д'
        44: 2,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 2,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 2,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 2,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 1,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 2,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 1,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 3,  # 'т'
        19: 2,  # 'у'
        29: 2,  # 'ф'
        25: 1,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    34: {  # 'Т'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 2,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 2,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 2,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 1,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 1,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 1,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 1,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 3,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    51: {  # 'У'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 2,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 0,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 2,  # 'Т'
        51: 0,  # 'У'
        48: 1,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 1,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 2,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 2,  # 'и'
        26: 1,  # 'й'
        12: 2,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 2,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 2,  # 'с'
        5: 1,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    48: {  # 'Ф'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 2,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 1,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 2,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    49: {  # 'Х'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 1,  # 'П'
        39: 1,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 1,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    53: {  # 'Ц'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 2,  # 'И'
        59: 0,  # 'Й'
        33: 2,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 1,  # 'Р'
        28: 2,  # 'С'
        34: 0,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 2,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 1,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    50: {  # 'Ч'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 2,  # 'А'
        32: 1,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 1,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 2,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 1,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    54: {  # 'Ш'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 1,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 1,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 2,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 2,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 1,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    57: {  # 'Щ'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 1,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 1,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 1,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 1,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    61: {  # 'Ъ'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 1,  # 'Д'
        44: 0,  # 'Е'
        55: 1,  # 'Ж'
        47: 1,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 2,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 0,  # 'О'
        30: 1,  # 'П'
        39: 2,  # 'Р'
        28: 1,  # 'С'
        34: 1,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 1,  # 'Х'
        53: 1,  # 'Ц'
        50: 1,  # 'Ч'
        54: 1,  # 'Ш'
        57: 1,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 0,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 1,  # 'л'
        14: 0,  # 'м'
        6: 1,  # 'н'
        4: 0,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    60: {  # 'Ю'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 1,  # 'Б'
        35: 0,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 0,  # 'Е'
        55: 1,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 0,  # 'М'
        36: 1,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 1,  # 'Р'
        28: 1,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 2,  # 'г'
        11: 1,  # 'д'
        3: 0,  # 'е'
        23: 2,  # 'ж'
        15: 1,  # 'з'
        2: 1,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 0,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    56: {  # 'Я'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 1,  # 'Б'
        35: 1,  # 'В'
        43: 1,  # 'Г'
        37: 1,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 1,  # 'Л'
        38: 1,  # 'М'
        36: 1,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 1,  # 'С'
        34: 2,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 0,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 1,  # 'и'
        26: 1,  # 'й'
        12: 1,  # 'к'
        10: 1,  # 'л'
        14: 2,  # 'м'
        6: 2,  # 'н'
        4: 0,  # 'о'
        13: 2,  # 'п'
        7: 1,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    1: {  # 'а'
        63: 1,  # 'e'
        45: 1,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 1,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 1,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 3,  # 'и'
        26: 3,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 3,  # 'ф'
        25: 3,  # 'х'
        22: 3,  # 'ц'
        21: 3,  # 'ч'
        27: 3,  # 'ш'
        24: 3,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    18: {  # 'б'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 3,  # 'в'
        20: 1,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 0,  # 'т'
        19: 3,  # 'у'
        29: 0,  # 'ф'
        25: 2,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 3,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    9: {  # 'в'
        63: 1,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 1,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 0,  # 'в'
        20: 2,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 3,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 2,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 3,  # 'ч'
        27: 2,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    20: {  # 'г'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 3,  # 'л'
        14: 1,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 3,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    11: {  # 'д'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 2,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 1,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    3: {  # 'е'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 2,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 2,  # 'и'
        26: 3,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 2,  # 'у'
        29: 3,  # 'ф'
        25: 3,  # 'х'
        22: 3,  # 'ц'
        21: 3,  # 'ч'
        27: 3,  # 'ш'
        24: 3,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    23: {  # 'ж'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 2,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    15: {  # 'з'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 2,  # 'ш'
        24: 1,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    2: {  # 'и'
        63: 1,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 1,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 1,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 1,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 1,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 3,  # 'и'
        26: 3,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 2,  # 'у'
        29: 3,  # 'ф'
        25: 3,  # 'х'
        22: 3,  # 'ц'
        21: 3,  # 'ч'
        27: 3,  # 'ш'
        24: 3,  # 'щ'
        17: 2,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    26: {  # 'й'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 1,  # 'а'
        18: 2,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 2,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 2,  # 'з'
        2: 1,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 1,  # 'у'
        29: 2,  # 'ф'
        25: 1,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    12: {  # 'к'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 1,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 1,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 3,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    10: {  # 'л'
        63: 1,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 1,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 1,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 2,  # 'п'
        7: 2,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 2,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 2,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 2,  # 'ь'
        42: 3,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    14: {  # 'м'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 1,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 2,  # 'к'
        10: 3,  # 'л'
        14: 1,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 1,  # 'т'
        19: 3,  # 'у'
        29: 2,  # 'ф'
        25: 1,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 2,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    6: {  # 'н'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 1,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 2,  # 'б'
        9: 2,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 2,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 3,  # 'ф'
        25: 2,  # 'х'
        22: 3,  # 'ц'
        21: 3,  # 'ч'
        27: 2,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 2,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    4: {  # 'о'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 2,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 3,  # 'и'
        26: 3,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 2,  # 'у'
        29: 3,  # 'ф'
        25: 3,  # 'х'
        22: 3,  # 'ц'
        21: 3,  # 'ч'
        27: 3,  # 'ш'
        24: 3,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    13: {  # 'п'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 1,  # 'й'
        12: 2,  # 'к'
        10: 3,  # 'л'
        14: 1,  # 'м'
        6: 2,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 3,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    7: {  # 'р'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 3,  # 'е'
        23: 3,  # 'ж'
        15: 2,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 2,  # 'п'
        7: 1,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 2,  # 'ф'
        25: 3,  # 'х'
        22: 3,  # 'ц'
        21: 2,  # 'ч'
        27: 3,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 1,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    8: {  # 'с'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 2,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 1,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 2,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 2,  # 'ш'
        24: 0,  # 'щ'
        17: 3,  # 'ъ'
        52: 2,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    5: {  # 'т'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 2,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 2,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 3,  # 'у'
        29: 1,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 2,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 3,  # 'ъ'
        52: 2,  # 'ь'
        42: 2,  # 'ю'
        16: 3,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    19: {  # 'у'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 2,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 2,  # 'и'
        26: 2,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 1,  # 'у'
        29: 2,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 3,  # 'ч'
        27: 3,  # 'ш'
        24: 2,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    29: {  # 'ф'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 1,  # 'в'
        20: 1,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 2,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 2,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 2,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    25: {  # 'х'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 3,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 2,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 1,  # 'п'
        7: 3,  # 'р'
        8: 1,  # 'с'
        5: 2,  # 'т'
        19: 3,  # 'у'
        29: 0,  # 'ф'
        25: 1,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    22: {  # 'ц'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 2,  # 'в'
        20: 1,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 1,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 2,  # 'к'
        10: 1,  # 'л'
        14: 1,  # 'м'
        6: 1,  # 'н'
        4: 2,  # 'о'
        13: 1,  # 'п'
        7: 1,  # 'р'
        8: 1,  # 'с'
        5: 1,  # 'т'
        19: 2,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 1,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 0,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    21: {  # 'ч'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 1,  # 'б'
        9: 3,  # 'в'
        20: 1,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 1,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 2,  # 'р'
        8: 0,  # 'с'
        5: 2,  # 'т'
        19: 3,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 1,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    27: {  # 'ш'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 2,  # 'в'
        20: 0,  # 'г'
        11: 1,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 3,  # 'к'
        10: 2,  # 'л'
        14: 1,  # 'м'
        6: 3,  # 'н'
        4: 2,  # 'о'
        13: 2,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 1,  # 'т'
        19: 2,  # 'у'
        29: 1,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 1,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 2,  # 'ъ'
        52: 1,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    24: {  # 'щ'
        63: 1,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 3,  # 'а'
        18: 0,  # 'б'
        9: 1,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 3,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 3,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 2,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 1,  # 'р'
        8: 0,  # 'с'
        5: 2,  # 'т'
        19: 3,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 1,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 2,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    17: {  # 'ъ'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 1,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 3,  # 'г'
        11: 3,  # 'д'
        3: 2,  # 'е'
        23: 3,  # 'ж'
        15: 3,  # 'з'
        2: 1,  # 'и'
        26: 2,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 3,  # 'о'
        13: 3,  # 'п'
        7: 3,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 1,  # 'у'
        29: 1,  # 'ф'
        25: 2,  # 'х'
        22: 2,  # 'ц'
        21: 3,  # 'ч'
        27: 2,  # 'ш'
        24: 3,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 2,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    52: {  # 'ь'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 1,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 1,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 1,  # 'н'
        4: 3,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 1,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 1,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 1,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    42: {  # 'ю'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 1,  # 'а'
        18: 2,  # 'б'
        9: 1,  # 'в'
        20: 2,  # 'г'
        11: 2,  # 'д'
        3: 1,  # 'е'
        23: 2,  # 'ж'
        15: 2,  # 'з'
        2: 1,  # 'и'
        26: 1,  # 'й'
        12: 2,  # 'к'
        10: 2,  # 'л'
        14: 2,  # 'м'
        6: 2,  # 'н'
        4: 1,  # 'о'
        13: 1,  # 'п'
        7: 2,  # 'р'
        8: 2,  # 'с'
        5: 2,  # 'т'
        19: 1,  # 'у'
        29: 1,  # 'ф'
        25: 1,  # 'х'
        22: 2,  # 'ц'
        21: 3,  # 'ч'
        27: 1,  # 'ш'
        24: 1,  # 'щ'
        17: 1,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    16: {  # 'я'
        63: 0,  # 'e'
        45: 1,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 3,  # 'б'
        9: 3,  # 'в'
        20: 2,  # 'г'
        11: 3,  # 'д'
        3: 2,  # 'е'
        23: 1,  # 'ж'
        15: 2,  # 'з'
        2: 1,  # 'и'
        26: 2,  # 'й'
        12: 3,  # 'к'
        10: 3,  # 'л'
        14: 3,  # 'м'
        6: 3,  # 'н'
        4: 1,  # 'о'
        13: 2,  # 'п'
        7: 2,  # 'р'
        8: 3,  # 'с'
        5: 3,  # 'т'
        19: 1,  # 'у'
        29: 1,  # 'ф'
        25: 3,  # 'х'
        22: 2,  # 'ц'
        21: 1,  # 'ч'
        27: 1,  # 'ш'
        24: 2,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 1,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    58: {  # 'є'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 0,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 0,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
    62: {  # '№'
        63: 0,  # 'e'
        45: 0,  # '\xad'
        31: 0,  # 'А'
        32: 0,  # 'Б'
        35: 0,  # 'В'
        43: 0,  # 'Г'
        37: 0,  # 'Д'
        44: 0,  # 'Е'
        55: 0,  # 'Ж'
        47: 0,  # 'З'
        40: 0,  # 'И'
        59: 0,  # 'Й'
        33: 0,  # 'К'
        46: 0,  # 'Л'
        38: 0,  # 'М'
        36: 0,  # 'Н'
        41: 0,  # 'О'
        30: 0,  # 'П'
        39: 0,  # 'Р'
        28: 0,  # 'С'
        34: 0,  # 'Т'
        51: 0,  # 'У'
        48: 0,  # 'Ф'
        49: 0,  # 'Х'
        53: 0,  # 'Ц'
        50: 0,  # 'Ч'
        54: 0,  # 'Ш'
        57: 0,  # 'Щ'
        61: 0,  # 'Ъ'
        60: 0,  # 'Ю'
        56: 0,  # 'Я'
        1: 0,  # 'а'
        18: 0,  # 'б'
        9: 0,  # 'в'
        20: 0,  # 'г'
        11: 0,  # 'д'
        3: 0,  # 'е'
        23: 0,  # 'ж'
        15: 0,  # 'з'
        2: 0,  # 'и'
        26: 0,  # 'й'
        12: 0,  # 'к'
        10: 0,  # 'л'
        14: 0,  # 'м'
        6: 0,  # 'н'
        4: 0,  # 'о'
        13: 0,  # 'п'
        7: 0,  # 'р'
        8: 0,  # 'с'
        5: 0,  # 'т'
        19: 0,  # 'у'
        29: 0,  # 'ф'
        25: 0,  # 'х'
        22: 0,  # 'ц'
        21: 0,  # 'ч'
        27: 0,  # 'ш'
        24: 0,  # 'щ'
        17: 0,  # 'ъ'
        52: 0,  # 'ь'
        42: 0,  # 'ю'
        16: 0,  # 'я'
        58: 0,  # 'є'
        62: 0,  # '№'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
ISO_8859_5_BULGARIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 77,  # 'A'
     66: 90,  # 'B'
     67: 99,  # 'C'
     68: 100,  # 'D'
     69: 72,  # 'E'
     70: 109,  # 'F'
     71: 107,  # 'G'
     72: 101,  # 'H'
     73: 79,  # 'I'
     74: 185,  # 'J'
     75: 81,  # 'K'
     76: 102,  # 'L'
     77: 76,  # 'M'
     78: 94,  # 'N'
     79: 82,  # 'O'
     80: 110,  # 'P'
     81: 186,  # 'Q'
     82: 108,  # 'R'
     83: 91,  # 'S'
     84: 74,  # 'T'
     85: 119,  # 'U'
     86: 84,  # 'V'
     87: 96,  # 'W'
     88: 111,  # 'X'
     89: 187,  # 'Y'
     90: 115,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 65,  # 'a'
     98: 69,  # 'b'
     99: 70,  # 'c'
     100: 66,  # 'd'
     101: 63,  # 'e'
     102: 68,  # 'f'
     103: 112,  # 'g'
     104: 103,  # 'h'
     105: 92,  # 'i'
     106: 194,  # 'j'
     107: 104,  # 'k'
     108: 95,  # 'l'
     109: 86,  # 'm'
     110: 87,  # 'n'
     111: 71,  # 'o'
     112: 116,  # 'p'
     113: 195,  # 'q'
     114: 85,  # 'r'
     115: 93,  # 's'
     116: 97,  # 't'
     117: 113,  # 'u'
     118: 196,  # 'v'
     119: 197,  # 'w'
     120: 198,  # 'x'
     121: 199,  # 'y'
     122: 200,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 194,  # '\x80'
     129: 195,  # '\x81'
     130: 196,  # '\x82'
     131: 197,  # '\x83'
     132: 198,  # '\x84'
     133: 199,  # '\x85'
     134: 200,  # '\x86'
     135: 201,  # '\x87'
     136: 202,  # '\x88'
     137: 203,  # '\x89'
     138: 204,  # '\x8a'
     139: 205,  # '\x8b'
     140: 206,  # '\x8c'
     141: 207,  # '\x8d'
     142: 208,  # '\x8e'
     143: 209,  # '\x8f'
     144: 210,  # '\x90'
     145: 211,  # '\x91'
     146: 212,  # '\x92'
     147: 213,  # '\x93'
     148: 214,  # '\x94'
     149: 215,  # '\x95'
     150: 216,  # '\x96'
     151: 217,  # '\x97'
     152: 218,  # '\x98'
     153: 219,  # '\x99'
     154: 220,  # '\x9a'
     155: 221,  # '\x9b'
     156: 222,  # '\x9c'
     157: 223,  # '\x9d'
     158: 224,  # '\x9e'
     159: 225,  # '\x9f'
     160: 81,  # '\xa0'
     161: 226,  # 'Ё'
     162: 227,  # 'Ђ'
     163: 228,  # 'Ѓ'
     164: 229,  # 'Є'
     165: 230,  # 'Ѕ'
     166: 105,  # 'І'
     167: 231,  # 'Ї'
     168: 232,  # 'Ј'
     169: 233,  # 'Љ'
     170: 234,  # 'Њ'
     171: 235,  # 'Ћ'
     172: 236,  # 'Ќ'
     173: 45,  # '\xad'
     174: 237,  # 'Ў'
     175: 238,  # 'Џ'
     176: 31,  # 'А'
     177: 32,  # 'Б'
     178: 35,  # 'В'
     179: 43,  # 'Г'
     180: 37,  # 'Д'
     181: 44,  # 'Е'
     182: 55,  # 'Ж'
     183: 47,  # 'З'
     184: 40,  # 'И'
     185: 59,  # 'Й'
     186: 33,  # 'К'
     187: 46,  # 'Л'
     188: 38,  # 'М'
     189: 36,  # 'Н'
     190: 41,  # 'О'
     191: 30,  # 'П'
     192: 39,  # 'Р'
     193: 28,  # 'С'
     194: 34,  # 'Т'
     195: 51,  # 'У'
     196: 48,  # 'Ф'
     197: 49,  # 'Х'
     198: 53,  # 'Ц'
     199: 50,  # 'Ч'
     200: 54,  # 'Ш'
     201: 57,  # 'Щ'
     202: 61,  # 'Ъ'
     203: 239,  # 'Ы'
     204: 67,  # 'Ь'
     205: 240,  # 'Э'
     206: 60,  # 'Ю'
     207: 56,  # 'Я'
     208: 1,  # 'а'
     209: 18,  # 'б'
     210: 9,  # 'в'
     211: 20,  # 'г'
     212: 11,  # 'д'
     213: 3,  # 'е'
     214: 23,  # 'ж'
     215: 15,  # 'з'
     216: 2,  # 'и'
     217: 26,  # 'й'
     218: 12,  # 'к'
     219: 10,  # 'л'
     220: 14,  # 'м'
     221: 6,  # 'н'
     222: 4,  # 'о'
     223: 13,  # 'п'
     224: 7,  # 'р'
     225: 8,  # 'с'
     226: 5,  # 'т'
     227: 19,  # 'у'
     228: 29,  # 'ф'
     229: 25,  # 'х'
     230: 22,  # 'ц'
     231: 21,  # 'ч'
     232: 27,  # 'ш'
     233: 24,  # 'щ'
     234: 17,  # 'ъ'
     235: 75,  # 'ы'
     236: 52,  # 'ь'
     237: 241,  # 'э'
     238: 42,  # 'ю'
     239: 16,  # 'я'
     240: 62,  # '№'
     241: 242,  # 'ё'
     242: 243,  # 'ђ'
     243: 244,  # 'ѓ'
     244: 58,  # 'є'
     245: 245,  # 'ѕ'
     246: 98,  # 'і'
     247: 246,  # 'ї'
     248: 247,  # 'ј'
     249: 248,  # 'љ'
     250: 249,  # 'њ'
     251: 250,  # 'ћ'
     252: 251,  # 'ќ'
     253: 91,  # '§'
     254: 252,  # 'ў'
     255: 253,  # 'џ'
}

ISO_8859_5_BULGARIAN_MODEL = SingleByteCharSetModel(charset_name='ISO-8859-5',
                                                    language='Bulgarian',
                                                    char_to_order_map=ISO_8859_5_BULGARIAN_CHAR_TO_ORDER,
                                                    language_model=BULGARIAN_LANG_MODEL,
                                                    typical_positive_ratio=0.969392,
                                                    keep_ascii_letters=False,
                                                    alphabet='АБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЬЮЯабвгдежзийклмнопрстуфхцчшщъьюя')

WINDOWS_1251_BULGARIAN_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 77,  # 'A'
     66: 90,  # 'B'
     67: 99,  # 'C'
     68: 100,  # 'D'
     69: 72,  # 'E'
     70: 109,  # 'F'
     71: 107,  # 'G'
     72: 101,  # 'H'
     73: 79,  # 'I'
     74: 185,  # 'J'
     75: 81,  # 'K'
     76: 102,  # 'L'
     77: 76,  # 'M'
     78: 94,  # 'N'
     79: 82,  # 'O'
     80: 110,  # 'P'
     81: 186,  # 'Q'
     82: 108,  # 'R'
     83: 91,  # 'S'
     84: 74,  # 'T'
     85: 119,  # 'U'
     86: 84,  # 'V'
     87: 96,  # 'W'
     88: 111,  # 'X'
     89: 187,  # 'Y'
     90: 115,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 65,  # 'a'
     98: 69,  # 'b'
     99: 70,  # 'c'
     100: 66,  # 'd'
     101: 63,  # 'e'
     102: 68,  # 'f'
     103: 112,  # 'g'
     104: 103,  # 'h'
     105: 92,  # 'i'
     106: 194,  # 'j'
     107: 104,  # 'k'
     108: 95,  # 'l'
     109: 86,  # 'm'
     110: 87,  # 'n'
     111: 71,  # 'o'
     112: 116,  # 'p'
     113: 195,  # 'q'
     114: 85,  # 'r'
     115: 93,  # 's'
     116: 97,  # 't'
     117: 113,  # 'u'
     118: 196,  # 'v'
     119: 197,  # 'w'
     120: 198,  # 'x'
     121: 199,  # 'y'
     122: 200,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 206,  # 'Ђ'
     129: 207,  # 'Ѓ'
     130: 208,  # '‚'
     131: 209,  # 'ѓ'
     132: 210,  # '„'
     133: 211,  # '…'
     134: 212,  # '†'
     135: 213,  # '‡'
     136: 120,  # '€'
     137: 214,  # '‰'
     138: 215,  # 'Љ'
     139: 216,  # '‹'
     140: 217,  # 'Њ'
     141: 218,  # 'Ќ'
     142: 219,  # 'Ћ'
     143: 220,  # 'Џ'
     144: 221,  # 'ђ'
     145: 78,  # '‘'
     146: 64,  # '’'
     147: 83,  # '“'
     148: 121,  # '”'
     149: 98,  # '•'
     150: 117,  # '–'
     151: 105,  # '—'
     152: 222,  # None
     153: 223,  # '™'
     154: 224,  # 'љ'
     155: 225,  # '›'
     156: 226,  # 'њ'
     157: 227,  # 'ќ'
     158: 228,  # 'ћ'
     159: 229,  # 'џ'
     160: 88,  # '\xa0'
     161: 230,  # 'Ў'
     162: 231,  # 'ў'
     163: 232,  # 'Ј'
     164: 233,  # '¤'
     165: 122,  # 'Ґ'
     166: 89,  # '¦'
     167: 106,  # '§'
     168: 234,  # 'Ё'
     169: 235,  # '©'
     170: 236,  # 'Є'
     171: 237,  # '«'
     172: 238,  # '¬'
     173: 45,  # '\xad'
     174: 239,  # '®'
     175: 240,  # 'Ї'
     176: 73,  # '°'
     177: 80,  # '±'
     178: 118,  # 'І'
     179: 114,  # 'і'
     180: 241,  # 'ґ'
     181: 242,  # 'µ'
     182: 243,  # '¶'
     183: 244,  # '·'
     184: 245,  # 'ё'
     185: 62,  # '№'
     186: 58,  # 'є'
     187: 246,  # '»'
     188: 247,  # 'ј'
     189: 248,  # 'Ѕ'
     190: 249,  # 'ѕ'
     191: 250,  # 'ї'
     192: 31,  # 'А'
     193: 32,  # 'Б'
     194: 35,  # 'В'
     195: 43,  # 'Г'
     196: 37,  # 'Д'
     197: 44,  # 'Е'
     198: 55,  # 'Ж'
     199: 47,  # 'З'
     200: 40,  # 'И'
     201: 59,  # 'Й'
     202: 33,  # 'К'
     203: 46,  # 'Л'
     204: 38,  # 'М'
     205: 36,  # 'Н'
     206: 41,  # 'О'
     207: 30,  # 'П'
     208: 39,  # 'Р'
     209: 28,  # 'С'
     210: 34,  # 'Т'
     211: 51,  # 'У'
     212: 48,  # 'Ф'
     213: 49,  # 'Х'
     214: 53,  # 'Ц'
     215: 50,  # 'Ч'
     216: 54,  # 'Ш'
     217: 57,  # 'Щ'
     218: 61,  # 'Ъ'
     219: 251,  # 'Ы'
     220: 67,  # 'Ь'
     221: 252,  # 'Э'
     222: 60,  # 'Ю'
     223: 56,  # 'Я'
     224: 1,  # 'а'
     225: 18,  # 'б'
     226: 9,  # 'в'
     227: 20,  # 'г'
     228: 11,  # 'д'
     229: 3,  # 'е'
     230: 23,  # 'ж'
     231: 15,  # 'з'
     232: 2,  # 'и'
     233: 26,  # 'й'
     234: 12,  # 'к'
     235: 10,  # 'л'
     236: 14,  # 'м'
     237: 6,  # 'н'
     238: 4,  # 'о'
     239: 13,  # 'п'
     240: 7,  # 'р'
     241: 8,  # 'с'
     242: 5,  # 'т'
     243: 19,  # 'у'
     244: 29,  # 'ф'
     245: 25,  # 'х'
     246: 22,  # 'ц'
     247: 21,  # 'ч'
     248: 27,  # 'ш'
     249: 24,  # 'щ'
     250: 17,  # 'ъ'
     251: 75,  # 'ы'
     252: 52,  # 'ь'
     253: 253,  # 'э'
     254: 42,  # 'ю'
     255: 16,  # 'я'
}

WINDOWS_1251_BULGARIAN_MODEL = SingleByteCharSetModel(charset_name='windows-1251',
                                                      language='Bulgarian',
                                                      char_to_order_map=WINDOWS_1251_BULGARIAN_CHAR_TO_ORDER,
                                                      language_model=BULGARIAN_LANG_MODEL,
                                                      typical_positive_ratio=0.969392,
                                                      keep_ascii_letters=False,
                                                      alphabet='АБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЬЮЯабвгдежзийклмнопрстуфхцчшщъьюя')

site-packages/pip/_vendor/chardet/sbcharsetprober.py000064400000013770150564745010016703 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from collections import namedtuple

from .charsetprober import CharSetProber
from .enums import CharacterCategory, ProbingState, SequenceLikelihood


SingleByteCharSetModel = namedtuple('SingleByteCharSetModel',
                                    ['charset_name',
                                     'language',
                                     'char_to_order_map',
                                     'language_model',
                                     'typical_positive_ratio',
                                     'keep_ascii_letters',
                                     'alphabet'])


class SingleByteCharSetProber(CharSetProber):
    SAMPLE_SIZE = 64
    SB_ENOUGH_REL_THRESHOLD = 1024  #  0.25 * SAMPLE_SIZE^2
    POSITIVE_SHORTCUT_THRESHOLD = 0.95
    NEGATIVE_SHORTCUT_THRESHOLD = 0.05

    def __init__(self, model, reversed=False, name_prober=None):
        super(SingleByteCharSetProber, self).__init__()
        self._model = model
        # TRUE if we need to reverse every pair in the model lookup
        self._reversed = reversed
        # Optional auxiliary prober for name decision
        self._name_prober = name_prober
        self._last_order = None
        self._seq_counters = None
        self._total_seqs = None
        self._total_char = None
        self._freq_char = None
        self.reset()

    def reset(self):
        super(SingleByteCharSetProber, self).reset()
        # char order of last character
        self._last_order = 255
        self._seq_counters = [0] * SequenceLikelihood.get_num_categories()
        self._total_seqs = 0
        self._total_char = 0
        # characters that fall in our sampling range
        self._freq_char = 0

    @property
    def charset_name(self):
        if self._name_prober:
            return self._name_prober.charset_name
        else:
            return self._model.charset_name

    @property
    def language(self):
        if self._name_prober:
            return self._name_prober.language
        else:
            return self._model.language

    def feed(self, byte_str):
        # TODO: Make filter_international_words keep things in self.alphabet
        if not self._model.keep_ascii_letters:
            byte_str = self.filter_international_words(byte_str)
        if not byte_str:
            return self.state
        char_to_order_map = self._model.char_to_order_map
        language_model = self._model.language_model
        for char in byte_str:
            order = char_to_order_map.get(char, CharacterCategory.UNDEFINED)
            # XXX: This was SYMBOL_CAT_ORDER before, with a value of 250, but
            #      CharacterCategory.SYMBOL is actually 253, so we use CONTROL
            #      to make it closer to the original intent. The only difference
            #      is whether or not we count digits and control characters for
            #      _total_char purposes.
            if order < CharacterCategory.CONTROL:
                self._total_char += 1
            # TODO: Follow uchardet's lead and discount confidence for frequent
            #       control characters.
            #       See https://github.com/BYVoid/uchardet/commit/55b4f23971db61
            if order < self.SAMPLE_SIZE:
                self._freq_char += 1
                if self._last_order < self.SAMPLE_SIZE:
                    self._total_seqs += 1
                    if not self._reversed:
                        lm_cat = language_model[self._last_order][order]
                    else:
                        lm_cat = language_model[order][self._last_order]
                    self._seq_counters[lm_cat] += 1
            self._last_order = order

        charset_name = self._model.charset_name
        if self.state == ProbingState.DETECTING:
            if self._total_seqs > self.SB_ENOUGH_REL_THRESHOLD:
                confidence = self.get_confidence()
                if confidence > self.POSITIVE_SHORTCUT_THRESHOLD:
                    self.logger.debug('%s confidence = %s, we have a winner',
                                      charset_name, confidence)
                    self._state = ProbingState.FOUND_IT
                elif confidence < self.NEGATIVE_SHORTCUT_THRESHOLD:
                    self.logger.debug('%s confidence = %s, below negative '
                                      'shortcut threshhold %s', charset_name,
                                      confidence,
                                      self.NEGATIVE_SHORTCUT_THRESHOLD)
                    self._state = ProbingState.NOT_ME

        return self.state

    def get_confidence(self):
        r = 0.01
        if self._total_seqs > 0:
            r = ((1.0 * self._seq_counters[SequenceLikelihood.POSITIVE]) /
                 self._total_seqs / self._model.typical_positive_ratio)
            r = r * self._freq_char / self._total_char
            if r >= 1.0:
                r = 0.99
        return r
site-packages/pip/_vendor/chardet/langgreekmodel.py000064400000302335150564745010016471 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

GREEK_LANG_MODEL = {
    60: {  # 'e'
        60: 2,  # 'e'
        55: 1,  # 'o'
        58: 2,  # 't'
        36: 1,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    55: {  # 'o'
        60: 0,  # 'e'
        55: 2,  # 'o'
        58: 2,  # 't'
        36: 1,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 1,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 1,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    58: {  # 't'
        60: 2,  # 'e'
        55: 1,  # 'o'
        58: 1,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 1,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    36: {  # '·'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    61: {  # 'Ά'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 1,  # 'γ'
        21: 2,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 1,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    46: {  # 'Έ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 2,  # 'β'
        20: 2,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 2,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 0,  # 'ο'
        9: 2,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 1,  # 'σ'
        2: 2,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    54: {  # 'Ό'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 2,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 2,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 2,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    31: {  # 'Α'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 2,  # 'Β'
        43: 2,  # 'Γ'
        41: 1,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 2,  # 'Θ'
        47: 2,  # 'Ι'
        44: 2,  # 'Κ'
        53: 2,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 1,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 2,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 2,  # 'Υ'
        56: 2,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 2,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 1,  # 'θ'
        5: 0,  # 'ι'
        11: 2,  # 'κ'
        16: 3,  # 'λ'
        10: 2,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 2,  # 'ς'
        7: 2,  # 'σ'
        2: 0,  # 'τ'
        12: 3,  # 'υ'
        28: 2,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    51: {  # 'Β'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 1,  # 'Ε'
        40: 1,  # 'Η'
        52: 0,  # 'Θ'
        47: 1,  # 'Ι'
        44: 0,  # 'Κ'
        53: 1,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 2,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    43: {  # 'Γ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 1,  # 'Α'
        51: 0,  # 'Β'
        43: 2,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 1,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 1,  # 'Κ'
        53: 1,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 1,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 2,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 2,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    41: {  # 'Δ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 2,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 2,  # 'ή'
        15: 2,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 1,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    34: {  # 'Ε'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 2,  # 'Γ'
        41: 2,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 2,  # 'Κ'
        53: 2,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 1,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 2,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 2,  # 'Υ'
        56: 0,  # 'Φ'
        50: 2,  # 'Χ'
        57: 2,  # 'Ω'
        17: 3,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 3,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 1,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 1,  # 'θ'
        5: 2,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 2,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 2,  # 'σ'
        2: 2,  # 'τ'
        12: 2,  # 'υ'
        28: 2,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 1,  # 'ύ'
        27: 0,  # 'ώ'
    },
    40: {  # 'Η'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 1,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 2,  # 'Θ'
        47: 0,  # 'Ι'
        44: 2,  # 'Κ'
        53: 0,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 2,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 1,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 1,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 1,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    52: {  # 'Θ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 1,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 1,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    47: {  # 'Ι'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 1,  # 'Β'
        43: 1,  # 'Γ'
        41: 2,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 2,  # 'Κ'
        53: 2,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 0,  # 'Υ'
        56: 2,  # 'Φ'
        50: 0,  # 'Χ'
        57: 2,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 2,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 1,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 2,  # 'σ'
        2: 1,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 1,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    44: {  # 'Κ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 1,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 1,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 1,  # 'Τ'
        45: 2,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 1,  # 'Ω'
        17: 3,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 2,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    53: {  # 'Λ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 2,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 0,  # 'Τ'
        45: 2,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 2,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 1,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 2,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    38: {  # 'Μ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 2,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 2,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 2,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 2,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 3,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 2,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    49: {  # 'Ν'
        60: 2,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 2,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 2,  # 'Ω'
        17: 0,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 1,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 1,  # 'ω'
        19: 2,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    59: {  # 'Ξ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 1,  # 'Ε'
        40: 1,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 1,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    39: {  # 'Ο'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 1,  # 'Β'
        43: 2,  # 'Γ'
        41: 2,  # 'Δ'
        34: 2,  # 'Ε'
        40: 1,  # 'Η'
        52: 2,  # 'Θ'
        47: 2,  # 'Ι'
        44: 2,  # 'Κ'
        53: 2,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 2,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 2,  # 'Υ'
        56: 2,  # 'Φ'
        50: 2,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 2,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 2,  # 'κ'
        16: 2,  # 'λ'
        10: 2,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 2,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 2,  # 'τ'
        12: 2,  # 'υ'
        28: 1,  # 'φ'
        23: 1,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    35: {  # 'Π'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 2,  # 'Λ'
        38: 1,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 1,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 2,  # 'Ω'
        17: 2,  # 'ά'
        18: 1,  # 'έ'
        22: 1,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 2,  # 'ό'
        26: 0,  # 'ύ'
        27: 3,  # 'ώ'
    },
    48: {  # 'Ρ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 1,  # 'Γ'
        41: 1,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 1,  # 'Τ'
        45: 1,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 1,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 2,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 1,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    37: {  # 'Σ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 1,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 2,  # 'Κ'
        53: 0,  # 'Λ'
        38: 2,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 2,  # 'Υ'
        56: 0,  # 'Φ'
        50: 2,  # 'Χ'
        57: 2,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 2,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 2,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 2,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 2,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    33: {  # 'Τ'
        60: 0,  # 'e'
        55: 1,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 2,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 2,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 1,  # 'Τ'
        45: 1,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 2,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 2,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 2,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 2,  # 'ό'
        26: 2,  # 'ύ'
        27: 3,  # 'ώ'
    },
    45: {  # 'Υ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 2,  # 'Γ'
        41: 0,  # 'Δ'
        34: 1,  # 'Ε'
        40: 2,  # 'Η'
        52: 2,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 1,  # 'Λ'
        38: 2,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 2,  # 'Π'
        48: 1,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    56: {  # 'Φ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 1,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 1,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 2,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 2,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 1,  # 'ύ'
        27: 1,  # 'ώ'
    },
    50: {  # 'Χ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 1,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 2,  # 'Ε'
        40: 2,  # 'Η'
        52: 0,  # 'Θ'
        47: 2,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 1,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 1,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 1,  # 'Χ'
        57: 1,  # 'Ω'
        17: 2,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 2,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 2,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    57: {  # 'Ω'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 1,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 1,  # 'Λ'
        38: 0,  # 'Μ'
        49: 2,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 2,  # 'Ρ'
        37: 2,  # 'Σ'
        33: 2,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 2,  # 'ρ'
        14: 2,  # 'ς'
        7: 2,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 1,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    17: {  # 'ά'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 3,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 2,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 3,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    18: {  # 'έ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 3,  # 'α'
        29: 2,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 3,  # 'ε'
        32: 2,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 3,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    22: {  # 'ή'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 1,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 2,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    15: {  # 'ί'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 3,  # 'α'
        29: 2,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 3,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 1,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 3,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    1: {  # 'α'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 2,  # 'έ'
        22: 0,  # 'ή'
        15: 3,  # 'ί'
        1: 0,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 2,  # 'ε'
        32: 3,  # 'ζ'
        13: 1,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 2,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 0,  # 'ω'
        19: 2,  # 'ό'
        26: 2,  # 'ύ'
        27: 0,  # 'ώ'
    },
    29: {  # 'β'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 2,  # 'έ'
        22: 3,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 2,  # 'γ'
        21: 2,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 3,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 2,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    20: {  # 'γ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 3,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 3,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 2,  # 'ύ'
        27: 3,  # 'ώ'
    },
    21: {  # 'δ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    3: {  # 'ε'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 3,  # 'ί'
        1: 2,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 2,  # 'ε'
        32: 2,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 2,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 3,  # 'ω'
        19: 2,  # 'ό'
        26: 3,  # 'ύ'
        27: 2,  # 'ώ'
    },
    32: {  # 'ζ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 2,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 1,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 2,  # 'ό'
        26: 0,  # 'ύ'
        27: 2,  # 'ώ'
    },
    13: {  # 'η'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 0,  # 'ο'
        9: 2,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    25: {  # 'θ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 1,  # 'λ'
        10: 3,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    5: {  # 'ι'
        60: 0,  # 'e'
        55: 1,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 1,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 0,  # 'ί'
        1: 3,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 2,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 0,  # 'ύ'
        27: 3,  # 'ώ'
    },
    11: {  # 'κ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 2,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 2,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 2,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    16: {  # 'λ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 1,  # 'β'
        20: 2,  # 'γ'
        21: 1,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 2,  # 'θ'
        5: 3,  # 'ι'
        11: 2,  # 'κ'
        16: 3,  # 'λ'
        10: 2,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 2,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    10: {  # 'μ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 1,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 3,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 2,  # 'υ'
        28: 3,  # 'φ'
        23: 0,  # 'χ'
        42: 2,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    6: {  # 'ν'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 2,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 1,  # 'λ'
        10: 0,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    30: {  # 'ξ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 2,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 2,  # 'ό'
        26: 3,  # 'ύ'
        27: 1,  # 'ώ'
    },
    4: {  # 'ο'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 2,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 2,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 2,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 2,  # 'ω'
        19: 1,  # 'ό'
        26: 3,  # 'ύ'
        27: 2,  # 'ώ'
    },
    9: {  # 'π'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 3,  # 'λ'
        10: 0,  # 'μ'
        6: 2,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 2,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 2,  # 'ύ'
        27: 3,  # 'ώ'
    },
    8: {  # 'ρ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 2,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 1,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 3,  # 'ο'
        9: 2,  # 'π'
        8: 2,  # 'ρ'
        14: 0,  # 'ς'
        7: 2,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    14: {  # 'ς'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 2,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 0,  # 'θ'
        5: 0,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 0,  # 'τ'
        12: 0,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    7: {  # 'σ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 3,  # 'β'
        20: 0,  # 'γ'
        21: 2,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 3,  # 'θ'
        5: 3,  # 'ι'
        11: 3,  # 'κ'
        16: 2,  # 'λ'
        10: 3,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 3,  # 'φ'
        23: 3,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 2,  # 'ώ'
    },
    2: {  # 'τ'
        60: 0,  # 'e'
        55: 2,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 2,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 3,  # 'ι'
        11: 2,  # 'κ'
        16: 2,  # 'λ'
        10: 3,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 2,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    12: {  # 'υ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 3,  # 'ή'
        15: 2,  # 'ί'
        1: 3,  # 'α'
        29: 2,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 2,  # 'ε'
        32: 2,  # 'ζ'
        13: 2,  # 'η'
        25: 3,  # 'θ'
        5: 2,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 3,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 2,  # 'ω'
        19: 2,  # 'ό'
        26: 0,  # 'ύ'
        27: 2,  # 'ώ'
    },
    28: {  # 'φ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 3,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 2,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 0,  # 'μ'
        6: 1,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 1,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 2,  # 'ύ'
        27: 2,  # 'ώ'
    },
    23: {  # 'χ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 3,  # 'ά'
        18: 2,  # 'έ'
        22: 3,  # 'ή'
        15: 3,  # 'ί'
        1: 3,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 2,  # 'θ'
        5: 3,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 2,  # 'μ'
        6: 3,  # 'ν'
        30: 0,  # 'ξ'
        4: 3,  # 'ο'
        9: 0,  # 'π'
        8: 3,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 3,  # 'τ'
        12: 3,  # 'υ'
        28: 0,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 3,  # 'ω'
        19: 3,  # 'ό'
        26: 3,  # 'ύ'
        27: 3,  # 'ώ'
    },
    42: {  # 'ψ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 2,  # 'ά'
        18: 2,  # 'έ'
        22: 1,  # 'ή'
        15: 2,  # 'ί'
        1: 2,  # 'α'
        29: 0,  # 'β'
        20: 0,  # 'γ'
        21: 0,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 3,  # 'η'
        25: 0,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 0,  # 'λ'
        10: 0,  # 'μ'
        6: 0,  # 'ν'
        30: 0,  # 'ξ'
        4: 2,  # 'ο'
        9: 0,  # 'π'
        8: 0,  # 'ρ'
        14: 0,  # 'ς'
        7: 0,  # 'σ'
        2: 2,  # 'τ'
        12: 1,  # 'υ'
        28: 0,  # 'φ'
        23: 0,  # 'χ'
        42: 0,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    24: {  # 'ω'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 1,  # 'ά'
        18: 0,  # 'έ'
        22: 2,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 2,  # 'β'
        20: 3,  # 'γ'
        21: 2,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 0,  # 'η'
        25: 3,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 0,  # 'ξ'
        4: 0,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 2,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    19: {  # 'ό'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 3,  # 'β'
        20: 3,  # 'γ'
        21: 3,  # 'δ'
        3: 1,  # 'ε'
        32: 2,  # 'ζ'
        13: 2,  # 'η'
        25: 2,  # 'θ'
        5: 2,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 1,  # 'ξ'
        4: 2,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 3,  # 'χ'
        42: 2,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    26: {  # 'ύ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 2,  # 'α'
        29: 2,  # 'β'
        20: 2,  # 'γ'
        21: 1,  # 'δ'
        3: 3,  # 'ε'
        32: 0,  # 'ζ'
        13: 2,  # 'η'
        25: 3,  # 'θ'
        5: 0,  # 'ι'
        11: 3,  # 'κ'
        16: 3,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 2,  # 'ξ'
        4: 3,  # 'ο'
        9: 3,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 2,  # 'φ'
        23: 2,  # 'χ'
        42: 2,  # 'ψ'
        24: 2,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
    27: {  # 'ώ'
        60: 0,  # 'e'
        55: 0,  # 'o'
        58: 0,  # 't'
        36: 0,  # '·'
        61: 0,  # 'Ά'
        46: 0,  # 'Έ'
        54: 0,  # 'Ό'
        31: 0,  # 'Α'
        51: 0,  # 'Β'
        43: 0,  # 'Γ'
        41: 0,  # 'Δ'
        34: 0,  # 'Ε'
        40: 0,  # 'Η'
        52: 0,  # 'Θ'
        47: 0,  # 'Ι'
        44: 0,  # 'Κ'
        53: 0,  # 'Λ'
        38: 0,  # 'Μ'
        49: 0,  # 'Ν'
        59: 0,  # 'Ξ'
        39: 0,  # 'Ο'
        35: 0,  # 'Π'
        48: 0,  # 'Ρ'
        37: 0,  # 'Σ'
        33: 0,  # 'Τ'
        45: 0,  # 'Υ'
        56: 0,  # 'Φ'
        50: 0,  # 'Χ'
        57: 0,  # 'Ω'
        17: 0,  # 'ά'
        18: 0,  # 'έ'
        22: 0,  # 'ή'
        15: 0,  # 'ί'
        1: 0,  # 'α'
        29: 1,  # 'β'
        20: 0,  # 'γ'
        21: 3,  # 'δ'
        3: 0,  # 'ε'
        32: 0,  # 'ζ'
        13: 1,  # 'η'
        25: 2,  # 'θ'
        5: 2,  # 'ι'
        11: 0,  # 'κ'
        16: 2,  # 'λ'
        10: 3,  # 'μ'
        6: 3,  # 'ν'
        30: 1,  # 'ξ'
        4: 0,  # 'ο'
        9: 2,  # 'π'
        8: 3,  # 'ρ'
        14: 3,  # 'ς'
        7: 3,  # 'σ'
        2: 3,  # 'τ'
        12: 0,  # 'υ'
        28: 1,  # 'φ'
        23: 1,  # 'χ'
        42: 0,  # 'ψ'
        24: 0,  # 'ω'
        19: 0,  # 'ό'
        26: 0,  # 'ύ'
        27: 0,  # 'ώ'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
WINDOWS_1253_GREEK_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 82,  # 'A'
     66: 100,  # 'B'
     67: 104,  # 'C'
     68: 94,  # 'D'
     69: 98,  # 'E'
     70: 101,  # 'F'
     71: 116,  # 'G'
     72: 102,  # 'H'
     73: 111,  # 'I'
     74: 187,  # 'J'
     75: 117,  # 'K'
     76: 92,  # 'L'
     77: 88,  # 'M'
     78: 113,  # 'N'
     79: 85,  # 'O'
     80: 79,  # 'P'
     81: 118,  # 'Q'
     82: 105,  # 'R'
     83: 83,  # 'S'
     84: 67,  # 'T'
     85: 114,  # 'U'
     86: 119,  # 'V'
     87: 95,  # 'W'
     88: 99,  # 'X'
     89: 109,  # 'Y'
     90: 188,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 72,  # 'a'
     98: 70,  # 'b'
     99: 80,  # 'c'
     100: 81,  # 'd'
     101: 60,  # 'e'
     102: 96,  # 'f'
     103: 93,  # 'g'
     104: 89,  # 'h'
     105: 68,  # 'i'
     106: 120,  # 'j'
     107: 97,  # 'k'
     108: 77,  # 'l'
     109: 86,  # 'm'
     110: 69,  # 'n'
     111: 55,  # 'o'
     112: 78,  # 'p'
     113: 115,  # 'q'
     114: 65,  # 'r'
     115: 66,  # 's'
     116: 58,  # 't'
     117: 76,  # 'u'
     118: 106,  # 'v'
     119: 103,  # 'w'
     120: 87,  # 'x'
     121: 107,  # 'y'
     122: 112,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 255,  # '€'
     129: 255,  # None
     130: 255,  # '‚'
     131: 255,  # 'ƒ'
     132: 255,  # '„'
     133: 255,  # '…'
     134: 255,  # '†'
     135: 255,  # '‡'
     136: 255,  # None
     137: 255,  # '‰'
     138: 255,  # None
     139: 255,  # '‹'
     140: 255,  # None
     141: 255,  # None
     142: 255,  # None
     143: 255,  # None
     144: 255,  # None
     145: 255,  # '‘'
     146: 255,  # '’'
     147: 255,  # '“'
     148: 255,  # '”'
     149: 255,  # '•'
     150: 255,  # '–'
     151: 255,  # '—'
     152: 255,  # None
     153: 255,  # '™'
     154: 255,  # None
     155: 255,  # '›'
     156: 255,  # None
     157: 255,  # None
     158: 255,  # None
     159: 255,  # None
     160: 253,  # '\xa0'
     161: 233,  # '΅'
     162: 61,  # 'Ά'
     163: 253,  # '£'
     164: 253,  # '¤'
     165: 253,  # '¥'
     166: 253,  # '¦'
     167: 253,  # '§'
     168: 253,  # '¨'
     169: 253,  # '©'
     170: 253,  # None
     171: 253,  # '«'
     172: 253,  # '¬'
     173: 74,  # '\xad'
     174: 253,  # '®'
     175: 253,  # '―'
     176: 253,  # '°'
     177: 253,  # '±'
     178: 253,  # '²'
     179: 253,  # '³'
     180: 247,  # '΄'
     181: 253,  # 'µ'
     182: 253,  # '¶'
     183: 36,  # '·'
     184: 46,  # 'Έ'
     185: 71,  # 'Ή'
     186: 73,  # 'Ί'
     187: 253,  # '»'
     188: 54,  # 'Ό'
     189: 253,  # '½'
     190: 108,  # 'Ύ'
     191: 123,  # 'Ώ'
     192: 110,  # 'ΐ'
     193: 31,  # 'Α'
     194: 51,  # 'Β'
     195: 43,  # 'Γ'
     196: 41,  # 'Δ'
     197: 34,  # 'Ε'
     198: 91,  # 'Ζ'
     199: 40,  # 'Η'
     200: 52,  # 'Θ'
     201: 47,  # 'Ι'
     202: 44,  # 'Κ'
     203: 53,  # 'Λ'
     204: 38,  # 'Μ'
     205: 49,  # 'Ν'
     206: 59,  # 'Ξ'
     207: 39,  # 'Ο'
     208: 35,  # 'Π'
     209: 48,  # 'Ρ'
     210: 250,  # None
     211: 37,  # 'Σ'
     212: 33,  # 'Τ'
     213: 45,  # 'Υ'
     214: 56,  # 'Φ'
     215: 50,  # 'Χ'
     216: 84,  # 'Ψ'
     217: 57,  # 'Ω'
     218: 120,  # 'Ϊ'
     219: 121,  # 'Ϋ'
     220: 17,  # 'ά'
     221: 18,  # 'έ'
     222: 22,  # 'ή'
     223: 15,  # 'ί'
     224: 124,  # 'ΰ'
     225: 1,  # 'α'
     226: 29,  # 'β'
     227: 20,  # 'γ'
     228: 21,  # 'δ'
     229: 3,  # 'ε'
     230: 32,  # 'ζ'
     231: 13,  # 'η'
     232: 25,  # 'θ'
     233: 5,  # 'ι'
     234: 11,  # 'κ'
     235: 16,  # 'λ'
     236: 10,  # 'μ'
     237: 6,  # 'ν'
     238: 30,  # 'ξ'
     239: 4,  # 'ο'
     240: 9,  # 'π'
     241: 8,  # 'ρ'
     242: 14,  # 'ς'
     243: 7,  # 'σ'
     244: 2,  # 'τ'
     245: 12,  # 'υ'
     246: 28,  # 'φ'
     247: 23,  # 'χ'
     248: 42,  # 'ψ'
     249: 24,  # 'ω'
     250: 64,  # 'ϊ'
     251: 75,  # 'ϋ'
     252: 19,  # 'ό'
     253: 26,  # 'ύ'
     254: 27,  # 'ώ'
     255: 253,  # None
}

WINDOWS_1253_GREEK_MODEL = SingleByteCharSetModel(charset_name='windows-1253',
                                                  language='Greek',
                                                  char_to_order_map=WINDOWS_1253_GREEK_CHAR_TO_ORDER,
                                                  language_model=GREEK_LANG_MODEL,
                                                  typical_positive_ratio=0.982851,
                                                  keep_ascii_letters=False,
                                                  alphabet='ΆΈΉΊΌΎΏΑΒΓΔΕΖΗΘΙΚΛΜΝΞΟΠΡΣΤΥΦΧΨΩάέήίαβγδεζηθικλμνξοπρςστυφχψωόύώ')

ISO_8859_7_GREEK_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 254,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 254,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 253,  # ' '
     33: 253,  # '!'
     34: 253,  # '"'
     35: 253,  # '#'
     36: 253,  # '$'
     37: 253,  # '%'
     38: 253,  # '&'
     39: 253,  # "'"
     40: 253,  # '('
     41: 253,  # ')'
     42: 253,  # '*'
     43: 253,  # '+'
     44: 253,  # ','
     45: 253,  # '-'
     46: 253,  # '.'
     47: 253,  # '/'
     48: 252,  # '0'
     49: 252,  # '1'
     50: 252,  # '2'
     51: 252,  # '3'
     52: 252,  # '4'
     53: 252,  # '5'
     54: 252,  # '6'
     55: 252,  # '7'
     56: 252,  # '8'
     57: 252,  # '9'
     58: 253,  # ':'
     59: 253,  # ';'
     60: 253,  # '<'
     61: 253,  # '='
     62: 253,  # '>'
     63: 253,  # '?'
     64: 253,  # '@'
     65: 82,  # 'A'
     66: 100,  # 'B'
     67: 104,  # 'C'
     68: 94,  # 'D'
     69: 98,  # 'E'
     70: 101,  # 'F'
     71: 116,  # 'G'
     72: 102,  # 'H'
     73: 111,  # 'I'
     74: 187,  # 'J'
     75: 117,  # 'K'
     76: 92,  # 'L'
     77: 88,  # 'M'
     78: 113,  # 'N'
     79: 85,  # 'O'
     80: 79,  # 'P'
     81: 118,  # 'Q'
     82: 105,  # 'R'
     83: 83,  # 'S'
     84: 67,  # 'T'
     85: 114,  # 'U'
     86: 119,  # 'V'
     87: 95,  # 'W'
     88: 99,  # 'X'
     89: 109,  # 'Y'
     90: 188,  # 'Z'
     91: 253,  # '['
     92: 253,  # '\\'
     93: 253,  # ']'
     94: 253,  # '^'
     95: 253,  # '_'
     96: 253,  # '`'
     97: 72,  # 'a'
     98: 70,  # 'b'
     99: 80,  # 'c'
     100: 81,  # 'd'
     101: 60,  # 'e'
     102: 96,  # 'f'
     103: 93,  # 'g'
     104: 89,  # 'h'
     105: 68,  # 'i'
     106: 120,  # 'j'
     107: 97,  # 'k'
     108: 77,  # 'l'
     109: 86,  # 'm'
     110: 69,  # 'n'
     111: 55,  # 'o'
     112: 78,  # 'p'
     113: 115,  # 'q'
     114: 65,  # 'r'
     115: 66,  # 's'
     116: 58,  # 't'
     117: 76,  # 'u'
     118: 106,  # 'v'
     119: 103,  # 'w'
     120: 87,  # 'x'
     121: 107,  # 'y'
     122: 112,  # 'z'
     123: 253,  # '{'
     124: 253,  # '|'
     125: 253,  # '}'
     126: 253,  # '~'
     127: 253,  # '\x7f'
     128: 255,  # '\x80'
     129: 255,  # '\x81'
     130: 255,  # '\x82'
     131: 255,  # '\x83'
     132: 255,  # '\x84'
     133: 255,  # '\x85'
     134: 255,  # '\x86'
     135: 255,  # '\x87'
     136: 255,  # '\x88'
     137: 255,  # '\x89'
     138: 255,  # '\x8a'
     139: 255,  # '\x8b'
     140: 255,  # '\x8c'
     141: 255,  # '\x8d'
     142: 255,  # '\x8e'
     143: 255,  # '\x8f'
     144: 255,  # '\x90'
     145: 255,  # '\x91'
     146: 255,  # '\x92'
     147: 255,  # '\x93'
     148: 255,  # '\x94'
     149: 255,  # '\x95'
     150: 255,  # '\x96'
     151: 255,  # '\x97'
     152: 255,  # '\x98'
     153: 255,  # '\x99'
     154: 255,  # '\x9a'
     155: 255,  # '\x9b'
     156: 255,  # '\x9c'
     157: 255,  # '\x9d'
     158: 255,  # '\x9e'
     159: 255,  # '\x9f'
     160: 253,  # '\xa0'
     161: 233,  # '‘'
     162: 90,  # '’'
     163: 253,  # '£'
     164: 253,  # '€'
     165: 253,  # '₯'
     166: 253,  # '¦'
     167: 253,  # '§'
     168: 253,  # '¨'
     169: 253,  # '©'
     170: 253,  # 'ͺ'
     171: 253,  # '«'
     172: 253,  # '¬'
     173: 74,  # '\xad'
     174: 253,  # None
     175: 253,  # '―'
     176: 253,  # '°'
     177: 253,  # '±'
     178: 253,  # '²'
     179: 253,  # '³'
     180: 247,  # '΄'
     181: 248,  # '΅'
     182: 61,  # 'Ά'
     183: 36,  # '·'
     184: 46,  # 'Έ'
     185: 71,  # 'Ή'
     186: 73,  # 'Ί'
     187: 253,  # '»'
     188: 54,  # 'Ό'
     189: 253,  # '½'
     190: 108,  # 'Ύ'
     191: 123,  # 'Ώ'
     192: 110,  # 'ΐ'
     193: 31,  # 'Α'
     194: 51,  # 'Β'
     195: 43,  # 'Γ'
     196: 41,  # 'Δ'
     197: 34,  # 'Ε'
     198: 91,  # 'Ζ'
     199: 40,  # 'Η'
     200: 52,  # 'Θ'
     201: 47,  # 'Ι'
     202: 44,  # 'Κ'
     203: 53,  # 'Λ'
     204: 38,  # 'Μ'
     205: 49,  # 'Ν'
     206: 59,  # 'Ξ'
     207: 39,  # 'Ο'
     208: 35,  # 'Π'
     209: 48,  # 'Ρ'
     210: 250,  # None
     211: 37,  # 'Σ'
     212: 33,  # 'Τ'
     213: 45,  # 'Υ'
     214: 56,  # 'Φ'
     215: 50,  # 'Χ'
     216: 84,  # 'Ψ'
     217: 57,  # 'Ω'
     218: 120,  # 'Ϊ'
     219: 121,  # 'Ϋ'
     220: 17,  # 'ά'
     221: 18,  # 'έ'
     222: 22,  # 'ή'
     223: 15,  # 'ί'
     224: 124,  # 'ΰ'
     225: 1,  # 'α'
     226: 29,  # 'β'
     227: 20,  # 'γ'
     228: 21,  # 'δ'
     229: 3,  # 'ε'
     230: 32,  # 'ζ'
     231: 13,  # 'η'
     232: 25,  # 'θ'
     233: 5,  # 'ι'
     234: 11,  # 'κ'
     235: 16,  # 'λ'
     236: 10,  # 'μ'
     237: 6,  # 'ν'
     238: 30,  # 'ξ'
     239: 4,  # 'ο'
     240: 9,  # 'π'
     241: 8,  # 'ρ'
     242: 14,  # 'ς'
     243: 7,  # 'σ'
     244: 2,  # 'τ'
     245: 12,  # 'υ'
     246: 28,  # 'φ'
     247: 23,  # 'χ'
     248: 42,  # 'ψ'
     249: 24,  # 'ω'
     250: 64,  # 'ϊ'
     251: 75,  # 'ϋ'
     252: 19,  # 'ό'
     253: 26,  # 'ύ'
     254: 27,  # 'ώ'
     255: 253,  # None
}

ISO_8859_7_GREEK_MODEL = SingleByteCharSetModel(charset_name='ISO-8859-7',
                                                language='Greek',
                                                char_to_order_map=ISO_8859_7_GREEK_CHAR_TO_ORDER,
                                                language_model=GREEK_LANG_MODEL,
                                                typical_positive_ratio=0.982851,
                                                keep_ascii_letters=False,
                                                alphabet='ΆΈΉΊΌΎΏΑΒΓΔΕΖΗΘΙΚΛΜΝΞΟΠΡΣΤΥΦΧΨΩάέήίαβγδεζηθικλμνξοπρςστυφχψωόύώ')

site-packages/pip/_vendor/chardet/hebrewprober.py000064400000033016150564745010016174 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
#          Shy Shalom
# Portions created by the Initial Developer are Copyright (C) 2005
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetprober import CharSetProber
from .enums import ProbingState

# This prober doesn't actually recognize a language or a charset.
# It is a helper prober for the use of the Hebrew model probers

### General ideas of the Hebrew charset recognition ###
#
# Four main charsets exist in Hebrew:
# "ISO-8859-8" - Visual Hebrew
# "windows-1255" - Logical Hebrew
# "ISO-8859-8-I" - Logical Hebrew
# "x-mac-hebrew" - ?? Logical Hebrew ??
#
# Both "ISO" charsets use a completely identical set of code points, whereas
# "windows-1255" and "x-mac-hebrew" are two different proper supersets of
# these code points. windows-1255 defines additional characters in the range
# 0x80-0x9F as some misc punctuation marks as well as some Hebrew-specific
# diacritics and additional 'Yiddish' ligature letters in the range 0xc0-0xd6.
# x-mac-hebrew defines similar additional code points but with a different
# mapping.
#
# As far as an average Hebrew text with no diacritics is concerned, all four
# charsets are identical with respect to code points. Meaning that for the
# main Hebrew alphabet, all four map the same values to all 27 Hebrew letters
# (including final letters).
#
# The dominant difference between these charsets is their directionality.
# "Visual" directionality means that the text is ordered as if the renderer is
# not aware of a BIDI rendering algorithm. The renderer sees the text and
# draws it from left to right. The text itself when ordered naturally is read
# backwards. A buffer of Visual Hebrew generally looks like so:
# "[last word of first line spelled backwards] [whole line ordered backwards
# and spelled backwards] [first word of first line spelled backwards]
# [end of line] [last word of second line] ... etc' "
# adding punctuation marks, numbers and English text to visual text is
# naturally also "visual" and from left to right.
#
# "Logical" directionality means the text is ordered "naturally" according to
# the order it is read. It is the responsibility of the renderer to display
# the text from right to left. A BIDI algorithm is used to place general
# punctuation marks, numbers and English text in the text.
#
# Texts in x-mac-hebrew are almost impossible to find on the Internet. From
# what little evidence I could find, it seems that its general directionality
# is Logical.
#
# To sum up all of the above, the Hebrew probing mechanism knows about two
# charsets:
# Visual Hebrew - "ISO-8859-8" - backwards text - Words and sentences are
#    backwards while line order is natural. For charset recognition purposes
#    the line order is unimportant (In fact, for this implementation, even
#    word order is unimportant).
# Logical Hebrew - "windows-1255" - normal, naturally ordered text.
#
# "ISO-8859-8-I" is a subset of windows-1255 and doesn't need to be
#    specifically identified.
# "x-mac-hebrew" is also identified as windows-1255. A text in x-mac-hebrew
#    that contain special punctuation marks or diacritics is displayed with
#    some unconverted characters showing as question marks. This problem might
#    be corrected using another model prober for x-mac-hebrew. Due to the fact
#    that x-mac-hebrew texts are so rare, writing another model prober isn't
#    worth the effort and performance hit.
#
#### The Prober ####
#
# The prober is divided between two SBCharSetProbers and a HebrewProber,
# all of which are managed, created, fed data, inquired and deleted by the
# SBCSGroupProber. The two SBCharSetProbers identify that the text is in
# fact some kind of Hebrew, Logical or Visual. The final decision about which
# one is it is made by the HebrewProber by combining final-letter scores
# with the scores of the two SBCharSetProbers to produce a final answer.
#
# The SBCSGroupProber is responsible for stripping the original text of HTML
# tags, English characters, numbers, low-ASCII punctuation characters, spaces
# and new lines. It reduces any sequence of such characters to a single space.
# The buffer fed to each prober in the SBCS group prober is pure text in
# high-ASCII.
# The two SBCharSetProbers (model probers) share the same language model:
# Win1255Model.
# The first SBCharSetProber uses the model normally as any other
# SBCharSetProber does, to recognize windows-1255, upon which this model was
# built. The second SBCharSetProber is told to make the pair-of-letter
# lookup in the language model backwards. This in practice exactly simulates
# a visual Hebrew model using the windows-1255 logical Hebrew model.
#
# The HebrewProber is not using any language model. All it does is look for
# final-letter evidence suggesting the text is either logical Hebrew or visual
# Hebrew. Disjointed from the model probers, the results of the HebrewProber
# alone are meaningless. HebrewProber always returns 0.00 as confidence
# since it never identifies a charset by itself. Instead, the pointer to the
# HebrewProber is passed to the model probers as a helper "Name Prober".
# When the Group prober receives a positive identification from any prober,
# it asks for the name of the charset identified. If the prober queried is a
# Hebrew model prober, the model prober forwards the call to the
# HebrewProber to make the final decision. In the HebrewProber, the
# decision is made according to the final-letters scores maintained and Both
# model probers scores. The answer is returned in the form of the name of the
# charset identified, either "windows-1255" or "ISO-8859-8".

class HebrewProber(CharSetProber):
    # windows-1255 / ISO-8859-8 code points of interest
    FINAL_KAF = 0xea
    NORMAL_KAF = 0xeb
    FINAL_MEM = 0xed
    NORMAL_MEM = 0xee
    FINAL_NUN = 0xef
    NORMAL_NUN = 0xf0
    FINAL_PE = 0xf3
    NORMAL_PE = 0xf4
    FINAL_TSADI = 0xf5
    NORMAL_TSADI = 0xf6

    # Minimum Visual vs Logical final letter score difference.
    # If the difference is below this, don't rely solely on the final letter score
    # distance.
    MIN_FINAL_CHAR_DISTANCE = 5

    # Minimum Visual vs Logical model score difference.
    # If the difference is below this, don't rely at all on the model score
    # distance.
    MIN_MODEL_DISTANCE = 0.01

    VISUAL_HEBREW_NAME = "ISO-8859-8"
    LOGICAL_HEBREW_NAME = "windows-1255"

    def __init__(self):
        super(HebrewProber, self).__init__()
        self._final_char_logical_score = None
        self._final_char_visual_score = None
        self._prev = None
        self._before_prev = None
        self._logical_prober = None
        self._visual_prober = None
        self.reset()

    def reset(self):
        self._final_char_logical_score = 0
        self._final_char_visual_score = 0
        # The two last characters seen in the previous buffer,
        # mPrev and mBeforePrev are initialized to space in order to simulate
        # a word delimiter at the beginning of the data
        self._prev = ' '
        self._before_prev = ' '
        # These probers are owned by the group prober.

    def set_model_probers(self, logicalProber, visualProber):
        self._logical_prober = logicalProber
        self._visual_prober = visualProber

    def is_final(self, c):
        return c in [self.FINAL_KAF, self.FINAL_MEM, self.FINAL_NUN,
                     self.FINAL_PE, self.FINAL_TSADI]

    def is_non_final(self, c):
        # The normal Tsadi is not a good Non-Final letter due to words like
        # 'lechotet' (to chat) containing an apostrophe after the tsadi. This
        # apostrophe is converted to a space in FilterWithoutEnglishLetters
        # causing the Non-Final tsadi to appear at an end of a word even
        # though this is not the case in the original text.
        # The letters Pe and Kaf rarely display a related behavior of not being
        # a good Non-Final letter. Words like 'Pop', 'Winamp' and 'Mubarak'
        # for example legally end with a Non-Final Pe or Kaf. However, the
        # benefit of these letters as Non-Final letters outweighs the damage
        # since these words are quite rare.
        return c in [self.NORMAL_KAF, self.NORMAL_MEM,
                     self.NORMAL_NUN, self.NORMAL_PE]

    def feed(self, byte_str):
        # Final letter analysis for logical-visual decision.
        # Look for evidence that the received buffer is either logical Hebrew
        # or visual Hebrew.
        # The following cases are checked:
        # 1) A word longer than 1 letter, ending with a final letter. This is
        #    an indication that the text is laid out "naturally" since the
        #    final letter really appears at the end. +1 for logical score.
        # 2) A word longer than 1 letter, ending with a Non-Final letter. In
        #    normal Hebrew, words ending with Kaf, Mem, Nun, Pe or Tsadi,
        #    should not end with the Non-Final form of that letter. Exceptions
        #    to this rule are mentioned above in isNonFinal(). This is an
        #    indication that the text is laid out backwards. +1 for visual
        #    score
        # 3) A word longer than 1 letter, starting with a final letter. Final
        #    letters should not appear at the beginning of a word. This is an
        #    indication that the text is laid out backwards. +1 for visual
        #    score.
        #
        # The visual score and logical score are accumulated throughout the
        # text and are finally checked against each other in GetCharSetName().
        # No checking for final letters in the middle of words is done since
        # that case is not an indication for either Logical or Visual text.
        #
        # We automatically filter out all 7-bit characters (replace them with
        # spaces) so the word boundary detection works properly. [MAP]

        if self.state == ProbingState.NOT_ME:
            # Both model probers say it's not them. No reason to continue.
            return ProbingState.NOT_ME

        byte_str = self.filter_high_byte_only(byte_str)

        for cur in byte_str:
            if cur == ' ':
                # We stand on a space - a word just ended
                if self._before_prev != ' ':
                    # next-to-last char was not a space so self._prev is not a
                    # 1 letter word
                    if self.is_final(self._prev):
                        # case (1) [-2:not space][-1:final letter][cur:space]
                        self._final_char_logical_score += 1
                    elif self.is_non_final(self._prev):
                        # case (2) [-2:not space][-1:Non-Final letter][
                        #  cur:space]
                        self._final_char_visual_score += 1
            else:
                # Not standing on a space
                if ((self._before_prev == ' ') and
                        (self.is_final(self._prev)) and (cur != ' ')):
                    # case (3) [-2:space][-1:final letter][cur:not space]
                    self._final_char_visual_score += 1
            self._before_prev = self._prev
            self._prev = cur

        # Forever detecting, till the end or until both model probers return
        # ProbingState.NOT_ME (handled above)
        return ProbingState.DETECTING

    @property
    def charset_name(self):
        # Make the decision: is it Logical or Visual?
        # If the final letter score distance is dominant enough, rely on it.
        finalsub = self._final_char_logical_score - self._final_char_visual_score
        if finalsub >= self.MIN_FINAL_CHAR_DISTANCE:
            return self.LOGICAL_HEBREW_NAME
        if finalsub <= -self.MIN_FINAL_CHAR_DISTANCE:
            return self.VISUAL_HEBREW_NAME

        # It's not dominant enough, try to rely on the model scores instead.
        modelsub = (self._logical_prober.get_confidence()
                    - self._visual_prober.get_confidence())
        if modelsub > self.MIN_MODEL_DISTANCE:
            return self.LOGICAL_HEBREW_NAME
        if modelsub < -self.MIN_MODEL_DISTANCE:
            return self.VISUAL_HEBREW_NAME

        # Still no good, back to final letter distance, maybe it'll save the
        # day.
        if finalsub < 0.0:
            return self.VISUAL_HEBREW_NAME

        # (finalsub > 0 - Logical) or (don't know what to do) default to
        # Logical.
        return self.LOGICAL_HEBREW_NAME

    @property
    def language(self):
        return 'Hebrew'

    @property
    def state(self):
        # Remain active as long as any of the model probers are active.
        if (self._logical_prober.state == ProbingState.NOT_ME) and \
           (self._visual_prober.state == ProbingState.NOT_ME):
            return ProbingState.NOT_ME
        return ProbingState.DETECTING
site-packages/pip/_vendor/chardet/version.py000064400000000362150564745010015171 0ustar00"""
This module exists only to simplify retrieving the version number of chardet
from within setup.py and from chardet subpackages.

:author: Dan Blanchard (dan.blanchard@gmail.com)
"""

__version__ = "4.0.0"
VERSION = __version__.split('.')
site-packages/pip/_vendor/chardet/escprober.py000064400000007556150564745010015504 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetprober import CharSetProber
from .codingstatemachine import CodingStateMachine
from .enums import LanguageFilter, ProbingState, MachineState
from .escsm import (HZ_SM_MODEL, ISO2022CN_SM_MODEL, ISO2022JP_SM_MODEL,
                    ISO2022KR_SM_MODEL)


class EscCharSetProber(CharSetProber):
    """
    This CharSetProber uses a "code scheme" approach for detecting encodings,
    whereby easily recognizable escape or shift sequences are relied on to
    identify these encodings.
    """

    def __init__(self, lang_filter=None):
        super(EscCharSetProber, self).__init__(lang_filter=lang_filter)
        self.coding_sm = []
        if self.lang_filter & LanguageFilter.CHINESE_SIMPLIFIED:
            self.coding_sm.append(CodingStateMachine(HZ_SM_MODEL))
            self.coding_sm.append(CodingStateMachine(ISO2022CN_SM_MODEL))
        if self.lang_filter & LanguageFilter.JAPANESE:
            self.coding_sm.append(CodingStateMachine(ISO2022JP_SM_MODEL))
        if self.lang_filter & LanguageFilter.KOREAN:
            self.coding_sm.append(CodingStateMachine(ISO2022KR_SM_MODEL))
        self.active_sm_count = None
        self._detected_charset = None
        self._detected_language = None
        self._state = None
        self.reset()

    def reset(self):
        super(EscCharSetProber, self).reset()
        for coding_sm in self.coding_sm:
            if not coding_sm:
                continue
            coding_sm.active = True
            coding_sm.reset()
        self.active_sm_count = len(self.coding_sm)
        self._detected_charset = None
        self._detected_language = None

    @property
    def charset_name(self):
        return self._detected_charset

    @property
    def language(self):
        return self._detected_language

    def get_confidence(self):
        if self._detected_charset:
            return 0.99
        else:
            return 0.00

    def feed(self, byte_str):
        for c in byte_str:
            for coding_sm in self.coding_sm:
                if not coding_sm or not coding_sm.active:
                    continue
                coding_state = coding_sm.next_state(c)
                if coding_state == MachineState.ERROR:
                    coding_sm.active = False
                    self.active_sm_count -= 1
                    if self.active_sm_count <= 0:
                        self._state = ProbingState.NOT_ME
                        return self.state
                elif coding_state == MachineState.ITS_ME:
                    self._state = ProbingState.FOUND_IT
                    self._detected_charset = coding_sm.get_coding_state_machine()
                    self._detected_language = coding_sm.language
                    return self.state

        return self.state
site-packages/pip/_vendor/chardet/gb2312prober.py000064400000003332150564745010015616 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import GB2312DistributionAnalysis
from .mbcssm import GB2312_SM_MODEL

class GB2312Prober(MultiByteCharSetProber):
    def __init__(self):
        super(GB2312Prober, self).__init__()
        self.coding_sm = CodingStateMachine(GB2312_SM_MODEL)
        self.distribution_analyzer = GB2312DistributionAnalysis()
        self.reset()

    @property
    def charset_name(self):
        return "GB2312"

    @property
    def language(self):
        return "Chinese"
site-packages/pip/_vendor/chardet/enums.py000064400000003175150564745010014640 0ustar00"""
All of the Enums that are used throughout the chardet package.

:author: Dan Blanchard (dan.blanchard@gmail.com)
"""


class InputState(object):
    """
    This enum represents the different states a universal detector can be in.
    """
    PURE_ASCII = 0
    ESC_ASCII = 1
    HIGH_BYTE = 2


class LanguageFilter(object):
    """
    This enum represents the different language filters we can apply to a
    ``UniversalDetector``.
    """
    CHINESE_SIMPLIFIED = 0x01
    CHINESE_TRADITIONAL = 0x02
    JAPANESE = 0x04
    KOREAN = 0x08
    NON_CJK = 0x10
    ALL = 0x1F
    CHINESE = CHINESE_SIMPLIFIED | CHINESE_TRADITIONAL
    CJK = CHINESE | JAPANESE | KOREAN


class ProbingState(object):
    """
    This enum represents the different states a prober can be in.
    """
    DETECTING = 0
    FOUND_IT = 1
    NOT_ME = 2


class MachineState(object):
    """
    This enum represents the different states a state machine can be in.
    """
    START = 0
    ERROR = 1
    ITS_ME = 2


class SequenceLikelihood(object):
    """
    This enum represents the likelihood of a character following the previous one.
    """
    NEGATIVE = 0
    UNLIKELY = 1
    LIKELY = 2
    POSITIVE = 3

    @classmethod
    def get_num_categories(cls):
        """:returns: The number of likelihood categories in the enum."""
        return 4


class CharacterCategory(object):
    """
    This enum represents the different categories language models for
    ``SingleByteCharsetProber`` put characters into.

    Anything less than CONTROL is considered a letter.
    """
    UNDEFINED = 255
    LINE_BREAK = 254
    SYMBOL = 253
    DIGIT = 252
    CONTROL = 251
site-packages/pip/_vendor/chardet/chardistribution.py000064400000022303150564745010017060 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .euctwfreq import (EUCTW_CHAR_TO_FREQ_ORDER, EUCTW_TABLE_SIZE,
                        EUCTW_TYPICAL_DISTRIBUTION_RATIO)
from .euckrfreq import (EUCKR_CHAR_TO_FREQ_ORDER, EUCKR_TABLE_SIZE,
                        EUCKR_TYPICAL_DISTRIBUTION_RATIO)
from .gb2312freq import (GB2312_CHAR_TO_FREQ_ORDER, GB2312_TABLE_SIZE,
                         GB2312_TYPICAL_DISTRIBUTION_RATIO)
from .big5freq import (BIG5_CHAR_TO_FREQ_ORDER, BIG5_TABLE_SIZE,
                       BIG5_TYPICAL_DISTRIBUTION_RATIO)
from .jisfreq import (JIS_CHAR_TO_FREQ_ORDER, JIS_TABLE_SIZE,
                      JIS_TYPICAL_DISTRIBUTION_RATIO)


class CharDistributionAnalysis(object):
    ENOUGH_DATA_THRESHOLD = 1024
    SURE_YES = 0.99
    SURE_NO = 0.01
    MINIMUM_DATA_THRESHOLD = 3

    def __init__(self):
        # Mapping table to get frequency order from char order (get from
        # GetOrder())
        self._char_to_freq_order = None
        self._table_size = None  # Size of above table
        # This is a constant value which varies from language to language,
        # used in calculating confidence.  See
        # http://www.mozilla.org/projects/intl/UniversalCharsetDetection.html
        # for further detail.
        self.typical_distribution_ratio = None
        self._done = None
        self._total_chars = None
        self._freq_chars = None
        self.reset()

    def reset(self):
        """reset analyser, clear any state"""
        # If this flag is set to True, detection is done and conclusion has
        # been made
        self._done = False
        self._total_chars = 0  # Total characters encountered
        # The number of characters whose frequency order is less than 512
        self._freq_chars = 0

    def feed(self, char, char_len):
        """feed a character with known length"""
        if char_len == 2:
            # we only care about 2-bytes character in our distribution analysis
            order = self.get_order(char)
        else:
            order = -1
        if order >= 0:
            self._total_chars += 1
            # order is valid
            if order < self._table_size:
                if 512 > self._char_to_freq_order[order]:
                    self._freq_chars += 1

    def get_confidence(self):
        """return confidence based on existing data"""
        # if we didn't receive any character in our consideration range,
        # return negative answer
        if self._total_chars <= 0 or self._freq_chars <= self.MINIMUM_DATA_THRESHOLD:
            return self.SURE_NO

        if self._total_chars != self._freq_chars:
            r = (self._freq_chars / ((self._total_chars - self._freq_chars)
                 * self.typical_distribution_ratio))
            if r < self.SURE_YES:
                return r

        # normalize confidence (we don't want to be 100% sure)
        return self.SURE_YES

    def got_enough_data(self):
        # It is not necessary to receive all data to draw conclusion.
        # For charset detection, certain amount of data is enough
        return self._total_chars > self.ENOUGH_DATA_THRESHOLD

    def get_order(self, byte_str):
        # We do not handle characters based on the original encoding string,
        # but convert this encoding string to a number, here called order.
        # This allows multiple encodings of a language to share one frequency
        # table.
        return -1


class EUCTWDistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(EUCTWDistributionAnalysis, self).__init__()
        self._char_to_freq_order = EUCTW_CHAR_TO_FREQ_ORDER
        self._table_size = EUCTW_TABLE_SIZE
        self.typical_distribution_ratio = EUCTW_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for euc-TW encoding, we are interested
        #   first  byte range: 0xc4 -- 0xfe
        #   second byte range: 0xa1 -- 0xfe
        # no validation needed here. State machine has done that
        first_char = byte_str[0]
        if first_char >= 0xC4:
            return 94 * (first_char - 0xC4) + byte_str[1] - 0xA1
        else:
            return -1


class EUCKRDistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(EUCKRDistributionAnalysis, self).__init__()
        self._char_to_freq_order = EUCKR_CHAR_TO_FREQ_ORDER
        self._table_size = EUCKR_TABLE_SIZE
        self.typical_distribution_ratio = EUCKR_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for euc-KR encoding, we are interested
        #   first  byte range: 0xb0 -- 0xfe
        #   second byte range: 0xa1 -- 0xfe
        # no validation needed here. State machine has done that
        first_char = byte_str[0]
        if first_char >= 0xB0:
            return 94 * (first_char - 0xB0) + byte_str[1] - 0xA1
        else:
            return -1


class GB2312DistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(GB2312DistributionAnalysis, self).__init__()
        self._char_to_freq_order = GB2312_CHAR_TO_FREQ_ORDER
        self._table_size = GB2312_TABLE_SIZE
        self.typical_distribution_ratio = GB2312_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for GB2312 encoding, we are interested
        #  first  byte range: 0xb0 -- 0xfe
        #  second byte range: 0xa1 -- 0xfe
        # no validation needed here. State machine has done that
        first_char, second_char = byte_str[0], byte_str[1]
        if (first_char >= 0xB0) and (second_char >= 0xA1):
            return 94 * (first_char - 0xB0) + second_char - 0xA1
        else:
            return -1


class Big5DistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(Big5DistributionAnalysis, self).__init__()
        self._char_to_freq_order = BIG5_CHAR_TO_FREQ_ORDER
        self._table_size = BIG5_TABLE_SIZE
        self.typical_distribution_ratio = BIG5_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for big5 encoding, we are interested
        #   first  byte range: 0xa4 -- 0xfe
        #   second byte range: 0x40 -- 0x7e , 0xa1 -- 0xfe
        # no validation needed here. State machine has done that
        first_char, second_char = byte_str[0], byte_str[1]
        if first_char >= 0xA4:
            if second_char >= 0xA1:
                return 157 * (first_char - 0xA4) + second_char - 0xA1 + 63
            else:
                return 157 * (first_char - 0xA4) + second_char - 0x40
        else:
            return -1


class SJISDistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(SJISDistributionAnalysis, self).__init__()
        self._char_to_freq_order = JIS_CHAR_TO_FREQ_ORDER
        self._table_size = JIS_TABLE_SIZE
        self.typical_distribution_ratio = JIS_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for sjis encoding, we are interested
        #   first  byte range: 0x81 -- 0x9f , 0xe0 -- 0xfe
        #   second byte range: 0x40 -- 0x7e,  0x81 -- oxfe
        # no validation needed here. State machine has done that
        first_char, second_char = byte_str[0], byte_str[1]
        if (first_char >= 0x81) and (first_char <= 0x9F):
            order = 188 * (first_char - 0x81)
        elif (first_char >= 0xE0) and (first_char <= 0xEF):
            order = 188 * (first_char - 0xE0 + 31)
        else:
            return -1
        order = order + second_char - 0x40
        if second_char > 0x7F:
            order = -1
        return order


class EUCJPDistributionAnalysis(CharDistributionAnalysis):
    def __init__(self):
        super(EUCJPDistributionAnalysis, self).__init__()
        self._char_to_freq_order = JIS_CHAR_TO_FREQ_ORDER
        self._table_size = JIS_TABLE_SIZE
        self.typical_distribution_ratio = JIS_TYPICAL_DISTRIBUTION_RATIO

    def get_order(self, byte_str):
        # for euc-JP encoding, we are interested
        #   first  byte range: 0xa0 -- 0xfe
        #   second byte range: 0xa1 -- 0xfe
        # no validation needed here. State machine has done that
        char = byte_str[0]
        if char >= 0xA0:
            return 94 * (char - 0xA1) + byte_str[1] - 0xa1
        else:
            return -1
site-packages/pip/_vendor/chardet/__init__.py000064400000006307150564745010015250 0ustar00######################## BEGIN LICENSE BLOCK ########################
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################


from .universaldetector import UniversalDetector
from .enums import InputState
from .version import __version__, VERSION


__all__ = ['UniversalDetector', 'detect', 'detect_all', '__version__', 'VERSION']


def detect(byte_str):
    """
    Detect the encoding of the given byte string.

    :param byte_str:     The byte sequence to examine.
    :type byte_str:      ``bytes`` or ``bytearray``
    """
    if not isinstance(byte_str, bytearray):
        if not isinstance(byte_str, bytes):
            raise TypeError('Expected object of type bytes or bytearray, got: '
                            '{}'.format(type(byte_str)))
        else:
            byte_str = bytearray(byte_str)
    detector = UniversalDetector()
    detector.feed(byte_str)
    return detector.close()


def detect_all(byte_str):
    """
    Detect all the possible encodings of the given byte string.

    :param byte_str:     The byte sequence to examine.
    :type byte_str:      ``bytes`` or ``bytearray``
    """
    if not isinstance(byte_str, bytearray):
        if not isinstance(byte_str, bytes):
            raise TypeError('Expected object of type bytes or bytearray, got: '
                            '{}'.format(type(byte_str)))
        else:
            byte_str = bytearray(byte_str)

    detector = UniversalDetector()
    detector.feed(byte_str)
    detector.close()

    if detector._input_state == InputState.HIGH_BYTE:
        results = []
        for prober in detector._charset_probers:
            if prober.get_confidence() > detector.MINIMUM_THRESHOLD:
                charset_name = prober.charset_name
                lower_charset_name = prober.charset_name.lower()
                # Use Windows encoding name instead of ISO-8859 if we saw any
                # extra Windows-specific bytes
                if lower_charset_name.startswith('iso-8859'):
                    if detector._has_win_bytes:
                        charset_name = detector.ISO_WIN_MAP.get(lower_charset_name,
                                                            charset_name)
                results.append({
                    'encoding': charset_name,
                    'confidence': prober.get_confidence(),
                    'language': prober.language,
                })
        if len(results) > 0:
            return sorted(results, key=lambda result: -result['confidence'])

    return [detector.result]
site-packages/pip/_vendor/chardet/utf8prober.py000064400000005316150564745010015610 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetprober import CharSetProber
from .enums import ProbingState, MachineState
from .codingstatemachine import CodingStateMachine
from .mbcssm import UTF8_SM_MODEL



class UTF8Prober(CharSetProber):
    ONE_CHAR_PROB = 0.5

    def __init__(self):
        super(UTF8Prober, self).__init__()
        self.coding_sm = CodingStateMachine(UTF8_SM_MODEL)
        self._num_mb_chars = None
        self.reset()

    def reset(self):
        super(UTF8Prober, self).reset()
        self.coding_sm.reset()
        self._num_mb_chars = 0

    @property
    def charset_name(self):
        return "utf-8"

    @property
    def language(self):
        return ""

    def feed(self, byte_str):
        for c in byte_str:
            coding_state = self.coding_sm.next_state(c)
            if coding_state == MachineState.ERROR:
                self._state = ProbingState.NOT_ME
                break
            elif coding_state == MachineState.ITS_ME:
                self._state = ProbingState.FOUND_IT
                break
            elif coding_state == MachineState.START:
                if self.coding_sm.get_current_charlen() >= 2:
                    self._num_mb_chars += 1

        if self.state == ProbingState.DETECTING:
            if self.get_confidence() > self.SHORTCUT_THRESHOLD:
                self._state = ProbingState.FOUND_IT

        return self.state

    def get_confidence(self):
        unlike = 0.99
        if self._num_mb_chars < 6:
            unlike *= self.ONE_CHAR_PROB ** self._num_mb_chars
            return 1.0 - unlike
        else:
            return unlike
site-packages/pip/_vendor/chardet/langturkishmodel.py000064400000273264150564745010017075 0ustar00# -*- coding: utf-8 -*-

from pip._vendor.chardet.sbcharsetprober import SingleByteCharSetModel


# 3: Positive
# 2: Likely
# 1: Unlikely
# 0: Negative

TURKISH_LANG_MODEL = {
    23: {  # 'A'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 1,  # 'i'
        24: 0,  # 'j'
        10: 2,  # 'k'
        5: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 1,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    37: {  # 'B'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 2,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
    47: {  # 'C'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 1,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 2,  # 'l'
        13: 2,  # 'm'
        4: 2,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 2,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 1,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    39: {  # 'D'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 1,  # 'l'
        13: 3,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 1,  # 'Ş'
        19: 0,  # 'ş'
    },
    29: {  # 'E'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 1,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 0,  # 'h'
        3: 1,  # 'i'
        24: 1,  # 'j'
        10: 0,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 1,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    52: {  # 'F'
        23: 0,  # 'A'
        37: 1,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 1,  # 'E'
        52: 2,  # 'F'
        36: 0,  # 'G'
        45: 2,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 1,  # 'b'
        28: 1,  # 'c'
        12: 1,  # 'd'
        2: 0,  # 'e'
        18: 1,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 2,  # 'i'
        24: 1,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 2,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 2,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 1,  # 'Ö'
        55: 2,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 2,  # 'ş'
    },
    36: {  # 'G'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 2,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 2,  # 'N'
        42: 1,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 1,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 1,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 0,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 1,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 2,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    45: {  # 'H'
        23: 0,  # 'A'
        37: 1,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 2,  # 'G'
        45: 1,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 1,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 2,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 2,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 1,  # 'p'
        7: 1,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 2,  # 'ğ'
        41: 1,  # 'İ'
        6: 0,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    53: {  # 'I'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 2,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
    60: {  # 'J'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 0,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 1,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 1,  # 's'
        9: 0,  # 't'
        14: 0,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    16: {  # 'K'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 1,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 0,  # 'u'
        32: 3,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 1,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    49: {  # 'L'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 2,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 2,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 0,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 2,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 2,  # 'n'
        15: 1,  # 'o'
        26: 1,  # 'p'
        7: 1,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 0,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 2,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 1,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    20: {  # 'M'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 2,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 0,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    46: {  # 'N'
        23: 0,  # 'A'
        37: 1,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 1,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 1,  # 'o'
        26: 1,  # 'p'
        7: 1,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 1,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 1,  # 'İ'
        6: 2,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
    42: {  # 'O'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 1,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 2,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 2,  # 'İ'
        6: 1,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
    48: {  # 'P'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 2,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 2,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 0,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    44: {  # 'R'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 1,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 2,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 1,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
    35: {  # 'S'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 1,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 1,  # 'l'
        13: 2,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 1,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 2,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    31: {  # 'T'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 3,  # 'e'
        18: 2,  # 'f'
        27: 2,  # 'g'
        25: 0,  # 'h'
        3: 1,  # 'i'
        24: 1,  # 'j'
        10: 2,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 2,  # 'r'
        8: 0,  # 's'
        9: 2,  # 't'
        14: 2,  # 'u'
        32: 1,  # 'v'
        57: 1,  # 'w'
        58: 1,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    51: {  # 'U'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 1,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 1,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    38: {  # 'V'
        23: 1,  # 'A'
        37: 1,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 2,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 1,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 1,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 3,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    62: {  # 'W'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 0,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 0,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 0,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    43: {  # 'Y'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 0,  # 'G'
        45: 1,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 2,  # 'N'
        42: 0,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 1,  # 'j'
        10: 1,  # 'k'
        5: 1,  # 'l'
        13: 3,  # 'm'
        4: 0,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 1,  # 'Ü'
        59: 1,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 0,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    56: {  # 'Z'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 2,  # 'Z'
        1: 2,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 2,  # 'i'
        24: 1,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 1,  # 'r'
        8: 1,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    1: {  # 'a'
        23: 3,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 3,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 2,  # 'Z'
        1: 2,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 2,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 3,  # 'v'
        57: 2,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 1,  # 'î'
        34: 1,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    21: {  # 'b'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 3,  # 'g'
        25: 1,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 3,  # 'p'
        7: 1,  # 'r'
        8: 2,  # 's'
        9: 2,  # 't'
        14: 2,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    28: {  # 'c'
        23: 0,  # 'A'
        37: 1,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 2,  # 'E'
        52: 0,  # 'F'
        36: 2,  # 'G'
        45: 2,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 2,  # 'T'
        51: 2,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 3,  # 'Y'
        56: 0,  # 'Z'
        1: 1,  # 'a'
        21: 1,  # 'b'
        28: 2,  # 'c'
        12: 2,  # 'd'
        2: 1,  # 'e'
        18: 1,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 1,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 2,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 1,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 1,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 1,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 1,  # 'î'
        34: 2,  # 'ö'
        17: 2,  # 'ü'
        30: 2,  # 'ğ'
        41: 1,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 2,  # 'ş'
    },
    12: {  # 'd'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 2,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 1,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 2,  # 'i'
        24: 3,  # 'j'
        10: 2,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 2,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 1,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    2: {  # 'e'
        23: 2,  # 'A'
        37: 0,  # 'B'
        47: 2,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 3,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 2,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 3,  # 'v'
        57: 2,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 1,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    18: {  # 'f'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 2,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 1,  # 'i'
        24: 1,  # 'j'
        10: 1,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 1,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 1,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    27: {  # 'g'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 1,  # 'h'
        3: 2,  # 'i'
        24: 3,  # 'j'
        10: 2,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 2,  # 'r'
        8: 2,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    25: {  # 'h'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 2,  # 'h'
        3: 2,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 1,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    3: {  # 'i'
        23: 2,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 1,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 2,  # 'f'
        27: 3,  # 'g'
        25: 1,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 1,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 1,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 1,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 1,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    24: {  # 'j'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 2,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 1,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 2,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 2,  # 'i'
        24: 1,  # 'j'
        10: 2,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 2,  # 'r'
        8: 3,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 2,  # 'x'
        11: 1,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    10: {  # 'k'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 3,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 3,  # 'e'
        18: 1,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 2,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 3,  # 'p'
        7: 2,  # 'r'
        8: 2,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 3,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    5: {  # 'l'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 1,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 1,  # 'l'
        13: 1,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 2,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    13: {  # 'm'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 3,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 2,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 2,  # 'u'
        32: 2,  # 'v'
        57: 1,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    4: {  # 'n'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 2,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 1,  # 'f'
        27: 2,  # 'g'
        25: 3,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 3,  # 'p'
        7: 2,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 2,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    15: {  # 'o'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 2,  # 'L'
        20: 0,  # 'M'
        46: 2,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 1,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 1,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 2,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 2,  # 'ğ'
        41: 2,  # 'İ'
        6: 3,  # 'ı'
        40: 2,  # 'Ş'
        19: 2,  # 'ş'
    },
    26: {  # 'p'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 2,  # 'i'
        24: 3,  # 'j'
        10: 1,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 2,  # 'r'
        8: 1,  # 's'
        9: 1,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 1,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    7: {  # 'r'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 1,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 2,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 1,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 3,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    8: {  # 's'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 2,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 2,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    9: {  # 't'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 2,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 3,  # 'v'
        57: 0,  # 'w'
        58: 2,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    14: {  # 'u'
        23: 3,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 2,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 3,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 2,  # 'Z'
        1: 2,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 2,  # 'e'
        18: 2,  # 'f'
        27: 3,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 2,  # 'v'
        57: 2,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 3,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    32: {  # 'v'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 1,  # 'j'
        10: 1,  # 'k'
        5: 3,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 1,  # 'r'
        8: 2,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    57: {  # 'w'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 1,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 1,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 1,  # 's'
        9: 0,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 2,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    58: {  # 'x'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 1,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 1,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 2,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 1,  # 'r'
        8: 2,  # 's'
        9: 1,  # 't'
        14: 0,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    11: {  # 'y'
        23: 1,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 2,  # 'i'
        24: 1,  # 'j'
        10: 2,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 2,  # 'r'
        8: 1,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 3,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    22: {  # 'z'
        23: 2,  # 'A'
        37: 2,  # 'B'
        47: 1,  # 'C'
        39: 2,  # 'D'
        29: 3,  # 'E'
        52: 1,  # 'F'
        36: 2,  # 'G'
        45: 2,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 2,  # 'N'
        42: 2,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 3,  # 'T'
        51: 2,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 1,  # 'Z'
        1: 1,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 2,  # 'd'
        2: 2,  # 'e'
        18: 3,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 2,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 0,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 3,  # 'y'
        22: 2,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 2,  # 'Ü'
        59: 1,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 2,  # 'ü'
        30: 2,  # 'ğ'
        41: 1,  # 'İ'
        6: 3,  # 'ı'
        40: 1,  # 'Ş'
        19: 2,  # 'ş'
    },
    63: {  # '·'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 1,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    54: {  # 'Ç'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 1,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 1,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 0,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 0,  # 'h'
        3: 3,  # 'i'
        24: 0,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 2,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 2,  # 'r'
        8: 0,  # 's'
        9: 1,  # 't'
        14: 0,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 2,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    50: {  # 'Ö'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 2,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 2,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 1,  # 'N'
        42: 2,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 2,  # 'd'
        2: 0,  # 'e'
        18: 1,  # 'f'
        27: 1,  # 'g'
        25: 1,  # 'h'
        3: 2,  # 'i'
        24: 0,  # 'j'
        10: 2,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 3,  # 'n'
        15: 2,  # 'o'
        26: 2,  # 'p'
        7: 3,  # 'r'
        8: 1,  # 's'
        9: 2,  # 't'
        14: 0,  # 'u'
        32: 1,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 2,  # 'ü'
        30: 1,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    55: {  # 'Ü'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 1,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 1,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 1,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 1,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 1,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 0,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    59: {  # 'â'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 0,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 2,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 2,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 1,  # 'Ş'
        19: 0,  # 'ş'
    },
    33: {  # 'ç'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 3,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 0,  # 'Z'
        1: 0,  # 'a'
        21: 3,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 0,  # 'e'
        18: 2,  # 'f'
        27: 1,  # 'g'
        25: 3,  # 'h'
        3: 3,  # 'i'
        24: 0,  # 'j'
        10: 3,  # 'k'
        5: 0,  # 'l'
        13: 0,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 2,  # 's'
        9: 3,  # 't'
        14: 0,  # 'u'
        32: 2,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 1,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    61: {  # 'î'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 0,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 0,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 2,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 1,  # 'j'
        10: 0,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 1,  # 'n'
        15: 0,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 1,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 1,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 1,  # 'î'
        34: 0,  # 'ö'
        17: 0,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 1,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    34: {  # 'ö'
        23: 0,  # 'A'
        37: 1,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 1,  # 'G'
        45: 1,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 1,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 2,  # 'c'
        12: 1,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 2,  # 'h'
        3: 1,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 2,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 0,  # 'r'
        8: 3,  # 's'
        9: 1,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 1,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 0,  # 'ü'
        30: 2,  # 'ğ'
        41: 1,  # 'İ'
        6: 1,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    17: {  # 'ü'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 0,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 1,  # 'J'
        16: 1,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 0,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 0,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 0,  # 'c'
        12: 1,  # 'd'
        2: 3,  # 'e'
        18: 1,  # 'f'
        27: 2,  # 'g'
        25: 0,  # 'h'
        3: 1,  # 'i'
        24: 1,  # 'j'
        10: 2,  # 'k'
        5: 3,  # 'l'
        13: 2,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 2,  # 'p'
        7: 2,  # 'r'
        8: 3,  # 's'
        9: 2,  # 't'
        14: 3,  # 'u'
        32: 1,  # 'v'
        57: 1,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 2,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    30: {  # 'ğ'
        23: 0,  # 'A'
        37: 2,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 1,  # 'M'
        46: 2,  # 'N'
        42: 2,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 0,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 2,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 0,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 2,  # 'e'
        18: 0,  # 'f'
        27: 0,  # 'g'
        25: 0,  # 'h'
        3: 0,  # 'i'
        24: 3,  # 'j'
        10: 1,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 0,  # 'n'
        15: 1,  # 'o'
        26: 0,  # 'p'
        7: 1,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 2,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 2,  # 'İ'
        6: 2,  # 'ı'
        40: 2,  # 'Ş'
        19: 1,  # 'ş'
    },
    41: {  # 'İ'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 1,  # 'E'
        52: 0,  # 'F'
        36: 2,  # 'G'
        45: 2,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 0,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 0,  # 'Z'
        1: 1,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 2,  # 'd'
        2: 1,  # 'e'
        18: 0,  # 'f'
        27: 3,  # 'g'
        25: 2,  # 'h'
        3: 2,  # 'i'
        24: 2,  # 'j'
        10: 2,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 3,  # 'n'
        15: 1,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 2,  # 't'
        14: 0,  # 'u'
        32: 0,  # 'v'
        57: 1,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 1,  # 'Ü'
        59: 1,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 1,  # 'ö'
        17: 1,  # 'ü'
        30: 2,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 1,  # 'ş'
    },
    6: {  # 'ı'
        23: 2,  # 'A'
        37: 0,  # 'B'
        47: 0,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 2,  # 'J'
        16: 3,  # 'K'
        49: 0,  # 'L'
        20: 3,  # 'M'
        46: 1,  # 'N'
        42: 0,  # 'O'
        48: 0,  # 'P'
        44: 0,  # 'R'
        35: 0,  # 'S'
        31: 2,  # 'T'
        51: 0,  # 'U'
        38: 0,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 1,  # 'Z'
        1: 3,  # 'a'
        21: 2,  # 'b'
        28: 1,  # 'c'
        12: 3,  # 'd'
        2: 3,  # 'e'
        18: 3,  # 'f'
        27: 3,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 3,  # 'j'
        10: 3,  # 'k'
        5: 3,  # 'l'
        13: 3,  # 'm'
        4: 3,  # 'n'
        15: 0,  # 'o'
        26: 3,  # 'p'
        7: 3,  # 'r'
        8: 3,  # 's'
        9: 3,  # 't'
        14: 3,  # 'u'
        32: 3,  # 'v'
        57: 1,  # 'w'
        58: 1,  # 'x'
        11: 3,  # 'y'
        22: 0,  # 'z'
        63: 1,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 2,  # 'ç'
        61: 0,  # 'î'
        34: 0,  # 'ö'
        17: 3,  # 'ü'
        30: 0,  # 'ğ'
        41: 0,  # 'İ'
        6: 3,  # 'ı'
        40: 0,  # 'Ş'
        19: 0,  # 'ş'
    },
    40: {  # 'Ş'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 1,  # 'D'
        29: 1,  # 'E'
        52: 0,  # 'F'
        36: 1,  # 'G'
        45: 2,  # 'H'
        53: 1,  # 'I'
        60: 0,  # 'J'
        16: 0,  # 'K'
        49: 0,  # 'L'
        20: 2,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 2,  # 'P'
        44: 2,  # 'R'
        35: 1,  # 'S'
        31: 1,  # 'T'
        51: 0,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 2,  # 'Y'
        56: 1,  # 'Z'
        1: 0,  # 'a'
        21: 2,  # 'b'
        28: 0,  # 'c'
        12: 2,  # 'd'
        2: 0,  # 'e'
        18: 3,  # 'f'
        27: 0,  # 'g'
        25: 2,  # 'h'
        3: 3,  # 'i'
        24: 2,  # 'j'
        10: 1,  # 'k'
        5: 0,  # 'l'
        13: 1,  # 'm'
        4: 3,  # 'n'
        15: 2,  # 'o'
        26: 0,  # 'p'
        7: 3,  # 'r'
        8: 2,  # 's'
        9: 2,  # 't'
        14: 1,  # 'u'
        32: 3,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 2,  # 'y'
        22: 0,  # 'z'
        63: 0,  # '·'
        54: 0,  # 'Ç'
        50: 0,  # 'Ö'
        55: 1,  # 'Ü'
        59: 0,  # 'â'
        33: 0,  # 'ç'
        61: 0,  # 'î'
        34: 2,  # 'ö'
        17: 1,  # 'ü'
        30: 2,  # 'ğ'
        41: 0,  # 'İ'
        6: 2,  # 'ı'
        40: 1,  # 'Ş'
        19: 2,  # 'ş'
    },
    19: {  # 'ş'
        23: 0,  # 'A'
        37: 0,  # 'B'
        47: 1,  # 'C'
        39: 0,  # 'D'
        29: 0,  # 'E'
        52: 2,  # 'F'
        36: 1,  # 'G'
        45: 0,  # 'H'
        53: 0,  # 'I'
        60: 0,  # 'J'
        16: 3,  # 'K'
        49: 2,  # 'L'
        20: 0,  # 'M'
        46: 1,  # 'N'
        42: 1,  # 'O'
        48: 1,  # 'P'
        44: 1,  # 'R'
        35: 1,  # 'S'
        31: 0,  # 'T'
        51: 1,  # 'U'
        38: 1,  # 'V'
        62: 0,  # 'W'
        43: 1,  # 'Y'
        56: 0,  # 'Z'
        1: 3,  # 'a'
        21: 1,  # 'b'
        28: 2,  # 'c'
        12: 0,  # 'd'
        2: 3,  # 'e'
        18: 0,  # 'f'
        27: 2,  # 'g'
        25: 1,  # 'h'
        3: 1,  # 'i'
        24: 0,  # 'j'
        10: 2,  # 'k'
        5: 2,  # 'l'
        13: 3,  # 'm'
        4: 0,  # 'n'
        15: 0,  # 'o'
        26: 1,  # 'p'
        7: 3,  # 'r'
        8: 0,  # 's'
        9: 0,  # 't'
        14: 3,  # 'u'
        32: 0,  # 'v'
        57: 0,  # 'w'
        58: 0,  # 'x'
        11: 0,  # 'y'
        22: 2,  # 'z'
        63: 0,  # '·'
        54: 1,  # 'Ç'
        50: 2,  # 'Ö'
        55: 0,  # 'Ü'
        59: 0,  # 'â'
        33: 1,  # 'ç'
        61: 1,  # 'î'
        34: 2,  # 'ö'
        17: 0,  # 'ü'
        30: 1,  # 'ğ'
        41: 1,  # 'İ'
        6: 1,  # 'ı'
        40: 1,  # 'Ş'
        19: 1,  # 'ş'
    },
}

# 255: Undefined characters that did not exist in training text
# 254: Carriage/Return
# 253: symbol (punctuation) that does not belong to word
# 252: 0 - 9
# 251: Control characters

# Character Mapping Table(s):
ISO_8859_9_TURKISH_CHAR_TO_ORDER = {
     0: 255,  # '\x00'
     1: 255,  # '\x01'
     2: 255,  # '\x02'
     3: 255,  # '\x03'
     4: 255,  # '\x04'
     5: 255,  # '\x05'
     6: 255,  # '\x06'
     7: 255,  # '\x07'
     8: 255,  # '\x08'
     9: 255,  # '\t'
     10: 255,  # '\n'
     11: 255,  # '\x0b'
     12: 255,  # '\x0c'
     13: 255,  # '\r'
     14: 255,  # '\x0e'
     15: 255,  # '\x0f'
     16: 255,  # '\x10'
     17: 255,  # '\x11'
     18: 255,  # '\x12'
     19: 255,  # '\x13'
     20: 255,  # '\x14'
     21: 255,  # '\x15'
     22: 255,  # '\x16'
     23: 255,  # '\x17'
     24: 255,  # '\x18'
     25: 255,  # '\x19'
     26: 255,  # '\x1a'
     27: 255,  # '\x1b'
     28: 255,  # '\x1c'
     29: 255,  # '\x1d'
     30: 255,  # '\x1e'
     31: 255,  # '\x1f'
     32: 255,  # ' '
     33: 255,  # '!'
     34: 255,  # '"'
     35: 255,  # '#'
     36: 255,  # '$'
     37: 255,  # '%'
     38: 255,  # '&'
     39: 255,  # "'"
     40: 255,  # '('
     41: 255,  # ')'
     42: 255,  # '*'
     43: 255,  # '+'
     44: 255,  # ','
     45: 255,  # '-'
     46: 255,  # '.'
     47: 255,  # '/'
     48: 255,  # '0'
     49: 255,  # '1'
     50: 255,  # '2'
     51: 255,  # '3'
     52: 255,  # '4'
     53: 255,  # '5'
     54: 255,  # '6'
     55: 255,  # '7'
     56: 255,  # '8'
     57: 255,  # '9'
     58: 255,  # ':'
     59: 255,  # ';'
     60: 255,  # '<'
     61: 255,  # '='
     62: 255,  # '>'
     63: 255,  # '?'
     64: 255,  # '@'
     65: 23,  # 'A'
     66: 37,  # 'B'
     67: 47,  # 'C'
     68: 39,  # 'D'
     69: 29,  # 'E'
     70: 52,  # 'F'
     71: 36,  # 'G'
     72: 45,  # 'H'
     73: 53,  # 'I'
     74: 60,  # 'J'
     75: 16,  # 'K'
     76: 49,  # 'L'
     77: 20,  # 'M'
     78: 46,  # 'N'
     79: 42,  # 'O'
     80: 48,  # 'P'
     81: 69,  # 'Q'
     82: 44,  # 'R'
     83: 35,  # 'S'
     84: 31,  # 'T'
     85: 51,  # 'U'
     86: 38,  # 'V'
     87: 62,  # 'W'
     88: 65,  # 'X'
     89: 43,  # 'Y'
     90: 56,  # 'Z'
     91: 255,  # '['
     92: 255,  # '\\'
     93: 255,  # ']'
     94: 255,  # '^'
     95: 255,  # '_'
     96: 255,  # '`'
     97: 1,  # 'a'
     98: 21,  # 'b'
     99: 28,  # 'c'
     100: 12,  # 'd'
     101: 2,  # 'e'
     102: 18,  # 'f'
     103: 27,  # 'g'
     104: 25,  # 'h'
     105: 3,  # 'i'
     106: 24,  # 'j'
     107: 10,  # 'k'
     108: 5,  # 'l'
     109: 13,  # 'm'
     110: 4,  # 'n'
     111: 15,  # 'o'
     112: 26,  # 'p'
     113: 64,  # 'q'
     114: 7,  # 'r'
     115: 8,  # 's'
     116: 9,  # 't'
     117: 14,  # 'u'
     118: 32,  # 'v'
     119: 57,  # 'w'
     120: 58,  # 'x'
     121: 11,  # 'y'
     122: 22,  # 'z'
     123: 255,  # '{'
     124: 255,  # '|'
     125: 255,  # '}'
     126: 255,  # '~'
     127: 255,  # '\x7f'
     128: 180,  # '\x80'
     129: 179,  # '\x81'
     130: 178,  # '\x82'
     131: 177,  # '\x83'
     132: 176,  # '\x84'
     133: 175,  # '\x85'
     134: 174,  # '\x86'
     135: 173,  # '\x87'
     136: 172,  # '\x88'
     137: 171,  # '\x89'
     138: 170,  # '\x8a'
     139: 169,  # '\x8b'
     140: 168,  # '\x8c'
     141: 167,  # '\x8d'
     142: 166,  # '\x8e'
     143: 165,  # '\x8f'
     144: 164,  # '\x90'
     145: 163,  # '\x91'
     146: 162,  # '\x92'
     147: 161,  # '\x93'
     148: 160,  # '\x94'
     149: 159,  # '\x95'
     150: 101,  # '\x96'
     151: 158,  # '\x97'
     152: 157,  # '\x98'
     153: 156,  # '\x99'
     154: 155,  # '\x9a'
     155: 154,  # '\x9b'
     156: 153,  # '\x9c'
     157: 152,  # '\x9d'
     158: 151,  # '\x9e'
     159: 106,  # '\x9f'
     160: 150,  # '\xa0'
     161: 149,  # '¡'
     162: 148,  # '¢'
     163: 147,  # '£'
     164: 146,  # '¤'
     165: 145,  # '¥'
     166: 144,  # '¦'
     167: 100,  # '§'
     168: 143,  # '¨'
     169: 142,  # '©'
     170: 141,  # 'ª'
     171: 140,  # '«'
     172: 139,  # '¬'
     173: 138,  # '\xad'
     174: 137,  # '®'
     175: 136,  # '¯'
     176: 94,  # '°'
     177: 80,  # '±'
     178: 93,  # '²'
     179: 135,  # '³'
     180: 105,  # '´'
     181: 134,  # 'µ'
     182: 133,  # '¶'
     183: 63,  # '·'
     184: 132,  # '¸'
     185: 131,  # '¹'
     186: 130,  # 'º'
     187: 129,  # '»'
     188: 128,  # '¼'
     189: 127,  # '½'
     190: 126,  # '¾'
     191: 125,  # '¿'
     192: 124,  # 'À'
     193: 104,  # 'Á'
     194: 73,  # 'Â'
     195: 99,  # 'Ã'
     196: 79,  # 'Ä'
     197: 85,  # 'Å'
     198: 123,  # 'Æ'
     199: 54,  # 'Ç'
     200: 122,  # 'È'
     201: 98,  # 'É'
     202: 92,  # 'Ê'
     203: 121,  # 'Ë'
     204: 120,  # 'Ì'
     205: 91,  # 'Í'
     206: 103,  # 'Î'
     207: 119,  # 'Ï'
     208: 68,  # 'Ğ'
     209: 118,  # 'Ñ'
     210: 117,  # 'Ò'
     211: 97,  # 'Ó'
     212: 116,  # 'Ô'
     213: 115,  # 'Õ'
     214: 50,  # 'Ö'
     215: 90,  # '×'
     216: 114,  # 'Ø'
     217: 113,  # 'Ù'
     218: 112,  # 'Ú'
     219: 111,  # 'Û'
     220: 55,  # 'Ü'
     221: 41,  # 'İ'
     222: 40,  # 'Ş'
     223: 86,  # 'ß'
     224: 89,  # 'à'
     225: 70,  # 'á'
     226: 59,  # 'â'
     227: 78,  # 'ã'
     228: 71,  # 'ä'
     229: 82,  # 'å'
     230: 88,  # 'æ'
     231: 33,  # 'ç'
     232: 77,  # 'è'
     233: 66,  # 'é'
     234: 84,  # 'ê'
     235: 83,  # 'ë'
     236: 110,  # 'ì'
     237: 75,  # 'í'
     238: 61,  # 'î'
     239: 96,  # 'ï'
     240: 30,  # 'ğ'
     241: 67,  # 'ñ'
     242: 109,  # 'ò'
     243: 74,  # 'ó'
     244: 87,  # 'ô'
     245: 102,  # 'õ'
     246: 34,  # 'ö'
     247: 95,  # '÷'
     248: 81,  # 'ø'
     249: 108,  # 'ù'
     250: 76,  # 'ú'
     251: 72,  # 'û'
     252: 17,  # 'ü'
     253: 6,  # 'ı'
     254: 19,  # 'ş'
     255: 107,  # 'ÿ'
}

ISO_8859_9_TURKISH_MODEL = SingleByteCharSetModel(charset_name='ISO-8859-9',
                                                  language='Turkish',
                                                  char_to_order_map=ISO_8859_9_TURKISH_CHAR_TO_ORDER,
                                                  language_model=TURKISH_LANG_MODEL,
                                                  typical_positive_ratio=0.97029,
                                                  keep_ascii_letters=True,
                                                  alphabet='ABCDEFGHIJKLMNOPRSTUVYZabcdefghijklmnoprstuvyzÂÇÎÖÛÜâçîöûüĞğİıŞş')

site-packages/pip/_vendor/chardet/codingstatemachine.py000064400000007006150564745010017337 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

import logging

from .enums import MachineState


class CodingStateMachine(object):
    """
    A state machine to verify a byte sequence for a particular encoding. For
    each byte the detector receives, it will feed that byte to every active
    state machine available, one byte at a time. The state machine changes its
    state based on its previous state and the byte it receives. There are 3
    states in a state machine that are of interest to an auto-detector:

    START state: This is the state to start with, or a legal byte sequence
                 (i.e. a valid code point) for character has been identified.

    ME state:  This indicates that the state machine identified a byte sequence
               that is specific to the charset it is designed for and that
               there is no other possible encoding which can contain this byte
               sequence. This will to lead to an immediate positive answer for
               the detector.

    ERROR state: This indicates the state machine identified an illegal byte
                 sequence for that encoding. This will lead to an immediate
                 negative answer for this encoding. Detector will exclude this
                 encoding from consideration from here on.
    """
    def __init__(self, sm):
        self._model = sm
        self._curr_byte_pos = 0
        self._curr_char_len = 0
        self._curr_state = None
        self.logger = logging.getLogger(__name__)
        self.reset()

    def reset(self):
        self._curr_state = MachineState.START

    def next_state(self, c):
        # for each byte we get its class
        # if it is first byte, we also get byte length
        byte_class = self._model['class_table'][c]
        if self._curr_state == MachineState.START:
            self._curr_byte_pos = 0
            self._curr_char_len = self._model['char_len_table'][byte_class]
        # from byte's class and state_table, we get its next state
        curr_state = (self._curr_state * self._model['class_factor']
                      + byte_class)
        self._curr_state = self._model['state_table'][curr_state]
        self._curr_byte_pos += 1
        return self._curr_state

    def get_current_charlen(self):
        return self._curr_char_len

    def get_coding_state_machine(self):
        return self._model['name']

    @property
    def language(self):
        return self._model['language']
site-packages/pip/_vendor/chardet/metadata/__pycache__/languages.cpython-39.pyc000064400000017442150564745010023530 0ustar00a

��Re�K�&@sdZddlmZmZddlmZGdd�de�Zedddgd	�d
dgd�ed
ddgd�ddgd�edddgd�ddgd�edddddgddgd�edddgd �d!d"gd�ed#d$dd%d&gd'd(gd�ed)d*dd+d,gd-d.gd�ed/d0dd%d&gd1gd2�ed3d4dd5gd6d7gd�ed8d9dgd �d:d;gd�ed<d=dgd>�d?d@gd�edAdBdgd �dCdDgd�edEdFdgd �dGdHdIgd�edJdKddLdMgdNdOgd�edPdQdddgdRdSgd�edTdUdddgdVdWgd�edXdYdgd �dZd[gd�ed\d]dgd^�d_d`gd�edadbdgd^�dcddgd�ededfdgdg�dhdigd�edjdkdd%d&gdlgd2�edmdndddgdodpgd�edqdrdgd �dsdtgd�edudvdddgdwdxgd�edydzdgd{�d|d}gd�ed~ddddgd�d�gd�ed�d�dddgd�d�gd�ed�d�d�gdg�d�gd��ed�d�dgd��d�d�gd�ed�d�dgd��d�d�gd�ed�d�dd�gd�d�gd�d��Zd�S)�z�
Metadata about languages used by our model training code for our
SingleByteCharSetProbers.  Could be used for other things in the future.

This code is based on the language metadata from the uchardet project.
�)�absolute_import�print_function)�
ascii_letterscs*eZdZdZd�fdd�	Zdd�Z�ZS)	�Languagea�Metadata about a language useful for training models

    :ivar name: The human name for the language, in English.
    :type name: str
    :ivar iso_code: 2-letter ISO 639-1 if possible, 3-letter ISO code otherwise,
                    or use another catalog as a last resort.
    :type iso_code: str
    :ivar use_ascii: Whether or not ASCII letters should be included in trained
                     models.
    :type use_ascii: bool
    :ivar charsets: The charsets we want to support and create data for.
    :type charsets: list of str
    :ivar alphabet: The characters in the language's alphabet. If `use_ascii` is
                    `True`, you only need to add those not in the ASCII set.
    :type alphabet: str
    :ivar wiki_start_pages: The Wikipedia pages to start from if we're crawling
                            Wikipedia for training data.
    :type wiki_start_pages: list of str
    NTcsrtt|���||_||_||_||_|jr@|r:|t7}qLt}n|sLtd��|rbd�	t
t|���nd|_||_
dS)Nz*Must supply alphabet if use_ascii is False�)�superr�__init__�name�iso_code�	use_ascii�charsetsr�
ValueError�join�sorted�set�alphabet�wiki_start_pages)�selfr	r
rrrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/metadata/languages.pyr#s
zLanguage.__init__cCs&d�|jjd�dd�|j��D���S)Nz{}({})z, css(|] \}}|�d�sd�||�VqdS)�_z{}={!r}N)�
startswith�format)�.0�k�vrrr�	<genexpr>6s
�z$Language.__repr__.<locals>.<genexpr>)rr�__name__r�__dict__�items)rrrr�__repr__4s


��zLanguage.__repr__)NNTNNN)r�
__module__�__qualname__�__doc__rr"�
__classcell__rrrrrs
�r�Arabic�arF)z
ISO-8859-6zWINDOWS-1256ZCP720ZCP864ubءآأؤإئابةتثجحخدذرزسشصضطظعغػؼؽؾؿـفقكلمنهوىيًٌٍَُِّuالصفحة_الرئيسية)r	r
rrrr�
Belarusian�be)�
ISO-8859-5�WINDOWS-1251�IBM866�MacCyrillicu�АБВГДЕЁЖЗІЙКЛМНОПРСТУЎФХЦЧШЫЬЭЮЯабвгдеёжзійклмнопрстуўфхцчшыьэюяʼu!Галоўная_старонка�	Bulgarian�bg)r+r,�IBM855uxАБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЬЮЯабвгдежзийклмнопрстуфхцчшщъьюяuНачална_страница�Czech�czTz
ISO-8859-2zWINDOWS-1250u<áčďéěíňóřšťúůýžÁČĎÉĚÍŇÓŘŠŤÚŮÝŽuHlavní_strana�Danish�da)�
ISO-8859-1zISO-8859-15�WINDOWS-1252uæøåÆØÅZForside�German�der6r7uäöüßÄÖÜzWikipedia:Hauptseite�Greek�elz
ISO-8859-7zWINDOWS-1253u�αβγδεζηθικλμνξοπρσςτυφχψωάέήίόύώΑΒΓΔΕΖΗΘΙΚΛΜΝΞΟΠΡΣΣΤΥΦΧΨΩΆΈΉΊΌΎΏuΠύλη:Κύρια�English�enZ	Main_Page)r	r
rrr�	Esperanto�eo�
ISO-8859-3uDabcĉdefgĝhĥijĵklmnoprsŝtuŭvzABCĈDEFGĜHĤIJĴKLMNOPRSŜTUŬVZuVikipedio:Ĉefpaĝo�Spanish�esuñáéíóúüÑÁÉÍÓÚÜzWikipedia:Portada�Estonian�et)�
ISO-8859-4�ISO-8859-13�WINDOWS-1257u6ABDEGHIJKLMNOPRSTUVÕÄÖÜabdeghijklmnoprstuvõäöüZEsileht�Finnish�fiuÅÄÖŠŽåäöšžzWikipedia:Etusivu�French�fru,œàâçèéîïùûêŒÀÂÇÈÉÎÏÙÛÊuWikipédia:Accueil_principaluBœuf (animal)�Hebrew�hez
ISO-8859-8zWINDOWS-1255u<אבגדהוזחטיךכלםמןנסעףפץצקרשתװױײuעמוד_ראשי�Croatian�hru@abcčćdđefghijklmnoprsštuvzžABCČĆDĐEFGHIJKLMNOPRSŠTUVZŽZGlavna_stranica�	Hungarian�huuPabcdefghijklmnoprstuvzáéíóöőúüűABCDEFGHIJKLMNOPRSTUVZÁÉÍÓÖŐÚÜŰu	Kezdőlap�Italian�ituÀÈÉÌÒÓÙàèéìòóùZPagina_principale�
Lithuanian�lt)rFrGrEuRAĄBCČDEĘĖFGHIĮYJKLMNOPRSŠTUŲŪVZŽaąbcčdeęėfghiįyjklmnoprsštuųūvzžZPagrindinis_puslapis�Latvian�lvuXAĀBCČDEĒFGĢHIĪJKĶLĻMNŅOPRSŠTUŪVZŽaābcčdeēfgģhiījkķlļmnņoprsštuūvzžu
Sākumlapa�
Macedonian�mk)r+r,r.r1u|АБВГДЃЕЖЗЅИЈКЛЉМНЊОПРСТЌУФХЦЧЏШабвгдѓежзѕијклљмнњопрстќуфхцчџшuГлавна_страница�Dutch�nlZHoofdpagina�Polish�pluRAĄBCĆDEĘFGHIJKLŁMNŃOÓPRSŚTUWYZŹŻaąbcćdeęfghijklłmnńoóprsśtuwyzźżuWikipedia:Strona_główna�
Portuguese�ptu0ÁÂÃÀÇÉÊÍÓÔÕÚáâãàçéêíóôõúuWikipédia:Página_principal�Romanian�rouăâîșțĂÂÎȘȚuPagina_principală�Russian�ru)r+r,zKOI8-Rr.r-r1u�абвгдеёжзийклмнопрстуфхцчшщъыьэюяАБВГДЕЁЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯu#Заглавная_страница�Slovak�skuDáäčďéíĺľňóôŕšťúýžÁÄČĎÉÍĹĽŇÓÔŔŠŤÚÝŽuHlavná_stránka�Slovene�slu8abcčdefghijklmnoprsštuvzžABCČDEFGHIJKLMNOPRSŠTUVZŽZGlavna_stran�Serbian�sruxАБВГДЂЕЖЗИЈКЛЉМНЊОПРСТЋУФХЦЧЏШабвгдђежзијклљмнњопрстћуфхцчџшuГлавна_страна)r	r
rrr�Thai�th)zISO-8859-11zTIS-620ZCP874uกขฃคฅฆงจฉชซฌญฎฏฐฑฒณดตถทธนบปผฝพฟภมยรฤลฦวศษสหฬอฮฯะัาำิีึืฺุู฿เแโใไๅๆ็่้๊๋์ํ๎๏๐๑๒๓๔๕๖๗๘๙๚๛uหน้าหลัก�Turkish�tr)r@z
ISO-8859-9zWINDOWS-1254uRabcçdefgğhıijklmnoöprsştuüvyzâîûABCÇDEFGĞHIİJKLMNOÖPRSŞTUÜVYZÂÎÛZ	Ana_Sayfa�
Vietnamese�vizWINDOWS-1258uHaăâbcdđeêghiklmnoôơpqrstuưvxyAĂÂBCDĐEÊGHIKLMNOÔƠPQRSTUƯVXYuChữ_Quốc_ngữ)r'r)r/r2r4r8r:r<r>rArCrHrJrLrNrPrRrTrVrXrZr\r^r`rbrdrfrhrjrlrnN)	r%�
__future__rr�stringr�objectrZ	LANGUAGESrrrr�<module>s�,�����������
��������	�	������	�����	���site-packages/pip/_vendor/chardet/metadata/__pycache__/__init__.cpython-39.pyc000064400000000357150564745010023316 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/chardet/metadata/__init__.py�<module>�site-packages/pip/_vendor/chardet/metadata/__init__.py000064400000000000150564745010017010 0ustar00site-packages/pip/_vendor/chardet/metadata/languages.py000064400000045774150564745010017252 0ustar00# -*- coding: utf-8 -*-
"""
Metadata about languages used by our model training code for our
SingleByteCharSetProbers.  Could be used for other things in the future.

This code is based on the language metadata from the uchardet project.
"""
from __future__ import absolute_import, print_function

from string import ascii_letters


# TODO: Add Ukranian (KOI8-U)

class Language(object):
    """Metadata about a language useful for training models

    :ivar name: The human name for the language, in English.
    :type name: str
    :ivar iso_code: 2-letter ISO 639-1 if possible, 3-letter ISO code otherwise,
                    or use another catalog as a last resort.
    :type iso_code: str
    :ivar use_ascii: Whether or not ASCII letters should be included in trained
                     models.
    :type use_ascii: bool
    :ivar charsets: The charsets we want to support and create data for.
    :type charsets: list of str
    :ivar alphabet: The characters in the language's alphabet. If `use_ascii` is
                    `True`, you only need to add those not in the ASCII set.
    :type alphabet: str
    :ivar wiki_start_pages: The Wikipedia pages to start from if we're crawling
                            Wikipedia for training data.
    :type wiki_start_pages: list of str
    """
    def __init__(self, name=None, iso_code=None, use_ascii=True, charsets=None,
                 alphabet=None, wiki_start_pages=None):
        super(Language, self).__init__()
        self.name = name
        self.iso_code = iso_code
        self.use_ascii = use_ascii
        self.charsets = charsets
        if self.use_ascii:
            if alphabet:
                alphabet += ascii_letters
            else:
                alphabet = ascii_letters
        elif not alphabet:
            raise ValueError('Must supply alphabet if use_ascii is False')
        self.alphabet = ''.join(sorted(set(alphabet))) if alphabet else None
        self.wiki_start_pages = wiki_start_pages

    def __repr__(self):
        return '{}({})'.format(self.__class__.__name__,
                               ', '.join('{}={!r}'.format(k, v)
                                         for k, v in self.__dict__.items()
                                         if not k.startswith('_')))


LANGUAGES = {'Arabic': Language(name='Arabic',
                                iso_code='ar',
                                use_ascii=False,
                                # We only support encodings that use isolated
                                # forms, because the current recommendation is
                                # that the rendering system handles presentation
                                # forms. This means we purposefully skip IBM864.
                                charsets=['ISO-8859-6', 'WINDOWS-1256',
                                          'CP720', 'CP864'],
                                alphabet=u'ءآأؤإئابةتثجحخدذرزسشصضطظعغػؼؽؾؿـفقكلمنهوىيًٌٍَُِّ',
                                wiki_start_pages=[u'الصفحة_الرئيسية']),
             'Belarusian': Language(name='Belarusian',
                                    iso_code='be',
                                    use_ascii=False,
                                    charsets=['ISO-8859-5', 'WINDOWS-1251',
                                              'IBM866', 'MacCyrillic'],
                                    alphabet=(u'АБВГДЕЁЖЗІЙКЛМНОПРСТУЎФХЦЧШЫЬЭЮЯ'
                                              u'абвгдеёжзійклмнопрстуўфхцчшыьэюяʼ'),
                                    wiki_start_pages=[u'Галоўная_старонка']),
             'Bulgarian': Language(name='Bulgarian',
                                   iso_code='bg',
                                   use_ascii=False,
                                   charsets=['ISO-8859-5', 'WINDOWS-1251',
                                             'IBM855'],
                                   alphabet=(u'АБВГДЕЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЬЮЯ'
                                             u'абвгдежзийклмнопрстуфхцчшщъьюя'),
                                   wiki_start_pages=[u'Начална_страница']),
             'Czech': Language(name='Czech',
                               iso_code='cz',
                               use_ascii=True,
                               charsets=['ISO-8859-2', 'WINDOWS-1250'],
                               alphabet=u'áčďéěíňóřšťúůýžÁČĎÉĚÍŇÓŘŠŤÚŮÝŽ',
                               wiki_start_pages=[u'Hlavní_strana']),
             'Danish': Language(name='Danish',
                                iso_code='da',
                                use_ascii=True,
                                charsets=['ISO-8859-1', 'ISO-8859-15',
                                          'WINDOWS-1252'],
                                alphabet=u'æøåÆØÅ',
                                wiki_start_pages=[u'Forside']),
             'German': Language(name='German',
                                iso_code='de',
                                use_ascii=True,
                                charsets=['ISO-8859-1', 'WINDOWS-1252'],
                                alphabet=u'äöüßÄÖÜ',
                                wiki_start_pages=[u'Wikipedia:Hauptseite']),
             'Greek': Language(name='Greek',
                               iso_code='el',
                               use_ascii=False,
                               charsets=['ISO-8859-7', 'WINDOWS-1253'],
                               alphabet=(u'αβγδεζηθικλμνξοπρσςτυφχψωάέήίόύώ'
                                         u'ΑΒΓΔΕΖΗΘΙΚΛΜΝΞΟΠΡΣΣΤΥΦΧΨΩΆΈΉΊΌΎΏ'),
                               wiki_start_pages=[u'Πύλη:Κύρια']),
             'English': Language(name='English',
                                 iso_code='en',
                                 use_ascii=True,
                                 charsets=['ISO-8859-1', 'WINDOWS-1252'],
                                 wiki_start_pages=[u'Main_Page']),
             'Esperanto': Language(name='Esperanto',
                                   iso_code='eo',
                                   # Q, W, X, and Y not used at all
                                   use_ascii=False,
                                   charsets=['ISO-8859-3'],
                                   alphabet=(u'abcĉdefgĝhĥijĵklmnoprsŝtuŭvz'
                                             u'ABCĈDEFGĜHĤIJĴKLMNOPRSŜTUŬVZ'),
                                   wiki_start_pages=[u'Vikipedio:Ĉefpaĝo']),
             'Spanish': Language(name='Spanish',
                                 iso_code='es',
                                 use_ascii=True,
                                 charsets=['ISO-8859-1', 'ISO-8859-15',
                                           'WINDOWS-1252'],
                                 alphabet=u'ñáéíóúüÑÁÉÍÓÚÜ',
                                 wiki_start_pages=[u'Wikipedia:Portada']),
             'Estonian': Language(name='Estonian',
                                  iso_code='et',
                                  use_ascii=False,
                                  charsets=['ISO-8859-4', 'ISO-8859-13',
                                            'WINDOWS-1257'],
                                  # C, F, Š, Q, W, X, Y, Z, Ž are only for
                                  # loanwords
                                  alphabet=(u'ABDEGHIJKLMNOPRSTUVÕÄÖÜ'
                                            u'abdeghijklmnoprstuvõäöü'),
                                  wiki_start_pages=[u'Esileht']),
             'Finnish': Language(name='Finnish',
                                 iso_code='fi',
                                 use_ascii=True,
                                 charsets=['ISO-8859-1', 'ISO-8859-15',
                                           'WINDOWS-1252'],
                                 alphabet=u'ÅÄÖŠŽåäöšž',
                                 wiki_start_pages=[u'Wikipedia:Etusivu']),
             'French': Language(name='French',
                                iso_code='fr',
                                use_ascii=True,
                                charsets=['ISO-8859-1', 'ISO-8859-15',
                                          'WINDOWS-1252'],
                                alphabet=u'œàâçèéîïùûêŒÀÂÇÈÉÎÏÙÛÊ',
                                wiki_start_pages=[u'Wikipédia:Accueil_principal',
                                                  u'Bœuf (animal)']),
             'Hebrew': Language(name='Hebrew',
                                iso_code='he',
                                use_ascii=False,
                                charsets=['ISO-8859-8', 'WINDOWS-1255'],
                                alphabet=u'אבגדהוזחטיךכלםמןנסעףפץצקרשתװױײ',
                                wiki_start_pages=[u'עמוד_ראשי']),
             'Croatian': Language(name='Croatian',
                                  iso_code='hr',
                                  # Q, W, X, Y are only used for foreign words.
                                  use_ascii=False,
                                  charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                  alphabet=(u'abcčćdđefghijklmnoprsštuvzž'
                                            u'ABCČĆDĐEFGHIJKLMNOPRSŠTUVZŽ'),
                                  wiki_start_pages=[u'Glavna_stranica']),
             'Hungarian': Language(name='Hungarian',
                                   iso_code='hu',
                                   # Q, W, X, Y are only used for foreign words.
                                   use_ascii=False,
                                   charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                   alphabet=(u'abcdefghijklmnoprstuvzáéíóöőúüű'
                                             u'ABCDEFGHIJKLMNOPRSTUVZÁÉÍÓÖŐÚÜŰ'),
                                   wiki_start_pages=[u'Kezdőlap']),
             'Italian': Language(name='Italian',
                                 iso_code='it',
                                 use_ascii=True,
                                 charsets=['ISO-8859-1', 'ISO-8859-15',
                                           'WINDOWS-1252'],
                                 alphabet=u'ÀÈÉÌÒÓÙàèéìòóù',
                                 wiki_start_pages=[u'Pagina_principale']),
             'Lithuanian': Language(name='Lithuanian',
                                    iso_code='lt',
                                    use_ascii=False,
                                    charsets=['ISO-8859-13', 'WINDOWS-1257',
                                              'ISO-8859-4'],
                                    # Q, W, and X not used at all
                                    alphabet=(u'AĄBCČDEĘĖFGHIĮYJKLMNOPRSŠTUŲŪVZŽ'
                                              u'aąbcčdeęėfghiįyjklmnoprsštuųūvzž'),
                                    wiki_start_pages=[u'Pagrindinis_puslapis']),
             'Latvian': Language(name='Latvian',
                                 iso_code='lv',
                                 use_ascii=False,
                                 charsets=['ISO-8859-13', 'WINDOWS-1257',
                                           'ISO-8859-4'],
                                 # Q, W, X, Y are only for loanwords
                                 alphabet=(u'AĀBCČDEĒFGĢHIĪJKĶLĻMNŅOPRSŠTUŪVZŽ'
                                           u'aābcčdeēfgģhiījkķlļmnņoprsštuūvzž'),
                                 wiki_start_pages=[u'Sākumlapa']),
             'Macedonian': Language(name='Macedonian',
                                    iso_code='mk',
                                    use_ascii=False,
                                    charsets=['ISO-8859-5', 'WINDOWS-1251',
                                              'MacCyrillic', 'IBM855'],
                                    alphabet=(u'АБВГДЃЕЖЗЅИЈКЛЉМНЊОПРСТЌУФХЦЧЏШ'
                                              u'абвгдѓежзѕијклљмнњопрстќуфхцчџш'),
                                    wiki_start_pages=[u'Главна_страница']),
             'Dutch': Language(name='Dutch',
                               iso_code='nl',
                               use_ascii=True,
                               charsets=['ISO-8859-1', 'WINDOWS-1252'],
                               wiki_start_pages=[u'Hoofdpagina']),
             'Polish': Language(name='Polish',
                                iso_code='pl',
                                # Q and X are only used for foreign words.
                                use_ascii=False,
                                charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                alphabet=(u'AĄBCĆDEĘFGHIJKLŁMNŃOÓPRSŚTUWYZŹŻ'
                                          u'aąbcćdeęfghijklłmnńoóprsśtuwyzźż'),
                                wiki_start_pages=[u'Wikipedia:Strona_główna']),
             'Portuguese': Language(name='Portuguese',
                                 iso_code='pt',
                                 use_ascii=True,
                                 charsets=['ISO-8859-1', 'ISO-8859-15',
                                           'WINDOWS-1252'],
                                 alphabet=u'ÁÂÃÀÇÉÊÍÓÔÕÚáâãàçéêíóôõú',
                                 wiki_start_pages=[u'Wikipédia:Página_principal']),
             'Romanian': Language(name='Romanian',
                                  iso_code='ro',
                                  use_ascii=True,
                                  charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                  alphabet=u'ăâîșțĂÂÎȘȚ',
                                  wiki_start_pages=[u'Pagina_principală']),
             'Russian': Language(name='Russian',
                                 iso_code='ru',
                                 use_ascii=False,
                                 charsets=['ISO-8859-5', 'WINDOWS-1251',
                                           'KOI8-R', 'MacCyrillic', 'IBM866',
                                           'IBM855'],
                                 alphabet=(u'абвгдеёжзийклмнопрстуфхцчшщъыьэюя'
                                           u'АБВГДЕЁЖЗИЙКЛМНОПРСТУФХЦЧШЩЪЫЬЭЮЯ'),
                                 wiki_start_pages=[u'Заглавная_страница']),
             'Slovak': Language(name='Slovak',
                                iso_code='sk',
                                use_ascii=True,
                                charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                alphabet=u'áäčďéíĺľňóôŕšťúýžÁÄČĎÉÍĹĽŇÓÔŔŠŤÚÝŽ',
                                wiki_start_pages=[u'Hlavná_stránka']),
             'Slovene': Language(name='Slovene',
                                 iso_code='sl',
                                 # Q, W, X, Y are only used for foreign words.
                                 use_ascii=False,
                                 charsets=['ISO-8859-2', 'WINDOWS-1250'],
                                 alphabet=(u'abcčdefghijklmnoprsštuvzž'
                                           u'ABCČDEFGHIJKLMNOPRSŠTUVZŽ'),
                                 wiki_start_pages=[u'Glavna_stran']),
             # Serbian can be written in both Latin and Cyrillic, but there's no
             # simple way to get the Latin alphabet pages from Wikipedia through
             # the API, so for now we just support Cyrillic.
             'Serbian': Language(name='Serbian',
                                 iso_code='sr',
                                 alphabet=(u'АБВГДЂЕЖЗИЈКЛЉМНЊОПРСТЋУФХЦЧЏШ'
                                           u'абвгдђежзијклљмнњопрстћуфхцчџш'),
                                 charsets=['ISO-8859-5', 'WINDOWS-1251',
                                           'MacCyrillic', 'IBM855'],
                                 wiki_start_pages=[u'Главна_страна']),
             'Thai': Language(name='Thai',
                              iso_code='th',
                              use_ascii=False,
                              charsets=['ISO-8859-11', 'TIS-620', 'CP874'],
                              alphabet=u'กขฃคฅฆงจฉชซฌญฎฏฐฑฒณดตถทธนบปผฝพฟภมยรฤลฦวศษสหฬอฮฯะัาำิีึืฺุู฿เแโใไๅๆ็่้๊๋์ํ๎๏๐๑๒๓๔๕๖๗๘๙๚๛',
                              wiki_start_pages=[u'หน้าหลัก']),
             'Turkish': Language(name='Turkish',
                                 iso_code='tr',
                                 # Q, W, and X are not used by Turkish
                                 use_ascii=False,
                                 charsets=['ISO-8859-3', 'ISO-8859-9',
                                           'WINDOWS-1254'],
                                 alphabet=(u'abcçdefgğhıijklmnoöprsştuüvyzâîû'
                                           u'ABCÇDEFGĞHIİJKLMNOÖPRSŞTUÜVYZÂÎÛ'),
                                 wiki_start_pages=[u'Ana_Sayfa']),
             'Vietnamese': Language(name='Vietnamese',
                                    iso_code='vi',
                                    use_ascii=False,
                                    # Windows-1258 is the only common 8-bit
                                    # Vietnamese encoding supported by Python.
                                    # From Wikipedia:
                                    # For systems that lack support for Unicode,
                                    # dozens of 8-bit Vietnamese code pages are
                                    # available.[1] The most common are VISCII
                                    # (TCVN 5712:1993), VPS, and Windows-1258.[3]
                                    # Where ASCII is required, such as when
                                    # ensuring readability in plain text e-mail,
                                    # Vietnamese letters are often encoded
                                    # according to Vietnamese Quoted-Readable
                                    # (VIQR) or VSCII Mnemonic (VSCII-MNEM),[4]
                                    # though usage of either variable-width
                                    # scheme has declined dramatically following
                                    # the adoption of Unicode on the World Wide
                                    # Web.
                                    charsets=['WINDOWS-1258'],
                                    alphabet=(u'aăâbcdđeêghiklmnoôơpqrstuưvxy'
                                              u'AĂÂBCDĐEÊGHIKLMNOÔƠPQRSTUƯVXY'),
                                    wiki_start_pages=[u'Chữ_Quốc_ngữ']),
            }
site-packages/pip/_vendor/chardet/universaldetector.py000064400000030327150564745010017252 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################
"""
Module containing the UniversalDetector detector class, which is the primary
class a user of ``chardet`` should use.

:author: Mark Pilgrim (initial port to Python)
:author: Shy Shalom (original C code)
:author: Dan Blanchard (major refactoring for 3.0)
:author: Ian Cordasco
"""


import codecs
import logging
import re

from .charsetgroupprober import CharSetGroupProber
from .enums import InputState, LanguageFilter, ProbingState
from .escprober import EscCharSetProber
from .latin1prober import Latin1Prober
from .mbcsgroupprober import MBCSGroupProber
from .sbcsgroupprober import SBCSGroupProber


class UniversalDetector(object):
    """
    The ``UniversalDetector`` class underlies the ``chardet.detect`` function
    and coordinates all of the different charset probers.

    To get a ``dict`` containing an encoding and its confidence, you can simply
    run:

    .. code::

            u = UniversalDetector()
            u.feed(some_bytes)
            u.close()
            detected = u.result

    """

    MINIMUM_THRESHOLD = 0.20
    HIGH_BYTE_DETECTOR = re.compile(b'[\x80-\xFF]')
    ESC_DETECTOR = re.compile(b'(\033|~{)')
    WIN_BYTE_DETECTOR = re.compile(b'[\x80-\x9F]')
    ISO_WIN_MAP = {'iso-8859-1': 'Windows-1252',
                   'iso-8859-2': 'Windows-1250',
                   'iso-8859-5': 'Windows-1251',
                   'iso-8859-6': 'Windows-1256',
                   'iso-8859-7': 'Windows-1253',
                   'iso-8859-8': 'Windows-1255',
                   'iso-8859-9': 'Windows-1254',
                   'iso-8859-13': 'Windows-1257'}

    def __init__(self, lang_filter=LanguageFilter.ALL):
        self._esc_charset_prober = None
        self._charset_probers = []
        self.result = None
        self.done = None
        self._got_data = None
        self._input_state = None
        self._last_char = None
        self.lang_filter = lang_filter
        self.logger = logging.getLogger(__name__)
        self._has_win_bytes = None
        self.reset()

    def reset(self):
        """
        Reset the UniversalDetector and all of its probers back to their
        initial states.  This is called by ``__init__``, so you only need to
        call this directly in between analyses of different documents.
        """
        self.result = {'encoding': None, 'confidence': 0.0, 'language': None}
        self.done = False
        self._got_data = False
        self._has_win_bytes = False
        self._input_state = InputState.PURE_ASCII
        self._last_char = b''
        if self._esc_charset_prober:
            self._esc_charset_prober.reset()
        for prober in self._charset_probers:
            prober.reset()

    def feed(self, byte_str):
        """
        Takes a chunk of a document and feeds it through all of the relevant
        charset probers.

        After calling ``feed``, you can check the value of the ``done``
        attribute to see if you need to continue feeding the
        ``UniversalDetector`` more data, or if it has made a prediction
        (in the ``result`` attribute).

        .. note::
           You should always call ``close`` when you're done feeding in your
           document if ``done`` is not already ``True``.
        """
        if self.done:
            return

        if not len(byte_str):
            return

        if not isinstance(byte_str, bytearray):
            byte_str = bytearray(byte_str)

        # First check for known BOMs, since these are guaranteed to be correct
        if not self._got_data:
            # If the data starts with BOM, we know it is UTF
            if byte_str.startswith(codecs.BOM_UTF8):
                # EF BB BF  UTF-8 with BOM
                self.result = {'encoding': "UTF-8-SIG",
                               'confidence': 1.0,
                               'language': ''}
            elif byte_str.startswith((codecs.BOM_UTF32_LE,
                                      codecs.BOM_UTF32_BE)):
                # FF FE 00 00  UTF-32, little-endian BOM
                # 00 00 FE FF  UTF-32, big-endian BOM
                self.result = {'encoding': "UTF-32",
                               'confidence': 1.0,
                               'language': ''}
            elif byte_str.startswith(b'\xFE\xFF\x00\x00'):
                # FE FF 00 00  UCS-4, unusual octet order BOM (3412)
                self.result = {'encoding': "X-ISO-10646-UCS-4-3412",
                               'confidence': 1.0,
                               'language': ''}
            elif byte_str.startswith(b'\x00\x00\xFF\xFE'):
                # 00 00 FF FE  UCS-4, unusual octet order BOM (2143)
                self.result = {'encoding': "X-ISO-10646-UCS-4-2143",
                               'confidence': 1.0,
                               'language': ''}
            elif byte_str.startswith((codecs.BOM_LE, codecs.BOM_BE)):
                # FF FE  UTF-16, little endian BOM
                # FE FF  UTF-16, big endian BOM
                self.result = {'encoding': "UTF-16",
                               'confidence': 1.0,
                               'language': ''}

            self._got_data = True
            if self.result['encoding'] is not None:
                self.done = True
                return

        # If none of those matched and we've only see ASCII so far, check
        # for high bytes and escape sequences
        if self._input_state == InputState.PURE_ASCII:
            if self.HIGH_BYTE_DETECTOR.search(byte_str):
                self._input_state = InputState.HIGH_BYTE
            elif self._input_state == InputState.PURE_ASCII and \
                    self.ESC_DETECTOR.search(self._last_char + byte_str):
                self._input_state = InputState.ESC_ASCII

        self._last_char = byte_str[-1:]

        # If we've seen escape sequences, use the EscCharSetProber, which
        # uses a simple state machine to check for known escape sequences in
        # HZ and ISO-2022 encodings, since those are the only encodings that
        # use such sequences.
        if self._input_state == InputState.ESC_ASCII:
            if not self._esc_charset_prober:
                self._esc_charset_prober = EscCharSetProber(self.lang_filter)
            if self._esc_charset_prober.feed(byte_str) == ProbingState.FOUND_IT:
                self.result = {'encoding':
                               self._esc_charset_prober.charset_name,
                               'confidence':
                               self._esc_charset_prober.get_confidence(),
                               'language':
                               self._esc_charset_prober.language}
                self.done = True
        # If we've seen high bytes (i.e., those with values greater than 127),
        # we need to do more complicated checks using all our multi-byte and
        # single-byte probers that are left.  The single-byte probers
        # use character bigram distributions to determine the encoding, whereas
        # the multi-byte probers use a combination of character unigram and
        # bigram distributions.
        elif self._input_state == InputState.HIGH_BYTE:
            if not self._charset_probers:
                self._charset_probers = [MBCSGroupProber(self.lang_filter)]
                # If we're checking non-CJK encodings, use single-byte prober
                if self.lang_filter & LanguageFilter.NON_CJK:
                    self._charset_probers.append(SBCSGroupProber())
                self._charset_probers.append(Latin1Prober())
            for prober in self._charset_probers:
                if prober.feed(byte_str) == ProbingState.FOUND_IT:
                    self.result = {'encoding': prober.charset_name,
                                   'confidence': prober.get_confidence(),
                                   'language': prober.language}
                    self.done = True
                    break
            if self.WIN_BYTE_DETECTOR.search(byte_str):
                self._has_win_bytes = True

    def close(self):
        """
        Stop analyzing the current document and come up with a final
        prediction.

        :returns:  The ``result`` attribute, a ``dict`` with the keys
                   `encoding`, `confidence`, and `language`.
        """
        # Don't bother with checks if we're already done
        if self.done:
            return self.result
        self.done = True

        if not self._got_data:
            self.logger.debug('no data received!')

        # Default to ASCII if it is all we've seen so far
        elif self._input_state == InputState.PURE_ASCII:
            self.result = {'encoding': 'ascii',
                           'confidence': 1.0,
                           'language': ''}

        # If we have seen non-ASCII, return the best that met MINIMUM_THRESHOLD
        elif self._input_state == InputState.HIGH_BYTE:
            prober_confidence = None
            max_prober_confidence = 0.0
            max_prober = None
            for prober in self._charset_probers:
                if not prober:
                    continue
                prober_confidence = prober.get_confidence()
                if prober_confidence > max_prober_confidence:
                    max_prober_confidence = prober_confidence
                    max_prober = prober
            if max_prober and (max_prober_confidence > self.MINIMUM_THRESHOLD):
                charset_name = max_prober.charset_name
                lower_charset_name = max_prober.charset_name.lower()
                confidence = max_prober.get_confidence()
                # Use Windows encoding name instead of ISO-8859 if we saw any
                # extra Windows-specific bytes
                if lower_charset_name.startswith('iso-8859'):
                    if self._has_win_bytes:
                        charset_name = self.ISO_WIN_MAP.get(lower_charset_name,
                                                            charset_name)
                self.result = {'encoding': charset_name,
                               'confidence': confidence,
                               'language': max_prober.language}

        # Log all prober confidences if none met MINIMUM_THRESHOLD
        if self.logger.getEffectiveLevel() <= logging.DEBUG:
            if self.result['encoding'] is None:
                self.logger.debug('no probers hit minimum threshold')
                for group_prober in self._charset_probers:
                    if not group_prober:
                        continue
                    if isinstance(group_prober, CharSetGroupProber):
                        for prober in group_prober.probers:
                            self.logger.debug('%s %s confidence = %s',
                                              prober.charset_name,
                                              prober.language,
                                              prober.get_confidence())
                    else:
                        self.logger.debug('%s %s confidence = %s',
                                          group_prober.charset_name,
                                          group_prober.language,
                                          group_prober.get_confidence())
        return self.result
site-packages/pip/_vendor/chardet/mbcharsetprober.py000064400000006525150564745010016675 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Universal charset detector code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 2001
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#   Shy Shalom - original C code
#   Proofpoint, Inc.
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .charsetprober import CharSetProber
from .enums import ProbingState, MachineState


class MultiByteCharSetProber(CharSetProber):
    """
    MultiByteCharSetProber
    """

    def __init__(self, lang_filter=None):
        super(MultiByteCharSetProber, self).__init__(lang_filter=lang_filter)
        self.distribution_analyzer = None
        self.coding_sm = None
        self._last_char = [0, 0]

    def reset(self):
        super(MultiByteCharSetProber, self).reset()
        if self.coding_sm:
            self.coding_sm.reset()
        if self.distribution_analyzer:
            self.distribution_analyzer.reset()
        self._last_char = [0, 0]

    @property
    def charset_name(self):
        raise NotImplementedError

    @property
    def language(self):
        raise NotImplementedError

    def feed(self, byte_str):
        for i in range(len(byte_str)):
            coding_state = self.coding_sm.next_state(byte_str[i])
            if coding_state == MachineState.ERROR:
                self.logger.debug('%s %s prober hit error at byte %s',
                                  self.charset_name, self.language, i)
                self._state = ProbingState.NOT_ME
                break
            elif coding_state == MachineState.ITS_ME:
                self._state = ProbingState.FOUND_IT
                break
            elif coding_state == MachineState.START:
                char_len = self.coding_sm.get_current_charlen()
                if i == 0:
                    self._last_char[1] = byte_str[0]
                    self.distribution_analyzer.feed(self._last_char, char_len)
                else:
                    self.distribution_analyzer.feed(byte_str[i - 1:i + 1],
                                                    char_len)

        self._last_char[0] = byte_str[-1]

        if self.state == ProbingState.DETECTING:
            if (self.distribution_analyzer.got_enough_data() and
                    (self.get_confidence() > self.SHORTCUT_THRESHOLD)):
                self._state = ProbingState.FOUND_IT

        return self.state

    def get_confidence(self):
        return self.distribution_analyzer.get_confidence()
site-packages/pip/_vendor/chardet/charsetgroupprober.py000064400000007377150564745010017441 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .enums import ProbingState
from .charsetprober import CharSetProber


class CharSetGroupProber(CharSetProber):
    def __init__(self, lang_filter=None):
        super(CharSetGroupProber, self).__init__(lang_filter=lang_filter)
        self._active_num = 0
        self.probers = []
        self._best_guess_prober = None

    def reset(self):
        super(CharSetGroupProber, self).reset()
        self._active_num = 0
        for prober in self.probers:
            if prober:
                prober.reset()
                prober.active = True
                self._active_num += 1
        self._best_guess_prober = None

    @property
    def charset_name(self):
        if not self._best_guess_prober:
            self.get_confidence()
            if not self._best_guess_prober:
                return None
        return self._best_guess_prober.charset_name

    @property
    def language(self):
        if not self._best_guess_prober:
            self.get_confidence()
            if not self._best_guess_prober:
                return None
        return self._best_guess_prober.language

    def feed(self, byte_str):
        for prober in self.probers:
            if not prober:
                continue
            if not prober.active:
                continue
            state = prober.feed(byte_str)
            if not state:
                continue
            if state == ProbingState.FOUND_IT:
                self._best_guess_prober = prober
                self._state = ProbingState.FOUND_IT
                return self.state
            elif state == ProbingState.NOT_ME:
                prober.active = False
                self._active_num -= 1
                if self._active_num <= 0:
                    self._state = ProbingState.NOT_ME
                    return self.state
        return self.state

    def get_confidence(self):
        state = self.state
        if state == ProbingState.FOUND_IT:
            return 0.99
        elif state == ProbingState.NOT_ME:
            return 0.01
        best_conf = 0.0
        self._best_guess_prober = None
        for prober in self.probers:
            if not prober:
                continue
            if not prober.active:
                self.logger.debug('%s not active', prober.charset_name)
                continue
            conf = prober.get_confidence()
            self.logger.debug('%s %s confidence = %s', prober.charset_name, prober.language, conf)
            if best_conf < conf:
                best_conf = conf
                self._best_guess_prober = prober
        if not self._best_guess_prober:
            return 0.0
        return best_conf
site-packages/pip/_vendor/chardet/cp949prober.py000064400000003477150564745010015600 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .chardistribution import EUCKRDistributionAnalysis
from .codingstatemachine import CodingStateMachine
from .mbcharsetprober import MultiByteCharSetProber
from .mbcssm import CP949_SM_MODEL


class CP949Prober(MultiByteCharSetProber):
    def __init__(self):
        super(CP949Prober, self).__init__()
        self.coding_sm = CodingStateMachine(CP949_SM_MODEL)
        # NOTE: CP949 is a superset of EUC-KR, so the distribution should be
        #       not different.
        self.distribution_analyzer = EUCKRDistributionAnalysis()
        self.reset()

    @property
    def charset_name(self):
        return "CP949"

    @property
    def language(self):
        return "Korean"
site-packages/pip/_vendor/chardet/euctwprober.py000064400000003323150564745010016045 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import EUCTWDistributionAnalysis
from .mbcssm import EUCTW_SM_MODEL

class EUCTWProber(MultiByteCharSetProber):
    def __init__(self):
        super(EUCTWProber, self).__init__()
        self.coding_sm = CodingStateMachine(EUCTW_SM_MODEL)
        self.distribution_analyzer = EUCTWDistributionAnalysis()
        self.reset()

    @property
    def charset_name(self):
        return "EUC-TW"

    @property
    def language(self):
        return "Taiwan"
site-packages/pip/_vendor/chardet/sjisprober.py000064400000007276150564745010015701 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .mbcharsetprober import MultiByteCharSetProber
from .codingstatemachine import CodingStateMachine
from .chardistribution import SJISDistributionAnalysis
from .jpcntx import SJISContextAnalysis
from .mbcssm import SJIS_SM_MODEL
from .enums import ProbingState, MachineState


class SJISProber(MultiByteCharSetProber):
    def __init__(self):
        super(SJISProber, self).__init__()
        self.coding_sm = CodingStateMachine(SJIS_SM_MODEL)
        self.distribution_analyzer = SJISDistributionAnalysis()
        self.context_analyzer = SJISContextAnalysis()
        self.reset()

    def reset(self):
        super(SJISProber, self).reset()
        self.context_analyzer.reset()

    @property
    def charset_name(self):
        return self.context_analyzer.charset_name

    @property
    def language(self):
        return "Japanese"

    def feed(self, byte_str):
        for i in range(len(byte_str)):
            coding_state = self.coding_sm.next_state(byte_str[i])
            if coding_state == MachineState.ERROR:
                self.logger.debug('%s %s prober hit error at byte %s',
                                  self.charset_name, self.language, i)
                self._state = ProbingState.NOT_ME
                break
            elif coding_state == MachineState.ITS_ME:
                self._state = ProbingState.FOUND_IT
                break
            elif coding_state == MachineState.START:
                char_len = self.coding_sm.get_current_charlen()
                if i == 0:
                    self._last_char[1] = byte_str[0]
                    self.context_analyzer.feed(self._last_char[2 - char_len:],
                                               char_len)
                    self.distribution_analyzer.feed(self._last_char, char_len)
                else:
                    self.context_analyzer.feed(byte_str[i + 1 - char_len:i + 3
                                                        - char_len], char_len)
                    self.distribution_analyzer.feed(byte_str[i - 1:i + 1],
                                                    char_len)

        self._last_char[0] = byte_str[-1]

        if self.state == ProbingState.DETECTING:
            if (self.context_analyzer.got_enough_data() and
               (self.get_confidence() > self.SHORTCUT_THRESHOLD)):
                self._state = ProbingState.FOUND_IT

        return self.state

    def get_confidence(self):
        context_conf = self.context_analyzer.get_confidence()
        distrib_conf = self.distribution_analyzer.get_confidence()
        return max(context_conf, distrib_conf)
site-packages/pip/_vendor/chardet/escsm.py000064400000024416150564745010014624 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .enums import MachineState

HZ_CLS = (
1,0,0,0,0,0,0,0,  # 00 - 07
0,0,0,0,0,0,0,0,  # 08 - 0f
0,0,0,0,0,0,0,0,  # 10 - 17
0,0,0,1,0,0,0,0,  # 18 - 1f
0,0,0,0,0,0,0,0,  # 20 - 27
0,0,0,0,0,0,0,0,  # 28 - 2f
0,0,0,0,0,0,0,0,  # 30 - 37
0,0,0,0,0,0,0,0,  # 38 - 3f
0,0,0,0,0,0,0,0,  # 40 - 47
0,0,0,0,0,0,0,0,  # 48 - 4f
0,0,0,0,0,0,0,0,  # 50 - 57
0,0,0,0,0,0,0,0,  # 58 - 5f
0,0,0,0,0,0,0,0,  # 60 - 67
0,0,0,0,0,0,0,0,  # 68 - 6f
0,0,0,0,0,0,0,0,  # 70 - 77
0,0,0,4,0,5,2,0,  # 78 - 7f
1,1,1,1,1,1,1,1,  # 80 - 87
1,1,1,1,1,1,1,1,  # 88 - 8f
1,1,1,1,1,1,1,1,  # 90 - 97
1,1,1,1,1,1,1,1,  # 98 - 9f
1,1,1,1,1,1,1,1,  # a0 - a7
1,1,1,1,1,1,1,1,  # a8 - af
1,1,1,1,1,1,1,1,  # b0 - b7
1,1,1,1,1,1,1,1,  # b8 - bf
1,1,1,1,1,1,1,1,  # c0 - c7
1,1,1,1,1,1,1,1,  # c8 - cf
1,1,1,1,1,1,1,1,  # d0 - d7
1,1,1,1,1,1,1,1,  # d8 - df
1,1,1,1,1,1,1,1,  # e0 - e7
1,1,1,1,1,1,1,1,  # e8 - ef
1,1,1,1,1,1,1,1,  # f0 - f7
1,1,1,1,1,1,1,1,  # f8 - ff
)

HZ_ST = (
MachineState.START,MachineState.ERROR,     3,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,# 00-07
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 08-0f
MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,     4,MachineState.ERROR,# 10-17
     5,MachineState.ERROR,     6,MachineState.ERROR,     5,     5,     4,MachineState.ERROR,# 18-1f
     4,MachineState.ERROR,     4,     4,     4,MachineState.ERROR,     4,MachineState.ERROR,# 20-27
     4,MachineState.ITS_ME,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 28-2f
)

HZ_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0)

HZ_SM_MODEL = {'class_table': HZ_CLS,
               'class_factor': 6,
               'state_table': HZ_ST,
               'char_len_table': HZ_CHAR_LEN_TABLE,
               'name': "HZ-GB-2312",
               'language': 'Chinese'}

ISO2022CN_CLS = (
2,0,0,0,0,0,0,0,  # 00 - 07
0,0,0,0,0,0,0,0,  # 08 - 0f
0,0,0,0,0,0,0,0,  # 10 - 17
0,0,0,1,0,0,0,0,  # 18 - 1f
0,0,0,0,0,0,0,0,  # 20 - 27
0,3,0,0,0,0,0,0,  # 28 - 2f
0,0,0,0,0,0,0,0,  # 30 - 37
0,0,0,0,0,0,0,0,  # 38 - 3f
0,0,0,4,0,0,0,0,  # 40 - 47
0,0,0,0,0,0,0,0,  # 48 - 4f
0,0,0,0,0,0,0,0,  # 50 - 57
0,0,0,0,0,0,0,0,  # 58 - 5f
0,0,0,0,0,0,0,0,  # 60 - 67
0,0,0,0,0,0,0,0,  # 68 - 6f
0,0,0,0,0,0,0,0,  # 70 - 77
0,0,0,0,0,0,0,0,  # 78 - 7f
2,2,2,2,2,2,2,2,  # 80 - 87
2,2,2,2,2,2,2,2,  # 88 - 8f
2,2,2,2,2,2,2,2,  # 90 - 97
2,2,2,2,2,2,2,2,  # 98 - 9f
2,2,2,2,2,2,2,2,  # a0 - a7
2,2,2,2,2,2,2,2,  # a8 - af
2,2,2,2,2,2,2,2,  # b0 - b7
2,2,2,2,2,2,2,2,  # b8 - bf
2,2,2,2,2,2,2,2,  # c0 - c7
2,2,2,2,2,2,2,2,  # c8 - cf
2,2,2,2,2,2,2,2,  # d0 - d7
2,2,2,2,2,2,2,2,  # d8 - df
2,2,2,2,2,2,2,2,  # e0 - e7
2,2,2,2,2,2,2,2,  # e8 - ef
2,2,2,2,2,2,2,2,  # f0 - f7
2,2,2,2,2,2,2,2,  # f8 - ff
)

ISO2022CN_ST = (
MachineState.START,     3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 00-07
MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 08-0f
MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 10-17
MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     4,MachineState.ERROR,# 18-1f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 20-27
     5,     6,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 28-2f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 30-37
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,# 38-3f
)

ISO2022CN_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0, 0, 0, 0)

ISO2022CN_SM_MODEL = {'class_table': ISO2022CN_CLS,
                      'class_factor': 9,
                      'state_table': ISO2022CN_ST,
                      'char_len_table': ISO2022CN_CHAR_LEN_TABLE,
                      'name': "ISO-2022-CN",
                      'language': 'Chinese'}

ISO2022JP_CLS = (
2,0,0,0,0,0,0,0,  # 00 - 07
0,0,0,0,0,0,2,2,  # 08 - 0f
0,0,0,0,0,0,0,0,  # 10 - 17
0,0,0,1,0,0,0,0,  # 18 - 1f
0,0,0,0,7,0,0,0,  # 20 - 27
3,0,0,0,0,0,0,0,  # 28 - 2f
0,0,0,0,0,0,0,0,  # 30 - 37
0,0,0,0,0,0,0,0,  # 38 - 3f
6,0,4,0,8,0,0,0,  # 40 - 47
0,9,5,0,0,0,0,0,  # 48 - 4f
0,0,0,0,0,0,0,0,  # 50 - 57
0,0,0,0,0,0,0,0,  # 58 - 5f
0,0,0,0,0,0,0,0,  # 60 - 67
0,0,0,0,0,0,0,0,  # 68 - 6f
0,0,0,0,0,0,0,0,  # 70 - 77
0,0,0,0,0,0,0,0,  # 78 - 7f
2,2,2,2,2,2,2,2,  # 80 - 87
2,2,2,2,2,2,2,2,  # 88 - 8f
2,2,2,2,2,2,2,2,  # 90 - 97
2,2,2,2,2,2,2,2,  # 98 - 9f
2,2,2,2,2,2,2,2,  # a0 - a7
2,2,2,2,2,2,2,2,  # a8 - af
2,2,2,2,2,2,2,2,  # b0 - b7
2,2,2,2,2,2,2,2,  # b8 - bf
2,2,2,2,2,2,2,2,  # c0 - c7
2,2,2,2,2,2,2,2,  # c8 - cf
2,2,2,2,2,2,2,2,  # d0 - d7
2,2,2,2,2,2,2,2,  # d8 - df
2,2,2,2,2,2,2,2,  # e0 - e7
2,2,2,2,2,2,2,2,  # e8 - ef
2,2,2,2,2,2,2,2,  # f0 - f7
2,2,2,2,2,2,2,2,  # f8 - ff
)

ISO2022JP_ST = (
MachineState.START,     3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 00-07
MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 08-0f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 10-17
MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,# 18-1f
MachineState.ERROR,     5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     4,MachineState.ERROR,MachineState.ERROR,# 20-27
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     6,MachineState.ITS_ME,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,# 28-2f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,# 30-37
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 38-3f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,MachineState.START,# 40-47
)

ISO2022JP_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0, 0, 0, 0, 0)

ISO2022JP_SM_MODEL = {'class_table': ISO2022JP_CLS,
                      'class_factor': 10,
                      'state_table': ISO2022JP_ST,
                      'char_len_table': ISO2022JP_CHAR_LEN_TABLE,
                      'name': "ISO-2022-JP",
                      'language': 'Japanese'}

ISO2022KR_CLS = (
2,0,0,0,0,0,0,0,  # 00 - 07
0,0,0,0,0,0,0,0,  # 08 - 0f
0,0,0,0,0,0,0,0,  # 10 - 17
0,0,0,1,0,0,0,0,  # 18 - 1f
0,0,0,0,3,0,0,0,  # 20 - 27
0,4,0,0,0,0,0,0,  # 28 - 2f
0,0,0,0,0,0,0,0,  # 30 - 37
0,0,0,0,0,0,0,0,  # 38 - 3f
0,0,0,5,0,0,0,0,  # 40 - 47
0,0,0,0,0,0,0,0,  # 48 - 4f
0,0,0,0,0,0,0,0,  # 50 - 57
0,0,0,0,0,0,0,0,  # 58 - 5f
0,0,0,0,0,0,0,0,  # 60 - 67
0,0,0,0,0,0,0,0,  # 68 - 6f
0,0,0,0,0,0,0,0,  # 70 - 77
0,0,0,0,0,0,0,0,  # 78 - 7f
2,2,2,2,2,2,2,2,  # 80 - 87
2,2,2,2,2,2,2,2,  # 88 - 8f
2,2,2,2,2,2,2,2,  # 90 - 97
2,2,2,2,2,2,2,2,  # 98 - 9f
2,2,2,2,2,2,2,2,  # a0 - a7
2,2,2,2,2,2,2,2,  # a8 - af
2,2,2,2,2,2,2,2,  # b0 - b7
2,2,2,2,2,2,2,2,  # b8 - bf
2,2,2,2,2,2,2,2,  # c0 - c7
2,2,2,2,2,2,2,2,  # c8 - cf
2,2,2,2,2,2,2,2,  # d0 - d7
2,2,2,2,2,2,2,2,  # d8 - df
2,2,2,2,2,2,2,2,  # e0 - e7
2,2,2,2,2,2,2,2,  # e8 - ef
2,2,2,2,2,2,2,2,  # f0 - f7
2,2,2,2,2,2,2,2,  # f8 - ff
)

ISO2022KR_ST = (
MachineState.START,     3,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,# 00-07
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,# 08-0f
MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     4,MachineState.ERROR,MachineState.ERROR,# 10-17
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,# 18-1f
MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.START,MachineState.START,MachineState.START,MachineState.START,# 20-27
)

ISO2022KR_CHAR_LEN_TABLE = (0, 0, 0, 0, 0, 0)

ISO2022KR_SM_MODEL = {'class_table': ISO2022KR_CLS,
                      'class_factor': 6,
                      'state_table': ISO2022KR_ST,
                      'char_len_table': ISO2022KR_CHAR_LEN_TABLE,
                      'name': "ISO-2022-KR",
                      'language': 'Korean'}


site-packages/pip/_vendor/chardet/mbcssm.py000064400000061611150564745010014774 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is mozilla.org code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

from .enums import MachineState

# BIG5

BIG5_CLS = (
    1,1,1,1,1,1,1,1,  # 00 - 07    #allow 0x00 as legal value
    1,1,1,1,1,1,0,0,  # 08 - 0f
    1,1,1,1,1,1,1,1,  # 10 - 17
    1,1,1,0,1,1,1,1,  # 18 - 1f
    1,1,1,1,1,1,1,1,  # 20 - 27
    1,1,1,1,1,1,1,1,  # 28 - 2f
    1,1,1,1,1,1,1,1,  # 30 - 37
    1,1,1,1,1,1,1,1,  # 38 - 3f
    2,2,2,2,2,2,2,2,  # 40 - 47
    2,2,2,2,2,2,2,2,  # 48 - 4f
    2,2,2,2,2,2,2,2,  # 50 - 57
    2,2,2,2,2,2,2,2,  # 58 - 5f
    2,2,2,2,2,2,2,2,  # 60 - 67
    2,2,2,2,2,2,2,2,  # 68 - 6f
    2,2,2,2,2,2,2,2,  # 70 - 77
    2,2,2,2,2,2,2,1,  # 78 - 7f
    4,4,4,4,4,4,4,4,  # 80 - 87
    4,4,4,4,4,4,4,4,  # 88 - 8f
    4,4,4,4,4,4,4,4,  # 90 - 97
    4,4,4,4,4,4,4,4,  # 98 - 9f
    4,3,3,3,3,3,3,3,  # a0 - a7
    3,3,3,3,3,3,3,3,  # a8 - af
    3,3,3,3,3,3,3,3,  # b0 - b7
    3,3,3,3,3,3,3,3,  # b8 - bf
    3,3,3,3,3,3,3,3,  # c0 - c7
    3,3,3,3,3,3,3,3,  # c8 - cf
    3,3,3,3,3,3,3,3,  # d0 - d7
    3,3,3,3,3,3,3,3,  # d8 - df
    3,3,3,3,3,3,3,3,  # e0 - e7
    3,3,3,3,3,3,3,3,  # e8 - ef
    3,3,3,3,3,3,3,3,  # f0 - f7
    3,3,3,3,3,3,3,0  # f8 - ff
)

BIG5_ST = (
    MachineState.ERROR,MachineState.START,MachineState.START,     3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07
    MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,#08-0f
    MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START#10-17
)

BIG5_CHAR_LEN_TABLE = (0, 1, 1, 2, 0)

BIG5_SM_MODEL = {'class_table': BIG5_CLS,
                 'class_factor': 5,
                 'state_table': BIG5_ST,
                 'char_len_table': BIG5_CHAR_LEN_TABLE,
                 'name': 'Big5'}

# CP949

CP949_CLS  = (
    1,1,1,1,1,1,1,1, 1,1,1,1,1,1,0,0,  # 00 - 0f
    1,1,1,1,1,1,1,1, 1,1,1,0,1,1,1,1,  # 10 - 1f
    1,1,1,1,1,1,1,1, 1,1,1,1,1,1,1,1,  # 20 - 2f
    1,1,1,1,1,1,1,1, 1,1,1,1,1,1,1,1,  # 30 - 3f
    1,4,4,4,4,4,4,4, 4,4,4,4,4,4,4,4,  # 40 - 4f
    4,4,5,5,5,5,5,5, 5,5,5,1,1,1,1,1,  # 50 - 5f
    1,5,5,5,5,5,5,5, 5,5,5,5,5,5,5,5,  # 60 - 6f
    5,5,5,5,5,5,5,5, 5,5,5,1,1,1,1,1,  # 70 - 7f
    0,6,6,6,6,6,6,6, 6,6,6,6,6,6,6,6,  # 80 - 8f
    6,6,6,6,6,6,6,6, 6,6,6,6,6,6,6,6,  # 90 - 9f
    6,7,7,7,7,7,7,7, 7,7,7,7,7,8,8,8,  # a0 - af
    7,7,7,7,7,7,7,7, 7,7,7,7,7,7,7,7,  # b0 - bf
    7,7,7,7,7,7,9,2, 2,3,2,2,2,2,2,2,  # c0 - cf
    2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,2,  # d0 - df
    2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,2,  # e0 - ef
    2,2,2,2,2,2,2,2, 2,2,2,2,2,2,2,0,  # f0 - ff
)

CP949_ST = (
#cls=    0      1      2      3      4      5      6      7      8      9  # previous state =
    MachineState.ERROR,MachineState.START,     3,MachineState.ERROR,MachineState.START,MachineState.START,     4,     5,MachineState.ERROR,     6, # MachineState.START
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR, # MachineState.ERROR
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME, # MachineState.ITS_ME
    MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START, # 3
    MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START, # 4
    MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START, # 5
    MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START, # 6
)

CP949_CHAR_LEN_TABLE = (0, 1, 2, 0, 1, 1, 2, 2, 0, 2)

CP949_SM_MODEL = {'class_table': CP949_CLS,
                  'class_factor': 10,
                  'state_table': CP949_ST,
                  'char_len_table': CP949_CHAR_LEN_TABLE,
                  'name': 'CP949'}

# EUC-JP

EUCJP_CLS = (
    4,4,4,4,4,4,4,4,  # 00 - 07
    4,4,4,4,4,4,5,5,  # 08 - 0f
    4,4,4,4,4,4,4,4,  # 10 - 17
    4,4,4,5,4,4,4,4,  # 18 - 1f
    4,4,4,4,4,4,4,4,  # 20 - 27
    4,4,4,4,4,4,4,4,  # 28 - 2f
    4,4,4,4,4,4,4,4,  # 30 - 37
    4,4,4,4,4,4,4,4,  # 38 - 3f
    4,4,4,4,4,4,4,4,  # 40 - 47
    4,4,4,4,4,4,4,4,  # 48 - 4f
    4,4,4,4,4,4,4,4,  # 50 - 57
    4,4,4,4,4,4,4,4,  # 58 - 5f
    4,4,4,4,4,4,4,4,  # 60 - 67
    4,4,4,4,4,4,4,4,  # 68 - 6f
    4,4,4,4,4,4,4,4,  # 70 - 77
    4,4,4,4,4,4,4,4,  # 78 - 7f
    5,5,5,5,5,5,5,5,  # 80 - 87
    5,5,5,5,5,5,1,3,  # 88 - 8f
    5,5,5,5,5,5,5,5,  # 90 - 97
    5,5,5,5,5,5,5,5,  # 98 - 9f
    5,2,2,2,2,2,2,2,  # a0 - a7
    2,2,2,2,2,2,2,2,  # a8 - af
    2,2,2,2,2,2,2,2,  # b0 - b7
    2,2,2,2,2,2,2,2,  # b8 - bf
    2,2,2,2,2,2,2,2,  # c0 - c7
    2,2,2,2,2,2,2,2,  # c8 - cf
    2,2,2,2,2,2,2,2,  # d0 - d7
    2,2,2,2,2,2,2,2,  # d8 - df
    0,0,0,0,0,0,0,0,  # e0 - e7
    0,0,0,0,0,0,0,0,  # e8 - ef
    0,0,0,0,0,0,0,0,  # f0 - f7
    0,0,0,0,0,0,0,5  # f8 - ff
)

EUCJP_ST = (
          3,     4,     3,     5,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07
     MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
     MachineState.ITS_ME,MachineState.ITS_ME,MachineState.START,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#10-17
     MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     3,MachineState.ERROR,#18-1f
          3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START#20-27
)

EUCJP_CHAR_LEN_TABLE = (2, 2, 2, 3, 1, 0)

EUCJP_SM_MODEL = {'class_table': EUCJP_CLS,
                  'class_factor': 6,
                  'state_table': EUCJP_ST,
                  'char_len_table': EUCJP_CHAR_LEN_TABLE,
                  'name': 'EUC-JP'}

# EUC-KR

EUCKR_CLS  = (
    1,1,1,1,1,1,1,1,  # 00 - 07
    1,1,1,1,1,1,0,0,  # 08 - 0f
    1,1,1,1,1,1,1,1,  # 10 - 17
    1,1,1,0,1,1,1,1,  # 18 - 1f
    1,1,1,1,1,1,1,1,  # 20 - 27
    1,1,1,1,1,1,1,1,  # 28 - 2f
    1,1,1,1,1,1,1,1,  # 30 - 37
    1,1,1,1,1,1,1,1,  # 38 - 3f
    1,1,1,1,1,1,1,1,  # 40 - 47
    1,1,1,1,1,1,1,1,  # 48 - 4f
    1,1,1,1,1,1,1,1,  # 50 - 57
    1,1,1,1,1,1,1,1,  # 58 - 5f
    1,1,1,1,1,1,1,1,  # 60 - 67
    1,1,1,1,1,1,1,1,  # 68 - 6f
    1,1,1,1,1,1,1,1,  # 70 - 77
    1,1,1,1,1,1,1,1,  # 78 - 7f
    0,0,0,0,0,0,0,0,  # 80 - 87
    0,0,0,0,0,0,0,0,  # 88 - 8f
    0,0,0,0,0,0,0,0,  # 90 - 97
    0,0,0,0,0,0,0,0,  # 98 - 9f
    0,2,2,2,2,2,2,2,  # a0 - a7
    2,2,2,2,2,3,3,3,  # a8 - af
    2,2,2,2,2,2,2,2,  # b0 - b7
    2,2,2,2,2,2,2,2,  # b8 - bf
    2,2,2,2,2,2,2,2,  # c0 - c7
    2,3,2,2,2,2,2,2,  # c8 - cf
    2,2,2,2,2,2,2,2,  # d0 - d7
    2,2,2,2,2,2,2,2,  # d8 - df
    2,2,2,2,2,2,2,2,  # e0 - e7
    2,2,2,2,2,2,2,2,  # e8 - ef
    2,2,2,2,2,2,2,2,  # f0 - f7
    2,2,2,2,2,2,2,0   # f8 - ff
)

EUCKR_ST = (
    MachineState.ERROR,MachineState.START,     3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START #08-0f
)

EUCKR_CHAR_LEN_TABLE = (0, 1, 2, 0)

EUCKR_SM_MODEL = {'class_table': EUCKR_CLS,
                'class_factor': 4,
                'state_table': EUCKR_ST,
                'char_len_table': EUCKR_CHAR_LEN_TABLE,
                'name': 'EUC-KR'}

# EUC-TW

EUCTW_CLS = (
    2,2,2,2,2,2,2,2,  # 00 - 07
    2,2,2,2,2,2,0,0,  # 08 - 0f
    2,2,2,2,2,2,2,2,  # 10 - 17
    2,2,2,0,2,2,2,2,  # 18 - 1f
    2,2,2,2,2,2,2,2,  # 20 - 27
    2,2,2,2,2,2,2,2,  # 28 - 2f
    2,2,2,2,2,2,2,2,  # 30 - 37
    2,2,2,2,2,2,2,2,  # 38 - 3f
    2,2,2,2,2,2,2,2,  # 40 - 47
    2,2,2,2,2,2,2,2,  # 48 - 4f
    2,2,2,2,2,2,2,2,  # 50 - 57
    2,2,2,2,2,2,2,2,  # 58 - 5f
    2,2,2,2,2,2,2,2,  # 60 - 67
    2,2,2,2,2,2,2,2,  # 68 - 6f
    2,2,2,2,2,2,2,2,  # 70 - 77
    2,2,2,2,2,2,2,2,  # 78 - 7f
    0,0,0,0,0,0,0,0,  # 80 - 87
    0,0,0,0,0,0,6,0,  # 88 - 8f
    0,0,0,0,0,0,0,0,  # 90 - 97
    0,0,0,0,0,0,0,0,  # 98 - 9f
    0,3,4,4,4,4,4,4,  # a0 - a7
    5,5,1,1,1,1,1,1,  # a8 - af
    1,1,1,1,1,1,1,1,  # b0 - b7
    1,1,1,1,1,1,1,1,  # b8 - bf
    1,1,3,1,3,3,3,3,  # c0 - c7
    3,3,3,3,3,3,3,3,  # c8 - cf
    3,3,3,3,3,3,3,3,  # d0 - d7
    3,3,3,3,3,3,3,3,  # d8 - df
    3,3,3,3,3,3,3,3,  # e0 - e7
    3,3,3,3,3,3,3,3,  # e8 - ef
    3,3,3,3,3,3,3,3,  # f0 - f7
    3,3,3,3,3,3,3,0   # f8 - ff
)

EUCTW_ST = (
    MachineState.ERROR,MachineState.ERROR,MachineState.START,     3,     3,     3,     4,MachineState.ERROR,#00-07
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.START,MachineState.ERROR,#10-17
    MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f
         5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.START,MachineState.START,#20-27
    MachineState.START,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START #28-2f
)

EUCTW_CHAR_LEN_TABLE = (0, 0, 1, 2, 2, 2, 3)

EUCTW_SM_MODEL = {'class_table': EUCTW_CLS,
                'class_factor': 7,
                'state_table': EUCTW_ST,
                'char_len_table': EUCTW_CHAR_LEN_TABLE,
                'name': 'x-euc-tw'}

# GB2312

GB2312_CLS = (
    1,1,1,1,1,1,1,1,  # 00 - 07
    1,1,1,1,1,1,0,0,  # 08 - 0f
    1,1,1,1,1,1,1,1,  # 10 - 17
    1,1,1,0,1,1,1,1,  # 18 - 1f
    1,1,1,1,1,1,1,1,  # 20 - 27
    1,1,1,1,1,1,1,1,  # 28 - 2f
    3,3,3,3,3,3,3,3,  # 30 - 37
    3,3,1,1,1,1,1,1,  # 38 - 3f
    2,2,2,2,2,2,2,2,  # 40 - 47
    2,2,2,2,2,2,2,2,  # 48 - 4f
    2,2,2,2,2,2,2,2,  # 50 - 57
    2,2,2,2,2,2,2,2,  # 58 - 5f
    2,2,2,2,2,2,2,2,  # 60 - 67
    2,2,2,2,2,2,2,2,  # 68 - 6f
    2,2,2,2,2,2,2,2,  # 70 - 77
    2,2,2,2,2,2,2,4,  # 78 - 7f
    5,6,6,6,6,6,6,6,  # 80 - 87
    6,6,6,6,6,6,6,6,  # 88 - 8f
    6,6,6,6,6,6,6,6,  # 90 - 97
    6,6,6,6,6,6,6,6,  # 98 - 9f
    6,6,6,6,6,6,6,6,  # a0 - a7
    6,6,6,6,6,6,6,6,  # a8 - af
    6,6,6,6,6,6,6,6,  # b0 - b7
    6,6,6,6,6,6,6,6,  # b8 - bf
    6,6,6,6,6,6,6,6,  # c0 - c7
    6,6,6,6,6,6,6,6,  # c8 - cf
    6,6,6,6,6,6,6,6,  # d0 - d7
    6,6,6,6,6,6,6,6,  # d8 - df
    6,6,6,6,6,6,6,6,  # e0 - e7
    6,6,6,6,6,6,6,6,  # e8 - ef
    6,6,6,6,6,6,6,6,  # f0 - f7
    6,6,6,6,6,6,6,0   # f8 - ff
)

GB2312_ST = (
    MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,     3,MachineState.ERROR,#00-07
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,#10-17
         4,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f
    MachineState.ERROR,MachineState.ERROR,     5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,#20-27
    MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.START #28-2f
)

# To be accurate, the length of class 6 can be either 2 or 4.
# But it is not necessary to discriminate between the two since
# it is used for frequency analysis only, and we are validating
# each code range there as well. So it is safe to set it to be
# 2 here.
GB2312_CHAR_LEN_TABLE = (0, 1, 1, 1, 1, 1, 2)

GB2312_SM_MODEL = {'class_table': GB2312_CLS,
                   'class_factor': 7,
                   'state_table': GB2312_ST,
                   'char_len_table': GB2312_CHAR_LEN_TABLE,
                   'name': 'GB2312'}

# Shift_JIS

SJIS_CLS = (
    1,1,1,1,1,1,1,1,  # 00 - 07
    1,1,1,1,1,1,0,0,  # 08 - 0f
    1,1,1,1,1,1,1,1,  # 10 - 17
    1,1,1,0,1,1,1,1,  # 18 - 1f
    1,1,1,1,1,1,1,1,  # 20 - 27
    1,1,1,1,1,1,1,1,  # 28 - 2f
    1,1,1,1,1,1,1,1,  # 30 - 37
    1,1,1,1,1,1,1,1,  # 38 - 3f
    2,2,2,2,2,2,2,2,  # 40 - 47
    2,2,2,2,2,2,2,2,  # 48 - 4f
    2,2,2,2,2,2,2,2,  # 50 - 57
    2,2,2,2,2,2,2,2,  # 58 - 5f
    2,2,2,2,2,2,2,2,  # 60 - 67
    2,2,2,2,2,2,2,2,  # 68 - 6f
    2,2,2,2,2,2,2,2,  # 70 - 77
    2,2,2,2,2,2,2,1,  # 78 - 7f
    3,3,3,3,3,2,2,3,  # 80 - 87
    3,3,3,3,3,3,3,3,  # 88 - 8f
    3,3,3,3,3,3,3,3,  # 90 - 97
    3,3,3,3,3,3,3,3,  # 98 - 9f
    #0xa0 is illegal in sjis encoding, but some pages does
    #contain such byte. We need to be more error forgiven.
    2,2,2,2,2,2,2,2,  # a0 - a7
    2,2,2,2,2,2,2,2,  # a8 - af
    2,2,2,2,2,2,2,2,  # b0 - b7
    2,2,2,2,2,2,2,2,  # b8 - bf
    2,2,2,2,2,2,2,2,  # c0 - c7
    2,2,2,2,2,2,2,2,  # c8 - cf
    2,2,2,2,2,2,2,2,  # d0 - d7
    2,2,2,2,2,2,2,2,  # d8 - df
    3,3,3,3,3,3,3,3,  # e0 - e7
    3,3,3,3,3,4,4,4,  # e8 - ef
    3,3,3,3,3,3,3,3,  # f0 - f7
    3,3,3,3,3,0,0,0)  # f8 - ff


SJIS_ST = (
    MachineState.ERROR,MachineState.START,MachineState.START,     3,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#00-07
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START #10-17
)

SJIS_CHAR_LEN_TABLE = (0, 1, 1, 2, 0, 0)

SJIS_SM_MODEL = {'class_table': SJIS_CLS,
               'class_factor': 6,
               'state_table': SJIS_ST,
               'char_len_table': SJIS_CHAR_LEN_TABLE,
               'name': 'Shift_JIS'}

# UCS2-BE

UCS2BE_CLS = (
    0,0,0,0,0,0,0,0,  # 00 - 07
    0,0,1,0,0,2,0,0,  # 08 - 0f
    0,0,0,0,0,0,0,0,  # 10 - 17
    0,0,0,3,0,0,0,0,  # 18 - 1f
    0,0,0,0,0,0,0,0,  # 20 - 27
    0,3,3,3,3,3,0,0,  # 28 - 2f
    0,0,0,0,0,0,0,0,  # 30 - 37
    0,0,0,0,0,0,0,0,  # 38 - 3f
    0,0,0,0,0,0,0,0,  # 40 - 47
    0,0,0,0,0,0,0,0,  # 48 - 4f
    0,0,0,0,0,0,0,0,  # 50 - 57
    0,0,0,0,0,0,0,0,  # 58 - 5f
    0,0,0,0,0,0,0,0,  # 60 - 67
    0,0,0,0,0,0,0,0,  # 68 - 6f
    0,0,0,0,0,0,0,0,  # 70 - 77
    0,0,0,0,0,0,0,0,  # 78 - 7f
    0,0,0,0,0,0,0,0,  # 80 - 87
    0,0,0,0,0,0,0,0,  # 88 - 8f
    0,0,0,0,0,0,0,0,  # 90 - 97
    0,0,0,0,0,0,0,0,  # 98 - 9f
    0,0,0,0,0,0,0,0,  # a0 - a7
    0,0,0,0,0,0,0,0,  # a8 - af
    0,0,0,0,0,0,0,0,  # b0 - b7
    0,0,0,0,0,0,0,0,  # b8 - bf
    0,0,0,0,0,0,0,0,  # c0 - c7
    0,0,0,0,0,0,0,0,  # c8 - cf
    0,0,0,0,0,0,0,0,  # d0 - d7
    0,0,0,0,0,0,0,0,  # d8 - df
    0,0,0,0,0,0,0,0,  # e0 - e7
    0,0,0,0,0,0,0,0,  # e8 - ef
    0,0,0,0,0,0,0,0,  # f0 - f7
    0,0,0,0,0,0,4,5   # f8 - ff
)

UCS2BE_ST  = (
          5,     7,     7,MachineState.ERROR,     4,     3,MachineState.ERROR,MachineState.ERROR,#00-07
     MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
     MachineState.ITS_ME,MachineState.ITS_ME,     6,     6,     6,     6,MachineState.ERROR,MachineState.ERROR,#10-17
          6,     6,     6,     6,     6,MachineState.ITS_ME,     6,     6,#18-1f
          6,     6,     6,     6,     5,     7,     7,MachineState.ERROR,#20-27
          5,     8,     6,     6,MachineState.ERROR,     6,     6,     6,#28-2f
          6,     6,     6,     6,MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START #30-37
)

UCS2BE_CHAR_LEN_TABLE = (2, 2, 2, 0, 2, 2)

UCS2BE_SM_MODEL = {'class_table': UCS2BE_CLS,
                   'class_factor': 6,
                   'state_table': UCS2BE_ST,
                   'char_len_table': UCS2BE_CHAR_LEN_TABLE,
                   'name': 'UTF-16BE'}

# UCS2-LE

UCS2LE_CLS = (
    0,0,0,0,0,0,0,0,  # 00 - 07
    0,0,1,0,0,2,0,0,  # 08 - 0f
    0,0,0,0,0,0,0,0,  # 10 - 17
    0,0,0,3,0,0,0,0,  # 18 - 1f
    0,0,0,0,0,0,0,0,  # 20 - 27
    0,3,3,3,3,3,0,0,  # 28 - 2f
    0,0,0,0,0,0,0,0,  # 30 - 37
    0,0,0,0,0,0,0,0,  # 38 - 3f
    0,0,0,0,0,0,0,0,  # 40 - 47
    0,0,0,0,0,0,0,0,  # 48 - 4f
    0,0,0,0,0,0,0,0,  # 50 - 57
    0,0,0,0,0,0,0,0,  # 58 - 5f
    0,0,0,0,0,0,0,0,  # 60 - 67
    0,0,0,0,0,0,0,0,  # 68 - 6f
    0,0,0,0,0,0,0,0,  # 70 - 77
    0,0,0,0,0,0,0,0,  # 78 - 7f
    0,0,0,0,0,0,0,0,  # 80 - 87
    0,0,0,0,0,0,0,0,  # 88 - 8f
    0,0,0,0,0,0,0,0,  # 90 - 97
    0,0,0,0,0,0,0,0,  # 98 - 9f
    0,0,0,0,0,0,0,0,  # a0 - a7
    0,0,0,0,0,0,0,0,  # a8 - af
    0,0,0,0,0,0,0,0,  # b0 - b7
    0,0,0,0,0,0,0,0,  # b8 - bf
    0,0,0,0,0,0,0,0,  # c0 - c7
    0,0,0,0,0,0,0,0,  # c8 - cf
    0,0,0,0,0,0,0,0,  # d0 - d7
    0,0,0,0,0,0,0,0,  # d8 - df
    0,0,0,0,0,0,0,0,  # e0 - e7
    0,0,0,0,0,0,0,0,  # e8 - ef
    0,0,0,0,0,0,0,0,  # f0 - f7
    0,0,0,0,0,0,4,5   # f8 - ff
)

UCS2LE_ST = (
          6,     6,     7,     6,     4,     3,MachineState.ERROR,MachineState.ERROR,#00-07
     MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#08-0f
     MachineState.ITS_ME,MachineState.ITS_ME,     5,     5,     5,MachineState.ERROR,MachineState.ITS_ME,MachineState.ERROR,#10-17
          5,     5,     5,MachineState.ERROR,     5,MachineState.ERROR,     6,     6,#18-1f
          7,     6,     8,     8,     5,     5,     5,MachineState.ERROR,#20-27
          5,     5,     5,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     5,     5,#28-2f
          5,     5,     5,MachineState.ERROR,     5,MachineState.ERROR,MachineState.START,MachineState.START #30-37
)

UCS2LE_CHAR_LEN_TABLE = (2, 2, 2, 2, 2, 2)

UCS2LE_SM_MODEL = {'class_table': UCS2LE_CLS,
                 'class_factor': 6,
                 'state_table': UCS2LE_ST,
                 'char_len_table': UCS2LE_CHAR_LEN_TABLE,
                 'name': 'UTF-16LE'}

# UTF-8

UTF8_CLS = (
    1,1,1,1,1,1,1,1,  # 00 - 07  #allow 0x00 as a legal value
    1,1,1,1,1,1,0,0,  # 08 - 0f
    1,1,1,1,1,1,1,1,  # 10 - 17
    1,1,1,0,1,1,1,1,  # 18 - 1f
    1,1,1,1,1,1,1,1,  # 20 - 27
    1,1,1,1,1,1,1,1,  # 28 - 2f
    1,1,1,1,1,1,1,1,  # 30 - 37
    1,1,1,1,1,1,1,1,  # 38 - 3f
    1,1,1,1,1,1,1,1,  # 40 - 47
    1,1,1,1,1,1,1,1,  # 48 - 4f
    1,1,1,1,1,1,1,1,  # 50 - 57
    1,1,1,1,1,1,1,1,  # 58 - 5f
    1,1,1,1,1,1,1,1,  # 60 - 67
    1,1,1,1,1,1,1,1,  # 68 - 6f
    1,1,1,1,1,1,1,1,  # 70 - 77
    1,1,1,1,1,1,1,1,  # 78 - 7f
    2,2,2,2,3,3,3,3,  # 80 - 87
    4,4,4,4,4,4,4,4,  # 88 - 8f
    4,4,4,4,4,4,4,4,  # 90 - 97
    4,4,4,4,4,4,4,4,  # 98 - 9f
    5,5,5,5,5,5,5,5,  # a0 - a7
    5,5,5,5,5,5,5,5,  # a8 - af
    5,5,5,5,5,5,5,5,  # b0 - b7
    5,5,5,5,5,5,5,5,  # b8 - bf
    0,0,6,6,6,6,6,6,  # c0 - c7
    6,6,6,6,6,6,6,6,  # c8 - cf
    6,6,6,6,6,6,6,6,  # d0 - d7
    6,6,6,6,6,6,6,6,  # d8 - df
    7,8,8,8,8,8,8,8,  # e0 - e7
    8,8,8,8,8,9,8,8,  # e8 - ef
    10,11,11,11,11,11,11,11,  # f0 - f7
    12,13,13,13,14,15,0,0    # f8 - ff
)

UTF8_ST = (
    MachineState.ERROR,MachineState.START,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     12,   10,#00-07
         9,     11,     8,     7,     6,     5,     4,    3,#08-0f
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#10-17
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#18-1f
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#20-27
    MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,MachineState.ITS_ME,#28-2f
    MachineState.ERROR,MachineState.ERROR,     5,     5,     5,     5,MachineState.ERROR,MachineState.ERROR,#30-37
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#38-3f
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     5,     5,     5,MachineState.ERROR,MachineState.ERROR,#40-47
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#48-4f
    MachineState.ERROR,MachineState.ERROR,     7,     7,     7,     7,MachineState.ERROR,MachineState.ERROR,#50-57
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#58-5f
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     7,     7,MachineState.ERROR,MachineState.ERROR,#60-67
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#68-6f
    MachineState.ERROR,MachineState.ERROR,     9,     9,     9,     9,MachineState.ERROR,MachineState.ERROR,#70-77
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#78-7f
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,     9,MachineState.ERROR,MachineState.ERROR,#80-87
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#88-8f
    MachineState.ERROR,MachineState.ERROR,    12,    12,    12,    12,MachineState.ERROR,MachineState.ERROR,#90-97
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#98-9f
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,    12,MachineState.ERROR,MachineState.ERROR,#a0-a7
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#a8-af
    MachineState.ERROR,MachineState.ERROR,    12,    12,    12,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#b0-b7
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,#b8-bf
    MachineState.ERROR,MachineState.ERROR,MachineState.START,MachineState.START,MachineState.START,MachineState.START,MachineState.ERROR,MachineState.ERROR,#c0-c7
    MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR,MachineState.ERROR #c8-cf
)

UTF8_CHAR_LEN_TABLE = (0, 1, 0, 0, 0, 0, 2, 3, 3, 3, 4, 4, 5, 5, 6, 6)

UTF8_SM_MODEL = {'class_table': UTF8_CLS,
                 'class_factor': 16,
                 'state_table': UTF8_ST,
                 'char_len_table': UTF8_CHAR_LEN_TABLE,
                 'name': 'UTF-8'}
site-packages/pip/_vendor/chardet/gb2312freq.py000064400000050353150564745010015267 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

# GB2312 most frequently used character table
#
# Char to FreqOrder table , from hz6763

# 512  --> 0.79  -- 0.79
# 1024 --> 0.92  -- 0.13
# 2048 --> 0.98  -- 0.06
# 6768 --> 1.00  -- 0.02
#
# Ideal Distribution Ratio = 0.79135/(1-0.79135) = 3.79
# Random Distribution Ration = 512 / (3755 - 512) = 0.157
#
# Typical Distribution Ratio about 25% of Ideal one, still much higher that RDR

GB2312_TYPICAL_DISTRIBUTION_RATIO = 0.9

GB2312_TABLE_SIZE = 3760

GB2312_CHAR_TO_FREQ_ORDER = (
1671, 749,1443,2364,3924,3807,2330,3921,1704,3463,2691,1511,1515, 572,3191,2205,
2361, 224,2558, 479,1711, 963,3162, 440,4060,1905,2966,2947,3580,2647,3961,3842,
2204, 869,4207, 970,2678,5626,2944,2956,1479,4048, 514,3595, 588,1346,2820,3409,
 249,4088,1746,1873,2047,1774, 581,1813, 358,1174,3590,1014,1561,4844,2245, 670,
1636,3112, 889,1286, 953, 556,2327,3060,1290,3141, 613, 185,3477,1367, 850,3820,
1715,2428,2642,2303,2732,3041,2562,2648,3566,3946,1349, 388,3098,2091,1360,3585,
 152,1687,1539, 738,1559,  59,1232,2925,2267,1388,1249,1741,1679,2960, 151,1566,
1125,1352,4271, 924,4296, 385,3166,4459, 310,1245,2850,  70,3285,2729,3534,3575,
2398,3298,3466,1960,2265, 217,3647, 864,1909,2084,4401,2773,1010,3269,5152, 853,
3051,3121,1244,4251,1895, 364,1499,1540,2313,1180,3655,2268, 562, 715,2417,3061,
 544, 336,3768,2380,1752,4075, 950, 280,2425,4382, 183,2759,3272, 333,4297,2155,
1688,2356,1444,1039,4540, 736,1177,3349,2443,2368,2144,2225, 565, 196,1482,3406,
 927,1335,4147, 692, 878,1311,1653,3911,3622,1378,4200,1840,2969,3149,2126,1816,
2534,1546,2393,2760, 737,2494,  13, 447, 245,2747,  38,2765,2129,2589,1079, 606,
 360, 471,3755,2890, 404, 848, 699,1785,1236, 370,2221,1023,3746,2074,2026,2023,
2388,1581,2119, 812,1141,3091,2536,1519, 804,2053, 406,1596,1090, 784, 548,4414,
1806,2264,2936,1100, 343,4114,5096, 622,3358, 743,3668,1510,1626,5020,3567,2513,
3195,4115,5627,2489,2991,  24,2065,2697,1087,2719,  48,1634, 315,  68, 985,2052,
 198,2239,1347,1107,1439, 597,2366,2172, 871,3307, 919,2487,2790,1867, 236,2570,
1413,3794, 906,3365,3381,1701,1982,1818,1524,2924,1205, 616,2586,2072,2004, 575,
 253,3099,  32,1365,1182, 197,1714,2454,1201, 554,3388,3224,2748, 756,2587, 250,
2567,1507,1517,3529,1922,2761,2337,3416,1961,1677,2452,2238,3153, 615, 911,1506,
1474,2495,1265,1906,2749,3756,3280,2161, 898,2714,1759,3450,2243,2444, 563,  26,
3286,2266,3769,3344,2707,3677, 611,1402, 531,1028,2871,4548,1375, 261,2948, 835,
1190,4134, 353, 840,2684,1900,3082,1435,2109,1207,1674, 329,1872,2781,4055,2686,
2104, 608,3318,2423,2957,2768,1108,3739,3512,3271,3985,2203,1771,3520,1418,2054,
1681,1153, 225,1627,2929, 162,2050,2511,3687,1954, 124,1859,2431,1684,3032,2894,
 585,4805,3969,2869,2704,2088,2032,2095,3656,2635,4362,2209, 256, 518,2042,2105,
3777,3657, 643,2298,1148,1779, 190, 989,3544, 414,  11,2135,2063,2979,1471, 403,
3678, 126, 770,1563, 671,2499,3216,2877, 600,1179, 307,2805,4937,1268,1297,2694,
 252,4032,1448,1494,1331,1394, 127,2256, 222,1647,1035,1481,3056,1915,1048, 873,
3651, 210,  33,1608,2516, 200,1520, 415, 102,   0,3389,1287, 817,  91,3299,2940,
 836,1814, 549,2197,1396,1669,2987,3582,2297,2848,4528,1070, 687,  20,1819, 121,
1552,1364,1461,1968,2617,3540,2824,2083, 177, 948,4938,2291, 110,4549,2066, 648,
3359,1755,2110,2114,4642,4845,1693,3937,3308,1257,1869,2123, 208,1804,3159,2992,
2531,2549,3361,2418,1350,2347,2800,2568,1291,2036,2680,  72, 842,1990, 212,1233,
1154,1586,  75,2027,3410,4900,1823,1337,2710,2676, 728,2810,1522,3026,4995, 157,
 755,1050,4022, 710, 785,1936,2194,2085,1406,2777,2400, 150,1250,4049,1206, 807,
1910, 534, 529,3309,1721,1660, 274,  39,2827, 661,2670,1578, 925,3248,3815,1094,
4278,4901,4252,  41,1150,3747,2572,2227,4501,3658,4902,3813,3357,3617,2884,2258,
 887, 538,4187,3199,1294,2439,3042,2329,2343,2497,1255, 107, 543,1527, 521,3478,
3568, 194,5062,  15, 961,3870,1241,1192,2664,  66,5215,3260,2111,1295,1127,2152,
3805,4135, 901,1164,1976, 398,1278, 530,1460, 748, 904,1054,1966,1426,  53,2909,
 509, 523,2279,1534, 536,1019, 239,1685, 460,2353, 673,1065,2401,3600,4298,2272,
1272,2363, 284,1753,3679,4064,1695,  81, 815,2677,2757,2731,1386, 859, 500,4221,
2190,2566, 757,1006,2519,2068,1166,1455, 337,2654,3203,1863,1682,1914,3025,1252,
1409,1366, 847, 714,2834,2038,3209, 964,2970,1901, 885,2553,1078,1756,3049, 301,
1572,3326, 688,2130,1996,2429,1805,1648,2930,3421,2750,3652,3088, 262,1158,1254,
 389,1641,1812, 526,1719, 923,2073,1073,1902, 468, 489,4625,1140, 857,2375,3070,
3319,2863, 380, 116,1328,2693,1161,2244, 273,1212,1884,2769,3011,1775,1142, 461,
3066,1200,2147,2212, 790, 702,2695,4222,1601,1058, 434,2338,5153,3640,  67,2360,
4099,2502, 618,3472,1329, 416,1132, 830,2782,1807,2653,3211,3510,1662, 192,2124,
 296,3979,1739,1611,3684,  23, 118, 324, 446,1239,1225, 293,2520,3814,3795,2535,
3116,  17,1074, 467,2692,2201, 387,2922,  45,1326,3055,1645,3659,2817, 958, 243,
1903,2320,1339,2825,1784,3289, 356, 576, 865,2315,2381,3377,3916,1088,3122,1713,
1655, 935, 628,4689,1034,1327, 441, 800, 720, 894,1979,2183,1528,5289,2702,1071,
4046,3572,2399,1571,3281,  79, 761,1103, 327, 134, 758,1899,1371,1615, 879, 442,
 215,2605,2579, 173,2048,2485,1057,2975,3317,1097,2253,3801,4263,1403,1650,2946,
 814,4968,3487,1548,2644,1567,1285,   2, 295,2636,  97, 946,3576, 832, 141,4257,
3273, 760,3821,3521,3156,2607, 949,1024,1733,1516,1803,1920,2125,2283,2665,3180,
1501,2064,3560,2171,1592, 803,3518,1416, 732,3897,4258,1363,1362,2458, 119,1427,
 602,1525,2608,1605,1639,3175, 694,3064,  10, 465,  76,2000,4846,4208, 444,3781,
1619,3353,2206,1273,3796, 740,2483, 320,1723,2377,3660,2619,1359,1137,1762,1724,
2345,2842,1850,1862, 912, 821,1866, 612,2625,1735,2573,3369,1093, 844,  89, 937,
 930,1424,3564,2413,2972,1004,3046,3019,2011, 711,3171,1452,4178, 428, 801,1943,
 432, 445,2811, 206,4136,1472, 730, 349,  73, 397,2802,2547, 998,1637,1167, 789,
 396,3217, 154,1218, 716,1120,1780,2819,4826,1931,3334,3762,2139,1215,2627, 552,
3664,3628,3232,1405,2383,3111,1356,2652,3577,3320,3101,1703, 640,1045,1370,1246,
4996, 371,1575,2436,1621,2210, 984,4033,1734,2638,  16,4529, 663,2755,3255,1451,
3917,2257,1253,1955,2234,1263,2951, 214,1229, 617, 485, 359,1831,1969, 473,2310,
 750,2058, 165,  80,2864,2419, 361,4344,2416,2479,1134, 796,3726,1266,2943, 860,
2715, 938, 390,2734,1313,1384, 248, 202, 877,1064,2854, 522,3907, 279,1602, 297,
2357, 395,3740, 137,2075, 944,4089,2584,1267,3802,  62,1533,2285, 178, 176, 780,
2440, 201,3707, 590, 478,1560,4354,2117,1075,  30,  74,4643,4004,1635,1441,2745,
 776,2596, 238,1077,1692,1912,2844, 605, 499,1742,3947, 241,3053, 980,1749, 936,
2640,4511,2582, 515,1543,2162,5322,2892,2993, 890,2148,1924, 665,1827,3581,1032,
 968,3163, 339,1044,1896, 270, 583,1791,1720,4367,1194,3488,3669,  43,2523,1657,
 163,2167, 290,1209,1622,3378, 550, 634,2508,2510, 695,2634,2384,2512,1476,1414,
 220,1469,2341,2138,2852,3183,2900,4939,2865,3502,1211,3680, 854,3227,1299,2976,
3172, 186,2998,1459, 443,1067,3251,1495, 321,1932,3054, 909, 753,1410,1828, 436,
2441,1119,1587,3164,2186,1258, 227, 231,1425,1890,3200,3942, 247, 959, 725,5254,
2741, 577,2158,2079, 929, 120, 174, 838,2813, 591,1115, 417,2024,  40,3240,1536,
1037, 291,4151,2354, 632,1298,2406,2500,3535,1825,1846,3451, 205,1171, 345,4238,
  18,1163, 811, 685,2208,1217, 425,1312,1508,1175,4308,2552,1033, 587,1381,3059,
2984,3482, 340,1316,4023,3972, 792,3176, 519, 777,4690, 918, 933,4130,2981,3741,
  90,3360,2911,2200,5184,4550, 609,3079,2030, 272,3379,2736, 363,3881,1130,1447,
 286, 779, 357,1169,3350,3137,1630,1220,2687,2391, 747,1277,3688,2618,2682,2601,
1156,3196,5290,4034,3102,1689,3596,3128, 874, 219,2783, 798, 508,1843,2461, 269,
1658,1776,1392,1913,2983,3287,2866,2159,2372, 829,4076,  46,4253,2873,1889,1894,
 915,1834,1631,2181,2318, 298, 664,2818,3555,2735, 954,3228,3117, 527,3511,2173,
 681,2712,3033,2247,2346,3467,1652, 155,2164,3382, 113,1994, 450, 899, 494, 994,
1237,2958,1875,2336,1926,3727, 545,1577,1550, 633,3473, 204,1305,3072,2410,1956,
2471, 707,2134, 841,2195,2196,2663,3843,1026,4940, 990,3252,4997, 368,1092, 437,
3212,3258,1933,1829, 675,2977,2893, 412, 943,3723,4644,3294,3283,2230,2373,5154,
2389,2241,2661,2323,1404,2524, 593, 787, 677,3008,1275,2059, 438,2709,2609,2240,
2269,2246,1446,  36,1568,1373,3892,1574,2301,1456,3962, 693,2276,5216,2035,1143,
2720,1919,1797,1811,2763,4137,2597,1830,1699,1488,1198,2090, 424,1694, 312,3634,
3390,4179,3335,2252,1214, 561,1059,3243,2295,2561, 975,5155,2321,2751,3772, 472,
1537,3282,3398,1047,2077,2348,2878,1323,3340,3076, 690,2906,  51, 369, 170,3541,
1060,2187,2688,3670,2541,1083,1683, 928,3918, 459, 109,4427, 599,3744,4286, 143,
2101,2730,2490,  82,1588,3036,2121, 281,1860, 477,4035,1238,2812,3020,2716,3312,
1530,2188,2055,1317, 843, 636,1808,1173,3495, 649, 181,1002, 147,3641,1159,2414,
3750,2289,2795, 813,3123,2610,1136,4368,   5,3391,4541,2174, 420, 429,1728, 754,
1228,2115,2219, 347,2223,2733, 735,1518,3003,2355,3134,1764,3948,3329,1888,2424,
1001,1234,1972,3321,3363,1672,1021,1450,1584, 226, 765, 655,2526,3404,3244,2302,
3665, 731, 594,2184, 319,1576, 621, 658,2656,4299,2099,3864,1279,2071,2598,2739,
 795,3086,3699,3908,1707,2352,2402,1382,3136,2475,1465,4847,3496,3865,1085,3004,
2591,1084, 213,2287,1963,3565,2250, 822, 793,4574,3187,1772,1789,3050, 595,1484,
1959,2770,1080,2650, 456, 422,2996, 940,3322,4328,4345,3092,2742, 965,2784, 739,
4124, 952,1358,2498,2949,2565, 332,2698,2378, 660,2260,2473,4194,3856,2919, 535,
1260,2651,1208,1428,1300,1949,1303,2942, 433,2455,2450,1251,1946, 614,1269, 641,
1306,1810,2737,3078,2912, 564,2365,1419,1415,1497,4460,2367,2185,1379,3005,1307,
3218,2175,1897,3063, 682,1157,4040,4005,1712,1160,1941,1399, 394, 402,2952,1573,
1151,2986,2404, 862, 299,2033,1489,3006, 346, 171,2886,3401,1726,2932, 168,2533,
  47,2507,1030,3735,1145,3370,1395,1318,1579,3609,4560,2857,4116,1457,2529,1965,
 504,1036,2690,2988,2405, 745,5871, 849,2397,2056,3081, 863,2359,3857,2096,  99,
1397,1769,2300,4428,1643,3455,1978,1757,3718,1440,  35,4879,3742,1296,4228,2280,
 160,5063,1599,2013, 166, 520,3479,1646,3345,3012, 490,1937,1545,1264,2182,2505,
1096,1188,1369,1436,2421,1667,2792,2460,1270,2122, 727,3167,2143, 806,1706,1012,
1800,3037, 960,2218,1882, 805, 139,2456,1139,1521, 851,1052,3093,3089, 342,2039,
 744,5097,1468,1502,1585,2087, 223, 939, 326,2140,2577, 892,2481,1623,4077, 982,
3708, 135,2131,  87,2503,3114,2326,1106, 876,1616, 547,2997,2831,2093,3441,4530,
4314,   9,3256,4229,4148, 659,1462,1986,1710,2046,2913,2231,4090,4880,5255,3392,
3274,1368,3689,4645,1477, 705,3384,3635,1068,1529,2941,1458,3782,1509, 100,1656,
2548, 718,2339, 408,1590,2780,3548,1838,4117,3719,1345,3530, 717,3442,2778,3220,
2898,1892,4590,3614,3371,2043,1998,1224,3483, 891, 635, 584,2559,3355, 733,1766,
1729,1172,3789,1891,2307, 781,2982,2271,1957,1580,5773,2633,2005,4195,3097,1535,
3213,1189,1934,5693,3262, 586,3118,1324,1598, 517,1564,2217,1868,1893,4445,3728,
2703,3139,1526,1787,1992,3882,2875,1549,1199,1056,2224,1904,2711,5098,4287, 338,
1993,3129,3489,2689,1809,2815,1997, 957,1855,3898,2550,3275,3057,1105,1319, 627,
1505,1911,1883,3526, 698,3629,3456,1833,1431, 746,  77,1261,2017,2296,1977,1885,
 125,1334,1600, 525,1798,1109,2222,1470,1945, 559,2236,1186,3443,2476,1929,1411,
2411,3135,1777,3372,2621,1841,1613,3229, 668,1430,1839,2643,2916, 195,1989,2671,
2358,1387, 629,3205,2293,5256,4439, 123,1310, 888,1879,4300,3021,3605,1003,1162,
3192,2910,2010, 140,2395,2859,  55,1082,2012,2901, 662, 419,2081,1438, 680,2774,
4654,3912,1620,1731,1625,5035,4065,2328, 512,1344, 802,5443,2163,2311,2537, 524,
3399,  98,1155,2103,1918,2606,3925,2816,1393,2465,1504,3773,2177,3963,1478,4346,
 180,1113,4655,3461,2028,1698, 833,2696,1235,1322,1594,4408,3623,3013,3225,2040,
3022, 541,2881, 607,3632,2029,1665,1219, 639,1385,1686,1099,2803,3231,1938,3188,
2858, 427, 676,2772,1168,2025, 454,3253,2486,3556, 230,1950, 580, 791,1991,1280,
1086,1974,2034, 630, 257,3338,2788,4903,1017,  86,4790, 966,2789,1995,1696,1131,
 259,3095,4188,1308, 179,1463,5257, 289,4107,1248,  42,3413,1725,2288, 896,1947,
 774,4474,4254, 604,3430,4264, 392,2514,2588, 452, 237,1408,3018, 988,4531,1970,
3034,3310, 540,2370,1562,1288,2990, 502,4765,1147,   4,1853,2708, 207, 294,2814,
4078,2902,2509, 684,  34,3105,3532,2551, 644, 709,2801,2344, 573,1727,3573,3557,
2021,1081,3100,4315,2100,3681, 199,2263,1837,2385, 146,3484,1195,2776,3949, 997,
1939,3973,1008,1091,1202,1962,1847,1149,4209,5444,1076, 493, 117,5400,2521, 972,
1490,2934,1796,4542,2374,1512,2933,2657, 413,2888,1135,2762,2314,2156,1355,2369,
 766,2007,2527,2170,3124,2491,2593,2632,4757,2437, 234,3125,3591,1898,1750,1376,
1942,3468,3138, 570,2127,2145,3276,4131, 962, 132,1445,4196,  19, 941,3624,3480,
3366,1973,1374,4461,3431,2629, 283,2415,2275, 808,2887,3620,2112,2563,1353,3610,
 955,1089,3103,1053,  96,  88,4097, 823,3808,1583, 399, 292,4091,3313, 421,1128,
 642,4006, 903,2539,1877,2082, 596,  29,4066,1790, 722,2157, 130, 995,1569, 769,
1485, 464, 513,2213, 288,1923,1101,2453,4316, 133, 486,2445,  50, 625, 487,2207,
  57, 423, 481,2962, 159,3729,1558, 491, 303, 482, 501, 240,2837, 112,3648,2392,
1783, 362,   8,3433,3422, 610,2793,3277,1390,1284,1654,  21,3823, 734, 367, 623,
 193, 287, 374,1009,1483, 816, 476, 313,2255,2340,1262,2150,2899,1146,2581, 782,
2116,1659,2018,1880, 255,3586,3314,1110,2867,2137,2564, 986,2767,5185,2006, 650,
 158, 926, 762, 881,3157,2717,2362,3587, 306,3690,3245,1542,3077,2427,1691,2478,
2118,2985,3490,2438, 539,2305, 983, 129,1754, 355,4201,2386, 827,2923, 104,1773,
2838,2771, 411,2905,3919, 376, 767, 122,1114, 828,2422,1817,3506, 266,3460,1007,
1609,4998, 945,2612,4429,2274, 726,1247,1964,2914,2199,2070,4002,4108, 657,3323,
1422, 579, 455,2764,4737,1222,2895,1670, 824,1223,1487,2525, 558, 861,3080, 598,
2659,2515,1967, 752,2583,2376,2214,4180, 977, 704,2464,4999,2622,4109,1210,2961,
 819,1541, 142,2284,  44, 418, 457,1126,3730,4347,4626,1644,1876,3671,1864, 302,
1063,5694, 624, 723,1984,3745,1314,1676,2488,1610,1449,3558,3569,2166,2098, 409,
1011,2325,3704,2306, 818,1732,1383,1824,1844,3757, 999,2705,3497,1216,1423,2683,
2426,2954,2501,2726,2229,1475,2554,5064,1971,1794,1666,2014,1343, 783, 724, 191,
2434,1354,2220,5065,1763,2752,2472,4152, 131, 175,2885,3434,  92,1466,4920,2616,
3871,3872,3866, 128,1551,1632, 669,1854,3682,4691,4125,1230, 188,2973,3290,1302,
1213, 560,3266, 917, 763,3909,3249,1760, 868,1958, 764,1782,2097, 145,2277,3774,
4462,  64,1491,3062, 971,2132,3606,2442, 221,1226,1617, 218, 323,1185,3207,3147,
 571, 619,1473,1005,1744,2281, 449,1887,2396,3685, 275, 375,3816,1743,3844,3731,
 845,1983,2350,4210,1377, 773, 967,3499,3052,3743,2725,4007,1697,1022,3943,1464,
3264,2855,2722,1952,1029,2839,2467,  84,4383,2215, 820,1391,2015,2448,3672, 377,
1948,2168, 797,2545,3536,2578,2645,  94,2874,1678, 405,1259,3071, 771, 546,1315,
 470,1243,3083, 895,2468, 981, 969,2037, 846,4181, 653,1276,2928,  14,2594, 557,
3007,2474, 156, 902,1338,1740,2574, 537,2518, 973,2282,2216,2433,1928, 138,2903,
1293,2631,1612, 646,3457, 839,2935, 111, 496,2191,2847, 589,3186, 149,3994,2060,
4031,2641,4067,3145,1870,  37,3597,2136,1025,2051,3009,3383,3549,1121,1016,3261,
1301, 251,2446,2599,2153, 872,3246, 637, 334,3705, 831, 884, 921,3065,3140,4092,
2198,1944, 246,2964, 108,2045,1152,1921,2308,1031, 203,3173,4170,1907,3890, 810,
1401,2003,1690, 506, 647,1242,2828,1761,1649,3208,2249,1589,3709,2931,5156,1708,
 498, 666,2613, 834,3817,1231, 184,2851,1124, 883,3197,2261,3710,1765,1553,2658,
1178,2639,2351,  93,1193, 942,2538,2141,4402, 235,1821, 870,1591,2192,1709,1871,
3341,1618,4126,2595,2334, 603, 651,  69, 701, 268,2662,3411,2555,1380,1606, 503,
 448, 254,2371,2646, 574,1187,2309,1770, 322,2235,1292,1801, 305, 566,1133, 229,
2067,2057, 706, 167, 483,2002,2672,3295,1820,3561,3067, 316, 378,2746,3452,1112,
 136,1981, 507,1651,2917,1117, 285,4591, 182,2580,3522,1304, 335,3303,1835,2504,
1795,1792,2248, 674,1018,2106,2449,1857,2292,2845, 976,3047,1781,2600,2727,1389,
1281,  52,3152, 153, 265,3950, 672,3485,3951,4463, 430,1183, 365, 278,2169,  27,
1407,1336,2304, 209,1340,1730,2202,1852,2403,2883, 979,1737,1062, 631,2829,2542,
3876,2592, 825,2086,2226,3048,3625, 352,1417,3724, 542, 991, 431,1351,3938,1861,
2294, 826,1361,2927,3142,3503,1738, 463,2462,2723, 582,1916,1595,2808, 400,3845,
3891,2868,3621,2254,  58,2492,1123, 910,2160,2614,1372,1603,1196,1072,3385,1700,
3267,1980, 696, 480,2430, 920, 799,1570,2920,1951,2041,4047,2540,1321,4223,2469,
3562,2228,1271,2602, 401,2833,3351,2575,5157, 907,2312,1256, 410, 263,3507,1582,
 996, 678,1849,2316,1480, 908,3545,2237, 703,2322, 667,1826,2849,1531,2604,2999,
2407,3146,2151,2630,1786,3711, 469,3542, 497,3899,2409, 858, 837,4446,3393,1274,
 786, 620,1845,2001,3311, 484, 308,3367,1204,1815,3691,2332,1532,2557,1842,2020,
2724,1927,2333,4440, 567,  22,1673,2728,4475,1987,1858,1144,1597, 101,1832,3601,
  12, 974,3783,4391, 951,1412,   1,3720, 453,4608,4041, 528,1041,1027,3230,2628,
1129, 875,1051,3291,1203,2262,1069,2860,2799,2149,2615,3278, 144,1758,3040,  31,
 475,1680, 366,2685,3184, 311,1642,4008,2466,5036,1593,1493,2809, 216,1420,1668,
 233, 304,2128,3284, 232,1429,1768,1040,2008,3407,2740,2967,2543, 242,2133, 778,
1565,2022,2620, 505,2189,2756,1098,2273, 372,1614, 708, 553,2846,2094,2278, 169,
3626,2835,4161, 228,2674,3165, 809,1454,1309, 466,1705,1095, 900,3423, 880,2667,
3751,5258,2317,3109,2571,4317,2766,1503,1342, 866,4447,1118,  63,2076, 314,1881,
1348,1061, 172, 978,3515,1747, 532, 511,3970,   6, 601, 905,2699,3300,1751, 276,
1467,3725,2668,  65,4239,2544,2779,2556,1604, 578,2451,1802, 992,2331,2624,1320,
3446, 713,1513,1013, 103,2786,2447,1661, 886,1702, 916, 654,3574,2031,1556, 751,
2178,2821,2179,1498,1538,2176, 271, 914,2251,2080,1325, 638,1953,2937,3877,2432,
2754,  95,3265,1716, 260,1227,4083, 775, 106,1357,3254, 426,1607, 555,2480, 772,
1985, 244,2546, 474, 495,1046,2611,1851,2061,  71,2089,1675,2590, 742,3758,2843,
3222,1433, 267,2180,2576,2826,2233,2092,3913,2435, 956,1745,3075, 856,2113,1116,
 451,   3,1988,2896,1398, 993,2463,1878,2049,1341,2718,2721,2870,2108, 712,2904,
4363,2753,2324, 277,2872,2349,2649, 384, 987, 435, 691,3000, 922, 164,3939, 652,
1500,1184,4153,2482,3373,2165,4848,2335,3775,3508,3154,2806,2830,1554,2102,1664,
2530,1434,2408, 893,1547,2623,3447,2832,2242,2532,3169,2856,3223,2078,  49,3770,
3469, 462, 318, 656,2259,3250,3069, 679,1629,2758, 344,1138,1104,3120,1836,1283,
3115,2154,1437,4448, 934, 759,1999, 794,2862,1038, 533,2560,1722,2342, 855,2626,
1197,1663,4476,3127,  85,4240,2528,  25,1111,1181,3673, 407,3470,4561,2679,2713,
 768,1925,2841,3986,1544,1165, 932, 373,1240,2146,1930,2673, 721,4766, 354,4333,
 391,2963, 187,  61,3364,1442,1102, 330,1940,1767, 341,3809,4118, 393,2496,2062,
2211, 105, 331, 300, 439, 913,1332, 626, 379,3304,1557, 328, 689,3952, 309,1555,
 931, 317,2517,3027, 325, 569, 686,2107,3084,  60,1042,1333,2794, 264,3177,4014,
1628, 258,3712,   7,4464,1176,1043,1778, 683, 114,1975,  78,1492, 383,1886, 510,
 386, 645,5291,2891,2069,3305,4138,3867,2939,2603,2493,1935,1066,1848,3588,1015,
1282,1289,4609, 697,1453,3044,2666,3611,1856,2412,  54, 719,1330, 568,3778,2459,
1748, 788, 492, 551,1191,1000, 488,3394,3763, 282,1799, 348,2016,1523,3155,2390,
1049, 382,2019,1788,1170, 729,2968,3523, 897,3926,2785,2938,3292, 350,2319,3238,
1718,1717,2655,3453,3143,4465, 161,2889,2980,2009,1421,  56,1908,1640,2387,2232,
1917,1874,2477,4921, 148,  83,3438, 592,4245,2882,1822,1055, 741, 115,1496,1624,
 381,1638,4592,1020, 516,3214, 458, 947,4575,1432, 211,1514,2926,1865,2142, 189,
 852,1221,1400,1486, 882,2299,4036, 351,  28,1122, 700,6479,6480,6481,6482,6483,  #last 512
)

site-packages/pip/_vendor/chardet/jpcntx.py000064400000046273150564745010015025 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################


# This is hiragana 2-char sequence table, the number in each cell represents its frequency category
jp2CharContext = (
(0,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,0,0,0,1),
(2,4,0,4,0,3,0,4,0,3,4,4,4,2,4,3,3,4,3,2,3,3,4,2,3,3,3,2,4,1,4,3,3,1,5,4,3,4,3,4,3,5,3,0,3,5,4,2,0,3,1,0,3,3,0,3,3,0,1,1,0,4,3,0,3,3,0,4,0,2,0,3,5,5,5,5,4,0,4,1,0,3,4),
(0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2),
(0,4,0,5,0,5,0,4,0,4,5,4,4,3,5,3,5,1,5,3,4,3,4,4,3,4,3,3,4,3,5,4,4,3,5,5,3,5,5,5,3,5,5,3,4,5,5,3,1,3,2,0,3,4,0,4,2,0,4,2,1,5,3,2,3,5,0,4,0,2,0,5,4,4,5,4,5,0,4,0,0,4,4),
(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0),
(0,3,0,4,0,3,0,3,0,4,5,4,3,3,3,3,4,3,5,4,4,3,5,4,4,3,4,3,4,4,4,4,5,3,4,4,3,4,5,5,4,5,5,1,4,5,4,3,0,3,3,1,3,3,0,4,4,0,3,3,1,5,3,3,3,5,0,4,0,3,0,4,4,3,4,3,3,0,4,1,1,3,4),
(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0),
(0,4,0,3,0,3,0,4,0,3,4,4,3,2,2,1,2,1,3,1,3,3,3,3,3,4,3,1,3,3,5,3,3,0,4,3,0,5,4,3,3,5,4,4,3,4,4,5,0,1,2,0,1,2,0,2,2,0,1,0,0,5,2,2,1,4,0,3,0,1,0,4,4,3,5,4,3,0,2,1,0,4,3),
(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0),
(0,3,0,5,0,4,0,2,1,4,4,2,4,1,4,2,4,2,4,3,3,3,4,3,3,3,3,1,4,2,3,3,3,1,4,4,1,1,1,4,3,3,2,0,2,4,3,2,0,3,3,0,3,1,1,0,0,0,3,3,0,4,2,2,3,4,0,4,0,3,0,4,4,5,3,4,4,0,3,0,0,1,4),
(1,4,0,4,0,4,0,4,0,3,5,4,4,3,4,3,5,4,3,3,4,3,5,4,4,4,4,3,4,2,4,3,3,1,5,4,3,2,4,5,4,5,5,4,4,5,4,4,0,3,2,2,3,3,0,4,3,1,3,2,1,4,3,3,4,5,0,3,0,2,0,4,5,5,4,5,4,0,4,0,0,5,4),
(0,5,0,5,0,4,0,3,0,4,4,3,4,3,3,3,4,0,4,4,4,3,4,3,4,3,3,1,4,2,4,3,4,0,5,4,1,4,5,4,4,5,3,2,4,3,4,3,2,4,1,3,3,3,2,3,2,0,4,3,3,4,3,3,3,4,0,4,0,3,0,4,5,4,4,4,3,0,4,1,0,1,3),
(0,3,1,4,0,3,0,2,0,3,4,4,3,1,4,2,3,3,4,3,4,3,4,3,4,4,3,2,3,1,5,4,4,1,4,4,3,5,4,4,3,5,5,4,3,4,4,3,1,2,3,1,2,2,0,3,2,0,3,1,0,5,3,3,3,4,3,3,3,3,4,4,4,4,5,4,2,0,3,3,2,4,3),
(0,2,0,3,0,1,0,1,0,0,3,2,0,0,2,0,1,0,2,1,3,3,3,1,2,3,1,0,1,0,4,2,1,1,3,3,0,4,3,3,1,4,3,3,0,3,3,2,0,0,0,0,1,0,0,2,0,0,0,0,0,4,1,0,2,3,2,2,2,1,3,3,3,4,4,3,2,0,3,1,0,3,3),
(0,4,0,4,0,3,0,3,0,4,4,4,3,3,3,3,3,3,4,3,4,2,4,3,4,3,3,2,4,3,4,5,4,1,4,5,3,5,4,5,3,5,4,0,3,5,5,3,1,3,3,2,2,3,0,3,4,1,3,3,2,4,3,3,3,4,0,4,0,3,0,4,5,4,4,5,3,0,4,1,0,3,4),
(0,2,0,3,0,3,0,0,0,2,2,2,1,0,1,0,0,0,3,0,3,0,3,0,1,3,1,0,3,1,3,3,3,1,3,3,3,0,1,3,1,3,4,0,0,3,1,1,0,3,2,0,0,0,0,1,3,0,1,0,0,3,3,2,0,3,0,0,0,0,0,3,4,3,4,3,3,0,3,0,0,2,3),
(2,3,0,3,0,2,0,1,0,3,3,4,3,1,3,1,1,1,3,1,4,3,4,3,3,3,0,0,3,1,5,4,3,1,4,3,2,5,5,4,4,4,4,3,3,4,4,4,0,2,1,1,3,2,0,1,2,0,0,1,0,4,1,3,3,3,0,3,0,1,0,4,4,4,5,5,3,0,2,0,0,4,4),
(0,2,0,1,0,3,1,3,0,2,3,3,3,0,3,1,0,0,3,0,3,2,3,1,3,2,1,1,0,0,4,2,1,0,2,3,1,4,3,2,0,4,4,3,1,3,1,3,0,1,0,0,1,0,0,0,1,0,0,0,0,4,1,1,1,2,0,3,0,0,0,3,4,2,4,3,2,0,1,0,0,3,3),
(0,1,0,4,0,5,0,4,0,2,4,4,2,3,3,2,3,3,5,3,3,3,4,3,4,2,3,0,4,3,3,3,4,1,4,3,2,1,5,5,3,4,5,1,3,5,4,2,0,3,3,0,1,3,0,4,2,0,1,3,1,4,3,3,3,3,0,3,0,1,0,3,4,4,4,5,5,0,3,0,1,4,5),
(0,2,0,3,0,3,0,0,0,2,3,1,3,0,4,0,1,1,3,0,3,4,3,2,3,1,0,3,3,2,3,1,3,0,2,3,0,2,1,4,1,2,2,0,0,3,3,0,0,2,0,0,0,1,0,0,0,0,2,2,0,3,2,1,3,3,0,2,0,2,0,0,3,3,1,2,4,0,3,0,2,2,3),
(2,4,0,5,0,4,0,4,0,2,4,4,4,3,4,3,3,3,1,2,4,3,4,3,4,4,5,0,3,3,3,3,2,0,4,3,1,4,3,4,1,4,4,3,3,4,4,3,1,2,3,0,4,2,0,4,1,0,3,3,0,4,3,3,3,4,0,4,0,2,0,3,5,3,4,5,2,0,3,0,0,4,5),
(0,3,0,4,0,1,0,1,0,1,3,2,2,1,3,0,3,0,2,0,2,0,3,0,2,0,0,0,1,0,1,1,0,0,3,1,0,0,0,4,0,3,1,0,2,1,3,0,0,0,0,0,0,3,0,0,0,0,0,0,0,4,2,2,3,1,0,3,0,0,0,1,4,4,4,3,0,0,4,0,0,1,4),
(1,4,1,5,0,3,0,3,0,4,5,4,4,3,5,3,3,4,4,3,4,1,3,3,3,3,2,1,4,1,5,4,3,1,4,4,3,5,4,4,3,5,4,3,3,4,4,4,0,3,3,1,2,3,0,3,1,0,3,3,0,5,4,4,4,4,4,4,3,3,5,4,4,3,3,5,4,0,3,2,0,4,4),
(0,2,0,3,0,1,0,0,0,1,3,3,3,2,4,1,3,0,3,1,3,0,2,2,1,1,0,0,2,0,4,3,1,0,4,3,0,4,4,4,1,4,3,1,1,3,3,1,0,2,0,0,1,3,0,0,0,0,2,0,0,4,3,2,4,3,5,4,3,3,3,4,3,3,4,3,3,0,2,1,0,3,3),
(0,2,0,4,0,3,0,2,0,2,5,5,3,4,4,4,4,1,4,3,3,0,4,3,4,3,1,3,3,2,4,3,0,3,4,3,0,3,4,4,2,4,4,0,4,5,3,3,2,2,1,1,1,2,0,1,5,0,3,3,2,4,3,3,3,4,0,3,0,2,0,4,4,3,5,5,0,0,3,0,2,3,3),
(0,3,0,4,0,3,0,1,0,3,4,3,3,1,3,3,3,0,3,1,3,0,4,3,3,1,1,0,3,0,3,3,0,0,4,4,0,1,5,4,3,3,5,0,3,3,4,3,0,2,0,1,1,1,0,1,3,0,1,2,1,3,3,2,3,3,0,3,0,1,0,1,3,3,4,4,1,0,1,2,2,1,3),
(0,1,0,4,0,4,0,3,0,1,3,3,3,2,3,1,1,0,3,0,3,3,4,3,2,4,2,0,1,0,4,3,2,0,4,3,0,5,3,3,2,4,4,4,3,3,3,4,0,1,3,0,0,1,0,0,1,0,0,0,0,4,2,3,3,3,0,3,0,0,0,4,4,4,5,3,2,0,3,3,0,3,5),
(0,2,0,3,0,0,0,3,0,1,3,0,2,0,0,0,1,0,3,1,1,3,3,0,0,3,0,0,3,0,2,3,1,0,3,1,0,3,3,2,0,4,2,2,0,2,0,0,0,4,0,0,0,0,0,0,0,0,0,0,0,2,1,2,0,1,0,1,0,0,0,1,3,1,2,0,0,0,1,0,0,1,4),
(0,3,0,3,0,5,0,1,0,2,4,3,1,3,3,2,1,1,5,2,1,0,5,1,2,0,0,0,3,3,2,2,3,2,4,3,0,0,3,3,1,3,3,0,2,5,3,4,0,3,3,0,1,2,0,2,2,0,3,2,0,2,2,3,3,3,0,2,0,1,0,3,4,4,2,5,4,0,3,0,0,3,5),
(0,3,0,3,0,3,0,1,0,3,3,3,3,0,3,0,2,0,2,1,1,0,2,0,1,0,0,0,2,1,0,0,1,0,3,2,0,0,3,3,1,2,3,1,0,3,3,0,0,1,0,0,0,0,0,2,0,0,0,0,0,2,3,1,2,3,0,3,0,1,0,3,2,1,0,4,3,0,1,1,0,3,3),
(0,4,0,5,0,3,0,3,0,4,5,5,4,3,5,3,4,3,5,3,3,2,5,3,4,4,4,3,4,3,4,5,5,3,4,4,3,4,4,5,4,4,4,3,4,5,5,4,2,3,4,2,3,4,0,3,3,1,4,3,2,4,3,3,5,5,0,3,0,3,0,5,5,5,5,4,4,0,4,0,1,4,4),
(0,4,0,4,0,3,0,3,0,3,5,4,4,2,3,2,5,1,3,2,5,1,4,2,3,2,3,3,4,3,3,3,3,2,5,4,1,3,3,5,3,4,4,0,4,4,3,1,1,3,1,0,2,3,0,2,3,0,3,0,0,4,3,1,3,4,0,3,0,2,0,4,4,4,3,4,5,0,4,0,0,3,4),
(0,3,0,3,0,3,1,2,0,3,4,4,3,3,3,0,2,2,4,3,3,1,3,3,3,1,1,0,3,1,4,3,2,3,4,4,2,4,4,4,3,4,4,3,2,4,4,3,1,3,3,1,3,3,0,4,1,0,2,2,1,4,3,2,3,3,5,4,3,3,5,4,4,3,3,0,4,0,3,2,2,4,4),
(0,2,0,1,0,0,0,0,0,1,2,1,3,0,0,0,0,0,2,0,1,2,1,0,0,1,0,0,0,0,3,0,0,1,0,1,1,3,1,0,0,0,1,1,0,1,1,0,0,0,0,0,2,0,0,0,0,0,0,0,0,1,1,2,2,0,3,4,0,0,0,1,1,0,0,1,0,0,0,0,0,1,1),
(0,1,0,0,0,1,0,0,0,0,4,0,4,1,4,0,3,0,4,0,3,0,4,0,3,0,3,0,4,1,5,1,4,0,0,3,0,5,0,5,2,0,1,0,0,0,2,1,4,0,1,3,0,0,3,0,0,3,1,1,4,1,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0),
(1,4,0,5,0,3,0,2,0,3,5,4,4,3,4,3,5,3,4,3,3,0,4,3,3,3,3,3,3,2,4,4,3,1,3,4,4,5,4,4,3,4,4,1,3,5,4,3,3,3,1,2,2,3,3,1,3,1,3,3,3,5,3,3,4,5,0,3,0,3,0,3,4,3,4,4,3,0,3,0,2,4,3),
(0,1,0,4,0,0,0,0,0,1,4,0,4,1,4,2,4,0,3,0,1,0,1,0,0,0,0,0,2,0,3,1,1,1,0,3,0,0,0,1,2,1,0,0,1,1,1,1,0,1,0,0,0,1,0,0,3,0,0,0,0,3,2,0,2,2,0,1,0,0,0,2,3,2,3,3,0,0,0,0,2,1,0),
(0,5,1,5,0,3,0,3,0,5,4,4,5,1,5,3,3,0,4,3,4,3,5,3,4,3,3,2,4,3,4,3,3,0,3,3,1,4,4,3,4,4,4,3,4,5,5,3,2,3,1,1,3,3,1,3,1,1,3,3,2,4,5,3,3,5,0,4,0,3,0,4,4,3,5,3,3,0,3,4,0,4,3),
(0,5,0,5,0,3,0,2,0,4,4,3,5,2,4,3,3,3,4,4,4,3,5,3,5,3,3,1,4,0,4,3,3,0,3,3,0,4,4,4,4,5,4,3,3,5,5,3,2,3,1,2,3,2,0,1,0,0,3,2,2,4,4,3,1,5,0,4,0,3,0,4,3,1,3,2,1,0,3,3,0,3,3),
(0,4,0,5,0,5,0,4,0,4,5,5,5,3,4,3,3,2,5,4,4,3,5,3,5,3,4,0,4,3,4,4,3,2,4,4,3,4,5,4,4,5,5,0,3,5,5,4,1,3,3,2,3,3,1,3,1,0,4,3,1,4,4,3,4,5,0,4,0,2,0,4,3,4,4,3,3,0,4,0,0,5,5),
(0,4,0,4,0,5,0,1,1,3,3,4,4,3,4,1,3,0,5,1,3,0,3,1,3,1,1,0,3,0,3,3,4,0,4,3,0,4,4,4,3,4,4,0,3,5,4,1,0,3,0,0,2,3,0,3,1,0,3,1,0,3,2,1,3,5,0,3,0,1,0,3,2,3,3,4,4,0,2,2,0,4,4),
(2,4,0,5,0,4,0,3,0,4,5,5,4,3,5,3,5,3,5,3,5,2,5,3,4,3,3,4,3,4,5,3,2,1,5,4,3,2,3,4,5,3,4,1,2,5,4,3,0,3,3,0,3,2,0,2,3,0,4,1,0,3,4,3,3,5,0,3,0,1,0,4,5,5,5,4,3,0,4,2,0,3,5),
(0,5,0,4,0,4,0,2,0,5,4,3,4,3,4,3,3,3,4,3,4,2,5,3,5,3,4,1,4,3,4,4,4,0,3,5,0,4,4,4,4,5,3,1,3,4,5,3,3,3,3,3,3,3,0,2,2,0,3,3,2,4,3,3,3,5,3,4,1,3,3,5,3,2,0,0,0,0,4,3,1,3,3),
(0,1,0,3,0,3,0,1,0,1,3,3,3,2,3,3,3,0,3,0,0,0,3,1,3,0,0,0,2,2,2,3,0,0,3,2,0,1,2,4,1,3,3,0,0,3,3,3,0,1,0,0,2,1,0,0,3,0,3,1,0,3,0,0,1,3,0,2,0,1,0,3,3,1,3,3,0,0,1,1,0,3,3),
(0,2,0,3,0,2,1,4,0,2,2,3,1,1,3,1,1,0,2,0,3,1,2,3,1,3,0,0,1,0,4,3,2,3,3,3,1,4,2,3,3,3,3,1,0,3,1,4,0,1,1,0,1,2,0,1,1,0,1,1,0,3,1,3,2,2,0,1,0,0,0,2,3,3,3,1,0,0,0,0,0,2,3),
(0,5,0,4,0,5,0,2,0,4,5,5,3,3,4,3,3,1,5,4,4,2,4,4,4,3,4,2,4,3,5,5,4,3,3,4,3,3,5,5,4,5,5,1,3,4,5,3,1,4,3,1,3,3,0,3,3,1,4,3,1,4,5,3,3,5,0,4,0,3,0,5,3,3,1,4,3,0,4,0,1,5,3),
(0,5,0,5,0,4,0,2,0,4,4,3,4,3,3,3,3,3,5,4,4,4,4,4,4,5,3,3,5,2,4,4,4,3,4,4,3,3,4,4,5,5,3,3,4,3,4,3,3,4,3,3,3,3,1,2,2,1,4,3,3,5,4,4,3,4,0,4,0,3,0,4,4,4,4,4,1,0,4,2,0,2,4),
(0,4,0,4,0,3,0,1,0,3,5,2,3,0,3,0,2,1,4,2,3,3,4,1,4,3,3,2,4,1,3,3,3,0,3,3,0,0,3,3,3,5,3,3,3,3,3,2,0,2,0,0,2,0,0,2,0,0,1,0,0,3,1,2,2,3,0,3,0,2,0,4,4,3,3,4,1,0,3,0,0,2,4),
(0,0,0,4,0,0,0,0,0,0,1,0,1,0,2,0,0,0,0,0,1,0,2,0,1,0,0,0,0,0,3,1,3,0,3,2,0,0,0,1,0,3,2,0,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,4,0,2,0,0,0,0,0,0,2),
(0,2,1,3,0,2,0,2,0,3,3,3,3,1,3,1,3,3,3,3,3,3,4,2,2,1,2,1,4,0,4,3,1,3,3,3,2,4,3,5,4,3,3,3,3,3,3,3,0,1,3,0,2,0,0,1,0,0,1,0,0,4,2,0,2,3,0,3,3,0,3,3,4,2,3,1,4,0,1,2,0,2,3),
(0,3,0,3,0,1,0,3,0,2,3,3,3,0,3,1,2,0,3,3,2,3,3,2,3,2,3,1,3,0,4,3,2,0,3,3,1,4,3,3,2,3,4,3,1,3,3,1,1,0,1,1,0,1,0,1,0,1,0,0,0,4,1,1,0,3,0,3,1,0,2,3,3,3,3,3,1,0,0,2,0,3,3),
(0,0,0,0,0,0,0,0,0,0,3,0,2,0,3,0,0,0,0,0,0,0,3,0,0,0,0,0,0,0,3,0,3,0,3,1,0,1,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,2,0,2,3,0,0,0,0,0,0,0,0,3),
(0,2,0,3,1,3,0,3,0,2,3,3,3,1,3,1,3,1,3,1,3,3,3,1,3,0,2,3,1,1,4,3,3,2,3,3,1,2,2,4,1,3,3,0,1,4,2,3,0,1,3,0,3,0,0,1,3,0,2,0,0,3,3,2,1,3,0,3,0,2,0,3,4,4,4,3,1,0,3,0,0,3,3),
(0,2,0,1,0,2,0,0,0,1,3,2,2,1,3,0,1,1,3,0,3,2,3,1,2,0,2,0,1,1,3,3,3,0,3,3,1,1,2,3,2,3,3,1,2,3,2,0,0,1,0,0,0,0,0,0,3,0,1,0,0,2,1,2,1,3,0,3,0,0,0,3,4,4,4,3,2,0,2,0,0,2,4),
(0,0,0,1,0,1,0,0,0,0,1,0,0,0,1,0,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,2,2,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,1,3,1,0,0,0,0,0,0,0,3),
(0,3,0,3,0,2,0,3,0,3,3,3,2,3,2,2,2,0,3,1,3,3,3,2,3,3,0,0,3,0,3,2,2,0,2,3,1,4,3,4,3,3,2,3,1,5,4,4,0,3,1,2,1,3,0,3,1,1,2,0,2,3,1,3,1,3,0,3,0,1,0,3,3,4,4,2,1,0,2,1,0,2,4),
(0,1,0,3,0,1,0,2,0,1,4,2,5,1,4,0,2,0,2,1,3,1,4,0,2,1,0,0,2,1,4,1,1,0,3,3,0,5,1,3,2,3,3,1,0,3,2,3,0,1,0,0,0,0,0,0,1,0,0,0,0,4,0,1,0,3,0,2,0,1,0,3,3,3,4,3,3,0,0,0,0,2,3),
(0,0,0,1,0,0,0,0,0,0,2,0,1,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,2,1,0,0,1,0,0,0,0,0,3),
(0,1,0,3,0,4,0,3,0,2,4,3,1,0,3,2,2,1,3,1,2,2,3,1,1,1,2,1,3,0,1,2,0,1,3,2,1,3,0,5,5,1,0,0,1,3,2,1,0,3,0,0,1,0,0,0,0,0,3,4,0,1,1,1,3,2,0,2,0,1,0,2,3,3,1,2,3,0,1,0,1,0,4),
(0,0,0,1,0,3,0,3,0,2,2,1,0,0,4,0,3,0,3,1,3,0,3,0,3,0,1,0,3,0,3,1,3,0,3,3,0,0,1,2,1,1,1,0,1,2,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,2,2,1,2,0,0,2,0,0,0,0,2,3,3,3,3,0,0,0,0,1,4),
(0,0,0,3,0,3,0,0,0,0,3,1,1,0,3,0,1,0,2,0,1,0,0,0,0,0,0,0,1,0,3,0,2,0,2,3,0,0,2,2,3,1,2,0,0,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,2,0,0,0,0,2,3),
(2,4,0,5,0,5,0,4,0,3,4,3,3,3,4,3,3,3,4,3,4,4,5,4,5,5,5,2,3,0,5,5,4,1,5,4,3,1,5,4,3,4,4,3,3,4,3,3,0,3,2,0,2,3,0,3,0,0,3,3,0,5,3,2,3,3,0,3,0,3,0,3,4,5,4,5,3,0,4,3,0,3,4),
(0,3,0,3,0,3,0,3,0,3,3,4,3,2,3,2,3,0,4,3,3,3,3,3,3,3,3,0,3,2,4,3,3,1,3,4,3,4,4,4,3,4,4,3,2,4,4,1,0,2,0,0,1,1,0,2,0,0,3,1,0,5,3,2,1,3,0,3,0,1,2,4,3,2,4,3,3,0,3,2,0,4,4),
(0,3,0,3,0,1,0,0,0,1,4,3,3,2,3,1,3,1,4,2,3,2,4,2,3,4,3,0,2,2,3,3,3,0,3,3,3,0,3,4,1,3,3,0,3,4,3,3,0,1,1,0,1,0,0,0,4,0,3,0,0,3,1,2,1,3,0,4,0,1,0,4,3,3,4,3,3,0,2,0,0,3,3),
(0,3,0,4,0,1,0,3,0,3,4,3,3,0,3,3,3,1,3,1,3,3,4,3,3,3,0,0,3,1,5,3,3,1,3,3,2,5,4,3,3,4,5,3,2,5,3,4,0,1,0,0,0,0,0,2,0,0,1,1,0,4,2,2,1,3,0,3,0,2,0,4,4,3,5,3,2,0,1,1,0,3,4),
(0,5,0,4,0,5,0,2,0,4,4,3,3,2,3,3,3,1,4,3,4,1,5,3,4,3,4,0,4,2,4,3,4,1,5,4,0,4,4,4,4,5,4,1,3,5,4,2,1,4,1,1,3,2,0,3,1,0,3,2,1,4,3,3,3,4,0,4,0,3,0,4,4,4,3,3,3,0,4,2,0,3,4),
(1,4,0,4,0,3,0,1,0,3,3,3,1,1,3,3,2,2,3,3,1,0,3,2,2,1,2,0,3,1,2,1,2,0,3,2,0,2,2,3,3,4,3,0,3,3,1,2,0,1,1,3,1,2,0,0,3,0,1,1,0,3,2,2,3,3,0,3,0,0,0,2,3,3,4,3,3,0,1,0,0,1,4),
(0,4,0,4,0,4,0,0,0,3,4,4,3,1,4,2,3,2,3,3,3,1,4,3,4,0,3,0,4,2,3,3,2,2,5,4,2,1,3,4,3,4,3,1,3,3,4,2,0,2,1,0,3,3,0,0,2,0,3,1,0,4,4,3,4,3,0,4,0,1,0,2,4,4,4,4,4,0,3,2,0,3,3),
(0,0,0,1,0,4,0,0,0,0,0,0,1,1,1,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,1,0,3,2,0,0,1,0,0,0,1,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,2),
(0,2,0,3,0,4,0,4,0,1,3,3,3,0,4,0,2,1,2,1,1,1,2,0,3,1,1,0,1,0,3,1,0,0,3,3,2,0,1,1,0,0,0,0,0,1,0,2,0,2,2,0,3,1,0,0,1,0,1,1,0,1,2,0,3,0,0,0,0,1,0,0,3,3,4,3,1,0,1,0,3,0,2),
(0,0,0,3,0,5,0,0,0,0,1,0,2,0,3,1,0,1,3,0,0,0,2,0,0,0,1,0,0,0,1,1,0,0,4,0,0,0,2,3,0,1,4,1,0,2,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,0,0,0,0,1,0,0,0,0,0,0,0,2,0,0,3,0,0,0,0,0,3),
(0,2,0,5,0,5,0,1,0,2,4,3,3,2,5,1,3,2,3,3,3,0,4,1,2,0,3,0,4,0,2,2,1,1,5,3,0,0,1,4,2,3,2,0,3,3,3,2,0,2,4,1,1,2,0,1,1,0,3,1,0,1,3,1,2,3,0,2,0,0,0,1,3,5,4,4,4,0,3,0,0,1,3),
(0,4,0,5,0,4,0,4,0,4,5,4,3,3,4,3,3,3,4,3,4,4,5,3,4,5,4,2,4,2,3,4,3,1,4,4,1,3,5,4,4,5,5,4,4,5,5,5,2,3,3,1,4,3,1,3,3,0,3,3,1,4,3,4,4,4,0,3,0,4,0,3,3,4,4,5,0,0,4,3,0,4,5),
(0,4,0,4,0,3,0,3,0,3,4,4,4,3,3,2,4,3,4,3,4,3,5,3,4,3,2,1,4,2,4,4,3,1,3,4,2,4,5,5,3,4,5,4,1,5,4,3,0,3,2,2,3,2,1,3,1,0,3,3,3,5,3,3,3,5,4,4,2,3,3,4,3,3,3,2,1,0,3,2,1,4,3),
(0,4,0,5,0,4,0,3,0,3,5,5,3,2,4,3,4,0,5,4,4,1,4,4,4,3,3,3,4,3,5,5,2,3,3,4,1,2,5,5,3,5,5,2,3,5,5,4,0,3,2,0,3,3,1,1,5,1,4,1,0,4,3,2,3,5,0,4,0,3,0,5,4,3,4,3,0,0,4,1,0,4,4),
(1,3,0,4,0,2,0,2,0,2,5,5,3,3,3,3,3,0,4,2,3,4,4,4,3,4,0,0,3,4,5,4,3,3,3,3,2,5,5,4,5,5,5,4,3,5,5,5,1,3,1,0,1,0,0,3,2,0,4,2,0,5,2,3,2,4,1,3,0,3,0,4,5,4,5,4,3,0,4,2,0,5,4),
(0,3,0,4,0,5,0,3,0,3,4,4,3,2,3,2,3,3,3,3,3,2,4,3,3,2,2,0,3,3,3,3,3,1,3,3,3,0,4,4,3,4,4,1,1,4,4,2,0,3,1,0,1,1,0,4,1,0,2,3,1,3,3,1,3,4,0,3,0,1,0,3,1,3,0,0,1,0,2,0,0,4,4),
(0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0),
(0,3,0,3,0,2,0,3,0,1,5,4,3,3,3,1,4,2,1,2,3,4,4,2,4,4,5,0,3,1,4,3,4,0,4,3,3,3,2,3,2,5,3,4,3,2,2,3,0,0,3,0,2,1,0,1,2,0,0,0,0,2,1,1,3,1,0,2,0,4,0,3,4,4,4,5,2,0,2,0,0,1,3),
(0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0,0,1,1,1,0,0,1,1,0,0,0,4,2,1,1,0,1,0,3,2,0,0,3,1,1,1,2,2,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,0,1,0,0,0,2,0,0,0,1,4,0,4,2,1,0,0,0,0,0,1),
(0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,1,0,0,0,0,0,0,1,0,1,0,0,0,0,3,1,0,0,0,2,0,2,1,0,0,1,2,1,0,1,1,0,0,3,0,0,0,0,0,0,0,0,0,0,0,1,3,1,0,0,0,0,0,1,0,0,2,1,0,0,0,0,0,0,0,0,2),
(0,4,0,4,0,4,0,3,0,4,4,3,4,2,4,3,2,0,4,4,4,3,5,3,5,3,3,2,4,2,4,3,4,3,1,4,0,2,3,4,4,4,3,3,3,4,4,4,3,4,1,3,4,3,2,1,2,1,3,3,3,4,4,3,3,5,0,4,0,3,0,4,3,3,3,2,1,0,3,0,0,3,3),
(0,4,0,3,0,3,0,3,0,3,5,5,3,3,3,3,4,3,4,3,3,3,4,4,4,3,3,3,3,4,3,5,3,3,1,3,2,4,5,5,5,5,4,3,4,5,5,3,2,2,3,3,3,3,2,3,3,1,2,3,2,4,3,3,3,4,0,4,0,2,0,4,3,2,2,1,2,0,3,0,0,4,1),
)

class JapaneseContextAnalysis(object):
    NUM_OF_CATEGORY = 6
    DONT_KNOW = -1
    ENOUGH_REL_THRESHOLD = 100
    MAX_REL_THRESHOLD = 1000
    MINIMUM_DATA_THRESHOLD = 4

    def __init__(self):
        self._total_rel = None
        self._rel_sample = None
        self._need_to_skip_char_num = None
        self._last_char_order = None
        self._done = None
        self.reset()

    def reset(self):
        self._total_rel = 0  # total sequence received
        # category counters, each integer counts sequence in its category
        self._rel_sample = [0] * self.NUM_OF_CATEGORY
        # if last byte in current buffer is not the last byte of a character,
        # we need to know how many bytes to skip in next buffer
        self._need_to_skip_char_num = 0
        self._last_char_order = -1  # The order of previous char
        # If this flag is set to True, detection is done and conclusion has
        # been made
        self._done = False

    def feed(self, byte_str, num_bytes):
        if self._done:
            return

        # The buffer we got is byte oriented, and a character may span in more than one
        # buffers. In case the last one or two byte in last buffer is not
        # complete, we record how many byte needed to complete that character
        # and skip these bytes here.  We can choose to record those bytes as
        # well and analyse the character once it is complete, but since a
        # character will not make much difference, by simply skipping
        # this character will simply our logic and improve performance.
        i = self._need_to_skip_char_num
        while i < num_bytes:
            order, char_len = self.get_order(byte_str[i:i + 2])
            i += char_len
            if i > num_bytes:
                self._need_to_skip_char_num = i - num_bytes
                self._last_char_order = -1
            else:
                if (order != -1) and (self._last_char_order != -1):
                    self._total_rel += 1
                    if self._total_rel > self.MAX_REL_THRESHOLD:
                        self._done = True
                        break
                    self._rel_sample[jp2CharContext[self._last_char_order][order]] += 1
                self._last_char_order = order

    def got_enough_data(self):
        return self._total_rel > self.ENOUGH_REL_THRESHOLD

    def get_confidence(self):
        # This is just one way to calculate confidence. It works well for me.
        if self._total_rel > self.MINIMUM_DATA_THRESHOLD:
            return (self._total_rel - self._rel_sample[0]) / self._total_rel
        else:
            return self.DONT_KNOW

    def get_order(self, byte_str):
        return -1, 1

class SJISContextAnalysis(JapaneseContextAnalysis):
    def __init__(self):
        super(SJISContextAnalysis, self).__init__()
        self._charset_name = "SHIFT_JIS"

    @property
    def charset_name(self):
        return self._charset_name

    def get_order(self, byte_str):
        if not byte_str:
            return -1, 1
        # find out current char's byte length
        first_char = byte_str[0]
        if (0x81 <= first_char <= 0x9F) or (0xE0 <= first_char <= 0xFC):
            char_len = 2
            if (first_char == 0x87) or (0xFA <= first_char <= 0xFC):
                self._charset_name = "CP932"
        else:
            char_len = 1

        # return its order if it is hiragana
        if len(byte_str) > 1:
            second_char = byte_str[1]
            if (first_char == 202) and (0x9F <= second_char <= 0xF1):
                return second_char - 0x9F, char_len

        return -1, char_len

class EUCJPContextAnalysis(JapaneseContextAnalysis):
    def get_order(self, byte_str):
        if not byte_str:
            return -1, 1
        # find out current char's byte length
        first_char = byte_str[0]
        if (first_char == 0x8E) or (0xA1 <= first_char <= 0xFE):
            char_len = 2
        elif first_char == 0x8F:
            char_len = 3
        else:
            char_len = 1

        # return its order if it is hiragana
        if len(byte_str) > 1:
            second_char = byte_str[1]
            if (first_char == 0xA4) and (0xA1 <= second_char <= 0xF3):
                return second_char - 0xA1, char_len

        return -1, char_len


site-packages/pip/_vendor/chardet/euckrfreq.py000064400000032352150564745010015477 0ustar00######################## BEGIN LICENSE BLOCK ########################
# The Original Code is Mozilla Communicator client code.
#
# The Initial Developer of the Original Code is
# Netscape Communications Corporation.
# Portions created by the Initial Developer are Copyright (C) 1998
# the Initial Developer. All Rights Reserved.
#
# Contributor(s):
#   Mark Pilgrim - port to Python
#
# This library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# This library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with this library; if not, write to the Free Software
# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA
# 02110-1301  USA
######################### END LICENSE BLOCK #########################

# Sampling from about 20M text materials include literature and computer technology

# 128  --> 0.79
# 256  --> 0.92
# 512  --> 0.986
# 1024 --> 0.99944
# 2048 --> 0.99999
#
# Idea Distribution Ratio = 0.98653 / (1-0.98653) = 73.24
# Random Distribution Ration = 512 / (2350-512) = 0.279.
#
# Typical Distribution Ratio

EUCKR_TYPICAL_DISTRIBUTION_RATIO = 6.0

EUCKR_TABLE_SIZE = 2352

# Char to FreqOrder table ,
EUCKR_CHAR_TO_FREQ_ORDER = (
  13, 130, 120,1396, 481,1719,1720, 328, 609, 212,1721, 707, 400, 299,1722,  87,
1397,1723, 104, 536,1117,1203,1724,1267, 685,1268, 508,1725,1726,1727,1728,1398,
1399,1729,1730,1731, 141, 621, 326,1057, 368,1732, 267, 488,  20,1733,1269,1734,
 945,1400,1735,  47, 904,1270,1736,1737, 773, 248,1738, 409, 313, 786, 429,1739,
 116, 987, 813,1401, 683,  75,1204, 145,1740,1741,1742,1743,  16, 847, 667, 622,
 708,1744,1745,1746, 966, 787, 304, 129,1747,  60, 820, 123, 676,1748,1749,1750,
1751, 617,1752, 626,1753,1754,1755,1756, 653,1757,1758,1759,1760,1761,1762, 856,
 344,1763,1764,1765,1766,  89, 401, 418, 806, 905, 848,1767,1768,1769, 946,1205,
 709,1770,1118,1771, 241,1772,1773,1774,1271,1775, 569,1776, 999,1777,1778,1779,
1780, 337, 751,1058,  28, 628, 254,1781, 177, 906, 270, 349, 891,1079,1782,  19,
1783, 379,1784, 315,1785, 629, 754,1402, 559,1786, 636, 203,1206,1787, 710, 567,
1788, 935, 814,1789,1790,1207, 766, 528,1791,1792,1208,1793,1794,1795,1796,1797,
1403,1798,1799, 533,1059,1404,1405,1156,1406, 936, 884,1080,1800, 351,1801,1802,
1803,1804,1805, 801,1806,1807,1808,1119,1809,1157, 714, 474,1407,1810, 298, 899,
 885,1811,1120, 802,1158,1812, 892,1813,1814,1408, 659,1815,1816,1121,1817,1818,
1819,1820,1821,1822, 319,1823, 594, 545,1824, 815, 937,1209,1825,1826, 573,1409,
1022,1827,1210,1828,1829,1830,1831,1832,1833, 556, 722, 807,1122,1060,1834, 697,
1835, 900, 557, 715,1836,1410, 540,1411, 752,1159, 294, 597,1211, 976, 803, 770,
1412,1837,1838,  39, 794,1413, 358,1839, 371, 925,1840, 453, 661, 788, 531, 723,
 544,1023,1081, 869,  91,1841, 392, 430, 790, 602,1414, 677,1082, 457,1415,1416,
1842,1843, 475, 327,1024,1417, 795, 121,1844, 733, 403,1418,1845,1846,1847, 300,
 119, 711,1212, 627,1848,1272, 207,1849,1850, 796,1213, 382,1851, 519,1852,1083,
 893,1853,1854,1855, 367, 809, 487, 671,1856, 663,1857,1858, 956, 471, 306, 857,
1859,1860,1160,1084,1861,1862,1863,1864,1865,1061,1866,1867,1868,1869,1870,1871,
 282,  96, 574,1872, 502,1085,1873,1214,1874, 907,1875,1876, 827, 977,1419,1420,
1421, 268,1877,1422,1878,1879,1880, 308,1881,   2, 537,1882,1883,1215,1884,1885,
 127, 791,1886,1273,1423,1887,  34, 336, 404, 643,1888, 571, 654, 894, 840,1889,
   0, 886,1274, 122, 575, 260, 908, 938,1890,1275, 410, 316,1891,1892, 100,1893,
1894,1123,  48,1161,1124,1025,1895, 633, 901,1276,1896,1897, 115, 816,1898, 317,
1899, 694,1900, 909, 734,1424, 572, 866,1425, 691,  85, 524,1010, 543, 394, 841,
1901,1902,1903,1026,1904,1905,1906,1907,1908,1909,  30, 451, 651, 988, 310,1910,
1911,1426, 810,1216,  93,1912,1913,1277,1217,1914, 858, 759,  45,  58, 181, 610,
 269,1915,1916, 131,1062, 551, 443,1000, 821,1427, 957, 895,1086,1917,1918, 375,
1919, 359,1920, 687,1921, 822,1922, 293,1923,1924,  40, 662, 118, 692,  29, 939,
 887, 640, 482, 174,1925,  69,1162, 728,1428, 910,1926,1278,1218,1279, 386, 870,
 217, 854,1163, 823,1927,1928,1929,1930, 834,1931,  78,1932, 859,1933,1063,1934,
1935,1936,1937, 438,1164, 208, 595,1938,1939,1940,1941,1219,1125,1942, 280, 888,
1429,1430,1220,1431,1943,1944,1945,1946,1947,1280, 150, 510,1432,1948,1949,1950,
1951,1952,1953,1954,1011,1087,1955,1433,1043,1956, 881,1957, 614, 958,1064,1065,
1221,1958, 638,1001, 860, 967, 896,1434, 989, 492, 553,1281,1165,1959,1282,1002,
1283,1222,1960,1961,1962,1963,  36, 383, 228, 753, 247, 454,1964, 876, 678,1965,
1966,1284, 126, 464, 490, 835, 136, 672, 529, 940,1088,1435, 473,1967,1968, 467,
  50, 390, 227, 587, 279, 378, 598, 792, 968, 240, 151, 160, 849, 882,1126,1285,
 639,1044, 133, 140, 288, 360, 811, 563,1027, 561, 142, 523,1969,1970,1971,   7,
 103, 296, 439, 407, 506, 634, 990,1972,1973,1974,1975, 645,1976,1977,1978,1979,
1980,1981, 236,1982,1436,1983,1984,1089, 192, 828, 618, 518,1166, 333,1127,1985,
 818,1223,1986,1987,1988,1989,1990,1991,1992,1993, 342,1128,1286, 746, 842,1994,
1995, 560, 223,1287,  98,   8, 189, 650, 978,1288,1996,1437,1997,  17, 345, 250,
 423, 277, 234, 512, 226,  97, 289,  42, 167,1998, 201,1999,2000, 843, 836, 824,
 532, 338, 783,1090, 182, 576, 436,1438,1439, 527, 500,2001, 947, 889,2002,2003,
2004,2005, 262, 600, 314, 447,2006, 547,2007, 693, 738,1129,2008,  71,1440, 745,
 619, 688,2009, 829,2010,2011, 147,2012,  33, 948,2013,2014,  74, 224,2015,  61,
 191, 918, 399, 637,2016,1028,1130, 257, 902,2017,2018,2019,2020,2021,2022,2023,
2024,2025,2026, 837,2027,2028,2029,2030, 179, 874, 591,  52, 724, 246,2031,2032,
2033,2034,1167, 969,2035,1289, 630, 605, 911,1091,1168,2036,2037,2038,1441, 912,
2039, 623,2040,2041, 253,1169,1290,2042,1442, 146, 620, 611, 577, 433,2043,1224,
 719,1170, 959, 440, 437, 534,  84, 388, 480,1131, 159, 220, 198, 679,2044,1012,
 819,1066,1443, 113,1225, 194, 318,1003,1029,2045,2046,2047,2048,1067,2049,2050,
2051,2052,2053,  59, 913, 112,2054, 632,2055, 455, 144, 739,1291,2056, 273, 681,
 499,2057, 448,2058,2059, 760,2060,2061, 970, 384, 169, 245,1132,2062,2063, 414,
1444,2064,2065,  41, 235,2066, 157, 252, 877, 568, 919, 789, 580,2067, 725,2068,
2069,1292,2070,2071,1445,2072,1446,2073,2074,  55, 588,  66,1447, 271,1092,2075,
1226,2076, 960,1013, 372,2077,2078,2079,2080,2081,1293,2082,2083,2084,2085, 850,
2086,2087,2088,2089,2090, 186,2091,1068, 180,2092,2093,2094, 109,1227, 522, 606,
2095, 867,1448,1093, 991,1171, 926, 353,1133,2096, 581,2097,2098,2099,1294,1449,
1450,2100, 596,1172,1014,1228,2101,1451,1295,1173,1229,2102,2103,1296,1134,1452,
 949,1135,2104,2105,1094,1453,1454,1455,2106,1095,2107,2108,2109,2110,2111,2112,
2113,2114,2115,2116,2117, 804,2118,2119,1230,1231, 805,1456, 405,1136,2120,2121,
2122,2123,2124, 720, 701,1297, 992,1457, 927,1004,2125,2126,2127,2128,2129,2130,
  22, 417,2131, 303,2132, 385,2133, 971, 520, 513,2134,1174,  73,1096, 231, 274,
 962,1458, 673,2135,1459,2136, 152,1137,2137,2138,2139,2140,1005,1138,1460,1139,
2141,2142,2143,2144,  11, 374, 844,2145, 154,1232,  46,1461,2146, 838, 830, 721,
1233, 106,2147,  90, 428, 462, 578, 566,1175, 352,2148,2149, 538,1234, 124,1298,
2150,1462, 761, 565,2151, 686,2152, 649,2153,  72, 173,2154, 460, 415,2155,1463,
2156,1235, 305,2157,2158,2159,2160,2161,2162, 579,2163,2164,2165,2166,2167, 747,
2168,2169,2170,2171,1464, 669,2172,2173,2174,2175,2176,1465,2177,  23, 530, 285,
2178, 335, 729,2179, 397,2180,2181,2182,1030,2183,2184, 698,2185,2186, 325,2187,
2188, 369,2189, 799,1097,1015, 348,2190,1069, 680,2191, 851,1466,2192,2193,  10,
2194, 613, 424,2195, 979, 108, 449, 589,  27, 172,  81,1031,  80, 774, 281, 350,
1032, 525, 301, 582,1176,2196, 674,1045,2197,2198,1467, 730, 762,2199,2200,2201,
2202,1468,2203, 993,2204,2205, 266,1070, 963,1140,2206,2207,2208, 664,1098, 972,
2209,2210,2211,1177,1469,1470, 871,2212,2213,2214,2215,2216,1471,2217,2218,2219,
2220,2221,2222,2223,2224,2225,2226,2227,1472,1236,2228,2229,2230,2231,2232,2233,
2234,2235,1299,2236,2237, 200,2238, 477, 373,2239,2240, 731, 825, 777,2241,2242,
2243, 521, 486, 548,2244,2245,2246,1473,1300,  53, 549, 137, 875,  76, 158,2247,
1301,1474, 469, 396,1016, 278, 712,2248, 321, 442, 503, 767, 744, 941,1237,1178,
1475,2249,  82, 178,1141,1179, 973,2250,1302,2251, 297,2252,2253, 570,2254,2255,
2256,  18, 450, 206,2257, 290, 292,1142,2258, 511, 162,  99, 346, 164, 735,2259,
1476,1477,   4, 554, 343, 798,1099,2260,1100,2261,  43, 171,1303, 139, 215,2262,
2263, 717, 775,2264,1033, 322, 216,2265, 831,2266, 149,2267,1304,2268,2269, 702,
1238, 135, 845, 347, 309,2270, 484,2271, 878, 655, 238,1006,1478,2272,  67,2273,
 295,2274,2275, 461,2276, 478, 942, 412,2277,1034,2278,2279,2280, 265,2281, 541,
2282,2283,2284,2285,2286,  70, 852,1071,2287,2288,2289,2290,  21,  56, 509, 117,
 432,2291,2292, 331, 980, 552,1101, 148, 284, 105, 393,1180,1239, 755,2293, 187,
2294,1046,1479,2295, 340,2296,  63,1047, 230,2297,2298,1305, 763,1306, 101, 800,
 808, 494,2299,2300,2301, 903,2302,  37,1072,  14,   5,2303,  79, 675,2304, 312,
2305,2306,2307,2308,2309,1480,   6,1307,2310,2311,2312,   1, 470,  35,  24, 229,
2313, 695, 210,  86, 778,  15, 784, 592, 779,  32,  77, 855, 964,2314, 259,2315,
 501, 380,2316,2317,  83, 981, 153, 689,1308,1481,1482,1483,2318,2319, 716,1484,
2320,2321,2322,2323,2324,2325,1485,2326,2327, 128,  57,  68, 261,1048, 211, 170,
1240,  31,2328,  51, 435, 742,2329,2330,2331, 635,2332, 264, 456,2333,2334,2335,
 425,2336,1486, 143, 507, 263, 943,2337, 363, 920,1487, 256,1488,1102, 243, 601,
1489,2338,2339,2340,2341,2342,2343,2344, 861,2345,2346,2347,2348,2349,2350, 395,
2351,1490,1491,  62, 535, 166, 225,2352,2353, 668, 419,1241, 138, 604, 928,2354,
1181,2355,1492,1493,2356,2357,2358,1143,2359, 696,2360, 387, 307,1309, 682, 476,
2361,2362, 332,  12, 222, 156,2363, 232,2364, 641, 276, 656, 517,1494,1495,1035,
 416, 736,1496,2365,1017, 586,2366,2367,2368,1497,2369, 242,2370,2371,2372,1498,
2373, 965, 713,2374,2375,2376,2377, 740, 982,1499, 944,1500,1007,2378,2379,1310,
1501,2380,2381,2382, 785, 329,2383,2384,1502,2385,2386,2387, 932,2388,1503,2389,
2390,2391,2392,1242,2393,2394,2395,2396,2397, 994, 950,2398,2399,2400,2401,1504,
1311,2402,2403,2404,2405,1049, 749,2406,2407, 853, 718,1144,1312,2408,1182,1505,
2409,2410, 255, 516, 479, 564, 550, 214,1506,1507,1313, 413, 239, 444, 339,1145,
1036,1508,1509,1314,1037,1510,1315,2411,1511,2412,2413,2414, 176, 703, 497, 624,
 593, 921, 302,2415, 341, 165,1103,1512,2416,1513,2417,2418,2419, 376,2420, 700,
2421,2422,2423, 258, 768,1316,2424,1183,2425, 995, 608,2426,2427,2428,2429, 221,
2430,2431,2432,2433,2434,2435,2436,2437, 195, 323, 726, 188, 897, 983,1317, 377,
 644,1050, 879,2438, 452,2439,2440,2441,2442,2443,2444, 914,2445,2446,2447,2448,
 915, 489,2449,1514,1184,2450,2451, 515,  64, 427, 495,2452, 583,2453, 483, 485,
1038, 562, 213,1515, 748, 666,2454,2455,2456,2457, 334,2458, 780, 996,1008, 705,
1243,2459,2460,2461,2462,2463, 114,2464, 493,1146, 366, 163,1516, 961,1104,2465,
 291,2466,1318,1105,2467,1517, 365,2468, 355, 951,1244,2469,1319,2470, 631,2471,
2472, 218,1320, 364, 320, 756,1518,1519,1321,1520,1322,2473,2474,2475,2476, 997,
2477,2478,2479,2480, 665,1185,2481, 916,1521,2482,2483,2484, 584, 684,2485,2486,
 797,2487,1051,1186,2488,2489,2490,1522,2491,2492, 370,2493,1039,1187,  65,2494,
 434, 205, 463,1188,2495, 125, 812, 391, 402, 826, 699, 286, 398, 155, 781, 771,
 585,2496, 590, 505,1073,2497, 599, 244, 219, 917,1018, 952, 646,1523,2498,1323,
2499,2500,  49, 984, 354, 741,2501, 625,2502,1324,2503,1019, 190, 357, 757, 491,
  95, 782, 868,2504,2505,2506,2507,2508,2509, 134,1524,1074, 422,1525, 898,2510,
 161,2511,2512,2513,2514, 769,2515,1526,2516,2517, 411,1325,2518, 472,1527,2519,
2520,2521,2522,2523,2524, 985,2525,2526,2527,2528,2529,2530, 764,2531,1245,2532,
2533,  25, 204, 311,2534, 496,2535,1052,2536,2537,2538,2539,2540,2541,2542, 199,
 704, 504, 468, 758, 657,1528, 196,  44, 839,1246, 272, 750,2543, 765, 862,2544,
2545,1326,2546, 132, 615, 933,2547, 732,2548,2549,2550,1189,1529,2551, 283,1247,
1053, 607, 929,2552,2553,2554, 930, 183, 872, 616,1040,1147,2555,1148,1020, 441,
 249,1075,2556,2557,2558, 466, 743,2559,2560,2561,  92, 514, 426, 420, 526,2562,
2563,2564,2565,2566,2567,2568, 185,2569,2570,2571,2572, 776,1530, 658,2573, 362,
2574, 361, 922,1076, 793,2575,2576,2577,2578,2579,2580,1531, 251,2581,2582,2583,
2584,1532,  54, 612, 237,1327,2585,2586, 275, 408, 647, 111,2587,1533,1106, 465,
   3, 458,   9,  38,2588, 107, 110, 890, 209,  26, 737, 498,2589,1534,2590, 431,
 202,  88,1535, 356, 287,1107, 660,1149,2591, 381,1536, 986,1150, 445,1248,1151,
 974,2592,2593, 846,2594, 446, 953, 184,1249,1250, 727,2595, 923, 193, 883,2596,
2597,2598, 102, 324, 539, 817,2599, 421,1041,2600, 832,2601,  94, 175, 197, 406,
2602, 459,2603,2604,2605,2606,2607, 330, 555,2608,2609,2610, 706,1108, 389,2611,
2612,2613,2614, 233,2615, 833, 558, 931, 954,1251,2616,2617,1537, 546,2618,2619,
1009,2620,2621,2622,1538, 690,1328,2623, 955,2624,1539,2625,2626, 772,2627,2628,
2629,2630,2631, 924, 648, 863, 603,2632,2633, 934,1540, 864, 865,2634, 642,1042,
 670,1190,2635,2636,2637,2638, 168,2639, 652, 873, 542,1054,1541,2640,2641,2642,  # 512, 256
)

site-packages/pip/_vendor/pkg_resources/__pycache__/__init__.cpython-39.pyc000064400000304027150564745010023000 0ustar00a

��Re���@sdZddlmZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlZddlmZzddlZWne�yddlZYn0ze Wne!�y&e"Z Yn0ddl#m$Z$ddl%m&Z&m'Z'm(Z(ddlm)Z)zddlm*Z*m+Z+m,Z,d	Z-Wne�y�d
Z-Yn0ddlm.Z/ddl0m1Z1m2Z2zddl3m4Z5e5j6Wne�y�dZ5Yn0d
dl7m8Z8ddl#m9Z9ddl#m:Z:e;d�e;d�e;d�e;d�e<Z=dej>k�r8dk�rDnne?d��e$j@�rTdZAdZBdZCdZDdZEdZFdZGdZHdZIdZJdZKdZLdZMdZNdZOdZPdZQdZRdZSGdd�deT�ZUdd�ZViZWdd�ZXdd�ZYd d!�ZZd"d#�Z[d$d%�Z\d&d'�Z]d(d)�Z^d*d+�Z_Z`d,d-�Zagd.�ZbGd/d0�d0ec�ZdGd1d2�d2ed�ZeGd3d4�d4ee�ZfGd5d6�d6ed�ZgGd7d8�d8ed�ZhiZid9jjej>�Zkd:Zld;Zmd
ZndZod<Zpd=d>�Zqd?d@�ZrgfdAdB�ZsdCdD�ZtdEdF�Zue�vdG�Zwe�vdH�ZxeuZydIdJ�ZzdKdL�Z{e{Z|dMdN�Z}dOdP�Z~d�dQdR�ZdSdT�Z�GdUdV�dV�Z�GdWdX�dXe��Z�GdYdZ�dZ�Z�Gd[d\�d\e��Z�Gd]d^�d^�Z�e�Z�Gd_d`�d`e?�Z�Gdadb�db�Z�dcdd�Z�dedf�Z�dgdh�Z�didj�Z�dkdl�Z�dmdn�Z�d�dodp�Z�Gdqdr�dr�Z�eqe�e��Gdsdt�dte��Z�Gdudv�dve��Z�e����Gdwdx�dxe��Z�e��Z�Gdydz�dze��Z�Gd{d|�d|e��Z�Gd}d~�d~e��Z�eqe
j�e��Gdd��d�e��Z�Gd�d��d�e��Z�Gd�d��d�e��Z�eXd�id��d�d��Z�d�d�d��Z�d�d�d��Z�e�e
j�e��d�d�d��Z�e�e�e��d�d��Z�d�d�d��Z�d�d��Z�Gd�d��d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�e�ej�e��e�e5d���re�e5j�e��eXd�id��eXd�id��d�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d�d��Z�d�d��Z�e�ej�e��e�e
j�e��e�e5d���r�e�e5j�e��d�d��Z�e�e�e��d�d��Z�d�d��Z�ifd�d��Z�d�d��Z�d�d��Z�d�d��Z�d�d��Z�e�vd��j�Z�e�vd�ej�ej�B�j�Z�Gd�d��d��Z�d�dÄZ�d�dńZ�Gd�dDŽdǃZ�Gd�dɄd�eăZ�Gd�d˄d�eăZ�e�e�e�d̜Z�d�d΄Z�Gd�dЄd�eɃZ�d�d҄Z�Gd�dԄd�e:j�j̓Z�d�dքZ�d�d؄Z�d�dڄZ�d�d܄Z�d�dބZ�d�d�Z�ej�d�eUd	d�d�d�Z�e�eփfd�d��Z�e�d�d��Z�Gd�d�d�eكZ�dS)�aZ
Package resource API
--------------------

A resource is a logical file contained within a package, or a logical
subdirectory thereof.  The package resource API expects resource names
to have their path parts separated with ``/``, *not* whatever the local
path separator is.  Do not use os.path operations to manipulate resource
names being passed into the API.

The package resource API is designed to work with normal filesystem packages,
.egg files, and unpacked .egg files.  It can also work in a limited way with
.zip files and with custom PEP 302 loaders that support the ``get_data()``
method.
�)�absolute_importN)�get_importer)�six)�urllib�map�filter)�utime)�mkdir�rename�unlinkTF)�open)�isdir�split�)�
py31compat)�platformdirs)�	packagingzpip._vendor.packaging.versionz pip._vendor.packaging.specifiersz"pip._vendor.packaging.requirementszpip._vendor.packaging.markers)�r)r�zPython 3.5 or later is requiredc@seZdZdZdS)�
PEP440Warningza
    Used when there is an issue with a version or specifier not complying with
    PEP 440.
    N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pkg_resources/__init__.pyrxsrcCs6ztj�|�WStjjy0tj�|�YS0dS�N)r�version�Version�InvalidVersion�
LegacyVersion)�vrrr�
parse_versionsr#cKs"t��|�t�t�||��dSr)�globals�update�_state_vars�dict�fromkeys)�vartype�kwrrr�_declare_state�sr+cCs8i}t�}t��D] \}}|d|||�||<q|S)N�_sget_)r$r&�items��state�g�kr"rrr�__getstate__�s
r2cCs8t�}|��D]$\}}|dt|||||�q|S)N�_sset_)r$r-r&r.rrr�__setstate__�sr4cCs|��Sr)�copy��valrrr�
_sget_dict�sr8cCs|��|�|�dSr)�clearr%��key�obr/rrr�
_sset_dict�sr=cCs|��Sr)r2r6rrr�_sget_object�sr>cCs|�|�dSr)r4r:rrr�_sset_object�sr?cGsdSrr��argsrrr�<lambda>��rBcCs`t�}t�|�}|dur\tjdkr\z&dd�t�dd��|�d�f}WntyZYn0|S)aZReturn this platform's maximum compatible version.

    distutils.util.get_platform() normally reports the minimum version
    of Mac OS X that would be required to *use* extensions produced by
    distutils.  But what we want when checking compatibility is to know the
    version of Mac OS X that we are *running*.  To allow usage of packages that
    explicitly require a newer version of Mac OS X, we must also know the
    current version of the OS.

    If this condition occurs for any other platform with a version in its
    platform strings, this function should be extended accordingly.
    N�darwinzmacosx-%s-%s�.�r)	�get_build_platform�macosVersionString�match�sys�platform�join�_macosx_vers�group�
ValueError)�plat�mrrr�get_supported_platform�s

&rR)G�require�
run_script�get_provider�get_distribution�load_entry_point�
get_entry_map�get_entry_info�iter_entry_points�resource_string�resource_stream�resource_filename�resource_listdir�resource_exists�resource_isdir�declare_namespace�working_set�add_activation_listener�find_distributions�set_extraction_path�cleanup_resources�get_default_cache�Environment�
WorkingSet�ResourceManager�Distribution�Requirement�
EntryPoint�ResolutionError�VersionConflict�DistributionNotFound�UnknownExtra�ExtractionErrorr�parse_requirementsr#�	safe_name�safe_version�get_platform�compatible_platforms�yield_lines�split_sections�
safe_extra�to_filename�invalid_marker�evaluate_marker�ensure_directory�normalize_path�EGG_DIST�BINARY_DIST�SOURCE_DIST�
CHECKOUT_DIST�DEVELOP_DIST�IMetadataProvider�IResourceProvider�FileMetadata�PathMetadata�EggMetadata�
EmptyProvider�empty_provider�NullProvider�EggProvider�DefaultProvider�ZipProvider�register_finder�register_namespace_handler�register_loader_type�fixup_namespace_packagesr�PkgResourcesDeprecationWarning�run_main�AvailableDistributionsc@seZdZdZdd�ZdS)rnz.Abstract base for dependency resolution errorscCs|jjt|j�Sr)�	__class__r�reprrA��selfrrr�__repr__�szResolutionError.__repr__N)rrrrr�rrrrrn�srnc@s<eZdZdZdZedd��Zedd��Zdd�Zd	d
�Z	dS)roz�
    An already-installed version conflicts with the requested version.

    Should be initialized with the installed Distribution and the requested
    Requirement.
    z3{self.dist} is installed but {self.req} is requiredcCs
|jdS�Nrr@r�rrr�distszVersionConflict.distcCs
|jdS�Nrr@r�rrr�reqszVersionConflict.reqcCs|jjfit���Sr��	_template�format�localsr�rrr�reportszVersionConflict.reportcCs|s|S|j|f}t|�S)zt
        If required_by is non-empty, return a version of self that is a
        ContextualVersionConflict.
        )rA�ContextualVersionConflict)r��required_byrArrr�with_contextszVersionConflict.with_contextN)
rrrrr��propertyr�r�r�r�rrrrros

roc@s&eZdZdZejdZedd��ZdS)r�z�
    A VersionConflict that accepts a third parameter, the set of the
    requirements that required the installed Distribution.
    z by {self.required_by}cCs
|jdS)NrFr@r�rrrr�*sz%ContextualVersionConflict.required_byN)rrrrror�r�r�rrrrr�"s
r�c@sHeZdZdZdZedd��Zedd��Zedd��Zd	d
�Z	dd�Z
d
S)rpz&A requested distribution was not foundzSThe '{self.req}' distribution was not found and is required by {self.requirers_str}cCs
|jdSr�r@r�rrrr�5szDistributionNotFound.reqcCs
|jdSr�r@r�rrr�	requirers9szDistributionNotFound.requirerscCs|js
dSd�|j�S)Nzthe applicationz, )r�rLr�rrr�
requirers_str=sz"DistributionNotFound.requirers_strcCs|jjfit���Srr�r�rrrr�CszDistributionNotFound.reportcCs|��Sr)r�r�rrr�__str__FszDistributionNotFound.__str__N)rrrrr�r�r�r�r�r�r�rrrrrp/s


rpc@seZdZdZdS)rqz>Distribution doesn't have an "extra feature" of the given nameNrrrrrrqJsrqz{}.{}rrF���cCs|t|<dS)aRegister `provider_factory` to make providers for `loader_type`

    `loader_type` is the type or class of a PEP 302 ``module.__loader__``,
    and `provider_factory` is a function that, passed a *module* object,
    returns an ``IResourceProvider`` for that module.
    N)�_provider_factories)�loader_type�provider_factoryrrrr�Xsr�cCsrt|t�r$t�|�p"tt|��dSztj|}Wn$tyVt	|�tj|}Yn0t
|dd�}tt|�|�S)z?Return an IResourceProvider for the named module or requirementr�
__loader__N)
�
isinstancerlrb�findrS�strrJ�modules�KeyError�
__import__�getattr�
_find_adapterr�)�moduleOrReq�module�loaderrrrrUbs
rUcCsd|s\t��d}|dkrLd}tj�|�rLttd�rLt�|�}d|vrL|d}|�|�	d��|dS)Nr�z0/System/Library/CoreServices/SystemVersion.plist�	readPlist�ProductVersionrE)
rK�mac_ver�os�path�exists�hasattr�plistlibr��appendr)�_cacher�plist�
plist_contentrrrrMos

rMcCsddd��||�S)N�ppc)�PowerPC�Power_Macintosh)�get)�machinerrr�_macosx_archsr�cCs|ddlm}|�}tjdkrx|�d�sxz>t�}t��d�dd�}dt	|d�t	|d	�t
|�fWStyvYn0|S)
z�Return this platform's string for platform-specific distributions

    XXX Currently this is the same as ``distutils.util.get_platform()``, but it
    needs some hacks for Linux and Mac OS X.
    r)rvrDzmacosx-�� �_zmacosx-%d.%d-%sr)�	sysconfigrvrJrK�
startswithrMr��uname�replace�intr�rO)rvrPrr�rrrrG�s�rGzmacosx-(\d+)\.(\d+)-(.*)zdarwin-(\d+)\.(\d+)\.(\d+)-(.*)cCs�|dus|dus||krdSt�|�}|r�t�|�}|s�t�|�}|r�t|�d��}d|�d�|�d�f}|dkr||dks�|dkr�|d	kr�dSd
S|�d�|�d�ks�|�d�|�d�kr�d
St|�d��t|�d��kr�d
SdSd
S)z�Can code for the `provided` platform run on the `required` platform?

    Returns true if either platform is ``None``, or the platforms are equal.

    XXX Needs compatibility checks for Linux and other unixy OSes.
    NTrz%s.%srF�z10.3�z10.4Fr)rHrI�darwinVersionStringr�rN)�provided�required�reqMac�provMac�
provDarwin�dversion�macosversionrrrrw�s2


���rwcCs<t�d�j}|d}|��||d<t|�d�||�dS)z@Locate distribution `dist_spec` and run its `script_name` scriptrrrN�rJ�	_getframe�	f_globalsr9rSrT)Z	dist_spec�script_name�ns�namerrrrT�s
rTcCs@t|tj�rt�|�}t|t�r(t|�}t|t�s<td|��|S)z@Return a current distribution object for a Requirement or stringz-Expected string, Requirement, or Distribution)r�r�string_typesrl�parserUrk�	TypeError�r�rrrrV�s



rVcCst|��||�S)zDReturn `name` entry point of `group` for `dist` or raise ImportError)rVrW�r�rNr�rrrrW�srWcCst|��|�S)�=Return the entry point map for `group`, or the full entry map)rVrX)r�rNrrrrX�srXcCst|��||�S�z<Return the EntryPoint object for `group`+`name`, or ``None``)rVrYr�rrrrY�srYc@s<eZdZdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Zd
S)r�cCsdS)z;Does the package's distribution contain the named metadata?Nr�r�rrr�has_metadata�szIMetadataProvider.has_metadatacCsdS)z'The named metadata resource as a stringNrr�rrr�get_metadata�szIMetadataProvider.get_metadatacCsdS)z�Yield named metadata resource as list of non-blank non-comment lines

       Leading and trailing whitespace is stripped from each line, and lines
       with ``#`` as the first non-blank character are omitted.Nrr�rrr�get_metadata_lines�sz$IMetadataProvider.get_metadata_linescCsdS)z>Is the named metadata a directory?  (like ``os.path.isdir()``)Nrr�rrr�metadata_isdirsz IMetadataProvider.metadata_isdircCsdS)z?List of metadata names in the directory (like ``os.listdir()``)Nrr�rrr�metadata_listdirsz"IMetadataProvider.metadata_listdircCsdS)z=Execute the named script in the supplied namespace dictionaryNr)r��	namespacerrrrT	szIMetadataProvider.run_scriptN)	rrrr�r�r�r�r�rTrrrrr��sr�c@s@eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dS)r�z3An object that provides access to package resourcescCsdS)zdReturn a true filesystem path for `resource_name`

        `manager` must be an ``IResourceManager``Nr��manager�
resource_namerrr�get_resource_filenamesz'IResourceProvider.get_resource_filenamecCsdS)ziReturn a readable file-like object for `resource_name`

        `manager` must be an ``IResourceManager``Nrr�rrr�get_resource_streamsz%IResourceProvider.get_resource_streamcCsdS)zmReturn a string containing the contents of `resource_name`

        `manager` must be an ``IResourceManager``Nrr�rrr�get_resource_stringsz%IResourceProvider.get_resource_stringcCsdS)z,Does the package contain the named resource?Nr�r�rrr�has_resourceszIResourceProvider.has_resourcecCsdS)z>Is the named resource a directory?  (like ``os.path.isdir()``)Nrr�rrrr`"sz IResourceProvider.resource_isdircCsdS)z?List of resource names in the directory (like ``os.listdir()``)Nrr�rrrr^%sz"IResourceProvider.resource_listdirN)
rrrrr�r�r�r�r`r^rrrrr�
sr�c@s�eZdZdZd'dd�Zedd��Zedd��Zd	d
�Zdd�Z	d
d�Z
d(dd�Zdd�Zdd�Z
d)dd�Zd*dd�Zd+dd�Zdd�Zd,dd �Zd!d"�Zd#d$�Zd%d&�ZdS)-rizDA collection of active distributions on sys.path (or a similar list)NcCs>g|_i|_i|_g|_|dur&tj}|D]}|�|�q*dS)z?Create working set from list of path entries (default=sys.path)N)�entries�
entry_keys�by_key�	callbacksrJr��	add_entry)r�r��entryrrr�__init__,szWorkingSet.__init__cCs^|�}zddlm}Wnty,|YS0z|�|�WntyX|�|�YS0|S)z1
        Prepare the master working set.
        r)�__requires__)�__main__r�ImportErrorrSro�_build_from_requirements)�cls�wsrrrr�
_build_master9s
zWorkingSet._build_mastercCsf|g�}t|�}|�|t��}|D]}|�|�q"tjD]}||jvr8|�|�q8|jtjdd�<|S)zQ
        Build a working set from a requirement spec. Rewrites sys.path.
        N)rs�resolverh�addrJr�r�r)r�req_specr	�reqs�distsr�rrrrrMs

z#WorkingSet._build_from_requirementscCs<|j�|g�|j�|�t|d�D]}|�||d�q$dS)a�Add a path item to ``.entries``, finding any distributions on it

        ``find_distributions(entry, True)`` is used to find distributions
        corresponding to the path entry, and they are added.  `entry` is
        always appended to ``.entries``, even if it is already present.
        (This is because ``sys.path`` can contain the same value more than
        once, and the ``.entries`` of the ``sys.path`` WorkingSet should always
        equal ``sys.path``.)
        TFN)r��
setdefaultr�r�rdr)r�rr�rrrrcs
zWorkingSet.add_entrycCs|j�|j�|kS)z9True if `dist` is the active distribution for its project)r�r�r;�r�r�rrr�__contains__rszWorkingSet.__contains__cCs,|j�|j�}|dur(||vr(t||��|S)a�Find a distribution matching requirement `req`

        If there is an active distribution for the requested project, this
        returns it as long as it meets the version requirement specified by
        `req`.  But, if there is an active distribution for the project and it
        does *not* meet the `req` requirement, ``VersionConflict`` is raised.
        If there is no active distribution for the requested project, ``None``
        is returned.
        N)r�r�r;ro)r�r�r�rrrr�vs

zWorkingSet.findcs��fdd�|D�S)aYield entry point objects from `group` matching `name`

        If `name` is None, yields all entry points in `group` from all
        distributions in the working set, otherwise only ones matching
        both `group` and `name` are yielded (in distribution order).
        c3s8|]0}|�����D]}�dus*�|jkr|VqqdSr)rX�valuesr�)�.0r�r�rNr�rr�	<genexpr>�s�z/WorkingSet.iter_entry_points.<locals>.<genexpr>r�r�rNr�rrrrZ�s�zWorkingSet.iter_entry_pointscCs>t�d�j}|d}|��||d<|�|�d�||�dS)z?Locate distribution for `requires` and run `script_name` scriptrrrNr�)r��requiresr�r�r�rrrrT�s
zWorkingSet.run_scriptccsLi}|jD]<}||jvrq
|j|D] }||vr$d||<|j|Vq$q
dS)z�Yield distributions for non-duplicate projects in the working set

        The yield order is the order in which the items' path entries were
        added to the working set.
        rN)r�r�r�)r��seen�itemr;rrr�__iter__�s

zWorkingSet.__iter__TFcCs�|r|j|j||d�|dur$|j}|j�|g�}|j�|jg�}|sV|j|jvrVdS||j|j<|j|vrx|�|j�|j|vr�|�|j�|�|�dS)aAdd `dist` to working set, associated with `entry`

        If `entry` is unspecified, it defaults to the ``.location`` of `dist`.
        On exit from this routine, `entry` is added to the end of the working
        set's ``.entries`` (if it wasn't already present).

        `dist` is only added to the working set if it's for a project that
        doesn't already have a distribution in the set, unless `replace=True`.
        If it's added, any callbacks registered with the ``subscribe()`` method
        will be called.
        �r�N)	�	insert_onr��locationr�rr;r�r��
_added_new)r�r�r�insertr��keys�keys2rrrr�s

zWorkingSet.addcCsxt|�ddd�}i}i}g}t�}	t�t�}
|�rt|�d�}||vrHq.|	�||�sVq.|�|j�}|du�r|j	�|j�}|dus�||vr�|r�|}
|dur�|dur�t
|j�}nt
g�}tg�}
|j
||
||d�}||j<|dur�|
�|d�}t||��|�|�||v�r$|
|}t||��|��|�|j�ddd�}|�|�|D] }|
|�|j�|j|	|<�qHd||<q.|S)a�List all distributions needed to (recursively) meet `requirements`

        `requirements` must be a sequence of ``Requirement`` objects.  `env`,
        if supplied, should be an ``Environment`` instance.  If
        not supplied, it defaults to all distributions available within any
        entry or distribution in the working set.  `installer`, if supplied,
        will be invoked with each requirement that cannot be met by an
        already-installed distribution; it should return a ``Distribution`` or
        ``None``.

        Unless `replace_conflicting=True`, raises a VersionConflict exception
        if
        any requirements are found on the path that have the correct name but
        the wrong version.  Otherwise, if an `installer` is supplied it will be
        invoked to obtain the correct version of the requirement and activate
        it.

        `extras` is a list of the extras to be used with these requirements.
        This is important because extra requirements may look like `my_req;
        extra = "my_extra"`, which would otherwise be interpreted as a purely
        optional requirement.  Instead, we want to be able to assert that these
        requirements are truly required.
        Nr�r)�replace_conflictingT)�list�
_ReqExtras�collections�defaultdict�set�pop�markers_passr�r;r�rhr�ri�
best_matchrpr�ror�r�extras�extendr�project_name)r��requirements�env�	installerr#r,�	processed�best�to_activate�
req_extrasr�r�r�r	r��
dependent_req�new_requirements�new_requirementrrrr�sP


�




zWorkingSet.resolvecCst|�}|��i}i}|dur4t|j�}||7}n||}|�g�}	tt|	j|��|D]�}
||
D]�}|��g}z|	�|||�}
WnLt	y�}z4|||<|r�WYd}~qfnWYd}~qZWYd}~qfd}~00tt|	j|
��|�
t�|
��qZqfqZt|�}|��||fS)asFind all activatable distributions in `plugin_env`

        Example usage::

            distributions, errors = working_set.find_plugins(
                Environment(plugin_dirlist)
            )
            # add plugins+libs to sys.path
            map(working_set.add, distributions)
            # display errors
            print('Could not load', errors)

        The `plugin_env` should be an ``Environment`` instance that contains
        only distributions that are in the project's "plugin directory" or
        directories. The `full_env`, if supplied, should be an ``Environment``
        contains all currently-available distributions.  If `full_env` is not
        supplied, one is created automatically from the ``WorkingSet`` this
        method is called on, which will typically mean that every directory on
        ``sys.path`` will be scanned for distributions.

        `installer` is a standard installer callback as used by the
        ``resolve()`` method. The `fallback` flag indicates whether we should
        attempt to resolve older versions of a plugin if the newest version
        cannot be resolved.

        This method returns a 2-tuple: (`distributions`, `error_info`), where
        `distributions` is a list of the distributions found in `plugin_env`
        that were loadable, along with any other distributions that are needed
        to resolve their dependencies.  `error_info` is a dictionary mapping
        unloadable plugin distributions to an exception instance describing the
        error that occurred. Usually this will be a ``DistributionNotFound`` or
        ``VersionConflict`` instance.
        N)
r$�sortrhr�r�rr�as_requirementrrnr%r'r()r��
plugin_env�full_envr1�fallback�plugin_projects�
error_info�
distributionsr0�
shadow_setr.r�r��	resolveesr"rrr�find_plugins's4$



$zWorkingSet.find_pluginscGs&|�t|��}|D]}|�|�q|S)a�Ensure that distributions matching `requirements` are activated

        `requirements` must be a string or a (possibly-nested) sequence
        thereof, specifying the distributions and versions required.  The
        return value is a sequence of the distributions that needed to be
        activated to fulfill the requirements; all relevant distributions are
        included, even if they were already activated in this working set.
        )rrsr)r�r/�neededr�rrrrS{s	zWorkingSet.requirecCs8||jvrdS|j�|�|s"dS|D]}||�q&dS)z�Invoke `callback` for all distributions

        If `existing=True` (default),
        call on all existing ones, as well.
        N)rr�)r��callback�existingr�rrr�	subscribe�s
zWorkingSet.subscribecCs|jD]}||�qdSr)r)r�r�rErrrr�s
zWorkingSet._added_newcCs,|jdd�|j��|j��|jdd�fSr)r�r�r5r�rr�rrrr2�s�zWorkingSet.__getstate__cCs@|\}}}}|dd�|_|��|_|��|_|dd�|_dSr)r�r5r�r�r)r��e_k_b_cr�r!r�rrrrr4�s


zWorkingSet.__setstate__)N)N)NTF)NNFN)NNT)T)rrrrr�classmethodr
rrrr�rZrTrrrrCrSrGrr2r4rrrrri)s.





�
]�
T
ric@seZdZdZddd�ZdS)r%z>
    Map each requirement to the extras that demanded it.
    Ncs2�fdd�|��d�|pdD�}�jp0t|�S)z�
        Evaluate markers for req against each extra that
        demanded it.

        Return False if the req has a marker and fails
        evaluation. Otherwise, return True.
        c3s|]}�j�d|i�VqdS)�extraN��marker�evaluate)rrJ�r�rrr�s�z*_ReqExtras.markers_pass.<locals>.<genexpr>rr)r�rL�any)r�r�r,�extra_evalsrrNrr*�s
�z_ReqExtras.markers_pass)N)rrrrr*rrrrr%�sr%c@sxeZdZdZde�efdd�Zdd�Zdd�Zdd	d
�Z	dd�Z
d
d�Zddd�Zddd�Z
dd�Zdd�Zdd�ZdS)rhz5Searchable snapshot of distributions on a search pathNcCs i|_||_||_|�|�dS)a!Snapshot distributions available on a search path

        Any distributions found on `search_path` are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.

        `platform` is an optional string specifying the name of the platform
        that platform-specific distributions must be compatible with.  If
        unspecified, it defaults to the current platform.  `python` is an
        optional string naming the desired version of Python (e.g. ``'3.6'``);
        it defaults to the current version.

        You may explicitly set `platform` (and/or `python`) to ``None`` if you
        wish to map *all* distributions, not just those compatible with the
        running platform or Python version.
        N)�_distmaprK�python�scan)r��search_pathrKrRrrrr�szEnvironment.__init__cCs2|jdup|jdup|j|jk}|o0t|j|j�S)z�Is distribution `dist` acceptable for this environment?

        The distribution must match the platform and python version
        requirements specified when this environment was created, or False
        is returned.
        N)rR�
py_versionrwrK)r�r��	py_compatrrr�can_add�s
�
�zEnvironment.can_addcCs|j|j�|�dS)z"Remove `dist` from the environmentN)rQr;�removerrrrrX�szEnvironment.removecCs4|durtj}|D]}t|�D]}|�|�qqdS)adScan `search_path` for distributions usable in this environment

        Any distributions found are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.  Only distributions conforming to
        the platform/python version defined at initialization are added.
        N)rJr�rdr)r�rTrr�rrrrS�s
zEnvironment.scancCs|��}|j�|g�S)aReturn a newest-to-oldest list of distributions for `project_name`

        Uses case-insensitive `project_name` comparison, assuming all the
        project's distributions use their project's name converted to all
        lowercase as their key.

        )�lowerrQr�)r�r.�distribution_keyrrr�__getitem__�szEnvironment.__getitem__cCsL|�|�rH|��rH|j�|jg�}||vrH|�|�|jt�d�dd�dS)zLAdd `dist` if we ``can_add()`` it and it has not already been added
        �hashcmpT�r;�reverseN)	rW�has_versionrQrr;r�r9�operator�
attrgetter)r�r�rrrrrs

zEnvironment.addFcCsdz|�|�}Wnty*|s"�d}Yn0|dur8|S||jD]}||vrB|SqB|�||�S)a�Find distribution best matching `req` and usable on `working_set`

        This calls the ``find(req)`` method of the `working_set` to see if a
        suitable distribution is already active.  (This may raise
        ``VersionConflict`` if an unsuitable version of the project is already
        active in the specified `working_set`.)  If a suitable distribution
        isn't active, this method returns the newest distribution in the
        environment that meets the ``Requirement`` in `req`.  If no suitable
        distribution is found, and `installer` is supplied, then the result of
        calling the environment's ``obtain(req, installer)`` method will be
        returned.
        N)r�ror;�obtain)r�r�rbr1r#r�rrrr+s

zEnvironment.best_matchcCs|dur||�SdS)a�Obtain a distribution matching `requirement` (e.g. via download)

        Obtain a distro that matches requirement (e.g. via download).  In the
        base ``Environment`` class, this routine just returns
        ``installer(requirement)``, unless `installer` is None, in which case
        None is returned instead.  This method is a hook that allows subclasses
        to attempt other ways of obtaining a distribution before falling back
        to the `installer` argument.Nr)r��requirementr1rrrrb+s	zEnvironment.obtainccs"|j��D]}||r
|Vq
dS)z=Yield the unique project names of the available distributionsN)rQr!�r�r;rrrr7szEnvironment.__iter__cCsVt|t�r|�|�n<t|t�rD|D]}||D]}|�|�q0q$ntd|f��|S)z2In-place addition of a distribution or environmentzCan't add %r to environment)r�rkrrhr�)r��other�projectr�rrr�__iadd__=s

zEnvironment.__iadd__cCs*|jgddd�}||fD]}||7}q|S)z4Add an environment or distribution to an environmentN)rKrR�r�)r�re�newr0rrr�__add__Is
zEnvironment.__add__)N)NF)N)rrrrrR�PY_MAJORrrWrXrSr[rr+rbrrgrjrrrrrh�s�


�

rhc@seZdZdZdS)rraTAn error occurred extracting a resource

    The following attributes are available from instances of this exception:

    manager
        The resource manager that raised this exception

    cache_path
        The base directory for resource extraction

    original_error
        The exception instance that caused extraction to fail
    NrrrrrrrUsrrc@s�eZdZdZdZdd�Zdd�Zdd�Zd	d
�Zdd�Z	d
d�Z
dd�Zdd�Zddd�Z
edd��Zdd�Zdd�Zd dd�ZdS)!rjz'Manage resource extraction and packagesNcCs
i|_dSr)�cached_filesr�rrrriszResourceManager.__init__cCst|��|�S)zDoes the named resource exist?)rUr��r��package_or_requirementr�rrrr_lszResourceManager.resource_existscCst|��|�S)z,Is the named resource an existing directory?)rUr`rmrrrr`ps�zResourceManager.resource_isdircCst|��||�S)z4Return a true filesystem path for specified resource)rUr�rmrrrr]vs�z!ResourceManager.resource_filenamecCst|��||�S)z9Return a readable file-like object for specified resource)rUr�rmrrrr\|s�zResourceManager.resource_streamcCst|��||�S)z%Return specified resource as a string)rUr�rmrrrr[�s�zResourceManager.resource_stringcCst|��|�S)z1List the contents of the named resource directory)rUr^rmrrrr^�s�z ResourceManager.resource_listdircCsVt��d}|jpt�}t�d���}t|jfit	����}||_
||_||_|�dS)z5Give an error message for problems extracting file(s)ra
            Can't extract file(s) to egg cache

            The following error occurred while trying to extract file(s)
            to the Python egg cache:

              {old_exc}

            The Python egg cache directory is currently set to:

              {cache_path}

            Perhaps your account does not have write access to this directory?
            You can change the cache directory by setting the PYTHON_EGG_CACHE
            environment variable to point to an accessible directory.
            N)
rJ�exc_info�extraction_pathrg�textwrap�dedent�lstriprrr�r�r��
cache_path�original_error)r��old_excrt�tmpl�errrrr�extraction_error�sz ResourceManager.extraction_errorrcCsf|jp
t�}tjj||dg|�R�}zt|�WntyL|��Yn0|�|�d|j	|<|S)a�Return absolute location in cache for `archive_name` and `names`

        The parent directory of the resulting path will be created if it does
        not already exist.  `archive_name` should be the base filename of the
        enclosing egg (which may not be the name of the enclosing zipfile!),
        including its ".egg" extension.  `names`, if provided, should be a
        sequence of path name parts "under" the egg's extraction location.

        This method should only be called by resource providers that need to
        obtain an extraction location, and only for names they intend to
        extract, as it tracks the generated names for possible cleanup later.
        z-tmpr)
rprgr�r�rL�_bypass_ensure_directory�	Exceptionry�_warn_unsafe_extraction_pathrl)r��archive_name�names�extract_path�target_pathrrr�get_cache_path�s


zResourceManager.get_cache_pathcCsVtjdkr|�tjd�sdSt�|�j}|tj@s>|tj@rRd|}t�	|t
�dS)aN
        If the default extraction path is overridden and set to an insecure
        location, such as /tmp, it opens up an opportunity for an attacker to
        replace an extracted file with an unauthorized payload. Warn the user
        if a known insecure location is used.

        See Distribute #375 for more details.
        �nt�windirNz�%s is writable by group/others and vulnerable to attack when used with get_resource_filename. Consider a more secure location (set with .set_extraction_path or the PYTHON_EGG_CACHE environment variable).)r�r�r��environ�stat�st_mode�S_IWOTH�S_IWGRP�warnings�warn�UserWarning)r��mode�msgrrrr|�s
��z,ResourceManager._warn_unsafe_extraction_pathcCs.tjdkr*t�|�jdBd@}t�||�dS)a4Perform any platform-specific postprocessing of `tempname`

        This is where Mac header rewrites should be done; other platforms don't
        have anything special they should do.

        Resource providers should call this method ONLY after successfully
        extracting a compressed resource.  They must NOT call it on resources
        that are already in the filesystem.

        `tempname` is the current (temporary) name of the file, and `filename`
        is the name it will be renamed to by the caller after this routine
        returns.
        �posiximi�N)r�r�r�r��chmod)r��tempname�filenamer�rrr�postprocess�s
zResourceManager.postprocesscCs|jrtd��||_dS)a�Set the base path where resources will be extracted to, if needed.

        If you do not call this routine before any extractions take place, the
        path defaults to the return value of ``get_default_cache()``.  (Which
        is based on the ``PYTHON_EGG_CACHE`` environment variable, with various
        platform-specific fallbacks.  See that routine's documentation for more
        details.)

        Resources are extracted to subdirectories of this path based upon
        information given by the ``IResourceProvider``.  You may set this to a
        temporary directory, but then you must call ``cleanup_resources()`` to
        delete the extracted files when done.  There is no guarantee that
        ``cleanup_resources()`` will be able to remove all extracted files.

        (Note: you may not change the extraction path for a given resource
        manager once resources have been extracted, unless you first call
        ``cleanup_resources()``.)
        z5Can't change extraction path, files already extractedN)rlrOrp�r�r�rrrre�s
�z#ResourceManager.set_extraction_pathFcCsdS)aB
        Delete all extracted resource files and directories, returning a list
        of the file and directory names that could not be successfully removed.
        This function does not have any concurrency protection, so it should
        generally only be called when the extraction path is a temporary
        directory exclusive to a single process.  This method is not
        automatically called; you must call it explicitly or register it as an
        ``atexit`` function if you wish to ensure cleanup of a temporary
        directory used for extractions.
        Nr)r��forcerrrrfsz!ResourceManager.cleanup_resources)r)F)rrrrrprr_r`r]r\r[r^ryr��staticmethodr|r�rerfrrrrrjes 

rjcCstj�d�ptjdd�S)z�
    Return the ``PYTHON_EGG_CACHE`` environment variable
    or a platform-relevant user cache dir for an app
    named "Python-Eggs".
    �PYTHON_EGG_CACHEzPython-Eggs)�appname)r�r�r�r�user_cache_dirrrrrrgs
�rgcCst�dd|�S)z�Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    �[^A-Za-z0-9.]+�-)�re�subr�rrrrt%srtcCsHzttj�|��WStjjyB|�dd�}t�dd|�YS0dS)zB
    Convert an arbitrary string to a standard version string
    r�rEr�r�N)r�rrrr r�r�r�)rrrrru-s
rucCst�dd|���S)z�Convert an arbitrary string to a standard 'extra' name

    Any runs of non-alphanumeric characters are replaced with a single '_',
    and the result is always lowercased.
    z[^A-Za-z0-9.-]+r�)r�r�rY)rJrrrrz9srzcCs|�dd�S)z|Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    r�r�rr�rrrr{Bsr{c
CsFzt|�Wn4ty@}zd|_d|_|WYd}~Sd}~00dS)zo
    Validate text as a PEP 508 environment marker; return an exception
    if invalid or False otherwise.
    NF)r}�SyntaxErrorr��lineno)�text�errrr|Jsr|c
CsLztj�|�}|��WStjjyF}zt|��WYd}~n
d}~00dS)z�
    Evaluate a PEP 508 environment marker.
    Return a boolean indicating the marker result in this environment.
    Raise SyntaxError if marker is invalid.

    This implementation uses the 'pyparsing' module.
    N)r�markers�MarkerrM�
InvalidMarkerr�)r�rJrLr�rrrr}Xs

r}c@s�eZdZdZdZdZdZdd�Zdd�Zdd�Z	d	d
�Z
dd�Zd
d�Zdd�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd%d&�Zed'd(��Zd)d*�ZdS)+r�zETry to implement resources and metadata for arbitrary PEP 302 loadersNcCs(t|dd�|_tj�t|dd��|_dS)Nr��__file__r�)r�r�r�r��dirname�module_path�r�r�rrrrnszNullProvider.__init__cCs|�|j|�Sr)�_fnr��r�r�r�rrrr�rsz"NullProvider.get_resource_filenamecCst�|�||��Sr)�io�BytesIOr�r�rrrr�usz NullProvider.get_resource_streamcCs|�|�|j|��Sr)�_getr�r�r�rrrr�xsz NullProvider.get_resource_stringcCs|�|�|j|��Sr)�_hasr�r��r�r�rrrr�{szNullProvider.has_resourcecCs|�|j|�Sr)r��egg_info�r�r�rrr�_get_metadata_path~szNullProvider._get_metadata_pathcCs |js|jS|�|�}|�|�Sr)r�r�r��r�r�r�rrrr��s
zNullProvider.has_metadatac
Csv|js
dS|�|�}|�|�}tjr(|Sz|�d�WStyp}z$|jd�||�7_�WYd}~n
d}~00dS)Nr��utf-8z in {} file at path: {})	r�r�r�r�PY2�decode�UnicodeDecodeError�reasonr�)r�r�r��value�excrrrr��s

zNullProvider.get_metadatacCst|�|��Sr�rxr�r�rrrr��szNullProvider.get_metadata_linescCs|�|�|j|��Sr)�_isdirr�r�r�rrrr`�szNullProvider.resource_isdircCs|jo|�|�|j|��Sr)r�r�r�r�rrrr��szNullProvider.metadata_isdircCs|�|�|j|��Sr)�_listdirr�r�r�rrrr^�szNullProvider.resource_listdircCs|jr|�|�|j|��SgSr)r�r�r�r�rrrr��szNullProvider.metadata_listdirc
Cs�d|}|�|�s(tdjfit�����|�|��dd�}|�dd�}|�|j|�}||d<tj	�
|�r�t|���}t
||d�}t|||�n>dd	lm}t|�d|�d�|f||<t
||d�}	t|	||�dS)
Nzscripts/z<Script {script!r} not found in metadata at {self.egg_info!r}z
�
�
r��execr)�cache)r�rnr�r�r�r�r�r�r�r�r�r�read�compiler��	linecacher��lenr)
r�r�r��script�script_text�script_filename�source�coder��script_coderrrrT�s(
���zNullProvider.run_scriptcCstd��dS�Nz9Can't perform this operation for unregistered loader type��NotImplementedErrorr�rrrr��s�zNullProvider._hascCstd��dSr�r�r�rrrr��s�zNullProvider._isdircCstd��dSr�r�r�rrrr��s�zNullProvider._listdircCs,|�|�|r(tjj|g|�d��R�S|S)N�/)�_validate_resource_pathr�r�rLr)r��baser�rrrr��s
zNullProvider._fncCsptjj|�tj�vp&t�|�p&t�|�}|s0dSd}t�|�rPt�|�sPt|��t	j
|dd�dtdd�dS)aO
        Validate the resource paths according to the docs.
        https://setuptools.readthedocs.io/en/latest/pkg_resources.html#basic-resource-access

        >>> warned = getfixture('recwarn')
        >>> warnings.simplefilter('always')
        >>> vrp = NullProvider._validate_resource_path
        >>> vrp('foo/bar.txt')
        >>> bool(warned)
        False
        >>> vrp('../foo/bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('/foo/bar.txt')
        >>> bool(warned)
        True
        >>> vrp('foo/../../bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('foo/f../bar.txt')
        >>> bool(warned)
        False

        Windows path separators are straight-up disallowed.
        >>> vrp(r'\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path is not allowed.

        >>> vrp(r'C:\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path is not allowed.

        Blank values are allowed

        >>> vrp('')
        >>> bool(warned)
        False

        Non-string values are not.

        >>> vrp(None)
        Traceback (most recent call last):
        ...
        AttributeError: ...
        Nz=Use of .. or absolute path in a resource path is not allowed.r�z/ and will raise exceptions in a future release.r���
stacklevel)r�r��pardirr�	posixpath�sep�isabs�ntpathrOr�r��DeprecationWarning)r��invalidr�rrrr��s6���z$NullProvider._validate_resource_pathcCs$t|jd�r|j�|�Std��dS)N�get_dataz=Can't perform this operation for loaders without 'get_data()')r�r�r�r�r�rrrr�s
�zNullProvider._get)rrrr�egg_namer�r�rr�r�r�r�r�r�r�r�r`r�r^r�rTr�r�r�r�r�r�r�rrrrr�gs2
Jr�c@s eZdZdZdd�Zdd�ZdS)r�z&Provider based on a virtual filesystemcCst�||�|��dSr)r�r�
_setup_prefixr�rrrr-szEggProvider.__init__cCsZ|j}d}||krVt|�r@tj�|�|_tj�|d�|_||_qV|}tj�	|�\}}q
dS)N�EGG-INFO)
r��_is_egg_pathr�r��basenamer�rLr��egg_rootr)r�r��oldr�rrrr�1szEggProvider._setup_prefixN)rrrrrr�rrrrr�*sr�c@sDeZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Ze	dd
��Z
dS)r�z6Provides access to package resources in the filesystemcCstj�|�Sr)r�r�r�r�rrrr�CszDefaultProvider._hascCstj�|�Sr)r�r�r
r�rrrr�FszDefaultProvider._isdircCs
t�|�Sr)r��listdirr�rrrr�IszDefaultProvider._listdircCst|�|j|�d�S�N�rb)rr�r�r�rrrr�Lsz#DefaultProvider.get_resource_streamcCs6t|d��}|��Wd�S1s(0YdSr�)rr�)r�r��streamrrrr�OszDefaultProvider._getcCs,d}|D]}tt|td��}t||�qdS)N)�SourceFileLoader�SourcelessFileLoader)r��importlib_machinery�typer�)r�loader_namesr��
loader_clsrrr�	_registerSszDefaultProvider._registerN)rrrrr�r�r�r�r�rIr�rrrrr�@sr�c@s8eZdZdZdZdd�ZZdd�Zdd�Zd	d
�Z	dS)r�z.Provider that returns nothing for all requestsNcCsdS�NFrr�rrrrBcrCzEmptyProvider.<lambda>cCsdS�Nr�rr�rrrr�eszEmptyProvider._getcCsgSrrr�rrrr�hszEmptyProvider._listdircCsdSrrr�rrrrkszEmptyProvider.__init__)
rrrrr�r�r�r�r�rrrrrr�^sr�c@s eZdZdZedd��ZeZdS)�ZipManifestsz
    zip manifest builder
    csLt�|��.��fdd����D�}t|�Wd�S1s>0YdS)a
        Build a dictionary similar to the zipimport directory
        caches, except instead of tuples, store ZipInfo objects.

        Use a platform-specific path separator (os.sep) for the path keys
        for compatibility with pypy on Windows.
        c3s&|]}|�dtj���|�fVqdS)r�N)r�r�r��getinfo�rr���zfilerrr�s��z%ZipManifests.build.<locals>.<genexpr>N)�zipfile�ZipFile�namelistr')rr�r-rr�r�buildws
	
�zZipManifests.buildN)rrrrrIr��loadrrrrr�rs
r�c@s$eZdZdZe�dd�Zdd�ZdS)�MemoizedZipManifestsz%
    Memoized zipfile manifests.
    �manifest_modzmanifest mtimecCsRtj�|�}t�|�j}||vs.||j|krH|�|�}|�||�||<||jS)zW
        Load a manifest at path or return a suitable manifest already loaded.
        )	r�r��normpathr��st_mtime�mtimer�r��manifest)r�r�r�r�rrrr��s
zMemoizedZipManifests.loadN)rrrrr&�
namedtupler�r�rrrrr��sr�c@s�eZdZdZdZe�Zdd�Zdd�Zdd�Z	e
d	d
��Zdd�Ze
d
d��Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd �ZdS)!r�z"Resource support for zips and eggsNcCs t�||�|jjtj|_dSr)r�rr��archiver�r��zip_prer�rrrr�szZipProvider.__init__cCsP|�tj�}||jjkrdS|�|j�r:|t|j�d�Std||jf��dS)Nr��%s is not a subpath of %s)	�rstripr�r�r�rr�rr��AssertionError�r��fspathrrr�
_zipinfo_name�s�zZipProvider._zipinfo_namecCsP|j|}|�|jtj�r:|t|j�dd��tj�Std||jf��dS)Nrr)rr�r�r�r�r�rr)r��zip_pathrrrr�_parts�s
�zZipProvider._partscCs|j�|jj�Sr)�_zip_manifestsr�r�rr�rrr�zipinfo�szZipProvider.zipinfocCs\|jstd��|�|�}|��}d�|�|��|vrP|D]}|�||�|��q8|�||�S)Nz5resource_filename() only supported for .egg, not .zipr�)r�r��_resource_to_zip�_get_eager_resourcesrLr	�_extract_resource�
_eager_to_zip)r�r�r�r�eagersr�rrrr��s�
z!ZipProvider.get_resource_filenamecCs"|j}|jd}t�|�}||fS)N)rrr�)�	file_size�	date_time�time�mktime)�zip_stat�sizer�	timestamprrr�_get_date_and_size�s

zZipProvider._get_date_and_sizec
Cst||��vr@|��|D]}|�|tj�||��}qtj�|�S|�|j|�\}}ts`t	d��z�|�
|j|�|��}|�
||�r�|WStdtj�|�d�\}}	t�||j�|��t�|�t|	||f�|�|	|�zt|	|�Wnftj�yLtj�|��rF|�
||��r|YWStjdk�rFt|�t|	|�|YWS�Yn0Wntj�yn|��Yn0|S)Nz>"os.rename" and "os.unlink" are not supported on this platformz	.$extract)�dirr�)�_indexrr�r�rLr�rr�
WRITE_SUPPORT�IOErrorr�r�r	�_is_current�_mkstemp�writer�r��closerr�r
�error�isfiler�rry)
r�r�rr��lastrr�	real_path�outf�tmpnamrrrr�sJ��
�




zZipProvider._extract_resourcec	Cs�|�|j|�\}}tj�|�s$dSt�|�}|j|ksB|j|krFdS|j�	|�}t
|d��}|��}Wd�n1sz0Y||kS)zK
        Return True if the file_path is current for this zip_path
        Fr�N)rrr�r�r"r��st_sizer�r�r�rr�)	r��	file_pathrrrr��zip_contents�f�
file_contentsrrrrs
&zZipProvider._is_currentcCs>|jdur8g}dD]}|�|�r|�|�|��q||_|jS)N)znative_libs.txtzeager_resources.txt)rr�r-r�)r�rr�rrrr
"s

z ZipProvider._get_eager_resourcesc	Cs�z|jWSty�i}|jD]V}|�tj�}|r tj�|dd��}||vrf||�|d�q q0|��g||<q0q ||_|YS0dS)Nr�)	�	_dirindex�AttributeErrorrrr�r�rLr�r))r��indr��parts�parentrrrr+s
zZipProvider._indexcCs |�|�}||jvp||��vSr)rrr)r�rrrrrr�<s
zZipProvider._hascCs|�|�|��vSr)rrrrrrr�@szZipProvider._isdircCst|���|�|�d��S�Nr)r$rr�rrrrrr�CszZipProvider._listdircCs|�|�|j|��Sr)rr�r�r�rrrrFszZipProvider._eager_to_zipcCs|�|�|j|��Sr)rr�r�r�rrrrIszZipProvider._resource_to_zip)rrrrrr�r
rrr	r�rr�r�rrrr
rr�r�r�rrrrrrr��s(



7	r�c@s@eZdZdZdd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Z	dS)r�a*Metadata handler for standalone PKG-INFO files

    Usage::

        metadata = FileMetadata("/path/to/PKG-INFO")

    This provider rejects all data and metadata requests except for PKG-INFO,
    which is treated as existing, and will be the contents of the file at
    the provided location.
    cCs
||_dSr�r�r�rrrr\szFileMetadata.__init__cCs|jSrr2r�rrrr�_szFileMetadata._get_metadata_pathcCs|dkotj�|j�S)N�PKG-INFO)r�r�r"r�rrrr�bszFileMetadata.has_metadatacCsX|dkrtd��tj|jddd��}|��}Wd�n1s@0Y|�|�|S)Nr3z(No metadata except PKG-INFO is availabler�r�)�encoding�errors)r�r�rr�r��_warn_on_replacement)r�r�r*�metadatarrrr�es&
zFileMetadata.get_metadatacCs6d�d�}||vr2d}|jfit���}t�|�dS)Ns�r�z2{self.path} could not be properly decoded in UTF-8)r�r�r�r�r�)r�r7�replacement_charrwr�rrrr6ns

z!FileMetadata._warn_on_replacementcCst|�|��Srr�r�rrrr�vszFileMetadata.get_metadata_linesN)
rrrrrr�r�r�r6r�rrrrr�Ps	r�c@seZdZdZdd�ZdS)r�asMetadata provider for egg directories

    Usage::

        # Development eggs:

        egg_info = "/path/to/PackageName.egg-info"
        base_dir = os.path.dirname(egg_info)
        metadata = PathMetadata(base_dir, egg_info)
        dist_name = os.path.splitext(os.path.basename(egg_info))[0]
        dist = Distribution(basedir, project_name=dist_name, metadata=metadata)

        # Unpacked egg directories:

        egg_path = "/path/to/PackageName-ver-pyver-etc.egg"
        metadata = PathMetadata(egg_path, os.path.join(egg_path,'EGG-INFO'))
        dist = Distribution.from_filename(egg_path, metadata=metadata)
    cCs||_||_dSr)r�r�)r�r�r�rrrr�szPathMetadata.__init__N�rrrrrrrrrr�zsr�c@seZdZdZdd�ZdS)r�z Metadata provider for .egg filescCsD|jtj|_||_|jr0tj�|j|j�|_n|j|_|�	�dS)z-Create a metadata provider from a zipimporterN)
rr�r�rr��prefixr�rLr�r�)r��importerrrrr�szEggMetadata.__init__Nr9rrrrr��sr�r'��_distribution_finderscCs|t|<dS)axRegister `distribution_finder` to find distributions in sys.path items

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `distribution_finder` is a callable that, passed a path
    item and the importer instance, yields ``Distribution`` instances found on
    that path item.  See ``pkg_resources.find_on_path`` for an example.Nr<)�
importer_type�distribution_finderrrrr��sr�cCst|�}tt|�}||||�S)z.Yield distributions accessible via `path_item`)rr�r=)�	path_item�onlyr;�finderrrrrd�s
rdc	cs�|j�d�rdSt|�}|�d�r2tj||d�V|r:dS|�d�D]|}t|�r�tj	�
||�}tt�
|�|�}|D]
}|VqrqD|���d�rDtj	�
||�}tt�
|��}||_t�|||�VqDdS)z@
    Find eggs in zip files; possibly multiple nested eggs.
    z.whlNr3�r7r��
.dist-info)r�endswithr�r�rk�
from_filenamer^r�r�r�rL�find_eggs_in_zip�	zipimport�zipimporterrYr��
from_location)	r;r@rAr7�subitem�subpathrr��submetarrrrG�s$

rGcCsdSr1r)r;r@rArrr�find_nothing�srNcCsdd�}t||dd�S)aL
    Given a list of filenames, return them in descending order
    by version number.

    >>> names = 'bar', 'foo', 'Python-2.7.10.egg', 'Python-2.7.2.egg'
    >>> _by_version_descending(names)
    ['Python-2.7.10.egg', 'Python-2.7.2.egg', 'foo', 'bar']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.egg', 'Setuptools-1.2.3b1.egg']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.post1.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.post1.egg', 'Setuptools-1.2.3b1.egg']
    cSs2tj�|�\}}t�|�d�|g�}dd�|D�S)z6
        Parse each component of the filename
        r�cSsg|]}tj�|��qSr)rrr�)r�partrrr�
<listcomp>�rCz?_by_version_descending.<locals>._by_version.<locals>.<listcomp>)r�r��splitext�	itertools�chainr)r��extr/rrr�_by_version�sz+_by_version_descending.<locals>._by_versionTr])�sorted)r~rUrrr�_by_version_descending�srWc
#s�t���t��r4tj�t�tj��d��d�VdSt��}��fdd�|D�}t	|�}|D]2}tj��|�}t
�|��}||�D]
}	|	Vq�q\dS)z6Yield distributions accessible on a sys.path directoryr�rCNc3s|]}t�|��r|VqdSr)�dist_factory)rr�rAr@rrrs�zfind_on_path.<locals>.<genexpr>)�_normalize_cached�_is_unpacked_eggrkrFr�r�r�rL�safe_listdirrWrX)
r;r@rAr��filtered�path_item_entriesr�fullpath�factoryr�rrYr�find_on_path�s$���racCsH|��}tt|jd��}|r tS|s0t|�r0tS|sB|�d�rBtSt�S)z9
    Return a dist_factory for a path_item and entry
    )�	.egg-inforDz	.egg-link)	rYrOrrE�distributions_from_metadatar�rd�resolve_egg_link�NoDists)r@rrArY�is_metarrrrXs������rXc@s*eZdZdZdd�ZejreZdd�ZdS)rezS
    >>> bool(NoDists())
    False

    >>> list(NoDists()('anything'))
    []
    cCsdSr�rr�rrr�__bool__.szNoDists.__bool__cCstd�Sr1)�iter)r�r_rrr�__call__3szNoDists.__call__N)	rrrrrgrr��__nonzero__rirrrrre&s
rec
Csvzt�|�WSttfy"YnPtyp}z8|jtjtjtjfvpTt	|dd�dk}|s\�WYd}~n
d}~00dS)zI
    Attempt to list contents of path, but suppress some exceptions.
    �winerrorNir)
r�r��PermissionError�NotADirectoryError�OSError�errno�ENOTDIR�EACCES�ENOENTr�)r�r��	ignorablerrrr\7s�r\ccsftj�|�}tj�|�r:tt�|��dkr.dSt||�}nt|�}tj�|�}t	j
|||td�VdS)Nr)�
precedence)r�r�r�r
r�r�r�r�r�rkrJr�)r��rootr7rrrrrcLs�rcccsHt|��,}|D]}|��}|r|VqWd�n1s:0YdS)z1
    Yield non-empty lines from file at path
    N)r�strip)r�r*�linerrr�non_empty_lines[s

rxcs.t��}�fdd�|D�}tt|�}t|d�S)za
    Given a path to an .egg-link, resolve distributions
    present in the referenced path.
    c3s$|]}tj�tj���|�VqdSr)r�r�rLr�)r�refr2rrrls�z#resolve_egg_link.<locals>.<genexpr>r)rxrrd�next)r��referenced_paths�resolved_paths�dist_groupsrr2rrdfs
�
rd�
FileFinder��_namespace_handlers)�_namespace_packagescCs|t|<dS)a�Register `namespace_handler` to declare namespace packages

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `namespace_handler` is a callable like this::

        def namespace_handler(importer, path_entry, moduleName, module):
            # return a path_entry to use for child packages

    Namespace handlers are only called if the importer object has already
    agreed that it can handle the relevant path item, and they should only
    return a subpath if the module __path__ does not already contain an
    equivalent subpath.  For an example namespace handler, see
    ``pkg_resources.file_ns_handler``.
    Nr)r>�namespace_handlerrrrr�}sr�cCs�t|�}|durdSt���$t�d�|�|�}Wd�n1sF0Y|dur\dStj�|�}|dur�t�	|�}tj|<g|_
t|�nt|d�s�t
d|��tt|�}|||||�}|dur�|j
}|�|�|�|�t|||�|S)zEEnsure that named package includes a subpath of path_item (if needed)N�ignore�__path__�Not a package:)rr��catch_warnings�simplefilter�find_modulerJr�r��types�
ModuleTyper��_set_parent_nsr�r�r�r�r��load_module�_rebuild_mod_path)�packageNamer@r;r�r��handlerrLr�rrr�
_handle_ns�s.

(





r�csjdd�tjD���fdd����fdd�}t||d�}dd�|D�}t|jt�r`||jd	d	�<n||_d	S)
zq
    Rebuild module.__path__ ensuring that all entries are ordered
    corresponding to their sys.path order
    cSsg|]}t|��qSr�rZ�r�prrrrP�rCz%_rebuild_mod_path.<locals>.<listcomp>cs,z��|�WSty&td�YS0dS)z/
        Workaround for #520 and #513.
        �infN)�indexrO�float)r)�sys_pathrr�safe_sys_path_index�sz._rebuild_mod_path.<locals>.safe_sys_path_indexcs<|�tj�}��d�d}|d|�}�ttj�|���S)zR
        Return the ordinal of the path based on its position in sys.path
        rErN)rr�r��countrZrL)r��
path_parts�module_partsr/)�package_namer�rr�position_in_sys_path�sz/_rebuild_mod_path.<locals>.position_in_sys_path)r;cSsg|]}t|��qSrr�r�rrrrP�rCN)rJr�rVr�r�r$)�	orig_pathr�r�r��new_pathr)r�r�r�rr��s		r�cCs�t��z�|tvr Wt��dStj}|�d�\}}}|r�t|�|tvrRt|�ztj	|j
}Wnty~td|��Yn0t�
|p�dg��|�t�
|g�|D]}t||�q�Wt��n
t��0dS)z9Declare that package 'packageName' is a namespace packageNrEr�)�_imp�acquire_lockr��release_lockrJr��
rpartitionrar�r�r�r-r�rr�r�)r�r�r0r�r@rrrra�s*�racCsNt��z6t�|d�D]}t||�}|rt||�qWt��n
t��0dS)zDEnsure that previously-declared namespace packages include path_itemrN)r�r�r�r�r�r�r�)r@r0�packagerLrrrr��s
r�cCsDtj�||�d�d�}t|�}|jD]}t|�|kr&q@q&|SdS)zBCompute an ns-package subpath for a filesystem or zipfile importerrEr�N)r�r�rLrrZr�)r;r@r�r�rL�
normalizedrrrr�file_ns_handler�s
r�cCsdSrr)r;r@r�r�rrr�null_ns_handler	sr�cCs tj�tj�tj�t|����S)z1Normalize a file/dir name for comparison purposes)r�r��normcase�realpathr��
_cygwin_patch�r�rrrr	srcCstjdkrtj�|�S|S)a
    Contrary to POSIX 2008, on Cygwin, getcwd (3) contains
    symlink components. Using
    os.path.abspath() works around this limitation. A fix in os.getcwd()
    would probably better, in Cygwin even more so, except
    that this seems to be by design...
    �cygwin)rJrKr�r��abspathr�rrrr�	sr�cCs6z
||WSty0t|�||<}|YS0dSr)r�r)r�r��resultrrrrZ)	s

rZcCs|���d�S)z7
    Determine if given path appears to be an egg.
    �.egg)rYrEr2rrrr�1	sr�cCs t|�otj�tj�|dd��S)z@
    Determine if given path appears to be an unpacked egg.
    r�r3)r�r�r�r"rLr2rrrr[8	s�r[cCs<|�d�}|��}|r8d�|�}ttj||tj|�dS)NrE)rr)rL�setattrrJr�)r�r/r�r0rrrr�B	s


r�ccsZt|tj�r8|��D] }|��}|r|�d�s|Vqn|D]}t|�D]
}|VqHq<dS)z9Yield non-empty/non-comment lines of a string or sequence�#N)r�rr��
splitlinesrvr�rx)�strs�s�ssrrrrxJ	s
rxz\w+(\.\w+)*$z�
    (?P<name>[^-]+) (
        -(?P<ver>[^-]+) (
            -py(?P<pyver>[^-]+) (
                -(?P<plat>.+)
            )?
        )?
    )?
    c@s�eZdZdZddd�Zdd�Zdd	�Zddd�Zd
d�Zddd�Z	e
�d�Ze
ddd��Ze
dd��Ze
ddd��Ze
ddd��ZdS) rmz3Object representing an advertised importable objectrNcCs<t|�std|��||_||_t|�|_t|�|_||_dS)NzInvalid module name)�MODULErOr��module_name�tuple�attrsr,r�)r�r�r�r�r,r�rrrrj	s


zEntryPoint.__init__cCsHd|j|jf}|jr*|dd�|j�7}|jrD|dd�|j�7}|S)Nz%s = %s�:rEz [%s]�,)r�r�r�rLr,)r�r�rrrr�s	szEntryPoint.__str__cCsdt|�S)NzEntryPoint.parse(%r)�r�r�rrrr�{	szEntryPoint.__repr__TcOs8|r|s|rtjdtdd�|r0|j|i|��|��S)zH
        Require packages for this EntryPoint, then resolve it.
        zJParameters to load are deprecated.  Call .resolve and .require separately.rFr�)r�r�r�rSr)r�rSrA�kwargsrrrr�~	s�zEntryPoint.loadc
CsZt|jdgdd�}zt�t|j|�WStyT}ztt|���WYd}~n
d}~00dS)zD
        Resolve the entry point from its module and attrs.
        rr)�fromlist�levelN)	r�r��	functools�reducer�r�r-rr�)r�r�r�rrrr�	s
zEntryPoint.resolvecCsL|jr|jstd|��|j�|j�}tj||||jd�}tttj|��dS)Nz&Can't require() without a distribution)r,)	r,r�rqrrbrr$rr)r�r0r1rr-rrrrS�	s

zEntryPoint.requirez]\s*(?P<name>.+?)\s*=\s*(?P<module>[\w.]+)\s*(:\s*(?P<attr>[\w.]+))?\s*(?P<extras>\[.*\])?\s*$cCsf|j�|�}|sd}t||��|��}|�|d�}|drJ|d�d�nd}||d|d|||�S)aParse a single entry point from string `src`

        Entry point syntax follows the form::

            name = some.module:some.attr [extra1, extra2]

        The entry name and module name are required, but the ``:attrs`` and
        ``[extras]`` parts are optional
        z9EntryPoint must be in 'name=module:attrs [extras]' formatr,�attrrErr�r�)�patternrIrO�	groupdict�
_parse_extrasr)r�srcr�rQr��resr,r�rrrr��	s
zEntryPoint.parsecCs(|sdSt�d|�}|jr"t��|jS)Nr�x)rlr��specsrOr,)r�extras_specr�rrrr��	szEntryPoint._parse_extrascCsVt|�std|��i}t|�D]2}|�||�}|j|vrFtd||j��|||j<q|S)zParse an entry point groupzInvalid group namezDuplicate entry point)r�rOrxr�r�)rrN�linesr��thisrw�eprrr�parse_group�	s

zEntryPoint.parse_groupcCstt|t�r|��}nt|�}i}|D]J\}}|durB|s:q$td��|��}||vr\td|��|�|||�||<q$|S)z!Parse a map of entry point groupsNz%Entry points must be listed in groupszDuplicate group name)r�r'r-ryrOrvr�)r�datar��mapsrNr�rrr�	parse_map�	s


zEntryPoint.parse_map)rrN)T)NN)N)N)N)rrrrrr�r�r�rrSr�r�r�rIr�r�r�r�rrrrrmg	s$
	



�	
rmcCs>|sdStj�|�}|d�d�r:tj�|dd�d�S|S)Nr�r�zmd5=)r�)rr��urlparser��
urlunparse)r�parsedrrr�_remove_md5_fragment�	sr�cCs@dd�}t||�}tt|�d�}|�d�\}}}t|���p>dS)z�
    Given an iterable of lines from a Metadata file, return
    the value of the Version field, if present, or None otherwise.
    cSs|���d�S)Nzversion:)rYr�)rwrrr�is_version_line�	sz+_version_from_file.<locals>.is_version_liner�r�N)rrzrh�	partitionrurv)r�r��
version_linesrwr�r�rrr�_version_from_file�	s

r�cs�eZdZdZdZddddedefdd�ZedSdd��Z	dd	�Z
ed
d��Zdd
�Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zedd��Zedd��Zdd�Zed d!��Zed"d#��Zed$d%��Zd&d'�ZdTd)d*�Zd+d,�Zd-d.�Zd/d0�ZdUd2d3�Z d4d5�Z!d6d7�Z"d8d9�Z#d:d;�Z$�fd<d=�Z%e&e'd>��s4[%edVd?d@��Z(dAdB�Z)dCdD�Z*dWdEdF�Z+dGdH�Z,dXdIdJ�Z-dKdL�Z.dMdN�Z/dOdP�Z0edQdR��Z1�Z2S)Yrkz5Wrap an actual or potential sys.path entry w/metadatar3NcCsFt|pd�|_|dur t|�|_||_||_||_||_|p>t|_	dS)N�Unknown)
rtr.ru�_versionrUrKrrtr��	_provider)r�rr7r.rrUrKrtrrrr
s
zDistribution.__init__cKs~dgd\}}}}tj�|�\}}	|	��tvr^t|	��}t|�}
|
r^|
�dddd�\}}}}|||f||||d�|����S)Nr�r��ver�pyverrP)r.rrUrK)r�r�rQrY�_distributionImpl�EGG_NAMErN�_reload_version)rrr�r7r*r.rrUrKrTrIrrrrJ
s"����zDistribution.from_locationcCs|Srrr�rrrr�#
szDistribution._reload_versioncCs(|j|j|jt|j�|jpd|jp$dfSr�)�parsed_versionrtr;r�rrUrKr�rrrr\&
s�zDistribution.hashcmpcCs
t|j�Sr)�hashr\r�rrr�__hash__1
szDistribution.__hash__cCs|j|jkSr�r\�r�rerrr�__lt__4
szDistribution.__lt__cCs|j|jkSrr�r�rrr�__le__7
szDistribution.__le__cCs|j|jkSrr�r�rrr�__gt__:
szDistribution.__gt__cCs|j|jkSrr�r�rrr�__ge__=
szDistribution.__ge__cCst||j�sdS|j|jkSr�)r�r�r\r�rrr�__eq__@
szDistribution.__eq__cCs
||kSrrr�rrr�__ne__F
szDistribution.__ne__cCs4z|jWSty.|j��|_}|YS0dSr)�_keyr-r.rYrdrrrr;M
s
zDistribution.keycCst|d�st|j�|_|jS)N�_parsed_version)r�r#rr�r�rrrr�U
s
zDistribution.parsed_versioncCs\tjj}t|j|�}|sdS|js&dSt�d����dd�}t	�
|jfit|���t
�dS)Na>
            '{project_name} ({version})' is being parsed as a legacy,
            non PEP 440,
            version. You may find odd behavior and sort order.
            In particular it will be sorted as less than 0.0. It
            is recommended to migrate to PEP 440 compatible
            versions.
            r�r�)rrr!r�r�rqrrrvr�r�r�r��varsr)r��LV�	is_legacyrwrrr�_warn_legacy_version\
s�	z!Distribution._warn_legacy_versioncCsXz|jWStyR|��}|durJ|�|j�}d�|j|�}t||��|YS0dS)Nz4Missing 'Version:' header and/or {} file at path: {})r�r-�_get_version�_get_metadata_path_for_display�PKG_INFOr�rO)r�rr�r�rrrrv
s��
zDistribution.versioncCs2z|jWSty*|�|���|_Yn0|jS)z~
        A map of extra to its list of (direct) requirements
        for this distribution, including the null extra.
        )�_Distribution__dep_mapr-�_filter_extras�_build_dep_mapr�rrr�_dep_map�
s
zDistribution._dep_mapcCsrttd|��D]^}|}|�|�}|�d�\}}}|oDt|�pDt|�}|rNg}t|�pXd}|�|g��|�q|S)z�
        Given a mapping of extras to dependencies, strip off
        environment markers and filter out any dependencies
        not matching the markers.
        Nr�)	r$rr)r�r|r}rzrr-)�dmrJ�	new_extrarr�rL�fails_markerrrrr��
s
�zDistribution._filter_extrascCs@i}dD]2}t|�|��D]\}}|�|g��t|��qq|S)N)zrequires.txtzdepends.txt)ry�
_get_metadatarr-rs)r�r�r�rJrrrrr��
s
zDistribution._build_dep_maprc	Csd|j}g}|�|�dd��|D]>}z|�|t|��Wq ty\td||f��Yq 0q |S)z@List of Requirements needed for this distro if `extras` are usedNrz%s has no such extra feature %r)r�r-r�rzr�rq)r�r,r��depsrTrrrr�
s
�zDistribution.requirescCs*z|j�|�}Wnty$YdS0|S)zK
        Return the path to the given metadata file, if available.
        z[could not detect])r�r�r{r�rrrr��
s
z+Distribution._get_metadata_path_for_displayccs$|�|�r |�|�D]
}|VqdSr)r�r�)r�r�rwrrrr��
s
zDistribution._get_metadatacCs|�|j�}t|�}|Sr)r�r�r�)r�r�rrrrr��
szDistribution._get_versionFcCsV|durtj}|j||d�|tjurRt|j�|�d�D]}|tjvr:t|�q:dS)z>Ensure distribution is importable on `path` (default=sys.path)Nr�namespace_packages.txt)rJr�rr�rr�r�ra)r�r�r��pkgrrr�activate�
s


zDistribution.activatecCs8dt|j�t|j�|jptf}|jr4|d|j7}|S)z@Return what this distribution's standard .egg filename should bez
%s-%s-py%sr�)r{r.rrUrkrK)r�r�rrrr��
s�zDistribution.egg_namecCs |jrd||jfSt|�SdS)Nz%s (%s))rr�r�rrrr��
szDistribution.__repr__cCs>zt|dd�}Wnty&d}Yn0|p.d}d|j|fS)Nrz[unknown version]z%s %s)r�rOr.)r�rrrrr��
s
zDistribution.__str__cCs|�d�rt|��t|j|�S)zADelegate all unrecognized public attributes to .metadata providerr�)r�r-r�r�)r�r�rrr�__getattr__�
s
zDistribution.__getattr__cs.tttt|����tdd�|j��D��B�S)Ncss|]}|�d�s|VqdS�r�N)r�)rr�rrrrs
�z'Distribution.__dir__.<locals>.<genexpr>)r$r(�superrk�__dir__r�r�rhrrrs���zDistribution.__dir__rcKs"|jt|�tj�|�|fi|��Sr)rJrZr�r�r�)rr�r7r*rrrrFs
��zDistribution.from_filenamecCs<t|jtjj�r"d|j|jf}nd|j|jf}t�|�S)z?Return a ``Requirement`` that matches this distribution exactlyz%s==%sz%s===%s)r�r�rrrr.rlr�)r��specrrrr:szDistribution.as_requirementcCs.|�||�}|dur&td||ff��|��S)z=Return the `name` entry point of `group` or raise ImportErrorNzEntry point %r not found)rYrr�)r�rNr�r�rrrrWszDistribution.load_entry_pointcCsNz
|j}Wn*ty4t�|�d�|�}|_Yn0|durJ|�|i�S|S)r�zentry_points.txtN)�_ep_mapr-rmr�r�r�)r�rN�ep_maprrrrX&s

�zDistribution.get_entry_mapcCs|�|��|�Sr�)rXr�rrrrrY2szDistribution.get_entry_infoc
Cs2|p|j}|sdSt|�}tj�|�}dd�|D�}t|�D]|\}}||kr^|rVq�q�dSq<||kr<|jtkr<|s�|||d�vr�dS|tjur�|�	�|�
||�|�
||�q�q<|tjur�|�	�|r�|�
d|�n
|�|�dSz|�||d�}	Wnt
�yY�q.Yq�0||	=||	=|	}q�dS)a�Ensure self.location is on path

        If replace=False (default):
            - If location is already in path anywhere, do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent.
              - Else: add to the end of path.
        If replace=True:
            - If location is already on path anywhere (not eggs)
              or higher priority than its parent (eggs)
              do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent,
                removing any lower-priority entries.
              - Else: add it to the front of path.
        NcSsg|]}|rt|�p|�qSrr�r�rrrrPPrCz*Distribution.insert_on.<locals>.<listcomp>rr)rrZr�r�r��	enumeratertr�rJ�check_version_conflictr r�r�rO)
r�r��locr��nloc�bdir�npathr�r�nprrrr6s@



zDistribution.insert_oncCs�|jdkrdSt�|�d��}t|j�}|�d�D]p}|tjvs2||vs2|tvrRq2|dvr\q2t	tj|dd�}|r�t|��
|�s2|�
|j�r�q2td|||jf�q2dS)N�
setuptoolsr�z
top_level.txt)�
pkg_resourcesr�siter�zIModule %s was already imported from %s, but %s is being added to sys.path)r;r'r(r�rrrJr�r�r�r��
issue_warning)r��nspr�modname�fnrrrr
zs*

�
�
��z#Distribution.check_version_conflictcCs4z
|jWn$ty.tdt|��YdS0dS)NzUnbuilt egg for FT)rrOrr�r�rrrr_�s
zDistribution.has_versioncKsDd}|��D]}|�|t||d��q|�d|j�|jfi|��S)z@Copy this distribution, substituting in any changed keyword argsz<project_name version py_version platform location precedenceNr7)rrr�r�r�)r�r*r~r�rrr�clone�s
zDistribution.clonecCsdd�|jD�S)NcSsg|]}|r|�qSrr)r�deprrrrP�rCz'Distribution.extras.<locals>.<listcomp>)r�r�rrrr,�szDistribution.extras)N)r)NF)N)N)NF)3rrrrr�rkr�rrIrJr�r�r\r�r�r�r�r�r�r�r;r�r�rr�r�r�r�rr�r�r�rr�r�r�rrr��objectrFr:rWrXrYrr
r_rr,�
__classcell__rrrhrrk
sl�










		

Drkc@seZdZdd�ZdS)�EggInfoDistributioncCs|��}|r||_|S)a�
        Packages installed by distutils (e.g. numpy or scipy),
        which uses an old safe_version, and so
        their version numbers can get mangled when
        converted to filenames (e.g., 1.11.0.dev0+2329eae to
        1.11.0.dev0_2329eae). These distributions will not be
        parsed properly
        downstream by Distribution and safe_version, so
        take an extra step and try to get the version number from
        the metadata file itself instead of the filename.
        )r�r�)r��
md_versionrrrr��sz#EggInfoDistribution._reload_versionN)rrrr�rrrrr�src@s>eZdZdZdZe�d�Zedd��Z	edd��Z
dd	�Zd
S)�DistInfoDistributionzV
    Wrap an actual or potential sys.path entry
    w/metadata, .dist-info style.
    �METADATAz([\(,])\s*(\d.*?)\s*([,\)])cCsDz|jWSty>|�|j�}tj���|�|_|jYS0dS)zParse and cache metadataN)�	_pkg_infor-r�r��email�parser�Parser�parsestr)r�r7rrr�_parsed_pkg_info�sz%DistInfoDistribution._parsed_pkg_infocCs0z|jWSty*|��|_|jYS0dSr)�_DistInfoDistribution__dep_mapr-�_compute_dependenciesr�rrrr��s

zDistInfoDistribution._dep_mapcs�dgi}|_g�|j�d�p gD]}��t|��q"�fdd�}t|d��}|d�|�|j�d�pjgD](}t|���}tt||��|�||<ql|S)z+Recompute this distribution's dependencies.Nz
Requires-Distc3s*�D] }|jr|j�d|i�r|VqdS)NrJrK)rJr��rrr�reqs_for_extra�szBDistInfoDistribution._compute_dependencies.<locals>.reqs_for_extrazProvides-Extra)	r%r$�get_allr-rs�	frozensetrzrvr$)r�r�r�r(�commonrJ�s_extrarr'rr&�sz*DistInfoDistribution._compute_dependenciesN)rrrrr�r�r��EQEQr�r$r�r&rrrrr�s

	
r)r�rbrDcOsXd}t�}zt�|�j|ur&|d7}qWnty:Yn0tj|d|di|��dS)Nrr�)r$rJr�r�rOr�r�)rAr*r�r0rrrr�src@seZdZdd�ZdS)�RequirementParseErrorcCsd�|j�S)Nr�)rLrAr�rrrr��szRequirementParseError.__str__N)rrrr�rrrrr.�sr.c	cs�tt|��}|D]j}d|vr.|d|�d��}|�d�rp|dd���}z|t|�7}WntynYdS0t|�VqdS)z�Yield ``Requirement`` objects for each specification in `strs`

    `strs` must be a string, or a (possibly-nested) iterable thereof.
    z #N�\���)rhrxr�rErvrz�
StopIterationrl)r�r�rwrrrrss

rscsPeZdZ�fdd�Zdd�Zdd�Zdd�Zd	d
�Zdd�Ze	d
d��Z
�ZS)rlc
s�ztt|��|�Wn4tjjyH}ztt|���WYd}~n
d}~00|j|_	t
|j�}||��|_|_
dd�|jD�|_ttt|j��|_|j
|j|jt|j�|jr�t|j�ndf|_t|j�|_dS)z>DO NOT CALL THIS UNDOCUMENTED METHOD; use Requirement.parse()!NcSsg|]}|j|jf�qSr)r`r)rrrrrrP#sz(Requirement.__init__.<locals>.<listcomp>)rrlrrr/�InvalidRequirementr.r�r��unsafe_namertrYr.r;�	specifierr�r�rrzr,�urlr*rL�hashCmpr��_Requirement__hash)r��requirement_stringr�r.rhrrrs$"
��zRequirement.__init__cCst|t�o|j|jkSr)r�rlr6r�rrrr�/s

�zRequirement.__eq__cCs
||kSrrr�rrrr�5szRequirement.__ne__cCs0t|t�r |j|jkrdS|j}|jj|dd�S)NFT)�prereleases)r�rkr;rr4�contains)r�rrrrr8s

zRequirement.__contains__cCs|jSr)r7r�rrrr�DszRequirement.__hash__cCsdt|�S)NzRequirement.parse(%r)r�r�rrrr�GszRequirement.__repr__cCst|�\}|Sr)rs)r�r�rrrr�Js
zRequirement.parse)rrrrr�r�rr�r�r�r�rrrrhrrlsrlcCst|vr|tfS|S)zJ
    Ensure object appears in the mro even
    for old-style classes.
    )r)�classesrrr�_always_objectPs
r<cCs<tt�t|dt|����}|D]}||vr||SqdS)z2Return an adapter factory for `ob` from `registry`r�N)r<�inspect�getmror�r�)�registryr<r��trrrr�Zsr�cCstj�|�}tj|dd�dS)z1Ensure that the parent directory of `path` existsT)�exist_okN)r�r�r�r�makedirs)r�r�rrrr~bsr~cCsVtstd��t|�\}}|rR|rRt|�sRt|�zt|d�WntyPYn0dS)z/Sandbox-bypassing version of ensure_directory()z*"os.mkdir" not supported on this platform.i�N)rrrr
rzr	�FileExistsError)r�r�r�rrrrzhsrzccsvd}g}t|�D]V}|�d�r\|�d�rP|s0|r:||fV|dd���}g}qftd|��q|�|�q||fVdS)asSplit a string or iterable thereof into (section, content) pairs

    Each ``section`` is a stripped version of the section header ("[section]")
    and each ``content`` is a list of stripped lines excluding blank lines and
    comment-only lines.  If there are any such lines before the first section
    header, they're returned in a first ``section`` of ``None``.
    N�[�]rr�zInvalid section heading)rxr�rErvrOr�)r��section�contentrwrrrryus


rycOs2tj}ztt_tj|i|��W|t_S|t_0dSr)r�r�os_open�tempfile�mkstemp)rAr*�old_openrrrr�s�rr�)�categoryr�cOs||i|��|Srr)r*rAr�rrr�_call_aside�srMcs.t���|d<|��fdd�t��D��dS)z=Set up global resource manager (deliberately not state-saved)�_managerc3s&|]}|�d�s|t�|�fVqdSr)r�r�r��r�rrr�s
�z_initialize.<locals>.<genexpr>N)rjr%r)r0rrOr�_initialize�s
�rPcCs|t��}td|d�|j}|j}|j}|j}|}tdd�|D��|dd�dd�g|_t	t
|jtj
��t��t��d	S)
aE
    Prepare the master working set and make the ``require()``
    API available.

    This function has explicit effects on the global state
    of pkg_resources. It is intended to be invoked once at
    the initialization of this module.

    Invocation by other packages is unsupported and done
    at their own risk.
    r)rbcss|]}|jdd�VqdS)FrN�r)rr�rrrr�s�z1_initialize_master_working_set.<locals>.<genexpr>cSs|jdd�S)NTrrQr�rrrrB�rCz0_initialize_master_working_set.<locals>.<lambda>F)rFN)rir
r+rSrZrGrTr�r�r$rrrJr�r$r%r�)rbrSrZrcrTr�rrr�_initialize_master_working_set�s"
��rRc@seZdZdZdS)r�z�
    Base class for warning about deprecations in ``pkg_resources``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    Nrrrrrr��sr�)N)N)F)F)F)F)N)�r�
__future__rrJr�r�rr�r�r�rHr�r�r��pkgutilr`rKr&r��email.parserr rorIrqrRr=r�r�rr�r�imprC�	NameErrorrn�pip._vendorr�pip._vendor.six.movesrrrrr	r
rrrrH�os.pathr
r�importlib.machinery�	machineryr�rr�rrrr�r��
__metaclass__�version_info�RuntimeErrorr�rlrmrSrbrc�resources_streamrf�resource_dirr\rer`r[rZr^r]r_r=r�r��RuntimeWarningrr#r&r+r2r4r8r=r>r?�
_sget_none�
_sset_nonerR�__all__r{rnror�rprqr�r�rkr�r�r�r�r�r�rUrMr�rGr�rHr�rvrwrTr�rVrWrXrYr�r�rir'r%rhr�rrrjrgrtrurzr{r|r}r�rr�r�r�r�r�r�r�r�rIr�r�r�r�rdrGrNrWrarXrer\rcrxrd�ImpImporterr�r~r�r�r�rar�r�r�rr�rZr�r[r�rxrIr��VERBOSE�
IGNORECASEr�rmr�r�rkrrr�rrOr.rsr/rlr<r�r~rzryr�filterwarningsrMr$rPrR�Warningr�rrrr�<module>s�



2 




.

5	
A
-*





 ""



	
�	
'3�
7


&site-packages/pip/_vendor/pkg_resources/__pycache__/py31compat.cpython-39.pyc000064400000001261150564745010023233 0ustar00a

��Re2�@s`ddlZddlZddlZddlmZddd�ZejpLdejkoHdknZerVenej	Z	dS)	�N)�sixFc
CsJzt�|�Wn6tyD}z|r.|jtjkr0�WYd}~n
d}~00dS)N)�os�makedirs�OSError�errno�EEXIST)�path�exist_ok�exc�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/pkg_resources/py31compat.py�_makedirs_31s
r
)��)rr�)F)
rr�sys�pip._vendorrr
�PY2�version_info�needs_makedirsrrrrr�<module>s
�site-packages/pip/_vendor/pkg_resources/py31compat.py000064400000001062150564745010016743 0ustar00import os
import errno
import sys

from pip._vendor import six


def _makedirs_31(path, exist_ok=False):
    try:
        os.makedirs(path)
    except OSError as exc:
        if not exist_ok or exc.errno != errno.EEXIST:
            raise


# rely on compatibility behavior until mode considerations
#  and exists_ok considerations are disentangled.
# See https://github.com/pypa/setuptools/pull/1083#issuecomment-315168663
needs_makedirs = (
    six.PY2 or
    (3, 4) <= sys.version_info < (3, 4, 1)
)
makedirs = _makedirs_31 if needs_makedirs else os.makedirs
site-packages/pip/_vendor/pkg_resources/__init__.py000064400000323377150564745010016522 0ustar00# coding: utf-8
"""
Package resource API
--------------------

A resource is a logical file contained within a package, or a logical
subdirectory thereof.  The package resource API expects resource names
to have their path parts separated with ``/``, *not* whatever the local
path separator is.  Do not use os.path operations to manipulate resource
names being passed into the API.

The package resource API is designed to work with normal filesystem packages,
.egg files, and unpacked .egg files.  It can also work in a limited way with
.zip files and with custom PEP 302 loaders that support the ``get_data()``
method.
"""

from __future__ import absolute_import

import sys
import os
import io
import time
import re
import types
import zipfile
import zipimport
import warnings
import stat
import functools
import pkgutil
import operator
import platform
import collections
import plistlib
import email.parser
import errno
import tempfile
import textwrap
import itertools
import inspect
import ntpath
import posixpath
from pkgutil import get_importer

try:
    import _imp
except ImportError:
    # Python 3.2 compatibility
    import imp as _imp

try:
    FileExistsError
except NameError:
    FileExistsError = OSError

from pip._vendor import six
from pip._vendor.six.moves import urllib, map, filter

# capture these to bypass sandboxing
from os import utime
try:
    from os import mkdir, rename, unlink
    WRITE_SUPPORT = True
except ImportError:
    # no write support, probably under GAE
    WRITE_SUPPORT = False

from os import open as os_open
from os.path import isdir, split

try:
    import importlib.machinery as importlib_machinery
    # access attribute to force import under delayed import mechanisms.
    importlib_machinery.__name__
except ImportError:
    importlib_machinery = None

from . import py31compat
from pip._vendor import platformdirs
from pip._vendor import packaging
__import__('pip._vendor.packaging.version')
__import__('pip._vendor.packaging.specifiers')
__import__('pip._vendor.packaging.requirements')
__import__('pip._vendor.packaging.markers')


__metaclass__ = type


if (3, 0) < sys.version_info < (3, 5):
    raise RuntimeError("Python 3.5 or later is required")

if six.PY2:
    # Those builtin exceptions are only defined in Python 3
    PermissionError = None
    NotADirectoryError = None

# declare some globals that will be defined later to
# satisfy the linters.
require = None
working_set = None
add_activation_listener = None
resources_stream = None
cleanup_resources = None
resource_dir = None
resource_stream = None
set_extraction_path = None
resource_isdir = None
resource_string = None
iter_entry_points = None
resource_listdir = None
resource_filename = None
resource_exists = None
_distribution_finders = None
_namespace_handlers = None
_namespace_packages = None


class PEP440Warning(RuntimeWarning):
    """
    Used when there is an issue with a version or specifier not complying with
    PEP 440.
    """


def parse_version(v):
    try:
        return packaging.version.Version(v)
    except packaging.version.InvalidVersion:
        return packaging.version.LegacyVersion(v)


_state_vars = {}


def _declare_state(vartype, **kw):
    globals().update(kw)
    _state_vars.update(dict.fromkeys(kw, vartype))


def __getstate__():
    state = {}
    g = globals()
    for k, v in _state_vars.items():
        state[k] = g['_sget_' + v](g[k])
    return state


def __setstate__(state):
    g = globals()
    for k, v in state.items():
        g['_sset_' + _state_vars[k]](k, g[k], v)
    return state


def _sget_dict(val):
    return val.copy()


def _sset_dict(key, ob, state):
    ob.clear()
    ob.update(state)


def _sget_object(val):
    return val.__getstate__()


def _sset_object(key, ob, state):
    ob.__setstate__(state)


_sget_none = _sset_none = lambda *args: None


def get_supported_platform():
    """Return this platform's maximum compatible version.

    distutils.util.get_platform() normally reports the minimum version
    of Mac OS X that would be required to *use* extensions produced by
    distutils.  But what we want when checking compatibility is to know the
    version of Mac OS X that we are *running*.  To allow usage of packages that
    explicitly require a newer version of Mac OS X, we must also know the
    current version of the OS.

    If this condition occurs for any other platform with a version in its
    platform strings, this function should be extended accordingly.
    """
    plat = get_build_platform()
    m = macosVersionString.match(plat)
    if m is not None and sys.platform == "darwin":
        try:
            plat = 'macosx-%s-%s' % ('.'.join(_macosx_vers()[:2]), m.group(3))
        except ValueError:
            # not Mac OS X
            pass
    return plat


__all__ = [
    # Basic resource access and distribution/entry point discovery
    'require', 'run_script', 'get_provider', 'get_distribution',
    'load_entry_point', 'get_entry_map', 'get_entry_info',
    'iter_entry_points',
    'resource_string', 'resource_stream', 'resource_filename',
    'resource_listdir', 'resource_exists', 'resource_isdir',

    # Environmental control
    'declare_namespace', 'working_set', 'add_activation_listener',
    'find_distributions', 'set_extraction_path', 'cleanup_resources',
    'get_default_cache',

    # Primary implementation classes
    'Environment', 'WorkingSet', 'ResourceManager',
    'Distribution', 'Requirement', 'EntryPoint',

    # Exceptions
    'ResolutionError', 'VersionConflict', 'DistributionNotFound',
    'UnknownExtra', 'ExtractionError',

    # Warnings
    'PEP440Warning',

    # Parsing functions and string utilities
    'parse_requirements', 'parse_version', 'safe_name', 'safe_version',
    'get_platform', 'compatible_platforms', 'yield_lines', 'split_sections',
    'safe_extra', 'to_filename', 'invalid_marker', 'evaluate_marker',

    # filesystem utilities
    'ensure_directory', 'normalize_path',

    # Distribution "precedence" constants
    'EGG_DIST', 'BINARY_DIST', 'SOURCE_DIST', 'CHECKOUT_DIST', 'DEVELOP_DIST',

    # "Provider" interfaces, implementations, and registration/lookup APIs
    'IMetadataProvider', 'IResourceProvider', 'FileMetadata',
    'PathMetadata', 'EggMetadata', 'EmptyProvider', 'empty_provider',
    'NullProvider', 'EggProvider', 'DefaultProvider', 'ZipProvider',
    'register_finder', 'register_namespace_handler', 'register_loader_type',
    'fixup_namespace_packages', 'get_importer',

    # Warnings
    'PkgResourcesDeprecationWarning',

    # Deprecated/backward compatibility only
    'run_main', 'AvailableDistributions',
]


class ResolutionError(Exception):
    """Abstract base for dependency resolution errors"""

    def __repr__(self):
        return self.__class__.__name__ + repr(self.args)


class VersionConflict(ResolutionError):
    """
    An already-installed version conflicts with the requested version.

    Should be initialized with the installed Distribution and the requested
    Requirement.
    """

    _template = "{self.dist} is installed but {self.req} is required"

    @property
    def dist(self):
        return self.args[0]

    @property
    def req(self):
        return self.args[1]

    def report(self):
        return self._template.format(**locals())

    def with_context(self, required_by):
        """
        If required_by is non-empty, return a version of self that is a
        ContextualVersionConflict.
        """
        if not required_by:
            return self
        args = self.args + (required_by,)
        return ContextualVersionConflict(*args)


class ContextualVersionConflict(VersionConflict):
    """
    A VersionConflict that accepts a third parameter, the set of the
    requirements that required the installed Distribution.
    """

    _template = VersionConflict._template + ' by {self.required_by}'

    @property
    def required_by(self):
        return self.args[2]


class DistributionNotFound(ResolutionError):
    """A requested distribution was not found"""

    _template = ("The '{self.req}' distribution was not found "
                 "and is required by {self.requirers_str}")

    @property
    def req(self):
        return self.args[0]

    @property
    def requirers(self):
        return self.args[1]

    @property
    def requirers_str(self):
        if not self.requirers:
            return 'the application'
        return ', '.join(self.requirers)

    def report(self):
        return self._template.format(**locals())

    def __str__(self):
        return self.report()


class UnknownExtra(ResolutionError):
    """Distribution doesn't have an "extra feature" of the given name"""


_provider_factories = {}

PY_MAJOR = '{}.{}'.format(*sys.version_info)
EGG_DIST = 3
BINARY_DIST = 2
SOURCE_DIST = 1
CHECKOUT_DIST = 0
DEVELOP_DIST = -1


def register_loader_type(loader_type, provider_factory):
    """Register `provider_factory` to make providers for `loader_type`

    `loader_type` is the type or class of a PEP 302 ``module.__loader__``,
    and `provider_factory` is a function that, passed a *module* object,
    returns an ``IResourceProvider`` for that module.
    """
    _provider_factories[loader_type] = provider_factory


def get_provider(moduleOrReq):
    """Return an IResourceProvider for the named module or requirement"""
    if isinstance(moduleOrReq, Requirement):
        return working_set.find(moduleOrReq) or require(str(moduleOrReq))[0]
    try:
        module = sys.modules[moduleOrReq]
    except KeyError:
        __import__(moduleOrReq)
        module = sys.modules[moduleOrReq]
    loader = getattr(module, '__loader__', None)
    return _find_adapter(_provider_factories, loader)(module)


def _macosx_vers(_cache=[]):
    if not _cache:
        version = platform.mac_ver()[0]
        # fallback for MacPorts
        if version == '':
            plist = '/System/Library/CoreServices/SystemVersion.plist'
            if os.path.exists(plist):
                if hasattr(plistlib, 'readPlist'):
                    plist_content = plistlib.readPlist(plist)
                    if 'ProductVersion' in plist_content:
                        version = plist_content['ProductVersion']

        _cache.append(version.split('.'))
    return _cache[0]


def _macosx_arch(machine):
    return {'PowerPC': 'ppc', 'Power_Macintosh': 'ppc'}.get(machine, machine)


def get_build_platform():
    """Return this platform's string for platform-specific distributions

    XXX Currently this is the same as ``distutils.util.get_platform()``, but it
    needs some hacks for Linux and Mac OS X.
    """
    from sysconfig import get_platform

    plat = get_platform()
    if sys.platform == "darwin" and not plat.startswith('macosx-'):
        try:
            version = _macosx_vers()
            machine = os.uname()[4].replace(" ", "_")
            return "macosx-%d.%d-%s" % (
                int(version[0]), int(version[1]),
                _macosx_arch(machine),
            )
        except ValueError:
            # if someone is running a non-Mac darwin system, this will fall
            # through to the default implementation
            pass
    return plat


macosVersionString = re.compile(r"macosx-(\d+)\.(\d+)-(.*)")
darwinVersionString = re.compile(r"darwin-(\d+)\.(\d+)\.(\d+)-(.*)")
# XXX backward compat
get_platform = get_build_platform


def compatible_platforms(provided, required):
    """Can code for the `provided` platform run on the `required` platform?

    Returns true if either platform is ``None``, or the platforms are equal.

    XXX Needs compatibility checks for Linux and other unixy OSes.
    """
    if provided is None or required is None or provided == required:
        # easy case
        return True

    # Mac OS X special cases
    reqMac = macosVersionString.match(required)
    if reqMac:
        provMac = macosVersionString.match(provided)

        # is this a Mac package?
        if not provMac:
            # this is backwards compatibility for packages built before
            # setuptools 0.6. All packages built after this point will
            # use the new macosx designation.
            provDarwin = darwinVersionString.match(provided)
            if provDarwin:
                dversion = int(provDarwin.group(1))
                macosversion = "%s.%s" % (reqMac.group(1), reqMac.group(2))
                if dversion == 7 and macosversion >= "10.3" or \
                        dversion == 8 and macosversion >= "10.4":
                    return True
            # egg isn't macosx or legacy darwin
            return False

        # are they the same major version and machine type?
        if provMac.group(1) != reqMac.group(1) or \
                provMac.group(3) != reqMac.group(3):
            return False

        # is the required OS major update >= the provided one?
        if int(provMac.group(2)) > int(reqMac.group(2)):
            return False

        return True

    # XXX Linux and other platforms' special cases should go here
    return False


def run_script(dist_spec, script_name):
    """Locate distribution `dist_spec` and run its `script_name` script"""
    ns = sys._getframe(1).f_globals
    name = ns['__name__']
    ns.clear()
    ns['__name__'] = name
    require(dist_spec)[0].run_script(script_name, ns)


# backward compatibility
run_main = run_script


def get_distribution(dist):
    """Return a current distribution object for a Requirement or string"""
    if isinstance(dist, six.string_types):
        dist = Requirement.parse(dist)
    if isinstance(dist, Requirement):
        dist = get_provider(dist)
    if not isinstance(dist, Distribution):
        raise TypeError("Expected string, Requirement, or Distribution", dist)
    return dist


def load_entry_point(dist, group, name):
    """Return `name` entry point of `group` for `dist` or raise ImportError"""
    return get_distribution(dist).load_entry_point(group, name)


def get_entry_map(dist, group=None):
    """Return the entry point map for `group`, or the full entry map"""
    return get_distribution(dist).get_entry_map(group)


def get_entry_info(dist, group, name):
    """Return the EntryPoint object for `group`+`name`, or ``None``"""
    return get_distribution(dist).get_entry_info(group, name)


class IMetadataProvider:
    def has_metadata(name):
        """Does the package's distribution contain the named metadata?"""

    def get_metadata(name):
        """The named metadata resource as a string"""

    def get_metadata_lines(name):
        """Yield named metadata resource as list of non-blank non-comment lines

       Leading and trailing whitespace is stripped from each line, and lines
       with ``#`` as the first non-blank character are omitted."""

    def metadata_isdir(name):
        """Is the named metadata a directory?  (like ``os.path.isdir()``)"""

    def metadata_listdir(name):
        """List of metadata names in the directory (like ``os.listdir()``)"""

    def run_script(script_name, namespace):
        """Execute the named script in the supplied namespace dictionary"""


class IResourceProvider(IMetadataProvider):
    """An object that provides access to package resources"""

    def get_resource_filename(manager, resource_name):
        """Return a true filesystem path for `resource_name`

        `manager` must be an ``IResourceManager``"""

    def get_resource_stream(manager, resource_name):
        """Return a readable file-like object for `resource_name`

        `manager` must be an ``IResourceManager``"""

    def get_resource_string(manager, resource_name):
        """Return a string containing the contents of `resource_name`

        `manager` must be an ``IResourceManager``"""

    def has_resource(resource_name):
        """Does the package contain the named resource?"""

    def resource_isdir(resource_name):
        """Is the named resource a directory?  (like ``os.path.isdir()``)"""

    def resource_listdir(resource_name):
        """List of resource names in the directory (like ``os.listdir()``)"""


class WorkingSet:
    """A collection of active distributions on sys.path (or a similar list)"""

    def __init__(self, entries=None):
        """Create working set from list of path entries (default=sys.path)"""
        self.entries = []
        self.entry_keys = {}
        self.by_key = {}
        self.callbacks = []

        if entries is None:
            entries = sys.path

        for entry in entries:
            self.add_entry(entry)

    @classmethod
    def _build_master(cls):
        """
        Prepare the master working set.
        """
        ws = cls()
        try:
            from __main__ import __requires__
        except ImportError:
            # The main program does not list any requirements
            return ws

        # ensure the requirements are met
        try:
            ws.require(__requires__)
        except VersionConflict:
            return cls._build_from_requirements(__requires__)

        return ws

    @classmethod
    def _build_from_requirements(cls, req_spec):
        """
        Build a working set from a requirement spec. Rewrites sys.path.
        """
        # try it without defaults already on sys.path
        # by starting with an empty path
        ws = cls([])
        reqs = parse_requirements(req_spec)
        dists = ws.resolve(reqs, Environment())
        for dist in dists:
            ws.add(dist)

        # add any missing entries from sys.path
        for entry in sys.path:
            if entry not in ws.entries:
                ws.add_entry(entry)

        # then copy back to sys.path
        sys.path[:] = ws.entries
        return ws

    def add_entry(self, entry):
        """Add a path item to ``.entries``, finding any distributions on it

        ``find_distributions(entry, True)`` is used to find distributions
        corresponding to the path entry, and they are added.  `entry` is
        always appended to ``.entries``, even if it is already present.
        (This is because ``sys.path`` can contain the same value more than
        once, and the ``.entries`` of the ``sys.path`` WorkingSet should always
        equal ``sys.path``.)
        """
        self.entry_keys.setdefault(entry, [])
        self.entries.append(entry)
        for dist in find_distributions(entry, True):
            self.add(dist, entry, False)

    def __contains__(self, dist):
        """True if `dist` is the active distribution for its project"""
        return self.by_key.get(dist.key) == dist

    def find(self, req):
        """Find a distribution matching requirement `req`

        If there is an active distribution for the requested project, this
        returns it as long as it meets the version requirement specified by
        `req`.  But, if there is an active distribution for the project and it
        does *not* meet the `req` requirement, ``VersionConflict`` is raised.
        If there is no active distribution for the requested project, ``None``
        is returned.
        """
        dist = self.by_key.get(req.key)
        if dist is not None and dist not in req:
            # XXX add more info
            raise VersionConflict(dist, req)
        return dist

    def iter_entry_points(self, group, name=None):
        """Yield entry point objects from `group` matching `name`

        If `name` is None, yields all entry points in `group` from all
        distributions in the working set, otherwise only ones matching
        both `group` and `name` are yielded (in distribution order).
        """
        return (
            entry
            for dist in self
            for entry in dist.get_entry_map(group).values()
            if name is None or name == entry.name
        )

    def run_script(self, requires, script_name):
        """Locate distribution for `requires` and run `script_name` script"""
        ns = sys._getframe(1).f_globals
        name = ns['__name__']
        ns.clear()
        ns['__name__'] = name
        self.require(requires)[0].run_script(script_name, ns)

    def __iter__(self):
        """Yield distributions for non-duplicate projects in the working set

        The yield order is the order in which the items' path entries were
        added to the working set.
        """
        seen = {}
        for item in self.entries:
            if item not in self.entry_keys:
                # workaround a cache issue
                continue

            for key in self.entry_keys[item]:
                if key not in seen:
                    seen[key] = 1
                    yield self.by_key[key]

    def add(self, dist, entry=None, insert=True, replace=False):
        """Add `dist` to working set, associated with `entry`

        If `entry` is unspecified, it defaults to the ``.location`` of `dist`.
        On exit from this routine, `entry` is added to the end of the working
        set's ``.entries`` (if it wasn't already present).

        `dist` is only added to the working set if it's for a project that
        doesn't already have a distribution in the set, unless `replace=True`.
        If it's added, any callbacks registered with the ``subscribe()`` method
        will be called.
        """
        if insert:
            dist.insert_on(self.entries, entry, replace=replace)

        if entry is None:
            entry = dist.location
        keys = self.entry_keys.setdefault(entry, [])
        keys2 = self.entry_keys.setdefault(dist.location, [])
        if not replace and dist.key in self.by_key:
            # ignore hidden distros
            return

        self.by_key[dist.key] = dist
        if dist.key not in keys:
            keys.append(dist.key)
        if dist.key not in keys2:
            keys2.append(dist.key)
        self._added_new(dist)

    def resolve(self, requirements, env=None, installer=None,
                replace_conflicting=False, extras=None):
        """List all distributions needed to (recursively) meet `requirements`

        `requirements` must be a sequence of ``Requirement`` objects.  `env`,
        if supplied, should be an ``Environment`` instance.  If
        not supplied, it defaults to all distributions available within any
        entry or distribution in the working set.  `installer`, if supplied,
        will be invoked with each requirement that cannot be met by an
        already-installed distribution; it should return a ``Distribution`` or
        ``None``.

        Unless `replace_conflicting=True`, raises a VersionConflict exception
        if
        any requirements are found on the path that have the correct name but
        the wrong version.  Otherwise, if an `installer` is supplied it will be
        invoked to obtain the correct version of the requirement and activate
        it.

        `extras` is a list of the extras to be used with these requirements.
        This is important because extra requirements may look like `my_req;
        extra = "my_extra"`, which would otherwise be interpreted as a purely
        optional requirement.  Instead, we want to be able to assert that these
        requirements are truly required.
        """

        # set up the stack
        requirements = list(requirements)[::-1]
        # set of processed requirements
        processed = {}
        # key -> dist
        best = {}
        to_activate = []

        req_extras = _ReqExtras()

        # Mapping of requirement to set of distributions that required it;
        # useful for reporting info about conflicts.
        required_by = collections.defaultdict(set)

        while requirements:
            # process dependencies breadth-first
            req = requirements.pop(0)
            if req in processed:
                # Ignore cyclic or redundant dependencies
                continue

            if not req_extras.markers_pass(req, extras):
                continue

            dist = best.get(req.key)
            if dist is None:
                # Find the best distribution and add it to the map
                dist = self.by_key.get(req.key)
                if dist is None or (dist not in req and replace_conflicting):
                    ws = self
                    if env is None:
                        if dist is None:
                            env = Environment(self.entries)
                        else:
                            # Use an empty environment and workingset to avoid
                            # any further conflicts with the conflicting
                            # distribution
                            env = Environment([])
                            ws = WorkingSet([])
                    dist = best[req.key] = env.best_match(
                        req, ws, installer,
                        replace_conflicting=replace_conflicting
                    )
                    if dist is None:
                        requirers = required_by.get(req, None)
                        raise DistributionNotFound(req, requirers)
                to_activate.append(dist)
            if dist not in req:
                # Oops, the "best" so far conflicts with a dependency
                dependent_req = required_by[req]
                raise VersionConflict(dist, req).with_context(dependent_req)

            # push the new requirements onto the stack
            new_requirements = dist.requires(req.extras)[::-1]
            requirements.extend(new_requirements)

            # Register the new requirements needed by req
            for new_requirement in new_requirements:
                required_by[new_requirement].add(req.project_name)
                req_extras[new_requirement] = req.extras

            processed[req] = True

        # return list of distros to activate
        return to_activate

    def find_plugins(
            self, plugin_env, full_env=None, installer=None, fallback=True):
        """Find all activatable distributions in `plugin_env`

        Example usage::

            distributions, errors = working_set.find_plugins(
                Environment(plugin_dirlist)
            )
            # add plugins+libs to sys.path
            map(working_set.add, distributions)
            # display errors
            print('Could not load', errors)

        The `plugin_env` should be an ``Environment`` instance that contains
        only distributions that are in the project's "plugin directory" or
        directories. The `full_env`, if supplied, should be an ``Environment``
        contains all currently-available distributions.  If `full_env` is not
        supplied, one is created automatically from the ``WorkingSet`` this
        method is called on, which will typically mean that every directory on
        ``sys.path`` will be scanned for distributions.

        `installer` is a standard installer callback as used by the
        ``resolve()`` method. The `fallback` flag indicates whether we should
        attempt to resolve older versions of a plugin if the newest version
        cannot be resolved.

        This method returns a 2-tuple: (`distributions`, `error_info`), where
        `distributions` is a list of the distributions found in `plugin_env`
        that were loadable, along with any other distributions that are needed
        to resolve their dependencies.  `error_info` is a dictionary mapping
        unloadable plugin distributions to an exception instance describing the
        error that occurred. Usually this will be a ``DistributionNotFound`` or
        ``VersionConflict`` instance.
        """

        plugin_projects = list(plugin_env)
        # scan project names in alphabetic order
        plugin_projects.sort()

        error_info = {}
        distributions = {}

        if full_env is None:
            env = Environment(self.entries)
            env += plugin_env
        else:
            env = full_env + plugin_env

        shadow_set = self.__class__([])
        # put all our entries in shadow_set
        list(map(shadow_set.add, self))

        for project_name in plugin_projects:

            for dist in plugin_env[project_name]:

                req = [dist.as_requirement()]

                try:
                    resolvees = shadow_set.resolve(req, env, installer)

                except ResolutionError as v:
                    # save error info
                    error_info[dist] = v
                    if fallback:
                        # try the next older version of project
                        continue
                    else:
                        # give up on this project, keep going
                        break

                else:
                    list(map(shadow_set.add, resolvees))
                    distributions.update(dict.fromkeys(resolvees))

                    # success, no need to try any more versions of this project
                    break

        distributions = list(distributions)
        distributions.sort()

        return distributions, error_info

    def require(self, *requirements):
        """Ensure that distributions matching `requirements` are activated

        `requirements` must be a string or a (possibly-nested) sequence
        thereof, specifying the distributions and versions required.  The
        return value is a sequence of the distributions that needed to be
        activated to fulfill the requirements; all relevant distributions are
        included, even if they were already activated in this working set.
        """
        needed = self.resolve(parse_requirements(requirements))

        for dist in needed:
            self.add(dist)

        return needed

    def subscribe(self, callback, existing=True):
        """Invoke `callback` for all distributions

        If `existing=True` (default),
        call on all existing ones, as well.
        """
        if callback in self.callbacks:
            return
        self.callbacks.append(callback)
        if not existing:
            return
        for dist in self:
            callback(dist)

    def _added_new(self, dist):
        for callback in self.callbacks:
            callback(dist)

    def __getstate__(self):
        return (
            self.entries[:], self.entry_keys.copy(), self.by_key.copy(),
            self.callbacks[:]
        )

    def __setstate__(self, e_k_b_c):
        entries, keys, by_key, callbacks = e_k_b_c
        self.entries = entries[:]
        self.entry_keys = keys.copy()
        self.by_key = by_key.copy()
        self.callbacks = callbacks[:]


class _ReqExtras(dict):
    """
    Map each requirement to the extras that demanded it.
    """

    def markers_pass(self, req, extras=None):
        """
        Evaluate markers for req against each extra that
        demanded it.

        Return False if the req has a marker and fails
        evaluation. Otherwise, return True.
        """
        extra_evals = (
            req.marker.evaluate({'extra': extra})
            for extra in self.get(req, ()) + (extras or (None,))
        )
        return not req.marker or any(extra_evals)


class Environment:
    """Searchable snapshot of distributions on a search path"""

    def __init__(
            self, search_path=None, platform=get_supported_platform(),
            python=PY_MAJOR):
        """Snapshot distributions available on a search path

        Any distributions found on `search_path` are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.

        `platform` is an optional string specifying the name of the platform
        that platform-specific distributions must be compatible with.  If
        unspecified, it defaults to the current platform.  `python` is an
        optional string naming the desired version of Python (e.g. ``'3.6'``);
        it defaults to the current version.

        You may explicitly set `platform` (and/or `python`) to ``None`` if you
        wish to map *all* distributions, not just those compatible with the
        running platform or Python version.
        """
        self._distmap = {}
        self.platform = platform
        self.python = python
        self.scan(search_path)

    def can_add(self, dist):
        """Is distribution `dist` acceptable for this environment?

        The distribution must match the platform and python version
        requirements specified when this environment was created, or False
        is returned.
        """
        py_compat = (
            self.python is None
            or dist.py_version is None
            or dist.py_version == self.python
        )
        return py_compat and compatible_platforms(dist.platform, self.platform)

    def remove(self, dist):
        """Remove `dist` from the environment"""
        self._distmap[dist.key].remove(dist)

    def scan(self, search_path=None):
        """Scan `search_path` for distributions usable in this environment

        Any distributions found are added to the environment.
        `search_path` should be a sequence of ``sys.path`` items.  If not
        supplied, ``sys.path`` is used.  Only distributions conforming to
        the platform/python version defined at initialization are added.
        """
        if search_path is None:
            search_path = sys.path

        for item in search_path:
            for dist in find_distributions(item):
                self.add(dist)

    def __getitem__(self, project_name):
        """Return a newest-to-oldest list of distributions for `project_name`

        Uses case-insensitive `project_name` comparison, assuming all the
        project's distributions use their project's name converted to all
        lowercase as their key.

        """
        distribution_key = project_name.lower()
        return self._distmap.get(distribution_key, [])

    def add(self, dist):
        """Add `dist` if we ``can_add()`` it and it has not already been added
        """
        if self.can_add(dist) and dist.has_version():
            dists = self._distmap.setdefault(dist.key, [])
            if dist not in dists:
                dists.append(dist)
                dists.sort(key=operator.attrgetter('hashcmp'), reverse=True)

    def best_match(
            self, req, working_set, installer=None, replace_conflicting=False):
        """Find distribution best matching `req` and usable on `working_set`

        This calls the ``find(req)`` method of the `working_set` to see if a
        suitable distribution is already active.  (This may raise
        ``VersionConflict`` if an unsuitable version of the project is already
        active in the specified `working_set`.)  If a suitable distribution
        isn't active, this method returns the newest distribution in the
        environment that meets the ``Requirement`` in `req`.  If no suitable
        distribution is found, and `installer` is supplied, then the result of
        calling the environment's ``obtain(req, installer)`` method will be
        returned.
        """
        try:
            dist = working_set.find(req)
        except VersionConflict:
            if not replace_conflicting:
                raise
            dist = None
        if dist is not None:
            return dist
        for dist in self[req.key]:
            if dist in req:
                return dist
        # try to download/install
        return self.obtain(req, installer)

    def obtain(self, requirement, installer=None):
        """Obtain a distribution matching `requirement` (e.g. via download)

        Obtain a distro that matches requirement (e.g. via download).  In the
        base ``Environment`` class, this routine just returns
        ``installer(requirement)``, unless `installer` is None, in which case
        None is returned instead.  This method is a hook that allows subclasses
        to attempt other ways of obtaining a distribution before falling back
        to the `installer` argument."""
        if installer is not None:
            return installer(requirement)

    def __iter__(self):
        """Yield the unique project names of the available distributions"""
        for key in self._distmap.keys():
            if self[key]:
                yield key

    def __iadd__(self, other):
        """In-place addition of a distribution or environment"""
        if isinstance(other, Distribution):
            self.add(other)
        elif isinstance(other, Environment):
            for project in other:
                for dist in other[project]:
                    self.add(dist)
        else:
            raise TypeError("Can't add %r to environment" % (other,))
        return self

    def __add__(self, other):
        """Add an environment or distribution to an environment"""
        new = self.__class__([], platform=None, python=None)
        for env in self, other:
            new += env
        return new


# XXX backward compatibility
AvailableDistributions = Environment


class ExtractionError(RuntimeError):
    """An error occurred extracting a resource

    The following attributes are available from instances of this exception:

    manager
        The resource manager that raised this exception

    cache_path
        The base directory for resource extraction

    original_error
        The exception instance that caused extraction to fail
    """


class ResourceManager:
    """Manage resource extraction and packages"""
    extraction_path = None

    def __init__(self):
        self.cached_files = {}

    def resource_exists(self, package_or_requirement, resource_name):
        """Does the named resource exist?"""
        return get_provider(package_or_requirement).has_resource(resource_name)

    def resource_isdir(self, package_or_requirement, resource_name):
        """Is the named resource an existing directory?"""
        return get_provider(package_or_requirement).resource_isdir(
            resource_name
        )

    def resource_filename(self, package_or_requirement, resource_name):
        """Return a true filesystem path for specified resource"""
        return get_provider(package_or_requirement).get_resource_filename(
            self, resource_name
        )

    def resource_stream(self, package_or_requirement, resource_name):
        """Return a readable file-like object for specified resource"""
        return get_provider(package_or_requirement).get_resource_stream(
            self, resource_name
        )

    def resource_string(self, package_or_requirement, resource_name):
        """Return specified resource as a string"""
        return get_provider(package_or_requirement).get_resource_string(
            self, resource_name
        )

    def resource_listdir(self, package_or_requirement, resource_name):
        """List the contents of the named resource directory"""
        return get_provider(package_or_requirement).resource_listdir(
            resource_name
        )

    def extraction_error(self):
        """Give an error message for problems extracting file(s)"""

        old_exc = sys.exc_info()[1]
        cache_path = self.extraction_path or get_default_cache()

        tmpl = textwrap.dedent("""
            Can't extract file(s) to egg cache

            The following error occurred while trying to extract file(s)
            to the Python egg cache:

              {old_exc}

            The Python egg cache directory is currently set to:

              {cache_path}

            Perhaps your account does not have write access to this directory?
            You can change the cache directory by setting the PYTHON_EGG_CACHE
            environment variable to point to an accessible directory.
            """).lstrip()
        err = ExtractionError(tmpl.format(**locals()))
        err.manager = self
        err.cache_path = cache_path
        err.original_error = old_exc
        raise err

    def get_cache_path(self, archive_name, names=()):
        """Return absolute location in cache for `archive_name` and `names`

        The parent directory of the resulting path will be created if it does
        not already exist.  `archive_name` should be the base filename of the
        enclosing egg (which may not be the name of the enclosing zipfile!),
        including its ".egg" extension.  `names`, if provided, should be a
        sequence of path name parts "under" the egg's extraction location.

        This method should only be called by resource providers that need to
        obtain an extraction location, and only for names they intend to
        extract, as it tracks the generated names for possible cleanup later.
        """
        extract_path = self.extraction_path or get_default_cache()
        target_path = os.path.join(extract_path, archive_name + '-tmp', *names)
        try:
            _bypass_ensure_directory(target_path)
        except Exception:
            self.extraction_error()

        self._warn_unsafe_extraction_path(extract_path)

        self.cached_files[target_path] = 1
        return target_path

    @staticmethod
    def _warn_unsafe_extraction_path(path):
        """
        If the default extraction path is overridden and set to an insecure
        location, such as /tmp, it opens up an opportunity for an attacker to
        replace an extracted file with an unauthorized payload. Warn the user
        if a known insecure location is used.

        See Distribute #375 for more details.
        """
        if os.name == 'nt' and not path.startswith(os.environ['windir']):
            # On Windows, permissions are generally restrictive by default
            #  and temp directories are not writable by other users, so
            #  bypass the warning.
            return
        mode = os.stat(path).st_mode
        if mode & stat.S_IWOTH or mode & stat.S_IWGRP:
            msg = (
                "%s is writable by group/others and vulnerable to attack "
                "when "
                "used with get_resource_filename. Consider a more secure "
                "location (set with .set_extraction_path or the "
                "PYTHON_EGG_CACHE environment variable)." % path
            )
            warnings.warn(msg, UserWarning)

    def postprocess(self, tempname, filename):
        """Perform any platform-specific postprocessing of `tempname`

        This is where Mac header rewrites should be done; other platforms don't
        have anything special they should do.

        Resource providers should call this method ONLY after successfully
        extracting a compressed resource.  They must NOT call it on resources
        that are already in the filesystem.

        `tempname` is the current (temporary) name of the file, and `filename`
        is the name it will be renamed to by the caller after this routine
        returns.
        """

        if os.name == 'posix':
            # Make the resource executable
            mode = ((os.stat(tempname).st_mode) | 0o555) & 0o7777
            os.chmod(tempname, mode)

    def set_extraction_path(self, path):
        """Set the base path where resources will be extracted to, if needed.

        If you do not call this routine before any extractions take place, the
        path defaults to the return value of ``get_default_cache()``.  (Which
        is based on the ``PYTHON_EGG_CACHE`` environment variable, with various
        platform-specific fallbacks.  See that routine's documentation for more
        details.)

        Resources are extracted to subdirectories of this path based upon
        information given by the ``IResourceProvider``.  You may set this to a
        temporary directory, but then you must call ``cleanup_resources()`` to
        delete the extracted files when done.  There is no guarantee that
        ``cleanup_resources()`` will be able to remove all extracted files.

        (Note: you may not change the extraction path for a given resource
        manager once resources have been extracted, unless you first call
        ``cleanup_resources()``.)
        """
        if self.cached_files:
            raise ValueError(
                "Can't change extraction path, files already extracted"
            )

        self.extraction_path = path

    def cleanup_resources(self, force=False):
        """
        Delete all extracted resource files and directories, returning a list
        of the file and directory names that could not be successfully removed.
        This function does not have any concurrency protection, so it should
        generally only be called when the extraction path is a temporary
        directory exclusive to a single process.  This method is not
        automatically called; you must call it explicitly or register it as an
        ``atexit`` function if you wish to ensure cleanup of a temporary
        directory used for extractions.
        """
        # XXX


def get_default_cache():
    """
    Return the ``PYTHON_EGG_CACHE`` environment variable
    or a platform-relevant user cache dir for an app
    named "Python-Eggs".
    """
    return (
        os.environ.get('PYTHON_EGG_CACHE')
        or platformdirs.user_cache_dir(appname='Python-Eggs')
    )


def safe_name(name):
    """Convert an arbitrary string to a standard distribution name

    Any runs of non-alphanumeric/. characters are replaced with a single '-'.
    """
    return re.sub('[^A-Za-z0-9.]+', '-', name)


def safe_version(version):
    """
    Convert an arbitrary string to a standard version string
    """
    try:
        # normalize the version
        return str(packaging.version.Version(version))
    except packaging.version.InvalidVersion:
        version = version.replace(' ', '.')
        return re.sub('[^A-Za-z0-9.]+', '-', version)


def safe_extra(extra):
    """Convert an arbitrary string to a standard 'extra' name

    Any runs of non-alphanumeric characters are replaced with a single '_',
    and the result is always lowercased.
    """
    return re.sub('[^A-Za-z0-9.-]+', '_', extra).lower()


def to_filename(name):
    """Convert a project or version name to its filename-escaped form

    Any '-' characters are currently replaced with '_'.
    """
    return name.replace('-', '_')


def invalid_marker(text):
    """
    Validate text as a PEP 508 environment marker; return an exception
    if invalid or False otherwise.
    """
    try:
        evaluate_marker(text)
    except SyntaxError as e:
        e.filename = None
        e.lineno = None
        return e
    return False


def evaluate_marker(text, extra=None):
    """
    Evaluate a PEP 508 environment marker.
    Return a boolean indicating the marker result in this environment.
    Raise SyntaxError if marker is invalid.

    This implementation uses the 'pyparsing' module.
    """
    try:
        marker = packaging.markers.Marker(text)
        return marker.evaluate()
    except packaging.markers.InvalidMarker as e:
        raise SyntaxError(e)


class NullProvider:
    """Try to implement resources and metadata for arbitrary PEP 302 loaders"""

    egg_name = None
    egg_info = None
    loader = None

    def __init__(self, module):
        self.loader = getattr(module, '__loader__', None)
        self.module_path = os.path.dirname(getattr(module, '__file__', ''))

    def get_resource_filename(self, manager, resource_name):
        return self._fn(self.module_path, resource_name)

    def get_resource_stream(self, manager, resource_name):
        return io.BytesIO(self.get_resource_string(manager, resource_name))

    def get_resource_string(self, manager, resource_name):
        return self._get(self._fn(self.module_path, resource_name))

    def has_resource(self, resource_name):
        return self._has(self._fn(self.module_path, resource_name))

    def _get_metadata_path(self, name):
        return self._fn(self.egg_info, name)

    def has_metadata(self, name):
        if not self.egg_info:
            return self.egg_info

        path = self._get_metadata_path(name)
        return self._has(path)

    def get_metadata(self, name):
        if not self.egg_info:
            return ""
        path = self._get_metadata_path(name)
        value = self._get(path)
        if six.PY2:
            return value
        try:
            return value.decode('utf-8')
        except UnicodeDecodeError as exc:
            # Include the path in the error message to simplify
            # troubleshooting, and without changing the exception type.
            exc.reason += ' in {} file at path: {}'.format(name, path)
            raise

    def get_metadata_lines(self, name):
        return yield_lines(self.get_metadata(name))

    def resource_isdir(self, resource_name):
        return self._isdir(self._fn(self.module_path, resource_name))

    def metadata_isdir(self, name):
        return self.egg_info and self._isdir(self._fn(self.egg_info, name))

    def resource_listdir(self, resource_name):
        return self._listdir(self._fn(self.module_path, resource_name))

    def metadata_listdir(self, name):
        if self.egg_info:
            return self._listdir(self._fn(self.egg_info, name))
        return []

    def run_script(self, script_name, namespace):
        script = 'scripts/' + script_name
        if not self.has_metadata(script):
            raise ResolutionError(
                "Script {script!r} not found in metadata at {self.egg_info!r}"
                .format(**locals()),
            )
        script_text = self.get_metadata(script).replace('\r\n', '\n')
        script_text = script_text.replace('\r', '\n')
        script_filename = self._fn(self.egg_info, script)
        namespace['__file__'] = script_filename
        if os.path.exists(script_filename):
            source = open(script_filename).read()
            code = compile(source, script_filename, 'exec')
            exec(code, namespace, namespace)
        else:
            from linecache import cache
            cache[script_filename] = (
                len(script_text), 0, script_text.split('\n'), script_filename
            )
            script_code = compile(script_text, script_filename, 'exec')
            exec(script_code, namespace, namespace)

    def _has(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _isdir(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _listdir(self, path):
        raise NotImplementedError(
            "Can't perform this operation for unregistered loader type"
        )

    def _fn(self, base, resource_name):
        self._validate_resource_path(resource_name)
        if resource_name:
            return os.path.join(base, *resource_name.split('/'))
        return base

    @staticmethod
    def _validate_resource_path(path):
        """
        Validate the resource paths according to the docs.
        https://setuptools.readthedocs.io/en/latest/pkg_resources.html#basic-resource-access

        >>> warned = getfixture('recwarn')
        >>> warnings.simplefilter('always')
        >>> vrp = NullProvider._validate_resource_path
        >>> vrp('foo/bar.txt')
        >>> bool(warned)
        False
        >>> vrp('../foo/bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('/foo/bar.txt')
        >>> bool(warned)
        True
        >>> vrp('foo/../../bar.txt')
        >>> bool(warned)
        True
        >>> warned.clear()
        >>> vrp('foo/f../bar.txt')
        >>> bool(warned)
        False

        Windows path separators are straight-up disallowed.
        >>> vrp(r'\\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path \
is not allowed.

        >>> vrp(r'C:\\foo/bar.txt')
        Traceback (most recent call last):
        ...
        ValueError: Use of .. or absolute path in a resource path \
is not allowed.

        Blank values are allowed

        >>> vrp('')
        >>> bool(warned)
        False

        Non-string values are not.

        >>> vrp(None)
        Traceback (most recent call last):
        ...
        AttributeError: ...
        """
        invalid = (
            os.path.pardir in path.split(posixpath.sep) or
            posixpath.isabs(path) or
            ntpath.isabs(path)
        )
        if not invalid:
            return

        msg = "Use of .. or absolute path in a resource path is not allowed."

        # Aggressively disallow Windows absolute paths
        if ntpath.isabs(path) and not posixpath.isabs(path):
            raise ValueError(msg)

        # for compatibility, warn; in future
        # raise ValueError(msg)
        warnings.warn(
            msg[:-1] + " and will raise exceptions in a future release.",
            DeprecationWarning,
            stacklevel=4,
        )

    def _get(self, path):
        if hasattr(self.loader, 'get_data'):
            return self.loader.get_data(path)
        raise NotImplementedError(
            "Can't perform this operation for loaders without 'get_data()'"
        )


register_loader_type(object, NullProvider)


class EggProvider(NullProvider):
    """Provider based on a virtual filesystem"""

    def __init__(self, module):
        NullProvider.__init__(self, module)
        self._setup_prefix()

    def _setup_prefix(self):
        # we assume here that our metadata may be nested inside a "basket"
        # of multiple eggs; that's why we use module_path instead of .archive
        path = self.module_path
        old = None
        while path != old:
            if _is_egg_path(path):
                self.egg_name = os.path.basename(path)
                self.egg_info = os.path.join(path, 'EGG-INFO')
                self.egg_root = path
                break
            old = path
            path, base = os.path.split(path)


class DefaultProvider(EggProvider):
    """Provides access to package resources in the filesystem"""

    def _has(self, path):
        return os.path.exists(path)

    def _isdir(self, path):
        return os.path.isdir(path)

    def _listdir(self, path):
        return os.listdir(path)

    def get_resource_stream(self, manager, resource_name):
        return open(self._fn(self.module_path, resource_name), 'rb')

    def _get(self, path):
        with open(path, 'rb') as stream:
            return stream.read()

    @classmethod
    def _register(cls):
        loader_names = 'SourceFileLoader', 'SourcelessFileLoader',
        for name in loader_names:
            loader_cls = getattr(importlib_machinery, name, type(None))
            register_loader_type(loader_cls, cls)


DefaultProvider._register()


class EmptyProvider(NullProvider):
    """Provider that returns nothing for all requests"""

    module_path = None

    _isdir = _has = lambda self, path: False

    def _get(self, path):
        return ''

    def _listdir(self, path):
        return []

    def __init__(self):
        pass


empty_provider = EmptyProvider()


class ZipManifests(dict):
    """
    zip manifest builder
    """

    @classmethod
    def build(cls, path):
        """
        Build a dictionary similar to the zipimport directory
        caches, except instead of tuples, store ZipInfo objects.

        Use a platform-specific path separator (os.sep) for the path keys
        for compatibility with pypy on Windows.
        """
        with zipfile.ZipFile(path) as zfile:
            items = (
                (
                    name.replace('/', os.sep),
                    zfile.getinfo(name),
                )
                for name in zfile.namelist()
            )
            return dict(items)

    load = build


class MemoizedZipManifests(ZipManifests):
    """
    Memoized zipfile manifests.
    """
    manifest_mod = collections.namedtuple('manifest_mod', 'manifest mtime')

    def load(self, path):
        """
        Load a manifest at path or return a suitable manifest already loaded.
        """
        path = os.path.normpath(path)
        mtime = os.stat(path).st_mtime

        if path not in self or self[path].mtime != mtime:
            manifest = self.build(path)
            self[path] = self.manifest_mod(manifest, mtime)

        return self[path].manifest


class ZipProvider(EggProvider):
    """Resource support for zips and eggs"""

    eagers = None
    _zip_manifests = MemoizedZipManifests()

    def __init__(self, module):
        EggProvider.__init__(self, module)
        self.zip_pre = self.loader.archive + os.sep

    def _zipinfo_name(self, fspath):
        # Convert a virtual filename (full path to file) into a zipfile subpath
        # usable with the zipimport directory cache for our target archive
        fspath = fspath.rstrip(os.sep)
        if fspath == self.loader.archive:
            return ''
        if fspath.startswith(self.zip_pre):
            return fspath[len(self.zip_pre):]
        raise AssertionError(
            "%s is not a subpath of %s" % (fspath, self.zip_pre)
        )

    def _parts(self, zip_path):
        # Convert a zipfile subpath into an egg-relative path part list.
        # pseudo-fs path
        fspath = self.zip_pre + zip_path
        if fspath.startswith(self.egg_root + os.sep):
            return fspath[len(self.egg_root) + 1:].split(os.sep)
        raise AssertionError(
            "%s is not a subpath of %s" % (fspath, self.egg_root)
        )

    @property
    def zipinfo(self):
        return self._zip_manifests.load(self.loader.archive)

    def get_resource_filename(self, manager, resource_name):
        if not self.egg_name:
            raise NotImplementedError(
                "resource_filename() only supported for .egg, not .zip"
            )
        # no need to lock for extraction, since we use temp names
        zip_path = self._resource_to_zip(resource_name)
        eagers = self._get_eager_resources()
        if '/'.join(self._parts(zip_path)) in eagers:
            for name in eagers:
                self._extract_resource(manager, self._eager_to_zip(name))
        return self._extract_resource(manager, zip_path)

    @staticmethod
    def _get_date_and_size(zip_stat):
        size = zip_stat.file_size
        # ymdhms+wday, yday, dst
        date_time = zip_stat.date_time + (0, 0, -1)
        # 1980 offset already done
        timestamp = time.mktime(date_time)
        return timestamp, size

    def _extract_resource(self, manager, zip_path):

        if zip_path in self._index():
            for name in self._index()[zip_path]:
                last = self._extract_resource(
                    manager, os.path.join(zip_path, name)
                )
            # return the extracted directory name
            return os.path.dirname(last)

        timestamp, size = self._get_date_and_size(self.zipinfo[zip_path])

        if not WRITE_SUPPORT:
            raise IOError('"os.rename" and "os.unlink" are not supported '
                          'on this platform')
        try:

            real_path = manager.get_cache_path(
                self.egg_name, self._parts(zip_path)
            )

            if self._is_current(real_path, zip_path):
                return real_path

            outf, tmpnam = _mkstemp(
                ".$extract",
                dir=os.path.dirname(real_path),
            )
            os.write(outf, self.loader.get_data(zip_path))
            os.close(outf)
            utime(tmpnam, (timestamp, timestamp))
            manager.postprocess(tmpnam, real_path)

            try:
                rename(tmpnam, real_path)

            except os.error:
                if os.path.isfile(real_path):
                    if self._is_current(real_path, zip_path):
                        # the file became current since it was checked above,
                        #  so proceed.
                        return real_path
                    # Windows, del old file and retry
                    elif os.name == 'nt':
                        unlink(real_path)
                        rename(tmpnam, real_path)
                        return real_path
                raise

        except os.error:
            # report a user-friendly error
            manager.extraction_error()

        return real_path

    def _is_current(self, file_path, zip_path):
        """
        Return True if the file_path is current for this zip_path
        """
        timestamp, size = self._get_date_and_size(self.zipinfo[zip_path])
        if not os.path.isfile(file_path):
            return False
        stat = os.stat(file_path)
        if stat.st_size != size or stat.st_mtime != timestamp:
            return False
        # check that the contents match
        zip_contents = self.loader.get_data(zip_path)
        with open(file_path, 'rb') as f:
            file_contents = f.read()
        return zip_contents == file_contents

    def _get_eager_resources(self):
        if self.eagers is None:
            eagers = []
            for name in ('native_libs.txt', 'eager_resources.txt'):
                if self.has_metadata(name):
                    eagers.extend(self.get_metadata_lines(name))
            self.eagers = eagers
        return self.eagers

    def _index(self):
        try:
            return self._dirindex
        except AttributeError:
            ind = {}
            for path in self.zipinfo:
                parts = path.split(os.sep)
                while parts:
                    parent = os.sep.join(parts[:-1])
                    if parent in ind:
                        ind[parent].append(parts[-1])
                        break
                    else:
                        ind[parent] = [parts.pop()]
            self._dirindex = ind
            return ind

    def _has(self, fspath):
        zip_path = self._zipinfo_name(fspath)
        return zip_path in self.zipinfo or zip_path in self._index()

    def _isdir(self, fspath):
        return self._zipinfo_name(fspath) in self._index()

    def _listdir(self, fspath):
        return list(self._index().get(self._zipinfo_name(fspath), ()))

    def _eager_to_zip(self, resource_name):
        return self._zipinfo_name(self._fn(self.egg_root, resource_name))

    def _resource_to_zip(self, resource_name):
        return self._zipinfo_name(self._fn(self.module_path, resource_name))


register_loader_type(zipimport.zipimporter, ZipProvider)


class FileMetadata(EmptyProvider):
    """Metadata handler for standalone PKG-INFO files

    Usage::

        metadata = FileMetadata("/path/to/PKG-INFO")

    This provider rejects all data and metadata requests except for PKG-INFO,
    which is treated as existing, and will be the contents of the file at
    the provided location.
    """

    def __init__(self, path):
        self.path = path

    def _get_metadata_path(self, name):
        return self.path

    def has_metadata(self, name):
        return name == 'PKG-INFO' and os.path.isfile(self.path)

    def get_metadata(self, name):
        if name != 'PKG-INFO':
            raise KeyError("No metadata except PKG-INFO is available")

        with io.open(self.path, encoding='utf-8', errors="replace") as f:
            metadata = f.read()
        self._warn_on_replacement(metadata)
        return metadata

    def _warn_on_replacement(self, metadata):
        # Python 2.7 compat for: replacement_char = '�'
        replacement_char = b'\xef\xbf\xbd'.decode('utf-8')
        if replacement_char in metadata:
            tmpl = "{self.path} could not be properly decoded in UTF-8"
            msg = tmpl.format(**locals())
            warnings.warn(msg)

    def get_metadata_lines(self, name):
        return yield_lines(self.get_metadata(name))


class PathMetadata(DefaultProvider):
    """Metadata provider for egg directories

    Usage::

        # Development eggs:

        egg_info = "/path/to/PackageName.egg-info"
        base_dir = os.path.dirname(egg_info)
        metadata = PathMetadata(base_dir, egg_info)
        dist_name = os.path.splitext(os.path.basename(egg_info))[0]
        dist = Distribution(basedir, project_name=dist_name, metadata=metadata)

        # Unpacked egg directories:

        egg_path = "/path/to/PackageName-ver-pyver-etc.egg"
        metadata = PathMetadata(egg_path, os.path.join(egg_path,'EGG-INFO'))
        dist = Distribution.from_filename(egg_path, metadata=metadata)
    """

    def __init__(self, path, egg_info):
        self.module_path = path
        self.egg_info = egg_info


class EggMetadata(ZipProvider):
    """Metadata provider for .egg files"""

    def __init__(self, importer):
        """Create a metadata provider from a zipimporter"""

        self.zip_pre = importer.archive + os.sep
        self.loader = importer
        if importer.prefix:
            self.module_path = os.path.join(importer.archive, importer.prefix)
        else:
            self.module_path = importer.archive
        self._setup_prefix()


_declare_state('dict', _distribution_finders={})


def register_finder(importer_type, distribution_finder):
    """Register `distribution_finder` to find distributions in sys.path items

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `distribution_finder` is a callable that, passed a path
    item and the importer instance, yields ``Distribution`` instances found on
    that path item.  See ``pkg_resources.find_on_path`` for an example."""
    _distribution_finders[importer_type] = distribution_finder


def find_distributions(path_item, only=False):
    """Yield distributions accessible via `path_item`"""
    importer = get_importer(path_item)
    finder = _find_adapter(_distribution_finders, importer)
    return finder(importer, path_item, only)


def find_eggs_in_zip(importer, path_item, only=False):
    """
    Find eggs in zip files; possibly multiple nested eggs.
    """
    if importer.archive.endswith('.whl'):
        # wheels are not supported with this finder
        # they don't have PKG-INFO metadata, and won't ever contain eggs
        return
    metadata = EggMetadata(importer)
    if metadata.has_metadata('PKG-INFO'):
        yield Distribution.from_filename(path_item, metadata=metadata)
    if only:
        # don't yield nested distros
        return
    for subitem in metadata.resource_listdir(''):
        if _is_egg_path(subitem):
            subpath = os.path.join(path_item, subitem)
            dists = find_eggs_in_zip(zipimport.zipimporter(subpath), subpath)
            for dist in dists:
                yield dist
        elif subitem.lower().endswith('.dist-info'):
            subpath = os.path.join(path_item, subitem)
            submeta = EggMetadata(zipimport.zipimporter(subpath))
            submeta.egg_info = subpath
            yield Distribution.from_location(path_item, subitem, submeta)


register_finder(zipimport.zipimporter, find_eggs_in_zip)


def find_nothing(importer, path_item, only=False):
    return ()


register_finder(object, find_nothing)


def _by_version_descending(names):
    """
    Given a list of filenames, return them in descending order
    by version number.

    >>> names = 'bar', 'foo', 'Python-2.7.10.egg', 'Python-2.7.2.egg'
    >>> _by_version_descending(names)
    ['Python-2.7.10.egg', 'Python-2.7.2.egg', 'foo', 'bar']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.egg', 'Setuptools-1.2.3b1.egg']
    >>> names = 'Setuptools-1.2.3b1.egg', 'Setuptools-1.2.3.post1.egg'
    >>> _by_version_descending(names)
    ['Setuptools-1.2.3.post1.egg', 'Setuptools-1.2.3b1.egg']
    """
    def _by_version(name):
        """
        Parse each component of the filename
        """
        name, ext = os.path.splitext(name)
        parts = itertools.chain(name.split('-'), [ext])
        return [packaging.version.parse(part) for part in parts]

    return sorted(names, key=_by_version, reverse=True)


def find_on_path(importer, path_item, only=False):
    """Yield distributions accessible on a sys.path directory"""
    path_item = _normalize_cached(path_item)

    if _is_unpacked_egg(path_item):
        yield Distribution.from_filename(
            path_item, metadata=PathMetadata(
                path_item, os.path.join(path_item, 'EGG-INFO')
            )
        )
        return

    entries = safe_listdir(path_item)

    # for performance, before sorting by version,
    # screen entries for only those that will yield
    # distributions
    filtered = (
        entry
        for entry in entries
        if dist_factory(path_item, entry, only)
    )

    # scan for .egg and .egg-info in directory
    path_item_entries = _by_version_descending(filtered)
    for entry in path_item_entries:
        fullpath = os.path.join(path_item, entry)
        factory = dist_factory(path_item, entry, only)
        for dist in factory(fullpath):
            yield dist


def dist_factory(path_item, entry, only):
    """
    Return a dist_factory for a path_item and entry
    """
    lower = entry.lower()
    is_meta = any(map(lower.endswith, ('.egg-info', '.dist-info')))
    return (
        distributions_from_metadata
        if is_meta else
        find_distributions
        if not only and _is_egg_path(entry) else
        resolve_egg_link
        if not only and lower.endswith('.egg-link') else
        NoDists()
    )


class NoDists:
    """
    >>> bool(NoDists())
    False

    >>> list(NoDists()('anything'))
    []
    """
    def __bool__(self):
        return False
    if six.PY2:
        __nonzero__ = __bool__

    def __call__(self, fullpath):
        return iter(())


def safe_listdir(path):
    """
    Attempt to list contents of path, but suppress some exceptions.
    """
    try:
        return os.listdir(path)
    except (PermissionError, NotADirectoryError):
        pass
    except OSError as e:
        # Ignore the directory if does not exist, not a directory or
        # permission denied
        ignorable = (
            e.errno in (errno.ENOTDIR, errno.EACCES, errno.ENOENT)
            # Python 2 on Windows needs to be handled this way :(
            or getattr(e, "winerror", None) == 267
        )
        if not ignorable:
            raise
    return ()


def distributions_from_metadata(path):
    root = os.path.dirname(path)
    if os.path.isdir(path):
        if len(os.listdir(path)) == 0:
            # empty metadata dir; skip
            return
        metadata = PathMetadata(root, path)
    else:
        metadata = FileMetadata(path)
    entry = os.path.basename(path)
    yield Distribution.from_location(
        root, entry, metadata, precedence=DEVELOP_DIST,
    )


def non_empty_lines(path):
    """
    Yield non-empty lines from file at path
    """
    with open(path) as f:
        for line in f:
            line = line.strip()
            if line:
                yield line


def resolve_egg_link(path):
    """
    Given a path to an .egg-link, resolve distributions
    present in the referenced path.
    """
    referenced_paths = non_empty_lines(path)
    resolved_paths = (
        os.path.join(os.path.dirname(path), ref)
        for ref in referenced_paths
    )
    dist_groups = map(find_distributions, resolved_paths)
    return next(dist_groups, ())


register_finder(pkgutil.ImpImporter, find_on_path)

if hasattr(importlib_machinery, 'FileFinder'):
    register_finder(importlib_machinery.FileFinder, find_on_path)

_declare_state('dict', _namespace_handlers={})
_declare_state('dict', _namespace_packages={})


def register_namespace_handler(importer_type, namespace_handler):
    """Register `namespace_handler` to declare namespace packages

    `importer_type` is the type or class of a PEP 302 "Importer" (sys.path item
    handler), and `namespace_handler` is a callable like this::

        def namespace_handler(importer, path_entry, moduleName, module):
            # return a path_entry to use for child packages

    Namespace handlers are only called if the importer object has already
    agreed that it can handle the relevant path item, and they should only
    return a subpath if the module __path__ does not already contain an
    equivalent subpath.  For an example namespace handler, see
    ``pkg_resources.file_ns_handler``.
    """
    _namespace_handlers[importer_type] = namespace_handler


def _handle_ns(packageName, path_item):
    """Ensure that named package includes a subpath of path_item (if needed)"""

    importer = get_importer(path_item)
    if importer is None:
        return None

    # capture warnings due to #1111
    with warnings.catch_warnings():
        warnings.simplefilter("ignore")
        loader = importer.find_module(packageName)

    if loader is None:
        return None
    module = sys.modules.get(packageName)
    if module is None:
        module = sys.modules[packageName] = types.ModuleType(packageName)
        module.__path__ = []
        _set_parent_ns(packageName)
    elif not hasattr(module, '__path__'):
        raise TypeError("Not a package:", packageName)
    handler = _find_adapter(_namespace_handlers, importer)
    subpath = handler(importer, path_item, packageName, module)
    if subpath is not None:
        path = module.__path__
        path.append(subpath)
        loader.load_module(packageName)
        _rebuild_mod_path(path, packageName, module)
    return subpath


def _rebuild_mod_path(orig_path, package_name, module):
    """
    Rebuild module.__path__ ensuring that all entries are ordered
    corresponding to their sys.path order
    """
    sys_path = [_normalize_cached(p) for p in sys.path]

    def safe_sys_path_index(entry):
        """
        Workaround for #520 and #513.
        """
        try:
            return sys_path.index(entry)
        except ValueError:
            return float('inf')

    def position_in_sys_path(path):
        """
        Return the ordinal of the path based on its position in sys.path
        """
        path_parts = path.split(os.sep)
        module_parts = package_name.count('.') + 1
        parts = path_parts[:-module_parts]
        return safe_sys_path_index(_normalize_cached(os.sep.join(parts)))

    new_path = sorted(orig_path, key=position_in_sys_path)
    new_path = [_normalize_cached(p) for p in new_path]

    if isinstance(module.__path__, list):
        module.__path__[:] = new_path
    else:
        module.__path__ = new_path


def declare_namespace(packageName):
    """Declare that package 'packageName' is a namespace package"""

    _imp.acquire_lock()
    try:
        if packageName in _namespace_packages:
            return

        path = sys.path
        parent, _, _ = packageName.rpartition('.')

        if parent:
            declare_namespace(parent)
            if parent not in _namespace_packages:
                __import__(parent)
            try:
                path = sys.modules[parent].__path__
            except AttributeError:
                raise TypeError("Not a package:", parent)

        # Track what packages are namespaces, so when new path items are added,
        # they can be updated
        _namespace_packages.setdefault(parent or None, []).append(packageName)
        _namespace_packages.setdefault(packageName, [])

        for path_item in path:
            # Ensure all the parent's path items are reflected in the child,
            # if they apply
            _handle_ns(packageName, path_item)

    finally:
        _imp.release_lock()


def fixup_namespace_packages(path_item, parent=None):
    """Ensure that previously-declared namespace packages include path_item"""
    _imp.acquire_lock()
    try:
        for package in _namespace_packages.get(parent, ()):
            subpath = _handle_ns(package, path_item)
            if subpath:
                fixup_namespace_packages(subpath, package)
    finally:
        _imp.release_lock()


def file_ns_handler(importer, path_item, packageName, module):
    """Compute an ns-package subpath for a filesystem or zipfile importer"""

    subpath = os.path.join(path_item, packageName.split('.')[-1])
    normalized = _normalize_cached(subpath)
    for item in module.__path__:
        if _normalize_cached(item) == normalized:
            break
    else:
        # Only return the path if it's not already there
        return subpath


register_namespace_handler(pkgutil.ImpImporter, file_ns_handler)
register_namespace_handler(zipimport.zipimporter, file_ns_handler)

if hasattr(importlib_machinery, 'FileFinder'):
    register_namespace_handler(importlib_machinery.FileFinder, file_ns_handler)


def null_ns_handler(importer, path_item, packageName, module):
    return None


register_namespace_handler(object, null_ns_handler)


def normalize_path(filename):
    """Normalize a file/dir name for comparison purposes"""
    return os.path.normcase(os.path.realpath(os.path.normpath(_cygwin_patch(filename))))


def _cygwin_patch(filename):  # pragma: nocover
    """
    Contrary to POSIX 2008, on Cygwin, getcwd (3) contains
    symlink components. Using
    os.path.abspath() works around this limitation. A fix in os.getcwd()
    would probably better, in Cygwin even more so, except
    that this seems to be by design...
    """
    return os.path.abspath(filename) if sys.platform == 'cygwin' else filename


def _normalize_cached(filename, _cache={}):
    try:
        return _cache[filename]
    except KeyError:
        _cache[filename] = result = normalize_path(filename)
        return result


def _is_egg_path(path):
    """
    Determine if given path appears to be an egg.
    """
    return path.lower().endswith('.egg')


def _is_unpacked_egg(path):
    """
    Determine if given path appears to be an unpacked egg.
    """
    return (
        _is_egg_path(path) and
        os.path.isfile(os.path.join(path, 'EGG-INFO', 'PKG-INFO'))
    )


def _set_parent_ns(packageName):
    parts = packageName.split('.')
    name = parts.pop()
    if parts:
        parent = '.'.join(parts)
        setattr(sys.modules[parent], name, sys.modules[packageName])


def yield_lines(strs):
    """Yield non-empty/non-comment lines of a string or sequence"""
    if isinstance(strs, six.string_types):
        for s in strs.splitlines():
            s = s.strip()
            # skip blank lines/comments
            if s and not s.startswith('#'):
                yield s
    else:
        for ss in strs:
            for s in yield_lines(ss):
                yield s


MODULE = re.compile(r"\w+(\.\w+)*$").match
EGG_NAME = re.compile(
    r"""
    (?P<name>[^-]+) (
        -(?P<ver>[^-]+) (
            -py(?P<pyver>[^-]+) (
                -(?P<plat>.+)
            )?
        )?
    )?
    """,
    re.VERBOSE | re.IGNORECASE,
).match


class EntryPoint:
    """Object representing an advertised importable object"""

    def __init__(self, name, module_name, attrs=(), extras=(), dist=None):
        if not MODULE(module_name):
            raise ValueError("Invalid module name", module_name)
        self.name = name
        self.module_name = module_name
        self.attrs = tuple(attrs)
        self.extras = tuple(extras)
        self.dist = dist

    def __str__(self):
        s = "%s = %s" % (self.name, self.module_name)
        if self.attrs:
            s += ':' + '.'.join(self.attrs)
        if self.extras:
            s += ' [%s]' % ','.join(self.extras)
        return s

    def __repr__(self):
        return "EntryPoint.parse(%r)" % str(self)

    def load(self, require=True, *args, **kwargs):
        """
        Require packages for this EntryPoint, then resolve it.
        """
        if not require or args or kwargs:
            warnings.warn(
                "Parameters to load are deprecated.  Call .resolve and "
                ".require separately.",
                PkgResourcesDeprecationWarning,
                stacklevel=2,
            )
        if require:
            self.require(*args, **kwargs)
        return self.resolve()

    def resolve(self):
        """
        Resolve the entry point from its module and attrs.
        """
        module = __import__(self.module_name, fromlist=['__name__'], level=0)
        try:
            return functools.reduce(getattr, self.attrs, module)
        except AttributeError as exc:
            raise ImportError(str(exc))

    def require(self, env=None, installer=None):
        if self.extras and not self.dist:
            raise UnknownExtra("Can't require() without a distribution", self)

        # Get the requirements for this entry point with all its extras and
        # then resolve them. We have to pass `extras` along when resolving so
        # that the working set knows what extras we want. Otherwise, for
        # dist-info distributions, the working set will assume that the
        # requirements for that extra are purely optional and skip over them.
        reqs = self.dist.requires(self.extras)
        items = working_set.resolve(reqs, env, installer, extras=self.extras)
        list(map(working_set.add, items))

    pattern = re.compile(
        r'\s*'
        r'(?P<name>.+?)\s*'
        r'=\s*'
        r'(?P<module>[\w.]+)\s*'
        r'(:\s*(?P<attr>[\w.]+))?\s*'
        r'(?P<extras>\[.*\])?\s*$'
    )

    @classmethod
    def parse(cls, src, dist=None):
        """Parse a single entry point from string `src`

        Entry point syntax follows the form::

            name = some.module:some.attr [extra1, extra2]

        The entry name and module name are required, but the ``:attrs`` and
        ``[extras]`` parts are optional
        """
        m = cls.pattern.match(src)
        if not m:
            msg = "EntryPoint must be in 'name=module:attrs [extras]' format"
            raise ValueError(msg, src)
        res = m.groupdict()
        extras = cls._parse_extras(res['extras'])
        attrs = res['attr'].split('.') if res['attr'] else ()
        return cls(res['name'], res['module'], attrs, extras, dist)

    @classmethod
    def _parse_extras(cls, extras_spec):
        if not extras_spec:
            return ()
        req = Requirement.parse('x' + extras_spec)
        if req.specs:
            raise ValueError()
        return req.extras

    @classmethod
    def parse_group(cls, group, lines, dist=None):
        """Parse an entry point group"""
        if not MODULE(group):
            raise ValueError("Invalid group name", group)
        this = {}
        for line in yield_lines(lines):
            ep = cls.parse(line, dist)
            if ep.name in this:
                raise ValueError("Duplicate entry point", group, ep.name)
            this[ep.name] = ep
        return this

    @classmethod
    def parse_map(cls, data, dist=None):
        """Parse a map of entry point groups"""
        if isinstance(data, dict):
            data = data.items()
        else:
            data = split_sections(data)
        maps = {}
        for group, lines in data:
            if group is None:
                if not lines:
                    continue
                raise ValueError("Entry points must be listed in groups")
            group = group.strip()
            if group in maps:
                raise ValueError("Duplicate group name", group)
            maps[group] = cls.parse_group(group, lines, dist)
        return maps


def _remove_md5_fragment(location):
    if not location:
        return ''
    parsed = urllib.parse.urlparse(location)
    if parsed[-1].startswith('md5='):
        return urllib.parse.urlunparse(parsed[:-1] + ('',))
    return location


def _version_from_file(lines):
    """
    Given an iterable of lines from a Metadata file, return
    the value of the Version field, if present, or None otherwise.
    """
    def is_version_line(line):
        return line.lower().startswith('version:')
    version_lines = filter(is_version_line, lines)
    line = next(iter(version_lines), '')
    _, _, value = line.partition(':')
    return safe_version(value.strip()) or None


class Distribution:
    """Wrap an actual or potential sys.path entry w/metadata"""
    PKG_INFO = 'PKG-INFO'

    def __init__(
            self, location=None, metadata=None, project_name=None,
            version=None, py_version=PY_MAJOR, platform=None,
            precedence=EGG_DIST):
        self.project_name = safe_name(project_name or 'Unknown')
        if version is not None:
            self._version = safe_version(version)
        self.py_version = py_version
        self.platform = platform
        self.location = location
        self.precedence = precedence
        self._provider = metadata or empty_provider

    @classmethod
    def from_location(cls, location, basename, metadata=None, **kw):
        project_name, version, py_version, platform = [None] * 4
        basename, ext = os.path.splitext(basename)
        if ext.lower() in _distributionImpl:
            cls = _distributionImpl[ext.lower()]

            match = EGG_NAME(basename)
            if match:
                project_name, version, py_version, platform = match.group(
                    'name', 'ver', 'pyver', 'plat'
                )
        return cls(
            location, metadata, project_name=project_name, version=version,
            py_version=py_version, platform=platform, **kw
        )._reload_version()

    def _reload_version(self):
        return self

    @property
    def hashcmp(self):
        return (
            self.parsed_version,
            self.precedence,
            self.key,
            _remove_md5_fragment(self.location),
            self.py_version or '',
            self.platform or '',
        )

    def __hash__(self):
        return hash(self.hashcmp)

    def __lt__(self, other):
        return self.hashcmp < other.hashcmp

    def __le__(self, other):
        return self.hashcmp <= other.hashcmp

    def __gt__(self, other):
        return self.hashcmp > other.hashcmp

    def __ge__(self, other):
        return self.hashcmp >= other.hashcmp

    def __eq__(self, other):
        if not isinstance(other, self.__class__):
            # It's not a Distribution, so they are not equal
            return False
        return self.hashcmp == other.hashcmp

    def __ne__(self, other):
        return not self == other

    # These properties have to be lazy so that we don't have to load any
    # metadata until/unless it's actually needed.  (i.e., some distributions
    # may not know their name or version without loading PKG-INFO)

    @property
    def key(self):
        try:
            return self._key
        except AttributeError:
            self._key = key = self.project_name.lower()
            return key

    @property
    def parsed_version(self):
        if not hasattr(self, "_parsed_version"):
            self._parsed_version = parse_version(self.version)

        return self._parsed_version

    def _warn_legacy_version(self):
        LV = packaging.version.LegacyVersion
        is_legacy = isinstance(self._parsed_version, LV)
        if not is_legacy:
            return

        # While an empty version is technically a legacy version and
        # is not a valid PEP 440 version, it's also unlikely to
        # actually come from someone and instead it is more likely that
        # it comes from setuptools attempting to parse a filename and
        # including it in the list. So for that we'll gate this warning
        # on if the version is anything at all or not.
        if not self.version:
            return

        tmpl = textwrap.dedent("""
            '{project_name} ({version})' is being parsed as a legacy,
            non PEP 440,
            version. You may find odd behavior and sort order.
            In particular it will be sorted as less than 0.0. It
            is recommended to migrate to PEP 440 compatible
            versions.
            """).strip().replace('\n', ' ')

        warnings.warn(tmpl.format(**vars(self)), PEP440Warning)

    @property
    def version(self):
        try:
            return self._version
        except AttributeError:
            version = self._get_version()
            if version is None:
                path = self._get_metadata_path_for_display(self.PKG_INFO)
                msg = (
                    "Missing 'Version:' header and/or {} file at path: {}"
                ).format(self.PKG_INFO, path)
                raise ValueError(msg, self)

            return version

    @property
    def _dep_map(self):
        """
        A map of extra to its list of (direct) requirements
        for this distribution, including the null extra.
        """
        try:
            return self.__dep_map
        except AttributeError:
            self.__dep_map = self._filter_extras(self._build_dep_map())
        return self.__dep_map

    @staticmethod
    def _filter_extras(dm):
        """
        Given a mapping of extras to dependencies, strip off
        environment markers and filter out any dependencies
        not matching the markers.
        """
        for extra in list(filter(None, dm)):
            new_extra = extra
            reqs = dm.pop(extra)
            new_extra, _, marker = extra.partition(':')
            fails_marker = marker and (
                invalid_marker(marker)
                or not evaluate_marker(marker)
            )
            if fails_marker:
                reqs = []
            new_extra = safe_extra(new_extra) or None

            dm.setdefault(new_extra, []).extend(reqs)
        return dm

    def _build_dep_map(self):
        dm = {}
        for name in 'requires.txt', 'depends.txt':
            for extra, reqs in split_sections(self._get_metadata(name)):
                dm.setdefault(extra, []).extend(parse_requirements(reqs))
        return dm

    def requires(self, extras=()):
        """List of Requirements needed for this distro if `extras` are used"""
        dm = self._dep_map
        deps = []
        deps.extend(dm.get(None, ()))
        for ext in extras:
            try:
                deps.extend(dm[safe_extra(ext)])
            except KeyError:
                raise UnknownExtra(
                    "%s has no such extra feature %r" % (self, ext)
                )
        return deps

    def _get_metadata_path_for_display(self, name):
        """
        Return the path to the given metadata file, if available.
        """
        try:
            # We need to access _get_metadata_path() on the provider object
            # directly rather than through this class's __getattr__()
            # since _get_metadata_path() is marked private.
            path = self._provider._get_metadata_path(name)

        # Handle exceptions e.g. in case the distribution's metadata
        # provider doesn't support _get_metadata_path().
        except Exception:
            return '[could not detect]'

        return path

    def _get_metadata(self, name):
        if self.has_metadata(name):
            for line in self.get_metadata_lines(name):
                yield line

    def _get_version(self):
        lines = self._get_metadata(self.PKG_INFO)
        version = _version_from_file(lines)

        return version

    def activate(self, path=None, replace=False):
        """Ensure distribution is importable on `path` (default=sys.path)"""
        if path is None:
            path = sys.path
        self.insert_on(path, replace=replace)
        if path is sys.path:
            fixup_namespace_packages(self.location)
            for pkg in self._get_metadata('namespace_packages.txt'):
                if pkg in sys.modules:
                    declare_namespace(pkg)

    def egg_name(self):
        """Return what this distribution's standard .egg filename should be"""
        filename = "%s-%s-py%s" % (
            to_filename(self.project_name), to_filename(self.version),
            self.py_version or PY_MAJOR
        )

        if self.platform:
            filename += '-' + self.platform
        return filename

    def __repr__(self):
        if self.location:
            return "%s (%s)" % (self, self.location)
        else:
            return str(self)

    def __str__(self):
        try:
            version = getattr(self, 'version', None)
        except ValueError:
            version = None
        version = version or "[unknown version]"
        return "%s %s" % (self.project_name, version)

    def __getattr__(self, attr):
        """Delegate all unrecognized public attributes to .metadata provider"""
        if attr.startswith('_'):
            raise AttributeError(attr)
        return getattr(self._provider, attr)

    def __dir__(self):
        return list(
            set(super(Distribution, self).__dir__())
            | set(
                attr for attr in self._provider.__dir__()
                if not attr.startswith('_')
            )
        )

    if not hasattr(object, '__dir__'):
        # python 2.7 not supported
        del __dir__

    @classmethod
    def from_filename(cls, filename, metadata=None, **kw):
        return cls.from_location(
            _normalize_cached(filename), os.path.basename(filename), metadata,
            **kw
        )

    def as_requirement(self):
        """Return a ``Requirement`` that matches this distribution exactly"""
        if isinstance(self.parsed_version, packaging.version.Version):
            spec = "%s==%s" % (self.project_name, self.parsed_version)
        else:
            spec = "%s===%s" % (self.project_name, self.parsed_version)

        return Requirement.parse(spec)

    def load_entry_point(self, group, name):
        """Return the `name` entry point of `group` or raise ImportError"""
        ep = self.get_entry_info(group, name)
        if ep is None:
            raise ImportError("Entry point %r not found" % ((group, name),))
        return ep.load()

    def get_entry_map(self, group=None):
        """Return the entry point map for `group`, or the full entry map"""
        try:
            ep_map = self._ep_map
        except AttributeError:
            ep_map = self._ep_map = EntryPoint.parse_map(
                self._get_metadata('entry_points.txt'), self
            )
        if group is not None:
            return ep_map.get(group, {})
        return ep_map

    def get_entry_info(self, group, name):
        """Return the EntryPoint object for `group`+`name`, or ``None``"""
        return self.get_entry_map(group).get(name)

    def insert_on(self, path, loc=None, replace=False):
        """Ensure self.location is on path

        If replace=False (default):
            - If location is already in path anywhere, do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent.
              - Else: add to the end of path.
        If replace=True:
            - If location is already on path anywhere (not eggs)
              or higher priority than its parent (eggs)
              do nothing.
            - Else:
              - If it's an egg and its parent directory is on path,
                insert just ahead of the parent,
                removing any lower-priority entries.
              - Else: add it to the front of path.
        """

        loc = loc or self.location
        if not loc:
            return

        nloc = _normalize_cached(loc)
        bdir = os.path.dirname(nloc)
        npath = [(p and _normalize_cached(p) or p) for p in path]

        for p, item in enumerate(npath):
            if item == nloc:
                if replace:
                    break
                else:
                    # don't modify path (even removing duplicates) if
                    # found and not replace
                    return
            elif item == bdir and self.precedence == EGG_DIST:
                # if it's an .egg, give it precedence over its directory
                # UNLESS it's already been added to sys.path and replace=False
                if (not replace) and nloc in npath[p:]:
                    return
                if path is sys.path:
                    self.check_version_conflict()
                path.insert(p, loc)
                npath.insert(p, nloc)
                break
        else:
            if path is sys.path:
                self.check_version_conflict()
            if replace:
                path.insert(0, loc)
            else:
                path.append(loc)
            return

        # p is the spot where we found or inserted loc; now remove duplicates
        while True:
            try:
                np = npath.index(nloc, p + 1)
            except ValueError:
                break
            else:
                del npath[np], path[np]
                # ha!
                p = np

        return

    def check_version_conflict(self):
        if self.key == 'setuptools':
            # ignore the inevitable setuptools self-conflicts  :(
            return

        nsp = dict.fromkeys(self._get_metadata('namespace_packages.txt'))
        loc = normalize_path(self.location)
        for modname in self._get_metadata('top_level.txt'):
            if (modname not in sys.modules or modname in nsp
                    or modname in _namespace_packages):
                continue
            if modname in ('pkg_resources', 'setuptools', 'site'):
                continue
            fn = getattr(sys.modules[modname], '__file__', None)
            if fn and (normalize_path(fn).startswith(loc) or
                       fn.startswith(self.location)):
                continue
            issue_warning(
                "Module %s was already imported from %s, but %s is being added"
                " to sys.path" % (modname, fn, self.location),
            )

    def has_version(self):
        try:
            self.version
        except ValueError:
            issue_warning("Unbuilt egg for " + repr(self))
            return False
        return True

    def clone(self, **kw):
        """Copy this distribution, substituting in any changed keyword args"""
        names = 'project_name version py_version platform location precedence'
        for attr in names.split():
            kw.setdefault(attr, getattr(self, attr, None))
        kw.setdefault('metadata', self._provider)
        return self.__class__(**kw)

    @property
    def extras(self):
        return [dep for dep in self._dep_map if dep]


class EggInfoDistribution(Distribution):
    def _reload_version(self):
        """
        Packages installed by distutils (e.g. numpy or scipy),
        which uses an old safe_version, and so
        their version numbers can get mangled when
        converted to filenames (e.g., 1.11.0.dev0+2329eae to
        1.11.0.dev0_2329eae). These distributions will not be
        parsed properly
        downstream by Distribution and safe_version, so
        take an extra step and try to get the version number from
        the metadata file itself instead of the filename.
        """
        md_version = self._get_version()
        if md_version:
            self._version = md_version
        return self


class DistInfoDistribution(Distribution):
    """
    Wrap an actual or potential sys.path entry
    w/metadata, .dist-info style.
    """
    PKG_INFO = 'METADATA'
    EQEQ = re.compile(r"([\(,])\s*(\d.*?)\s*([,\)])")

    @property
    def _parsed_pkg_info(self):
        """Parse and cache metadata"""
        try:
            return self._pkg_info
        except AttributeError:
            metadata = self.get_metadata(self.PKG_INFO)
            self._pkg_info = email.parser.Parser().parsestr(metadata)
            return self._pkg_info

    @property
    def _dep_map(self):
        try:
            return self.__dep_map
        except AttributeError:
            self.__dep_map = self._compute_dependencies()
            return self.__dep_map

    def _compute_dependencies(self):
        """Recompute this distribution's dependencies."""
        dm = self.__dep_map = {None: []}

        reqs = []
        # Including any condition expressions
        for req in self._parsed_pkg_info.get_all('Requires-Dist') or []:
            reqs.extend(parse_requirements(req))

        def reqs_for_extra(extra):
            for req in reqs:
                if not req.marker or req.marker.evaluate({'extra': extra}):
                    yield req

        common = frozenset(reqs_for_extra(None))
        dm[None].extend(common)

        for extra in self._parsed_pkg_info.get_all('Provides-Extra') or []:
            s_extra = safe_extra(extra.strip())
            dm[s_extra] = list(frozenset(reqs_for_extra(extra)) - common)

        return dm


_distributionImpl = {
    '.egg': Distribution,
    '.egg-info': EggInfoDistribution,
    '.dist-info': DistInfoDistribution,
}


def issue_warning(*args, **kw):
    level = 1
    g = globals()
    try:
        # find the first stack frame that is *not* code in
        # the pkg_resources module, to use for the warning
        while sys._getframe(level).f_globals is g:
            level += 1
    except ValueError:
        pass
    warnings.warn(stacklevel=level + 1, *args, **kw)


class RequirementParseError(ValueError):
    def __str__(self):
        return ' '.join(self.args)


def parse_requirements(strs):
    """Yield ``Requirement`` objects for each specification in `strs`

    `strs` must be a string, or a (possibly-nested) iterable thereof.
    """
    # create a steppable iterator, so we can handle \-continuations
    lines = iter(yield_lines(strs))

    for line in lines:
        # Drop comments -- a hash without a space may be in a URL.
        if ' #' in line:
            line = line[:line.find(' #')]
        # If there is a line continuation, drop it, and append the next line.
        if line.endswith('\\'):
            line = line[:-2].strip()
            try:
                line += next(lines)
            except StopIteration:
                return
        yield Requirement(line)


class Requirement(packaging.requirements.Requirement):
    def __init__(self, requirement_string):
        """DO NOT CALL THIS UNDOCUMENTED METHOD; use Requirement.parse()!"""
        try:
            super(Requirement, self).__init__(requirement_string)
        except packaging.requirements.InvalidRequirement as e:
            raise RequirementParseError(str(e))
        self.unsafe_name = self.name
        project_name = safe_name(self.name)
        self.project_name, self.key = project_name, project_name.lower()
        self.specs = [
            (spec.operator, spec.version) for spec in self.specifier]
        self.extras = tuple(map(safe_extra, self.extras))
        self.hashCmp = (
            self.key,
            self.url,
            self.specifier,
            frozenset(self.extras),
            str(self.marker) if self.marker else None,
        )
        self.__hash = hash(self.hashCmp)

    def __eq__(self, other):
        return (
            isinstance(other, Requirement) and
            self.hashCmp == other.hashCmp
        )

    def __ne__(self, other):
        return not self == other

    def __contains__(self, item):
        if isinstance(item, Distribution):
            if item.key != self.key:
                return False

            item = item.version

        # Allow prereleases always in order to match the previous behavior of
        # this method. In the future this should be smarter and follow PEP 440
        # more accurately.
        return self.specifier.contains(item, prereleases=True)

    def __hash__(self):
        return self.__hash

    def __repr__(self):
        return "Requirement.parse(%r)" % str(self)

    @staticmethod
    def parse(s):
        req, = parse_requirements(s)
        return req


def _always_object(classes):
    """
    Ensure object appears in the mro even
    for old-style classes.
    """
    if object not in classes:
        return classes + (object,)
    return classes


def _find_adapter(registry, ob):
    """Return an adapter factory for `ob` from `registry`"""
    types = _always_object(inspect.getmro(getattr(ob, '__class__', type(ob))))
    for t in types:
        if t in registry:
            return registry[t]


def ensure_directory(path):
    """Ensure that the parent directory of `path` exists"""
    dirname = os.path.dirname(path)
    py31compat.makedirs(dirname, exist_ok=True)


def _bypass_ensure_directory(path):
    """Sandbox-bypassing version of ensure_directory()"""
    if not WRITE_SUPPORT:
        raise IOError('"os.mkdir" not supported on this platform.')
    dirname, filename = split(path)
    if dirname and filename and not isdir(dirname):
        _bypass_ensure_directory(dirname)
        try:
            mkdir(dirname, 0o755)
        except FileExistsError:
            pass


def split_sections(s):
    """Split a string or iterable thereof into (section, content) pairs

    Each ``section`` is a stripped version of the section header ("[section]")
    and each ``content`` is a list of stripped lines excluding blank lines and
    comment-only lines.  If there are any such lines before the first section
    header, they're returned in a first ``section`` of ``None``.
    """
    section = None
    content = []
    for line in yield_lines(s):
        if line.startswith("["):
            if line.endswith("]"):
                if section or content:
                    yield section, content
                section = line[1:-1].strip()
                content = []
            else:
                raise ValueError("Invalid section heading", line)
        else:
            content.append(line)

    # wrap up last segment
    yield section, content


def _mkstemp(*args, **kw):
    old_open = os.open
    try:
        # temporarily bypass sandboxing
        os.open = os_open
        return tempfile.mkstemp(*args, **kw)
    finally:
        # and then put it back
        os.open = old_open


# Silence the PEP440Warning by default, so that end users don't get hit by it
# randomly just because they use pkg_resources. We want to append the rule
# because we want earlier uses of filterwarnings to take precedence over this
# one.
warnings.filterwarnings("ignore", category=PEP440Warning, append=True)


# from jaraco.functools 1.3
def _call_aside(f, *args, **kwargs):
    f(*args, **kwargs)
    return f


@_call_aside
def _initialize(g=globals()):
    "Set up global resource manager (deliberately not state-saved)"
    manager = ResourceManager()
    g['_manager'] = manager
    g.update(
        (name, getattr(manager, name))
        for name in dir(manager)
        if not name.startswith('_')
    )


@_call_aside
def _initialize_master_working_set():
    """
    Prepare the master working set and make the ``require()``
    API available.

    This function has explicit effects on the global state
    of pkg_resources. It is intended to be invoked once at
    the initialization of this module.

    Invocation by other packages is unsupported and done
    at their own risk.
    """
    working_set = WorkingSet._build_master()
    _declare_state('object', working_set=working_set)

    require = working_set.require
    iter_entry_points = working_set.iter_entry_points
    add_activation_listener = working_set.subscribe
    run_script = working_set.run_script
    # backward compatibility
    run_main = run_script
    # Activate all distributions already on sys.path with replace=False and
    # ensure that all distributions added to the working set in the future
    # (e.g. by calling ``require()``) will get activated as well,
    # with higher priority (replace=True).
    tuple(
        dist.activate(replace=False)
        for dist in working_set
    )
    add_activation_listener(
        lambda dist: dist.activate(replace=True),
        existing=False,
    )
    working_set.entries = []
    # match order
    list(map(working_set.add_entry, sys.path))
    globals().update(locals())

class PkgResourcesDeprecationWarning(Warning):
    """
    Base class for warning about deprecations in ``pkg_resources``

    This class is not derived from ``DeprecationWarning``, and as such is
    visible by default.
    """
site-packages/pip/_vendor/urllib3/request.py000064400000013541150564745010015141 0ustar00from __future__ import absolute_import

from .filepost import encode_multipart_formdata
from .packages.six.moves.urllib.parse import urlencode

__all__ = ["RequestMethods"]


class RequestMethods(object):
    """
    Convenience mixin for classes who implement a :meth:`urlopen` method, such
    as :class:`urllib3.HTTPConnectionPool` and
    :class:`urllib3.PoolManager`.

    Provides behavior for making common types of HTTP request methods and
    decides which type of request field encoding to use.

    Specifically,

    :meth:`.request_encode_url` is for sending requests whose fields are
    encoded in the URL (such as GET, HEAD, DELETE).

    :meth:`.request_encode_body` is for sending requests whose fields are
    encoded in the *body* of the request using multipart or www-form-urlencoded
    (such as for POST, PUT, PATCH).

    :meth:`.request` is for making any kind of request, it will look up the
    appropriate encoding format and use one of the above two methods to make
    the request.

    Initializer parameters:

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.
    """

    _encode_url_methods = {"DELETE", "GET", "HEAD", "OPTIONS"}

    def __init__(self, headers=None):
        self.headers = headers or {}

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        encode_multipart=True,
        multipart_boundary=None,
        **kw
    ):  # Abstract
        raise NotImplementedError(
            "Classes extending RequestMethods must implement "
            "their own ``urlopen`` method."
        )

    def request(self, method, url, fields=None, headers=None, **urlopen_kw):
        """
        Make a request using :meth:`urlopen` with the appropriate encoding of
        ``fields`` based on the ``method`` used.

        This is a convenience method that requires the least amount of manual
        effort. It can be used in most situations, while still having the
        option to drop down to more specific methods when necessary, such as
        :meth:`request_encode_url`, :meth:`request_encode_body`,
        or even the lowest level :meth:`urlopen`.
        """
        method = method.upper()

        urlopen_kw["request_url"] = url

        if method in self._encode_url_methods:
            return self.request_encode_url(
                method, url, fields=fields, headers=headers, **urlopen_kw
            )
        else:
            return self.request_encode_body(
                method, url, fields=fields, headers=headers, **urlopen_kw
            )

    def request_encode_url(self, method, url, fields=None, headers=None, **urlopen_kw):
        """
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the url. This is useful for request methods like GET, HEAD, DELETE, etc.
        """
        if headers is None:
            headers = self.headers

        extra_kw = {"headers": headers}
        extra_kw.update(urlopen_kw)

        if fields:
            url += "?" + urlencode(fields)

        return self.urlopen(method, url, **extra_kw)

    def request_encode_body(
        self,
        method,
        url,
        fields=None,
        headers=None,
        encode_multipart=True,
        multipart_boundary=None,
        **urlopen_kw
    ):
        """
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the body. This is useful for request methods like POST, PUT, PATCH, etc.

        When ``encode_multipart=True`` (default), then
        :func:`urllib3.encode_multipart_formdata` is used to encode
        the payload with the appropriate content type. Otherwise
        :func:`urllib.parse.urlencode` is used with the
        'application/x-www-form-urlencoded' content type.

        Multipart encoding must be used when posting files, and it's reasonably
        safe to use it in other times too. However, it may break request
        signing, such as with OAuth.

        Supports an optional ``fields`` parameter of key/value strings AND
        key/filetuple. A filetuple is a (filename, data, MIME type) tuple where
        the MIME type is optional. For example::

            fields = {
                'foo': 'bar',
                'fakefile': ('foofile.txt', 'contents of foofile'),
                'realfile': ('barfile.txt', open('realfile').read()),
                'typedfile': ('bazfile.bin', open('bazfile').read(),
                              'image/jpeg'),
                'nonamefile': 'contents of nonamefile field',
            }

        When uploading a file, providing a filename (the first parameter of the
        tuple) is optional but recommended to best mimic behavior of browsers.

        Note that if ``headers`` are supplied, the 'Content-Type' header will
        be overwritten because it depends on the dynamic random boundary string
        which is used to compose the body of the request. The random boundary
        string can be explicitly set with the ``multipart_boundary`` parameter.
        """
        if headers is None:
            headers = self.headers

        extra_kw = {"headers": {}}

        if fields:
            if "body" in urlopen_kw:
                raise TypeError(
                    "request got values for both 'fields' and 'body', can only specify one."
                )

            if encode_multipart:
                body, content_type = encode_multipart_formdata(
                    fields, boundary=multipart_boundary
                )
            else:
                body, content_type = (
                    urlencode(fields),
                    "application/x-www-form-urlencoded",
                )

            extra_kw["body"] = body
            extra_kw["headers"] = {"Content-Type": content_type}

        extra_kw["headers"].update(headers)
        extra_kw.update(urlopen_kw)

        return self.urlopen(method, url, **extra_kw)
site-packages/pip/_vendor/urllib3/response.py000064400000067053150564745010015316 0ustar00from __future__ import absolute_import

import io
import logging
import zlib
from contextlib import contextmanager
from socket import error as SocketError
from socket import timeout as SocketTimeout

try:
    import brotli
except ImportError:
    brotli = None

from ._collections import HTTPHeaderDict
from .connection import BaseSSLError, HTTPException
from .exceptions import (
    BodyNotHttplibCompatible,
    DecodeError,
    HTTPError,
    IncompleteRead,
    InvalidChunkLength,
    InvalidHeader,
    ProtocolError,
    ReadTimeoutError,
    ResponseNotChunked,
    SSLError,
)
from .packages import six
from .util.response import is_fp_closed, is_response_to_head

log = logging.getLogger(__name__)


class DeflateDecoder(object):
    def __init__(self):
        self._first_try = True
        self._data = b""
        self._obj = zlib.decompressobj()

    def __getattr__(self, name):
        return getattr(self._obj, name)

    def decompress(self, data):
        if not data:
            return data

        if not self._first_try:
            return self._obj.decompress(data)

        self._data += data
        try:
            decompressed = self._obj.decompress(data)
            if decompressed:
                self._first_try = False
                self._data = None
            return decompressed
        except zlib.error:
            self._first_try = False
            self._obj = zlib.decompressobj(-zlib.MAX_WBITS)
            try:
                return self.decompress(self._data)
            finally:
                self._data = None


class GzipDecoderState(object):

    FIRST_MEMBER = 0
    OTHER_MEMBERS = 1
    SWALLOW_DATA = 2


class GzipDecoder(object):
    def __init__(self):
        self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)
        self._state = GzipDecoderState.FIRST_MEMBER

    def __getattr__(self, name):
        return getattr(self._obj, name)

    def decompress(self, data):
        ret = bytearray()
        if self._state == GzipDecoderState.SWALLOW_DATA or not data:
            return bytes(ret)
        while True:
            try:
                ret += self._obj.decompress(data)
            except zlib.error:
                previous_state = self._state
                # Ignore data after the first error
                self._state = GzipDecoderState.SWALLOW_DATA
                if previous_state == GzipDecoderState.OTHER_MEMBERS:
                    # Allow trailing garbage acceptable in other gzip clients
                    return bytes(ret)
                raise
            data = self._obj.unused_data
            if not data:
                return bytes(ret)
            self._state = GzipDecoderState.OTHER_MEMBERS
            self._obj = zlib.decompressobj(16 + zlib.MAX_WBITS)


if brotli is not None:

    class BrotliDecoder(object):
        # Supports both 'brotlipy' and 'Brotli' packages
        # since they share an import name. The top branches
        # are for 'brotlipy' and bottom branches for 'Brotli'
        def __init__(self):
            self._obj = brotli.Decompressor()
            if hasattr(self._obj, "decompress"):
                self.decompress = self._obj.decompress
            else:
                self.decompress = self._obj.process

        def flush(self):
            if hasattr(self._obj, "flush"):
                return self._obj.flush()
            return b""


class MultiDecoder(object):
    """
    From RFC7231:
        If one or more encodings have been applied to a representation, the
        sender that applied the encodings MUST generate a Content-Encoding
        header field that lists the content codings in the order in which
        they were applied.
    """

    def __init__(self, modes):
        self._decoders = [_get_decoder(m.strip()) for m in modes.split(",")]

    def flush(self):
        return self._decoders[0].flush()

    def decompress(self, data):
        for d in reversed(self._decoders):
            data = d.decompress(data)
        return data


def _get_decoder(mode):
    if "," in mode:
        return MultiDecoder(mode)

    if mode == "gzip":
        return GzipDecoder()

    if brotli is not None and mode == "br":
        return BrotliDecoder()

    return DeflateDecoder()


class HTTPResponse(io.IOBase):
    """
    HTTP Response container.

    Backwards-compatible with :class:`http.client.HTTPResponse` but the response ``body`` is
    loaded and decoded on-demand when the ``data`` property is accessed.  This
    class is also compatible with the Python standard library's :mod:`io`
    module, and can hence be treated as a readable object in the context of that
    framework.

    Extra parameters for behaviour not present in :class:`http.client.HTTPResponse`:

    :param preload_content:
        If True, the response's body will be preloaded during construction.

    :param decode_content:
        If True, will attempt to decode the body based on the
        'content-encoding' header.

    :param original_response:
        When this HTTPResponse wrapper is generated from an :class:`http.client.HTTPResponse`
        object, it's convenient to include the original for debug purposes. It's
        otherwise unused.

    :param retries:
        The retries contains the last :class:`~urllib3.util.retry.Retry` that
        was used during the request.

    :param enforce_content_length:
        Enforce content length checking. Body returned by server must match
        value of Content-Length header, if present. Otherwise, raise error.
    """

    CONTENT_DECODERS = ["gzip", "deflate"]
    if brotli is not None:
        CONTENT_DECODERS += ["br"]
    REDIRECT_STATUSES = [301, 302, 303, 307, 308]

    def __init__(
        self,
        body="",
        headers=None,
        status=0,
        version=0,
        reason=None,
        strict=0,
        preload_content=True,
        decode_content=True,
        original_response=None,
        pool=None,
        connection=None,
        msg=None,
        retries=None,
        enforce_content_length=False,
        request_method=None,
        request_url=None,
        auto_close=True,
    ):

        if isinstance(headers, HTTPHeaderDict):
            self.headers = headers
        else:
            self.headers = HTTPHeaderDict(headers)
        self.status = status
        self.version = version
        self.reason = reason
        self.strict = strict
        self.decode_content = decode_content
        self.retries = retries
        self.enforce_content_length = enforce_content_length
        self.auto_close = auto_close

        self._decoder = None
        self._body = None
        self._fp = None
        self._original_response = original_response
        self._fp_bytes_read = 0
        self.msg = msg
        self._request_url = request_url

        if body and isinstance(body, (six.string_types, bytes)):
            self._body = body

        self._pool = pool
        self._connection = connection

        if hasattr(body, "read"):
            self._fp = body

        # Are we using the chunked-style of transfer encoding?
        self.chunked = False
        self.chunk_left = None
        tr_enc = self.headers.get("transfer-encoding", "").lower()
        # Don't incur the penalty of creating a list and then discarding it
        encodings = (enc.strip() for enc in tr_enc.split(","))
        if "chunked" in encodings:
            self.chunked = True

        # Determine length of response
        self.length_remaining = self._init_length(request_method)

        # If requested, preload the body.
        if preload_content and not self._body:
            self._body = self.read(decode_content=decode_content)

    def get_redirect_location(self):
        """
        Should we redirect and where to?

        :returns: Truthy redirect location string if we got a redirect status
            code and valid location. ``None`` if redirect status and no
            location. ``False`` if not a redirect status code.
        """
        if self.status in self.REDIRECT_STATUSES:
            return self.headers.get("location")

        return False

    def release_conn(self):
        if not self._pool or not self._connection:
            return

        self._pool._put_conn(self._connection)
        self._connection = None

    def drain_conn(self):
        """
        Read and discard any remaining HTTP response data in the response connection.

        Unread data in the HTTPResponse connection blocks the connection from being released back to the pool.
        """
        try:
            self.read()
        except (HTTPError, SocketError, BaseSSLError, HTTPException):
            pass

    @property
    def data(self):
        # For backwards-compat with earlier urllib3 0.4 and earlier.
        if self._body:
            return self._body

        if self._fp:
            return self.read(cache_content=True)

    @property
    def connection(self):
        return self._connection

    def isclosed(self):
        return is_fp_closed(self._fp)

    def tell(self):
        """
        Obtain the number of bytes pulled over the wire so far. May differ from
        the amount of content returned by :meth:``urllib3.response.HTTPResponse.read``
        if bytes are encoded on the wire (e.g, compressed).
        """
        return self._fp_bytes_read

    def _init_length(self, request_method):
        """
        Set initial length value for Response content if available.
        """
        length = self.headers.get("content-length")

        if length is not None:
            if self.chunked:
                # This Response will fail with an IncompleteRead if it can't be
                # received as chunked. This method falls back to attempt reading
                # the response before raising an exception.
                log.warning(
                    "Received response with both Content-Length and "
                    "Transfer-Encoding set. This is expressly forbidden "
                    "by RFC 7230 sec 3.3.2. Ignoring Content-Length and "
                    "attempting to process response as Transfer-Encoding: "
                    "chunked."
                )
                return None

            try:
                # RFC 7230 section 3.3.2 specifies multiple content lengths can
                # be sent in a single Content-Length header
                # (e.g. Content-Length: 42, 42). This line ensures the values
                # are all valid ints and that as long as the `set` length is 1,
                # all values are the same. Otherwise, the header is invalid.
                lengths = set([int(val) for val in length.split(",")])
                if len(lengths) > 1:
                    raise InvalidHeader(
                        "Content-Length contained multiple "
                        "unmatching values (%s)" % length
                    )
                length = lengths.pop()
            except ValueError:
                length = None
            else:
                if length < 0:
                    length = None

        # Convert status to int for comparison
        # In some cases, httplib returns a status of "_UNKNOWN"
        try:
            status = int(self.status)
        except ValueError:
            status = 0

        # Check for responses that shouldn't include a body
        if status in (204, 304) or 100 <= status < 200 or request_method == "HEAD":
            length = 0

        return length

    def _init_decoder(self):
        """
        Set-up the _decoder attribute if necessary.
        """
        # Note: content-encoding value should be case-insensitive, per RFC 7230
        # Section 3.2
        content_encoding = self.headers.get("content-encoding", "").lower()
        if self._decoder is None:
            if content_encoding in self.CONTENT_DECODERS:
                self._decoder = _get_decoder(content_encoding)
            elif "," in content_encoding:
                encodings = [
                    e.strip()
                    for e in content_encoding.split(",")
                    if e.strip() in self.CONTENT_DECODERS
                ]
                if len(encodings):
                    self._decoder = _get_decoder(content_encoding)

    DECODER_ERROR_CLASSES = (IOError, zlib.error)
    if brotli is not None:
        DECODER_ERROR_CLASSES += (brotli.error,)

    def _decode(self, data, decode_content, flush_decoder):
        """
        Decode the data passed in and potentially flush the decoder.
        """
        if not decode_content:
            return data

        try:
            if self._decoder:
                data = self._decoder.decompress(data)
        except self.DECODER_ERROR_CLASSES as e:
            content_encoding = self.headers.get("content-encoding", "").lower()
            raise DecodeError(
                "Received response with content-encoding: %s, but "
                "failed to decode it." % content_encoding,
                e,
            )
        if flush_decoder:
            data += self._flush_decoder()

        return data

    def _flush_decoder(self):
        """
        Flushes the decoder. Should only be called if the decoder is actually
        being used.
        """
        if self._decoder:
            buf = self._decoder.decompress(b"")
            return buf + self._decoder.flush()

        return b""

    @contextmanager
    def _error_catcher(self):
        """
        Catch low-level python exceptions, instead re-raising urllib3
        variants, so that low-level exceptions are not leaked in the
        high-level api.

        On exit, release the connection back to the pool.
        """
        clean_exit = False

        try:
            try:
                yield

            except SocketTimeout:
                # FIXME: Ideally we'd like to include the url in the ReadTimeoutError but
                # there is yet no clean way to get at it from this context.
                raise ReadTimeoutError(self._pool, None, "Read timed out.")

            except BaseSSLError as e:
                # FIXME: Is there a better way to differentiate between SSLErrors?
                if "read operation timed out" not in str(e):
                    # SSL errors related to framing/MAC get wrapped and reraised here
                    raise SSLError(e)

                raise ReadTimeoutError(self._pool, None, "Read timed out.")

            except (HTTPException, SocketError) as e:
                # This includes IncompleteRead.
                raise ProtocolError("Connection broken: %r" % e, e)

            # If no exception is thrown, we should avoid cleaning up
            # unnecessarily.
            clean_exit = True
        finally:
            # If we didn't terminate cleanly, we need to throw away our
            # connection.
            if not clean_exit:
                # The response may not be closed but we're not going to use it
                # anymore so close it now to ensure that the connection is
                # released back to the pool.
                if self._original_response:
                    self._original_response.close()

                # Closing the response may not actually be sufficient to close
                # everything, so if we have a hold of the connection close that
                # too.
                if self._connection:
                    self._connection.close()

            # If we hold the original response but it's closed now, we should
            # return the connection back to the pool.
            if self._original_response and self._original_response.isclosed():
                self.release_conn()

    def read(self, amt=None, decode_content=None, cache_content=False):
        """
        Similar to :meth:`http.client.HTTPResponse.read`, but with two additional
        parameters: ``decode_content`` and ``cache_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.

        :param cache_content:
            If True, will save the returned data such that the same result is
            returned despite of the state of the underlying file object. This
            is useful if you want the ``.data`` property to continue working
            after having ``.read()`` the file object. (Overridden if ``amt`` is
            set.)
        """
        self._init_decoder()
        if decode_content is None:
            decode_content = self.decode_content

        if self._fp is None:
            return

        flush_decoder = False
        fp_closed = getattr(self._fp, "closed", False)

        with self._error_catcher():
            if amt is None:
                # cStringIO doesn't like amt=None
                data = self._fp.read() if not fp_closed else b""
                flush_decoder = True
            else:
                cache_content = False
                data = self._fp.read(amt) if not fp_closed else b""
                if (
                    amt != 0 and not data
                ):  # Platform-specific: Buggy versions of Python.
                    # Close the connection when no data is returned
                    #
                    # This is redundant to what httplib/http.client _should_
                    # already do.  However, versions of python released before
                    # December 15, 2012 (http://bugs.python.org/issue16298) do
                    # not properly close the connection in all cases. There is
                    # no harm in redundantly calling close.
                    self._fp.close()
                    flush_decoder = True
                    if self.enforce_content_length and self.length_remaining not in (
                        0,
                        None,
                    ):
                        # This is an edge case that httplib failed to cover due
                        # to concerns of backward compatibility. We're
                        # addressing it here to make sure IncompleteRead is
                        # raised during streaming, so all calls with incorrect
                        # Content-Length are caught.
                        raise IncompleteRead(self._fp_bytes_read, self.length_remaining)

        if data:
            self._fp_bytes_read += len(data)
            if self.length_remaining is not None:
                self.length_remaining -= len(data)

            data = self._decode(data, decode_content, flush_decoder)

            if cache_content:
                self._body = data

        return data

    def stream(self, amt=2 ** 16, decode_content=None):
        """
        A generator wrapper for the read() method. A call will block until
        ``amt`` bytes have been read from the connection or until the
        connection is closed.

        :param amt:
            How much of the content to read. The generator will return up to
            much data per iteration, but may return less. This is particularly
            likely when using compressed data. However, the empty string will
            never be returned.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        """
        if self.chunked and self.supports_chunked_reads():
            for line in self.read_chunked(amt, decode_content=decode_content):
                yield line
        else:
            while not is_fp_closed(self._fp):
                data = self.read(amt=amt, decode_content=decode_content)

                if data:
                    yield data

    @classmethod
    def from_httplib(ResponseCls, r, **response_kw):
        """
        Given an :class:`http.client.HTTPResponse` instance ``r``, return a
        corresponding :class:`urllib3.response.HTTPResponse` object.

        Remaining parameters are passed to the HTTPResponse constructor, along
        with ``original_response=r``.
        """
        headers = r.msg

        if not isinstance(headers, HTTPHeaderDict):
            if six.PY2:
                # Python 2.7
                headers = HTTPHeaderDict.from_httplib(headers)
            else:
                headers = HTTPHeaderDict(headers.items())

        # HTTPResponse objects in Python 3 don't have a .strict attribute
        strict = getattr(r, "strict", 0)
        resp = ResponseCls(
            body=r,
            headers=headers,
            status=r.status,
            version=r.version,
            reason=r.reason,
            strict=strict,
            original_response=r,
            **response_kw
        )
        return resp

    # Backwards-compatibility methods for http.client.HTTPResponse
    def getheaders(self):
        return self.headers

    def getheader(self, name, default=None):
        return self.headers.get(name, default)

    # Backwards compatibility for http.cookiejar
    def info(self):
        return self.headers

    # Overrides from io.IOBase
    def close(self):
        if not self.closed:
            self._fp.close()

        if self._connection:
            self._connection.close()

        if not self.auto_close:
            io.IOBase.close(self)

    @property
    def closed(self):
        if not self.auto_close:
            return io.IOBase.closed.__get__(self)
        elif self._fp is None:
            return True
        elif hasattr(self._fp, "isclosed"):
            return self._fp.isclosed()
        elif hasattr(self._fp, "closed"):
            return self._fp.closed
        else:
            return True

    def fileno(self):
        if self._fp is None:
            raise IOError("HTTPResponse has no file to get a fileno from")
        elif hasattr(self._fp, "fileno"):
            return self._fp.fileno()
        else:
            raise IOError(
                "The file-like object this HTTPResponse is wrapped "
                "around has no file descriptor"
            )

    def flush(self):
        if (
            self._fp is not None
            and hasattr(self._fp, "flush")
            and not getattr(self._fp, "closed", False)
        ):
            return self._fp.flush()

    def readable(self):
        # This method is required for `io` module compatibility.
        return True

    def readinto(self, b):
        # This method is required for `io` module compatibility.
        temp = self.read(len(b))
        if len(temp) == 0:
            return 0
        else:
            b[: len(temp)] = temp
            return len(temp)

    def supports_chunked_reads(self):
        """
        Checks if the underlying file-like object looks like a
        :class:`http.client.HTTPResponse` object. We do this by testing for
        the fp attribute. If it is present we assume it returns raw chunks as
        processed by read_chunked().
        """
        return hasattr(self._fp, "fp")

    def _update_chunk_length(self):
        # First, we'll figure out length of a chunk and then
        # we'll try to read it from socket.
        if self.chunk_left is not None:
            return
        line = self._fp.fp.readline()
        line = line.split(b";", 1)[0]
        try:
            self.chunk_left = int(line, 16)
        except ValueError:
            # Invalid chunked protocol response, abort.
            self.close()
            raise InvalidChunkLength(self, line)

    def _handle_chunk(self, amt):
        returned_chunk = None
        if amt is None:
            chunk = self._fp._safe_read(self.chunk_left)
            returned_chunk = chunk
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
        elif amt < self.chunk_left:
            value = self._fp._safe_read(amt)
            self.chunk_left = self.chunk_left - amt
            returned_chunk = value
        elif amt == self.chunk_left:
            value = self._fp._safe_read(amt)
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
            returned_chunk = value
        else:  # amt > self.chunk_left
            returned_chunk = self._fp._safe_read(self.chunk_left)
            self._fp._safe_read(2)  # Toss the CRLF at the end of the chunk.
            self.chunk_left = None
        return returned_chunk

    def read_chunked(self, amt=None, decode_content=None):
        """
        Similar to :meth:`HTTPResponse.read`, but with an additional
        parameter: ``decode_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        """
        self._init_decoder()
        # FIXME: Rewrite this method and make it a class with a better structured logic.
        if not self.chunked:
            raise ResponseNotChunked(
                "Response is not chunked. "
                "Header 'transfer-encoding: chunked' is missing."
            )
        if not self.supports_chunked_reads():
            raise BodyNotHttplibCompatible(
                "Body should be http.client.HTTPResponse like. "
                "It should have have an fp attribute which returns raw chunks."
            )

        with self._error_catcher():
            # Don't bother reading the body of a HEAD request.
            if self._original_response and is_response_to_head(self._original_response):
                self._original_response.close()
                return

            # If a response is already read and closed
            # then return immediately.
            if self._fp.fp is None:
                return

            while True:
                self._update_chunk_length()
                if self.chunk_left == 0:
                    break
                chunk = self._handle_chunk(amt)
                decoded = self._decode(
                    chunk, decode_content=decode_content, flush_decoder=False
                )
                if decoded:
                    yield decoded

            if decode_content:
                # On CPython and PyPy, we should never need to flush the
                # decoder. However, on Jython we *might* need to, so
                # lets defensively do it anyway.
                decoded = self._flush_decoder()
                if decoded:  # Platform-specific: Jython.
                    yield decoded

            # Chunk content ends with \r\n: discard it.
            while True:
                line = self._fp.fp.readline()
                if not line:
                    # Some sites may not end with '\r\n'.
                    break
                if line == b"\r\n":
                    break

            # We read everything; close the "file".
            if self._original_response:
                self._original_response.close()

    def geturl(self):
        """
        Returns the URL that was the source of this response.
        If the request that generated this response redirected, this method
        will return the final redirect location.
        """
        if self.retries is not None and len(self.retries.history):
            return self.retries.history[-1].redirect_location
        else:
            return self._request_url

    def __iter__(self):
        buffer = []
        for chunk in self.stream(decode_content=True):
            if b"\n" in chunk:
                chunk = chunk.split(b"\n")
                yield b"".join(buffer) + chunk[0] + b"\n"
                for x in chunk[1:-1]:
                    yield x + b"\n"
                if chunk[-1]:
                    buffer = [chunk[-1]]
                else:
                    buffer = []
            else:
                buffer.append(chunk)
        if buffer:
            yield b"".join(buffer)
site-packages/pip/_vendor/urllib3/packages/__pycache__/six.cpython-39.pyc000064400000065746150564745010022337 0ustar00a

��Rej��E@sRdZddlmZddlZddlZddlZddlZddlZdZdZ	ej
ddkZej
ddkZej
dd�dkZ
er�efZefZefZeZeZejZn|efZeefZeejfZeZeZej�d	�r�ed
�ZnFGdd�de�Z ze!e ��Wne"�yed
�ZYn
0ed
�Z[ e
�r$ddl#m$Z$ndZ$dd�Z%dd�Z&Gdd�de�Z'Gdd�de'�Z(Gdd�dej)�Z*Gdd�de'�Z+Gdd�de�Z,e,e-�Z.Gdd�de*�Z/e+ddd d!�e+d"d#d$d%d"�e+d&d#d#d'd&�e+d(d)d$d*d(�e+d+d)d,�e+d-d#d$d.d-�e+d/d0d0d1d/�e+d2d0d0d/d2�e+d3d4d5�e+d6d)d$d7d6�e+d8d)e
�r4d9nd:d;�e+d<d)d=�e+d>d?d@dA�e+d!d!d �e+dBdBdC�e+dDdDdC�e+dEdEdC�e+d7d)d$d7d6�e+dFd#d$dGdF�e+dHd#d#dIdH�e(d$d)�e(dJdK�e(dLdCej
dMk�r�dNndC�e(dOdP�e(dQdRdS�e(dTdUdV�e(dWdXej
dYk�r�dWndZ�e(d[d\d]�e(d^d_d`�e(dadbdc�e(dddedf�e(dgdhdi�e(djdkdl�e(dmdndo�e(dpdqdr�e(dsdtdu�e(dvdwdx�e(dydydz�e(d{d{dz�e(d|d|dz�e(d}d}d~�e(dd��e(d�d��e(d�d��e(dZd�dZ�e(d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�d�d��e(d�e-d�d��e(d�e-d�d��e(d�e-d�e-d��e(d�d�d��e(d�d�d��e(d�d�d��gBZ0ejd�k�r�e0e(d�dÃg7Z0e0D]2Z1e2e/e1j3e1�e4e1e(��r�e.�5e1d�e1j3��q�[1e0e/_0e/e-d��Z6e.�5e6dơGd�dȄd�e*�Z7e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+dAd�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d�d�dփe+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��e+d�d�d��gZ8e8D]Z1e2e7e1j3e1��q[1e8e7_0e.�5e7e-d��d�d�Gd�d�d�e*�Z9e+d�d�d��e+d�d�d��e+d�d�d��gZ:e:D]Z1e2e9e1j3e1��q�[1e:e9_0e.�5e9e-d��d�d�Gd�d�d�e*�Z;e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+d�d�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�d	d�d�e+�d
d�d�e+�dd�d�e+�dd�d�e+�d
d�d�e+�dd�d�e+�dd�d�e+�dd�d�e+�dd�d�g#Z<e<D]Z1e2e;e1j3e1��qV[1e<e;_0e.�5e;e-�d��d�d�G�d�d��de*�Z=e+�dd��d�e+�dd��d�e+�dd��d�e+�dd��d�gZ>e>D]Z1e2e=e1j3e1��q�[1e>e=_0e.�5e=e-�d��d�d�G�d�d ��d e*�Z?e+�d!d�d��gZ@e@D]Z1e2e?e1j3e1��	qJ[1e@e?_0e.�5e?e-�d"��d#�d$�G�d%�d&��d&ej)�ZAe.�5eAe-d���d'��d(�d)�ZB�d*�d+�ZCe�	r��d,ZD�d-ZE�d.ZF�d/ZG�d0ZH�d1ZIn$�d2ZD�d3ZE�d4ZF�d5ZG�d6ZH�d7ZIzeJZKWn eL�
yB�d8�d9�ZKYn0eKZJzeMZMWn eL�
yp�d:�d;�ZMYn0e�
r��d<�d=�ZNejOZP�d>�d?�ZQeZRn>�d@�d=�ZN�dA�dB�ZP�dC�d?�ZQG�dD�dE��dEe�ZReMZMe%eN�dF�e�SeD�ZTe�SeE�ZUe�SeF�ZVe�SeG�ZWe�SeH�ZXe�SeI�ZYe�r~�dG�dH�ZZ�dI�dJ�Z[�dK�dL�Z\�dM�dN�Z]e�^�dO�Z_e�^�dP�Z`e�^�dQ�ZanT�dR�dH�ZZ�dS�dJ�Z[�dT�dL�Z\�dU�dN�Z]e�^�dV�Z_e�^�dW�Z`e�^�dX�Zae%eZ�dY�e%e[�dZ�e%e\�d[�e%e]�d\�e�r��d]�d^�Zb�d_�d`�ZcedZeddlfZfef�g�da�jhZi[fe�jd�ZkejlZmenZoddlpZpepjqZqepjrZr[p�dbZsej
�dc�dck�r��ddZt�deZu�dfZvn�dgZt�dhZu�diZvnp�dj�d^�Zb�dk�d`�ZceeZeedZi�dl�dm�Zk�dn�do�Zme�wejxey�ZoddlqZqeqjqZqZr�dpZs�ddZt�deZu�dfZve%eb�dq�e%ec�dr��ds�db�Zz�dt�dg�Z{�du�dh�Z|�dv�di�Z}e�
r�e~e6j�dw�Z��d��dx�dy�Z�n�d��dz�d{�Z�e��d|�ej
dd��d}k�
r�e��d~�n�d�d��Z�e~e6j�d�d�Z�e�du�
r�d��d��Z�ej
dd�dMk�re�Z��d��d��Z�e%e��d��ej
dd�dk�rzej�ej�f�d��d��Z�ej�je�_ej�ej�f�d��d��Z�ej�je�_nej�Z��d��d��Z��d��d��Z��d��d��d��Z��d��d��d��Z��d��d��d��Z��d��d��Z�gZ�e-Z�e�����d��du�r�ge�_�ej��rBe�ej��D]4\Z�Z�ee��j-dk�re�j3e-k�rej�e�=�q>�q[�[�ej���e.�dS(�z6Utilities for writing code that runs on Python 2 and 3�)�absolute_importNz'Benjamin Peterson <benjamin@python.org>z1.16.0��)r��javai���c@seZdZdd�ZdS)�XcCsdS)Nl���selfrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/six.py�__len__=sz	X.__len__N)�__name__�
__module__�__qualname__rrrrrr<srl����)�spec_from_loadercCs
||_dS)z Add documentation to a function.N)�__doc__)�func�docrrr�_add_docPsrcCst|�tj|S)z7Import module, returning the module after the last dot.)�
__import__�sys�modules��namerrr�_import_moduleUsrc@seZdZdd�Zdd�ZdS)�
_LazyDescrcCs
||_dS�Nr�r
rrrr�__init__\sz_LazyDescr.__init__cCs@|��}t||j|�zt|j|j�Wnty:Yn0|Sr)�_resolve�setattrr�delattr�	__class__�AttributeError)r
�obj�tp�resultrrr�__get___sz_LazyDescr.__get__N)r
rrrr'rrrrrrcs.eZdZd�fdd�	Zdd�Zdd�Z�ZS)	�MovedModuleNcs2tt|��|�tr(|dur |}||_n||_dSr)�superr(r�PY3�mod)r
r�old�new�r"rrrlszMovedModule.__init__cCs
t|j�Sr)rr+r	rrrruszMovedModule._resolvecCs"|��}t||�}t|||�|Sr)r�getattrr )r
�attr�_module�valuerrr�__getattr__xs
zMovedModule.__getattr__)N)r
rrrrr3�
__classcell__rrr.rr(ks	r(cs(eZdZ�fdd�Zdd�ZgZ�ZS)�_LazyModulecstt|��|�|jj|_dSr)r)r5rr"rrr.rrr�sz_LazyModule.__init__cCs ddg}|dd�|jD�7}|S)Nrr
cSsg|]
}|j�qSrr)�.0r0rrr�
<listcomp>��z'_LazyModule.__dir__.<locals>.<listcomp>)�_moved_attributes)r
�attrsrrr�__dir__�sz_LazyModule.__dir__)r
rrrr;r9r4rrr.rr5sr5cs&eZdZd�fdd�	Zdd�Z�ZS)�MovedAttributeNcsdtt|��|�trH|dur |}||_|dur@|dur<|}n|}||_n||_|durZ|}||_dSr)r)r<rr*r+r0)r
r�old_mod�new_mod�old_attr�new_attrr.rrr�szMovedAttribute.__init__cCst|j�}t||j�Sr)rr+r/r0�r
�modulerrrr�s
zMovedAttribute._resolve)NN)r
rrrrr4rrr.rr<�sr<c@speZdZdZdd�Zdd�Zdd�Zdd	d
�Zddd�Zd
d�Z	dd�Z
dd�Zdd�ZeZ
dd�Zdd�ZdS)�_SixMetaPathImporterz�
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    cCs||_i|_dSr)r�
known_modules)r
�six_module_namerrrr�sz_SixMetaPathImporter.__init__cGs"|D]}||j|jd|<qdS�N�.�rDr)r
r+�	fullnames�fullnamerrr�_add_module�sz _SixMetaPathImporter._add_modulecCs|j|jd|SrFrH�r
rJrrr�_get_module�sz _SixMetaPathImporter._get_moduleNcCs||jvr|SdSr)rD)r
rJ�pathrrr�find_module�s
z _SixMetaPathImporter.find_modulecCs||jvrt||�SdSr)rDr)r
rJrN�targetrrr�	find_spec�s

z_SixMetaPathImporter.find_speccCs0z|j|WSty*td|��Yn0dS)Nz!This loader does not know module )rD�KeyError�ImportErrorrLrrr�__get_module�sz!_SixMetaPathImporter.__get_modulecCsRztj|WStyYn0|�|�}t|t�r>|��}n||_|tj|<|Sr)rrrR� _SixMetaPathImporter__get_module�
isinstancer(r�
__loader__)r
rJr+rrr�load_module�s



z _SixMetaPathImporter.load_modulecCst|�|�d�S)z�
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        �__path__)�hasattrrUrLrrr�
is_package�sz_SixMetaPathImporter.is_packagecCs|�|�dS)z;Return None

        Required, if is_package is implementedN)rUrLrrr�get_code�s
z_SixMetaPathImporter.get_codecCs|�|j�Sr)rXr)r
�specrrr�
create_module�sz"_SixMetaPathImporter.create_modulecCsdSrrrArrr�exec_module�sz _SixMetaPathImporter.exec_module)N)N)r
rrrrrKrMrOrQrUrXr[r\�
get_sourcer^r_rrrrrC�s

	rCc@seZdZdZgZdS)�_MovedItemszLazy loading of moved objectsN)r
rrrrYrrrrra�sra�	cStringIO�io�StringIO�filter�	itertools�builtins�ifilter�filterfalse�ifilterfalse�input�__builtin__�	raw_input�internr�map�imap�getcwd�osZgetcwdu�getcwdb�	getoutput�commands�
subprocess�range�xrange�
reload_module�	importlib�imp�reload�reduce�	functools�shlex_quoteZpipes�shlex�quote�UserDict�collections�UserList�
UserString�zip�izip�zip_longest�izip_longest�configparser�ConfigParser�collections_abc)rrzcollections.abc�copyreg�copy_reg�dbm_gnu�gdbmzdbm.gnu�dbm_ndbm�dbmzdbm.ndbm�
_dummy_thread�dummy_thread)r�	�_thread�http_cookiejar�	cookielibzhttp.cookiejar�http_cookies�Cookiezhttp.cookies�
html_entities�htmlentitydefsz
html.entities�html_parser�
HTMLParserzhtml.parser�http_client�httplibzhttp.client�email_mime_basezemail.MIMEBasezemail.mime.base�email_mime_imagezemail.MIMEImagezemail.mime.image�email_mime_multipartzemail.MIMEMultipartzemail.mime.multipart�email_mime_nonmultipartzemail.MIMENonMultipartzemail.mime.nonmultipart�email_mime_textzemail.MIMETextzemail.mime.text�BaseHTTPServerzhttp.server�
CGIHTTPServer�SimpleHTTPServer�cPickle�pickle�queue�Queue�reprlib�repr�socketserver�SocketServer�thread�tkinter�Tkinter�tkinter_dialog�Dialogztkinter.dialog�tkinter_filedialog�
FileDialogztkinter.filedialog�tkinter_scrolledtext�ScrolledTextztkinter.scrolledtext�tkinter_simpledialog�SimpleDialogztkinter.simpledialog�tkinter_tix�Tixztkinter.tix�tkinter_ttk�ttkztkinter.ttk�tkinter_constants�Tkconstantsztkinter.constants�tkinter_dnd�Tkdndztkinter.dnd�tkinter_colorchooser�tkColorChooserztkinter.colorchooser�tkinter_commondialog�tkCommonDialogztkinter.commondialog�tkinter_tkfiledialog�tkFileDialog�tkinter_font�tkFontztkinter.font�tkinter_messagebox�tkMessageBoxztkinter.messagebox�tkinter_tksimpledialog�tkSimpleDialog�urllib_parsez.moves.urllib_parsezurllib.parse�urllib_errorz.moves.urllib_errorzurllib.error�urllibz
.moves.urllib�urllib_robotparser�robotparserzurllib.robotparser�
xmlrpc_client�	xmlrpclibz
xmlrpc.client�
xmlrpc_server�SimpleXMLRPCServerz
xmlrpc.server�win32�winreg�_winregzmoves.z.moves�movesc@seZdZdZdS)�Module_six_moves_urllib_parsez7Lazy loading of moved objects in six.moves.urllib_parseN�r
rrrrrrrr�`sr��ParseResult�urlparse�SplitResult�parse_qs�	parse_qsl�	urldefrag�urljoin�urlsplit�
urlunparse�
urlunsplit�
quote_plus�unquote�unquote_plus�unquote_to_bytes�	urlencode�
splitquery�splittag�	splituser�
splitvalue�
uses_fragment�uses_netloc�uses_params�
uses_query�
uses_relative�moves.urllib_parsezmoves.urllib.parsec@seZdZdZdS)�Module_six_moves_urllib_errorz7Lazy loading of moved objects in six.moves.urllib_errorNr�rrrrr��sr��URLError�urllib2�	HTTPError�ContentTooShortErrorz.moves.urllib.error�moves.urllib_errorzmoves.urllib.errorc@seZdZdZdS)�Module_six_moves_urllib_requestz9Lazy loading of moved objects in six.moves.urllib_requestNr�rrrrr��sr��urlopenzurllib.request�install_opener�build_opener�pathname2url�url2pathname�
getproxies�Request�OpenerDirector�HTTPDefaultErrorHandler�HTTPRedirectHandler�HTTPCookieProcessor�ProxyHandler�BaseHandler�HTTPPasswordMgr�HTTPPasswordMgrWithDefaultRealm�AbstractBasicAuthHandler�HTTPBasicAuthHandler�ProxyBasicAuthHandler�AbstractDigestAuthHandler�HTTPDigestAuthHandler�ProxyDigestAuthHandler�HTTPHandler�HTTPSHandler�FileHandler�
FTPHandler�CacheFTPHandler�UnknownHandler�HTTPErrorProcessor�urlretrieve�
urlcleanup�	URLopener�FancyURLopener�proxy_bypass�parse_http_list�parse_keqv_listz.moves.urllib.request�moves.urllib_requestzmoves.urllib.requestc@seZdZdZdS)� Module_six_moves_urllib_responsez:Lazy loading of moved objects in six.moves.urllib_responseNr�rrrrr#�sr#�addbasezurllib.response�addclosehook�addinfo�
addinfourlz.moves.urllib.response�moves.urllib_responsezmoves.urllib.responsec@seZdZdZdS)�#Module_six_moves_urllib_robotparserz=Lazy loading of moved objects in six.moves.urllib_robotparserNr�rrrrr)�sr)�RobotFileParserz.moves.urllib.robotparser�moves.urllib_robotparserzmoves.urllib.robotparserc@sNeZdZdZgZe�d�Ze�d�Ze�d�Z	e�d�Z
e�d�Zdd�Zd	S)
�Module_six_moves_urllibzICreate a six.moves.urllib namespace that resembles the Python 3 namespacer�r�r"r(r+cCsgd�S)N)�parse�error�request�responser�rr	rrrr;szModule_six_moves_urllib.__dir__N)
r
rrrrY�	_importerrMr-r.r/r0r�r;rrrrr,s




r,zmoves.urllibcCstt|j|�dS)zAdd an item to six.moves.N)r rar)�moverrr�add_move sr3cCsTztt|�Wn@tyNztj|=Wn tyHtd|f��Yn0Yn0dS)zRemove item from six.moves.zno such move, %rN)r!rar#r��__dict__rRrrrr�remove_move%sr5�__func__�__self__�__closure__�__code__�__defaults__�__globals__�im_funcZim_selfZfunc_closure�	func_codeZ
func_defaultsZfunc_globalscCs|��Sr)�next)�itrrr�advance_iteratorFsr@cCstdd�t|�jD��S)Ncss|]}d|jvVqdS)�__call__N)r4)r6�klassrrr�	<genexpr>Rr8zcallable.<locals>.<genexpr>)�any�type�__mro__)r$rrr�callableQsrGcCs|Srr��unboundrrr�get_unbound_functionWsrJcCs|Srr�r�clsrrr�create_unbound_method\srMcCs|jSr)r<rHrrrrJbscCst�|||j�Sr)�types�
MethodTyper")rr$rrr�create_bound_methodesrPcCst�|d|�Sr)rNrOrKrrrrMhsc@seZdZdd�ZdS)�IteratorcCst|��|�Sr)rE�__next__r	rrrr>lsz
Iterator.nextN)r
rrr>rrrrrQksrQz3Get the function out of a possibly unbound functioncKst|jfi|���Sr)�iter�keys��d�kwrrr�iterkeyssrXcKst|jfi|���Sr)rS�valuesrUrrr�
itervalues�srZcKst|jfi|���Sr)rS�itemsrUrrr�	iteritems�sr\cKst|jfi|���Sr)rS�listsrUrrr�	iterlists�sr^rTrYr[cKs|jfi|��Sr)rXrUrrrrX�scKs|jfi|��Sr)rZrUrrrrZ�scKs|jfi|��Sr)r\rUrrrr\�scKs|jfi|��Sr)r^rUrrrr^�s�viewkeys�
viewvalues�	viewitemsz1Return an iterator over the keys of a dictionary.z3Return an iterator over the values of a dictionary.z?Return an iterator over the (key, value) pairs of a dictionary.zBReturn an iterator over the (key, [values]) pairs of a dictionary.cCs
|�d�S)Nzlatin-1)�encode��srrr�b�srecCs|Srrrcrrr�u�srfz>B�assertCountEqual�ZassertRaisesRegexpZassertRegexpMatchesZassertNotRegexpMatches�assertRaisesRegex�assertRegex�assertNotRegexcCs|Srrrcrrrre�scCst|�dd�d�S)Nz\\z\\\\�unicode_escape)�unicode�replacercrrrrf�scCst|d�S)Nr��ord)�bsrrr�byte2int�srrcCst||�Srro)�buf�irrr�
indexbytes�sruZassertItemsEqualzByte literalzText literalcOst|t�|i|��Sr)r/�_assertCountEqual�r
�args�kwargsrrrrg�scOst|t�|i|��Sr)r/�_assertRaisesRegexrwrrrri�scOst|t�|i|��Sr)r/�_assertRegexrwrrrrj�scOst|t�|i|��Sr)r/�_assertNotRegexrwrrrrk�s�execcCsBz2|dur|�}|j|ur$|�|��|�Wd}d}n
d}d}0dSr)�
__traceback__�with_traceback)r%r2�tbrrr�reraise�s

�r�cCsB|dur*t�d�}|j}|dur&|j}~n|dur6|}td�dS)zExecute code in a namespace.Nrhzexec _code_ in _globs_, _locs_)r�	_getframe�	f_globals�f_localsr})Z_code_Z_globs_Z_locs_�framerrr�exec_
s
r�zedef reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
)rzrdef raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
cCs|�dSrr)r2�
from_valuerrr�
raise_from+sr��printc
s.|�dtj���durdS�fdd�}d}|�dd�}|dur`t|t�rNd}nt|t�s`td��|�d	d�}|dur�t|t�r�d}nt|t�s�td
��|r�td��|s�|D]}t|t�r�d}q�q�|r�td�}td
�}nd}d
}|dur�|}|dur�|}t|�D] \}	}|	�r||�||��q||�dS)z4The new-style print function for Python 2.4 and 2.5.�fileNcsdt|t�st|�}t�t�rVt|t�rV�jdurVt�dd�}|durHd}|��j|�}��|�dS)N�errors�strict)	rV�
basestring�strr�rm�encodingr/rb�write)�datar���fprrr�8s
���zprint_.<locals>.writeF�sepTzsep must be None or a string�endzend must be None or a stringz$invalid keyword arguments to print()�
� )�popr�stdoutrVrmr��	TypeError�	enumerate)
rxryr�Zwant_unicoder�r��arg�newline�spacertrr�r�print_2sL





r�cOs@|�dtj�}|�dd�}t|i|��|r<|dur<|��dS)Nr��flushF)�getrr�r��_printr�)rxryr�r�rrrr�os
zReraise an exception.c	Csh|D]6}zt||�}Wnty,YqYq0t|||�q|D]}t||��t||i��q@||_|Sr)r/r#r �update�__wrapped__)�wrapper�wrapped�assigned�updatedr0r2rrr�_update_wrappers
r�cCstjt|||d�S)N�r�r�r�)r~�partialr�r�rrr�wraps�s�r�cs&G��fdd�dt�}t�|ddi�S)z%Create a base class with a metaclass.cs,eZdZ��fdd�Ze��fdd��ZdS)z!with_metaclass.<locals>.metaclasscs>tjdd�dkr.t���}|�ur2�|d<n�}�|||�S)Nr)r��__orig_bases__)r�version_inforN�
resolve_bases)rLr�
this_basesrV�resolved_bases��bases�metarr�__new__�s

z)with_metaclass.<locals>.metaclass.__new__cs��|��Sr)�__prepare__)rLrr�r�rrr��sz-with_metaclass.<locals>.metaclass.__prepare__N)r
rrr��classmethodr�rr�rr�	metaclass�sr��temporary_classr)rEr�)r�r�r�rr�r�with_metaclass�sr�cs�fdd�}|S)z6Class decorator for creating a class with a metaclass.cs||j��}|�d�}|dur@t|t�r,|g}|D]}|�|�q0|�dd�|�dd�t|d�rl|j|d<�|j|j	|�S)N�	__slots__r4�__weakref__r)
r4�copyr�rVr�r�rZrr
�	__bases__)rL�	orig_vars�slots�	slots_var�r�rrr��s




zadd_metaclass.<locals>.wrapperr)r�r�rr�r�
add_metaclass�sr��utf-8r�cCs8t|t�r|St|t�r$|�||�Stdt|���dS)z�Coerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    �not expecting type '%s'N)rV�binary_type�	text_typerbr�rE�rdr�r�rrr�
ensure_binary�s


r�cCsft|�tur|Str*t|t�r*|�||�StrDt|t�rD|�||�St|ttf�sbt	dt|���|S)z�Coerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r�)
rEr��PY2rVr�rbr*r��decoder�r�rrr�
ensure_str�sr�cCs8t|t�r|�||�St|t�r$|Stdt|���dS)z�Coerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    r�N)rVr�r�r�r�rEr�rrr�ensure_text�s


r�cCs2tr.d|jvrtd|j��|j|_dd�|_|S)a
    A class decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    �__str__zY@python_2_unicode_compatible cannot be applied to %s because it doesn't define __str__().cSs|���d�S)Nr�)�__unicode__rbr	rrr�<lambda>r8z-python_2_unicode_compatible.<locals>.<lambda>)r�r4�
ValueErrorr
r�r�)rBrrr�python_2_unicode_compatible
s
��
r��__spec__)N)NN)r�r�)r�r�)r�r�)�r�
__future__rr~rf�operatorrrN�
__author__�__version__r�r�r*�PY34r��string_types�int�
integer_typesrE�class_typesr��bytesr��maxsize�MAXSIZEr��longZ	ClassTyperm�platform�
startswith�objectr�len�
OverflowError�importlib.utilrrrrr(�
ModuleTyper5r<rCr
r1rar9r0r rrVrKr�r��_urllib_parse_moved_attributesr��_urllib_error_moved_attributesr�� _urllib_request_moved_attributesr#�!_urllib_response_moved_attributesr)�$_urllib_robotparser_moved_attributesr,r3r5�
_meth_func�
_meth_self�
_func_closure�
_func_code�_func_defaults�
_func_globalsr>r@�	NameErrorrGrJrOrPrMrQ�
attrgetter�get_method_function�get_method_self�get_function_closure�get_function_code�get_function_defaults�get_function_globalsrXrZr\r^�methodcallerr_r`rarerf�chr�unichr�struct�Struct�pack�int2byte�
itemgetterrr�getitemrurS�	iterbytesrcrd�BytesIOrvrzr{r|r�rprprgrirjrkr/rgr�r�r�r�r��WRAPPER_ASSIGNMENTS�WRAPPER_UPDATESr��update_wrapperr�r�r�r�r�r�r�rY�__package__�globalsr�r��submodule_search_locations�	meta_pathr�rt�importer�appendrrrr�<module>sN

K
�

�





��

�







�






















�U�














�









�
�


�
�
















�%��������





�

�
�

:	�
�	��
site-packages/pip/_vendor/urllib3/packages/__pycache__/__init__.cpython-39.pyc000064400000000540150564745010023250 0ustar00a

��Rel�@s ddlmZddlmZdZdS)�)�absolute_import�)�ssl_match_hostnameN)�
__future__r�r�__all__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/__init__.py�<module>ssite-packages/pip/_vendor/urllib3/packages/backports/makefile.py000064400000002611150564745010020770 0ustar00# -*- coding: utf-8 -*-
"""
backports.makefile
~~~~~~~~~~~~~~~~~~

Backports the Python 3 ``socket.makefile`` method for use with anything that
wants to create a "fake" socket object.
"""
import io
from socket import SocketIO


def backport_makefile(
    self, mode="r", buffering=None, encoding=None, errors=None, newline=None
):
    """
    Backport of ``socket.makefile`` from Python 3.5.
    """
    if not set(mode) <= {"r", "w", "b"}:
        raise ValueError("invalid mode %r (only r, w, b allowed)" % (mode,))
    writing = "w" in mode
    reading = "r" in mode or not writing
    assert reading or writing
    binary = "b" in mode
    rawmode = ""
    if reading:
        rawmode += "r"
    if writing:
        rawmode += "w"
    raw = SocketIO(self, rawmode)
    self._makefile_refs += 1
    if buffering is None:
        buffering = -1
    if buffering < 0:
        buffering = io.DEFAULT_BUFFER_SIZE
    if buffering == 0:
        if not binary:
            raise ValueError("unbuffered streams must be binary")
        return raw
    if reading and writing:
        buffer = io.BufferedRWPair(raw, raw, buffering)
    elif reading:
        buffer = io.BufferedReader(raw, buffering)
    else:
        assert writing
        buffer = io.BufferedWriter(raw, buffering)
    if binary:
        return buffer
    text = io.TextIOWrapper(buffer, encoding, errors, newline)
    text.mode = mode
    return text
site-packages/pip/_vendor/urllib3/packages/backports/__pycache__/__init__.cpython-39.pyc000064400000000371150564745010025242 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/backports/__init__.py�<module>�site-packages/pip/_vendor/urllib3/packages/backports/__pycache__/makefile.cpython-39.pyc000064400000002503150564745010025257 0ustar00a

��Re��@s&dZddlZddlmZddd�ZdS)z�
backports.makefile
~~~~~~~~~~~~~~~~~~

Backports the Python 3 ``socket.makefile`` method for use with anything that
wants to create a "fake" socket object.
�N)�SocketIO�rc
Cst|�hd�kstd|f��d|v}d|vp2|}|s@|s@J�d|v}d}	|rX|	d7}	|rd|	d7}	t||	�}
|jd7_|dur�d	}|d
kr�tj}|d
kr�|s�td��|
S|r�|r�t�|
|
|�}n&|r�t�|
|�}n|s�J�t�|
|�}|r�|St�	||||�}||_
|S)z:
    Backport of ``socket.makefile`` from Python 3.5.
    >r�w�bz&invalid mode %r (only r, w, b allowed)rrr��N���rz!unbuffered streams must be binary)�set�
ValueErrorrZ_makefile_refs�io�DEFAULT_BUFFER_SIZE�BufferedRWPair�BufferedReader�BufferedWriter�
TextIOWrapper�mode)
�selfr�	buffering�encoding�errors�newline�writing�reading�binary�rawmode�raw�buffer�text�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/backports/makefile.py�backport_makefile
s@
r )rNNNN)�__doc__r�socketrr rrrr�<module>s�site-packages/pip/_vendor/urllib3/packages/backports/__init__.py000064400000000000150564745010020740 0ustar00site-packages/pip/_vendor/urllib3/packages/six.py000064400000103552150564745010016034 0ustar00# Copyright (c) 2010-2020 Benjamin Peterson
#
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
# copies of the Software, and to permit persons to whom the Software is
# furnished to do so, subject to the following conditions:
#
# The above copyright notice and this permission notice shall be included in all
# copies or substantial portions of the Software.
#
# THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
# SOFTWARE.

"""Utilities for writing code that runs on Python 2 and 3"""

from __future__ import absolute_import

import functools
import itertools
import operator
import sys
import types

__author__ = "Benjamin Peterson <benjamin@python.org>"
__version__ = "1.16.0"


# Useful for very coarse version differentiation.
PY2 = sys.version_info[0] == 2
PY3 = sys.version_info[0] == 3
PY34 = sys.version_info[0:2] >= (3, 4)

if PY3:
    string_types = (str,)
    integer_types = (int,)
    class_types = (type,)
    text_type = str
    binary_type = bytes

    MAXSIZE = sys.maxsize
else:
    string_types = (basestring,)
    integer_types = (int, long)
    class_types = (type, types.ClassType)
    text_type = unicode
    binary_type = str

    if sys.platform.startswith("java"):
        # Jython always uses 32 bits.
        MAXSIZE = int((1 << 31) - 1)
    else:
        # It's possible to have sizeof(long) != sizeof(Py_ssize_t).
        class X(object):
            def __len__(self):
                return 1 << 31

        try:
            len(X())
        except OverflowError:
            # 32-bit
            MAXSIZE = int((1 << 31) - 1)
        else:
            # 64-bit
            MAXSIZE = int((1 << 63) - 1)
        del X

if PY34:
    from importlib.util import spec_from_loader
else:
    spec_from_loader = None


def _add_doc(func, doc):
    """Add documentation to a function."""
    func.__doc__ = doc


def _import_module(name):
    """Import module, returning the module after the last dot."""
    __import__(name)
    return sys.modules[name]


class _LazyDescr(object):
    def __init__(self, name):
        self.name = name

    def __get__(self, obj, tp):
        result = self._resolve()
        setattr(obj, self.name, result)  # Invokes __set__.
        try:
            # This is a bit ugly, but it avoids running this again by
            # removing this descriptor.
            delattr(obj.__class__, self.name)
        except AttributeError:
            pass
        return result


class MovedModule(_LazyDescr):
    def __init__(self, name, old, new=None):
        super(MovedModule, self).__init__(name)
        if PY3:
            if new is None:
                new = name
            self.mod = new
        else:
            self.mod = old

    def _resolve(self):
        return _import_module(self.mod)

    def __getattr__(self, attr):
        _module = self._resolve()
        value = getattr(_module, attr)
        setattr(self, attr, value)
        return value


class _LazyModule(types.ModuleType):
    def __init__(self, name):
        super(_LazyModule, self).__init__(name)
        self.__doc__ = self.__class__.__doc__

    def __dir__(self):
        attrs = ["__doc__", "__name__"]
        attrs += [attr.name for attr in self._moved_attributes]
        return attrs

    # Subclasses should override this
    _moved_attributes = []


class MovedAttribute(_LazyDescr):
    def __init__(self, name, old_mod, new_mod, old_attr=None, new_attr=None):
        super(MovedAttribute, self).__init__(name)
        if PY3:
            if new_mod is None:
                new_mod = name
            self.mod = new_mod
            if new_attr is None:
                if old_attr is None:
                    new_attr = name
                else:
                    new_attr = old_attr
            self.attr = new_attr
        else:
            self.mod = old_mod
            if old_attr is None:
                old_attr = name
            self.attr = old_attr

    def _resolve(self):
        module = _import_module(self.mod)
        return getattr(module, self.attr)


class _SixMetaPathImporter(object):

    """
    A meta path importer to import six.moves and its submodules.

    This class implements a PEP302 finder and loader. It should be compatible
    with Python 2.5 and all existing versions of Python3
    """

    def __init__(self, six_module_name):
        self.name = six_module_name
        self.known_modules = {}

    def _add_module(self, mod, *fullnames):
        for fullname in fullnames:
            self.known_modules[self.name + "." + fullname] = mod

    def _get_module(self, fullname):
        return self.known_modules[self.name + "." + fullname]

    def find_module(self, fullname, path=None):
        if fullname in self.known_modules:
            return self
        return None

    def find_spec(self, fullname, path, target=None):
        if fullname in self.known_modules:
            return spec_from_loader(fullname, self)
        return None

    def __get_module(self, fullname):
        try:
            return self.known_modules[fullname]
        except KeyError:
            raise ImportError("This loader does not know module " + fullname)

    def load_module(self, fullname):
        try:
            # in case of a reload
            return sys.modules[fullname]
        except KeyError:
            pass
        mod = self.__get_module(fullname)
        if isinstance(mod, MovedModule):
            mod = mod._resolve()
        else:
            mod.__loader__ = self
        sys.modules[fullname] = mod
        return mod

    def is_package(self, fullname):
        """
        Return true, if the named module is a package.

        We need this method to get correct spec objects with
        Python 3.4 (see PEP451)
        """
        return hasattr(self.__get_module(fullname), "__path__")

    def get_code(self, fullname):
        """Return None

        Required, if is_package is implemented"""
        self.__get_module(fullname)  # eventually raises ImportError
        return None

    get_source = get_code  # same as get_code

    def create_module(self, spec):
        return self.load_module(spec.name)

    def exec_module(self, module):
        pass


_importer = _SixMetaPathImporter(__name__)


class _MovedItems(_LazyModule):

    """Lazy loading of moved objects"""

    __path__ = []  # mark as package


_moved_attributes = [
    MovedAttribute("cStringIO", "cStringIO", "io", "StringIO"),
    MovedAttribute("filter", "itertools", "builtins", "ifilter", "filter"),
    MovedAttribute(
        "filterfalse", "itertools", "itertools", "ifilterfalse", "filterfalse"
    ),
    MovedAttribute("input", "__builtin__", "builtins", "raw_input", "input"),
    MovedAttribute("intern", "__builtin__", "sys"),
    MovedAttribute("map", "itertools", "builtins", "imap", "map"),
    MovedAttribute("getcwd", "os", "os", "getcwdu", "getcwd"),
    MovedAttribute("getcwdb", "os", "os", "getcwd", "getcwdb"),
    MovedAttribute("getoutput", "commands", "subprocess"),
    MovedAttribute("range", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute(
        "reload_module", "__builtin__", "importlib" if PY34 else "imp", "reload"
    ),
    MovedAttribute("reduce", "__builtin__", "functools"),
    MovedAttribute("shlex_quote", "pipes", "shlex", "quote"),
    MovedAttribute("StringIO", "StringIO", "io"),
    MovedAttribute("UserDict", "UserDict", "collections"),
    MovedAttribute("UserList", "UserList", "collections"),
    MovedAttribute("UserString", "UserString", "collections"),
    MovedAttribute("xrange", "__builtin__", "builtins", "xrange", "range"),
    MovedAttribute("zip", "itertools", "builtins", "izip", "zip"),
    MovedAttribute(
        "zip_longest", "itertools", "itertools", "izip_longest", "zip_longest"
    ),
    MovedModule("builtins", "__builtin__"),
    MovedModule("configparser", "ConfigParser"),
    MovedModule(
        "collections_abc",
        "collections",
        "collections.abc" if sys.version_info >= (3, 3) else "collections",
    ),
    MovedModule("copyreg", "copy_reg"),
    MovedModule("dbm_gnu", "gdbm", "dbm.gnu"),
    MovedModule("dbm_ndbm", "dbm", "dbm.ndbm"),
    MovedModule(
        "_dummy_thread",
        "dummy_thread",
        "_dummy_thread" if sys.version_info < (3, 9) else "_thread",
    ),
    MovedModule("http_cookiejar", "cookielib", "http.cookiejar"),
    MovedModule("http_cookies", "Cookie", "http.cookies"),
    MovedModule("html_entities", "htmlentitydefs", "html.entities"),
    MovedModule("html_parser", "HTMLParser", "html.parser"),
    MovedModule("http_client", "httplib", "http.client"),
    MovedModule("email_mime_base", "email.MIMEBase", "email.mime.base"),
    MovedModule("email_mime_image", "email.MIMEImage", "email.mime.image"),
    MovedModule("email_mime_multipart", "email.MIMEMultipart", "email.mime.multipart"),
    MovedModule(
        "email_mime_nonmultipart", "email.MIMENonMultipart", "email.mime.nonmultipart"
    ),
    MovedModule("email_mime_text", "email.MIMEText", "email.mime.text"),
    MovedModule("BaseHTTPServer", "BaseHTTPServer", "http.server"),
    MovedModule("CGIHTTPServer", "CGIHTTPServer", "http.server"),
    MovedModule("SimpleHTTPServer", "SimpleHTTPServer", "http.server"),
    MovedModule("cPickle", "cPickle", "pickle"),
    MovedModule("queue", "Queue"),
    MovedModule("reprlib", "repr"),
    MovedModule("socketserver", "SocketServer"),
    MovedModule("_thread", "thread", "_thread"),
    MovedModule("tkinter", "Tkinter"),
    MovedModule("tkinter_dialog", "Dialog", "tkinter.dialog"),
    MovedModule("tkinter_filedialog", "FileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_scrolledtext", "ScrolledText", "tkinter.scrolledtext"),
    MovedModule("tkinter_simpledialog", "SimpleDialog", "tkinter.simpledialog"),
    MovedModule("tkinter_tix", "Tix", "tkinter.tix"),
    MovedModule("tkinter_ttk", "ttk", "tkinter.ttk"),
    MovedModule("tkinter_constants", "Tkconstants", "tkinter.constants"),
    MovedModule("tkinter_dnd", "Tkdnd", "tkinter.dnd"),
    MovedModule("tkinter_colorchooser", "tkColorChooser", "tkinter.colorchooser"),
    MovedModule("tkinter_commondialog", "tkCommonDialog", "tkinter.commondialog"),
    MovedModule("tkinter_tkfiledialog", "tkFileDialog", "tkinter.filedialog"),
    MovedModule("tkinter_font", "tkFont", "tkinter.font"),
    MovedModule("tkinter_messagebox", "tkMessageBox", "tkinter.messagebox"),
    MovedModule("tkinter_tksimpledialog", "tkSimpleDialog", "tkinter.simpledialog"),
    MovedModule("urllib_parse", __name__ + ".moves.urllib_parse", "urllib.parse"),
    MovedModule("urllib_error", __name__ + ".moves.urllib_error", "urllib.error"),
    MovedModule("urllib", __name__ + ".moves.urllib", __name__ + ".moves.urllib"),
    MovedModule("urllib_robotparser", "robotparser", "urllib.robotparser"),
    MovedModule("xmlrpc_client", "xmlrpclib", "xmlrpc.client"),
    MovedModule("xmlrpc_server", "SimpleXMLRPCServer", "xmlrpc.server"),
]
# Add windows specific modules.
if sys.platform == "win32":
    _moved_attributes += [
        MovedModule("winreg", "_winreg"),
    ]

for attr in _moved_attributes:
    setattr(_MovedItems, attr.name, attr)
    if isinstance(attr, MovedModule):
        _importer._add_module(attr, "moves." + attr.name)
del attr

_MovedItems._moved_attributes = _moved_attributes

moves = _MovedItems(__name__ + ".moves")
_importer._add_module(moves, "moves")


class Module_six_moves_urllib_parse(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_parse"""


_urllib_parse_moved_attributes = [
    MovedAttribute("ParseResult", "urlparse", "urllib.parse"),
    MovedAttribute("SplitResult", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qs", "urlparse", "urllib.parse"),
    MovedAttribute("parse_qsl", "urlparse", "urllib.parse"),
    MovedAttribute("urldefrag", "urlparse", "urllib.parse"),
    MovedAttribute("urljoin", "urlparse", "urllib.parse"),
    MovedAttribute("urlparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlsplit", "urlparse", "urllib.parse"),
    MovedAttribute("urlunparse", "urlparse", "urllib.parse"),
    MovedAttribute("urlunsplit", "urlparse", "urllib.parse"),
    MovedAttribute("quote", "urllib", "urllib.parse"),
    MovedAttribute("quote_plus", "urllib", "urllib.parse"),
    MovedAttribute("unquote", "urllib", "urllib.parse"),
    MovedAttribute("unquote_plus", "urllib", "urllib.parse"),
    MovedAttribute(
        "unquote_to_bytes", "urllib", "urllib.parse", "unquote", "unquote_to_bytes"
    ),
    MovedAttribute("urlencode", "urllib", "urllib.parse"),
    MovedAttribute("splitquery", "urllib", "urllib.parse"),
    MovedAttribute("splittag", "urllib", "urllib.parse"),
    MovedAttribute("splituser", "urllib", "urllib.parse"),
    MovedAttribute("splitvalue", "urllib", "urllib.parse"),
    MovedAttribute("uses_fragment", "urlparse", "urllib.parse"),
    MovedAttribute("uses_netloc", "urlparse", "urllib.parse"),
    MovedAttribute("uses_params", "urlparse", "urllib.parse"),
    MovedAttribute("uses_query", "urlparse", "urllib.parse"),
    MovedAttribute("uses_relative", "urlparse", "urllib.parse"),
]
for attr in _urllib_parse_moved_attributes:
    setattr(Module_six_moves_urllib_parse, attr.name, attr)
del attr

Module_six_moves_urllib_parse._moved_attributes = _urllib_parse_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_parse(__name__ + ".moves.urllib_parse"),
    "moves.urllib_parse",
    "moves.urllib.parse",
)


class Module_six_moves_urllib_error(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_error"""


_urllib_error_moved_attributes = [
    MovedAttribute("URLError", "urllib2", "urllib.error"),
    MovedAttribute("HTTPError", "urllib2", "urllib.error"),
    MovedAttribute("ContentTooShortError", "urllib", "urllib.error"),
]
for attr in _urllib_error_moved_attributes:
    setattr(Module_six_moves_urllib_error, attr.name, attr)
del attr

Module_six_moves_urllib_error._moved_attributes = _urllib_error_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_error(__name__ + ".moves.urllib.error"),
    "moves.urllib_error",
    "moves.urllib.error",
)


class Module_six_moves_urllib_request(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_request"""


_urllib_request_moved_attributes = [
    MovedAttribute("urlopen", "urllib2", "urllib.request"),
    MovedAttribute("install_opener", "urllib2", "urllib.request"),
    MovedAttribute("build_opener", "urllib2", "urllib.request"),
    MovedAttribute("pathname2url", "urllib", "urllib.request"),
    MovedAttribute("url2pathname", "urllib", "urllib.request"),
    MovedAttribute("getproxies", "urllib", "urllib.request"),
    MovedAttribute("Request", "urllib2", "urllib.request"),
    MovedAttribute("OpenerDirector", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDefaultErrorHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPRedirectHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPCookieProcessor", "urllib2", "urllib.request"),
    MovedAttribute("ProxyHandler", "urllib2", "urllib.request"),
    MovedAttribute("BaseHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgr", "urllib2", "urllib.request"),
    MovedAttribute("HTTPPasswordMgrWithDefaultRealm", "urllib2", "urllib.request"),
    MovedAttribute("AbstractBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyBasicAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("AbstractDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("ProxyDigestAuthHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPSHandler", "urllib2", "urllib.request"),
    MovedAttribute("FileHandler", "urllib2", "urllib.request"),
    MovedAttribute("FTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("CacheFTPHandler", "urllib2", "urllib.request"),
    MovedAttribute("UnknownHandler", "urllib2", "urllib.request"),
    MovedAttribute("HTTPErrorProcessor", "urllib2", "urllib.request"),
    MovedAttribute("urlretrieve", "urllib", "urllib.request"),
    MovedAttribute("urlcleanup", "urllib", "urllib.request"),
    MovedAttribute("URLopener", "urllib", "urllib.request"),
    MovedAttribute("FancyURLopener", "urllib", "urllib.request"),
    MovedAttribute("proxy_bypass", "urllib", "urllib.request"),
    MovedAttribute("parse_http_list", "urllib2", "urllib.request"),
    MovedAttribute("parse_keqv_list", "urllib2", "urllib.request"),
]
for attr in _urllib_request_moved_attributes:
    setattr(Module_six_moves_urllib_request, attr.name, attr)
del attr

Module_six_moves_urllib_request._moved_attributes = _urllib_request_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_request(__name__ + ".moves.urllib.request"),
    "moves.urllib_request",
    "moves.urllib.request",
)


class Module_six_moves_urllib_response(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_response"""


_urllib_response_moved_attributes = [
    MovedAttribute("addbase", "urllib", "urllib.response"),
    MovedAttribute("addclosehook", "urllib", "urllib.response"),
    MovedAttribute("addinfo", "urllib", "urllib.response"),
    MovedAttribute("addinfourl", "urllib", "urllib.response"),
]
for attr in _urllib_response_moved_attributes:
    setattr(Module_six_moves_urllib_response, attr.name, attr)
del attr

Module_six_moves_urllib_response._moved_attributes = _urllib_response_moved_attributes

_importer._add_module(
    Module_six_moves_urllib_response(__name__ + ".moves.urllib.response"),
    "moves.urllib_response",
    "moves.urllib.response",
)


class Module_six_moves_urllib_robotparser(_LazyModule):

    """Lazy loading of moved objects in six.moves.urllib_robotparser"""


_urllib_robotparser_moved_attributes = [
    MovedAttribute("RobotFileParser", "robotparser", "urllib.robotparser"),
]
for attr in _urllib_robotparser_moved_attributes:
    setattr(Module_six_moves_urllib_robotparser, attr.name, attr)
del attr

Module_six_moves_urllib_robotparser._moved_attributes = (
    _urllib_robotparser_moved_attributes
)

_importer._add_module(
    Module_six_moves_urllib_robotparser(__name__ + ".moves.urllib.robotparser"),
    "moves.urllib_robotparser",
    "moves.urllib.robotparser",
)


class Module_six_moves_urllib(types.ModuleType):

    """Create a six.moves.urllib namespace that resembles the Python 3 namespace"""

    __path__ = []  # mark as package
    parse = _importer._get_module("moves.urllib_parse")
    error = _importer._get_module("moves.urllib_error")
    request = _importer._get_module("moves.urllib_request")
    response = _importer._get_module("moves.urllib_response")
    robotparser = _importer._get_module("moves.urllib_robotparser")

    def __dir__(self):
        return ["parse", "error", "request", "response", "robotparser"]


_importer._add_module(
    Module_six_moves_urllib(__name__ + ".moves.urllib"), "moves.urllib"
)


def add_move(move):
    """Add an item to six.moves."""
    setattr(_MovedItems, move.name, move)


def remove_move(name):
    """Remove item from six.moves."""
    try:
        delattr(_MovedItems, name)
    except AttributeError:
        try:
            del moves.__dict__[name]
        except KeyError:
            raise AttributeError("no such move, %r" % (name,))


if PY3:
    _meth_func = "__func__"
    _meth_self = "__self__"

    _func_closure = "__closure__"
    _func_code = "__code__"
    _func_defaults = "__defaults__"
    _func_globals = "__globals__"
else:
    _meth_func = "im_func"
    _meth_self = "im_self"

    _func_closure = "func_closure"
    _func_code = "func_code"
    _func_defaults = "func_defaults"
    _func_globals = "func_globals"


try:
    advance_iterator = next
except NameError:

    def advance_iterator(it):
        return it.next()


next = advance_iterator


try:
    callable = callable
except NameError:

    def callable(obj):
        return any("__call__" in klass.__dict__ for klass in type(obj).__mro__)


if PY3:

    def get_unbound_function(unbound):
        return unbound

    create_bound_method = types.MethodType

    def create_unbound_method(func, cls):
        return func

    Iterator = object
else:

    def get_unbound_function(unbound):
        return unbound.im_func

    def create_bound_method(func, obj):
        return types.MethodType(func, obj, obj.__class__)

    def create_unbound_method(func, cls):
        return types.MethodType(func, None, cls)

    class Iterator(object):
        def next(self):
            return type(self).__next__(self)

    callable = callable
_add_doc(
    get_unbound_function, """Get the function out of a possibly unbound function"""
)


get_method_function = operator.attrgetter(_meth_func)
get_method_self = operator.attrgetter(_meth_self)
get_function_closure = operator.attrgetter(_func_closure)
get_function_code = operator.attrgetter(_func_code)
get_function_defaults = operator.attrgetter(_func_defaults)
get_function_globals = operator.attrgetter(_func_globals)


if PY3:

    def iterkeys(d, **kw):
        return iter(d.keys(**kw))

    def itervalues(d, **kw):
        return iter(d.values(**kw))

    def iteritems(d, **kw):
        return iter(d.items(**kw))

    def iterlists(d, **kw):
        return iter(d.lists(**kw))

    viewkeys = operator.methodcaller("keys")

    viewvalues = operator.methodcaller("values")

    viewitems = operator.methodcaller("items")
else:

    def iterkeys(d, **kw):
        return d.iterkeys(**kw)

    def itervalues(d, **kw):
        return d.itervalues(**kw)

    def iteritems(d, **kw):
        return d.iteritems(**kw)

    def iterlists(d, **kw):
        return d.iterlists(**kw)

    viewkeys = operator.methodcaller("viewkeys")

    viewvalues = operator.methodcaller("viewvalues")

    viewitems = operator.methodcaller("viewitems")

_add_doc(iterkeys, "Return an iterator over the keys of a dictionary.")
_add_doc(itervalues, "Return an iterator over the values of a dictionary.")
_add_doc(iteritems, "Return an iterator over the (key, value) pairs of a dictionary.")
_add_doc(
    iterlists, "Return an iterator over the (key, [values]) pairs of a dictionary."
)


if PY3:

    def b(s):
        return s.encode("latin-1")

    def u(s):
        return s

    unichr = chr
    import struct

    int2byte = struct.Struct(">B").pack
    del struct
    byte2int = operator.itemgetter(0)
    indexbytes = operator.getitem
    iterbytes = iter
    import io

    StringIO = io.StringIO
    BytesIO = io.BytesIO
    del io
    _assertCountEqual = "assertCountEqual"
    if sys.version_info[1] <= 1:
        _assertRaisesRegex = "assertRaisesRegexp"
        _assertRegex = "assertRegexpMatches"
        _assertNotRegex = "assertNotRegexpMatches"
    else:
        _assertRaisesRegex = "assertRaisesRegex"
        _assertRegex = "assertRegex"
        _assertNotRegex = "assertNotRegex"
else:

    def b(s):
        return s

    # Workaround for standalone backslash

    def u(s):
        return unicode(s.replace(r"\\", r"\\\\"), "unicode_escape")

    unichr = unichr
    int2byte = chr

    def byte2int(bs):
        return ord(bs[0])

    def indexbytes(buf, i):
        return ord(buf[i])

    iterbytes = functools.partial(itertools.imap, ord)
    import StringIO

    StringIO = BytesIO = StringIO.StringIO
    _assertCountEqual = "assertItemsEqual"
    _assertRaisesRegex = "assertRaisesRegexp"
    _assertRegex = "assertRegexpMatches"
    _assertNotRegex = "assertNotRegexpMatches"
_add_doc(b, """Byte literal""")
_add_doc(u, """Text literal""")


def assertCountEqual(self, *args, **kwargs):
    return getattr(self, _assertCountEqual)(*args, **kwargs)


def assertRaisesRegex(self, *args, **kwargs):
    return getattr(self, _assertRaisesRegex)(*args, **kwargs)


def assertRegex(self, *args, **kwargs):
    return getattr(self, _assertRegex)(*args, **kwargs)


def assertNotRegex(self, *args, **kwargs):
    return getattr(self, _assertNotRegex)(*args, **kwargs)


if PY3:
    exec_ = getattr(moves.builtins, "exec")

    def reraise(tp, value, tb=None):
        try:
            if value is None:
                value = tp()
            if value.__traceback__ is not tb:
                raise value.with_traceback(tb)
            raise value
        finally:
            value = None
            tb = None


else:

    def exec_(_code_, _globs_=None, _locs_=None):
        """Execute code in a namespace."""
        if _globs_ is None:
            frame = sys._getframe(1)
            _globs_ = frame.f_globals
            if _locs_ is None:
                _locs_ = frame.f_locals
            del frame
        elif _locs_ is None:
            _locs_ = _globs_
        exec ("""exec _code_ in _globs_, _locs_""")

    exec_(
        """def reraise(tp, value, tb=None):
    try:
        raise tp, value, tb
    finally:
        tb = None
"""
    )


if sys.version_info[:2] > (3,):
    exec_(
        """def raise_from(value, from_value):
    try:
        raise value from from_value
    finally:
        value = None
"""
    )
else:

    def raise_from(value, from_value):
        raise value


print_ = getattr(moves.builtins, "print", None)
if print_ is None:

    def print_(*args, **kwargs):
        """The new-style print function for Python 2.4 and 2.5."""
        fp = kwargs.pop("file", sys.stdout)
        if fp is None:
            return

        def write(data):
            if not isinstance(data, basestring):
                data = str(data)
            # If the file has an encoding, encode unicode with it.
            if (
                isinstance(fp, file)
                and isinstance(data, unicode)
                and fp.encoding is not None
            ):
                errors = getattr(fp, "errors", None)
                if errors is None:
                    errors = "strict"
                data = data.encode(fp.encoding, errors)
            fp.write(data)

        want_unicode = False
        sep = kwargs.pop("sep", None)
        if sep is not None:
            if isinstance(sep, unicode):
                want_unicode = True
            elif not isinstance(sep, str):
                raise TypeError("sep must be None or a string")
        end = kwargs.pop("end", None)
        if end is not None:
            if isinstance(end, unicode):
                want_unicode = True
            elif not isinstance(end, str):
                raise TypeError("end must be None or a string")
        if kwargs:
            raise TypeError("invalid keyword arguments to print()")
        if not want_unicode:
            for arg in args:
                if isinstance(arg, unicode):
                    want_unicode = True
                    break
        if want_unicode:
            newline = unicode("\n")
            space = unicode(" ")
        else:
            newline = "\n"
            space = " "
        if sep is None:
            sep = space
        if end is None:
            end = newline
        for i, arg in enumerate(args):
            if i:
                write(sep)
            write(arg)
        write(end)


if sys.version_info[:2] < (3, 3):
    _print = print_

    def print_(*args, **kwargs):
        fp = kwargs.get("file", sys.stdout)
        flush = kwargs.pop("flush", False)
        _print(*args, **kwargs)
        if flush and fp is not None:
            fp.flush()


_add_doc(reraise, """Reraise an exception.""")

if sys.version_info[0:2] < (3, 4):
    # This does exactly the same what the :func:`py3:functools.update_wrapper`
    # function does on Python versions after 3.2. It sets the ``__wrapped__``
    # attribute on ``wrapper`` object and it doesn't raise an error if any of
    # the attributes mentioned in ``assigned`` and ``updated`` are missing on
    # ``wrapped`` object.
    def _update_wrapper(
        wrapper,
        wrapped,
        assigned=functools.WRAPPER_ASSIGNMENTS,
        updated=functools.WRAPPER_UPDATES,
    ):
        for attr in assigned:
            try:
                value = getattr(wrapped, attr)
            except AttributeError:
                continue
            else:
                setattr(wrapper, attr, value)
        for attr in updated:
            getattr(wrapper, attr).update(getattr(wrapped, attr, {}))
        wrapper.__wrapped__ = wrapped
        return wrapper

    _update_wrapper.__doc__ = functools.update_wrapper.__doc__

    def wraps(
        wrapped,
        assigned=functools.WRAPPER_ASSIGNMENTS,
        updated=functools.WRAPPER_UPDATES,
    ):
        return functools.partial(
            _update_wrapper, wrapped=wrapped, assigned=assigned, updated=updated
        )

    wraps.__doc__ = functools.wraps.__doc__

else:
    wraps = functools.wraps


def with_metaclass(meta, *bases):
    """Create a base class with a metaclass."""
    # This requires a bit of explanation: the basic idea is to make a dummy
    # metaclass for one level of class instantiation that replaces itself with
    # the actual metaclass.
    class metaclass(type):
        def __new__(cls, name, this_bases, d):
            if sys.version_info[:2] >= (3, 7):
                # This version introduced PEP 560 that requires a bit
                # of extra care (we mimic what is done by __build_class__).
                resolved_bases = types.resolve_bases(bases)
                if resolved_bases is not bases:
                    d["__orig_bases__"] = bases
            else:
                resolved_bases = bases
            return meta(name, resolved_bases, d)

        @classmethod
        def __prepare__(cls, name, this_bases):
            return meta.__prepare__(name, bases)

    return type.__new__(metaclass, "temporary_class", (), {})


def add_metaclass(metaclass):
    """Class decorator for creating a class with a metaclass."""

    def wrapper(cls):
        orig_vars = cls.__dict__.copy()
        slots = orig_vars.get("__slots__")
        if slots is not None:
            if isinstance(slots, str):
                slots = [slots]
            for slots_var in slots:
                orig_vars.pop(slots_var)
        orig_vars.pop("__dict__", None)
        orig_vars.pop("__weakref__", None)
        if hasattr(cls, "__qualname__"):
            orig_vars["__qualname__"] = cls.__qualname__
        return metaclass(cls.__name__, cls.__bases__, orig_vars)

    return wrapper


def ensure_binary(s, encoding="utf-8", errors="strict"):
    """Coerce **s** to six.binary_type.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> encoded to `bytes`
      - `bytes` -> `bytes`
    """
    if isinstance(s, binary_type):
        return s
    if isinstance(s, text_type):
        return s.encode(encoding, errors)
    raise TypeError("not expecting type '%s'" % type(s))


def ensure_str(s, encoding="utf-8", errors="strict"):
    """Coerce *s* to `str`.

    For Python 2:
      - `unicode` -> encoded to `str`
      - `str` -> `str`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    # Optimization: Fast return for the common case.
    if type(s) is str:
        return s
    if PY2 and isinstance(s, text_type):
        return s.encode(encoding, errors)
    elif PY3 and isinstance(s, binary_type):
        return s.decode(encoding, errors)
    elif not isinstance(s, (text_type, binary_type)):
        raise TypeError("not expecting type '%s'" % type(s))
    return s


def ensure_text(s, encoding="utf-8", errors="strict"):
    """Coerce *s* to six.text_type.

    For Python 2:
      - `unicode` -> `unicode`
      - `str` -> `unicode`

    For Python 3:
      - `str` -> `str`
      - `bytes` -> decoded to `str`
    """
    if isinstance(s, binary_type):
        return s.decode(encoding, errors)
    elif isinstance(s, text_type):
        return s
    else:
        raise TypeError("not expecting type '%s'" % type(s))


def python_2_unicode_compatible(klass):
    """
    A class decorator that defines __unicode__ and __str__ methods under Python 2.
    Under Python 3 it does nothing.

    To support Python 2 and 3 with a single code base, define a __str__ method
    returning text and apply this decorator to the class.
    """
    if PY2:
        if "__str__" not in klass.__dict__:
            raise ValueError(
                "@python_2_unicode_compatible cannot be applied "
                "to %s because it doesn't define __str__()." % klass.__name__
            )
        klass.__unicode__ = klass.__str__
        klass.__str__ = lambda self: self.__unicode__().encode("utf-8")
    return klass


# Complete the moves implementation.
# This code is at the end of this module to speed up module loading.
# Turn this module into a package.
__path__ = []  # required for PEP 302 and PEP 451
__package__ = __name__  # see PEP 366 @ReservedAssignment
if globals().get("__spec__") is not None:
    __spec__.submodule_search_locations = []  # PEP 451 @UndefinedVariable
# Remove other six meta path importers, since they cause problems. This can
# happen if six is removed from sys.modules and then reloaded. (Setuptools does
# this for some reason.)
if sys.meta_path:
    for i, importer in enumerate(sys.meta_path):
        # Here's some real nastiness: Another "instance" of the six module might
        # be floating around. Therefore, we can't use isinstance() to check for
        # the six meta path importer, since the other six instance will have
        # inserted an importer with different class.
        if (
            type(importer).__name__ == "_SixMetaPathImporter"
            and importer.name == __name__
        ):
            del sys.meta_path[i]
            break
    del i, importer
# Finally, add the importer to the meta path import hook.
sys.meta_path.append(_importer)
site-packages/pip/_vendor/urllib3/packages/__init__.py000064400000000154150564745010016762 0ustar00from __future__ import absolute_import

from . import ssl_match_hostname

__all__ = ("ssl_match_hostname",)
pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/_implementation.cpython-39.pyc000064400000006416150564745010030501 0ustar00site-packagesa

��Re/�@srdZddlZddlZzddlZWney6dZYn0dZGdd�de�Zddd�Zd	d
�Z	dd�Z
d
d�ZdS)zJThe match_hostname() function from Python 3.3.3, essential when using SSL.�Nz3.5.0.1c@seZdZdS)�CertificateErrorN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.pyrsr�c
Cs�g}|sdS|�d�}|d}|dd�}|�d�}||krLtdt|���|s`|��|��kS|dkrt|�d�n>|�d	�s�|�d	�r�|�t�|��n|�t�|��	d
d��|D]}|�t�|��q�t�
dd
�|�dtj�}	|	�
|�S)zhMatching according to RFC 6125, section 6.4.3

    http://tools.ietf.org/html/rfc6125#section-6.4.3
    F�.rrN�*z,too many wildcards in certificate DNS name: z[^.]+zxn--z\*z[^.]*z\Az\.z\Z)�split�countr�repr�lower�append�
startswith�re�escape�replace�compile�join�
IGNORECASE�match)
�dn�hostnameZ
max_wildcardsZpats�partsZleftmost�	remainder�	wildcards�frag�patrrr�_dnsname_matchs,


�rcCs&t|t�r"tjdkr"t|ddd�}|S)N)��ascii�strict)�encoding�errors)�
isinstance�str�sys�version_info�unicode)�objrrr�_to_unicodeOsr+cCst�t|����}||kS)z�Exact matching of IP addresses.

    RFC 6125 explicitly doesn't define an algorithm for this
    (section 1.7.2 - "Out of Scope").
    )�	ipaddress�
ip_addressr+�rstrip)�ipname�host_ip�iprrr�_ipaddress_matchUsr2cCs�|std��zt�t|��}WnJty4d}Yn6tyHd}Yn"tyhtdurbd}n�Yn0g}|�dd�}|D]^\}}|dkr�|dur�t||�r�dS|�|�q~|dkr~|dur�t	||�r�dS|�|�q~|�s*|�dd�D]8}|D].\}}|dkr�t||��rdS|�|�q�q�t
|�d	k�rVtd
|d�t
t|��f��n,t
|�d	k�rztd||d
f��ntd��dS)a)Verify that *cert* (in decoded format as returned by
    SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
    rules are followed, but IP addresses are not accepted for *hostname*.

    CertificateError is raised on failure. On success, the function
    returns nothing.
    ztempty or no certificate, match_hostname needs a SSL socket or SSL context with either CERT_OPTIONAL or CERT_REQUIREDN�subjectAltNamer�DNSz
IP Address�subject�
commonNamerz&hostname %r doesn't match either of %sz, zhostname %r doesn't match %rrz=no appropriate commonName or subjectAltName fields were found)�
ValueErrorr,r-r+�UnicodeError�AttributeError�getrrr2�lenrr�mapr
)�certrr0�dnsnames�san�key�value�subrrr�match_hostnameasV����rC)r)�__doc__rr'r,�ImportError�__version__r7rrr+r2rCrrrr�<module>s

6site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__pycache__/__init__.cpython-39.pyc000064400000001143150564745010027123 0ustar00a

��Re��@s�ddlZz0ejdksejdkr&ed��ddlmZmZWnJey�zddlmZmZWn"ey|ddlmZmZYn0Yn0dZdS)�N)��)r�
zFallback to vendored code)�CertificateError�match_hostname�)	�sys�version_info�ImportError�sslrrZbackports.ssl_match_hostnameZ_implementation�__all__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py�<module>ssite-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/_implementation.py000064400000013057150564745010024270 0ustar00"""The match_hostname() function from Python 3.3.3, essential when using SSL."""

# Note: This file is under the PSF license as the code comes from the python
# stdlib.   http://docs.python.org/3/license.html

import re
import sys

# ipaddress has been backported to 2.6+ in pypi.  If it is installed on the
# system, use it to handle IPAddress ServerAltnames (this was added in
# python-3.5) otherwise only do DNS matching.  This allows
# backports.ssl_match_hostname to continue to be used in Python 2.7.
try:
    import ipaddress
except ImportError:
    ipaddress = None

__version__ = "3.5.0.1"


class CertificateError(ValueError):
    pass


def _dnsname_match(dn, hostname, max_wildcards=1):
    """Matching according to RFC 6125, section 6.4.3

    http://tools.ietf.org/html/rfc6125#section-6.4.3
    """
    pats = []
    if not dn:
        return False

    # Ported from python3-syntax:
    # leftmost, *remainder = dn.split(r'.')
    parts = dn.split(r".")
    leftmost = parts[0]
    remainder = parts[1:]

    wildcards = leftmost.count("*")
    if wildcards > max_wildcards:
        # Issue #17980: avoid denials of service by refusing more
        # than one wildcard per fragment.  A survey of established
        # policy among SSL implementations showed it to be a
        # reasonable choice.
        raise CertificateError(
            "too many wildcards in certificate DNS name: " + repr(dn)
        )

    # speed up common case w/o wildcards
    if not wildcards:
        return dn.lower() == hostname.lower()

    # RFC 6125, section 6.4.3, subitem 1.
    # The client SHOULD NOT attempt to match a presented identifier in which
    # the wildcard character comprises a label other than the left-most label.
    if leftmost == "*":
        # When '*' is a fragment by itself, it matches a non-empty dotless
        # fragment.
        pats.append("[^.]+")
    elif leftmost.startswith("xn--") or hostname.startswith("xn--"):
        # RFC 6125, section 6.4.3, subitem 3.
        # The client SHOULD NOT attempt to match a presented identifier
        # where the wildcard character is embedded within an A-label or
        # U-label of an internationalized domain name.
        pats.append(re.escape(leftmost))
    else:
        # Otherwise, '*' matches any dotless string, e.g. www*
        pats.append(re.escape(leftmost).replace(r"\*", "[^.]*"))

    # add the remaining fragments, ignore any wildcards
    for frag in remainder:
        pats.append(re.escape(frag))

    pat = re.compile(r"\A" + r"\.".join(pats) + r"\Z", re.IGNORECASE)
    return pat.match(hostname)


def _to_unicode(obj):
    if isinstance(obj, str) and sys.version_info < (3,):
        obj = unicode(obj, encoding="ascii", errors="strict")
    return obj


def _ipaddress_match(ipname, host_ip):
    """Exact matching of IP addresses.

    RFC 6125 explicitly doesn't define an algorithm for this
    (section 1.7.2 - "Out of Scope").
    """
    # OpenSSL may add a trailing newline to a subjectAltName's IP address
    # Divergence from upstream: ipaddress can't handle byte str
    ip = ipaddress.ip_address(_to_unicode(ipname).rstrip())
    return ip == host_ip


def match_hostname(cert, hostname):
    """Verify that *cert* (in decoded format as returned by
    SSLSocket.getpeercert()) matches the *hostname*.  RFC 2818 and RFC 6125
    rules are followed, but IP addresses are not accepted for *hostname*.

    CertificateError is raised on failure. On success, the function
    returns nothing.
    """
    if not cert:
        raise ValueError(
            "empty or no certificate, match_hostname needs a "
            "SSL socket or SSL context with either "
            "CERT_OPTIONAL or CERT_REQUIRED"
        )
    try:
        # Divergence from upstream: ipaddress can't handle byte str
        host_ip = ipaddress.ip_address(_to_unicode(hostname))
    except ValueError:
        # Not an IP address (common case)
        host_ip = None
    except UnicodeError:
        # Divergence from upstream: Have to deal with ipaddress not taking
        # byte strings.  addresses should be all ascii, so we consider it not
        # an ipaddress in this case
        host_ip = None
    except AttributeError:
        # Divergence from upstream: Make ipaddress library optional
        if ipaddress is None:
            host_ip = None
        else:
            raise
    dnsnames = []
    san = cert.get("subjectAltName", ())
    for key, value in san:
        if key == "DNS":
            if host_ip is None and _dnsname_match(value, hostname):
                return
            dnsnames.append(value)
        elif key == "IP Address":
            if host_ip is not None and _ipaddress_match(value, host_ip):
                return
            dnsnames.append(value)
    if not dnsnames:
        # The subject is only checked when there is no dNSName entry
        # in subjectAltName
        for sub in cert.get("subject", ()):
            for key, value in sub:
                # XXX according to RFC 2818, the most specific Common Name
                # must be used.
                if key == "commonName":
                    if _dnsname_match(value, hostname):
                        return
                    dnsnames.append(value)
    if len(dnsnames) > 1:
        raise CertificateError(
            "hostname %r "
            "doesn't match either of %s" % (hostname, ", ".join(map(repr, dnsnames)))
        )
    elif len(dnsnames) == 1:
        raise CertificateError("hostname %r doesn't match %r" % (hostname, dnsnames[0]))
    else:
        raise CertificateError(
            "no appropriate commonName or subjectAltName fields were found"
        )
site-packages/pip/_vendor/urllib3/packages/ssl_match_hostname/__init__.py000064400000001637150564745010022644 0ustar00import sys

try:
    # Our match_hostname function is the same as 3.10's, so we only want to
    # import the match_hostname function if it's at least that good.
    # We also fallback on Python 3.10+ because our code doesn't emit
    # deprecation warnings and is the same as Python 3.10 otherwise.
    if sys.version_info < (3, 5) or sys.version_info >= (3, 10):
        raise ImportError("Fallback to vendored code")

    from ssl import CertificateError, match_hostname
except ImportError:
    try:
        # Backport of the function from a pypi module
        from backports.ssl_match_hostname import (  # type: ignore
            CertificateError,
            match_hostname,
        )
    except ImportError:
        # Our vendored copy
        from ._implementation import CertificateError, match_hostname  # type: ignore

# Not needed, but documenting what we provide.
__all__ = ("CertificateError", "match_hostname")
site-packages/pip/_vendor/urllib3/connectionpool.py000064400000111323150564745010016477 0ustar00from __future__ import absolute_import

import errno
import logging
import socket
import sys
import warnings
from socket import error as SocketError
from socket import timeout as SocketTimeout

from .connection import (
    BaseSSLError,
    BrokenPipeError,
    DummyConnection,
    HTTPConnection,
    HTTPException,
    HTTPSConnection,
    VerifiedHTTPSConnection,
    port_by_scheme,
)
from .exceptions import (
    ClosedPoolError,
    EmptyPoolError,
    HeaderParsingError,
    HostChangedError,
    InsecureRequestWarning,
    LocationValueError,
    MaxRetryError,
    NewConnectionError,
    ProtocolError,
    ProxyError,
    ReadTimeoutError,
    SSLError,
    TimeoutError,
)
from .packages import six
from .packages.six.moves import queue
from .packages.ssl_match_hostname import CertificateError
from .request import RequestMethods
from .response import HTTPResponse
from .util.connection import is_connection_dropped
from .util.proxy import connection_requires_http_tunnel
from .util.queue import LifoQueue
from .util.request import set_file_position
from .util.response import assert_header_parsing
from .util.retry import Retry
from .util.timeout import Timeout
from .util.url import Url, _encode_target
from .util.url import _normalize_host as normalize_host
from .util.url import get_host, parse_url

xrange = six.moves.xrange

log = logging.getLogger(__name__)

_Default = object()


# Pool objects
class ConnectionPool(object):
    """
    Base class for all connection pools, such as
    :class:`.HTTPConnectionPool` and :class:`.HTTPSConnectionPool`.

    .. note::
       ConnectionPool.urlopen() does not normalize or percent-encode target URIs
       which is useful if your target server doesn't support percent-encoded
       target URIs.
    """

    scheme = None
    QueueCls = LifoQueue

    def __init__(self, host, port=None):
        if not host:
            raise LocationValueError("No host specified.")

        self.host = _normalize_host(host, scheme=self.scheme)
        self._proxy_host = host.lower()
        self.port = port

    def __str__(self):
        return "%s(host=%r, port=%r)" % (type(self).__name__, self.host, self.port)

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        self.close()
        # Return False to re-raise any potential exceptions
        return False

    def close(self):
        """
        Close all pooled connections and disable the pool.
        """
        pass


# This is taken from http://hg.python.org/cpython/file/7aaba721ebc0/Lib/socket.py#l252
_blocking_errnos = {errno.EAGAIN, errno.EWOULDBLOCK}


class HTTPConnectionPool(ConnectionPool, RequestMethods):
    """
    Thread-safe connection pool for one host.

    :param host:
        Host used for this HTTP Connection (e.g. "localhost"), passed into
        :class:`http.client.HTTPConnection`.

    :param port:
        Port used for this HTTP Connection (None is equivalent to 80), passed
        into :class:`http.client.HTTPConnection`.

    :param strict:
        Causes BadStatusLine to be raised if the status line can't be parsed
        as a valid HTTP/1.0 or 1.1 status line, passed into
        :class:`http.client.HTTPConnection`.

        .. note::
           Only works in Python 2. This parameter is ignored in Python 3.

    :param timeout:
        Socket timeout in seconds for each individual connection. This can
        be a float or integer, which sets the timeout for the HTTP request,
        or an instance of :class:`urllib3.util.Timeout` which gives you more
        fine-grained control over request timeouts. After the constructor has
        been parsed, this is always a `urllib3.util.Timeout` object.

    :param maxsize:
        Number of connections to save that can be reused. More than 1 is useful
        in multithreaded situations. If ``block`` is set to False, more
        connections will be created but they will not be saved once they've
        been used.

    :param block:
        If set to True, no more than ``maxsize`` connections will be used at
        a time. When no free connections are available, the call will block
        until a connection has been released. This is a useful side effect for
        particular multithreaded situations where one does not want to use more
        than maxsize connections per host to prevent flooding.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param retries:
        Retry configuration to use by default with requests in this pool.

    :param _proxy:
        Parsed proxy URL, should not be used directly, instead, see
        :class:`urllib3.ProxyManager`

    :param _proxy_headers:
        A dictionary with proxy headers, should not be used directly,
        instead, see :class:`urllib3.ProxyManager`

    :param \\**conn_kw:
        Additional parameters are used to create fresh :class:`urllib3.connection.HTTPConnection`,
        :class:`urllib3.connection.HTTPSConnection` instances.
    """

    scheme = "http"
    ConnectionCls = HTTPConnection
    ResponseCls = HTTPResponse

    def __init__(
        self,
        host,
        port=None,
        strict=False,
        timeout=Timeout.DEFAULT_TIMEOUT,
        maxsize=1,
        block=False,
        headers=None,
        retries=None,
        _proxy=None,
        _proxy_headers=None,
        _proxy_config=None,
        **conn_kw
    ):
        ConnectionPool.__init__(self, host, port)
        RequestMethods.__init__(self, headers)

        self.strict = strict

        if not isinstance(timeout, Timeout):
            timeout = Timeout.from_float(timeout)

        if retries is None:
            retries = Retry.DEFAULT

        self.timeout = timeout
        self.retries = retries

        self.pool = self.QueueCls(maxsize)
        self.block = block

        self.proxy = _proxy
        self.proxy_headers = _proxy_headers or {}
        self.proxy_config = _proxy_config

        # Fill the queue up so that doing get() on it will block properly
        for _ in xrange(maxsize):
            self.pool.put(None)

        # These are mostly for testing and debugging purposes.
        self.num_connections = 0
        self.num_requests = 0
        self.conn_kw = conn_kw

        if self.proxy:
            # Enable Nagle's algorithm for proxies, to avoid packet fragmentation.
            # We cannot know if the user has added default socket options, so we cannot replace the
            # list.
            self.conn_kw.setdefault("socket_options", [])

            self.conn_kw["proxy"] = self.proxy
            self.conn_kw["proxy_config"] = self.proxy_config

    def _new_conn(self):
        """
        Return a fresh :class:`HTTPConnection`.
        """
        self.num_connections += 1
        log.debug(
            "Starting new HTTP connection (%d): %s:%s",
            self.num_connections,
            self.host,
            self.port or "80",
        )

        conn = self.ConnectionCls(
            host=self.host,
            port=self.port,
            timeout=self.timeout.connect_timeout,
            strict=self.strict,
            **self.conn_kw
        )
        return conn

    def _get_conn(self, timeout=None):
        """
        Get a connection. Will return a pooled connection if one is available.

        If no connections are available and :prop:`.block` is ``False``, then a
        fresh connection is returned.

        :param timeout:
            Seconds to wait before giving up and raising
            :class:`urllib3.exceptions.EmptyPoolError` if the pool is empty and
            :prop:`.block` is ``True``.
        """
        conn = None
        try:
            conn = self.pool.get(block=self.block, timeout=timeout)

        except AttributeError:  # self.pool is None
            raise ClosedPoolError(self, "Pool is closed.")

        except queue.Empty:
            if self.block:
                raise EmptyPoolError(
                    self,
                    "Pool reached maximum size and no more connections are allowed.",
                )
            pass  # Oh well, we'll create a new connection then

        # If this is a persistent connection, check if it got disconnected
        if conn and is_connection_dropped(conn):
            log.debug("Resetting dropped connection: %s", self.host)
            conn.close()
            if getattr(conn, "auto_open", 1) == 0:
                # This is a proxied connection that has been mutated by
                # http.client._tunnel() and cannot be reused (since it would
                # attempt to bypass the proxy)
                conn = None

        return conn or self._new_conn()

    def _put_conn(self, conn):
        """
        Put a connection back into the pool.

        :param conn:
            Connection object for the current host and port as returned by
            :meth:`._new_conn` or :meth:`._get_conn`.

        If the pool is already full, the connection is closed and discarded
        because we exceeded maxsize. If connections are discarded frequently,
        then maxsize should be increased.

        If the pool is closed, then the connection will be closed and discarded.
        """
        try:
            self.pool.put(conn, block=False)
            return  # Everything is dandy, done.
        except AttributeError:
            # self.pool is None.
            pass
        except queue.Full:
            # This should never happen if self.block == True
            log.warning("Connection pool is full, discarding connection: %s", self.host)

        # Connection never got put back into the pool, close it.
        if conn:
            conn.close()

    def _validate_conn(self, conn):
        """
        Called right before a request is made, after the socket is created.
        """
        pass

    def _prepare_proxy(self, conn):
        # Nothing to do for HTTP connections.
        pass

    def _get_timeout(self, timeout):
        """Helper that always returns a :class:`urllib3.util.Timeout`"""
        if timeout is _Default:
            return self.timeout.clone()

        if isinstance(timeout, Timeout):
            return timeout.clone()
        else:
            # User passed us an int/float. This is for backwards compatibility,
            # can be removed later
            return Timeout.from_float(timeout)

    def _raise_timeout(self, err, url, timeout_value):
        """Is the error actually a timeout? Will raise a ReadTimeout or pass"""

        if isinstance(err, SocketTimeout):
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

        # See the above comment about EAGAIN in Python 3. In Python 2 we have
        # to specifically catch it and throw the timeout error
        if hasattr(err, "errno") and err.errno in _blocking_errnos:
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

        # Catch possible read timeouts thrown as SSL errors. If not the
        # case, rethrow the original. We need to do this because of:
        # http://bugs.python.org/issue10272
        if "timed out" in str(err) or "did not complete (read)" in str(
            err
        ):  # Python < 2.7.4
            raise ReadTimeoutError(
                self, url, "Read timed out. (read timeout=%s)" % timeout_value
            )

    def _make_request(
        self, conn, method, url, timeout=_Default, chunked=False, **httplib_request_kw
    ):
        """
        Perform a request on a given urllib connection object taken from our
        pool.

        :param conn:
            a connection from one of our connection pools

        :param timeout:
            Socket timeout in seconds for the request. This can be a
            float or integer, which will set the same timeout value for
            the socket connect and the socket read, or an instance of
            :class:`urllib3.util.Timeout`, which gives you more fine-grained
            control over your timeouts.
        """
        self.num_requests += 1

        timeout_obj = self._get_timeout(timeout)
        timeout_obj.start_connect()
        conn.timeout = timeout_obj.connect_timeout

        # Trigger any extra validation we need to do.
        try:
            self._validate_conn(conn)
        except (SocketTimeout, BaseSSLError) as e:
            # Py2 raises this as a BaseSSLError, Py3 raises it as socket timeout.
            self._raise_timeout(err=e, url=url, timeout_value=conn.timeout)
            raise

        # conn.request() calls http.client.*.request, not the method in
        # urllib3.request. It also calls makefile (recv) on the socket.
        try:
            if chunked:
                conn.request_chunked(method, url, **httplib_request_kw)
            else:
                conn.request(method, url, **httplib_request_kw)

        # We are swallowing BrokenPipeError (errno.EPIPE) since the server is
        # legitimately able to close the connection after sending a valid response.
        # With this behaviour, the received response is still readable.
        except BrokenPipeError:
            # Python 3
            pass
        except IOError as e:
            # Python 2 and macOS/Linux
            # EPIPE and ESHUTDOWN are BrokenPipeError on Python 2, and EPROTOTYPE is needed on macOS
            # https://erickt.github.io/blog/2014/11/19/adventures-in-debugging-a-potential-osx-kernel-bug/
            if e.errno not in {
                errno.EPIPE,
                errno.ESHUTDOWN,
                errno.EPROTOTYPE,
            }:
                raise

        # Reset the timeout for the recv() on the socket
        read_timeout = timeout_obj.read_timeout

        # App Engine doesn't have a sock attr
        if getattr(conn, "sock", None):
            # In Python 3 socket.py will catch EAGAIN and return None when you
            # try and read into the file pointer created by http.client, which
            # instead raises a BadStatusLine exception. Instead of catching
            # the exception and assuming all BadStatusLine exceptions are read
            # timeouts, check for a zero timeout before making the request.
            if read_timeout == 0:
                raise ReadTimeoutError(
                    self, url, "Read timed out. (read timeout=%s)" % read_timeout
                )
            if read_timeout is Timeout.DEFAULT_TIMEOUT:
                conn.sock.settimeout(socket.getdefaulttimeout())
            else:  # None or a value
                conn.sock.settimeout(read_timeout)

        # Receive the response from the server
        try:
            try:
                # Python 2.7, use buffering of HTTP responses
                httplib_response = conn.getresponse(buffering=True)
            except TypeError:
                # Python 3
                try:
                    httplib_response = conn.getresponse()
                except BaseException as e:
                    # Remove the TypeError from the exception chain in
                    # Python 3 (including for exceptions like SystemExit).
                    # Otherwise it looks like a bug in the code.
                    six.raise_from(e, None)
        except (SocketTimeout, BaseSSLError, SocketError) as e:
            self._raise_timeout(err=e, url=url, timeout_value=read_timeout)
            raise

        # AppEngine doesn't have a version attr.
        http_version = getattr(conn, "_http_vsn_str", "HTTP/?")
        log.debug(
            '%s://%s:%s "%s %s %s" %s %s',
            self.scheme,
            self.host,
            self.port,
            method,
            url,
            http_version,
            httplib_response.status,
            httplib_response.length,
        )

        try:
            assert_header_parsing(httplib_response.msg)
        except (HeaderParsingError, TypeError) as hpe:  # Platform-specific: Python 3
            log.warning(
                "Failed to parse headers (url=%s): %s",
                self._absolute_url(url),
                hpe,
                exc_info=True,
            )

        return httplib_response

    def _absolute_url(self, path):
        return Url(scheme=self.scheme, host=self.host, port=self.port, path=path).url

    def close(self):
        """
        Close all pooled connections and disable the pool.
        """
        if self.pool is None:
            return
        # Disable access to the pool
        old_pool, self.pool = self.pool, None

        try:
            while True:
                conn = old_pool.get(block=False)
                if conn:
                    conn.close()

        except queue.Empty:
            pass  # Done.

    def is_same_host(self, url):
        """
        Check if the given ``url`` is a member of the same host as this
        connection pool.
        """
        if url.startswith("/"):
            return True

        # TODO: Add optional support for socket.gethostbyname checking.
        scheme, host, port = get_host(url)
        if host is not None:
            host = _normalize_host(host, scheme=scheme)

        # Use explicit default port for comparison when none is given
        if self.port and not port:
            port = port_by_scheme.get(scheme)
        elif not self.port and port == port_by_scheme.get(scheme):
            port = None

        return (scheme, host, port) == (self.scheme, self.host, self.port)

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=None,
        redirect=True,
        assert_same_host=True,
        timeout=_Default,
        pool_timeout=None,
        release_conn=None,
        chunked=False,
        body_pos=None,
        **response_kw
    ):
        """
        Get a connection from the pool and perform an HTTP request. This is the
        lowest level call for making a request, so you'll need to specify all
        the raw details.

        .. note::

           More commonly, it's appropriate to use a convenience method provided
           by :class:`.RequestMethods`, such as :meth:`request`.

        .. note::

           `release_conn` will only behave as expected if
           `preload_content=False` because we want to make
           `preload_content=False` the default behaviour someday soon without
           breaking backwards compatibility.

        :param method:
            HTTP request method (such as GET, POST, PUT, etc.)

        :param url:
            The URL to perform the request on.

        :param body:
            Data to send in the request body, either :class:`str`, :class:`bytes`,
            an iterable of :class:`str`/:class:`bytes`, or a file-like object.

        :param headers:
            Dictionary of custom headers to send, such as User-Agent,
            If-None-Match, etc. If None, pool headers are used. If provided,
            these headers completely replace any pool-specific headers.

        :param retries:
            Configure the number of retries to allow before raising a
            :class:`~urllib3.exceptions.MaxRetryError` exception.

            Pass ``None`` to retry until you receive a response. Pass a
            :class:`~urllib3.util.retry.Retry` object for fine-grained control
            over different types of retries.
            Pass an integer number to retry connection errors that many times,
            but no other types of errors. Pass zero to never retry.

            If ``False``, then retries are disabled and any exception is raised
            immediately. Also, instead of raising a MaxRetryError on redirects,
            the redirect response will be returned.

        :type retries: :class:`~urllib3.util.retry.Retry`, False, or an int.

        :param redirect:
            If True, automatically handle redirects (status codes 301, 302,
            303, 307, 308). Each redirect counts as a retry. Disabling retries
            will disable redirect, too.

        :param assert_same_host:
            If ``True``, will make sure that the host of the pool requests is
            consistent else will raise HostChangedError. When ``False``, you can
            use the pool on an HTTP proxy and request foreign hosts.

        :param timeout:
            If specified, overrides the default timeout for this one
            request. It may be a float (in seconds) or an instance of
            :class:`urllib3.util.Timeout`.

        :param pool_timeout:
            If set and the pool is set to block=True, then this method will
            block for ``pool_timeout`` seconds and raise EmptyPoolError if no
            connection is available within the time period.

        :param release_conn:
            If False, then the urlopen call will not release the connection
            back into the pool once a response is received (but will release if
            you read the entire contents of the response such as when
            `preload_content=True`). This is useful if you're not preloading
            the response's content immediately. You will need to call
            ``r.release_conn()`` on the response ``r`` to return the connection
            back into the pool. If None, it takes the value of
            ``response_kw.get('preload_content', True)``.

        :param chunked:
            If True, urllib3 will send the body using chunked transfer
            encoding. Otherwise, urllib3 will send the body using the standard
            content-length form. Defaults to False.

        :param int body_pos:
            Position to seek to in file-like body in the event of a retry or
            redirect. Typically this won't need to be set because urllib3 will
            auto-populate the value when needed.

        :param \\**response_kw:
            Additional parameters are passed to
            :meth:`urllib3.response.HTTPResponse.from_httplib`
        """

        parsed_url = parse_url(url)
        destination_scheme = parsed_url.scheme

        if headers is None:
            headers = self.headers

        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect, default=self.retries)

        if release_conn is None:
            release_conn = response_kw.get("preload_content", True)

        # Check host
        if assert_same_host and not self.is_same_host(url):
            raise HostChangedError(self, url, retries)

        # Ensure that the URL we're connecting to is properly encoded
        if url.startswith("/"):
            url = six.ensure_str(_encode_target(url))
        else:
            url = six.ensure_str(parsed_url.url)

        conn = None

        # Track whether `conn` needs to be released before
        # returning/raising/recursing. Update this variable if necessary, and
        # leave `release_conn` constant throughout the function. That way, if
        # the function recurses, the original value of `release_conn` will be
        # passed down into the recursive call, and its value will be respected.
        #
        # See issue #651 [1] for details.
        #
        # [1] <https://github.com/urllib3/urllib3/issues/651>
        release_this_conn = release_conn

        http_tunnel_required = connection_requires_http_tunnel(
            self.proxy, self.proxy_config, destination_scheme
        )

        # Merge the proxy headers. Only done when not using HTTP CONNECT. We
        # have to copy the headers dict so we can safely change it without those
        # changes being reflected in anyone else's copy.
        if not http_tunnel_required:
            headers = headers.copy()
            headers.update(self.proxy_headers)

        # Must keep the exception bound to a separate variable or else Python 3
        # complains about UnboundLocalError.
        err = None

        # Keep track of whether we cleanly exited the except block. This
        # ensures we do proper cleanup in finally.
        clean_exit = False

        # Rewind body position, if needed. Record current position
        # for future rewinds in the event of a redirect/retry.
        body_pos = set_file_position(body, body_pos)

        try:
            # Request a connection from the queue.
            timeout_obj = self._get_timeout(timeout)
            conn = self._get_conn(timeout=pool_timeout)

            conn.timeout = timeout_obj.connect_timeout

            is_new_proxy_conn = self.proxy is not None and not getattr(
                conn, "sock", None
            )
            if is_new_proxy_conn and http_tunnel_required:
                self._prepare_proxy(conn)

            # Make the request on the httplib connection object.
            httplib_response = self._make_request(
                conn,
                method,
                url,
                timeout=timeout_obj,
                body=body,
                headers=headers,
                chunked=chunked,
            )

            # If we're going to release the connection in ``finally:``, then
            # the response doesn't need to know about the connection. Otherwise
            # it will also try to release it and we'll have a double-release
            # mess.
            response_conn = conn if not release_conn else None

            # Pass method to Response for length checking
            response_kw["request_method"] = method

            # Import httplib's response into our own wrapper object
            response = self.ResponseCls.from_httplib(
                httplib_response,
                pool=self,
                connection=response_conn,
                retries=retries,
                **response_kw
            )

            # Everything went great!
            clean_exit = True

        except EmptyPoolError:
            # Didn't get a connection from the pool, no need to clean up
            clean_exit = True
            release_this_conn = False
            raise

        except (
            TimeoutError,
            HTTPException,
            SocketError,
            ProtocolError,
            BaseSSLError,
            SSLError,
            CertificateError,
        ) as e:
            # Discard the connection for these exceptions. It will be
            # replaced during the next _get_conn() call.
            clean_exit = False
            if isinstance(e, (BaseSSLError, CertificateError)):
                e = SSLError(e)
            elif isinstance(e, (SocketError, NewConnectionError)) and self.proxy:
                e = ProxyError("Cannot connect to proxy.", e)
            elif isinstance(e, (SocketError, HTTPException)):
                e = ProtocolError("Connection aborted.", e)

            retries = retries.increment(
                method, url, error=e, _pool=self, _stacktrace=sys.exc_info()[2]
            )
            retries.sleep()

            # Keep track of the error for the retry warning.
            err = e

        finally:
            if not clean_exit:
                # We hit some kind of exception, handled or otherwise. We need
                # to throw the connection away unless explicitly told not to.
                # Close the connection, set the variable to None, and make sure
                # we put the None back in the pool to avoid leaking it.
                conn = conn and conn.close()
                release_this_conn = True

            if release_this_conn:
                # Put the connection back to be reused. If the connection is
                # expired then it will be None, which will get replaced with a
                # fresh connection during _get_conn.
                self._put_conn(conn)

        if not conn:
            # Try again
            log.warning(
                "Retrying (%r) after connection broken by '%r': %s", retries, err, url
            )
            return self.urlopen(
                method,
                url,
                body,
                headers,
                retries,
                redirect,
                assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                chunked=chunked,
                body_pos=body_pos,
                **response_kw
            )

        # Handle redirect?
        redirect_location = redirect and response.get_redirect_location()
        if redirect_location:
            if response.status == 303:
                method = "GET"

            try:
                retries = retries.increment(method, url, response=response, _pool=self)
            except MaxRetryError:
                if retries.raise_on_redirect:
                    response.drain_conn()
                    raise
                return response

            response.drain_conn()
            retries.sleep_for_retry(response)
            log.debug("Redirecting %s -> %s", url, redirect_location)
            return self.urlopen(
                method,
                redirect_location,
                body,
                headers,
                retries=retries,
                redirect=redirect,
                assert_same_host=assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                chunked=chunked,
                body_pos=body_pos,
                **response_kw
            )

        # Check if we should retry the HTTP response.
        has_retry_after = bool(response.getheader("Retry-After"))
        if retries.is_retry(method, response.status, has_retry_after):
            try:
                retries = retries.increment(method, url, response=response, _pool=self)
            except MaxRetryError:
                if retries.raise_on_status:
                    response.drain_conn()
                    raise
                return response

            response.drain_conn()
            retries.sleep(response)
            log.debug("Retry: %s", url)
            return self.urlopen(
                method,
                url,
                body,
                headers,
                retries=retries,
                redirect=redirect,
                assert_same_host=assert_same_host,
                timeout=timeout,
                pool_timeout=pool_timeout,
                release_conn=release_conn,
                chunked=chunked,
                body_pos=body_pos,
                **response_kw
            )

        return response


class HTTPSConnectionPool(HTTPConnectionPool):
    """
    Same as :class:`.HTTPConnectionPool`, but HTTPS.

    :class:`.HTTPSConnection` uses one of ``assert_fingerprint``,
    ``assert_hostname`` and ``host`` in this order to verify connections.
    If ``assert_hostname`` is False, no verification is done.

    The ``key_file``, ``cert_file``, ``cert_reqs``, ``ca_certs``,
    ``ca_cert_dir``, ``ssl_version``, ``key_password`` are only used if :mod:`ssl`
    is available and are fed into :meth:`urllib3.util.ssl_wrap_socket` to upgrade
    the connection socket into an SSL socket.
    """

    scheme = "https"
    ConnectionCls = HTTPSConnection

    def __init__(
        self,
        host,
        port=None,
        strict=False,
        timeout=Timeout.DEFAULT_TIMEOUT,
        maxsize=1,
        block=False,
        headers=None,
        retries=None,
        _proxy=None,
        _proxy_headers=None,
        key_file=None,
        cert_file=None,
        cert_reqs=None,
        key_password=None,
        ca_certs=None,
        ssl_version=None,
        assert_hostname=None,
        assert_fingerprint=None,
        ca_cert_dir=None,
        **conn_kw
    ):

        HTTPConnectionPool.__init__(
            self,
            host,
            port,
            strict,
            timeout,
            maxsize,
            block,
            headers,
            retries,
            _proxy,
            _proxy_headers,
            **conn_kw
        )

        self.key_file = key_file
        self.cert_file = cert_file
        self.cert_reqs = cert_reqs
        self.key_password = key_password
        self.ca_certs = ca_certs
        self.ca_cert_dir = ca_cert_dir
        self.ssl_version = ssl_version
        self.assert_hostname = assert_hostname
        self.assert_fingerprint = assert_fingerprint

    def _prepare_conn(self, conn):
        """
        Prepare the ``connection`` for :meth:`urllib3.util.ssl_wrap_socket`
        and establish the tunnel if proxy is used.
        """

        if isinstance(conn, VerifiedHTTPSConnection):
            conn.set_cert(
                key_file=self.key_file,
                key_password=self.key_password,
                cert_file=self.cert_file,
                cert_reqs=self.cert_reqs,
                ca_certs=self.ca_certs,
                ca_cert_dir=self.ca_cert_dir,
                assert_hostname=self.assert_hostname,
                assert_fingerprint=self.assert_fingerprint,
            )
            conn.ssl_version = self.ssl_version
        return conn

    def _prepare_proxy(self, conn):
        """
        Establishes a tunnel connection through HTTP CONNECT.

        Tunnel connection is established early because otherwise httplib would
        improperly set Host: header to proxy's IP:port.
        """

        conn.set_tunnel(self._proxy_host, self.port, self.proxy_headers)

        if self.proxy.scheme == "https":
            conn.tls_in_tls_required = True

        conn.connect()

    def _new_conn(self):
        """
        Return a fresh :class:`http.client.HTTPSConnection`.
        """
        self.num_connections += 1
        log.debug(
            "Starting new HTTPS connection (%d): %s:%s",
            self.num_connections,
            self.host,
            self.port or "443",
        )

        if not self.ConnectionCls or self.ConnectionCls is DummyConnection:
            raise SSLError(
                "Can't connect to HTTPS URL because the SSL module is not available."
            )

        actual_host = self.host
        actual_port = self.port
        if self.proxy is not None:
            actual_host = self.proxy.host
            actual_port = self.proxy.port

        conn = self.ConnectionCls(
            host=actual_host,
            port=actual_port,
            timeout=self.timeout.connect_timeout,
            strict=self.strict,
            cert_file=self.cert_file,
            key_file=self.key_file,
            key_password=self.key_password,
            **self.conn_kw
        )

        return self._prepare_conn(conn)

    def _validate_conn(self, conn):
        """
        Called right before a request is made, after the socket is created.
        """
        super(HTTPSConnectionPool, self)._validate_conn(conn)

        # Force connect early to allow us to validate the connection.
        if not getattr(conn, "sock", None):  # AppEngine might not have  `.sock`
            conn.connect()

        if not conn.is_verified:
            warnings.warn(
                (
                    "Unverified HTTPS request is being made to host '%s'. "
                    "Adding certificate verification is strongly advised. See: "
                    "https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html"
                    "#ssl-warnings" % conn.host
                ),
                InsecureRequestWarning,
            )

        if getattr(conn, "proxy_is_verified", None) is False:
            warnings.warn(
                (
                    "Unverified HTTPS connection done to an HTTPS proxy. "
                    "Adding certificate verification is strongly advised. See: "
                    "https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html"
                    "#ssl-warnings"
                ),
                InsecureRequestWarning,
            )


def connection_from_url(url, **kw):
    """
    Given a url, return an :class:`.ConnectionPool` instance of its host.

    This is a shortcut for not having to parse out the scheme, host, and port
    of the url before creating an :class:`.ConnectionPool` instance.

    :param url:
        Absolute URL string that must include the scheme. Port is optional.

    :param \\**kw:
        Passes additional parameters to the constructor of the appropriate
        :class:`.ConnectionPool`. Useful for specifying things like
        timeout, maxsize, headers, etc.

    Example::

        >>> conn = connection_from_url('http://google.com/')
        >>> r = conn.request('GET', '/')
    """
    scheme, host, port = get_host(url)
    port = port or port_by_scheme.get(scheme, 80)
    if scheme == "https":
        return HTTPSConnectionPool(host, port=port, **kw)
    else:
        return HTTPConnectionPool(host, port=port, **kw)


def _normalize_host(host, scheme):
    """
    Normalize hosts for comparisons and use with sockets.
    """

    host = normalize_host(host, scheme)

    # httplib doesn't like it when we include brackets in IPv6 addresses
    # Specifically, if we include brackets but also pass the port then
    # httplib crazily doubles up the square brackets on the Host header.
    # Instead, we need to make sure we never pass ``None`` as the port.
    # However, for backward compatibility reasons we can't actually
    # *assert* that.  See http://bugs.python.org/issue28539
    if host.startswith("[") and host.endswith("]"):
        host = host[1:-1]
    return host
site-packages/pip/_vendor/urllib3/__pycache__/request.cpython-39.pyc000064400000013040150564745010021422 0ustar00a

��Rea�@s>ddlmZddlmZddlmZdgZGdd�de�ZdS)�)�absolute_import�)�encode_multipart_formdata)�	urlencode�RequestMethodsc@sJeZdZdZhd�Zddd�Zddd�Zdd	d
�Zddd�Zdd
d�Z	dS)ra�
    Convenience mixin for classes who implement a :meth:`urlopen` method, such
    as :class:`urllib3.HTTPConnectionPool` and
    :class:`urllib3.PoolManager`.

    Provides behavior for making common types of HTTP request methods and
    decides which type of request field encoding to use.

    Specifically,

    :meth:`.request_encode_url` is for sending requests whose fields are
    encoded in the URL (such as GET, HEAD, DELETE).

    :meth:`.request_encode_body` is for sending requests whose fields are
    encoded in the *body* of the request using multipart or www-form-urlencoded
    (such as for POST, PUT, PATCH).

    :meth:`.request` is for making any kind of request, it will look up the
    appropriate encoding format and use one of the above two methods to make
    the request.

    Initializer parameters:

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.
    >�DELETE�HEAD�OPTIONS�GETNcCs|pi|_dS)N)�headers)�selfr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/request.py�__init__(szRequestMethods.__init__TcKstd��dS)NzMClasses extending RequestMethods must implement their own ``urlopen`` method.)�NotImplementedError)r�method�url�bodyr�encode_multipart�multipart_boundary�kwr
r
r�urlopen+s
�zRequestMethods.urlopencKsR|��}||d<||jvr4|j||f||d�|��S|j||f||d�|��SdS)a�
        Make a request using :meth:`urlopen` with the appropriate encoding of
        ``fields`` based on the ``method`` used.

        This is a convenience method that requires the least amount of manual
        effort. It can be used in most situations, while still having the
        option to drop down to more specific methods when necessary, such as
        :meth:`request_encode_url`, :meth:`request_encode_body`,
        or even the lowest level :meth:`urlopen`.
        �request_url)�fieldsrN)�upper�_encode_url_methods�request_encode_url�request_encode_body)rrrrr�
urlopen_kwr
r
r�request:s"
������zRequestMethods.requestcKsH|dur|j}d|i}|�|�|r4|dt|�7}|j||fi|��S)z�
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the url. This is useful for request methods like GET, HEAD, DELETE, etc.
        Nr�?)r�updaterr)rrrrrr�extra_kwr
r
rrRs
z!RequestMethods.request_encode_urlcKs�|dur|j}dii}|rbd|vr*td��|r@t||d�\}	}
nt|�d}	}
|	|d<d|
i|d<|d�|�|�|�|j||fi|��S)a�
        Make a request using :meth:`urlopen` with the ``fields`` encoded in
        the body. This is useful for request methods like POST, PUT, PATCH, etc.

        When ``encode_multipart=True`` (default), then
        :func:`urllib3.encode_multipart_formdata` is used to encode
        the payload with the appropriate content type. Otherwise
        :func:`urllib.parse.urlencode` is used with the
        'application/x-www-form-urlencoded' content type.

        Multipart encoding must be used when posting files, and it's reasonably
        safe to use it in other times too. However, it may break request
        signing, such as with OAuth.

        Supports an optional ``fields`` parameter of key/value strings AND
        key/filetuple. A filetuple is a (filename, data, MIME type) tuple where
        the MIME type is optional. For example::

            fields = {
                'foo': 'bar',
                'fakefile': ('foofile.txt', 'contents of foofile'),
                'realfile': ('barfile.txt', open('realfile').read()),
                'typedfile': ('bazfile.bin', open('bazfile').read(),
                              'image/jpeg'),
                'nonamefile': 'contents of nonamefile field',
            }

        When uploading a file, providing a filename (the first parameter of the
        tuple) is optional but recommended to best mimic behavior of browsers.

        Note that if ``headers`` are supplied, the 'Content-Type' header will
        be overwritten because it depends on the dynamic random boundary string
        which is used to compose the body of the request. The random boundary
        string can be explicitly set with the ``multipart_boundary`` parameter.
        NrrzFrequest got values for both 'fields' and 'body', can only specify one.)�boundaryz!application/x-www-form-urlencodedzContent-Type)r�	TypeErrorrrr!r)rrrrrrrrr"r�content_typer
r
rrbs(-���
z"RequestMethods.request_encode_body)N)NNTN)NN)NN)NNTN)
�__name__�
__module__�__qualname__�__doc__rrrrrrr
r
r
rr	s
�


�N)	�
__future__r�filepostrZpackages.six.moves.urllib.parser�__all__�objectrr
r
r
r�<module>ssite-packages/pip/_vendor/urllib3/__pycache__/fields.cpython-39.pyc000064400000020010150564745010021173 0ustar00a

��Re�!�@s�ddlmZddlZddlZddlZddlmZddd�Zdd	�Z	d
dd�Z
e
�d
d�edd�D��dd�Z
dd�ZeZGdd�de�ZdS)�)�absolute_importN�)�six�application/octet-streamcCs|rt�|�dp|S|S)z�
    Guess the "Content-Type" of a file.

    :param filename:
        The filename to guess the "Content-Type" of using :mod:`mimetypes`.
    :param default:
        If no "Content-Type" can be guessed, default to `default`.
    r)�	mimetypes�
guess_type)�filename�default�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/fields.py�guess_content_type
s	rc	s�t�tj�r��d��t�fdd�dD��sbd|�f}z|�d�Wnttfy\Yn0|Stjrr��d��t	j
��d��d|�f�tjr���d���S)a!
    Helper function to format and quote a single header parameter using the
    strategy defined in RFC 2231.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows
    `RFC 2388 Section 4.4 <https://tools.ietf.org/html/rfc2388#section-4.4>`_.

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        An RFC-2231-formatted unicode string.
    �utf-8c3s|]}|�vVqdS�Nr
)�.0�ch��valuer
r�	<genexpr>+�z.format_header_param_rfc2231.<locals>.<genexpr>z"\
�%s="%s"�asciiz%s*=%s)�
isinstancer�binary_type�decode�any�encode�UnicodeEncodeError�UnicodeDecodeError�PY2�email�utils�encode_rfc2231)�namer�resultr
rr�format_header_param_rfc2231s 


r$z%22z\\)�"�\cCs&i|]}|dvrt�|�d�|��qS))�z%{:02X})r�unichr�format)r�ccr
r
r�
<dictcomp>Js�r+� cs:�fdd�}t�d�dd����D���}|�||�}|S)Ncs�|�d�S)Nr)�group)�match��needles_and_replacementsr
r�replacerSsz#_replace_multiple.<locals>.replacer�|cSsg|]}t�|��qSr
)�re�escape)r�needler
r
r�
<listcomp>Wrz%_replace_multiple.<locals>.<listcomp>)r3�compile�join�keys�sub)rr0r1�patternr#r
r/r�_replace_multipleRs�r<cCs,t|tj�r|�d�}t|t�}d||fS)a�
    Helper function to format and quote a single header parameter using the
    HTML5 strategy.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows the `HTML5 Working Draft
    Section 4.10.22.7`_ and matches the behavior of curl and modern browsers.

    .. _HTML5 Working Draft Section 4.10.22.7:
        https://w3c.github.io/html/sec-forms.html#multipart-form-data

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        A unicode string, stripped of troublesome characters.
    r
r)rrrrr<�_HTML5_REPLACEMENTS)r"rr
r
r�format_header_param_html5_s

r>c@sReZdZdZddefdd�Zeefdd��Zdd�Zd	d
�Z	dd�Z
dd
d�ZdS)�RequestFielda
    A data container for request body parameters.

    :param name:
        The name of this request field. Must be unicode.
    :param data:
        The data/value body.
    :param filename:
        An optional filename of the request field. Must be unicode.
    :param headers:
        An optional dict-like object of headers to initially use for the field.
    :param header_formatter:
        An optional callable that is used to encode and format the headers. By
        default, this is :func:`format_header_param_html5`.
    NcCs0||_||_||_i|_|r&t|�|_||_dSr)�_name�	_filename�data�headers�dict�header_formatter)�selfr"rBrrCrEr
r
r�__init__�s
zRequestField.__init__cCs`t|t�r4t|�dkr"|\}}}q@|\}}t|�}nd}d}|}|||||d�}|j|d�|S)a�
        A :class:`~urllib3.fields.RequestField` factory from old-style tuple parameters.

        Supports constructing :class:`~urllib3.fields.RequestField` from
        parameter of key/value strings AND key/filetuple. A filetuple is a
        (filename, data, MIME type) tuple where the MIME type is optional.
        For example::

            'foo': 'bar',
            'fakefile': ('foofile.txt', 'contents of foofile'),
            'realfile': ('barfile.txt', open('realfile').read()),
            'typedfile': ('bazfile.bin', open('bazfile').read(), 'image/jpeg'),
            'nonamefile': 'contents of nonamefile field',

        Field names and filenames must be unicode.
        �N)rrE)�content_type)r�tuple�lenr�make_multipart)�cls�	fieldnamerrErrBrI�
request_paramr
r
r�from_tuples�s

�zRequestField.from_tuplescCs|�||�S)aI
        Overridable helper function to format a single header parameter. By
        default, this calls ``self.header_formatter``.

        :param name:
            The name of the parameter, a string expected to be ASCII only.
        :param value:
            The value of the parameter, provided as a unicode string.
        )rE)rFr"rr
r
r�_render_part�szRequestField._render_partcCsLg}|}t|t�r|��}|D]"\}}|dur|�|�||��qd�|�S)aO
        Helper function to format and quote a single header.

        Useful for single headers that are composed of multiple items. E.g.,
        'Content-Disposition' fields.

        :param header_parts:
            A sequence of (k, v) tuples or a :class:`dict` of (k, v) to format
            as `k1="v1"; k2="v2"; ...`.
        N�; )rrD�items�appendrQr8)rF�header_parts�parts�iterabler"rr
r
r�
_render_parts�s
zRequestField._render_partscCs�g}gd�}|D]*}|j�|d�r|�d||j|f�q|j��D]&\}}||vrF|rF|�d||f�qF|�d�d�|�S)z=
        Renders the headers for this request field.
        )�Content-Disposition�Content-Type�Content-LocationFz%s: %sz
)rC�getrTrSr8)rF�lines�	sort_keys�sort_key�header_name�header_valuer
r
r�render_headers�s
zRequestField.render_headerscCsX|pd|jd<|jdd�d|�d|jfd|jff�g�7<||jd<||jd<d	S)
a|
        Makes this request field into a multipart request field.

        This method overrides "Content-Disposition", "Content-Type" and
        "Content-Location" headers to the request parameter.

        :param content_type:
            The 'Content-Type' of the request body.
        :param content_location:
            The 'Content-Location' of the request body.

        z	form-datarYrR�r"rrZr[N)rCr8rXr@rA)rF�content_dispositionrI�content_locationr
r
rrL�s���
zRequestField.make_multipart)NNN)�__name__�
__module__�__qualname__�__doc__r>rG�classmethodrPrQrXrbrLr
r
r
rr?~s�
#
�r?)r)�
__future__r�email.utilsrrr3�packagesrrr$r=�update�ranger<r>�format_header_param�objectr?r
r
r
r�<module>s$
+���	
site-packages/pip/_vendor/urllib3/__pycache__/filepost.cpython-39.pyc000064400000005361150564745010021566 0ustar00a

��Re�	�@s�ddlmZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe�
d�d	Zd
d�Zdd
�Zdd�Zddd�ZdS)�)�absolute_importN)�BytesIO�)�RequestField)�six)�bzutf-8�cCs$t�t�d��}tjs |�d�}|S)zN
    Our embarrassingly-simple replacement for mimetools.choose_boundary.
    ��ascii)�binascii�hexlify�os�urandomr�PY2�decode)�boundary�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/filepost.py�choose_boundarys
rccsJt|t�rt�|�}nt|�}|D]"}t|t�r8|Vq"tj|�Vq"dS)z�
    Iterate over fields.

    Supports list of (k, v) tuples and dicts, and lists of
    :class:`~urllib3.fields.RequestField`.

    N)�
isinstance�dictr�	iteritems�iterr�from_tuples)�fields�i�fieldrrr�iter_field_objectss

rcCs,t|t�rdd�t�|�D�Sdd�|D�S)a-
    .. deprecated:: 1.6

    Iterate over fields.

    The addition of :class:`~urllib3.fields.RequestField` makes this function
    obsolete. Instead, use :func:`iter_field_objects`, which returns
    :class:`~urllib3.fields.RequestField` objects.

    Supports list of (k, v) tuples and dicts.
    css|]\}}||fVqdS�Nr��.0�k�vrrr�	<genexpr>:�ziter_fields.<locals>.<genexpr>css|]\}}||fVqdSrrrrrrr#<r$)rrrr)rrrr�iter_fields-s
r%cCs�t�}|durt�}t|�D]p}|�td|��t|��|���|j}t|t	�r\t
|�}t|tj�rxt|��|�n
|�|�|�d�q|�td|��t
d|�}|�
�|fS)aQ
    Encode a dictionary of ``fields`` using the multipart/form-data MIME format.

    :param fields:
        Dictionary of fields or list of (key, :class:`~urllib3.fields.RequestField`).

    :param boundary:
        If not specified, then a random boundary will be generated using
        :func:`urllib3.filepost.choose_boundary`.
    Nz--%s
s
z--%s--
z multipart/form-data; boundary=%s)rrr�writer�writer�render_headers�datar�int�strr�	text_type�getvalue)rr�bodyrr)�content_typerrr�encode_multipart_formdata?s 

r0)N)�
__future__rr�codecsr
�iorrr�packagesrZpackages.sixr�lookupr'rrr%r0rrrr�<module>s
site-packages/pip/_vendor/urllib3/__pycache__/exceptions.cpython-39.pyc000064400000026645150564745010022132 0ustar00a

��Re �@s~ddlmZddlmZGdd�de�ZGdd�de�ZGdd	�d	e�Z	Gd
d�de	�Z
Gdd
�d
e�ZGdd�de�ZGdd�de�Z
Gdd�de�ZeZGdd�de
�ZGdd�de
�ZGdd�de�ZGdd�de�ZGdd�dee
�ZGdd�de�ZGd d!�d!ee	�ZGd"d#�d#e	�ZGd$d%�d%e	�ZGd&d'�d'ee�ZGd(d)�d)e�ZGd*d+�d+e�ZGd,d-�d-e�ZGd.d/�d/e�ZGd0d1�d1e�ZGd2d3�d3e�Z Gd4d5�d5e�Z!Gd6d7�d7e�Z"Gd8d9�d9e�Z#Gd:d;�d;e�Z$Gd<d=�d=ee�Z%Gd>d?�d?e�Z&Gd@dA�dAee�ZGdBdC�dCee�Z'GdDdE�dEe�Z(GdFdG�dGe)e�Z*GdHdI�dIe�Z+GdJdK�dKe�Z,GdLdM�dMe�Z-dNS)O�)�absolute_import�)�IncompleteReadc@seZdZdZdS)�	HTTPErrorz#Base exception used by this module.N��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/exceptions.pyrsrc@seZdZdZdS)�HTTPWarningz!Base warning used by this module.Nrrrrrr
sr
c@s eZdZdZdd�Zdd�ZdS)�	PoolErrorz/Base exception for errors caused within a pool.cCs||_t�|d||f�dS)Nz%s: %s)�poolr�__init__)�selfr�messagerrrrszPoolError.__init__cCs
|jdfS)N)NN��	__class__�rrrr�
__reduce__szPoolError.__reduce__N�rrr	r
rrrrrrrsrc@s eZdZdZdd�Zdd�ZdS)�RequestErrorz8Base exception for PoolErrors that have associated URLs.cCs||_t�|||�dS�N)�urlrr)rrrrrrrr#szRequestError.__init__cCs|jd|jdffSr)rrrrrrr'szRequestError.__reduce__Nrrrrrr src@seZdZdZdS)�SSLErrorz9Raised when SSL certificate fails in an HTTPS connection.Nrrrrrr,srcs eZdZdZ�fdd�Z�ZS)�
ProxyErrorz,Raised when the connection to a proxy fails.cs$tt|�j||g|�R�||_dSr)�superrr�original_error)rr�error�argsrrrr5szProxyError.__init__�rrr	r
r�
__classcell__rrrrr2src@seZdZdZdS)�DecodeErrorz;Raised when automatic decoding based on Content-Type fails.Nrrrrrr#:sr#c@seZdZdZdS)�
ProtocolErrorz>Raised when something unexpected happens mid-request/response.Nrrrrrr$@sr$c@seZdZdZddd�ZdS)�
MaxRetryErroraRaised when the maximum number of retries is exceeded.

    :param pool: The connection pool
    :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool`
    :param string url: The requested Url
    :param exceptions.Exception reason: The underlying error

    NcCs&||_d||f}t�||||�dS)Nz0Max retries exceeded with url: %s (Caused by %r))�reasonrr)rrrr&rrrrrWszMaxRetryError.__init__)N�rrr	r
rrrrrr%Ms	r%c@seZdZdZddd�ZdS)�HostChangedErrorz?Raised when an existing pool gets a request for a foreign host.�cCs"d|}t�||||�||_dS)Nz)Tried to open a foreign host with url: %s)rr�retries)rrrr*rrrrrbszHostChangedError.__init__N)r)r'rrrrr(_sr(c@seZdZdZdS)�TimeoutStateErrorz1Raised when passing an invalid state to a timeoutNrrrrrr+hsr+c@seZdZdZdS)�TimeoutErrorz�Raised when a socket timeout error occurs.

    Catching this error will catch both :exc:`ReadTimeoutErrors
    <ReadTimeoutError>` and :exc:`ConnectTimeoutErrors <ConnectTimeoutError>`.
    Nrrrrrr,nsr,c@seZdZdZdS)�ReadTimeoutErrorzFRaised when a socket timeout occurs while receiving data from a serverNrrrrrr-xsr-c@seZdZdZdS)�ConnectTimeoutErrorz@Raised when a socket timeout occurs while connecting to a serverNrrrrrr.�sr.c@seZdZdZdS)�NewConnectionErrorzHRaised when we fail to establish a new connection. Usually ECONNREFUSED.Nrrrrrr/�sr/c@seZdZdZdS)�EmptyPoolErrorzCRaised when a pool runs out of connections and no more are allowed.Nrrrrrr0�sr0c@seZdZdZdS)�ClosedPoolErrorzCRaised when a request enters a pool after the pool has been closed.Nrrrrrr1�sr1c@seZdZdZdS)�LocationValueErrorz<Raised when there is something wrong with a given URL input.Nrrrrrr2�sr2c@seZdZdZdd�ZdS)�LocationParseErrorz=Raised when get_host or similar fails to parse the URL input.cCsd|}t�||�||_dS)NzFailed to parse: %s)rr�location)rr4rrrrr�szLocationParseError.__init__Nr'rrrrr3�sr3cs eZdZdZ�fdd�Z�ZS)�URLSchemeUnknownz2Raised when a URL input has an unsupported scheme.cs"d|}tt|��|�||_dS)NzNot supported URL scheme %s)rr5r�scheme�rr6rrrrr�szURLSchemeUnknown.__init__r!rrrrr5�sr5c@seZdZdZdZdZdS)�
ResponseErrorzDUsed as a container for an error reason supplied in a MaxRetryError.ztoo many error responsesz&too many {status_code} error responsesN)rrr	r
�
GENERIC_ERROR�SPECIFIC_ERRORrrrrr8�sr8c@seZdZdZdS)�SecurityWarningz0Warned when performing security reducing actionsNrrrrrr;�sr;c@seZdZdZdS)�SubjectAltNameWarningzBWarned when connecting to a host with a certificate missing a SAN.Nrrrrrr<�sr<c@seZdZdZdS)�InsecureRequestWarningz/Warned when making an unverified HTTPS request.Nrrrrrr=�sr=c@seZdZdZdS)�SystemTimeWarningz0Warned when system time is suspected to be wrongNrrrrrr>�sr>c@seZdZdZdS)�InsecurePlatformWarningzIWarned when certain TLS/SSL configuration is not available on a platform.Nrrrrrr?�sr?c@seZdZdZdS)�SNIMissingWarningz9Warned when making a HTTPS request without SNI available.Nrrrrrr@�sr@c@seZdZdZdS)�DependencyWarningzc
    Warned when an attempt is made to import a module with missing optional
    dependencies.
    NrrrrrrA�srAc@seZdZdZdS)�ResponseNotChunkedz;Response needs to be chunked in order to read it as chunks.NrrrrrrB�srBc@seZdZdZdS)�BodyNotHttplibCompatiblez�
    Body should be :class:`http.client.HTTPResponse` like
    (have an fp attribute which returns raw chunks) for read_chunked().
    NrrrrrrC�srCcs(eZdZdZ�fdd�Zdd�Z�ZS)rz�
    Response length doesn't match expected Content-Length

    Subclass of :class:`http.client.IncompleteRead` to allow int value
    for ``partial`` to avoid creating large objects on streamed reads.
    cstt|��||�dSr)rrr)r�partial�expectedrrrr�szIncompleteRead.__init__cCsd|j|jfS)Nz/IncompleteRead(%i bytes read, %i more expected))rDrErrrr�__repr__s�zIncompleteRead.__repr__�rrr	r
rrFr"rrrrr�srcs(eZdZdZ�fdd�Zdd�Z�ZS)�InvalidChunkLengthz+Invalid chunk length in a chunked response.cs(tt|��|��|j�||_||_dSr)rrHr�tell�length_remaining�response�length)rrKrLrrrr
s


�zInvalidChunkLength.__init__cCsd|j|jfS)Nz0InvalidChunkLength(got length %r, %i bytes read))rLrDrrrrrFs�zInvalidChunkLength.__repr__rGrrrrrHsrHc@seZdZdZdS)�
InvalidHeaderz(The header provided was somehow invalid.NrrrrrrMsrMcs eZdZdZ�fdd�Z�ZS)�ProxySchemeUnknownz1ProxyManager does not support the supplied schemecs6|dkrd}|durd}nd|}tt|��|�dS)N�	localhostz>Proxy URL had no scheme, should start with http:// or https://zCProxy URL had unsupported scheme %s, should use http:// or https://)rrNrr7rrrr#s��zProxySchemeUnknown.__init__r!rrrrrNsrNc@seZdZdZdS)�ProxySchemeUnsupportedz=Fetching HTTPS resources through HTTPS proxies is unsupportedNrrrrrrP2srPcs eZdZdZ�fdd�Z�ZS)�HeaderParsingErrorzNRaised by assert_header_parsing, but we convert it to a log.warning statement.cs$d|pd|f}tt|��|�dS)Nz%s, unparsed data: %r�Unknown)rrQr)r�defects�
unparsed_datarrrrr;szHeaderParsingError.__init__r!rrrrrQ8srQc@seZdZdZdS)�UnrewindableBodyErrorz9urllib3 encountered an error when trying to rewind a bodyNrrrrrrU@srUN).�
__future__rZpackages.six.moves.http_clientr�httplib_IncompleteRead�	Exceptionr�Warningr
rrrrr#r$�ConnectionErrorr%r(r+r,r-r.r/r0r1�
ValueErrorr2r3r5r8r;r<r=r>r?r@rArBrCrHrM�AssertionErrorrNrPrQrUrrrr�<module>sN	


		site-packages/pip/_vendor/urllib3/__pycache__/_collections.cpython-39.pyc000064400000025107150564745010022416 0ustar00a

��Re;*�@s�ddlmZzddlmZmZWn"eyBddlmZmZYn0zddlmZWn eytGdd�d�ZYn0ddlm	Z	ddl
mZdd	lm
Z
dd
lmZmZddgZe�ZGd
d�de�ZGdd�de�ZdS)�)�absolute_import)�Mapping�MutableMapping)�RLockc@seZdZdd�Zdd�ZdS)rcCsdS�N���selfrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/_collections.py�	__enter__szRLock.__enter__cCsdSrr)r	�exc_type�	exc_value�	tracebackrrr
�__exit__szRLock.__exit__N)�__name__�
__module__�__qualname__rrrrrr
rsr)�OrderedDict�)�
InvalidHeader)�six)�iterkeys�
itervalues�RecentlyUsedContainer�HTTPHeaderDictc@sVeZdZdZeZddd�Zdd�Zdd	�Zd
d�Z	dd
�Z
dd�Zdd�Zdd�Z
dS)ra�
    Provides a thread-safe dict-like container which maintains up to
    ``maxsize`` keys while throwing away the least-recently-used keys beyond
    ``maxsize``.

    :param maxsize:
        Maximum number of recent elements to retain.

    :param dispose_func:
        Every time an item is evicted from the container,
        ``dispose_func(value)`` is called.  Callback which will get called
    �
NcCs"||_||_|��|_t�|_dSr)�_maxsize�dispose_func�ContainerCls�
_containerr�lock)r	�maxsizerrrr
�__init__/s
zRecentlyUsedContainer.__init__cCsD|j�*|j�|�}||j|<|Wd�S1s60YdSr)r r�pop)r	�key�itemrrr
�__getitem__6s
z!RecentlyUsedContainer.__getitem__cCs�t}|j�J|j�|t�}||j|<t|j�|jkrF|jjdd�\}}Wd�n1sZ0Y|jr||tur||�|�dS)NF)�last)�_Nullr r�get�lenr�popitemr)r	r$�value�
evicted_value�_keyrrr
�__setitem__=s
0z!RecentlyUsedContainer.__setitem__cCsF|j�|j�|�}Wd�n1s(0Y|jrB|�|�dSr)r rr#r)r	r$r,rrr
�__delitem__Ls*z!RecentlyUsedContainer.__delitem__cCs4|j�t|j�Wd�S1s&0YdSr)r r*rrrrr
�__len__SszRecentlyUsedContainer.__len__cCstd��dS)Nz7Iteration over this class is unlikely to be threadsafe.)�NotImplementedErrorrrrr
�__iter__Ws�zRecentlyUsedContainer.__iter__cCs\|j�(tt|j��}|j��Wd�n1s40Y|jrX|D]}|�|�qHdSr)r �listrr�clearr)r	�valuesr,rrr
r5\s(zRecentlyUsedContainer.clearcCs8|j�tt|j��Wd�S1s*0YdSr)r r4rrrrrr
�keysfszRecentlyUsedContainer.keys)rN)rrr�__doc__rrr"r&r/r0r1r3r5r7rrrr
rs


cs�eZdZdZd-�fdd�	Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
ejr\e
jZe
jZe�Zdd�Zdd�Zefdd�Zdd�Zdd�Zdd�Zefdd�ZeZeZeZeZdd �Zd!d"�Zd#d$�Zd%d&�Z d'd(�Z!d)d*�Z"e#d+d,��Z$�Z%S).rap
    :param headers:
        An iterable of field-value pairs. Must not contain multiple field names
        when compared case-insensitively.

    :param kwargs:
        Additional field-value pairs to pass in to ``dict.update``.

    A ``dict`` like container for storing HTTP Headers.

    Field names are stored and compared case-insensitively in compliance with
    RFC 7230. Iteration provides the first case-sensitive key seen for each
    case-insensitive pair.

    Using ``__setitem__`` syntax overwrites fields that compare equal
    case-insensitively in order to maintain ``dict``'s api. For fields that
    compare equal, instead create a new ``HTTPHeaderDict`` and use ``.add``
    in a loop.

    If multiple fields that are equal case-insensitively are passed to the
    constructor or ``.update``, the behavior is undefined and some will be
    lost.

    >>> headers = HTTPHeaderDict()
    >>> headers.add('Set-Cookie', 'foo=bar')
    >>> headers.add('set-cookie', 'baz=quxx')
    >>> headers['content-length'] = '7'
    >>> headers['SET-cookie']
    'foo=bar, baz=quxx'
    >>> headers['Content-Length']
    '7'
    NcsPtt|���t�|_|dur>t|t�r4|�|�n
|�|�|rL|�|�dSr)�superrr"rr�
isinstance�
_copy_from�extend)r	�headers�kwargs��	__class__rr
r"�s

zHTTPHeaderDict.__init__cCs ||g|j|��<|j|��Sr�r�lower�r	r$�valrrr
r/�szHTTPHeaderDict.__setitem__cCs |j|��}d�|dd��S)N�, r�rrB�joinrCrrr
r&�szHTTPHeaderDict.__getitem__cCs|j|��=dSrrA�r	r$rrr
r0�szHTTPHeaderDict.__delitem__cCs|��|jvSr)rBrrHrrr
�__contains__�szHTTPHeaderDict.__contains__cCs^t|t�st|d�sdSt|t|��s2t|�|�}tdd�|��D��tdd�|��D��kS)Nr7Fcss|]\}}|��|fVqdSr)rB)�.0�k�vrrr
�	<genexpr>��z(HTTPHeaderDict.__eq__.<locals>.<genexpr>)r:r�hasattr�type�dict�
itermerged�r	�otherrrr
�__eq__�s�zHTTPHeaderDict.__eq__cCs|�|�Sr)rUrSrrr
�__ne__�szHTTPHeaderDict.__ne__cCs
t|j�Sr)r*rrrrr
r1�szHTTPHeaderDict.__len__ccs|j��D]}|dVq
dS)Nr)rr6)r	�valsrrr
r3�szHTTPHeaderDict.__iter__cCs>z||}Wn"ty.||jur&�|YS0||=|SdS)z�D.pop(k[,d]) -> v, remove specified key and return the corresponding value.
        If key is not found, d is returned if given, otherwise KeyError is raised.
        N)�KeyError�_HTTPHeaderDict__marker)r	r$�defaultr,rrr
r#�s

zHTTPHeaderDict.popcCs"z
||=WntyYn0dSr)rXrHrrr
�discard�s
zHTTPHeaderDict.discardcCs4|��}||g}|j�||�}||ur0|�|�dS)z�Adds a (name, value) pair, doesn't overwrite the value if it already
        exists.

        >>> headers = HTTPHeaderDict(foo='bar')
        >>> headers.add('Foo', 'baz')
        >>> headers['foo']
        'bar, baz'
        N)rBr�
setdefault�append)r	r$rD�	key_lower�new_valsrWrrr
�add�s
	zHTTPHeaderDict.addcOs�t|�dkrtd�t|����t|�dkr2|dnd}t|t�r`|��D]\}}|�||�qHnjt|t�r�|D]}|�|||�qnnDt|d�r�|�	�D]}|�|||�q�n|D]\}}|�||�q�|�
�D]\}}|�||�q�dS)z�Generic import function for any type of header-like object.
        Adapted version of MutableMapping.update in order to insert items
        with self.add instead of self.__setitem__
        rz9extend() takes at most 1 positional arguments ({0} given)rrr7N)r*�	TypeError�formatr:r�	iteritemsr`rrOr7�items)r	�argsr>rTr$rDr,rrr
r<�s(��


zHTTPHeaderDict.extendcCsLz|j|��}Wn(ty:||jur2gYS|YS0|dd�SdS)zmReturns a list of all the values for the named field. Returns an
        empty list if the key doesn't exist.rN)rrBrXrY)r	r$rZrWrrr
�getlists

zHTTPHeaderDict.getlistcCsdt|�jt|���fS)Nz%s(%s))rPrrQrRrrrr
�__repr__szHTTPHeaderDict.__repr__cCs>|D]4}|�|�}t|t�r$t|�}|g||j|��<qdSr)rfr:r4rrB)r	rTr$rDrrr
r;s


zHTTPHeaderDict._copy_fromcCst|��}|�|�|Sr)rPr;)r	�clonerrr
�copy"s

zHTTPHeaderDict.copyccs<|D]2}|j|��}|dd�D]}|d|fVq"qdS)z8Iterate over all header lines, including duplicate ones.rNrrA)r	r$rWrDrrr
rc'szHTTPHeaderDict.iteritemsccs8|D].}|j|��}|dd�|dd��fVqdS)z:Iterate over all headers, merging duplicate ones together.rrErNrFrCrrr
rR.szHTTPHeaderDict.itermergedcCst|���Sr)r4rcrrrr
rd4szHTTPHeaderDict.itemscCs�d}g}|jD]h}|�|�rT|s.td|��n&|d\}}||d|��f|d<q|�dd�\}}|�||��f�q||�S)z4Read headers from a Python 2 httplib message object.)� �	z/Header continuation with no previous header: %s���rj�:r)r=�
startswithr�strip�splitr])�cls�message�obs_fold_continued_leadersr=�liner$r,rrr
�from_httplib7s

�zHTTPHeaderDict.from_httplib)N)&rrrr8r"r/r&r0rIrUrVr�PY2rrr�objectrYr1r3r#r[r`r<rf�
getheaders�getallmatchingheaders�iget�get_allrgr;rircrRrd�classmethodru�
__classcell__rrr?r
rks>!	
N)�
__future__r�collections.abcrr�ImportError�collections�	threadingrr�
exceptionsr�packagesrZpackages.sixrr�__all__rwr(rrrrrr
�<module>s Lsite-packages/pip/_vendor/urllib3/__pycache__/response.cpython-39.pyc000064400000050613150564745010021577 0ustar00a

��Re+n�@sVddlmZddlZddlZddlZddlmZddlmZ	ddlm
ZzddlZWne
yjdZYn0ddlmZddlmZmZdd	lmZmZmZmZmZmZmZmZmZmZdd
lmZddl m!Z!m"Z"e�#e$�Z%Gdd
�d
e&�Z'Gdd�de&�Z(Gdd�de&�Z)edu�r(Gdd�de&�Z*Gdd�de&�Z+dd�Z,Gdd�dej-�Z.dS)�)�absolute_importN)�contextmanager)�error)�timeout�)�HTTPHeaderDict)�BaseSSLError�
HTTPException)
�BodyNotHttplibCompatible�DecodeError�	HTTPError�IncompleteRead�InvalidChunkLength�
InvalidHeader�
ProtocolError�ReadTimeoutError�ResponseNotChunked�SSLError)�six)�is_fp_closed�is_response_to_headc@s$eZdZdd�Zdd�Zdd�ZdS)�DeflateDecodercCsd|_d|_t��|_dS)NT�)�
_first_try�_data�zlib�
decompressobj�_obj��self�r ��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/response.py�__init__$szDeflateDecoder.__init__cCst|j|�S�N��getattrr�r�namer r r!�__getattr__)szDeflateDecoder.__getattr__cCs�|s|S|js|j�|�S|j|7_z"|j�|�}|rFd|_d|_|WStjy�d|_t�tj�|_z|�|j�Wd|_YSd|_0Yn0dS)NF)rr�
decompressrrrr�	MAX_WBITS)r�data�decompressedr r r!r),s&�zDeflateDecoder.decompressN��__name__�
__module__�__qualname__r"r(r)r r r r!r#src@seZdZdZdZdZdS)�GzipDecoderStaterr�N)r.r/r0�FIRST_MEMBER�
OTHER_MEMBERS�SWALLOW_DATAr r r r!r1Csr1c@s$eZdZdd�Zdd�Zdd�ZdS)�GzipDecodercCst�dtj�|_tj|_dS�N�)rrr*rr1r3�_staterr r r!r"KszGzipDecoder.__init__cCst|j|�Sr#r$r&r r r!r(OszGzipDecoder.__getattr__cCs�t�}|jtjks|st|�Sz||j�|�7}Wn:tjyl|j}tj|_|tj	krft|�YS�Yn0|jj
}|s�t|�Stj	|_t�dtj�|_qdSr7)
�	bytearrayr9r1r5�bytesrr)rrr4�unused_datarr*)rr+�ret�previous_stater r r!r)Rs 
zGzipDecoder.decompressNr-r r r r!r6Jsr6c@seZdZdd�Zdd�ZdS)�
BrotliDecodercCs0t��|_t|jd�r"|jj|_n
|jj|_dS)Nr))�brotliZDecompressorr�hasattrr)�processrr r r!r"ns
zBrotliDecoder.__init__cCst|jd�r|j��SdS)N�flushr)rArrCrr r r!rCus
zBrotliDecoder.flushN)r.r/r0r"rCr r r r!r?jsr?c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�MultiDecodera
    From RFC7231:
        If one or more encodings have been applied to a representation, the
        sender that applied the encodings MUST generate a Content-Encoding
        header field that lists the content codings in the order in which
        they were applied.
    cCsdd�|�d�D�|_dS)NcSsg|]}t|����qSr )�_get_decoder�strip)�.0�mr r r!�
<listcomp>�rz)MultiDecoder.__init__.<locals>.<listcomp>�,)�split�	_decoders)r�modesr r r!r"�szMultiDecoder.__init__cCs|jd��S�Nr)rLrCrr r r!rC�szMultiDecoder.flushcCst|j�D]}|�|�}q
|Sr#)�reversedrLr))rr+�dr r r!r)�szMultiDecoder.decompressN)r.r/r0�__doc__r"rCr)r r r r!rD{srDcCs:d|vrt|�S|dkrt�Stdur4|dkr4t�St�S)NrJ�gzip�br)rDr6r@r?r)�moder r r!rE�srEc@sfeZdZdZddgZedur&edg7Zgd�ZdJdd�Zd
d�Zdd�Z	dd�Z
edd��Zedd��Z
dd�Zdd�Zdd�Zdd�ZeejfZedur�eejf7Zdd �Zd!d"�Zed#d$��ZdKd%d&�ZdLd(d)�Zed*d+��Zd,d-�ZdMd.d/�Zd0d1�Z d2d3�Z!ed4d5��Z"d6d7�Z#d8d9�Z$d:d;�Z%d<d=�Z&d>d?�Z'd@dA�Z(dBdC�Z)dNdDdE�Z*dFdG�Z+dHdI�Z,dS)O�HTTPResponsea�
    HTTP Response container.

    Backwards-compatible with :class:`http.client.HTTPResponse` but the response ``body`` is
    loaded and decoded on-demand when the ``data`` property is accessed.  This
    class is also compatible with the Python standard library's :mod:`io`
    module, and can hence be treated as a readable object in the context of that
    framework.

    Extra parameters for behaviour not present in :class:`http.client.HTTPResponse`:

    :param preload_content:
        If True, the response's body will be preloaded during construction.

    :param decode_content:
        If True, will attempt to decode the body based on the
        'content-encoding' header.

    :param original_response:
        When this HTTPResponse wrapper is generated from an :class:`http.client.HTTPResponse`
        object, it's convenient to include the original for debug purposes. It's
        otherwise unused.

    :param retries:
        The retries contains the last :class:`~urllib3.util.retry.Retry` that
        was used during the request.

    :param enforce_content_length:
        Enforce content length checking. Body returned by server must match
        value of Content-Length header, if present. Otherwise, raise error.
    rR�deflateNrS)i-i.i/i3i4�rTFcCst|t�r||_n
t|�|_||_||_||_||_||_|
|_||_	||_
d|_d|_d|_
|	|_d|_||_||_|r�t|tjtf�r�||_|
|_||_t|d�r�||_
d|_d|_|j�dd���}dd�|�d�D�}d	|vr�d
|_|�|�|_|�r|j�s|j|d�|_dS)Nr�readFztransfer-encodingrWcss|]}|��VqdSr#)rF)rG�encr r r!�	<genexpr>�rz(HTTPResponse.__init__.<locals>.<genexpr>rJ�chunkedT��decode_content) �
isinstancer�headers�status�version�reason�strictr]�retries�enforce_content_length�
auto_close�_decoder�_body�_fp�_original_response�_fp_bytes_read�msg�_request_urlr�string_typesr;�_pool�_connectionrAr[�
chunk_left�get�lowerrK�_init_length�length_remainingrX)r�bodyr_r`rarbrc�preload_contentr]�original_response�pool�
connectionrlrdre�request_method�request_urlrf�tr_enc�	encodingsr r r!r"�sB


zHTTPResponse.__init__cCs|j|jvr|j�d�SdS)a
        Should we redirect and where to?

        :returns: Truthy redirect location string if we got a redirect status
            code and valid location. ``None`` if redirect status and no
            location. ``False`` if not a redirect status code.
        �locationF)r`�REDIRECT_STATUSESr_rrrr r r!�get_redirect_locationsz"HTTPResponse.get_redirect_locationcCs(|jr|jsdS|j�|j�d|_dSr#)rorp�	_put_connrr r r!�release_connszHTTPResponse.release_conncCs,z|��Wnttttfy&Yn0dS)z�
        Read and discard any remaining HTTP response data in the response connection.

        Unread data in the HTTPResponse connection blocks the connection from being released back to the pool.
        N)rXr�SocketErrorrr	rr r r!�
drain_connszHTTPResponse.drain_conncCs"|jr|jS|jr|jdd�SdS)NT)�
cache_content)rhrirXrr r r!r+%szHTTPResponse.datacCs|jSr#)rprr r r!rz.szHTTPResponse.connectioncCs
t|j�Sr#)rrirr r r!�isclosed2szHTTPResponse.isclosedcCs|jS)z�
        Obtain the number of bytes pulled over the wire so far. May differ from
        the amount of content returned by :meth:``urllib3.response.HTTPResponse.read``
        if bytes are encoded on the wire (e.g, compressed).
        )rkrr r r!�tell5szHTTPResponse.tellcCs�|j�d�}|dur�|jr(t�d�dSz<tdd�|�d�D��}t|�dkrZtd|��|�	�}Wnt
yzd}Yn0|d	kr�d}zt|j�}Wnt
y�d	}Yn0|d
vs�d|kr�dks�n|d
kr�d	}|S)zM
        Set initial length value for Response content if available.
        zcontent-lengthNz�Received response with both Content-Length and Transfer-Encoding set. This is expressly forbidden by RFC 7230 sec 3.3.2. Ignoring Content-Length and attempting to process response as Transfer-Encoding: chunked.cSsg|]}t|��qSr )�int)rG�valr r r!rIWrz-HTTPResponse._init_length.<locals>.<listcomp>rJrz8Content-Length contained multiple unmatching values (%s)r)��i0�d���HEAD)
r_rrr[�log�warning�setrK�lenr�pop�
ValueErrorr�r`)rr{�length�lengthsr`r r r!rt=s6���

&zHTTPResponse._init_lengthcsh�j�dd���}�jdurd|�jvr2t|��_n2d|vrd�fdd�|�d�D�}t|�rdt|��_dS)z=
        Set-up the _decoder attribute if necessary.
        �content-encodingrWNrJcs"g|]}|���jvr|���qSr )rF�CONTENT_DECODERS)rG�err r!rI|s�z.HTTPResponse._init_decoder.<locals>.<listcomp>)r_rrrsrgr�rErKr�)r�content_encodingr~r rr!�
_init_decoderqs


�zHTTPResponse._init_decoderc
Csz|s|Sz|jr|j�|�}WnF|jyd}z,|j�dd���}td||��WYd}~n
d}~00|rv||��7}|S)zN
        Decode the data passed in and potentially flush the decoder.
        r�rWzEReceived response with content-encoding: %s, but failed to decode it.N)rgr)�DECODER_ERROR_CLASSESr_rrrsr�_flush_decoder)rr+r]�
flush_decoderr�r�r r r!�_decode�s ��zHTTPResponse._decodecCs$|jr |j�d�}||j��SdS)zk
        Flushes the decoder. Should only be called if the decoder is actually
        being used.
        r)rgr)rC)r�bufr r r!r��szHTTPResponse._flush_decoderc
cs:d}z�z
dVWn�ty0t|jdd��Ynztyt}z.dt|�vrRt|��t|jdd��WYd}~n>d}~0ttfy�}ztd||��WYd}~n
d}~00d}W|s�|j	r�|j	�
�|jr�|j�
�|j	r�|j	��r�|�
�nH|�s|j	�r|j	�
�|j�r|j�
�|j	�r4|j	���r4|�
�0dS)z�
        Catch low-level python exceptions, instead re-raising urllib3
        variants, so that low-level exceptions are not leaked in the
        high-level api.

        On exit, release the connection back to the pool.
        FNzRead timed out.zread operation timed outzConnection broken: %rT)�
SocketTimeoutrror�strrr	r�rrj�closerpr�r�)r�
clean_exitr�r r r!�_error_catcher�s6	
"$


�

zHTTPResponse._error_catchercCs"|��|dur|j}|jdur$dSd}t|jdd�}|����|dur`|sV|j��nd}d}nPd}|st|j�|�nd}|dkr�|s�|j��d}|jr�|jdvr�t	|j
|j��Wd�n1s�0Y|�r|j
t|�7_
|jdu�r|jt|�8_|�|||�}|�r||_
|S)aT
        Similar to :meth:`http.client.HTTPResponse.read`, but with two additional
        parameters: ``decode_content`` and ``cache_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.

        :param cache_content:
            If True, will save the returned data such that the same result is
            returned despite of the state of the underlying file object. This
            is useful if you want the ``.data`` property to continue working
            after having ``.read()`` the file object. (Overridden if ``amt`` is
            set.)
        NF�closedrTr)rN)r�r]rir%r�rXr�rerur
rkr�r�rh)r�amtr]r�r��	fp_closedr+r r r!rX�s:

��

	,zHTTPResponse.read�ccsR|jr*|��r*|j||d�D]
}|Vqn$t|j�sN|j||d�}|r*|Vq*dS)a_
        A generator wrapper for the read() method. A call will block until
        ``amt`` bytes have been read from the connection or until the
        connection is closed.

        :param amt:
            How much of the content to read. The generator will return up to
            much data per iteration, but may return less. This is particularly
            likely when using compressed data. However, the empty string will
            never be returned.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        r\)r�r]N)r[�supports_chunked_reads�read_chunkedrrirX)rr�r]�liner+r r r!�stream+s

zHTTPResponse.streamc
Ksb|j}t|t�s.tjr"t�|�}nt|���}t|dd�}|f|||j|j	|j
||d�|��}|S)a
        Given an :class:`http.client.HTTPResponse` instance ``r``, return a
        corresponding :class:`urllib3.response.HTTPResponse` object.

        Remaining parameters are passed to the HTTPResponse constructor, along
        with ``original_response=r``.
        rcr)rvr_r`rarbrcrx)rlr^rr�PY2�from_httplib�itemsr%r`rarb)�ResponseCls�r�response_kwr_rc�respr r r!r�Es$	
��
zHTTPResponse.from_httplibcCs|jSr#�r_rr r r!�
getheadersfszHTTPResponse.getheaderscCs|j�||�Sr#)r_rr)rr'�defaultr r r!�	getheaderiszHTTPResponse.getheadercCs|jSr#r�rr r r!�infomszHTTPResponse.infocCs6|js|j��|jr |j��|js2tj�|�dSr#)r�rir�rprf�io�IOBaserr r r!r�qs

zHTTPResponse.closecCsT|jstjj�|�S|jdur"dSt|jd�r8|j��St|jd�rL|jjSdSdS)NTr�r�)rfr�r�r��__get__rirAr�rr r r!r�{s

zHTTPResponse.closedcCs6|jdurtd��nt|jd�r*|j��Std��dS)Nz-HTTPResponse has no file to get a fileno from�filenozOThe file-like object this HTTPResponse is wrapped around has no file descriptor)ri�IOErrorrAr�rr r r!r��s


�zHTTPResponse.filenocCs2|jdur.t|jd�r.t|jdd�s.|j��SdS)NrCr�F)rirAr%rCrr r r!rC�s�
��zHTTPResponse.flushcCsdS)NTr rr r r!�readable�szHTTPResponse.readablecCs:|�t|��}t|�dkrdS||dt|��<t|�SdSrN)rXr�)r�b�tempr r r!�readinto�s
zHTTPResponse.readintocCst|jd�S)a
        Checks if the underlying file-like object looks like a
        :class:`http.client.HTTPResponse` object. We do this by testing for
        the fp attribute. If it is present we assume it returns raw chunks as
        processed by read_chunked().
        �fp)rArirr r r!r��sz#HTTPResponse.supports_chunked_readscCsd|jdurdS|jj��}|�dd�d}zt|d�|_Wn$ty^|��t||��Yn0dS)N�;rrr8)	rqrir��readlinerKr�r�r�r)rr�r r r!�_update_chunk_length�s
z!HTTPResponse._update_chunk_lengthcCs�d}|dur2|j�|j�}|}|j�d�d|_nv||jkrZ|j�|�}|j||_|}nN||jkr�|j�|�}|j�d�d|_|}n |j�|j�}|j�d�d|_|S)Nr2)ri�
_safe_readrq)rr��returned_chunk�chunk�valuer r r!�
_handle_chunk�s&

zHTTPResponse._handle_chunkccs|��|jstd��|��s&td��|����|jrZt|j�rZ|j��Wd�dS|j	j
durvWd�dS|��|jdkr�q�|�
|�}|j||dd�}|rv|Vqv|r�|��}|r�|V|j	j
��}|s�q�|dkr�q�q�|jr�|j��Wd�n1�s
0YdS)a�
        Similar to :meth:`HTTPResponse.read`, but with an additional
        parameter: ``decode_content``.

        :param amt:
            How much of the content to read. If specified, caching is skipped
            because it doesn't make sense to cache partial content as the full
            response.

        :param decode_content:
            If True, will attempt to decode the body based on the
            'content-encoding' header.
        zHResponse is not chunked. Header 'transfer-encoding: chunked' is missing.zkBody should be http.client.HTTPResponse like. It should have have an fp attribute which returns raw chunks.NrF)r]r�s
)r�r[rr�r
r�rjrr�rir�r�rqr�r�r�r�)rr�r]r��decodedr�r r r!r��sF��



�zHTTPResponse.read_chunkedcCs.|jdur$t|jj�r$|jjdjS|jSdS)z�
        Returns the URL that was the source of this response.
        If the request that generated this response redirected, this method
        will return the final redirect location.
        N���)rdr��history�redirect_locationrmrr r r!�geturlszHTTPResponse.geturlccs�g}|jdd�D]n}d|vrt|�d�}d�|�|ddV|dd�D]}|dVqJ|drn|dg}q~g}q|�|�q|r�d�|�VdS)NTr\�
rrrr�)r�rK�join�append)r�bufferr��xr r r!�__iter__&s
zHTTPResponse.__iter__)rWNrrNrTTNNNNNFNNT)NNF)r�N)N)NN)-r.r/r0rQr�r@r�r"r�r�r��propertyr+rzr�r�rtr�r�rrr�r�r�rr�rXr��classmethodr�r�r�r�r�r�r�rCr�r�r�r�r�r�r�r�r r r r!rU�sz 
�
C


4

7
J

 


		
FrU)/�
__future__rr��loggingr�
contextlibr�socketrr�rr�r@�ImportError�_collectionsrrzrr	�
exceptionsr
rrr
rrrrrr�packagesrZ
util.responserr�	getLoggerr.r��objectrr1r6r?rDrEr�rUr r r r!�<module>s0
0
 

site-packages/pip/_vendor/urllib3/__pycache__/connectionpool.cpython-39.pyc000064400000060254150564745010022774 0ustar00a

��ReӒ�@s�ddlmZddlZddlZddlZddlZddlZddlmZddlm	Z
ddlmZm
Z
mZmZmZmZmZmZddlmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!ddl"m#Z#dd	l$m%Z%dd
l&m'Z'ddl(m)Z)ddl*m+Z+dd
l,m-Z-ddl.m/Z/ddl0m1Z1ddl2m3Z3ddl4m5Z5ddl6m7Z7ddl8m9Z9ddl:m;Z;m<Z<ddl:m=Z>ddl:m?Z?m@Z@e#jAjBZBe�CeD�ZEeF�ZGGdd�deF�ZHejIejJhZKGdd�deHe)�ZLGdd�deL�ZMdd�ZNdd �Z=dS)!�)�absolute_importN)�error��timeout�)�BaseSSLError�BrokenPipeError�DummyConnection�HTTPConnection�
HTTPException�HTTPSConnection�VerifiedHTTPSConnection�port_by_scheme)
�ClosedPoolError�EmptyPoolError�HeaderParsingError�HostChangedError�InsecureRequestWarning�LocationValueError�
MaxRetryError�NewConnectionError�
ProtocolError�
ProxyError�ReadTimeoutError�SSLError�TimeoutError)�six)�queue)�CertificateError)�RequestMethods)�HTTPResponse)�is_connection_dropped)�connection_requires_http_tunnel)�	LifoQueue)�set_file_position)�assert_header_parsing)�Retry)�Timeout)�Url�_encode_target)�_normalize_host)�get_host�	parse_urlc@sBeZdZdZdZeZd
dd�Zdd�Zdd�Z	d	d
�Z
dd�ZdS)�ConnectionPoola;
    Base class for all connection pools, such as
    :class:`.HTTPConnectionPool` and :class:`.HTTPSConnectionPool`.

    .. note::
       ConnectionPool.urlopen() does not normalize or percent-encode target URIs
       which is useful if your target server doesn't support percent-encoded
       target URIs.
    NcCs0|std��t||jd�|_|��|_||_dS)NzNo host specified.��scheme)rr*r/�host�lower�_proxy_host�port)�selfr0r3�r5��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/connectionpool.py�__init__Js

zConnectionPool.__init__cCsdt|�j|j|jfS)Nz%s(host=%r, port=%r))�type�__name__r0r3�r4r5r5r6�__str__RszConnectionPool.__str__cCs|S�Nr5r:r5r5r6�	__enter__UszConnectionPool.__enter__cCs|��dS)NF)�close)r4�exc_type�exc_val�exc_tbr5r5r6�__exit__XszConnectionPool.__exit__cCsdS)�D
        Close all pooled connections and disable the pool.
        Nr5r:r5r5r6r>]szConnectionPool.close)N)r9�
__module__�__qualname__�__doc__r/r#�QueueClsr7r;r=rBr>r5r5r5r6r-<s

r-c
@s�eZdZdZdZeZeZdde	j
dddddddf
dd�Zdd	�Zd!d
d�Z
dd
�Zdd�Zdd�Zdd�Zdd�Zedfdd�Zdd�Zdd�Zdd�Zdddddeddddf
dd �ZdS)"�HTTPConnectionPoola:	
    Thread-safe connection pool for one host.

    :param host:
        Host used for this HTTP Connection (e.g. "localhost"), passed into
        :class:`http.client.HTTPConnection`.

    :param port:
        Port used for this HTTP Connection (None is equivalent to 80), passed
        into :class:`http.client.HTTPConnection`.

    :param strict:
        Causes BadStatusLine to be raised if the status line can't be parsed
        as a valid HTTP/1.0 or 1.1 status line, passed into
        :class:`http.client.HTTPConnection`.

        .. note::
           Only works in Python 2. This parameter is ignored in Python 3.

    :param timeout:
        Socket timeout in seconds for each individual connection. This can
        be a float or integer, which sets the timeout for the HTTP request,
        or an instance of :class:`urllib3.util.Timeout` which gives you more
        fine-grained control over request timeouts. After the constructor has
        been parsed, this is always a `urllib3.util.Timeout` object.

    :param maxsize:
        Number of connections to save that can be reused. More than 1 is useful
        in multithreaded situations. If ``block`` is set to False, more
        connections will be created but they will not be saved once they've
        been used.

    :param block:
        If set to True, no more than ``maxsize`` connections will be used at
        a time. When no free connections are available, the call will block
        until a connection has been released. This is a useful side effect for
        particular multithreaded situations where one does not want to use more
        than maxsize connections per host to prevent flooding.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param retries:
        Retry configuration to use by default with requests in this pool.

    :param _proxy:
        Parsed proxy URL, should not be used directly, instead, see
        :class:`urllib3.ProxyManager`

    :param _proxy_headers:
        A dictionary with proxy headers, should not be used directly,
        instead, see :class:`urllib3.ProxyManager`

    :param \**conn_kw:
        Additional parameters are used to create fresh :class:`urllib3.connection.HTTPConnection`,
        :class:`urllib3.connection.HTTPSConnection` instances.
    �httpNFrcKs�t�|||�t�||�||_t|t�s4t�|�}|durBtj}||_	||_
|�|�|_||_
|	|_|
pli|_||_t|�D]}
|j�d�q~d|_d|_||_|jr�|j�dg�|j|jd<|j|jd<dS)Nr�socket_options�proxy�proxy_config)r-r7r�strict�
isinstancer'�
from_floatr&�DEFAULTr�retriesrG�pool�blockrK�
proxy_headersrL�xrange�put�num_connections�num_requests�conn_kw�
setdefault)r4r0r3rMr�maxsizerS�headersrQ�_proxy�_proxy_headers�
_proxy_configrY�_r5r5r6r7�s.


zHTTPConnectionPool.__init__cCsR|jd7_t�d|j|j|jp$d�|jf|j|j|jj|jd�|j	��}|S)z9
        Return a fresh :class:`HTTPConnection`.
        rz(Starting new HTTP connection (%d): %s:%s�80)r0r3rrM)
rW�log�debugr0r3�
ConnectionClsr�connect_timeoutrMrY�r4�connr5r5r6�	_new_conn�s ���zHTTPConnectionPool._new_conncCs�d}z|jj|j|d�}Wn>ty6t|d��Yn$tjyX|jrTt|d��Yn0|r�t|�r�t	�
d|j�|��t
|dd�dkr�d}|p�|��S)	a�
        Get a connection. Will return a pooled connection if one is available.

        If no connections are available and :prop:`.block` is ``False``, then a
        fresh connection is returned.

        :param timeout:
            Seconds to wait before giving up and raising
            :class:`urllib3.exceptions.EmptyPoolError` if the pool is empty and
            :prop:`.block` is ``True``.
        N)rSrzPool is closed.z>Pool reached maximum size and no more connections are allowed.z Resetting dropped connection: %s�	auto_openrr)rR�getrS�AttributeErrorrr�Emptyrr!rbrcr0r>�getattrrh)r4rrgr5r5r6�	_get_conn�s$�zHTTPConnectionPool._get_conncCsZz|jj|dd�WdSty(Yn"tjyHt�d|j�Yn0|rV|��dS)a�
        Put a connection back into the pool.

        :param conn:
            Connection object for the current host and port as returned by
            :meth:`._new_conn` or :meth:`._get_conn`.

        If the pool is already full, the connection is closed and discarded
        because we exceeded maxsize. If connections are discarded frequently,
        then maxsize should be increased.

        If the pool is closed, then the connection will be closed and discarded.
        F�rSNz2Connection pool is full, discarding connection: %s)	rRrVrkr�Fullrb�warningr0r>rfr5r5r6�	_put_connszHTTPConnectionPool._put_conncCsdS)�U
        Called right before a request is made, after the socket is created.
        Nr5rfr5r5r6�_validate_conn6sz!HTTPConnectionPool._validate_conncCsdSr<r5rfr5r5r6�_prepare_proxy<sz!HTTPConnectionPool._prepare_proxycCs2|tur|j��St|t�r$|��St�|�SdS)z:Helper that always returns a :class:`urllib3.util.Timeout`N)�_Defaultr�clonerNr'rO)r4rr5r5r6�_get_timeout@s


zHTTPConnectionPool._get_timeoutcCsjt|t�rt||d|��t|d�r>|jtvr>t||d|��dt|�vsVdt|�vrft||d|��dS)zAIs the error actually a timeout? Will raise a ReadTimeout or pass�!Read timed out. (read timeout=%s)�errnoz	timed outzdid not complete (read)N)rN�
SocketTimeoutr�hasattrrz�_blocking_errnos�str)r4�err�url�
timeout_valuer5r5r6�_raise_timeoutLs

�
��
�z!HTTPConnectionPool._raise_timeoutc

Ks||jd7_|�|�}|��|j|_z|�|�Wn<ttfyr}z |j|||jd��WYd}~n
d}~00z2|r�|j	||fi|��n|j
||fi|��WnLty�Yn<ty�}z$|j
t
jt
jt
jhvrނWYd}~n
d}~00|j}	t|dd��rL|	dk�r"t||d|	��|	tju�r@|j�t���n|j�|	�zjz|jdd�}
WnTt�y�z|��}
Wn2t�y�}zt�|d�WYd}~n
d}~00Yn0Wn>tttf�y�}z|j|||	d��WYd}~n
d}~00t|d	d
�}t �!d|j"|j#|j$||||
j%|
j&�	zt'|
j(�WnBt)tf�yv}z$t j*d|�+|�|dd
�WYd}~n
d}~00|
S)a
        Perform a request on a given urllib connection object taken from our
        pool.

        :param conn:
            a connection from one of our connection pools

        :param timeout:
            Socket timeout in seconds for the request. This can be a
            float or integer, which will set the same timeout value for
            the socket connect and the socket read, or an instance of
            :class:`urllib3.util.Timeout`, which gives you more fine-grained
            control over your timeouts.
        r)rr�r�N�sockrryT)�	buffering�
_http_vsn_strzHTTP/?z%s://%s:%s "%s %s %s" %s %sz$Failed to parse headers (url=%s): %s)�exc_info),rXrx�
start_connectrerrtr{rr��request_chunked�requestr�IOErrorrz�EPIPE�	ESHUTDOWN�
EPROTOTYPE�read_timeoutrmrr'�DEFAULT_TIMEOUTr��
settimeout�socket�getdefaulttimeout�getresponse�	TypeError�
BaseExceptionr�
raise_from�SocketErrorrbrcr/r0r3�status�lengthr%�msgrrq�
_absolute_url)
r4rg�methodr�r�chunked�httplib_request_kw�timeout_obj�er��httplib_response�http_version�hper5r5r6�
_make_requestes�
�

�,��z HTTPConnectionPool._make_requestcCst|j|j|j|d�jS)N)r/r0r3�path)r(r/r0r3r�)r4r�r5r5r6r��sz HTTPConnectionPool._absolute_urlcCsT|jdurdS|jd}|_z|jdd�}|r|��qWntjyNYn0dS)rCNFro)rRrjr>rrl)r4�old_poolrgr5r5r6r>�s
zHTTPConnectionPool.closecCsx|�d�rdSt|�\}}}|dur0t||d�}|jrF|sFt�|�}n|js^|t�|�kr^d}|||f|j|j|jfkS)zj
        Check if the given ``url`` is a member of the same host as this
        connection pool.
        �/TNr.)�
startswithr+r*r3rrjr/r0)r4r�r/r0r3r5r5r6�is_same_host�s

zHTTPConnectionPool.is_same_hostTc
Ks<t|�}|j}|dur|j}t|t�s8tj|||jd�}|
durL|
�dd�}
|rf|�|�sft	|||��|�
d�r�t�t
|��}nt�|j�}d}|
}t|j|j|�}|s�|��}|�|j�d}d}t||�}�z�z�|�|�}|j|	d�}|j|_|jdu�o
t|dd�}|�r"|�r"|�|�|j|||||||d	�}|
�sD|nd}||
d
<|jj|f|||d�|
��}d}Wn�t�y�d}d}�Yn�t t!t"t#t$t%t&f�yF}z�d}t|t$t&f��r�t%|�}n>t|t"t'f��r�|j�r�t(d|�}nt|t"t!f��r
t#d
|�}|j)||||t*�+�dd�}|�,�|}WYd}~n
d}~00W|�sb|�o\|�-�}d}|�r�|�.|�n*|�s�|�o�|�-�}d}|�r�|�.|�0|�s�t/�0d|||�|j1|||||||f||	|
||d�|
��S|�o�|�2�}|�r�|j3dk�rd}z|j)||||d�}Wn*t4�yB|j5�r:|�6��|YS0|�6�|�7|�t/�8d||�|j1||||f|||||	|
||d�|
��St9|�:d��}|�;||j3|��r8z|j)||||d�}Wn*t4�y�|j<�r�|�6��|YS0|�6�|�,|�t/�8d|�|j1||||f|||||	|
||d�|
��S|S)a+
        Get a connection from the pool and perform an HTTP request. This is the
        lowest level call for making a request, so you'll need to specify all
        the raw details.

        .. note::

           More commonly, it's appropriate to use a convenience method provided
           by :class:`.RequestMethods`, such as :meth:`request`.

        .. note::

           `release_conn` will only behave as expected if
           `preload_content=False` because we want to make
           `preload_content=False` the default behaviour someday soon without
           breaking backwards compatibility.

        :param method:
            HTTP request method (such as GET, POST, PUT, etc.)

        :param url:
            The URL to perform the request on.

        :param body:
            Data to send in the request body, either :class:`str`, :class:`bytes`,
            an iterable of :class:`str`/:class:`bytes`, or a file-like object.

        :param headers:
            Dictionary of custom headers to send, such as User-Agent,
            If-None-Match, etc. If None, pool headers are used. If provided,
            these headers completely replace any pool-specific headers.

        :param retries:
            Configure the number of retries to allow before raising a
            :class:`~urllib3.exceptions.MaxRetryError` exception.

            Pass ``None`` to retry until you receive a response. Pass a
            :class:`~urllib3.util.retry.Retry` object for fine-grained control
            over different types of retries.
            Pass an integer number to retry connection errors that many times,
            but no other types of errors. Pass zero to never retry.

            If ``False``, then retries are disabled and any exception is raised
            immediately. Also, instead of raising a MaxRetryError on redirects,
            the redirect response will be returned.

        :type retries: :class:`~urllib3.util.retry.Retry`, False, or an int.

        :param redirect:
            If True, automatically handle redirects (status codes 301, 302,
            303, 307, 308). Each redirect counts as a retry. Disabling retries
            will disable redirect, too.

        :param assert_same_host:
            If ``True``, will make sure that the host of the pool requests is
            consistent else will raise HostChangedError. When ``False``, you can
            use the pool on an HTTP proxy and request foreign hosts.

        :param timeout:
            If specified, overrides the default timeout for this one
            request. It may be a float (in seconds) or an instance of
            :class:`urllib3.util.Timeout`.

        :param pool_timeout:
            If set and the pool is set to block=True, then this method will
            block for ``pool_timeout`` seconds and raise EmptyPoolError if no
            connection is available within the time period.

        :param release_conn:
            If False, then the urlopen call will not release the connection
            back into the pool once a response is received (but will release if
            you read the entire contents of the response such as when
            `preload_content=True`). This is useful if you're not preloading
            the response's content immediately. You will need to call
            ``r.release_conn()`` on the response ``r`` to return the connection
            back into the pool. If None, it takes the value of
            ``response_kw.get('preload_content', True)``.

        :param chunked:
            If True, urllib3 will send the body using chunked transfer
            encoding. Otherwise, urllib3 will send the body using the standard
            content-length form. Defaults to False.

        :param int body_pos:
            Position to seek to in file-like body in the event of a retry or
            redirect. Typically this won't need to be set because urllib3 will
            auto-populate the value when needed.

        :param \**response_kw:
            Additional parameters are passed to
            :meth:`urllib3.response.HTTPResponse.from_httplib`
        N)�redirect�default�preload_contentTr�Frr�)r�bodyr\r��request_method)rR�
connectionrQzCannot connect to proxy.zConnection aborted.�)r�_pool�_stacktracez1Retrying (%r) after connection broken by '%r': %s)r�pool_timeout�release_connr��body_posi/�GET)�responser�zRedirecting %s -> %s)rQr��assert_same_hostrr�r�r�r�zRetry-Afterz	Retry: %s)=r,r/r\rNr&�from_intrQrjr�rr�r�
ensure_strr)r�r"rKrL�copy�updaterTr$rxrnrerrmrur��ResponseCls�from_httplibrrrr�rrrrrr�	increment�sysr��sleepr>rrrbrq�urlopen�get_redirect_locationr�r�raise_on_redirect�
drain_conn�sleep_for_retryrc�bool�	getheader�is_retry�raise_on_status)r4r�r�r�r\rQr�r�rr�r�r�r��response_kw�
parsed_url�destination_schemerg�release_this_conn�http_tunnel_requiredr�
clean_exitr��is_new_proxy_connr��
response_connr�r��redirect_location�has_retry_afterr5r5r6r�sVm


�

�
����	�

�����
�

��
�

��
�zHTTPConnectionPool.urlopen)N)r9rDrErFr/r
rdr r�r'r�r7rhrnrrrtrurxr�rvr�r�r>r�r�r5r5r5r6rHhsL;�
6
'�
w�rHcspeZdZdZdZeZddejdddddddddddddddfdd�Z	dd	�Z
d
d�Zdd
�Z�fdd�Z
�ZS)�HTTPSConnectionPoola
    Same as :class:`.HTTPConnectionPool`, but HTTPS.

    :class:`.HTTPSConnection` uses one of ``assert_fingerprint``,
    ``assert_hostname`` and ``host`` in this order to verify connections.
    If ``assert_hostname`` is False, no verification is done.

    The ``key_file``, ``cert_file``, ``cert_reqs``, ``ca_certs``,
    ``ca_cert_dir``, ``ssl_version``, ``key_password`` are only used if :mod:`ssl`
    is available and are fed into :meth:`urllib3.util.ssl_wrap_socket` to upgrade
    the connection socket into an SSL socket.
    �httpsNFrcKs`tj||||||||||	|
fi|��||_||_|
|_||_||_||_||_||_	||_
dSr<)rHr7�key_file�	cert_file�	cert_reqs�key_password�ca_certs�ca_cert_dir�ssl_version�assert_hostname�assert_fingerprint)r4r0r3rMrr[rSr\rQr]r^r�r�r�r�r�r�r�r�r�rYr5r5r6r7rs0��zHTTPSConnectionPool.__init__c
Cs@t|t�r<|j|j|j|j|j|j|j|j	|j
d�|j|_|S)z�
        Prepare the ``connection`` for :meth:`urllib3.util.ssl_wrap_socket`
        and establish the tunnel if proxy is used.
        )r�r�r�r�r�r�r�r�)rNr
�set_certr�r�r�r�r�r�r�r�r�rfr5r5r6�
_prepare_conn�s
�
z!HTTPSConnectionPool._prepare_conncCs2|�|j|j|j�|jjdkr&d|_|��dS)z�
        Establishes a tunnel connection through HTTP CONNECT.

        Tunnel connection is established early because otherwise httplib would
        improperly set Host: header to proxy's IP:port.
        r�TN)�
set_tunnelr2r3rTrKr/�tls_in_tls_required�connectrfr5r5r6ru�sz"HTTPSConnectionPool._prepare_proxyc
Cs�|jd7_t�d|j|j|jp$d�|jr8|jtur@td��|j}|j}|jdurf|jj}|jj}|jf|||j	j
|j|j|j
|jd�|j��}|�|�S)zF
        Return a fresh :class:`http.client.HTTPSConnection`.
        rz)Starting new HTTPS connection (%d): %s:%s�443zCCan't connect to HTTPS URL because the SSL module is not available.N)r0r3rrMr�r�r�)rWrbrcr0r3rdr	rrKrrerMr�r�r�rYr�)r4�actual_host�actual_portrgr5r5r6rh�s8��
��zHTTPSConnectionPool._new_conncs\tt|��|�t|dd�s$|��|js<t�d|jt	�t|dd�durXt�dt	�dS)rsr�Nz�Unverified HTTPS request is being made to host '%s'. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html#ssl-warnings�proxy_is_verifiedFz�Unverified HTTPS connection done to an HTTPS proxy. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html#ssl-warnings)
�superr�rtrmr��is_verified�warnings�warnr0rrf��	__class__r5r6rt�s��
�z"HTTPSConnectionPool._validate_conn)r9rDrErFr/rrdr'r�r7r�rurhrt�
__classcell__r5r5r�r6r�as4
�
1$r�cKsRt|�\}}}|pt�|d�}|dkr:t|fd|i|��St|fd|i|��SdS)a�
    Given a url, return an :class:`.ConnectionPool` instance of its host.

    This is a shortcut for not having to parse out the scheme, host, and port
    of the url before creating an :class:`.ConnectionPool` instance.

    :param url:
        Absolute URL string that must include the scheme. Port is optional.

    :param \**kw:
        Passes additional parameters to the constructor of the appropriate
        :class:`.ConnectionPool`. Useful for specifying things like
        timeout, maxsize, headers, etc.

    Example::

        >>> conn = connection_from_url('http://google.com/')
        >>> r = conn.request('GET', '/')
    �Pr�r3N)r+rrjr�rH)r��kwr/r0r3r5r5r6�connection_from_urls
r�cCs.t||�}|�d�r*|�d�r*|dd�}|S)z?
    Normalize hosts for comparisons and use with sockets.
    �[�]r���)�normalize_hostr��endswith)r0r/r5r5r6r*'s
r*)O�
__future__rrz�loggingr�r�r�rr�rr{r�rrr	r
rrr
r�
exceptionsrrrrrrrrrrrrr�packagesrZpackages.six.movesrZpackages.ssl_match_hostnamerr�rr�r Zutil.connectionr!Z
util.proxyr"Z
util.queuer#Zutil.requestr$Z
util.responser%Z
util.retryr&Zutil.timeoutr'Zutil.urlr(r)r*r�r+r,�movesrU�	getLoggerr9rb�objectrvr-�EAGAIN�EWOULDBLOCKr}rHr�r�r5r5r5r6�<module>sN(
<
)~+site-packages/pip/_vendor/urllib3/__pycache__/poolmanager.cpython-39.pyc000064400000035543150564745010022252 0ustar00a

��Re3M�@s6ddlmZddlZddlZddlZddlmZddlmZm	Z	m
Z
ddlmZm
Z
mZmZmZddlmZddlmZdd	lmZdd
lmZddlmZddlmZgd
�Ze�e�Z dZ!dZ"e�#de"�Z$dZ%e�#de%�Z&dd�Z'e�(e'e$�e�(e'e$�d�Z)ee	d�Z*Gdd�de�Z+Gdd�de+�Z,dd�Z-dS)�)�absolute_importN�)�RecentlyUsedContainer)�HTTPConnectionPool�HTTPSConnectionPool�port_by_scheme)�LocationValueError�
MaxRetryError�ProxySchemeUnknown�ProxySchemeUnsupported�URLSchemeUnknown)�six)�urljoin)�RequestMethods)�connection_requires_http_tunnel)�Retry)�	parse_url)�PoolManager�ProxyManager�proxy_from_url)�key_file�	cert_file�	cert_reqs�ca_certs�ssl_version�ca_cert_dir�ssl_context�key_password)�
key_scheme�key_host�key_port�key_timeout�key_retries�
key_strict�	key_block�key_source_address�key_key_file�key_key_password�
key_cert_file�
key_cert_reqs�key_ca_certs�key_ssl_version�key_ca_cert_dir�key_ssl_context�key_maxsize�key_headers�
key__proxy�key__proxy_headers�key__proxy_config�key_socket_options�key__socks_options�key_assert_hostname�key_assert_fingerprint�key_server_hostname�PoolKey)r�use_forwarding_for_https�ProxyConfigcCs�|��}|d��|d<|d��|d<dD],}||vr,||dur,t||���||<q,|�d�}|durxt|�|d<t|���D]}|�|�|d|<q�|j	D]}||vr�d||<q�|fi|��S)a�
    Create a pool key out of a request context dictionary.

    According to RFC 3986, both the scheme and host are case-insensitive.
    Therefore, this function normalizes both before constructing the pool
    key for an HTTPS request. If you wish to change this behaviour, provide
    alternate callables to ``key_fn_by_scheme``.

    :param key_class:
        The class to use when constructing the key. This should be a namedtuple
        with the ``scheme`` and ``host`` keys at a minimum.
    :type  key_class: namedtuple
    :param request_context:
        A dictionary-like object that contain the context for a request.
    :type  request_context: dict

    :return: A namedtuple that can be used as a connection pool key.
    :rtype:  PoolKey
    �scheme�host)�headers�_proxy_headers�_socks_optionsN�socket_options�key_)
�copy�lower�	frozenset�items�get�tuple�list�keys�pop�_fields)�	key_class�request_context�context�key�socket_opts�field�rR��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/poolmanager.py�_default_key_normalizerNs


rT��http�httpsc@s�eZdZdZdZdZd dd�Zdd�Zdd	�Zd!d
d�Z	dd
�Z
d"dd�Zdd�Zd#dd�Z
d$dd�Zdd�Zdd�Zdd�Zd%dd�ZdS)&ra$
    Allows for arbitrary requests while transparently keeping track of
    necessary connection pools for you.

    :param num_pools:
        Number of connection pools to cache before discarding the least
        recently used pool.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param \**connection_pool_kw:
        Additional parameters are used to create fresh
        :class:`urllib3.connectionpool.ConnectionPool` instances.

    Example::

        >>> manager = PoolManager(num_pools=2)
        >>> r = manager.request('GET', 'http://google.com/')
        >>> r = manager.request('GET', 'http://google.com/mail')
        >>> r = manager.request('GET', 'http://yahoo.com/')
        >>> len(manager.pools)
        2

    N�
cKs8t�||�||_t|dd�d�|_t|_t��|_dS)NcSs|��S�N)�close)�prRrRrS�<lambda>��z&PoolManager.__init__.<locals>.<lambda>)�dispose_func)r�__init__�connection_pool_kwr�pools�pool_classes_by_scheme�key_fn_by_schemerB)�self�	num_poolsr=r`rRrRrSr_�s
zPoolManager.__init__cCs|SrYrR�rdrRrRrS�	__enter__�szPoolManager.__enter__cCs|��dS)NF)�clear)rd�exc_type�exc_val�exc_tbrRrRrS�__exit__�szPoolManager.__exit__cCsb|j|}|dur|j��}dD]}|�|d�q |dkrPtD]}|�|d�q>|||fi|��S)a�
        Create a new :class:`urllib3.connectionpool.ConnectionPool` based on host, port, scheme, and
        any additional pool keyword arguments.

        If ``request_context`` is provided, it is provided as keyword arguments
        to the pool class used. This method is used to actually create the
        connection pools handed out by :meth:`connection_from_url` and
        companion methods. It is intended to be overridden for customization.
        N)r;r<�portrV)rbr`rBrJ�SSL_KEYWORDS)rdr;r<rmrM�pool_clsrO�kwrRrRrS�	_new_pool�s


zPoolManager._new_poolcCs|j��dS)z�
        Empty our store of pools and direct them all to close.

        This will not affect in-flight connections, but they will not be
        re-used after completion.
        N)rarhrfrRrRrSrh�szPoolManager.clearrVcCsT|std��|�|�}|pd|d<|s:t�|d��d�}||d<||d<|�|�S)a�
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the host, port, and scheme.

        If ``port`` isn't given, it will be derived from the ``scheme`` using
        ``urllib3.connectionpool.port_by_scheme``. If ``pool_kwargs`` is
        provided, it is merged with the instance's ``connection_pool_kw``
        variable and used to create the new connection pool, if one is
        needed.
        zNo host specified.rVr;�Prmr<)r�_merge_pool_kwargsrrFrC�connection_from_context)rdr<rmr;�pool_kwargsrMrRrRrS�connection_from_host�s
z PoolManager.connection_from_hostcCs:|d��}|j�|�}|s$t|��||�}|j||d�S)z�
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the request context.

        ``request_context`` must at least contain the ``scheme`` key and its
        value must be a key in ``key_fn_by_scheme`` instance variable.
        r;�rM)rCrcrFr�connection_from_pool_key)rdrMr;�pool_key_constructor�pool_keyrRrRrSrt�sz#PoolManager.connection_from_contextcCs�|jj�f|j�|�}|r,|Wd�S|d}|d}|d}|j||||d�}||j|<Wd�n1st0Y|S)a
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the provided pool key.

        ``pool_key`` should be a namedtuple that only contains immutable
        objects. At a minimum it must have the ``scheme``, ``host``, and
        ``port`` fields.
        Nr;r<rmrw)ra�lockrFrq)rdrzrM�poolr;r<rmrRrRrSrxs
(z$PoolManager.connection_from_pool_keycCs t|�}|j|j|j|j|d�S)a�
        Similar to :func:`urllib3.connectionpool.connection_from_url`.

        If ``pool_kwargs`` is not provided and a new pool needs to be
        constructed, ``self.connection_pool_kw`` is used to initialize
        the :class:`urllib3.connectionpool.ConnectionPool`. If ``pool_kwargs``
        is provided, it is used instead. Note that if a new pool does not
        need to be created for the request, the provided ``pool_kwargs`` are
        not used.
        )rmr;ru)rrvr<rmr;)rd�urlru�urRrRrS�connection_from_urls�zPoolManager.connection_from_urlc	CsT|j��}|rP|��D]8\}}|durFz
||=WqNtyBYqN0q|||<q|S)a
        Merge a dictionary of override values for self.connection_pool_kw.

        This does not modify self.connection_pool_kw and returns a new dict.
        Any keys in the override dictionary with a value of ``None`` are
        removed from the merged dictionary.
        N)r`rBrE�KeyError)rd�override�base_pool_kwargsrO�valuerRrRrSrs.s


zPoolManager._merge_pool_kwargscCs"|jdurdSt|j|j|j�S)z�
        Indicates if the proxy requires the complete destination URL in the
        request.  Normally this is only needed when not using an HTTP CONNECT
        tunnel.
        NF)�proxyr�proxy_configr;)rd�
parsed_urlrRrRrS�!_proxy_requires_url_absolute_formBs

�z-PoolManager._proxy_requires_url_absolute_formcCs@|jdus|dkrdS|jjdkr&dStjr<|jjs<td��dS)z�
        Validates that were not attempting to do TLS in TLS connections on
        Python2 or with unsupported SSL implementations.
        NrWzfContacting HTTPS destinations through HTTPS proxies 'via CONNECT tunnels' is not supported in Python 2)r�r;r
�PY2r�r9r)rd�
url_schemerRrRrS�$_validate_proxy_scheme_url_selectionOs�z0PoolManager._validate_proxy_scheme_url_selectionTcKs�t|�}|�|j�|j|j|j|jd�}d|d<d|d<d|vrP|j��|d<|�|�rp|j	||fi|��}n|j	||j
fi|��}|o�|��}|s�|St||�}|j
dkr�d}|�d�}	t|	t�s�tj|	|d	�}	|	j�r"|�|��s"tt�|d��}
|
D]"}|��|	jvr�|d�|d
�q�z|	j||||d�}	Wn*t�yb|	j�rZ|���|YS0|	|d<||d<t�d||�|��|j	||fi|��S)
aN
        Same as :meth:`urllib3.HTTPConnectionPool.urlopen`
        with custom cross-host redirect logic and only sends the request-uri
        portion of the ``url``.

        The given ``url`` parameter must be absolute, such that an appropriate
        :class:`urllib3.connectionpool.ConnectionPool` can be chosen for it.
        )rmr;F�assert_same_host�redirectr=i/�GET�retries)r�N)�response�_poolzRedirecting %s -> %s)rr�r;rvr<rmr=rBr��urlopen�request_uri�get_redirect_locationr�statusrF�
isinstancer�from_int�remove_headers_on_redirect�is_same_hostrHr
�iterkeysrCrJ�	incrementr	�raise_on_redirect�
drain_conn�log�info)rd�methodr}r�rpr~�connr��redirect_locationr�r=�headerrRrRrSr�`sL	




�
zPoolManager.urlopen)rXN)N)NrVN)N)N)T)�__name__�
__module__�__qualname__�__doc__r�r�r_rgrlrqrhrvrtrxrrsr�r�r�rRrRrRrSr�s 


	



rcsHeZdZdZd�fdd�	Zd�fdd	�	Zdd
d�Zd�fd
d�	Z�ZS)ra>
    Behaves just like :class:`PoolManager`, but sends all requests through
    the defined proxy, using the CONNECT method for HTTPS URLs.

    :param proxy_url:
        The URL of the proxy to be used.

    :param proxy_headers:
        A dictionary containing headers that will be sent to the proxy. In case
        of HTTP they are being sent with each request, while in the
        HTTPS/CONNECT case they are sent only once. Could be used for proxy
        authentication.

    :param proxy_ssl_context:
        The proxy SSL context is used to establish the TLS connection to the
        proxy when using HTTPS proxies.

    :param use_forwarding_for_https:
        (Defaults to False) If set to True will forward requests to the HTTPS
        proxy to be made on behalf of the client instead of creating a TLS
        tunnel via the CONNECT method. **Enabling this flag means that request
        and response headers and content will be visible from the HTTPS proxy**
        whereas tunneling keeps request and response headers and content
        private.  IP address, target hostname, SNI, and port are always visible
        to an HTTPS proxy even when this flag is disabled.

    Example:
        >>> proxy = urllib3.ProxyManager('http://localhost:3128/')
        >>> r1 = proxy.request('GET', 'http://google.com/')
        >>> r2 = proxy.request('GET', 'http://httpbin.org/')
        >>> len(proxy.pools)
        1
        >>> r3 = proxy.request('GET', 'https://httpbin.org/')
        >>> r4 = proxy.request('GET', 'https://twitter.com/')
        >>> len(proxy.pools)
        3

    rXNFc
s�t|t�rd|j|j|jf}t|�}|jdvr:t|j��|jsZt�|jd�}	|j	|	d�}||_
|pfi|_||_t
||�|_|j
|d<|j|d<|j|d<tt|�j||fi|��dS)Nz
%s://%s:%irUrr)rm�_proxyr>�
_proxy_config)r�rr;r<rmrr
rrF�_replacer��
proxy_headers�proxy_ssl_contextr:r��superrr_)
rd�	proxy_urlrer=r�r�r9r`r�rm��	__class__rRrSr_�s(
�





zProxyManager.__init__rVcsD|dkr tt|�j||||d�Stt|�j|jj|jj|jj|d�S)NrW)ru)r�rrvr�r<rmr;)rdr<rmr;rur�rRrSrv�s
�
�z!ProxyManager.connection_from_hostcCs0ddi}t|�j}|r||d<|r,|�|�|S)z�
        Sets headers needed by proxies: specifically, the Accept and Host
        headers. Only sets headers not provided by the user.
        �Acceptz*/*�Host)r�netloc�update)rdr}r=�headers_r�rRrRrS�_set_proxy_headers�s

zProxyManager._set_proxy_headersTcsVt|�}t|j|j|j�s8|�d|j�}|�||�|d<tt	|�j
||fd|i|��S)z@Same as HTTP(S)ConnectionPool.urlopen, ``url`` must be absolute.r=r�)rrr�r�r;rFr=r�r�rr�)rdr�r}r�rpr~r=r�rRrSr�
s
zProxyManager.urlopen)rXNNNF)NrVN)N)T)	r�r�r�r�r_rvr�r��
__classcell__rRrRr�rSr�s*�%

rcKstfd|i|��S)Nr�)r)r}rprRrRrSrsr).�
__future__r�collections�	functools�logging�_collectionsr�connectionpoolrrr�
exceptionsrr	r
rr�packagesr
Zpackages.six.moves.urllib.parser�requestrZ
util.proxyrZ
util.retryrZutil.urlr�__all__�	getLoggerr�r�rn�_key_fields�
namedtupler8�_proxy_config_fieldsr:rT�partialrcrbrrrrRrRrRrS�<module>s:

6

�
ssite-packages/pip/_vendor/urllib3/__pycache__/__init__.cpython-39.pyc000064400000004264150564745010021501 0ustar00a

��Re�
�@s@dZddlmZddlZddlZddlmZddlmZddlm	Z	ddl
mZmZm
Z
dd	lmZdd
lmZmZmZddlmZddlmZdd
lmZddlmZddlmZdZdZe	Z	dZ e�!e"��#e��ej$fdd�Z%[ej&dej'dd�ej&dej(dd�ej&dej)dd�ej&dej*dd�ej+fdd�Z,dS)ze
Python HTTP library with thread-safe connection pooling, file post support, user friendly, and more
�)�absolute_importN)�NullHandler�)�
exceptions)�__version__)�HTTPConnectionPool�HTTPSConnectionPool�connection_from_url)�encode_multipart_formdata)�PoolManager�ProxyManager�proxy_from_url)�HTTPResponse)�make_headers)�Retry)�Timeout)�get_hostz(Andrey Petrov (andrey.petrov@shazow.net)�MIT)rrrrrrr�add_stderr_loggerr	�disable_warningsr
rrr
cCsFt�t�}t��}|�t�d��|�|�|�|�|�dt�|S)z�
    Helper for quickly adding a StreamHandler to the logger. Useful for
    debugging.

    Returns the handler after adding it.
    z%%(asctime)s %(levelname)s %(message)sz,Added a stderr logging handler to logger: %s)	�logging�	getLogger�__name__�
StreamHandler�setFormatter�	Formatter�
addHandler�setLevel�debug)�level�logger�handler�r"��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/__init__.pyr.s	


r�alwaysT)�append�defaultcCst�d|�dS)z<
    Helper for quickly disabling all urllib3 warnings.
    �ignoreN)�warnings�simplefilter)�categoryr"r"r#rQsr)-�__doc__�
__future__rrr(r�r�_versionr�connectionpoolrrr	�filepostr
�poolmanagerrrr
�responserZutil.requestrZ
util.retryrZutil.timeoutrZutil.urlr�
__author__�__license__�__all__rrr�DEBUGrr)�SecurityWarning�SubjectAltNameWarning�InsecurePlatformWarning�SNIMissingWarning�HTTPWarningrr"r"r"r#�<module>s4site-packages/pip/_vendor/urllib3/__pycache__/_version.cpython-39.pyc000064400000000374150564745010021564 0ustar00a

��Re?�@sdZdS)z1.26.7N)�__version__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/_version.py�<module>�site-packages/pip/_vendor/urllib3/__pycache__/connection.cpython-39.pyc000064400000032631150564745010022100 0ustar00a

��RepN�	@sddlmZddlZddlZddlZddlZddlZddlZddlmZ	ddlm
Zddlm
Z
ddlmZddlmZdd	lmZzddlZejZWn*eefy�dZGd
d�de�ZYn0zeZWn"ey�Gdd
�d
e�ZYn0zeZWn$e�yGdd�de�ZYn0ddlmZddl m!Z!ddl"m#Z#m$Z$m%Z%m&Z&ddl'm(Z(m)Z)ddl*m+Z+m,Z,m-Z-ddl.m/Z/m0Z0m1Z1m2Z2m3Z3m4Z4e�5e6�Z7ddd�Z8e�9ddd�Z:e�;d�Z<Gdd�dee=�ZGdd�de�Z>d d!�Z?d"d#�Z@Gd$d%�d%e=�ZAe�seAZ>e>ZBdS)&�)�absolute_importN)�error)�timeout�)�six)�HTTPConnection)�
HTTPException)�create_proxy_ssl_contextc@seZdZdS)�BaseSSLErrorN��__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/connection.pyr
sr
c@seZdZdS)�ConnectionErrorNrrrrrr!src@seZdZdS)�BrokenPipeErrorNrrrrrr*sr)�HTTPHeaderDict��__version__)�ConnectTimeoutError�NewConnectionError�SubjectAltNameWarning�SystemTimeWarning)�CertificateError�match_hostname)�SKIP_HEADER�SKIPPABLE_HEADERS�
connection)�assert_fingerprint�create_urllib3_context�is_ipaddress�resolve_cert_reqs�resolve_ssl_version�ssl_wrap_socket�Pi�)�http�httpsi��z[^-!#$%&'*+.^_`|~0-9a-zA-Z]cs�eZdZdZedZejejdfgZ	dZ
dZdd�Ze
dd	��Zejd
d	��Zdd�Zd
d�Zdd�Zdd�Zdd�Zdd�Zd�fdd�	Zddd�Z�ZS)ra5
    Based on :class:`http.client.HTTPConnection` but provides an extra constructor
    backwards-compatibility layer between older and newer Pythons.

    Additional keyword parameters are used to configure attributes of the connection.
    Accepted parameters include:

    - ``strict``: See the documentation on :class:`urllib3.connectionpool.HTTPConnectionPool`
    - ``source_address``: Set the source address for the current connection.
    - ``socket_options``: Set specific options on the underlying socket. If not specified, then
      defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
      Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.

      For example, if you wish to enable TCP Keep Alive in addition to the defaults,
      you might pass:

      .. code-block:: python

         HTTPConnection.default_socket_options + [
             (socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
         ]

      Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
    r&rFNcOsftjs|�dd�|�d�|_|�d|j�|_|�dd�|_|�dd�|_t	j
|g|�Ri|��dS)N�strict�source_address�socket_options�proxy�proxy_config)r�PY2�pop�getr*�default_socket_optionsr+r,r-�_HTTPConnection�__init__)�self�args�kwrrrr3sszHTTPConnection.__init__cCs|j�d�S)a�
        Getter method to remove any trailing dots that indicate the hostname is an FQDN.

        In general, SSL certificates don't include the trailing dot indicating a
        fully-qualified domain name, and thus, they don't validate properly when
        checked against a domain name that includes the dot. In addition, some
        servers may not expect to receive the trailing dot when provided.

        However, the hostname with trailing dot is critical to DNS resolution; doing a
        lookup with the trailing dot will properly only resolve the appropriate FQDN,
        whereas a lookup without a trailing dot will search the system's search domain
        list. Thus, it's important to keep the original host around for use only in
        those cases where it's appropriate (i.e., when doing DNS lookup to establish the
        actual TCP connection across which we're going to send HTTP requests).
        �.)�	_dns_host�rstrip�r4rrr�host�szHTTPConnection.hostcCs
||_dS)z�
        Setter for the `host` property.

        We assume that only urllib3 uses the _dns_host attribute; httplib itself
        only uses `host`, and it seems reasonable that other libraries follow suit.
        N)r8)r4�valuerrrr;�sc
Cs�i}|jr|j|d<|jr$|j|d<z"tj|j|jf|jfi|��}WnXtynt|d|j	|jf��Yn2t
y�}zt|d|��WYd}~n
d}~00|S)zoEstablish a socket connection and set nodelay settings on it.

        :return: New socket connection.
        r*r+z0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)r*r+r�create_connectionr8�portr�
SocketTimeoutrr;�SocketErrorr)r4�extra_kw�conn�errr�	_new_conn�s.

��

���zHTTPConnection._new_conncCst|dd�S)N�_tunnel_host)�getattrr:rrr�_is_using_tunnel�szHTTPConnection._is_using_tunnelcCs ||_|��r|��d|_dS)Nr)�sockrG�_tunnel�	auto_open�r4rBrrr�
_prepare_conn�szHTTPConnection._prepare_conncCs|��}|�|�dS�N)rDrLrKrrr�connect�szHTTPConnection.connectcOs>t�|�}|r"td||��f��tj|||g|�Ri|��S)� zAMethod cannot contain non-token characters %r (found at least %r))�_CONTAINS_CONTROL_CHAR_RE�search�
ValueError�groupr2�
putrequest)r4�method�urlr5�kwargs�matchrrrrT�s

��zHTTPConnection.putrequestcGs^tdd�|D��s(tj||g|�R�n2t�|���tvrZtdd�t	t
jtt���f��dS)rOcss |]}t|t�o|tkVqdSrM)�
isinstance�strr)�.0�vrrr�	<genexpr>��z+HTTPConnection.putheader.<locals>.<genexpr>z+urllib3.util.SKIP_HEADER only supports '%s'z', 'N)
�anyr2�	putheaderr�
ensure_str�lowerrrR�join�maprZ�title�sorted)r4�header�valuesrrrr`�s��zHTTPConnection.putheadercsN|duri}n|��}ddd�|D�vr2t�|d<tt|�j||||d�dS)N�
user-agentcss|]}t�|���VqdSrM�rrarb�r[�krrrr]�r^z)HTTPConnection.request.<locals>.<genexpr>�
User-Agent)�body�headers)�copy�_get_default_user_agent�superr�request)r4rUrVrnro��	__class__rrrs�s
zHTTPConnection.requestcCs&|pi}tdd�|D��}d|v}d|v}|j||||d�d|vrR|�dt��|��D]\}}	|�||	�qZd|vr�|�d	d
�|��|du�rtjtf}
t	||
�r�|f}|D]`}|s�q�t	|t�s�|�
d�}tt|��d
d�}t
|�
��}
|
d7}
|
|7}
|
d7}
|�|
�q�|�d�dS)z�
        Alternative to the common request method, which sends the
        body with chunked encoding and not as one block
        cSsg|]}t�|����qSrrjrkrrr�
<listcomp>�r^z2HTTPConnection.request_chunked.<locals>.<listcomp>zaccept-encodingr;)�skip_accept_encoding�	skip_hostrirmztransfer-encodingzTransfer-Encoding�chunkedN�utf8�s
s0

)�setrTr`rq�items�
endheadersr�string_types�bytesrY�encode�hex�len�	bytearray�send)r4rUrVrnro�header_keysrwrxrgr<�stringish_types�chunk�len_str�to_sendrrr�request_chunked�s<�



zHTTPConnection.request_chunked)NN)NN)rr
r�__doc__�port_by_scheme�default_port�socket�IPPROTO_TCP�TCP_NODELAYr1�is_verified�proxy_is_verifiedr3�propertyr;�setterrDrGrLrNrTr`rsr��
__classcell__rrrtrrLs$

	


rc	@sjeZdZdZedZdZdZdZdZ	dZ
dZdZddddde
jddfdd�Zd
dd�Zd	d
�Zdd�ZdS)�HTTPSConnectionz�
    Many of the parameters to this constructor are passed to the underlying SSL
    socket by means of :py:func:`urllib3.util.ssl_wrap_socket`.
    r'NFc
KsDtj|||f||d�|
��||_||_||_||_|	|_d|_dS)N)r)rr')rr3�key_file�	cert_file�key_password�ssl_context�server_hostname�	_protocol)r4r;r>r�r�r�r)rr�r�r6rrrr3)szHTTPSConnection.__init__c

Csv|dur$|jdur|jj}ntd�}||_||_||_||_||_||_|oVt	j
�|�|_|oht	j
�|�|_
|	|_dS)zX
        This method should only be called once, before the connection is used.
        N)r��verify_moder"r�r��	cert_reqsr��assert_hostnamer�os�path�
expanduser�ca_certs�ca_cert_dir�ca_cert_data)
r4r�r�r�r�r�r�rr�r�rrr�set_certCs

zHTTPSConnection.set_certc	Cs|��}|j}d}|��rJ|jr0|�||�}d}||_|��d|_|j}|}|j	dur^|j	}t
j��t
k}|r�t�d�t
�t�d}|jdur�d}tt|j�t|j�d�|_|j}t|j�|_|js�|js�|js�|r�t|d�r�|��t||j|j |j!|j|j|j|||d�
|_|�r^|jdu�r^t|jd��r^|j�"�d	v�r^t�d
|j|j�"�ft#�|j$�r~t$|jj%dd�|j$�nd|jt&j'k�r�t(|dd��s�|j)du�r�|j�%�}|�*d
d��s�t�d�|�t+�t,||j)�p�|�|jt&j-k�p�|j$du|_.dS)NFTrzWSystem time is way off (before {0}). This will probably lead to SSL verification errors)�ssl_versionr��load_default_certs)
rH�keyfile�certfiler�r�r�r�r�r��
tls_in_tls�version>�TLSv1zTLSv1.1z�Negotiating TLSv1/TLSv1.1 by default is deprecated and will be disabled in urllib3 v2.0.0. Connecting to '%s' with '%s' can be enabled by explicitly opting-in with 'ssl_version')�binary_form�check_hostname�subjectAltNamer��Certificate for {0} has no `subjectAltName`, falling back to check for a `commonName` for now. This feature is being removed by major browsers and deprecated by RFC 2818. (See https://github.com/urllib3/urllib3/issues/497 for details.))/rDr;rG�tls_in_tls_required�_connect_tls_proxyrHrIrJrEr��datetime�date�today�RECENT_DATE�warnings�warn�formatrr�r r#r�r"r�r�r�r�r��hasattrr�r$r�r�r�r��DeprecationWarningr�getpeercert�ssl�	CERT_NONErFr�r0r�_match_hostname�
CERT_REQUIREDr�)	r4rB�hostnamer�r��is_time_off�default_ssl_context�context�certrrrrNds�
���

���������
�����
�
��
��	�zHTTPSConnection.connectcCs�|j}|j}|rt|||d�St|j|j|j|j|j�}t||j|j|j||d�}|j	t
jkr�t|dd�s�|�
�}|�dd�s�t�d�|�t�t||�|j	t
jk|_|S)zY
        Establish a TLS connection to the proxy using the provided SSL context.
        )rHr�r�)rHr�r�r�r�r�r�Fr�rr�)r-r�r$r	r�r�r�r�r�r�r�r�rFr�r0r�r�r�rr�r�r�)r4r�rBr-r�r�r�rrrr��sJ��
�	���	
z"HTTPSConnection._connect_tls_proxy)	NNNNNNNNN)rr
rr�r�r�r�r�r�r�r�rr�r��_GLOBAL_DEFAULT_TIMEOUTr3r�rNr�rrrrr�s<�
�
!xr�c
Csd|�d�}t|�r|}zt||�Wn:ty^}z"t�d||�||_�WYd}~n
d}~00dS)Nzu[]z@Certificate did not match expected hostname: %s. Certificate: %s)�stripr!rr�log�warning�
_peer_cert)r��asserted_hostname�stripped_hostnamerCrrrr�s
�r�cCsdtS)Nzpython-urllib3/%srrrrrrq+srqc@seZdZdZdS)�DummyConnectionz-Used to detect a failed ConnectionCls import.N)rr
rr�rrrrr�/sr�)C�
__future__rr��loggingr��rer�r�rr@rr?�packagesrZpackages.six.moves.http_clientrr2rZ
util.proxyr	r��SSLErrorr
�ImportError�AttributeError�
BaseExceptionr�	NameError�	Exceptionr�_collectionsr�_versionr�
exceptionsrrrrZpackages.ssl_match_hostnamerr�utilrrrZ	util.ssl_rr r!r"r#r$�	getLoggerrr�r�r�r��compilerP�objectr�r�rqr��VerifiedHTTPSConnectionrrrr�<module>s\
 	


N}site-packages/pip/_vendor/urllib3/exceptions.py000064400000020031150564745010015622 0ustar00from __future__ import absolute_import

from .packages.six.moves.http_client import IncompleteRead as httplib_IncompleteRead

# Base Exceptions


class HTTPError(Exception):
    """Base exception used by this module."""

    pass


class HTTPWarning(Warning):
    """Base warning used by this module."""

    pass


class PoolError(HTTPError):
    """Base exception for errors caused within a pool."""

    def __init__(self, pool, message):
        self.pool = pool
        HTTPError.__init__(self, "%s: %s" % (pool, message))

    def __reduce__(self):
        # For pickling purposes.
        return self.__class__, (None, None)


class RequestError(PoolError):
    """Base exception for PoolErrors that have associated URLs."""

    def __init__(self, pool, url, message):
        self.url = url
        PoolError.__init__(self, pool, message)

    def __reduce__(self):
        # For pickling purposes.
        return self.__class__, (None, self.url, None)


class SSLError(HTTPError):
    """Raised when SSL certificate fails in an HTTPS connection."""

    pass


class ProxyError(HTTPError):
    """Raised when the connection to a proxy fails."""

    def __init__(self, message, error, *args):
        super(ProxyError, self).__init__(message, error, *args)
        self.original_error = error


class DecodeError(HTTPError):
    """Raised when automatic decoding based on Content-Type fails."""

    pass


class ProtocolError(HTTPError):
    """Raised when something unexpected happens mid-request/response."""

    pass


#: Renamed to ProtocolError but aliased for backwards compatibility.
ConnectionError = ProtocolError


# Leaf Exceptions


class MaxRetryError(RequestError):
    """Raised when the maximum number of retries is exceeded.

    :param pool: The connection pool
    :type pool: :class:`~urllib3.connectionpool.HTTPConnectionPool`
    :param string url: The requested Url
    :param exceptions.Exception reason: The underlying error

    """

    def __init__(self, pool, url, reason=None):
        self.reason = reason

        message = "Max retries exceeded with url: %s (Caused by %r)" % (url, reason)

        RequestError.__init__(self, pool, url, message)


class HostChangedError(RequestError):
    """Raised when an existing pool gets a request for a foreign host."""

    def __init__(self, pool, url, retries=3):
        message = "Tried to open a foreign host with url: %s" % url
        RequestError.__init__(self, pool, url, message)
        self.retries = retries


class TimeoutStateError(HTTPError):
    """Raised when passing an invalid state to a timeout"""

    pass


class TimeoutError(HTTPError):
    """Raised when a socket timeout error occurs.

    Catching this error will catch both :exc:`ReadTimeoutErrors
    <ReadTimeoutError>` and :exc:`ConnectTimeoutErrors <ConnectTimeoutError>`.
    """

    pass


class ReadTimeoutError(TimeoutError, RequestError):
    """Raised when a socket timeout occurs while receiving data from a server"""

    pass


# This timeout error does not have a URL attached and needs to inherit from the
# base HTTPError
class ConnectTimeoutError(TimeoutError):
    """Raised when a socket timeout occurs while connecting to a server"""

    pass


class NewConnectionError(ConnectTimeoutError, PoolError):
    """Raised when we fail to establish a new connection. Usually ECONNREFUSED."""

    pass


class EmptyPoolError(PoolError):
    """Raised when a pool runs out of connections and no more are allowed."""

    pass


class ClosedPoolError(PoolError):
    """Raised when a request enters a pool after the pool has been closed."""

    pass


class LocationValueError(ValueError, HTTPError):
    """Raised when there is something wrong with a given URL input."""

    pass


class LocationParseError(LocationValueError):
    """Raised when get_host or similar fails to parse the URL input."""

    def __init__(self, location):
        message = "Failed to parse: %s" % location
        HTTPError.__init__(self, message)

        self.location = location


class URLSchemeUnknown(LocationValueError):
    """Raised when a URL input has an unsupported scheme."""

    def __init__(self, scheme):
        message = "Not supported URL scheme %s" % scheme
        super(URLSchemeUnknown, self).__init__(message)

        self.scheme = scheme


class ResponseError(HTTPError):
    """Used as a container for an error reason supplied in a MaxRetryError."""

    GENERIC_ERROR = "too many error responses"
    SPECIFIC_ERROR = "too many {status_code} error responses"


class SecurityWarning(HTTPWarning):
    """Warned when performing security reducing actions"""

    pass


class SubjectAltNameWarning(SecurityWarning):
    """Warned when connecting to a host with a certificate missing a SAN."""

    pass


class InsecureRequestWarning(SecurityWarning):
    """Warned when making an unverified HTTPS request."""

    pass


class SystemTimeWarning(SecurityWarning):
    """Warned when system time is suspected to be wrong"""

    pass


class InsecurePlatformWarning(SecurityWarning):
    """Warned when certain TLS/SSL configuration is not available on a platform."""

    pass


class SNIMissingWarning(HTTPWarning):
    """Warned when making a HTTPS request without SNI available."""

    pass


class DependencyWarning(HTTPWarning):
    """
    Warned when an attempt is made to import a module with missing optional
    dependencies.
    """

    pass


class ResponseNotChunked(ProtocolError, ValueError):
    """Response needs to be chunked in order to read it as chunks."""

    pass


class BodyNotHttplibCompatible(HTTPError):
    """
    Body should be :class:`http.client.HTTPResponse` like
    (have an fp attribute which returns raw chunks) for read_chunked().
    """

    pass


class IncompleteRead(HTTPError, httplib_IncompleteRead):
    """
    Response length doesn't match expected Content-Length

    Subclass of :class:`http.client.IncompleteRead` to allow int value
    for ``partial`` to avoid creating large objects on streamed reads.
    """

    def __init__(self, partial, expected):
        super(IncompleteRead, self).__init__(partial, expected)

    def __repr__(self):
        return "IncompleteRead(%i bytes read, %i more expected)" % (
            self.partial,
            self.expected,
        )


class InvalidChunkLength(HTTPError, httplib_IncompleteRead):
    """Invalid chunk length in a chunked response."""

    def __init__(self, response, length):
        super(InvalidChunkLength, self).__init__(
            response.tell(), response.length_remaining
        )
        self.response = response
        self.length = length

    def __repr__(self):
        return "InvalidChunkLength(got length %r, %i bytes read)" % (
            self.length,
            self.partial,
        )


class InvalidHeader(HTTPError):
    """The header provided was somehow invalid."""

    pass


class ProxySchemeUnknown(AssertionError, URLSchemeUnknown):
    """ProxyManager does not support the supplied scheme"""

    # TODO(t-8ch): Stop inheriting from AssertionError in v2.0.

    def __init__(self, scheme):
        # 'localhost' is here because our URL parser parses
        # localhost:8080 -> scheme=localhost, remove if we fix this.
        if scheme == "localhost":
            scheme = None
        if scheme is None:
            message = "Proxy URL had no scheme, should start with http:// or https://"
        else:
            message = (
                "Proxy URL had unsupported scheme %s, should use http:// or https://"
                % scheme
            )
        super(ProxySchemeUnknown, self).__init__(message)


class ProxySchemeUnsupported(ValueError):
    """Fetching HTTPS resources through HTTPS proxies is unsupported"""

    pass


class HeaderParsingError(HTTPError):
    """Raised by assert_header_parsing, but we convert it to a log.warning statement."""

    def __init__(self, defects, unparsed_data):
        message = "%s, unparsed data: %r" % (defects or "Unknown", unparsed_data)
        super(HeaderParsingError, self).__init__(message)


class UnrewindableBodyError(HTTPError):
    """urllib3 encountered an error when trying to rewind a body"""

    pass
site-packages/pip/_vendor/urllib3/util/request.py000064400000010033150564745010016107 0ustar00from __future__ import absolute_import

from base64 import b64encode

from ..exceptions import UnrewindableBodyError
from ..packages.six import b, integer_types

# Pass as a value within ``headers`` to skip
# emitting some HTTP headers that are added automatically.
# The only headers that are supported are ``Accept-Encoding``,
# ``Host``, and ``User-Agent``.
SKIP_HEADER = "@@@SKIP_HEADER@@@"
SKIPPABLE_HEADERS = frozenset(["accept-encoding", "host", "user-agent"])

ACCEPT_ENCODING = "gzip,deflate"
try:
    import brotli as _unused_module_brotli  # noqa: F401
except ImportError:
    pass
else:
    ACCEPT_ENCODING += ",br"

_FAILEDTELL = object()


def make_headers(
    keep_alive=None,
    accept_encoding=None,
    user_agent=None,
    basic_auth=None,
    proxy_basic_auth=None,
    disable_cache=None,
):
    """
    Shortcuts for generating request headers.

    :param keep_alive:
        If ``True``, adds 'connection: keep-alive' header.

    :param accept_encoding:
        Can be a boolean, list, or string.
        ``True`` translates to 'gzip,deflate'.
        List will get joined by comma.
        String will be used as provided.

    :param user_agent:
        String representing the user-agent you want, such as
        "python-urllib3/0.6"

    :param basic_auth:
        Colon-separated username:password string for 'authorization: basic ...'
        auth header.

    :param proxy_basic_auth:
        Colon-separated username:password string for 'proxy-authorization: basic ...'
        auth header.

    :param disable_cache:
        If ``True``, adds 'cache-control: no-cache' header.

    Example::

        >>> make_headers(keep_alive=True, user_agent="Batman/1.0")
        {'connection': 'keep-alive', 'user-agent': 'Batman/1.0'}
        >>> make_headers(accept_encoding=True)
        {'accept-encoding': 'gzip,deflate'}
    """
    headers = {}
    if accept_encoding:
        if isinstance(accept_encoding, str):
            pass
        elif isinstance(accept_encoding, list):
            accept_encoding = ",".join(accept_encoding)
        else:
            accept_encoding = ACCEPT_ENCODING
        headers["accept-encoding"] = accept_encoding

    if user_agent:
        headers["user-agent"] = user_agent

    if keep_alive:
        headers["connection"] = "keep-alive"

    if basic_auth:
        headers["authorization"] = "Basic " + b64encode(b(basic_auth)).decode("utf-8")

    if proxy_basic_auth:
        headers["proxy-authorization"] = "Basic " + b64encode(
            b(proxy_basic_auth)
        ).decode("utf-8")

    if disable_cache:
        headers["cache-control"] = "no-cache"

    return headers


def set_file_position(body, pos):
    """
    If a position is provided, move file to that point.
    Otherwise, we'll attempt to record a position for future use.
    """
    if pos is not None:
        rewind_body(body, pos)
    elif getattr(body, "tell", None) is not None:
        try:
            pos = body.tell()
        except (IOError, OSError):
            # This differentiates from None, allowing us to catch
            # a failed `tell()` later when trying to rewind the body.
            pos = _FAILEDTELL

    return pos


def rewind_body(body, body_pos):
    """
    Attempt to rewind body to a certain position.
    Primarily used for request redirects and retries.

    :param body:
        File-like object that supports seek.

    :param int pos:
        Position to seek to in file.
    """
    body_seek = getattr(body, "seek", None)
    if body_seek is not None and isinstance(body_pos, integer_types):
        try:
            body_seek(body_pos)
        except (IOError, OSError):
            raise UnrewindableBodyError(
                "An error occurred when rewinding request body for redirect/retry."
            )
    elif body_pos is _FAILEDTELL:
        raise UnrewindableBodyError(
            "Unable to record file position for rewinding "
            "request body during a redirect/retry."
        )
    else:
        raise ValueError(
            "body_pos must be of type integer, instead it was %s." % type(body_pos)
        )
site-packages/pip/_vendor/urllib3/util/response.py000064400000006666150564745010016276 0ustar00from __future__ import absolute_import

from email.errors import MultipartInvariantViolationDefect, StartBoundaryNotFoundDefect

from ..exceptions import HeaderParsingError
from ..packages.six.moves import http_client as httplib


def is_fp_closed(obj):
    """
    Checks whether a given file-like object is closed.

    :param obj:
        The file-like object to check.
    """

    try:
        # Check `isclosed()` first, in case Python3 doesn't set `closed`.
        # GH Issue #928
        return obj.isclosed()
    except AttributeError:
        pass

    try:
        # Check via the official file-like-object way.
        return obj.closed
    except AttributeError:
        pass

    try:
        # Check if the object is a container for another file-like object that
        # gets released on exhaustion (e.g. HTTPResponse).
        return obj.fp is None
    except AttributeError:
        pass

    raise ValueError("Unable to determine whether fp is closed.")


def assert_header_parsing(headers):
    """
    Asserts whether all headers have been successfully parsed.
    Extracts encountered errors from the result of parsing headers.

    Only works on Python 3.

    :param http.client.HTTPMessage headers: Headers to verify.

    :raises urllib3.exceptions.HeaderParsingError:
        If parsing errors are found.
    """

    # This will fail silently if we pass in the wrong kind of parameter.
    # To make debugging easier add an explicit check.
    if not isinstance(headers, httplib.HTTPMessage):
        raise TypeError("expected httplib.Message, got {0}.".format(type(headers)))

    defects = getattr(headers, "defects", None)
    get_payload = getattr(headers, "get_payload", None)

    unparsed_data = None
    if get_payload:
        # get_payload is actually email.message.Message.get_payload;
        # we're only interested in the result if it's not a multipart message
        if not headers.is_multipart():
            payload = get_payload()

            if isinstance(payload, (bytes, str)):
                unparsed_data = payload
    if defects:
        # httplib is assuming a response body is available
        # when parsing headers even when httplib only sends
        # header data to parse_headers() This results in
        # defects on multipart responses in particular.
        # See: https://github.com/urllib3/urllib3/issues/800

        # So we ignore the following defects:
        # - StartBoundaryNotFoundDefect:
        #     The claimed start boundary was never found.
        # - MultipartInvariantViolationDefect:
        #     A message claimed to be a multipart but no subparts were found.
        defects = [
            defect
            for defect in defects
            if not isinstance(
                defect, (StartBoundaryNotFoundDefect, MultipartInvariantViolationDefect)
            )
        ]

    if defects or unparsed_data:
        raise HeaderParsingError(defects=defects, unparsed_data=unparsed_data)


def is_response_to_head(response):
    """
    Checks whether the request of a response has been a HEAD-request.
    Handles the quirks of AppEngine.

    :param http.client.HTTPResponse response:
        Response to check if the originating request
        used 'HEAD' as a method.
    """
    # FIXME: Can we do this somehow without accessing private httplib _method?
    method = response._method
    if isinstance(method, int):  # Platform-specific: Appengine
        return method == 3
    return method.upper() == "HEAD"
site-packages/pip/_vendor/urllib3/util/ssl_.py000064400000041431150564745010015365 0ustar00from __future__ import absolute_import

import hmac
import os
import sys
import warnings
from binascii import hexlify, unhexlify
from hashlib import md5, sha1, sha256

from ..exceptions import (
    InsecurePlatformWarning,
    ProxySchemeUnsupported,
    SNIMissingWarning,
    SSLError,
)
from ..packages import six
from .url import BRACELESS_IPV6_ADDRZ_RE, IPV4_RE

SSLContext = None
SSLTransport = None
HAS_SNI = False
IS_PYOPENSSL = False
IS_SECURETRANSPORT = False
ALPN_PROTOCOLS = ["http/1.1"]

# Maps the length of a digest to a possible hash function producing this digest
HASHFUNC_MAP = {32: md5, 40: sha1, 64: sha256}


def _const_compare_digest_backport(a, b):
    """
    Compare two digests of equal length in constant time.

    The digests must be of type str/bytes.
    Returns True if the digests match, and False otherwise.
    """
    result = abs(len(a) - len(b))
    for left, right in zip(bytearray(a), bytearray(b)):
        result |= left ^ right
    return result == 0


_const_compare_digest = getattr(hmac, "compare_digest", _const_compare_digest_backport)

try:  # Test for SSL features
    import ssl
    from ssl import CERT_REQUIRED, wrap_socket
except ImportError:
    pass

try:
    from ssl import HAS_SNI  # Has SNI?
except ImportError:
    pass

try:
    from .ssltransport import SSLTransport
except ImportError:
    pass


try:  # Platform-specific: Python 3.6
    from ssl import PROTOCOL_TLS

    PROTOCOL_SSLv23 = PROTOCOL_TLS
except ImportError:
    try:
        from ssl import PROTOCOL_SSLv23 as PROTOCOL_TLS

        PROTOCOL_SSLv23 = PROTOCOL_TLS
    except ImportError:
        PROTOCOL_SSLv23 = PROTOCOL_TLS = 2

try:
    from ssl import PROTOCOL_TLS_CLIENT
except ImportError:
    PROTOCOL_TLS_CLIENT = PROTOCOL_TLS


try:
    from ssl import OP_NO_COMPRESSION, OP_NO_SSLv2, OP_NO_SSLv3
except ImportError:
    OP_NO_SSLv2, OP_NO_SSLv3 = 0x1000000, 0x2000000
    OP_NO_COMPRESSION = 0x20000


try:  # OP_NO_TICKET was added in Python 3.6
    from ssl import OP_NO_TICKET
except ImportError:
    OP_NO_TICKET = 0x4000


# A secure default.
# Sources for more information on TLS ciphers:
#
# - https://wiki.mozilla.org/Security/Server_Side_TLS
# - https://www.ssllabs.com/projects/best-practices/index.html
# - https://hynek.me/articles/hardening-your-web-servers-ssl-ciphers/
#
# The general intent is:
# - prefer cipher suites that offer perfect forward secrecy (DHE/ECDHE),
# - prefer ECDHE over DHE for better performance,
# - prefer any AES-GCM and ChaCha20 over any AES-CBC for better performance and
#   security,
# - prefer AES-GCM over ChaCha20 because hardware-accelerated AES is common,
# - disable NULL authentication, MD5 MACs, DSS, and other
#   insecure ciphers for security reasons.
# - NOTE: TLS 1.3 cipher suites are managed through a different interface
#   not exposed by CPython (yet!) and are enabled by default if they're available.
DEFAULT_CIPHERS = ":".join(
    [
        "ECDHE+AESGCM",
        "ECDHE+CHACHA20",
        "DHE+AESGCM",
        "DHE+CHACHA20",
        "ECDH+AESGCM",
        "DH+AESGCM",
        "ECDH+AES",
        "DH+AES",
        "RSA+AESGCM",
        "RSA+AES",
        "!aNULL",
        "!eNULL",
        "!MD5",
        "!DSS",
    ]
)

try:
    from ssl import SSLContext  # Modern SSL?
except ImportError:

    class SSLContext(object):  # Platform-specific: Python 2
        def __init__(self, protocol_version):
            self.protocol = protocol_version
            # Use default values from a real SSLContext
            self.check_hostname = False
            self.verify_mode = ssl.CERT_NONE
            self.ca_certs = None
            self.options = 0
            self.certfile = None
            self.keyfile = None
            self.ciphers = None

        def load_cert_chain(self, certfile, keyfile):
            self.certfile = certfile
            self.keyfile = keyfile

        def load_verify_locations(self, cafile=None, capath=None, cadata=None):
            self.ca_certs = cafile

            if capath is not None:
                raise SSLError("CA directories not supported in older Pythons")

            if cadata is not None:
                raise SSLError("CA data not supported in older Pythons")

        def set_ciphers(self, cipher_suite):
            self.ciphers = cipher_suite

        def wrap_socket(self, socket, server_hostname=None, server_side=False):
            warnings.warn(
                "A true SSLContext object is not available. This prevents "
                "urllib3 from configuring SSL appropriately and may cause "
                "certain SSL connections to fail. You can upgrade to a newer "
                "version of Python to solve this. For more information, see "
                "https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html"
                "#ssl-warnings",
                InsecurePlatformWarning,
            )
            kwargs = {
                "keyfile": self.keyfile,
                "certfile": self.certfile,
                "ca_certs": self.ca_certs,
                "cert_reqs": self.verify_mode,
                "ssl_version": self.protocol,
                "server_side": server_side,
            }
            return wrap_socket(socket, ciphers=self.ciphers, **kwargs)


def assert_fingerprint(cert, fingerprint):
    """
    Checks if given fingerprint matches the supplied certificate.

    :param cert:
        Certificate as bytes object.
    :param fingerprint:
        Fingerprint as string of hexdigits, can be interspersed by colons.
    """

    fingerprint = fingerprint.replace(":", "").lower()
    digest_length = len(fingerprint)
    hashfunc = HASHFUNC_MAP.get(digest_length)
    if not hashfunc:
        raise SSLError("Fingerprint of invalid length: {0}".format(fingerprint))

    # We need encode() here for py32; works on py2 and p33.
    fingerprint_bytes = unhexlify(fingerprint.encode())

    cert_digest = hashfunc(cert).digest()

    if not _const_compare_digest(cert_digest, fingerprint_bytes):
        raise SSLError(
            'Fingerprints did not match. Expected "{0}", got "{1}".'.format(
                fingerprint, hexlify(cert_digest)
            )
        )


def resolve_cert_reqs(candidate):
    """
    Resolves the argument to a numeric constant, which can be passed to
    the wrap_socket function/method from the ssl module.
    Defaults to :data:`ssl.CERT_REQUIRED`.
    If given a string it is assumed to be the name of the constant in the
    :mod:`ssl` module or its abbreviation.
    (So you can specify `REQUIRED` instead of `CERT_REQUIRED`.
    If it's neither `None` nor a string we assume it is already the numeric
    constant which can directly be passed to wrap_socket.
    """
    if candidate is None:
        return CERT_REQUIRED

    if isinstance(candidate, str):
        res = getattr(ssl, candidate, None)
        if res is None:
            res = getattr(ssl, "CERT_" + candidate)
        return res

    return candidate


def resolve_ssl_version(candidate):
    """
    like resolve_cert_reqs
    """
    if candidate is None:
        return PROTOCOL_TLS

    if isinstance(candidate, str):
        res = getattr(ssl, candidate, None)
        if res is None:
            res = getattr(ssl, "PROTOCOL_" + candidate)
        return res

    return candidate


def create_urllib3_context(
    ssl_version=None, cert_reqs=None, options=None, ciphers=None
):
    """All arguments have the same meaning as ``ssl_wrap_socket``.

    By default, this function does a lot of the same work that
    ``ssl.create_default_context`` does on Python 3.4+. It:

    - Disables SSLv2, SSLv3, and compression
    - Sets a restricted set of server ciphers

    If you wish to enable SSLv3, you can do::

        from pip._vendor.urllib3.util import ssl_
        context = ssl_.create_urllib3_context()
        context.options &= ~ssl_.OP_NO_SSLv3

    You can do the same to enable compression (substituting ``COMPRESSION``
    for ``SSLv3`` in the last line above).

    :param ssl_version:
        The desired protocol version to use. This will default to
        PROTOCOL_SSLv23 which will negotiate the highest protocol that both
        the server and your installation of OpenSSL support.
    :param cert_reqs:
        Whether to require the certificate verification. This defaults to
        ``ssl.CERT_REQUIRED``.
    :param options:
        Specific OpenSSL options. These default to ``ssl.OP_NO_SSLv2``,
        ``ssl.OP_NO_SSLv3``, ``ssl.OP_NO_COMPRESSION``, and ``ssl.OP_NO_TICKET``.
    :param ciphers:
        Which cipher suites to allow the server to select.
    :returns:
        Constructed SSLContext object with specified options
    :rtype: SSLContext
    """
    # PROTOCOL_TLS is deprecated in Python 3.10
    if not ssl_version or ssl_version == PROTOCOL_TLS:
        ssl_version = PROTOCOL_TLS_CLIENT

    context = SSLContext(ssl_version)

    context.set_ciphers(ciphers or DEFAULT_CIPHERS)

    # Setting the default here, as we may have no ssl module on import
    cert_reqs = ssl.CERT_REQUIRED if cert_reqs is None else cert_reqs

    if options is None:
        options = 0
        # SSLv2 is easily broken and is considered harmful and dangerous
        options |= OP_NO_SSLv2
        # SSLv3 has several problems and is now dangerous
        options |= OP_NO_SSLv3
        # Disable compression to prevent CRIME attacks for OpenSSL 1.0+
        # (issue #309)
        options |= OP_NO_COMPRESSION
        # TLSv1.2 only. Unless set explicitly, do not request tickets.
        # This may save some bandwidth on wire, and although the ticket is encrypted,
        # there is a risk associated with it being on wire,
        # if the server is not rotating its ticketing keys properly.
        options |= OP_NO_TICKET

    context.options |= options

    # Enable post-handshake authentication for TLS 1.3, see GH #1634. PHA is
    # necessary for conditional client cert authentication with TLS 1.3.
    # The attribute is None for OpenSSL <= 1.1.0 or does not exist in older
    # versions of Python.  We only enable on Python 3.7.4+ or if certificate
    # verification is enabled to work around Python issue #37428
    # See: https://bugs.python.org/issue37428
    if (cert_reqs == ssl.CERT_REQUIRED or sys.version_info >= (3, 7, 4)) and getattr(
        context, "post_handshake_auth", None
    ) is not None:
        context.post_handshake_auth = True

    def disable_check_hostname():
        if (
            getattr(context, "check_hostname", None) is not None
        ):  # Platform-specific: Python 3.2
            # We do our own verification, including fingerprints and alternative
            # hostnames. So disable it here
            context.check_hostname = False

    # The order of the below lines setting verify_mode and check_hostname
    # matter due to safe-guards SSLContext has to prevent an SSLContext with
    # check_hostname=True, verify_mode=NONE/OPTIONAL. This is made even more
    # complex because we don't know whether PROTOCOL_TLS_CLIENT will be used
    # or not so we don't know the initial state of the freshly created SSLContext.
    if cert_reqs == ssl.CERT_REQUIRED:
        context.verify_mode = cert_reqs
        disable_check_hostname()
    else:
        disable_check_hostname()
        context.verify_mode = cert_reqs

    # Enable logging of TLS session keys via defacto standard environment variable
    # 'SSLKEYLOGFILE', if the feature is available (Python 3.8+). Skip empty values.
    if hasattr(context, "keylog_filename"):
        sslkeylogfile = os.environ.get("SSLKEYLOGFILE")
        if sslkeylogfile:
            context.keylog_filename = sslkeylogfile

    return context


def ssl_wrap_socket(
    sock,
    keyfile=None,
    certfile=None,
    cert_reqs=None,
    ca_certs=None,
    server_hostname=None,
    ssl_version=None,
    ciphers=None,
    ssl_context=None,
    ca_cert_dir=None,
    key_password=None,
    ca_cert_data=None,
    tls_in_tls=False,
):
    """
    All arguments except for server_hostname, ssl_context, and ca_cert_dir have
    the same meaning as they do when using :func:`ssl.wrap_socket`.

    :param server_hostname:
        When SNI is supported, the expected hostname of the certificate
    :param ssl_context:
        A pre-made :class:`SSLContext` object. If none is provided, one will
        be created using :func:`create_urllib3_context`.
    :param ciphers:
        A string of ciphers we wish the client to support.
    :param ca_cert_dir:
        A directory containing CA certificates in multiple separate files, as
        supported by OpenSSL's -CApath flag or the capath argument to
        SSLContext.load_verify_locations().
    :param key_password:
        Optional password if the keyfile is encrypted.
    :param ca_cert_data:
        Optional string containing CA certificates in PEM format suitable for
        passing as the cadata parameter to SSLContext.load_verify_locations()
    :param tls_in_tls:
        Use SSLTransport to wrap the existing socket.
    """
    context = ssl_context
    if context is None:
        # Note: This branch of code and all the variables in it are no longer
        # used by urllib3 itself. We should consider deprecating and removing
        # this code.
        context = create_urllib3_context(ssl_version, cert_reqs, ciphers=ciphers)

    if ca_certs or ca_cert_dir or ca_cert_data:
        try:
            context.load_verify_locations(ca_certs, ca_cert_dir, ca_cert_data)
        except (IOError, OSError) as e:
            raise SSLError(e)

    elif ssl_context is None and hasattr(context, "load_default_certs"):
        # try to load OS default certs; works well on Windows (require Python3.4+)
        context.load_default_certs()

    # Attempt to detect if we get the goofy behavior of the
    # keyfile being encrypted and OpenSSL asking for the
    # passphrase via the terminal and instead error out.
    if keyfile and key_password is None and _is_key_file_encrypted(keyfile):
        raise SSLError("Client private key is encrypted, password is required")

    if certfile:
        if key_password is None:
            context.load_cert_chain(certfile, keyfile)
        else:
            context.load_cert_chain(certfile, keyfile, key_password)

    try:
        if hasattr(context, "set_alpn_protocols"):
            context.set_alpn_protocols(ALPN_PROTOCOLS)
    except NotImplementedError:  # Defensive: in CI, we always have set_alpn_protocols
        pass

    # If we detect server_hostname is an IP address then the SNI
    # extension should not be used according to RFC3546 Section 3.1
    use_sni_hostname = server_hostname and not is_ipaddress(server_hostname)
    # SecureTransport uses server_hostname in certificate verification.
    send_sni = (use_sni_hostname and HAS_SNI) or (
        IS_SECURETRANSPORT and server_hostname
    )
    # Do not warn the user if server_hostname is an invalid SNI hostname.
    if not HAS_SNI and use_sni_hostname:
        warnings.warn(
            "An HTTPS request has been made, but the SNI (Server Name "
            "Indication) extension to TLS is not available on this platform. "
            "This may cause the server to present an incorrect TLS "
            "certificate, which can cause validation failures. You can upgrade to "
            "a newer version of Python to solve this. For more information, see "
            "https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html"
            "#ssl-warnings",
            SNIMissingWarning,
        )

    if send_sni:
        ssl_sock = _ssl_wrap_socket_impl(
            sock, context, tls_in_tls, server_hostname=server_hostname
        )
    else:
        ssl_sock = _ssl_wrap_socket_impl(sock, context, tls_in_tls)
    return ssl_sock


def is_ipaddress(hostname):
    """Detects whether the hostname given is an IPv4 or IPv6 address.
    Also detects IPv6 addresses with Zone IDs.

    :param str hostname: Hostname to examine.
    :return: True if the hostname is an IP address, False otherwise.
    """
    if not six.PY2 and isinstance(hostname, bytes):
        # IDN A-label bytes are ASCII compatible.
        hostname = hostname.decode("ascii")
    return bool(IPV4_RE.match(hostname) or BRACELESS_IPV6_ADDRZ_RE.match(hostname))


def _is_key_file_encrypted(key_file):
    """Detects if a key file is encrypted or not."""
    with open(key_file, "r") as f:
        for line in f:
            # Look for Proc-Type: 4,ENCRYPTED
            if "ENCRYPTED" in line:
                return True

    return False


def _ssl_wrap_socket_impl(sock, ssl_context, tls_in_tls, server_hostname=None):
    if tls_in_tls:
        if not SSLTransport:
            # Import error, ssl is not available.
            raise ProxySchemeUnsupported(
                "TLS in TLS requires support for the 'ssl' module"
            )

        SSLTransport._validate_ssl_context_for_tls_in_tls(ssl_context)
        return SSLTransport(sock, ssl_context, server_hostname)

    if server_hostname:
        return ssl_context.wrap_socket(sock, server_hostname=server_hostname)
    else:
        return ssl_context.wrap_socket(sock)
site-packages/pip/_vendor/urllib3/util/wait.py000064400000012434150564745010015372 0ustar00import errno
import select
import sys
from functools import partial

try:
    from time import monotonic
except ImportError:
    from time import time as monotonic

__all__ = ["NoWayToWaitForSocketError", "wait_for_read", "wait_for_write"]


class NoWayToWaitForSocketError(Exception):
    pass


# How should we wait on sockets?
#
# There are two types of APIs you can use for waiting on sockets: the fancy
# modern stateful APIs like epoll/kqueue, and the older stateless APIs like
# select/poll. The stateful APIs are more efficient when you have a lots of
# sockets to keep track of, because you can set them up once and then use them
# lots of times. But we only ever want to wait on a single socket at a time
# and don't want to keep track of state, so the stateless APIs are actually
# more efficient. So we want to use select() or poll().
#
# Now, how do we choose between select() and poll()? On traditional Unixes,
# select() has a strange calling convention that makes it slow, or fail
# altogether, for high-numbered file descriptors. The point of poll() is to fix
# that, so on Unixes, we prefer poll().
#
# On Windows, there is no poll() (or at least Python doesn't provide a wrapper
# for it), but that's OK, because on Windows, select() doesn't have this
# strange calling convention; plain select() works fine.
#
# So: on Windows we use select(), and everywhere else we use poll(). We also
# fall back to select() in case poll() is somehow broken or missing.

if sys.version_info >= (3, 5):
    # Modern Python, that retries syscalls by default
    def _retry_on_intr(fn, timeout):
        return fn(timeout)


else:
    # Old and broken Pythons.
    def _retry_on_intr(fn, timeout):
        if timeout is None:
            deadline = float("inf")
        else:
            deadline = monotonic() + timeout

        while True:
            try:
                return fn(timeout)
            # OSError for 3 <= pyver < 3.5, select.error for pyver <= 2.7
            except (OSError, select.error) as e:
                # 'e.args[0]' incantation works for both OSError and select.error
                if e.args[0] != errno.EINTR:
                    raise
                else:
                    timeout = deadline - monotonic()
                    if timeout < 0:
                        timeout = 0
                    if timeout == float("inf"):
                        timeout = None
                    continue


def select_wait_for_socket(sock, read=False, write=False, timeout=None):
    if not read and not write:
        raise RuntimeError("must specify at least one of read=True, write=True")
    rcheck = []
    wcheck = []
    if read:
        rcheck.append(sock)
    if write:
        wcheck.append(sock)
    # When doing a non-blocking connect, most systems signal success by
    # marking the socket writable. Windows, though, signals success by marked
    # it as "exceptional". We paper over the difference by checking the write
    # sockets for both conditions. (The stdlib selectors module does the same
    # thing.)
    fn = partial(select.select, rcheck, wcheck, wcheck)
    rready, wready, xready = _retry_on_intr(fn, timeout)
    return bool(rready or wready or xready)


def poll_wait_for_socket(sock, read=False, write=False, timeout=None):
    if not read and not write:
        raise RuntimeError("must specify at least one of read=True, write=True")
    mask = 0
    if read:
        mask |= select.POLLIN
    if write:
        mask |= select.POLLOUT
    poll_obj = select.poll()
    poll_obj.register(sock, mask)

    # For some reason, poll() takes timeout in milliseconds
    def do_poll(t):
        if t is not None:
            t *= 1000
        return poll_obj.poll(t)

    return bool(_retry_on_intr(do_poll, timeout))


def null_wait_for_socket(*args, **kwargs):
    raise NoWayToWaitForSocketError("no select-equivalent available")


def _have_working_poll():
    # Apparently some systems have a select.poll that fails as soon as you try
    # to use it, either due to strange configuration or broken monkeypatching
    # from libraries like eventlet/greenlet.
    try:
        poll_obj = select.poll()
        _retry_on_intr(poll_obj.poll, 0)
    except (AttributeError, OSError):
        return False
    else:
        return True


def wait_for_socket(*args, **kwargs):
    # We delay choosing which implementation to use until the first time we're
    # called. We could do it at import time, but then we might make the wrong
    # decision if someone goes wild with monkeypatching select.poll after
    # we're imported.
    global wait_for_socket
    if _have_working_poll():
        wait_for_socket = poll_wait_for_socket
    elif hasattr(select, "select"):
        wait_for_socket = select_wait_for_socket
    else:  # Platform-specific: Appengine.
        wait_for_socket = null_wait_for_socket
    return wait_for_socket(*args, **kwargs)


def wait_for_read(sock, timeout=None):
    """Waits for reading to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    """
    return wait_for_socket(sock, read=True, timeout=timeout)


def wait_for_write(sock, timeout=None):
    """Waits for writing to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    """
    return wait_for_socket(sock, write=True, timeout=timeout)
site-packages/pip/_vendor/urllib3/util/__pycache__/request.cpython-39.pyc000064400000006643150564745010022412 0ustar00a

��Re�@s�ddlmZddlmZddlmZddlmZmZdZ	e
gd��ZdZzdd	l
ZWneyfYn
0ed
7Ze�Zddd�Zd
d�Zdd�Zd	S)�)�absolute_import)�	b64encode�)�UnrewindableBodyError)�b�
integer_typesz@@@SKIP_HEADER@@@)�accept-encoding�host�
user-agentzgzip,deflateNz,brcCs�i}|r6t|t�rnt|t�r*d�|�}nt}||d<|rB||d<|rNd|d<|rldtt|���d�|d<|r�dtt|���d�|d	<|r�d
|d<|S)a
    Shortcuts for generating request headers.

    :param keep_alive:
        If ``True``, adds 'connection: keep-alive' header.

    :param accept_encoding:
        Can be a boolean, list, or string.
        ``True`` translates to 'gzip,deflate'.
        List will get joined by comma.
        String will be used as provided.

    :param user_agent:
        String representing the user-agent you want, such as
        "python-urllib3/0.6"

    :param basic_auth:
        Colon-separated username:password string for 'authorization: basic ...'
        auth header.

    :param proxy_basic_auth:
        Colon-separated username:password string for 'proxy-authorization: basic ...'
        auth header.

    :param disable_cache:
        If ``True``, adds 'cache-control: no-cache' header.

    Example::

        >>> make_headers(keep_alive=True, user_agent="Batman/1.0")
        {'connection': 'keep-alive', 'user-agent': 'Batman/1.0'}
        >>> make_headers(accept_encoding=True)
        {'accept-encoding': 'gzip,deflate'}
    �,rr
z
keep-alive�
connectionzBasic zutf-8�
authorizationzproxy-authorizationzno-cachez
cache-control)�
isinstance�str�list�join�ACCEPT_ENCODINGrr�decode)�
keep_alive�accept_encoding�
user_agent�
basic_auth�proxy_basic_auth�
disable_cache�headers�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/request.py�make_headerss.*

��
rc	CsP|durt||�n8t|dd�durLz|��}WnttfyJt}Yn0|S)z
    If a position is provided, move file to that point.
    Otherwise, we'll attempt to record a position for future use.
    N�tell)�rewind_body�getattrr�IOError�OSError�_FAILEDTELL)�body�posrrr�set_file_positionbs
r&c	Csrt|dd�}|durLt|t�rLz||�WqnttfyHtd��Yqn0n"|tur^td��ntdt|���dS)z�
    Attempt to rewind body to a certain position.
    Primarily used for request redirects and retries.

    :param body:
        File-like object that supports seek.

    :param int pos:
        Position to seek to in file.
    �seekNzAAn error occurred when rewinding request body for redirect/retry.zRUnable to record file position for rewinding request body during a redirect/retry.z4body_pos must be of type integer, instead it was %s.)	r rrr!r"rr#�
ValueError�type)r$�body_pos�	body_seekrrrrts��
�r)NNNNNN)�
__future__r�base64r�
exceptionsrZpackages.sixrr�SKIP_HEADER�	frozenset�SKIPPABLE_HEADERSr�brotliZ_unused_module_brotli�ImportError�objectr#rr&rrrrr�<module>s*�
Hsite-packages/pip/_vendor/urllib3/util/__pycache__/wait.cpython-39.pyc000064400000006143150564745010021661 0ustar00a

��Re�@s�ddlZddlZddlZddlmZzddlmZWneyRddlmZYn0gd�ZGdd�de	�Z
ejdkr�d	d
�Zndd
�Zdd
d�Z
ddd�Zdd�Zdd�Zdd�addd�Zddd�ZdS)�N)�partial)�	monotonic)�time)�NoWayToWaitForSocketError�
wait_for_read�wait_for_writec@seZdZdS)rN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/wait.pyrsr)��cCs||�S)Nr)�fn�timeoutrrr�_retry_on_intr*src
Cs�|durtd�}n
t�|}z
||�WSttjfy�}zR|jdtjkrP�n2|t�}|dkrfd}|td�krvd}WYd}~qWYd}~qd}~00qdS)N�infr)�floatr�OSError�select�error�args�errnoZEINTR)rr�deadline�errrr0s



Fc
Csd|s|std��g}g}|r&|�|�|r4|�|�ttj|||�}t||�\}}}	t|p`|p`|	�S)N�2must specify at least one of read=True, write=True)�RuntimeError�appendrrr�bool)
�sock�read�writer�rcheck�wcheckr�rready�wready�xreadyrrr�select_wait_for_socketGs

r'cs^|s|std��d}|r"|tjO}|r0|tjO}t�����||��fdd�}tt||��S)Nrrcs|dur|d9}��|�S)Ni�)�poll)�t��poll_objrr�do_pollfsz%poll_wait_for_socket.<locals>.do_poll)rr�POLLIN�POLLOUTr(�registerrr)rr r!r�maskr,rr*r�poll_wait_for_socketZs

r1cOstd��dS)Nzno select-equivalent available)r�r�kwargsrrr�null_wait_for_socketnsr4c	Cs:zt��}t|jd�Wnttfy0YdS0dSdS)NrFT)rr(r�AttributeErrorrr*rrr�_have_working_pollrsr6cOs.t�rtanttd�rtantat|i|��S)Nr)r6r1�wait_for_socket�hasattrrr'r4r2rrrr7s
r7cCst|d|d�S)z�Waits for reading to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)r r�r7�rrrrrr�srcCst|d|d�S)z�Waits for writing to be available on a given socket.
    Returns True if the socket is readable, or False if the timeout expired.
    T)r!rr9r:rrrr�sr)FFN)FFN)N)N)rr�sys�	functoolsrrr�ImportError�__all__�	Exceptionr�version_inforr'r1r4r6r7rrrrrr�<module>s&





site-packages/pip/_vendor/urllib3/util/__pycache__/retry.cpython-39.pyc000064400000037003150564745010022061 0ustar00a

��Re�S�@s�ddlmZddlZddlZddlZddlZddlZddlmZddl	m
Z
ddlmZm
Z
mZmZmZmZmZddlmZe�e�Zedgd	��Ze�ZGd
d�de�Ze�e�Gdd
�d
e��Zed�e_dS)�)�absolute_importN)�
namedtuple)�	takewhile�)�ConnectTimeoutError�
InvalidHeader�
MaxRetryError�
ProtocolError�
ProxyError�ReadTimeoutError�
ResponseError)�six�RequestHistory)�method�url�error�status�redirect_locationc@s@eZdZedd��Zejdd��Zedd��Zejdd��ZdS)�
_RetryMetacCst�dt�|jS�Nz}Using 'Retry.DEFAULT_METHOD_WHITELIST' is deprecated and will be removed in v2.0. Use 'Retry.DEFAULT_ALLOWED_METHODS' instead��warnings�warn�DeprecationWarning�DEFAULT_ALLOWED_METHODS��cls�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/retry.py�DEFAULT_METHOD_WHITELIST$s
�z#_RetryMeta.DEFAULT_METHOD_WHITELISTcCst�dt�||_dSrr�r�valuerrrr-s
�cCst�dt�|jS�Nz�Using 'Retry.DEFAULT_REDIRECT_HEADERS_BLACKLIST' is deprecated and will be removed in v2.0. Use 'Retry.DEFAULT_REMOVE_HEADERS_ON_REDIRECT' instead�rrr�"DEFAULT_REMOVE_HEADERS_ON_REDIRECTrrrr�"DEFAULT_REDIRECT_HEADERS_BLACKLIST6s
�z-_RetryMeta.DEFAULT_REDIRECT_HEADERS_BLACKLISTcCst�dt�||_dSr"r#r rrrr%?s
�N)�__name__�
__module__�__qualname__�propertyr�setterr%rrrrr#s


rcs�eZdZdZegd��Zegd��Zedg�ZdZdddddde	ddd	d	dd	e	e	fd
d�Z
dd
�Zed-dd��Z
dd�Zdd�Zdd�Zd.dd�Zdd�Zd/dd�Zdd�Zdd�Zd d!�Zd0d#d$�Zd%d&�Zd1d'd(�Zd)d*�Z�fd+d,�Z�ZS)2�RetryaBRetry configuration.

    Each retry attempt will create a new Retry object with updated values, so
    they can be safely reused.

    Retries can be defined as a default for a pool::

        retries = Retry(connect=5, read=2, redirect=5)
        http = PoolManager(retries=retries)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', retries=Retry(10))

    Retries can be disabled by passing ``False``::

        response = http.request('GET', 'http://example.com/', retries=False)

    Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
    retries are disabled, in which case the causing exception will be raised.

    :param int total:
        Total number of retries to allow. Takes precedence over other counts.

        Set to ``None`` to remove this constraint and fall back on other
        counts.

        Set to ``0`` to fail on the first retry.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int connect:
        How many connection-related errors to retry on.

        These are errors raised before the request is sent to the remote server,
        which we assume has not triggered the server to process the request.

        Set to ``0`` to fail on the first retry of this type.

    :param int read:
        How many times to retry on read errors.

        These errors are raised after the request was sent to the server, so the
        request may have side-effects.

        Set to ``0`` to fail on the first retry of this type.

    :param int redirect:
        How many redirects to perform. Limit this to avoid infinite redirect
        loops.

        A redirect is a HTTP response with a status code 301, 302, 303, 307 or
        308.

        Set to ``0`` to fail on the first retry of this type.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int status:
        How many times to retry on bad status codes.

        These are retries made on responses, where status code matches
        ``status_forcelist``.

        Set to ``0`` to fail on the first retry of this type.

    :param int other:
        How many times to retry on other errors.

        Other errors are errors that are not connect, read, redirect or status errors.
        These errors might be raised after the request was sent to the server, so the
        request might have side-effects.

        Set to ``0`` to fail on the first retry of this type.

        If ``total`` is not set, it's a good idea to set this to 0 to account
        for unexpected edge cases and avoid infinite retry loops.

    :param iterable allowed_methods:
        Set of uppercased HTTP method verbs that we should retry on.

        By default, we only retry on methods which are considered to be
        idempotent (multiple requests with the same parameters end with the
        same state). See :attr:`Retry.DEFAULT_ALLOWED_METHODS`.

        Set to a ``False`` value to retry on any verb.

        .. warning::

            Previously this parameter was named ``method_whitelist``, that
            usage is deprecated in v1.26.0 and will be removed in v2.0.

    :param iterable status_forcelist:
        A set of integer HTTP status codes that we should force a retry on.
        A retry is initiated if the request method is in ``allowed_methods``
        and the response status code is in ``status_forcelist``.

        By default, this is disabled with ``None``.

    :param float backoff_factor:
        A backoff factor to apply between attempts after the second try
        (most errors are resolved immediately by a second try without a
        delay). urllib3 will sleep for::

            {backoff factor} * (2 ** ({number of total retries} - 1))

        seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
        for [0.0s, 0.2s, 0.4s, ...] between retries. It will never be longer
        than :attr:`Retry.BACKOFF_MAX`.

        By default, backoff is disabled (set to 0).

    :param bool raise_on_redirect: Whether, if the number of redirects is
        exhausted, to raise a MaxRetryError, or to return a response with a
        response code in the 3xx range.

    :param bool raise_on_status: Similar meaning to ``raise_on_redirect``:
        whether we should raise an exception, or return a response,
        if status falls in ``status_forcelist`` range and retries have
        been exhausted.

    :param tuple history: The history of the request encountered during
        each call to :meth:`~Retry.increment`. The list is in the order
        the requests occurred. Each list item is of class :class:`RequestHistory`.

    :param bool respect_retry_after_header:
        Whether to respect Retry-After header on status codes defined as
        :attr:`Retry.RETRY_AFTER_STATUS_CODES` or not.

    :param iterable remove_headers_on_redirect:
        Sequence of headers to remove from the request when a response
        indicating a redirect is returned before firing off the redirected
        request.
    )�HEAD�GET�PUT�DELETE�OPTIONS�TRACE)i�i�i��
Authorization�x�
NrTcCs�|tur,|turtd��tjdtdd�|}|tur:|j}|turH|j}||_||_||_	||_
||_|dusv|dur~d}d}
||_|p�t
�|_||_|	|_|
|_||_|p�t�|_|
|_tdd�|D��|_dS)	NzoUsing both 'allowed_methods' and 'method_whitelist' together is not allowed. Instead only use 'allowed_methods'�lUsing 'method_whitelist' with Retry is deprecated and will be removed in v2.0. Use 'allowed_methods' insteadr)�
stacklevelFrcSsg|]}|���qSr)�lower)�.0�hrrr�
<listcomp>�z"Retry.__init__.<locals>.<listcomp>)�_Default�
ValueErrorrrrrr$�total�connect�readr�other�redirect�set�status_forcelist�allowed_methods�backoff_factor�raise_on_redirect�raise_on_status�tuple�history�respect_retry_after_header�	frozenset�remove_headers_on_redirect)�selfr>r?r@rBrrArErDrFrGrHrJrKrM�method_whitelistrrr�__init__�sD���zRetry.__init__cKs�t|j|j|j|j|j|j|j|j|j	|j
|j|j|j
d�
}d|vrxd|vrxd|jvrnt�dt�|j|d<n
|j|d<|�|�t|�fi|��S)N)
r>r?r@rBrrArDrFrGrHrJrMrKrOrEr5)�dictr>r?r@rBrrArDrFrGrHrJrMrK�__dict__rrrrE�update�type)rN�kw�paramsrrr�news2�
�

z	Retry.newcCsR|dur|dur|n|j}t|t�r(|St|�o2d}|||d�}t�d||�|S)z3Backwards-compatibility for the old retries format.N)rBz!Converted retries value: %r -> %r)�DEFAULT�
isinstancer+�bool�log�debug)r�retriesrB�default�new_retriesrrr�from_intBs
zRetry.from_intcCsFtttdd�t|j����}|dkr(dS|jd|d}t|j|�S)zIFormula for computing the current backoff

        :rtype: float
        cSs
|jduS)N)r)�xrrr�<lambda>Xr;z(Retry.get_backoff_time.<locals>.<lambda>�rr)�len�listr�reversedrJrF�min�BACKOFF_MAX)rN�consecutive_errors_len�
backoff_valuerrr�get_backoff_timePs��zRetry.get_backoff_timecCs�t�d|�rt|�}n`tj�|�}|dur6td|��|ddur^|dd�d|dd�}tj�|�}|t��}|dkr�d}|S)Nz^\s*[0-9]+\s*$zInvalid Retry-After header: %s�	)rr4r)	�re�match�int�email�utils�parsedate_tzr�	mktime_tz�time)rN�retry_after�seconds�retry_date_tuple�
retry_daterrr�parse_retry_afteras
zRetry.parse_retry_aftercCs |�d�}|durdS|�|�S)z(Get the value of Retry-After in seconds.zRetry-AfterN)�	getheaderry�rN�responserurrr�get_retry_afterxs
zRetry.get_retry_aftercCs |�|�}|rt�|�dSdS)NTF)r}rt�sleepr{rrr�sleep_for_retry�s


zRetry.sleep_for_retrycCs"|��}|dkrdSt�|�dS)Nr)rkrtr~)rN�backoffrrr�_sleep_backoff�szRetry._sleep_backoffcCs(|jr|r|�|�}|rdS|��dS)aBSleep between retry attempts.

        This method will respect a server's ``Retry-After`` response header
        and sleep the duration of the time requested. If that is not present, it
        will use an exponential backoff. By default, the backoff factor is 0 and
        this method will return immediately.
        N)rKrr�)rNr|�sleptrrrr~�s
	

zRetry.sleepcCst|t�r|j}t|t�S)zzErrors when we're fairly sure that the server did not receive the
        request, so it should be safe to retry.
        )rYr
�original_errorr�rN�errrrr�_is_connection_error�s
zRetry._is_connection_errorcCst|ttf�S)zErrors that occur after the request has been started, so we should
        assume that the server began processing it.
        )rYrr	r�rrr�_is_read_error�szRetry._is_read_errorcCs<d|jvrt�dt�|j}n|j}|r8|��|vr8dSdS)zyChecks if a given HTTP method should be retried upon, depending if
        it is included in the allowed_methods
        rOr5FT)rRrrrrOrE�upper)rNrrErrr�_is_method_retryable�s
�zRetry._is_method_retryableFcCs<|�|�sdS|jr"||jvr"dS|jo:|jo:|o:||jvS)awIs this method/status code retryable? (Based on allowlists and control
        variables such as the number of total retries to allow, whether to
        respect the Retry-After header, whether this header is present, and
        whether the returned status code is on the list of status codes to
        be retried upon on the presence of the aforementioned header)
        FT)r�rDr>rK�RETRY_AFTER_STATUS_CODES)rNr�status_code�has_retry_afterrrr�is_retry�s
���zRetry.is_retrycCs>|j|j|j|j|j|jf}ttd|��}|s2dSt|�dkS)zAre we out of retries?NFr)	r>r?r@rBrrAre�filterrg)rN�retry_countsrrr�is_exhausted�s�zRetry.is_exhaustedc	Cs�|jdur |r t�t|�||��|j}|dur6|d8}|j}|j}	|j}
|j}|j}d}
d}d}|r�|�	|�r�|dur�t�t|�||��n|dur�|d8}n�|r�|�
|�r�|	dus�|�|�s�t�t|�||��n|	dur�|	d8}	n�|r�|dur�|d8}nr|�r.|���r.|
du�r|
d8}
d}
|��}|j}n<t
j}
|�rj|j�rj|du�rT|d8}t
jj|jd�}
|j}|jt|||||�f}|j|||	|
|||d�}|���r�t|||�p�t
|
���t�d||�|S)	a�Return a new Retry object with incremented retry counters.

        :param response: A response object, or None, if the server did not
            return a response.
        :type response: :class:`~urllib3.response.HTTPResponse`
        :param Exception error: An error encountered during the request, or
            None if the response was received successfully.

        :return: A new ``Retry`` object.
        FNrc�unknownztoo many redirects)r�)r>r?r@rBrrArJz$Incremented Retry for (url='%s'): %r)r>r
�reraiserTr?r@rBrrAr�r�r��get_redirect_locationr�
GENERIC_ERROR�SPECIFIC_ERROR�formatrJrrWr�rr[r\)rNrrr|r�_pool�_stacktracer>r?r@rB�status_countrA�causerrrJ�	new_retryrrr�	increment�sl




��

zRetry.incrementcCsdjt|�|d�S)Nz|{cls.__name__}(total={self.total}, connect={self.connect}, read={self.read}, redirect={self.redirect}, status={self.status}))rrN)r�rT)rNrrr�__repr__Ds��zRetry.__repr__csN|dkrt�dt�|jSzttt|�|�WStyHtt|�YS0dS)NrOr5)rrrrE�getattr�superr+�AttributeError)rN�item��	__class__rr�__getattr__Js�zRetry.__getattr__)TN)N)N)F)NNNNNN)r&r'r(�__doc__rLrr�r$rhr<rPrW�classmethodr`rkryr}rr�r~r�r�r�r�r�r�r�r��
__classcell__rrr�rr+Is^
�
�
=$




�
^r+�) �
__future__rrp�loggingrmrtr�collectionsr�	itertoolsr�
exceptionsrrrr	r
rr�packagesr
�	getLoggerr&r[r�objectr<rTr�
add_metaclassr+rXrrrr�<module>s,$	
�&site-packages/pip/_vendor/urllib3/util/__pycache__/queue.cpython-39.pyc000064400000002117150564745010022036 0ustar00a

��Re��@sDddlZddlmZddlmZejr.ddlZGdd�dej�ZdS)�N�)�six��queuec@s0eZdZdd�Zefdd�Zdd�Zdd�Zd	S)
�	LifoQueuecCst��|_dS�N)�collections�dequer)�self�_�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/queue.py�_initszLifoQueue._initcCs
||j�Srr)r
�lenrrr
�_qsizeszLifoQueue._qsizecCs|j�|�dSr)r�append)r
�itemrrr
�_putszLifoQueue._putcCs
|j��Sr)r�pop)r
rrr
�_getszLifoQueue._getN)�__name__�
__module__�__qualname__rrrrrrrrr
rsr)	r�packagesrZpackages.six.movesr�PY2�QueueZ_unused_module_Queuerrrrr
�<module>s
site-packages/pip/_vendor/urllib3/util/__pycache__/proxy.cpython-39.pyc000064400000002550150564745010022074 0ustar00a

��ReE�@s,ddlmZmZmZddd�Zddd�ZdS)	�)�create_urllib3_context�resolve_cert_reqs�resolve_ssl_versionNcCs4|durdS|dkrdS|jdkr0|r0|jr0dSdS)a?
    Returns True if the connection requires an HTTP CONNECT through the proxy.

    :param URL proxy_url:
        URL of the proxy.
    :param ProxyConfig proxy_config:
        Proxy configuration from poolmanager.py
    :param str destination_scheme:
        The scheme of the destination. (i.e https, http, etc)
    NF�http�httpsT)�scheme�use_forwarding_for_https)�	proxy_url�proxy_config�destination_scheme�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/proxy.py�connection_requires_http_tunnels���rcCs6tt|�t|�d�}|s2|s2|s2t|d�r2|��|S)z\
    Generates a default proxy ssl context if one hasn't been provided by the
    user.
    )�ssl_version�	cert_reqs�load_default_certs)rrr�hasattrr)rr�ca_certs�ca_cert_dir�ca_cert_data�ssl_contextrrr
�create_proxy_ssl_context%s�����r)NNN)NNN)�ssl_rrrrrrrrr
�<module>s�
"�site-packages/pip/_vendor/urllib3/util/__pycache__/response.cpython-39.pyc000064400000004524150564745010022554 0ustar00a

��Re�
�@sPddlmZddlmZmZddlmZddlmZ	dd�Z
dd	�Zd
d�ZdS)
�)�absolute_import)�!MultipartInvariantViolationDefect�StartBoundaryNotFoundDefect�)�HeaderParsingError)�http_clientcCsfz
|��WStyYn0z|jWSty8Yn0z|jduWStyXYn0td��dS)zt
    Checks whether a given file-like object is closed.

    :param obj:
        The file-like object to check.
    Nz)Unable to determine whether fp is closed.)�isclosed�AttributeError�closed�fp�
ValueError)�obj�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/response.py�is_fp_closed	s
rcCs�t|tj�std�t|����t|dd�}t|dd�}d}|r^|��s^|�}t|tt	f�r^|}|rpdd�|D�}|sx|r�t
||d��dS)a>
    Asserts whether all headers have been successfully parsed.
    Extracts encountered errors from the result of parsing headers.

    Only works on Python 3.

    :param http.client.HTTPMessage headers: Headers to verify.

    :raises urllib3.exceptions.HeaderParsingError:
        If parsing errors are found.
    z"expected httplib.Message, got {0}.�defectsN�get_payloadcSsg|]}t|ttf�s|�qSr)�
isinstancerr)�.0�defectrrr�
<listcomp>Rs
��z)assert_header_parsing.<locals>.<listcomp>)r�
unparsed_data)r�httplib�HTTPMessage�	TypeError�format�type�getattr�is_multipart�bytes�strr)�headersrrr�payloadrrr�assert_header_parsing(s �r#cCs$|j}t|t�r|dkS|��dkS)z�
    Checks whether the request of a response has been a HEAD-request.
    Handles the quirks of AppEngine.

    :param http.client.HTTPResponse response:
        Response to check if the originating request
        used 'HEAD' as a method.
    ��HEAD)�_methodr�int�upper)�response�methodrrr�is_response_to_head^s

r+N)
�
__future__rZemail.errorsrr�
exceptionsrZpackages.six.movesrrrr#r+rrrr�<module>s6site-packages/pip/_vendor/urllib3/util/__pycache__/ssl_.cpython-39.pyc000064400000026154150564745010021661 0ustar00a

��ReC�@s�ddlmZddlZddlZddlZddlZddlmZmZddl	m
Z
mZmZddl
mZmZmZmZddlmZdd	lmZmZdZdZd
Zd
Zd
ZdgZe
eed�Zd
d�Zeede�Z zddl!Z!ddl!m"Z"m#Z#Wne$y�Yn0zddl!mZWne$�yYn0zddl%mZWne$�y8Yn0zddl!m&Z&e&Z'WnFe$�y�zddl!m'Z&e&Z'Wne$�y�dZ'Z&Yn0Yn0zddl!m(Z(Wne$�y�e&Z(Yn0zddl!m)Z)m*Z*m+Z+Wn e$�y�d\Z*Z+dZ)Yn0zddl!m,Z,Wne$�y"dZ,Yn0d�-gd��Z.zddl!mZWn$e$�yfGdd�de/�ZYn0d d!�Z0d"d#�Z1d$d%�Z2d0d&d'�Z3d1d(d)�Z4d*d+�Z5d,d-�Z6d2d.d/�Z7dS)3�)�absolute_importN)�hexlify�	unhexlify)�md5�sha1�sha256�)�InsecurePlatformWarning�ProxySchemeUnsupported�SNIMissingWarning�SSLError)�six�)�BRACELESS_IPV6_ADDRZ_RE�IPV4_REFzhttp/1.1)� �(�@cCsDtt|�t|��}tt|�t|��D]\}}|||AO}q&|dkS)z�
    Compare two digests of equal length in constant time.

    The digests must be of type str/bytes.
    Returns True if the digests match, and False otherwise.
    r)�abs�len�zip�	bytearray)�a�b�result�left�right�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/ssl_.py�_const_compare_digest_backportsr�compare_digest)�
CERT_REQUIRED�wrap_socket)�HAS_SNI)�SSLTransport)�PROTOCOL_TLS)�PROTOCOL_SSLv23)�PROTOCOL_TLS_CLIENT)�OP_NO_COMPRESSION�OP_NO_SSLv2�OP_NO_SSLv3)iii)�OP_NO_TICKETi@�:)zECDHE+AESGCMzECDHE+CHACHA20z
DHE+AESGCMzDHE+CHACHA20zECDH+AESGCMz	DH+AESGCMzECDH+AESzDH+AESz
RSA+AESGCMzRSA+AESz!aNULLz!eNULLz!MD5z!DSS)�
SSLContextc@s8eZdZdd�Zdd�Zd
dd�Zdd	�Zddd�ZdS)r-cCs6||_d|_tj|_d|_d|_d|_d|_d|_	dS)NFr)
�protocol�check_hostname�ssl�	CERT_NONE�verify_mode�ca_certs�options�certfile�keyfile�ciphers)�selfZprotocol_versionrrr�__init__�szSSLContext.__init__cCs||_||_dS�N)r5r6)r8r5r6rrr�load_cert_chain�szSSLContext.load_cert_chainNcCs*||_|durtd��|dur&td��dS)Nz-CA directories not supported in older Pythonsz&CA data not supported in older Pythons)r3r)r8�cafile�capath�cadatarrr�load_verify_locations�s
z SSLContext.load_verify_locationscCs
||_dSr:�r7)r8Zcipher_suiterrr�set_ciphers�szSSLContext.set_ciphersFcCs>t�dt�|j|j|j|j|j|d�}t|fd|j	i|��S)Na2A true SSLContext object is not available. This prevents urllib3 from configuring SSL appropriately and may cause certain SSL connections to fail. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html#ssl-warnings)r6r5r3�	cert_reqs�ssl_version�server_sider7)
�warnings�warnr	r6r5r3r2r.r"r7)r8�socket�server_hostnamerD�kwargsrrrr"�s�
�zSSLContext.wrap_socket)NNN)NF)�__name__�
__module__�__qualname__r9r;r?rAr"rrrrr-�s

	r-cCsn|�dd���}t|�}t�|�}|s4td�|���t|���}||��	�}t
||�sjtd�|t|����dS)z�
    Checks if given fingerprint matches the supplied certificate.

    :param cert:
        Certificate as bytes object.
    :param fingerprint:
        Fingerprint as string of hexdigits, can be interspersed by colons.
    r,�z"Fingerprint of invalid length: {0}z6Fingerprints did not match. Expected "{0}", got "{1}".N)�replace�lowerr�HASHFUNC_MAP�getr�formatr�encode�digest�_const_compare_digestr)�cert�fingerprint�
digest_length�hashfunc�fingerprint_bytes�cert_digestrrr�assert_fingerprint�s


��r\cCs@|durtSt|t�r<tt|d�}|dur8ttd|�}|S|S)a�
    Resolves the argument to a numeric constant, which can be passed to
    the wrap_socket function/method from the ssl module.
    Defaults to :data:`ssl.CERT_REQUIRED`.
    If given a string it is assumed to be the name of the constant in the
    :mod:`ssl` module or its abbreviation.
    (So you can specify `REQUIRED` instead of `CERT_REQUIRED`.
    If it's neither `None` nor a string we assume it is already the numeric
    constant which can directly be passed to wrap_socket.
    N�CERT_)r!�
isinstance�str�getattrr0��	candidate�resrrr�resolve_cert_reqs�s
rdcCs@|durtSt|t�r<tt|d�}|dur8ttd|�}|S|S)z 
    like resolve_cert_reqs
    N�	PROTOCOL_)r%r^r_r`r0rarrr�resolve_ssl_version�s
rfcs�|r|tkrt}t|����|p"t�|dur4tjn|}|durdd}|tO}|tO}|t	O}|t
O}�j|O_|tjks�tj
dkr�t�dd�dur�d�_�fdd�}|tjkr�|�_|�n|�|�_t�d�r�tj�d	�}|r�|�_�S)
aAll arguments have the same meaning as ``ssl_wrap_socket``.

    By default, this function does a lot of the same work that
    ``ssl.create_default_context`` does on Python 3.4+. It:

    - Disables SSLv2, SSLv3, and compression
    - Sets a restricted set of server ciphers

    If you wish to enable SSLv3, you can do::

        from pip._vendor.urllib3.util import ssl_
        context = ssl_.create_urllib3_context()
        context.options &= ~ssl_.OP_NO_SSLv3

    You can do the same to enable compression (substituting ``COMPRESSION``
    for ``SSLv3`` in the last line above).

    :param ssl_version:
        The desired protocol version to use. This will default to
        PROTOCOL_SSLv23 which will negotiate the highest protocol that both
        the server and your installation of OpenSSL support.
    :param cert_reqs:
        Whether to require the certificate verification. This defaults to
        ``ssl.CERT_REQUIRED``.
    :param options:
        Specific OpenSSL options. These default to ``ssl.OP_NO_SSLv2``,
        ``ssl.OP_NO_SSLv3``, ``ssl.OP_NO_COMPRESSION``, and ``ssl.OP_NO_TICKET``.
    :param ciphers:
        Which cipher suites to allow the server to select.
    :returns:
        Constructed SSLContext object with specified options
    :rtype: SSLContext
    Nr)����post_handshake_authTcst�dd�durd�_dS)Nr/F)r`r/r��contextrr�disable_check_hostnameEs�z6create_urllib3_context.<locals>.disable_check_hostname�keylog_filename�
SSLKEYLOGFILE)r%r'r-rA�DEFAULT_CIPHERSr0r!r)r*r(r+r4�sys�version_infor`rjr2�hasattr�os�environrQrn)rCrBr4r7rm�
sslkeylogfilerrkr�create_urllib3_context�s:%��

rwc

Cs\|}
|
durt|||d�}
|s&|	s&|rlz|
�||	|�Wq�ttfyh}zt|��WYd}~q�d}~00n|dur�t|
d�r�|
��|r�|
dur�t|�r�td��|r�|
dur�|
�||�n|
�|||
�zt|
d�r�|
�	t
�Wnty�Yn0|�ot|�}|�rt
�pt�o|}t
�s4|�r4t�dt�|�rLt||
||d�}nt||
|�}|S)a
    All arguments except for server_hostname, ssl_context, and ca_cert_dir have
    the same meaning as they do when using :func:`ssl.wrap_socket`.

    :param server_hostname:
        When SNI is supported, the expected hostname of the certificate
    :param ssl_context:
        A pre-made :class:`SSLContext` object. If none is provided, one will
        be created using :func:`create_urllib3_context`.
    :param ciphers:
        A string of ciphers we wish the client to support.
    :param ca_cert_dir:
        A directory containing CA certificates in multiple separate files, as
        supported by OpenSSL's -CApath flag or the capath argument to
        SSLContext.load_verify_locations().
    :param key_password:
        Optional password if the keyfile is encrypted.
    :param ca_cert_data:
        Optional string containing CA certificates in PEM format suitable for
        passing as the cadata parameter to SSLContext.load_verify_locations()
    :param tls_in_tls:
        Use SSLTransport to wrap the existing socket.
    Nr@�load_default_certsz5Client private key is encrypted, password is required�set_alpn_protocolsa�An HTTPS request has been made, but the SNI (Server Name Indication) extension to TLS is not available on this platform. This may cause the server to present an incorrect TLS certificate, which can cause validation failures. You can upgrade to a newer version of Python to solve this. For more information, see https://urllib3.readthedocs.io/en/1.26.x/advanced-usage.html#ssl-warnings�rH)rwr?�IOError�OSErrorrrsrx�_is_key_file_encryptedr;ry�ALPN_PROTOCOLS�NotImplementedError�is_ipaddressr#�IS_SECURETRANSPORTrErFr�_ssl_wrap_socket_impl)�sockr6r5rBr3rHrCr7�ssl_context�ca_cert_dir�key_password�ca_cert_data�
tls_in_tlsrl�e�use_sni_hostname�send_sni�ssl_sockrrr�ssl_wrap_socketcsH& 
���r�cCs2tjst|t�r|�d�}tt�|�p.t�|��S)z�Detects whether the hostname given is an IPv4 or IPv6 address.
    Also detects IPv6 addresses with Zone IDs.

    :param str hostname: Hostname to examine.
    :return: True if the hostname is an IP address, False otherwise.
    �ascii)	r
�PY2r^�bytes�decode�boolr�matchr)�hostnamerrrr��s
r�cCsRt|d��4}|D]}d|vrWd�dSqWd�n1sD0YdS)z*Detects if a key file is encrypted or not.�r�	ENCRYPTEDNTF)�open)�key_file�f�linerrrr}�s
2r}cCsF|r&tstd��t�|�t|||�S|r8|j||d�S|�|�SdS)Nz0TLS in TLS requires support for the 'ssl' modulerz)r$r
�$_validate_ssl_context_for_tls_in_tlsr")r�r�r�rHrrrr��s�
r�)NNNN)NNNNNNNNNNNF)N)8�
__future__r�hmacrtrqrE�binasciirr�hashlibrrr�
exceptionsr	r
rr�packagesr
�urlrrr-r$r#�IS_PYOPENSSLr�r~rPrr`rUr0r!r"�ImportError�ssltransportr%r&r'r(r)r*r+�joinrp�objectr\rdrfrwr�r�r}r�rrrr�<module>s�



�1�
k�
f
site-packages/pip/_vendor/urllib3/util/__pycache__/ssltransport.cpython-39.pyc000064400000016601150564745010023473 0ustar00a

��Re�@sFddlZddlZddlZddlmZddlmZdZGdd�d�ZdS)�N)�ProxySchemeUnsupported)�sixi@c@s�eZdZdZedd��Zd:dd�Zdd	�Zd
d�Zdd
�Z	d;dd�Z
d<dd�Zd=dd�Zd>dd�Z
d?dd�Zd@dd�Zdd�Zdd �ZdAd"d#�Zd$d%�Zd&d'�Zd(d)�Zd*d+�Zd,d-�Zd.d/�Zd0d1�Zd2d3�Zd4d5�ZdBd6d7�Zd8d9�ZdS)C�SSLTransportaL
    The SSLTransport wraps an existing socket and establishes an SSL connection.

    Contrary to Python's implementation of SSLSocket, it allows you to chain
    multiple TLS connections together. It's particularly useful if you need to
    implement TLS within TLS.

    The class supports most of the socket API operations.
    cCs&t|d�s"tjrtd��ntd��dS)z�
        Raises a ProxySchemeUnsupported if the provided ssl_context can't be used
        for TLS in TLS.

        The only requirement is that the ssl_context provides the 'wrap_bio'
        methods.
        �wrap_biozKTLS in TLS requires SSLContext.wrap_bio() which isn't supported on Python 2zXTLS in TLS requires SSLContext.wrap_bio() which isn't available on non-native SSLContextN)�hasattrr�PY2r)�ssl_context�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/ssltransport.py�$_validate_ssl_context_for_tls_in_tlss

��z1SSLTransport._validate_ssl_context_for_tls_in_tlsNTcCsHt��|_t��|_||_||_|j|j|j|d�|_|�|jj	�dS)zV
        Create an SSLTransport around socket using the provided ssl_context.
        )�server_hostnameN)
�ssl�	MemoryBIO�incoming�outgoing�suppress_ragged_eofs�socketr�sslobj�_ssl_io_loop�do_handshake)�selfrrrrr	r	r
�__init__,s


�zSSLTransport.__init__cCs|S�Nr	�rr	r	r
�	__enter__?szSSLTransport.__enter__cGs|��dSr)�close)r�_r	r	r
�__exit__BszSSLTransport.__exit__cCs
|j��Sr)r�filenorr	r	r
rEszSSLTransport.fileno�cCs|�||�Sr)�_wrap_ssl_read)r�len�bufferr	r	r
�readHszSSLTransport.readrcCs|dkrtd��|�|�S)Nrz+non-zero flags not allowed in calls to recv)�
ValueErrorr )rr!�flagsr	r	r
�recvKszSSLTransport.recvcCs>|dkrtd��|r&|dur&t|�}n|dur2d}|�||�S)Nrz0non-zero flags not allowed in calls to recv_intor)r$r!r#)rr"�nbytesr%r	r	r
�	recv_intoPs
zSSLTransport.recv_intoc	Cs�|dkrtd��d}t|��f}|�d��<}t|�}||krV|�||d��}||7}q2Wd�n1sj0YWd�n1s�0YdS)Nrz.non-zero flags not allowed in calls to sendall�B)r$�
memoryview�castr!�send)r�datar%�count�view�	byte_view�amount�vr	r	r
�sendallYszSSLTransport.sendallcCs$|dkrtd��|�|jj|�}|S)Nrz+non-zero flags not allowed in calls to send)r$rr�write)rr-r%�responser	r	r
r,cszSSLTransport.send�rc
Cst|�hd�kstd|f��d|v}d|vp2|}|s@|s@J�d|v}d}	|rX|	d7}	|rd|	d7}	t�||	�}
|jjd7_|dur�d	}|d
kr�tj}|d
kr�|s�td��|
S|r�|r�t�|
|
|�}n&|r�t�|
|�}n|s�J�t�	|
|�}|r�|St�
||||�}||_|S)a
        Python's httpclient uses makefile and buffered io when reading HTTP
        messages and we need to support it.

        This is unfortunately a copy and paste of socket.py makefile with small
        changes to point to the socket directly.
        >r6�w�bz&invalid mode %r (only r, w, b allowed)r7r6r8��N���rz!unbuffered streams must be binary)�setr$r�SocketIO�_io_refs�io�DEFAULT_BUFFER_SIZE�BufferedRWPair�BufferedReader�BufferedWriter�
TextIOWrapper�mode)
rrE�	buffering�encoding�errors�newline�writing�reading�binary�rawmode�rawr"�textr	r	r
�makefileis@
zSSLTransport.makefilecCs|�|jj�dSr)rr�unwraprr	r	r
rQ�szSSLTransport.unwrapcCs|j��dSr)rrrr	r	r
r�szSSLTransport.closeFcCs|j�|�Sr)r�getpeercert)r�binary_formr	r	r
rR�szSSLTransport.getpeercertcCs
|j��Sr)r�versionrr	r	r
rT�szSSLTransport.versioncCs
|j��Sr)r�cipherrr	r	r
rU�szSSLTransport.ciphercCs
|j��Sr)r�selected_alpn_protocolrr	r	r
rV�sz#SSLTransport.selected_alpn_protocolcCs
|j��Sr)r�selected_npn_protocolrr	r	r
rW�sz"SSLTransport.selected_npn_protocolcCs
|j��Sr)r�shared_ciphersrr	r	r
rX�szSSLTransport.shared_cipherscCs
|j��Sr)r�compressionrr	r	r
rY�szSSLTransport.compressioncCs|j�|�dSr)r�
settimeout)r�valuer	r	r
rZ�szSSLTransport.settimeoutcCs
|j��Sr)r�
gettimeoutrr	r	r
r\�szSSLTransport.gettimeoutcCs|j��dSr)r�_decref_socketiosrr	r	r
r]�szSSLTransport._decref_socketiosc
Csbz|�|jj||�WStjy\}z.|jtjkrF|jrFWYd}~dS�WYd}~n
d}~00dS)Nr)rrr#r
�SSLError�errno�
SSL_ERROR_EOFr)rr!r"�er	r	r
r �szSSLTransport._wrap_ssl_readc
Gs�d}d}|r�d}z||�}WnBtjy^}z(|jtjtjfvrD|�|j}WYd}~n
d}~00|j��}|j�|�|dur�d}q|tjkr|j�	t
�}|r�|j�|�q|j�
�q|S)z>Performs an I/O loop between incoming/outgoing and the socket.TNF)r
r^r_�SSL_ERROR_WANT_READ�SSL_ERROR_WANT_WRITErr#rr3r&�
SSL_BLOCKSIZErr4�	write_eof)r�func�args�should_loop�retr_ra�bufr	r	r
r�s(

zSSLTransport._ssl_io_loop)NT)rN)rr)Nr)r)r)r6NNNN)F)N)�__name__�
__module__�__qualname__�__doc__�staticmethodrrrrrr#r&r(r3r,rPrQrrRrTrUrVrWrXrYrZr\r]r rr	r	r	r
rs:

�



	


�
-

	r)	r?rr
Zpip._vendor.urllib3.exceptionsrZpip._vendor.urllib3.packagesrrdrr	r	r	r
�<module>ssite-packages/pip/_vendor/urllib3/util/__pycache__/url.cpython-39.pyc000064400000024727150564745010021527 0ustar00a

��Re�6�@s�ddlmZddlZddlmZddlmZddlmZgd�Z	dZ
e�d	�Ze�d
�Z
e�dejejB�ZdZd
Zdjeed�Zeed�Zgd�ZdZdd�dd�eD��dZdedZdededZdZe�d�Ze�ded�Ze�ded�Ze�ded�Z e�dedd �d�Z!e�d!ed"�Z"d#eeefZ#e�e#ejejB�Z$e%d$�Z&e%d%�Z'e&e'Bd&hBZ(e(d'd(hBZ)e)d)hBZ*Z+Gd*d+�d+ed+e	��Z,d,d-�Z-d=d/d0�Z.d1d2�Z/d3d4�Z0d5d6�Z1d7d8�Z2d9d:�Z3d;d<�Z4dS)>�)�absolute_importN)�
namedtuple�)�LocationParseError)�six��scheme�auth�host�port�path�query�fragment)�http�httpsNz%[a-fA-F0-9]{2}z^(?:[a-zA-Z][a-zA-Z0-9+-]*:|/)zS^(?:([a-zA-Z][a-zA-Z0-9+.-]*):)?(?://([^\\/?#]*))?([^?#]*)(?:\?([^#]*))?(?:#(.*))?$z(?:[0-9]{1,3}\.){3}[0-9]{1,3}z[0-9A-Fa-f]{1,4}z(?:{hex}:{hex}|{ipv4}))�hexZipv4)r�ls32)	z(?:%(hex)s:){6}%(ls32)sz::(?:%(hex)s:){5}%(ls32)sz%(?:%(hex)s)?::(?:%(hex)s:){4}%(ls32)sz2(?:(?:%(hex)s:)?%(hex)s)?::(?:%(hex)s:){3}%(ls32)sz6(?:(?:%(hex)s:){0,2}%(hex)s)?::(?:%(hex)s:){2}%(ls32)sz/(?:(?:%(hex)s:){0,3}%(hex)s)?::%(hex)s:%(ls32)sz'(?:(?:%(hex)s:){0,4}%(hex)s)?::%(ls32)sz&(?:(?:%(hex)s:){0,5}%(hex)s)?::%(hex)sz(?:(?:%(hex)s:){0,6}%(hex)s)?::zDABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._!\-~z(?:�|cCsg|]}|t�qS�)�_subs��.0�xrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/url.py�
<listcomp>6�r�)z
(?:%25|%)(?:[z]|%[a-fA-F0-9]{2})+z\[z)?\]z!(?:[^\[\]%:/?#]|%[a-fA-F0-9]{2})*z!^(/[^?#]*)(?:\?([^#]*))?(?:#.*)?$�^�$����(z)\]$z^(%s|%s|%s)(?::([0-9]{0,5}))?$zBABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._-~z!$&'()*+,;=�:�@�/�?cs^eZdZdZdZd�fdd�	Zedd��Zedd	��Zed
d��Z	edd
��Z
dd�Z�ZS)�Urlz�
    Data structure for representing an HTTP URL. Used as a return value for
    :func:`parse_url`. Both the scheme and host are normalized as they are
    both case-insensitive according to RFC 3986.
    rNc
sD|r|�d�sd|}|dur&|��}tt|��||||||||�S)Nr#)�
startswith�lower�superr%�__new__)�clsrr	r
rrr
r��	__class__rrr)[s

�zUrl.__new__cCs|jS)z@For backwards-compatibility with urlparse. We're nice like that.)r
��selfrrr�hostnamemszUrl.hostnamecCs&|jpd}|jdur"|d|j7}|S)z)Absolute path including the query string.r#Nr$)rr
)r.�urirrr�request_urirs

zUrl.request_uricCs|jrd|j|jfS|jS)z(Network location including host and portz%s:%d)rr
r-rrr�netloc|sz
Url.netlocc	Cs�|\}}}}}}}d}|dur*||d7}|dur>||d7}|durN||7}|durf|dt|�7}|durv||7}|dur�|d|7}|dur�|d|7}|S)a�
        Convert self into a url

        This function should more or less round-trip with :func:`.parse_url`. The
        returned url may not be exactly the same as the url inputted to
        :func:`.parse_url`, but it should be equivalent by the RFC (e.g., urls
        with a blank port will have : removed).

        Example: ::

            >>> U = parse_url('http://google.com/mail/')
            >>> U.url
            'http://google.com/mail/'
            >>> Url('http', 'username:password', 'host.com', 80,
            ... '/path', 'query', 'fragment').url
            'http://username:password@host.com:80/path?query#fragment'
        �Nz://r"r!r$�#)�str)	r.rr	r
rrr
r�urlrrrr6�s"zUrl.urlcCs|jS�N)r6r-rrr�__str__�szUrl.__str__)NNNNNNN)
�__name__�
__module__�__qualname__�__doc__�	__slots__r)�propertyr/r1r2r6r8�
__classcell__rrr+rr%Rs&�

	

'r%cCsvd}d}|D]0}|�|�}|dkr$q|dus4||kr|}|}q|dusN|dkrX|ddfS|d|�||dd�|fS)a�
    .. deprecated:: 1.25

    Given a string and an iterable of delimiters, split on the first found
    delimiter. Return two split parts and the matched delimiter.

    If not found, then the first part is the full input string.

    Example::

        >>> split_first('foo/bar?baz', '?/=')
        ('foo', 'bar?baz', '/')
        >>> split_first('foo/bar?baz', '123')
        ('foo/bar?baz', '', None)

    Scales linearly with number of delims. Not ideal for large number of delims.
    Nrr3�)�find)�s�delims�min_idx�	min_delim�d�idxrrr�split_first�s

rH�utf-8c
Cs�|dur|St�|�}t�dd�|�\}}|�dd�}||�d�k}t�}tdt|��D]n}|||d�}t	|�}	|r�|dks�|	d	kr�|�
�|vr�||7}qX|�dt|	�d
d����
d
����qX|�
|�S)zfPercent-encodes a URI component without reapplying
    onto an already percent-encoded component.
    NcSs|�d���S)Nr)�group�upper)�matchrrr�<lambda>�rz'_encode_invalid_chars.<locals>.<lambda>rI�
surrogatepass�%rr@�r)r�ensure_text�
PERCENT_RE�subn�encode�count�	bytearray�range�len�ord�decode�extendr�zfillrK)
�	component�
allowed_chars�encoding�percent_encodings�	uri_bytes�is_percent_encoded�encoded_component�i�byte�byte_ordrrr�_encode_invalid_chars�s*
��
�*rgcCs�|�d�}g}|D]0}|dkr"qq|dkr6|�|�q|r|��q|�d�rf|rZ|drf|�dd�|�d�rz|�d�d�|�S)Nr#�.z..rr3)z/.z/..)�split�append�popr&�insert�endswith�join)r�segments�output�segmentrrr�_remove_path_dot_segments�s



rrcCs�|r�t|tj�rt�|�}|tvr�t�|�}|r�t�|�}|r�|�	d�\}}|||�}|�
d�rx|dkrx|dd�}n|dd�}dt|t�}|d|��
�|||d�S|�
�Sn*t�|�s�t�d�dd�|�d�D���S|S)	Nr@z%25��%�.cSsg|]}t|��qSr)�_idna_encode)r�labelrrrr)rz#_normalize_host.<locals>.<listcomp>rh)�
isinstancer�binary_type�
ensure_str�NORMALIZABLE_SCHEMES�
IPV6_ADDRZ_RErL�
ZONE_ID_RE�search�spanr&rg�UNRESERVED_CHARSr'�IPV4_RErnri)r
r�is_ipv6rL�start�end�zone_idrrr�_normalize_hosts*


 

�r�cCs�|r�tdd�|D��r�zddlm}Wn"tyHt�td�d�Yn0z|j|��ddd�WS|j	y�t�td|�d�Yn0|���d	�S)
NcSsg|]}t|�dk�qS)rP)rYrrrrr/rz _idna_encode.<locals>.<listcomp>r)�idnaz-Unable to parse URL without the 'idna' moduleT)�strict�
std3_rulesz#Name '%s' is not a valid IDNA label�ascii)
�any�pip._vendorr��ImportErrorr�
raise_fromrrTr'�	IDNAError)�namer�rrrrv.s�
�
rvcCs>t�|���\}}t|t�}t|t�}|dur:|d|7}|S)zHPercent-encodes a request target so that there are no invalid charactersNr$)�	TARGET_RErL�groupsrg�
PATH_CHARS�QUERY_CHARS)�targetrr
rrr�_encode_target@s

r�c	s�|s
t�S|}t�|�s d|}�zt�|���\}}}}}|dupN|��tv}|r\|��}|r�|�d�\}}	}
|pvd}t	�|
���\}}|r�|r�t
|t�}|dkr�d}n
d\}}}|dur�t|�}d|kr�dks�nt
|��t||�}|�r
|�r
t|�}t
|t�}|�r |�r t
|t�}|�r6|�r6t
|t�}Wn(ttf�y`t�t
|�d�YS0|�s�|du�s||du�r�d}nd}t|tj��r�tj�ntj��fdd	�}
t|
|�|
|�|
|�||
|�|
|�|
|�d
�S)a

    Given a url, return a parsed :class:`.Url` namedtuple. Best-effort is
    performed to parse incomplete urls. Fields not provided will be None.
    This parser is RFC 3986 compliant.

    The parser logic and helper functions are based heavily on
    work done in the ``rfc3986`` module.

    :param str url: URL to parse into a :class:`.Url` namedtuple.

    Partly backwards-compatible with :mod:`urlparse`.

    Example::

        >>> parse_url('http://google.com/mail/')
        Url(scheme='http', host='google.com', port=None, path='/mail/', ...)
        >>> parse_url('google.com:80')
        Url(scheme=None, host='google.com', port=80, path=None, ...)
        >>> parse_url('/foo?bar')
        Url(scheme=None, host=None, port=None, path='/foo', query='bar', ...)
    z//Nr"r3)NNNri��cs|dur|S�|�Sr7r)r��ensure_funcrr�ensure_type�szparse_url.<locals>.ensure_typer)r%�	SCHEME_REr~�URI_RErLr�r'r{�
rpartition�
_HOST_PORT_RErg�USERINFO_CHARS�intrr�rrr�r��FRAGMENT_CHARS�
ValueError�AttributeErrorrr�rx�	text_typerQrz)r6�
source_urlr�	authorityrr
r�
normalize_urir	�_�	host_portr
rr�rr�r�	parse_urlJsd





�r�cCst|�}|jpd|j|jfS)z4
    Deprecated. Use :func:`parse_url` instead.
    r)r�rr/r)r6�prrr�get_host�sr�)rI)5�
__future__r�re�collectionsr�
exceptionsr�packagesr�	url_attrsr{�compilerRr��UNICODE�DOTALLr��IPV4_PAT�HEX_PAT�format�LS32_PATr�_variations�UNRESERVED_PATrn�IPV6_PAT�ZONE_ID_PAT�IPV6_ADDRZ_PAT�REG_NAME_PATr�r��IPV6_REr|�BRACELESS_IPV6_ADDRZ_REr}�_HOST_PORT_PATr��setr��SUB_DELIM_CHARSr�r�r�r�r%rHrgrrr�rvr�r�r�rrrr�<module>sd


�	

��]#
"
asite-packages/pip/_vendor/urllib3/util/__pycache__/__init__.cpython-39.pyc000064400000002174150564745010022454 0ustar00a

��Re��@s�ddlmZddlmZddlmZmZmZddlm	Z	ddl
mZddlm
Z
mZmZmZmZmZmZmZmZmZddlmZmZdd	lmZmZmZmZdd
lm Z m!Z!dZ"dS)
�)�absolute_import�)�is_connection_dropped)�SKIP_HEADER�SKIPPABLE_HEADERS�make_headers)�is_fp_closed)�Retry)
�ALPN_PROTOCOLS�HAS_SNI�IS_PYOPENSSL�IS_SECURETRANSPORT�PROTOCOL_TLS�
SSLContext�assert_fingerprint�resolve_cert_reqs�resolve_ssl_version�ssl_wrap_socket)�Timeout�current_time)�Url�get_host�	parse_url�split_first)�
wait_for_read�wait_for_write)rrr
rrr
r	rrrrrrrrrrrrrrrrrN)#�
__future__r�
connectionr�requestrrr�responser�retryr	�ssl_r
rrr
rrrrrr�timeoutrr�urlrrrr�waitrr�__all__�r&r&��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/__init__.py�<module>s0site-packages/pip/_vendor/urllib3/util/__pycache__/timeout.cpython-39.pyc000064400000021432150564745010022401 0ustar00a

��Re'�@sTddlmZddlZddlmZddlmZe�Ze	edej�Z
Gdd�de�ZdS)	�)�absolute_importN)�_GLOBAL_DEFAULT_TIMEOUT�)�TimeoutStateError�	monotonicc@sxeZdZdZeZdeefdd�Zdd�ZeZ	e
dd��Ze
d	d
��Zdd�Z
d
d�Zdd�Zedd��Zedd��ZdS)�Timeouta2Timeout configuration.

    Timeouts can be defined as a default for a pool:

    .. code-block:: python

       timeout = Timeout(connect=2.0, read=7.0)
       http = PoolManager(timeout=timeout)
       response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool):

    .. code-block:: python

       response = http.request('GET', 'http://example.com/', timeout=Timeout(10))

    Timeouts can be disabled by setting all the parameters to ``None``:

    .. code-block:: python

       no_timeout = Timeout(connect=None, read=None)
       response = http.request('GET', 'http://example.com/, timeout=no_timeout)


    :param total:
        This combines the connect and read timeouts into one; the read timeout
        will be set to the time leftover from the connect attempt. In the
        event that both a connect timeout and a total are specified, or a read
        timeout and a total are specified, the shorter timeout will be applied.

        Defaults to None.

    :type total: int, float, or None

    :param connect:
        The maximum amount of time (in seconds) to wait for a connection
        attempt to a server to succeed. Omitting the parameter will default the
        connect timeout to the system default, probably `the global default
        timeout in socket.py
        <http://hg.python.org/cpython/file/603b4d593758/Lib/socket.py#l535>`_.
        None will set an infinite timeout for connection attempts.

    :type connect: int, float, or None

    :param read:
        The maximum amount of time (in seconds) to wait between consecutive
        read operations for a response from the server. Omitting the parameter
        will default the read timeout to the system default, probably `the
        global default timeout in socket.py
        <http://hg.python.org/cpython/file/603b4d593758/Lib/socket.py#l535>`_.
        None will set an infinite timeout.

    :type read: int, float, or None

    .. note::

        Many factors can affect the total amount of time for urllib3 to return
        an HTTP response.

        For example, Python's DNS resolver does not obey the timeout specified
        on the socket. Other factors that can affect total request time include
        high CPU load, high swap, the program running at a low priority level,
        or other behaviors.

        In addition, the read and total timeouts only measure the time between
        read operations on the socket connecting the client and the server,
        not the total amount of time for the request to return a complete
        response. For most requests, the timeout is raised because the server
        has not sent the first byte in the specified time. This is not always
        the case; if a server streams one byte every fifteen seconds, a timeout
        of 20 seconds will not trigger, even though the request will take
        several minutes to complete.

        If your goal is to cut off any request after a set amount of wall clock
        time, consider having a second "watcher" thread to cut off a slow
        request.
    NcCs4|�|d�|_|�|d�|_|�|d�|_d|_dS)N�connect�read�total)�_validate_timeout�_connect�_readr
�_start_connect)�selfr
rr	�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/timeout.py�__init__fszTimeout.__init__cCsdt|�j|j|j|jfS)Nz!%s(connect=%r, read=%r, total=%r))�type�__name__rr
r
�rrrr�__repr__ls�zTimeout.__repr__c	Cs�|tur|jS|dus ||jur$|St|t�r6td��zt|�Wn&ttfyhtd||f��Yn0z|dkr�td||f��Wn"ty�td||f��Yn0|S)a�Check that a timeout attribute is valid.

        :param value: The timeout value to validate
        :param name: The name of the timeout attribute to validate. This is
            used to specify in error messages.
        :return: The validated and casted version of the given value.
        :raises ValueError: If it is a numeric value less than or equal to
            zero, or the type is not an integer, float, or None.
        NzDTimeout cannot be a boolean value. It must be an int, float or None.z>Timeout value %s was %s, but it must be an int, float or None.rzdAttempted to set %s timeout to %s, but the timeout cannot be set to a value less than or equal to 0.)�_Default�DEFAULT_TIMEOUT�
isinstance�bool�
ValueError�float�	TypeError)�cls�value�namerrrrws<
���
����
zTimeout._validate_timeoutcCst||d�S)a�Create a new Timeout from a legacy timeout value.

        The timeout value used by httplib.py sets the same timeout on the
        connect(), and recv() socket requests. This creates a :class:`Timeout`
        object that sets the individual timeouts to the ``timeout`` value
        passed to this function.

        :param timeout: The legacy timeout value.
        :type timeout: integer, float, sentinel default object, or None
        :return: Timeout object
        :rtype: :class:`Timeout`
        )r	r)r)r�timeoutrrr�
from_float�szTimeout.from_floatcCst|j|j|jd�S)aCreate a copy of the timeout object

        Timeout properties are stored per-pool but each request needs a fresh
        Timeout object to ensure each one has its own start/stop configured.

        :return: a copy of the timeout object
        :rtype: :class:`Timeout`
        )rr	r
)rrr
r
rrrr�clone�sz
Timeout.clonecCs |jdurtd��t�|_|jS)z�Start the timeout clock, used during a connect() attempt

        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to start a timer that has been started already.
        Nz'Timeout timer has already been started.�rr�current_timerrrr�
start_connect�s
zTimeout.start_connectcCs|jdurtd��t�|jS)aGets the time elapsed since the call to :meth:`start_connect`.

        :return: Elapsed time in seconds.
        :rtype: float
        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to get duration for a timer that hasn't been started.
        Nz:Can't get connect duration for timer that has not started.r$rrrr�get_connect_duration�s

�zTimeout.get_connect_durationcCs:|jdur|jS|jdus&|j|jur,|jSt|j|j�S)a!Get the value to use when setting a connection timeout.

        This will be a positive float or integer, the value None
        (never timeout), or the default system timeout.

        :return: Connect timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        N)r
rr�minrrrr�connect_timeout�s


zTimeout.connect_timeoutcCs�|jdurX|j|jurX|jdurX|j|jurX|jdur<|jStdt|j|��|j��S|jdur�|j|jur�td|j|���S|jSdS)a�Get the value for the read timeout.

        This assumes some time has elapsed in the connection timeout and
        computes the read timeout appropriately.

        If self.total is set, the read timeout is dependent on the amount of
        time taken by the connect timeout. If the connection time has not been
        established, a :exc:`~urllib3.exceptions.TimeoutStateError` will be
        raised.

        :return: Value to use for the read timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        :raises urllib3.exceptions.TimeoutStateError: If :meth:`start_connect`
            has not yet been called on this object.
        Nr)r
rr
r�maxr(r'rrrr�read_timeout�s�
��
�
zTimeout.read_timeout)r�
__module__�__qualname__�__doc__rrrrr�__str__�classmethodrr"r#r&r'�propertyr)r+rrrrrs O	
-

r)�
__future__r�time�socketr�
exceptionsr�objectr�getattrr%rrrrr�<module>ssite-packages/pip/_vendor/urllib3/util/__pycache__/connection.cpython-39.pyc000064400000006654150564745010023063 0ustar00a

��Re8�@s�ddlmZddlZddlmZddlmZddlmZddl	m
Z
mZd	d
�Zej
ddfdd�Zd
d�Zdd�Zdd�Zed�ZdS)�)�absolute_importN)�LocationParseError�)�_appengine_environ)�six�)�NoWayToWaitForSocketError�
wait_for_readcCsLt|dd�}|durdS|dur$dSzt|dd�WStyFYdS0dS)a$
    Returns True if the connection is dropped and should be closed.

    :param conn:
        :class:`http.client.HTTPConnection` object.

    Note: For platforms like AppEngine, this will always return ``False`` to
    let the platform handle connection recycling transparently for us.
    �sockFNTg)�timeout)�getattrr	r)�connr
�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/util/connection.py�is_connection_droppeds
rcCs4|\}}|�d�r|�d�}d}t�}z|�d�Wn&tyZt�td|�d�YS0t�	|||tj
�D]�}|\}	}
}}}
d}zNt�|	|
|�}t||�|tjur�|�
|�|r�|�|�|�|
�|WStj�y}z&|}|du�r|��d}WYd}~qnd}~00qn|du�r&|�t�d��dS)akConnect to *address* and return the socket object.

    Convenience function.  Connect to *address* (a 2-tuple ``(host,
    port)``) and return the socket object.  Passing the optional
    *timeout* parameter will set the timeout on the socket instance
    before attempting to connect.  If no *timeout* is supplied, the
    global default timeout setting returned by :func:`socket.getdefaulttimeout`
    is used.  If *source_address* is set it must be a tuple of (host, port)
    for the socket to bind as a source address before making the connection.
    An host of '' or port 0 tells the OS to use the default.
    �[z[]N�idnaz'%s', label empty or too longz!getaddrinfo returns an empty list)�
startswith�strip�allowed_gai_family�encode�UnicodeErrorr�
raise_fromr�socket�getaddrinfo�SOCK_STREAM�_set_socket_options�_GLOBAL_DEFAULT_TIMEOUT�
settimeout�bind�connect�error�close)�addressr�source_address�socket_options�host�port�err�family�res�af�socktype�proto�	canonname�sar
�errr�create_connection&s>

�








r1cCs$|durdS|D]}|j|�qdS)N)�
setsockopt)r
�options�optrrrresrcCstj}trtj}|S)z�This function is designed to work in the context of
    getaddrinfo, where family=socket.AF_UNSPEC is the default and
    will perform a DNS search for both IPv6 and IPv4 records.)r�AF_INET�HAS_IPV6�	AF_UNSPEC)r)rrrrmsrcCs`d}d}t��rdStjrPz"t�tj�}|�|df�d}WntyNYn0|r\|��|S)z4Returns True if the system can bind an IPv6 address.NFrT)r�is_appengine_sandboxr�has_ipv6�AF_INET6r�	Exceptionr")r&r
r9rrr�	_has_ipv6xsr<z::1)�
__future__rrZpip._vendor.urllib3.exceptionsr�contribr�packagesr�waitrr	rrr1rrr<r6rrrr�<module>s�
?site-packages/pip/_vendor/urllib3/util/timeout.py000064400000023423150564745010016114 0ustar00from __future__ import absolute_import

import time

# The default socket timeout, used by httplib to indicate that no timeout was
# specified by the user
from socket import _GLOBAL_DEFAULT_TIMEOUT

from ..exceptions import TimeoutStateError

# A sentinel value to indicate that no timeout was specified by the user in
# urllib3
_Default = object()


# Use time.monotonic if available.
current_time = getattr(time, "monotonic", time.time)


class Timeout(object):
    """Timeout configuration.

    Timeouts can be defined as a default for a pool:

    .. code-block:: python

       timeout = Timeout(connect=2.0, read=7.0)
       http = PoolManager(timeout=timeout)
       response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool):

    .. code-block:: python

       response = http.request('GET', 'http://example.com/', timeout=Timeout(10))

    Timeouts can be disabled by setting all the parameters to ``None``:

    .. code-block:: python

       no_timeout = Timeout(connect=None, read=None)
       response = http.request('GET', 'http://example.com/, timeout=no_timeout)


    :param total:
        This combines the connect and read timeouts into one; the read timeout
        will be set to the time leftover from the connect attempt. In the
        event that both a connect timeout and a total are specified, or a read
        timeout and a total are specified, the shorter timeout will be applied.

        Defaults to None.

    :type total: int, float, or None

    :param connect:
        The maximum amount of time (in seconds) to wait for a connection
        attempt to a server to succeed. Omitting the parameter will default the
        connect timeout to the system default, probably `the global default
        timeout in socket.py
        <http://hg.python.org/cpython/file/603b4d593758/Lib/socket.py#l535>`_.
        None will set an infinite timeout for connection attempts.

    :type connect: int, float, or None

    :param read:
        The maximum amount of time (in seconds) to wait between consecutive
        read operations for a response from the server. Omitting the parameter
        will default the read timeout to the system default, probably `the
        global default timeout in socket.py
        <http://hg.python.org/cpython/file/603b4d593758/Lib/socket.py#l535>`_.
        None will set an infinite timeout.

    :type read: int, float, or None

    .. note::

        Many factors can affect the total amount of time for urllib3 to return
        an HTTP response.

        For example, Python's DNS resolver does not obey the timeout specified
        on the socket. Other factors that can affect total request time include
        high CPU load, high swap, the program running at a low priority level,
        or other behaviors.

        In addition, the read and total timeouts only measure the time between
        read operations on the socket connecting the client and the server,
        not the total amount of time for the request to return a complete
        response. For most requests, the timeout is raised because the server
        has not sent the first byte in the specified time. This is not always
        the case; if a server streams one byte every fifteen seconds, a timeout
        of 20 seconds will not trigger, even though the request will take
        several minutes to complete.

        If your goal is to cut off any request after a set amount of wall clock
        time, consider having a second "watcher" thread to cut off a slow
        request.
    """

    #: A sentinel object representing the default timeout value
    DEFAULT_TIMEOUT = _GLOBAL_DEFAULT_TIMEOUT

    def __init__(self, total=None, connect=_Default, read=_Default):
        self._connect = self._validate_timeout(connect, "connect")
        self._read = self._validate_timeout(read, "read")
        self.total = self._validate_timeout(total, "total")
        self._start_connect = None

    def __repr__(self):
        return "%s(connect=%r, read=%r, total=%r)" % (
            type(self).__name__,
            self._connect,
            self._read,
            self.total,
        )

    # __str__ provided for backwards compatibility
    __str__ = __repr__

    @classmethod
    def _validate_timeout(cls, value, name):
        """Check that a timeout attribute is valid.

        :param value: The timeout value to validate
        :param name: The name of the timeout attribute to validate. This is
            used to specify in error messages.
        :return: The validated and casted version of the given value.
        :raises ValueError: If it is a numeric value less than or equal to
            zero, or the type is not an integer, float, or None.
        """
        if value is _Default:
            return cls.DEFAULT_TIMEOUT

        if value is None or value is cls.DEFAULT_TIMEOUT:
            return value

        if isinstance(value, bool):
            raise ValueError(
                "Timeout cannot be a boolean value. It must "
                "be an int, float or None."
            )
        try:
            float(value)
        except (TypeError, ValueError):
            raise ValueError(
                "Timeout value %s was %s, but it must be an "
                "int, float or None." % (name, value)
            )

        try:
            if value <= 0:
                raise ValueError(
                    "Attempted to set %s timeout to %s, but the "
                    "timeout cannot be set to a value less "
                    "than or equal to 0." % (name, value)
                )
        except TypeError:
            # Python 3
            raise ValueError(
                "Timeout value %s was %s, but it must be an "
                "int, float or None." % (name, value)
            )

        return value

    @classmethod
    def from_float(cls, timeout):
        """Create a new Timeout from a legacy timeout value.

        The timeout value used by httplib.py sets the same timeout on the
        connect(), and recv() socket requests. This creates a :class:`Timeout`
        object that sets the individual timeouts to the ``timeout`` value
        passed to this function.

        :param timeout: The legacy timeout value.
        :type timeout: integer, float, sentinel default object, or None
        :return: Timeout object
        :rtype: :class:`Timeout`
        """
        return Timeout(read=timeout, connect=timeout)

    def clone(self):
        """Create a copy of the timeout object

        Timeout properties are stored per-pool but each request needs a fresh
        Timeout object to ensure each one has its own start/stop configured.

        :return: a copy of the timeout object
        :rtype: :class:`Timeout`
        """
        # We can't use copy.deepcopy because that will also create a new object
        # for _GLOBAL_DEFAULT_TIMEOUT, which socket.py uses as a sentinel to
        # detect the user default.
        return Timeout(connect=self._connect, read=self._read, total=self.total)

    def start_connect(self):
        """Start the timeout clock, used during a connect() attempt

        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to start a timer that has been started already.
        """
        if self._start_connect is not None:
            raise TimeoutStateError("Timeout timer has already been started.")
        self._start_connect = current_time()
        return self._start_connect

    def get_connect_duration(self):
        """Gets the time elapsed since the call to :meth:`start_connect`.

        :return: Elapsed time in seconds.
        :rtype: float
        :raises urllib3.exceptions.TimeoutStateError: if you attempt
            to get duration for a timer that hasn't been started.
        """
        if self._start_connect is None:
            raise TimeoutStateError(
                "Can't get connect duration for timer that has not started."
            )
        return current_time() - self._start_connect

    @property
    def connect_timeout(self):
        """Get the value to use when setting a connection timeout.

        This will be a positive float or integer, the value None
        (never timeout), or the default system timeout.

        :return: Connect timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        """
        if self.total is None:
            return self._connect

        if self._connect is None or self._connect is self.DEFAULT_TIMEOUT:
            return self.total

        return min(self._connect, self.total)

    @property
    def read_timeout(self):
        """Get the value for the read timeout.

        This assumes some time has elapsed in the connection timeout and
        computes the read timeout appropriately.

        If self.total is set, the read timeout is dependent on the amount of
        time taken by the connect timeout. If the connection time has not been
        established, a :exc:`~urllib3.exceptions.TimeoutStateError` will be
        raised.

        :return: Value to use for the read timeout.
        :rtype: int, float, :attr:`Timeout.DEFAULT_TIMEOUT` or None
        :raises urllib3.exceptions.TimeoutStateError: If :meth:`start_connect`
            has not yet been called on this object.
        """
        if (
            self.total is not None
            and self.total is not self.DEFAULT_TIMEOUT
            and self._read is not None
            and self._read is not self.DEFAULT_TIMEOUT
        ):
            # In case the connect timeout has not yet been established.
            if self._start_connect is None:
                return self._read
            return max(0, min(self.total - self.get_connect_duration(), self._read))
        elif self.total is not None and self.total is not self.DEFAULT_TIMEOUT:
            return max(0, self.total - self.get_connect_duration())
        else:
            return self._read
site-packages/pip/_vendor/urllib3/util/queue.py000064400000000762150564745010015553 0ustar00import collections

from ..packages import six
from ..packages.six.moves import queue

if six.PY2:
    # Queue is imported for side effects on MS Windows. See issue #229.
    import Queue as _unused_module_Queue  # noqa: F401


class LifoQueue(queue.Queue):
    def _init(self, _):
        self.queue = collections.deque()

    def _qsize(self, len=len):
        return len(self.queue)

    def _put(self, item):
        self.queue.append(item)

    def _get(self):
        return self.queue.pop()
site-packages/pip/_vendor/urllib3/util/connection.py000064400000011470150564745010016564 0ustar00from __future__ import absolute_import

import socket

from pip._vendor.urllib3.exceptions import LocationParseError

from ..contrib import _appengine_environ
from ..packages import six
from .wait import NoWayToWaitForSocketError, wait_for_read


def is_connection_dropped(conn):  # Platform-specific
    """
    Returns True if the connection is dropped and should be closed.

    :param conn:
        :class:`http.client.HTTPConnection` object.

    Note: For platforms like AppEngine, this will always return ``False`` to
    let the platform handle connection recycling transparently for us.
    """
    sock = getattr(conn, "sock", False)
    if sock is False:  # Platform-specific: AppEngine
        return False
    if sock is None:  # Connection already closed (such as by httplib).
        return True
    try:
        # Returns True if readable, which here means it's been dropped
        return wait_for_read(sock, timeout=0.0)
    except NoWayToWaitForSocketError:  # Platform-specific: AppEngine
        return False


# This function is copied from socket.py in the Python 2.7 standard
# library test suite. Added to its signature is only `socket_options`.
# One additional modification is that we avoid binding to IPv6 servers
# discovered in DNS if the system doesn't have IPv6 functionality.
def create_connection(
    address,
    timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
    source_address=None,
    socket_options=None,
):
    """Connect to *address* and return the socket object.

    Convenience function.  Connect to *address* (a 2-tuple ``(host,
    port)``) and return the socket object.  Passing the optional
    *timeout* parameter will set the timeout on the socket instance
    before attempting to connect.  If no *timeout* is supplied, the
    global default timeout setting returned by :func:`socket.getdefaulttimeout`
    is used.  If *source_address* is set it must be a tuple of (host, port)
    for the socket to bind as a source address before making the connection.
    An host of '' or port 0 tells the OS to use the default.
    """

    host, port = address
    if host.startswith("["):
        host = host.strip("[]")
    err = None

    # Using the value from allowed_gai_family() in the context of getaddrinfo lets
    # us select whether to work with IPv4 DNS records, IPv6 records, or both.
    # The original create_connection function always returns all records.
    family = allowed_gai_family()

    try:
        host.encode("idna")
    except UnicodeError:
        return six.raise_from(
            LocationParseError(u"'%s', label empty or too long" % host), None
        )

    for res in socket.getaddrinfo(host, port, family, socket.SOCK_STREAM):
        af, socktype, proto, canonname, sa = res
        sock = None
        try:
            sock = socket.socket(af, socktype, proto)

            # If provided, set socket level options before connecting.
            _set_socket_options(sock, socket_options)

            if timeout is not socket._GLOBAL_DEFAULT_TIMEOUT:
                sock.settimeout(timeout)
            if source_address:
                sock.bind(source_address)
            sock.connect(sa)
            return sock

        except socket.error as e:
            err = e
            if sock is not None:
                sock.close()
                sock = None

    if err is not None:
        raise err

    raise socket.error("getaddrinfo returns an empty list")


def _set_socket_options(sock, options):
    if options is None:
        return

    for opt in options:
        sock.setsockopt(*opt)


def allowed_gai_family():
    """This function is designed to work in the context of
    getaddrinfo, where family=socket.AF_UNSPEC is the default and
    will perform a DNS search for both IPv6 and IPv4 records."""

    family = socket.AF_INET
    if HAS_IPV6:
        family = socket.AF_UNSPEC
    return family


def _has_ipv6(host):
    """Returns True if the system can bind an IPv6 address."""
    sock = None
    has_ipv6 = False

    # App Engine doesn't support IPV6 sockets and actually has a quota on the
    # number of sockets that can be used, so just early out here instead of
    # creating a socket needlessly.
    # See https://github.com/urllib3/urllib3/issues/1446
    if _appengine_environ.is_appengine_sandbox():
        return False

    if socket.has_ipv6:
        # has_ipv6 returns true if cPython was compiled with IPv6 support.
        # It does not tell us if the system has IPv6 support enabled. To
        # determine that we must bind to an IPv6 address.
        # https://github.com/urllib3/urllib3/pull/611
        # https://bugs.python.org/issue658327
        try:
            sock = socket.socket(socket.AF_INET6)
            sock.bind((host, 0))
            has_ipv6 = True
        except Exception:
            pass

    if sock:
        sock.close()
    return has_ipv6


HAS_IPV6 = _has_ipv6("::1")
site-packages/pip/_vendor/urllib3/util/__init__.py000064400000002203150564745010016156 0ustar00from __future__ import absolute_import

# For backwards compatibility, provide imports that used to be here.
from .connection import is_connection_dropped
from .request import SKIP_HEADER, SKIPPABLE_HEADERS, make_headers
from .response import is_fp_closed
from .retry import Retry
from .ssl_ import (
    ALPN_PROTOCOLS,
    HAS_SNI,
    IS_PYOPENSSL,
    IS_SECURETRANSPORT,
    PROTOCOL_TLS,
    SSLContext,
    assert_fingerprint,
    resolve_cert_reqs,
    resolve_ssl_version,
    ssl_wrap_socket,
)
from .timeout import Timeout, current_time
from .url import Url, get_host, parse_url, split_first
from .wait import wait_for_read, wait_for_write

__all__ = (
    "HAS_SNI",
    "IS_PYOPENSSL",
    "IS_SECURETRANSPORT",
    "SSLContext",
    "PROTOCOL_TLS",
    "ALPN_PROTOCOLS",
    "Retry",
    "Timeout",
    "Url",
    "assert_fingerprint",
    "current_time",
    "is_connection_dropped",
    "is_fp_closed",
    "get_host",
    "parse_url",
    "make_headers",
    "resolve_cert_reqs",
    "resolve_ssl_version",
    "split_first",
    "ssl_wrap_socket",
    "wait_for_read",
    "wait_for_write",
    "SKIP_HEADER",
    "SKIPPABLE_HEADERS",
)
site-packages/pip/_vendor/urllib3/util/retry.py000064400000051617150564745010015601 0ustar00from __future__ import absolute_import

import email
import logging
import re
import time
import warnings
from collections import namedtuple
from itertools import takewhile

from ..exceptions import (
    ConnectTimeoutError,
    InvalidHeader,
    MaxRetryError,
    ProtocolError,
    ProxyError,
    ReadTimeoutError,
    ResponseError,
)
from ..packages import six

log = logging.getLogger(__name__)


# Data structure for representing the metadata of requests that result in a retry.
RequestHistory = namedtuple(
    "RequestHistory", ["method", "url", "error", "status", "redirect_location"]
)


# TODO: In v2 we can remove this sentinel and metaclass with deprecated options.
_Default = object()


class _RetryMeta(type):
    @property
    def DEFAULT_METHOD_WHITELIST(cls):
        warnings.warn(
            "Using 'Retry.DEFAULT_METHOD_WHITELIST' is deprecated and "
            "will be removed in v2.0. Use 'Retry.DEFAULT_ALLOWED_METHODS' instead",
            DeprecationWarning,
        )
        return cls.DEFAULT_ALLOWED_METHODS

    @DEFAULT_METHOD_WHITELIST.setter
    def DEFAULT_METHOD_WHITELIST(cls, value):
        warnings.warn(
            "Using 'Retry.DEFAULT_METHOD_WHITELIST' is deprecated and "
            "will be removed in v2.0. Use 'Retry.DEFAULT_ALLOWED_METHODS' instead",
            DeprecationWarning,
        )
        cls.DEFAULT_ALLOWED_METHODS = value

    @property
    def DEFAULT_REDIRECT_HEADERS_BLACKLIST(cls):
        warnings.warn(
            "Using 'Retry.DEFAULT_REDIRECT_HEADERS_BLACKLIST' is deprecated and "
            "will be removed in v2.0. Use 'Retry.DEFAULT_REMOVE_HEADERS_ON_REDIRECT' instead",
            DeprecationWarning,
        )
        return cls.DEFAULT_REMOVE_HEADERS_ON_REDIRECT

    @DEFAULT_REDIRECT_HEADERS_BLACKLIST.setter
    def DEFAULT_REDIRECT_HEADERS_BLACKLIST(cls, value):
        warnings.warn(
            "Using 'Retry.DEFAULT_REDIRECT_HEADERS_BLACKLIST' is deprecated and "
            "will be removed in v2.0. Use 'Retry.DEFAULT_REMOVE_HEADERS_ON_REDIRECT' instead",
            DeprecationWarning,
        )
        cls.DEFAULT_REMOVE_HEADERS_ON_REDIRECT = value


@six.add_metaclass(_RetryMeta)
class Retry(object):
    """Retry configuration.

    Each retry attempt will create a new Retry object with updated values, so
    they can be safely reused.

    Retries can be defined as a default for a pool::

        retries = Retry(connect=5, read=2, redirect=5)
        http = PoolManager(retries=retries)
        response = http.request('GET', 'http://example.com/')

    Or per-request (which overrides the default for the pool)::

        response = http.request('GET', 'http://example.com/', retries=Retry(10))

    Retries can be disabled by passing ``False``::

        response = http.request('GET', 'http://example.com/', retries=False)

    Errors will be wrapped in :class:`~urllib3.exceptions.MaxRetryError` unless
    retries are disabled, in which case the causing exception will be raised.

    :param int total:
        Total number of retries to allow. Takes precedence over other counts.

        Set to ``None`` to remove this constraint and fall back on other
        counts.

        Set to ``0`` to fail on the first retry.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int connect:
        How many connection-related errors to retry on.

        These are errors raised before the request is sent to the remote server,
        which we assume has not triggered the server to process the request.

        Set to ``0`` to fail on the first retry of this type.

    :param int read:
        How many times to retry on read errors.

        These errors are raised after the request was sent to the server, so the
        request may have side-effects.

        Set to ``0`` to fail on the first retry of this type.

    :param int redirect:
        How many redirects to perform. Limit this to avoid infinite redirect
        loops.

        A redirect is a HTTP response with a status code 301, 302, 303, 307 or
        308.

        Set to ``0`` to fail on the first retry of this type.

        Set to ``False`` to disable and imply ``raise_on_redirect=False``.

    :param int status:
        How many times to retry on bad status codes.

        These are retries made on responses, where status code matches
        ``status_forcelist``.

        Set to ``0`` to fail on the first retry of this type.

    :param int other:
        How many times to retry on other errors.

        Other errors are errors that are not connect, read, redirect or status errors.
        These errors might be raised after the request was sent to the server, so the
        request might have side-effects.

        Set to ``0`` to fail on the first retry of this type.

        If ``total`` is not set, it's a good idea to set this to 0 to account
        for unexpected edge cases and avoid infinite retry loops.

    :param iterable allowed_methods:
        Set of uppercased HTTP method verbs that we should retry on.

        By default, we only retry on methods which are considered to be
        idempotent (multiple requests with the same parameters end with the
        same state). See :attr:`Retry.DEFAULT_ALLOWED_METHODS`.

        Set to a ``False`` value to retry on any verb.

        .. warning::

            Previously this parameter was named ``method_whitelist``, that
            usage is deprecated in v1.26.0 and will be removed in v2.0.

    :param iterable status_forcelist:
        A set of integer HTTP status codes that we should force a retry on.
        A retry is initiated if the request method is in ``allowed_methods``
        and the response status code is in ``status_forcelist``.

        By default, this is disabled with ``None``.

    :param float backoff_factor:
        A backoff factor to apply between attempts after the second try
        (most errors are resolved immediately by a second try without a
        delay). urllib3 will sleep for::

            {backoff factor} * (2 ** ({number of total retries} - 1))

        seconds. If the backoff_factor is 0.1, then :func:`.sleep` will sleep
        for [0.0s, 0.2s, 0.4s, ...] between retries. It will never be longer
        than :attr:`Retry.BACKOFF_MAX`.

        By default, backoff is disabled (set to 0).

    :param bool raise_on_redirect: Whether, if the number of redirects is
        exhausted, to raise a MaxRetryError, or to return a response with a
        response code in the 3xx range.

    :param bool raise_on_status: Similar meaning to ``raise_on_redirect``:
        whether we should raise an exception, or return a response,
        if status falls in ``status_forcelist`` range and retries have
        been exhausted.

    :param tuple history: The history of the request encountered during
        each call to :meth:`~Retry.increment`. The list is in the order
        the requests occurred. Each list item is of class :class:`RequestHistory`.

    :param bool respect_retry_after_header:
        Whether to respect Retry-After header on status codes defined as
        :attr:`Retry.RETRY_AFTER_STATUS_CODES` or not.

    :param iterable remove_headers_on_redirect:
        Sequence of headers to remove from the request when a response
        indicating a redirect is returned before firing off the redirected
        request.
    """

    #: Default methods to be used for ``allowed_methods``
    DEFAULT_ALLOWED_METHODS = frozenset(
        ["HEAD", "GET", "PUT", "DELETE", "OPTIONS", "TRACE"]
    )

    #: Default status codes to be used for ``status_forcelist``
    RETRY_AFTER_STATUS_CODES = frozenset([413, 429, 503])

    #: Default headers to be used for ``remove_headers_on_redirect``
    DEFAULT_REMOVE_HEADERS_ON_REDIRECT = frozenset(["Authorization"])

    #: Maximum backoff time.
    BACKOFF_MAX = 120

    def __init__(
        self,
        total=10,
        connect=None,
        read=None,
        redirect=None,
        status=None,
        other=None,
        allowed_methods=_Default,
        status_forcelist=None,
        backoff_factor=0,
        raise_on_redirect=True,
        raise_on_status=True,
        history=None,
        respect_retry_after_header=True,
        remove_headers_on_redirect=_Default,
        # TODO: Deprecated, remove in v2.0
        method_whitelist=_Default,
    ):

        if method_whitelist is not _Default:
            if allowed_methods is not _Default:
                raise ValueError(
                    "Using both 'allowed_methods' and "
                    "'method_whitelist' together is not allowed. "
                    "Instead only use 'allowed_methods'"
                )
            warnings.warn(
                "Using 'method_whitelist' with Retry is deprecated and "
                "will be removed in v2.0. Use 'allowed_methods' instead",
                DeprecationWarning,
                stacklevel=2,
            )
            allowed_methods = method_whitelist
        if allowed_methods is _Default:
            allowed_methods = self.DEFAULT_ALLOWED_METHODS
        if remove_headers_on_redirect is _Default:
            remove_headers_on_redirect = self.DEFAULT_REMOVE_HEADERS_ON_REDIRECT

        self.total = total
        self.connect = connect
        self.read = read
        self.status = status
        self.other = other

        if redirect is False or total is False:
            redirect = 0
            raise_on_redirect = False

        self.redirect = redirect
        self.status_forcelist = status_forcelist or set()
        self.allowed_methods = allowed_methods
        self.backoff_factor = backoff_factor
        self.raise_on_redirect = raise_on_redirect
        self.raise_on_status = raise_on_status
        self.history = history or tuple()
        self.respect_retry_after_header = respect_retry_after_header
        self.remove_headers_on_redirect = frozenset(
            [h.lower() for h in remove_headers_on_redirect]
        )

    def new(self, **kw):
        params = dict(
            total=self.total,
            connect=self.connect,
            read=self.read,
            redirect=self.redirect,
            status=self.status,
            other=self.other,
            status_forcelist=self.status_forcelist,
            backoff_factor=self.backoff_factor,
            raise_on_redirect=self.raise_on_redirect,
            raise_on_status=self.raise_on_status,
            history=self.history,
            remove_headers_on_redirect=self.remove_headers_on_redirect,
            respect_retry_after_header=self.respect_retry_after_header,
        )

        # TODO: If already given in **kw we use what's given to us
        # If not given we need to figure out what to pass. We decide
        # based on whether our class has the 'method_whitelist' property
        # and if so we pass the deprecated 'method_whitelist' otherwise
        # we use 'allowed_methods'. Remove in v2.0
        if "method_whitelist" not in kw and "allowed_methods" not in kw:
            if "method_whitelist" in self.__dict__:
                warnings.warn(
                    "Using 'method_whitelist' with Retry is deprecated and "
                    "will be removed in v2.0. Use 'allowed_methods' instead",
                    DeprecationWarning,
                )
                params["method_whitelist"] = self.allowed_methods
            else:
                params["allowed_methods"] = self.allowed_methods

        params.update(kw)
        return type(self)(**params)

    @classmethod
    def from_int(cls, retries, redirect=True, default=None):
        """Backwards-compatibility for the old retries format."""
        if retries is None:
            retries = default if default is not None else cls.DEFAULT

        if isinstance(retries, Retry):
            return retries

        redirect = bool(redirect) and None
        new_retries = cls(retries, redirect=redirect)
        log.debug("Converted retries value: %r -> %r", retries, new_retries)
        return new_retries

    def get_backoff_time(self):
        """Formula for computing the current backoff

        :rtype: float
        """
        # We want to consider only the last consecutive errors sequence (Ignore redirects).
        consecutive_errors_len = len(
            list(
                takewhile(lambda x: x.redirect_location is None, reversed(self.history))
            )
        )
        if consecutive_errors_len <= 1:
            return 0

        backoff_value = self.backoff_factor * (2 ** (consecutive_errors_len - 1))
        return min(self.BACKOFF_MAX, backoff_value)

    def parse_retry_after(self, retry_after):
        # Whitespace: https://tools.ietf.org/html/rfc7230#section-3.2.4
        if re.match(r"^\s*[0-9]+\s*$", retry_after):
            seconds = int(retry_after)
        else:
            retry_date_tuple = email.utils.parsedate_tz(retry_after)
            if retry_date_tuple is None:
                raise InvalidHeader("Invalid Retry-After header: %s" % retry_after)
            if retry_date_tuple[9] is None:  # Python 2
                # Assume UTC if no timezone was specified
                # On Python2.7, parsedate_tz returns None for a timezone offset
                # instead of 0 if no timezone is given, where mktime_tz treats
                # a None timezone offset as local time.
                retry_date_tuple = retry_date_tuple[:9] + (0,) + retry_date_tuple[10:]

            retry_date = email.utils.mktime_tz(retry_date_tuple)
            seconds = retry_date - time.time()

        if seconds < 0:
            seconds = 0

        return seconds

    def get_retry_after(self, response):
        """Get the value of Retry-After in seconds."""

        retry_after = response.getheader("Retry-After")

        if retry_after is None:
            return None

        return self.parse_retry_after(retry_after)

    def sleep_for_retry(self, response=None):
        retry_after = self.get_retry_after(response)
        if retry_after:
            time.sleep(retry_after)
            return True

        return False

    def _sleep_backoff(self):
        backoff = self.get_backoff_time()
        if backoff <= 0:
            return
        time.sleep(backoff)

    def sleep(self, response=None):
        """Sleep between retry attempts.

        This method will respect a server's ``Retry-After`` response header
        and sleep the duration of the time requested. If that is not present, it
        will use an exponential backoff. By default, the backoff factor is 0 and
        this method will return immediately.
        """

        if self.respect_retry_after_header and response:
            slept = self.sleep_for_retry(response)
            if slept:
                return

        self._sleep_backoff()

    def _is_connection_error(self, err):
        """Errors when we're fairly sure that the server did not receive the
        request, so it should be safe to retry.
        """
        if isinstance(err, ProxyError):
            err = err.original_error
        return isinstance(err, ConnectTimeoutError)

    def _is_read_error(self, err):
        """Errors that occur after the request has been started, so we should
        assume that the server began processing it.
        """
        return isinstance(err, (ReadTimeoutError, ProtocolError))

    def _is_method_retryable(self, method):
        """Checks if a given HTTP method should be retried upon, depending if
        it is included in the allowed_methods
        """
        # TODO: For now favor if the Retry implementation sets its own method_whitelist
        # property outside of our constructor to avoid breaking custom implementations.
        if "method_whitelist" in self.__dict__:
            warnings.warn(
                "Using 'method_whitelist' with Retry is deprecated and "
                "will be removed in v2.0. Use 'allowed_methods' instead",
                DeprecationWarning,
            )
            allowed_methods = self.method_whitelist
        else:
            allowed_methods = self.allowed_methods

        if allowed_methods and method.upper() not in allowed_methods:
            return False
        return True

    def is_retry(self, method, status_code, has_retry_after=False):
        """Is this method/status code retryable? (Based on allowlists and control
        variables such as the number of total retries to allow, whether to
        respect the Retry-After header, whether this header is present, and
        whether the returned status code is on the list of status codes to
        be retried upon on the presence of the aforementioned header)
        """
        if not self._is_method_retryable(method):
            return False

        if self.status_forcelist and status_code in self.status_forcelist:
            return True

        return (
            self.total
            and self.respect_retry_after_header
            and has_retry_after
            and (status_code in self.RETRY_AFTER_STATUS_CODES)
        )

    def is_exhausted(self):
        """Are we out of retries?"""
        retry_counts = (
            self.total,
            self.connect,
            self.read,
            self.redirect,
            self.status,
            self.other,
        )
        retry_counts = list(filter(None, retry_counts))
        if not retry_counts:
            return False

        return min(retry_counts) < 0

    def increment(
        self,
        method=None,
        url=None,
        response=None,
        error=None,
        _pool=None,
        _stacktrace=None,
    ):
        """Return a new Retry object with incremented retry counters.

        :param response: A response object, or None, if the server did not
            return a response.
        :type response: :class:`~urllib3.response.HTTPResponse`
        :param Exception error: An error encountered during the request, or
            None if the response was received successfully.

        :return: A new ``Retry`` object.
        """
        if self.total is False and error:
            # Disabled, indicate to re-raise the error.
            raise six.reraise(type(error), error, _stacktrace)

        total = self.total
        if total is not None:
            total -= 1

        connect = self.connect
        read = self.read
        redirect = self.redirect
        status_count = self.status
        other = self.other
        cause = "unknown"
        status = None
        redirect_location = None

        if error and self._is_connection_error(error):
            # Connect retry?
            if connect is False:
                raise six.reraise(type(error), error, _stacktrace)
            elif connect is not None:
                connect -= 1

        elif error and self._is_read_error(error):
            # Read retry?
            if read is False or not self._is_method_retryable(method):
                raise six.reraise(type(error), error, _stacktrace)
            elif read is not None:
                read -= 1

        elif error:
            # Other retry?
            if other is not None:
                other -= 1

        elif response and response.get_redirect_location():
            # Redirect retry?
            if redirect is not None:
                redirect -= 1
            cause = "too many redirects"
            redirect_location = response.get_redirect_location()
            status = response.status

        else:
            # Incrementing because of a server error like a 500 in
            # status_forcelist and the given method is in the allowed_methods
            cause = ResponseError.GENERIC_ERROR
            if response and response.status:
                if status_count is not None:
                    status_count -= 1
                cause = ResponseError.SPECIFIC_ERROR.format(status_code=response.status)
                status = response.status

        history = self.history + (
            RequestHistory(method, url, error, status, redirect_location),
        )

        new_retry = self.new(
            total=total,
            connect=connect,
            read=read,
            redirect=redirect,
            status=status_count,
            other=other,
            history=history,
        )

        if new_retry.is_exhausted():
            raise MaxRetryError(_pool, url, error or ResponseError(cause))

        log.debug("Incremented Retry for (url='%s'): %r", url, new_retry)

        return new_retry

    def __repr__(self):
        return (
            "{cls.__name__}(total={self.total}, connect={self.connect}, "
            "read={self.read}, redirect={self.redirect}, status={self.status})"
        ).format(cls=type(self), self=self)

    def __getattr__(self, item):
        if item == "method_whitelist":
            # TODO: Remove this deprecated alias in v2.0
            warnings.warn(
                "Using 'method_whitelist' with Retry is deprecated and "
                "will be removed in v2.0. Use 'allowed_methods' instead",
                DeprecationWarning,
            )
            return self.allowed_methods
        try:
            return getattr(super(Retry, self), item)
        except AttributeError:
            return getattr(Retry, item)


# For backwards compatibility (equivalent to pre-v1.9):
Retry.DEFAULT = Retry(3)
site-packages/pip/_vendor/urllib3/util/proxy.py000064400000003105150564745010015602 0ustar00from .ssl_ import create_urllib3_context, resolve_cert_reqs, resolve_ssl_version


def connection_requires_http_tunnel(
    proxy_url=None, proxy_config=None, destination_scheme=None
):
    """
    Returns True if the connection requires an HTTP CONNECT through the proxy.

    :param URL proxy_url:
        URL of the proxy.
    :param ProxyConfig proxy_config:
        Proxy configuration from poolmanager.py
    :param str destination_scheme:
        The scheme of the destination. (i.e https, http, etc)
    """
    # If we're not using a proxy, no way to use a tunnel.
    if proxy_url is None:
        return False

    # HTTP destinations never require tunneling, we always forward.
    if destination_scheme == "http":
        return False

    # Support for forwarding with HTTPS proxies and HTTPS destinations.
    if (
        proxy_url.scheme == "https"
        and proxy_config
        and proxy_config.use_forwarding_for_https
    ):
        return False

    # Otherwise always use a tunnel.
    return True


def create_proxy_ssl_context(
    ssl_version, cert_reqs, ca_certs=None, ca_cert_dir=None, ca_cert_data=None
):
    """
    Generates a default proxy ssl context if one hasn't been provided by the
    user.
    """
    ssl_context = create_urllib3_context(
        ssl_version=resolve_ssl_version(ssl_version),
        cert_reqs=resolve_cert_reqs(cert_reqs),
    )

    if (
        not ca_certs
        and not ca_cert_dir
        and not ca_cert_data
        and hasattr(ssl_context, "load_default_certs")
    ):
        ssl_context.load_default_certs()

    return ssl_context
site-packages/pip/_vendor/urllib3/util/url.py000064400000033337150564745010015235 0ustar00from __future__ import absolute_import

import re
from collections import namedtuple

from ..exceptions import LocationParseError
from ..packages import six

url_attrs = ["scheme", "auth", "host", "port", "path", "query", "fragment"]

# We only want to normalize urls with an HTTP(S) scheme.
# urllib3 infers URLs without a scheme (None) to be http.
NORMALIZABLE_SCHEMES = ("http", "https", None)

# Almost all of these patterns were derived from the
# 'rfc3986' module: https://github.com/python-hyper/rfc3986
PERCENT_RE = re.compile(r"%[a-fA-F0-9]{2}")
SCHEME_RE = re.compile(r"^(?:[a-zA-Z][a-zA-Z0-9+-]*:|/)")
URI_RE = re.compile(
    r"^(?:([a-zA-Z][a-zA-Z0-9+.-]*):)?"
    r"(?://([^\\/?#]*))?"
    r"([^?#]*)"
    r"(?:\?([^#]*))?"
    r"(?:#(.*))?$",
    re.UNICODE | re.DOTALL,
)

IPV4_PAT = r"(?:[0-9]{1,3}\.){3}[0-9]{1,3}"
HEX_PAT = "[0-9A-Fa-f]{1,4}"
LS32_PAT = "(?:{hex}:{hex}|{ipv4})".format(hex=HEX_PAT, ipv4=IPV4_PAT)
_subs = {"hex": HEX_PAT, "ls32": LS32_PAT}
_variations = [
    #                            6( h16 ":" ) ls32
    "(?:%(hex)s:){6}%(ls32)s",
    #                       "::" 5( h16 ":" ) ls32
    "::(?:%(hex)s:){5}%(ls32)s",
    # [               h16 ] "::" 4( h16 ":" ) ls32
    "(?:%(hex)s)?::(?:%(hex)s:){4}%(ls32)s",
    # [ *1( h16 ":" ) h16 ] "::" 3( h16 ":" ) ls32
    "(?:(?:%(hex)s:)?%(hex)s)?::(?:%(hex)s:){3}%(ls32)s",
    # [ *2( h16 ":" ) h16 ] "::" 2( h16 ":" ) ls32
    "(?:(?:%(hex)s:){0,2}%(hex)s)?::(?:%(hex)s:){2}%(ls32)s",
    # [ *3( h16 ":" ) h16 ] "::"    h16 ":"   ls32
    "(?:(?:%(hex)s:){0,3}%(hex)s)?::%(hex)s:%(ls32)s",
    # [ *4( h16 ":" ) h16 ] "::"              ls32
    "(?:(?:%(hex)s:){0,4}%(hex)s)?::%(ls32)s",
    # [ *5( h16 ":" ) h16 ] "::"              h16
    "(?:(?:%(hex)s:){0,5}%(hex)s)?::%(hex)s",
    # [ *6( h16 ":" ) h16 ] "::"
    "(?:(?:%(hex)s:){0,6}%(hex)s)?::",
]

UNRESERVED_PAT = r"ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._!\-~"
IPV6_PAT = "(?:" + "|".join([x % _subs for x in _variations]) + ")"
ZONE_ID_PAT = "(?:%25|%)(?:[" + UNRESERVED_PAT + "]|%[a-fA-F0-9]{2})+"
IPV6_ADDRZ_PAT = r"\[" + IPV6_PAT + r"(?:" + ZONE_ID_PAT + r")?\]"
REG_NAME_PAT = r"(?:[^\[\]%:/?#]|%[a-fA-F0-9]{2})*"
TARGET_RE = re.compile(r"^(/[^?#]*)(?:\?([^#]*))?(?:#.*)?$")

IPV4_RE = re.compile("^" + IPV4_PAT + "$")
IPV6_RE = re.compile("^" + IPV6_PAT + "$")
IPV6_ADDRZ_RE = re.compile("^" + IPV6_ADDRZ_PAT + "$")
BRACELESS_IPV6_ADDRZ_RE = re.compile("^" + IPV6_ADDRZ_PAT[2:-2] + "$")
ZONE_ID_RE = re.compile("(" + ZONE_ID_PAT + r")\]$")

_HOST_PORT_PAT = ("^(%s|%s|%s)(?::([0-9]{0,5}))?$") % (
    REG_NAME_PAT,
    IPV4_PAT,
    IPV6_ADDRZ_PAT,
)
_HOST_PORT_RE = re.compile(_HOST_PORT_PAT, re.UNICODE | re.DOTALL)

UNRESERVED_CHARS = set(
    "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789._-~"
)
SUB_DELIM_CHARS = set("!$&'()*+,;=")
USERINFO_CHARS = UNRESERVED_CHARS | SUB_DELIM_CHARS | {":"}
PATH_CHARS = USERINFO_CHARS | {"@", "/"}
QUERY_CHARS = FRAGMENT_CHARS = PATH_CHARS | {"?"}


class Url(namedtuple("Url", url_attrs)):
    """
    Data structure for representing an HTTP URL. Used as a return value for
    :func:`parse_url`. Both the scheme and host are normalized as they are
    both case-insensitive according to RFC 3986.
    """

    __slots__ = ()

    def __new__(
        cls,
        scheme=None,
        auth=None,
        host=None,
        port=None,
        path=None,
        query=None,
        fragment=None,
    ):
        if path and not path.startswith("/"):
            path = "/" + path
        if scheme is not None:
            scheme = scheme.lower()
        return super(Url, cls).__new__(
            cls, scheme, auth, host, port, path, query, fragment
        )

    @property
    def hostname(self):
        """For backwards-compatibility with urlparse. We're nice like that."""
        return self.host

    @property
    def request_uri(self):
        """Absolute path including the query string."""
        uri = self.path or "/"

        if self.query is not None:
            uri += "?" + self.query

        return uri

    @property
    def netloc(self):
        """Network location including host and port"""
        if self.port:
            return "%s:%d" % (self.host, self.port)
        return self.host

    @property
    def url(self):
        """
        Convert self into a url

        This function should more or less round-trip with :func:`.parse_url`. The
        returned url may not be exactly the same as the url inputted to
        :func:`.parse_url`, but it should be equivalent by the RFC (e.g., urls
        with a blank port will have : removed).

        Example: ::

            >>> U = parse_url('http://google.com/mail/')
            >>> U.url
            'http://google.com/mail/'
            >>> Url('http', 'username:password', 'host.com', 80,
            ... '/path', 'query', 'fragment').url
            'http://username:password@host.com:80/path?query#fragment'
        """
        scheme, auth, host, port, path, query, fragment = self
        url = u""

        # We use "is not None" we want things to happen with empty strings (or 0 port)
        if scheme is not None:
            url += scheme + u"://"
        if auth is not None:
            url += auth + u"@"
        if host is not None:
            url += host
        if port is not None:
            url += u":" + str(port)
        if path is not None:
            url += path
        if query is not None:
            url += u"?" + query
        if fragment is not None:
            url += u"#" + fragment

        return url

    def __str__(self):
        return self.url


def split_first(s, delims):
    """
    .. deprecated:: 1.25

    Given a string and an iterable of delimiters, split on the first found
    delimiter. Return two split parts and the matched delimiter.

    If not found, then the first part is the full input string.

    Example::

        >>> split_first('foo/bar?baz', '?/=')
        ('foo', 'bar?baz', '/')
        >>> split_first('foo/bar?baz', '123')
        ('foo/bar?baz', '', None)

    Scales linearly with number of delims. Not ideal for large number of delims.
    """
    min_idx = None
    min_delim = None
    for d in delims:
        idx = s.find(d)
        if idx < 0:
            continue

        if min_idx is None or idx < min_idx:
            min_idx = idx
            min_delim = d

    if min_idx is None or min_idx < 0:
        return s, "", None

    return s[:min_idx], s[min_idx + 1 :], min_delim


def _encode_invalid_chars(component, allowed_chars, encoding="utf-8"):
    """Percent-encodes a URI component without reapplying
    onto an already percent-encoded component.
    """
    if component is None:
        return component

    component = six.ensure_text(component)

    # Normalize existing percent-encoded bytes.
    # Try to see if the component we're encoding is already percent-encoded
    # so we can skip all '%' characters but still encode all others.
    component, percent_encodings = PERCENT_RE.subn(
        lambda match: match.group(0).upper(), component
    )

    uri_bytes = component.encode("utf-8", "surrogatepass")
    is_percent_encoded = percent_encodings == uri_bytes.count(b"%")
    encoded_component = bytearray()

    for i in range(0, len(uri_bytes)):
        # Will return a single character bytestring on both Python 2 & 3
        byte = uri_bytes[i : i + 1]
        byte_ord = ord(byte)
        if (is_percent_encoded and byte == b"%") or (
            byte_ord < 128 and byte.decode() in allowed_chars
        ):
            encoded_component += byte
            continue
        encoded_component.extend(b"%" + (hex(byte_ord)[2:].encode().zfill(2).upper()))

    return encoded_component.decode(encoding)


def _remove_path_dot_segments(path):
    # See http://tools.ietf.org/html/rfc3986#section-5.2.4 for pseudo-code
    segments = path.split("/")  # Turn the path into a list of segments
    output = []  # Initialize the variable to use to store output

    for segment in segments:
        # '.' is the current directory, so ignore it, it is superfluous
        if segment == ".":
            continue
        # Anything other than '..', should be appended to the output
        elif segment != "..":
            output.append(segment)
        # In this case segment == '..', if we can, we should pop the last
        # element
        elif output:
            output.pop()

    # If the path starts with '/' and the output is empty or the first string
    # is non-empty
    if path.startswith("/") and (not output or output[0]):
        output.insert(0, "")

    # If the path starts with '/.' or '/..' ensure we add one more empty
    # string to add a trailing '/'
    if path.endswith(("/.", "/..")):
        output.append("")

    return "/".join(output)


def _normalize_host(host, scheme):
    if host:
        if isinstance(host, six.binary_type):
            host = six.ensure_str(host)

        if scheme in NORMALIZABLE_SCHEMES:
            is_ipv6 = IPV6_ADDRZ_RE.match(host)
            if is_ipv6:
                match = ZONE_ID_RE.search(host)
                if match:
                    start, end = match.span(1)
                    zone_id = host[start:end]

                    if zone_id.startswith("%25") and zone_id != "%25":
                        zone_id = zone_id[3:]
                    else:
                        zone_id = zone_id[1:]
                    zone_id = "%" + _encode_invalid_chars(zone_id, UNRESERVED_CHARS)
                    return host[:start].lower() + zone_id + host[end:]
                else:
                    return host.lower()
            elif not IPV4_RE.match(host):
                return six.ensure_str(
                    b".".join([_idna_encode(label) for label in host.split(".")])
                )
    return host


def _idna_encode(name):
    if name and any([ord(x) > 128 for x in name]):
        try:
            from pip._vendor import idna
        except ImportError:
            six.raise_from(
                LocationParseError("Unable to parse URL without the 'idna' module"),
                None,
            )
        try:
            return idna.encode(name.lower(), strict=True, std3_rules=True)
        except idna.IDNAError:
            six.raise_from(
                LocationParseError(u"Name '%s' is not a valid IDNA label" % name), None
            )
    return name.lower().encode("ascii")


def _encode_target(target):
    """Percent-encodes a request target so that there are no invalid characters"""
    path, query = TARGET_RE.match(target).groups()
    target = _encode_invalid_chars(path, PATH_CHARS)
    query = _encode_invalid_chars(query, QUERY_CHARS)
    if query is not None:
        target += "?" + query
    return target


def parse_url(url):
    """
    Given a url, return a parsed :class:`.Url` namedtuple. Best-effort is
    performed to parse incomplete urls. Fields not provided will be None.
    This parser is RFC 3986 compliant.

    The parser logic and helper functions are based heavily on
    work done in the ``rfc3986`` module.

    :param str url: URL to parse into a :class:`.Url` namedtuple.

    Partly backwards-compatible with :mod:`urlparse`.

    Example::

        >>> parse_url('http://google.com/mail/')
        Url(scheme='http', host='google.com', port=None, path='/mail/', ...)
        >>> parse_url('google.com:80')
        Url(scheme=None, host='google.com', port=80, path=None, ...)
        >>> parse_url('/foo?bar')
        Url(scheme=None, host=None, port=None, path='/foo', query='bar', ...)
    """
    if not url:
        # Empty
        return Url()

    source_url = url
    if not SCHEME_RE.search(url):
        url = "//" + url

    try:
        scheme, authority, path, query, fragment = URI_RE.match(url).groups()
        normalize_uri = scheme is None or scheme.lower() in NORMALIZABLE_SCHEMES

        if scheme:
            scheme = scheme.lower()

        if authority:
            auth, _, host_port = authority.rpartition("@")
            auth = auth or None
            host, port = _HOST_PORT_RE.match(host_port).groups()
            if auth and normalize_uri:
                auth = _encode_invalid_chars(auth, USERINFO_CHARS)
            if port == "":
                port = None
        else:
            auth, host, port = None, None, None

        if port is not None:
            port = int(port)
            if not (0 <= port <= 65535):
                raise LocationParseError(url)

        host = _normalize_host(host, scheme)

        if normalize_uri and path:
            path = _remove_path_dot_segments(path)
            path = _encode_invalid_chars(path, PATH_CHARS)
        if normalize_uri and query:
            query = _encode_invalid_chars(query, QUERY_CHARS)
        if normalize_uri and fragment:
            fragment = _encode_invalid_chars(fragment, FRAGMENT_CHARS)

    except (ValueError, AttributeError):
        return six.raise_from(LocationParseError(source_url), None)

    # For the sake of backwards compatibility we put empty
    # string values for path if there are any defined values
    # beyond the path in the URL.
    # TODO: Remove this when we break backwards compatibility.
    if not path:
        if query is not None or fragment is not None:
            path = ""
        else:
            path = None

    # Ensure that each part of the URL is a `str` for
    # backwards compatibility.
    if isinstance(url, six.text_type):
        ensure_func = six.ensure_text
    else:
        ensure_func = six.ensure_str

    def ensure_type(x):
        return x if x is None else ensure_func(x)

    return Url(
        scheme=ensure_type(scheme),
        auth=ensure_type(auth),
        host=ensure_type(host),
        port=port,
        path=ensure_type(path),
        query=ensure_type(query),
        fragment=ensure_type(fragment),
    )


def get_host(url):
    """
    Deprecated. Use :func:`parse_url` instead.
    """
    p = parse_url(url)
    return p.scheme or "http", p.hostname, p.port
site-packages/pip/_vendor/urllib3/util/ssltransport.py000064400000015423150564745010017205 0ustar00import io
import socket
import ssl

from pip._vendor.urllib3.exceptions import ProxySchemeUnsupported
from pip._vendor.urllib3.packages import six

SSL_BLOCKSIZE = 16384


class SSLTransport:
    """
    The SSLTransport wraps an existing socket and establishes an SSL connection.

    Contrary to Python's implementation of SSLSocket, it allows you to chain
    multiple TLS connections together. It's particularly useful if you need to
    implement TLS within TLS.

    The class supports most of the socket API operations.
    """

    @staticmethod
    def _validate_ssl_context_for_tls_in_tls(ssl_context):
        """
        Raises a ProxySchemeUnsupported if the provided ssl_context can't be used
        for TLS in TLS.

        The only requirement is that the ssl_context provides the 'wrap_bio'
        methods.
        """

        if not hasattr(ssl_context, "wrap_bio"):
            if six.PY2:
                raise ProxySchemeUnsupported(
                    "TLS in TLS requires SSLContext.wrap_bio() which isn't "
                    "supported on Python 2"
                )
            else:
                raise ProxySchemeUnsupported(
                    "TLS in TLS requires SSLContext.wrap_bio() which isn't "
                    "available on non-native SSLContext"
                )

    def __init__(
        self, socket, ssl_context, server_hostname=None, suppress_ragged_eofs=True
    ):
        """
        Create an SSLTransport around socket using the provided ssl_context.
        """
        self.incoming = ssl.MemoryBIO()
        self.outgoing = ssl.MemoryBIO()

        self.suppress_ragged_eofs = suppress_ragged_eofs
        self.socket = socket

        self.sslobj = ssl_context.wrap_bio(
            self.incoming, self.outgoing, server_hostname=server_hostname
        )

        # Perform initial handshake.
        self._ssl_io_loop(self.sslobj.do_handshake)

    def __enter__(self):
        return self

    def __exit__(self, *_):
        self.close()

    def fileno(self):
        return self.socket.fileno()

    def read(self, len=1024, buffer=None):
        return self._wrap_ssl_read(len, buffer)

    def recv(self, len=1024, flags=0):
        if flags != 0:
            raise ValueError("non-zero flags not allowed in calls to recv")
        return self._wrap_ssl_read(len)

    def recv_into(self, buffer, nbytes=None, flags=0):
        if flags != 0:
            raise ValueError("non-zero flags not allowed in calls to recv_into")
        if buffer and (nbytes is None):
            nbytes = len(buffer)
        elif nbytes is None:
            nbytes = 1024
        return self.read(nbytes, buffer)

    def sendall(self, data, flags=0):
        if flags != 0:
            raise ValueError("non-zero flags not allowed in calls to sendall")
        count = 0
        with memoryview(data) as view, view.cast("B") as byte_view:
            amount = len(byte_view)
            while count < amount:
                v = self.send(byte_view[count:])
                count += v

    def send(self, data, flags=0):
        if flags != 0:
            raise ValueError("non-zero flags not allowed in calls to send")
        response = self._ssl_io_loop(self.sslobj.write, data)
        return response

    def makefile(
        self, mode="r", buffering=None, encoding=None, errors=None, newline=None
    ):
        """
        Python's httpclient uses makefile and buffered io when reading HTTP
        messages and we need to support it.

        This is unfortunately a copy and paste of socket.py makefile with small
        changes to point to the socket directly.
        """
        if not set(mode) <= {"r", "w", "b"}:
            raise ValueError("invalid mode %r (only r, w, b allowed)" % (mode,))

        writing = "w" in mode
        reading = "r" in mode or not writing
        assert reading or writing
        binary = "b" in mode
        rawmode = ""
        if reading:
            rawmode += "r"
        if writing:
            rawmode += "w"
        raw = socket.SocketIO(self, rawmode)
        self.socket._io_refs += 1
        if buffering is None:
            buffering = -1
        if buffering < 0:
            buffering = io.DEFAULT_BUFFER_SIZE
        if buffering == 0:
            if not binary:
                raise ValueError("unbuffered streams must be binary")
            return raw
        if reading and writing:
            buffer = io.BufferedRWPair(raw, raw, buffering)
        elif reading:
            buffer = io.BufferedReader(raw, buffering)
        else:
            assert writing
            buffer = io.BufferedWriter(raw, buffering)
        if binary:
            return buffer
        text = io.TextIOWrapper(buffer, encoding, errors, newline)
        text.mode = mode
        return text

    def unwrap(self):
        self._ssl_io_loop(self.sslobj.unwrap)

    def close(self):
        self.socket.close()

    def getpeercert(self, binary_form=False):
        return self.sslobj.getpeercert(binary_form)

    def version(self):
        return self.sslobj.version()

    def cipher(self):
        return self.sslobj.cipher()

    def selected_alpn_protocol(self):
        return self.sslobj.selected_alpn_protocol()

    def selected_npn_protocol(self):
        return self.sslobj.selected_npn_protocol()

    def shared_ciphers(self):
        return self.sslobj.shared_ciphers()

    def compression(self):
        return self.sslobj.compression()

    def settimeout(self, value):
        self.socket.settimeout(value)

    def gettimeout(self):
        return self.socket.gettimeout()

    def _decref_socketios(self):
        self.socket._decref_socketios()

    def _wrap_ssl_read(self, len, buffer=None):
        try:
            return self._ssl_io_loop(self.sslobj.read, len, buffer)
        except ssl.SSLError as e:
            if e.errno == ssl.SSL_ERROR_EOF and self.suppress_ragged_eofs:
                return 0  # eof, return 0.
            else:
                raise

    def _ssl_io_loop(self, func, *args):
        """Performs an I/O loop between incoming/outgoing and the socket."""
        should_loop = True
        ret = None

        while should_loop:
            errno = None
            try:
                ret = func(*args)
            except ssl.SSLError as e:
                if e.errno not in (ssl.SSL_ERROR_WANT_READ, ssl.SSL_ERROR_WANT_WRITE):
                    # WANT_READ, and WANT_WRITE are expected, others are not.
                    raise e
                errno = e.errno

            buf = self.outgoing.read()
            self.socket.sendall(buf)

            if errno is None:
                should_loop = False
            elif errno == ssl.SSL_ERROR_WANT_READ:
                buf = self.socket.recv(SSL_BLOCKSIZE)
                if buf:
                    self.incoming.write(buf)
                else:
                    self.incoming.write_eof()
        return ret
site-packages/pip/_vendor/urllib3/poolmanager.py000064400000046463150564745010015766 0ustar00from __future__ import absolute_import

import collections
import functools
import logging

from ._collections import RecentlyUsedContainer
from .connectionpool import HTTPConnectionPool, HTTPSConnectionPool, port_by_scheme
from .exceptions import (
    LocationValueError,
    MaxRetryError,
    ProxySchemeUnknown,
    ProxySchemeUnsupported,
    URLSchemeUnknown,
)
from .packages import six
from .packages.six.moves.urllib.parse import urljoin
from .request import RequestMethods
from .util.proxy import connection_requires_http_tunnel
from .util.retry import Retry
from .util.url import parse_url

__all__ = ["PoolManager", "ProxyManager", "proxy_from_url"]


log = logging.getLogger(__name__)

SSL_KEYWORDS = (
    "key_file",
    "cert_file",
    "cert_reqs",
    "ca_certs",
    "ssl_version",
    "ca_cert_dir",
    "ssl_context",
    "key_password",
)

# All known keyword arguments that could be provided to the pool manager, its
# pools, or the underlying connections. This is used to construct a pool key.
_key_fields = (
    "key_scheme",  # str
    "key_host",  # str
    "key_port",  # int
    "key_timeout",  # int or float or Timeout
    "key_retries",  # int or Retry
    "key_strict",  # bool
    "key_block",  # bool
    "key_source_address",  # str
    "key_key_file",  # str
    "key_key_password",  # str
    "key_cert_file",  # str
    "key_cert_reqs",  # str
    "key_ca_certs",  # str
    "key_ssl_version",  # str
    "key_ca_cert_dir",  # str
    "key_ssl_context",  # instance of ssl.SSLContext or urllib3.util.ssl_.SSLContext
    "key_maxsize",  # int
    "key_headers",  # dict
    "key__proxy",  # parsed proxy url
    "key__proxy_headers",  # dict
    "key__proxy_config",  # class
    "key_socket_options",  # list of (level (int), optname (int), value (int or str)) tuples
    "key__socks_options",  # dict
    "key_assert_hostname",  # bool or string
    "key_assert_fingerprint",  # str
    "key_server_hostname",  # str
)

#: The namedtuple class used to construct keys for the connection pool.
#: All custom key schemes should include the fields in this key at a minimum.
PoolKey = collections.namedtuple("PoolKey", _key_fields)

_proxy_config_fields = ("ssl_context", "use_forwarding_for_https")
ProxyConfig = collections.namedtuple("ProxyConfig", _proxy_config_fields)


def _default_key_normalizer(key_class, request_context):
    """
    Create a pool key out of a request context dictionary.

    According to RFC 3986, both the scheme and host are case-insensitive.
    Therefore, this function normalizes both before constructing the pool
    key for an HTTPS request. If you wish to change this behaviour, provide
    alternate callables to ``key_fn_by_scheme``.

    :param key_class:
        The class to use when constructing the key. This should be a namedtuple
        with the ``scheme`` and ``host`` keys at a minimum.
    :type  key_class: namedtuple
    :param request_context:
        A dictionary-like object that contain the context for a request.
    :type  request_context: dict

    :return: A namedtuple that can be used as a connection pool key.
    :rtype:  PoolKey
    """
    # Since we mutate the dictionary, make a copy first
    context = request_context.copy()
    context["scheme"] = context["scheme"].lower()
    context["host"] = context["host"].lower()

    # These are both dictionaries and need to be transformed into frozensets
    for key in ("headers", "_proxy_headers", "_socks_options"):
        if key in context and context[key] is not None:
            context[key] = frozenset(context[key].items())

    # The socket_options key may be a list and needs to be transformed into a
    # tuple.
    socket_opts = context.get("socket_options")
    if socket_opts is not None:
        context["socket_options"] = tuple(socket_opts)

    # Map the kwargs to the names in the namedtuple - this is necessary since
    # namedtuples can't have fields starting with '_'.
    for key in list(context.keys()):
        context["key_" + key] = context.pop(key)

    # Default to ``None`` for keys missing from the context
    for field in key_class._fields:
        if field not in context:
            context[field] = None

    return key_class(**context)


#: A dictionary that maps a scheme to a callable that creates a pool key.
#: This can be used to alter the way pool keys are constructed, if desired.
#: Each PoolManager makes a copy of this dictionary so they can be configured
#: globally here, or individually on the instance.
key_fn_by_scheme = {
    "http": functools.partial(_default_key_normalizer, PoolKey),
    "https": functools.partial(_default_key_normalizer, PoolKey),
}

pool_classes_by_scheme = {"http": HTTPConnectionPool, "https": HTTPSConnectionPool}


class PoolManager(RequestMethods):
    """
    Allows for arbitrary requests while transparently keeping track of
    necessary connection pools for you.

    :param num_pools:
        Number of connection pools to cache before discarding the least
        recently used pool.

    :param headers:
        Headers to include with all requests, unless other headers are given
        explicitly.

    :param \\**connection_pool_kw:
        Additional parameters are used to create fresh
        :class:`urllib3.connectionpool.ConnectionPool` instances.

    Example::

        >>> manager = PoolManager(num_pools=2)
        >>> r = manager.request('GET', 'http://google.com/')
        >>> r = manager.request('GET', 'http://google.com/mail')
        >>> r = manager.request('GET', 'http://yahoo.com/')
        >>> len(manager.pools)
        2

    """

    proxy = None
    proxy_config = None

    def __init__(self, num_pools=10, headers=None, **connection_pool_kw):
        RequestMethods.__init__(self, headers)
        self.connection_pool_kw = connection_pool_kw
        self.pools = RecentlyUsedContainer(num_pools, dispose_func=lambda p: p.close())

        # Locally set the pool classes and keys so other PoolManagers can
        # override them.
        self.pool_classes_by_scheme = pool_classes_by_scheme
        self.key_fn_by_scheme = key_fn_by_scheme.copy()

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        self.clear()
        # Return False to re-raise any potential exceptions
        return False

    def _new_pool(self, scheme, host, port, request_context=None):
        """
        Create a new :class:`urllib3.connectionpool.ConnectionPool` based on host, port, scheme, and
        any additional pool keyword arguments.

        If ``request_context`` is provided, it is provided as keyword arguments
        to the pool class used. This method is used to actually create the
        connection pools handed out by :meth:`connection_from_url` and
        companion methods. It is intended to be overridden for customization.
        """
        pool_cls = self.pool_classes_by_scheme[scheme]
        if request_context is None:
            request_context = self.connection_pool_kw.copy()

        # Although the context has everything necessary to create the pool,
        # this function has historically only used the scheme, host, and port
        # in the positional args. When an API change is acceptable these can
        # be removed.
        for key in ("scheme", "host", "port"):
            request_context.pop(key, None)

        if scheme == "http":
            for kw in SSL_KEYWORDS:
                request_context.pop(kw, None)

        return pool_cls(host, port, **request_context)

    def clear(self):
        """
        Empty our store of pools and direct them all to close.

        This will not affect in-flight connections, but they will not be
        re-used after completion.
        """
        self.pools.clear()

    def connection_from_host(self, host, port=None, scheme="http", pool_kwargs=None):
        """
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the host, port, and scheme.

        If ``port`` isn't given, it will be derived from the ``scheme`` using
        ``urllib3.connectionpool.port_by_scheme``. If ``pool_kwargs`` is
        provided, it is merged with the instance's ``connection_pool_kw``
        variable and used to create the new connection pool, if one is
        needed.
        """

        if not host:
            raise LocationValueError("No host specified.")

        request_context = self._merge_pool_kwargs(pool_kwargs)
        request_context["scheme"] = scheme or "http"
        if not port:
            port = port_by_scheme.get(request_context["scheme"].lower(), 80)
        request_context["port"] = port
        request_context["host"] = host

        return self.connection_from_context(request_context)

    def connection_from_context(self, request_context):
        """
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the request context.

        ``request_context`` must at least contain the ``scheme`` key and its
        value must be a key in ``key_fn_by_scheme`` instance variable.
        """
        scheme = request_context["scheme"].lower()
        pool_key_constructor = self.key_fn_by_scheme.get(scheme)
        if not pool_key_constructor:
            raise URLSchemeUnknown(scheme)
        pool_key = pool_key_constructor(request_context)

        return self.connection_from_pool_key(pool_key, request_context=request_context)

    def connection_from_pool_key(self, pool_key, request_context=None):
        """
        Get a :class:`urllib3.connectionpool.ConnectionPool` based on the provided pool key.

        ``pool_key`` should be a namedtuple that only contains immutable
        objects. At a minimum it must have the ``scheme``, ``host``, and
        ``port`` fields.
        """
        with self.pools.lock:
            # If the scheme, host, or port doesn't match existing open
            # connections, open a new ConnectionPool.
            pool = self.pools.get(pool_key)
            if pool:
                return pool

            # Make a fresh ConnectionPool of the desired type
            scheme = request_context["scheme"]
            host = request_context["host"]
            port = request_context["port"]
            pool = self._new_pool(scheme, host, port, request_context=request_context)
            self.pools[pool_key] = pool

        return pool

    def connection_from_url(self, url, pool_kwargs=None):
        """
        Similar to :func:`urllib3.connectionpool.connection_from_url`.

        If ``pool_kwargs`` is not provided and a new pool needs to be
        constructed, ``self.connection_pool_kw`` is used to initialize
        the :class:`urllib3.connectionpool.ConnectionPool`. If ``pool_kwargs``
        is provided, it is used instead. Note that if a new pool does not
        need to be created for the request, the provided ``pool_kwargs`` are
        not used.
        """
        u = parse_url(url)
        return self.connection_from_host(
            u.host, port=u.port, scheme=u.scheme, pool_kwargs=pool_kwargs
        )

    def _merge_pool_kwargs(self, override):
        """
        Merge a dictionary of override values for self.connection_pool_kw.

        This does not modify self.connection_pool_kw and returns a new dict.
        Any keys in the override dictionary with a value of ``None`` are
        removed from the merged dictionary.
        """
        base_pool_kwargs = self.connection_pool_kw.copy()
        if override:
            for key, value in override.items():
                if value is None:
                    try:
                        del base_pool_kwargs[key]
                    except KeyError:
                        pass
                else:
                    base_pool_kwargs[key] = value
        return base_pool_kwargs

    def _proxy_requires_url_absolute_form(self, parsed_url):
        """
        Indicates if the proxy requires the complete destination URL in the
        request.  Normally this is only needed when not using an HTTP CONNECT
        tunnel.
        """
        if self.proxy is None:
            return False

        return not connection_requires_http_tunnel(
            self.proxy, self.proxy_config, parsed_url.scheme
        )

    def _validate_proxy_scheme_url_selection(self, url_scheme):
        """
        Validates that were not attempting to do TLS in TLS connections on
        Python2 or with unsupported SSL implementations.
        """
        if self.proxy is None or url_scheme != "https":
            return

        if self.proxy.scheme != "https":
            return

        if six.PY2 and not self.proxy_config.use_forwarding_for_https:
            raise ProxySchemeUnsupported(
                "Contacting HTTPS destinations through HTTPS proxies "
                "'via CONNECT tunnels' is not supported in Python 2"
            )

    def urlopen(self, method, url, redirect=True, **kw):
        """
        Same as :meth:`urllib3.HTTPConnectionPool.urlopen`
        with custom cross-host redirect logic and only sends the request-uri
        portion of the ``url``.

        The given ``url`` parameter must be absolute, such that an appropriate
        :class:`urllib3.connectionpool.ConnectionPool` can be chosen for it.
        """
        u = parse_url(url)
        self._validate_proxy_scheme_url_selection(u.scheme)

        conn = self.connection_from_host(u.host, port=u.port, scheme=u.scheme)

        kw["assert_same_host"] = False
        kw["redirect"] = False

        if "headers" not in kw:
            kw["headers"] = self.headers.copy()

        if self._proxy_requires_url_absolute_form(u):
            response = conn.urlopen(method, url, **kw)
        else:
            response = conn.urlopen(method, u.request_uri, **kw)

        redirect_location = redirect and response.get_redirect_location()
        if not redirect_location:
            return response

        # Support relative URLs for redirecting.
        redirect_location = urljoin(url, redirect_location)

        # RFC 7231, Section 6.4.4
        if response.status == 303:
            method = "GET"

        retries = kw.get("retries")
        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect)

        # Strip headers marked as unsafe to forward to the redirected location.
        # Check remove_headers_on_redirect to avoid a potential network call within
        # conn.is_same_host() which may use socket.gethostbyname() in the future.
        if retries.remove_headers_on_redirect and not conn.is_same_host(
            redirect_location
        ):
            headers = list(six.iterkeys(kw["headers"]))
            for header in headers:
                if header.lower() in retries.remove_headers_on_redirect:
                    kw["headers"].pop(header, None)

        try:
            retries = retries.increment(method, url, response=response, _pool=conn)
        except MaxRetryError:
            if retries.raise_on_redirect:
                response.drain_conn()
                raise
            return response

        kw["retries"] = retries
        kw["redirect"] = redirect

        log.info("Redirecting %s -> %s", url, redirect_location)

        response.drain_conn()
        return self.urlopen(method, redirect_location, **kw)


class ProxyManager(PoolManager):
    """
    Behaves just like :class:`PoolManager`, but sends all requests through
    the defined proxy, using the CONNECT method for HTTPS URLs.

    :param proxy_url:
        The URL of the proxy to be used.

    :param proxy_headers:
        A dictionary containing headers that will be sent to the proxy. In case
        of HTTP they are being sent with each request, while in the
        HTTPS/CONNECT case they are sent only once. Could be used for proxy
        authentication.

    :param proxy_ssl_context:
        The proxy SSL context is used to establish the TLS connection to the
        proxy when using HTTPS proxies.

    :param use_forwarding_for_https:
        (Defaults to False) If set to True will forward requests to the HTTPS
        proxy to be made on behalf of the client instead of creating a TLS
        tunnel via the CONNECT method. **Enabling this flag means that request
        and response headers and content will be visible from the HTTPS proxy**
        whereas tunneling keeps request and response headers and content
        private.  IP address, target hostname, SNI, and port are always visible
        to an HTTPS proxy even when this flag is disabled.

    Example:
        >>> proxy = urllib3.ProxyManager('http://localhost:3128/')
        >>> r1 = proxy.request('GET', 'http://google.com/')
        >>> r2 = proxy.request('GET', 'http://httpbin.org/')
        >>> len(proxy.pools)
        1
        >>> r3 = proxy.request('GET', 'https://httpbin.org/')
        >>> r4 = proxy.request('GET', 'https://twitter.com/')
        >>> len(proxy.pools)
        3

    """

    def __init__(
        self,
        proxy_url,
        num_pools=10,
        headers=None,
        proxy_headers=None,
        proxy_ssl_context=None,
        use_forwarding_for_https=False,
        **connection_pool_kw
    ):

        if isinstance(proxy_url, HTTPConnectionPool):
            proxy_url = "%s://%s:%i" % (
                proxy_url.scheme,
                proxy_url.host,
                proxy_url.port,
            )
        proxy = parse_url(proxy_url)

        if proxy.scheme not in ("http", "https"):
            raise ProxySchemeUnknown(proxy.scheme)

        if not proxy.port:
            port = port_by_scheme.get(proxy.scheme, 80)
            proxy = proxy._replace(port=port)

        self.proxy = proxy
        self.proxy_headers = proxy_headers or {}
        self.proxy_ssl_context = proxy_ssl_context
        self.proxy_config = ProxyConfig(proxy_ssl_context, use_forwarding_for_https)

        connection_pool_kw["_proxy"] = self.proxy
        connection_pool_kw["_proxy_headers"] = self.proxy_headers
        connection_pool_kw["_proxy_config"] = self.proxy_config

        super(ProxyManager, self).__init__(num_pools, headers, **connection_pool_kw)

    def connection_from_host(self, host, port=None, scheme="http", pool_kwargs=None):
        if scheme == "https":
            return super(ProxyManager, self).connection_from_host(
                host, port, scheme, pool_kwargs=pool_kwargs
            )

        return super(ProxyManager, self).connection_from_host(
            self.proxy.host, self.proxy.port, self.proxy.scheme, pool_kwargs=pool_kwargs
        )

    def _set_proxy_headers(self, url, headers=None):
        """
        Sets headers needed by proxies: specifically, the Accept and Host
        headers. Only sets headers not provided by the user.
        """
        headers_ = {"Accept": "*/*"}

        netloc = parse_url(url).netloc
        if netloc:
            headers_["Host"] = netloc

        if headers:
            headers_.update(headers)
        return headers_

    def urlopen(self, method, url, redirect=True, **kw):
        "Same as HTTP(S)ConnectionPool.urlopen, ``url`` must be absolute."
        u = parse_url(url)
        if not connection_requires_http_tunnel(self.proxy, self.proxy_config, u.scheme):
            # For connections using HTTP CONNECT, httplib sets the necessary
            # headers on the CONNECT to the proxy. If we're not using CONNECT,
            # we'll definitely need to set 'Host' at the very least.
            headers = kw.get("headers", self.headers)
            kw["headers"] = self._set_proxy_headers(url, headers)

        return super(ProxyManager, self).urlopen(method, url, redirect=redirect, **kw)


def proxy_from_url(url, **kw):
    return ProxyManager(proxy_url=url, **kw)
site-packages/pip/_vendor/urllib3/connection.py000064400000047160150564745010015614 0ustar00from __future__ import absolute_import

import datetime
import logging
import os
import re
import socket
import warnings
from socket import error as SocketError
from socket import timeout as SocketTimeout

from .packages import six
from .packages.six.moves.http_client import HTTPConnection as _HTTPConnection
from .packages.six.moves.http_client import HTTPException  # noqa: F401
from .util.proxy import create_proxy_ssl_context

try:  # Compiled with SSL?
    import ssl

    BaseSSLError = ssl.SSLError
except (ImportError, AttributeError):  # Platform-specific: No SSL.
    ssl = None

    class BaseSSLError(BaseException):
        pass


try:
    # Python 3: not a no-op, we're adding this to the namespace so it can be imported.
    ConnectionError = ConnectionError
except NameError:
    # Python 2
    class ConnectionError(Exception):
        pass


try:  # Python 3:
    # Not a no-op, we're adding this to the namespace so it can be imported.
    BrokenPipeError = BrokenPipeError
except NameError:  # Python 2:

    class BrokenPipeError(Exception):
        pass


from ._collections import HTTPHeaderDict  # noqa (historical, removed in v2)
from ._version import __version__
from .exceptions import (
    ConnectTimeoutError,
    NewConnectionError,
    SubjectAltNameWarning,
    SystemTimeWarning,
)
from .packages.ssl_match_hostname import CertificateError, match_hostname
from .util import SKIP_HEADER, SKIPPABLE_HEADERS, connection
from .util.ssl_ import (
    assert_fingerprint,
    create_urllib3_context,
    is_ipaddress,
    resolve_cert_reqs,
    resolve_ssl_version,
    ssl_wrap_socket,
)

log = logging.getLogger(__name__)

port_by_scheme = {"http": 80, "https": 443}

# When it comes time to update this value as a part of regular maintenance
# (ie test_recent_date is failing) update it to ~6 months before the current date.
RECENT_DATE = datetime.date(2020, 7, 1)

_CONTAINS_CONTROL_CHAR_RE = re.compile(r"[^-!#$%&'*+.^_`|~0-9a-zA-Z]")


class HTTPConnection(_HTTPConnection, object):
    """
    Based on :class:`http.client.HTTPConnection` but provides an extra constructor
    backwards-compatibility layer between older and newer Pythons.

    Additional keyword parameters are used to configure attributes of the connection.
    Accepted parameters include:

    - ``strict``: See the documentation on :class:`urllib3.connectionpool.HTTPConnectionPool`
    - ``source_address``: Set the source address for the current connection.
    - ``socket_options``: Set specific options on the underlying socket. If not specified, then
      defaults are loaded from ``HTTPConnection.default_socket_options`` which includes disabling
      Nagle's algorithm (sets TCP_NODELAY to 1) unless the connection is behind a proxy.

      For example, if you wish to enable TCP Keep Alive in addition to the defaults,
      you might pass:

      .. code-block:: python

         HTTPConnection.default_socket_options + [
             (socket.SOL_SOCKET, socket.SO_KEEPALIVE, 1),
         ]

      Or you may want to disable the defaults by passing an empty list (e.g., ``[]``).
    """

    default_port = port_by_scheme["http"]

    #: Disable Nagle's algorithm by default.
    #: ``[(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)]``
    default_socket_options = [(socket.IPPROTO_TCP, socket.TCP_NODELAY, 1)]

    #: Whether this connection verifies the host's certificate.
    is_verified = False

    #: Whether this proxy connection (if used) verifies the proxy host's
    #: certificate.
    proxy_is_verified = None

    def __init__(self, *args, **kw):
        if not six.PY2:
            kw.pop("strict", None)

        # Pre-set source_address.
        self.source_address = kw.get("source_address")

        #: The socket options provided by the user. If no options are
        #: provided, we use the default options.
        self.socket_options = kw.pop("socket_options", self.default_socket_options)

        # Proxy options provided by the user.
        self.proxy = kw.pop("proxy", None)
        self.proxy_config = kw.pop("proxy_config", None)

        _HTTPConnection.__init__(self, *args, **kw)

    @property
    def host(self):
        """
        Getter method to remove any trailing dots that indicate the hostname is an FQDN.

        In general, SSL certificates don't include the trailing dot indicating a
        fully-qualified domain name, and thus, they don't validate properly when
        checked against a domain name that includes the dot. In addition, some
        servers may not expect to receive the trailing dot when provided.

        However, the hostname with trailing dot is critical to DNS resolution; doing a
        lookup with the trailing dot will properly only resolve the appropriate FQDN,
        whereas a lookup without a trailing dot will search the system's search domain
        list. Thus, it's important to keep the original host around for use only in
        those cases where it's appropriate (i.e., when doing DNS lookup to establish the
        actual TCP connection across which we're going to send HTTP requests).
        """
        return self._dns_host.rstrip(".")

    @host.setter
    def host(self, value):
        """
        Setter for the `host` property.

        We assume that only urllib3 uses the _dns_host attribute; httplib itself
        only uses `host`, and it seems reasonable that other libraries follow suit.
        """
        self._dns_host = value

    def _new_conn(self):
        """Establish a socket connection and set nodelay settings on it.

        :return: New socket connection.
        """
        extra_kw = {}
        if self.source_address:
            extra_kw["source_address"] = self.source_address

        if self.socket_options:
            extra_kw["socket_options"] = self.socket_options

        try:
            conn = connection.create_connection(
                (self._dns_host, self.port), self.timeout, **extra_kw
            )

        except SocketTimeout:
            raise ConnectTimeoutError(
                self,
                "Connection to %s timed out. (connect timeout=%s)"
                % (self.host, self.timeout),
            )

        except SocketError as e:
            raise NewConnectionError(
                self, "Failed to establish a new connection: %s" % e
            )

        return conn

    def _is_using_tunnel(self):
        # Google App Engine's httplib does not define _tunnel_host
        return getattr(self, "_tunnel_host", None)

    def _prepare_conn(self, conn):
        self.sock = conn
        if self._is_using_tunnel():
            # TODO: Fix tunnel so it doesn't depend on self.sock state.
            self._tunnel()
            # Mark this connection as not reusable
            self.auto_open = 0

    def connect(self):
        conn = self._new_conn()
        self._prepare_conn(conn)

    def putrequest(self, method, url, *args, **kwargs):
        """ """
        # Empty docstring because the indentation of CPython's implementation
        # is broken but we don't want this method in our documentation.
        match = _CONTAINS_CONTROL_CHAR_RE.search(method)
        if match:
            raise ValueError(
                "Method cannot contain non-token characters %r (found at least %r)"
                % (method, match.group())
            )

        return _HTTPConnection.putrequest(self, method, url, *args, **kwargs)

    def putheader(self, header, *values):
        """ """
        if not any(isinstance(v, str) and v == SKIP_HEADER for v in values):
            _HTTPConnection.putheader(self, header, *values)
        elif six.ensure_str(header.lower()) not in SKIPPABLE_HEADERS:
            raise ValueError(
                "urllib3.util.SKIP_HEADER only supports '%s'"
                % ("', '".join(map(str.title, sorted(SKIPPABLE_HEADERS))),)
            )

    def request(self, method, url, body=None, headers=None):
        if headers is None:
            headers = {}
        else:
            # Avoid modifying the headers passed into .request()
            headers = headers.copy()
        if "user-agent" not in (six.ensure_str(k.lower()) for k in headers):
            headers["User-Agent"] = _get_default_user_agent()
        super(HTTPConnection, self).request(method, url, body=body, headers=headers)

    def request_chunked(self, method, url, body=None, headers=None):
        """
        Alternative to the common request method, which sends the
        body with chunked encoding and not as one block
        """
        headers = headers or {}
        header_keys = set([six.ensure_str(k.lower()) for k in headers])
        skip_accept_encoding = "accept-encoding" in header_keys
        skip_host = "host" in header_keys
        self.putrequest(
            method, url, skip_accept_encoding=skip_accept_encoding, skip_host=skip_host
        )
        if "user-agent" not in header_keys:
            self.putheader("User-Agent", _get_default_user_agent())
        for header, value in headers.items():
            self.putheader(header, value)
        if "transfer-encoding" not in header_keys:
            self.putheader("Transfer-Encoding", "chunked")
        self.endheaders()

        if body is not None:
            stringish_types = six.string_types + (bytes,)
            if isinstance(body, stringish_types):
                body = (body,)
            for chunk in body:
                if not chunk:
                    continue
                if not isinstance(chunk, bytes):
                    chunk = chunk.encode("utf8")
                len_str = hex(len(chunk))[2:]
                to_send = bytearray(len_str.encode())
                to_send += b"\r\n"
                to_send += chunk
                to_send += b"\r\n"
                self.send(to_send)

        # After the if clause, to always have a closed body
        self.send(b"0\r\n\r\n")


class HTTPSConnection(HTTPConnection):
    """
    Many of the parameters to this constructor are passed to the underlying SSL
    socket by means of :py:func:`urllib3.util.ssl_wrap_socket`.
    """

    default_port = port_by_scheme["https"]

    cert_reqs = None
    ca_certs = None
    ca_cert_dir = None
    ca_cert_data = None
    ssl_version = None
    assert_fingerprint = None
    tls_in_tls_required = False

    def __init__(
        self,
        host,
        port=None,
        key_file=None,
        cert_file=None,
        key_password=None,
        strict=None,
        timeout=socket._GLOBAL_DEFAULT_TIMEOUT,
        ssl_context=None,
        server_hostname=None,
        **kw
    ):

        HTTPConnection.__init__(self, host, port, strict=strict, timeout=timeout, **kw)

        self.key_file = key_file
        self.cert_file = cert_file
        self.key_password = key_password
        self.ssl_context = ssl_context
        self.server_hostname = server_hostname

        # Required property for Google AppEngine 1.9.0 which otherwise causes
        # HTTPS requests to go out as HTTP. (See Issue #356)
        self._protocol = "https"

    def set_cert(
        self,
        key_file=None,
        cert_file=None,
        cert_reqs=None,
        key_password=None,
        ca_certs=None,
        assert_hostname=None,
        assert_fingerprint=None,
        ca_cert_dir=None,
        ca_cert_data=None,
    ):
        """
        This method should only be called once, before the connection is used.
        """
        # If cert_reqs is not provided we'll assume CERT_REQUIRED unless we also
        # have an SSLContext object in which case we'll use its verify_mode.
        if cert_reqs is None:
            if self.ssl_context is not None:
                cert_reqs = self.ssl_context.verify_mode
            else:
                cert_reqs = resolve_cert_reqs(None)

        self.key_file = key_file
        self.cert_file = cert_file
        self.cert_reqs = cert_reqs
        self.key_password = key_password
        self.assert_hostname = assert_hostname
        self.assert_fingerprint = assert_fingerprint
        self.ca_certs = ca_certs and os.path.expanduser(ca_certs)
        self.ca_cert_dir = ca_cert_dir and os.path.expanduser(ca_cert_dir)
        self.ca_cert_data = ca_cert_data

    def connect(self):
        # Add certificate verification
        conn = self._new_conn()
        hostname = self.host
        tls_in_tls = False

        if self._is_using_tunnel():
            if self.tls_in_tls_required:
                conn = self._connect_tls_proxy(hostname, conn)
                tls_in_tls = True

            self.sock = conn

            # Calls self._set_hostport(), so self.host is
            # self._tunnel_host below.
            self._tunnel()
            # Mark this connection as not reusable
            self.auto_open = 0

            # Override the host with the one we're requesting data from.
            hostname = self._tunnel_host

        server_hostname = hostname
        if self.server_hostname is not None:
            server_hostname = self.server_hostname

        is_time_off = datetime.date.today() < RECENT_DATE
        if is_time_off:
            warnings.warn(
                (
                    "System time is way off (before {0}). This will probably "
                    "lead to SSL verification errors"
                ).format(RECENT_DATE),
                SystemTimeWarning,
            )

        # Wrap socket using verification with the root certs in
        # trusted_root_certs
        default_ssl_context = False
        if self.ssl_context is None:
            default_ssl_context = True
            self.ssl_context = create_urllib3_context(
                ssl_version=resolve_ssl_version(self.ssl_version),
                cert_reqs=resolve_cert_reqs(self.cert_reqs),
            )

        context = self.ssl_context
        context.verify_mode = resolve_cert_reqs(self.cert_reqs)

        # Try to load OS default certs if none are given.
        # Works well on Windows (requires Python3.4+)
        if (
            not self.ca_certs
            and not self.ca_cert_dir
            and not self.ca_cert_data
            and default_ssl_context
            and hasattr(context, "load_default_certs")
        ):
            context.load_default_certs()

        self.sock = ssl_wrap_socket(
            sock=conn,
            keyfile=self.key_file,
            certfile=self.cert_file,
            key_password=self.key_password,
            ca_certs=self.ca_certs,
            ca_cert_dir=self.ca_cert_dir,
            ca_cert_data=self.ca_cert_data,
            server_hostname=server_hostname,
            ssl_context=context,
            tls_in_tls=tls_in_tls,
        )

        # If we're using all defaults and the connection
        # is TLSv1 or TLSv1.1 we throw a DeprecationWarning
        # for the host.
        if (
            default_ssl_context
            and self.ssl_version is None
            and hasattr(self.sock, "version")
            and self.sock.version() in {"TLSv1", "TLSv1.1"}
        ):
            warnings.warn(
                "Negotiating TLSv1/TLSv1.1 by default is deprecated "
                "and will be disabled in urllib3 v2.0.0. Connecting to "
                "'%s' with '%s' can be enabled by explicitly opting-in "
                "with 'ssl_version'" % (self.host, self.sock.version()),
                DeprecationWarning,
            )

        if self.assert_fingerprint:
            assert_fingerprint(
                self.sock.getpeercert(binary_form=True), self.assert_fingerprint
            )
        elif (
            context.verify_mode != ssl.CERT_NONE
            and not getattr(context, "check_hostname", False)
            and self.assert_hostname is not False
        ):
            # While urllib3 attempts to always turn off hostname matching from
            # the TLS library, this cannot always be done. So we check whether
            # the TLS Library still thinks it's matching hostnames.
            cert = self.sock.getpeercert()
            if not cert.get("subjectAltName", ()):
                warnings.warn(
                    (
                        "Certificate for {0} has no `subjectAltName`, falling back to check for a "
                        "`commonName` for now. This feature is being removed by major browsers and "
                        "deprecated by RFC 2818. (See https://github.com/urllib3/urllib3/issues/497 "
                        "for details.)".format(hostname)
                    ),
                    SubjectAltNameWarning,
                )
            _match_hostname(cert, self.assert_hostname or server_hostname)

        self.is_verified = (
            context.verify_mode == ssl.CERT_REQUIRED
            or self.assert_fingerprint is not None
        )

    def _connect_tls_proxy(self, hostname, conn):
        """
        Establish a TLS connection to the proxy using the provided SSL context.
        """
        proxy_config = self.proxy_config
        ssl_context = proxy_config.ssl_context
        if ssl_context:
            # If the user provided a proxy context, we assume CA and client
            # certificates have already been set
            return ssl_wrap_socket(
                sock=conn,
                server_hostname=hostname,
                ssl_context=ssl_context,
            )

        ssl_context = create_proxy_ssl_context(
            self.ssl_version,
            self.cert_reqs,
            self.ca_certs,
            self.ca_cert_dir,
            self.ca_cert_data,
        )

        # If no cert was provided, use only the default options for server
        # certificate validation
        socket = ssl_wrap_socket(
            sock=conn,
            ca_certs=self.ca_certs,
            ca_cert_dir=self.ca_cert_dir,
            ca_cert_data=self.ca_cert_data,
            server_hostname=hostname,
            ssl_context=ssl_context,
        )

        if ssl_context.verify_mode != ssl.CERT_NONE and not getattr(
            ssl_context, "check_hostname", False
        ):
            # While urllib3 attempts to always turn off hostname matching from
            # the TLS library, this cannot always be done. So we check whether
            # the TLS Library still thinks it's matching hostnames.
            cert = socket.getpeercert()
            if not cert.get("subjectAltName", ()):
                warnings.warn(
                    (
                        "Certificate for {0} has no `subjectAltName`, falling back to check for a "
                        "`commonName` for now. This feature is being removed by major browsers and "
                        "deprecated by RFC 2818. (See https://github.com/urllib3/urllib3/issues/497 "
                        "for details.)".format(hostname)
                    ),
                    SubjectAltNameWarning,
                )
            _match_hostname(cert, hostname)

        self.proxy_is_verified = ssl_context.verify_mode == ssl.CERT_REQUIRED
        return socket


def _match_hostname(cert, asserted_hostname):
    # Our upstream implementation of ssl.match_hostname()
    # only applies this normalization to IP addresses so it doesn't
    # match DNS SANs so we do the same thing!
    stripped_hostname = asserted_hostname.strip("u[]")
    if is_ipaddress(stripped_hostname):
        asserted_hostname = stripped_hostname

    try:
        match_hostname(cert, asserted_hostname)
    except CertificateError as e:
        log.warning(
            "Certificate did not match expected hostname: %s. Certificate: %s",
            asserted_hostname,
            cert,
        )
        # Add cert to exception and reraise so client code can inspect
        # the cert when catching the exception, if they want to
        e._peer_cert = cert
        raise


def _get_default_user_agent():
    return "python-urllib3/%s" % __version__


class DummyConnection(object):
    """Used to detect a failed ConnectionCls import."""

    pass


if not ssl:
    HTTPSConnection = DummyConnection  # noqa: F811


VerifiedHTTPSConnection = HTTPSConnection
site-packages/pip/_vendor/urllib3/contrib/__pycache__/ntlmpool.cpython-39.pyc000064400000007121150564745010023241 0ustar00a

��Re��@spdZddlmZddlZddlmZddlmZddlmZddl	m
Z
e�d	e�ee
�ZGd
d�de�ZdS)z
NTLM authenticating pool, contributed by erikcederstran

Issue #10, see: http://code.google.com/p/urllib3/issues/detail?id=10
�)�absolute_importN)�	getLogger)�ntlm�)�HTTPSConnectionPool)�HTTPSConnectiona#The 'urllib3.contrib.ntlmpool' module is deprecated and will be removed in urllib3 v2.0 release, urllib3 is not able to support it properly due to reasons listed in issue: https://github.com/urllib3/urllib3/issues/2282. If you are a user of this module please comment in the mentioned issue.cs:eZdZdZdZ�fdd�Zdd�Zd�fd
d�	Z�ZS)
�NTLMConnectionPoolzQ
    Implements an NTLM authentication version of an urllib3 connection pool
    �httpscsPtt|�j|i|��||_||_|�dd�}|d��|_|d|_||_	dS)z�
        authurl is a random URL on the server that is protected by NTLM.
        user is the Windows user, probably in the DOMAIN\username format.
        pw is the password for the user.
        �\�rN)
�superr�__init__�authurl�rawuser�split�upper�domain�user�pw)�selfrrr�args�kwargsZ
user_parts��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/ntlmpool.pyr
"s
zNTLMConnectionPool.__init__c
Cs�|jd7_t�d|j|j|j�ddi}d}d}t|j|jd�}dt�|j	�||<t�d	|�|�
d
|jd|�|��}t|�
��}t�d|j|j�t�d|�t�d
|�d��d|_||�d�}d}|D] }	|	dd�dkr�|	dd�}q�|du�rtd|||f��t�|�\}
}t�|
|j|j|j|�}d|||<t�d	|�|�
d
|jd|�|��}t�d|j|j�t�dt|�
���t�d
|��dd��|jdk�r�|jdk�r�td��td|j|jf��d|_t�d�|S)Nrz3Starting NTLM HTTPS connection no. %d: https://%s%s�
Connection�
Keep-Alive�
Authorizationzwww-authenticate)�host�portzNTLM %szRequest headers: %s�GETzResponse status: %s %szResponse headers: %szResponse data: %s [...]�dz, �zNTLM z!Unexpected %s response header: %s��i�z3Server rejected request: wrong username or passwordzWrong server response: %s %szConnection established)�num_connections�log�debugrrrr rZcreate_NTLM_NEGOTIATE_MESSAGEr�request�getresponse�dict�
getheaders�status�reason�read�fpr�	ExceptionZparse_NTLM_CHALLENGE_MESSAGEZ create_NTLM_AUTHENTICATE_MESSAGErrr)
r�headersZ
req_headerZresp_header�conn�resZreshdrZauth_header_valuesZauth_header_value�sZServerChallengeZNegotiateFlagsZauth_msgrrr�	_new_conn0sf��

���
zNTLMConnectionPool._new_connN�Tc	s0|duri}d|d<tt|��|||||||�S)Nrr)rr�urlopen)r�method�url�bodyr1�retries�redirect�assert_same_hostrrrr7ss

�zNTLMConnectionPool.urlopen)NNr6TT)	�__name__�
__module__�__qualname__�__doc__�schemer
r5r7�
__classcell__rrrrrsG�r)rA�
__future__r�warnings�loggingrr�rZpackages.six.moves.http_clientr�warn�DeprecationWarningr>r&rrrrr�<module>s�site-packages/pip/_vendor/urllib3/contrib/__pycache__/pyopenssl.cpython-39.pyc000064400000036432150564745010023440 0ustar00a

��ReB�@stdZddlmZddlZddlmZddlmZ	ddl
mZzddlm
Z
Wn"eynGdd	�d	e�Z
Yn0dd
lmZddlmZddlmZzdd
lmZWn"ey�dZddlmZYn0ddlZddlZddlZddlmZddlmZddl m!Z!ddgZ"dZ#ej$ej%j&e!ej%j&ej'ej%j(iZ)e*ed��rXe*ej%d��rXej%j+e)ej,<e*ed��r�e*ej%d��r�ej%j-e)ej.<e*ed��r�e*ej%d��r�ej%j/e)ej0<ej1ej%j2ej3ej%j4ej5ej%j4ej%j6iZ7e8dd�e7�9�D��Z:dZ;ej#Z<ej=j>Z?e�@eA�ZBdd�ZCd d�ZDd!d"�ZEd#d$�ZFd%d&�ZGGd'd(�d(eH�ZIe�rNd0d*d+�ZJneZJeJeI_JGd,d-�d-eH�ZKd.d/�ZLdS)1a�
TLS with SNI_-support for Python 2. Follow these instructions if you would
like to verify TLS certificates in Python 2. Note, the default libraries do
*not* do certificate checking; you need to do additional work to validate
certificates yourself.

This needs the following packages installed:

* `pyOpenSSL`_ (tested with 16.0.0)
* `cryptography`_ (minimum 1.3.4, from pyopenssl)
* `idna`_ (minimum 2.0, from cryptography)

However, pyopenssl depends on cryptography, which depends on idna, so while we
use all three directly here we end up having relatively few packages required.

You can install them with the following command:

.. code-block:: bash

    $ python -m pip install pyopenssl cryptography idna

To activate certificate checking, call
:func:`~urllib3.contrib.pyopenssl.inject_into_urllib3` from your Python code
before you begin making HTTP requests. This can be done in a ``sitecustomize``
module, or at any other time before your application begins using ``urllib3``,
like this:

.. code-block:: python

    try:
        import pip._vendor.urllib3.contrib.pyopenssl as pyopenssl
        pyopenssl.inject_into_urllib3()
    except ImportError:
        pass

Now you can use :mod:`urllib3` as you normally would, and it will support SNI
when the required modules are installed.

Activating this module also has the positive side effect of disabling SSL/TLS
compression in Python 2 (see `CRIME attack`_).

.. _sni: https://en.wikipedia.org/wiki/Server_Name_Indication
.. _crime attack: https://en.wikipedia.org/wiki/CRIME_(security_exploit)
.. _pyopenssl: https://www.pyopenssl.org
.. _cryptography: https://cryptography.io
.. _idna: https://github.com/kjd/idna
�)�absolute_importN)�x509)�backend)�_Certificate)�UnsupportedExtensionc@seZdZdS)rN)�__name__�
__module__�__qualname__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/pyopenssl.pyr;sr)�BytesIO)�error)�timeout)�_fileobject�)�backport_makefile)�util)�six)�PROTOCOL_TLS_CLIENT�inject_into_urllib3�extract_from_urllib3T�PROTOCOL_SSLv3�SSLv3_METHOD�PROTOCOL_TLSv1_1�TLSv1_1_METHOD�PROTOCOL_TLSv1_2�TLSv1_2_METHODccs|]\}}||fVqdS�Nr
)�.0�k�vr
r
r�	<genexpr>m�r!i@cCs4t�tt_ttj_tt_ttj_dt_dtj_dS)z7Monkey-patch urllib3 with PyOpenSSL-backed SSL-support.TN)�_validate_dependencies_met�PyOpenSSLContextr�
SSLContext�ssl_�HAS_SNI�IS_PYOPENSSLr
r
r
rryscCs.tt_ttj_tt_ttj_dt_dtj_dS)z4Undo monkey-patching by :func:`inject_into_urllib3`.FN)�orig_util_SSLContextrr%r&�orig_util_HAS_SNIr'r(r
r
r
rr�scCsRddlm}t|dd�dur$td��ddlm}|�}t|dd�durNtd��dS)	z{
    Verifies that PyOpenSSL's package-level dependencies have been met.
    Throws `ImportError` if they are not met.
    r)�
Extensions�get_extension_for_classNzX'cryptography' module missing required functionality.  Try upgrading to v1.3.4 or newer.)�X509�_x509zS'pyOpenSSL' module missing required functionality. Try upgrading to v0.14 or newer.)Zcryptography.x509.extensionsr+�getattr�ImportErrorZOpenSSL.cryptor-)r+r-rr
r
rr#�s��r#cCs@dd�}d|vr|S||�}|dur(dStjdkr<|�d�}|S)a%
    Converts a dNSName SubjectAlternativeName field to the form used by the
    standard library on the given Python version.

    Cryptography produces a dNSName as a unicode string that was idna-decoded
    from ASCII bytes. We need to idna-encode that string to get it back, and
    then on Python 3 we also need to convert to unicode via UTF-8 (the stdlib
    uses PyUnicode_FromStringAndSize on it, which decodes via UTF-8).

    If the name cannot be idna-encoded then we return None signalling that
    the name given should be skipped.
    cSstddlm}zJdD]8}|�|�r|t|�d�}|�d�|�|�WSq|�|�WS|jjynYdS0dS)z�
        Borrowed wholesale from the Python Cryptography Project. It turns out
        that we can't just safely call `idna.encode`: it can explode for
        wildcard names. This avoids that problem.
        r)�idna)z*.�.N�ascii)�pip._vendorr1�
startswith�len�encode�core�	IDNAError)�namer1�prefixr
r
r�idna_encode�s
z'_dnsname_to_stdlib.<locals>.idna_encode�:N)�r�utf-8)�sys�version_info�decode)r:r<r
r
r�_dnsname_to_stdlib�s

rCcCs�t|d�r|��}ntt|j�}z|j�tj�j	}WnVtj
yLgYStjttj
tfy�}zt�d|�gWYd}~Sd}~00dd�tt|�tj��D�}|�dd�|�tj�D��|S)zU
    Given an PyOpenSSL certificate, provides all the subject alternative names.
    �to_cryptographyz�A problem was encountered with the certificate that prevented urllib3 from finding the SubjectAlternativeName field. This can affect certificate validation. The error was %sNcSsg|]}|durd|f�qS)N�DNSr
�rr:r
r
r�
<listcomp>s�z%get_subj_alt_name.<locals>.<listcomp>css|]}dt|�fVqdS)z
IP AddressN)�strrFr
r
rr!sz$get_subj_alt_name.<locals>.<genexpr>)�hasattrrDr�openssl_backendr.�
extensionsr,rZSubjectAlternativeName�valueZExtensionNotFoundZDuplicateExtensionrZUnsupportedGeneralNameType�UnicodeError�log�warning�maprCZget_values_for_typeZDNSName�extendZ	IPAddress)Z	peer_cert�cert�ext�e�namesr
r
r�get_subj_alt_name�s2

��	�

�rVc@s�eZdZdZd!dd�Zdd�Zdd�Zd	d
�Zdd�Zd
d�Z	dd�Z
dd�Zdd�Zdd�Z
d"dd�Zdd�Zdd�Zdd�Zd S)#�
WrappedSocketz�API-compatibility wrapper for Python OpenSSL's Connection-class.

    Note: _makefile_refs, _drop() and _reuse() are needed for the garbage
    collector of pypy.
    TcCs"||_||_||_d|_d|_dS�NrF)�
connection�socket�suppress_ragged_eofs�_makefile_refs�_closed)�selfrYrZr[r
r
r�__init__s
zWrappedSocket.__init__cCs
|j��Sr)rZ�fileno�r^r
r
rr`szWrappedSocket.filenocCs*|jdkr|jd8_|jr&|��dS)Nr�)r\r]�closerar
r
r�_decref_socketioss
zWrappedSocket._decref_socketiosc
Osz|jj|i|��}W�n�tjjyj}z6|jrJ|jdkrJWYd}~dStt|���WYd}~n�d}~0tjj	y�|j�
�tjjkr�YdS�Yn�tjjy�t
�|j|j���s�td��n|j|i|��YSYn<tjj�y}zt�d|��WYd}~nd}~00|SdS)N����zUnexpected EOFr"�The read operation timed out�read error: %r)rY�recv�OpenSSL�SSL�SysCallErrorr[�args�SocketErrorrH�ZeroReturnError�get_shutdown�RECEIVED_SHUTDOWN�
WantReadErrorr�
wait_for_readrZ�
gettimeoutr�Error�ssl�SSLError)r^rm�kwargs�datarTr
r
rri$s" 
$zWrappedSocket.recvc
Osz|jj|i|��WStjjyf}z6|jrF|jdkrFWYd}~dStt|���WYd}~n�d}~0tjj	y�|j�
�tjjkr�YdS�Yn~tjjy�t
�|j|j���s�td��n|j|i|��YSYn8tjj�y}zt�d|��WYd}~n
d}~00dS)Nrerrgrh)rY�	recv_intorjrkrlr[rmrnrHrorprqrrrrsrZrtrrurvrw)r^rmrxrTr
r
rrz=s  
zWrappedSocket.recv_intocCs|j�|�Sr)rZ�
settimeout)r^rr
r
rr{TszWrappedSocket.settimeoutc
Cs|z|j�|�WStjjyBt�|j|j���s:t	��YqYqtjj
yt}ztt|���WYd}~qd}~00qdSr)
rY�sendrjrkZWantWriteErrorr�wait_for_writerZrtrrlrnrH)r^ryrTr
r
r�_send_until_doneWszWrappedSocket._send_until_donecCs4d}|t|�kr0|�|||t��}||7}qdS�Nr)r6r~�SSL_WRITE_BLOCKSIZE)r^ry�
total_sent�sentr
r
r�sendallbs�zWrappedSocket.sendallcCs|j��dSr)rY�shutdownrar
r
rr�jszWrappedSocket.shutdowncCsJ|jdkr8zd|_|j��WStjjy4YdS0n|jd8_dS)NrbT)r\r]rYrcrjrkrurar
r
rrcns

zWrappedSocket.closeFcCsD|j��}|s|S|r(tj�tjj|�Sd|��jffft|�d�S)N�
commonName)�subject�subjectAltName)	rYZget_peer_certificaterjZcryptoZdump_certificateZ
FILETYPE_ASN1Zget_subjectZCNrV)r^�binary_formrr
r
r�getpeercertxs
�zWrappedSocket.getpeercertcCs
|j��Sr)rYZget_protocol_version_namerar
r
r�version�szWrappedSocket.versioncCs|jd7_dS�Nrb)r\rar
r
r�_reuse�szWrappedSocket._reusecCs&|jdkr|��n|jd8_dSr�)r\rcrar
r
r�_drop�s

zWrappedSocket._dropN)T)F)rrr	�__doc__r_r`rdrirzr{r~r�r�rcr�r�r�r�r
r
r
rrWs


rWrfcCs|jd7_t|||dd�S)NrbT)rc)r\r)r^�mode�bufsizer
r
r�makefile�sr�c@s�eZdZdZdd�Zedd��Zejdd��Zedd��Zejd	d��Zd
d�Z	dd
�Z
ddd�Zddd�Zdd�Z
ddd�ZdS)r$z�
    I am a wrapper class for the PyOpenSSL ``Context`` object. I am responsible
    for translating the interface of the standard library ``SSLContext`` object
    to calls into PyOpenSSL.
    cCs*t||_tj�|j�|_d|_d|_dSrX)�_openssl_versions�protocolrjrk�Context�_ctx�_options�check_hostname)r^r�r
r
rr_�s
zPyOpenSSLContext.__init__cCs|jSr)r�rar
r
r�options�szPyOpenSSLContext.optionscCs||_|j�|�dSr)r�r�Zset_options�r^rLr
r
rr��scCst|j��Sr)�_openssl_to_stdlib_verifyr�Zget_verify_moderar
r
r�verify_mode�szPyOpenSSLContext.verify_modecCs|j�t|t�dSr)r�Z
set_verify�_stdlib_to_openssl_verify�_verify_callbackr�r
r
rr��scCs|j��dSr)r��set_default_verify_pathsrar
r
rr��sz)PyOpenSSLContext.set_default_verify_pathscCs&t|tj�r|�d�}|j�|�dS)Nr?)�
isinstancer�	text_typer7r�Zset_cipher_list)r^�ciphersr
r
r�set_ciphers�s
zPyOpenSSLContext.set_ciphersNc
Cs�|dur|�d�}|dur$|�d�}z*|j�||�|durL|j�t|��Wn6tjjy�}zt�d|��WYd}~n
d}~00dS)Nr?z'unable to load trusted certificates: %r)	r7r��load_verify_locationsrrjrkrurvrw)r^�cafile�capath�cadatarTr
r
rr��s

z&PyOpenSSLContext.load_verify_locationscsR|j�|��dur>t�tj�s*��d��|j��fdd��|j�|pJ|�dS)Nr?cs�Srr
)�_��passwordr
r�<lambda>�r"z2PyOpenSSLContext.load_cert_chain.<locals>.<lambda>)r�Zuse_certificate_chain_filer�r�binary_typer7Z
set_passwd_cbZuse_privatekey_file)r^�certfile�keyfiler�r
r�r�load_cert_chain�s
z PyOpenSSLContext.load_cert_chaincCsdd�|D�}|j�|�S)NcSsg|]}t�|��qSr
)r�
ensure_binary)r�pr
r
rrG�r"z7PyOpenSSLContext.set_alpn_protocols.<locals>.<listcomp>)r�Zset_alpn_protos)r^�	protocolsr
r
r�set_alpn_protocols�sz#PyOpenSSLContext.set_alpn_protocolsFTc
Cs�tj�|j|�}t|tj�r&|�d�}|dur8|�|�|�	�z|�
�Wq�tjjy~t�
||���svtd��Yq@Yq�tjjy�}zt�d|��WYd}~q�d}~00q�q@t||�S)Nr?zselect timed outzbad handshake: %r)rjrk�
Connectionr�r�rr�r7Zset_tlsext_host_nameZset_connect_state�do_handshakerrrrsrtrrurvrwrW)r^�sock�server_side�do_handshake_on_connectr[�server_hostname�cnxrTr
r
r�wrap_socket�s 

$zPyOpenSSLContext.wrap_socket)NNN)NN)FTTN)rrr	r�r_�propertyr��setterr�r�r�r�r�r�r�r
r
r
rr$�s(





�r$cCs|dkSrr
)r�rZerr_noZ	err_depthZreturn_coder
r
rr��sr�)rf)Mr��
__future__rZOpenSSL.SSLrjZcryptographyrZ$cryptography.hazmat.backends.opensslrrJZ)cryptography.hazmat.backends.openssl.x509rZcryptography.x509rr0�	Exception�iorrZr
rnrrZpackages.backports.makefiler�loggingrvr@�r�packagesrZ	util.ssl_r�__all__r'�PROTOCOL_TLSrkZ
SSLv23_METHOD�PROTOCOL_TLSv1ZTLSv1_METHODr�rIrrrrrr�	CERT_NONEZVERIFY_NONE�
CERT_OPTIONALZVERIFY_PEER�
CERT_REQUIREDZVERIFY_FAIL_IF_NO_PEER_CERTr��dict�itemsr�r�r*r&r%r)�	getLoggerrrNrrr#rCrV�objectrWr�r$r�r
r
r
r�<module>sv/

�


��

+6^site-packages/pip/_vendor/urllib3/contrib/__pycache__/_appengine_environ.cpython-39.pyc000064400000002662150564745010025247 0ustar00a

��Re��@s8dZddlZdd�Zdd�Zdd�Zd	d
�Zdd�ZdS)
zB
This module provides means to detect the App Engine environment.
�NcCst�p
t�S)N)�is_local_appengine�is_prod_appengine�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/_appengine_environ.py�is_appenginesrcCst�otjddkS)a#Reports if the app is running in the first generation sandbox.

    The second generation runtimes are technically still in a sandbox, but it
    is much less restrictive, so generally you shouldn't need to check for it.
    see https://cloud.google.com/appengine/docs/standard/runtimes
    �APPENGINE_RUNTIME�python27)r�os�environrrrr�is_appengine_sandboxsrcCsdtjvotj�dd��d�S)Nr�SERVER_SOFTWARE�zDevelopment/�r	r
�get�
startswithrrrrrs
��rcCsdtjvotj�dd��d�S)Nrrr
zGoogle App Engine/rrrrrrs
��rcCsdS)zDeprecated.Frrrrr�is_prod_appengine_mvms"sr)�__doc__r	rrrrrrrrr�<module>s
site-packages/pip/_vendor/urllib3/contrib/__pycache__/socks.cpython-39.pyc000064400000013052150564745010022517 0ustar00a

��Re��@s,dZddlmZzddlZWn4eyPddlZddlmZe�de��Yn0ddl	m
Zddl	mZ
dd	lmZmZdd
lmZmZddlmZmZddlmZdd
lmZzddlZWney�dZYn0Gdd�de�ZGdd�dee�ZGdd�de�ZGdd�de�ZGdd�de�ZdS)a�
This module contains provisional support for SOCKS proxies from within
urllib3. This module supports SOCKS4, SOCKS4A (an extension of SOCKS4), and
SOCKS5. To enable its functionality, either install PySocks or install this
module with the ``socks`` extra.

The SOCKS implementation supports the full range of urllib3 features. It also
supports the following SOCKS features:

- SOCKS4A (``proxy_url='socks4a://...``)
- SOCKS4 (``proxy_url='socks4://...``)
- SOCKS5 with remote DNS (``proxy_url='socks5h://...``)
- SOCKS5 with local DNS (``proxy_url='socks5://...``)
- Usernames and passwords for the SOCKS proxy

.. note::
   It is recommended to use ``socks5h://`` or ``socks4a://`` schemes in
   your ``proxy_url`` to ensure that DNS resolution is done from the remote
   server instead of client-side when connecting to a domain name.

SOCKS4 supports IPv4 and domain names with the SOCKS4A extension. SOCKS5
supports IPv4, IPv6, and domain names.

When connecting to a SOCKS4 proxy the ``username`` portion of the ``proxy_url``
will be sent as the ``userid`` section of the SOCKS request:

.. code-block:: python

    proxy_url="socks4a://<userid>@proxy-host"

When connecting to a SOCKS5 proxy the ``username`` and ``password`` portion
of the ``proxy_url`` will be sent as the username/password to authenticate
with the proxy:

.. code-block:: python

    proxy_url="socks5h://<username>:<password>@proxy-host"

�)�absolute_importN�)�DependencyWarningz�SOCKS support in urllib3 requires the installation of optional dependencies: specifically, PySocks.  For more information, see https://urllib3.readthedocs.io/en/1.26.x/contrib.html#socks-proxies)�error)�timeout)�HTTPConnection�HTTPSConnection)�HTTPConnectionPool�HTTPSConnectionPool)�ConnectTimeoutError�NewConnectionError)�PoolManager)�	parse_urlcs(eZdZdZ�fdd�Zdd�Z�ZS)�SOCKSConnectionzG
    A plain-text HTTP connection that connects via a SOCKS proxy.
    cs&|�d�|_tt|�j|i|��dS)N�_socks_options)�popr�superr�__init__)�self�args�kwargs��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/socks.pyrPszSOCKSConnection.__init__c
CsJi}|jr|j|d<|jr$|j|d<zTtj|j|jff|jd|jd|jd|jd|jd|jd|jd	�|��}Wn�ty�t	|d
|j|jf��Yn�tj
�y}zX|jr�|j}t|t�r�t	|d
|j|jf��q�t
|d|��nt
|d|��WYd}~n<d}~0t�yD}zt
|d|��WYd}~n
d}~00|S)
zA
        Establish a new connection via the SOCKS proxy.
        �source_address�socket_options�
socks_version�
proxy_host�
proxy_port�username�password�rdns)�
proxy_typeZ
proxy_addrrZproxy_usernameZproxy_passwordZ
proxy_rdnsrz0Connection to %s timed out. (connect timeout=%s)z(Failed to establish a new connection: %sN)rr�socks�create_connection�host�portrr�
SocketTimeoutr�
ProxyErrorZ
socket_err�
isinstancer�SocketError)r�extra_kw�conn�errrr�	_new_connTs^


��	�

��

�����zSOCKSConnection._new_conn)�__name__�
__module__�__qualname__�__doc__rr/�
__classcell__rrrrrKsrc@seZdZdS)�SOCKSHTTPSConnectionN)r0r1r2rrrrr5�sr5c@seZdZeZdS)�SOCKSHTTPConnectionPoolN)r0r1r2r�
ConnectionClsrrrrr6�sr6c@seZdZeZdS)�SOCKSHTTPSConnectionPoolN)r0r1r2r5r7rrrrr8�sr8cs,eZdZdZeed�Zd�fdd�	Z�ZS)�SOCKSProxyManagerzh
    A version of the urllib3 ProxyManager that routes connections via the
    defined SOCKS proxy.
    )�http�httpsN�
cs�t|�}|durB|durB|jdurB|j�d�}t|�dkrB|\}}|jdkrXtj}	d}
nN|jdkrntj}	d}
n8|jdkr�tj}	d}
n"|jdkr�tj}	d}
ntd	|��||_	|	|j
|j|||
d
�}||d<tt
|�j||fi|��t
j|_dS)N�:rZsocks5FZsocks5hTZsocks4Zsocks4az)Unable to determine SOCKS version from %s)rrrr r!r"r)r�auth�split�len�schemer$ZPROXY_TYPE_SOCKS5ZPROXY_TYPE_SOCKS4�
ValueError�	proxy_urlr&r'rr9r�pool_classes_by_scheme)rrCr r!�	num_pools�headers�connection_pool_kw�parsedr?rr"Z
socks_optionsrrrr�sB	



�
��zSOCKSProxyManager.__init__)NNr<N)	r0r1r2r3r6r8rDrr4rrrrr9�s��r9) r3�
__future__rr$�ImportError�warnings�
exceptionsr�warn�socketrr+rr(�
connectionrr�connectionpoolr	r
rr�poolmanagerr
Zutil.urlr�sslrr5r6r8r9rrrr�<module>s6'�
Hsite-packages/pip/_vendor/urllib3/contrib/__pycache__/securetransport.cpython-39.pyc000064400000052714150564745010024650 0ustar00a

��Re���@s�dZddlmZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
Z
ddlmZddlmZddlmZdd	lmZmZmZdd
lmZmZmZmZmZmZzddl	mZWn"e y�dZddl!m"Z"Yn0d
dgZ#dZ$ej$Z%ej&j'Z(e
�)�Z*e�+�Z,dZ-ej.ej/ej0ej1ej2ej3ej4ej5ej6ej7ej8ej9ej:ej;ej<ej=ej>ej?ej@ejAejBejCejDejEejFejGejHejIejJejKgZLejMejNejOfeejNejOfiZPeQe
d��r�ejRejRfePe
jS<eQe
d��r�ejTejTfePe
jU<eQe
d��rejNejNfePe
jV<eQe
d��r,ejWejWfePe
jX<eQe
d��rJejOejOfePe
jY<dd
�ZZdd�Z[dd�Z\dd�Z]e�^e\�Z_e�`e]�ZaGdd�deb�Zce�r�d%dd �Zdn
d&d"d �Zdedec_dGd#d$�d$eb�ZedS)'a�	
SecureTranport support for urllib3 via ctypes.

This makes platform-native TLS available to urllib3 users on macOS without the
use of a compiler. This is an important feature because the Python Package
Index is moving to become a TLSv1.2-or-higher server, and the default OpenSSL
that ships with macOS is not capable of doing TLSv1.2. The only way to resolve
this is to give macOS users an alternative solution to the problem, and that
solution is to use SecureTransport.

We use ctypes here because this solution must not require a compiler. That's
because pip is not allowed to require a compiler either.

This is not intended to be a seriously long-term solution to this problem.
The hope is that PEP 543 will eventually solve this issue for us, at which
point we can retire this contrib module. But in the short term, we need to
solve the impending tire fire that is Python on Mac without this kind of
contrib module. So...here we are.

To use this module, simply import and inject it::

    import pip._vendor.urllib3.contrib.securetransport as securetransport
    securetransport.inject_into_urllib3()

Happy TLSing!

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

.. code-block::

    Copyright (c) 2015-2016 Will Bond <will@wbond.net>

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
�)�absolute_importN)�six�)�util)�PROTOCOL_TLS_CLIENT�)�CoreFoundation�Security�
SecurityConst)�_assert_no_error�_build_tls_unknown_ca_alert�_cert_array_from_pem�_create_cfstring_array�_load_client_cert_chain�_temporary_keychain)�_fileobject��backport_makefile�inject_into_urllib3�extract_from_urllib3Ti@�PROTOCOL_SSLv2�PROTOCOL_SSLv3�PROTOCOL_TLSv1�PROTOCOL_TLSv1_1�PROTOCOL_TLSv1_2cCs.tt_ttj_tt_ttj_dt_dtj_dS)zG
    Monkey-patch urllib3 with SecureTransport-backed SSL-support.
    TN)�SecureTransportContextr�
SSLContext�ssl_�HAS_SNI�IS_SECURETRANSPORT�r r ��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/securetransport.pyr�scCs.tt_ttj_tt_ttj_dt_dtj_dS)z>
    Undo monkey-patching by :func:`inject_into_urllib3`.
    FN)�orig_util_SSLContextrrr�orig_util_HAS_SNIrrr r r r!r�sc

Cs�d}�zPt�|�}|dur"tjWS|j}|d}|��}d}d}z|||kr�|dusZ|dkrtt�||�stt�t	j
d��||}	tj|	�
||�}
|�|
|	�}||7}|sB|s�tjWWSq�qBWnztj�y6}z^|j	}|du�r"|t	j
k�r"||d<|t	jk�s|t	jk�r tjWYd}~WS�WYd}~n
d}~00||d<||k�rRtjWSWdSt�y�}z"|du�rx||_tjWYd}~Sd}~00dS)zs
    SecureTransport read callback. This is called by ST to request that data
    be returned from the socket.
    Nr�	timed out)�_connection_refs�getr
�errSSLInternal�socket�
gettimeoutr�
wait_for_read�error�errno�EAGAIN�ctypes�c_char�from_address�	recv_into�errSSLClosedGraceful�
ECONNRESET�EPIPE�errSSLClosedAbort�errSSLWouldBlock�	Exception�
_exception)
�
connection_id�data_buffer�data_length_pointer�wrapped_socket�base_socketZrequested_length�timeoutr+Z
read_count�	remaining�buffer�
chunk_size�er r r!�_read_callback�sP

�


rCc
Cs|d}�z4t�|�}|dur"tjWS|j}|d}t�||�}|��}d}d}	zV|	|kr�|dusf|dkr�t�	||�s�t�
tjd��|�
|�}
|	|
7}	||
d�}qNWnxtj
�y}z\|j}|du�r|tjk�r|	|d<|tjks�|tjk�rtjWYd}~WS�WYd}~n
d}~00|	|d<|	|k�r6tjWSWdSt�yv}z"|du�r\||_tjWYd}~Sd}~00dS)zx
    SecureTransport write callback. This is called by ST to request that data
    actually be sent on the network.
    Nrr$)r%r&r
r'r(r.�	string_atr)r�wait_for_writer+r,r-�sendr3r4r5r6r7r8)r9r:r;r<r=Zbytes_to_write�datar>r+�sentZ
chunk_sentrBr r r!�_write_callbacksD



rIc@s�eZdZdZdd�Zejdd��Zdd�Zdd	�Z	d
d�Z
dd
�Zdd�Zdd�Z
dd�Zdd�Zd.dd�Zdd�Zdd�Zdd�Zdd �Zd!d"�Zd#d$�Zd/d&d'�Zd(d)�Zd*d+�Zd,d-�ZdS)0�
WrappedSocketz�
    API-compatibility wrapper for Python's OpenSSL wrapped socket object.

    Note: _makefile_refs, _drop(), and _reuse() are needed for the garbage
    collector of PyPy.
    cCsL||_d|_d|_d|_d|_d|_d|_d|_|j��|_	|j�
d�dS�NrF)r(�context�_makefile_refs�_closedr8�	_keychain�
_keychain_dir�_client_cert_chainr)�_timeout�
settimeout)�selfr(r r r!�__init__NszWrappedSocket.__init__ccs4d|_dV|jdur0|jd}|_|��|�dS)a]
        A context manager that can be used to wrap calls that do I/O from
        SecureTransport. If any of the I/O callbacks hit an exception, this
        context manager will correctly propagate the exception after the fact.
        This avoids silently swallowing those exceptions.

        It also correctly forces the socket closed.
        N)r8�close)rT�	exceptionr r r!�_raise_on_error`s

zWrappedSocket._raise_on_errorcCs2tjtt�t�}t�|j|tt��}t|�dS)a4
        Sets up the allowed ciphers. By default this matches the set in
        util.ssl_.DEFAULT_CIPHERS, at least as supported by macOS. This is done
        custom and doesn't allow changing at this time, mostly because parsing
        OpenSSL cipher strings is going to be a freaking nightmare.
        N)r	ZSSLCipherSuite�len�
CIPHER_SUITESZSSLSetEnabledCiphersrLr)rT�ciphers�resultr r r!�_set_ciphersus
�zWrappedSocket._set_ciphersc	CsF|sdSt|�}z$t�|j|�}t|�Wt�|�nt�|�0dS)z<
        Sets up the ALPN protocols on the context.
        N)rr	�SSLSetALPNProtocolsrLrr�	CFRelease)rT�	protocolsZ
protocols_arrr\r r r!�_set_alpn_protocols�s
z!WrappedSocket._set_alpn_protocolsc	
Cs�|sdStjtjf}z&|�|�}||vr.WdSd|f}Wn.tyh}zd|f}WYd}~n
d}~00t|���}|j�|�t	�
ddd�}|j�tjtj
|�|��t�d|��dS)z�
        Called when we have set custom validation. We do this in two cases:
        first, when cert validation is entirely disabled; and second, when
        using a custom trust DB.
        Raises an SSLError if the connection is not trusted.
        Nzerror code: %dz
exception: %r�iirrzcertificate verify failed, %s)r
ZkSecTrustResultUnspecifiedZkSecTrustResultProceed�_evaluate_trustr7r�versionr(�sendall�struct�pack�
setsockopt�
SOL_SOCKET�	SO_LINGERrV�ssl�SSLError)	rT�verify�trust_bundleZ	successes�trust_result�reasonrBZrec�optsr r r!�_custom_validate�s$�
 zWrappedSocket._custom_validatec	Cstj�|�r>t|d��}|��}Wd�n1s40Yd}t��}z�t|�}t�|j	t
�|��}t|�|s~t
�d��t�||�}t|�t�|d�}t|�t��}t�|t
�|��}t|�W|r�t�|�|dur�t�|�n$|r�t�|�|du�rt�|�0|jS)N�rbzFailed to copy trust referenceT)�os�path�isfile�open�readr	�SecTrustRefr
�SSLCopyPeerTrustrLr.�byrefrrkrlZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZSecTrustResultTypeZSecTrustEvaluaterr_�value)rTrn�fZ
cert_array�trustr\ror r r!rc�s6&


�

zWrappedSocket._evaluate_trustc
Cst�dtjtj�|_t�|jtt�}
t	|
�t
�:t|�d}|tvrT|dd}q>|t|<Wd�n1sp0Yt�
|j|�}
t	|
�|r�t|t�s�|�d�}t�|j|t|��}
t	|
�|��|�|	�t�|j|�}
t	|
�t�|j|�}
t	|
�|�r|du�r,t�|jtjd�}
t	|
�|�rht�\|_|_t|j||�|_t�|j|j�}
t	|
�|���vt� |j�}
|
tj!k�r�t"�#d��nB|
tj$k�r�|�%||�Wd��qhnt	|
�Wd��q�Wd�n1�s�0Y�qhdS)z�
        Actually performs the TLS handshake. This is run automatically by
        wrapped socket, and shouldn't be needed in user code.
        Ni���rzutf-8Tzhandshake timed out)&r	ZSSLCreateContextr
ZkSSLClientSideZkSSLStreamTyperLZ
SSLSetIOFuncs�_read_callback_pointer�_write_callback_pointerr�_connection_ref_lock�idr%ZSSLSetConnection�
isinstance�bytes�encodeZSSLSetPeerDomainNamerYr]raZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxZSSLSetSessionOptionZ"kSSLSessionOptionBreakOnServerAuthrrOrPrrQZSSLSetCertificaterXZSSLHandshaker6r(r>ZerrSSLServerAuthCompletedrr)rT�server_hostnamermrn�min_versionZmax_version�client_certZ
client_keyZclient_key_passphrase�alpn_protocolsr\�handler r r!�	handshake�s`
��&

�

��
zWrappedSocket.handshakecCs
|j��S�N)r(�fileno�rTr r r!r�7szWrappedSocket.filenocCs*|jdkr|jd8_|jr&|��dS)Nrr)rMrNrVr�r r r!�_decref_socketios;s
zWrappedSocket._decref_socketioscCs&t�|�}|�||�}|d|�}|Sr�)r.�create_string_bufferr1)rTZbufsizr@�
bytes_readrGr r r!�recvAs
zWrappedSocket.recvNc	Cs�|jr
dS|durt|�}tj|�|�}t�d�}|���(t�|j	||t�
|��}Wd�n1sj0Y|tjkr�|j
dkr�t�d��n"|tjtjfvr�|��nt|�|j
S)Nrzrecv timed out)rNrYr.r/�from_buffer�c_size_trXr	ZSSLReadrLr{r
r6r|r(r>r2ZerrSSLClosedNoNotifyrVr)rTr@�nbytes�processed_bytesr\r r r!r1Gs(

�"

�
zWrappedSocket.recv_intocCs
||_dSr��rR)rTr>r r r!rSrszWrappedSocket.settimeoutcCs|jSr�r�r�r r r!r)uszWrappedSocket.gettimeoutc	Cs|t�d�}|���,t�|j|t|�t�|��}Wd�n1sD0Y|tj	krn|j
dkrnt�d��nt
|�|j
S)Nrzsend timed out)r.r�rXr	ZSSLWriterLrYr{r
r6r|r(r>r)rTrGr�r\r r r!rFxs

�"zWrappedSocket.sendcCs4d}|t|�kr0|�|||t��}||7}qdS�Nr)rYrF�SSL_WRITE_BLOCKSIZE)rTrG�
total_sentrHr r r!re�szWrappedSocket.sendallcCs8|���t�|j�Wd�n1s*0YdSr�)rXr	ZSSLCloserLr�r r r!�shutdown�s
zWrappedSocket.shutdowncCs�|jdkr�d|_|jr(t�|j�d|_|jr@t�|j�d|_|jrvt�|j�t�|j�t	�
|j�d|_|_|j�
�S|jd8_dS)NrT)rMrNrLrr_rQrOr	ZSecKeychainDelete�shutil�rmtreerPr(rVr�r r r!rV�s

zWrappedSocket.closeFc
	Cs|std��t��}d}d}z�t�|jt�|��}t|�|s`W|rNt�	|�|r\t�	|�dSt�
|�}|s�W|r~t�	|�|r�t�	|�dSt�|d�}|s�J�t�|�}|s�J�t�
|�}t�|�}	t�|	|�}W|r�t�	|�|r�t�	|�n"|�rt�	|�|�rt�	|�0|S)Nz2SecureTransport only supports dumping binary certsr)�
ValueErrorr	ryrzrLr.r{rrr_ZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexZSecCertificateCopyDataZCFDataGetLengthZCFDataGetBytePtrrD)
rT�binary_formr~ZcertdataZ	der_bytesr\Z
cert_countZleafZdata_lengthr:r r r!�getpeercert�sN

�



�



�
zWrappedSocket.getpeercertcCs�t��}t�|jt�|��}t|�|jtj	kr<t
�d��n^|jtjkrLdS|jtj
kr\dS|jtjkrldS|jtjkr|dS|jtjkr�dSt
�d|��dS)Nz(SecureTransport does not support TLS 1.3zTLSv1.2zTLSv1.1�TLSv1�SSLv3ZSSLv2zUnknown TLS version: %r)r	�SSLProtocolZSSLGetNegotiatedProtocolVersionrLr.r{rr|r
ZkTLSProtocol13rkrl�kTLSProtocol12�kTLSProtocol11�
kTLSProtocol1�
kSSLProtocol3�
kSSLProtocol2)rT�protocolr\r r r!rd�s$�zWrappedSocket.versioncCs|jd7_dS�Nr)rMr�r r r!�_reuse�szWrappedSocket._reusecCs&|jdkr|��n|jd8_dSr�)rMrVr�r r r!�_drop�s

zWrappedSocket._drop)N)F)�__name__�
__module__�__qualname__�__doc__rU�
contextlib�contextmanagerrXr]rarrrcr�r�r�r�r1rSr)rFrer�rVr�rdr�r�r r r r!rJFs.


#(]
+
:rJ���cCs|jd7_t|||dd�S)NrT)rV)rMr)rT�mode�bufsizer r r!�makefilesr��rcOsd}t|||g|�Ri|��Sr�r)rTr��	buffering�args�kwargsr r r!r�sc@s�eZdZdZdd�Zedd��Zejdd��Zedd��Zejd	d��Zed
d��Z	e	jdd��Z	d
d�Z
dd�Zdd�Zddd�Z
ddd�Zdd�Zd dd�ZdS)!rz�
    I am a wrapper class for the SecureTransport library, to translate the
    interface of the standard library ``SSLContext`` object to calls into
    SecureTransport.
    cCs>t|\|_|_d|_d|_d|_d|_d|_d|_d|_	dSrK)
�_protocol_to_min_max�_min_version�_max_version�_options�_verify�
_trust_bundle�_client_cert�_client_key�_client_key_passphrase�_alpn_protocols)rTr�r r r!rUszSecureTransportContext.__init__cCsdS)��
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        Tr r�r r r!�check_hostname#sz%SecureTransportContext.check_hostnamecCsdS)r�Nr �rTr|r r r!r�+scCs|jSr��r�r�r r r!�options3szSecureTransportContext.optionscCs
||_dSr�r�r�r r r!r�=scCs|jrtjStjSr�)r�rk�
CERT_REQUIRED�	CERT_NONEr�r r r!�verify_modeBsz"SecureTransportContext.verify_modecCs|tjkrdnd|_dS)NTF)rkr�r�r�r r r!r�FscCsdSr�r r�r r r!�set_default_verify_pathsJs
z/SecureTransportContext.set_default_verify_pathscCs|��Sr�)r�r�r r r!�load_default_certsVsz)SecureTransportContext.load_default_certscCs|tjjkrtd��dS)Nz5SecureTransport doesn't support custom cipher strings)rr�DEFAULT_CIPHERSr�)rTr[r r r!�set_ciphersYsz"SecureTransportContext.set_ciphersNcCsN|durtd��|dur@t|��Wd�n1s60Y|pF||_dS)Nz1SecureTransport does not support cert directories)r�rwr�)rT�cafile�capath�cadatar r r!�load_verify_locations^s
z,SecureTransportContext.load_verify_locationscCs||_||_||_dSr�)r�r�Z_client_cert_passphrase)rT�certfile�keyfile�passwordr r r!�load_cert_chainjsz&SecureTransportContext.load_cert_chaincCs&ttd�std��dd�|D�|_dS)z�
        Sets the ALPN protocols that will later be set on the context.

        Raises a NotImplementedError if ALPN is not supported.
        r^z2SecureTransport supports ALPN only in macOS 10.12+cSsg|]}t�|��qSr )r�
ensure_binary)�.0�pr r r!�
<listcomp>y�z=SecureTransportContext.set_alpn_protocols.<locals>.<listcomp>N)�hasattrr	�NotImplementedErrorr�)rTr`r r r!�set_alpn_protocolsos

�z)SecureTransportContext.set_alpn_protocolsFTcCsN|rJ�|sJ�|sJ�t|�}|�||j|j|j|j|j|j|j|j	�	|Sr�)
rJr�r�r�r�r�r�r�r�r�)rT�sock�server_side�do_handshake_on_connect�suppress_ragged_eofsr�r<r r r!�wrap_socket{s �z"SecureTransportContext.wrap_socket)NNN)NN)FTTN)r�r�r�r�rU�propertyr��setterr�r�r�r�r�r�r�r�r�r r r r!rs2



	




�r)r�)r�N)fr��
__future__rr�r.r,�os.pathrtr�r(rkrf�	threading�weakref�pip._vendorr�rZ	util.ssl_rZ_securetransport.bindingsrr	r
Z_securetransport.low_levelrrr
rrrr�ImportErrorZpackages.backports.makefiler�__all__rr#rrr"�WeakValueDictionaryr%�Lockr�r�Z'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_AES_256_GCM_SHA384ZTLS_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHArZ�PROTOCOL_TLSr�r�r�r�r�rr�rrr�rrrrrCrIZSSLReadFuncrZSSLWriteFuncr��objectrJr�rr r r r!�<module>s�5 	�%��
�
�
�
�
76

<
site-packages/pip/_vendor/urllib3/contrib/__pycache__/__init__.cpython-39.pyc000064400000000356150564745010023137 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/__init__.py�<module>�site-packages/pip/_vendor/urllib3/contrib/__pycache__/appengine.cpython-39.pyc000064400000020167150564745010023350 0ustar00a

��Re+�@sdZddlmZddlZddlZddlZddlmZmZm	Z	m
Z
mZmZddl
mZddlmZddlmZdd	lmZdd
lmZddlmZzdd
lmZWney�dZYn0e�e�ZGdd�de�ZGdd�de�Z Gdd�de�Z!ej"Z"ej#Z#ej$Z$ej%Z%ej&Z&dS)a[
This module provides a pool manager that uses Google App Engine's
`URLFetch Service <https://cloud.google.com/appengine/docs/python/urlfetch>`_.

Example usage::

    from pip._vendor.urllib3 import PoolManager
    from pip._vendor.urllib3.contrib.appengine import AppEngineManager, is_appengine_sandbox

    if is_appengine_sandbox():
        # AppEngineManager uses AppEngine's URLFetch API behind the scenes
        http = AppEngineManager()
    else:
        # PoolManager uses a socket-level API behind the scenes
        http = PoolManager()

    r = http.request('GET', 'https://google.com/')

There are `limitations <https://cloud.google.com/appengine/docs/python/urlfetch/#Python_Quotas_and_limits>`_ to the URLFetch service and it may not be
the best choice for your application. There are three options for using
urllib3 on Google App Engine:

1. You can use :class:`AppEngineManager` with URLFetch. URLFetch is
   cost-effective in many circumstances as long as your usage is within the
   limitations.
2. You can use a normal :class:`~urllib3.PoolManager` by enabling sockets.
   Sockets also have `limitations and restrictions
   <https://cloud.google.com/appengine/docs/python/sockets/   #limitations-and-restrictions>`_ and have a lower free quota than URLFetch.
   To use sockets, be sure to specify the following in your ``app.yaml``::

        env_variables:
            GAE_USE_SOCKETS_HTTPLIB : 'true'

3. If you are using `App Engine Flexible
<https://cloud.google.com/appengine/docs/flexible/>`_, you can use the standard
:class:`PoolManager` without any configuration or special environment variables.
�)�absolute_importN�)�	HTTPError�HTTPWarning�
MaxRetryError�
ProtocolError�SSLError�TimeoutError)�urljoin)�RequestMethods)�HTTPResponse)�Retry)�Timeout�)�_appengine_environ)�urlfetchc@seZdZdS)�AppEnginePlatformWarningN��__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/appengine.pyrGsrc@seZdZdS)�AppEnginePlatformErrorNrrrrrrKsrc@sXeZdZdZddd�Zdd�Zdd	�Zddddejfd
d�Z	dd
�Z
dd�Zdd�ZdS)�AppEngineManagera
    Connection manager for Google App Engine sandbox applications.

    This manager uses the URLFetch service directly instead of using the
    emulated httplib, and is subject to URLFetch limitations as described in
    the App Engine documentation `here
    <https://cloud.google.com/appengine/docs/python/urlfetch>`_.

    Notably it will raise an :class:`AppEnginePlatformError` if:
        * URLFetch is not available.
        * If you attempt to use this on App Engine Flexible, as full socket
          support is available.
        * If a request size is more than 10 megabytes.
        * If a response size is more than 32 megabytes.
        * If you use an unsupported request method such as OPTIONS.

    Beyond those cases, it will raise normal urllib3 errors.
    NTcCs@tstd��t�dt�t�||�||_||_|p8t	j
|_dS)Nz.URLFetch is not available in this environment.z�urllib3 is using URLFetch on Google App Engine sandbox instead of sockets. To use sockets directly instead of URLFetch see https://urllib3.readthedocs.io/en/1.26.x/reference/urllib3.contrib.html.)rr�warnings�warnrr�__init__�validate_certificate�urlfetch_retriesr
�DEFAULT�retries)�self�headersr!rrrrrrcs��zAppEngineManager.__init__cCs|S)Nr)r"rrr�	__enter__|szAppEngineManager.__enter__cCsdS)NFr)r"�exc_type�exc_val�exc_tbrrr�__exit__szAppEngineManager.__exit__c
Ks�|�||�}zF|o |jdko |j}	tj||||p2id|jo<|	|�|�|jd�}
W�nNtjy�}zt	||��WYd}~�n&d}~0tj
y�}z*dt|�vr�td|��t
|��WYd}~n�d}~0tj�y}z.dt|�vr�t|||d��t
|��WYd}~n�d}~0tj�y>}ztd|��WYd}~nld}~0tj�yl}zt|��WYd}~n>d}~0tj�y�}ztd	||��WYd}~n
d}~00|j|
fd
|i|��}|�o�|��}
|
�r�|j�r�|j�r�t||d��n�|jdk�r�d
}z|j||||d�}Wn,t�y<|j�r4t||d��|YS0|�|�t�d||
�t||
�}|j||||f|||d�|��St|�d��}|� ||j|��r�|j||||d�}t�d|�|�!|�|j||f|||||d�|��S|S)NrF)�payload�methodr#Zallow_truncated�follow_redirects�deadlinerz	too largezOURLFetch request too large, URLFetch only supports requests up to 10mb in size.zToo many redirects)�reasonzPURLFetch response too large, URLFetch only supportsresponses up to 32mb in size.z$URLFetch does not support method: %sr!ztoo many redirectsi/�GET)�response�_poolzRedirecting %s -> %s)r!�redirect�timeoutzRetry-Afterz	Retry: %s)�bodyr#r!r1r2)"�_get_retriesr1�totalr�fetchr�_get_absolute_timeoutrZDeadlineExceededErrorr	ZInvalidURLError�strrrZ
DownloadErrorrZResponseTooLargeErrorZSSLCertificateErrorrZInvalidMethodError�#_urlfetch_response_to_http_response�get_redirect_location�raise_on_redirect�status�	increment�sleep_for_retry�log�debugr
�urlopen�bool�	getheader�is_retry�sleep)r"r*�urlr3r#r!r1r2�response_kwr+r/�e�
http_response�redirect_locationZredirect_url�has_retry_afterrrrrA�s��
 �������



���
���zAppEngineManager.urlopencKs�t�r"|j�d�}|dkr"|jd=|j�d�}|dkrZ|�d�}|�d�d�|�|jd<tft�|j	�|j
|j|jd�|��}tft�|j	�|j|j|d�|��S)Nzcontent-encoding�deflateztransfer-encoding�chunked�,)r3�msgr#r<)r3r#r<�original_response)�is_prod_appenginer#�get�split�remove�joinr�io�BytesIO�contentZ
header_msg�status_code)r"Z
urlfetch_resprG�content_encodingZtransfer_encoding�	encodingsrPrrrr9�s2


��

��z4AppEngineManager._urlfetch_response_to_http_responsecCsB|tjurdSt|t�r>|jdus,|jdur8t�dt�|jS|S)NzdURLFetch does not support granular timeout settings, reverting to total or default URLFetch timeout.)	r�DEFAULT_TIMEOUT�
isinstance�_read�_connectrrrr5)r"r2rrrr7s

�z&AppEngineManager._get_absolute_timeoutcCs>t|t�stj|||jd�}|js.|js.|jr:t�dt	�|S)N)r1�defaultzhURLFetch only supports total retries and does not recognize connect, read, or redirect retry parameters.)
r]r
�from_intr!�connect�readr1rrr)r"r!r1rrrr4&s
�zAppEngineManager._get_retries)NNTT)
rrr�__doc__rr$r(rr\rAr9r7r4rrrrrOs"�
�
r$
r)'rd�
__future__rrV�loggingr�
exceptionsrrrrrr	Zpackages.six.moves.urllib.parser
�requestrr/rZ
util.retryr
Zutil.timeoutr�rZgoogle.appengine.apir�ImportError�	getLoggerrr?rrr�is_appengine�is_appengine_sandbox�is_local_appenginerQ�is_prod_appengine_mvmsrrrr�<module>s2( 

hsite-packages/pip/_vendor/urllib3/contrib/securetransport.py000064400000103221150564745010020347 0ustar00"""
SecureTranport support for urllib3 via ctypes.

This makes platform-native TLS available to urllib3 users on macOS without the
use of a compiler. This is an important feature because the Python Package
Index is moving to become a TLSv1.2-or-higher server, and the default OpenSSL
that ships with macOS is not capable of doing TLSv1.2. The only way to resolve
this is to give macOS users an alternative solution to the problem, and that
solution is to use SecureTransport.

We use ctypes here because this solution must not require a compiler. That's
because pip is not allowed to require a compiler either.

This is not intended to be a seriously long-term solution to this problem.
The hope is that PEP 543 will eventually solve this issue for us, at which
point we can retire this contrib module. But in the short term, we need to
solve the impending tire fire that is Python on Mac without this kind of
contrib module. So...here we are.

To use this module, simply import and inject it::

    import pip._vendor.urllib3.contrib.securetransport as securetransport
    securetransport.inject_into_urllib3()

Happy TLSing!

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

.. code-block::

    Copyright (c) 2015-2016 Will Bond <will@wbond.net>

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
"""
from __future__ import absolute_import

import contextlib
import ctypes
import errno
import os.path
import shutil
import socket
import ssl
import struct
import threading
import weakref

from pip._vendor import six

from .. import util
from ..util.ssl_ import PROTOCOL_TLS_CLIENT
from ._securetransport.bindings import CoreFoundation, Security, SecurityConst
from ._securetransport.low_level import (
    _assert_no_error,
    _build_tls_unknown_ca_alert,
    _cert_array_from_pem,
    _create_cfstring_array,
    _load_client_cert_chain,
    _temporary_keychain,
)

try:  # Platform-specific: Python 2
    from socket import _fileobject
except ImportError:  # Platform-specific: Python 3
    _fileobject = None
    from ..packages.backports.makefile import backport_makefile

__all__ = ["inject_into_urllib3", "extract_from_urllib3"]

# SNI always works
HAS_SNI = True

orig_util_HAS_SNI = util.HAS_SNI
orig_util_SSLContext = util.ssl_.SSLContext

# This dictionary is used by the read callback to obtain a handle to the
# calling wrapped socket. This is a pretty silly approach, but for now it'll
# do. I feel like I should be able to smuggle a handle to the wrapped socket
# directly in the SSLConnectionRef, but for now this approach will work I
# guess.
#
# We need to lock around this structure for inserts, but we don't do it for
# reads/writes in the callbacks. The reasoning here goes as follows:
#
#    1. It is not possible to call into the callbacks before the dictionary is
#       populated, so once in the callback the id must be in the dictionary.
#    2. The callbacks don't mutate the dictionary, they only read from it, and
#       so cannot conflict with any of the insertions.
#
# This is good: if we had to lock in the callbacks we'd drastically slow down
# the performance of this code.
_connection_refs = weakref.WeakValueDictionary()
_connection_ref_lock = threading.Lock()

# Limit writes to 16kB. This is OpenSSL's limit, but we'll cargo-cult it over
# for no better reason than we need *a* limit, and this one is right there.
SSL_WRITE_BLOCKSIZE = 16384

# This is our equivalent of util.ssl_.DEFAULT_CIPHERS, but expanded out to
# individual cipher suites. We need to do this because this is how
# SecureTransport wants them.
CIPHER_SUITES = [
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_CBC_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_DHE_RSA_WITH_AES_128_CBC_SHA,
    SecurityConst.TLS_AES_256_GCM_SHA384,
    SecurityConst.TLS_AES_128_GCM_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_GCM_SHA384,
    SecurityConst.TLS_RSA_WITH_AES_128_GCM_SHA256,
    SecurityConst.TLS_AES_128_CCM_8_SHA256,
    SecurityConst.TLS_AES_128_CCM_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_CBC_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_128_CBC_SHA256,
    SecurityConst.TLS_RSA_WITH_AES_256_CBC_SHA,
    SecurityConst.TLS_RSA_WITH_AES_128_CBC_SHA,
]

# Basically this is simple: for PROTOCOL_SSLv23 we turn it into a low of
# TLSv1 and a high of TLSv1.2. For everything else, we pin to that version.
# TLSv1 to 1.2 are supported on macOS 10.8+
_protocol_to_min_max = {
    util.PROTOCOL_TLS: (SecurityConst.kTLSProtocol1, SecurityConst.kTLSProtocol12),
    PROTOCOL_TLS_CLIENT: (SecurityConst.kTLSProtocol1, SecurityConst.kTLSProtocol12),
}

if hasattr(ssl, "PROTOCOL_SSLv2"):
    _protocol_to_min_max[ssl.PROTOCOL_SSLv2] = (
        SecurityConst.kSSLProtocol2,
        SecurityConst.kSSLProtocol2,
    )
if hasattr(ssl, "PROTOCOL_SSLv3"):
    _protocol_to_min_max[ssl.PROTOCOL_SSLv3] = (
        SecurityConst.kSSLProtocol3,
        SecurityConst.kSSLProtocol3,
    )
if hasattr(ssl, "PROTOCOL_TLSv1"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1] = (
        SecurityConst.kTLSProtocol1,
        SecurityConst.kTLSProtocol1,
    )
if hasattr(ssl, "PROTOCOL_TLSv1_1"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1_1] = (
        SecurityConst.kTLSProtocol11,
        SecurityConst.kTLSProtocol11,
    )
if hasattr(ssl, "PROTOCOL_TLSv1_2"):
    _protocol_to_min_max[ssl.PROTOCOL_TLSv1_2] = (
        SecurityConst.kTLSProtocol12,
        SecurityConst.kTLSProtocol12,
    )


def inject_into_urllib3():
    """
    Monkey-patch urllib3 with SecureTransport-backed SSL-support.
    """
    util.SSLContext = SecureTransportContext
    util.ssl_.SSLContext = SecureTransportContext
    util.HAS_SNI = HAS_SNI
    util.ssl_.HAS_SNI = HAS_SNI
    util.IS_SECURETRANSPORT = True
    util.ssl_.IS_SECURETRANSPORT = True


def extract_from_urllib3():
    """
    Undo monkey-patching by :func:`inject_into_urllib3`.
    """
    util.SSLContext = orig_util_SSLContext
    util.ssl_.SSLContext = orig_util_SSLContext
    util.HAS_SNI = orig_util_HAS_SNI
    util.ssl_.HAS_SNI = orig_util_HAS_SNI
    util.IS_SECURETRANSPORT = False
    util.ssl_.IS_SECURETRANSPORT = False


def _read_callback(connection_id, data_buffer, data_length_pointer):
    """
    SecureTransport read callback. This is called by ST to request that data
    be returned from the socket.
    """
    wrapped_socket = None
    try:
        wrapped_socket = _connection_refs.get(connection_id)
        if wrapped_socket is None:
            return SecurityConst.errSSLInternal
        base_socket = wrapped_socket.socket

        requested_length = data_length_pointer[0]

        timeout = wrapped_socket.gettimeout()
        error = None
        read_count = 0

        try:
            while read_count < requested_length:
                if timeout is None or timeout >= 0:
                    if not util.wait_for_read(base_socket, timeout):
                        raise socket.error(errno.EAGAIN, "timed out")

                remaining = requested_length - read_count
                buffer = (ctypes.c_char * remaining).from_address(
                    data_buffer + read_count
                )
                chunk_size = base_socket.recv_into(buffer, remaining)
                read_count += chunk_size
                if not chunk_size:
                    if not read_count:
                        return SecurityConst.errSSLClosedGraceful
                    break
        except (socket.error) as e:
            error = e.errno

            if error is not None and error != errno.EAGAIN:
                data_length_pointer[0] = read_count
                if error == errno.ECONNRESET or error == errno.EPIPE:
                    return SecurityConst.errSSLClosedAbort
                raise

        data_length_pointer[0] = read_count

        if read_count != requested_length:
            return SecurityConst.errSSLWouldBlock

        return 0
    except Exception as e:
        if wrapped_socket is not None:
            wrapped_socket._exception = e
        return SecurityConst.errSSLInternal


def _write_callback(connection_id, data_buffer, data_length_pointer):
    """
    SecureTransport write callback. This is called by ST to request that data
    actually be sent on the network.
    """
    wrapped_socket = None
    try:
        wrapped_socket = _connection_refs.get(connection_id)
        if wrapped_socket is None:
            return SecurityConst.errSSLInternal
        base_socket = wrapped_socket.socket

        bytes_to_write = data_length_pointer[0]
        data = ctypes.string_at(data_buffer, bytes_to_write)

        timeout = wrapped_socket.gettimeout()
        error = None
        sent = 0

        try:
            while sent < bytes_to_write:
                if timeout is None or timeout >= 0:
                    if not util.wait_for_write(base_socket, timeout):
                        raise socket.error(errno.EAGAIN, "timed out")
                chunk_sent = base_socket.send(data)
                sent += chunk_sent

                # This has some needless copying here, but I'm not sure there's
                # much value in optimising this data path.
                data = data[chunk_sent:]
        except (socket.error) as e:
            error = e.errno

            if error is not None and error != errno.EAGAIN:
                data_length_pointer[0] = sent
                if error == errno.ECONNRESET or error == errno.EPIPE:
                    return SecurityConst.errSSLClosedAbort
                raise

        data_length_pointer[0] = sent

        if sent != bytes_to_write:
            return SecurityConst.errSSLWouldBlock

        return 0
    except Exception as e:
        if wrapped_socket is not None:
            wrapped_socket._exception = e
        return SecurityConst.errSSLInternal


# We need to keep these two objects references alive: if they get GC'd while
# in use then SecureTransport could attempt to call a function that is in freed
# memory. That would be...uh...bad. Yeah, that's the word. Bad.
_read_callback_pointer = Security.SSLReadFunc(_read_callback)
_write_callback_pointer = Security.SSLWriteFunc(_write_callback)


class WrappedSocket(object):
    """
    API-compatibility wrapper for Python's OpenSSL wrapped socket object.

    Note: _makefile_refs, _drop(), and _reuse() are needed for the garbage
    collector of PyPy.
    """

    def __init__(self, socket):
        self.socket = socket
        self.context = None
        self._makefile_refs = 0
        self._closed = False
        self._exception = None
        self._keychain = None
        self._keychain_dir = None
        self._client_cert_chain = None

        # We save off the previously-configured timeout and then set it to
        # zero. This is done because we use select and friends to handle the
        # timeouts, but if we leave the timeout set on the lower socket then
        # Python will "kindly" call select on that socket again for us. Avoid
        # that by forcing the timeout to zero.
        self._timeout = self.socket.gettimeout()
        self.socket.settimeout(0)

    @contextlib.contextmanager
    def _raise_on_error(self):
        """
        A context manager that can be used to wrap calls that do I/O from
        SecureTransport. If any of the I/O callbacks hit an exception, this
        context manager will correctly propagate the exception after the fact.
        This avoids silently swallowing those exceptions.

        It also correctly forces the socket closed.
        """
        self._exception = None

        # We explicitly don't catch around this yield because in the unlikely
        # event that an exception was hit in the block we don't want to swallow
        # it.
        yield
        if self._exception is not None:
            exception, self._exception = self._exception, None
            self.close()
            raise exception

    def _set_ciphers(self):
        """
        Sets up the allowed ciphers. By default this matches the set in
        util.ssl_.DEFAULT_CIPHERS, at least as supported by macOS. This is done
        custom and doesn't allow changing at this time, mostly because parsing
        OpenSSL cipher strings is going to be a freaking nightmare.
        """
        ciphers = (Security.SSLCipherSuite * len(CIPHER_SUITES))(*CIPHER_SUITES)
        result = Security.SSLSetEnabledCiphers(
            self.context, ciphers, len(CIPHER_SUITES)
        )
        _assert_no_error(result)

    def _set_alpn_protocols(self, protocols):
        """
        Sets up the ALPN protocols on the context.
        """
        if not protocols:
            return
        protocols_arr = _create_cfstring_array(protocols)
        try:
            result = Security.SSLSetALPNProtocols(self.context, protocols_arr)
            _assert_no_error(result)
        finally:
            CoreFoundation.CFRelease(protocols_arr)

    def _custom_validate(self, verify, trust_bundle):
        """
        Called when we have set custom validation. We do this in two cases:
        first, when cert validation is entirely disabled; and second, when
        using a custom trust DB.
        Raises an SSLError if the connection is not trusted.
        """
        # If we disabled cert validation, just say: cool.
        if not verify:
            return

        successes = (
            SecurityConst.kSecTrustResultUnspecified,
            SecurityConst.kSecTrustResultProceed,
        )
        try:
            trust_result = self._evaluate_trust(trust_bundle)
            if trust_result in successes:
                return
            reason = "error code: %d" % (trust_result,)
        except Exception as e:
            # Do not trust on error
            reason = "exception: %r" % (e,)

        # SecureTransport does not send an alert nor shuts down the connection.
        rec = _build_tls_unknown_ca_alert(self.version())
        self.socket.sendall(rec)
        # close the connection immediately
        # l_onoff = 1, activate linger
        # l_linger = 0, linger for 0 seoncds
        opts = struct.pack("ii", 1, 0)
        self.socket.setsockopt(socket.SOL_SOCKET, socket.SO_LINGER, opts)
        self.close()
        raise ssl.SSLError("certificate verify failed, %s" % reason)

    def _evaluate_trust(self, trust_bundle):
        # We want data in memory, so load it up.
        if os.path.isfile(trust_bundle):
            with open(trust_bundle, "rb") as f:
                trust_bundle = f.read()

        cert_array = None
        trust = Security.SecTrustRef()

        try:
            # Get a CFArray that contains the certs we want.
            cert_array = _cert_array_from_pem(trust_bundle)

            # Ok, now the hard part. We want to get the SecTrustRef that ST has
            # created for this connection, shove our CAs into it, tell ST to
            # ignore everything else it knows, and then ask if it can build a
            # chain. This is a buuuunch of code.
            result = Security.SSLCopyPeerTrust(self.context, ctypes.byref(trust))
            _assert_no_error(result)
            if not trust:
                raise ssl.SSLError("Failed to copy trust reference")

            result = Security.SecTrustSetAnchorCertificates(trust, cert_array)
            _assert_no_error(result)

            result = Security.SecTrustSetAnchorCertificatesOnly(trust, True)
            _assert_no_error(result)

            trust_result = Security.SecTrustResultType()
            result = Security.SecTrustEvaluate(trust, ctypes.byref(trust_result))
            _assert_no_error(result)
        finally:
            if trust:
                CoreFoundation.CFRelease(trust)

            if cert_array is not None:
                CoreFoundation.CFRelease(cert_array)

        return trust_result.value

    def handshake(
        self,
        server_hostname,
        verify,
        trust_bundle,
        min_version,
        max_version,
        client_cert,
        client_key,
        client_key_passphrase,
        alpn_protocols,
    ):
        """
        Actually performs the TLS handshake. This is run automatically by
        wrapped socket, and shouldn't be needed in user code.
        """
        # First, we do the initial bits of connection setup. We need to create
        # a context, set its I/O funcs, and set the connection reference.
        self.context = Security.SSLCreateContext(
            None, SecurityConst.kSSLClientSide, SecurityConst.kSSLStreamType
        )
        result = Security.SSLSetIOFuncs(
            self.context, _read_callback_pointer, _write_callback_pointer
        )
        _assert_no_error(result)

        # Here we need to compute the handle to use. We do this by taking the
        # id of self modulo 2**31 - 1. If this is already in the dictionary, we
        # just keep incrementing by one until we find a free space.
        with _connection_ref_lock:
            handle = id(self) % 2147483647
            while handle in _connection_refs:
                handle = (handle + 1) % 2147483647
            _connection_refs[handle] = self

        result = Security.SSLSetConnection(self.context, handle)
        _assert_no_error(result)

        # If we have a server hostname, we should set that too.
        if server_hostname:
            if not isinstance(server_hostname, bytes):
                server_hostname = server_hostname.encode("utf-8")

            result = Security.SSLSetPeerDomainName(
                self.context, server_hostname, len(server_hostname)
            )
            _assert_no_error(result)

        # Setup the ciphers.
        self._set_ciphers()

        # Setup the ALPN protocols.
        self._set_alpn_protocols(alpn_protocols)

        # Set the minimum and maximum TLS versions.
        result = Security.SSLSetProtocolVersionMin(self.context, min_version)
        _assert_no_error(result)

        result = Security.SSLSetProtocolVersionMax(self.context, max_version)
        _assert_no_error(result)

        # If there's a trust DB, we need to use it. We do that by telling
        # SecureTransport to break on server auth. We also do that if we don't
        # want to validate the certs at all: we just won't actually do any
        # authing in that case.
        if not verify or trust_bundle is not None:
            result = Security.SSLSetSessionOption(
                self.context, SecurityConst.kSSLSessionOptionBreakOnServerAuth, True
            )
            _assert_no_error(result)

        # If there's a client cert, we need to use it.
        if client_cert:
            self._keychain, self._keychain_dir = _temporary_keychain()
            self._client_cert_chain = _load_client_cert_chain(
                self._keychain, client_cert, client_key
            )
            result = Security.SSLSetCertificate(self.context, self._client_cert_chain)
            _assert_no_error(result)

        while True:
            with self._raise_on_error():
                result = Security.SSLHandshake(self.context)

                if result == SecurityConst.errSSLWouldBlock:
                    raise socket.timeout("handshake timed out")
                elif result == SecurityConst.errSSLServerAuthCompleted:
                    self._custom_validate(verify, trust_bundle)
                    continue
                else:
                    _assert_no_error(result)
                    break

    def fileno(self):
        return self.socket.fileno()

    # Copy-pasted from Python 3.5 source code
    def _decref_socketios(self):
        if self._makefile_refs > 0:
            self._makefile_refs -= 1
        if self._closed:
            self.close()

    def recv(self, bufsiz):
        buffer = ctypes.create_string_buffer(bufsiz)
        bytes_read = self.recv_into(buffer, bufsiz)
        data = buffer[:bytes_read]
        return data

    def recv_into(self, buffer, nbytes=None):
        # Read short on EOF.
        if self._closed:
            return 0

        if nbytes is None:
            nbytes = len(buffer)

        buffer = (ctypes.c_char * nbytes).from_buffer(buffer)
        processed_bytes = ctypes.c_size_t(0)

        with self._raise_on_error():
            result = Security.SSLRead(
                self.context, buffer, nbytes, ctypes.byref(processed_bytes)
            )

        # There are some result codes that we want to treat as "not always
        # errors". Specifically, those are errSSLWouldBlock,
        # errSSLClosedGraceful, and errSSLClosedNoNotify.
        if result == SecurityConst.errSSLWouldBlock:
            # If we didn't process any bytes, then this was just a time out.
            # However, we can get errSSLWouldBlock in situations when we *did*
            # read some data, and in those cases we should just read "short"
            # and return.
            if processed_bytes.value == 0:
                # Timed out, no data read.
                raise socket.timeout("recv timed out")
        elif result in (
            SecurityConst.errSSLClosedGraceful,
            SecurityConst.errSSLClosedNoNotify,
        ):
            # The remote peer has closed this connection. We should do so as
            # well. Note that we don't actually return here because in
            # principle this could actually be fired along with return data.
            # It's unlikely though.
            self.close()
        else:
            _assert_no_error(result)

        # Ok, we read and probably succeeded. We should return whatever data
        # was actually read.
        return processed_bytes.value

    def settimeout(self, timeout):
        self._timeout = timeout

    def gettimeout(self):
        return self._timeout

    def send(self, data):
        processed_bytes = ctypes.c_size_t(0)

        with self._raise_on_error():
            result = Security.SSLWrite(
                self.context, data, len(data), ctypes.byref(processed_bytes)
            )

        if result == SecurityConst.errSSLWouldBlock and processed_bytes.value == 0:
            # Timed out
            raise socket.timeout("send timed out")
        else:
            _assert_no_error(result)

        # We sent, and probably succeeded. Tell them how much we sent.
        return processed_bytes.value

    def sendall(self, data):
        total_sent = 0
        while total_sent < len(data):
            sent = self.send(data[total_sent : total_sent + SSL_WRITE_BLOCKSIZE])
            total_sent += sent

    def shutdown(self):
        with self._raise_on_error():
            Security.SSLClose(self.context)

    def close(self):
        # TODO: should I do clean shutdown here? Do I have to?
        if self._makefile_refs < 1:
            self._closed = True
            if self.context:
                CoreFoundation.CFRelease(self.context)
                self.context = None
            if self._client_cert_chain:
                CoreFoundation.CFRelease(self._client_cert_chain)
                self._client_cert_chain = None
            if self._keychain:
                Security.SecKeychainDelete(self._keychain)
                CoreFoundation.CFRelease(self._keychain)
                shutil.rmtree(self._keychain_dir)
                self._keychain = self._keychain_dir = None
            return self.socket.close()
        else:
            self._makefile_refs -= 1

    def getpeercert(self, binary_form=False):
        # Urgh, annoying.
        #
        # Here's how we do this:
        #
        # 1. Call SSLCopyPeerTrust to get hold of the trust object for this
        #    connection.
        # 2. Call SecTrustGetCertificateAtIndex for index 0 to get the leaf.
        # 3. To get the CN, call SecCertificateCopyCommonName and process that
        #    string so that it's of the appropriate type.
        # 4. To get the SAN, we need to do something a bit more complex:
        #    a. Call SecCertificateCopyValues to get the data, requesting
        #       kSecOIDSubjectAltName.
        #    b. Mess about with this dictionary to try to get the SANs out.
        #
        # This is gross. Really gross. It's going to be a few hundred LoC extra
        # just to repeat something that SecureTransport can *already do*. So my
        # operating assumption at this time is that what we want to do is
        # instead to just flag to urllib3 that it shouldn't do its own hostname
        # validation when using SecureTransport.
        if not binary_form:
            raise ValueError("SecureTransport only supports dumping binary certs")
        trust = Security.SecTrustRef()
        certdata = None
        der_bytes = None

        try:
            # Grab the trust store.
            result = Security.SSLCopyPeerTrust(self.context, ctypes.byref(trust))
            _assert_no_error(result)
            if not trust:
                # Probably we haven't done the handshake yet. No biggie.
                return None

            cert_count = Security.SecTrustGetCertificateCount(trust)
            if not cert_count:
                # Also a case that might happen if we haven't handshaked.
                # Handshook? Handshaken?
                return None

            leaf = Security.SecTrustGetCertificateAtIndex(trust, 0)
            assert leaf

            # Ok, now we want the DER bytes.
            certdata = Security.SecCertificateCopyData(leaf)
            assert certdata

            data_length = CoreFoundation.CFDataGetLength(certdata)
            data_buffer = CoreFoundation.CFDataGetBytePtr(certdata)
            der_bytes = ctypes.string_at(data_buffer, data_length)
        finally:
            if certdata:
                CoreFoundation.CFRelease(certdata)
            if trust:
                CoreFoundation.CFRelease(trust)

        return der_bytes

    def version(self):
        protocol = Security.SSLProtocol()
        result = Security.SSLGetNegotiatedProtocolVersion(
            self.context, ctypes.byref(protocol)
        )
        _assert_no_error(result)
        if protocol.value == SecurityConst.kTLSProtocol13:
            raise ssl.SSLError("SecureTransport does not support TLS 1.3")
        elif protocol.value == SecurityConst.kTLSProtocol12:
            return "TLSv1.2"
        elif protocol.value == SecurityConst.kTLSProtocol11:
            return "TLSv1.1"
        elif protocol.value == SecurityConst.kTLSProtocol1:
            return "TLSv1"
        elif protocol.value == SecurityConst.kSSLProtocol3:
            return "SSLv3"
        elif protocol.value == SecurityConst.kSSLProtocol2:
            return "SSLv2"
        else:
            raise ssl.SSLError("Unknown TLS version: %r" % protocol)

    def _reuse(self):
        self._makefile_refs += 1

    def _drop(self):
        if self._makefile_refs < 1:
            self.close()
        else:
            self._makefile_refs -= 1


if _fileobject:  # Platform-specific: Python 2

    def makefile(self, mode, bufsize=-1):
        self._makefile_refs += 1
        return _fileobject(self, mode, bufsize, close=True)


else:  # Platform-specific: Python 3

    def makefile(self, mode="r", buffering=None, *args, **kwargs):
        # We disable buffering with SecureTransport because it conflicts with
        # the buffering that ST does internally (see issue #1153 for more).
        buffering = 0
        return backport_makefile(self, mode, buffering, *args, **kwargs)


WrappedSocket.makefile = makefile


class SecureTransportContext(object):
    """
    I am a wrapper class for the SecureTransport library, to translate the
    interface of the standard library ``SSLContext`` object to calls into
    SecureTransport.
    """

    def __init__(self, protocol):
        self._min_version, self._max_version = _protocol_to_min_max[protocol]
        self._options = 0
        self._verify = False
        self._trust_bundle = None
        self._client_cert = None
        self._client_key = None
        self._client_key_passphrase = None
        self._alpn_protocols = None

    @property
    def check_hostname(self):
        """
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        """
        return True

    @check_hostname.setter
    def check_hostname(self, value):
        """
        SecureTransport cannot have its hostname checking disabled. For more,
        see the comment on getpeercert() in this file.
        """
        pass

    @property
    def options(self):
        # TODO: Well, crap.
        #
        # So this is the bit of the code that is the most likely to cause us
        # trouble. Essentially we need to enumerate all of the SSL options that
        # users might want to use and try to see if we can sensibly translate
        # them, or whether we should just ignore them.
        return self._options

    @options.setter
    def options(self, value):
        # TODO: Update in line with above.
        self._options = value

    @property
    def verify_mode(self):
        return ssl.CERT_REQUIRED if self._verify else ssl.CERT_NONE

    @verify_mode.setter
    def verify_mode(self, value):
        self._verify = True if value == ssl.CERT_REQUIRED else False

    def set_default_verify_paths(self):
        # So, this has to do something a bit weird. Specifically, what it does
        # is nothing.
        #
        # This means that, if we had previously had load_verify_locations
        # called, this does not undo that. We need to do that because it turns
        # out that the rest of the urllib3 code will attempt to load the
        # default verify paths if it hasn't been told about any paths, even if
        # the context itself was sometime earlier. We resolve that by just
        # ignoring it.
        pass

    def load_default_certs(self):
        return self.set_default_verify_paths()

    def set_ciphers(self, ciphers):
        # For now, we just require the default cipher string.
        if ciphers != util.ssl_.DEFAULT_CIPHERS:
            raise ValueError("SecureTransport doesn't support custom cipher strings")

    def load_verify_locations(self, cafile=None, capath=None, cadata=None):
        # OK, we only really support cadata and cafile.
        if capath is not None:
            raise ValueError("SecureTransport does not support cert directories")

        # Raise if cafile does not exist.
        if cafile is not None:
            with open(cafile):
                pass

        self._trust_bundle = cafile or cadata

    def load_cert_chain(self, certfile, keyfile=None, password=None):
        self._client_cert = certfile
        self._client_key = keyfile
        self._client_cert_passphrase = password

    def set_alpn_protocols(self, protocols):
        """
        Sets the ALPN protocols that will later be set on the context.

        Raises a NotImplementedError if ALPN is not supported.
        """
        if not hasattr(Security, "SSLSetALPNProtocols"):
            raise NotImplementedError(
                "SecureTransport supports ALPN only in macOS 10.12+"
            )
        self._alpn_protocols = [six.ensure_binary(p) for p in protocols]

    def wrap_socket(
        self,
        sock,
        server_side=False,
        do_handshake_on_connect=True,
        suppress_ragged_eofs=True,
        server_hostname=None,
    ):
        # So, what do we do here? Firstly, we assert some properties. This is a
        # stripped down shim, so there is some functionality we don't support.
        # See PEP 543 for the real deal.
        assert not server_side
        assert do_handshake_on_connect
        assert suppress_ragged_eofs

        # Ok, we're good to go. Now we want to create the wrapped socket object
        # and store it in the appropriate place.
        wrapped_socket = WrappedSocket(sock)

        # Now we can handshake
        wrapped_socket.handshake(
            server_hostname,
            self._verify,
            self._trust_bundle,
            self._min_version,
            self._max_version,
            self._client_cert,
            self._client_key,
            self._client_key_passphrase,
            self._alpn_protocols,
        )
        return wrapped_socket
site-packages/pip/_vendor/urllib3/contrib/__init__.py000064400000000000150564745010016632 0ustar00site-packages/pip/_vendor/urllib3/contrib/pyopenssl.py000064400000041004150564745010017140 0ustar00"""
TLS with SNI_-support for Python 2. Follow these instructions if you would
like to verify TLS certificates in Python 2. Note, the default libraries do
*not* do certificate checking; you need to do additional work to validate
certificates yourself.

This needs the following packages installed:

* `pyOpenSSL`_ (tested with 16.0.0)
* `cryptography`_ (minimum 1.3.4, from pyopenssl)
* `idna`_ (minimum 2.0, from cryptography)

However, pyopenssl depends on cryptography, which depends on idna, so while we
use all three directly here we end up having relatively few packages required.

You can install them with the following command:

.. code-block:: bash

    $ python -m pip install pyopenssl cryptography idna

To activate certificate checking, call
:func:`~urllib3.contrib.pyopenssl.inject_into_urllib3` from your Python code
before you begin making HTTP requests. This can be done in a ``sitecustomize``
module, or at any other time before your application begins using ``urllib3``,
like this:

.. code-block:: python

    try:
        import pip._vendor.urllib3.contrib.pyopenssl as pyopenssl
        pyopenssl.inject_into_urllib3()
    except ImportError:
        pass

Now you can use :mod:`urllib3` as you normally would, and it will support SNI
when the required modules are installed.

Activating this module also has the positive side effect of disabling SSL/TLS
compression in Python 2 (see `CRIME attack`_).

.. _sni: https://en.wikipedia.org/wiki/Server_Name_Indication
.. _crime attack: https://en.wikipedia.org/wiki/CRIME_(security_exploit)
.. _pyopenssl: https://www.pyopenssl.org
.. _cryptography: https://cryptography.io
.. _idna: https://github.com/kjd/idna
"""
from __future__ import absolute_import

import OpenSSL.SSL
from cryptography import x509
from cryptography.hazmat.backends.openssl import backend as openssl_backend
from cryptography.hazmat.backends.openssl.x509 import _Certificate

try:
    from cryptography.x509 import UnsupportedExtension
except ImportError:
    # UnsupportedExtension is gone in cryptography >= 2.1.0
    class UnsupportedExtension(Exception):
        pass


from io import BytesIO
from socket import error as SocketError
from socket import timeout

try:  # Platform-specific: Python 2
    from socket import _fileobject
except ImportError:  # Platform-specific: Python 3
    _fileobject = None
    from ..packages.backports.makefile import backport_makefile

import logging
import ssl
import sys

from .. import util
from ..packages import six
from ..util.ssl_ import PROTOCOL_TLS_CLIENT

__all__ = ["inject_into_urllib3", "extract_from_urllib3"]

# SNI always works.
HAS_SNI = True

# Map from urllib3 to PyOpenSSL compatible parameter-values.
_openssl_versions = {
    util.PROTOCOL_TLS: OpenSSL.SSL.SSLv23_METHOD,
    PROTOCOL_TLS_CLIENT: OpenSSL.SSL.SSLv23_METHOD,
    ssl.PROTOCOL_TLSv1: OpenSSL.SSL.TLSv1_METHOD,
}

if hasattr(ssl, "PROTOCOL_SSLv3") and hasattr(OpenSSL.SSL, "SSLv3_METHOD"):
    _openssl_versions[ssl.PROTOCOL_SSLv3] = OpenSSL.SSL.SSLv3_METHOD

if hasattr(ssl, "PROTOCOL_TLSv1_1") and hasattr(OpenSSL.SSL, "TLSv1_1_METHOD"):
    _openssl_versions[ssl.PROTOCOL_TLSv1_1] = OpenSSL.SSL.TLSv1_1_METHOD

if hasattr(ssl, "PROTOCOL_TLSv1_2") and hasattr(OpenSSL.SSL, "TLSv1_2_METHOD"):
    _openssl_versions[ssl.PROTOCOL_TLSv1_2] = OpenSSL.SSL.TLSv1_2_METHOD


_stdlib_to_openssl_verify = {
    ssl.CERT_NONE: OpenSSL.SSL.VERIFY_NONE,
    ssl.CERT_OPTIONAL: OpenSSL.SSL.VERIFY_PEER,
    ssl.CERT_REQUIRED: OpenSSL.SSL.VERIFY_PEER
    + OpenSSL.SSL.VERIFY_FAIL_IF_NO_PEER_CERT,
}
_openssl_to_stdlib_verify = dict((v, k) for k, v in _stdlib_to_openssl_verify.items())

# OpenSSL will only write 16K at a time
SSL_WRITE_BLOCKSIZE = 16384

orig_util_HAS_SNI = util.HAS_SNI
orig_util_SSLContext = util.ssl_.SSLContext


log = logging.getLogger(__name__)


def inject_into_urllib3():
    "Monkey-patch urllib3 with PyOpenSSL-backed SSL-support."

    _validate_dependencies_met()

    util.SSLContext = PyOpenSSLContext
    util.ssl_.SSLContext = PyOpenSSLContext
    util.HAS_SNI = HAS_SNI
    util.ssl_.HAS_SNI = HAS_SNI
    util.IS_PYOPENSSL = True
    util.ssl_.IS_PYOPENSSL = True


def extract_from_urllib3():
    "Undo monkey-patching by :func:`inject_into_urllib3`."

    util.SSLContext = orig_util_SSLContext
    util.ssl_.SSLContext = orig_util_SSLContext
    util.HAS_SNI = orig_util_HAS_SNI
    util.ssl_.HAS_SNI = orig_util_HAS_SNI
    util.IS_PYOPENSSL = False
    util.ssl_.IS_PYOPENSSL = False


def _validate_dependencies_met():
    """
    Verifies that PyOpenSSL's package-level dependencies have been met.
    Throws `ImportError` if they are not met.
    """
    # Method added in `cryptography==1.1`; not available in older versions
    from cryptography.x509.extensions import Extensions

    if getattr(Extensions, "get_extension_for_class", None) is None:
        raise ImportError(
            "'cryptography' module missing required functionality.  "
            "Try upgrading to v1.3.4 or newer."
        )

    # pyOpenSSL 0.14 and above use cryptography for OpenSSL bindings. The _x509
    # attribute is only present on those versions.
    from OpenSSL.crypto import X509

    x509 = X509()
    if getattr(x509, "_x509", None) is None:
        raise ImportError(
            "'pyOpenSSL' module missing required functionality. "
            "Try upgrading to v0.14 or newer."
        )


def _dnsname_to_stdlib(name):
    """
    Converts a dNSName SubjectAlternativeName field to the form used by the
    standard library on the given Python version.

    Cryptography produces a dNSName as a unicode string that was idna-decoded
    from ASCII bytes. We need to idna-encode that string to get it back, and
    then on Python 3 we also need to convert to unicode via UTF-8 (the stdlib
    uses PyUnicode_FromStringAndSize on it, which decodes via UTF-8).

    If the name cannot be idna-encoded then we return None signalling that
    the name given should be skipped.
    """

    def idna_encode(name):
        """
        Borrowed wholesale from the Python Cryptography Project. It turns out
        that we can't just safely call `idna.encode`: it can explode for
        wildcard names. This avoids that problem.
        """
        from pip._vendor import idna

        try:
            for prefix in [u"*.", u"."]:
                if name.startswith(prefix):
                    name = name[len(prefix) :]
                    return prefix.encode("ascii") + idna.encode(name)
            return idna.encode(name)
        except idna.core.IDNAError:
            return None

    # Don't send IPv6 addresses through the IDNA encoder.
    if ":" in name:
        return name

    name = idna_encode(name)
    if name is None:
        return None
    elif sys.version_info >= (3, 0):
        name = name.decode("utf-8")
    return name


def get_subj_alt_name(peer_cert):
    """
    Given an PyOpenSSL certificate, provides all the subject alternative names.
    """
    # Pass the cert to cryptography, which has much better APIs for this.
    if hasattr(peer_cert, "to_cryptography"):
        cert = peer_cert.to_cryptography()
    else:
        # This is technically using private APIs, but should work across all
        # relevant versions before PyOpenSSL got a proper API for this.
        cert = _Certificate(openssl_backend, peer_cert._x509)

    # We want to find the SAN extension. Ask Cryptography to locate it (it's
    # faster than looping in Python)
    try:
        ext = cert.extensions.get_extension_for_class(x509.SubjectAlternativeName).value
    except x509.ExtensionNotFound:
        # No such extension, return the empty list.
        return []
    except (
        x509.DuplicateExtension,
        UnsupportedExtension,
        x509.UnsupportedGeneralNameType,
        UnicodeError,
    ) as e:
        # A problem has been found with the quality of the certificate. Assume
        # no SAN field is present.
        log.warning(
            "A problem was encountered with the certificate that prevented "
            "urllib3 from finding the SubjectAlternativeName field. This can "
            "affect certificate validation. The error was %s",
            e,
        )
        return []

    # We want to return dNSName and iPAddress fields. We need to cast the IPs
    # back to strings because the match_hostname function wants them as
    # strings.
    # Sadly the DNS names need to be idna encoded and then, on Python 3, UTF-8
    # decoded. This is pretty frustrating, but that's what the standard library
    # does with certificates, and so we need to attempt to do the same.
    # We also want to skip over names which cannot be idna encoded.
    names = [
        ("DNS", name)
        for name in map(_dnsname_to_stdlib, ext.get_values_for_type(x509.DNSName))
        if name is not None
    ]
    names.extend(
        ("IP Address", str(name)) for name in ext.get_values_for_type(x509.IPAddress)
    )

    return names


class WrappedSocket(object):
    """API-compatibility wrapper for Python OpenSSL's Connection-class.

    Note: _makefile_refs, _drop() and _reuse() are needed for the garbage
    collector of pypy.
    """

    def __init__(self, connection, socket, suppress_ragged_eofs=True):
        self.connection = connection
        self.socket = socket
        self.suppress_ragged_eofs = suppress_ragged_eofs
        self._makefile_refs = 0
        self._closed = False

    def fileno(self):
        return self.socket.fileno()

    # Copy-pasted from Python 3.5 source code
    def _decref_socketios(self):
        if self._makefile_refs > 0:
            self._makefile_refs -= 1
        if self._closed:
            self.close()

    def recv(self, *args, **kwargs):
        try:
            data = self.connection.recv(*args, **kwargs)
        except OpenSSL.SSL.SysCallError as e:
            if self.suppress_ragged_eofs and e.args == (-1, "Unexpected EOF"):
                return b""
            else:
                raise SocketError(str(e))
        except OpenSSL.SSL.ZeroReturnError:
            if self.connection.get_shutdown() == OpenSSL.SSL.RECEIVED_SHUTDOWN:
                return b""
            else:
                raise
        except OpenSSL.SSL.WantReadError:
            if not util.wait_for_read(self.socket, self.socket.gettimeout()):
                raise timeout("The read operation timed out")
            else:
                return self.recv(*args, **kwargs)

        # TLS 1.3 post-handshake authentication
        except OpenSSL.SSL.Error as e:
            raise ssl.SSLError("read error: %r" % e)
        else:
            return data

    def recv_into(self, *args, **kwargs):
        try:
            return self.connection.recv_into(*args, **kwargs)
        except OpenSSL.SSL.SysCallError as e:
            if self.suppress_ragged_eofs and e.args == (-1, "Unexpected EOF"):
                return 0
            else:
                raise SocketError(str(e))
        except OpenSSL.SSL.ZeroReturnError:
            if self.connection.get_shutdown() == OpenSSL.SSL.RECEIVED_SHUTDOWN:
                return 0
            else:
                raise
        except OpenSSL.SSL.WantReadError:
            if not util.wait_for_read(self.socket, self.socket.gettimeout()):
                raise timeout("The read operation timed out")
            else:
                return self.recv_into(*args, **kwargs)

        # TLS 1.3 post-handshake authentication
        except OpenSSL.SSL.Error as e:
            raise ssl.SSLError("read error: %r" % e)

    def settimeout(self, timeout):
        return self.socket.settimeout(timeout)

    def _send_until_done(self, data):
        while True:
            try:
                return self.connection.send(data)
            except OpenSSL.SSL.WantWriteError:
                if not util.wait_for_write(self.socket, self.socket.gettimeout()):
                    raise timeout()
                continue
            except OpenSSL.SSL.SysCallError as e:
                raise SocketError(str(e))

    def sendall(self, data):
        total_sent = 0
        while total_sent < len(data):
            sent = self._send_until_done(
                data[total_sent : total_sent + SSL_WRITE_BLOCKSIZE]
            )
            total_sent += sent

    def shutdown(self):
        # FIXME rethrow compatible exceptions should we ever use this
        self.connection.shutdown()

    def close(self):
        if self._makefile_refs < 1:
            try:
                self._closed = True
                return self.connection.close()
            except OpenSSL.SSL.Error:
                return
        else:
            self._makefile_refs -= 1

    def getpeercert(self, binary_form=False):
        x509 = self.connection.get_peer_certificate()

        if not x509:
            return x509

        if binary_form:
            return OpenSSL.crypto.dump_certificate(OpenSSL.crypto.FILETYPE_ASN1, x509)

        return {
            "subject": ((("commonName", x509.get_subject().CN),),),
            "subjectAltName": get_subj_alt_name(x509),
        }

    def version(self):
        return self.connection.get_protocol_version_name()

    def _reuse(self):
        self._makefile_refs += 1

    def _drop(self):
        if self._makefile_refs < 1:
            self.close()
        else:
            self._makefile_refs -= 1


if _fileobject:  # Platform-specific: Python 2

    def makefile(self, mode, bufsize=-1):
        self._makefile_refs += 1
        return _fileobject(self, mode, bufsize, close=True)


else:  # Platform-specific: Python 3
    makefile = backport_makefile

WrappedSocket.makefile = makefile


class PyOpenSSLContext(object):
    """
    I am a wrapper class for the PyOpenSSL ``Context`` object. I am responsible
    for translating the interface of the standard library ``SSLContext`` object
    to calls into PyOpenSSL.
    """

    def __init__(self, protocol):
        self.protocol = _openssl_versions[protocol]
        self._ctx = OpenSSL.SSL.Context(self.protocol)
        self._options = 0
        self.check_hostname = False

    @property
    def options(self):
        return self._options

    @options.setter
    def options(self, value):
        self._options = value
        self._ctx.set_options(value)

    @property
    def verify_mode(self):
        return _openssl_to_stdlib_verify[self._ctx.get_verify_mode()]

    @verify_mode.setter
    def verify_mode(self, value):
        self._ctx.set_verify(_stdlib_to_openssl_verify[value], _verify_callback)

    def set_default_verify_paths(self):
        self._ctx.set_default_verify_paths()

    def set_ciphers(self, ciphers):
        if isinstance(ciphers, six.text_type):
            ciphers = ciphers.encode("utf-8")
        self._ctx.set_cipher_list(ciphers)

    def load_verify_locations(self, cafile=None, capath=None, cadata=None):
        if cafile is not None:
            cafile = cafile.encode("utf-8")
        if capath is not None:
            capath = capath.encode("utf-8")
        try:
            self._ctx.load_verify_locations(cafile, capath)
            if cadata is not None:
                self._ctx.load_verify_locations(BytesIO(cadata))
        except OpenSSL.SSL.Error as e:
            raise ssl.SSLError("unable to load trusted certificates: %r" % e)

    def load_cert_chain(self, certfile, keyfile=None, password=None):
        self._ctx.use_certificate_chain_file(certfile)
        if password is not None:
            if not isinstance(password, six.binary_type):
                password = password.encode("utf-8")
            self._ctx.set_passwd_cb(lambda *_: password)
        self._ctx.use_privatekey_file(keyfile or certfile)

    def set_alpn_protocols(self, protocols):
        protocols = [six.ensure_binary(p) for p in protocols]
        return self._ctx.set_alpn_protos(protocols)

    def wrap_socket(
        self,
        sock,
        server_side=False,
        do_handshake_on_connect=True,
        suppress_ragged_eofs=True,
        server_hostname=None,
    ):
        cnx = OpenSSL.SSL.Connection(self._ctx, sock)

        if isinstance(server_hostname, six.text_type):  # Platform-specific: Python 3
            server_hostname = server_hostname.encode("utf-8")

        if server_hostname is not None:
            cnx.set_tlsext_host_name(server_hostname)

        cnx.set_connect_state()

        while True:
            try:
                cnx.do_handshake()
            except OpenSSL.SSL.WantReadError:
                if not util.wait_for_read(sock, sock.gettimeout()):
                    raise timeout("select timed out")
                continue
            except OpenSSL.SSL.Error as e:
                raise ssl.SSLError("bad handshake: %r" % e)
            break

        return WrappedSocket(cnx, sock)


def _verify_callback(cnx, x509, err_no, err_depth, return_code):
    return err_no == 0
site-packages/pip/_vendor/urllib3/contrib/ntlmpool.py000064400000010672150564745010016757 0ustar00"""
NTLM authenticating pool, contributed by erikcederstran

Issue #10, see: http://code.google.com/p/urllib3/issues/detail?id=10
"""
from __future__ import absolute_import

import warnings
from logging import getLogger

from ntlm import ntlm

from .. import HTTPSConnectionPool
from ..packages.six.moves.http_client import HTTPSConnection

warnings.warn(
    "The 'urllib3.contrib.ntlmpool' module is deprecated and will be removed "
    "in urllib3 v2.0 release, urllib3 is not able to support it properly due "
    "to reasons listed in issue: https://github.com/urllib3/urllib3/issues/2282. "
    "If you are a user of this module please comment in the mentioned issue.",
    DeprecationWarning,
)

log = getLogger(__name__)


class NTLMConnectionPool(HTTPSConnectionPool):
    """
    Implements an NTLM authentication version of an urllib3 connection pool
    """

    scheme = "https"

    def __init__(self, user, pw, authurl, *args, **kwargs):
        """
        authurl is a random URL on the server that is protected by NTLM.
        user is the Windows user, probably in the DOMAIN\\username format.
        pw is the password for the user.
        """
        super(NTLMConnectionPool, self).__init__(*args, **kwargs)
        self.authurl = authurl
        self.rawuser = user
        user_parts = user.split("\\", 1)
        self.domain = user_parts[0].upper()
        self.user = user_parts[1]
        self.pw = pw

    def _new_conn(self):
        # Performs the NTLM handshake that secures the connection. The socket
        # must be kept open while requests are performed.
        self.num_connections += 1
        log.debug(
            "Starting NTLM HTTPS connection no. %d: https://%s%s",
            self.num_connections,
            self.host,
            self.authurl,
        )

        headers = {"Connection": "Keep-Alive"}
        req_header = "Authorization"
        resp_header = "www-authenticate"

        conn = HTTPSConnection(host=self.host, port=self.port)

        # Send negotiation message
        headers[req_header] = "NTLM %s" % ntlm.create_NTLM_NEGOTIATE_MESSAGE(
            self.rawuser
        )
        log.debug("Request headers: %s", headers)
        conn.request("GET", self.authurl, None, headers)
        res = conn.getresponse()
        reshdr = dict(res.getheaders())
        log.debug("Response status: %s %s", res.status, res.reason)
        log.debug("Response headers: %s", reshdr)
        log.debug("Response data: %s [...]", res.read(100))

        # Remove the reference to the socket, so that it can not be closed by
        # the response object (we want to keep the socket open)
        res.fp = None

        # Server should respond with a challenge message
        auth_header_values = reshdr[resp_header].split(", ")
        auth_header_value = None
        for s in auth_header_values:
            if s[:5] == "NTLM ":
                auth_header_value = s[5:]
        if auth_header_value is None:
            raise Exception(
                "Unexpected %s response header: %s" % (resp_header, reshdr[resp_header])
            )

        # Send authentication message
        ServerChallenge, NegotiateFlags = ntlm.parse_NTLM_CHALLENGE_MESSAGE(
            auth_header_value
        )
        auth_msg = ntlm.create_NTLM_AUTHENTICATE_MESSAGE(
            ServerChallenge, self.user, self.domain, self.pw, NegotiateFlags
        )
        headers[req_header] = "NTLM %s" % auth_msg
        log.debug("Request headers: %s", headers)
        conn.request("GET", self.authurl, None, headers)
        res = conn.getresponse()
        log.debug("Response status: %s %s", res.status, res.reason)
        log.debug("Response headers: %s", dict(res.getheaders()))
        log.debug("Response data: %s [...]", res.read()[:100])
        if res.status != 200:
            if res.status == 401:
                raise Exception("Server rejected request: wrong username or password")
            raise Exception("Wrong server response: %s %s" % (res.status, res.reason))

        res.fp = None
        log.debug("Connection established")
        return conn

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=3,
        redirect=True,
        assert_same_host=True,
    ):
        if headers is None:
            headers = {}
        headers["Connection"] = "Keep-Alive"
        return super(NTLMConnectionPool, self).urlopen(
            method, url, body, headers, retries, redirect, assert_same_host
        )
site-packages/pip/_vendor/urllib3/contrib/socks.py000064400000015671150564745010016241 0ustar00# -*- coding: utf-8 -*-
"""
This module contains provisional support for SOCKS proxies from within
urllib3. This module supports SOCKS4, SOCKS4A (an extension of SOCKS4), and
SOCKS5. To enable its functionality, either install PySocks or install this
module with the ``socks`` extra.

The SOCKS implementation supports the full range of urllib3 features. It also
supports the following SOCKS features:

- SOCKS4A (``proxy_url='socks4a://...``)
- SOCKS4 (``proxy_url='socks4://...``)
- SOCKS5 with remote DNS (``proxy_url='socks5h://...``)
- SOCKS5 with local DNS (``proxy_url='socks5://...``)
- Usernames and passwords for the SOCKS proxy

.. note::
   It is recommended to use ``socks5h://`` or ``socks4a://`` schemes in
   your ``proxy_url`` to ensure that DNS resolution is done from the remote
   server instead of client-side when connecting to a domain name.

SOCKS4 supports IPv4 and domain names with the SOCKS4A extension. SOCKS5
supports IPv4, IPv6, and domain names.

When connecting to a SOCKS4 proxy the ``username`` portion of the ``proxy_url``
will be sent as the ``userid`` section of the SOCKS request:

.. code-block:: python

    proxy_url="socks4a://<userid>@proxy-host"

When connecting to a SOCKS5 proxy the ``username`` and ``password`` portion
of the ``proxy_url`` will be sent as the username/password to authenticate
with the proxy:

.. code-block:: python

    proxy_url="socks5h://<username>:<password>@proxy-host"

"""
from __future__ import absolute_import

try:
    import socks
except ImportError:
    import warnings

    from ..exceptions import DependencyWarning

    warnings.warn(
        (
            "SOCKS support in urllib3 requires the installation of optional "
            "dependencies: specifically, PySocks.  For more information, see "
            "https://urllib3.readthedocs.io/en/1.26.x/contrib.html#socks-proxies"
        ),
        DependencyWarning,
    )
    raise

from socket import error as SocketError
from socket import timeout as SocketTimeout

from ..connection import HTTPConnection, HTTPSConnection
from ..connectionpool import HTTPConnectionPool, HTTPSConnectionPool
from ..exceptions import ConnectTimeoutError, NewConnectionError
from ..poolmanager import PoolManager
from ..util.url import parse_url

try:
    import ssl
except ImportError:
    ssl = None


class SOCKSConnection(HTTPConnection):
    """
    A plain-text HTTP connection that connects via a SOCKS proxy.
    """

    def __init__(self, *args, **kwargs):
        self._socks_options = kwargs.pop("_socks_options")
        super(SOCKSConnection, self).__init__(*args, **kwargs)

    def _new_conn(self):
        """
        Establish a new connection via the SOCKS proxy.
        """
        extra_kw = {}
        if self.source_address:
            extra_kw["source_address"] = self.source_address

        if self.socket_options:
            extra_kw["socket_options"] = self.socket_options

        try:
            conn = socks.create_connection(
                (self.host, self.port),
                proxy_type=self._socks_options["socks_version"],
                proxy_addr=self._socks_options["proxy_host"],
                proxy_port=self._socks_options["proxy_port"],
                proxy_username=self._socks_options["username"],
                proxy_password=self._socks_options["password"],
                proxy_rdns=self._socks_options["rdns"],
                timeout=self.timeout,
                **extra_kw
            )

        except SocketTimeout:
            raise ConnectTimeoutError(
                self,
                "Connection to %s timed out. (connect timeout=%s)"
                % (self.host, self.timeout),
            )

        except socks.ProxyError as e:
            # This is fragile as hell, but it seems to be the only way to raise
            # useful errors here.
            if e.socket_err:
                error = e.socket_err
                if isinstance(error, SocketTimeout):
                    raise ConnectTimeoutError(
                        self,
                        "Connection to %s timed out. (connect timeout=%s)"
                        % (self.host, self.timeout),
                    )
                else:
                    raise NewConnectionError(
                        self, "Failed to establish a new connection: %s" % error
                    )
            else:
                raise NewConnectionError(
                    self, "Failed to establish a new connection: %s" % e
                )

        except SocketError as e:  # Defensive: PySocks should catch all these.
            raise NewConnectionError(
                self, "Failed to establish a new connection: %s" % e
            )

        return conn


# We don't need to duplicate the Verified/Unverified distinction from
# urllib3/connection.py here because the HTTPSConnection will already have been
# correctly set to either the Verified or Unverified form by that module. This
# means the SOCKSHTTPSConnection will automatically be the correct type.
class SOCKSHTTPSConnection(SOCKSConnection, HTTPSConnection):
    pass


class SOCKSHTTPConnectionPool(HTTPConnectionPool):
    ConnectionCls = SOCKSConnection


class SOCKSHTTPSConnectionPool(HTTPSConnectionPool):
    ConnectionCls = SOCKSHTTPSConnection


class SOCKSProxyManager(PoolManager):
    """
    A version of the urllib3 ProxyManager that routes connections via the
    defined SOCKS proxy.
    """

    pool_classes_by_scheme = {
        "http": SOCKSHTTPConnectionPool,
        "https": SOCKSHTTPSConnectionPool,
    }

    def __init__(
        self,
        proxy_url,
        username=None,
        password=None,
        num_pools=10,
        headers=None,
        **connection_pool_kw
    ):
        parsed = parse_url(proxy_url)

        if username is None and password is None and parsed.auth is not None:
            split = parsed.auth.split(":")
            if len(split) == 2:
                username, password = split
        if parsed.scheme == "socks5":
            socks_version = socks.PROXY_TYPE_SOCKS5
            rdns = False
        elif parsed.scheme == "socks5h":
            socks_version = socks.PROXY_TYPE_SOCKS5
            rdns = True
        elif parsed.scheme == "socks4":
            socks_version = socks.PROXY_TYPE_SOCKS4
            rdns = False
        elif parsed.scheme == "socks4a":
            socks_version = socks.PROXY_TYPE_SOCKS4
            rdns = True
        else:
            raise ValueError("Unable to determine SOCKS version from %s" % proxy_url)

        self.proxy_url = proxy_url

        socks_options = {
            "socks_version": socks_version,
            "proxy_host": parsed.host,
            "proxy_port": parsed.port,
            "username": username,
            "password": password,
            "rdns": rdns,
        }
        connection_pool_kw["_socks_options"] = socks_options

        super(SOCKSProxyManager, self).__init__(
            num_pools, headers, **connection_pool_kw
        )

        self.pool_classes_by_scheme = SOCKSProxyManager.pool_classes_by_scheme
site-packages/pip/_vendor/urllib3/contrib/appengine.py000064400000025432150564745010017061 0ustar00"""
This module provides a pool manager that uses Google App Engine's
`URLFetch Service <https://cloud.google.com/appengine/docs/python/urlfetch>`_.

Example usage::

    from pip._vendor.urllib3 import PoolManager
    from pip._vendor.urllib3.contrib.appengine import AppEngineManager, is_appengine_sandbox

    if is_appengine_sandbox():
        # AppEngineManager uses AppEngine's URLFetch API behind the scenes
        http = AppEngineManager()
    else:
        # PoolManager uses a socket-level API behind the scenes
        http = PoolManager()

    r = http.request('GET', 'https://google.com/')

There are `limitations <https://cloud.google.com/appengine/docs/python/\
urlfetch/#Python_Quotas_and_limits>`_ to the URLFetch service and it may not be
the best choice for your application. There are three options for using
urllib3 on Google App Engine:

1. You can use :class:`AppEngineManager` with URLFetch. URLFetch is
   cost-effective in many circumstances as long as your usage is within the
   limitations.
2. You can use a normal :class:`~urllib3.PoolManager` by enabling sockets.
   Sockets also have `limitations and restrictions
   <https://cloud.google.com/appengine/docs/python/sockets/\
   #limitations-and-restrictions>`_ and have a lower free quota than URLFetch.
   To use sockets, be sure to specify the following in your ``app.yaml``::

        env_variables:
            GAE_USE_SOCKETS_HTTPLIB : 'true'

3. If you are using `App Engine Flexible
<https://cloud.google.com/appengine/docs/flexible/>`_, you can use the standard
:class:`PoolManager` without any configuration or special environment variables.
"""

from __future__ import absolute_import

import io
import logging
import warnings

from ..exceptions import (
    HTTPError,
    HTTPWarning,
    MaxRetryError,
    ProtocolError,
    SSLError,
    TimeoutError,
)
from ..packages.six.moves.urllib.parse import urljoin
from ..request import RequestMethods
from ..response import HTTPResponse
from ..util.retry import Retry
from ..util.timeout import Timeout
from . import _appengine_environ

try:
    from google.appengine.api import urlfetch
except ImportError:
    urlfetch = None


log = logging.getLogger(__name__)


class AppEnginePlatformWarning(HTTPWarning):
    pass


class AppEnginePlatformError(HTTPError):
    pass


class AppEngineManager(RequestMethods):
    """
    Connection manager for Google App Engine sandbox applications.

    This manager uses the URLFetch service directly instead of using the
    emulated httplib, and is subject to URLFetch limitations as described in
    the App Engine documentation `here
    <https://cloud.google.com/appengine/docs/python/urlfetch>`_.

    Notably it will raise an :class:`AppEnginePlatformError` if:
        * URLFetch is not available.
        * If you attempt to use this on App Engine Flexible, as full socket
          support is available.
        * If a request size is more than 10 megabytes.
        * If a response size is more than 32 megabytes.
        * If you use an unsupported request method such as OPTIONS.

    Beyond those cases, it will raise normal urllib3 errors.
    """

    def __init__(
        self,
        headers=None,
        retries=None,
        validate_certificate=True,
        urlfetch_retries=True,
    ):
        if not urlfetch:
            raise AppEnginePlatformError(
                "URLFetch is not available in this environment."
            )

        warnings.warn(
            "urllib3 is using URLFetch on Google App Engine sandbox instead "
            "of sockets. To use sockets directly instead of URLFetch see "
            "https://urllib3.readthedocs.io/en/1.26.x/reference/urllib3.contrib.html.",
            AppEnginePlatformWarning,
        )

        RequestMethods.__init__(self, headers)
        self.validate_certificate = validate_certificate
        self.urlfetch_retries = urlfetch_retries

        self.retries = retries or Retry.DEFAULT

    def __enter__(self):
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        # Return False to re-raise any potential exceptions
        return False

    def urlopen(
        self,
        method,
        url,
        body=None,
        headers=None,
        retries=None,
        redirect=True,
        timeout=Timeout.DEFAULT_TIMEOUT,
        **response_kw
    ):

        retries = self._get_retries(retries, redirect)

        try:
            follow_redirects = redirect and retries.redirect != 0 and retries.total
            response = urlfetch.fetch(
                url,
                payload=body,
                method=method,
                headers=headers or {},
                allow_truncated=False,
                follow_redirects=self.urlfetch_retries and follow_redirects,
                deadline=self._get_absolute_timeout(timeout),
                validate_certificate=self.validate_certificate,
            )
        except urlfetch.DeadlineExceededError as e:
            raise TimeoutError(self, e)

        except urlfetch.InvalidURLError as e:
            if "too large" in str(e):
                raise AppEnginePlatformError(
                    "URLFetch request too large, URLFetch only "
                    "supports requests up to 10mb in size.",
                    e,
                )
            raise ProtocolError(e)

        except urlfetch.DownloadError as e:
            if "Too many redirects" in str(e):
                raise MaxRetryError(self, url, reason=e)
            raise ProtocolError(e)

        except urlfetch.ResponseTooLargeError as e:
            raise AppEnginePlatformError(
                "URLFetch response too large, URLFetch only supports"
                "responses up to 32mb in size.",
                e,
            )

        except urlfetch.SSLCertificateError as e:
            raise SSLError(e)

        except urlfetch.InvalidMethodError as e:
            raise AppEnginePlatformError(
                "URLFetch does not support method: %s" % method, e
            )

        http_response = self._urlfetch_response_to_http_response(
            response, retries=retries, **response_kw
        )

        # Handle redirect?
        redirect_location = redirect and http_response.get_redirect_location()
        if redirect_location:
            # Check for redirect response
            if self.urlfetch_retries and retries.raise_on_redirect:
                raise MaxRetryError(self, url, "too many redirects")
            else:
                if http_response.status == 303:
                    method = "GET"

                try:
                    retries = retries.increment(
                        method, url, response=http_response, _pool=self
                    )
                except MaxRetryError:
                    if retries.raise_on_redirect:
                        raise MaxRetryError(self, url, "too many redirects")
                    return http_response

                retries.sleep_for_retry(http_response)
                log.debug("Redirecting %s -> %s", url, redirect_location)
                redirect_url = urljoin(url, redirect_location)
                return self.urlopen(
                    method,
                    redirect_url,
                    body,
                    headers,
                    retries=retries,
                    redirect=redirect,
                    timeout=timeout,
                    **response_kw
                )

        # Check if we should retry the HTTP response.
        has_retry_after = bool(http_response.getheader("Retry-After"))
        if retries.is_retry(method, http_response.status, has_retry_after):
            retries = retries.increment(method, url, response=http_response, _pool=self)
            log.debug("Retry: %s", url)
            retries.sleep(http_response)
            return self.urlopen(
                method,
                url,
                body=body,
                headers=headers,
                retries=retries,
                redirect=redirect,
                timeout=timeout,
                **response_kw
            )

        return http_response

    def _urlfetch_response_to_http_response(self, urlfetch_resp, **response_kw):

        if is_prod_appengine():
            # Production GAE handles deflate encoding automatically, but does
            # not remove the encoding header.
            content_encoding = urlfetch_resp.headers.get("content-encoding")

            if content_encoding == "deflate":
                del urlfetch_resp.headers["content-encoding"]

        transfer_encoding = urlfetch_resp.headers.get("transfer-encoding")
        # We have a full response's content,
        # so let's make sure we don't report ourselves as chunked data.
        if transfer_encoding == "chunked":
            encodings = transfer_encoding.split(",")
            encodings.remove("chunked")
            urlfetch_resp.headers["transfer-encoding"] = ",".join(encodings)

        original_response = HTTPResponse(
            # In order for decoding to work, we must present the content as
            # a file-like object.
            body=io.BytesIO(urlfetch_resp.content),
            msg=urlfetch_resp.header_msg,
            headers=urlfetch_resp.headers,
            status=urlfetch_resp.status_code,
            **response_kw
        )

        return HTTPResponse(
            body=io.BytesIO(urlfetch_resp.content),
            headers=urlfetch_resp.headers,
            status=urlfetch_resp.status_code,
            original_response=original_response,
            **response_kw
        )

    def _get_absolute_timeout(self, timeout):
        if timeout is Timeout.DEFAULT_TIMEOUT:
            return None  # Defer to URLFetch's default.
        if isinstance(timeout, Timeout):
            if timeout._read is not None or timeout._connect is not None:
                warnings.warn(
                    "URLFetch does not support granular timeout settings, "
                    "reverting to total or default URLFetch timeout.",
                    AppEnginePlatformWarning,
                )
            return timeout.total
        return timeout

    def _get_retries(self, retries, redirect):
        if not isinstance(retries, Retry):
            retries = Retry.from_int(retries, redirect=redirect, default=self.retries)

        if retries.connect or retries.read or retries.redirect:
            warnings.warn(
                "URLFetch only supports total retries and does not "
                "recognize connect, read, or redirect retry parameters.",
                AppEnginePlatformWarning,
            )

        return retries


# Alias methods from _appengine_environ to maintain public API interface.

is_appengine = _appengine_environ.is_appengine
is_appengine_sandbox = _appengine_environ.is_appengine_sandbox
is_local_appengine = _appengine_environ.is_local_appengine
is_prod_appengine = _appengine_environ.is_prod_appengine
is_prod_appengine_mvms = _appengine_environ.is_prod_appengine_mvms
site-packages/pip/_vendor/urllib3/contrib/_appengine_environ.py000064400000001675150564745010020763 0ustar00"""
This module provides means to detect the App Engine environment.
"""

import os


def is_appengine():
    return is_local_appengine() or is_prod_appengine()


def is_appengine_sandbox():
    """Reports if the app is running in the first generation sandbox.

    The second generation runtimes are technically still in a sandbox, but it
    is much less restrictive, so generally you shouldn't need to check for it.
    see https://cloud.google.com/appengine/docs/standard/runtimes
    """
    return is_appengine() and os.environ["APPENGINE_RUNTIME"] == "python27"


def is_local_appengine():
    return "APPENGINE_RUNTIME" in os.environ and os.environ.get(
        "SERVER_SOFTWARE", ""
    ).startswith("Development/")


def is_prod_appengine():
    return "APPENGINE_RUNTIME" in os.environ and os.environ.get(
        "SERVER_SOFTWARE", ""
    ).startswith("Google App Engine/")


def is_prod_appengine_mvms():
    """Deprecated."""
    return False
site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/low_level.cpython-39.pyc000064400000021777150564745010027004 0ustar00a

��Reb6�@s�dZddlZddlZddlZddlZddlZddlZddlZddlZddl	m
Z
mZmZe�
dej�Zdd�Zdd	�Zd
d�Zdd
�Zdd�Zd&dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zdd�Zddd d!d"d#�Zd$d%�ZdS)'a�
Low-level helpers for the SecureTransport bindings.

These are Python functions that are not directly related to the high-level APIs
but are necessary to get them to work. They include a whole bunch of low-level
CoreFoundation messing about and memory management. The concerns in this module
are almost entirely about trying to avoid memory leaks and providing
appropriate and useful assistance to the higher-level code.
�N�)�CFConst�CoreFoundation�Securitys;-----BEGIN CERTIFICATE-----
(.*?)
-----END CERTIFICATE-----cCst�tj|t|��S)zv
    Given a bytestring, create a CFData object from it. This CFData object must
    be CFReleased by the caller.
    )r�CFDataCreate�kCFAllocatorDefault�len)Z
bytestring�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/_securetransport/low_level.py�_cf_data_from_bytess�rcCsZt|�}dd�|D�}dd�|D�}tj||�}tj||�}t�tj|||tjtj�S)zK
    Given a list of Python tuples, create an associated CFDictionary.
    css|]}|dVqdS)rNr	��.0�tr	r	r
�	<genexpr>,�z-_cf_dictionary_from_tuples.<locals>.<genexpr>css|]}|dVqdS)rNr	rr	r	r
r-r)rr�	CFTypeRefZCFDictionaryCreaterZkCFTypeDictionaryKeyCallBacksZkCFTypeDictionaryValueCallBacks)ZtuplesZdictionary_size�keys�valuesZcf_keysZ	cf_valuesr	r	r
�_cf_dictionary_from_tuples%s�rcCs t�|�}t�tj|tj�}|S)zi
    Given a Python binary data, create a CFString.
    The string must be CFReleased by the caller.
    )�ctypes�c_char_prZCFStringCreateWithCStringrr�kCFStringEncodingUTF8)Zpy_bstrZc_str�cf_strr	r	r
�_cfstr;s
�rc
Cs�d}znt�tjdt�tj��}|s*td��|D]@}t|�}|sFtd��zt�||�Wt�	|�q.t�	|�0q.WnBt
y�}z*|r�t�	|�t�d|f��WYd}~n
d}~00|S)z�
    Given a list of Python binary data, create an associated CFMutableArray.
    The array must be CFReleased by the caller.

    Raises an ssl.SSLError on failure.
    Nr�Unable to allocate memory!zUnable to allocate array: %s)
r�CFArrayCreateMutablerr�byref�kCFTypeArrayCallBacks�MemoryErrorr�CFArrayAppendValue�	CFRelease�
BaseException�ssl�SSLError)�lstZcf_arr�itemr�er	r	r
�_create_cfstring_arrayIs*
�
&r'cCsnt�|t�tj��}t�|tj�}|durXt�d�}t�	||dtj�}|sRt
d��|j}|durj|�d�}|S)z�
    Creates a Unicode string from a CFString object. Used entirely for error
    reporting.

    Yes, it annoys me quite a lot that this function is this complex.
    Niz'Error copying C string from CFStringRef�utf-8)
r�cast�POINTER�c_void_prZCFStringGetCStringPtrrr�create_string_bufferZCFStringGetCString�OSError�value�decode)r.Zvalue_as_void_p�string�buffer�resultr	r	r
�_cf_string_to_unicodehs�

�
r3cCs\|dkrdSt�|d�}t|�}t�|�|dus:|dkrBd|}|durPtj}||��dS)z[
    Checks the return code and throws an exception if there is an error to
    report
    rN�zOSStatus %s)rZSecCopyErrorMessageStringr3rr r"r#)�errorZexception_classZcf_error_string�outputr	r	r
�_assert_no_error�s
r7cCs�|�dd�}dd�t�|�D�}|s.t�d��t�tjdt�	tj
��}|sTt�d��z`|D]V}t|�}|stt�d��t�
tj|�}t�|�|s�t�d��t�||�t�|�qZWnty�t�|��Yn0|S)	z�
    Given a bundle of certs in PEM format, turns them into a CFArray of certs
    that can be used to validate a cert chain.
    s
�
cSsg|]}t�|�d���qS)r)�base64�	b64decode�group)r
�matchr	r	r
�
<listcomp>�sz(_cert_array_from_pem.<locals>.<listcomp>zNo root certificates specifiedrrzUnable to build cert object!)�replace�
_PEM_CERTS_RE�finditerr"r#rrrrrrrrZSecCertificateCreateWithDatar r�	Exception)Z
pem_bundleZ	der_certsZ
cert_arrayZ	der_bytesZcertdata�certr	r	r
�_cert_array_from_pem�s<�

�

�


rCcCst��}t�|�|kS)z=
    Returns True if a given CFTypeRef is a certificate.
    )rZSecCertificateGetTypeIDr�CFGetTypeID�r%�expectedr	r	r
�_is_cert�srGcCst��}t�|�|kS)z;
    Returns True if a given CFTypeRef is an identity.
    )rZSecIdentityGetTypeIDrrDrEr	r	r
�_is_identity�srHc
Cs�t�d�}t�|dd���d�}t�|dd��}t��}tj�||��	d�}t
��}t
�|t
|�|ddt�|��}t|�||fS)a�
    This function creates a temporary Mac keychain that we can use to work with
    credentials. This keychain uses a one-time password and a temporary file to
    store the data. We expect to have one keychain per socket. The returned
    SecKeychainRef must be freed by the caller, including calling
    SecKeychainDelete.

    Returns a tuple of the SecKeychainRef and the path to the temporary
    directory that contains it.
    �(N�r(F)�os�urandomr9�	b16encoder/�tempfile�mkdtemp�path�join�encoderZSecKeychainRefZSecKeychainCreaterrrr7)Zrandom_bytes�filename�passwordZ
tempdirectoryZ
keychain_path�keychain�statusr	r	r
�_temporary_keychain�s
�rWcCs.g}g}d}t|d��}|��}Wd�n1s40Yz�t�tj|t|��}t��}t�|ddddd|t	�
|��}t|�t�|�}	t
|	�D]X}
t�||
�}t	�|tj�}t|�r�t�|�|�|�q�t|�r�t�|�|�|�q�W|r�t�|�t�|�n|�rt�|�t�|�0||fS)z�
    Given a single file, loads all the trust objects from it into arrays and
    the keychain.
    Returns a tuple of lists: the first list is a list of identities, the
    second a list of certs.
    N�rbr)�open�readrrrrZ
CFArrayRefrZ
SecItemImportrrr7ZCFArrayGetCount�rangeZCFArrayGetValueAtIndexr)rrGZCFRetain�appendrHr )rUrP�certificates�
identitiesZresult_array�fZraw_filedataZfiledatar2Zresult_count�indexr%r	r	r
�_load_items_from_file�sL&��




�
rac
Gsg}g}dd�|D�}z�|D]&}t||�\}}|�|�|�|�q|s�t��}t�||dt�|��}t|�|�|�t	�
|�d��t	�t	j
dt�t	j��}	t�||�D]}
t	�|	|
�q�|	Wt�||�D]}t	�
|�q�S]}t	�
|�q�n t�||�D]}t	�
|��q0dS)z�
    Load certificates and maybe keys from a number of files. Has the end goal
    of returning a CFArray containing one SecIdentityRef, and then zero or more
    SecCertificateRef objects, suitable for use as a client certificate trust
    chain.
    css|]}|r|VqdS)Nr	)r
rPr	r	r
rRrz*_load_client_cert_chain.<locals>.<genexpr>rN)ra�extendrZSecIdentityRefZ SecIdentityCreateWithCertificaterrr7r\rr �poprrr�	itertools�chainr)rU�pathsr]r^�	file_pathZnew_identitiesZ	new_certsZnew_identityrVZtrust_chainr%�objr	r	r
�_load_client_cert_chain.s> 
�

���ri)r�)�r)rkr)rkrj)rkrk)ZSSLv2�SSLv3�TLSv1zTLSv1.1zTLSv1.2c	CsHt|\}}d}d}t�d||�}t|�}d}t�d||||�|}|S)z6
    Builds a TLS alert record for an unknown CA.
    rj�0z>BB�z>BBBH)�TLS_PROTOCOL_VERSIONS�struct�packr)	�versionZver_majZver_minZseverity_fatalZdescription_unknown_ca�msgZmsg_lenZrecord_type_alert�recordr	r	r
�_build_tls_unknown_ca_alert�srv)N)�__doc__r9rrdrK�rer"rqrNZbindingsrrr�compile�DOTALLr?rrrr'r3r7rCrGrHrWrarirprvr	r	r	r
�<module>s>	�

.#7L�	site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/__init__.cpython-39.pyc000064400000000377150564745010026544 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/_securetransport/__init__.py�<module>�site-packages/pip/_vendor/urllib3/contrib/_securetransport/__pycache__/bindings.cpython-39.pyc000064400000025012150564745010026573 0ustar00a

��Re�D�	@s�dZddlmZddlZddlmZmZmZmZm	Z	m
Z
mZmZm
Z
mZmZmZddlmZddlmZe��dkr|ed��e��dZeeee�d	���Zed
kr�edededf��d
d�Zedd�Z edd�Z!eZ"eZ#eZ$eZ%eZ&eZ'eZ(eZ)eZ*eZ+eZ,ee+�Z-eZ.eZ/ee%�Z0ee&�Z1ee'�Z2ee(�Z3ee)�Z4eZ5eZ6eZ7ee�Z8eZ9eZ:ee�Z;eZ<eZ=ee�Z>eZ?eZ@ee�ZAee�ZBeZCeZDeZEeZFeZGeZH�z�e0e1ee9�ee:�e<ee=�e>ee2�ge jI_Je/e jI_Kge jL_Je,e jL_Kge jM_Je,e jM_Kge jN_Je,e jN_Ke.e0ge jO_Je8e jO_Ke8ge jP_Je0e jP_Ke/ege jQ_Je1e jQ_Ke-e8ee;�ge jR_Je/e jR_Ke
eee"eee>�ge jS_Je/e jS_Ke>ge jT_Je/e jT_Ke0e4ee2�ge jU_Je/e jU_Kee/eCeee
��ZVee/eCee	�ee
��ZWeAeVeWge jX_Je/e jX_KeAe
e
ge jY_Je/e jY_KeAe2ge jZ_Je/e jZ_KeAe-e"ge j[_Je/e j[_KeAeCge j\_Je/e j\_KeAe
e
ge j]_Je/e j]_KeAge j^_Je/e j^_KeAe
e
ee
�ge j__Je/e j__KeAe
e
ee
�ge j`_Je/e j`_KeAge ja_Je/e ja_KeAee
�ge jb_Je/e jb_KeAee@�ee
�ge jc_Je/e jc_KeAee@�e
ge jd_Je/e jd_KeAee
�ge je_fe/e je_KeAee@�ee
�ge jg_Je/e jg_KeAee@�ge jh_Je/e jh_KeAee?�ge ji_Je/e ji_KeAeeB�ge jj_Je/e jj_KeBe2ge jk_Je/e jk_KeBe"ge jl_me/e jl_KeBeeD�ge jn_Je/e jn_KeBge jo_Je#e jo_KeBe#ge jp_Je8e jp_Ke.eFeGge jq_JeAe jq_KeAeHe"ge jr_Je/e jr_KeAe?ge js_Je/e js_KeAe?ge jt_Je/e jt_KzeAe2ge ju_Je/e ju_KWnev�y\Yn0e/ege jQ_Je1e jQ_KeVe _VeWe _WeAe _Ae?e _?e@e _@e;e _;e>e _>eBe _BeDe _De9e _9e/e _/e1�we d�e _xe1�we d�e _ye-ge!jz_Je-e!jz_Ke-ge!j{_Jde!j{_Ke-ge!j|_Je,e!j|_Ke.e
e$ge!j}_Je1e!j}_Ke1e$ge!j~_Je
e!j~_Ke1e
e#e$ge!j_Jee!j_Ke.e
e#ge!j�_Je0e!j�_Ke0ge!j�_Je#e!j�_Ke0ge!j�_Jee!j�_Ke.ee-�ee-�e#e6e7ge!j�_Je4e!j�_Ke4e-ge!j�_Je-e!j�_Ke.ee-�e#e5ge!j�_Je2e!j�_Ke.e#e5ge!j�_Je3e!j�_Ke3ege!j�_Jde!j�_Ke2ge!j�_Je#e!j�_Ke2e#ge!j�_Jee!j�_Ke.�we!d�e!_�e�we!d�e!_�e�we!d�e!_�e�we!d�e!_�e-e!_-e2e!_2e1e!_1e4e!_4Wnev�y�ed��Yn0Gdd�de��Z�Gdd�de��Z�dS)ay
This module uses ctypes to bind a whole bunch of functions and constants from
SecureTransport. The goal here is to provide the low-level API to
SecureTransport. These are essentially the C-level functions and constants, and
they're pretty gross to work with.

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond <will@wbond.net>

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
�)�absolute_importN)�CDLL�	CFUNCTYPE�POINTER�c_bool�c_byte�c_char_p�c_int32�c_long�c_size_t�c_uint32�c_ulong�c_void_p)�find_library)�
raise_from�DarwinzOnly macOS is supported�.)�
�z1Only OS X 10.8 and newer are supported, not %s.%s�cCsVz,tdkr|}nt|�}|s t�t|dd�WStyPttd|�d�Yn0dS)z:Loads a CDLL by name, falling back to known path on 10.16+)r�T)�	use_errnozThe library %s failed to loadN)�version_infor�OSErrorrr�ImportError)�nameZmacos10_16_path�path�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/urllib3/contrib/_securetransport/bindings.py�	load_cdllAsr�Securityz6/System/Library/Frameworks/Security.framework/Security�CoreFoundationzB/System/Library/Frameworks/CoreFoundation.framework/CoreFoundation�kSecImportExportPassphrase�kSecImportItemIdentity�kCFAllocatorDefault�kCFTypeArrayCallBacks�kCFTypeDictionaryKeyCallBacks�kCFTypeDictionaryValueCallBackszError initializing ctypesc@seZdZdZed�ZdS)�CFConstz_
    A class object that acts as essentially a namespace for CoreFoundation
    constants.
    iN)�__name__�
__module__�__qualname__�__doc__�CFStringEncodingZkCFStringEncodingUTF8rrrrr(�sr(c@s(eZdZdZdZdZdZdZdZdZ	dZ
d	ZdZdZ
dZdZdZd
ZdZdZdZdZd
ZdZdZdZdZdZdZdZdZdZdZ dZ!dZ"dZ#dZ$dZ%dZ&dZ'dZ(d Z)d!Z*d"Z+d#Z,d$Z-d%Z.d&Z/d'Z0d(Z1d)Z2d*Z3d+Z4d,Z5d-Z6d.Z7d/Z8d0Z9d1Z:d2Z;d3Z<d4Z=d5Z>d6Z?d7Z@d8ZAd9ZBd:ZCd;ZDd<ZEd=ZFd>ZGd?ZHd@ZIdAS)B�
SecurityConstzU
    A class object that acts as essentially a namespace for Security constants.
    rr���rri����i���i���i���i���i���i���i���i���i���i���i���i���i���i���i���i���i���i���i ���iQ���i,���iR���i,�i0�i+�i/�i��i����i$�i(�i
�i��k�9i#�i'�i	�i��g�3���=�<�5�/iiiiN)Jr)r*r+r,Z"kSSLSessionOptionBreakOnServerAuthZ
kSSLProtocol2Z
kSSLProtocol3Z
kTLSProtocol1ZkTLSProtocol11ZkTLSProtocol12ZkTLSProtocol13ZkTLSProtocolMaxSupportedZkSSLClientSideZkSSLStreamTypeZkSecFormatPEMSequenceZkSecTrustResultInvalidZkSecTrustResultProceedZkSecTrustResultDenyZkSecTrustResultUnspecifiedZ&kSecTrustResultRecoverableTrustFailureZ kSecTrustResultFatalTrustFailureZkSecTrustResultOtherErrorZerrSSLProtocolZerrSSLWouldBlockZerrSSLClosedGracefulZerrSSLClosedNoNotifyZerrSSLClosedAbortZerrSSLXCertChainInvalidZerrSSLCryptoZerrSSLInternalZerrSSLCertExpiredZerrSSLCertNotYetValidZerrSSLUnknownRootCertZerrSSLNoRootCertZerrSSLHostNameMismatchZerrSSLPeerHandshakeFailZerrSSLPeerUserCancelledZerrSSLWeakPeerEphemeralDHKeyZerrSSLServerAuthCompletedZerrSSLRecordOverflowZerrSecVerifyFailedZerrSecNoTrustSettingsZerrSecItemNotFoundZerrSecInvalidTrustSettingsZ'TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384Z'TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256Z-TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256Z+TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256Z#TLS_DHE_RSA_WITH_AES_256_GCM_SHA384Z#TLS_DHE_RSA_WITH_AES_128_GCM_SHA256Z'TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384Z%TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384Z$TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_256_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_256_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_256_CBC_SHAZ'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256Z%TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256Z$TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHAZ"TLS_ECDHE_RSA_WITH_AES_128_CBC_SHAZ#TLS_DHE_RSA_WITH_AES_128_CBC_SHA256Z TLS_DHE_RSA_WITH_AES_128_CBC_SHAZTLS_RSA_WITH_AES_256_GCM_SHA384ZTLS_RSA_WITH_AES_128_GCM_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHA256ZTLS_RSA_WITH_AES_128_CBC_SHA256ZTLS_RSA_WITH_AES_256_CBC_SHAZTLS_RSA_WITH_AES_128_CBC_SHAZTLS_AES_128_GCM_SHA256ZTLS_AES_256_GCM_SHA384ZTLS_AES_128_CCM_8_SHA256ZTLS_AES_128_CCM_SHA256rrrrr.�s�r.)�r,�
__future__r�platform�ctypesrrrrrrr	r
rrr
rZctypes.utilrZ pip._vendor.urllib3.packages.sixr�systemr�mac_ver�version�tuple�map�int�splitrrrr r!�BooleanZCFIndexr-ZCFDataZCFStringZCFArrayZCFMutableArrayZCFDictionaryZCFErrorZCFTypeZCFTypeIDZ	CFTypeRefZCFAllocatorRefZOSStatusZ	CFDataRefZCFStringRefZ
CFArrayRefZCFMutableArrayRefZCFDictionaryRefZCFArrayCallBacksZCFDictionaryKeyCallBacksZCFDictionaryValueCallBacksZSecCertificateRefZSecExternalFormatZSecExternalItemTypeZSecIdentityRefZSecItemImportExportFlagsZ SecItemImportExportKeyParametersZSecKeychainRef�SSLProtocolZSSLCipherSuiteZ
SSLContextRefZSecTrustRefZSSLConnectionRefZSecTrustResultTypeZSecTrustOptionFlagsZSSLProtocolSideZSSLConnectionTypeZSSLSessionOptionZ
SecItemImport�argtypes�restypeZSecCertificateGetTypeIDZSecIdentityGetTypeIDZSecKeyGetTypeIDZSecCertificateCreateWithDataZSecCertificateCopyDataZSecCopyErrorMessageStringZ SecIdentityCreateWithCertificateZSecKeychainCreateZSecKeychainDeleteZSecPKCS12ImportZSSLReadFuncZSSLWriteFuncZ
SSLSetIOFuncsZSSLSetPeerIDZSSLSetCertificateZSSLSetCertificateAuthoritiesZSSLSetConnectionZSSLSetPeerDomainNameZSSLHandshakeZSSLReadZSSLWriteZSSLCloseZSSLGetNumberSupportedCiphersZSSLGetSupportedCiphersZSSLSetEnabledCiphersZSSLGetNumberEnabledCiphersZargtypeZSSLGetEnabledCiphersZSSLGetNegotiatedCipherZSSLGetNegotiatedProtocolVersionZSSLCopyPeerTrustZSecTrustSetAnchorCertificatesZ!SecTrustSetAnchorCertificatesOnlyZ	argstypesZSecTrustEvaluateZSecTrustGetCertificateCountZSecTrustGetCertificateAtIndexZSSLCreateContextZSSLSetSessionOptionZSSLSetProtocolVersionMinZSSLSetProtocolVersionMaxZSSLSetALPNProtocols�AttributeError�in_dllr"r#ZCFRetainZ	CFReleaseZCFGetTypeIDZCFStringCreateWithCStringZCFStringGetCStringPtrZCFStringGetCStringZCFDataCreateZCFDataGetLengthZCFDataGetBytePtrZCFDictionaryCreateZCFDictionaryGetValueZ
CFArrayCreateZCFArrayCreateMutableZCFArrayAppendValueZCFArrayGetCountZCFArrayGetValueAtIndexr$r%r&r'�objectr(r.rrrr�<module>s"8�����

��
��

����
���


��

���
����
	site-packages/pip/_vendor/urllib3/contrib/_securetransport/bindings.py000064400000042361150564745010022312 0ustar00"""
This module uses ctypes to bind a whole bunch of functions and constants from
SecureTransport. The goal here is to provide the low-level API to
SecureTransport. These are essentially the C-level functions and constants, and
they're pretty gross to work with.

This code is a bastardised version of the code found in Will Bond's oscrypto
library. An enormous debt is owed to him for blazing this trail for us. For
that reason, this code should be considered to be covered both by urllib3's
license and by oscrypto's:

    Copyright (c) 2015-2016 Will Bond <will@wbond.net>

    Permission is hereby granted, free of charge, to any person obtaining a
    copy of this software and associated documentation files (the "Software"),
    to deal in the Software without restriction, including without limitation
    the rights to use, copy, modify, merge, publish, distribute, sublicense,
    and/or sell copies of the Software, and to permit persons to whom the
    Software is furnished to do so, subject to the following conditions:

    The above copyright notice and this permission notice shall be included in
    all copies or substantial portions of the Software.

    THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
    IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
    FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
    AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
    LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
    FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
    DEALINGS IN THE SOFTWARE.
"""
from __future__ import absolute_import

import platform
from ctypes import (
    CDLL,
    CFUNCTYPE,
    POINTER,
    c_bool,
    c_byte,
    c_char_p,
    c_int32,
    c_long,
    c_size_t,
    c_uint32,
    c_ulong,
    c_void_p,
)
from ctypes.util import find_library

from pip._vendor.urllib3.packages.six import raise_from

if platform.system() != "Darwin":
    raise ImportError("Only macOS is supported")

version = platform.mac_ver()[0]
version_info = tuple(map(int, version.split(".")))
if version_info < (10, 8):
    raise OSError(
        "Only OS X 10.8 and newer are supported, not %s.%s"
        % (version_info[0], version_info[1])
    )


def load_cdll(name, macos10_16_path):
    """Loads a CDLL by name, falling back to known path on 10.16+"""
    try:
        # Big Sur is technically 11 but we use 10.16 due to the Big Sur
        # beta being labeled as 10.16.
        if version_info >= (10, 16):
            path = macos10_16_path
        else:
            path = find_library(name)
        if not path:
            raise OSError  # Caught and reraised as 'ImportError'
        return CDLL(path, use_errno=True)
    except OSError:
        raise_from(ImportError("The library %s failed to load" % name), None)


Security = load_cdll(
    "Security", "/System/Library/Frameworks/Security.framework/Security"
)
CoreFoundation = load_cdll(
    "CoreFoundation",
    "/System/Library/Frameworks/CoreFoundation.framework/CoreFoundation",
)


Boolean = c_bool
CFIndex = c_long
CFStringEncoding = c_uint32
CFData = c_void_p
CFString = c_void_p
CFArray = c_void_p
CFMutableArray = c_void_p
CFDictionary = c_void_p
CFError = c_void_p
CFType = c_void_p
CFTypeID = c_ulong

CFTypeRef = POINTER(CFType)
CFAllocatorRef = c_void_p

OSStatus = c_int32

CFDataRef = POINTER(CFData)
CFStringRef = POINTER(CFString)
CFArrayRef = POINTER(CFArray)
CFMutableArrayRef = POINTER(CFMutableArray)
CFDictionaryRef = POINTER(CFDictionary)
CFArrayCallBacks = c_void_p
CFDictionaryKeyCallBacks = c_void_p
CFDictionaryValueCallBacks = c_void_p

SecCertificateRef = POINTER(c_void_p)
SecExternalFormat = c_uint32
SecExternalItemType = c_uint32
SecIdentityRef = POINTER(c_void_p)
SecItemImportExportFlags = c_uint32
SecItemImportExportKeyParameters = c_void_p
SecKeychainRef = POINTER(c_void_p)
SSLProtocol = c_uint32
SSLCipherSuite = c_uint32
SSLContextRef = POINTER(c_void_p)
SecTrustRef = POINTER(c_void_p)
SSLConnectionRef = c_uint32
SecTrustResultType = c_uint32
SecTrustOptionFlags = c_uint32
SSLProtocolSide = c_uint32
SSLConnectionType = c_uint32
SSLSessionOption = c_uint32


try:
    Security.SecItemImport.argtypes = [
        CFDataRef,
        CFStringRef,
        POINTER(SecExternalFormat),
        POINTER(SecExternalItemType),
        SecItemImportExportFlags,
        POINTER(SecItemImportExportKeyParameters),
        SecKeychainRef,
        POINTER(CFArrayRef),
    ]
    Security.SecItemImport.restype = OSStatus

    Security.SecCertificateGetTypeID.argtypes = []
    Security.SecCertificateGetTypeID.restype = CFTypeID

    Security.SecIdentityGetTypeID.argtypes = []
    Security.SecIdentityGetTypeID.restype = CFTypeID

    Security.SecKeyGetTypeID.argtypes = []
    Security.SecKeyGetTypeID.restype = CFTypeID

    Security.SecCertificateCreateWithData.argtypes = [CFAllocatorRef, CFDataRef]
    Security.SecCertificateCreateWithData.restype = SecCertificateRef

    Security.SecCertificateCopyData.argtypes = [SecCertificateRef]
    Security.SecCertificateCopyData.restype = CFDataRef

    Security.SecCopyErrorMessageString.argtypes = [OSStatus, c_void_p]
    Security.SecCopyErrorMessageString.restype = CFStringRef

    Security.SecIdentityCreateWithCertificate.argtypes = [
        CFTypeRef,
        SecCertificateRef,
        POINTER(SecIdentityRef),
    ]
    Security.SecIdentityCreateWithCertificate.restype = OSStatus

    Security.SecKeychainCreate.argtypes = [
        c_char_p,
        c_uint32,
        c_void_p,
        Boolean,
        c_void_p,
        POINTER(SecKeychainRef),
    ]
    Security.SecKeychainCreate.restype = OSStatus

    Security.SecKeychainDelete.argtypes = [SecKeychainRef]
    Security.SecKeychainDelete.restype = OSStatus

    Security.SecPKCS12Import.argtypes = [
        CFDataRef,
        CFDictionaryRef,
        POINTER(CFArrayRef),
    ]
    Security.SecPKCS12Import.restype = OSStatus

    SSLReadFunc = CFUNCTYPE(OSStatus, SSLConnectionRef, c_void_p, POINTER(c_size_t))
    SSLWriteFunc = CFUNCTYPE(
        OSStatus, SSLConnectionRef, POINTER(c_byte), POINTER(c_size_t)
    )

    Security.SSLSetIOFuncs.argtypes = [SSLContextRef, SSLReadFunc, SSLWriteFunc]
    Security.SSLSetIOFuncs.restype = OSStatus

    Security.SSLSetPeerID.argtypes = [SSLContextRef, c_char_p, c_size_t]
    Security.SSLSetPeerID.restype = OSStatus

    Security.SSLSetCertificate.argtypes = [SSLContextRef, CFArrayRef]
    Security.SSLSetCertificate.restype = OSStatus

    Security.SSLSetCertificateAuthorities.argtypes = [SSLContextRef, CFTypeRef, Boolean]
    Security.SSLSetCertificateAuthorities.restype = OSStatus

    Security.SSLSetConnection.argtypes = [SSLContextRef, SSLConnectionRef]
    Security.SSLSetConnection.restype = OSStatus

    Security.SSLSetPeerDomainName.argtypes = [SSLContextRef, c_char_p, c_size_t]
    Security.SSLSetPeerDomainName.restype = OSStatus

    Security.SSLHandshake.argtypes = [SSLContextRef]
    Security.SSLHandshake.restype = OSStatus

    Security.SSLRead.argtypes = [SSLContextRef, c_char_p, c_size_t, POINTER(c_size_t)]
    Security.SSLRead.restype = OSStatus

    Security.SSLWrite.argtypes = [SSLContextRef, c_char_p, c_size_t, POINTER(c_size_t)]
    Security.SSLWrite.restype = OSStatus

    Security.SSLClose.argtypes = [SSLContextRef]
    Security.SSLClose.restype = OSStatus

    Security.SSLGetNumberSupportedCiphers.argtypes = [SSLContextRef, POINTER(c_size_t)]
    Security.SSLGetNumberSupportedCiphers.restype = OSStatus

    Security.SSLGetSupportedCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        POINTER(c_size_t),
    ]
    Security.SSLGetSupportedCiphers.restype = OSStatus

    Security.SSLSetEnabledCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        c_size_t,
    ]
    Security.SSLSetEnabledCiphers.restype = OSStatus

    Security.SSLGetNumberEnabledCiphers.argtype = [SSLContextRef, POINTER(c_size_t)]
    Security.SSLGetNumberEnabledCiphers.restype = OSStatus

    Security.SSLGetEnabledCiphers.argtypes = [
        SSLContextRef,
        POINTER(SSLCipherSuite),
        POINTER(c_size_t),
    ]
    Security.SSLGetEnabledCiphers.restype = OSStatus

    Security.SSLGetNegotiatedCipher.argtypes = [SSLContextRef, POINTER(SSLCipherSuite)]
    Security.SSLGetNegotiatedCipher.restype = OSStatus

    Security.SSLGetNegotiatedProtocolVersion.argtypes = [
        SSLContextRef,
        POINTER(SSLProtocol),
    ]
    Security.SSLGetNegotiatedProtocolVersion.restype = OSStatus

    Security.SSLCopyPeerTrust.argtypes = [SSLContextRef, POINTER(SecTrustRef)]
    Security.SSLCopyPeerTrust.restype = OSStatus

    Security.SecTrustSetAnchorCertificates.argtypes = [SecTrustRef, CFArrayRef]
    Security.SecTrustSetAnchorCertificates.restype = OSStatus

    Security.SecTrustSetAnchorCertificatesOnly.argstypes = [SecTrustRef, Boolean]
    Security.SecTrustSetAnchorCertificatesOnly.restype = OSStatus

    Security.SecTrustEvaluate.argtypes = [SecTrustRef, POINTER(SecTrustResultType)]
    Security.SecTrustEvaluate.restype = OSStatus

    Security.SecTrustGetCertificateCount.argtypes = [SecTrustRef]
    Security.SecTrustGetCertificateCount.restype = CFIndex

    Security.SecTrustGetCertificateAtIndex.argtypes = [SecTrustRef, CFIndex]
    Security.SecTrustGetCertificateAtIndex.restype = SecCertificateRef

    Security.SSLCreateContext.argtypes = [
        CFAllocatorRef,
        SSLProtocolSide,
        SSLConnectionType,
    ]
    Security.SSLCreateContext.restype = SSLContextRef

    Security.SSLSetSessionOption.argtypes = [SSLContextRef, SSLSessionOption, Boolean]
    Security.SSLSetSessionOption.restype = OSStatus

    Security.SSLSetProtocolVersionMin.argtypes = [SSLContextRef, SSLProtocol]
    Security.SSLSetProtocolVersionMin.restype = OSStatus

    Security.SSLSetProtocolVersionMax.argtypes = [SSLContextRef, SSLProtocol]
    Security.SSLSetProtocolVersionMax.restype = OSStatus

    try:
        Security.SSLSetALPNProtocols.argtypes = [SSLContextRef, CFArrayRef]
        Security.SSLSetALPNProtocols.restype = OSStatus
    except AttributeError:
        # Supported only in 10.12+
        pass

    Security.SecCopyErrorMessageString.argtypes = [OSStatus, c_void_p]
    Security.SecCopyErrorMessageString.restype = CFStringRef

    Security.SSLReadFunc = SSLReadFunc
    Security.SSLWriteFunc = SSLWriteFunc
    Security.SSLContextRef = SSLContextRef
    Security.SSLProtocol = SSLProtocol
    Security.SSLCipherSuite = SSLCipherSuite
    Security.SecIdentityRef = SecIdentityRef
    Security.SecKeychainRef = SecKeychainRef
    Security.SecTrustRef = SecTrustRef
    Security.SecTrustResultType = SecTrustResultType
    Security.SecExternalFormat = SecExternalFormat
    Security.OSStatus = OSStatus

    Security.kSecImportExportPassphrase = CFStringRef.in_dll(
        Security, "kSecImportExportPassphrase"
    )
    Security.kSecImportItemIdentity = CFStringRef.in_dll(
        Security, "kSecImportItemIdentity"
    )

    # CoreFoundation time!
    CoreFoundation.CFRetain.argtypes = [CFTypeRef]
    CoreFoundation.CFRetain.restype = CFTypeRef

    CoreFoundation.CFRelease.argtypes = [CFTypeRef]
    CoreFoundation.CFRelease.restype = None

    CoreFoundation.CFGetTypeID.argtypes = [CFTypeRef]
    CoreFoundation.CFGetTypeID.restype = CFTypeID

    CoreFoundation.CFStringCreateWithCString.argtypes = [
        CFAllocatorRef,
        c_char_p,
        CFStringEncoding,
    ]
    CoreFoundation.CFStringCreateWithCString.restype = CFStringRef

    CoreFoundation.CFStringGetCStringPtr.argtypes = [CFStringRef, CFStringEncoding]
    CoreFoundation.CFStringGetCStringPtr.restype = c_char_p

    CoreFoundation.CFStringGetCString.argtypes = [
        CFStringRef,
        c_char_p,
        CFIndex,
        CFStringEncoding,
    ]
    CoreFoundation.CFStringGetCString.restype = c_bool

    CoreFoundation.CFDataCreate.argtypes = [CFAllocatorRef, c_char_p, CFIndex]
    CoreFoundation.CFDataCreate.restype = CFDataRef

    CoreFoundation.CFDataGetLength.argtypes = [CFDataRef]
    CoreFoundation.CFDataGetLength.restype = CFIndex

    CoreFoundation.CFDataGetBytePtr.argtypes = [CFDataRef]
    CoreFoundation.CFDataGetBytePtr.restype = c_void_p

    CoreFoundation.CFDictionaryCreate.argtypes = [
        CFAllocatorRef,
        POINTER(CFTypeRef),
        POINTER(CFTypeRef),
        CFIndex,
        CFDictionaryKeyCallBacks,
        CFDictionaryValueCallBacks,
    ]
    CoreFoundation.CFDictionaryCreate.restype = CFDictionaryRef

    CoreFoundation.CFDictionaryGetValue.argtypes = [CFDictionaryRef, CFTypeRef]
    CoreFoundation.CFDictionaryGetValue.restype = CFTypeRef

    CoreFoundation.CFArrayCreate.argtypes = [
        CFAllocatorRef,
        POINTER(CFTypeRef),
        CFIndex,
        CFArrayCallBacks,
    ]
    CoreFoundation.CFArrayCreate.restype = CFArrayRef

    CoreFoundation.CFArrayCreateMutable.argtypes = [
        CFAllocatorRef,
        CFIndex,
        CFArrayCallBacks,
    ]
    CoreFoundation.CFArrayCreateMutable.restype = CFMutableArrayRef

    CoreFoundation.CFArrayAppendValue.argtypes = [CFMutableArrayRef, c_void_p]
    CoreFoundation.CFArrayAppendValue.restype = None

    CoreFoundation.CFArrayGetCount.argtypes = [CFArrayRef]
    CoreFoundation.CFArrayGetCount.restype = CFIndex

    CoreFoundation.CFArrayGetValueAtIndex.argtypes = [CFArrayRef, CFIndex]
    CoreFoundation.CFArrayGetValueAtIndex.restype = c_void_p

    CoreFoundation.kCFAllocatorDefault = CFAllocatorRef.in_dll(
        CoreFoundation, "kCFAllocatorDefault"
    )
    CoreFoundation.kCFTypeArrayCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeArrayCallBacks"
    )
    CoreFoundation.kCFTypeDictionaryKeyCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeDictionaryKeyCallBacks"
    )
    CoreFoundation.kCFTypeDictionaryValueCallBacks = c_void_p.in_dll(
        CoreFoundation, "kCFTypeDictionaryValueCallBacks"
    )

    CoreFoundation.CFTypeRef = CFTypeRef
    CoreFoundation.CFArrayRef = CFArrayRef
    CoreFoundation.CFStringRef = CFStringRef
    CoreFoundation.CFDictionaryRef = CFDictionaryRef

except (AttributeError):
    raise ImportError("Error initializing ctypes")


class CFConst(object):
    """
    A class object that acts as essentially a namespace for CoreFoundation
    constants.
    """

    kCFStringEncodingUTF8 = CFStringEncoding(0x08000100)


class SecurityConst(object):
    """
    A class object that acts as essentially a namespace for Security constants.
    """

    kSSLSessionOptionBreakOnServerAuth = 0

    kSSLProtocol2 = 1
    kSSLProtocol3 = 2
    kTLSProtocol1 = 4
    kTLSProtocol11 = 7
    kTLSProtocol12 = 8
    # SecureTransport does not support TLS 1.3 even if there's a constant for it
    kTLSProtocol13 = 10
    kTLSProtocolMaxSupported = 999

    kSSLClientSide = 1
    kSSLStreamType = 0

    kSecFormatPEMSequence = 10

    kSecTrustResultInvalid = 0
    kSecTrustResultProceed = 1
    # This gap is present on purpose: this was kSecTrustResultConfirm, which
    # is deprecated.
    kSecTrustResultDeny = 3
    kSecTrustResultUnspecified = 4
    kSecTrustResultRecoverableTrustFailure = 5
    kSecTrustResultFatalTrustFailure = 6
    kSecTrustResultOtherError = 7

    errSSLProtocol = -9800
    errSSLWouldBlock = -9803
    errSSLClosedGraceful = -9805
    errSSLClosedNoNotify = -9816
    errSSLClosedAbort = -9806

    errSSLXCertChainInvalid = -9807
    errSSLCrypto = -9809
    errSSLInternal = -9810
    errSSLCertExpired = -9814
    errSSLCertNotYetValid = -9815
    errSSLUnknownRootCert = -9812
    errSSLNoRootCert = -9813
    errSSLHostNameMismatch = -9843
    errSSLPeerHandshakeFail = -9824
    errSSLPeerUserCancelled = -9839
    errSSLWeakPeerEphemeralDHKey = -9850
    errSSLServerAuthCompleted = -9841
    errSSLRecordOverflow = -9847

    errSecVerifyFailed = -67808
    errSecNoTrustSettings = -25263
    errSecItemNotFound = -25300
    errSecInvalidTrustSettings = -25262

    # Cipher suites. We only pick the ones our default cipher string allows.
    # Source: https://developer.apple.com/documentation/security/1550981-ssl_cipher_suite_values
    TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384 = 0xC02C
    TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 = 0xC030
    TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 = 0xC02B
    TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256 = 0xC02F
    TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256 = 0xCCA9
    TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 = 0xCCA8
    TLS_DHE_RSA_WITH_AES_256_GCM_SHA384 = 0x009F
    TLS_DHE_RSA_WITH_AES_128_GCM_SHA256 = 0x009E
    TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA384 = 0xC024
    TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA384 = 0xC028
    TLS_ECDHE_ECDSA_WITH_AES_256_CBC_SHA = 0xC00A
    TLS_ECDHE_RSA_WITH_AES_256_CBC_SHA = 0xC014
    TLS_DHE_RSA_WITH_AES_256_CBC_SHA256 = 0x006B
    TLS_DHE_RSA_WITH_AES_256_CBC_SHA = 0x0039
    TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256 = 0xC023
    TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256 = 0xC027
    TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA = 0xC009
    TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA = 0xC013
    TLS_DHE_RSA_WITH_AES_128_CBC_SHA256 = 0x0067
    TLS_DHE_RSA_WITH_AES_128_CBC_SHA = 0x0033
    TLS_RSA_WITH_AES_256_GCM_SHA384 = 0x009D
    TLS_RSA_WITH_AES_128_GCM_SHA256 = 0x009C
    TLS_RSA_WITH_AES_256_CBC_SHA256 = 0x003D
    TLS_RSA_WITH_AES_128_CBC_SHA256 = 0x003C
    TLS_RSA_WITH_AES_256_CBC_SHA = 0x0035
    TLS_RSA_WITH_AES_128_CBC_SHA = 0x002F
    TLS_AES_128_GCM_SHA256 = 0x1301
    TLS_AES_256_GCM_SHA384 = 0x1302
    TLS_AES_128_CCM_8_SHA256 = 0x1305
    TLS_AES_128_CCM_SHA256 = 0x1304
site-packages/pip/_vendor/urllib3/contrib/_securetransport/__init__.py000064400000000000150564745010022234 0ustar00site-packages/pip/_vendor/urllib3/contrib/_securetransport/low_level.py000064400000033142150564745010022502 0ustar00"""
Low-level helpers for the SecureTransport bindings.

These are Python functions that are not directly related to the high-level APIs
but are necessary to get them to work. They include a whole bunch of low-level
CoreFoundation messing about and memory management. The concerns in this module
are almost entirely about trying to avoid memory leaks and providing
appropriate and useful assistance to the higher-level code.
"""
import base64
import ctypes
import itertools
import os
import re
import ssl
import struct
import tempfile

from .bindings import CFConst, CoreFoundation, Security

# This regular expression is used to grab PEM data out of a PEM bundle.
_PEM_CERTS_RE = re.compile(
    b"-----BEGIN CERTIFICATE-----\n(.*?)\n-----END CERTIFICATE-----", re.DOTALL
)


def _cf_data_from_bytes(bytestring):
    """
    Given a bytestring, create a CFData object from it. This CFData object must
    be CFReleased by the caller.
    """
    return CoreFoundation.CFDataCreate(
        CoreFoundation.kCFAllocatorDefault, bytestring, len(bytestring)
    )


def _cf_dictionary_from_tuples(tuples):
    """
    Given a list of Python tuples, create an associated CFDictionary.
    """
    dictionary_size = len(tuples)

    # We need to get the dictionary keys and values out in the same order.
    keys = (t[0] for t in tuples)
    values = (t[1] for t in tuples)
    cf_keys = (CoreFoundation.CFTypeRef * dictionary_size)(*keys)
    cf_values = (CoreFoundation.CFTypeRef * dictionary_size)(*values)

    return CoreFoundation.CFDictionaryCreate(
        CoreFoundation.kCFAllocatorDefault,
        cf_keys,
        cf_values,
        dictionary_size,
        CoreFoundation.kCFTypeDictionaryKeyCallBacks,
        CoreFoundation.kCFTypeDictionaryValueCallBacks,
    )


def _cfstr(py_bstr):
    """
    Given a Python binary data, create a CFString.
    The string must be CFReleased by the caller.
    """
    c_str = ctypes.c_char_p(py_bstr)
    cf_str = CoreFoundation.CFStringCreateWithCString(
        CoreFoundation.kCFAllocatorDefault,
        c_str,
        CFConst.kCFStringEncodingUTF8,
    )
    return cf_str


def _create_cfstring_array(lst):
    """
    Given a list of Python binary data, create an associated CFMutableArray.
    The array must be CFReleased by the caller.

    Raises an ssl.SSLError on failure.
    """
    cf_arr = None
    try:
        cf_arr = CoreFoundation.CFArrayCreateMutable(
            CoreFoundation.kCFAllocatorDefault,
            0,
            ctypes.byref(CoreFoundation.kCFTypeArrayCallBacks),
        )
        if not cf_arr:
            raise MemoryError("Unable to allocate memory!")
        for item in lst:
            cf_str = _cfstr(item)
            if not cf_str:
                raise MemoryError("Unable to allocate memory!")
            try:
                CoreFoundation.CFArrayAppendValue(cf_arr, cf_str)
            finally:
                CoreFoundation.CFRelease(cf_str)
    except BaseException as e:
        if cf_arr:
            CoreFoundation.CFRelease(cf_arr)
        raise ssl.SSLError("Unable to allocate array: %s" % (e,))
    return cf_arr


def _cf_string_to_unicode(value):
    """
    Creates a Unicode string from a CFString object. Used entirely for error
    reporting.

    Yes, it annoys me quite a lot that this function is this complex.
    """
    value_as_void_p = ctypes.cast(value, ctypes.POINTER(ctypes.c_void_p))

    string = CoreFoundation.CFStringGetCStringPtr(
        value_as_void_p, CFConst.kCFStringEncodingUTF8
    )
    if string is None:
        buffer = ctypes.create_string_buffer(1024)
        result = CoreFoundation.CFStringGetCString(
            value_as_void_p, buffer, 1024, CFConst.kCFStringEncodingUTF8
        )
        if not result:
            raise OSError("Error copying C string from CFStringRef")
        string = buffer.value
    if string is not None:
        string = string.decode("utf-8")
    return string


def _assert_no_error(error, exception_class=None):
    """
    Checks the return code and throws an exception if there is an error to
    report
    """
    if error == 0:
        return

    cf_error_string = Security.SecCopyErrorMessageString(error, None)
    output = _cf_string_to_unicode(cf_error_string)
    CoreFoundation.CFRelease(cf_error_string)

    if output is None or output == u"":
        output = u"OSStatus %s" % error

    if exception_class is None:
        exception_class = ssl.SSLError

    raise exception_class(output)


def _cert_array_from_pem(pem_bundle):
    """
    Given a bundle of certs in PEM format, turns them into a CFArray of certs
    that can be used to validate a cert chain.
    """
    # Normalize the PEM bundle's line endings.
    pem_bundle = pem_bundle.replace(b"\r\n", b"\n")

    der_certs = [
        base64.b64decode(match.group(1)) for match in _PEM_CERTS_RE.finditer(pem_bundle)
    ]
    if not der_certs:
        raise ssl.SSLError("No root certificates specified")

    cert_array = CoreFoundation.CFArrayCreateMutable(
        CoreFoundation.kCFAllocatorDefault,
        0,
        ctypes.byref(CoreFoundation.kCFTypeArrayCallBacks),
    )
    if not cert_array:
        raise ssl.SSLError("Unable to allocate memory!")

    try:
        for der_bytes in der_certs:
            certdata = _cf_data_from_bytes(der_bytes)
            if not certdata:
                raise ssl.SSLError("Unable to allocate memory!")
            cert = Security.SecCertificateCreateWithData(
                CoreFoundation.kCFAllocatorDefault, certdata
            )
            CoreFoundation.CFRelease(certdata)
            if not cert:
                raise ssl.SSLError("Unable to build cert object!")

            CoreFoundation.CFArrayAppendValue(cert_array, cert)
            CoreFoundation.CFRelease(cert)
    except Exception:
        # We need to free the array before the exception bubbles further.
        # We only want to do that if an error occurs: otherwise, the caller
        # should free.
        CoreFoundation.CFRelease(cert_array)
        raise

    return cert_array


def _is_cert(item):
    """
    Returns True if a given CFTypeRef is a certificate.
    """
    expected = Security.SecCertificateGetTypeID()
    return CoreFoundation.CFGetTypeID(item) == expected


def _is_identity(item):
    """
    Returns True if a given CFTypeRef is an identity.
    """
    expected = Security.SecIdentityGetTypeID()
    return CoreFoundation.CFGetTypeID(item) == expected


def _temporary_keychain():
    """
    This function creates a temporary Mac keychain that we can use to work with
    credentials. This keychain uses a one-time password and a temporary file to
    store the data. We expect to have one keychain per socket. The returned
    SecKeychainRef must be freed by the caller, including calling
    SecKeychainDelete.

    Returns a tuple of the SecKeychainRef and the path to the temporary
    directory that contains it.
    """
    # Unfortunately, SecKeychainCreate requires a path to a keychain. This
    # means we cannot use mkstemp to use a generic temporary file. Instead,
    # we're going to create a temporary directory and a filename to use there.
    # This filename will be 8 random bytes expanded into base64. We also need
    # some random bytes to password-protect the keychain we're creating, so we
    # ask for 40 random bytes.
    random_bytes = os.urandom(40)
    filename = base64.b16encode(random_bytes[:8]).decode("utf-8")
    password = base64.b16encode(random_bytes[8:])  # Must be valid UTF-8
    tempdirectory = tempfile.mkdtemp()

    keychain_path = os.path.join(tempdirectory, filename).encode("utf-8")

    # We now want to create the keychain itself.
    keychain = Security.SecKeychainRef()
    status = Security.SecKeychainCreate(
        keychain_path, len(password), password, False, None, ctypes.byref(keychain)
    )
    _assert_no_error(status)

    # Having created the keychain, we want to pass it off to the caller.
    return keychain, tempdirectory


def _load_items_from_file(keychain, path):
    """
    Given a single file, loads all the trust objects from it into arrays and
    the keychain.
    Returns a tuple of lists: the first list is a list of identities, the
    second a list of certs.
    """
    certificates = []
    identities = []
    result_array = None

    with open(path, "rb") as f:
        raw_filedata = f.read()

    try:
        filedata = CoreFoundation.CFDataCreate(
            CoreFoundation.kCFAllocatorDefault, raw_filedata, len(raw_filedata)
        )
        result_array = CoreFoundation.CFArrayRef()
        result = Security.SecItemImport(
            filedata,  # cert data
            None,  # Filename, leaving it out for now
            None,  # What the type of the file is, we don't care
            None,  # what's in the file, we don't care
            0,  # import flags
            None,  # key params, can include passphrase in the future
            keychain,  # The keychain to insert into
            ctypes.byref(result_array),  # Results
        )
        _assert_no_error(result)

        # A CFArray is not very useful to us as an intermediary
        # representation, so we are going to extract the objects we want
        # and then free the array. We don't need to keep hold of keys: the
        # keychain already has them!
        result_count = CoreFoundation.CFArrayGetCount(result_array)
        for index in range(result_count):
            item = CoreFoundation.CFArrayGetValueAtIndex(result_array, index)
            item = ctypes.cast(item, CoreFoundation.CFTypeRef)

            if _is_cert(item):
                CoreFoundation.CFRetain(item)
                certificates.append(item)
            elif _is_identity(item):
                CoreFoundation.CFRetain(item)
                identities.append(item)
    finally:
        if result_array:
            CoreFoundation.CFRelease(result_array)

        CoreFoundation.CFRelease(filedata)

    return (identities, certificates)


def _load_client_cert_chain(keychain, *paths):
    """
    Load certificates and maybe keys from a number of files. Has the end goal
    of returning a CFArray containing one SecIdentityRef, and then zero or more
    SecCertificateRef objects, suitable for use as a client certificate trust
    chain.
    """
    # Ok, the strategy.
    #
    # This relies on knowing that macOS will not give you a SecIdentityRef
    # unless you have imported a key into a keychain. This is a somewhat
    # artificial limitation of macOS (for example, it doesn't necessarily
    # affect iOS), but there is nothing inside Security.framework that lets you
    # get a SecIdentityRef without having a key in a keychain.
    #
    # So the policy here is we take all the files and iterate them in order.
    # Each one will use SecItemImport to have one or more objects loaded from
    # it. We will also point at a keychain that macOS can use to work with the
    # private key.
    #
    # Once we have all the objects, we'll check what we actually have. If we
    # already have a SecIdentityRef in hand, fab: we'll use that. Otherwise,
    # we'll take the first certificate (which we assume to be our leaf) and
    # ask the keychain to give us a SecIdentityRef with that cert's associated
    # key.
    #
    # We'll then return a CFArray containing the trust chain: one
    # SecIdentityRef and then zero-or-more SecCertificateRef objects. The
    # responsibility for freeing this CFArray will be with the caller. This
    # CFArray must remain alive for the entire connection, so in practice it
    # will be stored with a single SSLSocket, along with the reference to the
    # keychain.
    certificates = []
    identities = []

    # Filter out bad paths.
    paths = (path for path in paths if path)

    try:
        for file_path in paths:
            new_identities, new_certs = _load_items_from_file(keychain, file_path)
            identities.extend(new_identities)
            certificates.extend(new_certs)

        # Ok, we have everything. The question is: do we have an identity? If
        # not, we want to grab one from the first cert we have.
        if not identities:
            new_identity = Security.SecIdentityRef()
            status = Security.SecIdentityCreateWithCertificate(
                keychain, certificates[0], ctypes.byref(new_identity)
            )
            _assert_no_error(status)
            identities.append(new_identity)

            # We now want to release the original certificate, as we no longer
            # need it.
            CoreFoundation.CFRelease(certificates.pop(0))

        # We now need to build a new CFArray that holds the trust chain.
        trust_chain = CoreFoundation.CFArrayCreateMutable(
            CoreFoundation.kCFAllocatorDefault,
            0,
            ctypes.byref(CoreFoundation.kCFTypeArrayCallBacks),
        )
        for item in itertools.chain(identities, certificates):
            # ArrayAppendValue does a CFRetain on the item. That's fine,
            # because the finally block will release our other refs to them.
            CoreFoundation.CFArrayAppendValue(trust_chain, item)

        return trust_chain
    finally:
        for obj in itertools.chain(identities, certificates):
            CoreFoundation.CFRelease(obj)


TLS_PROTOCOL_VERSIONS = {
    "SSLv2": (0, 2),
    "SSLv3": (3, 0),
    "TLSv1": (3, 1),
    "TLSv1.1": (3, 2),
    "TLSv1.2": (3, 3),
}


def _build_tls_unknown_ca_alert(version):
    """
    Builds a TLS alert record for an unknown CA.
    """
    ver_maj, ver_min = TLS_PROTOCOL_VERSIONS[version]
    severity_fatal = 0x02
    description_unknown_ca = 0x30
    msg = struct.pack(">BB", severity_fatal, description_unknown_ca)
    msg_len = len(msg)
    record_type_alert = 0x15
    record = struct.pack(">BBBH", record_type_alert, ver_maj, ver_min, msg_len) + msg
    return record
site-packages/pip/_vendor/urllib3/fields.py000064400000020603150564745010014714 0ustar00from __future__ import absolute_import

import email.utils
import mimetypes
import re

from .packages import six


def guess_content_type(filename, default="application/octet-stream"):
    """
    Guess the "Content-Type" of a file.

    :param filename:
        The filename to guess the "Content-Type" of using :mod:`mimetypes`.
    :param default:
        If no "Content-Type" can be guessed, default to `default`.
    """
    if filename:
        return mimetypes.guess_type(filename)[0] or default
    return default


def format_header_param_rfc2231(name, value):
    """
    Helper function to format and quote a single header parameter using the
    strategy defined in RFC 2231.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows
    `RFC 2388 Section 4.4 <https://tools.ietf.org/html/rfc2388#section-4.4>`_.

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        An RFC-2231-formatted unicode string.
    """
    if isinstance(value, six.binary_type):
        value = value.decode("utf-8")

    if not any(ch in value for ch in '"\\\r\n'):
        result = u'%s="%s"' % (name, value)
        try:
            result.encode("ascii")
        except (UnicodeEncodeError, UnicodeDecodeError):
            pass
        else:
            return result

    if six.PY2:  # Python 2:
        value = value.encode("utf-8")

    # encode_rfc2231 accepts an encoded string and returns an ascii-encoded
    # string in Python 2 but accepts and returns unicode strings in Python 3
    value = email.utils.encode_rfc2231(value, "utf-8")
    value = "%s*=%s" % (name, value)

    if six.PY2:  # Python 2:
        value = value.decode("utf-8")

    return value


_HTML5_REPLACEMENTS = {
    u"\u0022": u"%22",
    # Replace "\" with "\\".
    u"\u005C": u"\u005C\u005C",
}

# All control characters from 0x00 to 0x1F *except* 0x1B.
_HTML5_REPLACEMENTS.update(
    {
        six.unichr(cc): u"%{:02X}".format(cc)
        for cc in range(0x00, 0x1F + 1)
        if cc not in (0x1B,)
    }
)


def _replace_multiple(value, needles_and_replacements):
    def replacer(match):
        return needles_and_replacements[match.group(0)]

    pattern = re.compile(
        r"|".join([re.escape(needle) for needle in needles_and_replacements.keys()])
    )

    result = pattern.sub(replacer, value)

    return result


def format_header_param_html5(name, value):
    """
    Helper function to format and quote a single header parameter using the
    HTML5 strategy.

    Particularly useful for header parameters which might contain
    non-ASCII values, like file names. This follows the `HTML5 Working Draft
    Section 4.10.22.7`_ and matches the behavior of curl and modern browsers.

    .. _HTML5 Working Draft Section 4.10.22.7:
        https://w3c.github.io/html/sec-forms.html#multipart-form-data

    :param name:
        The name of the parameter, a string expected to be ASCII only.
    :param value:
        The value of the parameter, provided as ``bytes`` or `str``.
    :ret:
        A unicode string, stripped of troublesome characters.
    """
    if isinstance(value, six.binary_type):
        value = value.decode("utf-8")

    value = _replace_multiple(value, _HTML5_REPLACEMENTS)

    return u'%s="%s"' % (name, value)


# For backwards-compatibility.
format_header_param = format_header_param_html5


class RequestField(object):
    """
    A data container for request body parameters.

    :param name:
        The name of this request field. Must be unicode.
    :param data:
        The data/value body.
    :param filename:
        An optional filename of the request field. Must be unicode.
    :param headers:
        An optional dict-like object of headers to initially use for the field.
    :param header_formatter:
        An optional callable that is used to encode and format the headers. By
        default, this is :func:`format_header_param_html5`.
    """

    def __init__(
        self,
        name,
        data,
        filename=None,
        headers=None,
        header_formatter=format_header_param_html5,
    ):
        self._name = name
        self._filename = filename
        self.data = data
        self.headers = {}
        if headers:
            self.headers = dict(headers)
        self.header_formatter = header_formatter

    @classmethod
    def from_tuples(cls, fieldname, value, header_formatter=format_header_param_html5):
        """
        A :class:`~urllib3.fields.RequestField` factory from old-style tuple parameters.

        Supports constructing :class:`~urllib3.fields.RequestField` from
        parameter of key/value strings AND key/filetuple. A filetuple is a
        (filename, data, MIME type) tuple where the MIME type is optional.
        For example::

            'foo': 'bar',
            'fakefile': ('foofile.txt', 'contents of foofile'),
            'realfile': ('barfile.txt', open('realfile').read()),
            'typedfile': ('bazfile.bin', open('bazfile').read(), 'image/jpeg'),
            'nonamefile': 'contents of nonamefile field',

        Field names and filenames must be unicode.
        """
        if isinstance(value, tuple):
            if len(value) == 3:
                filename, data, content_type = value
            else:
                filename, data = value
                content_type = guess_content_type(filename)
        else:
            filename = None
            content_type = None
            data = value

        request_param = cls(
            fieldname, data, filename=filename, header_formatter=header_formatter
        )
        request_param.make_multipart(content_type=content_type)

        return request_param

    def _render_part(self, name, value):
        """
        Overridable helper function to format a single header parameter. By
        default, this calls ``self.header_formatter``.

        :param name:
            The name of the parameter, a string expected to be ASCII only.
        :param value:
            The value of the parameter, provided as a unicode string.
        """

        return self.header_formatter(name, value)

    def _render_parts(self, header_parts):
        """
        Helper function to format and quote a single header.

        Useful for single headers that are composed of multiple items. E.g.,
        'Content-Disposition' fields.

        :param header_parts:
            A sequence of (k, v) tuples or a :class:`dict` of (k, v) to format
            as `k1="v1"; k2="v2"; ...`.
        """
        parts = []
        iterable = header_parts
        if isinstance(header_parts, dict):
            iterable = header_parts.items()

        for name, value in iterable:
            if value is not None:
                parts.append(self._render_part(name, value))

        return u"; ".join(parts)

    def render_headers(self):
        """
        Renders the headers for this request field.
        """
        lines = []

        sort_keys = ["Content-Disposition", "Content-Type", "Content-Location"]
        for sort_key in sort_keys:
            if self.headers.get(sort_key, False):
                lines.append(u"%s: %s" % (sort_key, self.headers[sort_key]))

        for header_name, header_value in self.headers.items():
            if header_name not in sort_keys:
                if header_value:
                    lines.append(u"%s: %s" % (header_name, header_value))

        lines.append(u"\r\n")
        return u"\r\n".join(lines)

    def make_multipart(
        self, content_disposition=None, content_type=None, content_location=None
    ):
        """
        Makes this request field into a multipart request field.

        This method overrides "Content-Disposition", "Content-Type" and
        "Content-Location" headers to the request parameter.

        :param content_type:
            The 'Content-Type' of the request body.
        :param content_location:
            The 'Content-Location' of the request body.

        """
        self.headers["Content-Disposition"] = content_disposition or u"form-data"
        self.headers["Content-Disposition"] += u"; ".join(
            [
                u"",
                self._render_parts(
                    ((u"name", self._name), (u"filename", self._filename))
                ),
            ]
        )
        self.headers["Content-Type"] = content_type
        self.headers["Content-Location"] = content_location
site-packages/pip/_vendor/urllib3/__init__.py000064400000005313150564745010015206 0ustar00"""
Python HTTP library with thread-safe connection pooling, file post support, user friendly, and more
"""
from __future__ import absolute_import

# Set default logging handler to avoid "No handler found" warnings.
import logging
import warnings
from logging import NullHandler

from . import exceptions
from ._version import __version__
from .connectionpool import HTTPConnectionPool, HTTPSConnectionPool, connection_from_url
from .filepost import encode_multipart_formdata
from .poolmanager import PoolManager, ProxyManager, proxy_from_url
from .response import HTTPResponse
from .util.request import make_headers
from .util.retry import Retry
from .util.timeout import Timeout
from .util.url import get_host

__author__ = "Andrey Petrov (andrey.petrov@shazow.net)"
__license__ = "MIT"
__version__ = __version__

__all__ = (
    "HTTPConnectionPool",
    "HTTPSConnectionPool",
    "PoolManager",
    "ProxyManager",
    "HTTPResponse",
    "Retry",
    "Timeout",
    "add_stderr_logger",
    "connection_from_url",
    "disable_warnings",
    "encode_multipart_formdata",
    "get_host",
    "make_headers",
    "proxy_from_url",
)

logging.getLogger(__name__).addHandler(NullHandler())


def add_stderr_logger(level=logging.DEBUG):
    """
    Helper for quickly adding a StreamHandler to the logger. Useful for
    debugging.

    Returns the handler after adding it.
    """
    # This method needs to be in this __init__.py to get the __name__ correct
    # even if urllib3 is vendored within another package.
    logger = logging.getLogger(__name__)
    handler = logging.StreamHandler()
    handler.setFormatter(logging.Formatter("%(asctime)s %(levelname)s %(message)s"))
    logger.addHandler(handler)
    logger.setLevel(level)
    logger.debug("Added a stderr logging handler to logger: %s", __name__)
    return handler


# ... Clean up.
del NullHandler


# All warning filters *must* be appended unless you're really certain that they
# shouldn't be: otherwise, it's very hard for users to use most Python
# mechanisms to silence them.
# SecurityWarning's always go off by default.
warnings.simplefilter("always", exceptions.SecurityWarning, append=True)
# SubjectAltNameWarning's should go off once per host
warnings.simplefilter("default", exceptions.SubjectAltNameWarning, append=True)
# InsecurePlatformWarning's don't vary between requests, so we keep it default.
warnings.simplefilter("default", exceptions.InsecurePlatformWarning, append=True)
# SNIMissingWarnings should go off only once.
warnings.simplefilter("default", exceptions.SNIMissingWarning, append=True)


def disable_warnings(category=exceptions.HTTPWarning):
    """
    Helper for quickly disabling all urllib3 warnings.
    """
    warnings.simplefilter("ignore", category)
site-packages/pip/_vendor/urllib3/_version.py000064400000000077150564745010015275 0ustar00# This file is protected via CODEOWNERS
__version__ = "1.26.7"
site-packages/pip/_vendor/urllib3/filepost.py000064400000004610150564745010015273 0ustar00from __future__ import absolute_import

import binascii
import codecs
import os
from io import BytesIO

from .fields import RequestField
from .packages import six
from .packages.six import b

writer = codecs.lookup("utf-8")[3]


def choose_boundary():
    """
    Our embarrassingly-simple replacement for mimetools.choose_boundary.
    """
    boundary = binascii.hexlify(os.urandom(16))
    if not six.PY2:
        boundary = boundary.decode("ascii")
    return boundary


def iter_field_objects(fields):
    """
    Iterate over fields.

    Supports list of (k, v) tuples and dicts, and lists of
    :class:`~urllib3.fields.RequestField`.

    """
    if isinstance(fields, dict):
        i = six.iteritems(fields)
    else:
        i = iter(fields)

    for field in i:
        if isinstance(field, RequestField):
            yield field
        else:
            yield RequestField.from_tuples(*field)


def iter_fields(fields):
    """
    .. deprecated:: 1.6

    Iterate over fields.

    The addition of :class:`~urllib3.fields.RequestField` makes this function
    obsolete. Instead, use :func:`iter_field_objects`, which returns
    :class:`~urllib3.fields.RequestField` objects.

    Supports list of (k, v) tuples and dicts.
    """
    if isinstance(fields, dict):
        return ((k, v) for k, v in six.iteritems(fields))

    return ((k, v) for k, v in fields)


def encode_multipart_formdata(fields, boundary=None):
    """
    Encode a dictionary of ``fields`` using the multipart/form-data MIME format.

    :param fields:
        Dictionary of fields or list of (key, :class:`~urllib3.fields.RequestField`).

    :param boundary:
        If not specified, then a random boundary will be generated using
        :func:`urllib3.filepost.choose_boundary`.
    """
    body = BytesIO()
    if boundary is None:
        boundary = choose_boundary()

    for field in iter_field_objects(fields):
        body.write(b("--%s\r\n" % (boundary)))

        writer(body).write(field.render_headers())
        data = field.data

        if isinstance(data, int):
            data = str(data)  # Backwards compatibility

        if isinstance(data, six.text_type):
            writer(body).write(data)
        else:
            body.write(data)

        body.write(b"\r\n")

    body.write(b("--%s--\r\n" % (boundary)))

    content_type = str("multipart/form-data; boundary=%s" % boundary)

    return body.getvalue(), content_type
site-packages/pip/_vendor/urllib3/_collections.py000064400000025073150564745010016131 0ustar00from __future__ import absolute_import

try:
    from collections.abc import Mapping, MutableMapping
except ImportError:
    from collections import Mapping, MutableMapping
try:
    from threading import RLock
except ImportError:  # Platform-specific: No threads available

    class RLock:
        def __enter__(self):
            pass

        def __exit__(self, exc_type, exc_value, traceback):
            pass


from collections import OrderedDict

from .exceptions import InvalidHeader
from .packages import six
from .packages.six import iterkeys, itervalues

__all__ = ["RecentlyUsedContainer", "HTTPHeaderDict"]


_Null = object()


class RecentlyUsedContainer(MutableMapping):
    """
    Provides a thread-safe dict-like container which maintains up to
    ``maxsize`` keys while throwing away the least-recently-used keys beyond
    ``maxsize``.

    :param maxsize:
        Maximum number of recent elements to retain.

    :param dispose_func:
        Every time an item is evicted from the container,
        ``dispose_func(value)`` is called.  Callback which will get called
    """

    ContainerCls = OrderedDict

    def __init__(self, maxsize=10, dispose_func=None):
        self._maxsize = maxsize
        self.dispose_func = dispose_func

        self._container = self.ContainerCls()
        self.lock = RLock()

    def __getitem__(self, key):
        # Re-insert the item, moving it to the end of the eviction line.
        with self.lock:
            item = self._container.pop(key)
            self._container[key] = item
            return item

    def __setitem__(self, key, value):
        evicted_value = _Null
        with self.lock:
            # Possibly evict the existing value of 'key'
            evicted_value = self._container.get(key, _Null)
            self._container[key] = value

            # If we didn't evict an existing value, we might have to evict the
            # least recently used item from the beginning of the container.
            if len(self._container) > self._maxsize:
                _key, evicted_value = self._container.popitem(last=False)

        if self.dispose_func and evicted_value is not _Null:
            self.dispose_func(evicted_value)

    def __delitem__(self, key):
        with self.lock:
            value = self._container.pop(key)

        if self.dispose_func:
            self.dispose_func(value)

    def __len__(self):
        with self.lock:
            return len(self._container)

    def __iter__(self):
        raise NotImplementedError(
            "Iteration over this class is unlikely to be threadsafe."
        )

    def clear(self):
        with self.lock:
            # Copy pointers to all values, then wipe the mapping
            values = list(itervalues(self._container))
            self._container.clear()

        if self.dispose_func:
            for value in values:
                self.dispose_func(value)

    def keys(self):
        with self.lock:
            return list(iterkeys(self._container))


class HTTPHeaderDict(MutableMapping):
    """
    :param headers:
        An iterable of field-value pairs. Must not contain multiple field names
        when compared case-insensitively.

    :param kwargs:
        Additional field-value pairs to pass in to ``dict.update``.

    A ``dict`` like container for storing HTTP Headers.

    Field names are stored and compared case-insensitively in compliance with
    RFC 7230. Iteration provides the first case-sensitive key seen for each
    case-insensitive pair.

    Using ``__setitem__`` syntax overwrites fields that compare equal
    case-insensitively in order to maintain ``dict``'s api. For fields that
    compare equal, instead create a new ``HTTPHeaderDict`` and use ``.add``
    in a loop.

    If multiple fields that are equal case-insensitively are passed to the
    constructor or ``.update``, the behavior is undefined and some will be
    lost.

    >>> headers = HTTPHeaderDict()
    >>> headers.add('Set-Cookie', 'foo=bar')
    >>> headers.add('set-cookie', 'baz=quxx')
    >>> headers['content-length'] = '7'
    >>> headers['SET-cookie']
    'foo=bar, baz=quxx'
    >>> headers['Content-Length']
    '7'
    """

    def __init__(self, headers=None, **kwargs):
        super(HTTPHeaderDict, self).__init__()
        self._container = OrderedDict()
        if headers is not None:
            if isinstance(headers, HTTPHeaderDict):
                self._copy_from(headers)
            else:
                self.extend(headers)
        if kwargs:
            self.extend(kwargs)

    def __setitem__(self, key, val):
        self._container[key.lower()] = [key, val]
        return self._container[key.lower()]

    def __getitem__(self, key):
        val = self._container[key.lower()]
        return ", ".join(val[1:])

    def __delitem__(self, key):
        del self._container[key.lower()]

    def __contains__(self, key):
        return key.lower() in self._container

    def __eq__(self, other):
        if not isinstance(other, Mapping) and not hasattr(other, "keys"):
            return False
        if not isinstance(other, type(self)):
            other = type(self)(other)
        return dict((k.lower(), v) for k, v in self.itermerged()) == dict(
            (k.lower(), v) for k, v in other.itermerged()
        )

    def __ne__(self, other):
        return not self.__eq__(other)

    if six.PY2:  # Python 2
        iterkeys = MutableMapping.iterkeys
        itervalues = MutableMapping.itervalues

    __marker = object()

    def __len__(self):
        return len(self._container)

    def __iter__(self):
        # Only provide the originally cased names
        for vals in self._container.values():
            yield vals[0]

    def pop(self, key, default=__marker):
        """D.pop(k[,d]) -> v, remove specified key and return the corresponding value.
        If key is not found, d is returned if given, otherwise KeyError is raised.
        """
        # Using the MutableMapping function directly fails due to the private marker.
        # Using ordinary dict.pop would expose the internal structures.
        # So let's reinvent the wheel.
        try:
            value = self[key]
        except KeyError:
            if default is self.__marker:
                raise
            return default
        else:
            del self[key]
            return value

    def discard(self, key):
        try:
            del self[key]
        except KeyError:
            pass

    def add(self, key, val):
        """Adds a (name, value) pair, doesn't overwrite the value if it already
        exists.

        >>> headers = HTTPHeaderDict(foo='bar')
        >>> headers.add('Foo', 'baz')
        >>> headers['foo']
        'bar, baz'
        """
        key_lower = key.lower()
        new_vals = [key, val]
        # Keep the common case aka no item present as fast as possible
        vals = self._container.setdefault(key_lower, new_vals)
        if new_vals is not vals:
            vals.append(val)

    def extend(self, *args, **kwargs):
        """Generic import function for any type of header-like object.
        Adapted version of MutableMapping.update in order to insert items
        with self.add instead of self.__setitem__
        """
        if len(args) > 1:
            raise TypeError(
                "extend() takes at most 1 positional "
                "arguments ({0} given)".format(len(args))
            )
        other = args[0] if len(args) >= 1 else ()

        if isinstance(other, HTTPHeaderDict):
            for key, val in other.iteritems():
                self.add(key, val)
        elif isinstance(other, Mapping):
            for key in other:
                self.add(key, other[key])
        elif hasattr(other, "keys"):
            for key in other.keys():
                self.add(key, other[key])
        else:
            for key, value in other:
                self.add(key, value)

        for key, value in kwargs.items():
            self.add(key, value)

    def getlist(self, key, default=__marker):
        """Returns a list of all the values for the named field. Returns an
        empty list if the key doesn't exist."""
        try:
            vals = self._container[key.lower()]
        except KeyError:
            if default is self.__marker:
                return []
            return default
        else:
            return vals[1:]

    # Backwards compatibility for httplib
    getheaders = getlist
    getallmatchingheaders = getlist
    iget = getlist

    # Backwards compatibility for http.cookiejar
    get_all = getlist

    def __repr__(self):
        return "%s(%s)" % (type(self).__name__, dict(self.itermerged()))

    def _copy_from(self, other):
        for key in other:
            val = other.getlist(key)
            if isinstance(val, list):
                # Don't need to convert tuples
                val = list(val)
            self._container[key.lower()] = [key] + val

    def copy(self):
        clone = type(self)()
        clone._copy_from(self)
        return clone

    def iteritems(self):
        """Iterate over all header lines, including duplicate ones."""
        for key in self:
            vals = self._container[key.lower()]
            for val in vals[1:]:
                yield vals[0], val

    def itermerged(self):
        """Iterate over all headers, merging duplicate ones together."""
        for key in self:
            val = self._container[key.lower()]
            yield val[0], ", ".join(val[1:])

    def items(self):
        return list(self.iteritems())

    @classmethod
    def from_httplib(cls, message):  # Python 2
        """Read headers from a Python 2 httplib message object."""
        # python2.7 does not expose a proper API for exporting multiheaders
        # efficiently. This function re-reads raw lines from the message
        # object and extracts the multiheaders properly.
        obs_fold_continued_leaders = (" ", "\t")
        headers = []

        for line in message.headers:
            if line.startswith(obs_fold_continued_leaders):
                if not headers:
                    # We received a header line that starts with OWS as described
                    # in RFC-7230 S3.2.4. This indicates a multiline header, but
                    # there exists no previous header to which we can attach it.
                    raise InvalidHeader(
                        "Header continuation with no previous header: %s" % line
                    )
                else:
                    key, value = headers[-1]
                    headers[-1] = (key, value + " " + line.strip())
                    continue

            key, value = line.split(":", 1)
            headers.append((key, value.strip()))

        return cls(headers)
site-packages/pip/_vendor/progress/__pycache__/colors.cpython-39.pyc000064400000003003150564745010021521 0ustar00a

��Re_
�@s�ddlmZdZdZddd�Zeedd�Zeed	d�Zeed
d�Zeedd�Zeedd�Z	eed
d�Z
eedd�Zeedd�Zeedd�Z
eedd�Zeedd�Zeedd�Zeedd�Zeedd�Zeedd�Zeedd�Zeedd�ZdS)�)�partial)�black�red�green�yellow�blue�magenta�cyan�white)	�bold�faint�italic�	underline�blink�blink2�negative�	concealed�crossedNcCsFg}|rl|tvr*|�tdt�|���nBt|t�r`d|krHdkr`nn|�dt|��ntd|��|r�|tvr�|�tdt�|���n>t|t�r�d|kr�dkr�nn|�d|�ntd|��|�r|�d�D]4}|tv�r|�td	t�|���q�td
|��q�|�r>dd�	|�d
}d}|||S|SdS)N�r�z38;5;%dzInvalid color "%s"�(z48;5;%d�+�zInvalid style "%s"z[�;�mz)
�COLORS�append�str�index�
isinstance�int�	Exception�split�STYLES�join)�s�fg�bg�style�sgr�st�prefix�suffix�r-��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/progress/colors.py�colors.""
r/r)r&rrrrrr	r
r)r(rr
rrrrrr)NNN)�	functoolsrrr#r/rrrrrrr	r
rrr
rrrrrrr-r-r-r.�<module>s(
#site-packages/pip/_vendor/progress/__pycache__/counter.cpython-39.pyc000064400000003200150564745010021676 0ustar00a

��ReM�@s`ddlmZddlmZmZGdd�de�ZGdd�de�ZGdd	�d	e�ZGd
d�de�ZdS)
�)�unicode_literals�)�Infinite�Progressc@seZdZdd�ZdS)�CountercCs,|j|}d�|t|j�g�}|�|�dS�N�)�message�join�str�index�writeln��selfr	�line�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/progress/counter.py�updates
zCounter.updateN��__name__�
__module__�__qualname__rrrrrrsrc@seZdZdd�ZdS)�	CountdowncCs,|j|}d�|t|j�g�}|�|�dSr)r	r
r�	remainingr
rrrrrs
zCountdown.updateNrrrrrrsrc@seZdZdZdd�ZdS)�Stack)	� u▁u▂u▃u▄u▅u▆u▇u█cCsNt|j�}t|dt|j|��}|j|}d�||j|g�}|�|�dS)Nrr)�len�phases�min�int�progressr	r
r
)r�nphases�ir	rrrrr&s


zStack.updateN)rrrrrrrrrr#src@seZdZdZdS)�Pie)u○u◔u◑u◕u●N)rrrrrrrrr#.sr#N)	�
__future__rrrrrrrr#rrrr�<module>s
site-packages/pip/_vendor/progress/__pycache__/bar.cpython-39.pyc000064400000005316150564745010020775 0ustar00a

��Re~�@s�ddlmZddlZddlmZddlmZGdd�de�ZGdd	�d	e�ZGd
d�de�Z	Gdd
�d
e�Z
Gdd�de�ZGdd�de�ZGdd�de�Z
dS)�)�unicode_literalsN�)�Progress)�colorc@s0eZdZdZdZdZdZdZdZdZ	dd	�Z
dS)
�Bar� z%(index)d/%(max)dz |z| � �#NcCstt|j|j�}|j|}|j|}t|j||jd�}|j|}|j|}d�||j	|||j
|g�}|�|�dS)N��fg�)�int�width�progress�messager�fill�
empty_fill�suffix�join�
bar_prefix�
bar_suffix�writeln)�self�
filled_length�empty_lengthr�bar�emptyr�line�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/progress/bar.py�update"s



�z
Bar.update)�__name__�
__module__�__qualname__rrrrrrrr rrrrrsrc@s eZdZdZdZdZdZdZdS)�ChargingBarz
%(percent)d%%ru∙�█N)r!r"r#rrrrrrrrrr$/s
r$c@seZdZdZdZdS)�FillingSquaresBaru▢u▣N�r!r"r#rrrrrrr&7sr&c@seZdZdZdZdS)�FillingCirclesBaru◯u◉Nr'rrrrr(<sr(c@s*eZdZej�d�rdZndZdd�ZdS)�IncrementalBar�win)r�▌r%)	ru▏u▎u▍r+u▋u▊u▉r%c	Cs�t|j�}|j|j}t|�}t|||�}|j|}|j|}t|jd||jd�}|dkrl|j|nd}|jtd|t|��}	|j	|}
d�
||j|||	|j|
g�}|�
|�dS)N���r
rr)�len�phasesrrr
rrr�maxrrrrr)r�nphases�
filled_len�nfull�phase�nemptyrr�currentrrrrrrr Gs



�zIncrementalBar.updateN)r!r"r#�sys�platform�
startswithr.r rrrrr)Asr)c@seZdZdZdS)�PixelBar)u⡀u⡄u⡆u⡇u⣇u⣧u⣷u⣿N�r!r"r#r.rrrrr9Xsr9c@seZdZdZdS)�ShadyBar)ru░u▒u▓r%Nr:rrrrr;\sr;)�
__future__rr6rr�colorsrrr$r&r(r)r9r;rrrr�<module>ssite-packages/pip/_vendor/progress/__pycache__/spinner.cpython-39.pyc000064400000002722150564745010021705 0ustar00a

��Re��@slddlmZddlmZGdd�de�ZGdd�de�ZGdd	�d	e�ZGd
d�de�ZGdd
�d
e�ZdS)�)�unicode_literals�)�Infinitec@seZdZdZdZdd�ZdS)�Spinner)�-�\�|�/TcCs<|jt|j�}|j|}d�||j|g�}|�|�dS)N�)�index�len�phases�message�join�writeln)�self�ir�line�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/progress/spinner.py�updates
zSpinner.updateN)�__name__�
__module__�__qualname__r
�hide_cursorrrrrrrsrc@seZdZgd�ZdS)�
PieSpinner)u◷u◶u◵u◴N�rrrr
rrrrr src@seZdZgd�ZdS)�MoonSpinner)u◑u◒u◐u◓Nrrrrrr$src@seZdZgd�ZdS)�LineSpinner)u⎺�⎻�⎼u⎽r rNrrrrrr(src@seZdZgd�ZdS)�PixelSpinner)u⣾u⣷u⣯u⣟u⡿u⢿u⣻u⣽Nrrrrrr!,sr!N)	�
__future__rr
rrrrrr!rrrr�<module>ssite-packages/pip/_vendor/progress/__pycache__/__init__.cpython-39.pyc000064400000013131150564745010021762 0ustar00a

��Re��@s�ddlmZmZddlmZddlmZddlmZddl	m
Z
zddlmZWne
ynddlmZYn0dZd	Zd
ZGdd�de�ZGd
d�de�ZdS)�)�division�print_function)�deque)�	timedelta)�ceil)�stderr)�	monotonic)�timez1.6z[?25lz[?25hc@s�eZdZeZdZdZdZd$dd�Zdd�Z	dd	�Z
ed
d��Zedd
��Z
dd�Zdd�Zdd�Zdd�Zdd�Zdd�Zd%dd�Zdd�Zdd �Zd!d"�Zd#S)&�Infinite�
T�cKs�d|_t�|_d|_|j|_|j|_t|jd�|_|�	�D]\}}t
|||�q:d|_d|_||_
|jr�|��r�|jr�ttd|jd�d|_|�d�dS)Nr)�maxlenFr��end�fileT)�indexr�start_ts�avg�_avg_update_ts�_tsr�
sma_window�_xput�items�setattr�
_max_width�_hidden_cursor�messager�is_tty�hide_cursor�print�HIDE_CURSOR�writeln)�selfr�kwargs�key�val�r&��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/progress/__init__.py�__init__'s zInfinite.__init__cCs|jrttd|jd�dS)Nrr)rr�SHOW_CURSORr�r"r&r&r'�__del__;szInfinite.__del__cCs|�d�rdSt||d�S)N�_)�
startswith�getattr)r"r$r&r&r'�__getitem__?s
zInfinite.__getitem__cCstt�|j�S�N)�intrrr*r&r&r'�elapsedDszInfinite.elapsedcCst|jd�S�N)�seconds)rr2r*r&r&r'�
elapsed_tdHszInfinite.elapsed_tdcCs`|dkr\t|j�}|j�||�t�}||jks@||jdkr\t|j�t|j�|_||_dS�Nr�)�lenr�appendrrr�sumr)r"�n�dt�xput_len�nowr&r&r'�
update_avgLs

�zInfinite.update_avgcCsdSr0r&r*r&r&r'�updateWszInfinite.updatecCsdSr0r&r*r&r&r'�startZszInfinite.startcCs\|jrX|��rXt|�}||jkr4|d|j|7}n||_td|d|jd�|j��dS)N� �
rr)rrr8rr�flush)r"�line�widthr&r&r'r!]s
zInfinite.writelncCs:|jr6|��r6t|jd�|jr6ttd|jd�d|_dS)N)rrrF)rrrrr)r*r&r&r'�finishhs
zInfinite.finishcCs>z|jr|j��ndWSty8d|}t|��Yn0dS)NTz:%s has no attribute 'isatty'. Try setting check_tty=False.)�	check_ttyr�isatty�AttributeError)r"�msgr&r&r'ros
zInfinite.is_ttyr7cCs:t�}||j}|�||�||_|j||_|��dSr0)rrr?rr@)r"r;r>r<r&r&r'�nextvs
z
Infinite.nextccsPd|_|�.|D]}||_|V|��qWd�n1s>0Y|`dSr0)�
iter_valuerL�r"�it�xr&r&r'�iter~s(z
Infinite.itercCs|��|Sr0)rAr*r&r&r'�	__enter__�szInfinite.__enter__cCs|��dSr0)rG)r"�exc_type�exc_val�exc_tbr&r&r'�__exit__�szInfinite.__exit__N)r)r7)�__name__�
__module__�__qualname__rrrrHrr(r+r/�propertyr2r5r?r@rAr!rGrrLrQrRrVr&r&r&r'r
!s*



	r
cspeZdZ�fdd�Zedd��Zedd��Zedd��Zed	d
��Zedd��Z	d
d�Z
dd�Zdd�Z�Z
S)�Progresscs(tt|�j|i|��|�dd�|_dS)N�max�d)�superr[r(�getr\)r"�argsr#��	__class__r&r'r(�szProgress.__init__cCstt|j|j��Sr0)r1rr�	remainingr*r&r&r'�eta�szProgress.etacCst|jd�Sr3)rrdr*r&r&r'�eta_td�szProgress.eta_tdcCs
|jdS)Nr])�progressr*r&r&r'�percent�szProgress.percentcCs |jdkrdStd|j|j�Sr6)r\�minrr*r&r&r'rf�s
zProgress.progresscCst|j|jd�S)Nr)r\rr*r&r&r'rc�szProgress.remainingcCs|��dSr0)r@r*r&r&r'rA�szProgress.startcCs||j}|�|�dSr0)rrL)r"r�incrr&r&r'�goto�s
z
Progress.gotoccsrzt|�|_Wnty Yn0d|_|�.|D]}||_|V|��q2Wd�n1s`0Y|`dSr0)r8r\�	TypeErrorrMrLrNr&r&r'rQ�s(z
Progress.iter)rWrXrYr(rZrdrergrfrcrArjrQ�
__classcell__r&r&rar'r[�s




r[N)�
__future__rr�collectionsr�datetimer�mathr�sysrr	r�ImportError�__version__r r)�objectr
r[r&r&r&r'�<module>snsite-packages/pip/_vendor/progress/colors.py000064400000005137150564745010015244 0ustar00# -*- coding: utf-8 -*-

# Copyright (c) 2020 Georgios Verigakis <verigak@gmail.com>
#
# Permission to use, copy, modify, and distribute this software for any
# purpose with or without fee is hereby granted, provided that the above
# copyright notice and this permission notice appear in all copies.
#
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
# ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
# OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.

from functools import partial


COLORS = ('black', 'red', 'green', 'yellow', 'blue', 'magenta', 'cyan',
          'white')
STYLES = ('bold', 'faint', 'italic', 'underline', 'blink', 'blink2',
          'negative', 'concealed', 'crossed')


def color(s, fg=None, bg=None, style=None):
    sgr = []

    if fg:
        if fg in COLORS:
            sgr.append(str(30 + COLORS.index(fg)))
        elif isinstance(fg, int) and 0 <= fg <= 255:
            sgr.append('38;5;%d' % int(fg))
        else:
            raise Exception('Invalid color "%s"' % fg)

    if bg:
        if bg in COLORS:
            sgr.append(str(40 + COLORS.index(bg)))
        elif isinstance(bg, int) and 0 <= bg <= 255:
            sgr.append('48;5;%d' % bg)
        else:
            raise Exception('Invalid color "%s"' % bg)

    if style:
        for st in style.split('+'):
            if st in STYLES:
                sgr.append(str(1 + STYLES.index(st)))
            else:
                raise Exception('Invalid style "%s"' % st)

    if sgr:
        prefix = '\x1b[' + ';'.join(sgr) + 'm'
        suffix = '\x1b[0m'
        return prefix + s + suffix
    else:
        return s


# Foreground shortcuts
black = partial(color, fg='black')
red = partial(color, fg='red')
green = partial(color, fg='green')
yellow = partial(color, fg='yellow')
blue = partial(color, fg='blue')
magenta = partial(color, fg='magenta')
cyan = partial(color, fg='cyan')
white = partial(color, fg='white')

# Style shortcuts
bold = partial(color, style='bold')
faint = partial(color, style='faint')
italic = partial(color, style='italic')
underline = partial(color, style='underline')
blink = partial(color, style='blink')
blink2 = partial(color, style='blink2')
negative = partial(color, style='negative')
concealed = partial(color, style='concealed')
crossed = partial(color, style='crossed')
site-packages/pip/_vendor/progress/bar.py000064400000005576150564745010014516 0ustar00# -*- coding: utf-8 -*-

# Copyright (c) 2012 Georgios Verigakis <verigak@gmail.com>
#
# Permission to use, copy, modify, and distribute this software for any
# purpose with or without fee is hereby granted, provided that the above
# copyright notice and this permission notice appear in all copies.
#
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
# ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
# OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.

from __future__ import unicode_literals

import sys

from . import Progress
from .colors import color


class Bar(Progress):
    width = 32
    suffix = '%(index)d/%(max)d'
    bar_prefix = ' |'
    bar_suffix = '| '
    empty_fill = ' '
    fill = '#'
    color = None

    def update(self):
        filled_length = int(self.width * self.progress)
        empty_length = self.width - filled_length

        message = self.message % self
        bar = color(self.fill * filled_length, fg=self.color)
        empty = self.empty_fill * empty_length
        suffix = self.suffix % self
        line = ''.join([message, self.bar_prefix, bar, empty, self.bar_suffix,
                        suffix])
        self.writeln(line)


class ChargingBar(Bar):
    suffix = '%(percent)d%%'
    bar_prefix = ' '
    bar_suffix = ' '
    empty_fill = '∙'
    fill = '█'


class FillingSquaresBar(ChargingBar):
    empty_fill = '▢'
    fill = '▣'


class FillingCirclesBar(ChargingBar):
    empty_fill = '◯'
    fill = '◉'


class IncrementalBar(Bar):
    if sys.platform.startswith('win'):
        phases = (u' ', u'▌', u'█')
    else:
        phases = (' ', '▏', '▎', '▍', '▌', '▋', '▊', '▉', '█')

    def update(self):
        nphases = len(self.phases)
        filled_len = self.width * self.progress
        nfull = int(filled_len)                      # Number of full chars
        phase = int((filled_len - nfull) * nphases)  # Phase of last char
        nempty = self.width - nfull                  # Number of empty chars

        message = self.message % self
        bar = color(self.phases[-1] * nfull, fg=self.color)
        current = self.phases[phase] if phase > 0 else ''
        empty = self.empty_fill * max(0, nempty - len(current))
        suffix = self.suffix % self
        line = ''.join([message, self.bar_prefix, bar, current, empty,
                        self.bar_suffix, suffix])
        self.writeln(line)


class PixelBar(IncrementalBar):
    phases = ('⡀', '⡄', '⡆', '⡇', '⣇', '⣧', '⣷', '⣿')


class ShadyBar(IncrementalBar):
    phases = (' ', '░', '▒', '▓', '█')
site-packages/pip/_vendor/progress/counter.py000064400000003115150564745010015414 0ustar00# -*- coding: utf-8 -*-

# Copyright (c) 2012 Georgios Verigakis <verigak@gmail.com>
#
# Permission to use, copy, modify, and distribute this software for any
# purpose with or without fee is hereby granted, provided that the above
# copyright notice and this permission notice appear in all copies.
#
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
# ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
# OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.

from __future__ import unicode_literals
from . import Infinite, Progress


class Counter(Infinite):
    def update(self):
        message = self.message % self
        line = ''.join([message, str(self.index)])
        self.writeln(line)


class Countdown(Progress):
    def update(self):
        message = self.message % self
        line = ''.join([message, str(self.remaining)])
        self.writeln(line)


class Stack(Progress):
    phases = (' ', '▁', '▂', '▃', '▄', '▅', '▆', '▇', '█')

    def update(self):
        nphases = len(self.phases)
        i = min(nphases - 1, int(self.progress * nphases))
        message = self.message % self
        line = ''.join([message, self.phases[i]])
        self.writeln(line)


class Pie(Stack):
    phases = ('○', '◔', '◑', '◕', '●')
site-packages/pip/_vendor/progress/__init__.py000064400000012256150564745010015502 0ustar00# Copyright (c) 2012 Georgios Verigakis <verigak@gmail.com>
#
# Permission to use, copy, modify, and distribute this software for any
# purpose with or without fee is hereby granted, provided that the above
# copyright notice and this permission notice appear in all copies.
#
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
# ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
# OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.

from __future__ import division, print_function

from collections import deque
from datetime import timedelta
from math import ceil
from sys import stderr
try:
    from time import monotonic
except ImportError:
    from time import time as monotonic


__version__ = '1.6'

HIDE_CURSOR = '\x1b[?25l'
SHOW_CURSOR = '\x1b[?25h'


class Infinite(object):
    file = stderr
    sma_window = 10         # Simple Moving Average window
    check_tty = True
    hide_cursor = True

    def __init__(self, message='', **kwargs):
        self.index = 0
        self.start_ts = monotonic()
        self.avg = 0
        self._avg_update_ts = self.start_ts
        self._ts = self.start_ts
        self._xput = deque(maxlen=self.sma_window)
        for key, val in kwargs.items():
            setattr(self, key, val)

        self._max_width = 0
        self._hidden_cursor = False
        self.message = message

        if self.file and self.is_tty():
            if self.hide_cursor:
                print(HIDE_CURSOR, end='', file=self.file)
                self._hidden_cursor = True
        self.writeln('')

    def __del__(self):
        if self._hidden_cursor:
            print(SHOW_CURSOR, end='', file=self.file)

    def __getitem__(self, key):
        if key.startswith('_'):
            return None
        return getattr(self, key, None)

    @property
    def elapsed(self):
        return int(monotonic() - self.start_ts)

    @property
    def elapsed_td(self):
        return timedelta(seconds=self.elapsed)

    def update_avg(self, n, dt):
        if n > 0:
            xput_len = len(self._xput)
            self._xput.append(dt / n)
            now = monotonic()
            # update when we're still filling _xput, then after every second
            if (xput_len < self.sma_window or
                    now - self._avg_update_ts > 1):
                self.avg = sum(self._xput) / len(self._xput)
                self._avg_update_ts = now

    def update(self):
        pass

    def start(self):
        pass

    def writeln(self, line):
        if self.file and self.is_tty():
            width = len(line)
            if width < self._max_width:
                # Add padding to cover previous contents
                line += ' ' * (self._max_width - width)
            else:
                self._max_width = width
            print('\r' + line, end='', file=self.file)
            self.file.flush()

    def finish(self):
        if self.file and self.is_tty():
            print(file=self.file)
            if self._hidden_cursor:
                print(SHOW_CURSOR, end='', file=self.file)
                self._hidden_cursor = False

    def is_tty(self):
        try:
            return self.file.isatty() if self.check_tty else True
        except AttributeError:
            msg = "%s has no attribute 'isatty'. Try setting check_tty=False." % self
            raise AttributeError(msg)

    def next(self, n=1):
        now = monotonic()
        dt = now - self._ts
        self.update_avg(n, dt)
        self._ts = now
        self.index = self.index + n
        self.update()

    def iter(self, it):
        self.iter_value = None
        with self:
            for x in it:
                self.iter_value = x
                yield x
                self.next()
        del self.iter_value

    def __enter__(self):
        self.start()
        return self

    def __exit__(self, exc_type, exc_val, exc_tb):
        self.finish()


class Progress(Infinite):
    def __init__(self, *args, **kwargs):
        super(Progress, self).__init__(*args, **kwargs)
        self.max = kwargs.get('max', 100)

    @property
    def eta(self):
        return int(ceil(self.avg * self.remaining))

    @property
    def eta_td(self):
        return timedelta(seconds=self.eta)

    @property
    def percent(self):
        return self.progress * 100

    @property
    def progress(self):
        if self.max == 0:
            return 0
        return min(1, self.index / self.max)

    @property
    def remaining(self):
        return max(self.max - self.index, 0)

    def start(self):
        self.update()

    def goto(self, index):
        incr = index - self.index
        self.next(incr)

    def iter(self, it):
        try:
            self.max = len(it)
        except TypeError:
            pass

        self.iter_value = None
        with self:
            for x in it:
                self.iter_value = x
                yield x
                self.next()
        del self.iter_value
site-packages/pip/_vendor/progress/spinner.py000064400000002665150564745010015424 0ustar00# -*- coding: utf-8 -*-

# Copyright (c) 2012 Georgios Verigakis <verigak@gmail.com>
#
# Permission to use, copy, modify, and distribute this software for any
# purpose with or without fee is hereby granted, provided that the above
# copyright notice and this permission notice appear in all copies.
#
# THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
# WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
# MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
# ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
# WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
# ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
# OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.

from __future__ import unicode_literals
from . import Infinite


class Spinner(Infinite):
    phases = ('-', '\\', '|', '/')
    hide_cursor = True

    def update(self):
        i = self.index % len(self.phases)
        message = self.message % self
        line = ''.join([message, self.phases[i]])
        self.writeln(line)


class PieSpinner(Spinner):
    phases = ['◷', '◶', '◵', '◴']


class MoonSpinner(Spinner):
    phases = ['◑', '◒', '◐', '◓']


class LineSpinner(Spinner):
    phases = ['⎺', '⎻', '⎼', '⎽', '⎼', '⎻']


class PixelSpinner(Spinner):
    phases = ['⣾', '⣷', '⣯', '⣟', '⡿', '⢿', '⣻', '⣽']
site-packages/pip/_vendor/tenacity/stop.py000064400000005346150564745010014706 0ustar00# Copyright 2016–2021 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import abc
import typing

if typing.TYPE_CHECKING:
    import threading

    from pip._vendor.tenacity import RetryCallState


class stop_base(abc.ABC):
    """Abstract base class for stop strategies."""

    @abc.abstractmethod
    def __call__(self, retry_state: "RetryCallState") -> bool:
        pass

    def __and__(self, other: "stop_base") -> "stop_all":
        return stop_all(self, other)

    def __or__(self, other: "stop_base") -> "stop_any":
        return stop_any(self, other)


class stop_any(stop_base):
    """Stop if any of the stop condition is valid."""

    def __init__(self, *stops: stop_base) -> None:
        self.stops = stops

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return any(x(retry_state) for x in self.stops)


class stop_all(stop_base):
    """Stop if all the stop conditions are valid."""

    def __init__(self, *stops: stop_base) -> None:
        self.stops = stops

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return all(x(retry_state) for x in self.stops)


class _stop_never(stop_base):
    """Never stop."""

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return False


stop_never = _stop_never()


class stop_when_event_set(stop_base):
    """Stop when the given event is set."""

    def __init__(self, event: "threading.Event") -> None:
        self.event = event

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return self.event.is_set()


class stop_after_attempt(stop_base):
    """Stop when the previous attempt >= max_attempt."""

    def __init__(self, max_attempt_number: int) -> None:
        self.max_attempt_number = max_attempt_number

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return retry_state.attempt_number >= self.max_attempt_number


class stop_after_delay(stop_base):
    """Stop when the time from the first attempt >= limit."""

    def __init__(self, max_delay: float) -> None:
        self.max_delay = max_delay

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return retry_state.seconds_since_start >= self.max_delay
site-packages/pip/_vendor/tenacity/after.py000064400000002730150564745010015014 0ustar00# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import typing

from pip._vendor.tenacity import _utils

if typing.TYPE_CHECKING:
    import logging

    from pip._vendor.tenacity import RetryCallState


def after_nothing(retry_state: "RetryCallState") -> None:
    """After call strategy that does nothing."""


def after_log(
    logger: "logging.Logger",
    log_level: int,
    sec_format: str = "%0.3f",
) -> typing.Callable[["RetryCallState"], None]:
    """After call strategy that logs to some logger the finished attempt."""

    def log_it(retry_state: "RetryCallState") -> None:
        logger.log(
            log_level,
            f"Finished call to '{_utils.get_callback_name(retry_state.fn)}' "
            f"after {sec_format % retry_state.seconds_since_start}(s), "
            f"this was the {_utils.to_ordinal(retry_state.attempt_number)} time calling it.",
        )

    return log_it
site-packages/pip/_vendor/tenacity/wait.py000064400000015043150564745010014660 0ustar00# Copyright 2016–2021 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import abc
import random
import typing

from pip._vendor.tenacity import _utils

if typing.TYPE_CHECKING:
    from pip._vendor.tenacity import RetryCallState


class wait_base(abc.ABC):
    """Abstract base class for wait strategies."""

    @abc.abstractmethod
    def __call__(self, retry_state: "RetryCallState") -> float:
        pass

    def __add__(self, other: "wait_base") -> "wait_combine":
        return wait_combine(self, other)

    def __radd__(self, other: "wait_base") -> typing.Union["wait_combine", "wait_base"]:
        # make it possible to use multiple waits with the built-in sum function
        if other == 0:
            return self
        return self.__add__(other)


class wait_fixed(wait_base):
    """Wait strategy that waits a fixed amount of time between each retry."""

    def __init__(self, wait: float) -> None:
        self.wait_fixed = wait

    def __call__(self, retry_state: "RetryCallState") -> float:
        return self.wait_fixed


class wait_none(wait_fixed):
    """Wait strategy that doesn't wait at all before retrying."""

    def __init__(self) -> None:
        super().__init__(0)


class wait_random(wait_base):
    """Wait strategy that waits a random amount of time between min/max."""

    def __init__(self, min: typing.Union[int, float] = 0, max: typing.Union[int, float] = 1) -> None:  # noqa
        self.wait_random_min = min
        self.wait_random_max = max

    def __call__(self, retry_state: "RetryCallState") -> float:
        return self.wait_random_min + (random.random() * (self.wait_random_max - self.wait_random_min))


class wait_combine(wait_base):
    """Combine several waiting strategies."""

    def __init__(self, *strategies: wait_base) -> None:
        self.wait_funcs = strategies

    def __call__(self, retry_state: "RetryCallState") -> float:
        return sum(x(retry_state=retry_state) for x in self.wait_funcs)


class wait_chain(wait_base):
    """Chain two or more waiting strategies.

    If all strategies are exhausted, the very last strategy is used
    thereafter.

    For example::

        @retry(wait=wait_chain(*[wait_fixed(1) for i in range(3)] +
                               [wait_fixed(2) for j in range(5)] +
                               [wait_fixed(5) for k in range(4)))
        def wait_chained():
            print("Wait 1s for 3 attempts, 2s for 5 attempts and 5s
                   thereafter.")
    """

    def __init__(self, *strategies: wait_base) -> None:
        self.strategies = strategies

    def __call__(self, retry_state: "RetryCallState") -> float:
        wait_func_no = min(max(retry_state.attempt_number, 1), len(self.strategies))
        wait_func = self.strategies[wait_func_no - 1]
        return wait_func(retry_state=retry_state)


class wait_incrementing(wait_base):
    """Wait an incremental amount of time after each attempt.

    Starting at a starting value and incrementing by a value for each attempt
    (and restricting the upper limit to some maximum value).
    """

    def __init__(
        self,
        start: typing.Union[int, float] = 0,
        increment: typing.Union[int, float] = 100,
        max: typing.Union[int, float] = _utils.MAX_WAIT,  # noqa
    ) -> None:
        self.start = start
        self.increment = increment
        self.max = max

    def __call__(self, retry_state: "RetryCallState") -> float:
        result = self.start + (self.increment * (retry_state.attempt_number - 1))
        return max(0, min(result, self.max))


class wait_exponential(wait_base):
    """Wait strategy that applies exponential backoff.

    It allows for a customized multiplier and an ability to restrict the
    upper and lower limits to some maximum and minimum value.

    The intervals are fixed (i.e. there is no jitter), so this strategy is
    suitable for balancing retries against latency when a required resource is
    unavailable for an unknown duration, but *not* suitable for resolving
    contention between multiple processes for a shared resource. Use
    wait_random_exponential for the latter case.
    """

    def __init__(
        self,
        multiplier: typing.Union[int, float] = 1,
        max: typing.Union[int, float] = _utils.MAX_WAIT,  # noqa
        exp_base: typing.Union[int, float] = 2,
        min: typing.Union[int, float] = 0,  # noqa
    ) -> None:
        self.multiplier = multiplier
        self.min = min
        self.max = max
        self.exp_base = exp_base

    def __call__(self, retry_state: "RetryCallState") -> float:
        try:
            exp = self.exp_base ** (retry_state.attempt_number - 1)
            result = self.multiplier * exp
        except OverflowError:
            return self.max
        return max(max(0, self.min), min(result, self.max))


class wait_random_exponential(wait_exponential):
    """Random wait with exponentially widening window.

    An exponential backoff strategy used to mediate contention between multiple
    uncoordinated processes for a shared resource in distributed systems. This
    is the sense in which "exponential backoff" is meant in e.g. Ethernet
    networking, and corresponds to the "Full Jitter" algorithm described in
    this blog post:

    https://aws.amazon.com/blogs/architecture/exponential-backoff-and-jitter/

    Each retry occurs at a random time in a geometrically expanding interval.
    It allows for a custom multiplier and an ability to restrict the upper
    limit of the random interval to some maximum value.

    Example::

        wait_random_exponential(multiplier=0.5,  # initial window 0.5s
                                max=60)          # max 60s timeout

    When waiting for an unavailable resource to become available again, as
    opposed to trying to resolve contention for a shared resource, the
    wait_exponential strategy (which uses a fixed interval) may be preferable.

    """

    def __call__(self, retry_state: "RetryCallState") -> float:
        high = super().__call__(retry_state=retry_state)
        return random.uniform(0, high)
site-packages/pip/_vendor/tenacity/__pycache__/after.cpython-39.pyc000064400000002341150564745010021301 0ustar00a

��Re��@sdddlZddlmZejr.ddlZddlmZddd�dd�Zd
d	eeej	dgdfd
�dd�Z
dS)�N)�_utils)�RetryCallStater��retry_state�returncCsdS)z&After call strategy that does nothing.N��rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/after.py�
after_nothingsr
�%0.3fzlogging.Logger)�logger�	log_level�
sec_formatrcsddd����fdd�}|S)zBAfter call strategy that logs to some logger the finished attempt.rNrcs:���dt�|j��d�|j�dt�|j��d��dS)NzFinished call to 'z' after z(s), this was the z time calling it.)�logr�get_callback_name�fn�seconds_since_start�
to_ordinal�attempt_numberr�r
rrrr	�log_it&s�
��zafter_log.<locals>.log_itr)rr
rrrrr	�	after_logsr)r)�typing�pip._vendor.tenacityr�
TYPE_CHECKING�loggingrr
�int�str�Callablerrrrr	�<module>s��site-packages/pip/_vendor/tenacity/__pycache__/nap.cpython-39.pyc000064400000002317150564745010020761 0ustar00a

��Reg�@s@ddlZddlZejrddlZedd�dd�ZGdd�d�ZdS)�N)�seconds�returncCst�|�dS)z�
    Sleep strategy that delays execution for a given number of seconds.

    This is the default strategy, and may be mocked out for unit testing.
    N)�time�sleep)r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/nap.pyrsrc@s6eZdZdZddd�dd�Zejedd�dd	�ZdS)
�sleep_using_eventz0Sleep strategy that waits on an event to be set.zthreading.EventN)�eventrcCs
||_dS)N)r	)�selfr	rrr�__init__%szsleep_using_event.__init__)�timeoutrcCs|jj|d�dS)N)r)r	�wait)r
rrrr�__call__(szsleep_using_event.__call__)	�__name__�
__module__�__qualname__�__doc__r�typing�Optional�floatrrrrrr"sr)rr�
TYPE_CHECKING�	threadingrrrrrrr�<module>s
	site-packages/pip/_vendor/tenacity/__pycache__/wait.cpython-39.pyc000064400000017473150564745010021160 0ustar00a

��Re#�@s�ddlZddlZddlZddlmZejr6ddlmZGdd�dej�ZGdd�de�Z	Gdd	�d	e	�Z
Gd
d�de�ZGdd
�d
e�ZGdd�de�Z
Gdd�de�ZGdd�de�ZGdd�de�ZdS)�N)�_utils)�RetryCallStatec@sLeZdZdZejded�dd��Zddd�dd	�Zde	j
d
d�dd�Zd
S)�	wait_basez(Abstract base class for wait strategies.r��retry_state�returncCsdS�N���selfrr	r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/wait.py�__call__szwait_base.__call__�wait_combine)�otherrcCs
t||�Sr)r�rrr	r	r�__add__"szwait_base.__add__)rrcCs|dkr|S|�|�S�Nr)rrr	r	r�__radd__%szwait_base.__radd__N)�__name__�
__module__�__qualname__�__doc__�abc�abstractmethod�floatr
r�typing�Unionrr	r	r	rrs
rc@s0eZdZdZedd�dd�Zded�dd	�ZdS)
�
wait_fixedzCWait strategy that waits a fixed amount of time between each retry.N)�waitrcCs
||_dSr�r)rrr	r	r�__init__/szwait_fixed.__init__rrcCs|jSrrr
r	r	rr
2szwait_fixed.__call__)rrrrrr r
r	r	r	rr,srcs&eZdZdZdd��fdd�Z�ZS)�	wait_nonez7Wait strategy that doesn't wait at all before retrying.N)rcst��d�dSr)�superr )r��	__class__r	rr 9szwait_none.__init__)rrrrr �
__classcell__r	r	r#rr!6sr!c@sHeZdZdZdejeefejeefdd�dd�Zded	�d
d�Z	dS)
�wait_randomzAWait strategy that waits a random amount of time between min/max.r�N)�min�maxrcCs||_||_dSr)�wait_random_min�wait_random_max)rr(r)r	r	rr @szwait_random.__init__rrcCs|jt��|j|jSr)r*�randomr+r
r	r	rr
Dszwait_random.__call__)rr')
rrrrrr�intrr r
r	r	r	rr&=s(r&c@s0eZdZdZedd�dd�Zded�dd	�ZdS)
rz#Combine several waiting strategies.N��
strategiesrcGs
||_dSr)�
wait_funcs�rr/r	r	rr Kszwait_combine.__init__rrcst�fdd�|jD��S)Nc3s|]}|�d�VqdS)�rNr	)�.0�xr2r	r�	<genexpr>O�z(wait_combine.__call__.<locals>.<genexpr>)�sumr0r
r	r2rr
Nszwait_combine.__call__�rrrrrr rr
r	r	r	rrHsrc@s0eZdZdZedd�dd�Zded�dd	�ZdS)
�
wait_chaina�Chain two or more waiting strategies.

    If all strategies are exhausted, the very last strategy is used
    thereafter.

    For example::

        @retry(wait=wait_chain(*[wait_fixed(1) for i in range(3)] +
                               [wait_fixed(2) for j in range(5)] +
                               [wait_fixed(5) for k in range(4)))
        def wait_chained():
            print("Wait 1s for 3 attempts, 2s for 5 attempts and 5s
                   thereafter.")
    Nr.cGs
||_dSr)r/r1r	r	rr bszwait_chain.__init__rrcCs0tt|jd�t|j��}|j|d}||d�S)Nr'r2)r(r)�attempt_number�lenr/)rr�wait_func_no�	wait_funcr	r	rr
eszwait_chain.__call__r8r	r	r	rr9Rsr9c@s\eZdZdZddejfejee	fejee	fejee	fdd�dd�Z
de	d	�d
d�ZdS)�wait_incrementingz�Wait an incremental amount of time after each attempt.

    Starting at a starting value and incrementing by a value for each attempt
    (and restricting the upper limit to some maximum value).
    r�dN)�start�	incrementr)rcCs||_||_||_dSr)r@rAr))rr@rAr)r	r	rr rszwait_incrementing.__init__rrcCs(|j|j|jd}tdt||j��S�Nr'r)r@rAr:r)r()rr�resultr	r	rr
|szwait_incrementing.__call__�rrrrr�MAX_WAITrrr-rr r
r	r	r	rr>ks��
r>c@sjeZdZdZdejddfejee	fejee	fejee	fejee	fdd�dd�Z
d	e	d
�dd�ZdS)
�wait_exponentialaWait strategy that applies exponential backoff.

    It allows for a customized multiplier and an ability to restrict the
    upper and lower limits to some maximum and minimum value.

    The intervals are fixed (i.e. there is no jitter), so this strategy is
    suitable for balancing retries against latency when a required resource is
    unavailable for an unknown duration, but *not* suitable for resolving
    contention between multiple processes for a shared resource. Use
    wait_random_exponential for the latter case.
    r'�rN)�
multiplierr)�exp_baser(rcCs||_||_||_||_dSr)rHr(r)rI)rrHr)rIr(r	r	rr �szwait_exponential.__init__rrcCsRz|j|jd}|j|}Wnty6|jYS0ttd|j�t||j��SrB)rIr:rH�
OverflowErrorr)r()rr�exprCr	r	rr
�szwait_exponential.__call__rDr	r	r	rrF�s��rFcs(eZdZdZded��fdd�Z�ZS)�wait_random_exponentialaRandom wait with exponentially widening window.

    An exponential backoff strategy used to mediate contention between multiple
    uncoordinated processes for a shared resource in distributed systems. This
    is the sense in which "exponential backoff" is meant in e.g. Ethernet
    networking, and corresponds to the "Full Jitter" algorithm described in
    this blog post:

    https://aws.amazon.com/blogs/architecture/exponential-backoff-and-jitter/

    Each retry occurs at a random time in a geometrically expanding interval.
    It allows for a custom multiplier and an ability to restrict the upper
    limit of the random interval to some maximum value.

    Example::

        wait_random_exponential(multiplier=0.5,  # initial window 0.5s
                                max=60)          # max 60s timeout

    When waiting for an unavailable resource to become available again, as
    opposed to trying to resolve contention for a shared resource, the
    wait_exponential strategy (which uses a fixed interval) may be preferable.

    rrcst�j|d�}t�d|�S)Nr2r)r"r
r,�uniform)rr�highr#r	rr
�sz wait_random_exponential.__call__)rrrrrr
r%r	r	r#rrL�srL)rr,r�pip._vendor.tenacityr�
TYPE_CHECKINGr�ABCrrr!r&rr9r>rFrLr	r	r	r�<module>s

"site-packages/pip/_vendor/tenacity/__pycache__/retry.cpython-39.pyc000064400000021163150564745010021350 0ustar00a

��Re��@sddlZddlZddlZejr*ddlmZGdd�dej�ZGdd�de�Ze�Z	Gdd�de�Z
e
�ZGd	d
�d
e�ZGdd�de�Z
Gd
d�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZdS)�N)�RetryCallStatec@sFeZdZdZejded�dd��Zddd�dd	�Zdd
d�dd�Z	d
S)�
retry_basez)Abstract base class for retry strategies.r��retry_state�returncCsdS�N���selfrrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/retry.py�__call__szretry_base.__call__�	retry_all)�otherrcCs
t||�Sr)r
�r
rrrr�__and__ szretry_base.__and__�	retry_anycCs
t||�Sr)rrrrr�__or__#szretry_base.__or__N)
�__name__�
__module__�__qualname__�__doc__�abc�abstractmethod�boolrrrrrrrrs
rc@s eZdZdZded�dd�ZdS)�_retry_neverz-Retry strategy that never rejects any result.rrcCsdS�NFrr	rrrr*sz_retry_never.__call__N�rrrrrrrrrrr'src@s eZdZdZded�dd�ZdS)�
_retry_alwaysz.Retry strategy that always rejects any result.rrcCsdS�NTrr	rrrr4sz_retry_always.__call__Nrrrrrr1src@s<eZdZdZejegefdd�dd�Zded�dd	�Z	dS)
�retry_if_exceptionzARetry strategy that retries if an exception verifies a predicate.N��	predicatercCs
||_dSr�r!�r
r!rrr�__init__>szretry_if_exception.__init__rrcCs |jjr|�|j���SdSdSr��outcome�failedr!�	exceptionr	rrrrAszretry_if_exception.__call__)
rrrr�typing�Callable�
BaseExceptionrr$rrrrrr;srcsLeZdZdZefejejeej	ejedffdd��fdd�
Z
�ZS)�retry_if_exception_typez=Retries if an exception has been raised of one or more types..N��exception_typesrcs�|_t���fdd��dS)Ncs
t|��Sr��
isinstance��e�r.rr�<lambda>S�z2retry_if_exception_type.__init__.<locals>.<lambda>�r.�superr$�r
r.��	__class__r3rr$Ksz retry_if_exception_type.__init__�rrrr�	Exceptionr)�Union�Typer+�Tupler$�
__classcell__rrr9rr,Hs����r,csLeZdZdZefejejeej	ejedffdd��fdd�
Z
�ZS)�retry_if_not_exception_typezARetries except an exception has been raised of one or more types..Nr-cs�|_t���fdd��dS)Ncst|��Srr/r1r3rrr4ar5z6retry_if_not_exception_type.__init__.<locals>.<lambda>r6r8r9r3rr$Ysz$retry_if_not_exception_type.__init__r;rrr9rrAVs����rAcs\eZdZdZefejejeej	ejedffdd��fdd�
Z
ded�d	d
�Z�Z
S)�retry_unless_exception_typez:Retries until an exception is raised of one or more types..Nr-cs�|_t���fdd��dS)Ncst|��Srr/r1r3rrr4or5z6retry_unless_exception_type.__init__.<locals>.<lambda>r6r8r9r3rr$gsz$retry_unless_exception_type.__init__rrcCs|jjsdS|�|j���Srr%r	rrrrqsz$retry_unless_exception_type.__call__)rrrrr<r)r=r>r+r?r$rrr@rrr9rrBds����
rBc@s>eZdZdZejejgefdd�dd�Zded�dd	�Z	dS)
�retry_if_resultz+Retries if the result verifies a predicate.Nr cCs
||_dSrr"r#rrrr${szretry_if_result.__init__rrcCs |jjs|�|j���SdSdSr�r&r'r!�resultr	rrrr~szretry_if_result.__call__�
rrrrr)r*�Anyrr$rrrrrrCxsrCc@s>eZdZdZejejgefdd�dd�Zded�dd	�Z	dS)
�retry_if_not_resultz*Retries if the result refutes a predicate.Nr cCs
||_dSrr"r#rrrr$�szretry_if_not_result.__init__rrcCs"|jjs|�|j���SdSdSrrDr	rrrr�szretry_if_not_result.__call__rFrrrrrH�srHcs8eZdZdZdejeejedd��fdd�
Z�ZS)�retry_if_exception_messagez2Retries if an exception message equals or matches.N��message�matchrcs��r|rt|jj�d����r8ttd��fdd�}|}n:|r`t�|��ttd��fdd�}|}nt|jj�d���t��|�dS)Nz.() takes either 'message' or 'match', not both)r(rcs�t|�kSr)�str�r()rKrr�message_fnc�sz8retry_if_exception_message.__init__.<locals>.message_fnccst��t|���Sr)rrLrMrN)�progrr�	match_fnc�sz6retry_if_exception_message.__init__.<locals>.match_fncz3() missing 1 required argument 'message' or 'match')	�	TypeErrorr:rr+r�re�compiler7r$)r
rKrLrOr!rQr9)rKrPrr$�s
z#retry_if_exception_message.__init__)NN)	rrrrr)�OptionalrMr$r@rrr9rrI�s��rIcsHeZdZdZd
ejeejedd��fdd�
Zded�dd	�Z	�Z
S)�retry_if_not_exception_messagez5Retries until an exception message equals or matches.NrJcs&t��||�|j��fdd�|_dS)Ncs�|i|��Srr)�args_�kwargs_��if_predicaterrr4�r5z9retry_if_not_exception_message.__init__.<locals>.<lambda>)r7r$r!)r
rKrLr9rYrr$�sz'retry_if_not_exception_message.__init__rrcCs|jjsdS|�|j���Srr%r	rrrr�sz'retry_if_not_exception_message.__call__)NN)rrrrr)rUrMr$rrr@rrr9rrV�s��
rVc@s0eZdZdZedd�dd�Zded�dd	�ZdS)
rz1Retries if any of the retries condition is valid.N��retriesrcGs
||_dSr�r\�r
r\rrrr$�szretry_any.__init__rrcst�fdd�|jD��S)Nc3s|]}|��VqdSrr��.0�r�rrr�	<genexpr>�r5z%retry_any.__call__.<locals>.<genexpr>)�anyr\r	rrbrr�szretry_any.__call__�rrrrrr$rrrrrrr�src@s0eZdZdZedd�dd�Zded�dd	�ZdS)
r
z/Retries if all the retries condition are valid.Nr[cGs
||_dSrr]r^rrrr$�szretry_all.__init__rrcst�fdd�|jD��S)Nc3s|]}|��VqdSrrr_rbrrrc�r5z%retry_all.__call__.<locals>.<genexpr>)�allr\r	rrbrr�szretry_all.__call__rerrrrr
�sr
)rrSr)�
TYPE_CHECKING�pip._vendor.tenacityr�ABCrr�retry_neverr�retry_alwaysrr,rArBrCrHrIrVrr
rrrr�<module>s&



site-packages/pip/_vendor/tenacity/__pycache__/stop.cpython-39.pyc000064400000010273150564745010021170 0ustar00a

��Re�
�@s�ddlZddlZejr*ddlZddlmZGdd�dej�ZGdd�de�ZGdd�de�Z	Gd	d
�d
e�Z
e
�ZGdd�de�ZGd
d�de�Z
Gdd�de�ZdS)�N)�RetryCallStatec@sFeZdZdZejded�dd��Zddd�dd	�Zdd
d�dd�Z	d
S)�	stop_basez(Abstract base class for stop strategies.r��retry_state�returncCsdS�N���selfrrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/stop.py�__call__szstop_base.__call__�stop_all)�otherrcCs
t||�Sr)r
�r
rrrr�__and__ szstop_base.__and__�stop_anycCs
t||�Sr)rrrrr�__or__#szstop_base.__or__N)
�__name__�
__module__�__qualname__�__doc__�abc�abstractmethod�boolrrrrrrrrs
rc@s0eZdZdZedd�dd�Zded�dd	�ZdS)
rz+Stop if any of the stop condition is valid.N��stopsrcGs
||_dSr�r�r
rrrr�__init__*szstop_any.__init__rrcst�fdd�|jD��S)Nc3s|]}|��VqdSrr��.0�x�rrr�	<genexpr>.�z$stop_any.__call__.<locals>.<genexpr>)�anyrr	rr"rr-szstop_any.__call__�rrrrrrrrrrrrr'src@s0eZdZdZedd�dd�Zded�dd	�ZdS)
r
z*Stop if all the stop conditions are valid.NrcGs
||_dSrrrrrrr4szstop_all.__init__rrcst�fdd�|jD��S)Nc3s|]}|��VqdSrrrr"rrr#8r$z$stop_all.__call__.<locals>.<genexpr>)�allrr	rr"rr7szstop_all.__call__r&rrrrr
1sr
c@s eZdZdZded�dd�ZdS)�_stop_neverzNever stop.rrcCsdS)NFrr	rrrr>sz_stop_never.__call__N)rrrrrrrrrrr(;sr(c@s0eZdZdZddd�dd�Zded�d	d
�ZdS)�stop_when_event_setz!Stop when the given event is set.zthreading.EventN)�eventrcCs
||_dSr)r*)r
r*rrrrHszstop_when_event_set.__init__rrcCs
|j��Sr)r*�is_setr	rrrrKszstop_when_event_set.__call__)rrrrrrrrrrrr)Esr)c@s0eZdZdZedd�dd�Zded�dd	�ZdS)
�stop_after_attemptz.Stop when the previous attempt >= max_attempt.N)�max_attempt_numberrcCs
||_dSr)r-)r
r-rrrrRszstop_after_attempt.__init__rrcCs|j|jkSr)�attempt_numberr-r	rrrrUszstop_after_attempt.__call__)rrrr�intrrrrrrrr,Osr,c@s0eZdZdZedd�dd�Zded�dd	�ZdS)
�stop_after_delayz3Stop when the time from the first attempt >= limit.N)�	max_delayrcCs
||_dSr)r1)r
r1rrrr\szstop_after_delay.__init__rrcCs|j|jkSr)�seconds_since_startr1r	rrrr_szstop_after_delay.__call__)rrrr�floatrrrrrrrr0Ysr0)r�typing�
TYPE_CHECKING�	threading�pip._vendor.tenacityr�ABCrrr
r(�
stop_neverr)r,r0rrrr�<module>s



site-packages/pip/_vendor/tenacity/__pycache__/tornadoweb.cpython-39.pyc000064400000003367150564745010022355 0ustar00a

��Rea�@s|ddlZddlZddlmZddlmZddlmZddlmZddlmZej	r^ddl
mZe�d�Z
Gd	d
�d
e�ZdS)�N)�BaseRetrying)�	DoAttempt)�DoSleep)�RetryCallState)�gen)�Future�_RetValTcsLeZdZejfdejdd��fdd�
Zejdejejdd�d	d
��Z	�Z
S)�TornadoRetryingz&typing.Callable[[float], Future[None]]N)�sleep�kwargs�returncst�jfi|��||_dS)N)�super�__init__r
)�selfr
r��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/tornadoweb.pyr szTornadoRetrying.__init__zhtyping.Callable[..., typing.Union[typing.Generator[typing.Any, typing.Any, _RetValT], Future[_RetValT]]]z2typing.Generator[typing.Any, typing.Any, _RetValT])�fn�argsrrcos�|��t||||d�}|j|d�}t|t�rpz||i|��V}Wn tyb|�t���Yq�0|�	|�qt|t
�r�|��|�|�Vqt
�|��qdS)N)�retry_objectrrr)�retry_state)�beginr�iter�
isinstancer�
BaseException�
set_exception�sys�exc_info�
set_resultr�prepare_for_next_attemptr
r�Return)rrrrr�do�resultrrr�__call__$s

zTornadoRetrying.__call__)�__name__�
__module__�__qualname__rr
�typing�Anyr�	coroutiner$�
__classcell__rrrrr	s�r	)rr(�pip._vendor.tenacityrrrr�tornador�
TYPE_CHECKINGZtornado.concurrentr�TypeVarrr	rrrr�<module>s
site-packages/pip/_vendor/tenacity/__pycache__/before_sleep.cpython-39.pyc000064400000002625150564745010022637 0ustar00a

��Ret�@sdddlZddlmZejr.ddlZddlmZddd�dd�Zd
d	eeej	dgdfd
�dd�Z
dS)�N)�_utils)�RetryCallStater��retry_state�returncCsdS)z'Before call strategy that does nothing.N�)rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/before_sleep.py�before_sleep_nothingsr	Fzlogging.Logger)�logger�	log_level�exc_inforcsddd����fdd�}|S)z:Before call strategy that logs to some logger the attempt.rNrcs�|jjr@|j��}d|jj�d|��}}�r:|j��}qTd}nd|j��}}d}�j�dt�|j	��d|j
j�d|�d|�d	�	|d
�dS)N�raisedz: F�returnedz	Retrying z in z seconds as it � �.)r)�outcome�failed�	exception�	__class__�__name__�result�logr�get_callback_name�fn�next_action�sleep)r�ex�verb�value�local_exc_info�rrr
rr�log_it&s&
����z before_sleep_log.<locals>.log_itr)r
rrr!rr r�before_sleep_logsr")F)�typing�pip._vendor.tenacityr�
TYPE_CHECKING�loggingrr	�int�bool�Callabler"rrrr�<module>s��site-packages/pip/_vendor/tenacity/__pycache__/_asyncio.cpython-39.pyc000064400000005111150564745010022002 0ustar00a

��Re��@s�ddlZddlZddlZddlmZddlmZddlmZddlmZddlm	Z	ddlm
Z
ejdejd	�Z
e�d
�ZGdd�de�ZdS)
�N)�sleep)�AttemptManager)�BaseRetrying)�	DoAttempt)�DoSleep)�RetryCallState�	WrappedFn)�bound�_RetValTcs�eZdZefejegejfejdd��fdd�
Z	ejdeje
fejeje
d�dd�Zdd	�d
d�Zej
eejfd	�dd
�Zeed��fdd�Z�ZS)�
AsyncRetryingN)r�kwargs�returncst�jfi|��||_dS�N)�super�__init__r)�selfrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/_asyncio.pyr"szAsyncRetrying.__init__.)�fn�argsrr
c�s�|��t||||d�}|j|d�}t|t�rtz||i|��IdH}Wn tyf|�t���Yq�0|�	|�qt|t
�r�|��|�|�IdHq|SqdS)N)�retry_objectrrr��retry_state)
�beginr�iter�
isinstancer�
BaseException�
set_exception�sys�exc_info�
set_resultr�prepare_for_next_attemptr)rrrrr�do�resultrrr�__call__&s

zAsyncRetrying.__call__)r
cCs|��t|ddid�|_|S)Nr)rrr)rr�_retry_state)rrrr�	__aiter__>szAsyncRetrying.__aiter__c�sb|j|jd�}|durt�qt|t�r2t|jd�St|t�rX|j��|�|�IdHq|SqdS)Nr)	rr'�StopAsyncIterationrrrrr#r)rr$rrr�	__anext__Cs


zAsyncRetrying.__anext__)rr
csFt�����t���tjtjtjd��fdd��}�j|_�j|_|S)N)rrr
c�s�|i|��IdHSrr)rr�rrr�
async_wrappedTsz*AsyncRetrying.wraps.<locals>.async_wrapped)r�wraps�	functools�typing�Any�retry�
retry_with)rrr,rr+rr-PszAsyncRetrying.wraps)�__name__�
__module__�__qualname__rr/�Callable�float�	Awaitabler0rr
r&r(�Unionrr*rr-�
__classcell__rrrrr!s*�
r)r.r r/�asyncior�pip._vendor.tenacityrrrrr�TypeVarr6rr
rrrrr�<module>s
site-packages/pip/_vendor/tenacity/__pycache__/__init__.cpython-39.pyc000064400000037676150564745010021762 0ustar00a

��ReQG�@s�ddlZddlZddlZddlZddlZddlZddlmZm	Z	ddl
mZddlm
Z
ddlmZddlmZddlmZdd	lmZdd
lmZddlmZddlmZdd
lmZddlmZddlmZddlmZddlmZddlmZddlmZddlmZddlm Z ddlm!Z!ddlm"Z"ddlm#Z#ddlm$Z$ddlm%Z%ddl&m'Z'ddl&m(Z(ddl&m)Z)ddl&m*Z*ddl&m+Z+dd l&m,Z,dd!l&m-Z-dd"l&m.Z.dd"l&m.Z/dd#l0m1Z1dd$l0m2Z2dd%l3m4Z4dd&l3m5Z5dd'l6m7Z7dd(l6m8Z8dZ9ej:�r4ddl;Z;dd)l&m<Z<dd*lm=Z=ej>d+ej?d,�Z@e�>d-�ZAejBe@e@d.�d/d0��ZejBejCejCej?e@ge@fd1�d2d0��ZejCejCejDe@ej?e@ge@ffd1�d3d0�ZGd4d5�d5eE�ZFeG�ZHGd6d7�d7�ZIGd8d9�d9eJ�ZKGd:d;�d;�ZLGd<d=�d=eL�ZMeG�ZNejDejCeGfejCejCd>�d?d@�ZOGdAdB�dBeE�ZPGdCdD�dD�ZQGdEdF�dFe�ZRGdGdH�dHeR�ZSGdIdJ�dJejT�ZTGdKdL�dL�ZUddMlVmWZWe9�r�ddNlXmYZYdS)O�N)�ABC�abstractmethod)�futures)�iscoroutinefunction�)�
retry_base)�	retry_all)�retry_always)�	retry_any)�retry_if_exception)�retry_if_exception_type)�retry_if_not_exception_type)�retry_if_not_result)�retry_if_result)�retry_never)�retry_unless_exception_type)�retry_if_exception_message)�retry_if_not_exception_message��sleep)�sleep_using_event)�stop_after_attempt)�stop_after_delay)�stop_all)�stop_any)�
stop_never)�stop_when_event_set)�
wait_chain)�wait_combine)�wait_exponential)�
wait_fixed)�wait_incrementing)�	wait_none)�wait_random)�wait_random_exponential)�
before_log)�before_nothing)�	after_log)�
after_nothing)�before_sleep_log)�before_sleep_nothing)�	wait_base)�	stop_base�	WrappedFn)�bound�_RetValT)�fn�returncCsdS�N�)r0r3r3��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/__init__.py�retryasr5)�dargs�dkwr1cOsdSr2r3�r6r7r3r3r4r5fscsDt��dkr&t�d�r&t��d�Sttd���fdd�}|SdS)z�Wrap a function with a new `Retrying` object.

    :param dargs: positional arguments passed to Retrying object
    :param dkw: keyword arguments passed to the Retrying object
    rr��fr1cs�t|t�r*t�d|jj�d|jj�d��t|�rBt�i���}n:trnt	tj
d�rntj
�|�rnt�i���}nt
�i���}|�|�S)NzGot retry_base instance (zQ) as callable argument, this will probably hang indefinitely (did you mean retry=z(...)?)�is_coroutine_function)�
isinstancer�warnings�warn�	__class__�__name__r�
AsyncRetrying�tornado�hasattr�genr;�TornadoRetrying�Retrying�wraps)r:�rr8r3r4�wrapvs
��zretry.<locals>.wrapN)�len�callabler5r-)r6r7rIr3r8r4r5ksc@seZdZdZdS)�TryAgainz/Always retry the executed function when raised.N)r@�
__module__�__qualname__�__doc__r3r3r3r4rL�srLc@seZdZdS)�	DoAttemptN�r@rMrNr3r3r3r4rP�srPc@seZdZdS)�DoSleepNrQr3r3r3r4rR�srRc@sReZdZUdZdZejeed<dZ	ej
eed<ed�dd�Zed�d	d
�ZdS)�
BaseActiona5Base class for representing actions to take by retry object.

    Concrete implementations must define:
    - __init__: to initialize all necessary fields
    - REPR_FIELDS: class variable specifying attributes to include in repr(self)
    - NAME: for identification in retry object methods and callbacks
    r3�REPR_FIELDSN�NAME�r1cs.d��fdd��jD��}�jj�d|�d�S)Nz, c3s"|]}|�dt�|���VqdS)�=N)�getattr)�.0�field��selfr3r4�	<genexpr>��z&BaseAction.__repr__.<locals>.<genexpr>�(�))�joinrTr?r@)r\�	state_strr3r[r4�__repr__�szBaseAction.__repr__cCst|�Sr2)�reprr[r3r3r4�__str__�szBaseAction.__str__)
r@rMrNrOrT�t�Sequence�str�__annotations__rU�Optionalrcrer3r3r3r4rS�s

rSc@s&eZdZdZdZejdd�dd�ZdS)�RetryActionrr5N)rr1cCst|�|_dSr2)�floatr)r\rr3r3r4�__init__�szRetryAction.__init__)r@rMrNrTrUrf�
SupportsFloatrmr3r3r3r4rk�srk)�first�secondr1cCs|tur|S|Sr2)�_unset)rorpr3r3r4�
_first_set�srrcsDeZdZdZddd��fdd�Zdd�d	d
�Zed�dd�Z�ZS)
�
RetryErrorz>Encapsulates the last attempt instance right before giving up.�FutureN)�last_attemptr1cs||_t��|�dSr2)ru�superrm)r\ru�r?r3r4rm�szRetryError.__init__z
t.NoReturnrVcCs|jjr|j���|�dSr2)ru�failed�resultr[r3r3r4�reraise�s
zRetryError.reraisecCs|jj�d|j�d�S)N�[�])r?r@rur[r3r3r4re�szRetryError.__str__)	r@rMrNrOrmrzrhre�
__classcell__r3r3rwr4rs�srsc@s^eZdZdZdd�dd�Zdd�dd	�Zejeje	eje	ejd
eje
d�dd
�ZdS)�AttemptManagerzManage attempt context.�RetryCallState��retry_statecCs
||_dSr2r�)r\r�r3r3r4rm�szAttemptManager.__init__NrVcCsdSr2r3r[r3r3r4�	__enter__�szAttemptManager.__enter__�types.TracebackType)�exc_type�	exc_value�	tracebackr1cCs4t|t�r |j�|||f�dS|j�d�dSdS)NT)r<�
BaseExceptionr��
set_exception�
set_result)r\r�r�r�r3r3r4�__exit__�s

zAttemptManager.__exit__)r@rMrNrOrmr�rfrj�Typer��boolr�r3r3r3r4r~�s�r~c@s>eZdZeee�e�eedde	df
e
je
je
efgdfddee
jdgdfe
jdgdfe
je
jdgdfee
je	e
je
jdge
jfd�
dd�Zeeeeeeeeeef
e
je
je
je
efgdfefe
jdefe
jdefe
jeefe
je
jdgdfefe
je
jdgdfefe
je
je
jdgdfefe
jeefe
je
je	efe
je
je
jdge
jfefdd	�d
d�Zed�d
d�Zee
jee
jfd�dd��Zeed�dd�Zdd�dd�Zde
je e!e
jfd�dd�Z"e
j#e$ddfd�dd�Z%e&e
jde'fe
je
je'd�dd��Z(dS)�BaseRetryingNFr,r+r�
r�stop�waitr5�before�after�before_sleeprz�retry_error_cls�retry_error_callbackcCsJ||_||_||_||_||_||_||_||_t�	�|_
|	|_|
|_dSr2)
rr�r�r5r�r�r�rz�	threading�local�_localr�r��r\rr�r�r5r�r�r�rzr�r�r3r3r4rm�s

zBaseRetrying.__init__)rr�r�r5r�r�r�rzr�r�r1c
Csn|jt||j�t||j�t||j�t||j�t||j�t||j�t||j�t||j	�t|	|j
�t|
|j�d�
S)z8Copy this object with some parameters changed if needed.r�)r?rrrr�r�r5r�r�r�rzr�r�r�r3r3r4�copy�s









�zBaseRetrying.copyrVcCsLd|jj�dt|�d�d|j�d|j�d|j�d|j�d|j�d	|j�d
�S)N�<z
 object at 0x�xz (stop=z, wait=z, sleep=z, retry=z	, before=z, after=z)>)	r?r@�idr�r�rr5r�r�r[r3r3r4rcs�������zBaseRetrying.__repr__cCs2z
|jjWSty,i|j_|jjYS0dS)aSReturn a dictionary of runtime statistics.

        This dictionary will be empty when the controller has never been
        ran. When it is running or has ran previously it should have (but
        may not) have useful and/or informational keys and values when
        running is underway and/or completed.

        .. warning:: The keys in this dictionary **should** be some what
                     stable (not changing), but there existence **may**
                     change between major releases as new statistics are
                     gathered or removed so before accessing keys ensure that
                     they actually exist and handle when they do not.

        .. note:: The values in this dictionary are local to the thread
                  running call (so if multiple threads share the same retrying
                  object - either directly or indirectly) they will each have
                  there own view of statistics they have collected (in the
                  future we may provide a way to aggregate the various
                  statistics from each thread).
        N)r��
statistics�AttributeErrorr[r3r3r4r�"s

zBaseRetrying.statisticsr9csTt���tjtjtjd���fdd��}tjtjtd���fdd�}�|_||_|S)z[Wrap a function for retrying.

        :param f: A function to wraps for retrying.
        )�args�kwr1cs��g|�Ri|��Sr2r3)r�r��r:r\r3r4�	wrapped_fDsz%BaseRetrying.wraps.<locals>.wrapped_f)r��kwargsr1cs�j|i|�����Sr2)r�rG)r�r�r�r3r4�
retry_withHsz&BaseRetrying.wraps.<locals>.retry_with)�	functoolsrGrf�Anyr-r5r�)r\r:r�r�r3r�r4rG>s zBaseRetrying.wrapscCs0|j��t��|jd<d|jd<d|jd<dS)N�
start_timer�attempt_numberr�idle_for)r��clear�time�	monotonicr[r3r3r4�beginPs

zBaseRetrying.begin)r�r1cCs(|j}|dur(|jdur"|�|�t�S|jjo>t|j��t�}|sX|j|d�sX|��S|j	durl|�	|�|j
|jd<|j|d�r�|j
r�|�
|�S|�|�}|jr�|���||���|jr�|j|d�}nd}t|�|_|j|7_|jd|7<|jdd7<|jdu�r |�|�t|�S)Nr��delay_since_first_attempt�r�r�r)�outcomer�rPrxr<�	exceptionrLr5ryr��seconds_since_startr�r�r�r�rzr�rk�next_actionr�r�rR)r\r��fut�is_explicit_retry�	retry_excrr3r3r4�iterVs8








zBaseRetrying.iterccsb|��t|ddid�}|j|d�}t|t�r<t|d�Vqt|t�r^|��|�|�qq^qdS)Nr3)r0r�r�r�)	r�rr�r<rPr~rR�prepare_for_next_attemptr)r\r��dor3r3r4�__iter__{s

zBaseRetrying.__iter__.�r0r�r�r1cOsdSr2r3)r\r0r�r�r3r3r4�__call__�szBaseRetrying.__call__))r@rMrNrrr"rr&r(rsrf�Callable�Union�intrlrrjr�r�r�rmrq�objectr�rhrc�property�Dictr�r-rGr�rPrRr��	Generatorr~r�rr/r�r3r3r3r4r��sl���" �%r�c@s2eZdZdZejdefejejed�dd�ZdS)rFzRetrying controller..r�cOs�|��t||||d�}|j|d�}t|t�rnz||i|��}Wn ty`|�t���Yq�0|�	|�qt|t
�r�|��|�|�q|SqdS)N)�retry_objectr0r�r�r�)
r�rr�r<rPr�r��sys�exc_infor�rRr�r)r\r0r�r�r�r�ryr3r3r4r��s

zRetrying.__call__N)	r@rMrNrOrfr�r/r�r�r3r3r3r4rF�srFcsTeZdZdZedd��fdd�Zeed�dd��Ze	ee
jedd	�d
d��Z�Z
S)rtzDEncapsulates a (future or past) attempted call to a target function.N)r�r1cst���||_dSr2)rvrmr�)r\r�rwr3r4rm�s
zFuture.__init__rVcCs|��duS)z8Return whether a exception is being held in this future.N)r�r[r3r3r4rx�sz
Future.failed)r��value�
has_exceptionr1cCs&||�}|r|�|�n
|�|�|S)zConstruct a new Future object.)r�r�)�clsr�r�r�r�r3r3r4�	construct�s

zFuture.construct)r@rMrNrOr�rmr�r�rx�classmethodrfr�r�r}r3r3rwr4rt�srtc@s�eZdZdZeejeejejdd�dd�Z	e
ejed�dd��Zdd�d	d
�Z
ejdd�dd
�Zejejeedfdd�dd�Zdd�ZdS)rz5State related to a single call wrapped with Retrying.N)r�r0r�r�r1cCsDt��|_||_||_||_||_d|_d|_d|_	d|_
d|_dS)Nrr�)r�r�r�r�r0r�r�r�r��outcome_timestampr�r�)r\r�r0r�r�r3r3r4rm�s
zRetryCallState.__init__rVcCs|jdurdS|j|jSr2)r�r�r[r3r3r4r��s
z"RetryCallState.seconds_since_startcCs$d|_d|_|jd7_d|_dS�Nr)r�r�r�r�r[r3r3r4r��sz'RetryCallState.prepare_for_next_attempt)�valr1cCs.t��}t|j�}|�|�|||_|_dSr2)r�r�rtr�r�r�r�)r\r��tsr�r3r3r4r��s

zRetryCallState.set_resultr�)r�r1cCs2t��}t|j�}|�|d�|||_|_dSr�)r�r�rtr�r�r�r�)r\r�r�r�r3r3r4r��s
zRetryCallState.set_exceptioncCs�|jdurd}n:|jjr:|j��}d|jj�d|�d�}nd|j����}tt|jd��}|jj}d|�dt	|��d|j
�d	|�d
|�d�S)Nznone yetzfailed (� r`z	returned �r�z: attempt #z; slept for z; last result: �>)r�rxr�r?r@ryrl�roundr�r�r�)r\ryr��slept�clsnamer3r3r4rc�s

zRetryCallState.__repr__)r@rMrNrOr�rfrjr-r�rmr�rlr�r�r��Tupler�r�r�rcr3r3r3r4r�s�"r)rA)rE)Zr�r�r�r��typingrfr=�abcrr�
concurrentr�inspectrr5rrr	r
rrr
rrrrrr�naprrr�rrrrrrr�rrrr r!r"r#r$�wait_full_jitterr�r%r&r�r'r(r�r)r*rB�
TYPE_CHECKING�typesr+r,�TypeVarr�r-r/�overloadr�r��	ExceptionrLr��	NO_RESULTrPrlrRrSrkrqrrrsr~r�rFrtr�pip._vendor.tenacity._asynciorAZpip._vendor.tenacity.tornadowebrEr3r3r3r4�<module>s�
$,"-Fsite-packages/pip/_vendor/tenacity/__pycache__/before.cpython-39.pyc000064400000002161150564745010021442 0ustar00a

��Re`�@s`ddlZddlmZejr.ddlZddlmZddd�dd�Zdeejdgdfd	�d
d�Z	dS)�N)�_utils)�RetryCallStater��retry_state�returncCsdS)z'Before call strategy that does nothing.N��rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/before.py�before_nothingsr
zlogging.Logger)�logger�	log_levelrcsddd���fdd�}|S)z:Before call strategy that logs to some logger the attempt.rNrc	s.���dt�|j��dt�|j��d��dS)NzStarting call to 'z', this is the z time calling it.)�logr�get_callback_name�fn�
to_ordinal�attempt_numberr�rrrr	�log_it"s
��zbefore_log.<locals>.log_itr)rrrrrr	�
before_logsr)
�typing�pip._vendor.tenacityr�
TYPE_CHECKING�loggingrr
�int�Callablerrrrr	�<module>ssite-packages/pip/_vendor/tenacity/__pycache__/_utils.cpython-39.pyc000064400000002366150564745010021506 0ustar00a

��Re��@sZddlZddlZejdZeed�dd�Zeed�dd�Zejdej	fed	�d
d�Z
dS)�N�)�pos_num�returncCs\|dkrdS|dkrdS|dkr$dS|dkr0dSd	|krDd
krLnndSt|d�SdS)Nr�th��str�nd��rd���
��find_ordinal�r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/tenacity/_utils.pyrsrcCs|�t|���S)Nrrrrr�
to_ordinal*sr.)�cbrcCs�g}z|�|j�Wn6tyJz|�|j�WntyDYn0Yn0|sXt|�Sz|jrn|�d|j�Wnty�Yn0d�|�SdS)znGet a callback fully-qualified name.

    If no name can be produced ``repr(cb)`` is called and returned.
    r�.N)�append�__qualname__�AttributeError�__name__�repr�
__module__�insert�join)r�segmentsrrr�get_callback_name.s r)�sys�typing�maxsize�MAX_WAIT�int�strrr�Callable�Anyrrrrr�<module>s

site-packages/pip/_vendor/tenacity/before_sleep.py000064400000003564150564745010016353 0ustar00# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import typing

from pip._vendor.tenacity import _utils

if typing.TYPE_CHECKING:
    import logging

    from pip._vendor.tenacity import RetryCallState


def before_sleep_nothing(retry_state: "RetryCallState") -> None:
    """Before call strategy that does nothing."""


def before_sleep_log(
    logger: "logging.Logger",
    log_level: int,
    exc_info: bool = False,
) -> typing.Callable[["RetryCallState"], None]:
    """Before call strategy that logs to some logger the attempt."""

    def log_it(retry_state: "RetryCallState") -> None:
        if retry_state.outcome.failed:
            ex = retry_state.outcome.exception()
            verb, value = "raised", f"{ex.__class__.__name__}: {ex}"

            if exc_info:
                local_exc_info = retry_state.outcome.exception()
            else:
                local_exc_info = False
        else:
            verb, value = "returned", retry_state.outcome.result()
            local_exc_info = False  # exc_info does not apply when no exception

        logger.log(
            log_level,
            f"Retrying {_utils.get_callback_name(retry_state.fn)} "
            f"in {retry_state.next_action.sleep} seconds as it {verb} {value}.",
            exc_info=local_exc_info,
        )

    return log_it
site-packages/pip/_vendor/tenacity/__init__.py000064400000043521150564745010015455 0ustar00# Copyright 2016-2018 Julien Danjou
# Copyright 2017 Elisey Zanko
# Copyright 2016 Étienne Bersac
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import functools
import sys
import threading
import time
import typing as t
import warnings
from abc import ABC, abstractmethod
from concurrent import futures
from inspect import iscoroutinefunction

# Import all built-in retry strategies for easier usage.
from .retry import retry_base  # noqa
from .retry import retry_all  # noqa
from .retry import retry_always  # noqa
from .retry import retry_any  # noqa
from .retry import retry_if_exception  # noqa
from .retry import retry_if_exception_type  # noqa
from .retry import retry_if_not_exception_type  # noqa
from .retry import retry_if_not_result  # noqa
from .retry import retry_if_result  # noqa
from .retry import retry_never  # noqa
from .retry import retry_unless_exception_type  # noqa
from .retry import retry_if_exception_message  # noqa
from .retry import retry_if_not_exception_message  # noqa

# Import all nap strategies for easier usage.
from .nap import sleep  # noqa
from .nap import sleep_using_event  # noqa

# Import all built-in stop strategies for easier usage.
from .stop import stop_after_attempt  # noqa
from .stop import stop_after_delay  # noqa
from .stop import stop_all  # noqa
from .stop import stop_any  # noqa
from .stop import stop_never  # noqa
from .stop import stop_when_event_set  # noqa

# Import all built-in wait strategies for easier usage.
from .wait import wait_chain  # noqa
from .wait import wait_combine  # noqa
from .wait import wait_exponential  # noqa
from .wait import wait_fixed  # noqa
from .wait import wait_incrementing  # noqa
from .wait import wait_none  # noqa
from .wait import wait_random  # noqa
from .wait import wait_random_exponential  # noqa
from .wait import wait_random_exponential as wait_full_jitter  # noqa

# Import all built-in before strategies for easier usage.
from .before import before_log  # noqa
from .before import before_nothing  # noqa

# Import all built-in after strategies for easier usage.
from .after import after_log  # noqa
from .after import after_nothing  # noqa

# Import all built-in after strategies for easier usage.
from .before_sleep import before_sleep_log  # noqa
from .before_sleep import before_sleep_nothing  # noqa

# Replace a conditional import with a hard-coded None so that pip does
# not attempt to use tornado even if it is present in the environment.
# If tornado is non-None, tenacity will attempt to execute some code
# that is sensitive to the version of tornado, which could break pip
# if an old version is found.
tornado = None  # type: ignore

if t.TYPE_CHECKING:
    import types

    from .wait import wait_base
    from .stop import stop_base


WrappedFn = t.TypeVar("WrappedFn", bound=t.Callable)
_RetValT = t.TypeVar("_RetValT")


@t.overload
def retry(fn: WrappedFn) -> WrappedFn:
    pass


@t.overload
def retry(*dargs: t.Any, **dkw: t.Any) -> t.Callable[[WrappedFn], WrappedFn]:  # noqa
    pass


def retry(*dargs: t.Any, **dkw: t.Any) -> t.Union[WrappedFn, t.Callable[[WrappedFn], WrappedFn]]:  # noqa
    """Wrap a function with a new `Retrying` object.

    :param dargs: positional arguments passed to Retrying object
    :param dkw: keyword arguments passed to the Retrying object
    """
    # support both @retry and @retry() as valid syntax
    if len(dargs) == 1 and callable(dargs[0]):
        return retry()(dargs[0])
    else:

        def wrap(f: WrappedFn) -> WrappedFn:
            if isinstance(f, retry_base):
                warnings.warn(
                    f"Got retry_base instance ({f.__class__.__name__}) as callable argument, "
                    f"this will probably hang indefinitely (did you mean retry={f.__class__.__name__}(...)?)"
                )
            if iscoroutinefunction(f):
                r: "BaseRetrying" = AsyncRetrying(*dargs, **dkw)
            elif tornado and hasattr(tornado.gen, "is_coroutine_function") and tornado.gen.is_coroutine_function(f):
                r = TornadoRetrying(*dargs, **dkw)
            else:
                r = Retrying(*dargs, **dkw)

            return r.wraps(f)

        return wrap


class TryAgain(Exception):
    """Always retry the executed function when raised."""


NO_RESULT = object()


class DoAttempt:
    pass


class DoSleep(float):
    pass


class BaseAction:
    """Base class for representing actions to take by retry object.

    Concrete implementations must define:
    - __init__: to initialize all necessary fields
    - REPR_FIELDS: class variable specifying attributes to include in repr(self)
    - NAME: for identification in retry object methods and callbacks
    """

    REPR_FIELDS: t.Sequence[str] = ()
    NAME: t.Optional[str] = None

    def __repr__(self) -> str:
        state_str = ", ".join(f"{field}={getattr(self, field)!r}" for field in self.REPR_FIELDS)
        return f"{self.__class__.__name__}({state_str})"

    def __str__(self) -> str:
        return repr(self)


class RetryAction(BaseAction):
    REPR_FIELDS = ("sleep",)
    NAME = "retry"

    def __init__(self, sleep: t.SupportsFloat) -> None:
        self.sleep = float(sleep)


_unset = object()


def _first_set(first: t.Union[t.Any, object], second: t.Any) -> t.Any:
    return second if first is _unset else first


class RetryError(Exception):
    """Encapsulates the last attempt instance right before giving up."""

    def __init__(self, last_attempt: "Future") -> None:
        self.last_attempt = last_attempt
        super().__init__(last_attempt)

    def reraise(self) -> "t.NoReturn":
        if self.last_attempt.failed:
            raise self.last_attempt.result()
        raise self

    def __str__(self) -> str:
        return f"{self.__class__.__name__}[{self.last_attempt}]"


class AttemptManager:
    """Manage attempt context."""

    def __init__(self, retry_state: "RetryCallState"):
        self.retry_state = retry_state

    def __enter__(self) -> None:
        pass

    def __exit__(
        self,
        exc_type: t.Optional[t.Type[BaseException]],
        exc_value: t.Optional[BaseException],
        traceback: t.Optional["types.TracebackType"],
    ) -> t.Optional[bool]:
        if isinstance(exc_value, BaseException):
            self.retry_state.set_exception((exc_type, exc_value, traceback))
            return True  # Swallow exception.
        else:
            # We don't have the result, actually.
            self.retry_state.set_result(None)
            return None


class BaseRetrying(ABC):
    def __init__(
        self,
        sleep: t.Callable[[t.Union[int, float]], None] = sleep,
        stop: "stop_base" = stop_never,
        wait: "wait_base" = wait_none(),
        retry: retry_base = retry_if_exception_type(),
        before: t.Callable[["RetryCallState"], None] = before_nothing,
        after: t.Callable[["RetryCallState"], None] = after_nothing,
        before_sleep: t.Optional[t.Callable[["RetryCallState"], None]] = None,
        reraise: bool = False,
        retry_error_cls: t.Type[RetryError] = RetryError,
        retry_error_callback: t.Optional[t.Callable[["RetryCallState"], t.Any]] = None,
    ):
        self.sleep = sleep
        self.stop = stop
        self.wait = wait
        self.retry = retry
        self.before = before
        self.after = after
        self.before_sleep = before_sleep
        self.reraise = reraise
        self._local = threading.local()
        self.retry_error_cls = retry_error_cls
        self.retry_error_callback = retry_error_callback

    def copy(
        self,
        sleep: t.Union[t.Callable[[t.Union[int, float]], None], object] = _unset,
        stop: t.Union["stop_base", object] = _unset,
        wait: t.Union["wait_base", object] = _unset,
        retry: t.Union[retry_base, object] = _unset,
        before: t.Union[t.Callable[["RetryCallState"], None], object] = _unset,
        after: t.Union[t.Callable[["RetryCallState"], None], object] = _unset,
        before_sleep: t.Union[t.Optional[t.Callable[["RetryCallState"], None]], object] = _unset,
        reraise: t.Union[bool, object] = _unset,
        retry_error_cls: t.Union[t.Type[RetryError], object] = _unset,
        retry_error_callback: t.Union[t.Optional[t.Callable[["RetryCallState"], t.Any]], object] = _unset,
    ) -> "BaseRetrying":
        """Copy this object with some parameters changed if needed."""
        return self.__class__(
            sleep=_first_set(sleep, self.sleep),
            stop=_first_set(stop, self.stop),
            wait=_first_set(wait, self.wait),
            retry=_first_set(retry, self.retry),
            before=_first_set(before, self.before),
            after=_first_set(after, self.after),
            before_sleep=_first_set(before_sleep, self.before_sleep),
            reraise=_first_set(reraise, self.reraise),
            retry_error_cls=_first_set(retry_error_cls, self.retry_error_cls),
            retry_error_callback=_first_set(retry_error_callback, self.retry_error_callback),
        )

    def __repr__(self) -> str:
        return (
            f"<{self.__class__.__name__} object at 0x{id(self):x} ("
            f"stop={self.stop}, "
            f"wait={self.wait}, "
            f"sleep={self.sleep}, "
            f"retry={self.retry}, "
            f"before={self.before}, "
            f"after={self.after})>"
        )

    @property
    def statistics(self) -> t.Dict[str, t.Any]:
        """Return a dictionary of runtime statistics.

        This dictionary will be empty when the controller has never been
        ran. When it is running or has ran previously it should have (but
        may not) have useful and/or informational keys and values when
        running is underway and/or completed.

        .. warning:: The keys in this dictionary **should** be some what
                     stable (not changing), but there existence **may**
                     change between major releases as new statistics are
                     gathered or removed so before accessing keys ensure that
                     they actually exist and handle when they do not.

        .. note:: The values in this dictionary are local to the thread
                  running call (so if multiple threads share the same retrying
                  object - either directly or indirectly) they will each have
                  there own view of statistics they have collected (in the
                  future we may provide a way to aggregate the various
                  statistics from each thread).
        """
        try:
            return self._local.statistics
        except AttributeError:
            self._local.statistics = {}
            return self._local.statistics

    def wraps(self, f: WrappedFn) -> WrappedFn:
        """Wrap a function for retrying.

        :param f: A function to wraps for retrying.
        """

        @functools.wraps(f)
        def wrapped_f(*args: t.Any, **kw: t.Any) -> t.Any:
            return self(f, *args, **kw)

        def retry_with(*args: t.Any, **kwargs: t.Any) -> WrappedFn:
            return self.copy(*args, **kwargs).wraps(f)

        wrapped_f.retry = self
        wrapped_f.retry_with = retry_with

        return wrapped_f

    def begin(self) -> None:
        self.statistics.clear()
        self.statistics["start_time"] = time.monotonic()
        self.statistics["attempt_number"] = 1
        self.statistics["idle_for"] = 0

    def iter(self, retry_state: "RetryCallState") -> t.Union[DoAttempt, DoSleep, t.Any]:  # noqa
        fut = retry_state.outcome
        if fut is None:
            if self.before is not None:
                self.before(retry_state)
            return DoAttempt()

        is_explicit_retry = retry_state.outcome.failed and isinstance(retry_state.outcome.exception(), TryAgain)
        if not (is_explicit_retry or self.retry(retry_state=retry_state)):
            return fut.result()

        if self.after is not None:
            self.after(retry_state)

        self.statistics["delay_since_first_attempt"] = retry_state.seconds_since_start
        if self.stop(retry_state=retry_state):
            if self.retry_error_callback:
                return self.retry_error_callback(retry_state)
            retry_exc = self.retry_error_cls(fut)
            if self.reraise:
                raise retry_exc.reraise()
            raise retry_exc from fut.exception()

        if self.wait:
            sleep = self.wait(retry_state=retry_state)
        else:
            sleep = 0.0
        retry_state.next_action = RetryAction(sleep)
        retry_state.idle_for += sleep
        self.statistics["idle_for"] += sleep
        self.statistics["attempt_number"] += 1

        if self.before_sleep is not None:
            self.before_sleep(retry_state)

        return DoSleep(sleep)

    def __iter__(self) -> t.Generator[AttemptManager, None, None]:
        self.begin()

        retry_state = RetryCallState(self, fn=None, args=(), kwargs={})
        while True:
            do = self.iter(retry_state=retry_state)
            if isinstance(do, DoAttempt):
                yield AttemptManager(retry_state=retry_state)
            elif isinstance(do, DoSleep):
                retry_state.prepare_for_next_attempt()
                self.sleep(do)
            else:
                break

    @abstractmethod
    def __call__(self, fn: t.Callable[..., _RetValT], *args: t.Any, **kwargs: t.Any) -> _RetValT:
        pass


class Retrying(BaseRetrying):
    """Retrying controller."""

    def __call__(self, fn: t.Callable[..., _RetValT], *args: t.Any, **kwargs: t.Any) -> _RetValT:
        self.begin()

        retry_state = RetryCallState(retry_object=self, fn=fn, args=args, kwargs=kwargs)
        while True:
            do = self.iter(retry_state=retry_state)
            if isinstance(do, DoAttempt):
                try:
                    result = fn(*args, **kwargs)
                except BaseException:  # noqa: B902
                    retry_state.set_exception(sys.exc_info())
                else:
                    retry_state.set_result(result)
            elif isinstance(do, DoSleep):
                retry_state.prepare_for_next_attempt()
                self.sleep(do)
            else:
                return do


class Future(futures.Future):
    """Encapsulates a (future or past) attempted call to a target function."""

    def __init__(self, attempt_number: int) -> None:
        super().__init__()
        self.attempt_number = attempt_number

    @property
    def failed(self) -> bool:
        """Return whether a exception is being held in this future."""
        return self.exception() is not None

    @classmethod
    def construct(cls, attempt_number: int, value: t.Any, has_exception: bool) -> "Future":
        """Construct a new Future object."""
        fut = cls(attempt_number)
        if has_exception:
            fut.set_exception(value)
        else:
            fut.set_result(value)
        return fut


class RetryCallState:
    """State related to a single call wrapped with Retrying."""

    def __init__(
        self,
        retry_object: BaseRetrying,
        fn: t.Optional[WrappedFn],
        args: t.Any,
        kwargs: t.Any,
    ) -> None:
        #: Retry call start timestamp
        self.start_time = time.monotonic()
        #: Retry manager object
        self.retry_object = retry_object
        #: Function wrapped by this retry call
        self.fn = fn
        #: Arguments of the function wrapped by this retry call
        self.args = args
        #: Keyword arguments of the function wrapped by this retry call
        self.kwargs = kwargs

        #: The number of the current attempt
        self.attempt_number: int = 1
        #: Last outcome (result or exception) produced by the function
        self.outcome: t.Optional[Future] = None
        #: Timestamp of the last outcome
        self.outcome_timestamp: t.Optional[float] = None
        #: Time spent sleeping in retries
        self.idle_for: float = 0.0
        #: Next action as decided by the retry manager
        self.next_action: t.Optional[RetryAction] = None

    @property
    def seconds_since_start(self) -> t.Optional[float]:
        if self.outcome_timestamp is None:
            return None
        return self.outcome_timestamp - self.start_time

    def prepare_for_next_attempt(self) -> None:
        self.outcome = None
        self.outcome_timestamp = None
        self.attempt_number += 1
        self.next_action = None

    def set_result(self, val: t.Any) -> None:
        ts = time.monotonic()
        fut = Future(self.attempt_number)
        fut.set_result(val)
        self.outcome, self.outcome_timestamp = fut, ts

    def set_exception(self, exc_info: t.Tuple[t.Type[BaseException], BaseException, "types.TracebackType"]) -> None:
        ts = time.monotonic()
        fut = Future(self.attempt_number)
        fut.set_exception(exc_info[1])
        self.outcome, self.outcome_timestamp = fut, ts

    def __repr__(self):
        if self.outcome is None:
            result = "none yet"
        elif self.outcome.failed:
            exception = self.outcome.exception()
            result = f"failed ({exception.__class__.__name__} {exception})"
        else:
            result = f"returned {self.outcome.result()}"

        slept = float(round(self.idle_for, 2))
        clsname = self.__class__.__name__
        return f"<{clsname} {id(self)}: attempt #{self.attempt_number}; slept for {slept}; last result: {result}>"


from pip._vendor.tenacity._asyncio import AsyncRetrying  # noqa:E402,I100

if tornado:
    from pip._vendor.tenacity.tornadoweb import TornadoRetrying
site-packages/pip/_vendor/tenacity/retry.py000064400000014765150564745010015073 0ustar00# Copyright 2016–2021 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import abc
import re
import typing

if typing.TYPE_CHECKING:
    from pip._vendor.tenacity import RetryCallState


class retry_base(abc.ABC):
    """Abstract base class for retry strategies."""

    @abc.abstractmethod
    def __call__(self, retry_state: "RetryCallState") -> bool:
        pass

    def __and__(self, other: "retry_base") -> "retry_all":
        return retry_all(self, other)

    def __or__(self, other: "retry_base") -> "retry_any":
        return retry_any(self, other)


class _retry_never(retry_base):
    """Retry strategy that never rejects any result."""

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return False


retry_never = _retry_never()


class _retry_always(retry_base):
    """Retry strategy that always rejects any result."""

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return True


retry_always = _retry_always()


class retry_if_exception(retry_base):
    """Retry strategy that retries if an exception verifies a predicate."""

    def __init__(self, predicate: typing.Callable[[BaseException], bool]) -> None:
        self.predicate = predicate

    def __call__(self, retry_state: "RetryCallState") -> bool:
        if retry_state.outcome.failed:
            return self.predicate(retry_state.outcome.exception())
        else:
            return False


class retry_if_exception_type(retry_if_exception):
    """Retries if an exception has been raised of one or more types."""

    def __init__(
        self,
        exception_types: typing.Union[
            typing.Type[BaseException],
            typing.Tuple[typing.Type[BaseException], ...],
        ] = Exception,
    ) -> None:
        self.exception_types = exception_types
        super().__init__(lambda e: isinstance(e, exception_types))


class retry_if_not_exception_type(retry_if_exception):
    """Retries except an exception has been raised of one or more types."""

    def __init__(
        self,
        exception_types: typing.Union[
            typing.Type[BaseException],
            typing.Tuple[typing.Type[BaseException], ...],
        ] = Exception,
    ) -> None:
        self.exception_types = exception_types
        super().__init__(lambda e: not isinstance(e, exception_types))


class retry_unless_exception_type(retry_if_exception):
    """Retries until an exception is raised of one or more types."""

    def __init__(
        self,
        exception_types: typing.Union[
            typing.Type[BaseException],
            typing.Tuple[typing.Type[BaseException], ...],
        ] = Exception,
    ) -> None:
        self.exception_types = exception_types
        super().__init__(lambda e: not isinstance(e, exception_types))

    def __call__(self, retry_state: "RetryCallState") -> bool:
        # always retry if no exception was raised
        if not retry_state.outcome.failed:
            return True
        return self.predicate(retry_state.outcome.exception())


class retry_if_result(retry_base):
    """Retries if the result verifies a predicate."""

    def __init__(self, predicate: typing.Callable[[typing.Any], bool]) -> None:
        self.predicate = predicate

    def __call__(self, retry_state: "RetryCallState") -> bool:
        if not retry_state.outcome.failed:
            return self.predicate(retry_state.outcome.result())
        else:
            return False


class retry_if_not_result(retry_base):
    """Retries if the result refutes a predicate."""

    def __init__(self, predicate: typing.Callable[[typing.Any], bool]) -> None:
        self.predicate = predicate

    def __call__(self, retry_state: "RetryCallState") -> bool:
        if not retry_state.outcome.failed:
            return not self.predicate(retry_state.outcome.result())
        else:
            return False


class retry_if_exception_message(retry_if_exception):
    """Retries if an exception message equals or matches."""

    def __init__(
        self,
        message: typing.Optional[str] = None,
        match: typing.Optional[str] = None,
    ) -> None:
        if message and match:
            raise TypeError(f"{self.__class__.__name__}() takes either 'message' or 'match', not both")

        # set predicate
        if message:

            def message_fnc(exception: BaseException) -> bool:
                return message == str(exception)

            predicate = message_fnc
        elif match:
            prog = re.compile(match)

            def match_fnc(exception: BaseException) -> bool:
                return bool(prog.match(str(exception)))

            predicate = match_fnc
        else:
            raise TypeError(f"{self.__class__.__name__}() missing 1 required argument 'message' or 'match'")

        super().__init__(predicate)


class retry_if_not_exception_message(retry_if_exception_message):
    """Retries until an exception message equals or matches."""

    def __init__(
        self,
        message: typing.Optional[str] = None,
        match: typing.Optional[str] = None,
    ) -> None:
        super().__init__(message, match)
        # invert predicate
        if_predicate = self.predicate
        self.predicate = lambda *args_, **kwargs_: not if_predicate(*args_, **kwargs_)

    def __call__(self, retry_state: "RetryCallState") -> bool:
        if not retry_state.outcome.failed:
            return True
        return self.predicate(retry_state.outcome.exception())


class retry_any(retry_base):
    """Retries if any of the retries condition is valid."""

    def __init__(self, *retries: retry_base) -> None:
        self.retries = retries

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return any(r(retry_state) for r in self.retries)


class retry_all(retry_base):
    """Retries if all the retries condition are valid."""

    def __init__(self, *retries: retry_base) -> None:
        self.retries = retries

    def __call__(self, retry_state: "RetryCallState") -> bool:
        return all(r(retry_state) for r in self.retries)
site-packages/pip/_vendor/tenacity/before.py000064400000002540150564745010015154 0ustar00# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import typing

from pip._vendor.tenacity import _utils

if typing.TYPE_CHECKING:
    import logging

    from pip._vendor.tenacity import RetryCallState


def before_nothing(retry_state: "RetryCallState") -> None:
    """Before call strategy that does nothing."""


def before_log(logger: "logging.Logger", log_level: int) -> typing.Callable[["RetryCallState"], None]:
    """Before call strategy that logs to some logger the attempt."""

    def log_it(retry_state: "RetryCallState") -> None:
        logger.log(
            log_level,
            f"Starting call to '{_utils.get_callback_name(retry_state.fn)}', "
            f"this is the {_utils.to_ordinal(retry_state.attempt_number)} time calling it.",
        )

    return log_it
site-packages/pip/_vendor/tenacity/tornadoweb.py000064400000004141150564745010016055 0ustar00# Copyright 2017 Elisey Zanko
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import sys
import typing

from pip._vendor.tenacity import BaseRetrying
from pip._vendor.tenacity import DoAttempt
from pip._vendor.tenacity import DoSleep
from pip._vendor.tenacity import RetryCallState

from tornado import gen

if typing.TYPE_CHECKING:
    from tornado.concurrent import Future

_RetValT = typing.TypeVar("_RetValT")


class TornadoRetrying(BaseRetrying):
    def __init__(self, sleep: "typing.Callable[[float], Future[None]]" = gen.sleep, **kwargs: typing.Any) -> None:
        super().__init__(**kwargs)
        self.sleep = sleep

    @gen.coroutine
    def __call__(  # type: ignore  # Change signature from supertype
        self,
        fn: "typing.Callable[..., typing.Union[typing.Generator[typing.Any, typing.Any, _RetValT], Future[_RetValT]]]",
        *args: typing.Any,
        **kwargs: typing.Any,
    ) -> "typing.Generator[typing.Any, typing.Any, _RetValT]":
        self.begin()

        retry_state = RetryCallState(retry_object=self, fn=fn, args=args, kwargs=kwargs)
        while True:
            do = self.iter(retry_state=retry_state)
            if isinstance(do, DoAttempt):
                try:
                    result = yield fn(*args, **kwargs)
                except BaseException:  # noqa: B902
                    retry_state.set_exception(sys.exc_info())
                else:
                    retry_state.set_result(result)
            elif isinstance(do, DoSleep):
                retry_state.prepare_for_next_attempt()
                yield self.sleep(do)
            else:
                raise gen.Return(do)
site-packages/pip/_vendor/tenacity/nap.py000064400000002547150564745010014477 0ustar00# Copyright 2016 Étienne Bersac
# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import time
import typing

if typing.TYPE_CHECKING:
    import threading


def sleep(seconds: float) -> None:
    """
    Sleep strategy that delays execution for a given number of seconds.

    This is the default strategy, and may be mocked out for unit testing.
    """
    time.sleep(seconds)


class sleep_using_event:
    """Sleep strategy that waits on an event to be set."""

    def __init__(self, event: "threading.Event") -> None:
        self.event = event

    def __call__(self, timeout: typing.Optional[float]) -> None:
        # NOTE(harlowja): this may *not* actually wait for timeout
        # seconds if the event is set (ie this may eject out early).
        self.event.wait(timeout=timeout)
site-packages/pip/_vendor/tenacity/_utils.py000064400000003630150564745010015212 0ustar00# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import sys
import typing


# sys.maxsize:
# An integer giving the maximum value a variable of type Py_ssize_t can take.
MAX_WAIT = sys.maxsize / 2


def find_ordinal(pos_num: int) -> str:
    # See: https://en.wikipedia.org/wiki/English_numerals#Ordinal_numbers
    if pos_num == 0:
        return "th"
    elif pos_num == 1:
        return "st"
    elif pos_num == 2:
        return "nd"
    elif pos_num == 3:
        return "rd"
    elif 4 <= pos_num <= 20:
        return "th"
    else:
        return find_ordinal(pos_num % 10)


def to_ordinal(pos_num: int) -> str:
    return f"{pos_num}{find_ordinal(pos_num)}"


def get_callback_name(cb: typing.Callable[..., typing.Any]) -> str:
    """Get a callback fully-qualified name.

    If no name can be produced ``repr(cb)`` is called and returned.
    """
    segments = []
    try:
        segments.append(cb.__qualname__)
    except AttributeError:
        try:
            segments.append(cb.__name__)
        except AttributeError:
            pass
    if not segments:
        return repr(cb)
    else:
        try:
            # When running under sphinx it appears this can be none?
            if cb.__module__:
                segments.insert(0, cb.__module__)
        except AttributeError:
            pass
        return ".".join(segments)
site-packages/pip/_vendor/tenacity/_asyncio.py000064400000006362150564745010015524 0ustar00# Copyright 2016 Étienne Bersac
# Copyright 2016 Julien Danjou
# Copyright 2016 Joshua Harlow
# Copyright 2013-2014 Ray Holder
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

import functools
import sys
import typing
from asyncio import sleep

from pip._vendor.tenacity import AttemptManager
from pip._vendor.tenacity import BaseRetrying
from pip._vendor.tenacity import DoAttempt
from pip._vendor.tenacity import DoSleep
from pip._vendor.tenacity import RetryCallState

WrappedFn = typing.TypeVar("WrappedFn", bound=typing.Callable)
_RetValT = typing.TypeVar("_RetValT")


class AsyncRetrying(BaseRetrying):
    def __init__(self, sleep: typing.Callable[[float], typing.Awaitable] = sleep, **kwargs: typing.Any) -> None:
        super().__init__(**kwargs)
        self.sleep = sleep

    async def __call__(  # type: ignore  # Change signature from supertype
        self,
        fn: typing.Callable[..., typing.Awaitable[_RetValT]],
        *args: typing.Any,
        **kwargs: typing.Any,
    ) -> _RetValT:
        self.begin()

        retry_state = RetryCallState(retry_object=self, fn=fn, args=args, kwargs=kwargs)
        while True:
            do = self.iter(retry_state=retry_state)
            if isinstance(do, DoAttempt):
                try:
                    result = await fn(*args, **kwargs)
                except BaseException:  # noqa: B902
                    retry_state.set_exception(sys.exc_info())
                else:
                    retry_state.set_result(result)
            elif isinstance(do, DoSleep):
                retry_state.prepare_for_next_attempt()
                await self.sleep(do)
            else:
                return do

    def __aiter__(self) -> "AsyncRetrying":
        self.begin()
        self._retry_state = RetryCallState(self, fn=None, args=(), kwargs={})
        return self

    async def __anext__(self) -> typing.Union[AttemptManager, typing.Any]:
        while True:
            do = self.iter(retry_state=self._retry_state)
            if do is None:
                raise StopAsyncIteration
            elif isinstance(do, DoAttempt):
                return AttemptManager(retry_state=self._retry_state)
            elif isinstance(do, DoSleep):
                self._retry_state.prepare_for_next_attempt()
                await self.sleep(do)
            else:
                return do

    def wraps(self, fn: WrappedFn) -> WrappedFn:
        fn = super().wraps(fn)
        # Ensure wrapper is recognized as a coroutine function.

        @functools.wraps(fn)
        async def async_wrapped(*args: typing.Any, **kwargs: typing.Any) -> typing.Any:
            return await fn(*args, **kwargs)

        # Preserve attributes
        async_wrapped.retry = fn.retry
        async_wrapped.retry_with = fn.retry_with

        return async_wrapped
site-packages/pip/_vendor/cachecontrol/caches/__pycache__/file_cache.cpython-39.pyc000064400000006424150564745010024322 0ustar00a

��Re9�@s|ddlZddlZddlmZddlmZddlmZzeWne	yVe
efZYn0dd�ZGdd	�d	e�Z
d
d�ZdS)�N)�dedent�)�	BaseCache)�CacheControllerc	Cs�tj}|tjtjBO}ttd�r*|tjO}ttd�r>|tjO}zt�|�Wntt	fybYn0t�
|||�}zt�|d�WSt�|��Yn0dS)N�
O_NOFOLLOW�O_BINARY�wb)
�os�O_WRONLY�O_CREAT�O_EXCL�hasattrrr�remove�IOError�OSError�open�fdopen�close)�filename�fmode�flags�fd�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/caches/file_cache.py�_secure_open_writes 




rc@sBeZdZddd�Zedd��Zd	d
�Zdd�Zd
d�Zdd�Z	dS)�	FileCacheF��Nc
Cs�|dur|durtd��zddlm}ddlm}Wn"tyVtd�}	t|	��Yn0|rb|}n|durn|}||_||_||_	||_
||_dS)Nz/Cannot use use_dir_lock and lock_class togetherr)�LockFile)�
MkdirLockFilez�
            NOTE: In order to use the FileCache you must have
            lockfile installed. You can install it via pip:
              pip install lockfile
            )�
ValueError�lockfiler�lockfile.mkdirlockfiler�ImportErrorr�	directory�forever�filemode�dirmode�
lock_class)
�selfr$r%r&r'�use_dir_lockr(rr�noticerrr�__init__:s&
�zFileCache.__init__cCst�|�����S�N)�hashlib�sha224�encode�	hexdigest)�xrrrr0aszFileCache.encodecCs6|�|�}t|dd��|g}tjj|jg|�R�S)N�)r0�listr	�path�joinr$)r)�name�hashed�partsrrr�_fnes
z
FileCache._fncCs\|�|�}z8t|d��}|��Wd�WS1s60YWntyVYdS0dS)N�rb)r:r�read�FileNotFoundError)r)�keyr7�fhrrr�getls
,z
FileCache.getc	Cs�|�|�}zt�tj�|�|j�Wnttfy:Yn0|�|��H}t	|j|j
��}|�|�Wd�n1sv0YWd�n1s�0YdSr-)r:r	�makedirsr5�dirnamer'rrr(rr&�write)r)r>�valuer7�lockr?rrr�setus
z
FileCache.setcCs6|�|�}|js2zt�|�Wnty0Yn0dSr-)r:r%r	rr=)r)r>r7rrr�delete�s
zFileCache.delete)FrrNN)
�__name__�
__module__�__qualname__r,�staticmethodr0r:r@rFrGrrrrr8s�
'
	rcCst�|�}|�|�S)z\Return the file cache path based on the URL.

    This does not ensure the file exists!
    )r�	cache_urlr:)�url�	filecacher>rrr�url_to_file_path�s
rO)r.r	�textwrapr�cacher�
controllerrr=�	NameErrorrrrrrOrrrr�<module>s)Tsite-packages/pip/_vendor/cachecontrol/caches/__pycache__/redis_cache.cpython-39.pyc000064400000003104150564745010024501 0ustar00a

��ReX�@s8ddlmZddlmZddlmZGdd�de�ZdS)�)�division)�datetime)�	BaseCachec@s>eZdZdd�Zdd�Zddd�Zdd	�Zd
d�Zdd
�ZdS)�
RedisCachecCs
||_dS�N)�conn)�selfr�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/caches/redis_cache.py�__init__	szRedisCache.__init__cCs|j�|�Sr)r�get�r�keyr	r	r
rszRedisCache.getNcCs<|s|j�||�n$|t��}|j�|t|���|�dSr)r�setr�utcnow�setex�int�
total_seconds)rr�value�expiresr	r	r
rszRedisCache.setcCs|j�|�dSr)r�deleter
r	r	r
rszRedisCache.deletecCs |j��D]}|j�|�q
dS)zIHelper for clearing all the keys in a database. Use with
        caution!N)r�keysrr
r	r	r
�clearszRedisCache.clearcCsdS)z?Redis uses connection pooling, no need to close the connection.Nr	)rr	r	r
�closeszRedisCache.close)N)	�__name__�
__module__�__qualname__rrrrrrr	r	r	r
rs
rN)�
__future__rrZpip._vendor.cachecontrol.cacherrr	r	r	r
�<module>ssite-packages/pip/_vendor/cachecontrol/caches/__pycache__/__init__.cpython-39.pyc000064400000000514150564745010024011 0ustar00a

��ReV�@sddlmZddlmZdS)�)�	FileCache)�
RedisCacheN)�
file_cacher�redis_cacher�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/caches/__init__.py�<module>ssite-packages/pip/_vendor/cachecontrol/caches/file_cache.py000064400000010071150564745010020024 0ustar00import hashlib
import os
from textwrap import dedent

from ..cache import BaseCache
from ..controller import CacheController

try:
    FileNotFoundError
except NameError:
    # py2.X
    FileNotFoundError = (IOError, OSError)


def _secure_open_write(filename, fmode):
    # We only want to write to this file, so open it in write only mode
    flags = os.O_WRONLY

    # os.O_CREAT | os.O_EXCL will fail if the file already exists, so we only
    #  will open *new* files.
    # We specify this because we want to ensure that the mode we pass is the
    # mode of the file.
    flags |= os.O_CREAT | os.O_EXCL

    # Do not follow symlinks to prevent someone from making a symlink that
    # we follow and insecurely open a cache file.
    if hasattr(os, "O_NOFOLLOW"):
        flags |= os.O_NOFOLLOW

    # On Windows we'll mark this file as binary
    if hasattr(os, "O_BINARY"):
        flags |= os.O_BINARY

    # Before we open our file, we want to delete any existing file that is
    # there
    try:
        os.remove(filename)
    except (IOError, OSError):
        # The file must not exist already, so we can just skip ahead to opening
        pass

    # Open our file, the use of os.O_CREAT | os.O_EXCL will ensure that if a
    # race condition happens between the os.remove and this line, that an
    # error will be raised. Because we utilize a lockfile this should only
    # happen if someone is attempting to attack us.
    fd = os.open(filename, flags, fmode)
    try:
        return os.fdopen(fd, "wb")

    except:
        # An error occurred wrapping our FD in a file object
        os.close(fd)
        raise


class FileCache(BaseCache):

    def __init__(
        self,
        directory,
        forever=False,
        filemode=0o0600,
        dirmode=0o0700,
        use_dir_lock=None,
        lock_class=None,
    ):

        if use_dir_lock is not None and lock_class is not None:
            raise ValueError("Cannot use use_dir_lock and lock_class together")

        try:
            from lockfile import LockFile
            from lockfile.mkdirlockfile import MkdirLockFile
        except ImportError:
            notice = dedent(
                """
            NOTE: In order to use the FileCache you must have
            lockfile installed. You can install it via pip:
              pip install lockfile
            """
            )
            raise ImportError(notice)

        else:
            if use_dir_lock:
                lock_class = MkdirLockFile

            elif lock_class is None:
                lock_class = LockFile

        self.directory = directory
        self.forever = forever
        self.filemode = filemode
        self.dirmode = dirmode
        self.lock_class = lock_class

    @staticmethod
    def encode(x):
        return hashlib.sha224(x.encode()).hexdigest()

    def _fn(self, name):
        # NOTE: This method should not change as some may depend on it.
        #       See: https://github.com/ionrock/cachecontrol/issues/63
        hashed = self.encode(name)
        parts = list(hashed[:5]) + [hashed]
        return os.path.join(self.directory, *parts)

    def get(self, key):
        name = self._fn(key)
        try:
            with open(name, "rb") as fh:
                return fh.read()

        except FileNotFoundError:
            return None

    def set(self, key, value):
        name = self._fn(key)

        # Make sure the directory exists
        try:
            os.makedirs(os.path.dirname(name), self.dirmode)
        except (IOError, OSError):
            pass

        with self.lock_class(name) as lock:
            # Write our actual file
            with _secure_open_write(lock.path, self.filemode) as fh:
                fh.write(value)

    def delete(self, key):
        name = self._fn(key)
        if not self.forever:
            try:
                os.remove(name)
            except FileNotFoundError:
                pass


def url_to_file_path(url, filecache):
    """Return the file cache path based on the URL.

    This does not ensure the file exists!
    """
    key = CacheController.cache_url(url)
    return filecache._fn(key)
site-packages/pip/_vendor/cachecontrol/caches/redis_cache.py000064400000001530150564745010020213 0ustar00from __future__ import division

from datetime import datetime
from pip._vendor.cachecontrol.cache import BaseCache


class RedisCache(BaseCache):

    def __init__(self, conn):
        self.conn = conn

    def get(self, key):
        return self.conn.get(key)

    def set(self, key, value, expires=None):
        if not expires:
            self.conn.set(key, value)
        else:
            expires = expires - datetime.utcnow()
            self.conn.setex(key, int(expires.total_seconds()), value)

    def delete(self, key):
        self.conn.delete(key)

    def clear(self):
        """Helper for clearing all the keys in a database. Use with
        caution!"""
        for key in self.conn.keys():
            self.conn.delete(key)

    def close(self):
        """Redis uses connection pooling, no need to close the connection."""
        pass
site-packages/pip/_vendor/cachecontrol/caches/__init__.py000064400000000126150564745010017521 0ustar00from .file_cache import FileCache  # noqa
from .redis_cache import RedisCache  # noqa
site-packages/pip/_vendor/cachecontrol/adapter.py000064400000011422150564745010016155 0ustar00import types
import functools
import zlib

from pip._vendor.requests.adapters import HTTPAdapter

from .controller import CacheController
from .cache import DictCache
from .filewrapper import CallbackFileWrapper


class CacheControlAdapter(HTTPAdapter):
    invalidating_methods = {"PUT", "DELETE"}

    def __init__(
        self,
        cache=None,
        cache_etags=True,
        controller_class=None,
        serializer=None,
        heuristic=None,
        cacheable_methods=None,
        *args,
        **kw
    ):
        super(CacheControlAdapter, self).__init__(*args, **kw)
        self.cache = DictCache() if cache is None else cache
        self.heuristic = heuristic
        self.cacheable_methods = cacheable_methods or ("GET",)

        controller_factory = controller_class or CacheController
        self.controller = controller_factory(
            self.cache, cache_etags=cache_etags, serializer=serializer
        )

    def send(self, request, cacheable_methods=None, **kw):
        """
        Send a request. Use the request information to see if it
        exists in the cache and cache the response if we need to and can.
        """
        cacheable = cacheable_methods or self.cacheable_methods
        if request.method in cacheable:
            try:
                cached_response = self.controller.cached_request(request)
            except zlib.error:
                cached_response = None
            if cached_response:
                return self.build_response(request, cached_response, from_cache=True)

            # check for etags and add headers if appropriate
            request.headers.update(self.controller.conditional_headers(request))

        resp = super(CacheControlAdapter, self).send(request, **kw)

        return resp

    def build_response(
        self, request, response, from_cache=False, cacheable_methods=None
    ):
        """
        Build a response by making a request or using the cache.

        This will end up calling send and returning a potentially
        cached response
        """
        cacheable = cacheable_methods or self.cacheable_methods
        if not from_cache and request.method in cacheable:
            # Check for any heuristics that might update headers
            # before trying to cache.
            if self.heuristic:
                response = self.heuristic.apply(response)

            # apply any expiration heuristics
            if response.status == 304:
                # We must have sent an ETag request. This could mean
                # that we've been expired already or that we simply
                # have an etag. In either case, we want to try and
                # update the cache if that is the case.
                cached_response = self.controller.update_cached_response(
                    request, response
                )

                if cached_response is not response:
                    from_cache = True

                # We are done with the server response, read a
                # possible response body (compliant servers will
                # not return one, but we cannot be 100% sure) and
                # release the connection back to the pool.
                response.read(decode_content=False)
                response.release_conn()

                response = cached_response

            # We always cache the 301 responses
            elif response.status == 301:
                self.controller.cache_response(request, response)
            else:
                # Wrap the response file with a wrapper that will cache the
                #   response when the stream has been consumed.
                response._fp = CallbackFileWrapper(
                    response._fp,
                    functools.partial(
                        self.controller.cache_response, request, response
                    ),
                )
                if response.chunked:
                    super_update_chunk_length = response._update_chunk_length

                    def _update_chunk_length(self):
                        super_update_chunk_length()
                        if self.chunk_left == 0:
                            self._fp._close()

                    response._update_chunk_length = types.MethodType(
                        _update_chunk_length, response
                    )

        resp = super(CacheControlAdapter, self).build_response(request, response)

        # See if we should invalidate the cache.
        if request.method in self.invalidating_methods and resp.ok:
            cache_url = self.controller.cache_url(request.url)
            self.cache.delete(cache_url)

        # Give the request a from_cache attr to let people use it
        resp.from_cache = from_cache

        return resp

    def close(self):
        self.cache.close()
        super(CacheControlAdapter, self).close()
site-packages/pip/_vendor/cachecontrol/__pycache__/heuristics.cpython-39.pyc000064400000011206150564745010023206 0ustar00a

��Re��@s�ddlZddlZddlmZmZmZddlmZmZdZddd�Z	dd�Z
Gd	d
�d
e�ZGdd�de�Z
Gd
d�de�ZGdd�de�ZdS)�N)�
formatdate�	parsedate�parsedate_tz)�datetime�	timedeltaz%a, %d %b %Y %H:%M:%S GMTcCs|p
t��}||S�N)r�utcnow)�delta�date�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/heuristics.py�expire_aftersr
cCstt�|����Sr)r�calendar�timegm�	timetuple)�dtrrr�datetime_to_headersrc@s$eZdZdd�Zdd�Zdd�ZdS)�
BaseHeuristiccCsdS)a!
        Return a valid 1xx warning header value describing the cache
        adjustments.

        The response is provided too allow warnings like 113
        http://tools.ietf.org/html/rfc7234#section-5.5.4 where we need
        to explicitly say response is over 24 hours old.
        z110 - "Response is Stale"r��self�responserrr�warnings	zBaseHeuristic.warningcCsiS)z�Update the response headers with any new headers.

        NOTE: This SHOULD always include some Warning header to
              signify that the response was cached by the client, not
              by way of the provided headers.
        rrrrr�update_headers!szBaseHeuristic.update_headerscCs@|�|�}|r<|j�|�|�|�}|dur<|j�d|i�|S)N�Warning)r�headers�updater)rrZupdated_headersZwarning_header_valuerrr�apply*s

zBaseHeuristic.applyN)�__name__�
__module__�__qualname__rrrrrrrrs	rc@seZdZdZdd�ZdS)�OneDayCachezM
    Cache the response by providing an expires 1 day in the
    future.
    cCsRi}d|jvrNt|jd�}ttdd�t|dd��d�}t|�|d<d|d<|S)	N�expiresr
�)�days�)r
�public�
cache-control)rrr
rrr)rrrr
r!rrrr<s
zOneDayCache.update_headersN)rrr�__doc__rrrrrr 6sr c@s(eZdZdZdd�Zdd�Zdd�ZdS)	�ExpiresAfterz;
    Cache **all** requests for a defined time period.
    cKstfi|��|_dSr)rr	)r�kwrrr�__init__LszExpiresAfter.__init__cCst|j�}t|�dd�S)Nr%)r!r&)r
r	r)rrr!rrrrOs
zExpiresAfter.update_headerscCsd}||jS)Nz:110 - Automatically cached for %s. Response might be stale)r	)rr�tmplrrrrSszExpiresAfter.warningN)rrrr'r*rrrrrrr(Gsr(c@s(eZdZdZhd�Zdd�Zdd�ZdS)�LastModifieda�
    If there is no Expires header already, fall back on Last-Modified
    using the heuristic from
    http://tools.ietf.org/html/rfc7234#section-4.2.2
    to calculate a reasonable value.

    Firefox also does something like this per
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Caching_FAQ
    http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397
    Unlike mozilla we limit this to 24-hr.
    >������i,��i-i�i�i�i�i�c
Cs�|j}d|vriSd|vr*|ddkr*iS|j|jvr:iSd|vsJd|vrNiSt�t|d��}t|d�}|dus||dur�iSt��}td||�}|t�|�}tdt	|dd��}||kr�iS||}	dt�
tt�|	��iS)	Nr!r&r%r
z
last-modifiedr�
i�Q)
r�status�cacheable_by_default_statusesrrrr�time�max�min�strftime�TIME_FMT�gmtime)
r�resprr
Z
last_modified�now�current_ager	�freshness_lifetimer!rrrrhs*zLastModified.update_headerscCsdSrr)rr:rrrr�szLastModified.warningN)rrrr'r3rrrrrrr,Xsr,)N)rr4�email.utilsrrrrrr8r
r�objectrr r(r,rrrr�<module>s
"site-packages/pip/_vendor/cachecontrol/__pycache__/wrapper.cpython-39.pyc000064400000001310150564745010022477 0ustar00a

��Re��@s&ddlmZddlmZddd�ZdS)�)�CacheControlAdapter)�	DictCacheNTc	CsJ|durt�n|}|pt}|||||||d�}|�d|�|�d|�|S)N)�cache_etags�
serializer�	heuristic�controller_class�cacheable_methodszhttp://zhttps://)rr�mount)	�sess�cacherrrr�
adapter_classr�adapter�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/wrapper.py�CacheControls�r)NTNNNNN)r
rrrrrrrr�<module>s�site-packages/pip/_vendor/cachecontrol/__pycache__/adapter.cpython-39.pyc000064400000006052150564745010022447 0ustar00a

��Re�@s\ddlZddlZddlZddlmZddlmZddlmZddl	m
Z
Gdd�de�ZdS)	�N)�HTTPAdapter�)�CacheController)�	DictCache)�CallbackFileWrappercsNeZdZddhZd�fdd�	Zd�fdd�	Zd�fd
d�	Z�fdd
�Z�ZS)�CacheControlAdapter�PUT�DELETENTc
sXtt|�j|i|��|dur$t�n||_||_|p6d|_|p@t}	|	|j||d�|_dS)N)�GET)�cache_etags�
serializer)	�superr�__init__r�cache�	heuristic�cacheable_methodsr�
controller)
�selfrr�controller_classrrr�args�kw�controller_factory��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/adapter.pyrs
�zCacheControlAdapter.__init__cs�|p|j}|j|vrfz|j�|�}Wntjy<d}Yn0|rR|j||dd�S|j�|j�	|��t
t|�j|fi|��}|S)z�
        Send a request. Use the request information to see if it
        exists in the cache and cache the response if we need to and can.
        NT)�
from_cache)
r�methodr�cached_request�zlib�error�build_response�headers�update�conditional_headersr
r�send)r�requestrr�	cacheable�cached_response�resprrrr%$s


zCacheControlAdapter.sendFc
s|p|j}|s�|j|vr�|jr*|j�|�}|jdkrh|j�||�}||urNd}|jdd�|��|}n\|jdkr�|j�	||�nBt
|jt�
|jj	||��|_|jr�|j��fdd�}t�||�|_tt|��||�}|j|jv�r|j�r|j�|j�}	|j�|	�||_|S)z�
        Build a response by making a request or using the cache.

        This will end up calling send and returning a potentially
        cached response
        i0TF)�decode_contenti-cs��|jdkr|j��dS)Nr)�
chunk_left�_fp�_close�r��super_update_chunk_lengthrr�_update_chunk_lengthns
z@CacheControlAdapter.build_response.<locals>._update_chunk_length)rrr�apply�statusr�update_cached_response�read�release_conn�cache_responserr,�	functools�partial�chunkedr1�types�
MethodTyper
rr!�invalidating_methods�ok�	cache_url�urlr�deleter)
rr&�responserrr'r(r1r)r?rr/rr!9sB	

�

���z"CacheControlAdapter.build_responsecs|j��tt|���dS)N)r�closer
rr.rrrrC�s
zCacheControlAdapter.close)NTNNNN)N)FN)	�__name__�
__module__�__qualname__r=rr%r!rC�
__classcell__rrrrrs��Jr)r;r8rZpip._vendor.requests.adaptersrrrrr�filewrapperrrrrrr�<module>ssite-packages/pip/_vendor/cachecontrol/__pycache__/serialize.cpython-39.pyc000064400000010245150564745010023015 0ustar00a

��Re��@spddlZddlZddlZddlZddlmZddlmZddlm	Z	m
Z
mZdd�Zdd	�Z
Gd
d�de�ZdS)�N)�msgpack)�CaseInsensitiveDict�)�HTTPResponse�pickle�	text_typecCst�|�d��S)N�ascii)�base64�	b64decode�encode)�b�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/serialize.py�_b64_decode_bytessrcCst|��d�S)N�utf8)r�decode)�sr
r
r�_b64_decode_strsrc@sNeZdZddd�Zdd�Zdd�Zdd	�Zd
d�Zdd
�Zdd�Z	dd�Z
dS)�
SerializerNc		Cs�t|j�}|dur*|jdd�}t�|�|_d|tdd�|j��D��|j|j	t
|j�|j|j
d�i}i|d<d|vr�|d�d�}|D]:}t
|���}|j�|d�}|dur�t
|�}||d|<q�d	�d
tj|dd�g�S)
NF)�decode_content�responsecss"|]\}}t|�t|�fVqdS�N)r��.0�k�vr
r
r�	<genexpr>2sz#Serializer.dumps.<locals>.<genexpr>)�body�headers�status�version�reason�strictr�vary�,�,scc=4T)�use_bin_type)rr�read�io�BytesIO�_fp�dict�itemsrr rr!r"r�split�strip�get�joinr�dumps)	�self�requestrr�response_headers�data�varied_headers�header�header_valuer
r
rr1s4

���zSerializer.dumpscCs�|sdSz|�dd�\}}Wnty2d}Yn0|dd�dkrP||}d}|�dd�d�d�}zt|d	�|��||�WSty�YdS0dS)
Nr%rscc=0�scc=�=���rz
_loads_v{})r-�
ValueErrorr�getattr�format�AttributeError)r2r3r5�verr
r
r�loadsJs
zSerializer.loadscCs�d|�di�vrdS|�di���D] \}}|j�|d�|kr$dSq$|d�d�}t|ddd�}|�dd	�d
kr�|�d�||dd<zt�|�}Wn"ty�t�|�d��}Yn0t	f|dd
�|d��S)z`Verify our vary headers match and construct a real urllib3
        HTTPResponse object.
        �*r#Nrrr)r5ztransfer-encoding��chunkedrF)r�preload_content)
r/r,r�poprr(r)�	TypeErrorrr)r2r3�cachedr7�value�body_rawrrr
r
r�prepare_responsehs
zSerializer.prepare_responsecCsdSrr
�r2r3r5r
r
r�	_loads_v0�szSerializer._loads_v0cCs0zt�|�}Wnty"YdS0|�||�Sr)rrAr<rK�r2r3r5rHr
r
r�	_loads_v1�s
zSerializer._loads_v1c	Cs�zt�t�|��d��}Wnttjfy4YdS0t|dd�|dd<tdd�|dd�	�D��|dd<t
|dd�|dd<tdd�|d	�	�D��|d	<|�||�S)
Nrrrcss"|]\}}t|�t|�fVqdSr�rrr
r
rr�s�z'Serializer._loads_v2.<locals>.<genexpr>rr!css.|]&\}}t|�|dur t|�n|fVqdSrrPrr
r
rr�s�r#)�jsonrA�zlib�
decompressrr<�errorrr+r,rrKrNr
r
r�	_loads_v2�s�
�zSerializer._loads_v2cCsdSrr
rLr
r
r�	_loads_v3�szSerializer._loads_v3cCs4ztj|dd�}Wnty&YdS0|�||�S)NF)�raw)rrAr<rKrNr
r
r�	_loads_v4�s
zSerializer._loads_v4)N)�__name__�
__module__�__qualname__r1rArKrMrOrUrVrXr
r
r
rrs
4&r)r	r(rQrR�pip._vendorrZpip._vendor.requests.structuresr�compatrrrrr�objectrr
r
r
r�<module>ssite-packages/pip/_vendor/cachecontrol/__pycache__/filewrapper.cpython-39.pyc000064400000004242150564745010023346 0ustar00a

��Re�	�@s ddlmZGdd�de�ZdS)�)�BytesIOc@sBeZdZdZdd�Zdd�Zdd�Zdd	�Zddd�Zd
d�Z	d
S)�CallbackFileWrapperav
    Small wrapper around a fp object which will tee everything read into a
    buffer, and when that file is closed it will execute a callback with the
    contents of that buffer.

    All attributes are proxied to the underlying file object.

    This class uses members with a double underscore (__) leading prefix so as
    not to accidentally shadow an attribute.
    cCst�|_||_||_dS�N)r�_CallbackFileWrapper__buf�_CallbackFileWrapper__fp�_CallbackFileWrapper__callback)�self�fp�callback�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/filewrapper.py�__init__szCallbackFileWrapper.__init__cCs|�d�}t||�S)Nr)�__getattribute__�getattr)r�namer	rrr�__getattr__s	
zCallbackFileWrapper.__getattr__cCsDz|jjduWSty Yn0z
|jjWSty>Yn0dS)NF)rr	�AttributeError�closed�rrrr�__is_fp_closed!s
z"CallbackFileWrapper.__is_fp_closedcCs |jr|�|j���d|_dSr)rr�getvaluerrrr�_close2szCallbackFileWrapper._closeNcCs,|j�|�}|j�|�|��r(|��|Sr)r�readr�write�"_CallbackFileWrapper__is_fp_closedr�r�amt�datarrrr=s
zCallbackFileWrapper.readcCs@|j�|�}|dkr |dkr |S|j�|�|��r<|��|S)N�s
)r�
_safe_readrrrrrrrrrEszCallbackFileWrapper._safe_read)N)
�__name__�
__module__�__qualname__�__doc__r
rrrrrrrrrrs
rN)�ior�objectrrrrr�<module>ssite-packages/pip/_vendor/cachecontrol/__pycache__/compat.cpython-39.pyc000064400000001421150564745010022305 0ustar00a

��Re��@s�zddlmZWney.ddlmZYn0zddlZWneyVddlZYn0zddlmZWney�ddlmZYn0zddl	m
Z
Wney�ddlm
Z
Yn0zeZ
Wney�eZ
Yn0dS)�)�urljoinN)�HTTPResponse)�is_fp_closed)�urllib.parser�ImportError�urlparse�cPickle�pickleZ.pip._vendor.requests.packages.urllib3.responserZpip._vendor.urllib3.responseZ*pip._vendor.requests.packages.urllib3.utilr�pip._vendor.urllib3.util�unicode�	text_type�	NameError�str�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/compat.py�<module>s&site-packages/pip/_vendor/cachecontrol/__pycache__/__init__.cpython-39.pyc000064400000001114150564745010022560 0ustar00a

��Re.�@s8dZdZdZdZddlmZddlmZddlm	Z	dS)	zbCacheControl import Interface.

Make it easy to import from cachecontrol without long namespaces.
zEric Larsonzeric@ionrock.orgz0.12.6�)�CacheControl)�CacheControlAdapter)�CacheControllerN)
�__doc__�
__author__�	__email__�__version__�wrapperr�adapterr�
controllerr�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/__init__.py�<module>ssite-packages/pip/_vendor/cachecontrol/__pycache__/controller.cpython-39.pyc000064400000017172150564745010023217 0ustar00a

��ReE7�@s�dZddlZddlZddlZddlZddlmZddlmZddl	m
Z
ddlmZe�
e�Ze�d�Zd	d
�ZGdd�de�ZdS)
z7
The httplib2 algorithms ported for use with requests.
�N)�parsedate_tz)�CaseInsensitiveDict�)�	DictCache)�
Serializerz9^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?cCs0t�|���}|d|d|d|d|dfS)z�Parses a URI using the regex given in Appendix B of RFC 3986.

        (scheme, authority, path, query, fragment) = parse_uri(uri)
    r����)�URI�match�groups)�urir
�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/controller.py�	parse_urisrc@s\eZdZdZddd�Zedd��Zedd	��Zd
d�Zdd
�Z	dd�Z
ddd�Zdd�ZdS)�CacheControllerz9An interface to see if request should cached or not.
    NTcCs4|durt�n||_||_|p"t�|_|p,d|_dS)N)����i,�-)r�cache�cache_etagsr�
serializer�cacheable_status_codes)�selfrrr�status_codesrrr�__init__"szCacheController.__init__c	Csht|�\}}}}}|r|s&td|��|��}|��}|s>d}|rPd�||g�pR|}|d||}|S)z4Normalize the URL to create a safe key for the cachez(Only absolute URIs are allowed. uri = %s�/�?z://)r�	Exception�lower�join)	�clsr�scheme�	authority�path�query�fragment�request_uri�
defrag_urirrr�_urlnorm*szCacheController._urlnormcCs
|�|�S�N)r*)r"rrrr�	cache_url>szCacheController.cache_urlc

Cs"tdftdftdfddddddddtdfd�}|�d|�dd��}i}|�d�D]�}|��s^qP|�d	d
�}|d��}z||\}}	Wn"ty�t�d|�YqPYn0|r�|	s�d||<|rPz||d
���||<WqPty�|	r�t�d
|�YqPt�yt�d||j	�YqP0qP|S)NTF)NF)�max-agez	max-stale�	min-fresh�no-cache�no-storezno-transformzonly-if-cachedzmust-revalidate�public�privatezproxy-revalidatezs-maxagez
cache-controlz
Cache-Control��,�=rrz,Ignoring unknown cache-control directive: %sz-Missing value for cache-control directive: %sz8Invalid value for cache-control directive %s, must be %s)
�int�get�split�strip�KeyError�logger�debug�
IndexError�
ValueError�__name__)
r�headers�known_directives�
cc_headers�retval�cc_directive�parts�	directive�typ�requiredrrr�parse_cache_controlBsV�
��z#CacheController.parse_cache_controlcCs0|�|j�}t�d|�|�|j�}d|vr:t�d�dSd|vr\|ddkr\t�d�dS|j�|�}|dur~t�d	�dS|j�	||�}|s�t�
d
�dS|jdkr�d}t�|�|St|j�}|r�d
|vr�d|vr�t�d�|j�
|�t�d�dSt��}t�t|d
��}	td||	�}
t�d|
�|�|�}d}d|v�r^|d}t�d|�nDd|v�r�t|d�}
|
du�r�t�|
�|	}td|�}t�d|�d|v�r�|d}t�d|�d|v�r�|d}|
|7}
t�d|
�||
k�rt�d�t�d||
�|Sd|v�r,t�d�|j�
|�dS)ze
        Return a cached response if it exists in the cache, otherwise
        return False.
        zLooking up "%s" in the cacher/z-Request header has "no-cache", cache bypassedFr-rz1Request header has "max_age" as 0, cache bypassedNzNo cache entry availablez1Cache entry deserialization failed, entry ignoredrzVReturning cached "301 Moved Permanently" response (ignoring date and etag information)�date�etagz(Purging cached response: no date or etagz!Ignoring cached response: no datezCurrent age based on date: %iz#Freshness lifetime from max-age: %i�expiresz#Freshness lifetime from expires: %iz+Freshness lifetime from request max-age: %ir.z'Adjusted current age from min-fresh: %iz2The response is "fresh", returning cached responsez%i > %iz4The cached response is "stale" with no etag, purging)r,�urlr;r<rIr@rr7r�loads�warning�statusr�delete�time�calendar�timegmr�max)r�requestr,�cc�
cache_data�resp�msgr@�nowrJ�current_age�resp_cc�freshness_lifetimerL�expire_time�	min_freshrrr�cached_requestxsz





�









�




zCacheController.cached_requestcCs`|�|j�}|j�||j�|��}i}|r\t|j�}d|vrH|d|d<d|vr\|d|d<|S)NrK�ETagz
If-None-Matchz
last-modifiedz
Last-ModifiedzIf-Modified-Since)r,rMrrNrr7rr@)rrVr,rY�new_headersr@rrr�conditional_headers�s
z#CacheController.conditional_headerscCs|p|j}|j|vr(t�d|j|�dSt|j�}|durfd|vrf|d��rft|d�t|�krfdS|�	|j�}|�	|�}|�
|j�}	t�d|	�d}
d|vr�d}
t�d�d|vr�d}
t�d	�|
r�|j�
|	�r�t�d
�|j�|	�|
r�dSd|�
dd
�v�rt�d�dS|j�rLd|v�rLt�d�|j�|	|jj|||d��n�|jdk�r|t�d�|j�|	|j�||��n�d|v�rd|v�r�|ddk�r�t�d�|j�|	|jj|||d��n:d|v�r|d�rt�d�|j�|	|jj|||d��dS)zc
        Algorithm for caching requests.

        This assumes a requests Response object.
        zStatus code %s not in %sN�content-lengthz&Updating cache with response from "%s"Fr0TzResponse header has "no-store"zRequest header has "no-store"z0Purging existing cache entry to honor "no-store"�*�varyr3zResponse header has "Vary: *"rKzCaching due to etag)�bodyrzCaching permanant redirectrJr-rz'Caching b/c date exists and max-age > 0rLzCaching b/c of expires header)rrPr;r<rr@�isdigitr6�lenrIr,rMrr7rQr�setr�dumps)rrV�responserhrr�response_headers�cc_reqrWr,�no_storerrr�cache_response�sn

�
��
��





�


�


�zCacheController.cache_responsecsv|�|j�}|j�||j�|��}|s*|Sdg�|j�t�fdd�|j�	�D���d|_
|j�||j�||��|S)z�On a 304 we will get a new set of headers that we want to
        update our cached value with, assuming we have one.

        This should only ever be called when we've sent an ETag and
        gotten a 304 as the response.
        rec3s&|]\}}|���vr||fVqdSr+)r )�.0�k�v��excluded_headersrr�	<genexpr>ks�z9CacheController.update_cached_response.<locals>.<genexpr>r)
r,rMrrNrr7r@�update�dict�itemsrPrkrl)rrVrmr,�cached_responserrur�update_cached_responseRs	��	z&CacheController.update_cached_response)NTNN)NN)
r?�
__module__�__qualname__�__doc__r�classmethodr*r,rIrardrqr|rrrrrs�


6o
[r)r�logging�rerSrR�email.utilsrZpip._vendor.requests.structuresrrr�	serializer�	getLoggerr?r;�compilerr�objectrrrrr�<module>s

	site-packages/pip/_vendor/cachecontrol/__pycache__/cache.cpython-39.pyc000064400000003505150564745010022072 0ustar00a

��Re%�@s4dZddlmZGdd�de�ZGdd�de�ZdS)zb
The cache object API for implementing caches. The default is a thread
safe in-memory dictionary.
�)�Lockc@s,eZdZdd�Zdd�Zdd�Zdd�Zd	S)
�	BaseCachecCs
t��dS�N��NotImplementedError��self�key�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/cache.py�get
sz
BaseCache.getcCs
t��dSrr�rr	�valuer
r
r�set
sz
BaseCache.setcCs
t��dSrrrr
r
r�deleteszBaseCache.deletecCsdSrr
)rr
r
r�closeszBaseCache.closeN)�__name__�
__module__�__qualname__rrrrr
r
r
rrsrc@s.eZdZd
dd�Zdd�Zdd�Zdd	�ZdS)�	DictCacheNcCst�|_|pi|_dSr)r�lock�data)r�	init_dictr
r
r�__init__szDictCache.__init__cCs|j�|d�Sr)rrrr
r
rrsz
DictCache.getcCs:|j� |j�||i�Wd�n1s,0YdSr)rr�updater
r
r
rr sz
DictCache.setcCs@|j�&||jvr|j�|�Wd�n1s20YdSr)rr�poprr
r
rr$s
zDictCache.delete)N)rrrrrrrr
r
r
rrs
rN)�__doc__�	threadingr�objectrrr
r
r
r�<module>ssite-packages/pip/_vendor/cachecontrol/__pycache__/_cmd.cpython-39.pyc000064400000003111150564745010021722 0ustar00a

��Re�@sxddlZddlmZddlmZddlmZddlmZddl	m
Z
dd�Zd	d
�Zdd�Z
dd
d�Zedkrte�dS)�N)�requests)�CacheControlAdapter)�	DictCache)�logger)�ArgumentParsercCs"t�tj�t��}t�|�dS)N)r�setLevel�logging�DEBUG�
StreamHandler�
addHandler)�handler�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_vendor/cachecontrol/_cmd.py�
setup_loggingsrcCs>tt�dddd�}t��}|�d|�|�d|�|j|_|S)NT)�cache_etags�
serializer�	heuristiczhttp://zhttps://)rrr�Session�mount�
controller�cache_controller)�adapter�sessr
r
r�get_sessions
�rcCst�}|jddd�|��S)N�urlzThe URL to try and cache)�help)r�add_argument�
parse_args)�parserr
r
r�get_argssrcCsTt�}t�}|�|j�}t�|j�|j|j�|j�	|j�rHt
d�nt
d�dS)NzCached!z
Not cached :()rr�getrrr�cache_response�request�raw�cached_request�print)�argsr�respr
r
r�main$s
r(�__main__)N)r�pip._vendorrZ pip._vendor.cachecontrol.adapterrZpip._vendor.cachecontrol.cacherZ#pip._vendor.cachecontrol.controllerr�argparserrrrr(�__name__r
r
r
r�<module>s
site-packages/pip/_vendor/cachecontrol/wrapper.py000064400000001262150564745010016216 0ustar00from .adapter import CacheControlAdapter
from .cache import DictCache


def CacheControl(
    sess,
    cache=None,
    cache_etags=True,
    serializer=None,
    heuristic=None,
    controller_class=None,
    adapter_class=None,
    cacheable_methods=None,
):

    cache = DictCache() if cache is None else cache
    adapter_class = adapter_class or CacheControlAdapter
    adapter = adapter_class(
        cache,
        cache_etags=cache_etags,
        serializer=serializer,
        heuristic=heuristic,
        controller_class=controller_class,
        cacheable_methods=cacheable_methods,
    )
    sess.mount("http://", adapter)
    sess.mount("https://", adapter)

    return sess
site-packages/pip/_vendor/cachecontrol/_cmd.py000064400000002417150564745010015443 0ustar00import logging

from pip._vendor import requests

from pip._vendor.cachecontrol.adapter import CacheControlAdapter
from pip._vendor.cachecontrol.cache import DictCache
from pip._vendor.cachecontrol.controller import logger

from argparse import ArgumentParser


def setup_logging():
    logger.setLevel(logging.DEBUG)
    handler = logging.StreamHandler()
    logger.addHandler(handler)


def get_session():
    adapter = CacheControlAdapter(
        DictCache(), cache_etags=True, serializer=None, heuristic=None
    )
    sess = requests.Session()
    sess.mount("http://", adapter)
    sess.mount("https://", adapter)

    sess.cache_controller = adapter.controller
    return sess


def get_args():
    parser = ArgumentParser()
    parser.add_argument("url", help="The URL to try and cache")
    return parser.parse_args()


def main(args=None):
    args = get_args()
    sess = get_session()

    # Make a request to get a response
    resp = sess.get(args.url)

    # Turn on logging
    setup_logging()

    # try setting the cache
    sess.cache_controller.cache_response(resp.request, resp.raw)

    # Now try to get it
    if sess.cache_controller.cached_request(resp.request):
        print("Cached!")
    else:
        print("Not cached :(")


if __name__ == "__main__":
    main()
site-packages/pip/_vendor/cachecontrol/filewrapper.py000064400000004745150564745010017067 0ustar00from io import BytesIO


class CallbackFileWrapper(object):
    """
    Small wrapper around a fp object which will tee everything read into a
    buffer, and when that file is closed it will execute a callback with the
    contents of that buffer.

    All attributes are proxied to the underlying file object.

    This class uses members with a double underscore (__) leading prefix so as
    not to accidentally shadow an attribute.
    """

    def __init__(self, fp, callback):
        self.__buf = BytesIO()
        self.__fp = fp
        self.__callback = callback

    def __getattr__(self, name):
        # The vaguaries of garbage collection means that self.__fp is
        # not always set.  By using __getattribute__ and the private
        # name[0] allows looking up the attribute value and raising an
        # AttributeError when it doesn't exist. This stop thigns from
        # infinitely recursing calls to getattr in the case where
        # self.__fp hasn't been set.
        #
        # [0] https://docs.python.org/2/reference/expressions.html#atom-identifiers
        fp = self.__getattribute__("_CallbackFileWrapper__fp")
        return getattr(fp, name)

    def __is_fp_closed(self):
        try:
            return self.__fp.fp is None

        except AttributeError:
            pass

        try:
            return self.__fp.closed

        except AttributeError:
            pass

        # We just don't cache it then.
        # TODO: Add some logging here...
        return False

    def _close(self):
        if self.__callback:
            self.__callback(self.__buf.getvalue())

        # We assign this to None here, because otherwise we can get into
        # really tricky problems where the CPython interpreter dead locks
        # because the callback is holding a reference to something which
        # has a __del__ method. Setting this to None breaks the cycle
        # and allows the garbage collector to do it's thing normally.
        self.__callback = None

    def read(self, amt=None):
        data = self.__fp.read(amt)
        self.__buf.write(data)
        if self.__is_fp_closed():
            self._close()

        return data

    def _safe_read(self, amt):
        data = self.__fp._safe_read(amt)
        if amt == 2 and data == b"\r\n":
            # urllib executes this read to toss the CRLF at the end
            # of the chunk.
            return data

        self.__buf.write(data)
        if self.__is_fp_closed():
            self._close()

        return data
site-packages/pip/_vendor/cachecontrol/compat.py000064400000001267150564745010016026 0ustar00try:
    from urllib.parse import urljoin
except ImportError:
    from urlparse import urljoin


try:
    import cPickle as pickle
except ImportError:
    import pickle


# Handle the case where the requests module has been patched to not have
# urllib3 bundled as part of its source.
try:
    from pip._vendor.requests.packages.urllib3.response import HTTPResponse
except ImportError:
    from pip._vendor.urllib3.response import HTTPResponse

try:
    from pip._vendor.requests.packages.urllib3.util import is_fp_closed
except ImportError:
    from pip._vendor.urllib3.util import is_fp_closed

# Replicate some six behaviour
try:
    text_type = unicode
except NameError:
    text_type = str
site-packages/pip/_vendor/cachecontrol/__init__.py000064400000000456150564745010016301 0ustar00"""CacheControl import Interface.

Make it easy to import from cachecontrol without long namespaces.
"""
__author__ = "Eric Larson"
__email__ = "eric@ionrock.org"
__version__ = "0.12.6"

from .wrapper import CacheControl
from .adapter import CacheControlAdapter
from .controller import CacheController
site-packages/pip/_vendor/cachecontrol/controller.py000064400000033505150564745010016726 0ustar00"""
The httplib2 algorithms ported for use with requests.
"""
import logging
import re
import calendar
import time
from email.utils import parsedate_tz

from pip._vendor.requests.structures import CaseInsensitiveDict

from .cache import DictCache
from .serialize import Serializer


logger = logging.getLogger(__name__)

URI = re.compile(r"^(([^:/?#]+):)?(//([^/?#]*))?([^?#]*)(\?([^#]*))?(#(.*))?")


def parse_uri(uri):
    """Parses a URI using the regex given in Appendix B of RFC 3986.

        (scheme, authority, path, query, fragment) = parse_uri(uri)
    """
    groups = URI.match(uri).groups()
    return (groups[1], groups[3], groups[4], groups[6], groups[8])


class CacheController(object):
    """An interface to see if request should cached or not.
    """

    def __init__(
        self, cache=None, cache_etags=True, serializer=None, status_codes=None
    ):
        self.cache = DictCache() if cache is None else cache
        self.cache_etags = cache_etags
        self.serializer = serializer or Serializer()
        self.cacheable_status_codes = status_codes or (200, 203, 300, 301)

    @classmethod
    def _urlnorm(cls, uri):
        """Normalize the URL to create a safe key for the cache"""
        (scheme, authority, path, query, fragment) = parse_uri(uri)
        if not scheme or not authority:
            raise Exception("Only absolute URIs are allowed. uri = %s" % uri)

        scheme = scheme.lower()
        authority = authority.lower()

        if not path:
            path = "/"

        # Could do syntax based normalization of the URI before
        # computing the digest. See Section 6.2.2 of Std 66.
        request_uri = query and "?".join([path, query]) or path
        defrag_uri = scheme + "://" + authority + request_uri

        return defrag_uri

    @classmethod
    def cache_url(cls, uri):
        return cls._urlnorm(uri)

    def parse_cache_control(self, headers):
        known_directives = {
            # https://tools.ietf.org/html/rfc7234#section-5.2
            "max-age": (int, True),
            "max-stale": (int, False),
            "min-fresh": (int, True),
            "no-cache": (None, False),
            "no-store": (None, False),
            "no-transform": (None, False),
            "only-if-cached": (None, False),
            "must-revalidate": (None, False),
            "public": (None, False),
            "private": (None, False),
            "proxy-revalidate": (None, False),
            "s-maxage": (int, True),
        }

        cc_headers = headers.get("cache-control", headers.get("Cache-Control", ""))

        retval = {}

        for cc_directive in cc_headers.split(","):
            if not cc_directive.strip():
                continue

            parts = cc_directive.split("=", 1)
            directive = parts[0].strip()

            try:
                typ, required = known_directives[directive]
            except KeyError:
                logger.debug("Ignoring unknown cache-control directive: %s", directive)
                continue

            if not typ or not required:
                retval[directive] = None
            if typ:
                try:
                    retval[directive] = typ(parts[1].strip())
                except IndexError:
                    if required:
                        logger.debug(
                            "Missing value for cache-control " "directive: %s",
                            directive,
                        )
                except ValueError:
                    logger.debug(
                        "Invalid value for cache-control directive " "%s, must be %s",
                        directive,
                        typ.__name__,
                    )

        return retval

    def cached_request(self, request):
        """
        Return a cached response if it exists in the cache, otherwise
        return False.
        """
        cache_url = self.cache_url(request.url)
        logger.debug('Looking up "%s" in the cache', cache_url)
        cc = self.parse_cache_control(request.headers)

        # Bail out if the request insists on fresh data
        if "no-cache" in cc:
            logger.debug('Request header has "no-cache", cache bypassed')
            return False

        if "max-age" in cc and cc["max-age"] == 0:
            logger.debug('Request header has "max_age" as 0, cache bypassed')
            return False

        # Request allows serving from the cache, let's see if we find something
        cache_data = self.cache.get(cache_url)
        if cache_data is None:
            logger.debug("No cache entry available")
            return False

        # Check whether it can be deserialized
        resp = self.serializer.loads(request, cache_data)
        if not resp:
            logger.warning("Cache entry deserialization failed, entry ignored")
            return False

        # If we have a cached 301, return it immediately. We don't
        # need to test our response for other headers b/c it is
        # intrinsically "cacheable" as it is Permanent.
        # See:
        #   https://tools.ietf.org/html/rfc7231#section-6.4.2
        #
        # Client can try to refresh the value by repeating the request
        # with cache busting headers as usual (ie no-cache).
        if resp.status == 301:
            msg = (
                'Returning cached "301 Moved Permanently" response '
                "(ignoring date and etag information)"
            )
            logger.debug(msg)
            return resp

        headers = CaseInsensitiveDict(resp.headers)
        if not headers or "date" not in headers:
            if "etag" not in headers:
                # Without date or etag, the cached response can never be used
                # and should be deleted.
                logger.debug("Purging cached response: no date or etag")
                self.cache.delete(cache_url)
            logger.debug("Ignoring cached response: no date")
            return False

        now = time.time()
        date = calendar.timegm(parsedate_tz(headers["date"]))
        current_age = max(0, now - date)
        logger.debug("Current age based on date: %i", current_age)

        # TODO: There is an assumption that the result will be a
        #       urllib3 response object. This may not be best since we
        #       could probably avoid instantiating or constructing the
        #       response until we know we need it.
        resp_cc = self.parse_cache_control(headers)

        # determine freshness
        freshness_lifetime = 0

        # Check the max-age pragma in the cache control header
        if "max-age" in resp_cc:
            freshness_lifetime = resp_cc["max-age"]
            logger.debug("Freshness lifetime from max-age: %i", freshness_lifetime)

        # If there isn't a max-age, check for an expires header
        elif "expires" in headers:
            expires = parsedate_tz(headers["expires"])
            if expires is not None:
                expire_time = calendar.timegm(expires) - date
                freshness_lifetime = max(0, expire_time)
                logger.debug("Freshness lifetime from expires: %i", freshness_lifetime)

        # Determine if we are setting freshness limit in the
        # request. Note, this overrides what was in the response.
        if "max-age" in cc:
            freshness_lifetime = cc["max-age"]
            logger.debug(
                "Freshness lifetime from request max-age: %i", freshness_lifetime
            )

        if "min-fresh" in cc:
            min_fresh = cc["min-fresh"]
            # adjust our current age by our min fresh
            current_age += min_fresh
            logger.debug("Adjusted current age from min-fresh: %i", current_age)

        # Return entry if it is fresh enough
        if freshness_lifetime > current_age:
            logger.debug('The response is "fresh", returning cached response')
            logger.debug("%i > %i", freshness_lifetime, current_age)
            return resp

        # we're not fresh. If we don't have an Etag, clear it out
        if "etag" not in headers:
            logger.debug('The cached response is "stale" with no etag, purging')
            self.cache.delete(cache_url)

        # return the original handler
        return False

    def conditional_headers(self, request):
        cache_url = self.cache_url(request.url)
        resp = self.serializer.loads(request, self.cache.get(cache_url))
        new_headers = {}

        if resp:
            headers = CaseInsensitiveDict(resp.headers)

            if "etag" in headers:
                new_headers["If-None-Match"] = headers["ETag"]

            if "last-modified" in headers:
                new_headers["If-Modified-Since"] = headers["Last-Modified"]

        return new_headers

    def cache_response(self, request, response, body=None, status_codes=None):
        """
        Algorithm for caching requests.

        This assumes a requests Response object.
        """
        # From httplib2: Don't cache 206's since we aren't going to
        #                handle byte range requests
        cacheable_status_codes = status_codes or self.cacheable_status_codes
        if response.status not in cacheable_status_codes:
            logger.debug(
                "Status code %s not in %s", response.status, cacheable_status_codes
            )
            return

        response_headers = CaseInsensitiveDict(response.headers)

        # If we've been given a body, our response has a Content-Length, that
        # Content-Length is valid then we can check to see if the body we've
        # been given matches the expected size, and if it doesn't we'll just
        # skip trying to cache it.
        if (
            body is not None
            and "content-length" in response_headers
            and response_headers["content-length"].isdigit()
            and int(response_headers["content-length"]) != len(body)
        ):
            return

        cc_req = self.parse_cache_control(request.headers)
        cc = self.parse_cache_control(response_headers)

        cache_url = self.cache_url(request.url)
        logger.debug('Updating cache with response from "%s"', cache_url)

        # Delete it from the cache if we happen to have it stored there
        no_store = False
        if "no-store" in cc:
            no_store = True
            logger.debug('Response header has "no-store"')
        if "no-store" in cc_req:
            no_store = True
            logger.debug('Request header has "no-store"')
        if no_store and self.cache.get(cache_url):
            logger.debug('Purging existing cache entry to honor "no-store"')
            self.cache.delete(cache_url)
        if no_store:
            return

        # https://tools.ietf.org/html/rfc7234#section-4.1:
        # A Vary header field-value of "*" always fails to match.
        # Storing such a response leads to a deserialization warning
        # during cache lookup and is not allowed to ever be served,
        # so storing it can be avoided.
        if "*" in response_headers.get("vary", ""):
            logger.debug('Response header has "Vary: *"')
            return

        # If we've been given an etag, then keep the response
        if self.cache_etags and "etag" in response_headers:
            logger.debug("Caching due to etag")
            self.cache.set(
                cache_url, self.serializer.dumps(request, response, body=body)
            )

        # Add to the cache any 301s. We do this before looking that
        # the Date headers.
        elif response.status == 301:
            logger.debug("Caching permanant redirect")
            self.cache.set(cache_url, self.serializer.dumps(request, response))

        # Add to the cache if the response headers demand it. If there
        # is no date header then we can't do anything about expiring
        # the cache.
        elif "date" in response_headers:
            # cache when there is a max-age > 0
            if "max-age" in cc and cc["max-age"] > 0:
                logger.debug("Caching b/c date exists and max-age > 0")
                self.cache.set(
                    cache_url, self.serializer.dumps(request, response, body=body)
                )

            # If the request can expire, it means we should cache it
            # in the meantime.
            elif "expires" in response_headers:
                if response_headers["expires"]:
                    logger.debug("Caching b/c of expires header")
                    self.cache.set(
                        cache_url, self.serializer.dumps(request, response, body=body)
                    )

    def update_cached_response(self, request, response):
        """On a 304 we will get a new set of headers that we want to
        update our cached value with, assuming we have one.

        This should only ever be called when we've sent an ETag and
        gotten a 304 as the response.
        """
        cache_url = self.cache_url(request.url)

        cached_response = self.serializer.loads(request, self.cache.get(cache_url))

        if not cached_response:
            # we didn't have a cached response
            return response

        # Lets update our headers with the headers from the new request:
        # http://tools.ietf.org/html/draft-ietf-httpbis-p4-conditional-26#section-4.1
        #
        # The server isn't supposed to send headers that would make
        # the cached body invalid. But... just in case, we'll be sure
        # to strip out ones we know that might be problmatic due to
        # typical assumptions.
        excluded_headers = ["content-length"]

        cached_response.headers.update(
            dict(
                (k, v)
                for k, v in response.headers.items()
                if k.lower() not in excluded_headers
            )
        )

        # we want a 200 b/c we have content via the cache
        cached_response.status = 200

        # update our cache
        self.cache.set(cache_url, self.serializer.dumps(request, cached_response))

        return cached_response
site-packages/pip/_vendor/cachecontrol/heuristics.py000064400000007746150564745010016735 0ustar00import calendar
import time

from email.utils import formatdate, parsedate, parsedate_tz

from datetime import datetime, timedelta

TIME_FMT = "%a, %d %b %Y %H:%M:%S GMT"


def expire_after(delta, date=None):
    date = date or datetime.utcnow()
    return date + delta


def datetime_to_header(dt):
    return formatdate(calendar.timegm(dt.timetuple()))


class BaseHeuristic(object):

    def warning(self, response):
        """
        Return a valid 1xx warning header value describing the cache
        adjustments.

        The response is provided too allow warnings like 113
        http://tools.ietf.org/html/rfc7234#section-5.5.4 where we need
        to explicitly say response is over 24 hours old.
        """
        return '110 - "Response is Stale"'

    def update_headers(self, response):
        """Update the response headers with any new headers.

        NOTE: This SHOULD always include some Warning header to
              signify that the response was cached by the client, not
              by way of the provided headers.
        """
        return {}

    def apply(self, response):
        updated_headers = self.update_headers(response)

        if updated_headers:
            response.headers.update(updated_headers)
            warning_header_value = self.warning(response)
            if warning_header_value is not None:
                response.headers.update({"Warning": warning_header_value})

        return response


class OneDayCache(BaseHeuristic):
    """
    Cache the response by providing an expires 1 day in the
    future.
    """

    def update_headers(self, response):
        headers = {}

        if "expires" not in response.headers:
            date = parsedate(response.headers["date"])
            expires = expire_after(timedelta(days=1), date=datetime(*date[:6]))
            headers["expires"] = datetime_to_header(expires)
            headers["cache-control"] = "public"
        return headers


class ExpiresAfter(BaseHeuristic):
    """
    Cache **all** requests for a defined time period.
    """

    def __init__(self, **kw):
        self.delta = timedelta(**kw)

    def update_headers(self, response):
        expires = expire_after(self.delta)
        return {"expires": datetime_to_header(expires), "cache-control": "public"}

    def warning(self, response):
        tmpl = "110 - Automatically cached for %s. Response might be stale"
        return tmpl % self.delta


class LastModified(BaseHeuristic):
    """
    If there is no Expires header already, fall back on Last-Modified
    using the heuristic from
    http://tools.ietf.org/html/rfc7234#section-4.2.2
    to calculate a reasonable value.

    Firefox also does something like this per
    https://developer.mozilla.org/en-US/docs/Web/HTTP/Caching_FAQ
    http://lxr.mozilla.org/mozilla-release/source/netwerk/protocol/http/nsHttpResponseHead.cpp#397
    Unlike mozilla we limit this to 24-hr.
    """
    cacheable_by_default_statuses = {
        200, 203, 204, 206, 300, 301, 404, 405, 410, 414, 501
    }

    def update_headers(self, resp):
        headers = resp.headers

        if "expires" in headers:
            return {}

        if "cache-control" in headers and headers["cache-control"] != "public":
            return {}

        if resp.status not in self.cacheable_by_default_statuses:
            return {}

        if "date" not in headers or "last-modified" not in headers:
            return {}

        date = calendar.timegm(parsedate_tz(headers["date"]))
        last_modified = parsedate(headers["last-modified"])
        if date is None or last_modified is None:
            return {}

        now = time.time()
        current_age = max(0, now - date)
        delta = date - calendar.timegm(last_modified)
        freshness_lifetime = max(0, min(delta / 10, 24 * 3600))
        if freshness_lifetime <= current_age:
            return {}

        expires = date + freshness_lifetime
        return {"expires": time.strftime(TIME_FMT, time.gmtime(expires))}

    def warning(self, resp):
        return None
site-packages/pip/_vendor/cachecontrol/serialize.py000064400000015663150564745010016537 0ustar00import base64
import io
import json
import zlib

from pip._vendor import msgpack
from pip._vendor.requests.structures import CaseInsensitiveDict

from .compat import HTTPResponse, pickle, text_type


def _b64_decode_bytes(b):
    return base64.b64decode(b.encode("ascii"))


def _b64_decode_str(s):
    return _b64_decode_bytes(s).decode("utf8")


class Serializer(object):

    def dumps(self, request, response, body=None):
        response_headers = CaseInsensitiveDict(response.headers)

        if body is None:
            body = response.read(decode_content=False)

            # NOTE: 99% sure this is dead code. I'm only leaving it
            #       here b/c I don't have a test yet to prove
            #       it. Basically, before using
            #       `cachecontrol.filewrapper.CallbackFileWrapper`,
            #       this made an effort to reset the file handle. The
            #       `CallbackFileWrapper` short circuits this code by
            #       setting the body as the content is consumed, the
            #       result being a `body` argument is *always* passed
            #       into cache_response, and in turn,
            #       `Serializer.dump`.
            response._fp = io.BytesIO(body)

        # NOTE: This is all a bit weird, but it's really important that on
        #       Python 2.x these objects are unicode and not str, even when
        #       they contain only ascii. The problem here is that msgpack
        #       understands the difference between unicode and bytes and we
        #       have it set to differentiate between them, however Python 2
        #       doesn't know the difference. Forcing these to unicode will be
        #       enough to have msgpack know the difference.
        data = {
            u"response": {
                u"body": body,
                u"headers": dict(
                    (text_type(k), text_type(v)) for k, v in response.headers.items()
                ),
                u"status": response.status,
                u"version": response.version,
                u"reason": text_type(response.reason),
                u"strict": response.strict,
                u"decode_content": response.decode_content,
            }
        }

        # Construct our vary headers
        data[u"vary"] = {}
        if u"vary" in response_headers:
            varied_headers = response_headers[u"vary"].split(",")
            for header in varied_headers:
                header = text_type(header).strip()
                header_value = request.headers.get(header, None)
                if header_value is not None:
                    header_value = text_type(header_value)
                data[u"vary"][header] = header_value

        return b",".join([b"cc=4", msgpack.dumps(data, use_bin_type=True)])

    def loads(self, request, data):
        # Short circuit if we've been given an empty set of data
        if not data:
            return

        # Determine what version of the serializer the data was serialized
        # with
        try:
            ver, data = data.split(b",", 1)
        except ValueError:
            ver = b"cc=0"

        # Make sure that our "ver" is actually a version and isn't a false
        # positive from a , being in the data stream.
        if ver[:3] != b"cc=":
            data = ver + data
            ver = b"cc=0"

        # Get the version number out of the cc=N
        ver = ver.split(b"=", 1)[-1].decode("ascii")

        # Dispatch to the actual load method for the given version
        try:
            return getattr(self, "_loads_v{}".format(ver))(request, data)

        except AttributeError:
            # This is a version we don't have a loads function for, so we'll
            # just treat it as a miss and return None
            return

    def prepare_response(self, request, cached):
        """Verify our vary headers match and construct a real urllib3
        HTTPResponse object.
        """
        # Special case the '*' Vary value as it means we cannot actually
        # determine if the cached response is suitable for this request.
        # This case is also handled in the controller code when creating
        # a cache entry, but is left here for backwards compatibility.
        if "*" in cached.get("vary", {}):
            return

        # Ensure that the Vary headers for the cached response match our
        # request
        for header, value in cached.get("vary", {}).items():
            if request.headers.get(header, None) != value:
                return

        body_raw = cached["response"].pop("body")

        headers = CaseInsensitiveDict(data=cached["response"]["headers"])
        if headers.get("transfer-encoding", "") == "chunked":
            headers.pop("transfer-encoding")

        cached["response"]["headers"] = headers

        try:
            body = io.BytesIO(body_raw)
        except TypeError:
            # This can happen if cachecontrol serialized to v1 format (pickle)
            # using Python 2. A Python 2 str(byte string) will be unpickled as
            # a Python 3 str (unicode string), which will cause the above to
            # fail with:
            #
            #     TypeError: 'str' does not support the buffer interface
            body = io.BytesIO(body_raw.encode("utf8"))

        return HTTPResponse(body=body, preload_content=False, **cached["response"])

    def _loads_v0(self, request, data):
        # The original legacy cache data. This doesn't contain enough
        # information to construct everything we need, so we'll treat this as
        # a miss.
        return

    def _loads_v1(self, request, data):
        try:
            cached = pickle.loads(data)
        except ValueError:
            return

        return self.prepare_response(request, cached)

    def _loads_v2(self, request, data):
        try:
            cached = json.loads(zlib.decompress(data).decode("utf8"))
        except (ValueError, zlib.error):
            return

        # We need to decode the items that we've base64 encoded
        cached["response"]["body"] = _b64_decode_bytes(cached["response"]["body"])
        cached["response"]["headers"] = dict(
            (_b64_decode_str(k), _b64_decode_str(v))
            for k, v in cached["response"]["headers"].items()
        )
        cached["response"]["reason"] = _b64_decode_str(cached["response"]["reason"])
        cached["vary"] = dict(
            (_b64_decode_str(k), _b64_decode_str(v) if v is not None else v)
            for k, v in cached["vary"].items()
        )

        return self.prepare_response(request, cached)

    def _loads_v3(self, request, data):
        # Due to Python 2 encoding issues, it's impossible to know for sure
        # exactly how to load v3 entries, thus we'll treat these as a miss so
        # that they get rewritten out as v4 entries.
        return

    def _loads_v4(self, request, data):
        try:
            cached = msgpack.loads(data, raw=False)
        except ValueError:
            return

        return self.prepare_response(request, cached)
site-packages/pip/_vendor/cachecontrol/cache.py000064400000001445150564745010015604 0ustar00"""
The cache object API for implementing caches. The default is a thread
safe in-memory dictionary.
"""
from threading import Lock


class BaseCache(object):

    def get(self, key):
        raise NotImplementedError()

    def set(self, key, value):
        raise NotImplementedError()

    def delete(self, key):
        raise NotImplementedError()

    def close(self):
        pass


class DictCache(BaseCache):

    def __init__(self, init_dict=None):
        self.lock = Lock()
        self.data = init_dict or {}

    def get(self, key):
        return self.data.get(key, None)

    def set(self, key, value):
        with self.lock:
            self.data.update({key: value})

    def delete(self, key):
        with self.lock:
            if key in self.data:
                self.data.pop(key)
site-packages/pip/_vendor/distro.py000064400000136436150564745010013412 0ustar00# Copyright 2015,2016,2017 Nir Cohen
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.

"""
The ``distro`` package (``distro`` stands for Linux Distribution) provides
information about the Linux distribution it runs on, such as a reliable
machine-readable distro ID, or version information.

It is the recommended replacement for Python's original
:py:func:`platform.linux_distribution` function, but it provides much more
functionality. An alternative implementation became necessary because Python
3.5 deprecated this function, and Python 3.8 removed it altogether. Its
predecessor function :py:func:`platform.dist` was already deprecated since
Python 2.6 and removed in Python 3.8. Still, there are many cases in which
access to OS distribution information is needed. See `Python issue 1322
<https://bugs.python.org/issue1322>`_ for more information.
"""

import argparse
import json
import logging
import os
import re
import shlex
import subprocess
import sys
import warnings

__version__ = "1.6.0"

# Use `if False` to avoid an ImportError on Python 2. After dropping Python 2
# support, can use typing.TYPE_CHECKING instead. See:
# https://docs.python.org/3/library/typing.html#typing.TYPE_CHECKING
if False:  # pragma: nocover
    from typing import (
        Any,
        Callable,
        Dict,
        Iterable,
        Optional,
        Sequence,
        TextIO,
        Tuple,
        Type,
        TypedDict,
        Union,
    )

    VersionDict = TypedDict(
        "VersionDict", {"major": str, "minor": str, "build_number": str}
    )
    InfoDict = TypedDict(
        "InfoDict",
        {
            "id": str,
            "version": str,
            "version_parts": VersionDict,
            "like": str,
            "codename": str,
        },
    )


_UNIXCONFDIR = os.environ.get("UNIXCONFDIR", "/etc")
_UNIXUSRLIBDIR = os.environ.get("UNIXUSRLIBDIR", "/usr/lib")
_OS_RELEASE_BASENAME = "os-release"

#: Translation table for normalizing the "ID" attribute defined in os-release
#: files, for use by the :func:`distro.id` method.
#:
#: * Key: Value as defined in the os-release file, translated to lower case,
#:   with blanks translated to underscores.
#:
#: * Value: Normalized value.
NORMALIZED_OS_ID = {
    "ol": "oracle",  # Oracle Linux
}

#: Translation table for normalizing the "Distributor ID" attribute returned by
#: the lsb_release command, for use by the :func:`distro.id` method.
#:
#: * Key: Value as returned by the lsb_release command, translated to lower
#:   case, with blanks translated to underscores.
#:
#: * Value: Normalized value.
NORMALIZED_LSB_ID = {
    "enterpriseenterpriseas": "oracle",  # Oracle Enterprise Linux 4
    "enterpriseenterpriseserver": "oracle",  # Oracle Linux 5
    "redhatenterpriseworkstation": "rhel",  # RHEL 6, 7 Workstation
    "redhatenterpriseserver": "rhel",  # RHEL 6, 7 Server
    "redhatenterprisecomputenode": "rhel",  # RHEL 6 ComputeNode
}

#: Translation table for normalizing the distro ID derived from the file name
#: of distro release files, for use by the :func:`distro.id` method.
#:
#: * Key: Value as derived from the file name of a distro release file,
#:   translated to lower case, with blanks translated to underscores.
#:
#: * Value: Normalized value.
NORMALIZED_DISTRO_ID = {
    "redhat": "rhel",  # RHEL 6.x, 7.x
}

# Pattern for content of distro release file (reversed)
_DISTRO_RELEASE_CONTENT_REVERSED_PATTERN = re.compile(
    r"(?:[^)]*\)(.*)\()? *(?:STL )?([\d.+\-a-z]*\d) *(?:esaeler *)?(.+)"
)

# Pattern for base file name of distro release file
_DISTRO_RELEASE_BASENAME_PATTERN = re.compile(r"(\w+)[-_](release|version)$")

# Base file names to be ignored when searching for distro release file
_DISTRO_RELEASE_IGNORE_BASENAMES = (
    "debian_version",
    "lsb-release",
    "oem-release",
    _OS_RELEASE_BASENAME,
    "system-release",
    "plesk-release",
    "iredmail-release",
)


def linux_distribution(full_distribution_name=True):
    # type: (bool) -> Tuple[str, str, str]
    """
    .. deprecated:: 1.6.0

        :func:`distro.linux_distribution()` is deprecated. It should only be
        used as a compatibility shim with Python's
        :py:func:`platform.linux_distribution()`. Please use :func:`distro.id`,
        :func:`distro.version` and :func:`distro.name` instead.

    Return information about the current OS distribution as a tuple
    ``(id_name, version, codename)`` with items as follows:

    * ``id_name``:  If *full_distribution_name* is false, the result of
      :func:`distro.id`. Otherwise, the result of :func:`distro.name`.

    * ``version``:  The result of :func:`distro.version`.

    * ``codename``:  The result of :func:`distro.codename`.

    The interface of this function is compatible with the original
    :py:func:`platform.linux_distribution` function, supporting a subset of
    its parameters.

    The data it returns may not exactly be the same, because it uses more data
    sources than the original function, and that may lead to different data if
    the OS distribution is not consistent across multiple data sources it
    provides (there are indeed such distributions ...).

    Another reason for differences is the fact that the :func:`distro.id`
    method normalizes the distro ID string to a reliable machine-readable value
    for a number of popular OS distributions.
    """
    warnings.warn(
        "distro.linux_distribution() is deprecated. It should only be used as a "
        "compatibility shim with Python's platform.linux_distribution(). Please use "
        "distro.id(), distro.version() and distro.name() instead.",
        DeprecationWarning,
        stacklevel=2,
    )
    return _distro.linux_distribution(full_distribution_name)


def id():
    # type: () -> str
    """
    Return the distro ID of the current distribution, as a
    machine-readable string.

    For a number of OS distributions, the returned distro ID value is
    *reliable*, in the sense that it is documented and that it does not change
    across releases of the distribution.

    This package maintains the following reliable distro ID values:

    ==============  =========================================
    Distro ID       Distribution
    ==============  =========================================
    "ubuntu"        Ubuntu
    "debian"        Debian
    "rhel"          RedHat Enterprise Linux
    "centos"        CentOS
    "fedora"        Fedora
    "sles"          SUSE Linux Enterprise Server
    "opensuse"      openSUSE
    "amazon"        Amazon Linux
    "arch"          Arch Linux
    "cloudlinux"    CloudLinux OS
    "exherbo"       Exherbo Linux
    "gentoo"        GenToo Linux
    "ibm_powerkvm"  IBM PowerKVM
    "kvmibm"        KVM for IBM z Systems
    "linuxmint"     Linux Mint
    "mageia"        Mageia
    "mandriva"      Mandriva Linux
    "parallels"     Parallels
    "pidora"        Pidora
    "raspbian"      Raspbian
    "oracle"        Oracle Linux (and Oracle Enterprise Linux)
    "scientific"    Scientific Linux
    "slackware"     Slackware
    "xenserver"     XenServer
    "openbsd"       OpenBSD
    "netbsd"        NetBSD
    "freebsd"       FreeBSD
    "midnightbsd"   MidnightBSD
    ==============  =========================================

    If you have a need to get distros for reliable IDs added into this set,
    or if you find that the :func:`distro.id` function returns a different
    distro ID for one of the listed distros, please create an issue in the
    `distro issue tracker`_.

    **Lookup hierarchy and transformations:**

    First, the ID is obtained from the following sources, in the specified
    order. The first available and non-empty value is used:

    * the value of the "ID" attribute of the os-release file,

    * the value of the "Distributor ID" attribute returned by the lsb_release
      command,

    * the first part of the file name of the distro release file,

    The so determined ID value then passes the following transformations,
    before it is returned by this method:

    * it is translated to lower case,

    * blanks (which should not be there anyway) are translated to underscores,

    * a normalization of the ID is performed, based upon
      `normalization tables`_. The purpose of this normalization is to ensure
      that the ID is as reliable as possible, even across incompatible changes
      in the OS distributions. A common reason for an incompatible change is
      the addition of an os-release file, or the addition of the lsb_release
      command, with ID values that differ from what was previously determined
      from the distro release file name.
    """
    return _distro.id()


def name(pretty=False):
    # type: (bool) -> str
    """
    Return the name of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the name is returned without version or codename.
    (e.g. "CentOS Linux")

    If *pretty* is true, the version and codename are appended.
    (e.g. "CentOS Linux 7.1.1503 (Core)")

    **Lookup hierarchy:**

    The name is obtained from the following sources, in the specified order.
    The first available and non-empty value is used:

    * If *pretty* is false:

      - the value of the "NAME" attribute of the os-release file,

      - the value of the "Distributor ID" attribute returned by the lsb_release
        command,

      - the value of the "<name>" field of the distro release file.

    * If *pretty* is true:

      - the value of the "PRETTY_NAME" attribute of the os-release file,

      - the value of the "Description" attribute returned by the lsb_release
        command,

      - the value of the "<name>" field of the distro release file, appended
        with the value of the pretty version ("<version_id>" and "<codename>"
        fields) of the distro release file, if available.
    """
    return _distro.name(pretty)


def version(pretty=False, best=False):
    # type: (bool, bool) -> str
    """
    Return the version of the current OS distribution, as a human-readable
    string.

    If *pretty* is false, the version is returned without codename (e.g.
    "7.0").

    If *pretty* is true, the codename in parenthesis is appended, if the
    codename is non-empty (e.g. "7.0 (Maipo)").

    Some distributions provide version numbers with different precisions in
    the different sources of distribution information. Examining the different
    sources in a fixed priority order does not always yield the most precise
    version (e.g. for Debian 8.2, or CentOS 7.1).

    The *best* parameter can be used to control the approach for the returned
    version:

    If *best* is false, the first non-empty version number in priority order of
    the examined sources is returned.

    If *best* is true, the most precise version number out of all examined
    sources is returned.

    **Lookup hierarchy:**

    In all cases, the version number is obtained from the following sources.
    If *best* is false, this order represents the priority order:

    * the value of the "VERSION_ID" attribute of the os-release file,
    * the value of the "Release" attribute returned by the lsb_release
      command,
    * the version number parsed from the "<version_id>" field of the first line
      of the distro release file,
    * the version number parsed from the "PRETTY_NAME" attribute of the
      os-release file, if it follows the format of the distro release files.
    * the version number parsed from the "Description" attribute returned by
      the lsb_release command, if it follows the format of the distro release
      files.
    """
    return _distro.version(pretty, best)


def version_parts(best=False):
    # type: (bool) -> Tuple[str, str, str]
    """
    Return the version of the current OS distribution as a tuple
    ``(major, minor, build_number)`` with items as follows:

    * ``major``:  The result of :func:`distro.major_version`.

    * ``minor``:  The result of :func:`distro.minor_version`.

    * ``build_number``:  The result of :func:`distro.build_number`.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    """
    return _distro.version_parts(best)


def major_version(best=False):
    # type: (bool) -> str
    """
    Return the major version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The major version is the first
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    """
    return _distro.major_version(best)


def minor_version(best=False):
    # type: (bool) -> str
    """
    Return the minor version of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The minor version is the second
    part of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    """
    return _distro.minor_version(best)


def build_number(best=False):
    # type: (bool) -> str
    """
    Return the build number of the current OS distribution, as a string,
    if provided.
    Otherwise, the empty string is returned. The build number is the third part
    of the dot-separated version string.

    For a description of the *best* parameter, see the :func:`distro.version`
    method.
    """
    return _distro.build_number(best)


def like():
    # type: () -> str
    """
    Return a space-separated list of distro IDs of distributions that are
    closely related to the current OS distribution in regards to packaging
    and programming interfaces, for example distributions the current
    distribution is a derivative from.

    **Lookup hierarchy:**

    This information item is only provided by the os-release file.
    For details, see the description of the "ID_LIKE" attribute in the
    `os-release man page
    <http://www.freedesktop.org/software/systemd/man/os-release.html>`_.
    """
    return _distro.like()


def codename():
    # type: () -> str
    """
    Return the codename for the release of the current OS distribution,
    as a string.

    If the distribution does not have a codename, an empty string is returned.

    Note that the returned codename is not always really a codename. For
    example, openSUSE returns "x86_64". This function does not handle such
    cases in any special way and just returns the string it finds, if any.

    **Lookup hierarchy:**

    * the codename within the "VERSION" attribute of the os-release file, if
      provided,

    * the value of the "Codename" attribute returned by the lsb_release
      command,

    * the value of the "<codename>" field of the distro release file.
    """
    return _distro.codename()


def info(pretty=False, best=False):
    # type: (bool, bool) -> InfoDict
    """
    Return certain machine-readable information items about the current OS
    distribution in a dictionary, as shown in the following example:

    .. sourcecode:: python

        {
            'id': 'rhel',
            'version': '7.0',
            'version_parts': {
                'major': '7',
                'minor': '0',
                'build_number': ''
            },
            'like': 'fedora',
            'codename': 'Maipo'
        }

    The dictionary structure and keys are always the same, regardless of which
    information items are available in the underlying data sources. The values
    for the various keys are as follows:

    * ``id``:  The result of :func:`distro.id`.

    * ``version``:  The result of :func:`distro.version`.

    * ``version_parts -> major``:  The result of :func:`distro.major_version`.

    * ``version_parts -> minor``:  The result of :func:`distro.minor_version`.

    * ``version_parts -> build_number``:  The result of
      :func:`distro.build_number`.

    * ``like``:  The result of :func:`distro.like`.

    * ``codename``:  The result of :func:`distro.codename`.

    For a description of the *pretty* and *best* parameters, see the
    :func:`distro.version` method.
    """
    return _distro.info(pretty, best)


def os_release_info():
    # type: () -> Dict[str, str]
    """
    Return a dictionary containing key-value pairs for the information items
    from the os-release file data source of the current OS distribution.

    See `os-release file`_ for details about these information items.
    """
    return _distro.os_release_info()


def lsb_release_info():
    # type: () -> Dict[str, str]
    """
    Return a dictionary containing key-value pairs for the information items
    from the lsb_release command data source of the current OS distribution.

    See `lsb_release command output`_ for details about these information
    items.
    """
    return _distro.lsb_release_info()


def distro_release_info():
    # type: () -> Dict[str, str]
    """
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.

    See `distro release file`_ for details about these information items.
    """
    return _distro.distro_release_info()


def uname_info():
    # type: () -> Dict[str, str]
    """
    Return a dictionary containing key-value pairs for the information items
    from the distro release file data source of the current OS distribution.
    """
    return _distro.uname_info()


def os_release_attr(attribute):
    # type: (str) -> str
    """
    Return a single named information item from the os-release file data source
    of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `os-release file`_ for details about these information items.
    """
    return _distro.os_release_attr(attribute)


def lsb_release_attr(attribute):
    # type: (str) -> str
    """
    Return a single named information item from the lsb_release command output
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `lsb_release command output`_ for details about these information
    items.
    """
    return _distro.lsb_release_attr(attribute)


def distro_release_attr(attribute):
    # type: (str) -> str
    """
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
      The empty string, if the item does not exist.

    See `distro release file`_ for details about these information items.
    """
    return _distro.distro_release_attr(attribute)


def uname_attr(attribute):
    # type: (str) -> str
    """
    Return a single named information item from the distro release file
    data source of the current OS distribution.

    Parameters:

    * ``attribute`` (string): Key of the information item.

    Returns:

    * (string): Value of the information item, if the item exists.
                The empty string, if the item does not exist.
    """
    return _distro.uname_attr(attribute)


try:
    from functools import cached_property
except ImportError:
    # Python < 3.8
    class cached_property(object):  # type: ignore
        """A version of @property which caches the value.  On access, it calls the
        underlying function and sets the value in `__dict__` so future accesses
        will not re-call the property.
        """

        def __init__(self, f):
            # type: (Callable[[Any], Any]) -> None
            self._fname = f.__name__
            self._f = f

        def __get__(self, obj, owner):
            # type: (Any, Type[Any]) -> Any
            assert obj is not None, "call {} on an instance".format(self._fname)
            ret = obj.__dict__[self._fname] = self._f(obj)
            return ret


class LinuxDistribution(object):
    """
    Provides information about a OS distribution.

    This package creates a private module-global instance of this class with
    default initialization arguments, that is used by the
    `consolidated accessor functions`_ and `single source accessor functions`_.
    By using default initialization arguments, that module-global instance
    returns data about the current OS distribution (i.e. the distro this
    package runs on).

    Normally, it is not necessary to create additional instances of this class.
    However, in situations where control is needed over the exact data sources
    that are used, instances of this class can be created with a specific
    distro release file, or a specific os-release file, or without invoking the
    lsb_release command.
    """

    def __init__(
        self,
        include_lsb=True,
        os_release_file="",
        distro_release_file="",
        include_uname=True,
        root_dir=None,
    ):
        # type: (bool, str, str, bool, Optional[str]) -> None
        """
        The initialization method of this class gathers information from the
        available data sources, and stores that in private instance attributes.
        Subsequent access to the information items uses these private instance
        attributes, so that the data sources are read only once.

        Parameters:

        * ``include_lsb`` (bool): Controls whether the
          `lsb_release command output`_ is included as a data source.

          If the lsb_release command is not available in the program execution
          path, the data source for the lsb_release command will be empty.

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is to be used as a data source.

          An empty string (the default) will cause the default path name to
          be used (see `os-release file`_ for details).

          If the specified or defaulted os-release file does not exist, the
          data source for the os-release file will be empty.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is to be used as a data source.

          An empty string (the default) will cause a default search algorithm
          to be used (see `distro release file`_ for details).

          If the specified distro release file does not exist, or if no default
          distro release file can be found, the data source for the distro
          release file will be empty.

        * ``include_uname`` (bool): Controls whether uname command output is
          included as a data source. If the uname command is not available in
          the program execution path the data source for the uname command will
          be empty.

        * ``root_dir`` (string): The absolute path to the root directory to use
          to find distro-related information files.

        Public instance attributes:

        * ``os_release_file`` (string): The path name of the
          `os-release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``distro_release_file`` (string): The path name of the
          `distro release file`_ that is actually used as a data source. The
          empty string if no distro release file is used as a data source.

        * ``include_lsb`` (bool): The result of the ``include_lsb`` parameter.
          This controls whether the lsb information will be loaded.

        * ``include_uname`` (bool): The result of the ``include_uname``
          parameter. This controls whether the uname information will
          be loaded.

        Raises:

        * :py:exc:`IOError`: Some I/O issue with an os-release file or distro
          release file.

        * :py:exc:`subprocess.CalledProcessError`: The lsb_release command had
          some issue (other than not being available in the program execution
          path).

        * :py:exc:`UnicodeError`: A data source has unexpected characters or
          uses an unexpected encoding.
        """
        self.root_dir = root_dir
        self.etc_dir = os.path.join(root_dir, "etc") if root_dir else _UNIXCONFDIR
        self.usr_lib_dir = (
            os.path.join(root_dir, "usr/lib") if root_dir else _UNIXUSRLIBDIR
        )

        if os_release_file:
            self.os_release_file = os_release_file
        else:
            etc_dir_os_release_file = os.path.join(self.etc_dir, _OS_RELEASE_BASENAME)
            usr_lib_os_release_file = os.path.join(
                self.usr_lib_dir, _OS_RELEASE_BASENAME
            )

            # NOTE: The idea is to respect order **and** have it set
            #       at all times for API backwards compatibility.
            if os.path.isfile(etc_dir_os_release_file) or not os.path.isfile(
                usr_lib_os_release_file
            ):
                self.os_release_file = etc_dir_os_release_file
            else:
                self.os_release_file = usr_lib_os_release_file

        self.distro_release_file = distro_release_file or ""  # updated later
        self.include_lsb = include_lsb
        self.include_uname = include_uname

    def __repr__(self):
        # type: () -> str
        """Return repr of all info"""
        return (
            "LinuxDistribution("
            "os_release_file={self.os_release_file!r}, "
            "distro_release_file={self.distro_release_file!r}, "
            "include_lsb={self.include_lsb!r}, "
            "include_uname={self.include_uname!r}, "
            "_os_release_info={self._os_release_info!r}, "
            "_lsb_release_info={self._lsb_release_info!r}, "
            "_distro_release_info={self._distro_release_info!r}, "
            "_uname_info={self._uname_info!r})".format(self=self)
        )

    def linux_distribution(self, full_distribution_name=True):
        # type: (bool) -> Tuple[str, str, str]
        """
        Return information about the OS distribution that is compatible
        with Python's :func:`platform.linux_distribution`, supporting a subset
        of its parameters.

        For details, see :func:`distro.linux_distribution`.
        """
        return (
            self.name() if full_distribution_name else self.id(),
            self.version(),
            self.codename(),
        )

    def id(self):
        # type: () -> str
        """Return the distro ID of the OS distribution, as a string.

        For details, see :func:`distro.id`.
        """

        def normalize(distro_id, table):
            # type: (str, Dict[str, str]) -> str
            distro_id = distro_id.lower().replace(" ", "_")
            return table.get(distro_id, distro_id)

        distro_id = self.os_release_attr("id")
        if distro_id:
            return normalize(distro_id, NORMALIZED_OS_ID)

        distro_id = self.lsb_release_attr("distributor_id")
        if distro_id:
            return normalize(distro_id, NORMALIZED_LSB_ID)

        distro_id = self.distro_release_attr("id")
        if distro_id:
            return normalize(distro_id, NORMALIZED_DISTRO_ID)

        distro_id = self.uname_attr("id")
        if distro_id:
            return normalize(distro_id, NORMALIZED_DISTRO_ID)

        return ""

    def name(self, pretty=False):
        # type: (bool) -> str
        """
        Return the name of the OS distribution, as a string.

        For details, see :func:`distro.name`.
        """
        name = (
            self.os_release_attr("name")
            or self.lsb_release_attr("distributor_id")
            or self.distro_release_attr("name")
            or self.uname_attr("name")
        )
        if pretty:
            name = self.os_release_attr("pretty_name") or self.lsb_release_attr(
                "description"
            )
            if not name:
                name = self.distro_release_attr("name") or self.uname_attr("name")
                version = self.version(pretty=True)
                if version:
                    name = name + " " + version
        return name or ""

    def version(self, pretty=False, best=False):
        # type: (bool, bool) -> str
        """
        Return the version of the OS distribution, as a string.

        For details, see :func:`distro.version`.
        """
        versions = [
            self.os_release_attr("version_id"),
            self.lsb_release_attr("release"),
            self.distro_release_attr("version_id"),
            self._parse_distro_release_content(self.os_release_attr("pretty_name")).get(
                "version_id", ""
            ),
            self._parse_distro_release_content(
                self.lsb_release_attr("description")
            ).get("version_id", ""),
            self.uname_attr("release"),
        ]
        version = ""
        if best:
            # This algorithm uses the last version in priority order that has
            # the best precision. If the versions are not in conflict, that
            # does not matter; otherwise, using the last one instead of the
            # first one might be considered a surprise.
            for v in versions:
                if v.count(".") > version.count(".") or version == "":
                    version = v
        else:
            for v in versions:
                if v != "":
                    version = v
                    break
        if pretty and version and self.codename():
            version = "{0} ({1})".format(version, self.codename())
        return version

    def version_parts(self, best=False):
        # type: (bool) -> Tuple[str, str, str]
        """
        Return the version of the OS distribution, as a tuple of version
        numbers.

        For details, see :func:`distro.version_parts`.
        """
        version_str = self.version(best=best)
        if version_str:
            version_regex = re.compile(r"(\d+)\.?(\d+)?\.?(\d+)?")
            matches = version_regex.match(version_str)
            if matches:
                major, minor, build_number = matches.groups()
                return major, minor or "", build_number or ""
        return "", "", ""

    def major_version(self, best=False):
        # type: (bool) -> str
        """
        Return the major version number of the current distribution.

        For details, see :func:`distro.major_version`.
        """
        return self.version_parts(best)[0]

    def minor_version(self, best=False):
        # type: (bool) -> str
        """
        Return the minor version number of the current distribution.

        For details, see :func:`distro.minor_version`.
        """
        return self.version_parts(best)[1]

    def build_number(self, best=False):
        # type: (bool) -> str
        """
        Return the build number of the current distribution.

        For details, see :func:`distro.build_number`.
        """
        return self.version_parts(best)[2]

    def like(self):
        # type: () -> str
        """
        Return the IDs of distributions that are like the OS distribution.

        For details, see :func:`distro.like`.
        """
        return self.os_release_attr("id_like") or ""

    def codename(self):
        # type: () -> str
        """
        Return the codename of the OS distribution.

        For details, see :func:`distro.codename`.
        """
        try:
            # Handle os_release specially since distros might purposefully set
            # this to empty string to have no codename
            return self._os_release_info["codename"]
        except KeyError:
            return (
                self.lsb_release_attr("codename")
                or self.distro_release_attr("codename")
                or ""
            )

    def info(self, pretty=False, best=False):
        # type: (bool, bool) -> InfoDict
        """
        Return certain machine-readable information about the OS
        distribution.

        For details, see :func:`distro.info`.
        """
        return dict(
            id=self.id(),
            version=self.version(pretty, best),
            version_parts=dict(
                major=self.major_version(best),
                minor=self.minor_version(best),
                build_number=self.build_number(best),
            ),
            like=self.like(),
            codename=self.codename(),
        )

    def os_release_info(self):
        # type: () -> Dict[str, str]
        """
        Return a dictionary containing key-value pairs for the information
        items from the os-release file data source of the OS distribution.

        For details, see :func:`distro.os_release_info`.
        """
        return self._os_release_info

    def lsb_release_info(self):
        # type: () -> Dict[str, str]
        """
        Return a dictionary containing key-value pairs for the information
        items from the lsb_release command data source of the OS
        distribution.

        For details, see :func:`distro.lsb_release_info`.
        """
        return self._lsb_release_info

    def distro_release_info(self):
        # type: () -> Dict[str, str]
        """
        Return a dictionary containing key-value pairs for the information
        items from the distro release file data source of the OS
        distribution.

        For details, see :func:`distro.distro_release_info`.
        """
        return self._distro_release_info

    def uname_info(self):
        # type: () -> Dict[str, str]
        """
        Return a dictionary containing key-value pairs for the information
        items from the uname command data source of the OS distribution.

        For details, see :func:`distro.uname_info`.
        """
        return self._uname_info

    def os_release_attr(self, attribute):
        # type: (str) -> str
        """
        Return a single named information item from the os-release file data
        source of the OS distribution.

        For details, see :func:`distro.os_release_attr`.
        """
        return self._os_release_info.get(attribute, "")

    def lsb_release_attr(self, attribute):
        # type: (str) -> str
        """
        Return a single named information item from the lsb_release command
        output data source of the OS distribution.

        For details, see :func:`distro.lsb_release_attr`.
        """
        return self._lsb_release_info.get(attribute, "")

    def distro_release_attr(self, attribute):
        # type: (str) -> str
        """
        Return a single named information item from the distro release file
        data source of the OS distribution.

        For details, see :func:`distro.distro_release_attr`.
        """
        return self._distro_release_info.get(attribute, "")

    def uname_attr(self, attribute):
        # type: (str) -> str
        """
        Return a single named information item from the uname command
        output data source of the OS distribution.

        For details, see :func:`distro.uname_attr`.
        """
        return self._uname_info.get(attribute, "")

    @cached_property
    def _os_release_info(self):
        # type: () -> Dict[str, str]
        """
        Get the information items from the specified os-release file.

        Returns:
            A dictionary containing all information items.
        """
        if os.path.isfile(self.os_release_file):
            with open(self.os_release_file) as release_file:
                return self._parse_os_release_content(release_file)
        return {}

    @staticmethod
    def _parse_os_release_content(lines):
        # type: (TextIO) -> Dict[str, str]
        """
        Parse the lines of an os-release file.

        Parameters:

        * lines: Iterable through the lines in the os-release file.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        """
        props = {}
        lexer = shlex.shlex(lines, posix=True)
        lexer.whitespace_split = True

        # The shlex module defines its `wordchars` variable using literals,
        # making it dependent on the encoding of the Python source file.
        # In Python 2.6 and 2.7, the shlex source file is encoded in
        # 'iso-8859-1', and the `wordchars` variable is defined as a byte
        # string. This causes a UnicodeDecodeError to be raised when the
        # parsed content is a unicode object. The following fix resolves that
        # (... but it should be fixed in shlex...):
        if sys.version_info[0] == 2 and isinstance(lexer.wordchars, bytes):
            lexer.wordchars = lexer.wordchars.decode("iso-8859-1")

        tokens = list(lexer)
        for token in tokens:
            # At this point, all shell-like parsing has been done (i.e.
            # comments processed, quotes and backslash escape sequences
            # processed, multi-line values assembled, trailing newlines
            # stripped, etc.), so the tokens are now either:
            # * variable assignments: var=value
            # * commands or their arguments (not allowed in os-release)
            if "=" in token:
                k, v = token.split("=", 1)
                props[k.lower()] = v
            else:
                # Ignore any tokens that are not variable assignments
                pass

        if "version_codename" in props:
            # os-release added a version_codename field.  Use that in
            # preference to anything else Note that some distros purposefully
            # do not have code names.  They should be setting
            # version_codename=""
            props["codename"] = props["version_codename"]
        elif "ubuntu_codename" in props:
            # Same as above but a non-standard field name used on older Ubuntus
            props["codename"] = props["ubuntu_codename"]
        elif "version" in props:
            # If there is no version_codename, parse it from the version
            match = re.search(r"(\(\D+\))|,(\s+)?\D+", props["version"])
            if match:
                codename = match.group()
                codename = codename.strip("()")
                codename = codename.strip(",")
                codename = codename.strip()
                # codename appears within paranthese.
                props["codename"] = codename

        return props

    @cached_property
    def _lsb_release_info(self):
        # type: () -> Dict[str, str]
        """
        Get the information items from the lsb_release command output.

        Returns:
            A dictionary containing all information items.
        """
        if not self.include_lsb:
            return {}
        with open(os.devnull, "wb") as devnull:
            try:
                cmd = ("lsb_release", "-a")
                stdout = subprocess.check_output(cmd, stderr=devnull)
            # Command not found or lsb_release returned error
            except (OSError, subprocess.CalledProcessError):
                return {}
        content = self._to_str(stdout).splitlines()
        return self._parse_lsb_release_content(content)

    @staticmethod
    def _parse_lsb_release_content(lines):
        # type: (Iterable[str]) -> Dict[str, str]
        """
        Parse the output of the lsb_release command.

        Parameters:

        * lines: Iterable through the lines of the lsb_release output.
                 Each line must be a unicode string or a UTF-8 encoded byte
                 string.

        Returns:
            A dictionary containing all information items.
        """
        props = {}
        for line in lines:
            kv = line.strip("\n").split(":", 1)
            if len(kv) != 2:
                # Ignore lines without colon.
                continue
            k, v = kv
            props.update({k.replace(" ", "_").lower(): v.strip()})
        return props

    @cached_property
    def _uname_info(self):
        # type: () -> Dict[str, str]
        with open(os.devnull, "wb") as devnull:
            try:
                cmd = ("uname", "-rs")
                stdout = subprocess.check_output(cmd, stderr=devnull)
            except OSError:
                return {}
        content = self._to_str(stdout).splitlines()
        return self._parse_uname_content(content)

    @staticmethod
    def _parse_uname_content(lines):
        # type: (Sequence[str]) -> Dict[str, str]
        props = {}
        match = re.search(r"^([^\s]+)\s+([\d\.]+)", lines[0].strip())
        if match:
            name, version = match.groups()

            # This is to prevent the Linux kernel version from
            # appearing as the 'best' version on otherwise
            # identifiable distributions.
            if name == "Linux":
                return {}
            props["id"] = name.lower()
            props["name"] = name
            props["release"] = version
        return props

    @staticmethod
    def _to_str(text):
        # type: (Union[bytes, str]) -> str
        encoding = sys.getfilesystemencoding()
        encoding = "utf-8" if encoding == "ascii" else encoding

        if sys.version_info[0] >= 3:
            if isinstance(text, bytes):
                return text.decode(encoding)
        else:
            if isinstance(text, unicode):  # noqa
                return text.encode(encoding)

        return text

    @cached_property
    def _distro_release_info(self):
        # type: () -> Dict[str, str]
        """
        Get the information items from the specified distro release file.

        Returns:
            A dictionary containing all information items.
        """
        if self.distro_release_file:
            # If it was specified, we use it and parse what we can, even if
            # its file name or content does not match the expected pattern.
            distro_info = self._parse_distro_release_file(self.distro_release_file)
            basename = os.path.basename(self.distro_release_file)
            # The file name pattern for user-specified distro release files
            # is somewhat more tolerant (compared to when searching for the
            # file), because we want to use what was specified as best as
            # possible.
            match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename)
            if "name" in distro_info and "cloudlinux" in distro_info["name"].lower():
                distro_info["id"] = "cloudlinux"
            elif match:
                distro_info["id"] = match.group(1)
            return distro_info
        else:
            try:
                basenames = os.listdir(self.etc_dir)
                # We sort for repeatability in cases where there are multiple
                # distro specific files; e.g. CentOS, Oracle, Enterprise all
                # containing `redhat-release` on top of their own.
                basenames.sort()
            except OSError:
                # This may occur when /etc is not readable but we can't be
                # sure about the *-release files. Check common entries of
                # /etc for information. If they turn out to not be there the
                # error is handled in `_parse_distro_release_file()`.
                basenames = [
                    "SuSE-release",
                    "arch-release",
                    "base-release",
                    "centos-release",
                    "fedora-release",
                    "gentoo-release",
                    "mageia-release",
                    "mandrake-release",
                    "mandriva-release",
                    "mandrivalinux-release",
                    "manjaro-release",
                    "oracle-release",
                    "redhat-release",
                    "sl-release",
                    "slackware-version",
                ]
            for basename in basenames:
                if basename in _DISTRO_RELEASE_IGNORE_BASENAMES:
                    continue
                match = _DISTRO_RELEASE_BASENAME_PATTERN.match(basename)
                if match:
                    filepath = os.path.join(self.etc_dir, basename)
                    distro_info = self._parse_distro_release_file(filepath)
                    if "name" in distro_info:
                        # The name is always present if the pattern matches
                        self.distro_release_file = filepath
                        distro_info["id"] = match.group(1)
                        if "cloudlinux" in distro_info["name"].lower():
                            distro_info["id"] = "cloudlinux"
                        return distro_info
            return {}

    def _parse_distro_release_file(self, filepath):
        # type: (str) -> Dict[str, str]
        """
        Parse a distro release file.

        Parameters:

        * filepath: Path name of the distro release file.

        Returns:
            A dictionary containing all information items.
        """
        try:
            with open(filepath) as fp:
                # Only parse the first line. For instance, on SLES there
                # are multiple lines. We don't want them...
                return self._parse_distro_release_content(fp.readline())
        except (OSError, IOError):
            # Ignore not being able to read a specific, seemingly version
            # related file.
            # See https://github.com/python-distro/distro/issues/162
            return {}

    @staticmethod
    def _parse_distro_release_content(line):
        # type: (str) -> Dict[str, str]
        """
        Parse a line from a distro release file.

        Parameters:
        * line: Line from the distro release file. Must be a unicode string
                or a UTF-8 encoded byte string.

        Returns:
            A dictionary containing all information items.
        """
        matches = _DISTRO_RELEASE_CONTENT_REVERSED_PATTERN.match(line.strip()[::-1])
        distro_info = {}
        if matches:
            # regexp ensures non-None
            distro_info["name"] = matches.group(3)[::-1]
            if matches.group(2):
                distro_info["version_id"] = matches.group(2)[::-1]
            if matches.group(1):
                distro_info["codename"] = matches.group(1)[::-1]
        elif line:
            distro_info["name"] = line.strip()
        return distro_info


_distro = LinuxDistribution()


def main():
    # type: () -> None
    logger = logging.getLogger(__name__)
    logger.setLevel(logging.DEBUG)
    logger.addHandler(logging.StreamHandler(sys.stdout))

    parser = argparse.ArgumentParser(description="OS distro info tool")
    parser.add_argument(
        "--json", "-j", help="Output in machine readable format", action="store_true"
    )

    parser.add_argument(
        "--root-dir",
        "-r",
        type=str,
        dest="root_dir",
        help="Path to the root filesystem directory (defaults to /)",
    )

    args = parser.parse_args()

    if args.root_dir:
        dist = LinuxDistribution(
            include_lsb=False, include_uname=False, root_dir=args.root_dir
        )
    else:
        dist = _distro

    if args.json:
        logger.info(json.dumps(dist.info(), indent=4, sort_keys=True))
    else:
        logger.info("Name: %s", dist.name(pretty=True))
        distribution_version = dist.version(pretty=True)
        logger.info("Version: %s", distribution_version)
        distribution_codename = dist.codename()
        logger.info("Codename: %s", distribution_codename)


if __name__ == "__main__":
    main()
site-packages/pip/_internal/main.py000064400000000524150564745010013335 0ustar00from typing import List, Optional


def main(args: Optional[List[str]] = None) -> int:
    """This is preserved for old console scripts that may still be referencing
    it.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    """
    from pip._internal.utils.entrypoints import _wrapper

    return _wrapper(args)
site-packages/pip/_internal/self_outdated_check.py000064400000014371150564745010016375 0ustar00import datetime
import hashlib
import json
import logging
import optparse
import os.path
import sys
from typing import Any, Dict

from pip._vendor.packaging.version import parse as parse_version

from pip._internal.index.collector import LinkCollector
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import get_default_environment
from pip._internal.models.selection_prefs import SelectionPreferences
from pip._internal.network.session import PipSession
from pip._internal.utils.filesystem import adjacent_tmp_file, check_path_owner, replace
from pip._internal.utils.misc import ensure_dir

SELFCHECK_DATE_FMT = "%Y-%m-%dT%H:%M:%SZ"


logger = logging.getLogger(__name__)


def _get_statefile_name(key: str) -> str:
    key_bytes = key.encode()
    name = hashlib.sha224(key_bytes).hexdigest()
    return name


class SelfCheckState:
    def __init__(self, cache_dir: str) -> None:
        self.state: Dict[str, Any] = {}
        self.statefile_path = None

        # Try to load the existing state
        if cache_dir:
            self.statefile_path = os.path.join(
                cache_dir, "selfcheck", _get_statefile_name(self.key)
            )
            try:
                with open(self.statefile_path, encoding="utf-8") as statefile:
                    self.state = json.load(statefile)
            except (OSError, ValueError, KeyError):
                # Explicitly suppressing exceptions, since we don't want to
                # error out if the cache file is invalid.
                pass

    @property
    def key(self) -> str:
        return sys.prefix

    def save(self, pypi_version: str, current_time: datetime.datetime) -> None:
        # If we do not have a path to cache in, don't bother saving.
        if not self.statefile_path:
            return

        # Check to make sure that we own the directory
        if not check_path_owner(os.path.dirname(self.statefile_path)):
            return

        # Now that we've ensured the directory is owned by this user, we'll go
        # ahead and make sure that all our directories are created.
        ensure_dir(os.path.dirname(self.statefile_path))

        state = {
            # Include the key so it's easy to tell which pip wrote the
            # file.
            "key": self.key,
            "last_check": current_time.strftime(SELFCHECK_DATE_FMT),
            "pypi_version": pypi_version,
        }

        text = json.dumps(state, sort_keys=True, separators=(",", ":"))

        with adjacent_tmp_file(self.statefile_path) as f:
            f.write(text.encode())

        try:
            # Since we have a prefix-specific state file, we can just
            # overwrite whatever is there, no need to check.
            replace(f.name, self.statefile_path)
        except OSError:
            # Best effort.
            pass


def was_installed_by_pip(pkg: str) -> bool:
    """Checks whether pkg was installed by pip

    This is used not to display the upgrade message when pip is in fact
    installed by system package manager, such as dnf on Fedora.
    """
    dist = get_default_environment().get_distribution(pkg)
    return dist is not None and "pip" == dist.installer


def pip_self_version_check(session: PipSession, options: optparse.Values) -> None:
    """Check for an update for pip.

    Limit the frequency of checks to once per week. State is stored either in
    the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix
    of the pip script path.
    """
    installed_dist = get_default_environment().get_distribution("pip")
    if not installed_dist:
        return

    pip_version = installed_dist.version
    pypi_version = None

    try:
        state = SelfCheckState(cache_dir=options.cache_dir)

        current_time = datetime.datetime.utcnow()
        # Determine if we need to refresh the state
        if "last_check" in state.state and "pypi_version" in state.state:
            last_check = datetime.datetime.strptime(
                state.state["last_check"], SELFCHECK_DATE_FMT
            )
            if (current_time - last_check).total_seconds() < 7 * 24 * 60 * 60:
                pypi_version = state.state["pypi_version"]

        # Refresh the version if we need to or just see if we need to warn
        if pypi_version is None:
            # Lets use PackageFinder to see what the latest pip version is
            link_collector = LinkCollector.create(
                session,
                options=options,
                suppress_no_index=True,
            )

            # Pass allow_yanked=False so we don't suggest upgrading to a
            # yanked version.
            selection_prefs = SelectionPreferences(
                allow_yanked=False,
                allow_all_prereleases=False,  # Explicitly set to False
            )

            finder = PackageFinder.create(
                link_collector=link_collector,
                selection_prefs=selection_prefs,
            )
            best_candidate = finder.find_best_candidate("pip").best_candidate
            if best_candidate is None:
                return
            pypi_version = str(best_candidate.version)

            # save that we've performed a check
            state.save(pypi_version, current_time)

        remote_version = parse_version(pypi_version)

        local_version_is_older = (
            pip_version < remote_version
            and pip_version.base_version != remote_version.base_version
            and was_installed_by_pip("pip")
        )

        # Determine if our pypi_version is older
        if not local_version_is_older:
            return

        # We cannot tell how the current pip is available in the current
        # command context, so be pragmatic here and suggest the command
        # that's always available. This does not accommodate spaces in
        # `sys.executable`.
        pip_cmd = f"{sys.executable} -m pip"
        logger.warning(
            "You are using pip version %s; however, version %s is "
            "available.\nYou should consider upgrading via the "
            "'%s install --upgrade pip' command.",
            pip_version,
            pypi_version,
            pip_cmd,
        )
    except Exception:
        logger.debug(
            "There was an error checking the latest version of pip",
            exc_info=True,
        )
site-packages/pip/_internal/__pycache__/main.cpython-39.pyc000064400000001220150564745010017616 0ustar00a

��ReT�@s.ddlmZmZdeeeed�dd�ZdS)�)�List�OptionalN)�args�returncCsddlm}||�S)z�This is preserved for old console scripts that may still be referencing
    it.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    r)�_wrapper)�pip._internal.utils.entrypointsr)rr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/main.py�mainsr
)N)�typingrr�str�intr
rrrr	�<module>ssite-packages/pip/_internal/__pycache__/exceptions.cpython-39.pyc000064400000040032150564745010021057 0ustar00a

��Re�1�@s4dZddlZddlmZmZmZddlmZmZm	Z	m
Z
mZddlm
Z
ddlmZmZer�ddlmZddlmZdd	lmZGd
d�de�ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�ZGdd�de�Z Gdd�de�Z!Gdd�de�Z"Gd d!�d!e�Z#Gd"d#�d#e�Z$Gd$d%�d%e�Z%Gd&d'�d'e�Z&Gd(d)�d)e�Z'Gd*d+�d+e�Z(Gd,d-�d-e�Z)Gd.d/�d/e�Z*Gd0d1�d1e�Z+Gd2d3�d3e+�Z,Gd4d5�d5e+�Z-Gd6d7�d7e+�Z.Gd8d9�d9e+�Z/Gd:d;�d;e+�Z0Gd<d=�d=e�Z1Gd>d?�d?e�Z2dS)@z"Exceptions used throughout package�N)�chain�groupby�repeat)�
TYPE_CHECKING�Dict�List�Optional�Union)�Distribution)�Request�Response)�_Hash)�BaseDistribution)�InstallRequirementc@seZdZdZdS)�PipErrorzBase pip exceptionN��__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/exceptions.pyrsrc@seZdZdZdS)�ConfigurationErrorz"General exception in configurationNrrrrrrsrc@seZdZdZdS)�InstallationErrorz%General exception during installationNrrrrrrsrc@seZdZdZdS)�UninstallationErrorz'General exception during uninstallationNrrrrrrsrc@s8eZdZdZeedfedd�dd�Zed�dd	�ZdS)
�NoneMetadataErrora
    Raised when accessing "METADATA" or "PKG-INFO" metadata for a
    pip._vendor.pkg_resources.Distribution object and
    `dist.has_metadata('METADATA')` returns True but
    `dist.get_metadata('METADATA')` returns None (and similarly for
    "PKG-INFO").
    rN)�dist�
metadata_name�returncCs||_||_dS)z�
        :param dist: A Distribution object.
        :param metadata_name: The name of the metadata being accessed
            (can be "METADATA" or "PKG-INFO").
        N)rr)�selfrrrrr�__init__*s
zNoneMetadataError.__init__�rcCsd�|j|j�S)Nz+None {} metadata found for distribution: {})�formatrr�rrrr�__str__7s�zNoneMetadataError.__str__)	rrrrr	r
�strr r$rrrrr!s

�
rc@seZdZdZed�dd�ZdS)�UserInstallationInvalidzBA --user install is requested on an environment without user site.r!cCsdS)Nz$User base directory is not specifiedrr#rrrr$CszUserInstallationInvalid.__str__N)rrrrr%r$rrrrr&@sr&c@seZdZed�dd�ZdS)�InvalidSchemeCombinationr!cCs6d�dd�|jdd�D��}d|�d|jd�d�S)Nz, css|]}t|�VqdS�N)r%)�.0�arrr�	<genexpr>I�z3InvalidSchemeCombination.__str__.<locals>.<genexpr>���zCannot set z and z	 together)�join�args)r�beforerrrr$Hsz InvalidSchemeCombination.__str__N)rrrr%r$rrrrr'Gsr'c@seZdZdZdS)�DistributionNotFoundzCRaised when a distribution cannot be found to satisfy a requirementNrrrrrr1Msr1c@seZdZdZdS)�RequirementsFileParseErrorzDRaised when a general error occurs parsing a requirements file line.Nrrrrrr2Qsr2c@seZdZdZdS)�BestVersionAlreadyInstalledzNRaised when the most up-to-date version of a package is already
    installed.Nrrrrrr3Usr3c@seZdZdZdS)�
BadCommandz0Raised when virtualenv or a command is not foundNrrrrrr4Zsr4c@seZdZdZdS)�CommandErrorz7Raised when there is an error in command-line argumentsNrrrrrr5^sr5c@seZdZdZdS)�PreviousBuildDirErrorz:Raised when there's a previous conflicting build directoryNrrrrrr6bsr6cs<eZdZdZd	eeedd��fdd�
Zed�dd�Z�Z	S)
�NetworkConnectionErrorzHTTP connection errorN)�	error_msg�response�requestrcsJ||_||_||_|jdur6|js6t|d�r6|jj|_t��|||�dS)zc
        Initialize NetworkConnectionError with  `request` and `response`
        objects.
        Nr:)r9r:r8�hasattr�superr )rr8r9r:��	__class__rrr is���
zNetworkConnectionError.__init__r!cCs
t|j�Sr()r%r8r#rrrr${szNetworkConnectionError.__str__)NN)
rrrrr%rrr r$�
__classcell__rrr=rr7fs��r7c@seZdZdZdS)�InvalidWheelFilenamezInvalid wheel filename.Nrrrrrr@sr@c@seZdZdZdS)�UnsupportedWheelzUnsupported wheel.NrrrrrrA�srAc@s4eZdZdZdeeedd�dd�Zed�dd	�ZdS)
�MetadataInconsistentz�Built metadata contains inconsistent information.

    This is raised when the metadata contains values (e.g. name and version)
    that do not match the information previously obtained from sdist filename
    or user-supplied ``#egg=`` value.
    rN)�ireq�field�f_val�m_valrcCs||_||_||_||_dSr()rCrDrErF)rrCrDrErFrrrr �szMetadataInconsistent.__init__r!cCsd}|�|j|j|j|j�S)NzJRequested {} has inconsistent {}: filename has {!r}, but metadata has {!r})r"rCrDrErF)r�templaterrrr$�s�zMetadataInconsistent.__str__)rrrrr%r r$rrrrrB�s
�rBc@s0eZdZdZeedd�dd�Zed�dd�ZdS)	�InstallationSubprocessErrorz-A subprocess call failed during installation.N)�
returncode�descriptionrcCs||_||_dSr()rIrJ)rrIrJrrrr �sz$InstallationSubprocessError.__init__r!cCsd�|j|j�S)NzSCommand errored out with exit status {}: {} Check the logs for full command output.)r"rIrJr#rrrr$�s��z#InstallationSubprocessError.__str__)rrrr�intr%r r$rrrrrH�srHc@sJeZdZdZdd�dd�Zddd�dd	�Zed�d
d�Zed�dd
�Z	dS)�
HashErrorsz:Multiple HashError instances rolled into one for reportingNr!cCs
g|_dSr()�errorsr#rrrr �szHashErrors.__init__�	HashError)�errorrcCs|j�|�dSr()rM�append)rrOrrrrP�szHashErrors.appendcCsbg}|jjdd�d�t|jdd��D](\}}|�|j�|�dd�|D��q&|r^d�|�SdS)	NcSs|jSr()�order��errr�<lambda>�r,z$HashErrors.__str__.<locals>.<lambda>)�keycSs|jSr(r=rRrrrrT�r,css|]}|��VqdSr()�body�r)rSrrrr+�r,z%HashErrors.__str__.<locals>.<genexpr>�
�)rM�sortrrP�head�extendr.)r�lines�cls�
errors_of_clsrrrr$�s
zHashErrors.__str__cCs
t|j�Sr()�boolrMr#rrr�__bool__�szHashErrors.__bool__)
rrrrr rPr%r$r`rarrrrrL�s

rLc@s\eZdZUdZdZeded<dZdZe	ed<e
d�d	d
�Ze
d�dd�Ze
d�d
d�Z
dS)rNa�
    A failure to verify a package against known-good hashes

    :cvar order: An int sorting hash exception classes by difficulty of
        recovery (lower being harder), so the user doesn't bother fretting
        about unpinned packages when he has deeper issues, like VCS
        dependencies, to deal with. Also keeps error reports in a
        deterministic order.
    :cvar head: A section heading for display above potentially many
        exceptions of this kind
    :ivar req: The InstallRequirement that triggered this error. This is
        pasted on after the exception is instantiated, because it's not
        typically available earlier.

    Nr�reqrYr-rQr!cCsd|����S)a=Return a summary of me for display under the heading.

        This default implementation simply prints a description of the
        triggering requirement.

        :param req: The InstallRequirement that provoked this error, with
            its link already populated by the resolver's _populate_link().

        z    )�_requirement_namer#rrrrV�s
zHashError.bodycCs|j�d|����S)NrX)r[rVr#rrrr$�szHashError.__str__cCs|jrt|j�SdS)z�Return a description of the requirement that triggered me.

        This default implementation returns long description of the req, with
        line numbers

        �unknown package)rbr%r#rrrrc�szHashError._requirement_name)rrrrrbr�__annotations__r[rQrKr%rVr$rcrrrrrN�s
rNc@seZdZdZdZdZdS)�VcsHashUnsupported�uA hash was provided for a version-control-system-based requirement, but
    we don't have a method for hashing those.rzlCan't verify hashes for these requirements because we don't have a way to hash version control repositories:N�rrrrrQr[rrrrrf�s�rfc@seZdZdZdZdZdS)�DirectoryUrlHashUnsupportedrg�zUCan't verify hashes for these file:// requirements because they point to directories:Nrhrrrrri�s�ric@s6eZdZdZdZdZedd�dd�Zed�d	d
�ZdS)�HashMissingz2A hash was needed for a requirement but is absent.�awHashes are required in --require-hashes mode, but they are missing from some requirements. Here is a list of those requirements along with the hashes their downloaded archives actually had. Add lines like these to your requirements files to prevent tampering. (If you did not enable --require-hashes manually, note that it turns on automatically when any package has a hash.)N)�gotten_hashrcCs
||_dS)zq
        :param gotten_hash: The hash of the (possibly malicious) archive we
            just downloaded
        N)rm)rrmrrrr szHashMissing.__init__r!cCsHddlm}d}|jr4|jjr&|jjnt|jdd�}d�|p>d||j�S)Nr)�
FAVORITE_HASHrbz    {} --hash={}:{}rd)�pip._internal.utils.hashesrnrb�
original_link�getattrr"rm)rrn�packagerrrrVs�
��zHashMissing.body)	rrrrrQr[r%r rVrrrrrks�
rkc@seZdZdZdZdZdS)�HashUnpinnedzPA requirement had a hash specified but was not pinned to a specific
    version.�zaIn --require-hashes mode, all requirements must have their versions pinned with ==. These do not:Nrhrrrrrs2s�rsc@sZeZdZdZdZdZeeeefeedfdd�dd�Z	ed	�d
d�Z
ed	�dd
�ZdS)�HashMismatchz�
    Distribution file hash values don't match.

    :ivar package_name: The name of the package that triggered the hash
        mismatch. Feel free to write to this after the exception is raise to
        improve its error message.

    �z�THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS FILE. If you have updated the package versions, please update the hashes. Otherwise, examine the package contents carefully; someone may have tampered with them.r
N)�allowed�gotsrcCs||_||_dS)z�
        :param allowed: A dict of algorithm names pointing to lists of allowed
            hex digests
        :param gots: A dict of algorithm names pointing to hashes we
            actually got from the files under suspicion
        N)rwrx)rrwrxrrrr OszHashMismatch.__init__r!cCsd�|��|���S)Nz
    {}:
{})r"rc�_hash_comparisonr#rrrrVYszHashMismatch.bodycsltdd�dd�}g}|j��D]B\}}||��|��fdd�|D��|�d�|j|����qd�|�S)	aE
        Return a comparison of actual and expected hash values.

        Example::

               Expected sha256 abcdeabcdeabcdeabcdeabcdeabcdeabcdeabcdeabcde
                            or 123451234512345123451234512345123451234512345
                    Got        bcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdef

        z
chain[str])�	hash_namercSst|gtd��S)Nz    or)rr)rzrrr�hash_then_orhsz3HashMismatch._hash_comparison.<locals>.hash_then_orc3s|]}d�t��|�VqdS)z        Expected {} {}N)r"�nextrW��prefixrrr+psz0HashMismatch._hash_comparison.<locals>.<genexpr>z             Got        {}
rX)	r%rw�itemsr\rPr"rx�	hexdigestr.)rr{r]rz�	expectedsrr}rry\s��zHashMismatch._hash_comparison)rrrrrQr[rr%rr rVryrrrrru=s	�&
ruc@seZdZdZdS)�UnsupportedPythonVersionzMUnsupported python version according to Requires-Python package
    metadata.Nrrrrrr�ysr�csFeZdZdZd
eeeeejdd��fdd�
Zed�dd	�Z	�Z
S)�!ConfigurationFileCouldNotBeLoadedz8When there are errors while loading a configuration file�could not be loadedN)�reason�fnamerOrcs"t��|�||_||_||_dSr()r<r r�r�rO)rr�r�rOr=rrr �sz*ConfigurationFileCouldNotBeLoaded.__init__r!cCsF|jdurd|j�d�}n|jdus(J�d|j�d�}d|j�|��S)Nz in �.z.
rXzConfiguration file )r�rOr�)r�message_partrrrr$�s

z)ConfigurationFileCouldNotBeLoaded.__str__)r�NN)rrrrr%r�configparser�Errorr r$r?rrr=rr�~s��r�)3rr��	itertoolsrrr�typingrrrrr	�pip._vendor.pkg_resourcesr
Zpip._vendor.requests.modelsrr�hashlibr
�pip._internal.metadatarZpip._internal.req.req_installr�	Exceptionrrrrrr&r'r1r2r3r4r5r6r7r@rArBrHrLrNrfrirkrsrur�r�rrrr�<module>sH.*<site-packages/pip/_internal/__pycache__/pyproject.cpython-39.pyc000064400000007222150564745010020721 0ustar00a

��Re/�@s�ddlZddlmZddlmZmZmZddlmZddl	m
Z
mZddlm
Z
eed�dd	�Zeed
�dd�Zed
gd��Zeeeeeeed�dd�ZdS)�N)�
namedtuple)�Any�List�Optional)�tomli)�InvalidRequirement�Requirement)�InstallationError)�obj�returncCst|t�otdd�|D��S)Ncss|]}t|t�VqdS)N)�
isinstance�str)�.0�item�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/pyproject.py�	<genexpr>�z"_is_list_of_str.<locals>.<genexpr>)r�list�all)r
rrr�_is_list_of_strsr)�unpacked_source_directoryrcCstj�|d�S)Nzpyproject.toml)�os�path�join)rrrr�make_pyproject_pathsr�BuildSystemDetails)�requires�backend�check�backend_path)�
use_pep517�pyproject_toml�setup_py�req_namerc
Cs�tj�|�}tj�|�}|s.|s.t|�d���|rtt|dd��}t�|�}Wd�n1s^0Y|�d�}nd}|r�|s�|dur�|s�td��d}n<|r�d|vr�|dur�|s�td	�|d���d}n|dur�|}|dus�J�|s�dS|du�rd
dgdd
�}|du�sJ�d}	d|v�r0t|	j|dd���|d}
t	|
��sTt|	j|dd���|
D]@}zt
|�Wn,t�y�t|	j|d�|�d���Yn0�qX|�d�}|�dg�}
g}|du�r�d}d
dg}t|
|||
�S)aBLoad the pyproject.toml file.

    Parameters:
        use_pep517 - Has the user requested PEP 517 processing? None
                     means the user hasn't explicitly specified.
        pyproject_toml - Location of the project's pyproject.toml file
        setup_py - Location of the project's setup.py file
        req_name - The name of the requirement we're processing (for
                   error reporting)

    Returns:
        None if we should use the legacy code path, otherwise a tuple
        (
            requirements from pyproject.toml,
            name of PEP 517 backend,
            requirements we should check are installed after setting
                up the build environment
            directory paths to import the backend from (backend-path),
                relative to the project root.
        )
    zW does not appear to be a Python project: neither 'setup.py' nor 'pyproject.toml' found.zutf-8)�encodingNzbuild-systemzIDisabling PEP 517 processing is invalid: project does not have a setup.pyT�
build-backendzbDisabling PEP 517 processing is invalid: project specifies a build backend of {} in pyproject.tomlzsetuptools>=40.8.0�wheelz setuptools.build_meta:__legacy__)rr&zO{package} has a pyproject.toml file that does not comply with PEP 518: {reason}rz]it has a 'build-system' table but not 'build-system.requires' which is mandatory in the table)�package�reasonz1'build-system.requires' is not a list of strings.z='build-system.requires' contains an invalid requirement: {!r}zbackend-path)
rr�isfiler	�openr�load�get�formatrrrr)r!r"r#r$�
has_pyproject�	has_setup�f�pp_toml�build_system�error_templater�requirementrr rrrr�load_pyproject_tomls��(���

�	�
��
�����



r6)r�collectionsr�typingrrr�pip._vendorrZ"pip._vendor.packaging.requirementsrr�pip._internal.exceptionsr	�boolrr
rrr6rrrr�<module>s��site-packages/pip/_internal/__pycache__/configuration.cpython-39.pyc000064400000025634150564745010021560 0ustar00a

��Rea3�@s>dZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
mZmZddl
mZmZddlmZddlmZddlmZddlmZmZejZe
d	e�Zer�d
ndZdZed
ddddd�Zejej ej!ej"ej#fZ$ej ejej!fZ%ee&�Z'eed�dd�Z(ee	ed�dd�Z)eee	efd�dd�Z*Gdd�d�Z+dS)aConfiguration management setup

Some terminology:
- name
  As written in config files.
- value
  Value associated with a name
- key
  Name combined with it's section (section.name)
- variant
  A single word describing where the configuration key-value pair came from
�N)�Any�Dict�Iterable�List�NewType�Optional�Tuple)�ConfigurationError�!ConfigurationFileCouldNotBeLoaded)�appdirs)�WINDOWS)�	getLogger)�
ensure_dir�enum�Kindzpip.inizpip.conf)�version�help�user�global�site�envzenv-var)�USER�GLOBAL�SITE�ENV�ENV_VAR)�name�returncCs*|���dd�}|�d�r&|dd�}|S)zAMake a name consistent regardless of source (environment or file)�_�-z--�N)�lower�replace�
startswith)r�r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/configuration.py�_normalize_name2s
r&cCs&d|vrd�|�}t|��|�dd�S)N�.zbKey does not contain dot separated section and key. Perhaps you wanted to use 'global.{}' instead?�)�formatr	�split)r�
error_messager$r$r%�_disassemble_key:s��r,�rcCstdd�t�d�D�}tj�tjt�}tj�tj�d�t	r<dndt�}tj�t�
d�t�}tj|tj
|gtj||giS)NcSsg|]}tj�|t��qSr$)�os�path�join�CONFIG_BASENAME)�.0r/r$r$r%�
<listcomp>Esz+get_configuration_files.<locals>.<listcomp>�pip�~z.pip)r�site_config_dirsr.r/r0�sys�prefixr1�
expanduserr�user_config_dir�kindsrrr)�global_config_files�site_config_file�legacy_config_file�new_config_filer$r$r%�get_configuration_filesDs�

�
�r@cs�eZdZdZd6eeedd��fdd�
Zdd�dd�Zee	d�d	d
�Z
eee	e
fd�dd�Ze	e
d
�dd�Ze	e
dd�dd�Ze	dd
�dd�Zdd�dd�Zdd�dd�Zeee	e
fd�dd��Zdd�dd�Zee	ed�dd�Ze	ed �d!d"�Zdd�d#d$�Ze	eee	e
fee	e
fd%�d&d'�Zeee	e	fd�d(d)�Zeeeee	fd�d*d+�Zeee	e
fd,�d-d.�Z ee	efd�d/d0�Z!e	edd1�d2d3�Z"e	d�d4d5�Z#�Z$S)7�
Configurationa�Handles management of configuration.

    Provides an interface to accessing and managing configuration files.

    This class converts provides an API that takes "section.key-name" style
    keys and stores the value associated with it as "key-name" under the
    section "section".

    This allows for a clean interface wherein the both the section and the
    key-name are preserved in an easy to manage form in the configuration files
    and the data stored is also nice.
    N)�isolated�	load_onlyrcsjt���|dur4|tvr4td�d�ttt�����||_||_	dd�t
D�|_dd�t
D�|_g|_
dS)Nz5Got invalid value for load_only - should be one of {}z, cSsi|]
}|g�qSr$r$�r2�variantr$r$r%�
<dictcomp>rsz*Configuration.__init__.<locals>.<dictcomp>cSsi|]
}|i�qSr$r$rDr$r$r%rFus)�super�__init__�VALID_LOAD_ONLYr	r)r0�map�reprrBrC�OVERRIDE_ORDER�_parsers�_config�_modified_parsers)�selfrBrC��	__class__r$r%rHes 
����zConfiguration.__init__r-cCs|��|js|��dS)z<Loads configuration from configuration files and environmentN)�_load_config_filesrB�_load_environment_vars�rPr$r$r%�loadzszConfiguration.loadcCs:|jdusJd��z|��dWSty4YdS0dS)z7Returns the file with highest priority in configurationNz)Need to be specified a file to be editingr)rC�_get_parser_to_modify�
IndexErrorrUr$r$r%�get_file_to_edit�s
zConfiguration.get_file_to_editcCs
|j��S)z`Returns key-value pairs like dict.items() representing the loaded
        configuration
        )�_dictionary�itemsrUr$r$r%r[�szConfiguration.items)�keyrcCs2z|j|WSty,td|����Yn0dS)z#Get a value from the configuration.�No such key - N)rZ�KeyErrorr	)rPr\r$r$r%�	get_value�szConfiguration.get_value)r\�valuercCst|��|jsJ�|��\}}|durTt|�\}}|�|�sF|�|�|�|||�||j|j|<|�||�dS)z$Modify a value in the configuration.N)	�_ensure_have_load_onlyrCrWr,�has_section�add_section�setrN�_mark_as_modified)rPr\r`�fname�parser�sectionrr$r$r%�	set_value�s


zConfiguration.set_valuecCs�|��|jsJ�||j|jvr0td|����|��\}}|dur�t|�\}}|�|�rf|�||�sntd��|�|�s�|�	|�|�
||�|j|j|=dS)z#Unset a value in the configuration.r]Nz4Fatal Internal error [id=1]. Please report as a bug.)rarCrNr	rWr,rb�
remove_optionr[�remove_sectionre)rPr\rfrgrhrr$r$r%�unset_value�s$
�
��

zConfiguration.unset_valuec	Csl|��|jD]X\}}t�d|�ttj�|��t|d��}|�	|�Wd�q1s\0YqdS)z!Save the current in-memory state.z
Writing to %s�wN)
rarO�logger�inforr.r/�dirname�open�write)rPrfrg�fr$r$r%�save�szConfiguration.savecCs$|jdurtd��t�d|j�dS)Nz'Needed a specific file to be modifying.z$Will be working with %s variant only)rCr	rn�debugrUr$r$r%ra�s
z$Configuration._ensure_have_load_onlycCs"i}tD]}|�|j|�q|S)z3A dictionary representing the loaded configuration.)rL�updaterN)rP�retvalrEr$r$r%rZ�szConfiguration._dictionarycCs�t|���}|tjdd�tjgkr4t�d�dS|��D]V\}}|D]H}|j	durp||j	krpt�d||�qH|�
||�}|j|�||f�qHq<dS)z,Loads configuration from configuration filesrr(zZSkipping loading configuration files due to environment's PIP_CONFIG_FILE being os.devnullNz Skipping file '%s' (variant: %s))
�dict�iter_config_filesr;rr.�devnullrnrur[rC�
_load_filerM�append)rP�config_filesrE�filesrfrgr$r$r%rS�s�z Configuration._load_config_files)rErfrcCsLt�d||�|�|�}|��D]&}|�|�}|j|�|�||��q |S)Nz'For variant '%s', will try loading '%s')rn�verbose�_construct_parser�sectionsr[rNrv�_normalized_keys)rPrErfrgrhr[r$r$r%r{�s

zConfiguration._load_file)rfrc
Cs�t��}tj�|�r|z|�|�WnXtyLtd�t	�
d��|d��Yn0tjyz}zt|d��WYd}~n
d}~00|S)Nzcontains invalid {} charactersF)�reasonrf)�error)�configparser�RawConfigParserr.r/�exists�read�UnicodeDecodeErrorr
r)�locale�getpreferredencoding�Error)rPrfrgr�r$r$r%r�s��
 zConfiguration._construct_parsercCs"|jtj�|�d|����dS)z.Loads configuration from environment variablesz:env:N)rNr;rrvr��get_environ_varsrUr$r$r%rTs�z$Configuration._load_environment_vars)rhr[rcCs.i}|D] \}}|dt|�}|||<q|S)z�Normalizes items to construct a dictionary with normalized keys.

        This routine is where the names become keys and are made the same
        regardless of source - configuration files or environment.
        r')r&)rPrhr[�
normalizedr�valr\r$r$r%r�"s

zConfiguration._normalized_keysccsDtj��D]4\}}|�d�r
|dd���}|tvr
||fVq
dS)z@Returns a generator with all environmental vars with prefix PIP_�PIP_�N)r.�environr[r#r!�ENV_NAMES_IGNORED)rPr\r�rr$r$r%r�0s

zConfiguration.get_environ_varsccs�tj�dd�}|dur&tj|gfVntjgfVt�}tj|tjfV|job|o`tj�	|�}|rztj
|tj
fVtj|tjfVdS)zYields variant and configuration files associated with it.

        This should be treated like items of a dictionary.
        �PIP_CONFIG_FILEN)r.r��getr;rr@rrBr/r�rr)rP�config_filer}�should_load_user_configr$r$r%ry9s�zConfiguration.iter_config_files)rErcCs
|j|S)z#Get values present in a config file)rN)rPrEr$r$r%�get_values_in_configWsz"Configuration.get_values_in_configcCs*|js
J�|j|j}|s"td��|dS)Nz4Fatal Internal error [id=2]. Please report as a bug.���)rCrMr	)rP�parsersr$r$r%rW[s
�z#Configuration._get_parser_to_modify)rfrgrcCs"||f}||jvr|j�|�dS)N)rOr|)rPrfrg�file_parser_tupler$r$r%reis
zConfiguration._mark_as_modifiedcCs|jj�d|j�d�S)N�(�))rR�__name__rZrUr$r$r%�__repr__nszConfiguration.__repr__)N)%r��
__module__�__qualname__�__doc__�boolrrrHrV�strrYrrrr[r_rirlrtra�propertyrrZrSr�r{r�rTr�r�rryr�rWrer��
__classcell__r$r$rQr%rAWs2
	

�	rA),r�r�r�r.r7�typingrrrrrrr�pip._internal.exceptionsr	r
Zpip._internal.utilsr�pip._internal.utils.compatr�pip._internal.utils.loggingr
�pip._internal.utils.miscrrr�r�rr1r�r;rrrrrrLrIr�rnr&r,r@rAr$r$r$r%�<module>s8
$
�
site-packages/pip/_internal/__pycache__/build_env.cpython-39.pyc000064400000022443150564745010020653 0ustar00a

��Re�&�@sddZddlZddlZddlZddlZddlZddlZddlZddlm	Z	ddl
mZddlm
Z
ddlmZmZmZmZmZmZmZmZddlmZddlmZdd	lmZdd
lmZddl m!Z!ddl"m#Z#m$Z$m%Z%dd
l&m'Z'ddl(m)Z)ddl*m+Z+m,Z,e�rddl-m.Z.e�/e0�Z1Gdd�d�Z2ej3ee4d�dd��Z5Gdd�d�Z6Gdd�de6�Z7dS)z;Build Environment used for isolation during sdist building
�N)�OrderedDict)�	get_paths)�
TracebackType)�
TYPE_CHECKING�Iterable�Iterator�List�Optional�Set�Tuple�Type)�where)�Requirement)�Version)�__file__)�open_spinner)�get_platlib�get_prefixed_libs�get_purelib)�get_environment)�call_subprocess)�
TempDirectory�
tempdir_kinds)�
PackageFinderc@seZdZedd�dd�ZdS)�_PrefixN)�path�returncCs@||_d|_ttjdkrdnd||d�d�d|_t|�|_dS)NF�nt�posix_prefix)�base�platbase)�vars�scripts)r�setupr�os�name�bin_dirr�lib_dirs)�selfr�r)��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/build_env.py�__init__"s��z_Prefix.__init__)�__name__�
__module__�__qualname__�strr+r)r)r)r*r!sr�rc	cs�t�t���j}|��s&t|�VdStdd���}tj	�
|j	d�}i}tjdkrXd|d<t
j|dfi|���8}|�d	�D]}|�||�|j����qxWd�n1s�0Ytj	�
|d
�VWd�n1s�0YdS)z�Create a "standalone pip" zip file.

    The zip file's content is identical to the currently-running pip.
    It will be used to install requirements into the build environment.
    Nzstandalone-pip)�kindz__env_pip__.zip)��F�strict_timestamps�w�*�pip)�pathlib�Path�pip_location�resolve�parent�is_dirr/rr$r�join�sys�version_info�zipfile�ZipFile�rglob�write�relative_to�as_posix)�source�tmp_dir�pip_zip�kwargs�zf�childr)r)r*�_create_standalone_pip,s

8rMc@s�eZdZdZdd�dd�Zdd�dd�Zeeeeeee	dd�d	d
�Z
eee
ee
eefeefd�dd
�Zdeeeedd�dd�Zeedeeeedd�dd��ZdS)�BuildEnvironmentzACreates and manages an isolated environment to install build depsNr0cs�ttjdd��t�fdd�dD��|_g|_g|_tt|j�	���D] }|j�
|j�|j�|j
�qDdd�t�t�fD�}tj��jd�|_tj�|j�s�t�|j�ttj�|jd	�d
��,}|�t�d�j||jd��Wd�n1s�0YdS)
NT)r1�globally_managedc3s&|]}|ttj��j|��fVqdS�N)rr$rr>��.0r%��temp_dirr)r*�	<genexpr>Ls�z,BuildEnvironment.__init__.<locals>.<genexpr>)�normal�overlaycSsh|]}tj�|��qSr))r$r�normcase)rR�siter)r)r*�	<setcomp>Zsz,BuildEnvironment.__init__.<locals>.<setcomp>rYzsitecustomize.pyr5a�
                import os, site, sys

                # First, drop system-sites related paths.
                original_sys_path = sys.path[:]
                known_paths = set()
                for path in {system_sites!r}:
                    site.addsitedir(path, known_paths=known_paths)
                system_paths = set(
                    os.path.normcase(path)
                    for path in sys.path[len(original_sys_path):]
                )
                original_sys_path = [
                    path for path in original_sys_path
                    if os.path.normcase(path) not in system_paths
                ]
                sys.path = original_sys_path

                # Second, add lib directories.
                # ensuring .pth file are processed.
                for path in {lib_dirs!r}:
                    assert not path in sys.path
                    site.addsitedir(path)
                )�system_sitesr')rr�	BUILD_ENVr�	_prefixes�	_bin_dirs�	_lib_dirs�reversed�list�values�appendr&�extendr'rrr$rr>�	_site_dir�exists�mkdir�openrD�textwrap�dedent�format)r(�prefixr[�fpr)rSr*r+Is.�

����zBuildEnvironment.__init__cCsndd�dD�|_|jdd�}|jd}|r>|�|�tj��|jg}tj�tj�	|�dtj�	|�d��dS)NcSsi|]}|tj�|d��qSrP)r$�environ�getrQr)r)r*�
<dictcomp>s�z.BuildEnvironment.__enter__.<locals>.<dictcomp>)�PATH�PYTHONNOUSERSITE�
PYTHONPATHrq�1)
�	_save_envr^rd�splitr$�pathseprern�updater>)r(r�old_path�
pythonpathr)r)r*�	__enter__~s�


��zBuildEnvironment.__enter__��exc_type�exc_val�exc_tbrcCs:|j��D]*\}}|dur*tj�|d�q
|tj|<q
dSrP)ru�itemsr$rn�pop)r(r}r~r�varname�	old_valuer)r)r*�__exit__�szBuildEnvironment.__exit__)�reqsrc	Cs�t�}t�}|r�t|j�}|D]t}t|�}|�|j�}|sF|�|�qt|jt	�rf|j�d|j��}n|j�d|j��}|j|j
vr|�||f�q||fS)z�Return 2 sets:
        - conflicting requirements: set of (installed, wanted) reqs tuples
        - missing requirements: set of reqs
        z==z===)�setrr_r�get_distributionr%�add�
isinstance�versionr�	specifier)	r(r��missing�conflicting�env�req_str�req�dist�installed_req_strr)r)r*�check_requirements�s 

z#BuildEnvironment.check_requirementsr��finder�requirements�prefix_as_string�messagercCs�|j|}|jrJ�d|_|s"dSt���F}tjdkrDtj�t	�}n|�
t��}|�|||||�Wd�n1sv0YdS)NT)r2�)
r]r#�
contextlib�	ExitStackr?r@r$r�dirnamer:�
enter_contextrM�_install_requirements)r(r�r�r�r�rl�ctx�pip_runnabler)r)r*�install_requirements�s 



�z%BuildEnvironment.install_requirements)r�r�r�rlr�rcCshtj|dddd|jdg}t��tjkr0|�d�dD]:}t|j	|�}|�
d|�d	d
�d�t
|pddh��f�q4|j}|r�|�
d
|dg�|dd�D]}	|�
d|	g�q�n
|�d�|jD]}
|�
d|
g�q�|jD]}|�
d|g�q�|jr�|�d�|j�r|�d�|�d�|�
|�dt�i}t|��}
t||
|d�Wd�n1�sZ0YdS)N�installz--ignore-installedz	--no-userz--prefixz--no-warn-script-locationz-v)�	no_binary�only_binaryz--�_�-�,z:none:z-ir�z--extra-index-urlz
--no-indexz--find-linksz--trusted-hostz--prez--prefer-binary�_PIP_STANDALONE_CERT)�spinner�
extra_environ)r?�
executabler�logger�getEffectiveLevel�logging�DEBUGrc�getattr�format_controlrd�replacer>�sorted�
index_urls�
find_links�
trusted_hosts�allow_all_prereleases�
prefer_binaryr
rr)r�r�r�rlr��argsr��formatsr��extra_index�link�hostr�r�r)r)r*r��sJ	�

��








z&BuildEnvironment._install_requirements)r,r-r.�__doc__r+r{r	r�
BaseExceptionrr�rr/rr
r�r��staticmethodrr�r)r)r)r*rNFs25
�
���rNc@sxeZdZdZdd�dd�Zdd�dd�Zeeeeeee	dd�d	d
�Z
dd�dd�Zd
ee
e
e
dd�dd�ZdS)�NoOpBuildEnvironmentz0A no-op drop-in replacement for BuildEnvironmentNr0cCsdSrPr)�r(r)r)r*r+
szNoOpBuildEnvironment.__init__cCsdSrPr)r�r)r)r*r{szNoOpBuildEnvironment.__enter__r|cCsdSrPr))r(r}r~rr)r)r*r�szNoOpBuildEnvironment.__exit__cCsdSrPr)r�r)r)r*�cleanupszNoOpBuildEnvironment.cleanuprr�cCs
t��dSrP)�NotImplementedError)r(r�r�r�r�r)r)r*r�sz)NoOpBuildEnvironment.install_requirements)r,r-r.r�r+r{r	rr�rr�r�rr/r�r)r)r)r*r�
s
��r�)8r�r�r�r$r8r?rirA�collectionsr�	sysconfigr�typesr�typingrrrrr	r
rr�pip._vendor.certifir
Z"pip._vendor.packaging.requirementsrZpip._vendor.packaging.versionrr7rr:�pip._internal.cli.spinnersr�pip._internal.locationsrrr�pip._internal.metadatar�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirrr�"pip._internal.index.package_finderr�	getLoggerr,r�r�contextmanagerr/rMrNr�r)r)r)r*�<module>s:(
Esite-packages/pip/_internal/__pycache__/wheel_builder.cpython-39.pyc000064400000021705150564745010021516 0ustar00a

��Re�/�@s�dZddlZddlZddlZddlZddlmZmZm	Z	m
Z
mZmZddl
mZmZddlmZmZddlmZddlmZmZddlmZmZdd	lmZdd
lmZddlm Z ddl!m"Z"dd
l#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*m+Z+m,Z,ddl-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6e�7e8�Z9e�:dej;�Z<ee&ge=fZ>ee
e&e
e&fZ?e@e=d�dd�ZAe&e=e>e=d�dd�ZBe&e=d�dd�ZCe&e>e=d �d!d"�ZDe&ee=d�d#d$�ZEe&ee@d%�d&d'�ZFee=d(�d)d*�ZGe&e@dd+�d,d-�ZHe&e@e=e
e@e
e@e=ee@d.�d/d0�ZIe&e@e
e@e
e@e=ee@d1�d2d3�ZJe&e
e@e=d4�d5d6�ZKe	e&ee=e
e@e
e@e?d7�d8d9�ZLdS):z;Orchestrator for building wheels from InstallRequirements.
�N)�Any�Callable�Iterable�List�Optional�Tuple)�canonicalize_name�canonicalize_version)�InvalidVersion�Version)�
WheelCache)�InvalidWheelFilename�UnsupportedWheel)�FilesystemWheel�get_wheel_distribution)�Link)�Wheel)�build_wheel_pep517)�build_wheel_editable)�build_wheel_legacy)�InstallRequirement)�
indent_log)�
ensure_dir�	hash_file�is_wheel_installed)�make_setuptools_clean_args)�call_subprocess)�
TempDirectory)�path_to_url)�vcsz([a-z0-9_.]+)-([a-z0-9_.!+-]+))�s�returncCstt�|��S)zjDetermine whether the string looks like an egg_info.

    :param s: The string to parse. E.g. foo-2.1
    )�bool�_egg_info_re�search)r �r%��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/wheel_builder.py�_contains_egg_info&sr')�req�
need_wheel�check_binary_allowedr!cCs�|jr
dS|jr&|r"t�d|j�dS|r.dS|js8dS|jrF|��S|jrPdS||�sjt�d|j�dSt	�s�t�d|j�dSdS)zBReturn whether an InstallRequirement should be built into a wheel.Fz(Skipping %s, due to already being wheel.TzCSkipping wheel build for %s, due to binaries being disabled for it.zOUsing legacy 'setup.py install' for %s, since package 'wheel' is not installed.)
�
constraint�is_wheel�logger�info�name�
source_dir�editable�supports_pyproject_editable�
use_pep517r)r(r)r*r%r%r&�
_should_build.s<���r4)r(r!cCst|dtd�S)NT�r)r*)r4�_always_true)r(r%r%r&�should_build_for_wheel_commandcsr7)r(r*r!cCst|d|d�S)NFr5)r4)r(r*r%r%r&� should_build_for_install_commandis�r8cCs�|js|jsdS|jrb|jjrb|jr(J�|js2J�t�|jj�}|sHJ�|�|jj|j�r^dSdS|jslJ�|j�	�\}}t
|�r�dSdS)z�
    Return whether a built InstallRequirement can be stored in the persistent
    wheel cache, assuming the wheel cache is available, and _should_build()
    has determined a wheel needs to be built.
    FT)r1r0�link�is_vcsr�get_backend_for_scheme�scheme�is_immutable_rev_checkout�url�splitextr')r(�vcs_backend�base�extr%r%r&�
_should_cachers


rC)r(�wheel_cacher!cCs>t|j�}|jsJ�|r.t|�r.|�|j�}n|�|j�}|S)zdReturn the persistent or temporary cache directory where the built
    wheel need to be stored.
    )r"�	cache_dirr9rC�get_path_for_link�get_ephem_path_for_link)r(rD�cache_availablerEr%r%r&�_get_cache_dir�s

rI)�_r!cCsdS)NTr%)rJr%r%r&r6�sr6)r(�
wheel_pathr!c	Cs�t|jp
d�}ttj�|��}t|j�|kr>td�||j���tt	|�|�}t
|j�}t|�t|j�krztd�||j���|j
}|dur�td��zt|�}Wn$ty�d|��}t|��Yn0|td�kr�t|jt�s�td�|���dS)N�z7Wheel has unexpected file name: expected {!r}, got {!r}zMissing Metadata-VersionzInvalid Metadata-Version: z1.2z6Metadata 1.2 mandates PEP 440 version, but {!r} is not)rr/r�os�path�basenamer
�formatrr�str�versionr	�metadata_versionrrr
�
isinstance)	r(rK�canonical_name�w�dist�dist_verstr�metadata_version_valuerS�msgr%r%r&�_verify_one�s<��
��
��r[)r(�
output_dir�verify�
build_options�global_optionsr1r!c	
Cs�|rdnd}zt|�Wn8tyP}z t�d||j|�WYd}~dSd}~00|j� t|||||�}Wd�n1s~0Y|r�|r�zt||�Wn<tt	fy�}z t�d||j|�WYd}~dSd}~00|S)zaBuild one wheel.

    :return: The filename of the built wheel, or None if the build failed.
    r1�wheelzBuilding %s for %s failed: %sNzBuilt %s for %s is invalid: %s)
r�OSErrorr-�warningr/�	build_env�_build_one_inside_envr[r
r)	r(r\r]r^r_r1�artifact�erKr%r%r&�
_build_one�s,�
�"rg)r(r\r^r_r1r!cCs�tdd���l}|jsJ�|jr�|js(J�|js2J�|rDt�d|j�|rVt�d|j�|rtt|j|j|j|jd�}q�t	|j|j|j|jd�}nt
|j|j|j|||jd�}|du�rVt
j�|�}t
j�||�}zPt|�\}	}
t�||�t�d|j||
|	���t�d|�|WWd�St�yT}zt�d	|j|�WYd}~n
d}~00|j�sht||�Wd�dS1�s�0YdS)
Nr`)�kindz7Ignoring --global-option when building %s using PEP 517z6Ignoring --build-option when building %s using PEP 517)r/�backend�metadata_directory�tempd)r/�
setup_py_pathr0r_r^rkz3Created wheel for %s: filename=%s size=%d sha256=%szStored in directory: %sz Building wheel for %s failed: %s)rr/r3rj�pep517_backendr-rbrrNrrrl�unpacked_source_directoryrMrO�joinr�shutil�mover.�	hexdigest�	Exception�_clean_one_legacy)r(r\r^r_r1�temp_dirrK�
wheel_name�	dest_path�
wheel_hash�lengthrfr%r%r&rd�st


�����	
��
rd)r(r_r!cCsXt|j|d�}t�d|j�zt||jd�WdStyRt�d|j�YdS0dS)N)r_zRunning setup.py clean for %s)�cwdTz Failed cleaning build dir for %sF)	rrlr-r.r/rr0rs�error)r(r_�
clean_argsr%r%r&rt1s�rt)�requirementsrDr]r^r_r!c
	Cs|sggfSt�dd�dd�|D���t���gg}}|D]p}|jsLJ�t||�}t||||||jol|j�}	|	r�t	t
|	��|_|jj|_
|jjs�J�|�|�q>|�|�q>Wd�n1s�0Y|r�t�dd�dd	�|D���|�rt�d
d�dd	�|D���||fS)z�Build wheels.

    :return: The list of InstallRequirement that succeeded to build and
        the list of InstallRequirement that failed to build.
    z*Building wheels for collected packages: %sz, css|]}|jVqdS)N�r/��.0r(r%r%r&�	<genexpr>R�zbuild.<locals>.<genexpr>NzSuccessfully built %s� cSsg|]
}|j�qSr%r~rr%r%r&�
<listcomp>or�zbuild.<locals>.<listcomp>zFailed to build %scSsg|]
}|j�qSr%r~rr%r%r&r�tr�)r-r.rorr/rIrgr1�permit_editable_wheelsrrr9�	file_path�local_file_pathr,�append)
r}rDr]r^r_�build_successes�build_failuresr(rE�
wheel_filer%r%r&�build@sH�



�
*��r�)M�__doc__�logging�os.pathrM�rerp�typingrrrrrr�pip._vendor.packaging.utilsrr	Zpip._vendor.packaging.versionr
r�pip._internal.cacher�pip._internal.exceptionsr
r�pip._internal.metadatarr�pip._internal.models.linkr�pip._internal.models.wheelr�$pip._internal.operations.build.wheelr�-pip._internal.operations.build.wheel_editabler�+pip._internal.operations.build.wheel_legacyrZpip._internal.req.req_installr�pip._internal.utils.loggingr�pip._internal.utils.miscrrr�$pip._internal.utils.setuptools_buildr�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirr�pip._internal.utils.urlsr�pip._internal.vcsr�	getLogger�__name__r-�compile�
IGNORECASEr#r"�BinaryAllowedPredicate�BuildResultrQr'r4r7r8rCrIr6r[rgrdrtr�r%r%r%r&�<module>s� 
	�6��
�!��'�G�site-packages/pip/_internal/__pycache__/__init__.cpython-39.pyc000064400000001447150564745010020444 0ustar00a

��ReK�@sJddlmZmZddlZddlmZe��deeee	d�dd�Z
dS)�)�List�OptionalN)�_log)�args�returncCsddlm}||dd�S)z�This is preserved for old console scripts that may still be referencing
    it.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    r)�_wrapperT)�_nowarn)�pip._internal.utils.entrypointsr)rr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/__init__.py�mainsr)N)�typingrr�*pip._internal.utils.inject_securetransport�pipZpip._internal.utilsr�init_logging�str�intrr
r
r
r�<module>ssite-packages/pip/_internal/__pycache__/self_outdated_check.cpython-39.pyc000064400000010677150564745010022671 0ustar00a

��Re��@sddlZddlZddlZddlZddlZddlZddlZddlm	Z	m
Z
ddlmZ
ddlmZddlmZddlmZddlmZddlmZdd	lmZmZmZdd
lmZdZe�e �Z!e"e"d�d
d�Z#Gdd�d�Z$e"e%d�dd�Z&eej'dd�dd�Z(dS)�N)�Any�Dict)�parse)�
LinkCollector)�
PackageFinder)�get_default_environment)�SelectionPreferences)�
PipSession)�adjacent_tmp_file�check_path_owner�replace)�
ensure_dirz%Y-%m-%dT%H:%M:%SZ)�key�returncCs|��}t�|���}|S�N)�encode�hashlib�sha224�	hexdigest)r�	key_bytes�name�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/self_outdated_check.py�_get_statefile_namesrc@sBeZdZedd�dd�Zeed�dd��Zeejdd�d	d
�ZdS)�SelfCheckStateN)�	cache_dirrc
Cs�i|_d|_|r�tj�|dt|j��|_z>t|jdd��}t�	|�|_Wd�n1sZ0YWnt
ttfy~Yn0dS)N�	selfcheckzutf-8)�encoding)
�state�statefile_path�os�path�joinrr�open�json�load�OSError�
ValueError�KeyError)�selfr�	statefilerrr�__init__!s�.zSelfCheckState.__init__)rcCstjSr)�sys�prefix)r)rrrr2szSelfCheckState.key)�pypi_version�current_timercCs�|js
dSttj�|j��s dSttj�|j��|j|�t�|d�}t	j
|ddd�}t|j��}|�|�
��Wd�n1s�0Yzt|j|j�Wnty�Yn0dS)N)r�
last_checkr.T)�,�:)�	sort_keys�
separators)rrr r!�dirnamer
r�strftime�SELFCHECK_DATE_FMTr$�dumpsr
�writerrrr&)r)r.r/r�text�frrr�save6s �,zSelfCheckState.save)	�__name__�
__module__�__qualname__�strr+�propertyr�datetimer<rrrrr sr)�pkgrcCst��|�}|duod|jkS)z�Checks whether pkg was installed by pip

    This is used not to display the upgrade message when pip is in fact
    installed by system package manager, such as dnf on Fedora.
    N�pip)r�get_distribution�	installer)rC�distrrr�was_installed_by_pipYsrH)�session�optionsrcCsRt��d�}|sdS|j}d}�z
t|jd�}tj��}d|jvrzd|jvrztj�|jdt	�}||�
�dkrz|jd}|dur�tj||dd�}t
d	d	d
�}	tj||	d�}
|
�d�j}|dur�WdSt|j�}|�||�t|�}||ko�|j|jko�td�}
|
�sWdStj�d�}t�d
|||�Wn"t�yLtjddd�Yn0dS)z�Check for an update for pip.

    Limit the frequency of checks to once per week. State is stored either in
    the active virtualenv or in the user's USER_CACHE_DIR keyed off the prefix
    of the pip script path.
    rDN)rr0r.i�:	T)rJ�suppress_no_indexF)�allow_yanked�allow_all_prereleases)�link_collector�selection_prefsz -m pipz�You are using pip version %s; however, version %s is available.
You should consider upgrading via the '%s install --upgrade pip' command.z5There was an error checking the latest version of pip)�exc_info)rrE�versionrrrB�utcnowr�strptimer7�
total_secondsr�createrr�find_best_candidate�best_candidater@r<�
parse_version�base_versionrHr,�
executable�logger�warning�	Exception�debug)rIrJ�installed_dist�pip_versionr.rr/r0rNrO�finderrW�remote_version�local_version_is_older�pip_cmdrrr�pip_self_version_checkcsj

�
���

����re))rBrr$�logging�optparse�os.pathr r,�typingrrZpip._vendor.packaging.versionrrX�pip._internal.index.collectorr�"pip._internal.index.package_finderr�pip._internal.metadatar�$pip._internal.models.selection_prefsr�pip._internal.network.sessionr	�pip._internal.utils.filesystemr
rr�pip._internal.utils.miscr
r7�	getLoggerr=r[r@rr�boolrH�Valuesrerrrr�<module>s*
9
site-packages/pip/_internal/__pycache__/cache.cpython-39.pyc000064400000020234150564745010017743 0ustar00a

��Re�$�@sdZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
ddlmZm
Z
mZddlmZddlmZddlmZddlmZdd	lmZdd
lmZmZddlmZe�e�Z ee!e!fe!d�d
d�Z"Gdd�d�Z#Gdd�de#�Z$Gdd�de$�Z%Gdd�d�Z&Gdd�de#�Z'dS)zCache Management
�N)�Any�Dict�List�Optional�Set)�Tag�interpreter_name�interpreter_version)�canonicalize_name)�InvalidWheelFilename)�
FormatControl)�Link)�Wheel)�
TempDirectory�
tempdir_kinds)�path_to_url)�d�returncCs&tj|dddd�}t�|�d����S)z'Return a stable sha224 of a dictionary.T)�,�:)�	sort_keys�
separators�ensure_ascii�ascii)�json�dumps�hashlib�sha224�encode�	hexdigest)r�s�r!��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cache.py�
_hash_dictsr#cs�eZdZdZeeeedd��fdd�Zee	ed�dd�Z
eee	ed	�d
d�Zeed�dd
�Z
eeee	eed�dd�Z�ZS)�CacheanAn abstract class - provides cache directories for data from links


    :param cache_dir: The root of the cache.
    :param format_control: An object of FormatControl class to limit
        binaries being read from the cache.
    :param allowed_formats: which formats of files the cache should store.
        ('binary' and 'source' are the only allowed values)
    N)�	cache_dir�format_control�allowed_formatsrcsTt���|rtj�|�sJ�|p$d|_||_||_ddh}|j�|�|ksPJ�dS)N�source�binary)	�super�__init__�os�path�isabsr%r&r'�union)�selfr%r&r'�_valid_formats��	__class__r!r"r+(s

zCache.__init__��linkrcCs�d|ji}|jdur*|jdur*|j||j<|jr:|j|d<t�|d<t�|d<t|�}|dd�|dd�|dd�|dd�g}|S)	z<Get parts of part that must be os.path.joined with cache_dir�urlN�subdirectoryrr	���)�url_without_fragment�	hash_name�hash�subdirectory_fragmentrr	r#)r0r5�	key_parts�hashed�partsr!r!r"�_get_cache_path_parts4s



,zCache._get_cache_path_parts)r5�canonical_package_namercCst|jp|p|}|rgS|j�|�}|j�|�s8gSg}|�|�}tj�|�rpt�	|�D]}|�
||f�q\|S�N)r%r&�get_allowed_formatsr'�intersection�get_path_for_linkr,r-�isdir�listdir�append)r0r5rC�
can_not_cache�formats�
candidatesr-�	candidater!r!r"�_get_candidatesUs
zCache._get_candidatescCs
t��dS)z5Return a directory to store cached items in for link.N��NotImplementedError�r0r5r!r!r"rGeszCache.get_path_for_link�r5�package_name�supported_tagsrcCs
t��dS)zaReturns a link to a cached item if it exists, otherwise returns the
        passed link.
        NrP)r0r5rTrUr!r!r"�getis	z	Cache.get)�__name__�
__module__�__qualname__�__doc__�strrrr+r
rrBrrOrGrrrV�
__classcell__r!r!r2r"r$s
�!�r$csVeZdZdZeedd��fdd�Zeed�dd�Zee	ee
eed	�d
d�Z�Z
S)�SimpleWheelCachez&A cache of wheels for future installs.N�r%r&rcst��||dh�dS�Nr))r*r+�r0r%r&r2r!r"r+xszSimpleWheelCache.__init__r4cCs,|�|�}|jsJ�tjj|jdg|�R�S)a�Return a directory to store cached wheels for link

        Because there are M wheels for any one sdist, we provide a directory
        to cache them in, and then consult that directory when looking up
        cache hits.

        We only insert things into the cache if they have plausible version
        numbers, so that we don't contaminate the cache with things that were
        not unique. E.g. ./package might have dozens of installs done for it
        and build a version of 0.0...and if we built and cached a wheel, we'd
        end up using the same wheel even if the source has been edited.

        :param link: The link of the sdist for which this will cache wheels.
        �wheels)rBr%r,r-�join)r0r5rAr!r!r"rG{s

z"SimpleWheelCache.get_path_for_linkrSc
	Cs�g}|s|St|�}|�||�D]n\}}zt|�}WntyJYq Yn0t|j�|krlt�d|||�q |�|�sxq |�|�	|�||f�q |s�|St
|�\}	}}ttt
j�||���S)NzWIgnoring cached wheel %s for %s as it does not match the expected distribution name %s.)r
rOrr�name�logger�debug�	supportedrJ�support_index_min�minr
rr,r-rb)
r0r5rTrUrMrC�
wheel_name�	wheel_dir�wheel�_r!r!r"rV�s:
�
��zSimpleWheelCache.get)rWrXrYrZr[rr+r
rGrrrrVr\r!r!r2r"r]us�r]cs(eZdZdZedd��fdd�Z�ZS)�EphemWheelCachezBA SimpleWheelCache that creates it's own temporary cache directoryN)r&rcs&ttjdd�|_t��|jj|�dS)NT)�kind�globally_managed)rr�EPHEM_WHEEL_CACHE�	_temp_dirr*r+r-)r0r&r2r!r"r+�s
�zEphemWheelCache.__init__)rWrXrYrZrr+r\r!r!r2r"rm�srmc@seZdZeed�dd�ZdS)�
CacheEntry�r5�
persistentcCs||_||_dSrDrs)r0r5rtr!r!r"r+�szCacheEntry.__init__N)rWrXrYr
�boolr+r!r!r!r"rr�s�rrcs�eZdZdZeedd��fdd�Zeed�dd�Zeed�d	d
�Z	ee
eeeed�dd
�Z
ee
eeee
ed�dd�Z�ZS)�
WheelCachez�Wraps EphemWheelCache and SimpleWheelCache into a single Cache

    This Cache allows for gracefully degradation, using the ephem wheel cache
    when a certain link is not found in the simple wheel cache first.
    Nr^cs,t��||dh�t||�|_t|�|_dSr_)r*r+r]�_wheel_cacherm�_ephem_cacher`r2r!r"r+�szWheelCache.__init__r4cCs|j�|�SrD)rwrGrRr!r!r"rG�szWheelCache.get_path_for_linkcCs|j�|�SrD)rxrGrRr!r!r"�get_ephem_path_for_link�sz"WheelCache.get_ephem_path_for_linkrScCs |�|||�}|dur|S|jSrD)�get_cache_entryr5)r0r5rTrU�cache_entryr!r!r"rV�szWheelCache.getcCsP|jj|||d�}||ur&t|dd�S|jj|||d�}||urLt|dd�SdS)z�Returns a CacheEntry with a link to a cached item if it exists or
        None. The cache entry indicates if the item was found in the persistent
        or ephemeral cache.
        )r5rTrUT)rtFN)rwrVrrrx)r0r5rTrU�retvalr!r!r"rz�s
��zWheelCache.get_cache_entry)rWrXrYrZr[rr+r
rGryrrrrVrrrzr\r!r!r2r"rv�s�
�rv)(rZrr�loggingr,�typingrrrrrZpip._vendor.packaging.tagsrrr	�pip._vendor.packaging.utilsr
�pip._internal.exceptionsr�#pip._internal.models.format_controlr�pip._internal.models.linkr
�pip._internal.models.wheelr�pip._internal.utils.temp_dirrr�pip._internal.utils.urlsr�	getLoggerrWrdr[r#r$r]rmrrrvr!r!r!r"�<module>s(
XF
site-packages/pip/_internal/operations/__pycache__/prepare.cpython-39.pyc000064400000034543150564745010022531 0ustar00a

��Re]�@sRdZddlZddlZddlZddlZddlmZmZmZm	Z	ddl
mZddlm
Z
ddlmZddlmZmZmZmZmZmZmZddlmZdd	lmZdd
lmZddlmZddl m!Z!m"Z"dd
l#m$Z$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+ddl,m-Z-ddl.m/Z/m0Z0ddl1m2Z2ddl3m4Z4m5Z5m6Z6m7Z7ddl8m9Z9ddl:m;Z;ddl<m=Z=e�>e?�Z@e)e+eeAed�dd�ZBeeCdd�dd�ZDGdd�d�ZEd3ee"e	eCe	e/eEd �d!d"�ZFeCeCdd#�d$d%�ZGeCeCdd&�d'd(�ZHd4ee	eCe	e/eEd)�d*d+�ZId5eeCe"e	eCe	e/e	eEd,�d-d.�ZJeeCe	e/e	eCd)�d/d0�ZKGd1d2�d2�ZLdS)6z)Prepares a distribution for installation
�N)�Dict�Iterable�List�Optional)�canonicalize_name)�)make_distribution_for_install_requirement)�InstalledDistribution)�DirectoryUrlHashUnsupported�HashMismatch�HashUnpinned�InstallationError�NetworkConnectionError�PreviousBuildDirError�VcsHashUnsupported)�
PackageFinder)�BaseDistribution)�Link)�Wheel)�BatchDownloader�
Downloader)�HTTPRangeRequestUnsupported�dist_from_wheel_url)�
PipSession)�InstallRequirement)�RequirementTracker)�copy2_fixed)�Hashes�
MissingHashes)�
indent_log)�display_path�hide_url�is_installable_dir�rmtree)�
TempDirectory)�unpack_file)�vcs)�req�req_tracker�finder�build_isolation�returncCsFt|�}|�|��|�||�Wd�n1s40Y|��S)z(Prepare a distribution for installation.N)r�track�prepare_distribution_metadata�get_metadata_distribution)r&r'r(r)�
abstract_dist�r/��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/prepare.py�_get_prepared_distribution1s*r1)�link�locationr*cCs0t�|j�}|dusJ�|j|t|j�d�dS)N)�url)r%�get_backend_for_scheme�scheme�unpackr r4)r2r3�vcs_backendr/r/r0�unpack_vcs_link>sr9c@s"eZdZeeedd�dd�ZdS)�FileN)�path�content_typer*cCs*||_|dur t�|�d|_n||_dS)Nr)r;�	mimetypes�
guess_typer<)�selfr;r<r/r/r0�__init__Esz
File.__init__)�__name__�
__module__�__qualname__�strrr@r/r/r/r0r:Dsr:)r2�download�download_dir�hashesr*cCsVtddd�}d}|r t|||�}|r.|}d}n|||j�\}}|rL|�|�t||�S)Nr7T��kind�globally_managed)r#�_check_download_dirr;�check_against_pathr:)r2rErFrG�temp_dir�already_downloaded_path�	from_pathr<r/r/r0�get_http_urlMs
rP)�src�destr*c
CsNzt||�Wn:tjyH}z t�dt|�||�WYd}~n
d}~00dS)z�Copying special files is not supported, but as a convenience to users
    we skip errors copying them. This supports tools that may create e.g.
    socket files in the project source directory.
    z>Ignoring special file error '%s' encountered copying %s to %s.N)r�shutil�SpecialFileError�logger�warningrD)rQrR�er/r/r0�_copy2_ignoring_special_fileses�rX)�source�targetr*cs^tj�|�}tj�|��tj�|��tttttd����fdd�}tj�||dt	d�dS)N)�d�namesr*cs6g}|�kr|ddg7}tj�|��kr2|�g7}|S)Nz.toxz.nox)�osr;�abspath)r[r\�skipped�rY�target_basename�target_dirnamer/r0�ignore~s
z!_copy_source_tree.<locals>.ignoreT)rc�symlinks�
copy_function)
r]r;r^�basename�dirnamerDrrS�copytreerX)rYrZ�target_abspathrcr/r`r0�_copy_source_treeys"�rj)r2rFrGr*cCs<d}|rt|||�}|r|}n|j}|r2|�|�t|d�S)z'Get file and optionally check its hash.N)rK�	file_pathrLr:)r2rFrGrNrOr/r/r0�get_file_url�s
rl)r2r3rErFrGr*cCs�|jrt||�dS|��r@tj�|�r0t|�t|j|�dS|j	rVt
|||d�}nt||||d�}|js|t
|j||j�|S)a_Unpack link into location, downloading if required.

    :param hashes: A Hashes object, one of whose embedded hashes must match,
        or HashMismatch will be raised. If the Hashes is empty, no matches are
        required, and unhashable types of requirements (like VCS ones, which
        would ordinarily raise HashUnsupported) are allowed.
    N)rG)�is_vcsr9�is_existing_dirr]r;�isdirr"rjrk�is_filerlrP�is_wheelr$r<)r2r3rErFrG�filer/r/r0�
unpack_url�s&

�	rscCsntj�||j�}tj�|�s dSt�d|�|rjz|�|�Wn*tyht�	d|�t�
|�YdS0|S)z�Check download_dir for previously downloaded file with correct hash
    If a correct file is found return its path else None
    NzFile was already downloaded %sz;Previously-downloaded file %s has bad hash. Re-downloading.)r]r;�join�filename�existsrU�inforLr
rV�unlink)r2rFrG�
download_pathr/r/r0rK�s�
rKcseZdZdZeeeeeeeee	eeeedd�
�fdd�Z
edd�dd�Zeedd	�d
d�Z
eed�dd
�Zeeed�dd�Zd#eeedd�dd�Zd$eeed	�dd�Zd%eeedd�dd�Zeeed	�dd�Zedd�dd�Zeed�dd�Zeeed �d!d"�Z�ZS)&�RequirementPreparerzPrepares a RequirementN)
�	build_dirrF�src_dirr)r'�session�progress_barr(�require_hashes�
use_user_site�
lazy_wheel�
in_tree_buildr*c

stt���||_||_||_||_t||�|_t||�|_	||_
||_||_|	|_
|
|_||_||_i|_d|_dS)N)�r�)�superr@r|r{r'�_sessionr�	_downloadr�_batch_downloadr(rFr)rr��use_lazy_wheelr��_downloaded�_previous_requirement_header)
r?r{rFr|r)r'r}r~r(rr�r�r���	__class__r/r0r@s 
zRequirementPreparer.__init__)r&r*cCs�|jjr$|js$d}tt|jj��}nd}t|jp2|�}||f|jkrZ||f|_t�	||�|jr�t
�� t�	d|jj�Wd�n1s�0YdS)z3Provide context for the requirement being prepared.z
Processing %sz
Collecting %szUsing cached %sN)r2rp�original_link_is_in_wheel_cacherDrrkr&r�rUrwrru)r?r&�message�informationr/r/r0�_log_preparing_link3s
z'RequirementPreparer._log_preparing_link)r&�parallel_buildsr*cCsj|jjrdS|jdusJ�|j��r8|jr8|jj|_dS|j|jd|d�t|j�rft	d�
||j���dS)z1Ensure source_dir of a linked InstallRequirement.NT)�
autodeleter�z�pip can't proceed with requirements '{}' due to apre-existing build directory ({}). This is likely due to a previous installation that failed . pip is being responsible and not assuming it can delete this. Please delete it and try again.)r2rq�
source_dirrnr�rk�ensure_has_source_dirr{r!r�format)r?r&r�r/r/r0�_ensure_link_req_src_dirDs"
�
��z,RequirementPreparer._ensure_link_req_src_dircCsX|js|jdd�S|jjr t��|j��r0t��|jdurF|jsFt	��|jdd�pVt
�S)NT)�trust_internetF)rrGr2rmrrnr	�
original_link�	is_pinnedrr)r?r&r/r/r0�_get_linked_req_hashesis
z*RequirementPreparer._get_linked_req_hashes)r2r*cCs�|js
dS|jrt�d�dS|js*|js:t�d|�dSt|j�}t|j	�}t�
d||j�|j�
dd�d}zt|||j�WSty�t�d|�YdS0dS)	z-Fetch metadata using lazy wheel, if possible.Nz3Lazy wheel is not used as hash checking is requiredz>Lazy wheel is not used as %r does not points to a remote wheelz+Obtaining dependency information from %s %s�#�rz"%s does not support range requests)r�rrU�debugrprqrrur�namerw�versionr4�splitrr�r)r?r2�wheelr�r4r/r/r0� _fetch_metadata_using_lazy_wheel�s0
�

�z4RequirementPreparer._fetch_metadata_using_lazy_wheelF)�partially_downloaded_reqsr�r*c
Cs�tddd�j}i}|D]}|js$J�|||j<q|�|��|�}|D](\}\}}	t�d||�||}||_qD|D]}|�||�qrdS)z>Download any requirements which were only fetched by metadata.r7TrHzDownloading link %s to %sN)	r#r;r2r��keysrUr��local_file_path�_prepare_linked_requirement)
r?r�r�rM�links_to_fully_downloadr&�batch_downloadr2�filepath�_r/r/r0�_complete_partial_requirements�s
�z2RequirementPreparer._complete_partial_requirementscCs�|js
J�|j}|�|�t���d}|jdurP|jrP|�|�}t|j|j|�}|durh||j|jj<n*|�	|�}|dur�d|_
|Wd�S|�||�Wd�S1s�0YdS)z3Prepare a requirement to be obtained from req.link.NT)r2r�rrFrqr�rKr�r4r��needs_more_preparationr�)r?r&r�r2rkrG�
wheel_distr/r/r0�prepare_linked_requirement�s



z.RequirementPreparer.prepare_linked_requirement)�reqsr�r*cCs�dd�|D�}|D]L}|jdur|jjr|�|�}t|j|j|�}|dur||j|jj<d|_qg}|D]"}|jr~|�|�qh|�	||�qh|j
||d�dS)z,Prepare linked requirements more, if needed.cSsg|]}|jr|�qSr/)r�)�.0r&r/r/r0�
<listcomp>��zHRequirementPreparer.prepare_linked_requirements_more.<locals>.<listcomp>NF)r�)rFr2rqr�rKr�r4r��appendr�r�)r?r�r�r&rGrkr�r/r/r0� prepare_linked_requirements_more�s"
�z4RequirementPreparer.prepare_linked_requirements_morec	
Cs�|js
J�|j}|�||�|�|�}|��r:|jr:d}n�|j|jvr�zt||j|j	|j
|�}Wq�ty�}ztd�
|||���WYd}~q�d}~00n&|j|j}|r�|�|�t|dd�}|r�|j|_t||j|j|j�}|S)NzDCould not install requirement {} because of HTTP error {} for URL {})r<)r2r�r�rnr�r4r�rsr�r�rFr
rr�rLr:r;r�r1r'r(r))	r?r&r�r2rG�
local_file�excrk�distr/r/r0r�s<

���
�z/RequirementPreparer._prepare_linked_requirementcCs�|jdusJ�|jdusJ�|j}|js6|��rF|jrF|�|j�dS|��r^t�d|�dS|jdurldSt	j
�|j|j�}t	j
�
|�s�t�|j|�t|�}t�d|�dS)NzENot copying link to destination directory since it is a directory: %szSaved %s)rFr2rmrn�editable�archiverUr�r�r]r;rtrurvrS�copyrrw)r?r&r2�download_locationryr/r/r0�save_linked_requirement*s&�
z+RequirementPreparer.save_linked_requirementcCs�|jsJd��t�d|�t��X|jr6td�|���|�|j�|�	�t
||j|j|j
�}|�|j�Wd�n1s~0Y|S)z Prepare an editable requirement.z-cannot prepare a non-editable req as editablezObtaining %szoThe editable requirement {} cannot be installed when requiring hashes, because there is no single file to hash.N)r�rUrwrrrr�r�r|�update_editabler1r'r(r)�check_if_existsr�)r?r&r�r/r/r0�prepare_editable_requirementDs&���*z0RequirementPreparer.prepare_editable_requirement)r&�skip_reasonr*cCs�|jsJd��|dus&Jd�|j���t�d|||jj�t��,|jrRt�d�t|��	�Wd�S1sr0YdS)z)Prepare an already-installed requirement.z(req should have been satisfied but isn'tNzAdid not get skip reason skipped but req.satisfied_by is set to {}zRequirement %s: %s (%s)z�Since it is already installed, we are trusting this package without checking its hash. To ensure a completely repeatable environment, install into an empty virtualenv.)
�satisfied_byr�rUrwr�rrr�rr-)r?r&r�r/r/r0�prepare_installed_requirementbs
����z1RequirementPreparer.prepare_installed_requirement)F)F)F)rArBrC�__doc__rDr�boolrrrr@rr�r�rr�rrr�rr�r�r�r�r�r�r��
__classcell__r/r/r�r0rz�sd�2�%"�"��!�����(� �rz)NN)NN)NN)Mr��loggingr=r]rS�typingrrrr�pip._vendor.packaging.utilsr�pip._internal.distributionsr�%pip._internal.distributions.installedr�pip._internal.exceptionsr	r
rrr
rr�"pip._internal.index.package_finderr�pip._internal.metadatar�pip._internal.models.linkr�pip._internal.models.wheelr�pip._internal.network.downloadrr� pip._internal.network.lazy_wheelrr�pip._internal.network.sessionrZpip._internal.req.req_installr�pip._internal.req.req_trackerr�pip._internal.utils.filesystemr�pip._internal.utils.hashesrr�pip._internal.utils.loggingr�pip._internal.utils.miscrr r!r"�pip._internal.utils.temp_dirr#�pip._internal.utils.unpackingr$�pip._internal.vcsr%�	getLoggerrArUr�r1rDr9r:rPrXrjrlrsrKrzr/r/r/r0�<module>s�$	
�
������7
�site-packages/pip/_internal/operations/__pycache__/check.cpython-39.pyc000064400000007707150564745010022152 0ustar00a

��Re��@sxdZddlZddlmZmZmZmZmZmZm	Z	ddl
mZddlm
Z
mZddlmZddlmZddlmZdd	lmZe�e�ZGd
d�de�Zee
efZe	e
efZe	e
eefZee
eefZee
eefZe	eefZ e	ee fZ!e	ee"fd�d
d�Z#deeee$ge"fe d�dd�Z%eee!d�dd�Z&eeeee
d�dd�Z'ee
eee
d�dd�Z(dS)z'Validation of dependencies of packages
�N)�Callable�Dict�List�
NamedTuple�Optional�Set�Tuple)�Requirement)�NormalizedName�canonicalize_name)�)make_distribution_for_install_requirement)�get_default_environment)�DistributionVersion)�InstallRequirementc@s"eZdZUeed<eeed<dS)�PackageDetails�version�dependenciesN)�__name__�
__module__�__qualname__r�__annotations__rr	�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/check.pyrs
r)�returncCs�i}d}t�}|jddd�D]f}|j}z t|���}t|j|�||<Wqttfy�}zt	�
d||�d}WYd}~qd}~00q||fS)z3Converts a list of distributions into a PackageSet.Fr)�
local_only�skipz%Error parsing requirements for %s: %sTN)r
�iter_installed_distributions�canonical_name�list�iter_dependenciesrr�OSError�
ValueError�logger�warning)�package_set�problems�env�dist�namer�errr�!create_package_set_from_installed"sr*)r$�
should_ignorercCs�i}i}|��D]�\}}t�}t�}|r2||�r2q|jD]l}t|j�}	|	|vrzd}
|jdurf|j��}
|
r8|�|	|f�q8||	j}|j	j
|dd�s8|�|	||f�q8|r�t|td�||<|rt|td�||<q||fS)z�Check if a package set is consistent

    If should_ignore is passed, it should be a callable that takes a
    package name and returns a boolean.
    TN)�prereleases)�key)
�items�setrrr(�marker�evaluate�addr�	specifier�contains�sorted�str)r$r+�missing�conflicting�package_name�package_detail�missing_deps�conflicting_deps�reqr(�missedrrrr�check_package_set3s0	




r?)�
to_installrcs6t�\}}t||�}t||��|t|�fdd�d�fS)zeFor checking if the dependency graph would be consistent after     installing given requirements
    cs|�vS�Nr)r(��	whitelistrr�<lambda>o�z)check_install_conflicts.<locals>.<lambda>)r+)r*�_simulate_installation_of�_create_whitelistr?)r@r$�_�would_be_installedrrBr�check_install_conflicts`s


��rJ)r@r$rcCsLt�}|D]<}t|�}|��}|j}t|jt|����||<|�|�q
|S)z=Computes the version of packages after installing to_install.)	r/r�get_metadata_distributionrrrrrr2)r@r$�	installed�inst_req�
abstract_distr'r(rrrrFtsrF)rIr$rcCsLt|�}|D]:}||vrq||jD] }t|j�|vr$|�|�qq$q|SrA)r/rrr(r2)rIr$�packages_affectedr9r=rrrrG�s
rG)N))�__doc__�logging�typingrrrrrrrZ"pip._vendor.packaging.requirementsr	�pip._vendor.packaging.utilsr
r�pip._internal.distributionsr�pip._internal.metadatar
Zpip._internal.metadata.baserZpip._internal.req.req_installr�	getLoggerrr"r�
PackageSet�Missing�Conflicting�MissingDict�ConflictingDict�CheckResult�ConflictDetails�boolr*r6r?rJrFrGrrrr�<module>s<$
��-��site-packages/pip/_internal/operations/__pycache__/__init__.cpython-39.pyc000064400000000353150564745010022622 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/__init__.py�<module>�site-packages/pip/_internal/operations/__pycache__/freeze.cpython-39.pyc000064400000014045150564745010022346 0ustar00a

��Re*&�
@s"ddlZddlZddlZddlmZmZmZmZmZm	Z	m
Z
mZddlm
Z
ddlmZddlmZmZddlmZmZddlmZmZddlmZdd	lmZe�e�ZGd
d�de	�Z de
ee!e"e"e
ee!e"e"ee!ee!d�dd�Z#ee!d�dd�Z$ee d�dd�Z%Gdd�d�Z&dS)�N)�	Container�Dict�Iterable�Iterator�List�
NamedTuple�Optional�Set)�canonicalize_name)�Version)�
BadCommand�InstallationError)�BaseDistribution�get_environment)�install_req_from_editable�install_req_from_line)�
COMMENT_RE)�%direct_url_as_pep440_direct_referencec@s"eZdZUeed<eeed<dS)�
_EditableInfo�requirement�commentsN)�__name__�
__module__�__qualname__�str�__annotations__r�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/freeze.pyrs
rFr)r�
local_only�	user_only�paths�isolated�exclude_editable�skip�returnccsni}t|�j|d|d�}|D]$}	t�|	�}
|r6|
jr6q|
||
j<q|�r4t�}t�t	�}|D�]�}
t
|
���p}|D�]X}|��r�|���d�s�|�d�r�|�
�}||vrr|�|�|Vqr|�d�s�|�d��r|�d�r�|dd���}n|td�d����d�}t||d	�}ntt�d
|���|d	�}|j�sRt�d|
|���t�d�qrt|j�}||v�r�||j�s�t�d
|
t�d
|���|j�n||j�|
�qrt||��
�V||=||j�|
�qrWd�q\1�s�0Yq\|��D]4\}}t|�dk�r�t�d|d�tt|�����q�dVt|��dd�d�D] }|j|v�rHt|��
�V�qHdS)Nr)rr#r�#)z-rz
--requirementz-fz--find-linksz-iz--index-urlz--prez--trusted-hostz--process-dependency-linksz--extra-index-urlz
--use-featurez-ez
--editable��=)r!�zWSkipping line in requirement file [%s] because it's not clear what it would install: %sz9  (add #egg=PackageName to the URL to avoid this warning)zBRequirement file [%s] contains %s, but package %r is not installed�z+Requirement %s included multiple times [%s]z, z7## The following requirements were added by pip freeze:cSs
|j��S�N)�name�lower)�xrrr�<lambda>��zfreeze.<locals>.<lambda>)�key) r�iter_installed_distributions�FrozenRequirement�	from_dist�editable�canonical_name�set�collections�defaultdict�list�open�strip�
startswith�rstrip�add�len�lstriprrr�subr+�logger�infor
�warning�appendr�items�join�sorted�values)rrrr r!r"r#Z
installations�dists�dist�reqZemitted_optionsZ	req_filesZ
req_file_path�req_file�lineZline_reqZline_req_canonical_namer+�filesZinstallationrrr�freezes�	�




����

����

�4�rP�rKr$cCs0t|jt�r|j�d|j��S|j�d|j��S)Nz==z===)�
isinstance�versionr�raw_name)rKrrr�_format_as_name_version�srUc
Cs�|j}|sJ�tj�tj�|��}ddlm}m}m}|�	|�}|durtt
|�}t�d||�t
|d|�d�gd�St|�j}z|�||j�}	Wn�|y�t
|�}t
|d|�d	|�d�gd�YS|�y}
z<t
|�}t
|d|�d
|�d�d|
j�d
�gd�WYd}
~
Sd}
~
0t�yHt�d||j�t
|gd�YSt�yx}zt�d|�WYd}~nd}~00t
|	gd�St�d|�t
|dgd�S)za
    Compute and return values (req, comments) for use in
    FrozenRequirement.from_dist().
    r)�RemoteNotFoundError�RemoteNotValidError�vcsNz1No VCS found for editable requirement "%s" in: %rz,# Editable install with no version control (�))rrz# Editable z install with no remote (z
 install (z4) with either a deleted local remote or invalid URI:z# '�'zPcannot determine version of editable source in %s (%s command not found in path)z6Error when trying to get requirement for VCS system %sz-Could not determine repository location of %sz-## !! Could not determine repository location)�editable_project_location�os�path�normcase�abspath�pip._internal.vcsrVrWrX�get_backend_for_dirrUrB�debugr�typer�get_src_requirementrT�urlrrDr+r
)rKr[�locationrVrWrX�vcs_backendZdisplayZvcs_namerL�ex�excrrr�_get_editable_info�s`
��
�
���"�rjc@sJeZdZdeeeeedd�dd�Zeedd�dd��Z	ed	�d
d�Z
dS)
r2rN)r+rLr4rr$cCs&||_t|�|_||_||_||_dSr*)r+r
r5rLr4r)�selfr+rLr4rrrr�__init__�s

zFrozenRequirement.__init__rQcCsN|j}|rt|�\}}n$g}|j}|r4t||j�}nt|�}||j|||d�S)N)r)r4rj�
direct_urlrrTrU)�clsrKr4rLrrmrrrr3�szFrozenRequirement.from_dist)r$cCs4|j}|jrd|��}d�t|j�t|�g�dS)Nz-e �
)rLr4rGr9rr)rkrLrrr�__str__�s
zFrozenRequirement.__str__)r)rrrr�boolrrl�classmethodrr3rprrrrr2�s��
r2)NFFNFFr)'r7�loggingr\�typingrrrrrrrr	�pip._vendor.packaging.utilsr
Zpip._vendor.packaging.versionr�pip._internal.exceptionsrr
�pip._internal.metadatarr�pip._internal.req.constructorsrrZpip._internal.req.req_filer�&pip._internal.utils.direct_url_helpersr�	getLoggerrrBrrrqrPrUrjr2rrrr�<module>s@(
�

�yBsite-packages/pip/_internal/operations/check.py000064400000011765150564745010015662 0ustar00"""Validation of dependencies of packages
"""

import logging
from typing import Callable, Dict, List, NamedTuple, Optional, Set, Tuple

from pip._vendor.packaging.requirements import Requirement
from pip._vendor.packaging.utils import NormalizedName, canonicalize_name

from pip._internal.distributions import make_distribution_for_install_requirement
from pip._internal.metadata import get_default_environment
from pip._internal.metadata.base import DistributionVersion
from pip._internal.req.req_install import InstallRequirement

logger = logging.getLogger(__name__)


class PackageDetails(NamedTuple):
    version: DistributionVersion
    dependencies: List[Requirement]


# Shorthands
PackageSet = Dict[NormalizedName, PackageDetails]
Missing = Tuple[NormalizedName, Requirement]
Conflicting = Tuple[NormalizedName, DistributionVersion, Requirement]

MissingDict = Dict[NormalizedName, List[Missing]]
ConflictingDict = Dict[NormalizedName, List[Conflicting]]
CheckResult = Tuple[MissingDict, ConflictingDict]
ConflictDetails = Tuple[PackageSet, CheckResult]


def create_package_set_from_installed() -> Tuple[PackageSet, bool]:
    """Converts a list of distributions into a PackageSet."""
    package_set = {}
    problems = False
    env = get_default_environment()
    for dist in env.iter_installed_distributions(local_only=False, skip=()):
        name = dist.canonical_name
        try:
            dependencies = list(dist.iter_dependencies())
            package_set[name] = PackageDetails(dist.version, dependencies)
        except (OSError, ValueError) as e:
            # Don't crash on unreadable or broken metadata.
            logger.warning("Error parsing requirements for %s: %s", name, e)
            problems = True
    return package_set, problems


def check_package_set(
    package_set: PackageSet, should_ignore: Optional[Callable[[str], bool]] = None
) -> CheckResult:
    """Check if a package set is consistent

    If should_ignore is passed, it should be a callable that takes a
    package name and returns a boolean.
    """

    missing = {}
    conflicting = {}

    for package_name, package_detail in package_set.items():
        # Info about dependencies of package_name
        missing_deps: Set[Missing] = set()
        conflicting_deps: Set[Conflicting] = set()

        if should_ignore and should_ignore(package_name):
            continue

        for req in package_detail.dependencies:
            name = canonicalize_name(req.name)

            # Check if it's missing
            if name not in package_set:
                missed = True
                if req.marker is not None:
                    missed = req.marker.evaluate()
                if missed:
                    missing_deps.add((name, req))
                continue

            # Check if there's a conflict
            version = package_set[name].version
            if not req.specifier.contains(version, prereleases=True):
                conflicting_deps.add((name, version, req))

        if missing_deps:
            missing[package_name] = sorted(missing_deps, key=str)
        if conflicting_deps:
            conflicting[package_name] = sorted(conflicting_deps, key=str)

    return missing, conflicting


def check_install_conflicts(to_install: List[InstallRequirement]) -> ConflictDetails:
    """For checking if the dependency graph would be consistent after \
    installing given requirements
    """
    # Start from the current state
    package_set, _ = create_package_set_from_installed()
    # Install packages
    would_be_installed = _simulate_installation_of(to_install, package_set)

    # Only warn about directly-dependent packages; create a whitelist of them
    whitelist = _create_whitelist(would_be_installed, package_set)

    return (
        package_set,
        check_package_set(
            package_set, should_ignore=lambda name: name not in whitelist
        ),
    )


def _simulate_installation_of(
    to_install: List[InstallRequirement], package_set: PackageSet
) -> Set[NormalizedName]:
    """Computes the version of packages after installing to_install."""
    # Keep track of packages that were installed
    installed = set()

    # Modify it as installing requirement_set would (assuming no errors)
    for inst_req in to_install:
        abstract_dist = make_distribution_for_install_requirement(inst_req)
        dist = abstract_dist.get_metadata_distribution()
        name = dist.canonical_name
        package_set[name] = PackageDetails(dist.version, list(dist.iter_dependencies()))

        installed.add(name)

    return installed


def _create_whitelist(
    would_be_installed: Set[NormalizedName], package_set: PackageSet
) -> Set[NormalizedName]:
    packages_affected = set(would_be_installed)

    for package_name in package_set:
        if package_name in packages_affected:
            continue

        for req in package_set[package_name].dependencies:
            if canonicalize_name(req.name) in packages_affected:
                packages_affected.add(package_name)
                break

    return packages_affected
site-packages/pip/_internal/operations/install/__pycache__/wheel.cpython-39.pyc000064400000051011150564745010023632 0ustar00a

��Rek�@s�dZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
mZddlmZddlmZmZmZddlmZmZmZmZmZmZmZmZmZmZm Z m!Z!m"Z"m#Z#m$Z$m%Z%ddl&m'Z'm(Z(ddl)m*Z*dd	l+m,Z,dd
l-m.Z.ddl/m0Z0ddl1m2Z2dd
l3m4Z4m5Z5m6Z6ddl7m8Z8m9Z9ddl:m;Z;m<Z<ddl=m>Z>m?Z?ddl@mAZAmBZBmCZCmDZDddlEmFZFmGZGmHZHmIZIddlJmKZKe�r�ddlmLZLGdd�deL�ZMe�NeO�ZPedeQ�ZRe#eReQe$eSeQffZTdPeQeSe#eQeQfd�dd�ZUeQeeQefd�dd�ZVeQeWd�d d!�ZXeeWd"�d#d$�ZYe4e#eeQeQfeeQeQffd%�d&d'�ZZe!eQe eQd(�d)d*�Z[eeTee#eQeQeQfd+�d,d-�Z\eReQd.�d/d0�Z]dQeQe eQeRd1�d2d3�Z^eeeQeeReRfe"eReeQeQeeTd4�d5d6�Z_eeQeQfeeQd7�d8d9�Z`Gd:d;�d;�ZaGd<d=�d=�ZbGd>d?�d?e0�ZceQdd@�dAdB�ZdGdCdD�dDe*�ZedReQe'eQe<eWeWe e9eWddG�	dHdI�ZfejgeQeddJ�dKdL��ZhdSeQeQe<eQeWeWe e9eWddM�	dNdO�ZidS)TzGSupport for installing and building the "wheel" binary package format.
�N)�urlsafe_b64encode)�Message)�chain�filterfalse�starmap)�IO�
TYPE_CHECKING�Any�BinaryIO�Callable�Dict�Iterable�Iterator�List�NewType�Optional�Sequence�Set�Tuple�Union�cast)�ZipFile�ZipInfo)�ScriptMaker)�get_export_entry)�canonicalize_name)�InstallationError)�get_major_minor_version)�BaseDistribution�FilesystemWheel�get_wheel_distribution)�DIRECT_URL_METADATA_NAME�	DirectUrl)�SCHEME_KEYS�Scheme)�adjacent_tmp_file�replace)�captured_stdout�
ensure_dir�	hash_file�	partition)�
current_umask�is_within_directory�2set_extracted_file_to_default_mode_plus_executable�zip_item_is_executable)�parse_wheel)�Protocolc@s4eZdZUded<eed<eed<dd�dd�ZdS)	�File�
RecordPath�src_record_path�	dest_path�changedN��returncCsdS�N���selfr9r9��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/install/wheel.py�saveEsz	File.save)�__name__�
__module__�__qualname__�__annotations__�str�boolr=r9r9r9r<r1@s
r1r2�)�path�	blocksizer7cCs6t||�\}}dt|����d��d�}|t|�fS)z?Return (encoded_digest, length) for path using hashlib.sha256()zsha256=�latin1�=)r)r�digest�decode�rstriprB)rErF�h�lengthrIr9r9r<�rehashOsrN)�moder7cCs|ddd�S)zPReturn keyword arguments to properly open a CSV file
    in the given mode.
    ��utf-8)rO�newline�encodingr9)rOr9r9r<�
csv_io_kwargsVsrT�rEr7cCs�tj�|�sJ�t|d��^}|��}|�d�s>Wd�dStj�t�	��}d|tj
�d�}|��}Wd�n1s~0Yt|d��$}|�|�|�|�Wd�n1s�0YdS)	zQReplace #!python with #!/path/to/python
    Return True if file was changed.
    �rbs#!pythonNFs#!�ascii�wbT)
�osrE�isfile�open�readline�
startswith�sys�
executable�encode�getfilesystemencoding�linesep�read�write)rE�script�	firstline�exename�restr9r9r<�
fix_script]s
&
(ri)�metadatar7cCs|�dd���dkS)NzRoot-Is-PurelibrP�true)�get�lower)rjr9r9r<�wheel_root_is_purelibqsrn)�distr7cCsLi}i}|��D]2}|jdkr,|j||j<q|jdkr|j||j<q||fS)N�console_scripts�gui_scripts)�iter_entry_points�group�value�name)rorprq�entry_pointr9r9r<�get_entrypointsus

rw)�scriptsr7c
s�|sdSt�t�}|D]*}tj�|�}tj�|�}||�|�qdd�tj�	dd��
tj�D����tj�
tj�tj����fdd�|��D�}|s�dSg}|��D]b\}}t|�}t|�dkr�d	�|d
�}	n$d�d�|dd
��d|d
�}	|�d�|	|��q�d}
t|�dk�r0|�|
�d��n|�|
�d��tdd�tj�	dd��
tj�D��}|�rxd}|�|�d�|�S)z�Determine if any scripts are not on PATH and format a warning.
    Returns a warning message if one or more scripts are not on PATH,
    otherwise None.
    NcSs g|]}tj�|��tj��qSr9)rYrE�normcaserK�sep��.0�ir9r9r<�
<listcomp>�s�z5message_about_scripts_not_on_PATH.<locals>.<listcomp>�PATHrPcs&i|]\}}tj�|��vr||�qSr9)rYrEry)r|�
parent_dirrx��
not_warn_dirsr9r<�
<dictcomp>�s�z5message_about_scripts_not_on_PATH.<locals>.<dictcomp>�zscript {} isrzscripts {} are�, ���z and z.The {} installed in '{}' which is not on PATH.zeConsider adding {} to PATH or, if you prefer to suppress this warning, use --no-warn-script-location.zthis directoryzthese directoriescss|]}|r|ddkVqdS)r�~Nr9r{r9r9r<�	<genexpr>�sz4message_about_scripts_not_on_PATH.<locals>.<genexpr>ziNOTE: The current PATH contains path(s) starting with `~`, which may not be expanded by all applications.�
)�collections�defaultdict�setrYrE�dirname�basename�add�environrl�split�pathsep�appendryr^r_�items�sorted�len�format�join�any)
rx�grouped_by_dir�destfiler��script_name�warn_for�	msg_lines�dir_scripts�sorted_scripts�
start_text�
last_line_fmt�warn_for_tilde�tilde_warning_msgr9r�r<�!message_about_scripts_not_on_PATH�sT
�
�������
r�)�outrowsr7cCstdd�|D��S)a�Normalize the given rows of a RECORD file.

    Items in each row are converted into str. Rows are then sorted to make
    the value more predictable for tests.

    Each row is a 3-tuple (path, hash, size) and corresponds to a record of
    a RECORD file (see PEP 376 and PEP 427 for details).  For the rows
    passed to this function, the size can be an integer as an int or string,
    or the empty string.
    css"|]\}}}||t|�fVqdSr8)rB)r|�record_path�hash_�sizer9r9r<r��sz&_normalized_outrows.<locals>.<genexpr>)r�)r�r9r9r<�_normalized_outrows�s�r��r�r7cCs|Sr8r9)r�r9r9r<�_record_to_fs_path�sr�)rE�relative_tor7cCsX|dur>tj�|�d��tj�|�d��kr>tj�||�}|�tjjd�}td|�S)Nr�/r2)rYrE�
splitdriverm�relpathr&rzr)rEr�r9r9r<�_fs_to_record_path�s��r�)�old_csv_rows�	installedr5�	generated�lib_dirr7cCs�g}|D]�}t|�dkr$t�d|�td|d�}|�||�}||vrXtt|��\}	}
n0t|�dkrl|dnd}	t|�dkr�|dnd}
|�||	|
f�q|D]*}t||�}t|�\}	}
|�||	|
f�q�|�	�D]}
|�|
ddf�q�|S)z_
    :param installed: A map from archive RECORD path to installation RECORD
        path.
    �z,RECORD line has more than three elements: %sr2rr�rP�)
r��logger�warningr�poprNr�r�r��values)r�r�r5r�r��installed_rows�row�old_record_path�new_record_pathrIrM�frE�installed_record_pathr9r9r<�get_csv_rows_for_installed�s$
r�)�consoler7cCs|��}g}|�dd�}|r�dtjvr4|�d|�tj�dd�dkr^|�d�tjd|��|�dt	��d	|���d
d�|D�}|D]
}||=q�|�dd�}|r�dtjvr�|�d
|�|�d�t	�|��dd�|D�}|D]
}||=q�|�
tdj|����|S)zk
    Given the mapping from entrypoint name to callable, return the relevant
    console script specs.
    �pipN�ENSUREPIP_OPTIONSzpip = rP�
altinstallz
pip{} = {}rz = cSsg|]}t�d|�r|�qS)zpip(\d(\.\d)?)?$��re�match�r|�kr9r9r<r~J�z,get_console_script_specs.<locals>.<listcomp>�easy_installzeasy_install = zeasy_install-{} = {}cSsg|]}t�d|�r|�qS)zeasy_install(-\d\.\d)?$r�r�r9r9r<r~Xs�{} = {})
�copyr�rYr�r�rlr�r^�version_infor�extendrr�)r��scripts_to_generate�
pip_script�pip_epr��easy_install_script�easy_install_epr9r9r<�get_console_script_specss<#
�
���r�c@s<eZdZeeedd�dd�Zed�dd�Zdd�dd	�Z	dS)
�
ZipBackedFileN)r3r4�zip_filer7cCs||_||_||_d|_dS�NF)r3r4�	_zip_filer5)r;r3r4r�r9r9r<�__init__eszZipBackedFile.__init__r6cCs|j�|j�Sr8)r��getinfor3r:r9r9r<�_getinfomszZipBackedFile._getinfoc	Cs�tj�|j�}t|�tj�|j�r0t�|j�|��}|j�	|��H}t	|jd��}t
�||�Wd�n1st0YWd�n1s�0Yt|�r�t
|j�dS)NrX)rYrEr�r4r(�exists�unlinkr�r�r[�shutil�copyfileobjr.r-)r;r��zipinfor��destr9r9r<r=ps
HzZipBackedFile.save)
r>r?r@r2rBrr�rr�r=r9r9r9r<r�ds
�r�c@s*eZdZddd�dd�Zdd�dd�ZdS)	�
ScriptFiler1N��filer7cCs$||_|jj|_|jj|_d|_dSr�)�_filer3r4r5)r;r�r9r9r<r��s

zScriptFile.__init__r6cCs|j��t|j�|_dSr8)r�r=rir4r5r:r9r9r<r=�s
zScriptFile.save)r>r?r@r�r=r9r9r9r<r��sr�cs$eZdZedd��fdd�Z�ZS)�MissingCallableSuffixN)rvr7cst��d�|��dS)Nz�Invalid script entry point: {} - A callable suffix is required. Cf https://packaging.python.org/specifications/entry-points/#use-for-scripts for more information.)�superr�r�)r;rv��	__class__r9r<r��s
��zMissingCallableSuffix.__init__)r>r?r@rBr��
__classcell__r9r9r�r<r��sr�)�
specificationr7cCs*t|�}|dur&|jdur&tt|���dSr8)r�suffixr�rB)r��entryr9r9r<�_raise_for_invalid_entrypoint�sr�cs4eZdZdeeeefeed��fdd�
Z�ZS)�PipScriptMakerN)r��optionsr7cst|�t��||�Sr8)r�r��make)r;r�r�r�r9r<r��szPipScriptMaker.make)N)	r>r?r@rBrr	rr�r�r9r9r�r<r��sr�TF)	ru�	wheel_zip�
wheel_path�scheme�	pycompile�warn_script_location�
direct_url�	requestedr7c8	st||�\}}	t|	�r|j�n|j�i�t��g}
d4tttdd����fdd�
}ttd�dd�}ttdd	��fd
d��ttt	tgdfd
��fdd�}
tt
t	tgdfd���fdd�}ttd�dd�}ttt|�
��}t||�}t||�\}}|
|��}t||�}ttd�dd�}t||�\}}|||�}t||�}t||�}tt��t|��}t|�\��dtd���fdd�}t||�}t||�}tt|�}t||�}|D] }|��||j|j|j��q�ttd���fdd�}ttd�dd�} |�r�t���}!t���xt�d�|�D]V}"t j!|"d d d!�}#|#�r| |"�}$t"j#�$|$��sHJ�td"|$�%t"j#j&d#��}%||%|$��qWd�n1�s�0YWd�n1�s�0Yt'�(|!�)��t*d|j+�}&d |&_,d$h|&_-d |&_.t/��}'t0t1d%j2��3���}(|&�4|'�})|
�5|)�|
�5|&�4|(d&d i��|�rBt6|)�}*|*du�rBt'�7|*�d't8�@�t9j:tt;tt<d(��fd)d*��}+t"j#�=�|�},t"j#�=|,d+�}-|+|-��}.|.�>d,�Wd�n1�s�0Y|
�?|-�|du�r(t"j#�=|,t@�}/|+|/��$}0|0�>|�A��Bd-��Wd�n1�s0Y|
�?|/�|�rrt"j#�=|,d.�}1tC|1d/��Wd�n1�s^0Y|
�?|1�|�Dd0�}2t0tE�F|2�G���}3tH|3��|
�d1�}4t"j#�=|,d0�}5|+|5fitId2����.}6tE�Jtd3|6��}7|7�KtL|4��Wd�n1�s�0YdS)5a�Install a wheel.

    :param name: Name of the project to install
    :param wheel_zip: open ZipFile for wheel being installed
    :param scheme: Distutils scheme dictating the install directories
    :param req_description: String used in place of the requirement, for
        logging
    :param pycompile: Whether to byte-compile installed Python files
    :param warn_script_location: Whether to check that scripts are installed
        into a directory on PATH
    :raises UnsupportedWheel:
        * when the directory holds an unpacked wheel with incompatible
          Wheel-Version
        * when the .dist-info dir does not match the wheel
    FN)�srcfiler��modifiedr7cs(t|��}|�|<|r$��t|��dS)z6Map archive RECORD paths to installation RECORD paths.N)r�r�)r�r�r��newpath)r5r�r�r9r<�record_installed�s
z(_install_wheel.<locals>.record_installedrUcSs
|�d�S)Nr�)�endswith�rEr9r9r<�is_dir_path�sz#_install_wheel.<locals>.is_dir_path)�
dest_dir_path�target_pathr7cs$t||�s d}t|��||���dS)NzRThe wheel {!r} has a file {!r} trying to install outside the target directory {!r})r,rr�)rr�message)r�r9r<�assert_no_path_traversal�s
��z0_install_wheel.<locals>.assert_no_path_traversalr1)r�r�r7cstdd����fdd�}|S)Nr1r�cs0tj�|�}tj��|�}��|�t||��Sr8)rYrE�normpathr�r�)r��normed_pathr4)rr�r�r9r<�make_root_scheme_file�s
zM_install_wheel.<locals>.root_scheme_file_maker.<locals>.make_root_scheme_file)r2)r�r�r	)r)r�r�r<�root_scheme_file_maker�sz._install_wheel.<locals>.root_scheme_file_maker)r�r�r7cs0�fdd�tD��tdd�����fdd�}|S)Ncsi|]}|t�|��qSr9)�getattr)r|�key)r�r9r<r��r�zB_install_wheel.<locals>.data_scheme_file_maker.<locals>.<dictcomp>r1r�c		s�tj�|�}z|�tjjd�\}}}Wn&tyLd��|�}t|��Yn0z�|}Wn8ty�d�	t
���}d��|||�}t|��Yn0tj�	||�}�||�t||��S)Nr�zbUnexpected file in {}: {!r}. .data directory contents should be named like: '<scheme key>/<path>'.r�z�Unknown scheme key used in {}: {} (for file {!r}). .data directory contents should be in subdirectories named with a valid scheme key ({}))rYrErr�rz�
ValueErrorr�r�KeyErrorr�r�r�)	r�r�_�
scheme_key�dest_subpathr�scheme_path�valid_scheme_keysr4)r�scheme_pathsr�r�r9r<�make_data_scheme_file�s*����
zM_install_wheel.<locals>.data_scheme_file_maker.<locals>.make_data_scheme_file)r#r2)r�r�r)rr�)r�rr�r<�data_scheme_file_maker�sz._install_wheel.<locals>.data_scheme_file_makercSs|�dd�d�d�S)Nr�r�r�.data)r�rrr9r9r<�is_data_scheme_pathsz+_install_wheel.<locals>.is_data_scheme_pathcSs2|�dd�}t|�dko0|d�d�o0|ddkS)Nr�r�rrr�rx)r�r�r)rE�partsr9r9r<�is_script_scheme_path"sz-_install_wheel.<locals>.is_script_scheme_pathr�csz|j}tj�|�}|���d�r.|dd�}n<|���d�rJ|dd�}n |���d�rf|dd�}n|}|�vpx|�vS)Nz.exe���z
-script.pyi����z.pya)r4rYrEr�rmr)r�rEru�	matchname)r��guir9r<�is_entrypoint_wrapper5sz-_install_wheel.<locals>.is_entrypoint_wrapperr6c3sHtt�����D]2}tj��|�}tj�|�s0q|�d�s<q|VqdS)Nz.py)r�r�r�rYrEr�rZr)�installed_path�full_installed_path)r�r�r9r<�pyc_source_file_pathsPs
z-_install_wheel.<locals>.pyc_source_file_pathscSstj�|�S)z8Return the path the pyc file would have been written to.)�	importlib�util�cache_from_sourcerr9r9r<�pyc_output_path]sz'_install_wheel.<locals>.pyc_output_path�ignoreT)�force�quietr2r�rPr�ri�)rE�kwargsr7c;sTt|fi|���}|VWd�n1s,0Yt�|j��t|j|�dSr8)r%rY�chmodrur&)rEr)r�)�generated_file_moder9r<�_generate_file�s$z&_install_wheel.<locals>._generate_file�	INSTALLERspip
rQ�	REQUESTEDrX�RECORD)r�r5r�r��wzIO[str])F)Mr/rn�purelib�platlibr�r2rBrCrrr$rr�namelistrr*�maprr rrrwr�r=r3r4r5rr'�warnings�catch_warnings�filterwarnings�
compileall�compile_filerYrEr�r&rzr��debug�getvaluer�rx�clobber�variants�set_moder��listrr�r��
make_multipler�r�r�r+�
contextlib�contextmanagerr	r
r�rdr�r!�to_jsonr`r[�	read_text�csv�reader�
splitlinesr�rT�writer�	writerowsr�)8rur�r�r�r�r�r�r��info_dirrjr�r�rr
rr�paths�
file_paths�root_scheme_paths�data_scheme_pathsr	�filesr�other_scheme_paths�script_scheme_pathsr�other_scheme_files�distributionr�script_scheme_filesr�r!r%�stdoutrE�success�pyc_path�pyc_record_path�makerr��gui_scripts_to_generate�generated_console_scripts�msgr,�
dest_info_dir�installer_path�installer_file�direct_url_path�direct_url_file�requested_path�record_text�record_rows�rowsr��record_filerHr9)rr5r�r+rr�r�r�r<�_install_wheel�s���	��!


�


��






�N




*


4
 

�	rg)�req_descriptionr7c
csPz
dVWn@tyJ}z(d�||jd�}t|�|�WYd}~n
d}~00dS)NzFor req: {}. {}r)rr��args)rh�err9r9r<�req_error_context�s

rk)	rur�r�rhr�r�r�r�r7c	Cspt|dd��P}t|��(t||||||||d�Wd�n1sD0YWd�n1sb0YdS)NT)�
allowZip64)rur�r�r�r�r�r�r�)rrkrg)	rur�r�rhr�r�r�r��zr9r9r<�
install_wheel�s

�rn)rD)N)TTNF)TTNF)j�__doc__r�r8rArEr"�logging�os.pathrYr�r�r^r5�base64r�
email.messager�	itertoolsrrr�typingrrr	r
rrr
rrrrrrrrr�zipfilerr�pip._vendor.distlib.scriptsrZpip._vendor.distlib.utilr�pip._vendor.packaging.utilsr�pip._internal.exceptionsr�pip._internal.locationsr�pip._internal.metadatarrr �pip._internal.models.direct_urlr!r"�pip._internal.models.schemer#r$�pip._internal.utils.filesystemr%r&�pip._internal.utils.miscr'r(r)r*�pip._internal.utils.unpackingr+r,r-r.�pip._internal.utils.wheelr/r0r1�	getLoggerr>r�rBr2�int�InstalledCSVRowrNrTrCrirnrwr�r�r�r�r�r�r�r�r�r�r�rgrBrkrnr9r9r9r<�<module>s�H	

(I�

� Q(
��
��site-packages/pip/_internal/operations/install/__pycache__/editable_legacy.cpython-39.pyc000064400000002705150564745010025631 0ustar00a

��Re�@s�dZddlZddlmZmZmZddlmZddlm	Z	ddl
mZddlm
Z
e�e�Zeeeeeeeeeeeeeedd�d	d
�ZdS)z?Legacy editable installation process, i.e. `setup.py develop`.
�N)�List�Optional�Sequence)�BuildEnvironment)�
indent_log)�make_setuptools_develop_args)�call_subprocess)�install_options�global_options�prefix�home�
use_user_site�name�
setup_py_path�isolated�	build_env�unpacked_source_directory�returnc
	Cs|t�d|�t|||||||d�}
t��@|�t|
|	d�Wd�n1sP0YWd�n1sn0YdS)z[Install a package in editable mode. Most arguments are pass-through
    to setuptools.
    zRunning setup.py develop for %s)r
r	�no_user_configrrr
)�cwdN)�logger�inforrr)r	r
rrr
rrrrr�args�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/install/editable_legacy.py�install_editables �
�r)�__doc__�logging�typingrrr�pip._internal.build_envr�pip._internal.utils.loggingr�$pip._internal.utils.setuptools_buildr�pip._internal.utils.subprocessr�	getLogger�__name__r�str�boolrrrrr�<module>s&
�site-packages/pip/_internal/operations/install/__pycache__/legacy.cpython-39.pyc000064400000006721150564745010024002 0ustar00a

��Re>�@sdZddlZddlZddlmZddlmZmZmZddl	m
Z
ddlmZddl
mZddlmZdd	lmZdd
lmZddlmZddlmZe�e�ZGd
d�de�Zeeeeedd�dd�Zeeeeeeeeeee e eee ee
eee d�dd�Z!dS)z6Legacy installation process, i.e. `setup.py install`.
�N)�change_root)�List�Optional�Sequence)�BuildEnvironment)�InstallationError)�Scheme)�
indent_log)�
ensure_dir)�make_setuptools_install_args)�runner_with_spinner_message)�
TempDirectoryc@seZdZdS)�LegacyInstallFailureN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/install/legacy.pyrsr)�record_lines�root�req_description�returncs�ttd��fdd�}|D]&}tj�|�}|�d�r||�}qRqd�|�}t|��g}|D]<}|��}	tj�|	�r~|	tjj	7}	|�
tj�||	�|��qZ|��t
|�tj�|d�}
t|
d��$}|�d�|�d�Wd�n1s�0YdS)	N)�pathrcs&�dustj�|�s|St�|�SdS)N)�osr�isabsr)r�rrr�prepend_rootszBwrite_installed_files_from_setuptools_record.<locals>.prepend_rootz	.egg-infoz�{} did not indicate that it installed an .egg-info directory. Only setup.py projects generating .egg-info directories are supported.zinstalled-files.txt�w�
)�strrr�dirname�endswith�formatr�strip�isdir�sep�append�relpath�sortr
�join�open�write)rrrr�line�	directory�egg_info_dir�message�	new_lines�filename�inst_files_path�frrr�,write_installed_files_from_setuptools_records,
��r4)�install_options�global_optionsr�home�prefix�
use_user_site�	pycompile�scheme�
setup_py_path�isolated�req_name�	build_env�unpacked_source_directoryrrcCsf|j}tdd���2}z�tj�|jd�}t||||||||||	|d�}td|
���}t��@|�|||d�Wd�n1s�0YWd�n1s�0Ytj�|�s�t	�
d|�WWd�dSWn*ty�}zt|�WYd}~n
d}~00t
|��}|����}Wd�n1�s,0YWd�n1�sL0Yt|||
�d	S)
N�record)�kindzinstall-record.txt)
r6r5�record_filenamerr8�
header_dirr7r9�no_user_configr:zRunning setup.py install for )�cmd�cwdzRecord file %s not foundFT)�headersr
rrr)rrr	�exists�logger�debug�	Exceptionrr*�read�
splitlinesr4)r5r6rr7r8r9r:r;r<r=r>r?r@rrD�temp_dirrC�install_args�runner�er3rrrr�install>sD���B
LrS)"�__doc__�loggingr�distutils.utilr�typingrrr�pip._internal.build_envr�pip._internal.exceptionsr�pip._internal.models.schemer�pip._internal.utils.loggingr	�pip._internal.utils.miscr
�$pip._internal.utils.setuptools_buildr�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirr
�	getLoggerrrJrLrrr4�boolrSrrrr�<module>sF
�&�site-packages/pip/_internal/operations/install/__pycache__/__init__.cpython-39.pyc000064400000000453150564745010024271 0ustar00a

��Re3�@sdZdS)z,For modules related to installing packages.
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/install/__init__.py�<module>�site-packages/pip/_internal/operations/install/wheel.py000064400000065424150564745010017360 0ustar00"""Support for installing and building the "wheel" binary package format.
"""

import collections
import compileall
import contextlib
import csv
import importlib
import logging
import os.path
import re
import shutil
import sys
import warnings
from base64 import urlsafe_b64encode
from email.message import Message
from itertools import chain, filterfalse, starmap
from typing import (
    IO,
    TYPE_CHECKING,
    Any,
    BinaryIO,
    Callable,
    Dict,
    Iterable,
    Iterator,
    List,
    NewType,
    Optional,
    Sequence,
    Set,
    Tuple,
    Union,
    cast,
)
from zipfile import ZipFile, ZipInfo

from pip._vendor.distlib.scripts import ScriptMaker
from pip._vendor.distlib.util import get_export_entry
from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.exceptions import InstallationError
from pip._internal.locations import get_major_minor_version
from pip._internal.metadata import (
    BaseDistribution,
    FilesystemWheel,
    get_wheel_distribution,
)
from pip._internal.models.direct_url import DIRECT_URL_METADATA_NAME, DirectUrl
from pip._internal.models.scheme import SCHEME_KEYS, Scheme
from pip._internal.utils.filesystem import adjacent_tmp_file, replace
from pip._internal.utils.misc import captured_stdout, ensure_dir, hash_file, partition
from pip._internal.utils.unpacking import (
    current_umask,
    is_within_directory,
    set_extracted_file_to_default_mode_plus_executable,
    zip_item_is_executable,
)
from pip._internal.utils.wheel import parse_wheel

if TYPE_CHECKING:
    from typing import Protocol

    class File(Protocol):
        src_record_path: "RecordPath"
        dest_path: str
        changed: bool

        def save(self) -> None:
            pass


logger = logging.getLogger(__name__)

RecordPath = NewType("RecordPath", str)
InstalledCSVRow = Tuple[RecordPath, str, Union[int, str]]


def rehash(path: str, blocksize: int = 1 << 20) -> Tuple[str, str]:
    """Return (encoded_digest, length) for path using hashlib.sha256()"""
    h, length = hash_file(path, blocksize)
    digest = "sha256=" + urlsafe_b64encode(h.digest()).decode("latin1").rstrip("=")
    return (digest, str(length))


def csv_io_kwargs(mode: str) -> Dict[str, Any]:
    """Return keyword arguments to properly open a CSV file
    in the given mode.
    """
    return {"mode": mode, "newline": "", "encoding": "utf-8"}


def fix_script(path: str) -> bool:
    """Replace #!python with #!/path/to/python
    Return True if file was changed.
    """
    # XXX RECORD hashes will need to be updated
    assert os.path.isfile(path)

    with open(path, "rb") as script:
        firstline = script.readline()
        if not firstline.startswith(b"#!python"):
            return False
        exename = sys.executable.encode(sys.getfilesystemencoding())
        firstline = b"#!" + exename + os.linesep.encode("ascii")
        rest = script.read()
    with open(path, "wb") as script:
        script.write(firstline)
        script.write(rest)
    return True


def wheel_root_is_purelib(metadata: Message) -> bool:
    return metadata.get("Root-Is-Purelib", "").lower() == "true"


def get_entrypoints(dist: BaseDistribution) -> Tuple[Dict[str, str], Dict[str, str]]:
    console_scripts = {}
    gui_scripts = {}
    for entry_point in dist.iter_entry_points():
        if entry_point.group == "console_scripts":
            console_scripts[entry_point.name] = entry_point.value
        elif entry_point.group == "gui_scripts":
            gui_scripts[entry_point.name] = entry_point.value
    return console_scripts, gui_scripts


def message_about_scripts_not_on_PATH(scripts: Sequence[str]) -> Optional[str]:
    """Determine if any scripts are not on PATH and format a warning.
    Returns a warning message if one or more scripts are not on PATH,
    otherwise None.
    """
    if not scripts:
        return None

    # Group scripts by the path they were installed in
    grouped_by_dir: Dict[str, Set[str]] = collections.defaultdict(set)
    for destfile in scripts:
        parent_dir = os.path.dirname(destfile)
        script_name = os.path.basename(destfile)
        grouped_by_dir[parent_dir].add(script_name)

    # We don't want to warn for directories that are on PATH.
    not_warn_dirs = [
        os.path.normcase(i).rstrip(os.sep)
        for i in os.environ.get("PATH", "").split(os.pathsep)
    ]
    # If an executable sits with sys.executable, we don't warn for it.
    #     This covers the case of venv invocations without activating the venv.
    not_warn_dirs.append(os.path.normcase(os.path.dirname(sys.executable)))
    warn_for: Dict[str, Set[str]] = {
        parent_dir: scripts
        for parent_dir, scripts in grouped_by_dir.items()
        if os.path.normcase(parent_dir) not in not_warn_dirs
    }
    if not warn_for:
        return None

    # Format a message
    msg_lines = []
    for parent_dir, dir_scripts in warn_for.items():
        sorted_scripts: List[str] = sorted(dir_scripts)
        if len(sorted_scripts) == 1:
            start_text = "script {} is".format(sorted_scripts[0])
        else:
            start_text = "scripts {} are".format(
                ", ".join(sorted_scripts[:-1]) + " and " + sorted_scripts[-1]
            )

        msg_lines.append(
            "The {} installed in '{}' which is not on PATH.".format(
                start_text, parent_dir
            )
        )

    last_line_fmt = (
        "Consider adding {} to PATH or, if you prefer "
        "to suppress this warning, use --no-warn-script-location."
    )
    if len(msg_lines) == 1:
        msg_lines.append(last_line_fmt.format("this directory"))
    else:
        msg_lines.append(last_line_fmt.format("these directories"))

    # Add a note if any directory starts with ~
    warn_for_tilde = any(
        i[0] == "~" for i in os.environ.get("PATH", "").split(os.pathsep) if i
    )
    if warn_for_tilde:
        tilde_warning_msg = (
            "NOTE: The current PATH contains path(s) starting with `~`, "
            "which may not be expanded by all applications."
        )
        msg_lines.append(tilde_warning_msg)

    # Returns the formatted multiline message
    return "\n".join(msg_lines)


def _normalized_outrows(
    outrows: Iterable[InstalledCSVRow],
) -> List[Tuple[str, str, str]]:
    """Normalize the given rows of a RECORD file.

    Items in each row are converted into str. Rows are then sorted to make
    the value more predictable for tests.

    Each row is a 3-tuple (path, hash, size) and corresponds to a record of
    a RECORD file (see PEP 376 and PEP 427 for details).  For the rows
    passed to this function, the size can be an integer as an int or string,
    or the empty string.
    """
    # Normally, there should only be one row per path, in which case the
    # second and third elements don't come into play when sorting.
    # However, in cases in the wild where a path might happen to occur twice,
    # we don't want the sort operation to trigger an error (but still want
    # determinism).  Since the third element can be an int or string, we
    # coerce each element to a string to avoid a TypeError in this case.
    # For additional background, see--
    # https://github.com/pypa/pip/issues/5868
    return sorted(
        (record_path, hash_, str(size)) for record_path, hash_, size in outrows
    )


def _record_to_fs_path(record_path: RecordPath) -> str:
    return record_path


def _fs_to_record_path(path: str, relative_to: Optional[str] = None) -> RecordPath:
    if relative_to is not None:
        # On Windows, do not handle relative paths if they belong to different
        # logical disks
        if (
            os.path.splitdrive(path)[0].lower()
            == os.path.splitdrive(relative_to)[0].lower()
        ):
            path = os.path.relpath(path, relative_to)
    path = path.replace(os.path.sep, "/")
    return cast("RecordPath", path)


def get_csv_rows_for_installed(
    old_csv_rows: List[List[str]],
    installed: Dict[RecordPath, RecordPath],
    changed: Set[RecordPath],
    generated: List[str],
    lib_dir: str,
) -> List[InstalledCSVRow]:
    """
    :param installed: A map from archive RECORD path to installation RECORD
        path.
    """
    installed_rows: List[InstalledCSVRow] = []
    for row in old_csv_rows:
        if len(row) > 3:
            logger.warning("RECORD line has more than three elements: %s", row)
        old_record_path = cast("RecordPath", row[0])
        new_record_path = installed.pop(old_record_path, old_record_path)
        if new_record_path in changed:
            digest, length = rehash(_record_to_fs_path(new_record_path))
        else:
            digest = row[1] if len(row) > 1 else ""
            length = row[2] if len(row) > 2 else ""
        installed_rows.append((new_record_path, digest, length))
    for f in generated:
        path = _fs_to_record_path(f, lib_dir)
        digest, length = rehash(f)
        installed_rows.append((path, digest, length))
    for installed_record_path in installed.values():
        installed_rows.append((installed_record_path, "", ""))
    return installed_rows


def get_console_script_specs(console: Dict[str, str]) -> List[str]:
    """
    Given the mapping from entrypoint name to callable, return the relevant
    console script specs.
    """
    # Don't mutate caller's version
    console = console.copy()

    scripts_to_generate = []

    # Special case pip and setuptools to generate versioned wrappers
    #
    # The issue is that some projects (specifically, pip and setuptools) use
    # code in setup.py to create "versioned" entry points - pip2.7 on Python
    # 2.7, pip3.3 on Python 3.3, etc. But these entry points are baked into
    # the wheel metadata at build time, and so if the wheel is installed with
    # a *different* version of Python the entry points will be wrong. The
    # correct fix for this is to enhance the metadata to be able to describe
    # such versioned entry points, but that won't happen till Metadata 2.0 is
    # available.
    # In the meantime, projects using versioned entry points will either have
    # incorrect versioned entry points, or they will not be able to distribute
    # "universal" wheels (i.e., they will need a wheel per Python version).
    #
    # Because setuptools and pip are bundled with _ensurepip and virtualenv,
    # we need to use universal wheels. So, as a stopgap until Metadata 2.0, we
    # override the versioned entry points in the wheel and generate the
    # correct ones. This code is purely a short-term measure until Metadata 2.0
    # is available.
    #
    # To add the level of hack in this section of code, in order to support
    # ensurepip this code will look for an ``ENSUREPIP_OPTIONS`` environment
    # variable which will control which version scripts get installed.
    #
    # ENSUREPIP_OPTIONS=altinstall
    #   - Only pipX.Y and easy_install-X.Y will be generated and installed
    # ENSUREPIP_OPTIONS=install
    #   - pipX.Y, pipX, easy_install-X.Y will be generated and installed. Note
    #     that this option is technically if ENSUREPIP_OPTIONS is set and is
    #     not altinstall
    # DEFAULT
    #   - The default behavior is to install pip, pipX, pipX.Y, easy_install
    #     and easy_install-X.Y.
    pip_script = console.pop("pip", None)
    if pip_script:
        if "ENSUREPIP_OPTIONS" not in os.environ:
            scripts_to_generate.append("pip = " + pip_script)

        if os.environ.get("ENSUREPIP_OPTIONS", "") != "altinstall":
            scripts_to_generate.append(
                "pip{} = {}".format(sys.version_info[0], pip_script)
            )

        scripts_to_generate.append(f"pip{get_major_minor_version()} = {pip_script}")
        # Delete any other versioned pip entry points
        pip_ep = [k for k in console if re.match(r"pip(\d(\.\d)?)?$", k)]
        for k in pip_ep:
            del console[k]
    easy_install_script = console.pop("easy_install", None)
    if easy_install_script:
        if "ENSUREPIP_OPTIONS" not in os.environ:
            scripts_to_generate.append("easy_install = " + easy_install_script)

        scripts_to_generate.append(
            "easy_install-{} = {}".format(
                get_major_minor_version(), easy_install_script
            )
        )
        # Delete any other versioned easy_install entry points
        easy_install_ep = [
            k for k in console if re.match(r"easy_install(-\d\.\d)?$", k)
        ]
        for k in easy_install_ep:
            del console[k]

    # Generate the console entry points specified in the wheel
    scripts_to_generate.extend(starmap("{} = {}".format, console.items()))

    return scripts_to_generate


class ZipBackedFile:
    def __init__(
        self, src_record_path: RecordPath, dest_path: str, zip_file: ZipFile
    ) -> None:
        self.src_record_path = src_record_path
        self.dest_path = dest_path
        self._zip_file = zip_file
        self.changed = False

    def _getinfo(self) -> ZipInfo:
        return self._zip_file.getinfo(self.src_record_path)

    def save(self) -> None:
        # directory creation is lazy and after file filtering
        # to ensure we don't install empty dirs; empty dirs can't be
        # uninstalled.
        parent_dir = os.path.dirname(self.dest_path)
        ensure_dir(parent_dir)

        # When we open the output file below, any existing file is truncated
        # before we start writing the new contents. This is fine in most
        # cases, but can cause a segfault if pip has loaded a shared
        # object (e.g. from pyopenssl through its vendored urllib3)
        # Since the shared object is mmap'd an attempt to call a
        # symbol in it will then cause a segfault. Unlinking the file
        # allows writing of new contents while allowing the process to
        # continue to use the old copy.
        if os.path.exists(self.dest_path):
            os.unlink(self.dest_path)

        zipinfo = self._getinfo()

        with self._zip_file.open(zipinfo) as f:
            with open(self.dest_path, "wb") as dest:
                shutil.copyfileobj(f, dest)

        if zip_item_is_executable(zipinfo):
            set_extracted_file_to_default_mode_plus_executable(self.dest_path)


class ScriptFile:
    def __init__(self, file: "File") -> None:
        self._file = file
        self.src_record_path = self._file.src_record_path
        self.dest_path = self._file.dest_path
        self.changed = False

    def save(self) -> None:
        self._file.save()
        self.changed = fix_script(self.dest_path)


class MissingCallableSuffix(InstallationError):
    def __init__(self, entry_point: str) -> None:
        super().__init__(
            "Invalid script entry point: {} - A callable "
            "suffix is required. Cf https://packaging.python.org/"
            "specifications/entry-points/#use-for-scripts for more "
            "information.".format(entry_point)
        )


def _raise_for_invalid_entrypoint(specification: str) -> None:
    entry = get_export_entry(specification)
    if entry is not None and entry.suffix is None:
        raise MissingCallableSuffix(str(entry))


class PipScriptMaker(ScriptMaker):
    def make(self, specification: str, options: Dict[str, Any] = None) -> List[str]:
        _raise_for_invalid_entrypoint(specification)
        return super().make(specification, options)


def _install_wheel(
    name: str,
    wheel_zip: ZipFile,
    wheel_path: str,
    scheme: Scheme,
    pycompile: bool = True,
    warn_script_location: bool = True,
    direct_url: Optional[DirectUrl] = None,
    requested: bool = False,
) -> None:
    """Install a wheel.

    :param name: Name of the project to install
    :param wheel_zip: open ZipFile for wheel being installed
    :param scheme: Distutils scheme dictating the install directories
    :param req_description: String used in place of the requirement, for
        logging
    :param pycompile: Whether to byte-compile installed Python files
    :param warn_script_location: Whether to check that scripts are installed
        into a directory on PATH
    :raises UnsupportedWheel:
        * when the directory holds an unpacked wheel with incompatible
          Wheel-Version
        * when the .dist-info dir does not match the wheel
    """
    info_dir, metadata = parse_wheel(wheel_zip, name)

    if wheel_root_is_purelib(metadata):
        lib_dir = scheme.purelib
    else:
        lib_dir = scheme.platlib

    # Record details of the files moved
    #   installed = files copied from the wheel to the destination
    #   changed = files changed while installing (scripts #! line typically)
    #   generated = files newly generated during the install (script wrappers)
    installed: Dict[RecordPath, RecordPath] = {}
    changed: Set[RecordPath] = set()
    generated: List[str] = []

    def record_installed(
        srcfile: RecordPath, destfile: str, modified: bool = False
    ) -> None:
        """Map archive RECORD paths to installation RECORD paths."""
        newpath = _fs_to_record_path(destfile, lib_dir)
        installed[srcfile] = newpath
        if modified:
            changed.add(_fs_to_record_path(destfile))

    def is_dir_path(path: RecordPath) -> bool:
        return path.endswith("/")

    def assert_no_path_traversal(dest_dir_path: str, target_path: str) -> None:
        if not is_within_directory(dest_dir_path, target_path):
            message = (
                "The wheel {!r} has a file {!r} trying to install"
                " outside the target directory {!r}"
            )
            raise InstallationError(
                message.format(wheel_path, target_path, dest_dir_path)
            )

    def root_scheme_file_maker(
        zip_file: ZipFile, dest: str
    ) -> Callable[[RecordPath], "File"]:
        def make_root_scheme_file(record_path: RecordPath) -> "File":
            normed_path = os.path.normpath(record_path)
            dest_path = os.path.join(dest, normed_path)
            assert_no_path_traversal(dest, dest_path)
            return ZipBackedFile(record_path, dest_path, zip_file)

        return make_root_scheme_file

    def data_scheme_file_maker(
        zip_file: ZipFile, scheme: Scheme
    ) -> Callable[[RecordPath], "File"]:
        scheme_paths = {key: getattr(scheme, key) for key in SCHEME_KEYS}

        def make_data_scheme_file(record_path: RecordPath) -> "File":
            normed_path = os.path.normpath(record_path)
            try:
                _, scheme_key, dest_subpath = normed_path.split(os.path.sep, 2)
            except ValueError:
                message = (
                    "Unexpected file in {}: {!r}. .data directory contents"
                    " should be named like: '<scheme key>/<path>'."
                ).format(wheel_path, record_path)
                raise InstallationError(message)

            try:
                scheme_path = scheme_paths[scheme_key]
            except KeyError:
                valid_scheme_keys = ", ".join(sorted(scheme_paths))
                message = (
                    "Unknown scheme key used in {}: {} (for file {!r}). .data"
                    " directory contents should be in subdirectories named"
                    " with a valid scheme key ({})"
                ).format(wheel_path, scheme_key, record_path, valid_scheme_keys)
                raise InstallationError(message)

            dest_path = os.path.join(scheme_path, dest_subpath)
            assert_no_path_traversal(scheme_path, dest_path)
            return ZipBackedFile(record_path, dest_path, zip_file)

        return make_data_scheme_file

    def is_data_scheme_path(path: RecordPath) -> bool:
        return path.split("/", 1)[0].endswith(".data")

    paths = cast(List[RecordPath], wheel_zip.namelist())
    file_paths = filterfalse(is_dir_path, paths)
    root_scheme_paths, data_scheme_paths = partition(is_data_scheme_path, file_paths)

    make_root_scheme_file = root_scheme_file_maker(wheel_zip, lib_dir)
    files: Iterator[File] = map(make_root_scheme_file, root_scheme_paths)

    def is_script_scheme_path(path: RecordPath) -> bool:
        parts = path.split("/", 2)
        return len(parts) > 2 and parts[0].endswith(".data") and parts[1] == "scripts"

    other_scheme_paths, script_scheme_paths = partition(
        is_script_scheme_path, data_scheme_paths
    )

    make_data_scheme_file = data_scheme_file_maker(wheel_zip, scheme)
    other_scheme_files = map(make_data_scheme_file, other_scheme_paths)
    files = chain(files, other_scheme_files)

    # Get the defined entry points
    distribution = get_wheel_distribution(
        FilesystemWheel(wheel_path),
        canonicalize_name(name),
    )
    console, gui = get_entrypoints(distribution)

    def is_entrypoint_wrapper(file: "File") -> bool:
        # EP, EP.exe and EP-script.py are scripts generated for
        # entry point EP by setuptools
        path = file.dest_path
        name = os.path.basename(path)
        if name.lower().endswith(".exe"):
            matchname = name[:-4]
        elif name.lower().endswith("-script.py"):
            matchname = name[:-10]
        elif name.lower().endswith(".pya"):
            matchname = name[:-4]
        else:
            matchname = name
        # Ignore setuptools-generated scripts
        return matchname in console or matchname in gui

    script_scheme_files: Iterator[File] = map(
        make_data_scheme_file, script_scheme_paths
    )
    script_scheme_files = filterfalse(is_entrypoint_wrapper, script_scheme_files)
    script_scheme_files = map(ScriptFile, script_scheme_files)
    files = chain(files, script_scheme_files)

    for file in files:
        file.save()
        record_installed(file.src_record_path, file.dest_path, file.changed)

    def pyc_source_file_paths() -> Iterator[str]:
        # We de-duplicate installation paths, since there can be overlap (e.g.
        # file in .data maps to same location as file in wheel root).
        # Sorting installation paths makes it easier to reproduce and debug
        # issues related to permissions on existing files.
        for installed_path in sorted(set(installed.values())):
            full_installed_path = os.path.join(lib_dir, installed_path)
            if not os.path.isfile(full_installed_path):
                continue
            if not full_installed_path.endswith(".py"):
                continue
            yield full_installed_path

    def pyc_output_path(path: str) -> str:
        """Return the path the pyc file would have been written to."""
        return importlib.util.cache_from_source(path)

    # Compile all of the pyc files for the installed files
    if pycompile:
        with captured_stdout() as stdout:
            with warnings.catch_warnings():
                warnings.filterwarnings("ignore")
                for path in pyc_source_file_paths():
                    success = compileall.compile_file(path, force=True, quiet=True)
                    if success:
                        pyc_path = pyc_output_path(path)
                        assert os.path.exists(pyc_path)
                        pyc_record_path = cast(
                            "RecordPath", pyc_path.replace(os.path.sep, "/")
                        )
                        record_installed(pyc_record_path, pyc_path)
        logger.debug(stdout.getvalue())

    maker = PipScriptMaker(None, scheme.scripts)

    # Ensure old scripts are overwritten.
    # See https://github.com/pypa/pip/issues/1800
    maker.clobber = True

    # Ensure we don't generate any variants for scripts because this is almost
    # never what somebody wants.
    # See https://bitbucket.org/pypa/distlib/issue/35/
    maker.variants = {""}

    # This is required because otherwise distlib creates scripts that are not
    # executable.
    # See https://bitbucket.org/pypa/distlib/issue/32/
    maker.set_mode = True

    # Generate the console and GUI entry points specified in the wheel
    scripts_to_generate = get_console_script_specs(console)

    gui_scripts_to_generate = list(starmap("{} = {}".format, gui.items()))

    generated_console_scripts = maker.make_multiple(scripts_to_generate)
    generated.extend(generated_console_scripts)

    generated.extend(maker.make_multiple(gui_scripts_to_generate, {"gui": True}))

    if warn_script_location:
        msg = message_about_scripts_not_on_PATH(generated_console_scripts)
        if msg is not None:
            logger.warning(msg)

    generated_file_mode = 0o666 & ~current_umask()

    @contextlib.contextmanager
    def _generate_file(path: str, **kwargs: Any) -> Iterator[BinaryIO]:
        with adjacent_tmp_file(path, **kwargs) as f:
            yield f
        os.chmod(f.name, generated_file_mode)
        replace(f.name, path)

    dest_info_dir = os.path.join(lib_dir, info_dir)

    # Record pip as the installer
    installer_path = os.path.join(dest_info_dir, "INSTALLER")
    with _generate_file(installer_path) as installer_file:
        installer_file.write(b"pip\n")
    generated.append(installer_path)

    # Record the PEP 610 direct URL reference
    if direct_url is not None:
        direct_url_path = os.path.join(dest_info_dir, DIRECT_URL_METADATA_NAME)
        with _generate_file(direct_url_path) as direct_url_file:
            direct_url_file.write(direct_url.to_json().encode("utf-8"))
        generated.append(direct_url_path)

    # Record the REQUESTED file
    if requested:
        requested_path = os.path.join(dest_info_dir, "REQUESTED")
        with open(requested_path, "wb"):
            pass
        generated.append(requested_path)

    record_text = distribution.read_text("RECORD")
    record_rows = list(csv.reader(record_text.splitlines()))

    rows = get_csv_rows_for_installed(
        record_rows,
        installed=installed,
        changed=changed,
        generated=generated,
        lib_dir=lib_dir,
    )

    # Record details of all files installed
    record_path = os.path.join(dest_info_dir, "RECORD")

    with _generate_file(record_path, **csv_io_kwargs("w")) as record_file:
        # Explicitly cast to typing.IO[str] as a workaround for the mypy error:
        # "writer" has incompatible type "BinaryIO"; expected "_Writer"
        writer = csv.writer(cast("IO[str]", record_file))
        writer.writerows(_normalized_outrows(rows))


@contextlib.contextmanager
def req_error_context(req_description: str) -> Iterator[None]:
    try:
        yield
    except InstallationError as e:
        message = "For req: {}. {}".format(req_description, e.args[0])
        raise InstallationError(message) from e


def install_wheel(
    name: str,
    wheel_path: str,
    scheme: Scheme,
    req_description: str,
    pycompile: bool = True,
    warn_script_location: bool = True,
    direct_url: Optional[DirectUrl] = None,
    requested: bool = False,
) -> None:
    with ZipFile(wheel_path, allowZip64=True) as z:
        with req_error_context(req_description):
            _install_wheel(
                name=name,
                wheel_zip=z,
                wheel_path=wheel_path,
                scheme=scheme,
                pycompile=pycompile,
                warn_script_location=warn_script_location,
                direct_url=direct_url,
                requested=requested,
            )
site-packages/pip/_internal/operations/install/legacy.py000064400000010076150564745010017511 0ustar00"""Legacy installation process, i.e. `setup.py install`.
"""

import logging
import os
from distutils.util import change_root
from typing import List, Optional, Sequence

from pip._internal.build_env import BuildEnvironment
from pip._internal.exceptions import InstallationError
from pip._internal.models.scheme import Scheme
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import ensure_dir
from pip._internal.utils.setuptools_build import make_setuptools_install_args
from pip._internal.utils.subprocess import runner_with_spinner_message
from pip._internal.utils.temp_dir import TempDirectory

logger = logging.getLogger(__name__)


class LegacyInstallFailure(Exception):
    pass


def write_installed_files_from_setuptools_record(
    record_lines: List[str],
    root: Optional[str],
    req_description: str,
) -> None:
    def prepend_root(path: str) -> str:
        if root is None or not os.path.isabs(path):
            return path
        else:
            return change_root(root, path)

    for line in record_lines:
        directory = os.path.dirname(line)
        if directory.endswith(".egg-info"):
            egg_info_dir = prepend_root(directory)
            break
    else:
        message = (
            "{} did not indicate that it installed an "
            ".egg-info directory. Only setup.py projects "
            "generating .egg-info directories are supported."
        ).format(req_description)
        raise InstallationError(message)

    new_lines = []
    for line in record_lines:
        filename = line.strip()
        if os.path.isdir(filename):
            filename += os.path.sep
        new_lines.append(os.path.relpath(prepend_root(filename), egg_info_dir))
    new_lines.sort()
    ensure_dir(egg_info_dir)
    inst_files_path = os.path.join(egg_info_dir, "installed-files.txt")
    with open(inst_files_path, "w") as f:
        f.write("\n".join(new_lines) + "\n")


def install(
    install_options: List[str],
    global_options: Sequence[str],
    root: Optional[str],
    home: Optional[str],
    prefix: Optional[str],
    use_user_site: bool,
    pycompile: bool,
    scheme: Scheme,
    setup_py_path: str,
    isolated: bool,
    req_name: str,
    build_env: BuildEnvironment,
    unpacked_source_directory: str,
    req_description: str,
) -> bool:

    header_dir = scheme.headers

    with TempDirectory(kind="record") as temp_dir:
        try:
            record_filename = os.path.join(temp_dir.path, "install-record.txt")
            install_args = make_setuptools_install_args(
                setup_py_path,
                global_options=global_options,
                install_options=install_options,
                record_filename=record_filename,
                root=root,
                prefix=prefix,
                header_dir=header_dir,
                home=home,
                use_user_site=use_user_site,
                no_user_config=isolated,
                pycompile=pycompile,
            )

            runner = runner_with_spinner_message(
                f"Running setup.py install for {req_name}"
            )
            with indent_log(), build_env:
                runner(
                    cmd=install_args,
                    cwd=unpacked_source_directory,
                )

            if not os.path.exists(record_filename):
                logger.debug("Record file %s not found", record_filename)
                # Signal to the caller that we didn't install the new package
                return False

        except Exception as e:
            # Signal to the caller that we didn't install the new package
            raise LegacyInstallFailure from e

        # At this point, we have successfully installed the requirement.

        # We intentionally do not use any encoding to read the file because
        # setuptools writes the file using distutils.file_util.write_file,
        # which does not specify an encoding.
        with open(record_filename) as f:
            record_lines = f.read().splitlines()

    write_installed_files_from_setuptools_record(record_lines, root, req_description)
    return True
site-packages/pip/_internal/operations/install/__init__.py000064400000000063150564745010017777 0ustar00"""For modules related to installing packages.
"""
site-packages/pip/_internal/operations/install/editable_legacy.py000064400000002422150564745010021336 0ustar00"""Legacy editable installation process, i.e. `setup.py develop`.
"""
import logging
from typing import List, Optional, Sequence

from pip._internal.build_env import BuildEnvironment
from pip._internal.utils.logging import indent_log
from pip._internal.utils.setuptools_build import make_setuptools_develop_args
from pip._internal.utils.subprocess import call_subprocess

logger = logging.getLogger(__name__)


def install_editable(
    install_options: List[str],
    global_options: Sequence[str],
    prefix: Optional[str],
    home: Optional[str],
    use_user_site: bool,
    name: str,
    setup_py_path: str,
    isolated: bool,
    build_env: BuildEnvironment,
    unpacked_source_directory: str,
) -> None:
    """Install a package in editable mode. Most arguments are pass-through
    to setuptools.
    """
    logger.info("Running setup.py develop for %s", name)

    args = make_setuptools_develop_args(
        setup_py_path,
        global_options=global_options,
        install_options=install_options,
        no_user_config=isolated,
        prefix=prefix,
        home=home,
        use_user_site=use_user_site,
    )

    with indent_log():
        with build_env:
            call_subprocess(
                args,
                cwd=unpacked_source_directory,
            )
site-packages/pip/_internal/operations/prepare.py000064400000056436150564745010016247 0ustar00"""Prepares a distribution for installation
"""

# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import logging
import mimetypes
import os
import shutil
from typing import Dict, Iterable, List, Optional

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.distributions import make_distribution_for_install_requirement
from pip._internal.distributions.installed import InstalledDistribution
from pip._internal.exceptions import (
    DirectoryUrlHashUnsupported,
    HashMismatch,
    HashUnpinned,
    InstallationError,
    NetworkConnectionError,
    PreviousBuildDirError,
    VcsHashUnsupported,
)
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import BaseDistribution
from pip._internal.models.link import Link
from pip._internal.models.wheel import Wheel
from pip._internal.network.download import BatchDownloader, Downloader
from pip._internal.network.lazy_wheel import (
    HTTPRangeRequestUnsupported,
    dist_from_wheel_url,
)
from pip._internal.network.session import PipSession
from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_tracker import RequirementTracker
from pip._internal.utils.filesystem import copy2_fixed
from pip._internal.utils.hashes import Hashes, MissingHashes
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import display_path, hide_url, is_installable_dir, rmtree
from pip._internal.utils.temp_dir import TempDirectory
from pip._internal.utils.unpacking import unpack_file
from pip._internal.vcs import vcs

logger = logging.getLogger(__name__)


def _get_prepared_distribution(
    req: InstallRequirement,
    req_tracker: RequirementTracker,
    finder: PackageFinder,
    build_isolation: bool,
) -> BaseDistribution:
    """Prepare a distribution for installation."""
    abstract_dist = make_distribution_for_install_requirement(req)
    with req_tracker.track(req):
        abstract_dist.prepare_distribution_metadata(finder, build_isolation)
    return abstract_dist.get_metadata_distribution()


def unpack_vcs_link(link: Link, location: str) -> None:
    vcs_backend = vcs.get_backend_for_scheme(link.scheme)
    assert vcs_backend is not None
    vcs_backend.unpack(location, url=hide_url(link.url))


class File:
    def __init__(self, path: str, content_type: Optional[str]) -> None:
        self.path = path
        if content_type is None:
            self.content_type = mimetypes.guess_type(path)[0]
        else:
            self.content_type = content_type


def get_http_url(
    link: Link,
    download: Downloader,
    download_dir: Optional[str] = None,
    hashes: Optional[Hashes] = None,
) -> File:
    temp_dir = TempDirectory(kind="unpack", globally_managed=True)
    # If a download dir is specified, is the file already downloaded there?
    already_downloaded_path = None
    if download_dir:
        already_downloaded_path = _check_download_dir(link, download_dir, hashes)

    if already_downloaded_path:
        from_path = already_downloaded_path
        content_type = None
    else:
        # let's download to a tmp dir
        from_path, content_type = download(link, temp_dir.path)
        if hashes:
            hashes.check_against_path(from_path)

    return File(from_path, content_type)


def _copy2_ignoring_special_files(src: str, dest: str) -> None:
    """Copying special files is not supported, but as a convenience to users
    we skip errors copying them. This supports tools that may create e.g.
    socket files in the project source directory.
    """
    try:
        copy2_fixed(src, dest)
    except shutil.SpecialFileError as e:
        # SpecialFileError may be raised due to either the source or
        # destination. If the destination was the cause then we would actually
        # care, but since the destination directory is deleted prior to
        # copy we ignore all of them assuming it is caused by the source.
        logger.warning(
            "Ignoring special file error '%s' encountered copying %s to %s.",
            str(e),
            src,
            dest,
        )


def _copy_source_tree(source: str, target: str) -> None:
    target_abspath = os.path.abspath(target)
    target_basename = os.path.basename(target_abspath)
    target_dirname = os.path.dirname(target_abspath)

    def ignore(d: str, names: List[str]) -> List[str]:
        skipped: List[str] = []
        if d == source:
            # Pulling in those directories can potentially be very slow,
            # exclude the following directories if they appear in the top
            # level dir (and only it).
            # See discussion at https://github.com/pypa/pip/pull/6770
            skipped += [".tox", ".nox"]
        if os.path.abspath(d) == target_dirname:
            # Prevent an infinite recursion if the target is in source.
            # This can happen when TMPDIR is set to ${PWD}/...
            # and we copy PWD to TMPDIR.
            skipped += [target_basename]
        return skipped

    shutil.copytree(
        source,
        target,
        ignore=ignore,
        symlinks=True,
        copy_function=_copy2_ignoring_special_files,
    )


def get_file_url(
    link: Link, download_dir: Optional[str] = None, hashes: Optional[Hashes] = None
) -> File:
    """Get file and optionally check its hash."""
    # If a download dir is specified, is the file already there and valid?
    already_downloaded_path = None
    if download_dir:
        already_downloaded_path = _check_download_dir(link, download_dir, hashes)

    if already_downloaded_path:
        from_path = already_downloaded_path
    else:
        from_path = link.file_path

    # If --require-hashes is off, `hashes` is either empty, the
    # link's embedded hash, or MissingHashes; it is required to
    # match. If --require-hashes is on, we are satisfied by any
    # hash in `hashes` matching: a URL-based or an option-based
    # one; no internet-sourced hash will be in `hashes`.
    if hashes:
        hashes.check_against_path(from_path)
    return File(from_path, None)


def unpack_url(
    link: Link,
    location: str,
    download: Downloader,
    download_dir: Optional[str] = None,
    hashes: Optional[Hashes] = None,
) -> Optional[File]:
    """Unpack link into location, downloading if required.

    :param hashes: A Hashes object, one of whose embedded hashes must match,
        or HashMismatch will be raised. If the Hashes is empty, no matches are
        required, and unhashable types of requirements (like VCS ones, which
        would ordinarily raise HashUnsupported) are allowed.
    """
    # non-editable vcs urls
    if link.is_vcs:
        unpack_vcs_link(link, location)
        return None

    # Once out-of-tree-builds are no longer supported, could potentially
    # replace the below condition with `assert not link.is_existing_dir`
    # - unpack_url does not need to be called for in-tree-builds.
    #
    # As further cleanup, _copy_source_tree and accompanying tests can
    # be removed.
    #
    # TODO when use-deprecated=out-of-tree-build is removed
    if link.is_existing_dir():
        if os.path.isdir(location):
            rmtree(location)
        _copy_source_tree(link.file_path, location)
        return None

    # file urls
    if link.is_file:
        file = get_file_url(link, download_dir, hashes=hashes)

    # http urls
    else:
        file = get_http_url(
            link,
            download,
            download_dir,
            hashes=hashes,
        )

    # unpack the archive to the build dir location. even when only downloading
    # archives, they have to be unpacked to parse dependencies, except wheels
    if not link.is_wheel:
        unpack_file(file.path, location, file.content_type)

    return file


def _check_download_dir(
    link: Link, download_dir: str, hashes: Optional[Hashes]
) -> Optional[str]:
    """Check download_dir for previously downloaded file with correct hash
    If a correct file is found return its path else None
    """
    download_path = os.path.join(download_dir, link.filename)

    if not os.path.exists(download_path):
        return None

    # If already downloaded, does its hash match?
    logger.info("File was already downloaded %s", download_path)
    if hashes:
        try:
            hashes.check_against_path(download_path)
        except HashMismatch:
            logger.warning(
                "Previously-downloaded file %s has bad hash. Re-downloading.",
                download_path,
            )
            os.unlink(download_path)
            return None
    return download_path


class RequirementPreparer:
    """Prepares a Requirement"""

    def __init__(
        self,
        build_dir: str,
        download_dir: Optional[str],
        src_dir: str,
        build_isolation: bool,
        req_tracker: RequirementTracker,
        session: PipSession,
        progress_bar: str,
        finder: PackageFinder,
        require_hashes: bool,
        use_user_site: bool,
        lazy_wheel: bool,
        in_tree_build: bool,
    ) -> None:
        super().__init__()

        self.src_dir = src_dir
        self.build_dir = build_dir
        self.req_tracker = req_tracker
        self._session = session
        self._download = Downloader(session, progress_bar)
        self._batch_download = BatchDownloader(session, progress_bar)
        self.finder = finder

        # Where still-packed archives should be written to. If None, they are
        # not saved, and are deleted immediately after unpacking.
        self.download_dir = download_dir

        # Is build isolation allowed?
        self.build_isolation = build_isolation

        # Should hash-checking be required?
        self.require_hashes = require_hashes

        # Should install in user site-packages?
        self.use_user_site = use_user_site

        # Should wheels be downloaded lazily?
        self.use_lazy_wheel = lazy_wheel

        # Should in-tree builds be used for local paths?
        self.in_tree_build = in_tree_build

        # Memoized downloaded files, as mapping of url: path.
        self._downloaded: Dict[str, str] = {}

        # Previous "header" printed for a link-based InstallRequirement
        self._previous_requirement_header = ("", "")

    def _log_preparing_link(self, req: InstallRequirement) -> None:
        """Provide context for the requirement being prepared."""
        if req.link.is_file and not req.original_link_is_in_wheel_cache:
            message = "Processing %s"
            information = str(display_path(req.link.file_path))
        else:
            message = "Collecting %s"
            information = str(req.req or req)

        if (message, information) != self._previous_requirement_header:
            self._previous_requirement_header = (message, information)
            logger.info(message, information)

        if req.original_link_is_in_wheel_cache:
            with indent_log():
                logger.info("Using cached %s", req.link.filename)

    def _ensure_link_req_src_dir(
        self, req: InstallRequirement, parallel_builds: bool
    ) -> None:
        """Ensure source_dir of a linked InstallRequirement."""
        # Since source_dir is only set for editable requirements.
        if req.link.is_wheel:
            # We don't need to unpack wheels, so no need for a source
            # directory.
            return
        assert req.source_dir is None
        if req.link.is_existing_dir() and self.in_tree_build:
            # build local directories in-tree
            req.source_dir = req.link.file_path
            return

        # We always delete unpacked sdists after pip runs.
        req.ensure_has_source_dir(
            self.build_dir,
            autodelete=True,
            parallel_builds=parallel_builds,
        )

        # If a checkout exists, it's unwise to keep going.  version
        # inconsistencies are logged later, but do not fail the
        # installation.
        # FIXME: this won't upgrade when there's an existing
        # package unpacked in `req.source_dir`
        # TODO: this check is now probably dead code
        if is_installable_dir(req.source_dir):
            raise PreviousBuildDirError(
                "pip can't proceed with requirements '{}' due to a"
                "pre-existing build directory ({}). This is likely "
                "due to a previous installation that failed . pip is "
                "being responsible and not assuming it can delete this. "
                "Please delete it and try again.".format(req, req.source_dir)
            )

    def _get_linked_req_hashes(self, req: InstallRequirement) -> Hashes:
        # By the time this is called, the requirement's link should have
        # been checked so we can tell what kind of requirements req is
        # and raise some more informative errors than otherwise.
        # (For example, we can raise VcsHashUnsupported for a VCS URL
        # rather than HashMissing.)
        if not self.require_hashes:
            return req.hashes(trust_internet=True)

        # We could check these first 2 conditions inside unpack_url
        # and save repetition of conditions, but then we would
        # report less-useful error messages for unhashable
        # requirements, complaining that there's no hash provided.
        if req.link.is_vcs:
            raise VcsHashUnsupported()
        if req.link.is_existing_dir():
            raise DirectoryUrlHashUnsupported()

        # Unpinned packages are asking for trouble when a new version
        # is uploaded.  This isn't a security check, but it saves users
        # a surprising hash mismatch in the future.
        # file:/// URLs aren't pinnable, so don't complain about them
        # not being pinned.
        if req.original_link is None and not req.is_pinned:
            raise HashUnpinned()

        # If known-good hashes are missing for this requirement,
        # shim it with a facade object that will provoke hash
        # computation and then raise a HashMissing exception
        # showing the user what the hash should be.
        return req.hashes(trust_internet=False) or MissingHashes()

    def _fetch_metadata_using_lazy_wheel(
        self,
        link: Link,
    ) -> Optional[BaseDistribution]:
        """Fetch metadata using lazy wheel, if possible."""
        if not self.use_lazy_wheel:
            return None
        if self.require_hashes:
            logger.debug("Lazy wheel is not used as hash checking is required")
            return None
        if link.is_file or not link.is_wheel:
            logger.debug(
                "Lazy wheel is not used as %r does not points to a remote wheel",
                link,
            )
            return None

        wheel = Wheel(link.filename)
        name = canonicalize_name(wheel.name)
        logger.info(
            "Obtaining dependency information from %s %s",
            name,
            wheel.version,
        )
        url = link.url.split("#", 1)[0]
        try:
            return dist_from_wheel_url(name, url, self._session)
        except HTTPRangeRequestUnsupported:
            logger.debug("%s does not support range requests", url)
            return None

    def _complete_partial_requirements(
        self,
        partially_downloaded_reqs: Iterable[InstallRequirement],
        parallel_builds: bool = False,
    ) -> None:
        """Download any requirements which were only fetched by metadata."""
        # Download to a temporary directory. These will be copied over as
        # needed for downstream 'download', 'wheel', and 'install' commands.
        temp_dir = TempDirectory(kind="unpack", globally_managed=True).path

        # Map each link to the requirement that owns it. This allows us to set
        # `req.local_file_path` on the appropriate requirement after passing
        # all the links at once into BatchDownloader.
        links_to_fully_download: Dict[Link, InstallRequirement] = {}
        for req in partially_downloaded_reqs:
            assert req.link
            links_to_fully_download[req.link] = req

        batch_download = self._batch_download(
            links_to_fully_download.keys(),
            temp_dir,
        )
        for link, (filepath, _) in batch_download:
            logger.debug("Downloading link %s to %s", link, filepath)
            req = links_to_fully_download[link]
            req.local_file_path = filepath

        # This step is necessary to ensure all lazy wheels are processed
        # successfully by the 'download', 'wheel', and 'install' commands.
        for req in partially_downloaded_reqs:
            self._prepare_linked_requirement(req, parallel_builds)

    def prepare_linked_requirement(
        self, req: InstallRequirement, parallel_builds: bool = False
    ) -> BaseDistribution:
        """Prepare a requirement to be obtained from req.link."""
        assert req.link
        link = req.link
        self._log_preparing_link(req)
        with indent_log():
            # Check if the relevant file is already available
            # in the download directory
            file_path = None
            if self.download_dir is not None and link.is_wheel:
                hashes = self._get_linked_req_hashes(req)
                file_path = _check_download_dir(req.link, self.download_dir, hashes)

            if file_path is not None:
                # The file is already available, so mark it as downloaded
                self._downloaded[req.link.url] = file_path
            else:
                # The file is not available, attempt to fetch only metadata
                wheel_dist = self._fetch_metadata_using_lazy_wheel(link)
                if wheel_dist is not None:
                    req.needs_more_preparation = True
                    return wheel_dist

            # None of the optimizations worked, fully prepare the requirement
            return self._prepare_linked_requirement(req, parallel_builds)

    def prepare_linked_requirements_more(
        self, reqs: Iterable[InstallRequirement], parallel_builds: bool = False
    ) -> None:
        """Prepare linked requirements more, if needed."""
        reqs = [req for req in reqs if req.needs_more_preparation]
        for req in reqs:
            # Determine if any of these requirements were already downloaded.
            if self.download_dir is not None and req.link.is_wheel:
                hashes = self._get_linked_req_hashes(req)
                file_path = _check_download_dir(req.link, self.download_dir, hashes)
                if file_path is not None:
                    self._downloaded[req.link.url] = file_path
                    req.needs_more_preparation = False

        # Prepare requirements we found were already downloaded for some
        # reason. The other downloads will be completed separately.
        partially_downloaded_reqs: List[InstallRequirement] = []
        for req in reqs:
            if req.needs_more_preparation:
                partially_downloaded_reqs.append(req)
            else:
                self._prepare_linked_requirement(req, parallel_builds)

        # TODO: separate this part out from RequirementPreparer when the v1
        # resolver can be removed!
        self._complete_partial_requirements(
            partially_downloaded_reqs,
            parallel_builds=parallel_builds,
        )

    def _prepare_linked_requirement(
        self, req: InstallRequirement, parallel_builds: bool
    ) -> BaseDistribution:
        assert req.link
        link = req.link

        self._ensure_link_req_src_dir(req, parallel_builds)
        hashes = self._get_linked_req_hashes(req)

        if link.is_existing_dir() and self.in_tree_build:
            local_file = None
        elif link.url not in self._downloaded:
            try:
                local_file = unpack_url(
                    link, req.source_dir, self._download, self.download_dir, hashes
                )
            except NetworkConnectionError as exc:
                raise InstallationError(
                    "Could not install requirement {} because of HTTP "
                    "error {} for URL {}".format(req, exc, link)
                )
        else:
            file_path = self._downloaded[link.url]
            if hashes:
                hashes.check_against_path(file_path)
            local_file = File(file_path, content_type=None)

        # For use in later processing,
        # preserve the file path on the requirement.
        if local_file:
            req.local_file_path = local_file.path

        dist = _get_prepared_distribution(
            req,
            self.req_tracker,
            self.finder,
            self.build_isolation,
        )
        return dist

    def save_linked_requirement(self, req: InstallRequirement) -> None:
        assert self.download_dir is not None
        assert req.link is not None
        link = req.link
        if link.is_vcs or (link.is_existing_dir() and req.editable):
            # Make a .zip of the source_dir we already created.
            req.archive(self.download_dir)
            return

        if link.is_existing_dir():
            logger.debug(
                "Not copying link to destination directory "
                "since it is a directory: %s",
                link,
            )
            return
        if req.local_file_path is None:
            # No distribution was downloaded for this requirement.
            return

        download_location = os.path.join(self.download_dir, link.filename)
        if not os.path.exists(download_location):
            shutil.copy(req.local_file_path, download_location)
            download_path = display_path(download_location)
            logger.info("Saved %s", download_path)

    def prepare_editable_requirement(
        self,
        req: InstallRequirement,
    ) -> BaseDistribution:
        """Prepare an editable requirement."""
        assert req.editable, "cannot prepare a non-editable req as editable"

        logger.info("Obtaining %s", req)

        with indent_log():
            if self.require_hashes:
                raise InstallationError(
                    "The editable requirement {} cannot be installed when "
                    "requiring hashes, because there is no single file to "
                    "hash.".format(req)
                )
            req.ensure_has_source_dir(self.src_dir)
            req.update_editable()

            dist = _get_prepared_distribution(
                req,
                self.req_tracker,
                self.finder,
                self.build_isolation,
            )

            req.check_if_exists(self.use_user_site)

        return dist

    def prepare_installed_requirement(
        self,
        req: InstallRequirement,
        skip_reason: str,
    ) -> BaseDistribution:
        """Prepare an already-installed requirement."""
        assert req.satisfied_by, "req should have been satisfied but isn't"
        assert skip_reason is not None, (
            "did not get skip reason skipped but req.satisfied_by "
            "is set to {}".format(req.satisfied_by)
        )
        logger.info(
            "Requirement %s: %s (%s)", skip_reason, req, req.satisfied_by.version
        )
        with indent_log():
            if self.require_hashes:
                logger.debug(
                    "Since it is already installed, we are trusting this "
                    "package without checking its hash. To ensure a "
                    "completely repeatable environment, install into an "
                    "empty virtualenv."
                )
            return InstalledDistribution(req).get_metadata_distribution()
site-packages/pip/_internal/operations/__init__.py000064400000000000150564745010016320 0ustar00site-packages/pip/_internal/operations/freeze.py000064400000023052150564745010016055 0ustar00import collections
import logging
import os
from typing import Container, Dict, Iterable, Iterator, List, NamedTuple, Optional, Set

from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.packaging.version import Version

from pip._internal.exceptions import BadCommand, InstallationError
from pip._internal.metadata import BaseDistribution, get_environment
from pip._internal.req.constructors import (
    install_req_from_editable,
    install_req_from_line,
)
from pip._internal.req.req_file import COMMENT_RE
from pip._internal.utils.direct_url_helpers import direct_url_as_pep440_direct_reference

logger = logging.getLogger(__name__)


class _EditableInfo(NamedTuple):
    requirement: str
    comments: List[str]


def freeze(
    requirement: Optional[List[str]] = None,
    local_only: bool = False,
    user_only: bool = False,
    paths: Optional[List[str]] = None,
    isolated: bool = False,
    exclude_editable: bool = False,
    skip: Container[str] = (),
) -> Iterator[str]:
    installations: Dict[str, FrozenRequirement] = {}

    dists = get_environment(paths).iter_installed_distributions(
        local_only=local_only,
        skip=(),
        user_only=user_only,
    )
    for dist in dists:
        req = FrozenRequirement.from_dist(dist)
        if exclude_editable and req.editable:
            continue
        installations[req.canonical_name] = req

    if requirement:
        # the options that don't get turned into an InstallRequirement
        # should only be emitted once, even if the same option is in multiple
        # requirements files, so we need to keep track of what has been emitted
        # so that we don't emit it again if it's seen again
        emitted_options: Set[str] = set()
        # keep track of which files a requirement is in so that we can
        # give an accurate warning if a requirement appears multiple times.
        req_files: Dict[str, List[str]] = collections.defaultdict(list)
        for req_file_path in requirement:
            with open(req_file_path) as req_file:
                for line in req_file:
                    if (
                        not line.strip()
                        or line.strip().startswith("#")
                        or line.startswith(
                            (
                                "-r",
                                "--requirement",
                                "-f",
                                "--find-links",
                                "-i",
                                "--index-url",
                                "--pre",
                                "--trusted-host",
                                "--process-dependency-links",
                                "--extra-index-url",
                                "--use-feature",
                            )
                        )
                    ):
                        line = line.rstrip()
                        if line not in emitted_options:
                            emitted_options.add(line)
                            yield line
                        continue

                    if line.startswith("-e") or line.startswith("--editable"):
                        if line.startswith("-e"):
                            line = line[2:].strip()
                        else:
                            line = line[len("--editable") :].strip().lstrip("=")
                        line_req = install_req_from_editable(
                            line,
                            isolated=isolated,
                        )
                    else:
                        line_req = install_req_from_line(
                            COMMENT_RE.sub("", line).strip(),
                            isolated=isolated,
                        )

                    if not line_req.name:
                        logger.info(
                            "Skipping line in requirement file [%s] because "
                            "it's not clear what it would install: %s",
                            req_file_path,
                            line.strip(),
                        )
                        logger.info(
                            "  (add #egg=PackageName to the URL to avoid"
                            " this warning)"
                        )
                    else:
                        line_req_canonical_name = canonicalize_name(line_req.name)
                        if line_req_canonical_name not in installations:
                            # either it's not installed, or it is installed
                            # but has been processed already
                            if not req_files[line_req.name]:
                                logger.warning(
                                    "Requirement file [%s] contains %s, but "
                                    "package %r is not installed",
                                    req_file_path,
                                    COMMENT_RE.sub("", line).strip(),
                                    line_req.name,
                                )
                            else:
                                req_files[line_req.name].append(req_file_path)
                        else:
                            yield str(installations[line_req_canonical_name]).rstrip()
                            del installations[line_req_canonical_name]
                            req_files[line_req.name].append(req_file_path)

        # Warn about requirements that were included multiple times (in a
        # single requirements file or in different requirements files).
        for name, files in req_files.items():
            if len(files) > 1:
                logger.warning(
                    "Requirement %s included multiple times [%s]",
                    name,
                    ", ".join(sorted(set(files))),
                )

        yield ("## The following requirements were added by pip freeze:")
    for installation in sorted(installations.values(), key=lambda x: x.name.lower()):
        if installation.canonical_name not in skip:
            yield str(installation).rstrip()


def _format_as_name_version(dist: BaseDistribution) -> str:
    if isinstance(dist.version, Version):
        return f"{dist.raw_name}=={dist.version}"
    return f"{dist.raw_name}==={dist.version}"


def _get_editable_info(dist: BaseDistribution) -> _EditableInfo:
    """
    Compute and return values (req, comments) for use in
    FrozenRequirement.from_dist().
    """
    editable_project_location = dist.editable_project_location
    assert editable_project_location
    location = os.path.normcase(os.path.abspath(editable_project_location))

    from pip._internal.vcs import RemoteNotFoundError, RemoteNotValidError, vcs

    vcs_backend = vcs.get_backend_for_dir(location)

    if vcs_backend is None:
        display = _format_as_name_version(dist)
        logger.debug(
            'No VCS found for editable requirement "%s" in: %r',
            display,
            location,
        )
        return _EditableInfo(
            requirement=location,
            comments=[f"# Editable install with no version control ({display})"],
        )

    vcs_name = type(vcs_backend).__name__

    try:
        req = vcs_backend.get_src_requirement(location, dist.raw_name)
    except RemoteNotFoundError:
        display = _format_as_name_version(dist)
        return _EditableInfo(
            requirement=location,
            comments=[f"# Editable {vcs_name} install with no remote ({display})"],
        )
    except RemoteNotValidError as ex:
        display = _format_as_name_version(dist)
        return _EditableInfo(
            requirement=location,
            comments=[
                f"# Editable {vcs_name} install ({display}) with either a deleted "
                f"local remote or invalid URI:",
                f"# '{ex.url}'",
            ],
        )
    except BadCommand:
        logger.warning(
            "cannot determine version of editable source in %s "
            "(%s command not found in path)",
            location,
            vcs_backend.name,
        )
        return _EditableInfo(requirement=location, comments=[])
    except InstallationError as exc:
        logger.warning("Error when trying to get requirement for VCS system %s", exc)
    else:
        return _EditableInfo(requirement=req, comments=[])

    logger.warning("Could not determine repository location of %s", location)

    return _EditableInfo(
        requirement=location,
        comments=["## !! Could not determine repository location"],
    )


class FrozenRequirement:
    def __init__(
        self,
        name: str,
        req: str,
        editable: bool,
        comments: Iterable[str] = (),
    ) -> None:
        self.name = name
        self.canonical_name = canonicalize_name(name)
        self.req = req
        self.editable = editable
        self.comments = comments

    @classmethod
    def from_dist(cls, dist: BaseDistribution) -> "FrozenRequirement":
        editable = dist.editable
        if editable:
            req, comments = _get_editable_info(dist)
        else:
            comments = []
            direct_url = dist.direct_url
            if direct_url:
                # if PEP 610 metadata is present, use it
                req = direct_url_as_pep440_direct_reference(direct_url, dist.raw_name)
            else:
                # name==version requirement
                req = _format_as_name_version(dist)

        return cls(dist.raw_name, req, editable, comments=comments)

    def __str__(self) -> str:
        req = self.req
        if self.editable:
            req = f"-e {req}"
        return "\n".join(list(self.comments) + [str(req)]) + "\n"
site-packages/pip/_internal/operations/build/__pycache__/metadata_legacy.cpython-39.pyc000064400000004221150564745010025264 0ustar00a

��Re��@s�dZddlZddlZddlmZddlmZddlmZddl	m
Z
ddlmZddl
mZe�e�Zeed	�d
d�Zeeeeeed�d
d�ZdS)z;Metadata generation logic for legacy source distributions.
�N)�BuildEnvironment)�open_spinner)�InstallationError)�make_setuptools_egg_info_args)�call_subprocess)�
TempDirectory)�	directory�returncCsRdd�t�|�D�}|s&td|����t|�dkr@td�|���tj�||d�S)z.Find an .egg-info subdirectory in `directory`.cSsg|]}|�d�r|�qS)z	.egg-info)�endswith)�.0�f�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/metadata_legacy.py�
<listcomp>�z"_find_egg_info.<locals>.<listcomp>z No .egg-info directory found in �z-More than one .egg-info directory found in {}r)�os�listdirr�len�format�path�join)r�	filenamesr
r
r�_find_egg_infos�r)�	build_env�
setup_py_path�
source_dir�isolated�detailsr	c	Cs�t�d||�tddd�j}t|||d�}|�Htd�� }t||d|d�Wd	�n1s^0YWd	�n1s|0Yt|�S)
znGenerate metadata using setup.py-based defacto mechanisms.

    Returns the generated metadata directory.
    z2Running setup.py (path:%s) egg_info for package %szpip-egg-infoT)�kind�globally_managed)�egg_info_dir�no_user_configzPreparing metadata (setup.py)zpython setup.py egg_info)�cwd�command_desc�spinnerN)�logger�debugrrrrrr)rrrrrr!�argsr%r
r
r�generate_metadata s(��
�Br))�__doc__�loggingr�pip._internal.build_envr�pip._internal.cli.spinnersr�pip._internal.exceptionsr�$pip._internal.utils.setuptools_buildr�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirr�	getLogger�__name__r&�strr�boolr)r
r
r
r�<module>s"
�site-packages/pip/_internal/operations/build/__pycache__/wheel_legacy.cpython-39.pyc000064400000005300150564745010024607 0ustar00a

��Re��@s�ddlZddlZddlmZmZddlmZddlm	Z	ddl
mZmZm
Z
e�e�Zeeeed�dd�Zeeeeeeeeed	�d
d�Zeeeeeeeeeed�d
d�ZdS)�N)�List�Optional)�open_spinner)� make_setuptools_bdist_wheel_args)�LOG_DIVIDER�call_subprocess�format_command_args)�command_args�command_output�returncCsbt|�}d|�d�}|s"|d7}n<t��tjkr:|d7}n$|�d�sL|d7}|d|�t��7}|S)z'Format command information for logging.zCommand arguments: �
zCommand output: Nonez'Command output: [use --verbose to show]zCommand output:
)r�logger�getEffectiveLevel�logging�DEBUG�endswithr)r	r
�command_desc�text�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/wheel_legacy.py�format_command_results


r)�names�temp_dir�namer	r
rcCstt|�}|s2d�|�}|t||�7}t�|�dSt|�dkrbd�||�}|t||�7}t�|�tj�||d�S)z>Return the path to the wheel in the temporary build directory.z1Legacy build of wheel for {!r} created no files.
N�zZLegacy build of wheel for {!r} created more than one file.
Filenames (choosing first): {}
r)	�sorted�formatrr
�warning�len�os�path�join)rrrr	r
�msgrrr�get_legacy_build_wheel_path$s	

��
r#)r�
setup_py_path�
source_dir�global_options�
build_options�tempdrc	Cs�t||||d�}d|�d�}t|���}t�d|�zt|||d�}	Wn6tyz|�d�t�d|�YWd�dS0t�	|�}
t
|
||||	d	�}|Wd�S1s�0YdS)
z�Build one unpacked package using the "legacy" build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    )r&r'�destination_dirzBuilding wheel for z (setup.py)zDestination directory: %s)�cwd�spinner�errorzFailed building wheel for %sN)rrrr	r
)rrr
�debugr�	Exception�finishr,r�listdirr#)rr$r%r&r'r(�
wheel_args�spin_messager+�outputr�
wheel_pathrrr�build_wheel_legacy?s8�
�


�r5)r�os.pathr�typingrr�pip._internal.cli.spinnersr�$pip._internal.utils.setuptools_buildr�pip._internal.utils.subprocessrrr�	getLogger�__name__r
�strrr#r5rrrr�<module>s2
���site-packages/pip/_internal/operations/build/__pycache__/wheel.cpython-39.pyc000064400000002324150564745010023266 0ustar00a

��Re'�@s\ddlZddlZddlmZddlmZddlmZe�e	�Z
eeeeeed�dd�ZdS)�N)�Optional)�Pep517HookCaller)�runner_with_spinner_message)�name�backend�metadata_directory�tempd�returncCs�|dusJ�zXt�d|�td|�d��}|�|��|j||d�}Wd�n1sX0YWn ty�t�d|�YdS0tj�	||�S)z�Build one InstallRequirement using the PEP 517 build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    NzDestination directory: %szBuilding wheel for z (pyproject.toml))rzFailed building wheel for %s)
�logger�debugr�subprocess_runner�build_wheel�	Exception�error�os�path�join)rrrr�runner�
wheel_name�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/wheel.py�build_wheel_pep517s

��(r)
�loggingr�typingrZpip._vendor.pep517.wrappersr�pip._internal.utils.subprocessr�	getLogger�__name__r
�strrrrrr�<module>s
�site-packages/pip/_internal/operations/build/__pycache__/metadata_editable.cpython-39.pyc000064400000002357150564745010025601 0ustar00a

��Re��@sRdZddlZddlmZddlmZddlmZddlm	Z	eee
d�dd	�ZdS)
z4Metadata generation logic for source distributions.
�N)�Pep517HookCaller)�BuildEnvironment)�runner_with_spinner_message)�
TempDirectory)�	build_env�backend�returnc	Cs�tddd�}|j}|�Ltd�}|�|��|�|�}Wd�n1sJ0YWd�n1sh0Ytj�||�S)zlGenerate metadata using mechanisms described in PEP 660.

    Returns the generated metadata directory.
    zmodern-metadataT)�kind�globally_managedz,Preparing editable metadata (pyproject.toml)N)r�pathr�subprocess_runner�#prepare_metadata_for_build_editable�os�join)rr�metadata_tmpdir�metadata_dir�runner�distinfo_dir�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/metadata_editable.py�generate_editable_metadata
s�Fr)�__doc__rZpip._vendor.pep517.wrappersr�pip._internal.build_envr�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirr�strrrrrr�<module>s�site-packages/pip/_internal/operations/build/__pycache__/wheel_editable.cpython-39.pyc000064400000002654150564745010025125 0ustar00a

��Re}�@s`ddlZddlZddlmZddlmZmZddlmZe�	e
�Zeeeeeed�dd�Z
dS)�N)�Optional)�HookMissing�Pep517HookCaller)�runner_with_spinner_message)�name�backend�metadata_directory�tempd�returncCs�|dusJ�z�t�d|�td|�d��}|�|��fz|j||d�}WnBty�}z*t�d||�WYd}~Wd�WdSd}~00Wd�n1s�0YWn ty�t�d|�YdS0tj	�
||�S)z�Build one InstallRequirement using the PEP 660 build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    NzDestination directory: %szBuilding editable for z (pyproject.toml))rzLCannot build editable %s because the build backend does not have the %s hookzFailed building editable for %s)�logger�debugr�subprocess_runner�build_editabler�error�	Exception�os�path�join)rrrr	�runner�
wheel_name�e�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/wheel_editable.py�build_wheel_editables.

��
�Hr)�loggingr�typingrZpip._vendor.pep517.wrappersrr�pip._internal.utils.subprocessr�	getLogger�__name__r�strrrrrr�<module>s
�site-packages/pip/_internal/operations/build/__pycache__/__init__.cpython-39.pyc000064400000000361150564745010023720 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/__init__.py�<module>�site-packages/pip/_internal/operations/build/__pycache__/metadata.cpython-39.pyc000064400000002311150564745010023736 0ustar00a

��Re_�@sRdZddlZddlmZddlmZddlmZddlm	Z	eee
d�dd	�ZdS)
z4Metadata generation logic for source distributions.
�N)�Pep517HookCaller)�BuildEnvironment)�runner_with_spinner_message)�
TempDirectory)�	build_env�backend�returnc	Cs�tddd�}|j}|�Ltd�}|�|��|�|�}Wd�n1sJ0YWd�n1sh0Ytj�||�S)zlGenerate metadata using mechanisms described in PEP 517.

    Returns the generated metadata directory.
    zmodern-metadataT)�kind�globally_managedz#Preparing metadata (pyproject.toml)N)r�pathr�subprocess_runner� prepare_metadata_for_build_wheel�os�join)rr�metadata_tmpdir�metadata_dir�runner�distinfo_dir�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/operations/build/metadata.py�generate_metadata
sFr)�__doc__rZpip._vendor.pep517.wrappersr�pip._internal.build_envr�pip._internal.utils.subprocessr�pip._internal.utils.temp_dirr�strrrrrr�<module>ssite-packages/pip/_internal/operations/build/metadata.py000064400000002137150564745010017455 0ustar00"""Metadata generation logic for source distributions.
"""

import os

from pip._vendor.pep517.wrappers import Pep517HookCaller

from pip._internal.build_env import BuildEnvironment
from pip._internal.utils.subprocess import runner_with_spinner_message
from pip._internal.utils.temp_dir import TempDirectory


def generate_metadata(build_env: BuildEnvironment, backend: Pep517HookCaller) -> str:
    """Generate metadata using mechanisms described in PEP 517.

    Returns the generated metadata directory.
    """
    metadata_tmpdir = TempDirectory(kind="modern-metadata", globally_managed=True)

    metadata_dir = metadata_tmpdir.path

    with build_env:
        # Note that Pep517HookCaller implements a fallback for
        # prepare_metadata_for_build_wheel, so we don't have to
        # consider the possibility that this hook doesn't exist.
        runner = runner_with_spinner_message("Preparing metadata (pyproject.toml)")
        with backend.subprocess_runner(runner):
            distinfo_dir = backend.prepare_metadata_for_build_wheel(metadata_dir)

    return os.path.join(metadata_dir, distinfo_dir)
site-packages/pip/_internal/operations/build/wheel_legacy.py000064400000005747150564745010020337 0ustar00import logging
import os.path
from typing import List, Optional

from pip._internal.cli.spinners import open_spinner
from pip._internal.utils.setuptools_build import make_setuptools_bdist_wheel_args
from pip._internal.utils.subprocess import (
    LOG_DIVIDER,
    call_subprocess,
    format_command_args,
)

logger = logging.getLogger(__name__)


def format_command_result(
    command_args: List[str],
    command_output: str,
) -> str:
    """Format command information for logging."""
    command_desc = format_command_args(command_args)
    text = f"Command arguments: {command_desc}\n"

    if not command_output:
        text += "Command output: None"
    elif logger.getEffectiveLevel() > logging.DEBUG:
        text += "Command output: [use --verbose to show]"
    else:
        if not command_output.endswith("\n"):
            command_output += "\n"
        text += f"Command output:\n{command_output}{LOG_DIVIDER}"

    return text


def get_legacy_build_wheel_path(
    names: List[str],
    temp_dir: str,
    name: str,
    command_args: List[str],
    command_output: str,
) -> Optional[str]:
    """Return the path to the wheel in the temporary build directory."""
    # Sort for determinism.
    names = sorted(names)
    if not names:
        msg = ("Legacy build of wheel for {!r} created no files.\n").format(name)
        msg += format_command_result(command_args, command_output)
        logger.warning(msg)
        return None

    if len(names) > 1:
        msg = (
            "Legacy build of wheel for {!r} created more than one file.\n"
            "Filenames (choosing first): {}\n"
        ).format(name, names)
        msg += format_command_result(command_args, command_output)
        logger.warning(msg)

    return os.path.join(temp_dir, names[0])


def build_wheel_legacy(
    name: str,
    setup_py_path: str,
    source_dir: str,
    global_options: List[str],
    build_options: List[str],
    tempd: str,
) -> Optional[str]:
    """Build one unpacked package using the "legacy" build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    """
    wheel_args = make_setuptools_bdist_wheel_args(
        setup_py_path,
        global_options=global_options,
        build_options=build_options,
        destination_dir=tempd,
    )

    spin_message = f"Building wheel for {name} (setup.py)"
    with open_spinner(spin_message) as spinner:
        logger.debug("Destination directory: %s", tempd)

        try:
            output = call_subprocess(
                wheel_args,
                cwd=source_dir,
                spinner=spinner,
            )
        except Exception:
            spinner.finish("error")
            logger.error("Failed building wheel for %s", name)
            return None

        names = os.listdir(tempd)
        wheel_path = get_legacy_build_wheel_path(
            names=names,
            temp_dir=tempd,
            name=name,
            command_args=wheel_args,
            command_output=output,
        )
        return wheel_path
site-packages/pip/_internal/operations/build/metadata_legacy.py000064400000003631150564745010021001 0ustar00"""Metadata generation logic for legacy source distributions.
"""

import logging
import os

from pip._internal.build_env import BuildEnvironment
from pip._internal.cli.spinners import open_spinner
from pip._internal.exceptions import InstallationError
from pip._internal.utils.setuptools_build import make_setuptools_egg_info_args
from pip._internal.utils.subprocess import call_subprocess
from pip._internal.utils.temp_dir import TempDirectory

logger = logging.getLogger(__name__)


def _find_egg_info(directory: str) -> str:
    """Find an .egg-info subdirectory in `directory`."""
    filenames = [f for f in os.listdir(directory) if f.endswith(".egg-info")]

    if not filenames:
        raise InstallationError(f"No .egg-info directory found in {directory}")

    if len(filenames) > 1:
        raise InstallationError(
            "More than one .egg-info directory found in {}".format(directory)
        )

    return os.path.join(directory, filenames[0])


def generate_metadata(
    build_env: BuildEnvironment,
    setup_py_path: str,
    source_dir: str,
    isolated: bool,
    details: str,
) -> str:
    """Generate metadata using setup.py-based defacto mechanisms.

    Returns the generated metadata directory.
    """
    logger.debug(
        "Running setup.py (path:%s) egg_info for package %s",
        setup_py_path,
        details,
    )

    egg_info_dir = TempDirectory(kind="pip-egg-info", globally_managed=True).path

    args = make_setuptools_egg_info_args(
        setup_py_path,
        egg_info_dir=egg_info_dir,
        no_user_config=isolated,
    )

    with build_env:
        with open_spinner("Preparing metadata (setup.py)") as spinner:
            call_subprocess(
                args,
                cwd=source_dir,
                command_desc="python setup.py egg_info",
                spinner=spinner,
            )

    # Return the .egg-info directory.
    return _find_egg_info(egg_info_dir)
site-packages/pip/_internal/operations/build/wheel.py000064400000002047150564745010017001 0ustar00import logging
import os
from typing import Optional

from pip._vendor.pep517.wrappers import Pep517HookCaller

from pip._internal.utils.subprocess import runner_with_spinner_message

logger = logging.getLogger(__name__)


def build_wheel_pep517(
    name: str,
    backend: Pep517HookCaller,
    metadata_directory: str,
    tempd: str,
) -> Optional[str]:
    """Build one InstallRequirement using the PEP 517 build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    """
    assert metadata_directory is not None
    try:
        logger.debug("Destination directory: %s", tempd)

        runner = runner_with_spinner_message(
            f"Building wheel for {name} (pyproject.toml)"
        )
        with backend.subprocess_runner(runner):
            wheel_name = backend.build_wheel(
                tempd,
                metadata_directory=metadata_directory,
            )
    except Exception:
        logger.error("Failed building wheel for %s", name)
        return None
    return os.path.join(tempd, wheel_name)
site-packages/pip/_internal/operations/build/metadata_editable.py000064400000002231150564745010021301 0ustar00"""Metadata generation logic for source distributions.
"""

import os

from pip._vendor.pep517.wrappers import Pep517HookCaller

from pip._internal.build_env import BuildEnvironment
from pip._internal.utils.subprocess import runner_with_spinner_message
from pip._internal.utils.temp_dir import TempDirectory


def generate_editable_metadata(
    build_env: BuildEnvironment, backend: Pep517HookCaller
) -> str:
    """Generate metadata using mechanisms described in PEP 660.

    Returns the generated metadata directory.
    """
    metadata_tmpdir = TempDirectory(kind="modern-metadata", globally_managed=True)

    metadata_dir = metadata_tmpdir.path

    with build_env:
        # Note that Pep517HookCaller implements a fallback for
        # prepare_metadata_for_build_wheel/editable, so we don't have to
        # consider the possibility that this hook doesn't exist.
        runner = runner_with_spinner_message(
            "Preparing editable metadata (pyproject.toml)"
        )
        with backend.subprocess_runner(runner):
            distinfo_dir = backend.prepare_metadata_for_build_editable(metadata_dir)

    return os.path.join(metadata_dir, distinfo_dir)
site-packages/pip/_internal/operations/build/__init__.py000064400000000000150564745010017417 0ustar00site-packages/pip/_internal/operations/build/wheel_editable.py000064400000002575150564745010020640 0ustar00import logging
import os
from typing import Optional

from pip._vendor.pep517.wrappers import HookMissing, Pep517HookCaller

from pip._internal.utils.subprocess import runner_with_spinner_message

logger = logging.getLogger(__name__)


def build_wheel_editable(
    name: str,
    backend: Pep517HookCaller,
    metadata_directory: str,
    tempd: str,
) -> Optional[str]:
    """Build one InstallRequirement using the PEP 660 build process.

    Returns path to wheel if successfully built. Otherwise, returns None.
    """
    assert metadata_directory is not None
    try:
        logger.debug("Destination directory: %s", tempd)

        runner = runner_with_spinner_message(
            f"Building editable for {name} (pyproject.toml)"
        )
        with backend.subprocess_runner(runner):
            try:
                wheel_name = backend.build_editable(
                    tempd,
                    metadata_directory=metadata_directory,
                )
            except HookMissing as e:
                logger.error(
                    "Cannot build editable %s because the build "
                    "backend does not have the %s hook",
                    name,
                    e,
                )
                return None
    except Exception:
        logger.error("Failed building editable for %s", name)
        return None
    return os.path.join(tempd, wheel_name)
site-packages/pip/_internal/commands/__pycache__/help.cpython-39.pyc000064400000002476150564745010021441 0ustar00a

��Rel�@sPddlmZddlmZddlmZddlmZddlm	Z	Gdd�de�Z
dS)	�)�Values)�List)�Command)�SUCCESS)�CommandErrorc@s.eZdZdZdZdZeeee	d�dd�Z
dS)�HelpCommandzShow help for commandsz
      %prog <command>T)�options�args�returnc
Cs�ddlm}m}m}z|d}Wnty6tYS0||vrz||�}d|�d�g}|rl|�d|�d��td�|���||�}	|	j	�
�tS)Nr)�
commands_dict�create_command�get_similar_commandszunknown command "�"zmaybe you meant "z - )�pip._internal.commandsrrr
�
IndexErrorr�appendr�join�parser�
print_help)
�selfrr	rrr
�cmd_name�guess�msg�command�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/help.py�runs

zHelpCommand.runN)�__name__�
__module__�__qualname__�__doc__�usage�ignore_require_venvrr�str�intrrrrrr	srN)�optparser�typingr�pip._internal.cli.base_commandr�pip._internal.cli.status_codesr�pip._internal.exceptionsrrrrrr�<module>s
site-packages/pip/_internal/commands/__pycache__/download.cpython-39.pyc000064400000007637150564745010022324 0ustar00a

��Re(�@s�ddlZddlZddlmZddlmZddlmZddlm	Z	ddl
mZmZddl
mZddlmZdd	lmZmZmZdd
lmZe�e�ZGdd�de�ZdS)
�N)�Values)�List)�
cmdoptions)�make_target_python)�RequirementCommand�with_cleanup)�SUCCESS)�get_requirement_tracker)�
ensure_dir�normalize_path�write_output)�
TempDirectoryc@s<eZdZdZdZdd�dd�Zeeee	e
d�dd	��ZdS)
�DownloadCommandaL
    Download packages from:

    - PyPI (and other indexes) using requirement specifiers.
    - VCS project urls.
    - Local project directories.
    - Local or remote source archives.

    pip also supports downloading from "requirements files", which provide
    an easy way to specify a whole environment to be downloaded.
    a
      %prog [options] <requirement specifier> [package-index-options] ...
      %prog [options] -r <requirements file> [package-index-options] ...
      %prog [options] <vcs project url> ...
      %prog [options] <local project path> ...
      %prog [options] <archive url/path> ...N)�returnc
CsL|j�t���|j�t���|j�t���|j�t���|j�t���|j�t���|j�t�	��|j�t�
��|j�t���|j�t���|j�t�
��|j�t���|j�t���|j�t���|j�t���|jjddddddtjdd�t�|j�t�tj|j�}|j�d	|�|j�d	|j�dS)
Nz-dz--destz--destination-dirz--destination-directory�download_dir�dirzDownload packages into <dir>.)�dest�metavar�default�helpr)�cmd_opts�
add_optionr�constraints�requirements�no_deps�global_options�	no_binary�only_binary�
prefer_binary�src�pre�require_hashes�progress_bar�no_build_isolation�
use_pep517�
no_use_pep517�ignore_requires_python�os�curdir�add_target_python_options�make_option_group�index_group�parser�insert_option_group)�self�
index_opts�r0��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/download.py�add_options%s@��zDownloadCommand.add_options)�options�argsrc	Cs&d|_g|_t�|�t|j�|_t|j�|�|�}t|�}|j	||||j
d�}|�t��}t
|jddd�}|�||||�}|j||||||jdd�}	|j|	|||j
|jd�}
|�|�|
j|dd�}g}|j��D]2}
|
jdur�|
jdus�J�|	�|
�|�|
j�q�|�r"td	d
�|��tS)NT)r3�session�
target_pythonr&�download)�delete�kind�globally_managedF)�temp_build_dirr3�req_trackerr5�finderr�
use_user_site)�preparerr=r3r&�py_version_info)�check_supported_wheelszSuccessfully downloaded %s� )�ignore_installed�	editablesr�check_dist_restrictionrrr
�get_default_sessionr�_build_package_finderr&�
enter_contextr	r
�no_clean�get_requirements�make_requirement_preparer�
make_resolver�python_version�trace_basic_info�resolver�values�satisfied_by�name�save_linked_requirement�appendr�joinr)r.r3r4r5r6r=r<�	directory�reqsr?�resolver�requirement_set�
downloaded�reqr0r0r1�runKs^


���
�


zDownloadCommand.run)�__name__�
__module__�__qualname__�__doc__�usager2rrr�str�intr\r0r0r0r1rs
&r)�loggingr'�optparser�typingrZpip._internal.clirZpip._internal.cli.cmdoptionsr�pip._internal.cli.req_commandrr�pip._internal.cli.status_codesr�pip._internal.req.req_trackerr	�pip._internal.utils.miscr
rr�pip._internal.utils.temp_dirr
�	getLoggerr]�loggerrr0r0r0r1�<module>s
site-packages/pip/_internal/commands/__pycache__/show.cpython-39.pyc000064400000020471150564745010021464 0ustar00a

��Re��@sddlZddlZddlZddlmZddlmZmZmZm	Z	m
Z
ddlmZddl
mZddlmZmZddlmZmZddlmZe�e�ZGd	d
�d
e�ZGdd�de�Ze
ed
fe
ed
fed�dd�Zeeeed�dd�Zeeeeed�dd�Z dS)�N)�Values)�Iterator�List�
NamedTuple�Optional�Tuple��canonicalize_name)�Command)�ERROR�SUCCESS)�BaseDistribution�get_default_environment)�write_outputc@s<eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�ZdS)�ShowCommandzx
    Show information about one or more installed packages.

    The output is in RFC-compliant mail header format.
    z$
      %prog [options] <package> ...TN)�returncCs,|jjddddddd�|j�d|j�dS)	Nz-fz--files�files�
store_trueFz7Show the full list of installed files for each package.)�dest�action�default�helpr)�cmd_opts�
add_option�parser�insert_option_group)�self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/show.py�add_optionss�	zShowCommand.add_options)�options�argsrcCs8|st�d�tS|}t|�}t||j|jd�s4tStS)Nz.ERROR: Please provide a package name or names.)�
list_files�verbose)�logger�warningr�search_packages_info�
print_resultsrr#r)rr r!�query�resultsrrr�run(s

�zShowCommand.run)�__name__�
__module__�__qualname__�__doc__�usage�ignore_require_venvrrr�str�intr*rrrrrs
rc@s�eZdZUeed<eed<eed<eeed<eeed<eed<eed<eeed<eed	<eed
<eed<eed<eed
<eeed<eeeed<dS)�_PackageInfo�name�version�location�requires�required_by�	installer�metadata_version�classifiers�summary�homepage�author�author_email�license�entry_pointsrN)r+r,r-r1�__annotations__rrrrrrr36s
r3.)�entry�inforcCs\|rD|ddkrD|r |ddkr*|d7}n|dd�}|dd�}qttjg|�|�R��S)a�Convert a legacy installed-files.txt path into modern RECORD path.

    The legacy format stores paths relative to the info directory, while the
    modern format stores paths relative to the package root, e.g. the
    site-packages directory.

    :param entry: Path parts of the installed-files.txt entry.
    :param info: Path parts of the egg-info directory relative to package root.
    :returns: The converted entry.

    For best compatibility with symlinks, this does not use ``abspath()`` or
    ``Path.resolve()``, but tries to work with path parts:

    1. While ``entry`` starts with ``..``, remove the equal amounts of parts
       from ``info``; if ``info`` is empty, start appending ``..`` instead.
    2. Join the two directly.
    r�..���)rEN��r1�pathlib�Path)rCrDrrr�_convert_legacy_entryHs
rK)r(rc#s�t�}dd�|��D��dd�|D�}t�fdd�t||�D��}|rXt�dd�|��ttt	d��fd	d
�}tt
tt	d�dd
�}tt
tt	d�dd�}|D�],}z�|}Wnty�Yq�Yn0tdd�|��D�t	j
d�}	t||�t	j
d�}
z|�d�}|jdd�}Wnt�y.g}Yn0||��p@||�}
|
du�rRd}nt|
�}|j}t|jt	|j�|j�pxd|	|
|j|j�p�d|�dg�|�dd�|�dd�|�dd�|�dd�|�dd�||d�Vq�dS)z�
    Gather details from installed distributions. Print distribution name,
    version, location, and installed files. Installed files requires a
    pip generated 'installed-files.txt' in the distributions '.egg-info'
    directory.
    cSsi|]}|j|�qSr)�canonical_name��.0�distrrr�
<dictcomp>l�z(search_packages_info.<locals>.<dictcomp>cSsg|]}t|��qSrr)rNr4rrr�
<listcomp>mrQz(search_packages_info.<locals>.<listcomp>csg|]\}}|�vr|�qSrr)rNr4�pkg��	installedrrrRorQzPackage(s) not found: %s�, )�current_distrcs�fdd����D�S)Nc3s4|],}�jdd�|��D�vr|jdp*dVqdS)cSsh|]}t|j��qSr)r	r4)rN�drrr�	<setcomp>yrQzRsearch_packages_info.<locals>._get_requiring_packages.<locals>.<genexpr>.<setcomp>�Name�UNKNOWNN)rL�iter_dependencies�metadatarM�rWrr�	<genexpr>us
��zHsearch_packages_info.<locals>._get_requiring_packages.<locals>.<genexpr>)�valuesr^rTr^r�_get_requiring_packagests
�z5search_packages_info.<locals>._get_requiring_packages)rOrcSs<z|�d�}Wnty"YdS0dd�t�|���D�S)N�RECORDcss |]}tt�|d��VqdS)rNrH)rN�rowrrrr_�rQzCsearch_packages_info.<locals>._files_from_record.<locals>.<genexpr>)�	read_text�FileNotFoundError�csv�reader�
splitlines)rO�textrrr�_files_from_record|s
z0search_packages_info.<locals>._files_from_recordcs�z|�d�}Wnty"YdS0dd�|jdd�D�}|j}|j}|dusV|durZ|Szt�|��|��Wnty�|YS0�j	s�|S�fdd�|D�S)Nzinstalled-files.txtcss|]}|r|VqdS�Nr�rN�prrrr_�rQzCsearch_packages_info.<locals>._files_from_legacy.<locals>.<genexpr>F��keependsc3s"|]}tt�|�j�j�VqdSrk)rKrIrJ�partsrl�Zinfo_relrrr_�s)
rdrerhr6�info_directoryrIrJ�relative_to�
ValueErrorrp)rOri�paths�rootrDrrqr�_files_from_legacy�s$

�z0search_packages_info.<locals>._files_from_legacycss|]}|jVqdSrk)r4)rN�reqrrrr_�rQz'search_packages_info.<locals>.<genexpr>)�keyzentry_points.txtFrnN��
ClassifierZSummaryz	Home-pageZAuthorzAuthor-emailZLicense)r4r5r6r7r8r9r:r;r<r=r>r?r@rAr)r�iter_distributions�sorted�zipr$r%�joinr
rr1r�KeyErrorr\�lowerrdrhrer]r3�raw_namer5r6r9r:�get_all�get)r(�envZquery_names�missingrarjrwZ
query_namerOr7r8Zentry_points_textrAZ
files_iterrr]rrTrr&csZ�












�r&)�
distributionsr"r#rc	CsPd}t|�D�]<\}}d}|dkr*td�td|j�td|j�td|j�td|j�td	|j�td
|j�td|j�td|j	�td
d�
|j��tdd�
|j��|�rtd|j
�td|j�td�|jD]}td|�q�td�|jD]}td|���q�|rtd�|jdu�r.td�q|jD]}td|����q4q|S)zC
    Print the information from installed distributions found.
    FTrz---zName: %szVersion: %szSummary: %sz
Home-page: %sz
Author: %szAuthor-email: %szLicense: %szLocation: %szRequires: %srVzRequired-by: %szMetadata-Version: %sz
Installer: %szClassifiers:z  %sz
Entry-points:zFiles:Nz+Cannot locate RECORD or installed-files.txt)�	enumeraterr4r5r<r=r>r?r@r6rr7r8r:r9r;rA�stripr)	r�r"r#Zresults_printed�irO�
classifierrC�linerrrr'�s>



r')!rf�loggingrI�optparser�typingrrrrr�pip._vendor.packaging.utilsr	�pip._internal.cli.base_commandr
�pip._internal.cli.status_codesrr�pip._internal.metadatar
r�pip._internal.utils.miscr�	getLoggerr+r$rr3r1rKr&�boolr'rrrr�<module>s&
%"`�site-packages/pip/_internal/commands/__pycache__/install.cpython-39.pyc000064400000042472150564745010022157 0ustar00a

��Re�l�@sHddlZddlZddlZddlZddlZddlmZmZddlm	Z	m
Z
mZddlm
Z
ddlmZddlmZddlmZddlmZmZmZdd	lmZmZdd
lmZmZddlmZddl m!Z!dd
l"m#Z#ddl$m%Z%m&Z&ddl'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6m7Z7m8Z8m9Z9ddl:m;Z;ddl<m=Z=m>Z>ddl?m@Z@mAZAmBZBe4eC�ZDe#e@d�dd�ZEGdd�de�ZFd/eGeeHeeHeGeeHe
eHd �d!d"�ZIeeHeGeGd#�d$d%�ZJd0eeGeeHeeHeeHeGeGd&�d'd(�ZKe
e*ee
eHdd)�d*d+�ZLeMeGeGeHd,�d-d.�ZNdS)1�N)�
SUPPRESS_HELP�Values)�Iterable�List�Optional)�canonicalize_name)�
WheelCache)�
cmdoptions)�make_target_python)�RequirementCommand�warn_if_run_as_root�with_cleanup)�ERROR�SUCCESS)�CommandError�InstallationError)�
get_scheme)�get_environment)�
FormatControl)�ConflictDetails�check_install_conflicts)�install_given_reqs)�InstallRequirement)�get_requirement_tracker)�WINDOWS)�parse_distutils_args��test_writable_dir)�	getLogger)�
ensure_dir�get_pip_version�(protect_pip_from_modification_on_windows�write_output)�
TempDirectory)�running_under_virtualenv�virtualenv_no_global)�BinaryAllowedPredicate�build� should_build_for_install_command)�format_control�returncsttd��fdd�}|S)N)�reqr*cs t|jp
d�}��|�}d|vS)N��binary)r�name�get_allowed_formats)r+�canonical_name�allowed_formats�r)���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/install.py�check_binary_allowed5s
z6get_check_binary_allowed.<locals>.check_binary_allowed)r�bool)r)r5r3r2r4�get_check_binary_allowed4sr7c@szeZdZdZdZdd�dd�Zeeee	e
d�dd	��Ze	ee
dd
�dd�Zeeeed
�dd�Zee	dd�dd�ZdS)�InstallCommandaI
    Install packages from:

    - PyPI (and other indexes) using requirement specifiers.
    - VCS project urls.
    - Local project directories.
    - Local or remote source archives.

    pip also supports installing from "requirements files", which provide
    an easy way to specify a whole environment to be installed.
    a%
      %prog [options] <requirement specifier> [package-index-options] ...
      %prog [options] -r <requirements file> [package-index-options] ...
      %prog [options] [-e] <vcs project url> ...
      %prog [options] [-e] <local project path> ...
      %prog [options] <archive url/path> ...N)r*cCsj|j�t���|j�t���|j�t���|j�t���|j�t���|jjddddddd�t�|j�|jjddd	d
d�|jjddd
t	d�|jjdddddd�|jjdddddd�|j�t�
��|jjdddd	dd�|jjdddddgdd�|jjddd	d d�|jjd!d"d#d	d$d�|j�t���|j�t���|j�t�
��|j�t���|j�t���|j�t���|jjd%d	d&d'd(d)�|jjd*d
d&d+d,�|jjd-d
d.d'd/d)�|jjd0d
d1d'd2d)�|j�t���|j�t���|j�t���|j�t���|j�t���t�tj|j�}|j�d3|�|j�d3|j�dS)4Nz-tz--target�
target_dir�dirz�Install packages into <dir>. By default this will not replace existing files/folders in <dir>. Use --upgrade to replace existing packages in <dir> with new versions.)�dest�metavar�default�helpz--user�
use_user_site�
store_truez�Install to the Python user install directory for your platform. Typically ~/.local/, or %APPDATA%\Python on Windows. (See the Python documentation for site.USER_BASE for full details.))r;�actionr>z	--no-user�store_falsez--root�	root_pathz=Install everything relative to this alternate root directory.z--prefix�prefix_pathzIInstallation prefix where lib, bin and other top-level folders are placedz-Uz	--upgrade�upgradez�Upgrade all specified packages to the newest available version. The handling of dependencies depends on the upgrade-strategy used.z--upgrade-strategy�upgrade_strategyzonly-if-needed�eageraGDetermines how dependency upgrading should be handled [default: %default]. "eager" - dependencies are upgraded regardless of whether the currently installed version satisfies the requirements of the upgraded package(s). "only-if-needed" -  are upgraded only when they do not satisfy the requirements of the upgraded package(s).)r;r=�choicesr>z--force-reinstall�force_reinstallz;Reinstall all packages even if they are already up-to-date.z-Iz--ignore-installed�ignore_installedz�Ignore the installed packages, overwriting them. This can break your system if the existing package is of a different version or was installed with a different package manager!z	--compile�compileTz'Compile Python source files to bytecode)rAr;r=r>z--no-compilez.Do not compile Python source files to bytecode)rAr;r>z--no-warn-script-location�warn_script_locationz0Do not warn when installing scripts outside PATHz--no-warn-conflicts�warn_about_conflictsz%Do not warn about broken dependenciesr)�cmd_opts�
add_optionr	�requirements�constraints�no_deps�pre�editable�add_target_python_optionsr�src�ignore_requires_python�no_build_isolation�
use_pep517�
no_use_pep517�install_options�global_options�	no_binary�only_binary�
prefer_binary�require_hashes�progress_bar�make_option_group�index_group�parser�insert_option_group)�self�
index_optsr3r3r4�add_optionsQs��
��������
�����zInstallCommand.add_options)�options�argsr*c)s.|jr|jdurtd��t�|�d}|jr2|j}tj|dd�|jpHg}t	�
dt��t|j|j
|j|j|jd�|_d}d}|jr�d|_tj�|j�|_tj�|j�r�tj�|j�s�td��tdd	�}|j}|�|�|jp�g}|�|�}t|�}	|j|||	|jd
�}
t|j|j�}|�t ��}t|j!ddd�}
�zx|�"|||
|�}|D]}d|_#�qLt$||j�|j%|
||||
|jd
�}|j&||
|||j|j|j|j'||j(d�
}|�)|
�|j*||jd�}z|�+d�}Wnt,�y�d}Yn0|j-du}t.|d�t/|
j���fdd�|j0�1�D�}t2||dggd�\}}dd�|D�}|�r^t3d�4d�5|����|D]}|j(�sbd|_6�qb|�7|�}d}|j8�o�|j9}|�r�|�:|�}|j;}|j�s�|j
�r�d}t<||||j||j
||j|j=d�	}t>|j||j|j
|jd�}t?|�} |j@tA�Bd�d�g}!|D]T}"|"jC}#z(| �D|#�}$|$du�rN|#�d|$jE��}#WntF�ydYn0|!�G|#��q|du�r�|jH||�I|�d�d �5|!�}%|%�r�tJd!|%�WnPtK�y�}&z6|jLd"k}'tM|&|'|j�}(t	jN|(|'d#�tOWYd}&~&Sd}&~&00|j�r$|�sJ�|�P|j||j�tQ�tRS)$Nz'Can not combine '--user' and '--target'zto-satisfy-onlyT)�check_targetzUsing %s)rDr9rC�
isolated_modez=Target path exists but is not a directory, will not continue.�target)�kind)ri�session�
target_pythonrW�install)�deletern�globally_managed)�temp_build_dirri�req_trackerro�finderr?)
�preparerrvri�wheel_cacher?rJrWrIrFrY)�check_supported_wheels�pipF)�
modifying_pipcsg|]}t|��r|�qSr3)r(��.0�r�r5r3r4�
<listcomp>bs
�z&InstallCommand.run.<locals>.<listcomp>)rx�verify�
build_optionsr\cSsg|]}|jr|j�qSr3)rYr.r|r3r3r4r�rszYCould not build wheels for {}, which is required to install pyproject.toml-based projectsz, i� )�root�home�prefixrLr?�	pycompile)�userr�r�r��isolatedr.)�key�-)�resolver_variant� zSuccessfully installed %s�)�exc_info)Sr?r9rr	�check_install_build_globalrErF�check_dist_restrictionr[�logger�verboser �decide_user_installrDrCrlrJ�os�path�abspath�exists�isdirr#�
enter_contextr\�get_default_sessionr
�_build_package_finderrWr�	cache_dirr)r�no_clean�get_requirements�permit_editable_wheels�'reject_location_related_install_options�make_requirement_preparer�
make_resolverrIrY�trace_basic_info�resolve�get_requirement�KeyError�satisfied_byr!r7rP�valuesr'r�format�join�legacy_install_reason�get_installation_order�ignore_dependenciesrM�_determine_conflictsrLrrK�get_lib_location_guessesr�sort�operator�
attrgetterr.�get_distribution�version�	Exception�append�_warn_about_conflicts�determine_resolver_variantr"�OSError�	verbosity�create_os_error_message�errorr�_handle_target_dirrr))rfrirjrFr[�target_temp_dir�target_temp_dir_pathr\rorprvrxru�	directory�reqsr+rw�resolver�requirement_set�pip_reqr{�
reqs_to_build�_�build_failures�pep517_build_failure_namesr~�
to_install�	conflicts�should_warn_about_conflictsrL�	installed�
lib_locations�env�items�result�item�installed_dist�installed_descr��show_traceback�messager3rr4�run�sV

����



��
��
�




��

���


�
��


�
�
�

�zInstallCommand.run)r9r�rEr*csNt|�g}td|jd�}|j}|j}|j}tj�|�rB|�|�tj�|�r`||kr`|�|�tj�|�rv|�|�|D]�}	t�	|	�D]�}
|	|kr�tj�
||
��t�fdd�|dd�D��r�q�tj�
||
�}tj�|��r0|s�t�
d|�q�tj�|��rt�
d|�q�tj�|��r&t�|�n
t�|�t�tj�
|	|
�|�q�qzdS)Nr,)r�c3s|]}|���VqdS�N)�
startswith)r}�s��ddirr3r4�	<genexpr>��z4InstallCommand._handle_target_dir.<locals>.<genexpr>���zKTarget directory %s already exists. Specify --upgrade to force replacement.z�Target directory %s already exists and is a link. pip will not automatically replace links, please remove if replacement is desired.)rrr��purelib�platlib�datar�r�r��listdirr��anyr��warning�islinkr��shutil�rmtree�remove�move)rfr9r�rE�lib_dir_list�scheme�purelib_dir�platlib_dir�data_dir�lib_dirr��target_item_dirr3r�r4r��sH


��
z!InstallCommand._handle_target_dir)r�r*cCs.z
t|�WSty(t�d�YdS0dS)NzwError while checking for conflicts. Please file an issue on pip's issue tracker: https://github.com/pypa/pip/issues/new)rr�r��	exception)rfr�r3r3r4r�s
�z#InstallCommand._determine_conflicts)�conflict_detailsr�r*c
Cs�|\}\}}|s|sdSg}|dkr0|�d�n|dks<J�|�d�|D]<}||d}||D]"}	dj|||	dd�}
|�|
�qbqJ|D]P}||d}||D]6\}}}
d	j|||
|||dkr�d
ndd�}
|�|
�q�q�t�d
�|��dS)N�legacyz�pip's legacy dependency resolver does not consider dependency conflicts when selecting packages. This behaviour is the source of the following dependency conflicts.z
2020-resolverz�pip's dependency resolver does not currently take into account all the packages that are installed. This behaviour is the source of the following dependency conflicts.rz@{name} {version} requires {requirement}, which is not installed.r�)r.r��requirementzg{name} {version} requires {requirement}, but {you} have {dep_name} {dep_version} which is incompatible.�youzyou'll)r.r�r��dep_name�dep_versionr��
)r�r�r��criticalr�)rfr�r��package_set�missing�conflicting�parts�project_namer��
dependencyr�r�r�r+r3r3r4r�sH������z$InstallCommand._warn_about_conflicts)�__name__�
__module__�__qualname__�__doc__�usagerhr
rr�str�intr�r#r6r�rrrr�r�r3r3r3r4r8=s  g�7�
�r8F)r�r�r�r�r�r*cCs td|||||d�}|j|jgS)Nr,)r�r�r�r�r�)rr�r�)r�r�r�r�r�r�r3r3r4r�Os�r�)r�r�r*cCstdd�tt||d��D��S)Ncss|]}t|�VqdSr�r)r}�dr3r3r4r�bs�z)site_packages_writable.<locals>.<genexpr>�r�r�)�all�setr�rr3r3r4�site_packages_writableas�r)r?rDr9rCrlr*cCs�|dur|st�d�dS|rF|r*td��t�r8td��t�d�dS|dusRJ�|sZ|rht�d�dStjs|t�d	�dSt||d
�r�t�d�dSt�d�dS)
aZDetermine whether to do a user install based on the input options.

    If use_user_site is False, no additional checks are done.
    If use_user_site is True, it is checked for compatibility with other
    options.
    If use_user_site is None, the default behaviour depends on the environment,
    which is provided by the other arguments.
    Nz$Non-user install by explicit requestFzVCan not combine '--user' and '--prefix' as they imply different installation locationszZCan not perform a '--user' install. User site-packages are not visible in this virtualenv.z User install by explicit requestTz3Non-user install due to --prefix or --target optionz4Non-user install because user site-packages disabledrz0Non-user install because site-packages writeablezMDefaulting to user installation because normal site-packages is not writeable)	r��debugrr%r�site�ENABLE_USER_SITEr�info)r?rDr9rCrlr3r3r4r�hs8
��



�r�)rPrir*cCs�ttttd�dd�}g}|D]0}|j}t|�}|r |�d�||���|��q |rzt|�}|rz|�d�||�����|s�dStd�d�	|����dS)	z�If any location-changing --install-option arguments were passed for
    requirements or on the command-line, then show a deprecation warning.
    )�option_namesr*cSsdd�|D�S)NcSsg|]}d�|�dd���qS)z--{}r�r�)r��replace)r}r.r3r3r4r��r�zSreject_location_related_install_options.<locals>.format_options.<locals>.<listcomp>r3)rr3r3r4�format_options�sz?reject_location_related_install_options.<locals>.format_optionsz{!r} from {}z{!r} from command lineNz�Location-changing options found in --install-option: {}. This is unsupported, use pip-level options like --user, --prefix, --root, and --target instead.z; )
rrrr[rr�r��keysrr�)rPrir�	offendersr�r[�location_optionsr3r3r4r��s0�����r�)r�r��using_user_siter*cCs�g}|�d�|s,|�d�|�t|��n
|�d�|dd7<|jtjkr�d}d}t�sz|sz|�|d|��g�n
|�|�|�d	�tr�|jtjkr�|j	r�t
|j	�d
kr�|�d�d�|���dS)
zrFormat an error message for an OSError

    It may occur anytime during the execution of the install command.
    z,Could not install packages due to an OSErrorz: �.r�r�z"Consider using the `--user` optionzCheck the permissionsz or z.
iz�HINT: This error might have occurred since this system does not have Windows Long Path support enabled. You can find information on how to enable this at https://pip.pypa.io/warnings/enable-long-paths
r,)
r�r�errno�EACCESr$�extend�lowerr�ENOENT�filename�lenr��strip)r�r�rr��user_option_part�permissions_partr3r3r4r��s>



��

�
����r�)FNNFN)NNNF)Orr�r�r�r
�optparserr�typingrrr�pip._vendor.packaging.utilsr�pip._internal.cacherZpip._internal.clir	Zpip._internal.cli.cmdoptionsr
�pip._internal.cli.req_commandrrr
�pip._internal.cli.status_codesrr�pip._internal.exceptionsrr�pip._internal.locationsr�pip._internal.metadatar�#pip._internal.models.format_controlr�pip._internal.operations.checkrr�pip._internal.reqrZpip._internal.req.req_installr�pip._internal.req.req_trackerr�pip._internal.utils.compatr�"pip._internal.utils.distutils_argsr�pip._internal.utils.filesystemr�pip._internal.utils.loggingr�pip._internal.utils.miscrr r!r"�pip._internal.utils.temp_dirr#�pip._internal.utils.virtualenvr$r%�pip._internal.wheel_builderr&r'r(rr�r7r8r6rr�rr�r�r�r�r3r3r3r4�<module>s�	��	��>�(�site-packages/pip/_internal/commands/__pycache__/uninstall.cpython-39.pyc000064400000006102150564745010022510 0ustar00a

��Re�
�@s�ddlZddlmZddlmZddlmZddlmZddl	m
Z
mZddlm
Z
ddlmZdd	lmZdd
lmZmZddlmZe�e�ZGdd
�d
ee
�ZdS)�N)�Values)�List)�canonicalize_name)�Command)�SessionCommandMixin�warn_if_run_as_root)�SUCCESS)�InstallationError)�parse_requirements)�install_req_from_line�#install_req_from_parsed_requirement)�(protect_pip_from_modification_on_windowsc@s8eZdZdZdZdd�dd�Zeeee	d�dd	�Z
dS)
�UninstallCommandaB
    Uninstall packages.

    pip is able to uninstall most installed packages. Known exceptions are:

    - Pure distutils packages installed with ``python setup.py install``, which
      leave behind no metadata to determine what files were installed.
    - Script wrappers installed by ``python setup.py develop``.
    zU
      %prog [options] <package> ...
      %prog [options] -r <requirements file> ...N)�returnc	CsD|jjddddgddd�|jjdd	d
ddd
�|j�d|j�dS)Nz-rz
--requirement�requirements�append�filezjUninstall all the packages listed in the given requirements file.  This option can be used multiple times.)�dest�action�default�metavar�helpz-yz--yes�yes�
store_truez2Don't ask for confirmation of uninstall deletions.)rrrr)�cmd_opts�
add_option�parser�insert_option_group)�self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/uninstall.py�add_options$s"��zUninstallCommand.add_options)�options�argsrc
Cs�|�|�}i}|D]4}t||jd�}|jr:||t|j�<qt�d|�q|jD]:}t|||d�D]&}t	||jd�}|jr`||t|j�<q`qN|s�t
d|j�d|j�d���td|vd�|��D]&}|j
|j|jd	kd
�}	|	r�|	��q�t�tS)N)�isolatedzSInvalid requirement: %r ignored - the uninstall command expects named requirements.)r"�sessionz*You must give at least one requirement to z (see "pip help z")�pip)�
modifying_pipr)�auto_confirm�verbose)�get_default_sessionr�
isolated_mode�namer�logger�warningrr
rr	r
�values�	uninstallr�	verbosity�commitrr)
rr"r#r%Zreqs_to_uninstallr,�req�filename�
parsed_reqZuninstall_pathsetrrr �run;sP
��
�
�
����
zUninstallCommand.run)�__name__�
__module__�__qualname__�__doc__�usager!rr�str�intr6rrrr rs
r)�logging�optparser�typingr�pip._vendor.packaging.utilsr�pip._internal.cli.base_commandr�pip._internal.cli.req_commandrr�pip._internal.cli.status_codesr�pip._internal.exceptionsr	�pip._internal.reqr
�pip._internal.req.constructorsrr�pip._internal.utils.miscr
�	getLoggerr7r-rrrrr �<module>s
site-packages/pip/_internal/commands/__pycache__/debug.cpython-39.pyc000064400000015045150564745010021573 0ustar00a

��Re��@s�ddlZddlZddlZddlZddlmZddlmZddlm	Z	m
Z
mZmZddl
ZddlmZddlmZddlmZddlmZdd	lmZdd
lmZddlmZddlmZdd
l m!Z!ddl"m#Z#ddl$m%Z%e�&e'�Z(e)e	dd�dd�Z*dd�dd�Z+e
e)e)fd�dd�Z,e)ed�dd�Z-e)ee)d�dd�Z.e
e)e)fdd�dd�Z/dd�d d!�Z0edd"�d#d$�Z1ee)d%�d&d'�Z2Gd(d)�d)e�Z3dS)*�N)�Values)�
ModuleType)�Any�Dict�List�Optional)�where)�parse)�__file__)�
cmdoptions)�Command)�make_target_python)�SUCCESS)�
Configuration)�get_environment)�
indent_log)�get_pip_version)�name�value�returncCst�d||�dS)Nz%s: %s)�logger�info)rr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/debug.py�
show_valuesr�rcCsFt�d�tjj}t��td|�Wd�n1s80YdS)Nzsys.implementation:r)rr�sys�implementationrrr)�implementation_namerrr�show_sys_implementations
rcCsdtj�tj�t�dd�}t|��"}dd�|��D�}Wd�n1sH0Ytdd�|D��S)N�_vendorz
vendor.txtcSs(g|] }d|vr|���dd�d�qS)�==� �r)�strip�split��.0�linerrr�
<listcomp>-sz)create_vendor_txt_map.<locals>.<listcomp>css|]}|�dd�VqdS)r!r#N)r%r&rrr�	<genexpr>2�z(create_vendor_txt_map.<locals>.<genexpr>)�os�path�join�dirname�pip_location�open�	readlines�dict)Zvendor_txt_path�f�linesrrr�create_vendor_txt_map%s�
�$r6)�module_namercCs:|��}|dkrd}td|��t�t�dd�ttj|�S)N�
setuptools�
pkg_resourceszpip._vendor.r)�level)�lower�
__import__�globals�locals�getattr�pipr )r7rrr�get_module_from_module_name5s
rAcCsHt|�}t|dd�}|sDttj�|j�g�}|�|�}|rDt|j	�}|S)N�__version__)
rAr?rr,r-r/r
�get_distribution�str�version)r7�modulerE�env�distrrr�get_vendor_version_from_module@s

rI)�vendor_txt_versionsrcCsZ|��D]L\}}d}t|�}|s*d}|}nt|�t|�krDd�|�}t�d|||�qdS)z{Log the actual version and print extra info if there is
    a conflict or if the actual version could not be imported.
    �zM (Unable to locate actual module version, using vendor.txt specified version)z5 (CONFLICT: vendor.txt suggests version should be {})z%s==%s%sN)�itemsrI�
parse_version�formatrr)rJr7Zexpected_versionZ
extra_messageZactual_versionrrr�show_actual_vendor_versionsNs���rOcCsBt�d�t�}t��t|�Wd�n1s40YdS)Nzvendored library versions:)rrr6rrO)rJrrr�show_vendor_versionscs
rP)�optionsrc	Cs�d}t|�}|��}|��}d}|r0d|�d�}d�t|�|�}t�|�|jdkrrt|�|krrd}|d|�}nd}t��B|D]}t�t	|��q�|r�d	j|d
�}t�|�Wd�n1s�0YdS)N�
rKz
 (target: �)zCompatible tags: {}{}r#TFz?...
[First {tag_limit} tags shown. Pass --verbose to show all.])�	tag_limit)
r
�get_tags�format_givenrN�lenrr�verboserrD)	rQrT�
target_python�tagsZformatted_target�suffix�msgZtags_limited�tagrrr�	show_tagsks,
��r^)�configrcsrt�}|��D]\}}|�|�d�d�q|s4dSgd���fdd�|D�}|sVdSd|vrh|�d�d�|�S)	N�.rz
Not specified)�install�wheel�downloadcsg|]}|�vr|�qSrr)r'r:�Zlevels_that_override_globalrrr)�sz"ca_bundle_info.<locals>.<listcomp>�globalz, )�setrL�addr%�remover.)r_Zlevels�key�_Zglobal_overriding_levelrrdr�ca_bundle_info�s
�
rkc@s<eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�ZdS)�DebugCommandz$
    Display debug information.
    z
      %prog <options>TNrcCs,t�|j�|j�d|j�|jj��dS)Nr)r�add_target_python_options�cmd_opts�parser�insert_option_groupr_�load)�selfrrr�add_options�szDebugCommand.add_options)rQ�argsrcCs�t�d�tdt��tdtj�tdtj�tdt���tdt���tdt	�
��tdtj�t�td	t
|jj��td
tj�d
��tdtj�d��tdt��td
tjj�t�t|�tS)Nz�This command is only meant for debugging. Do not use this with automation for parsing and getting these details, since the output and options of this command may change without notice.zpip versionzsys.versionzsys.executablezsys.getdefaultencodingzsys.getfilesystemencodingzlocale.getpreferredencodingzsys.platformz'cert' config value�REQUESTS_CA_BUNDLE�CURL_CA_BUNDLEzpip._vendor.certifi.where()zpip._vendor.DEBUNDLED)r�warningrrrrE�
executable�getdefaultencoding�getfilesystemencoding�locale�getpreferredencoding�platformrrkror_r,�environ�getrr@r �	DEBUNDLEDrPr^r)rrrQrtrrr�run�s,��zDebugCommand.run)�__name__�
__module__�__qualname__�__doc__�usage�ignore_require_venvrsrrrD�intr�rrrrrl�s
rl)4r{�loggingr,r�optparser�typesr�typingrrrr�pip._vendorr@�pip._vendor.certifirZpip._vendor.packaging.versionr	rMr
r0Zpip._internal.clir�pip._internal.cli.base_commandrZpip._internal.cli.cmdoptionsr
�pip._internal.cli.status_codesr�pip._internal.configurationr�pip._internal.metadatar�pip._internal.utils.loggingr�pip._internal.utils.miscr�	getLoggerr�rrDrrr6rArIrOrPr^rkrlrrrr�<module>s:
 site-packages/pip/_internal/commands/__pycache__/check.cpython-39.pyc000064400000003104150564745010021553 0ustar00a

��Re��@svddlZddlmZddlmZddlmZddlmZm	Z	ddl
mZmZddl
mZe�e�ZGdd	�d	e�ZdS)
�N)�Values)�List)�Command)�ERROR�SUCCESS)�check_package_set�!create_package_set_from_installed)�write_outputc@s*eZdZdZdZeeeed�dd�Z	dS)�CheckCommandz7Verify installed packages have compatible dependencies.z
      %prog [options])�options�args�returnc
	Cs�t�\}}t|�\}}|D].}||j}||D]}	td|||	d�q0q|D]4}||j}||D]\}
}}td||||
|�qdqN|s�|s�|r�tStd�tSdS)Nz*%s %s requires %s, which is not installed.rz-%s %s has requirement %s, but you have %s %s.zNo broken requirements found.)rr�versionr	rr)
�selfrr�package_setZ
parsing_probs�missing�conflicting�project_namer�
dependency�dep_name�dep_version�req�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/check.py�runs4

�
�	zCheckCommand.runN)
�__name__�
__module__�__qualname__�__doc__�usagerr�str�intrrrrrr
sr
)�logging�optparser�typingr�pip._internal.cli.base_commandr�pip._internal.cli.status_codesrr�pip._internal.operations.checkrr�pip._internal.utils.miscr	�	getLoggerr�loggerr
rrrr�<module>s
site-packages/pip/_internal/commands/__pycache__/wheel.cpython-39.pyc000064400000011402150564745010021602 0ustar00a

��Re�@s�ddlZddlZddlZddlmZddlmZddlmZddl	m
Z
ddlmZm
Z
ddlmZddlmZdd	lmZdd
lmZddlmZmZddlmZdd
lmZmZe�e�Z Gdd�de�Z!dS)�N)�Values)�List)�
WheelCache)�
cmdoptions)�RequirementCommand�with_cleanup)�SUCCESS)�CommandError)�InstallRequirement)�get_requirement_tracker)�
ensure_dir�normalize_path)�
TempDirectory)�build�should_build_for_wheel_commandc@s<eZdZdZdZdd�dd�Zeeee	e
d�dd	��ZdS)
�WheelCommanda�
    Build Wheel archives for your requirements and dependencies.

    Wheel is a built-package format, and offers the advantage of not
    recompiling your software during every install. For more details, see the
    wheel docs: https://wheel.readthedocs.io/en/latest/

    Requirements: setuptools>=0.8, and wheel.

    'pip wheel' uses the bdist_wheel setuptools extension from the wheel
    package to build individual wheels.

    z�
      %prog [options] <requirement specifier> ...
      %prog [options] -r <requirements file> ...
      %prog [options] [-e] <vcs project url> ...
      %prog [options] [-e] <local project path> ...
      %prog [options] <archive url/path> ...N)�returncCsv|jjddddtjdd�|j�t���|j�t���|j�t���|j�t���|j�t�	��|j�t�
��|j�t���|j�t���|j�t�
��|j�t���|j�t���|j�t���|j�t���|jjddd	d
dd�|j�t���|j�t���|jjd
d	d
dd�|j�t���t�tj|j�}|j�d|�|j�d|j�dS)Nz-wz--wheel-dir�	wheel_dir�dirzLBuild wheels into <dir>, where the default is the current working directory.)�dest�metavar�default�helpz--no-verify�	no_verify�
store_trueFz%Don't verify if built wheel is valid.)r�actionrrz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.)rrrr)�cmd_opts�
add_option�os�curdirr�	no_binary�only_binary�
prefer_binary�no_build_isolation�
use_pep517�
no_use_pep517�constraints�editable�requirements�src�ignore_requires_python�no_deps�progress_bar�
build_options�global_options�require_hashes�make_option_group�index_group�parser�insert_option_group)�self�
index_opts�r6��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/wheel.py�add_options+sV���
�zWheelCommand.add_options)�options�argsrcCs�t�|�|�|�}|�||�}t|j|j�}t|j�|_t	|j�|�
t��}t|j
ddd�}|�||||�}|j||||||jdd�}	|j|	||||j|jd�}
|�|�|
j|dd�}g}|j��D](}
|
jr�|	�|
�q�t|
�r�|�|
�q�t|||j|j�pg|j�pgd�\}}|D]~}
|
j�r:|
jj�s>J�|
j �sJJ�zt!�"|
j |j�Wn@t#�y�}z&t$�%d	|
j&|�|�|
�WYd}~n
d}~00�q$t'|�d
k�r�t(d��t)S)N�wheelT)�delete�kind�globally_managedF)�temp_build_dirr9�req_tracker�session�finder�download_dir�
use_user_site)�preparerrBr9�wheel_cacher*r$)�check_supported_wheels)rF�verifyr-r.z Building wheel for %s failed: %srz"Failed to build one or more wheels)*r�check_install_build_global�get_default_session�_build_package_finderr�	cache_dir�format_controlr
rr�
enter_contextrr�no_clean�get_requirements�make_requirement_preparer�
make_resolverr*r$�trace_basic_info�resolver(�values�is_wheel�save_linked_requirementr�appendrrr-r.�link�local_file_path�shutil�copy�OSError�logger�warning�name�lenr	r)r4r9r:rArBrFr@�	directory�reqsrE�resolver�requirement_set�
reqs_to_build�req�build_successes�build_failures�er6r6r7�runesx


��
�	


�
�$zWheelCommand.run)�__name__�
__module__�__qualname__�__doc__�usager8rrr�str�intrkr6r6r6r7rs
:r)"�loggingrr[�optparser�typingr�pip._internal.cacherZpip._internal.clir�pip._internal.cli.req_commandrr�pip._internal.cli.status_codesr�pip._internal.exceptionsr	Zpip._internal.req.req_installr
�pip._internal.req.req_trackerr�pip._internal.utils.miscrr
�pip._internal.utils.temp_dirr�pip._internal.wheel_builderrr�	getLoggerrlr^rr6r6r6r7�<module>s 
site-packages/pip/_internal/commands/__pycache__/completion.cpython-39.pyc000064400000006151150564745010022654 0ustar00a

��Re��@spddlZddlZddlmZddlmZddlmZddlm	Z	ddl
mZdZdd	d
d�Z
Gdd
�d
e�ZdS)�N)�Values)�List)�Command)�SUCCESS)�get_progzD
# pip {shell} completion start{script}# pip {shell} completion end
a
        _pip_completion()
        {{
            COMPREPLY=( $( COMP_WORDS="${{COMP_WORDS[*]}}" \
                           COMP_CWORD=$COMP_CWORD \
                           PIP_AUTO_COMPLETE=1 $1 2>/dev/null ) )
        }}
        complete -o default -F _pip_completion {prog}
    aM
        function _pip_completion {{
          local words cword
          read -Ac words
          read -cn cword
          reply=( $( COMP_WORDS="$words[*]" \
                     COMP_CWORD=$(( cword-1 )) \
                     PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null ))
        }}
        compctl -K _pip_completion {prog}
    au
        function __fish_complete_pip
            set -lx COMP_WORDS (commandline -o) ""
            set -lx COMP_CWORD ( \
                math (contains -i -- (commandline -t) $COMP_WORDS)-1 \
            )
            set -lx PIP_AUTO_COMPLETE 1
            string split \  -- (eval $COMP_WORDS[1])
        end
        complete -fa "(__fish_complete_pip)" -c {prog}
    )�bash�zsh�fishc@s8eZdZdZdZdd�dd�Zeeee	d�dd	�Z
dS)
�CompletionCommandz3A helper command to be used for command completion.TN)�returncCs\|jjddddddd�|jjdd	dd
ddd�|jjdd
ddddd�|j�d|j�dS)Nz--bashz-b�store_constr�shellzEmit completion code for bash)�action�const�dest�helpz--zshz-zrzEmit completion code for zshz--fishz-fr	zEmit completion code for fishr)�cmd_opts�
add_option�parser�insert_option_group)�self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/completion.py�add_options6s2���	zCompletionCommand.add_options)�options�argsrcCszt��}dd�t|�D�}|j|vrZt�t�|jd�jt�d��}t	t
j||jd��tStj
�d�d�|���tSdS)	z-Prints the completion code of the given shellcSsg|]}d|�qS)z--r)�.0r
rrr�
<listcomp>U�z)CompletionCommand.run.<locals>.<listcomp>�)�prog)�scriptr
zERROR: You must pass {}
z or N)�COMPLETION_SCRIPTS�keys�sortedr
�textwrap�dedent�get�formatr�print�BASE_COMPLETIONr�sys�stderr�write�join)rrrZshellsZ
shell_optionsr!rrr�runRs
��zCompletionCommand.run)�__name__�
__module__�__qualname__�__doc__�ignore_require_venvrrr�str�intr/rrrrr
1sr
)r+r%�optparser�typingr�pip._internal.cli.base_commandr�pip._internal.cli.status_codesr�pip._internal.utils.miscrr*r"r
rrrr�<module>s	�#site-packages/pip/_internal/commands/__pycache__/configuration.cpython-39.pyc000064400000020246150564745010023353 0ustar00a

��Re�"�@s�ddlZddlZddlZddlmZddlmZmZmZddl	m
Z
ddlmZm
Z
ddlmZmZmZmZddlmZddlmZdd	lmZmZe�e�ZGd
d�de
�ZdS)�N)�Values)�Any�List�Optional)�Command)�ERROR�SUCCESS)�
Configuration�Kind�get_configuration_files�kinds)�PipError)�
indent_log)�get_prog�write_outputc@s*eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�Zeee
ed�dd
�Zeee	dd�dd�Zeee	dd�dd�Zeee	dd�dd�Zeee	dd�dd�Zeee	dd�dd�Zedd�dd�Zdd�dd�Zeee	dd�dd�Zee	e	e
ed�d d!�Zdd�d"d#�Zee	d$�d%d&�ZdS)'�ConfigurationCommandaa
    Manage local and global configuration.

    Subcommands:

    - list: List the active configuration (or from the file specified)
    - edit: Edit the configuration file in an editor
    - get: Get the value associated with name
    - set: Set the name=value
    - unset: Unset the value associated with name
    - debug: List the configuration files and values defined under them

    If none of --user, --global and --site are passed, a virtual
    environment configuration file is used if one is active and the file
    exists. Otherwise, all modifications happen to the user file by
    default.
    Ta
        %prog [<file-option>] list
        %prog [<file-option>] [--editor <editor-path>] edit

        %prog [<file-option>] get name
        %prog [<file-option>] set name value
        %prog [<file-option>] unset name
        %prog [<file-option>] debug
    N)�returncCsl|jjdddddd�|jjdddd	d
d�|jjdddd	d
d�|jjdddd	dd�|j�d|j�dS)Nz--editor�editor�storez\Editor to use to edit the file. Uses VISUAL or EDITOR environment variables if not provided.)�dest�action�default�helpz--global�global_file�
store_trueFz+Use the system-wide configuration file onlyz--user�	user_filez$Use the user configuration file onlyz--site�	site_filez3Use the current environment configuration file onlyr)�cmd_opts�
add_option�parser�insert_option_group��self�r#��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/configuration.py�add_options4s:����z ConfigurationCommand.add_options)�options�argsrc
Cs|j|j|j|j|j|jd�}|r.|d|vrHt�dd�t	|���t
S|d}z|j||dvd�}Wn8ty�}z t�|j
d�t
WYd}~Sd}~00t|j|d�|_|j��z||||dd��Wn:t�y}z t�|j
d�t
WYd}~Sd}~00tS)	N)�list�edit�get�set�unset�debugrzNeed an action (%s) to perform.z, )r*r+r,r))�
need_value)�isolated�	load_only�)�list_values�open_in_editor�get_name�set_name_value�
unset_name�list_config_values�logger�error�join�sortedr�_determine_filer
r'r	�
isolated_mode�
configuration�loadr)r"r&r'�handlersrr0�er#r#r$�runZs>�
��
�
zConfigurationCommand.run)r&r.rcCs�dd�tj|jftj|jftj|jffD�}|s`|s8dStdd�t�tjD��rXtjStjSnt	|�dkrt|dSt
d��dS)NcSsg|]\}}|r|�qSr#r#)�.0�key�valuer#r#r$�
<listcomp>�s�z8ConfigurationCommand._determine_file.<locals>.<listcomp>css|]}tj�|�VqdS)N)�os�path�exists)rC�site_config_filer#r#r$�	<genexpr>�s�z7ConfigurationCommand._determine_file.<locals>.<genexpr>r1rzLNeed exactly one file to operate upon (--user, --site, --global) to perform.)r�USERr�GLOBALr�SITEr�anyr�lenr
)r"r&r.Zfile_optionsr#r#r$r<�s&


��

��z$ConfigurationCommand._determine_filecCs8|j|ddd�t|j���D]\}}td||�qdS)Nr(r��n�%s=%r)�_get_n_argsr;r>�itemsr�r"r&r'rDrEr#r#r$r2�sz ConfigurationCommand.list_valuescCs*|j|ddd�}|j�|�}td|�dS)Nz
get [name]r1rQz%s)rTr>�	get_valuerrVr#r#r$r4�szConfigurationCommand.get_namecCs.|j|ddd�\}}|j�||�|��dS)Nzset [name] [value]�rQ)rTr>�	set_value�_save_configurationrVr#r#r$r5�sz#ConfigurationCommand.set_name_valuecCs(|j|ddd�}|j�|�|��dS)Nzunset [name]r1rQ)rTr>�unset_valuerZ)r"r&r'rDr#r#r$r6�szConfigurationCommand.unset_namec
Cs�|j|ddd�|��t|j���D]h\}}td|�|D]P}t��6tj�	|�}td||�|rn|�
|�Wd�q<1s�0Yq<q&dS)z9List config key-value pairs across different config filesr-rrQ�%s:z%s, exists: %rN)rT�print_env_var_valuesr;r>�iter_config_filesrrrGrHrI�print_config_file_values)r"r&r'�variant�files�fnameZfile_existsr#r#r$r7�s
z'ConfigurationCommand.list_config_values)r`rc	CsP|j�|���D]:\}}t��td||�Wd�q1s@0YqdS)z.Get key-value pairs from the file of a variantz%s: %sN)r>�get_values_in_configrUrr)r"r`�namerEr#r#r$r_�sz-ConfigurationCommand.print_config_file_valuescCsftdd�t��Bt|j���D]"\}}d|����}td||�q Wd�n1sX0YdS)z5Get key-values pairs present as environment variablesr\�env_var�PIP_rSN)rrr;r>�get_environ_vars�upper)r"rDrErer#r#r$r]�s

z)ConfigurationCommand.print_env_var_valuesc
Csr|�|�}|j��}|dur$td��zt�||g�Wn6tjyl}ztd�|j���WYd}~n
d}~00dS)Nz%Could not determine appropriate file.z*Editor Subprocess exited with exit code {})	�_determine_editorr>�get_file_to_editr
�
subprocess�
check_call�CalledProcessError�format�
returncode)r"r&r'rrbrAr#r#r$r3�s


�z#ConfigurationCommand.open_in_editor)r'�examplerRrcCs<t|�|kr$d�|t�|�}t|��|dkr4|dS|SdS)zAHelper to make sure the command got the right number of argumentszJGot unexpected number of arguments, expected {}. (example: "{} config {}")r1rN)rPrnrr
)r"r'rprR�msgr#r#r$rT�s��z ConfigurationCommand._get_n_argscCs8z|j��Wn$ty2t�d�td��Yn0dS)Nz:Unable to save configuration. Please report this as a bug.zInternal Error.)r>�save�	Exceptionr8�	exceptionr
r!r#r#r$rZ�s�z(ConfigurationCommand._save_configuration)r&rcCsD|jdur|jSdtjvr$tjdSdtjvr8tjdStd��dS)NZVISUALZEDITORz"Could not determine editor to use.)rrG�environr
)r"r&r#r#r$ris




z&ConfigurationCommand._determine_editor)�__name__�
__module__�__qualname__�__doc__�ignore_require_venv�usager%rr�str�intrB�boolrr
r<r2r4r5r6r7r_r]r3rrTrZrir#r#r#r$rs"
&-r)�loggingrGrk�optparser�typingrrr�pip._internal.cli.base_commandr�pip._internal.cli.status_codesrr�pip._internal.configurationr	r
rr�pip._internal.exceptionsr
�pip._internal.utils.loggingr�pip._internal.utils.miscrr�	getLoggerrvr8rr#r#r#r$�<module>s
site-packages/pip/_internal/commands/__pycache__/search.cpython-39.pyc000064400000012360150564745010021747 0ustar00a

��ReA�@s|ddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZmZm
Z
mZddlmZddlmZddlmZddlmZmZdd	lmZdd
lmZddlmZddlm Z dd
l!m"Z"ddl#m$Z$er�ddl
m%Z%Gdd�de%�Z&e�'e(�Z)Gdd�dee�Z*e
ee+e+fe
dd�dd�Z,e+e+dd�dd�Z-d e
dee.ee.dd�dd�Z/e
e+e+d�dd�Z0dS)!�N)�OrderedDict)�Values)�
TYPE_CHECKING�Dict�List�Optional)�parse)�Command)�SessionCommandMixin)�NO_MATCHES_FOUND�SUCCESS)�CommandError)�get_default_environment)�PyPI)�PipXmlrpcTransport)�
indent_log)�write_output)�	TypedDictc@s*eZdZUeed<eed<eeed<dS)�TransformedHit�name�summary�versionsN)�__name__�
__module__�__qualname__�str�__annotations__r�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/search.pyrs
rc@s^eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�Zee	eeee	e	fd�dd
�Z
dS)�
SearchCommandz@Search for PyPI packages whose name or summary contains <query>.z
      %prog [options] <query>TN)�returncCs.|jjddddtjdd�|j�d|j�dS)Nz-iz--index�index�URLz3Base URL of Python Package Index (default %default))�dest�metavar�default�helpr)�cmd_opts�
add_optionr�pypi_url�parser�insert_option_group)�selfrrr�add_options)s�	zSearchCommand.add_options)�options�argsr cCsV|std��|}|�||�}t|�}d}tj��r>t��d}t||d�|rRt	St
S)Nz)Missing required argument (search query).r)�terminal_width)r
�search�transform_hits�sys�stdout�isatty�shutil�get_terminal_size�
print_resultsrr)r,r.r/�queryZ	pypi_hits�hitsr0rrr�run5s
zSearchCommand.run)r9r.r c

Cs�|j}|�|�}t||�}tj�||�}z|�||d�d�}WnBtjjy�}z&dj|j	|j
d�}	t|	��WYd}~n
d}~00t|t
�s�J�|S)N)rr�orz-XMLRPC request failed [code: {code}]
{string})�code�string)r!�get_default_sessionr�xmlrpc�client�ServerProxyr1�Fault�format�	faultCode�faultStringr
�
isinstance�list)
r,r9r.�	index_url�session�	transport�pypir:�fault�messagerrrr1Es

�zSearchCommand.search)rrr�__doc__�usage�ignore_require_venvr-rrr�intr;rr1rrrrr"sr)r:r cCs�t�}|D]n}|d}|d}|d}||��vrF|||gd�||<q
||d�|�|t||d�kr
|||d<q
t|���S)z�
    The list from pypi is really a list of versions. We want a list of
    packages with the list of versions stored inline. This converts the
    list from pypi into one we can use.
    rr�version)rrrr)r�keys�append�highest_versionrH�values)r:�packages�hitrrrSrrrr2Xs�r2)r�latestr cCs�t�}|�|�}|dur�t��T|j|kr8td|j�n,td|j�t|�jrZtd|�n
td|�Wd�n1sx0YdS)NzINSTALLED: %s (latest)z
INSTALLED: %sz=LATEST:    %s (pre-release; install with `pip install --pre`)z
LATEST:    %s)r�get_distributionrrSr�
parse_version�pre)rrZ�env�distrrr�print_dist_installation_infots


�r`)r:�name_column_widthr0r c	Cs�|sdS|dur&tdd�|D��d}|D]�}|d}|dp@d}t|�ddg��}|dur�||d	}|d
kr�t�||�}dd|d
�|�}|�d|�d�}	|	|��d|��}
zt|
�t||�Wq*ty�Yq*0q*dS)Nc	Ss.g|]&}t|d�tt|�ddg����qS)rr�-)�lenrV�get)�.0rYrrr�
<listcomp>�s�z!print_results.<locals>.<listcomp>�rr�rrb��
�
� �z (�)z - )	�maxrVrd�textwrap�wrap�joinrr`�UnicodeEncodeError)r:rar0rYrrrZZtarget_widthZ
summary_linesZname_latest�linerrrr8�s6����
r8)rr cCst|td�S)N)�key)ror\)rrrrrV�srV)NN)1�loggingr6r3rp�
xmlrpc.clientr@�collectionsr�optparser�typingrrrrZpip._vendor.packaging.versionrr\�pip._internal.cli.base_commandr	�pip._internal.cli.req_commandr
�pip._internal.cli.status_codesrr�pip._internal.exceptionsr
�pip._internal.metadatar�pip._internal.models.indexrZpip._internal.network.xmlrpcr�pip._internal.utils.loggingr�pip._internal.utils.miscrrr�	getLoggerr�loggerrrr2r`rRr8rVrrrr�<module>sB
6 ��&site-packages/pip/_internal/commands/__pycache__/index.cpython-39.pyc000064400000010716150564745010021614 0ustar00a

��Re��@s�ddlZddlmZddlmZmZmZmZmZddl	m
Z
mZddlm
Z
ddlmZddlmZmZddlmZdd	lmZmZmZdd
lmZddlmZddlmZdd
lm Z ddl!m"Z"ddl#m$Z$e�%e&�Z'Gdd�de�Z(dS)�N)�Values)�Any�Iterable�List�Optional�Union)�
LegacyVersion�Version)�
cmdoptions)�IndexGroupCommand)�ERROR�SUCCESS)�print_dist_installation_info)�CommandError�DistributionNotFound�PipError)�
LinkCollector)�
PackageFinder)�SelectionPreferences)�TargetPython)�
PipSession)�write_outputc@sneZdZdZdZdd�dd�Zeeee	d�dd	�Z
deeee
eeed
�dd�Zeeedd�d
d�ZdS)�IndexCommandz=
    Inspect information available from package indexes.
    z&
        %prog versions <package>
    N)�returncCs~t�|j�|j�t���|j�t���|j�t���|j�t���t�tj	|j
�}|j
�d|�|j
�d|j�dS)Nr)r
�add_target_python_options�cmd_opts�
add_option�ignore_requires_python�pre�	no_binary�only_binary�make_option_group�index_group�parser�insert_option_group)�self�
index_opts�r'��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/index.py�add_optionss�zIndexCommand.add_options)�options�argsrc
Cs�d|ji}t�d�|r$|d|vr>t�dd�t|���tS|d}z||||dd��Wn8ty�}z t�|jd�tWYd}~Sd}~00t	S)N�versionsztpip index is currently an experimental command. It may be removed/changed in a future release without prior warning.rzNeed an action (%s) to perform.�, �)
�get_available_package_versions�logger�warning�error�join�sortedrrr+r
)r%r*r+�handlers�action�er'r'r(�run/s$���zIndexCommand.run)r*�session�
target_pythonrrcCs.tj||d�}td|j|d�}tj|||d�S)zK
        Create a package finder appropriate to the index command.
        )r*F)�allow_yanked�allow_all_prereleasesr)�link_collector�selection_prefsr:)r�createrrr)r%r*r9r:rr=r>r'r'r(�_build_package_finderMs
��z"IndexCommand._build_package_finderc
Cs�t|�dkrtd��t�|�}|d}|�|���}|j||||jd�}dd�|�|�D�}|jsndd�|D�}t	|�}|s�t
d�|���d	d
�t|dd�D�}|d}	Wd�n1s�0Yt
d
�||	��t
d�d�|���t||	�dS)Nr.z(You need to specify exactly one argumentr)r*r9r:rcss|]}|jVqdS�N)�version)�.0�	candidater'r'r(�	<genexpr>usz>IndexCommand.get_available_package_versions.<locals>.<genexpr>css|]}|js|VqdSrA)�
is_prerelease)rCrBr'r'r(rE{sz%No matching distribution found for {}cSsg|]}t|��qSr')�str)rC�verr'r'r(�
<listcomp>��z?IndexCommand.get_available_package_versions.<locals>.<listcomp>T)�reversez{} ({})zAvailable versions: {}r-)�lenrr
�make_target_python�_build_sessionr@r�find_all_candidatesr�setr�formatr4rr3r)
r%r*r+r:�queryr9�finderr,Zformatted_versionsZlatestr'r'r(r/fs8
����&z+IndexCommand.get_available_package_versions)NN)�__name__�
__module__�__qualname__�__doc__�usager)rrrG�intr8rrr�boolrr@rr/r'r'r'r(rs"��r))�logging�optparser�typingrrrrrZpip._vendor.packaging.versionrr	Zpip._internal.clir
�pip._internal.cli.req_commandr�pip._internal.cli.status_codesrr
Zpip._internal.commands.searchr�pip._internal.exceptionsrrr�pip._internal.index.collectorr�"pip._internal.index.package_finderr�$pip._internal.models.selection_prefsr�"pip._internal.models.target_pythonr�pip._internal.network.sessionr�pip._internal.utils.miscr�	getLoggerrTr0rr'r'r'r(�<module>s 
site-packages/pip/_internal/commands/__pycache__/hash.cpython-39.pyc000064400000004172150564745010021427 0ustar00a

��Re��@s�ddlZddlZddlZddlmZddlmZddlmZddl	m
Z
mZddlm
Z
mZddlmZmZe�e�ZGdd	�d	e�Zeeed
�dd�ZdS)
�N)�Values)�List)�Command)�ERROR�SUCCESS)�
FAVORITE_HASH�
STRONG_HASHES)�read_chunks�write_outputc@s<eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�ZdS)�HashCommandz�
    Compute a hash of a local package archive.

    These can be used with --hash in a requirements file to do repeatable
    installs.
    z%prog [options] <file> ...TN)�returncCs:|jjdddtdtd�d�t��d�|j�d|j�dS)	Nz-az--algorithm�	algorithm�storez$The hash algorithm to use: one of {}z, )�dest�choices�action�default�helpr)�cmd_opts�
add_optionrr�format�join�parser�insert_option_group)�self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/hash.py�add_optionss��zHashCommand.add_options)�options�argsrcCs>|s|j�tj�tS|j}|D]}td||t||��q tS)Nz%s:
--hash=%s:%s)	r�print_usage�sys�stderrrr
r
�
_hash_of_filer)rrrr
�pathrrr�run(s�zHashCommand.run)�__name__�
__module__�__qualname__�__doc__�usage�ignore_require_venvrrr�str�intr%rrrrrs
r)r$r
rcCsTt|d��2}t�|�}t|�D]}|�|�qWd�n1sB0Y|��S)z!Return the hash digest of a file.�rbN)�open�hashlib�newr	�update�	hexdigest)r$r
�archive�hash�chunkrrrr#5s

*r#)r0�loggingr!�optparser�typingr�pip._internal.cli.base_commandr�pip._internal.cli.status_codesrr�pip._internal.utils.hashesrr�pip._internal.utils.miscr	r
�	getLoggerr&�loggerrr,r#rrrr�<module>s
&site-packages/pip/_internal/commands/__pycache__/list.cpython-39.pyc000064400000023712150564745010021460 0ustar00a

��Re�/�@sdddlZddlZddlmZddlmZmZmZmZm	Z	m
Z
mZddlm
Z
ddlmZddlmZddlmZddlmZdd	lmZdd
lmZddlmZmZddlmZdd
lm Z ddl!m"Z"ddl#m$Z$m%Z%ddl&m'Z'e�rddl(m)Z)Gdd�de�Z*e	e*Z+e�,e-�Z.Gdd�de�Z/dee
eee0ee0fd�dd�Z1dee0d�dd�Z2dS)�N)�Values)�
TYPE_CHECKING�Iterator�List�Optional�Sequence�Tuple�cast��canonicalize_name)�
cmdoptions)�IndexGroupCommand)�SUCCESS)�CommandError)�
LinkCollector)�
PackageFinder)�BaseDistribution�get_environment)�SelectionPreferences)�
PipSession)�stdlib_pkgs)�tabulate�write_output)�map_multithread)�DistributionVersionc@s"eZdZUdZeed<eed<dS)�_DistWithLatestInfoz�Give the distribution object a couple of extra fields.

        These will be populated during ``get_outdated()``. This is dirty but
        makes the rest of the code much cleaner.
        �latest_version�latest_filetypeN)�__name__�
__module__�__qualname__�__doc__r�__annotations__�str�r$r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/list.pyrs
rc@s�eZdZdZdZdZdd�dd�Zeee	d�d	d
�Z
eeee
d�dd
�Zdedd�dd�Zdedd�dd�Zdedd�dd�Zdeedd�dd�Zdedd�dd�Zeeeeedd�dd�ZdS)�ListCommandzt
    List installed packages, including editables.

    Packages are listed in a case-insensitive sorted order.
    Tz
      %prog [options]N)�returncCs*|jjdddddd�|jjddddd	d�|jjd
ddddd�|jjd
ddddd�|jjdddddd�|j�t���|jjddddd�|jjddddddd�|jjddddd �|jjd!d"d#d$d �|jjd%dd#d&d'd(�|j�t���t�tj|j�}|j�d)|�|j�d)|j�dS)*Nz-oz
--outdated�
store_trueFzList outdated packages)�action�default�helpz-uz
--uptodatezList uptodate packagesz-ez
--editablezList editable projects.z-lz--localzSIf in a virtualenv that has global access, do not list globally-installed packages.z--user�userz,Only output packages installed in user-site.)�destr)r*r+z--prezYInclude pre-release and development versions. By default, pip only finds stable versions.z--format�store�list_format�columns)r0�freeze�jsonzBSelect the output format among: columns (default), freeze, or json)r)r-r*�choicesr+z--not-required�not_requiredz>List packages that are not dependencies of installed packages.)r)r-r+z--exclude-editable�store_false�include_editablez%Exclude editable package from output.z--include-editablez%Include editable package from output.T)r)r-r+r*r)	�cmd_opts�
add_optionr�	list_path�list_exclude�make_option_group�index_group�parser�insert_option_group)�self�
index_optsr$r$r%�add_options3s�����
��
�	���zListCommand.add_options)�options�sessionr'cCs*tj||d�}td|jd�}tj||d�S)zK
        Create a package finder appropriate to this list command.
        )rBF)�allow_yanked�allow_all_prereleases)�link_collector�selection_prefs)r�creater�prer)r?rBrCrFrGr$r$r%�_build_package_finder�s��z!ListCommand._build_package_finder)rB�argsr'cCs�|jr|jrtd��t�|�tt�}|jrB|�dd�|jD��dd�t	|j
�j|j|j
|j|j|d�D�}|jr�|�||�}|jr�|�||�}n|jr�|�||�}|�||�tS)Nz5Options --outdated and --uptodate cannot be combined.css|]}t|�VqdS�Nr
)�.0�nr$r$r%�	<genexpr>��z"ListCommand.run.<locals>.<genexpr>cSsg|]}td|��qS)r)r	)rM�dr$r$r%�
<listcomp>�s�z#ListCommand.run.<locals>.<listcomp>)�
local_only�	user_only�editables_only�include_editables�skip)�outdatedZuptodaterr�check_list_path_option�setr�excludes�updater�path�iter_installed_distributions�localr,�editabler6r4�get_not_required�get_outdated�get_uptodate�output_package_listingr)r?rBrKrW�packagesr$r$r%�run�s.

��zListCommand.run�_ProcessedDists�rerBr'cCsdd�|�||�D�S)NcSsg|]}|j|jkr|�qSr$�r�version�rM�distr$r$r%rR�s�z,ListCommand.get_outdated.<locals>.<listcomp>��iter_packages_latest_infos�r?rerBr$r$r%rb�s
�zListCommand.get_outdatedcCsdd�|�||�D�S)NcSsg|]}|j|jkr|�qSr$rirkr$r$r%rR�s�z,ListCommand.get_uptodate.<locals>.<listcomp>rmror$r$r%rc�s
�zListCommand.get_uptodatecs$dd�|D��t�fdd�|D��S)NcSs(h|] }|��pdD]}t|j��qqS)r$)�iter_dependenciesr�name)rMrl�depr$r$r%�	<setcomp>�s�z/ListCommand.get_not_required.<locals>.<setcomp>csh|]}|j�vr|�qSr$��canonical_name)rM�pkg�Zdep_keysr$r%rs�rP)�listror$rwr%ra�s�	zListCommand.get_not_requiredrc#sr|����T}|��|��dtdd���fdd�}t||�D]}|dur<|Vq<Wd�n1sd0YdS)Nr)rlr'csl��|j�}�js dd�|D�}�j|jd�}|�|�}|durDdS|j}|jjrXd}nd}||_||_	|S)NcSsg|]}|jjs|�qSr$)rj�
is_prerelease)rM�	candidater$r$r%rR�s�zOListCommand.iter_packages_latest_infos.<locals>.latest_info.<locals>.<listcomp>)�project_name�wheel�sdist)
�find_all_candidatesrurI�make_candidate_evaluator�sort_best_candidaterj�link�is_wheelrr)rlZall_candidatesZ	evaluator�best_candidate�remote_version�typ��finderrBr$r%�latest_info�s$��
z;ListCommand.iter_packages_latest_infos.<locals>.latest_info)�_build_sessionrJrr)r?rerBrCr�rlr$r�r%rn�s�z&ListCommand.iter_packages_latest_infoscCs�t|dd�d�}|jdkr:|r:t||�\}}|�||�n^|jdkr�|D]4}|jdkrltd|j|j|j�qHtd|j|j�qHn|jd	kr�tt	||��dS)
NcSs|jSrLrt)rlr$r$r%�<lambda>
rPz4ListCommand.output_package_listing.<locals>.<lambda>)�keyr0r1�z%s==%s (%s)z%s==%sr2)
�sortedr/�format_for_columns�output_package_listing_columns�verboser�raw_namerj�location�format_for_json)r?rerB�data�headerrlr$r$r%rds �

�
z"ListCommand.output_package_listing)r�r�r'cCsbt|�dkr|�d|�t|�\}}t|�dkrL|�dd�tdd�|���|D]}t|�qPdS)Nrr�� cSsd|S)N�-r$)�xr$r$r%r�%rPz<ListCommand.output_package_listing_columns.<locals>.<lambda>)�len�insertr�join�mapr)r?r�r�Zpkg_strings�sizes�valr$r$r%r�sz*ListCommand.output_package_listing_columns)rrr r!�ignore_require_venv�usagerArrrrJrr#�intrfrbrcrarrnrdr�r$r$r$r%r&(s4V�%�
�
��'��r&rg)�pkgsrBr'cCs�ddg}|j}|r |�ddg�tdd�|D��}|r@|�d�|jdkrT|�d	�|jdkrh|�d
�g}|D]�}|jt|j�g}|r�|�t|j��|�|j	�|r�|�|j
p�d�|jdkr�|�|jp�d�|jdkr�|�|j�|�|�qp||fS)z_
    Convert the package data into something usable
    by output_package_listing_columns.
    �Package�VersionZLatest�Typecss|]}|jVqdSrL)r`)rMr�r$r$r%rO8rPz%format_for_columns.<locals>.<genexpr>zEditable project locationr�ZLocationZ	Installer�)
rX�extend�any�appendr�r�r#rjrr�editable_project_locationr��	installer)r�rBr�Zrunning_outdatedZ
has_editablesr�Zproj�rowr$r$r%r�+s2






r�rhcCs�g}|D]r}|jt|j�d�}|jdkr@|jp0d|d<|j|d<|jr^t|j�|d<|j|d<|j	}|rp||d<|�
|�qt�|�S)	N)rqrjr�r�r�r�rrr�)
r�r#rjr�r�r�rXrrr�r�r2�dumps)rerBr�rl�infor�r$r$r%r�Xs �


r�)3r2�logging�optparser�typingrrrrrrr	�pip._vendor.packaging.utilsrZpip._internal.clir�pip._internal.cli.req_commandr
�pip._internal.cli.status_codesr�pip._internal.exceptionsr�pip._internal.index.collectorr�"pip._internal.index.package_finderr�pip._internal.metadatarr�$pip._internal.models.selection_prefsr�pip._internal.network.sessionr�pip._internal.utils.compatr�pip._internal.utils.miscrrZpip._internal.utils.parallelrZpip._internal.metadata.baserrrg�	getLoggerr�loggerr&r#r�r�r$r$r$r%�<module>s8$

�-site-packages/pip/_internal/commands/__pycache__/__init__.cpython-39.pyc000064400000006020150564745010022235 0ustar00a

��Re��@s$UdZddlZddlmZddlmZmZmZddlm	Z	edd�Z
e
dd	d
�e
ddd
�e
ddd�e
ddd�e
ddd�e
ddd�e
ddd�e
ddd�e
d d!d"�e
d#d$d%�e
d&d'd(�e
d)d*d+�e
d,d-d.�e
d/d0d1�e
d2d3d4�e
d5d6d7�d8�Zeee
fe
d9<eee	d:�d;d<�Zeeed=�d>d?�ZdS)@z%
Package containing all pip commands
�N)�
namedtuple)�Any�Dict�Optional)�Command�CommandInfoz module_path, class_name, summaryzpip._internal.commands.install�InstallCommandzInstall packages.zpip._internal.commands.download�DownloadCommandzDownload packages.z pip._internal.commands.uninstall�UninstallCommandzUninstall packages.zpip._internal.commands.freeze�
FreezeCommandz1Output installed packages in requirements format.zpip._internal.commands.list�ListCommandzList installed packages.zpip._internal.commands.show�ShowCommandz*Show information about installed packages.zpip._internal.commands.check�CheckCommandz7Verify installed packages have compatible dependencies.z$pip._internal.commands.configuration�ConfigurationCommandz&Manage local and global configuration.zpip._internal.commands.search�
SearchCommandzSearch PyPI for packages.zpip._internal.commands.cache�CacheCommandz%Inspect and manage pip's wheel cache.zpip._internal.commands.index�IndexCommandz3Inspect information available from package indexes.zpip._internal.commands.wheel�WheelCommandz$Build wheels from your requirements.zpip._internal.commands.hash�HashCommandz#Compute hashes of package archives.z!pip._internal.commands.completion�CompletionCommandz-A helper command used for command completion.zpip._internal.commands.debug�DebugCommandz&Show information useful for debugging.zpip._internal.commands.help�HelpCommandzShow help for commands.)�install�download�	uninstall�freeze�list�show�check�config�search�cache�index�wheel�hash�
completion�debug�help�
commands_dict)�name�kwargs�returncKs:t|\}}}t�|�}t||�}|f||d�|��}|S)zF
    Create an instance of the Command class with the given name.
    )r)�summary)r(�	importlib�
import_module�getattr)r)r*�module_path�
class_namer,�module�
command_class�command�r5��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/__init__.py�create_commandhs


r7)r)r+cCs6ddlm}|��}||t���}|r.|dSdSdS)zCommand name auto-correct.r)�get_close_matchesN)�difflibr8�lowerr(�keys)r)r8�close_commandsr5r5r6�get_similar_commandstsr=)�__doc__r-�collectionsr�typingrrr�pip._internal.cli.base_commandrrr(�str�__annotations__r7r=r5r5r5r6�<module>s�

�����������������Tsite-packages/pip/_internal/commands/__pycache__/freeze.cpython-39.pyc000064400000005162150564745010021764 0ustar00a

��Re��@sxddlZddlmZddlmZddlmZddlmZddl	m
Z
ddlmZddl
mZhd	�ZGd
d�de�ZdS)�N)�Values)�List)�
cmdoptions)�Command)�SUCCESS)�freeze)�stdlib_pkgs>�
setuptools�wheelZ
distribute�pipc@s<eZdZdZdZdZdd�dd�Zeee	e
d�d	d
�ZdS)�
FreezeCommandzx
    Output installed packages in requirements format.

    packages are listed in a case-insensitive sorted order.
    z
      %prog [options])�ext://sys.stderrr
N)�returnc	Cs�|jjddddgddd�|jjdd	d
ddd
d�|jjdddddd�|j�t���|jjdddd�d�t��d�|jjddddd�|j�t���|j�	d|j�dS)Nz-rz
--requirement�requirements�append�filez}Use the order in the given requirements file and its comments when generating output. This option can be used multiple times.)�dest�action�default�metavar�helpz-lz--local�local�
store_trueFzUIf in a virtualenv that has global access, do not output globally-installed packages.)rrrrz--user�userz,Only output packages installed in user-site.z--all�
freeze_allz,Do not skip these packages in the output: {}z, )rrrz--exclude-editable�exclude_editablez%Exclude editable package from output.r)
�cmd_opts�
add_optionr�	list_path�format�join�DEV_PKGS�list_exclude�parser�insert_option_group)�self�r&��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/freeze.py�add_optionssR�
����	�zFreezeCommand.add_options)�options�argsrc	Csptt�}|js|�t�|jr*|�|j�t�|�t|j	|j
|j|j|j
||jd�D]}tj�|d�qVtS)N)�requirement�
local_only�	user_only�paths�isolated�skipr�
)�setrr�updater!�excludesr�check_list_path_optionrrrr�path�
isolated_moder�sys�stdout�writer)r%r)r*r0�liner&r&r'�runMs"

�
	zFreezeCommand.run)�__name__�
__module__�__qualname__�__doc__�usage�log_streamsr(rr�str�intr<r&r&r&r'rs
4r)r8�optparser�typingrZpip._internal.clir�pip._internal.cli.base_commandr�pip._internal.cli.status_codesrZpip._internal.operations.freezer�pip._internal.utils.compatrr!rr&r&r&r'�<module>ssite-packages/pip/_internal/commands/__pycache__/cache.cpython-39.pyc000064400000014107150564745010021546 0ustar00a

��Red�@s�ddlZddlZddlmZddlmZmZddlmm	m
Z
ddlmZddl
mZmZddlmZmZddlmZee�ZGdd	�d	e�ZdS)
�N)�Values)�Any�List)�Command)�ERROR�SUCCESS)�CommandError�PipError)�	getLoggerc@seZdZdZdZdZdd�dd�Zeee	e
d�d	d
�Zeeedd�dd�Z
eeedd�d
d�Zeeedd�dd�Zee	dd�dd�Zee	dd�dd�Zeeedd�dd�Zeeedd�dd�Zee	e	d�dd�Zeee	d�dd�Zee	ee	d �d!d"�ZdS)#�CacheCommandaw
    Inspect and manage pip's wheel cache.

    Subcommands:

    - dir: Show the cache directory.
    - info: Show information about the cache.
    - list: List filenames of packages stored in the cache.
    - remove: Remove one or more package from the cache.
    - purge: Remove all items from the cache.

    ``<pattern>`` can be a glob expression or a package name.
    Tz�
        %prog dir
        %prog info
        %prog list [<pattern>] [--format=[human, abspath]]
        %prog remove <pattern>
        %prog purge
    N)�returncCs,|jjddddddd�|j�d|j�dS)	Nz--format�store�list_format�human)r�abspathz:Select the output format among: human (default) or abspath)�action�dest�default�choices�helpr)�cmd_opts�
add_option�parser�insert_option_group)�self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/commands/cache.py�add_options's�	zCacheCommand.add_options)�options�argsrc
Cs�|j|j|j|j|jd�}|js.t�d�tS|r>|d|vrXt�dd�	t
|���tS|d}z||||dd��Wn8ty�}z t�|jd�tWYd}~Sd}~00t
S)N)�dir�info�list�remove�purgez<pip cache commands can not function since cache is disabled.rzNeed an action (%s) to perform.z, �)�
get_cache_dir�get_cache_info�list_cache_items�remove_cache_items�purge_cache�	cache_dir�logger�errorr�join�sortedr	rr)rrr�handlersr�errr�run4s,�
�zCacheCommand.runcCs|rtd��t�|j�dS)N�Too many arguments)rr,r!r+�rrrrrrr&TszCacheCommand.get_cache_dirc
Cs�|rtd��t|�|��}t|�|d��}|�|d�}|�|d�}t�|�}t�|�}t�d�j	||||||d��
�}	t�|	�dS)Nr3�*�http�wheelsa�
                    Package index page cache location: {http_cache_location}
                    Package index page cache size: {http_cache_size}
                    Number of HTTP files: {num_http_files}
                    Wheels location: {wheels_cache_location}
                    Wheels size: {wheels_cache_size}
                    Number of wheels: {package_count}
                )�http_cache_location�http_cache_size�num_http_files�wheels_cache_locationZ
package_count�wheels_cache_size)
r�len�_find_http_files�_find_wheels�
_cache_dir�
filesystem�format_directory_size�textwrap�dedent�format�stripr,r!)
rrrr:Znum_packagesr8r;r9r<�messagerrrr'Zs(

���zCacheCommand.get_cache_infocCsVt|�dkrtd��|r"|d}nd}|�||�}|jdkrH|�|�n
|�|�dS)Nr%r3rr5r)r=rr?r�format_for_human�format_for_abspath)rrr�pattern�filesrrrr(~s

zCacheCommand.list_cache_items)rKrcCsp|st�d�dSg}|D]2}tj�|�}t�|�}|�d|�d|�d��qt�d�t�d�t	|���dS)NzNothing cached.z - z (�)zCache contents:
�
)
r,r!�os�path�basenamerA�format_file_size�appendr.r/)rrK�results�filename�wheel�sizerrrrH�s


zCacheCommand.format_for_humancCs8|sdSg}|D]}|�|�qt�d�t|���dS)NrM)rRr,r!r.r/)rrKrSrTrrrrI�szCacheCommand.format_for_abspathcCs�t|�dkrtd��|s td��|�||d�}d}|ddkrP||�|�7}n|d�|d�7}|spt�|�|D]}t�|�t�	d|�qtt�
d	t|��dS)
Nr%r3zPlease provide a patternrzNo matching packagesr5z for pattern "{}"z
Removed %szFiles removed: %s)r=rr?r>rEr,�warningrN�unlink�verboser!)rrrrKZno_matching_msgrTrrrr)�s

zCacheCommand.remove_cache_itemscCs|rtd��|�|dg�S)Nr3r5)rr)r4rrrr*�szCacheCommand.purge_cache)r�subdirrcCstj�|j|�S)N)rNrOr.r+)rrrZrrrr@�szCacheCommand._cache_dir)rrcCs|�|d�}t�|d�S)Nr6r5�r@rA�
find_files)rrZhttp_dirrrrr>�szCacheCommand._find_http_files)rrJrcCs,|�|d�}|d|vrdnd}t�||�S)Nr7�-z*.whlz-*.whlr[)rrrJ�	wheel_dirrrrr?�szCacheCommand._find_wheels)�__name__�
__module__�__qualname__�__doc__�ignore_require_venv�usagerrr�str�intr2rr&r'r(rHrIr)r*r@r>r?rrrrrs
 $

r)rNrC�optparser�typingrr�pip._internal.utils.filesystem�	_internal�utilsrA�pip._internal.cli.base_commandr�pip._internal.cli.status_codesrr�pip._internal.exceptionsrr	�pip._internal.utils.loggingr
r_r,rrrrr�<module>ssite-packages/pip/_internal/commands/check.py000064400000003225150564745010015270 0ustar00import logging
from optparse import Values
from typing import List

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.operations.check import (
    check_package_set,
    create_package_set_from_installed,
)
from pip._internal.utils.misc import write_output

logger = logging.getLogger(__name__)


class CheckCommand(Command):
    """Verify installed packages have compatible dependencies."""

    usage = """
      %prog [options]"""

    def run(self, options: Values, args: List[str]) -> int:

        package_set, parsing_probs = create_package_set_from_installed()
        missing, conflicting = check_package_set(package_set)

        for project_name in missing:
            version = package_set[project_name].version
            for dependency in missing[project_name]:
                write_output(
                    "%s %s requires %s, which is not installed.",
                    project_name,
                    version,
                    dependency[0],
                )

        for project_name in conflicting:
            version = package_set[project_name].version
            for dep_name, dep_version, req in conflicting[project_name]:
                write_output(
                    "%s %s has requirement %s, but you have %s %s.",
                    project_name,
                    version,
                    req,
                    dep_name,
                    dep_version,
                )

        if missing or conflicting or parsing_probs:
            return ERROR
        else:
            write_output("No broken requirements found.")
            return SUCCESS
site-packages/pip/_internal/commands/uninstall.py000064400000006706150564745010016233 0ustar00import logging
from optparse import Values
from typing import List

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.cli.base_command import Command
from pip._internal.cli.req_command import SessionCommandMixin, warn_if_run_as_root
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.exceptions import InstallationError
from pip._internal.req import parse_requirements
from pip._internal.req.constructors import (
    install_req_from_line,
    install_req_from_parsed_requirement,
)
from pip._internal.utils.misc import protect_pip_from_modification_on_windows

logger = logging.getLogger(__name__)


class UninstallCommand(Command, SessionCommandMixin):
    """
    Uninstall packages.

    pip is able to uninstall most installed packages. Known exceptions are:

    - Pure distutils packages installed with ``python setup.py install``, which
      leave behind no metadata to determine what files were installed.
    - Script wrappers installed by ``python setup.py develop``.
    """

    usage = """
      %prog [options] <package> ...
      %prog [options] -r <requirements file> ..."""

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-r",
            "--requirement",
            dest="requirements",
            action="append",
            default=[],
            metavar="file",
            help=(
                "Uninstall all the packages listed in the given requirements "
                "file.  This option can be used multiple times."
            ),
        )
        self.cmd_opts.add_option(
            "-y",
            "--yes",
            dest="yes",
            action="store_true",
            help="Don't ask for confirmation of uninstall deletions.",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        session = self.get_default_session(options)

        reqs_to_uninstall = {}
        for name in args:
            req = install_req_from_line(
                name,
                isolated=options.isolated_mode,
            )
            if req.name:
                reqs_to_uninstall[canonicalize_name(req.name)] = req
            else:
                logger.warning(
                    "Invalid requirement: %r ignored -"
                    " the uninstall command expects named"
                    " requirements.",
                    name,
                )
        for filename in options.requirements:
            for parsed_req in parse_requirements(
                filename, options=options, session=session
            ):
                req = install_req_from_parsed_requirement(
                    parsed_req, isolated=options.isolated_mode
                )
                if req.name:
                    reqs_to_uninstall[canonicalize_name(req.name)] = req
        if not reqs_to_uninstall:
            raise InstallationError(
                f"You must give at least one requirement to {self.name} (see "
                f'"pip help {self.name}")'
            )

        protect_pip_from_modification_on_windows(
            modifying_pip="pip" in reqs_to_uninstall
        )

        for req in reqs_to_uninstall.values():
            uninstall_pathset = req.uninstall(
                auto_confirm=options.yes,
                verbose=self.verbosity > 0,
            )
            if uninstall_pathset:
                uninstall_pathset.commit()

        warn_if_run_as_root()
        return SUCCESS
site-packages/pip/_internal/commands/index.py000064400000011232150564745010015317 0ustar00import logging
from optparse import Values
from typing import Any, Iterable, List, Optional, Union

from pip._vendor.packaging.version import LegacyVersion, Version

from pip._internal.cli import cmdoptions
from pip._internal.cli.req_command import IndexGroupCommand
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.commands.search import print_dist_installation_info
from pip._internal.exceptions import CommandError, DistributionNotFound, PipError
from pip._internal.index.collector import LinkCollector
from pip._internal.index.package_finder import PackageFinder
from pip._internal.models.selection_prefs import SelectionPreferences
from pip._internal.models.target_python import TargetPython
from pip._internal.network.session import PipSession
from pip._internal.utils.misc import write_output

logger = logging.getLogger(__name__)


class IndexCommand(IndexGroupCommand):
    """
    Inspect information available from package indexes.
    """

    usage = """
        %prog versions <package>
    """

    def add_options(self) -> None:
        cmdoptions.add_target_python_options(self.cmd_opts)

        self.cmd_opts.add_option(cmdoptions.ignore_requires_python())
        self.cmd_opts.add_option(cmdoptions.pre())
        self.cmd_opts.add_option(cmdoptions.no_binary())
        self.cmd_opts.add_option(cmdoptions.only_binary())

        index_opts = cmdoptions.make_option_group(
            cmdoptions.index_group,
            self.parser,
        )

        self.parser.insert_option_group(0, index_opts)
        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        handlers = {
            "versions": self.get_available_package_versions,
        }

        logger.warning(
            "pip index is currently an experimental command. "
            "It may be removed/changed in a future release "
            "without prior warning."
        )

        # Determine action
        if not args or args[0] not in handlers:
            logger.error(
                "Need an action (%s) to perform.",
                ", ".join(sorted(handlers)),
            )
            return ERROR

        action = args[0]

        # Error handling happens here, not in the action-handlers.
        try:
            handlers[action](options, args[1:])
        except PipError as e:
            logger.error(e.args[0])
            return ERROR

        return SUCCESS

    def _build_package_finder(
        self,
        options: Values,
        session: PipSession,
        target_python: Optional[TargetPython] = None,
        ignore_requires_python: Optional[bool] = None,
    ) -> PackageFinder:
        """
        Create a package finder appropriate to the index command.
        """
        link_collector = LinkCollector.create(session, options=options)

        # Pass allow_yanked=False to ignore yanked versions.
        selection_prefs = SelectionPreferences(
            allow_yanked=False,
            allow_all_prereleases=options.pre,
            ignore_requires_python=ignore_requires_python,
        )

        return PackageFinder.create(
            link_collector=link_collector,
            selection_prefs=selection_prefs,
            target_python=target_python,
        )

    def get_available_package_versions(self, options: Values, args: List[Any]) -> None:
        if len(args) != 1:
            raise CommandError("You need to specify exactly one argument")

        target_python = cmdoptions.make_target_python(options)
        query = args[0]

        with self._build_session(options) as session:
            finder = self._build_package_finder(
                options=options,
                session=session,
                target_python=target_python,
                ignore_requires_python=options.ignore_requires_python,
            )

            versions: Iterable[Union[LegacyVersion, Version]] = (
                candidate.version for candidate in finder.find_all_candidates(query)
            )

            if not options.pre:
                # Remove prereleases
                versions = (
                    version for version in versions if not version.is_prerelease
                )
            versions = set(versions)

            if not versions:
                raise DistributionNotFound(
                    "No matching distribution found for {}".format(query)
                )

            formatted_versions = [str(ver) for ver in sorted(versions, reverse=True)]
            latest = formatted_versions[0]

        write_output("{} ({})".format(query, latest))
        write_output("Available versions: {}".format(", ".join(formatted_versions)))
        print_dist_installation_info(query, latest)
site-packages/pip/_internal/commands/help.py000064400000002154150564745010015143 0ustar00from optparse import Values
from typing import List

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.exceptions import CommandError


class HelpCommand(Command):
    """Show help for commands"""

    usage = """
      %prog <command>"""
    ignore_require_venv = True

    def run(self, options: Values, args: List[str]) -> int:
        from pip._internal.commands import (
            commands_dict,
            create_command,
            get_similar_commands,
        )

        try:
            # 'pip help' with no args is handled by pip.__init__.parseopt()
            cmd_name = args[0]  # the command we need help for
        except IndexError:
            return SUCCESS

        if cmd_name not in commands_dict:
            guess = get_similar_commands(cmd_name)

            msg = [f'unknown command "{cmd_name}"']
            if guess:
                msg.append(f'maybe you meant "{guess}"')

            raise CommandError(" - ".join(msg))

        command = create_command(cmd_name)
        command.parser.print_help()

        return SUCCESS
site-packages/pip/_internal/commands/wheel.py000064400000014030150564745010015313 0ustar00import logging
import os
import shutil
from optparse import Values
from typing import List

from pip._internal.cache import WheelCache
from pip._internal.cli import cmdoptions
from pip._internal.cli.req_command import RequirementCommand, with_cleanup
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.exceptions import CommandError
from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_tracker import get_requirement_tracker
from pip._internal.utils.misc import ensure_dir, normalize_path
from pip._internal.utils.temp_dir import TempDirectory
from pip._internal.wheel_builder import build, should_build_for_wheel_command

logger = logging.getLogger(__name__)


class WheelCommand(RequirementCommand):
    """
    Build Wheel archives for your requirements and dependencies.

    Wheel is a built-package format, and offers the advantage of not
    recompiling your software during every install. For more details, see the
    wheel docs: https://wheel.readthedocs.io/en/latest/

    Requirements: setuptools>=0.8, and wheel.

    'pip wheel' uses the bdist_wheel setuptools extension from the wheel
    package to build individual wheels.

    """

    usage = """
      %prog [options] <requirement specifier> ...
      %prog [options] -r <requirements file> ...
      %prog [options] [-e] <vcs project url> ...
      %prog [options] [-e] <local project path> ...
      %prog [options] <archive url/path> ..."""

    def add_options(self) -> None:

        self.cmd_opts.add_option(
            "-w",
            "--wheel-dir",
            dest="wheel_dir",
            metavar="dir",
            default=os.curdir,
            help=(
                "Build wheels into <dir>, where the default is the "
                "current working directory."
            ),
        )
        self.cmd_opts.add_option(cmdoptions.no_binary())
        self.cmd_opts.add_option(cmdoptions.only_binary())
        self.cmd_opts.add_option(cmdoptions.prefer_binary())
        self.cmd_opts.add_option(cmdoptions.no_build_isolation())
        self.cmd_opts.add_option(cmdoptions.use_pep517())
        self.cmd_opts.add_option(cmdoptions.no_use_pep517())
        self.cmd_opts.add_option(cmdoptions.constraints())
        self.cmd_opts.add_option(cmdoptions.editable())
        self.cmd_opts.add_option(cmdoptions.requirements())
        self.cmd_opts.add_option(cmdoptions.src())
        self.cmd_opts.add_option(cmdoptions.ignore_requires_python())
        self.cmd_opts.add_option(cmdoptions.no_deps())
        self.cmd_opts.add_option(cmdoptions.progress_bar())

        self.cmd_opts.add_option(
            "--no-verify",
            dest="no_verify",
            action="store_true",
            default=False,
            help="Don't verify if built wheel is valid.",
        )

        self.cmd_opts.add_option(cmdoptions.build_options())
        self.cmd_opts.add_option(cmdoptions.global_options())

        self.cmd_opts.add_option(
            "--pre",
            action="store_true",
            default=False,
            help=(
                "Include pre-release and development versions. By default, "
                "pip only finds stable versions."
            ),
        )

        self.cmd_opts.add_option(cmdoptions.require_hashes())

        index_opts = cmdoptions.make_option_group(
            cmdoptions.index_group,
            self.parser,
        )

        self.parser.insert_option_group(0, index_opts)
        self.parser.insert_option_group(0, self.cmd_opts)

    @with_cleanup
    def run(self, options: Values, args: List[str]) -> int:
        cmdoptions.check_install_build_global(options)

        session = self.get_default_session(options)

        finder = self._build_package_finder(options, session)
        wheel_cache = WheelCache(options.cache_dir, options.format_control)

        options.wheel_dir = normalize_path(options.wheel_dir)
        ensure_dir(options.wheel_dir)

        req_tracker = self.enter_context(get_requirement_tracker())

        directory = TempDirectory(
            delete=not options.no_clean,
            kind="wheel",
            globally_managed=True,
        )

        reqs = self.get_requirements(args, options, finder, session)

        preparer = self.make_requirement_preparer(
            temp_build_dir=directory,
            options=options,
            req_tracker=req_tracker,
            session=session,
            finder=finder,
            download_dir=options.wheel_dir,
            use_user_site=False,
        )

        resolver = self.make_resolver(
            preparer=preparer,
            finder=finder,
            options=options,
            wheel_cache=wheel_cache,
            ignore_requires_python=options.ignore_requires_python,
            use_pep517=options.use_pep517,
        )

        self.trace_basic_info(finder)

        requirement_set = resolver.resolve(reqs, check_supported_wheels=True)

        reqs_to_build: List[InstallRequirement] = []
        for req in requirement_set.requirements.values():
            if req.is_wheel:
                preparer.save_linked_requirement(req)
            elif should_build_for_wheel_command(req):
                reqs_to_build.append(req)

        # build wheels
        build_successes, build_failures = build(
            reqs_to_build,
            wheel_cache=wheel_cache,
            verify=(not options.no_verify),
            build_options=options.build_options or [],
            global_options=options.global_options or [],
        )
        for req in build_successes:
            assert req.link and req.link.is_wheel
            assert req.local_file_path
            # copy from cache to target directory
            try:
                shutil.copy(req.local_file_path, options.wheel_dir)
            except OSError as e:
                logger.warning(
                    "Building wheel for %s failed: %s",
                    req.name,
                    e,
                )
                build_failures.append(req)
        if len(build_failures) != 0:
            raise CommandError("Failed to build one or more wheels")

        return SUCCESS
site-packages/pip/_internal/commands/list.py000064400000027653150564745010015201 0ustar00import json
import logging
from optparse import Values
from typing import TYPE_CHECKING, Iterator, List, Optional, Sequence, Tuple, cast

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.cli import cmdoptions
from pip._internal.cli.req_command import IndexGroupCommand
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.exceptions import CommandError
from pip._internal.index.collector import LinkCollector
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import BaseDistribution, get_environment
from pip._internal.models.selection_prefs import SelectionPreferences
from pip._internal.network.session import PipSession
from pip._internal.utils.compat import stdlib_pkgs
from pip._internal.utils.misc import tabulate, write_output
from pip._internal.utils.parallel import map_multithread

if TYPE_CHECKING:
    from pip._internal.metadata.base import DistributionVersion

    class _DistWithLatestInfo(BaseDistribution):
        """Give the distribution object a couple of extra fields.

        These will be populated during ``get_outdated()``. This is dirty but
        makes the rest of the code much cleaner.
        """

        latest_version: DistributionVersion
        latest_filetype: str

    _ProcessedDists = Sequence[_DistWithLatestInfo]


logger = logging.getLogger(__name__)


class ListCommand(IndexGroupCommand):
    """
    List installed packages, including editables.

    Packages are listed in a case-insensitive sorted order.
    """

    ignore_require_venv = True
    usage = """
      %prog [options]"""

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-o",
            "--outdated",
            action="store_true",
            default=False,
            help="List outdated packages",
        )
        self.cmd_opts.add_option(
            "-u",
            "--uptodate",
            action="store_true",
            default=False,
            help="List uptodate packages",
        )
        self.cmd_opts.add_option(
            "-e",
            "--editable",
            action="store_true",
            default=False,
            help="List editable projects.",
        )
        self.cmd_opts.add_option(
            "-l",
            "--local",
            action="store_true",
            default=False,
            help=(
                "If in a virtualenv that has global access, do not list "
                "globally-installed packages."
            ),
        )
        self.cmd_opts.add_option(
            "--user",
            dest="user",
            action="store_true",
            default=False,
            help="Only output packages installed in user-site.",
        )
        self.cmd_opts.add_option(cmdoptions.list_path())
        self.cmd_opts.add_option(
            "--pre",
            action="store_true",
            default=False,
            help=(
                "Include pre-release and development versions. By default, "
                "pip only finds stable versions."
            ),
        )

        self.cmd_opts.add_option(
            "--format",
            action="store",
            dest="list_format",
            default="columns",
            choices=("columns", "freeze", "json"),
            help="Select the output format among: columns (default), freeze, or json",
        )

        self.cmd_opts.add_option(
            "--not-required",
            action="store_true",
            dest="not_required",
            help="List packages that are not dependencies of installed packages.",
        )

        self.cmd_opts.add_option(
            "--exclude-editable",
            action="store_false",
            dest="include_editable",
            help="Exclude editable package from output.",
        )
        self.cmd_opts.add_option(
            "--include-editable",
            action="store_true",
            dest="include_editable",
            help="Include editable package from output.",
            default=True,
        )
        self.cmd_opts.add_option(cmdoptions.list_exclude())
        index_opts = cmdoptions.make_option_group(cmdoptions.index_group, self.parser)

        self.parser.insert_option_group(0, index_opts)
        self.parser.insert_option_group(0, self.cmd_opts)

    def _build_package_finder(
        self, options: Values, session: PipSession
    ) -> PackageFinder:
        """
        Create a package finder appropriate to this list command.
        """
        link_collector = LinkCollector.create(session, options=options)

        # Pass allow_yanked=False to ignore yanked versions.
        selection_prefs = SelectionPreferences(
            allow_yanked=False,
            allow_all_prereleases=options.pre,
        )

        return PackageFinder.create(
            link_collector=link_collector,
            selection_prefs=selection_prefs,
        )

    def run(self, options: Values, args: List[str]) -> int:
        if options.outdated and options.uptodate:
            raise CommandError("Options --outdated and --uptodate cannot be combined.")

        cmdoptions.check_list_path_option(options)

        skip = set(stdlib_pkgs)
        if options.excludes:
            skip.update(canonicalize_name(n) for n in options.excludes)

        packages: "_ProcessedDists" = [
            cast("_DistWithLatestInfo", d)
            for d in get_environment(options.path).iter_installed_distributions(
                local_only=options.local,
                user_only=options.user,
                editables_only=options.editable,
                include_editables=options.include_editable,
                skip=skip,
            )
        ]

        # get_not_required must be called firstly in order to find and
        # filter out all dependencies correctly. Otherwise a package
        # can't be identified as requirement because some parent packages
        # could be filtered out before.
        if options.not_required:
            packages = self.get_not_required(packages, options)

        if options.outdated:
            packages = self.get_outdated(packages, options)
        elif options.uptodate:
            packages = self.get_uptodate(packages, options)

        self.output_package_listing(packages, options)
        return SUCCESS

    def get_outdated(
        self, packages: "_ProcessedDists", options: Values
    ) -> "_ProcessedDists":
        return [
            dist
            for dist in self.iter_packages_latest_infos(packages, options)
            if dist.latest_version > dist.version
        ]

    def get_uptodate(
        self, packages: "_ProcessedDists", options: Values
    ) -> "_ProcessedDists":
        return [
            dist
            for dist in self.iter_packages_latest_infos(packages, options)
            if dist.latest_version == dist.version
        ]

    def get_not_required(
        self, packages: "_ProcessedDists", options: Values
    ) -> "_ProcessedDists":
        dep_keys = {
            canonicalize_name(dep.name)
            for dist in packages
            for dep in (dist.iter_dependencies() or ())
        }

        # Create a set to remove duplicate packages, and cast it to a list
        # to keep the return type consistent with get_outdated and
        # get_uptodate
        return list({pkg for pkg in packages if pkg.canonical_name not in dep_keys})

    def iter_packages_latest_infos(
        self, packages: "_ProcessedDists", options: Values
    ) -> Iterator["_DistWithLatestInfo"]:
        with self._build_session(options) as session:
            finder = self._build_package_finder(options, session)

            def latest_info(
                dist: "_DistWithLatestInfo",
            ) -> Optional["_DistWithLatestInfo"]:
                all_candidates = finder.find_all_candidates(dist.canonical_name)
                if not options.pre:
                    # Remove prereleases
                    all_candidates = [
                        candidate
                        for candidate in all_candidates
                        if not candidate.version.is_prerelease
                    ]

                evaluator = finder.make_candidate_evaluator(
                    project_name=dist.canonical_name,
                )
                best_candidate = evaluator.sort_best_candidate(all_candidates)
                if best_candidate is None:
                    return None

                remote_version = best_candidate.version
                if best_candidate.link.is_wheel:
                    typ = "wheel"
                else:
                    typ = "sdist"
                dist.latest_version = remote_version
                dist.latest_filetype = typ
                return dist

            for dist in map_multithread(latest_info, packages):
                if dist is not None:
                    yield dist

    def output_package_listing(
        self, packages: "_ProcessedDists", options: Values
    ) -> None:
        packages = sorted(
            packages,
            key=lambda dist: dist.canonical_name,
        )
        if options.list_format == "columns" and packages:
            data, header = format_for_columns(packages, options)
            self.output_package_listing_columns(data, header)
        elif options.list_format == "freeze":
            for dist in packages:
                if options.verbose >= 1:
                    write_output(
                        "%s==%s (%s)", dist.raw_name, dist.version, dist.location
                    )
                else:
                    write_output("%s==%s", dist.raw_name, dist.version)
        elif options.list_format == "json":
            write_output(format_for_json(packages, options))

    def output_package_listing_columns(
        self, data: List[List[str]], header: List[str]
    ) -> None:
        # insert the header first: we need to know the size of column names
        if len(data) > 0:
            data.insert(0, header)

        pkg_strings, sizes = tabulate(data)

        # Create and add a separator.
        if len(data) > 0:
            pkg_strings.insert(1, " ".join(map(lambda x: "-" * x, sizes)))

        for val in pkg_strings:
            write_output(val)


def format_for_columns(
    pkgs: "_ProcessedDists", options: Values
) -> Tuple[List[List[str]], List[str]]:
    """
    Convert the package data into something usable
    by output_package_listing_columns.
    """
    header = ["Package", "Version"]

    running_outdated = options.outdated
    if running_outdated:
        header.extend(["Latest", "Type"])

    has_editables = any(x.editable for x in pkgs)
    if has_editables:
        header.append("Editable project location")

    if options.verbose >= 1:
        header.append("Location")
    if options.verbose >= 1:
        header.append("Installer")

    data = []
    for proj in pkgs:
        # if we're working on the 'outdated' list, separate out the
        # latest_version and type
        row = [proj.raw_name, str(proj.version)]

        if running_outdated:
            row.append(str(proj.latest_version))
            row.append(proj.latest_filetype)

        if has_editables:
            row.append(proj.editable_project_location or "")

        if options.verbose >= 1:
            row.append(proj.location or "")
        if options.verbose >= 1:
            row.append(proj.installer)

        data.append(row)

    return data, header


def format_for_json(packages: "_ProcessedDists", options: Values) -> str:
    data = []
    for dist in packages:
        info = {
            "name": dist.raw_name,
            "version": str(dist.version),
        }
        if options.verbose >= 1:
            info["location"] = dist.location or ""
            info["installer"] = dist.installer
        if options.outdated:
            info["latest_version"] = str(dist.latest_version)
            info["latest_filetype"] = dist.latest_filetype
        editable_project_location = dist.editable_project_location
        if editable_project_location:
            info["editable_project_location"] = editable_project_location
        data.append(info)
    return json.dumps(data)
site-packages/pip/_internal/commands/debug.py000064400000014745150564745010015312 0ustar00import locale
import logging
import os
import sys
from optparse import Values
from types import ModuleType
from typing import Any, Dict, List, Optional

import pip._vendor
from pip._vendor.certifi import where
from pip._vendor.packaging.version import parse as parse_version

from pip import __file__ as pip_location
from pip._internal.cli import cmdoptions
from pip._internal.cli.base_command import Command
from pip._internal.cli.cmdoptions import make_target_python
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.configuration import Configuration
from pip._internal.metadata import get_environment
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import get_pip_version

logger = logging.getLogger(__name__)


def show_value(name: str, value: Any) -> None:
    logger.info("%s: %s", name, value)


def show_sys_implementation() -> None:
    logger.info("sys.implementation:")
    implementation_name = sys.implementation.name
    with indent_log():
        show_value("name", implementation_name)


def create_vendor_txt_map() -> Dict[str, str]:
    vendor_txt_path = os.path.join(
        os.path.dirname(pip_location), "_vendor", "vendor.txt"
    )

    with open(vendor_txt_path) as f:
        # Purge non version specifying lines.
        # Also, remove any space prefix or suffixes (including comments).
        lines = [
            line.strip().split(" ", 1)[0] for line in f.readlines() if "==" in line
        ]

    # Transform into "module" -> version dict.
    return dict(line.split("==", 1) for line in lines)  # type: ignore


def get_module_from_module_name(module_name: str) -> ModuleType:
    # Module name can be uppercase in vendor.txt for some reason...
    module_name = module_name.lower()
    # PATCH: setuptools is actually only pkg_resources.
    if module_name == "setuptools":
        module_name = "pkg_resources"

    __import__(f"pip._vendor.{module_name}", globals(), locals(), level=0)
    return getattr(pip._vendor, module_name)


def get_vendor_version_from_module(module_name: str) -> Optional[str]:
    module = get_module_from_module_name(module_name)
    version = getattr(module, "__version__", None)

    if not version:
        # Try to find version in debundled module info.
        env = get_environment([os.path.dirname(module.__file__)])
        dist = env.get_distribution(module_name)
        if dist:
            version = str(dist.version)

    return version


def show_actual_vendor_versions(vendor_txt_versions: Dict[str, str]) -> None:
    """Log the actual version and print extra info if there is
    a conflict or if the actual version could not be imported.
    """
    for module_name, expected_version in vendor_txt_versions.items():
        extra_message = ""
        actual_version = get_vendor_version_from_module(module_name)
        if not actual_version:
            extra_message = (
                " (Unable to locate actual module version, using"
                " vendor.txt specified version)"
            )
            actual_version = expected_version
        elif parse_version(actual_version) != parse_version(expected_version):
            extra_message = (
                " (CONFLICT: vendor.txt suggests version should"
                " be {})".format(expected_version)
            )
        logger.info("%s==%s%s", module_name, actual_version, extra_message)


def show_vendor_versions() -> None:
    logger.info("vendored library versions:")

    vendor_txt_versions = create_vendor_txt_map()
    with indent_log():
        show_actual_vendor_versions(vendor_txt_versions)


def show_tags(options: Values) -> None:
    tag_limit = 10

    target_python = make_target_python(options)
    tags = target_python.get_tags()

    # Display the target options that were explicitly provided.
    formatted_target = target_python.format_given()
    suffix = ""
    if formatted_target:
        suffix = f" (target: {formatted_target})"

    msg = "Compatible tags: {}{}".format(len(tags), suffix)
    logger.info(msg)

    if options.verbose < 1 and len(tags) > tag_limit:
        tags_limited = True
        tags = tags[:tag_limit]
    else:
        tags_limited = False

    with indent_log():
        for tag in tags:
            logger.info(str(tag))

        if tags_limited:
            msg = (
                "...\n[First {tag_limit} tags shown. Pass --verbose to show all.]"
            ).format(tag_limit=tag_limit)
            logger.info(msg)


def ca_bundle_info(config: Configuration) -> str:
    levels = set()
    for key, _ in config.items():
        levels.add(key.split(".")[0])

    if not levels:
        return "Not specified"

    levels_that_override_global = ["install", "wheel", "download"]
    global_overriding_level = [
        level for level in levels if level in levels_that_override_global
    ]
    if not global_overriding_level:
        return "global"

    if "global" in levels:
        levels.remove("global")
    return ", ".join(levels)


class DebugCommand(Command):
    """
    Display debug information.
    """

    usage = """
      %prog <options>"""
    ignore_require_venv = True

    def add_options(self) -> None:
        cmdoptions.add_target_python_options(self.cmd_opts)
        self.parser.insert_option_group(0, self.cmd_opts)
        self.parser.config.load()

    def run(self, options: Values, args: List[str]) -> int:
        logger.warning(
            "This command is only meant for debugging. "
            "Do not use this with automation for parsing and getting these "
            "details, since the output and options of this command may "
            "change without notice."
        )
        show_value("pip version", get_pip_version())
        show_value("sys.version", sys.version)
        show_value("sys.executable", sys.executable)
        show_value("sys.getdefaultencoding", sys.getdefaultencoding())
        show_value("sys.getfilesystemencoding", sys.getfilesystemencoding())
        show_value(
            "locale.getpreferredencoding",
            locale.getpreferredencoding(),
        )
        show_value("sys.platform", sys.platform)
        show_sys_implementation()

        show_value("'cert' config value", ca_bundle_info(self.parser.config))
        show_value("REQUESTS_CA_BUNDLE", os.environ.get("REQUESTS_CA_BUNDLE"))
        show_value("CURL_CA_BUNDLE", os.environ.get("CURL_CA_BUNDLE"))
        show_value("pip._vendor.certifi.where()", where())
        show_value("pip._vendor.DEBUNDLED", pip._vendor.DEBUNDLED)

        show_vendor_versions()

        show_tags(options)

        return SUCCESS
site-packages/pip/_internal/commands/configuration.py000064400000021360150564745010017062 0ustar00import logging
import os
import subprocess
from optparse import Values
from typing import Any, List, Optional

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.configuration import (
    Configuration,
    Kind,
    get_configuration_files,
    kinds,
)
from pip._internal.exceptions import PipError
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import get_prog, write_output

logger = logging.getLogger(__name__)


class ConfigurationCommand(Command):
    """
    Manage local and global configuration.

    Subcommands:

    - list: List the active configuration (or from the file specified)
    - edit: Edit the configuration file in an editor
    - get: Get the value associated with name
    - set: Set the name=value
    - unset: Unset the value associated with name
    - debug: List the configuration files and values defined under them

    If none of --user, --global and --site are passed, a virtual
    environment configuration file is used if one is active and the file
    exists. Otherwise, all modifications happen to the user file by
    default.
    """

    ignore_require_venv = True
    usage = """
        %prog [<file-option>] list
        %prog [<file-option>] [--editor <editor-path>] edit

        %prog [<file-option>] get name
        %prog [<file-option>] set name value
        %prog [<file-option>] unset name
        %prog [<file-option>] debug
    """

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "--editor",
            dest="editor",
            action="store",
            default=None,
            help=(
                "Editor to use to edit the file. Uses VISUAL or EDITOR "
                "environment variables if not provided."
            ),
        )

        self.cmd_opts.add_option(
            "--global",
            dest="global_file",
            action="store_true",
            default=False,
            help="Use the system-wide configuration file only",
        )

        self.cmd_opts.add_option(
            "--user",
            dest="user_file",
            action="store_true",
            default=False,
            help="Use the user configuration file only",
        )

        self.cmd_opts.add_option(
            "--site",
            dest="site_file",
            action="store_true",
            default=False,
            help="Use the current environment configuration file only",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        handlers = {
            "list": self.list_values,
            "edit": self.open_in_editor,
            "get": self.get_name,
            "set": self.set_name_value,
            "unset": self.unset_name,
            "debug": self.list_config_values,
        }

        # Determine action
        if not args or args[0] not in handlers:
            logger.error(
                "Need an action (%s) to perform.",
                ", ".join(sorted(handlers)),
            )
            return ERROR

        action = args[0]

        # Determine which configuration files are to be loaded
        #    Depends on whether the command is modifying.
        try:
            load_only = self._determine_file(
                options, need_value=(action in ["get", "set", "unset", "edit"])
            )
        except PipError as e:
            logger.error(e.args[0])
            return ERROR

        # Load a new configuration
        self.configuration = Configuration(
            isolated=options.isolated_mode, load_only=load_only
        )
        self.configuration.load()

        # Error handling happens here, not in the action-handlers.
        try:
            handlers[action](options, args[1:])
        except PipError as e:
            logger.error(e.args[0])
            return ERROR

        return SUCCESS

    def _determine_file(self, options: Values, need_value: bool) -> Optional[Kind]:
        file_options = [
            key
            for key, value in (
                (kinds.USER, options.user_file),
                (kinds.GLOBAL, options.global_file),
                (kinds.SITE, options.site_file),
            )
            if value
        ]

        if not file_options:
            if not need_value:
                return None
            # Default to user, unless there's a site file.
            elif any(
                os.path.exists(site_config_file)
                for site_config_file in get_configuration_files()[kinds.SITE]
            ):
                return kinds.SITE
            else:
                return kinds.USER
        elif len(file_options) == 1:
            return file_options[0]

        raise PipError(
            "Need exactly one file to operate upon "
            "(--user, --site, --global) to perform."
        )

    def list_values(self, options: Values, args: List[str]) -> None:
        self._get_n_args(args, "list", n=0)

        for key, value in sorted(self.configuration.items()):
            write_output("%s=%r", key, value)

    def get_name(self, options: Values, args: List[str]) -> None:
        key = self._get_n_args(args, "get [name]", n=1)
        value = self.configuration.get_value(key)

        write_output("%s", value)

    def set_name_value(self, options: Values, args: List[str]) -> None:
        key, value = self._get_n_args(args, "set [name] [value]", n=2)
        self.configuration.set_value(key, value)

        self._save_configuration()

    def unset_name(self, options: Values, args: List[str]) -> None:
        key = self._get_n_args(args, "unset [name]", n=1)
        self.configuration.unset_value(key)

        self._save_configuration()

    def list_config_values(self, options: Values, args: List[str]) -> None:
        """List config key-value pairs across different config files"""
        self._get_n_args(args, "debug", n=0)

        self.print_env_var_values()
        # Iterate over config files and print if they exist, and the
        # key-value pairs present in them if they do
        for variant, files in sorted(self.configuration.iter_config_files()):
            write_output("%s:", variant)
            for fname in files:
                with indent_log():
                    file_exists = os.path.exists(fname)
                    write_output("%s, exists: %r", fname, file_exists)
                    if file_exists:
                        self.print_config_file_values(variant)

    def print_config_file_values(self, variant: Kind) -> None:
        """Get key-value pairs from the file of a variant"""
        for name, value in self.configuration.get_values_in_config(variant).items():
            with indent_log():
                write_output("%s: %s", name, value)

    def print_env_var_values(self) -> None:
        """Get key-values pairs present as environment variables"""
        write_output("%s:", "env_var")
        with indent_log():
            for key, value in sorted(self.configuration.get_environ_vars()):
                env_var = f"PIP_{key.upper()}"
                write_output("%s=%r", env_var, value)

    def open_in_editor(self, options: Values, args: List[str]) -> None:
        editor = self._determine_editor(options)

        fname = self.configuration.get_file_to_edit()
        if fname is None:
            raise PipError("Could not determine appropriate file.")

        try:
            subprocess.check_call([editor, fname])
        except subprocess.CalledProcessError as e:
            raise PipError(
                "Editor Subprocess exited with exit code {}".format(e.returncode)
            )

    def _get_n_args(self, args: List[str], example: str, n: int) -> Any:
        """Helper to make sure the command got the right number of arguments"""
        if len(args) != n:
            msg = (
                "Got unexpected number of arguments, expected {}. "
                '(example: "{} config {}")'
            ).format(n, get_prog(), example)
            raise PipError(msg)

        if n == 1:
            return args[0]
        else:
            return args

    def _save_configuration(self) -> None:
        # We successfully ran a modifying command. Need to save the
        # configuration.
        try:
            self.configuration.save()
        except Exception:
            logger.exception(
                "Unable to save configuration. Please report this as a bug."
            )
            raise PipError("Internal Error.")

    def _determine_editor(self, options: Values) -> str:
        if options.editor is not None:
            return options.editor
        elif "VISUAL" in os.environ:
            return os.environ["VISUAL"]
        elif "EDITOR" in os.environ:
            return os.environ["EDITOR"]
        else:
            raise PipError("Could not determine editor to use.")
site-packages/pip/_internal/commands/search.py000064400000013101150564745010015452 0ustar00import logging
import shutil
import sys
import textwrap
import xmlrpc.client
from collections import OrderedDict
from optparse import Values
from typing import TYPE_CHECKING, Dict, List, Optional

from pip._vendor.packaging.version import parse as parse_version

from pip._internal.cli.base_command import Command
from pip._internal.cli.req_command import SessionCommandMixin
from pip._internal.cli.status_codes import NO_MATCHES_FOUND, SUCCESS
from pip._internal.exceptions import CommandError
from pip._internal.metadata import get_default_environment
from pip._internal.models.index import PyPI
from pip._internal.network.xmlrpc import PipXmlrpcTransport
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import write_output

if TYPE_CHECKING:
    from typing import TypedDict

    class TransformedHit(TypedDict):
        name: str
        summary: str
        versions: List[str]


logger = logging.getLogger(__name__)


class SearchCommand(Command, SessionCommandMixin):
    """Search for PyPI packages whose name or summary contains <query>."""

    usage = """
      %prog [options] <query>"""
    ignore_require_venv = True

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-i",
            "--index",
            dest="index",
            metavar="URL",
            default=PyPI.pypi_url,
            help="Base URL of Python Package Index (default %default)",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        if not args:
            raise CommandError("Missing required argument (search query).")
        query = args
        pypi_hits = self.search(query, options)
        hits = transform_hits(pypi_hits)

        terminal_width = None
        if sys.stdout.isatty():
            terminal_width = shutil.get_terminal_size()[0]

        print_results(hits, terminal_width=terminal_width)
        if pypi_hits:
            return SUCCESS
        return NO_MATCHES_FOUND

    def search(self, query: List[str], options: Values) -> List[Dict[str, str]]:
        index_url = options.index

        session = self.get_default_session(options)

        transport = PipXmlrpcTransport(index_url, session)
        pypi = xmlrpc.client.ServerProxy(index_url, transport)
        try:
            hits = pypi.search({"name": query, "summary": query}, "or")
        except xmlrpc.client.Fault as fault:
            message = "XMLRPC request failed [code: {code}]\n{string}".format(
                code=fault.faultCode,
                string=fault.faultString,
            )
            raise CommandError(message)
        assert isinstance(hits, list)
        return hits


def transform_hits(hits: List[Dict[str, str]]) -> List["TransformedHit"]:
    """
    The list from pypi is really a list of versions. We want a list of
    packages with the list of versions stored inline. This converts the
    list from pypi into one we can use.
    """
    packages: Dict[str, "TransformedHit"] = OrderedDict()
    for hit in hits:
        name = hit["name"]
        summary = hit["summary"]
        version = hit["version"]

        if name not in packages.keys():
            packages[name] = {
                "name": name,
                "summary": summary,
                "versions": [version],
            }
        else:
            packages[name]["versions"].append(version)

            # if this is the highest version, replace summary and score
            if version == highest_version(packages[name]["versions"]):
                packages[name]["summary"] = summary

    return list(packages.values())


def print_dist_installation_info(name: str, latest: str) -> None:
    env = get_default_environment()
    dist = env.get_distribution(name)
    if dist is not None:
        with indent_log():
            if dist.version == latest:
                write_output("INSTALLED: %s (latest)", dist.version)
            else:
                write_output("INSTALLED: %s", dist.version)
                if parse_version(latest).pre:
                    write_output(
                        "LATEST:    %s (pre-release; install"
                        " with `pip install --pre`)",
                        latest,
                    )
                else:
                    write_output("LATEST:    %s", latest)


def print_results(
    hits: List["TransformedHit"],
    name_column_width: Optional[int] = None,
    terminal_width: Optional[int] = None,
) -> None:
    if not hits:
        return
    if name_column_width is None:
        name_column_width = (
            max(
                [
                    len(hit["name"]) + len(highest_version(hit.get("versions", ["-"])))
                    for hit in hits
                ]
            )
            + 4
        )

    for hit in hits:
        name = hit["name"]
        summary = hit["summary"] or ""
        latest = highest_version(hit.get("versions", ["-"]))
        if terminal_width is not None:
            target_width = terminal_width - name_column_width - 5
            if target_width > 10:
                # wrap and indent summary to fit terminal
                summary_lines = textwrap.wrap(summary, target_width)
                summary = ("\n" + " " * (name_column_width + 3)).join(summary_lines)

        name_latest = f"{name} ({latest})"
        line = f"{name_latest:{name_column_width}} - {summary}"
        try:
            write_output(line)
            print_dist_installation_info(name, latest)
        except UnicodeEncodeError:
            pass


def highest_version(versions: List[str]) -> str:
    return max(versions, key=parse_version)
site-packages/pip/_internal/commands/install.py000064400000066313150564745010015670 0ustar00import errno
import operator
import os
import shutil
import site
from optparse import SUPPRESS_HELP, Values
from typing import Iterable, List, Optional

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.cache import WheelCache
from pip._internal.cli import cmdoptions
from pip._internal.cli.cmdoptions import make_target_python
from pip._internal.cli.req_command import (
    RequirementCommand,
    warn_if_run_as_root,
    with_cleanup,
)
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.exceptions import CommandError, InstallationError
from pip._internal.locations import get_scheme
from pip._internal.metadata import get_environment
from pip._internal.models.format_control import FormatControl
from pip._internal.operations.check import ConflictDetails, check_install_conflicts
from pip._internal.req import install_given_reqs
from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_tracker import get_requirement_tracker
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.distutils_args import parse_distutils_args
from pip._internal.utils.filesystem import test_writable_dir
from pip._internal.utils.logging import getLogger
from pip._internal.utils.misc import (
    ensure_dir,
    get_pip_version,
    protect_pip_from_modification_on_windows,
    write_output,
)
from pip._internal.utils.temp_dir import TempDirectory
from pip._internal.utils.virtualenv import (
    running_under_virtualenv,
    virtualenv_no_global,
)
from pip._internal.wheel_builder import (
    BinaryAllowedPredicate,
    build,
    should_build_for_install_command,
)

logger = getLogger(__name__)


def get_check_binary_allowed(format_control: FormatControl) -> BinaryAllowedPredicate:
    def check_binary_allowed(req: InstallRequirement) -> bool:
        canonical_name = canonicalize_name(req.name or "")
        allowed_formats = format_control.get_allowed_formats(canonical_name)
        return "binary" in allowed_formats

    return check_binary_allowed


class InstallCommand(RequirementCommand):
    """
    Install packages from:

    - PyPI (and other indexes) using requirement specifiers.
    - VCS project urls.
    - Local project directories.
    - Local or remote source archives.

    pip also supports installing from "requirements files", which provide
    an easy way to specify a whole environment to be installed.
    """

    usage = """
      %prog [options] <requirement specifier> [package-index-options] ...
      %prog [options] -r <requirements file> [package-index-options] ...
      %prog [options] [-e] <vcs project url> ...
      %prog [options] [-e] <local project path> ...
      %prog [options] <archive url/path> ..."""

    def add_options(self) -> None:
        self.cmd_opts.add_option(cmdoptions.requirements())
        self.cmd_opts.add_option(cmdoptions.constraints())
        self.cmd_opts.add_option(cmdoptions.no_deps())
        self.cmd_opts.add_option(cmdoptions.pre())

        self.cmd_opts.add_option(cmdoptions.editable())
        self.cmd_opts.add_option(
            "-t",
            "--target",
            dest="target_dir",
            metavar="dir",
            default=None,
            help=(
                "Install packages into <dir>. "
                "By default this will not replace existing files/folders in "
                "<dir>. Use --upgrade to replace existing packages in <dir> "
                "with new versions."
            ),
        )
        cmdoptions.add_target_python_options(self.cmd_opts)

        self.cmd_opts.add_option(
            "--user",
            dest="use_user_site",
            action="store_true",
            help=(
                "Install to the Python user install directory for your "
                "platform. Typically ~/.local/, or %APPDATA%\\Python on "
                "Windows. (See the Python documentation for site.USER_BASE "
                "for full details.)"
            ),
        )
        self.cmd_opts.add_option(
            "--no-user",
            dest="use_user_site",
            action="store_false",
            help=SUPPRESS_HELP,
        )
        self.cmd_opts.add_option(
            "--root",
            dest="root_path",
            metavar="dir",
            default=None,
            help="Install everything relative to this alternate root directory.",
        )
        self.cmd_opts.add_option(
            "--prefix",
            dest="prefix_path",
            metavar="dir",
            default=None,
            help=(
                "Installation prefix where lib, bin and other top-level "
                "folders are placed"
            ),
        )

        self.cmd_opts.add_option(cmdoptions.src())

        self.cmd_opts.add_option(
            "-U",
            "--upgrade",
            dest="upgrade",
            action="store_true",
            help=(
                "Upgrade all specified packages to the newest available "
                "version. The handling of dependencies depends on the "
                "upgrade-strategy used."
            ),
        )

        self.cmd_opts.add_option(
            "--upgrade-strategy",
            dest="upgrade_strategy",
            default="only-if-needed",
            choices=["only-if-needed", "eager"],
            help=(
                "Determines how dependency upgrading should be handled "
                "[default: %default]. "
                '"eager" - dependencies are upgraded regardless of '
                "whether the currently installed version satisfies the "
                "requirements of the upgraded package(s). "
                '"only-if-needed" -  are upgraded only when they do not '
                "satisfy the requirements of the upgraded package(s)."
            ),
        )

        self.cmd_opts.add_option(
            "--force-reinstall",
            dest="force_reinstall",
            action="store_true",
            help="Reinstall all packages even if they are already up-to-date.",
        )

        self.cmd_opts.add_option(
            "-I",
            "--ignore-installed",
            dest="ignore_installed",
            action="store_true",
            help=(
                "Ignore the installed packages, overwriting them. "
                "This can break your system if the existing package "
                "is of a different version or was installed "
                "with a different package manager!"
            ),
        )

        self.cmd_opts.add_option(cmdoptions.ignore_requires_python())
        self.cmd_opts.add_option(cmdoptions.no_build_isolation())
        self.cmd_opts.add_option(cmdoptions.use_pep517())
        self.cmd_opts.add_option(cmdoptions.no_use_pep517())

        self.cmd_opts.add_option(cmdoptions.install_options())
        self.cmd_opts.add_option(cmdoptions.global_options())

        self.cmd_opts.add_option(
            "--compile",
            action="store_true",
            dest="compile",
            default=True,
            help="Compile Python source files to bytecode",
        )

        self.cmd_opts.add_option(
            "--no-compile",
            action="store_false",
            dest="compile",
            help="Do not compile Python source files to bytecode",
        )

        self.cmd_opts.add_option(
            "--no-warn-script-location",
            action="store_false",
            dest="warn_script_location",
            default=True,
            help="Do not warn when installing scripts outside PATH",
        )
        self.cmd_opts.add_option(
            "--no-warn-conflicts",
            action="store_false",
            dest="warn_about_conflicts",
            default=True,
            help="Do not warn about broken dependencies",
        )

        self.cmd_opts.add_option(cmdoptions.no_binary())
        self.cmd_opts.add_option(cmdoptions.only_binary())
        self.cmd_opts.add_option(cmdoptions.prefer_binary())
        self.cmd_opts.add_option(cmdoptions.require_hashes())
        self.cmd_opts.add_option(cmdoptions.progress_bar())

        index_opts = cmdoptions.make_option_group(
            cmdoptions.index_group,
            self.parser,
        )

        self.parser.insert_option_group(0, index_opts)
        self.parser.insert_option_group(0, self.cmd_opts)

    @with_cleanup
    def run(self, options: Values, args: List[str]) -> int:
        if options.use_user_site and options.target_dir is not None:
            raise CommandError("Can not combine '--user' and '--target'")

        cmdoptions.check_install_build_global(options)
        upgrade_strategy = "to-satisfy-only"
        if options.upgrade:
            upgrade_strategy = options.upgrade_strategy

        cmdoptions.check_dist_restriction(options, check_target=True)

        install_options = options.install_options or []

        logger.verbose("Using %s", get_pip_version())
        options.use_user_site = decide_user_install(
            options.use_user_site,
            prefix_path=options.prefix_path,
            target_dir=options.target_dir,
            root_path=options.root_path,
            isolated_mode=options.isolated_mode,
        )

        target_temp_dir: Optional[TempDirectory] = None
        target_temp_dir_path: Optional[str] = None
        if options.target_dir:
            options.ignore_installed = True
            options.target_dir = os.path.abspath(options.target_dir)
            if (
                # fmt: off
                os.path.exists(options.target_dir) and
                not os.path.isdir(options.target_dir)
                # fmt: on
            ):
                raise CommandError(
                    "Target path exists but is not a directory, will not continue."
                )

            # Create a target directory for using with the target option
            target_temp_dir = TempDirectory(kind="target")
            target_temp_dir_path = target_temp_dir.path
            self.enter_context(target_temp_dir)

        global_options = options.global_options or []

        session = self.get_default_session(options)

        target_python = make_target_python(options)
        finder = self._build_package_finder(
            options=options,
            session=session,
            target_python=target_python,
            ignore_requires_python=options.ignore_requires_python,
        )
        wheel_cache = WheelCache(options.cache_dir, options.format_control)

        req_tracker = self.enter_context(get_requirement_tracker())

        directory = TempDirectory(
            delete=not options.no_clean,
            kind="install",
            globally_managed=True,
        )

        try:
            reqs = self.get_requirements(args, options, finder, session)

            # Only when installing is it permitted to use PEP 660.
            # In other circumstances (pip wheel, pip download) we generate
            # regular (i.e. non editable) metadata and wheels.
            for req in reqs:
                req.permit_editable_wheels = True

            reject_location_related_install_options(reqs, options.install_options)

            preparer = self.make_requirement_preparer(
                temp_build_dir=directory,
                options=options,
                req_tracker=req_tracker,
                session=session,
                finder=finder,
                use_user_site=options.use_user_site,
            )
            resolver = self.make_resolver(
                preparer=preparer,
                finder=finder,
                options=options,
                wheel_cache=wheel_cache,
                use_user_site=options.use_user_site,
                ignore_installed=options.ignore_installed,
                ignore_requires_python=options.ignore_requires_python,
                force_reinstall=options.force_reinstall,
                upgrade_strategy=upgrade_strategy,
                use_pep517=options.use_pep517,
            )

            self.trace_basic_info(finder)

            requirement_set = resolver.resolve(
                reqs, check_supported_wheels=not options.target_dir
            )

            try:
                pip_req = requirement_set.get_requirement("pip")
            except KeyError:
                modifying_pip = False
            else:
                # If we're not replacing an already installed pip,
                # we're not modifying it.
                modifying_pip = pip_req.satisfied_by is None
            protect_pip_from_modification_on_windows(modifying_pip=modifying_pip)

            check_binary_allowed = get_check_binary_allowed(finder.format_control)

            reqs_to_build = [
                r
                for r in requirement_set.requirements.values()
                if should_build_for_install_command(r, check_binary_allowed)
            ]

            _, build_failures = build(
                reqs_to_build,
                wheel_cache=wheel_cache,
                verify=True,
                build_options=[],
                global_options=[],
            )

            # If we're using PEP 517, we cannot do a legacy setup.py install
            # so we fail here.
            pep517_build_failure_names: List[str] = [
                r.name for r in build_failures if r.use_pep517  # type: ignore
            ]
            if pep517_build_failure_names:
                raise InstallationError(
                    "Could not build wheels for {}, which is required to "
                    "install pyproject.toml-based projects".format(
                        ", ".join(pep517_build_failure_names)
                    )
                )

            # For now, we just warn about failures building legacy
            # requirements, as we'll fall through to a setup.py install for
            # those.
            for r in build_failures:
                if not r.use_pep517:
                    r.legacy_install_reason = 8368

            to_install = resolver.get_installation_order(requirement_set)

            # Check for conflicts in the package set we're installing.
            conflicts: Optional[ConflictDetails] = None
            should_warn_about_conflicts = (
                not options.ignore_dependencies and options.warn_about_conflicts
            )
            if should_warn_about_conflicts:
                conflicts = self._determine_conflicts(to_install)

            # Don't warn about script install locations if
            # --target or --prefix has been specified
            warn_script_location = options.warn_script_location
            if options.target_dir or options.prefix_path:
                warn_script_location = False

            installed = install_given_reqs(
                to_install,
                install_options,
                global_options,
                root=options.root_path,
                home=target_temp_dir_path,
                prefix=options.prefix_path,
                warn_script_location=warn_script_location,
                use_user_site=options.use_user_site,
                pycompile=options.compile,
            )

            lib_locations = get_lib_location_guesses(
                user=options.use_user_site,
                home=target_temp_dir_path,
                root=options.root_path,
                prefix=options.prefix_path,
                isolated=options.isolated_mode,
            )
            env = get_environment(lib_locations)

            installed.sort(key=operator.attrgetter("name"))
            items = []
            for result in installed:
                item = result.name
                try:
                    installed_dist = env.get_distribution(item)
                    if installed_dist is not None:
                        item = f"{item}-{installed_dist.version}"
                except Exception:
                    pass
                items.append(item)

            if conflicts is not None:
                self._warn_about_conflicts(
                    conflicts,
                    resolver_variant=self.determine_resolver_variant(options),
                )

            installed_desc = " ".join(items)
            if installed_desc:
                write_output(
                    "Successfully installed %s",
                    installed_desc,
                )
        except OSError as error:
            show_traceback = self.verbosity >= 1

            message = create_os_error_message(
                error,
                show_traceback,
                options.use_user_site,
            )
            logger.error(message, exc_info=show_traceback)  # noqa

            return ERROR

        if options.target_dir:
            assert target_temp_dir
            self._handle_target_dir(
                options.target_dir, target_temp_dir, options.upgrade
            )

        warn_if_run_as_root()
        return SUCCESS

    def _handle_target_dir(
        self, target_dir: str, target_temp_dir: TempDirectory, upgrade: bool
    ) -> None:
        ensure_dir(target_dir)

        # Checking both purelib and platlib directories for installed
        # packages to be moved to target directory
        lib_dir_list = []

        # Checking both purelib and platlib directories for installed
        # packages to be moved to target directory
        scheme = get_scheme("", home=target_temp_dir.path)
        purelib_dir = scheme.purelib
        platlib_dir = scheme.platlib
        data_dir = scheme.data

        if os.path.exists(purelib_dir):
            lib_dir_list.append(purelib_dir)
        if os.path.exists(platlib_dir) and platlib_dir != purelib_dir:
            lib_dir_list.append(platlib_dir)
        if os.path.exists(data_dir):
            lib_dir_list.append(data_dir)

        for lib_dir in lib_dir_list:
            for item in os.listdir(lib_dir):
                if lib_dir == data_dir:
                    ddir = os.path.join(data_dir, item)
                    if any(s.startswith(ddir) for s in lib_dir_list[:-1]):
                        continue
                target_item_dir = os.path.join(target_dir, item)
                if os.path.exists(target_item_dir):
                    if not upgrade:
                        logger.warning(
                            "Target directory %s already exists. Specify "
                            "--upgrade to force replacement.",
                            target_item_dir,
                        )
                        continue
                    if os.path.islink(target_item_dir):
                        logger.warning(
                            "Target directory %s already exists and is "
                            "a link. pip will not automatically replace "
                            "links, please remove if replacement is "
                            "desired.",
                            target_item_dir,
                        )
                        continue
                    if os.path.isdir(target_item_dir):
                        shutil.rmtree(target_item_dir)
                    else:
                        os.remove(target_item_dir)

                shutil.move(os.path.join(lib_dir, item), target_item_dir)

    def _determine_conflicts(
        self, to_install: List[InstallRequirement]
    ) -> Optional[ConflictDetails]:
        try:
            return check_install_conflicts(to_install)
        except Exception:
            logger.exception(
                "Error while checking for conflicts. Please file an issue on "
                "pip's issue tracker: https://github.com/pypa/pip/issues/new"
            )
            return None

    def _warn_about_conflicts(
        self, conflict_details: ConflictDetails, resolver_variant: str
    ) -> None:
        package_set, (missing, conflicting) = conflict_details
        if not missing and not conflicting:
            return

        parts: List[str] = []
        if resolver_variant == "legacy":
            parts.append(
                "pip's legacy dependency resolver does not consider dependency "
                "conflicts when selecting packages. This behaviour is the "
                "source of the following dependency conflicts."
            )
        else:
            assert resolver_variant == "2020-resolver"
            parts.append(
                "pip's dependency resolver does not currently take into account "
                "all the packages that are installed. This behaviour is the "
                "source of the following dependency conflicts."
            )

        # NOTE: There is some duplication here, with commands/check.py
        for project_name in missing:
            version = package_set[project_name][0]
            for dependency in missing[project_name]:
                message = (
                    "{name} {version} requires {requirement}, "
                    "which is not installed."
                ).format(
                    name=project_name,
                    version=version,
                    requirement=dependency[1],
                )
                parts.append(message)

        for project_name in conflicting:
            version = package_set[project_name][0]
            for dep_name, dep_version, req in conflicting[project_name]:
                message = (
                    "{name} {version} requires {requirement}, but {you} have "
                    "{dep_name} {dep_version} which is incompatible."
                ).format(
                    name=project_name,
                    version=version,
                    requirement=req,
                    dep_name=dep_name,
                    dep_version=dep_version,
                    you=("you" if resolver_variant == "2020-resolver" else "you'll"),
                )
                parts.append(message)

        logger.critical("\n".join(parts))


def get_lib_location_guesses(
    user: bool = False,
    home: Optional[str] = None,
    root: Optional[str] = None,
    isolated: bool = False,
    prefix: Optional[str] = None,
) -> List[str]:
    scheme = get_scheme(
        "",
        user=user,
        home=home,
        root=root,
        isolated=isolated,
        prefix=prefix,
    )
    return [scheme.purelib, scheme.platlib]


def site_packages_writable(root: Optional[str], isolated: bool) -> bool:
    return all(
        test_writable_dir(d)
        for d in set(get_lib_location_guesses(root=root, isolated=isolated))
    )


def decide_user_install(
    use_user_site: Optional[bool],
    prefix_path: Optional[str] = None,
    target_dir: Optional[str] = None,
    root_path: Optional[str] = None,
    isolated_mode: bool = False,
) -> bool:
    """Determine whether to do a user install based on the input options.

    If use_user_site is False, no additional checks are done.
    If use_user_site is True, it is checked for compatibility with other
    options.
    If use_user_site is None, the default behaviour depends on the environment,
    which is provided by the other arguments.
    """
    # In some cases (config from tox), use_user_site can be set to an integer
    # rather than a bool, which 'use_user_site is False' wouldn't catch.
    if (use_user_site is not None) and (not use_user_site):
        logger.debug("Non-user install by explicit request")
        return False

    if use_user_site:
        if prefix_path:
            raise CommandError(
                "Can not combine '--user' and '--prefix' as they imply "
                "different installation locations"
            )
        if virtualenv_no_global():
            raise InstallationError(
                "Can not perform a '--user' install. User site-packages "
                "are not visible in this virtualenv."
            )
        logger.debug("User install by explicit request")
        return True

    # If we are here, user installs have not been explicitly requested/avoided
    assert use_user_site is None

    # user install incompatible with --prefix/--target
    if prefix_path or target_dir:
        logger.debug("Non-user install due to --prefix or --target option")
        return False

    # If user installs are not enabled, choose a non-user install
    if not site.ENABLE_USER_SITE:
        logger.debug("Non-user install because user site-packages disabled")
        return False

    # If we have permission for a non-user install, do that,
    # otherwise do a user install.
    if site_packages_writable(root=root_path, isolated=isolated_mode):
        logger.debug("Non-user install because site-packages writeable")
        return False

    logger.info(
        "Defaulting to user installation because normal site-packages "
        "is not writeable"
    )
    return True


def reject_location_related_install_options(
    requirements: List[InstallRequirement], options: Optional[List[str]]
) -> None:
    """If any location-changing --install-option arguments were passed for
    requirements or on the command-line, then show a deprecation warning.
    """

    def format_options(option_names: Iterable[str]) -> List[str]:
        return ["--{}".format(name.replace("_", "-")) for name in option_names]

    offenders = []

    for requirement in requirements:
        install_options = requirement.install_options
        location_options = parse_distutils_args(install_options)
        if location_options:
            offenders.append(
                "{!r} from {}".format(
                    format_options(location_options.keys()), requirement
                )
            )

    if options:
        location_options = parse_distutils_args(options)
        if location_options:
            offenders.append(
                "{!r} from command line".format(format_options(location_options.keys()))
            )

    if not offenders:
        return

    raise CommandError(
        "Location-changing options found in --install-option: {}."
        " This is unsupported, use pip-level options like --user,"
        " --prefix, --root, and --target instead.".format("; ".join(offenders))
    )


def create_os_error_message(
    error: OSError, show_traceback: bool, using_user_site: bool
) -> str:
    """Format an error message for an OSError

    It may occur anytime during the execution of the install command.
    """
    parts = []

    # Mention the error if we are not going to show a traceback
    parts.append("Could not install packages due to an OSError")
    if not show_traceback:
        parts.append(": ")
        parts.append(str(error))
    else:
        parts.append(".")

    # Spilt the error indication from a helper message (if any)
    parts[-1] += "\n"

    # Suggest useful actions to the user:
    #  (1) using user site-packages or (2) verifying the permissions
    if error.errno == errno.EACCES:
        user_option_part = "Consider using the `--user` option"
        permissions_part = "Check the permissions"

        if not running_under_virtualenv() and not using_user_site:
            parts.extend(
                [
                    user_option_part,
                    " or ",
                    permissions_part.lower(),
                ]
            )
        else:
            parts.append(permissions_part)
        parts.append(".\n")

    # Suggest the user to enable Long Paths if path length is
    # more than 260
    if (
        WINDOWS
        and error.errno == errno.ENOENT
        and error.filename
        and len(error.filename) > 260
    ):
        parts.append(
            "HINT: This error might have occurred since "
            "this system does not have Windows Long Path "
            "support enabled. You can find information on "
            "how to enable this at "
            "https://pip.pypa.io/warnings/enable-long-paths\n"
        )

    return "".join(parts).strip() + "\n"
site-packages/pip/_internal/commands/show.py000064400000017600150564745010015175 0ustar00import csv
import logging
import pathlib
from optparse import Values
from typing import Iterator, List, NamedTuple, Optional, Tuple

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.metadata import BaseDistribution, get_default_environment
from pip._internal.utils.misc import write_output

logger = logging.getLogger(__name__)


class ShowCommand(Command):
    """
    Show information about one or more installed packages.

    The output is in RFC-compliant mail header format.
    """

    usage = """
      %prog [options] <package> ..."""
    ignore_require_venv = True

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-f",
            "--files",
            dest="files",
            action="store_true",
            default=False,
            help="Show the full list of installed files for each package.",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        if not args:
            logger.warning("ERROR: Please provide a package name or names.")
            return ERROR
        query = args

        results = search_packages_info(query)
        if not print_results(
            results, list_files=options.files, verbose=options.verbose
        ):
            return ERROR
        return SUCCESS


class _PackageInfo(NamedTuple):
    name: str
    version: str
    location: str
    requires: List[str]
    required_by: List[str]
    installer: str
    metadata_version: str
    classifiers: List[str]
    summary: str
    homepage: str
    author: str
    author_email: str
    license: str
    entry_points: List[str]
    files: Optional[List[str]]


def _convert_legacy_entry(entry: Tuple[str, ...], info: Tuple[str, ...]) -> str:
    """Convert a legacy installed-files.txt path into modern RECORD path.

    The legacy format stores paths relative to the info directory, while the
    modern format stores paths relative to the package root, e.g. the
    site-packages directory.

    :param entry: Path parts of the installed-files.txt entry.
    :param info: Path parts of the egg-info directory relative to package root.
    :returns: The converted entry.

    For best compatibility with symlinks, this does not use ``abspath()`` or
    ``Path.resolve()``, but tries to work with path parts:

    1. While ``entry`` starts with ``..``, remove the equal amounts of parts
       from ``info``; if ``info`` is empty, start appending ``..`` instead.
    2. Join the two directly.
    """
    while entry and entry[0] == "..":
        if not info or info[-1] == "..":
            info += ("..",)
        else:
            info = info[:-1]
        entry = entry[1:]
    return str(pathlib.Path(*info, *entry))


def search_packages_info(query: List[str]) -> Iterator[_PackageInfo]:
    """
    Gather details from installed distributions. Print distribution name,
    version, location, and installed files. Installed files requires a
    pip generated 'installed-files.txt' in the distributions '.egg-info'
    directory.
    """
    env = get_default_environment()

    installed = {dist.canonical_name: dist for dist in env.iter_distributions()}
    query_names = [canonicalize_name(name) for name in query]
    missing = sorted(
        [name for name, pkg in zip(query, query_names) if pkg not in installed]
    )
    if missing:
        logger.warning("Package(s) not found: %s", ", ".join(missing))

    def _get_requiring_packages(current_dist: BaseDistribution) -> Iterator[str]:
        return (
            dist.metadata["Name"] or "UNKNOWN"
            for dist in installed.values()
            if current_dist.canonical_name
            in {canonicalize_name(d.name) for d in dist.iter_dependencies()}
        )

    def _files_from_record(dist: BaseDistribution) -> Optional[Iterator[str]]:
        try:
            text = dist.read_text("RECORD")
        except FileNotFoundError:
            return None
        # This extra Path-str cast normalizes entries.
        return (str(pathlib.Path(row[0])) for row in csv.reader(text.splitlines()))

    def _files_from_legacy(dist: BaseDistribution) -> Optional[Iterator[str]]:
        try:
            text = dist.read_text("installed-files.txt")
        except FileNotFoundError:
            return None
        paths = (p for p in text.splitlines(keepends=False) if p)
        root = dist.location
        info = dist.info_directory
        if root is None or info is None:
            return paths
        try:
            info_rel = pathlib.Path(info).relative_to(root)
        except ValueError:  # info is not relative to root.
            return paths
        if not info_rel.parts:  # info *is* root.
            return paths
        return (
            _convert_legacy_entry(pathlib.Path(p).parts, info_rel.parts) for p in paths
        )

    for query_name in query_names:
        try:
            dist = installed[query_name]
        except KeyError:
            continue

        requires = sorted((req.name for req in dist.iter_dependencies()), key=str.lower)
        required_by = sorted(_get_requiring_packages(dist), key=str.lower)

        try:
            entry_points_text = dist.read_text("entry_points.txt")
            entry_points = entry_points_text.splitlines(keepends=False)
        except FileNotFoundError:
            entry_points = []

        files_iter = _files_from_record(dist) or _files_from_legacy(dist)
        if files_iter is None:
            files: Optional[List[str]] = None
        else:
            files = sorted(files_iter)

        metadata = dist.metadata

        yield _PackageInfo(
            name=dist.raw_name,
            version=str(dist.version),
            location=dist.location or "",
            requires=requires,
            required_by=required_by,
            installer=dist.installer,
            metadata_version=dist.metadata_version or "",
            classifiers=metadata.get_all("Classifier", []),
            summary=metadata.get("Summary", ""),
            homepage=metadata.get("Home-page", ""),
            author=metadata.get("Author", ""),
            author_email=metadata.get("Author-email", ""),
            license=metadata.get("License", ""),
            entry_points=entry_points,
            files=files,
        )


def print_results(
    distributions: Iterator[_PackageInfo],
    list_files: bool,
    verbose: bool,
) -> bool:
    """
    Print the information from installed distributions found.
    """
    results_printed = False
    for i, dist in enumerate(distributions):
        results_printed = True
        if i > 0:
            write_output("---")

        write_output("Name: %s", dist.name)
        write_output("Version: %s", dist.version)
        write_output("Summary: %s", dist.summary)
        write_output("Home-page: %s", dist.homepage)
        write_output("Author: %s", dist.author)
        write_output("Author-email: %s", dist.author_email)
        write_output("License: %s", dist.license)
        write_output("Location: %s", dist.location)
        write_output("Requires: %s", ", ".join(dist.requires))
        write_output("Required-by: %s", ", ".join(dist.required_by))

        if verbose:
            write_output("Metadata-Version: %s", dist.metadata_version)
            write_output("Installer: %s", dist.installer)
            write_output("Classifiers:")
            for classifier in dist.classifiers:
                write_output("  %s", classifier)
            write_output("Entry-points:")
            for entry in dist.entry_points:
                write_output("  %s", entry.strip())
        if list_files:
            write_output("Files:")
            if dist.files is None:
                write_output("Cannot locate RECORD or installed-files.txt")
            else:
                for line in dist.files:
                    write_output("  %s", line.strip())
    return results_printed
site-packages/pip/_internal/commands/hash.py000064400000003247150564745010015142 0ustar00import hashlib
import logging
import sys
from optparse import Values
from typing import List

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.utils.hashes import FAVORITE_HASH, STRONG_HASHES
from pip._internal.utils.misc import read_chunks, write_output

logger = logging.getLogger(__name__)


class HashCommand(Command):
    """
    Compute a hash of a local package archive.

    These can be used with --hash in a requirements file to do repeatable
    installs.
    """

    usage = "%prog [options] <file> ..."
    ignore_require_venv = True

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-a",
            "--algorithm",
            dest="algorithm",
            choices=STRONG_HASHES,
            action="store",
            default=FAVORITE_HASH,
            help="The hash algorithm to use: one of {}".format(
                ", ".join(STRONG_HASHES)
            ),
        )
        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        if not args:
            self.parser.print_usage(sys.stderr)
            return ERROR

        algorithm = options.algorithm
        for path in args:
            write_output(
                "%s:\n--hash=%s:%s", path, algorithm, _hash_of_file(path, algorithm)
            )
        return SUCCESS


def _hash_of_file(path: str, algorithm: str) -> str:
    """Return the hash digest of a file."""
    with open(path, "rb") as archive:
        hash = hashlib.new(algorithm)
        for chunk in read_chunks(archive):
            hash.update(chunk)
    return hash.hexdigest()
site-packages/pip/_internal/commands/__init__.py000064400000007230150564745010015752 0ustar00"""
Package containing all pip commands
"""

import importlib
from collections import namedtuple
from typing import Any, Dict, Optional

from pip._internal.cli.base_command import Command

CommandInfo = namedtuple("CommandInfo", "module_path, class_name, summary")

# This dictionary does a bunch of heavy lifting for help output:
# - Enables avoiding additional (costly) imports for presenting `--help`.
# - The ordering matters for help display.
#
# Even though the module path starts with the same "pip._internal.commands"
# prefix, the full path makes testing easier (specifically when modifying
# `commands_dict` in test setup / teardown).
commands_dict: Dict[str, CommandInfo] = {
    "install": CommandInfo(
        "pip._internal.commands.install",
        "InstallCommand",
        "Install packages.",
    ),
    "download": CommandInfo(
        "pip._internal.commands.download",
        "DownloadCommand",
        "Download packages.",
    ),
    "uninstall": CommandInfo(
        "pip._internal.commands.uninstall",
        "UninstallCommand",
        "Uninstall packages.",
    ),
    "freeze": CommandInfo(
        "pip._internal.commands.freeze",
        "FreezeCommand",
        "Output installed packages in requirements format.",
    ),
    "list": CommandInfo(
        "pip._internal.commands.list",
        "ListCommand",
        "List installed packages.",
    ),
    "show": CommandInfo(
        "pip._internal.commands.show",
        "ShowCommand",
        "Show information about installed packages.",
    ),
    "check": CommandInfo(
        "pip._internal.commands.check",
        "CheckCommand",
        "Verify installed packages have compatible dependencies.",
    ),
    "config": CommandInfo(
        "pip._internal.commands.configuration",
        "ConfigurationCommand",
        "Manage local and global configuration.",
    ),
    "search": CommandInfo(
        "pip._internal.commands.search",
        "SearchCommand",
        "Search PyPI for packages.",
    ),
    "cache": CommandInfo(
        "pip._internal.commands.cache",
        "CacheCommand",
        "Inspect and manage pip's wheel cache.",
    ),
    "index": CommandInfo(
        "pip._internal.commands.index",
        "IndexCommand",
        "Inspect information available from package indexes.",
    ),
    "wheel": CommandInfo(
        "pip._internal.commands.wheel",
        "WheelCommand",
        "Build wheels from your requirements.",
    ),
    "hash": CommandInfo(
        "pip._internal.commands.hash",
        "HashCommand",
        "Compute hashes of package archives.",
    ),
    "completion": CommandInfo(
        "pip._internal.commands.completion",
        "CompletionCommand",
        "A helper command used for command completion.",
    ),
    "debug": CommandInfo(
        "pip._internal.commands.debug",
        "DebugCommand",
        "Show information useful for debugging.",
    ),
    "help": CommandInfo(
        "pip._internal.commands.help",
        "HelpCommand",
        "Show help for commands.",
    ),
}


def create_command(name: str, **kwargs: Any) -> Command:
    """
    Create an instance of the Command class with the given name.
    """
    module_path, class_name, summary = commands_dict[name]
    module = importlib.import_module(module_path)
    command_class = getattr(module, class_name)
    command = command_class(name=name, summary=summary, **kwargs)

    return command


def get_similar_commands(name: str) -> Optional[str]:
    """Command name auto-correct."""
    from difflib import get_close_matches

    name = name.lower()

    close_commands = get_close_matches(name, commands_dict.keys())

    if close_commands:
        return close_commands[0]
    else:
        return None
site-packages/pip/_internal/commands/download.py000064400000011450150564745010016021 0ustar00import logging
import os
from optparse import Values
from typing import List

from pip._internal.cli import cmdoptions
from pip._internal.cli.cmdoptions import make_target_python
from pip._internal.cli.req_command import RequirementCommand, with_cleanup
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.req.req_tracker import get_requirement_tracker
from pip._internal.utils.misc import ensure_dir, normalize_path, write_output
from pip._internal.utils.temp_dir import TempDirectory

logger = logging.getLogger(__name__)


class DownloadCommand(RequirementCommand):
    """
    Download packages from:

    - PyPI (and other indexes) using requirement specifiers.
    - VCS project urls.
    - Local project directories.
    - Local or remote source archives.

    pip also supports downloading from "requirements files", which provide
    an easy way to specify a whole environment to be downloaded.
    """

    usage = """
      %prog [options] <requirement specifier> [package-index-options] ...
      %prog [options] -r <requirements file> [package-index-options] ...
      %prog [options] <vcs project url> ...
      %prog [options] <local project path> ...
      %prog [options] <archive url/path> ..."""

    def add_options(self) -> None:
        self.cmd_opts.add_option(cmdoptions.constraints())
        self.cmd_opts.add_option(cmdoptions.requirements())
        self.cmd_opts.add_option(cmdoptions.no_deps())
        self.cmd_opts.add_option(cmdoptions.global_options())
        self.cmd_opts.add_option(cmdoptions.no_binary())
        self.cmd_opts.add_option(cmdoptions.only_binary())
        self.cmd_opts.add_option(cmdoptions.prefer_binary())
        self.cmd_opts.add_option(cmdoptions.src())
        self.cmd_opts.add_option(cmdoptions.pre())
        self.cmd_opts.add_option(cmdoptions.require_hashes())
        self.cmd_opts.add_option(cmdoptions.progress_bar())
        self.cmd_opts.add_option(cmdoptions.no_build_isolation())
        self.cmd_opts.add_option(cmdoptions.use_pep517())
        self.cmd_opts.add_option(cmdoptions.no_use_pep517())
        self.cmd_opts.add_option(cmdoptions.ignore_requires_python())

        self.cmd_opts.add_option(
            "-d",
            "--dest",
            "--destination-dir",
            "--destination-directory",
            dest="download_dir",
            metavar="dir",
            default=os.curdir,
            help="Download packages into <dir>.",
        )

        cmdoptions.add_target_python_options(self.cmd_opts)

        index_opts = cmdoptions.make_option_group(
            cmdoptions.index_group,
            self.parser,
        )

        self.parser.insert_option_group(0, index_opts)
        self.parser.insert_option_group(0, self.cmd_opts)

    @with_cleanup
    def run(self, options: Values, args: List[str]) -> int:

        options.ignore_installed = True
        # editable doesn't really make sense for `pip download`, but the bowels
        # of the RequirementSet code require that property.
        options.editables = []

        cmdoptions.check_dist_restriction(options)

        options.download_dir = normalize_path(options.download_dir)
        ensure_dir(options.download_dir)

        session = self.get_default_session(options)

        target_python = make_target_python(options)
        finder = self._build_package_finder(
            options=options,
            session=session,
            target_python=target_python,
            ignore_requires_python=options.ignore_requires_python,
        )

        req_tracker = self.enter_context(get_requirement_tracker())

        directory = TempDirectory(
            delete=not options.no_clean,
            kind="download",
            globally_managed=True,
        )

        reqs = self.get_requirements(args, options, finder, session)

        preparer = self.make_requirement_preparer(
            temp_build_dir=directory,
            options=options,
            req_tracker=req_tracker,
            session=session,
            finder=finder,
            download_dir=options.download_dir,
            use_user_site=False,
        )

        resolver = self.make_resolver(
            preparer=preparer,
            finder=finder,
            options=options,
            ignore_requires_python=options.ignore_requires_python,
            py_version_info=options.python_version,
        )

        self.trace_basic_info(finder)

        requirement_set = resolver.resolve(reqs, check_supported_wheels=True)

        downloaded: List[str] = []
        for req in requirement_set.requirements.values():
            if req.satisfied_by is None:
                assert req.name is not None
                preparer.save_linked_requirement(req)
                downloaded.append(req.name)
        if downloaded:
            write_output("Successfully downloaded %s", " ".join(downloaded))

        return SUCCESS
site-packages/pip/_internal/commands/completion.py000064400000005616150564745010016372 0ustar00import sys
import textwrap
from optparse import Values
from typing import List

from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.utils.misc import get_prog

BASE_COMPLETION = """
# pip {shell} completion start{script}# pip {shell} completion end
"""

COMPLETION_SCRIPTS = {
    "bash": """
        _pip_completion()
        {{
            COMPREPLY=( $( COMP_WORDS="${{COMP_WORDS[*]}}" \\
                           COMP_CWORD=$COMP_CWORD \\
                           PIP_AUTO_COMPLETE=1 $1 2>/dev/null ) )
        }}
        complete -o default -F _pip_completion {prog}
    """,
    "zsh": """
        function _pip_completion {{
          local words cword
          read -Ac words
          read -cn cword
          reply=( $( COMP_WORDS="$words[*]" \\
                     COMP_CWORD=$(( cword-1 )) \\
                     PIP_AUTO_COMPLETE=1 $words[1] 2>/dev/null ))
        }}
        compctl -K _pip_completion {prog}
    """,
    "fish": """
        function __fish_complete_pip
            set -lx COMP_WORDS (commandline -o) ""
            set -lx COMP_CWORD ( \\
                math (contains -i -- (commandline -t) $COMP_WORDS)-1 \\
            )
            set -lx PIP_AUTO_COMPLETE 1
            string split \\  -- (eval $COMP_WORDS[1])
        end
        complete -fa "(__fish_complete_pip)" -c {prog}
    """,
}


class CompletionCommand(Command):
    """A helper command to be used for command completion."""

    ignore_require_venv = True

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "--bash",
            "-b",
            action="store_const",
            const="bash",
            dest="shell",
            help="Emit completion code for bash",
        )
        self.cmd_opts.add_option(
            "--zsh",
            "-z",
            action="store_const",
            const="zsh",
            dest="shell",
            help="Emit completion code for zsh",
        )
        self.cmd_opts.add_option(
            "--fish",
            "-f",
            action="store_const",
            const="fish",
            dest="shell",
            help="Emit completion code for fish",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        """Prints the completion code of the given shell"""
        shells = COMPLETION_SCRIPTS.keys()
        shell_options = ["--" + shell for shell in sorted(shells)]
        if options.shell in shells:
            script = textwrap.dedent(
                COMPLETION_SCRIPTS.get(options.shell, "").format(prog=get_prog())
            )
            print(BASE_COMPLETION.format(script=script, shell=options.shell))
            return SUCCESS
        else:
            sys.stderr.write(
                "ERROR: You must pass {}\n".format(" or ".join(shell_options))
            )
            return SUCCESS
site-packages/pip/_internal/commands/freeze.py000064400000005607150564745010015501 0ustar00import sys
from optparse import Values
from typing import List

from pip._internal.cli import cmdoptions
from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import SUCCESS
from pip._internal.operations.freeze import freeze
from pip._internal.utils.compat import stdlib_pkgs

DEV_PKGS = {"pip", "setuptools", "distribute", "wheel"}


class FreezeCommand(Command):
    """
    Output installed packages in requirements format.

    packages are listed in a case-insensitive sorted order.
    """

    usage = """
      %prog [options]"""
    log_streams = ("ext://sys.stderr", "ext://sys.stderr")

    def add_options(self) -> None:
        self.cmd_opts.add_option(
            "-r",
            "--requirement",
            dest="requirements",
            action="append",
            default=[],
            metavar="file",
            help=(
                "Use the order in the given requirements file and its "
                "comments when generating output. This option can be "
                "used multiple times."
            ),
        )
        self.cmd_opts.add_option(
            "-l",
            "--local",
            dest="local",
            action="store_true",
            default=False,
            help=(
                "If in a virtualenv that has global access, do not output "
                "globally-installed packages."
            ),
        )
        self.cmd_opts.add_option(
            "--user",
            dest="user",
            action="store_true",
            default=False,
            help="Only output packages installed in user-site.",
        )
        self.cmd_opts.add_option(cmdoptions.list_path())
        self.cmd_opts.add_option(
            "--all",
            dest="freeze_all",
            action="store_true",
            help=(
                "Do not skip these packages in the output:"
                " {}".format(", ".join(DEV_PKGS))
            ),
        )
        self.cmd_opts.add_option(
            "--exclude-editable",
            dest="exclude_editable",
            action="store_true",
            help="Exclude editable package from output.",
        )
        self.cmd_opts.add_option(cmdoptions.list_exclude())

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        skip = set(stdlib_pkgs)
        if not options.freeze_all:
            skip.update(DEV_PKGS)

        if options.excludes:
            skip.update(options.excludes)

        cmdoptions.check_list_path_option(options)

        for line in freeze(
            requirement=options.requirements,
            local_only=options.local,
            user_only=options.user,
            paths=options.path,
            isolated=options.isolated_mode,
            skip=skip,
            exclude_editable=options.exclude_editable,
        ):
            sys.stdout.write(line + "\n")
        return SUCCESS
site-packages/pip/_internal/commands/cache.py000064400000016544150564745010015266 0ustar00import os
import textwrap
from optparse import Values
from typing import Any, List

import pip._internal.utils.filesystem as filesystem
from pip._internal.cli.base_command import Command
from pip._internal.cli.status_codes import ERROR, SUCCESS
from pip._internal.exceptions import CommandError, PipError
from pip._internal.utils.logging import getLogger

logger = getLogger(__name__)


class CacheCommand(Command):
    """
    Inspect and manage pip's wheel cache.

    Subcommands:

    - dir: Show the cache directory.
    - info: Show information about the cache.
    - list: List filenames of packages stored in the cache.
    - remove: Remove one or more package from the cache.
    - purge: Remove all items from the cache.

    ``<pattern>`` can be a glob expression or a package name.
    """

    ignore_require_venv = True
    usage = """
        %prog dir
        %prog info
        %prog list [<pattern>] [--format=[human, abspath]]
        %prog remove <pattern>
        %prog purge
    """

    def add_options(self) -> None:

        self.cmd_opts.add_option(
            "--format",
            action="store",
            dest="list_format",
            default="human",
            choices=("human", "abspath"),
            help="Select the output format among: human (default) or abspath",
        )

        self.parser.insert_option_group(0, self.cmd_opts)

    def run(self, options: Values, args: List[str]) -> int:
        handlers = {
            "dir": self.get_cache_dir,
            "info": self.get_cache_info,
            "list": self.list_cache_items,
            "remove": self.remove_cache_items,
            "purge": self.purge_cache,
        }

        if not options.cache_dir:
            logger.error("pip cache commands can not function since cache is disabled.")
            return ERROR

        # Determine action
        if not args or args[0] not in handlers:
            logger.error(
                "Need an action (%s) to perform.",
                ", ".join(sorted(handlers)),
            )
            return ERROR

        action = args[0]

        # Error handling happens here, not in the action-handlers.
        try:
            handlers[action](options, args[1:])
        except PipError as e:
            logger.error(e.args[0])
            return ERROR

        return SUCCESS

    def get_cache_dir(self, options: Values, args: List[Any]) -> None:
        if args:
            raise CommandError("Too many arguments")

        logger.info(options.cache_dir)

    def get_cache_info(self, options: Values, args: List[Any]) -> None:
        if args:
            raise CommandError("Too many arguments")

        num_http_files = len(self._find_http_files(options))
        num_packages = len(self._find_wheels(options, "*"))

        http_cache_location = self._cache_dir(options, "http")
        wheels_cache_location = self._cache_dir(options, "wheels")
        http_cache_size = filesystem.format_directory_size(http_cache_location)
        wheels_cache_size = filesystem.format_directory_size(wheels_cache_location)

        message = (
            textwrap.dedent(
                """
                    Package index page cache location: {http_cache_location}
                    Package index page cache size: {http_cache_size}
                    Number of HTTP files: {num_http_files}
                    Wheels location: {wheels_cache_location}
                    Wheels size: {wheels_cache_size}
                    Number of wheels: {package_count}
                """
            )
            .format(
                http_cache_location=http_cache_location,
                http_cache_size=http_cache_size,
                num_http_files=num_http_files,
                wheels_cache_location=wheels_cache_location,
                package_count=num_packages,
                wheels_cache_size=wheels_cache_size,
            )
            .strip()
        )

        logger.info(message)

    def list_cache_items(self, options: Values, args: List[Any]) -> None:
        if len(args) > 1:
            raise CommandError("Too many arguments")

        if args:
            pattern = args[0]
        else:
            pattern = "*"

        files = self._find_wheels(options, pattern)
        if options.list_format == "human":
            self.format_for_human(files)
        else:
            self.format_for_abspath(files)

    def format_for_human(self, files: List[str]) -> None:
        if not files:
            logger.info("Nothing cached.")
            return

        results = []
        for filename in files:
            wheel = os.path.basename(filename)
            size = filesystem.format_file_size(filename)
            results.append(f" - {wheel} ({size})")
        logger.info("Cache contents:\n")
        logger.info("\n".join(sorted(results)))

    def format_for_abspath(self, files: List[str]) -> None:
        if not files:
            return

        results = []
        for filename in files:
            results.append(filename)

        logger.info("\n".join(sorted(results)))

    def remove_cache_items(self, options: Values, args: List[Any]) -> None:
        if len(args) > 1:
            raise CommandError("Too many arguments")

        if not args:
            raise CommandError("Please provide a pattern")

        files = self._find_wheels(options, args[0])

        no_matching_msg = "No matching packages"
        if args[0] == "*":
            # Only fetch http files if no specific pattern given
            files += self._find_http_files(options)
        else:
            # Add the pattern to the log message
            no_matching_msg += ' for pattern "{}"'.format(args[0])

        if not files:
            logger.warning(no_matching_msg)

        for filename in files:
            os.unlink(filename)
            logger.verbose("Removed %s", filename)
        logger.info("Files removed: %s", len(files))

    def purge_cache(self, options: Values, args: List[Any]) -> None:
        if args:
            raise CommandError("Too many arguments")

        return self.remove_cache_items(options, ["*"])

    def _cache_dir(self, options: Values, subdir: str) -> str:
        return os.path.join(options.cache_dir, subdir)

    def _find_http_files(self, options: Values) -> List[str]:
        http_dir = self._cache_dir(options, "http")
        return filesystem.find_files(http_dir, "*")

    def _find_wheels(self, options: Values, pattern: str) -> List[str]:
        wheel_dir = self._cache_dir(options, "wheels")

        # The wheel filename format, as specified in PEP 427, is:
        #     {distribution}-{version}(-{build})?-{python}-{abi}-{platform}.whl
        #
        # Additionally, non-alphanumeric values in the distribution are
        # normalized to underscores (_), meaning hyphens can never occur
        # before `-{version}`.
        #
        # Given that information:
        # - If the pattern we're given contains a hyphen (-), the user is
        #   providing at least the version. Thus, we can just append `*.whl`
        #   to match the rest of it.
        # - If the pattern we're given doesn't contain a hyphen (-), the
        #   user is only providing the name. Thus, we append `-*.whl` to
        #   match the hyphen before the version, followed by anything else.
        #
        # PEP 427: https://www.python.org/dev/peps/pep-0427/
        pattern = pattern + ("*.whl" if "-" in pattern else "-*.whl")

        return filesystem.find_files(wheel_dir, pattern)
site-packages/pip/_internal/cli/main.py000064400000004650150564745010014110 0ustar00"""Primary application entrypoint.
"""
import locale
import logging
import os
import sys
from typing import List, Optional

from pip._internal.cli.autocompletion import autocomplete
from pip._internal.cli.main_parser import parse_command
from pip._internal.commands import create_command
from pip._internal.exceptions import PipError
from pip._internal.utils import deprecation

logger = logging.getLogger(__name__)


# Do not import and use main() directly! Using it directly is actively
# discouraged by pip's maintainers. The name, location and behavior of
# this function is subject to change, so calling it directly is not
# portable across different pip versions.

# In addition, running pip in-process is unsupported and unsafe. This is
# elaborated in detail at
# https://pip.pypa.io/en/stable/user_guide/#using-pip-from-your-program.
# That document also provides suggestions that should work for nearly
# all users that are considering importing and using main() directly.

# However, we know that certain users will still want to invoke pip
# in-process. If you understand and accept the implications of using pip
# in an unsupported manner, the best approach is to use runpy to avoid
# depending on the exact location of this entry point.

# The following example shows how to use runpy to invoke pip in that
# case:
#
#     sys.argv = ["pip", your, args, here]
#     runpy.run_module("pip", run_name="__main__")
#
# Note that this will exit the process after running, unlike a direct
# call to main. As it is not safe to do any processing after calling
# main, this should not be an issue in practice.


def main(args: Optional[List[str]] = None) -> int:
    if args is None:
        args = sys.argv[1:]

    # Configure our deprecation warnings to be sent through loggers
    deprecation.install_warning_logger()

    autocomplete()

    try:
        cmd_name, cmd_args = parse_command(args)
    except PipError as exc:
        sys.stderr.write(f"ERROR: {exc}")
        sys.stderr.write(os.linesep)
        sys.exit(1)

    # Needed for locale.getpreferredencoding(False) to work
    # in pip._internal.utils.encoding.auto_decode
    try:
        locale.setlocale(locale.LC_ALL, "")
    except locale.Error as e:
        # setlocale can apparently crash if locale are uninitialized
        logger.debug("Ignoring error %s when setting locale", e)
    command = create_command(cmd_name, isolated=("--isolated" in cmd_args))

    return command.main(cmd_args)
site-packages/pip/_internal/cli/__pycache__/base_command.cpython-39.pyc000064400000014036150564745010022062 0ustar00a

��Ren�@s<dZddlZddlZddlZddlZddlZddlZddlZddlmZddl	m
Z
mZmZm
Z
mZddlmZddlmZddlmZmZddlmZmZmZmZdd	lmZmZmZmZm Z m!Z!dd
l"m#Z#ddl$m%Z%m&Z&ddl'm(Z(m)Z)dd
l*m+Z,ddl*m-Z-m.Z.ddl/m0Z0dgZ1e�2e3�Z4Gdd�de�Z5dS)z(Base Command class, and related routines�N)�Values)�Any�Callable�List�Optional�Tuple)�
cmdoptions)�CommandContextMixIn)�ConfigOptionParser�UpdatingDefaultsHelpFormatter)�ERROR�PREVIOUS_BUILD_DIR_ERROR�
UNKNOWN_ERROR�VIRTUALENV_NOT_FOUND)�
BadCommand�CommandError�InstallationError�NetworkConnectionError�PreviousBuildDirError�UninstallationError)�check_path_owner)�BrokenStdoutLoggingError�
setup_logging)�get_prog�normalize_path)�TempDirectoryTypeRegistry)�global_tempdir_manager�tempdir_registry)�running_under_virtualenv�Commandcs�eZdZUdZeed<dZeed<deeedd��fdd�
Zdd	�d
d�Z	e
dd�d
d�Ze
eee
d�dd�Zeeee
eefd�dd�Zeee
d�dd�Zeee
d�dd�Z�ZS)r��usageF�ignore_require_venvN)�name�summary�isolated�returnc	s�t���||_||_t|jt��d|��t�d||j|d�|_	d|_
|j���d�}t�
|j	|�|_t�tj|j	�}|j	�|�|��dS)N� F)r!�prog�	formatter�add_help_optionr#�descriptionr%z Options)�super�__init__r#r$r
r!rr�__doc__�parserr�
capitalize�optparse�OptionGroup�cmd_optsr�make_option_group�
general_group�add_option_group�add_options)�selfr#r$r%�
optgroup_name�gen_opts��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/base_command.pyr-.s*
�
�zCommand.__init__)r&cCsdS�Nr=)r8r=r=r>r7LszCommand.add_options)�optionsr&cCst|d�rJ�dS)zf
        This is a no-op so that commands by default do not do the pip version
        check.
        �no_indexN)�hasattr)r8r@r=r=r>�handle_pip_version_checkOsz Command.handle_pip_version_check)r@�argsr&cCst�dSr?)�NotImplementedError)r8r@rDr=r=r>�runXszCommand.run�rDr&cCs|j�|�Sr?)r/�
parse_args�r8rDr=r=r>rH[szCommand.parse_argscCsXzH|���$|�|�Wd�Wt��S1s40YWt��n
t��0dSr?)�main_context�_main�logging�shutdownrIr=r=r>�main_s
�zCommand.mainc	sB|�t��|_|�t��|�|�\}}|j|j|_t|j|j|j	d��|j
rZdtjd<|j
rrd�|j
�tjd<|jr�|js�t�s�t�d�t�t�|jr�t|j�|_t|j�s�t�d|j�d|_d|jvr�t�d	�td
tftd
tfd��fdd
�}z0|j�s||j�}n|j}|||�W|� |�S|� |�0dS)N)�	verbosity�no_color�
user_log_file�1�PIP_NO_INPUTr'�PIP_EXISTS_ACTIONz2Could not find an activated virtualenv (required).z�The directory '%s' or its parent directory is not owned or is not writable by the current user. The cache has been disabled. Check the permissions and owner of that directory. If executing pip with sudo, you should use sudo's -H flag.z
2020-resolverz�--use-feature=2020-resolver no longer has any effect, since it is now the default dependency resolver in pip. This will become an error in pip 21.0..)�run_funcr&cs$t���ttd���fdd��}|S)NrGcs�z�|�}t|t�sJ�|WSty`}z,t�t|��tjddd�tWYd}~Sd}~0tt	t
tfy�}z,t�t|��tjddd�tWYd}~Sd}~0t
�y�}z*t�d|�tjddd�tWYd}~Sd}~0t�y*tdtjd��tjk�r"tjtjd�tYSt�yXt�d�tjddd�tYSt�y|tjddd�tYS0dS)	NzException information:T)�exc_infoz%sz ERROR: Pipe to stdout was broken)�filezOperation cancelled by userz
Exception:)�
isinstance�intr�logger�critical�str�debugr
rrrrrrr�print�sys�stderrrL�DEBUG�	traceback�	print_exc�KeyboardInterrupt�
BaseExceptionr)rD�status�exc)�level_numberrUr=r>�exc_logging_wrapper�sB�
zLCommand._main.<locals>.intercepts_unhandled_exc.<locals>.exc_logging_wrapper)�	functools�wrapsrrY)rUri�rh)rUr>�intercepts_unhandled_exc�s+z/Command._main.<locals>.intercepts_unhandled_exc)!�
enter_contextrrrH�verbose�quietrOrrP�log�no_input�os�environ�
exists_action�join�require_venvr"rrZr[r_�exitr�	cache_dirrr�warning�features_enabledrrY�
debug_moderFrC)r8rDr@rmrFr=rlr>rKfsP�




�
�

�1

�z
Command._main)F)�__name__�
__module__�__qualname__r!r\�__annotations__r"�boolr-r7rrCrrYrFrrHrNrK�
__classcell__r=r=r;r>r*s
	 )6r.rjrL�logging.configr1rsr_rbr�typingrrrrrZpip._internal.clir�!pip._internal.cli.command_contextr	�pip._internal.cli.parserr
r�pip._internal.cli.status_codesrr
rr�pip._internal.exceptionsrrrrrr�pip._internal.utils.filesystemr�pip._internal.utils.loggingrr�pip._internal.utils.miscrr�pip._internal.utils.temp_dirr�TempDirRegistryrr�pip._internal.utils.virtualenvr�__all__�	getLoggerr}rZrr=r=r=r>�<module>s. 
site-packages/pip/_internal/cli/__pycache__/main.cpython-39.pyc000064400000002571150564745010020377 0ustar00a

��Re�	�@s�dZddlZddlZddlZddlZddlmZmZddlm	Z	ddl
mZddlm
Z
ddlmZddlmZe�e�Zdeeeed	�d
d�ZdS)
z Primary application entrypoint.
�N)�List�Optional)�autocomplete)�
parse_command)�create_command)�PipError)�deprecation)�args�returnc
Cs�|durtjdd�}t��t�zt|�\}}WnNty�}z6tj�d|���tj�t	j
�t�d�WYd}~n
d}~00zt�
tjd�Wn2tjy�}zt�d|�WYd}~n
d}~00t|d|vd�}|�|�S)N�zERROR: �z%Ignoring error %s when setting localez
--isolated)�isolated)�sys�argvr�install_warning_loggerrrr�stderr�write�os�linesep�exit�locale�	setlocale�LC_ALL�Error�logger�debugr�main)r	�cmd_name�cmd_args�exc�e�command�r"��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/main.pyr-s  "r)N)�__doc__r�loggingrr�typingrr� pip._internal.cli.autocompletionr�pip._internal.cli.main_parserr�pip._internal.commandsr�pip._internal.exceptionsrZpip._internal.utilsr�	getLogger�__name__r�str�intrr"r"r"r#�<module>s
site-packages/pip/_internal/cli/__pycache__/command_context.cpython-39.pyc000064400000002456150564745010022637 0ustar00a

��Re��@sBddlmZmZddlmZmZmZeddd�ZGdd�d�ZdS)	�)�	ExitStack�contextmanager)�ContextManager�Iterator�TypeVar�_TT)�	covariantcsLeZdZdd��fdd�Zeedd�dd��Zeeed�dd	�Z	�Z
S)
�CommandContextMixInN)�returncst���d|_t�|_dS)NF)�super�__init__�_in_main_contextr�
_main_context��self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/command_context.pyrs
zCommandContextMixIn.__init__ccsT|jr
J�d|_z6|j�dVWd�n1s40YWd|_nd|_0dS)NTF)r
rrrrr�main_context
s
&z CommandContextMixIn.main_context)�context_providerr
cCs|js
J�|j�|�S)N)r
r�
enter_context)rrrrrrs
z!CommandContextMixIn.enter_context)�__name__�
__module__�__qualname__rrrrrrr�
__classcell__rrrrr	s
r	N)	�
contextlibrr�typingrrrrr	rrrr�<module>ssite-packages/pip/_internal/cli/__pycache__/progress_bars.cpython-39.pyc000064400000016752150564745010022334 0ustar00a

��Rel �@s�UddlZddlZddlmZmZmZddlmZddlmZm	Z	m
Z
ddlmZddl
mZddlmZddlmZzdd	lmZWney�dZYn0eeed
�dd�Zee
e�Zeed
<Gdd�d�ZGdd�de�ZGdd�de
�ZGdd�d�ZGdd�d�ZGdd�deee�ZGdd�dee�ZGdd�dee�Z Gdd�dee�Z!Gd d!�d!ee	�Z"Gd"d#�d#ee�Z#Gd$d%�d%eeee�Z$e e fee$fe!e$fe"e$fe#e$fd&�Z%d)d'd(�Z&dS)*�N)�SIGINT�default_int_handler�signal)�Any)�Bar�FillingCirclesBar�IncrementalBar)�Spinner)�WINDOWS)�get_indentation)�format_size)�colorama)�	preferred�fallback�returncCsvt|jdd�}|s|St|dd�t|dd�g}|tt|dg��7}zd�|��|�Wntyl|YS0|SdS)N�encoding�
empty_fill��fill�phases)�getattr�file�list�join�encode�UnicodeEncodeError)rrr�
characters�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/progress_bars.py�_select_progress_classs

�
r�_BaseBarcsDeZdZdZeedd��fdd�Zdd��fdd�Zd	d
�Z�ZS)�InterruptibleMixina�
    Helper to ensure that self.finish() gets called on keyboard interrupt.

    This allows downloads to be interrupted without leaving temporary state
    (like hidden cursors) behind.

    This class is similar to the progress library's existing SigIntMixin
    helper, but as of version 1.2, that helper has the following problems:

    1. It calls sys.exit().
    2. It discards the existing SIGINT handler completely.
    3. It leaves its own handler in place even after an uninterrupted finish,
       which will have unexpected delayed effects if the user triggers an
       unrelated keyboard interrupt some time after a progress-displaying
       download has already completed, for example.
    N��args�kwargsrcs4t�j|i|��tt|j�|_|jdur0t|_dS)z=
        Save the original SIGINT handler for later.
        N)�super�__init__rr�
handle_sigint�original_handlerr��selfr#r$��	__class__rrr&Es
zInterruptibleMixin.__init__�rcst���tt|j�dS)z�
        Restore the original SIGINT handler after finishing.

        This should happen regardless of whether the progress display finishes
        normally, or gets interrupted.
        N)r%�finishrrr(�r*r+rrr.Vs
zInterruptibleMixin.finishcCs|��|�||�dS)z�
        Call self.finish() before delegating to the original SIGINT handler.

        This handler should only be in place while the progress display is
        active.
        N)r.r()r*�signum�framerrrr'`sz InterruptibleMixin.handle_sigint)	�__name__�
__module__�__qualname__�__doc__rr&r.r'�
__classcell__rrr+rr!3s
r!c@seZdZdd�dd�ZdS)�	SilentBarNr-cCsdS�Nrr/rrr�updatelszSilentBar.update)r2r3r4r9rrrrr7ksr7c@seZdZdZdZdZdZdS)�BlueEmojiBar�
%(percent)d%%� )u🔹u🔷u🔵N)r2r3r4�suffix�
bar_prefix�
bar_suffixrrrrrr:psr:csdeZdZeedd��fdd�Zeed�dd��Zeed�dd	��Zeed�d
d��Z	dd
�Z
�ZS)�DownloadProgressMixinNr"cs,t�j|i|��dt�d|j|_dS)Nr<�)r%r&r�messager)r+rrr&yszDownloadProgressMixin.__init__r-cCs
t|j�Sr8)r�indexr/rrr�
downloaded~sz DownloadProgressMixin.downloadedcCs |jdkrdStd|j�dS)Ngz...�z/s)�avgrr/rrr�download_speed�s
z$DownloadProgressMixin.download_speedcCs|jrd|j��SdS)Nzeta r)�eta�eta_tdr/rrr�
pretty_eta�sz DownloadProgressMixin.pretty_etaccs*|D]}|V|�t|��q|��dSr8)�next�lenr.)r*�it�xrrr�iter�szDownloadProgressMixin.iter)r2r3r4rr&�property�strrDrGrJrOr6rrr+rr@xsr@cs&eZdZeedd��fdd�Z�ZS)�WindowsMixinNr"cs\tr�jrd�_t�j|i|��trXtrXt��j��_�fdd��j_�fdd��j_dS)NFcs�jj��Sr8)r�wrapped�isattyrr/rr�<lambda>��z'WindowsMixin.__init__.<locals>.<lambda>cs�jj��Sr8)rrS�flushrr/rrrU�rV)	r
�hide_cursorr%r&r
�AnsiToWin32rrTrWr)r+r/rr&�s
zWindowsMixin.__init__)r2r3r4rr&r6rrr+rrR�srRc@seZdZejZdZdZdS)�BaseDownloadProgressBarr;z0%(downloaded)s %(download_speed)s %(pretty_eta)sN)r2r3r4�sys�stdoutrrBr=rrrrrZ�srZc@seZdZdS)�DefaultDownloadProgressBarN�r2r3r4rrrrr]�sr]c@seZdZdS)�DownloadSilentBarNr^rrrrr_�sr_c@seZdZdS)�DownloadBarNr^rrrrr`�sr`c@seZdZdS)�DownloadFillingCirclesBarNr^rrrrra�srac@seZdZdS)�DownloadBlueEmojiProgressBarNr^rrrrrb�srbc@s2eZdZejZdZed�dd�Zdd�dd�Z	dS)�DownloadProgressSpinnerz!%(downloaded)s %(download_speed)sr-cCs"t|d�st�|j�|_t|j�S)N�_phaser)�hasattr�	itertools�cyclerrdrKr/rrr�
next_phase�s
z"DownloadProgressSpinner.next_phaseNcCsN|j|}|��}|j|}d�||r*dnd||r6dnd|g�}|�|�dS)Nrr<)rBrhr=r�writeln)r*rB�phaser=�linerrrr9�s



��
zDownloadProgressSpinner.update)
r2r3r4r[r\rr=rQrhr9rrrrrc�src)�off�on�ascii�pretty�emojicCs8|dus|dkr t|d�jSt|d|d�jSdS)NrrE)�max)�	BAR_TYPESrO)�progress_barrqrrr�DownloadProgressProvider�srt)N)'rfr[rrr�typingr�pip._vendor.progress.barrrr�pip._vendor.progress.spinnerr	�pip._internal.utils.compatr
�pip._internal.utils.loggingr�pip._internal.utils.miscr�pip._vendorr
�	Exceptionrr �__annotations__r!r7r:r@rRrZr]r_r`rarbrcrrrtrrrr�<module>sF

8 
��	site-packages/pip/_internal/cli/__pycache__/spinners.cpython-39.pyc000064400000011565150564745010021317 0ustar00a

��Re��@s�ddlZddlZddlZddlZddlZddlmZmZddlm	Z	m
Z
ddlmZddl
mZe�e�ZGdd�d�ZGdd	�d	e�ZGd
d�de�ZGdd
�d
�Zejeeed�dd��Zejeeedd�dd��ZdS)�N)�IO�Iterator)�HIDE_CURSOR�SHOW_CURSOR)�WINDOWS)�get_indentationc@s*eZdZdd�dd�Zedd�dd�ZdS)�SpinnerInterfaceN��returncCs
t��dS�N��NotImplementedError��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/spinners.py�spinszSpinnerInterface.spin��final_statusr
cCs
t��dSrr�rrrrr�finishszSpinnerInterface.finish)�__name__�
__module__�__qualname__r�strrrrrrrsrc@sTeZdZdeeeeed�dd�Zedd�dd	�Zdd
�dd�Zedd
�dd�Z	dS)�InteractiveSpinnerN�-\|/��?)�message�file�
spin_chars�min_update_interval_secondscCs\||_|durtj}||_t|�|_d|_t�|�|_	|j�
dt�|jd�d|_dS)NF� z ... r)
�_message�sys�stdout�_file�RateLimiter�
_rate_limiter�	_finished�	itertools�cycle�_spin_cycle�writer�_width)rrrr r!rrr�__init__s
zInteractiveSpinner.__init__��statusr
cCs\|jr
J�d|j}|j�|d|j|�|j�|�t|�|_|j��|j��dS)N�r")r)r.r&r-�len�flushr(�reset)rr1�backuprrr�_write-s



zInteractiveSpinner._writer	cCs,|jr
dS|j��sdS|�t|j��dSr)r)r(�readyr7�nextr,rrrrr9s

zInteractiveSpinner.spinrcCs4|jr
dS|�|�|j�d�|j��d|_dS)N�
T)r)r7r&r-r4rrrrr@s

zInteractiveSpinner.finish)Nrr)
rrrrr�floatr/r7rrrrrrrs��rc@sNeZdZdeedd�dd�Zedd�dd�Zdd	�d
d�Zedd�d
d�ZdS)�NonInteractiveSpinner�N@N)rr!r
cCs$||_d|_t|�|_|�d�dS)NF�started)r#r)r'r(�_update)rrr!rrrr/Ns
zNonInteractiveSpinner.__init__r0cCs(|jr
J�|j��t�d|j|�dS)Nz%s: %s)r)r(r5�logger�infor#)rr1rrrr?Ts

zNonInteractiveSpinner._updater	cCs&|jr
dS|j��sdS|�d�dS)Nzstill running...)r)r(r8r?rrrrrYs

zNonInteractiveSpinner.spinrcCs&|jr
dS|�d|�d��d|_dS)Nzfinished with status '�'T)r)r?rrrrr`szNonInteractiveSpinner.finish)r=)	rrrrr;r/r?rrrrrrr<Msr<c@s8eZdZedd�dd�Zed�dd�Zdd�dd	�ZdS)
r'N)r!r
cCs||_d|_dS)Nr)�_min_update_interval_seconds�_last_update)rr!rrrr/hszRateLimiter.__init__r	cCst��}||j}||jkSr)�timerDrC)r�now�deltarrrr8ls
zRateLimiter.readycCst��|_dSr)rErDrrrrr5qszRateLimiter.reset)rrrr;r/�boolr8r5rrrrr'gsr')rr
ccs�tj��r"t��tjkr"t|�}nt|�}z4t	tj��|VWd�n1sR0YWn:t
y||�d��Yn(ty�|�d��Yn0|�d�dS)N�canceled�error�done)
r$r%�isattyr@�getEffectiveLevel�logging�INFOrr<�
hidden_cursor�KeyboardInterruptr�	Exception)r�spinnerrrr�open_spinnerus
(

rT)rr
c	csZtrdVnJ|��r"t��tjkr*dVn,|�t�zdVW|�t�n|�t�0dSr)	rrLr@rMrNrOr-rr)rrrrrP�s
rP)�
contextlibr*rNr$rE�typingrrZpip._vendor.progressrr�pip._internal.utils.compatr�pip._internal.utils.loggingr�	getLoggerrr@rrr<r'�contextmanagerrrTrPrrrr�<module>s"
5site-packages/pip/_internal/cli/__pycache__/cmdoptions.cpython-39.pyc000064400000054114150564745010021632 0ustar00a

��Re�n�@s2
UdZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
mZddlmZddl
mZmZmZmZmZddlmZddlmZdd	lmZdd
lmZddlmZmZddlmZdd
l m!Z!ddl"m#Z#ddl$m%Z%ddl&m'Z'e
ee(dd�dd�Z)ee(efee	d�dd�Z*�d(eeedd�dd�Z+�d)ee,dd�dd�Z-ee(e(e(d�dd �Z.ee(e(e(d�d!d"�Z/Gd#d$�d$e�Z0eed%d&d'd'd(d)�Z1ed*efe2d+<eed,d-d.dd/d0�Z3ed*efe2d-<eed1d2d.dd3d0�Z4ed*efe2d2<eed4d5d6d.ded0�Z5ed*efe2d7<eed8d9d:d;dd<d0�Z6ed*efe2d:<eed=d>d.dd?d0�Z7ed*efe2d><eed@dAdBd.dCd)�Z8ed*efe2dB<eedDdEdFd;ddGd0�Z9ed*efe2dF<eedHdIdJe:e�;��dKdLdM�<e�;��dNdO�Z=ed*efe2dI<ee0dPdQdRdSdTdTdUdV�Z>ed*efe2dS<eedWdXd.ddYd0�Z?ed*efe2dX<eedZd[d\d]d^d_�Z@ed*efe2d[<eed`dadbdcddd_�ZAed*efe2da<eededfdgdhdidjdkdl�ZBed*efe2dh<edm�dndo�ZCee0dpdqdTdTdrds�ZDed*efe2dq<ee0dtdudTddTdvdw�ZEed*efe2du<eedxdydzd{d|e!jFd}d~�ZGed*efe2d{<edm�dd��ZHeed�d�d.dd�d0�ZIed*efe2d�<edm�d�d��ZJedm�d�d��ZKedm�d�d��ZLedm�d�d��ZMedm�d�d��ZNee(e(e
dd��d�d��ZOee0d�d�d�d�d�dTd�e�d�eOd�d��ZPed*efe2d�<eeed��d�d��ZQee(e(e
dd��d�d��ZRee(e(e
dd��d�d��ZSedm�d�d��ZTedm�d�d��ZUeed�d�d�d�dd�d��ZVed*efe2d�<e(eeeWd*fee(fd��d�d��ZXee(e(e
dd��d�d��ZYeed�d�d�d�eYd\ded��d��	ZZed*efe2d�<eed�d�d�dd�d~�Z[ed*efe2d�<eed�d�d�d�dd�d��Z\ed*efe2d�<e	dd��d�d��Z]ee#d��d�d��Z^edm�d�dÄZ_ee0d�d�ed�dTd�dǍZ`ed*efe2d�<ee(e(e
ddȜd�dʄZaeed�d�d�ead�d͍Zbed*efe2d�<eed�d�d�d.dd�d0�Zced*efe2d�<eed�d�d.d�d)�Zded*efe2d�<eed�d�d�d�d�d0�Zeed*efe2d�<ee(e(e
ddȜd�dބZfeed�d�d.dd�d0�Zgee2d�<eed�d�d�efded�Zhee2d�<eed�d�d�d�d�d�Zied*efe2d�<eed�d�d�d�d�d�Zjed*efe2d�<eed�d�d�d�d�d�Zked*efe2d�<eed�d.dd�d�Zled*efe2d�<eed�d.dd�d�Zmed*efe2d�<eed�d�d.dd�d0�Zned*efe2d�<ee(e(e
dd��d�d��Zoeed�d�d�eod��d�d�Zped*efe2�d<ee�d�dd.d�dd0�Zqed*efe2�d<ee0�ddTdTd��d�d�Zred*efe2�d	<edd���d
�d�Zsee0�d�d
d��d�d�d�d�Zted*efe2�d<ee�d�dd.d�dd0�Zued*efe2�d<ee�d�d�dd�gg�d��d�d�Zved*efe2�d<ee�d�d�dd�g�d�d g�d!�d�Zwed*efe2�d"<�d#e1e3e4e5e6e8e9e>e?e@eAeBeCeKeDeEe`ebene7euevewg�d$�Zxee(efe2�d%<�d&eGeHeIeJg�d$�Zyee(efe2�d'<dS(*aC
shared options and groups

The principle here is to define options once, but *not* instantiate them
globally. One reason being that options with action='append' can carry state
between parses. pip parses general options twice internally, and shouldn't
pass on state. To be consistent, all options will follow this design.
�N)�partial)�
SUPPRESS_HELP�Option�OptionGroup�OptionParser�Values)�dedent)�Any�Callable�Dict�Optional�Tuple��canonicalize_name)�ConfigOptionParser)�	BAR_TYPES)�CommandError)�USER_CACHE_DIR�get_src_prefix)�
FormatControl)�PyPI)�TargetPython)�
STRONG_HASHES)�	strtobool)�parser�option�msg�returncCs0|�d|��}t�d�|����}|�|�dS)z�
    Raise an option parsing error using parser.error().

    Args:
      parser: an OptionParser instance.
      option: an Option instance.
      msg: the error text.
    z error: � N)�textwrap�fill�join�split�error)rrr�r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/cmdoptions.py�raise_option_error"s	r&)�grouprrcCs,t||d�}|dD]}|�|��q|S)z�
    Return an OptionGroup object
    group  -- assumed to be dict with 'name' and 'options' keys
    parser -- an optparse Parser
    �name�options)r�
add_option)r'r�option_grouprr$r$r%�make_option_group0sr,)r)�
check_optionsrcsZ�dur|�tttd��fdd�}gd�}tt||��rV|j}|��tjddd�dS)	z�Disable wheels if per-setup.py call options are set.

    :param options: The OptionParser options to update.
    :param check_options: The options to check, if not supplied defaults to
        options.
    N)�nrcst�|d�S�N)�getattr)r.�r-r$r%�getnameHsz+check_install_build_global.<locals>.getname)�
build_options�global_options�install_optionszbDisabling all use of wheels due to the use of --build-option / --global-option / --install-option.�)�
stacklevel)	�strrr	�any�map�format_control�disallow_binaries�warnings�warn)r)r-r2�names�controlr$r1r%�check_install_build_global<s	�rAF)r)�check_targetrcCsbt|j|j|j|jg�}tt�dh�}|j|ko6|j}|rH|rHt	d��|r^|r^|j
s^t	d��dS)z�Function for determining if custom platform options are allowed.

    :param options: The OptionParser options.
    :param check_target: Whether or not to check if --target is being used.
    z:all:z�When restricting platform and interpreter constraints using --python-version, --platform, --abi, or --implementation, either --no-deps must be set, or --only-binary=:all: must be set and --no-binary must not be set (or must be set to :none:).zQCan not use any platform or abi specific options unless installing via '--target'N)r9�python_version�	platforms�abis�implementationr�setr;�ignore_dependenciesr�
target_dir)r)rB�dist_restriction_set�binary_only�sdist_dependencies_allowedr$r$r%�check_dist_restrictionVs&��	��
�rM)r�opt�valuercCstj�|�Sr/)�os�path�
expanduser�rrNrOr$r$r%�_path_option_check~srTcCst|�Sr/rrSr$r$r%�_package_name_option_check�srUc@s0eZdZejdZej��Zeed<eed<dS)�	PipOption)rQ�package_namerWrQN)	�__name__�
__module__�__qualname__r�TYPES�TYPE_CHECKER�copyrUrTr$r$r$r%rV�s

rVz-hz--help�helpz
Show help.)�dest�actionr^.�help_z--debug�
debug_mode�
store_truezbLet unhandled exceptions propagate outside the main subroutine, instead of logging them to stderr.�r_r`�defaultr^z
--isolated�
isolated_modezSRun pip in an isolated mode, ignoring environment variables and user configuration.z--require-virtualenvz--require-venv�require_venv�require_virtualenvz-vz	--verbose�verbose�countzDGive more output. Option is additive, and can be used up to 3 times.z
--no-color�no_colorzSuppress colored output.z-Vz	--version�versionzShow version and exit.z-qz--quiet�quietz�Give less output. Option is additive, and can be used up to 3 times (corresponding to WARNING, ERROR, and CRITICAL logging levels).z--progress-bar�progress_bar�choice�onz*Specify type of progress to be displayed [�|z] (default: %default))r_�type�choicesrer^z--logz
--log-filez--local-log�logrQz Path to a verbose appending log.)r_�metavarrrr^z
--no-input�no_inputzDisable prompting for input.z--proxy�proxyr8�z<Specify a proxy in the form [user:passwd@]proxy.server:port.)r_rrrer^z	--retries�retries�int�zRMaximum number of retries each connection should attempt (default %default times).z	--timeoutz--default-timeout�sec�timeout�float�z2Set the socket timeout (default %default seconds).)rur_rrrer^)rc
Cstdddgd�gdddd�S)	Nz--exists-action�
exists_actionro)�s�i�w�b�a�appendr`zYDefault action when a path already exists: (s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.)r_rrrsrer`rur^�rr$r$r$r%r�)s�r�z--cert�certz�Path to PEM-encoded CA certificate bundle. If provided, overrides the default. See 'SSL Certificate Verification' in pip documentation for more information.)r_rrrur^z
--client-cert�client_certzkPath to SSL client certificate, a single file containing the private key and the certificate in PEM format.)r_rrrerur^z-iz--index-urlz
--pypi-url�	index_url�URLz�Base URL of the Python Package Index (default %default). This should point to a repository compliant with PEP 503 (the simple repository API) or a local directory laid out in the same format.)r_rurer^cCstddddgdd�S)Nz--extra-index-url�extra_index_urlsr�r�zmExtra URLs of package indexes to use in addition to --index-url. Should follow the same rules as --index-url.�r_rur`rer^r�r$r$r$r%�extra_index_url`s�r�z
--no-index�no_indexzAIgnore package index (only looking at --find-links URLs instead).c	Cstddddgddd�S)Nz-fz--find-links�
find_linksr��urlaIf a URL or path to an html file, then parse for links to archives such as sdist (.tar.gz) or wheel (.whl) files. If a local path or file:// URL that's a directory, then look for archives in the directory listing. Links to VCS project URLs are not supported.�r_r`rerur^r�r$r$r$r%r�ws�r�cCstddddgdd�S)Nz--trusted-host�
trusted_hostsr��HOSTNAMEz]Mark this host or host:port pair as trusted, even though it does not have valid or any HTTPS.)r_r`rurer^r�r$r$r$r%�trusted_host�s�r�c	Cstddddgddd�S)Nz-cz--constraint�constraintsr��filez\Constrain versions using the given constraints file. This option can be used multiple times.r�r�r$r$r$r%r��s�r�c	Cstddddgddd�S)Nz-rz
--requirement�requirementsr�r�zQInstall from the given requirements file. This option can be used multiple times.r�r�r$r$r$r%r��s�r�c	Cstddddgddd�S)Nz-ez
--editable�	editablesr�zpath/urlzkInstall a project in editable mode (i.e. setuptools "develop mode") from a local project path or a VCS url.r�r�r$r$r$r%�editable�s�r�)r�opt_strrOrrcCs tj�|�}t|j|j|�dSr/)rPrQ�abspath�setattr�valuesr_)rr�rOrr$r$r%�_handle_src�sr�z--srcz--sourcez--source-dirz--source-directory�src_dir�dir�callbackz�Directory to check out editable projects into. The default in a virtualenv is "<venv path>/src". The default for global installs is "<current dir>/src".)r_rrrurer`r�r^�src)r�rrcCst||j�S)zGet a format_control object.)r0r_)r�rr$r$r%�_get_format_control�sr�cCs"t|j|�}t�||j|j�dSr/)r�r�r�handle_mutual_excludes�	no_binary�only_binary�rr�rOr�existingr$r$r%�_handle_no_binary�s�r�cCs"t|j|�}t�||j|j�dSr/)r�r�rr�r�r�r�r$r$r%�_handle_only_binary�s�r�c	Cs$tt�t��}tdddtd|dd�S)Nz--no-binaryr;r�r8avDo not use binary packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either ":all:" to disable all binary packages, ":none:" to empty the set (notice the colons), or one or more package names with commas between them (no colons). Note that some packages are tricky to compile and may fail to install when this option is used on them.�r_r`r�rrrer^)rrGrr��r;r$r$r%r��s�r�c	Cs$tt�t��}tdddtd|dd�S)Nz
--only-binaryr;r�r8aKDo not use source packages. Can be supplied multiple times, and each time adds to the existing value. Accepts either ":all:" to disable all source packages, ":none:" to empty the set, or one or more package names with commas between them. Packages without binary distributions will fail to install when this option is used on them.r�)rrGrr�r�r$r$r%r�s�r�z
--platformrD�platformr�z�Only use wheels compatible with <platform>. Defaults to the platform of the running system. Use this option multiple times to specify multiple platforms supported by the target interpreter.r�)rOrcCs�|sdS|�d�}t|�dkr"dSt|�dkrV|d}t|�dkrV|d|dd�g}ztdd	�|D��}Wnty�Yd
S0|dfS)z�
    Convert a version string like "3", "37", or "3.7.3" into a tuple of ints.

    :return: A 2-tuple (version_info, error_msg), where `error_msg` is
        non-None if and only if there was a parsing error.
    )NN�.�)r$z'at most three version parts are allowed�rNcss|]}t|�VqdSr/)rz)�.0�partr$r$r%�	<genexpr>8�z*_convert_python_version.<locals>.<genexpr>)r$z$each version part must be an integer)r"�len�tuple�
ValueError)rO�parts�version_infor$r$r%�_convert_python_version"s
r�cCs:t|�\}}|dur.d�||�}t|||d�||j_dS)z3
    Handle a provided --python-version value.
    Nz(invalid --python-version value: {!r}: {}�rr)r��formatr&r�rC)rr�rOrr��	error_msgrr$r$r%�_handle_python_version?s�r�z--python-versionrCa�    The Python interpreter version to use for wheel and "Requires-Python"
    compatibility checks. Defaults to a version derived from the running
    interpreter. The version can be specified using up to three dot-separated
    integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor
    version can also be given as a string without dots (e.g. "37" for 3.7.0).
    )r_rur`r�rrrer^z--implementationrFz�Only use wheels compatible with Python implementation <implementation>, e.g. 'pp', 'jy', 'cp',  or 'ip'. If not specified, then the current interpreter implementation is used.  Use 'py' to force implementation-agnostic wheels.z--abirE�abiaLOnly use wheels compatible with Python abi <abi>, e.g. 'pypy_41'. If not specified, then the current interpreter abi tag is used. Use this option multiple times to specify multiple abis supported by the target interpreter. Generally you will need to specify --implementation, --platform, and --python-version when using this option.)�cmd_optsrcCs4|�t��|�t��|�t��|�t��dSr/)r*rDrCrFrE)r�r$r$r%�add_target_python_options�sr�)r)rcCst|j|j|j|jd�}|S)N)rD�py_version_inforErF)rrDrCrErF)r)�
target_pythonr$r$r%�make_target_python�s�r�cCstdddddd�S)Nz--prefer-binary�
prefer_binaryrcFz8Prefer older binary packages over newer source packages.rdr�r$r$r$r%r��s�r�z--cache-dir�	cache_dirzStore the cache data in <dir>.)r_rerurrr^)rrNrOrrc
CsX|durLzt|�Wn6tyJ}zt||t|�d�WYd}~n
d}~00d|j_dS)z�
    Process a value provided for the --no-cache-dir option.

    This is an optparse.Option callback for the --no-cache-dir option.
    Nr�F)rr�r&r8r�r�)rrNrOr�excr$r$r%�_handle_no_cache_dir�s(	r�z--no-cache-dirzDisable the cache.)r_r`r�r^�no_cachez	--no-depsz--no-dependenciesrHz#Don't install package dependencies.�no_depsz--ignore-requires-python�ignore_requires_pythonz'Ignore the Requires-Python information.z--no-build-isolation�build_isolation�store_falseTz�Disable isolation when building a modern source distribution. Build dependencies specified by PEP 518 must be already installed if this option is used.�no_build_isolationcCs&|durd}t|||d�d|j_dS)z�
    Process a value provided for the --no-use-pep517 option.

    This is an optparse.Option callback for the no_use_pep517 option.
    Na0A value was passed for --no-use-pep517,
        probably using either the PIP_NO_USE_PEP517 environment variable
        or the "no-use-pep517" config file option. Use an appropriate value
        of the PIP_USE_PEP517 environment variable or the "use-pep517"
        config file option instead.
        r�F)r&r��
use_pep517)rrNrOrrr$r$r%�_handle_no_use_pep517�sr�z--use-pep517r�z^Use PEP 517 for building source distributions (use --no-use-pep517 to force legacy behaviour).z--no-use-pep517)r_r`r�rer^�
no_use_pep517z--install-optionr5r)a"Extra arguments to be supplied to the setup.py install command (use like --install-option="--install-scripts=/usr/local/bin"). Use multiple --install-option options to pass multiple options to setup.py install. If you are using an option with a directory path, be sure to use absolute path.)r_r`rur^z--build-optionr3z9Extra arguments to be supplied to 'setup.py bdist_wheel'.)r_rur`r^z--global-optionr4zcExtra global options to be supplied to the setup.py call before the install or bdist_wheel command.z
--no-cleanz!Don't clean up build directories.)r`rer^�no_cleanz--prezYInclude pre-release and development versions. By default, pip only finds stable versions.�prez--disable-pip-version-check�disable_pip_version_checkz{Don't periodically check PyPI to determine whether a new version of pip is available for download. Implied with --no-index.cCs�|jjsi|j_z|�dd�\}}Wn"tyF|�d�|��Yn0|tvrh|�d�|d�t���|jj�|g��	|�dS)zkGiven a value spelled "algo:digest", append the digest to a list
    pointed to in a dict by the algo name.�:r�zTArguments to {} must be a hash name followed by a value, like --hash=sha256:abcde...z&Allowed hash algorithms for {} are {}.z, N)
r��hashesr"r�r#r�rr!�
setdefaultr�)rr�rOr�algo�digestr$r$r%�_handle_merge_hash[s"��

��r�z--hashr��stringzgVerify that the package's archive matches this hash before installing. Example: --hash=sha256:abcdef...)r_r`r�rrr^�hashz--require-hashes�require_hashesz�Require a hash to check each requirement against, for repeatable installs. This option is implied when any package in a requirements file has a --hash option.z--pathz^Restrict to the specified installation path for listing packages (can be used multiple times).)r_rrr`r^�	list_pathcCs|jr|js|jrtd��dS)Nz2Cannot combine '--path' with '--user' or '--local')rQ�user�localr)r)r$r$r%�check_list_path_option�sr�z	--exclude�excludes�packagerWz)Exclude specified package from the output)r_r`rurrr^�list_excludez--no-python-version-warning�no_python_version_warningz>Silence deprecation warnings for upcoming unsupported Pythons.z
--use-feature�features_enabled�feature)z
2020-resolverz	fast-depsz
in-tree-buildz<Enable new functionality, that may be backward incompatible.)r_rur`rersr^�use_new_featurez--use-deprecated�deprecated_features_enabledzlegacy-resolverzout-of-tree-buildzDEnable deprecated functionality, that will be removed in the future.�use_deprecated_featurezGeneral Options)r(r)�
general_groupzPackage Index Options�index_group)N)F)z�__doc__rPrr=�	functoolsr�optparserrrrrr�typingr	r
rrr
�pip._vendor.packaging.utilsr�pip._internal.cli.parserr�pip._internal.cli.progress_barsr�pip._internal.exceptionsr�pip._internal.locationsrr�#pip._internal.models.format_controlr�pip._internal.models.indexr�"pip._internal.models.target_pythonr�pip._internal.utils.hashesr�pip._internal.utils.miscrr8r&r,rA�boolrMrTrUrVra�__annotations__rbrfrhrirkrlrm�list�keysr!rnrtrvrwryr}r�r�r��
simple_urlr�r�r�r�r�r�r�r�r�r�r�r�r�r�r�rDrzr�r�rCrFrEr�r�r�r�r�r�r�r�r�r�r�r�r5r3r4r�r�r�r�r�r�r�r�r�r�r�r�r�r�r$r$r$r%�<module>s�
��(�	����
�	�	�
�����
�	�
����
�


����$�����
���	�
��
��
�
�
�	�
��	�������
�
�����site-packages/pip/_internal/cli/__pycache__/req_command.cpython-39.pyc000064400000030151150564745010021733 0ustar00a

��Re�B�@s�dZddlZddlZddlZddlmZddlmZddlm	Z	m
Z
mZmZddl
mZddlmZddlmZdd	lmZdd
lmZmZddlmZddlmZdd
lmZddlmZddl m!Z!ddl"m#Z#ddl$m%Z%m&Z&m'Z'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6m7Z7m8Z8ddl9m:Z:e�;e<�Z=Gdd�de�Z>Gdd�dee>�Z?e8j@e8jAe8jBgZCdd�dd �ZDe	e	d!�d"d#�ZEGd$d%�d%e?�ZFdS)&aContains the Command base classes that depend on PipSession.

The classes in this module are in a separate module so the commands not
needing download / PackageFinder capability don't unnecessarily import the
PackageFinder machinery and all its vendored dependencies, etc.
�N)�partial)�Values)�Any�List�Optional�Tuple)�
WheelCache)�
cmdoptions)�Command)�CommandContextMixIn)�CommandError�PreviousBuildDirError)�
LinkCollector)�
PackageFinder)�SelectionPreferences)�TargetPython)�
PipSession)�RequirementPreparer)�install_req_from_editable�install_req_from_line�#install_req_from_parsed_requirement�install_req_from_req_string)�parse_requirements)�InstallRequirement)�RequirementTracker)�BaseResolver)�pip_self_version_check)�
deprecated)�
TempDirectory�TempDirectoryTypeRegistry�
tempdir_kinds)�running_under_virtualenvcspeZdZdZdd��fdd�Zeeeee	d�dd��Z
eed�d	d
�Zdeee
ee
ed�dd
�Z�ZS)�SessionCommandMixinzE
    A class mixin for command classes needing _build_session().
    N��returncst���d|_dS�N)�super�__init__�_session)�self��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/req_command.pyr'6s
zSessionCommandMixin.__init__��optionsr$cCsLg}t|dd�s*t|dd�}|r*|�|�t|dd�}|rD|�|�|pJdS)z7Return a list of index urls from user-provided options.�no_indexF�	index_urlN�extra_index_urls)�getattr�append�extend)�clsr/�
index_urls�url�urlsr,r,r-�_get_index_urls:s

z#SessionCommandMixin._get_index_urlscCs0|jdur*|�|�|��|_|jdus*J�|jS)zGet a default-managed session.N)r(�
enter_context�_build_session)r)r/r,r,r-�get_default_sessionHs
z'SessionCommandMixin.get_default_session)r/�retries�timeoutr$cCs�|jrtj�|j�sJ�t|jr0tj�|jd�nd|dur>|n|j|j|�|�d�}|j	rb|j	|_
|jrp|j|_	|jsz|r�|dur�|n|j|_|j
r�|j
|j
d�|_|j|j_|S)N�http)�cacher>�
trusted_hostsr7)r@�https)�	cache_dir�os�path�isabsr�joinr>rBr:�cert�verify�client_certr?�proxy�proxies�no_input�auth�	prompting)r)r/r>r?�sessionr,r,r-r<Rs&�

�z"SessionCommandMixin._build_session)NN)�__name__�
__module__�__qualname__�__doc__r'�classmethodrrr�strr:rr=�intr<�
__classcell__r,r,r*r-r"0s

��r"c@s eZdZdZedd�dd�ZdS)�IndexGroupCommandz�
    Abstract base class for commands with the index_group options.

    This also corresponds to the commands that permit the pip version check.
    Nr.cCsht|d�sJ�|js|jrdS|j|dtd|j�d�}|�t||�Wd�n1sZ0YdS)z�
        Do the pip version check if not disabled.

        This overrides the default behavior of not doing the check.
        r0Nr�)r>r?)�hasattr�disable_pip_version_checkr0r<�minr?r)r)r/rQr,r,r-�handle_pip_version_check�s�z*IndexGroupCommand.handle_pip_version_check)rRrSrTrUrr_r,r,r,r-rZ{srZr#cCsNt�r
dSttd�sdStjdks,tjdkr0dSt��dkr@dSt�d�dS)z�Output a warning for sudo users on Unix.

    In a virtual environment, sudo pip still writes to virtualenv.
    On Windows, users may run pip as Administrator without issues.
    This warning only applies to Unix root users outside of virtualenv.
    N�getuid�win32�cygwinrz�Running pip as the 'root' user can result in broken permissions and conflicting behaviour with the system package manager. It is recommended to use a virtual environment instead: https://pip.pypa.io/warnings/venv)r!r\rE�sys�platformr`�logger�warningr,r,r,r-�warn_if_run_as_root�s
�rg)�funcr$cs6tdd�dd��ttttttd���fdd�}|S)zNDecorator for common logic related to managing temporary
    directories.
    N)�registryr$cSstD]}|�|d�qdS)NF)�KEEPABLE_TEMPDIR_TYPES�
set_delete)ri�tr,r,r-�configure_tempdir_registry�sz0with_cleanup.<locals>.configure_tempdir_registry)r)r/�argsr$csP|jdusJ�|jr�|j�z�|||�WStyJ�|j��Yn0dSr%)�tempdir_registry�no_cleanr
)r)r/rn�rmrhr,r-�wrapper�s

zwith_cleanup.<locals>.wrapper)r�RequirementCommandrrrrrX)rhrrr,rqr-�with_cleanup�s

�rtcs�eZdZeedd��fdd�Zeeed�dd��Ze	de
eeee
eeeed�d	d
��Ze	dee
eeeeeeeeeeeeedfed�dd��Zeeee
eeed�dd�Zee
dd�dd��Zdeeeeeee
d�dd�Z�ZS)rsN)rn�kwr$cs&t�j|i|��|j�t���dSr%)r&r'�cmd_opts�
add_optionr	rp)r)rnrur*r,r-r'�szRequirementCommand.__init__r.cCsd|jvrdSdS)zEDetermines which resolver should be used, based on the given options.zlegacy-resolver�legacy�
2020-resolver)�deprecated_features_enabled�r/r,r,r-�determine_resolver_variant�s
z-RequirementCommand.determine_resolver_variant)�temp_build_dirr/�req_trackerrQ�finder�
use_user_site�download_dirr$cCs�|j}|dusJ�|�|�}	|	dkr>d|jv}
|
rVt�d�nd}
d|jvrVt�d�d|jv}d|jvrxtd	d
dd�d|jvr�td
ddd�t||j||j	|||j
||j||
|d�S)zQ
        Create a RequirementPreparer instance for the given parameters.
        Nryz	fast-depsz�pip is using lazily downloaded wheels using HTTP range requests to obtain dependency information. This experimental feature is enabled through --use-feature=fast-deps and it is not ready for production.Fz;fast-deps has no effect when used with the legacy resolver.zout-of-tree-buildz
in-tree-buildz#In-tree builds are now the default.z.to remove the --use-feature=in-tree-build flagz22.1)�reason�replacement�gone_inz"Out-of-tree builds are deprecated.)�	build_dir�src_dirr��build_isolationr~rQ�progress_barr�require_hashesr��
lazy_wheel�
in_tree_build)rFr|�features_enabledrerfrzrrr�r�r�r�)r6r}r/r~rQrr�r��temp_build_dir_path�resolver_variantr�r�r,r,r-�make_requirement_preparer�sR

�
�

�
��z,RequirementCommand.make_requirement_preparerFT�to-satisfy-only.)�preparerrr/�wheel_cacher��ignore_installed�ignore_requires_python�force_reinstall�upgrade_strategy�
use_pep517�py_version_infor$c
Cs�tt|j|
d�}|�|�}
|
dkrTddl}|jjjjj	||||||j
||||	|d�Sddl}|jjjjj	||||||j
||||	|d�S)zF
        Create a Resolver instance for the given parameters.
        )�isolatedr�ryrN)r�rr��make_install_reqr��ignore_dependenciesr�r�r�r�r�)
rr�
isolated_moder|�,pip._internal.resolution.resolvelib.resolver�	_internal�
resolution�
resolvelib�resolver�Resolverr��(pip._internal.resolution.legacy.resolverrx)r6r�rr/r�r�r�r�r�r�r�r�r�r��pipr,r,r-�
make_resolver&sF�
�
�z RequirementCommand.make_resolver)rnr/rrQr$c	CsTg}|jD]6}t|d|||d�D]}t||jdd�}|�|�q q
|D]$}	t|	d|j|jdd�}|�|�qF|jD]"}	t|	d|j|jd�}|�|�qr|j	D]8}t||||d�D]"}t||j|jdd�}|�|�q�q�t
d	d
�|D��r�d|_|�sP|j�sP|j	�sPd|ji}
|j
�r<tdjfit|
d
�|j
�d�����ntdjfi|
����|S)zS
        Parse command-line arguments into the corresponding requirements.
        T)�
constraintrr/rQF)r��
user_suppliedN)r�r�r�)r�r�r�)rr/rQcss|]}|jVqdSr%)�has_hash_options)�.0�reqr,r,r-�	<genexpr>��z6RequirementCommand.get_requirements.<locals>.<genexpr>�namezXYou must give at least one requirement to {name} (maybe you meant "pip {name} {links}"?)� )�linkszHYou must give at least one requirement to {name} (see "pip help {name}"))�constraintsrrr�r4rr��	editablesr�requirements�anyr�r��
find_linksr�format�dictrH)r)rnr/rrQr��filename�
parsed_req�
req_to_addr��optsr,r,r-�get_requirementsasv

�
��
�
�
�
����z#RequirementCommand.get_requirements)rr$cCs |j}|��}|rt�|�dS)zE
        Trace basic information about the provided objects.
        N)�search_scope�get_formatted_locationsre�info)rr��	locationsr,r,r-�trace_basic_info�sz#RequirementCommand.trace_basic_info)r/rQ�
target_pythonr�r$cCs6tj||d�}td|j|j|j|d�}tj|||d�S)z�
        Create a package finder appropriate to this requirement command.

        :param ignore_requires_python: Whether to ignore incompatible
            "Requires-Python" values in links. Defaults to False.
        r{T)�allow_yanked�format_control�allow_all_prereleases�
prefer_binaryr�)�link_collector�selection_prefsr�)r�createrr��prer�r)r)r/rQr�r�r�r�r,r,r-�_build_package_finder�s
��z(RequirementCommand._build_package_finder)N)NFTFFr�NN)NN)rRrSrTrr'�staticmethodrrWr|rVrrrr�boolrrr�rrrXrr�rrr�r�rr�rYr,r,r*r-rs�sn	��?��<�O��rs)GrU�loggingrErc�	functoolsr�optparser�typingrrrr�pip._internal.cacherZpip._internal.clir	�pip._internal.cli.base_commandr
�!pip._internal.cli.command_contextr�pip._internal.exceptionsrr
�pip._internal.index.collectorr�"pip._internal.index.package_finderr�$pip._internal.models.selection_prefsr�"pip._internal.models.target_pythonr�pip._internal.network.sessionr� pip._internal.operations.preparer�pip._internal.req.constructorsrrrrZpip._internal.req.req_filerZpip._internal.req.req_installr�pip._internal.req.req_trackerr�pip._internal.resolution.baser�!pip._internal.self_outdated_checkr�pip._internal.utils.deprecationr�pip._internal.utils.temp_dirrrr �pip._internal.utils.virtualenvr!�	getLoggerrRrer"rZ�	BUILD_ENV�EPHEM_WHEEL_CACHE�	REQ_BUILDrjrgrtrsr,r,r,r-�<module>sH
K�site-packages/pip/_internal/cli/__pycache__/status_codes.cpython-39.pyc000064400000000606150564745010022150 0ustar00a

��Ret�@sdZdZdZdZdZdZdS)������N)�SUCCESS�ERROR�
UNKNOWN_ERROR�VIRTUALENV_NOT_FOUND�PREVIOUS_BUILD_DIR_ERROR�NO_MATCHES_FOUND�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/status_codes.py�<module>s
site-packages/pip/_internal/cli/__pycache__/parser.cpython-39.pyc000064400000023376150564745010020755 0ustar00a

��Re$*�@s�dZddlZddlZddlZddlZddlZddlmZddlm	Z	m
Z
mZmZm
Z
ddlmZddlmZmZddlmZmZe�e�ZGdd	�d	ej�ZGd
d�de�ZGdd
�d
ej�ZGdd�de�ZdS)zBase option parser setup�N)�suppress)�Any�Dict�Iterator�List�Tuple)�
UNKNOWN_ERROR)�
Configuration�ConfigurationError)�redact_auth_from_url�	strtoboolcs�eZdZdZeedd��fdd�Zejed�dd�Z	dejeeed�dd
�Z
eed�dd�Zeed�dd�Zeed�dd�Z
eed�dd�Zeeed�dd�Z�ZS)�PrettyHelpFormatterz4A prettier/less verbose help formatter for optparse.N)�args�kwargs�returncs:d|d<d|d<t��dd|d<t�j|i|��dS)N��max_help_position��indent_incrementr��width)�shutil�get_terminal_size�super�__init__)�selfrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/parser.pyrszPrettyHelpFormatter.__init__��optionrcCs
|�|�S�N)�_format_option_strings)rr!rrr�format_option_stringssz)PrettyHelpFormatter.format_option_strings� <{}>�, )r!�mvarfmt�optseprcCs�g}|jr|�|jd�|jr0|�|jd�t|�dkrH|�d|�|��r�|jdus^J�|jpl|j��}|�|�	|����d�
|�S)z�
        Return a comma-separated list of option strings and metavars.

        :param option:  tuple of (short opt, long opt), e.g: ('-f', '--format')
        :param mvarfmt: metavar format string
        :param optsep:  separator
        rrN�)�_short_opts�append�
_long_opts�len�insert�takes_value�dest�metavar�lower�format�join)rr!r'r(�optsr1rrrr#s
z*PrettyHelpFormatter._format_option_strings)�headingrcCs|dkrdS|dS)N�Optionsr)�:
r)rr6rrr�format_heading9sz"PrettyHelpFormatter.format_heading)�usagercCsd�|�t�|�d��}|S)zz
        Ensure there is only one newline between usage and the first heading
        if there is no description.
        z
Usage: {}
�  )r3�indent_lines�textwrap�dedent)rr:�msgrrr�format_usage>sz PrettyHelpFormatter.format_usage)�descriptionrcCsZ|rRt|jd�rd}nd}|�d�}|��}|�t�|�d�}|�d|�d�}|SdSdS)N�main�Commands�Description�
r;r8r))�hasattr�parser�lstrip�rstripr<r=r>)rrA�labelrrr�format_descriptionFs
z&PrettyHelpFormatter.format_description)�epilogrcCs|r|SdSdS)Nr)r)rrLrrr�
format_epilogXsz!PrettyHelpFormatter.format_epilog)�text�indentrcs"�fdd�|�d�D�}d�|�S)Ncsg|]}�|�qSrr)�.0�line�rOrr�
<listcomp>`�z4PrettyHelpFormatter.indent_lines.<locals>.<listcomp>rE)�splitr4)rrNrO�	new_linesrrRrr<_sz PrettyHelpFormatter.indent_lines)r%r&)�__name__�
__module__�__qualname__�__doc__rr�optparse�Option�strr$r#r9r@rKrMr<�
__classcell__rrrrr
s��r
cs*eZdZdZejed��fdd�Z�ZS)�UpdatingDefaultsHelpFormatterz�Custom help formatter for use in ConfigOptionParser.

    This is updates the defaults before expanding them, allowing
    them to show up correctly in the help listing.

    Also redact auth from url type options
    r cs�d}|jdurLt|jt�sJ�|j�|jj�|jdus<J�|jj�|j�}t��|�}|r�|j	dkr�t|t
�rv|g}t|t�s�g}|D]}|�|t
|��}q�|S)N�URL)rG�
isinstance�ConfigOptionParser�_update_defaults�defaultsr0�getr�expand_defaultr1r]�list�replacer)rr!�default_values�	help_text�valrrrrfms


z,UpdatingDefaultsHelpFormatter.expand_default)	rWrXrYrZr[r\r]rfr^rrrrr_dsr_c@s:eZdZeeeejd�dd�Zee	ej
d�dd��ZdS)�CustomOptionParser)�idxrrrcOs,|j|i|��}|j��|j�||�|S)z*Insert an OptionGroup at a given position.)�add_option_group�
option_groups�popr.)rrmrr�grouprrr�insert_option_group�s
z&CustomOptionParser.insert_option_group�rcCs*|jdd�}|jD]}|�|j�q|S)z<Get a list of all options, including those in option groups.N)�option_listro�extend)r�res�irrr�option_list_all�s
z"CustomOptionParser.option_list_allN)rWrXrY�intrr[�OptionGrouprr�propertyrr\rxrrrrrl�s
�rlcs�eZdZdZdd�eeeedd��fdd�Zej	eeed�d	d
�Z
eeeefd�dd
�Z
eeefeeefd�dd�Zejd�dd�Zedd�dd�Z�ZS)rbzsCustom option parser which updates its defaults by checking the
    configuration files and environmental variablesF)�isolatedN)r�namer|rrcs0||_t|�|_|jsJ�t�j|i|��dSr")r}r	�configrr)rr}r|rrrrrr�s

zConfigOptionParser.__init__)r!�keyrkrc
CsRz|�||�WStjyL}z$td|���t�d�WYd}~n
d}~00dS)Nz(An error occurred during configuration: �)�check_valuer[�OptionValueError�print�sys�exit)rr!rrk�excrrr�
check_default�s
z ConfigOptionParser.check_defaultrsccs�d|jdg}dd�|D�}|j��D]D\}}|s>t�d|�q$|�dd�\}}||vr$||�||f�q$|D] }||D]\}}||fVqzqndS)N�globalz:env:cSsi|]
}|g�qSrr)rPr}rrr�
<dictcomp>�szGConfigOptionParser._get_ordered_configuration_items.<locals>.<dictcomp>z7Ignoring configuration key '%s' as it's value is empty.�.r)r}r~�items�logger�debugrUr+)r�override_order�
section_items�section_keyrk�sectionrrrr� _get_ordered_configuration_items�s"��z3ConfigOptionParser._get_ordered_configuration_items)rdrc
st��j��_t�}���D�]�\�}��d����dur>q�jdusLJ��jdvr�zt	|�}Wn$t
y���d�|���Yn0�nD�jdk�r"t
t
��t	|�}Wd�n1s�0Yt
t
��t|�}Wd�n1s�0Yt|t��r|dk�rЈ�d�|���n��jdk�rN|��}���fd	d
�|D�}n��jdk�rˆjdu�sjJ�|��j����}��||�}�j�p�d}�j�p�i}�j�||�g|�Ri|��n����|�}||�j<q|D]�t�j��|�<�q�d�_|S)
z�Updates the given defaults with values from the config files and
        the environ. Does a little special handling for certain types of
        options (lists).z--N)�
store_true�store_falsezm{} is not a valid value for {} option, please specify a boolean value like yes/no, true/false or 1/0 instead.�countrz�{} is not a valid value for {} option, please instead specify either a non-negative integer or a boolean value like yes/no or false/true which is equivalent to 1/0.r+csg|]}����|��qSr)r�)rP�v�rr!rrrrS�rTz7ConfigOptionParser._update_defaults.<locals>.<listcomp>�callbackr)r[�Valuesrd�values�setr��
get_optionr0�actionr�
ValueError�errorr3rryrarUr��add�get_opt_string�
convert_value�
callback_args�callback_kwargsr��getattr)rrd�	late_evalrk�opt_strrrrr�rrc�sX
��
&
&�� z#ConfigOptionParser._update_defaultsc
Cs�|jst�|j�Sz|j��Wn4tyT}z|�tt	|��WYd}~n
d}~00|�
|j���}|��D]B}|j
dus�J�|�|j
�}t|t	�rn|��}|�||�||j
<qnt�|�S)z�Overriding to make updating the defaults after instantiation of
        the option parser possible, _update_defaults() does the dirty work.N)�process_default_valuesr[r�rdr~�loadr
r�rr]rc�copy�_get_all_optionsr0rerar�r�)r�errrdr!�defaultr�rrr�get_default_valuess&
z%ConfigOptionParser.get_default_values)r?rcCs"|�tj�|�t|�d��dS)NrE)�print_usager��stderrr�r)rr?rrrr�"szConfigOptionParser.error)rWrXrYrZrr]�boolrr[r\r�rrr�rrcr�r�r�r^rrrrrb�s��
 @rb)rZ�loggingr[rr�r=�
contextlibr�typingrrrrr�pip._internal.cli.status_codesr�pip._internal.configurationr	r
�pip._internal.utils.miscrr�	getLoggerrWr��IndentedHelpFormatterr
r_�OptionParserrlrbrrrr�<module>s
R site-packages/pip/_internal/cli/__pycache__/autocompletion.cpython-39.pyc000064400000012060150564745010022507 0ustar00a

��Re��@s�dZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
ddlmZddl
mZmZddlmZdd�d	d
�Ze	eeeee
ed�dd
�Zeeeed�dd�ZdS)zBLogic that powers autocompletion installed by ``pip completion``.
�N)�chain)�Any�Iterable�List�Optional)�create_main_parser)�
commands_dict�create_command)�get_default_environment)�returncs�dtjvrdStjd��dd��ttjd�}z�|d�WntyXd�Yn0t�}tt�}g}d}�D]}||vrt|}q�qt|du�r|dkr�t�	d���
d�o�|d	v}|�rt�}������fd
d�|j
dd
�D�}|�r|D]}	t|	�q�t�	d�t|�}
|
jjD]8}|jtjk�r|j|jD]}|�||jf��q:�qdd��d|d�D���fdd�|D�}�fdd�|D�}t�||
jj�}
|
�r�t�|
�}dd�|D�}|D]>}|d}|d�r�|ddd�dk�r�|d7}t|��q�n�dd�|jD�}|�|j�t�|�}��
d��rf|D]$}|jtjk�r>||j|j7}�q>n t�||�}
|
�r�tt�|
��}td��fdd�|D���t�	d�dS)z:Entry Point for completion of main and subcommand options.�PIP_AUTO_COMPLETEN�
COMP_WORDS��
COMP_CWORD��help�-)�show�	uninstallcs0g|](}|j���r|j�dd�vr|j�qS)rN)�canonical_name�
startswith)�.0�dist)�cwords�lc���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/autocompletion.py�
<listcomp>2s�z autocomplete.<locals>.<listcomp>T)�
local_onlycSsg|]}|�d�d�qS)�=r)�split�r�xrrrrF�cs g|]\}}|�vr||f�qSrr)rr"�v)�	prev_optsrrrGr#cs"g|]\}}|���r||f�qSr�r)r�kr$��currentrrrIr#cSsg|]}|df�qS)rr)r�pathrrrrTr#r�z--rcSsg|]
}|j�qSr)�option_list)r�irrrr^r#� csg|]}|���r|�qSrr&r!r(rrrkr#) �os�environr �int�
IndexErrorr�listr�sys�exitrr
�lower�iter_installed_distributions�printr	�parser�option_list_allr�optparse�
SUPPRESS_HELP�
_long_opts�_short_opts�append�nargs�get_path_completion_type�auto_complete_paths�
option_groupsr,r�
from_iterable�join)�cwordr9�subcommands�options�subcommand_name�word�should_list_installed�env�	installedr�
subcommand�opt�opt_str�completion_type�paths�option�	opt_label�opts�flattened_optsr)r)rrr%r�autocompletes~




�

�
 
rW)rrFrUrcCs�|dks||d�d�sdS|D]n}|jtjkr4q"t|��d�D]L}||d�d�d|krB|jr�tdd�|j�d�D��rB|jSqBq"dS)	aLGet the type of path completion (``file``, ``dir``, ``path`` or None)

    :param cwords: same as the environmental variable ``COMP_WORDS``
    :param cword: same as the environmental variable ``COMP_CWORD``
    :param opts: The available options to check
    :return: path completion type (``file``, ``dir``, ``path`` or None)
    r+rN�/rrcss|]}|dvVqdS))r*�file�dirNrr!rrr�	<genexpr>�sz+get_path_completion_type.<locals>.<genexpr>)rrr;r<�strr �metavar�any)rrFrUrO�orrrrAos

�rA)r)rQrc#s�tj�|�\}�tj�|�}t�|tj�s.dStj�����fdd�t�|�D�}|D]`}tj�||�}tj�tj�||��}|dkr�tj�	|�r�|VqVtj�
|�rVtj�|d�VqVdS)aoIf ``completion_type`` is ``file`` or ``path``, list all regular files
    and directories starting with ``current``; otherwise only list directories
    starting with ``current``.

    :param current: The word to be completed
    :param completion_type: path completion type(`file`, `path` or `dir`)i
    :return: A generator of regular files and/or directories
    Nc3s$|]}tj�|����r|VqdS)N)r/r*�normcaserr!��filenamerrr[�sz&auto_complete_paths.<locals>.<genexpr>rZr)r/r*r �abspath�access�R_OKr`�listdirrE�isfile�isdir)r)rQ�	directory�current_path�	file_list�frO�	comp_filerrarrB�s	
�rB)�__doc__r;r/r4�	itertoolsr�typingrrrr�pip._internal.cli.main_parserr�pip._internal.commandsrr	�pip._internal.metadatar
rWr\r1rArBrrrr�<module>sa�site-packages/pip/_internal/cli/__pycache__/main_parser.cpython-39.pyc000064400000004226150564745010021752 0ustar00a

��Re6
�@s�dZddlZddlZddlmZmZddlmZddlm	Z	m
Z
ddlmZm
Z
ddlmZddlmZmZd	d
gZe	d�dd	�Zeeeeeefd
�dd
�ZdS)z=A single place for constructing and exposing the main parser
�N)�List�Tuple)�
cmdoptions)�ConfigOptionParser�UpdatingDefaultsHelpFormatter)�
commands_dict�get_similar_commands)�CommandError)�get_pip_version�get_prog�create_main_parser�
parse_command)�returncCsltddt�dt�d�}|��t�|_t�tj|�}|�	|�d|_
dgdd�t��D�}d	�
|�|_|S)
z1Creates and returns the main parser for pip's CLIz
%prog <command> [options]F�global)�usage�add_help_option�	formatter�name�progT�cSs"g|]\}}|d�d|j���qS)�27� )�summary)�.0r�command_info�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/main_parser.py�
<listcomp>'s�z&create_main_parser.<locals>.<listcomp>�
)rrr�disable_interspersed_argsr
�versionr�make_option_group�
general_group�add_option_group�mainr�items�join�description)�parser�gen_optsr'rrrrs"�

�)�argsrcCs�t�}|�|�\}}|jr>tj�|j�tj�tj�t��|rZ|ddkrjt	|�dkrj|�
�t��|d}|tvr�t|�}d|�d�g}|r�|�
d|�d��td�|���|dd�}|�|�||fS)Nr�help�zunknown command "�"zmaybe you meant "z - )r�
parse_argsr �sys�stdout�write�os�linesep�exit�len�
print_helprr�appendr	r&�remove)r*r(�general_options�	args_else�cmd_name�guess�msg�cmd_argsrrrr
0s&	
)�__doc__r2r/�typingrrZpip._internal.clir�pip._internal.cli.parserrr�pip._internal.commandsrr�pip._internal.exceptionsr	�pip._internal.utils.miscr
r�__all__r�strr
rrrr�<module>ssite-packages/pip/_internal/cli/__pycache__/__init__.cpython-39.pyc000064400000000467150564745010021214 0ustar00a

��Re��@sdZdS)zGSubpackage containing all of pip's command line interface related code
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/cli/__init__.py�<module>�site-packages/pip/_internal/cli/command_context.py000064400000001370150564745010016342 0ustar00from contextlib import ExitStack, contextmanager
from typing import ContextManager, Iterator, TypeVar

_T = TypeVar("_T", covariant=True)


class CommandContextMixIn:
    def __init__(self) -> None:
        super().__init__()
        self._in_main_context = False
        self._main_context = ExitStack()

    @contextmanager
    def main_context(self) -> Iterator[None]:
        assert not self._in_main_context

        self._in_main_context = True
        try:
            with self._main_context:
                yield
        finally:
            self._in_main_context = False

    def enter_context(self, context_provider: ContextManager[_T]) -> _T:
        assert self._in_main_context

        return self._main_context.enter_context(context_provider)
site-packages/pip/_internal/cli/main_parser.py000064400000005066150564745010015466 0ustar00"""A single place for constructing and exposing the main parser
"""

import os
import sys
from typing import List, Tuple

from pip._internal.cli import cmdoptions
from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter
from pip._internal.commands import commands_dict, get_similar_commands
from pip._internal.exceptions import CommandError
from pip._internal.utils.misc import get_pip_version, get_prog

__all__ = ["create_main_parser", "parse_command"]


def create_main_parser() -> ConfigOptionParser:
    """Creates and returns the main parser for pip's CLI"""

    parser = ConfigOptionParser(
        usage="\n%prog <command> [options]",
        add_help_option=False,
        formatter=UpdatingDefaultsHelpFormatter(),
        name="global",
        prog=get_prog(),
    )
    parser.disable_interspersed_args()

    parser.version = get_pip_version()

    # add the general options
    gen_opts = cmdoptions.make_option_group(cmdoptions.general_group, parser)
    parser.add_option_group(gen_opts)

    # so the help formatter knows
    parser.main = True  # type: ignore

    # create command listing for description
    description = [""] + [
        f"{name:27} {command_info.summary}"
        for name, command_info in commands_dict.items()
    ]
    parser.description = "\n".join(description)

    return parser


def parse_command(args: List[str]) -> Tuple[str, List[str]]:
    parser = create_main_parser()

    # Note: parser calls disable_interspersed_args(), so the result of this
    # call is to split the initial args into the general options before the
    # subcommand and everything else.
    # For example:
    #  args: ['--timeout=5', 'install', '--user', 'INITools']
    #  general_options: ['--timeout==5']
    #  args_else: ['install', '--user', 'INITools']
    general_options, args_else = parser.parse_args(args)

    # --version
    if general_options.version:
        sys.stdout.write(parser.version)
        sys.stdout.write(os.linesep)
        sys.exit()

    # pip || pip help -> print_help()
    if not args_else or (args_else[0] == "help" and len(args_else) == 1):
        parser.print_help()
        sys.exit()

    # the subcommand name
    cmd_name = args_else[0]

    if cmd_name not in commands_dict:
        guess = get_similar_commands(cmd_name)

        msg = [f'unknown command "{cmd_name}"']
        if guess:
            msg.append(f'maybe you meant "{guess}"')

        raise CommandError(" - ".join(msg))

    # all the args without the subcommand
    cmd_args = args[:]
    cmd_args.remove(cmd_name)

    return cmd_name, cmd_args
site-packages/pip/_internal/cli/autocompletion.py000064400000014377150564745010016235 0ustar00"""Logic that powers autocompletion installed by ``pip completion``.
"""

import optparse
import os
import sys
from itertools import chain
from typing import Any, Iterable, List, Optional

from pip._internal.cli.main_parser import create_main_parser
from pip._internal.commands import commands_dict, create_command
from pip._internal.metadata import get_default_environment


def autocomplete() -> None:
    """Entry Point for completion of main and subcommand options."""
    # Don't complete if user hasn't sourced bash_completion file.
    if "PIP_AUTO_COMPLETE" not in os.environ:
        return
    cwords = os.environ["COMP_WORDS"].split()[1:]
    cword = int(os.environ["COMP_CWORD"])
    try:
        current = cwords[cword - 1]
    except IndexError:
        current = ""

    parser = create_main_parser()
    subcommands = list(commands_dict)
    options = []

    # subcommand
    subcommand_name: Optional[str] = None
    for word in cwords:
        if word in subcommands:
            subcommand_name = word
            break
    # subcommand options
    if subcommand_name is not None:
        # special case: 'help' subcommand has no options
        if subcommand_name == "help":
            sys.exit(1)
        # special case: list locally installed dists for show and uninstall
        should_list_installed = not current.startswith("-") and subcommand_name in [
            "show",
            "uninstall",
        ]
        if should_list_installed:
            env = get_default_environment()
            lc = current.lower()
            installed = [
                dist.canonical_name
                for dist in env.iter_installed_distributions(local_only=True)
                if dist.canonical_name.startswith(lc)
                and dist.canonical_name not in cwords[1:]
            ]
            # if there are no dists installed, fall back to option completion
            if installed:
                for dist in installed:
                    print(dist)
                sys.exit(1)

        subcommand = create_command(subcommand_name)

        for opt in subcommand.parser.option_list_all:
            if opt.help != optparse.SUPPRESS_HELP:
                for opt_str in opt._long_opts + opt._short_opts:
                    options.append((opt_str, opt.nargs))

        # filter out previously specified options from available options
        prev_opts = [x.split("=")[0] for x in cwords[1 : cword - 1]]
        options = [(x, v) for (x, v) in options if x not in prev_opts]
        # filter options by current input
        options = [(k, v) for k, v in options if k.startswith(current)]
        # get completion type given cwords and available subcommand options
        completion_type = get_path_completion_type(
            cwords,
            cword,
            subcommand.parser.option_list_all,
        )
        # get completion files and directories if ``completion_type`` is
        # ``<file>``, ``<dir>`` or ``<path>``
        if completion_type:
            paths = auto_complete_paths(current, completion_type)
            options = [(path, 0) for path in paths]
        for option in options:
            opt_label = option[0]
            # append '=' to options which require args
            if option[1] and option[0][:2] == "--":
                opt_label += "="
            print(opt_label)
    else:
        # show main parser options only when necessary

        opts = [i.option_list for i in parser.option_groups]
        opts.append(parser.option_list)
        flattened_opts = chain.from_iterable(opts)
        if current.startswith("-"):
            for opt in flattened_opts:
                if opt.help != optparse.SUPPRESS_HELP:
                    subcommands += opt._long_opts + opt._short_opts
        else:
            # get completion type given cwords and all available options
            completion_type = get_path_completion_type(cwords, cword, flattened_opts)
            if completion_type:
                subcommands = list(auto_complete_paths(current, completion_type))

        print(" ".join([x for x in subcommands if x.startswith(current)]))
    sys.exit(1)


def get_path_completion_type(
    cwords: List[str], cword: int, opts: Iterable[Any]
) -> Optional[str]:
    """Get the type of path completion (``file``, ``dir``, ``path`` or None)

    :param cwords: same as the environmental variable ``COMP_WORDS``
    :param cword: same as the environmental variable ``COMP_CWORD``
    :param opts: The available options to check
    :return: path completion type (``file``, ``dir``, ``path`` or None)
    """
    if cword < 2 or not cwords[cword - 2].startswith("-"):
        return None
    for opt in opts:
        if opt.help == optparse.SUPPRESS_HELP:
            continue
        for o in str(opt).split("/"):
            if cwords[cword - 2].split("=")[0] == o:
                if not opt.metavar or any(
                    x in ("path", "file", "dir") for x in opt.metavar.split("/")
                ):
                    return opt.metavar
    return None


def auto_complete_paths(current: str, completion_type: str) -> Iterable[str]:
    """If ``completion_type`` is ``file`` or ``path``, list all regular files
    and directories starting with ``current``; otherwise only list directories
    starting with ``current``.

    :param current: The word to be completed
    :param completion_type: path completion type(`file`, `path` or `dir`)i
    :return: A generator of regular files and/or directories
    """
    directory, filename = os.path.split(current)
    current_path = os.path.abspath(directory)
    # Don't complete paths if they can't be accessed
    if not os.access(current_path, os.R_OK):
        return
    filename = os.path.normcase(filename)
    # list all files that start with ``filename``
    file_list = (
        x for x in os.listdir(current_path) if os.path.normcase(x).startswith(filename)
    )
    for f in file_list:
        opt = os.path.join(current_path, f)
        comp_file = os.path.normcase(os.path.join(directory, f))
        # complete regular files when there is not ``<dir>`` after option
        # complete directories when there is ``<file>``, ``<path>`` or
        # ``<dir>``after option
        if completion_type != "dir" and os.path.isfile(opt):
            yield comp_file
        elif os.path.isdir(opt):
            yield os.path.join(comp_file, "")
site-packages/pip/_internal/cli/status_codes.py000064400000000164150564745010015660 0ustar00SUCCESS = 0
ERROR = 1
UNKNOWN_ERROR = 2
VIRTUALENV_NOT_FOUND = 3
PREVIOUS_BUILD_DIR_ERROR = 4
NO_MATCHES_FOUND = 23
site-packages/pip/_internal/cli/parser.py000064400000025044150564745010014460 0ustar00"""Base option parser setup"""

import logging
import optparse
import shutil
import sys
import textwrap
from contextlib import suppress
from typing import Any, Dict, Iterator, List, Tuple

from pip._internal.cli.status_codes import UNKNOWN_ERROR
from pip._internal.configuration import Configuration, ConfigurationError
from pip._internal.utils.misc import redact_auth_from_url, strtobool

logger = logging.getLogger(__name__)


class PrettyHelpFormatter(optparse.IndentedHelpFormatter):
    """A prettier/less verbose help formatter for optparse."""

    def __init__(self, *args: Any, **kwargs: Any) -> None:
        # help position must be aligned with __init__.parseopts.description
        kwargs["max_help_position"] = 30
        kwargs["indent_increment"] = 1
        kwargs["width"] = shutil.get_terminal_size()[0] - 2
        super().__init__(*args, **kwargs)

    def format_option_strings(self, option: optparse.Option) -> str:
        return self._format_option_strings(option)

    def _format_option_strings(
        self, option: optparse.Option, mvarfmt: str = " <{}>", optsep: str = ", "
    ) -> str:
        """
        Return a comma-separated list of option strings and metavars.

        :param option:  tuple of (short opt, long opt), e.g: ('-f', '--format')
        :param mvarfmt: metavar format string
        :param optsep:  separator
        """
        opts = []

        if option._short_opts:
            opts.append(option._short_opts[0])
        if option._long_opts:
            opts.append(option._long_opts[0])
        if len(opts) > 1:
            opts.insert(1, optsep)

        if option.takes_value():
            assert option.dest is not None
            metavar = option.metavar or option.dest.lower()
            opts.append(mvarfmt.format(metavar.lower()))

        return "".join(opts)

    def format_heading(self, heading: str) -> str:
        if heading == "Options":
            return ""
        return heading + ":\n"

    def format_usage(self, usage: str) -> str:
        """
        Ensure there is only one newline between usage and the first heading
        if there is no description.
        """
        msg = "\nUsage: {}\n".format(self.indent_lines(textwrap.dedent(usage), "  "))
        return msg

    def format_description(self, description: str) -> str:
        # leave full control over description to us
        if description:
            if hasattr(self.parser, "main"):
                label = "Commands"
            else:
                label = "Description"
            # some doc strings have initial newlines, some don't
            description = description.lstrip("\n")
            # some doc strings have final newlines and spaces, some don't
            description = description.rstrip()
            # dedent, then reindent
            description = self.indent_lines(textwrap.dedent(description), "  ")
            description = f"{label}:\n{description}\n"
            return description
        else:
            return ""

    def format_epilog(self, epilog: str) -> str:
        # leave full control over epilog to us
        if epilog:
            return epilog
        else:
            return ""

    def indent_lines(self, text: str, indent: str) -> str:
        new_lines = [indent + line for line in text.split("\n")]
        return "\n".join(new_lines)


class UpdatingDefaultsHelpFormatter(PrettyHelpFormatter):
    """Custom help formatter for use in ConfigOptionParser.

    This is updates the defaults before expanding them, allowing
    them to show up correctly in the help listing.

    Also redact auth from url type options
    """

    def expand_default(self, option: optparse.Option) -> str:
        default_values = None
        if self.parser is not None:
            assert isinstance(self.parser, ConfigOptionParser)
            self.parser._update_defaults(self.parser.defaults)
            assert option.dest is not None
            default_values = self.parser.defaults.get(option.dest)
        help_text = super().expand_default(option)

        if default_values and option.metavar == "URL":
            if isinstance(default_values, str):
                default_values = [default_values]

            # If its not a list, we should abort and just return the help text
            if not isinstance(default_values, list):
                default_values = []

            for val in default_values:
                help_text = help_text.replace(val, redact_auth_from_url(val))

        return help_text


class CustomOptionParser(optparse.OptionParser):
    def insert_option_group(
        self, idx: int, *args: Any, **kwargs: Any
    ) -> optparse.OptionGroup:
        """Insert an OptionGroup at a given position."""
        group = self.add_option_group(*args, **kwargs)

        self.option_groups.pop()
        self.option_groups.insert(idx, group)

        return group

    @property
    def option_list_all(self) -> List[optparse.Option]:
        """Get a list of all options, including those in option groups."""
        res = self.option_list[:]
        for i in self.option_groups:
            res.extend(i.option_list)

        return res


class ConfigOptionParser(CustomOptionParser):
    """Custom option parser which updates its defaults by checking the
    configuration files and environmental variables"""

    def __init__(
        self,
        *args: Any,
        name: str,
        isolated: bool = False,
        **kwargs: Any,
    ) -> None:
        self.name = name
        self.config = Configuration(isolated)

        assert self.name
        super().__init__(*args, **kwargs)

    def check_default(self, option: optparse.Option, key: str, val: Any) -> Any:
        try:
            return option.check_value(key, val)
        except optparse.OptionValueError as exc:
            print(f"An error occurred during configuration: {exc}")
            sys.exit(3)

    def _get_ordered_configuration_items(self) -> Iterator[Tuple[str, Any]]:
        # Configuration gives keys in an unordered manner. Order them.
        override_order = ["global", self.name, ":env:"]

        # Pool the options into different groups
        section_items: Dict[str, List[Tuple[str, Any]]] = {
            name: [] for name in override_order
        }
        for section_key, val in self.config.items():
            # ignore empty values
            if not val:
                logger.debug(
                    "Ignoring configuration key '%s' as it's value is empty.",
                    section_key,
                )
                continue

            section, key = section_key.split(".", 1)
            if section in override_order:
                section_items[section].append((key, val))

        # Yield each group in their override order
        for section in override_order:
            for key, val in section_items[section]:
                yield key, val

    def _update_defaults(self, defaults: Dict[str, Any]) -> Dict[str, Any]:
        """Updates the given defaults with values from the config files and
        the environ. Does a little special handling for certain types of
        options (lists)."""

        # Accumulate complex default state.
        self.values = optparse.Values(self.defaults)
        late_eval = set()
        # Then set the options with those values
        for key, val in self._get_ordered_configuration_items():
            # '--' because configuration supports only long names
            option = self.get_option("--" + key)

            # Ignore options not present in this parser. E.g. non-globals put
            # in [global] by users that want them to apply to all applicable
            # commands.
            if option is None:
                continue

            assert option.dest is not None

            if option.action in ("store_true", "store_false"):
                try:
                    val = strtobool(val)
                except ValueError:
                    self.error(
                        "{} is not a valid value for {} option, "  # noqa
                        "please specify a boolean value like yes/no, "
                        "true/false or 1/0 instead.".format(val, key)
                    )
            elif option.action == "count":
                with suppress(ValueError):
                    val = strtobool(val)
                with suppress(ValueError):
                    val = int(val)
                if not isinstance(val, int) or val < 0:
                    self.error(
                        "{} is not a valid value for {} option, "  # noqa
                        "please instead specify either a non-negative integer "
                        "or a boolean value like yes/no or false/true "
                        "which is equivalent to 1/0.".format(val, key)
                    )
            elif option.action == "append":
                val = val.split()
                val = [self.check_default(option, key, v) for v in val]
            elif option.action == "callback":
                assert option.callback is not None
                late_eval.add(option.dest)
                opt_str = option.get_opt_string()
                val = option.convert_value(opt_str, val)
                # From take_action
                args = option.callback_args or ()
                kwargs = option.callback_kwargs or {}
                option.callback(option, opt_str, val, self, *args, **kwargs)
            else:
                val = self.check_default(option, key, val)

            defaults[option.dest] = val

        for key in late_eval:
            defaults[key] = getattr(self.values, key)
        self.values = None
        return defaults

    def get_default_values(self) -> optparse.Values:
        """Overriding to make updating the defaults after instantiation of
        the option parser possible, _update_defaults() does the dirty work."""
        if not self.process_default_values:
            # Old, pre-Optik 1.5 behaviour.
            return optparse.Values(self.defaults)

        # Load the configuration, or error out in case of an error
        try:
            self.config.load()
        except ConfigurationError as err:
            self.exit(UNKNOWN_ERROR, str(err))

        defaults = self._update_defaults(self.defaults.copy())  # ours
        for option in self._get_all_options():
            assert option.dest is not None
            default = defaults.get(option.dest)
            if isinstance(default, str):
                opt_str = option.get_opt_string()
                defaults[option.dest] = option.check_value(opt_str, default)
        return optparse.Values(defaults)

    def error(self, msg: str) -> None:
        self.print_usage(sys.stderr)
        self.exit(UNKNOWN_ERROR, f"{msg}\n")
site-packages/pip/_internal/cli/cmdoptions.py000064400000067347150564745010015357 0ustar00"""
shared options and groups

The principle here is to define options once, but *not* instantiate them
globally. One reason being that options with action='append' can carry state
between parses. pip parses general options twice internally, and shouldn't
pass on state. To be consistent, all options will follow this design.
"""

# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import os
import textwrap
import warnings
from functools import partial
from optparse import SUPPRESS_HELP, Option, OptionGroup, OptionParser, Values
from textwrap import dedent
from typing import Any, Callable, Dict, Optional, Tuple

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.cli.parser import ConfigOptionParser
from pip._internal.cli.progress_bars import BAR_TYPES
from pip._internal.exceptions import CommandError
from pip._internal.locations import USER_CACHE_DIR, get_src_prefix
from pip._internal.models.format_control import FormatControl
from pip._internal.models.index import PyPI
from pip._internal.models.target_python import TargetPython
from pip._internal.utils.hashes import STRONG_HASHES
from pip._internal.utils.misc import strtobool


def raise_option_error(parser: OptionParser, option: Option, msg: str) -> None:
    """
    Raise an option parsing error using parser.error().

    Args:
      parser: an OptionParser instance.
      option: an Option instance.
      msg: the error text.
    """
    msg = f"{option} error: {msg}"
    msg = textwrap.fill(" ".join(msg.split()))
    parser.error(msg)


def make_option_group(group: Dict[str, Any], parser: ConfigOptionParser) -> OptionGroup:
    """
    Return an OptionGroup object
    group  -- assumed to be dict with 'name' and 'options' keys
    parser -- an optparse Parser
    """
    option_group = OptionGroup(parser, group["name"])
    for option in group["options"]:
        option_group.add_option(option())
    return option_group


def check_install_build_global(
    options: Values, check_options: Optional[Values] = None
) -> None:
    """Disable wheels if per-setup.py call options are set.

    :param options: The OptionParser options to update.
    :param check_options: The options to check, if not supplied defaults to
        options.
    """
    if check_options is None:
        check_options = options

    def getname(n: str) -> Optional[Any]:
        return getattr(check_options, n, None)

    names = ["build_options", "global_options", "install_options"]
    if any(map(getname, names)):
        control = options.format_control
        control.disallow_binaries()
        warnings.warn(
            "Disabling all use of wheels due to the use of --build-option "
            "/ --global-option / --install-option.",
            stacklevel=2,
        )


def check_dist_restriction(options: Values, check_target: bool = False) -> None:
    """Function for determining if custom platform options are allowed.

    :param options: The OptionParser options.
    :param check_target: Whether or not to check if --target is being used.
    """
    dist_restriction_set = any(
        [
            options.python_version,
            options.platforms,
            options.abis,
            options.implementation,
        ]
    )

    binary_only = FormatControl(set(), {":all:"})
    sdist_dependencies_allowed = (
        options.format_control != binary_only and not options.ignore_dependencies
    )

    # Installations or downloads using dist restrictions must not combine
    # source distributions and dist-specific wheels, as they are not
    # guaranteed to be locally compatible.
    if dist_restriction_set and sdist_dependencies_allowed:
        raise CommandError(
            "When restricting platform and interpreter constraints using "
            "--python-version, --platform, --abi, or --implementation, "
            "either --no-deps must be set, or --only-binary=:all: must be "
            "set and --no-binary must not be set (or must be set to "
            ":none:)."
        )

    if check_target:
        if dist_restriction_set and not options.target_dir:
            raise CommandError(
                "Can not use any platform or abi specific options unless "
                "installing via '--target'"
            )


def _path_option_check(option: Option, opt: str, value: str) -> str:
    return os.path.expanduser(value)


def _package_name_option_check(option: Option, opt: str, value: str) -> str:
    return canonicalize_name(value)


class PipOption(Option):
    TYPES = Option.TYPES + ("path", "package_name")
    TYPE_CHECKER = Option.TYPE_CHECKER.copy()
    TYPE_CHECKER["package_name"] = _package_name_option_check
    TYPE_CHECKER["path"] = _path_option_check


###########
# options #
###########

help_: Callable[..., Option] = partial(
    Option,
    "-h",
    "--help",
    dest="help",
    action="help",
    help="Show help.",
)

debug_mode: Callable[..., Option] = partial(
    Option,
    "--debug",
    dest="debug_mode",
    action="store_true",
    default=False,
    help=(
        "Let unhandled exceptions propagate outside the main subroutine, "
        "instead of logging them to stderr."
    ),
)

isolated_mode: Callable[..., Option] = partial(
    Option,
    "--isolated",
    dest="isolated_mode",
    action="store_true",
    default=False,
    help=(
        "Run pip in an isolated mode, ignoring environment variables and user "
        "configuration."
    ),
)

require_virtualenv: Callable[..., Option] = partial(
    Option,
    # Run only if inside a virtualenv, bail if not.
    "--require-virtualenv",
    "--require-venv",
    dest="require_venv",
    action="store_true",
    default=False,
    help=SUPPRESS_HELP,
)

verbose: Callable[..., Option] = partial(
    Option,
    "-v",
    "--verbose",
    dest="verbose",
    action="count",
    default=0,
    help="Give more output. Option is additive, and can be used up to 3 times.",
)

no_color: Callable[..., Option] = partial(
    Option,
    "--no-color",
    dest="no_color",
    action="store_true",
    default=False,
    help="Suppress colored output.",
)

version: Callable[..., Option] = partial(
    Option,
    "-V",
    "--version",
    dest="version",
    action="store_true",
    help="Show version and exit.",
)

quiet: Callable[..., Option] = partial(
    Option,
    "-q",
    "--quiet",
    dest="quiet",
    action="count",
    default=0,
    help=(
        "Give less output. Option is additive, and can be used up to 3"
        " times (corresponding to WARNING, ERROR, and CRITICAL logging"
        " levels)."
    ),
)

progress_bar: Callable[..., Option] = partial(
    Option,
    "--progress-bar",
    dest="progress_bar",
    type="choice",
    choices=list(BAR_TYPES.keys()),
    default="on",
    help=(
        "Specify type of progress to be displayed ["
        + "|".join(BAR_TYPES.keys())
        + "] (default: %default)"
    ),
)

log: Callable[..., Option] = partial(
    PipOption,
    "--log",
    "--log-file",
    "--local-log",
    dest="log",
    metavar="path",
    type="path",
    help="Path to a verbose appending log.",
)

no_input: Callable[..., Option] = partial(
    Option,
    # Don't ask for input
    "--no-input",
    dest="no_input",
    action="store_true",
    default=False,
    help="Disable prompting for input.",
)

proxy: Callable[..., Option] = partial(
    Option,
    "--proxy",
    dest="proxy",
    type="str",
    default="",
    help="Specify a proxy in the form [user:passwd@]proxy.server:port.",
)

retries: Callable[..., Option] = partial(
    Option,
    "--retries",
    dest="retries",
    type="int",
    default=5,
    help="Maximum number of retries each connection should attempt "
    "(default %default times).",
)

timeout: Callable[..., Option] = partial(
    Option,
    "--timeout",
    "--default-timeout",
    metavar="sec",
    dest="timeout",
    type="float",
    default=15,
    help="Set the socket timeout (default %default seconds).",
)


def exists_action() -> Option:
    return Option(
        # Option when path already exist
        "--exists-action",
        dest="exists_action",
        type="choice",
        choices=["s", "i", "w", "b", "a"],
        default=[],
        action="append",
        metavar="action",
        help="Default action when a path already exists: "
        "(s)witch, (i)gnore, (w)ipe, (b)ackup, (a)bort.",
    )


cert: Callable[..., Option] = partial(
    PipOption,
    "--cert",
    dest="cert",
    type="path",
    metavar="path",
    help=(
        "Path to PEM-encoded CA certificate bundle. "
        "If provided, overrides the default. "
        "See 'SSL Certificate Verification' in pip documentation "
        "for more information."
    ),
)

client_cert: Callable[..., Option] = partial(
    PipOption,
    "--client-cert",
    dest="client_cert",
    type="path",
    default=None,
    metavar="path",
    help="Path to SSL client certificate, a single file containing the "
    "private key and the certificate in PEM format.",
)

index_url: Callable[..., Option] = partial(
    Option,
    "-i",
    "--index-url",
    "--pypi-url",
    dest="index_url",
    metavar="URL",
    default=PyPI.simple_url,
    help="Base URL of the Python Package Index (default %default). "
    "This should point to a repository compliant with PEP 503 "
    "(the simple repository API) or a local directory laid out "
    "in the same format.",
)


def extra_index_url() -> Option:
    return Option(
        "--extra-index-url",
        dest="extra_index_urls",
        metavar="URL",
        action="append",
        default=[],
        help="Extra URLs of package indexes to use in addition to "
        "--index-url. Should follow the same rules as "
        "--index-url.",
    )


no_index: Callable[..., Option] = partial(
    Option,
    "--no-index",
    dest="no_index",
    action="store_true",
    default=False,
    help="Ignore package index (only looking at --find-links URLs instead).",
)


def find_links() -> Option:
    return Option(
        "-f",
        "--find-links",
        dest="find_links",
        action="append",
        default=[],
        metavar="url",
        help="If a URL or path to an html file, then parse for links to "
        "archives such as sdist (.tar.gz) or wheel (.whl) files. "
        "If a local path or file:// URL that's a directory, "
        "then look for archives in the directory listing. "
        "Links to VCS project URLs are not supported.",
    )


def trusted_host() -> Option:
    return Option(
        "--trusted-host",
        dest="trusted_hosts",
        action="append",
        metavar="HOSTNAME",
        default=[],
        help="Mark this host or host:port pair as trusted, even though it "
        "does not have valid or any HTTPS.",
    )


def constraints() -> Option:
    return Option(
        "-c",
        "--constraint",
        dest="constraints",
        action="append",
        default=[],
        metavar="file",
        help="Constrain versions using the given constraints file. "
        "This option can be used multiple times.",
    )


def requirements() -> Option:
    return Option(
        "-r",
        "--requirement",
        dest="requirements",
        action="append",
        default=[],
        metavar="file",
        help="Install from the given requirements file. "
        "This option can be used multiple times.",
    )


def editable() -> Option:
    return Option(
        "-e",
        "--editable",
        dest="editables",
        action="append",
        default=[],
        metavar="path/url",
        help=(
            "Install a project in editable mode (i.e. setuptools "
            '"develop mode") from a local project path or a VCS url.'
        ),
    )


def _handle_src(option: Option, opt_str: str, value: str, parser: OptionParser) -> None:
    value = os.path.abspath(value)
    setattr(parser.values, option.dest, value)


src: Callable[..., Option] = partial(
    PipOption,
    "--src",
    "--source",
    "--source-dir",
    "--source-directory",
    dest="src_dir",
    type="path",
    metavar="dir",
    default=get_src_prefix(),
    action="callback",
    callback=_handle_src,
    help="Directory to check out editable projects into. "
    'The default in a virtualenv is "<venv path>/src". '
    'The default for global installs is "<current dir>/src".',
)


def _get_format_control(values: Values, option: Option) -> Any:
    """Get a format_control object."""
    return getattr(values, option.dest)


def _handle_no_binary(
    option: Option, opt_str: str, value: str, parser: OptionParser
) -> None:
    existing = _get_format_control(parser.values, option)
    FormatControl.handle_mutual_excludes(
        value,
        existing.no_binary,
        existing.only_binary,
    )


def _handle_only_binary(
    option: Option, opt_str: str, value: str, parser: OptionParser
) -> None:
    existing = _get_format_control(parser.values, option)
    FormatControl.handle_mutual_excludes(
        value,
        existing.only_binary,
        existing.no_binary,
    )


def no_binary() -> Option:
    format_control = FormatControl(set(), set())
    return Option(
        "--no-binary",
        dest="format_control",
        action="callback",
        callback=_handle_no_binary,
        type="str",
        default=format_control,
        help="Do not use binary packages. Can be supplied multiple times, and "
        'each time adds to the existing value. Accepts either ":all:" to '
        'disable all binary packages, ":none:" to empty the set (notice '
        "the colons), or one or more package names with commas between "
        "them (no colons). Note that some packages are tricky to compile "
        "and may fail to install when this option is used on them.",
    )


def only_binary() -> Option:
    format_control = FormatControl(set(), set())
    return Option(
        "--only-binary",
        dest="format_control",
        action="callback",
        callback=_handle_only_binary,
        type="str",
        default=format_control,
        help="Do not use source packages. Can be supplied multiple times, and "
        'each time adds to the existing value. Accepts either ":all:" to '
        'disable all source packages, ":none:" to empty the set, or one '
        "or more package names with commas between them. Packages "
        "without binary distributions will fail to install when this "
        "option is used on them.",
    )


platforms: Callable[..., Option] = partial(
    Option,
    "--platform",
    dest="platforms",
    metavar="platform",
    action="append",
    default=None,
    help=(
        "Only use wheels compatible with <platform>. Defaults to the "
        "platform of the running system. Use this option multiple times to "
        "specify multiple platforms supported by the target interpreter."
    ),
)


# This was made a separate function for unit-testing purposes.
def _convert_python_version(value: str) -> Tuple[Tuple[int, ...], Optional[str]]:
    """
    Convert a version string like "3", "37", or "3.7.3" into a tuple of ints.

    :return: A 2-tuple (version_info, error_msg), where `error_msg` is
        non-None if and only if there was a parsing error.
    """
    if not value:
        # The empty string is the same as not providing a value.
        return (None, None)

    parts = value.split(".")
    if len(parts) > 3:
        return ((), "at most three version parts are allowed")

    if len(parts) == 1:
        # Then we are in the case of "3" or "37".
        value = parts[0]
        if len(value) > 1:
            parts = [value[0], value[1:]]

    try:
        version_info = tuple(int(part) for part in parts)
    except ValueError:
        return ((), "each version part must be an integer")

    return (version_info, None)


def _handle_python_version(
    option: Option, opt_str: str, value: str, parser: OptionParser
) -> None:
    """
    Handle a provided --python-version value.
    """
    version_info, error_msg = _convert_python_version(value)
    if error_msg is not None:
        msg = "invalid --python-version value: {!r}: {}".format(
            value,
            error_msg,
        )
        raise_option_error(parser, option=option, msg=msg)

    parser.values.python_version = version_info


python_version: Callable[..., Option] = partial(
    Option,
    "--python-version",
    dest="python_version",
    metavar="python_version",
    action="callback",
    callback=_handle_python_version,
    type="str",
    default=None,
    help=dedent(
        """\
    The Python interpreter version to use for wheel and "Requires-Python"
    compatibility checks. Defaults to a version derived from the running
    interpreter. The version can be specified using up to three dot-separated
    integers (e.g. "3" for 3.0.0, "3.7" for 3.7.0, or "3.7.3"). A major-minor
    version can also be given as a string without dots (e.g. "37" for 3.7.0).
    """
    ),
)


implementation: Callable[..., Option] = partial(
    Option,
    "--implementation",
    dest="implementation",
    metavar="implementation",
    default=None,
    help=(
        "Only use wheels compatible with Python "
        "implementation <implementation>, e.g. 'pp', 'jy', 'cp', "
        " or 'ip'. If not specified, then the current "
        "interpreter implementation is used.  Use 'py' to force "
        "implementation-agnostic wheels."
    ),
)


abis: Callable[..., Option] = partial(
    Option,
    "--abi",
    dest="abis",
    metavar="abi",
    action="append",
    default=None,
    help=(
        "Only use wheels compatible with Python abi <abi>, e.g. 'pypy_41'. "
        "If not specified, then the current interpreter abi tag is used. "
        "Use this option multiple times to specify multiple abis supported "
        "by the target interpreter. Generally you will need to specify "
        "--implementation, --platform, and --python-version when using this "
        "option."
    ),
)


def add_target_python_options(cmd_opts: OptionGroup) -> None:
    cmd_opts.add_option(platforms())
    cmd_opts.add_option(python_version())
    cmd_opts.add_option(implementation())
    cmd_opts.add_option(abis())


def make_target_python(options: Values) -> TargetPython:
    target_python = TargetPython(
        platforms=options.platforms,
        py_version_info=options.python_version,
        abis=options.abis,
        implementation=options.implementation,
    )

    return target_python


def prefer_binary() -> Option:
    return Option(
        "--prefer-binary",
        dest="prefer_binary",
        action="store_true",
        default=False,
        help="Prefer older binary packages over newer source packages.",
    )


cache_dir: Callable[..., Option] = partial(
    PipOption,
    "--cache-dir",
    dest="cache_dir",
    default=USER_CACHE_DIR,
    metavar="dir",
    type="path",
    help="Store the cache data in <dir>.",
)


def _handle_no_cache_dir(
    option: Option, opt: str, value: str, parser: OptionParser
) -> None:
    """
    Process a value provided for the --no-cache-dir option.

    This is an optparse.Option callback for the --no-cache-dir option.
    """
    # The value argument will be None if --no-cache-dir is passed via the
    # command-line, since the option doesn't accept arguments.  However,
    # the value can be non-None if the option is triggered e.g. by an
    # environment variable, like PIP_NO_CACHE_DIR=true.
    if value is not None:
        # Then parse the string value to get argument error-checking.
        try:
            strtobool(value)
        except ValueError as exc:
            raise_option_error(parser, option=option, msg=str(exc))

    # Originally, setting PIP_NO_CACHE_DIR to a value that strtobool()
    # converted to 0 (like "false" or "no") caused cache_dir to be disabled
    # rather than enabled (logic would say the latter).  Thus, we disable
    # the cache directory not just on values that parse to True, but (for
    # backwards compatibility reasons) also on values that parse to False.
    # In other words, always set it to False if the option is provided in
    # some (valid) form.
    parser.values.cache_dir = False


no_cache: Callable[..., Option] = partial(
    Option,
    "--no-cache-dir",
    dest="cache_dir",
    action="callback",
    callback=_handle_no_cache_dir,
    help="Disable the cache.",
)

no_deps: Callable[..., Option] = partial(
    Option,
    "--no-deps",
    "--no-dependencies",
    dest="ignore_dependencies",
    action="store_true",
    default=False,
    help="Don't install package dependencies.",
)

ignore_requires_python: Callable[..., Option] = partial(
    Option,
    "--ignore-requires-python",
    dest="ignore_requires_python",
    action="store_true",
    help="Ignore the Requires-Python information.",
)

no_build_isolation: Callable[..., Option] = partial(
    Option,
    "--no-build-isolation",
    dest="build_isolation",
    action="store_false",
    default=True,
    help="Disable isolation when building a modern source distribution. "
    "Build dependencies specified by PEP 518 must be already installed "
    "if this option is used.",
)


def _handle_no_use_pep517(
    option: Option, opt: str, value: str, parser: OptionParser
) -> None:
    """
    Process a value provided for the --no-use-pep517 option.

    This is an optparse.Option callback for the no_use_pep517 option.
    """
    # Since --no-use-pep517 doesn't accept arguments, the value argument
    # will be None if --no-use-pep517 is passed via the command-line.
    # However, the value can be non-None if the option is triggered e.g.
    # by an environment variable, for example "PIP_NO_USE_PEP517=true".
    if value is not None:
        msg = """A value was passed for --no-use-pep517,
        probably using either the PIP_NO_USE_PEP517 environment variable
        or the "no-use-pep517" config file option. Use an appropriate value
        of the PIP_USE_PEP517 environment variable or the "use-pep517"
        config file option instead.
        """
        raise_option_error(parser, option=option, msg=msg)

    # Otherwise, --no-use-pep517 was passed via the command-line.
    parser.values.use_pep517 = False


use_pep517: Any = partial(
    Option,
    "--use-pep517",
    dest="use_pep517",
    action="store_true",
    default=None,
    help="Use PEP 517 for building source distributions "
    "(use --no-use-pep517 to force legacy behaviour).",
)

no_use_pep517: Any = partial(
    Option,
    "--no-use-pep517",
    dest="use_pep517",
    action="callback",
    callback=_handle_no_use_pep517,
    default=None,
    help=SUPPRESS_HELP,
)

install_options: Callable[..., Option] = partial(
    Option,
    "--install-option",
    dest="install_options",
    action="append",
    metavar="options",
    help="Extra arguments to be supplied to the setup.py install "
    'command (use like --install-option="--install-scripts=/usr/local/'
    'bin"). Use multiple --install-option options to pass multiple '
    "options to setup.py install. If you are using an option with a "
    "directory path, be sure to use absolute path.",
)

build_options: Callable[..., Option] = partial(
    Option,
    "--build-option",
    dest="build_options",
    metavar="options",
    action="append",
    help="Extra arguments to be supplied to 'setup.py bdist_wheel'.",
)

global_options: Callable[..., Option] = partial(
    Option,
    "--global-option",
    dest="global_options",
    action="append",
    metavar="options",
    help="Extra global options to be supplied to the setup.py "
    "call before the install or bdist_wheel command.",
)

no_clean: Callable[..., Option] = partial(
    Option,
    "--no-clean",
    action="store_true",
    default=False,
    help="Don't clean up build directories.",
)

pre: Callable[..., Option] = partial(
    Option,
    "--pre",
    action="store_true",
    default=False,
    help="Include pre-release and development versions. By default, "
    "pip only finds stable versions.",
)

disable_pip_version_check: Callable[..., Option] = partial(
    Option,
    "--disable-pip-version-check",
    dest="disable_pip_version_check",
    action="store_true",
    default=False,
    help="Don't periodically check PyPI to determine whether a new version "
    "of pip is available for download. Implied with --no-index.",
)


def _handle_merge_hash(
    option: Option, opt_str: str, value: str, parser: OptionParser
) -> None:
    """Given a value spelled "algo:digest", append the digest to a list
    pointed to in a dict by the algo name."""
    if not parser.values.hashes:
        parser.values.hashes = {}
    try:
        algo, digest = value.split(":", 1)
    except ValueError:
        parser.error(
            "Arguments to {} must be a hash name "  # noqa
            "followed by a value, like --hash=sha256:"
            "abcde...".format(opt_str)
        )
    if algo not in STRONG_HASHES:
        parser.error(
            "Allowed hash algorithms for {} are {}.".format(  # noqa
                opt_str, ", ".join(STRONG_HASHES)
            )
        )
    parser.values.hashes.setdefault(algo, []).append(digest)


hash: Callable[..., Option] = partial(
    Option,
    "--hash",
    # Hash values eventually end up in InstallRequirement.hashes due to
    # __dict__ copying in process_line().
    dest="hashes",
    action="callback",
    callback=_handle_merge_hash,
    type="string",
    help="Verify that the package's archive matches this "
    "hash before installing. Example: --hash=sha256:abcdef...",
)


require_hashes: Callable[..., Option] = partial(
    Option,
    "--require-hashes",
    dest="require_hashes",
    action="store_true",
    default=False,
    help="Require a hash to check each requirement against, for "
    "repeatable installs. This option is implied when any package in a "
    "requirements file has a --hash option.",
)


list_path: Callable[..., Option] = partial(
    PipOption,
    "--path",
    dest="path",
    type="path",
    action="append",
    help="Restrict to the specified installation path for listing "
    "packages (can be used multiple times).",
)


def check_list_path_option(options: Values) -> None:
    if options.path and (options.user or options.local):
        raise CommandError("Cannot combine '--path' with '--user' or '--local'")


list_exclude: Callable[..., Option] = partial(
    PipOption,
    "--exclude",
    dest="excludes",
    action="append",
    metavar="package",
    type="package_name",
    help="Exclude specified package from the output",
)


no_python_version_warning: Callable[..., Option] = partial(
    Option,
    "--no-python-version-warning",
    dest="no_python_version_warning",
    action="store_true",
    default=False,
    help="Silence deprecation warnings for upcoming unsupported Pythons.",
)


use_new_feature: Callable[..., Option] = partial(
    Option,
    "--use-feature",
    dest="features_enabled",
    metavar="feature",
    action="append",
    default=[],
    choices=["2020-resolver", "fast-deps", "in-tree-build"],
    help="Enable new functionality, that may be backward incompatible.",
)

use_deprecated_feature: Callable[..., Option] = partial(
    Option,
    "--use-deprecated",
    dest="deprecated_features_enabled",
    metavar="feature",
    action="append",
    default=[],
    choices=["legacy-resolver", "out-of-tree-build"],
    help=("Enable deprecated functionality, that will be removed in the future."),
)


##########
# groups #
##########

general_group: Dict[str, Any] = {
    "name": "General Options",
    "options": [
        help_,
        debug_mode,
        isolated_mode,
        require_virtualenv,
        verbose,
        version,
        quiet,
        log,
        no_input,
        proxy,
        retries,
        timeout,
        exists_action,
        trusted_host,
        cert,
        client_cert,
        cache_dir,
        no_cache,
        disable_pip_version_check,
        no_color,
        no_python_version_warning,
        use_new_feature,
        use_deprecated_feature,
    ],
}

index_group: Dict[str, Any] = {
    "name": "Package Index Options",
    "options": [
        index_url,
        extra_index_url,
        no_index,
        find_links,
    ],
}
site-packages/pip/_internal/cli/req_command.py000064400000041311150564745010015444 0ustar00"""Contains the Command base classes that depend on PipSession.

The classes in this module are in a separate module so the commands not
needing download / PackageFinder capability don't unnecessarily import the
PackageFinder machinery and all its vendored dependencies, etc.
"""

import logging
import os
import sys
from functools import partial
from optparse import Values
from typing import Any, List, Optional, Tuple

from pip._internal.cache import WheelCache
from pip._internal.cli import cmdoptions
from pip._internal.cli.base_command import Command
from pip._internal.cli.command_context import CommandContextMixIn
from pip._internal.exceptions import CommandError, PreviousBuildDirError
from pip._internal.index.collector import LinkCollector
from pip._internal.index.package_finder import PackageFinder
from pip._internal.models.selection_prefs import SelectionPreferences
from pip._internal.models.target_python import TargetPython
from pip._internal.network.session import PipSession
from pip._internal.operations.prepare import RequirementPreparer
from pip._internal.req.constructors import (
    install_req_from_editable,
    install_req_from_line,
    install_req_from_parsed_requirement,
    install_req_from_req_string,
)
from pip._internal.req.req_file import parse_requirements
from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_tracker import RequirementTracker
from pip._internal.resolution.base import BaseResolver
from pip._internal.self_outdated_check import pip_self_version_check
from pip._internal.utils.deprecation import deprecated
from pip._internal.utils.temp_dir import (
    TempDirectory,
    TempDirectoryTypeRegistry,
    tempdir_kinds,
)
from pip._internal.utils.virtualenv import running_under_virtualenv

logger = logging.getLogger(__name__)


class SessionCommandMixin(CommandContextMixIn):

    """
    A class mixin for command classes needing _build_session().
    """

    def __init__(self) -> None:
        super().__init__()
        self._session: Optional[PipSession] = None

    @classmethod
    def _get_index_urls(cls, options: Values) -> Optional[List[str]]:
        """Return a list of index urls from user-provided options."""
        index_urls = []
        if not getattr(options, "no_index", False):
            url = getattr(options, "index_url", None)
            if url:
                index_urls.append(url)
        urls = getattr(options, "extra_index_urls", None)
        if urls:
            index_urls.extend(urls)
        # Return None rather than an empty list
        return index_urls or None

    def get_default_session(self, options: Values) -> PipSession:
        """Get a default-managed session."""
        if self._session is None:
            self._session = self.enter_context(self._build_session(options))
            # there's no type annotation on requests.Session, so it's
            # automatically ContextManager[Any] and self._session becomes Any,
            # then https://github.com/python/mypy/issues/7696 kicks in
            assert self._session is not None
        return self._session

    def _build_session(
        self,
        options: Values,
        retries: Optional[int] = None,
        timeout: Optional[int] = None,
    ) -> PipSession:
        assert not options.cache_dir or os.path.isabs(options.cache_dir)
        session = PipSession(
            cache=(
                os.path.join(options.cache_dir, "http") if options.cache_dir else None
            ),
            retries=retries if retries is not None else options.retries,
            trusted_hosts=options.trusted_hosts,
            index_urls=self._get_index_urls(options),
        )

        # Handle custom ca-bundles from the user
        if options.cert:
            session.verify = options.cert

        # Handle SSL client certificate
        if options.client_cert:
            session.cert = options.client_cert

        # Handle timeouts
        if options.timeout or timeout:
            session.timeout = timeout if timeout is not None else options.timeout

        # Handle configured proxies
        if options.proxy:
            session.proxies = {
                "http": options.proxy,
                "https": options.proxy,
            }

        # Determine if we can prompt the user for authentication or not
        session.auth.prompting = not options.no_input

        return session


class IndexGroupCommand(Command, SessionCommandMixin):

    """
    Abstract base class for commands with the index_group options.

    This also corresponds to the commands that permit the pip version check.
    """

    def handle_pip_version_check(self, options: Values) -> None:
        """
        Do the pip version check if not disabled.

        This overrides the default behavior of not doing the check.
        """
        # Make sure the index_group options are present.
        assert hasattr(options, "no_index")

        if options.disable_pip_version_check or options.no_index:
            return

        # Otherwise, check if we're using the latest version of pip available.
        session = self._build_session(
            options, retries=0, timeout=min(5, options.timeout)
        )
        with session:
            pip_self_version_check(session, options)


KEEPABLE_TEMPDIR_TYPES = [
    tempdir_kinds.BUILD_ENV,
    tempdir_kinds.EPHEM_WHEEL_CACHE,
    tempdir_kinds.REQ_BUILD,
]


def warn_if_run_as_root() -> None:
    """Output a warning for sudo users on Unix.

    In a virtual environment, sudo pip still writes to virtualenv.
    On Windows, users may run pip as Administrator without issues.
    This warning only applies to Unix root users outside of virtualenv.
    """
    if running_under_virtualenv():
        return
    if not hasattr(os, "getuid"):
        return
    # On Windows, there are no "system managed" Python packages. Installing as
    # Administrator via pip is the correct way of updating system environments.
    #
    # We choose sys.platform over utils.compat.WINDOWS here to enable Mypy platform
    # checks: https://mypy.readthedocs.io/en/stable/common_issues.html
    if sys.platform == "win32" or sys.platform == "cygwin":
        return

    if os.getuid() != 0:
        return

    logger.warning(
        "Running pip as the 'root' user can result in broken permissions and "
        "conflicting behaviour with the system package manager. "
        "It is recommended to use a virtual environment instead: "
        "https://pip.pypa.io/warnings/venv"
    )


def with_cleanup(func: Any) -> Any:
    """Decorator for common logic related to managing temporary
    directories.
    """

    def configure_tempdir_registry(registry: TempDirectoryTypeRegistry) -> None:
        for t in KEEPABLE_TEMPDIR_TYPES:
            registry.set_delete(t, False)

    def wrapper(
        self: RequirementCommand, options: Values, args: List[Any]
    ) -> Optional[int]:
        assert self.tempdir_registry is not None
        if options.no_clean:
            configure_tempdir_registry(self.tempdir_registry)

        try:
            return func(self, options, args)
        except PreviousBuildDirError:
            # This kind of conflict can occur when the user passes an explicit
            # build directory with a pre-existing folder. In that case we do
            # not want to accidentally remove it.
            configure_tempdir_registry(self.tempdir_registry)
            raise

    return wrapper


class RequirementCommand(IndexGroupCommand):
    def __init__(self, *args: Any, **kw: Any) -> None:
        super().__init__(*args, **kw)

        self.cmd_opts.add_option(cmdoptions.no_clean())

    @staticmethod
    def determine_resolver_variant(options: Values) -> str:
        """Determines which resolver should be used, based on the given options."""
        if "legacy-resolver" in options.deprecated_features_enabled:
            return "legacy"

        return "2020-resolver"

    @classmethod
    def make_requirement_preparer(
        cls,
        temp_build_dir: TempDirectory,
        options: Values,
        req_tracker: RequirementTracker,
        session: PipSession,
        finder: PackageFinder,
        use_user_site: bool,
        download_dir: Optional[str] = None,
    ) -> RequirementPreparer:
        """
        Create a RequirementPreparer instance for the given parameters.
        """
        temp_build_dir_path = temp_build_dir.path
        assert temp_build_dir_path is not None

        resolver_variant = cls.determine_resolver_variant(options)
        if resolver_variant == "2020-resolver":
            lazy_wheel = "fast-deps" in options.features_enabled
            if lazy_wheel:
                logger.warning(
                    "pip is using lazily downloaded wheels using HTTP "
                    "range requests to obtain dependency information. "
                    "This experimental feature is enabled through "
                    "--use-feature=fast-deps and it is not ready for "
                    "production."
                )
        else:
            lazy_wheel = False
            if "fast-deps" in options.features_enabled:
                logger.warning(
                    "fast-deps has no effect when used with the legacy resolver."
                )

        in_tree_build = "out-of-tree-build" not in options.deprecated_features_enabled
        if "in-tree-build" in options.features_enabled:
            deprecated(
                reason="In-tree builds are now the default.",
                replacement="to remove the --use-feature=in-tree-build flag",
                gone_in="22.1",
            )
        if "out-of-tree-build" in options.deprecated_features_enabled:
            deprecated(
                reason="Out-of-tree builds are deprecated.",
                replacement=None,
                gone_in="22.1",
            )

        return RequirementPreparer(
            build_dir=temp_build_dir_path,
            src_dir=options.src_dir,
            download_dir=download_dir,
            build_isolation=options.build_isolation,
            req_tracker=req_tracker,
            session=session,
            progress_bar=options.progress_bar,
            finder=finder,
            require_hashes=options.require_hashes,
            use_user_site=use_user_site,
            lazy_wheel=lazy_wheel,
            in_tree_build=in_tree_build,
        )

    @classmethod
    def make_resolver(
        cls,
        preparer: RequirementPreparer,
        finder: PackageFinder,
        options: Values,
        wheel_cache: Optional[WheelCache] = None,
        use_user_site: bool = False,
        ignore_installed: bool = True,
        ignore_requires_python: bool = False,
        force_reinstall: bool = False,
        upgrade_strategy: str = "to-satisfy-only",
        use_pep517: Optional[bool] = None,
        py_version_info: Optional[Tuple[int, ...]] = None,
    ) -> BaseResolver:
        """
        Create a Resolver instance for the given parameters.
        """
        make_install_req = partial(
            install_req_from_req_string,
            isolated=options.isolated_mode,
            use_pep517=use_pep517,
        )
        resolver_variant = cls.determine_resolver_variant(options)
        # The long import name and duplicated invocation is needed to convince
        # Mypy into correctly typechecking. Otherwise it would complain the
        # "Resolver" class being redefined.
        if resolver_variant == "2020-resolver":
            import pip._internal.resolution.resolvelib.resolver

            return pip._internal.resolution.resolvelib.resolver.Resolver(
                preparer=preparer,
                finder=finder,
                wheel_cache=wheel_cache,
                make_install_req=make_install_req,
                use_user_site=use_user_site,
                ignore_dependencies=options.ignore_dependencies,
                ignore_installed=ignore_installed,
                ignore_requires_python=ignore_requires_python,
                force_reinstall=force_reinstall,
                upgrade_strategy=upgrade_strategy,
                py_version_info=py_version_info,
            )
        import pip._internal.resolution.legacy.resolver

        return pip._internal.resolution.legacy.resolver.Resolver(
            preparer=preparer,
            finder=finder,
            wheel_cache=wheel_cache,
            make_install_req=make_install_req,
            use_user_site=use_user_site,
            ignore_dependencies=options.ignore_dependencies,
            ignore_installed=ignore_installed,
            ignore_requires_python=ignore_requires_python,
            force_reinstall=force_reinstall,
            upgrade_strategy=upgrade_strategy,
            py_version_info=py_version_info,
        )

    def get_requirements(
        self,
        args: List[str],
        options: Values,
        finder: PackageFinder,
        session: PipSession,
    ) -> List[InstallRequirement]:
        """
        Parse command-line arguments into the corresponding requirements.
        """
        requirements: List[InstallRequirement] = []
        for filename in options.constraints:
            for parsed_req in parse_requirements(
                filename,
                constraint=True,
                finder=finder,
                options=options,
                session=session,
            ):
                req_to_add = install_req_from_parsed_requirement(
                    parsed_req,
                    isolated=options.isolated_mode,
                    user_supplied=False,
                )
                requirements.append(req_to_add)

        for req in args:
            req_to_add = install_req_from_line(
                req,
                None,
                isolated=options.isolated_mode,
                use_pep517=options.use_pep517,
                user_supplied=True,
            )
            requirements.append(req_to_add)

        for req in options.editables:
            req_to_add = install_req_from_editable(
                req,
                user_supplied=True,
                isolated=options.isolated_mode,
                use_pep517=options.use_pep517,
            )
            requirements.append(req_to_add)

        # NOTE: options.require_hashes may be set if --require-hashes is True
        for filename in options.requirements:
            for parsed_req in parse_requirements(
                filename, finder=finder, options=options, session=session
            ):
                req_to_add = install_req_from_parsed_requirement(
                    parsed_req,
                    isolated=options.isolated_mode,
                    use_pep517=options.use_pep517,
                    user_supplied=True,
                )
                requirements.append(req_to_add)

        # If any requirement has hash options, enable hash checking.
        if any(req.has_hash_options for req in requirements):
            options.require_hashes = True

        if not (args or options.editables or options.requirements):
            opts = {"name": self.name}
            if options.find_links:
                raise CommandError(
                    "You must give at least one requirement to {name} "
                    '(maybe you meant "pip {name} {links}"?)'.format(
                        **dict(opts, links=" ".join(options.find_links))
                    )
                )
            else:
                raise CommandError(
                    "You must give at least one requirement to {name} "
                    '(see "pip help {name}")'.format(**opts)
                )

        return requirements

    @staticmethod
    def trace_basic_info(finder: PackageFinder) -> None:
        """
        Trace basic information about the provided objects.
        """
        # Display where finder is looking for packages
        search_scope = finder.search_scope
        locations = search_scope.get_formatted_locations()
        if locations:
            logger.info(locations)

    def _build_package_finder(
        self,
        options: Values,
        session: PipSession,
        target_python: Optional[TargetPython] = None,
        ignore_requires_python: Optional[bool] = None,
    ) -> PackageFinder:
        """
        Create a package finder appropriate to this requirement command.

        :param ignore_requires_python: Whether to ignore incompatible
            "Requires-Python" values in links. Defaults to False.
        """
        link_collector = LinkCollector.create(session, options=options)
        selection_prefs = SelectionPreferences(
            allow_yanked=True,
            format_control=options.format_control,
            allow_all_prereleases=options.pre,
            prefer_binary=options.prefer_binary,
            ignore_requires_python=ignore_requires_python,
        )

        return PackageFinder.create(
            link_collector=link_collector,
            selection_prefs=selection_prefs,
            target_python=target_python,
        )
site-packages/pip/_internal/cli/progress_bars.py000064400000020154150564745010016034 0ustar00import itertools
import sys
from signal import SIGINT, default_int_handler, signal
from typing import Any

from pip._vendor.progress.bar import Bar, FillingCirclesBar, IncrementalBar
from pip._vendor.progress.spinner import Spinner

from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.logging import get_indentation
from pip._internal.utils.misc import format_size

try:
    from pip._vendor import colorama
# Lots of different errors can come from this, including SystemError and
# ImportError.
except Exception:
    colorama = None


def _select_progress_class(preferred: Bar, fallback: Bar) -> Bar:
    encoding = getattr(preferred.file, "encoding", None)

    # If we don't know what encoding this file is in, then we'll just assume
    # that it doesn't support unicode and use the ASCII bar.
    if not encoding:
        return fallback

    # Collect all of the possible characters we want to use with the preferred
    # bar.
    characters = [
        getattr(preferred, "empty_fill", ""),
        getattr(preferred, "fill", ""),
    ]
    characters += list(getattr(preferred, "phases", []))

    # Try to decode the characters we're using for the bar using the encoding
    # of the given file, if this works then we'll assume that we can use the
    # fancier bar and if not we'll fall back to the plaintext bar.
    try:
        "".join(characters).encode(encoding)
    except UnicodeEncodeError:
        return fallback
    else:
        return preferred


_BaseBar: Any = _select_progress_class(IncrementalBar, Bar)


class InterruptibleMixin:
    """
    Helper to ensure that self.finish() gets called on keyboard interrupt.

    This allows downloads to be interrupted without leaving temporary state
    (like hidden cursors) behind.

    This class is similar to the progress library's existing SigIntMixin
    helper, but as of version 1.2, that helper has the following problems:

    1. It calls sys.exit().
    2. It discards the existing SIGINT handler completely.
    3. It leaves its own handler in place even after an uninterrupted finish,
       which will have unexpected delayed effects if the user triggers an
       unrelated keyboard interrupt some time after a progress-displaying
       download has already completed, for example.
    """

    def __init__(self, *args: Any, **kwargs: Any) -> None:
        """
        Save the original SIGINT handler for later.
        """
        # https://github.com/python/mypy/issues/5887
        super().__init__(*args, **kwargs)  # type: ignore

        self.original_handler = signal(SIGINT, self.handle_sigint)

        # If signal() returns None, the previous handler was not installed from
        # Python, and we cannot restore it. This probably should not happen,
        # but if it does, we must restore something sensible instead, at least.
        # The least bad option should be Python's default SIGINT handler, which
        # just raises KeyboardInterrupt.
        if self.original_handler is None:
            self.original_handler = default_int_handler

    def finish(self) -> None:
        """
        Restore the original SIGINT handler after finishing.

        This should happen regardless of whether the progress display finishes
        normally, or gets interrupted.
        """
        super().finish()  # type: ignore
        signal(SIGINT, self.original_handler)

    def handle_sigint(self, signum, frame):  # type: ignore
        """
        Call self.finish() before delegating to the original SIGINT handler.

        This handler should only be in place while the progress display is
        active.
        """
        self.finish()
        self.original_handler(signum, frame)


class SilentBar(Bar):
    def update(self) -> None:
        pass


class BlueEmojiBar(IncrementalBar):

    suffix = "%(percent)d%%"
    bar_prefix = " "
    bar_suffix = " "
    phases = ("\U0001F539", "\U0001F537", "\U0001F535")


class DownloadProgressMixin:
    def __init__(self, *args: Any, **kwargs: Any) -> None:
        # https://github.com/python/mypy/issues/5887
        super().__init__(*args, **kwargs)  # type: ignore
        self.message: str = (" " * (get_indentation() + 2)) + self.message

    @property
    def downloaded(self) -> str:
        return format_size(self.index)  # type: ignore

    @property
    def download_speed(self) -> str:
        # Avoid zero division errors...
        if self.avg == 0.0:  # type: ignore
            return "..."
        return format_size(1 / self.avg) + "/s"  # type: ignore

    @property
    def pretty_eta(self) -> str:
        if self.eta:  # type: ignore
            return f"eta {self.eta_td}"  # type: ignore
        return ""

    def iter(self, it):  # type: ignore
        for x in it:
            yield x
            # B305 is incorrectly raised here
            # https://github.com/PyCQA/flake8-bugbear/issues/59
            self.next(len(x))  # noqa: B305
        self.finish()


class WindowsMixin:
    def __init__(self, *args: Any, **kwargs: Any) -> None:
        # The Windows terminal does not support the hide/show cursor ANSI codes
        # even with colorama. So we'll ensure that hide_cursor is False on
        # Windows.
        # This call needs to go before the super() call, so that hide_cursor
        # is set in time. The base progress bar class writes the "hide cursor"
        # code to the terminal in its init, so if we don't set this soon
        # enough, we get a "hide" with no corresponding "show"...
        if WINDOWS and self.hide_cursor:  # type: ignore
            self.hide_cursor = False

        # https://github.com/python/mypy/issues/5887
        super().__init__(*args, **kwargs)  # type: ignore

        # Check if we are running on Windows and we have the colorama module,
        # if we do then wrap our file with it.
        if WINDOWS and colorama:
            self.file = colorama.AnsiToWin32(self.file)  # type: ignore
            # The progress code expects to be able to call self.file.isatty()
            # but the colorama.AnsiToWin32() object doesn't have that, so we'll
            # add it.
            self.file.isatty = lambda: self.file.wrapped.isatty()
            # The progress code expects to be able to call self.file.flush()
            # but the colorama.AnsiToWin32() object doesn't have that, so we'll
            # add it.
            self.file.flush = lambda: self.file.wrapped.flush()


class BaseDownloadProgressBar(WindowsMixin, InterruptibleMixin, DownloadProgressMixin):

    file = sys.stdout
    message = "%(percent)d%%"
    suffix = "%(downloaded)s %(download_speed)s %(pretty_eta)s"


class DefaultDownloadProgressBar(BaseDownloadProgressBar, _BaseBar):
    pass


class DownloadSilentBar(BaseDownloadProgressBar, SilentBar):
    pass


class DownloadBar(BaseDownloadProgressBar, Bar):
    pass


class DownloadFillingCirclesBar(BaseDownloadProgressBar, FillingCirclesBar):
    pass


class DownloadBlueEmojiProgressBar(BaseDownloadProgressBar, BlueEmojiBar):
    pass


class DownloadProgressSpinner(
    WindowsMixin, InterruptibleMixin, DownloadProgressMixin, Spinner
):

    file = sys.stdout
    suffix = "%(downloaded)s %(download_speed)s"

    def next_phase(self) -> str:
        if not hasattr(self, "_phaser"):
            self._phaser = itertools.cycle(self.phases)
        return next(self._phaser)

    def update(self) -> None:
        message = self.message % self
        phase = self.next_phase()
        suffix = self.suffix % self
        line = "".join(
            [
                message,
                " " if message else "",
                phase,
                " " if suffix else "",
                suffix,
            ]
        )

        self.writeln(line)


BAR_TYPES = {
    "off": (DownloadSilentBar, DownloadSilentBar),
    "on": (DefaultDownloadProgressBar, DownloadProgressSpinner),
    "ascii": (DownloadBar, DownloadProgressSpinner),
    "pretty": (DownloadFillingCirclesBar, DownloadProgressSpinner),
    "emoji": (DownloadBlueEmojiProgressBar, DownloadProgressSpinner),
}


def DownloadProgressProvider(progress_bar, max=None):  # type: ignore
    if max is None or max == 0:
        return BAR_TYPES[progress_bar][1]().iter
    else:
        return BAR_TYPES[progress_bar][0](max=max).iter
site-packages/pip/_internal/cli/spinners.py000064400000011724150564745010015025 0ustar00import contextlib
import itertools
import logging
import sys
import time
from typing import IO, Iterator

from pip._vendor.progress import HIDE_CURSOR, SHOW_CURSOR

from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.logging import get_indentation

logger = logging.getLogger(__name__)


class SpinnerInterface:
    def spin(self) -> None:
        raise NotImplementedError()

    def finish(self, final_status: str) -> None:
        raise NotImplementedError()


class InteractiveSpinner(SpinnerInterface):
    def __init__(
        self,
        message: str,
        file: IO[str] = None,
        spin_chars: str = "-\\|/",
        # Empirically, 8 updates/second looks nice
        min_update_interval_seconds: float = 0.125,
    ):
        self._message = message
        if file is None:
            file = sys.stdout
        self._file = file
        self._rate_limiter = RateLimiter(min_update_interval_seconds)
        self._finished = False

        self._spin_cycle = itertools.cycle(spin_chars)

        self._file.write(" " * get_indentation() + self._message + " ... ")
        self._width = 0

    def _write(self, status: str) -> None:
        assert not self._finished
        # Erase what we wrote before by backspacing to the beginning, writing
        # spaces to overwrite the old text, and then backspacing again
        backup = "\b" * self._width
        self._file.write(backup + " " * self._width + backup)
        # Now we have a blank slate to add our status
        self._file.write(status)
        self._width = len(status)
        self._file.flush()
        self._rate_limiter.reset()

    def spin(self) -> None:
        if self._finished:
            return
        if not self._rate_limiter.ready():
            return
        self._write(next(self._spin_cycle))

    def finish(self, final_status: str) -> None:
        if self._finished:
            return
        self._write(final_status)
        self._file.write("\n")
        self._file.flush()
        self._finished = True


# Used for dumb terminals, non-interactive installs (no tty), etc.
# We still print updates occasionally (once every 60 seconds by default) to
# act as a keep-alive for systems like Travis-CI that take lack-of-output as
# an indication that a task has frozen.
class NonInteractiveSpinner(SpinnerInterface):
    def __init__(self, message: str, min_update_interval_seconds: float = 60.0) -> None:
        self._message = message
        self._finished = False
        self._rate_limiter = RateLimiter(min_update_interval_seconds)
        self._update("started")

    def _update(self, status: str) -> None:
        assert not self._finished
        self._rate_limiter.reset()
        logger.info("%s: %s", self._message, status)

    def spin(self) -> None:
        if self._finished:
            return
        if not self._rate_limiter.ready():
            return
        self._update("still running...")

    def finish(self, final_status: str) -> None:
        if self._finished:
            return
        self._update(f"finished with status '{final_status}'")
        self._finished = True


class RateLimiter:
    def __init__(self, min_update_interval_seconds: float) -> None:
        self._min_update_interval_seconds = min_update_interval_seconds
        self._last_update: float = 0

    def ready(self) -> bool:
        now = time.time()
        delta = now - self._last_update
        return delta >= self._min_update_interval_seconds

    def reset(self) -> None:
        self._last_update = time.time()


@contextlib.contextmanager
def open_spinner(message: str) -> Iterator[SpinnerInterface]:
    # Interactive spinner goes directly to sys.stdout rather than being routed
    # through the logging system, but it acts like it has level INFO,
    # i.e. it's only displayed if we're at level INFO or better.
    # Non-interactive spinner goes through the logging system, so it is always
    # in sync with logging configuration.
    if sys.stdout.isatty() and logger.getEffectiveLevel() <= logging.INFO:
        spinner: SpinnerInterface = InteractiveSpinner(message)
    else:
        spinner = NonInteractiveSpinner(message)
    try:
        with hidden_cursor(sys.stdout):
            yield spinner
    except KeyboardInterrupt:
        spinner.finish("canceled")
        raise
    except Exception:
        spinner.finish("error")
        raise
    else:
        spinner.finish("done")


@contextlib.contextmanager
def hidden_cursor(file: IO[str]) -> Iterator[None]:
    # The Windows terminal does not support the hide/show cursor ANSI codes,
    # even via colorama. So don't even try.
    if WINDOWS:
        yield
    # We don't want to clutter the output with control characters if we're
    # writing to a file, or if the user is running with --quiet.
    # See https://github.com/pypa/pip/issues/3418
    elif not file.isatty() or logger.getEffectiveLevel() > logging.INFO:
        yield
    else:
        file.write(HIDE_CURSOR)
        try:
            yield
        finally:
            file.write(SHOW_CURSOR)
site-packages/pip/_internal/cli/__init__.py000064400000000204150564745010014712 0ustar00"""Subpackage containing all of pip's command line interface related code
"""

# This file intentionally does not import submodules
site-packages/pip/_internal/cli/base_command.py000064400000017156150564745010015601 0ustar00"""Base Command class, and related routines"""

import functools
import logging
import logging.config
import optparse
import os
import sys
import traceback
from optparse import Values
from typing import Any, Callable, List, Optional, Tuple

from pip._internal.cli import cmdoptions
from pip._internal.cli.command_context import CommandContextMixIn
from pip._internal.cli.parser import ConfigOptionParser, UpdatingDefaultsHelpFormatter
from pip._internal.cli.status_codes import (
    ERROR,
    PREVIOUS_BUILD_DIR_ERROR,
    UNKNOWN_ERROR,
    VIRTUALENV_NOT_FOUND,
)
from pip._internal.exceptions import (
    BadCommand,
    CommandError,
    InstallationError,
    NetworkConnectionError,
    PreviousBuildDirError,
    UninstallationError,
)
from pip._internal.utils.filesystem import check_path_owner
from pip._internal.utils.logging import BrokenStdoutLoggingError, setup_logging
from pip._internal.utils.misc import get_prog, normalize_path
from pip._internal.utils.temp_dir import TempDirectoryTypeRegistry as TempDirRegistry
from pip._internal.utils.temp_dir import global_tempdir_manager, tempdir_registry
from pip._internal.utils.virtualenv import running_under_virtualenv

__all__ = ["Command"]

logger = logging.getLogger(__name__)


class Command(CommandContextMixIn):
    usage: str = ""
    ignore_require_venv: bool = False

    def __init__(self, name: str, summary: str, isolated: bool = False) -> None:
        super().__init__()

        self.name = name
        self.summary = summary
        self.parser = ConfigOptionParser(
            usage=self.usage,
            prog=f"{get_prog()} {name}",
            formatter=UpdatingDefaultsHelpFormatter(),
            add_help_option=False,
            name=name,
            description=self.__doc__,
            isolated=isolated,
        )

        self.tempdir_registry: Optional[TempDirRegistry] = None

        # Commands should add options to this option group
        optgroup_name = f"{self.name.capitalize()} Options"
        self.cmd_opts = optparse.OptionGroup(self.parser, optgroup_name)

        # Add the general options
        gen_opts = cmdoptions.make_option_group(
            cmdoptions.general_group,
            self.parser,
        )
        self.parser.add_option_group(gen_opts)

        self.add_options()

    def add_options(self) -> None:
        pass

    def handle_pip_version_check(self, options: Values) -> None:
        """
        This is a no-op so that commands by default do not do the pip version
        check.
        """
        # Make sure we do the pip version check if the index_group options
        # are present.
        assert not hasattr(options, "no_index")

    def run(self, options: Values, args: List[str]) -> int:
        raise NotImplementedError

    def parse_args(self, args: List[str]) -> Tuple[Values, List[str]]:
        # factored out for testability
        return self.parser.parse_args(args)

    def main(self, args: List[str]) -> int:
        try:
            with self.main_context():
                return self._main(args)
        finally:
            logging.shutdown()

    def _main(self, args: List[str]) -> int:
        # We must initialize this before the tempdir manager, otherwise the
        # configuration would not be accessible by the time we clean up the
        # tempdir manager.
        self.tempdir_registry = self.enter_context(tempdir_registry())
        # Intentionally set as early as possible so globally-managed temporary
        # directories are available to the rest of the code.
        self.enter_context(global_tempdir_manager())

        options, args = self.parse_args(args)

        # Set verbosity so that it can be used elsewhere.
        self.verbosity = options.verbose - options.quiet

        level_number = setup_logging(
            verbosity=self.verbosity,
            no_color=options.no_color,
            user_log_file=options.log,
        )

        # TODO: Try to get these passing down from the command?
        #       without resorting to os.environ to hold these.
        #       This also affects isolated builds and it should.

        if options.no_input:
            os.environ["PIP_NO_INPUT"] = "1"

        if options.exists_action:
            os.environ["PIP_EXISTS_ACTION"] = " ".join(options.exists_action)

        if options.require_venv and not self.ignore_require_venv:
            # If a venv is required check if it can really be found
            if not running_under_virtualenv():
                logger.critical("Could not find an activated virtualenv (required).")
                sys.exit(VIRTUALENV_NOT_FOUND)

        if options.cache_dir:
            options.cache_dir = normalize_path(options.cache_dir)
            if not check_path_owner(options.cache_dir):
                logger.warning(
                    "The directory '%s' or its parent directory is not owned "
                    "or is not writable by the current user. The cache "
                    "has been disabled. Check the permissions and owner of "
                    "that directory. If executing pip with sudo, you should "
                    "use sudo's -H flag.",
                    options.cache_dir,
                )
                options.cache_dir = None

        if "2020-resolver" in options.features_enabled:
            logger.warning(
                "--use-feature=2020-resolver no longer has any effect, "
                "since it is now the default dependency resolver in pip. "
                "This will become an error in pip 21.0."
            )

        def intercepts_unhandled_exc(
            run_func: Callable[..., int]
        ) -> Callable[..., int]:
            @functools.wraps(run_func)
            def exc_logging_wrapper(*args: Any) -> int:
                try:
                    status = run_func(*args)
                    assert isinstance(status, int)
                    return status
                except PreviousBuildDirError as exc:
                    logger.critical(str(exc))
                    logger.debug("Exception information:", exc_info=True)

                    return PREVIOUS_BUILD_DIR_ERROR
                except (
                    InstallationError,
                    UninstallationError,
                    BadCommand,
                    NetworkConnectionError,
                ) as exc:
                    logger.critical(str(exc))
                    logger.debug("Exception information:", exc_info=True)

                    return ERROR
                except CommandError as exc:
                    logger.critical("%s", exc)
                    logger.debug("Exception information:", exc_info=True)

                    return ERROR
                except BrokenStdoutLoggingError:
                    # Bypass our logger and write any remaining messages to
                    # stderr because stdout no longer works.
                    print("ERROR: Pipe to stdout was broken", file=sys.stderr)
                    if level_number <= logging.DEBUG:
                        traceback.print_exc(file=sys.stderr)

                    return ERROR
                except KeyboardInterrupt:
                    logger.critical("Operation cancelled by user")
                    logger.debug("Exception information:", exc_info=True)

                    return ERROR
                except BaseException:
                    logger.critical("Exception:", exc_info=True)

                    return UNKNOWN_ERROR

            return exc_logging_wrapper

        try:
            if not options.debug_mode:
                run = intercepts_unhandled_exc(self.run)
            else:
                run = self.run
            return run(options, args)
        finally:
            self.handle_pip_version_check(options)
site-packages/pip/_internal/index/__pycache__/collector.cpython-39.pyc000064400000037057150564745010022010 0ustar00a

��Re~D�@sdZddlZddlZddlZddlZddlZddlZddlZddlZ	ddl
Z	ddlZddl
mZddlmZmZmZmZmZmZmZmZddlmZmZddlmZddlmZmZddl m!Z!dd	l"m#Z#dd
l$m%Z%ddl&m'Z'ddl(m)Z)dd
l*m+Z+ddl,m-Z-m.Z.ddl/m0Z0ddl1m2Z2m3Z3m4Z4e�5e6�Z7ej8j9j:Z;ee<e<fZ=e<ee<d�dd�Z>Gdd�de?�Z@edd�dd�ZAGdd�de?�ZBe<e'dd�dd�ZCe<e'ed�dd �ZDe=ee<d!�d"d#�ZEe;e<e<d$�d%d&�ZFe<e<d'�d(d)�ZGe<e<d'�d*d+�ZHe�Id,ejJ�ZKe<eLe<d-�d.d/�ZMe<e<d�d0d1�ZNe;e<e<ee#d2�d3d4�ZOGd5d6�d6�ZPed7gee#fed7gee#fd8�d9d:�ZQeQd7ee#d;�d<d=��ZRGd>d7�d7�ZSdNe#ee<e?feed?dd@�dAdB�ZTdOeeLeSdD�dEdF�ZUdPe#ee'ed7dG�dHdI�ZVGdJdK�dKe�ZWGdLdM�dM�ZXdS)QzO
The main purpose of this module is to expose LinkCollector.collect_sources().
�N)�Values)�Callable�Iterable�List�MutableMapping�
NamedTuple�Optional�Sequence�Union)�html5lib�requests)�Response)�
RetryError�SSLError)�NetworkConnectionError)�Link)�SearchScope)�
PipSession)�raise_for_status)�is_archive_file)�pairwise�redact_auth_from_url)�vcs�)�CandidatesFromPage�
LinkSource�build_source)�url�returncCs6tjD]*}|���|�r|t|�dvr|SqdS)zgLook for VCS schemes in the URL.

    Returns the matched VCS scheme, or None if there's no match.
    z+:N)r�schemes�lower�
startswith�len)r�scheme�r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/index/collector.py�_match_vcs_scheme0s

r&cs&eZdZeedd��fdd�Z�ZS)�_NotHTMLN)�content_type�request_descrcst��||�||_||_dS�N)�super�__init__r(r))�selfr(r)��	__class__r$r%r,<sz_NotHTML.__init__)�__name__�
__module__�__qualname__�strr,�
__classcell__r$r$r.r%r';sr')�responsercCs.|j�dd�}|���d�s*t||jj��dS)z�Check the Content-Type header to ensure the response contains HTML.

    Raises `_NotHTML` if the content type is not text/html.
    �Content-Type��	text/htmlN)�headers�getr r!r'�request�method)r5r(r$r$r%�_ensure_html_headerBsr=c@seZdZdS)�_NotHTTPN)r0r1r2r$r$r$r%r>Lsr>)r�sessionrcCsFtj�|�\}}}}}|dvr$t��|j|dd�}t|�t|�dS)z�Send a HEAD request to the URL, and ensure the response contains HTML.

    Raises `_NotHTTP` if the URL is not available for a HEAD request, or
    `_NotHTML` if the content type is not text/html.
    >�http�httpsT)�allow_redirectsN)�urllib�parse�urlsplitr>�headrr=)rr?r#�netloc�path�query�fragment�respr$r$r%�_ensure_html_responsePsrLcCsRtt|�j�rt||d�t�dt|��|j|ddd�d�}t|�t	|�|S)aAccess an HTML page with GET, and return the response.

    This consists of three parts:

    1. If the URL looks suspiciously like an archive, send a HEAD first to
       check the Content-Type is HTML, to avoid downloading a large file.
       Raise `_NotHTTP` if the content type cannot be determined, or
       `_NotHTML` if it is not HTML.
    2. Actually perform the request. Raise HTTP exceptions on network failures.
    3. Check the Content-Type header to make sure we got HTML, and raise
       `_NotHTML` otherwise.
    �r?zGetting page %sr8z	max-age=0)�Acceptz
Cache-Control)r9)
rr�filenamerL�logger�debugrr:rr=)rr?rKr$r$r%�_get_html_response`s
��rR)r9rcCs2|r.d|vr.t�|d�\}}d|vr.|dSdS)z=Determine if we have any encoding information in our headers.r6�charsetN)�cgi�parse_header)r9r(�paramsr$r$r%�_get_encoding_from_headers�s
rW)�document�page_urlrcCs.|�d�D]}|�d�}|dur
|Sq
|S)a�Determine the HTML document's base URL.

    This looks for a ``<base>`` tag in the HTML document. If present, its href
    attribute denotes the base URL of anchor tags in the document. If there is
    no such tag (or if it does not have a valid href attribute), the HTML
    file's URL is used as the base URL.

    :param document: An HTML document representation. The current
        implementation expects the result of ``html5lib.parse()``.
    :param page_url: The URL of the HTML document.
    z.//base�hrefN)�findallr:)rXrY�baserZr$r$r%�_determine_base_url�s


r])�partrcCstj�tj�|��S)zP
    Clean a "part" of a URL path (i.e. after splitting on "@" characters).
    )rCrD�quote�unquote�r^r$r$r%�_clean_url_path_part�srbcCstj�tj�|��S)z�
    Clean the first part of a URL path that corresponds to a local
    filesystem path (i.e. the first part after splitting on "@" characters).
    )rCr;�pathname2url�url2pathnamerar$r$r%�_clean_file_url_path�s
rez(@|%2F))rH�
is_local_pathrcCs^|r
t}nt}t�|�}g}tt�|dg��D]$\}}|�||��|�|���q.d�	|�S)z*
    Clean the path portion of a URL.
    r7)
rerb�_reserved_chars_re�splitr�	itertools�chain�append�upper�join)rHrf�
clean_func�parts�
cleaned_parts�to_clean�reservedr$r$r%�_clean_url_path�s
rscCs6tj�|�}|j}t|j|d�}tj�|j|d��S)z�
    Make sure a link is fully quoted.
    For example, if ' ' occurs in the URL, it will be replaced with "%20",
    and without double-quoting other characters.
    )rf)rH)rCrD�urlparserGrsrH�
urlunparse�_replace)r�resultrfrHr$r$r%�_clean_link�srx)�anchorrY�base_urlrcCsL|�d�}|sdSttj�||��}|�d�}|�d�}t||||d�}|S)zJ
    Convert an anchor element in a simple repository page to a Link.
    rZNzdata-requires-pythonzdata-yanked)�
comes_from�requires_python�
yanked_reason)r:rxrCrD�urljoinr)ryrYrzrZr�	pyrequirer}�linkr$r$r%�_create_link_from_element�s


�r�c@s:eZdZddd�dd�Zeed�dd�Zed	�d
d�ZdS)�CacheablePageContent�HTMLPageN��pagercCs|js
J�||_dSr*)�cache_link_parsingr�)r-r�r$r$r%r,s
zCacheablePageContent.__init__)�otherrcCst|t|��o|jj|jjkSr*)�
isinstance�typer�r)r-r�r$r$r%�__eq__szCacheablePageContent.__eq__�rcCst|jj�Sr*)�hashr�r�r-r$r$r%�__hash__szCacheablePageContent.__hash__)	r0r1r2r,�object�boolr��intr�r$r$r$r%r�sr�r�)�fnrcsLtjdd�tttd��fdd���t���dttd���fdd	��}|S)
z�
    Given a function that parses an Iterable[Link] from an HTMLPage, cache the
    function's result (keyed by CacheablePageContent), unless the HTMLPage
    `page` has `page.cache_link_parsing == False`.
    N)�maxsize)�cacheable_pagercst�|j��Sr*)�listr�)r�)r�r$r%�wrappersz'with_cached_html_pages.<locals>.wrapperr�r�cs|jr�t|��St�|��Sr*)r�r�r�)r��r�r�r$r%�wrapper_wrappersz/with_cached_html_pages.<locals>.wrapper_wrapper)�	functools�	lru_cacher�rr�wraps)r�r�r$r�r%�with_cached_html_pagess
	
r�r�ccsVtj|j|jdd�}|j}t||�}|�d�D]"}t|||d�}|durJq.|Vq.dS)zP
    Parse an HTML document, and yield its anchor elements as Link objects.
    F)�transport_encoding�namespaceHTMLElementsz.//a)rYrzN)rrD�content�encodingrr]r[r�)r�rXrrzryr�r$r$r%�parse_links(s �
�r�c@s:eZdZdZd
eeeeedd�dd�Zed�dd	�Z	dS)r�z'Represents one page, along with its URLTN)r�r�rr�rcCs||_||_||_||_dS)am
        :param encoding: the encoding to decode the given content.
        :param url: the URL from which the HTML was downloaded.
        :param cache_link_parsing: whether links parsed from this page's url
                                   should be cached. PyPI index urls should
                                   have this set to False, for example.
        N)r�r�rr�)r-r�r�rr�r$r$r%r,CszHTMLPage.__init__r�cCs
t|j�Sr*)rrr�r$r$r%�__str__VszHTMLPage.__str__)T)
r0r1r2�__doc__�bytesrr3r�r,r�r$r$r$r%r�@s��).N)r��reason�methrcCs|durtj}|d||�dS)Nz%Could not fetch URL %s: %s - skipping)rPrQ)r�r�r�r$r$r%�_handle_get_page_failZsr�T)r5r�rcCst|j�}t|j||j|d�S)N)r�rr�)rWr9r�r�r)r5r�r�r$r$r%�_make_html_pageds
�r�)r�r?rc

Cs|durtd��|j�dd�d}t|�}|r@t�d||�dStj�|�\}}}}}}|dkr�t	j
�tj�
|��r�|�d�s�|d7}tj�|d�}t�d	|�zt||d
�}W�nNty�t�d|�Y�n>t�y}z"t�d||j|j�WYd}~�nd}~0t�y:}zt||�WYd}~n�d}~0t�yh}zt||�WYd}~n�d}~0t�y�}z,d
}	|	t|�7}	t||	tjd�WYd}~nld}~0tj�y�}zt|d|���WYd}~n6d}~0tj�yt|d�Yn0t||j d�SdS)Nz?_get_html_page() missing 1 required keyword argument: 'session'�#rrzICannot look at %s URL %s because it does not support lookup as web pages.�file�/z
index.htmlz# file: URL is directory, getting %srMz`Skipping page %s because it looks like an archive, and cannot be checked by a HTTP HEAD request.ziSkipping page %s because the %s request got Content-Type: %s.The only supported Content-Type is text/htmlz4There was a problem confirming the ssl certificate: )r�zconnection error: z	timed out)r�)!�	TypeErrorrrhr&rP�warningrCrDrt�osrH�isdirr;rd�endswithr~rQrRr>r'r)r(rr�rrr3�infor�ConnectionError�Timeoutr�r�)
r�r?r�
vcs_schemer#�_rHrK�excr�r$r$r%�_get_html_pagens^��
�
�$$r�c@s.eZdZUeeeed<eeeed<dS)�CollectedSources�
find_links�
index_urlsN)r0r1r2r	rr�__annotations__r$r$r$r%r��s
r�c@sxeZdZdZeedd�dd�Zedeee	dd�dd	��Z
eee
d
�dd��Zeeed
�dd�Ze
eed�dd�ZdS)�
LinkCollectorz�
    Responsible for collecting Link objects from all configured locations,
    making network requests as needed.

    The class's main method is its collect_sources() method.
    N)r?�search_scopercCs||_||_dSr*)r�r?)r-r?r�r$r$r%r,�szLinkCollector.__init__F)r?�options�suppress_no_indexrcCs`|jg|j}|jr8|s8t�dd�dd�|D���g}|jp@g}tj||d�}t	||d�}|S)z�
        :param session: The Session to use to make requests.
        :param suppress_no_index: Whether to ignore the --no-index option
            when constructing the SearchScope object.
        zIgnoring indexes: %s�,css|]}t|�VqdSr*)r)�.0rr$r$r%�	<genexpr>��z'LinkCollector.create.<locals>.<genexpr>�r�r�)r?r�)
�	index_url�extra_index_urls�no_indexrPrQrmr�r�creater�)�clsr?r�r�r�r�r��link_collectorr$r$r%r��s"
�
��zLinkCollector.creater�cCs|jjSr*)r�r�r�r$r$r%r��szLinkCollector.find_links)�locationrcCst||jd�S)z>
        Fetch an HTML page containing package links.
        rM)r�r?)r-r�r$r$r%�
fetch_page�szLinkCollector.fetch_page)�project_name�candidates_from_pagercs�t���fdd��j�|�D����}t���fdd��jD����}t�tj	�r�dd�t
�||�D�}t|��d|�d�g|}t�
d�|��tt|�t|�d	�S)
Nc3s$|]}t|��jjddd�VqdS)F�r��page_validator�
expand_dirr�N�rr?�is_secure_origin�r��loc�r�r-r$r%r��s��z0LinkCollector.collect_sources.<locals>.<genexpr>c3s$|]}t|��jjddd�VqdS)Tr�Nr�r�r�r$r%r��s��cSs*g|]"}|dur|jdurd|j���qS)Nz* )r�)r��sr$r$r%�
<listcomp>
s�z1LinkCollector.collect_sources.<locals>.<listcomp>z' location(s) to search for versions of �:�
r�)�collections�OrderedDictr��get_index_urls_locations�valuesr�rP�isEnabledFor�logging�DEBUGrirjr"rQrmr�r�)r-r�r��index_url_sources�find_links_sources�linesr$r�r%�collect_sources�s*
�
�
�
����zLinkCollector.collect_sources)F)r0r1r2r�rrr,�classmethodrr�r��propertyrr3r�rrr�r�rr�r�r$r$r$r%r��s(	��� �r�)N)T)N)Yr�rTr�r�rir�r��re�urllib.parserC�urllib.request�xml.etree.ElementTree�xml�optparser�typingrrrrrrr	r
�pip._vendorrrZpip._vendor.requestsr
Zpip._vendor.requests.exceptionsrr�pip._internal.exceptionsr�pip._internal.models.linkr�!pip._internal.models.search_scoper�pip._internal.network.sessionr�pip._internal.network.utilsr�pip._internal.utils.filetypesr�pip._internal.utils.miscrr�pip._internal.vcsr�sourcesrrr�	getLoggerr0rP�etree�ElementTree�Element�HTMLElementr3�ResponseHeadersr&�	Exceptionr'r=r>rLrRrWr]rbre�compile�
IGNORECASErgr�rsrxr�r�r�r�r�r�r�r�r�r�r$r$r$r%�<module>s�(


2	�
��

�
��?site-packages/pip/_internal/index/__pycache__/package_finder.cpython-39.pyc000064400000066746150564745010022753 0ustar00a

��Re���@sZdZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
mZmZddl
mZddlmZddlmZddlmZddlmZdd	lmZmZmZmZdd
lmZmZddlm Z ddl!m"Z"dd
l#m$Z$ddl%m&Z&ddl'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:ddl;m<Z<ddl=m>Z>gd�Z?e0e@�ZAeedeeBeCffZDeeBeBeBee	eBeDfZEd3e$eeBeBeBfeFeFd�dd �ZGGd!d"�d"�ZHee e4eCee d#�d$d%�ZIGd&d'�d'�ZJGd(d)�d)�ZKGd*d+�d+�ZLGd,d-�d-�ZMeCeCeBd.�d/d0�ZNeCeCe	eCd.�d1d2�ZOdS)4z!Routines related to PyPI, indexes�N)�	FrozenSet�Iterable�List�Optional�Set�Tuple�Union)�
specifiers)�Tag)�canonicalize_name)�_BaseVersion)�parse)�BestVersionAlreadyInstalled�DistributionNotFound�InvalidWheelFilename�UnsupportedWheel)�
LinkCollector�parse_links)�InstallationCandidate)�
FormatControl)�Link)�SearchScope)�SelectionPreferences)�TargetPython)�Wheel)�InstallRequirement)�	getLogger)�WHEEL_EXTENSION)�Hashes)�
indent_log)�build_netloc)�check_requires_python)�SUPPORTED_EXTENSIONS)�url_to_path)r�BestCandidateResult�
PackageFinder�F)�link�version_info�ignore_requires_python�returncCs|zt|j|d�}Wn$tjy6t�d|j|�YnB0|sxd�tt|��}|sft�	d||j|�dSt�d||j|�dS)aa
    Return whether the given Python version is compatible with a link's
    "Requires-Python" value.

    :param version_info: A 3-tuple of ints representing the Python
        major-minor-micro version to check.
    :param ignore_requires_python: Whether to ignore the "Requires-Python"
        value if the given Python version isn't compatible.
    )r(z2Ignoring invalid Requires-Python (%r) for link: %s�.z4Link requires a different Python (%s not in: %r): %sFzBIgnoring failed Requires-Python check (%s not in: %r) for link: %sT)
r!�requires_pythonr	�InvalidSpecifier�logger�debug�join�map�str�verbose)r'r(r)�
is_compatible�versionr&r&��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/index/package_finder.py�_check_link_requires_python3s8�
�
��r7c	@sZeZdZdZe�d�Zd
eeeee	e
ee
dd�dd�Ze
ee
eefd�dd	�ZdS)�
LinkEvaluatorzD
    Responsible for evaluating links for a particular project.
    z-py([123]\.?[0-9]?)$N)�project_name�canonical_name�formats�
target_python�allow_yankedr)r*cCs4|durd}||_||_||_||_||_||_dS)a�
        :param project_name: The user supplied package name.
        :param canonical_name: The canonical package name.
        :param formats: The formats allowed for this package. Should be a set
            with 'binary' or 'source' or both in it.
        :param target_python: The target Python interpreter to use when
            evaluating link compatibility. This is used, for example, to
            check wheel compatibility, as well as when checking the Python
            version, e.g. the Python version embedded in a link filename
            (or egg fragment) and against an HTML link's optional PEP 503
            "data-requires-python" attribute.
        :param allow_yanked: Whether files marked as yanked (in the sense
            of PEP 592) are permitted to be candidates for install.
        :param ignore_requires_python: Whether to ignore incompatible
            PEP 503 "data-requires-python" values in HTML links. Defaults
            to False.
        NF)�
_allow_yanked�_canonical_name�_ignore_requires_python�_formats�_target_pythonr9)�selfr9r:r;r<r=r)r&r&r6�__init__nszLinkEvaluator.__init__)r'r*cCs�d}|jr(|js(|jpd}dd|��fS|jr<|j}|j}n�|��\}}|sPdS|tvrfdd|��fSd|jvr�|tkr�d�	|j
�}d|fSd	|jvr�|d
kr�dS|tk�r0zt|j
�}Wnty�YdS0t|j�|jkr�d
�	|j
�}d|fS|j��}|�|��s*|��}d�	d�|��}d|fS|j}d|jv�rZ|tk�rZd|j
��}d|fS|�slt||j�}|�s�d|j
��}d|fS|j�|�}	|	�r�|d|	���}|	�d�}
|
|jjk�r�dSt||jj|j d�}|�s�dSt!�"d||�d|fS)aG
        Determine whether a link is a candidate for installation.

        :return: A tuple (is_candidate, result), where `result` is (1) a
            version string if `is_candidate` is True, and (2) if
            `is_candidate` is False, an optional string to log the reason
            the link fails to qualify.
        Nz<none given>Fzyanked for reason: )Fz
not a filezunsupported archive format: �binaryzNo binaries permitted for {}�macosx10z.zip)Fzmacosx10 one)Fzinvalid wheel filenamezwrong project name (not {})z^none of the wheel's tags ({}) are compatible (run pip debug --verbose to show compatible tags)�, �sourcezNo sources permitted for zMissing project version for �)FzPython version is incorrect)r(r))FNzFound link %s, version: %sT)#�	is_yankedr>�
yanked_reason�egg_fragment�ext�splitextr"rAr�formatr9�pathr�filenamerr�namer?rB�get_tags�	supported�get_formatted_file_tagsr0r5�_extract_version_from_fragment�_py_version_re�search�start�group�
py_versionr7�py_version_infor@r.r/)rCr'r5�reason�egg_inforM�wheel�supported_tags�	file_tags�matchr[�supports_pythonr&r&r6�
evaluate_link�sx	


���
�zLinkEvaluator.evaluate_link)N)�__name__�
__module__�__qualname__�__doc__�re�compilerWr2rr�boolrrDrrrdr&r&r&r6r8bs

��%r8)�
candidates�hashesr9r*c
	Cs�|st�dt|�|�t|�Sg}g}d}|D]>}|j}|js@n"|j|d�rV|d7}n|�|�q.|�|�q.|rx|}nt|�}t|�t|�kr�d}	n d�t|�d�	dd	�|D���}	t�d
t|�||j
|t|�||	�|S)a�
    Filter out candidates whose hashes aren't allowed, and return a new
    list of candidates.

    If at least one candidate has an allowed hash, then all candidates with
    either an allowed hash or no hash specified are returned.  Otherwise,
    the given candidates are returned.

    Including the candidates with no hash specified when there is a match
    allows a warning to be logged if there is a more preferred candidate
    with no hash specified.  Returning all candidates in the case of no
    matches lets pip report the hash of the candidate that would otherwise
    have been installed (e.g. permitting the user to more easily update
    their requirements file with the desired hash).
    zJGiven no hashes to check %s links for project %r: discarding no candidatesr)rmrIzdiscarding no candidateszdiscarding {} non-matches:
  {}z
  css|]}t|j�VqdS�N)r2r')�.0�	candidater&r&r6�	<genexpr>$�z*filter_unallowed_hashes.<locals>.<genexpr>zPChecked %s links for project %r against %s hashes (%s matches, %s no digest): %s)r.r/�len�listr'�has_hash�is_hash_allowed�appendrOr0�digest_count)
rlrmr9�matches_or_no_digest�non_matches�match_countrpr'�filtered�discard_messager&r&r6�filter_unallowed_hashes�sL�

�
�r~c@s$eZdZdZdeedd�dd�ZdS)�CandidatePreferenceszk
    Encapsulates some of the preferences for filtering and sorting
    InstallationCandidate objects.
    FN)�
prefer_binary�allow_all_prereleasesr*cCs||_||_dS)zR
        :param allow_all_prereleases: Whether to allow all pre-releases.
        N)r�r�)rCr�r�r&r&r6rD<szCandidatePreferences.__init__)FF)rerfrgrhrkrDr&r&r&r6r5s��rc@sTeZdZdZeeeeeedd�dd�Zeed�dd�Z	eed�d	d
�Z
dS)r$z�A collection of candidates, returned by `PackageFinder.find_best_candidate`.

    This class is only intended to be instantiated by CandidateEvaluator's
    `compute_best_candidate()` method.
    N)rl�applicable_candidates�best_candidater*cCsHt|�t|�ksJ�|dur&|r2J�n||vs2J�||_||_||_dS)a
        :param candidates: A sequence of all available candidates found.
        :param applicable_candidates: The applicable candidates.
        :param best_candidate: The most preferred candidate found, or None
            if no applicable candidates were found.
        N)�set�_applicable_candidates�_candidatesr��rCrlr�r�r&r&r6rDOs
zBestCandidateResult.__init__�r*cCs
t|j�S)zIterate through all candidates.)�iterr��rCr&r&r6�iter_allgszBestCandidateResult.iter_allcCs
t|j�S)z*Iterate through the applicable candidates.)r�r�r�r&r&r6�iter_applicableksz#BestCandidateResult.iter_applicable)rerfrgrhrrrrDrr�r�r&r&r&r6r$Hs�r$c
@s�eZdZdZedeeeeeee	j
eedd�dd��Zdee
ee	j
eeeedd�dd	�Ze
ee
ed
�dd�Zeed
�dd�Ze
eeed
�dd�Ze
eed
�dd�ZdS)�CandidateEvaluatorzm
    Responsible for filtering and sorting candidates for installation based
    on what tags are valid.
    NF)r9r<r�r��	specifierrmr*cCs:|durt�}|durt��}|��}|||||||d�S)a�Create a CandidateEvaluator object.

        :param target_python: The target Python interpreter to use when
            checking compatibility. If None (the default), a TargetPython
            object will be constructed from the running Python.
        :param specifier: An optional object implementing `filter`
            (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable
            versions.
        :param hashes: An optional collection of allowed hashes.
        N)r9r`r�r�r�rm)rr	�SpecifierSetrS)�clsr9r<r�r�r�rmr`r&r&r6�createws�zCandidateEvaluator.create)r9r`r�r�r�rmr*cCs<||_||_||_||_||_||_dd�t|�D�|_dS)z�
        :param supported_tags: The PEP 425 tags supported by the target
            Python in order of preference (most preferred first).
        cSsi|]\}}||�qSr&r&)ro�idx�tagr&r&r6�
<dictcomp>�sz/CandidateEvaluator.__init__.<locals>.<dictcomp>N)�_allow_all_prereleases�_hashes�_prefer_binary�
_project_name�
_specifier�_supported_tags�	enumerate�_wheel_tag_preferences)rCr9r`r�r�r�rmr&r&r6rD�s
�zCandidateEvaluator.__init__)rlr*csd|jpd}|j}dd�|jdd�|D�|d�D���fdd�|D�}t||j|jd	�}t||jd
�S)zM
        Return the applicable candidates from a list of candidates.
        NcSsh|]}t|��qSr&)r2)ro�vr&r&r6�	<setcomp>�s�z?CandidateEvaluator.get_applicable_candidates.<locals>.<setcomp>css|]}t|j�VqdSrn�r2r5�ro�cr&r&r6rq�rrz?CandidateEvaluator.get_applicable_candidates.<locals>.<genexpr>)�prereleasescsg|]}t|j��vr|�qSr&r�r���versionsr&r6�
<listcomp>�rrz@CandidateEvaluator.get_applicable_candidates.<locals>.<listcomp>)rlrmr9��key)r�r��filterr~r�r��sorted�	_sort_key)rCrl�allow_prereleasesr�r��filtered_applicable_candidatesr&r�r6�get_applicable_candidates�s
���z,CandidateEvaluator.get_applicable_candidates)rpr*c
Cs�|j}t|�}d}d}|j}|jr�t|j�}z|�||j�}Wn"tybt	d�
|j���Yn0|jrnd}|jdur�t
�d|j�}	|	��}
t|
d�|
df}n|}t|�|j��}dt|j�}||||j||fS)a)
        Function to pass as the `key` argument to a call to sorted() to sort
        InstallationCandidates by preference.

        Returns a tuple such that tuples sorting as greater using Python's
        default comparison operator are more preferred.

        The preference is as follows:

        First and foremost, candidates with allowed (matching) hashes are
        always preferred over candidates without matching hashes. This is
        because e.g. if the only candidate with an allowed hash is yanked,
        we still want to use that candidate.

        Second, excepting hash considerations, candidates that have been
        yanked (in the sense of PEP 592) are always less preferred than
        candidates that haven't been yanked. Then:

        If not finding wheels, they are sorted by version only.
        If finding wheels, then the sort order is by version, then:
          1. existing installs
          2. wheels ordered via Wheel.support_index_min(self._supported_tags)
          3. source archives
        If prefer_binary was set, then all wheels are sorted above sources.

        Note: it was considered to embed this logic into the Link
              comparison operators, but then different sdist links
              with the same version, would have to be considered equal
        r&rzB{} is not a supported wheel for this platform. It can't be sorted.rINz^(\d+)(.*)$���)r�rsr'�is_wheelrrQ�find_most_preferred_tagr��
ValueErrorrrOr��	build_tagrirb�groups�intrvr�rJr5)
rCrp�
valid_tags�support_numr��binary_preferencer'r_�prirb�build_tag_groups�has_allowed_hash�
yank_valuer&r&r6r��sD
����

�zCandidateEvaluator._sort_keycCs|sdSt||jd�}|S)zy
        Return the best candidate per the instance's sort order, or None if
        no candidate is acceptable.
        Nr�)�maxr�)rCrlr�r&r&r6�sort_best_candidatesz&CandidateEvaluator.sort_best_candidatecCs"|�|�}|�|�}t|||d�S)zF
        Compute and return a `BestCandidateResult` instance.
        )r�r�)r�r�r$r�r&r&r6�compute_best_candidate*s

�z)CandidateEvaluator.compute_best_candidate)NFFNN)FFN)rerfrgrh�classmethodr2rrrkr	�
BaseSpecifierrr�rr
rDrr��CandidateSortingKeyr�r�r$r�r&r&r&r6r�psL��(���$F��r�c	@seZdZdZd;eeeeeee	eedd�dd�Z
ed<eeeedd�dd��Z
eed	�d
d��Zeed	�dd
��Zejedd�dd
��Zeeed	�dd��Zeeed	�dd��Zeeed	�dd��Zeed	�dd��Zdd	�dd�Zeed	�dd��Zdd	�dd�Zeed�dd �Zeeeed!�d"d#�Z eedd$�d%d&�Z!eeee"d'�d(d)�Z#eeeee"d*�d+d,�Z$eeee"d-�d.d/�Z%e&j'dd0�eee"d�d1d2��Z(d=eee)j*ee+e,d3�d4d5�Z-e&j'dd0�d>eee)j*ee+e.d3�d6d7��Z/e0eee"d8�d9d:�Z1dS)?r%z�This finds packages.

    This is meant to match easy_install's technique for looking for
    packages, by reading pages and looking for appropriate links.
    N)�link_collectorr<r=�format_control�candidate_prefsr)r*cCsP|durt�}|ptt�t��}||_||_||_||_||_||_t�|_	dS)a�
        This constructor is primarily meant to be used by the create() class
        method and from tests.

        :param format_control: A FormatControl object, used to control
            the selection of source packages / binary packages when consulting
            the index and links.
        :param candidate_prefs: Options to use when creating a
            CandidateEvaluator object.
        N)
rrr�r>�_candidate_prefsr@�_link_collectorrBr��
_logged_links)rCr�r<r=r�r�r)r&r&r6rDCszPackageFinder.__init__)r��selection_prefsr<r*cCs8|durt�}t|j|jd�}|||||j|j|jd�S)afCreate a PackageFinder.

        :param selection_prefs: The candidate selection preferences, as a
            SelectionPreferences object.
        :param target_python: The target Python interpreter to use when
            checking compatibility. If None (the default), a TargetPython
            object will be constructed from the running Python.
        N)r�r�)r�r�r<r=r�r))rrr�r�r=r�r))r�r�r�r<r�r&r&r6r�js��zPackageFinder.creater�cCs|jSrn)rBr�r&r&r6r<�szPackageFinder.target_pythoncCs|jjSrn�r��search_scoper�r&r&r6r��szPackageFinder.search_scope)r�r*cCs||j_dSrnr�)rCr�r&r&r6r��scCs|jjSrn)r��
find_linksr�r&r&r6r��szPackageFinder.find_linkscCs|jjSrn)r��
index_urlsr�r&r&r6r��szPackageFinder.index_urlsccs|jjjD]}t|�Vq
dSrn)r��session�pip_trusted_originsr )rC�	host_portr&r&r6�
trusted_hosts�szPackageFinder.trusted_hostscCs|jjSrn�r�r�r�r&r&r6r��sz#PackageFinder.allow_all_prereleasescCsd|j_dS�NTr�r�r&r&r6�set_allow_all_prereleases�sz'PackageFinder.set_allow_all_prereleasescCs|jjSrn�r�r�r�r&r&r6r��szPackageFinder.prefer_binarycCsd|j_dSr�r�r�r&r&r6�set_prefer_binary�szPackageFinder.set_prefer_binary)r9r*cCs.t|�}|j�|�}t||||j|j|jd�S)N)r9r:r;r<r=r))rr��get_allowed_formatsr8rBr>r@)rCr9r:r;r&r&r6�make_link_evaluator�s�z!PackageFinder.make_link_evaluator)�linksr*cCsPgg}}t�}|D]2}||vr|�|�|jr<|�|�q|�|�q||S)z�
        Returns elements of links in order, non-egg links first, egg links
        second, while eliminating duplicates
        )r��addrLrw)rCr��eggs�no_eggs�seenr'r&r&r6�_sort_links�s

zPackageFinder._sort_links)r'r]r*cCs(||jvr$t�d||�|j�|�dS)NzSkipping link: %s: %s)r�r.r/r�)rCr'r]r&r&r6�_log_skipped_link�s
zPackageFinder._log_skipped_link)�link_evaluatorr'r*cCs8|�|�\}}|s(|r$|j||d�dSt|j||d�S)z�
        If the link is a candidate for install, convert it to an
        InstallationCandidate and return it. Otherwise, return None.
        )r]N)rRr'r5)rdr�rr9)rCr�r'�is_candidate�resultr&r&r6�get_install_candidate�s�z#PackageFinder.get_install_candidate)r�r�r*cCs6g}|�|�D]"}|�||�}|dur|�|�q|S)zU
        Convert links that are candidates to InstallationCandidate objects.
        N)r�r�rw)rCr�r�rlr'rpr&r&r6�evaluate_links�szPackageFinder.evaluate_links)�project_urlr�r*cCsht�d|�|j�|�}|dur$gStt|��}t��|j||d�}Wd�n1sZ0Y|S)Nz-Fetching project page and analyzing links: %s)r�)r.r/r��
fetch_pagertrrr�)rCr�r��	html_page�
page_links�
package_linksr&r&r6�process_project_url�s��$z!PackageFinder.process_project_url)�maxsizec	Cs�|�|�}|jj|tj|j|d�d�}tj�dd�|D��}t	|�}tj�dd�|D��}|�
|t|dd��}t�
tj�r�|r�dd	�|D�}t�d
d�|��||S)aFind all available InstallationCandidate for project_name

        This checks index_urls and find_links.
        All versions found are returned as an InstallationCandidate list.

        See LinkEvaluator.evaluate_link() for details on which files
        are accepted.
        )r�)r9�candidates_from_pagecss(|] }|D]}|dur
|��Vq
qdSrn)�page_candidates�ro�sourcesrHr&r&r6rqs�z4PackageFinder.find_all_candidates.<locals>.<genexpr>css(|] }|D]}|dur
|��Vq
qdSrn)�
file_linksr�r&r&r6rq's�T)�reversecSsg|]}t|jj��qSr&)r#r'�urlr�r&r&r6r�3rrz5PackageFinder.find_all_candidates.<locals>.<listcomp>zLocal files found: %srG)r�r��collect_sources�	functools�partialr��	itertools�chain�
from_iterablertr�r�r.�isEnabledFor�logging�DEBUGr/r0)	rCr9r��collected_sources�page_candidates_itr��
file_links_it�file_candidates�pathsr&r&r6�find_all_candidatess.

����
�z!PackageFinder.find_all_candidates)r9r�rmr*cCs"|j}tj||j|j|j||d�S)z*Create a CandidateEvaluator object to use.)r9r<r�r�r�rm)r�r�r�rBr�r�)rCr9r�rmr�r&r&r6�make_candidate_evaluator9s�z&PackageFinder.make_candidate_evaluatorcCs$|�|�}|j|||d�}|�|�S)aFind matches for the given project and specifier.

        :param specifier: An optional object implementing `filter`
            (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable
            versions.

        :return: A `BestCandidateResult` instance.
        )r9r�rm)rrr�)rCr9r�rmrl�candidate_evaluatorr&r&r6�find_best_candidateJs
�z!PackageFinder.find_best_candidate)�req�upgrader*c	Cs|jdd�}|j|j|j|d�}|j}d}|jdur@t|jj�}tt	t
d�dd�}|dur�|dur�t�d|||�
���td	�|���d}|r�|dus�|j|kr�d
}|s�|dur�|r�t�d|�nt�d||j�dS|r�t�d
|||����t�t�d|j||����|S)z�Try to find a Link matching req

        Expects req, an InstallRequirement and upgrade, a boolean
        Returns a InstallationCandidate if found,
        Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise
        F)�trust_internet)r�rmN)�	cand_iterr*cSs d�tdd�|D�td��pdS)NrGcSsh|]}t|j��qSr&r�r�r&r&r6r�~rrzKPackageFinder.find_requirement.<locals>._format_versions.<locals>.<setcomp>r��none)r0r��
parse_version)rr&r&r6�_format_versionsvs���z8PackageFinder.find_requirement.<locals>._format_versionszNCould not find a version that satisfies the requirement %s (from versions: %s)z%No matching distribution found for {}TzLExisting installed version (%s) is most up-to-date and satisfies requirementzUExisting installed version (%s) satisfies requirement (most up-to-date version is %s)z=Installed version (%s) is most up-to-date (past versions: %s)z)Using version %s (newest of versions: %s))rmrrRr�r��satisfied_byr	r5rrr2r.�criticalr�rrOr/r�r)	rCrrrm�best_candidate_resultr��installed_versionr
�best_installedr&r&r6�find_requirementash	�

������
�
�zPackageFinder.find_requirement)NNN)N)NN)NN)2rerfrgrhrrrkrrrrDr�rr��propertyr<rr��setterrr2r�r�rr�r�r�r�r�r8r�rr�r�rr�r�r�r��	lru_cacherr	r�rr�rr$rrrr&r&r&r6r%<s���'��
���
0��
���r%)�fragmentr:r*cCsNt|�D].\}}|dkrqt|d|��|kr|Sqt|�d|����dS)a�Find the separator's index based on the package's canonical name.

    :param fragment: A <package>+<version> filename "fragment" (stem) or
        egg fragment.
    :param canonical_name: The package's canonical name.

    This function is needed since the canonicalized name does not necessarily
    have the same length as the egg info's name part. An example::

    >>> fragment = 'foo__bar-1.0'
    >>> canonical_name = 'foo-bar'
    >>> _find_name_version_sep(fragment, canonical_name)
    8
    �-Nz does not match )r�rr�)rr:�ir�r&r&r6�_find_name_version_sep�s
rcCs@zt||�d}Wnty&YdS0||d�}|s<dS|S)z�Parse the version string from a <package>+<version> filename
    "fragment" (stem) or egg fragment.

    :param fragment: The string to parse. E.g. foo-2.1
    :param canonical_name: The canonicalized name of the package this
        belongs to.
    rIN)rr�)rr:�
version_startr5r&r&r6rV�srV)F)Prhr�r�r�ri�typingrrrrrrrZpip._vendor.packagingr	Zpip._vendor.packaging.tagsr
�pip._vendor.packaging.utilsrZpip._vendor.packaging.versionrr
r	�pip._internal.exceptionsrrrr�pip._internal.index.collectorrr�pip._internal.models.candidater�#pip._internal.models.format_controlr�pip._internal.models.linkr�!pip._internal.models.search_scoper�$pip._internal.models.selection_prefsr�"pip._internal.models.target_pythonr�pip._internal.models.wheelr�pip._internal.reqrZpip._internal.utils._logr�pip._internal.utils.filetypesr�pip._internal.utils.hashesr�pip._internal.utils.loggingr�pip._internal.utils.miscr �pip._internal.utils.packagingr!�pip._internal.utils.unpackingr"�pip._internal.utils.urlsr#�__all__rer.r�r2�BuildTagr�rkr7r8r~rr$r�r%rrVr&r&r&r6�<module>sn$��/�J(M~site-packages/pip/_internal/index/__pycache__/sources.cpython-39.pyc000064400000016076150564745010021503 0ustar00a

��Re��	@s0ddlZddlZddlZddlZddlmZmZmZmZddl	m
Z
ddlmZddl
mZmZddlmZe�e�Zee
ZeeZeegee
fZeegefZGdd�d�Zeed	�d
d�ZGdd
�d
e�ZGdd�de�ZGdd�de�ZGdd�de�Z eeeeeeeeeefd�dd�Z!dS)�N)�Callable�Iterable�Optional�Tuple)�InstallationCandidate)�Link)�path_to_url�url_to_path)�is_urlc@s>eZdZeeed�dd��Zed�dd�Ze	d�dd�Z
dS)	�
LinkSource��returncCs
t��dS)z,Returns the underlying link, if there's one.N��NotImplementedError��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/index/sources.py�linkszLinkSource.linkcCs
t��dS)z9Candidates found by parsing an archive listing HTML file.Nrrrrr�page_candidatesszLinkSource.page_candidatescCs
t��dS)z,Links found by specifying archives directly.Nrrrrr�
file_linksszLinkSource.file_linksN)�__name__�
__module__�__qualname__�propertyrrr�FoundCandidatesr�
FoundLinksrrrrrrsr)�file_urlr
cCstj|dd�ddkS)NF)�strictrz	text/html)�	mimetypes�
guess_type)rrrr�
_is_html_file#sr!c@sTeZdZdZeedd�dd�Zeee	d�dd��Z
ed�d	d
�Ze
d�dd�ZdS)
�_FlatDirectorySourcez�Link source specified by ``--find-links=<path-to-dir>``.

    This looks the content of the directory, and returns:

    * ``page_candidates``: Links listed on each HTML file in the directory.
    * ``file_candidates``: Archives in the directory.
    N)�candidates_from_page�pathr
cCs||_t�tj�|��|_dS�N)�_candidates_from_page�pathlib�Path�osr$�realpath�_path)rr#r$rrr�__init__0sz_FlatDirectorySource.__init__rcCsdSr%rrrrrr8sz_FlatDirectorySource.linkccs>|j��D].}tt|��}t|�s$q
|�t|��EdHq
dSr%)r+�iterdirr�strr!r&r�rr$�urlrrrr<s
z$_FlatDirectorySource.page_candidatesccs4|j��D]$}tt|��}t|�r$q
t|�Vq
dSr%)r+r-rr.r!rr/rrrrCs
z_FlatDirectorySource.file_links)rrr�__doc__�CandidatesFromPager.r,rrrrrrrrrrrrr"'s
�r"c@sTeZdZdZeedd�dd�Zeeed�dd��Z	e
d�d	d
�Zed�dd�Z
dS)
�_LocalFileSourceaC``--find-links=<path-or-url>`` or ``--[extra-]index-url=<path-or-url>``.

    If a URL is supplied, it must be a ``file:`` URL. If a path is supplied to
    the option, it is converted to a URL first. This returns:

    * ``page_candidates``: Links listed on an HTML file.
    * ``file_candidates``: The non-HTML file.
    N�r#rr
cCs||_||_dSr%�r&�_link�rr#rrrrr,Usz_LocalFileSource.__init__rcCs|jSr%�r6rrrrr]sz_LocalFileSource.linkccs&t|jj�sdS|�|j�EdHdSr%)r!r6r0r&rrrrrasz _LocalFileSource.page_candidatesccst|jj�rdS|jVdSr%)r!r6r0rrrrrfsz_LocalFileSource.file_links�rrrr1r2rr,rrrrrrrrrrrr3Ks�r3c@sVeZdZdZeeedd�dd�Zee	ed�dd��Z
ed�d	d
�Ze
d�dd�ZdS)
�_RemoteFileSourcez�``--find-links=<url>`` or ``--[extra-]index-url=<url>``.

    This returns:

    * ``page_candidates``: Links listed on an HTML file.
    * ``file_candidates``: The non-HTML file.
    N)r#�page_validatorrr
cCs||_||_||_dSr%)r&�_page_validatorr6)rr#r;rrrrr,usz_RemoteFileSource.__init__rcCs|jSr%r8rrrrrsz_RemoteFileSource.linkccs&|�|j�sdS|�|j�EdHdSr%)r<r6r&rrrrr�sz!_RemoteFileSource.page_candidatesccs|jVdSr%r8rrrrr�sz_RemoteFileSource.file_links)rrrr1r2�
PageValidatorrr,rrrrrrrrrrrr:ls
�
r:c@sTeZdZdZeedd�dd�Zeeed�dd��Z	e
d�d	d
�Zed�dd�Z
dS)
�_IndexDirectorySourcez�``--[extra-]index-url=<path-to-directory>``.

    This is treated like a remote URL; ``candidates_from_page`` contains logic
    for this by appending ``index.html`` to the link.
    Nr4cCs||_||_dSr%r5r7rrrr,�sz_IndexDirectorySource.__init__rcCs|jSr%r8rrrrr�sz_IndexDirectorySource.linkccs|�|j�EdHdSr%r5rrrrr�sz%_IndexDirectorySource.page_candidatescCsdS)Nrrrrrrr�sz _IndexDirectorySource.file_linksr9rrrrr>�s�r>)�locationr#r;�
expand_dir�cache_link_parsingr
c	Cs�d}d}tj�|�r"t|�}|}n$|�d�r:|}t|�}nt|�rF|}|durbd}t�||�dS|dur�t	||t
||d�d�}||fStj�|�r�|r�t||d�}nt
|t
||d�d�}||fStj�|�r�t|t
||d�d�}||fSt�d|�|dfS)	Nzfile:zVLocation '%s' is ignored: it is either a non-existing path or lacks a specific scheme.)NN)rA)r#r;r)r#r$)r#rz?Location '%s' is ignored: it is neither a file nor a directory.)r)r$�existsr�
startswithr	r
�logger�warningr:r�isdirr"r>�isfiler3)	r?r#r;r@rAr$r0�msg�sourcerrr�build_source�sX	

�
��
�
��rJ)"�loggingrr)r'�typingrrrr�pip._internal.models.candidater�pip._internal.models.linkr�pip._internal.utils.urlsrr	�pip._internal.vcsr
�	getLoggerrrDrrr2�boolr=rr.r!r"r3r:r>rJrrrr�<module>s4
$! �site-packages/pip/_internal/index/__pycache__/__init__.cpython-39.pyc000064400000000411150564745010021541 0ustar00a

��Re�@sdZdS)zIndex interaction code
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/index/__init__.py�<module>�site-packages/pip/_internal/index/package_finder.py000064400000106770150564745010016454 0ustar00"""Routines related to PyPI, indexes"""

# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import functools
import itertools
import logging
import re
from typing import FrozenSet, Iterable, List, Optional, Set, Tuple, Union

from pip._vendor.packaging import specifiers
from pip._vendor.packaging.tags import Tag
from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.packaging.version import _BaseVersion
from pip._vendor.packaging.version import parse as parse_version

from pip._internal.exceptions import (
    BestVersionAlreadyInstalled,
    DistributionNotFound,
    InvalidWheelFilename,
    UnsupportedWheel,
)
from pip._internal.index.collector import LinkCollector, parse_links
from pip._internal.models.candidate import InstallationCandidate
from pip._internal.models.format_control import FormatControl
from pip._internal.models.link import Link
from pip._internal.models.search_scope import SearchScope
from pip._internal.models.selection_prefs import SelectionPreferences
from pip._internal.models.target_python import TargetPython
from pip._internal.models.wheel import Wheel
from pip._internal.req import InstallRequirement
from pip._internal.utils._log import getLogger
from pip._internal.utils.filetypes import WHEEL_EXTENSION
from pip._internal.utils.hashes import Hashes
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import build_netloc
from pip._internal.utils.packaging import check_requires_python
from pip._internal.utils.unpacking import SUPPORTED_EXTENSIONS
from pip._internal.utils.urls import url_to_path

__all__ = ["FormatControl", "BestCandidateResult", "PackageFinder"]


logger = getLogger(__name__)

BuildTag = Union[Tuple[()], Tuple[int, str]]
CandidateSortingKey = Tuple[int, int, int, _BaseVersion, Optional[int], BuildTag]


def _check_link_requires_python(
    link: Link,
    version_info: Tuple[int, int, int],
    ignore_requires_python: bool = False,
) -> bool:
    """
    Return whether the given Python version is compatible with a link's
    "Requires-Python" value.

    :param version_info: A 3-tuple of ints representing the Python
        major-minor-micro version to check.
    :param ignore_requires_python: Whether to ignore the "Requires-Python"
        value if the given Python version isn't compatible.
    """
    try:
        is_compatible = check_requires_python(
            link.requires_python,
            version_info=version_info,
        )
    except specifiers.InvalidSpecifier:
        logger.debug(
            "Ignoring invalid Requires-Python (%r) for link: %s",
            link.requires_python,
            link,
        )
    else:
        if not is_compatible:
            version = ".".join(map(str, version_info))
            if not ignore_requires_python:
                logger.verbose(
                    "Link requires a different Python (%s not in: %r): %s",
                    version,
                    link.requires_python,
                    link,
                )
                return False

            logger.debug(
                "Ignoring failed Requires-Python check (%s not in: %r) for link: %s",
                version,
                link.requires_python,
                link,
            )

    return True


class LinkEvaluator:

    """
    Responsible for evaluating links for a particular project.
    """

    _py_version_re = re.compile(r"-py([123]\.?[0-9]?)$")

    # Don't include an allow_yanked default value to make sure each call
    # site considers whether yanked releases are allowed. This also causes
    # that decision to be made explicit in the calling code, which helps
    # people when reading the code.
    def __init__(
        self,
        project_name: str,
        canonical_name: str,
        formats: FrozenSet[str],
        target_python: TargetPython,
        allow_yanked: bool,
        ignore_requires_python: Optional[bool] = None,
    ) -> None:
        """
        :param project_name: The user supplied package name.
        :param canonical_name: The canonical package name.
        :param formats: The formats allowed for this package. Should be a set
            with 'binary' or 'source' or both in it.
        :param target_python: The target Python interpreter to use when
            evaluating link compatibility. This is used, for example, to
            check wheel compatibility, as well as when checking the Python
            version, e.g. the Python version embedded in a link filename
            (or egg fragment) and against an HTML link's optional PEP 503
            "data-requires-python" attribute.
        :param allow_yanked: Whether files marked as yanked (in the sense
            of PEP 592) are permitted to be candidates for install.
        :param ignore_requires_python: Whether to ignore incompatible
            PEP 503 "data-requires-python" values in HTML links. Defaults
            to False.
        """
        if ignore_requires_python is None:
            ignore_requires_python = False

        self._allow_yanked = allow_yanked
        self._canonical_name = canonical_name
        self._ignore_requires_python = ignore_requires_python
        self._formats = formats
        self._target_python = target_python

        self.project_name = project_name

    def evaluate_link(self, link: Link) -> Tuple[bool, Optional[str]]:
        """
        Determine whether a link is a candidate for installation.

        :return: A tuple (is_candidate, result), where `result` is (1) a
            version string if `is_candidate` is True, and (2) if
            `is_candidate` is False, an optional string to log the reason
            the link fails to qualify.
        """
        version = None
        if link.is_yanked and not self._allow_yanked:
            reason = link.yanked_reason or "<none given>"
            return (False, f"yanked for reason: {reason}")

        if link.egg_fragment:
            egg_info = link.egg_fragment
            ext = link.ext
        else:
            egg_info, ext = link.splitext()
            if not ext:
                return (False, "not a file")
            if ext not in SUPPORTED_EXTENSIONS:
                return (False, f"unsupported archive format: {ext}")
            if "binary" not in self._formats and ext == WHEEL_EXTENSION:
                reason = "No binaries permitted for {}".format(self.project_name)
                return (False, reason)
            if "macosx10" in link.path and ext == ".zip":
                return (False, "macosx10 one")
            if ext == WHEEL_EXTENSION:
                try:
                    wheel = Wheel(link.filename)
                except InvalidWheelFilename:
                    return (False, "invalid wheel filename")
                if canonicalize_name(wheel.name) != self._canonical_name:
                    reason = "wrong project name (not {})".format(self.project_name)
                    return (False, reason)

                supported_tags = self._target_python.get_tags()
                if not wheel.supported(supported_tags):
                    # Include the wheel's tags in the reason string to
                    # simplify troubleshooting compatibility issues.
                    file_tags = wheel.get_formatted_file_tags()
                    reason = (
                        "none of the wheel's tags ({}) are compatible "
                        "(run pip debug --verbose to show compatible tags)".format(
                            ", ".join(file_tags)
                        )
                    )
                    return (False, reason)

                version = wheel.version

        # This should be up by the self.ok_binary check, but see issue 2700.
        if "source" not in self._formats and ext != WHEEL_EXTENSION:
            reason = f"No sources permitted for {self.project_name}"
            return (False, reason)

        if not version:
            version = _extract_version_from_fragment(
                egg_info,
                self._canonical_name,
            )
        if not version:
            reason = f"Missing project version for {self.project_name}"
            return (False, reason)

        match = self._py_version_re.search(version)
        if match:
            version = version[: match.start()]
            py_version = match.group(1)
            if py_version != self._target_python.py_version:
                return (False, "Python version is incorrect")

        supports_python = _check_link_requires_python(
            link,
            version_info=self._target_python.py_version_info,
            ignore_requires_python=self._ignore_requires_python,
        )
        if not supports_python:
            # Return None for the reason text to suppress calling
            # _log_skipped_link().
            return (False, None)

        logger.debug("Found link %s, version: %s", link, version)

        return (True, version)


def filter_unallowed_hashes(
    candidates: List[InstallationCandidate],
    hashes: Hashes,
    project_name: str,
) -> List[InstallationCandidate]:
    """
    Filter out candidates whose hashes aren't allowed, and return a new
    list of candidates.

    If at least one candidate has an allowed hash, then all candidates with
    either an allowed hash or no hash specified are returned.  Otherwise,
    the given candidates are returned.

    Including the candidates with no hash specified when there is a match
    allows a warning to be logged if there is a more preferred candidate
    with no hash specified.  Returning all candidates in the case of no
    matches lets pip report the hash of the candidate that would otherwise
    have been installed (e.g. permitting the user to more easily update
    their requirements file with the desired hash).
    """
    if not hashes:
        logger.debug(
            "Given no hashes to check %s links for project %r: "
            "discarding no candidates",
            len(candidates),
            project_name,
        )
        # Make sure we're not returning back the given value.
        return list(candidates)

    matches_or_no_digest = []
    # Collect the non-matches for logging purposes.
    non_matches = []
    match_count = 0
    for candidate in candidates:
        link = candidate.link
        if not link.has_hash:
            pass
        elif link.is_hash_allowed(hashes=hashes):
            match_count += 1
        else:
            non_matches.append(candidate)
            continue

        matches_or_no_digest.append(candidate)

    if match_count:
        filtered = matches_or_no_digest
    else:
        # Make sure we're not returning back the given value.
        filtered = list(candidates)

    if len(filtered) == len(candidates):
        discard_message = "discarding no candidates"
    else:
        discard_message = "discarding {} non-matches:\n  {}".format(
            len(non_matches),
            "\n  ".join(str(candidate.link) for candidate in non_matches),
        )

    logger.debug(
        "Checked %s links for project %r against %s hashes "
        "(%s matches, %s no digest): %s",
        len(candidates),
        project_name,
        hashes.digest_count,
        match_count,
        len(matches_or_no_digest) - match_count,
        discard_message,
    )

    return filtered


class CandidatePreferences:

    """
    Encapsulates some of the preferences for filtering and sorting
    InstallationCandidate objects.
    """

    def __init__(
        self,
        prefer_binary: bool = False,
        allow_all_prereleases: bool = False,
    ) -> None:
        """
        :param allow_all_prereleases: Whether to allow all pre-releases.
        """
        self.allow_all_prereleases = allow_all_prereleases
        self.prefer_binary = prefer_binary


class BestCandidateResult:
    """A collection of candidates, returned by `PackageFinder.find_best_candidate`.

    This class is only intended to be instantiated by CandidateEvaluator's
    `compute_best_candidate()` method.
    """

    def __init__(
        self,
        candidates: List[InstallationCandidate],
        applicable_candidates: List[InstallationCandidate],
        best_candidate: Optional[InstallationCandidate],
    ) -> None:
        """
        :param candidates: A sequence of all available candidates found.
        :param applicable_candidates: The applicable candidates.
        :param best_candidate: The most preferred candidate found, or None
            if no applicable candidates were found.
        """
        assert set(applicable_candidates) <= set(candidates)

        if best_candidate is None:
            assert not applicable_candidates
        else:
            assert best_candidate in applicable_candidates

        self._applicable_candidates = applicable_candidates
        self._candidates = candidates

        self.best_candidate = best_candidate

    def iter_all(self) -> Iterable[InstallationCandidate]:
        """Iterate through all candidates."""
        return iter(self._candidates)

    def iter_applicable(self) -> Iterable[InstallationCandidate]:
        """Iterate through the applicable candidates."""
        return iter(self._applicable_candidates)


class CandidateEvaluator:

    """
    Responsible for filtering and sorting candidates for installation based
    on what tags are valid.
    """

    @classmethod
    def create(
        cls,
        project_name: str,
        target_python: Optional[TargetPython] = None,
        prefer_binary: bool = False,
        allow_all_prereleases: bool = False,
        specifier: Optional[specifiers.BaseSpecifier] = None,
        hashes: Optional[Hashes] = None,
    ) -> "CandidateEvaluator":
        """Create a CandidateEvaluator object.

        :param target_python: The target Python interpreter to use when
            checking compatibility. If None (the default), a TargetPython
            object will be constructed from the running Python.
        :param specifier: An optional object implementing `filter`
            (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable
            versions.
        :param hashes: An optional collection of allowed hashes.
        """
        if target_python is None:
            target_python = TargetPython()
        if specifier is None:
            specifier = specifiers.SpecifierSet()

        supported_tags = target_python.get_tags()

        return cls(
            project_name=project_name,
            supported_tags=supported_tags,
            specifier=specifier,
            prefer_binary=prefer_binary,
            allow_all_prereleases=allow_all_prereleases,
            hashes=hashes,
        )

    def __init__(
        self,
        project_name: str,
        supported_tags: List[Tag],
        specifier: specifiers.BaseSpecifier,
        prefer_binary: bool = False,
        allow_all_prereleases: bool = False,
        hashes: Optional[Hashes] = None,
    ) -> None:
        """
        :param supported_tags: The PEP 425 tags supported by the target
            Python in order of preference (most preferred first).
        """
        self._allow_all_prereleases = allow_all_prereleases
        self._hashes = hashes
        self._prefer_binary = prefer_binary
        self._project_name = project_name
        self._specifier = specifier
        self._supported_tags = supported_tags
        # Since the index of the tag in the _supported_tags list is used
        # as a priority, precompute a map from tag to index/priority to be
        # used in wheel.find_most_preferred_tag.
        self._wheel_tag_preferences = {
            tag: idx for idx, tag in enumerate(supported_tags)
        }

    def get_applicable_candidates(
        self,
        candidates: List[InstallationCandidate],
    ) -> List[InstallationCandidate]:
        """
        Return the applicable candidates from a list of candidates.
        """
        # Using None infers from the specifier instead.
        allow_prereleases = self._allow_all_prereleases or None
        specifier = self._specifier
        versions = {
            str(v)
            for v in specifier.filter(
                # We turn the version object into a str here because otherwise
                # when we're debundled but setuptools isn't, Python will see
                # packaging.version.Version and
                # pkg_resources._vendor.packaging.version.Version as different
                # types. This way we'll use a str as a common data interchange
                # format. If we stop using the pkg_resources provided specifier
                # and start using our own, we can drop the cast to str().
                (str(c.version) for c in candidates),
                prereleases=allow_prereleases,
            )
        }

        # Again, converting version to str to deal with debundling.
        applicable_candidates = [c for c in candidates if str(c.version) in versions]

        filtered_applicable_candidates = filter_unallowed_hashes(
            candidates=applicable_candidates,
            hashes=self._hashes,
            project_name=self._project_name,
        )

        return sorted(filtered_applicable_candidates, key=self._sort_key)

    def _sort_key(self, candidate: InstallationCandidate) -> CandidateSortingKey:
        """
        Function to pass as the `key` argument to a call to sorted() to sort
        InstallationCandidates by preference.

        Returns a tuple such that tuples sorting as greater using Python's
        default comparison operator are more preferred.

        The preference is as follows:

        First and foremost, candidates with allowed (matching) hashes are
        always preferred over candidates without matching hashes. This is
        because e.g. if the only candidate with an allowed hash is yanked,
        we still want to use that candidate.

        Second, excepting hash considerations, candidates that have been
        yanked (in the sense of PEP 592) are always less preferred than
        candidates that haven't been yanked. Then:

        If not finding wheels, they are sorted by version only.
        If finding wheels, then the sort order is by version, then:
          1. existing installs
          2. wheels ordered via Wheel.support_index_min(self._supported_tags)
          3. source archives
        If prefer_binary was set, then all wheels are sorted above sources.

        Note: it was considered to embed this logic into the Link
              comparison operators, but then different sdist links
              with the same version, would have to be considered equal
        """
        valid_tags = self._supported_tags
        support_num = len(valid_tags)
        build_tag: BuildTag = ()
        binary_preference = 0
        link = candidate.link
        if link.is_wheel:
            # can raise InvalidWheelFilename
            wheel = Wheel(link.filename)
            try:
                pri = -(
                    wheel.find_most_preferred_tag(
                        valid_tags, self._wheel_tag_preferences
                    )
                )
            except ValueError:
                raise UnsupportedWheel(
                    "{} is not a supported wheel for this platform. It "
                    "can't be sorted.".format(wheel.filename)
                )
            if self._prefer_binary:
                binary_preference = 1
            if wheel.build_tag is not None:
                match = re.match(r"^(\d+)(.*)$", wheel.build_tag)
                build_tag_groups = match.groups()
                build_tag = (int(build_tag_groups[0]), build_tag_groups[1])
        else:  # sdist
            pri = -(support_num)
        has_allowed_hash = int(link.is_hash_allowed(self._hashes))
        yank_value = -1 * int(link.is_yanked)  # -1 for yanked.
        return (
            has_allowed_hash,
            yank_value,
            binary_preference,
            candidate.version,
            pri,
            build_tag,
        )

    def sort_best_candidate(
        self,
        candidates: List[InstallationCandidate],
    ) -> Optional[InstallationCandidate]:
        """
        Return the best candidate per the instance's sort order, or None if
        no candidate is acceptable.
        """
        if not candidates:
            return None
        best_candidate = max(candidates, key=self._sort_key)
        return best_candidate

    def compute_best_candidate(
        self,
        candidates: List[InstallationCandidate],
    ) -> BestCandidateResult:
        """
        Compute and return a `BestCandidateResult` instance.
        """
        applicable_candidates = self.get_applicable_candidates(candidates)

        best_candidate = self.sort_best_candidate(applicable_candidates)

        return BestCandidateResult(
            candidates,
            applicable_candidates=applicable_candidates,
            best_candidate=best_candidate,
        )


class PackageFinder:
    """This finds packages.

    This is meant to match easy_install's technique for looking for
    packages, by reading pages and looking for appropriate links.
    """

    def __init__(
        self,
        link_collector: LinkCollector,
        target_python: TargetPython,
        allow_yanked: bool,
        format_control: Optional[FormatControl] = None,
        candidate_prefs: Optional[CandidatePreferences] = None,
        ignore_requires_python: Optional[bool] = None,
    ) -> None:
        """
        This constructor is primarily meant to be used by the create() class
        method and from tests.

        :param format_control: A FormatControl object, used to control
            the selection of source packages / binary packages when consulting
            the index and links.
        :param candidate_prefs: Options to use when creating a
            CandidateEvaluator object.
        """
        if candidate_prefs is None:
            candidate_prefs = CandidatePreferences()

        format_control = format_control or FormatControl(set(), set())

        self._allow_yanked = allow_yanked
        self._candidate_prefs = candidate_prefs
        self._ignore_requires_python = ignore_requires_python
        self._link_collector = link_collector
        self._target_python = target_python

        self.format_control = format_control

        # These are boring links that have already been logged somehow.
        self._logged_links: Set[Link] = set()

    # Don't include an allow_yanked default value to make sure each call
    # site considers whether yanked releases are allowed. This also causes
    # that decision to be made explicit in the calling code, which helps
    # people when reading the code.
    @classmethod
    def create(
        cls,
        link_collector: LinkCollector,
        selection_prefs: SelectionPreferences,
        target_python: Optional[TargetPython] = None,
    ) -> "PackageFinder":
        """Create a PackageFinder.

        :param selection_prefs: The candidate selection preferences, as a
            SelectionPreferences object.
        :param target_python: The target Python interpreter to use when
            checking compatibility. If None (the default), a TargetPython
            object will be constructed from the running Python.
        """
        if target_python is None:
            target_python = TargetPython()

        candidate_prefs = CandidatePreferences(
            prefer_binary=selection_prefs.prefer_binary,
            allow_all_prereleases=selection_prefs.allow_all_prereleases,
        )

        return cls(
            candidate_prefs=candidate_prefs,
            link_collector=link_collector,
            target_python=target_python,
            allow_yanked=selection_prefs.allow_yanked,
            format_control=selection_prefs.format_control,
            ignore_requires_python=selection_prefs.ignore_requires_python,
        )

    @property
    def target_python(self) -> TargetPython:
        return self._target_python

    @property
    def search_scope(self) -> SearchScope:
        return self._link_collector.search_scope

    @search_scope.setter
    def search_scope(self, search_scope: SearchScope) -> None:
        self._link_collector.search_scope = search_scope

    @property
    def find_links(self) -> List[str]:
        return self._link_collector.find_links

    @property
    def index_urls(self) -> List[str]:
        return self.search_scope.index_urls

    @property
    def trusted_hosts(self) -> Iterable[str]:
        for host_port in self._link_collector.session.pip_trusted_origins:
            yield build_netloc(*host_port)

    @property
    def allow_all_prereleases(self) -> bool:
        return self._candidate_prefs.allow_all_prereleases

    def set_allow_all_prereleases(self) -> None:
        self._candidate_prefs.allow_all_prereleases = True

    @property
    def prefer_binary(self) -> bool:
        return self._candidate_prefs.prefer_binary

    def set_prefer_binary(self) -> None:
        self._candidate_prefs.prefer_binary = True

    def make_link_evaluator(self, project_name: str) -> LinkEvaluator:
        canonical_name = canonicalize_name(project_name)
        formats = self.format_control.get_allowed_formats(canonical_name)

        return LinkEvaluator(
            project_name=project_name,
            canonical_name=canonical_name,
            formats=formats,
            target_python=self._target_python,
            allow_yanked=self._allow_yanked,
            ignore_requires_python=self._ignore_requires_python,
        )

    def _sort_links(self, links: Iterable[Link]) -> List[Link]:
        """
        Returns elements of links in order, non-egg links first, egg links
        second, while eliminating duplicates
        """
        eggs, no_eggs = [], []
        seen: Set[Link] = set()
        for link in links:
            if link not in seen:
                seen.add(link)
                if link.egg_fragment:
                    eggs.append(link)
                else:
                    no_eggs.append(link)
        return no_eggs + eggs

    def _log_skipped_link(self, link: Link, reason: str) -> None:
        if link not in self._logged_links:
            # Put the link at the end so the reason is more visible and because
            # the link string is usually very long.
            logger.debug("Skipping link: %s: %s", reason, link)
            self._logged_links.add(link)

    def get_install_candidate(
        self, link_evaluator: LinkEvaluator, link: Link
    ) -> Optional[InstallationCandidate]:
        """
        If the link is a candidate for install, convert it to an
        InstallationCandidate and return it. Otherwise, return None.
        """
        is_candidate, result = link_evaluator.evaluate_link(link)
        if not is_candidate:
            if result:
                self._log_skipped_link(link, reason=result)
            return None

        return InstallationCandidate(
            name=link_evaluator.project_name,
            link=link,
            version=result,
        )

    def evaluate_links(
        self, link_evaluator: LinkEvaluator, links: Iterable[Link]
    ) -> List[InstallationCandidate]:
        """
        Convert links that are candidates to InstallationCandidate objects.
        """
        candidates = []
        for link in self._sort_links(links):
            candidate = self.get_install_candidate(link_evaluator, link)
            if candidate is not None:
                candidates.append(candidate)

        return candidates

    def process_project_url(
        self, project_url: Link, link_evaluator: LinkEvaluator
    ) -> List[InstallationCandidate]:
        logger.debug(
            "Fetching project page and analyzing links: %s",
            project_url,
        )
        html_page = self._link_collector.fetch_page(project_url)
        if html_page is None:
            return []

        page_links = list(parse_links(html_page))

        with indent_log():
            package_links = self.evaluate_links(
                link_evaluator,
                links=page_links,
            )

        return package_links

    @functools.lru_cache(maxsize=None)
    def find_all_candidates(self, project_name: str) -> List[InstallationCandidate]:
        """Find all available InstallationCandidate for project_name

        This checks index_urls and find_links.
        All versions found are returned as an InstallationCandidate list.

        See LinkEvaluator.evaluate_link() for details on which files
        are accepted.
        """
        link_evaluator = self.make_link_evaluator(project_name)

        collected_sources = self._link_collector.collect_sources(
            project_name=project_name,
            candidates_from_page=functools.partial(
                self.process_project_url,
                link_evaluator=link_evaluator,
            ),
        )

        page_candidates_it = itertools.chain.from_iterable(
            source.page_candidates()
            for sources in collected_sources
            for source in sources
            if source is not None
        )
        page_candidates = list(page_candidates_it)

        file_links_it = itertools.chain.from_iterable(
            source.file_links()
            for sources in collected_sources
            for source in sources
            if source is not None
        )
        file_candidates = self.evaluate_links(
            link_evaluator,
            sorted(file_links_it, reverse=True),
        )

        if logger.isEnabledFor(logging.DEBUG) and file_candidates:
            paths = [url_to_path(c.link.url) for c in file_candidates]
            logger.debug("Local files found: %s", ", ".join(paths))

        # This is an intentional priority ordering
        return file_candidates + page_candidates

    def make_candidate_evaluator(
        self,
        project_name: str,
        specifier: Optional[specifiers.BaseSpecifier] = None,
        hashes: Optional[Hashes] = None,
    ) -> CandidateEvaluator:
        """Create a CandidateEvaluator object to use."""
        candidate_prefs = self._candidate_prefs
        return CandidateEvaluator.create(
            project_name=project_name,
            target_python=self._target_python,
            prefer_binary=candidate_prefs.prefer_binary,
            allow_all_prereleases=candidate_prefs.allow_all_prereleases,
            specifier=specifier,
            hashes=hashes,
        )

    @functools.lru_cache(maxsize=None)
    def find_best_candidate(
        self,
        project_name: str,
        specifier: Optional[specifiers.BaseSpecifier] = None,
        hashes: Optional[Hashes] = None,
    ) -> BestCandidateResult:
        """Find matches for the given project and specifier.

        :param specifier: An optional object implementing `filter`
            (e.g. `packaging.specifiers.SpecifierSet`) to filter applicable
            versions.

        :return: A `BestCandidateResult` instance.
        """
        candidates = self.find_all_candidates(project_name)
        candidate_evaluator = self.make_candidate_evaluator(
            project_name=project_name,
            specifier=specifier,
            hashes=hashes,
        )
        return candidate_evaluator.compute_best_candidate(candidates)

    def find_requirement(
        self, req: InstallRequirement, upgrade: bool
    ) -> Optional[InstallationCandidate]:
        """Try to find a Link matching req

        Expects req, an InstallRequirement and upgrade, a boolean
        Returns a InstallationCandidate if found,
        Raises DistributionNotFound or BestVersionAlreadyInstalled otherwise
        """
        hashes = req.hashes(trust_internet=False)
        best_candidate_result = self.find_best_candidate(
            req.name,
            specifier=req.specifier,
            hashes=hashes,
        )
        best_candidate = best_candidate_result.best_candidate

        installed_version: Optional[_BaseVersion] = None
        if req.satisfied_by is not None:
            installed_version = parse_version(req.satisfied_by.version)

        def _format_versions(cand_iter: Iterable[InstallationCandidate]) -> str:
            # This repeated parse_version and str() conversion is needed to
            # handle different vendoring sources from pip and pkg_resources.
            # If we stop using the pkg_resources provided specifier and start
            # using our own, we can drop the cast to str().
            return (
                ", ".join(
                    sorted(
                        {str(c.version) for c in cand_iter},
                        key=parse_version,
                    )
                )
                or "none"
            )

        if installed_version is None and best_candidate is None:
            logger.critical(
                "Could not find a version that satisfies the requirement %s "
                "(from versions: %s)",
                req,
                _format_versions(best_candidate_result.iter_all()),
            )

            raise DistributionNotFound(
                "No matching distribution found for {}".format(req)
            )

        best_installed = False
        if installed_version and (
            best_candidate is None or best_candidate.version <= installed_version
        ):
            best_installed = True

        if not upgrade and installed_version is not None:
            if best_installed:
                logger.debug(
                    "Existing installed version (%s) is most up-to-date and "
                    "satisfies requirement",
                    installed_version,
                )
            else:
                logger.debug(
                    "Existing installed version (%s) satisfies requirement "
                    "(most up-to-date version is %s)",
                    installed_version,
                    best_candidate.version,
                )
            return None

        if best_installed:
            # We have an existing version, and its the best version
            logger.debug(
                "Installed version (%s) is most up-to-date (past versions: %s)",
                installed_version,
                _format_versions(best_candidate_result.iter_applicable()),
            )
            raise BestVersionAlreadyInstalled

        logger.debug(
            "Using version %s (newest of versions: %s)",
            best_candidate.version,
            _format_versions(best_candidate_result.iter_applicable()),
        )
        return best_candidate


def _find_name_version_sep(fragment: str, canonical_name: str) -> int:
    """Find the separator's index based on the package's canonical name.

    :param fragment: A <package>+<version> filename "fragment" (stem) or
        egg fragment.
    :param canonical_name: The package's canonical name.

    This function is needed since the canonicalized name does not necessarily
    have the same length as the egg info's name part. An example::

    >>> fragment = 'foo__bar-1.0'
    >>> canonical_name = 'foo-bar'
    >>> _find_name_version_sep(fragment, canonical_name)
    8
    """
    # Project name and version must be separated by one single dash. Find all
    # occurrences of dashes; if the string in front of it matches the canonical
    # name, this is the one separating the name and version parts.
    for i, c in enumerate(fragment):
        if c != "-":
            continue
        if canonicalize_name(fragment[:i]) == canonical_name:
            return i
    raise ValueError(f"{fragment} does not match {canonical_name}")


def _extract_version_from_fragment(fragment: str, canonical_name: str) -> Optional[str]:
    """Parse the version string from a <package>+<version> filename
    "fragment" (stem) or egg fragment.

    :param fragment: The string to parse. E.g. foo-2.1
    :param canonical_name: The canonicalized name of the package this
        belongs to.
    """
    try:
        version_start = _find_name_version_sep(fragment, canonical_name) + 1
    except ValueError:
        return None
    version = fragment[version_start:]
    if not version:
        return None
    return version
site-packages/pip/_internal/index/__init__.py000064400000000036150564745010015255 0ustar00"""Index interaction code
"""
site-packages/pip/_internal/index/sources.py000064400000014635150564745010015213 0ustar00import logging
import mimetypes
import os
import pathlib
from typing import Callable, Iterable, Optional, Tuple

from pip._internal.models.candidate import InstallationCandidate
from pip._internal.models.link import Link
from pip._internal.utils.urls import path_to_url, url_to_path
from pip._internal.vcs import is_url

logger = logging.getLogger(__name__)

FoundCandidates = Iterable[InstallationCandidate]
FoundLinks = Iterable[Link]
CandidatesFromPage = Callable[[Link], Iterable[InstallationCandidate]]
PageValidator = Callable[[Link], bool]


class LinkSource:
    @property
    def link(self) -> Optional[Link]:
        """Returns the underlying link, if there's one."""
        raise NotImplementedError()

    def page_candidates(self) -> FoundCandidates:
        """Candidates found by parsing an archive listing HTML file."""
        raise NotImplementedError()

    def file_links(self) -> FoundLinks:
        """Links found by specifying archives directly."""
        raise NotImplementedError()


def _is_html_file(file_url: str) -> bool:
    return mimetypes.guess_type(file_url, strict=False)[0] == "text/html"


class _FlatDirectorySource(LinkSource):
    """Link source specified by ``--find-links=<path-to-dir>``.

    This looks the content of the directory, and returns:

    * ``page_candidates``: Links listed on each HTML file in the directory.
    * ``file_candidates``: Archives in the directory.
    """

    def __init__(
        self,
        candidates_from_page: CandidatesFromPage,
        path: str,
    ) -> None:
        self._candidates_from_page = candidates_from_page
        self._path = pathlib.Path(os.path.realpath(path))

    @property
    def link(self) -> Optional[Link]:
        return None

    def page_candidates(self) -> FoundCandidates:
        for path in self._path.iterdir():
            url = path_to_url(str(path))
            if not _is_html_file(url):
                continue
            yield from self._candidates_from_page(Link(url))

    def file_links(self) -> FoundLinks:
        for path in self._path.iterdir():
            url = path_to_url(str(path))
            if _is_html_file(url):
                continue
            yield Link(url)


class _LocalFileSource(LinkSource):
    """``--find-links=<path-or-url>`` or ``--[extra-]index-url=<path-or-url>``.

    If a URL is supplied, it must be a ``file:`` URL. If a path is supplied to
    the option, it is converted to a URL first. This returns:

    * ``page_candidates``: Links listed on an HTML file.
    * ``file_candidates``: The non-HTML file.
    """

    def __init__(
        self,
        candidates_from_page: CandidatesFromPage,
        link: Link,
    ) -> None:
        self._candidates_from_page = candidates_from_page
        self._link = link

    @property
    def link(self) -> Optional[Link]:
        return self._link

    def page_candidates(self) -> FoundCandidates:
        if not _is_html_file(self._link.url):
            return
        yield from self._candidates_from_page(self._link)

    def file_links(self) -> FoundLinks:
        if _is_html_file(self._link.url):
            return
        yield self._link


class _RemoteFileSource(LinkSource):
    """``--find-links=<url>`` or ``--[extra-]index-url=<url>``.

    This returns:

    * ``page_candidates``: Links listed on an HTML file.
    * ``file_candidates``: The non-HTML file.
    """

    def __init__(
        self,
        candidates_from_page: CandidatesFromPage,
        page_validator: PageValidator,
        link: Link,
    ) -> None:
        self._candidates_from_page = candidates_from_page
        self._page_validator = page_validator
        self._link = link

    @property
    def link(self) -> Optional[Link]:
        return self._link

    def page_candidates(self) -> FoundCandidates:
        if not self._page_validator(self._link):
            return
        yield from self._candidates_from_page(self._link)

    def file_links(self) -> FoundLinks:
        yield self._link


class _IndexDirectorySource(LinkSource):
    """``--[extra-]index-url=<path-to-directory>``.

    This is treated like a remote URL; ``candidates_from_page`` contains logic
    for this by appending ``index.html`` to the link.
    """

    def __init__(
        self,
        candidates_from_page: CandidatesFromPage,
        link: Link,
    ) -> None:
        self._candidates_from_page = candidates_from_page
        self._link = link

    @property
    def link(self) -> Optional[Link]:
        return self._link

    def page_candidates(self) -> FoundCandidates:
        yield from self._candidates_from_page(self._link)

    def file_links(self) -> FoundLinks:
        return ()


def build_source(
    location: str,
    *,
    candidates_from_page: CandidatesFromPage,
    page_validator: PageValidator,
    expand_dir: bool,
    cache_link_parsing: bool,
) -> Tuple[Optional[str], Optional[LinkSource]]:

    path: Optional[str] = None
    url: Optional[str] = None
    if os.path.exists(location):  # Is a local path.
        url = path_to_url(location)
        path = location
    elif location.startswith("file:"):  # A file: URL.
        url = location
        path = url_to_path(location)
    elif is_url(location):
        url = location

    if url is None:
        msg = (
            "Location '%s' is ignored: "
            "it is either a non-existing path or lacks a specific scheme."
        )
        logger.warning(msg, location)
        return (None, None)

    if path is None:
        source: LinkSource = _RemoteFileSource(
            candidates_from_page=candidates_from_page,
            page_validator=page_validator,
            link=Link(url, cache_link_parsing=cache_link_parsing),
        )
        return (url, source)

    if os.path.isdir(path):
        if expand_dir:
            source = _FlatDirectorySource(
                candidates_from_page=candidates_from_page,
                path=path,
            )
        else:
            source = _IndexDirectorySource(
                candidates_from_page=candidates_from_page,
                link=Link(url, cache_link_parsing=cache_link_parsing),
            )
        return (url, source)
    elif os.path.isfile(path):
        source = _LocalFileSource(
            candidates_from_page=candidates_from_page,
            link=Link(url, cache_link_parsing=cache_link_parsing),
        )
        return (url, source)
    logger.warning(
        "Location '%s' is ignored: it is neither a file nor a directory.",
        location,
    )
    return (url, None)
site-packages/pip/_internal/index/collector.py000064400000042176150564745010015517 0ustar00"""
The main purpose of this module is to expose LinkCollector.collect_sources().
"""

import cgi
import collections
import functools
import itertools
import logging
import os
import re
import urllib.parse
import urllib.request
import xml.etree.ElementTree
from optparse import Values
from typing import (
    Callable,
    Iterable,
    List,
    MutableMapping,
    NamedTuple,
    Optional,
    Sequence,
    Union,
)

from pip._vendor import html5lib, requests
from pip._vendor.requests import Response
from pip._vendor.requests.exceptions import RetryError, SSLError

from pip._internal.exceptions import NetworkConnectionError
from pip._internal.models.link import Link
from pip._internal.models.search_scope import SearchScope
from pip._internal.network.session import PipSession
from pip._internal.network.utils import raise_for_status
from pip._internal.utils.filetypes import is_archive_file
from pip._internal.utils.misc import pairwise, redact_auth_from_url
from pip._internal.vcs import vcs

from .sources import CandidatesFromPage, LinkSource, build_source

logger = logging.getLogger(__name__)

HTMLElement = xml.etree.ElementTree.Element
ResponseHeaders = MutableMapping[str, str]


def _match_vcs_scheme(url: str) -> Optional[str]:
    """Look for VCS schemes in the URL.

    Returns the matched VCS scheme, or None if there's no match.
    """
    for scheme in vcs.schemes:
        if url.lower().startswith(scheme) and url[len(scheme)] in "+:":
            return scheme
    return None


class _NotHTML(Exception):
    def __init__(self, content_type: str, request_desc: str) -> None:
        super().__init__(content_type, request_desc)
        self.content_type = content_type
        self.request_desc = request_desc


def _ensure_html_header(response: Response) -> None:
    """Check the Content-Type header to ensure the response contains HTML.

    Raises `_NotHTML` if the content type is not text/html.
    """
    content_type = response.headers.get("Content-Type", "")
    if not content_type.lower().startswith("text/html"):
        raise _NotHTML(content_type, response.request.method)


class _NotHTTP(Exception):
    pass


def _ensure_html_response(url: str, session: PipSession) -> None:
    """Send a HEAD request to the URL, and ensure the response contains HTML.

    Raises `_NotHTTP` if the URL is not available for a HEAD request, or
    `_NotHTML` if the content type is not text/html.
    """
    scheme, netloc, path, query, fragment = urllib.parse.urlsplit(url)
    if scheme not in {"http", "https"}:
        raise _NotHTTP()

    resp = session.head(url, allow_redirects=True)
    raise_for_status(resp)

    _ensure_html_header(resp)


def _get_html_response(url: str, session: PipSession) -> Response:
    """Access an HTML page with GET, and return the response.

    This consists of three parts:

    1. If the URL looks suspiciously like an archive, send a HEAD first to
       check the Content-Type is HTML, to avoid downloading a large file.
       Raise `_NotHTTP` if the content type cannot be determined, or
       `_NotHTML` if it is not HTML.
    2. Actually perform the request. Raise HTTP exceptions on network failures.
    3. Check the Content-Type header to make sure we got HTML, and raise
       `_NotHTML` otherwise.
    """
    if is_archive_file(Link(url).filename):
        _ensure_html_response(url, session=session)

    logger.debug("Getting page %s", redact_auth_from_url(url))

    resp = session.get(
        url,
        headers={
            "Accept": "text/html",
            # We don't want to blindly returned cached data for
            # /simple/, because authors generally expecting that
            # twine upload && pip install will function, but if
            # they've done a pip install in the last ~10 minutes
            # it won't. Thus by setting this to zero we will not
            # blindly use any cached data, however the benefit of
            # using max-age=0 instead of no-cache, is that we will
            # still support conditional requests, so we will still
            # minimize traffic sent in cases where the page hasn't
            # changed at all, we will just always incur the round
            # trip for the conditional GET now instead of only
            # once per 10 minutes.
            # For more information, please see pypa/pip#5670.
            "Cache-Control": "max-age=0",
        },
    )
    raise_for_status(resp)

    # The check for archives above only works if the url ends with
    # something that looks like an archive. However that is not a
    # requirement of an url. Unless we issue a HEAD request on every
    # url we cannot know ahead of time for sure if something is HTML
    # or not. However we can check after we've downloaded it.
    _ensure_html_header(resp)

    return resp


def _get_encoding_from_headers(headers: ResponseHeaders) -> Optional[str]:
    """Determine if we have any encoding information in our headers."""
    if headers and "Content-Type" in headers:
        content_type, params = cgi.parse_header(headers["Content-Type"])
        if "charset" in params:
            return params["charset"]
    return None


def _determine_base_url(document: HTMLElement, page_url: str) -> str:
    """Determine the HTML document's base URL.

    This looks for a ``<base>`` tag in the HTML document. If present, its href
    attribute denotes the base URL of anchor tags in the document. If there is
    no such tag (or if it does not have a valid href attribute), the HTML
    file's URL is used as the base URL.

    :param document: An HTML document representation. The current
        implementation expects the result of ``html5lib.parse()``.
    :param page_url: The URL of the HTML document.
    """
    for base in document.findall(".//base"):
        href = base.get("href")
        if href is not None:
            return href
    return page_url


def _clean_url_path_part(part: str) -> str:
    """
    Clean a "part" of a URL path (i.e. after splitting on "@" characters).
    """
    # We unquote prior to quoting to make sure nothing is double quoted.
    return urllib.parse.quote(urllib.parse.unquote(part))


def _clean_file_url_path(part: str) -> str:
    """
    Clean the first part of a URL path that corresponds to a local
    filesystem path (i.e. the first part after splitting on "@" characters).
    """
    # We unquote prior to quoting to make sure nothing is double quoted.
    # Also, on Windows the path part might contain a drive letter which
    # should not be quoted. On Linux where drive letters do not
    # exist, the colon should be quoted. We rely on urllib.request
    # to do the right thing here.
    return urllib.request.pathname2url(urllib.request.url2pathname(part))


# percent-encoded:                   /
_reserved_chars_re = re.compile("(@|%2F)", re.IGNORECASE)


def _clean_url_path(path: str, is_local_path: bool) -> str:
    """
    Clean the path portion of a URL.
    """
    if is_local_path:
        clean_func = _clean_file_url_path
    else:
        clean_func = _clean_url_path_part

    # Split on the reserved characters prior to cleaning so that
    # revision strings in VCS URLs are properly preserved.
    parts = _reserved_chars_re.split(path)

    cleaned_parts = []
    for to_clean, reserved in pairwise(itertools.chain(parts, [""])):
        cleaned_parts.append(clean_func(to_clean))
        # Normalize %xx escapes (e.g. %2f -> %2F)
        cleaned_parts.append(reserved.upper())

    return "".join(cleaned_parts)


def _clean_link(url: str) -> str:
    """
    Make sure a link is fully quoted.
    For example, if ' ' occurs in the URL, it will be replaced with "%20",
    and without double-quoting other characters.
    """
    # Split the URL into parts according to the general structure
    # `scheme://netloc/path;parameters?query#fragment`.
    result = urllib.parse.urlparse(url)
    # If the netloc is empty, then the URL refers to a local filesystem path.
    is_local_path = not result.netloc
    path = _clean_url_path(result.path, is_local_path=is_local_path)
    return urllib.parse.urlunparse(result._replace(path=path))


def _create_link_from_element(
    anchor: HTMLElement,
    page_url: str,
    base_url: str,
) -> Optional[Link]:
    """
    Convert an anchor element in a simple repository page to a Link.
    """
    href = anchor.get("href")
    if not href:
        return None

    url = _clean_link(urllib.parse.urljoin(base_url, href))
    pyrequire = anchor.get("data-requires-python")
    yanked_reason = anchor.get("data-yanked")

    link = Link(
        url,
        comes_from=page_url,
        requires_python=pyrequire,
        yanked_reason=yanked_reason,
    )

    return link


class CacheablePageContent:
    def __init__(self, page: "HTMLPage") -> None:
        assert page.cache_link_parsing
        self.page = page

    def __eq__(self, other: object) -> bool:
        return isinstance(other, type(self)) and self.page.url == other.page.url

    def __hash__(self) -> int:
        return hash(self.page.url)


def with_cached_html_pages(
    fn: Callable[["HTMLPage"], Iterable[Link]],
) -> Callable[["HTMLPage"], List[Link]]:
    """
    Given a function that parses an Iterable[Link] from an HTMLPage, cache the
    function's result (keyed by CacheablePageContent), unless the HTMLPage
    `page` has `page.cache_link_parsing == False`.
    """

    @functools.lru_cache(maxsize=None)
    def wrapper(cacheable_page: CacheablePageContent) -> List[Link]:
        return list(fn(cacheable_page.page))

    @functools.wraps(fn)
    def wrapper_wrapper(page: "HTMLPage") -> List[Link]:
        if page.cache_link_parsing:
            return wrapper(CacheablePageContent(page))
        return list(fn(page))

    return wrapper_wrapper


@with_cached_html_pages
def parse_links(page: "HTMLPage") -> Iterable[Link]:
    """
    Parse an HTML document, and yield its anchor elements as Link objects.
    """
    document = html5lib.parse(
        page.content,
        transport_encoding=page.encoding,
        namespaceHTMLElements=False,
    )

    url = page.url
    base_url = _determine_base_url(document, url)
    for anchor in document.findall(".//a"):
        link = _create_link_from_element(
            anchor,
            page_url=url,
            base_url=base_url,
        )
        if link is None:
            continue
        yield link


class HTMLPage:
    """Represents one page, along with its URL"""

    def __init__(
        self,
        content: bytes,
        encoding: Optional[str],
        url: str,
        cache_link_parsing: bool = True,
    ) -> None:
        """
        :param encoding: the encoding to decode the given content.
        :param url: the URL from which the HTML was downloaded.
        :param cache_link_parsing: whether links parsed from this page's url
                                   should be cached. PyPI index urls should
                                   have this set to False, for example.
        """
        self.content = content
        self.encoding = encoding
        self.url = url
        self.cache_link_parsing = cache_link_parsing

    def __str__(self) -> str:
        return redact_auth_from_url(self.url)


def _handle_get_page_fail(
    link: Link,
    reason: Union[str, Exception],
    meth: Optional[Callable[..., None]] = None,
) -> None:
    if meth is None:
        meth = logger.debug
    meth("Could not fetch URL %s: %s - skipping", link, reason)


def _make_html_page(response: Response, cache_link_parsing: bool = True) -> HTMLPage:
    encoding = _get_encoding_from_headers(response.headers)
    return HTMLPage(
        response.content,
        encoding=encoding,
        url=response.url,
        cache_link_parsing=cache_link_parsing,
    )


def _get_html_page(
    link: Link, session: Optional[PipSession] = None
) -> Optional["HTMLPage"]:
    if session is None:
        raise TypeError(
            "_get_html_page() missing 1 required keyword argument: 'session'"
        )

    url = link.url.split("#", 1)[0]

    # Check for VCS schemes that do not support lookup as web pages.
    vcs_scheme = _match_vcs_scheme(url)
    if vcs_scheme:
        logger.warning(
            "Cannot look at %s URL %s because it does not support lookup as web pages.",
            vcs_scheme,
            link,
        )
        return None

    # Tack index.html onto file:// URLs that point to directories
    scheme, _, path, _, _, _ = urllib.parse.urlparse(url)
    if scheme == "file" and os.path.isdir(urllib.request.url2pathname(path)):
        # add trailing slash if not present so urljoin doesn't trim
        # final segment
        if not url.endswith("/"):
            url += "/"
        url = urllib.parse.urljoin(url, "index.html")
        logger.debug(" file: URL is directory, getting %s", url)

    try:
        resp = _get_html_response(url, session=session)
    except _NotHTTP:
        logger.warning(
            "Skipping page %s because it looks like an archive, and cannot "
            "be checked by a HTTP HEAD request.",
            link,
        )
    except _NotHTML as exc:
        logger.warning(
            "Skipping page %s because the %s request got Content-Type: %s."
            "The only supported Content-Type is text/html",
            link,
            exc.request_desc,
            exc.content_type,
        )
    except NetworkConnectionError as exc:
        _handle_get_page_fail(link, exc)
    except RetryError as exc:
        _handle_get_page_fail(link, exc)
    except SSLError as exc:
        reason = "There was a problem confirming the ssl certificate: "
        reason += str(exc)
        _handle_get_page_fail(link, reason, meth=logger.info)
    except requests.ConnectionError as exc:
        _handle_get_page_fail(link, f"connection error: {exc}")
    except requests.Timeout:
        _handle_get_page_fail(link, "timed out")
    else:
        return _make_html_page(resp, cache_link_parsing=link.cache_link_parsing)
    return None


class CollectedSources(NamedTuple):
    find_links: Sequence[Optional[LinkSource]]
    index_urls: Sequence[Optional[LinkSource]]


class LinkCollector:

    """
    Responsible for collecting Link objects from all configured locations,
    making network requests as needed.

    The class's main method is its collect_sources() method.
    """

    def __init__(
        self,
        session: PipSession,
        search_scope: SearchScope,
    ) -> None:
        self.search_scope = search_scope
        self.session = session

    @classmethod
    def create(
        cls,
        session: PipSession,
        options: Values,
        suppress_no_index: bool = False,
    ) -> "LinkCollector":
        """
        :param session: The Session to use to make requests.
        :param suppress_no_index: Whether to ignore the --no-index option
            when constructing the SearchScope object.
        """
        index_urls = [options.index_url] + options.extra_index_urls
        if options.no_index and not suppress_no_index:
            logger.debug(
                "Ignoring indexes: %s",
                ",".join(redact_auth_from_url(url) for url in index_urls),
            )
            index_urls = []

        # Make sure find_links is a list before passing to create().
        find_links = options.find_links or []

        search_scope = SearchScope.create(
            find_links=find_links,
            index_urls=index_urls,
        )
        link_collector = LinkCollector(
            session=session,
            search_scope=search_scope,
        )
        return link_collector

    @property
    def find_links(self) -> List[str]:
        return self.search_scope.find_links

    def fetch_page(self, location: Link) -> Optional[HTMLPage]:
        """
        Fetch an HTML page containing package links.
        """
        return _get_html_page(location, session=self.session)

    def collect_sources(
        self,
        project_name: str,
        candidates_from_page: CandidatesFromPage,
    ) -> CollectedSources:
        # The OrderedDict calls deduplicate sources by URL.
        index_url_sources = collections.OrderedDict(
            build_source(
                loc,
                candidates_from_page=candidates_from_page,
                page_validator=self.session.is_secure_origin,
                expand_dir=False,
                cache_link_parsing=False,
            )
            for loc in self.search_scope.get_index_urls_locations(project_name)
        ).values()
        find_links_sources = collections.OrderedDict(
            build_source(
                loc,
                candidates_from_page=candidates_from_page,
                page_validator=self.session.is_secure_origin,
                expand_dir=True,
                cache_link_parsing=True,
            )
            for loc in self.find_links
        ).values()

        if logger.isEnabledFor(logging.DEBUG):
            lines = [
                f"* {s.link}"
                for s in itertools.chain(find_links_sources, index_url_sources)
                if s is not None and s.link is not None
            ]
            lines = [
                f"{len(lines)} location(s) to search "
                f"for versions of {project_name}:"
            ] + lines
            logger.debug("\n".join(lines))

        return CollectedSources(
            find_links=list(find_links_sources),
            index_urls=list(index_url_sources),
        )
site-packages/pip/_internal/exceptions.py000064400000030732150564745010014576 0ustar00"""Exceptions used throughout package"""

import configparser
from itertools import chain, groupby, repeat
from typing import TYPE_CHECKING, Dict, List, Optional, Union

from pip._vendor.pkg_resources import Distribution
from pip._vendor.requests.models import Request, Response

if TYPE_CHECKING:
    from hashlib import _Hash

    from pip._internal.metadata import BaseDistribution
    from pip._internal.req.req_install import InstallRequirement


class PipError(Exception):
    """Base pip exception"""


class ConfigurationError(PipError):
    """General exception in configuration"""


class InstallationError(PipError):
    """General exception during installation"""


class UninstallationError(PipError):
    """General exception during uninstallation"""


class NoneMetadataError(PipError):
    """
    Raised when accessing "METADATA" or "PKG-INFO" metadata for a
    pip._vendor.pkg_resources.Distribution object and
    `dist.has_metadata('METADATA')` returns True but
    `dist.get_metadata('METADATA')` returns None (and similarly for
    "PKG-INFO").
    """

    def __init__(
        self,
        dist: Union[Distribution, "BaseDistribution"],
        metadata_name: str,
    ) -> None:
        """
        :param dist: A Distribution object.
        :param metadata_name: The name of the metadata being accessed
            (can be "METADATA" or "PKG-INFO").
        """
        self.dist = dist
        self.metadata_name = metadata_name

    def __str__(self) -> str:
        # Use `dist` in the error message because its stringification
        # includes more information, like the version and location.
        return "None {} metadata found for distribution: {}".format(
            self.metadata_name,
            self.dist,
        )


class UserInstallationInvalid(InstallationError):
    """A --user install is requested on an environment without user site."""

    def __str__(self) -> str:
        return "User base directory is not specified"


class InvalidSchemeCombination(InstallationError):
    def __str__(self) -> str:
        before = ", ".join(str(a) for a in self.args[:-1])
        return f"Cannot set {before} and {self.args[-1]} together"


class DistributionNotFound(InstallationError):
    """Raised when a distribution cannot be found to satisfy a requirement"""


class RequirementsFileParseError(InstallationError):
    """Raised when a general error occurs parsing a requirements file line."""


class BestVersionAlreadyInstalled(PipError):
    """Raised when the most up-to-date version of a package is already
    installed."""


class BadCommand(PipError):
    """Raised when virtualenv or a command is not found"""


class CommandError(PipError):
    """Raised when there is an error in command-line arguments"""


class PreviousBuildDirError(PipError):
    """Raised when there's a previous conflicting build directory"""


class NetworkConnectionError(PipError):
    """HTTP connection error"""

    def __init__(
        self, error_msg: str, response: Response = None, request: Request = None
    ) -> None:
        """
        Initialize NetworkConnectionError with  `request` and `response`
        objects.
        """
        self.response = response
        self.request = request
        self.error_msg = error_msg
        if (
            self.response is not None
            and not self.request
            and hasattr(response, "request")
        ):
            self.request = self.response.request
        super().__init__(error_msg, response, request)

    def __str__(self) -> str:
        return str(self.error_msg)


class InvalidWheelFilename(InstallationError):
    """Invalid wheel filename."""


class UnsupportedWheel(InstallationError):
    """Unsupported wheel."""


class MetadataInconsistent(InstallationError):
    """Built metadata contains inconsistent information.

    This is raised when the metadata contains values (e.g. name and version)
    that do not match the information previously obtained from sdist filename
    or user-supplied ``#egg=`` value.
    """

    def __init__(
        self, ireq: "InstallRequirement", field: str, f_val: str, m_val: str
    ) -> None:
        self.ireq = ireq
        self.field = field
        self.f_val = f_val
        self.m_val = m_val

    def __str__(self) -> str:
        template = (
            "Requested {} has inconsistent {}: "
            "filename has {!r}, but metadata has {!r}"
        )
        return template.format(self.ireq, self.field, self.f_val, self.m_val)


class InstallationSubprocessError(InstallationError):
    """A subprocess call failed during installation."""

    def __init__(self, returncode: int, description: str) -> None:
        self.returncode = returncode
        self.description = description

    def __str__(self) -> str:
        return (
            "Command errored out with exit status {}: {} "
            "Check the logs for full command output."
        ).format(self.returncode, self.description)


class HashErrors(InstallationError):
    """Multiple HashError instances rolled into one for reporting"""

    def __init__(self) -> None:
        self.errors: List["HashError"] = []

    def append(self, error: "HashError") -> None:
        self.errors.append(error)

    def __str__(self) -> str:
        lines = []
        self.errors.sort(key=lambda e: e.order)
        for cls, errors_of_cls in groupby(self.errors, lambda e: e.__class__):
            lines.append(cls.head)
            lines.extend(e.body() for e in errors_of_cls)
        if lines:
            return "\n".join(lines)
        return ""

    def __bool__(self) -> bool:
        return bool(self.errors)


class HashError(InstallationError):
    """
    A failure to verify a package against known-good hashes

    :cvar order: An int sorting hash exception classes by difficulty of
        recovery (lower being harder), so the user doesn't bother fretting
        about unpinned packages when he has deeper issues, like VCS
        dependencies, to deal with. Also keeps error reports in a
        deterministic order.
    :cvar head: A section heading for display above potentially many
        exceptions of this kind
    :ivar req: The InstallRequirement that triggered this error. This is
        pasted on after the exception is instantiated, because it's not
        typically available earlier.

    """

    req: Optional["InstallRequirement"] = None
    head = ""
    order: int = -1

    def body(self) -> str:
        """Return a summary of me for display under the heading.

        This default implementation simply prints a description of the
        triggering requirement.

        :param req: The InstallRequirement that provoked this error, with
            its link already populated by the resolver's _populate_link().

        """
        return f"    {self._requirement_name()}"

    def __str__(self) -> str:
        return f"{self.head}\n{self.body()}"

    def _requirement_name(self) -> str:
        """Return a description of the requirement that triggered me.

        This default implementation returns long description of the req, with
        line numbers

        """
        return str(self.req) if self.req else "unknown package"


class VcsHashUnsupported(HashError):
    """A hash was provided for a version-control-system-based requirement, but
    we don't have a method for hashing those."""

    order = 0
    head = (
        "Can't verify hashes for these requirements because we don't "
        "have a way to hash version control repositories:"
    )


class DirectoryUrlHashUnsupported(HashError):
    """A hash was provided for a version-control-system-based requirement, but
    we don't have a method for hashing those."""

    order = 1
    head = (
        "Can't verify hashes for these file:// requirements because they "
        "point to directories:"
    )


class HashMissing(HashError):
    """A hash was needed for a requirement but is absent."""

    order = 2
    head = (
        "Hashes are required in --require-hashes mode, but they are "
        "missing from some requirements. Here is a list of those "
        "requirements along with the hashes their downloaded archives "
        "actually had. Add lines like these to your requirements files to "
        "prevent tampering. (If you did not enable --require-hashes "
        "manually, note that it turns on automatically when any package "
        "has a hash.)"
    )

    def __init__(self, gotten_hash: str) -> None:
        """
        :param gotten_hash: The hash of the (possibly malicious) archive we
            just downloaded
        """
        self.gotten_hash = gotten_hash

    def body(self) -> str:
        # Dodge circular import.
        from pip._internal.utils.hashes import FAVORITE_HASH

        package = None
        if self.req:
            # In the case of URL-based requirements, display the original URL
            # seen in the requirements file rather than the package name,
            # so the output can be directly copied into the requirements file.
            package = (
                self.req.original_link
                if self.req.original_link
                # In case someone feeds something downright stupid
                # to InstallRequirement's constructor.
                else getattr(self.req, "req", None)
            )
        return "    {} --hash={}:{}".format(
            package or "unknown package", FAVORITE_HASH, self.gotten_hash
        )


class HashUnpinned(HashError):
    """A requirement had a hash specified but was not pinned to a specific
    version."""

    order = 3
    head = (
        "In --require-hashes mode, all requirements must have their "
        "versions pinned with ==. These do not:"
    )


class HashMismatch(HashError):
    """
    Distribution file hash values don't match.

    :ivar package_name: The name of the package that triggered the hash
        mismatch. Feel free to write to this after the exception is raise to
        improve its error message.

    """

    order = 4
    head = (
        "THESE PACKAGES DO NOT MATCH THE HASHES FROM THE REQUIREMENTS "
        "FILE. If you have updated the package versions, please update "
        "the hashes. Otherwise, examine the package contents carefully; "
        "someone may have tampered with them."
    )

    def __init__(self, allowed: Dict[str, List[str]], gots: Dict[str, "_Hash"]) -> None:
        """
        :param allowed: A dict of algorithm names pointing to lists of allowed
            hex digests
        :param gots: A dict of algorithm names pointing to hashes we
            actually got from the files under suspicion
        """
        self.allowed = allowed
        self.gots = gots

    def body(self) -> str:
        return "    {}:\n{}".format(self._requirement_name(), self._hash_comparison())

    def _hash_comparison(self) -> str:
        """
        Return a comparison of actual and expected hash values.

        Example::

               Expected sha256 abcdeabcdeabcdeabcdeabcdeabcdeabcdeabcdeabcde
                            or 123451234512345123451234512345123451234512345
                    Got        bcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdefbcdef

        """

        def hash_then_or(hash_name: str) -> "chain[str]":
            # For now, all the decent hashes have 6-char names, so we can get
            # away with hard-coding space literals.
            return chain([hash_name], repeat("    or"))

        lines: List[str] = []
        for hash_name, expecteds in self.allowed.items():
            prefix = hash_then_or(hash_name)
            lines.extend(
                ("        Expected {} {}".format(next(prefix), e)) for e in expecteds
            )
            lines.append(
                "             Got        {}\n".format(self.gots[hash_name].hexdigest())
            )
        return "\n".join(lines)


class UnsupportedPythonVersion(InstallationError):
    """Unsupported python version according to Requires-Python package
    metadata."""


class ConfigurationFileCouldNotBeLoaded(ConfigurationError):
    """When there are errors while loading a configuration file"""

    def __init__(
        self,
        reason: str = "could not be loaded",
        fname: Optional[str] = None,
        error: Optional[configparser.Error] = None,
    ) -> None:
        super().__init__(error)
        self.reason = reason
        self.fname = fname
        self.error = error

    def __str__(self) -> str:
        if self.fname is not None:
            message_part = f" in {self.fname}."
        else:
            assert self.error is not None
            message_part = f".\n{self.error}\n"
        return f"Configuration file {self.reason}{message_part}"
site-packages/pip/_internal/pyproject.py000064400000016057150564745010014440 0ustar00import os
from collections import namedtuple
from typing import Any, List, Optional

from pip._vendor import tomli
from pip._vendor.packaging.requirements import InvalidRequirement, Requirement

from pip._internal.exceptions import InstallationError


def _is_list_of_str(obj: Any) -> bool:
    return isinstance(obj, list) and all(isinstance(item, str) for item in obj)


def make_pyproject_path(unpacked_source_directory: str) -> str:
    return os.path.join(unpacked_source_directory, "pyproject.toml")


BuildSystemDetails = namedtuple(
    "BuildSystemDetails", ["requires", "backend", "check", "backend_path"]
)


def load_pyproject_toml(
    use_pep517: Optional[bool], pyproject_toml: str, setup_py: str, req_name: str
) -> Optional[BuildSystemDetails]:
    """Load the pyproject.toml file.

    Parameters:
        use_pep517 - Has the user requested PEP 517 processing? None
                     means the user hasn't explicitly specified.
        pyproject_toml - Location of the project's pyproject.toml file
        setup_py - Location of the project's setup.py file
        req_name - The name of the requirement we're processing (for
                   error reporting)

    Returns:
        None if we should use the legacy code path, otherwise a tuple
        (
            requirements from pyproject.toml,
            name of PEP 517 backend,
            requirements we should check are installed after setting
                up the build environment
            directory paths to import the backend from (backend-path),
                relative to the project root.
        )
    """
    has_pyproject = os.path.isfile(pyproject_toml)
    has_setup = os.path.isfile(setup_py)

    if not has_pyproject and not has_setup:
        raise InstallationError(
            f"{req_name} does not appear to be a Python project: "
            f"neither 'setup.py' nor 'pyproject.toml' found."
        )

    if has_pyproject:
        with open(pyproject_toml, encoding="utf-8") as f:
            pp_toml = tomli.load(f)
        build_system = pp_toml.get("build-system")
    else:
        build_system = None

    # The following cases must use PEP 517
    # We check for use_pep517 being non-None and falsey because that means
    # the user explicitly requested --no-use-pep517.  The value 0 as
    # opposed to False can occur when the value is provided via an
    # environment variable or config file option (due to the quirk of
    # strtobool() returning an integer in pip's configuration code).
    if has_pyproject and not has_setup:
        if use_pep517 is not None and not use_pep517:
            raise InstallationError(
                "Disabling PEP 517 processing is invalid: "
                "project does not have a setup.py"
            )
        use_pep517 = True
    elif build_system and "build-backend" in build_system:
        if use_pep517 is not None and not use_pep517:
            raise InstallationError(
                "Disabling PEP 517 processing is invalid: "
                "project specifies a build backend of {} "
                "in pyproject.toml".format(build_system["build-backend"])
            )
        use_pep517 = True

    # If we haven't worked out whether to use PEP 517 yet,
    # and the user hasn't explicitly stated a preference,
    # we do so if the project has a pyproject.toml file.
    elif use_pep517 is None:
        use_pep517 = has_pyproject

    # At this point, we know whether we're going to use PEP 517.
    assert use_pep517 is not None

    # If we're using the legacy code path, there is nothing further
    # for us to do here.
    if not use_pep517:
        return None

    if build_system is None:
        # Either the user has a pyproject.toml with no build-system
        # section, or the user has no pyproject.toml, but has opted in
        # explicitly via --use-pep517.
        # In the absence of any explicit backend specification, we
        # assume the setuptools backend that most closely emulates the
        # traditional direct setup.py execution, and require wheel and
        # a version of setuptools that supports that backend.

        build_system = {
            "requires": ["setuptools>=40.8.0", "wheel"],
            "build-backend": "setuptools.build_meta:__legacy__",
        }

    # If we're using PEP 517, we have build system information (either
    # from pyproject.toml, or defaulted by the code above).
    # Note that at this point, we do not know if the user has actually
    # specified a backend, though.
    assert build_system is not None

    # Ensure that the build-system section in pyproject.toml conforms
    # to PEP 518.
    error_template = (
        "{package} has a pyproject.toml file that does not comply "
        "with PEP 518: {reason}"
    )

    # Specifying the build-system table but not the requires key is invalid
    if "requires" not in build_system:
        raise InstallationError(
            error_template.format(
                package=req_name,
                reason=(
                    "it has a 'build-system' table but not "
                    "'build-system.requires' which is mandatory in the table"
                ),
            )
        )

    # Error out if requires is not a list of strings
    requires = build_system["requires"]
    if not _is_list_of_str(requires):
        raise InstallationError(
            error_template.format(
                package=req_name,
                reason="'build-system.requires' is not a list of strings.",
            )
        )

    # Each requirement must be valid as per PEP 508
    for requirement in requires:
        try:
            Requirement(requirement)
        except InvalidRequirement:
            raise InstallationError(
                error_template.format(
                    package=req_name,
                    reason=(
                        "'build-system.requires' contains an invalid "
                        "requirement: {!r}".format(requirement)
                    ),
                )
            )

    backend = build_system.get("build-backend")
    backend_path = build_system.get("backend-path", [])
    check: List[str] = []
    if backend is None:
        # If the user didn't specify a backend, we assume they want to use
        # the setuptools backend. But we can't be sure they have included
        # a version of setuptools which supplies the backend, or wheel
        # (which is needed by the backend) in their requirements. So we
        # make a note to check that those requirements are present once
        # we have set up the environment.
        # This is quite a lot of work to check for a very specific case. But
        # the problem is, that case is potentially quite common - projects that
        # adopted PEP 518 early for the ability to specify requirements to
        # execute setup.py, but never considered needing to mention the build
        # tools themselves. The original PEP 518 code had a similar check (but
        # implemented in a different way).
        backend = "setuptools.build_meta:__legacy__"
        check = ["setuptools>=40.8.0", "wheel"]

    return BuildSystemDetails(requires, backend, check, backend_path)
site-packages/pip/_internal/configuration.py000064400000031541150564745010015263 0ustar00"""Configuration management setup

Some terminology:
- name
  As written in config files.
- value
  Value associated with a name
- key
  Name combined with it's section (section.name)
- variant
  A single word describing where the configuration key-value pair came from
"""

import configparser
import locale
import os
import sys
from typing import Any, Dict, Iterable, List, NewType, Optional, Tuple

from pip._internal.exceptions import (
    ConfigurationError,
    ConfigurationFileCouldNotBeLoaded,
)
from pip._internal.utils import appdirs
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.logging import getLogger
from pip._internal.utils.misc import ensure_dir, enum

RawConfigParser = configparser.RawConfigParser  # Shorthand
Kind = NewType("Kind", str)

CONFIG_BASENAME = "pip.ini" if WINDOWS else "pip.conf"
ENV_NAMES_IGNORED = "version", "help"

# The kinds of configurations there are.
kinds = enum(
    USER="user",  # User Specific
    GLOBAL="global",  # System Wide
    SITE="site",  # [Virtual] Environment Specific
    ENV="env",  # from PIP_CONFIG_FILE
    ENV_VAR="env-var",  # from Environment Variables
)
OVERRIDE_ORDER = kinds.GLOBAL, kinds.USER, kinds.SITE, kinds.ENV, kinds.ENV_VAR
VALID_LOAD_ONLY = kinds.USER, kinds.GLOBAL, kinds.SITE

logger = getLogger(__name__)


# NOTE: Maybe use the optionx attribute to normalize keynames.
def _normalize_name(name: str) -> str:
    """Make a name consistent regardless of source (environment or file)"""
    name = name.lower().replace("_", "-")
    if name.startswith("--"):
        name = name[2:]  # only prefer long opts
    return name


def _disassemble_key(name: str) -> List[str]:
    if "." not in name:
        error_message = (
            "Key does not contain dot separated section and key. "
            "Perhaps you wanted to use 'global.{}' instead?"
        ).format(name)
        raise ConfigurationError(error_message)
    return name.split(".", 1)


def get_configuration_files() -> Dict[Kind, List[str]]:
    global_config_files = [
        os.path.join(path, CONFIG_BASENAME) for path in appdirs.site_config_dirs("pip")
    ]

    site_config_file = os.path.join(sys.prefix, CONFIG_BASENAME)
    legacy_config_file = os.path.join(
        os.path.expanduser("~"),
        "pip" if WINDOWS else ".pip",
        CONFIG_BASENAME,
    )
    new_config_file = os.path.join(appdirs.user_config_dir("pip"), CONFIG_BASENAME)
    return {
        kinds.GLOBAL: global_config_files,
        kinds.SITE: [site_config_file],
        kinds.USER: [legacy_config_file, new_config_file],
    }


class Configuration:
    """Handles management of configuration.

    Provides an interface to accessing and managing configuration files.

    This class converts provides an API that takes "section.key-name" style
    keys and stores the value associated with it as "key-name" under the
    section "section".

    This allows for a clean interface wherein the both the section and the
    key-name are preserved in an easy to manage form in the configuration files
    and the data stored is also nice.
    """

    def __init__(self, isolated: bool, load_only: Optional[Kind] = None) -> None:
        super().__init__()

        if load_only is not None and load_only not in VALID_LOAD_ONLY:
            raise ConfigurationError(
                "Got invalid value for load_only - should be one of {}".format(
                    ", ".join(map(repr, VALID_LOAD_ONLY))
                )
            )
        self.isolated = isolated
        self.load_only = load_only

        # Because we keep track of where we got the data from
        self._parsers: Dict[Kind, List[Tuple[str, RawConfigParser]]] = {
            variant: [] for variant in OVERRIDE_ORDER
        }
        self._config: Dict[Kind, Dict[str, Any]] = {
            variant: {} for variant in OVERRIDE_ORDER
        }
        self._modified_parsers: List[Tuple[str, RawConfigParser]] = []

    def load(self) -> None:
        """Loads configuration from configuration files and environment"""
        self._load_config_files()
        if not self.isolated:
            self._load_environment_vars()

    def get_file_to_edit(self) -> Optional[str]:
        """Returns the file with highest priority in configuration"""
        assert self.load_only is not None, "Need to be specified a file to be editing"

        try:
            return self._get_parser_to_modify()[0]
        except IndexError:
            return None

    def items(self) -> Iterable[Tuple[str, Any]]:
        """Returns key-value pairs like dict.items() representing the loaded
        configuration
        """
        return self._dictionary.items()

    def get_value(self, key: str) -> Any:
        """Get a value from the configuration."""
        try:
            return self._dictionary[key]
        except KeyError:
            raise ConfigurationError(f"No such key - {key}")

    def set_value(self, key: str, value: Any) -> None:
        """Modify a value in the configuration."""
        self._ensure_have_load_only()

        assert self.load_only
        fname, parser = self._get_parser_to_modify()

        if parser is not None:
            section, name = _disassemble_key(key)

            # Modify the parser and the configuration
            if not parser.has_section(section):
                parser.add_section(section)
            parser.set(section, name, value)

        self._config[self.load_only][key] = value
        self._mark_as_modified(fname, parser)

    def unset_value(self, key: str) -> None:
        """Unset a value in the configuration."""
        self._ensure_have_load_only()

        assert self.load_only
        if key not in self._config[self.load_only]:
            raise ConfigurationError(f"No such key - {key}")

        fname, parser = self._get_parser_to_modify()

        if parser is not None:
            section, name = _disassemble_key(key)
            if not (
                parser.has_section(section) and parser.remove_option(section, name)
            ):
                # The option was not removed.
                raise ConfigurationError(
                    "Fatal Internal error [id=1]. Please report as a bug."
                )

            # The section may be empty after the option was removed.
            if not parser.items(section):
                parser.remove_section(section)
            self._mark_as_modified(fname, parser)

        del self._config[self.load_only][key]

    def save(self) -> None:
        """Save the current in-memory state."""
        self._ensure_have_load_only()

        for fname, parser in self._modified_parsers:
            logger.info("Writing to %s", fname)

            # Ensure directory exists.
            ensure_dir(os.path.dirname(fname))

            with open(fname, "w") as f:
                parser.write(f)

    #
    # Private routines
    #

    def _ensure_have_load_only(self) -> None:
        if self.load_only is None:
            raise ConfigurationError("Needed a specific file to be modifying.")
        logger.debug("Will be working with %s variant only", self.load_only)

    @property
    def _dictionary(self) -> Dict[str, Any]:
        """A dictionary representing the loaded configuration."""
        # NOTE: Dictionaries are not populated if not loaded. So, conditionals
        #       are not needed here.
        retval = {}

        for variant in OVERRIDE_ORDER:
            retval.update(self._config[variant])

        return retval

    def _load_config_files(self) -> None:
        """Loads configuration from configuration files"""
        config_files = dict(self.iter_config_files())
        if config_files[kinds.ENV][0:1] == [os.devnull]:
            logger.debug(
                "Skipping loading configuration files due to "
                "environment's PIP_CONFIG_FILE being os.devnull"
            )
            return

        for variant, files in config_files.items():
            for fname in files:
                # If there's specific variant set in `load_only`, load only
                # that variant, not the others.
                if self.load_only is not None and variant != self.load_only:
                    logger.debug("Skipping file '%s' (variant: %s)", fname, variant)
                    continue

                parser = self._load_file(variant, fname)

                # Keeping track of the parsers used
                self._parsers[variant].append((fname, parser))

    def _load_file(self, variant: Kind, fname: str) -> RawConfigParser:
        logger.verbose("For variant '%s', will try loading '%s'", variant, fname)
        parser = self._construct_parser(fname)

        for section in parser.sections():
            items = parser.items(section)
            self._config[variant].update(self._normalized_keys(section, items))

        return parser

    def _construct_parser(self, fname: str) -> RawConfigParser:
        parser = configparser.RawConfigParser()
        # If there is no such file, don't bother reading it but create the
        # parser anyway, to hold the data.
        # Doing this is useful when modifying and saving files, where we don't
        # need to construct a parser.
        if os.path.exists(fname):
            try:
                parser.read(fname)
            except UnicodeDecodeError:
                # See https://github.com/pypa/pip/issues/4963
                raise ConfigurationFileCouldNotBeLoaded(
                    reason="contains invalid {} characters".format(
                        locale.getpreferredencoding(False)
                    ),
                    fname=fname,
                )
            except configparser.Error as error:
                # See https://github.com/pypa/pip/issues/4893
                raise ConfigurationFileCouldNotBeLoaded(error=error)
        return parser

    def _load_environment_vars(self) -> None:
        """Loads configuration from environment variables"""
        self._config[kinds.ENV_VAR].update(
            self._normalized_keys(":env:", self.get_environ_vars())
        )

    def _normalized_keys(
        self, section: str, items: Iterable[Tuple[str, Any]]
    ) -> Dict[str, Any]:
        """Normalizes items to construct a dictionary with normalized keys.

        This routine is where the names become keys and are made the same
        regardless of source - configuration files or environment.
        """
        normalized = {}
        for name, val in items:
            key = section + "." + _normalize_name(name)
            normalized[key] = val
        return normalized

    def get_environ_vars(self) -> Iterable[Tuple[str, str]]:
        """Returns a generator with all environmental vars with prefix PIP_"""
        for key, val in os.environ.items():
            if key.startswith("PIP_"):
                name = key[4:].lower()
                if name not in ENV_NAMES_IGNORED:
                    yield name, val

    # XXX: This is patched in the tests.
    def iter_config_files(self) -> Iterable[Tuple[Kind, List[str]]]:
        """Yields variant and configuration files associated with it.

        This should be treated like items of a dictionary.
        """
        # SMELL: Move the conditions out of this function

        # environment variables have the lowest priority
        config_file = os.environ.get("PIP_CONFIG_FILE", None)
        if config_file is not None:
            yield kinds.ENV, [config_file]
        else:
            yield kinds.ENV, []

        config_files = get_configuration_files()

        # at the base we have any global configuration
        yield kinds.GLOBAL, config_files[kinds.GLOBAL]

        # per-user configuration next
        should_load_user_config = not self.isolated and not (
            config_file and os.path.exists(config_file)
        )
        if should_load_user_config:
            # The legacy config file is overridden by the new config file
            yield kinds.USER, config_files[kinds.USER]

        # finally virtualenv configuration first trumping others
        yield kinds.SITE, config_files[kinds.SITE]

    def get_values_in_config(self, variant: Kind) -> Dict[str, Any]:
        """Get values present in a config file"""
        return self._config[variant]

    def _get_parser_to_modify(self) -> Tuple[str, RawConfigParser]:
        # Determine which parser to modify
        assert self.load_only
        parsers = self._parsers[self.load_only]
        if not parsers:
            # This should not happen if everything works correctly.
            raise ConfigurationError(
                "Fatal Internal error [id=2]. Please report as a bug."
            )

        # Use the highest priority parser.
        return parsers[-1]

    # XXX: This is patched in the tests.
    def _mark_as_modified(self, fname: str, parser: RawConfigParser) -> None:
        file_parser_tuple = (fname, parser)
        if file_parser_tuple not in self._modified_parsers:
            self._modified_parsers.append(file_parser_tuple)

    def __repr__(self) -> str:
        return f"{self.__class__.__name__}({self._dictionary!r})"
site-packages/pip/_internal/resolution/__pycache__/base.cpython-39.pyc000064400000002104150564745010022011 0ustar00a

��ReG�@sRddlmZmZmZddlmZddlmZeeeegefZ	Gdd�d�Z
dS)�)�Callable�List�Optional)�InstallRequirement)�RequirementSetc@s6eZdZeeeed�dd�Zeeed�dd�ZdS)�BaseResolver)�	root_reqs�check_supported_wheels�returncCs
t��dS�N��NotImplementedError)�selfrr	�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/base.py�resolveszBaseResolver.resolve)�req_setr
cCs
t��dSrr)rrrrr�get_installation_ordersz#BaseResolver.get_installation_orderN)	�__name__�
__module__�__qualname__rr�boolrrrrrrrrs��rN)�typingrrrZpip._internal.req.req_installrZpip._internal.req.req_setr�str�InstallRequirementProviderrrrrr�<module>s�site-packages/pip/_internal/resolution/__pycache__/__init__.cpython-39.pyc000064400000000353150564745010022642 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/__init__.py�<module>�site-packages/pip/_internal/resolution/base.py000064400000001107150564745010015524 0ustar00from typing import Callable, List, Optional

from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_set import RequirementSet

InstallRequirementProvider = Callable[
    [str, Optional[InstallRequirement]], InstallRequirement
]


class BaseResolver:
    def resolve(
        self, root_reqs: List[InstallRequirement], check_supported_wheels: bool
    ) -> RequirementSet:
        raise NotImplementedError()

    def get_installation_order(
        self, req_set: RequirementSet
    ) -> List[InstallRequirement]:
        raise NotImplementedError()
site-packages/pip/_internal/resolution/resolvelib/found_candidates.py000064400000013111150564745010022250 0ustar00"""Utilities to lazily create and visit candidates found.

Creating and visiting a candidate is a *very* costly operation. It involves
fetching, extracting, potentially building modules from source, and verifying
distribution metadata. It is therefore crucial for performance to keep
everything here lazy all the way down, so we only touch candidates that we
absolutely need, and not "download the world" when we only need one version of
something.
"""

import functools
from collections.abc import Sequence
from typing import TYPE_CHECKING, Any, Callable, Iterator, Optional, Set, Tuple

from pip._vendor.packaging.version import _BaseVersion

from .base import Candidate

IndexCandidateInfo = Tuple[_BaseVersion, Callable[[], Optional[Candidate]]]

if TYPE_CHECKING:
    SequenceCandidate = Sequence[Candidate]
else:
    # For compatibility: Python before 3.9 does not support using [] on the
    # Sequence class.
    #
    # >>> from collections.abc import Sequence
    # >>> Sequence[str]
    # Traceback (most recent call last):
    #   File "<stdin>", line 1, in <module>
    # TypeError: 'ABCMeta' object is not subscriptable
    #
    # TODO: Remove this block after dropping Python 3.8 support.
    SequenceCandidate = Sequence


def _iter_built(infos: Iterator[IndexCandidateInfo]) -> Iterator[Candidate]:
    """Iterator for ``FoundCandidates``.

    This iterator is used when the package is not already installed. Candidates
    from index come later in their normal ordering.
    """
    versions_found: Set[_BaseVersion] = set()
    for version, func in infos:
        if version in versions_found:
            continue
        candidate = func()
        if candidate is None:
            continue
        yield candidate
        versions_found.add(version)


def _iter_built_with_prepended(
    installed: Candidate, infos: Iterator[IndexCandidateInfo]
) -> Iterator[Candidate]:
    """Iterator for ``FoundCandidates``.

    This iterator is used when the resolver prefers the already-installed
    candidate and NOT to upgrade. The installed candidate is therefore
    always yielded first, and candidates from index come later in their
    normal ordering, except skipped when the version is already installed.
    """
    yield installed
    versions_found: Set[_BaseVersion] = {installed.version}
    for version, func in infos:
        if version in versions_found:
            continue
        candidate = func()
        if candidate is None:
            continue
        yield candidate
        versions_found.add(version)


def _iter_built_with_inserted(
    installed: Candidate, infos: Iterator[IndexCandidateInfo]
) -> Iterator[Candidate]:
    """Iterator for ``FoundCandidates``.

    This iterator is used when the resolver prefers to upgrade an
    already-installed package. Candidates from index are returned in their
    normal ordering, except replaced when the version is already installed.

    The implementation iterates through and yields other candidates, inserting
    the installed candidate exactly once before we start yielding older or
    equivalent candidates, or after all other candidates if they are all newer.
    """
    versions_found: Set[_BaseVersion] = set()
    for version, func in infos:
        if version in versions_found:
            continue
        # If the installed candidate is better, yield it first.
        if installed.version >= version:
            yield installed
            versions_found.add(installed.version)
        candidate = func()
        if candidate is None:
            continue
        yield candidate
        versions_found.add(version)

    # If the installed candidate is older than all other candidates.
    if installed.version not in versions_found:
        yield installed


class FoundCandidates(SequenceCandidate):
    """A lazy sequence to provide candidates to the resolver.

    The intended usage is to return this from `find_matches()` so the resolver
    can iterate through the sequence multiple times, but only access the index
    page when remote packages are actually needed. This improve performances
    when suitable candidates are already installed on disk.
    """

    def __init__(
        self,
        get_infos: Callable[[], Iterator[IndexCandidateInfo]],
        installed: Optional[Candidate],
        prefers_installed: bool,
        incompatible_ids: Set[int],
    ):
        self._get_infos = get_infos
        self._installed = installed
        self._prefers_installed = prefers_installed
        self._incompatible_ids = incompatible_ids

    def __getitem__(self, index: Any) -> Any:
        # Implemented to satisfy the ABC check. This is not needed by the
        # resolver, and should not be used by the provider either (for
        # performance reasons).
        raise NotImplementedError("don't do this")

    def __iter__(self) -> Iterator[Candidate]:
        infos = self._get_infos()
        if not self._installed:
            iterator = _iter_built(infos)
        elif self._prefers_installed:
            iterator = _iter_built_with_prepended(self._installed, infos)
        else:
            iterator = _iter_built_with_inserted(self._installed, infos)
        return (c for c in iterator if id(c) not in self._incompatible_ids)

    def __len__(self) -> int:
        # Implemented to satisfy the ABC check. This is not needed by the
        # resolver, and should not be used by the provider either (for
        # performance reasons).
        raise NotImplementedError("don't do this")

    @functools.lru_cache(maxsize=1)
    def __bool__(self) -> bool:
        if self._prefers_installed and self._installed:
            return True
        return any(self)
site-packages/pip/_internal/resolution/resolvelib/__pycache__/resolver.cpython-39.pyc000064400000016244150564745010025120 0ustar00a

��Rel%�@s~ddlZddlZddlZddlmZmZmZmZmZm	Z	m
Z
ddlmZddl
mZmZddl
mZddlmZddlmZddlmZdd	lmZdd
lmZddlmZddlmZm Z dd
l!m"Z"ddl#m$Z$m%Z%ddl&m'Z'm(Z(ddl)m*Z*e�rddl+m,Z-e-e(e'e.fZ,e�/e0�Z1Gdd�de�Zde2eee.e2fd�dd�Z3e	e.efeee.e2fe	e2e.fd�dd�Z4dS)�N)�
TYPE_CHECKING�Dict�List�Optional�Set�Tuple�cast��canonicalize_name)�BaseReporter�ResolutionImpossible)�Resolver)�
DirectedGraph)�
WheelCache)�
PackageFinder)�RequirementPreparer)�InstallRequirement)�RequirementSet)�BaseResolver�InstallRequirementProvider)�PipProvider)�PipDebuggingReporter�PipReporter�)�	Candidate�Requirement)�Factory)�ResultcszeZdZhd�Zd
eeeeee	e	e	e	e	e
eeedfd��fdd�
Z
eee	ed�dd	�Zeeed
�dd�Z�ZS)r
>zto-satisfy-only�eagerzonly-if-neededN.)�preparer�finder�wheel_cache�make_install_req�
use_user_site�ignore_dependencies�ignore_installed�ignore_requires_python�force_reinstall�upgrade_strategy�py_version_infocsJt���|
|jvsJ�t||||||	|||d�	|_||_|
|_d|_dS)N)	r rr"r!r#r'r%r&r))�super�__init__�_allowed_strategiesr�factoryr$r(�_result)�selfrr r!r"r#r$r%r&r'r(r)��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/resolver.pyr+&s 
�zResolver.__init__)�	root_reqs�check_supported_wheels�returnc
Cs�|j�|�}t|j|j|j|j|jd�}dtjvr:t	�}nt
�}t||�}zd}|j|j
|d�}|_Wn@ty�}	z(|j�td|	�|j�}
|
|	�WYd}	~	n
d}	~	00t|d�}|j��D]�}|��}
|
dur�q�|j�|�}|dur�d|
_nr|jj�rd|
_n`|j|jk�rd|
_nJ|j�s(|j�r0d|
_n2|jr�|jjr�|jj�rXt� d	|
j!�q�d|
_nq�|j}|�r�|j"�r�d
j#|j!|j||j$�p�dd�}t�%|�|�&|
�q�|j'}|jj(�)|�|S)
N)r-�constraintsr$r(�user_requested�PIP_RESOLVER_DEBUGi��)�
max_roundsz,ResolutionImpossible[Requirement, Candidate])r5FTz�%s is already installed with the same version as the provided wheel. Use --force-reinstall to force an installation of the wheel.z�The candidate selected for download or install is a yanked version: {name!r} candidate (version {version} at {link})
Reason for being yanked: {reason}z<none given>)�name�version�link�reason)*r-�collect_root_requirementsrr7r$r(r8�os�environrr�
RLResolver�resolve�requirementsr.r�get_installation_errorrr�mapping�values�get_install_requirement�get_dist_to_uninstall�should_reinstallr'r<�is_editable�editable�source_link�is_file�is_wheel�logger�infor;�	is_yanked�format�
yanked_reason�warning�add_named_requirement�all_requirementsr� prepare_linked_requirements_more)r/r4r5�	collected�provider�reporter�resolver� try_to_avoid_resolution_too_deep�result�e�error�req_set�	candidate�ireq�installed_distr=�msg�reqsr2r2r3rCFs~�
���


��
�

zResolver.resolve)rar6cCs^|jdusJd��|jj}t|t|jj�dd�}t|j��tj	t
|d�dd�}dd	�|D�S)
aGet order for installation of requirements in RequirementSet.

        The returned list contains a requirement before another that depends on
        it. This helps ensure that the environment is kept consistent as they
        get installed one-by-one.

        The current implementation creates a topological ordering of the
        dependency graph, while breaking any cycles in the graph at arbitrary
        points. We make no guarantees about where the cycle would be broken,
        other than they would be broken.
        Nzmust call resolve() firstr)�expected_node_count)�weightsT)�key�reversecSsg|]\}}|�qSr2r2)�.0�_rcr2r2r3�
<listcomp>��z3Resolver.get_installation_order.<locals>.<listcomp>)r.�graph�get_topological_weights�lenrF�sortedrD�items�	functools�partial�_req_set_item_sorter)r/rarorh�sorted_itemsr2r2r3�get_installation_order�s��zResolver.get_installation_order)N)�__name__�
__module__�__qualname__r,rrrrr�bool�strr�intr+rrrrCrx�
__classcell__r2r2r0r3r
#s*��!�_�r
zDirectedGraph[Optional[str]])rorgr6csTt��i�ttdd�����fdd���d��ddks@J�t��|ksPJ��S)a�Assign weights to each node based on how "deep" they are.

    This implementation may change at any point in the future without prior
    notice.

    We take the length for the longest path to any node from root, ignoring any
    paths that contain a single node twice (i.e. cycles). This is done through
    a depth-first search through the graph, while keeping track of the path to
    the node.

    Cycles in the graph result would result in node being revisited while also
    being it's own path. In this case, take no action. This helps ensure we
    don't get stuck in a cycle.

    When assigning weight, the longer path (i.e. larger length) is preferred.
    N)�noder6csZ|�vrdS��|���|�D]}�|�q ��|���|d�}t|t����|<dS)Nr)�add�
iter_children�remove�get�maxrq)r��child�last_known_parent_count�ro�path�visitrhr2r3r��s


z&get_topological_weights.<locals>.visitr)�setrr}rq)rorgr2r�r3rp�srp)�itemrhr6cCst|d�}|||fS)a)Key function used to sort install requirements for installation.

    Based on the "weight" mapping calculated in ``get_installation_order()``.
    The canonical package name is returned as the second member as a tie-
    breaker to ensure the result is predictable, which is useful in tests.
    rr	)r�rhr;r2r2r3rv�s
rv)5rt�loggingr@�typingrrrrrrr�pip._vendor.packaging.utilsr
�pip._vendor.resolvelibrrr
rBZpip._vendor.resolvelib.structsr�pip._internal.cacher�"pip._internal.index.package_finderr� pip._internal.operations.preparerZpip._internal.req.req_installrZpip._internal.req.req_setr�pip._internal.resolution.baserr�,pip._internal.resolution.resolvelib.providerr�,pip._internal.resolution.resolvelib.reporterrr�baserrr-rZ pip._vendor.resolvelib.resolversrZRLResultr}�	getLoggerryrPr~rprvr2r2r2r3�<module>s<$
!�/

�site-packages/pip/_internal/resolution/resolvelib/__pycache__/factory.cpython-39.pyc000064400000044512150564745010024725 0ustar00a

��Re�j�@sBddlZddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
mZmZm
Z
mZmZmZmZmZmZddlmZddlmZddlmZmZddlmZddlmZmZddl m!Z!m"Z"m#Z#m$Z$m%Z%m&Z&dd	l'm(Z(dd
l)m*Z*ddl+m,Z,m-Z-ddl.m/Z/dd
l0m1Z1ddl2m3Z3ddl4m5Z5ddl6m7Z7m8Z8ddl9m:Z:ddl;m<Z<ddl=m>Z>ddl?m@Z@ddlAmBZBddlCmDZDddlEmFZFmGZGmHZHmIZIddlJmKZKmLZLmMZMmNZNmOZOmPZPmQZQddlRmSZSmTZTddlUmVZVmWZWmXZXmYZYe�rddlmZZZGdd�deZ�Z[e�\e]�Z^ed�Z_ee/e_fZ`Gd d!�d!e
�ZaGd"d#�d#�ZbdS)$�N)�
TYPE_CHECKING�Dict�	FrozenSet�Iterable�Iterator�List�Mapping�
NamedTuple�Optional�Sequence�Set�Tuple�TypeVar�cast)�InvalidRequirement)�SpecifierSet)�NormalizedName�canonicalize_name)�ResolutionImpossible)�
CacheEntry�
WheelCache)�DistributionNotFound�InstallationError�InstallationSubprocessError�MetadataInconsistent�UnsupportedPythonVersion�UnsupportedWheel)�
PackageFinder)�
get_scheme)�BaseDistribution�get_default_environment)�Link)�Wheel)�RequirementPreparer)�install_req_from_link_and_ireq)�InstallRequirement�check_invalid_constraint_type)�InstallRequirementProvider)�
get_supported)�Hashes)�
dist_location)�get_requirement)�running_under_virtualenv�)�	Candidate�CandidateVersion�
Constraint�Requirement)�AlreadyInstalledCandidate�
BaseCandidate�EditableCandidate�ExtrasCandidate�
LinkCandidate�RequiresPythonCandidate�as_base_candidate)�FoundCandidates�IndexCandidateInfo)�ExplicitRequirement�RequiresPythonRequirement�SpecifierRequirement�UnsatisfiableRequirement)�Protocolc@seZdZUeed<eed<dS)�
ConflictCause�requirement�parentN)�__name__�
__module__�__qualname__r<�__annotations__r.�rGrG��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/factory.pyr@Ls
r@�Cc@s:eZdZUeeed<eeefed<eee	fed<dS)�CollectedRootRequirements�requirements�constraints�user_requestedN)
rCrDrErr1rFr�strr0�intrGrGrGrHrJWs
rJc
@s*eZdZdAeeeeeeeeeee	e
dfdd�
dd�Zeed�dd��Z
edd	�d
d�Zeeeed�d
d�Zeeeeed�dd�Zeeeeeeeeeed�dd�Zeeeeeee
e ed�dd�Z!e e"eee#ed�dd�Z$ee%ee#ed�dd�Z&ee'ee e"fe'ee#efe%ee ed�dd �Z(ee eee"d!�d"d#�Z)e*ee+d$�d%d&�Z,ee-d'�d(d)�Z.dBeeee eee"d+�d,d-�Z/eee"d.�d/d0�Z0eeeee1d1�d2d3�Z2eeed'�d4d5�Z3ed6e4d7�d8d9�Z5e"eee6d:�d;d<�Z7d=e8ee%fe9d>�d?d@�Z:dS)C�FactoryN.)
�finder�preparer�make_install_req�wheel_cache�
use_user_site�force_reinstall�ignore_installed�ignore_requires_python�py_version_info�returnc
Cs�||_||_||_t|	�|_||_||_||_||_i|_	i|_
i|_i|_i|_
|svt�}
dd�|
jdd�D�|_ni|_dS)NcSsi|]}|j|�qSrG)�canonical_name)�.0�distrGrGrH�
<dictcomp>}s�z$Factory.__init__.<locals>.<dictcomp>F)�
local_only)�_finderrR�_wheel_cacher7�_python_candidate�_make_install_req_from_spec�_use_user_site�_force_reinstall�_ignore_requires_python�_build_failures�_link_candidate_cache�_editable_candidate_cache�_installed_candidate_cache�_extras_candidate_cacher �iter_installed_distributions�_installed_dists)�selfrQrRrSrTrUrVrWrXrY�envrGrGrH�__init__^s(
�
�
zFactory.__init__�rZcCs|jS�N)re�rnrGrGrHrV�szFactory.force_reinstall)�linkrZcCsB|js
dSt|j�}|�|jj���r*dS|j�d�}t|��dS)Nz+ is not a supported wheel on this platform.)�is_wheelr"�filename�	supportedr`�
target_python�get_tagsr)rnrt�wheel�msgrGrGrH�"_fail_if_link_is_unsupported_wheel�s
z*Factory._fail_if_link_is_unsupported_wheel)�base�extrasrZcCsFt|�|f}z|j|}Wn&ty@t||�}||j|<Yn0|Srr)�idrk�KeyErrorr5)rnr}r~�	cache_key�	candidaterGrGrH�_make_extras_candidate�s
zFactory._make_extras_candidate)r]r~�templaterZcCsRz|j|j}Wn,ty<t|||d�}||j|j<Yn0|sF|S|�||�S)N)�factory)rjr[r�r2r�)rnr]r~r�r}rGrGrH�_make_candidate_from_dist�sz!Factory._make_candidate_from_dist)rtr~r��name�versionrZc
Cs||jvrdS|jr�||jvr~zt|||||d�|j|<WnBttfy|}z&t�d||�||j|<WYd}~dSd}~00|j|}nx||jv�r�zt	|||||d�|j|<WnDttf�y�}z&t�d||�||j|<WYd}~dSd}~00|j|}|�s|S|�
||�S)N)r�r�r�zDiscarding %s. %s)rg�editablerir4rr�logger�warningrhr6r�)rnrtr~r�r�r��er}rGrGrH�_make_candidate_from_link�sD

�
�

z!Factory._make_candidate_from_link)�ireqs�	specifier�hashes�prefers_installed�incompatible_idsrZc	s�|sdS|d��jsJd��t�jj��t��|D]<}|jsFJd���|jjM��|jdd�M��t|j�O�q4ttd�������fdd�}t	t
d�������fd	d
�}t||�|��S)NrGrz)Candidates found on index must be PEP 508F)�trust_internetrqcsf�jr
dSz�j�}Wnty,YdS0�j|jdd�sBdS�j|��d�}t|��vrbdS|S)z6Get the candidate for the currently-installed version.NT)�prereleases)r]r~r�)rermr��containsr�r�r)�installed_distr�)r~r�r�rnr�r�rGrH�_get_installed_candidate�s �z@Factory._iter_found_candidates.<locals>._get_installed_candidatec	3sx�jj���d�}t|���}tdd�|D��}t|�D]:}|sJ|jjrJq8tj	�j
|j���|jd�}|j|fVq8dS)N)�project_namer�r�css|]}|jjVqdSrr)rt�	is_yanked)r\�icanrGrGrH�	<genexpr>�zUFactory._iter_found_candidates.<locals>.iter_index_candidate_infos.<locals>.<genexpr>)rtr~r�r�r�)r`�find_best_candidate�list�iter_applicable�all�reversedrtr��	functools�partialr�r�)�result�icans�
all_yankedr��func)r~r�r�rnr�r�rGrH�iter_index_candidate_infoss&��zBFactory._iter_found_candidates.<locals>.iter_index_candidate_infos)�reqrr��	frozensetr�r�r~r
r.rr:r9)	rnr�r�r�r�r��ireqr�r�rG)r~r�r�r�rnr�r�rH�_iter_found_candidates�s&  �zFactory._iter_found_candidates)�base_requirementsr~rZccsJ|D]@}|��\}}|durqt|�}|dus6Jd��|�||�VqdS)a8Produce explicit candidates from the base given an extra-ed package.

        :param base_requirements: Requirements known to the resolver. The
            requirements are guaranteed to not have extras.
        :param extras: The extras to inject into the explicit requirements'
            candidates.
        Nzno extras here)�get_candidate_lookupr8r�)rnr�r~r��lookup_cand�_�	base_candrGrGrH�#_iter_explicit_candidates_from_base2sz+Factory._iter_explicit_candidates_from_base)�
identifier�
constraintr�rZccsD|jD]8}|�|�|j|t�t||�t|�dd�}|r|VqdS)z�Produce explicit candidates from constraints.

        This creates "fake" InstallRequirement objects that are basically clones
        of what "should" be the template, but with original_link set to link.
        N�r~r�r�r�)�linksr|r�r�r$r)rnr�r�r�rtr�rGrGrH�!_iter_candidates_from_constraintsHs

�z)Factory._iter_candidates_from_constraints)r�rK�incompatibilitiesr�r�rZc	st�}g}��D]4}|��\}	}
|	dur4|�|	�|
dur|�|
�qt�t��:t��}|�|�	��
|jd�t|j
���Wd�n1s�0Y|r�z|�|j��|dd��Wnty�YdS0dd�|�
�d�D��|�s|�|�j�j|��S����fdd�|D�S)NrGr)r�cSsh|]}t|��qSrG)r�r\�crGrGrH�	<setcomp>�r�z*Factory.find_candidates.<locals>.<setcomp>c3sB|]:�t���vr����rt�fdd���D��r�VqdS)c3s|]}|���VqdSrr)�is_satisfied_by)r\r��r�rGrHr��r�z4Factory.find_candidates.<locals>.<genexpr>.<genexpr>N)rr�r�)r\�r�r��incompat_idsrKr�rHr��s

�z*Factory.find_candidates.<locals>.<genexpr>)�setr��add�append�
contextlib�suppressrr+�updater��getr�r�r~r�rr�r�r�)rnr�rKr�r�r��explicit_candidatesr�r��candr��parsed_requirementrGr�rH�find_candidates_sN	
��"����zFactory.find_candidates)r��requested_extrasrZcCs�|�|�s t�d|j|j�dS|js.t|�S|�|j�|j|jt	|j
�||jr\t|j�nddd�}|dur�|js�|j|j�t
t|j��S|�|�S)Nz6Ignoring %s: markers '%s' don't match your environmentr�)�
match_markersr��infor��markersrtr=r|r�r�r~rrgr>�make_requirement_from_candidate)rnr�r�r�rGrGrH�"_make_requirement_from_install_req�s,
��z*Factory._make_requirement_from_install_req)�
root_ireqsrZcCs�tgii�}t|�D]�\}}|jr�t|�}|r6t|��|��s@q|jsNJd��t|j�}||jvrv|j||M<q�t	�
|�|j|<q|j|dd�}|dur�q|jr�|j|j
vr�||j
|j<|j�|�q|S)NzConstraint must be namedrG)r�)rJ�	enumerater�r&rr�r�rrLr0�	from_ireqr��
user_suppliedrMrKr�)rnr��	collected�ir��problemr�r�rGrGrH�collect_root_requirements�s.

�z!Factory.collect_root_requirements)r�rZcCst|�Srr)r;)rnr�rGrGrHr��sz'Factory.make_requirement_from_candidaterG)r��
comes_fromr�rZcCs|�||�}|�||�Srr)rcr�)rnr�r�r�r�rGrGrH�make_requirement_from_spec�sz"Factory.make_requirement_from_spec)r�rZcCs"|jr
dSt|�sdSt||j�Srr)rfrNr<rb)rnr�rGrGrH� make_requires_python_requirement�s
z(Factory.make_requires_python_requirement)rtr�rZcCs*|jdus|jjrdS|jj||t�d�S)a�Look up the link in the wheel cache.

        If ``preparer.require_hashes`` is True, don't use the wheel cache,
        because cached wheels, always built locally, have different hashes
        than the files downloaded from the index server and thus throw false
        hash mismatches. Furthermore, cached wheels at present have
        nondeterministic contents due to file modification times.
        N)rt�package_name�supported_tags)rarR�require_hashes�get_cache_entryr()rnrtr�rGrGrH�get_wheel_cache_entry�s�zFactory.get_wheel_cache_entrycCs�|j�|j�}|durdSz@t|j�tjdddtjid�tjdddtjid�fvrXWdSWnt	ylYn0|j
sx|S|jr�|St�r�|j
r�d|j�d|j��}t|��dS)N�purelib�
rpm_prefixr})�scheme�vars�platlibzNWill not install to the user site because it will lack sys.path precedence to z in )rmr�r�r*�_dist�	sysconfig�get_path�sys�base_prefixr�rd�in_usersiter,�in_site_packages�raw_name�locationr)rnr�r]�messagerGrGrH�get_dist_to_uninstalls0�
���zFactory.get_dist_to_uninstallr@)�causesrZcCs�|sJd��|jj}t|�dkrVt|djj�}d|djj�d|�d|��}t|�Sd|�d�}|D]0}|j�	�}t|jj�}|d	|�d
|�d�7}qft|�S)Nz,Requires-Python error reported with no causer-rzPackage z requires a different Python: z not in z%Packages require a different Python. z not in:�
z (required by �))
rbr��lenrNrAr�rBr�r�format_for_error)rnr�r�r�r��cause�packagerGrGrH�_report_requires_python_error5s"���
z%Factory._report_requires_python_error)r�rBrZcCs�|durt|�}n|�d|j�d�}|j�|j�}dd�tdd�|D��D�}t�d|d�|�pbd	�t|�d
kr|t�	d�t
d|���S)
Nz (from r�cSsg|]}t|��qSrG)rN)r\�vrGrGrH�
<listcomp>Tr�z?Factory._report_single_requirement_conflict.<locals>.<listcomp>cSsh|]
}|j�qSrG)r�r�rGrGrHr�Tr�z>Factory._report_single_requirement_conflict.<locals>.<setcomp>zNCould not find a version that satisfies the requirement %s (from versions: %s)�, �nonezrequirements.txtz�HINT: You are attempting to install a package literally named "requirements.txt" (which cannot exist). Consider using the '-r' flag to install the packages listed in requirements.txtz#No matching distribution found for )rNr�r`�find_all_candidatesr��sortedr��critical�joinr�r)rnr�rB�req_disp�cands�versionsrGrGrH�#_report_single_requirement_conflictKs
��z+Factory._report_single_requirement_conflictz,ResolutionImpossible[Requirement, Candidate])r�rLrZcs�|jsJd���fdd�|jD�}|r6��td|��St|j�dkrh|jd\}}|j|vrh��||�Stttd�dd	�}ttd
�dd�}t	�}|jD],\}}|dur�|�
�}	n||�}	|�|	�q�|r�|t|��}
nd
}
d�
|
�}t�|�d}t	�}|jD]^\}}|j|v�r |�|j�|d}|�rH||j�d|j�d�}n|d}||�
�}�q|D]"}
||
j}|d|
�|��7}�qd|ddddd}t�|�td�S)Nz)Installation error reported with no causecs*g|]"}t|jt�r|j��j�s|�qSrG)�
isinstancerAr<r�rb)r\r�rsrGrHr�ps�z2Factory.get_installation_error.<locals>.<listcomp>zSequence[ConflictCause]r-r)�partsrZcSs2t|�dkr|dSd�|dd��d|dS)Nr-rr����z and )r�r)rrGrGrH�	text_join�sz1Factory.get_installation_error.<locals>.text_join)rBrZcSsF|��}|r|js$|j�d|j��St|jt�r<t|jj�St|j�S)Nz==)�get_install_requirementr�r�r�rr%rN)rBr�rGrGrH�describe_trigger�s
z8Factory.get_installation_error.<locals>.describe_triggerzthe requested packageszOCannot install {} because these package versions have conflicting dependencies.z
The conflict is caused by:z
    � z depends on zThe user requested z%
    The user requested (constraint) z

zTo fix this you could try to:
z91. loosen the range of package versions you've specified
z92. remove package versions to allow pip attempt to solve zthe dependency conflict
znResolutionImpossible: for help visit https://pip.pypa.io/en/latest/user_guide/#fixing-conflicting-dependencies)r�r�rr�r�rrrNr.r�r�r�r��formatr�r�r�r�r�r)rnr�rL�requires_python_causesr�rBrr
�triggers�triggerr�r{�relevant_constraints�key�specrGrsrH�get_installation_errorfsr
��	

��

������	
�zFactory.get_installation_error)N)rG);rCrDrErr#r'r
r�boolr
rOrp�propertyrVr!r|r3rrNr5r�rr%r.r�rr/r�rrr)rrr�r1rr�r0r�rr�r�rrJr�r;r�r�r�rr�r�rr�rrrrrrGrGrGrHrP]s���&
�
��5�V���G�!��	����'��
�rP)cr�r��loggingr�r��typingrrrrrrrr	r
rrr
rrZ"pip._vendor.packaging.requirementsrZ pip._vendor.packaging.specifiersr�pip._vendor.packaging.utilsrr�pip._vendor.resolvelibr�pip._internal.cacherr�pip._internal.exceptionsrrrrrr�"pip._internal.index.package_finderr�pip._internal.locationsr�pip._internal.metadatarr �pip._internal.models.linkr!�pip._internal.models.wheelr"� pip._internal.operations.preparer#�pip._internal.req.constructorsr$Zpip._internal.req.req_installr%r&�pip._internal.resolution.baser'�&pip._internal.utils.compatibility_tagsr(�pip._internal.utils.hashesr)�pip._internal.utils.miscr*�pip._internal.utils.packagingr+�pip._internal.utils.virtualenvr,r}r.r/r0r1�
candidatesr2r3r4r5r6r7r8�found_candidatesr9r:rKr;r<r=r>r?r@�	getLoggerrCr�rI�CacherJrPrGrGrGrH�<module>sJ@ $	
site-packages/pip/_internal/resolution/resolvelib/__pycache__/found_candidates.cpython-39.pyc000064400000011415150564745010026544 0ustar00a

��ReI�@s�dZddlZddlmZddlmZmZmZmZm	Z	m
Z
mZddlm
Z
ddlmZee
ege	effZerzeeZneZeeeed�d	d
�Zeeeeed�dd
�Zeeeeed�dd�ZGdd�de�ZdS)a�Utilities to lazily create and visit candidates found.

Creating and visiting a candidate is a *very* costly operation. It involves
fetching, extracting, potentially building modules from source, and verifying
distribution metadata. It is therefore crucial for performance to keep
everything here lazy all the way down, so we only touch candidates that we
absolutely need, and not "download the world" when we only need one version of
something.
�N)�Sequence)�
TYPE_CHECKING�Any�Callable�Iterator�Optional�Set�Tuple)�_BaseVersion�)�	Candidate)�infos�returnccsBt�}|D]2\}}||vrq
|�}|dur,q
|V|�|�q
dS)z�Iterator for ``FoundCandidates``.

    This iterator is used when the package is not already installed. Candidates
    from index come later in their normal ordering.
    N)�set�add)r
�versions_found�version�func�	candidate�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/found_candidates.py�_iter_built%sr)�	installedr
rccsJ|V|jh}|D]2\}}||vr$q|�}|dur4q|V|�|�qdS)aKIterator for ``FoundCandidates``.

    This iterator is used when the resolver prefers the already-installed
    candidate and NOT to upgrade. The installed candidate is therefore
    always yielded first, and candidates from index come later in their
    normal ordering, except skipped when the version is already installed.
    N)rr�rr
rrrrrrr�_iter_built_with_prepended6s
rccsnt�}|D]N\}}||vrq
|j|kr8|V|�|j�|�}|durHq
|V|�|�q
|j|vrj|VdS)a�Iterator for ``FoundCandidates``.

    This iterator is used when the resolver prefers to upgrade an
    already-installed package. Candidates from index are returned in their
    normal ordering, except replaced when the version is already installed.

    The implementation iterates through and yields other candidates, inserting
    the installed candidate exactly once before we start yielding older or
    equivalent candidates, or after all other candidates if they are all newer.
    N)rrrrrrr�_iter_built_with_insertedLs


rc@s�eZdZdZegeefeee	e
ed�dd�Ze
e
d�dd�Zeed�d	d
�Zed�dd�Zejd
d�e	d�dd��ZdS)�FoundCandidatesacA lazy sequence to provide candidates to the resolver.

    The intended usage is to return this from `find_matches()` so the resolver
    can iterate through the sequence multiple times, but only access the index
    page when remote packages are actually needed. This improve performances
    when suitable candidates are already installed on disk.
    )�	get_infosr�prefers_installed�incompatible_idscCs||_||_||_||_dS�N)�
_get_infos�
_installed�_prefers_installed�_incompatible_ids)�selfrrrrrrr�__init__uszFoundCandidates.__init__)�indexrcCstd��dS�Nz
don't do this��NotImplementedError)r%r'rrr�__getitem__�szFoundCandidates.__getitem__)rcsJ���}�jst|�}n �jr,t�j|�}nt�j|�}�fdd�|D�S)Nc3s |]}t|��jvr|VqdSr )�idr$)�.0�c�r%rr�	<genexpr>��z+FoundCandidates.__iter__.<locals>.<genexpr>)r!r"rr#rr)r%r
�iteratorrr/r�__iter__�s
zFoundCandidates.__iter__cCstd��dSr(r)r/rrr�__len__�szFoundCandidates.__len__r)�maxsizecCs|jr|jrdSt|�S)NT)r#r"�anyr/rrr�__bool__�szFoundCandidates.__bool__N)�__name__�
__module__�__qualname__�__doc__rr�IndexCandidateInforr�boolr�intr&rr+r3r4�	functools�	lru_cacher7rrrrrls
�

r)r;r?�collections.abcr�typingrrrrrrr	Zpip._vendor.packaging.versionr
�baserr<�SequenceCandidaterrrrrrrr�<module>s"
$
�� site-packages/pip/_internal/resolution/resolvelib/__pycache__/requirements.cpython-39.pyc000064400000016700150564745010025777 0ustar00a

��ReO�@s�ddlmZddlmZmZddlmZddlmZm	Z	m
Z
mZGdd�de
�ZGdd	�d	e
�Z
Gd
d�de
�ZGdd
�d
e
�ZdS)�)�SpecifierSet)�NormalizedName�canonicalize_name)�InstallRequirement�)�	Candidate�CandidateLookup�Requirement�format_namec@s�eZdZedd�dd�Zed�dd�Zed�dd	�Zee	d�d
d��Z
eed�dd
��Zed�dd�Ze
d�dd�Zeed�dd�ZdS)�ExplicitRequirementN��	candidate�returncCs
||_dS�N�r
��selfr
�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/requirements.py�__init__
szExplicitRequirement.__init__�rcCs
t|j�Sr)�strr
�rrrr�__str__
szExplicitRequirement.__str__cCsdj|jj|jd�S)Nz{class_name}({candidate!r}))�
class_namer
)�format�	__class__�__name__r
rrrr�__repr__s�zExplicitRequirement.__repr__cCs|jjSr)r
�project_namerrrrrsz ExplicitRequirement.project_namecCs|jjSr)r
�namerrrrr szExplicitRequirement.namecCs
|j��Sr)r
�format_for_errorrrrrr! sz$ExplicitRequirement.format_for_errorcCs
|jdfSrrrrrr�get_candidate_lookup#sz(ExplicitRequirement.get_candidate_lookupcCs
||jkSrrrrrr�is_satisfied_by&sz#ExplicitRequirement.is_satisfied_by)r�
__module__�__qualname__rrrrr�propertyrrr r!rr"�boolr#rrrrr	src@s�eZdZedd�dd�Zed�dd�Zed�dd	�Zee	d�d
d��Z
eed�dd
��Zed�dd�Ze
d�dd�Zeed�dd�ZdS)�SpecifierRequirementN)�ireqrcCs(|jdusJd��||_t|j�|_dS)NzThis is a link, not a specifier)�link�_ireq�	frozenset�extras�_extras)rr)rrrr+szSpecifierRequirement.__init__rcCst|jj�Sr)rr+�reqrrrrr0szSpecifierRequirement.__str__cCsdj|jjt|jj�d�S)Nz{class_name}({requirement!r}))r�requirement)rrrrr+r/rrrrr3s
�zSpecifierRequirement.__repr__cCs|jjsJd��t|jjj�S)N�'Specifier-backed ireq is always PEP 508)r+r/rr rrrrr9sz!SpecifierRequirement.project_namecCst|j|j�Sr)r
rr.rrrrr >szSpecifierRequirement.namecCsZdd�t|��d�D�}t|�dkr(dSt|�dkr<|dSd�|dd��d	|dS)
NcSsg|]}|���qSr)�strip)�.0�srrr�
<listcomp>H�z9SpecifierRequirement.format_for_error.<locals>.<listcomp>�,r�rz, ���z and )r�split�len�join)r�partsrrrr!Bsz%SpecifierRequirement.format_for_errorcCs
d|jfSr)r+rrrrr"Psz)SpecifierRequirement.get_candidate_lookuprcCsN|j|jks$Jd|j�d|j����|jjs4Jd��|jjj}|j|jdd�S)Nz6Internal issue: Candidate is not for this requirement z vs r1T��prereleases)r r+r/�	specifier�contains�version)rr
�specrrrr#Ss���
z$SpecifierRequirement.is_satisfied_by)rr$r%rrrrrr&rrr r!rr"rr'r#rrrrr(*sr(c@s�eZdZdZeedd�dd�Zed�dd�Zed�d	d
�Z	e
ed�dd��Ze
ed�d
d��Z
ed�dd�Zed�dd�Zeed�dd�ZdS)�RequiresPythonRequirementz4A requirement representing Requires-Python metadata.N)r@�matchrcCs||_||_dSr)r@�
_candidate)rr@rErrrrcsz"RequiresPythonRequirement.__init__rcCsd|j��S)NzPython )r@rrrrrgsz!RequiresPythonRequirement.__str__cCsdj|jjt|j�d�S)Nz{class_name}({specifier!r}))rr@)rrrrr@rrrrrjs�z"RequiresPythonRequirement.__repr__cCs|jjSr)rFrrrrrrpsz&RequiresPythonRequirement.project_namecCs|jjSr)rFr rrrrr tszRequiresPythonRequirement.namecCst|�Sr�rrrrrr!xsz*RequiresPythonRequirement.format_for_errorcCs"|jj|jjdd�r|jdfSdS)NTr>�NN)r@rArFrBrrrrr"{s
z.RequiresPythonRequirement.get_candidate_lookuprcCs(|j|jjksJd��|jj|jdd�S)NzNot Python candidateTr>)r rFr@rArBrrrrr#�sz)RequiresPythonRequirement.is_satisfied_by)rr$r%�__doc__rrrrrrr&rrr r!rr"r'r#rrrrrD`srDc@s�eZdZdZedd�dd�Zed�dd�Zed�d	d
�Ze	ed�dd��Z
e	ed�d
d��Zed�dd�Ze
d�dd�Zeed�dd�ZdS)�UnsatisfiableRequirementz'A requirement that cannot be satisfied.N)r rcCs
||_dSr��_name)rr rrrr�sz!UnsatisfiableRequirement.__init__rcCs|j�d�S)Nz (unavailable)rKrrrrr�sz UnsatisfiableRequirement.__str__cCsdj|jjt|j�d�S)Nz{class_name}({name!r}))rr )rrrrrLrrrrr�s�z!UnsatisfiableRequirement.__repr__cCs|jSrrKrrrrr�sz%UnsatisfiableRequirement.project_namecCs|jSrrKrrrrr �szUnsatisfiableRequirement.namecCst|�SrrGrrrrr!�sz)UnsatisfiableRequirement.format_for_errorcCsdS)NrHrrrrrr"�sz-UnsatisfiableRequirement.get_candidate_lookuprcCsdS)NFrrrrrr#�sz(UnsatisfiableRequirement.is_satisfied_by)rr$r%rIrrrrrr&rr r!rr"rr'r#rrrrrJ�srJN)Z pip._vendor.packaging.specifiersr�pip._vendor.packaging.utilsrrZpip._internal.req.req_installr�baserrr	r
rr(rDrJrrrr�<module>s!6(site-packages/pip/_internal/resolution/resolvelib/__pycache__/provider.cpython-39.pyc000064400000015754150564745010025116 0ustar00a

��Re�#�@s�ddlZddlZddlmZmZmZmZmZmZm	Z	ddl
mZddlm
Z
mZmZddlmZddlmZer�ddl
mZdd	lmZeee
fZeee
efZneZGd
d�de�ZdS)�N)�
TYPE_CHECKING�Dict�Iterable�Iterator�Mapping�Sequence�Union)�AbstractProvider�)�	Candidate�
Constraint�Requirement)�REQUIRES_PYTHON_IDENTIFIER)�Factory)�
Preference)�RequirementInformationc@seZdZdZeeeefeeeee	fdd�dd�Z
eee
fed�dd�Zeeee
feeee
feeed	fed	d
d�dd
�Zeed�dd�Zeeeeefeeee
fee
d�dd�Zee
ed�dd�Ze
eed�dd�Zeeed	ed�dd��ZdS)�PipProvidera�Pip's provider implementation for resolvelib.

    :params constraints: A mapping of constraints specified by the user. Keys
        are canonicalized project names.
    :params ignore_dependencies: Whether the user specified ``--no-deps``.
    :params upgrade_strategy: The user-specified upgrade strategy.
    :params user_requested: A set of canonicalized package names that the user
        supplied for pip to install/upgrade.
    N)�factory�constraints�ignore_dependencies�upgrade_strategy�user_requested�returncCs2||_||_||_||_||_t�dd��|_dS)NcSstjS�N)�math�inf�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/provider.py�<lambda>@�z&PipProvider.__init__.<locals>.<lambda>)�_factory�_constraints�_ignore_dependencies�_upgrade_strategy�_user_requested�collections�defaultdict�
_known_depths)�selfrrrrrrrr�__init__3szPipProvider.__init__)�requirement_or_candidatercCs|jSr��name)r(r*rrr�identifyBszPipProvider.identify�PreferenceInformationr)�
identifier�resolutions�
candidates�information�backtrack_causesrc	sdd�||D�}t|�\}}dd�dd�|D�D�}	|du}
tdd�|	D��}t|	�}z�j|}
WnDty�tj}
�fdd�||D�}td	d�|D��d
}Yn0d
}|�j|<�j�	|tj�}
|t
k}|dk}��||�}|||
||||
||f	S)a[Produce a sort key for given requirement based on preference.

        The lower the return value is, the more preferred this group of
        arguments is.

        Currently pip considers the followings in order:

        * Prefer if any of the known requirements is "direct", e.g. points to an
          explicit URL.
        * If equal, prefer if any requirement is "pinned", i.e. contains
          operator ``===`` or ``==``.
        * If equal, calculate an approximate "depth" and resolve requirements
          closer to the user-specified requirements first.
        * Order user-specified requirements by the order they are specified.
        * If equal, prefers "non-free" requirements, i.e. contains at least one
          operator, such as ``>=`` or ``<``.
        * If equal, order alphabetically for consistency (helps debuggability).
        css|]\}}|��VqdSr)�get_candidate_lookup)�.0�r�_rrr�	<genexpr>_rz-PipProvider.get_preference.<locals>.<genexpr>cSsg|]}|D]
}|j�qqSr)�operator)r5�
specifier_set�	specifierrrr�
<listcomp>as�z.PipProvider.get_preference.<locals>.<listcomp>css|]}|r|jVqdSr)r;)r5�ireqrrrr8crNcss|]}|dd�dkVqdS)N�z==r)r5�oprrrr8hrc3s*|]"\}}|dur�j|jndVqdS)Ng)r'r,)r5r7�parent�r(rrr8os�css|]
}|VqdSrr)r5�drrrr8srg�?�
setuptools)�zip�any�boolr$�KeyErrorrr�minr'�getr�is_backtrack_cause)r(r/r0r1r2r3�lookups�	candidate�ireqs�	operators�direct�pinned�unfree�requested_order�
parent_depths�inferred_depth�requires_python�
delay_this�backtrack_causerrAr�get_preferenceEs@�
�

�zPipProvider.get_preference)r/rcCsD||jvr|j|S|�d�\}}}|r<||jvr<|j|St��S)N�[)r!�	partitionr�empty)r(r/r,�open_bracketr7rrr�_get_constraint�s


zPipProvider._get_constraint)r/�requirements�incompatibilitiesrcs6ttd��fdd�}�jj||��|�||�|d�S)N)r,rcs&�jdkrdS�jdkr"|�jvSdS)a�Are upgrades allowed for this project?

            This checks the upgrade strategy, and whether the project was one
            that the user specified in the command line, in order to decide
            whether we should upgrade if there's a newer version available.

            (Note that we don't need access to the `--upgrade` flag, because
            an upgrade strategy of "to-satisfy-only" means that `--upgrade`
            was not specified).
            �eagerTzonly-if-neededF)r#r$r+rArr�_eligible_for_upgrade�s



z7PipProvider.find_matches.<locals>._eligible_for_upgrade)r/r^�
constraint�prefers_installedr_)�strrFr �find_candidatesr])r(r/r^r_rarrAr�find_matches�s�zPipProvider.find_matches)�requirementrLrcCs
|�|�Sr)�is_satisfied_by)r(rgrLrrrrh�szPipProvider.is_satisfied_by)rLrcCs|j}dd�|�|�D�S)NcSsg|]}|dur|�qSrr)r5r6rrrr<�rz0PipProvider.get_dependencies.<locals>.<listcomp>)r"�iter_dependencies)r(rL�
with_requiresrrr�get_dependencies�szPipProvider.get_dependencies)r/r3rcCs8|D].}||jjkrdS|jr||jjkrdSqdS)NTF)rgr,r@)r/r3rWrrrrJ�szPipProvider.is_backtrack_cause)�__name__�
__module__�__qualname__�__doc__rrrdrrF�intr)rr
rr-rrrrrXr]rfrhrk�staticmethodrJrrrrr(s8

�
�T��r)r%r�typingrrrrrrrZ pip._vendor.resolvelib.providersr	�baserrr
r1rrrrZ pip._vendor.resolvelib.resolversrr.rd�
_ProviderBaserrrrr�<module>s$site-packages/pip/_internal/resolution/resolvelib/__pycache__/base.cpython-39.pyc000064400000014763150564745010024175 0ustar00a

��Red�@s�ddlmZmZmZmZmZddlmZddlm	Z	m
Z
ddlmZm
Z
ddlmZmZddlmZddlmZeedeefZeee
fZeeeed	�d
d�ZGdd
�d
�ZGdd�d�Zeded�dd�ZGdd�d�ZdS)�)�	FrozenSet�Iterable�Optional�Tuple�Union)�SpecifierSet)�NormalizedName�canonicalize_name)�
LegacyVersion�Version)�Link�links_equivalent)�InstallRequirement)�Hashes�	Candidate)�project�extras�returncCs,|s|Stdd�|D��}d�|d�|��S)Ncss|]}t|�VqdS�N)r	)�.0�e�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/base.py�	<genexpr>�zformat_name.<locals>.<genexpr>z{}[{}]�,)�sorted�format�join)rr�canonical_extrasrrr�format_namesr c@sxeZdZeeeedd�dd�Zedd�dd��Z	ee
dd�d	d
��Zed�dd�Z
e
dd
�dd�Zded�dd�ZdS)�
ConstraintN)�	specifier�hashes�linksrcCs||_||_||_dSr)r"r#r$)�selfr"r#r$rrr�__init__szConstraint.__init__�rcCstt�t�t��Sr)r!rr�	frozenset)�clsrrr�emptyszConstraint.empty)�ireqrcCs.|jrt|jg�nt�}t|j|jdd�|�S�NF)�trust_internet)�linkr(r!r"r#)r)r+r$rrr�	from_ireq"szConstraint.from_ireqcCst|j�pt|j�pt|j�Sr)�boolr"r#r$�r%rrr�__bool__'szConstraint.__bool__)�otherrcCsRt|t�stS|j|j@}|j|jdd�@}|j}|jrF|�|jg�}t|||�Sr,)	�
isinstancer�NotImplementedr"r#r$r.�unionr!)r%r3r"r#r$rrr�__and__*s
zConstraint.__and__r��	candidatercs4|jr"t�fdd�|jD��s"dS|jj�jdd�S)Nc3s|]}t|��VqdSr)�_match_link)rr.�r9rrr6rz-Constraint.is_satisfied_by.<locals>.<genexpr>FT)�prereleases)r$�allr"�contains�version�r%r9rr;r�is_satisfied_by4szConstraint.is_satisfied_by)�__name__�
__module__�__qualname__rrrrr&�classmethodr*rr/r0r2r7rArrrrr!s
�
r!c@s\eZdZeed�dd��Zeed�dd��Zded�dd	�Z	e
d�d
d�Zed�dd
�ZdS)�Requirementr'cCstd��dS)z�The "project name" of a requirement.

        This is different from ``name`` if this requirement contains extras,
        in which case ``name`` would contain the ``[...]`` part, while this
        refers to the name of the project.
        �Subclass should overrideN��NotImplementedErrorr1rrr�project_name?szRequirement.project_namecCstd��dS)z�The name identifying this requirement in the resolver.

        This is different from ``project_name`` if this requirement contains
        extras, where ``project_name`` would not contain the ``[...]`` part.
        rGNrHr1rrr�nameIszRequirement.namerr8cCsdS�NFrr@rrrrARszRequirement.is_satisfied_bycCstd��dS�NrGrHr1rrr�get_candidate_lookupUsz Requirement.get_candidate_lookupcCstd��dSrMrHr1rrr�format_for_errorXszRequirement.format_for_errorN)
rBrCrD�propertyrrJ�strrKr0rA�CandidateLookuprNrOrrrrrF>s	rF)r.r9rcCs|jrt||j�SdSrL)�source_linkr
)r.r9rrrr:\sr:c@s�eZdZeed�dd��Zeed�dd��Zeed�dd��Z	ee
d�dd	��Zee
d�d
d��Zee
ed�dd
��Ze
ee
ed�dd�Ze
ed�dd�Zed�dd�ZdS)rr'cCstd��dS)z�The "project name" of the candidate.

        This is different from ``name`` if this candidate contains extras,
        in which case ``name`` would contain the ``[...]`` part, while this
        refers to the name of the project.
        �Override in subclassNrHr1rrrrJcszCandidate.project_namecCstd��dS)z�The name identifying this candidate in the resolver.

        This is different from ``project_name`` if this candidate contains
        extras, where ``project_name`` would not contain the ``[...]`` part.
        rTNrHr1rrrrKmszCandidate.namecCstd��dS�NrTrHr1rrrr?vszCandidate.versioncCstd��dSrUrHr1rrr�is_installedzszCandidate.is_installedcCstd��dSrUrHr1rrr�is_editable~szCandidate.is_editablecCstd��dSrUrHr1rrrrS�szCandidate.source_link)�
with_requiresrcCstd��dSrUrH)r%rXrrr�iter_dependencies�szCandidate.iter_dependenciescCstd��dSrUrHr1rrr�get_install_requirement�sz!Candidate.get_install_requirementcCstd��dSrMrHr1rrrrO�szCandidate.format_for_errorN)rBrCrDrPrrJrQrK�CandidateVersionr?r0rVrWrrrSrrFrYrrZrOrrrrrbs	N)�typingrrrrrZ pip._vendor.packaging.specifiersr�pip._vendor.packaging.utilsrr	Zpip._vendor.packaging.versionr
r�pip._internal.models.linkrr
Zpip._internal.req.req_installr�pip._internal.utils.hashesrrRr[rQr r!rFr0r:rrrrr�<module>s(site-packages/pip/_internal/resolution/resolvelib/__pycache__/__init__.cpython-39.pyc000064400000000366150564745010025014 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/__init__.py�<module>�site-packages/pip/_internal/resolution/resolvelib/__pycache__/candidates.cpython-39.pyc000064400000044157150564745010025362 0ustar00a

��Re"G�@s�ddlZddlZddlmZmZmZmZmZmZm	Z	m
Z
ddlmZm
Z
ddlmZddlmZmZddlmZddlmZmZddlmZdd	lmZmZdd
lmZddlm Z dd
l!m"Z"m#Z#m$Z$m%Z%er�ddl&m'Z'e�(e)�Z*e	dZ+e
ed�Z,e"ee+d�dd�Z-eeed�dd�Z.eeed�dd�Z/eeed�dd�Z0Gdd�de"�Z1Gdd�de1�Z2Gd d!�d!e1�Z3Gd"d#�d#e"�Z4Gd$d%�d%e"�Z5Gd&d'�d'e"�Z6dS)(�N)�
TYPE_CHECKING�Any�	FrozenSet�Iterable�Optional�Tuple�Union�cast)�NormalizedName�canonicalize_name)�Version)�	HashError�MetadataInconsistent)�BaseDistribution)�Link�links_equivalent)�Wheel)�install_req_from_editable�install_req_from_line)�InstallRequirement)�normalize_version_info�)�	Candidate�CandidateVersion�Requirement�format_name)�Factory)�AlreadyInstalledCandidate�EditableCandidate�
LinkCandidatez<Python from Requires-Python>)�	candidate�returncCstttf}t||�r|SdS)z%The runtime version of BaseCandidate.N)rrr�
isinstance)r �base_candidate_classes�r$��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/candidates.py�as_base_candidate$s�
r&)�link�templater!cCsh|jrJd��|jr t|j�}n|j}t||j|j|j|j|j	t
|j|j|j
d�d�}|j|_||_|S)Nztemplate is editable��install_options�global_options�hashes��
user_supplied�
comes_from�
use_pep517�isolated�
constraint�options)�editable�req�str�urlrr.r/r0r1r2�dictr*r+�hash_options�
original_linkr')r'r(�line�ireqr$r$r%�make_install_req_from_link0s(��
r=c
CsD|jsJd��t|j|j|j|j|j|j|jt	|j
|j|jd�d�S)Nztemplate not editabler))r.r/r0r1r2�permit_editable_wheelsr3)
r4rr7r.r/r0r1r2r>r8r*r+r9)r'r(r$r$r%�make_install_req_from_editableJs��r?)�distr(r!cCs�ddlm}|jrt|j�}n.|jr:|j�d|jj��}n|j�d|j��}t||j	|j
|j|j|j
t|j|j|jd�d�}t||�j|_|S)Nr)�Distributionz @ z==r)r-)�$pip._internal.metadata.pkg_resourcesrAr5r6r'�canonical_namer7�versionrr.r/r0r1r2r8r*r+r9r	�_dist�satisfied_by)r@r(�_Distr;r<r$r$r%�_make_install_req_from_dist^s*��
rHc	@s,eZdZUdZeed<dZd)eeede	e
e	edd�dd�Ze
d	�d
d�Ze
d	�dd
�Zed	�dd�Zeed�dd�Zee	ed	�dd��Zee
d	�dd��Zee
d	�dd��Zeed	�dd��Ze
d	�dd�Zed	�dd�Zedd�d d!�Zed	�d"d#�Zeee	ed$�d%d&�Z e	ed	�d'd(�Z!dS)*�"_InstallRequirementBackedCandidatea�A candidate backed by an ``InstallRequirement``.

    This represents a package request with the target not being already
    in the environment, and needs to be fetched and installed. The backing
    ``InstallRequirement`` is responsible for most of the leg work; this
    class exposes appropriate information to the resolver.

    :param link: The link passed to the ``InstallRequirement``. The backing
        ``InstallRequirement`` will use this link to fetch the distribution.
    :param source_link: The link this candidate "originates" from. This is
        different from ``link`` when the link is found in the wheel cache.
        ``link`` would point to the wheel cache, while this points to the
        found remote link (e.g. from pypi.org).
    r@FNr)r'�source_linkr<�factory�namerDr!cCs2||_||_||_||_||_||_|��|_dS�N)�_link�_source_link�_factory�_ireq�_name�_version�_preparer@)�selfr'rJr<rKrLrDr$r$r%�__init__�s	z+_InstallRequirementBackedCandidate.__init__�r!cCs|j�d|j��S)N� �rLrD�rUr$r$r%�__str__�sz*_InstallRequirementBackedCandidate.__str__cCsdj|jjt|j�d�S)Nz{class_name}({link!r}))�
class_namer')�format�	__class__�__name__r6rNrZr$r$r%�__repr__�s�z+_InstallRequirementBackedCandidate.__repr__cCst|j|jf�SrM)�hashr^rNrZr$r$r%�__hash__�sz+_InstallRequirementBackedCandidate.__hash__��otherr!cCst||j�rt|j|j�SdS�NF)r"r^rrN�rUrdr$r$r%�__eq__�sz)_InstallRequirementBackedCandidate.__eq__cCs|jSrM)rOrZr$r$r%rJ�sz._InstallRequirementBackedCandidate.source_linkcCs|jdur|jj|_|jS)�:The normalised name of the project the candidate refers toN)rRr@rCrZr$r$r%�project_name�s

z/_InstallRequirementBackedCandidate.project_namecCs|jSrM�rirZr$r$r%rL�sz'_InstallRequirementBackedCandidate.namecCs|jdur|jj|_|jSrM)rSr@rDrZr$r$r%rD�s

z*_InstallRequirementBackedCandidate.versioncCs$d�|j|j|jjr|jjn|j�S)Nz{} {} (from {}))r]rLrDrN�is_file�	file_pathrZr$r$r%�format_for_error�s
�z3_InstallRequirementBackedCandidate.format_for_errorcCstd��dS)NzOverride in subclass)�NotImplementedErrorrZr$r$r%�_prepare_distribution�sz8_InstallRequirementBackedCandidate._prepare_distribution)r@r!cCs`|jdur*|j|jkr*t|jd|j|j��|jdur\|j|jkr\t|jdt|j�t|j���dS)z:Check for consistency of project name and version of dist.NrLrD)rRrCrrQrSrDr6)rUr@r$r$r%�_check_metadata_consistency�s��z>_InstallRequirementBackedCandidate._check_metadata_consistencyc
CsJz|��}Wn.ty:}z|j|_�WYd}~n
d}~00|�|�|SrM)ror
rQr5rp)rUr@�er$r$r%rT�s
z+_InstallRequirementBackedCandidate._prepare��
with_requiresr!ccsH|r|j��nd}|D]}|j�t|�|j�Vq|j�|jj�VdS�Nr$)r@�iter_dependenciesrP�make_requirement_from_specr6rQ� make_requires_python_requirement�requires_python)rUrs�requires�rr$r$r%ru�sz4_InstallRequirementBackedCandidate.iter_dependenciescCs|jSrM)rQrZr$r$r%�get_install_requirement�sz:_InstallRequirementBackedCandidate.get_install_requirement)NN)"r_�
__module__�__qualname__�__doc__r�__annotations__�is_installedrrrr
rrVr6r[r`�intrbr�boolrg�propertyrJrirLrDrmrorprTrrrur{r$r$r$r%rIzs@
��rIcsHeZdZdZd
eedeeeedd��fdd�
Z	e
d�dd	�Z�ZS)rFNr�r'r(rKrLrDr!cs�|}|�||�}|dur,t�d|j�|j}t||�}|j|ksDJ�|jjr�|jjs�t|jj�}	t	|	j
�}
||
ks�J|�d|
�d���|dur�t|	j�}||ks�Jd�
|||���|dur�|jr�|j|jur�d|_t�j||||||d�dS)NzUsing cached wheel link: %sz != z
 for wheelz{!r} != {!r} for wheel {}T�r'rJr<rKrLrD)�get_wheel_cache_entry�logger�debugr'r=�is_wheelrkr�filenamerrLrrDr]�
persistentr:�original_link_is_in_wheel_cache�superrV)rUr'r(rKrLrDrJ�cache_entryr<�wheel�
wheel_name�
wheel_version�r^r$r%rV�s>


���
��zLinkCandidate.__init__rWcCs|jj}|j|jdd�S)NT)�parallel_builds)rP�preparer�prepare_linked_requirementrQ)rUr�r$r$r%ro"sz#LinkCandidate._prepare_distribution)NN�
r_r|r}�is_editablerrrr
rrVrro�
__classcell__r$r$r�r%r�s��*rcsHeZdZdZd
eedeeeedd��fdd�
Z	e
d�dd	�Z�ZS)rTNrr�cs"t�j||t||�|||d�dS)Nr�)r�rVr?)rUr'r(rKrLrDr�r$r%rV*s�zEditableCandidate.__init__rWcCs|jj�|j�SrM)rPr��prepare_editable_requirementrQrZr$r$r%ro;sz'EditableCandidate._prepare_distribution)NNr�r$r$r�r%r's��rc@s�eZdZdZdZeeddd�dd�Zed�dd	�Z	ed�d
d�Z
ed�dd
�Ze
ed�dd�Zeed�dd��Zeed�dd��Zeed�dd��Zeed�dd��Zed�dd�Zeeeed�dd�Zeed�dd�ZdS) rTNr)r@r(rKr!cCs0||_t||�|_||_d}|j�|j|�dS)Nzalready satisfied)r@rHrQrPr��prepare_installed_requirement)rUr@r(rK�skip_reasonr$r$r%rVCs
z"AlreadyInstalledCandidate.__init__rWcCs
t|j�SrM)r6r@rZr$r$r%r[Tsz!AlreadyInstalledCandidate.__str__cCsdj|jj|jd�S)Nz{class_name}({distribution!r}))r\�distribution)r]r^r_r@rZr$r$r%r`Ws�z"AlreadyInstalledCandidate.__repr__cCst|j|j|jf�SrM)rar^rLrDrZr$r$r%rb]sz"AlreadyInstalledCandidate.__hash__rccCs(t||j�r$|j|jko"|j|jkSdSre)r"r^rLrDrfr$r$r%rg`sz AlreadyInstalledCandidate.__eq__cCs|jjSrM)r@rCrZr$r$r%riesz&AlreadyInstalledCandidate.project_namecCs|jSrMrjrZr$r$r%rLiszAlreadyInstalledCandidate.namecCs|jjSrM)r@rDrZr$r$r%rDmsz!AlreadyInstalledCandidate.versioncCs|jjSrM)r@r4rZr$r$r%r�qsz%AlreadyInstalledCandidate.is_editablecCs|j�d|j�d�S)NrXz (Installed)rYrZr$r$r%rmusz*AlreadyInstalledCandidate.format_for_errorrrccs2|sdS|j��D]}|j�t|�|j�VqdSrM)r@rurPrvr6rQ)rUrsrzr$r$r%ruxsz+AlreadyInstalledCandidate.iter_dependenciescCsdSrMr$rZr$r$r%r{~sz1AlreadyInstalledCandidate.get_install_requirement)r_r|r}r�rJrrrVr6r[r`r�rbrr�rgr�r
rirLrrDr�rmrrrrur{r$r$r$r%r?s,�rc@seZdZdZeeedd�dd�Zed�dd�Zed�d	d
�Z	e
d�dd�Zee
d
�dd�Zeed�dd��Zeed�dd��Zeed�dd��Zed�dd�Zee
d�dd��Zee
d�dd��Zeeed�dd��Ze
eeed�dd �Zeed�d!d"�ZdS)#�ExtrasCandidatea�A candidate that has 'extras', indicating additional dependencies.

    Requirements can be for a project with dependencies, something like
    foo[extra].  The extras don't affect the project/version being installed
    directly, but indicate that we need additional dependencies. We model that
    by having an artificial ExtrasCandidate that wraps the "base" candidate.

    The ExtrasCandidate differs from the base in the following ways:

    1. It has a unique name, of the form foo[extra]. This causes the resolver
       to treat it as a separate node in the dependency graph.
    2. When we're getting the candidate's dependencies,
       a) We specify that we want the extra dependencies as well.
       b) We add a dependency on the base candidate.
          See below for why this is needed.
    3. We return None for the underlying InstallRequirement, as the base
       candidate will provide it, and we don't want to end up with duplicates.

    The dependency on the base candidate is needed so that the resolver can't
    decide that it should recommend foo[extra1] version 1.0 and foo[extra2]
    version 2.0. Having those candidates depend on foo=1.0 and foo=2.0
    respectively forces the resolver to recognise that this is a conflict.
    N)�base�extrasr!cCs||_||_dSrM)r�r�)rUr�r�r$r$r%rV�szExtrasCandidate.__init__rWcCs,t|j��dd�\}}d�|d�|j�|�S)NrXrz	{}[{}] {}�,)r6r��splitr]�joinr�)rUrL�restr$r$r%r[�szExtrasCandidate.__str__cCsdj|jj|j|jd�S)Nz.{class_name}(base={base!r}, extras={extras!r}))r\r�r�)r]r^r_r�r�rZr$r$r%r`�s
�zExtrasCandidate.__repr__cCst|j|jf�SrM)rar�r�rZr$r$r%rb�szExtrasCandidate.__hash__rccCs(t||j�r$|j|jko"|j|jkSdSre)r"r^r�r�rfr$r$r%rg�szExtrasCandidate.__eq__cCs|jjSrM)r�rirZr$r$r%ri�szExtrasCandidate.project_namecCst|jj|j�S)rh)rr�rir�rZr$r$r%rL�szExtrasCandidate.namecCs|jjSrM)r�rDrZr$r$r%rD�szExtrasCandidate.versioncCsd�|j��d�t|j���S)Nz{} [{}]z, )r]r�rmr��sortedr�rZr$r$r%rm�s�z ExtrasCandidate.format_for_errorcCs|jjSrM)r�r�rZr$r$r%r��szExtrasCandidate.is_installedcCs|jjSrM)r�r�rZr$r$r%r��szExtrasCandidate.is_editablecCs|jjSrM)r�rJrZr$r$r%rJ�szExtrasCandidate.source_linkrrccs�|jj}|�|j�V|sdS|j�|jj���}|j�|jj���}t|�D]}t	�
d|jj|j|�qN|jj�
|�D]$}|�t|�|jj|�}|rx|VqxdS)Nz%%s %s does not provide the extra '%s')r�rP�make_requirement_from_candidater��intersectionr@�iter_provided_extras�
differencer�r��warningrLrDrurvr6rQ)rUrsrK�valid_extras�invalid_extras�extrarz�requirementr$r$r%ru�s&��z!ExtrasCandidate.iter_dependenciescCsdSrMr$rZr$r$r%r{�sz'ExtrasCandidate.get_install_requirement) r_r|r}r~�
BaseCandidaterr6rVr[r`r�rbrr�rgr�r
rirLrrDrmr�r�rrrJrrrurr{r$r$r$r%r��s0�r�c@s�eZdZdZdZeeedfdd�dd�Ze	d�dd	�Z
eed�d
d��Z
ee	d�dd
��Zeed�dd��Ze	d�dd�Zeeeed�dd�Zeed�dd�ZdS)�RequiresPythonCandidateFN.)�py_version_infor!cCs>|durt|�}ntjdd�}td�dd�|D���|_dS)N��.css|]}t|�VqdSrM)r6)�.0�cr$r$r%�	<genexpr>�z3RequiresPythonCandidate.__init__.<locals>.<genexpr>)r�sys�version_inforr�rS)rUr�r�r$r$r%rV�s
z RequiresPythonCandidate.__init__rWcCsd|j��S�NzPython �rSrZr$r$r%r[szRequiresPythonCandidate.__str__cCstSrM��REQUIRES_PYTHON_IDENTIFIERrZr$r$r%ri	sz$RequiresPythonCandidate.project_namecCstSrMr�rZr$r$r%rL
szRequiresPythonCandidate.namecCs|jSrMr�rZr$r$r%rDszRequiresPythonCandidate.versioncCsd|j��Sr�)rDrZr$r$r%rmsz(RequiresPythonCandidate.format_for_errorrrcCsdSrtr$)rUrsr$r$r%rusz)RequiresPythonCandidate.iter_dependenciescCsdSrMr$rZr$r$r%r{sz/RequiresPythonCandidate.get_install_requirement)r_r|r}r�rJrrr�rVr6r[r�r
rirLrrDrmr�rrrurr{r$r$r$r%r��sr�)7�loggingr��typingrrrrrrrr	�pip._vendor.packaging.utilsr
rZpip._vendor.packaging.versionr�pip._internal.exceptionsr
r�pip._internal.metadatar�pip._internal.models.linkrr�pip._internal.models.wheelr�pip._internal.req.constructorsrrZpip._internal.req.req_installr�pip._internal.utils.miscrr�rrrrrKr�	getLoggerr_r�r�r�r&r=r?rHrIrrrr�r�r$r$r$r%�<module>sF(
�

���{2Cusite-packages/pip/_internal/resolution/resolvelib/__pycache__/reporter.cpython-39.pyc000064400000006353150564745010025121 0ustar00a

��Re�	�@spddlmZddlmZddlmZmZddlmZddl	m
Z
mZee�Z
Gdd�de�ZGd	d
�d
e�ZdS)�)�defaultdict)�	getLogger)�Any�DefaultDict)�BaseReporter�)�	Candidate�Requirementc@s*eZdZdd�dd�Zedd�dd�ZdS)�PipReporterN��returncCstt�|_dddd�|_dS)Nz�pip is looking at multiple versions of {package_name} to determine which version is compatible with other requirements. This could take a while.z�This is taking longer than usual. You might need to provide the dependency resolver with stricter constraints to reduce runtime. See https://pip.pypa.io/warnings/backtracking for guidance. If you want to abort this run, press Ctrl + C.)r��
)r�int�backtracks_by_package�_messages_at_backtrack��self�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/resolvelib/reporter.py�__init__
s

�zPipReporter.__init__��	candidatercCsR|j|jd7<|j|j}||jvr.dS|j|}t�d|j|jd��dS)NrzINFO: %s)�package_name)r�namer�logger�info�format)rr�count�messagerrr�backtracking#s

zPipReporter.backtracking)�__name__�
__module__�__qualname__rrr rrrrr
sr
c@s�eZdZdZdd�dd�Zedd�dd�Zeedd	�d
d�Zedd�d
d�Z	e
edd�dd�Zedd�dd�Z
edd�dd�ZdS)�PipDebuggingReporterz9A reporter that does an info log for every event it sees.NrcCst�d�dS)NzReporter.starting()�rrrrrr�starting1szPipDebuggingReporter.starting)�indexrcCst�d|�dS)NzReporter.starting_round(%r)r%)rr'rrr�starting_round4sz#PipDebuggingReporter.starting_round)r'�statercCst�d|�dS)Nz Reporter.ending_round(%r, state)r%)rr'r)rrr�ending_round7sz!PipDebuggingReporter.ending_round)r)rcCst�d|�dS)NzReporter.ending(%r)r%)rr)rrr�ending:szPipDebuggingReporter.ending)�requirement�parentrcCst�d||�dS)Nz#Reporter.adding_requirement(%r, %r)r%)rr,r-rrr�adding_requirement=sz'PipDebuggingReporter.adding_requirementrcCst�d|�dS)NzReporter.backtracking(%r)r%�rrrrrr @sz!PipDebuggingReporter.backtrackingcCst�d|�dS)NzReporter.pinning(%r)r%r/rrr�pinningCszPipDebuggingReporter.pinning)r!r"r#�__doc__r&rr(rr*r+r	rr.r r0rrrrr$.sr$N)�collectionsr�loggingr�typingrrZ pip._vendor.resolvelib.reportersr�baserr	r!rr
r$rrrr�<module>s"site-packages/pip/_internal/resolution/resolvelib/base.py000064400000012144150564745010017675 0ustar00from typing import FrozenSet, Iterable, Optional, Tuple, Union

from pip._vendor.packaging.specifiers import SpecifierSet
from pip._vendor.packaging.utils import NormalizedName, canonicalize_name
from pip._vendor.packaging.version import LegacyVersion, Version

from pip._internal.models.link import Link, links_equivalent
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils.hashes import Hashes

CandidateLookup = Tuple[Optional["Candidate"], Optional[InstallRequirement]]
CandidateVersion = Union[LegacyVersion, Version]


def format_name(project: str, extras: FrozenSet[str]) -> str:
    if not extras:
        return project
    canonical_extras = sorted(canonicalize_name(e) for e in extras)
    return "{}[{}]".format(project, ",".join(canonical_extras))


class Constraint:
    def __init__(
        self, specifier: SpecifierSet, hashes: Hashes, links: FrozenSet[Link]
    ) -> None:
        self.specifier = specifier
        self.hashes = hashes
        self.links = links

    @classmethod
    def empty(cls) -> "Constraint":
        return Constraint(SpecifierSet(), Hashes(), frozenset())

    @classmethod
    def from_ireq(cls, ireq: InstallRequirement) -> "Constraint":
        links = frozenset([ireq.link]) if ireq.link else frozenset()
        return Constraint(ireq.specifier, ireq.hashes(trust_internet=False), links)

    def __bool__(self) -> bool:
        return bool(self.specifier) or bool(self.hashes) or bool(self.links)

    def __and__(self, other: InstallRequirement) -> "Constraint":
        if not isinstance(other, InstallRequirement):
            return NotImplemented
        specifier = self.specifier & other.specifier
        hashes = self.hashes & other.hashes(trust_internet=False)
        links = self.links
        if other.link:
            links = links.union([other.link])
        return Constraint(specifier, hashes, links)

    def is_satisfied_by(self, candidate: "Candidate") -> bool:
        # Reject if there are any mismatched URL constraints on this package.
        if self.links and not all(_match_link(link, candidate) for link in self.links):
            return False
        # We can safely always allow prereleases here since PackageFinder
        # already implements the prerelease logic, and would have filtered out
        # prerelease candidates if the user does not expect them.
        return self.specifier.contains(candidate.version, prereleases=True)


class Requirement:
    @property
    def project_name(self) -> NormalizedName:
        """The "project name" of a requirement.

        This is different from ``name`` if this requirement contains extras,
        in which case ``name`` would contain the ``[...]`` part, while this
        refers to the name of the project.
        """
        raise NotImplementedError("Subclass should override")

    @property
    def name(self) -> str:
        """The name identifying this requirement in the resolver.

        This is different from ``project_name`` if this requirement contains
        extras, where ``project_name`` would not contain the ``[...]`` part.
        """
        raise NotImplementedError("Subclass should override")

    def is_satisfied_by(self, candidate: "Candidate") -> bool:
        return False

    def get_candidate_lookup(self) -> CandidateLookup:
        raise NotImplementedError("Subclass should override")

    def format_for_error(self) -> str:
        raise NotImplementedError("Subclass should override")


def _match_link(link: Link, candidate: "Candidate") -> bool:
    if candidate.source_link:
        return links_equivalent(link, candidate.source_link)
    return False


class Candidate:
    @property
    def project_name(self) -> NormalizedName:
        """The "project name" of the candidate.

        This is different from ``name`` if this candidate contains extras,
        in which case ``name`` would contain the ``[...]`` part, while this
        refers to the name of the project.
        """
        raise NotImplementedError("Override in subclass")

    @property
    def name(self) -> str:
        """The name identifying this candidate in the resolver.

        This is different from ``project_name`` if this candidate contains
        extras, where ``project_name`` would not contain the ``[...]`` part.
        """
        raise NotImplementedError("Override in subclass")

    @property
    def version(self) -> CandidateVersion:
        raise NotImplementedError("Override in subclass")

    @property
    def is_installed(self) -> bool:
        raise NotImplementedError("Override in subclass")

    @property
    def is_editable(self) -> bool:
        raise NotImplementedError("Override in subclass")

    @property
    def source_link(self) -> Optional[Link]:
        raise NotImplementedError("Override in subclass")

    def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]:
        raise NotImplementedError("Override in subclass")

    def get_install_requirement(self) -> Optional[InstallRequirement]:
        raise NotImplementedError("Override in subclass")

    def format_for_error(self) -> str:
        raise NotImplementedError("Subclass should override")
site-packages/pip/_internal/resolution/resolvelib/reporter.py000064400000004736150564745010020635 0ustar00from collections import defaultdict
from logging import getLogger
from typing import Any, DefaultDict

from pip._vendor.resolvelib.reporters import BaseReporter

from .base import Candidate, Requirement

logger = getLogger(__name__)


class PipReporter(BaseReporter):
    def __init__(self) -> None:
        self.backtracks_by_package: DefaultDict[str, int] = defaultdict(int)

        self._messages_at_backtrack = {
            1: (
                "pip is looking at multiple versions of {package_name} to "
                "determine which version is compatible with other "
                "requirements. This could take a while."
            ),
            8: (
                "pip is looking at multiple versions of {package_name} to "
                "determine which version is compatible with other "
                "requirements. This could take a while."
            ),
            13: (
                "This is taking longer than usual. You might need to provide "
                "the dependency resolver with stricter constraints to reduce "
                "runtime. See https://pip.pypa.io/warnings/backtracking for "
                "guidance. If you want to abort this run, press Ctrl + C."
            ),
        }

    def backtracking(self, candidate: Candidate) -> None:
        self.backtracks_by_package[candidate.name] += 1

        count = self.backtracks_by_package[candidate.name]
        if count not in self._messages_at_backtrack:
            return

        message = self._messages_at_backtrack[count]
        logger.info("INFO: %s", message.format(package_name=candidate.name))


class PipDebuggingReporter(BaseReporter):
    """A reporter that does an info log for every event it sees."""

    def starting(self) -> None:
        logger.info("Reporter.starting()")

    def starting_round(self, index: int) -> None:
        logger.info("Reporter.starting_round(%r)", index)

    def ending_round(self, index: int, state: Any) -> None:
        logger.info("Reporter.ending_round(%r, state)", index)

    def ending(self, state: Any) -> None:
        logger.info("Reporter.ending(%r)", state)

    def adding_requirement(self, requirement: Requirement, parent: Candidate) -> None:
        logger.info("Reporter.adding_requirement(%r, %r)", requirement, parent)

    def backtracking(self, candidate: Candidate) -> None:
        logger.info("Reporter.backtracking(%r)", candidate)

    def pinning(self, candidate: Candidate) -> None:
        logger.info("Reporter.pinning(%r)", candidate)
site-packages/pip/_internal/resolution/resolvelib/__init__.py000064400000000000150564745010020506 0ustar00site-packages/pip/_internal/resolution/resolvelib/provider.py000064400000021765150564745010020626 0ustar00import collections
import math
from typing import TYPE_CHECKING, Dict, Iterable, Iterator, Mapping, Sequence, Union

from pip._vendor.resolvelib.providers import AbstractProvider

from .base import Candidate, Constraint, Requirement
from .candidates import REQUIRES_PYTHON_IDENTIFIER
from .factory import Factory

if TYPE_CHECKING:
    from pip._vendor.resolvelib.providers import Preference
    from pip._vendor.resolvelib.resolvers import RequirementInformation

    PreferenceInformation = RequirementInformation[Requirement, Candidate]

    _ProviderBase = AbstractProvider[Requirement, Candidate, str]
else:
    _ProviderBase = AbstractProvider

# Notes on the relationship between the provider, the factory, and the
# candidate and requirement classes.
#
# The provider is a direct implementation of the resolvelib class. Its role
# is to deliver the API that resolvelib expects.
#
# Rather than work with completely abstract "requirement" and "candidate"
# concepts as resolvelib does, pip has concrete classes implementing these two
# ideas. The API of Requirement and Candidate objects are defined in the base
# classes, but essentially map fairly directly to the equivalent provider
# methods. In particular, `find_matches` and `is_satisfied_by` are
# requirement methods, and `get_dependencies` is a candidate method.
#
# The factory is the interface to pip's internal mechanisms. It is stateless,
# and is created by the resolver and held as a property of the provider. It is
# responsible for creating Requirement and Candidate objects, and provides
# services to those objects (access to pip's finder and preparer).


class PipProvider(_ProviderBase):
    """Pip's provider implementation for resolvelib.

    :params constraints: A mapping of constraints specified by the user. Keys
        are canonicalized project names.
    :params ignore_dependencies: Whether the user specified ``--no-deps``.
    :params upgrade_strategy: The user-specified upgrade strategy.
    :params user_requested: A set of canonicalized package names that the user
        supplied for pip to install/upgrade.
    """

    def __init__(
        self,
        factory: Factory,
        constraints: Dict[str, Constraint],
        ignore_dependencies: bool,
        upgrade_strategy: str,
        user_requested: Dict[str, int],
    ) -> None:
        self._factory = factory
        self._constraints = constraints
        self._ignore_dependencies = ignore_dependencies
        self._upgrade_strategy = upgrade_strategy
        self._user_requested = user_requested
        self._known_depths: Dict[str, float] = collections.defaultdict(lambda: math.inf)

    def identify(self, requirement_or_candidate: Union[Requirement, Candidate]) -> str:
        return requirement_or_candidate.name

    def get_preference(  # type: ignore
        self,
        identifier: str,
        resolutions: Mapping[str, Candidate],
        candidates: Mapping[str, Iterator[Candidate]],
        information: Mapping[str, Iterable["PreferenceInformation"]],
        backtrack_causes: Sequence["PreferenceInformation"],
    ) -> "Preference":
        """Produce a sort key for given requirement based on preference.

        The lower the return value is, the more preferred this group of
        arguments is.

        Currently pip considers the followings in order:

        * Prefer if any of the known requirements is "direct", e.g. points to an
          explicit URL.
        * If equal, prefer if any requirement is "pinned", i.e. contains
          operator ``===`` or ``==``.
        * If equal, calculate an approximate "depth" and resolve requirements
          closer to the user-specified requirements first.
        * Order user-specified requirements by the order they are specified.
        * If equal, prefers "non-free" requirements, i.e. contains at least one
          operator, such as ``>=`` or ``<``.
        * If equal, order alphabetically for consistency (helps debuggability).
        """
        lookups = (r.get_candidate_lookup() for r, _ in information[identifier])
        candidate, ireqs = zip(*lookups)
        operators = [
            specifier.operator
            for specifier_set in (ireq.specifier for ireq in ireqs if ireq)
            for specifier in specifier_set
        ]

        direct = candidate is not None
        pinned = any(op[:2] == "==" for op in operators)
        unfree = bool(operators)

        try:
            requested_order: Union[int, float] = self._user_requested[identifier]
        except KeyError:
            requested_order = math.inf
            parent_depths = (
                self._known_depths[parent.name] if parent is not None else 0.0
                for _, parent in information[identifier]
            )
            inferred_depth = min(d for d in parent_depths) + 1.0
        else:
            inferred_depth = 1.0
        self._known_depths[identifier] = inferred_depth

        requested_order = self._user_requested.get(identifier, math.inf)

        # Requires-Python has only one candidate and the check is basically
        # free, so we always do it first to avoid needless work if it fails.
        requires_python = identifier == REQUIRES_PYTHON_IDENTIFIER

        # HACK: Setuptools have a very long and solid backward compatibility
        # track record, and extremely few projects would request a narrow,
        # non-recent version range of it since that would break a lot things.
        # (Most projects specify it only to request for an installer feature,
        # which does not work, but that's another topic.) Intentionally
        # delaying Setuptools helps reduce branches the resolver has to check.
        # This serves as a temporary fix for issues like "apache-airlfow[all]"
        # while we work on "proper" branch pruning techniques.
        delay_this = identifier == "setuptools"

        # Prefer the causes of backtracking on the assumption that the problem
        # resolving the dependency tree is related to the failures that caused
        # the backtracking
        backtrack_cause = self.is_backtrack_cause(identifier, backtrack_causes)

        return (
            not requires_python,
            delay_this,
            not direct,
            not pinned,
            not backtrack_cause,
            inferred_depth,
            requested_order,
            not unfree,
            identifier,
        )

    def _get_constraint(self, identifier: str) -> Constraint:
        if identifier in self._constraints:
            return self._constraints[identifier]

        # HACK: Theoratically we should check whether this identifier is a valid
        # "NAME[EXTRAS]" format, and parse out the name part with packaging or
        # some regular expression. But since pip's resolver only spits out
        # three kinds of identifiers: normalized PEP 503 names, normalized names
        # plus extras, and Requires-Python, we can cheat a bit here.
        name, open_bracket, _ = identifier.partition("[")
        if open_bracket and name in self._constraints:
            return self._constraints[name]

        return Constraint.empty()

    def find_matches(
        self,
        identifier: str,
        requirements: Mapping[str, Iterator[Requirement]],
        incompatibilities: Mapping[str, Iterator[Candidate]],
    ) -> Iterable[Candidate]:
        def _eligible_for_upgrade(name: str) -> bool:
            """Are upgrades allowed for this project?

            This checks the upgrade strategy, and whether the project was one
            that the user specified in the command line, in order to decide
            whether we should upgrade if there's a newer version available.

            (Note that we don't need access to the `--upgrade` flag, because
            an upgrade strategy of "to-satisfy-only" means that `--upgrade`
            was not specified).
            """
            if self._upgrade_strategy == "eager":
                return True
            elif self._upgrade_strategy == "only-if-needed":
                return name in self._user_requested
            return False

        return self._factory.find_candidates(
            identifier=identifier,
            requirements=requirements,
            constraint=self._get_constraint(identifier),
            prefers_installed=(not _eligible_for_upgrade(identifier)),
            incompatibilities=incompatibilities,
        )

    def is_satisfied_by(self, requirement: Requirement, candidate: Candidate) -> bool:
        return requirement.is_satisfied_by(candidate)

    def get_dependencies(self, candidate: Candidate) -> Sequence[Requirement]:
        with_requires = not self._ignore_dependencies
        return [r for r in candidate.iter_dependencies(with_requires) if r is not None]

    @staticmethod
    def is_backtrack_cause(
        identifier: str, backtrack_causes: Sequence["PreferenceInformation"]
    ) -> bool:
        for backtrack_cause in backtrack_causes:
            if identifier == backtrack_cause.requirement.name:
                return True
            if backtrack_cause.parent and identifier == backtrack_cause.parent.name:
                return True
        return False
site-packages/pip/_internal/resolution/resolvelib/candidates.py000064400000043442150564745010021067 0ustar00import logging
import sys
from typing import TYPE_CHECKING, Any, FrozenSet, Iterable, Optional, Tuple, Union, cast

from pip._vendor.packaging.utils import NormalizedName, canonicalize_name
from pip._vendor.packaging.version import Version

from pip._internal.exceptions import HashError, MetadataInconsistent
from pip._internal.metadata import BaseDistribution
from pip._internal.models.link import Link, links_equivalent
from pip._internal.models.wheel import Wheel
from pip._internal.req.constructors import (
    install_req_from_editable,
    install_req_from_line,
)
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils.misc import normalize_version_info

from .base import Candidate, CandidateVersion, Requirement, format_name

if TYPE_CHECKING:
    from .factory import Factory

logger = logging.getLogger(__name__)

BaseCandidate = Union[
    "AlreadyInstalledCandidate",
    "EditableCandidate",
    "LinkCandidate",
]

# Avoid conflicting with the PyPI package "Python".
REQUIRES_PYTHON_IDENTIFIER = cast(NormalizedName, "<Python from Requires-Python>")


def as_base_candidate(candidate: Candidate) -> Optional[BaseCandidate]:
    """The runtime version of BaseCandidate."""
    base_candidate_classes = (
        AlreadyInstalledCandidate,
        EditableCandidate,
        LinkCandidate,
    )
    if isinstance(candidate, base_candidate_classes):
        return candidate
    return None


def make_install_req_from_link(
    link: Link, template: InstallRequirement
) -> InstallRequirement:
    assert not template.editable, "template is editable"
    if template.req:
        line = str(template.req)
    else:
        line = link.url
    ireq = install_req_from_line(
        line,
        user_supplied=template.user_supplied,
        comes_from=template.comes_from,
        use_pep517=template.use_pep517,
        isolated=template.isolated,
        constraint=template.constraint,
        options=dict(
            install_options=template.install_options,
            global_options=template.global_options,
            hashes=template.hash_options,
        ),
    )
    ireq.original_link = template.original_link
    ireq.link = link
    return ireq


def make_install_req_from_editable(
    link: Link, template: InstallRequirement
) -> InstallRequirement:
    assert template.editable, "template not editable"
    return install_req_from_editable(
        link.url,
        user_supplied=template.user_supplied,
        comes_from=template.comes_from,
        use_pep517=template.use_pep517,
        isolated=template.isolated,
        constraint=template.constraint,
        permit_editable_wheels=template.permit_editable_wheels,
        options=dict(
            install_options=template.install_options,
            global_options=template.global_options,
            hashes=template.hash_options,
        ),
    )


def _make_install_req_from_dist(
    dist: BaseDistribution, template: InstallRequirement
) -> InstallRequirement:
    from pip._internal.metadata.pkg_resources import Distribution as _Dist

    if template.req:
        line = str(template.req)
    elif template.link:
        line = f"{dist.canonical_name} @ {template.link.url}"
    else:
        line = f"{dist.canonical_name}=={dist.version}"
    ireq = install_req_from_line(
        line,
        user_supplied=template.user_supplied,
        comes_from=template.comes_from,
        use_pep517=template.use_pep517,
        isolated=template.isolated,
        constraint=template.constraint,
        options=dict(
            install_options=template.install_options,
            global_options=template.global_options,
            hashes=template.hash_options,
        ),
    )
    ireq.satisfied_by = cast(_Dist, dist)._dist
    return ireq


class _InstallRequirementBackedCandidate(Candidate):
    """A candidate backed by an ``InstallRequirement``.

    This represents a package request with the target not being already
    in the environment, and needs to be fetched and installed. The backing
    ``InstallRequirement`` is responsible for most of the leg work; this
    class exposes appropriate information to the resolver.

    :param link: The link passed to the ``InstallRequirement``. The backing
        ``InstallRequirement`` will use this link to fetch the distribution.
    :param source_link: The link this candidate "originates" from. This is
        different from ``link`` when the link is found in the wheel cache.
        ``link`` would point to the wheel cache, while this points to the
        found remote link (e.g. from pypi.org).
    """

    dist: BaseDistribution
    is_installed = False

    def __init__(
        self,
        link: Link,
        source_link: Link,
        ireq: InstallRequirement,
        factory: "Factory",
        name: Optional[NormalizedName] = None,
        version: Optional[CandidateVersion] = None,
    ) -> None:
        self._link = link
        self._source_link = source_link
        self._factory = factory
        self._ireq = ireq
        self._name = name
        self._version = version
        self.dist = self._prepare()

    def __str__(self) -> str:
        return f"{self.name} {self.version}"

    def __repr__(self) -> str:
        return "{class_name}({link!r})".format(
            class_name=self.__class__.__name__,
            link=str(self._link),
        )

    def __hash__(self) -> int:
        return hash((self.__class__, self._link))

    def __eq__(self, other: Any) -> bool:
        if isinstance(other, self.__class__):
            return links_equivalent(self._link, other._link)
        return False

    @property
    def source_link(self) -> Optional[Link]:
        return self._source_link

    @property
    def project_name(self) -> NormalizedName:
        """The normalised name of the project the candidate refers to"""
        if self._name is None:
            self._name = self.dist.canonical_name
        return self._name

    @property
    def name(self) -> str:
        return self.project_name

    @property
    def version(self) -> CandidateVersion:
        if self._version is None:
            self._version = self.dist.version
        return self._version

    def format_for_error(self) -> str:
        return "{} {} (from {})".format(
            self.name,
            self.version,
            self._link.file_path if self._link.is_file else self._link,
        )

    def _prepare_distribution(self) -> BaseDistribution:
        raise NotImplementedError("Override in subclass")

    def _check_metadata_consistency(self, dist: BaseDistribution) -> None:
        """Check for consistency of project name and version of dist."""
        if self._name is not None and self._name != dist.canonical_name:
            raise MetadataInconsistent(
                self._ireq,
                "name",
                self._name,
                dist.canonical_name,
            )
        if self._version is not None and self._version != dist.version:
            raise MetadataInconsistent(
                self._ireq,
                "version",
                str(self._version),
                str(dist.version),
            )

    def _prepare(self) -> BaseDistribution:
        try:
            dist = self._prepare_distribution()
        except HashError as e:
            # Provide HashError the underlying ireq that caused it. This
            # provides context for the resulting error message to show the
            # offending line to the user.
            e.req = self._ireq
            raise
        self._check_metadata_consistency(dist)
        return dist

    def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]:
        requires = self.dist.iter_dependencies() if with_requires else ()
        for r in requires:
            yield self._factory.make_requirement_from_spec(str(r), self._ireq)
        yield self._factory.make_requires_python_requirement(self.dist.requires_python)

    def get_install_requirement(self) -> Optional[InstallRequirement]:
        return self._ireq


class LinkCandidate(_InstallRequirementBackedCandidate):
    is_editable = False

    def __init__(
        self,
        link: Link,
        template: InstallRequirement,
        factory: "Factory",
        name: Optional[NormalizedName] = None,
        version: Optional[CandidateVersion] = None,
    ) -> None:
        source_link = link
        cache_entry = factory.get_wheel_cache_entry(link, name)
        if cache_entry is not None:
            logger.debug("Using cached wheel link: %s", cache_entry.link)
            link = cache_entry.link
        ireq = make_install_req_from_link(link, template)
        assert ireq.link == link
        if ireq.link.is_wheel and not ireq.link.is_file:
            wheel = Wheel(ireq.link.filename)
            wheel_name = canonicalize_name(wheel.name)
            assert name == wheel_name, f"{name!r} != {wheel_name!r} for wheel"
            # Version may not be present for PEP 508 direct URLs
            if version is not None:
                wheel_version = Version(wheel.version)
                assert version == wheel_version, "{!r} != {!r} for wheel {}".format(
                    version, wheel_version, name
                )

        if (
            cache_entry is not None
            and cache_entry.persistent
            and template.link is template.original_link
        ):
            ireq.original_link_is_in_wheel_cache = True

        super().__init__(
            link=link,
            source_link=source_link,
            ireq=ireq,
            factory=factory,
            name=name,
            version=version,
        )

    def _prepare_distribution(self) -> BaseDistribution:
        preparer = self._factory.preparer
        return preparer.prepare_linked_requirement(self._ireq, parallel_builds=True)


class EditableCandidate(_InstallRequirementBackedCandidate):
    is_editable = True

    def __init__(
        self,
        link: Link,
        template: InstallRequirement,
        factory: "Factory",
        name: Optional[NormalizedName] = None,
        version: Optional[CandidateVersion] = None,
    ) -> None:
        super().__init__(
            link=link,
            source_link=link,
            ireq=make_install_req_from_editable(link, template),
            factory=factory,
            name=name,
            version=version,
        )

    def _prepare_distribution(self) -> BaseDistribution:
        return self._factory.preparer.prepare_editable_requirement(self._ireq)


class AlreadyInstalledCandidate(Candidate):
    is_installed = True
    source_link = None

    def __init__(
        self,
        dist: BaseDistribution,
        template: InstallRequirement,
        factory: "Factory",
    ) -> None:
        self.dist = dist
        self._ireq = _make_install_req_from_dist(dist, template)
        self._factory = factory

        # This is just logging some messages, so we can do it eagerly.
        # The returned dist would be exactly the same as self.dist because we
        # set satisfied_by in _make_install_req_from_dist.
        # TODO: Supply reason based on force_reinstall and upgrade_strategy.
        skip_reason = "already satisfied"
        factory.preparer.prepare_installed_requirement(self._ireq, skip_reason)

    def __str__(self) -> str:
        return str(self.dist)

    def __repr__(self) -> str:
        return "{class_name}({distribution!r})".format(
            class_name=self.__class__.__name__,
            distribution=self.dist,
        )

    def __hash__(self) -> int:
        return hash((self.__class__, self.name, self.version))

    def __eq__(self, other: Any) -> bool:
        if isinstance(other, self.__class__):
            return self.name == other.name and self.version == other.version
        return False

    @property
    def project_name(self) -> NormalizedName:
        return self.dist.canonical_name

    @property
    def name(self) -> str:
        return self.project_name

    @property
    def version(self) -> CandidateVersion:
        return self.dist.version

    @property
    def is_editable(self) -> bool:
        return self.dist.editable

    def format_for_error(self) -> str:
        return f"{self.name} {self.version} (Installed)"

    def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]:
        if not with_requires:
            return
        for r in self.dist.iter_dependencies():
            yield self._factory.make_requirement_from_spec(str(r), self._ireq)

    def get_install_requirement(self) -> Optional[InstallRequirement]:
        return None


class ExtrasCandidate(Candidate):
    """A candidate that has 'extras', indicating additional dependencies.

    Requirements can be for a project with dependencies, something like
    foo[extra].  The extras don't affect the project/version being installed
    directly, but indicate that we need additional dependencies. We model that
    by having an artificial ExtrasCandidate that wraps the "base" candidate.

    The ExtrasCandidate differs from the base in the following ways:

    1. It has a unique name, of the form foo[extra]. This causes the resolver
       to treat it as a separate node in the dependency graph.
    2. When we're getting the candidate's dependencies,
       a) We specify that we want the extra dependencies as well.
       b) We add a dependency on the base candidate.
          See below for why this is needed.
    3. We return None for the underlying InstallRequirement, as the base
       candidate will provide it, and we don't want to end up with duplicates.

    The dependency on the base candidate is needed so that the resolver can't
    decide that it should recommend foo[extra1] version 1.0 and foo[extra2]
    version 2.0. Having those candidates depend on foo=1.0 and foo=2.0
    respectively forces the resolver to recognise that this is a conflict.
    """

    def __init__(
        self,
        base: BaseCandidate,
        extras: FrozenSet[str],
    ) -> None:
        self.base = base
        self.extras = extras

    def __str__(self) -> str:
        name, rest = str(self.base).split(" ", 1)
        return "{}[{}] {}".format(name, ",".join(self.extras), rest)

    def __repr__(self) -> str:
        return "{class_name}(base={base!r}, extras={extras!r})".format(
            class_name=self.__class__.__name__,
            base=self.base,
            extras=self.extras,
        )

    def __hash__(self) -> int:
        return hash((self.base, self.extras))

    def __eq__(self, other: Any) -> bool:
        if isinstance(other, self.__class__):
            return self.base == other.base and self.extras == other.extras
        return False

    @property
    def project_name(self) -> NormalizedName:
        return self.base.project_name

    @property
    def name(self) -> str:
        """The normalised name of the project the candidate refers to"""
        return format_name(self.base.project_name, self.extras)

    @property
    def version(self) -> CandidateVersion:
        return self.base.version

    def format_for_error(self) -> str:
        return "{} [{}]".format(
            self.base.format_for_error(), ", ".join(sorted(self.extras))
        )

    @property
    def is_installed(self) -> bool:
        return self.base.is_installed

    @property
    def is_editable(self) -> bool:
        return self.base.is_editable

    @property
    def source_link(self) -> Optional[Link]:
        return self.base.source_link

    def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]:
        factory = self.base._factory

        # Add a dependency on the exact base
        # (See note 2b in the class docstring)
        yield factory.make_requirement_from_candidate(self.base)
        if not with_requires:
            return

        # The user may have specified extras that the candidate doesn't
        # support. We ignore any unsupported extras here.
        valid_extras = self.extras.intersection(self.base.dist.iter_provided_extras())
        invalid_extras = self.extras.difference(self.base.dist.iter_provided_extras())
        for extra in sorted(invalid_extras):
            logger.warning(
                "%s %s does not provide the extra '%s'",
                self.base.name,
                self.version,
                extra,
            )

        for r in self.base.dist.iter_dependencies(valid_extras):
            requirement = factory.make_requirement_from_spec(
                str(r), self.base._ireq, valid_extras
            )
            if requirement:
                yield requirement

    def get_install_requirement(self) -> Optional[InstallRequirement]:
        # We don't return anything here, because we always
        # depend on the base candidate, and we'll get the
        # install requirement from that.
        return None


class RequiresPythonCandidate(Candidate):
    is_installed = False
    source_link = None

    def __init__(self, py_version_info: Optional[Tuple[int, ...]]) -> None:
        if py_version_info is not None:
            version_info = normalize_version_info(py_version_info)
        else:
            version_info = sys.version_info[:3]
        self._version = Version(".".join(str(c) for c in version_info))

    # We don't need to implement __eq__() and __ne__() since there is always
    # only one RequiresPythonCandidate in a resolution, i.e. the host Python.
    # The built-in object.__eq__() and object.__ne__() do exactly what we want.

    def __str__(self) -> str:
        return f"Python {self._version}"

    @property
    def project_name(self) -> NormalizedName:
        return REQUIRES_PYTHON_IDENTIFIER

    @property
    def name(self) -> str:
        return REQUIRES_PYTHON_IDENTIFIER

    @property
    def version(self) -> CandidateVersion:
        return self._version

    def format_for_error(self) -> str:
        return f"Python {self.version}"

    def iter_dependencies(self, with_requires: bool) -> Iterable[Optional[Requirement]]:
        return ()

    def get_install_requirement(self) -> Optional[InstallRequirement]:
        return None
site-packages/pip/_internal/resolution/resolvelib/resolver.py000064400000022554150564745010020632 0ustar00import functools
import logging
import os
from typing import TYPE_CHECKING, Dict, List, Optional, Set, Tuple, cast

from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.resolvelib import BaseReporter, ResolutionImpossible
from pip._vendor.resolvelib import Resolver as RLResolver
from pip._vendor.resolvelib.structs import DirectedGraph

from pip._internal.cache import WheelCache
from pip._internal.index.package_finder import PackageFinder
from pip._internal.operations.prepare import RequirementPreparer
from pip._internal.req.req_install import InstallRequirement
from pip._internal.req.req_set import RequirementSet
from pip._internal.resolution.base import BaseResolver, InstallRequirementProvider
from pip._internal.resolution.resolvelib.provider import PipProvider
from pip._internal.resolution.resolvelib.reporter import (
    PipDebuggingReporter,
    PipReporter,
)

from .base import Candidate, Requirement
from .factory import Factory

if TYPE_CHECKING:
    from pip._vendor.resolvelib.resolvers import Result as RLResult

    Result = RLResult[Requirement, Candidate, str]


logger = logging.getLogger(__name__)


class Resolver(BaseResolver):
    _allowed_strategies = {"eager", "only-if-needed", "to-satisfy-only"}

    def __init__(
        self,
        preparer: RequirementPreparer,
        finder: PackageFinder,
        wheel_cache: Optional[WheelCache],
        make_install_req: InstallRequirementProvider,
        use_user_site: bool,
        ignore_dependencies: bool,
        ignore_installed: bool,
        ignore_requires_python: bool,
        force_reinstall: bool,
        upgrade_strategy: str,
        py_version_info: Optional[Tuple[int, ...]] = None,
    ):
        super().__init__()
        assert upgrade_strategy in self._allowed_strategies

        self.factory = Factory(
            finder=finder,
            preparer=preparer,
            make_install_req=make_install_req,
            wheel_cache=wheel_cache,
            use_user_site=use_user_site,
            force_reinstall=force_reinstall,
            ignore_installed=ignore_installed,
            ignore_requires_python=ignore_requires_python,
            py_version_info=py_version_info,
        )
        self.ignore_dependencies = ignore_dependencies
        self.upgrade_strategy = upgrade_strategy
        self._result: Optional[Result] = None

    def resolve(
        self, root_reqs: List[InstallRequirement], check_supported_wheels: bool
    ) -> RequirementSet:
        collected = self.factory.collect_root_requirements(root_reqs)
        provider = PipProvider(
            factory=self.factory,
            constraints=collected.constraints,
            ignore_dependencies=self.ignore_dependencies,
            upgrade_strategy=self.upgrade_strategy,
            user_requested=collected.user_requested,
        )
        if "PIP_RESOLVER_DEBUG" in os.environ:
            reporter: BaseReporter = PipDebuggingReporter()
        else:
            reporter = PipReporter()
        resolver: RLResolver[Requirement, Candidate, str] = RLResolver(
            provider,
            reporter,
        )

        try:
            try_to_avoid_resolution_too_deep = 2000000
            result = self._result = resolver.resolve(
                collected.requirements, max_rounds=try_to_avoid_resolution_too_deep
            )

        except ResolutionImpossible as e:
            error = self.factory.get_installation_error(
                cast("ResolutionImpossible[Requirement, Candidate]", e),
                collected.constraints,
            )
            raise error from e

        req_set = RequirementSet(check_supported_wheels=check_supported_wheels)
        for candidate in result.mapping.values():
            ireq = candidate.get_install_requirement()
            if ireq is None:
                continue

            # Check if there is already an installation under the same name,
            # and set a flag for later stages to uninstall it, if needed.
            installed_dist = self.factory.get_dist_to_uninstall(candidate)
            if installed_dist is None:
                # There is no existing installation -- nothing to uninstall.
                ireq.should_reinstall = False
            elif self.factory.force_reinstall:
                # The --force-reinstall flag is set -- reinstall.
                ireq.should_reinstall = True
            elif installed_dist.version != candidate.version:
                # The installation is different in version -- reinstall.
                ireq.should_reinstall = True
            elif candidate.is_editable or installed_dist.editable:
                # The incoming distribution is editable, or different in
                # editable-ness to installation -- reinstall.
                ireq.should_reinstall = True
            elif candidate.source_link and candidate.source_link.is_file:
                # The incoming distribution is under file://
                if candidate.source_link.is_wheel:
                    # is a local wheel -- do nothing.
                    logger.info(
                        "%s is already installed with the same version as the "
                        "provided wheel. Use --force-reinstall to force an "
                        "installation of the wheel.",
                        ireq.name,
                    )
                    continue

                # is a local sdist or path -- reinstall
                ireq.should_reinstall = True
            else:
                continue

            link = candidate.source_link
            if link and link.is_yanked:
                # The reason can contain non-ASCII characters, Unicode
                # is required for Python 2.
                msg = (
                    "The candidate selected for download or install is a "
                    "yanked version: {name!r} candidate (version {version} "
                    "at {link})\nReason for being yanked: {reason}"
                ).format(
                    name=candidate.name,
                    version=candidate.version,
                    link=link,
                    reason=link.yanked_reason or "<none given>",
                )
                logger.warning(msg)

            req_set.add_named_requirement(ireq)

        reqs = req_set.all_requirements
        self.factory.preparer.prepare_linked_requirements_more(reqs)
        return req_set

    def get_installation_order(
        self, req_set: RequirementSet
    ) -> List[InstallRequirement]:
        """Get order for installation of requirements in RequirementSet.

        The returned list contains a requirement before another that depends on
        it. This helps ensure that the environment is kept consistent as they
        get installed one-by-one.

        The current implementation creates a topological ordering of the
        dependency graph, while breaking any cycles in the graph at arbitrary
        points. We make no guarantees about where the cycle would be broken,
        other than they would be broken.
        """
        assert self._result is not None, "must call resolve() first"

        graph = self._result.graph
        weights = get_topological_weights(
            graph,
            expected_node_count=len(self._result.mapping) + 1,
        )

        sorted_items = sorted(
            req_set.requirements.items(),
            key=functools.partial(_req_set_item_sorter, weights=weights),
            reverse=True,
        )
        return [ireq for _, ireq in sorted_items]


def get_topological_weights(
    graph: "DirectedGraph[Optional[str]]", expected_node_count: int
) -> Dict[Optional[str], int]:
    """Assign weights to each node based on how "deep" they are.

    This implementation may change at any point in the future without prior
    notice.

    We take the length for the longest path to any node from root, ignoring any
    paths that contain a single node twice (i.e. cycles). This is done through
    a depth-first search through the graph, while keeping track of the path to
    the node.

    Cycles in the graph result would result in node being revisited while also
    being it's own path. In this case, take no action. This helps ensure we
    don't get stuck in a cycle.

    When assigning weight, the longer path (i.e. larger length) is preferred.
    """
    path: Set[Optional[str]] = set()
    weights: Dict[Optional[str], int] = {}

    def visit(node: Optional[str]) -> None:
        if node in path:
            # We hit a cycle, so we'll break it here.
            return

        # Time to visit the children!
        path.add(node)
        for child in graph.iter_children(node):
            visit(child)
        path.remove(node)

        last_known_parent_count = weights.get(node, 0)
        weights[node] = max(last_known_parent_count, len(path))

    # `None` is guaranteed to be the root node by resolvelib.
    visit(None)

    # Sanity checks
    assert weights[None] == 0
    assert len(weights) == expected_node_count

    return weights


def _req_set_item_sorter(
    item: Tuple[str, InstallRequirement],
    weights: Dict[Optional[str], int],
) -> Tuple[int, str]:
    """Key function used to sort install requirements for installation.

    Based on the "weight" mapping calculated in ``get_installation_order()``.
    The canonical package name is returned as the second member as a tie-
    breaker to ensure the result is predictable, which is useful in tests.
    """
    name = canonicalize_name(item[0])
    return weights[name], name
site-packages/pip/_internal/resolution/resolvelib/factory.py000064400000065377150564745010020452 0ustar00import contextlib
import functools
import logging
import sys
import sysconfig
from typing import (
    TYPE_CHECKING,
    Dict,
    FrozenSet,
    Iterable,
    Iterator,
    List,
    Mapping,
    NamedTuple,
    Optional,
    Sequence,
    Set,
    Tuple,
    TypeVar,
    cast,
)

from pip._vendor.packaging.requirements import InvalidRequirement
from pip._vendor.packaging.specifiers import SpecifierSet
from pip._vendor.packaging.utils import NormalizedName, canonicalize_name
from pip._vendor.resolvelib import ResolutionImpossible

from pip._internal.cache import CacheEntry, WheelCache
from pip._internal.exceptions import (
    DistributionNotFound,
    InstallationError,
    InstallationSubprocessError,
    MetadataInconsistent,
    UnsupportedPythonVersion,
    UnsupportedWheel,
)
from pip._internal.index.package_finder import PackageFinder
from pip._internal.locations import get_scheme
from pip._internal.metadata import BaseDistribution, get_default_environment
from pip._internal.models.link import Link
from pip._internal.models.wheel import Wheel
from pip._internal.operations.prepare import RequirementPreparer
from pip._internal.req.constructors import install_req_from_link_and_ireq
from pip._internal.req.req_install import (
    InstallRequirement,
    check_invalid_constraint_type,
)
from pip._internal.resolution.base import InstallRequirementProvider
from pip._internal.utils.compatibility_tags import get_supported
from pip._internal.utils.hashes import Hashes
from pip._internal.utils.misc import dist_location
from pip._internal.utils.packaging import get_requirement
from pip._internal.utils.virtualenv import running_under_virtualenv

from .base import Candidate, CandidateVersion, Constraint, Requirement
from .candidates import (
    AlreadyInstalledCandidate,
    BaseCandidate,
    EditableCandidate,
    ExtrasCandidate,
    LinkCandidate,
    RequiresPythonCandidate,
    as_base_candidate,
)
from .found_candidates import FoundCandidates, IndexCandidateInfo
from .requirements import (
    ExplicitRequirement,
    RequiresPythonRequirement,
    SpecifierRequirement,
    UnsatisfiableRequirement,
)

if TYPE_CHECKING:
    from typing import Protocol

    class ConflictCause(Protocol):
        requirement: RequiresPythonRequirement
        parent: Candidate


logger = logging.getLogger(__name__)

C = TypeVar("C")
Cache = Dict[Link, C]


class CollectedRootRequirements(NamedTuple):
    requirements: List[Requirement]
    constraints: Dict[str, Constraint]
    user_requested: Dict[str, int]


class Factory:
    def __init__(
        self,
        finder: PackageFinder,
        preparer: RequirementPreparer,
        make_install_req: InstallRequirementProvider,
        wheel_cache: Optional[WheelCache],
        use_user_site: bool,
        force_reinstall: bool,
        ignore_installed: bool,
        ignore_requires_python: bool,
        py_version_info: Optional[Tuple[int, ...]] = None,
    ) -> None:
        self._finder = finder
        self.preparer = preparer
        self._wheel_cache = wheel_cache
        self._python_candidate = RequiresPythonCandidate(py_version_info)
        self._make_install_req_from_spec = make_install_req
        self._use_user_site = use_user_site
        self._force_reinstall = force_reinstall
        self._ignore_requires_python = ignore_requires_python

        self._build_failures: Cache[InstallationError] = {}
        self._link_candidate_cache: Cache[LinkCandidate] = {}
        self._editable_candidate_cache: Cache[EditableCandidate] = {}
        self._installed_candidate_cache: Dict[str, AlreadyInstalledCandidate] = {}
        self._extras_candidate_cache: Dict[
            Tuple[int, FrozenSet[str]], ExtrasCandidate
        ] = {}

        if not ignore_installed:
            env = get_default_environment()
            self._installed_dists = {
                dist.canonical_name: dist
                for dist in env.iter_installed_distributions(local_only=False)
            }
        else:
            self._installed_dists = {}

    @property
    def force_reinstall(self) -> bool:
        return self._force_reinstall

    def _fail_if_link_is_unsupported_wheel(self, link: Link) -> None:
        if not link.is_wheel:
            return
        wheel = Wheel(link.filename)
        if wheel.supported(self._finder.target_python.get_tags()):
            return
        msg = f"{link.filename} is not a supported wheel on this platform."
        raise UnsupportedWheel(msg)

    def _make_extras_candidate(
        self, base: BaseCandidate, extras: FrozenSet[str]
    ) -> ExtrasCandidate:
        cache_key = (id(base), extras)
        try:
            candidate = self._extras_candidate_cache[cache_key]
        except KeyError:
            candidate = ExtrasCandidate(base, extras)
            self._extras_candidate_cache[cache_key] = candidate
        return candidate

    def _make_candidate_from_dist(
        self,
        dist: BaseDistribution,
        extras: FrozenSet[str],
        template: InstallRequirement,
    ) -> Candidate:
        try:
            base = self._installed_candidate_cache[dist.canonical_name]
        except KeyError:
            base = AlreadyInstalledCandidate(dist, template, factory=self)
            self._installed_candidate_cache[dist.canonical_name] = base
        if not extras:
            return base
        return self._make_extras_candidate(base, extras)

    def _make_candidate_from_link(
        self,
        link: Link,
        extras: FrozenSet[str],
        template: InstallRequirement,
        name: Optional[NormalizedName],
        version: Optional[CandidateVersion],
    ) -> Optional[Candidate]:
        # TODO: Check already installed candidate, and use it if the link and
        # editable flag match.

        if link in self._build_failures:
            # We already tried this candidate before, and it does not build.
            # Don't bother trying again.
            return None

        if template.editable:
            if link not in self._editable_candidate_cache:
                try:
                    self._editable_candidate_cache[link] = EditableCandidate(
                        link,
                        template,
                        factory=self,
                        name=name,
                        version=version,
                    )
                except (InstallationSubprocessError, MetadataInconsistent) as e:
                    logger.warning("Discarding %s. %s", link, e)
                    self._build_failures[link] = e
                    return None
            base: BaseCandidate = self._editable_candidate_cache[link]
        else:
            if link not in self._link_candidate_cache:
                try:
                    self._link_candidate_cache[link] = LinkCandidate(
                        link,
                        template,
                        factory=self,
                        name=name,
                        version=version,
                    )
                except (InstallationSubprocessError, MetadataInconsistent) as e:
                    logger.warning("Discarding %s. %s", link, e)
                    self._build_failures[link] = e
                    return None
            base = self._link_candidate_cache[link]

        if not extras:
            return base
        return self._make_extras_candidate(base, extras)

    def _iter_found_candidates(
        self,
        ireqs: Sequence[InstallRequirement],
        specifier: SpecifierSet,
        hashes: Hashes,
        prefers_installed: bool,
        incompatible_ids: Set[int],
    ) -> Iterable[Candidate]:
        if not ireqs:
            return ()

        # The InstallRequirement implementation requires us to give it a
        # "template". Here we just choose the first requirement to represent
        # all of them.
        # Hopefully the Project model can correct this mismatch in the future.
        template = ireqs[0]
        assert template.req, "Candidates found on index must be PEP 508"
        name = canonicalize_name(template.req.name)

        extras: FrozenSet[str] = frozenset()
        for ireq in ireqs:
            assert ireq.req, "Candidates found on index must be PEP 508"
            specifier &= ireq.req.specifier
            hashes &= ireq.hashes(trust_internet=False)
            extras |= frozenset(ireq.extras)

        def _get_installed_candidate() -> Optional[Candidate]:
            """Get the candidate for the currently-installed version."""
            # If --force-reinstall is set, we want the version from the index
            # instead, so we "pretend" there is nothing installed.
            if self._force_reinstall:
                return None
            try:
                installed_dist = self._installed_dists[name]
            except KeyError:
                return None
            # Don't use the installed distribution if its version does not fit
            # the current dependency graph.
            if not specifier.contains(installed_dist.version, prereleases=True):
                return None
            candidate = self._make_candidate_from_dist(
                dist=installed_dist,
                extras=extras,
                template=template,
            )
            # The candidate is a known incompatiblity. Don't use it.
            if id(candidate) in incompatible_ids:
                return None
            return candidate

        def iter_index_candidate_infos() -> Iterator[IndexCandidateInfo]:
            result = self._finder.find_best_candidate(
                project_name=name,
                specifier=specifier,
                hashes=hashes,
            )
            icans = list(result.iter_applicable())

            # PEP 592: Yanked releases must be ignored unless only yanked
            # releases can satisfy the version range. So if this is false,
            # all yanked icans need to be skipped.
            all_yanked = all(ican.link.is_yanked for ican in icans)

            # PackageFinder returns earlier versions first, so we reverse.
            for ican in reversed(icans):
                if not all_yanked and ican.link.is_yanked:
                    continue
                func = functools.partial(
                    self._make_candidate_from_link,
                    link=ican.link,
                    extras=extras,
                    template=template,
                    name=name,
                    version=ican.version,
                )
                yield ican.version, func

        return FoundCandidates(
            iter_index_candidate_infos,
            _get_installed_candidate(),
            prefers_installed,
            incompatible_ids,
        )

    def _iter_explicit_candidates_from_base(
        self,
        base_requirements: Iterable[Requirement],
        extras: FrozenSet[str],
    ) -> Iterator[Candidate]:
        """Produce explicit candidates from the base given an extra-ed package.

        :param base_requirements: Requirements known to the resolver. The
            requirements are guaranteed to not have extras.
        :param extras: The extras to inject into the explicit requirements'
            candidates.
        """
        for req in base_requirements:
            lookup_cand, _ = req.get_candidate_lookup()
            if lookup_cand is None:  # Not explicit.
                continue
            # We've stripped extras from the identifier, and should always
            # get a BaseCandidate here, unless there's a bug elsewhere.
            base_cand = as_base_candidate(lookup_cand)
            assert base_cand is not None, "no extras here"
            yield self._make_extras_candidate(base_cand, extras)

    def _iter_candidates_from_constraints(
        self,
        identifier: str,
        constraint: Constraint,
        template: InstallRequirement,
    ) -> Iterator[Candidate]:
        """Produce explicit candidates from constraints.

        This creates "fake" InstallRequirement objects that are basically clones
        of what "should" be the template, but with original_link set to link.
        """
        for link in constraint.links:
            self._fail_if_link_is_unsupported_wheel(link)
            candidate = self._make_candidate_from_link(
                link,
                extras=frozenset(),
                template=install_req_from_link_and_ireq(link, template),
                name=canonicalize_name(identifier),
                version=None,
            )
            if candidate:
                yield candidate

    def find_candidates(
        self,
        identifier: str,
        requirements: Mapping[str, Iterable[Requirement]],
        incompatibilities: Mapping[str, Iterator[Candidate]],
        constraint: Constraint,
        prefers_installed: bool,
    ) -> Iterable[Candidate]:
        # Collect basic lookup information from the requirements.
        explicit_candidates: Set[Candidate] = set()
        ireqs: List[InstallRequirement] = []
        for req in requirements[identifier]:
            cand, ireq = req.get_candidate_lookup()
            if cand is not None:
                explicit_candidates.add(cand)
            if ireq is not None:
                ireqs.append(ireq)

        # If the current identifier contains extras, add explicit candidates
        # from entries from extra-less identifier.
        with contextlib.suppress(InvalidRequirement):
            parsed_requirement = get_requirement(identifier)
            explicit_candidates.update(
                self._iter_explicit_candidates_from_base(
                    requirements.get(parsed_requirement.name, ()),
                    frozenset(parsed_requirement.extras),
                ),
            )

        # Add explicit candidates from constraints. We only do this if there are
        # kown ireqs, which represent requirements not already explicit. If
        # there are no ireqs, we're constraining already-explicit requirements,
        # which is handled later when we return the explicit candidates.
        if ireqs:
            try:
                explicit_candidates.update(
                    self._iter_candidates_from_constraints(
                        identifier,
                        constraint,
                        template=ireqs[0],
                    ),
                )
            except UnsupportedWheel:
                # If we're constrained to install a wheel incompatible with the
                # target architecture, no candidates will ever be valid.
                return ()

        # Since we cache all the candidates, incompatibility identification
        # can be made quicker by comparing only the id() values.
        incompat_ids = {id(c) for c in incompatibilities.get(identifier, ())}

        # If none of the requirements want an explicit candidate, we can ask
        # the finder for candidates.
        if not explicit_candidates:
            return self._iter_found_candidates(
                ireqs,
                constraint.specifier,
                constraint.hashes,
                prefers_installed,
                incompat_ids,
            )

        return (
            c
            for c in explicit_candidates
            if id(c) not in incompat_ids
            and constraint.is_satisfied_by(c)
            and all(req.is_satisfied_by(c) for req in requirements[identifier])
        )

    def _make_requirement_from_install_req(
        self, ireq: InstallRequirement, requested_extras: Iterable[str]
    ) -> Optional[Requirement]:
        if not ireq.match_markers(requested_extras):
            logger.info(
                "Ignoring %s: markers '%s' don't match your environment",
                ireq.name,
                ireq.markers,
            )
            return None
        if not ireq.link:
            return SpecifierRequirement(ireq)
        self._fail_if_link_is_unsupported_wheel(ireq.link)
        cand = self._make_candidate_from_link(
            ireq.link,
            extras=frozenset(ireq.extras),
            template=ireq,
            name=canonicalize_name(ireq.name) if ireq.name else None,
            version=None,
        )
        if cand is None:
            # There's no way we can satisfy a URL requirement if the underlying
            # candidate fails to build. An unnamed URL must be user-supplied, so
            # we fail eagerly. If the URL is named, an unsatisfiable requirement
            # can make the resolver do the right thing, either backtrack (and
            # maybe find some other requirement that's buildable) or raise a
            # ResolutionImpossible eventually.
            if not ireq.name:
                raise self._build_failures[ireq.link]
            return UnsatisfiableRequirement(canonicalize_name(ireq.name))
        return self.make_requirement_from_candidate(cand)

    def collect_root_requirements(
        self, root_ireqs: List[InstallRequirement]
    ) -> CollectedRootRequirements:
        collected = CollectedRootRequirements([], {}, {})
        for i, ireq in enumerate(root_ireqs):
            if ireq.constraint:
                # Ensure we only accept valid constraints
                problem = check_invalid_constraint_type(ireq)
                if problem:
                    raise InstallationError(problem)
                if not ireq.match_markers():
                    continue
                assert ireq.name, "Constraint must be named"
                name = canonicalize_name(ireq.name)
                if name in collected.constraints:
                    collected.constraints[name] &= ireq
                else:
                    collected.constraints[name] = Constraint.from_ireq(ireq)
            else:
                req = self._make_requirement_from_install_req(
                    ireq,
                    requested_extras=(),
                )
                if req is None:
                    continue
                if ireq.user_supplied and req.name not in collected.user_requested:
                    collected.user_requested[req.name] = i
                collected.requirements.append(req)
        return collected

    def make_requirement_from_candidate(
        self, candidate: Candidate
    ) -> ExplicitRequirement:
        return ExplicitRequirement(candidate)

    def make_requirement_from_spec(
        self,
        specifier: str,
        comes_from: Optional[InstallRequirement],
        requested_extras: Iterable[str] = (),
    ) -> Optional[Requirement]:
        ireq = self._make_install_req_from_spec(specifier, comes_from)
        return self._make_requirement_from_install_req(ireq, requested_extras)

    def make_requires_python_requirement(
        self,
        specifier: SpecifierSet,
    ) -> Optional[Requirement]:
        if self._ignore_requires_python:
            return None
        # Don't bother creating a dependency for an empty Requires-Python.
        if not str(specifier):
            return None
        return RequiresPythonRequirement(specifier, self._python_candidate)

    def get_wheel_cache_entry(
        self, link: Link, name: Optional[str]
    ) -> Optional[CacheEntry]:
        """Look up the link in the wheel cache.

        If ``preparer.require_hashes`` is True, don't use the wheel cache,
        because cached wheels, always built locally, have different hashes
        than the files downloaded from the index server and thus throw false
        hash mismatches. Furthermore, cached wheels at present have
        nondeterministic contents due to file modification times.
        """
        if self._wheel_cache is None or self.preparer.require_hashes:
            return None
        return self._wheel_cache.get_cache_entry(
            link=link,
            package_name=name,
            supported_tags=get_supported(),
        )

    def get_dist_to_uninstall(self, candidate: Candidate) -> Optional[BaseDistribution]:
        # TODO: Are there more cases this needs to return True? Editable?
        dist = self._installed_dists.get(candidate.project_name)
        if dist is None:  # Not installed, no uninstallation required.
            return None

        # Prevent uninstalling packages from /usr
        try:
            if dist_location(dist._dist) in (
                    sysconfig.get_path('purelib', scheme='rpm_prefix', vars={'base': sys.base_prefix}),
                    sysconfig.get_path('platlib', scheme='rpm_prefix', vars={'base': sys.base_prefix}),
                ):
                return None
        except KeyError:  # this Python doesn't have 'rpm_prefix' scheme yet
            pass

        # We're installing into global site. The current installation must
        # be uninstalled, no matter it's in global or user site, because the
        # user site installation has precedence over global.
        if not self._use_user_site:
            return dist

        # We're installing into user site. Remove the user site installation.
        if dist.in_usersite:
            return dist

        # We're installing into user site, but the installed incompatible
        # package is in global site. We can't uninstall that, and would let
        # the new user installation to "shadow" it. But shadowing won't work
        # in virtual environments, so we error out.
        if running_under_virtualenv() and dist.in_site_packages:
            message = (
                f"Will not install to the user site because it will lack "
                f"sys.path precedence to {dist.raw_name} in {dist.location}"
            )
            raise InstallationError(message)
        return None

    def _report_requires_python_error(
        self, causes: Sequence["ConflictCause"]
    ) -> UnsupportedPythonVersion:
        assert causes, "Requires-Python error reported with no cause"

        version = self._python_candidate.version

        if len(causes) == 1:
            specifier = str(causes[0].requirement.specifier)
            message = (
                f"Package {causes[0].parent.name!r} requires a different "
                f"Python: {version} not in {specifier!r}"
            )
            return UnsupportedPythonVersion(message)

        message = f"Packages require a different Python. {version} not in:"
        for cause in causes:
            package = cause.parent.format_for_error()
            specifier = str(cause.requirement.specifier)
            message += f"\n{specifier!r} (required by {package})"
        return UnsupportedPythonVersion(message)

    def _report_single_requirement_conflict(
        self, req: Requirement, parent: Optional[Candidate]
    ) -> DistributionNotFound:
        if parent is None:
            req_disp = str(req)
        else:
            req_disp = f"{req} (from {parent.name})"

        cands = self._finder.find_all_candidates(req.project_name)
        versions = [str(v) for v in sorted({c.version for c in cands})]

        logger.critical(
            "Could not find a version that satisfies the requirement %s "
            "(from versions: %s)",
            req_disp,
            ", ".join(versions) or "none",
        )
        if str(req) == "requirements.txt":
            logger.info(
                "HINT: You are attempting to install a package literally "
                'named "requirements.txt" (which cannot exist). Consider '
                "using the '-r' flag to install the packages listed in "
                "requirements.txt"
            )

        return DistributionNotFound(f"No matching distribution found for {req}")

    def get_installation_error(
        self,
        e: "ResolutionImpossible[Requirement, Candidate]",
        constraints: Dict[str, Constraint],
    ) -> InstallationError:

        assert e.causes, "Installation error reported with no cause"

        # If one of the things we can't solve is "we need Python X.Y",
        # that is what we report.
        requires_python_causes = [
            cause
            for cause in e.causes
            if isinstance(cause.requirement, RequiresPythonRequirement)
            and not cause.requirement.is_satisfied_by(self._python_candidate)
        ]
        if requires_python_causes:
            # The comprehension above makes sure all Requirement instances are
            # RequiresPythonRequirement, so let's cast for convinience.
            return self._report_requires_python_error(
                cast("Sequence[ConflictCause]", requires_python_causes),
            )

        # Otherwise, we have a set of causes which can't all be satisfied
        # at once.

        # The simplest case is when we have *one* cause that can't be
        # satisfied. We just report that case.
        if len(e.causes) == 1:
            req, parent = e.causes[0]
            if req.name not in constraints:
                return self._report_single_requirement_conflict(req, parent)

        # OK, we now have a list of requirements that can't all be
        # satisfied at once.

        # A couple of formatting helpers
        def text_join(parts: List[str]) -> str:
            if len(parts) == 1:
                return parts[0]

            return ", ".join(parts[:-1]) + " and " + parts[-1]

        def describe_trigger(parent: Candidate) -> str:
            ireq = parent.get_install_requirement()
            if not ireq or not ireq.comes_from:
                return f"{parent.name}=={parent.version}"
            if isinstance(ireq.comes_from, InstallRequirement):
                return str(ireq.comes_from.name)
            return str(ireq.comes_from)

        triggers = set()
        for req, parent in e.causes:
            if parent is None:
                # This is a root requirement, so we can report it directly
                trigger = req.format_for_error()
            else:
                trigger = describe_trigger(parent)
            triggers.add(trigger)

        if triggers:
            info = text_join(sorted(triggers))
        else:
            info = "the requested packages"

        msg = (
            "Cannot install {} because these package versions "
            "have conflicting dependencies.".format(info)
        )
        logger.critical(msg)
        msg = "\nThe conflict is caused by:"

        relevant_constraints = set()
        for req, parent in e.causes:
            if req.name in constraints:
                relevant_constraints.add(req.name)
            msg = msg + "\n    "
            if parent:
                msg = msg + f"{parent.name} {parent.version} depends on "
            else:
                msg = msg + "The user requested "
            msg = msg + req.format_for_error()
        for key in relevant_constraints:
            spec = constraints[key].specifier
            msg += f"\n    The user requested (constraint) {key}{spec}"

        msg = (
            msg
            + "\n\n"
            + "To fix this you could try to:\n"
            + "1. loosen the range of package versions you've specified\n"
            + "2. remove package versions to allow pip attempt to solve "
            + "the dependency conflict\n"
        )

        logger.info(msg)

        return DistributionNotFound(
            "ResolutionImpossible: for help visit "
            "https://pip.pypa.io/en/latest/user_guide/"
            "#fixing-conflicting-dependencies"
        )
site-packages/pip/_internal/resolution/resolvelib/requirements.py000064400000012517150564745010021512 0ustar00from pip._vendor.packaging.specifiers import SpecifierSet
from pip._vendor.packaging.utils import NormalizedName, canonicalize_name

from pip._internal.req.req_install import InstallRequirement

from .base import Candidate, CandidateLookup, Requirement, format_name


class ExplicitRequirement(Requirement):
    def __init__(self, candidate: Candidate) -> None:
        self.candidate = candidate

    def __str__(self) -> str:
        return str(self.candidate)

    def __repr__(self) -> str:
        return "{class_name}({candidate!r})".format(
            class_name=self.__class__.__name__,
            candidate=self.candidate,
        )

    @property
    def project_name(self) -> NormalizedName:
        # No need to canonicalise - the candidate did this
        return self.candidate.project_name

    @property
    def name(self) -> str:
        # No need to canonicalise - the candidate did this
        return self.candidate.name

    def format_for_error(self) -> str:
        return self.candidate.format_for_error()

    def get_candidate_lookup(self) -> CandidateLookup:
        return self.candidate, None

    def is_satisfied_by(self, candidate: Candidate) -> bool:
        return candidate == self.candidate


class SpecifierRequirement(Requirement):
    def __init__(self, ireq: InstallRequirement) -> None:
        assert ireq.link is None, "This is a link, not a specifier"
        self._ireq = ireq
        self._extras = frozenset(ireq.extras)

    def __str__(self) -> str:
        return str(self._ireq.req)

    def __repr__(self) -> str:
        return "{class_name}({requirement!r})".format(
            class_name=self.__class__.__name__,
            requirement=str(self._ireq.req),
        )

    @property
    def project_name(self) -> NormalizedName:
        assert self._ireq.req, "Specifier-backed ireq is always PEP 508"
        return canonicalize_name(self._ireq.req.name)

    @property
    def name(self) -> str:
        return format_name(self.project_name, self._extras)

    def format_for_error(self) -> str:

        # Convert comma-separated specifiers into "A, B, ..., F and G"
        # This makes the specifier a bit more "human readable", without
        # risking a change in meaning. (Hopefully! Not all edge cases have
        # been checked)
        parts = [s.strip() for s in str(self).split(",")]
        if len(parts) == 0:
            return ""
        elif len(parts) == 1:
            return parts[0]

        return ", ".join(parts[:-1]) + " and " + parts[-1]

    def get_candidate_lookup(self) -> CandidateLookup:
        return None, self._ireq

    def is_satisfied_by(self, candidate: Candidate) -> bool:
        assert candidate.name == self.name, (
            f"Internal issue: Candidate is not for this requirement "
            f"{candidate.name} vs {self.name}"
        )
        # We can safely always allow prereleases here since PackageFinder
        # already implements the prerelease logic, and would have filtered out
        # prerelease candidates if the user does not expect them.
        assert self._ireq.req, "Specifier-backed ireq is always PEP 508"
        spec = self._ireq.req.specifier
        return spec.contains(candidate.version, prereleases=True)


class RequiresPythonRequirement(Requirement):
    """A requirement representing Requires-Python metadata."""

    def __init__(self, specifier: SpecifierSet, match: Candidate) -> None:
        self.specifier = specifier
        self._candidate = match

    def __str__(self) -> str:
        return f"Python {self.specifier}"

    def __repr__(self) -> str:
        return "{class_name}({specifier!r})".format(
            class_name=self.__class__.__name__,
            specifier=str(self.specifier),
        )

    @property
    def project_name(self) -> NormalizedName:
        return self._candidate.project_name

    @property
    def name(self) -> str:
        return self._candidate.name

    def format_for_error(self) -> str:
        return str(self)

    def get_candidate_lookup(self) -> CandidateLookup:
        if self.specifier.contains(self._candidate.version, prereleases=True):
            return self._candidate, None
        return None, None

    def is_satisfied_by(self, candidate: Candidate) -> bool:
        assert candidate.name == self._candidate.name, "Not Python candidate"
        # We can safely always allow prereleases here since PackageFinder
        # already implements the prerelease logic, and would have filtered out
        # prerelease candidates if the user does not expect them.
        return self.specifier.contains(candidate.version, prereleases=True)


class UnsatisfiableRequirement(Requirement):
    """A requirement that cannot be satisfied."""

    def __init__(self, name: NormalizedName) -> None:
        self._name = name

    def __str__(self) -> str:
        return f"{self._name} (unavailable)"

    def __repr__(self) -> str:
        return "{class_name}({name!r})".format(
            class_name=self.__class__.__name__,
            name=str(self._name),
        )

    @property
    def project_name(self) -> NormalizedName:
        return self._name

    @property
    def name(self) -> str:
        return self._name

    def format_for_error(self) -> str:
        return str(self)

    def get_candidate_lookup(self) -> CandidateLookup:
        return None, None

    def is_satisfied_by(self, candidate: Candidate) -> bool:
        return False
site-packages/pip/_internal/resolution/__init__.py000064400000000000150564745010016340 0ustar00site-packages/pip/_internal/resolution/legacy/__pycache__/resolver.cpython-39.pyc000064400000030035150564745010024210 0ustar00a

��ReH�@szdZddlZddlZddlmZddlmZddlmZm	Z	m
Z
mZmZm
Z
ddlmZddlmZddlmZdd	lmZmZmZmZmZmZdd
lmZddlmZddlm Z dd
l!m"Z"ddl#m$Z$m%Z%ddl&m'Z'ddl(m)Z)m*Z*ddl+m,Z,ddl-m.Z.ddl/m0Z0m1Z1ddl/m2Z2ddl3m4Z4e�5e6�Z7ee8e
e$fZ9dee
e:e:e:fe;dd�dd�Z<Gdd�de)�Z=dS)ayDependency Resolution

The dependency resolution in pip is performed as follows:

for top-level requirements:
    a. only one spec allowed per project, regardless of conflicts or not.
       otherwise a "double requirement" exception is raised
    b. they override sub-dependency requirements.
for sub-dependencies
    a. "first found, wins" (where the order is breadth first)
�N)�defaultdict)�chain)�DefaultDict�Iterable�List�Optional�Set�Tuple)�
specifiers)�Requirement)�
WheelCache)�BestVersionAlreadyInstalled�DistributionNotFound�	HashError�
HashErrors�NoneMetadataError�UnsupportedPythonVersion)�
PackageFinder)�BaseDistribution)�Link)�RequirementPreparer)�InstallRequirement�check_invalid_constraint_type)�RequirementSet)�BaseResolver�InstallRequirementProvider)�
get_supported)�
indent_log)�dist_in_usersite�normalize_version_info)�dist_in_install_path)�check_requires_pythonF)�dist�version_info�ignore_requires_python�returnc
Cs�zt|j�}Wn2ty@}zt|t|���WYd}~n
d}~00zt||d�}Wn8tjy�}zt�d|j	|�WYd}~dSd}~00|r�dSd�
tt|��}|r�t�d|j	||�dSt
d�|j	||���dS)a�
    Check whether the given Python version is compatible with a distribution's
    "Requires-Python" value.

    :param version_info: A 3-tuple of ints representing the Python
        major-minor-micro version to check.
    :param ignore_requires_python: Whether to ignore the "Requires-Python"
        value if the given Python version isn't compatible.

    :raises UnsupportedPythonVersion: When the given Python version isn't
        compatible.
    N)r#z-Package %r has an invalid Requires-Python: %s�.zBIgnoring failed Requires-Python check for package %r: %s not in %rz8Package {!r} requires a different Python: {} not in {!r})�str�requires_python�FileNotFoundErrorrr!r
�InvalidSpecifier�logger�warning�raw_name�join�map�debugr�format)r"r#r$r(�e�
is_compatible�exc�version�r6��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/legacy/resolver.py�_check_dist_requires_python7s<$�
����r8cs�eZdZdZhd�Zdeeeee	e
e
e
e
e
eeee
dfdd��fdd�
Zeee
ed�d	d
�Zee
d�dd
�Zedd�dd�Zeeed�dd�Zeeed�dd�Zedd�dd�Zeed�dd�Zeeeed�dd�Zeeed�dd�Z�ZS) �Resolverz�Resolves which packages need to be installed/uninstalled to perform     the requested operation without breaking the requirements of any package.
    >�to-satisfy-only�eager�only-if-neededN.)�preparer�finder�wheel_cache�make_install_req�
use_user_site�ignore_dependencies�ignore_installedr$�force_reinstall�upgrade_strategy�py_version_infor%cs�t���|
|jvsJ�|dur0tjdd�}nt|�}||_||_||_||_	|
|_
|	|_||_||_
||_||_||_tt�|_dS)N�)�super�__init__�_allowed_strategies�sysr#r�_py_version_infor=r>r?rErDrBrCr$rA�_make_install_reqr�list�_discovered_dependencies)�selfr=r>r?r@rArBrCr$rDrErF��	__class__r6r7rIus"
zResolver.__init__)�	root_reqs�check_supported_wheelsr%cCs�t|d�}|D]}|jr t|�|�|�qg}t�}t|j|�D]P}z|�|�||��WqBt	y�}z||_
|�|�WYd}~qBd}~00qB|r�|�|S)a�Resolve what operations need to be done

        As a side-effect of this method, the packages (and their dependencies)
        are downloaded, unpacked and prepared for installation. This
        preparation is done by ``pip.operations.prepare``.

        Once PyPI has static dependency metadata available, it would be
        possible to move the preparation to become a step separated from
        dependency resolution.
        )rTN)r�
constraintr�add_requirementrr�all_requirements�extend�_resolve_oner�req�append)rPrSrT�requirement_setrZZdiscovered_reqsZhash_errorsr4r6r6r7�resolve�s 
"zResolver.resolve�rZr%cCs:|jdkrdS|jdkrdS|jdks*J�|jp4|jSdS)Nr:Fr;Tr<)rE�
user_suppliedrU�rPrZr6r6r7�_is_upgrade_allowed�s

zResolver._is_upgrade_allowedcCs*|jrt|j�r t|j�r d|_d|_dS)z4
        Set a requirement to be installed.
        TN)rAr�satisfied_byr �should_reinstallr`r6r6r7�_set_req_to_reinstall�s��zResolver._set_req_to_reinstall)�req_to_installr%cCs�|jr
dS|�|j�|js dS|jr4|�|�dS|�|�sP|jdkrLdSdS|js�z|j	j
|dd�Wn$ty~YdSty�Yn0|�|�dS)aCheck if req_to_install should be skipped.

        This will check if the req is installed, and whether we should upgrade
        or reinstall it, taking into account all the relevant user options.

        After calling this req_to_install will only have satisfied_by set to
        None if the req_to_install is to be upgraded/reinstalled etc. Any
        other value will be a dist recording the current thing installed that
        satisfies the requirement.

        Note that for vcs urls and the like we can't assess skipping in this
        routine - we simply identify that we need to pull the thing down,
        then later on it is pulled down and introspected to assess upgrade/
        reinstalls etc.

        :return: A text reason for why it was skipped, or None.
        Nr<z#already satisfied, skipping upgradezalready satisfiedT)�upgradezalready up-to-date)
rC�check_if_existsrArbrDrdrarE�linkr>�find_requirementr
r)rPrer6r6r7�_check_skip_installed�s*



zResolver._check_skip_installedcCsR|�|�}|j�||�}|s dS|j}|jrN|jp4d}dj||d�}t�|�|S)Nz<none given>zqThe candidate selected for download or install is a yanked version: {candidate}
Reason for being yanked: {reason})�	candidate�reason)	rar>rirh�	is_yanked�
yanked_reasonr1r+r,)rPrZrf�best_candidaterhrl�msgr6r6r7�_find_requirement_links

��
zResolver._find_requirement_linkcCs~|jdur|�|�|_|jdus(|jjr,dS|jj|j|jt�d�}|durzt�	d|j�|j|j
urr|jrrd|_|j|_dS)afEnsure that if a link can be found for this, that it is found.

        Note that req.link may still be None - if the requirement is already
        installed and not needed to be upgraded based on the return value of
        _is_upgrade_allowed().

        If preparer.require_hashes is True, don't use the wheel cache, because
        cached wheels, always built locally, have different hashes than the
        files downloaded from the index server and thus throw false hash
        mismatches. Furthermore, cached wheels at present have undeterministic
        contents due to file modification times.
        N)rh�package_name�supported_tagszUsing cached wheel link: %sT)
rhrqr?r=�require_hashes�get_cache_entry�namerr+r0�
original_link�
persistent�original_link_is_in_wheel_cache)rPrZ�cache_entryr6r6r7�_populate_link!s
�zResolver._populate_linkcCs�|jr|j�|�S|jdus J�|�|�}|jr>|j�||�S|�|�|j�|�}|jsf|�	|j
�|jr�|jdkp�|jp�|jp�|j
jdk}|r�|�|�nt�d|�|S)zzTakes a InstallRequirement and returns a single AbstractDist         representing a prepared variant of the same.
        Nr:�filez<Requirement already satisfied (use --upgrade to upgrade): %s)�editabler=�prepare_editable_requirementrbrj�prepare_installed_requirementr{�prepare_linked_requirementrCrgrArErDrh�schemerdr+�info)rPrZ�skip_reasonr"Z
should_modifyr6r6r7�
_get_dist_for>s2



��
��zResolver._get_dist_for)r\rer%c	s8�js�jrgSd�_����}t|�j�jd�g�tttdd�����fdd�}t	�����
�j�s��jstJ��j
�dd��j�s�jr�t�dd	��j��tt�j�t|����}|D]}t�d
|j|j|�q�tt|���t�j�@�}|�|�D]}|||d��qWd�n1�s*0Y�S)zxPrepare a single requirements file.

        :return: A list of additional InstallRequirements to also install.
        T)r#r$N)�subreq�extras_requestedr%csP��t|���}�j}�j|||d�\}}|rB|rB�j|�|���|�dS)N)�parent_req_namer�)rMr'rvrVrOr[rX)r�r�Zsub_install_reqr�Z
to_scan_againZ
add_to_parent�Z	more_reqsrer\rPr6r7�add_req�s�
z&Resolver._resolve_one.<locals>.add_req)r�z!Installing extra requirements: %r�,z%%s %s does not provide the extra '%s')r�)rU�preparedr�r8rLr$rrr'r�has_requirementrvr_rVrB�extrasr+r0r.�sorted�set�iter_provided_extrasr,r-r5�iter_dependencies)	rPr\rer"r�Zmissing_requested�missingZavailable_requestedr�r6r�r7rYlsJ
� 

����0zResolver._resolve_one)�req_setr%cs@g�t��tdd�����fdd��|j��D]}�|�q.�S)z�Create the installation order.

        The installation order is topological - requirements are installed
        before the requiring thing. We break cycles at an arbitrary point,
        and make no other guarantees.
        Nr^csN|js|�vrdS|jrdS��|��j|jD]}�|�q2��|�dS)N)rbrU�addrOrvr[)rZ�dep��orderZordered_reqs�schedulerPr6r7r��s

z1Resolver.get_installation_order.<locals>.schedule)r�r�requirements�values)rPr��install_reqr6r�r7�get_installation_order�s

zResolver.get_installation_order)N)�__name__�
__module__�__qualname__�__doc__rJrrrrr�boolr'r	�intrIrrrr]rardrjrrqr{rr�rYr��
__classcell__r6r6rQr7r9nsF��'�%	
�60�P�r9)F)>r��loggingrK�collectionsr�	itertoolsr�typingrrrrrr	Zpip._vendor.packagingr
Z"pip._vendor.packaging.requirementsr�pip._internal.cacher�pip._internal.exceptionsr
rrrrr�"pip._internal.index.package_finderr�pip._internal.metadatar�pip._internal.models.linkr� pip._internal.operations.preparerZpip._internal.req.req_installrrZpip._internal.req.req_setr�pip._internal.resolution.baserr�&pip._internal.utils.compatibility_tagsr�pip._internal.utils.loggingr�pip._internal.utils.miscrrr �pip._internal.utils.packagingr!�	getLoggerr�r+r'ZDiscoveredDependenciesr�r�r8r9r6r6r6r7�<module>s>  
��7site-packages/pip/_internal/resolution/legacy/__pycache__/__init__.cpython-39.pyc000064400000000362150564745010024106 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/resolution/legacy/__init__.py�<module>�site-packages/pip/_internal/resolution/legacy/__init__.py000064400000000000150564745010017604 0ustar00site-packages/pip/_internal/resolution/legacy/resolver.py000064400000044021150564745010017721 0ustar00"""Dependency Resolution

The dependency resolution in pip is performed as follows:

for top-level requirements:
    a. only one spec allowed per project, regardless of conflicts or not.
       otherwise a "double requirement" exception is raised
    b. they override sub-dependency requirements.
for sub-dependencies
    a. "first found, wins" (where the order is breadth first)
"""

# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import logging
import sys
from collections import defaultdict
from itertools import chain
from typing import DefaultDict, Iterable, List, Optional, Set, Tuple

from pip._vendor.packaging import specifiers
from pip._vendor.packaging.requirements import Requirement

from pip._internal.cache import WheelCache
from pip._internal.exceptions import (
    BestVersionAlreadyInstalled,
    DistributionNotFound,
    HashError,
    HashErrors,
    NoneMetadataError,
    UnsupportedPythonVersion,
)
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import BaseDistribution
from pip._internal.models.link import Link
from pip._internal.operations.prepare import RequirementPreparer
from pip._internal.req.req_install import (
    InstallRequirement,
    check_invalid_constraint_type,
)
from pip._internal.req.req_set import RequirementSet
from pip._internal.resolution.base import BaseResolver, InstallRequirementProvider
from pip._internal.utils.compatibility_tags import get_supported
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import dist_in_usersite, normalize_version_info
from pip._internal.utils.misc import dist_in_install_path
from pip._internal.utils.packaging import check_requires_python

logger = logging.getLogger(__name__)

DiscoveredDependencies = DefaultDict[str, List[InstallRequirement]]


def _check_dist_requires_python(
    dist: BaseDistribution,
    version_info: Tuple[int, int, int],
    ignore_requires_python: bool = False,
) -> None:
    """
    Check whether the given Python version is compatible with a distribution's
    "Requires-Python" value.

    :param version_info: A 3-tuple of ints representing the Python
        major-minor-micro version to check.
    :param ignore_requires_python: Whether to ignore the "Requires-Python"
        value if the given Python version isn't compatible.

    :raises UnsupportedPythonVersion: When the given Python version isn't
        compatible.
    """
    # This idiosyncratically converts the SpecifierSet to str and let
    # check_requires_python then parse it again into SpecifierSet. But this
    # is the legacy resolver so I'm just not going to bother refactoring.
    try:
        requires_python = str(dist.requires_python)
    except FileNotFoundError as e:
        raise NoneMetadataError(dist, str(e))
    try:
        is_compatible = check_requires_python(
            requires_python,
            version_info=version_info,
        )
    except specifiers.InvalidSpecifier as exc:
        logger.warning(
            "Package %r has an invalid Requires-Python: %s", dist.raw_name, exc
        )
        return

    if is_compatible:
        return

    version = ".".join(map(str, version_info))
    if ignore_requires_python:
        logger.debug(
            "Ignoring failed Requires-Python check for package %r: %s not in %r",
            dist.raw_name,
            version,
            requires_python,
        )
        return

    raise UnsupportedPythonVersion(
        "Package {!r} requires a different Python: {} not in {!r}".format(
            dist.raw_name, version, requires_python
        )
    )


class Resolver(BaseResolver):
    """Resolves which packages need to be installed/uninstalled to perform \
    the requested operation without breaking the requirements of any package.
    """

    _allowed_strategies = {"eager", "only-if-needed", "to-satisfy-only"}

    def __init__(
        self,
        preparer: RequirementPreparer,
        finder: PackageFinder,
        wheel_cache: Optional[WheelCache],
        make_install_req: InstallRequirementProvider,
        use_user_site: bool,
        ignore_dependencies: bool,
        ignore_installed: bool,
        ignore_requires_python: bool,
        force_reinstall: bool,
        upgrade_strategy: str,
        py_version_info: Optional[Tuple[int, ...]] = None,
    ) -> None:
        super().__init__()
        assert upgrade_strategy in self._allowed_strategies

        if py_version_info is None:
            py_version_info = sys.version_info[:3]
        else:
            py_version_info = normalize_version_info(py_version_info)

        self._py_version_info = py_version_info

        self.preparer = preparer
        self.finder = finder
        self.wheel_cache = wheel_cache

        self.upgrade_strategy = upgrade_strategy
        self.force_reinstall = force_reinstall
        self.ignore_dependencies = ignore_dependencies
        self.ignore_installed = ignore_installed
        self.ignore_requires_python = ignore_requires_python
        self.use_user_site = use_user_site
        self._make_install_req = make_install_req

        self._discovered_dependencies: DiscoveredDependencies = defaultdict(list)

    def resolve(
        self, root_reqs: List[InstallRequirement], check_supported_wheels: bool
    ) -> RequirementSet:
        """Resolve what operations need to be done

        As a side-effect of this method, the packages (and their dependencies)
        are downloaded, unpacked and prepared for installation. This
        preparation is done by ``pip.operations.prepare``.

        Once PyPI has static dependency metadata available, it would be
        possible to move the preparation to become a step separated from
        dependency resolution.
        """
        requirement_set = RequirementSet(check_supported_wheels=check_supported_wheels)
        for req in root_reqs:
            if req.constraint:
                check_invalid_constraint_type(req)
            requirement_set.add_requirement(req)

        # Actually prepare the files, and collect any exceptions. Most hash
        # exceptions cannot be checked ahead of time, because
        # _populate_link() needs to be called before we can make decisions
        # based on link type.
        discovered_reqs: List[InstallRequirement] = []
        hash_errors = HashErrors()
        for req in chain(requirement_set.all_requirements, discovered_reqs):
            try:
                discovered_reqs.extend(self._resolve_one(requirement_set, req))
            except HashError as exc:
                exc.req = req
                hash_errors.append(exc)

        if hash_errors:
            raise hash_errors

        return requirement_set

    def _is_upgrade_allowed(self, req: InstallRequirement) -> bool:
        if self.upgrade_strategy == "to-satisfy-only":
            return False
        elif self.upgrade_strategy == "eager":
            return True
        else:
            assert self.upgrade_strategy == "only-if-needed"
            return req.user_supplied or req.constraint

    def _set_req_to_reinstall(self, req: InstallRequirement) -> None:
        """
        Set a requirement to be installed.
        """
        # Don't uninstall the conflict if doing a user install and the
        # conflict is not a user install.
        if ((not self.use_user_site
                or dist_in_usersite(req.satisfied_by))
                and dist_in_install_path(req.satisfied_by)):
            req.should_reinstall = True
        req.satisfied_by = None

    def _check_skip_installed(
        self, req_to_install: InstallRequirement
    ) -> Optional[str]:
        """Check if req_to_install should be skipped.

        This will check if the req is installed, and whether we should upgrade
        or reinstall it, taking into account all the relevant user options.

        After calling this req_to_install will only have satisfied_by set to
        None if the req_to_install is to be upgraded/reinstalled etc. Any
        other value will be a dist recording the current thing installed that
        satisfies the requirement.

        Note that for vcs urls and the like we can't assess skipping in this
        routine - we simply identify that we need to pull the thing down,
        then later on it is pulled down and introspected to assess upgrade/
        reinstalls etc.

        :return: A text reason for why it was skipped, or None.
        """
        if self.ignore_installed:
            return None

        req_to_install.check_if_exists(self.use_user_site)
        if not req_to_install.satisfied_by:
            return None

        if self.force_reinstall:
            self._set_req_to_reinstall(req_to_install)
            return None

        if not self._is_upgrade_allowed(req_to_install):
            if self.upgrade_strategy == "only-if-needed":
                return "already satisfied, skipping upgrade"
            return "already satisfied"

        # Check for the possibility of an upgrade.  For link-based
        # requirements we have to pull the tree down and inspect to assess
        # the version #, so it's handled way down.
        if not req_to_install.link:
            try:
                self.finder.find_requirement(req_to_install, upgrade=True)
            except BestVersionAlreadyInstalled:
                # Then the best version is installed.
                return "already up-to-date"
            except DistributionNotFound:
                # No distribution found, so we squash the error.  It will
                # be raised later when we re-try later to do the install.
                # Why don't we just raise here?
                pass

        self._set_req_to_reinstall(req_to_install)
        return None

    def _find_requirement_link(self, req: InstallRequirement) -> Optional[Link]:
        upgrade = self._is_upgrade_allowed(req)
        best_candidate = self.finder.find_requirement(req, upgrade)
        if not best_candidate:
            return None

        # Log a warning per PEP 592 if necessary before returning.
        link = best_candidate.link
        if link.is_yanked:
            reason = link.yanked_reason or "<none given>"
            msg = (
                # Mark this as a unicode string to prevent
                # "UnicodeEncodeError: 'ascii' codec can't encode character"
                # in Python 2 when the reason contains non-ascii characters.
                "The candidate selected for download or install is a "
                "yanked version: {candidate}\n"
                "Reason for being yanked: {reason}"
            ).format(candidate=best_candidate, reason=reason)
            logger.warning(msg)

        return link

    def _populate_link(self, req: InstallRequirement) -> None:
        """Ensure that if a link can be found for this, that it is found.

        Note that req.link may still be None - if the requirement is already
        installed and not needed to be upgraded based on the return value of
        _is_upgrade_allowed().

        If preparer.require_hashes is True, don't use the wheel cache, because
        cached wheels, always built locally, have different hashes than the
        files downloaded from the index server and thus throw false hash
        mismatches. Furthermore, cached wheels at present have undeterministic
        contents due to file modification times.
        """
        if req.link is None:
            req.link = self._find_requirement_link(req)

        if self.wheel_cache is None or self.preparer.require_hashes:
            return
        cache_entry = self.wheel_cache.get_cache_entry(
            link=req.link,
            package_name=req.name,
            supported_tags=get_supported(),
        )
        if cache_entry is not None:
            logger.debug("Using cached wheel link: %s", cache_entry.link)
            if req.link is req.original_link and cache_entry.persistent:
                req.original_link_is_in_wheel_cache = True
            req.link = cache_entry.link

    def _get_dist_for(self, req: InstallRequirement) -> BaseDistribution:
        """Takes a InstallRequirement and returns a single AbstractDist \
        representing a prepared variant of the same.
        """
        if req.editable:
            return self.preparer.prepare_editable_requirement(req)

        # satisfied_by is only evaluated by calling _check_skip_installed,
        # so it must be None here.
        assert req.satisfied_by is None
        skip_reason = self._check_skip_installed(req)

        if req.satisfied_by:
            return self.preparer.prepare_installed_requirement(req, skip_reason)

        # We eagerly populate the link, since that's our "legacy" behavior.
        self._populate_link(req)
        dist = self.preparer.prepare_linked_requirement(req)

        # NOTE
        # The following portion is for determining if a certain package is
        # going to be re-installed/upgraded or not and reporting to the user.
        # This should probably get cleaned up in a future refactor.

        # req.req is only avail after unpack for URL
        # pkgs repeat check_if_exists to uninstall-on-upgrade
        # (#14)
        if not self.ignore_installed:
            req.check_if_exists(self.use_user_site)

        if req.satisfied_by:
            should_modify = (
                self.upgrade_strategy != "to-satisfy-only"
                or self.force_reinstall
                or self.ignore_installed
                or req.link.scheme == "file"
            )
            if should_modify:
                self._set_req_to_reinstall(req)
            else:
                logger.info(
                    "Requirement already satisfied (use --upgrade to upgrade): %s",
                    req,
                )
        return dist

    def _resolve_one(
        self,
        requirement_set: RequirementSet,
        req_to_install: InstallRequirement,
    ) -> List[InstallRequirement]:
        """Prepare a single requirements file.

        :return: A list of additional InstallRequirements to also install.
        """
        # Tell user what we are doing for this requirement:
        # obtain (editable), skipping, processing (local url), collecting
        # (remote url or package name)
        if req_to_install.constraint or req_to_install.prepared:
            return []

        req_to_install.prepared = True

        # Parse and return dependencies
        dist = self._get_dist_for(req_to_install)
        # This will raise UnsupportedPythonVersion if the given Python
        # version isn't compatible with the distribution's Requires-Python.
        _check_dist_requires_python(
            dist,
            version_info=self._py_version_info,
            ignore_requires_python=self.ignore_requires_python,
        )

        more_reqs: List[InstallRequirement] = []

        def add_req(subreq: Requirement, extras_requested: Iterable[str]) -> None:
            # This idiosyncratically converts the Requirement to str and let
            # make_install_req then parse it again into Requirement. But this is
            # the legacy resolver so I'm just not going to bother refactoring.
            sub_install_req = self._make_install_req(str(subreq), req_to_install)
            parent_req_name = req_to_install.name
            to_scan_again, add_to_parent = requirement_set.add_requirement(
                sub_install_req,
                parent_req_name=parent_req_name,
                extras_requested=extras_requested,
            )
            if parent_req_name and add_to_parent:
                self._discovered_dependencies[parent_req_name].append(add_to_parent)
            more_reqs.extend(to_scan_again)

        with indent_log():
            # We add req_to_install before its dependencies, so that we
            # can refer to it when adding dependencies.
            if not requirement_set.has_requirement(req_to_install.name):
                # 'unnamed' requirements will get added here
                # 'unnamed' requirements can only come from being directly
                # provided by the user.
                assert req_to_install.user_supplied
                requirement_set.add_requirement(req_to_install, parent_req_name=None)

            if not self.ignore_dependencies:
                if req_to_install.extras:
                    logger.debug(
                        "Installing extra requirements: %r",
                        ",".join(req_to_install.extras),
                    )
                missing_requested = sorted(
                    set(req_to_install.extras) - set(dist.iter_provided_extras())
                )
                for missing in missing_requested:
                    logger.warning(
                        "%s %s does not provide the extra '%s'",
                        dist.raw_name,
                        dist.version,
                        missing,
                    )

                available_requested = sorted(
                    set(dist.iter_provided_extras()) & set(req_to_install.extras)
                )
                for subreq in dist.iter_dependencies(available_requested):
                    add_req(subreq, extras_requested=available_requested)

        return more_reqs

    def get_installation_order(
        self, req_set: RequirementSet
    ) -> List[InstallRequirement]:
        """Create the installation order.

        The installation order is topological - requirements are installed
        before the requiring thing. We break cycles at an arbitrary point,
        and make no other guarantees.
        """
        # The current implementation, which we may change at any point
        # installs the user specified things in the order given, except when
        # dependencies must come earlier to achieve topological order.
        order = []
        ordered_reqs: Set[InstallRequirement] = set()

        def schedule(req: InstallRequirement) -> None:
            if req.satisfied_by or req in ordered_reqs:
                return
            if req.constraint:
                return
            ordered_reqs.add(req)
            for dep in self._discovered_dependencies[req.name]:
                schedule(dep)
            order.append(req)

        for install_req in req_set.requirements.values():
            schedule(install_req)
        return order
site-packages/pip/_internal/vcs/__pycache__/mercurial.cpython-39.pyc000064400000011434150564745010021460 0ustar00a

��ReQ�@s�ddlZddlZddlZddlmZmZddlmZmZddl	m
Z
mZddlm
Z
ddlmZddlmZmZmZmZe�e�ZGdd	�d	e�Ze�e�dS)
�N)�List�Optional)�
BadCommand�InstallationError)�
HiddenText�display_path)�make_command)�path_to_url)�
RevOptions�VersionControl�(find_path_to_project_root_from_repo_root�vcscs�eZdZdZdZdZdZeee	ed�dd��Z
eeedd	�d
d�Z
eeedd	�dd
�Zeeedd	�dd�Zeeed�dd��Zeeed�dd��Zeeed�dd��Zeeeeed�dd��Zeeeed�dd��Zeeeed��fdd��Z�ZS)�	Mercurial�hgz.hg�clone)zhg+filezhg+httpzhg+httpszhg+sshzhg+static-http)�rev�returncCs|gS)N�)rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/mercurial.py�get_base_rev_args szMercurial.get_base_rev_argsN)�dest�url�rev_optionsrcCsP|��}t�d||t|��|�tddd||��|jtdd|���|d�dS)NzCloning hg %s%s to %srz
--noupdate�-q�update��cwd)�
to_display�logger�infor�run_commandr�to_args)�selfrrr�rev_displayrrr�	fetch_new$s��zMercurial.fetch_newc	
Cs�tj�||jd�}t��}zR|�|�|�dd|j�t	|d��}|�
|�Wd�n1s`0YWn8ttjfy�}zt
�d||�WYd}~n(d}~00tdd|���}|j||d�dS)	N�hgrc�paths�default�wz/Could not switch Mercurial repository to %s: %srrr)�os�path�join�dirname�configparser�RawConfigParser�read�set�secret�open�write�OSError�NoSectionErrorr�warningrr!r )	r"rrr�repo_config�config�config_file�exc�cmd_argsrrr�switch2s
,$zMercurial.switchcCs4|jddg|d�tdd|���}|j||d�dS)N�pullrrr)r rr!)r"rrrr;rrrr@szMercurial.update)�locationrcCs4|jddgdd|d���}|�|�r,t|�}|��S)N�
showconfigz
paths.defaultFT��show_stdout�stdout_onlyr)r �strip�_is_local_repositoryr	)�clsr>rrrr�get_remote_urlEs�

zMercurial.get_remote_urlcCs|jddgdd|d���}|S)zW
        Return the repository-local changeset revision number, as an integer.
        �parentsz--template={rev}FTr@�r rC)rEr>�current_revisionrrr�get_revisionQs�
zMercurial.get_revisioncCs|jddgdd|d���}|S)zh
        Return the changeset identification hash, as a 40-character
        hexadecimal string
        rGz--template={node}FTr@rH)rEr>�current_rev_hashrrr�get_requirement_revision^s�
z"Mercurial.get_requirement_revision)r�namercCsdS)z&Always assume the versions don't matchFr)rErrMrrr�is_commit_id_equallszMercurial.is_commit_id_equalcCsD|jdgdd|d���}tj�|�s:tj�tj�||��}t||�S)z�
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        �rootFTr@)r rCr)r*�isabs�abspathr+r)rEr>�	repo_rootrrr�get_subdirectoryqs
�
zMercurial.get_subdirectorycsvt��|�}|r|Sz|jdg|ddddd�}Wn2tyPt�d|�YdStybYdS0tj�	|�
d��S)NrOFT�raise)rrArB�
on_returncode�log_failed_cmdzIcould not determine if %s is under hg control because hg is not availablez
)�super�get_repository_rootr rr�debugrr)r*�normpath�rstrip)rEr>�loc�r��	__class__rrrXs*�
�zMercurial.get_repository_root)�__name__�
__module__�__qualname__rMr,�	repo_name�schemes�staticmethod�strrrrr
r$r<r�classmethodrFrJrLr�boolrNrSrX�
__classcell__rrr^rrs*

r)r-�loggingr)�typingrr�pip._internal.exceptionsrr�pip._internal.utils.miscrr�pip._internal.utils.subprocessr�pip._internal.utils.urlsr	� pip._internal.vcs.versioncontrolr
rrr
�	getLoggerr`rr�registerrrrr�<module>s
site-packages/pip/_internal/vcs/__pycache__/bazaar.cpython-39.pyc000064400000006277150564745010020746 0ustar00a

��Re)�@s�ddlZddlmZmZmZddlmZmZddlm	Z	ddl
mZddlm
Z
mZmZmZmZe�e�ZGdd�de�Ze�e�dS)	�N)�List�Optional�Tuple)�
HiddenText�display_path)�make_command)�path_to_url)�AuthInfo�RemoteNotFoundError�
RevOptions�VersionControl�vcscs�eZdZdZdZdZdZeee	ed�dd��Z
eeedd	�d
d�Z
eeedd	�dd
�Zeeedd	�dd�Zeeeeeeefd��fdd��Zeeed�dd��Zeeed�dd��Zeeeeed�dd��Z�ZS)�Bazaar�bzrz.bzr�branch)zbzr+httpz	bzr+httpszbzr+sshzbzr+sftpzbzr+ftpzbzr+lpzbzr+file)�rev�returncCsd|gS)Nz-r�)rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/bazaar.py�get_base_rev_args szBazaar.get_base_rev_argsN)�dest�url�rev_optionsrcCs>|��}t�d||t|��tdd|��||�}|�|�dS)NzChecking out %s%s to %sr�-q)�
to_display�logger�inforr�to_args�run_command)�selfrrr�rev_display�cmd_argsrrr�	fetch_new$s�zBazaar.fetch_newcCs|jtd|�|d�dS)N�switch��cwd)rr)rrrrrrrr#/sz
Bazaar.switchcCs"tdd|���}|j||d�dS)N�pullrr$)rrr)rrrrr!rrr�update2sz
Bazaar.update)rrcs.t��|�\}}}|�d�r$d|}|||fS)Nzssh://zbzr+)�super�get_url_rev_and_auth�
startswith)�clsrr�	user_pass��	__class__rrr)6s
zBazaar.get_url_rev_and_auth)�locationrcCsz|jdgdd|d�}|��D]T}|��}dD]B}|�|�r,|�|�d}|�|�rbt|�S|Sq,qt�dS)NrFT��show_stdout�stdout_onlyr%)zcheckout of branch: zparent branch: �)r�
splitlines�stripr*�split�_is_local_repositoryrr
)r+r/�urls�line�x�reporrr�get_remote_url>s
�

zBazaar.get_remote_urlcCs |jdgdd|d�}|��dS)N�revnoFTr0���)rr4)r+r/�revisionrrr�get_revisionMs�zBazaar.get_revision)r�namercCsdS)z&Always assume the versions don't matchFr)r+rrArrr�is_commit_id_equalWszBazaar.is_commit_id_equal)�__name__�
__module__�__qualname__rA�dirname�	repo_name�schemes�staticmethod�strrrrrr"r#r'�classmethodrrr	r)r<r@�boolrB�
__classcell__rrr-rrs"
$	r)�logging�typingrrr�pip._internal.utils.miscrr�pip._internal.utils.subprocessr�pip._internal.utils.urlsr� pip._internal.vcs.versioncontrolr	r
rrr
�	getLoggerrCrr�registerrrrr�<module>s
Ksite-packages/pip/_internal/vcs/__pycache__/git.cpython-39.pyc000064400000030143150564745010020256 0ustar00a

��Re�E�@sddlZddlZddlZddlZddlZddlZddlm	Z	m
Z
mZddlm
Z
mZddlmZmZmZddlmZddlmZmZmZmZmZmZmZejjZejjZe� e!�Z"e�#d�Z$e�#d�Z%e�#d	ej&�Z'e(e)d
�dd�Z*Gd
d�de�Z+e�,e+�dS)�N)�List�Optional�Tuple)�
BadCommand�InstallationError)�
HiddenText�display_path�hide_url)�make_command)�AuthInfo�RemoteNotFoundError�RemoteNotValidError�
RevOptions�VersionControl�(find_path_to_project_root_from_repo_root�vcsz(^git version (\d+)\.(\d+)(?:\.(\d+))?.*$z^[a-fA-F0-9]{40}$a/^
    # Optional user, e.g. 'git@'
    (\w+@)?
    # Server, e.g. 'github.com'.
    ([^/:]+):
    # The server-side path. e.g. 'user/project.git'. Must start with an
    # alphanumeric character so as not to be confusable with a Windows paths
    # like 'C:/foo/bar' or 'C:\foo\bar'.
    (\w[^:]*)
    $)�sha�returncCstt�|��S�N)�bool�
HASH_REGEX�match)r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/git.py�looks_like_hash7srcseZdZdZdZdZdZdZdZe	e
ee
d�dd	��Ze
e
e
d
�dd�Zeed
fd�dd�Zee
ee
d�dd��Zee
e
eee
e
fd�dd��Zee
e
e
d�dd��Zee
eeed�dd��Zee
ee
e
d�dd��Ze
eedd�d d!�Ze
eedd�d"d#�Ze
eedd�d$d%�Zee
e
d�d&d'��Ze	e
e
d(�d)d*��Zee
e
e
d+�d,d-��Z ed;e
ee
e
d+�d.d/��Z!ee
ee
d�d0d1��Z"ee
ee
ee
e#fd(��fd2d3��Z$ee
dd�d4d5��Z%ee
ee
d��fd6d7��Z&e	e
e
d8�d9d:��Z'�Z(S)<�Git�gitz.git�clone)zgit+httpz	git+httpszgit+sshzgit+gitzgit+file)�GIT_DIR�
GIT_WORK_TREE�HEAD)�revrcCs|gSrr�r!rrr�get_base_rev_argsKszGit.get_base_rev_args)�url�destrcCsJ|�t|��\}}|jsdS|�||j�s.dSt|�||j�d�}|S)NFr)�get_url_rev_optionsr	r!�is_commit_id_equalr�get_revision_sha)�selfr$r%�_�rev_options�is_tag_or_branchrrr�is_immutable_rev_checkoutOszGit.is_immutable_rev_checkout.)rcCsF|jdgddd�}t�|�}|s0t�d|�dStdd�|��D��S)	N�versionFT)�show_stdout�stdout_onlyzCan't parse git version: %srcss|]}t|�VqdSr)�int)�.0�crrr�	<genexpr>c�z&Git.get_git_version.<locals>.<genexpr>)�run_command�GIT_VERSION_REGEXr�logger�warning�tuple�groups)r)r.rrrr�get_git_version]s
zGit.get_git_version)�locationrcCsBgd�}|j|ddd|d�}|��}|�d�r>|td�d�SdS)zl
        Return the current branch, or None if HEAD isn't at a branch
        (e.g. detached HEAD).
        )zsymbolic-ref�-qr ��FT��extra_ok_returncodesr/r0�cwdzrefs/heads/N)r6�strip�
startswith�len)�clsr=�args�output�refrrr�get_current_branches
�
zGit.get_current_branch)r%r!rc	Cs�|jd|g|dddd�}i}|���d�D]T}|�d�}|s>q*z|jdd	d
�\}}Wn tyttd|����Yn0|||<q*d|��}d
|��}	|�|�}
|
dur�|
dfS|�|	�}
|
dfS)z�
        Return (sha_or_none, is_branch), where sha_or_none is a commit hash
        if the revision names a remote branch or tag, otherwise None.

        Args:
          dest: the repository directory.
          rev: the revision name.
        zshow-refFT�ignore)rCr/r0�
on_returncode�
�
� �)�maxsplitzunexpected show-ref line: zrefs/remotes/origin/z
refs/tags/N)r6rD�split�rstrip�
ValueError�get)rGr%r!rI�refs�line�ref_sha�ref_name�
branch_ref�tag_refrrrrr(~s0�





zGit.get_revision_shacCs.|�d�rdSt|�sdS|�||�r*dSdS)a$
        Return true if rev is a ref or is a commit that we don't have locally.

        Branches and tags are not considered in this method because they are
        assumed to be always available locally (which is a normal outcome of
        ``git clone`` and ``git fetch --tags``).
        zrefs/TF)rEr�
has_commit)rGr%r!rrr�
_should_fetch�s	
zGit._should_fetch)r%r$r+rcCs�|j}|dusJ�|�||�\}}|durF|�|�}|r<|nd|_|St|�sZt�d|�|�||�sj|S|jt	dd||�
��|d�|j|dd�}|�|�}|S)z�
        Resolve a revision to a new RevOptions object with the SHA1 of the
        branch, tag, or ref if found.

        Args:
          rev_options: a RevOptions object.
        Nz:Did not find branch or tag '%s', assuming revision or ref.�fetchr>�rC�
FETCH_HEADr")�arg_revr(�make_new�branch_namerr8r9r^r6r
�to_args�get_revision)rGr%r$r+r!r�	is_branchrrr�resolve_revision�s*
��
zGit.resolve_revision)r%�namercCs|sdS|�|�|kS)z�
        Return whether the current commit hash equals the given name.

        Args:
          dest: the repository directory.
          name: a string name.
        F)rf)rGr%rirrrr'�s	zGit.is_commit_id_equalNc	Cs |��}t�d||t|��|��dkr@|�tddd||��n|�tdd||��|jr�|�|||�}t	|dd�}t�
d||�|dur�|�||j�s�tdd|���}|j||d	�n4|�
|�|kr�d
|��}dd|d|g}|j||d	�n|�|�}|�|�}t�d
||j�|�|�dS)NzCloning %s%s to %s)rQ�rz--filter=blob:noner>rdzRev options %s, branch_name %s�checkoutr`zorigin/z-bz--trackzResolved %s to commit %s)�
to_displayr8�inforr<r6r
r!rh�getattr�debugr'rerKrfrc�update_submodules)	r)r%r$r+�rev_displayrd�cmd_args�track_branchrrrr�	fetch_newsL��
�
�

z
Git.fetch_newcCsB|jtdd|�|d�tdd|���}|j||d�|�|�dS)N�configzremote.origin.urlr`rkr>)r6r
rerp�r)r%r$r+rrrrr�switch7s
�z
Git.switchcCsn|��dkr |jgd�|d�n|jddg|d�|�|||�}tddd|���}|j||d�|�|�dS)N)r@�	)r_r>z--tagsr`r_r>�resetz--hard)r<r6rhr
rerprvrrr�updateAsz
Git.updatecCs||jgd�ddd|d�}|��}z|d}WntyBt�Yn0|D]}|�d�rH|}q`qH|�d�d	}|�|���S)
z�
        Return URL of the first remote encountered.

        Raises RemoteNotFoundError if the repository does not have a remote
        url configured.
        )ruz--get-regexpzremote\..*\.urlr?FTrArzremote.origin.url rPr@)r6�
splitlines�
IndexErrorrrErS�_git_remote_to_pip_urlrD)rGr=�stdout�remotes�found_remote�remoter$rrr�get_remote_urlOs$
�

zGit.get_remote_url)r$rcCsNt�d|�r|Stj�|�r*t�|���St�|�}|rB|�	d�St
|��dS)a8
        Convert a remote url from what git uses to what pip accepts.

        There are 3 legal forms **url** may take:

            1. A fully qualified url: ssh://git@example.com/foo/bar.git
            2. A local project.git folder: /path/to/bare/repository.git
            3. SCP shorthand for form 1: git@example.com:foo/bar.git

        Form 1 is output as-is. Form 2 must be converted to URI and form 3 must
        be converted to form 1.

        See the corresponding test test_git_remote_url_to_pip() for examples of
        sample inputs/outputs.
        z\w+://z
ssh://\1\2/\3N)�rer�os�path�exists�pathlib�PurePath�as_uri�	SCP_REGEX�expandr
)r$�	scp_matchrrrr}ms

zGit._git_remote_to_pip_url)r=r!rcCs>z |jdddd|g|dd�Wnty4YdS0dSdS)	zU
        Check if rev is a commit that is available in the local repository.
        �	rev-parser>z--verifyzsha^F)rC�log_failed_cmdTN)r6r)rGr=r!rrrr]�s�
zGit.has_commitcCs*|durd}|jd|gdd|d�}|��S)Nr r�FT�r/r0rC)r6rD)rGr=r!�current_revrrrrf�s�zGit.get_revisioncCsT|jddgdd|d���}tj�|�s4tj�||�}tj�tj�|d��}t||�S)z�
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        r�z	--git-dirFTr�z..)r6rDr�r��isabs�join�abspathr)rGr=�git_dir�	repo_rootrrr�get_subdirectory�s�
zGit.get_subdirectorycs�t|�\}}}}}|�d�r�|dt|�d���}|tj�|��dd��d�}|�d�d}	|d|	�t	||	d�||||f�}d|vr�d|vs�J�|�d	d
�}t
��|�\}}
}|�dd�}nt
��|�\}}
}||
|fS)
a9
        Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'.
        That's required because although they use SSH they sometimes don't
        work with a ssh:// scheme (e.g. GitHub). But we need a scheme for
        parsing. Hence we remove it again afterwards and return it as a stub.
        �fileN�/�\�+r@z://zfile:zgit+z
git+ssh://zssh://�)�urlsplit�endswithrF�lstrip�urllib�request�url2pathname�replace�find�
urlunsplit�super�get_url_rev_and_auth)rGr$�scheme�netlocr��query�fragment�initial_slashes�newpath�
after_plusr!�	user_pass��	__class__rrr��s&

���zGit.get_url_rev_and_authcCs0tj�tj�|d��sdS|jgd�|d�dS)Nz.gitmodules)�	submodulerzz--initz--recursiver>r`)r�r�r�r�r6)rGr=rrrrp�s�zGit.update_submodulescsxt��|�}|r|Sz|jddg|ddddd�}Wn2tyRt�d|�YdStydYdS0tj�	|�
d��S)	Nr�z--show-toplevelFT�raise)rCr/r0rMr�zKcould not determine if %s is under git control because git is not availablez
)r��get_repository_rootr6rr8rorr�r��normpathrT)rGr=�loc�rr�rrr��s*�
�zGit.get_repository_root)�repo_urlrcCsdS)zEIn either https or ssh form, requirements must be prefixed with git+.Tr)r�rrr�should_add_vcs_url_prefix�szGit.should_add_vcs_url_prefix)N))�__name__�
__module__�__qualname__ri�dirname�	repo_name�schemes�
unset_environ�default_arg_rev�staticmethod�strrr#rr-rr1r<�classmethodrrKr(r^rrrhr'rtrwrzr�r}r]rfr�rr�rpr�r��
__classcell__rrr�rr;sV	 -�-7
$r)-�logging�os.pathr�r�r��urllib.parser��urllib.request�typingrrr�pip._internal.exceptionsrr�pip._internal.utils.miscrrr	�pip._internal.utils.subprocessr
� pip._internal.vcs.versioncontrolrrr
rrrr�parser�r��	getLoggerr�r8�compiler7r�VERBOSEr�r�rrr�registerrrrr�<module>s6$

�

�Isite-packages/pip/_internal/vcs/__pycache__/versioncontrol.cpython-39.pyc000064400000050675150564745010022575 0ustar00a

��Re�W�@s�dZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
mZmZm
Z
mZmZmZmZmZddlmZddlmZmZddlmZmZmZmZmZmZmZm Z ddl!m"Z"m#Z#m$Z$ddl%m&Z&er�dd	lm'Z'd
gZ(e�)e*�Z+eee,ee,fZ-e,e.d�dd
�Z/de,e,e,ee,e,d�dd�Z0e,e,ee,d�dd�Z1Gdd�de2�Z3Gdd�de2�Z4Gdd�d�Z5Gdd�d�Z6e6�Z7Gdd�d�Z8dS)z)Handles all VCS (version control) support�N)�
TYPE_CHECKING�Any�Dict�Iterable�Iterator�List�Mapping�Optional�Tuple�Type�Union)�SpinnerInterface)�
BadCommand�InstallationError)�
HiddenText�ask_path_exists�
backup_dir�display_path�hide_url�
hide_value�is_installable_dir�rmtree)�CommandArgs�call_subprocess�make_command)�get_url_scheme)�Literal�vcs��name�returncCs&t|�}|durdS|gd�tjvS)z3
    Return true if the name looks like a URL.
    NF)�http�https�file�ftp)rr�all_schemes)r�scheme�r'��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/versioncontrol.py�is_url4sr))�repo_url�rev�project_name�subdirr cCs6|�dd�}|�d|�d|��}|r2|d|��7}|S)z�
    Return the URL for a VCS requirement.

    Args:
      repo_url: the remote VCS url, with any needed VCS prefix (e.g. "git+").
      project_name: the (unescaped) project name.
    �-�_�@z#egg=z&subdirectory=)�replace)r*r+r,r-�egg_project_name�reqr'r'r(�make_vcs_requirement_url>s

r4)�location�	repo_rootr cCsV|}t|�s6|}tj�|�}||krt�d|�dSqtj�||�rHdStj�||�S)z�
    Find the the Python project's root by searching up the filesystem from
    `location`. Return the path to project root relative to `repo_root`.
    Return None if the project root is `repo_root`, or cannot be found.
    zOCould not find a Python project for directory %s (tried all parent directories)N)r�os�path�dirname�logger�warning�samefile�relpath)r5r6�
orig_location�
last_locationr'r'r(�(find_path_to_project_root_from_repo_rootPs	�r@c@seZdZdS)�RemoteNotFoundErrorN)�__name__�
__module__�__qualname__r'r'r'r(rAmsrAcs"eZdZed��fdd�Z�ZS)�RemoteNotValidError��urlcst��|�||_dS�N)�super�__init__rG)�selfrG��	__class__r'r(rJrszRemoteNotValidError.__init__)rBrCrD�strrJ�
__classcell__r'r'rLr(rEqsrEc@s�eZdZdZdedeeeedd�dd�Zed�dd	�Z	e
eed�d
d��Zed�dd
�Zed�dd�Z
edd�dd�ZdS)�
RevOptionsz�
    Encapsulates a VCS-specific revision to install, along with any VCS
    install options.

    Instances of this class should be treated as if immutable.
    N�VersionControl)�vc_classr+�
extra_argsr cCs(|durg}||_||_||_d|_dS)z�
        Args:
          vc_class: a VersionControl subclass.
          rev: the name of the revision to install.
          extra_args: a list of extra options.
        N)rSr+rR�branch_name)rKrRr+rSr'r'r(rJ�szRevOptions.__init__�r cCsd|jj�d|j�d�S)Nz<RevOptions z: rev=�>)rRrr+�rKr'r'r(�__repr__�szRevOptions.__repr__cCs|jdur|jjS|jSrH)r+rR�default_arg_revrWr'r'r(�arg_rev�s
zRevOptions.arg_revcCs0g}|j}|dur"||j�|�7}||j7}|S)z<
        Return the VCS-specific command arguments.
        N)rZrR�get_base_rev_argsrS)rK�argsr+r'r'r(�to_args�s
zRevOptions.to_argscCs|js
dSd|j�d�S)N�z (to revision �)�r+rWr'r'r(�
to_display�szRevOptions.to_display�r+r cCs|jj||jd�S)z�
        Make a copy of the current instance, but with a new rev.

        Args:
          rev: the name of the revision for the new object.
        �rS)rR�make_rev_optionsrS)rKr+r'r'r(�make_new�szRevOptions.make_new)NN)rBrCrD�__doc__rr	rNrrJrX�propertyrZr]rarer'r'r'r(rPws
��rPcs�eZdZUiZeedfed<gd�Zdd��fdd�Ze	ed�dd	�Z
eedd�d
d��Z
eeed�dd
��Zeeed�dd��Zeddd�dd�Zedd�dd�Zeedd�dd�Zeedd�dd�Zeedd�dd�Z�ZS)�
VcsSupportrQ�	_registry)�ssh�git�hg�bzr�sftp�svnNrUcstjj�|j�t���dSrH)�urllib�parse�uses_netloc�extend�schemesrIrJrWrLr'r(rJ�szVcsSupport.__init__cCs
|j��SrH)ri�__iter__rWr'r'r(ru�szVcsSupport.__iter__cCst|j���SrH)�listri�valuesrWr'r'r(�backends�szVcsSupport.backendscCsdd�|jD�S)NcSsg|]
}|j�qSr')r9)�.0�backendr'r'r(�
<listcomp>��z'VcsSupport.dirnames.<locals>.<listcomp>)rxrWr'r'r(�dirnames�szVcsSupport.dirnamescCs g}|jD]}|�|j�q
|SrH)rxrsrt)rKrtrzr'r'r(r%�s
zVcsSupport.all_schemes)�clsr cCsHt|d�st�d|j�dS|j|jvrD|�|j|j<t�d|j�dS)NrzCannot register VCS %szRegistered VCS backend: %s)�hasattrr:r;rBrri�debug)rKr~r'r'r(�register�s
zVcsSupport.registerrcCs||jvr|j|=dSrH)ri�rKrr'r'r(�
unregister�s
zVcsSupport.unregister�r5r cCsXi}|j��D],}|�|�}|s"qt�d||j�|||<q|sDdSt|td�}||S)zv
        Return a VersionControl object if a repository of that type is found
        at the given directory.
        zDetermine that %s uses VCS: %sN)�key)rirw�get_repository_rootr:r�r�max�len)rKr5�vcs_backends�vcs_backend�	repo_path�inner_most_repo_pathr'r'r(�get_backend_for_dir�s

zVcsSupport.get_backend_for_dir)r&r cCs&|j��D]}||jvr
|Sq
dS)�9
        Return a VersionControl object or None.
        N)rirwrt)rKr&r�r'r'r(�get_backend_for_scheme�s

z!VcsSupport.get_backend_for_schemecCs|��}|j�|�S)r�)�lowerri�getr�r'r'r(�get_backendszVcsSupport.get_backend)rBrCrDrirrN�__annotations__rtrJrrurgrrxr}r%rr�r�r	r�r�r�rOr'r'rLr(rh�s
	rhc@s�eZdZUdZdZdZdZeedfe	d<dZ
eedfe	d<dZeee	d<e
eed�d	d
��Ze
eeed�dd
��Ze
eed�dd��Ze
eeed�dd��Zeeeed�dd��Zeeed�dd�Ze
dPeeeeed�dd��Ze
eed�dd��Ze
eeeeeeeeeffd �d!d"��Ze
eeeeeefd#�d$d%��Zeeeeeed&�d'd(��Zeeeefd#�d)d*�Z eeed#�d+d,��Z!e
eeed-�d.d/��Z"eeedd0�d1d2�Z#eeedd0�d3d4�Z$eeedd0�d5d6�Z%e
eeeed7�d8d9��Z&eedd:�d;d<�Z'eedd=�d>d?�Z(e
eed�d@dA��Z)e
eed�dBdC��Z*e
dQe+eeefeeedGee,e-eeee.ee/fee0eeedH�dIdJ��Z1e
eedK�dLdM��Z2e
eeed�dNdO��Z3dS)RrQr^r'.rt�
unset_environNrY)�
remote_urlr cCs|���|j�d��S)z�
        Return whether the vcs prefix (e.g. "git+") should be added to a
        repository's remote url when used in a requirement.
        �:)r��
startswithr)r~r�r'r'r(�should_add_vcs_url_prefixsz(VersionControl.should_add_vcs_url_prefixr�cCsdS)z�
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        Nr'�r~r5r'r'r(�get_subdirectoryszVersionControl.get_subdirectory)�repo_dirr cCs
|�|�S)zR
        Return the revision string that should be used in a requirement.
        )�get_revision)r~r�r'r'r(�get_requirement_revision'sz'VersionControl.get_requirement_revision)r�r,r cCsL|�|�}|�|�r$|j�d|��}|�|�}|�|�}t||||d�}|S)aC
        Return the requirement string to use to redownload the files
        currently at the given repository directory.

        Args:
          project_name: the (unescaped) project name.

        The return value has a form similar to the following:

            {repository_url}@{revision}#egg={project_name}
        �+)r-)�get_remote_urlr�rr�r�r4)r~r�r,r*�revisionr-r3r'r'r(�get_src_requirement.s



z"VersionControl.get_src_requirementrbcCst�dS)z�
        Return the base revision arguments for a vcs command.

        Args:
          rev: the name of a revision to install.  Cannot be None.
        N��NotImplementedErrorr`r'r'r(r[Fsz VersionControl.get_base_rev_args)rG�destr cCsdS)aZ
        Return true if the commit hash checked out at dest matches
        the revision in url.

        Always return False, if the VCS does not support immutable commit
        hashes.

        This method does not check if there are local uncommitted changes
        in dest after checkout, as pip currently has no use case for that.
        Fr')rKrGr�r'r'r(�is_immutable_rev_checkoutPsz(VersionControl.is_immutable_rev_checkout)r+rSr cCst|||d�S)z�
        Return a RevOptions object.

        Args:
          rev: the name of a revision to install.
          extra_args: a list of extra options.
        rc)rP)r~r+rSr'r'r(rd]szVersionControl.make_rev_options)�repor cCs&tj�|�\}}|�tjj�p$t|�S)zs
        posix absolute paths start with os.path.sep,
        win32 ones start with drive (like c:\folder)
        )r7r8�
splitdriver��sep�bool)r~r��drive�tailr'r'r(�_is_local_repositoryjsz#VersionControl._is_local_repository)�netlocr&r cCs|dfS)aZ
        Parse the repository URL's netloc, and return the new netloc to use
        along with auth information.

        Args:
          netloc: the original repository URL netloc.
          scheme: the repository URL's scheme without the vcs prefix.

        This is mainly for the Subversion class to override, so that auth
        information can be provided via the --username and --password options
        instead of through the URL.  For other subclasses like Git without
        such an option, auth information must stay in the URL.

        Returns: (netloc, (username, password)).
        )NNr')r~r�r&r'r'r(�get_netloc_and_authssz"VersionControl.get_netloc_and_auth)rGr c	Cs�tj�|�\}}}}}d|vr,td�|���|�dd�d}|�||�\}}d}d|vrz|�dd�\}}|sztd�|���tj�	||||df�}|||fS)z�
        Parse the repository URL to use, and return the URL, revision,
        and auth info to use.

        Returns: (url, rev, (username, password)).
        r�zvSorry, {!r} is a malformed VCS url. The format is <vcs>+<protocol>://<url>, e.g. svn+http://myrepo/svn/MyApp#egg=MyApp�Nr0zyThe URL {!r} has an empty revision (after @) which is not supported. Include a revision after @ or remove @ from the URL.r^)
rprq�urlsplit�
ValueError�format�splitr��rsplitr�
urlunsplit)	r~rGr&r�r8�query�frag�	user_passr+r'r'r(�get_url_rev_and_auth�s(����z#VersionControl.get_url_rev_and_auth)�username�passwordr cCsgS)zM
        Return the RevOptions "extra arguments" to use in obtain().
        r')r�r�r'r'r(�
make_rev_args�szVersionControl.make_rev_argsc
CsT|�|j�\}}}|\}}d}|dur.t|�}|�||�}|j||d�}	t|�|	fS)zq
        Return the URL and RevOptions object to use in obtain(),
        as a tuple (url, rev_options).
        Nrc)r��secretrr�rdr)
rKrG�
secret_urlr+r�r��secret_passwordr�rS�rev_optionsr'r'r(�get_url_rev_options�sz"VersionControl.get_url_rev_optionscCstj�|��d�S)zi
        Normalize a URL for comparison by unquoting it and removing any
        trailing slash.
        �/)rprq�unquote�rstriprFr'r'r(�
normalize_url�szVersionControl.normalize_url)�url1�url2r cCs|�|�|�|�kS)zV
        Compare two repo URLs for identity, ignoring incidental differences.
        )r�)r~r�r�r'r'r(�compare_urls�szVersionControl.compare_urls)r�rGr�r cCst�dS)z�
        Fetch a revision from a repository, in the case that this is the
        first fetch from the repository.

        Args:
          dest: the directory to fetch the repository to.
          rev_options: a RevOptions object.
        Nr��rKr�rGr�r'r'r(�	fetch_new�s	zVersionControl.fetch_newcCst�dS)z}
        Switch the repo at ``dest`` to point to ``URL``.

        Args:
          rev_options: a RevOptions object.
        Nr�r�r'r'r(�switch�szVersionControl.switchcCst�dS)z�
        Update an already-existing repo to the given ``rev_options``.

        Args:
          rev_options: a RevOptions object.
        Nr�r�r'r'r(�update�szVersionControl.update)r�rr cCst�dS)z�
        Return whether the id of the current commit equals the given name.

        Args:
          dest: the repository directory.
          name: a string name.
        Nr�)r~r�rr'r'r(�is_commit_id_equal�s	z!VersionControl.is_commit_id_equal)r�rGr c	Cs�|�|�\}}tj�|�s,|�|||�dS|��}|�|�r�|�|�}|�||j	�r�t
�d|j�
�t|�|�|�||j�s�t
�dt|�|j|�|�|||�n
t
�d�dSt
�d|j|jt|�|�d}nt
�d||j|j�d}t
�d	|j|�td
�|d�|d�}|d
k�r&t�d�|dk�rZt
�dt|��t|�|�|||�dS|dk�r�t|�}t
�dt|�|�t�||�|�|||�dS|dk�r�t
�d|jt|�||�|�|||�dS)a
        Install or update in editable mode the package represented by this
        VersionControl object.

        :param dest: the repository directory in which to install or update.
        :param url: the repository URL starting with a vcs prefix.
        Nz)%s in %s exists, and has correct URL (%s)zUpdating %s %s%sz$Skipping because already up-to-date.z%s %s in %s exists with URL %s)z%(s)witch, (i)gnore, (w)ipe, (b)ackup )�s�i�w�bz0Directory %s already exists, and is not a %s %s.)z(i)gnore, (w)ipe, (b)ackup )r�r�r�z+The plan is to install the %s repository %szWhat to do?  {}rr��a���r�zDeleting %sr�zBacking up %s to %sr�zSwitching %s %s to %s%s)r�r7r8�existsr�ra�is_repository_directoryr�r�r�r:r��	repo_name�titlerr�r+�infor�r;rrr��sys�exitrr�shutil�mover�)	rKr�rGr��rev_display�existing_url�prompt�response�dest_dirr'r'r(�obtain�s�

��
���




�zVersionControl.obtain)r5rGr cCs&tj�|�rt|�|j||d�dS)z�
        Clean up current location and download the url repository
        (and vcs infos) into location

        :param url: the repository URL starting with a vcs prefix.
        rFN)r7r8r�rr�)rKr5rGr'r'r(�unpackMszVersionControl.unpackcCst�dS)z�
        Return the url used at location

        Raises RemoteNotFoundError if the repository does not have a remote
        url configured.
        Nr�r�r'r'r(r�XszVersionControl.get_remote_urlcCst�dS)zR
        Return the current commit id of the files at the given location.
        Nr�r�r'r'r(r�bszVersionControl.get_revisionT�raiseFz"Literal["raise", "warn", "ignore"])�cmd�show_stdout�cwd�
on_returncode�extra_ok_returncodes�command_desc�
extra_environ�spinner�log_failed_cmd�stdout_onlyr c
Cs�t|jg|�R�}z"t||||||||j||	|
d�WSty`td|j�d|j�d���Yn$ty�td|j�d���Yn0dS)z�
        Run a VCS subcommand
        This is simply a wrapper around call_subprocess that adds the VCS
        command name, and checks that the VCS is available
        )r�r�r�r�r�r�r�r�zCannot find command z - do you have z installed and in your PATH?zNo permission to execute z� - install it locally, globally (ask admin), or check your PATH. See possible solutions at https://pip.pypa.io/en/latest/reference/pip_freeze/#fixing-permission-denied.N)rrrr��FileNotFoundErrorr�PermissionError)r~r�r�r�r�r�r�r�r�r�r�r'r'r(�run_commandis2�

���zVersionControl.run_command)r8r cCs,t�d||j|j�tj�tj�||j��S)zL
        Return whether a directory path is a repository directory.
        zChecking in %s for %s (%s)...)r:r�r9rr7r8r��join)r~r8r'r'r(r��sz&VersionControl.is_repository_directorycCs|�|�r|SdS)ay
        Return the "root" (top-level) directory controlled by the vcs,
        or `None` if the directory is not in any.

        It is meant to be overridden to implement smarter detection
        mechanisms for specific vcs.

        This can do more than is_repository_directory() alone. For
        example, the Git override checks that Git is actually available.
        N)r�r�r'r'r(r��s
z"VersionControl.get_repository_root)NN)	TNr�NNNNTF)4rBrCrDrr9r�rtr
rNr�r�rYr	�classmethodr�r�r�r�r��staticmethodrr[r�rrPrdr�r��AuthInfor�rr�r�r�r�r�r�r�r�r�r�r�r�rr�intrrr
r�r�r�r'r'r'r(rQ
s�
	
��� �		
X	�
�5rQ)N)9rf�loggingr7r�r��urllib.parserp�typingrrrrrrrr	r
rr�pip._internal.cli.spinnersr
�pip._internal.exceptionsrr�pip._internal.utils.miscrrrrrrrr�pip._internal.utils.subprocessrrr�pip._internal.utils.urlsrr�__all__�	getLoggerrBr:rNr�r�r)r4r@�	ExceptionrArErPrhrrQr'r'r'r(�<module>s>4(

���CPsite-packages/pip/_internal/vcs/__pycache__/subversion.cpython-39.pyc000064400000020351150564745010021672 0ustar00a

��ReL-�@s�ddlZddlZddlZddlmZmZmZddlmZm	Z	m
Z
mZmZddl
mZmZddlmZmZmZmZmZe�e�Ze�d�Ze�d�Ze�d�Ze�d	�ZGd
d�de�Ze�e�dS)�N)�List�Optional�Tuple)�
HiddenText�display_path�is_console_interactive�is_installable_dir�split_auth_from_netloc)�CommandArgs�make_command)�AuthInfo�RemoteNotFoundError�
RevOptions�VersionControl�vcsz
url="([^"]+)"zcommitted-rev="(\d+)"z\s*revision="(\d+)"z<url>(.*)</url>c	s�eZdZdZdZdZdZeee	d�dd��Z
eeeed�d	d
��Z
eeed�dd
��Zeeeeeeeeeeffd��fdd��Zeeeeeeefd��fdd��Zeeeeeed�dd��Zeeed�dd��Zeeeeeefd�dd��Zeeeee	d�dd��Zd1e	dd��fd d!�
Zeed"fd#�d$d%�Zeed"fd#�d&d'�Zed#�d(d)�Zeeedd*�d+d,�Z eeedd*�d-d.�Z!eeedd*�d/d0�Z"�Z#S)2�
Subversion�svnz.svn�checkout)zsvn+sshzsvn+httpz	svn+httpszsvn+svnzsvn+file)�
remote_url�returncCsdS)NT�)�clsrrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/subversion.py�should_add_vcs_url_prefix$sz$Subversion.should_add_vcs_url_prefix)�revrcCsd|gS)Nz-rr)rrrr�get_base_rev_args(szSubversion.get_base_rev_args)�locationrc	Cs�d}t�|�D]�\}}}|j|vr0g|dd�<q|�|j�tj�||jd�}tj�|�s\q|�|�\}}||kr�|dus~J�|d}n|r�|�|�s�g|dd�<qt	||�}qt
|�S)zR
        Return the maximum revision for all files under a given location
        rN�entries�/)�os�walk�dirname�remove�path�join�exists�_get_svn_url_rev�
startswith�max�str)	rr�revision�base�dirs�_�
entries_fn�dirurl�localrevrrr�get_revision,s$

zSubversion.get_revision)�netloc�schemercs|dkrt��||�St|�S)z�
        This override allows the auth information to be passed to svn via the
        --username and --password options instead of via the URL.
        �ssh)�super�get_netloc_and_authr	)rr2r3��	__class__rrr6IszSubversion.get_netloc_and_auth)�urlrcs.t��|�\}}}|�d�r$d|}|||fS)Nzssh://zsvn+)r5�get_url_rev_and_authr')rr9r�	user_passr7rrr:Xs
zSubversion.get_url_rev_and_auth)�username�passwordrcCs(g}|r|d|g7}|r$|d|g7}|S)Nz
--usernamez
--passwordr)r<r=�
extra_argsrrr�
make_rev_args`szSubversion.make_rev_argscCsT|}t|�s6|}tj�|�}||krt�d|�t�q|�|�\}}|durPt�|S)NzMCould not find Python project for directory %s (tried all parent directories))rrr#r!�logger�warningr
r&)rr�
orig_location�
last_locationr9�_revrrr�get_remote_urlls�zSubversion.get_remote_urlcCs�ddlm}tj�||jd�}tj�|�r\t|��}|��}Wd�q`1sP0Ynd}d}|�	d�s�|�	d�s�|�	d�r�t
ttj
|�d���}|dd=|dd	}d
d�|D�dg}n�|�	d��rt�|�}	|	s�td
|����|	�d�}dd�t�|�D�dg}npzP|jdd|gddd�}
t�|
�}	|	du�sBJ�|	�d�}dd�t�|
�D�}Wn|�y�dg}}Yn0|�r�t|�}nd}||fS)Nr)�InstallationErrorr��8�9�10z

�cSs,g|]$}t|�dkr|drt|d��qS)�	)�len�int)�.0�drrr�
<listcomp>��z/Subversion._get_svn_url_rev.<locals>.<listcomp>z<?xmlzBadly formatted data: �cSsg|]}t|�d���qS�rS�rN�group�rO�mrrrrQ�rR�infoz--xmlFT��show_stdout�stdout_onlycSsg|]}t|�d���qSrTrUrWrrrrQ�rR)�pip._internal.exceptionsrFrr#r$r!r%�open�readr'�list�mapr)�
splitlines�split�_svn_xml_url_re�search�
ValueErrorrV�_svn_rev_re�finditer�run_command�_svn_info_xml_url_re�_svn_info_xml_rev_rer()rrrF�entries_path�f�datar9r�revs�match�xmlrrrrr&�sD
(


�


zSubversion._get_svn_url_rev)�dest�namercCsdS)z&Always assume the versions don't matchFr)rrrrsrrr�is_commit_id_equal�szSubversion.is_commit_id_equalN)�use_interactivercs(|durt�}||_d|_t���dS)N)rru�_vcs_versionr5�__init__)�selfrur7rrrw�s
zSubversion.__init__.)rcCs|d}|jdgddd�}|�|�s$dS|t|�d���d}|�d	�d�d
�}zttt|��}WntyvYdS0|S)aQuery the version of the currently installed Subversion client.

        :return: A tuple containing the parts of the version information or
            ``()`` if the version returned from ``svn`` could not be parsed.
        :raises: BadCommand: If ``svn`` is not installed.
        z
svn, version z	--versionFTrZrNr�-�.)	rir'rMrc�	partition�tuplerarNrf)rx�version_prefix�version�version_list�parsed_versionrrr�call_vcs_version�s
zSubversion.call_vcs_versioncCs"|jdur|jS|��}||_|S)a�Return the version of the currently installed Subversion client.

        If the version of the Subversion client has already been queried,
        a cached value will be used.

        :return: A tuple containing the parts of the version information or
            ``()`` if the version returned from ``svn`` could not be parsed.
        :raises: BadCommand: If ``svn`` is not installed.
        N)rvr�)rx�vcs_versionrrr�get_vcs_version�s


zSubversion.get_vcs_versioncCs&|jsdgS|��}|dkr"dgSgS)aCReturn options to be used on calls to Subversion that contact the server.

        These options are applicable for the following ``svn`` subcommands used
        in this class.

            - checkout
            - switch
            - update

        :return: A list of command line arguments to pass to ``svn``.
        z--non-interactive)rS�z--force-interactive)rur�)rx�svn_versionrrr�get_remote_call_options�s	z"Subversion.get_remote_call_options)rrr9�rev_optionsrcCsD|��}t�d||t|��tdd|��|��||�}|�|�dS)NzChecking out %s%s to %srz-q)�
to_displayr@rYrrr��to_argsri)rxrrr9r��rev_display�cmd_argsrrr�	fetch_news ��zSubversion.fetch_newcCs&td|��|��||�}|�|�dS)N�switch�rr�r�ri�rxrrr9r�r�rrrr�*s�zSubversion.switchcCs$td|��|��|�}|�|�dS)N�updater�r�rrrr�4s�zSubversion.update)N)$�__name__�
__module__�__qualname__rsr!�	repo_name�schemes�classmethodr)�boolr�staticmethodrrr1rrr6rr:rr
r?rErNr&rtrwr�r�r�rr�r�r��
__classcell__rrr7rrsB�$�1
r) �loggingr�re�typingrrr�pip._internal.utils.miscrrrrr	�pip._internal.utils.subprocessr
r� pip._internal.vcs.versioncontrolrr
rrr�	getLoggerr�r@�compilerdrgrkrjr�registerrrrr�<module>s




"site-packages/pip/_internal/vcs/__pycache__/__init__.cpython-39.pyc000064400000001051150564745010021226 0ustar00a

��ReT�@s@ddlZddlZddlZddlZddlmZmZmZm	Z	m
Z
dS)�N)�RemoteNotFoundError�RemoteNotValidError�is_url�make_vcs_requirement_url�vcs)�pip._internal.vcs.bazaar�pip�pip._internal.vcs.git�pip._internal.vcs.mercurial�pip._internal.vcs.subversion� pip._internal.vcs.versioncontrolrrrrr�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/vcs/__init__.py�<module>ssite-packages/pip/_internal/vcs/versioncontrol.py000064400000053616150564745010016304 0ustar00"""Handles all VCS (version control) support"""

import logging
import os
import shutil
import sys
import urllib.parse
from typing import (
    TYPE_CHECKING,
    Any,
    Dict,
    Iterable,
    Iterator,
    List,
    Mapping,
    Optional,
    Tuple,
    Type,
    Union,
)

from pip._internal.cli.spinners import SpinnerInterface
from pip._internal.exceptions import BadCommand, InstallationError
from pip._internal.utils.misc import (
    HiddenText,
    ask_path_exists,
    backup_dir,
    display_path,
    hide_url,
    hide_value,
    is_installable_dir,
    rmtree,
)
from pip._internal.utils.subprocess import CommandArgs, call_subprocess, make_command
from pip._internal.utils.urls import get_url_scheme

if TYPE_CHECKING:
    # Literal was introduced in Python 3.8.
    #
    # TODO: Remove `if TYPE_CHECKING` when dropping support for Python 3.7.
    from typing import Literal


__all__ = ["vcs"]


logger = logging.getLogger(__name__)

AuthInfo = Tuple[Optional[str], Optional[str]]


def is_url(name: str) -> bool:
    """
    Return true if the name looks like a URL.
    """
    scheme = get_url_scheme(name)
    if scheme is None:
        return False
    return scheme in ["http", "https", "file", "ftp"] + vcs.all_schemes


def make_vcs_requirement_url(
    repo_url: str, rev: str, project_name: str, subdir: Optional[str] = None
) -> str:
    """
    Return the URL for a VCS requirement.

    Args:
      repo_url: the remote VCS url, with any needed VCS prefix (e.g. "git+").
      project_name: the (unescaped) project name.
    """
    egg_project_name = project_name.replace("-", "_")
    req = f"{repo_url}@{rev}#egg={egg_project_name}"
    if subdir:
        req += f"&subdirectory={subdir}"

    return req


def find_path_to_project_root_from_repo_root(
    location: str, repo_root: str
) -> Optional[str]:
    """
    Find the the Python project's root by searching up the filesystem from
    `location`. Return the path to project root relative to `repo_root`.
    Return None if the project root is `repo_root`, or cannot be found.
    """
    # find project root.
    orig_location = location
    while not is_installable_dir(location):
        last_location = location
        location = os.path.dirname(location)
        if location == last_location:
            # We've traversed up to the root of the filesystem without
            # finding a Python project.
            logger.warning(
                "Could not find a Python project for directory %s (tried all "
                "parent directories)",
                orig_location,
            )
            return None

    if os.path.samefile(repo_root, location):
        return None

    return os.path.relpath(location, repo_root)


class RemoteNotFoundError(Exception):
    pass


class RemoteNotValidError(Exception):
    def __init__(self, url: str):
        super().__init__(url)
        self.url = url


class RevOptions:

    """
    Encapsulates a VCS-specific revision to install, along with any VCS
    install options.

    Instances of this class should be treated as if immutable.
    """

    def __init__(
        self,
        vc_class: Type["VersionControl"],
        rev: Optional[str] = None,
        extra_args: Optional[CommandArgs] = None,
    ) -> None:
        """
        Args:
          vc_class: a VersionControl subclass.
          rev: the name of the revision to install.
          extra_args: a list of extra options.
        """
        if extra_args is None:
            extra_args = []

        self.extra_args = extra_args
        self.rev = rev
        self.vc_class = vc_class
        self.branch_name: Optional[str] = None

    def __repr__(self) -> str:
        return f"<RevOptions {self.vc_class.name}: rev={self.rev!r}>"

    @property
    def arg_rev(self) -> Optional[str]:
        if self.rev is None:
            return self.vc_class.default_arg_rev

        return self.rev

    def to_args(self) -> CommandArgs:
        """
        Return the VCS-specific command arguments.
        """
        args: CommandArgs = []
        rev = self.arg_rev
        if rev is not None:
            args += self.vc_class.get_base_rev_args(rev)
        args += self.extra_args

        return args

    def to_display(self) -> str:
        if not self.rev:
            return ""

        return f" (to revision {self.rev})"

    def make_new(self, rev: str) -> "RevOptions":
        """
        Make a copy of the current instance, but with a new rev.

        Args:
          rev: the name of the revision for the new object.
        """
        return self.vc_class.make_rev_options(rev, extra_args=self.extra_args)


class VcsSupport:
    _registry: Dict[str, "VersionControl"] = {}
    schemes = ["ssh", "git", "hg", "bzr", "sftp", "svn"]

    def __init__(self) -> None:
        # Register more schemes with urlparse for various version control
        # systems
        urllib.parse.uses_netloc.extend(self.schemes)
        super().__init__()

    def __iter__(self) -> Iterator[str]:
        return self._registry.__iter__()

    @property
    def backends(self) -> List["VersionControl"]:
        return list(self._registry.values())

    @property
    def dirnames(self) -> List[str]:
        return [backend.dirname for backend in self.backends]

    @property
    def all_schemes(self) -> List[str]:
        schemes: List[str] = []
        for backend in self.backends:
            schemes.extend(backend.schemes)
        return schemes

    def register(self, cls: Type["VersionControl"]) -> None:
        if not hasattr(cls, "name"):
            logger.warning("Cannot register VCS %s", cls.__name__)
            return
        if cls.name not in self._registry:
            self._registry[cls.name] = cls()
            logger.debug("Registered VCS backend: %s", cls.name)

    def unregister(self, name: str) -> None:
        if name in self._registry:
            del self._registry[name]

    def get_backend_for_dir(self, location: str) -> Optional["VersionControl"]:
        """
        Return a VersionControl object if a repository of that type is found
        at the given directory.
        """
        vcs_backends = {}
        for vcs_backend in self._registry.values():
            repo_path = vcs_backend.get_repository_root(location)
            if not repo_path:
                continue
            logger.debug("Determine that %s uses VCS: %s", location, vcs_backend.name)
            vcs_backends[repo_path] = vcs_backend

        if not vcs_backends:
            return None

        # Choose the VCS in the inner-most directory. Since all repository
        # roots found here would be either `location` or one of its
        # parents, the longest path should have the most path components,
        # i.e. the backend representing the inner-most repository.
        inner_most_repo_path = max(vcs_backends, key=len)
        return vcs_backends[inner_most_repo_path]

    def get_backend_for_scheme(self, scheme: str) -> Optional["VersionControl"]:
        """
        Return a VersionControl object or None.
        """
        for vcs_backend in self._registry.values():
            if scheme in vcs_backend.schemes:
                return vcs_backend
        return None

    def get_backend(self, name: str) -> Optional["VersionControl"]:
        """
        Return a VersionControl object or None.
        """
        name = name.lower()
        return self._registry.get(name)


vcs = VcsSupport()


class VersionControl:
    name = ""
    dirname = ""
    repo_name = ""
    # List of supported schemes for this Version Control
    schemes: Tuple[str, ...] = ()
    # Iterable of environment variable names to pass to call_subprocess().
    unset_environ: Tuple[str, ...] = ()
    default_arg_rev: Optional[str] = None

    @classmethod
    def should_add_vcs_url_prefix(cls, remote_url: str) -> bool:
        """
        Return whether the vcs prefix (e.g. "git+") should be added to a
        repository's remote url when used in a requirement.
        """
        return not remote_url.lower().startswith(f"{cls.name}:")

    @classmethod
    def get_subdirectory(cls, location: str) -> Optional[str]:
        """
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        """
        return None

    @classmethod
    def get_requirement_revision(cls, repo_dir: str) -> str:
        """
        Return the revision string that should be used in a requirement.
        """
        return cls.get_revision(repo_dir)

    @classmethod
    def get_src_requirement(cls, repo_dir: str, project_name: str) -> str:
        """
        Return the requirement string to use to redownload the files
        currently at the given repository directory.

        Args:
          project_name: the (unescaped) project name.

        The return value has a form similar to the following:

            {repository_url}@{revision}#egg={project_name}
        """
        repo_url = cls.get_remote_url(repo_dir)

        if cls.should_add_vcs_url_prefix(repo_url):
            repo_url = f"{cls.name}+{repo_url}"

        revision = cls.get_requirement_revision(repo_dir)
        subdir = cls.get_subdirectory(repo_dir)
        req = make_vcs_requirement_url(repo_url, revision, project_name, subdir=subdir)

        return req

    @staticmethod
    def get_base_rev_args(rev: str) -> List[str]:
        """
        Return the base revision arguments for a vcs command.

        Args:
          rev: the name of a revision to install.  Cannot be None.
        """
        raise NotImplementedError

    def is_immutable_rev_checkout(self, url: str, dest: str) -> bool:
        """
        Return true if the commit hash checked out at dest matches
        the revision in url.

        Always return False, if the VCS does not support immutable commit
        hashes.

        This method does not check if there are local uncommitted changes
        in dest after checkout, as pip currently has no use case for that.
        """
        return False

    @classmethod
    def make_rev_options(
        cls, rev: Optional[str] = None, extra_args: Optional[CommandArgs] = None
    ) -> RevOptions:
        """
        Return a RevOptions object.

        Args:
          rev: the name of a revision to install.
          extra_args: a list of extra options.
        """
        return RevOptions(cls, rev, extra_args=extra_args)

    @classmethod
    def _is_local_repository(cls, repo: str) -> bool:
        """
        posix absolute paths start with os.path.sep,
        win32 ones start with drive (like c:\\folder)
        """
        drive, tail = os.path.splitdrive(repo)
        return repo.startswith(os.path.sep) or bool(drive)

    @classmethod
    def get_netloc_and_auth(
        cls, netloc: str, scheme: str
    ) -> Tuple[str, Tuple[Optional[str], Optional[str]]]:
        """
        Parse the repository URL's netloc, and return the new netloc to use
        along with auth information.

        Args:
          netloc: the original repository URL netloc.
          scheme: the repository URL's scheme without the vcs prefix.

        This is mainly for the Subversion class to override, so that auth
        information can be provided via the --username and --password options
        instead of through the URL.  For other subclasses like Git without
        such an option, auth information must stay in the URL.

        Returns: (netloc, (username, password)).
        """
        return netloc, (None, None)

    @classmethod
    def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]:
        """
        Parse the repository URL to use, and return the URL, revision,
        and auth info to use.

        Returns: (url, rev, (username, password)).
        """
        scheme, netloc, path, query, frag = urllib.parse.urlsplit(url)
        if "+" not in scheme:
            raise ValueError(
                "Sorry, {!r} is a malformed VCS url. "
                "The format is <vcs>+<protocol>://<url>, "
                "e.g. svn+http://myrepo/svn/MyApp#egg=MyApp".format(url)
            )
        # Remove the vcs prefix.
        scheme = scheme.split("+", 1)[1]
        netloc, user_pass = cls.get_netloc_and_auth(netloc, scheme)
        rev = None
        if "@" in path:
            path, rev = path.rsplit("@", 1)
            if not rev:
                raise InstallationError(
                    "The URL {!r} has an empty revision (after @) "
                    "which is not supported. Include a revision after @ "
                    "or remove @ from the URL.".format(url)
                )
        url = urllib.parse.urlunsplit((scheme, netloc, path, query, ""))
        return url, rev, user_pass

    @staticmethod
    def make_rev_args(
        username: Optional[str], password: Optional[HiddenText]
    ) -> CommandArgs:
        """
        Return the RevOptions "extra arguments" to use in obtain().
        """
        return []

    def get_url_rev_options(self, url: HiddenText) -> Tuple[HiddenText, RevOptions]:
        """
        Return the URL and RevOptions object to use in obtain(),
        as a tuple (url, rev_options).
        """
        secret_url, rev, user_pass = self.get_url_rev_and_auth(url.secret)
        username, secret_password = user_pass
        password: Optional[HiddenText] = None
        if secret_password is not None:
            password = hide_value(secret_password)
        extra_args = self.make_rev_args(username, password)
        rev_options = self.make_rev_options(rev, extra_args=extra_args)

        return hide_url(secret_url), rev_options

    @staticmethod
    def normalize_url(url: str) -> str:
        """
        Normalize a URL for comparison by unquoting it and removing any
        trailing slash.
        """
        return urllib.parse.unquote(url).rstrip("/")

    @classmethod
    def compare_urls(cls, url1: str, url2: str) -> bool:
        """
        Compare two repo URLs for identity, ignoring incidental differences.
        """
        return cls.normalize_url(url1) == cls.normalize_url(url2)

    def fetch_new(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        """
        Fetch a revision from a repository, in the case that this is the
        first fetch from the repository.

        Args:
          dest: the directory to fetch the repository to.
          rev_options: a RevOptions object.
        """
        raise NotImplementedError

    def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        """
        Switch the repo at ``dest`` to point to ``URL``.

        Args:
          rev_options: a RevOptions object.
        """
        raise NotImplementedError

    def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        """
        Update an already-existing repo to the given ``rev_options``.

        Args:
          rev_options: a RevOptions object.
        """
        raise NotImplementedError

    @classmethod
    def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool:
        """
        Return whether the id of the current commit equals the given name.

        Args:
          dest: the repository directory.
          name: a string name.
        """
        raise NotImplementedError

    def obtain(self, dest: str, url: HiddenText) -> None:
        """
        Install or update in editable mode the package represented by this
        VersionControl object.

        :param dest: the repository directory in which to install or update.
        :param url: the repository URL starting with a vcs prefix.
        """
        url, rev_options = self.get_url_rev_options(url)

        if not os.path.exists(dest):
            self.fetch_new(dest, url, rev_options)
            return

        rev_display = rev_options.to_display()
        if self.is_repository_directory(dest):
            existing_url = self.get_remote_url(dest)
            if self.compare_urls(existing_url, url.secret):
                logger.debug(
                    "%s in %s exists, and has correct URL (%s)",
                    self.repo_name.title(),
                    display_path(dest),
                    url,
                )
                if not self.is_commit_id_equal(dest, rev_options.rev):
                    logger.info(
                        "Updating %s %s%s",
                        display_path(dest),
                        self.repo_name,
                        rev_display,
                    )
                    self.update(dest, url, rev_options)
                else:
                    logger.info("Skipping because already up-to-date.")
                return

            logger.warning(
                "%s %s in %s exists with URL %s",
                self.name,
                self.repo_name,
                display_path(dest),
                existing_url,
            )
            prompt = ("(s)witch, (i)gnore, (w)ipe, (b)ackup ", ("s", "i", "w", "b"))
        else:
            logger.warning(
                "Directory %s already exists, and is not a %s %s.",
                dest,
                self.name,
                self.repo_name,
            )
            # https://github.com/python/mypy/issues/1174
            prompt = ("(i)gnore, (w)ipe, (b)ackup ", ("i", "w", "b"))  # type: ignore

        logger.warning(
            "The plan is to install the %s repository %s",
            self.name,
            url,
        )
        response = ask_path_exists("What to do?  {}".format(prompt[0]), prompt[1])

        if response == "a":
            sys.exit(-1)

        if response == "w":
            logger.warning("Deleting %s", display_path(dest))
            rmtree(dest)
            self.fetch_new(dest, url, rev_options)
            return

        if response == "b":
            dest_dir = backup_dir(dest)
            logger.warning("Backing up %s to %s", display_path(dest), dest_dir)
            shutil.move(dest, dest_dir)
            self.fetch_new(dest, url, rev_options)
            return

        # Do nothing if the response is "i".
        if response == "s":
            logger.info(
                "Switching %s %s to %s%s",
                self.repo_name,
                display_path(dest),
                url,
                rev_display,
            )
            self.switch(dest, url, rev_options)

    def unpack(self, location: str, url: HiddenText) -> None:
        """
        Clean up current location and download the url repository
        (and vcs infos) into location

        :param url: the repository URL starting with a vcs prefix.
        """
        if os.path.exists(location):
            rmtree(location)
        self.obtain(location, url=url)

    @classmethod
    def get_remote_url(cls, location: str) -> str:
        """
        Return the url used at location

        Raises RemoteNotFoundError if the repository does not have a remote
        url configured.
        """
        raise NotImplementedError

    @classmethod
    def get_revision(cls, location: str) -> str:
        """
        Return the current commit id of the files at the given location.
        """
        raise NotImplementedError

    @classmethod
    def run_command(
        cls,
        cmd: Union[List[str], CommandArgs],
        show_stdout: bool = True,
        cwd: Optional[str] = None,
        on_returncode: 'Literal["raise", "warn", "ignore"]' = "raise",
        extra_ok_returncodes: Optional[Iterable[int]] = None,
        command_desc: Optional[str] = None,
        extra_environ: Optional[Mapping[str, Any]] = None,
        spinner: Optional[SpinnerInterface] = None,
        log_failed_cmd: bool = True,
        stdout_only: bool = False,
    ) -> str:
        """
        Run a VCS subcommand
        This is simply a wrapper around call_subprocess that adds the VCS
        command name, and checks that the VCS is available
        """
        cmd = make_command(cls.name, *cmd)
        try:
            return call_subprocess(
                cmd,
                show_stdout,
                cwd,
                on_returncode=on_returncode,
                extra_ok_returncodes=extra_ok_returncodes,
                command_desc=command_desc,
                extra_environ=extra_environ,
                unset_environ=cls.unset_environ,
                spinner=spinner,
                log_failed_cmd=log_failed_cmd,
                stdout_only=stdout_only,
            )
        except FileNotFoundError:
            # errno.ENOENT = no such file or directory
            # In other words, the VCS executable isn't available
            raise BadCommand(
                f"Cannot find command {cls.name!r} - do you have "
                f"{cls.name!r} installed and in your PATH?"
            )
        except PermissionError:
            # errno.EACCES = Permission denied
            # This error occurs, for instance, when the command is installed
            # only for another user. So, the current user don't have
            # permission to call the other user command.
            raise BadCommand(
                f"No permission to execute {cls.name!r} - install it "
                f"locally, globally (ask admin), or check your PATH. "
                f"See possible solutions at "
                f"https://pip.pypa.io/en/latest/reference/pip_freeze/"
                f"#fixing-permission-denied."
            )

    @classmethod
    def is_repository_directory(cls, path: str) -> bool:
        """
        Return whether a directory path is a repository directory.
        """
        logger.debug("Checking in %s for %s (%s)...", path, cls.dirname, cls.name)
        return os.path.exists(os.path.join(path, cls.dirname))

    @classmethod
    def get_repository_root(cls, location: str) -> Optional[str]:
        """
        Return the "root" (top-level) directory controlled by the vcs,
        or `None` if the directory is not in any.

        It is meant to be overridden to implement smarter detection
        mechanisms for specific vcs.

        This can do more than is_repository_directory() alone. For
        example, the Git override checks that Git is actually available.
        """
        if cls.is_repository_directory(location):
            return location
        return None
site-packages/pip/_internal/vcs/bazaar.py000064400000005451150564745010014450 0ustar00import logging
from typing import List, Optional, Tuple

from pip._internal.utils.misc import HiddenText, display_path
from pip._internal.utils.subprocess import make_command
from pip._internal.utils.urls import path_to_url
from pip._internal.vcs.versioncontrol import (
    AuthInfo,
    RemoteNotFoundError,
    RevOptions,
    VersionControl,
    vcs,
)

logger = logging.getLogger(__name__)


class Bazaar(VersionControl):
    name = "bzr"
    dirname = ".bzr"
    repo_name = "branch"
    schemes = (
        "bzr+http",
        "bzr+https",
        "bzr+ssh",
        "bzr+sftp",
        "bzr+ftp",
        "bzr+lp",
        "bzr+file",
    )

    @staticmethod
    def get_base_rev_args(rev: str) -> List[str]:
        return ["-r", rev]

    def fetch_new(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        rev_display = rev_options.to_display()
        logger.info(
            "Checking out %s%s to %s",
            url,
            rev_display,
            display_path(dest),
        )
        cmd_args = make_command("branch", "-q", rev_options.to_args(), url, dest)
        self.run_command(cmd_args)

    def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        self.run_command(make_command("switch", url), cwd=dest)

    def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        cmd_args = make_command("pull", "-q", rev_options.to_args())
        self.run_command(cmd_args, cwd=dest)

    @classmethod
    def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]:
        # hotfix the URL scheme after removing bzr+ from bzr+ssh:// readd it
        url, rev, user_pass = super().get_url_rev_and_auth(url)
        if url.startswith("ssh://"):
            url = "bzr+" + url
        return url, rev, user_pass

    @classmethod
    def get_remote_url(cls, location: str) -> str:
        urls = cls.run_command(
            ["info"], show_stdout=False, stdout_only=True, cwd=location
        )
        for line in urls.splitlines():
            line = line.strip()
            for x in ("checkout of branch: ", "parent branch: "):
                if line.startswith(x):
                    repo = line.split(x)[1]
                    if cls._is_local_repository(repo):
                        return path_to_url(repo)
                    return repo
        raise RemoteNotFoundError

    @classmethod
    def get_revision(cls, location: str) -> str:
        revision = cls.run_command(
            ["revno"],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        )
        return revision.splitlines()[-1]

    @classmethod
    def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool:
        """Always assume the versions don't match"""
        return False


vcs.register(Bazaar)
site-packages/pip/_internal/vcs/subversion.py000064400000026514150564745010015412 0ustar00import logging
import os
import re
from typing import List, Optional, Tuple

from pip._internal.utils.misc import (
    HiddenText,
    display_path,
    is_console_interactive,
    is_installable_dir,
    split_auth_from_netloc,
)
from pip._internal.utils.subprocess import CommandArgs, make_command
from pip._internal.vcs.versioncontrol import (
    AuthInfo,
    RemoteNotFoundError,
    RevOptions,
    VersionControl,
    vcs,
)

logger = logging.getLogger(__name__)

_svn_xml_url_re = re.compile('url="([^"]+)"')
_svn_rev_re = re.compile(r'committed-rev="(\d+)"')
_svn_info_xml_rev_re = re.compile(r'\s*revision="(\d+)"')
_svn_info_xml_url_re = re.compile(r"<url>(.*)</url>")


class Subversion(VersionControl):
    name = "svn"
    dirname = ".svn"
    repo_name = "checkout"
    schemes = ("svn+ssh", "svn+http", "svn+https", "svn+svn", "svn+file")

    @classmethod
    def should_add_vcs_url_prefix(cls, remote_url: str) -> bool:
        return True

    @staticmethod
    def get_base_rev_args(rev: str) -> List[str]:
        return ["-r", rev]

    @classmethod
    def get_revision(cls, location: str) -> str:
        """
        Return the maximum revision for all files under a given location
        """
        # Note: taken from setuptools.command.egg_info
        revision = 0

        for base, dirs, _ in os.walk(location):
            if cls.dirname not in dirs:
                dirs[:] = []
                continue  # no sense walking uncontrolled subdirs
            dirs.remove(cls.dirname)
            entries_fn = os.path.join(base, cls.dirname, "entries")
            if not os.path.exists(entries_fn):
                # FIXME: should we warn?
                continue

            dirurl, localrev = cls._get_svn_url_rev(base)

            if base == location:
                assert dirurl is not None
                base = dirurl + "/"  # save the root url
            elif not dirurl or not dirurl.startswith(base):
                dirs[:] = []
                continue  # not part of the same svn tree, skip it
            revision = max(revision, localrev)
        return str(revision)

    @classmethod
    def get_netloc_and_auth(
        cls, netloc: str, scheme: str
    ) -> Tuple[str, Tuple[Optional[str], Optional[str]]]:
        """
        This override allows the auth information to be passed to svn via the
        --username and --password options instead of via the URL.
        """
        if scheme == "ssh":
            # The --username and --password options can't be used for
            # svn+ssh URLs, so keep the auth information in the URL.
            return super().get_netloc_and_auth(netloc, scheme)

        return split_auth_from_netloc(netloc)

    @classmethod
    def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]:
        # hotfix the URL scheme after removing svn+ from svn+ssh:// readd it
        url, rev, user_pass = super().get_url_rev_and_auth(url)
        if url.startswith("ssh://"):
            url = "svn+" + url
        return url, rev, user_pass

    @staticmethod
    def make_rev_args(
        username: Optional[str], password: Optional[HiddenText]
    ) -> CommandArgs:
        extra_args: CommandArgs = []
        if username:
            extra_args += ["--username", username]
        if password:
            extra_args += ["--password", password]

        return extra_args

    @classmethod
    def get_remote_url(cls, location: str) -> str:
        # In cases where the source is in a subdirectory, we have to look up in
        # the location until we find a valid project root.
        orig_location = location
        while not is_installable_dir(location):
            last_location = location
            location = os.path.dirname(location)
            if location == last_location:
                # We've traversed up to the root of the filesystem without
                # finding a Python project.
                logger.warning(
                    "Could not find Python project for directory %s (tried all "
                    "parent directories)",
                    orig_location,
                )
                raise RemoteNotFoundError

        url, _rev = cls._get_svn_url_rev(location)
        if url is None:
            raise RemoteNotFoundError

        return url

    @classmethod
    def _get_svn_url_rev(cls, location: str) -> Tuple[Optional[str], int]:
        from pip._internal.exceptions import InstallationError

        entries_path = os.path.join(location, cls.dirname, "entries")
        if os.path.exists(entries_path):
            with open(entries_path) as f:
                data = f.read()
        else:  # subversion >= 1.7 does not have the 'entries' file
            data = ""

        url = None
        if data.startswith("8") or data.startswith("9") or data.startswith("10"):
            entries = list(map(str.splitlines, data.split("\n\x0c\n")))
            del entries[0][0]  # get rid of the '8'
            url = entries[0][3]
            revs = [int(d[9]) for d in entries if len(d) > 9 and d[9]] + [0]
        elif data.startswith("<?xml"):
            match = _svn_xml_url_re.search(data)
            if not match:
                raise ValueError(f"Badly formatted data: {data!r}")
            url = match.group(1)  # get repository URL
            revs = [int(m.group(1)) for m in _svn_rev_re.finditer(data)] + [0]
        else:
            try:
                # subversion >= 1.7
                # Note that using get_remote_call_options is not necessary here
                # because `svn info` is being run against a local directory.
                # We don't need to worry about making sure interactive mode
                # is being used to prompt for passwords, because passwords
                # are only potentially needed for remote server requests.
                xml = cls.run_command(
                    ["info", "--xml", location],
                    show_stdout=False,
                    stdout_only=True,
                )
                match = _svn_info_xml_url_re.search(xml)
                assert match is not None
                url = match.group(1)
                revs = [int(m.group(1)) for m in _svn_info_xml_rev_re.finditer(xml)]
            except InstallationError:
                url, revs = None, []

        if revs:
            rev = max(revs)
        else:
            rev = 0

        return url, rev

    @classmethod
    def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool:
        """Always assume the versions don't match"""
        return False

    def __init__(self, use_interactive: bool = None) -> None:
        if use_interactive is None:
            use_interactive = is_console_interactive()
        self.use_interactive = use_interactive

        # This member is used to cache the fetched version of the current
        # ``svn`` client.
        # Special value definitions:
        #   None: Not evaluated yet.
        #   Empty tuple: Could not parse version.
        self._vcs_version: Optional[Tuple[int, ...]] = None

        super().__init__()

    def call_vcs_version(self) -> Tuple[int, ...]:
        """Query the version of the currently installed Subversion client.

        :return: A tuple containing the parts of the version information or
            ``()`` if the version returned from ``svn`` could not be parsed.
        :raises: BadCommand: If ``svn`` is not installed.
        """
        # Example versions:
        #   svn, version 1.10.3 (r1842928)
        #      compiled Feb 25 2019, 14:20:39 on x86_64-apple-darwin17.0.0
        #   svn, version 1.7.14 (r1542130)
        #      compiled Mar 28 2018, 08:49:13 on x86_64-pc-linux-gnu
        #   svn, version 1.12.0-SlikSvn (SlikSvn/1.12.0)
        #      compiled May 28 2019, 13:44:56 on x86_64-microsoft-windows6.2
        version_prefix = "svn, version "
        version = self.run_command(["--version"], show_stdout=False, stdout_only=True)
        if not version.startswith(version_prefix):
            return ()

        version = version[len(version_prefix) :].split()[0]
        version_list = version.partition("-")[0].split(".")
        try:
            parsed_version = tuple(map(int, version_list))
        except ValueError:
            return ()

        return parsed_version

    def get_vcs_version(self) -> Tuple[int, ...]:
        """Return the version of the currently installed Subversion client.

        If the version of the Subversion client has already been queried,
        a cached value will be used.

        :return: A tuple containing the parts of the version information or
            ``()`` if the version returned from ``svn`` could not be parsed.
        :raises: BadCommand: If ``svn`` is not installed.
        """
        if self._vcs_version is not None:
            # Use cached version, if available.
            # If parsing the version failed previously (empty tuple),
            # do not attempt to parse it again.
            return self._vcs_version

        vcs_version = self.call_vcs_version()
        self._vcs_version = vcs_version
        return vcs_version

    def get_remote_call_options(self) -> CommandArgs:
        """Return options to be used on calls to Subversion that contact the server.

        These options are applicable for the following ``svn`` subcommands used
        in this class.

            - checkout
            - switch
            - update

        :return: A list of command line arguments to pass to ``svn``.
        """
        if not self.use_interactive:
            # --non-interactive switch is available since Subversion 0.14.4.
            # Subversion < 1.8 runs in interactive mode by default.
            return ["--non-interactive"]

        svn_version = self.get_vcs_version()
        # By default, Subversion >= 1.8 runs in non-interactive mode if
        # stdin is not a TTY. Since that is how pip invokes SVN, in
        # call_subprocess(), pip must pass --force-interactive to ensure
        # the user can be prompted for a password, if required.
        #   SVN added the --force-interactive option in SVN 1.8. Since
        # e.g. RHEL/CentOS 7, which is supported until 2024, ships with
        # SVN 1.7, pip should continue to support SVN 1.7. Therefore, pip
        # can't safely add the option if the SVN version is < 1.8 (or unknown).
        if svn_version >= (1, 8):
            return ["--force-interactive"]

        return []

    def fetch_new(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        rev_display = rev_options.to_display()
        logger.info(
            "Checking out %s%s to %s",
            url,
            rev_display,
            display_path(dest),
        )
        cmd_args = make_command(
            "checkout",
            "-q",
            self.get_remote_call_options(),
            rev_options.to_args(),
            url,
            dest,
        )
        self.run_command(cmd_args)

    def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        cmd_args = make_command(
            "switch",
            self.get_remote_call_options(),
            rev_options.to_args(),
            url,
            dest,
        )
        self.run_command(cmd_args)

    def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        cmd_args = make_command(
            "update",
            self.get_remote_call_options(),
            rev_options.to_args(),
            dest,
        )
        self.run_command(cmd_args)


vcs.register(Subversion)
site-packages/pip/_internal/vcs/__init__.py000064400000001124150564745010014740 0ustar00# Expose a limited set of classes and functions so callers outside of
# the vcs package don't need to import deeper than `pip._internal.vcs`.
# (The test directory may still need to import from a vcs sub-package.)
# Import all vcs modules to register each VCS in the VcsSupport object.
import pip._internal.vcs.bazaar
import pip._internal.vcs.git
import pip._internal.vcs.mercurial
import pip._internal.vcs.subversion  # noqa: F401
from pip._internal.vcs.versioncontrol import (  # noqa: F401
    RemoteNotFoundError,
    RemoteNotValidError,
    is_url,
    make_vcs_requirement_url,
    vcs,
)
site-packages/pip/_internal/vcs/mercurial.py000064400000011521150564745010015166 0ustar00import configparser
import logging
import os
from typing import List, Optional

from pip._internal.exceptions import BadCommand, InstallationError
from pip._internal.utils.misc import HiddenText, display_path
from pip._internal.utils.subprocess import make_command
from pip._internal.utils.urls import path_to_url
from pip._internal.vcs.versioncontrol import (
    RevOptions,
    VersionControl,
    find_path_to_project_root_from_repo_root,
    vcs,
)

logger = logging.getLogger(__name__)


class Mercurial(VersionControl):
    name = "hg"
    dirname = ".hg"
    repo_name = "clone"
    schemes = (
        "hg+file",
        "hg+http",
        "hg+https",
        "hg+ssh",
        "hg+static-http",
    )

    @staticmethod
    def get_base_rev_args(rev: str) -> List[str]:
        return [rev]

    def fetch_new(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        rev_display = rev_options.to_display()
        logger.info(
            "Cloning hg %s%s to %s",
            url,
            rev_display,
            display_path(dest),
        )
        self.run_command(make_command("clone", "--noupdate", "-q", url, dest))
        self.run_command(
            make_command("update", "-q", rev_options.to_args()),
            cwd=dest,
        )

    def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        repo_config = os.path.join(dest, self.dirname, "hgrc")
        config = configparser.RawConfigParser()
        try:
            config.read(repo_config)
            config.set("paths", "default", url.secret)
            with open(repo_config, "w") as config_file:
                config.write(config_file)
        except (OSError, configparser.NoSectionError) as exc:
            logger.warning("Could not switch Mercurial repository to %s: %s", url, exc)
        else:
            cmd_args = make_command("update", "-q", rev_options.to_args())
            self.run_command(cmd_args, cwd=dest)

    def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        self.run_command(["pull", "-q"], cwd=dest)
        cmd_args = make_command("update", "-q", rev_options.to_args())
        self.run_command(cmd_args, cwd=dest)

    @classmethod
    def get_remote_url(cls, location: str) -> str:
        url = cls.run_command(
            ["showconfig", "paths.default"],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        ).strip()
        if cls._is_local_repository(url):
            url = path_to_url(url)
        return url.strip()

    @classmethod
    def get_revision(cls, location: str) -> str:
        """
        Return the repository-local changeset revision number, as an integer.
        """
        current_revision = cls.run_command(
            ["parents", "--template={rev}"],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        ).strip()
        return current_revision

    @classmethod
    def get_requirement_revision(cls, location: str) -> str:
        """
        Return the changeset identification hash, as a 40-character
        hexadecimal string
        """
        current_rev_hash = cls.run_command(
            ["parents", "--template={node}"],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        ).strip()
        return current_rev_hash

    @classmethod
    def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool:
        """Always assume the versions don't match"""
        return False

    @classmethod
    def get_subdirectory(cls, location: str) -> Optional[str]:
        """
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        """
        # find the repo root
        repo_root = cls.run_command(
            ["root"], show_stdout=False, stdout_only=True, cwd=location
        ).strip()
        if not os.path.isabs(repo_root):
            repo_root = os.path.abspath(os.path.join(location, repo_root))
        return find_path_to_project_root_from_repo_root(location, repo_root)

    @classmethod
    def get_repository_root(cls, location: str) -> Optional[str]:
        loc = super().get_repository_root(location)
        if loc:
            return loc
        try:
            r = cls.run_command(
                ["root"],
                cwd=location,
                show_stdout=False,
                stdout_only=True,
                on_returncode="raise",
                log_failed_cmd=False,
            )
        except BadCommand:
            logger.debug(
                "could not determine if %s is under hg control "
                "because hg is not available",
                location,
            )
            return None
        except InstallationError:
            return None
        return os.path.normpath(r.rstrip("\r\n"))


vcs.register(Mercurial)
site-packages/pip/_internal/vcs/git.py000064400000042614150564745010013775 0ustar00import logging
import os.path
import pathlib
import re
import urllib.parse
import urllib.request
from typing import List, Optional, Tuple

from pip._internal.exceptions import BadCommand, InstallationError
from pip._internal.utils.misc import HiddenText, display_path, hide_url
from pip._internal.utils.subprocess import make_command
from pip._internal.vcs.versioncontrol import (
    AuthInfo,
    RemoteNotFoundError,
    RemoteNotValidError,
    RevOptions,
    VersionControl,
    find_path_to_project_root_from_repo_root,
    vcs,
)

urlsplit = urllib.parse.urlsplit
urlunsplit = urllib.parse.urlunsplit


logger = logging.getLogger(__name__)


GIT_VERSION_REGEX = re.compile(
    r"^git version "  # Prefix.
    r"(\d+)"  # Major.
    r"\.(\d+)"  # Dot, minor.
    r"(?:\.(\d+))?"  # Optional dot, patch.
    r".*$"  # Suffix, including any pre- and post-release segments we don't care about.
)

HASH_REGEX = re.compile("^[a-fA-F0-9]{40}$")

# SCP (Secure copy protocol) shorthand. e.g. 'git@example.com:foo/bar.git'
SCP_REGEX = re.compile(
    r"""^
    # Optional user, e.g. 'git@'
    (\w+@)?
    # Server, e.g. 'github.com'.
    ([^/:]+):
    # The server-side path. e.g. 'user/project.git'. Must start with an
    # alphanumeric character so as not to be confusable with a Windows paths
    # like 'C:/foo/bar' or 'C:\foo\bar'.
    (\w[^:]*)
    $""",
    re.VERBOSE,
)


def looks_like_hash(sha: str) -> bool:
    return bool(HASH_REGEX.match(sha))


class Git(VersionControl):
    name = "git"
    dirname = ".git"
    repo_name = "clone"
    schemes = (
        "git+http",
        "git+https",
        "git+ssh",
        "git+git",
        "git+file",
    )
    # Prevent the user's environment variables from interfering with pip:
    # https://github.com/pypa/pip/issues/1130
    unset_environ = ("GIT_DIR", "GIT_WORK_TREE")
    default_arg_rev = "HEAD"

    @staticmethod
    def get_base_rev_args(rev: str) -> List[str]:
        return [rev]

    def is_immutable_rev_checkout(self, url: str, dest: str) -> bool:
        _, rev_options = self.get_url_rev_options(hide_url(url))
        if not rev_options.rev:
            return False
        if not self.is_commit_id_equal(dest, rev_options.rev):
            # the current commit is different from rev,
            # which means rev was something else than a commit hash
            return False
        # return False in the rare case rev is both a commit hash
        # and a tag or a branch; we don't want to cache in that case
        # because that branch/tag could point to something else in the future
        is_tag_or_branch = bool(self.get_revision_sha(dest, rev_options.rev)[0])
        return not is_tag_or_branch

    def get_git_version(self) -> Tuple[int, ...]:
        version = self.run_command(["version"], show_stdout=False, stdout_only=True)
        match = GIT_VERSION_REGEX.match(version)
        if not match:
            logger.warning("Can't parse git version: %s", version)
            return ()
        return tuple(int(c) for c in match.groups())

    @classmethod
    def get_current_branch(cls, location: str) -> Optional[str]:
        """
        Return the current branch, or None if HEAD isn't at a branch
        (e.g. detached HEAD).
        """
        # git-symbolic-ref exits with empty stdout if "HEAD" is a detached
        # HEAD rather than a symbolic ref.  In addition, the -q causes the
        # command to exit with status code 1 instead of 128 in this case
        # and to suppress the message to stderr.
        args = ["symbolic-ref", "-q", "HEAD"]
        output = cls.run_command(
            args,
            extra_ok_returncodes=(1,),
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        )
        ref = output.strip()

        if ref.startswith("refs/heads/"):
            return ref[len("refs/heads/") :]

        return None

    @classmethod
    def get_revision_sha(cls, dest: str, rev: str) -> Tuple[Optional[str], bool]:
        """
        Return (sha_or_none, is_branch), where sha_or_none is a commit hash
        if the revision names a remote branch or tag, otherwise None.

        Args:
          dest: the repository directory.
          rev: the revision name.
        """
        # Pass rev to pre-filter the list.
        output = cls.run_command(
            ["show-ref", rev],
            cwd=dest,
            show_stdout=False,
            stdout_only=True,
            on_returncode="ignore",
        )
        refs = {}
        # NOTE: We do not use splitlines here since that would split on other
        #       unicode separators, which can be maliciously used to install a
        #       different revision.
        for line in output.strip().split("\n"):
            line = line.rstrip("\r")
            if not line:
                continue
            try:
                ref_sha, ref_name = line.split(" ", maxsplit=2)
            except ValueError:
                # Include the offending line to simplify troubleshooting if
                # this error ever occurs.
                raise ValueError(f"unexpected show-ref line: {line!r}")

            refs[ref_name] = ref_sha

        branch_ref = f"refs/remotes/origin/{rev}"
        tag_ref = f"refs/tags/{rev}"

        sha = refs.get(branch_ref)
        if sha is not None:
            return (sha, True)

        sha = refs.get(tag_ref)

        return (sha, False)

    @classmethod
    def _should_fetch(cls, dest: str, rev: str) -> bool:
        """
        Return true if rev is a ref or is a commit that we don't have locally.

        Branches and tags are not considered in this method because they are
        assumed to be always available locally (which is a normal outcome of
        ``git clone`` and ``git fetch --tags``).
        """
        if rev.startswith("refs/"):
            # Always fetch remote refs.
            return True

        if not looks_like_hash(rev):
            # Git fetch would fail with abbreviated commits.
            return False

        if cls.has_commit(dest, rev):
            # Don't fetch if we have the commit locally.
            return False

        return True

    @classmethod
    def resolve_revision(
        cls, dest: str, url: HiddenText, rev_options: RevOptions
    ) -> RevOptions:
        """
        Resolve a revision to a new RevOptions object with the SHA1 of the
        branch, tag, or ref if found.

        Args:
          rev_options: a RevOptions object.
        """
        rev = rev_options.arg_rev
        # The arg_rev property's implementation for Git ensures that the
        # rev return value is always non-None.
        assert rev is not None

        sha, is_branch = cls.get_revision_sha(dest, rev)

        if sha is not None:
            rev_options = rev_options.make_new(sha)
            rev_options.branch_name = rev if is_branch else None

            return rev_options

        # Do not show a warning for the common case of something that has
        # the form of a Git commit hash.
        if not looks_like_hash(rev):
            logger.warning(
                "Did not find branch or tag '%s', assuming revision or ref.",
                rev,
            )

        if not cls._should_fetch(dest, rev):
            return rev_options

        # fetch the requested revision
        cls.run_command(
            make_command("fetch", "-q", url, rev_options.to_args()),
            cwd=dest,
        )
        # Change the revision to the SHA of the ref we fetched
        sha = cls.get_revision(dest, rev="FETCH_HEAD")
        rev_options = rev_options.make_new(sha)

        return rev_options

    @classmethod
    def is_commit_id_equal(cls, dest: str, name: Optional[str]) -> bool:
        """
        Return whether the current commit hash equals the given name.

        Args:
          dest: the repository directory.
          name: a string name.
        """
        if not name:
            # Then avoid an unnecessary subprocess call.
            return False

        return cls.get_revision(dest) == name

    def fetch_new(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        rev_display = rev_options.to_display()
        logger.info("Cloning %s%s to %s", url, rev_display, display_path(dest))
        if self.get_git_version() >= (2, 17):
            # Git added support for partial clone in 2.17
            # https://git-scm.com/docs/partial-clone
            # Speeds up cloning by functioning without a complete copy of repository
            self.run_command(
                make_command(
                    "clone",
                    "--filter=blob:none",
                    "-q",
                    url,
                    dest,
                )
            )
        else:
            self.run_command(make_command("clone", "-q", url, dest))

        if rev_options.rev:
            # Then a specific revision was requested.
            rev_options = self.resolve_revision(dest, url, rev_options)
            branch_name = getattr(rev_options, "branch_name", None)
            logger.debug("Rev options %s, branch_name %s", rev_options, branch_name)
            if branch_name is None:
                # Only do a checkout if the current commit id doesn't match
                # the requested revision.
                if not self.is_commit_id_equal(dest, rev_options.rev):
                    cmd_args = make_command(
                        "checkout",
                        "-q",
                        rev_options.to_args(),
                    )
                    self.run_command(cmd_args, cwd=dest)
            elif self.get_current_branch(dest) != branch_name:
                # Then a specific branch was requested, and that branch
                # is not yet checked out.
                track_branch = f"origin/{branch_name}"
                cmd_args = [
                    "checkout",
                    "-b",
                    branch_name,
                    "--track",
                    track_branch,
                ]
                self.run_command(cmd_args, cwd=dest)
        else:
            sha = self.get_revision(dest)
            rev_options = rev_options.make_new(sha)

        logger.info("Resolved %s to commit %s", url, rev_options.rev)

        #: repo may contain submodules
        self.update_submodules(dest)

    def switch(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        self.run_command(
            make_command("config", "remote.origin.url", url),
            cwd=dest,
        )
        cmd_args = make_command("checkout", "-q", rev_options.to_args())
        self.run_command(cmd_args, cwd=dest)

        self.update_submodules(dest)

    def update(self, dest: str, url: HiddenText, rev_options: RevOptions) -> None:
        # First fetch changes from the default remote
        if self.get_git_version() >= (1, 9):
            # fetch tags in addition to everything else
            self.run_command(["fetch", "-q", "--tags"], cwd=dest)
        else:
            self.run_command(["fetch", "-q"], cwd=dest)
        # Then reset to wanted revision (maybe even origin/master)
        rev_options = self.resolve_revision(dest, url, rev_options)
        cmd_args = make_command("reset", "--hard", "-q", rev_options.to_args())
        self.run_command(cmd_args, cwd=dest)
        #: update submodules
        self.update_submodules(dest)

    @classmethod
    def get_remote_url(cls, location: str) -> str:
        """
        Return URL of the first remote encountered.

        Raises RemoteNotFoundError if the repository does not have a remote
        url configured.
        """
        # We need to pass 1 for extra_ok_returncodes since the command
        # exits with return code 1 if there are no matching lines.
        stdout = cls.run_command(
            ["config", "--get-regexp", r"remote\..*\.url"],
            extra_ok_returncodes=(1,),
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        )
        remotes = stdout.splitlines()
        try:
            found_remote = remotes[0]
        except IndexError:
            raise RemoteNotFoundError

        for remote in remotes:
            if remote.startswith("remote.origin.url "):
                found_remote = remote
                break
        url = found_remote.split(" ")[1]
        return cls._git_remote_to_pip_url(url.strip())

    @staticmethod
    def _git_remote_to_pip_url(url: str) -> str:
        """
        Convert a remote url from what git uses to what pip accepts.

        There are 3 legal forms **url** may take:

            1. A fully qualified url: ssh://git@example.com/foo/bar.git
            2. A local project.git folder: /path/to/bare/repository.git
            3. SCP shorthand for form 1: git@example.com:foo/bar.git

        Form 1 is output as-is. Form 2 must be converted to URI and form 3 must
        be converted to form 1.

        See the corresponding test test_git_remote_url_to_pip() for examples of
        sample inputs/outputs.
        """
        if re.match(r"\w+://", url):
            # This is already valid. Pass it though as-is.
            return url
        if os.path.exists(url):
            # A local bare remote (git clone --mirror).
            # Needs a file:// prefix.
            return pathlib.PurePath(url).as_uri()
        scp_match = SCP_REGEX.match(url)
        if scp_match:
            # Add an ssh:// prefix and replace the ':' with a '/'.
            return scp_match.expand(r"ssh://\1\2/\3")
        # Otherwise, bail out.
        raise RemoteNotValidError(url)

    @classmethod
    def has_commit(cls, location: str, rev: str) -> bool:
        """
        Check if rev is a commit that is available in the local repository.
        """
        try:
            cls.run_command(
                ["rev-parse", "-q", "--verify", "sha^" + rev],
                cwd=location,
                log_failed_cmd=False,
            )
        except InstallationError:
            return False
        else:
            return True

    @classmethod
    def get_revision(cls, location: str, rev: Optional[str] = None) -> str:
        if rev is None:
            rev = "HEAD"
        current_rev = cls.run_command(
            ["rev-parse", rev],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        )
        return current_rev.strip()

    @classmethod
    def get_subdirectory(cls, location: str) -> Optional[str]:
        """
        Return the path to Python project root, relative to the repo root.
        Return None if the project root is in the repo root.
        """
        # find the repo root
        git_dir = cls.run_command(
            ["rev-parse", "--git-dir"],
            show_stdout=False,
            stdout_only=True,
            cwd=location,
        ).strip()
        if not os.path.isabs(git_dir):
            git_dir = os.path.join(location, git_dir)
        repo_root = os.path.abspath(os.path.join(git_dir, ".."))
        return find_path_to_project_root_from_repo_root(location, repo_root)

    @classmethod
    def get_url_rev_and_auth(cls, url: str) -> Tuple[str, Optional[str], AuthInfo]:
        """
        Prefixes stub URLs like 'user@hostname:user/repo.git' with 'ssh://'.
        That's required because although they use SSH they sometimes don't
        work with a ssh:// scheme (e.g. GitHub). But we need a scheme for
        parsing. Hence we remove it again afterwards and return it as a stub.
        """
        # Works around an apparent Git bug
        # (see https://article.gmane.org/gmane.comp.version-control.git/146500)
        scheme, netloc, path, query, fragment = urlsplit(url)
        if scheme.endswith("file"):
            initial_slashes = path[: -len(path.lstrip("/"))]
            newpath = initial_slashes + urllib.request.url2pathname(path).replace(
                "\\", "/"
            ).lstrip("/")
            after_plus = scheme.find("+") + 1
            url = scheme[:after_plus] + urlunsplit(
                (scheme[after_plus:], netloc, newpath, query, fragment),
            )

        if "://" not in url:
            assert "file:" not in url
            url = url.replace("git+", "git+ssh://")
            url, rev, user_pass = super().get_url_rev_and_auth(url)
            url = url.replace("ssh://", "")
        else:
            url, rev, user_pass = super().get_url_rev_and_auth(url)

        return url, rev, user_pass

    @classmethod
    def update_submodules(cls, location: str) -> None:
        if not os.path.exists(os.path.join(location, ".gitmodules")):
            return
        cls.run_command(
            ["submodule", "update", "--init", "--recursive", "-q"],
            cwd=location,
        )

    @classmethod
    def get_repository_root(cls, location: str) -> Optional[str]:
        loc = super().get_repository_root(location)
        if loc:
            return loc
        try:
            r = cls.run_command(
                ["rev-parse", "--show-toplevel"],
                cwd=location,
                show_stdout=False,
                stdout_only=True,
                on_returncode="raise",
                log_failed_cmd=False,
            )
        except BadCommand:
            logger.debug(
                "could not determine if %s is under git control "
                "because git is not available",
                location,
            )
            return None
        except InstallationError:
            return None
        return os.path.normpath(r.rstrip("\r\n"))

    @staticmethod
    def should_add_vcs_url_prefix(repo_url: str) -> bool:
        """In either https or ssh form, requirements must be prefixed with git+."""
        return True


vcs.register(Git)
site-packages/pip/_internal/req/__pycache__/req_file.cpython-39.pyc000064400000032144150564745010021260 0ustar00a

��Re
D�@sUdZddlZddlZddlZddlZddlZddlmZddlm	Z	m
Z
mZmZm
Z
mZmZmZmZddlmZddlmZmZddlmZddlmZdd	lmZdd
lmZddlm Z e	r�ddlm!Z!dd
l"m#Z#dgZ$e
ee%e&fZ'ee&gee&effZ(e�)dej*�Z+e�)d�Z,e�)d�Z-ej.ej/ej0ej1ej2ej3ej4ej5ej6ej7ej8ej9ej:ej;gZ<eedej=fe>d<ej?ej@ejAgZBeedej=fe>d<dd�eBD�ZCGdd�d�ZDGdd�d�ZEdBe&eedeejeFeeDd�dd�ZGe&e'd�d d!�ZHdCeEeejeDd"�d#d$�ZIdDee&e%edeejeedd%�d&d'�ZJdEeEeejedeeeeDd(�d)d*�ZKGd+d,�d,�ZLede(d-�d.d/�ZMe&ee&e&fd0�d1d2�ZNGd3d4�d4eO�ZPejQd5�d6d7�ZRe'e'd8�d9d:�ZSe'e'd8�d;d<�ZTe'e'd8�d=d>�ZUe&eee&e&fd?�d@dA�ZVdS)Fz
Requirements file parsing
�N)�Values)	�
TYPE_CHECKING�Any�Callable�Dict�Iterable�Iterator�List�Optional�Tuple)�
cmdoptions)�InstallationError�RequirementsFileParseError)�SearchScope)�
PipSession)�raise_for_status)�auto_decode)�get_url_scheme)�NoReturn)�
PackageFinder�parse_requirementsz^(http|https|file):z(^|\s+)#.*$z#(?P<var>\$\{(?P<name>[A-Z0-9_]+)\}).�SUPPORTED_OPTIONS�SUPPORTED_OPTIONS_REQcCsg|]}t|�j��qS�)�str�dest)�.0�orr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/req_file.py�
<listcomp>N�rc	@s8eZdZdeeeeeeeefeedd�dd�ZdS)�ParsedRequirementN)�requirement�is_editable�
comes_from�
constraint�options�line_source�returncCs(||_||_||_||_||_||_dS�N)r"r#r$r&r%r')�selfr"r#r$r%r&r'rrr�__init__Rs	zParsedRequirement.__init__)NN)	�__name__�
__module__�__qualname__r�boolr
rrr+rrrrr!Qs��r!c@s$eZdZeeeeedd�dd�ZdS)�
ParsedLineN)�filename�lineno�args�optsr%r(cCsZ||_||_||_||_|r0d|_d|_||_n&|jrPd|_d|_|jd|_nd|_dS)NTFr)r1r2r4r%�is_requirementr#r"�	editables)r*r1r2r3r4r%rrrr+dszParsedLine.__init__)r,r-r.r�intrr/r+rrrrr0cs�r0Fr)r1�session�finderr&r%r(c	csFt|�}t||�}|�||�D]"}t||||d�}|dur|VqdS)aqParse a requirements file and yield ParsedRequirement instances.

    :param filename:    Path or url of requirements file.
    :param session:     PipSession instance.
    :param finder:      Instance of pip.index.PackageFinder.
    :param options:     cli options.
    :param constraint:  If true, parsing a constraint file rather than
        requirements file.
    )r&r9r8N)�get_line_parser�RequirementsFileParser�parse�handle_line)	r1r8r9r&r%�line_parser�parser�parsed_line�
parsed_reqrrrr~s
�)�contentr(cCs,t|��dd�}t|�}t|�}t|�}|S)zxSplit, filter, and join lines, and return a line iterator

    :param content: the content of the requirements file
    �)�start)�	enumerate�
splitlines�
join_lines�ignore_comments�expand_env_variables)rB�
lines_enumrrr�
preprocess�s
rK)�liner&r(cCs�d�|jrdnd|j|j�}|js&J�|jrBt|j|j||jd�S|rTt�	||j
�i}tD],}||j
jvr\|j
j|r\|j
j|||<q\d|j�d|j��}t|j|j||j||d�SdS)Nz{} {} (line {})z-cz-r)r"r#r$r%�line � of )r"r#r$r%r&r')
�formatr%r1r2r5r#r!r"r�check_install_build_globalr4�SUPPORTED_OPTIONS_REQ_DEST�__dict__)rLr&�line_comes_from�req_optionsrr'rrr�handle_requirement_line�s8�
��rU)r4r1r2r9r&r8r(cs8�r4|jr|j�_|jr4�j��fdd�|jD��|�r4|j}|j}|jrT|jg}|jdurbg}|jrt|�|j�|jr�|jd}tj	�
tj	�|��}	tj	�|	|�}
tj	�
|
�r�|
}|�|�|r�|�|�t||d�}||_|jr�|��|jr�|��|�r4|j�pgD]$}d|�d|��}
|j||
d��qdS)	Nc3s|]}|�jvr|VqdSr))�features_enabled)r�f�r&rr�	<genexpr>�sz%handle_option_line.<locals>.<genexpr>Tr)�
find_links�
index_urlsrMrN)�source)�require_hashesrV�extendrZr[�	index_url�no_index�extra_index_urls�os�path�dirname�abspath�join�exists�append�update_index_urlsr�search_scope�pre�set_allow_all_prereleases�
prefer_binary�set_prefer_binary�
trusted_hosts�add_trusted_host)r4r1r2r9r&r8rZr[�value�req_dir�relative_to_reqs_filerj�hostr\rrXr�handle_option_line�sL	�



�ru)rLr&r9r8r(cCs4|jrt||�}|St|j|j|j|||�dSdS)aHandle a single parsed requirements line; This can result in
    creating/yielding requirements, or updating the finder.

    :param line:        The parsed line to be processed.
    :param options:     CLI options.
    :param finder:      The finder - updated by non-requirement lines.
    :param session:     The session - updated by non-requirement lines.

    Returns a ParsedRequirement object if the line is a requirement line,
    otherwise returns None.

    For lines that contain requirements, the only options that have an effect
    are from SUPPORTED_OPTIONS_REQ, and they are scoped to the
    requirement. Other options from SUPPORTED_OPTIONS may be present, but are
    ignored.

    For lines that do not contain requirements, the only options that have an
    effect are from SUPPORTED_OPTIONS. Options from SUPPORTED_OPTIONS_REQ may
    be present, but are ignored. These lines may contain multiple options
    (although our docs imply only one is supported), and all our parsed and
    affect the finder.
    N)r5rUrur4r1r2)rLr&r9r8rArrrr=s
�r=c@s`eZdZeedd�dd�Zeeee	d�dd�Z
eeee	d�dd	�Zeeee	d�d
d�ZdS)r;N)r8r>r(cCs||_||_dSr))�_session�_line_parser)r*r8r>rrrr+<szRequirementsFileParser.__init__)r1r%r(ccs|�||�EdHdS)z*Parse a given file, yielding parsed lines.N)�_parse_and_recurse)r*r1r%rrrr<DszRequirementsFileParser.parseccs�|�||�D]�}|js�|jjs&|jjr�|jjr@|jjd}d}n|jjd}d}t�|�rjtj�	||�}n t�|�s�t
j�t
j�
|�|�}|�||�EdHq|VqdS)NrFT)�_parse_filer5r4�requirements�constraints�	SCHEME_RE�search�urllibr<�urljoinrbrcrfrdrx)r*r1r%rL�req_path�nested_constraintrrrrxHs(��


�z)RequirementsFileParser._parse_and_recurseccs�t||j�\}}t|�}|D]l\}}z|�|�\}}	Wn>tyt}
z&d|�d|
j��}t|��WYd}
~
n
d}
~
00t||||	|�VqdS)NzInvalid requirement: �
)�get_file_contentrvrKrw�OptionParsingError�msgrr0)r*r1r%�_rBrJ�line_numberrL�args_strr4�er�rrrrygs�z"RequirementsFileParser._parse_file)
r,r-r.r�
LineParserr+rr/rr0r<rxryrrrrr;;s��r;)r9r(cs ttttfd��fdd�}|S)N�rLr(csJt�}|��}d|_�r �j|_t|�\}}|�t�|�|�\}}||fSr))�build_parser�get_default_valuesr_�format_control�break_args_options�
parse_args�shlex�split)rLr?�defaultsr��options_strr4r��r9rr�
parse_line~sz#get_line_parser.<locals>.parse_line)rrr)r9r�rr�rr:}sr:r�cCsf|�d�}g}|dd�}|D]2}|�d�s6|�d�r<qRq|�|�|�d�qd�|�d�|�fS)z�Break up the line into an args and options string.  We only want to shlex
    (and then optparse) the options, not the args.  args can contain markers
    which are corrupted by shlex.
    � N�-z--r)r��
startswithrh�poprf)rL�tokensr3r&�tokenrrrr��s

r�c@seZdZedd�dd�ZdS)r�N)r�r(cCs
||_dSr))r��r*r�rrrr+�szOptionParsingError.__init__)r,r-r.rr+rrrrr��sr�)r(cCsJtjdd�}tt}|D]}|�}|�|�qttdd�dd�}||_|S)z7
    Return a parser for parsing requirement lines
    F)�add_help_optionr)r*r�r(cSst|��dSr))r�r�rrr�parser_exit�sz!build_parser.<locals>.parser_exit)�optparse�OptionParserrr�
add_optionrr�exit)r?�option_factories�option_factory�optionr�rrrr��sr�)rJr(ccs�d}g}|D]�\}}|�d�r(t�|�rvt�|�r:d|}|rj|�|�|dusTJ�|d�|�fVg}q�||fVq|s~|}|�|�d��q|r�|dus�J�|d�|�fVdS)z�Joins a line ending in '' with the previous line (except when following
    comments).  The joined line takes on the index of the first line.
    N�\r��)�endswith�
COMMENT_RE�matchrhrf�strip)rJ�primary_line_number�new_liner�rLrrrrG�s$

rGccs4|D]*\}}t�d|�}|��}|r||fVqdS)z1
    Strips comments and filter empty lines.
    r�N)r��subr�)rJr�rLrrrrH�s
rHccsL|D]B\}}t�|�D]$\}}t�|�}|s.q|�||�}q||fVqdS)a�Replace all environment variables that can be retrieved via `os.getenv`.

    The only allowed format for environment variables defined in the
    requirement file is `${MY_VARIABLE_1}` to ensure two things:

    1. Strings that contain a `$` aren't accidentally (partially) expanded.
    2. Ensure consistency across platforms for requirement files.

    These points are the result of a discussion on the `github pull
    request #3514 <https://github.com/pypa/pip/pull/3514>`_.

    Valid characters in variable names follow the `POSIX standard
    <http://pubs.opengroup.org/onlinepubs/9699919799/>`_ and are limited
    to uppercase letter, digits and the `_` (underscore).
    N)�
ENV_VAR_RE�findallrb�getenv�replace)rJr�rL�env_var�var_namerqrrrrI�s
rI)�urlr8r(c
Cs�t|�}|dvr.|�|�}t|�|j|jfSz:t|d��}t|���}Wd�n1s\0YWn2ty�}zt	d|����WYd}~n
d}~00||fS)aGets the content of a file; it may be a filename, file: URL, or
    http: URL.  Returns (location, content).  Content is unicode.
    Respects # -*- coding: declarations on the retrieved files.

    :param url:         File path or url.
    :param session:     PipSession instance.
    )�http�https�file�rbNz"Could not open requirements file: )
r�getrr��text�openr�read�OSErrorr
)r�r8�scheme�resprWrB�excrrrr�s
.$r�)NNF)N)NNN)NNN)W�__doc__r�rb�rer��urllib.parser~r�typingrrrrrrr	r
rZpip._internal.clir�pip._internal.exceptionsr
r�!pip._internal.models.search_scoper�pip._internal.network.sessionr�pip._internal.network.utilsr�pip._internal.utils.encodingr�pip._internal.utils.urlsrr�"pip._internal.index.package_finderr�__all__r7r�ReqFileLinesr��compile�Ir|r�r�r_�extra_index_urlr`r{rz�editablerZ�	no_binary�only_binaryrmr]rk�trusted_host�use_new_featurer�Option�__annotations__�install_options�global_options�hashrrQr!r0r/rrKrUrur=r;r:r��	Exceptionr�r�r�rGrHrIr�rrrr�<module>s�,

������1��?��,Bsite-packages/pip/_internal/req/__pycache__/req_tracker.cpython-39.pyc000064400000010304150564745010021766 0ustar00a

��Re�@s�ddlZddlZddlZddlZddlmZddlmZmZm	Z	m
Z
mZmZddl
mZddlmZddlmZe�e�Zejeedd�dd	��Zejed
d�dd
��ZGdd
�d
�ZdS)�N)�
TracebackType)�Dict�Iterator�Optional�Set�Type�Union)�Link)�InstallRequirement)�
TempDirectory)�changes�returnc
ks�tj}t�}i}|��D]<\}}z||||<WntyJ|||<Yn0|||<qzBdVW|��D].\}}||ur�||=qht|t�s�J�|||<qhn:|��D].\}}||ur�||=q�t|t�s�J�|||<q�0dS�N)�os�environ�object�items�KeyError�
isinstance�str)r�target�non_existent_marker�saved_values�name�	new_value�original_value�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/req_tracker.py�update_env_context_managers*
�r�RequirementTracker�r
c	cs�tj�d�}t���t}|durL|�tdd��j}|�t|d��t	�
d|�t|��}|VWd�n1sp0YWd�n1s�0YdS)N�PIP_REQ_TRACKERzreq-tracker)�kind)r!z Initialized build tracking at %s)rr�get�
contextlib�	ExitStack�
enter_contextr�pathr�logger�debugr)�root�ctx�trackerrrr�get_requirement_tracker)s

r-c@s�eZdZedd�dd�Zdd�dd�Zeeeeeee	dd�d	d
�Z
eed�dd
�Ze
dd�dd�Ze
dd�dd�Zdd�dd�Zeje
edd�dd��ZdS)rN)r*r
cCs ||_t�|_t�d|j�dS)NzCreated build tracker: %s)�_root�set�_entriesr(r))�selfr*rrr�__init__7szRequirementTracker.__init__r cCst�d|j�|S)NzEntered build tracker: %s)r(r)r.)r1rrr�	__enter__<szRequirementTracker.__enter__)�exc_type�exc_val�exc_tbr
cCs|��dSr)�cleanup)r1r4r5r6rrr�__exit__@szRequirementTracker.__exit__)�linkr
cCs$t�|j�����}tj�|j|�Sr)	�hashlib�sha224�url_without_fragment�encode�	hexdigestrr'�joinr.)r1r9�hashedrrr�_entry_pathHszRequirementTracker._entry_path)�reqr
cCs�|js
J�|�|j�}z4t|��}|��}Wd�n1s>0YWnty\Yn0d�|j|�}t|��||jvs�J�t|ddd��}|�t	|��Wd�n1s�0Y|j�
|�t�d||j
�dS)z,Add an InstallRequirement to build tracking.Nz{} is already being built: {}�wzutf-8)�encodingzAdded %s to build tracker %r)r9rA�open�read�FileNotFoundError�format�LookupErrorr0�writer�addr(r)r.)r1rB�
entry_path�fp�contents�messagerrrrKLs

*,zRequirementTracker.addcCs<|js
J�t�|�|j��|j�|�t�d||j�dS)z1Remove an InstallRequirement from build tracking.z Removed %s from build tracker %rN)	r9r�unlinkrAr0�remover(r)r.�r1rBrrrrQhs
zRequirementTracker.removecCs,t|j�D]}|�|�q
t�d|j�dS)NzRemoved build tracker: %r)r/r0rQr(r)r.rRrrrr7rszRequirementTracker.cleanupccs|�|�dV|�|�dSr)rKrQrRrrr�trackxs
zRequirementTracker.track)�__name__�
__module__�__qualname__rr2r3rr�
BaseExceptionrr8r	rAr
rKrQr7r$�contextmanagerrrSrrrrr6s
�
)r$r:�loggingr�typesr�typingrrrrrr�pip._internal.models.linkr	Zpip._internal.req.req_installr
�pip._internal.utils.temp_dirr�	getLoggerrTr(rXrrr-rrrrr�<module>s 
site-packages/pip/_internal/req/__pycache__/req_install.cpython-39.pyc000064400000054465150564745010022021 0ustar00a

��ReJ��@s@ddlZddlZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
mZmZm
Z
mZmZddlmZddlmZddlmZddlmZddlmZddlmZdd	lmZdd
lmZddl m!Z!ddl"m#Z#m$Z$dd
l%m&Z&ddl'm(Z(ddl)m*Z*ddl+m,Z,ddl-m.Z.ddl/m,Z0ddl1m2Z3ddl4m5Z5ddl4m6Z7ddl8m9Z9ddl:m;Z;m<Z<ddl=m>Z>ddl?m@Z@ddlAmBZBmCZCddlDmEZEddlFmGZGmHZHmIZImJZJmKZKmLZLmMZMmNZNmOZOddlPmQZQddlRmSZSddlTmUZUmVZVddlWmXZXdd lYmZZZe�[e\�Z]e^e!d!�d"d#�Z_Gd$d%�d%�Z`e`e^d&�d'd(�ZadS))�N)�Any�
Collection�Dict�Iterable�List�Optional�Sequence�Union)�
pkg_resources)�Marker)�Requirement)�SpecifierSet)�canonicalize_name)�Version)�parse)�Pep517HookCaller)�Distribution)�BuildEnvironment�NoOpBuildEnvironment)�InstallationError)�
get_scheme)�Link)�generate_metadata)�generate_editable_metadata)�install_editable)�LegacyInstallFailure)�install)�
install_wheel)�load_pyproject_toml�make_pyproject_path)�UninstallPathSet)�
deprecated)�direct_url_for_editable�direct_url_from_link)�Hashes)	�ask_path_exists�
backup_dir�display_path�dist_in_install_path�dist_in_site_packages�dist_in_usersite�get_distribution�hide_url�redact_auth_from_url)�get_metadata)�runner_with_spinner_message)�
TempDirectory�
tempdir_kinds)�running_under_virtualenv)�vcs)�metadata_directory�returncCs�|�tj�}tj�|�\}}t�||�}|�d�rJtj}tj�	|�d}n.|�d�sXJ�tj
}tj�	|�d�d�d}||||d�S)zQReturn a pkg_resources.Distribution for the provided
    metadata directory.
    z	.egg-inforz
.dist-info�-)�project_name�metadata)�rstrip�os�sep�path�splitr
�PathMetadata�endswithr�splitext�DistInfoDistribution)r4�dist_dir�base_dir�
dist_dir_namer8�dist_cls�	dist_name�rG��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/req_install.py�	_get_distBs
�rIc@s�eZdZdZdSeeeeedfeee	ee
eeeeeeeeeeeeeefee
eeedd�dd�Zed�d	d
�Zed�dd�Zed�d
d�Zeeed�dd��Ze��ed�dd��Zeed�dd��Zeed�dd��ZdTeeeed�dd�Zeed�dd��ZdUeed�dd�Zeed�d d!�Zeeeed"�d#d$�Z dd�d%d&�Z!dd�d'd(�Z"edd)�d*d+�Z#eed�d,d-��Z$eed�d.d/��Z%eed�d0d1��Z&eed�d2d3��Z'eed�d4d5��Z(dd�d6d7�Z)dd�d8d9�Z*dd�d:d;�Z+ee,d�d<d=��Z-e.d�d>d?�Z/dd�d@dA�Z0dVeeeddB�dCdD�Z1dd�dEdF�Z2dWeeee3dG�dHdI�Z4eeeedJ�dKdL�Z5eeddM�dNdO�Z6dXeeee7eeeeeeeeeeddP�	dQdR�Z8dS)Y�InstallRequirementz�
    Represents something that may be installed later on, may have information
    about where to fetch the relevant requirement and also contains logic for
    installing the said requirement.
    FNrG)�req�
comes_from�editable�link�markers�
use_pep517�isolated�install_options�global_options�hash_options�
constraint�extras�
user_supplied�permit_editable_wheelsr5cCs�|dust|t�sJ|��||_||_||_||_||_d|_d|_|jrp|sRJ�|j	rpt
j�t
j�
|j��|_|dur�|r�|jr�t|j�}||_|_d|_d|_|jr�|jj	r�|jj|_|r�||_n |r�dd�|jD�|_nt�|_|dur�|r�|j}||_d|_d|_d|_d|_|�r"|ng|_|	�r2|	ng|_|
�rB|
ni|_d|_ |
|_!||_"t#�|_$d|_%d|_&g|_'d|_(||_)d|_*dS)NFcSsh|]}t�|��qSrG)r
�
safe_extra��.0�extrarGrGrH�	<setcomp>��z.InstallRequirement.__init__.<locals>.<setcomp>)+�
isinstancerrKrLrUrMrX�legacy_install_reason�
source_dir�is_filer:r<�normpath�abspath�	file_path�urlrrN�
original_link�original_link_is_in_wheel_cache�local_file_pathrV�set�markerrO�satisfied_by�should_reinstall�_temp_build_dir�install_succeededrRrSrT�preparedrWrQr�	build_envr4�pyproject_requires�requirements_to_check�pep517_backendrP�needs_more_preparation)�selfrKrLrMrNrOrPrQrRrSrTrUrVrWrXrGrGrH�__init__csX

zInstallRequirement.__init__)r5cCs�|jr.t|j�}|jrF|d�t|jj��7}n|jrBt|jj�}nd}|jdurf|d�t|jj��7}|j	r�t
|j	t�r�|j	}n
|j	��}|r�|d|�d�7}|S)Nz from {}z<InstallRequirement>z in {}z (from �))rK�strrN�formatr-rfrlr'�locationrLr_�	from_path�rv�srLrGrGrH�__str__�s 


zInstallRequirement.__str__cCsd�|jjt|�|j�S)Nz<{} object: {} editable={!r}>)rz�	__class__�__name__ryrM�rvrGrGrH�__repr__�s�zInstallRequirement.__repr__cs>t|��t��}�fdd�t|�D�}dj|jjd�|�d�S)z5An un-tested helper for getting state, for debugging.c3s|]}d�|�|�VqdS)z{}={!r}N)rz)r[�attr��
attributesrGrH�	<genexpr>�r^z2InstallRequirement.format_debug.<locals>.<genexpr>z<{name} object: {{{state}}}>z, )�name�state)�vars�sortedrzr�r��join)rv�namesr�rGr�rH�format_debug�s�zInstallRequirement.format_debugcCs|jdurdSt�|jj�S�N)rKr
�	safe_namer�r�rGrGrHr��s
zInstallRequirement.namec	Cs�|js
dS|jsJ�|j�`td�}|j�|��,d|j��vWd�Wd�S1sb0YWd�n1s�0YdS)NFz1Checking if build backend supports build_editable�build_editable)rPrtrqr/�subprocess_runner�_supported_features)rv�runnerrGrGrH�supports_pyproject_editable�s
�z.InstallRequirement.supports_pyproject_editablecCs|jjSr�)rK�	specifierr�rGrGrHr�szInstallRequirement.specifiercCs$|j}t|�dko"tt|��jdvS)z�Return whether I am pinned to an exact version.

        For example, some-package==1.2 is pinned; some-package>1.2 is not.
        �>�==�===)r��len�next�iter�operator)rv�
specifiersrGrGrH�	is_pinnedszInstallRequirement.is_pinned)�extras_requestedr5cs0|sd}�jdur(t�fdd�|D��SdSdS)N)�c3s|]}�j�d|i�VqdS)r\N)rO�evaluaterZr�rGrHr�sz3InstallRequirement.match_markers.<locals>.<genexpr>T)rO�any)rvr�rGr�rH�
match_markerss
�z InstallRequirement.match_markerscCs
t|j�S)z�Return whether any known-good hashes are specified as options.

        These activate --require-hashes mode; hashes specified as part of a
        URL do not.

        )�boolrTr�rGrGrH�has_hash_optionssz#InstallRequirement.has_hash_optionsT)�trust_internetr5cCsB|j��}|r|jn|j}|r:|jr:|�|jg��|j�t|�S)a�Return a hash-comparer that considers my option- and URL-based
        hashes to be known-good.

        Hashes in URLs--ones embedded in the requirements file, not ones
        downloaded from an index server--are almost peers with ones from
        flags. They satisfy --require-hashes (whether it was implicitly or
        explicitly activated) but do not activate it. md5 and sha224 are not
        allowed in flags, which should nudge people toward good algos. We
        always OR all hashes together, even ones from URLs.

        :param trust_internet: Whether to trust URL-based (#md5=...) hashes
            downloaded from the internet, as by populate_link()

        )	rT�copyrNrg�hash�
setdefault�	hash_name�appendr$)rvr��good_hashesrNrGrGrH�hashes's


zInstallRequirement.hashescCsP|jdurdSt|j�}|jrLt|jt�r2|j}n
|j��}|rL|d|7}|S)z8Format a nice indicator to show where this "comes from" Nz->)rKryrLr_r|r}rGrGrHr|<s


zInstallRequirement.from_path)�	build_dir�
autodelete�parallel_buildsr5cCs�|dusJ�|jdur*|jjs"J�|jjS|jdurLttjdd�|_|jjSt|j�}|rn|�dt�	�j
��}tj�|�s�t
�d|�t�|�tj�||�}|r�dnd}t||tjdd�jS)NT)�kind�globally_managed�_zCreating directory %sF)r<�deleter�r�)rnr<rKr0r1�	REQ_BUILDrr��uuid�uuid4�hexr:�exists�logger�debug�makedirsr�)rvr�r�r��dir_name�actual_build_dir�
delete_argrGrGrH�ensure_build_locationJs.

�

�z(InstallRequirement.ensure_build_locationcCsn|jdusJ�|jdusJ�|jdus*J�tt|jd�t�rDd}nd}td�|jd||jdg��|_dS)z*Set requirement after generating metadata.Nrr�r�r��Name)rKr8rar_�
parse_versionrrr�)rv�oprGrGrH�_set_requirementus���z#InstallRequirement._set_requirementcCsDt|jd�}t|jj�|kr"dSt�d|j||j�t|�|_dS)Nr�zeGenerating metadata for package %s produced metadata for project name %s. Fix your #egg=%s fragments.)rr8rKr�r��warningr)rv�
metadata_namerGrGrH�warn_on_mismatching_name�s�z+InstallRequirement.warn_on_mismatching_name)�
use_user_siter5cCs�|jdurdSt|jj�}|s"dS|j}|duo@|jjj|dd�}|s�d|_|r�t|�r`d|_q�t	�r�t
|�r�td�|j
|j���q�t|�r�d|_n|jr�d|_d|_n||_dS)z�Find an installed distribution that satisfies or conflicts
        with this requirement, and set self.satisfied_by or
        self.should_reinstall appropriately.
        NT)�prereleaseszVWill not install to the user site because it will lack sys.path precedence to {} in {})rKr+r��versionr��containsrlr*rmr2r)rrzr7r{r(rM)rvr��
existing_dist�existing_version�version_compatiblerGrGrH�check_if_exists�s8
����z"InstallRequirement.check_if_existscCs|js
dS|jjS)NF)rN�is_wheelr�rGrGrHr��szInstallRequirement.is_wheelcCstj�|j|jr|jjpd�S)Nr�)r:r<r�rarN�subdirectory_fragmentr�rGrGrH�unpacked_source_directory�s�z,InstallRequirement.unpacked_source_directorycCs(|jsJd|����tj�|jd�}|S)N�No source dir for zsetup.py�rar:r<r�r�)rv�setup_pyrGrGrH�
setup_py_path�sz InstallRequirement.setup_py_pathcCs(|jsJd|����tj�|jd�}|S)Nr�z	setup.cfgr�)rv�	setup_cfgrGrGrH�setup_cfg_path�sz!InstallRequirement.setup_cfg_pathcCs|jsJd|����t|j�S)Nr�)rarr�r�rGrGrH�pyproject_toml_path�sz&InstallRequirement.pyproject_toml_pathcCs^t|j|j|jt|��}|dur*d|_dSd|_|\}}}}||_||_t|j||d�|_	dS)aALoad the pyproject.toml file.

        After calling this routine, all of the attributes related to PEP 517
        processing for this requirement have been set. In particular, the
        use_pep517 attribute can be used to determine whether we should
        follow the PEP 517 or legacy (setup.py) code path.
        NFT)�backend_path)
rrPr�r�ryrsrrrr�rt)rv�pyproject_toml_data�requires�backend�checkr�rGrGrHr�s��z&InstallRequirement.load_pyproject_tomlcCsD|jr@|jr@|��s@tj�|j�s@tj�|j�s@td|�d���dS)z�Check that an editable requirement if valid for use with PEP 517/518.

        This verifies that an editable that has a pyproject.toml either supports PEP 660
        or as a setup.py or a setup.cfg
        zProject z� has a 'pyproject.toml' and its build backend is missing the 'build_editable' hook. Since it does not have a 'setup.py' nor a 'setup.cfg', it cannot be installed in editable mode. Consider using a build backend that supports PEP 660.N)	rMrPr�r:r<�isfiler�r�rr�rGrGrH�isolated_editable_sanity_checks�����
�z1InstallRequirement.isolated_editable_sanity_checkcCs�|js
J�|jrZ|jdusJ�|jrF|jrF|��rFt|j|jd�|_q�t	|j|jd�|_n*t
|j|j|j|j
|jp|d|j��d�|_|js�|��n|��|��dS)z�Ensure that project metadata is available.

        Under PEP 517 and PEP 660, call the backend hook to prepare the metadata.
        Under legacy processing, call setup.py egg-info.
        N)rqr�zfrom )rqr�rarQ�details)rarPrtrMrXr�rrqr4r�generate_metadata_legacyr�r�rQr�rNr�r��assert_source_matches_versionr�rGrGrH�prepare_metadatas8
����
�
�	
z#InstallRequirement.prepare_metadatacCst|d�st|���|_|jS)N�	_metadata)�hasattrr.�get_distr�r�rGrGrHr8Bs
zInstallRequirement.metadatacCs
t|j�Sr�)rIr4r�rGrGrHr�IszInstallRequirement.get_distcCsR|js
J�|jd}|jjr8||jjvr8t�d||�nt�dt|j�||�dS)Nr�z'Requested %s, but installing version %sz;Source in %s has version %s, which satisfies requirement %s)rar8rKr�r�r�r�r')rvr�rGrGrHr�Ls

��z0InstallRequirement.assert_source_matches_version)�
parent_dirr�r�r5cCs |jdur|j|||d�|_dS)aAEnsure that a source_dir is set.

        This will create a temporary build dir if the name of the requirement
        isn't known yet.

        :param parent_dir: The ideal pip parent_dir for the source_dir.
            Generally src_dir for editables and build_dir for sdists.
        :return: self.source_dir
        N)r�r�)rar�)rvr�r�r�rGrGrH�ensure_has_source_dir^s
�z(InstallRequirement.ensure_has_source_dircCs�|jst�d|j�dS|js"J�|js,J�|jjdkr<dSt�|jj�}|s`Jd|jj����t	|jj�}|j
|j|d�dS)Nz>Cannot update repository at %s; repository location is unknown�filezUnsupported VCS URL )rf)rNr�r�rarM�schemer3�get_backend_for_schemerfr,�obtain)rv�vcs_backend�
hidden_urlrGrGrH�update_editableus�

z"InstallRequirement.update_editable)�auto_confirm�verboser5cCsR|js
J�t|jj�}|s,t�d|j�dSt�d|�t�|�}|�||�|S)a�
        Uninstall the distribution currently satisfying this requirement.

        Prompts before removing or modifying files unless
        ``auto_confirm`` is True.

        Refuses to delete or modify files outside of ``sys.prefix`` -
        thus uninstallation within a virtual environment can only
        modify that virtual environment, even if the virtualenv is
        linked to global site-packages.

        z#Skipping %s as it is not installed.NzFound existing installation: %s)	rKr+r�r�r��infor �	from_dist�remove)rvr�r��dist�uninstalled_pathsetrGrGrH�	uninstall�s

zInstallRequirement.uninstall)r<�	parentdir�rootdirr5cCs8tttd�dd�}tj�||�}|||�}|jd|S)N)r��prefixr5cSsN|�|tjj�s&Jd|�d|����|t|�dd�}|�tjjd�}|S)Nzname z doesn't start with prefix r��/)�
startswithr:r<r;r��replace)r�r�rGrGrH�_clean_zip_name�s
��z=InstallRequirement._get_archive_name.<locals>._clean_zip_namer�)ryr:r<r�r�)rvr<r�r�rr�rGrGrH�_get_archive_name�s
z$InstallRequirement._get_archive_name)r�r5cCs�|js
J�|durdSd}d�|j|jd�}tj�||�}tj�|�r�td�t	|��d�}|dkrjd}nj|d	kr�t
�d
t	|��t�|�nF|dkr�t
|�}t
�dt	|�t	|��t�||�n|d
kr�t�d�|s�dStj|d	tjdd�}|��tj�tj�|j��}t�|�D]~\}	}
}|
D]6}|j||	|d�}
t�|
d�}d|_|�|d��q$|D]0}|j||	|d�}tj�|	|�}|�||��q`�qWd�n1�s�0Yt
�dt	|��dS)z}Saves archive to provided build_dir.

        Used for saving downloaded VCS requirements as part of `pip download`.
        NTz	{}-{}.zipr�z8The file {} exists. (i)gnore, (w)ipe, (b)ackup, (a)bort )�i�w�b�arFrzDeleting %srzBacking up %s to %sr���)�
allowZip64)r�r�r�i�r�zSaved %s)rarzr�r8r:r<r�r�r%r'r�r�r�r&�shutil�move�sys�exit�zipfile�ZipFile�ZIP_DEFLATED�normcaserdr��walkr�ZipInfo�
external_attr�writestr�writer�)rvr��create_archive�archive_name�archive_path�response�	dest_file�
zip_output�dir�dirpath�dirnames�	filenames�dirname�dir_arcname�zipdir�filename�file_arcnamerGrGrH�archive�sr
���
���4zInstallRequirement.archive)	rRrS�root�homer��warn_script_locationr��	pycompiler5c	
Cs�t|j||||j|d�}	|dur$|ng}|jrd|jsdt||||||j|j|j|j|jd�
d|_	dS|jr�|j
stJ�d}
|jr�t|j�}
n|jr�t
|j|j|j�}
t|j|j
|	t|j�|||
|jd�d|_	dSt|�|j}t|�|j}z8t||||||||	|j|j|j|j|jt|j�d�}WnLt�yV}zd|_	|j�WYd}~n$d}~0t�ypd|_	�Yn0||_	|�r�|jdk�r�td�|j�d	ddd
�dS)N)�userr&r%rQr�)r�r&r�r�r�rQrqr�T)r��req_descriptionr(r'�
direct_url�	requested)rRrSr%r&r�r�r(r�r�rQ�req_namerqr�r*Fi� zg{} was installed using the legacy 'setup.py install' method, because a wheel could not be built for it.z+to fix the wheel build issue reported above��reason�replacement�gone_in�issue)rr�rQrMr��install_editable_legacyr�rqr�rorir"rgr#rarhrryrKrW�listrSrR�install_legacyr�	__cause__�	Exceptionr`r!rz)
rvrRrSr%r&r�r'r�r(r�r+�success�excrGrGrHr�s��	�
��
	�
��zInstallRequirement.install)FNNNFNNNFrGFF)N)T)FF)FF)NNNNTFT)9r��
__module__�__qualname__�__doc__rrr	ryr�rrrrrrwrr�r��propertyr��	functools�	lru_cacher�r
r�r�rr�r�r$r�r|r�r�r�r�r�r�r�r�r�rr�r�rr8rr�r�r�r�r r�rr$rrrGrGrGrHrJ\s�
�

�j	�+.)����
F�
�rJ)rKr5cCs>d}|jsd}n|jrd}n
|jr&d}|r:tddddd�|S)	Nr�z3Unnamed requirements are not allowed as constraintsz4Editable requirements are not allowed as constraintszConstraints cannot have extrasaConstraints are only allowed to take the form of a package name and a version specifier. Other forms were originally permitted as an accident of the implementation, but were undocumented. The new implementation of the resolver no longer supports these forms.z+replacing the constraint with a requirementi r.)r�rMrVr!)rK�problemrGrGrH�check_invalid_constraint_typebs�rA)br>�loggingr:rr
r�r�typingrrrrrrrr	�pip._vendorr
Zpip._vendor.packaging.markersrZ"pip._vendor.packaging.requirementsrZ pip._vendor.packaging.specifiersr
�pip._vendor.packaging.utilsrZpip._vendor.packaging.versionrrr�Zpip._vendor.pep517.wrappersr�pip._vendor.pkg_resourcesr�pip._internal.build_envrr�pip._internal.exceptionsr�pip._internal.locationsr�pip._internal.models.linkr�'pip._internal.operations.build.metadatar�0pip._internal.operations.build.metadata_editabler�.pip._internal.operations.build.metadata_legacyr��0pip._internal.operations.install.editable_legacyrr3�'pip._internal.operations.install.legacyrrr5�&pip._internal.operations.install.wheelr�pip._internal.pyprojectrr�pip._internal.req.req_uninstallr �pip._internal.utils.deprecationr!�&pip._internal.utils.direct_url_helpersr"r#�pip._internal.utils.hashesr$�pip._internal.utils.miscr%r&r'r(r)r*r+r,r-�pip._internal.utils.packagingr.�pip._internal.utils.subprocessr/�pip._internal.utils.temp_dirr0r1�pip._internal.utils.virtualenvr2�pip._internal.vcsr3�	getLoggerr�r�ryrIrJrArGrGrGrH�<module>s`(,
site-packages/pip/_internal/req/__pycache__/req_set.cpython-39.pyc000064400000013454150564745010021137 0ustar00a

��Re��@s�ddlZddlmZddlmZmZmZmZmZddl	m
Z
ddlmZddl
mZddlmZddlmZe�e�ZGd	d
�d
�ZdS)�N)�OrderedDict)�Dict�Iterable�List�Optional�Tuple)�canonicalize_name)�InstallationError)�Wheel)�InstallRequirement)�compatibility_tagsc@s�eZdZdedd�dd�Zed�dd�Zed�d	d
�Zedd�dd
�Z	edd�dd�Z
deeeeeee
eeeefd�dd�Zeed�dd�Zeed�dd�Zeeed�dd��ZdS)�RequirementSetTN)�check_supported_wheels�returncCst�|_||_g|_dS)zCreate a RequirementSet.N)r�requirementsr�unnamed_requirements)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/req_set.py�__init__szRequirementSet.__init__)rcCs4tdd�|j��D�dd�d�}d�dd�|D��S)Ncss|]}|js|VqdS�N)�
comes_from��.0�reqrrr�	<genexpr>�z)RequirementSet.__str__.<locals>.<genexpr>cSst|jp
d�S�N��r�name�rrrr�<lambda>rz(RequirementSet.__str__.<locals>.<lambda>��key� css|]}t|j�VqdSr��strrrrrrrr)�sortedr�values�join)rrrrr�__str__s
�zRequirementSet.__str__cCsBt|j��dd�d�}d}|j|jjt|�d�dd�|D��d�S)	NcSst|jp
d�Srrr!rrrr""rz)RequirementSet.__repr__.<locals>.<lambda>r#z4<{classname} object; {count} requirement(s): {reqs}>z, css|]}t|j�VqdSrr&rrrrr)rz*RequirementSet.__repr__.<locals>.<genexpr>)�	classname�count�reqs)r(rr)�format�	__class__�__name__�lenr*)rr�
format_stringrrr�__repr__s��zRequirementSet.__repr__)�install_reqrcCs|jr
J�|j�|�dSr)r r�append)rr5rrr�add_unnamed_requirement,s
z&RequirementSet.add_unnamed_requirementcCs"|js
J�t|j�}||j|<dSr)r rr)rr5�project_namerrr�add_named_requirement0s

z$RequirementSet.add_named_requirement)r5�parent_req_name�extras_requestedrc	Cs�|�|�s$t�d|j|j�gdfS|jrf|jjrft|jj�}t	�
�}|jrf|�|�sft
d�|j���|jr||dus|Jd��|js�|�|�|gdfSz|�|j�}Wnty�d}Yn0|duo�|o�|jo�|j|jko�|jo�|jo�|jj|jjk}|�rt
d�|||j���|�s.|�|�|g|fS|j�s>|j�sFg|fS|j�of|j�od|jj|jjk}|�r~t
d�|j���d|_|j�r�d|_ttt|j�t|j�B��|_t�d	||j�|g|fS)
a&Add install_req as a requirement to install.

        :param parent_req_name: The name of the requirement that needed this
            added. The name is used because when multiple unnamed requirements
            resolve to the same name, we could otherwise end up with dependency
            links that point outside the Requirements set. parent_req must
            already be added. Note that None implies that this is a user
            supplied requirement, vs an inferred one.
        :param extras_requested: an iterable of extras used to evaluate the
            environment markers.
        :return: Additional requirements to scan. That is either [] if
            the requirement is not applicable, or [install_req] if the
            requirement is applicable and has just been added.
        z6Ignoring %s: markers '%s' don't match your environmentNz-{} is not a supported wheel on this platform.z+a user supplied req shouldn't have a parentz7Double requirement given: {} (already in {}, name={!r})zhCould not satisfy constraints for '{}': installation from path or url cannot be constrained to a versionFTzSetting %s extras to: %s)�
match_markers�logger�infor �markers�link�is_wheelr
�filenamer�
get_supportedr�	supportedr	r/�
user_suppliedr7�get_requirement�KeyError�
constraint�extrasr�	specifierr9�path�tupler(�set�debug)	rr5r:r;�wheel�tags�existing_req�has_conflicting_requirement�does_not_satisfy_constraintrrr�add_requirement6s�
������

�
��
����	��

�����zRequirementSet.add_requirement)r rcCs t|�}||jvo|j|jSr)rrrH�rr r8rrr�has_requirement�s
�zRequirementSet.has_requirementcCs.t|�}||jvr|j|Std|����dS)NzNo project with the name )rrrGrUrrrrF�s

zRequirementSet.get_requirementcCs|jt|j���Sr)r�listrr))rrrr�all_requirements�szRequirementSet.all_requirements)T)NN)r1�
__module__�__qualname__�boolrr'r+r4rr7r9rrrrrTrVrF�propertyrXrrrrr
s"
	�
�ur
)�logging�collectionsr�typingrrrrr�pip._vendor.packaging.utilsr�pip._internal.exceptionsr	�pip._internal.models.wheelr
Zpip._internal.req.req_installrZpip._internal.utilsr�	getLoggerr1r=r
rrrr�<module>s
site-packages/pip/_internal/req/__pycache__/constructors.cpython-39.pyc000064400000026130150564745010022240 0ustar00a

��Re�;�@s�dZddlZddlZddlZddlmZmZmZmZm	Z	m
Z
ddlmZddl
mZmZddlmZddlmZmZddlmZdd	lmZmZdd
lmZddlmZddlmZdd
l m!Z!ddl"m#Z#ddl$m%Z%ddl&m'Z'ddl(m)Z)ddl*m+Z+m,Z,gd�Z-e�.e/�Z0ej1�2�Z3e4e	e4ee4fd�dd�Z5ee4ee4d�dd�Z6e4e	ee4e4ee4fd�dd�Z7e4e4d�dd�Z8Gd d!�d!�Z9e4e9d�d"d#�Z:d=e4ee
e!e4fee;e;eee4efe;e;e;e!d%�	d&d'�Z<e4e;d(�d)d*�Z=e4e4ee4d+�d,d-�Z>e4ee4e9d.�d/d0�Z?d>e4ee
e4e!fee;e;eee4efe;ee4e;e!d1�	d2d3�Z@d?e4ee!e;ee;e;e!d4�d5d6�ZAd@ee;ee;e;e!d7�d8d9�ZBee!e!d:�d;d<�ZCdS)Aa~Backing implementation for InstallRequirement's various constructors

The idea here is that these formed a major chunk of InstallRequirement's size
so, moving them and support code dedicated to them outside of that class
helps creates for better understandability for the rest of the code.

These are meant to be used elsewhere within pip to create instances of
InstallRequirement.
�N)�Any�Dict�Optional�Set�Tuple�Union)�Marker)�InvalidRequirement�Requirement)�	Specifier)�RequirementParseError�parse_requirements)�InstallationError)�PyPI�TestPyPI)�Link)�Wheel)�ParsedRequirement)�InstallRequirement)�is_archive_file)�is_installable_dir)�get_requirement)�path_to_url)�is_url�vcs)�install_req_from_editable�install_req_from_line�parse_editable)�path�returncCs6t�d|�}d}|r*|�d�}|�d�}n|}||fS)Nz^(.+)(\[[^\]]+\])$��)�re�match�group)r�m�extras�path_no_extras�r(��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/constructors.py�
_strip_extras+s
r*)r&rcCs|s
t�Std|���jS)N�placeholder)�setr�lowerr&)r&r(r(r)�convert_extras7sr.)�editable_reqrcCs�|}t|�\}}tj�|�r$t|�}|���d�rdt|�j}|rX||t	d|���j
fS||t�fStD]*}|���|�d��rh|�d|��}q�qht|�}|j
s�d�tj�}t|�d|�d���|j}|s�td�|���||t�fS)	aParses an editable requirement into:
        - a requirement name
        - an URL
        - extras
        - editable options
    Accepted requirements:
        svn+http://blahblah@rev#egg=Foobar[baz]&subdirectory=version_subdir
        .[some_extra]
    zfile:r+�:�+z, zq is not a valid editable requirement. It should either be a path to a local project or a VCS URL (beginning with z).zZCould not detect requirement name for '{}', please specify one with #egg=your_package_name)r*�osr�isdirrr-�
startswithr�egg_fragmentrr&r,r�is_vcs�join�all_schemesr�format)r/�url�
url_no_extrasr&�package_name�version_control�link�backendsr(r(r)r=s@
�����r)�reqrcCs�d}tj�|�r�d}zJt|��.}tt|����|d�|�7}Wd�n1sR0YWq�ty�t	j
d|dd�Yq�0n|d|�d	�7}|S)
z�Returns helpful msg in case requirements file does not exist,
    or cannot be parsed.

    :params req: Requirements file path
    �z The path does exist. z�The argument you provided ({}) appears to be a requirements file. If that is the case, use the '-r' flag to install the packages specified within it.Nz&Cannot parse '%s' as requirements fileT)�exc_infoz File 'z' does not exist.)r2r�exists�open�nextr
�readr9r�logger�debug)r@�msg�fpr(r(r)�deduce_helpful_msgts
��(rKc@s0eZdZeeeeeeeed�dd�Z	dS)�RequirementParts��requirementr>�markersr&cCs||_||_||_||_dS�NrM)�selfrNr>rOr&r(r(r)�__init__�szRequirementParts.__init__N)
�__name__�
__module__�__qualname__rr
rrr�strrRr(r(r(r)rL�s
�rLcCsbt|�\}}}|durHzt|�}WqLtyDtd|�d���YqL0nd}t|�}t||d|�S)N�Invalid requirement: '�')rr
r	rrrL)r/�namer:�extras_overrider@r>r(r(r)�parse_req_from_editable�sr[F)	r/�
comes_from�
use_pep517�isolated�options�
constraint�
user_supplied�permit_editable_wheelsrc	Cs`t|�}t|j||d||j||||r0|�dg�ng|rB|�dg�ng|rT|�di�ni|jd�
S)NT�install_options�global_options�hashes)r\ra�editablerbr>r`r]r^rcrd�hash_optionsr&)r[rrNr>�getr&)	r/r\r]r^r_r`rarb�partsr(r(r)r�s �r)rYrcCs>tjj|vrdStjjdur,tjj|vr,dS|�d�r:dSdS)akChecks whether the string "looks like" a path on the filesystem.

    This does not check whether the target actually exists, only judge from the
    appearance.

    Returns true if any of the following conditions is true:
    * a path separator is found (either os.path.sep or os.path.altsep);
    * a dot is found (which represents the current directory).
    TN�.F)r2r�sep�altsepr4)rYr(r(r)�_looks_like_path�s

rm)rrYrcCs�t|�r4tj�|�r4t|�r$t|�Std|�d���t|�s@dStj�|�rTt|�S|�	dd�}t
|�dkr|t|d�s|dSt�d|�t|�S)	aK
    First, it checks whether a provided path is an installable directory. If it
    is, returns the path.

    If false, check if the path is an archive file (such as a .whl).
    The function checks if the path is a file. If false, if the path has
    an @, it will treat it as a PEP 440 URL requirement and return the path.
    z
Directory zC is not installable. Neither 'setup.py' nor 'pyproject.toml' found.N�@r r!rzARequirement %r looks like a filename, but the file does not exist)
rmr2rr3rrrr�isfile�split�lenrG�warning)rrY�urlreq_partsr(r(r)�_get_url_from_path�s$	
��rt)rY�line_sourcercsnt|�rd}nd}||vrF|�|d�\}}|��}|s<d}qJt|�}nd}|��}d}tj�tj�|��}d}d}t|�r�t|�}n&t	|�\}	}t
|	|�}
|
dur�t|
�}|�r|jdkr�t�
d|j�r�tttj�tj�|j����}|j�rt|j�}|j�d|j��}n|j}n|}t|�}ttd��fdd	��ttd
��fdd�}
|du�r\|
|�}nd}t||||�S)
Nz; �;r �filez\.\./z==)�textrcs�s|S|�d��d�S)Nz (from �)r()rx)rur(r)�with_source4sz(parse_req_from_line.<locals>.with_source)�
req_as_stringrcs�zt��}Wn�ty�tjj�vr8d}|t��7}n(d�vr\t�fdd�tD��s\d}nd}�d����}|r�|d|��7}t|��Yn40|j	D]*}t
|�}|�d	�r�d
|�d�}t|��q�|S)NzIt looks like a path.�=c3s|]}|�vVqdSrPr()�.0�op�r{r(r)�	<genexpr>@szAparse_req_from_line.<locals>._parse_req_string.<locals>.<genexpr>z,= is not a valid operator. Did you mean == ?rAzInvalid requirement: z
Hint: �]zExtras after version 'z'.)rr	r2rrkrK�any�	operatorsr�	specifierrV�endswith)r{r@�add_msgrI�spec�spec_str)rzrr)�_parse_req_string9s*�


z.parse_req_from_line.<locals>._parse_req_string)rrp�striprr2r�normpath�abspathrr*rt�schemer"�searchr:r�is_wheelr�filenamerY�versionr5r.rVr
rL)rYru�
marker_sep�markers_as_stringrOr{rr>�extras_as_string�pr:�wheelr&r�r@r()rurzr)�parse_req_from_linesH





r�)	rYr\r]r^r_r`rurarc	Csbt||�}t|j||j|j|||r.|�dg�ng|r@|�dg�ng|rR|�di�ni||j|d�S)aCreates an InstallRequirement from a name, which might be a
    requirement, directory containing 'setup.py', filename, or URL.

    :param line_source: An optional string describing where the line is from,
        for logging purposes in case of an error.
    rcrdre)
r>rOr]r^rcrdrgr`r&ra)r�rrNr>rOrhr&)	rYr\r]r^r_r`rurarir(r(r)r^s
�r)�
req_stringr\r^r]rarcCs|zt|�}Wn"ty.td|�d���Yn0tjtjg}|jrj|rj|jrj|jj|vrjtd�	|j
|���t|||||d�S)NrWrXzkPackages installed from PyPI cannot depend on packages which are not also hosted on PyPI.
{} depends on {} )r^r]ra)rr	rr�file_storage_domainrr:r>�netlocr9rYr)r�r\r^r]rar@�domains_not_allowedr(r(r)�install_req_from_req_string�s6����
����r�)�
parsed_reqr^r]rarc
CsH|jr"t|j|j||j||d�}n"t|j|j|||j|j|j|d�}|S)N)r\r]r`r^ra)r\r]r^r_r`rura)�is_editablerrNr\r`rr_ru)r�r^r]rar@r(r(r)�#install_req_from_parsed_requirement�s(�
�
r�)r>�ireqrcCs.t|j|j|j||j|j|j|j|j|j	d�
S)N)
r@r\rfr>rOr]r^rcrdrg)
rr@r\rfrOr]r^rcrdrg)r>r�r(r(r)�install_req_from_link_and_ireq�s�r�)NNFNFFF)NNFNFNF)NFNF)FNF)D�__doc__�loggingr2r"�typingrrrrrrZpip._vendor.packaging.markersrZ"pip._vendor.packaging.requirementsr	r
Z pip._vendor.packaging.specifiersr�pip._vendor.pkg_resourcesrr
�pip._internal.exceptionsr�pip._internal.models.indexrr�pip._internal.models.linkr�pip._internal.models.wheelrZpip._internal.req.req_filerZpip._internal.req.req_installr�pip._internal.utils.filetypesr�pip._internal.utils.miscr�pip._internal.utils.packagingr�pip._internal.utils.urlsr�pip._internal.vcsrr�__all__�	getLoggerrSrG�
_operators�keysr�rVr*r.rrKrLr[�boolrrmrtr�rr�r�r�r(r(r(r)�<module>s�
 

"7��"\��$��(���site-packages/pip/_internal/req/__pycache__/req_uninstall.cpython-39.pyc000064400000044610150564745010022353 0ustar00a

��Re�\�@s�ddlZddlZddlZddlZddlZddlmZddlmZm	Z	m
Z
mZmZm
Z
mZmZmZddlmZddlmZddlmZddlmZmZddlmZdd	lmZdd
lmZm Z ddl!m"Z"m#Z#m$Z$m%Z%m&Z&m'Z'm(Z(ddl)m*Z*m+Z+ee,�Z-ee.e/e
e.d
�dd�Z0e	deefe	deefd�dd�Z1e1eee.d�dd��Z2ee.ee.d�dd�Z3ee.ee.d�dd�Z4ee.eee.ee.fd�dd�Z5Gdd�d�Z6Gd d!�d!�Z7Gd"d#�d#�Z8dS)$�N)�cache_from_source)	�Any�Callable�Dict�Iterable�Iterator�List�Optional�Set�Tuple)�
pkg_resources)�Distribution)�UninstallationError)�get_bin_prefix�get_bin_user)�WINDOWS)�egg_link_path_from_location)�	getLogger�
indent_log)�ask�dist_in_usersite�
dist_is_local�is_local�normalize_path�renames�rmtree)�AdjacentTempDirectory�
TempDirectory)�dist�script_name�is_gui�returncCspt|�rt�}nt�}tj�||�}|g}trl|�|d�|�|d�|r^|�|d�n|�|d�|S)z�Create the fully qualified name of the files created by
    {console,gui}_scripts for the given ``dist``.
    Returns the list of file names
    z.exez
.exe.manifestz-script.pywz
-script.py)rrr�os�path�joinr�append)rrr �bin_dir�exe_name�paths_to_remove�r)��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/req_uninstall.py�
_script_namessr+.)�fnr!cs(t���ttttd��fdd��}|S)N)�args�kwr!c?s6t�}�|i|��D]}||vr|�|�|VqdS�N)�set�add)r-r.�seen�item�r,r)r*�unique5s

z_unique.<locals>.unique)�	functools�wrapsrr)r,r5r)r4r*�_unique4sr8�rr!c
cs8zt�|�d��}Wn�ty�}z�dj|d�}z$t|�d��}|rL|dkrRt��Wn6tttfy�d�|j	|j
�}|d�|�7}Yn0|d�|�7}t|�|�WYd	}~n
d	}~00|D]t}tj
�|j|d
�}|V|�d�r�tj
�|�\}}	|	d	d�}
tj
�||
d
�}|Vtj
�||
d�}|Vq�d	S)a�
    Yield all the uninstallation paths for dist based on RECORD-without-.py[co]

    Yield paths to all the files in RECORD. For each .py file in RECORD, add
    the .pyc and .pyo in the same directory.

    UninstallPathSet.add() takes care of the __pycache__ .py[co].

    If RECORD is not found, raises UninstallationError,
    with possible information from the INSTALLER file.

    https://packaging.python.org/specifications/recording-installed-packages/
    �RECORDz/Cannot uninstall {dist}, RECORD file not found.)r�	INSTALLER�pipz{}=={}zZ You might be able to recover from this via: 'pip install --force-reinstall --no-deps {}'.z' Hint: The package was installed by {}.Nr�.py����.pyc�.pyo)�csv�reader�get_metadata_lines�FileNotFoundError�format�next�
ValueError�OSError�
StopIteration�project_name�versionrr"r#r$�location�endswith�split)r�r�missing_record_exception�msg�	installer�dep�rowr#�dnr,�baser)r)r*�uninstallation_paths@s6
��
 
rW)�pathsr!csJtjj�t�}t|td�D]*�t��fdd�|D��}|s|���q|S)z�Compact a path set to contain the minimal number of paths
    necessary to contain all paths in the set. If /a/path/ and
    /a/path/to/a/file.txt are both in the set, leave only the
    shorter path.��keyc3s:|]2}��|�d��o0�t|�d������kVqdS)�*N)�
startswith�rstrip�len)�.0�	shortpath�r#�sepr)r*�	<genexpr>us�zcompact.<locals>.<genexpr>)r"r#rbr0�sortedr^�anyr1)rX�short_paths�should_skipr)rar*�compactls�rhc	s�dd�|D�}t|�}tdd�|��D�td�}t�}ttd�dd��|D]��t�fd	d
�|D��rfqJt�}t�}t���D]B\�}}|����fdd
�|D��|����fdd
�|D��q|||sJ|�	|�|�
�tj�qJtt|j
|��|BS)
z�Returns a set containing the paths that need to be renamed.

    This set may include directories when the original sequence of paths
    included every file on disk.
    cSsi|]}tj�|�|�qSr))r"r#�normcase�r_�pr)r)r*�
<dictcomp>��z'compress_for_rename.<locals>.<dictcomp>cSsh|]}tj�|�d�qS)r)r"r#rNrjr)r)r*�	<setcomp>�rmz&compress_for_rename.<locals>.<setcomp>rY)�ar!cWstj�tjj|��Sr/)r"r#rir$)ror)r)r*�	norm_join�sz&compress_for_rename.<locals>.norm_joinc3s |]}tj����|�VqdSr/)r"r#rir\)r_�w)�rootr)r*rc�rmz&compress_for_rename.<locals>.<genexpr>c3s|]}���|�VqdSr/r))r_�d��dirnamerprrr)r*rc�rmc3s|]}���|�VqdSr/r))r_�frtr)r*rc�rm)r0rd�valuesr^�strrer"�walk�update�difference_updater1rb�map�__getitem__)	rX�case_map�	remaining�	unchecked�	wildcards�	all_files�all_subdirs�subdirs�filesr)rtr*�compress_for_renames"
r�c
Cs�t|�}t�}t�}t�}|D]>}|�d�r.q|�d�s@d|vrR|�tj�|��|�|�qtttjj|��}t|�}|D]d}t�	|�D]T\}}	}
|
D]D}|�d�r�q�tj�
||�}tj�|�r�tj�|�|vr�|�|�q�q�q||dd�|D�B}||fS)asReturns a tuple of 2 sets of which paths to display to user

    The first set contains paths that would be deleted. Files of a package
    are not added and the top-level directory of the package has a '*' added
    at the end - to signify that all it's contents are removed.

    The second set contains files that would have been skipped in the above
    folders.
    r?z__init__.py�
.dist-infocSsh|]}tj�|d��qS)r[)r"r#r$)r_�folderr)r)r*rn�rmz.compress_for_output_listing.<locals>.<setcomp>)r0rMr1r"r#rur|rirhryr$�isfile)
rX�will_remove�	will_skip�foldersr�r#�_normcased_filesr��dirpath�_�dirfiles�fname�file_r)r)r*�compress_for_output_listing�s2


��r�c@s|eZdZdZdd�dd�Zeed�dd�Zeed�d	d
�Zeed�dd�Zdd�d
d�Z	dd�dd�Z
eed�dd��Z
dS)�StashedUninstallPathSetzWA set of file rename operations to stash files while
    tentatively uninstalling them.N�r!cCsi|_g|_dSr/)�
_save_dirs�_moves��selfr)r)r*�__init__�sz StashedUninstallPathSet.__init__�r#r!cCsBzt|�}Wnty(tdd�}Yn0||jtj�|�<|jS)z�Stashes a directory.

        Directories are stashed adjacent to their original location if
        possible, or else moved/copied into the user's temp dir.�	uninstall��kind)rrHrr�r"r#ri)r�r#�save_dirr)r)r*�_get_directory_stash�sz,StashedUninstallPathSet._get_directory_stashcCs�tj�|�}tj�|�d}}d}||krdz|j|}Wq�WntyNYn0tj�|�|}}q"tj�|�}tdd�}||j|<tj�||�}|r�|tjjkr�tj�	|j|�S|jS)z�Stashes a file.

        If no root has been provided, one will be created for the directory
        in the user's temp directory.Nr�r�)
r"r#rirur��KeyErrorr�relpath�curdirr$)r�r#�head�old_headr�r�r)r)r*�_get_file_stash�s"


z'StashedUninstallPathSet._get_file_stashcCsltj�|�otj�|�}|r*|�|�}n
|�|�}|j�||f�|r^tj�|�r^t�|�t	||�|S)z�Stashes the directory or file and returns its new location.
        Handle symlinks as files to avoid modifying the symlink targets.
        )
r"r#�isdir�islinkr�r�r�r%�rmdirr)r�r#�path_is_dir�new_pathr)r)r*�stashs


zStashedUninstallPathSet.stashcCs,|j��D]\}}|��q
g|_i|_dS)z0Commits the uninstall by removing stashed files.N)r��items�cleanupr�)r�r�r�r)r)r*�commits
zStashedUninstallPathSet.commitcCs�|jD]}tjdg|�R�q|jD]�\}}zTt�d||�tj�|�sTtj�|�r`t�|�ntj�	|�rtt
|�t||�Wq$ty�}z$t�
d|�t�d|�WYd}~q$d}~00q$|��dS)z2Undoes the uninstall by moving stashed files back.zMoving to %s
 from %szReplacing %s from %szFailed to restore %sz
Exception: %sN)r��logger�info�debugr"r#r�r��unlinkr�rrrH�errorr�)r�rkr�r#�exr)r)r*�rollback$s
$z StashedUninstallPathSet.rollbackcCs
t|j�Sr/)�boolr�r�r)r)r*�can_rollback7sz$StashedUninstallPathSet.can_rollback)�__name__�
__module__�__qualname__�__doc__r�rxr�r�r�r�r��propertyr�r�r)r)r)r*r��sr�c@s�eZdZdZedd�dd�Zeed�dd�Zedd�d	d
�Z	eedd�dd
�Z
deedd�dd�Zeed�dd�Zdd�dd�Z
dd�dd�Zeedd�dd��ZdS)�UninstallPathSetzMA set of file paths to be removed in the uninstallation of a
    requirement.Nr9cCs(t�|_t�|_i|_||_t�|_dSr/)r0rX�_refuse�pthrr��_moved_paths)r�rr)r)r*r�@s
zUninstallPathSet.__init__r�cCst|�S)zs
        Return True if the given path is one we are permitted to
        remove/modify, False otherwise.

        )r)r�r#r)r)r*�
_permittedGszUninstallPathSet._permittedcCs�tj�|�\}}tj�t|�tj�|��}tj�|�s:dS|�|�rR|j�	|�n|j
�	|�tj�|�ddkr�|�	t|��dS)N�r=)
r"r#rNr$rri�existsr�rXr1r��splitextr)r�r#r��tailr)r)r*r1Os
zUninstallPathSet.add)�pth_file�entryr!cCsLt|�}|�|�r<||jvr*t|�|j|<|j|�|�n|j�|�dSr/)rr�r��UninstallPthEntriesr1r�)r�r�r�r)r)r*�add_pthbs

zUninstallPathSet.add_pthF)�auto_confirm�verboser!cCs�|jst�d|jj�dS|jjd|jj}t�d|�t��z|sP|�|�r�|j}t	|j�}t
t|��D]}|�|�t�
d|�ql|j��D]}|��q�t�d|�Wd�n1s�0YdS)z[Remove paths in ``self.paths`` with confirmation (unless
        ``auto_confirm`` is True).z7Can't uninstall '%s'. No files were found to uninstall.N�-zUninstalling %s:zRemoving file or directory %szSuccessfully uninstalled %s)rXr�r�rrJrKr�_allowed_to_proceedr�r�rdrhr�r�r�rw�remove)r�r�r��dist_name_version�moved�
for_renamer#r�r)r)r*r�ks$�


zUninstallPathSet.remove)r�r!cCs|tttdd�dd�}|s*t|j�\}}nt|j�}t�}|d|�|d|�|d|j�|rn|dt|j��td	d
�dkS)z@Display which files would be deleted and prompt for confirmationN)rQrXr!cSsX|sdSt�|�t��,tt|��D]}t�|�q&Wd�n1sJ0YdSr/)r�r�rrdrh)rQrXr#r)r)r*�_display�s
z6UninstallPathSet._allowed_to_proceed.<locals>._displayz
Would remove:z+Would not remove (might be manually added):z%Would not remove (outside of prefix):zWill actually move:zProceed (Y/n)? )�y�n�r�)rxrr�rXr0r�r�r)r�r�r�r�r�r)r)r*r��s	


z$UninstallPathSet._allowed_to_proceedr�cCsR|jjst�d|jj�dSt�d|jj�|j��|j�	�D]}|��q@dS)z1Rollback the changes previously made by remove().z'Can't roll back %s; was not uninstalledNzRolling back uninstall of %s)
r�r�r�r�rrJr�r�r�rw)r�r�r)r)r*r��s�
zUninstallPathSet.rollbackcCs|j��dS)z?Remove temporary save dir: rollback will no longer be possible.N)r�r�r�r)r)r*r��szUninstallPathSet.commitcs�t|j�}t|�s.t�d|j|tj�||�S|dd�t�	d�t�	d�hD�vrht�d|j|�||�S||�}t
|j�}d�t
�|j��}|jo�tj�|j�}t|jdd�}|�r�|j�d	��r�|j�|��s�|�|j�|�d
��r |�d
���D]&}tj�tj�|j|��}	|�|	�q�n�|�d��r�|�d��rD|�d��ng��fd
d�|�d���D�D]J}
tj�|j|
�}	|�|	�|�|	d�|�|	d�|�|	d��qd�nF|�r�td�|j����n,|j�d��r$|�|j�tj�|j�d}tj�tj�|j�d�}|�|d|�n�|�rT|j�d��rTt |�D]}	|�|	��q@n�|�r�t!|��$}
tj�"|
�#��$��}Wd�n1�s�0Y||jk�s�Jd�||j|j���|�|�tj�tj�|�d�}|�||j�nt�%d||j�|�d��rn|�&d��rn|�'d�D]P}t(|��r2t)�}nt*�}|�tj�||��t+�r|�tj�||�d��qg}|j,dd�}|�-�D]}|�.t/||d���q�|j,dd�}|�-�D]}|�.t/||d ���q�|D]}|�|��q�|S)!Nz1Not uninstalling %s at %s, outside environment %scSsh|]}|r|�qSr)r)rjr)r)r*rn�s�z-UninstallPathSet.from_dist.<locals>.<setcomp>�stdlib�
platstdlibz<Not uninstalling %s at %s, as it is in the standard library.z{}.egg-infor#z	.egg-infozinstalled-files.txtz
top_level.txtznamespace_packages.txtcsg|]}|r|�vr|�qSr)r)rj��
namespacesr)r*�
<listcomp>�s�z.UninstallPathSet.from_dist.<locals>.<listcomp>r=r?r@z�Cannot uninstall {!r}. It is a distutils installed project and thus we cannot accurately determine which files belong to it which would lead to only a partial uninstall.z.eggr�zeasy-install.pthz./r�z;Egg-link {} does not match installed location of {} (at {})z)Not sure how to uninstall: %s - Check: %s�scriptsz.bat�console_scripts)�groupF�gui_scriptsT)0rrLrr�r�rZ�sys�prefix�	sysconfig�get_pathrrJrEr�to_filename�egg_infor"r#r��getattr�	_providerrMr1�has_metadata�get_metadata�
splitlines�normpathr$rrNrur�rW�openri�readline�stripr��metadata_isdir�metadata_listdirrrrr�
get_entry_map�keys�extendr+)�clsr�	dist_pathr(�develop_egg_link�develop_egg_link_egg_info�egg_info_exists�distutils_egg_info�installed_filer#�
top_level_pkg�easy_install_egg�easy_install_pth�fh�link_pointer�scriptr&�_scripts_to_remover��namer��sr)r�r*�	from_dist�s�
���

��
�
��
�

���
4�
��
��
zUninstallPathSet.from_dist)FF)r�r�r�r�r
r�rxr�r�r1r�r�r�r�r��classmethodrr)r)r)r*r�<s	
r�c@sHeZdZedd�dd�Zedd�dd�Zdd�d	d
�Zed�dd�ZdS)
r�N)r�r!cCs||_t�|_d|_dSr/)�filer0�entries�_saved_lines)r�r�r)r)r*r�CszUninstallPthEntries.__init__)r�r!cCs<tj�|�}tr,tj�|�ds,|�dd�}|j�|�dS)Nr�\�/)r"r#rir�
splitdrive�replacerr1)r�r�r)r)r*r1Hs
zUninstallPthEntries.addr�c	Cs2t�d|j�tj�|j�s.t�d|j�dSt|jd��}|��}||_	Wd�n1s^0Yt
dd�|D��r�d}nd}|r�|d�|�d	��s�|d|�d	�|d<|j
D]<}z$t�d
|�|�||�d	��Wq�ty�Yq�0q�t|jd��}|�|�Wd�n1�s$0YdS)NzRemoving pth entries from %s:z.Cannot remove entries from nonexistent file %s�rbcss|]}d|vVqdS)s
Nr))r_�liner)r)r*rcbrmz-UninstallPthEntries.remove.<locals>.<genexpr>z
�
���zutf-8zRemoving entry: %s�wb)r�r�rr"r#r��warningr��	readlinesrrerM�encoderr�rG�
writelines)r�r��lines�endliner�r)r)r*r�Ws($
zUninstallPthEntries.removecCsf|jdurt�d|j�dSt�d|j�t|jd��}|�|j�Wd�n1sX0YdS)Nz.Cannot roll back changes to %s, none were madeFz!Rolling %s back to previous staterT)rr�r�rr�r�r)r�r�r)r)r*r�rs
*zUninstallPthEntries.rollback)	r�r�r�rxr�r1r�r�r�r)r)r)r*r�Bsr�)9rAr6r"r�r��importlib.utilr�typingrrrrrrr	r
r�pip._vendorr�pip._vendor.pkg_resourcesr
�pip._internal.exceptionsr�pip._internal.locationsrr�pip._internal.utils.compatr�pip._internal.utils.egg_linkr�pip._internal.utils.loggingrr�pip._internal.utils.miscrrrrrrr�pip._internal.utils.temp_dirrrr�r�rxr�r+r8rWrhr�r�r�r�r�r)r)r)r*�<module>s8,$	(+"$2isite-packages/pip/_internal/req/__pycache__/__init__.cpython-39.pyc000064400000005042150564745010021226 0ustar00a

��Re�
�@s�ddlZddlZddlmZmZmZmZmZddlm	Z	ddl
mZddlm
Z
ddlmZgd�Ze�e�ZGd	d
�d
�Zee
eeee
fd�dd
�Zee
eeeeeeeeeeeeeeed�
dd�ZdS)�N)�Iterator�List�Optional�Sequence�Tuple)�
indent_log�)�parse_requirements)�InstallRequirement)�RequirementSet)rr
r	�install_given_reqsc@s*eZdZedd�dd�Zed�dd�ZdS)�InstallationResultN)�name�returncCs
||_dS)N�r)�selfr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/req/__init__.py�__init__szInstallationResult.__init__)rcCsd|j�d�S)NzInstallationResult(name=�)r)rrrr�__repr__szInstallationResult.__repr__)�__name__�
__module__�__qualname__�strrrrrrrr
sr
)�requirementsrccs.|D]$}|jsJd|����|j|fVqdS)Nz%invalid to-be-installed requirement: r)r�reqrrr�_validate_requirementssr)
r�install_options�global_options�root�home�prefix�warn_script_location�
use_user_site�	pycompilerc	Cst�t|��}	|	r(t�dd�|	����g}
t���|	��D]�\}}|j	r�t�d|�t��|j
dd�}
Wd�q�1s~0Ynd}
z|j||||||||d�Wn&ty�|
r�|j
s�|
���Yn0|
r�|j
r�|
��|
�t|��q<Wd�n1�s0Y|
S)zu
    Install everything in the given list.

    (to be called after having downloaded and unpacked the packages)
    z!Installing collected packages: %sz, zAttempting uninstall: %sT)�auto_confirmN)r r!r"r#r$r%)�collections�OrderedDictr�logger�info�join�keysr�items�should_reinstall�	uninstall�install�	Exception�install_succeeded�rollback�commit�appendr
)rrrr r!r"r#r$r%�
to_install�	installed�req_name�requirement�uninstalled_pathsetrrrr%sB�,�



0r)r'�logging�typingrrrrr�pip._internal.utils.loggingr�req_filer	�req_installr
�req_setr�__all__�	getLoggerrr)r
rr�boolrrrrr�<module>s.
	�	�site-packages/pip/_internal/req/req_set.py000064400000016640150564745010014650 0ustar00import logging
from collections import OrderedDict
from typing import Dict, Iterable, List, Optional, Tuple

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.exceptions import InstallationError
from pip._internal.models.wheel import Wheel
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils import compatibility_tags

logger = logging.getLogger(__name__)


class RequirementSet:
    def __init__(self, check_supported_wheels: bool = True) -> None:
        """Create a RequirementSet."""

        self.requirements: Dict[str, InstallRequirement] = OrderedDict()
        self.check_supported_wheels = check_supported_wheels

        self.unnamed_requirements: List[InstallRequirement] = []

    def __str__(self) -> str:
        requirements = sorted(
            (req for req in self.requirements.values() if not req.comes_from),
            key=lambda req: canonicalize_name(req.name or ""),
        )
        return " ".join(str(req.req) for req in requirements)

    def __repr__(self) -> str:
        requirements = sorted(
            self.requirements.values(),
            key=lambda req: canonicalize_name(req.name or ""),
        )

        format_string = "<{classname} object; {count} requirement(s): {reqs}>"
        return format_string.format(
            classname=self.__class__.__name__,
            count=len(requirements),
            reqs=", ".join(str(req.req) for req in requirements),
        )

    def add_unnamed_requirement(self, install_req: InstallRequirement) -> None:
        assert not install_req.name
        self.unnamed_requirements.append(install_req)

    def add_named_requirement(self, install_req: InstallRequirement) -> None:
        assert install_req.name

        project_name = canonicalize_name(install_req.name)
        self.requirements[project_name] = install_req

    def add_requirement(
        self,
        install_req: InstallRequirement,
        parent_req_name: Optional[str] = None,
        extras_requested: Optional[Iterable[str]] = None,
    ) -> Tuple[List[InstallRequirement], Optional[InstallRequirement]]:
        """Add install_req as a requirement to install.

        :param parent_req_name: The name of the requirement that needed this
            added. The name is used because when multiple unnamed requirements
            resolve to the same name, we could otherwise end up with dependency
            links that point outside the Requirements set. parent_req must
            already be added. Note that None implies that this is a user
            supplied requirement, vs an inferred one.
        :param extras_requested: an iterable of extras used to evaluate the
            environment markers.
        :return: Additional requirements to scan. That is either [] if
            the requirement is not applicable, or [install_req] if the
            requirement is applicable and has just been added.
        """
        # If the markers do not match, ignore this requirement.
        if not install_req.match_markers(extras_requested):
            logger.info(
                "Ignoring %s: markers '%s' don't match your environment",
                install_req.name,
                install_req.markers,
            )
            return [], None

        # If the wheel is not supported, raise an error.
        # Should check this after filtering out based on environment markers to
        # allow specifying different wheels based on the environment/OS, in a
        # single requirements file.
        if install_req.link and install_req.link.is_wheel:
            wheel = Wheel(install_req.link.filename)
            tags = compatibility_tags.get_supported()
            if self.check_supported_wheels and not wheel.supported(tags):
                raise InstallationError(
                    "{} is not a supported wheel on this platform.".format(
                        wheel.filename
                    )
                )

        # This next bit is really a sanity check.
        assert (
            not install_req.user_supplied or parent_req_name is None
        ), "a user supplied req shouldn't have a parent"

        # Unnamed requirements are scanned again and the requirement won't be
        # added as a dependency until after scanning.
        if not install_req.name:
            self.add_unnamed_requirement(install_req)
            return [install_req], None

        try:
            existing_req: Optional[InstallRequirement] = self.get_requirement(
                install_req.name
            )
        except KeyError:
            existing_req = None

        has_conflicting_requirement = (
            parent_req_name is None
            and existing_req
            and not existing_req.constraint
            and existing_req.extras == install_req.extras
            and existing_req.req
            and install_req.req
            and existing_req.req.specifier != install_req.req.specifier
        )
        if has_conflicting_requirement:
            raise InstallationError(
                "Double requirement given: {} (already in {}, name={!r})".format(
                    install_req, existing_req, install_req.name
                )
            )

        # When no existing requirement exists, add the requirement as a
        # dependency and it will be scanned again after.
        if not existing_req:
            self.add_named_requirement(install_req)
            # We'd want to rescan this requirement later
            return [install_req], install_req

        # Assume there's no need to scan, and that we've already
        # encountered this for scanning.
        if install_req.constraint or not existing_req.constraint:
            return [], existing_req

        does_not_satisfy_constraint = install_req.link and not (
            existing_req.link and install_req.link.path == existing_req.link.path
        )
        if does_not_satisfy_constraint:
            raise InstallationError(
                "Could not satisfy constraints for '{}': "
                "installation from path or url cannot be "
                "constrained to a version".format(install_req.name)
            )
        # If we're now installing a constraint, mark the existing
        # object for real installation.
        existing_req.constraint = False
        # If we're now installing a user supplied requirement,
        # mark the existing object as such.
        if install_req.user_supplied:
            existing_req.user_supplied = True
        existing_req.extras = tuple(
            sorted(set(existing_req.extras) | set(install_req.extras))
        )
        logger.debug(
            "Setting %s extras to: %s",
            existing_req,
            existing_req.extras,
        )
        # Return the existing requirement for addition to the parent and
        # scanning again.
        return [existing_req], existing_req

    def has_requirement(self, name: str) -> bool:
        project_name = canonicalize_name(name)

        return (
            project_name in self.requirements
            and not self.requirements[project_name].constraint
        )

    def get_requirement(self, name: str) -> InstallRequirement:
        project_name = canonicalize_name(name)

        if project_name in self.requirements:
            return self.requirements[project_name]

        raise KeyError(f"No project with the name {name!r}")

    @property
    def all_requirements(self) -> List[InstallRequirement]:
        return self.unnamed_requirements + list(self.requirements.values())
site-packages/pip/_internal/req/req_install.py000064400000102112150564745010015511 0ustar00# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import functools
import logging
import os
import shutil
import sys
import uuid
import zipfile
from typing import Any, Collection, Dict, Iterable, List, Optional, Sequence, Union

from pip._vendor import pkg_resources
from pip._vendor.packaging.markers import Marker
from pip._vendor.packaging.requirements import Requirement
from pip._vendor.packaging.specifiers import SpecifierSet
from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.packaging.version import Version
from pip._vendor.packaging.version import parse as parse_version
from pip._vendor.pep517.wrappers import Pep517HookCaller
from pip._vendor.pkg_resources import Distribution

from pip._internal.build_env import BuildEnvironment, NoOpBuildEnvironment
from pip._internal.exceptions import InstallationError
from pip._internal.locations import get_scheme
from pip._internal.models.link import Link
from pip._internal.operations.build.metadata import generate_metadata
from pip._internal.operations.build.metadata_editable import generate_editable_metadata
from pip._internal.operations.build.metadata_legacy import (
    generate_metadata as generate_metadata_legacy,
)
from pip._internal.operations.install.editable_legacy import (
    install_editable as install_editable_legacy,
)
from pip._internal.operations.install.legacy import LegacyInstallFailure
from pip._internal.operations.install.legacy import install as install_legacy
from pip._internal.operations.install.wheel import install_wheel
from pip._internal.pyproject import load_pyproject_toml, make_pyproject_path
from pip._internal.req.req_uninstall import UninstallPathSet
from pip._internal.utils.deprecation import deprecated
from pip._internal.utils.direct_url_helpers import (
    direct_url_for_editable,
    direct_url_from_link,
)
from pip._internal.utils.hashes import Hashes
from pip._internal.utils.misc import (
    ask_path_exists,
    backup_dir,
    display_path,
    dist_in_install_path,
    dist_in_site_packages,
    dist_in_usersite,
    get_distribution,
    hide_url,
    redact_auth_from_url,
)
from pip._internal.utils.packaging import get_metadata
from pip._internal.utils.subprocess import runner_with_spinner_message
from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds
from pip._internal.utils.virtualenv import running_under_virtualenv
from pip._internal.vcs import vcs

logger = logging.getLogger(__name__)


def _get_dist(metadata_directory: str) -> Distribution:
    """Return a pkg_resources.Distribution for the provided
    metadata directory.
    """
    dist_dir = metadata_directory.rstrip(os.sep)

    # Build a PathMetadata object, from path to metadata. :wink:
    base_dir, dist_dir_name = os.path.split(dist_dir)
    metadata = pkg_resources.PathMetadata(base_dir, dist_dir)

    # Determine the correct Distribution object type.
    if dist_dir.endswith(".egg-info"):
        dist_cls = pkg_resources.Distribution
        dist_name = os.path.splitext(dist_dir_name)[0]
    else:
        assert dist_dir.endswith(".dist-info")
        dist_cls = pkg_resources.DistInfoDistribution
        dist_name = os.path.splitext(dist_dir_name)[0].split("-")[0]

    return dist_cls(
        base_dir,
        project_name=dist_name,
        metadata=metadata,
    )


class InstallRequirement:
    """
    Represents something that may be installed later on, may have information
    about where to fetch the relevant requirement and also contains logic for
    installing the said requirement.
    """

    def __init__(
        self,
        req: Optional[Requirement],
        comes_from: Optional[Union[str, "InstallRequirement"]],
        editable: bool = False,
        link: Optional[Link] = None,
        markers: Optional[Marker] = None,
        use_pep517: Optional[bool] = None,
        isolated: bool = False,
        install_options: Optional[List[str]] = None,
        global_options: Optional[List[str]] = None,
        hash_options: Optional[Dict[str, List[str]]] = None,
        constraint: bool = False,
        extras: Collection[str] = (),
        user_supplied: bool = False,
        permit_editable_wheels: bool = False,
    ) -> None:
        assert req is None or isinstance(req, Requirement), req
        self.req = req
        self.comes_from = comes_from
        self.constraint = constraint
        self.editable = editable
        self.permit_editable_wheels = permit_editable_wheels
        self.legacy_install_reason: Optional[int] = None

        # source_dir is the local directory where the linked requirement is
        # located, or unpacked. In case unpacking is needed, creating and
        # populating source_dir is done by the RequirementPreparer. Note this
        # is not necessarily the directory where pyproject.toml or setup.py is
        # located - that one is obtained via unpacked_source_directory.
        self.source_dir: Optional[str] = None
        if self.editable:
            assert link
            if link.is_file:
                self.source_dir = os.path.normpath(os.path.abspath(link.file_path))

        if link is None and req and req.url:
            # PEP 508 URL requirement
            link = Link(req.url)
        self.link = self.original_link = link
        self.original_link_is_in_wheel_cache = False

        # Path to any downloaded or already-existing package.
        self.local_file_path: Optional[str] = None
        if self.link and self.link.is_file:
            self.local_file_path = self.link.file_path

        if extras:
            self.extras = extras
        elif req:
            self.extras = {pkg_resources.safe_extra(extra) for extra in req.extras}
        else:
            self.extras = set()
        if markers is None and req:
            markers = req.marker
        self.markers = markers

        # This holds the pkg_resources.Distribution object if this requirement
        # is already available:
        self.satisfied_by: Optional[Distribution] = None
        # Whether the installation process should try to uninstall an existing
        # distribution before installing this requirement.
        self.should_reinstall = False
        # Temporary build location
        self._temp_build_dir: Optional[TempDirectory] = None
        # Set to True after successful installation
        self.install_succeeded: Optional[bool] = None
        # Supplied options
        self.install_options = install_options if install_options else []
        self.global_options = global_options if global_options else []
        self.hash_options = hash_options if hash_options else {}
        # Set to True after successful preparation of this requirement
        self.prepared = False
        # User supplied requirement are explicitly requested for installation
        # by the user via CLI arguments or requirements files, as opposed to,
        # e.g. dependencies, extras or constraints.
        self.user_supplied = user_supplied

        self.isolated = isolated
        self.build_env: BuildEnvironment = NoOpBuildEnvironment()

        # For PEP 517, the directory where we request the project metadata
        # gets stored. We need this to pass to build_wheel, so the backend
        # can ensure that the wheel matches the metadata (see the PEP for
        # details).
        self.metadata_directory: Optional[str] = None

        # The static build requirements (from pyproject.toml)
        self.pyproject_requires: Optional[List[str]] = None

        # Build requirements that we will check are available
        self.requirements_to_check: List[str] = []

        # The PEP 517 backend we should use to build the project
        self.pep517_backend: Optional[Pep517HookCaller] = None

        # Are we using PEP 517 for this requirement?
        # After pyproject.toml has been loaded, the only valid values are True
        # and False. Before loading, None is valid (meaning "use the default").
        # Setting an explicit value before loading pyproject.toml is supported,
        # but after loading this flag should be treated as read only.
        self.use_pep517 = use_pep517

        # This requirement needs more preparation before it can be built
        self.needs_more_preparation = False

    def __str__(self) -> str:
        if self.req:
            s = str(self.req)
            if self.link:
                s += " from {}".format(redact_auth_from_url(self.link.url))
        elif self.link:
            s = redact_auth_from_url(self.link.url)
        else:
            s = "<InstallRequirement>"
        if self.satisfied_by is not None:
            s += " in {}".format(display_path(self.satisfied_by.location))
        if self.comes_from:
            if isinstance(self.comes_from, str):
                comes_from: Optional[str] = self.comes_from
            else:
                comes_from = self.comes_from.from_path()
            if comes_from:
                s += f" (from {comes_from})"
        return s

    def __repr__(self) -> str:
        return "<{} object: {} editable={!r}>".format(
            self.__class__.__name__, str(self), self.editable
        )

    def format_debug(self) -> str:
        """An un-tested helper for getting state, for debugging."""
        attributes = vars(self)
        names = sorted(attributes)

        state = ("{}={!r}".format(attr, attributes[attr]) for attr in sorted(names))
        return "<{name} object: {{{state}}}>".format(
            name=self.__class__.__name__,
            state=", ".join(state),
        )

    # Things that are valid for all kinds of requirements?
    @property
    def name(self) -> Optional[str]:
        if self.req is None:
            return None
        return pkg_resources.safe_name(self.req.name)

    @functools.lru_cache()  # use cached_property in python 3.8+
    def supports_pyproject_editable(self) -> bool:
        if not self.use_pep517:
            return False
        assert self.pep517_backend
        with self.build_env:
            runner = runner_with_spinner_message(
                "Checking if build backend supports build_editable"
            )
            with self.pep517_backend.subprocess_runner(runner):
                return "build_editable" in self.pep517_backend._supported_features()

    @property
    def specifier(self) -> SpecifierSet:
        return self.req.specifier

    @property
    def is_pinned(self) -> bool:
        """Return whether I am pinned to an exact version.

        For example, some-package==1.2 is pinned; some-package>1.2 is not.
        """
        specifiers = self.specifier
        return len(specifiers) == 1 and next(iter(specifiers)).operator in {"==", "==="}

    def match_markers(self, extras_requested: Optional[Iterable[str]] = None) -> bool:
        if not extras_requested:
            # Provide an extra to safely evaluate the markers
            # without matching any extra
            extras_requested = ("",)
        if self.markers is not None:
            return any(
                self.markers.evaluate({"extra": extra}) for extra in extras_requested
            )
        else:
            return True

    @property
    def has_hash_options(self) -> bool:
        """Return whether any known-good hashes are specified as options.

        These activate --require-hashes mode; hashes specified as part of a
        URL do not.

        """
        return bool(self.hash_options)

    def hashes(self, trust_internet: bool = True) -> Hashes:
        """Return a hash-comparer that considers my option- and URL-based
        hashes to be known-good.

        Hashes in URLs--ones embedded in the requirements file, not ones
        downloaded from an index server--are almost peers with ones from
        flags. They satisfy --require-hashes (whether it was implicitly or
        explicitly activated) but do not activate it. md5 and sha224 are not
        allowed in flags, which should nudge people toward good algos. We
        always OR all hashes together, even ones from URLs.

        :param trust_internet: Whether to trust URL-based (#md5=...) hashes
            downloaded from the internet, as by populate_link()

        """
        good_hashes = self.hash_options.copy()
        link = self.link if trust_internet else self.original_link
        if link and link.hash:
            good_hashes.setdefault(link.hash_name, []).append(link.hash)
        return Hashes(good_hashes)

    def from_path(self) -> Optional[str]:
        """Format a nice indicator to show where this "comes from" """
        if self.req is None:
            return None
        s = str(self.req)
        if self.comes_from:
            if isinstance(self.comes_from, str):
                comes_from = self.comes_from
            else:
                comes_from = self.comes_from.from_path()
            if comes_from:
                s += "->" + comes_from
        return s

    def ensure_build_location(
        self, build_dir: str, autodelete: bool, parallel_builds: bool
    ) -> str:
        assert build_dir is not None
        if self._temp_build_dir is not None:
            assert self._temp_build_dir.path
            return self._temp_build_dir.path
        if self.req is None:
            # Some systems have /tmp as a symlink which confuses custom
            # builds (such as numpy). Thus, we ensure that the real path
            # is returned.
            self._temp_build_dir = TempDirectory(
                kind=tempdir_kinds.REQ_BUILD, globally_managed=True
            )

            return self._temp_build_dir.path

        # This is the only remaining place where we manually determine the path
        # for the temporary directory. It is only needed for editables where
        # it is the value of the --src option.

        # When parallel builds are enabled, add a UUID to the build directory
        # name so multiple builds do not interfere with each other.
        dir_name: str = canonicalize_name(self.name)
        if parallel_builds:
            dir_name = f"{dir_name}_{uuid.uuid4().hex}"

        # FIXME: Is there a better place to create the build_dir? (hg and bzr
        # need this)
        if not os.path.exists(build_dir):
            logger.debug("Creating directory %s", build_dir)
            os.makedirs(build_dir)
        actual_build_dir = os.path.join(build_dir, dir_name)
        # `None` indicates that we respect the globally-configured deletion
        # settings, which is what we actually want when auto-deleting.
        delete_arg = None if autodelete else False
        return TempDirectory(
            path=actual_build_dir,
            delete=delete_arg,
            kind=tempdir_kinds.REQ_BUILD,
            globally_managed=True,
        ).path

    def _set_requirement(self) -> None:
        """Set requirement after generating metadata."""
        assert self.req is None
        assert self.metadata is not None
        assert self.source_dir is not None

        # Construct a Requirement object from the generated metadata
        if isinstance(parse_version(self.metadata["Version"]), Version):
            op = "=="
        else:
            op = "==="

        self.req = Requirement(
            "".join(
                [
                    self.metadata["Name"],
                    op,
                    self.metadata["Version"],
                ]
            )
        )

    def warn_on_mismatching_name(self) -> None:
        metadata_name = canonicalize_name(self.metadata["Name"])
        if canonicalize_name(self.req.name) == metadata_name:
            # Everything is fine.
            return

        # If we're here, there's a mismatch. Log a warning about it.
        logger.warning(
            "Generating metadata for package %s "
            "produced metadata for project name %s. Fix your "
            "#egg=%s fragments.",
            self.name,
            metadata_name,
            self.name,
        )
        self.req = Requirement(metadata_name)

    def check_if_exists(self, use_user_site: bool) -> None:
        """Find an installed distribution that satisfies or conflicts
        with this requirement, and set self.satisfied_by or
        self.should_reinstall appropriately.
        """
        if self.req is None:
            return
        existing_dist = get_distribution(self.req.name)
        if not existing_dist:
            return

        # pkg_resouces may contain a different copy of packaging.version from
        # pip in if the downstream distributor does a poor job debundling pip.
        # We avoid existing_dist.parsed_version and let SpecifierSet.contains
        # parses the version instead.
        existing_version = existing_dist.version
        version_compatible = (
            existing_version is not None
            and self.req.specifier.contains(existing_version, prereleases=True)
        )
        if not version_compatible:
            self.satisfied_by = None
            if use_user_site:
                if dist_in_usersite(existing_dist):
                    self.should_reinstall = True
                elif running_under_virtualenv() and dist_in_site_packages(
                    existing_dist
                ):
                    raise InstallationError(
                        "Will not install to the user site because it will "
                        "lack sys.path precedence to {} in {}".format(
                            existing_dist.project_name, existing_dist.location
                        )
                    )
            elif dist_in_install_path(existing_dist):
                self.should_reinstall = True
        else:
            if self.editable:
                self.should_reinstall = True
                # when installing editables, nothing pre-existing should ever
                # satisfy
                self.satisfied_by = None
            else:
                self.satisfied_by = existing_dist

    # Things valid for wheels
    @property
    def is_wheel(self) -> bool:
        if not self.link:
            return False
        return self.link.is_wheel

    # Things valid for sdists
    @property
    def unpacked_source_directory(self) -> str:
        return os.path.join(
            self.source_dir, self.link and self.link.subdirectory_fragment or ""
        )

    @property
    def setup_py_path(self) -> str:
        assert self.source_dir, f"No source dir for {self}"
        setup_py = os.path.join(self.unpacked_source_directory, "setup.py")

        return setup_py

    @property
    def setup_cfg_path(self) -> str:
        assert self.source_dir, f"No source dir for {self}"
        setup_cfg = os.path.join(self.unpacked_source_directory, "setup.cfg")

        return setup_cfg

    @property
    def pyproject_toml_path(self) -> str:
        assert self.source_dir, f"No source dir for {self}"
        return make_pyproject_path(self.unpacked_source_directory)

    def load_pyproject_toml(self) -> None:
        """Load the pyproject.toml file.

        After calling this routine, all of the attributes related to PEP 517
        processing for this requirement have been set. In particular, the
        use_pep517 attribute can be used to determine whether we should
        follow the PEP 517 or legacy (setup.py) code path.
        """
        pyproject_toml_data = load_pyproject_toml(
            self.use_pep517, self.pyproject_toml_path, self.setup_py_path, str(self)
        )

        if pyproject_toml_data is None:
            self.use_pep517 = False
            return

        self.use_pep517 = True
        requires, backend, check, backend_path = pyproject_toml_data
        self.requirements_to_check = check
        self.pyproject_requires = requires
        self.pep517_backend = Pep517HookCaller(
            self.unpacked_source_directory,
            backend,
            backend_path=backend_path,
        )

    def isolated_editable_sanity_check(self) -> None:
        """Check that an editable requirement if valid for use with PEP 517/518.

        This verifies that an editable that has a pyproject.toml either supports PEP 660
        or as a setup.py or a setup.cfg
        """
        if (
            self.editable
            and self.use_pep517
            and not self.supports_pyproject_editable()
            and not os.path.isfile(self.setup_py_path)
            and not os.path.isfile(self.setup_cfg_path)
        ):
            raise InstallationError(
                f"Project {self} has a 'pyproject.toml' and its build "
                f"backend is missing the 'build_editable' hook. Since it does not "
                f"have a 'setup.py' nor a 'setup.cfg', "
                f"it cannot be installed in editable mode. "
                f"Consider using a build backend that supports PEP 660."
            )

    def prepare_metadata(self) -> None:
        """Ensure that project metadata is available.

        Under PEP 517 and PEP 660, call the backend hook to prepare the metadata.
        Under legacy processing, call setup.py egg-info.
        """
        assert self.source_dir

        if self.use_pep517:
            assert self.pep517_backend is not None
            if (
                self.editable
                and self.permit_editable_wheels
                and self.supports_pyproject_editable()
            ):
                self.metadata_directory = generate_editable_metadata(
                    build_env=self.build_env,
                    backend=self.pep517_backend,
                )
            else:
                self.metadata_directory = generate_metadata(
                    build_env=self.build_env,
                    backend=self.pep517_backend,
                )
        else:
            self.metadata_directory = generate_metadata_legacy(
                build_env=self.build_env,
                setup_py_path=self.setup_py_path,
                source_dir=self.unpacked_source_directory,
                isolated=self.isolated,
                details=self.name or f"from {self.link}",
            )

        # Act on the newly generated metadata, based on the name and version.
        if not self.name:
            self._set_requirement()
        else:
            self.warn_on_mismatching_name()

        self.assert_source_matches_version()

    @property
    def metadata(self) -> Any:
        if not hasattr(self, "_metadata"):
            self._metadata = get_metadata(self.get_dist())

        return self._metadata

    def get_dist(self) -> Distribution:
        return _get_dist(self.metadata_directory)

    def assert_source_matches_version(self) -> None:
        assert self.source_dir
        version = self.metadata["version"]
        if self.req.specifier and version not in self.req.specifier:
            logger.warning(
                "Requested %s, but installing version %s",
                self,
                version,
            )
        else:
            logger.debug(
                "Source in %s has version %s, which satisfies requirement %s",
                display_path(self.source_dir),
                version,
                self,
            )

    # For both source distributions and editables
    def ensure_has_source_dir(
        self,
        parent_dir: str,
        autodelete: bool = False,
        parallel_builds: bool = False,
    ) -> None:
        """Ensure that a source_dir is set.

        This will create a temporary build dir if the name of the requirement
        isn't known yet.

        :param parent_dir: The ideal pip parent_dir for the source_dir.
            Generally src_dir for editables and build_dir for sdists.
        :return: self.source_dir
        """
        if self.source_dir is None:
            self.source_dir = self.ensure_build_location(
                parent_dir,
                autodelete=autodelete,
                parallel_builds=parallel_builds,
            )

    # For editable installations
    def update_editable(self) -> None:
        if not self.link:
            logger.debug(
                "Cannot update repository at %s; repository location is unknown",
                self.source_dir,
            )
            return
        assert self.editable
        assert self.source_dir
        if self.link.scheme == "file":
            # Static paths don't get updated
            return
        vcs_backend = vcs.get_backend_for_scheme(self.link.scheme)
        # Editable requirements are validated in Requirement constructors.
        # So here, if it's neither a path nor a valid VCS URL, it's a bug.
        assert vcs_backend, f"Unsupported VCS URL {self.link.url}"
        hidden_url = hide_url(self.link.url)
        vcs_backend.obtain(self.source_dir, url=hidden_url)

    # Top-level Actions
    def uninstall(
        self, auto_confirm: bool = False, verbose: bool = False
    ) -> Optional[UninstallPathSet]:
        """
        Uninstall the distribution currently satisfying this requirement.

        Prompts before removing or modifying files unless
        ``auto_confirm`` is True.

        Refuses to delete or modify files outside of ``sys.prefix`` -
        thus uninstallation within a virtual environment can only
        modify that virtual environment, even if the virtualenv is
        linked to global site-packages.

        """
        assert self.req
        dist = get_distribution(self.req.name)
        if not dist:
            logger.warning("Skipping %s as it is not installed.", self.name)
            return None
        logger.info("Found existing installation: %s", dist)

        uninstalled_pathset = UninstallPathSet.from_dist(dist)
        uninstalled_pathset.remove(auto_confirm, verbose)
        return uninstalled_pathset

    def _get_archive_name(self, path: str, parentdir: str, rootdir: str) -> str:
        def _clean_zip_name(name: str, prefix: str) -> str:
            assert name.startswith(
                prefix + os.path.sep
            ), f"name {name!r} doesn't start with prefix {prefix!r}"
            name = name[len(prefix) + 1 :]
            name = name.replace(os.path.sep, "/")
            return name

        path = os.path.join(parentdir, path)
        name = _clean_zip_name(path, rootdir)
        return self.name + "/" + name

    def archive(self, build_dir: Optional[str]) -> None:
        """Saves archive to provided build_dir.

        Used for saving downloaded VCS requirements as part of `pip download`.
        """
        assert self.source_dir
        if build_dir is None:
            return

        create_archive = True
        archive_name = "{}-{}.zip".format(self.name, self.metadata["version"])
        archive_path = os.path.join(build_dir, archive_name)

        if os.path.exists(archive_path):
            response = ask_path_exists(
                "The file {} exists. (i)gnore, (w)ipe, "
                "(b)ackup, (a)bort ".format(display_path(archive_path)),
                ("i", "w", "b", "a"),
            )
            if response == "i":
                create_archive = False
            elif response == "w":
                logger.warning("Deleting %s", display_path(archive_path))
                os.remove(archive_path)
            elif response == "b":
                dest_file = backup_dir(archive_path)
                logger.warning(
                    "Backing up %s to %s",
                    display_path(archive_path),
                    display_path(dest_file),
                )
                shutil.move(archive_path, dest_file)
            elif response == "a":
                sys.exit(-1)

        if not create_archive:
            return

        zip_output = zipfile.ZipFile(
            archive_path,
            "w",
            zipfile.ZIP_DEFLATED,
            allowZip64=True,
        )
        with zip_output:
            dir = os.path.normcase(os.path.abspath(self.unpacked_source_directory))
            for dirpath, dirnames, filenames in os.walk(dir):
                for dirname in dirnames:
                    dir_arcname = self._get_archive_name(
                        dirname,
                        parentdir=dirpath,
                        rootdir=dir,
                    )
                    zipdir = zipfile.ZipInfo(dir_arcname + "/")
                    zipdir.external_attr = 0x1ED << 16  # 0o755
                    zip_output.writestr(zipdir, "")
                for filename in filenames:
                    file_arcname = self._get_archive_name(
                        filename,
                        parentdir=dirpath,
                        rootdir=dir,
                    )
                    filename = os.path.join(dirpath, filename)
                    zip_output.write(filename, file_arcname)

        logger.info("Saved %s", display_path(archive_path))

    def install(
        self,
        install_options: List[str],
        global_options: Optional[Sequence[str]] = None,
        root: Optional[str] = None,
        home: Optional[str] = None,
        prefix: Optional[str] = None,
        warn_script_location: bool = True,
        use_user_site: bool = False,
        pycompile: bool = True,
    ) -> None:
        scheme = get_scheme(
            self.name,
            user=use_user_site,
            home=home,
            root=root,
            isolated=self.isolated,
            prefix=prefix,
        )

        global_options = global_options if global_options is not None else []
        if self.editable and not self.is_wheel:
            install_editable_legacy(
                install_options,
                global_options,
                prefix=prefix,
                home=home,
                use_user_site=use_user_site,
                name=self.name,
                setup_py_path=self.setup_py_path,
                isolated=self.isolated,
                build_env=self.build_env,
                unpacked_source_directory=self.unpacked_source_directory,
            )
            self.install_succeeded = True
            return

        if self.is_wheel:
            assert self.local_file_path
            direct_url = None
            if self.editable:
                direct_url = direct_url_for_editable(self.unpacked_source_directory)
            elif self.original_link:
                direct_url = direct_url_from_link(
                    self.original_link,
                    self.source_dir,
                    self.original_link_is_in_wheel_cache,
                )
            install_wheel(
                self.name,
                self.local_file_path,
                scheme=scheme,
                req_description=str(self.req),
                pycompile=pycompile,
                warn_script_location=warn_script_location,
                direct_url=direct_url,
                requested=self.user_supplied,
            )
            self.install_succeeded = True
            return

        # TODO: Why don't we do this for editable installs?

        # Extend the list of global and install options passed on to
        # the setup.py call with the ones from the requirements file.
        # Options specified in requirements file override those
        # specified on the command line, since the last option given
        # to setup.py is the one that is used.
        global_options = list(global_options) + self.global_options
        install_options = list(install_options) + self.install_options

        try:
            success = install_legacy(
                install_options=install_options,
                global_options=global_options,
                root=root,
                home=home,
                prefix=prefix,
                use_user_site=use_user_site,
                pycompile=pycompile,
                scheme=scheme,
                setup_py_path=self.setup_py_path,
                isolated=self.isolated,
                req_name=self.name,
                build_env=self.build_env,
                unpacked_source_directory=self.unpacked_source_directory,
                req_description=str(self.req),
            )
        except LegacyInstallFailure as exc:
            self.install_succeeded = False
            raise exc.__cause__
        except Exception:
            self.install_succeeded = True
            raise

        self.install_succeeded = success

        if success and self.legacy_install_reason == 8368:
            deprecated(
                reason=(
                    "{} was installed using the legacy 'setup.py install' "
                    "method, because a wheel could not be built for it.".format(
                        self.name
                    )
                ),
                replacement="to fix the wheel build issue reported above",
                gone_in=None,
                issue=8368,
            )


def check_invalid_constraint_type(req: InstallRequirement) -> str:

    # Check for unsupported forms
    problem = ""
    if not req.name:
        problem = "Unnamed requirements are not allowed as constraints"
    elif req.editable:
        problem = "Editable requirements are not allowed as constraints"
    elif req.extras:
        problem = "Constraints cannot have extras"

    if problem:
        deprecated(
            reason=(
                "Constraints are only allowed to take the form of a package "
                "name and a version specifier. Other forms were originally "
                "permitted as an accident of the implementation, but were "
                "undocumented. The new implementation of the resolver no "
                "longer supports these forms."
            ),
            replacement="replacing the constraint with a requirement",
            # No plan yet for when the new resolver becomes default
            gone_in=None,
            issue=8210,
        )

    return problem
site-packages/pip/_internal/req/req_tracker.py000064400000010025150564745010015477 0ustar00import contextlib
import hashlib
import logging
import os
from types import TracebackType
from typing import Dict, Iterator, Optional, Set, Type, Union

from pip._internal.models.link import Link
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils.temp_dir import TempDirectory

logger = logging.getLogger(__name__)


@contextlib.contextmanager
def update_env_context_manager(**changes: str) -> Iterator[None]:
    target = os.environ

    # Save values from the target and change them.
    non_existent_marker = object()
    saved_values: Dict[str, Union[object, str]] = {}
    for name, new_value in changes.items():
        try:
            saved_values[name] = target[name]
        except KeyError:
            saved_values[name] = non_existent_marker
        target[name] = new_value

    try:
        yield
    finally:
        # Restore original values in the target.
        for name, original_value in saved_values.items():
            if original_value is non_existent_marker:
                del target[name]
            else:
                assert isinstance(original_value, str)  # for mypy
                target[name] = original_value


@contextlib.contextmanager
def get_requirement_tracker() -> Iterator["RequirementTracker"]:
    root = os.environ.get("PIP_REQ_TRACKER")
    with contextlib.ExitStack() as ctx:
        if root is None:
            root = ctx.enter_context(TempDirectory(kind="req-tracker")).path
            ctx.enter_context(update_env_context_manager(PIP_REQ_TRACKER=root))
            logger.debug("Initialized build tracking at %s", root)

        with RequirementTracker(root) as tracker:
            yield tracker


class RequirementTracker:
    def __init__(self, root: str) -> None:
        self._root = root
        self._entries: Set[InstallRequirement] = set()
        logger.debug("Created build tracker: %s", self._root)

    def __enter__(self) -> "RequirementTracker":
        logger.debug("Entered build tracker: %s", self._root)
        return self

    def __exit__(
        self,
        exc_type: Optional[Type[BaseException]],
        exc_val: Optional[BaseException],
        exc_tb: Optional[TracebackType],
    ) -> None:
        self.cleanup()

    def _entry_path(self, link: Link) -> str:
        hashed = hashlib.sha224(link.url_without_fragment.encode()).hexdigest()
        return os.path.join(self._root, hashed)

    def add(self, req: InstallRequirement) -> None:
        """Add an InstallRequirement to build tracking."""

        assert req.link
        # Get the file to write information about this requirement.
        entry_path = self._entry_path(req.link)

        # Try reading from the file. If it exists and can be read from, a build
        # is already in progress, so a LookupError is raised.
        try:
            with open(entry_path) as fp:
                contents = fp.read()
        except FileNotFoundError:
            pass
        else:
            message = "{} is already being built: {}".format(req.link, contents)
            raise LookupError(message)

        # If we're here, req should really not be building already.
        assert req not in self._entries

        # Start tracking this requirement.
        with open(entry_path, "w", encoding="utf-8") as fp:
            fp.write(str(req))
        self._entries.add(req)

        logger.debug("Added %s to build tracker %r", req, self._root)

    def remove(self, req: InstallRequirement) -> None:
        """Remove an InstallRequirement from build tracking."""

        assert req.link
        # Delete the created file and the corresponding entries.
        os.unlink(self._entry_path(req.link))
        self._entries.remove(req)

        logger.debug("Removed %s from build tracker %r", req, self._root)

    def cleanup(self) -> None:
        for req in set(self._entries):
            self.remove(req)

        logger.debug("Removed build tracker: %r", self._root)

    @contextlib.contextmanager
    def track(self, req: InstallRequirement) -> Iterator[None]:
        self.add(req)
        yield
        self.remove(req)
site-packages/pip/_internal/req/__init__.py000064400000005351150564745010014742 0ustar00import collections
import logging
from typing import Iterator, List, Optional, Sequence, Tuple

from pip._internal.utils.logging import indent_log

from .req_file import parse_requirements
from .req_install import InstallRequirement
from .req_set import RequirementSet

__all__ = [
    "RequirementSet",
    "InstallRequirement",
    "parse_requirements",
    "install_given_reqs",
]

logger = logging.getLogger(__name__)


class InstallationResult:
    def __init__(self, name: str) -> None:
        self.name = name

    def __repr__(self) -> str:
        return f"InstallationResult(name={self.name!r})"


def _validate_requirements(
    requirements: List[InstallRequirement],
) -> Iterator[Tuple[str, InstallRequirement]]:
    for req in requirements:
        assert req.name, f"invalid to-be-installed requirement: {req}"
        yield req.name, req


def install_given_reqs(
    requirements: List[InstallRequirement],
    install_options: List[str],
    global_options: Sequence[str],
    root: Optional[str],
    home: Optional[str],
    prefix: Optional[str],
    warn_script_location: bool,
    use_user_site: bool,
    pycompile: bool,
) -> List[InstallationResult]:
    """
    Install everything in the given list.

    (to be called after having downloaded and unpacked the packages)
    """
    to_install = collections.OrderedDict(_validate_requirements(requirements))

    if to_install:
        logger.info(
            "Installing collected packages: %s",
            ", ".join(to_install.keys()),
        )

    installed = []

    with indent_log():
        for req_name, requirement in to_install.items():
            if requirement.should_reinstall:
                logger.info("Attempting uninstall: %s", req_name)
                with indent_log():
                    uninstalled_pathset = requirement.uninstall(auto_confirm=True)
            else:
                uninstalled_pathset = None

            try:
                requirement.install(
                    install_options,
                    global_options,
                    root=root,
                    home=home,
                    prefix=prefix,
                    warn_script_location=warn_script_location,
                    use_user_site=use_user_site,
                    pycompile=pycompile,
                )
            except Exception:
                # if install did not succeed, rollback previous uninstall
                if uninstalled_pathset and not requirement.install_succeeded:
                    uninstalled_pathset.rollback()
                raise
            else:
                if uninstalled_pathset and requirement.install_succeeded:
                    uninstalled_pathset.commit()

            installed.append(InstallationResult(req_name))

    return installed
site-packages/pip/_internal/req/constructors.py000064400000035665150564745010015766 0ustar00"""Backing implementation for InstallRequirement's various constructors

The idea here is that these formed a major chunk of InstallRequirement's size
so, moving them and support code dedicated to them outside of that class
helps creates for better understandability for the rest of the code.

These are meant to be used elsewhere within pip to create instances of
InstallRequirement.
"""

import logging
import os
import re
from typing import Any, Dict, Optional, Set, Tuple, Union

from pip._vendor.packaging.markers import Marker
from pip._vendor.packaging.requirements import InvalidRequirement, Requirement
from pip._vendor.packaging.specifiers import Specifier
from pip._vendor.pkg_resources import RequirementParseError, parse_requirements

from pip._internal.exceptions import InstallationError
from pip._internal.models.index import PyPI, TestPyPI
from pip._internal.models.link import Link
from pip._internal.models.wheel import Wheel
from pip._internal.req.req_file import ParsedRequirement
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils.filetypes import is_archive_file
from pip._internal.utils.misc import is_installable_dir
from pip._internal.utils.packaging import get_requirement
from pip._internal.utils.urls import path_to_url
from pip._internal.vcs import is_url, vcs

__all__ = [
    "install_req_from_editable",
    "install_req_from_line",
    "parse_editable",
]

logger = logging.getLogger(__name__)
operators = Specifier._operators.keys()


def _strip_extras(path: str) -> Tuple[str, Optional[str]]:
    m = re.match(r"^(.+)(\[[^\]]+\])$", path)
    extras = None
    if m:
        path_no_extras = m.group(1)
        extras = m.group(2)
    else:
        path_no_extras = path

    return path_no_extras, extras


def convert_extras(extras: Optional[str]) -> Set[str]:
    if not extras:
        return set()
    return get_requirement("placeholder" + extras.lower()).extras


def parse_editable(editable_req: str) -> Tuple[Optional[str], str, Set[str]]:
    """Parses an editable requirement into:
        - a requirement name
        - an URL
        - extras
        - editable options
    Accepted requirements:
        svn+http://blahblah@rev#egg=Foobar[baz]&subdirectory=version_subdir
        .[some_extra]
    """

    url = editable_req

    # If a file path is specified with extras, strip off the extras.
    url_no_extras, extras = _strip_extras(url)

    if os.path.isdir(url_no_extras):
        # Treating it as code that has already been checked out
        url_no_extras = path_to_url(url_no_extras)

    if url_no_extras.lower().startswith("file:"):
        package_name = Link(url_no_extras).egg_fragment
        if extras:
            return (
                package_name,
                url_no_extras,
                get_requirement("placeholder" + extras.lower()).extras,
            )
        else:
            return package_name, url_no_extras, set()

    for version_control in vcs:
        if url.lower().startswith(f"{version_control}:"):
            url = f"{version_control}+{url}"
            break

    link = Link(url)

    if not link.is_vcs:
        backends = ", ".join(vcs.all_schemes)
        raise InstallationError(
            f"{editable_req} is not a valid editable requirement. "
            f"It should either be a path to a local project or a VCS URL "
            f"(beginning with {backends})."
        )

    package_name = link.egg_fragment
    if not package_name:
        raise InstallationError(
            "Could not detect requirement name for '{}', please specify one "
            "with #egg=your_package_name".format(editable_req)
        )
    return package_name, url, set()


def deduce_helpful_msg(req: str) -> str:
    """Returns helpful msg in case requirements file does not exist,
    or cannot be parsed.

    :params req: Requirements file path
    """
    msg = ""
    if os.path.exists(req):
        msg = " The path does exist. "
        # Try to parse and check if it is a requirements file.
        try:
            with open(req) as fp:
                # parse first line only
                next(parse_requirements(fp.read()))
                msg += (
                    "The argument you provided "
                    "({}) appears to be a"
                    " requirements file. If that is the"
                    " case, use the '-r' flag to install"
                    " the packages specified within it."
                ).format(req)
        except RequirementParseError:
            logger.debug("Cannot parse '%s' as requirements file", req, exc_info=True)
    else:
        msg += f" File '{req}' does not exist."
    return msg


class RequirementParts:
    def __init__(
        self,
        requirement: Optional[Requirement],
        link: Optional[Link],
        markers: Optional[Marker],
        extras: Set[str],
    ):
        self.requirement = requirement
        self.link = link
        self.markers = markers
        self.extras = extras


def parse_req_from_editable(editable_req: str) -> RequirementParts:
    name, url, extras_override = parse_editable(editable_req)

    if name is not None:
        try:
            req: Optional[Requirement] = Requirement(name)
        except InvalidRequirement:
            raise InstallationError(f"Invalid requirement: '{name}'")
    else:
        req = None

    link = Link(url)

    return RequirementParts(req, link, None, extras_override)


# ---- The actual constructors follow ----


def install_req_from_editable(
    editable_req: str,
    comes_from: Optional[Union[InstallRequirement, str]] = None,
    use_pep517: Optional[bool] = None,
    isolated: bool = False,
    options: Optional[Dict[str, Any]] = None,
    constraint: bool = False,
    user_supplied: bool = False,
    permit_editable_wheels: bool = False,
) -> InstallRequirement:

    parts = parse_req_from_editable(editable_req)

    return InstallRequirement(
        parts.requirement,
        comes_from=comes_from,
        user_supplied=user_supplied,
        editable=True,
        permit_editable_wheels=permit_editable_wheels,
        link=parts.link,
        constraint=constraint,
        use_pep517=use_pep517,
        isolated=isolated,
        install_options=options.get("install_options", []) if options else [],
        global_options=options.get("global_options", []) if options else [],
        hash_options=options.get("hashes", {}) if options else {},
        extras=parts.extras,
    )


def _looks_like_path(name: str) -> bool:
    """Checks whether the string "looks like" a path on the filesystem.

    This does not check whether the target actually exists, only judge from the
    appearance.

    Returns true if any of the following conditions is true:
    * a path separator is found (either os.path.sep or os.path.altsep);
    * a dot is found (which represents the current directory).
    """
    if os.path.sep in name:
        return True
    if os.path.altsep is not None and os.path.altsep in name:
        return True
    if name.startswith("."):
        return True
    return False


def _get_url_from_path(path: str, name: str) -> Optional[str]:
    """
    First, it checks whether a provided path is an installable directory. If it
    is, returns the path.

    If false, check if the path is an archive file (such as a .whl).
    The function checks if the path is a file. If false, if the path has
    an @, it will treat it as a PEP 440 URL requirement and return the path.
    """
    if _looks_like_path(name) and os.path.isdir(path):
        if is_installable_dir(path):
            return path_to_url(path)
        # TODO: The is_installable_dir test here might not be necessary
        #       now that it is done in load_pyproject_toml too.
        raise InstallationError(
            f"Directory {name!r} is not installable. Neither 'setup.py' "
            "nor 'pyproject.toml' found."
        )
    if not is_archive_file(path):
        return None
    if os.path.isfile(path):
        return path_to_url(path)
    urlreq_parts = name.split("@", 1)
    if len(urlreq_parts) >= 2 and not _looks_like_path(urlreq_parts[0]):
        # If the path contains '@' and the part before it does not look
        # like a path, try to treat it as a PEP 440 URL req instead.
        return None
    logger.warning(
        "Requirement %r looks like a filename, but the file does not exist",
        name,
    )
    return path_to_url(path)


def parse_req_from_line(name: str, line_source: Optional[str]) -> RequirementParts:
    if is_url(name):
        marker_sep = "; "
    else:
        marker_sep = ";"
    if marker_sep in name:
        name, markers_as_string = name.split(marker_sep, 1)
        markers_as_string = markers_as_string.strip()
        if not markers_as_string:
            markers = None
        else:
            markers = Marker(markers_as_string)
    else:
        markers = None
    name = name.strip()
    req_as_string = None
    path = os.path.normpath(os.path.abspath(name))
    link = None
    extras_as_string = None

    if is_url(name):
        link = Link(name)
    else:
        p, extras_as_string = _strip_extras(path)
        url = _get_url_from_path(p, name)
        if url is not None:
            link = Link(url)

    # it's a local file, dir, or url
    if link:
        # Handle relative file URLs
        if link.scheme == "file" and re.search(r"\.\./", link.url):
            link = Link(path_to_url(os.path.normpath(os.path.abspath(link.path))))
        # wheel file
        if link.is_wheel:
            wheel = Wheel(link.filename)  # can raise InvalidWheelFilename
            req_as_string = f"{wheel.name}=={wheel.version}"
        else:
            # set the req to the egg fragment.  when it's not there, this
            # will become an 'unnamed' requirement
            req_as_string = link.egg_fragment

    # a requirement specifier
    else:
        req_as_string = name

    extras = convert_extras(extras_as_string)

    def with_source(text: str) -> str:
        if not line_source:
            return text
        return f"{text} (from {line_source})"

    def _parse_req_string(req_as_string: str) -> Requirement:
        try:
            req = get_requirement(req_as_string)
        except InvalidRequirement:
            if os.path.sep in req_as_string:
                add_msg = "It looks like a path."
                add_msg += deduce_helpful_msg(req_as_string)
            elif "=" in req_as_string and not any(
                op in req_as_string for op in operators
            ):
                add_msg = "= is not a valid operator. Did you mean == ?"
            else:
                add_msg = ""
            msg = with_source(f"Invalid requirement: {req_as_string!r}")
            if add_msg:
                msg += f"\nHint: {add_msg}"
            raise InstallationError(msg)
        else:
            # Deprecate extras after specifiers: "name>=1.0[extras]"
            # This currently works by accident because _strip_extras() parses
            # any extras in the end of the string and those are saved in
            # RequirementParts
            for spec in req.specifier:
                spec_str = str(spec)
                if spec_str.endswith("]"):
                    msg = f"Extras after version '{spec_str}'."
                    raise InstallationError(msg)
        return req

    if req_as_string is not None:
        req: Optional[Requirement] = _parse_req_string(req_as_string)
    else:
        req = None

    return RequirementParts(req, link, markers, extras)


def install_req_from_line(
    name: str,
    comes_from: Optional[Union[str, InstallRequirement]] = None,
    use_pep517: Optional[bool] = None,
    isolated: bool = False,
    options: Optional[Dict[str, Any]] = None,
    constraint: bool = False,
    line_source: Optional[str] = None,
    user_supplied: bool = False,
) -> InstallRequirement:
    """Creates an InstallRequirement from a name, which might be a
    requirement, directory containing 'setup.py', filename, or URL.

    :param line_source: An optional string describing where the line is from,
        for logging purposes in case of an error.
    """
    parts = parse_req_from_line(name, line_source)

    return InstallRequirement(
        parts.requirement,
        comes_from,
        link=parts.link,
        markers=parts.markers,
        use_pep517=use_pep517,
        isolated=isolated,
        install_options=options.get("install_options", []) if options else [],
        global_options=options.get("global_options", []) if options else [],
        hash_options=options.get("hashes", {}) if options else {},
        constraint=constraint,
        extras=parts.extras,
        user_supplied=user_supplied,
    )


def install_req_from_req_string(
    req_string: str,
    comes_from: Optional[InstallRequirement] = None,
    isolated: bool = False,
    use_pep517: Optional[bool] = None,
    user_supplied: bool = False,
) -> InstallRequirement:
    try:
        req = get_requirement(req_string)
    except InvalidRequirement:
        raise InstallationError(f"Invalid requirement: '{req_string}'")

    domains_not_allowed = [
        PyPI.file_storage_domain,
        TestPyPI.file_storage_domain,
    ]
    if (
        req.url
        and comes_from
        and comes_from.link
        and comes_from.link.netloc in domains_not_allowed
    ):
        # Explicitly disallow pypi packages that depend on external urls
        raise InstallationError(
            "Packages installed from PyPI cannot depend on packages "
            "which are not also hosted on PyPI.\n"
            "{} depends on {} ".format(comes_from.name, req)
        )

    return InstallRequirement(
        req,
        comes_from,
        isolated=isolated,
        use_pep517=use_pep517,
        user_supplied=user_supplied,
    )


def install_req_from_parsed_requirement(
    parsed_req: ParsedRequirement,
    isolated: bool = False,
    use_pep517: Optional[bool] = None,
    user_supplied: bool = False,
) -> InstallRequirement:
    if parsed_req.is_editable:
        req = install_req_from_editable(
            parsed_req.requirement,
            comes_from=parsed_req.comes_from,
            use_pep517=use_pep517,
            constraint=parsed_req.constraint,
            isolated=isolated,
            user_supplied=user_supplied,
        )

    else:
        req = install_req_from_line(
            parsed_req.requirement,
            comes_from=parsed_req.comes_from,
            use_pep517=use_pep517,
            isolated=isolated,
            options=parsed_req.options,
            constraint=parsed_req.constraint,
            line_source=parsed_req.line_source,
            user_supplied=user_supplied,
        )
    return req


def install_req_from_link_and_ireq(
    link: Link, ireq: InstallRequirement
) -> InstallRequirement:
    return InstallRequirement(
        req=ireq.req,
        comes_from=ireq.comes_from,
        editable=ireq.editable,
        link=link,
        markers=ireq.markers,
        use_pep517=ireq.use_pep517,
        isolated=ireq.isolated,
        install_options=ireq.install_options,
        global_options=ireq.global_options,
        hash_options=ireq.hash_options,
    )
site-packages/pip/_internal/req/req_uninstall.py000064400000056304150564745010016067 0ustar00import csv
import functools
import os
import sys
import sysconfig
from importlib.util import cache_from_source
from typing import Any, Callable, Dict, Iterable, Iterator, List, Optional, Set, Tuple

from pip._vendor import pkg_resources
from pip._vendor.pkg_resources import Distribution

from pip._internal.exceptions import UninstallationError
from pip._internal.locations import get_bin_prefix, get_bin_user
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.egg_link import egg_link_path_from_location
from pip._internal.utils.logging import getLogger, indent_log
from pip._internal.utils.misc import (
    ask,
    dist_in_usersite,
    dist_is_local,
    is_local,
    normalize_path,
    renames,
    rmtree,
)
from pip._internal.utils.temp_dir import AdjacentTempDirectory, TempDirectory

logger = getLogger(__name__)


def _script_names(dist: Distribution, script_name: str, is_gui: bool) -> List[str]:
    """Create the fully qualified name of the files created by
    {console,gui}_scripts for the given ``dist``.
    Returns the list of file names
    """
    if dist_in_usersite(dist):
        bin_dir = get_bin_user()
    else:
        bin_dir = get_bin_prefix()
    exe_name = os.path.join(bin_dir, script_name)
    paths_to_remove = [exe_name]
    if WINDOWS:
        paths_to_remove.append(exe_name + ".exe")
        paths_to_remove.append(exe_name + ".exe.manifest")
        if is_gui:
            paths_to_remove.append(exe_name + "-script.pyw")
        else:
            paths_to_remove.append(exe_name + "-script.py")
    return paths_to_remove


def _unique(fn: Callable[..., Iterator[Any]]) -> Callable[..., Iterator[Any]]:
    @functools.wraps(fn)
    def unique(*args: Any, **kw: Any) -> Iterator[Any]:
        seen: Set[Any] = set()
        for item in fn(*args, **kw):
            if item not in seen:
                seen.add(item)
                yield item

    return unique


@_unique
def uninstallation_paths(dist: Distribution) -> Iterator[str]:
    """
    Yield all the uninstallation paths for dist based on RECORD-without-.py[co]

    Yield paths to all the files in RECORD. For each .py file in RECORD, add
    the .pyc and .pyo in the same directory.

    UninstallPathSet.add() takes care of the __pycache__ .py[co].

    If RECORD is not found, raises UninstallationError,
    with possible information from the INSTALLER file.

    https://packaging.python.org/specifications/recording-installed-packages/
    """
    try:
        r = csv.reader(dist.get_metadata_lines("RECORD"))
    except FileNotFoundError as missing_record_exception:
        msg = "Cannot uninstall {dist}, RECORD file not found.".format(dist=dist)
        try:
            installer = next(dist.get_metadata_lines("INSTALLER"))
            if not installer or installer == "pip":
                raise ValueError()
        except (OSError, StopIteration, ValueError):
            dep = "{}=={}".format(dist.project_name, dist.version)
            msg += (
                " You might be able to recover from this via: "
                "'pip install --force-reinstall --no-deps {}'.".format(dep)
            )
        else:
            msg += " Hint: The package was installed by {}.".format(installer)
        raise UninstallationError(msg) from missing_record_exception
    for row in r:
        path = os.path.join(dist.location, row[0])
        yield path
        if path.endswith(".py"):
            dn, fn = os.path.split(path)
            base = fn[:-3]
            path = os.path.join(dn, base + ".pyc")
            yield path
            path = os.path.join(dn, base + ".pyo")
            yield path


def compact(paths: Iterable[str]) -> Set[str]:
    """Compact a path set to contain the minimal number of paths
    necessary to contain all paths in the set. If /a/path/ and
    /a/path/to/a/file.txt are both in the set, leave only the
    shorter path."""

    sep = os.path.sep
    short_paths: Set[str] = set()
    for path in sorted(paths, key=len):
        should_skip = any(
            path.startswith(shortpath.rstrip("*"))
            and path[len(shortpath.rstrip("*").rstrip(sep))] == sep
            for shortpath in short_paths
        )
        if not should_skip:
            short_paths.add(path)
    return short_paths


def compress_for_rename(paths: Iterable[str]) -> Set[str]:
    """Returns a set containing the paths that need to be renamed.

    This set may include directories when the original sequence of paths
    included every file on disk.
    """
    case_map = {os.path.normcase(p): p for p in paths}
    remaining = set(case_map)
    unchecked = sorted({os.path.split(p)[0] for p in case_map.values()}, key=len)
    wildcards: Set[str] = set()

    def norm_join(*a: str) -> str:
        return os.path.normcase(os.path.join(*a))

    for root in unchecked:
        if any(os.path.normcase(root).startswith(w) for w in wildcards):
            # This directory has already been handled.
            continue

        all_files: Set[str] = set()
        all_subdirs: Set[str] = set()
        for dirname, subdirs, files in os.walk(root):
            all_subdirs.update(norm_join(root, dirname, d) for d in subdirs)
            all_files.update(norm_join(root, dirname, f) for f in files)
        # If all the files we found are in our remaining set of files to
        # remove, then remove them from the latter set and add a wildcard
        # for the directory.
        if not (all_files - remaining):
            remaining.difference_update(all_files)
            wildcards.add(root + os.sep)

    return set(map(case_map.__getitem__, remaining)) | wildcards


def compress_for_output_listing(paths: Iterable[str]) -> Tuple[Set[str], Set[str]]:
    """Returns a tuple of 2 sets of which paths to display to user

    The first set contains paths that would be deleted. Files of a package
    are not added and the top-level directory of the package has a '*' added
    at the end - to signify that all it's contents are removed.

    The second set contains files that would have been skipped in the above
    folders.
    """

    will_remove = set(paths)
    will_skip = set()

    # Determine folders and files
    folders = set()
    files = set()
    for path in will_remove:
        if path.endswith(".pyc"):
            continue
        if path.endswith("__init__.py") or ".dist-info" in path:
            folders.add(os.path.dirname(path))
        files.add(path)

    # probably this one https://github.com/python/mypy/issues/390
    _normcased_files = set(map(os.path.normcase, files))  # type: ignore

    folders = compact(folders)

    # This walks the tree using os.walk to not miss extra folders
    # that might get added.
    for folder in folders:
        for dirpath, _, dirfiles in os.walk(folder):
            for fname in dirfiles:
                if fname.endswith(".pyc"):
                    continue

                file_ = os.path.join(dirpath, fname)
                if (
                    os.path.isfile(file_)
                    and os.path.normcase(file_) not in _normcased_files
                ):
                    # We are skipping this file. Add it to the set.
                    will_skip.add(file_)

    will_remove = files | {os.path.join(folder, "*") for folder in folders}

    return will_remove, will_skip


class StashedUninstallPathSet:
    """A set of file rename operations to stash files while
    tentatively uninstalling them."""

    def __init__(self) -> None:
        # Mapping from source file root to [Adjacent]TempDirectory
        # for files under that directory.
        self._save_dirs: Dict[str, TempDirectory] = {}
        # (old path, new path) tuples for each move that may need
        # to be undone.
        self._moves: List[Tuple[str, str]] = []

    def _get_directory_stash(self, path: str) -> str:
        """Stashes a directory.

        Directories are stashed adjacent to their original location if
        possible, or else moved/copied into the user's temp dir."""

        try:
            save_dir: TempDirectory = AdjacentTempDirectory(path)
        except OSError:
            save_dir = TempDirectory(kind="uninstall")
        self._save_dirs[os.path.normcase(path)] = save_dir

        return save_dir.path

    def _get_file_stash(self, path: str) -> str:
        """Stashes a file.

        If no root has been provided, one will be created for the directory
        in the user's temp directory."""
        path = os.path.normcase(path)
        head, old_head = os.path.dirname(path), None
        save_dir = None

        while head != old_head:
            try:
                save_dir = self._save_dirs[head]
                break
            except KeyError:
                pass
            head, old_head = os.path.dirname(head), head
        else:
            # Did not find any suitable root
            head = os.path.dirname(path)
            save_dir = TempDirectory(kind="uninstall")
            self._save_dirs[head] = save_dir

        relpath = os.path.relpath(path, head)
        if relpath and relpath != os.path.curdir:
            return os.path.join(save_dir.path, relpath)
        return save_dir.path

    def stash(self, path: str) -> str:
        """Stashes the directory or file and returns its new location.
        Handle symlinks as files to avoid modifying the symlink targets.
        """
        path_is_dir = os.path.isdir(path) and not os.path.islink(path)
        if path_is_dir:
            new_path = self._get_directory_stash(path)
        else:
            new_path = self._get_file_stash(path)

        self._moves.append((path, new_path))
        if path_is_dir and os.path.isdir(new_path):
            # If we're moving a directory, we need to
            # remove the destination first or else it will be
            # moved to inside the existing directory.
            # We just created new_path ourselves, so it will
            # be removable.
            os.rmdir(new_path)
        renames(path, new_path)
        return new_path

    def commit(self) -> None:
        """Commits the uninstall by removing stashed files."""
        for _, save_dir in self._save_dirs.items():
            save_dir.cleanup()
        self._moves = []
        self._save_dirs = {}

    def rollback(self) -> None:
        """Undoes the uninstall by moving stashed files back."""
        for p in self._moves:
            logger.info("Moving to %s\n from %s", *p)

        for new_path, path in self._moves:
            try:
                logger.debug("Replacing %s from %s", new_path, path)
                if os.path.isfile(new_path) or os.path.islink(new_path):
                    os.unlink(new_path)
                elif os.path.isdir(new_path):
                    rmtree(new_path)
                renames(path, new_path)
            except OSError as ex:
                logger.error("Failed to restore %s", new_path)
                logger.debug("Exception: %s", ex)

        self.commit()

    @property
    def can_rollback(self) -> bool:
        return bool(self._moves)


class UninstallPathSet:
    """A set of file paths to be removed in the uninstallation of a
    requirement."""

    def __init__(self, dist: Distribution) -> None:
        self.paths: Set[str] = set()
        self._refuse: Set[str] = set()
        self.pth: Dict[str, UninstallPthEntries] = {}
        self.dist = dist
        self._moved_paths = StashedUninstallPathSet()

    def _permitted(self, path: str) -> bool:
        """
        Return True if the given path is one we are permitted to
        remove/modify, False otherwise.

        """
        return is_local(path)

    def add(self, path: str) -> None:
        head, tail = os.path.split(path)

        # we normalize the head to resolve parent directory symlinks, but not
        # the tail, since we only want to uninstall symlinks, not their targets
        path = os.path.join(normalize_path(head), os.path.normcase(tail))

        if not os.path.exists(path):
            return
        if self._permitted(path):
            self.paths.add(path)
        else:
            self._refuse.add(path)

        # __pycache__ files can show up after 'installed-files.txt' is created,
        # due to imports
        if os.path.splitext(path)[1] == ".py":
            self.add(cache_from_source(path))

    def add_pth(self, pth_file: str, entry: str) -> None:
        pth_file = normalize_path(pth_file)
        if self._permitted(pth_file):
            if pth_file not in self.pth:
                self.pth[pth_file] = UninstallPthEntries(pth_file)
            self.pth[pth_file].add(entry)
        else:
            self._refuse.add(pth_file)

    def remove(self, auto_confirm: bool = False, verbose: bool = False) -> None:
        """Remove paths in ``self.paths`` with confirmation (unless
        ``auto_confirm`` is True)."""

        if not self.paths:
            logger.info(
                "Can't uninstall '%s'. No files were found to uninstall.",
                self.dist.project_name,
            )
            return

        dist_name_version = self.dist.project_name + "-" + self.dist.version
        logger.info("Uninstalling %s:", dist_name_version)

        with indent_log():
            if auto_confirm or self._allowed_to_proceed(verbose):
                moved = self._moved_paths

                for_rename = compress_for_rename(self.paths)

                for path in sorted(compact(for_rename)):
                    moved.stash(path)
                    logger.verbose("Removing file or directory %s", path)

                for pth in self.pth.values():
                    pth.remove()

                logger.info("Successfully uninstalled %s", dist_name_version)

    def _allowed_to_proceed(self, verbose: bool) -> bool:
        """Display which files would be deleted and prompt for confirmation"""

        def _display(msg: str, paths: Iterable[str]) -> None:
            if not paths:
                return

            logger.info(msg)
            with indent_log():
                for path in sorted(compact(paths)):
                    logger.info(path)

        if not verbose:
            will_remove, will_skip = compress_for_output_listing(self.paths)
        else:
            # In verbose mode, display all the files that are going to be
            # deleted.
            will_remove = set(self.paths)
            will_skip = set()

        _display("Would remove:", will_remove)
        _display("Would not remove (might be manually added):", will_skip)
        _display("Would not remove (outside of prefix):", self._refuse)
        if verbose:
            _display("Will actually move:", compress_for_rename(self.paths))

        return ask("Proceed (Y/n)? ", ("y", "n", "")) != "n"

    def rollback(self) -> None:
        """Rollback the changes previously made by remove()."""
        if not self._moved_paths.can_rollback:
            logger.error(
                "Can't roll back %s; was not uninstalled",
                self.dist.project_name,
            )
            return
        logger.info("Rolling back uninstall of %s", self.dist.project_name)
        self._moved_paths.rollback()
        for pth in self.pth.values():
            pth.rollback()

    def commit(self) -> None:
        """Remove temporary save dir: rollback will no longer be possible."""
        self._moved_paths.commit()

    @classmethod
    def from_dist(cls, dist: Distribution) -> "UninstallPathSet":
        dist_path = normalize_path(dist.location)
        if not dist_is_local(dist):
            logger.info(
                "Not uninstalling %s at %s, outside environment %s",
                dist.key,
                dist_path,
                sys.prefix,
            )
            return cls(dist)

        if dist_path in {
            p
            for p in {sysconfig.get_path("stdlib"), sysconfig.get_path("platstdlib")}
            if p
        }:
            logger.info(
                "Not uninstalling %s at %s, as it is in the standard library.",
                dist.key,
                dist_path,
            )
            return cls(dist)

        paths_to_remove = cls(dist)
        develop_egg_link = egg_link_path_from_location(dist.project_name)
        develop_egg_link_egg_info = "{}.egg-info".format(
            pkg_resources.to_filename(dist.project_name)
        )
        egg_info_exists = dist.egg_info and os.path.exists(dist.egg_info)
        # Special case for distutils installed package
        distutils_egg_info = getattr(dist._provider, "path", None)

        # Uninstall cases order do matter as in the case of 2 installs of the
        # same package, pip needs to uninstall the currently detected version
        if (
            egg_info_exists
            and dist.egg_info.endswith(".egg-info")
            and not dist.egg_info.endswith(develop_egg_link_egg_info)
        ):
            # if dist.egg_info.endswith(develop_egg_link_egg_info), we
            # are in fact in the develop_egg_link case
            paths_to_remove.add(dist.egg_info)
            if dist.has_metadata("installed-files.txt"):
                for installed_file in dist.get_metadata(
                    "installed-files.txt"
                ).splitlines():
                    path = os.path.normpath(os.path.join(dist.egg_info, installed_file))
                    paths_to_remove.add(path)
            # FIXME: need a test for this elif block
            # occurs with --single-version-externally-managed/--record outside
            # of pip
            elif dist.has_metadata("top_level.txt"):
                if dist.has_metadata("namespace_packages.txt"):
                    namespaces = dist.get_metadata("namespace_packages.txt")
                else:
                    namespaces = []
                for top_level_pkg in [
                    p
                    for p in dist.get_metadata("top_level.txt").splitlines()
                    if p and p not in namespaces
                ]:
                    path = os.path.join(dist.location, top_level_pkg)
                    paths_to_remove.add(path)
                    paths_to_remove.add(path + ".py")
                    paths_to_remove.add(path + ".pyc")
                    paths_to_remove.add(path + ".pyo")

        elif distutils_egg_info:
            raise UninstallationError(
                "Cannot uninstall {!r}. It is a distutils installed project "
                "and thus we cannot accurately determine which files belong "
                "to it which would lead to only a partial uninstall.".format(
                    dist.project_name,
                )
            )

        elif dist.location.endswith(".egg"):
            # package installed by easy_install
            # We cannot match on dist.egg_name because it can slightly vary
            # i.e. setuptools-0.6c11-py2.6.egg vs setuptools-0.6rc11-py2.6.egg
            paths_to_remove.add(dist.location)
            easy_install_egg = os.path.split(dist.location)[1]
            easy_install_pth = os.path.join(
                os.path.dirname(dist.location), "easy-install.pth"
            )
            paths_to_remove.add_pth(easy_install_pth, "./" + easy_install_egg)

        elif egg_info_exists and dist.egg_info.endswith(".dist-info"):
            for path in uninstallation_paths(dist):
                paths_to_remove.add(path)

        elif develop_egg_link:
            # develop egg
            with open(develop_egg_link) as fh:
                link_pointer = os.path.normcase(fh.readline().strip())
            assert (
                link_pointer == dist.location
            ), "Egg-link {} does not match installed location of {} (at {})".format(
                link_pointer, dist.project_name, dist.location
            )
            paths_to_remove.add(develop_egg_link)
            easy_install_pth = os.path.join(
                os.path.dirname(develop_egg_link), "easy-install.pth"
            )
            paths_to_remove.add_pth(easy_install_pth, dist.location)

        else:
            logger.debug(
                "Not sure how to uninstall: %s - Check: %s",
                dist,
                dist.location,
            )

        # find distutils scripts= scripts
        if dist.has_metadata("scripts") and dist.metadata_isdir("scripts"):
            for script in dist.metadata_listdir("scripts"):
                if dist_in_usersite(dist):
                    bin_dir = get_bin_user()
                else:
                    bin_dir = get_bin_prefix()
                paths_to_remove.add(os.path.join(bin_dir, script))
                if WINDOWS:
                    paths_to_remove.add(os.path.join(bin_dir, script) + ".bat")

        # find console_scripts
        _scripts_to_remove = []
        console_scripts = dist.get_entry_map(group="console_scripts")
        for name in console_scripts.keys():
            _scripts_to_remove.extend(_script_names(dist, name, False))
        # find gui_scripts
        gui_scripts = dist.get_entry_map(group="gui_scripts")
        for name in gui_scripts.keys():
            _scripts_to_remove.extend(_script_names(dist, name, True))

        for s in _scripts_to_remove:
            paths_to_remove.add(s)

        return paths_to_remove


class UninstallPthEntries:
    def __init__(self, pth_file: str) -> None:
        self.file = pth_file
        self.entries: Set[str] = set()
        self._saved_lines: Optional[List[bytes]] = None

    def add(self, entry: str) -> None:
        entry = os.path.normcase(entry)
        # On Windows, os.path.normcase converts the entry to use
        # backslashes.  This is correct for entries that describe absolute
        # paths outside of site-packages, but all the others use forward
        # slashes.
        # os.path.splitdrive is used instead of os.path.isabs because isabs
        # treats non-absolute paths with drive letter markings like c:foo\bar
        # as absolute paths. It also does not recognize UNC paths if they don't
        # have more than "\\sever\share". Valid examples: "\\server\share\" or
        # "\\server\share\folder".
        if WINDOWS and not os.path.splitdrive(entry)[0]:
            entry = entry.replace("\\", "/")
        self.entries.add(entry)

    def remove(self) -> None:
        logger.verbose("Removing pth entries from %s:", self.file)

        # If the file doesn't exist, log a warning and return
        if not os.path.isfile(self.file):
            logger.warning("Cannot remove entries from nonexistent file %s", self.file)
            return
        with open(self.file, "rb") as fh:
            # windows uses '\r\n' with py3k, but uses '\n' with py2.x
            lines = fh.readlines()
            self._saved_lines = lines
        if any(b"\r\n" in line for line in lines):
            endline = "\r\n"
        else:
            endline = "\n"
        # handle missing trailing newline
        if lines and not lines[-1].endswith(endline.encode("utf-8")):
            lines[-1] = lines[-1] + endline.encode("utf-8")
        for entry in self.entries:
            try:
                logger.verbose("Removing entry: %s", entry)
                lines.remove((entry + endline).encode("utf-8"))
            except ValueError:
                pass
        with open(self.file, "wb") as fh:
            fh.writelines(lines)

    def rollback(self) -> bool:
        if self._saved_lines is None:
            logger.error("Cannot roll back changes to %s, none were made", self.file)
            return False
        logger.debug("Rolling %s back to previous state", self.file)
        with open(self.file, "wb") as fh:
            fh.writelines(self._saved_lines)
        return True
site-packages/pip/_internal/req/req_file.py000064400000042015150564745010014767 0ustar00"""
Requirements file parsing
"""

import optparse
import os
import re
import shlex
import urllib.parse
from optparse import Values
from typing import (
    TYPE_CHECKING,
    Any,
    Callable,
    Dict,
    Iterable,
    Iterator,
    List,
    Optional,
    Tuple,
)

from pip._internal.cli import cmdoptions
from pip._internal.exceptions import InstallationError, RequirementsFileParseError
from pip._internal.models.search_scope import SearchScope
from pip._internal.network.session import PipSession
from pip._internal.network.utils import raise_for_status
from pip._internal.utils.encoding import auto_decode
from pip._internal.utils.urls import get_url_scheme

if TYPE_CHECKING:
    # NoReturn introduced in 3.6.2; imported only for type checking to maintain
    # pip compatibility with older patch versions of Python 3.6
    from typing import NoReturn

    from pip._internal.index.package_finder import PackageFinder

__all__ = ["parse_requirements"]

ReqFileLines = Iterable[Tuple[int, str]]

LineParser = Callable[[str], Tuple[str, Values]]

SCHEME_RE = re.compile(r"^(http|https|file):", re.I)
COMMENT_RE = re.compile(r"(^|\s+)#.*$")

# Matches environment variable-style values in '${MY_VARIABLE_1}' with the
# variable name consisting of only uppercase letters, digits or the '_'
# (underscore). This follows the POSIX standard defined in IEEE Std 1003.1,
# 2013 Edition.
ENV_VAR_RE = re.compile(r"(?P<var>\$\{(?P<name>[A-Z0-9_]+)\})")

SUPPORTED_OPTIONS: List[Callable[..., optparse.Option]] = [
    cmdoptions.index_url,
    cmdoptions.extra_index_url,
    cmdoptions.no_index,
    cmdoptions.constraints,
    cmdoptions.requirements,
    cmdoptions.editable,
    cmdoptions.find_links,
    cmdoptions.no_binary,
    cmdoptions.only_binary,
    cmdoptions.prefer_binary,
    cmdoptions.require_hashes,
    cmdoptions.pre,
    cmdoptions.trusted_host,
    cmdoptions.use_new_feature,
]

# options to be passed to requirements
SUPPORTED_OPTIONS_REQ: List[Callable[..., optparse.Option]] = [
    cmdoptions.install_options,
    cmdoptions.global_options,
    cmdoptions.hash,
]

# the 'dest' string values
SUPPORTED_OPTIONS_REQ_DEST = [str(o().dest) for o in SUPPORTED_OPTIONS_REQ]


class ParsedRequirement:
    def __init__(
        self,
        requirement: str,
        is_editable: bool,
        comes_from: str,
        constraint: bool,
        options: Optional[Dict[str, Any]] = None,
        line_source: Optional[str] = None,
    ) -> None:
        self.requirement = requirement
        self.is_editable = is_editable
        self.comes_from = comes_from
        self.options = options
        self.constraint = constraint
        self.line_source = line_source


class ParsedLine:
    def __init__(
        self,
        filename: str,
        lineno: int,
        args: str,
        opts: Values,
        constraint: bool,
    ) -> None:
        self.filename = filename
        self.lineno = lineno
        self.opts = opts
        self.constraint = constraint

        if args:
            self.is_requirement = True
            self.is_editable = False
            self.requirement = args
        elif opts.editables:
            self.is_requirement = True
            self.is_editable = True
            # We don't support multiple -e on one line
            self.requirement = opts.editables[0]
        else:
            self.is_requirement = False


def parse_requirements(
    filename: str,
    session: PipSession,
    finder: Optional["PackageFinder"] = None,
    options: Optional[optparse.Values] = None,
    constraint: bool = False,
) -> Iterator[ParsedRequirement]:
    """Parse a requirements file and yield ParsedRequirement instances.

    :param filename:    Path or url of requirements file.
    :param session:     PipSession instance.
    :param finder:      Instance of pip.index.PackageFinder.
    :param options:     cli options.
    :param constraint:  If true, parsing a constraint file rather than
        requirements file.
    """
    line_parser = get_line_parser(finder)
    parser = RequirementsFileParser(session, line_parser)

    for parsed_line in parser.parse(filename, constraint):
        parsed_req = handle_line(
            parsed_line, options=options, finder=finder, session=session
        )
        if parsed_req is not None:
            yield parsed_req


def preprocess(content: str) -> ReqFileLines:
    """Split, filter, and join lines, and return a line iterator

    :param content: the content of the requirements file
    """
    lines_enum: ReqFileLines = enumerate(content.splitlines(), start=1)
    lines_enum = join_lines(lines_enum)
    lines_enum = ignore_comments(lines_enum)
    lines_enum = expand_env_variables(lines_enum)
    return lines_enum


def handle_requirement_line(
    line: ParsedLine,
    options: Optional[optparse.Values] = None,
) -> ParsedRequirement:

    # preserve for the nested code path
    line_comes_from = "{} {} (line {})".format(
        "-c" if line.constraint else "-r",
        line.filename,
        line.lineno,
    )

    assert line.is_requirement

    if line.is_editable:
        # For editable requirements, we don't support per-requirement
        # options, so just return the parsed requirement.
        return ParsedRequirement(
            requirement=line.requirement,
            is_editable=line.is_editable,
            comes_from=line_comes_from,
            constraint=line.constraint,
        )
    else:
        if options:
            # Disable wheels if the user has specified build options
            cmdoptions.check_install_build_global(options, line.opts)

        # get the options that apply to requirements
        req_options = {}
        for dest in SUPPORTED_OPTIONS_REQ_DEST:
            if dest in line.opts.__dict__ and line.opts.__dict__[dest]:
                req_options[dest] = line.opts.__dict__[dest]

        line_source = f"line {line.lineno} of {line.filename}"
        return ParsedRequirement(
            requirement=line.requirement,
            is_editable=line.is_editable,
            comes_from=line_comes_from,
            constraint=line.constraint,
            options=req_options,
            line_source=line_source,
        )


def handle_option_line(
    opts: Values,
    filename: str,
    lineno: int,
    finder: Optional["PackageFinder"] = None,
    options: Optional[optparse.Values] = None,
    session: Optional[PipSession] = None,
) -> None:

    if options:
        # percolate options upward
        if opts.require_hashes:
            options.require_hashes = opts.require_hashes
        if opts.features_enabled:
            options.features_enabled.extend(
                f for f in opts.features_enabled if f not in options.features_enabled
            )

    # set finder options
    if finder:
        find_links = finder.find_links
        index_urls = finder.index_urls
        if opts.index_url:
            index_urls = [opts.index_url]
        if opts.no_index is True:
            index_urls = []
        if opts.extra_index_urls:
            index_urls.extend(opts.extra_index_urls)
        if opts.find_links:
            # FIXME: it would be nice to keep track of the source
            # of the find_links: support a find-links local path
            # relative to a requirements file.
            value = opts.find_links[0]
            req_dir = os.path.dirname(os.path.abspath(filename))
            relative_to_reqs_file = os.path.join(req_dir, value)
            if os.path.exists(relative_to_reqs_file):
                value = relative_to_reqs_file
            find_links.append(value)

        if session:
            # We need to update the auth urls in session
            session.update_index_urls(index_urls)

        search_scope = SearchScope(
            find_links=find_links,
            index_urls=index_urls,
        )
        finder.search_scope = search_scope

        if opts.pre:
            finder.set_allow_all_prereleases()

        if opts.prefer_binary:
            finder.set_prefer_binary()

        if session:
            for host in opts.trusted_hosts or []:
                source = f"line {lineno} of {filename}"
                session.add_trusted_host(host, source=source)


def handle_line(
    line: ParsedLine,
    options: Optional[optparse.Values] = None,
    finder: Optional["PackageFinder"] = None,
    session: Optional[PipSession] = None,
) -> Optional[ParsedRequirement]:
    """Handle a single parsed requirements line; This can result in
    creating/yielding requirements, or updating the finder.

    :param line:        The parsed line to be processed.
    :param options:     CLI options.
    :param finder:      The finder - updated by non-requirement lines.
    :param session:     The session - updated by non-requirement lines.

    Returns a ParsedRequirement object if the line is a requirement line,
    otherwise returns None.

    For lines that contain requirements, the only options that have an effect
    are from SUPPORTED_OPTIONS_REQ, and they are scoped to the
    requirement. Other options from SUPPORTED_OPTIONS may be present, but are
    ignored.

    For lines that do not contain requirements, the only options that have an
    effect are from SUPPORTED_OPTIONS. Options from SUPPORTED_OPTIONS_REQ may
    be present, but are ignored. These lines may contain multiple options
    (although our docs imply only one is supported), and all our parsed and
    affect the finder.
    """

    if line.is_requirement:
        parsed_req = handle_requirement_line(line, options)
        return parsed_req
    else:
        handle_option_line(
            line.opts,
            line.filename,
            line.lineno,
            finder,
            options,
            session,
        )
        return None


class RequirementsFileParser:
    def __init__(
        self,
        session: PipSession,
        line_parser: LineParser,
    ) -> None:
        self._session = session
        self._line_parser = line_parser

    def parse(self, filename: str, constraint: bool) -> Iterator[ParsedLine]:
        """Parse a given file, yielding parsed lines."""
        yield from self._parse_and_recurse(filename, constraint)

    def _parse_and_recurse(
        self, filename: str, constraint: bool
    ) -> Iterator[ParsedLine]:
        for line in self._parse_file(filename, constraint):
            if not line.is_requirement and (
                line.opts.requirements or line.opts.constraints
            ):
                # parse a nested requirements file
                if line.opts.requirements:
                    req_path = line.opts.requirements[0]
                    nested_constraint = False
                else:
                    req_path = line.opts.constraints[0]
                    nested_constraint = True

                # original file is over http
                if SCHEME_RE.search(filename):
                    # do a url join so relative paths work
                    req_path = urllib.parse.urljoin(filename, req_path)
                # original file and nested file are paths
                elif not SCHEME_RE.search(req_path):
                    # do a join so relative paths work
                    req_path = os.path.join(
                        os.path.dirname(filename),
                        req_path,
                    )

                yield from self._parse_and_recurse(req_path, nested_constraint)
            else:
                yield line

    def _parse_file(self, filename: str, constraint: bool) -> Iterator[ParsedLine]:
        _, content = get_file_content(filename, self._session)

        lines_enum = preprocess(content)

        for line_number, line in lines_enum:
            try:
                args_str, opts = self._line_parser(line)
            except OptionParsingError as e:
                # add offending line
                msg = f"Invalid requirement: {line}\n{e.msg}"
                raise RequirementsFileParseError(msg)

            yield ParsedLine(
                filename,
                line_number,
                args_str,
                opts,
                constraint,
            )


def get_line_parser(finder: Optional["PackageFinder"]) -> LineParser:
    def parse_line(line: str) -> Tuple[str, Values]:
        # Build new parser for each line since it accumulates appendable
        # options.
        parser = build_parser()
        defaults = parser.get_default_values()
        defaults.index_url = None
        if finder:
            defaults.format_control = finder.format_control

        args_str, options_str = break_args_options(line)

        opts, _ = parser.parse_args(shlex.split(options_str), defaults)

        return args_str, opts

    return parse_line


def break_args_options(line: str) -> Tuple[str, str]:
    """Break up the line into an args and options string.  We only want to shlex
    (and then optparse) the options, not the args.  args can contain markers
    which are corrupted by shlex.
    """
    tokens = line.split(" ")
    args = []
    options = tokens[:]
    for token in tokens:
        if token.startswith("-") or token.startswith("--"):
            break
        else:
            args.append(token)
            options.pop(0)
    return " ".join(args), " ".join(options)


class OptionParsingError(Exception):
    def __init__(self, msg: str) -> None:
        self.msg = msg


def build_parser() -> optparse.OptionParser:
    """
    Return a parser for parsing requirement lines
    """
    parser = optparse.OptionParser(add_help_option=False)

    option_factories = SUPPORTED_OPTIONS + SUPPORTED_OPTIONS_REQ
    for option_factory in option_factories:
        option = option_factory()
        parser.add_option(option)

    # By default optparse sys.exits on parsing errors. We want to wrap
    # that in our own exception.
    def parser_exit(self: Any, msg: str) -> "NoReturn":
        raise OptionParsingError(msg)

    # NOTE: mypy disallows assigning to a method
    #       https://github.com/python/mypy/issues/2427
    parser.exit = parser_exit  # type: ignore

    return parser


def join_lines(lines_enum: ReqFileLines) -> ReqFileLines:
    """Joins a line ending in '\' with the previous line (except when following
    comments).  The joined line takes on the index of the first line.
    """
    primary_line_number = None
    new_line: List[str] = []
    for line_number, line in lines_enum:
        if not line.endswith("\\") or COMMENT_RE.match(line):
            if COMMENT_RE.match(line):
                # this ensures comments are always matched later
                line = " " + line
            if new_line:
                new_line.append(line)
                assert primary_line_number is not None
                yield primary_line_number, "".join(new_line)
                new_line = []
            else:
                yield line_number, line
        else:
            if not new_line:
                primary_line_number = line_number
            new_line.append(line.strip("\\"))

    # last line contains \
    if new_line:
        assert primary_line_number is not None
        yield primary_line_number, "".join(new_line)

    # TODO: handle space after '\'.


def ignore_comments(lines_enum: ReqFileLines) -> ReqFileLines:
    """
    Strips comments and filter empty lines.
    """
    for line_number, line in lines_enum:
        line = COMMENT_RE.sub("", line)
        line = line.strip()
        if line:
            yield line_number, line


def expand_env_variables(lines_enum: ReqFileLines) -> ReqFileLines:
    """Replace all environment variables that can be retrieved via `os.getenv`.

    The only allowed format for environment variables defined in the
    requirement file is `${MY_VARIABLE_1}` to ensure two things:

    1. Strings that contain a `$` aren't accidentally (partially) expanded.
    2. Ensure consistency across platforms for requirement files.

    These points are the result of a discussion on the `github pull
    request #3514 <https://github.com/pypa/pip/pull/3514>`_.

    Valid characters in variable names follow the `POSIX standard
    <http://pubs.opengroup.org/onlinepubs/9699919799/>`_ and are limited
    to uppercase letter, digits and the `_` (underscore).
    """
    for line_number, line in lines_enum:
        for env_var, var_name in ENV_VAR_RE.findall(line):
            value = os.getenv(var_name)
            if not value:
                continue

            line = line.replace(env_var, value)

        yield line_number, line


def get_file_content(url: str, session: PipSession) -> Tuple[str, str]:
    """Gets the content of a file; it may be a filename, file: URL, or
    http: URL.  Returns (location, content).  Content is unicode.
    Respects # -*- coding: declarations on the retrieved files.

    :param url:         File path or url.
    :param session:     PipSession instance.
    """
    scheme = get_url_scheme(url)

    # Pip has special support for file:// URLs (LocalFSAdapter).
    if scheme in ["http", "https", "file"]:
        resp = session.get(url)
        raise_for_status(resp)
        return resp.url, resp.text

    # Assume this is a bare path.
    try:
        with open(url, "rb") as f:
            content = auto_decode(f.read())
    except OSError as exc:
        raise InstallationError(f"Could not open requirements file: {exc}")
    return url, content
site-packages/pip/_internal/__init__.py000064400000001113150564745010014143 0ustar00from typing import List, Optional

import pip._internal.utils.inject_securetransport  # noqa
from pip._internal.utils import _log

# init_logging() must be called before any call to logging.getLogger()
# which happens at import of most modules.
_log.init_logging()


def main(args: (Optional[List[str]]) = None) -> int:
    """This is preserved for old console scripts that may still be referencing
    it.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    """
    from pip._internal.utils.entrypoints import _wrapper

    return _wrapper(args, _nowarn=True)
site-packages/pip/_internal/distributions/__pycache__/base.cpython-39.pyc000064400000003565150564745010022524 0ustar00a

��Re��@sDddlZddlmZddlmZddlmZGdd�dejd�ZdS)�N)�
PackageFinder)�BaseDistribution)�InstallRequirementcsTeZdZdZedd��fdd�Zejed�dd��Z	eje
edd	�d
d��Z�Z
S)�AbstractDistributiona
A base class for handling installable artifacts.

    The requirements for anything installable are as follows:

     - we must be able to determine the requirement name
       (or we can't correctly handle the non-upgrade case).

     - for packages with setup requirements, we must also be able
       to determine their requirements without installing additional
       packages (for the same reason as run-time dependencies)

     - we must be able to create a Distribution object exposing the
       above metadata.
    N)�req�returncst���||_dS�N)�super�__init__r)�selfr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/distributions/base.pyr
s
zAbstractDistribution.__init__)rcCs
t��dSr��NotImplementedError)rrrr�get_metadata_distributionsz.AbstractDistribution.get_metadata_distribution)�finder�build_isolationrcCs
t��dSrr)rrrrrr�prepare_distribution_metadata sz2AbstractDistribution.prepare_distribution_metadata)�__name__�
__module__�__qualname__�__doc__rr
�abc�abstractmethodrrr�boolr�
__classcell__rrrrrs�r)�	metaclass)	r�"pip._internal.index.package_finderrZpip._internal.metadata.baser�pip._internal.reqr�ABCMetarrrrr�<module>ssite-packages/pip/_internal/distributions/__pycache__/wheel.cpython-39.pyc000064400000003142150564745010022705 0ustar00a

��Re[�@sLddlmZddlmZddlmZddlmZmZm	Z	Gdd�de�Z
dS)�)�canonicalize_name)�AbstractDistribution)�
PackageFinder)�BaseDistribution�FilesystemWheel�get_wheel_distributionc@s0eZdZdZed�dd�Zeedd�dd�ZdS)	�WheelDistributionzqRepresents a wheel distribution.

    This does not need any preparation as wheels can be directly unpacked.
    )�returncCs>|jjsJd��|jjs Jd��t|jj�}t|t|jj��S)z�Loads the metadata from the wheel file into memory and returns a
        Distribution that uses it, not relying on the wheel file or
        requirement.
        z*Set as part of preparation during downloadzWheels are never unnamed)�req�local_file_path�namerrr)�self�wheel�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/distributions/wheel.py�get_metadata_distributionsz+WheelDistribution.get_metadata_distributionN)�finder�build_isolationr	cCsdS)Nr)r
rrrrr�prepare_distribution_metadatasz/WheelDistribution.prepare_distribution_metadata)	�__name__�
__module__�__qualname__�__doc__rrr�boolrrrrrrs
�rN)�pip._vendor.packaging.utilsr� pip._internal.distributions.baser�"pip._internal.index.package_finderr�pip._internal.metadatarrrrrrrr�<module>ssite-packages/pip/_internal/distributions/__pycache__/sdist.cpython-39.pyc000064400000011112150564745010022723 0ustar00a

��Re��@s�ddlZddlmZmZmZddlmZddlmZddl	m
Z
ddlmZddl
mZddlmZe�e�ZGd	d
�d
e�ZdS)�N)�Iterable�Set�Tuple)�BuildEnvironment)�AbstractDistribution)�InstallationError)�
PackageFinder)�BaseDistribution)�runner_with_spinner_messagec@s�eZdZdZed�dd�Zeedd�dd�Zedd	�d
d�Z	e
ed�dd
�Ze
ed�dd�Z
edd	�dd�Zeeeeefdd�dd�ZdS)�SourceDistributionz�Represents a source distribution.

    The preparation step for these needs metadata for the packages to be
    generated, either using PEP 517 or using the legacy `setup.py egg_info`.
    )�returncCsddlm}||j���S)Nr)�Distribution)�$pip._internal.metadata.pkg_resourcesr
�req�get_dist)�self�_Dist�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/distributions/sdist.py�get_metadata_distributionsz,SourceDistribution.get_metadata_distributionN)�finder�build_isolationrcCsF|j��|jjo|}|r8|�|�|j��|�|�|j��dS)N)r�load_pyproject_toml�
use_pep517�_prepare_build_backend�isolated_editable_sanity_check�_install_build_reqs�prepare_metadata)rrr�should_isolaterrr�prepare_distribution_metadatas



z0SourceDistribution.prepare_distribution_metadata)rrc	Cs�|jj}|dusJ�t�|j_|jj�||dd�|jj�|jj�\}}|rX|�d|�|r�t�	d|j�t�	dd�
ttt
|����dS)N�overlayzInstalling build dependenciesz"PEP 517/518 supported requirementsz4Missing build requirements in pyproject.toml for %s.z`The project does not specify a build backend, and pip cannot fall back to setuptools without %s.z and )r�pyproject_requiresr�	build_env�install_requirements�check_requirements�requirements_to_check�_raise_conflicts�logger�warning�join�map�repr�sorted)rrr!�conflicting�missingrrrr3s(
����z)SourceDistribution._prepare_build_backendc	Cs�|jj�ltd�}|jj}|dus&J�|�|��&|��Wd�Wd�S1s\0YWd�n1sz0YdS)Nz#Getting requirements to build wheel)rr"r
�pep517_backend�subprocess_runner�get_requires_for_build_wheel�r�runner�backendrrr�_get_build_requires_wheelMs
z,SourceDistribution._get_build_requires_wheelc	Cs�|jj�ltd�}|jj}|dus&J�|�|��&|��Wd�Wd�S1s\0YWd�n1sz0YdS)Nz&Getting requirements to build editable)rr"r
r/r0�get_requires_for_build_editabler2rrr�_get_build_requires_editableUs
�z/SourceDistribution._get_build_requires_editablecCsf|jjr$|jjr$|j��r$|��}n|��}|jj�|�\}}|rN|�d|�|jj�	||dd�dS)Nzthe backend dependencies�normalzInstalling backend dependencies)
r�editable�permit_editable_wheels�supports_pyproject_editabler7r5r"r$r&r#)rr�
build_reqsr-r.rrrr_s���
�z&SourceDistribution._install_build_reqs)�conflicting_with�conflicting_reqsrcCs6d}|j|j|d�dd�t|�D��d�}t|��dS)NzZSome build dependencies for {requirement} conflict with {conflicting_with}: {description}.z, css |]\}}|�d|��VqdS)z is incompatible with Nr)�.0�	installed�wantedrrr�	<genexpr>|s�z6SourceDistribution._raise_conflicts.<locals>.<genexpr>)�requirementr=�description)�formatrr)r,r)rr=r>�
format_string�
error_messagerrrr&rs�
��z#SourceDistribution._raise_conflicts)�__name__�
__module__�__qualname__�__doc__r	rr�boolrrr�strr5r7rrrr&rrrrrs�
�r)�logging�typingrrr�pip._internal.build_envr� pip._internal.distributions.baser�pip._internal.exceptionsr�"pip._internal.index.package_finderr�pip._internal.metadatar	�pip._internal.utils.subprocessr
�	getLoggerrHr'rrrrr�<module>s
site-packages/pip/_internal/distributions/__pycache__/__init__.cpython-39.pyc000064400000001502150564745010023336 0ustar00a

��ReZ�@sDddlmZddlmZddlmZddlmZeed�dd�ZdS)	�)�AbstractDistribution)�SourceDistribution)�WheelDistribution)�InstallRequirement)�install_req�returncCs$|jrt|�S|jrt|�St|�S)z7Returns a Distribution for the given InstallRequirement)�editabler�is_wheelr)r�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/distributions/__init__.py�)make_distribution_for_install_requirements
rN)	� pip._internal.distributions.baser�!pip._internal.distributions.sdistr�!pip._internal.distributions.wheelrZpip._internal.req.req_installrrr
r
r
r�<module>s�site-packages/pip/_internal/distributions/__pycache__/installed.cpython-39.pyc000064400000002514150564745010023562 0ustar00a

��Re��@s8ddlmZddlmZddlmZGdd�de�ZdS)�)�AbstractDistribution)�
PackageFinder)�BaseDistributionc@s0eZdZdZed�dd�Zeedd�dd�ZdS)	�InstalledDistributionz�Represents an installed package.

    This does not need any preparation as the required information has already
    been computed.
    )�returncCs,ddlm}|jjdus Jd��||jj�S)Nr)�Distributionznot actually installed)�$pip._internal.metadata.pkg_resourcesr�req�satisfied_by)�self�_Dist�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/distributions/installed.py�get_metadata_distribution
sz/InstalledDistribution.get_metadata_distributionN)�finder�build_isolationrcCsdS)Nr
)rrrr
r
r�prepare_distribution_metadatasz3InstalledDistribution.prepare_distribution_metadata)	�__name__�
__module__�__qualname__�__doc__rrr�boolrr
r
r
rrs
�rN)� pip._internal.distributions.baser�"pip._internal.index.package_finderr�pip._internal.metadatarrr
r
r
r�<module>ssite-packages/pip/_internal/distributions/base.py000064400000002224150564745010016224 0ustar00import abc

from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata.base import BaseDistribution
from pip._internal.req import InstallRequirement


class AbstractDistribution(metaclass=abc.ABCMeta):
    """A base class for handling installable artifacts.

    The requirements for anything installable are as follows:

     - we must be able to determine the requirement name
       (or we can't correctly handle the non-upgrade case).

     - for packages with setup requirements, we must also be able
       to determine their requirements without installing additional
       packages (for the same reason as run-time dependencies)

     - we must be able to create a Distribution object exposing the
       above metadata.
    """

    def __init__(self, req: InstallRequirement) -> None:
        super().__init__()
        self.req = req

    @abc.abstractmethod
    def get_metadata_distribution(self) -> BaseDistribution:
        raise NotImplementedError()

    @abc.abstractmethod
    def prepare_distribution_metadata(
        self, finder: PackageFinder, build_isolation: bool
    ) -> None:
        raise NotImplementedError()
site-packages/pip/_internal/distributions/wheel.py000064400000002133150564745010016415 0ustar00from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.distributions.base import AbstractDistribution
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import (
    BaseDistribution,
    FilesystemWheel,
    get_wheel_distribution,
)


class WheelDistribution(AbstractDistribution):
    """Represents a wheel distribution.

    This does not need any preparation as wheels can be directly unpacked.
    """

    def get_metadata_distribution(self) -> BaseDistribution:
        """Loads the metadata from the wheel file into memory and returns a
        Distribution that uses it, not relying on the wheel file or
        requirement.
        """
        assert self.req.local_file_path, "Set as part of preparation during download"
        assert self.req.name, "Wheels are never unnamed"
        wheel = FilesystemWheel(self.req.local_file_path)
        return get_wheel_distribution(wheel, canonicalize_name(self.req.name))

    def prepare_distribution_metadata(
        self, finder: PackageFinder, build_isolation: bool
    ) -> None:
        pass
site-packages/pip/_internal/distributions/__init__.py000064400000001532150564745010017052 0ustar00from pip._internal.distributions.base import AbstractDistribution
from pip._internal.distributions.sdist import SourceDistribution
from pip._internal.distributions.wheel import WheelDistribution
from pip._internal.req.req_install import InstallRequirement


def make_distribution_for_install_requirement(
    install_req: InstallRequirement,
) -> AbstractDistribution:
    """Returns a Distribution for the given InstallRequirement"""
    # Editable requirements will always be source distributions. They use the
    # legacy logic until we create a modern standard for them.
    if install_req.editable:
        return SourceDistribution(install_req)

    # If it's a wheel, it's a WheelDistribution
    if install_req.is_wheel:
        return WheelDistribution(install_req)

    # Otherwise, a SourceDistribution
    return SourceDistribution(install_req)
site-packages/pip/_internal/distributions/installed.py000064400000001377150564745010017301 0ustar00from pip._internal.distributions.base import AbstractDistribution
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import BaseDistribution


class InstalledDistribution(AbstractDistribution):
    """Represents an installed package.

    This does not need any preparation as the required information has already
    been computed.
    """

    def get_metadata_distribution(self) -> BaseDistribution:
        from pip._internal.metadata.pkg_resources import Distribution as _Dist

        assert self.req.satisfied_by is not None, "not actually installed"
        return _Dist(self.req.satisfied_by)

    def prepare_distribution_metadata(
        self, finder: PackageFinder, build_isolation: bool
    ) -> None:
        pass
site-packages/pip/_internal/distributions/sdist.py000064400000012736150564745010016451 0ustar00import logging
from typing import Iterable, Set, Tuple

from pip._internal.build_env import BuildEnvironment
from pip._internal.distributions.base import AbstractDistribution
from pip._internal.exceptions import InstallationError
from pip._internal.index.package_finder import PackageFinder
from pip._internal.metadata import BaseDistribution
from pip._internal.utils.subprocess import runner_with_spinner_message

logger = logging.getLogger(__name__)


class SourceDistribution(AbstractDistribution):
    """Represents a source distribution.

    The preparation step for these needs metadata for the packages to be
    generated, either using PEP 517 or using the legacy `setup.py egg_info`.
    """

    def get_metadata_distribution(self) -> BaseDistribution:
        from pip._internal.metadata.pkg_resources import Distribution as _Dist

        return _Dist(self.req.get_dist())

    def prepare_distribution_metadata(
        self, finder: PackageFinder, build_isolation: bool
    ) -> None:
        # Load pyproject.toml, to determine whether PEP 517 is to be used
        self.req.load_pyproject_toml()

        # Set up the build isolation, if this requirement should be isolated
        should_isolate = self.req.use_pep517 and build_isolation
        if should_isolate:
            # Setup an isolated environment and install the build backend static
            # requirements in it.
            self._prepare_build_backend(finder)
            # Check that if the requirement is editable, it either supports PEP 660 or
            # has a setup.py or a setup.cfg. This cannot be done earlier because we need
            # to setup the build backend to verify it supports build_editable, nor can
            # it be done later, because we want to avoid installing build requirements
            # needlessly. Doing it here also works around setuptools generating
            # UNKNOWN.egg-info when running get_requires_for_build_wheel on a directory
            # without setup.py nor setup.cfg.
            self.req.isolated_editable_sanity_check()
            # Install the dynamic build requirements.
            self._install_build_reqs(finder)

        self.req.prepare_metadata()

    def _prepare_build_backend(self, finder: PackageFinder) -> None:
        # Isolate in a BuildEnvironment and install the build-time
        # requirements.
        pyproject_requires = self.req.pyproject_requires
        assert pyproject_requires is not None

        self.req.build_env = BuildEnvironment()
        self.req.build_env.install_requirements(
            finder, pyproject_requires, "overlay", "Installing build dependencies"
        )
        conflicting, missing = self.req.build_env.check_requirements(
            self.req.requirements_to_check
        )
        if conflicting:
            self._raise_conflicts("PEP 517/518 supported requirements", conflicting)
        if missing:
            logger.warning(
                "Missing build requirements in pyproject.toml for %s.",
                self.req,
            )
            logger.warning(
                "The project does not specify a build backend, and "
                "pip cannot fall back to setuptools without %s.",
                " and ".join(map(repr, sorted(missing))),
            )

    def _get_build_requires_wheel(self) -> Iterable[str]:
        with self.req.build_env:
            runner = runner_with_spinner_message("Getting requirements to build wheel")
            backend = self.req.pep517_backend
            assert backend is not None
            with backend.subprocess_runner(runner):
                return backend.get_requires_for_build_wheel()

    def _get_build_requires_editable(self) -> Iterable[str]:
        with self.req.build_env:
            runner = runner_with_spinner_message(
                "Getting requirements to build editable"
            )
            backend = self.req.pep517_backend
            assert backend is not None
            with backend.subprocess_runner(runner):
                return backend.get_requires_for_build_editable()

    def _install_build_reqs(self, finder: PackageFinder) -> None:
        # Install any extra build dependencies that the backend requests.
        # This must be done in a second pass, as the pyproject.toml
        # dependencies must be installed before we can call the backend.
        if (
            self.req.editable
            and self.req.permit_editable_wheels
            and self.req.supports_pyproject_editable()
        ):
            build_reqs = self._get_build_requires_editable()
        else:
            build_reqs = self._get_build_requires_wheel()
        conflicting, missing = self.req.build_env.check_requirements(build_reqs)
        if conflicting:
            self._raise_conflicts("the backend dependencies", conflicting)
        self.req.build_env.install_requirements(
            finder, missing, "normal", "Installing backend dependencies"
        )

    def _raise_conflicts(
        self, conflicting_with: str, conflicting_reqs: Set[Tuple[str, str]]
    ) -> None:
        format_string = (
            "Some build dependencies for {requirement} "
            "conflict with {conflicting_with}: {description}."
        )
        error_message = format_string.format(
            requirement=self.req,
            conflicting_with=conflicting_with,
            description=", ".join(
                f"{installed} is incompatible with {wanted}"
                for installed, wanted in sorted(conflicting_reqs)
            ),
        )
        raise InstallationError(error_message)
site-packages/pip/_internal/metadata/__pycache__/pkg_resources.cpython-39.pyc000064400000014003150564745010023330 0ustar00a

��Re��@s�ddlZddlZddlmZmZmZmZmZm	Z	ddl
mZddlm
Z
ddlmZmZddlmZddlmZddlmZmZdd	lmZd
dlmZmZmZmZm Z e�!e"�Z#Gdd
�d
e�Z$Gdd�de�Z%Gdd�de�Z&dS)�N)�
Collection�Iterable�Iterator�List�
NamedTuple�Optional)�
pkg_resources)�Requirement)�NormalizedName�canonicalize_name)�parse)�misc)�
get_installer�get_metadata)�$pkg_resources_distribution_for_wheel�)�BaseDistribution�BaseEntryPoint�BaseEnvironment�DistributionVersion�Wheelc@s&eZdZUeed<eed<eed<dS)�
EntryPoint�name�value�groupN)�__name__�
__module__�__qualname__�str�__annotations__�r r ��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/metadata/pkg_resources.pyrs
rc@s0eZdZejdd�dd�Zeeedd�dd��Z	e
eed�d	d
��Ze
eed�dd��Z
e
ed�d
d��Ze
ed�dd��Ze
ed�dd��Ze
ed�dd��Ze
ed�dd��Ze
ed�dd��Zeed�dd�Zeed�dd�Ze
ejjd�dd��Zd&eeee d!�d"d#�Z!eed�d$d%�Z"dS)'�DistributionN)�dist�returncCs
||_dS�N)�_dist��selfr#r r r!�__init__ szDistribution.__init__)�wheelrr$cCs>|���}t|||j�}Wd�n1s,0Y||�Sr%)�
as_zipfiler�location)�clsr*r�zfr#r r r!�
from_wheel#s
,zDistribution.from_wheel�r$cCs|jjSr%)r&r,�r(r r r!r,)szDistribution.locationcCs|jjSr%)r&�egg_infor1r r r!�info_directory-szDistribution.info_directorycCst|jj�Sr%)rr&�project_namer1r r r!�canonical_name1szDistribution.canonical_namecCst|jj�Sr%)�
parse_versionr&�versionr1r r r!r75szDistribution.versioncCs
t|j�Sr%)rr&r1r r r!�	installer9szDistribution.installercCst�|j�Sr%)r
�
dist_is_localr&r1r r r!�local=szDistribution.localcCst�|j�Sr%)r
�dist_in_usersiter&r1r r r!�in_usersiteAszDistribution.in_usersitecCst�|j�Sr%)r
�dist_in_site_packagesr&r1r r r!�in_site_packagesEszDistribution.in_site_packages�rr$cCs |j�|�st|��|j�|�Sr%)r&�has_metadata�FileNotFoundErrorr)r(rr r r!�	read_textIszDistribution.read_textccsZ|j����D]F\}}|��D]4\}}t|��d�\}}}t|��|��|d�VqqdS)N�=)rrr)r&�
get_entry_map�itemsr�	partitionr�strip)r(r�entriesr�entry_point�_rr r r!�iter_entry_pointsNszDistribution.iter_entry_pointscCs
t|j�Sr%)rr&r1r r r!�metadataTszDistribution.metadatar )�extrasr$cCs"|rt|��|jj�}|j�|�Sr%)�	frozenset�intersectionr&rM�requires)r(rMr r r!�iter_dependenciesXszDistribution.iter_dependenciescCs|jjSr%)r&rMr1r r r!�iter_provided_extras]sz!Distribution.iter_provided_extras)r )#rrrrr"r)�classmethodrrr/�propertyrr,r3r
r5rr7r8�boolr:r<r>rBrrrK�email�message�MessagerLrr	rQrRr r r r!r"s2r"c@s�eZdZejdd�dd�Zeed�dd��Zee	e
eed�d	d
��Zee	e
d�dd
�Zee	e
d�dd�Zee
d�dd�ZdS)�EnvironmentN)�wsr$cCs
||_dSr%)�_ws)r(rZr r r!r)bszEnvironment.__init__r0cCs
|tj�Sr%)r�working_set)r-r r r!�defaulteszEnvironment.default)�pathsr$cCs|t�|��Sr%)r�
WorkingSet)r-r^r r r!�
from_pathsiszEnvironment.from_pathsr?cCs,t|�}|��D]}|j|kr|SqdS)z�Find a distribution matching the ``name`` in the environment.

        This searches from *all* distributions available in the environment, to
        match the behavior of ``pkg_resources.get_distribution()``.
        N)r�iter_distributionsr5)r(rr5r#r r r!�_search_distributionms


z Environment._search_distributioncCsD|�|�}|r|Sz|j�|�Wntjy8YdS0|�|�Sr%)rbr[�requirer�DistributionNotFound)r(rr#r r r!�get_distributionys

zEnvironment.get_distributionccs|jD]}t|�VqdSr%)r[r"r'r r r!�_iter_distributions�s
zEnvironment._iter_distributions)rrrrr_r)rSrr]rrrr`rrbrerrfr r r r!rYasrY)'�
email.messagerV�logging�typingrrrrrr�pip._vendorrZ"pip._vendor.packaging.requirementsr	�pip._vendor.packaging.utilsr
rZpip._vendor.packaging.versionrr6Zpip._internal.utilsr
�pip._internal.utils.packagingrr�pip._internal.utils.wheelr�baserrrrr�	getLoggerr�loggerrr"rYr r r r!�<module>s 
Bsite-packages/pip/_internal/metadata/__pycache__/base.cpython-39.pyc000064400000031357150564745010021402 0ustar00a

��Re_+�@sRddlZddlZddlZddlZddlZddlmZmZm	Z	m
Z
mZmZm
Z
mZmZddlmZddlmZmZddlmZddlmZmZddlmZmZmZddlmZdd	l m!Z!dd
l"m#Z#er�ddlm$Z$ne%Z$eeefZ&e�'e(�Z)Gdd
�d
e$�Z*Gdd�de$�Z+Gdd�d�Z,Gdd�de$�Z-Gdd�de-�Z.Gdd�de-�Z/dS)�N)	�IO�
TYPE_CHECKING�
Collection�	Container�Iterable�Iterator�List�Optional�Union)�Requirement)�InvalidSpecifier�SpecifierSet)�NormalizedName)�
LegacyVersion�Version)�DIRECT_URL_METADATA_NAME�	DirectUrl�DirectUrlValidationError)�stdlib_pkgs)�egg_link_path_from_sys_path)�url_to_path)�Protocolc@sBeZdZeed�dd��Zeed�dd��Zeed�dd��ZdS)	�BaseEntryPoint��returncCs
t��dS�N��NotImplementedError��self�r ��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/metadata/base.py�name+szBaseEntryPoint.namecCs
t��dSrrrr r r!�value/szBaseEntryPoint.valuecCs
t��dSrrrr r r!�group3szBaseEntryPoint.groupN)�__name__�
__module__�__qualname__�property�strr"r#r$r r r r!r*src@s�eZdZed�dd�Zed�dd�Zeeed�dd��Zeeed�dd	��Z	eeed�d
d��Z
eed�dd
��Zee
d�dd��Zeeed�dd��Zeed�dd��Zeed�dd��Zeed�dd��Zeed�dd��Zeed�dd��Zeed�dd�Zeed�dd �Zeejjd�d!d"��Zeeed�d#d$��Zeed�d%d&��Z ee!d�d'd(��Z"d0e#eee$d*�d+d,�Z%eed�d-d.�Z&d/S)1�BaseDistributionrcCs|j�d|j�d|j�d�S)N� z (�))�raw_name�version�locationrr r r!�__repr__9szBaseDistribution.__repr__cCs|j�d|j��S)Nr+)r-r.rr r r!�__str__<szBaseDistribution.__str__cCs
t��dS)a�Where the distribution is loaded from.

        A string value is not necessarily a filesystem path, since distributions
        can be loaded from other sources, e.g. arbitrary zip archives. ``None``
        means the distribution is created in-memory.

        Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If
        this is a symbolic link, we want to preserve the relative path between
        it and files in the distribution.
        Nrrr r r!r/?szBaseDistribution.locationcCs6|j}|r|��r2t|j�Snt|j�}|r2|jSdS)z�The project location for editable distributions.

        This is the directory where pyproject.toml or setup.py is located.
        None if the distribution is not installed in editable mode.
        N)�
direct_url�is_local_editabler�urlrr-r/)rr2�
egg_link_pathr r r!�editable_project_locationMs
z*BaseDistribution.editable_project_locationcCs
t��dS)a'Location of the .[egg|dist]-info directory.

        Similarly to ``location``, a string value is not necessarily a
        filesystem path. ``None`` means the distribution is created in-memory.

        For a modern .dist-info installation on disk, this should be something
        like ``{location}/{raw_name}-{version}.dist-info``.

        Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If
        this is a symbolic link, we want to preserve the relative path between
        it and other files in the distribution.
        Nrrr r r!�info_directorycszBaseDistribution.info_directorycCs
t��dSrrrr r r!�canonical_namesszBaseDistribution.canonical_namecCs
t��dSrrrr r r!r.wszBaseDistribution.versionc
Csvz|�t�}Wnty"YdS0zt�|�WSttjtfyp}z t	�
dt|j|�WYd}~dSd}~00dS)z�Obtain a DirectUrl from this distribution.

        Returns None if the distribution has no `direct_url.json` metadata,
        or if `direct_url.json` is invalid.
        NzError parsing %s for %s: %s)�	read_textr�FileNotFoundErrorr�	from_json�UnicodeDecodeError�json�JSONDecodeErrorr�logger�warningr8)r�content�er r r!r2{s$��zBaseDistribution.direct_urlcCs
t��dSrrrr r r!�	installer�szBaseDistribution.installercCs
t|j�Sr)�boolr6rr r r!�editable�szBaseDistribution.editablecCs
t��dSrrrr r r!�local�szBaseDistribution.localcCs
t��dSrrrr r r!�in_usersite�szBaseDistribution.in_usersitecCs
t��dSrrrr r r!�in_site_packages�sz!BaseDistribution.in_site_packages�r"rcCs
t��dS)z�Read a file in the .dist-info (or .egg-info) directory.

        Should raise ``FileNotFoundError`` if ``name`` does not exist in the
        metadata directory.
        Nr�rr"r r r!r9�szBaseDistribution.read_textcCs
t��dSrrrr r r!�iter_entry_points�sz"BaseDistribution.iter_entry_pointscCs
t��dS)z?Metadata of distribution parsed from e.g. METADATA or PKG-INFO.Nrrr r r!�metadata�szBaseDistribution.metadatacCs|j�d�S)zDValue of "Metadata-Version:" in distribution metadata, if available.zMetadata-Version)rL�getrr r r!�metadata_version�sz!BaseDistribution.metadata_versioncCs|j�d|j�S)z*Value of "Name:" in distribution metadata.�Name)rLrMr8rr r r!r-�szBaseDistribution.raw_namec
Csn|j�d�}|durt�Sztt|��}Wn>tyh}z&d}t�||j|�t�WYd}~Sd}~00|S)z�Value of "Requires-Python:" in distribution metadata.

        If the key does not exist or contains an invalid value, an empty
        SpecifierSet should be returned.
        zRequires-PythonNz-Package %r has an invalid Requires-Python: %s)rLrMr
r)rr?r@r-)rr#�specrB�messager r r!�requires_python�sz BaseDistribution.requires_pythonr )�extrasrcCs
t��dS)z�Dependencies of this distribution.

        For modern .dist-info distributions, this is the collection of
        "Requires-Dist:" entries in distribution metadata.
        Nr)rrSr r r!�iter_dependencies�sz"BaseDistribution.iter_dependenciescCs
t��dS)z�Extras provided by this distribution.

        For modern .dist-info distributions, this is the collection of
        "Provides-Extra:" entries in distribution metadata.
        Nrrr r r!�iter_provided_extras�sz%BaseDistribution.iter_provided_extrasN)r )'r%r&r'r)r0r1r(r	r/r6r7rr8�DistributionVersionr.rr2rCrDrErFrGrHr9rrrK�emailrQ�MessagerLrNr-r
rRrrrTrUr r r r!r*8sH
r*c@s�eZdZdZedd�dd��Zeeeedd�dd��Z	eedd	�d
d�Z
edd�dd
�Zedd�dd�Z
dedddfeeeeeeeed�dd�ZdS)�BaseEnvironmentz6An environment containing distributions to introspect.rcCs
t��dSrr)�clsr r r!�default�szBaseEnvironment.default)�pathsrcCs
t��dSrr)rZr\r r r!�
from_paths�szBaseEnvironment.from_pathsr*rIcCs
t��dS)z=Given a requirement name, return the installed distributions.NrrJr r r!�get_distribution�sz BaseEnvironment.get_distributioncCs
t��dS)aIterate through installed distributions.

        This function should be implemented by subclass, but never called
        directly. Use the public ``iter_distribution()`` instead, which
        implements additional logic to make sure the distributions are valid.
        Nrrr r r!�_iter_distributions�sz#BaseEnvironment._iter_distributionsccsD|��D]6}tjd|jtjd�}|s8t�d|j|j�q|VqdS)z(Iterate through installed distributions.z)^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$)�flagsz%Ignoring invalid distribution %s (%s)N)r_�re�matchr8�
IGNORECASEr?r@r/)r�dist�project_name_validr r r!�iter_distributionss��z"BaseEnvironment.iter_distributionsTF)�
local_only�skip�include_editables�editables_only�	user_onlyrcsb|��}|rdd�|D�}|s,dd�|D�}|r>dd�|D�}|rPdd�|D�}�fdd�|D�S)aReturn a list of installed distributions.

        :param local_only: If True (default), only return installations
        local to the current virtualenv, if in a virtualenv.
        :param skip: An iterable of canonicalized project names to ignore;
            defaults to ``stdlib_pkgs``.
        :param include_editables: If False, don't report editables.
        :param editables_only: If True, only report editables.
        :param user_only: If True, only report installations in the user
        site directory.
        css|]}|jr|VqdSr)rF��.0�dr r r!�	<genexpr>+�z?BaseEnvironment.iter_installed_distributions.<locals>.<genexpr>css|]}|js|VqdSr�rErlr r r!ro-rpcss|]}|jr|VqdSrrqrlr r r!ro/rpcss|]}|jr|VqdSr)rGrlr r r!ro1rpc3s|]}|j�vr|VqdSr)r8rl�rhr r!ro2rp)rf)rrgrhrirjrk�itr rrr!�iter_installed_distributionssz,BaseEnvironment.iter_installed_distributionsN)r%r&r'�__doc__�classmethodr[r	rr)r]r^rr_rfrrDrr*rtr r r r!rY�s*	��rYc@s&eZdZUeed<ejd�dd�ZdS)�Wheelr/rcCs
t��dSrrrr r r!�
as_zipfile8szWheel.as_zipfileN)r%r&r'r)�__annotations__�zipfile�ZipFilerxr r r r!rw5s
rwc@s,eZdZedd�dd�Zejd�dd�ZdS)�FilesystemWheelN)r/rcCs
||_dSr)r/)rr/r r r!�__init__=szFilesystemWheel.__init__rcCstj|jdd�S�NT)�
allowZip64)rzr{r/rr r r!rx@szFilesystemWheel.as_zipfile)r%r&r'r)r}rzr{rxr r r r!r|<sr|c@s2eZdZeeedd�dd�Zejd�dd�Z	dS)�MemoryWheelN)r/�streamrcCs||_||_dSr)r/r�)rr/r�r r r!r}EszMemoryWheel.__init__rcCstj|jdd�Sr~)rzr{r�rr r r!rxIszMemoryWheel.as_zipfile)
r%r&r'r)r�bytesr}rzr{rxr r r r!r�Dsr�)0�
email.messagerWr=�loggingrarz�typingrrrrrrrr	r
Z"pip._vendor.packaging.requirementsrZ pip._vendor.packaging.specifiersrr
�pip._vendor.packaging.utilsrZpip._vendor.packaging.versionrr�pip._internal.models.direct_urlrrr�pip._internal.utils.compatr�pip._internal.utils.egg_linkr�pip._internal.utils.urlsrr�objectrV�	getLoggerr%r?rr*rYrwr|r�r r r r!�<module>s2,
2Lsite-packages/pip/_internal/metadata/__pycache__/__init__.cpython-39.pyc000064400000003650150564745010022222 0ustar00a

��Re|�@spddlmZmZddlmZmZmZmZmZgd�Z	ed�dd�Z
eeeed�d	d
�Zeeed�dd
�Z
dS)�)�List�Optional�)�BaseDistribution�BaseEnvironment�FilesystemWheel�MemoryWheel�Wheel)rrrrr	�get_default_environment�get_environment�get_wheel_distribution)�returncCsddlm}|��S)a	Get the default representation for the current environment.

    This returns an Environment instance from the chosen backend. The default
    Environment instance should be built from ``sys.path`` and may use caching
    to share instance state accorss calls.
    r��Environment)�
pkg_resourcesr�defaultr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/metadata/__init__.pyr
sr
)�pathsr
cCsddlm}|�|�S)a'Get a representation of the environment specified by ``paths``.

    This returns an Environment instance from the chosen backend based on the
    given import paths. The backend must build a fresh instance representing
    the state of installed distributions when this function is called.
    rr)rr�
from_paths)rrrrrrsr)�wheel�canonical_namer
cCsddlm}|�||�S)aGet the representation of the specified wheel's distribution metadata.

    This returns a Distribution instance from the chosen backend based on
    the given wheel's ``.dist-info`` directory.

    :param canonical_name: Normalized project name of the given wheel.
    r)�Distribution)rr�
from_wheel)rrrrrrr)srN)�typingrr�baserrrrr	�__all__r
�strrrrrrr�<module>s
site-packages/pip/_internal/metadata/base.py000064400000025537150564745010015116 0ustar00import email.message
import json
import logging
import re
import zipfile
from typing import (
    IO,
    TYPE_CHECKING,
    Collection,
    Container,
    Iterable,
    Iterator,
    List,
    Optional,
    Union,
)

from pip._vendor.packaging.requirements import Requirement
from pip._vendor.packaging.specifiers import InvalidSpecifier, SpecifierSet
from pip._vendor.packaging.utils import NormalizedName
from pip._vendor.packaging.version import LegacyVersion, Version

from pip._internal.models.direct_url import (
    DIRECT_URL_METADATA_NAME,
    DirectUrl,
    DirectUrlValidationError,
)
from pip._internal.utils.compat import stdlib_pkgs  # TODO: Move definition here.
from pip._internal.utils.egg_link import egg_link_path_from_sys_path
from pip._internal.utils.urls import url_to_path

if TYPE_CHECKING:
    from typing import Protocol
else:
    Protocol = object

DistributionVersion = Union[LegacyVersion, Version]

logger = logging.getLogger(__name__)


class BaseEntryPoint(Protocol):
    @property
    def name(self) -> str:
        raise NotImplementedError()

    @property
    def value(self) -> str:
        raise NotImplementedError()

    @property
    def group(self) -> str:
        raise NotImplementedError()


class BaseDistribution(Protocol):
    def __repr__(self) -> str:
        return f"{self.raw_name} {self.version} ({self.location})"

    def __str__(self) -> str:
        return f"{self.raw_name} {self.version}"

    @property
    def location(self) -> Optional[str]:
        """Where the distribution is loaded from.

        A string value is not necessarily a filesystem path, since distributions
        can be loaded from other sources, e.g. arbitrary zip archives. ``None``
        means the distribution is created in-memory.

        Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If
        this is a symbolic link, we want to preserve the relative path between
        it and files in the distribution.
        """
        raise NotImplementedError()

    @property
    def editable_project_location(self) -> Optional[str]:
        """The project location for editable distributions.

        This is the directory where pyproject.toml or setup.py is located.
        None if the distribution is not installed in editable mode.
        """
        # TODO: this property is relatively costly to compute, memoize it ?
        direct_url = self.direct_url
        if direct_url:
            if direct_url.is_local_editable():
                return url_to_path(direct_url.url)
        else:
            # Search for an .egg-link file by walking sys.path, as it was
            # done before by dist_is_editable().
            egg_link_path = egg_link_path_from_sys_path(self.raw_name)
            if egg_link_path:
                # TODO: get project location from second line of egg_link file
                #       (https://github.com/pypa/pip/issues/10243)
                return self.location
        return None

    @property
    def info_directory(self) -> Optional[str]:
        """Location of the .[egg|dist]-info directory.

        Similarly to ``location``, a string value is not necessarily a
        filesystem path. ``None`` means the distribution is created in-memory.

        For a modern .dist-info installation on disk, this should be something
        like ``{location}/{raw_name}-{version}.dist-info``.

        Do not canonicalize this value with e.g. ``pathlib.Path.resolve()``. If
        this is a symbolic link, we want to preserve the relative path between
        it and other files in the distribution.
        """
        raise NotImplementedError()

    @property
    def canonical_name(self) -> NormalizedName:
        raise NotImplementedError()

    @property
    def version(self) -> DistributionVersion:
        raise NotImplementedError()

    @property
    def direct_url(self) -> Optional[DirectUrl]:
        """Obtain a DirectUrl from this distribution.

        Returns None if the distribution has no `direct_url.json` metadata,
        or if `direct_url.json` is invalid.
        """
        try:
            content = self.read_text(DIRECT_URL_METADATA_NAME)
        except FileNotFoundError:
            return None
        try:
            return DirectUrl.from_json(content)
        except (
            UnicodeDecodeError,
            json.JSONDecodeError,
            DirectUrlValidationError,
        ) as e:
            logger.warning(
                "Error parsing %s for %s: %s",
                DIRECT_URL_METADATA_NAME,
                self.canonical_name,
                e,
            )
            return None

    @property
    def installer(self) -> str:
        raise NotImplementedError()

    @property
    def editable(self) -> bool:
        return bool(self.editable_project_location)

    @property
    def local(self) -> bool:
        raise NotImplementedError()

    @property
    def in_usersite(self) -> bool:
        raise NotImplementedError()

    @property
    def in_site_packages(self) -> bool:
        raise NotImplementedError()

    def read_text(self, name: str) -> str:
        """Read a file in the .dist-info (or .egg-info) directory.

        Should raise ``FileNotFoundError`` if ``name`` does not exist in the
        metadata directory.
        """
        raise NotImplementedError()

    def iter_entry_points(self) -> Iterable[BaseEntryPoint]:
        raise NotImplementedError()

    @property
    def metadata(self) -> email.message.Message:
        """Metadata of distribution parsed from e.g. METADATA or PKG-INFO."""
        raise NotImplementedError()

    @property
    def metadata_version(self) -> Optional[str]:
        """Value of "Metadata-Version:" in distribution metadata, if available."""
        return self.metadata.get("Metadata-Version")

    @property
    def raw_name(self) -> str:
        """Value of "Name:" in distribution metadata."""
        # The metadata should NEVER be missing the Name: key, but if it somehow
        # does, fall back to the known canonical name.
        return self.metadata.get("Name", self.canonical_name)

    @property
    def requires_python(self) -> SpecifierSet:
        """Value of "Requires-Python:" in distribution metadata.

        If the key does not exist or contains an invalid value, an empty
        SpecifierSet should be returned.
        """
        value = self.metadata.get("Requires-Python")
        if value is None:
            return SpecifierSet()
        try:
            # Convert to str to satisfy the type checker; this can be a Header object.
            spec = SpecifierSet(str(value))
        except InvalidSpecifier as e:
            message = "Package %r has an invalid Requires-Python: %s"
            logger.warning(message, self.raw_name, e)
            return SpecifierSet()
        return spec

    def iter_dependencies(self, extras: Collection[str] = ()) -> Iterable[Requirement]:
        """Dependencies of this distribution.

        For modern .dist-info distributions, this is the collection of
        "Requires-Dist:" entries in distribution metadata.
        """
        raise NotImplementedError()

    def iter_provided_extras(self) -> Iterable[str]:
        """Extras provided by this distribution.

        For modern .dist-info distributions, this is the collection of
        "Provides-Extra:" entries in distribution metadata.
        """
        raise NotImplementedError()


class BaseEnvironment:
    """An environment containing distributions to introspect."""

    @classmethod
    def default(cls) -> "BaseEnvironment":
        raise NotImplementedError()

    @classmethod
    def from_paths(cls, paths: Optional[List[str]]) -> "BaseEnvironment":
        raise NotImplementedError()

    def get_distribution(self, name: str) -> Optional["BaseDistribution"]:
        """Given a requirement name, return the installed distributions."""
        raise NotImplementedError()

    def _iter_distributions(self) -> Iterator["BaseDistribution"]:
        """Iterate through installed distributions.

        This function should be implemented by subclass, but never called
        directly. Use the public ``iter_distribution()`` instead, which
        implements additional logic to make sure the distributions are valid.
        """
        raise NotImplementedError()

    def iter_distributions(self) -> Iterator["BaseDistribution"]:
        """Iterate through installed distributions."""
        for dist in self._iter_distributions():
            # Make sure the distribution actually comes from a valid Python
            # packaging distribution. Pip's AdjacentTempDirectory leaves folders
            # e.g. ``~atplotlib.dist-info`` if cleanup was interrupted. The
            # valid project name pattern is taken from PEP 508.
            project_name_valid = re.match(
                r"^([A-Z0-9]|[A-Z0-9][A-Z0-9._-]*[A-Z0-9])$",
                dist.canonical_name,
                flags=re.IGNORECASE,
            )
            if not project_name_valid:
                logger.warning(
                    "Ignoring invalid distribution %s (%s)",
                    dist.canonical_name,
                    dist.location,
                )
                continue
            yield dist

    def iter_installed_distributions(
        self,
        local_only: bool = True,
        skip: Container[str] = stdlib_pkgs,
        include_editables: bool = True,
        editables_only: bool = False,
        user_only: bool = False,
    ) -> Iterator[BaseDistribution]:
        """Return a list of installed distributions.

        :param local_only: If True (default), only return installations
        local to the current virtualenv, if in a virtualenv.
        :param skip: An iterable of canonicalized project names to ignore;
            defaults to ``stdlib_pkgs``.
        :param include_editables: If False, don't report editables.
        :param editables_only: If True, only report editables.
        :param user_only: If True, only report installations in the user
        site directory.
        """
        it = self.iter_distributions()
        if local_only:
            it = (d for d in it if d.local)
        if not include_editables:
            it = (d for d in it if not d.editable)
        if editables_only:
            it = (d for d in it if d.editable)
        if user_only:
            it = (d for d in it if d.in_usersite)
        return (d for d in it if d.canonical_name not in skip)


class Wheel(Protocol):
    location: str

    def as_zipfile(self) -> zipfile.ZipFile:
        raise NotImplementedError()


class FilesystemWheel(Wheel):
    def __init__(self, location: str) -> None:
        self.location = location

    def as_zipfile(self) -> zipfile.ZipFile:
        return zipfile.ZipFile(self.location, allowZip64=True)


class MemoryWheel(Wheel):
    def __init__(self, location: str, stream: IO[bytes]) -> None:
        self.location = location
        self.stream = stream

    def as_zipfile(self) -> zipfile.ZipFile:
        return zipfile.ZipFile(self.stream, allowZip64=True)
site-packages/pip/_internal/metadata/pkg_resources.py000064400000011741150564745010017047 0ustar00import email.message
import logging
from typing import Collection, Iterable, Iterator, List, NamedTuple, Optional

from pip._vendor import pkg_resources
from pip._vendor.packaging.requirements import Requirement
from pip._vendor.packaging.utils import NormalizedName, canonicalize_name
from pip._vendor.packaging.version import parse as parse_version

from pip._internal.utils import misc  # TODO: Move definition here.
from pip._internal.utils.packaging import get_installer, get_metadata
from pip._internal.utils.wheel import pkg_resources_distribution_for_wheel

from .base import (
    BaseDistribution,
    BaseEntryPoint,
    BaseEnvironment,
    DistributionVersion,
    Wheel,
)

logger = logging.getLogger(__name__)


class EntryPoint(NamedTuple):
    name: str
    value: str
    group: str


class Distribution(BaseDistribution):
    def __init__(self, dist: pkg_resources.Distribution) -> None:
        self._dist = dist

    @classmethod
    def from_wheel(cls, wheel: Wheel, name: str) -> "Distribution":
        with wheel.as_zipfile() as zf:
            dist = pkg_resources_distribution_for_wheel(zf, name, wheel.location)
        return cls(dist)

    @property
    def location(self) -> Optional[str]:
        return self._dist.location

    @property
    def info_directory(self) -> Optional[str]:
        return self._dist.egg_info

    @property
    def canonical_name(self) -> NormalizedName:
        return canonicalize_name(self._dist.project_name)

    @property
    def version(self) -> DistributionVersion:
        return parse_version(self._dist.version)

    @property
    def installer(self) -> str:
        return get_installer(self._dist)

    @property
    def local(self) -> bool:
        return misc.dist_is_local(self._dist)

    @property
    def in_usersite(self) -> bool:
        return misc.dist_in_usersite(self._dist)

    @property
    def in_site_packages(self) -> bool:
        return misc.dist_in_site_packages(self._dist)

    def read_text(self, name: str) -> str:
        if not self._dist.has_metadata(name):
            raise FileNotFoundError(name)
        return self._dist.get_metadata(name)

    def iter_entry_points(self) -> Iterable[BaseEntryPoint]:
        for group, entries in self._dist.get_entry_map().items():
            for name, entry_point in entries.items():
                name, _, value = str(entry_point).partition("=")
                yield EntryPoint(name=name.strip(), value=value.strip(), group=group)

    @property
    def metadata(self) -> email.message.Message:
        return get_metadata(self._dist)

    def iter_dependencies(self, extras: Collection[str] = ()) -> Iterable[Requirement]:
        if extras:  # pkg_resources raises on invalid extras, so we sanitize.
            extras = frozenset(extras).intersection(self._dist.extras)
        return self._dist.requires(extras)

    def iter_provided_extras(self) -> Iterable[str]:
        return self._dist.extras


class Environment(BaseEnvironment):
    def __init__(self, ws: pkg_resources.WorkingSet) -> None:
        self._ws = ws

    @classmethod
    def default(cls) -> BaseEnvironment:
        return cls(pkg_resources.working_set)

    @classmethod
    def from_paths(cls, paths: Optional[List[str]]) -> BaseEnvironment:
        return cls(pkg_resources.WorkingSet(paths))

    def _search_distribution(self, name: str) -> Optional[BaseDistribution]:
        """Find a distribution matching the ``name`` in the environment.

        This searches from *all* distributions available in the environment, to
        match the behavior of ``pkg_resources.get_distribution()``.
        """
        canonical_name = canonicalize_name(name)
        for dist in self.iter_distributions():
            if dist.canonical_name == canonical_name:
                return dist
        return None

    def get_distribution(self, name: str) -> Optional[BaseDistribution]:

        # Search the distribution by looking through the working set.
        dist = self._search_distribution(name)
        if dist:
            return dist

        # If distribution could not be found, call working_set.require to
        # update the working set, and try to find the distribution again.
        # This might happen for e.g. when you install a package twice, once
        # using setup.py develop and again using setup.py install. Now when
        # running pip uninstall twice, the package gets removed from the
        # working set in the first uninstall, so we have to populate the
        # working set again so that pip knows about it and the packages gets
        # picked up and is successfully uninstalled the second time too.
        try:
            # We didn't pass in any version specifiers, so this can never
            # raise pkg_resources.VersionConflict.
            self._ws.require(name)
        except pkg_resources.DistributionNotFound:
            return None
        return self._search_distribution(name)

    def _iter_distributions(self) -> Iterator[BaseDistribution]:
        for dist in self._ws:
            yield Distribution(dist)
site-packages/pip/_internal/metadata/__init__.py000064400000003174150564745010015734 0ustar00from typing import List, Optional

from .base import BaseDistribution, BaseEnvironment, FilesystemWheel, MemoryWheel, Wheel

__all__ = [
    "BaseDistribution",
    "BaseEnvironment",
    "FilesystemWheel",
    "MemoryWheel",
    "Wheel",
    "get_default_environment",
    "get_environment",
    "get_wheel_distribution",
]


def get_default_environment() -> BaseEnvironment:
    """Get the default representation for the current environment.

    This returns an Environment instance from the chosen backend. The default
    Environment instance should be built from ``sys.path`` and may use caching
    to share instance state accorss calls.
    """
    from .pkg_resources import Environment

    return Environment.default()


def get_environment(paths: Optional[List[str]]) -> BaseEnvironment:
    """Get a representation of the environment specified by ``paths``.

    This returns an Environment instance from the chosen backend based on the
    given import paths. The backend must build a fresh instance representing
    the state of installed distributions when this function is called.
    """
    from .pkg_resources import Environment

    return Environment.from_paths(paths)


def get_wheel_distribution(wheel: Wheel, canonical_name: str) -> BaseDistribution:
    """Get the representation of the specified wheel's distribution metadata.

    This returns a Distribution instance from the chosen backend based on
    the given wheel's ``.dist-info`` directory.

    :param canonical_name: Normalized project name of the given wheel.
    """
    from .pkg_resources import Distribution

    return Distribution.from_wheel(wheel, canonical_name)
site-packages/pip/_internal/models/__pycache__/candidate.cpython-39.pyc000064400000002712150564745010022100 0ustar00a

��Re��@s8ddlmZddlmZddlmZGdd�de�ZdS)�)�parse)�Link)�KeyBasedCompareMixincsPeZdZdZgd�Zeeedd��fdd�Zed�dd	�Zed�d
d�Z	�Z
S)�InstallationCandidatez4Represents a potential "candidate" for installation.)�name�version�linkN)rrr�returncs6||_t|�|_||_t�j|j|j|jftd�dS)N)�key�defining_class)r�
parse_versionrr�super�__init__r)�selfrrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/candidate.pyrs
�zInstallationCandidate.__init__)r	cCsd�|j|j|j�S)Nz)<InstallationCandidate({!r}, {!r}, {!r})>��formatrrr�rrrr�__repr__s
�zInstallationCandidate.__repr__cCsd�|j|j|j�S)Nz!{!r} candidate (version {} at {})rrrrr�__str__s
�zInstallationCandidate.__str__)�__name__�
__module__�__qualname__�__doc__�	__slots__�strrrrr�
__classcell__rrrrrs

rN)Zpip._vendor.packaging.versionrr�pip._internal.models.linkr�pip._internal.utils.modelsrrrrrr�<module>ssite-packages/pip/_internal/models/__pycache__/target_python.cpython-39.pyc000064400000006567150564745010023067 0ustar00a

��Re�@sVddlZddlmZmZmZddlmZddlmZm	Z	ddl
mZGdd�d�ZdS)�N)�List�Optional�Tuple)�Tag)�
get_supported�version_info_to_nodot)�normalize_version_infoc@speZdZdZgd�Zd
eeeeee	dfeeeeedd�dd�Z
ed�d	d
�Zeed�dd�Z
dS)�TargetPythonzx
    Encapsulates the properties of a Python interpreter one is targeting
    for a package install, download, etc.
    )�_given_py_version_info�abis�implementation�	platforms�
py_version�py_version_info�_valid_tagsN.)r
rrr�returncCsf||_|durtjdd�}nt|�}d�tt|dd���}||_||_||_	||_
||_d|_dS)a<
        :param platforms: A list of strings or None. If None, searches for
            packages that are supported by the current system. Otherwise, will
            find packages that can be built on the platforms passed in. These
            packages will only be downloaded for distribution: they will
            not be built locally.
        :param py_version_info: An optional tuple of ints representing the
            Python version information to use (e.g. `sys.version_info[:3]`).
            This can have length 1, 2, or 3 when provided.
        :param abis: A list of strings or None. This is passed to
            compatibility_tags.py's get_supported() function as is.
        :param implementation: A string or None. This is passed to
            compatibility_tags.py's get_supported() function as is.
        N��.�)
r
�sys�version_infor�join�map�strrrr
rrr)�selfr
rrrr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/target_python.py�__init__szTargetPython.__init__)rcCsZd}|jdur$d�dd�|jD��}d|jfd|fd|jfd|jfg}d	�d
d�|D��S)zD
        Format the given, non-None attributes for display.
        Nrcss|]}t|�VqdS)N)r)�.0�partrrr�	<genexpr>Isz,TargetPython.format_given.<locals>.<genexpr>r
rrr� css(|] \}}|dur|�d|��VqdS)N�=r)r�key�valuerrrr Ss)r
rr
rr)r�display_version�
key_valuesrrr�format_givenCs

��
�zTargetPython.format_givencCsH|jdurB|j}|durd}nt|�}t||j|j|jd�}||_|jS)z�
        Return the supported PEP 425 tags to check wheel candidates against.

        The tags are returned in order of preference (most preferred first).
        N)�versionr
r�impl)rr
rrr
rr)rrr(�tagsrrr�get_tagsWs
�zTargetPython.get_tags)NNNN)�__name__�
__module__�__qualname__�__doc__�	__slots__rrrr�intrr'rr+rrrrr	
s�

�(r	)
r�typingrrrZpip._vendor.packaging.tagsr�&pip._internal.utils.compatibility_tagsrr�pip._internal.utils.miscrr	rrrr�<module>s
site-packages/pip/_internal/models/__pycache__/wheel.cpython-39.pyc000064400000010423150564745010021266 0ustar00a

��Re�
�@sJdZddlZddlmZmZmZddlmZddlm	Z	Gdd�d�Z
dS)z`Represents a wheel file and provides access to the various parts of the
name that have meaning.
�N)�Dict�Iterable�List)�Tag)�InvalidWheelFilenamec@s�eZdZdZe�dej�Zedd�dd�Z	e
ed�dd	�Ze
ee
d
�dd�Ze
eeee
fe
d
�dd�Zeeed
�dd�ZdS)�WheelzA wheel filez�^(?P<namever>(?P<name>.+?)-(?P<ver>.*?))
        ((-(?P<build>\d[^-]*?))?-(?P<pyver>.+?)-(?P<abi>.+?)-(?P<plat>.+?)
        \.whl|\.dist-info)$N)�filename�returncs��j�|�}|st|�d���|�_|�d��dd��_|�d��dd��_|�d��_|�d��	d��_
|�d	��	d��_|�d
��	d��_�fdd��j
D��_
d
S)zX
        :raises InvalidWheelFilename: when the filename is invalid for a wheel
        z is not a valid wheel filename.�name�_�-�ver�build�pyver�.�abi�platcs0h|](}�jD]}�jD]}t|||��qqqS�)�abis�platsr)�.0�x�y�z��selfr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/wheel.py�	<setcomp>(sz!Wheel.__init__.<locals>.<setcomp>N)�
wheel_file_re�matchrr�group�replacer
�version�	build_tag�split�
pyversionsrr�	file_tags)rr�
wheel_inforrr�__init__s
�zWheel.__init__)r	cCstdd�|jD��S)z4Return the wheel's tags as a sorted list of strings.css|]}t|�VqdS�N)�str�r�tagrrr�	<genexpr>.�z0Wheel.get_formatted_file_tags.<locals>.<genexpr>)�sortedr&rrrr�get_formatted_file_tags,szWheel.get_formatted_file_tags)�tagsr	cst�fdd�|jD��S)a�Return the lowest index that one of the wheel's file_tag combinations
        achieves in the given list of supported tags.

        For example, if there are 8 supported tags and one of the file tags
        is first in the list, then return 0.

        :param tags: the PEP 425 tags to check the wheel against, in order
            with most preferred first.

        :raises ValueError: If none of the wheel's file tags match one of
            the supported tags.
        c3s |]}|�vr��|�VqdSr))�indexr+�r1rrr-=r.z*Wheel.support_index_min.<locals>.<genexpr>��minr&�rr1rr3r�support_index_min0s
zWheel.support_index_min)r1�tag_to_priorityr	cst�fdd�|jD��S)a�Return the priority of the most preferred tag that one of the wheel's file
        tag combinations achieves in the given list of supported tags using the given
        tag_to_priority mapping, where lower priorities are more-preferred.

        This is used in place of support_index_min in some cases in order to avoid
        an expensive linear scan of a large list of tags.

        :param tags: the PEP 425 tags to check the wheel against.
        :param tag_to_priority: a mapping from tag to priority of that tag, where
            lower is more preferred.

        :raises ValueError: If none of the wheel's file tags match one of
            the supported tags.
        c3s|]}|�vr�|VqdSr)rr+�r8rrr-Psz0Wheel.find_most_preferred_tag.<locals>.<genexpr>r4)rr1r8rr9r�find_most_preferred_tag?s�zWheel.find_most_preferred_tagcCs|j�|�S)z�Return whether the wheel is compatible with one of the given tags.

        :param tags: the PEP 425 tags to check the wheel against.
        )r&�
isdisjointr6rrr�	supportedTszWheel.supported)�__name__�
__module__�__qualname__�__doc__�re�compile�VERBOSErr*r(rr0r�intr7rr:r�boolr<rrrrrs��r)r@rA�typingrrrZpip._vendor.packaging.tagsr�pip._internal.exceptionsrrrrrr�<module>s
site-packages/pip/_internal/models/__pycache__/link.cpython-39.pyc000064400000024065150564745010021126 0ustar00a

��ReY&�@sddlZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
mZmZm
Z
mZddlmZddlmZddlmZmZmZddlmZddlmZmZer�ddlmZe�e�Z d	Z!Gd
d�de�Z"Gdd
�d
e�Z#e"e#d�dd�Z$ej%dd�e"e"e&d�dd��Z'dS)�N)�
TYPE_CHECKING�Dict�List�
NamedTuple�Optional�Tuple�Union)�WHEEL_EXTENSION)�Hashes)�redact_auth_from_url�split_auth_from_netloc�splitext)�KeyBasedCompareMixin)�path_to_url�url_to_path)�HTMLPage)�sha1�sha224�sha384�sha256�sha512�md5cs"eZdZdZgd�Zd>eeeedfeeeeedd��fdd�
Z	ed	�d
d�Z
ed	�dd
�Zeed	�dd��Z
eed	�dd��Zeed	�dd��Zeed	�dd��Zeed	�dd��Zeed	�dd��Zeeefd	�dd�Zeed	�dd��Zeed	�dd��Ze�d �Zeeed	�d!d"��Ze�d#�Zeeed	�d$d%��Ze�d&jd'�e�d(��Z eeed	�d)d*��Z!eeed	�d+d,��Z"eed	�d-d.��Z#eed	�d/d0��Z$ed	�d1d2�Z%eed	�d3d4��Z&eed	�d5d6��Z'eed	�d7d8��Z(eed	�d9d:��Z)ee*ed;�d<d=�Z+�Z,S)?�Linkz:Represents a parsed link from a Package Index's simple URL)�_parsed_url�_url�
comes_from�requires_python�
yanked_reason�cache_link_parsingNTr)�urlrrrr�returncsZ|�d�rt|�}tj�|�|_||_||_|r4|nd|_||_	t
�j|td�||_
dS)a�
        :param url: url of the resource pointed to (href of the link)
        :param comes_from: instance of HTMLPage where the link was found,
            or string.
        :param requires_python: String containing the `Requires-Python`
            metadata field, specified in PEP 345. This may be specified by
            a data-requires-python attribute in the HTML link tag, as
            described in PEP 503.
        :param yanked_reason: the reason the file has been yanked, if the
            file has been yanked, or None if the file hasn't been yanked.
            This is the value of the "data-yanked" attribute, if present, in
            a simple repository HTML link. If the file has been yanked but
            no reason was provided, this should be the empty string. See
            PEP 592 for more information and the specification.
        :param cache_link_parsing: A flag that is used elsewhere to determine
                                   whether resources retrieved from this link
                                   should be cached. PyPI index urls should
                                   generally have this set to False, for
                                   example.
        z\\N)�key�defining_class)�
startswithr�urllib�parse�urlsplitrrrrr�super�__init__rr)�selfrrrrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/link.pyr((s
z
Link.__init__)r cCsH|jrd|j�d�}nd}|jr6d�t|j�|j|�Stt|j��SdS)Nz (requires-python:�)�z{} (from {}){})rr�formatrr�str)r)�rpr,r,r-�__str__Vs�zLink.__str__cCsd|�d�S)Nz<Link �>r,�r)r,r,r-�__repr__bsz
Link.__repr__cCs|jS�N)rr5r,r,r-reszLink.urlcCsR|j�d�}t�|�}|s,t|j�\}}|Stj�|�}|sNJd|j	�d���|S)N�/zURL z produced no filename)
�path�rstrip�	posixpath�basenamer�netlocr$r%�unquoter)r)r9�namer=�	user_passr,r,r-�filenameis
z
Link.filenamecCs
t|j�Sr7)rrr5r,r,r-�	file_pathwszLink.file_pathcCs|jjSr7)r�schemer5r,r,r-rC{szLink.schemecCs|jjS)z4
        This can contain auth information.
        )rr=r5r,r,r-r=szLink.netloccCstj�|jj�Sr7)r$r%r>rr9r5r,r,r-r9�sz	Link.pathcCstt�|j�d���S)Nr8)r
r;r<r9r:r5r,r,r-r
�sz
Link.splitextcCs|��dS�N�)r
r5r,r,r-�ext�szLink.extcCs&|j\}}}}}tj�||||df�S)Nr/)rr$r%�
urlunsplit)r)rCr=r9�query�fragmentr,r,r-�url_without_fragment�szLink.url_without_fragmentz[#&]egg=([^&]*)cCs |j�|j�}|sdS|�d�SrD)�_egg_fragment_re�searchr�group�r)�matchr,r,r-�egg_fragment�szLink.egg_fragmentz[#&]subdirectory=([^&]*)cCs |j�|j�}|sdS|�d�SrD)�_subdirectory_fragment_rerLrrMrNr,r,r-�subdirectory_fragment�szLink.subdirectory_fragmentz({choices})=([a-f0-9]+)�|)�choicescCs |j�|j�}|r|�d�SdS)N���_hash_rerLrrMrNr,r,r-�hash�s
z	Link.hashcCs |j�|j�}|r|�d�SdSrDrVrNr,r,r-�	hash_name�s
zLink.hash_namecCs$t�|j�dd�d�dd�d�S)N�#rEr�?)r;r<r�splitr5r,r,r-�show_url�sz
Link.show_urlcCs
|jdkS)N�file)rCr5r,r,r-�is_file�szLink.is_filecCs|jotj�|j�Sr7)r_�osr9�isdirrBr5r,r,r-�is_existing_dir�szLink.is_existing_dircCs
|jtkSr7)rFr	r5r,r,r-�is_wheel�sz
Link.is_wheelcCsddlm}|j|jvS)Nr)�vcs)�pip._internal.vcsrdrC�all_schemes)r)rdr,r,r-�is_vcs�szLink.is_vcscCs
|jduSr7)rr5r,r,r-�	is_yanked�szLink.is_yankedcCs
|jduSr7)rYr5r,r,r-�has_hash�sz
Link.has_hash)�hashesr cCs@|dus|jsdS|jdus J�|jdus.J�|j|j|jd�S)zG
        Return True if the link has a hash and it is allowed.
        NF)�
hex_digest)rirYrX�is_hash_allowed)r)rjr,r,r-rl�s
zLink.is_hash_allowed)NNNT)-�__name__�
__module__�__qualname__�__doc__�	__slots__r1rr�boolr(r3r6�propertyrrArBrCr=r9rr
rFrJ�re�compilerKrPrQrRr0�join�_SUPPORTED_HASHESrWrXrYr]r_rbrcrgrhrir
rl�
__classcell__r,r,r*r-rsx��.


�rc@sJeZdZUdZejjed<ee	e
e	fed<e	ed<ee	e	fed<dS)�_CleanResulta�Convert link for equivalency check.

    This is used in the resolver to check whether two URL-specified requirements
    likely point to the same distribution and can be considered equivalent. This
    equivalency logic avoids comparing URLs literally, which can be too strict
    (e.g. "a=1&b=2" vs "b=2&a=1") and produce conflicts unexpecting to users.

    Currently this does three things:

    1. Drop the basic auth part. This is technically wrong since a server can
       serve different content based on auth, but if it does that, it is even
       impossible to guarantee two URLs without auth are equivalent, since
       the user can input different auth information when prompted. So the
       practical solution is to assume the auth doesn't affect the response.
    2. Parse the query to avoid the ordering issue. Note that ordering under the
       same key in the query are NOT cleaned; i.e. "a=1&a=2" and "a=2&a=1" are
       still considered different.
    3. Explicitly drop most of the fragment part, except ``subdirectory=`` and
       hash values, since it should have no impact the downloaded content. Note
       that this drops the "egg=" part historically used to denote the requested
       project (and extras), which is wrong in the strictest sense, but too many
       people are supplying it inconsistently to cause superfluous resolution
       conflicts, so we choose to also ignore them.
    �parsedrH�subdirectoryrjN)rmrnrorpr$r%�SplitResult�__annotations__rr1rr,r,r,r-ry�s

ry)�linkr c	s�|j}|j�dd�d}|jdkr*|s*d}tj�|j��d�vrLt�	d|�z�dd	}Wnt
tfyvd
}Yn0�fdd�tD�}t
|j|d
d
d
�tj�|j�||d�S)N�@rE���r^�	localhost�eggzIgnoring egg= fragment in %sr{rr/cs"i|]}|�vr|�|d�qS)rr,)�.0�k�rIr,r-�
<dictcomp>�z_clean_link.<locals>.<dictcomp>)r=rHrI)rzrHr{rj)rr=�rsplitrCr$r%�parse_qsrI�logger�debug�
IndexError�KeyErrorrwry�_replacerH)r~rzr=r{rjr,r�r-�_clean_links$
�r�)�maxsize)�link1�link2r cCst|�t|�kSr7)r�)r�r�r,r,r-�links_equivalentsr�)(�	functools�loggingr`r;rt�urllib.parser$�typingrrrrrrr�pip._internal.utils.filetypesr	�pip._internal.utils.hashesr
�pip._internal.utils.miscrrr
�pip._internal.utils.modelsr�pip._internal.utils.urlsrr�pip._internal.index.collectorr�	getLoggerrmr�rwrryr��	lru_cacherrr�r,r,r,r-�<module>s*$
I 
site-packages/pip/_internal/models/__pycache__/scheme.cpython-39.pyc000064400000002025150564745010021425 0ustar00a

��Re��@sdZgd�ZGdd�d�ZdS)z�
For types associated with installation schemes.

For a general overview of available schemes and their context, see
https://docs.python.org/3/install/index.html#alternate-installation.
��platlib�purelib�headers�scripts�datac@s,eZdZdZeZeeeeedd�dd�ZdS)�SchemeztA Scheme holds paths which are used as the base directories for
    artifacts associated with a Python package.
    N)rrrrr�returncCs"||_||_||_||_||_dS)Nr)�selfrrrrr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/scheme.py�__init__s
zScheme.__init__)�__name__�
__module__�__qualname__�__doc__�SCHEME_KEYS�	__slots__�strrr
r
r
rrs�rN)rrrr
r
r
r�<module>ssite-packages/pip/_internal/models/__pycache__/index.cpython-39.pyc000064400000002365150564745010021277 0ustar00a

��Re�@s2ddlZGdd�d�Zeddd�Zeddd�ZdS)	�NcsBeZdZdZgd�Zeedd��fdd�Zeed�dd	�Z�ZS)
�PackageIndexzBRepresents a Package Index and provides easier access to endpoints)�url�netloc�
simple_url�pypi_url�file_storage_domainN)rr�returncsBt���||_tj�|�j|_|�d�|_|�d�|_	||_
dS)N�simple�pypi)�super�__init__r�urllib�parse�urlsplitr�
_url_for_pathrrr)�selfrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/index.pyr	s
zPackageIndex.__init__)�pathrcCstj�|j|�S)N)r
r�urljoinr)rrrrrrszPackageIndex._url_for_path)	�__name__�
__module__�__qualname__�__doc__�	__slots__�strrr�
__classcell__rrrrrsrzhttps://pypi.org/zfiles.pythonhosted.org)rzhttps://test.pypi.org/ztest-files.pythonhosted.org)�urllib.parser
r�PyPI�TestPyPIrrrr�<module>s
�site-packages/pip/_internal/models/__pycache__/format_control.cpython-39.pyc000064400000005301150564745010023211 0ustar00a

��Re�	�@s>ddlmZmZmZddlmZddlmZGdd�d�ZdS)�)�	FrozenSet�Optional�Set)�canonicalize_name)�CommandErrorc@s�eZdZdZddgZdeeeeeedd�dd�Ze	e
d�d	d
�Zed�dd
�Ze
eeeeedd�dd��Zeeed�dd�Zdd�dd�ZdS)�
FormatControlzBHelper for managing formats from which a package can be installed.�	no_binary�only_binaryN)rr	�returncCs,|durt�}|durt�}||_||_dS�N)�setrr	)�selfrr	�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/format_control.py�__init__
szFormatControl.__init__)�otherr
cs:t��j�stS�j�jkr dSt��fdd��jD��S)NFc3s"|]}t�|�t�|�kVqdSr)�getattr)�.0�k�rr
rr�	<genexpr>!�z'FormatControl.__eq__.<locals>.<genexpr>)�
isinstance�	__class__�NotImplemented�	__slots__�all)r
rrrr�__eq__s
zFormatControl.__eq__)r
cCsd�|jj|j|j�S)Nz
{}({}, {}))�formatr�__name__rr	�r
rrr�__repr__#s�zFormatControl.__repr__)�value�targetrr
cCs�|�d�rtd��|�d�}d|vr`|��|��|�d�|d|�d�d�=d|vrdSq|D]2}|dkrz|��qdt|�}|�|�|�|�qddS)N�-z7--no-binary / --only-binary option requires 1 argument.�,�:all:�z:none:)�
startswithr�split�clear�add�indexr�discard)r"r#r�new�namerrr�handle_mutual_excludes(s&
�


z$FormatControl.handle_mutual_excludes)�canonical_namer
cCsfddh}||jvr|�d�n@||jvr4|�d�n*d|jvrJ|�d�nd|jvr^|�d�t|�S)N�binary�sourcer&)r	r-r�	frozenset)r
r1�resultrrr�get_allowed_formats?s




z!FormatControl.get_allowed_formatscCs|�d|j|j�dS)Nr&)r0rr	r rrr�disallow_binariesKs
�zFormatControl.disallow_binaries)NN)r�
__module__�__qualname__�__doc__rrr�strr�object�boolrr!�staticmethodr0rr6r7rrrrrs�

�
	rN)	�typingrrr�pip._vendor.packaging.utilsr�pip._internal.exceptionsrrrrrr�<module>ssite-packages/pip/_internal/models/__pycache__/search_scope.cpython-39.pyc000064400000006672150564745010022633 0ustar00a

��Re��@s�ddlZddlZddlZddlZddlZddlmZddlm	Z	ddl
mZddlm
Z
ddlmZmZe�e�ZGdd�d�ZdS)	�N)�List)�canonicalize_name)�PyPI)�has_tls)�normalize_path�redact_auth_from_urlc@sreZdZdZddgZeeeeedd�dd��Zeeeedd�dd	�Z	ed
�dd�Z
eeed
�dd�ZdS)�SearchScopezF
    Encapsulates the locations that pip is configured to search.
    �
find_links�
index_urls)r	r
�returncCs�g}|D]0}|�d�r.t|�}tj�|�r.|}|�|�qt�svt�||�D](}t	j
�|�}|jdkrLt
�d�qvqL|||d�S)zQ
        Create a SearchScope object after normalizing the `find_links`.
        �~�httpszipip is configured with locations that require TLS/SSL, however the ssl module in Python is not available.�r	r
)�
startswithr�os�path�exists�appendr�	itertools�chain�urllib�parse�urlparse�scheme�logger�warning)�clsr	r
�built_find_links�link�new_link�parsed�r!��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/search_scope.py�creates&

��zSearchScope.createNcCs||_||_dS�Nr)�selfr	r
r!r!r"�__init__AszSearchScope.__init__)rcCs�g}g}|jrt|jtjgkrt|jD]:}t|�}tj�|�}|jsR|jsRt	�
d|�|�|�q"|�d�d�
|���|jr�|�d�d�
dd�|jD����d�
|�S)Nz:The index url "%s" seems invalid, please provide a scheme.zLooking in indexes: {}z, zLooking in links: {}css|]}t|�VqdSr$)r��.0�urlr!r!r"�	<genexpr>h�z6SearchScope.get_formatted_locations.<locals>.<genexpr>�
)r
r�
simple_urlrrr�urlsplitr�netlocrrr�format�joinr	)r%�lines�redacted_index_urlsr)�redacted_index_url�purlr!r!r"�get_formatted_locationsIs,
����z#SearchScope.get_formatted_locations)�project_namercs(ttd��fdd���fdd�|jD�S)z�Returns the locations found via self.index_urls

        Checks the url_name on the main (first in the list) index and
        use this url_name to produce all locations
        )r)rcs.t�|tj�t����}|�d�s*|d}|S)N�/)�	posixpathr1rr�quoter�endswith)r)�loc)r7r!r"�mkurl_pypi_urlts�
z<SearchScope.get_index_urls_locations.<locals>.mkurl_pypi_urlcsg|]}�|��qSr!r!r')r=r!r"�
<listcomp>�r+z8SearchScope.get_index_urls_locations.<locals>.<listcomp>)�strr
)r%r7r!)r=r7r"�get_index_urls_locationsms
z$SearchScope.get_index_urls_locations)�__name__�
__module__�__qualname__�__doc__�	__slots__�classmethodrr?r#r&r6r@r!r!r!r"rs�)�$r)r�loggingrr9�urllib.parser�typingr�pip._vendor.packaging.utilsr�pip._internal.models.indexr�pip._internal.utils.compatr�pip._internal.utils.miscrr�	getLoggerrArrr!r!r!r"�<module>s
site-packages/pip/_internal/models/__pycache__/selection_prefs.cpython-39.pyc000064400000003245150564745010023352 0ustar00a

��Res�@s*ddlmZddlmZGdd�d�ZdS)�)�Optional)�
FormatControlc@s:eZdZdZgd�Zdeeeeeeedd�dd�ZdS)	�SelectionPreferenceszd
    Encapsulates the candidate selection preferences for downloading
    and installing files.
    ��allow_yanked�allow_all_prereleases�format_control�
prefer_binary�ignore_requires_pythonFN)rrrr	r
�returncCs.|durd}||_||_||_||_||_dS)awCreate a SelectionPreferences object.

        :param allow_yanked: Whether files marked as yanked (in the sense
            of PEP 592) are permitted to be candidates for install.
        :param format_control: A FormatControl object or None. Used to control
            the selection of source packages / binary packages when consulting
            the index and links.
        :param prefer_binary: Whether to prefer an old, but valid, binary
            dist over a new source dist.
        :param ignore_requires_python: Whether to ignore incompatible
            "Requires-Python" values in links. Defaults to False.
        NFr)�selfrrrr	r
�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/selection_prefs.py�__init__szSelectionPreferences.__init__)FNFN)	�__name__�
__module__�__qualname__�__doc__�	__slots__�boolrrrr
r
r
rrs��rN)�typingr�#pip._internal.models.format_controlrrr
r
r
r�<module>ssite-packages/pip/_internal/models/__pycache__/__init__.cpython-39.pyc000064400000000453150564745010021723 0ustar00a

��Re?�@sdZdS)z8A package that contains models that represent entities.
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/__init__.py�<module>�site-packages/pip/_internal/models/__pycache__/direct_url.cpython-39.pyc000064400000016150150564745010022321 0ustar00a

��Re��@s<dZddlZddlZddlZddlmZmZmZm	Z	m
Z
mZmZgd�Z
ed�ZdZe�d�ZGdd	�d	e�Zdeeefe
eee	ee	ed
�dd�Zdeeefe
eee	eed
�d
d�Zee	ddd�dd�Zeeeefd�dd�ZGdd�d�ZGdd�d�ZGdd�d�ZeeeefZGdd�d�ZdS) z	 PEP 610 �N)�Any�Dict�Iterable�Optional�Type�TypeVar�Union)�	DirectUrl�DirectUrlValidationError�DirInfo�ArchiveInfo�VcsInfo�Tzdirect_url.jsonz.^\$\{[A-Za-z0-9-_]+\}(:\$\{[A-Za-z0-9-_]+\})?$c@seZdZdS)r
N)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/models/direct_url.pyr
sr
)�d�
expected_type�key�default�returncCs4||vr|S||}t||�s0td�|||���|S)z3Get value from dictionary and verify expected type.z-{!r} has unexpected type for {} (expected {}))�
isinstancer
�format�rrrr�valuerrr�_gets
��rcCs(t||||�}|dur$t|�d���|S)Nz must have a value)rr
rrrr�
_get_required)sr�InfoType)�infosrcCsFdd�|D�}|std��t|�dkr.td��|ddus>J�|dS)NcSsg|]}|dur|�qS�Nr)�.0�inforrr�
<listcomp>3�z#_exactly_one_of.<locals>.<listcomp>z/missing one of archive_info, dir_info, vcs_info�z1more than one of archive_info, dir_info, vcs_infor)r
�len)r rrr�_exactly_one_of2s��r()�kwargsrcKsdd�|��D�S)z Make dict excluding None values.cSsi|]\}}|dur||�qSr!r)r"�k�vrrr�
<dictcomp>Br%z _filter_none.<locals>.<dictcomp>)�items)r)rrr�_filter_none@sr.c@speZdZdZdeeeeeeeedd�dd�Zeeeee	fedd�dd��Z
eee	fd	�d
d�ZdS)
r
�vcs_infoN)�vcs�	commit_id�requested_revision�resolved_revision�resolved_revision_typercCs"||_||_||_||_||_dSr!�r0r2r1r3r4)�selfr0r1r2r3r4rrr�__init__Hs
zVcsInfo.__init__�rrc	CsF|durdS|t|td�t|td�t|td�t|td�t|td�d�S)Nr0r1r2r3r4)r0r1r2r3r4)r�strr��clsrrrr�
_from_dictVs




�zVcsInfo._from_dict�rcCst|j|j|j|j|jd�S)Nr5)r.r0r2r1r3r4�r6rrr�_to_dictbs�zVcsInfo._to_dict)NNN)rrr�namer9rr7�classmethodrrr<r?rrrrr
Es��"r
c@s`eZdZdZdeedd�dd�Zeeeee	fedd�dd��Z
eee	fd	�d
d�ZdS)
r�archive_infoN)�hashrcCs
||_dSr!�rC)r6rCrrrr7oszArchiveInfo.__init__r8cCs|durdS|t|td�d�S)NrCrD)rr9r:rrrr<uszArchiveInfo._from_dictr=cCst|jd�S)NrD)r.rCr>rrrr?{szArchiveInfo._to_dict)N)rrrr@rr9r7rArrr<r?rrrrrls��"rc@s\eZdZdZd
edd�dd�Zeeee	e
fedd�dd	��Zee	e
fd
�dd�ZdS)r�dir_infoFN)�editablercCs
||_dSr!�rF)r6rFrrrr7�szDirInfo.__init__r8cCs"|durdS|t|tddd�d�S)NrFF)rrG)r�boolr:rrrr<�szDirInfo._from_dictr=cCst|jp
dd�S)NrG)r.rFr>rrrr?�szDirInfo._to_dict)F)
rrrr@rHr7rArrr9rr<r?rrrrrs��"rc@s�eZdZdeeeedd�dd�Zeed�dd�Zeed�d	d
��Z	dd�dd�Z
eeee
fdd
�dd��Zeee
fd�dd�Zeedd�dd��Zed�dd�Zed�dd�ZdS)r	N)�urlr#�subdirectoryrcCs||_||_||_dSr!)rIr#rJ)r6rIr#rJrrrr7�szDirectUrl.__init__)�netlocrcCsRd|vr|S|�dd�\}}t|jt�r@|jjdkr@|dkr@|St�|�rN|S|S)N�@r&�git)�splitrr#r
r0�
ENV_VAR_RE�match)r6rK�	user_pass�netloc_no_user_passrrr�_remove_auth_from_netloc�s
�
��
z"DirectUrl._remove_auth_from_netlocr=cCs<tj�|j�}|�|j�}tj�|j||j|j	|j
f�}|S)z�url with user:password part removed unless it is formed with
        environment variables as specified in PEP 610, or it is ``git``
        in the case of a git URL.
        )�urllib�parse�urlsplitrIrSrK�
urlunsplit�scheme�path�query�fragment)r6�purlrK�surlrrr�redacted_url�s�zDirectUrl.redacted_urlcCs|�|���dSr!)�	from_dict�to_dictr>rrr�validate�szDirectUrl.validater8cCsRtt|td�t|td�tt�t|td��t�t|td��t	�t|td��g�d�S)NrIrJrBrEr/)rIrJr#)
r	rr9rr(rr<�dictrr
r:rrrr_�s

���zDirectUrl.from_dictcCs&t|j|jd�}|j��||jj<|S)N)rIrJ)r.r^rJr#r?r@)r6�resrrrr`�s�zDirectUrl.to_dict)�srcCs|�t�|��Sr!)r_�json�loads)r;rdrrr�	from_json�szDirectUrl.from_jsoncCstj|��dd�S)NT)�	sort_keys)re�dumpsr`r>rrr�to_json�szDirectUrl.to_jsoncCst|jt�o|jjSr!)rr#rrFr>rrr�is_local_editable�szDirectUrl.is_local_editable)N)rrrr9rrr7rS�propertyr^rarArrr_r`rgrjrHrkrrrrr	�s$��

r	)N)N)�__doc__re�re�urllib.parserT�typingrrrrrrr�__all__r�DIRECT_URL_METADATA_NAME�compilerO�	Exceptionr
r9rrr(r.r
rrrr	rrrr�<module>s4$
����	'site-packages/pip/_internal/models/format_control.py000064400000004730150564745010016727 0ustar00from typing import FrozenSet, Optional, Set

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.exceptions import CommandError


class FormatControl:
    """Helper for managing formats from which a package can be installed."""

    __slots__ = ["no_binary", "only_binary"]

    def __init__(
        self,
        no_binary: Optional[Set[str]] = None,
        only_binary: Optional[Set[str]] = None,
    ) -> None:
        if no_binary is None:
            no_binary = set()
        if only_binary is None:
            only_binary = set()

        self.no_binary = no_binary
        self.only_binary = only_binary

    def __eq__(self, other: object) -> bool:
        if not isinstance(other, self.__class__):
            return NotImplemented

        if self.__slots__ != other.__slots__:
            return False

        return all(getattr(self, k) == getattr(other, k) for k in self.__slots__)

    def __repr__(self) -> str:
        return "{}({}, {})".format(
            self.__class__.__name__, self.no_binary, self.only_binary
        )

    @staticmethod
    def handle_mutual_excludes(value: str, target: Set[str], other: Set[str]) -> None:
        if value.startswith("-"):
            raise CommandError(
                "--no-binary / --only-binary option requires 1 argument."
            )
        new = value.split(",")
        while ":all:" in new:
            other.clear()
            target.clear()
            target.add(":all:")
            del new[: new.index(":all:") + 1]
            # Without a none, we want to discard everything as :all: covers it
            if ":none:" not in new:
                return
        for name in new:
            if name == ":none:":
                target.clear()
                continue
            name = canonicalize_name(name)
            other.discard(name)
            target.add(name)

    def get_allowed_formats(self, canonical_name: str) -> FrozenSet[str]:
        result = {"binary", "source"}
        if canonical_name in self.only_binary:
            result.discard("source")
        elif canonical_name in self.no_binary:
            result.discard("binary")
        elif ":all:" in self.only_binary:
            result.discard("source")
        elif ":all:" in self.no_binary:
            result.discard("binary")
        return frozenset(result)

    def disallow_binaries(self) -> None:
        self.handle_mutual_excludes(
            ":all:",
            self.no_binary,
            self.only_binary,
        )
site-packages/pip/_internal/models/index.py000064400000002006150564745010015000 0ustar00import urllib.parse


class PackageIndex:
    """Represents a Package Index and provides easier access to endpoints"""

    __slots__ = ["url", "netloc", "simple_url", "pypi_url", "file_storage_domain"]

    def __init__(self, url: str, file_storage_domain: str) -> None:
        super().__init__()
        self.url = url
        self.netloc = urllib.parse.urlsplit(url).netloc
        self.simple_url = self._url_for_path("simple")
        self.pypi_url = self._url_for_path("pypi")

        # This is part of a temporary hack used to block installs of PyPI
        # packages which depend on external urls only necessary until PyPI can
        # block such packages themselves
        self.file_storage_domain = file_storage_domain

    def _url_for_path(self, path: str) -> str:
        return urllib.parse.urljoin(self.url, path)


PyPI = PackageIndex("https://pypi.org/", file_storage_domain="files.pythonhosted.org")
TestPyPI = PackageIndex(
    "https://test.pypi.org/", file_storage_domain="test-files.pythonhosted.org"
)
site-packages/pip/_internal/models/link.py000064400000023131150564745010014630 0ustar00import functools
import logging
import os
import posixpath
import re
import urllib.parse
from typing import TYPE_CHECKING, Dict, List, NamedTuple, Optional, Tuple, Union

from pip._internal.utils.filetypes import WHEEL_EXTENSION
from pip._internal.utils.hashes import Hashes
from pip._internal.utils.misc import (
    redact_auth_from_url,
    split_auth_from_netloc,
    splitext,
)
from pip._internal.utils.models import KeyBasedCompareMixin
from pip._internal.utils.urls import path_to_url, url_to_path

if TYPE_CHECKING:
    from pip._internal.index.collector import HTMLPage

logger = logging.getLogger(__name__)


_SUPPORTED_HASHES = ("sha1", "sha224", "sha384", "sha256", "sha512", "md5")


class Link(KeyBasedCompareMixin):
    """Represents a parsed link from a Package Index's simple URL"""

    __slots__ = [
        "_parsed_url",
        "_url",
        "comes_from",
        "requires_python",
        "yanked_reason",
        "cache_link_parsing",
    ]

    def __init__(
        self,
        url: str,
        comes_from: Optional[Union[str, "HTMLPage"]] = None,
        requires_python: Optional[str] = None,
        yanked_reason: Optional[str] = None,
        cache_link_parsing: bool = True,
    ) -> None:
        """
        :param url: url of the resource pointed to (href of the link)
        :param comes_from: instance of HTMLPage where the link was found,
            or string.
        :param requires_python: String containing the `Requires-Python`
            metadata field, specified in PEP 345. This may be specified by
            a data-requires-python attribute in the HTML link tag, as
            described in PEP 503.
        :param yanked_reason: the reason the file has been yanked, if the
            file has been yanked, or None if the file hasn't been yanked.
            This is the value of the "data-yanked" attribute, if present, in
            a simple repository HTML link. If the file has been yanked but
            no reason was provided, this should be the empty string. See
            PEP 592 for more information and the specification.
        :param cache_link_parsing: A flag that is used elsewhere to determine
                                   whether resources retrieved from this link
                                   should be cached. PyPI index urls should
                                   generally have this set to False, for
                                   example.
        """

        # url can be a UNC windows share
        if url.startswith("\\\\"):
            url = path_to_url(url)

        self._parsed_url = urllib.parse.urlsplit(url)
        # Store the url as a private attribute to prevent accidentally
        # trying to set a new value.
        self._url = url

        self.comes_from = comes_from
        self.requires_python = requires_python if requires_python else None
        self.yanked_reason = yanked_reason

        super().__init__(key=url, defining_class=Link)

        self.cache_link_parsing = cache_link_parsing

    def __str__(self) -> str:
        if self.requires_python:
            rp = f" (requires-python:{self.requires_python})"
        else:
            rp = ""
        if self.comes_from:
            return "{} (from {}){}".format(
                redact_auth_from_url(self._url), self.comes_from, rp
            )
        else:
            return redact_auth_from_url(str(self._url))

    def __repr__(self) -> str:
        return f"<Link {self}>"

    @property
    def url(self) -> str:
        return self._url

    @property
    def filename(self) -> str:
        path = self.path.rstrip("/")
        name = posixpath.basename(path)
        if not name:
            # Make sure we don't leak auth information if the netloc
            # includes a username and password.
            netloc, user_pass = split_auth_from_netloc(self.netloc)
            return netloc

        name = urllib.parse.unquote(name)
        assert name, f"URL {self._url!r} produced no filename"
        return name

    @property
    def file_path(self) -> str:
        return url_to_path(self.url)

    @property
    def scheme(self) -> str:
        return self._parsed_url.scheme

    @property
    def netloc(self) -> str:
        """
        This can contain auth information.
        """
        return self._parsed_url.netloc

    @property
    def path(self) -> str:
        return urllib.parse.unquote(self._parsed_url.path)

    def splitext(self) -> Tuple[str, str]:
        return splitext(posixpath.basename(self.path.rstrip("/")))

    @property
    def ext(self) -> str:
        return self.splitext()[1]

    @property
    def url_without_fragment(self) -> str:
        scheme, netloc, path, query, fragment = self._parsed_url
        return urllib.parse.urlunsplit((scheme, netloc, path, query, ""))

    _egg_fragment_re = re.compile(r"[#&]egg=([^&]*)")

    @property
    def egg_fragment(self) -> Optional[str]:
        match = self._egg_fragment_re.search(self._url)
        if not match:
            return None
        return match.group(1)

    _subdirectory_fragment_re = re.compile(r"[#&]subdirectory=([^&]*)")

    @property
    def subdirectory_fragment(self) -> Optional[str]:
        match = self._subdirectory_fragment_re.search(self._url)
        if not match:
            return None
        return match.group(1)

    _hash_re = re.compile(
        r"({choices})=([a-f0-9]+)".format(choices="|".join(_SUPPORTED_HASHES))
    )

    @property
    def hash(self) -> Optional[str]:
        match = self._hash_re.search(self._url)
        if match:
            return match.group(2)
        return None

    @property
    def hash_name(self) -> Optional[str]:
        match = self._hash_re.search(self._url)
        if match:
            return match.group(1)
        return None

    @property
    def show_url(self) -> str:
        return posixpath.basename(self._url.split("#", 1)[0].split("?", 1)[0])

    @property
    def is_file(self) -> bool:
        return self.scheme == "file"

    def is_existing_dir(self) -> bool:
        return self.is_file and os.path.isdir(self.file_path)

    @property
    def is_wheel(self) -> bool:
        return self.ext == WHEEL_EXTENSION

    @property
    def is_vcs(self) -> bool:
        from pip._internal.vcs import vcs

        return self.scheme in vcs.all_schemes

    @property
    def is_yanked(self) -> bool:
        return self.yanked_reason is not None

    @property
    def has_hash(self) -> bool:
        return self.hash_name is not None

    def is_hash_allowed(self, hashes: Optional[Hashes]) -> bool:
        """
        Return True if the link has a hash and it is allowed.
        """
        if hashes is None or not self.has_hash:
            return False
        # Assert non-None so mypy knows self.hash_name and self.hash are str.
        assert self.hash_name is not None
        assert self.hash is not None

        return hashes.is_hash_allowed(self.hash_name, hex_digest=self.hash)


class _CleanResult(NamedTuple):
    """Convert link for equivalency check.

    This is used in the resolver to check whether two URL-specified requirements
    likely point to the same distribution and can be considered equivalent. This
    equivalency logic avoids comparing URLs literally, which can be too strict
    (e.g. "a=1&b=2" vs "b=2&a=1") and produce conflicts unexpecting to users.

    Currently this does three things:

    1. Drop the basic auth part. This is technically wrong since a server can
       serve different content based on auth, but if it does that, it is even
       impossible to guarantee two URLs without auth are equivalent, since
       the user can input different auth information when prompted. So the
       practical solution is to assume the auth doesn't affect the response.
    2. Parse the query to avoid the ordering issue. Note that ordering under the
       same key in the query are NOT cleaned; i.e. "a=1&a=2" and "a=2&a=1" are
       still considered different.
    3. Explicitly drop most of the fragment part, except ``subdirectory=`` and
       hash values, since it should have no impact the downloaded content. Note
       that this drops the "egg=" part historically used to denote the requested
       project (and extras), which is wrong in the strictest sense, but too many
       people are supplying it inconsistently to cause superfluous resolution
       conflicts, so we choose to also ignore them.
    """

    parsed: urllib.parse.SplitResult
    query: Dict[str, List[str]]
    subdirectory: str
    hashes: Dict[str, str]


def _clean_link(link: Link) -> _CleanResult:
    parsed = link._parsed_url
    netloc = parsed.netloc.rsplit("@", 1)[-1]
    # According to RFC 8089, an empty host in file: means localhost.
    if parsed.scheme == "file" and not netloc:
        netloc = "localhost"
    fragment = urllib.parse.parse_qs(parsed.fragment)
    if "egg" in fragment:
        logger.debug("Ignoring egg= fragment in %s", link)
    try:
        # If there are multiple subdirectory values, use the first one.
        # This matches the behavior of Link.subdirectory_fragment.
        subdirectory = fragment["subdirectory"][0]
    except (IndexError, KeyError):
        subdirectory = ""
    # If there are multiple hash values under the same algorithm, use the
    # first one. This matches the behavior of Link.hash_value.
    hashes = {k: fragment[k][0] for k in _SUPPORTED_HASHES if k in fragment}
    return _CleanResult(
        parsed=parsed._replace(netloc=netloc, query="", fragment=""),
        query=urllib.parse.parse_qs(parsed.query),
        subdirectory=subdirectory,
        hashes=hashes,
    )


@functools.lru_cache(maxsize=None)
def links_equivalent(link1: Link, link2: Link) -> bool:
    return _clean_link(link1) == _clean_link(link2)
site-packages/pip/_internal/models/wheel.py000064400000006654150564745010015012 0ustar00"""Represents a wheel file and provides access to the various parts of the
name that have meaning.
"""
import re
from typing import Dict, Iterable, List

from pip._vendor.packaging.tags import Tag

from pip._internal.exceptions import InvalidWheelFilename


class Wheel:
    """A wheel file"""

    wheel_file_re = re.compile(
        r"""^(?P<namever>(?P<name>.+?)-(?P<ver>.*?))
        ((-(?P<build>\d[^-]*?))?-(?P<pyver>.+?)-(?P<abi>.+?)-(?P<plat>.+?)
        \.whl|\.dist-info)$""",
        re.VERBOSE,
    )

    def __init__(self, filename: str) -> None:
        """
        :raises InvalidWheelFilename: when the filename is invalid for a wheel
        """
        wheel_info = self.wheel_file_re.match(filename)
        if not wheel_info:
            raise InvalidWheelFilename(f"{filename} is not a valid wheel filename.")
        self.filename = filename
        self.name = wheel_info.group("name").replace("_", "-")
        # we'll assume "_" means "-" due to wheel naming scheme
        # (https://github.com/pypa/pip/issues/1150)
        self.version = wheel_info.group("ver").replace("_", "-")
        self.build_tag = wheel_info.group("build")
        self.pyversions = wheel_info.group("pyver").split(".")
        self.abis = wheel_info.group("abi").split(".")
        self.plats = wheel_info.group("plat").split(".")

        # All the tag combinations from this file
        self.file_tags = {
            Tag(x, y, z) for x in self.pyversions for y in self.abis for z in self.plats
        }

    def get_formatted_file_tags(self) -> List[str]:
        """Return the wheel's tags as a sorted list of strings."""
        return sorted(str(tag) for tag in self.file_tags)

    def support_index_min(self, tags: List[Tag]) -> int:
        """Return the lowest index that one of the wheel's file_tag combinations
        achieves in the given list of supported tags.

        For example, if there are 8 supported tags and one of the file tags
        is first in the list, then return 0.

        :param tags: the PEP 425 tags to check the wheel against, in order
            with most preferred first.

        :raises ValueError: If none of the wheel's file tags match one of
            the supported tags.
        """
        return min(tags.index(tag) for tag in self.file_tags if tag in tags)

    def find_most_preferred_tag(
        self, tags: List[Tag], tag_to_priority: Dict[Tag, int]
    ) -> int:
        """Return the priority of the most preferred tag that one of the wheel's file
        tag combinations achieves in the given list of supported tags using the given
        tag_to_priority mapping, where lower priorities are more-preferred.

        This is used in place of support_index_min in some cases in order to avoid
        an expensive linear scan of a large list of tags.

        :param tags: the PEP 425 tags to check the wheel against.
        :param tag_to_priority: a mapping from tag to priority of that tag, where
            lower is more preferred.

        :raises ValueError: If none of the wheel's file tags match one of
            the supported tags.
        """
        return min(
            tag_to_priority[tag] for tag in self.file_tags if tag in tag_to_priority
        )

    def supported(self, tags: Iterable[Tag]) -> bool:
        """Return whether the wheel is compatible with one of the given tags.

        :param tags: the PEP 425 tags to check the wheel against.
        """
        return not self.file_tags.isdisjoint(tags)
site-packages/pip/_internal/models/target_python.py000064400000007422150564745010016567 0ustar00import sys
from typing import List, Optional, Tuple

from pip._vendor.packaging.tags import Tag

from pip._internal.utils.compatibility_tags import get_supported, version_info_to_nodot
from pip._internal.utils.misc import normalize_version_info


class TargetPython:

    """
    Encapsulates the properties of a Python interpreter one is targeting
    for a package install, download, etc.
    """

    __slots__ = [
        "_given_py_version_info",
        "abis",
        "implementation",
        "platforms",
        "py_version",
        "py_version_info",
        "_valid_tags",
    ]

    def __init__(
        self,
        platforms: Optional[List[str]] = None,
        py_version_info: Optional[Tuple[int, ...]] = None,
        abis: Optional[List[str]] = None,
        implementation: Optional[str] = None,
    ) -> None:
        """
        :param platforms: A list of strings or None. If None, searches for
            packages that are supported by the current system. Otherwise, will
            find packages that can be built on the platforms passed in. These
            packages will only be downloaded for distribution: they will
            not be built locally.
        :param py_version_info: An optional tuple of ints representing the
            Python version information to use (e.g. `sys.version_info[:3]`).
            This can have length 1, 2, or 3 when provided.
        :param abis: A list of strings or None. This is passed to
            compatibility_tags.py's get_supported() function as is.
        :param implementation: A string or None. This is passed to
            compatibility_tags.py's get_supported() function as is.
        """
        # Store the given py_version_info for when we call get_supported().
        self._given_py_version_info = py_version_info

        if py_version_info is None:
            py_version_info = sys.version_info[:3]
        else:
            py_version_info = normalize_version_info(py_version_info)

        py_version = ".".join(map(str, py_version_info[:2]))

        self.abis = abis
        self.implementation = implementation
        self.platforms = platforms
        self.py_version = py_version
        self.py_version_info = py_version_info

        # This is used to cache the return value of get_tags().
        self._valid_tags: Optional[List[Tag]] = None

    def format_given(self) -> str:
        """
        Format the given, non-None attributes for display.
        """
        display_version = None
        if self._given_py_version_info is not None:
            display_version = ".".join(
                str(part) for part in self._given_py_version_info
            )

        key_values = [
            ("platforms", self.platforms),
            ("version_info", display_version),
            ("abis", self.abis),
            ("implementation", self.implementation),
        ]
        return " ".join(
            f"{key}={value!r}" for key, value in key_values if value is not None
        )

    def get_tags(self) -> List[Tag]:
        """
        Return the supported PEP 425 tags to check wheel candidates against.

        The tags are returned in order of preference (most preferred first).
        """
        if self._valid_tags is None:
            # Pass versions=None if no py_version_info was given since
            # versions=None uses special default logic.
            py_version_info = self._given_py_version_info
            if py_version_info is None:
                version = None
            else:
                version = version_info_to_nodot(py_version_info)

            tags = get_supported(
                version=version,
                platforms=self.platforms,
                abis=self.abis,
                impl=self.implementation,
            )
            self._valid_tags = tags

        return self._valid_tags
site-packages/pip/_internal/models/direct_url.py000064400000014316150564745010016034 0ustar00""" PEP 610 """
import json
import re
import urllib.parse
from typing import Any, Dict, Iterable, Optional, Type, TypeVar, Union

__all__ = [
    "DirectUrl",
    "DirectUrlValidationError",
    "DirInfo",
    "ArchiveInfo",
    "VcsInfo",
]

T = TypeVar("T")

DIRECT_URL_METADATA_NAME = "direct_url.json"
ENV_VAR_RE = re.compile(r"^\$\{[A-Za-z0-9-_]+\}(:\$\{[A-Za-z0-9-_]+\})?$")


class DirectUrlValidationError(Exception):
    pass


def _get(
    d: Dict[str, Any], expected_type: Type[T], key: str, default: Optional[T] = None
) -> Optional[T]:
    """Get value from dictionary and verify expected type."""
    if key not in d:
        return default
    value = d[key]
    if not isinstance(value, expected_type):
        raise DirectUrlValidationError(
            "{!r} has unexpected type for {} (expected {})".format(
                value, key, expected_type
            )
        )
    return value


def _get_required(
    d: Dict[str, Any], expected_type: Type[T], key: str, default: Optional[T] = None
) -> T:
    value = _get(d, expected_type, key, default)
    if value is None:
        raise DirectUrlValidationError(f"{key} must have a value")
    return value


def _exactly_one_of(infos: Iterable[Optional["InfoType"]]) -> "InfoType":
    infos = [info for info in infos if info is not None]
    if not infos:
        raise DirectUrlValidationError(
            "missing one of archive_info, dir_info, vcs_info"
        )
    if len(infos) > 1:
        raise DirectUrlValidationError(
            "more than one of archive_info, dir_info, vcs_info"
        )
    assert infos[0] is not None
    return infos[0]


def _filter_none(**kwargs: Any) -> Dict[str, Any]:
    """Make dict excluding None values."""
    return {k: v for k, v in kwargs.items() if v is not None}


class VcsInfo:
    name = "vcs_info"

    def __init__(
        self,
        vcs: str,
        commit_id: str,
        requested_revision: Optional[str] = None,
        resolved_revision: Optional[str] = None,
        resolved_revision_type: Optional[str] = None,
    ) -> None:
        self.vcs = vcs
        self.requested_revision = requested_revision
        self.commit_id = commit_id
        self.resolved_revision = resolved_revision
        self.resolved_revision_type = resolved_revision_type

    @classmethod
    def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["VcsInfo"]:
        if d is None:
            return None
        return cls(
            vcs=_get_required(d, str, "vcs"),
            commit_id=_get_required(d, str, "commit_id"),
            requested_revision=_get(d, str, "requested_revision"),
            resolved_revision=_get(d, str, "resolved_revision"),
            resolved_revision_type=_get(d, str, "resolved_revision_type"),
        )

    def _to_dict(self) -> Dict[str, Any]:
        return _filter_none(
            vcs=self.vcs,
            requested_revision=self.requested_revision,
            commit_id=self.commit_id,
            resolved_revision=self.resolved_revision,
            resolved_revision_type=self.resolved_revision_type,
        )


class ArchiveInfo:
    name = "archive_info"

    def __init__(
        self,
        hash: Optional[str] = None,
    ) -> None:
        self.hash = hash

    @classmethod
    def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["ArchiveInfo"]:
        if d is None:
            return None
        return cls(hash=_get(d, str, "hash"))

    def _to_dict(self) -> Dict[str, Any]:
        return _filter_none(hash=self.hash)


class DirInfo:
    name = "dir_info"

    def __init__(
        self,
        editable: bool = False,
    ) -> None:
        self.editable = editable

    @classmethod
    def _from_dict(cls, d: Optional[Dict[str, Any]]) -> Optional["DirInfo"]:
        if d is None:
            return None
        return cls(editable=_get_required(d, bool, "editable", default=False))

    def _to_dict(self) -> Dict[str, Any]:
        return _filter_none(editable=self.editable or None)


InfoType = Union[ArchiveInfo, DirInfo, VcsInfo]


class DirectUrl:
    def __init__(
        self,
        url: str,
        info: InfoType,
        subdirectory: Optional[str] = None,
    ) -> None:
        self.url = url
        self.info = info
        self.subdirectory = subdirectory

    def _remove_auth_from_netloc(self, netloc: str) -> str:
        if "@" not in netloc:
            return netloc
        user_pass, netloc_no_user_pass = netloc.split("@", 1)
        if (
            isinstance(self.info, VcsInfo)
            and self.info.vcs == "git"
            and user_pass == "git"
        ):
            return netloc
        if ENV_VAR_RE.match(user_pass):
            return netloc
        return netloc_no_user_pass

    @property
    def redacted_url(self) -> str:
        """url with user:password part removed unless it is formed with
        environment variables as specified in PEP 610, or it is ``git``
        in the case of a git URL.
        """
        purl = urllib.parse.urlsplit(self.url)
        netloc = self._remove_auth_from_netloc(purl.netloc)
        surl = urllib.parse.urlunsplit(
            (purl.scheme, netloc, purl.path, purl.query, purl.fragment)
        )
        return surl

    def validate(self) -> None:
        self.from_dict(self.to_dict())

    @classmethod
    def from_dict(cls, d: Dict[str, Any]) -> "DirectUrl":
        return DirectUrl(
            url=_get_required(d, str, "url"),
            subdirectory=_get(d, str, "subdirectory"),
            info=_exactly_one_of(
                [
                    ArchiveInfo._from_dict(_get(d, dict, "archive_info")),
                    DirInfo._from_dict(_get(d, dict, "dir_info")),
                    VcsInfo._from_dict(_get(d, dict, "vcs_info")),
                ]
            ),
        )

    def to_dict(self) -> Dict[str, Any]:
        res = _filter_none(
            url=self.redacted_url,
            subdirectory=self.subdirectory,
        )
        res[self.info.name] = self.info._to_dict()
        return res

    @classmethod
    def from_json(cls, s: str) -> "DirectUrl":
        return cls.from_dict(json.loads(s))

    def to_json(self) -> str:
        return json.dumps(self.to_dict(), sort_keys=True)

    def is_local_editable(self) -> bool:
        return isinstance(self.info, DirInfo) and self.info.editable
site-packages/pip/_internal/models/__init__.py000064400000000077150564745010015436 0ustar00"""A package that contains models that represent entities.
"""
site-packages/pip/_internal/models/search_scope.py000064400000010650150564745010016333 0ustar00import itertools
import logging
import os
import posixpath
import urllib.parse
from typing import List

from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.models.index import PyPI
from pip._internal.utils.compat import has_tls
from pip._internal.utils.misc import normalize_path, redact_auth_from_url

logger = logging.getLogger(__name__)


class SearchScope:

    """
    Encapsulates the locations that pip is configured to search.
    """

    __slots__ = ["find_links", "index_urls"]

    @classmethod
    def create(
        cls,
        find_links: List[str],
        index_urls: List[str],
    ) -> "SearchScope":
        """
        Create a SearchScope object after normalizing the `find_links`.
        """
        # Build find_links. If an argument starts with ~, it may be
        # a local file relative to a home directory. So try normalizing
        # it and if it exists, use the normalized version.
        # This is deliberately conservative - it might be fine just to
        # blindly normalize anything starting with a ~...
        built_find_links: List[str] = []
        for link in find_links:
            if link.startswith("~"):
                new_link = normalize_path(link)
                if os.path.exists(new_link):
                    link = new_link
            built_find_links.append(link)

        # If we don't have TLS enabled, then WARN if anyplace we're looking
        # relies on TLS.
        if not has_tls():
            for link in itertools.chain(index_urls, built_find_links):
                parsed = urllib.parse.urlparse(link)
                if parsed.scheme == "https":
                    logger.warning(
                        "pip is configured with locations that require "
                        "TLS/SSL, however the ssl module in Python is not "
                        "available."
                    )
                    break

        return cls(
            find_links=built_find_links,
            index_urls=index_urls,
        )

    def __init__(
        self,
        find_links: List[str],
        index_urls: List[str],
    ) -> None:
        self.find_links = find_links
        self.index_urls = index_urls

    def get_formatted_locations(self) -> str:
        lines = []
        redacted_index_urls = []
        if self.index_urls and self.index_urls != [PyPI.simple_url]:
            for url in self.index_urls:

                redacted_index_url = redact_auth_from_url(url)

                # Parse the URL
                purl = urllib.parse.urlsplit(redacted_index_url)

                # URL is generally invalid if scheme and netloc is missing
                # there are issues with Python and URL parsing, so this test
                # is a bit crude. See bpo-20271, bpo-23505. Python doesn't
                # always parse invalid URLs correctly - it should raise
                # exceptions for malformed URLs
                if not purl.scheme and not purl.netloc:
                    logger.warning(
                        'The index url "%s" seems invalid, please provide a scheme.',
                        redacted_index_url,
                    )

                redacted_index_urls.append(redacted_index_url)

            lines.append(
                "Looking in indexes: {}".format(", ".join(redacted_index_urls))
            )

        if self.find_links:
            lines.append(
                "Looking in links: {}".format(
                    ", ".join(redact_auth_from_url(url) for url in self.find_links)
                )
            )
        return "\n".join(lines)

    def get_index_urls_locations(self, project_name: str) -> List[str]:
        """Returns the locations found via self.index_urls

        Checks the url_name on the main (first in the list) index and
        use this url_name to produce all locations
        """

        def mkurl_pypi_url(url: str) -> str:
            loc = posixpath.join(
                url, urllib.parse.quote(canonicalize_name(project_name))
            )
            # For maximum compatibility with easy_install, ensure the path
            # ends in a trailing slash.  Although this isn't in the spec
            # (and PyPI can handle it without the slash) some other index
            # implementations might break if they relied on easy_install's
            # behavior.
            if not loc.endswith("/"):
                loc = loc + "/"
            return loc

        return [mkurl_pypi_url(url) for url in self.index_urls]
site-packages/pip/_internal/models/selection_prefs.py000064400000003563150564745010017066 0ustar00from typing import Optional

from pip._internal.models.format_control import FormatControl


class SelectionPreferences:
    """
    Encapsulates the candidate selection preferences for downloading
    and installing files.
    """

    __slots__ = [
        "allow_yanked",
        "allow_all_prereleases",
        "format_control",
        "prefer_binary",
        "ignore_requires_python",
    ]

    # Don't include an allow_yanked default value to make sure each call
    # site considers whether yanked releases are allowed. This also causes
    # that decision to be made explicit in the calling code, which helps
    # people when reading the code.
    def __init__(
        self,
        allow_yanked: bool,
        allow_all_prereleases: bool = False,
        format_control: Optional[FormatControl] = None,
        prefer_binary: bool = False,
        ignore_requires_python: Optional[bool] = None,
    ) -> None:
        """Create a SelectionPreferences object.

        :param allow_yanked: Whether files marked as yanked (in the sense
            of PEP 592) are permitted to be candidates for install.
        :param format_control: A FormatControl object or None. Used to control
            the selection of source packages / binary packages when consulting
            the index and links.
        :param prefer_binary: Whether to prefer an old, but valid, binary
            dist over a new source dist.
        :param ignore_requires_python: Whether to ignore incompatible
            "Requires-Python" values in links. Defaults to False.
        """
        if ignore_requires_python is None:
            ignore_requires_python = False

        self.allow_yanked = allow_yanked
        self.allow_all_prereleases = allow_all_prereleases
        self.format_control = format_control
        self.prefer_binary = prefer_binary
        self.ignore_requires_python = ignore_requires_python
site-packages/pip/_internal/models/scheme.py000064400000001342150564745010015137 0ustar00"""
For types associated with installation schemes.

For a general overview of available schemes and their context, see
https://docs.python.org/3/install/index.html#alternate-installation.
"""


SCHEME_KEYS = ["platlib", "purelib", "headers", "scripts", "data"]


class Scheme:
    """A Scheme holds paths which are used as the base directories for
    artifacts associated with a Python package.
    """

    __slots__ = SCHEME_KEYS

    def __init__(
        self,
        platlib: str,
        purelib: str,
        headers: str,
        scripts: str,
        data: str,
    ) -> None:
        self.platlib = platlib
        self.purelib = purelib
        self.headers = headers
        self.scripts = scripts
        self.data = data
site-packages/pip/_internal/models/candidate.py000064400000001736150564745010015616 0ustar00from pip._vendor.packaging.version import parse as parse_version

from pip._internal.models.link import Link
from pip._internal.utils.models import KeyBasedCompareMixin


class InstallationCandidate(KeyBasedCompareMixin):
    """Represents a potential "candidate" for installation."""

    __slots__ = ["name", "version", "link"]

    def __init__(self, name: str, version: str, link: Link) -> None:
        self.name = name
        self.version = parse_version(version)
        self.link = link

        super().__init__(
            key=(self.name, self.version, self.link),
            defining_class=InstallationCandidate,
        )

    def __repr__(self) -> str:
        return "<InstallationCandidate({!r}, {!r}, {!r})>".format(
            self.name,
            self.version,
            self.link,
        )

    def __str__(self) -> str:
        return "{!r} candidate (version {} at {})".format(
            self.name,
            self.version,
            self.link,
        )
site-packages/pip/_internal/network/__pycache__/utils.cpython-39.pyc000064400000002671150564745010021536 0ustar00a

��Re��@stUddlmZmZddlmZmZddlmZddiZee	e	fe
d<edd�d	d
�Zefeeee
d�dd
�ZdS)�)�Dict�Iterator)�CONTENT_CHUNK_SIZE�Response)�NetworkConnectionErrorzAccept-Encoding�identity�HEADERSN)�resp�returncCs�d}t|jt�rBz|j�d�}WqHty>|j�d�}YqH0n|j}d|jkr^dkr|nn|j�d|�d|j��}n2d|jkr�dkr�nn|j�d	|�d|j��}|r�t||d
��dS)N�zutf-8z
iso-8859-1i�i�z Client Error: z
 for url: iXz Server Error: )�response)�
isinstance�reason�bytes�decode�UnicodeDecodeError�status_code�urlr)r	�http_error_msgr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/utils.py�raise_for_statuss��r)r�
chunk_sizer
ccsRz |jj|dd�D]
}|VqWn,tyL|j�|�}|s@qH|Vq.Yn0dS)z3Given a requests Response, provide the data chunks.F)�decode_contentN)�raw�stream�AttributeError�read)rr�chunkrrr�response_chunks9s�
r)�typingrrZpip._vendor.requests.modelsrr�pip._internal.exceptionsrr�str�__annotations__r�intrrrrrr�<module>s��site-packages/pip/_internal/network/__pycache__/download.cpython-39.pyc000064400000012616150564745010022205 0ustar00a

��Re��@sTdZddlZddlZddlZddlZddlmZmZmZddl	m
Z
mZddlm
Z
ddlmZddlmZddlmZdd	lmZdd
lmZddlmZmZmZddlmZmZmZe� e!�Z"eee#d
�dd�Z$eee%ee&d�dd�Z'e%e%d�dd�Z(e%e%e%d�dd�Z)eee%d�dd�Z*eeed�dd�Z+Gdd �d �Z,Gd!d"�d"�Z-dS)#z)Download files with progress indicators.
�N)�Iterable�Optional�Tuple)�CONTENT_CHUNK_SIZE�Response)�DownloadProgressProvider)�NetworkConnectionError)�PyPI)�Link)�
is_from_cache)�
PipSession)�HEADERS�raise_for_status�response_chunks)�format_size�redact_auth_from_url�splitext)�resp�returnc
Cs0zt|jd�WStttfy*YdS0dS)Nzcontent-length)�int�headers�
ValueError�KeyError�	TypeError)r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/download.py�_get_http_response_sizesr)r�link�progress_barrcCs�t|�}|jtjkr|j}n|j}t|�}|r>d�|t|��}t	|�rTt
�d|�nt
�d|�t
��t
jkrtd}n*t	|�r�d}n|s�d}n|dkr�d}nd}t|t�}|s�|St||d�|�S)Nz{} ({})zUsing cached %szDownloading %sFTi@�)�max)r�netlocr	�file_storage_domain�show_url�url_without_fragmentr�formatrr�logger�info�getEffectiveLevel�logging�INFOrrr)rrr�total_length�url�
logged_url�
show_progress�chunksrrr�_prepare_downloads.
r/)�filenamercCstj�|�S)zJ
    Sanitize the "filename" value from a Content-Disposition header.
    )�os�path�basename)r0rrr�sanitize_content_filenameGsr4)�content_disposition�default_filenamercCs,t�|�\}}|�d�}|r$t|�}|p*|S)z�
    Parse the "filename" value from a Content-Disposition header, and
    return the default filename if the result is empty.
    r0)�cgi�parse_header�getr4)r5r6�_type�paramsr0rrr�parse_content_dispositionNs

r<)rrrcCs�|j}|j�d�}|r t||�}t|�d}|sPt�|j�dd��}|rP||7}|s~|j|jkr~tj	�|j�d}|r~||7}|S)zoGet an ideal filename from the given HTTP response, falling back to
    the link filename if not provided.
    zcontent-disposition�zcontent-type�)
r0rr9r<r�	mimetypes�guess_extensionr+r1r2)rrr0r5�extrrr�_get_http_response_filename\s
rB)�sessionrrcCs.|j�dd�d}|j|tdd�}t|�|S)N�#r=rT)r�stream)r+�splitr9r
r)rCr�
target_urlrrrr�_http_get_downloadqsrHc@s8eZdZeedd�dd�Zeeeeefd�dd�ZdS)�
DownloaderN�rCrrcCs||_||_dS�N��_session�
_progress_bar��selfrCrrrr�__init__yszDownloader.__init__)r�locationrc
Cs�zt|j|�}WnFtyV}z.|jdus.J�t�d|jj|��WYd}~n
d}~00t||�}tj	�
||�}t|||j�}t
|d��$}|D]}	|�|	�q�Wd�n1s�0Y|j�dd�}
||
fS)z.Download the file given by link into location.N�HTTP error %s while getting %s�wb�Content-Typer>�rHrMr�responser%�critical�status_coderBr1r2�joinr/rN�open�writerr9)rPrrRr�er0�filepathr.�content_file�chunk�content_typerrr�__call__�s 
�
*zDownloader.__call__)	�__name__�
__module__�__qualname__r�strrQr
rrbrrrrrIxs
�rIc@sHeZdZeedd�dd�Zeeeeeeeeeffd�dd�Z	dS)�BatchDownloaderNrJcCs||_||_dSrKrLrOrrrrQ�szBatchDownloader.__init__)�linksrRrccs�|D]�}zt|j|�}WnFty^}z.|jdus6J�t�d|jj|��WYd}~n
d}~00t||�}tj	�
||�}t|||j�}t
|d��$}	|D]}
|	�|
�q�Wd�n1s�0Y|j�dd�}|||ffVqdS)z0Download the files given by links into location.NrSrTrUr>rV)rPrhrRrrr]r0r^r.r_r`rarrrrb�s&�
*zBatchDownloader.__call__)
rcrdrerrfrQrr
rrbrrrrrg�s�	�rg).�__doc__r7r(r?r1�typingrrrZpip._vendor.requests.modelsrr�pip._internal.cli.progress_barsr�pip._internal.exceptionsr�pip._internal.models.indexr	�pip._internal.models.linkr
�pip._internal.network.cacher�pip._internal.network.sessionr�pip._internal.network.utilsr
rr�pip._internal.utils.miscrrr�	getLoggerrcr%rrrf�bytesr/r4r<rBrHrIrgrrrr�<module>s6
�)site-packages/pip/_internal/network/__pycache__/lazy_wheel.cpython-39.pyc000064400000020343150564745010022535 0ustar00a

��Re��@s�dZddgZddlmZmZddlmZddlmZddl	m
Z
mZmZm
Z
mZmZddlmZmZdd	lmZdd
lmZmZddlmZmZmZddlmZdd
lmZm Z m!Z!Gdd�de"�Z#e$e$eed�dd�Z%Gdd�d�Z&dS)zLazy ZIP over HTTP�HTTPRangeRequestUnsupported�dist_from_wheel_url�)�bisect_left�bisect_right)�contextmanager)�NamedTemporaryFile)�Any�Dict�Iterator�List�Optional�Tuple)�
BadZipfile�ZipFile)�canonicalize_name)�CONTENT_CHUNK_SIZE�Response)�BaseDistribution�MemoryWheel�get_wheel_distribution)�
PipSession)�HEADERS�raise_for_status�response_chunksc@seZdZdS)rN)�__name__�
__module__�__qualname__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/lazy_wheel.pyrs)�name�url�session�returncCsHt||��*}t|j|�}t|t|��Wd�S1s:0YdS)aReturn a distribution object from the given wheel URL.

    This uses HTTP range requests to only fetch the potion of the wheel
    containing metadata, just enough for the object to be constructed.
    If such requests are not supported, HTTPRangeRequestUnsupported
    is raised.
    N)�LazyZipOverHTTPrrrr)rr r!�zf�wheelrrrrsc@szeZdZdZefeeedd�dd�Ze	ed�dd��Z
e	ed�d	d
��Zed�dd�Z
dd�d
d�Ze	ed�dd��Zd3eed�dd�Zed�dd�Zd4eeed�dd�Zed�dd�Zd5eeed�dd�Zed�dd �Zdd�d!d"�Zeeed#�d$d%�Zeedd�d&d'��Zdd�d(d)�Zefeee eefe!d*�d+d,�Z"eeeeee#eefd-�d.d/�Z$eedd0�d1d2�Z%dS)6r#aFile-like object mapped to a ZIP file over HTTP.

    This uses HTTP range requests to lazily fetch the file's content,
    which is supposed to be fed to ZipFile.  If such requests are not
    supported by the server, raise HTTPRangeRequestUnsupported
    during initialization.
    N)r r!�
chunk_sizer"cCs�|j|td�}t|�|jdks$J�||||_|_|_t|jd�|_	t
�|_|�|j	�g|_
g|_d|j�dd�vr�td��|��dS)N)�headers��zContent-Length�bytesz
Accept-Ranges�nonezrange request is not supported)�headrr�status_code�_session�_url�_chunk_size�intr'�_lengthr�_file�truncate�_left�_right�getr�
_check_zip)�selfr r!r&r+rrr�__init__1szLazyZipOverHTTP.__init__)r"cCsdS)z!Opening mode, which is always rb.�rbr�r8rrr�modeAszLazyZipOverHTTP.modecCs|jjS)zPath to the underlying file.)r2rr;rrrrFszLazyZipOverHTTP.namecCsdS)z9Return whether random access is supported, which is True.Trr;rrr�seekableKszLazyZipOverHTTP.seekablecCs|j��dS)zClose the file.N)r2�closer;rrrr>OszLazyZipOverHTTP.closecCs|jjS)zWhether the file is closed.)r2�closedr;rrrr?SszLazyZipOverHTTP.closed���)�sizer"cCs`t||j�}|��|j}}|dkr(|nt|||�}td||�}|�||d�|j�|�S)z�Read up to size bytes from the object and return them.

        As a convenience, if size is unspecified or -1,
        all bytes until EOF are returned.  Fewer than
        size bytes may be returned if EOF is reached.
        r�)�maxr/�tellr1�min�	_downloadr2�read)r8rA�
download_size�start�length�stoprrrrGXszLazyZipOverHTTP.readcCsdS)z3Return whether the file is readable, which is True.Trr;rrr�readablefszLazyZipOverHTTP.readabler)�offset�whencer"cCs|j�||�S)a-Change stream position and return the new absolute position.

        Seek to offset relative position indicated by whence:
        * 0: Start of stream (the default).  pos should be >= 0;
        * 1: Current position - pos may be negative;
        * 2: End of stream - pos usually negative.
        )r2�seek)r8rMrNrrrrOjszLazyZipOverHTTP.seekcCs
|j��S)zReturn the current position.)r2rDr;rrrrDtszLazyZipOverHTTP.tellcCs|j�|�S)z�Resize the stream to the given size in bytes.

        If size is unspecified resize to the current position.
        The current stream position isn't changed.

        Return the new file size.
        )r2r3)r8rArrrr3xszLazyZipOverHTTP.truncatecCsdS)z
Return False.Frr;rrr�writable�szLazyZipOverHTTP.writablecCs|j��|S�N)r2�	__enter__r;rrrrR�s
zLazyZipOverHTTP.__enter__)�excr"cGs|jj|�SrQ)r2�__exit__)r8rSrrrrT�szLazyZipOverHTTP.__exit__c	cs.|��}zdVW|�|�n|�|�0dS)zyReturn a context manager keeping the position.

        At the end of the block, seek back to original position.
        N)rDrO)r8�posrrr�_stay�szLazyZipOverHTTP._stayc
Cs�|jd}ttd||j��D]h}|�||�|���@zt|�WntyTYn0Wd�q�Wd�q1sz0YqdS)z1Check and download until the file is a valid ZIP.rBrN)r1�reversed�ranger/rFrVrr)r8�endrIrrrr7�s

zLazyZipOverHTTP._check_zip)rIrY�base_headersr"cCs8|��}d|�d|��|d<d|d<|jj|j|dd�S)z:Return HTTP response to a range request from start to end.zbytes=�-�Rangezno-cachez
Cache-ControlT)r'�stream)�copyr-r6r.)r8rIrYrZr'rrr�_stream_response�sz LazyZipOverHTTP._stream_response)rIrY�left�rightr"c
cs�|j||�|j||�}}t|g|dd��}}t|g|dd��}t||�D]&\}}	||krv||dfV|	d}qX||kr�||fV|g|g|j||�<|j||�<dS)a/Return an iterator of intervals to be fetched.

        Args:
            start (int): Start of needed interval
            end (int): End of needed interval
            left (int): Index of first overlapping downloaded data
            right (int): Index after last overlapping downloaded data
        NrBr@)r4r5rErC�zip)
r8rIrYr`ra�lslice�rslice�i�j�krrr�_merge�s

zLazyZipOverHTTP._merge)rIrYr"cCs�|���~t|j|�}t|j|�}|�||||�D]D\}}|�||�}|��|�|�t	||j
�D]}|j�|�qdq2Wd�n1s�0YdS)z-Download bytes from start to end inclusively.N)
rVrr5rr4rhr_rrOrr/r2�write)r8rIrYr`ra�response�chunkrrrrF�s

zLazyZipOverHTTP._download)r@)r)N)&rrr�__doc__r�strrr0r9�propertyr<r�boolr=r>r?r)rGrLrOrDrr3rPrRrrTrr
rVr7rr	rr_r
rhrFrrrrr#(sD	��

���r#N)'rl�__all__�bisectrr�
contextlibr�tempfiler�typingrr	r
rrr
�zipfilerr�pip._vendor.packaging.utilsrZpip._vendor.requests.modelsrr�pip._internal.metadatarrr�pip._internal.network.sessionr�pip._internal.network.utilsrrr�	Exceptionrrmrr#rrrr�<module>s site-packages/pip/_internal/network/__pycache__/xmlrpc.cpython-39.pyc000064400000004046150564745010021701 0ustar00a

��Re��@s�dZddlZddlZddlZddlmZmZddl	m
Z
ddlmZddl
mZerdddlmZmZe�e�ZGdd	�d	ejj�ZdS)
z#xmlrpclib.Transport implementation
�N)�
TYPE_CHECKING�Tuple)�NetworkConnectionError)�
PipSession)�raise_for_status)�	_HostType�
_MarshallablecsJeZdZdZdeeedd��fdd�
Zd
deeee	dd	�d
d�Z
�ZS)�PipXmlrpcTransportzRProvide a `xmlrpclib.Transport` implementation via a `PipSession`
    object.
    FN)�	index_url�session�use_datetime�returncs*t��|�tj�|�}|j|_||_dS)N)�super�__init__�urllib�parse�urlparse�scheme�_scheme�_session)�selfr
rrZindex_parts��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/xmlrpc.pyrszPipXmlrpcTransport.__init__r)r.)�host�handler�request_body�verboser
c

Cs�t|t�sJ�|j||dddf}tj�|�}z8ddi}|jj|||dd�}t|�||_	|�
|j�WSty�}	z*|	j
s~J�t�d|	j
j|��WYd}	~	n
d}	~	00dS)NzContent-Typeztext/xmlT)�data�headers�streamzHTTP error %s while getting %s)�
isinstance�strrrr�
urlunparser�postrr�parse_response�rawr�response�logger�critical�status_code)
rrrrr�parts�urlr r(�excrrr�request s,�
�zPipXmlrpcTransport.request)F)F)�__name__�
__module__�__qualname__�__doc__r#r�boolr�bytesrr/�
__classcell__rrrrr	s��
��r	)r3�logging�urllib.parser�
xmlrpc.clientZxmlrpc�typingrr�pip._internal.exceptionsr�pip._internal.network.sessionr�pip._internal.network.utilsrrr�	getLoggerr0r)�client�	Transportr	rrrr�<module>s
site-packages/pip/_internal/network/__pycache__/auth.cpython-39.pyc000064400000016537150564745010021345 0ustar00a

��Re�/�
@s*dZddlZddlmZmZmZmZmZddl	m
Z
mZddlm
Z
mZddlmZddlmZddlmZmZmZmZmZdd	lmZee�ZeeeefZzddlaWnLe y�daYn8e!y�Z"z e�#d
ee"��daWYdZ"["n
dZ"["00eeeeeed�dd
�Z$Gdd�de
�Z%dS)z�Network Authentication Helpers

Contains interface (MultiDomainBasicAuth) and associated glue code for
providing credentials in the context of network requests.
�N)�Any�Dict�List�Optional�Tuple)�AuthBase�
HTTPBasicAuth)�Request�Response)�get_netrc_auth)�	getLogger)�ask�	ask_input�ask_password�remove_auth_from_url�split_auth_netloc_from_url)�AuthInfo�*Keyring is skipped due to an exception: %s)�url�username�returnc
Cs�|rtsdSz~z
tj}Wnty*Yn40t�d|�|||�}|durX|j|jfWSWdS|r�t�d|�t�||�}|r�||fWSWn8ty�}z t�	dt
|��daWYd}~n
d}~00dS)z3Return the tuple auth for a given url from keyring.Nz'Getting credentials from keyring for %sz$Getting password from keyring for %sr)�keyring�get_credential�AttributeError�logger�debugr�password�get_password�	Exception�warning�str)rrr�credr�exc�r#��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/auth.py�get_keyring_auth(s0

�r%c@s�eZdZd eeeedd�dd�Zeeed�dd�Zd!eeee	d
�dd�Z
eeeeeeefd
�dd�Ze
e
d�dd�Zeeeeeeefd�dd�Zed�dd�Zeeed�dd�Zeedd�dd�Zeedd�dd�ZdS)"�MultiDomainBasicAuthTN)�	prompting�
index_urlsrcCs||_||_i|_d|_dS)N)r'r(�	passwords�_credentials_to_save)�selfr'r(r#r#r$�__init__JszMultiDomainBasicAuth.__init__)rrcCsB|r
|jsdS|jD](}t|��d�d}|�|�r|SqdS)a�Return the original index URL matching the requested URL.

        Cached or dynamically generated credentials may work against
        the original index URL rather than just the netloc.

        The provided url should have had its username and password
        removed already. If the original index url had credentials then
        they will be included in the return value.

        Returns None if no matching index was found, or if --no-index
        was specified by the user.
        N�/)r(r�rstrip�
startswith)r+r�u�prefixr#r#r$�_get_index_urlWs



z#MultiDomainBasicAuth._get_index_urlF)�original_url�allow_netrc�
allow_keyringrcCs�t|�\}}}|\}}|dur6|dur6t�d|�|S|�|�}	|	rft|	�}
|
rf|
\}	}}t�d|	�|	r�|ddur�|\}}|dur�|dur�t�d|�|S|r�t|�}
|
r�t�d|�|
S|r�t|	|�p�t||�}|r�t�d|�|S||fS)z2Find and return credentials for the specified URL.NzFound credentials in url for %szFound index url %srz%Found credentials in index url for %sz!Found credentials in netrc for %sz#Found credentials in keyring for %s)rrrr2rr%)r+r3r4r5r�netloc�url_user_passwordrr�	index_url�
index_info�_�index_url_user_password�
netrc_auth�kr_authr#r#r$�_get_new_credentialsms>�



�z)MultiDomainBasicAuth._get_new_credentials)r3rc	Cs�t|�\}}}|�|�\}}|dus,|dur^||jvr^|j|\}}|dusT||kr^||}}|dusn|dur�|ptd}|p|d}||f|j|<|dur�|dus�|dur�|dus�Jd|����|||fS)a_Return the credentials to use for the provided URL.

        If allowed, netrc and keyring may be used to obtain the
        correct credentials.

        Returns (url_without_credentials, username, password). Note
        that even if the original URL contains credentials, this
        function may return a different username and password.
        N�z%Could not load credentials from url: )rr>r))	r+r3rr6r:rr�un�pwr#r#r$�_get_url_and_credentials�s*
�����z-MultiDomainBasicAuth._get_url_and_credentials)�reqrcCsH|�|j�\}}}||_|dur6|dur6t||�|�}|�d|j�|S)N�response)rBrr�
register_hook�
handle_401)r+rCrrrr#r#r$�__call__�szMultiDomainBasicAuth.__call__)r6rcCsbtd|�d��}|sdSt||�}|rP|ddurP|ddurP|d|ddfStd�}||dfS)	Nz	User for z: )NNFr�Fz
Password: T)rr%r)r+r6r�authrr#r#r$�_prompt_for_password�s
z)MultiDomainBasicAuth._prompt_for_password)rcCstsdStdddg�dkS)NFz#Save credentials to keyring [y/N]: �y�n)rr
)r+r#r#r$� _should_save_password_to_keyring�sz5MultiDomainBasicAuth._should_save_password_to_keyring)�resp�kwargsrc	Ks|jdkr|S|js|Stj�|j�}|j|jddd�\}}d}|sZ|sZ|�|j�\}}}d|_	|dur�|dur�||f|j
|j<|r�|��r�|j||f|_	|j|j
��t|p�d|p�d�|j�}|�d|j�|j	r�|�d|j�|jj|fi|��}|j�|�|S)N�FT)r4r5r?rD)�status_coder'�urllib�parse�urlparserr>rJr6r*r)rM�content�raw�release_connr�requestrE�warn_on_401�save_credentials�
connection�send�history�append)	r+rNrO�parsedrr�saverC�new_respr#r#r$rF�s6
�

zMultiDomainBasicAuth.handle_401cKs|jdkrt�d|jj�dS)z6Response callback to warn about incorrect credentials.rPz)401 Error, Credentials not correct for %sN)rQrrrXr)r+rNrOr#r#r$rY.s

�z MultiDomainBasicAuth.warn_on_401cKsltdusJd��tsdS|j}d|_|rh|jdkrhzt�d�tj|�Wntyft�d�Yn0dS)z1Response callback to save credentials on success.Nz'should never reach here without keyringi�zSaving credentials to keyringzFailed to save credentials)rr*rQr�info�set_passwordr�	exception)r+rNrO�credsr#r#r$rZ6s
z%MultiDomainBasicAuth.save_credentials)TN)TF)�__name__�
__module__�__qualname__�boolrrr r,r2rr>rrBr	rGrJrMr
rrFrYrZr#r#r#r$r&Is2��
��9�0�
6r&)&�__doc__�urllib.parserR�typingrrrrrZpip._vendor.requests.authrrZpip._vendor.requests.modelsr	r
Zpip._vendor.requests.utilsr�pip._internal.utils.loggingr�pip._internal.utils.miscr
rrrr� pip._internal.vcs.versioncontrolrrfrr �Credentialsr�ImportErrorrr"rr%r&r#r#r#r$�<module>s,�!site-packages/pip/_internal/network/__pycache__/__init__.cpython-39.pyc000064400000000437150564745010022133 0ustar00a

��Re2�@sdZdS)z+Contains purely network-related utilities.
N)�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/__init__.py�<module>�site-packages/pip/_internal/network/__pycache__/session.cpython-39.pyc000064400000025045150564745010022061 0ustar00a

��ReYA�@sUdZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlZddl
ZddlZddlmZmZmZmZmZmZmZmZmZddlmZmZddlmZddlm Z m!Z!ddl"m#Z#m$Z$ddl%m&Z&dd	l'm(Z(dd
l)m*Z*ddl+m,Z,ddl-m.Z.dd
l/m0Z0ddl1m2Z2ddl3m4Z4ddl5m6Z6ddl7m8Z8ddl9m:Z:m;Z;ddl<m=Z=e�>e?�Z@eeAeAeeeBeAffZCejDde*d�gd�ZEeeCeFd<dZGeHd�dd�ZIeAd�dd�ZJGdd�de �ZKGd d!�d!e!�ZLGd"d#�d#e�ZMGd$d%�d%ejN�ZOdS)&zhPipSession and supporting code, containing all pip-specific
network request configuration and behavior.
�N)	�Any�Dict�Iterator�List�Mapping�Optional�Sequence�Tuple�Union)�requests�urllib3)�CacheControlAdapter)�BaseAdapter�HTTPAdapter)�PreparedRequest�Response)�CaseInsensitiveDict)�ConnectionPool)�InsecureRequestWarning)�__version__)�get_default_environment)�Link)�MultiDomainBasicAuth)�
SafeFileCache)�has_tls)�libc_ver)�build_url_from_netloc�parse_netloc)�url_to_path�ignore)�category))�https�*r")r"�	localhostr")r"z127.0.0.0/8r")r"z::1/128r")�filer"N)�sshr"r"�SECURE_ORIGINS)�
BUILD_BUILDID�BUILD_ID�CI�	PIP_IS_CI��returncCstdd�tD��S)z?
    Return whether it looks like pip is running under CI.
    css|]}|tjvVqdS�N)�os�environ)�.0�name�r2��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/session.py�	<genexpr>X�z looks_like_ci.<locals>.<genexpr>)�any�CI_ENVIRONMENT_VARIABLESr2r2r2r3�
looks_like_ciQsr8c
Cs�dtd�t��dt��id�}|dddkr@t��|dd<n�|dddkr�tj}|jd	krl|d
d�}d�d
d�|D��|dd<nB|dddkr�t��|dd<n |dddkr�t��|dd<tj�d��rRddl	m
}|��|��|�
�f}ttdd�tgd�|���}ttdd�tddgt����}|�rD||d<|�rR||d<tj�d��r�t��d�r�dt��dd�|d<t���r�t��|�di�d<t���r�t��|�di�d<t���r�t��|d<t��r�dd
l}|j|d <t��d!�}|d
u�rt|j�|d"<t�d#�d
u�rvztj d#d$gtj!d%d&�}Wnt"�yRYn$0|�d'��rv|�#d(�d)�$�|d*<t%��r�d+nd
|d,<t&j'�(d-�}	|	d
u�r�|	|d.<d/j)|t*j+|d0d+d1�d2�S)3z6
    Return a string representing the user agent.
    �pip)r1�versionr1)�	installer�python�implementationr=�CPythonr:�PyPy�finalN��.cSsg|]}t|��qSr2)�str)r0�xr2r2r3�
<listcomp>nr5zuser_agent.<locals>.<listcomp>�Jython�
IronPython�linuxr)�distrocSs|dS�N�r2�rDr2r2r3�<lambda>}r5zuser_agent.<locals>.<lambda>)r1r:�idcSs|dSrJr2rLr2r2r3rM�r5�lib�libcrI�darwin�macOS�system�release�cpu�openssl_version�
setuptools�setuptools_version�rustcz	--versiong�?)�stderr�timeoutsrustc � rK�
rustc_versionT�ci�PIP_USER_AGENT_USER_DATA�	user_dataz9{data[installer][name]}/{data[installer][version]} {json})�,�:)�
separators�	sort_keys)�data�json),r�platform�python_version�python_implementation�sys�pypy_version_info�releaselevel�join�
startswith�pip._vendorrIr1r:�codename�dict�filter�zipr�mac_verrS�
setdefaultrT�machiner�_ssl�OPENSSL_VERSIONr�get_distributionrC�shutil�which�
subprocess�check_output�STDOUT�	Exception�split�decoder8r.r/�get�formatrf�dumps)
rerkrI�linux_distribution�distro_infosrP�ssl�setuptools_dist�rustc_outputr`r2r2r3�
user_agent[s���
�����




�

�r�c@sreZdZd
eeeeeeeeffeee	feee	ee	e	ffee
e	e	fed�dd�Zdd�dd	�Z
dS)�LocalFSAdapterFNT)�request�streamr[�verify�cert�proxiesr,c

Cs�t|j�}t�}d|_|j|_zt�|�}	WnTty�}
z<d|_t|
�j|_	t
�|j	�d|
���d��|_
WYd}
~
nXd}
~
00tjj|	jdd�}t�|�dp�d}t||	j|d	��|_t|d
�|_
|j
j|_|S)N��i�z: �utf8T)�usegmtrz
text/plain)zContent-TypezContent-Lengthz
Last-Modified�rb)r�urlr�status_coder.�stat�OSError�type�__name__�reason�io�BytesIO�encode�raw�email�utils�
formatdate�st_mtime�	mimetypes�
guess_typer�st_size�headers�open�close)
�selfr�r�r[r�r�r��pathname�resp�stats�exc�modified�content_typer2r2r3�send�s*	
4��
zLocalFSAdapter.sendr+cCsdSr-r2)r�r2r2r3r��szLocalFSAdapter.close)FNTNN)r��
__module__�__qualname__r�boolrr
�floatr	rCrrr�r�r2r2r2r3r��s�
�'r�c	sFeZdZeeeeefeeeeeeffdd��fdd�Z	�Z
S)�InsecureHTTPAdapterN��connr�r�r�r,cst�j||d|d�dS�NF)r�r�r�r���super�cert_verify�r�r�r�r�r���	__class__r2r3r��szInsecureHTTPAdapter.cert_verify�r�r�r�rrCr
r�rr	r��
__classcell__r2r2r�r3r��s
�r�c	sFeZdZeeeeefeeeeeeffdd��fdd�Z	�Z
S)�InsecureCacheControlAdapterNr�cst�j||d|d�dSr�r�r�r�r2r3r��sz'InsecureCacheControlAdapter.cert_verifyr�r2r2r�r3r��s
�r�c	s�eZdZUdZeeed<ddddd�eeeee	eee
eedd��fdd�Ze
edd	�d
d�Zdeeee
dd
�dd�Zeed�dd�Zee
d�dd�Zeeeeed��fdd�Z�ZS)�
PipSessionNr[rr2)�retries�cache�
trusted_hosts�
index_urls)�argsr�r�r�r��kwargsr,c
s�t�j|i|��g|_t�|jd<t|d�|_tj|gd�dd�}t	|d�}|rvt
t|�|d�}tt|�|d�|_
nt|d�}||_
|�d|�|�d	|�|�d
t��|D]}	|j|	dd�q�d
S)zj
        :param trusted_hosts: Domains not to emit warnings for when not using
            HTTPS.
        z
User-Agent)r�)i�i�iig�?)�total�status_forcelist�backoff_factor)�max_retries)r�r�zhttps://zhttp://zfile://T)�suppress_loggingN)r��__init__�pip_trusted_originsr�r�r�authr�Retryr�r
rr��_trusted_host_adapterr�mountr��add_trusted_host)
r�r�r�r�r�r�r��insecure_adapter�secure_adapter�hostr�r2r3r�s4
�
��

zPipSession.__init__)�new_index_urlsr,cCs||j_dS)zn
        :param new_index_urls: New index urls to update the authentication
            handler with.
        N)r�r�)r�r�r2r2r3�update_index_urlsSszPipSession.update_index_urlsF)r��sourcer�r,cCs�|s0d|��}|dur&|d|�d�7}t�|�t|�}||jvrN|j�|�|�t|dd�d|j�|�t|�d|j�|ds�|�t|dd�d	|j�|�t|�d	|j�dS)
z�
        :param host: It is okay to provide a host that has previously been
            added.
        :param source: An optional source string, for logging where the host
            string came from.
        zadding trusted host: Nz (from �)�http)�scheme�/rKrb)�logger�inforr��appendr�rr�)r�r�r�r��msg�	host_portr2r2r3r�Zs$	


��zPipSession.add_trusted_hostr+ccs6tEdH|jD] \}}d||dur(dn|fVqdS)Nr")r&r�)r�r��portr2r2r3�iter_secure_originsys
zPipSession.iter_secure_origins)�locationr,c	Cs�tj�t|��}|j|j|j}}}|�dd�d}|��D]�}|\}}}	||kr^|dkr^q>zt	�
|�}
t	�|�}Wn2ty�|r�|�
�|�
�kr�|dkr�Yq>Yn0|
|vr�q>||	kr�|	dkr�|	dur�q>dSt�d||�dS)N�+rK���r"Tz�The repository located at %s is not a trusted or secure host and is being ignored. If this repository is available via HTTPS we recommend you use HTTPS instead, otherwise you may silence this warning and allow it anyway with '--trusted-host %s'.F)�urllib�parse�urlparserCr��hostnamer��rsplitr��	ipaddress�
ip_address�
ip_network�
ValueError�lowerr��warning)r�r��parsed�origin_protocol�origin_host�origin_port�
secure_origin�secure_protocol�secure_host�secure_port�addr�networkr2r2r3�is_secure_origin~sJ�



���
����	zPipSession.is_secure_origin)�methodr�r�r�r,cs*|�d|j�t�j||g|�Ri|��S)Nr[)rur[r�r�)r�rr�r�r�r�r2r3r��szPipSession.request)NF)r�r�r�r[r�int�__annotations__rrCrrr�r�r�r�r�SecureOriginr�rr�rr�r�r2r2r�r3r�s.
�
�M�
�Cr�)P�__doc__�email.utilsr�r�r�rf�loggingr�r.rgrzr|rj�urllib.parser��warnings�typingrrrrrrrr	r
rorr�pip._vendor.cachecontrolr
Zpip._vendor.requests.adaptersrrZpip._vendor.requests.modelsrrZpip._vendor.requests.structuresrZ"pip._vendor.urllib3.connectionpoolrZpip._vendor.urllib3.exceptionsrr9r�pip._internal.metadatar�pip._internal.models.linkr�pip._internal.network.authr�pip._internal.network.cacher�pip._internal.utils.compatr�pip._internal.utils.glibcr�pip._internal.utils.miscrr�pip._internal.utils.urlsr�	getLoggerr�r�rCrr�filterwarningsr&rr7r�r8r�r�r�r��Sessionr�r2r2r2r3�<module>sR,

e,site-packages/pip/_internal/network/__pycache__/cache.cpython-39.pyc000064400000005561150564745010021442 0ustar00a

��Re4�@s�dZddlZddlmZddlmZmZddlmZddl	m
Z
ddlmZddl
mZmZdd	lmZeed
�dd�Zeedd
�dd��ZGdd�de�ZdS)zHTTP cache implementation.
�N)�contextmanager)�Iterator�Optional)�	BaseCache)�	FileCache)�Response)�adjacent_tmp_file�replace)�
ensure_dir)�response�returncCst|dd�S)N�
from_cacheF)�getattr)r�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/network/cache.py�
is_from_cachesr)rccs"z
dVWntyYn0dS)zvIf we can't access the cache then we can just skip caching and process
    requests as if caching wasn't enabled.
    N)�OSErrorrrrr�suppressed_cache_errorss
rcsneZdZdZedd��fdd�Zeed�dd�Zeeed	�d
d�Z	eedd�d
d�Z
edd	�dd�Z�ZS)�
SafeFileCachezw
    A file based cache which is safe to use even when the target directory may
    not be accessible or writable.
    N)�	directoryrcs$|dusJd��t���||_dS)Nz!Cache directory must not be None.)�super�__init__r)�selfr��	__class__rrr%s
zSafeFileCache.__init__)�namercCs6t�|�}t|dd��|g}tjj|jg|�R�S)N�)r�encode�list�os�path�joinr)rr�hashed�partsrrr�_get_cache_path*s
zSafeFileCache._get_cache_path)�keyrc	Cst|�|�}t��Pt|d��&}|��Wd�Wd�S1sH0YWd�n1sf0YdS)N�rb)r$r�open�read)rr%r �frrr�get2s
zSafeFileCache.get)r%�valuerc	Cs�|�|�}t��^ttj�|��t|��}|�|�Wd�n1sJ0Yt|j	|�Wd�n1st0YdS�N)
r$rr
rr �dirnamer�writer	r)rr%r+r r)rrr�set8s

(zSafeFileCache.setcCs>|�|�}t��t�|�Wd�n1s00YdSr,)r$rr�remove)rr%r rrr�deleteBs
zSafeFileCache.delete)
�__name__�
__module__�__qualname__�__doc__�strrr$r�bytesr*r/r1�
__classcell__rrrrrs
r)r5r�
contextlibr�typingrrZpip._vendor.cachecontrol.cacher�pip._vendor.cachecontrol.cachesrZpip._vendor.requests.modelsr�pip._internal.utils.filesystemrr	�pip._internal.utils.miscr
�boolrrrrrrr�<module>s
site-packages/pip/_internal/network/utils.py000064400000007733150564745010015253 0ustar00from typing import Dict, Iterator

from pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response

from pip._internal.exceptions import NetworkConnectionError

# The following comments and HTTP headers were originally added by
# Donald Stufft in git commit 22c562429a61bb77172039e480873fb239dd8c03.
#
# We use Accept-Encoding: identity here because requests defaults to
# accepting compressed responses. This breaks in a variety of ways
# depending on how the server is configured.
# - Some servers will notice that the file isn't a compressible file
#   and will leave the file alone and with an empty Content-Encoding
# - Some servers will notice that the file is already compressed and
#   will leave the file alone, adding a Content-Encoding: gzip header
# - Some servers won't notice anything at all and will take a file
#   that's already been compressed and compress it again, and set
#   the Content-Encoding: gzip header
# By setting this to request only the identity encoding we're hoping
# to eliminate the third case.  Hopefully there does not exist a server
# which when given a file will notice it is already compressed and that
# you're not asking for a compressed file and will then decompress it
# before sending because if that's the case I don't think it'll ever be
# possible to make this work.
HEADERS: Dict[str, str] = {"Accept-Encoding": "identity"}


def raise_for_status(resp: Response) -> None:
    http_error_msg = ""
    if isinstance(resp.reason, bytes):
        # We attempt to decode utf-8 first because some servers
        # choose to localize their reason strings. If the string
        # isn't utf-8, we fall back to iso-8859-1 for all other
        # encodings.
        try:
            reason = resp.reason.decode("utf-8")
        except UnicodeDecodeError:
            reason = resp.reason.decode("iso-8859-1")
    else:
        reason = resp.reason

    if 400 <= resp.status_code < 500:
        http_error_msg = (
            f"{resp.status_code} Client Error: {reason} for url: {resp.url}"
        )

    elif 500 <= resp.status_code < 600:
        http_error_msg = (
            f"{resp.status_code} Server Error: {reason} for url: {resp.url}"
        )

    if http_error_msg:
        raise NetworkConnectionError(http_error_msg, response=resp)


def response_chunks(
    response: Response, chunk_size: int = CONTENT_CHUNK_SIZE
) -> Iterator[bytes]:
    """Given a requests Response, provide the data chunks."""
    try:
        # Special case for urllib3.
        for chunk in response.raw.stream(
            chunk_size,
            # We use decode_content=False here because we don't
            # want urllib3 to mess with the raw bytes we get
            # from the server. If we decompress inside of
            # urllib3 then we cannot verify the checksum
            # because the checksum will be of the compressed
            # file. This breakage will only occur if the
            # server adds a Content-Encoding header, which
            # depends on how the server was configured:
            # - Some servers will notice that the file isn't a
            #   compressible file and will leave the file alone
            #   and with an empty Content-Encoding
            # - Some servers will notice that the file is
            #   already compressed and will leave the file
            #   alone and will add a Content-Encoding: gzip
            #   header
            # - Some servers won't notice anything at all and
            #   will take a file that's already been compressed
            #   and compress it again and set the
            #   Content-Encoding: gzip header
            #
            # By setting this not to decode automatically we
            # hope to eliminate problems with the second case.
            decode_content=False,
        ):
            yield chunk
    except AttributeError:
        # Standard file-like object.
        while True:
            chunk = response.raw.read(chunk_size)
            if not chunk:
                break
            yield chunk
site-packages/pip/_internal/network/session.py000064400000040531150564745010015567 0ustar00"""PipSession and supporting code, containing all pip-specific
network request configuration and behavior.
"""

import email.utils
import io
import ipaddress
import json
import logging
import mimetypes
import os
import platform
import shutil
import subprocess
import sys
import urllib.parse
import warnings
from typing import Any, Dict, Iterator, List, Mapping, Optional, Sequence, Tuple, Union

from pip._vendor import requests, urllib3
from pip._vendor.cachecontrol import CacheControlAdapter
from pip._vendor.requests.adapters import BaseAdapter, HTTPAdapter
from pip._vendor.requests.models import PreparedRequest, Response
from pip._vendor.requests.structures import CaseInsensitiveDict
from pip._vendor.urllib3.connectionpool import ConnectionPool
from pip._vendor.urllib3.exceptions import InsecureRequestWarning

from pip import __version__
from pip._internal.metadata import get_default_environment
from pip._internal.models.link import Link
from pip._internal.network.auth import MultiDomainBasicAuth
from pip._internal.network.cache import SafeFileCache

# Import ssl from compat so the initial import occurs in only one place.
from pip._internal.utils.compat import has_tls
from pip._internal.utils.glibc import libc_ver
from pip._internal.utils.misc import build_url_from_netloc, parse_netloc
from pip._internal.utils.urls import url_to_path

logger = logging.getLogger(__name__)

SecureOrigin = Tuple[str, str, Optional[Union[int, str]]]


# Ignore warning raised when using --trusted-host.
warnings.filterwarnings("ignore", category=InsecureRequestWarning)


SECURE_ORIGINS: List[SecureOrigin] = [
    # protocol, hostname, port
    # Taken from Chrome's list of secure origins (See: http://bit.ly/1qrySKC)
    ("https", "*", "*"),
    ("*", "localhost", "*"),
    ("*", "127.0.0.0/8", "*"),
    ("*", "::1/128", "*"),
    ("file", "*", None),
    # ssh is always secure.
    ("ssh", "*", "*"),
]


# These are environment variables present when running under various
# CI systems.  For each variable, some CI systems that use the variable
# are indicated.  The collection was chosen so that for each of a number
# of popular systems, at least one of the environment variables is used.
# This list is used to provide some indication of and lower bound for
# CI traffic to PyPI.  Thus, it is okay if the list is not comprehensive.
# For more background, see: https://github.com/pypa/pip/issues/5499
CI_ENVIRONMENT_VARIABLES = (
    # Azure Pipelines
    "BUILD_BUILDID",
    # Jenkins
    "BUILD_ID",
    # AppVeyor, CircleCI, Codeship, Gitlab CI, Shippable, Travis CI
    "CI",
    # Explicit environment variable.
    "PIP_IS_CI",
)


def looks_like_ci() -> bool:
    """
    Return whether it looks like pip is running under CI.
    """
    # We don't use the method of checking for a tty (e.g. using isatty())
    # because some CI systems mimic a tty (e.g. Travis CI).  Thus that
    # method doesn't provide definitive information in either direction.
    return any(name in os.environ for name in CI_ENVIRONMENT_VARIABLES)


def user_agent() -> str:
    """
    Return a string representing the user agent.
    """
    data: Dict[str, Any] = {
        "installer": {"name": "pip", "version": __version__},
        "python": platform.python_version(),
        "implementation": {
            "name": platform.python_implementation(),
        },
    }

    if data["implementation"]["name"] == "CPython":
        data["implementation"]["version"] = platform.python_version()
    elif data["implementation"]["name"] == "PyPy":
        pypy_version_info = sys.pypy_version_info  # type: ignore
        if pypy_version_info.releaselevel == "final":
            pypy_version_info = pypy_version_info[:3]
        data["implementation"]["version"] = ".".join(
            [str(x) for x in pypy_version_info]
        )
    elif data["implementation"]["name"] == "Jython":
        # Complete Guess
        data["implementation"]["version"] = platform.python_version()
    elif data["implementation"]["name"] == "IronPython":
        # Complete Guess
        data["implementation"]["version"] = platform.python_version()

    if sys.platform.startswith("linux"):
        from pip._vendor import distro

        linux_distribution = distro.name(), distro.version(), distro.codename()
        distro_infos: Dict[str, Any] = dict(
            filter(
                lambda x: x[1],
                zip(["name", "version", "id"], linux_distribution),
            )
        )
        libc = dict(
            filter(
                lambda x: x[1],
                zip(["lib", "version"], libc_ver()),
            )
        )
        if libc:
            distro_infos["libc"] = libc
        if distro_infos:
            data["distro"] = distro_infos

    if sys.platform.startswith("darwin") and platform.mac_ver()[0]:
        data["distro"] = {"name": "macOS", "version": platform.mac_ver()[0]}

    if platform.system():
        data.setdefault("system", {})["name"] = platform.system()

    if platform.release():
        data.setdefault("system", {})["release"] = platform.release()

    if platform.machine():
        data["cpu"] = platform.machine()

    if has_tls():
        import _ssl as ssl

        data["openssl_version"] = ssl.OPENSSL_VERSION

    setuptools_dist = get_default_environment().get_distribution("setuptools")
    if setuptools_dist is not None:
        data["setuptools_version"] = str(setuptools_dist.version)

    if shutil.which("rustc") is not None:
        # If for any reason `rustc --version` fails, silently ignore it
        try:
            rustc_output = subprocess.check_output(
                ["rustc", "--version"], stderr=subprocess.STDOUT, timeout=0.5
            )
        except Exception:
            pass
        else:
            if rustc_output.startswith(b"rustc "):
                # The format of `rustc --version` is:
                # `b'rustc 1.52.1 (9bc8c42bb 2021-05-09)\n'`
                # We extract just the middle (1.52.1) part
                data["rustc_version"] = rustc_output.split(b" ")[1].decode()

    # Use None rather than False so as not to give the impression that
    # pip knows it is not being run under CI.  Rather, it is a null or
    # inconclusive result.  Also, we include some value rather than no
    # value to make it easier to know that the check has been run.
    data["ci"] = True if looks_like_ci() else None

    user_data = os.environ.get("PIP_USER_AGENT_USER_DATA")
    if user_data is not None:
        data["user_data"] = user_data

    return "{data[installer][name]}/{data[installer][version]} {json}".format(
        data=data,
        json=json.dumps(data, separators=(",", ":"), sort_keys=True),
    )


class LocalFSAdapter(BaseAdapter):
    def send(
        self,
        request: PreparedRequest,
        stream: bool = False,
        timeout: Optional[Union[float, Tuple[float, float]]] = None,
        verify: Union[bool, str] = True,
        cert: Optional[Union[str, Tuple[str, str]]] = None,
        proxies: Optional[Mapping[str, str]] = None,
    ) -> Response:
        pathname = url_to_path(request.url)

        resp = Response()
        resp.status_code = 200
        resp.url = request.url

        try:
            stats = os.stat(pathname)
        except OSError as exc:
            # format the exception raised as a io.BytesIO object,
            # to return a better error message:
            resp.status_code = 404
            resp.reason = type(exc).__name__
            resp.raw = io.BytesIO(f"{resp.reason}: {exc}".encode("utf8"))
        else:
            modified = email.utils.formatdate(stats.st_mtime, usegmt=True)
            content_type = mimetypes.guess_type(pathname)[0] or "text/plain"
            resp.headers = CaseInsensitiveDict(
                {
                    "Content-Type": content_type,
                    "Content-Length": stats.st_size,
                    "Last-Modified": modified,
                }
            )

            resp.raw = open(pathname, "rb")
            resp.close = resp.raw.close

        return resp

    def close(self) -> None:
        pass


class InsecureHTTPAdapter(HTTPAdapter):
    def cert_verify(
        self,
        conn: ConnectionPool,
        url: str,
        verify: Union[bool, str],
        cert: Optional[Union[str, Tuple[str, str]]],
    ) -> None:
        super().cert_verify(conn=conn, url=url, verify=False, cert=cert)


class InsecureCacheControlAdapter(CacheControlAdapter):
    def cert_verify(
        self,
        conn: ConnectionPool,
        url: str,
        verify: Union[bool, str],
        cert: Optional[Union[str, Tuple[str, str]]],
    ) -> None:
        super().cert_verify(conn=conn, url=url, verify=False, cert=cert)


class PipSession(requests.Session):

    timeout: Optional[int] = None

    def __init__(
        self,
        *args: Any,
        retries: int = 0,
        cache: Optional[str] = None,
        trusted_hosts: Sequence[str] = (),
        index_urls: Optional[List[str]] = None,
        **kwargs: Any,
    ) -> None:
        """
        :param trusted_hosts: Domains not to emit warnings for when not using
            HTTPS.
        """
        super().__init__(*args, **kwargs)

        # Namespace the attribute with "pip_" just in case to prevent
        # possible conflicts with the base class.
        self.pip_trusted_origins: List[Tuple[str, Optional[int]]] = []

        # Attach our User Agent to the request
        self.headers["User-Agent"] = user_agent()

        # Attach our Authentication handler to the session
        self.auth = MultiDomainBasicAuth(index_urls=index_urls)

        # Create our urllib3.Retry instance which will allow us to customize
        # how we handle retries.
        retries = urllib3.Retry(
            # Set the total number of retries that a particular request can
            # have.
            total=retries,
            # A 503 error from PyPI typically means that the Fastly -> Origin
            # connection got interrupted in some way. A 503 error in general
            # is typically considered a transient error so we'll go ahead and
            # retry it.
            # A 500 may indicate transient error in Amazon S3
            # A 520 or 527 - may indicate transient error in CloudFlare
            status_forcelist=[500, 503, 520, 527],
            # Add a small amount of back off between failed requests in
            # order to prevent hammering the service.
            backoff_factor=0.25,
        )  # type: ignore

        # Our Insecure HTTPAdapter disables HTTPS validation. It does not
        # support caching so we'll use it for all http:// URLs.
        # If caching is disabled, we will also use it for
        # https:// hosts that we've marked as ignoring
        # TLS errors for (trusted-hosts).
        insecure_adapter = InsecureHTTPAdapter(max_retries=retries)

        # We want to _only_ cache responses on securely fetched origins or when
        # the host is specified as trusted. We do this because
        # we can't validate the response of an insecurely/untrusted fetched
        # origin, and we don't want someone to be able to poison the cache and
        # require manual eviction from the cache to fix it.
        if cache:
            secure_adapter = CacheControlAdapter(
                cache=SafeFileCache(cache),
                max_retries=retries,
            )
            self._trusted_host_adapter = InsecureCacheControlAdapter(
                cache=SafeFileCache(cache),
                max_retries=retries,
            )
        else:
            secure_adapter = HTTPAdapter(max_retries=retries)
            self._trusted_host_adapter = insecure_adapter

        self.mount("https://", secure_adapter)
        self.mount("http://", insecure_adapter)

        # Enable file:// urls
        self.mount("file://", LocalFSAdapter())

        for host in trusted_hosts:
            self.add_trusted_host(host, suppress_logging=True)

    def update_index_urls(self, new_index_urls: List[str]) -> None:
        """
        :param new_index_urls: New index urls to update the authentication
            handler with.
        """
        self.auth.index_urls = new_index_urls

    def add_trusted_host(
        self, host: str, source: Optional[str] = None, suppress_logging: bool = False
    ) -> None:
        """
        :param host: It is okay to provide a host that has previously been
            added.
        :param source: An optional source string, for logging where the host
            string came from.
        """
        if not suppress_logging:
            msg = f"adding trusted host: {host!r}"
            if source is not None:
                msg += f" (from {source})"
            logger.info(msg)

        host_port = parse_netloc(host)
        if host_port not in self.pip_trusted_origins:
            self.pip_trusted_origins.append(host_port)

        self.mount(
            build_url_from_netloc(host, scheme="http") + "/", self._trusted_host_adapter
        )
        self.mount(build_url_from_netloc(host) + "/", self._trusted_host_adapter)
        if not host_port[1]:
            self.mount(
                build_url_from_netloc(host, scheme="http") + ":",
                self._trusted_host_adapter,
            )
            # Mount wildcard ports for the same host.
            self.mount(build_url_from_netloc(host) + ":", self._trusted_host_adapter)

    def iter_secure_origins(self) -> Iterator[SecureOrigin]:
        yield from SECURE_ORIGINS
        for host, port in self.pip_trusted_origins:
            yield ("*", host, "*" if port is None else port)

    def is_secure_origin(self, location: Link) -> bool:
        # Determine if this url used a secure transport mechanism
        parsed = urllib.parse.urlparse(str(location))
        origin_protocol, origin_host, origin_port = (
            parsed.scheme,
            parsed.hostname,
            parsed.port,
        )

        # The protocol to use to see if the protocol matches.
        # Don't count the repository type as part of the protocol: in
        # cases such as "git+ssh", only use "ssh". (I.e., Only verify against
        # the last scheme.)
        origin_protocol = origin_protocol.rsplit("+", 1)[-1]

        # Determine if our origin is a secure origin by looking through our
        # hardcoded list of secure origins, as well as any additional ones
        # configured on this PackageFinder instance.
        for secure_origin in self.iter_secure_origins():
            secure_protocol, secure_host, secure_port = secure_origin
            if origin_protocol != secure_protocol and secure_protocol != "*":
                continue

            try:
                addr = ipaddress.ip_address(origin_host)
                network = ipaddress.ip_network(secure_host)
            except ValueError:
                # We don't have both a valid address or a valid network, so
                # we'll check this origin against hostnames.
                if (
                    origin_host
                    and origin_host.lower() != secure_host.lower()
                    and secure_host != "*"
                ):
                    continue
            else:
                # We have a valid address and network, so see if the address
                # is contained within the network.
                if addr not in network:
                    continue

            # Check to see if the port matches.
            if (
                origin_port != secure_port
                and secure_port != "*"
                and secure_port is not None
            ):
                continue

            # If we've gotten here, then this origin matches the current
            # secure origin and we should return True
            return True

        # If we've gotten to this point, then the origin isn't secure and we
        # will not accept it as a valid location to search. We will however
        # log a warning that we are ignoring it.
        logger.warning(
            "The repository located at %s is not a trusted or secure host and "
            "is being ignored. If this repository is available via HTTPS we "
            "recommend you use HTTPS instead, otherwise you may silence "
            "this warning and allow it anyway with '--trusted-host %s'.",
            origin_host,
            origin_host,
        )

        return False

    def request(self, method: str, url: str, *args: Any, **kwargs: Any) -> Response:
        # Allow setting a default timeout on a session
        kwargs.setdefault("timeout", self.timeout)

        # Dispatch the actual request
        return super().request(method, url, *args, **kwargs)
site-packages/pip/_internal/network/__init__.py000064400000000062150564745010015636 0ustar00"""Contains purely network-related utilities.
"""
site-packages/pip/_internal/network/download.py000064400000013600150564745010015710 0ustar00"""Download files with progress indicators.
"""
import cgi
import logging
import mimetypes
import os
from typing import Iterable, Optional, Tuple

from pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response

from pip._internal.cli.progress_bars import DownloadProgressProvider
from pip._internal.exceptions import NetworkConnectionError
from pip._internal.models.index import PyPI
from pip._internal.models.link import Link
from pip._internal.network.cache import is_from_cache
from pip._internal.network.session import PipSession
from pip._internal.network.utils import HEADERS, raise_for_status, response_chunks
from pip._internal.utils.misc import format_size, redact_auth_from_url, splitext

logger = logging.getLogger(__name__)


def _get_http_response_size(resp: Response) -> Optional[int]:
    try:
        return int(resp.headers["content-length"])
    except (ValueError, KeyError, TypeError):
        return None


def _prepare_download(
    resp: Response,
    link: Link,
    progress_bar: str,
) -> Iterable[bytes]:
    total_length = _get_http_response_size(resp)

    if link.netloc == PyPI.file_storage_domain:
        url = link.show_url
    else:
        url = link.url_without_fragment

    logged_url = redact_auth_from_url(url)

    if total_length:
        logged_url = "{} ({})".format(logged_url, format_size(total_length))

    if is_from_cache(resp):
        logger.info("Using cached %s", logged_url)
    else:
        logger.info("Downloading %s", logged_url)

    if logger.getEffectiveLevel() > logging.INFO:
        show_progress = False
    elif is_from_cache(resp):
        show_progress = False
    elif not total_length:
        show_progress = True
    elif total_length > (40 * 1000):
        show_progress = True
    else:
        show_progress = False

    chunks = response_chunks(resp, CONTENT_CHUNK_SIZE)

    if not show_progress:
        return chunks

    return DownloadProgressProvider(progress_bar, max=total_length)(chunks)


def sanitize_content_filename(filename: str) -> str:
    """
    Sanitize the "filename" value from a Content-Disposition header.
    """
    return os.path.basename(filename)


def parse_content_disposition(content_disposition: str, default_filename: str) -> str:
    """
    Parse the "filename" value from a Content-Disposition header, and
    return the default filename if the result is empty.
    """
    _type, params = cgi.parse_header(content_disposition)
    filename = params.get("filename")
    if filename:
        # We need to sanitize the filename to prevent directory traversal
        # in case the filename contains ".." path parts.
        filename = sanitize_content_filename(filename)
    return filename or default_filename


def _get_http_response_filename(resp: Response, link: Link) -> str:
    """Get an ideal filename from the given HTTP response, falling back to
    the link filename if not provided.
    """
    filename = link.filename  # fallback
    # Have a look at the Content-Disposition header for a better guess
    content_disposition = resp.headers.get("content-disposition")
    if content_disposition:
        filename = parse_content_disposition(content_disposition, filename)
    ext: Optional[str] = splitext(filename)[1]
    if not ext:
        ext = mimetypes.guess_extension(resp.headers.get("content-type", ""))
        if ext:
            filename += ext
    if not ext and link.url != resp.url:
        ext = os.path.splitext(resp.url)[1]
        if ext:
            filename += ext
    return filename


def _http_get_download(session: PipSession, link: Link) -> Response:
    target_url = link.url.split("#", 1)[0]
    resp = session.get(target_url, headers=HEADERS, stream=True)
    raise_for_status(resp)
    return resp


class Downloader:
    def __init__(
        self,
        session: PipSession,
        progress_bar: str,
    ) -> None:
        self._session = session
        self._progress_bar = progress_bar

    def __call__(self, link: Link, location: str) -> Tuple[str, str]:
        """Download the file given by link into location."""
        try:
            resp = _http_get_download(self._session, link)
        except NetworkConnectionError as e:
            assert e.response is not None
            logger.critical(
                "HTTP error %s while getting %s", e.response.status_code, link
            )
            raise

        filename = _get_http_response_filename(resp, link)
        filepath = os.path.join(location, filename)

        chunks = _prepare_download(resp, link, self._progress_bar)
        with open(filepath, "wb") as content_file:
            for chunk in chunks:
                content_file.write(chunk)
        content_type = resp.headers.get("Content-Type", "")
        return filepath, content_type


class BatchDownloader:
    def __init__(
        self,
        session: PipSession,
        progress_bar: str,
    ) -> None:
        self._session = session
        self._progress_bar = progress_bar

    def __call__(
        self, links: Iterable[Link], location: str
    ) -> Iterable[Tuple[Link, Tuple[str, str]]]:
        """Download the files given by links into location."""
        for link in links:
            try:
                resp = _http_get_download(self._session, link)
            except NetworkConnectionError as e:
                assert e.response is not None
                logger.critical(
                    "HTTP error %s while getting %s",
                    e.response.status_code,
                    link,
                )
                raise

            filename = _get_http_response_filename(resp, link)
            filepath = os.path.join(location, filename)

            chunks = _prepare_download(resp, link, self._progress_bar)
            with open(filepath, "wb") as content_file:
                for chunk in chunks:
                    content_file.write(chunk)
            content_type = resp.headers.get("Content-Type", "")
            yield link, (filepath, content_type)
site-packages/pip/_internal/network/auth.py000064400000027636150564745010015060 0ustar00"""Network Authentication Helpers

Contains interface (MultiDomainBasicAuth) and associated glue code for
providing credentials in the context of network requests.
"""

import urllib.parse
from typing import Any, Dict, List, Optional, Tuple

from pip._vendor.requests.auth import AuthBase, HTTPBasicAuth
from pip._vendor.requests.models import Request, Response
from pip._vendor.requests.utils import get_netrc_auth

from pip._internal.utils.logging import getLogger
from pip._internal.utils.misc import (
    ask,
    ask_input,
    ask_password,
    remove_auth_from_url,
    split_auth_netloc_from_url,
)
from pip._internal.vcs.versioncontrol import AuthInfo

logger = getLogger(__name__)

Credentials = Tuple[str, str, str]

try:
    import keyring
except ImportError:
    keyring = None  # type: ignore[assignment]
except Exception as exc:
    logger.warning(
        "Keyring is skipped due to an exception: %s",
        str(exc),
    )
    keyring = None  # type: ignore[assignment]


def get_keyring_auth(url: Optional[str], username: Optional[str]) -> Optional[AuthInfo]:
    """Return the tuple auth for a given url from keyring."""
    global keyring
    if not url or not keyring:
        return None

    try:
        try:
            get_credential = keyring.get_credential
        except AttributeError:
            pass
        else:
            logger.debug("Getting credentials from keyring for %s", url)
            cred = get_credential(url, username)
            if cred is not None:
                return cred.username, cred.password
            return None

        if username:
            logger.debug("Getting password from keyring for %s", url)
            password = keyring.get_password(url, username)
            if password:
                return username, password

    except Exception as exc:
        logger.warning(
            "Keyring is skipped due to an exception: %s",
            str(exc),
        )
        keyring = None  # type: ignore[assignment]
    return None


class MultiDomainBasicAuth(AuthBase):
    def __init__(
        self, prompting: bool = True, index_urls: Optional[List[str]] = None
    ) -> None:
        self.prompting = prompting
        self.index_urls = index_urls
        self.passwords: Dict[str, AuthInfo] = {}
        # When the user is prompted to enter credentials and keyring is
        # available, we will offer to save them. If the user accepts,
        # this value is set to the credentials they entered. After the
        # request authenticates, the caller should call
        # ``save_credentials`` to save these.
        self._credentials_to_save: Optional[Credentials] = None

    def _get_index_url(self, url: str) -> Optional[str]:
        """Return the original index URL matching the requested URL.

        Cached or dynamically generated credentials may work against
        the original index URL rather than just the netloc.

        The provided url should have had its username and password
        removed already. If the original index url had credentials then
        they will be included in the return value.

        Returns None if no matching index was found, or if --no-index
        was specified by the user.
        """
        if not url or not self.index_urls:
            return None

        for u in self.index_urls:
            prefix = remove_auth_from_url(u).rstrip("/") + "/"
            if url.startswith(prefix):
                return u
        return None

    def _get_new_credentials(
        self,
        original_url: str,
        allow_netrc: bool = True,
        allow_keyring: bool = False,
    ) -> AuthInfo:
        """Find and return credentials for the specified URL."""
        # Split the credentials and netloc from the url.
        url, netloc, url_user_password = split_auth_netloc_from_url(
            original_url,
        )

        # Start with the credentials embedded in the url
        username, password = url_user_password
        if username is not None and password is not None:
            logger.debug("Found credentials in url for %s", netloc)
            return url_user_password

        # Find a matching index url for this request
        index_url = self._get_index_url(url)
        if index_url:
            # Split the credentials from the url.
            index_info = split_auth_netloc_from_url(index_url)
            if index_info:
                index_url, _, index_url_user_password = index_info
                logger.debug("Found index url %s", index_url)

        # If an index URL was found, try its embedded credentials
        if index_url and index_url_user_password[0] is not None:
            username, password = index_url_user_password
            if username is not None and password is not None:
                logger.debug("Found credentials in index url for %s", netloc)
                return index_url_user_password

        # Get creds from netrc if we still don't have them
        if allow_netrc:
            netrc_auth = get_netrc_auth(original_url)
            if netrc_auth:
                logger.debug("Found credentials in netrc for %s", netloc)
                return netrc_auth

        # If we don't have a password and keyring is available, use it.
        if allow_keyring:
            # The index url is more specific than the netloc, so try it first
            # fmt: off
            kr_auth = (
                get_keyring_auth(index_url, username) or
                get_keyring_auth(netloc, username)
            )
            # fmt: on
            if kr_auth:
                logger.debug("Found credentials in keyring for %s", netloc)
                return kr_auth

        return username, password

    def _get_url_and_credentials(
        self, original_url: str
    ) -> Tuple[str, Optional[str], Optional[str]]:
        """Return the credentials to use for the provided URL.

        If allowed, netrc and keyring may be used to obtain the
        correct credentials.

        Returns (url_without_credentials, username, password). Note
        that even if the original URL contains credentials, this
        function may return a different username and password.
        """
        url, netloc, _ = split_auth_netloc_from_url(original_url)

        # Try to get credentials from original url
        username, password = self._get_new_credentials(original_url)

        # If credentials not found, use any stored credentials for this netloc.
        # Do this if either the username or the password is missing.
        # This accounts for the situation in which the user has specified
        # the username in the index url, but the password comes from keyring.
        if (username is None or password is None) and netloc in self.passwords:
            un, pw = self.passwords[netloc]
            # It is possible that the cached credentials are for a different username,
            # in which case the cache should be ignored.
            if username is None or username == un:
                username, password = un, pw

        if username is not None or password is not None:
            # Convert the username and password if they're None, so that
            # this netloc will show up as "cached" in the conditional above.
            # Further, HTTPBasicAuth doesn't accept None, so it makes sense to
            # cache the value that is going to be used.
            username = username or ""
            password = password or ""

            # Store any acquired credentials.
            self.passwords[netloc] = (username, password)

        assert (
            # Credentials were found
            (username is not None and password is not None)
            # Credentials were not found
            or (username is None and password is None)
        ), f"Could not load credentials from url: {original_url}"

        return url, username, password

    def __call__(self, req: Request) -> Request:
        # Get credentials for this request
        url, username, password = self._get_url_and_credentials(req.url)

        # Set the url of the request to the url without any credentials
        req.url = url

        if username is not None and password is not None:
            # Send the basic auth with this request
            req = HTTPBasicAuth(username, password)(req)

        # Attach a hook to handle 401 responses
        req.register_hook("response", self.handle_401)

        return req

    # Factored out to allow for easy patching in tests
    def _prompt_for_password(
        self, netloc: str
    ) -> Tuple[Optional[str], Optional[str], bool]:
        username = ask_input(f"User for {netloc}: ")
        if not username:
            return None, None, False
        auth = get_keyring_auth(netloc, username)
        if auth and auth[0] is not None and auth[1] is not None:
            return auth[0], auth[1], False
        password = ask_password("Password: ")
        return username, password, True

    # Factored out to allow for easy patching in tests
    def _should_save_password_to_keyring(self) -> bool:
        if not keyring:
            return False
        return ask("Save credentials to keyring [y/N]: ", ["y", "n"]) == "y"

    def handle_401(self, resp: Response, **kwargs: Any) -> Response:
        # We only care about 401 responses, anything else we want to just
        #   pass through the actual response
        if resp.status_code != 401:
            return resp

        # We are not able to prompt the user so simply return the response
        if not self.prompting:
            return resp

        parsed = urllib.parse.urlparse(resp.url)

        # Query the keyring for credentials:
        username, password = self._get_new_credentials(
            resp.url,
            allow_netrc=False,
            allow_keyring=True,
        )

        # Prompt the user for a new username and password
        save = False
        if not username and not password:
            username, password, save = self._prompt_for_password(parsed.netloc)

        # Store the new username and password to use for future requests
        self._credentials_to_save = None
        if username is not None and password is not None:
            self.passwords[parsed.netloc] = (username, password)

            # Prompt to save the password to keyring
            if save and self._should_save_password_to_keyring():
                self._credentials_to_save = (parsed.netloc, username, password)

        # Consume content and release the original connection to allow our new
        #   request to reuse the same one.
        resp.content
        resp.raw.release_conn()

        # Add our new username and password to the request
        req = HTTPBasicAuth(username or "", password or "")(resp.request)
        req.register_hook("response", self.warn_on_401)

        # On successful request, save the credentials that were used to
        # keyring. (Note that if the user responded "no" above, this member
        # is not set and nothing will be saved.)
        if self._credentials_to_save:
            req.register_hook("response", self.save_credentials)

        # Send our new request
        new_resp = resp.connection.send(req, **kwargs)
        new_resp.history.append(resp)

        return new_resp

    def warn_on_401(self, resp: Response, **kwargs: Any) -> None:
        """Response callback to warn about incorrect credentials."""
        if resp.status_code == 401:
            logger.warning(
                "401 Error, Credentials not correct for %s",
                resp.request.url,
            )

    def save_credentials(self, resp: Response, **kwargs: Any) -> None:
        """Response callback to save credentials on success."""
        assert keyring is not None, "should never reach here without keyring"
        if not keyring:
            return

        creds = self._credentials_to_save
        self._credentials_to_save = None
        if creds and resp.status_code < 400:
            try:
                logger.info("Saving credentials to keyring")
                keyring.set_password(*creds)
            except Exception:
                logger.exception("Failed to save credentials")
site-packages/pip/_internal/network/lazy_wheel.py000064400000016713150564745010016254 0ustar00"""Lazy ZIP over HTTP"""

__all__ = ["HTTPRangeRequestUnsupported", "dist_from_wheel_url"]

from bisect import bisect_left, bisect_right
from contextlib import contextmanager
from tempfile import NamedTemporaryFile
from typing import Any, Dict, Iterator, List, Optional, Tuple
from zipfile import BadZipfile, ZipFile

from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.requests.models import CONTENT_CHUNK_SIZE, Response

from pip._internal.metadata import BaseDistribution, MemoryWheel, get_wheel_distribution
from pip._internal.network.session import PipSession
from pip._internal.network.utils import HEADERS, raise_for_status, response_chunks


class HTTPRangeRequestUnsupported(Exception):
    pass


def dist_from_wheel_url(name: str, url: str, session: PipSession) -> BaseDistribution:
    """Return a distribution object from the given wheel URL.

    This uses HTTP range requests to only fetch the potion of the wheel
    containing metadata, just enough for the object to be constructed.
    If such requests are not supported, HTTPRangeRequestUnsupported
    is raised.
    """
    with LazyZipOverHTTP(url, session) as zf:
        # For read-only ZIP files, ZipFile only needs methods read,
        # seek, seekable and tell, not the whole IO protocol.
        wheel = MemoryWheel(zf.name, zf)  # type: ignore
        # After context manager exit, wheel.name
        # is an invalid file by intention.
        return get_wheel_distribution(wheel, canonicalize_name(name))


class LazyZipOverHTTP:
    """File-like object mapped to a ZIP file over HTTP.

    This uses HTTP range requests to lazily fetch the file's content,
    which is supposed to be fed to ZipFile.  If such requests are not
    supported by the server, raise HTTPRangeRequestUnsupported
    during initialization.
    """

    def __init__(
        self, url: str, session: PipSession, chunk_size: int = CONTENT_CHUNK_SIZE
    ) -> None:
        head = session.head(url, headers=HEADERS)
        raise_for_status(head)
        assert head.status_code == 200
        self._session, self._url, self._chunk_size = session, url, chunk_size
        self._length = int(head.headers["Content-Length"])
        self._file = NamedTemporaryFile()
        self.truncate(self._length)
        self._left: List[int] = []
        self._right: List[int] = []
        if "bytes" not in head.headers.get("Accept-Ranges", "none"):
            raise HTTPRangeRequestUnsupported("range request is not supported")
        self._check_zip()

    @property
    def mode(self) -> str:
        """Opening mode, which is always rb."""
        return "rb"

    @property
    def name(self) -> str:
        """Path to the underlying file."""
        return self._file.name

    def seekable(self) -> bool:
        """Return whether random access is supported, which is True."""
        return True

    def close(self) -> None:
        """Close the file."""
        self._file.close()

    @property
    def closed(self) -> bool:
        """Whether the file is closed."""
        return self._file.closed

    def read(self, size: int = -1) -> bytes:
        """Read up to size bytes from the object and return them.

        As a convenience, if size is unspecified or -1,
        all bytes until EOF are returned.  Fewer than
        size bytes may be returned if EOF is reached.
        """
        download_size = max(size, self._chunk_size)
        start, length = self.tell(), self._length
        stop = length if size < 0 else min(start + download_size, length)
        start = max(0, stop - download_size)
        self._download(start, stop - 1)
        return self._file.read(size)

    def readable(self) -> bool:
        """Return whether the file is readable, which is True."""
        return True

    def seek(self, offset: int, whence: int = 0) -> int:
        """Change stream position and return the new absolute position.

        Seek to offset relative position indicated by whence:
        * 0: Start of stream (the default).  pos should be >= 0;
        * 1: Current position - pos may be negative;
        * 2: End of stream - pos usually negative.
        """
        return self._file.seek(offset, whence)

    def tell(self) -> int:
        """Return the current position."""
        return self._file.tell()

    def truncate(self, size: Optional[int] = None) -> int:
        """Resize the stream to the given size in bytes.

        If size is unspecified resize to the current position.
        The current stream position isn't changed.

        Return the new file size.
        """
        return self._file.truncate(size)

    def writable(self) -> bool:
        """Return False."""
        return False

    def __enter__(self) -> "LazyZipOverHTTP":
        self._file.__enter__()
        return self

    def __exit__(self, *exc: Any) -> Optional[bool]:
        return self._file.__exit__(*exc)

    @contextmanager
    def _stay(self) -> Iterator[None]:
        """Return a context manager keeping the position.

        At the end of the block, seek back to original position.
        """
        pos = self.tell()
        try:
            yield
        finally:
            self.seek(pos)

    def _check_zip(self) -> None:
        """Check and download until the file is a valid ZIP."""
        end = self._length - 1
        for start in reversed(range(0, end, self._chunk_size)):
            self._download(start, end)
            with self._stay():
                try:
                    # For read-only ZIP files, ZipFile only needs
                    # methods read, seek, seekable and tell.
                    ZipFile(self)  # type: ignore
                except BadZipfile:
                    pass
                else:
                    break

    def _stream_response(
        self, start: int, end: int, base_headers: Dict[str, str] = HEADERS
    ) -> Response:
        """Return HTTP response to a range request from start to end."""
        headers = base_headers.copy()
        headers["Range"] = f"bytes={start}-{end}"
        # TODO: Get range requests to be correctly cached
        headers["Cache-Control"] = "no-cache"
        return self._session.get(self._url, headers=headers, stream=True)

    def _merge(
        self, start: int, end: int, left: int, right: int
    ) -> Iterator[Tuple[int, int]]:
        """Return an iterator of intervals to be fetched.

        Args:
            start (int): Start of needed interval
            end (int): End of needed interval
            left (int): Index of first overlapping downloaded data
            right (int): Index after last overlapping downloaded data
        """
        lslice, rslice = self._left[left:right], self._right[left:right]
        i = start = min([start] + lslice[:1])
        end = max([end] + rslice[-1:])
        for j, k in zip(lslice, rslice):
            if j > i:
                yield i, j - 1
            i = k + 1
        if i <= end:
            yield i, end
        self._left[left:right], self._right[left:right] = [start], [end]

    def _download(self, start: int, end: int) -> None:
        """Download bytes from start to end inclusively."""
        with self._stay():
            left = bisect_left(self._right, start)
            right = bisect_right(self._left, end)
            for start, end in self._merge(start, end, left, right):
                response = self._stream_response(start, end)
                response.raise_for_status()
                self.seek(start)
                for chunk in response_chunks(response, self._chunk_size):
                    self._file.write(chunk)
site-packages/pip/_internal/network/xmlrpc.py000064400000003377150564745010015420 0ustar00"""xmlrpclib.Transport implementation
"""

import logging
import urllib.parse
import xmlrpc.client
from typing import TYPE_CHECKING, Tuple

from pip._internal.exceptions import NetworkConnectionError
from pip._internal.network.session import PipSession
from pip._internal.network.utils import raise_for_status

if TYPE_CHECKING:
    from xmlrpc.client import _HostType, _Marshallable

logger = logging.getLogger(__name__)


class PipXmlrpcTransport(xmlrpc.client.Transport):
    """Provide a `xmlrpclib.Transport` implementation via a `PipSession`
    object.
    """

    def __init__(
        self, index_url: str, session: PipSession, use_datetime: bool = False
    ) -> None:
        super().__init__(use_datetime)
        index_parts = urllib.parse.urlparse(index_url)
        self._scheme = index_parts.scheme
        self._session = session

    def request(
        self,
        host: "_HostType",
        handler: str,
        request_body: bytes,
        verbose: bool = False,
    ) -> Tuple["_Marshallable", ...]:
        assert isinstance(host, str)
        parts = (self._scheme, host, handler, None, None, None)
        url = urllib.parse.urlunparse(parts)
        try:
            headers = {"Content-Type": "text/xml"}
            response = self._session.post(
                url,
                data=request_body,
                headers=headers,
                stream=True,
            )
            raise_for_status(response)
            self.verbose = verbose
            return self.parse_response(response.raw)
        except NetworkConnectionError as exc:
            assert exc.response
            logger.critical(
                "HTTP error %s while getting %s",
                exc.response.status_code,
                url,
            )
            raise
site-packages/pip/_internal/network/cache.py000064400000004064150564745010015150 0ustar00"""HTTP cache implementation.
"""

import os
from contextlib import contextmanager
from typing import Iterator, Optional

from pip._vendor.cachecontrol.cache import BaseCache
from pip._vendor.cachecontrol.caches import FileCache
from pip._vendor.requests.models import Response

from pip._internal.utils.filesystem import adjacent_tmp_file, replace
from pip._internal.utils.misc import ensure_dir


def is_from_cache(response: Response) -> bool:
    return getattr(response, "from_cache", False)


@contextmanager
def suppressed_cache_errors() -> Iterator[None]:
    """If we can't access the cache then we can just skip caching and process
    requests as if caching wasn't enabled.
    """
    try:
        yield
    except OSError:
        pass


class SafeFileCache(BaseCache):
    """
    A file based cache which is safe to use even when the target directory may
    not be accessible or writable.
    """

    def __init__(self, directory: str) -> None:
        assert directory is not None, "Cache directory must not be None."
        super().__init__()
        self.directory = directory

    def _get_cache_path(self, name: str) -> str:
        # From cachecontrol.caches.file_cache.FileCache._fn, brought into our
        # class for backwards-compatibility and to avoid using a non-public
        # method.
        hashed = FileCache.encode(name)
        parts = list(hashed[:5]) + [hashed]
        return os.path.join(self.directory, *parts)

    def get(self, key: str) -> Optional[bytes]:
        path = self._get_cache_path(key)
        with suppressed_cache_errors():
            with open(path, "rb") as f:
                return f.read()

    def set(self, key: str, value: bytes) -> None:
        path = self._get_cache_path(key)
        with suppressed_cache_errors():
            ensure_dir(os.path.dirname(path))

            with adjacent_tmp_file(path) as f:
                f.write(value)

            replace(f.name, path)

    def delete(self, key: str) -> None:
        path = self._get_cache_path(key)
        with suppressed_cache_errors():
            os.remove(path)
site-packages/pip/_internal/locations/_sysconfig.py000064400000017356150564745010016562 0ustar00import distutils.util  # FIXME: For change_root.
import logging
import os
import sys
import sysconfig
import typing

from pip._internal.exceptions import InvalidSchemeCombination, UserInstallationInvalid
from pip._internal.models.scheme import SCHEME_KEYS, Scheme
from pip._internal.utils.virtualenv import running_under_virtualenv

from .base import get_major_minor_version, is_osx_framework

logger = logging.getLogger(__name__)


# Notes on _infer_* functions.
# Unfortunately ``get_default_scheme()`` didn't exist before 3.10, so there's no
# way to ask things like "what is the '_prefix' scheme on this platform". These
# functions try to answer that with some heuristics while accounting for ad-hoc
# platforms not covered by CPython's default sysconfig implementation. If the
# ad-hoc implementation does not fully implement sysconfig, we'll fall back to
# a POSIX scheme.

_AVAILABLE_SCHEMES = set(sysconfig.get_scheme_names())

_PREFERRED_SCHEME_API = getattr(sysconfig, "get_preferred_scheme", None)


def _should_use_osx_framework_prefix() -> bool:
    """Check for Apple's ``osx_framework_library`` scheme.

    Python distributed by Apple's Command Line Tools has this special scheme
    that's used when:

    * This is a framework build.
    * We are installing into the system prefix.

    This does not account for ``pip install --prefix`` (also means we're not
    installing to the system prefix), which should use ``posix_prefix``, but
    logic here means ``_infer_prefix()`` outputs ``osx_framework_library``. But
    since ``prefix`` is not available for ``sysconfig.get_default_scheme()``,
    which is the stdlib replacement for ``_infer_prefix()``, presumably Apple
    wouldn't be able to magically switch between ``osx_framework_library`` and
    ``posix_prefix``. ``_infer_prefix()`` returning ``osx_framework_library``
    means its behavior is consistent whether we use the stdlib implementation
    or our own, and we deal with this special case in ``get_scheme()`` instead.
    """
    return (
        "osx_framework_library" in _AVAILABLE_SCHEMES
        and not running_under_virtualenv()
        and is_osx_framework()
    )


def _infer_prefix() -> str:
    """Try to find a prefix scheme for the current platform.

    This tries:

    * A special ``osx_framework_library`` for Python distributed by Apple's
      Command Line Tools, when not running in a virtual environment.
    * Implementation + OS, used by PyPy on Windows (``pypy_nt``).
    * Implementation without OS, used by PyPy on POSIX (``pypy``).
    * OS + "prefix", used by CPython on POSIX (``posix_prefix``).
    * Just the OS name, used by CPython on Windows (``nt``).

    If none of the above works, fall back to ``posix_prefix``.
    """
    if _PREFERRED_SCHEME_API:
        return _PREFERRED_SCHEME_API("prefix")
    if _should_use_osx_framework_prefix():
        return "osx_framework_library"
    implementation_suffixed = f"{sys.implementation.name}_{os.name}"
    if implementation_suffixed in _AVAILABLE_SCHEMES:
        return implementation_suffixed
    if sys.implementation.name in _AVAILABLE_SCHEMES:
        return sys.implementation.name
    suffixed = f"{os.name}_prefix"
    if suffixed in _AVAILABLE_SCHEMES:
        return suffixed
    if os.name in _AVAILABLE_SCHEMES:  # On Windows, prefx is just called "nt".
        return os.name
    return "posix_prefix"


def _infer_user() -> str:
    """Try to find a user scheme for the current platform."""
    if _PREFERRED_SCHEME_API:
        return _PREFERRED_SCHEME_API("user")
    if is_osx_framework() and not running_under_virtualenv():
        suffixed = "osx_framework_user"
    else:
        suffixed = f"{os.name}_user"
    if suffixed in _AVAILABLE_SCHEMES:
        return suffixed
    if "posix_user" not in _AVAILABLE_SCHEMES:  # User scheme unavailable.
        raise UserInstallationInvalid()
    return "posix_user"


def _infer_home() -> str:
    """Try to find a home for the current platform."""
    if _PREFERRED_SCHEME_API:
        return _PREFERRED_SCHEME_API("home")
    suffixed = f"{os.name}_home"
    if suffixed in _AVAILABLE_SCHEMES:
        return suffixed
    return "posix_home"


# Update these keys if the user sets a custom home.
_HOME_KEYS = [
    "installed_base",
    "base",
    "installed_platbase",
    "platbase",
    "prefix",
    "exec_prefix",
]
if sysconfig.get_config_var("userbase") is not None:
    _HOME_KEYS.append("userbase")


def get_scheme(
    dist_name: str,
    user: bool = False,
    home: typing.Optional[str] = None,
    root: typing.Optional[str] = None,
    isolated: bool = False,
    prefix: typing.Optional[str] = None,
) -> Scheme:
    """
    Get the "scheme" corresponding to the input parameters.

    :param dist_name: the name of the package to retrieve the scheme for, used
        in the headers scheme path
    :param user: indicates to use the "user" scheme
    :param home: indicates to use the "home" scheme
    :param root: root under which other directories are re-based
    :param isolated: ignored, but kept for distutils compatibility (where
        this controls whether the user-site pydistutils.cfg is honored)
    :param prefix: indicates to use the "prefix" scheme and provides the
        base directory for the same
    """
    if user and prefix:
        raise InvalidSchemeCombination("--user", "--prefix")
    if home and prefix:
        raise InvalidSchemeCombination("--home", "--prefix")

    if home is not None:
        scheme_name = _infer_home()
    elif user:
        scheme_name = _infer_user()
    else:
        scheme_name = _infer_prefix()

    # Special case: When installing into a custom prefix, use posix_prefix
    # instead of osx_framework_library. See _should_use_osx_framework_prefix()
    # docstring for details.
    if prefix is not None and scheme_name == "osx_framework_library":
        scheme_name = "posix_prefix"

    if home is not None:
        variables = {k: home for k in _HOME_KEYS}
    elif prefix is not None:
        variables = {k: prefix for k in _HOME_KEYS}
    else:
        variables = {}

    paths = sysconfig.get_paths(scheme=scheme_name, vars=variables)

    # Logic here is very arbitrary, we're doing it for compatibility, don't ask.
    # 1. Pip historically uses a special header path in virtual environments.
    # 2. If the distribution name is not known, distutils uses 'UNKNOWN'. We
    #    only do the same when not running in a virtual environment because
    #    pip's historical header path logic (see point 1) did not do this.
    if running_under_virtualenv():
        if user:
            base = variables.get("userbase", sys.prefix)
        else:
            base = variables.get("base", sys.prefix)
        python_xy = f"python{get_major_minor_version()}"
        paths["include"] = os.path.join(base, "include", "site", python_xy)
    elif not dist_name:
        dist_name = "UNKNOWN"

    scheme = Scheme(
        platlib=paths["platlib"],
        purelib=paths["purelib"],
        headers=os.path.join(paths["include"], dist_name),
        scripts=paths["scripts"],
        data=paths["data"],
    )
    if root is not None:
        for key in SCHEME_KEYS:
            value = distutils.util.change_root(root, getattr(scheme, key))
            setattr(scheme, key, value)
    return scheme


def get_bin_prefix() -> str:
    # Forcing to use /usr/local/bin for standard macOS framework installs.
    if sys.platform[:6] == "darwin" and sys.prefix[:16] == "/System/Library/":
        return "/usr/local/bin"
    return sysconfig.get_paths()["scripts"]


def get_purelib() -> str:
    return sysconfig.get_paths()["purelib"]


def get_platlib() -> str:
    return sysconfig.get_paths()["platlib"]


def get_prefixed_libs(prefix: str) -> typing.Tuple[str, str]:
    paths = sysconfig.get_paths(vars={"base": prefix, "platbase": prefix})
    return (paths["purelib"], paths["platlib"])
site-packages/pip/_internal/locations/__pycache__/_distutils.cpython-39.pyc000064400000011140150564745010023052 0ustar00a

��Re��@s@dZddlZddlZddlZddlmZddlmZddlm	Z
ddlmZddl
mZmZmZmZmZmZddlmZdd	lmZdd
lmZddlmZe�e�Zdd
d�ee eee ee eeefd�dd�Z!d ee eeeee eeed�dd�Z"ed�dd�Z#ed�dd�Z$ed�dd�Z%eeeefd�dd�Z&dS)!z7Locations where we look for configs, install stuff, etc�N)�Command)�SCHEME_KEYS)�install��get_python_lib)�Dict�List�Optional�Tuple�Union�cast)�Scheme)�WINDOWS)�running_under_virtualenv�)�get_major_minor_versionF)�ignore_config_files)�	dist_name�user�home�root�isolated�prefixr�returnc
Cs�ddlm}d|i}|r"dg|d<||�}	|srz|	��Wn6typ|	��}
t�dd�dd	�|
D���Yn0d
}|	jddd
�}|d
us�J�t	t
|�}|r�|r�Jd|�d|����|r�|r�Jd|�d|����|p�|j|_|s�|r�d|_|p�|j|_|�p|j
|_
|�p|j|_|��i}
tD]}t|d|�|
|<�q$d|	�d�v�rd|
�t|j|jd��t��r�|�rx|}n|�r�|j}n|j}tj�|dddt���|�|
d<|d
u�r�tj�tj�|
d��d}tj�||dd
��|
d<|
S)z+
    Return a distutils install scheme
    r)�Distribution�namez
--no-user-cfg�script_argsz6Ignore distutils configs in %s due to encoding errors.z, css|]}tj�|�VqdS)N)�os�path�basename)�.0�p�r"��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/locations/_distutils.py�	<genexpr>4�z#distutils_scheme.<locals>.<genexpr>NrT)�createzuser=z prefix=zhome=��install_�install_lib)�purelib�platlib�include�site�python�headersr)�distutils.distr�parse_config_files�UnicodeDecodeError�find_config_files�logger�warning�join�get_command_objr�distutils_install_commandrrrr�finalize_optionsr�getattr�get_option_dict�update�dictr)r�install_userbaserrr�
splitdrive�abspath)rrrrrrrr�	dist_args�d�paths�obj�i�scheme�key�
path_no_driver"r"r#�distutils_schemesb

�


�
rI)rrrrrrrcCs8t||||||�}t|d|d|d|d|dd�S)a;
    Get the "scheme" corresponding to the input parameters. The distutils
    documentation provides the context for the available schemes:
    https://docs.python.org/3/install/index.html#alternate-installation

    :param dist_name: the name of the package to retrieve the scheme for, used
        in the headers scheme path
    :param user: indicates to use the "user" scheme
    :param home: indicates to use the "home" scheme and provides the base
        directory for the same
    :param root: root under which other directories are re-based
    :param isolated: equivalent to --no-user-cfg, i.e. do not consider
        ~/.pydistutils.cfg (posix) or ~/pydistutils.cfg (non-posix) for
        scheme paths
    :param prefix: indicates to use the "prefix" scheme and provides the
        base directory for the same
    r+r*r/�scripts�data)r+r*r/rJrK)rIr
)rrrrrrrFr"r"r#�
get_schemeis�rL)rcCsrtj�tj�}tr>tj�|d�}tj�|�s:tj�|d�}|Stjdd�dkrd|dd�dkrddStj�|d�S)N�Scripts�bin��darwin�z/System/Library/z/usr/local/bin)	rr�normpath�sysrrr6�exists�platform)r�bin_pyr"r"r#�get_bin_prefix�s"rWcCs
tdd�S)NF��
plat_specificrr"r"r"r#�get_purelib�srZcCs
tdd�S)NTrXrr"r"r"r#�get_platlib�sr[)rrcCstd|d�td|d�fS)NF)rYrTr)rr"r"r#�get_prefixed_libs�s

�r\)FNNFN)FNNFN)'�__doc__�loggingrrS�
distutils.cmdr�DistutilsCommand�distutils.command.installrrr8�distutils.sysconfigr�typingrrr	r
rr�pip._internal.models.schemer
�pip._internal.utils.compatr�pip._internal.utils.virtualenvr�baser�	getLogger�__name__r4�str�boolrIrLrWrZr[r\r"r"r"r#�<module>s` 
��
�S��#site-packages/pip/_internal/locations/__pycache__/_sysconfig.cpython-39.pyc000064400000014227150564745010023043 0ustar00a

��Re��	@s\ddlZddlZddlZddlZddlZddlZddlmZm	Z	ddl
mZmZddl
mZddlmZmZe�e�Zee���Zeedd�Zed�d	d
�Zed�dd�Zed�d
d�Zed�dd�Zgd�Z e�!d�dur�e �"d�d eeej#eej#eeej#eed�dd�Z$ed�dd�Z%ed�dd�Z&ed�dd�Z'eej(eefd�dd�Z)dS)!�N)�InvalidSchemeCombination�UserInstallationInvalid)�SCHEME_KEYS�Scheme)�running_under_virtualenv�)�get_major_minor_version�is_osx_frameworkZget_preferred_scheme)�returncCsdtvot�ot�S)a�Check for Apple's ``osx_framework_library`` scheme.

    Python distributed by Apple's Command Line Tools has this special scheme
    that's used when:

    * This is a framework build.
    * We are installing into the system prefix.

    This does not account for ``pip install --prefix`` (also means we're not
    installing to the system prefix), which should use ``posix_prefix``, but
    logic here means ``_infer_prefix()`` outputs ``osx_framework_library``. But
    since ``prefix`` is not available for ``sysconfig.get_default_scheme()``,
    which is the stdlib replacement for ``_infer_prefix()``, presumably Apple
    wouldn't be able to magically switch between ``osx_framework_library`` and
    ``posix_prefix``. ``_infer_prefix()`` returning ``osx_framework_library``
    means its behavior is consistent whether we use the stdlib implementation
    or our own, and we deal with this special case in ``get_scheme()`` instead.
    �osx_framework_library)�_AVAILABLE_SCHEMESrr	�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/locations/_sysconfig.py� _should_use_osx_framework_prefixs
��rcCsvtrtd�St�rdStjj�dtj��}|tvr6|StjjtvrJtjjStj�d�}|tvrb|StjtvrrtjSdS)a!Try to find a prefix scheme for the current platform.

    This tries:

    * A special ``osx_framework_library`` for Python distributed by Apple's
      Command Line Tools, when not running in a virtual environment.
    * Implementation + OS, used by PyPy on Windows (``pypy_nt``).
    * Implementation without OS, used by PyPy on POSIX (``pypy``).
    * OS + "prefix", used by CPython on POSIX (``posix_prefix``).
    * Just the OS name, used by CPython on Windows (``nt``).

    If none of the above works, fall back to ``posix_prefix``.
    �prefixr�_�_prefix�posix_prefix)�_PREFERRED_SCHEME_APIr�sys�implementation�name�osr)�implementation_suffixed�suffixedr
r
r�
_infer_prefix8s
rcCsHtrtd�St�rt�sd}ntj�d�}|tvr6|SdtvrDt��dS)z3Try to find a user scheme for the current platform.�user�osx_framework_user�_user�
posix_user)rr	rrrrr�rr
r
r�_infer_userWsr!cCs(trtd�Stj�d�}|tvr$|SdS)z,Try to find a home for the current platform.�home�_home�
posix_home)rrrrr r
r
r�_infer_homefsr%)�installed_base�base�installed_platbase�platbaser�exec_prefix�userbaseF)�	dist_namerr"�root�isolatedrr
csb|r�rtdd���r$�r$tdd���dur4t�}n|r@t�}nt�}�durZ|dkrZd}�durv�fdd�tD�}n �dur��fd	d�tD�}ni}tj||d
�}t�r�|r�|�dt	j
�}	n|�dt	j
�}	d
t���}
tj
�|	dd|
�|d<n|s�d}t|d|dtj
�|d|�|d|dd�}|du�r^tD]&}tj�|t||��}
t|||
��q6|S)a\
    Get the "scheme" corresponding to the input parameters.

    :param dist_name: the name of the package to retrieve the scheme for, used
        in the headers scheme path
    :param user: indicates to use the "user" scheme
    :param home: indicates to use the "home" scheme
    :param root: root under which other directories are re-based
    :param isolated: ignored, but kept for distutils compatibility (where
        this controls whether the user-site pydistutils.cfg is honored)
    :param prefix: indicates to use the "prefix" scheme and provides the
        base directory for the same
    z--userz--prefixz--homeNrrcsi|]
}|��qSr
r
��.0�k)r"r
r�
<dictcomp>��zget_scheme.<locals>.<dictcomp>csi|]
}|��qSr
r
r/)rr
rr2�r3)�scheme�varsr+r'�python�include�site�UNKNOWN�platlib�purelib�scripts�data)r:r;�headersr<r=)rr%r!r�
_HOME_KEYS�	sysconfig�	get_pathsr�getrrrr�path�joinrr�	distutils�util�change_root�getattr�setattr)r,rr"r-r.r�scheme_name�	variables�pathsr'�	python_xyr4�key�valuer
)r"rr�
get_scheme}sJ

�
rPcCs4tjdd�dkr(tjdd�dkr(dSt��dS)N��darwin�z/System/Library/z/usr/local/binr<)r�platformrr@rAr
r
r
r�get_bin_prefix�s$rUcCst��dS)Nr;�r@rAr
r
r
r�get_purelib�srWcCst��dS)Nr:rVr
r
r
r�get_platlib�srX)rr
cCs"tj||d�d�}|d|dfS)N)r'r))r5r;r:rV)rrLr
r
r�get_prefixed_libs�srY)FNNFN)*�distutils.utilrE�loggingrrr@�typing�pip._internal.exceptionsrr�pip._internal.models.schemerr�pip._internal.utils.virtualenvrr'rr	�	getLogger�__name__�logger�set�get_scheme_namesrrHr�boolr�strrr!r%r?�get_config_var�append�OptionalrPrUrWrX�TuplerYr
r
r
r�<module>sJ

��Msite-packages/pip/_internal/locations/__pycache__/base.cpython-39.pyc000064400000003045150564745010021606 0ustar00a

��Re+�@s�UddlZddlZddlZddlZddlZddlZddlmZddlm	Z	e�
d�Ze�d�Z
ejeed<ed�dd	�Zed�d
d�Zze��Zejeed<Wney�ejZYn0ejdd
�ed�dd��ZdS)�N)�appdirs)�running_under_virtualenv�pip�purelib�
site_packages)�returncCsdjtj�S)ze
    Return the major-minor version of the current Python as a string, e.g.
    "3.7" or "3.10".
    z{}.{})�format�sys�version_info�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/locations/base.py�get_major_minor_versionsr
cCsXt�rtj�tjd�}n4ztj�t��d�}WntyJt�d�Yn0tj�	|�S)N�srcz=The folder you are executing pip from can no longer be found.)
r�os�path�joinr	�prefix�getcwd�OSError�exit�abspath)�
src_prefixrrr�get_src_prefixsr�	user_site)�maxsizecCstt�d��S)N�PYTHONFRAMEWORK)�bool�	sysconfig�get_config_varrrrr�is_osx_framework2sr)�	functoolsr�siter	r�typingZpip._internal.utilsr�pip._internal.utils.virtualenvr�user_cache_dir�USER_CACHE_DIR�get_pathr�Optional�str�__annotations__r
r�getusersitepackagesr�AttributeError�	USER_SITE�	lru_cacherrrrrr�<module>s"


site-packages/pip/_internal/locations/__pycache__/__init__.cpython-39.pyc000064400000024625150564745010022442 0ustar00a

��Rel8�	@s�UddlZddlZddlZddlZddlZddlZddlmZmZm	Z	m
Z
mZmZddl
mZmZddlmZddlmZddlmZddlmZmZdd	lmZmZmZmZmZmZgd
�Z e�!e"�Z#ej$�%d�r�ej&Z'nej(Z'e)edd
�Z*e+e,d<ej-dkZ.e/d�dd�Z0ee+e+fe/d�dd�Z1ej2dd�e/d�dd��Z3ej2dd�e/d�dd��Z4ej2dd�e/d�dd��Z5ej2dd�e/d�dd��Z6ee+e	e+d�d d!�Z7ej2dd�ej8ej8e+dd"�d#d$��Z9ej8ej8e+e/d"�d%d&�Z:ej2dd�d'dddd(�e/ee+ee+ee+dd)�d*d+��Z;d@e+e/ee+ee+e/ee+ed,�d-d.�Z<e+d�d/d0�Z=e+d�d1d2�Z>e+e/d3�d4d5�Z?e+d�d6d7�Z@e+d�d8d9�ZAe+e+e
e+d:�d;d<�ZBe+e
e+d=�d>d?�ZCdS)A�N)�Any�Dict�Iterator�List�Optional�Tuple)�SCHEME_KEYS�Scheme)�WINDOWS)�
deprecated)�running_under_virtualenv�)�
_distutils�
_sysconfig)�USER_CACHE_DIR�get_major_minor_version�get_src_prefix�is_osx_framework�
site_packages�	user_site)r�get_bin_prefix�get_bin_userr�get_platlib�get_prefixed_libs�get_purelib�
get_schemerrrZ"_PIP_LOCATIONS_NO_WARN_ON_MISMATCH�
platlibdir�lib�_PLATLIBDIR)��
)�returncCs:ddlm}z|dd}Wnty0YdS0|dkS)zsThe resolution to bpo-44860 will change this incorrect platlib.

    See <https://bugs.python.org/issue44860>.
    r��INSTALL_SCHEMES�	unix_user�platlibFz	$usersite)�distutils.command.installr#�KeyError)r#�unix_user_platlib�r)��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/locations/__init__.py�_looks_like_bpo_448603sr+)�schemer!cCs\|d}d|vr0ttd�r0|�ddtj�d��}d|vr<dS|�dd�}|�dd	�|d
kS)Nr%z
/$platlibdir/r�/z/lib64/Fz/lib/z
$platbase/z$base/�purelib)�hasattr�sys�replacer)r,r%�	unpatchedr)r)r*�+_looks_like_red_hat_patched_platlib_purelibAsr3)�maxsizecs"ddlm�t�fdd�dD��S)z�Red Hat patches platlib in unix_prefix and unix_home, but not purelib.

    This is the only way I can see to tell a Red Hat-patched Python.
    rr"c3s"|]}|�vot�|�VqdS�N)r3��.0�kr"r)r*�	<genexpr>Ss�z*_looks_like_red_hat_lib.<locals>.<genexpr>)�unix_prefix�	unix_home)r&r#�allr)r)r"r*�_looks_like_red_hat_libKs�r=cCsddlm}d|vod|vS)z#Debian adds two additional schemes.rr"�
deb_system�
unix_local)r&r#r"r)r)r*�_looks_like_debian_schemeZsr@cCs^ddlm}ddlm}||��}|��|jtj�t	j��d�ko\|j
tj�t	j
��d�kS)a\Red Hat patches ``sys.prefix`` and ``sys.exec_prefix``.

    Red Hat's ``00251-change-user-install-location.patch`` changes the install
    command's ``prefix`` and ``exec_prefix`` to append ``"/local"``. This is
    (fortunately?) done quite unconditionally, so we create a default command
    object without any configuration to detect this.
    r)�install)�Distributionz/local)r&rA�distutils.distrB�finalize_options�exec_prefix�os�path�normpathr0�prefix)rArB�cmdr)r)r*�_looks_like_red_hat_schemebs	
�rKcs.tjddd��tdd��fdd�dD�D��S)a�MSYS2 patches distutils and sysconfig to use a UNIX-like scheme.

    However, MSYS2 incorrectly patches sysconfig ``nt`` scheme. The fix is
    likely going to be included in their 3.10 release, so we ignore the warning.
    See msys2/MINGW-packages#9319.

    MSYS2 MINGW's patch uses lowercase ``"lib"`` instead of the usual uppercase,
    and is missing the final ``"site-packages"``.
    �ntF)�expandcss*|]"}d|vo d|vo |�d�VqdS)�Librz
site-packagesN)�endswith)r7�pr)r)r*r9�s�z1_looks_like_msys2_mingw_scheme.<locals>.<genexpr>c3s|]}�|VqdSr5r))r7�key��pathsr)r*r9���r%r.)�	sysconfig�	get_pathsr<r)r)rRr*�_looks_like_msys2_mingw_schemevs�rX)�partsr!ccsht�d�}ttdd�}|r(|r(|�|�s6|EdHdS|D](}|�|�r\|ddt|��}|Vq:dS)N�	LDVERSION�abiflagsr)rV�get_config_var�getattrr0rO�len)rY�	ldversionr[�partr)r)r*�
_fix_abiflags�s


ra)�old�newrQr!cCs d}d}t�t|||||�dS)Nz(https://github.com/pypa/pip/issues/10151zSValue for %s does not match. Please report this to <%s>
distutils: %s
sysconfig: %s)�logger�log�_MISMATCH_LEVEL)rbrcrQ�	issue_url�messager)r)r*�_warn_mismatched�s�ricCs||krdSt|||d�dS)NF�rQT)ri)rbrcrQr)r)r*�_warn_if_mismatch�srkF��user�home�rootrI)rmrnrorIr!cCs&gd�}t�td�|�||||�dS)N)zAdditional context:z	user = %rz	home = %rz	root = %rzprefix = %r�
)rdrerf�join)rmrnrorIrYr)r)r*�_log_context�srr)�	dist_namermrnro�isolatedrIr!c	s�tj||||||d�}tr|Stj||||||d��g}tD�]}t�t�|��}	t�t||��}
|	|
krlq<tj	j
dko�|duo�|dvo�|	j|
jko�|	j
�d�o�|
j
�d�}|r�q<|o�t
�o�|dko�|	jj|
jko�|	jj
�d�}|r�q<|dkr�t�r�q<|�o(|dk�o(t�o(tjdk�o(tdk�o(t�}
|
�r2q<|�pH|�pH|�pHt��o�|	jd	d
�dk�o�t|
j�d	k�o�|
jd	dk�o�t|
j�d
k�s�|
jd
dk�o�t��p�t�}|�r�q<tjdk�o�t�o�|dv�o�tt|	j��|
jk}|�r�q<t�o|dv�ot�}|�rq<tjdd��o,t�o,|dv}|�r6q<|�|	|
d|��f�q<|�sX�Stj||||||dd��t��fdd�tD���r�t dddd��S|D]\}	}
}t!|	|
|d��q�t"||||d��S)N)rmrnrortrI�pypyrU�python�headersr%)r�	rr
r)�usr�localry�rz)r�)rwr%r.T)�
check_home)rw�include�platincludezscheme.)�ignore_config_filesc3s |]}�|t�|�kVqdSr5)r]r6��default_oldrbr)r*r9OrTzget_scheme.<locals>.<genexpr>z�Configuring installation scheme with distutils config files is deprecated and will no longer work in the near future. If you are using a Homebrew or Linuxbrew Python, please see discussion at https://github.com/Homebrew/homebrew-core/issues/76621)�reason�replacement�gone_inrjrl)#rr�_USE_SYSCONFIGrr�pathlib�Pathr]r0�implementation�name�parent�
startswithrr=r
�version_inforr+rrYr^rKr@�tuplerarXrV�is_python_build�append�distutils_scheme�anyrrirr)rsrmrnrortrIrc�warning_contextsr8�old_v�new_v�skip_pypy_special_case�$skip_osx_framework_user_special_case�skip_bpo_44860�skip_linux_system_special_case�skip_sysconfig_abiflag_bug�skip_msys2_mingw_bug�skip_cpython_buildrQr)r�r*r�s���	
��
�
�
����������������������	�
rcCs<t��}tr|St��}tt�|�t�|�dd�r8t�|S)N�
bin_prefixrj)rrr�rrkr�r�rr�rcrbr)r)r*rdsrcCstjddd�jS)N�T)rm)rr�scriptsr)r)r)r*rosr)�valuer!cCst�s
dS|dkrdSdS)a�Check if the value is Debian's APT-controlled dist-packages.

    Debian's ``distutils.sysconfig.get_python_lib()`` implementation returns the
    default package path controlled by APT, but does not patch ``sysconfig`` to
    do the same. This is similar to the bug worked around in ``get_scheme()``,
    but here the default is ``deb_system`` instead of ``unix_local``. Ultimately
    we can't do anything about this Debian bug, and this detection allows us to
    skip the warning when needed.
    Fz/usr/lib/python3/dist-packagesT)r@)r�r)r)r*�$_looks_like_deb_system_dist_packagesss

r�cCsHt��}tr|St��}t|�r$|Stt�|�t�|�dd�rDt�|S)z,Return the default pure-Python lib location.r.rj)	rrr�rr�rkr�r�rrr�r)r)r*r�srcCsHt��}tr|St��}t|�r$|Stt�|�t�|�dd�rDt�|S)z0Return the default platform-shared lib location.r%rj)	rrr�rr�rkr�r�rrr�r)r)r*r�sr)�v1�v2r!cCs||kr|gS||gS)zDeduplicate values from a list.r))r�r�r)r)r*�
_deduplicated�sr�)rIr!cCszt�|�\}}trt||�St�|�\}}tt�|�t�|�dd�tt�|�t�|�dd�g}t|�rpt	|d�t||�S)z*Return the lib locations under ``prefix``.zprefixed-purelibrjzprefixed-platlib)rI)
rrr�r�rrkr�r�r�rr)rI�new_pure�new_plat�old_pure�old_plat�warnedr)r)r*r�s$
���
r)FNNFN)D�	functools�loggingrFr�r0rV�typingrrrrrr�pip._internal.models.schemerr	�pip._internal.utils.compatr
�pip._internal.utils.deprecationr�pip._internal.utils.virtualenvrr�rr�baserrrrrr�__all__�	getLogger�__name__rd�environ�get�DEBUGrf�WARNINGr]r�str�__annotations__r�r��boolr+r3�	lru_cacher=r@rKrXrar�rirkrrrrrr�rrr�rr)r)r)r*�<module>s�
  	









����(site-packages/pip/_internal/locations/base.py000064400000003053150564745010015316 0ustar00import functools
import os
import site
import sys
import sysconfig
import typing

from pip._internal.utils import appdirs
from pip._internal.utils.virtualenv import running_under_virtualenv

# Application Directories
USER_CACHE_DIR = appdirs.user_cache_dir("pip")

# FIXME doesn't account for venv linked to global site-packages
site_packages: typing.Optional[str] = sysconfig.get_path("purelib")


def get_major_minor_version() -> str:
    """
    Return the major-minor version of the current Python as a string, e.g.
    "3.7" or "3.10".
    """
    return "{}.{}".format(*sys.version_info)


def get_src_prefix() -> str:
    if running_under_virtualenv():
        src_prefix = os.path.join(sys.prefix, "src")
    else:
        # FIXME: keep src in cwd for now (it is not a temporary folder)
        try:
            src_prefix = os.path.join(os.getcwd(), "src")
        except OSError:
            # In case the current working directory has been renamed or deleted
            sys.exit("The folder you are executing pip from can no longer be found.")

    # under macOS + virtualenv sys.prefix is not properly resolved
    # it is something like /path/to/python/bin/..
    return os.path.abspath(src_prefix)


try:
    # Use getusersitepackages if this is present, as it ensures that the
    # value is initialised properly.
    user_site: typing.Optional[str] = site.getusersitepackages()
except AttributeError:
    user_site = site.USER_SITE


@functools.lru_cache(maxsize=None)
def is_osx_framework() -> bool:
    return bool(sysconfig.get_config_var("PYTHONFRAMEWORK"))
site-packages/pip/_internal/locations/__init__.py000064400000034154150564745010016151 0ustar00import functools
import logging
import os
import pathlib
import sys
import sysconfig
from typing import Any, Dict, Iterator, List, Optional, Tuple

from pip._internal.models.scheme import SCHEME_KEYS, Scheme
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.deprecation import deprecated
from pip._internal.utils.virtualenv import running_under_virtualenv

from . import _distutils, _sysconfig
from .base import (
    USER_CACHE_DIR,
    get_major_minor_version,
    get_src_prefix,
    is_osx_framework,
    site_packages,
    user_site,
)

__all__ = [
    "USER_CACHE_DIR",
    "get_bin_prefix",
    "get_bin_user",
    "get_major_minor_version",
    "get_platlib",
    "get_prefixed_libs",
    "get_purelib",
    "get_scheme",
    "get_src_prefix",
    "site_packages",
    "user_site",
]


logger = logging.getLogger(__name__)

if os.environ.get("_PIP_LOCATIONS_NO_WARN_ON_MISMATCH"):
    _MISMATCH_LEVEL = logging.DEBUG
else:
    _MISMATCH_LEVEL = logging.WARNING

_PLATLIBDIR: str = getattr(sys, "platlibdir", "lib")

_USE_SYSCONFIG = sys.version_info >= (3, 10)


def _looks_like_bpo_44860() -> bool:
    """The resolution to bpo-44860 will change this incorrect platlib.

    See <https://bugs.python.org/issue44860>.
    """
    from distutils.command.install import INSTALL_SCHEMES  # type: ignore

    try:
        unix_user_platlib = INSTALL_SCHEMES["unix_user"]["platlib"]
    except KeyError:
        return False
    return unix_user_platlib == "$usersite"


def _looks_like_red_hat_patched_platlib_purelib(scheme: Dict[str, str]) -> bool:
    platlib = scheme["platlib"]
    if "/$platlibdir/" in platlib and hasattr(sys, "platlibdir"):
        platlib = platlib.replace("/$platlibdir/", f"/{sys.platlibdir}/")
    if "/lib64/" not in platlib:
        return False
    unpatched = platlib.replace("/lib64/", "/lib/")
    return unpatched.replace("$platbase/", "$base/") == scheme["purelib"]


@functools.lru_cache(maxsize=None)
def _looks_like_red_hat_lib() -> bool:
    """Red Hat patches platlib in unix_prefix and unix_home, but not purelib.

    This is the only way I can see to tell a Red Hat-patched Python.
    """
    from distutils.command.install import INSTALL_SCHEMES  # type: ignore

    return all(
        k in INSTALL_SCHEMES
        and _looks_like_red_hat_patched_platlib_purelib(INSTALL_SCHEMES[k])
        for k in ("unix_prefix", "unix_home")
    )


@functools.lru_cache(maxsize=None)
def _looks_like_debian_scheme() -> bool:
    """Debian adds two additional schemes."""
    from distutils.command.install import INSTALL_SCHEMES  # type: ignore

    return "deb_system" in INSTALL_SCHEMES and "unix_local" in INSTALL_SCHEMES


@functools.lru_cache(maxsize=None)
def _looks_like_red_hat_scheme() -> bool:
    """Red Hat patches ``sys.prefix`` and ``sys.exec_prefix``.

    Red Hat's ``00251-change-user-install-location.patch`` changes the install
    command's ``prefix`` and ``exec_prefix`` to append ``"/local"``. This is
    (fortunately?) done quite unconditionally, so we create a default command
    object without any configuration to detect this.
    """
    from distutils.command.install import install
    from distutils.dist import Distribution

    cmd: Any = install(Distribution())
    cmd.finalize_options()
    return (
        cmd.exec_prefix == f"{os.path.normpath(sys.exec_prefix)}/local"
        and cmd.prefix == f"{os.path.normpath(sys.prefix)}/local"
    )


@functools.lru_cache(maxsize=None)
def _looks_like_msys2_mingw_scheme() -> bool:
    """MSYS2 patches distutils and sysconfig to use a UNIX-like scheme.

    However, MSYS2 incorrectly patches sysconfig ``nt`` scheme. The fix is
    likely going to be included in their 3.10 release, so we ignore the warning.
    See msys2/MINGW-packages#9319.

    MSYS2 MINGW's patch uses lowercase ``"lib"`` instead of the usual uppercase,
    and is missing the final ``"site-packages"``.
    """
    paths = sysconfig.get_paths("nt", expand=False)
    return all(
        "Lib" not in p and "lib" in p and not p.endswith("site-packages")
        for p in (paths[key] for key in ("platlib", "purelib"))
    )


def _fix_abiflags(parts: Tuple[str]) -> Iterator[str]:
    ldversion = sysconfig.get_config_var("LDVERSION")
    abiflags: str = getattr(sys, "abiflags", None)

    # LDVERSION does not end with sys.abiflags. Just return the path unchanged.
    if not ldversion or not abiflags or not ldversion.endswith(abiflags):
        yield from parts
        return

    # Strip sys.abiflags from LDVERSION-based path components.
    for part in parts:
        if part.endswith(ldversion):
            part = part[: (0 - len(abiflags))]
        yield part


@functools.lru_cache(maxsize=None)
def _warn_mismatched(old: pathlib.Path, new: pathlib.Path, *, key: str) -> None:
    issue_url = "https://github.com/pypa/pip/issues/10151"
    message = (
        "Value for %s does not match. Please report this to <%s>"
        "\ndistutils: %s"
        "\nsysconfig: %s"
    )
    logger.log(_MISMATCH_LEVEL, message, key, issue_url, old, new)


def _warn_if_mismatch(old: pathlib.Path, new: pathlib.Path, *, key: str) -> bool:
    if old == new:
        return False
    _warn_mismatched(old, new, key=key)
    return True


@functools.lru_cache(maxsize=None)
def _log_context(
    *,
    user: bool = False,
    home: Optional[str] = None,
    root: Optional[str] = None,
    prefix: Optional[str] = None,
) -> None:
    parts = [
        "Additional context:",
        "user = %r",
        "home = %r",
        "root = %r",
        "prefix = %r",
    ]

    logger.log(_MISMATCH_LEVEL, "\n".join(parts), user, home, root, prefix)


def get_scheme(
    dist_name: str,
    user: bool = False,
    home: Optional[str] = None,
    root: Optional[str] = None,
    isolated: bool = False,
    prefix: Optional[str] = None,
) -> Scheme:
    new = _sysconfig.get_scheme(
        dist_name,
        user=user,
        home=home,
        root=root,
        isolated=isolated,
        prefix=prefix,
    )
    if _USE_SYSCONFIG:
        return new

    old = _distutils.get_scheme(
        dist_name,
        user=user,
        home=home,
        root=root,
        isolated=isolated,
        prefix=prefix,
    )

    warning_contexts = []
    for k in SCHEME_KEYS:
        old_v = pathlib.Path(getattr(old, k))
        new_v = pathlib.Path(getattr(new, k))

        if old_v == new_v:
            continue

        # distutils incorrectly put PyPy packages under ``site-packages/python``
        # in the ``posix_home`` scheme, but PyPy devs said they expect the
        # directory name to be ``pypy`` instead. So we treat this as a bug fix
        # and not warn about it. See bpo-43307 and python/cpython#24628.
        skip_pypy_special_case = (
            sys.implementation.name == "pypy"
            and home is not None
            and k in ("platlib", "purelib")
            and old_v.parent == new_v.parent
            and old_v.name.startswith("python")
            and new_v.name.startswith("pypy")
        )
        if skip_pypy_special_case:
            continue

        # sysconfig's ``osx_framework_user`` does not include ``pythonX.Y`` in
        # the ``include`` value, but distutils's ``headers`` does. We'll let
        # CPython decide whether this is a bug or feature. See bpo-43948.
        skip_osx_framework_user_special_case = (
            user
            and is_osx_framework()
            and k == "headers"
            and old_v.parent.parent == new_v.parent
            and old_v.parent.name.startswith("python")
        )
        if skip_osx_framework_user_special_case:
            continue

        # On Red Hat and derived Linux distributions, distutils is patched to
        # use "lib64" instead of "lib" for platlib.
        if k == "platlib" and _looks_like_red_hat_lib():
            continue

        # On Python 3.9+, sysconfig's posix_user scheme sets platlib against
        # sys.platlibdir, but distutils's unix_user incorrectly coninutes
        # using the same $usersite for both platlib and purelib. This creates a
        # mismatch when sys.platlibdir is not "lib".
        skip_bpo_44860 = (
            user
            and k == "platlib"
            and not WINDOWS
            and sys.version_info >= (3, 9)
            and _PLATLIBDIR != "lib"
            and _looks_like_bpo_44860()
        )
        if skip_bpo_44860:
            continue

        # Both Debian and Red Hat patch Python to place the system site under
        # /usr/local instead of /usr. Debian also places lib in dist-packages
        # instead of site-packages, but the /usr/local check should cover it.
        skip_linux_system_special_case = (
            not (user or home or prefix or running_under_virtualenv())
            and old_v.parts[1:3] == ("usr", "local")
            and len(new_v.parts) > 1
            and new_v.parts[1] == "usr"
            and (len(new_v.parts) < 3 or new_v.parts[2] != "local")
            and (_looks_like_red_hat_scheme() or _looks_like_debian_scheme())
        )
        if skip_linux_system_special_case:
            continue

        # On Python 3.7 and earlier, sysconfig does not include sys.abiflags in
        # the "pythonX.Y" part of the path, but distutils does.
        skip_sysconfig_abiflag_bug = (
            sys.version_info < (3, 8)
            and not WINDOWS
            and k in ("headers", "platlib", "purelib")
            and tuple(_fix_abiflags(old_v.parts)) == new_v.parts
        )
        if skip_sysconfig_abiflag_bug:
            continue

        # MSYS2 MINGW's sysconfig patch does not include the "site-packages"
        # part of the path. This is incorrect and will be fixed in MSYS.
        skip_msys2_mingw_bug = (
            WINDOWS and k in ("platlib", "purelib") and _looks_like_msys2_mingw_scheme()
        )
        if skip_msys2_mingw_bug:
            continue

        # CPython's POSIX install script invokes pip (via ensurepip) against the
        # interpreter located in the source tree, not the install site. This
        # triggers special logic in sysconfig that's not present in distutils.
        # https://github.com/python/cpython/blob/8c21941ddaf/Lib/sysconfig.py#L178-L194
        skip_cpython_build = (
            sysconfig.is_python_build(check_home=True)
            and not WINDOWS
            and k in ("headers", "include", "platinclude")
        )
        if skip_cpython_build:
            continue

        warning_contexts.append((old_v, new_v, f"scheme.{k}"))

    if not warning_contexts:
        return old

    # Check if this path mismatch is caused by distutils config files. Those
    # files will no longer work once we switch to sysconfig, so this raises a
    # deprecation message for them.
    default_old = _distutils.distutils_scheme(
        dist_name,
        user,
        home,
        root,
        isolated,
        prefix,
        ignore_config_files=True,
    )
    if any(default_old[k] != getattr(old, k) for k in SCHEME_KEYS):
        deprecated(
            reason=(
                "Configuring installation scheme with distutils config files "
                "is deprecated and will no longer work in the near future. If you "
                "are using a Homebrew or Linuxbrew Python, please see discussion "
                "at https://github.com/Homebrew/homebrew-core/issues/76621"
            ),
            replacement=None,
            gone_in=None,
        )
        return old

    # Post warnings about this mismatch so user can report them back.
    for old_v, new_v, key in warning_contexts:
        _warn_mismatched(old_v, new_v, key=key)
    _log_context(user=user, home=home, root=root, prefix=prefix)

    return old


def get_bin_prefix() -> str:
    new = _sysconfig.get_bin_prefix()
    if _USE_SYSCONFIG:
        return new

    old = _distutils.get_bin_prefix()
    if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="bin_prefix"):
        _log_context()
    return old


def get_bin_user() -> str:
    return _sysconfig.get_scheme("", user=True).scripts


def _looks_like_deb_system_dist_packages(value: str) -> bool:
    """Check if the value is Debian's APT-controlled dist-packages.

    Debian's ``distutils.sysconfig.get_python_lib()`` implementation returns the
    default package path controlled by APT, but does not patch ``sysconfig`` to
    do the same. This is similar to the bug worked around in ``get_scheme()``,
    but here the default is ``deb_system`` instead of ``unix_local``. Ultimately
    we can't do anything about this Debian bug, and this detection allows us to
    skip the warning when needed.
    """
    if not _looks_like_debian_scheme():
        return False
    if value == "/usr/lib/python3/dist-packages":
        return True
    return False


def get_purelib() -> str:
    """Return the default pure-Python lib location."""
    new = _sysconfig.get_purelib()
    if _USE_SYSCONFIG:
        return new

    old = _distutils.get_purelib()
    if _looks_like_deb_system_dist_packages(old):
        return old
    if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="purelib"):
        _log_context()
    return old


def get_platlib() -> str:
    """Return the default platform-shared lib location."""
    new = _sysconfig.get_platlib()
    if _USE_SYSCONFIG:
        return new

    old = _distutils.get_platlib()
    if _looks_like_deb_system_dist_packages(old):
        return old
    if _warn_if_mismatch(pathlib.Path(old), pathlib.Path(new), key="platlib"):
        _log_context()
    return old


def _deduplicated(v1: str, v2: str) -> List[str]:
    """Deduplicate values from a list."""
    if v1 == v2:
        return [v1]
    return [v1, v2]


def get_prefixed_libs(prefix: str) -> List[str]:
    """Return the lib locations under ``prefix``."""
    new_pure, new_plat = _sysconfig.get_prefixed_libs(prefix)
    if _USE_SYSCONFIG:
        return _deduplicated(new_pure, new_plat)

    old_pure, old_plat = _distutils.get_prefixed_libs(prefix)

    warned = [
        _warn_if_mismatch(
            pathlib.Path(old_pure),
            pathlib.Path(new_pure),
            key="prefixed-purelib",
        ),
        _warn_if_mismatch(
            pathlib.Path(old_plat),
            pathlib.Path(new_plat),
            key="prefixed-platlib",
        ),
    ]
    if any(warned):
        _log_context(prefix=prefix)

    return _deduplicated(old_pure, old_plat)
site-packages/pip/_internal/locations/_distutils.py000064400000013357150564745010016577 0ustar00"""Locations where we look for configs, install stuff, etc"""

# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import logging
import os
import sys
from distutils.cmd import Command as DistutilsCommand
from distutils.command.install import SCHEME_KEYS
from distutils.command.install import install as distutils_install_command
from distutils.sysconfig import get_python_lib
from typing import Dict, List, Optional, Tuple, Union, cast

from pip._internal.models.scheme import Scheme
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.virtualenv import running_under_virtualenv

from .base import get_major_minor_version

logger = logging.getLogger(__name__)


def distutils_scheme(
    dist_name: str,
    user: bool = False,
    home: str = None,
    root: str = None,
    isolated: bool = False,
    prefix: str = None,
    *,
    ignore_config_files: bool = False,
) -> Dict[str, str]:
    """
    Return a distutils install scheme
    """
    from distutils.dist import Distribution

    dist_args: Dict[str, Union[str, List[str]]] = {"name": dist_name}
    if isolated:
        dist_args["script_args"] = ["--no-user-cfg"]

    d = Distribution(dist_args)
    if not ignore_config_files:
        try:
            d.parse_config_files()
        except UnicodeDecodeError:
            # Typeshed does not include find_config_files() for some reason.
            paths = d.find_config_files()  # type: ignore
            logger.warning(
                "Ignore distutils configs in %s due to encoding errors.",
                ", ".join(os.path.basename(p) for p in paths),
            )
    obj: Optional[DistutilsCommand] = None
    obj = d.get_command_obj("install", create=True)
    assert obj is not None
    i = cast(distutils_install_command, obj)
    # NOTE: setting user or home has the side-effect of creating the home dir
    # or user base for installations during finalize_options()
    # ideally, we'd prefer a scheme class that has no side-effects.
    assert not (user and prefix), f"user={user} prefix={prefix}"
    assert not (home and prefix), f"home={home} prefix={prefix}"
    i.user = user or i.user
    if user or home:
        i.prefix = ""
    i.prefix = prefix or i.prefix
    i.home = home or i.home
    i.root = root or i.root
    i.finalize_options()

    scheme = {}
    for key in SCHEME_KEYS:
        scheme[key] = getattr(i, "install_" + key)

    # install_lib specified in setup.cfg should install *everything*
    # into there (i.e. it takes precedence over both purelib and
    # platlib).  Note, i.install_lib is *always* set after
    # finalize_options(); we only want to override here if the user
    # has explicitly requested it hence going back to the config
    if "install_lib" in d.get_option_dict("install"):
        scheme.update(dict(purelib=i.install_lib, platlib=i.install_lib))

    if running_under_virtualenv():
        if home:
            prefix = home
        elif user:
            prefix = i.install_userbase  # type: ignore
        else:
            prefix = i.prefix
        scheme["headers"] = os.path.join(
            prefix,
            "include",
            "site",
            f"python{get_major_minor_version()}",
            dist_name,
        )

        if root is not None:
            path_no_drive = os.path.splitdrive(os.path.abspath(scheme["headers"]))[1]
            scheme["headers"] = os.path.join(root, path_no_drive[1:])

    return scheme


def get_scheme(
    dist_name: str,
    user: bool = False,
    home: Optional[str] = None,
    root: Optional[str] = None,
    isolated: bool = False,
    prefix: Optional[str] = None,
) -> Scheme:
    """
    Get the "scheme" corresponding to the input parameters. The distutils
    documentation provides the context for the available schemes:
    https://docs.python.org/3/install/index.html#alternate-installation

    :param dist_name: the name of the package to retrieve the scheme for, used
        in the headers scheme path
    :param user: indicates to use the "user" scheme
    :param home: indicates to use the "home" scheme and provides the base
        directory for the same
    :param root: root under which other directories are re-based
    :param isolated: equivalent to --no-user-cfg, i.e. do not consider
        ~/.pydistutils.cfg (posix) or ~/pydistutils.cfg (non-posix) for
        scheme paths
    :param prefix: indicates to use the "prefix" scheme and provides the
        base directory for the same
    """
    scheme = distutils_scheme(dist_name, user, home, root, isolated, prefix)
    return Scheme(
        platlib=scheme["platlib"],
        purelib=scheme["purelib"],
        headers=scheme["headers"],
        scripts=scheme["scripts"],
        data=scheme["data"],
    )


def get_bin_prefix() -> str:
    # XXX: In old virtualenv versions, sys.prefix can contain '..' components,
    # so we need to call normpath to eliminate them.
    prefix = os.path.normpath(sys.prefix)
    if WINDOWS:
        bin_py = os.path.join(prefix, "Scripts")
        # buildout uses 'bin' on Windows too?
        if not os.path.exists(bin_py):
            bin_py = os.path.join(prefix, "bin")
        return bin_py
    # Forcing to use /usr/local/bin for standard macOS framework installs
    # Also log to ~/Library/Logs/ for use with the Console.app log viewer
    if sys.platform[:6] == "darwin" and prefix[:16] == "/System/Library/":
        return "/usr/local/bin"
    return os.path.join(prefix, "bin")


def get_purelib() -> str:
    return get_python_lib(plat_specific=False)


def get_platlib() -> str:
    return get_python_lib(plat_specific=True)


def get_prefixed_libs(prefix: str) -> Tuple[str, str]:
    return (
        get_python_lib(plat_specific=False, prefix=prefix),
        get_python_lib(plat_specific=True, prefix=prefix),
    )
site-packages/pip/_internal/wheel_builder.py000064400000027727150564745010015241 0ustar00"""Orchestrator for building wheels from InstallRequirements.
"""

import logging
import os.path
import re
import shutil
from typing import Any, Callable, Iterable, List, Optional, Tuple

from pip._vendor.packaging.utils import canonicalize_name, canonicalize_version
from pip._vendor.packaging.version import InvalidVersion, Version

from pip._internal.cache import WheelCache
from pip._internal.exceptions import InvalidWheelFilename, UnsupportedWheel
from pip._internal.metadata import FilesystemWheel, get_wheel_distribution
from pip._internal.models.link import Link
from pip._internal.models.wheel import Wheel
from pip._internal.operations.build.wheel import build_wheel_pep517
from pip._internal.operations.build.wheel_editable import build_wheel_editable
from pip._internal.operations.build.wheel_legacy import build_wheel_legacy
from pip._internal.req.req_install import InstallRequirement
from pip._internal.utils.logging import indent_log
from pip._internal.utils.misc import ensure_dir, hash_file, is_wheel_installed
from pip._internal.utils.setuptools_build import make_setuptools_clean_args
from pip._internal.utils.subprocess import call_subprocess
from pip._internal.utils.temp_dir import TempDirectory
from pip._internal.utils.urls import path_to_url
from pip._internal.vcs import vcs

logger = logging.getLogger(__name__)

_egg_info_re = re.compile(r"([a-z0-9_.]+)-([a-z0-9_.!+-]+)", re.IGNORECASE)

BinaryAllowedPredicate = Callable[[InstallRequirement], bool]
BuildResult = Tuple[List[InstallRequirement], List[InstallRequirement]]


def _contains_egg_info(s: str) -> bool:
    """Determine whether the string looks like an egg_info.

    :param s: The string to parse. E.g. foo-2.1
    """
    return bool(_egg_info_re.search(s))


def _should_build(
    req: InstallRequirement,
    need_wheel: bool,
    check_binary_allowed: BinaryAllowedPredicate,
) -> bool:
    """Return whether an InstallRequirement should be built into a wheel."""
    if req.constraint:
        # never build requirements that are merely constraints
        return False
    if req.is_wheel:
        if need_wheel:
            logger.info(
                "Skipping %s, due to already being wheel.",
                req.name,
            )
        return False

    if need_wheel:
        # i.e. pip wheel, not pip install
        return True

    # From this point, this concerns the pip install command only
    # (need_wheel=False).

    if not req.source_dir:
        return False

    if req.editable:
        # we only build PEP 660 editable requirements
        return req.supports_pyproject_editable()

    if req.use_pep517:
        return True

    if not check_binary_allowed(req):
        logger.info(
            "Skipping wheel build for %s, due to binaries being disabled for it.",
            req.name,
        )
        return False

    if not is_wheel_installed():
        # we don't build legacy requirements if wheel is not installed
        logger.info(
            "Using legacy 'setup.py install' for %s, "
            "since package 'wheel' is not installed.",
            req.name,
        )
        return False

    return True


def should_build_for_wheel_command(
    req: InstallRequirement,
) -> bool:
    return _should_build(req, need_wheel=True, check_binary_allowed=_always_true)


def should_build_for_install_command(
    req: InstallRequirement,
    check_binary_allowed: BinaryAllowedPredicate,
) -> bool:
    return _should_build(
        req, need_wheel=False, check_binary_allowed=check_binary_allowed
    )


def _should_cache(
    req: InstallRequirement,
) -> Optional[bool]:
    """
    Return whether a built InstallRequirement can be stored in the persistent
    wheel cache, assuming the wheel cache is available, and _should_build()
    has determined a wheel needs to be built.
    """
    if req.editable or not req.source_dir:
        # never cache editable requirements
        return False

    if req.link and req.link.is_vcs:
        # VCS checkout. Do not cache
        # unless it points to an immutable commit hash.
        assert not req.editable
        assert req.source_dir
        vcs_backend = vcs.get_backend_for_scheme(req.link.scheme)
        assert vcs_backend
        if vcs_backend.is_immutable_rev_checkout(req.link.url, req.source_dir):
            return True
        return False

    assert req.link
    base, ext = req.link.splitext()
    if _contains_egg_info(base):
        return True

    # Otherwise, do not cache.
    return False


def _get_cache_dir(
    req: InstallRequirement,
    wheel_cache: WheelCache,
) -> str:
    """Return the persistent or temporary cache directory where the built
    wheel need to be stored.
    """
    cache_available = bool(wheel_cache.cache_dir)
    assert req.link
    if cache_available and _should_cache(req):
        cache_dir = wheel_cache.get_path_for_link(req.link)
    else:
        cache_dir = wheel_cache.get_ephem_path_for_link(req.link)
    return cache_dir


def _always_true(_: Any) -> bool:
    return True


def _verify_one(req: InstallRequirement, wheel_path: str) -> None:
    canonical_name = canonicalize_name(req.name or "")
    w = Wheel(os.path.basename(wheel_path))
    if canonicalize_name(w.name) != canonical_name:
        raise InvalidWheelFilename(
            "Wheel has unexpected file name: expected {!r}, "
            "got {!r}".format(canonical_name, w.name),
        )
    dist = get_wheel_distribution(FilesystemWheel(wheel_path), canonical_name)
    dist_verstr = str(dist.version)
    if canonicalize_version(dist_verstr) != canonicalize_version(w.version):
        raise InvalidWheelFilename(
            "Wheel has unexpected file name: expected {!r}, "
            "got {!r}".format(dist_verstr, w.version),
        )
    metadata_version_value = dist.metadata_version
    if metadata_version_value is None:
        raise UnsupportedWheel("Missing Metadata-Version")
    try:
        metadata_version = Version(metadata_version_value)
    except InvalidVersion:
        msg = f"Invalid Metadata-Version: {metadata_version_value}"
        raise UnsupportedWheel(msg)
    if metadata_version >= Version("1.2") and not isinstance(dist.version, Version):
        raise UnsupportedWheel(
            "Metadata 1.2 mandates PEP 440 version, "
            "but {!r} is not".format(dist_verstr)
        )


def _build_one(
    req: InstallRequirement,
    output_dir: str,
    verify: bool,
    build_options: List[str],
    global_options: List[str],
    editable: bool,
) -> Optional[str]:
    """Build one wheel.

    :return: The filename of the built wheel, or None if the build failed.
    """
    artifact = "editable" if editable else "wheel"
    try:
        ensure_dir(output_dir)
    except OSError as e:
        logger.warning(
            "Building %s for %s failed: %s",
            artifact,
            req.name,
            e,
        )
        return None

    # Install build deps into temporary directory (PEP 518)
    with req.build_env:
        wheel_path = _build_one_inside_env(
            req, output_dir, build_options, global_options, editable
        )
    if wheel_path and verify:
        try:
            _verify_one(req, wheel_path)
        except (InvalidWheelFilename, UnsupportedWheel) as e:
            logger.warning("Built %s for %s is invalid: %s", artifact, req.name, e)
            return None
    return wheel_path


def _build_one_inside_env(
    req: InstallRequirement,
    output_dir: str,
    build_options: List[str],
    global_options: List[str],
    editable: bool,
) -> Optional[str]:
    with TempDirectory(kind="wheel") as temp_dir:
        assert req.name
        if req.use_pep517:
            assert req.metadata_directory
            assert req.pep517_backend
            if global_options:
                logger.warning(
                    "Ignoring --global-option when building %s using PEP 517", req.name
                )
            if build_options:
                logger.warning(
                    "Ignoring --build-option when building %s using PEP 517", req.name
                )
            if editable:
                wheel_path = build_wheel_editable(
                    name=req.name,
                    backend=req.pep517_backend,
                    metadata_directory=req.metadata_directory,
                    tempd=temp_dir.path,
                )
            else:
                wheel_path = build_wheel_pep517(
                    name=req.name,
                    backend=req.pep517_backend,
                    metadata_directory=req.metadata_directory,
                    tempd=temp_dir.path,
                )
        else:
            wheel_path = build_wheel_legacy(
                name=req.name,
                setup_py_path=req.setup_py_path,
                source_dir=req.unpacked_source_directory,
                global_options=global_options,
                build_options=build_options,
                tempd=temp_dir.path,
            )

        if wheel_path is not None:
            wheel_name = os.path.basename(wheel_path)
            dest_path = os.path.join(output_dir, wheel_name)
            try:
                wheel_hash, length = hash_file(wheel_path)
                shutil.move(wheel_path, dest_path)
                logger.info(
                    "Created wheel for %s: filename=%s size=%d sha256=%s",
                    req.name,
                    wheel_name,
                    length,
                    wheel_hash.hexdigest(),
                )
                logger.info("Stored in directory: %s", output_dir)
                return dest_path
            except Exception as e:
                logger.warning(
                    "Building wheel for %s failed: %s",
                    req.name,
                    e,
                )
        # Ignore return, we can't do anything else useful.
        if not req.use_pep517:
            _clean_one_legacy(req, global_options)
        return None


def _clean_one_legacy(req: InstallRequirement, global_options: List[str]) -> bool:
    clean_args = make_setuptools_clean_args(
        req.setup_py_path,
        global_options=global_options,
    )

    logger.info("Running setup.py clean for %s", req.name)
    try:
        call_subprocess(clean_args, cwd=req.source_dir)
        return True
    except Exception:
        logger.error("Failed cleaning build dir for %s", req.name)
        return False


def build(
    requirements: Iterable[InstallRequirement],
    wheel_cache: WheelCache,
    verify: bool,
    build_options: List[str],
    global_options: List[str],
) -> BuildResult:
    """Build wheels.

    :return: The list of InstallRequirement that succeeded to build and
        the list of InstallRequirement that failed to build.
    """
    if not requirements:
        return [], []

    # Build the wheels.
    logger.info(
        "Building wheels for collected packages: %s",
        ", ".join(req.name for req in requirements),  # type: ignore
    )

    with indent_log():
        build_successes, build_failures = [], []
        for req in requirements:
            assert req.name
            cache_dir = _get_cache_dir(req, wheel_cache)
            wheel_file = _build_one(
                req,
                cache_dir,
                verify,
                build_options,
                global_options,
                req.editable and req.permit_editable_wheels,
            )
            if wheel_file:
                # Update the link for this.
                req.link = Link(path_to_url(wheel_file))
                req.local_file_path = req.link.file_path
                assert req.link.is_wheel
                build_successes.append(req)
            else:
                build_failures.append(req)

    # notify success/failure
    if build_successes:
        logger.info(
            "Successfully built %s",
            " ".join([req.name for req in build_successes]),  # type: ignore
        )
    if build_failures:
        logger.info(
            "Failed to build %s",
            " ".join([req.name for req in build_failures]),  # type: ignore
        )
    # Return a list of requirements that failed to build
    return build_successes, build_failures
site-packages/pip/_internal/cache.py000064400000022341150564745010013455 0ustar00"""Cache Management
"""

import hashlib
import json
import logging
import os
from typing import Any, Dict, List, Optional, Set

from pip._vendor.packaging.tags import Tag, interpreter_name, interpreter_version
from pip._vendor.packaging.utils import canonicalize_name

from pip._internal.exceptions import InvalidWheelFilename
from pip._internal.models.format_control import FormatControl
from pip._internal.models.link import Link
from pip._internal.models.wheel import Wheel
from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds
from pip._internal.utils.urls import path_to_url

logger = logging.getLogger(__name__)


def _hash_dict(d: Dict[str, str]) -> str:
    """Return a stable sha224 of a dictionary."""
    s = json.dumps(d, sort_keys=True, separators=(",", ":"), ensure_ascii=True)
    return hashlib.sha224(s.encode("ascii")).hexdigest()


class Cache:
    """An abstract class - provides cache directories for data from links


    :param cache_dir: The root of the cache.
    :param format_control: An object of FormatControl class to limit
        binaries being read from the cache.
    :param allowed_formats: which formats of files the cache should store.
        ('binary' and 'source' are the only allowed values)
    """

    def __init__(
        self, cache_dir: str, format_control: FormatControl, allowed_formats: Set[str]
    ) -> None:
        super().__init__()
        assert not cache_dir or os.path.isabs(cache_dir)
        self.cache_dir = cache_dir or None
        self.format_control = format_control
        self.allowed_formats = allowed_formats

        _valid_formats = {"source", "binary"}
        assert self.allowed_formats.union(_valid_formats) == _valid_formats

    def _get_cache_path_parts(self, link: Link) -> List[str]:
        """Get parts of part that must be os.path.joined with cache_dir"""

        # We want to generate an url to use as our cache key, we don't want to
        # just re-use the URL because it might have other items in the fragment
        # and we don't care about those.
        key_parts = {"url": link.url_without_fragment}
        if link.hash_name is not None and link.hash is not None:
            key_parts[link.hash_name] = link.hash
        if link.subdirectory_fragment:
            key_parts["subdirectory"] = link.subdirectory_fragment

        # Include interpreter name, major and minor version in cache key
        # to cope with ill-behaved sdists that build a different wheel
        # depending on the python version their setup.py is being run on,
        # and don't encode the difference in compatibility tags.
        # https://github.com/pypa/pip/issues/7296
        key_parts["interpreter_name"] = interpreter_name()
        key_parts["interpreter_version"] = interpreter_version()

        # Encode our key url with sha224, we'll use this because it has similar
        # security properties to sha256, but with a shorter total output (and
        # thus less secure). However the differences don't make a lot of
        # difference for our use case here.
        hashed = _hash_dict(key_parts)

        # We want to nest the directories some to prevent having a ton of top
        # level directories where we might run out of sub directories on some
        # FS.
        parts = [hashed[:2], hashed[2:4], hashed[4:6], hashed[6:]]

        return parts

    def _get_candidates(self, link: Link, canonical_package_name: str) -> List[Any]:
        can_not_cache = not self.cache_dir or not canonical_package_name or not link
        if can_not_cache:
            return []

        formats = self.format_control.get_allowed_formats(canonical_package_name)
        if not self.allowed_formats.intersection(formats):
            return []

        candidates = []
        path = self.get_path_for_link(link)
        if os.path.isdir(path):
            for candidate in os.listdir(path):
                candidates.append((candidate, path))
        return candidates

    def get_path_for_link(self, link: Link) -> str:
        """Return a directory to store cached items in for link."""
        raise NotImplementedError()

    def get(
        self,
        link: Link,
        package_name: Optional[str],
        supported_tags: List[Tag],
    ) -> Link:
        """Returns a link to a cached item if it exists, otherwise returns the
        passed link.
        """
        raise NotImplementedError()


class SimpleWheelCache(Cache):
    """A cache of wheels for future installs."""

    def __init__(self, cache_dir: str, format_control: FormatControl) -> None:
        super().__init__(cache_dir, format_control, {"binary"})

    def get_path_for_link(self, link: Link) -> str:
        """Return a directory to store cached wheels for link

        Because there are M wheels for any one sdist, we provide a directory
        to cache them in, and then consult that directory when looking up
        cache hits.

        We only insert things into the cache if they have plausible version
        numbers, so that we don't contaminate the cache with things that were
        not unique. E.g. ./package might have dozens of installs done for it
        and build a version of 0.0...and if we built and cached a wheel, we'd
        end up using the same wheel even if the source has been edited.

        :param link: The link of the sdist for which this will cache wheels.
        """
        parts = self._get_cache_path_parts(link)
        assert self.cache_dir
        # Store wheels within the root cache_dir
        return os.path.join(self.cache_dir, "wheels", *parts)

    def get(
        self,
        link: Link,
        package_name: Optional[str],
        supported_tags: List[Tag],
    ) -> Link:
        candidates = []

        if not package_name:
            return link

        canonical_package_name = canonicalize_name(package_name)
        for wheel_name, wheel_dir in self._get_candidates(link, canonical_package_name):
            try:
                wheel = Wheel(wheel_name)
            except InvalidWheelFilename:
                continue
            if canonicalize_name(wheel.name) != canonical_package_name:
                logger.debug(
                    "Ignoring cached wheel %s for %s as it "
                    "does not match the expected distribution name %s.",
                    wheel_name,
                    link,
                    package_name,
                )
                continue
            if not wheel.supported(supported_tags):
                # Built for a different python/arch/etc
                continue
            candidates.append(
                (
                    wheel.support_index_min(supported_tags),
                    wheel_name,
                    wheel_dir,
                )
            )

        if not candidates:
            return link

        _, wheel_name, wheel_dir = min(candidates)
        return Link(path_to_url(os.path.join(wheel_dir, wheel_name)))


class EphemWheelCache(SimpleWheelCache):
    """A SimpleWheelCache that creates it's own temporary cache directory"""

    def __init__(self, format_control: FormatControl) -> None:
        self._temp_dir = TempDirectory(
            kind=tempdir_kinds.EPHEM_WHEEL_CACHE,
            globally_managed=True,
        )

        super().__init__(self._temp_dir.path, format_control)


class CacheEntry:
    def __init__(
        self,
        link: Link,
        persistent: bool,
    ):
        self.link = link
        self.persistent = persistent


class WheelCache(Cache):
    """Wraps EphemWheelCache and SimpleWheelCache into a single Cache

    This Cache allows for gracefully degradation, using the ephem wheel cache
    when a certain link is not found in the simple wheel cache first.
    """

    def __init__(self, cache_dir: str, format_control: FormatControl) -> None:
        super().__init__(cache_dir, format_control, {"binary"})
        self._wheel_cache = SimpleWheelCache(cache_dir, format_control)
        self._ephem_cache = EphemWheelCache(format_control)

    def get_path_for_link(self, link: Link) -> str:
        return self._wheel_cache.get_path_for_link(link)

    def get_ephem_path_for_link(self, link: Link) -> str:
        return self._ephem_cache.get_path_for_link(link)

    def get(
        self,
        link: Link,
        package_name: Optional[str],
        supported_tags: List[Tag],
    ) -> Link:
        cache_entry = self.get_cache_entry(link, package_name, supported_tags)
        if cache_entry is None:
            return link
        return cache_entry.link

    def get_cache_entry(
        self,
        link: Link,
        package_name: Optional[str],
        supported_tags: List[Tag],
    ) -> Optional[CacheEntry]:
        """Returns a CacheEntry with a link to a cached item if it exists or
        None. The cache entry indicates if the item was found in the persistent
        or ephemeral cache.
        """
        retval = self._wheel_cache.get(
            link=link,
            package_name=package_name,
            supported_tags=supported_tags,
        )
        if retval is not link:
            return CacheEntry(retval, persistent=True)

        retval = self._ephem_cache.get(
            link=link,
            package_name=package_name,
            supported_tags=supported_tags,
        )
        if retval is not link:
            return CacheEntry(retval, persistent=False)

        return None
site-packages/pip/_internal/utils/datetime.py000064400000000362150564745010015345 0ustar00"""For when pip wants to check the date or time.
"""

import datetime


def today_is_later_than(year: int, month: int, day: int) -> bool:
    today = datetime.date.today()
    given = datetime.date(year, month, day)

    return today > given
site-packages/pip/_internal/utils/urls.py000064400000003337150564745010014543 0ustar00import os
import string
import urllib.parse
import urllib.request
from typing import Optional

from .compat import WINDOWS


def get_url_scheme(url: str) -> Optional[str]:
    if ":" not in url:
        return None
    return url.split(":", 1)[0].lower()


def path_to_url(path: str) -> str:
    """
    Convert a path to a file: URL.  The path will be made absolute and have
    quoted path parts.
    """
    path = os.path.normpath(os.path.abspath(path))
    url = urllib.parse.urljoin("file:", urllib.request.pathname2url(path))
    return url


def url_to_path(url: str) -> str:
    """
    Convert a file: URL to a path.
    """
    assert url.startswith(
        "file:"
    ), f"You can only turn file: urls into filenames (not {url!r})"

    _, netloc, path, _, _ = urllib.parse.urlsplit(url)

    if not netloc or netloc == "localhost":
        # According to RFC 8089, same as empty authority.
        netloc = ""
    elif WINDOWS:
        # If we have a UNC path, prepend UNC share notation.
        netloc = "\\\\" + netloc
    else:
        raise ValueError(
            f"non-local file URIs are not supported on this platform: {url!r}"
        )

    path = urllib.request.url2pathname(netloc + path)

    # On Windows, urlsplit parses the path as something like "/C:/Users/foo".
    # This creates issues for path-related functions like io.open(), so we try
    # to detect and strip the leading slash.
    if (
        WINDOWS
        and not netloc  # Not UNC.
        and len(path) >= 3
        and path[0] == "/"  # Leading slash to strip.
        and path[1] in string.ascii_letters  # Drive letter.
        and path[2:4] in (":", ":/")  # Colon + end of string, or colon + absolute path.
    ):
        path = path[1:]

    return path
site-packages/pip/_internal/utils/entrypoints.py000064400000002152150564745010016146 0ustar00import sys
from typing import List, Optional

from pip._internal.cli.main import main


def _wrapper(args: Optional[List[str]] = None, _nowarn: bool = False) -> int:
    """Central wrapper for all old entrypoints.

    Historically pip has had several entrypoints defined. Because of issues
    arising from PATH, sys.path, multiple Pythons, their interactions, and most
    of them having a pip installed, users suffer every time an entrypoint gets
    moved.

    To alleviate this pain, and provide a mechanism for warning users and
    directing them to an appropriate place for help, we now define all of
    our old entrypoints as wrappers for the current one.
    """
    if not _nowarn:
        sys.stderr.write(
            "WARNING: pip is being invoked by an old script wrapper. This will "
            "fail in a future version of pip.\n"
            "Please see https://github.com/pypa/pip/issues/5599 for advice on "
            "fixing the underlying issue.\n"
            "To avoid this problem you can invoke Python with '-m pip' instead of "
            "running pip directly.\n"
        )
    return main(args)
site-packages/pip/_internal/utils/hashes.py000064400000011313150564745010015022 0ustar00import hashlib
from typing import TYPE_CHECKING, BinaryIO, Dict, Iterator, List

from pip._internal.exceptions import HashMismatch, HashMissing, InstallationError
from pip._internal.utils.misc import read_chunks

if TYPE_CHECKING:
    from hashlib import _Hash

    # NoReturn introduced in 3.6.2; imported only for type checking to maintain
    # pip compatibility with older patch versions of Python 3.6
    from typing import NoReturn


# The recommended hash algo of the moment. Change this whenever the state of
# the art changes; it won't hurt backward compatibility.
FAVORITE_HASH = "sha256"


# Names of hashlib algorithms allowed by the --hash option and ``pip hash``
# Currently, those are the ones at least as collision-resistant as sha256.
STRONG_HASHES = ["sha256", "sha384", "sha512"]


class Hashes:
    """A wrapper that builds multiple hashes at once and checks them against
    known-good values

    """

    def __init__(self, hashes: Dict[str, List[str]] = None) -> None:
        """
        :param hashes: A dict of algorithm names pointing to lists of allowed
            hex digests
        """
        allowed = {}
        if hashes is not None:
            for alg, keys in hashes.items():
                # Make sure values are always sorted (to ease equality checks)
                allowed[alg] = sorted(keys)
        self._allowed = allowed

    def __and__(self, other: "Hashes") -> "Hashes":
        if not isinstance(other, Hashes):
            return NotImplemented

        # If either of the Hashes object is entirely empty (i.e. no hash
        # specified at all), all hashes from the other object are allowed.
        if not other:
            return self
        if not self:
            return other

        # Otherwise only hashes that present in both objects are allowed.
        new = {}
        for alg, values in other._allowed.items():
            if alg not in self._allowed:
                continue
            new[alg] = [v for v in values if v in self._allowed[alg]]
        return Hashes(new)

    @property
    def digest_count(self) -> int:
        return sum(len(digests) for digests in self._allowed.values())

    def is_hash_allowed(self, hash_name: str, hex_digest: str) -> bool:
        """Return whether the given hex digest is allowed."""
        return hex_digest in self._allowed.get(hash_name, [])

    def check_against_chunks(self, chunks: Iterator[bytes]) -> None:
        """Check good hashes against ones built from iterable of chunks of
        data.

        Raise HashMismatch if none match.

        """
        gots = {}
        for hash_name in self._allowed.keys():
            try:
                gots[hash_name] = hashlib.new(hash_name)
            except (ValueError, TypeError):
                raise InstallationError(f"Unknown hash name: {hash_name}")

        for chunk in chunks:
            for hash in gots.values():
                hash.update(chunk)

        for hash_name, got in gots.items():
            if got.hexdigest() in self._allowed[hash_name]:
                return
        self._raise(gots)

    def _raise(self, gots: Dict[str, "_Hash"]) -> "NoReturn":
        raise HashMismatch(self._allowed, gots)

    def check_against_file(self, file: BinaryIO) -> None:
        """Check good hashes against a file-like object

        Raise HashMismatch if none match.

        """
        return self.check_against_chunks(read_chunks(file))

    def check_against_path(self, path: str) -> None:
        with open(path, "rb") as file:
            return self.check_against_file(file)

    def __bool__(self) -> bool:
        """Return whether I know any known-good hashes."""
        return bool(self._allowed)

    def __eq__(self, other: object) -> bool:
        if not isinstance(other, Hashes):
            return NotImplemented
        return self._allowed == other._allowed

    def __hash__(self) -> int:
        return hash(
            ",".join(
                sorted(
                    ":".join((alg, digest))
                    for alg, digest_list in self._allowed.items()
                    for digest in digest_list
                )
            )
        )


class MissingHashes(Hashes):
    """A workalike for Hashes used when we're missing a hash for a requirement

    It computes the actual hash of the requirement and raises a HashMissing
    exception showing it to the user.

    """

    def __init__(self) -> None:
        """Don't offer the ``hashes`` kwarg."""
        # Pass our favorite hash in to generate a "gotten hash". With the
        # empty list, it will never match, so an error will always raise.
        super().__init__(hashes={FAVORITE_HASH: []})

    def _raise(self, gots: Dict[str, "_Hash"]) -> "NoReturn":
        raise HashMissing(gots[FAVORITE_HASH].hexdigest())
site-packages/pip/_internal/utils/subprocess.py000064400000023512150564745010015743 0ustar00import logging
import os
import shlex
import subprocess
from typing import (
    TYPE_CHECKING,
    Any,
    Callable,
    Iterable,
    List,
    Mapping,
    Optional,
    Union,
)

from pip._internal.cli.spinners import SpinnerInterface, open_spinner
from pip._internal.exceptions import InstallationSubprocessError
from pip._internal.utils.logging import VERBOSE, subprocess_logger
from pip._internal.utils.misc import HiddenText

if TYPE_CHECKING:
    # Literal was introduced in Python 3.8.
    #
    # TODO: Remove `if TYPE_CHECKING` when dropping support for Python 3.7.
    from typing import Literal

CommandArgs = List[Union[str, HiddenText]]


LOG_DIVIDER = "----------------------------------------"


def make_command(*args: Union[str, HiddenText, CommandArgs]) -> CommandArgs:
    """
    Create a CommandArgs object.
    """
    command_args: CommandArgs = []
    for arg in args:
        # Check for list instead of CommandArgs since CommandArgs is
        # only known during type-checking.
        if isinstance(arg, list):
            command_args.extend(arg)
        else:
            # Otherwise, arg is str or HiddenText.
            command_args.append(arg)

    return command_args


def format_command_args(args: Union[List[str], CommandArgs]) -> str:
    """
    Format command arguments for display.
    """
    # For HiddenText arguments, display the redacted form by calling str().
    # Also, we don't apply str() to arguments that aren't HiddenText since
    # this can trigger a UnicodeDecodeError in Python 2 if the argument
    # has type unicode and includes a non-ascii character.  (The type
    # checker doesn't ensure the annotations are correct in all cases.)
    return " ".join(
        shlex.quote(str(arg)) if isinstance(arg, HiddenText) else shlex.quote(arg)
        for arg in args
    )


def reveal_command_args(args: Union[List[str], CommandArgs]) -> List[str]:
    """
    Return the arguments in their raw, unredacted form.
    """
    return [arg.secret if isinstance(arg, HiddenText) else arg for arg in args]


def make_subprocess_output_error(
    cmd_args: Union[List[str], CommandArgs],
    cwd: Optional[str],
    lines: List[str],
    exit_status: int,
) -> str:
    """
    Create and return the error message to use to log a subprocess error
    with command output.

    :param lines: A list of lines, each ending with a newline.
    """
    command = format_command_args(cmd_args)

    # We know the joined output value ends in a newline.
    output = "".join(lines)
    msg = (
        # Use a unicode string to avoid "UnicodeEncodeError: 'ascii'
        # codec can't encode character ..." in Python 2 when a format
        # argument (e.g. `output`) has a non-ascii character.
        "Command errored out with exit status {exit_status}:\n"
        " command: {command_display}\n"
        "     cwd: {cwd_display}\n"
        "Complete output ({line_count} lines):\n{output}{divider}"
    ).format(
        exit_status=exit_status,
        command_display=command,
        cwd_display=cwd,
        line_count=len(lines),
        output=output,
        divider=LOG_DIVIDER,
    )
    return msg


def call_subprocess(
    cmd: Union[List[str], CommandArgs],
    show_stdout: bool = False,
    cwd: Optional[str] = None,
    on_returncode: 'Literal["raise", "warn", "ignore"]' = "raise",
    extra_ok_returncodes: Optional[Iterable[int]] = None,
    command_desc: Optional[str] = None,
    extra_environ: Optional[Mapping[str, Any]] = None,
    unset_environ: Optional[Iterable[str]] = None,
    spinner: Optional[SpinnerInterface] = None,
    log_failed_cmd: Optional[bool] = True,
    stdout_only: Optional[bool] = False,
) -> str:
    """
    Args:
      show_stdout: if true, use INFO to log the subprocess's stderr and
        stdout streams.  Otherwise, use DEBUG.  Defaults to False.
      extra_ok_returncodes: an iterable of integer return codes that are
        acceptable, in addition to 0. Defaults to None, which means [].
      unset_environ: an iterable of environment variable names to unset
        prior to calling subprocess.Popen().
      log_failed_cmd: if false, failed commands are not logged, only raised.
      stdout_only: if true, return only stdout, else return both. When true,
        logging of both stdout and stderr occurs when the subprocess has
        terminated, else logging occurs as subprocess output is produced.
    """
    if extra_ok_returncodes is None:
        extra_ok_returncodes = []
    if unset_environ is None:
        unset_environ = []
    # Most places in pip use show_stdout=False. What this means is--
    #
    # - We connect the child's output (combined stderr and stdout) to a
    #   single pipe, which we read.
    # - We log this output to stderr at DEBUG level as it is received.
    # - If DEBUG logging isn't enabled (e.g. if --verbose logging wasn't
    #   requested), then we show a spinner so the user can still see the
    #   subprocess is in progress.
    # - If the subprocess exits with an error, we log the output to stderr
    #   at ERROR level if it hasn't already been displayed to the console
    #   (e.g. if --verbose logging wasn't enabled).  This way we don't log
    #   the output to the console twice.
    #
    # If show_stdout=True, then the above is still done, but with DEBUG
    # replaced by INFO.
    if show_stdout:
        # Then log the subprocess output at INFO level.
        log_subprocess = subprocess_logger.info
        used_level = logging.INFO
    else:
        # Then log the subprocess output using VERBOSE.  This also ensures
        # it will be logged to the log file (aka user_log), if enabled.
        log_subprocess = subprocess_logger.verbose
        used_level = VERBOSE

    # Whether the subprocess will be visible in the console.
    showing_subprocess = subprocess_logger.getEffectiveLevel() <= used_level

    # Only use the spinner if we're not showing the subprocess output
    # and we have a spinner.
    use_spinner = not showing_subprocess and spinner is not None

    if command_desc is None:
        command_desc = format_command_args(cmd)

    log_subprocess("Running command %s", command_desc)
    env = os.environ.copy()
    if extra_environ:
        env.update(extra_environ)
    for name in unset_environ:
        env.pop(name, None)
    try:
        proc = subprocess.Popen(
            # Convert HiddenText objects to the underlying str.
            reveal_command_args(cmd),
            stdin=subprocess.PIPE,
            stdout=subprocess.PIPE,
            stderr=subprocess.STDOUT if not stdout_only else subprocess.PIPE,
            cwd=cwd,
            env=env,
            errors="backslashreplace",
        )
    except Exception as exc:
        if log_failed_cmd:
            subprocess_logger.critical(
                "Error %s while executing command %s",
                exc,
                command_desc,
            )
        raise
    all_output = []
    if not stdout_only:
        assert proc.stdout
        assert proc.stdin
        proc.stdin.close()
        # In this mode, stdout and stderr are in the same pipe.
        while True:
            line: str = proc.stdout.readline()
            if not line:
                break
            line = line.rstrip()
            all_output.append(line + "\n")

            # Show the line immediately.
            log_subprocess(line)
            # Update the spinner.
            if use_spinner:
                assert spinner
                spinner.spin()
        try:
            proc.wait()
        finally:
            if proc.stdout:
                proc.stdout.close()
        output = "".join(all_output)
    else:
        # In this mode, stdout and stderr are in different pipes.
        # We must use communicate() which is the only safe way to read both.
        out, err = proc.communicate()
        # log line by line to preserve pip log indenting
        for out_line in out.splitlines():
            log_subprocess(out_line)
        all_output.append(out)
        for err_line in err.splitlines():
            log_subprocess(err_line)
        all_output.append(err)
        output = out

    proc_had_error = proc.returncode and proc.returncode not in extra_ok_returncodes
    if use_spinner:
        assert spinner
        if proc_had_error:
            spinner.finish("error")
        else:
            spinner.finish("done")
    if proc_had_error:
        if on_returncode == "raise":
            if not showing_subprocess and log_failed_cmd:
                # Then the subprocess streams haven't been logged to the
                # console yet.
                msg = make_subprocess_output_error(
                    cmd_args=cmd,
                    cwd=cwd,
                    lines=all_output,
                    exit_status=proc.returncode,
                )
                subprocess_logger.error(msg)
            raise InstallationSubprocessError(proc.returncode, command_desc)
        elif on_returncode == "warn":
            subprocess_logger.warning(
                'Command "%s" had error code %s in %s',
                command_desc,
                proc.returncode,
                cwd,
            )
        elif on_returncode == "ignore":
            pass
        else:
            raise ValueError(f"Invalid value: on_returncode={on_returncode!r}")
    return output


def runner_with_spinner_message(message: str) -> Callable[..., None]:
    """Provide a subprocess_runner that shows a spinner message.

    Intended for use with for pep517's Pep517HookCaller. Thus, the runner has
    an API that matches what's expected by Pep517HookCaller.subprocess_runner.
    """

    def runner(
        cmd: List[str],
        cwd: Optional[str] = None,
        extra_environ: Optional[Mapping[str, Any]] = None,
    ) -> None:
        with open_spinner(message) as spinner:
            call_subprocess(
                cmd,
                cwd=cwd,
                extra_environ=extra_environ,
                spinner=spinner,
            )

    return runner
site-packages/pip/_internal/utils/__pycache__/subprocess.cpython-39.pyc000064400000014212150564745010022227 0ustar00a

��ReJ'�@s�ddlZddlZddlZddlZddlmZmZmZmZm	Z	m
Z
mZmZddl
mZmZddlmZddlmZmZddlmZer�ddlmZe	eeefZdZeeeefed	�d
d�Zee	eefed	�dd
�Zee	eefe	ed	�dd�Zee	eefeee	eeed�dd�Zdee	eefe eedeeeeeee
eefeeeeeee ee ed�dd�Z!eedd�dd�Z"dS)�N)�
TYPE_CHECKING�Any�Callable�Iterable�List�Mapping�Optional�Union)�SpinnerInterface�open_spinner)�InstallationSubprocessError)�VERBOSE�subprocess_logger)�
HiddenText)�Literalz(----------------------------------------)�args�returncGs2g}|D]$}t|t�r"|�|�q|�|�q|S)z&
    Create a CommandArgs object.
    )�
isinstance�list�extend�append)r�command_args�arg�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/subprocess.py�make_command!s
rcCsd�dd�|D��S)z/
    Format command arguments for display.
    � css0|](}t|t�rt�t|��nt�|�VqdS)N)rr�shlex�quote�str��.0rrrr�	<genexpr>;s�z&format_command_args.<locals>.<genexpr>)�join�rrrr�format_command_args2s	
�r%cCsdd�|D�S)z=
    Return the arguments in their raw, unredacted form.
    cSs g|]}t|t�r|jn|�qSr)rr�secretr rrr�
<listcomp>E�z'reveal_command_args.<locals>.<listcomp>rr$rrr�reveal_command_argsAsr))�cmd_args�cwd�lines�exit_statusrcCs0t|�}d�|�}dj|||t|�|td�}|S)z�
    Create and return the error message to use to log a subprocess error
    with command output.

    :param lines: A list of lines, each ending with a newline.
    �z�Command errored out with exit status {exit_status}:
 command: {command_display}
     cwd: {cwd_display}
Complete output ({line_count} lines):
{output}{divider})r-�command_display�cwd_display�
line_count�output�divider)r%r#�format�len�LOG_DIVIDER)r*r+r,r-�commandr2�msgrrr�make_subprocess_output_errorHs
�	�r9F�raiseTz"Literal["raise", "warn", "ignore"])�cmd�show_stdoutr+�
on_returncode�extra_ok_returncodes�command_desc�
extra_environ�
unset_environ�spinner�log_failed_cmd�stdout_onlyrc
Cs�|durg}|durg}|r*tj}tj}n
tj}t}t��|k}
|
oL|du}|dur^t|�}|d|�tj	�
�}|r�|�|�|D]}|�|d�q�z0t
jt|�t
jt
j|
s�t
jnt
j||dd�}Wn8ty�}z |	r�t�d||��WYd}~n
d}~00g}|
�s�|j�sJ�|j�s"J�|j��|j��}|�s@�qz|��}|�|d�||�|�r,|�snJ�|���q,z|��W|j�r�|j��n|j�r�|j��0d�|�}nT|��\}}|��D]}||��q�|�|�|��D]}||��q�|�|�|}|j�o|j|v}|�rL|�s0J�|�rB|� d�n
|� d	�|�r�|d
k�r�|
�s�|	�r�t!||||jd�}t�"|�t#|j|��n8|dk�r�t�$d
||j|�n|dk�r�nt%d|����|S)a�
    Args:
      show_stdout: if true, use INFO to log the subprocess's stderr and
        stdout streams.  Otherwise, use DEBUG.  Defaults to False.
      extra_ok_returncodes: an iterable of integer return codes that are
        acceptable, in addition to 0. Defaults to None, which means [].
      unset_environ: an iterable of environment variable names to unset
        prior to calling subprocess.Popen().
      log_failed_cmd: if false, failed commands are not logged, only raised.
      stdout_only: if true, return only stdout, else return both. When true,
        logging of both stdout and stderr occurs when the subprocess has
        terminated, else logging occurs as subprocess output is produced.
    NzRunning command %s�backslashreplace)�stdin�stdout�stderrr+�env�errorsz#Error %s while executing command %s�
r.�error�doner:)r*r+r,r-�warnz$Command "%s" had error code %s in %s�ignorezInvalid value: on_returncode=)&r�info�logging�INFO�verboser
�getEffectiveLevelr%�os�environ�copy�update�pop�
subprocess�Popenr)�PIPE�STDOUT�	Exception�criticalrGrF�close�readline�rstripr�spin�waitr#�communicate�
splitlines�
returncode�finishr9rLr�warning�
ValueError)r;r<r+r=r>r?r@rArBrCrD�log_subprocess�
used_level�showing_subprocess�use_spinnerrI�name�proc�exc�
all_output�liner2�out�err�out_line�err_line�proc_had_errorr8rrr�call_subprocessks�


�

�



�




�

�
ry).N)�messagercs2dttttttttfdd��fdd�
}|S)z�Provide a subprocess_runner that shows a spinner message.

    Intended for use with for pep517's Pep517HookCaller. Thus, the runner has
    an API that matches what's expected by Pep517HookCaller.subprocess_runner.
    N)r;r+r@rcs<t��� }t||||d�Wd�n1s.0YdS)N)r+r@rB)rry)r;r+r@rB�rzrr�runners
�z+runner_with_spinner_message.<locals>.runner)NN)rrrrr)rzr|rr{r�runner_with_spinner_message
s	��
r})
FNr:NNNNNTF)#rQrUrrZ�typingrrrrrrrr	�pip._internal.cli.spinnersr
r�pip._internal.exceptionsr�pip._internal.utils.loggingr
r�pip._internal.utils.miscrrr�CommandArgsr6rr%r)�intr9�boolryr}rrrr�<module>s^( �%�

�#site-packages/pip/_internal/utils/__pycache__/logging.cpython-39.pyc000064400000022155150564745010021472 0ustar00a

��Re-�@s�ddlZddlZddlZddlZddlZddlZddlmZddlmZm	Z	m
Z
mZmZm
Z
mZmZddlmZmZddlmZddlmZddlmZzddlZWney�ddlZYn0zddlmZWney�dZYn0e��Z ed	�Z!Gd
d�de�Z"ee#e#e$d�d
d�Z%ej&d&e'edd�dd��Z(e'd�dd�Z)Gdd�dej*�Z+e,e
e,ge,fd�dd�Z-Gdd�dej.�Z/Gdd�dej0j1�Z2Gdd �d e�Z3Gd!d"�d"e�Z4e'e$ee,e'd#�d$d%�Z5dS)'�N)�Filter)�IO�Any�Callable�Iterator�Optional�TextIO�Type�cast)�VERBOSE�	getLogger)�WINDOWS)�DEPRECATION_MSG_PREFIX)�
ensure_dir)�coloramazpip.subprocessorc@seZdZdZdS)�BrokenStdoutLoggingErrorzO
    Raised if BrokenPipeError occurs for the stdout stream while logging.
    N)�__name__�
__module__�__qualname__�__doc__�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/logging.pyr!sr)�	exc_class�exc�returncCs0|turdStsdSt|t�o.|jtjtjfvS)NTF)�BrokenPipeErrorr
�
isinstance�OSError�errno�EINVAL�EPIPE)rrrrr�_is_broken_pipe_error's
r!�)�numrc	csDt�t_tj|7_zdVWtj|8_ntj|8_0dS)zv
    A context manager which will cause the log output to be indented for any
    log messages emitted inside it.
    N)�get_indentation�
_log_state�indentation)r#rrr�
indent_log4s
r'�rcCsttdd�S)Nr&r)�getattrr%rrrrr$Csr$csZeZdZdZdd�eeedd��fdd�Zeeed�d	d
�Z	e
jed��fdd
�Z�Z
S)�IndentingFormatterz%Y-%m-%dT%H:%M:%SF)�
add_timestampN)�argsr+�kwargsrcs||_t�j|i|��dS)z�
        A logging.Formatter that obeys the indent_log() context manager.

        :param add_timestamp: A bool indicating output lines should be prefixed
            with their record's timestamp.
        N)r+�super�__init__)�selfr+r,r-��	__class__rrr/JszIndentingFormatter.__init__)�	formatted�levelnorcCs.|tjkrdS|�t�rdS|tjkr*dSdS)zv
        Return the start of the formatted log message (not counting the
        prefix to add to each line).
        �z	WARNING: zERROR: )�logging�WARNING�
startswithr�ERROR)r0r3r4rrr�get_message_startYs


z$IndentingFormatter.get_message_start��recordrcslt��|�}|�||j�}||}d�|jr<|�|��d���dt�7�d��fdd�|�d�D��}|S)z�
        Calls the standard formatter, but will indent all of the log message
        lines by our current indentation level.
        r5� csg|]}�|�qSrr)�.0�line��prefixrr�
<listcomp>v�z-IndentingFormatter.format.<locals>.<listcomp>T)	r.�formatr:r4r+�
formatTimer$�join�
splitlines)r0r<r3�
message_startr1r@rrDiszIndentingFormatter.format)rrr�default_time_formatr�boolr/�str�intr:r6�	LogRecordrD�
__classcell__rrr1rr*Gs��r*)�colorsrcsttd��fdd�}|S)N)�inprcsd�t��|tjjg�S)Nr5)rF�listr�Style�	RESET_ALL)rP�rOrr�wrapped{sz_color_wrap.<locals>.wrapped)rK)rOrUrrTr�_color_wrapzsrVcs�eZdZer2ejeejj�fej	eejj
�fgZngZdee
edd��fdd�
Zed�dd�Zed�dd	�Zejed
��fdd�Zejdd
��fd
d�Z�ZS)�ColorizedStreamHandlerN)�stream�no_colorrcs,t��|�||_tr(tr(t�|j�|_dS�N)r.r/�	_no_colorr
r�AnsiToWin32rX)r0rXrYr1rrr/�szColorizedStreamHandler.__init__r(cCs.tr"tr"ttj|j�}|jtjuS|jtjuS)zA
        Return whether the handler is using sys.stdout.
        )r
rr
r\rXrU�sys�stdout)r0rXrrr�
_using_stdout�sz$ColorizedStreamHandler._using_stdoutcCsXtr
|jrdSt|jtj�s"|jn|jj}t|d�r@|��r@dStj	�
d�dkrTdSdS)NF�isattyT�TERM�ANSI)rr[rrXr\rU�hasattrr`�os�environ�get)r0�real_streamrrr�should_color�s
��z#ColorizedStreamHandler.should_colorr;cs>t��|�}|��r:|jD]\}}|j|kr||�}q:q|SrZ)r.rDrh�COLORSr4)r0r<�msg�level�colorr1rrrD�s
zColorizedStreamHandler.formatcs@t��dd�\}}|r4|r4|��r4t||�r4t��t��|�S)Nr")r]�exc_infor_r!rr.�handleError)r0r<rrr1rrrn�s����z"ColorizedStreamHandler.handleError)NN)rrrrr6r9rV�Fore�REDr7�YELLOWrirrrJr/r_rhrMrKrDrnrNrrr1rrW�s�rWcs&eZdZeed��fdd�Z�ZS)�BetterRotatingFileHandlerr(csttj�|j��t���SrZ)rrd�path�dirname�baseFilenamer.�_open)r0r1rrrv�szBetterRotatingFileHandler._open)rrrrrrvrNrrr1rrr�srrc@s.eZdZedd�dd�Zejed�dd�ZdS)�MaxLevelFilterN)rkrcCs
||_dSrZ)rk)r0rkrrrr/�szMaxLevelFilter.__init__r;cCs|j|jkSrZ)r4rk�r0r<rrr�filter�szMaxLevelFilter.filter)	rrrrLr/r6rMrJryrrrrrw�srwcs*eZdZdZejed��fdd�Z�ZS)�ExcludeLoggerFilterzQ
    A logging Filter that excludes records from a logger (or its children).
    r;cst��|�SrZ)r.ryrxr1rrry�szExcludeLoggerFilter.filter)	rrrrr6rMrJryrNrrr1rrz�srz)�	verbosityrY�
user_log_filercCs~|dkrtj}nD|dkrt}n6|dkr.tj}n&|dkr>tj}n|dkrNtj}ntj}t�|�}|du}|rt|}d}nd}|}|d	vr�d
nd}ddd
�}	ddd�}
gd�|r�dgng}tj�	dddtjd�dt
jd�dt
jd�d�tdd�tddd�d�||
d||	d d!d"gd#d$�d
|
d||	d%d!gd#d$�||
d||	d%d&gd#d$�d|
d'|d(dd)d*�d+�||d,�d-d.|iid/��|S)0znConfigures and sets up all of the logging

    Returns the requested logging level, as its integer value.
    r"����������N�DEBUGz	/dev/null)�INFOr9r7zext://sys.stdoutzext://sys.stderr)r^�stderrz2pip._internal.utils.logging.ColorizedStreamHandlerz5pip._internal.utils.logging.BetterRotatingFileHandler)rX�file)�console�console_errors�console_subprocess�user_logFz*pip._internal.utils.logging.MaxLevelFilter)�()rkzlogging.Filter)r��namez/pip._internal.utils.logging.ExcludeLoggerFilter)�exclude_warnings�restrict_to_subprocess�exclude_subprocessz%(message)s)r�rDT)r�rDr+)�indent�indent_with_timestamprXr^r�r�r�)rk�classrYrX�filters�	formatterr�r�r�zutf-8r�)rkr��filename�encoding�delayr�)r�r�r�r�)rk�handlerszpip._vendorrk)�version�disable_existing_loggersr��
formattersr��root�loggers)
r6r�rr7r9�CRITICALr��getLevelName�config�
dictConfig�subprocess_loggerr�r*)r{rYr|�level_numberrk�include_user_log�additional_log_file�
root_level�vendored_log_level�log_streams�handler_classesr�rrr�
setup_logging�s�
����������
�	��	��%�
��Ir�)r")6�
contextlibrr6�logging.handlersrdr]r�typingrrrrrrr	r
Zpip._internal.utils._logrr�pip._internal.utils.compatr
�pip._internal.utils.deprecationr�pip._internal.utils.miscr�	threading�ImportErrorZdummy_threading�pip._vendorr�	Exception�localr%r�r�
BaseExceptionrJr!�contextmanagerrLr'r$�	Formatterr*rKrV�
StreamHandlerrWr��RotatingFileHandlerrrrwrzr�rrrr�<module>sB(

3Qsite-packages/pip/_internal/utils/__pycache__/deprecation.cpython-39.pyc000064400000006327150564745010022344 0ustar00a

��Re+�	@s�UdZddlZddlZddlmZmZmZmZmZddl	m
Z
ddlmZ
dZGdd�de�Zdaeed	<deeefeeeeeeeedd
�dd�Zdd
�dd�Zddd�eeeeeeeeedd�dd�ZdS)zN
A module that implements tooling to enable easy warnings about deprecations.
�N)�Any�Optional�TextIO�Type�Union)�parse)�__version__z
DEPRECATION: c@seZdZdS)�PipDeprecationWarningN)�__name__�
__module__�__qualname__�r
r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/deprecation.pyr	sr	�_original_showwarning)�message�category�filename�lineno�file�line�returncCsZ|dur$tdurVt||||||�n2t|t�rDt�d�}|�|�nt||||||�dS)Nzpip._internal.deprecations)r�
issubclassr	�logging�	getLogger�warning)rrrrrr�loggerr
r
r�_showwarnings

r)rcCs(tjdtdd�tdur$tjatt_dS)N�defaultT)�append)�warnings�simplefilterr	r�showwarningrr
r
r
r�install_warning_logger,sr")�feature_flag�issue)�reason�replacement�gone_inr#r$rcCs�|duott�t|�k}|t�d�f||s.dndf|df||sBdndf|dfg}d�d	d
�|D��}|rpt|��tj|tdd�dS)
aHelper to deprecate existing functionality.

    reason:
        Textual reason shown to the user about why this functionality has
        been deprecated. Should be a complete sentence.
    replacement:
        Textual suggestion shown to the user about what alternative
        functionality they can use.
    gone_in:
        The version of pip does this functionality should get removed in.
        Raises an error if pip's current version is greater than or equal to
        this.
    feature_flag:
        Command-line flag of the form --use-feature={feature_flag} for testing
        upcoming functionality.
    issue:
        Issue number on the tracker that would serve as a useful place for
        users to find related discussion and provide feedback.
    Nz{}z*pip {} will enforce this behaviour change.z*Since pip {}, this is no longer supported.zA possible replacement is {}.zEYou can use the flag --use-feature={} to test the upcoming behaviour.z@Discussion can be found at https://github.com/pypa/pip/issues/{}� css,|]$\}}|dur|dur|�|�VqdS)N)�format)�.0�value�
format_strr
r
r�	<genexpr>ns�zdeprecated.<locals>.<genexpr>�)r�
stacklevel)r�current_version�DEPRECATION_MSG_PREFIX�joinr	r�warn)r%r&r'r#r$�is_gone�
message_partsrr
r
r�
deprecated7s2�������
�r6)NN)�__doc__rr�typingrrrrrZpip._vendor.packaging.versionr�piprr0r1�Warningr	r�__annotations__�str�intrr"r6r
r
r
r�<module>s<	�
���site-packages/pip/_internal/utils/__pycache__/urls.cpython-39.pyc000064400000003142150564745010021024 0ustar00a

��Re��@spddlZddlZddlZddlZddlmZddlmZe	ee	d�dd�Z
e	e	d�d	d
�Ze	e	d�dd�ZdS)
�N)�Optional�)�WINDOWS)�url�returncCs d|vrdS|�dd�d��S)N�:rr)�split�lower)r�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/urls.py�get_url_scheme
sr)�pathrcCs.tj�tj�|��}tj�dtj�|��}|S)zh
    Convert a path to a file: URL.  The path will be made absolute and have
    quoted path parts.
    �file:)	�osr
�normpath�abspath�urllib�parse�urljoin�request�pathname2url)r
rr
r
r�path_to_urlsrcCs�|�d�sJd|�d���tj�|�\}}}}}|r<|dkrBd}ntrPd|}ntd|����tj�||�}tr�|s�t|�dkr�|d	d
kr�|dt	j
vr�|dd
�dvr�|dd�}|S)z(
    Convert a file: URL to a path.
    rz1You can only turn file: urls into filenames (not �)�	localhost�z\\z8non-local file URIs are not supported on this platform: �r�/r��)rz:/N)�
startswithrr�urlsplitr�
ValueErrorr�url2pathname�len�string�
ascii_letters)r�_�netlocr
r
r
r�url_to_paths8�
�
���
�
���r()
rr$�urllib.parser�urllib.request�typingr�compatr�strrrr(r
r
r
r�<module>s
site-packages/pip/_internal/utils/__pycache__/appdirs.cpython-39.pyc000064400000003175150564745010021507 0ustar00a

��Re��@s|dZddlZddlZddlmZddlmZeed�dd�Z	dee
ed	�d
d�Zdee
ed	�dd
�Zeeed�dd�Z
dS)z�
This code wraps the vendored appdirs module to so the return values are
compatible for the current pip code base.

The intention is to rewrite current usages gradually, keeping the tests pass,
and eventually drop this after all usages are changed.
�N)�List)�platformdirs)�appname�returncCstj|dd�S)NF)�	appauthor)�_appdirs�user_cache_dir)r�r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/appdirs.pyrsrT)r�roamingrcCsBtj|d|d�}tj�|�r |Sd}|r6tj�||�}tj�|�S)NF�rrz
~/.config/)r�
user_data_dir�os�path�isdir�join�
expanduser)rrr�linux_like_pathr	r	r
�_macos_user_config_dirsrcCs$tjdkrt||�Stj|d|d�S)N�darwinFr)�sys�platformrr�user_config_dir)rrr	r	r
r"s

rcCsNtjdkrtj|ddd�gStj|ddd�}tjdkr<|gS|�tj�dgS)NrFT)r�	multipath�win32z/etc)rrr�
site_data_dir�site_config_dir�splitr�pathsep)r�dirvalr	r	r
�site_config_dirs+s

r )T)T)�__doc__rr�typingr�pip._vendorrr�strr�boolrrr r	r	r	r
�<module>s	site-packages/pip/_internal/utils/__pycache__/virtualenv.cpython-39.pyc000064400000006374150564745010022250 0ustar00a

��Re�
�@s�ddlZddlZddlZddlZddlZddlmZmZe�e	�Z
e�d�Ze
d�dd�Ze
d�dd�Ze
d�d	d
�Zeeed�dd�Ze
d�d
d�Ze
d�dd�Ze
d�dd�ZdS)�N)�List�Optionalz8include-system-site-packages\s*=\s*(?P<value>true|false))�returncCstjttdtj�kS)znChecks if sys.base_prefix and sys.prefix match.

    This handles PEP 405 compliant virtual environments.
    �base_prefix)�sys�prefix�getattr�r	r	��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/virtualenv.py�_running_under_venvsrcCs
ttd�S)zmChecks if sys.real_prefix is set.

    This handles virtual environments created with pypa's virtualenv.
    �real_prefix)�hasattrrr	r	r	r
�!_running_under_regular_virtualenvsrcCst�p
t�S)zBReturn True if we're running inside a virtualenv, False otherwise.)rrr	r	r	r
�running_under_virtualenvsrcCshtj�tjd�}z>t|dd��}|����Wd�WS1sB0YWntybYdS0dS)z�Reads {sys.prefix}/pyvenv.cfg and returns its contents as list of lines

    Returns None, if it could not read/access the file.
    z
pyvenv.cfgzutf-8)�encodingN)	�os�path�joinrr�open�read�
splitlines�OSError)�pyvenv_cfg_file�fr	r	r
�_get_pyvenv_cfg_lines$s0rcCsPt�}|durt�d�dS|D]*}t�|�}|dur |�d�dkr dSq dS)aZCheck `{sys.prefix}/pyvenv.cfg` for system site-packages inclusion

    PEP 405 specifies that when system site-packages are not supposed to be
    visible from a virtual environment, `pyvenv.cfg` must contain the following
    line:

        include-system-site-packages = false

    Additionally, log a warning if accessing the file fails.
    Nz�Could not access 'pyvenv.cfg' despite a virtual environment being active. Assuming global site-packages is not accessible in this environment.T�value�falseF)r�logger�warning�#_INCLUDE_SYSTEM_SITE_PACKAGES_REGEX�match�group)�	cfg_lines�liner r	r	r
�_no_global_under_venv3s�
r$cCs0tj�tj�tj��}tj�|d�}tj�|�S)z�Check if "no-global-site-packages.txt" exists beside site.py

    This mirrors logic in pypa/virtualenv for determining whether system
    site-packages are visible in the virtual environment.
    zno-global-site-packages.txt)rr�dirname�abspath�site�__file__r�exists)�site_mod_dir�no_global_site_packages_filer	r	r
�#_no_global_under_regular_virtualenvPs�r,cCst�rt�St�rt�SdS)zHReturns a boolean, whether running in venv with no system site-packages.F)rr$rr,r	r	r	r
�virtualenv_no_global^s
r-)�loggingr�rer'r�typingrr�	getLogger�__name__r�compiler�boolrrr�strrr$r,r-r	r	r	r
�<module>s 
�	site-packages/pip/_internal/utils/__pycache__/datetime.cpython-39.pyc000064400000001045150564745010021633 0ustar00a

��Re��@s$dZddlZeeeed�dd�ZdS)z.For when pip wants to check the date or time.
�N)�year�month�day�returncCs tj��}t�|||�}||kS)N)�datetime�date�today)rrrr�given�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/datetime.py�today_is_later_thans
r)�__doc__r�int�boolrr
r
r
r�<module>ssite-packages/pip/_internal/utils/__pycache__/inject_securetransport.cpython-39.pyc000064400000001773150564745010024646 0ustar00a

��Re�@s$dZddlZdd�dd�Ze�dS)a-A helper module that injects SecureTransport, on import.

The import should be done as early as possible, to ensure all requests and
sessions (or whatever) are created after injecting SecureTransport.

Note that we only do the injection on macOS, when the linked OpenSSL is too
old to handle TLSv1.2.
�N)�returnc	CsttjdkrdSzddl}Wnty.YdS0|jdkr>dSzddlm}WnttfyfYdS0|��dS)N�darwinri)�securetransport)	�sys�platform�ssl�ImportError�OPENSSL_VERSION_NUMBER�pip._vendor.urllib3.contribr�OSError�inject_into_urllib3)rr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/inject_securetransport.py�inject_securetransport
s

r)�__doc__rrr
r
r
r�<module>s	site-packages/pip/_internal/utils/__pycache__/filetypes.cpython-39.pyc000064400000001724150564745010022047 0ustar00a

��Re��@s�UdZddlmZddlmZdZdZeedfed<dZ	eedfed	<d
efZ
eedfed<dZeedfed
<e
eee	Zee
d�dd�ZdS)zFiletype information.
�)�Tuple)�splitextz.whl)z.tar.bz2z.tbz.�BZ2_EXTENSIONS)z.tar.xzz.txzz.tlzz.tar.lzz	.tar.lzma�
XZ_EXTENSIONSz.zip�ZIP_EXTENSIONS)z.tar.gzz.tgzz.tar�TAR_EXTENSIONS)�name�returncCs t|�d��}|tvrdSdS)z9Return True if `name` is a considered as an archive file.�TF)r�lower�ARCHIVE_EXTENSIONS)r�ext�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/filetypes.py�is_archive_filesrN)�__doc__�typingr�pip._internal.utils.miscr�WHEEL_EXTENSIONr�str�__annotations__rrrr�boolrrrrr�<module>ssite-packages/pip/_internal/utils/__pycache__/compat.cpython-39.pyc000064400000003015150564745010021321 0ustar00a

��Re\�@sxdZddlZddlZddlZgd�Ze�e�Zed�dd�Z	e
ed�dd	�Zhd
�Z
ej�d�prejdkorejd
kZdS)zKStuff that differs in different Python versions and platform
distributions.�N)�get_path_uid�stdlib_pkgs�WINDOWS)�returncCs2zddl}WdSty Yn0ddlm}|S)NrT)�IS_PYOPENSSL)�_ssl�ImportError�pip._vendor.urllib3.utilr)rr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/compat.py�has_tlssr)�pathrcCsbttd�r6t�|tjtjB�}t�|�j}t�|�n(tj�	|�sPt�
|�j}nt|�d���|S)a)
    Return path's uid.

    Does not follow symlinks:
        https://github.com/pypa/pip/pull/935#discussion_r5307003

    Placed this function in compat due to differences on AIX and
    Jython, that should eventually go away.

    :raises OSError: When path is a symlink or can't be read.
    �
O_NOFOLLOWz/ is a symlink; Will not return uid for symlinks)�hasattr�os�open�O_RDONLYr�fstat�st_uid�closer
�islink�stat�OSError)r
�fd�file_uidr
r
rrs
r>�argparse�wsgiref�python�win�cli�nt)�__doc__�loggingr�sys�__all__�	getLogger�__name__�logger�boolr�str�intrr�platform�
startswith�namerr
r
r
r�<module>s

 site-packages/pip/_internal/utils/__pycache__/_log.cpython-39.pyc000064400000003023150564745010020755 0ustar00a

��Re��@sTdZddlZddlmZmZdZGdd�dej�Zeed�dd	�Z	dd
�dd�Z
dS)
z�Customize logging

Defines custom logger class for the `logger.verbose(...)` method.

init_logging() must be called before any other modules that call logging.getLogger.
�N)�Any�cast�c@s$eZdZdZeeedd�dd�ZdS)�
VerboseLoggerzXCustom Logger, defining a verbose log-level

    VERBOSE is between INFO and DEBUG.
    N)�msg�args�kwargs�returncOs|jt|g|�Ri|��S)N)�log�VERBOSE)�selfrrr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/_log.py�verboseszVerboseLogger.verbose)�__name__�
__module__�__qualname__�__doc__�strrrr
r
r
rrsr)�namer	cCsttt�|��S)zBlogging.getLogger, but ensures our VerboseLogger class is returned)rr�logging�	getLogger)rr
r
rrsr)r	cCst�t�t�td�dS)z�Register our VerboseLogger and VERBOSE log level.

    Should be called before any calls to getLogger(),
    i.e. in pip._internal.__init__
    rN)r�setLoggerClassr�addLevelNamerr
r
r
r�init_loggings
r)rr�typingrrr�Loggerrrrrr
r
r
r�<module>s
site-packages/pip/_internal/utils/__pycache__/egg_link.cpython-39.pyc000064400000004200150564745010021612 0ustar00a

��Re��@s�ddlZddlZddlZddlmZddlmZmZddlm	Z	m
Z
ddgZeed�dd	�Z
eeed�d
d�Zeeed�dd�ZdS)�N)�Optional)�
site_packages�	user_site)�running_under_virtualenv�virtualenv_no_global�egg_link_path_from_sys_path�egg_link_path_from_location)�raw_name�returncCst�dd|�dS)z�
    Convert a Name metadata value to a .egg-link name, by applying
    the same substitution as pkg_resources's safe_name function.
    Note: we cannot use canonicalize_name because it has a different logic.
    z[^A-Za-z0-9.]+�-z	.egg-link)�re�sub)r	�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/egg_link.py�_egg_link_namesrcCs:t|�}tjD]&}tj�||�}tj�|�r|SqdS)zJ
    Look for a .egg-link file for project name, by walking sys.path.
    N)r�sys�path�os�join�isfile)r	�
egg_link_name�	path_item�egg_linkrrrrs

cCszg}t�r*|�t�t�sBtrB|�t�ntr8|�t�|�t�t|�}|D]&}tj�||�}tj�	|�rN|SqNdS)a
    Return the path for the .egg-link file if it exists, otherwise, None.

    There's 3 scenarios:
    1) not in a virtualenv
       try to find in site.USER_SITE, then site_packages
    2) in a no-global virtualenv
       try to find in site_packages
    3) in a yes-global virtualenv
       try to find in site_packages, then site.USER_SITE
       (don't look in global location)

    For #1 and #3, there could be odd cases, where there's an egg-link in 2
    locations.

    This method will just return the first one found.
    N)
r�appendrrrrrrrr)r	�sitesr�site�egglinkrrrr*s




)rrr�typingr�pip._internal.locationsrr�pip._internal.utils.virtualenvrr�__all__�strrrrrrrr�<module>s�	site-packages/pip/_internal/utils/__pycache__/pkg_resources.cpython-39.pyc000064400000003563150564745010022721 0ustar00a

��Re��@s2ddlmZmZmZddlmZGdd�d�ZdS)�)�Dict�Iterable�List)�yield_linesc@s�eZdZdZeeefdd�dd�Zeed�dd�Z	eed�d	d
�Z
eeed�dd�Zeed�d
d�Z
eeed�dd�Zeedd�dd�ZdS)�DictMetadataz>IMetadataProvider that reads metadata files from a dictionary.N)�metadata�returncCs
||_dS�N��	_metadata)�selfr�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/pkg_resources.py�__init__	szDictMetadata.__init__)�namercCs
||jvSr	r
�rrr
r
r�has_metadataszDictMetadata.has_metadatac
CsRz|j|��WStyL}z$|jd|�d�7_�WYd}~n
d}~00dS)Nz in z file)r�decode�UnicodeDecodeError�reason)rr�er
r
r�get_metadatas
zDictMetadata.get_metadatacCst|�|��Sr	)rrrr
r
r�get_metadata_linesszDictMetadata.get_metadata_linescCsdS)NFr
rr
r
r�metadata_isdirszDictMetadata.metadata_isdircCsgSr	r
rr
r
r�metadata_listdirszDictMetadata.metadata_listdir)�script_name�	namespacercCsdSr	r
)rrrr
r
r�
run_script szDictMetadata.run_script)�__name__�
__module__�__qualname__�__doc__r�str�bytesr�boolrrrrrrrrr
r
r
rrsrN)�typingrrr�pip._vendor.pkg_resourcesrrr
r
r
r�<module>ssite-packages/pip/_internal/utils/__pycache__/entrypoints.cpython-39.pyc000064400000002532150564745010022437 0ustar00a

��Rej�@sDddlZddlmZmZddlmZdeeeeed�dd�Z	dS)	�N)�List�Optional)�mainF)�args�_nowarn�returncCs|stj�d�t|�S)a�Central wrapper for all old entrypoints.

    Historically pip has had several entrypoints defined. Because of issues
    arising from PATH, sys.path, multiple Pythons, their interactions, and most
    of them having a pip installed, users suffer every time an entrypoint gets
    moved.

    To alleviate this pain, and provide a mechanism for warning users and
    directing them to an appropriate place for help, we now define all of
    our old entrypoints as wrappers for the current one.
    aWARNING: pip is being invoked by an old script wrapper. This will fail in a future version of pip.
Please see https://github.com/pypa/pip/issues/5599 for advice on fixing the underlying issue.
To avoid this problem you can invoke Python with '-m pip' instead of running pip directly.
)�sys�stderr�writer)rr�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/entrypoints.py�_wrappers
�r
)NF)
r�typingrr�pip._internal.cli.mainr�str�bool�intr
rrrr�<module>ssite-packages/pip/_internal/utils/__pycache__/temp_dir.cpython-39.pyc000064400000016201150564745010021642 0ustar00a

��Re��@sUddlZddlZddlZddlZddlZddlmZmZddl	m
Z
mZmZm
Z
mZmZddlmZmZe�e�Zeddd�Zedd	d
d�Zdae
eed<eedd
�dd��ZGdd�d�Zdae
eed<eeed
�dd��ZGdd�d�Ze�ZGdd�d�Z Gdd�de �Z!dS)�N)�	ExitStack�contextmanager)�Any�Dict�Iterator�Optional�TypeVar�Union)�enum�rmtree�_T�
TempDirectory)�boundz	build-envzephem-wheel-cachez	req-build)�	BUILD_ENV�EPHEM_WHEEL_CACHE�	REQ_BUILD�_tempdir_manager��returnccsJt��0}t|}azdVW|an|a0Wd�n1s<0YdS�N)rr)�stack�old_tempdir_manager�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/temp_dir.py�global_tempdir_managers

rc@s@eZdZdZdd�dd�Zeedd�dd�Zeed	�d
d�ZdS)�TempDirectoryTypeRegistryzManages temp directory behaviorNrcCs
i|_dSr��_should_delete��selfrrr�__init__*sz"TempDirectoryTypeRegistry.__init__)�kind�valuercCs||j|<dS)z[Indicate whether a TempDirectory of the given kind should be
        auto-deleted.
        Nr)rr!r"rrr�
set_delete-sz$TempDirectoryTypeRegistry.set_delete�r!rcCs|j�|d�S)z^Get configured auto-delete flag for a given TempDirectory type,
        default True.
        T)r�get)rr!rrr�
get_delete3sz$TempDirectoryTypeRegistry.get_delete)	�__name__�
__module__�__qualname__�__doc__r �str�boolr#r&rrrrr'sr�_tempdir_registryccs$t}t�aztVW|an|a0dS)zuProvides a scoped global tempdir registry that can be used to dictate
    whether directories should be deleted.
    N)r-r)�old_tempdir_registryrrr�tempdir_registry=s
r/c@seZdZdS)�_DefaultN)r'r(r)rrrrr0Ksr0cs�eZdZdZdeddfeeeede	feed��fdd�
Z
eed�d	d
��Zed�dd�Z
eed
�dd�Zeeedd�dd�Zeed�dd�Zdd�dd�Z�ZS)r
aMHelper class that owns and cleans up a temporary directory.

    This class can be used as a context manager or as an OO representation of a
    temporary directory.

    Attributes:
        path
            Location to the created temporary directory
        delete
            Whether the directory should be deleted when exiting
            (when used as a contextmanager)

    Methods:
        cleanup()
            Deletes the temporary directory

    When used as a context manager, if the delete attribute is True, on
    exiting the context the temporary directory is deleted.
    N�tempF)�path�deleter!�globally_managedcslt���|tur$|dur d}nd}|dur6|�|�}||_d|_||_||_|rhtdus^J�t�	|�dS)NF)
�superr �_default�_create�_path�_deletedr3r!r�
enter_context)rr2r3r!r4��	__class__rrr gs

zTempDirectory.__init__rcCs|jrJd|j����|jS)Nz"Attempted to access deleted path: )r9r8rrrrr2�szTempDirectory.pathcCsd|jj�d|j�d�S)N�<� �>)r<r'r2rrrr�__repr__�szTempDirectory.__repr__)rrcCs|Srrrrrr�	__enter__�szTempDirectory.__enter__)�excr"�tbrcCs8|jdur|j}ntr$t�|j�}nd}|r4|��dS)NT)r3r-r&r!�cleanup)rrBr"rCr3rrr�__exit__�s
zTempDirectory.__exit__r$cCs,tj�tjd|�d�d��}t�d|�|S)z<Create a temporary directory and store its path in self.path�pip-�-��prefix�Created temporary directory: %s)�osr2�realpath�tempfile�mkdtemp�logger�debug)rr!r2rrrr7�szTempDirectory._createcCs&d|_tj�|j�sdSt|j�dS)z6Remove the temporary directory created and reset stateTN)r9rKr2�existsr8rrrrrrD�szTempDirectory.cleanup)r'r(r)r*r6rr+r	r,r0r �propertyr2r@rrArrEr7rD�
__classcell__rrr;rr
Rs$��!
cs\eZdZdZdZd
eeedd��fdd�
Ze	ee
ed�dd	��Zeed
�dd�Z�Z
S)�AdjacentTempDirectorya�Helper class that creates a temporary directory adjacent to a real one.

    Attributes:
        original
            The original directory to create a temp directory for.
        path
            After calling create() or entering, contains the full
            path to the temporary directory.
        delete
            Whether the directory should be deleted when exiting
            (when used as a contextmanager)

    z-~.=%0123456789N)�originalr3rcs|�d�|_t�j|d�dS)Nz/\)r3)�rstriprUr5r )rrUr3r;rrr �szAdjacentTempDirectory.__init__)�namerccs�tdt|��D]D}t�|j|d�D],}dd�|�||d�}||kr$|Vq$qtt|j��D]8}t�|j|�D]$}dd�|�|}||krt|VqtqbdS)aGenerates a series of temporary names.

        The algorithm replaces the leading characters in the name
        with ones that are valid filesystem characters, but are not
        valid package names (for both Python and pip definitions of
        package).
        ��~�N)�range�len�	itertools�combinations_with_replacement�
LEADING_CHARS�join)�clsrW�i�	candidate�new_namerrr�_generate_names�s	
�
�z%AdjacentTempDirectory._generate_namesr$cCs�tj�|j�\}}|�|�D]d}tj�||�}zt�|�Wn2tyn}z|jtj	krZ�WYd}~qd}~00tj�
|�}q�qtj�
tjd|�d�d��}t
�d|�|S)NrFrGrHrJ)rKr2�splitrUrer`�mkdir�OSError�errno�EEXISTrLrMrNrOrP)rr!�rootrWrcr2�exrrrr7�szAdjacentTempDirectory._create)N)r'r(r)r*r_r+rr,r �classmethodrrer7rSrrr;rrT�srT)"rir]�logging�os.pathrKrM�
contextlibrr�typingrrrrrr	�pip._internal.utils.miscr
r�	getLoggerr'rOr�
tempdir_kindsr�__annotations__rrr-r/r0r6r
rTrrrr�<module>s2
 
�

^site-packages/pip/_internal/utils/__pycache__/setuptools_build.cpython-39.pyc000064400000006615150564745010023447 0ustar00a

��ReY�
@sddlZddlmZmZmZdZdeeeeeeed�dd�Zeeeeeeeed�d	d
�Z	eeeeed�dd
�Z
eeeeeeeeeeeeed�dd�Zeeeeeed�dd�Zeeeeeeeeeeeeeeeeeeed�dd�Z
dS)�N)�List�Optional�Sequencea'import io, os, sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};f = getattr(tokenize, 'open', open)(__file__) if os.path.exists(__file__) else io.StringIO('from setuptools import setup; setup()');code = f.read().replace('\r\n', '\n');f.close();exec(compile(code, __file__, 'exec'))F)�
setup_py_path�global_options�no_user_config�unbuffered_output�returncCsFtjg}|r|dg7}|dt�|�g7}|r4||7}|rB|dg7}|S)ao
    Get setuptools command arguments with shim wrapped setup file invocation.

    :param setup_py_path: The path to setup.py to be wrapped.
    :param global_options: Additional global options.
    :param no_user_config: If True, disables personal user configuration.
    :param unbuffered_output: If True, adds the unbuffered switch to the
     argument list.
    z-uz-cz
--no-user-cfg)�sys�
executable�_SETUPTOOLS_SHIM�format)rrrr�args�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/setuptools_build.py�make_setuptools_shim_argss

r)rr�
build_options�destination_dirr	cCs(t||dd�}|dd|g7}||7}|S)NT�rr�bdist_wheelz-d�r)rrrrrrrr� make_setuptools_bdist_wheel_args/s
�r)rrr	cCst||dd�}|ddg7}|S)NTr�cleanz--allr)rrrrrr�make_setuptools_clean_argsAs
�r)rr�install_optionsr�prefix�home�
use_user_siter	cCsf|r|rJ�t|||d�}|ddg7}||7}|r>|d|g7}|durR|d|g7}|rb|ddg7}|S)N)rr�developz	--no-deps�--prefixz
--install-dir�--user�	--prefix=r)rrrrrrrrrrr�make_setuptools_develop_argsLs	�r")r�egg_info_dirrr	cCs*t||d�}|dg7}|r&|d|g7}|S)N)r�egg_infoz
--egg-baser)rr#rrrrr�make_setuptools_egg_info_argsls

r%)rrr�record_filename�rootr�
header_dirrrr�	pycompiler	cCs�|r|rJ�|r|rJ�t|||	dd�}|dd|g7}|dg7}|durT|d|g7}|durh|d|g7}|dur||d|g7}|r�|d	d
g7}|
r�|dg7}n
|dg7}|r�|d
|g7}||7}|S)NT)rrr�installz--recordz#--single-version-externally-managedz--rootrz--homer r!z	--compilez--no-compilez--install-headersr)rrrr&r'rr(rrrr)rrrr�make_setuptools_install_args{s2
�

r+)NFF)r
�typingrrrr�str�boolrrrr"r%r+rrrr�<module>sd	�
�����!��site-packages/pip/_internal/utils/__pycache__/compatibility_tags.cpython-39.pyc000064400000010005150564745010023722 0ustar00a

��Re�@s*dZddlZddlmZmZmZddlmZmZm	Z	m
Z
mZmZm
Z
mZe�d�Zeedfed�dd	�Zeeed
�dd�Zeeed
�d
d�Zeeed
�dd�Zeeeeeed�dd�Zeed�dd�Zdeeeeed�dd�Zdeeeeeeeeeeeed�dd�ZdS)z3Generate and work with PEP 425 Compatibility Tags.
�N)�List�Optional�Tuple)�
PythonVersion�Tag�compatible_tags�cpython_tags�generic_tags�interpreter_name�interpreter_version�
mac_platformsz(.+)_(\d+)_(\d+)_(.+).)�version_info�returncCsd�tt|dd���S)N��)�join�map�str)r
�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/compatibility_tags.py�version_info_to_nodotsr)�archrcsRt�|�}|rH|��\�}}}t|�t|�f}�fdd�t||�D�}n|g}|S)Ncs$g|]}d��|td�d���qS)z{}_{}�macosx_N)�format�len)�.0r��namerr�
<listcomp>s�z"_mac_platforms.<locals>.<listcomp>)�
_osx_arch_pat�match�groups�intr)rr �major�minor�actual_arch�mac_version�archesrrr�_mac_platformss

�r(cCsj|g}|�d�\}}}|dkrL|dvrf|�d||�|�d||�n|dkrf|�d||�|S)N�_�
manylinux2014>�i686�x86_64�
manylinux2010�
manylinux1)�	partition�append)rr'�arch_prefix�arch_sep�arch_suffixrrr�_custom_manylinux_platforms.sr4cCs@|�d�\}}}|�d�r$t|�}n|dvr6t|�}n|g}|S)Nr)�macosx)r*r-)r/�
startswithr(r4)rr1r2r3r'rrr�_get_custom_platformsCs


r7)�	platformsrcsT|sdSt��g}|D]8}|�vr$q�fdd�t|�D�}��|�|�|�q|S)Ncsg|]}|�vr|�qSrr)r�c��seenrrrX�z-_expand_allowed_platforms.<locals>.<listcomp>)�setr7�update�extend)r8�result�p�	additionsrr:r�_expand_allowed_platformsNs
rC)�versionrcCs:t|�dkr(t|d�t|dd��fSt|d�fSdS)N�r)rr")rDrrr�_get_python_version_srF)�implementationrDrcCs(|durt�}|durt�}|�|��S)N)r
r)rGrDrrr�_get_custom_interpreterfs
rH)rDr8�impl�abisrcCs~g}d}|durt|�}t||�}t|�}|p2t�dk}|rR|�t|||d��n|�t|||d��|�t|||d��|S)aVReturn a list of supported tags for each version specified in
    `versions`.

    :param version: a string version, of the form "33" or "32",
        or None. The version will be assumed to support our ABI.
    :param platform: specify a list of platforms you want valid
        tags for, or None. If None, use the local system platform.
    :param impl: specify the exact implementation you want valid
        tags for, or None. If None, use the local interpreter impl.
    :param abis: specify a list of abis you want valid
        tags for, or None. If None, use the local interpreter abi.
    N�cp)�python_versionrJr8)�interpreterrJr8)rLrMr8)rFrHrCr
r?rr	r)rDr8rIrJ�	supportedrLrM�
is_cpythonrrr�
get_supportedps<
������rP)NN)NNNN)�__doc__�re�typingrrrZpip._vendor.packaging.tagsrrrrr	r
rr�compilerr"rrr(r4r7rCrFrHrPrrrr�<module>s4(
 ���

�site-packages/pip/_internal/utils/__pycache__/filesystem.cpython-39.pyc000064400000012070150564745010022223 0ustar00a

��Re�@s�ddlZddlZddlZddlZddlZddlZddlZddlmZddl	m
Z
ddlmZm
Z
mZmZmZmZddlmZmZmZddlmZddlmZeed�d	d
�Zeedd�dd
�Zeed�dd�Zeeeee
d�dd��Zeded�ed�d�Z e ej!�Z!eed�dd�Z"eed�dd�Z#eeeed�dd�Z$eee%e&fd�dd�Z'eed�d d!�Z(eee%e&fd�d"d#�Z)eed�d$d%�Z*dS)&�N)�contextmanager)�NamedTemporaryFile)�Any�BinaryIO�Iterator�List�Union�cast)�retry�stop_after_delay�
wait_fixed)�get_path_uid)�format_size)�path�returncCs�tjdksttd�sdStj�|�s(J�d}||kr�tj�|�r�t��dkrvzt|�}Wnt	ylYdS0|dkSt�
|tj�Sq,|tj�|�}}q,dS)N�win32�geteuidTrF)
�sys�platform�hasattr�osr�isabs�lexistsrr
�OSError�access�W_OK�dirname)r�previous�path_uid�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/filesystem.py�check_path_ownersr!)�src�destrcCsnzt�||�WnXtyh||fD]:}zt|�}WntyHYq&0|r&t�d|�d���q&�Yn0dS)z�Wrap shutil.copy2() but map errors copying socket files to
    SpecialFileError as expected.

    See also https://bugs.python.org/issue37700.
    �`z
` is a socketN)�shutil�copy2r�	is_socket�SpecialFileError)r"r#�f�is_socket_filerrr �copy2_fixed-sr+cCst�t�|�j�S�N)�stat�S_ISSOCKr�lstat�st_mode�rrrr r'Dsr')r�kwargsrcks�tfdtj�|�tj�|�dd�|���T}tt|�}z |VW|��t�|�	��n|��t�|�	��0Wd�n1s�0YdS)a%Return a file-like object pointing to a tmp file next to path.

    The file is created securely and is ensured to be written to disk
    after the context reaches its end.

    kwargs will be passed to tempfile.NamedTemporaryFile to control
    the way the temporary file will be opened.
    Fz.tmp)�delete�dir�prefix�suffixN)
rrrr�basenamer	r�flush�fsync�fileno)rr2r)�resultrrr �adjacent_tmp_fileHs 


��
�r<T�g�?)�reraise�stop�waitcCsHtj�|�s(tj�|�}||kr"q(|}qtjdkr@t�|tj�St|�S)zgCheck if a directory is writable.

    Uses os.access() on POSIX, tries creating files on Windows.
    �posix)rr�isdirr�namerr�_test_writable_dir_win)r�parentrrr �test_writable_diris
rFc	s�d}d�td�D]�}|d��fdd�td�D��}tj�||�}zt�|tjtjBtjB�}Wn&tyrYqt	y�YdS0t�
|�t�|�d	Sqtd
��dS)N�(accesstest_deleteme_fishfingers_custard_�$abcdefghijklmnopqrstuvwxyz0123456789�
�c3s|]}t���VqdSr,)�random�choice)�.0�_��alphabetrr �	<genexpr>��z)_test_writable_dir_win.<locals>.<genexpr>�FTz3Unexpected condition testing for writable directory)
�range�joinrr�open�O_RDWR�O_CREAT�O_EXCL�FileExistsError�PermissionError�close�unlinkr)rr7rNrC�file�fdrrOr rD{s 


rD)r�patternrcsBg}t�|�D].\�}}t�||�}|��fdd�|D��q|S)z�Returns a list of absolute paths of files beneath path, recursively,
    with filenames which match the UNIX-style shell glob pattern.c3s|]}tj��|�VqdSr,)rrrU)rMr)��rootrr rQ�rRzfind_files.<locals>.<genexpr>)r�walk�fnmatch�filter�extend)rr`r;rN�files�matchesrrar �
find_files�s
ricCstj�|�rdStj�|�S)Nr)rr�islink�getsizer1rrr �	file_size�srlcCstt|��Sr,)rrlr1rrr �format_file_size�srmcCsBd}t�|�D].\}}}|D]}tj�||�}|t|�7}qq|S)Ng)rrcrrUrl)r�sizerb�_dirsrg�filename�	file_pathrrr �directory_size�srrcCstt|��Sr,)rrrr1rrr �format_directory_size�srs)+rdr�os.pathrKr%r-r�
contextlibr�tempfiler�typingrrrrrr	�pip._vendor.tenacityr
rr�pip._internal.utils.compatr
�pip._internal.utils.miscr�str�boolr!r+r'r<�_replace_retry�replacerFrDri�int�floatrlrmrrrsrrrr �<module>s4 

	site-packages/pip/_internal/utils/__pycache__/hashes.cpython-39.pyc000064400000012151150564745010021312 0ustar00a

��Re��@s�ddlZddlmZmZmZmZmZddlmZm	Z	m
Z
ddlmZer`ddlm
Z
ddlmZdZgd�ZGd	d
�d
�ZGdd�de�ZdS)
�N)�
TYPE_CHECKING�BinaryIO�Dict�Iterator�List)�HashMismatch�HashMissing�InstallationError)�read_chunks)�_Hash)�NoReturn�sha256)r
�sha384�sha512c@s�eZdZdZd#eeeefdd�dd�Zddd�dd�Ze	e
d	�d
d��Zeeed�d
d�Z
eedd�dd�Zeedfdd�dd�Zedd�dd�Zedd�dd�Zed	�dd�Zeed�dd �Ze
d	�d!d"�ZdS)$�HasheszaA wrapper that builds multiple hashes at once and checks them against
    known-good values

    N)�hashes�returncCs4i}|dur*|��D]\}}t|�||<q||_dS)zo
        :param hashes: A dict of algorithm names pointing to lists of allowed
            hex digests
        N)�items�sorted�_allowed)�selfr�allowed�alg�keys�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/hashes.py�__init__s
zHashes.__init__)�otherrcsbt|t�stS|s�S�s|Si}|j��D],\�}��jvr@q,��fdd�|D�|�<q,t|�S)Ncsg|]}|�j�vr|�qSr)r)�.0�v�rrrr�
<listcomp>;�z"Hashes.__and__.<locals>.<listcomp>)�
isinstancer�NotImplementedrr)rr�new�valuesrr r�__and__+s

zHashes.__and__�rcCstdd�|j��D��S)Ncss|]}t|�VqdS�N)�len)r�digestsrrr�	<genexpr>@r"z&Hashes.digest_count.<locals>.<genexpr>)�sumrr&�rrrr�digest_count>szHashes.digest_count)�	hash_name�
hex_digestrcCs||j�|g�vS)z/Return whether the given hex digest is allowed.)r�get)rr0r1rrr�is_hash_allowedBszHashes.is_hash_allowed)�chunksrc
Cs�i}|j��D]<}zt�|�||<WqttfyHtd|����Yq0q|D]}|��D]}|�|�q\qP|�	�D] \}}|�
�|j|vrvdSqv|�|�dS)z�Check good hashes against ones built from iterable of chunks of
        data.

        Raise HashMismatch if none match.

        zUnknown hash name: N)rr�hashlibr%�
ValueError�	TypeErrorr	r&�updater�	hexdigest�_raise)rr4�gotsr0�chunk�hash�gotrrr�check_against_chunksFszHashes.check_against_chunksrr�r;rcCst|j|��dSr))rr�rr;rrrr:]sz
Hashes._raise)�filercCs|�t|��S)zaCheck good hashes against a file-like object

        Raise HashMismatch if none match.

        )r?r
)rrBrrr�check_against_file`szHashes.check_against_file)�pathrcCs8t|d��}|�|�Wd�S1s*0YdS)N�rb)�openrC)rrDrBrrr�check_against_pathhszHashes.check_against_pathcCs
t|j�S)z,Return whether I know any known-good hashes.)�boolrr.rrr�__bool__lszHashes.__bool__cCst|t�stS|j|jkSr))r#rr$r)rrrrr�__eq__ps
z
Hashes.__eq__cCs"td�tdd�|j��D����S)N�,css*|]"\}}|D]}d�||f�VqqdS)�:N)�join)rr�digest_list�digestrrrr,xs�z"Hashes.__hash__.<locals>.<genexpr>)r=rMrrrr.rrr�__hash__us���zHashes.__hash__)N)�__name__�
__module__�__qualname__�__doc__r�strrrr'�property�intr/rHr3r�bytesr?r:rrCrGrI�objectrJrPrrrrrsrcs>eZdZdZdd��fdd�Zeedfdd�d	d
�Z�ZS)�
MissingHashesz�A workalike for Hashes used when we're missing a hash for a requirement

    It computes the actual hash of the requirement and raises a HashMissing
    exception showing it to the user.

    Nr(cst�jtgid�dS)z!Don't offer the ``hashes`` kwarg.)rN)�superr�
FAVORITE_HASHr.��	__class__rrr�szMissingHashes.__init__rrr@cCst|t����dSr))rr\r9rArrrr:�szMissingHashes._raise)	rQrRrSrTrrrUr:�
__classcell__rrr]rrZ�srZ)r5�typingrrrrr�pip._internal.exceptionsrrr	�pip._internal.utils.miscr
rrr\�
STRONG_HASHESrrZrrrr�<module>shsite-packages/pip/_internal/utils/__pycache__/wheel.cpython-39.pyc000064400000014251150564745010021146 0ustar00a

��Re�@s0dZddlZddlmZddlmZddlmZmZddl	m
Z
mZddlm
Z
ddlmZmZdd	lmZdd
lmZdZe�e�ZGdd
�d
e�Zeeeed�dd�Zeeeeefd�dd�Zeeed�dd�Zeeed�dd�Zeeed�dd�Z eee!dfd�dd �Z"ee!dfedd!�d"d#�Z#dS)$z0Support functions for working with wheel files.
�N)�Message)�Parser)�Dict�Tuple)�
BadZipFile�ZipFile)�canonicalize_name)�DistInfoDistribution�Distribution)�UnsupportedWheel)�DictMetadata)�rcsFeZdZdZeeefedd��fdd�Zeed��fdd�Z�Z	S)	�
WheelMetadatazaMetadata provider that maps metadata decoding exceptions to our
    internal exception type.
    N)�metadata�
wheel_name�returncst��|�||_dS)N)�super�__init__�_wheel_name)�selfrr��	__class__���/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/wheel.pyrszWheelMetadata.__init__)�namerc
sNzt��|�WStyH}z"td|j�d|����WYd}~n
d}~00dS)NzError decoding metadata for �: )r�get_metadata�UnicodeDecodeErrorrr)rr�errrrs�zWheelMetadata.get_metadata)
�__name__�
__module__�__qualname__�__doc__r�str�bytesrr�
__classcell__rrrrrsr)�	wheel_zipr�locationrc
s�t||�\�}�fdd�|��D�}i}|D]`}|�dd�\}}zt||�||<Wq,ty�}z td�|t|����WYd}~q,d}~00q,t||�}	t||	|d�S)zaGet a pkg_resources distribution given a wheel.

    :raises UnsupportedWheel: on any errors
    cs g|]}|���d��r|�qS)�/)�
startswith��.0�p��info_dirrr�
<listcomp>2�z8pkg_resources_distribution_for_wheel.<locals>.<listcomp>r(r
�{} has an invalid wheel, {}N)r'r�project_name)	�parse_wheel�namelist�split�read_wheel_metadata_filer�formatr#rr	)
r&rr'�_�metadata_files�
metadata_text�path�
metadata_namerrrr-r�$pkg_resources_distribution_for_wheel)s,
r=)r&rrc
Cslz t||�}t||�}t|�}Wn8tyX}z td�|t|����WYd}~n
d}~00t||�||fS)z�Extract information from the provided wheel, ensuring it meets basic
    standards.

    Returns the name of the .dist-info directory and the parsed WHEEL metadata.
    r1N)�wheel_dist_info_dir�wheel_metadata�
wheel_versionrr7r#�check_compatibility)r&rr.r�versionrrrrr3Bs

*
r3)�sourcerrcCs�dd�|��D�}dd�|D�}|s,td��t|�dkrLtd�d�|����|d	}t|�}t|�}|�|�s~td
�||���|S)z�Returns the name of the contained .dist-info directory.

    Raises AssertionError or UnsupportedWheel if not found, >1 found, or
    it doesn't match the provided name.
    cSsh|]}|�dd�d�qS)r(r
r)r5r*rrr�	<setcomp>[r0z&wheel_dist_info_dir.<locals>.<setcomp>cSsg|]}|�d�r|�qS)z
.dist-info)�endswith)r+�srrrr/]r0z'wheel_dist_info_dir.<locals>.<listcomp>z.dist-info directory not foundr
z)multiple .dist-info directories found: {}z, rz2.dist-info directory {!r} does not start with {!r})r4r�lenr7�joinrr))rCr�subdirs�	info_dirsr.�
info_dir_name�canonical_namerrrr>Ts$�
��r>)rCr;rc
CsPz|�|�WStttfyJ}z td|�d|����WYd}~n
d}~00dS)Nzcould not read z file: )�readr�KeyError�RuntimeErrorr)rCr;rrrrr6usr6)rC�
dist_info_dirrc
Csf|�d�}t||�}z|��}Wn8tyX}z td|�d|����WYd}~n
d}~00t��|�S)ziReturn the WHEEL metadata of an extracted wheel, if possible.
    Otherwise, raise UnsupportedWheel.
    z/WHEELzerror decoding rN)r6�decoderrr�parsestr)rCrPr;�wheel_contents�
wheel_textrrrrr?~s

*r?.)�
wheel_datarcCs\|d}|durtd��|��}zttt|�d���WStyVtd|����Yn0dS)zbGiven WHEEL metadata, return the parsed Wheel-Version.
    Otherwise, raise UnsupportedWheel.
    z
Wheel-VersionNzWHEEL is missing Wheel-Version�.zinvalid Wheel-Version: )r�strip�tuple�map�intr5�
ValueError)rU�version_textrBrrrr@�sr@)rBrrc	CsR|dtdkr.td�|d�tt|�����n |tkrNt�dd�tt|���dS)a�Raises errors or warns if called with an incompatible Wheel-Version.

    pip should refuse to install a Wheel-Version that's a major series
    ahead of what it's compatible with (e.g 2.0 > 1.1); and warn when
    installing a version only minor version ahead (e.g 1.2 > 1.1).

    version: a 2-tuple representing a Wheel-Version (Major, Minor)
    name: name of wheel or package to raise exception about

    :raises UnsupportedWheel: when an incompatible Wheel-Version is given
    rzB{}'s Wheel-Version ({}) is not compatible with this version of piprVz*Installing from a newer Wheel-Version (%s)N)�VERSION_COMPATIBLErr7rHrYr#�logger�warning)rBrrrrrA�s���rA)$r"�logging�
email.messager�email.parserr�typingrr�zipfilerr�pip._vendor.packaging.utilsr�pip._vendor.pkg_resourcesr	r
�pip._internal.exceptionsr�!pip._internal.utils.pkg_resourcesrr]�	getLoggerrr^rr#r=r3r>r$r6r?rZr@rArrrr�<module>s*
�!	site-packages/pip/_internal/utils/__pycache__/direct_url_helpers.cpython-39.pyc000064400000004110150564745010023711 0ustar00a

��Re��@s�ddlmZddlmZmZmZmZddlmZddl	m
Z
ddlmZee
e
d�dd�Ze
ed	�d
d�Zdeee
eed�dd�ZdS)�)�Optional)�ArchiveInfo�	DirectUrl�DirInfo�VcsInfo)�Link)�path_to_url)�vcs)�
direct_url�name�returncCs�|��|d}g}t|jt�r>|d�|jj|j|jj�7}nHt|jt�rl||j7}|jj	r�|�
|jj	�nt|jt�s|J�||j7}|jr�|�
d|j�|r�|dd�
|�7}|S)z0Convert a DirectUrl to a pip requirement string.z @ z{}+{}@{}z
subdirectory=�#�&)�validate�
isinstance�infor�formatr	�url�	commit_idr�hash�appendr�subdirectory�join)r
r�requirement�	fragments�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/direct_url_helpers.py�%direct_url_as_pep440_direct_reference	s$�

r)�
source_dirrcCstt|�tdd�d�S)NT)�editable)rr)rrr)rrrr�direct_url_for_editable s�r NF)�linkr�link_is_in_wheel_cacherc
Cs�|jrlt�|j�}|sJ�|�|j�\}}}|r>|s8J�|}n|sFJ�|�|�}t|t|j	||d�|j
d�S|��r�t|jt�|j
d�Sd}|j
}	|	r�|	�d|j��}t|jt|d�|j
d�SdS)N)r	r�requested_revision)rrr�=)r)�is_vcsr	�get_backend_for_scheme�scheme�get_url_rev_and_auth�url_without_fragment�get_revisionrrr�subdirectory_fragment�is_existing_dirr�	hash_namerr)
r!rr"�vcs_backendrr#�_rrr-rrr�direct_url_from_link'sF�

��	��r0)NF)�typingr�pip._internal.models.direct_urlrrrr�pip._internal.models.linkr�pip._internal.utils.urlsr�pip._internal.vcsr	�strrr �boolr0rrrr�<module>s�
�site-packages/pip/_internal/utils/__pycache__/packaging.cpython-39.pyc000064400000005162150564745010021767 0ustar00a

��Re��@s�ddlZddlZddlmZddlmZddlmZmZddl	m
Z
ddlmZm
Z
ddlmZddlmZdd	lmZdd
lmZe�e�Zeeeedfed�d
d�Zeed�dd�Zeed�dd�Zejdd�eed�dd��Z dS)�N)�Message)�
FeedParser)�Optional�Tuple)�
pkg_resources)�
specifiers�version��Requirement)�Distribution)�NoneMetadataError)�display_path.)�requires_python�version_info�returncCs4|durdSt�|�}t�d�tt|���}||vS)a�
    Check if the given Python version matches a "Requires-Python" specifier.

    :param version_info: A 3-tuple of ints representing a Python
        major-minor-micro version to check (e.g. `sys.version_info[:3]`).

    :return: `True` if the given Python version satisfies the requirement.
        Otherwise, return `False`.

    :raises InvalidSpecifier: If `requires_python` has an invalid format.
    NT�.)r�SpecifierSetr�parse�join�map�str)rr�requires_python_specifier�python_version�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/packaging.py�check_requires_pythons

r)�distrcCs�d}t|tj�r&|�|�r&|�|�}n0|�d�r@d}|�|�}nt�dt|j��d}|durht	||��t
�}|�|�|��S)z�
    :raises NoneMetadataError: if the distribution reports `has_metadata()`
        True but `get_metadata()` returns None.
    �METADATAzPKG-INFOzNo metadata found in %s�N)
�
isinstancer�DistInfoDistribution�has_metadata�get_metadata�logger�warningr
�locationrr�feed�close)r�
metadata_name�metadata�feed_parserrrrr")s�


r"cCs2|�d�r.|�d�D]}|��r|��SqdS)N�	INSTALLERr)r!�get_metadata_lines�strip)r�linerrr�
get_installerDs

r/i)�maxsize)�
req_stringrcCst|�S)z5Construct a packaging.Requirement object with cachingr	)r1rrr�get_requirementLsr2)!�	functools�logging�
email.messager�email.parserr�typingrr�pip._vendorrZpip._vendor.packagingrrZ"pip._vendor.packaging.requirementsr
�pip._vendor.pkg_resourcesr�pip._internal.exceptionsr�pip._internal.utils.miscr
�	getLogger�__name__r#r�int�boolrr"r/�	lru_cacher2rrrr�<module>s$
�
site-packages/pip/_internal/utils/__pycache__/encoding.cpython-39.pyc000064400000002473150564745010021633 0ustar00a

��Re��@s�UddlZddlZddlZddlZddlmZmZejdfejdfej	dfej
dfejdfejdfej
d	fgZeeeefed
<e�d�Zeed�d
d�ZdS)�N)�List�Tuplezutf-8zutf-16z	utf-16-bez	utf-16-lezutf-32z	utf-32-bez	utf-32-le�BOMSscoding[:=]\s*([-\w.]+))�data�returncCs�tD],\}}|�|�r|t|�d��|�Sq|�d�dd�D]T}|dd�dkrDt�|�rDt�|�}|dusxJ�|��d�d�}|�|�SqD|�t�	d�p�t
���S)	z�Check a bytes string for a BOM to correctly detect the encoding

    Fallback to locale.getpreferredencoding(False) like open() on Python3N�
�r��#�asciiF)r�
startswith�len�decode�split�ENCODING_RE�search�groups�locale�getpreferredencoding�sys�getdefaultencoding)r�bom�encoding�line�result�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/encoding.py�auto_decodes

�r)�codecsr�rer�typingrr�BOM_UTF8�	BOM_UTF16�BOM_UTF16_BE�BOM_UTF16_LE�	BOM_UTF32�BOM_UTF32_BE�BOM_UTF32_LEr�bytes�str�__annotations__�compilerrrrrr�<module>s
�

site-packages/pip/_internal/utils/__pycache__/distutils_args.cpython-39.pyc000064400000002157150564745010023104 0ustar00a

��Re��@sXddlmZddlmZddlmZmZgd�Zee�Zee	ee	e	fd�dd�Z
dS)	�)�DistutilsArgError)�FancyGetopt)�Dict�List))zexec-prefix=N�)zhome=Nr)z
install-base=Nr)z
install-data=Nr)zinstall-headers=Nr)zinstall-lib=Nr)zinstall-platlib=Nr)zinstall-purelib=Nr)zinstall-scripts=Nr)zprefix=Nr)zroot=Nr)�userNr)�args�returnc	CsHi}|D]:}ztj|gd�\}}Wnty4Yq0|�|j�q|S)z~Parse provided arguments, returning an object that has the
    matched arguments.

    Any unknown arguments are ignored.
    )r)�_distutils_getopt�getoptr�update�__dict__)r�result�arg�_�match�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/distutils_args.py�parse_distutils_argssrN)�distutils.errorsr�distutils.fancy_getoptr�typingrr�_optionsr
�strrrrrr�<module>s
site-packages/pip/_internal/utils/__pycache__/models.cpython-39.pyc000064400000004062150564745010021324 0ustar00a

��Re��@s2dZddlZddlmZmZmZGdd�d�ZdS)zUtilities for defining models
�N)�Any�Callable�Typec@s�eZdZdZddgZeeddd�dd�Zed�d	d
�Z	ee
d�dd
�Zee
d�dd�Zee
d�dd�Z
ee
d�dd�Zee
d�dd�Zeeeege
fe
d�dd�ZdS)�KeyBasedCompareMixinz7Provides comparison capabilities that is based on a key�_compare_key�_defining_classN)�key�defining_class�returncCs||_||_dS�N)rr)�selfrr	�r
��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/models.py�__init__
szKeyBasedCompareMixin.__init__)r
cCs
t|j�Sr)�hashr)rr
r
r�__hash__szKeyBasedCompareMixin.__hash__)�otherr
cCs|�|tj�Sr)�_compare�operator�__lt__�rrr
r
rrszKeyBasedCompareMixin.__lt__cCs|�|tj�Sr)rr�__le__rr
r
rrszKeyBasedCompareMixin.__le__cCs|�|tj�Sr)rr�__gt__rr
r
rrszKeyBasedCompareMixin.__gt__cCs|�|tj�Sr)rr�__ge__rr
r
rrszKeyBasedCompareMixin.__ge__cCs|�|tj�Sr)rr�__eq__rr
r
rr szKeyBasedCompareMixin.__eq__)r�methodr
cCst||j�stS||j|j�Sr)�
isinstancer�NotImplementedr)rrrr
r
rr#szKeyBasedCompareMixin._compare)�__name__�
__module__�__qualname__�__doc__�	__slots__rrr�intr�boolrrrrrrrr
r
r
rrsr)r!r�typingrrrrr
r
r
r�<module>ssite-packages/pip/_internal/utils/__pycache__/misc.cpython-39.pyc000064400000051521150564745010020776 0ustar00a

��Re*Q�@sddlZddlZddlZddlZddlZddlZddlZddlZddlZddl	Z	ddl
Z
ddlZddlm
Z
ddlmZmZmZddlmZddlmZmZmZmZmZmZmZmZmZmZmZm Z m!Z!ddl"m#Z#ddl$m%Z%m&Z&m'Z'ddl(m)Z)dd	l*m+Z+dd
l,m-Z-m.Z.m/Z/ddl,m0Z0ddl1m2Z2dd
l3m4Z4ddl5m6Z6gd�Z7e�8e9�Z:e d�Z;eee<e<efZ=ee>e>e>fZ?ee@eee@ee@ffZAe@d�dd�ZBee>dfee>e>e>fd�dd�ZCe@dd�dd�ZDe@d�dd�ZEe%de&d�e'd�d �d�e@eFdd"�d#d$��ZGedefe@e=dd%�d&d'�ZHe@e@d�d(d)�ZId�e@e@e@d+�d,d-�ZJe@ee@e@d.�d/d0�ZKe@dd1�d2d3�ZLe@ee@e@d.�d4d5�ZMe@e@d1�d6d7�ZNe@e@d1�d8d9�ZOe@e>d:�d;d<�ZPeQe@d=�d>d?�ZReeeeee@ee>fd@�dAdB�ZSe@eFd�dCdD�ZTejUfee>eeVdE�dFdG�ZWd�e@eFe@dH�dIdJ�ZXe@ee@e@fd�dKdL�ZYe@e@ddM�dNdO�ZZe@eFd�dPdQ�Z[e#eFdR�dSdT�Z\e#eFdR�dUdV�Z]e#eFdR�dWdX�Z^dYdZ�Z_e@ee#d[�d\d]�Z`e#e@dR�d^d_�Zaeedd`�dadb�ZbGdcdd�dde
�Zcejde@eecde�dfdg��Zeeecd�dhdi�Zfeecd�djdk�Zgeeeedl�dmdn�Zhe@ee>e@do�dpdq�Zid�e@e@e@ds�dtdu�Zje@ee@ee>fdv�dwdx�Zke@eAdv�dydz�Zle@e@dv�d{d|�Zme@ee@geedffee@eAfd}�d~d�Zne@eAdv�d�d��Zoe@ee@dv�d�d��Zpe@ee@e@ee@e@ffd��d�d��Zqe@e@d��d�d��Zre@e@d��d�d��ZsGd�d��d��Zte@etd��d�d��Zue@etd��d�d��ZveFdd��d�d��ZweFd�d�d��Zxd�e@e>eee>fd��d�d��ZyeFd�d�d��Zzeeeeeefd��d�d��Z{ee;geFfee;eee;ee;fd��d�d��Z|dS)��N)�StringIO)�filterfalse�tee�zip_longest)�
TracebackType)
�Any�BinaryIO�Callable�ContextManager�Iterable�Iterator�List�Optional�TextIO�Tuple�Type�TypeVar�cast��Distribution)�retry�stop_after_delay�
wait_fixed)�__version__)�CommandError)�get_major_minor_version�
site_packages�	user_site)�
get_scheme)�WINDOWS)�egg_link_path_from_location)�running_under_virtualenv)
�rmtree�display_path�
backup_dir�ask�splitext�format_size�is_installable_dir�normalize_path�renames�get_prog�captured_stdout�
ensure_dir�remove_auth_from_url�T��returncCs4tj�tj�t�dd�}tj�|�}d�t|t��S)Nz..zpip {} from {} (python {}))	�os�path�join�dirname�__file__�abspath�formatrr)�pip_pkg_dir�r:��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/misc.py�get_pip_versionGs�r<.)�py_version_infor1cCsDt|�dkr"|dt|�d7}nt|�dkr:|dd�}td|�S)ax
    Convert a tuple of ints representing a Python version to one of length
    three.

    :param py_version_info: a tuple of ints representing a Python version,
        or None to specify no version. The tuple can have any length.

    :return: a tuple of length three if `py_version_info` is non-None.
        Otherwise, return `py_version_info` unchanged (i.e. None).
    �)rN�VersionInfo)�lenr)r=r:r:r;�normalize_version_infoRs
rA)r3r1c
CsRzt�|�Wn>tyL}z&|jtjkr8|jtjkr8�WYd}~n
d}~00dS)z os.path.makedirs without EEXIST.N)r2�makedirs�OSError�errno�EEXIST�	ENOTEMPTY)r3�er:r:r;r-es
r-c
CsPz2tj�tjd�}|dvr*tj�d�WS|WSWntttfyJYn0dS)Nr)z__main__.pyz-cz -m pip�pip)	r2r3�basename�sys�argv�
executable�AttributeError�	TypeError�
IndexError)�progr:r:r;r+os
r+Tr>g�?)�reraise�stop�waitF)�dir�
ignore_errorsr1cCstj||td�dS)N)rU�onerror)�shutilr"�rmtree_errorhandler)rTrUr:r:r;r"}sr")�funcr3�exc_infor1cCsRzt�|�jtj@}Wnty,YdS0|rLt�|tj�||�dS�dS)z�On Windows, the files in .svn are read-only, so when rmtree() tries to
    remove them, an exception is thrown.  We catch that here, remove the
    read-only attribute, and hopefully continue without problems.N)r2�stat�st_mode�S_IWRITErC�chmod)rYr3rZ�has_attr_readonlyr:r:r;rX�srXcCsFtj�tj�|��}|�t��tjj�rBd|tt���d�}|S)zTGives the display value for a given path, making it relative to cwd
    if possible.�.N)r2r3�normcaser7�
startswith�getcwd�sepr@�r3r:r:r;r#�sr#�.bak)rT�extr1cCs6d}|}tj�||�r.|d7}|t|�}q||S)z\Figure out the name of a directory to back up the given dir to
    (adding .bak, .bak2, etc)�)r2r3�exists�str)rTrg�n�	extensionr:r:r;r$�sr$)�message�optionsr1cCs2tj�dd���D]}||vr|Sqt||�S)N�PIP_EXISTS_ACTION�)r2�environ�get�splitr%)rmrn�actionr:r:r;�ask_path_exists�s
ru)rmr1cCstj�d�rtd|����dS)z&Raise an error if no input is allowed.�PIP_NO_INPUTz5No input was expected ($PIP_NO_INPUT set); question: N)r2rqrr�	Exception�rmr:r:r;�_check_no_input�s�rycCsFt|�t|�}|����}||vr<td�|d�|���q|SqdS)z@Ask the message interactively, with the given possible responsesz>Your response ({!r}) was not one of the expected responses: {}z, N)ry�input�strip�lower�printr8r4)rmrn�responser:r:r;r%�s
��r%cCst|�t|�S)zAsk for input interactively.)ryrzrxr:r:r;�	ask_input�srcCst|�t�|�S)z!Ask for a password interactively.)ry�getpassrxr:r:r;�ask_password�sr�)�valr1cCs2|��}|dvrdS|dvr dStd|����dS)z�Convert a string representation of truth to true (1) or false (0).

    True values are 'y', 'yes', 't', 'true', 'on', and '1'; false values
    are 'n', 'no', 'f', 'false', 'off', and '0'.  Raises ValueError if
    'val' is anything else.
    )�y�yes�t�true�on�1rh)rk�no�f�false�off�0rzinvalid truth value N)r|�
ValueError)r�r:r:r;�	strtobool�sr�)�bytesr1cCs\|dkrd�|dd�S|dkr4d�t|d��S|dkrJd�|d�Sd�t|��SdS)	Ni@Bz	{:.1f} MBg@�@i�i'z{} kBz	{:.1f} kBz{} bytes)r8�int)r�r:r:r;r'�sr')�rowsr1cs@dd�|D�}dd�t|ddi�D���fdd�|D�}|�fS)z�Return a list of formatted rows and a list of column sizes.

    For example::

    >>> tabulate([['foobar', 2000], [0xdeadbeef]])
    (['foobar     2000', '3735928559'], [10, 4])
    cSsg|]}ttt|���qSr:)�tuple�maprj��.0�rowr:r:r;�
<listcomp>��ztabulate.<locals>.<listcomp>cSsg|]}ttt|���qSr:)�maxr�r@)r��colr:r:r;r��r��	fillvaluerpcs$g|]}d�ttj|������qS)� )r4r�rj�ljust�rstripr���sizesr:r;r��r�)r)r��tabler:r�r;�tabulate�sr�cCsHtj�|�sdStj�tj�|d��r*dStj�tj�|d��rDdSdS)atIs path is a directory containing pyproject.toml or setup.py?

    If pyproject.toml exists, this is a PEP 517 project. Otherwise we look for
    a legacy setuptools layout by identifying setup.py. We don't check for the
    setup.cfg because using it without setup.py is only available for PEP 517
    projects, which are already covered by the pyproject.toml check.
    Fzpyproject.tomlTzsetup.py)r2r3�isdir�isfiler4rer:r:r;r(�sr()�file�sizer1ccs|�|�}|sq|VqdS)z7Yield pieces of data from a file-like object until EOF.N)�read)r�r��chunkr:r:r;�read_chunkss
r�)r3�resolve_symlinksr1cCs6tj�|�}|rtj�|�}ntj�|�}tj�|�S)zN
    Convert a path to its canonical, case-normalized, absolute version.

    )r2r3�
expanduser�realpathr7ra)r3r�r:r:r;r)s
r)cCs@t�|�\}}|���d�r8|dd�|}|dd�}||fS)z,Like os.path.splitext, but take off .tar tooz.tar���N)�	posixpathr&r|�endswith)r3�basergr:r:r;r&$s
r&)�old�newr1cCsxtj�|�\}}|r.|r.tj�|�s.t�|�t�||�tj�|�\}}|rt|rtzt�|�WntyrYn0dS)z7Like os.renames(), but handles renaming across devices.N)	r2r3rsrirBrW�move�
removedirsrC)r�r��head�tailr:r:r;r*-s
r*cCst�s
dS|�ttj��S)z�
    Return True if path is within sys.prefix, if we're running in a virtualenv.

    If we're not in a virtualenv, all paths are considered "local."

    Caution: this function assumes the head of path has been normalized
    with normalize_path.
    T)r!rbr)rJ�prefixrer:r:r;�is_local>s	r�)�distr1cCstt|��S)z�
    Return True if given Distribution object is installed locally
    (i.e. within current virtualenv).

    Always True if we're not in a virtualenv.

    )r��
dist_location�r�r:r:r;�
dist_is_localLsr�cCst|��tt��S)zF
    Return True if given Distribution is installed in user site.
    )r�rbr)rr�r:r:r;�dist_in_usersiteWsr�cCst|��tt��S)z[
    Return True if given Distribution is installed in
    sysconfig.get_python_lib().
    )r�rbr)rr�r:r:r;�dist_in_site_packages^sr�cCs*tt|��}|�ttd�j�d�d��S)zf
    Return True if given Distribution is installed in
    path matching distutils_scheme layout.
    rp�pythonr)r)r�rbr�purelibrs)r��	norm_pathr:r:r;�dist_in_install_pathfs�r�)�req_namer1cCs<ddlm}ddlm}|��|�}|dur0dSt||�jS)a%Given a requirement name, return the installed Distribution object.

    This searches from *all* distributions available in the environment, to
    match the behavior of ``pkg_resources.get_distribution()``.

    Left for compatibility until direct pkg_resources uses are refactored out.
    r)�get_default_environmentrN)�pip._internal.metadatar��$pip._internal.metadata.pkg_resourcesr�get_distributionr�_dist)r�r��_Distr�r:r:r;r�psr�cCs t|j�}|rt|�St|j�S)aO
    Get the site-packages location of this distribution. Generally
    this is dist.location, except in the case of develop-installed
    packages, where dist.location is the source code location, and we
    want to know where the egg-link file is.

    The returned location is normalized (in particular, with symlinks removed).
    )r �project_namer)�location)r��egg_linkr:r:r;r��s	
r�)�msg�argsr1cGstj|g|�R�dS�N)�logger�info)r�r�r:r:r;�write_output�sr�c@s:eZdZUdZeed<eedd�dd��Zedd��Z	dS)�
StreamWrapperN�orig_stream)r�r1cCs||_|�Sr�)r�)�clsr�r:r:r;�from_stream�szStreamWrapper.from_streamcCs|jjSr�)r��encoding��selfr:r:r;r��szStreamWrapper.encoding)
�__name__�
__module__�__qualname__r�r�__annotations__�classmethodr��propertyr�r:r:r:r;r��s

r�)�stream_namer1c
csLtt|�}tt|t�|��ztt|�VWtt||�ntt||�0dS)z�Return a context manager used by captured_stdout/stdin/stderr
    that temporarily replaces the sys stream *stream_name* with a StringIO.

    Taken from Lib/support/__init__.py in the CPython repo.
    N)�getattrrJ�setattrr�r�)r��orig_stdoutr:r:r;�captured_output�s

r�cCstd�S)z�Capture the output of sys.stdout:

       with captured_stdout() as stdout:
           print('hello')
       self.assertEqual(stdout.getvalue(), 'hello
')

    Taken from Lib/support/__init__.py in the CPython repo.
    �stdout�r�r:r:r:r;r,�s	r,cCstd�S)z 
    See captured_stdout().
    �stderrr�r:r:r:r;�captured_stderr�sr�)�
sequential�namedr1cOsDtt|tt|���fi|��}dd�|��D�}||d<tdd|�S)NcSsi|]\}}||�qSr:r:)r��key�valuer:r:r;�
<dictcomp>�r�zenum.<locals>.<dictcomp>�reverse_mapping�Enumr:)�dict�zip�ranger@�items�type)r�r��enums�reverser:r:r;�enum�sr�)�host�portr1cCs.|dur|Sd|vr d|�d�}|�d|��S)z.
    Build a netloc from a host-port pair
    N�:�[�]r:)r�r�r:r:r;�build_netloc�s
r��https)�netloc�schemer1cCs8|�d�dkr*d|vr*d|vr*d|�d�}|�d|��S)z)
    Build a full URL from a netloc.
    r���@r�r�z://)�count)rrr:r:r;�build_url_from_netloc�sr)rr1cCs t|�}tj�|�}|j|jfS)z2
    Return the host-port pair from a netloc.
    )r�urllib�parse�urlparse�hostnamer�)r�url�parsedr:r:r;�parse_netloc�srcCstd|vr|dfS|�dd�\}}d}d|vr>|�dd�\}}n
|d}}tj�|�}|durhtj�|�}|||ffS)zp
    Parse out and remove the auth information from a netloc.

    Returns: (netloc, (username, password)).
    r)NNrhNr�)�rsplitrsrr�unquote)r�auth�pw�userr:r:r;�split_auth_from_netloc�s
rcCsNt|�\}\}}|dur|S|dur.d}d}ntj�|�}d}dj|||d�S)z�
    Replace the sensitive data in a netloc with "****", if it exists.

    For example:
        - "user:pass@example.com" returns "user:****@example.com"
        - "accesstoken@example.com" returns "****@example.com"
    N�****rpz:****z{user}{password}@{netloc})r�passwordr)rrr�quoter8)rrrr:r:r;�
redact_netloc	s�r)r
�transform_netlocr1cCsJtj�|�}||j�}|j|d|j|j|jf}tj�|�}|t	d|�fS)aRTransform and replace netloc in a url.

    transform_netloc is a function taking the netloc and returning a
    tuple. The first element of this tuple is the new netloc. The
    entire tuple is returned.

    Returns a tuple containing the transformed url as item 0 and the
    original tuple returned by transform_netloc as item 1.
    r�NetlocTuple)
rr�urlsplitrrr3�query�fragment�
urlunsplitr)r
r�purl�netloc_tuple�
url_pieces�surlr:r:r;�_transform_urls

r!cCst|�Sr�)r�rr:r:r;�_get_netloc3sr#cCs
t|�fSr�)rr"r:r:r;�_redact_netloc7sr$)r
r1cCst|t�\}\}}|||fS)z�
    Parse a url into separate netloc, auth, and url with no auth.

    Returns: (url_without_auth, netloc, (username, password))
    �r!r#)r
�url_without_authrrr:r:r;�split_auth_netloc_from_url;sr'cCst|t�dS)z7Return a copy of url with 'username:password@' removed.rr%�r
r:r:r;r.Esr.cCst|t�dS)z.Replace the password in a given url with ****.r)r!r$r(r:r:r;�redact_auth_from_urlLsr)c@sJeZdZeedd�dd�Zed�dd�Zed�dd	�Zeed
�dd�Z	dS)
�
HiddenTextN)�secret�redactedr1cCs||_||_dSr�)r+r,)r�r+r,r:r:r;�__init__RszHiddenText.__init__r0cCsd�t|��S)Nz<HiddenText {!r}>)r8rjr�r:r:r;�__repr__VszHiddenText.__repr__cCs|jSr��r,r�r:r:r;�__str__YszHiddenText.__str__)�otherr1cCs t|�t|�krdS|j|jkS)NF)r�r+)r�r1r:r:r;�__eq__]szHiddenText.__eq__)
r�r�r�rjr-r.r0r�boolr2r:r:r:r;r*Qsr*)r�r1cCst|dd�S)Nrr/)r*)r�r:r:r;�
hide_valuefsr4cCst|�}t||d�S)Nr/)r)r*)r
r,r:r:r;�hide_urljsr5)�
modifying_pipr1cCszdd�tjd�djtjdd��g}|oBtoBtj�tjd�|v}|rvtjddgtjd	d�}t	d
�d�
|����dS)z�Protection of pip.exe from modification on Windows

    On Windows, any operation modifying pip should be run as:
        python -m pip ...
    zpip.exez	pip{}.exerzpip{}.{}.exeNrz-mrHrhz3To modify pip, please run the following command:
{}r�)r8rJ�version_inforr2r3rIrKrLrr4)r6�	pip_names�should_show_use_python_msg�new_commandr:r:r;�(protect_pip_from_modification_on_windowsos����r;cCstjduotj��S)zIs this console interactive?N)rJ�stdin�isattyr:r:r:r;�is_console_interactive�sr>�)r3�	blocksizer1cCsft��}d}t|d��8}t||d�D]}|t|�7}|�|�q$Wd�n1sT0Y||fS)z5Return (hash, length) for path using hashlib.sha256()r�rb)r�N)�hashlib�sha256�openr�r@�update)r3r@�h�lengthr��blockr:r:r;�	hash_file�s*rIcCs&zddl}Wnty YdS0dS)z8
    Return whether the wheel package is installed.
    rNFT)�wheel�ImportError)rJr:r:r;�is_wheel_installed�s
rL)�iterabler1cCst|�}t||�S)zb
    Return paired elements.

    For example:
        s -> (s0, s1), (s2, s3), (s4, s5), ...
    )�iterr)rMr:r:r;�pairwise�srO)�predrMr1cCs t|�\}}t||�t||�fS)z�
    Use a predicate to partition entries into false entries and true entries,
    like

        partition(is_odd, range(10)) --> 0 2 4 6 8   and  1 3 5 7 9
    )rr�filter)rPrM�t1�t2r:r:r;�	partition�s
rT)F)rf)T)r�)r?)}�
contextlibrDr�rB�io�loggingr2r�rWr[rJ�urllib.parserr�	itertoolsrrr�typesr�typingrrr	r
rrr
rrrrrr�pip._vendor.pkg_resourcesr�pip._vendor.tenacityrrrrHr�pip._internal.exceptionsr�pip._internal.locationsrrrr�pip._internal.utils.compatr�pip._internal.utils.egg_linkr �pip._internal.utils.virtualenvr!�__all__�	getLoggerr�r�r/�
BaseException�ExcInfor�r?rjrr<rAr-r+r3r"rXr#r$ruryr%rr�r��floatr'r�r(�DEFAULT_BUFFER_SIZEr�r�r)r&r*r�r�r�r�r�r�r�r�r��contextmanagerr�r,r�r�r�rrrrr!r#r$r'r.r)r*r4r5r;r>rIrLrOrTr:r:r:r;�<module>s�<
"
	(	
	

	
�"
 �site-packages/pip/_internal/utils/__pycache__/unpacking.cpython-39.pyc000064400000015151150564745010022021 0ustar00a

��Re�"�@s�dZddlZddlZddlZddlZddlZddlZddlmZm	Z	m
Z
ddlmZddlm
Z
ddlmZmZmZmZddlmZe�e�ZeeZzddlZee7ZWney�e�d�Yn0zddlZee7ZWney�e�d	�Yn0ed
�dd�Zee	ed
�dd�Z eee!d�dd�Z"eee!d�dd�Z#edd
�dd�Z$ee!d�dd�Z%d%eee!dd�dd�Z&eedd�d d!�Z'd&eee
edd"�d#d$�Z(dS)'zUtilities related archives.
�N)�Iterable�List�Optional)�ZipInfo)�InstallationError)�BZ2_EXTENSIONS�TAR_EXTENSIONS�
XZ_EXTENSIONS�ZIP_EXTENSIONS)�
ensure_dirzbz2 module is not availablezlzma module is not available)�returncCst�d�}t�|�|S)zBGet the current umask which involves having to set it temporarily.r)�os�umask)�mask�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/unpacking.py�
current_umask+s

r)�pathrcCsh|�d��d�}d|vrHd|vr4|�d�|�d�ks<d|vrH|�dd�Sd|vr\|�dd�S|dgSdS)N�/�\��)�lstrip�find�split�rrrr�split_leading_dir2s���r)�pathsrcCsDd}|D]6}t|�\}}|s"dS|dur0|}q||krdSqdS)zyReturns true if all the paths have the same leading path name
    (i.e., everything is in one subdirectory in an archive)NFT)r)r�
common_prefixr�prefix�restrrr�has_leading_dir>sr!)�	directory�targetrcCs0tj�|�}tj�|�}tj�||g�}||kS)zL
    Return true if the absolute path of target is within the directory
    )r
r�abspath�commonprefix)r"r#�
abs_directory�
abs_targetrrrr�is_within_directoryMsr(cCst�|dt�@dB�dS)zx
    Make file present at path have execute for user/group/world
    (chmod +x) is no-op on windows per python docs
    i��IN)r
�chmodrrrrr�2set_extracted_file_to_default_mode_plus_executableXsr+)�inforcCs$|jd?}t|o t�|�o |d@�S)N�r))�
external_attr�bool�stat�S_ISREG)r,�moderrr�zip_item_is_executable`s
r3T)�filename�location�flattenrc
	CsRt|�t|d�}�z.tj|dd�}t|���o2|}|��D]�}|j}|}|rZt|�d}t	j
�||�}t	j
�|�}	t
||�s�d}
t|
�|||���|�d�s�|�d�r�t|�q<t|	�|�|�}zTt|d��}t�||�Wd	�n1s�0YW|��t|��r6t|�q<|��t|��r4t|�0q<W|��n
|��0d	S)
a�
    Unzip the file (with path `filename`) to the destination `location`.  All
    files are written based on system defaults and umask (i.e. permissions are
    not preserved), except that regular file members with any execute
    permissions (user, group, or world) have "chmod +x" applied after being
    written. Note that for windows, any execute changes using os.chmod are
    no-ops per the python docs.
    �rbT)�
allowZip64rzQThe zip file ({}) has a file ({}) trying to install outside target directory ({})rr�wbN)r�open�zipfile�ZipFiler!�namelist�infolistr4rr
r�join�dirnamer(r�format�endswith�shutil�copyfileobj�closer3r+)
r4r5r6�zipfp�zip�leadingr,�name�fn�dir�message�fp�destfprrr�
unzip_filegs<	

�

,

�
rO)r4r5rcCsXt|�|���d�s$|���d�r*d}nL|���t�r>d}n8|���t�rRd}n$|���d�rfd}nt�d|�d	}tj||d
d�}�z�t	dd
�|�
�D��}|�
�D�]�}|j}|r�t|�d}t
j�||�}t||�s�d}t|�|||���|���rt|�q�|���rhz|�||�WnDt�yd}	z*t�d||j|	�WYd}	~	q�WYd}	~	n
d}	~	00q�z|�|�}
WnHttf�y�}	z*t�d||j|	�WYd}	~	q�WYd}	~	n
d}	~	00tt
j�|��|
du�s�J�t|d��}t�|
|�Wd�n1�s0Y|
��|�||�|jd@r�t |�q�W|��n
|��0dS)a�
    Untar the file (with path `filename`) to the destination `location`.
    All files are written based on system defaults and umask (i.e. permissions
    are not preserved), except that regular file members with any execute
    permissions (user, group, or world) have "chmod +x" applied after being
    written.  Note that for windows, any execute changes using os.chmod are
    no-ops per the python docs.
    z.gzz.tgzzr:gzzr:bz2zr:xzz.tar�rz-Cannot determine compression type for file %szr:*zutf-8)�encodingcSsg|]
}|j�qSr)rI)�.0�memberrrr�
<listcomp>��zuntar_file.<locals>.<listcomp>rzQThe tar file ({}) has a file ({}) trying to install outside target directory ({})z/In the tar file %s the member %s is invalid: %sNr9r))!r�lowerrBrr	�logger�warning�tarfiler:r!�
getmembersrIrr
rr?r(rrA�isdir�issym�_extract_member�	Exception�extractfile�KeyError�AttributeErrorr@rCrDrE�utimer2r+)r4r5r2�tarrHrSrJrrL�excrMrNrrr�
untar_file�st	�
�


�$�",
re)r4r5�content_typercCs�tj�|�}|dks,|���t�s,t�|�rDt|||�d�d�nR|dkslt	�
|�sl|���ttt
�rxt||�nt�d|||�td|����dS)Nzapplication/zipz.whl)r6zapplication/x-gzipzZCannot unpack file %s (downloaded from %s, content-type: %s); cannot detect archive formatz#Cannot determine archive format of )r
r�realpathrVrBr
r;�
is_zipfilerOrY�
is_tarfilerrr	rerW�criticalr)r4r5rfrrr�unpack_file�s,�������rk)T)N))�__doc__�loggingr
rCr0rYr;�typingrrrr�pip._internal.exceptionsr�pip._internal.utils.filetypesrrr	r
�pip._internal.utils.miscr�	getLogger�__name__rW�SUPPORTED_EXTENSIONS�bz2�ImportError�debug�lzma�intr�strrr/r!r(r+r3rOrerkrrrr�<module>sL
.T��site-packages/pip/_internal/utils/__pycache__/parallel.cpython-39.pyc000064400000006255150564745010021643 0ustar00a

��Re|�@s\dZddgZddlmZddlmZddlmZddlmZ	ddl
mZmZm
Z
mZmZddlmZeejej	fZed	�Zed
�ZzddlZWney�dZYn0d
ZdZeee
ed�dd��Zdeegefeeee
ed�dd�Zdeegefeeee
ed�dd�Zdeegefeeee
ed�dd�Ze�rPeZZneZeZdS)abConvenient parallelization of higher order functions.

This module provides two helper functions, with appropriate fallbacks on
Python 2 and on systems lacking support for synchronization mechanisms:

- map_multiprocess
- map_multithread

These helpers work like Python 3's map, with two differences:

- They don't guarantee the order of processing of
  the elements of the iterable.
- The underlying process/thread pools chop the iterable into
  a number of chunks, so that for very long iterables using
  a large value for chunksize can make the job complete much faster
  than using the default value of 1.
�map_multiprocess�map_multithread�)�contextmanager)�Pool��pool)�Callable�Iterable�Iterator�TypeVar�Union)�DEFAULT_POOLSIZE�S�TNTFi��)r�returnccsBz"|VW|��|��|��n|��|��|��0dS)z>Return a context manager making sure the pool closes properly.N)�close�join�	terminater�r��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/parallel.py�closing.s
�r�)�func�iterable�	chunksizercCs
t||�S)z�Make an iterator applying func to each element in iterable.

    This function is the sequential fallback either on Python 2
    where Pool.imap* doesn't react to KeyboardInterrupt
    or when sem_open is unavailable.
    )�map)rrrrrr�
_map_fallback;s	rcCs<tt���}|�|||�Wd�S1s.0YdS)z�Chop iterable into chunks and submit them to a process pool.

    For very long iterables using a large value for chunksize can make
    the job complete much faster than using the default value of 1.

    Return an unordered iterator of the results.
    N)r�ProcessPool�imap_unordered�rrrrrrr�_map_multiprocessGs
r cCs>ttt���}|�|||�Wd�S1s00YdS)z�Chop iterable into chunks and submit them to a thread pool.

    For very long iterables using a large value for chunksize can make
    the job complete much faster than using the default value of 1.

    Return an unordered iterator of the results.
    N)r�
ThreadPoolr
rrrrr�_map_multithreadUs
r")r)r)r)�__doc__�__all__�
contextlibr�multiprocessingrrrZmultiprocessing.dummyr!�typingrr	r
rrZpip._vendor.requests.adaptersr
rrZmultiprocessing.synchronize�ImportErrorZ
LACK_SEM_OPEN�TIMEOUTr�intrr r"rrrrrr�<module>sJ

��
����
site-packages/pip/_internal/utils/__pycache__/__init__.cpython-39.pyc000064400000000346150564745010021601 0ustar00a

��Re�@sdS)N�rrr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/__init__.py�<module>�site-packages/pip/_internal/utils/__pycache__/glibc.cpython-39.pyc000064400000003257150564745010021126 0ustar00a

��Re&�@spddlZddlZddlmZmZeed�dd�Zeed�dd�Zeed�dd	�Zeeefd�d
d�Z	dS)�N)�Optional�Tuple)�returncCst�p
t�S)z9Returns glibc version string, or None if not using glibc.)�glibc_version_string_confstr�glibc_version_string_ctypes�rr��/builddir/build/BUILDROOT/alt-python39-pip-21.3.1-2.el8.x86_64/opt/alt/python39/lib/python3.9/site-packages/pip/_internal/utils/glibc.py�glibc_version_string	sr	c
CsDtjdkrdSzt�d���\}}Wntttfy>YdS0|S)z@Primary implementation of glibc_version_string using os.confstr.�win32N�CS_GNU_LIBC_VERSION)�sys�platform�os�confstr�split�AttributeError�OSError�
ValueError)�_�versionrrrrs
rcCsrzddl}Wnty YdS0|�d�}z
|j}WntyJYdS0|j|_|�}t|t�sn|�	d�}|S)z=Fallback implementation of glibc_version_string using ctypes.rN�ascii)
�ctypes�ImportError�CDLL�gnu_get_libc_versionr�c_char_p�restype�
isinstance�str�decode)r�process_namespacer�version_strrrrrs



rcCst�}|durdSd|fSdS)z�Try to determine the glibc version

    Returns a tuple of strings (lib, version) which default to empty strings
    in case the lookup fails.
    N)�r"�glibc)r	)�
glibc_versionrrr�libc_verNsr%)
rr�typingrrrr	rrr%rrrr�<module>s/site-packages/pip/_internal/utils/filesystem.py000064400000013405150564745010015737 0ustar00import fnmatch
import os
import os.path
import random
import shutil
import stat
import sys
from contextlib import contextmanager
from tempfile import NamedTemporaryFile
from typing import Any, BinaryIO, Iterator, List, Union, cast

from pip._vendor.tenacity import retry, stop_after_delay, wait_fixed

from pip._internal.utils.compat import get_path_uid
from pip._internal.utils.misc import format_size


def check_path_owner(path: str) -> bool:
    # If we don't have a way to check the effective uid of this process, then
    # we'll just assume that we own the directory.
    if sys.platform == "win32" or not hasattr(os, "geteuid"):
        return True

    assert os.path.isabs(path)

    previous = None
    while path != previous:
        if os.path.lexists(path):
            # Check if path is writable by current user.
            if os.geteuid() == 0:
                # Special handling for root user in order to handle properly
                # cases where users use sudo without -H flag.
                try:
                    path_uid = get_path_uid(path)
                except OSError:
                    return False
                return path_uid == 0
            else:
                return os.access(path, os.W_OK)
        else:
            previous, path = path, os.path.dirname(path)
    return False  # assume we don't own the path


def copy2_fixed(src: str, dest: str) -> None:
    """Wrap shutil.copy2() but map errors copying socket files to
    SpecialFileError as expected.

    See also https://bugs.python.org/issue37700.
    """
    try:
        shutil.copy2(src, dest)
    except OSError:
        for f in [src, dest]:
            try:
                is_socket_file = is_socket(f)
            except OSError:
                # An error has already occurred. Another error here is not
                # a problem and we can ignore it.
                pass
            else:
                if is_socket_file:
                    raise shutil.SpecialFileError(f"`{f}` is a socket")

        raise


def is_socket(path: str) -> bool:
    return stat.S_ISSOCK(os.lstat(path).st_mode)


@contextmanager
def adjacent_tmp_file(path: str, **kwargs: Any) -> Iterator[BinaryIO]:
    """Return a file-like object pointing to a tmp file next to path.

    The file is created securely and is ensured to be written to disk
    after the context reaches its end.

    kwargs will be passed to tempfile.NamedTemporaryFile to control
    the way the temporary file will be opened.
    """
    with NamedTemporaryFile(
        delete=False,
        dir=os.path.dirname(path),
        prefix=os.path.basename(path),
        suffix=".tmp",
        **kwargs,
    ) as f:
        result = cast(BinaryIO, f)
        try:
            yield result
        finally:
            result.flush()
            os.fsync(result.fileno())


# Tenacity raises RetryError by default, explicitly raise the original exception
_replace_retry = retry(reraise=True, stop=stop_after_delay(1), wait=wait_fixed(0.25))

replace = _replace_retry(os.replace)


# test_writable_dir and _test_writable_dir_win are copied from Flit,
# with the author's agreement to also place them under pip's license.
def test_writable_dir(path: str) -> bool:
    """Check if a directory is writable.

    Uses os.access() on POSIX, tries creating files on Windows.
    """
    # If the directory doesn't exist, find the closest parent that does.
    while not os.path.isdir(path):
        parent = os.path.dirname(path)
        if parent == path:
            break  # Should never get here, but infinite loops are bad
        path = parent

    if os.name == "posix":
        return os.access(path, os.W_OK)

    return _test_writable_dir_win(path)


def _test_writable_dir_win(path: str) -> bool:
    # os.access doesn't work on Windows: http://bugs.python.org/issue2528
    # and we can't use tempfile: http://bugs.python.org/issue22107
    basename = "accesstest_deleteme_fishfingers_custard_"
    alphabet = "abcdefghijklmnopqrstuvwxyz0123456789"
    for _ in range(10):
        name = basename + "".join(random.choice(alphabet) for _ in range(6))
        file = os.path.join(path, name)
        try:
            fd = os.open(file, os.O_RDWR | os.O_CREAT | os.O_EXCL)
        except FileExistsError:
            pass
        except PermissionError:
            # This could be because there's a directory with the same name.
            # But it's highly unlikely there's a directory called that,
            # so we'll assume it's because the parent dir is not writable.
            # This could as well be because the parent dir is not readable,
            # due to non-privileged user access.
            return False
        else:
            os.close(fd)
            os.unlink(file)
            return True

    # This should never be reached
    raise OSError("Unexpected condition testing for writable directory")


def find_files(path: str, pattern: str) -> List[str]:
    """Returns a list of absolute paths of files beneath path, recursively,
    with filenames which match the UNIX-style shell glob pattern."""
    result: List[str] = []
    for root, _, files in os.walk(path):
        matches = fnmatch.filter(files, pattern)
        result.extend(os.path.join(root, f) for f in matches)
    return result


def file_size(path: str) -> Union[int, float]:
    # If it's a symlink, return 0.
    if os.path.islink(path):
        return 0
    return os.path.getsize(path)


def format_file_size(path: str) -> str:
    return format_size(file_size(path))


def directory_size(path: str) -> Union[int, float]:
    size = 0.0
    for root, _dirs, files in os.walk(path):
        for filename in files:
            file_path = os.path.join(root, filename)
            size += file_size(file_path)
    return size


def format_directory_size(path: str) -> str:
    return format_size(directory_size(path))
site-packages/pip/_internal/utils/compatibility_tags.py000064400000012401150564745010017435 0ustar00"""Generate and work with PEP 425 Compatibility Tags.
"""

import re
from typing import List, Optional, Tuple

from pip._vendor.packaging.tags import (
    PythonVersion,
    Tag,
    compatible_tags,
    cpython_tags,
    generic_tags,
    interpreter_name,
    interpreter_version,
    mac_platforms,
)

_osx_arch_pat = re.compile(r"(.+)_(\d+)_(\d+)_(.+)")


def version_info_to_nodot(version_info: Tuple[int, ...]) -> str:
    # Only use up to the first two numbers.
    return "".join(map(str, version_info[:2]))


def _mac_platforms(arch: str) -> List[str]:
    match = _osx_arch_pat.match(arch)
    if match:
        name, major, minor, actual_arch = match.groups()
        mac_version = (int(major), int(minor))
        arches = [
            # Since we have always only checked that the platform starts
            # with "macosx", for backwards-compatibility we extract the
            # actual prefix provided by the user in case they provided
            # something like "macosxcustom_". It may be good to remove
            # this as undocumented or deprecate it in the future.
            "{}_{}".format(name, arch[len("macosx_") :])
            for arch in mac_platforms(mac_version, actual_arch)
        ]
    else:
        # arch pattern didn't match (?!)
        arches = [arch]
    return arches


def _custom_manylinux_platforms(arch: str) -> List[str]:
    arches = [arch]
    arch_prefix, arch_sep, arch_suffix = arch.partition("_")
    if arch_prefix == "manylinux2014":
        # manylinux1/manylinux2010 wheels run on most manylinux2014 systems
        # with the exception of wheels depending on ncurses. PEP 599 states
        # manylinux1/manylinux2010 wheels should be considered
        # manylinux2014 wheels:
        # https://www.python.org/dev/peps/pep-0599/#backwards-compatibility-with-manylinux2010-wheels
        if arch_suffix in {"i686", "x86_64"}:
            arches.append("manylinux2010" + arch_sep + arch_suffix)
            arches.append("manylinux1" + arch_sep + arch_suffix)
    elif arch_prefix == "manylinux2010":
        # manylinux1 wheels run on most manylinux2010 systems with the
        # exception of wheels depending on ncurses. PEP 571 states
        # manylinux1 wheels should be considered manylinux2010 wheels:
        # https://www.python.org/dev/peps/pep-0571/#backwards-compatibility-with-manylinux1-wheels
        arches.append("manylinux1" + arch_sep + arch_suffix)
    return arches


def _get_custom_platforms(arch: str) -> List[str]:
    arch_prefix, arch_sep, arch_suffix = arch.partition("_")
    if arch.startswith("macosx"):
        arches = _mac_platforms(arch)
    elif arch_prefix in ["manylinux2014", "manylinux2010"]:
        arches = _custom_manylinux_platforms(arch)
    else:
        arches = [arch]
    return arches


def _expand_allowed_platforms(platforms: Optional[List[str]]) -> Optional[List[str]]:
    if not platforms:
        return None

    seen = set()
    result = []

    for p in platforms:
        if p in seen:
            continue
        additions = [c for c in _get_custom_platforms(p) if c not in seen]
        seen.update(additions)
        result.extend(additions)

    return result


def _get_python_version(version: str) -> PythonVersion:
    if len(version) > 1:
        return int(version[0]), int(version[1:])
    else:
        return (int(version[0]),)


def _get_custom_interpreter(
    implementation: Optional[str] = None, version: Optional[str] = None
) -> str:
    if implementation is None:
        implementation = interpreter_name()
    if version is None:
        version = interpreter_version()
    return f"{implementation}{version}"


def get_supported(
    version: Optional[str] = None,
    platforms: Optional[List[str]] = None,
    impl: Optional[str] = None,
    abis: Optional[List[str]] = None,
) -> List[Tag]:
    """Return a list of supported tags for each version specified in
    `versions`.

    :param version: a string version, of the form "33" or "32",
        or None. The version will be assumed to support our ABI.
    :param platform: specify a list of platforms you want valid
        tags for, or None. If None, use the local system platform.
    :param impl: specify the exact implementation you want valid
        tags for, or None. If None, use the local interpreter impl.
    :param abis: specify a list of abis you want valid
        tags for, or None. If None, use the local interpreter abi.
    """
    supported: List[Tag] = []

    python_version: Optional[PythonVersion] = None
    if version is not None:
        python_version = _get_python_version(version)

    interpreter = _get_custom_interpreter(impl, version)

    platforms = _expand_allowed_platforms(platforms)

    is_cpython = (impl or interpreter_name()) == "cp"
    if is_cpython:
        supported.extend(
            cpython_tags(
                python_version=python_version,
                abis=abis,
                platforms=platforms,
            )
        )
    else:
        supported.extend(
            generic_tags(
                interpreter=interpreter,
                abis=abis,
                platforms=platforms,
            )
        )
    supported.extend(
        compatible_tags(
            python_version=python_version,
            interpreter=interpreter,
            platforms=platforms,
        )
    )

    return supported
site-packages/pip/_internal/utils/misc.py000064400000050452150564745010014511 0ustar00# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import contextlib
import errno
import getpass
import hashlib
import io
import logging
import os
import posixpath
import shutil
import stat
import sys
import urllib.parse
from io import StringIO
from itertools import filterfalse, tee, zip_longest
from types import TracebackType
from typing import (
    Any,
    BinaryIO,
    Callable,
    ContextManager,
    Iterable,
    Iterator,
    List,
    Optional,
    TextIO,
    Tuple,
    Type,
    TypeVar,
    cast,
)

from pip._vendor.pkg_resources import Distribution
from pip._vendor.tenacity import retry, stop_after_delay, wait_fixed

from pip import __version__
from pip._internal.exceptions import CommandError
from pip._internal.locations import get_major_minor_version, site_packages, user_site
from pip._internal.locations import get_scheme
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.egg_link import egg_link_path_from_location
from pip._internal.utils.virtualenv import running_under_virtualenv

__all__ = [
    "rmtree",
    "display_path",
    "backup_dir",
    "ask",
    "splitext",
    "format_size",
    "is_installable_dir",
    "normalize_path",
    "renames",
    "get_prog",
    "captured_stdout",
    "ensure_dir",
    "remove_auth_from_url",
]


logger = logging.getLogger(__name__)

T = TypeVar("T")
ExcInfo = Tuple[Type[BaseException], BaseException, TracebackType]
VersionInfo = Tuple[int, int, int]
NetlocTuple = Tuple[str, Tuple[Optional[str], Optional[str]]]


def get_pip_version() -> str:
    pip_pkg_dir = os.path.join(os.path.dirname(__file__), "..", "..")
    pip_pkg_dir = os.path.abspath(pip_pkg_dir)

    return "pip {} from {} (python {})".format(
        __version__,
        pip_pkg_dir,
        get_major_minor_version(),
    )


def normalize_version_info(py_version_info: Tuple[int, ...]) -> Tuple[int, int, int]:
    """
    Convert a tuple of ints representing a Python version to one of length
    three.

    :param py_version_info: a tuple of ints representing a Python version,
        or None to specify no version. The tuple can have any length.

    :return: a tuple of length three if `py_version_info` is non-None.
        Otherwise, return `py_version_info` unchanged (i.e. None).
    """
    if len(py_version_info) < 3:
        py_version_info += (3 - len(py_version_info)) * (0,)
    elif len(py_version_info) > 3:
        py_version_info = py_version_info[:3]

    return cast("VersionInfo", py_version_info)


def ensure_dir(path: str) -> None:
    """os.path.makedirs without EEXIST."""
    try:
        os.makedirs(path)
    except OSError as e:
        # Windows can raise spurious ENOTEMPTY errors. See #6426.
        if e.errno != errno.EEXIST and e.errno != errno.ENOTEMPTY:
            raise


def get_prog() -> str:
    try:
        prog = os.path.basename(sys.argv[0])
        if prog in ("__main__.py", "-c"):
            return f"{sys.executable} -m pip"
        else:
            return prog
    except (AttributeError, TypeError, IndexError):
        pass
    return "pip"


# Retry every half second for up to 3 seconds
# Tenacity raises RetryError by default, explicitly raise the original exception
@retry(reraise=True, stop=stop_after_delay(3), wait=wait_fixed(0.5))
def rmtree(dir: str, ignore_errors: bool = False) -> None:
    shutil.rmtree(dir, ignore_errors=ignore_errors, onerror=rmtree_errorhandler)


def rmtree_errorhandler(func: Callable[..., Any], path: str, exc_info: ExcInfo) -> None:
    """On Windows, the files in .svn are read-only, so when rmtree() tries to
    remove them, an exception is thrown.  We catch that here, remove the
    read-only attribute, and hopefully continue without problems."""
    try:
        has_attr_readonly = not (os.stat(path).st_mode & stat.S_IWRITE)
    except OSError:
        # it's equivalent to os.path.exists
        return

    if has_attr_readonly:
        # convert to read/write
        os.chmod(path, stat.S_IWRITE)
        # use the original function to repeat the operation
        func(path)
        return
    else:
        raise


def display_path(path: str) -> str:
    """Gives the display value for a given path, making it relative to cwd
    if possible."""
    path = os.path.normcase(os.path.abspath(path))
    if path.startswith(os.getcwd() + os.path.sep):
        path = "." + path[len(os.getcwd()) :]
    return path


def backup_dir(dir: str, ext: str = ".bak") -> str:
    """Figure out the name of a directory to back up the given dir to
    (adding .bak, .bak2, etc)"""
    n = 1
    extension = ext
    while os.path.exists(dir + extension):
        n += 1
        extension = ext + str(n)
    return dir + extension


def ask_path_exists(message: str, options: Iterable[str]) -> str:
    for action in os.environ.get("PIP_EXISTS_ACTION", "").split():
        if action in options:
            return action
    return ask(message, options)


def _check_no_input(message: str) -> None:
    """Raise an error if no input is allowed."""
    if os.environ.get("PIP_NO_INPUT"):
        raise Exception(
            f"No input was expected ($PIP_NO_INPUT set); question: {message}"
        )


def ask(message: str, options: Iterable[str]) -> str:
    """Ask the message interactively, with the given possible responses"""
    while 1:
        _check_no_input(message)
        response = input(message)
        response = response.strip().lower()
        if response not in options:
            print(
                "Your response ({!r}) was not one of the expected responses: "
                "{}".format(response, ", ".join(options))
            )
        else:
            return response


def ask_input(message: str) -> str:
    """Ask for input interactively."""
    _check_no_input(message)
    return input(message)


def ask_password(message: str) -> str:
    """Ask for a password interactively."""
    _check_no_input(message)
    return getpass.getpass(message)


def strtobool(val: str) -> int:
    """Convert a string representation of truth to true (1) or false (0).

    True values are 'y', 'yes', 't', 'true', 'on', and '1'; false values
    are 'n', 'no', 'f', 'false', 'off', and '0'.  Raises ValueError if
    'val' is anything else.
    """
    val = val.lower()
    if val in ("y", "yes", "t", "true", "on", "1"):
        return 1
    elif val in ("n", "no", "f", "false", "off", "0"):
        return 0
    else:
        raise ValueError(f"invalid truth value {val!r}")


def format_size(bytes: float) -> str:
    if bytes > 1000 * 1000:
        return "{:.1f} MB".format(bytes / 1000.0 / 1000)
    elif bytes > 10 * 1000:
        return "{} kB".format(int(bytes / 1000))
    elif bytes > 1000:
        return "{:.1f} kB".format(bytes / 1000.0)
    else:
        return "{} bytes".format(int(bytes))


def tabulate(rows: Iterable[Iterable[Any]]) -> Tuple[List[str], List[int]]:
    """Return a list of formatted rows and a list of column sizes.

    For example::

    >>> tabulate([['foobar', 2000], [0xdeadbeef]])
    (['foobar     2000', '3735928559'], [10, 4])
    """
    rows = [tuple(map(str, row)) for row in rows]
    sizes = [max(map(len, col)) for col in zip_longest(*rows, fillvalue="")]
    table = [" ".join(map(str.ljust, row, sizes)).rstrip() for row in rows]
    return table, sizes


def is_installable_dir(path: str) -> bool:
    """Is path is a directory containing pyproject.toml or setup.py?

    If pyproject.toml exists, this is a PEP 517 project. Otherwise we look for
    a legacy setuptools layout by identifying setup.py. We don't check for the
    setup.cfg because using it without setup.py is only available for PEP 517
    projects, which are already covered by the pyproject.toml check.
    """
    if not os.path.isdir(path):
        return False
    if os.path.isfile(os.path.join(path, "pyproject.toml")):
        return True
    if os.path.isfile(os.path.join(path, "setup.py")):
        return True
    return False


def read_chunks(file: BinaryIO, size: int = io.DEFAULT_BUFFER_SIZE) -> Iterator[bytes]:
    """Yield pieces of data from a file-like object until EOF."""
    while True:
        chunk = file.read(size)
        if not chunk:
            break
        yield chunk


def normalize_path(path: str, resolve_symlinks: bool = True) -> str:
    """
    Convert a path to its canonical, case-normalized, absolute version.

    """
    path = os.path.expanduser(path)
    if resolve_symlinks:
        path = os.path.realpath(path)
    else:
        path = os.path.abspath(path)
    return os.path.normcase(path)


def splitext(path: str) -> Tuple[str, str]:
    """Like os.path.splitext, but take off .tar too"""
    base, ext = posixpath.splitext(path)
    if base.lower().endswith(".tar"):
        ext = base[-4:] + ext
        base = base[:-4]
    return base, ext


def renames(old: str, new: str) -> None:
    """Like os.renames(), but handles renaming across devices."""
    # Implementation borrowed from os.renames().
    head, tail = os.path.split(new)
    if head and tail and not os.path.exists(head):
        os.makedirs(head)

    shutil.move(old, new)

    head, tail = os.path.split(old)
    if head and tail:
        try:
            os.removedirs(head)
        except OSError:
            pass


def is_local(path: str) -> bool:
    """
    Return True if path is within sys.prefix, if we're running in a virtualenv.

    If we're not in a virtualenv, all paths are considered "local."

    Caution: this function assumes the head of path has been normalized
    with normalize_path.
    """
    if not running_under_virtualenv():
        return True
    return path.startswith(normalize_path(sys.prefix))


def dist_is_local(dist: Distribution) -> bool:
    """
    Return True if given Distribution object is installed locally
    (i.e. within current virtualenv).

    Always True if we're not in a virtualenv.

    """
    return is_local(dist_location(dist))


def dist_in_usersite(dist: Distribution) -> bool:
    """
    Return True if given Distribution is installed in user site.
    """
    return dist_location(dist).startswith(normalize_path(user_site))


def dist_in_site_packages(dist: Distribution) -> bool:
    """
    Return True if given Distribution is installed in
    sysconfig.get_python_lib().
    """
    return dist_location(dist).startswith(normalize_path(site_packages))


def dist_in_install_path(dist):
    """
    Return True if given Distribution is installed in
    path matching distutils_scheme layout.
    """
    norm_path = normalize_path(dist_location(dist))
    return norm_path.startswith(normalize_path(
        get_scheme("").purelib.split('python')[0]))


def get_distribution(req_name: str) -> Optional[Distribution]:
    """Given a requirement name, return the installed Distribution object.

    This searches from *all* distributions available in the environment, to
    match the behavior of ``pkg_resources.get_distribution()``.

    Left for compatibility until direct pkg_resources uses are refactored out.
    """
    from pip._internal.metadata import get_default_environment
    from pip._internal.metadata.pkg_resources import Distribution as _Dist

    dist = get_default_environment().get_distribution(req_name)
    if dist is None:
        return None
    return cast(_Dist, dist)._dist


def dist_location(dist: Distribution) -> str:
    """
    Get the site-packages location of this distribution. Generally
    this is dist.location, except in the case of develop-installed
    packages, where dist.location is the source code location, and we
    want to know where the egg-link file is.

    The returned location is normalized (in particular, with symlinks removed).
    """
    egg_link = egg_link_path_from_location(dist.project_name)
    if egg_link:
        return normalize_path(egg_link)
    return normalize_path(dist.location)


def write_output(msg: Any, *args: Any) -> None:
    logger.info(msg, *args)


class StreamWrapper(StringIO):
    orig_stream: TextIO = None

    @classmethod
    def from_stream(cls, orig_stream: TextIO) -> "StreamWrapper":
        cls.orig_stream = orig_stream
        return cls()

    # compileall.compile_dir() needs stdout.encoding to print to stdout
    # https://github.com/python/mypy/issues/4125
    @property
    def encoding(self):  # type: ignore
        return self.orig_stream.encoding


@contextlib.contextmanager
def captured_output(stream_name: str) -> Iterator[StreamWrapper]:
    """Return a context manager used by captured_stdout/stdin/stderr
    that temporarily replaces the sys stream *stream_name* with a StringIO.

    Taken from Lib/support/__init__.py in the CPython repo.
    """
    orig_stdout = getattr(sys, stream_name)
    setattr(sys, stream_name, StreamWrapper.from_stream(orig_stdout))
    try:
        yield getattr(sys, stream_name)
    finally:
        setattr(sys, stream_name, orig_stdout)


def captured_stdout() -> ContextManager[StreamWrapper]:
    """Capture the output of sys.stdout:

       with captured_stdout() as stdout:
           print('hello')
       self.assertEqual(stdout.getvalue(), 'hello\n')

    Taken from Lib/support/__init__.py in the CPython repo.
    """
    return captured_output("stdout")


def captured_stderr() -> ContextManager[StreamWrapper]:
    """
    See captured_stdout().
    """
    return captured_output("stderr")


# Simulates an enum
def enum(*sequential: Any, **named: Any) -> Type[Any]:
    enums = dict(zip(sequential, range(len(sequential))), **named)
    reverse = {value: key for key, value in enums.items()}
    enums["reverse_mapping"] = reverse
    return type("Enum", (), enums)


def build_netloc(host: str, port: Optional[int]) -> str:
    """
    Build a netloc from a host-port pair
    """
    if port is None:
        return host
    if ":" in host:
        # Only wrap host with square brackets when it is IPv6
        host = f"[{host}]"
    return f"{host}:{port}"


def build_url_from_netloc(netloc: str, scheme: str = "https") -> str:
    """
    Build a full URL from a netloc.
    """
    if netloc.count(":") >= 2 and "@" not in netloc and "[" not in netloc:
        # It must be a bare IPv6 address, so wrap it with brackets.
        netloc = f"[{netloc}]"
    return f"{scheme}://{netloc}"


def parse_netloc(netloc: str) -> Tuple[str, Optional[int]]:
    """
    Return the host-port pair from a netloc.
    """
    url = build_url_from_netloc(netloc)
    parsed = urllib.parse.urlparse(url)
    return parsed.hostname, parsed.port


def split_auth_from_netloc(netloc: str) -> NetlocTuple:
    """
    Parse out and remove the auth information from a netloc.

    Returns: (netloc, (username, password)).
    """
    if "@" not in netloc:
        return netloc, (None, None)

    # Split from the right because that's how urllib.parse.urlsplit()
    # behaves if more than one @ is present (which can be checked using
    # the password attribute of urlsplit()'s return value).
    auth, netloc = netloc.rsplit("@", 1)
    pw: Optional[str] = None
    if ":" in auth:
        # Split from the left because that's how urllib.parse.urlsplit()
        # behaves if more than one : is present (which again can be checked
        # using the password attribute of the return value)
        user, pw = auth.split(":", 1)
    else:
        user, pw = auth, None

    user = urllib.parse.unquote(user)
    if pw is not None:
        pw = urllib.parse.unquote(pw)

    return netloc, (user, pw)


def redact_netloc(netloc: str) -> str:
    """
    Replace the sensitive data in a netloc with "****", if it exists.

    For example:
        - "user:pass@example.com" returns "user:****@example.com"
        - "accesstoken@example.com" returns "****@example.com"
    """
    netloc, (user, password) = split_auth_from_netloc(netloc)
    if user is None:
        return netloc
    if password is None:
        user = "****"
        password = ""
    else:
        user = urllib.parse.quote(user)
        password = ":****"
    return "{user}{password}@{netloc}".format(
        user=user, password=password, netloc=netloc
    )


def _transform_url(
    url: str, transform_netloc: Callable[[str], Tuple[Any, ...]]
) -> Tuple[str, NetlocTuple]:
    """Transform and replace netloc in a url.

    transform_netloc is a function taking the netloc and returning a
    tuple. The first element of this tuple is the new netloc. The
    entire tuple is returned.

    Returns a tuple containing the transformed url as item 0 and the
    original tuple returned by transform_netloc as item 1.
    """
    purl = urllib.parse.urlsplit(url)
    netloc_tuple = transform_netloc(purl.netloc)
    # stripped url
    url_pieces = (purl.scheme, netloc_tuple[0], purl.path, purl.query, purl.fragment)
    surl = urllib.parse.urlunsplit(url_pieces)
    return surl, cast("NetlocTuple", netloc_tuple)


def _get_netloc(netloc: str) -> NetlocTuple:
    return split_auth_from_netloc(netloc)


def _redact_netloc(netloc: str) -> Tuple[str]:
    return (redact_netloc(netloc),)


def split_auth_netloc_from_url(url: str) -> Tuple[str, str, Tuple[str, str]]:
    """
    Parse a url into separate netloc, auth, and url with no auth.

    Returns: (url_without_auth, netloc, (username, password))
    """
    url_without_auth, (netloc, auth) = _transform_url(url, _get_netloc)
    return url_without_auth, netloc, auth


def remove_auth_from_url(url: str) -> str:
    """Return a copy of url with 'username:password@' removed."""
    # username/pass params are passed to subversion through flags
    # and are not recognized in the url.
    return _transform_url(url, _get_netloc)[0]


def redact_auth_from_url(url: str) -> str:
    """Replace the password in a given url with ****."""
    return _transform_url(url, _redact_netloc)[0]


class HiddenText:
    def __init__(self, secret: str, redacted: str) -> None:
        self.secret = secret
        self.redacted = redacted

    def __repr__(self) -> str:
        return "<HiddenText {!r}>".format(str(self))

    def __str__(self) -> str:
        return self.redacted

    # This is useful for testing.
    def __eq__(self, other: Any) -> bool:
        if type(self) != type(other):
            return False

        # The string being used for redaction doesn't also have to match,
        # just the raw, original string.
        return self.secret == other.secret


def hide_value(value: str) -> HiddenText:
    return HiddenText(value, redacted="****")


def hide_url(url: str) -> HiddenText:
    redacted = redact_auth_from_url(url)
    return HiddenText(url, redacted=redacted)


def protect_pip_from_modification_on_windows(modifying_pip: bool) -> None:
    """Protection of pip.exe from modification on Windows

    On Windows, any operation modifying pip should be run as:
        python -m pip ...
    """
    pip_names = [
        "pip.exe",
        "pip{}.exe".format(sys.version_info[0]),
        "pip{}.{}.exe".format(*sys.version_info[:2]),
    ]

    # See https://github.com/pypa/pip/issues/1299 for more discussion
    should_show_use_python_msg = (
        modifying_pip and WINDOWS and os.path.basename(sys.argv[0]) in pip_names
    )

    if should_show_use_python_msg:
        new_command = [sys.executable, "-m", "pip"] + sys.argv[1:]
        raise CommandError(
            "To modify pip, please run the following command:\n{}".format(
                " ".join(new_command)
            )
        )


def is_console_interactive() -> bool:
    """Is this console interactive?"""
    return sys.stdin is not None and sys.stdin.isatty()


def hash_file(path: str, blocksize: int = 1 << 20) -> Tuple[Any, int]:
    """Return (hash, length) for path using hashlib.sha256()"""

    h = hashlib.sha256()
    length = 0
    with open(path, "rb") as f:
        for block in read_chunks(f, size=blocksize):
            length += len(block)
            h.update(block)
    return h, length


def is_wheel_installed() -> bool:
    """
    Return whether the wheel package is installed.
    """
    try:
        import wheel  # noqa: F401
    except ImportError:
        return False

    return True


def pairwise(iterable: Iterable[Any]) -> Iterator[Tuple[Any, Any]]:
    """
    Return paired elements.

    For example:
        s -> (s0, s1), (s2, s3), (s4, s5), ...
    """
    iterable = iter(iterable)
    return zip_longest(iterable, iterable)


def partition(
    pred: Callable[[T], bool],
    iterable: Iterable[T],
) -> Tuple[Iterable[T], Iterable[T]]:
    """
    Use a predicate to partition entries into false entries and true entries,
    like

        partition(is_odd, range(10)) --> 0 2 4 6 8   and  1 3 5 7 9
    """
    t1, t2 = tee(iterable)
    return filterfalse(pred, t1), filter(pred, t2)
site-packages/pip/_internal/utils/packaging.py000064400000005610150564745010015476 0ustar00import functools
import logging
from email.message import Message
from email.parser import FeedParser
from typing import Optional, Tuple

from pip._vendor import pkg_resources
from pip._vendor.packaging import specifiers, version
from pip._vendor.packaging.requirements import Requirement
from pip._vendor.pkg_resources import Distribution

from pip._internal.exceptions import NoneMetadataError
from pip._internal.utils.misc import display_path

logger = logging.getLogger(__name__)


def check_requires_python(
    requires_python: Optional[str], version_info: Tuple[int, ...]
) -> bool:
    """
    Check if the given Python version matches a "Requires-Python" specifier.

    :param version_info: A 3-tuple of ints representing a Python
        major-minor-micro version to check (e.g. `sys.version_info[:3]`).

    :return: `True` if the given Python version satisfies the requirement.
        Otherwise, return `False`.

    :raises InvalidSpecifier: If `requires_python` has an invalid format.
    """
    if requires_python is None:
        # The package provides no information
        return True
    requires_python_specifier = specifiers.SpecifierSet(requires_python)

    python_version = version.parse(".".join(map(str, version_info)))
    return python_version in requires_python_specifier


def get_metadata(dist: Distribution) -> Message:
    """
    :raises NoneMetadataError: if the distribution reports `has_metadata()`
        True but `get_metadata()` returns None.
    """
    metadata_name = "METADATA"
    if isinstance(dist, pkg_resources.DistInfoDistribution) and dist.has_metadata(
        metadata_name
    ):
        metadata = dist.get_metadata(metadata_name)
    elif dist.has_metadata("PKG-INFO"):
        metadata_name = "PKG-INFO"
        metadata = dist.get_metadata(metadata_name)
    else:
        logger.warning("No metadata found in %s", display_path(dist.location))
        metadata = ""

    if metadata is None:
        raise NoneMetadataError(dist, metadata_name)

    feed_parser = FeedParser()
    # The following line errors out if with a "NoneType" TypeError if
    # passed metadata=None.
    feed_parser.feed(metadata)
    return feed_parser.close()


def get_installer(dist: Distribution) -> str:
    if dist.has_metadata("INSTALLER"):
        for line in dist.get_metadata_lines("INSTALLER"):
            if line.strip():
                return line.strip()
    return ""


@functools.lru_cache(maxsize=512)
def get_requirement(req_string: str) -> Requirement:
    """Construct a packaging.Requirement object with caching"""
    # Parsing requirement strings is expensive, and is also expected to happen
    # with a low diversity of different arguments (at least relative the number
    # constructed). This method adds a cache to requirement object creation to
    # minimize repeated parsing of the same string to construct equivalent
    # Requirement objects.
    return Requirement(req_string)
site-packages/pip/_internal/utils/encoding.py000064400000002221150564745010015333 0ustar00import codecs
import locale
import re
import sys
from typing import List, Tuple

BOMS: List[Tuple[bytes, str]] = [
    (codecs.BOM_UTF8, "utf-8"),
    (codecs.BOM_UTF16, "utf-16"),
    (codecs.BOM_UTF16_BE, "utf-16-be"),
    (codecs.BOM_UTF16_LE, "utf-16-le"),
    (codecs.BOM_UTF32, "utf-32"),
    (codecs.BOM_UTF32_BE, "utf-32-be"),
    (codecs.BOM_UTF32_LE, "utf-32-le"),
]

ENCODING_RE = re.compile(br"coding[:=]\s*([-\w.]+)")


def auto_decode(data: bytes) -> str:
    """Check a bytes string for a BOM to correctly detect the encoding

    Fallback to locale.getpreferredencoding(False) like open() on Python3"""
    for bom, encoding in BOMS:
        if data.startswith(bom):
            return data[len(bom) :].decode(encoding)
    # Lets check the first two lines as in PEP263
    for line in data.split(b"\n")[:2]:
        if line[0:1] == b"#" and ENCODING_RE.search(line):
            result = ENCODING_RE.search(line)
            assert result is not None
            encoding = result.groups()[0].decode("ascii")
            return data.decode(encoding)
    return data.decode(
        locale.getpreferredencoding(False) or sys.getdefaultencoding(),
    )
site-packages/pip/_internal/utils/_log.py000064400000001767150564745010014503 0ustar00"""Customize logging

Defines custom logger class for the `logger.verbose(...)` method.

init_logging() must be called before any other modules that call logging.getLogger.
"""

import logging
from typing import Any, cast

# custom log level for `--verbose` output
# between DEBUG and INFO
VERBOSE = 15


class VerboseLogger(logging.Logger):
    """Custom Logger, defining a verbose log-level

    VERBOSE is between INFO and DEBUG.
    """

    def verbose(self, msg: str, *args: Any, **kwargs: Any) -> None:
        return self.log(VERBOSE, msg, *args, **kwargs)


def getLogger(name: str) -> VerboseLogger:
    """logging.getLogger, but ensures our VerboseLogger class is returned"""
    return cast(VerboseLogger, logging.getLogger(name))


def init_logging() -> None:
    """Register our VerboseLogger and VERBOSE log level.

    Should be called before any calls to getLogger(),
    i.e. in pip._internal.__init__
    """
    logging.setLoggerClass(VerboseLogger)
    logging.addLevelName(VERBOSE, "VERBOSE")
site-packages/pip/_internal/utils/wheel.py000064400000014023150564745010014654 0ustar00"""Support functions for working with wheel files.
"""

import logging
from email.message import Message
from email.parser import Parser
from typing import Dict, Tuple
from zipfile import BadZipFile, ZipFile

from pip._vendor.packaging.utils import canonicalize_name
from pip._vendor.pkg_resources import DistInfoDistribution, Distribution

from pip._internal.exceptions import UnsupportedWheel
from pip._internal.utils.pkg_resources import DictMetadata

VERSION_COMPATIBLE = (1, 0)


logger = logging.getLogger(__name__)


class WheelMetadata(DictMetadata):
    """Metadata provider that maps metadata decoding exceptions to our
    internal exception type.
    """

    def __init__(self, metadata: Dict[str, bytes], wheel_name: str) -> None:
        super().__init__(metadata)
        self._wheel_name = wheel_name

    def get_metadata(self, name: str) -> str:
        try:
            return super().get_metadata(name)
        except UnicodeDecodeError as e:
            # Augment the default error with the origin of the file.
            raise UnsupportedWheel(
                f"Error decoding metadata for {self._wheel_name}: {e}"
            )


def pkg_resources_distribution_for_wheel(
    wheel_zip: ZipFile, name: str, location: str
) -> Distribution:
    """Get a pkg_resources distribution given a wheel.

    :raises UnsupportedWheel: on any errors
    """
    info_dir, _ = parse_wheel(wheel_zip, name)

    metadata_files = [p for p in wheel_zip.namelist() if p.startswith(f"{info_dir}/")]

    metadata_text: Dict[str, bytes] = {}
    for path in metadata_files:
        _, metadata_name = path.split("/", 1)

        try:
            metadata_text[metadata_name] = read_wheel_metadata_file(wheel_zip, path)
        except UnsupportedWheel as e:
            raise UnsupportedWheel("{} has an invalid wheel, {}".format(name, str(e)))

    metadata = WheelMetadata(metadata_text, location)

    return DistInfoDistribution(location=location, metadata=metadata, project_name=name)


def parse_wheel(wheel_zip: ZipFile, name: str) -> Tuple[str, Message]:
    """Extract information from the provided wheel, ensuring it meets basic
    standards.

    Returns the name of the .dist-info directory and the parsed WHEEL metadata.
    """
    try:
        info_dir = wheel_dist_info_dir(wheel_zip, name)
        metadata = wheel_metadata(wheel_zip, info_dir)
        version = wheel_version(metadata)
    except UnsupportedWheel as e:
        raise UnsupportedWheel("{} has an invalid wheel, {}".format(name, str(e)))

    check_compatibility(version, name)

    return info_dir, metadata


def wheel_dist_info_dir(source: ZipFile, name: str) -> str:
    """Returns the name of the contained .dist-info directory.

    Raises AssertionError or UnsupportedWheel if not found, >1 found, or
    it doesn't match the provided name.
    """
    # Zip file path separators must be /
    subdirs = {p.split("/", 1)[0] for p in source.namelist()}

    info_dirs = [s for s in subdirs if s.endswith(".dist-info")]

    if not info_dirs:
        raise UnsupportedWheel(".dist-info directory not found")

    if len(info_dirs) > 1:
        raise UnsupportedWheel(
            "multiple .dist-info directories found: {}".format(", ".join(info_dirs))
        )

    info_dir = info_dirs[0]

    info_dir_name = canonicalize_name(info_dir)
    canonical_name = canonicalize_name(name)
    if not info_dir_name.startswith(canonical_name):
        raise UnsupportedWheel(
            ".dist-info directory {!r} does not start with {!r}".format(
                info_dir, canonical_name
            )
        )

    return info_dir


def read_wheel_metadata_file(source: ZipFile, path: str) -> bytes:
    try:
        return source.read(path)
        # BadZipFile for general corruption, KeyError for missing entry,
        # and RuntimeError for password-protected files
    except (BadZipFile, KeyError, RuntimeError) as e:
        raise UnsupportedWheel(f"could not read {path!r} file: {e!r}")


def wheel_metadata(source: ZipFile, dist_info_dir: str) -> Message:
    """Return the WHEEL metadata of an extracted wheel, if possible.
    Otherwise, raise UnsupportedWheel.
    """
    path = f"{dist_info_dir}/WHEEL"
    # Zip file path separators must be /
    wheel_contents = read_wheel_metadata_file(source, path)

    try:
        wheel_text = wheel_contents.decode()
    except UnicodeDecodeError as e:
        raise UnsupportedWheel(f"error decoding {path!r}: {e!r}")

    # FeedParser (used by Parser) does not raise any exceptions. The returned
    # message may have .defects populated, but for backwards-compatibility we
    # currently ignore them.
    return Parser().parsestr(wheel_text)


def wheel_version(wheel_data: Message) -> Tuple[int, ...]:
    """Given WHEEL metadata, return the parsed Wheel-Version.
    Otherwise, raise UnsupportedWheel.
    """
    version_text = wheel_data["Wheel-Version"]
    if version_text is None:
        raise UnsupportedWheel("WHEEL is missing Wheel-Version")

    version = version_text.strip()

    try:
        return tuple(map(int, version.split(".")))
    except ValueError:
        raise UnsupportedWheel(f"invalid Wheel-Version: {version!r}")


def check_compatibility(version: Tuple[int, ...], name: str) -> None:
    """Raises errors or warns if called with an incompatible Wheel-Version.

    pip should refuse to install a Wheel-Version that's a major series
    ahead of what it's compatible with (e.g 2.0 > 1.1); and warn when
    installing a version only minor version ahead (e.g 1.2 > 1.1).

    version: a 2-tuple representing a Wheel-Version (Major, Minor)
    name: name of wheel or package to raise exception about

    :raises UnsupportedWheel: when an incompatible Wheel-Version is given
    """
    if version[0] > VERSION_COMPATIBLE[0]:
        raise UnsupportedWheel(
            "{}'s Wheel-Version ({}) is not compatible with this version "
            "of pip".format(name, ".".join(map(str, version)))
        )
    elif version > VERSION_COMPATIBLE:
        logger.warning(
            "Installing from a newer Wheel-Version (%s)",
            ".".join(map(str, version)),
        )
site-packages/pip/_internal/utils/setuptools_build.py000064400000011131150564745010017145 0ustar00import sys
from typing import List, Optional, Sequence

# Shim to wrap setup.py invocation with setuptools
#
# We set sys.argv[0] to the path to the underlying setup.py file so
# setuptools / distutils don't take the path to the setup.py to be "-c" when
# invoking via the shim.  This avoids e.g. the following manifest_maker
# warning: "warning: manifest_maker: standard file '-c' not found".
_SETUPTOOLS_SHIM = (
    "import io, os, sys, setuptools, tokenize; sys.argv[0] = {0!r}; __file__={0!r};"
    "f = getattr(tokenize, 'open', open)(__file__) "
    "if os.path.exists(__file__) "
    "else io.StringIO('from setuptools import setup; setup()');"
    "code = f.read().replace('\\r\\n', '\\n');"
    "f.close();"
    "exec(compile(code, __file__, 'exec'))"
)


def make_setuptools_shim_args(
    setup_py_path: str,
    global_options: Sequence[str] = None,
    no_user_config: bool = False,
    unbuffered_output: bool = False,
) -> List[str]:
    """
    Get setuptools command arguments with shim wrapped setup file invocation.

    :param setup_py_path: The path to setup.py to be wrapped.
    :param global_options: Additional global options.
    :param no_user_config: If True, disables personal user configuration.
    :param unbuffered_output: If True, adds the unbuffered switch to the
     argument list.
    """
    args = [sys.executable]
    if unbuffered_output:
        args += ["-u"]
    args += ["-c", _SETUPTOOLS_SHIM.format(setup_py_path)]
    if global_options:
        args += global_options
    if no_user_config:
        args += ["--no-user-cfg"]
    return args


def make_setuptools_bdist_wheel_args(
    setup_py_path: str,
    global_options: Sequence[str],
    build_options: Sequence[str],
    destination_dir: str,
) -> List[str]:
    # NOTE: Eventually, we'd want to also -S to the flags here, when we're
    # isolating. Currently, it breaks Python in virtualenvs, because it
    # relies on site.py to find parts of the standard library outside the
    # virtualenv.
    args = make_setuptools_shim_args(
        setup_py_path, global_options=global_options, unbuffered_output=True
    )
    args += ["bdist_wheel", "-d", destination_dir]
    args += build_options
    return args


def make_setuptools_clean_args(
    setup_py_path: str,
    global_options: Sequence[str],
) -> List[str]:
    args = make_setuptools_shim_args(
        setup_py_path, global_options=global_options, unbuffered_output=True
    )
    args += ["clean", "--all"]
    return args


def make_setuptools_develop_args(
    setup_py_path: str,
    global_options: Sequence[str],
    install_options: Sequence[str],
    no_user_config: bool,
    prefix: Optional[str],
    home: Optional[str],
    use_user_site: bool,
) -> List[str]:
    assert not (use_user_site and prefix)

    args = make_setuptools_shim_args(
        setup_py_path,
        global_options=global_options,
        no_user_config=no_user_config,
    )

    args += ["develop", "--no-deps"]

    args += install_options

    if prefix:
        args += ["--prefix", prefix]
    if home is not None:
        args += ["--install-dir", home]

    if use_user_site:
        args += ["--user", "--prefix="]

    return args


def make_setuptools_egg_info_args(
    setup_py_path: str,
    egg_info_dir: Optional[str],
    no_user_config: bool,
) -> List[str]:
    args = make_setuptools_shim_args(setup_py_path, no_user_config=no_user_config)

    args += ["egg_info"]

    if egg_info_dir:
        args += ["--egg-base", egg_info_dir]

    return args


def make_setuptools_install_args(
    setup_py_path: str,
    global_options: Sequence[str],
    install_options: Sequence[str],
    record_filename: str,
    root: Optional[str],
    prefix: Optional[str],
    header_dir: Optional[str],
    home: Optional[str],
    use_user_site: bool,
    no_user_config: bool,
    pycompile: bool,
) -> List[str]:
    assert not (use_user_site and prefix)
    assert not (use_user_site and root)

    args = make_setuptools_shim_args(
        setup_py_path,
        global_options=global_options,
        no_user_config=no_user_config,
        unbuffered_output=True,
    )
    args += ["install", "--record", record_filename]
    args += ["--single-version-externally-managed"]

    if root is not None:
        args += ["--root", root]
    if prefix is not None:
        args += ["--prefix", prefix]
    if home is not None:
        args += ["--home", home]
    if use_user_site:
        args += ["--user", "--prefix="]

    if pycompile:
        args += ["--compile"]
    else:
        args += ["--no-compile"]

    if header_dir:
        args += ["--install-headers", header_dir]

    args += install_options

    return args
site-packages/pip/_internal/utils/compat.py000064400000003534150564745010015040 0ustar00"""Stuff that differs in different Python versions and platform
distributions."""

import logging
import os
import sys

__all__ = ["get_path_uid", "stdlib_pkgs", "WINDOWS"]


logger = logging.getLogger(__name__)


def has_tls() -> bool:
    try:
        import _ssl  # noqa: F401  # ignore unused

        return True
    except ImportError:
        pass

    from pip._vendor.urllib3.util import IS_PYOPENSSL

    return IS_PYOPENSSL


def get_path_uid(path: str) -> int:
    """
    Return path's uid.

    Does not follow symlinks:
        https://github.com/pypa/pip/pull/935#discussion_r5307003

    Placed this function in compat due to differences on AIX and
    Jython, that should eventually go away.

    :raises OSError: When path is a symlink or can't be read.
    """
    if hasattr(os, "O_NOFOLLOW"):
        fd = os.open(path, os.O_RDONLY | os.O_NOFOLLOW)
        file_uid = os.fstat(fd).st_uid
        os.close(fd)
    else:  # AIX and Jython
        # WARNING: time of check vulnerability, but best we can do w/o NOFOLLOW
        if not os.path.islink(path):
            # older versions of Jython don't have `os.fstat`
            file_uid = os.stat(path).st_uid
        else:
            # raise OSError for parity with os.O_NOFOLLOW above
            raise OSError(f"{path} is a symlink; Will not return uid for symlinks")
    return file_uid


# packages in the stdlib that may have installation metadata, but should not be
# considered 'installed'.  this theoretically could be determined based on
# dist.location (py27:`sysconfig.get_paths()['stdlib']`,
# py26:sysconfig.get_config_vars('LIBDEST')), but fear platform variation may
# make this ineffective, so hard-coding
stdlib_pkgs = {"python", "wsgiref", "argparse"}


# windows detection, covers cpython and ironpython
WINDOWS = sys.platform.startswith("win") or (sys.platform == "cli" and os.name == "nt")
site-packages/pip/_internal/utils/inject_securetransport.py000064400000001433150564745010020350 0ustar00"""A helper module that injects SecureTransport, on import.

The import should be done as early as possible, to ensure all requests and
sessions (or whatever) are created after injecting SecureTransport.

Note that we only do the injection on macOS, when the linked OpenSSL is too
old to handle TLSv1.2.
"""

import sys


def inject_securetransport() -> None:
    # Only relevant on macOS
    if sys.platform != "darwin":
        return

    try:
        import ssl
    except ImportError:
        return

    # Checks for OpenSSL 1.0.1
    if ssl.OPENSSL_VERSION_NUMBER >= 0x1000100F:
        return

    try:
        from pip._vendor.urllib3.contrib import securetransport
    except (ImportError, OSError):
        return

    securetransport.inject_into_urllib3()


inject_securetransport()
site-packages/pip/_internal/utils/pkg_resources.py000064400000001733150564745010016427 0ustar00from typing import Dict, Iterable, List

from pip._vendor.pkg_resources import yield_lines


class DictMetadata:
    """IMetadataProvider that reads metadata files from a dictionary."""

    def __init__(self, metadata: Dict[str, bytes]) -> None:
        self._metadata = metadata

    def has_metadata(self, name: str) -> bool:
        return name in self._metadata

    def get_metadata(self, name: str) -> str:
        try:
            return self._metadata[name].decode()
        except UnicodeDecodeError as e:
            # Mirrors handling done in pkg_resources.NullProvider.
            e.reason += f" in {name} file"
            raise

    def get_metadata_lines(self, name: str) -> Iterable[str]:
        return yield_lines(self.get_metadata(name))

    def metadata_isdir(self, name: str) -> bool:
        return False

    def metadata_listdir(self, name: str) -> List[str]:
        return []

    def run_script(self, script_name: str, namespace: str) -> None:
        pass
site-packages/pip/_internal/utils/filetypes.py000064400000001314150564745010015553 0ustar00"""Filetype information.
"""

from typing import Tuple

from pip._internal.utils.misc import splitext

WHEEL_EXTENSION = ".whl"
BZ2_EXTENSIONS: Tuple[str, ...] = (".tar.bz2", ".tbz")
XZ_EXTENSIONS: Tuple[str, ...] = (
    ".tar.xz",
    ".txz",
    ".tlz",
    ".tar.lz",
    ".tar.lzma",
)
ZIP_EXTENSIONS: Tuple[str, ...] = (".zip", WHEEL_EXTENSION)
TAR_EXTENSIONS: Tuple[str, ...] = (".tar.gz", ".tgz", ".tar")
ARCHIVE_EXTENSIONS = ZIP_EXTENSIONS + BZ2_EXTENSIONS + TAR_EXTENSIONS + XZ_EXTENSIONS


def is_archive_file(name: str) -> bool:
    """Return True if `name` is a considered as an archive file."""
    ext = splitext(name)[1].lower()
    if ext in ARCHIVE_EXTENSIONS:
        return True
    return False
site-packages/pip/_internal/utils/logging.py000064400000026414150564745010015205 0ustar00import contextlib
import errno
import logging
import logging.handlers
import os
import sys
from logging import Filter
from typing import IO, Any, Callable, Iterator, Optional, TextIO, Type, cast

from pip._internal.utils._log import VERBOSE, getLogger
from pip._internal.utils.compat import WINDOWS
from pip._internal.utils.deprecation import DEPRECATION_MSG_PREFIX
from pip._internal.utils.misc import ensure_dir

try:
    import threading
except ImportError:
    import dummy_threading as threading  # type: ignore


try:
    from pip._vendor import colorama
# Lots of different errors can come from this, including SystemError and
# ImportError.
except Exception:
    colorama = None


_log_state = threading.local()
subprocess_logger = getLogger("pip.subprocessor")


class BrokenStdoutLoggingError(Exception):
    """
    Raised if BrokenPipeError occurs for the stdout stream while logging.
    """


def _is_broken_pipe_error(exc_class: Type[BaseException], exc: BaseException) -> bool:
    if exc_class is BrokenPipeError:
        return True

    # On Windows, a broken pipe can show up as EINVAL rather than EPIPE:
    # https://bugs.python.org/issue19612
    # https://bugs.python.org/issue30418
    if not WINDOWS:
        return False

    return isinstance(exc, OSError) and exc.errno in (errno.EINVAL, errno.EPIPE)


@contextlib.contextmanager
def indent_log(num: int = 2) -> Iterator[None]:
    """
    A context manager which will cause the log output to be indented for any
    log messages emitted inside it.
    """
    # For thread-safety
    _log_state.indentation = get_indentation()
    _log_state.indentation += num
    try:
        yield
    finally:
        _log_state.indentation -= num


def get_indentation() -> int:
    return getattr(_log_state, "indentation", 0)


class IndentingFormatter(logging.Formatter):
    default_time_format = "%Y-%m-%dT%H:%M:%S"

    def __init__(
        self,
        *args: Any,
        add_timestamp: bool = False,
        **kwargs: Any,
    ) -> None:
        """
        A logging.Formatter that obeys the indent_log() context manager.

        :param add_timestamp: A bool indicating output lines should be prefixed
            with their record's timestamp.
        """
        self.add_timestamp = add_timestamp
        super().__init__(*args, **kwargs)

    def get_message_start(self, formatted: str, levelno: int) -> str:
        """
        Return the start of the formatted log message (not counting the
        prefix to add to each line).
        """
        if levelno < logging.WARNING:
            return ""
        if formatted.startswith(DEPRECATION_MSG_PREFIX):
            # Then the message already has a prefix.  We don't want it to
            # look like "WARNING: DEPRECATION: ...."
            return ""
        if levelno < logging.ERROR:
            return "WARNING: "

        return "ERROR: "

    def format(self, record: logging.LogRecord) -> str:
        """
        Calls the standard formatter, but will indent all of the log message
        lines by our current indentation level.
        """
        formatted = super().format(record)
        message_start = self.get_message_start(formatted, record.levelno)
        formatted = message_start + formatted

        prefix = ""
        if self.add_timestamp:
            prefix = f"{self.formatTime(record)} "
        prefix += " " * get_indentation()
        formatted = "".join([prefix + line for line in formatted.splitlines(True)])
        return formatted


def _color_wrap(*colors: str) -> Callable[[str], str]:
    def wrapped(inp: str) -> str:
        return "".join(list(colors) + [inp, colorama.Style.RESET_ALL])

    return wrapped


class ColorizedStreamHandler(logging.StreamHandler):

    # Don't build up a list of colors if we don't have colorama
    if colorama:
        COLORS = [
            # This needs to be in order from highest logging level to lowest.
            (logging.ERROR, _color_wrap(colorama.Fore.RED)),
            (logging.WARNING, _color_wrap(colorama.Fore.YELLOW)),
        ]
    else:
        COLORS = []

    def __init__(self, stream: Optional[TextIO] = None, no_color: bool = None) -> None:
        super().__init__(stream)
        self._no_color = no_color

        if WINDOWS and colorama:
            self.stream = colorama.AnsiToWin32(self.stream)

    def _using_stdout(self) -> bool:
        """
        Return whether the handler is using sys.stdout.
        """
        if WINDOWS and colorama:
            # Then self.stream is an AnsiToWin32 object.
            stream = cast(colorama.AnsiToWin32, self.stream)
            return stream.wrapped is sys.stdout

        return self.stream is sys.stdout

    def should_color(self) -> bool:
        # Don't colorize things if we do not have colorama or if told not to
        if not colorama or self._no_color:
            return False

        real_stream = (
            self.stream
            if not isinstance(self.stream, colorama.AnsiToWin32)
            else self.stream.wrapped
        )

        # If the stream is a tty we should color it
        if hasattr(real_stream, "isatty") and real_stream.isatty():
            return True

        # If we have an ANSI term we should color it
        if os.environ.get("TERM") == "ANSI":
            return True

        # If anything else we should not color it
        return False

    def format(self, record: logging.LogRecord) -> str:
        msg = super().format(record)

        if self.should_color():
            for level, color in self.COLORS:
                if record.levelno >= level:
                    msg = color(msg)
                    break

        return msg

    # The logging module says handleError() can be customized.
    def handleError(self, record: logging.LogRecord) -> None:
        exc_class, exc = sys.exc_info()[:2]
        # If a broken pipe occurred while calling write() or flush() on the
        # stdout stream in logging's Handler.emit(), then raise our special
        # exception so we can handle it in main() instead of logging the
        # broken pipe error and continuing.
        if (
            exc_class
            and exc
            and self._using_stdout()
            and _is_broken_pipe_error(exc_class, exc)
        ):
            raise BrokenStdoutLoggingError()

        return super().handleError(record)


class BetterRotatingFileHandler(logging.handlers.RotatingFileHandler):
    def _open(self) -> IO[Any]:
        ensure_dir(os.path.dirname(self.baseFilename))
        return super()._open()


class MaxLevelFilter(Filter):
    def __init__(self, level: int) -> None:
        self.level = level

    def filter(self, record: logging.LogRecord) -> bool:
        return record.levelno < self.level


class ExcludeLoggerFilter(Filter):

    """
    A logging Filter that excludes records from a logger (or its children).
    """

    def filter(self, record: logging.LogRecord) -> bool:
        # The base Filter class allows only records from a logger (or its
        # children).
        return not super().filter(record)


def setup_logging(verbosity: int, no_color: bool, user_log_file: Optional[str]) -> int:
    """Configures and sets up all of the logging

    Returns the requested logging level, as its integer value.
    """

    # Determine the level to be logging at.
    if verbosity >= 2:
        level_number = logging.DEBUG
    elif verbosity == 1:
        level_number = VERBOSE
    elif verbosity == -1:
        level_number = logging.WARNING
    elif verbosity == -2:
        level_number = logging.ERROR
    elif verbosity <= -3:
        level_number = logging.CRITICAL
    else:
        level_number = logging.INFO

    level = logging.getLevelName(level_number)

    # The "root" logger should match the "console" level *unless* we also need
    # to log to a user log file.
    include_user_log = user_log_file is not None
    if include_user_log:
        additional_log_file = user_log_file
        root_level = "DEBUG"
    else:
        additional_log_file = "/dev/null"
        root_level = level

    # Disable any logging besides WARNING unless we have DEBUG level logging
    # enabled for vendored libraries.
    vendored_log_level = "WARNING" if level in ["INFO", "ERROR"] else "DEBUG"

    # Shorthands for clarity
    log_streams = {
        "stdout": "ext://sys.stdout",
        "stderr": "ext://sys.stderr",
    }
    handler_classes = {
        "stream": "pip._internal.utils.logging.ColorizedStreamHandler",
        "file": "pip._internal.utils.logging.BetterRotatingFileHandler",
    }
    handlers = ["console", "console_errors", "console_subprocess"] + (
        ["user_log"] if include_user_log else []
    )

    logging.config.dictConfig(
        {
            "version": 1,
            "disable_existing_loggers": False,
            "filters": {
                "exclude_warnings": {
                    "()": "pip._internal.utils.logging.MaxLevelFilter",
                    "level": logging.WARNING,
                },
                "restrict_to_subprocess": {
                    "()": "logging.Filter",
                    "name": subprocess_logger.name,
                },
                "exclude_subprocess": {
                    "()": "pip._internal.utils.logging.ExcludeLoggerFilter",
                    "name": subprocess_logger.name,
                },
            },
            "formatters": {
                "indent": {
                    "()": IndentingFormatter,
                    "format": "%(message)s",
                },
                "indent_with_timestamp": {
                    "()": IndentingFormatter,
                    "format": "%(message)s",
                    "add_timestamp": True,
                },
            },
            "handlers": {
                "console": {
                    "level": level,
                    "class": handler_classes["stream"],
                    "no_color": no_color,
                    "stream": log_streams["stdout"],
                    "filters": ["exclude_subprocess", "exclude_warnings"],
                    "formatter": "indent",
                },
                "console_errors": {
                    "level": "WARNING",
                    "class": handler_classes["stream"],
                    "no_color": no_color,
                    "stream": log_streams["stderr"],
                    "filters": ["exclude_subprocess"],
                    "formatter": "indent",
                },
                # A handler responsible for logging to the console messages
                # from the "subprocessor" logger.
                "console_subprocess": {
                    "level": level,
                    "class": handler_classes["stream"],
                    "no_color": no_color,
                    "stream": log_streams["stderr"],
                    "filters": ["restrict_to_subprocess"],
                    "formatter": "indent",
                },
                "user_log": {
                    "level": "DEBUG",
                    "class": handler_classes["file"],
                    "filename": additional_log_file,
                    "encoding": "utf-8",
                    "delay": True,
                    "formatter": "indent_with_timestamp",
                },
            },
            "root": {
                "level": root_level,
                "handlers": handlers,
            },
            "loggers": {"pip._vendor": {"level": vendored_log_level}},
        }
    )

    return level_number
site-packages/pip/_internal/utils/__init__.py000064400000000000150564745010015275 0ustar00site-packages/pip/_internal/utils/distutils_args.py000064400000002341150564745010016610 0ustar00from distutils.errors import DistutilsArgError
from distutils.fancy_getopt import FancyGetopt
from typing import Dict, List

_options = [
    ("exec-prefix=", None, ""),
    ("home=", None, ""),
    ("install-base=", None, ""),
    ("install-data=", None, ""),
    ("install-headers=", None, ""),
    ("install-lib=", None, ""),
    ("install-platlib=", None, ""),
    ("install-purelib=", None, ""),
    ("install-scripts=", None, ""),
    ("prefix=", None, ""),
    ("root=", None, ""),
    ("user", None, ""),
]


# typeshed doesn't permit Tuple[str, None, str], see python/typeshed#3469.
_distutils_getopt = FancyGetopt(_options)  # type: ignore


def parse_distutils_args(args: List[str]) -> Dict[str, str]:
    """Parse provided arguments, returning an object that has the
    matched arguments.

    Any unknown arguments are ignored.
    """
    result = {}
    for arg in args:
        try:
            _, match = _distutils_getopt.getopt(args=[arg])
        except DistutilsArgError:
            # We don't care about any other options, which here may be
            # considered unrecognized since our option list is not
            # exhaustive.
            pass
        else:
            result.update(match.__dict__)
    return result
site-packages/pip/_internal/utils/glibc.py000064400000006046150564745010014636 0ustar00# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import os
import sys
from typing import Optional, Tuple


def glibc_version_string() -> Optional[str]:
    "Returns glibc version string, or None if not using glibc."
    return glibc_version_string_confstr() or glibc_version_string_ctypes()


def glibc_version_string_confstr() -> Optional[str]:
    "Primary implementation of glibc_version_string using os.confstr."
    # os.confstr is quite a bit faster than ctypes.DLL. It's also less likely
    # to be broken or missing. This strategy is used in the standard library
    # platform module:
    # https://github.com/python/cpython/blob/fcf1d003bf4f0100c9d0921ff3d70e1127ca1b71/Lib/platform.py#L175-L183
    if sys.platform == "win32":
        return None
    try:
        # os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17":
        _, version = os.confstr("CS_GNU_LIBC_VERSION").split()
    except (AttributeError, OSError, ValueError):
        # os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)...
        return None
    return version


def glibc_version_string_ctypes() -> Optional[str]:
    "Fallback implementation of glibc_version_string using ctypes."

    try:
        import ctypes
    except ImportError:
        return None

    # ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen
    # manpage says, "If filename is NULL, then the returned handle is for the
    # main program". This way we can let the linker do the work to figure out
    # which libc our process is actually using.
    process_namespace = ctypes.CDLL(None)
    try:
        gnu_get_libc_version = process_namespace.gnu_get_libc_version
    except AttributeError:
        # Symbol doesn't exist -> therefore, we are not linked to
        # glibc.
        return None

    # Call gnu_get_libc_version, which returns a string like "2.5"
    gnu_get_libc_version.restype = ctypes.c_char_p
    version_str = gnu_get_libc_version()
    # py2 / py3 compatibility:
    if not isinstance(version_str, str):
        version_str = version_str.decode("ascii")

    return version_str


# platform.libc_ver regularly returns completely nonsensical glibc
# versions. E.g. on my computer, platform says:
#
#   ~$ python2.7 -c 'import platform; print(platform.libc_ver())'
#   ('glibc', '2.7')
#   ~$ python3.5 -c 'import platform; print(platform.libc_ver())'
#   ('glibc', '2.9')
#
# But the truth is:
#
#   ~$ ldd --version
#   ldd (Debian GLIBC 2.22-11) 2.22
#
# This is unfortunate, because it means that the linehaul data on libc
# versions that was generated by pip 8.1.2 and earlier is useless and
# misleading. Solution: instead of using platform, use our code that actually
# works.
def libc_ver() -> Tuple[str, str]:
    """Try to determine the glibc version

    Returns a tuple of strings (lib, version) which default to empty strings
    in case the lookup fails.
    """
    glibc_version = glibc_version_string()
    if glibc_version is None:
        return ("", "")
    else:
        return ("glibc", glibc_version)
site-packages/pip/_internal/utils/temp_dir.py000064400000016756150564745010015372 0ustar00import errno
import itertools
import logging
import os.path
import tempfile
from contextlib import ExitStack, contextmanager
from typing import Any, Dict, Iterator, Optional, TypeVar, Union

from pip._internal.utils.misc import enum, rmtree

logger = logging.getLogger(__name__)

_T = TypeVar("_T", bound="TempDirectory")


# Kinds of temporary directories. Only needed for ones that are
# globally-managed.
tempdir_kinds = enum(
    BUILD_ENV="build-env",
    EPHEM_WHEEL_CACHE="ephem-wheel-cache",
    REQ_BUILD="req-build",
)


_tempdir_manager: Optional[ExitStack] = None


@contextmanager
def global_tempdir_manager() -> Iterator[None]:
    global _tempdir_manager
    with ExitStack() as stack:
        old_tempdir_manager, _tempdir_manager = _tempdir_manager, stack
        try:
            yield
        finally:
            _tempdir_manager = old_tempdir_manager


class TempDirectoryTypeRegistry:
    """Manages temp directory behavior"""

    def __init__(self) -> None:
        self._should_delete: Dict[str, bool] = {}

    def set_delete(self, kind: str, value: bool) -> None:
        """Indicate whether a TempDirectory of the given kind should be
        auto-deleted.
        """
        self._should_delete[kind] = value

    def get_delete(self, kind: str) -> bool:
        """Get configured auto-delete flag for a given TempDirectory type,
        default True.
        """
        return self._should_delete.get(kind, True)


_tempdir_registry: Optional[TempDirectoryTypeRegistry] = None


@contextmanager
def tempdir_registry() -> Iterator[TempDirectoryTypeRegistry]:
    """Provides a scoped global tempdir registry that can be used to dictate
    whether directories should be deleted.
    """
    global _tempdir_registry
    old_tempdir_registry = _tempdir_registry
    _tempdir_registry = TempDirectoryTypeRegistry()
    try:
        yield _tempdir_registry
    finally:
        _tempdir_registry = old_tempdir_registry


class _Default:
    pass


_default = _Default()


class TempDirectory:
    """Helper class that owns and cleans up a temporary directory.

    This class can be used as a context manager or as an OO representation of a
    temporary directory.

    Attributes:
        path
            Location to the created temporary directory
        delete
            Whether the directory should be deleted when exiting
            (when used as a contextmanager)

    Methods:
        cleanup()
            Deletes the temporary directory

    When used as a context manager, if the delete attribute is True, on
    exiting the context the temporary directory is deleted.
    """

    def __init__(
        self,
        path: Optional[str] = None,
        delete: Union[bool, None, _Default] = _default,
        kind: str = "temp",
        globally_managed: bool = False,
    ):
        super().__init__()

        if delete is _default:
            if path is not None:
                # If we were given an explicit directory, resolve delete option
                # now.
                delete = False
            else:
                # Otherwise, we wait until cleanup and see what
                # tempdir_registry says.
                delete = None

        # The only time we specify path is in for editables where it
        # is the value of the --src option.
        if path is None:
            path = self._create(kind)

        self._path = path
        self._deleted = False
        self.delete = delete
        self.kind = kind

        if globally_managed:
            assert _tempdir_manager is not None
            _tempdir_manager.enter_context(self)

    @property
    def path(self) -> str:
        assert not self._deleted, f"Attempted to access deleted path: {self._path}"
        return self._path

    def __repr__(self) -> str:
        return f"<{self.__class__.__name__} {self.path!r}>"

    def __enter__(self: _T) -> _T:
        return self

    def __exit__(self, exc: Any, value: Any, tb: Any) -> None:
        if self.delete is not None:
            delete = self.delete
        elif _tempdir_registry:
            delete = _tempdir_registry.get_delete(self.kind)
        else:
            delete = True

        if delete:
            self.cleanup()

    def _create(self, kind: str) -> str:
        """Create a temporary directory and store its path in self.path"""
        # We realpath here because some systems have their default tmpdir
        # symlinked to another directory.  This tends to confuse build
        # scripts, so we canonicalize the path by traversing potential
        # symlinks here.
        path = os.path.realpath(tempfile.mkdtemp(prefix=f"pip-{kind}-"))
        logger.debug("Created temporary directory: %s", path)
        return path

    def cleanup(self) -> None:
        """Remove the temporary directory created and reset state"""
        self._deleted = True
        if not os.path.exists(self._path):
            return
        rmtree(self._path)


class AdjacentTempDirectory(TempDirectory):
    """Helper class that creates a temporary directory adjacent to a real one.

    Attributes:
        original
            The original directory to create a temp directory for.
        path
            After calling create() or entering, contains the full
            path to the temporary directory.
        delete
            Whether the directory should be deleted when exiting
            (when used as a contextmanager)

    """

    # The characters that may be used to name the temp directory
    # We always prepend a ~ and then rotate through these until
    # a usable name is found.
    # pkg_resources raises a different error for .dist-info folder
    # with leading '-' and invalid metadata
    LEADING_CHARS = "-~.=%0123456789"

    def __init__(self, original: str, delete: Optional[bool] = None) -> None:
        self.original = original.rstrip("/\\")
        super().__init__(delete=delete)

    @classmethod
    def _generate_names(cls, name: str) -> Iterator[str]:
        """Generates a series of temporary names.

        The algorithm replaces the leading characters in the name
        with ones that are valid filesystem characters, but are not
        valid package names (for both Python and pip definitions of
        package).
        """
        for i in range(1, len(name)):
            for candidate in itertools.combinations_with_replacement(
                cls.LEADING_CHARS, i - 1
            ):
                new_name = "~" + "".join(candidate) + name[i:]
                if new_name != name:
                    yield new_name

        # If we make it this far, we will have to make a longer name
        for i in range(len(cls.LEADING_CHARS)):
            for candidate in itertools.combinations_with_replacement(
                cls.LEADING_CHARS, i
            ):
                new_name = "~" + "".join(candidate) + name
                if new_name != name:
                    yield new_name

    def _create(self, kind: str) -> str:
        root, name = os.path.split(self.original)
        for candidate in self._generate_names(name):
            path = os.path.join(root, candidate)
            try:
                os.mkdir(path)
            except OSError as ex:
                # Continue if the name exists already
                if ex.errno != errno.EEXIST:
                    raise
            else:
                path = os.path.realpath(path)
                break
        else:
            # Final fallback on the default behavior.
            path = os.path.realpath(tempfile.mkdtemp(prefix=f"pip-{kind}-"))

        logger.debug("Created temporary directory: %s", path)
        return path
site-packages/pip/_internal/utils/unpacking.py000064400000021312150564745010015526 0ustar00"""Utilities related archives.
"""

import logging
import os
import shutil
import stat
import tarfile
import zipfile
from typing import Iterable, List, Optional
from zipfile import ZipInfo

from pip._internal.exceptions import InstallationError
from pip._internal.utils.filetypes import (
    BZ2_EXTENSIONS,
    TAR_EXTENSIONS,
    XZ_EXTENSIONS,
    ZIP_EXTENSIONS,
)
from pip._internal.utils.misc import ensure_dir

logger = logging.getLogger(__name__)


SUPPORTED_EXTENSIONS = ZIP_EXTENSIONS + TAR_EXTENSIONS

try:
    import bz2  # noqa

    SUPPORTED_EXTENSIONS += BZ2_EXTENSIONS
except ImportError:
    logger.debug("bz2 module is not available")

try:
    # Only for Python 3.3+
    import lzma  # noqa

    SUPPORTED_EXTENSIONS += XZ_EXTENSIONS
except ImportError:
    logger.debug("lzma module is not available")


def current_umask() -> int:
    """Get the current umask which involves having to set it temporarily."""
    mask = os.umask(0)
    os.umask(mask)
    return mask


def split_leading_dir(path: str) -> List[str]:
    path = path.lstrip("/").lstrip("\\")
    if "/" in path and (
        ("\\" in path and path.find("/") < path.find("\\")) or "\\" not in path
    ):
        return path.split("/", 1)
    elif "\\" in path:
        return path.split("\\", 1)
    else:
        return [path, ""]


def has_leading_dir(paths: Iterable[str]) -> bool:
    """Returns true if all the paths have the same leading path name
    (i.e., everything is in one subdirectory in an archive)"""
    common_prefix = None
    for path in paths:
        prefix, rest = split_leading_dir(path)
        if not prefix:
            return False
        elif common_prefix is None:
            common_prefix = prefix
        elif prefix != common_prefix:
            return False
    return True


def is_within_directory(directory: str, target: str) -> bool:
    """
    Return true if the absolute path of target is within the directory
    """
    abs_directory = os.path.abspath(directory)
    abs_target = os.path.abspath(target)

    prefix = os.path.commonprefix([abs_directory, abs_target])
    return prefix == abs_directory


def set_extracted_file_to_default_mode_plus_executable(path: str) -> None:
    """
    Make file present at path have execute for user/group/world
    (chmod +x) is no-op on windows per python docs
    """
    os.chmod(path, (0o777 & ~current_umask() | 0o111))


def zip_item_is_executable(info: ZipInfo) -> bool:
    mode = info.external_attr >> 16
    # if mode and regular file and any execute permissions for
    # user/group/world?
    return bool(mode and stat.S_ISREG(mode) and mode & 0o111)


def unzip_file(filename: str, location: str, flatten: bool = True) -> None:
    """
    Unzip the file (with path `filename`) to the destination `location`.  All
    files are written based on system defaults and umask (i.e. permissions are
    not preserved), except that regular file members with any execute
    permissions (user, group, or world) have "chmod +x" applied after being
    written. Note that for windows, any execute changes using os.chmod are
    no-ops per the python docs.
    """
    ensure_dir(location)
    zipfp = open(filename, "rb")
    try:
        zip = zipfile.ZipFile(zipfp, allowZip64=True)
        leading = has_leading_dir(zip.namelist()) and flatten
        for info in zip.infolist():
            name = info.filename
            fn = name
            if leading:
                fn = split_leading_dir(name)[1]
            fn = os.path.join(location, fn)
            dir = os.path.dirname(fn)
            if not is_within_directory(location, fn):
                message = (
                    "The zip file ({}) has a file ({}) trying to install "
                    "outside target directory ({})"
                )
                raise InstallationError(message.format(filename, fn, location))
            if fn.endswith("/") or fn.endswith("\\"):
                # A directory
                ensure_dir(fn)
            else:
                ensure_dir(dir)
                # Don't use read() to avoid allocating an arbitrarily large
                # chunk of memory for the file's content
                fp = zip.open(name)
                try:
                    with open(fn, "wb") as destfp:
                        shutil.copyfileobj(fp, destfp)
                finally:
                    fp.close()
                    if zip_item_is_executable(info):
                        set_extracted_file_to_default_mode_plus_executable(fn)
    finally:
        zipfp.close()


def untar_file(filename: str, location: str) -> None:
    """
    Untar the file (with path `filename`) to the destination `location`.
    All files are written based on system defaults and umask (i.e. permissions
    are not preserved), except that regular file members with any execute
    permissions (user, group, or world) have "chmod +x" applied after being
    written.  Note that for windows, any execute changes using os.chmod are
    no-ops per the python docs.
    """
    ensure_dir(location)
    if filename.lower().endswith(".gz") or filename.lower().endswith(".tgz"):
        mode = "r:gz"
    elif filename.lower().endswith(BZ2_EXTENSIONS):
        mode = "r:bz2"
    elif filename.lower().endswith(XZ_EXTENSIONS):
        mode = "r:xz"
    elif filename.lower().endswith(".tar"):
        mode = "r"
    else:
        logger.warning(
            "Cannot determine compression type for file %s",
            filename,
        )
        mode = "r:*"
    tar = tarfile.open(filename, mode, encoding="utf-8")
    try:
        leading = has_leading_dir([member.name for member in tar.getmembers()])
        for member in tar.getmembers():
            fn = member.name
            if leading:
                fn = split_leading_dir(fn)[1]
            path = os.path.join(location, fn)
            if not is_within_directory(location, path):
                message = (
                    "The tar file ({}) has a file ({}) trying to install "
                    "outside target directory ({})"
                )
                raise InstallationError(message.format(filename, path, location))
            if member.isdir():
                ensure_dir(path)
            elif member.issym():
                try:
                    # https://github.com/python/typeshed/issues/2673
                    tar._extract_member(member, path)  # type: ignore
                except Exception as exc:
                    # Some corrupt tar files seem to produce this
                    # (specifically bad symlinks)
                    logger.warning(
                        "In the tar file %s the member %s is invalid: %s",
                        filename,
                        member.name,
                        exc,
                    )
                    continue
            else:
                try:
                    fp = tar.extractfile(member)
                except (KeyError, AttributeError) as exc:
                    # Some corrupt tar files seem to produce this
                    # (specifically bad symlinks)
                    logger.warning(
                        "In the tar file %s the member %s is invalid: %s",
                        filename,
                        member.name,
                        exc,
                    )
                    continue
                ensure_dir(os.path.dirname(path))
                assert fp is not None
                with open(path, "wb") as destfp:
                    shutil.copyfileobj(fp, destfp)
                fp.close()
                # Update the timestamp (useful for cython compiled files)
                tar.utime(member, path)
                # member have any execute permissions for user/group/world?
                if member.mode & 0o111:
                    set_extracted_file_to_default_mode_plus_executable(path)
    finally:
        tar.close()


def unpack_file(
    filename: str,
    location: str,
    content_type: Optional[str] = None,
) -> None:
    filename = os.path.realpath(filename)
    if (
        content_type == "application/zip"
        or filename.lower().endswith(ZIP_EXTENSIONS)
        or zipfile.is_zipfile(filename)
    ):
        unzip_file(filename, location, flatten=not filename.endswith(".whl"))
    elif (
        content_type == "application/x-gzip"
        or tarfile.is_tarfile(filename)
        or filename.lower().endswith(TAR_EXTENSIONS + BZ2_EXTENSIONS + XZ_EXTENSIONS)
    ):
        untar_file(filename, location)
    else:
        # FIXME: handle?
        # FIXME: magic signatures?
        logger.critical(
            "Cannot unpack file %s (downloaded from %s, content-type: %s); "
            "cannot detect archive format",
            filename,
            location,
            content_type,
        )
        raise InstallationError(f"Cannot determine archive format of {location}")
site-packages/pip/_internal/utils/egg_link.py000064400000004233150564745010015331 0ustar00# The following comment should be removed at some point in the future.
# mypy: strict-optional=False

import os
import re
import sys
from typing import Optional

from pip._internal.locations import site_packages, user_site
from pip._internal.utils.virtualenv import (
    running_under_virtualenv,
    virtualenv_no_global,
)

__all__ = [
    "egg_link_path_from_sys_path",
    "egg_link_path_from_location",
]


def _egg_link_name(raw_name: str) -> str:
    """
    Convert a Name metadata value to a .egg-link name, by applying
    the same substitution as pkg_resources's safe_name function.
    Note: we cannot use canonicalize_name because it has a different logic.
    """
    return re.sub("[^A-Za-z0-9.]+", "-", raw_name) + ".egg-link"


def egg_link_path_from_sys_path(raw_name: str) -> Optional[str]:
    """
    Look for a .egg-link file for project name, by walking sys.path.
    """
    egg_link_name = _egg_link_name(raw_name)
    for path_item in sys.path:
        egg_link = os.path.join(path_item, egg_link_name)
        if os.path.isfile(egg_link):
            return egg_link
    return None


def egg_link_path_from_location(raw_name: str) -> Optional[str]:
    """
    Return the path for the .egg-link file if it exists, otherwise, None.

    There's 3 scenarios:
    1) not in a virtualenv
       try to find in site.USER_SITE, then site_packages
    2) in a no-global virtualenv
       try to find in site_packages
    3) in a yes-global virtualenv
       try to find in site_packages, then site.USER_SITE
       (don't look in global location)

    For #1 and #3, there could be odd cases, where there's an egg-link in 2
    locations.

    This method will just return the first one found.
    """
    sites = []
    if running_under_virtualenv():
        sites.append(site_packages)
        if not virtualenv_no_global() and user_site:
            sites.append(user_site)
    else:
        if user_site:
            sites.append(user_site)
        sites.append(site_packages)

    egg_link_name = _egg_link_name(raw_name)
    for site in sites:
        egglink = os.path.join(site, egg_link_name)
        if os.path.isfile(egglink):
            return egglink
    return None
site-packages/pip/_internal/utils/direct_url_helpers.py000064400000006206150564745010017432 0ustar00from typing import Optional

from pip._internal.models.direct_url import ArchiveInfo, DirectUrl, DirInfo, VcsInfo
from pip._internal.models.link import Link
from pip._internal.utils.urls import path_to_url
from pip._internal.vcs import vcs


def direct_url_as_pep440_direct_reference(direct_url: DirectUrl, name: str) -> str:
    """Convert a DirectUrl to a pip requirement string."""
    direct_url.validate()  # if invalid, this is a pip bug
    requirement = name + " @ "
    fragments = []
    if isinstance(direct_url.info, VcsInfo):
        requirement += "{}+{}@{}".format(
            direct_url.info.vcs, direct_url.url, direct_url.info.commit_id
        )
    elif isinstance(direct_url.info, ArchiveInfo):
        requirement += direct_url.url
        if direct_url.info.hash:
            fragments.append(direct_url.info.hash)
    else:
        assert isinstance(direct_url.info, DirInfo)
        requirement += direct_url.url
    if direct_url.subdirectory:
        fragments.append("subdirectory=" + direct_url.subdirectory)
    if fragments:
        requirement += "#" + "&".join(fragments)
    return requirement


def direct_url_for_editable(source_dir: str) -> DirectUrl:
    return DirectUrl(
        url=path_to_url(source_dir),
        info=DirInfo(editable=True),
    )


def direct_url_from_link(
    link: Link, source_dir: Optional[str] = None, link_is_in_wheel_cache: bool = False
) -> DirectUrl:
    if link.is_vcs:
        vcs_backend = vcs.get_backend_for_scheme(link.scheme)
        assert vcs_backend
        url, requested_revision, _ = vcs_backend.get_url_rev_and_auth(
            link.url_without_fragment
        )
        # For VCS links, we need to find out and add commit_id.
        if link_is_in_wheel_cache:
            # If the requested VCS link corresponds to a cached
            # wheel, it means the requested revision was an
            # immutable commit hash, otherwise it would not have
            # been cached. In that case we don't have a source_dir
            # with the VCS checkout.
            assert requested_revision
            commit_id = requested_revision
        else:
            # If the wheel was not in cache, it means we have
            # had to checkout from VCS to build and we have a source_dir
            # which we can inspect to find out the commit id.
            assert source_dir
            commit_id = vcs_backend.get_revision(source_dir)
        return DirectUrl(
            url=url,
            info=VcsInfo(
                vcs=vcs_backend.name,
                commit_id=commit_id,
                requested_revision=requested_revision,
            ),
            subdirectory=link.subdirectory_fragment,
        )
    elif link.is_existing_dir():
        return DirectUrl(
            url=link.url_without_fragment,
            info=DirInfo(),
            subdirectory=link.subdirectory_fragment,
        )
    else:
        hash = None
        hash_name = link.hash_name
        if hash_name:
            hash = f"{hash_name}={link.hash}"
        return DirectUrl(
            url=link.url_without_fragment,
            info=ArchiveInfo(hash=hash),
            subdirectory=link.subdirectory_fragment,
        )
site-packages/pip/_internal/utils/deprecation.py000064400000007053150564745010016052 0ustar00"""
A module that implements tooling to enable easy warnings about deprecations.
"""

import logging
import warnings
from typing import Any, Optional, TextIO, Type, Union

from pip._vendor.packaging.version import parse

from pip import __version__ as current_version  # NOTE: tests patch this name.

DEPRECATION_MSG_PREFIX = "DEPRECATION: "


class PipDeprecationWarning(Warning):
    pass


_original_showwarning: Any = None


# Warnings <-> Logging Integration
def _showwarning(
    message: Union[Warning, str],
    category: Type[Warning],
    filename: str,
    lineno: int,
    file: Optional[TextIO] = None,
    line: Optional[str] = None,
) -> None:
    if file is not None:
        if _original_showwarning is not None:
            _original_showwarning(message, category, filename, lineno, file, line)
    elif issubclass(category, PipDeprecationWarning):
        # We use a specially named logger which will handle all of the
        # deprecation messages for pip.
        logger = logging.getLogger("pip._internal.deprecations")
        logger.warning(message)
    else:
        _original_showwarning(message, category, filename, lineno, file, line)


def install_warning_logger() -> None:
    # Enable our Deprecation Warnings
    warnings.simplefilter("default", PipDeprecationWarning, append=True)

    global _original_showwarning

    if _original_showwarning is None:
        _original_showwarning = warnings.showwarning
        warnings.showwarning = _showwarning


def deprecated(
    *,
    reason: str,
    replacement: Optional[str],
    gone_in: Optional[str],
    feature_flag: Optional[str] = None,
    issue: Optional[int] = None,
) -> None:
    """Helper to deprecate existing functionality.

    reason:
        Textual reason shown to the user about why this functionality has
        been deprecated. Should be a complete sentence.
    replacement:
        Textual suggestion shown to the user about what alternative
        functionality they can use.
    gone_in:
        The version of pip does this functionality should get removed in.
        Raises an error if pip's current version is greater than or equal to
        this.
    feature_flag:
        Command-line flag of the form --use-feature={feature_flag} for testing
        upcoming functionality.
    issue:
        Issue number on the tracker that would serve as a useful place for
        users to find related discussion and provide feedback.
    """

    # Determine whether or not the feature is already gone in this version.
    is_gone = gone_in is not None and parse(current_version) >= parse(gone_in)

    message_parts = [
        (reason, f"{DEPRECATION_MSG_PREFIX}{{}}"),
        (
            gone_in,
            "pip {} will enforce this behaviour change."
            if not is_gone
            else "Since pip {}, this is no longer supported.",
        ),
        (
            replacement,
            "A possible replacement is {}.",
        ),
        (
            feature_flag,
            "You can use the flag --use-feature={} to test the upcoming behaviour."
            if not is_gone
            else None,
        ),
        (
            issue,
            "Discussion can be found at https://github.com/pypa/pip/issues/{}",
        ),
    ]

    message = " ".join(
        format_str.format(value)
        for value, format_str in message_parts
        if format_str is not None and value is not None
    )

    # Raise as an error if this behaviour is deprecated.
    if is_gone:
        raise PipDeprecationWarning(message)

    warnings.warn(message, category=PipDeprecationWarning, stacklevel=2)
site-packages/pip/_internal/utils/parallel.py000064400000006174150564745010015354 0ustar00"""Convenient parallelization of higher order functions.

This module provides two helper functions, with appropriate fallbacks on
Python 2 and on systems lacking support for synchronization mechanisms:

- map_multiprocess
- map_multithread

These helpers work like Python 3's map, with two differences:

- They don't guarantee the order of processing of
  the elements of the iterable.
- The underlying process/thread pools chop the iterable into
  a number of chunks, so that for very long iterables using
  a large value for chunksize can make the job complete much faster
  than using the default value of 1.
"""

__all__ = ["map_multiprocess", "map_multithread"]

from contextlib import contextmanager
from multiprocessing import Pool as ProcessPool
from multiprocessing import pool
from multiprocessing.dummy import Pool as ThreadPool
from typing import Callable, Iterable, Iterator, TypeVar, Union

from pip._vendor.requests.adapters import DEFAULT_POOLSIZE

Pool = Union[pool.Pool, pool.ThreadPool]
S = TypeVar("S")
T = TypeVar("T")

# On platforms without sem_open, multiprocessing[.dummy] Pool
# cannot be created.
try:
    import multiprocessing.synchronize  # noqa
except ImportError:
    LACK_SEM_OPEN = True
else:
    LACK_SEM_OPEN = False

# Incredibly large timeout to work around bpo-8296 on Python 2.
TIMEOUT = 2000000


@contextmanager
def closing(pool: Pool) -> Iterator[Pool]:
    """Return a context manager making sure the pool closes properly."""
    try:
        yield pool
    finally:
        # For Pool.imap*, close and join are needed
        # for the returned iterator to begin yielding.
        pool.close()
        pool.join()
        pool.terminate()


def _map_fallback(
    func: Callable[[S], T], iterable: Iterable[S], chunksize: int = 1
) -> Iterator[T]:
    """Make an iterator applying func to each element in iterable.

    This function is the sequential fallback either on Python 2
    where Pool.imap* doesn't react to KeyboardInterrupt
    or when sem_open is unavailable.
    """
    return map(func, iterable)


def _map_multiprocess(
    func: Callable[[S], T], iterable: Iterable[S], chunksize: int = 1
) -> Iterator[T]:
    """Chop iterable into chunks and submit them to a process pool.

    For very long iterables using a large value for chunksize can make
    the job complete much faster than using the default value of 1.

    Return an unordered iterator of the results.
    """
    with closing(ProcessPool()) as pool:
        return pool.imap_unordered(func, iterable, chunksize)


def _map_multithread(
    func: Callable[[S], T], iterable: Iterable[S], chunksize: int = 1
) -> Iterator[T]:
    """Chop iterable into chunks and submit them to a thread pool.

    For very long iterables using a large value for chunksize can make
    the job complete much faster than using the default value of 1.

    Return an unordered iterator of the results.
    """
    with closing(ThreadPool(DEFAULT_POOLSIZE)) as pool:
        return pool.imap_unordered(func, iterable, chunksize)


if LACK_SEM_OPEN:
    map_multiprocess = map_multithread = _map_fallback
else:
    map_multiprocess = _map_multiprocess
    map_multithread = _map_multithread
site-packages/pip/_internal/utils/models.py000064400000002251150564745010015033 0ustar00"""Utilities for defining models
"""

import operator
from typing import Any, Callable, Type


class KeyBasedCompareMixin:
    """Provides comparison capabilities that is based on a key"""

    __slots__ = ["_compare_key", "_defining_class"]

    def __init__(self, key: Any, defining_class: Type["KeyBasedCompareMixin"]) -> None:
        self._compare_key = key
        self._defining_class = defining_class

    def __hash__(self) -> int:
        return hash(self._compare_key)

    def __lt__(self, other: Any) -> bool:
        return self._compare(other, operator.__lt__)

    def __le__(self, other: Any) -> bool:
        return self._compare(other, operator.__le__)

    def __gt__(self, other: Any) -> bool:
        return self._compare(other, operator.__gt__)

    def __ge__(self, other: Any) -> bool:
        return self._compare(other, operator.__ge__)

    def __eq__(self, other: Any) -> bool:
        return self._compare(other, operator.__eq__)

    def _compare(self, other: Any, method: Callable[[Any, Any], bool]) -> bool:
        if not isinstance(other, self._defining_class):
            return NotImplemented

        return method(self._compare_key, other._compare_key)
site-packages/pip/_internal/utils/virtualenv.py000064400000006603150564745010015754 0ustar00import logging
import os
import re
import site
import sys
from typing import List, Optional

logger = logging.getLogger(__name__)
_INCLUDE_SYSTEM_SITE_PACKAGES_REGEX = re.compile(
    r"include-system-site-packages\s*=\s*(?P<value>true|false)"
)


def _running_under_venv() -> bool:
    """Checks if sys.base_prefix and sys.prefix match.

    This handles PEP 405 compliant virtual environments.
    """
    return sys.prefix != getattr(sys, "base_prefix", sys.prefix)


def _running_under_regular_virtualenv() -> bool:
    """Checks if sys.real_prefix is set.

    This handles virtual environments created with pypa's virtualenv.
    """
    # pypa/virtualenv case
    return hasattr(sys, "real_prefix")


def running_under_virtualenv() -> bool:
    """Return True if we're running inside a virtualenv, False otherwise."""
    return _running_under_venv() or _running_under_regular_virtualenv()


def _get_pyvenv_cfg_lines() -> Optional[List[str]]:
    """Reads {sys.prefix}/pyvenv.cfg and returns its contents as list of lines

    Returns None, if it could not read/access the file.
    """
    pyvenv_cfg_file = os.path.join(sys.prefix, "pyvenv.cfg")
    try:
        # Although PEP 405 does not specify, the built-in venv module always
        # writes with UTF-8. (pypa/pip#8717)
        with open(pyvenv_cfg_file, encoding="utf-8") as f:
            return f.read().splitlines()  # avoids trailing newlines
    except OSError:
        return None


def _no_global_under_venv() -> bool:
    """Check `{sys.prefix}/pyvenv.cfg` for system site-packages inclusion

    PEP 405 specifies that when system site-packages are not supposed to be
    visible from a virtual environment, `pyvenv.cfg` must contain the following
    line:

        include-system-site-packages = false

    Additionally, log a warning if accessing the file fails.
    """
    cfg_lines = _get_pyvenv_cfg_lines()
    if cfg_lines is None:
        # We're not in a "sane" venv, so assume there is no system
        # site-packages access (since that's PEP 405's default state).
        logger.warning(
            "Could not access 'pyvenv.cfg' despite a virtual environment "
            "being active. Assuming global site-packages is not accessible "
            "in this environment."
        )
        return True

    for line in cfg_lines:
        match = _INCLUDE_SYSTEM_SITE_PACKAGES_REGEX.match(line)
        if match is not None and match.group("value") == "false":
            return True
    return False


def _no_global_under_regular_virtualenv() -> bool:
    """Check if "no-global-site-packages.txt" exists beside site.py

    This mirrors logic in pypa/virtualenv for determining whether system
    site-packages are visible in the virtual environment.
    """
    site_mod_dir = os.path.dirname(os.path.abspath(site.__file__))
    no_global_site_packages_file = os.path.join(
        site_mod_dir,
        "no-global-site-packages.txt",
    )
    return os.path.exists(no_global_site_packages_file)


def virtualenv_no_global() -> bool:
    """Returns a boolean, whether running in venv with no system site-packages."""
    # PEP 405 compliance needs to be checked first since virtualenv >=20 would
    # return True for both checks, but is only able to use the PEP 405 config.
    if _running_under_venv():
        return _no_global_under_venv()

    if _running_under_regular_virtualenv():
        return _no_global_under_regular_virtualenv()

    return False
site-packages/pip/_internal/utils/appdirs.py000064400000003201150564745010015206 0ustar00"""
This code wraps the vendored appdirs module to so the return values are
compatible for the current pip code base.

The intention is to rewrite current usages gradually, keeping the tests pass,
and eventually drop this after all usages are changed.
"""

import os
import sys
from typing import List

from pip._vendor import platformdirs as _appdirs


def user_cache_dir(appname: str) -> str:
    return _appdirs.user_cache_dir(appname, appauthor=False)


def _macos_user_config_dir(appname: str, roaming: bool = True) -> str:
    # Use ~/Application Support/pip, if the directory exists.
    path = _appdirs.user_data_dir(appname, appauthor=False, roaming=roaming)
    if os.path.isdir(path):
        return path

    # Use a Linux-like ~/.config/pip, by default.
    linux_like_path = "~/.config/"
    if appname:
        linux_like_path = os.path.join(linux_like_path, appname)

    return os.path.expanduser(linux_like_path)


def user_config_dir(appname: str, roaming: bool = True) -> str:
    if sys.platform == "darwin":
        return _macos_user_config_dir(appname, roaming)

    return _appdirs.user_config_dir(appname, appauthor=False, roaming=roaming)


# for the discussion regarding site_config_dir locations
# see <https://github.com/pypa/pip/issues/1733>
def site_config_dirs(appname: str) -> List[str]:
    if sys.platform == "darwin":
        return [_appdirs.site_data_dir(appname, appauthor=False, multipath=True)]

    dirval = _appdirs.site_config_dir(appname, appauthor=False, multipath=True)
    if sys.platform == "win32":
        return [dirval]

    # Unix-y system. Look in /etc as well.
    return dirval.split(os.pathsep) + ["/etc"]
site-packages/pip/_internal/build_env.py000064400000023336150564745010014366 0ustar00"""Build Environment used for isolation during sdist building
"""

import contextlib
import logging
import os
import pathlib
import sys
import textwrap
import zipfile
from collections import OrderedDict
from sysconfig import get_paths
from types import TracebackType
from typing import TYPE_CHECKING, Iterable, Iterator, List, Optional, Set, Tuple, Type

from pip._vendor.certifi import where
from pip._vendor.packaging.requirements import Requirement
from pip._vendor.packaging.version import Version

from pip import __file__ as pip_location
from pip._internal.cli.spinners import open_spinner
from pip._internal.locations import get_platlib, get_prefixed_libs, get_purelib
from pip._internal.metadata import get_environment
from pip._internal.utils.subprocess import call_subprocess
from pip._internal.utils.temp_dir import TempDirectory, tempdir_kinds

if TYPE_CHECKING:
    from pip._internal.index.package_finder import PackageFinder

logger = logging.getLogger(__name__)


class _Prefix:
    def __init__(self, path: str) -> None:
        self.path = path
        self.setup = False
        self.bin_dir = get_paths(
            "nt" if os.name == "nt" else "posix_prefix",
            vars={"base": path, "platbase": path},
        )["scripts"]
        self.lib_dirs = get_prefixed_libs(path)


@contextlib.contextmanager
def _create_standalone_pip() -> Iterator[str]:
    """Create a "standalone pip" zip file.

    The zip file's content is identical to the currently-running pip.
    It will be used to install requirements into the build environment.
    """
    source = pathlib.Path(pip_location).resolve().parent

    # Return the current instance if `source` is not a directory. We can't build
    # a zip from this, and it likely means the instance is already standalone.
    if not source.is_dir():
        yield str(source)
        return

    with TempDirectory(kind="standalone-pip") as tmp_dir:
        pip_zip = os.path.join(tmp_dir.path, "__env_pip__.zip")
        kwargs = {}
        if sys.version_info >= (3, 8):
            kwargs["strict_timestamps"] = False
        with zipfile.ZipFile(pip_zip, "w", **kwargs) as zf:
            for child in source.rglob("*"):
                zf.write(child, child.relative_to(source.parent).as_posix())
        yield os.path.join(pip_zip, "pip")


class BuildEnvironment:
    """Creates and manages an isolated environment to install build deps"""

    def __init__(self) -> None:
        temp_dir = TempDirectory(kind=tempdir_kinds.BUILD_ENV, globally_managed=True)

        self._prefixes = OrderedDict(
            (name, _Prefix(os.path.join(temp_dir.path, name)))
            for name in ("normal", "overlay")
        )

        self._bin_dirs: List[str] = []
        self._lib_dirs: List[str] = []
        for prefix in reversed(list(self._prefixes.values())):
            self._bin_dirs.append(prefix.bin_dir)
            self._lib_dirs.extend(prefix.lib_dirs)

        # Customize site to:
        # - ensure .pth files are honored
        # - prevent access to system site packages
        system_sites = {
            os.path.normcase(site) for site in (get_purelib(), get_platlib())
        }
        self._site_dir = os.path.join(temp_dir.path, "site")
        if not os.path.exists(self._site_dir):
            os.mkdir(self._site_dir)
        with open(os.path.join(self._site_dir, "sitecustomize.py"), "w") as fp:
            fp.write(
                textwrap.dedent(
                    """
                import os, site, sys

                # First, drop system-sites related paths.
                original_sys_path = sys.path[:]
                known_paths = set()
                for path in {system_sites!r}:
                    site.addsitedir(path, known_paths=known_paths)
                system_paths = set(
                    os.path.normcase(path)
                    for path in sys.path[len(original_sys_path):]
                )
                original_sys_path = [
                    path for path in original_sys_path
                    if os.path.normcase(path) not in system_paths
                ]
                sys.path = original_sys_path

                # Second, add lib directories.
                # ensuring .pth file are processed.
                for path in {lib_dirs!r}:
                    assert not path in sys.path
                    site.addsitedir(path)
                """
                ).format(system_sites=system_sites, lib_dirs=self._lib_dirs)
            )

    def __enter__(self) -> None:
        self._save_env = {
            name: os.environ.get(name, None)
            for name in ("PATH", "PYTHONNOUSERSITE", "PYTHONPATH")
        }

        path = self._bin_dirs[:]
        old_path = self._save_env["PATH"]
        if old_path:
            path.extend(old_path.split(os.pathsep))

        pythonpath = [self._site_dir]

        os.environ.update(
            {
                "PATH": os.pathsep.join(path),
                "PYTHONNOUSERSITE": "1",
                "PYTHONPATH": os.pathsep.join(pythonpath),
            }
        )

    def __exit__(
        self,
        exc_type: Optional[Type[BaseException]],
        exc_val: Optional[BaseException],
        exc_tb: Optional[TracebackType],
    ) -> None:
        for varname, old_value in self._save_env.items():
            if old_value is None:
                os.environ.pop(varname, None)
            else:
                os.environ[varname] = old_value

    def check_requirements(
        self, reqs: Iterable[str]
    ) -> Tuple[Set[Tuple[str, str]], Set[str]]:
        """Return 2 sets:
        - conflicting requirements: set of (installed, wanted) reqs tuples
        - missing requirements: set of reqs
        """
        missing = set()
        conflicting = set()
        if reqs:
            env = get_environment(self._lib_dirs)
            for req_str in reqs:
                req = Requirement(req_str)
                dist = env.get_distribution(req.name)
                if not dist:
                    missing.add(req_str)
                    continue
                if isinstance(dist.version, Version):
                    installed_req_str = f"{req.name}=={dist.version}"
                else:
                    installed_req_str = f"{req.name}==={dist.version}"
                if dist.version not in req.specifier:
                    conflicting.add((installed_req_str, req_str))
                # FIXME: Consider direct URL?
        return conflicting, missing

    def install_requirements(
        self,
        finder: "PackageFinder",
        requirements: Iterable[str],
        prefix_as_string: str,
        message: str,
    ) -> None:
        prefix = self._prefixes[prefix_as_string]
        assert not prefix.setup
        prefix.setup = True
        if not requirements:
            return
        with contextlib.ExitStack() as ctx:
            # TODO: Remove this block when dropping 3.6 support. Python 3.6
            # lacks importlib.resources and pep517 has issues loading files in
            # a zip, so we fallback to the "old" method by adding the current
            # pip directory to the child process's sys.path.
            if sys.version_info < (3, 7):
                pip_runnable = os.path.dirname(pip_location)
            else:
                pip_runnable = ctx.enter_context(_create_standalone_pip())
            self._install_requirements(
                pip_runnable,
                finder,
                requirements,
                prefix,
                message,
            )

    @staticmethod
    def _install_requirements(
        pip_runnable: str,
        finder: "PackageFinder",
        requirements: Iterable[str],
        prefix: _Prefix,
        message: str,
    ) -> None:
        args: List[str] = [
            sys.executable,
            pip_runnable,
            "install",
            "--ignore-installed",
            "--no-user",
            "--prefix",
            prefix.path,
            "--no-warn-script-location",
        ]
        if logger.getEffectiveLevel() <= logging.DEBUG:
            args.append("-v")
        for format_control in ("no_binary", "only_binary"):
            formats = getattr(finder.format_control, format_control)
            args.extend(
                (
                    "--" + format_control.replace("_", "-"),
                    ",".join(sorted(formats or {":none:"})),
                )
            )

        index_urls = finder.index_urls
        if index_urls:
            args.extend(["-i", index_urls[0]])
            for extra_index in index_urls[1:]:
                args.extend(["--extra-index-url", extra_index])
        else:
            args.append("--no-index")
        for link in finder.find_links:
            args.extend(["--find-links", link])

        for host in finder.trusted_hosts:
            args.extend(["--trusted-host", host])
        if finder.allow_all_prereleases:
            args.append("--pre")
        if finder.prefer_binary:
            args.append("--prefer-binary")
        args.append("--")
        args.extend(requirements)
        extra_environ = {"_PIP_STANDALONE_CERT": where()}
        with open_spinner(message) as spinner:
            call_subprocess(args, spinner=spinner, extra_environ=extra_environ)


class NoOpBuildEnvironment(BuildEnvironment):
    """A no-op drop-in replacement for BuildEnvironment"""

    def __init__(self) -> None:
        pass

    def __enter__(self) -> None:
        pass

    def __exit__(
        self,
        exc_type: Optional[Type[BaseException]],
        exc_val: Optional[BaseException],
        exc_tb: Optional[TracebackType],
    ) -> None:
        pass

    def cleanup(self) -> None:
        pass

    def install_requirements(
        self,
        finder: "PackageFinder",
        requirements: Iterable[str],
        prefix_as_string: str,
        message: str,
    ) -> None:
        raise NotImplementedError()
site-packages/pip/__init__.py000064400000000545150564745010012200 0ustar00from typing import List, Optional

__version__ = "21.3.1"


def main(args: Optional[List[str]] = None) -> int:
    """This is an internal API only meant for use by pip's own console scripts.

    For additional details, see https://github.com/pypa/pip/issues/7498.
    """
    from pip._internal.utils.entrypoints import _wrapper

    return _wrapper(args)
site-packages/pip/__main__.py000064400000002256150564745010012162 0ustar00import os
import sys
import warnings

# Remove '' and current working directory from the first entry
# of sys.path, if present to avoid using current directory
# in pip commands check, freeze, install, list and show,
# when invoked as python -m pip <command>
if sys.path[0] in ("", os.getcwd()):
    sys.path.pop(0)

# If we are running from a wheel, add the wheel to sys.path
# This allows the usage python pip-*.whl/pip install pip-*.whl
if __package__ == "":
    # __file__ is pip-*.whl/pip/__main__.py
    # first dirname call strips of '/__main__.py', second strips off '/pip'
    # Resulting path is the name of the wheel itself
    # Add that to sys.path so we can import pip
    path = os.path.dirname(os.path.dirname(__file__))
    sys.path.insert(0, path)

if __name__ == "__main__":
    # Work around the error reported in #9540, pending a proper fix.
    # Note: It is essential the warning filter is set *before* importing
    #       pip, as the deprecation happens at import time, not runtime.
    warnings.filterwarnings(
        "ignore", category=DeprecationWarning, module=".*packaging\\.version"
    )
    from pip._internal.cli.main import main as _main

    sys.exit(_main())
site-packages/pip/py.typed000064400000000436150564745010011565 0ustar00pip is a command line program. While it is implemented in Python, and so is
available for import, you must not use pip's internal APIs in this way. Typing
information is provided as a convenience only and is not a guarantee. Expect
unannounced changes to the API and types in releases.
opt/alt/python39/bin/python3.9000075500000037130150565763140012114 0ustar00ELF>`@7@8@@@@hh���hh��   8-8=8=��P-P=P=``���DDP�td   44Q�tdR�td8-8=8=��/lib64/ld-linux-x86-64.so.2GNU�x�o�iJ٘��RAF�GNU�Q! e�
BE��j	Cֻ�|:��K����2b���qX
��$ � @ �"�@�@@� @� 
Pe�
`/@�
�libpython3.9.so.1.0__gmon_start___ITM_deregisterTMCloneTable_ITM_registerTMCloneTablelibcrypt.so.1libdl.so.2libutil.so.1libm.so.6libpthread.so.0libc.so.6__cxa_finalize__libc_start_mainPy_BytesMain__libc_csu_fini_edata__data_start_IO_stdin_used__libc_csu_init_end__bss_startGLIBC_2.2.5/opt/alt/python39/lib64:/opt/alt/sqlite/usr/lib64�ui	&8=@@=H=H=�?�?�?�?�?�?�?��H��H��/H��t��H����5�/�%�/@�%�/h����%�/h��������f.����1�I��^H��H���PTL�FH�
�H�=����R/�H�=q/H�j/H9�tH�./H��t	�����H�=A/H�5:/H)�H��H��?H��H�H�tH�/H��t��fD�����=�.u+UH�=�.H��tH�=&,�����d�����.]������w������AWI��AVI��AUA��ATL�%�+UH�-�+SL)�H������H��t1��L��L��D��A��H��H9�u�H��[]A\A]A^A_�ff.������H��H���;0�dL�\�LL����zRx��/D$4��0FJw�?;*3$"\��Dt��eF�E�E �E(�H0�H8�G@n8A0A(B BBB���@H=Zhs���2
�8=@=���o�P
d�?08x�	���o���oX���o�o4���oP=6FGA$3a1�GA$3a1��GA$3p1113P�GA*GA$annobin gcc 8.5.0 20210514GA$plugin name: gcc-annobinGA$running gcc 8.5.0 20210514GA*GA*GA!
GA*FORTIFYGA+GLIBCXX_ASSERTIONSGA*GOW*GA*cf_protectionGA+omit_frame_pointerGA+stack_clashGA!stack_realign
GA*FORTIFYP�GA+GLIBCXX_ASSERTIONSpython3.9-3.9.23-1.el8.x86_64.debugk�k�7zXZ�ִF!t/���]?�E�h=��ڊ�2N��+�(�Y��4�5��[�2
֬y��}������7���OYIag�}a��].:����7�ft��xS�
�mt9Y?~���K�k�I��墲����_�R�P�j����"@�83�gtJR��j�D�
���^{I{�B��/4*A�z��;�$	�z݈�
���]��)W�����K=U擼4�52�}�|;=C��i�D1b����fNܩz�}���k:r�)��c�#*��\:vr���Χ�L��ʝƿ�N"ҋ3?8z�����c��<���:�`�� :��;��)'2�Z�^=-^=���vE�?K\��$Q8B��|���N!𽲓cwe�;h�D�
�|�5+��wx�i��� K:��9�ћ�1T>3�q�^l'F@|��ih�2l�465��N��V���:�A䉵9D�	�]�\�,R��
�|�E�c3Z1��"G�|_�E"wAAқ��(�˶�~cp�ʣ���s�\j(�R�4���`\�,Y�ȫ^�87�
�����K���Y��[��#�u�+)�� ��nJ��]qE�����-
���z��3�w}�#?MD�r�S��*���W�������&��⾹��R'��v�hä�Ҹ1����:�⥖R�"�1Nkv)E��U������VI�h��2nH�v%Ƙ�xlt���s��˪%�퇟�5��ś���_p�*-�=�E�V��݋�џmF���Tg�^ F�w��N�e�SO–:a
�O���p�Yy~0��BKM��U�/ױ�g�YZ.shstrtab.interp.note.gnu.build-id.note.ABI-tag.gnu.hash.dynsym.dynstr.gnu.version.gnu.version_r.rela.dyn.rela.plt.init.text.fini.rodata.eh_frame_hdr.eh_frame.init_array.fini_array.data.rel.ro.dynamic.got.data.bss.gnu.build.attributes.gnu_debuglink.gnu_debugdata����$&�� 4���oH>PP�F��dN���o44 [���oXX jxx�tB880~y  0�PPu���
�  �  4�8 8 ��8=8-�@=@-�H=H-�P=P-`��?�/P�@0�@0�`0lp2(�2X�5"